Skip to yearly menu bar Skip to main content


Poster

Increasing the Cost of Model Extraction with Calibrated Proof of Work

Adam Dziedzic · Muhammad Ahmad Kaleem · Yu Shen Lu · Nicolas Papernot

Keywords: [ Model stealing ] [ deep learning ] [ model extraction ] [ adversarial machine learning ]


Abstract:

In model extraction attacks, adversaries can steal a machine learning model exposed via a public API by repeatedly querying it and adjusting their own model based on obtained predictions. To prevent model stealing, existing defenses focus on detecting malicious queries, truncating, or distorting outputs, thus necessarily introducing a tradeoff between robustness and model utility for legitimate users. Instead, we propose to impede model extraction by requiring users to complete a proof-of-work before they can read the model's predictions. This deters attackers by greatly increasing (even up to 100x) the computational effort needed to leverage query access for model extraction. Since we calibrate the effort required to complete the proof-of-work to each query, this only introduces a slight overhead for regular users (up to 2x). To achieve this, our calibration applies tools from differential privacy to measure the information revealed by a query. Our method requires no modification of the victim model and can be applied by machine learning practitioners to guard their publicly exposed models against being easily stolen.

Chat is not available.