Poster
in
Workshop: Advances in Financial AI: Opportunities, Innovations, and Responsible AI
ExpProof : Operationalizing Explanations for Confidential Models with ZKPs
Chhavi Yadav · Evan Laufer · Dan Boneh · Kamalika Chaudhuri
In principle, explanations are intended as a way to increase trust in machine learn-ing models and are often obligated by regulations. However, many circumstanceswhere these are demanded are adversarial in nature, meaning the involved partieshave misaligned interests and are incentivized to manipulate explanations for theirpurpose. As a result, explainability methods fail to be operational in such settingsdespite the demand Bordt et al. (2022). In this paper, we take a step towards op-erationalizing explanations in adversarial scenarios with Zero-Knowledge Proofs(ZKPs), a cryptographic primitive. Specifically we explore ZKP-amenable ver-sions of the popular explainability algorithm LIME and evaluate their performanceon Neural Networks and Random Forests. Our code is publicly available at :.