In these years, we have seen principles and guidance relating to accountable and ethical use of artificial intelligence (AI) spring up around the globe. Specifically, Data Privacy, Accountability, Interpretability, {\bf R}obustness, and Reasoning have been broadly recognized as fundamental principles of using machine learning (ML) technologies on decision-critical and/or privacy-sensitive applications. On the other hand, in tremendous real-world applications, data itself can be well represented as various structured formalisms, such as graph-structured data (e.g., networks), grid-structured data (e.g., images), sequential data (e.g., text), etc. By exploiting the inherently structured knowledge, one can design plausible approaches to identify and use more relevant variables to make reliable decisions, thereby facilitating real-world deployments.In this workshop, we will examine the research progress towards accountable and ethical use of AI from diverse research communities, such as the ML community, security \& privacy community, and more. Specifically, we will focus on the limitations of existing notions on Privacy, Accountability, Interpretability, Robustness, and Reasoning. We aim to bring together researchers from various areas (e.g., ML, security \& privacy, computer vision, and healthcare) to facilitate discussions including related challenges, definitions, formalisms, and evaluation protocols regarding the accountable and ethical use of ML technologies in high-stake applications with structured data. In particular, we will discuss the interplay among the fundamental principles from theory to applications. We aim to identify new areas that call for additional research efforts. Additionally, we will seek possible solutions and associated interpretations from the notion of causation, which is an inherent property of systems. We hope that the proposed workshop is fruitful in building accountable and ethical use of AI systems in practice.
Fri 9:00 a.m. - 9:05 a.m.
|
Introduction and Opening Remark
SlidesLive Video » |
Hao Wang · Wanyu LIN 🔗 |
Fri 9:05 a.m. - 9:30 a.m.
|
On the Foundations of Causal Artificial Intelligence
(
Invited Talk
)
|
Elias Bareinboim 🔗 |
Fri 9:30 a.m. - 9:35 a.m.
|
Q&A with Elias Bareinboim
(
Q&A
)
|
Elias Bareinboim 🔗 |
Fri 9:35 a.m. - 10:05 a.m.
|
Privacy Meter Project: Towards Auditing Data Privacy and Q&A
(
Invited Talk
)
SlidesLive Video » |
Reza Shokri 🔗 |
Fri 10:05 a.m. - 10:15 a.m.
|
Rethinking Stability for Attribution-based Explanations
(
Oral
)
link »
SlidesLive Video » As attribution-based explanation methods are increasingly used to establish model trustworthiness in high-stakes situations, it is critical to ensure that these explanations are stable, e.g., robust to infinitesimal perturbations to an input. However, previous works have shown that state-of-the-art explanation methods generate unstable explanations. Here, we introduce metrics to quantify the stability of an explanation and show that several popular explanation methods are unstable. In particular, we propose new Relative Stability metrics that measure the change in output explanation with respect to change in input, model representation, or output of the underlying predictor. Finally, our experimental evaluation with three real-world datasets demonstrates interesting insights for seven explanation methods and different stability metrics. |
Chirag Agarwal · Nari Johnson · Martin Pawelczyk · Satyapriya Krishna · Eshika Saxena · Marinka Zitnik · Hima Lakkaraju 🔗 |
Fri 10:15 a.m. - 10:40 a.m.
|
Trustworthy Machine Learning via Logic Reasoning
(
Invited Talk
)
|
Bo Li 🔗 |
Fri 10:40 a.m. - 10:45 a.m.
|
Q&A with Bo Li
(
Q&A
)
|
Bo Li 🔗 |
Fri 10:45 a.m. - 11:10 a.m.
|
Quantifying Privacy Risks of Machine Learning Models
(
Invited Talk
)
|
Yang Zhang 🔗 |
Fri 11:10 a.m. - 11:15 a.m.
|
Q&A with Yang Zhang
(
Q&A
)
|
Yang Zhang 🔗 |
Fri 11:15 a.m. - 11:25 a.m.
|
Invariant Causal Representation Learning for Generalization in Imitation and Reinforcement Learning
(
Oral
)
link »
SlidesLive Video » A fundamental challenge in imitation and reinforcement learning is to learn policies, representations, or dynamics that do not build on spurious correlations and generalize beyond the specific environments that they were trained on. We investigate these generalization problems from a unified view. For this, we propose a general framework to tackle them with theoretical guarantees on both identifiability and generalizability under mild assumptions on environmental changes. By leveraging a diverse set of training environments, we construct a data representation that ignores any spurious features and consistently predicts target variables well across environments. Following this approach, we build invariant predictors in terms of policy, representations, and dynamics. We theoretically show that the resulting policies, representations, and dynamics are able to generalize to unseen environments. Extensive experiments on both synthetic and real-world datasets show that our methods attain improved generalization over a variety of baselines. |
Chaochao Lu · José Miguel Hernández Lobato · Bernhard Schoelkopf 🔗 |
Fri 11:25 a.m. - 1:30 p.m.
|
Poster Session 1 ( Poster Session ) link » | 🔗 |
Fri 1:30 p.m. - 1:55 p.m.
|
Interpretable AI for Medical Imaging
(
Invited Talk
)
|
Lei Xing 🔗 |
Fri 1:55 p.m. - 2:00 p.m.
|
Q&A with Lei Xing
(
Q&A
)
|
Lei Xing 🔗 |
Fri 2:00 p.m. - 2:25 p.m.
|
Learning Structured Dynamics Models for Physical Reasoning and Robot Manipulation
(
Invited Talk
)
|
Jiajun Wu 🔗 |
Fri 2:25 p.m. - 2:30 p.m.
|
Q&A with Jiajun Wu
(
Q&A
)
|
Jiajun Wu 🔗 |
Fri 2:30 p.m. - 2:40 p.m.
|
Maximizing Entropy on Adversarial Examples Can Improve Generalization
(
Oral
)
link »
SlidesLive Video » Supervised learning methods that directly optimize the cross entropy loss on training data often overfit. This overfitting is typically mitigated through regularizing the loss function (e.g., label smoothing) or by minimizing the same loss on new examples (e.g., data augmentation and adversarial training). In this work, we propose a complementary regularization strategy: Maximum Predictive Entropy (MPE) forcing the model to be uncertain on new, algorithmically-generated inputs. Across a range of tasks, we demonstrate that our computationally-efficient method improves test accuracy, and the benefits are complementary to methods such as label smoothing and data augmentation. |
Amrith Setlur · Benjamin Eysenbach 🔗 |
Fri 2:40 p.m. - 3:05 p.m.
|
Adapting Deep Predictors Under Causally Structured Shifts
(
Invited Talk
)
|
Zachary Lipton 🔗 |
Fri 3:05 p.m. - 3:10 p.m.
|
Q&A with Zachary Lipton
(
Q&A
)
|
Zachary Lipton 🔗 |
Fri 3:10 p.m. - 3:35 p.m.
|
Explainable AI in Practice: Challenges and Opportunities
(
Invited Talk
)
|
Himabindu Lakkaraju 🔗 |
Fri 3:35 p.m. - 3:40 p.m.
|
Q&A with Himabindu Lakkaraju
(
Q&A
)
|
Himabindu Lakkaraju 🔗 |
Fri 3:40 p.m. - 3:50 p.m.
|
Node-Level Differentially Private Graph Neural Networks
(
Oral
)
link »
SlidesLive Video » Graph Neural Networks (GNNs) are a popular technique for modelling graph-structured data and computing node-level representations via aggregation of information from the neighborhood of each node. However, this aggregation implies increased risk of revealing sensitive information, as a node can participate in the inference for multiple nodes. This implies that standard privacy preserving machine learning techniques, such as differentially private stochastic gradient descent (DP-SGD) - which are designed for situations where each data point participates in the inference for one point only - either do not apply, or lead to inaccurate solutions. In this work, we formally define the problem of learning GNN parameters with node-level privacy, and provide an algorithmic solution with a strong differential privacy guarantee. We employ a careful sensitivity analysis and provide a non-trivial extension of the privacy-by-amplification technique. An empirical evaluation on standard benchmarks datasets and architectures demonstrates that our method is indeed able to learn accurate privacy-preserving GNNs, while still outperforming standard non-private methods that completely ignore graph information. |
Ameya Daigavane · Gagan Madan · Aditya Sinha · Abhradeep Guha Thakurta · Gaurav Aggarwal · Prateek Jain 🔗 |
Fri 3:50 p.m. - 4:40 p.m.
|
Panel
|
🔗 |
Fri 4:40 p.m. - 6:00 p.m.
|
Poster Session 2 ( Poster Session ) link » | 🔗 |
-
|
REVERSING ADVERSARIAL ATTACKS WITH MULTIPLE SELF SUPERVISED TASKS
(
Poster
)
link »
Deep networks achieve state-of-the-art performance on computer vision tasks, yet they fail under adversarial attacks that are imperceptible to humans. In this paper, we propose a novel defense that can dynamically adapt the input using the intrinsic structure from multiple self-supervised tasks. By simultaneously using many self-supervised tasks, our defense avoids over-fitting the adapted image to one specific self-supervised task and restores more intrinsic structure in the image compared to a single self-supervised task approach. Our approach further improves robustness and clean accuracy significantly compared to the state-of-the-art single task self-supervised defense. Our work is the first to connect multiple self-supervised tasks to robustness, and suggests that we can achieve better robustness with more intrinsic signal from visual data. |
Matthew Lawhon · Chengzhi Mao · Gustave Ducrest · Junfeng Yang 🔗 |
-
|
Global Counterfactual Explanations: Investigations, Implementations and Improvements
(
Poster
)
link »
Counterfactual explanations have been widely studied in explainability, with a range of application dependent methods emerging in fairness, recourse and model understanding. However, the major shortcoming associated with these methods is their inability to provide explanations beyond the local or instance-level. While many works touch upon the notion of a global explanation, typically suggesting to aggregate masses of local explanations in the hope of ascertaining global properties, few provide frameworks that are either reliable or computationally tractable. Meanwhile, practitioners are requesting more efficient and interactive explainability tools. We take this opportunity to investigate existing global counterfactual methods, with a focus on implementing and improving Actionable Recourse Summaries (AReS), the only known global explanation framework for recourse. |
Dan Ley · Saumitra Mishra · Daniele Magazzeni 🔗 |
-
|
Saliency Maps Contain Network "Fingerprints"
(
Poster
)
link »
Explaining deep learning models and their predictions is an open question with many proposed, but difficult to validate, solutions. This difficulty in assessing explanation methods has raised the question on the validity of these methods: What are they showing and what are the factors influencing the explanations? Furthermore, how should one choose which one to use? Here, we explore saliency-type methods, finding that saliency maps contain network “fingerprints”, by which the network which generated the map can be uniquely identified. We test this by creating datasets made up of saliency maps from different “primary” networks, then training “secondary” networks on these saliency-map datasets. We find that secondary networks can learn to identify which primary network a saliency map comes from. Our findings hold across several saliency methods and for both CNN and ResNet "primary" architectures.Our analysis also reveals complex relationships between methods: a set of methods share fingerprints, while some contain unique fingerprints. We discuss a potentially related prior work that may explain some of these relationships; some methods are made up of 'higher order derivatives'.Our simple analytical framework is a first step towards understanding ingredients of and relationships between many saliency methods. |
Amy Widdicombe · Been Kim · Simon Julier 🔗 |
-
|
Geometrically Guided Saliency Maps
(
Poster
)
link »
Interpretability methods for deep neural networks mainly focus on modifying the rules of automatic differentiation or perturbing the input and observing the score drop to determine the most relevant features. Among them, gradient-based attribution methods, such as saliency maps, are arguably the most popular. Still, the produced saliency maps often may lack intelligibility. We address this problem based on recent discoveries in geometric properties of deep neural networks' loss landscape that reveal the existence of a multiplicity of local minima in the vicinity of a trained model's loss surface. We introduce two methods that leverage the geometry of the loss landscape to improve interpretability: 1) "Geometrically Guided Integrated Gradients," applying gradient ascent from each interpolation point of the linear path as a guide. 2) "Geometric Ensemble Gradients" that generates ensemble saliency maps by sampling proximal iso-loss models. Compared to vanilla and integrated gradients, these methods significantly improve saliency maps in quantitative and visual terms. We verify our findings on MNIST and Imagenet datasets across convolutional, ResNet, and Inception V3 architectures. |
Md Mahfuzur Rahman · Noah Lewis · Sergey Plis 🔗 |
-
|
ConceptDistil: Model-Agnostic Distillation of Concept Explanations
(
Poster
)
link »
Concept-based explainability aims to fill the model interpretability gap for non-technical decision-makers. Previous work has focused on providing concepts for specific models (e.g, neural networks) or data types (e.g., images), and by either trying to extract concepts from an already trained network or training self-explainable models through multi-task learning. In this work, we propose ConceptDistil, a method to bring concept explanations to any black-box classifier using knowledge distillation. Our method uses a surrogate neural network that approximates the predictions of a black-box classifier while producing concept explanations. We validate our proposed concept-based knowledge distillation explainer in a real world use-case, showing that it achieves alignment with the black-box classifier while attaining high performance on the explainability task, providing high-level domain explanations. |
João Pedro Sousa · Ricardo Moreira · Vladimir Balayan · Pedro Saleiro · Pedro Bizarro 🔗 |
-
|
Data Poisoning Attacks on Off-Policy Policy Evaluation Algorithms
(
Poster
)
link »
Off-policy Evaluation (OPE) methods are crucial for evaluating policies in high-stakes domains such as healthcare, where exploration is often infeasible or expensive. However, the extent to which such methods can be trusted under adversarial threats to data quality is largely unexplored. In this work, we make the first attempt at investigating the sensitivity of OPE methods to adversarial perturbations to the data.We design a data poisoning attack framework that leverages influence functions to construct perturbations that maximize error in the policy value estimates. Our experimental results show that many OPE methods are highly prone to data poisoning attacks, even for small adversarial perturbations. |
Elita Lobo · Harvineet Singh · Marek Petrik · Cynthia Rudin · Hima Lakkaraju 🔗 |
-
|
Efficient Privacy-Preserving Inference for Convolutional Neural Networks
(
Poster
)
link »
The processing of sensitive user data using deep learning models is an area that has gained recent traction. Existing work has leveraged homomorphic encryption (HE) schemes to enable computation on encrypted data. An early work was CryptoNets, which takes 250 seconds for one MNIST inference. The main limitation of such approaches is that of the expensive FFT-like operations required to perform operations on HE-encrypted ciphertext. Others have proposed the use of model pruning and efficient data representations to reduce the number of HE operations required. We focus on improving upon existing work by proposing changes to the representations of intermediate tensors during CNN inference. We construct and evaluate private CNNs on the MNIST and CIFAR-10 datasets, and achieve over a two-fold reduction in the number of operations used for inferences of the CryptoNets architecture. |
Han Xuanyuan · Francisco Vargas · Stephen Cummins 🔗 |
-
|
Post-hoc Concept Bottleneck Models
(
Poster
)
link »
Concept Bottleneck Models (CBMs) map the inputs onto a concept bottleneck and use the bottleneck to make a prediction. A concept bottleneck enhances interpretability since it can be investigated to understand what the model sees in an input, and which of these concepts are deemed important. However, CBMs are restrictive in practice as they require concept labels during training to learn the bottleneck. Additionally, it is questionable if CBMs can match the accuracy of an unrestricted neural network trained on a given domain, potentially reducing the incentive to deploy them in practice. In this work, we address these two key limitations by introducing Post-hoc Concept Bottleneck models (P-CBMs). We show that we can turn any neural network into a P-CBM, without sacrificing model performance and retaining interpretability benefits. Finally, we show that P-CBMs can provide significant performance gains with model editing without any fine-tuning and needing data from the target domain. |
Mert Yuksekgonul · Maggie Wang · James Y Zou 🔗 |
-
|
CLIP-Dissect: Automatic description of neuron representations in deep vision networks
(
Poster
)
link »
In this paper, we propose CLIP-Dissect, a new technique to automatically describe the function of individual hidden neurons inside vision networks. CLIP-Dissect leverages recent advances in multimodal vision/language models to label internal neurons with open-ended concepts without the need for any labeled data or human examples, which are required for existing tools to succeed. We show that CLIP-Dissect provides more accurate descriptions than existing methods for neurons where the ground-truth is available as well as qualitatively good descriptions for hidden layer neurons. In addition, our method is very flexible: it is model agnostic, can easily handle new concepts and can be extended to take advantage of better multimodal models in the future. Finally CLIP-Dissect is computationally efficient and labels all neurons of a layer in a large vision model in tens of minutes. |
Tuomas Oikarinen · Tsui-Wei Weng 🔗 |
-
|
Robust Randomized Smoothing via Two Cost-Effective Approaches
(
Poster
)
link »
Randomized smoothing has recently attracted attentions in the field of adversarial robustness to provide provable robustness guarantees on smoothed neural network classifiers. However, existing works show that vanilla randomized smoothing usually does not provide good robustness performance and often requires (re)training techniques on the base classifier in order to boost the robustness of the resulting smoothed classifier. In this work, we propose two cost-effective approaches to boost the robustness of randomized smoothing while preserving its standard performance. In the first approach, we propose a new robust training method AdvMacer that combines adversarial training and maximizing robustness certificate for randomized smoothing. We show that AdvMacer can improve the robustness performance of randomized smoothing classifiers compared to SOTA baselines. The second approach introduces a post-processing method named EsbRS which greatly improves the robustness certificate based on model ensembles. We explore different aspects of model ensembles that has not been studied by prior works and propose a mixed design strategy to further improve robustness of the ensemble. |
Linbo Liu · Trong Hoang · Lam Nguyen · Tsui-Wei Weng 🔗 |
-
|
Graphical Clusterability and Local Specialization in Deep Neural Networks
(
Poster
)
link »
The learned weights of deep neural networks have often been considered devoid of scrutable internal structure, and tools for studying them have not traditionally relied on techniques from network science. In this paper, we present methods for studying structure among a network's neurons by clustering them and for quantifying how well this reveals both graphical clusterability and local specialization -- the degree to which the network can be understood as having distinct, highly internally connected subsets of neurons that perform subtasks. We offer a pipeline for this analysis consisting of methods for (1) representing a network as a graph, (2) clustering that graph, and (3) performing statistical analysis to determine how graphically clusterable and (4) functionally specialized the clusters are. We demonstrate that image classification networks up to the ImageNet-scale are often highly clusterable and locally specialized. |
Stephen Casper · Shlomi Hod · Daniel Filan · Cody Wild · Andrew Critch · Stuart Russell 🔗 |
-
|
Sparse Logits Suffice to Fail Knowledge Distillation
(
Poster
)
link »
Knowledge distillation (KD) aims to transfer the power of pre-trained teacher models to (more lightweight) student models. However, KD also poses the risk of intellectual properties (IPs) leakage of teacher models. Even if the teacher model is released as a black box, it can still be cloned through KD by imitating input-output behaviors. To address this unwanted effect of KD, the concept of Nasty Teacher was proposed recently. It is a special network that achieves nearly the same accuracy as a normal one, but significantly degrades the accuracy of student models trying to imitate it. Previous work builds the nasty teacher by retraining a new model and distorting its output distribution from the normal one via an adversarial loss. With this design, the ``nasty" teacher tends to produce sparse and noisy logits. However, it is unclear why the distorted distribution is catastrophic to the student model, as the nasty logits still maintain the correct labels.In this paper, we provide a theoretical analysis of why the sparsity of logits is key to Nasty Teacher. Furthermore, we propose an ideal version of the nasty teacher to prevent imitation through KD, named \textit{Stingy Teacher}. The Stingy Teacher directly manipulates the logits of a standard pre-trained network by maintaining the values for a small subset of classes while zeroing out the rest. Extensive experiments on several datasets demonstrate that stingy teacher is more catastrophic to student models on both standard KD and data-free KD. Code and pretrained models will be released upon acceptance. |
Haoyu Ma · Yifan Huang · Hao Tang · Chenyu You · Deying Kong · Xiaohui Xie 🔗 |
-
|
User-Level Membership Inference Attack against Metric Embedding Learning
(
Poster
)
link »
Membership inference (MI) determines if a sample was part of a victim model training set. Recent development of MI attacks focus on record-level membership inference which limits their application in many real-world scenarios. For example, in the person re-identification task, the attacker (or investigator) is interested in determining if a user's images have been used during training or not. However, the exact training images might not be accessible to the attacker. In this paper, we develop a user-level MI attack where the goal is to find if any sample from the target user has been used during training even when no exact training sample is available to the attacker. We focus on metric embedding learning due to its dominance in person re-identification, where user-level MI attack is more sensible. We conduct an extensive evaluation on several datasets and show that our approach achieves high accuracy on user-level MI task. |
Guoyao Li · Shahbaz Rezaei · Xin Liu 🔗 |
-
|
Towards Differentially Private Query Release for Hierarchical Data
(
Poster
)
link »
While differentially private query release has been well-studied, research in this area is commonly restricted to data that do not exhibit hierarchical structure. However, in many real-world scenarios, individual data points can be grouped together (e.g., people within households, taxi trips per driver, etc.), begging the question---what statistical properties (or queries) are important when considering data of this form? In addition, although synthetic data generation approaches for private query release have grown increasingly popular, it is unclear how one can generate synthetic data at both the group and individual-level while capturing such statistical properties. In light of these challenges, we formalize the problem of hierarchical query release and provide a set of statistical queries that capture relationships between attributes at both the group and individual-level. Furthermore, we propose and implement a novel synthetic data generation algorithm, H-GEM, which outputs hierarchical data subject to differential privacy to answer such statistical queries. Finally, using the American Community Survey, we evaluate H-GEM, establishing a benchmark for future work to measure against |
Terrance Liu · Steven Wu 🔗 |
-
|
Sparse Neural Additive Model: Interpretable Deep Learning with Feature Selection via Group Sparsity
(
Poster
)
link »
Interpretable machine learning has demonstrated impressive performance while preserving explainability. In particular, neural additive models (NAM) offer the interpretability to the black-box deep learning and achieve state-of-the-art accuracy among the large family of generalized additive models. In order to empower NAM with feature selection and improve the generalization, we propose the sparse neural additive models (SNAM) that employ the group sparsity regularization (e.g. Group LASSO), where each feature is learned by a sub-network whose trainable parameters are clustered as a group. We study the theoretical properties for SNAM with novel techniques to tackle the non-parametric truth, thus extending from classical sparse linear models such as the LASSO, which only works on the parametric truth. Specifically, we show that the estimation error of SNAM vanishes asymptotically as $n\to\infty$. We also prove that SNAM, similar to LASSO, can have exact support recovery, i.e. perfect feature selection, with appropriate regularization. Moreover, we show that the SNAM can generalize well and preserve the `identifiability', recovering each feature's effect. We validate our theories via extensive experiments and further testify to the good accuracy and efficiency of SNAM.
|
Shiyun Xu · Zhiqi Bu · Pratik A Chaudhari · Ian Barnett 🔗 |
-
|
Neural Logic Analogy Learning
(
Poster
)
link »
Letter-string analogy is an important analogy learning task which seems to be easy for humans but very challenging for machines.The main idea behind current approaches to solving letter-string analogies is to design heuristic rules for extracting analogy structures and constructing analogy mappings. However, one key problem is that it is difficult to build a comprehensive and exhaustive set of analogy structures which can fully describe the subtlety of analogies. This problem makes current approaches unable to handle complicated letter-string analogy problems.In this paper, we propose Neural lOgic ANalogy learning (Noan), which is a dynamic neural architecture driven by differentiable logic reasoning to solve analogy problems. Each analogy problem is converted into logical expressions consisting of logical variables and basic logical operations (AND, OR, and NOT). More specifically, Noan learns the logical variables as vector embeddings and learns each logical operation as a neural module. In this way, the model builds computational graph integrating neural network with logical reasoning to capture the internal logical structure of the input letter strings. The analogy learning problem then becomes a True/False evaluation problem of the logical expressions. Experiments show that our machine learning-based Noan approach outperforms state-of-the-art approaches on standard letter-string analogy benchmark datasets. |
Yujia Fan · Yongfeng Zhang 🔗 |
-
|
Rethinking Stability for Attribution-based Explanations
(
Poster
)
link »
As attribution-based explanation methods are increasingly used to establish model trustworthiness in high-stakes situations, it is critical to ensure that these explanations are stable, e.g., robust to infinitesimal perturbations to an input. However, previous works have shown that state-of-the-art explanation methods generate unstable explanations. Here, we introduce metrics to quantify the stability of an explanation and show that several popular explanation methods are unstable. In particular, we propose new Relative Stability metrics that measure the change in output explanation with respect to change in input, model representation, or output of the underlying predictor. Finally, our experimental evaluation with three real-world datasets demonstrates interesting insights for seven explanation methods and different stability metrics. |
Chirag Agarwal · Nari Johnson · Martin Pawelczyk · Satyapriya Krishna · Eshika Saxena · Marinka Zitnik · Hima Lakkaraju 🔗 |
-
|
Maximizing entropy on adversarial examples can improve generalization
(
Poster
)
link »
Supervised learning methods that directly optimize the cross entropy loss on training data often overfit. This overfitting is typically mitigated through regularizing the loss function (e.g., label smoothing) or by minimizing the same loss on new examples (e.g., data augmentation and adversarial training). In this work, we propose a complementary regularization strategy: Maximum Predictive Entropy (MPE) forcing the model to be uncertain on new, algorithmically-generated inputs. Across a range of tasks, we demonstrate that our computationally-efficient method improves test accuracy, and the benefits are complementary to methods such as label smoothing and data augmentation. |
Amrith Setlur · Benjamin Eysenbach 🔗 |
-
|
Node-Level Differentially Private Graph Neural Networks
(
Poster
)
link »
Graph Neural Networks (GNNs) are a popular technique for modelling graph-structured data and computing node-level representations via aggregation of information from the neighborhood of each node. However, this aggregation implies increased risk of revealing sensitive information, as a node can participate in the inference for multiple nodes. This implies that standard privacy preserving machine learning techniques, such as differentially private stochastic gradient descent (DP-SGD) - which are designed for situations where each data point participates in the inference for one point only - either do not apply, or lead to inaccurate solutions. In this work, we formally define the problem of learning GNN parameters with node-level privacy, and provide an algorithmic solution with a strong differential privacy guarantee. We employ a careful sensitivity analysis and provide a non-trivial extension of the privacy-by-amplification technique. An empirical evaluation on standard benchmarks datasets and architectures demonstrates that our method is indeed able to learn accurate privacy-preserving GNNs, while still outperforming standard non-private methods that completely ignore graph information. |
Ameya Daigavane · Gagan Madan · Aditya Sinha · Abhradeep Guha Thakurta · Gaurav Aggarwal · Prateek Jain 🔗 |
-
|
Invariant Causal Representation Learning for Generalization in Imitation and Reinforcement Learning
(
Poster
)
link »
A fundamental challenge in imitation and reinforcement learning is to learn policies, representations, or dynamics that do not build on spurious correlations and generalize beyond the specific environments that they were trained on. We investigate these generalization problems from a unified view. For this, we propose a general framework to tackle them with theoretical guarantees on both identifiability and generalizability under mild assumptions on environmental changes. By leveraging a diverse set of training environments, we construct a data representation that ignores any spurious features and consistently predicts target variables well across environments. Following this approach, we build invariant predictors in terms of policy, representations, and dynamics. We theoretically show that the resulting policies, representations, and dynamics are able to generalize to unseen environments. Extensive experiments on both synthetic and real-world datasets show that our methods attain improved generalization over a variety of baselines. |
Chaochao Lu · José Miguel Hernández Lobato · Bernhard Schoelkopf 🔗 |