On the Privacy Risks of Algorithmic Recourse |
AISTATS |
2023 |
N/A |
`` |
|
Towards Bridging the Gaps between the Right to Explanation and the Right to be Forgotten |
ICML |
2023 |
N/A |
`` |
|
Tracr: Compiled Transformers as a Laboratory for Interpretability |
arXiv |
2023 |
Github |
DeepMind |
|
Probabilistically Robust Recourse: Navigating the Trade-offs between Costs and Robustness in Algorithmic Recourse |
ICLR |
2023 |
N/A |
`` |
|
Concept-level Debugging of Part-Prototype Networks |
ICLR |
2023 |
N/A |
`` |
|
Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning |
ICLR |
2023 |
N/A |
`` |
|
Re-calibrating Feature Attributions for Model Interpretation |
ICLR |
2023 |
N/A |
`` |
|
Post-hoc Concept Bottleneck Models |
ICLR |
2023 |
N/A |
`` |
|
Quantifying Memorization Across Neural Language Models |
ICLR |
2023 |
N/A |
`` |
|
STREET: A Multi-Task Structured Reasoning and Explanation Benchmark |
ICLR |
2023 |
N/A |
`` |
|
PIP-Net: Patch-Based Intuitive Prototypes for Interpretable Image Classification |
CVPR |
2023 |
N/A |
`` |
|
EVAL: Explainable Video Anomaly Localization |
CVPR |
2023 |
N/A |
`` |
|
Overlooked Factors in Concept-based Explanations: Dataset Choice, Concept Learnability, and Human Capability |
CVPR |
2023 |
Github |
`` |
|
Spatial-Temporal Concept Based Explanation of 3D ConvNets |
CVPR |
2023 |
Github |
`` |
|
Adversarial Counterfactual Visual Explanations |
CVPR |
2023 |
N/A |
`` |
|
Bridging the Gap Between Model Explanations in Partially Annotated Multi-Label Classification |
CVPR |
2023 |
N/A |
`` |
|
Explaining Image Classifiers With Multiscale Directional Image Representation |
CVPR |
2023 |
N/A |
`` |
|
CRAFT: Concept Recursive Activation FacTorization for Explainability |
CVPR |
2023 |
N/A |
`` |
|
SketchXAI: A First Look at Explainability for Human Sketches |
CVPR |
2023 |
N/A |
`` |
|
Don't Lie to Me! Robust and Efficient Explainability With Verified Perturbation Analysis |
CVPR |
2023 |
N/A |
`` |
|
Gradient-Based Uncertainty Attribution for Explainable Bayesian Deep Learning |
CVPR |
2023 |
N/A |
`` |
|
Language in a Bottle: Language Model Guided Concept Bottlenecks for Interpretable Image Classification |
CVPR |
2023 |
N/A |
`` |
|
Interpretable Neural-Symbolic Concept Reasoning |
ICML |
2023 |
Github |
|
|
Identifying Interpretable Subspaces in Image Representations |
ICML |
2023 |
N/A |
`` |
|
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models: Route, Interpret, Repeat |
ICML |
2023 |
N/A |
`` |
|
Explainability as statistical inference |
ICML |
2023 |
N/A |
`` |
|
On the Impact of Knowledge Distillation for Model Interpretability |
ICML |
2023 |
N/A |
`` |
|
NA2Q: Neural Attention Additive Model for Interpretable Multi-Agent Q-Learning |
ICML |
2023 |
N/A |
`` |
|
Explaining Reinforcement Learning with Shapley Values |
ICML |
2023 |
N/A |
`` |
|
Explainable Data-Driven Optimization: From Context to Decision and Back Again |
ICML |
2023 |
N/A |
`` |
|
Causal Proxy Models for Concept-based Model Explanations |
ICML |
2023 |
N/A |
`` |
|
Learning Perturbations to Explain Time Series Predictions |
ICML |
2023 |
N/A |
`` |
|
Rethinking Explaining Graph Neural Networks via Non-parametric Subgraph Matching |
ICML |
2023 |
N/A |
`` |
|
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models: Route, Interpret, Repeat |
ICML |
2023 |
Github |
`` |
|
Representer Point Selection for Explaining Regularized High-dimensional Models |
ICML |
2023 |
N/A |
`` |
|
Towards Explaining Distribution Shifts |
ICML |
2023 |
N/A |
`` |
|
Relevant Walk Search for Explaining Graph Neural Networks |
ICML |
2023 |
Github |
`` |
|
Concept-based Explanations for Out-of-Distribution Detectors |
ICML |
2023 |
N/A |
`` |
|
GLOBE-CE: A Translation Based Approach for Global Counterfactual Explanations |
ICML |
2023 |
N/A |
`` |
|
Robust Explanation for Free or At the Cost of Faithfulness |
ICML |
2023 |
N/A |
`` |
|
Learn to Accumulate Evidence from All Training Samples: Theory and Practice |
ICML |
2023 |
N/A |
`` |
|
Towards Trustworthy Explanation: On Causal Rationalization |
ICML |
2023 |
N/A |
`` |
|
Theoretical Behavior of XAI Methods in the Presence of Suppressor Variables |
ICML |
2023 |
N/A |
`` |
|
Probabilistic Concept Bottleneck Models |
ICML |
2023 |
N/A |
`` |
|
What do CNNs Learn in the First Layer and Why? A Linear Systems Perspective |
ICML |
2023 |
N/A |
`` |
|
Towards credible visual model interpretation with path attribution |
ICML |
2023 |
N/A |
`` |
|
Trainability, Expressivity and Interpretability in Gated Neural ODEs |
ICML |
2023 |
N/A |
`` |
|
Discover and Cure: Concept-aware Mitigation of Spurious Correlation |
ICML |
2023 |
N/A |
`` |
|
PWSHAP: A Path-Wise Explanation Model for Targeted Variables |
ICML |
2023 |
N/A |
`` |
|
A Closer Look at the Intervention Procedure of Concept Bottleneck Models |
ICML |
2023 |
N/A |
`` |
|
Rethinking Interpretation: Input-Agnostic Saliency Mapping of Deep Visual Classifiers |
AAAI |
2023 |
N/A |
`` |
|
TopicFM: Robust and Interpretable Topic-Assisted Feature Matching |
AAAI |
2023 |
N/A |
`` |
|
Solving Explainability Queries with Quantification: The Case of Feature Relevancy |
AAAI |
2023 |
N/A |
`` |
|
PEN: Prediction-Explanation Network to Forecast Stock Price Movement with Better Explainability |
AAAI |
2023 |
N/A |
`` |
|
KerPrint: Local-Global Knowledge Graph Enhanced Diagnosis Prediction for Retrospective and Prospective Interpretations |
AAAI |
2023 |
N/A |
`` |
|
Beyond Graph Convolutional Network: An Interpretable Regularizer-Centered Optimization Framework |
AAAI |
2023 |
N/A |
`` |
|
Learning to Select Prototypical Parts for Interpretable Sequential Data Modeling |
AAAI |
2023 |
N/A |
`` |
|
Learning Interpretable Temporal Properties from Positive Examples Only |
AAAI |
2023 |
N/A |
`` |
|
Symbolic Metamodels for Interpreting Black-Boxes Using Primitive Functions |
AAAI |
2023 |
N/A |
`` |
|
Towards More Robust Interpretation via Local Gradient Alignment |
AAAI |
2023 |
N/A |
`` |
|
Towards Fine-Grained Explainability for Heterogeneous Graph Neural Network |
AAAI |
2023 |
N/A |
`` |
|
XClusters: Explainability-First Clustering |
AAAI |
2023 |
N/A |
`` |
|
Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis |
AAAI |
2023 |
N/A |
`` |
|
Fairness and Explainability: Bridging the Gap towards Fair Model Explanations |
AAAI |
2023 |
N/A |
`` |
|
Explaining Model Confidence Using Counterfactuals |
AAAI |
2023 |
N/A |
`` |
|
SEAT: Stable and Explainable Attention |
AAAI |
2023 |
N/A |
`` |
|
Factual and Informative Review Generation for Explainable Recommendation |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability via Explicit Word Interaction Graph Layer |
AAAI |
2023 |
N/A |
`` |
|
Unveiling the Black Box of PLMs with Semantic Anchors: Towards Interpretable Neural Semantic Parsing |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability of Deep Sequential Knowledge Tracing Models with Question-centric Cognitive Representations |
AAAI |
2023 |
N/A |
`` |
|
Targeted Knowledge Infusion To Make Conversational AI Explainable and Safe |
AAAI |
2023 |
N/A |
`` |
|
eForecaster: Unifying Electricity Forecasting with Robust, Flexible, and Explainable Machine Learning Algorithms |
AAAI |
2023 |
N/A |
`` |
|
SolderNet: Towards Trustworthy Visual Inspection of Solder Joints in Electronics Manufacturing Using Explainable Artificial Intelligence |
AAAI |
2023 |
N/A |
`` |
|
Xaitk-Saliency: An Open Source Explainable AI Toolkit for Saliency |
AAAI |
2023 |
N/A |
`` |
|
Ripple: Concept-Based Interpretation for Raw Time Series Models in Education |
AAAI |
2023 |
N/A |
`` |
|
Semantics, Ontology and Explanation |
arXiv |
2023 |
N/A |
Ontological Unpacking |
|
|
|
|
N/A |
`` |
|