On the Privacy Risks of Algorithmic Recourse |
AISTATS |
2023 |
N/A |
`` |
|
Towards Bridging the Gaps between the Right to Explanation and the Right to be Forgotten |
ICML |
2023 |
N/A |
`` |
|
Tracr: Compiled Transformers as a Laboratory for Interpretability |
arXiv |
2023 |
Github |
DeepMind |
|
Probabilistically Robust Recourse: Navigating the Trade-offs between Costs and Robustness in Algorithmic Recourse |
ICLR |
2023 |
N/A |
`` |
|
Concept-level Debugging of Part-Prototype Networks |
ICLR |
2023 |
N/A |
`` |
|
Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning |
ICLR |
2023 |
N/A |
`` |
|
Re-calibrating Feature Attributions for Model Interpretation |
ICLR |
2023 |
N/A |
`` |
|
Post-hoc Concept Bottleneck Models |
ICLR |
2023 |
N/A |
`` |
|
Quantifying Memorization Across Neural Language Models |
ICLR |
2023 |
N/A |
`` |
|
STREET: A Multi-Task Structured Reasoning and Explanation Benchmark |
ICLR |
2023 |
N/A |
`` |
|
PIP-Net: Patch-Based Intuitive Prototypes for Interpretable Image Classification |
CVPR |
2023 |
N/A |
`` |
|
EVAL: Explainable Video Anomaly Localization |
CVPR |
2023 |
N/A |
`` |
|
Overlooked Factors in Concept-based Explanations: Dataset Choice, Concept Learnability, and Human Capability |
CVPR |
2023 |
Github |
`` |
|
Spatial-Temporal Concept Based Explanation of 3D ConvNets |
CVPR |
2023 |
Github |
`` |
|
Adversarial Counterfactual Visual Explanations |
CVPR |
2023 |
N/A |
`` |
|
Bridging the Gap Between Model Explanations in Partially Annotated Multi-Label Classification |
CVPR |
2023 |
N/A |
`` |
|
Explaining Image Classifiers With Multiscale Directional Image Representation |
CVPR |
2023 |
N/A |
`` |
|
CRAFT: Concept Recursive Activation FacTorization for Explainability |
CVPR |
2023 |
N/A |
`` |
|
SketchXAI: A First Look at Explainability for Human Sketches |
CVPR |
2023 |
N/A |
`` |
|
Don't Lie to Me! Robust and Efficient Explainability With Verified Perturbation Analysis |
CVPR |
2023 |
N/A |
`` |
|
Gradient-Based Uncertainty Attribution for Explainable Bayesian Deep Learning |
CVPR |
2023 |
N/A |
`` |
|
Learning Bottleneck Concepts in Image Classification |
CVPR |
2023 |
N/A |
`` |
|
Language in a Bottle: Language Model Guided Concept Bottlenecks for Interpretable Image Classification |
CVPR |
2023 |
N/A |
`` |
|
Interpretable Neural-Symbolic Concept Reasoning |
ICML |
2023 |
Github |
|
|
Identifying Interpretable Subspaces in Image Representations |
ICML |
2023 |
N/A |
`` |
|
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models: Route, Interpret, Repeat |
ICML |
2023 |
N/A |
`` |
|
Explainability as statistical inference |
ICML |
2023 |
N/A |
`` |
|
On the Impact of Knowledge Distillation for Model Interpretability |
ICML |
2023 |
N/A |
`` |
|
NA2Q: Neural Attention Additive Model for Interpretable Multi-Agent Q-Learning |
ICML |
2023 |
N/A |
`` |
|
Explaining Reinforcement Learning with Shapley Values |
ICML |
2023 |
N/A |
`` |
|
Explainable Data-Driven Optimization: From Context to Decision and Back Again |
ICML |
2023 |
N/A |
`` |
|
Causal Proxy Models for Concept-based Model Explanations |
ICML |
2023 |
N/A |
`` |
|
Learning Perturbations to Explain Time Series Predictions |
ICML |
2023 |
N/A |
`` |
|
Rethinking Explaining Graph Neural Networks via Non-parametric Subgraph Matching |
ICML |
2023 |
N/A |
`` |
|
Dividing and Conquering a BlackBox to a Mixture of Interpretable Models: Route, Interpret, Repeat |
ICML |
2023 |
Github |
`` |
|
Representer Point Selection for Explaining Regularized High-dimensional Models |
ICML |
2023 |
N/A |
`` |
|
Towards Explaining Distribution Shifts |
ICML |
2023 |
N/A |
`` |
|
Relevant Walk Search for Explaining Graph Neural Networks |
ICML |
2023 |
Github |
`` |
|
Concept-based Explanations for Out-of-Distribution Detectors |
ICML |
2023 |
N/A |
`` |
|
GLOBE-CE: A Translation Based Approach for Global Counterfactual Explanations |
ICML |
2023 |
Github |
`` |
|
Robust Explanation for Free or At the Cost of Faithfulness |
ICML |
2023 |
N/A |
`` |
|
Learn to Accumulate Evidence from All Training Samples: Theory and Practice |
ICML |
2023 |
N/A |
`` |
|
Towards Trustworthy Explanation: On Causal Rationalization |
ICML |
2023 |
N/A |
`` |
|
Theoretical Behavior of XAI Methods in the Presence of Suppressor Variables |
ICML |
2023 |
N/A |
`` |
|
Probabilistic Concept Bottleneck Models |
ICML |
2023 |
N/A |
`` |
|
What do CNNs Learn in the First Layer and Why? A Linear Systems Perspective |
ICML |
2023 |
N/A |
`` |
|
Towards credible visual model interpretation with path attribution |
ICML |
2023 |
N/A |
`` |
|
Trainability, Expressivity and Interpretability in Gated Neural ODEs |
ICML |
2023 |
N/A |
`` |
|
Discover and Cure: Concept-aware Mitigation of Spurious Correlation |
ICML |
2023 |
N/A |
`` |
|
PWSHAP: A Path-Wise Explanation Model for Targeted Variables |
ICML |
2023 |
N/A |
`` |
|
A Closer Look at the Intervention Procedure of Concept Bottleneck Models |
ICML |
2023 |
N/A |
`` |
|
Counterfactual Analysis in Dynamic Latent-State Models |
ICML |
2023 |
N/A |
`` |
|
Tackling Shortcut Learning in Deep Neural Networks: An Iterative Approach with Interpretable Models |
ICML Workshop |
2023 |
N/A |
`` |
|
Rethinking Interpretation: Input-Agnostic Saliency Mapping of Deep Visual Classifiers |
AAAI |
2023 |
N/A |
`` |
|
TopicFM: Robust and Interpretable Topic-Assisted Feature Matching |
AAAI |
2023 |
N/A |
`` |
|
Solving Explainability Queries with Quantification: The Case of Feature Relevancy |
AAAI |
2023 |
N/A |
`` |
|
PEN: Prediction-Explanation Network to Forecast Stock Price Movement with Better Explainability |
AAAI |
2023 |
N/A |
`` |
|
KerPrint: Local-Global Knowledge Graph Enhanced Diagnosis Prediction for Retrospective and Prospective Interpretations |
AAAI |
2023 |
N/A |
`` |
|
Beyond Graph Convolutional Network: An Interpretable Regularizer-Centered Optimization Framework |
AAAI |
2023 |
N/A |
`` |
|
Learning to Select Prototypical Parts for Interpretable Sequential Data Modeling |
AAAI |
2023 |
N/A |
`` |
|
Learning Interpretable Temporal Properties from Positive Examples Only |
AAAI |
2023 |
N/A |
`` |
|
Symbolic Metamodels for Interpreting Black-Boxes Using Primitive Functions |
AAAI |
2023 |
N/A |
`` |
|
Towards More Robust Interpretation via Local Gradient Alignment |
AAAI |
2023 |
N/A |
`` |
|
Towards Fine-Grained Explainability for Heterogeneous Graph Neural Network |
AAAI |
2023 |
N/A |
`` |
|
XClusters: Explainability-First Clustering |
AAAI |
2023 |
N/A |
`` |
|
Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis |
AAAI |
2023 |
N/A |
`` |
|
Fairness and Explainability: Bridging the Gap towards Fair Model Explanations |
AAAI |
2023 |
N/A |
`` |
|
Explaining Model Confidence Using Counterfactuals |
AAAI |
2023 |
N/A |
`` |
|
SEAT: Stable and Explainable Attention |
AAAI |
2023 |
N/A |
`` |
|
Factual and Informative Review Generation for Explainable Recommendation |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability via Explicit Word Interaction Graph Layer |
AAAI |
2023 |
N/A |
`` |
|
Unveiling the Black Box of PLMs with Semantic Anchors: Towards Interpretable Neural Semantic Parsing |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability of Deep Sequential Knowledge Tracing Models with Question-centric Cognitive Representations |
AAAI |
2023 |
N/A |
`` |
|
Targeted Knowledge Infusion To Make Conversational AI Explainable and Safe |
AAAI |
2023 |
N/A |
`` |
|
eForecaster: Unifying Electricity Forecasting with Robust, Flexible, and Explainable Machine Learning Algorithms |
AAAI |
2023 |
N/A |
`` |
|
SolderNet: Towards Trustworthy Visual Inspection of Solder Joints in Electronics Manufacturing Using Explainable Artificial Intelligence |
AAAI |
2023 |
N/A |
`` |
|
Xaitk-Saliency: An Open Source Explainable AI Toolkit for Saliency |
AAAI |
2023 |
N/A |
`` |
|
Ripple: Concept-Based Interpretation for Raw Time Series Models in Education |
AAAI |
2023 |
N/A |
`` |
|
Semantics, Ontology and Explanation |
arXiv |
2023 |
N/A |
Ontological Unpacking |
|
Post Hoc Explanations of Language Models Can Improve Language Models |
arXiv |
2023 |
N/A |
`` |
|
TopicFM: Robust and Interpretable Topic-Assisted Feature Matching |
AAAI |
2023 |
N/A |
`` |
|
Beyond Graph Convolutional Network: An Interpretable Regularizer-Centered Optimization Framework |
AAAI |
2023 |
N/A |
`` |
|
KerPrint: Local-Global Knowledge Graph Enhanced Diagnosis Prediction for Retrospective and Prospective Interpretations |
AAAI |
2023 |
N/A |
`` |
|
Solving Explainability Queries with Quantification: The Case of Feature Relevancy |
AAAI |
2023 |
N/A |
`` |
|
PEN: Prediction-Explanation Network to Forecast Stock Price Movement with Better Explainability |
AAAI |
2023 |
N/A |
`` |
|
Solving Explainability Queries with Quantification: The Case of Feature Relevancy |
AAAI |
2023 |
N/A |
`` |
|
Multi-Aspect Explainable Inductive Relation Prediction by Sentence Transformer |
AAAI |
2023 |
N/A |
`` |
|
Learning to Select Prototypical Parts for Interpretable Sequential Data Modeling |
AAAI |
2023 |
N/A |
`` |
|
Learning Interpretable Temporal Properties from Positive Examples Only |
AAAI |
2023 |
N/A |
`` |
|
Unfooling Perturbation-Based Post Hoc Explainers |
AAAI |
2023 |
N/A |
`` |
|
Very Fast, Approximate Counterfactual Explanations for Decision Forests |
AAAI |
2023 |
N/A |
`` |
|
Symbolic Metamodels for Interpreting Black-Boxes Using Primitive Functions |
AAAI |
2023 |
N/A |
`` |
|
Towards More Robust Interpretation via Local Gradient Alignment |
AAAI |
2023 |
N/A |
`` |
|
Towards Fine-Grained Explainability for Heterogeneous Graph Neural Network |
AAAI |
2023 |
N/A |
`` |
|
Local Explanations for Reinforcement Learning |
AAAI |
2023 |
N/A |
`` |
|
ConceptX: A Framework for Latent Concept Analysis |
AAAI |
2023 |
N/A |
`` |
|
XClusters: Explainability-First Clustering |
AAAI |
2023 |
N/A |
`` |
|
Explaining Random Forests Using Bipolar Argumentation and Markov Networks |
AAAI |
2023 |
N/A |
`` |
|
Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis |
AAAI |
2023 |
N/A |
`` |
|
Fairness and Explainability: Bridging the Gap towards Fair Model Explanations |
AAAI |
2023 |
N/A |
`` |
|
Explaining Model Confidence Using Counterfactuals |
AAAI |
2023 |
N/A |
`` |
|
XRand: Differentially Private Defense against Explanation-Guided Attacks |
AAAI |
2023 |
N/A |
`` |
|
Unsupervised Explanation Generation via Correct Instantiations |
AAAI |
2023 |
N/A |
`` |
|
SEAT: Stable and Explainable Attention |
AAAI |
2023 |
N/A |
`` |
|
Disentangled CVAEs with Contrastive Learning for Explainable Recommendation |
AAAI |
2023 |
N/A |
`` |
|
Factual and Informative Review Generation for Explainable Recommendation |
AAAI |
2023 |
N/A |
`` |
|
Unveiling the Black Box of PLMs with Semantic Anchors: Towards Interpretable Neural Semantic Parsing |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability via Explicit Word Interaction Graph Layer |
AAAI |
2023 |
N/A |
`` |
|
Improving Interpretability of Deep Sequential Knowledge Tracing Models with Question-centric Cognitive Representations |
AAAI |
2023 |
N/A |
`` |
|
Interpretable Chirality-Aware Graph Neural Network for Quantitative Structure Activity Relationship Modeling in Drug Discovery |
AAAI |
2023 |
N/A |
`` |
|
Monitoring Model Deterioration with Explainable Uncertainty Estimation via Non-parametric Bootstrap |
AAAI |
2023 |
N/A |
`` |
|
Interactive Concept Bottleneck Models |
AAAI |
2023 |
N/A |
`` |
|
Data-Efficient and Interpretable Tabular Anomaly Detection |
KDD |
2023 |
N/A |
`` |
|
Counterfactual Learning on Heterogeneous Graphs with Greedy Perturbation |
KDD |
2023 |
N/A |
`` |
|
Hands-on Tutorial: "Explanations in AI: Methods, Stakeholders and Pitfalls" |
KDD |
2023 |
N/A |
`` |
|
Feature-based Learning for Diverse and Privacy-Preserving Counterfactual Explanations |
KDD |
2023 |
N/A |
`` |
|
Generative AI meets Responsible AI: Practical Challenges and Opportunities |
KDD |
2023 |
N/A |
`` |
|
Empower Post-hoc Graph Explanations with Information Bottleneck: A Pre-training and Fine-tuning Perspective |
KDD |
2023 |
N/A |
`` |
|
MixupExplainer: Generalizing Explanations for Graph Neural Networks with Data Augmentation |
KDD |
2023 |
N/A |
`` |
|
CounterNet: End-to-End Training of Prediction Aware Counterfactual Explanations |
KDD |
2023 |
N/A |
`` |
|
Fire: An Optimization Approach for Fast Interpretable Rule Extraction |
KDD |
2023 |
N/A |
`` |
|
ESSA: Explanation Iterative Supervision via Saliency-guided Data Augmentation |
KDD |
2023 |
N/A |
`` |
|
A Causality Inspired Framework for Model Interpretation |
KDD |
2023 |
N/A |
`` |
|
Path-Specific Counterfactual Fairness for Recommender Systems |
KDD |
2023 |
N/A |
`` |
|
SURE: Robust, Explainable, and Fair Classification without Sensitive Attributes |
KDD |
2023 |
N/A |
`` |
|
Learning for Counterfactual Fairness from Observational Data |
KDD |
2023 |
N/A |
`` |
|
Interpretable Sparsification of Brain Graphs: Better Practices and Effective Designs for Graph Neural Networks |
KDD |
2023 |
N/A |
`` |
|
ExplainableFold: Understanding AlphaFold Prediction with Explainable AI |
KDD |
2023 |
N/A |
`` |
|
FLAMES2Graph: An Interpretable Federated Multivariate Time Series Classification Framework |
KDD |
2023 |
N/A |
`` |
|
Feature-based Learning for Diverse and Privacy-Preserving Counterfactual Explanations |
KDD |
2023 |
N/A |
`` |
|
ESSA: Explanation Iterative Supervision via Saliency-guided Data Augmentation |
KDD |
2023 |
N/A |
`` |
|
Counterfactual Explanations and Model Multiplicity: a Relational Verification View |
Proceedings of KR |
2023 |
N/A |
`` |
|
Explainable Representations for Relation Prediction in Knowledge Graphs |
Proceedings of KR |
2023 |
N/A |
`` |
|
Region-based Saliency Explanations on the Recognition of Facial Genetic Syndromes |
PMLR |
2023 |
N/A |
`` |
|
FunnyBirds: A Synthetic Vision Dataset for a Part-Based Analysis of Explainable AI Methods |
arXiv |
2023 |
N/A |
`` |
|
Diffusion-based Visual Counterfactual Explanations - Towards Systematic Quantitative Evaluation |
arXiv |
2023 |
N/A |
`` |
|
Testing methods of neural systems understanding |
Cognitive Systems Research |
2023 |
N/A |
`` |
|
Understanding CNN Hidden Neuron Activations Using Structured Background Knowledge and Deductive Reasoning |
arXiv |
2023 |
N/A |
`` |
|
An Explainable Federated Learning and Blockchain based Secure Credit Modeling Method |
EJOR |
2023 |
N/A |
`` |
|
i-Align: an interpretable knowledge graph alignment model |
DMKD |
2023 |
N/A |
`` |
|
Goodhart’s Law Applies to NLP’s Explanation Benchmarks |
arXiv |
2023 |
N/A |
`` |
|
DELELSTM: DECOMPOSITION-BASED LINEAR EXPLAINABLE LSTM TO CAPTURE INSTANTANEOUS AND LONG-TERM EFFECTS IN TIME SERIES |
arXiv |
2023 |
N/A |
`` |
|
BEYOND DISCRIMINATIVE REGIONS: SALIENCY MAPS AS ALTERNATIVES TO CAMS FOR WEAKLY SU- PERVISED SEMANTIC SEGMENTATION |
arXiv |
2023 |
N/A |
`` |
|
SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks |
arXiv |
2023 |
N/A |
`` |
|
Sparse Linear Concept Discovery Models |
arXiv |
2023 |
N/A |
`` |
|
Revisiting the Performance-Explainability Trade-Off in Explainable Artificial Intelligence (XAI) |
arXiv |
2023 |
N/A |
`` |
|
KGTN: Knowledge Graph Transformer Network for explainable multi-category item recommendation |
KBS |
2023 |
N/A |
`` |
|
SAFE: Saliency-Aware Counterfactual Explanations for DNN-based Automated Driving Systems |
arXiv |
2023 |
N/A |
`` |
|
Explainable Multi-Agent Reinforcement Learning for Temporal Queries |
IJCAI |
2023 |
N/A |
`` |
|
Advancing Post-Hoc Case-Based Explanation with Feature Highlighting |
IJCAI |
2023 |
N/A |
`` |
|
Explanation-Guided Reward Alignment |
IJCAI |
2023 |
N/A |
`` |
|
FEAMOE: Fair, Explainable and Adaptive Mixture of Experts |
IJCAI |
2023 |
N/A |
`` |
|
Statistically Significant Concept-based Explanation of Image Classifiers via Model Knockoffs |
IJCAI |
2023 |
N/A |
`` |
|
Learning Prototype Classifiers for Long-Tailed Recognition |
IJCAI |
2023 |
N/A |
`` |
|
On Translations between ML Models for XAI Purposes |
IJCAI |
2023 |
N/A |
`` |
|
The Parameterized Complexity of Finding Concise Local Explanations |
IJCAI |
2023 |
N/A |
`` |
|
Neuro-Symbolic Class Expression Learning |
IJCAI |
2023 |
N/A |
`` |
|
A Logic-based Approach to Contrastive Explainability for Neurosymbolic Visual Question Answering |
IJCAI |
2023 |
N/A |
`` |
|
Cardinality-Minimal Explanations for Monotonic Neural Networks |
IJCAI |
2023 |
N/A |
`` |
|
Unveiling Concepts Learned by a World-Class Chess-Playing Agent |
IJCAI |
2023 |
N/A |
`` |
|
Explainable Text Classification via Attentive and Targeted Mixing Data Augmentation |
IJCAI |
2023 |
N/A |
`` |
|
On the Complexity of Counterfactual Reasoning |
IJCAI |
2023 |
N/A |
`` |
|
Interpretable Local Concept-based Explanation with Human Feedback to Predict All-cause Mortality (Extended Abstract) |
IJCAI |
2023 |
N/A |
`` |
|
Good-looking but Lacking Faithfulness: Understanding Local Explanation Methods through Trend-based Testing |
arXiv |
2023 |
N/A |
`` |
|
Counterfactual Explanations via Locally-guided Sequential Algorithmic Recourse |
arXiv |
2023 |
N/A |
`` |
|
Flexible and Robust Counterfactual Explanations with Minimal Satisfiable Perturbations |
CIKM |
2023 |
N/A |
`` |
|
A Function Interpretation Benchmark for Evaluating Interpretability Methods |
arXiv |
2023 |
N/A |
`` |
|
Explaining through Transformer Input Sampling |
arXiv |
2023 |
N/A |
`` |
|
Backtracking Counterfactuals |
CLeaR |
2023 |
N/A |
`` |
|
Text2Concept: Concept Activation Vectors Directly from Text |
CVPR Workshop |
2023 |
N/A |
`` |
|
A Holistic Approach to Unifying Automatic Concept Extraction and Concept Importance Estimation |
arXiv |
2023 |
N/A |
`` |
|
Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance |
NeurIPS |
2023 |
Github |
`` |
|
CLIP-DISSECT: AUTOMATIC DESCRIPTION OF NEU- RON REPRESENTATIONS IN DEEP VISION NETWORKS |
ICLR |
2023 |
Github |
`` |
|
Label-free Concept Bottleneck Models |
ICLR |
2023 |
N/A |
`` |
|
Concept-level Debugging of Part-Prototype Networks |
ICLR |
2023 |
N/A |
`` |
|
Towards Interpretable Deep Reinforcement Learning with Human-Friendly Prototypes |
ICLR |
2023 |
N/A |
`` |
|
Re-calibrating Feature Attributions for Model Interpretation |
ICLR |
2023 |
N/A |
`` |
|
Post-hoc Concept Bottleneck Models |
ICLR |
2023 |
N/A |
`` |
|
Information Maximization Perspective of Orthogonal Matching Pursuit with Applications to Explainable AI |
NeurIPS |
2023 |
N/A |
`` |
|
Explaining Predictive Uncertainty with Information Theoretic Shapley Values |
NeurIPS |
2023 |
N/A |
`` |
|
REASONER: An Explainable Recommendation Dataset with Comprehensive Labeling Ground Truths |
NeurIPS |
2023 |
N/A |
`` |
|
Explain Any Concept: Segment Anything Meets Concept-Based Explanation |
NeurIPS |
2023 |
N/A |
`` |
|
VeriX: Towards Verified Explainability of Deep Neural Networks |
NeurIPS |
2023 |
N/A |
`` |
|
Explainable and Efficient Randomized Voting Rules |
NeurIPS |
2023 |
N/A |
`` |
|
TempME: Towards the Explainability of Temporal Graph Neural Networks via Motif Discovery |
NeurIPS |
2023 |
N/A |
`` |
|
Explaining the Uncertain: Stochastic Shapley Values for Gaussian Process Models |
NeurIPS |
2023 |
N/A |
`` |
|
V-InFoR: A Robust Graph Neural Networks Explainer for Structurally Corrupted Graphs |
NeurIPS |
2023 |
N/A |
`` |
|
Explainable Brain Age Prediction using coVariance Neural Networks |
NeurIPS |
2023 |
N/A |
`` |
|
TempME: Towards the Explainability of Temporal Graph Neural Networks via Motif Discovery |
NeurIPS |
2023 |
N/A |
`` |
|
D4Explainer: In-distribution Explanations of Graph Neural Network via Discrete Denoising Diffusion |
NeurIPS |
2023 |
N/A |
`` |
|
StateMask: Explaining Deep Reinforcement Learning through State Mask |
NeurIPS |
2023 |
N/A |
`` |
|
LICO: Explainable Models with Language-Image COnsistency |
NeurIPS |
2023 |
N/A |
`` |
|
On the explainable properties of 1-Lipschitz Neural Networks: An Optimal Transport Perspective |
NeurIPS |
2023 |
N/A |
`` |
|
Interpretable and Explainable Logical Policies via Neurally Guided Symbolic Abstraction |
NeurIPS |
2023 |
N/A |
`` |
|
Discriminative Feature Attributions: Bridging Post Hoc Explainability and Inherent Interpretability |
NeurIPS |
2023 |
N/A |
`` |
|
Train Once and Explain Everywhere: Pre-training Interpretable Graph Neural Networks |
NeurIPS |
2023 |
N/A |
`` |
|
Accountability in Offline Reinforcement Learning: Explaining Decisions with a Corpus of Examples |
NeurIPS |
2023 |
N/A |
`` |
|
HiBug: On Human-Interpretable Model Debug |
NeurIPS |
2023 |
N/A |
`` |
|
Towards Self-Interpretable Graph-Level Anomaly Detection |
NeurIPS |
2023 |
N/A |
`` |
|
Interpretable Graph Networks Formulate Universal Algebra Conjectures |
NeurIPS |
2023 |
N/A |
`` |
|
Towards Automated Circuit Discovery for Mechanistic Interpretabilit |
NeurIPS |
2023 |
N/A |
`` |
|
Interpretable Reward Redistribution in Reinforcement Learning: A Causal Approach |
NeurIPS |
2023 |
N/A |
`` |
|
DISCOVER: Making Vision Networks Interpretable via Competition and Dissection |
NeurIPS |
2023 |
N/A |
`` |
|
MultiMoDN—Multimodal, Multi-Task, Interpretable Modular Networks |
NeurIPS |
2023 |
N/A |
`` |
|
Causal Interpretation of Self-Attention in Pre-Trained Transformers |
NeurIPS |
2023 |
N/A |
`` |
|
Tracr: Compiled Transformers as a Laboratory for Interpretability |
NeurIPS |
2023 |
N/A |
`` |
|
Learning Interpretable Low-dimensional Representation via Physical Symmetry |
NeurIPS |
2023 |
N/A |
`` |
|
Scale Alone Does not Improve Mechanistic Interpretability in Vision Models |
NeurIPS |
2023 |
N/A |
`` |
|
Transitivity Recovering Decompositions: Interpretable and Robust Fine-Grained Relationships |
NeurIPS |
2023 |
N/A |
`` |
|
GRAND-SLAMIN’ Interpretable Additive Modeling with Structural Constraints |
NeurIPS |
2023 |
N/A |
`` |
|
Interpreting Unsupervised Anomaly Detection in Security via Rule Extraction |
NeurIPS |
2023 |
N/A |
`` |
|
GPEX, A Framework For Interpreting Artificial Neural Networks |
NeurIPS |
2023 |
N/A |
`` |
|
Dynamic Context Pruning for Efficient and Interpretable Autoregressive Transformers |
NeurIPS |
2023 |
N/A |
`` |
|
ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP |
NeurIPS |
2023 |
N/A |
`` |
|
On the Identifiability and Interpretability of Gaussian Process Models |
NeurIPS |
2023 |
N/A |
`` |
|
BasisFormer: Attention-based Time Series Forecasting with Learnable and Interpretable Basis |
NeurIPS |
2023 |
N/A |
`` |
|
Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance |
NeurIPS |
2023 |
N/A |
`` |
|
Evaluating Neuron Interpretation Methods of NLP Models |
NeurIPS |
2023 |
N/A |
`` |
|
FIND: A Function Description Benchmark for Evaluating Interpretability Methods |
NeurIPS |
2023 |
N/A |
`` |
|
How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model |
NeurIPS |
2023 |
N/A |
`` |
|
Interpretable Prototype-based Graph Information Bottleneck |
NeurIPS |
2023 |
N/A |
`` |
|
Interpretability at Scale: Identifying Causal Mechanisms in Alpaca |
NeurIPS |
2023 |
N/A |
`` |
|
M4: A Unified XAI Benchmark for Faithfulness Evaluation of Feature Attribution Methods across Metrics, Modalities and Models |
NeurIPS |
2023 |
N/A |
`` |
|
InstructSafety: A Unified Framework for Building Multidimensional and Explainable Safety Detector through Instruction Tuning |
EMNLP |
2023 |
N/A |
`` |
|
Towards Explainable and Accessible AI |
EMNLP |
2023 |
N/A |
`` |
|
KEBAP: Korean Error Explainable Benchmark Dataset for ASR and Post-processing |
EMNLP |
2023 |
N/A |
`` |
|
INSTRUCTSCORE: Towards Explainable Text Generation Evaluation with Automatic Feedback |
EMNLP |
2023 |
N/A |
`` |
|
Goal-Driven Explainable Clustering via Language Descriptions |
EMNLP |
2023 |
N/A |
`` |
|
VECHR: A Dataset for Explainable and Robust Classification of Vulnerability Type in the European Court of Human Rights |
EMNLP |
2023 |
N/A |
`` |
|
COFFEE: Counterfactual Fairness for Personalized Text Generation in Explainable Recommendation |
EMNLP |
2023 |
N/A |
`` |
|
Hop, Union, Generate: Explainable Multi-hop Reasoning without Rationale Supervision |
EMNLP |
2023 |
N/A |
`` |
|
GenEx: A Commonsense-aware Unified Generative Framework for Explainable Cyberbullying Detection |
EMNLP |
2023 |
N/A |
`` |
|
DRGCoder: Explainable Clinical Coding for the Early Prediction of Diagnostic-Related Groups |
EMNLP |
2023 |
N/A |
`` |
|
LLM4Vis: Explainable Visualization Recommendation using ChatGPT |
EMNLP |
2023 |
N/A |
`` |
|
Harnessing LLMs for Temporal Data - A Study on Explainable Financial Time Series Forecasting |
EMNLP |
2023 |
N/A |
`` |
|
HARE: Explainable Hate Speech Detection with Step-by-Step Reasoning |
EMNLP |
2023 |
N/A |
`` |
|
Distilling ChatGPT for Explainable Automated Student Answer Assessment |
EMNLP |
2023 |
N/A |
`` |
|
Explainable Claim Verification via Knowledge-Grounded Reasoning with Large Language Models |
EMNLP |
2023 |
N/A |
`` |
|
Leveraging Structured Information for Explainable Multi-hop Question Answering and Reasoning |
EMNLP |
2023 |
N/A |
`` |
|
InstructSafety: A Unified Framework for Building Multidimensional and Explainable Safety Detector through Instruction Tuning |
EMNLP |
2023 |
N/A |
`` |
|
Deep Integrated Explanations |
CIKM |
2023 |
N/A |
`` |
|
KG4Ex: An Explainable Knowledge Graph-Based Approach for Exercise Recommendation |
CIKM |
2023 |
N/A |
`` |
|
Interpretable Fake News Detection with Graph Evidence |
CIKM |
2023 |
N/A |
`` |
|
PriSHAP: Prior-guided Shapley Value Explanations for Correlated Features |
CIKM |
2023 |
N/A |
`` |
|
A Model-Agnostic Method to Interpret Link Prediction Evaluation of Knowledge Graph Embeddings |
CIKM |
2023 |
N/A |
`` |
|
ACGAN-GNNExplainer: Auxiliary Conditional Generative Explainer for Graph Neural Networks |
CIKM |
2023 |
N/A |
`` |
|
Concept Evolution in Deep Learning Training: A Unified Interpretation Framework and Discoveries |
CIKM |
2023 |
N/A |
`` |
|
Explainable Spatio-Temporal Graph Neural Networks |
CIKM |
2023 |
N/A |
`` |
|
Towards Deeper, Lighter and Interpretable Cross Network for CTR Prediction |
CIKM |
2023 |
N/A |
`` |
|
Flexible and Robust Counterfactual Explanations with Minimal Satisfiable Perturbations |
CIKM |
2023 |
N/A |
`` |
|
NOVO: Learnable and Interpretable Document Identifiers for Model-Based IR |
CIKM |
2023 |
N/A |
`` |
|
Counterfactual Monotonic Knowledge Tracing for Assessing Students' Dynamic Mastery of Knowledge Concepts |
CIKM |
2023 |
N/A |
`` |
|
Contrastive Counterfactual Learning for Causality-aware Interpretable Recommender Systems |
CIKM |
2023 |
N/A |
`` |
|
|
|
2023 |
N/A |
`` |
|