A maintained and curated list of practical and awesome responsible machine learning resources.
If you want to contribute to this list (and please do!), read over the contribution guidelines, send a pull request, or file an issue.
If something you contributed or found here is missing after our September 2023 redeux, please check the archive.
-
Community and Official Guidance Resources
-
Education Resources
-
Miscellaneous Resources
-
Technical Resources
This section is for responsible ML guidance put forward by organizations or individuals, not for official government guidance.
- 8 Principles of Responsible ML
- A Brief Overview of AI Governance for Responsible Machine Learning Systems
- ACL 2024 Tutorial: Vulnerabilities of Large Language Models to Adversarial Attacks
- Adversarial ML Threat Matrix
- AI Verify:
- AI Snake Oil
- AllenNLP Interpret: A Framework for Explaining Predictions of NLP Models
- Andreessen Horowitz (a16z) AI Canon
- Anthropic's Responsible Scaling Policy
- AuditBoard: 5 AI Auditing Frameworks to Encourage Accountability
- Auditing machine learning algorithms: A white paper for public auditors
- AWS Data Privacy FAQ
- AWS Privacy Notice
- AWS, What is Data Governance?
- Berryville Institute of Machine Learning, Architectural Risk Analysis of Large Language Models (requires free account login)
- Brendan Bycroft's LLM Visualization
- BIML Interactive Machine Learning Risk Framework
- Center for Security and Emerging Technology (CSET):
- Censius: AI Audit
- CivAI, GenAI Toolkit for the NIST AI Risk Management Framework: Thinking Through the Risks of a GenAI Chatbot
- Crowe LLP: Internal auditor's AI safety checklist
- DAIR Prompt Engineering Guide
- Data Provenance Explorer
- Data & Society, AI Red-Teaming Is Not a One-Stop Solution to AI Harms: Recommendations for Using Red-Teaming for AI Accountability
- Dealing with Bias and Fairness in AI/ML/Data Science Systems
- Debugging Machine Learning Models (ICLR workshop proceedings)
- Decision Points in AI Governance
- Distill
- Ethical and social risks of harm from Language Models
- Evaluating LLMs is a minefield
- Extracting Training Data from ChatGPT
- Fairly's Global AI Regulations Map
- FATML Principles and Best Practices
- ForHumanity Body of Knowledge (BOK)
- The Foundation Model Transparency Index
- From Principles to Practice: An interdisciplinary framework to operationalise AI ethics
- Frontier Model Forum: What is Red Teaming?
- Gage Repeatability and Reproducibility
- Georgetown University Library's Artificial Intelligence (Generative) Resources
- Google:
- Closing the AI accountability gap: defining an end-to-end framework for internal algorithmic auditing
- The Data Cards Playbook
- Data governance in the cloud - part 1 - People and processes
- Data Governance in the Cloud - part 2 - Tools
- Evaluating social and ethical risks from generative AI
- Generative AI Prohibited Use Policy
- Principles and best practices for data governance in the cloud
- Responsible AI Framework
- Responsible AI practices
- Testing and Debugging in Machine Learning
- H2O.ai Algorithms
- Haptic Networks: How to Perform an AI Audit for UK Organisations
- Hogan Lovells, The AI Act is coming: EU reaches political agreement on comprehensive regulation of artificial intelligence
- Hugging Face, The Landscape of ML Documentation Tools
- IAPP EU AI Act Cheat Sheet
- ICT Institute: A checklist for auditing AI systems
- IEEE:
- Independent Audit of AI Systems
- Identifying and Eliminating CSAM in Generative ML Training Data and Models
- Identifying and Overcoming Common Data Mining Mistakes
- Infocomm Media Development Authority (Singapore), First of its kind Generative AI Evaluation Sandbox for Trusted AI by AI Verify Foundation and IMDA
- Institute of Internal Auditors: Artificial Intelligence Auditing Framework, Practical Applications, Part A, Special Edition
- ISACA:
- Know Your Data
- Large language models, explained with a minimum of math and jargon
- Larry G. Wlosinski, April 30, 2021, Information System Contingency Planning Guidance
- Library of Congress, LC Labs AI Planning Framework
- Llama 2 Responsible Use Guide
- LLM Visualization
- Machine Learning Attack_Cheat_Sheet
- Machine Learning Quick Reference: Algorithms
- Machine Learning Quick Reference: Best Practices
- Manifest MLBOM Wiki
- Meta:
- Microsoft:
- model-cards-and-datasheets
- NewsGuard AI Tracking Center
- OpenAI Cookbook, How to implement LLM guardrails
- OpenAI, Evals
- OpenAI Red Teaming Network
- Open Sourcing Highly Capable Foundation Models
- Organization and Training of a Cyber Security Team
- Our Data Our Selves, Data Use Policy
- PAIR Explorables: Datasets Have Worldviews
- Partnership on AI, ABOUT ML Reference Document
- Partnership on AI, Responsible Practices for Synthetic Media: A Framework for Collective Action
- PwC's Responsible AI
- Real-World Strategies for Model Debugging
- RecoSense: Phases of an AI Data Audit – Assessing Opportunity in the Enterprise
- Red Teaming of Advanced Information Assurance Concepts
- Red Teaming Language Models to Reduce Harms: Methods, Scaling Behaviors, and Lessons Learned
- Robust ML
- Safe and Reliable Machine Learning
- Sample AI Incident Response Checklist
- SHRM Generative Artificial Intelligence (AI) Chatbot Usage Policy
- Stanford University, Responsible AI at Stanford: Enabling innovation through AI best practices
- The Rise of Generative AI and the Coming Era of Social Media Manipulation 3.0: Next-Generation Chinese Astroturfing and Coping with Ubiquitous AI
- Taskade: AI Audit PBC Request Checklist Template
- TechTarget: 9 questions to ask when auditing your AI systems
- Troubleshooting Deep Neural Networks
- Twitter Algorithmic Bias Bounty
- Unite.AI: How to perform an AI Audit in 2023
- University of California, Berkeley, Center for Long-Term Cybersecurity, A Taxonomy of Trustworthiness for Artificial Intelligence
- University of California, Berkeley, Information Security Office, How to Write an Effective Website Privacy Statement
- University of Washington Tech Policy Lab, Data Statements
- Warning Signs: The Future of Privacy and Security in an Age of Machine Learning
- When Not to Trust Your Explanations
- Why We Need to Know More: Exploring the State of AI Incident Documentation Practices
- You Created A Machine Learning Application Now Make Sure It's Secure
This section is for conferences, workshops and other major events related to responsible ML.
- AAAI Conference on Artificial Intelligence
- ACM FAccT (Fairness, Accountability, and Transparency)
- ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization (EAAMO)
- AIES (AAAI/ACM Conference on AI, Ethics, and Society)
- Black in AI
- Computer Vision and Pattern Recognition (CVPR)
- International Conference on Machine Learning (ICML)
- 2020:
- 2nd ICML Workshop on Human in the Loop Learning (HILL)
- 5th ICML Workshop on Human Interpretability in Machine Learning (WHI)
- Challenges in Deploying and Monitoring Machine Learning Systems
- Economics of privacy and data labor
- Federated Learning for User Privacy and Data Confidentiality
- Healthcare Systems, Population Health, and the Role of Health-tech
- Law & Machine Learning
- ML Interpretability for Scientific Discovery
- MLRetrospectives: A Venue for Self-Reflection in ML Research
- Participatory Approaches to Machine Learning
- XXAI: Extending Explainable AI Beyond Deep Models and Classifiers
- 2021:
- Human-AI Collaboration in Sequential Decision-Making
- Machine Learning for Data: Automated Creation, Privacy, Bias
- ICML Workshop on Algorithmic Recourse
- ICML Workshop on Human in the Loop Learning (HILL)
- ICML Workshop on Theoretic Foundation, Criticism, and Application Trend of Explainable AI
- Information-Theoretic Methods for Rigorous, Responsible, and Reliable Machine Learning (ITR3)
- International Workshop on Federated Learning for User Privacy and Data Confidentiality in Conjunction with ICML 2021 (FL-ICML'21)
- Interpretable Machine Learning in Healthcare
- Self-Supervised Learning for Reasoning and Perception
- The Neglected Assumptions In Causal Inference
- Theory and Practice of Differential Privacy
- Uncertainty and Robustness in Deep Learning
- Workshop on Computational Approaches to Mental Health @ ICML 2021
- Workshop on Distribution-Free Uncertainty Quantification
- Workshop on Socially Responsible Machine Learning
- 2022:
- 1st ICML 2022 Workshop on Safe Learning for Autonomous Driving (SL4AD)
- 2nd Workshop on Interpretable Machine Learning in Healthcare (IMLH)
- DataPerf: Benchmarking Data for Data-Centric AI
- Disinformation Countermeasures and Machine Learning (DisCoML)
- Responsible Decision Making in Dynamic Environments
- Spurious correlations, Invariance, and Stability (SCIS)
- The 1st Workshop on Healthcare AI and COVID-19
- Theory and Practice of Differential Privacy
- Workshop on Human-Machine Collaboration and Teaming
- 2023:
- 2nd ICML Workshop on New Frontiers in Adversarial Machine Learning
- 2nd Workshop on Formal Verification of Machine Learning
- 3rd Workshop on Interpretable Machine Learning in Healthcare (IMLH)
- Challenges in Deployable Generative AI
- “Could it have been different?” Counterfactuals in Minds and Machines
- Federated Learning and Analytics in Practice: Algorithms, Systems, Applications, and Opportunities
- Generative AI and Law (GenLaw)
- Interactive Learning with Implicit Human Feedback
- Neural Conversational AI Workshop - What’s left to TEACH (Trustworthy, Enhanced, Adaptable, Capable and Human-centric) chatbots?
- The Second Workshop on Spurious Correlations, Invariance and Stability
- 2020:
- Knowledge, Discovery, and Data Mining (KDD)
- Neural Information Processing Systems (NeurIPs)
- 2022:
- 5th Robot Learning Workshop: Trustworthy Robotics
- Algorithmic Fairness through the Lens of Causality and Privacy
- Causal Machine Learning for Real-World Impact
- Challenges in Deploying and Monitoring Machine Learning Systems
- Cultures of AI and AI for Culture
- Empowering Communities: A Participatory Approach to AI for Mental Health
- Federated Learning: Recent Advances and New Challenges
- Gaze meets ML
- HCAI@NeurIPS 2022, Human Centered AI
- Human Evaluation of Generative Models
- Human in the Loop Learning (HiLL) Workshop at NeurIPS 2022
- I Can’t Believe It’s Not Better: Understanding Deep Learning Through Empirical Falsification
- Learning Meaningful Representations of Life
- Machine Learning for Autonomous Driving
- Progress and Challenges in Building Trustworthy Embodied AI
- Tackling Climate Change with Machine Learning
- Trustworthy and Socially Responsible Machine Learning
- Workshop on Machine Learning Safety
- 2023:
- AI meets Moral Philosophy and Moral Psychology: An Interdisciplinary Dialogue about Computational Ethics
- Algorithmic Fairness through the Lens of Time
- Attributing Model Behavior at Scale (ATTRIB)
- Backdoors in Deep Learning: The Good, the Bad, and the Ugly
- Computational Sustainability: Promises and Pitfalls from Theory to Deployment
- I Can’t Believe It’s Not Better (ICBINB): Failure Modes in the Age of Foundation Models
- Socially Responsible Language Modelling Research (SoLaR)
- Regulatable ML: Towards Bridging the Gaps between Machine Learning Research and Regulations
- Workshop on Distribution Shifts: New Frontiers with Foundation Models
- XAI in Action: Past, Present, and Future Applications
- 2022:
- Oxford Generative AI Summit Slides
This section serves as a repository for policy documents, regulations, guidelines, and recommendations that govern the ethical and responsible use of artificial intelligence and machine learning technologies. From international legal frameworks to specific national laws, the resources cover a broad spectrum of topics such as fairness, privacy, ethics, and governance.
- 12 CFR Part 1002 - Equal Credit Opportunity Act (Regulation B)
- Aiming for truth, fairness, and equity in your company’s use of AI
- Algorithmic Accountability Act of 2023
- Algorithm Charter for Aotearoa New Zealand
- A Regulatory Framework for AI: Recommendations for PIPEDA Reform
- Artificial Intelligence (AI) in the Securities Industry
- Assessment List for Trustworthy Artificial Intelligence (ALTAI) for self-assessment - Shaping Europe’s digital future - European Commission
- Audit of Governance and Protection of Department of Defense Artificial Intelligence Data and Technology
- A Primer on Artificial Intelligence in Securities Markets
- Biometric Information Privacy Act
- Booker Wyden Health Care Letters
- California Consumer Privacy Act (CCPA)
- California Department of Justice, How to Read a Privacy Policy
- California Privacy Rights Act (CPRA)
- Can’t lose what you never had: Claims about digital ownership and creation in the age of generative AI
- Children's Online Privacy Protection Rule ("COPPA")
- Civil liability regime for artificial intelligence
- Congressional Research Service, Artificial Intelligence: Overview, Recent Advances, and Considerations for the 118th Congress
- Consumer Data Protection Act (Code of Virginia)
- Consumer Financial Protection Bureau (CFPB), Chatbots in consumer finance
- DARPA, Explainable Artificial Intelligence (XAI) (Archived)
- Data Availability and Transparency Act 2022 (Australia)
- data.gov, Privacy Policy and Data Policy
- Defense Technical Information Center, Computer Security Technology Planning Study, October 1, 1972
- De-identification Tools
- Department for Science, Innovation and Technology, The Bletchley Declaration by Countries Attending the AI Safety Summit, 1-2 November 2023
- Department for Science, Innovation and Technology, Frontier AI: capabilities and risks - discussion paper (United Kingdom)
- Department of Commerce, Intellectual property
- Department of Defense, AI Principles: Recommendations on the Ethical Use of Artificial Intelligence
- Department of Defense, Chief Data and Artificial Intelligence Officer (CDAO) Assessment and Assurance
- Developing Financial Sector Resilience in a Digital World: Selected Themes in Technology and Related Risks
- The Digital Services Act package (EU Digital Services Act and Digital Markets Act)
- Directive on Automated Decision Making (Canada)
- EEOC Letter (from U.S. senators re: hiring software)
- European Commission, Hiroshima Process International Guiding Principles for Advanced AI system
- Executive Order 13960 (2020-12-03), Promoting the Use of Trustworthy Artificial Intelligence in the Federal Government
- Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence
- Facial Recognition and Biometric Technology Moratorium Act of 2020
- FDA Artificial Intelligence/Machine Learning (AI/ML)-Based: Software as a Medical Device (SaMD) Action Plan, updated January 2021
- FDA Software as a Medical Device (SAMD) guidance (December 8, 2017)
- FDIC Supervisory Guidance on Model Risk Management
- Federal Consumer Online Privacy Rights Act (COPRA)
- Federal Reserve Bank of Dallas, Regulation B, Equal Credit Opportunity, Credit Scoring Interpretations: Withdrawl of Proposed Business Credit Amendments, June 3, 1982
- FHA model risk management/model governance guidance
- FTC Business Blog:
- 2020-04-08 Using Artificial Intelligence and Algorithms
- 2021-01-11 Facing the facts about facial recognition
- 2021-04-19 Aiming for truth, fairness, and equity in your company’s use of AI
- 2022-07-11 Location, health, and other sensitive information: FTC committed to fully enforcing the law against illegal use and sharing of highly sensitive data
- 2023-07-25 Protecting the privacy of health information: A baker’s dozen takeaways from FTC cases
- 2023-08-16 Can’t lose what you never had: Claims about digital ownership and creation in the age of generative AI
- 2023-08-22 For business opportunity sellers, FTC says “AI” stands for “allegedly inaccurate”
- 2023-09-15 Updated FTC-HHS publication outlines privacy and security laws and rules that impact consumer health data
- 2023-09-18 Companies warned about consequences of loose use of consumers’ confidential data
- 2023-09-27 Could PrivacyCon 2024 be the place to present your research on AI, privacy, or surveillance?
- 2022-05-20 Security Beyond Prevention: The Importance of Effective Breach Disclosures
- 2023-02-01 Security Principles: Addressing underlying causes of risk in complex systems
- 2023-06-29 Generative AI Raises Competition Concerns
- 2023-12-19 Coming face to face with Rite Aid’s allegedly unfair use of facial recognition technology
- FTC Privacy Policy
- Government Accountability Office: Artificial Intelligence: An Accountability Framework for Federal Agencies and Other Entities
- General Data Protection Regulation (GDPR)
- General principles for the use of Artificial Intelligence in the financial sector
- Gouvernance des algorithmes d’intelligence artificielle dans le secteur financier (France)
- Guidelines for secure AI system development
- IAPP Global AI Legislation Tracker
- IAPP US State Privacy Legislation Tracker
- Innovation spotlight: Providing adverse action notices when using AI/ML models
- Justice in Policing Act
- National Conference of State Legislatures (NCSL) 2020 Consumer Data Privacy Legislation
- National Institute of Standards and Technology (NIST), AI 100-1 Artificial Intelligence Risk Management Framework (NIST AI RMF 1.0)
- National Institute of Standards and Technology (NIST), Four Principles of Explainable Artificial Intelligence, Draft NISTIR 8312, 2020-08-17
- National Institute of Standards and Technology (NIST), Four Principles of Explainable Artificial Intelligence, NISTIR 8312, 2021-09-29
- National Institute of Standards and Technology (NIST), Measurement Uncertainty
- National Institute of Standards and Technology (NIST), NIST Special Publication 800-30 Revision 1, Guide for Conducting Risk Assessments
- National Science and Technology Council (NSTC), Select Committee on Artificial Intelligence, National Artificial Intelligence Research and Development Strategic Plan 2023 Update
- New York City Automated Decision Systems Task Force Report (November 2019)
- OECD, Open, Useful and Re-usable data (OURdata) Index: 2019 - Policy Paper
- Office of the Director of National Intelligence (ODNI), The AIM Initiative: A Strategy for Augmenting Intelligence Using Machines
- Office of Management and Budget, Guidance for Regulation of Artificial Intelligence Applications, finalized November 2020
- Office of Science and Technology Policy, Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People
- Office of the Comptroller of the Currency (OCC), 2021 Model Risk Management Handbook
- Online Harms White Paper: Full government response to the consultation (United Kingdom)
- Online Privacy Act of 2023
- Online Safety Bill (United Kingdom)
- Principles of Artificial Intelligence Ethics for the Intelligence Community
- Privacy Act 1988 (Australia)
- Proposal for a Regulation laying down harmonised rules on artificial intelligence (Artificial Intelligence Act)
- Psychological Foundations of Explainability and Interpretability in Artificial Intelligence
- The Public Sector Bodies (Websites and Mobile Applications) Accessibility Regulations 2018 (United Kingdom)
- Questions and Answers to Clarify and Provide a Common Interpretation of the Uniform Guidelines on Employee Selection Procedures
- Questions from the Commission on Protecting Privacy and Preventing Discrimination
- RE: Use of External Consumer Data and Information Sources in Underwriting for Life Insurance
- Singapore's Companion to the Model AI Governance Framework – Implementation and Self-Assessment Guide for Organizations
- Singapore's Compendium of Use Cases: Practical Illustrations of the Model AI Governance Framework
- Singapore's Model Artificial Intelligence Governance Framework (Second Edition)
- Supervisory Guidance on Model Risk Management
- Testing the Reliability, Validity, and Equity of Terrorism Risk Assessment Instruments
- UNESCO, Artificial Intelligence: examples of ethical dilemmas
- United States Department of Energy Artificial Intelligence and Technology Office:
- United States Department of Homeland Security, Use of Commercial Generative Artificial Intelligence (AI) Tools
- United States Department of Justice, Privacy Act of 1974
- United States Department of Justice, Overview of The Privacy Act of 1974 (2020 Edition)
- United States Patent and Trademark Office (USPTO), Public Views on Artificial Intelligence and Intellectual Property Policy
- Using Artificial Intelligence and Algorithms
- U.S. Army Concepts Analysis Agency, Proceedings of the Thirteenth Annual U.S. Army Operations Research Symposium, Volume 1, October 29 to November 1, 1974
- U.S. Web Design System (USWDS) Design principles
This section is a curated collection of guides and tutorials that simplify responsible ML implementation. It spans from basic model interpretability to advanced fairness techniques. Suitable for both novices and experts, the resources cover topics like COMPAS fairness analyses and explainable machine learning via counterfactuals.
- COMPAS Analysis Using Aequitas
- Explaining Quantitative Measures of Fairness (with SHAP)
- Getting a Window into your Black Box Model
- H20.ai, From GLM to GBM Part 1
- H20.ai, From GLM to GBM Part 2
- IML
- Interpretable Machine Learning with Python
- Interpreting Machine Learning Models with the iml Package
- Interpretable Machine Learning using Counterfactuals
- Machine Learning Explainability by Kaggle Learn
- Model Interpretability with DALEX
- Model Interpretation series by Dipanjan (DJ) Sarkar:
- Partial Dependence Plots in R
- PiML:
- Reliable-and-Trustworthy-AI-Notebooks
- Saliency Maps for Deep Learning
- Visualizing ML Models with LIME
- Visualizing and debugging deep convolutional networks
- What does a CNN see?
This section contains books that can be reasonably described as free, including some "historical" books dealing broadly with ethical and responsible tech.
- César A. Hidalgo, Diana Orghian, Jordi Albo-Canals, Filipa de Almeida, and Natalia Martin, 2021, How Humans Judge Machines
- Charles Perrow, 1984, Normal Accidents: Living with High-Risk Technologies
- Charles Perrow, 1999, Normal Accidents: Living with High-Risk Technologies with a New Afterword and a Postscript on the Y2K Problem
- Christoph Molnar, 2021, Interpretable Machine Learning: A Guide for Making Black Box Models Explainable
- Deborah G. Johnson and Keith W. Miller, 2009, Computer Ethics: Analyzing Information Technology, Fourth Edition
- Ed Dreby and Keith Helmuth (contributors) and Judy Lumb (editor), 2009, Fueling Our Future: A Dialogue about Technology, Ethics, Public Policy, and Remedial Action
- George Reynolds, 2002, Ethics in Information Technology
- George Reynolds, 2002, Ethics in Information Technology, Instructor's Edition
- Kenneth Vaux (editor), 1970, Who Shall Live? Medicine, Technology, Ethics
- Kush R. Varshney, 2022, Trustworthy Machine Learning: Concepts for Developing Accurate, Fair, Robust, Explainable, Transparent, Inclusive, Empowering, and Beneficial Machine Learning Systems
- Marsha Cook Woodbury, 2003, Computer and Information Ethics
- M. David Ermann, Mary B. Williams, and Claudio Gutierrez, 1990, Computers, Ethics, and Society
- Morton E. Winston and Ralph D. Edelbach, 2000, Society, Ethics, and Technology, First Edition
- Morton E. Winston and Ralph D. Edelbach, 2003, Society, Ethics, and Technology, Second Edition
- Morton E. Winston and Ralph D. Edelbach, 2006, Society, Ethics, and Technology, Third Edition
- Patrick Hall and Navdeep Gill, 2019, An Introduction to Machine Learning Interpretability: An Applied Perspective on Fairness, Accountability, Transparency, and Explainable AI, Second Edition
- Patrick Hall, Navdeep Gill, and Benjamin Cox, 2021, Responsible Machine Learning: Actionable Strategies for Mitigating Risks & Driving Adoption
- Patrick Hall, James Curtis, Parul Pandey, and Agus Sudjianto, 2023, Machine Learning for High-Risk Applications: Approaches to Responsible AI
- Paula Boddington, 2017, Towards a Code of Ethics for Artificial Intelligence
- Przemyslaw Biecek and Tomasz Burzykowski, 2020, Explanatory Model Analysis: Explore, Explain, and Examine Predictive Models. With examples in R and Python
- Przemyslaw Biecek, 2023, Adversarial Model Analysis
- Raymond E. Spier (editor), 2003, Science and Technology Ethics
- Richard A. Spinello, 1995, Ethical Aspects of Information Technology
- Richard A. Spinello, 1997, Case Studies in Information and Computer Ethics
- Richard A. Spinello, 2003, Case Studies in Information Technology Ethics, Second Edition
- Solon Barocas, Moritz Hardt, and Arvind Narayanan, 2022, Fairness and Machine Learning: Limitations and Opportunities
- Soraj Hongladarom and Charles Ess, 2007, Information Technology Ethics: Cultural Perspectives
- Stephen H. Unger, 1982, Controlling Technology: Ethics and the Responsible Engineer, First Edition
- Stephen H. Unger, 1994, Controlling Technology: Ethics and the Responsible Engineer, Second Edition
This section features a collection of glossaries and dictionaries that are geared toward defining terms in ML, including some "historical" dictionaries.
- A.I. For Anyone: The A-Z of AI
- Alan Turing Institute: Data science and AI glossary
- Appen Artificial Intelligence Glossary
- Brookings: The Brookings glossary of AI and emerging technologies
- Built In, Responsible AI Explained
- Center for Security and Emerging Technology: Glossary
- CompTIA: Artificial Intelligence (AI) Terminology: A Glossary for Beginners
- Council of Europe Artificial Intelligence Glossary
- Coursera: Artificial Intelligence (AI) Terms: A to Z Glossary
- Dataconomy: AI dictionary: Be a native speaker of Artificial Intelligence
- Dennis Mercadal, 1990, Dictionary of Artificial Intelligence
- G2: 70+ A to Z Artificial Intelligence Terms in Technology
- General Services Administration: AI Guide for Government: Key AI terminology
- Google Developers Machine Learning Glossary
- H2O.ai Glossary
- IAPP Glossary of Privacy Terms
- IAPP International Definitions of Artificial Intelligence
- IAPP Key Terms for AI Governance
- IBM: AI glossary
- ISO: Information technology — Artificial intelligence — Artificial intelligence concepts and terminology
- Jerry M. Rosenberg, 1986, Dictionary of Artificial Intelligence & Robotics
- MakeUseOf: A Glossary of AI Jargon: 29 AI Terms You Should Know
- Moveworks: AI Terms Glossary
- NIST AIRC: The Language of Trustworthy AI: An In-Depth Glossary of Terms
- Oliver Houdé, 2004, Dictionary of Cognitive Science: Neuroscience, Psychology, Artificial Intelligence, Linguistics, and Philosophy
- Otto Vollnhals, 1992, A Multilingual Dictionary of Artificial Intelligence (English, German, French, Spanish, Italian)
- Raoul Smith, 1989, The Facts on File Dictionary of Artificial Intelligence
- Raoul Smith, 1990, Collins Dictionary of Artificial Intelligence
- Salesforce: AI From A to Z: The Generative AI Glossary for Business Leaders
- Stanford University HAI Artificial Intelligence Definitions
- TechTarget: Artificial intelligence glossary: 60+ terms to know
- TELUS International: 50 AI terms every beginner should know
- University of New South Wales, Bill Wilson, The Machine Learning Dictionary
- VAIR (Vocabulary of AI Risks)
- Wikipedia: Glossary of artificial intelligence
- William J. Raynor, Jr, 1999, The International Dictionary of Artificial Intelligence, First Edition
- William J. Raynor, Jr, 2009, International Dictionary of Artificial Intelligence, Second Edition
This section features a selection of educational courses focused on ethical considerations and best practices in ML. The classes range from introductory courses on data ethics to specialized training in fairness and trustworthy deep learning.
- An Introduction to Data Ethics
- Certified Ethical Emerging Technologist
- Coursera, DeepLearning.AI, Generative AI for Everyone
- Coursera, DeepLearning.AI, Generative AI with Large Language Models
- Coursera, Google Cloud, Introduction to Generative AI
- Coursera, Vanderbilt University, Prompt Engineering for ChatGPT
- CS103F: Ethical Foundations of Computer Science
- ETH Zürich ReliableAI 2022 Course Project repository
- Fairness in Machine Learning
- Fast.ai Data Ethics course
- Human-Centered Machine Learning
- Introduction to AI Ethics
- INFO 4270: Ethics and Policy in Data Science
- Introduction to Responsible Machine Learning
- Machine Learning Fairness by Google
- Trustworthy Deep Learning
This section houses initiatives, networks, repositories, and publications that facilitate collective and interdisciplinary efforts to enhance AI safety. It includes platforms where experts and practitioners come together to share insights, identify potential vulnerabilities, and collaborate on developing robust safeguards for AI systems, including AI incident trackers.
- AI Incident Database (Responsible AI Collaborative)
- AI Vulnerability Database (AVID)
- AIAAIC
- George Washington University Law School's AI Litigation Database
- OECD AI Incidents Monitor
- Verica Open Incident Database (VOID)
This section contains challenges and competitions related to responsible ML.
- FICO Explainable Machine Learning Challenge
- National Fair Housing Alliance Hackathon
- Twitter Algorithmic Bias
We are seeking curated bibliographies related to responsible ML across various topics, see issue 115.
-
BibTeX:
-
Web:
This section links to other lists of responsible ML or related resources.
- A Living and Curated Collection of Explainable AI Methods
- AI Ethics Guidelines Global Inventory
- AI Ethics Resources
- AI Tools and Platforms
- Awesome AI Guidelines
- Awesome interpretable machine learning
- Awesome-explainable-AI
- Awesome-ML-Model-Governance
- Awesome MLOps
- Awesome Production Machine Learning
- Awful AI
- criticalML
- Evaluation Repository for 'Sociotechnical Safety Evaluation of Generative AI Systems'
- IMDA-BTG, LLM-Evals-Catalogue
- Machine Learning Ethics References
- Machine Learning Interpretability Resources
- OECD-NIST Catalogue of AI Tools and Metrics
- OpenAI Cookbook
- private-ai-resources
- ResponsibleAI
- Worldwide AI ethics: A review of 200 guidelines and recommendations for AI governance
- XAI Resources
- xaience
This section contains benchmarks or datasets used for benchmarks for ML systems, particularly those related to responsible ML desiderata.
Resource | Description |
---|---|
benchm-ml | "A minimal benchmark for scalability, speed and accuracy of commonly used open source implementations (R packages, Python scikit-learn, H2O, xgboost, Spark MLlib etc.) of the top machine learning algorithms for binary classification (random forests, gradient boosted trees, deep neural networks etc.)." |
Bias Benchmark for QA dataset (BBQ) | "Repository for the Bias Benchmark for QA dataset." |
GEM | "GEM is a benchmark environment for Natural Language Generation with a focus on its Evaluation, both through human annotations and automated Metrics." |
HELM | "A holistic framework for evaluating foundation models." |
i-gallegos, Fair-LLM-Benchmark | "Bias and Fairness in Large Language Models: A Survey" |
Nvidia MLPerf | "MLPerf™ benchmarks—developed by MLCommons, a consortium of AI leaders from academia, research labs, and industry—are designed to provide unbiased evaluations of training and inference performance for hardware, software, and services." |
OpenML Benchmarking Suites | OpenML's collection of over two dozen benchmarking suites. |
TrustLLM-Benchmark | "A Comprehensive Study of Trustworthiness in Large Language Models." |
Trust-LLM-Benchmark Leaderboard | A series of sortable leaderboards of LLMs based on different trustworthiness criteria. |
TruthfulQA | "TruthfulQA: Measuring How Models Imitate Human Falsehoods." |
Winogender Schemas | "Data for evaluating gender bias in coreference resolution systems." |
Real Toxicity Prompts (Allen Institute for AI) | "A dataset of 100k sentence snippets from the web for researchers to further address the risk of neural toxic degeneration in models." |
This section contains datasets that are commonly used in responsible ML evaulations or repositories of interesting/important data sources:
- Adult income dataset
- Balanced Faces in the Wild
- COMPAS Recidivism Risk Score Data and Analysis
- FANNIE MAE Single Family Loan Performance
- NYPD Stop, Question and Frisk Data
- socialfoundations / folktables
- Statlog (German Credit Data)
- Wikipedia Talk Labels: Personal Attacks
This section curates specialized software tools aimed at responsible ML within specific domains, such as in healthcare, finance, or social sciences.
This section contains open source or open access ML environment management software.
Resource | Description |
---|---|
dvc | "Manage and version images, audio, video, and text files in storage and organize your ML modeling process into a reproducible workflow." |
gigantum | "Building a better way to create, collaborate, and share data-driven science." |
mlflow | "An open source platform for the machine learning lifecycle." |
mlmd | "For recording and retrieving metadata associated with ML developer and data scientist workflows." |
modeldb | "Open Source ML Model Versioning, Metadata, and Experiment Management." |
neptune | "A single place to manage all your model metadata." |
This section contains open source or open access software used to implement responsible ML. As much as possible, descriptions are quoted verbatim from the respective repositories themselves. In rare instances, we provide our own descriptions (unmarked by quotes).
Name | Description |
---|---|
DiscriLens | "Discrimination in Machine Learning." |
Hugging Face, BiasAware: Dataset Bias Detection | "BiasAware is a specialized tool for detecting and quantifying biases within datasets used for Natural Language Processing (NLP) tasks." |
manifold | "A model-agnostic visual debugging tool for machine learning." |
PAIR-code / datacardsplaybook | "The Data Cards Playbook helps dataset producers and publishers adopt a people-centered approach to transparency in dataset documentation." |
PAIR-code / facets | "Visualizations for machine learning datasets." |
PAIR-code / knowyourdata | "A tool to help researchers and product teams understand datasets with the goal of improving data quality, and mitigating fairness and bias issues." |
TensorBoard Projector | "Using the TensorBoard Embedding Projector, you can graphically represent high dimensional embeddings. This can be helpful in visualizing, examining, and understanding your embedding layers." |
What-if Tool | "Visually probe the behavior of trained machine learning models, with minimal coding." |
Name | Description |
---|---|
Born-again Tree Ensembles | "Born-Again Tree Ensembles: Transforms a random forest into a single, minimal-size, tree with exactly the same prediction function in the entire feature space (ICML 2020)." |
Certifiably Optimal RulE ListS | "CORELS is a custom discrete optimization technique for building rule lists over a categorical feature space." |
Secure-ML | "Secure Linear Regression in the Semi-Honest Two-Party Setting." |
Name | Description |
---|---|
LDNOOBW | "List of Dirty, Naughty, Obscene, and Otherwise Bad Words" |
Name | Description |
---|---|
acd | "Produces hierarchical interpretations for a single prediction made by a pytorch neural network. Official code for Hierarchical interpretations for neural network predictions.” |
aequitas | "Aequitas is an open-source bias audit toolkit for data scientists, machine learning researchers, and policymakers to audit machine learning models for discrimination and bias, and to make informed and equitable decisions around developing and deploying predictive tools.” |
AI Fairness 360 | "A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.” |
AI Explainability 360 | "Interpretability and explainability of data and machine learning models.” |
ALEPython | "Python Accumulated Local Effects package.” |
Aletheia | "A Python package for unwrapping ReLU DNNs.” |
allennlp | "An open-source NLP research library, built on PyTorch.” |
algofairness | See [Algorithmic Fairness][http://fairness.haverford.edu/). |
Alibi | "Alibi is an open source Python library aimed at machine learning model inspection and interpretation. The focus of the library is to provide high-quality implementations of black-box, white-box, local and global explanation methods for classification and regression models.” |
anchor | "Code for 'High-Precision Model-Agnostic Explanations' paper.” |
Bayesian Case Model | |
Bayesian Ors-Of-Ands | "This code implements the Bayesian or-of-and algorithm as described in the BOA paper. We include the tictactoe dataset in the correct formatting to be used by this code.” |
Bayesian Rule List (BRL) | Rudin group at Duke Bayesian case model implementation |
BlackBoxAuditing | "Research code for auditing and exploring black box machine-learning models.” |
casme | "contains the code originally forked from the ImageNet training in PyTorch that is modified to present the performance of classifier-agnostic saliency map extraction, a practical algorithm to train a classifier-agnostic saliency mapping by simultaneously training a classifier and a saliency mapping.” |
Causal Discovery Toolbox | "Package for causal inference in graphs and in the pairwise settings. Tools for graph structure recovery and dependencies are included.” |
captum | "Model interpretability and understanding for PyTorch.” |
causalml | "Uplift modeling and causal inference with machine learning algorithms.” |
cdt15, Causal Discovery Lab., Shiga University | "LiNGAM is a new method for estimating structural equation models or linear causal Bayesian networks. It is based on using the non-Gaussianity of the data." |
checklist | "Beyond Accuracy: Behavioral Testing of NLP models with CheckList.” |
cleverhans | "An adversarial example library for constructing attacks, building defenses, and benchmarking both.” |
contextual-AI | "Contextual AI adds explainability to different stages of machine learning pipelines |
ContrastiveExplanation (Foil Trees) | "provides an explanation for why an instance had the current outcome (fact) rather than a targeted outcome of interest (foil). These counterfactual explanations limit the explanation to the features relevant in distinguishing fact from foil, thereby disregarding irrelevant features.” |
counterfit | "a CLI that provides a generic automation layer for assessing the security of ML models.” |
dalex | "moDel Agnostic Language for Exploration and eXplanation.” |
debiaswe | "Remove problematic gender bias from word embeddings.” |
DeepExplain | "provides a unified framework for state-of-the-art gradient and perturbation-based attribution methods. It can be used by researchers and practitioners for better undertanding the recommended existing models, as well for benchmarking other attribution methods.” |
DeepLIFT | "This repository implements the methods in 'Learning Important Features Through Propagating Activation Differences' by Shrikumar, Greenside & Kundaje, as well as other commonly-used methods such as gradients, gradient-times-input (equivalent to a version of Layerwise Relevance Propagation for ReLU networks), guided backprop and integrated gradients.” |
deepvis | "the code required to run the Deep Visualization Toolbox, as well as to generate the neuron-by-neuron visualizations using regularized optimization.” |
DIANNA | "DIANNA is a Python package that brings explainable AI (XAI) to your research project. It wraps carefully selected XAI methods in a simple, uniform interface. It's built by, with and for (academic) researchers and research software engineers working on machine learning projects.” |
DiCE | "Generate Diverse Counterfactual Explanations for any machine learning model.” |
DoWhy | "DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.” |
dtreeviz | "A python library for decision tree visualization and model interpretation.” |
ecco | "Explain, analyze, and visualize NLP language models. Ecco creates interactive visualizations directly in Jupyter notebooks explaining the behavior of Transformer-based language models (like GPT2, BERT, RoBERTA, T5, and T0).” |
eli5 | "A library for debugging/inspecting machine learning classifiers and explaining their predictions.” |
explabox | "aims to support data scientists and machine learning (ML) engineers in explaining, testing and documenting AI/ML models, developed in-house or acquired externally. The explabox turns your ingestibles (AI/ML model and/or dataset) into digestibles (statistics, explanations or sensitivity insights).” |
Explainable Boosting Machine (EBM)/GA2M | "an open-source package that incorporates state-of-the-art machine learning interpretability techniques under one roof. With this package, you can train interpretable glassbox models and explain blackbox systems. InterpretML helps you understand your model's global behavior, or understand the reasons behind individual predictions.” |
ExplainaBoard | "a tool that inspects your system outputs, identifies what is working and what is not working, and helps inspire you with ideas of where to go next.” |
explainerdashboard | "Quickly build Explainable AI dashboards that show the inner workings of so-called "blackbox" machine learning models.” |
explainX | "Explainable AI framework for data scientists. Explain & debug any blackbox machine learning model with a single line of code.” |
fair-classification | "Python code for training fair logistic regression classifiers.” |
fairml | "a python toolbox auditing the machine learning models for bias.” |
fairlearn | "a Python package that empowers developers of artificial intelligence (AI) systems to assess their system's fairness and mitigate any observed unfairness issues. Fairlearn contains mitigation algorithms as well as metrics for model assessment. Besides the source code, this repository also contains Jupyter notebooks with examples of Fairlearn usage.” |
fairness-comparison | "meant to facilitate the benchmarking of fairness aware machine learning algorithms.” |
fairness_measures_code | "contains implementations of measures used to quantify discrimination.” |
Falling Rule List (FRL) | Rudin group at Duke falling rule list implementation |
foolbox | "A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX.” |
Giskard | "The testing framework dedicated to ML models, from tabular to LLMs. Scan AI models to detect risks of biases, performance issues and errors. In 4 lines of code.” |
Grad-CAM (GitHub topic) | Grad-CAM is a technique for making convolutional neural networks more transparent by visualizing the regions of input that are important for predictions in computer vision models. |
gplearn | "implements Genetic Programming in Python, with a scikit-learn inspired and compatible API.” |
H2O-3 Penalized Generalized Linear Models | "Fits a generalized linear model, specified by a response variable, a set of predictors, and a description of the error distribution." |
H2O-3 Monotonic GBM | "Builds gradient boosted classification trees and gradient boosted regression trees on a parsed data set." |
H2O-3 Sparse Principal Components (GLRM) | "Builds a generalized low rank decomposition of an H2O data frame." |
h2o-LLM-eval | "Large-language Model Evaluation framework with Elo Leaderboard and A-B testing." |
hate-functional-tests | HateCheck: A dataset and test suite from an ACL 2021 paper, offering functional tests for hate speech detection models, including extensive case annotations and testing functionalities. |
imodels | "Python package for concise, transparent, and accurate predictive modeling. All sklearn-compatible and easy to use.” |
iNNvestigate neural nets | A comprehensive Python library to analyze and interpret neural network behaviors in Keras, featuring a variety of methods like Gradient, LRP, and Deep Taylor. |
Integrated-Gradients | "a variation on computing the gradient of the prediction output w.r.t. features of the input. It requires no modification to the original network, is simple to implement, and is applicable to a variety of deep models (sparse and dense, text and vision).” |
interpret | "an open-source package that incorporates state-of-the-art machine learning interpretability techniques under one roof.” |
interpret_with_rules | "induces rules to explain the predictions of a trained neural network, and optionally also to explain the patterns that the model captures from the training data, and the patterns that are present in the original dataset.” |
InterpretME | "integrates knowledge graphs (KG) with machine learning methods to generate interesting meaningful insights. It helps to generate human- and machine-readable decisions to provide assistance to users and enhance efficiency.” |
Keras-vis | "a high-level toolkit for visualizing and debugging your trained keras neural net models.” |
keract | Keract is a tool for visualizing activations and gradients in Keras models; it's meant to support a wide range of Tensorflow versions and to offer an intuitive API with Python examples. |
L2X | "Code for replicating the experiments in the paper Learning to Explain: An Information-Theoretic Perspective on Model Interpretation at ICML 2018, by Jianbo Chen, Mitchell Stern, Martin J. Wainwright, Michael I. Jordan.” |
langtest | "LangTest: Deliver Safe & Effective Language Models" |
learning-fair-representations | "Python numba implementation of Zemel et al. 2013 http://www.cs.toronto.edu/~toni/Papers/icml-final.pdf" |
lime | "explaining what machine learning classifiers (or models) are doing. At the moment, we support explaining individual predictions for text classifiers or classifiers that act on tables (numpy arrays of numerical or categorical data) or images, with a package called lime (short for local interpretable model-agnostic explanations).” |
LiFT | "The LinkedIn Fairness Toolkit (LiFT) is a Scala/Spark library that enables the measurement of fairness and the mitigation of bias in large-scale machine learning workflows. The measurement module includes measuring biases in training data, evaluating fairness metrics for ML models, and detecting statistically significant differences in their performance across different subgroups.” |
lit | "The Learning Interpretability Tool (LIT, formerly known as the Language Interpretability Tool) is a visual, interactive ML model-understanding tool that supports text, image, and tabular data. It can be run as a standalone server, or inside of notebook environments such as Colab, Jupyter, and Google Cloud Vertex AI notebooks.” |
lofo-importance | "LOFO (Leave One Feature Out) Importance calculates the importances of a set of features based on a metric of choice, for a model of choice, by iteratively removing each feature from the set, and evaluating the performance of the model, with a validation scheme of choice, based on the chosen metric.” |
lrp_toolbox | "The Layer-wise Relevance Propagation (LRP) algorithm explains a classifer's prediction specific to a given data point by attributing relevance scores to important components of the input by using the topology of the learned model itself.” |
MindsDB | "enables developers to build AI tools that need access to real-time data to perform their tasks.” |
MLextend | "Mlxtend (machine learning extensions) is a Python library of useful tools for the day-to-day data science tasks.” |
ml-fairness-gym | "a set of components for building simple simulations that explore the potential long-run impacts of deploying machine learning-based decision systems in social environments.” |
ml_privacy_meter | "an open-source library to audit data privacy in statistical and machine learning algorithms. The tool can help in the data protection impact assessment process by providing a quantitative analysis of the fundamental privacy risks of a (machine learning) model.” |
mllp | "This is a PyTorch implementation of Multilayer Logical Perceptrons (MLLP) and Random Binarization (RB) method to learn Concept Rule Sets (CRS) for transparent classification tasks, as described in our paper: Transparent Classification with Multilayer Logical Perceptrons and Random Binarization.” |
Monotonic Constraints | Guide on implementing and understanding monotonic constraints in XGBoost models to enhance predictive performance with practical Python examples. |
XGBoost | "an optimized distributed gradient boosting library designed to be highly efficient, flexible and portable.” |
Multilayer Logical Perceptron (MLLP) | "This is a PyTorch implementation of Multilayer Logical Perceptrons (MLLP) and Random Binarization (RB) method to learn Concept Rule Sets (CRS) for transparent classification tasks, as described in our paper: Transparent Classification with Multilayer Logical Perceptrons and Random Binarization.” |
OptBinning | "a library written in Python implementing a rigorous and flexible mathematical programming formulation to solve the optimal binning problem for a binary, continuous and multiclass target type, incorporating constraints not previously addressed.” |
Optimal Sparse Decision Trees | "This accompanies the paper, "Optimal Sparse Decision Trees" by Xiyang Hu, Cynthia Rudin, and Margo Seltzer.” |
parity-fairness | "This repository contains codes that demonstrate the use of fairness metrics, bias mitigations and explainability tool.” |
PDPbox | "Python Partial Dependence Plot toolbox. Visualize the influence of certain features on model predictions for supervised machine learning algorithms, utilizing partial dependence plots.” |
PiML-Toolbox | "a new Python toolbox for interpretable machine learning model development and validation. Through low-code interface and high-code APIs, PiML supports a growing list of inherently interpretable ML models.” |
Privacy-Preserving-ML | "Implementation of privacy-preserving SVM assuming public model private data scenario (data in encrypted but model parameters are unencrypted) using adequate partial homomorphic encryption.” |
ProtoPNet | "This code package implements the prototypical part network (ProtoPNet) from the paper "This Looks Like That: Deep Learning for Interpretable Image Recognition" (to appear at NeurIPS 2019), by Chaofan Chen (Duke University), Oscar Li |
pyBreakDown | See dalex. |
PyCEbox | "Python Individual Conditional Expectation Plot Toolbox.” |
pyGAM | "Generalized Additive Models in Python.” |
pymc3 | "PyMC (formerly PyMC3) is a Python package for Bayesian statistical modeling focusing on advanced Markov chain Monte Carlo (MCMC) and variational inference (VI) algorithms. Its flexibility and extensibility make it applicable to a large suite of problems.” |
pySS3 | "The SS3 text classifier is a novel and simple supervised machine learning model for text classification which is interpretable, that is, it has the ability to naturally (self)explain its rationale.” |
pytorch-grad-cam | "a package with state of the art methods for Explainable AI for computer vision. This can be used for diagnosing model predictions, either in production or while developing models. The aim is also to serve as a benchmark of algorithms and metrics for research of new explainability methods.” |
pytorch-innvestigate | "PyTorch implementation of Keras already existing project: https://github.com/albermax/innvestigate/.” |
Quantus | "Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanations." |
rationale | "This directory contains the code and resources of the following paper: "Rationalizing Neural Predictions". Tao Lei, Regina Barzilay and Tommi Jaakkola. EMNLP 2016. [PDF] [Slides]. The method learns to provide justifications, i.e. rationales, as supporting evidence of neural networks' prediction.” |
responsibly | "Toolkit for Auditing and Mitigating Bias and Fairness of Machine Learning Systems.” |
REVISE: REvealing VIsual biaSEs | "A tool that automatically detects possible forms of bias in a visual dataset along the axes of object-based, attribute-based, and geography-based patterns, and from which next steps for mitigation are suggested.” |
robustness | "a package we (students in the MadryLab) created to make training, evaluating, and exploring neural networks flexible and easy.” |
RISE | "contains source code necessary to reproduce some of the main results in the paper: Vitali Petsiuk, Abir Das, Kate Saenko (BMVC, 2018) [and] RISE: Randomized Input Sampling for Explanation of Black-box Models.” |
Risk-SLIM | "a machine learning method to fit simple customized risk scores in python.” |
SAGE | "SAGE (Shapley Additive Global importancE) is a game-theoretic approach for understanding black-box machine learning models. It quantifies each feature's importance based on how much predictive power it contributes, and it accounts for complex feature interactions using the Shapley value.” |
SALib | "Python implementations of commonly used sensitivity analysis methods. Useful in systems modeling to calculate the effects of model inputs or exogenous factors on outputs of interest.” |
Scikit-Explain | "User-friendly Python module for machine learning explainability," featuring PD and ALE plots, LIME, SHAP, permutation importance and Friedman's H, among other methods. |
Scikit-learn Decision Trees | "a non-parametric supervised learning method used for classification and regression.” |
Scikit-learn Generalized Linear Models | "a set of methods intended for regression in which the target value is expected to be a linear combination of the features.” |
Scikit-learn Sparse Principal Components | "a variant of [principal component analysis, PCA], with the goal of extracting the set of sparse components that best reconstruct the data.” |
scikit-fairness | Historical link. Merged with fairlearn. |
scikit-multiflow | "a machine learning package for streaming data in Python.” |
shap | "a game theoretic approach to explain the output of any machine learning model. It connects optimal credit allocation with local explanations using the classic Shapley values from game theory and their related extensions" |
shapley | "a Python library for evaluating binary classifiers in a machine learning ensemble.” |
sklearn-expertsys | "a scikit-learn compatible wrapper for the Bayesian Rule List classifier developed by Letham et al., 2015, extended by a minimum description length-based discretizer (Fayyad & Irani, 1993) for continuous data, and by an approach to subsample large datasets for better performance.” |
skope-rules | "a Python machine learning module built on top of scikit-learn and distributed under the 3-Clause BSD license.” |
solas-ai-disparity | "a collection of tools that allows modelers, compliance, and business stakeholders to test outcomes for bias or discrimination using widely accepted fairness metrics.” |
Super-sparse Linear Integer models (SLIMs) | "a package to learn customized scoring systems for decision-making problems.” |
tensorflow/lattice | "a library that implements constrained and interpretable lattice based models. It is an implementation of Monotonic Calibrated Interpolated Look-Up Tables in TensorFlow.” |
tensorflow/lucid | "a collection of infrastructure and tools for research in neural network interpretability.” |
tensorflow/fairness-indicators | "designed to support teams in evaluating, improving, and comparing models for fairness concerns in partnership with the broader Tensorflow toolkit.” |
tensorflow/model-analysis | "a library for evaluating TensorFlow models. It allows users to evaluate their models on large amounts of data in a distributed manner, using the same metrics defined in their trainer. These metrics can be computed over different slices of data and visualized in Jupyter notebooks.” |
tensorflow/model-card-toolkit | "streamlines and automates generation of Model Cards, machine learning documents that provide context and transparency into a model's development and performance. Integrating the MCT into your ML pipeline enables you to share model metadata and metrics with researchers, developers, reporters, and more.” |
tensorflow/model-remediation | "a library that provides solutions for machine learning practitioners working to create and train models in a way that reduces or eliminates user harm resulting from underlying performance biases.” |
tensorflow/privacy | "the source code for TensorFlow Privacy, a Python library that includes implementations of TensorFlow optimizers for training machine learning models with differential privacy. The library comes with tutorials and analysis tools for computing the privacy guarantees provided.” |
tensorflow/tcav | "Testing with Concept Activation Vectors (TCAV) is a new interpretability method to understand what signals your neural networks models uses for prediction.” |
tensorfuzz | "a library for performing coverage guided fuzzing of neural networks.” |
TensorWatch | "a debugging and visualization tool designed for data science, deep learning and reinforcement learning from Microsoft Research. It works in Jupyter Notebook to show real-time visualizations of your machine learning training and perform several other key analysis tasks for your models and data.” |
TextFooler | "A Model for Natural Language Attack on Text Classification and Inference" |
text_explainability | "text_explainability provides a generic architecture from which well-known state-of-the-art explainability approaches for text can be composed.” |
text_sensitivity | "Uses the generic architecture of text_explainability to also include tests of safety (how safe it the model in production, i.e. types of inputs it can handle), robustness (how generalizable the model is in production, e.g. stability when adding typos, or the effect of adding random unrelated data) and fairness (if equal individuals are treated equally by the model, e.g. subgroup fairness on sex and nationality).” |
tf-explain | "Implements interpretability methods as Tensorflow 2.x callbacks to ease neural network's understanding.” |
Themis | "A testing-based approach for measuring discrimination in a software system.” |
themis-ml | "A Python library built on top of pandas and sklearnthat implements fairness-aware machine learning algorithms.” |
TorchUncertainty | "A package designed to help you leverage uncertainty quantification techniques and make your deep neural networks more reliable.” |
treeinterpreter | "Package for interpreting scikit-learn's decision tree and random forest predictions.” |
TRIAGE | "This repository contains the implementation of TRIAGE, a "Data-Centric AI" framework for data characterization tailored for regression.” |
woe | "Tools for WoE Transformation mostly used in ScoreCard Model for credit rating.” |
xai | "A Machine Learning library that is designed with AI explainability in its core.” |
xdeep | "An open source Python library for Interpretable Machine Learning.” |
xplique | "A Python toolkit dedicated to explainability. The goal of this library is to gather the state of the art of Explainable AI to help you understand your complex neural network models.” |
ydata-profiling | "Provide[s] a one-line Exploratory Data Analysis (EDA) experience in a consistent and fast solution.” |
yellowbrick | "A suite of visual diagnostic tools called "Visualizers" that extend the scikit-learn API to allow human steering of the model selection process.” |
Name | Description |
---|---|
ALEPlot | "Visualizes the main effects of individual predictor variables and their second-order interaction effects in black-box supervised learning models." |
arules | "Provides the infrastructure for representing, manipulating and analyzing transaction data and patterns (frequent itemsets and association rules). Also provides C implementations of the association mining algorithms Apriori and Eclat. Hahsler, Gruen and Hornik (2005)." |
Causal SVM | "We present a new machine learning approach to estimate whether a treatment has an effect on an individual, in the setting of the classical potential outcomes framework with binary outcomes." |
DALEX | "moDel Agnostic Language for Exploration and eXplanation." |
DALEXtra: Extension for 'DALEX' Package | "Provides wrapper of various machine learning models." |
DrWhyAI | "DrWhy is [a] collection of tools for eXplainable AI (XAI). It's based on shared principles and simple grammar for exploration, explanation and visualisation of predictive models." |
elasticnet | "Provides functions for fitting the entire solution path of the Elastic-Net and also provides functions for doing sparse PCA." |
ExplainPrediction | "Generates explanations for classification and regression models and visualizes them." |
Explainable Boosting Machine (EBM)/GA2M | "Package for training interpretable machine learning models." |
fairmodels | "Flexible tool for bias detection, visualization, and mitigation. Use models explained with DALEX and calculate fairness classification metrics based on confusion matrices using fairness_check() or try newly developed module for regression models using fairness_check_regression()." |
fairness | "Offers calculation, visualization and comparison of algorithmic fairness metrics." |
fastshap | "The goal of fastshap is to provide an efficient and speedy approach (at least relative to other implementations) for computing approximate Shapley values, which help explain the predictions from any machine learning model." |
featureImportance | "An extension for the mlr package and allows to compute the permutation feature importance in a model-agnostic manner." |
flashlight | "The goal of this package is [to] shed light on black box machine learning models." |
forestmodel | "Produces forest plots using 'ggplot2' from models produced by functions such as stats::lm(), stats::glm() and survival::coxph()." |
fscaret | "Automated feature selection using variety of models provided by 'caret' package." |
gam | "Functions for fitting and working with generalized additive models, as described in chapter 7 of "Statistical Models in S" (Chambers and Hastie (eds), 1991), and "Generalized Additive Models" (Hastie and Tibshirani, 1990)." |
glm2 | "Fits generalized linear models using the same model specification as glm in the stats package, but with a modified default fitting method that provides greater stability for models that may fail to converge using glm." |
glmnet | "Extremely efficient procedures for fitting the entire lasso or elastic-net regularization path for linear regression, logistic and multinomial regression models, Poisson regression, Cox model, multiple-response Gaussian, and the grouped multinomial regression." |
H2O-3 Penalized Generalized Linear Models | "Fits a generalized linear model, specified by a response variable, a set of predictors, and a description of the error distribution." |
H2O-3 Monotonic GBM | "Builds gradient boosted classification trees and gradient boosted regression trees on a parsed data set." |
H2O-3 Sparse Principal Components (GLRM) | "Builds a generalized low rank decomposition of an H2O data frame." |
iBreakDown | "A model agnostic tool for explanation of predictions from black boxes ML models." |
ICEbox: Individual Conditional Expectation Plot Toolbox | "Implements Individual Conditional Expectation (ICE) plots, a tool for visualizing the model estimated by any supervised learning algorithm." |
iml | "An R package that interprets the behavior and explains predictions of machine learning models." |
ingredients | "A collection of tools for assessment of feature importance and feature effects." |
interpret: Fit Interpretable Machine Learning Models | "Package for training interpretable machine learning models." |
lightgbmExplainer | "An R package that makes LightGBM models fully interpretable." |
lime | "R port of the Python lime package." |
live | "Helps to understand key factors that drive the decision made by complicated predictive model (black box model)." |
mcr | "An R package for Model Reliance and Model Class Reliance." |
modelDown | "Website generator with HTML summaries for predictive models." |
modelOriented | GitHub repositories of Warsaw-based MI².AI. |
modelStudio | "Automates the explanatory analysis of machine learning predictive models." |
Monotonic XGBoost | Enforces consistent, directional relationships between features and predicted outcomes, enhancing model performance by aligning with prior data expectations. |
quantreg | "Estimation and inference methods for models for conditional quantile functions." |
rpart | "Recursive partitioning for classification, regression and survival trees." |
RuleFit | "Implements the learning method and interpretational tools described in Predictive Learning via Rule Ensembles." |
Scalable Bayesian Rule Lists (SBRL) | A more scalable implementation of Bayesian rule list from the Rudin group at Duke. |
shapFlex | Computes stochastic Shapley values for machine learning models to interpret them and evaluate fairness, including causal constraints in the feature space. |
shapleyR | "An R package that provides some functionality to use mlr tasks and models to generate shapley values." |
shapper | "Provides SHAP explanations of machine learning models." |
smbinning | "A set of functions to build a scoring model from beginning to end." |
vip | "An R package for constructing variable importance plots (VIPs)." |
xgboostExplainer | "An R package that makes xgboost models fully interpretable. |