A curated, but probably biased and incomplete, list of awesome machine learning interpretability resources.
If you want to contribute to this list (and please do!) read over the contribution guidelines, send a pull request, or contact me @hedema.
An incomplete, imperfect blueprint for a more human-centered, lower-risk machine learning. The resources in this repository can be used to do many of these things today. The resources in this repository should not be considered legal compliance advice.
Image credit: H2O.ai Machine Learning Interpretability team, https://github.com/h2oai/mli-resources.
- Comprehensive Software Examples and Tutorials
- Explainability- or Fairness-Enhancing Software Packages
- Free Books
- Other Interpretability and Fairness Resources and Lists
- Review and General Papers
- Teaching Resources
- Interpretable ("Whitebox") or Fair Modeling Packages
- Getting a Window into your Black Box Model
- IML
- Interpretable Machine Learning with Python
- Interpreting Machine Learning Models with the iml Package
- Interpretable Machine Learning using Counterfactuals
- Machine Learning Explainability by Kaggle Learn
- Model Interpretability with DALEX
- Model Interpretation series by Dipanjan (DJ) Sarkar:
- Partial Dependence Plots in R
- Saliency Maps for Deep Learning
- Visualizing ML Models with LIME
- What does a CNN see?
- Visualizing and debugging deep convolutional networks
- acd
- aequitas
- AI Fairness 360
- AI Explainability 360
- algofairness
- Alibi
- anchor
- BlackBoxAuditing
- casme
- ContrastiveExplanation (Foil Trees)
- DALEXtra
- DeepExplain
- deeplift
- deepvis
- eli5
- fairml
- fairness
- foolbox
- Grad-CAM
- iNNvestigate neural nets
- Integrated-Gradients
- interpret_with_rules
- Keras-vis
- keract
- lofo-importance
- L2X
- lime
- lrp_toolbox
- microsoft/interpret
- MLextend
- PDPbox
- pyBreakDown
- PyCEbox
- rationale
- robustness
- RISE
- shap
- Skater
- tensorfow/cleverhans
- tensorflow/lucid
- tensorflow/model-analysis
- tensorflow/privacy
- tensorfuzz
- TensorWatch
- tf-explain
- Themis
- themis-ml
- treeinterpreter
- woe
- xai
- yellowbrick
- ALEPlot
- breakDown
- DrWhyAI
- DALEX
- DALEXtra
- EloML
- ExplainPrediction
- featureImportance
- forestmodel
- fscaret
- ICEbox
- iml
- lightgbmExplainer
- lime
- live
- mcr
- modelDown
- modelOriented
- modelStudio
- pdp
- shapleyR
- shapper
- smbinning
- vip
- xgboostExplainer
- An Introduction to Machine Learning Interpretability
- Fairness and Machine Learning
- Interpretable Machine Learning
- 8 Principles of Responsible ML
- ACM FAT* 2019 Youtube Playlist
- AI Ethics Guidelines Global Inventory
- AllenNLP Interpret: A Framework for Explaining Predictions of NLP Models
- Awesome interpretable machine learning ;)
- Awesome machine learning operations
- algoaware
- Beyond Explainability: A Practical Guide to Managing Risk in Machine Learning Models
- criticalML
- Debugging Machine Learning Models (ICLR workshop proceedings)
- Deep Insights into Explainability and Interpretability of Machine Learning Algorithms and Applications to Risk Management
- Distill
- Fairness, Accountability, and Transparency in Machine Learning (FAT/ML) Scholarship
- Machine Learning Ethics References
- Machine Learning Interpretability Resources
- MIT AI Ethics Reading Group
- private-ai-resources
- Warning Signs: The Future of Privacy and Security in an Age of Machine Learning
- XAI Resources
- You Created A Machine Learning Application Now Make Sure It's Secure
- A Comparative Study of Fairness-Enhancing Interventions in Machine Learning
- A Survey Of Methods For Explaining Black Box Models
- A Marauder’s Map of Security and Privacy in Machine Learning
- Challenges for Transparency
- Explaining Explanations: An Overview of Interpretability of Machine Learning
- Explanation in Human-AI Systems: A Literature Meta-Review, Synopsis of Key Ideas and Publications, and Bibliography for Explainable AI
- Interpretable Machine Learning: Definitions, Methods, and Applications
- Machine Learning Explainability in Finance
- On the Art and Science of Machine Learning Explanations
- On the Responsibility of Technologists: A Prologue and Primer
- Please Stop Explaining Black Box Models for High-Stakes Decisions
- The Mythos of Model Interpretability
- Towards A Rigorous Science of Interpretable Machine Learning
- The Security of Machine Learning
- Techniques for Interpretable Machine Learning
- Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda
- An Introduction to Data Ethics
- Fairness in Machine Learning
- Human-Center Machine Learning
- Practical Model Interpretability
- Trustworthy Deep Learning
- Bayesian Case Model
- Bayesian Ors-Of-Ands
- Bayesian Rule List (BRL)
- Explainable Boosting Machine (EBM)
- fair-classification
- Falling Rule List (FRL)
- H2O-3
- Monotonic XGBoost
- pyGAM
- Risk-SLIM
- Scikit-learn
- sklearn-expertsys
- skope-rules
- Super-sparse Linear Integer models (SLIMs)
- tensorflow/lattice
- EconML: A Python Package for ML-Based Heterogeneous Treatment Effects Estimation