There are 57 repositories under interpretability topic.
A curated list of awesome open source libraries to deploy, monitor, version and scale your machine learning
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
Fit interpretable models. Explain blackbox machine learning.
Model interpretability and understanding for PyTorch
A collection of infrastructure and tools for research in neural network interpretability.
A curated list of awesome responsible machine learning resources.
StellarGraph - Machine Learning on Graphs
🔅 Shapash: User-friendly Explainability and Interpretability to Develop Reliable and Transparent Machine Learning Models
Class activation maps for your PyTorch models (CAM, Grad-CAM, Grad-CAM++, Smooth Grad-CAM++, Score-CAM, SS-CAM, IS-CAM, XGrad-CAM, Layer-CAM)
FedML - The Research and Production Integrated Federated Learning Library: https://fedml.ai
A JAX research toolkit for building, editing, and visualizing neural networks.
Responsible AI Toolbox is a suite of tools providing model and data exploration and assessment user interfaces and libraries that enable a better understanding of AI systems. These interfaces and libraries empower developers and stakeholders of AI systems to develop and monitor AI more responsibly, and take better data-driven actions.
A collection of research materials on explainable AI/ML
Stanford NLP Python library for Representation Finetuning (ReFT)
Interpretable ML package 🔍 for concise, transparent, and accurate predictive modeling (sklearn-compatible).
moDel Agnostic Language for Exploration and eXplanation (JMLR 2018; JMLR 2021)
Model explainability that works seamlessly with 🤗 transformers. Explain your transformers model in just 2 lines of code.
Interpretability Methods for tf.keras models with Tensorflow 2.x
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Public facing deeplift repo
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Interesting resources related to XAI (Explainable Artificial Intelligence)
Stanford NLP Python library for understanding and improving PyTorch models via interventions
深度学习近年来关于神经网络模型解释性的相关高引用/顶会论文(附带代码)
Visualization toolkit for neural networks in PyTorch! Demo -->
The nnsight package enables interpreting and manipulating the internals of deep learned models.
A collection of state-of-the-art algorithms for the training, serving and interpretation of Decision Forest models in Keras.
Examples of techniques for training interpretable ML models, explaining ML models, and debugging ML models for accuracy, discrimination, and security.