XMLX GitHub configuration
-
Updated
Jul 26, 2024
XMLX GitHub configuration
XMLX GitHub configuration
Default Risk Prediction from bank dataset with Interpretable Machine Learning
Demonstration of InterpretME, an interpretable machine learning pipeline
This project contains the data, code and results used in the paper title "On the relationship of novelty and value in digitalization patents: A machine learning approach".
INVASE: Instance-wise Variable Selection . For more details, read the paper "INVASE: Instance-wise Variable Selection using Neural Networks," International Conference on Learning Representations (ICLR), 2019.
Overview of machine learning interpretation techniques and their implementations
Rule Extraction from Bayesian Networks
TeleGam: Combining Visualization and Verbalization for Interpretable Machine Learning
Paper for 2018 Joint Statistical Meetings: https://ww2.amstat.org/meetings/jsm/2018/onlineprogram/AbstractDetails.cfm?abstractid=329539
Article for Special Edition of Information: Machine Learning with Python
Techniques & resources for training interpretable ML models, explaining ML models, and debugging ML models.
The code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Slides, videos and other potentially useful artifacts from various presentations on responsible machine learning.
An interpretable machine learning pipeline over knowledge graphs
Sample use case for Xavier AI in Healthcare conference: https://www.xavierhealth.org/ai-summit-day2/
Predicting the Likelihood to Purchase a Financial Product Following a Direct Marketing Campaign
Explainable AI framework for data scientists. Explain & debug any blackbox machine learning model with a single line of code. We are looking for co-authors to take this project forward. Reach out @ [email protected]
H2O.ai Machine Learning Interpretability Resources
Examples of techniques for training interpretable ML models, explaining ML models, and debugging ML models for accuracy, discrimination, and security.
Add a description, image, and links to the machine-learning-interpretability topic page so that developers can more easily learn about it.
To associate your repository with the machine-learning-interpretability topic, visit your repo's landing page and select "manage topics."