The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning. Learn more →
Interpretable-ml-book Alternatives
Similar projects and alternatives to interpretable-ml-book
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
neural_regression_discontinuity
In this repository, I modify a quasi-experimental statistical procedure for time-series inference using convolutional long short-term memory networks.
-
random-forest-importances
Code to compute permutation and drop-column importances in Python scikit-learn models
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
interpretable-ml-book reviews and mentions
- A Guide to Making Black Box Models Interpretable
-
So much for AI
If you're a student, I'd recommend this book :https://christophm.github.io/interpretable-ml-book/
-
Best way to make a random forest more explainable (need to know which features are driving the prediction)
Pretty much everyone shows SHAP plots now. Definitely the way to go. Check out the Christoph Molnar book. https://christophm.github.io/interpretable-ml-book/
-
Is there another way to determine the effect of the features other than the inbuilt features importance and SHAP values? [Research] [Discussion]
Yes, there are many techniques beyond the two you listed. I suggest doing a survey of techniques (hint: explainable AI or XAI), starting with the following book: Interpretable Machine Learning.
-
Which industry/profession/tasks require an aggregate analysis of data representing different physical objects (And how would you call that?)
Ah, alright. It sounds like you're looking for interpretability so I'd suggest this amazing overview of it by Christoph Molnar. If you choose the right models, or the right way of interpreting those, it can help a ton in communicating not only your results, but also what you did to obtain them.
-
What skills do I need to really work on?
Not necessarily; decision trees, Naive Bayes, etc., are interpretable. I'd refer to Molnar--specifically his Interpretable Machine Learning text--if you are interested in that subject.
-
Random forest vs multiple regression to determine predictor importance.
Consulting something like Interpretable Machine Learning or the documentation of a package like the vip package would also be a really, really good place to start.
-
The Rashomon Effect Explained — Does Truth Actually Exist? [13.46]
Just read a book called Interpretable Machine Learning which focuses on analyzing ML models and determine which inputs has more impact in the result.
- Interpretable Machine Learning
-
Saw this in my Linkedin feed - what are your thoughts?
Calling it a book on SHAP undersells it. https://christophm.github.io/interpretable-ml-book/
-
A note from our sponsor - WorkOS
workos.com | 25 Apr 2024
Stats
christophM/interpretable-ml-book is an open source project licensed under GNU General Public License v3.0 or later which is an OSI approved license.
The primary programming language of interpretable-ml-book is Jupyter Notebook.
Sponsored