SOTAVerified

Interpretable Machine Learning

The goal of Interpretable Machine Learning is to allow oversight and understanding of machine-learned decisions. Much of the work in Interpretable Machine Learning has come in the form of devising methods to better explain the predictions of machine learning models.

Source: Assessing the Local Interpretability of Machine Learning Models

Papers

Showing 201210 of 537 papers

TitleStatusHype
Explanations for Automatic Speech Recognition0
Structural Neural Additive Models: Enhanced Interpretable Machine LearningCode1
Applying BERT and ChatGPT for Sentiment Analysis of Lyme Disease in Scientific Literature0
The Contextual Lasso: Sparse Linear Models via Deep Neural Networks0
Data-driven Approach for Static Hedging of Exchange Traded Options0
SkinCon: A skin disease dataset densely annotated by domain experts for fine-grained model debugging and analysis0
Interpretable Classification of Early Stage Parkinson's Disease from EEG0
Explainable, Interpretable & Trustworthy AI for Intelligent Digital Twin: Case Study on Remaining Useful Life0
Learning Support and Trivial Prototypes for Interpretable Image ClassificationCode1
An interpretable machine learning system for colorectal cancer diagnosis from pathology slides0
Show:102550
← PrevPage 21 of 54Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Q-SENNTop 1 Accuracy85.9Unverified
2SLDD-ModelTop 1 Accuracy85.7Unverified