SOTAVerified

Interpretable Machine Learning

The goal of Interpretable Machine Learning is to allow oversight and understanding of machine-learned decisions. Much of the work in Interpretable Machine Learning has come in the form of devising methods to better explain the predictions of machine learning models.

Source: Assessing the Local Interpretability of Machine Learning Models

Papers

Showing 326350 of 537 papers

TitleStatusHype
Taming Waves: A Physically-Interpretable Machine Learning Framework for Realizable Control of Wave Dynamics0
Longitudinal Distance: Towards Accountable Instance Attribution0
Techniques for Interpretable Machine Learning0
Tell Me Why: Using Question Answering as Distant Supervision for Answer Justification0
Machine learning and Topological data analysis identify unique features of human papillae in 3D scans0
Machine Learning-Based Prediction of Mortality in Geriatric Traumatic Brain Injury Patients0
Machine Learning for Economic Forecasting: An Application to China's GDP Growth0
MAIR: Framework for mining relationships between research articles, strategies, and regulations in the field of explainable artificial intelligence0
Additive Higher-Order Factorization Machines0
A review of possible effects of cognitive biases on the interpretation of rule-based machine learning models0
Tensor Polynomial Additive Model0
The Contextual Lasso: Sparse Linear Models via Deep Neural Networks0
MAntRA: A framework for model agnostic reliability analysis0
The Doctor Just Won't Accept That!0
The explanation dialogues: an expert focus study to understand requirements towards explanations within the GDPR0
Using Interpretable Machine Learning to Massively Increase the Number of Antibody-Virus Interactions Across Studies0
Mathematics of statistical sequential decision-making: concentration, risk-awareness and modelling in stochastic bandits, with applications to bariatric surgery0
MCA-based Rule Mining Enables Interpretable Inference in Clinical Psychiatry0
MCCE: Missingness-aware Causal Concept Explainer0
Meaningful Models: Utilizing Conceptual Structure to Improve Machine Learning Interpretability0
Measuring, Interpreting, and Improving Fairness of Algorithms using Causal Inference and Randomized Experiments0
Measuring Perceived Trust in XAI-Assisted Decision-Making by Eliciting a Mental Model0
The Most Important Features in Generalized Additive Models Might Be Groups of Features0
Mining Meta-indicators of University Ranking: A Machine Learning Approach Based on SHAP0
Missing Values and Imputation in Healthcare Data: Can Interpretable Machine Learning Help?0
Show:102550
← PrevPage 14 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Q-SENNTop 1 Accuracy85.9Unverified
2SLDD-ModelTop 1 Accuracy85.7Unverified