SOTAVerified

Interpretable Machine Learning

The goal of Interpretable Machine Learning is to allow oversight and understanding of machine-learned decisions. Much of the work in Interpretable Machine Learning has come in the form of devising methods to better explain the predictions of machine learning models.

Source: Assessing the Local Interpretability of Machine Learning Models

Papers

Showing 151175 of 537 papers

TitleStatusHype
Generalized Convergence Analysis of Tsetlin Machines: A Probabilistic Approach to Concept Learning0
Neural Stochastic Differential Equations for Robust and Explainable Analysis of Electromagnetic Unintended Radiated Emissions0
Fast Approximation of the Shapley Values Based on Order-of-Addition Experimental Designs0
Causal Entropy and Information Gain for Measuring Causal Control0
Operator-Based Detecting, Learning, and Stabilizing Unstable Periodic Orbits of Chaotic Attractors0
Measuring, Interpreting, and Improving Fairness of Algorithms using Causal Inference and Randomized Experiments0
Expanding Mars Climate Modeling: Interpretable Machine Learning for Modeling MSL Relative Humidity0
Development and validation of an interpretable machine learning-based calculator for predicting 5-year weight trajectories after bariatric surgery: a multinational retrospective cohort SOPHIA study0
Structural Node Embeddings with Homomorphism Counts0
Hyperspectral Blind Unmixing using a Double Deep Image PriorCode0
Improving Clinical Decision Support through Interpretable Machine Learning and Error Handling in Electronic Health Records0
An Interpretable Machine Learning Model with Deep Learning-based Imaging Biomarkers for Diagnosis of Alzheimer's Disease0
Interpretable Machine Learning for Discovery: Statistical Challenges \& Opportunities0
Is Grad-CAM Explainable in Medical Images?0
Interpreting and Correcting Medical Image Classification with PIP-NetCode1
Measuring Perceived Trust in XAI-Assisted Decision-Making by Eliciting a Mental Model0
Machine learning and Topological data analysis identify unique features of human papillae in 3D scans0
A Deep Dive into Perturbations as Evaluation Technique for Time Series XAICode0
Worth of knowledge in deep learningCode0
Decoding Urban-health Nexus: Interpretable Machine Learning Illuminates Cancer Prevalence based on Intertwined City Features0
Explainable Representation Learning of Small Quantum StatesCode0
Genomic Interpreter: A Hierarchical Genomic Deep Neural Network with 1D Shifted Window TransformerCode1
Explainable AI using expressive Boolean formulas0
Learning Transformer ProgramsCode1
Loss-Optimal Classification Trees: A Generalized Framework and the Logistic CaseCode0
Show:102550
← PrevPage 7 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Q-SENNTop 1 Accuracy85.9Unverified
2SLDD-ModelTop 1 Accuracy85.7Unverified