SOTAVerified

Interpretable Machine Learning

The goal of Interpretable Machine Learning is to allow oversight and understanding of machine-learned decisions. Much of the work in Interpretable Machine Learning has come in the form of devising methods to better explain the predictions of machine learning models.

Source: Assessing the Local Interpretability of Machine Learning Models

Papers

Showing 376400 of 537 papers

TitleStatusHype
Neuro-symbolic Models for Interpretable Time Series Classification using Temporal Logic Description0
The Pros and Cons of Using Machine Learning and Interpretable Machine Learning Methods in psychiatry detection applications, specifically depression disorder: A Brief Review0
Nothing Else Matters: Model-Agnostic Explanations By Identifying Prediction Invariance0
Novel Topological Shapes of Model Interpretability0
An Interpretable Machine Learning Framework to Understand Bikeshare Demand before and during the COVID-19 Pandemic in New York City0
Using Explainable Boosting Machine to Compare Idiographic and Nomothetic Approaches for Ecological Momentary Assessment Data0
"Oh LLM, I'm Asking Thee, Please Give Me a Decision Tree": Zero-Shot Decision Tree Induction and Embedding with Large Language Models0
A hybrid machine learning framework for analyzing human decision making through learning preferences0
One Explanation Does Not Fit All: The Promise of Interactive Explanations for Machine Learning Transparency0
On Explaining Decision Trees0
On Interpretability and Similarity in Concept-Based Machine Learning0
Using Interpretable Machine Learning to Predict Maternal and Fetal Outcomes0
Online Product Feature Recommendations with Interpretable Machine Learning0
On quantitative aspects of model interpretability0
On the definition and importance of interpretability in scientific machine learning0
Look Who's Talking: Interpretable Machine Learning for Assessing Italian SMEs Credit Default0
On the Safety of Interpretable Machine Learning: A Maximum Deviation Approach0
On the Shape of Brainscores for Large Language Models (LLMs)0
On the Use of Interpretable Machine Learning for the Management of Data Quality0
Using Model-Based Trees with Boosting to Fit Low-Order Functional ANOVA Models0
Open Issues in Combating Fake News: Interpretability as an Opportunity0
Operator-Based Detecting, Learning, and Stabilizing Unstable Periodic Orbits of Chaotic Attractors0
OPTDTALS: Approximate Logic Synthesis via Optimal Decision Trees Approach0
An Interpretable Machine Learning Approach to Understanding the Relationships between Solar Flares and Source Active Regions0
Topological data analysis of zebrafish patterns0
Show:102550
← PrevPage 16 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Q-SENNTop 1 Accuracy85.9Unverified
2SLDD-ModelTop 1 Accuracy85.7Unverified