SOTAVerified

Concept Alignment

Concept Alignment aims to align the learned representations or concepts within a model with the intended or target concepts. It involves adjusting the model's parameters or training process to ensure that the learned concepts accurately reflect the underlying patterns in the data.

Papers

Showing 3136 of 36 papers

TitleStatusHype
Interpretable Concept-based Deep Learning Framework for Multimodal Human Behavior Modeling0
Language-based Action Concept Spaces Improve Video Self-Supervised Learning0
Natural Language Detectors Emerge in Individual Neurons0
NEUCORE: Neural Concept Reasoning for Composed Image Retrieval0
Replace in Translation: Boost Concept Alignment in Counterfactual Text-to-Image0
SNIFFER: Multimodal Large Language Model for Explainable Out-of-Context Misinformation Detection0
Show:102550
← PrevPage 4 of 4Next →

No leaderboard results yet.