SOTAVerified

Multimodal Deep Learning

Multimodal deep learning is a type of deep learning that combines information from multiple modalities, such as text, image, audio, and video, to make more accurate and comprehensive predictions. It involves training deep neural networks on data that includes multiple types of information and using the network to make predictions based on this combined data.

One of the key challenges in multimodal deep learning is how to effectively combine information from multiple modalities. This can be done using a variety of techniques, such as fusing the features extracted from each modality, or using attention mechanisms to weight the contribution of each modality based on its importance for the task at hand.

Multimodal deep learning has many applications, including image captioning, speech recognition, natural language processing, and autonomous vehicles. By combining information from multiple modalities, multimodal deep learning can improve the accuracy and robustness of models, enabling them to perform better in real-world scenarios where multiple types of information are present.

Papers

Showing 101125 of 213 papers

TitleStatusHype
P-Transformer: A Prompt-based Multimodal Transformer Architecture For Medical Tabular Data0
A Comprehensive and Versatile Multimodal Deep Learning Approach for Predicting Diverse Properties of Advanced Materials0
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init AttentionCode5
Multimodal Deep Learning to Differentiate Tumor Recurrence from Treatment Effect in Human Glioblastoma0
Show me your NFT and I tell you how it will perform: Multimodal representation learning for NFT selling price prediction0
Zorro: the masked multimodal transformerCode0
Multimodal Deep LearningCode1
Learning Semantic Relationship Among Instances for Image-Text MatchingCode1
Learning Multimodal Data Augmentation in Feature SpaceCode1
Common Practices and Taxonomy in Deep Multi-view Fusion for Remote Sensing ApplicationsCode1
Medical Diagnosis with Large Scale Multimodal Transformers: Leveraging Diverse Data for More Accurate DiagnosisCode1
A survey on knowledge-enhanced multimodal learning0
aiMotive Dataset: A Multimodal Dataset for Robust Autonomous Driving with Long-Range PerceptionCode1
Language-Assisted Deep Learning for Autistic Behaviors Recognition0
MultiCrossViT: Multimodal Vision Transformer for Schizophrenia Prediction using Structural MRI and Functional Network Connectivity Data0
Bayesian Prompt Learning for Image-Language Model GeneralizationCode1
LASP: Text-to-Text Optimization for Language-Aware Soft Prompting of Vision & Language ModelsCode1
Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question AnsweringCode2
LAVIS: A Library for Language-Vision Intelligence0
Identification of Cognitive Workload during Surgical Tasks with Multimodal Deep Learning0
TMSS: An End-to-End Transformer-based Multimodal Network for Segmentation and Survival PredictionCode1
Multi-Modal Experience Inspired AI CreationCode1
R2D2 at SemEval-2022 Task 5: Attention is only as good as its Values! A multimodal system for identifying misogynist memes0
Multimodal Attention-based Deep Learning for Alzheimer's Disease DiagnosisCode1
Vision-Aided Frame-Capture-Based CSI Recomposition for WiFi Sensing: A Multimodal Approach0
Show:102550
← PrevPage 5 of 9Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Two Branch Network (Text - Bert + Image - Nts-Net)Accuracy96.81Unverified