SOTAVerified

Audio captioning

Audio Captioning is the task of describing audio using text. The general approach is to use an audio encoder to encode the audio (example: PANN, CAV-MAE), and to use a decoder (example: transformer) to generate the text. To judge the quality of audio captions, though machine translation metrics (BLEU, METEOR, ROUGE) and image captioning metrics (SPICE, CIDER) are used, they are not very well-suited. Attempts have been made to use pretrained language model based metrics such as Sentence-BERT.

Papers

Showing 5175 of 119 papers

TitleStatusHype
AC/DC: LLM-based Audio Comprehension via Dialogue Continuation0
Auto-ACD: A Large-scale Dataset for Audio-Language Representation Learning0
An Attempt towards Interpretable Audio-Visual Video Captioning0
An investigation on selecting audio pre-trained models for audio captioning0
A Transformer-based Audio Captioning Model with Keyword Estimation0
AudioCaps: Generating Captions for Audios in The Wild0
Audio Captioning using Gated Recurrent Units0
Audio Captioning using Pre-Trained Large-Scale Language Model Guided by Audio-based Similar Caption Retrieval0
Enhancing Retrieval-Augmented Audio Captioning with Generation-Assisted Multimodal Querying and Progressive Learning0
Audio Captioning with Composition of Acoustic and Semantic Information0
Audio-CoT: Exploring Chain-of-Thought Reasoning in Large Audio Language Model0
Audio Dialogues: Dialogues dataset for audio and music understanding0
Audio Difference Learning for Audio Captioning0
Audio Flamingo 2: An Audio-Language Model with Long-Audio Understanding and Expert Reasoning Abilities0
Automated Audio Captioning: An Overview of Recent Progress and New Challenges0
Automated Audio Captioning using Transfer Learning and Reconstruction Latent Space Similarity Regularization0
Automated Audio Captioning via Fusion of Low- and High- Dimensional Features0
Automated Audio Captioning with Epochal Difficult Captions for Curriculum Learning0
Automated Audio Captioning with Recurrent Neural Networks0
Automatic Audio Captioning using Attention weighted Event based Embeddings0
CLAP-ART: Automated Audio Captioning with Semantic-rich Audio Representation Tokenizer0
Classifier-Guided Captioning Across Modalities0
CosyAudio: Improving Audio Generation with Confidence Scores and Synthetic Captions0
Diverse Audio Captioning via Adversarial Training0
Diversity and bias in audio captioning datasets0
Show:102550
← PrevPage 3 of 5Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1VASTCIDEr0.78Unverified
2VALORCIDEr0.74Unverified
3MQ-CapSPIDEr0.52Unverified
4SLAM-AACSPIDEr0.52Unverified
5LAVCapSPIDEr0.52Unverified
6EnCLAP++-largeSPIDEr0.51Unverified
7AutoCapSPIDEr0.51Unverified
8LOAESPIDEr0.51Unverified
9EnCLAP++-baseSPIDEr0.5Unverified
10EnCLAP-largeSPIDEr0.5Unverified
#ModelMetricClaimedVerifiedStatus
1VASTCIDEr0.52Unverified
2VALORCIDEr0.42Unverified
3SLAM-AACSPIDEr0.33Unverified
4LOAESPIDEr0.33Unverified
5MQ-CapSPIDEr0.32Unverified
6EnsembleSPIDEr0.32Unverified
7Audio Flamingo (Pengi trainset)SPIDEr0.31Unverified
8Ensemble-RLSPIDEr0.3Unverified
9Qwen-AudioSPIDEr0.29Unverified
10EnsembleSPIDEr0.21Unverified