Audio Classification
Audio Classification is a machine learning task that involves identifying and tagging audio signals into different classes or categories. The goal of audio classification is to enable machines to automatically recognize and distinguish between different types of audio, such as music, speech, and environmental sounds.
Papers
Showing 1–10 of 361 papers
All datasetsAudioSetESC-50ICBHI Respiratory Sound DatabaseVGGSoundSHDFSD50KBalanced Audio SetSpeech CommandsSSCBirdCLEF 2021DCASEEPIC-KITCHENS-100
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | Event-SSM | Percentage correct | 95.9 | — | Unverified |
| 2 | SNN with Dilated Convolution with Learnable Spacings | Percentage correct | 95.1 | — | Unverified |
| 3 | SNN featuring learnable axonal delays with adaptively delay caps | Percentage correct | 92.45 | — | Unverified |
| 4 | CNN | Percentage correct | 92.4 | — | Unverified |
| 5 | SNN with spatio-temporal filters and attention | Percentage correct | 92.4 | — | Unverified |
| 6 | SNN with temporal-wise attention | Percentage correct | 91.1 | — | Unverified |
| 7 | SNN | Percentage correct | 87 | — | Unverified |
| 8 | Recurrent convolutional SNN | Percentage correct | 83.5 | — | Unverified |
| 9 | Recurrent SNN | Percentage correct | 83.2 | — | Unverified |
| 10 | Sparse Spiking Gradient Descent | Percentage correct | 77.5 | — | Unverified |