SOTAVerified

VADMamba: Exploring State Space Models for Fast Video Anomaly Detection

2025-03-27Code Available1· sign in to hype

Jiahao Lyu, Minghua Zhao, Jing Hu, Xuewen Huang, Yifei Chen, Shuangli Du

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Video anomaly detection (VAD) methods are mostly CNN-based or Transformer-based, achieving impressive results, but the focus on detection accuracy often comes at the expense of inference speed. The emergence of state space models in computer vision, exemplified by the Mamba model, demonstrates improved computational efficiency through selective scans and showcases the great potential for long-range modeling. Our study pioneers the application of Mamba to VAD, dubbed VADMamba, which is based on multi-task learning for frame prediction and optical flow reconstruction. Specifically, we propose the VQ-Mamba Unet (VQ-MaU) framework, which incorporates a Vector Quantization (VQ) layer and Mamba-based Non-negative Visual State Space (NVSS) block. Furthermore, two individual VQ-MaU networks separately predict frames and reconstruct corresponding optical flows, further boosting accuracy through a clip-level fusion evaluation strategy. Experimental results validate the efficacy of the proposed VADMamba across three benchmark datasets, demonstrating superior performance in inference speed compared to previous work. Code is available at https://github.com/jLooo/VADMamba.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CUHK AvenueVADMambaAUC91.5Unverified
ShanghaiTech CampusVADMambaAUC77Unverified
UCSD Ped2VADMambaAUC98.5Unverified

Reproductions