SOTAVerified

Activity Recognition

Human Activity Recognition is the problem of identifying events performed by humans given a video input. It is formulated as a binary (or multiclass) classification problem of outputting activity class labels. Activity Recognition is an important problem with many societal applications including smart surveillance, video search/retrieval, intelligent robots, and other monitoring systems.

Source: Learning Latent Sub-events in Activity Videos Using Temporal Attention Filters

Papers

Showing 51100 of 1322 papers

TitleStatusHype
3D Human Shape and Pose from a Single Low-Resolution Image with Self-Supervised LearningCode1
Online Semi-Supervised Learning of Composite Event Rules by Combining Structure and Mass-Based Predicate SimilarityCode1
Optimization-Free Test-Time Adaptation for Cross-Person Activity RecognitionCode1
OSL𝛼: Online Structure Learning Using Background Knowledge AxiomatizationCode1
Deep Learning for Time Series Classification and Extrinsic Regression: A Current SurveyCode1
Quantized Distillation: Optimizing Driver Activity Recognition Models for Resource-Constrained EnvironmentsCode1
Real-world Anomaly Detection in Surveillance VideosCode1
RelCon: Relative Contrastive Learning for a Motion Foundation Model for Wearable DataCode1
SelfHAR: Improving Human Activity Recognition through Self-training with Unlabeled DataCode1
ESPRESSO: Entropy and ShaPe awaRe timE-Series SegmentatiOn for processing heterogeneous sensor dataCode1
Self-Supervised PPG Representation Learning Shows High Inter-Subject VariabilityCode1
Self-supervised transfer learning of physiological representations from free-living wearable dataCode1
Comparing Self-Supervised Learning Techniques for Wearable Human Activity RecognitionCode1
Challenges in Multi-centric Generalization: Phase and Step Recognition in Roux-en-Y Gastric Bypass SurgeryCode1
CALDA: Improving Multi-Source Time Series Domain Adaptation with Contrastive Adversarial LearningCode1
CholecTriplet2021: A benchmark challenge for surgical action triplet recognitionCode1
COMPOSER: Compositional Reasoning of Group Activity in Videos with Keypoint-Only ModalityCode1
Autoregressive Adaptive Hypergraph Transformer for Skeleton-based Activity RecognitionCode1
BASAR:Black-box Attack on Skeletal Action RecognitionCode1
COMODO: Cross-Modal Video-to-IMU Distillation for Efficient Egocentric Human Activity RecognitionCode1
Convolutional Tensor-Train LSTM for Spatio-temporal LearningCode1
CubeLearn: End-to-end Learning for Human Motion Recognition from Raw mmWave Radar SignalsCode1
Attention-Based Deep Learning Framework for Human Activity Recognition with User AdaptationCode1
A Review of Deep Learning Methods for Photoplethysmography DataCode1
DeepSense: A Unified Deep Learning Framework for Time-Series Mobile Sensing Data ProcessingCode1
Deep Unsupervised Domain Adaptation for Time Series Classification: a BenchmarkCode1
Action-slot: Visual Action-centric Representations for Multi-label Atomic Activity Recognition in Traffic ScenesCode1
Efficient Two-Stream Network for Violence Detection Using Separable Convolutional LSTMCode1
Exploring Few-Shot Adaptation for Activity Recognition on Diverse DomainsCode1
A Federated Learning Aggregation Algorithm for Pervasive Computing: Evaluation and ComparisonCode1
Fine-Grained Egocentric Hand-Object Segmentation: Dataset, Model, and ApplicationsCode1
Generating Virtual On-body Accelerometer Data from Virtual Textual Descriptions for Human Activity RecognitionCode1
Audio-Adaptive Activity Recognition Across Video DomainsCode1
Bridge-Prompt: Towards Ordinal Action Understanding in Instructional VideosCode1
HHAR-net: Hierarchical Human Activity Recognition using Neural NetworksCode1
Hierarchical Self Attention Based Autoencoder for Open-Set Human Activity RecognitionCode1
Human Activity Segmentation Challenge @ ECML/PKDD’23Code1
Human skeletons and change detection for efficient violence detection in surveillance videosCode1
IMU2CLIP: Multimodal Contrastive Learning for IMU Motion Sensors from Egocentric Videos and TextCode1
IMUGPT 2.0: Language-Based Cross Modality Transfer for Sensor-Based Human Activity RecognitionCode1
Knowledge Mining with Scene Text for Fine-Grained RecognitionCode1
LaMPP: Language Models as Probabilistic Priors for Perception and ActionCode1
Learning Generalizable Physiological Representations from Large-scale Wearable DataCode1
Learning Group Activities from Skeletons without Individual Action LabelsCode1
Contrastive Learning with Cross-Modal Knowledge Mining for Multimodal Human Activity RecognitionCode1
MOMA-LRG: Language-Refined Graphs for Multi-Object Multi-Actor Activity ParsingCode1
Multimodal Transformer for Nursing Activity RecognitionCode1
Multi-stage Learning for Radar Pulse Activity SegmentationCode1
Exploring Contrastive Learning in Human Activity Recognition for HealthcareCode1
Mobile Sensor Data AnonymizationCode1
Show:102550
← PrevPage 2 of 27Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Structured Keypoint PoolingAccuracy93.4Unverified
2Semi-Supervised Hard Attention (SSHA); pretrained on Deepmind Kinetics datasetAccuracy90.4Unverified
3Human Skeletons + Change DetectionAccuracy90.25Unverified
4Separable Convolutional LSTMAccuracy89.75Unverified
5SPIL ConvolutionAccuracy89.3Unverified
6Flow Gated NetworkAccuracy87.25Unverified
#ModelMetricClaimedVerifiedStatus
1FocusCLIPTop-3 Accuracy (%)10.47Unverified
2CLIPTop-3 Accuracy (%)6.49Unverified
#ModelMetricClaimedVerifiedStatus
1Boutaleb et al.1:1 Accuracy97.91Unverified
#ModelMetricClaimedVerifiedStatus
1all-landmark-modelActivity Recognition0.76Unverified