SOTAVerified

Scene Understanding

Scene understanding involves interpreting the visual information of a scene, including objects, their spatial relationships, and the overall layout. It goes beyond simple object recognition by considering the context and how objects relate to each other and the environment.

Papers

Showing 101125 of 1723 papers

TitleStatusHype
CLIP goes 3D: Leveraging Prompt Tuning for Language Grounded 3D RecognitionCode2
GALIP: Generative Adversarial CLIPs for Text-to-Image SynthesisCode2
Diffusion-based Generation, Optimization, and Planning in 3D ScenesCode2
Panoptic Lifting for 3D Scene Understanding with Neural FieldsCode2
PLA: Language-Driven Open-Vocabulary 3D Scene UnderstandingCode2
OpenScene: 3D Scene Understanding with Open VocabulariesCode2
Safety-Enhanced Autonomous Driving Using Interpretable Sensor Fusion TransformerCode2
Panoptic Scene Graph GenerationCode2
BinsFormer: Revisiting Adaptive Bins for Monocular Depth EstimationCode2
InvPT: Inverted Pyramid Multi-task Transformer for Dense Scene UnderstandingCode2
CMX: Cross-Modal Fusion for RGB-X Semantic Segmentation with TransformersCode2
GroupViT: Semantic Segmentation Emerges from Text SupervisionCode2
HAKE: A Knowledge Engine Foundation for Human Activity UnderstandingCode2
Panoptic nuScenes: A Large-Scale Benchmark for LiDAR Panoptic Segmentation and TrackingCode2
Hypersim: A Photorealistic Synthetic Dataset for Holistic Indoor Scene UnderstandingCode2
Multi-Task Learning as Multi-Objective OptimizationCode2
Learning to Tune Like an Expert: Interpretable and Scene-Aware Navigation via MLLM Reasoning and CVAE-Based AdaptationCode1
SurgTPGS: Semantic 3D Surgical Scene Understanding with Text Promptable Gaussian SplattingCode1
ReME: A Data-Centric Framework for Training-Free Open-Vocabulary SegmentationCode1
DIP: Unsupervised Dense In-Context Post-training of Visual RepresentationsCode1
STSBench: A Spatio-temporal Scenario Benchmark for Multi-modal Large Language Models in Autonomous DrivingCode1
OWMM-Agent: Open World Mobile Manipulation With Multi-modal Agentic Data SynthesisCode1
PhysGaia: A Physics-Aware Dataset of Multi-Body Interactions for Dynamic Novel View SynthesisCode1
CoNav: Collaborative Cross-Modal Reasoning for Embodied NavigationCode1
StoryReasoning Dataset: Using Chain-of-Thought for Scene Understanding and Grounded Story GenerationCode1
Show:102550
← PrevPage 5 of 69Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ACRV BaselineOMQ0.44Unverified
2Team VGAI (TCS Research)OMQ0.37Unverified
3Demo_semantic_SLAMOMQ0.11Unverified
#ModelMetricClaimedVerifiedStatus
1CPN(ResNet-101)Mean IoU46.3Unverified
#ModelMetricClaimedVerifiedStatus
1ACRV BaselineOMQ0.35Unverified