SOTAVerified

Language Modeling

Papers

Showing 9511000 of 14182 papers

TitleStatusHype
GOFA: A Generative One-For-All Model for Joint Graph Language ModelingCode2
Contrastive Decoding: Open-ended Text Generation as OptimizationCode2
Tell Me More! Towards Implicit User Intention Understanding of Language Model Driven AgentsCode2
Teola: Towards End-to-End Optimization of LLM-based ApplicationsCode2
Contrastive Search Is What You Need For Neural Text GenerationCode2
GoLLIE: Annotation Guidelines improve Zero-Shot Information-ExtractionCode2
CogView2: Faster and Better Text-to-Image Generation via Hierarchical TransformersCode2
TextHawk2: A Large Vision-Language Model Excels in Bilingual OCR and Grounding with 16x Fewer TokensCode2
GPT-Driver: Learning to Drive with GPTCode2
Continuous Diffusion Model for Language ModelingCode2
DiffArtist: Towards Structure and Appearance Controllable Image StylizationCode2
GLUS: Global-Local Reasoning Unified into A Single Large Language Model for Video SegmentationCode2
GIT: A Generative Image-to-text Transformer for Vision and LanguageCode2
The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single TransformerCode2
Enhancing Diagnostic Accuracy in Rare and Common Fundus Diseases with a Knowledge-Rich Vision-Language ModelCode2
Contextual Semantic Embeddings for Ontology Subsumption PredictionCode2
GMAI-VL & GMAI-VL-5.5M: A Large Vision-Language Model and A Comprehensive Multimodal Dataset Towards General Medical AICode2
GPT or BERT: why not both?Code2
How to Index Item IDs for Recommendation Foundation ModelsCode2
GenSim: A General Social Simulation Platform with Large Language Model based AgentsCode2
TIPO: Text to Image with Text Presampling for Prompt OptimizationCode2
Explore the Limits of Omni-modal Pretraining at ScaleCode2
Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNetCode2
MLLM-Tool: A Multimodal Large Language Model For Tool Agent LearningCode2
GeoChat: Grounded Large Vision-Language Model for Remote SensingCode2
Towards Interpreting Visual Information Processing in Vision-Language ModelsCode2
Generative Region-Language Pretraining for Open-Ended Object DetectionCode2
A Generalist AgentCode2
Generative Pretrained Structured Transformers: Unsupervised Syntactic Language Models at ScaleCode2
GeoGround: A Unified Large Vision-Language Model for Remote Sensing Visual GroundingCode2
Generating Benchmarks for Factuality Evaluation of Language ModelsCode2
TrafficVLM: A Controllable Visual Language Model for Traffic Video CaptioningCode2
Generate rather than Retrieve: Large Language Models are Strong Context GeneratorsCode2
Generative Modeling for Mathematical DiscoveryCode2
Generalized Interpolating Discrete DiffusionCode2
Generalized Few-shot 3D Point Cloud Segmentation with Vision-Language ModelCode2
Generative Pre-trained Speech Language Model with Efficient Hierarchical TransformerCode2
GeoReasoner: Geo-localization with Reasoning in Street Views using a Large Vision-Language ModelCode2
TrustRAG: Enhancing Robustness and Trustworthiness in RAGCode2
Composed Image Retrieval for Remote SensingCode2
G1: Bootstrapping Perception and Reasoning Abilities of Vision-Language Model via Reinforcement LearningCode2
Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking CapabilitiesCode2
Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model InferenceCode2
GAMA: A Large Audio-Language Model with Advanced Audio Understanding and Complex Reasoning AbilitiesCode2
ARAGOG: Advanced RAG Output GradingCode2
From Words to Numbers: Your Large Language Model Is Secretly A Capable Regressor When Given In-Context ExamplesCode2
Compression Represents Intelligence LinearlyCode2
Frontiers in Intelligent ColonoscopyCode2
FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows"Code2
GeoVision Labeler: Zero-Shot Geospatial Classification with Vision and Language ModelsCode2
Show:102550
← PrevPage 20 of 284Next →

No leaderboard results yet.