SOTAVerified

multimodal generation

Multimodal generation refers to the process of generating outputs that incorporate multiple modalities, such as images, text, and sound. This can be done using deep learning models that are trained on data that includes multiple modalities, allowing the models to generate output that is informed by more than one type of data.

For example, a multimodal generation model could be trained to generate captions for images that incorporate both text and visual information. The model could learn to identify objects in the image and generate descriptions of them in natural language, while also taking into account contextual information and the relationships between the objects in the image.

Multimodal generation can also be used in other applications, such as generating realistic images from textual descriptions or generating audio descriptions of video content. By combining multiple modalities in this way, multimodal generation models can produce more accurate and comprehensive output, making them useful for a wide range of applications.

Papers

Showing 7698 of 98 papers

TitleStatusHype
3D-VLA: A 3D Vision-Language-Action Generative World Model0
A Survey of Emerging Approaches and Advances in Video Generation0
Artificial Intelligence in Creative Industries: Advances Prior to 20250
ACDC: Autoregressive Coherent Multimodal Generation using Diffusion Correction0
PCQA: A Strong Baseline for AIGC Quality Assessment Based on Prompt Condition0
Pisces: An Auto-regressive Foundation Model for Image Understanding and Generation0
PlanMoGPT: Flow-Enhanced Progressive Planning for Text to Motion Synthesis0
Preliminary Explorations with GPT-4o(mni) Native Image Generation0
ProtTeX: Structure-In-Context Reasoning and Editing of Proteins with Large Language Models0
RDPM: Solve Diffusion Probabilistic Models via Recurrent Token Prediction0
Unconditional Image-Text Pair Generation with Multimodal Cross QuantizerCode0
PixelBytes: Catching Unified Embedding for Multimodal GenerationCode0
PixelBytes: Catching Unified Representation for Multimodal GenerationCode0
D-Judge: How Far Are We? Evaluating the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal GuidanceCode0
Multimedia Generative Script Learning for Task PlanningCode0
Pre-trained Encoder Inference: Revealing Upstream Encoders In Downstream Machine Learning ServicesCode0
Multimodal Generation of Novel Action Appearances for Synthetic-to-Real Recognition of Activities of Daily LivingCode0
Multi-modal Latent DiffusionCode0
Multimodal Latent Language Modeling with Next-Token DiffusionCode0
Consistent Multimodal Generation via A Unified GAN FrameworkCode0
Empathic Grounding: Explorations using Multimodal Interaction and Large Language Models with Conversational AgentsCode0
Continual and Multi-Task Architecture SearchCode0
Accountable Textual-Visual Chat Learns to Reject Human Instructions in Image Re-creationCode0
Show:102550
← PrevPage 4 of 4Next →

No leaderboard results yet.