SOTAVerified

multimodal generation

Multimodal generation refers to the process of generating outputs that incorporate multiple modalities, such as images, text, and sound. This can be done using deep learning models that are trained on data that includes multiple modalities, allowing the models to generate output that is informed by more than one type of data.

For example, a multimodal generation model could be trained to generate captions for images that incorporate both text and visual information. The model could learn to identify objects in the image and generate descriptions of them in natural language, while also taking into account contextual information and the relationships between the objects in the image.

Multimodal generation can also be used in other applications, such as generating realistic images from textual descriptions or generating audio descriptions of video content. By combining multiple modalities in this way, multimodal generation models can produce more accurate and comprehensive output, making them useful for a wide range of applications.

Papers

Showing 5160 of 98 papers

TitleStatusHype
Pre-trained Encoder Inference: Revealing Upstream Encoders In Downstream Machine Learning ServicesCode0
MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal ControlsCode2
Diffusion Models For Multi-Modal Generative Modeling0
Harmonizing Visual Text Comprehension and GenerationCode2
ANOLE: An Open, Autoregressive, Native Large Multimodal Models for Interleaved Image-Text GenerationCode4
Empathic Grounding: Explorations using Multimodal Interaction and Large Language Models with Conversational AgentsCode0
4M-21: An Any-to-Any Vision Model for Tens of Tasks and ModalitiesCode5
LLMs Meet Multimodal Generation and Editing: A SurveyCode4
The Evolution of Multimodal Model Architectures0
C3LLM: Conditional Multimodal Content Generation Using Large Language Models0
Show:102550
← PrevPage 6 of 10Next →

No leaderboard results yet.