SOTAVerified

LLaVAC: Fine-tuning LLaVA as a Multimodal Sentiment Classifier

2025-02-05Code Available0· sign in to hype

T. Chay-intr, Y. Chen, K. Viriyayudhakorn, T. Theeramunkong

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present LLaVAC, a method for constructing a classifier for multimodal sentiment analysis. This method leverages fine-tuning of the Large Language and Vision Assistant (LLaVA) to predict sentiment labels across both image and text modalities. Our approach involves designing a structured prompt that incorporates both unimodal and multimodal labels to fine-tune LLaVA, enabling it to perform sentiment classification effectively. Experiments on the MVSA-Single dataset demonstrate that LLaVAC outperforms existing methods in multimodal sentiment analysis across three data processing procedures. The implementation of LLaVAC is publicly available at https://github.com/tchayintr/llavac.

Tasks

Reproductions