Skip to content

YoanGab/intents_classification_for_neural_text_generation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

50 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MLP Architectures for Emotion and Sentiment Analysis: A Comparative Study of Fine-Tuned BERT and RoBERTa Models

Abstract

This paper presents a comparative study of multi-layer perceptron (MLP) architectures for Emotion and Sentiment (E/S) analysis using fine-tuned BERT and RoBERTa models. Both E/S recognition are important in understanding the opinions and attitudes of users towards products, services, and brands. Thus, it is essential to develop accurate and effective E/S detection approaches. We use five datasets from the SILICONE benchmark for training and we evaluated the performance using the F1 weighted average score. Our experiments show that BERT with MLP, consistently outperforms RoBERTa across all models and datasets. Our findings suggest that using an MLP architecture on top of a pre-trained LLM can improve the performance of emotion and sentiment analysis tasks.

Getting start

  1. Clone the repository:
git clone https://github.com/YoanGab/intents_classification_for_neural_text_generation.git
cd intents_classification_for_neural_text_generation
  1. Download dependencies
python -m venv intent_classification
source intent_classification/bin/activate
pip install --upgrade pip
pip install -r requirements.txt
  1. Run the package a. Fine-tuned a model on a dataset
python -m intents_classification --mode train

b. Test a fine-tuned model

python -m intents_classification --mode test --model_path path/to/model.pkl

Experimental results

Average $DyDA_e$ $IEMO$ $MELD_e$ $MELD_s$ $SEM$
BERT Baseline 0.231 0.204 0.095 0.182 0.324 0.350
Fine-tuned Baseline 0.386 0.420 0.170 0.399 0.518 0.423
Three Layers Perceptron 0.551 0.734 0.249 0.545 0.643 0.585
Four Layers Perceptron 0.448 0.542 0.195 0.443 0.581 0.479
RoBERTa Baseline 0.263 0.355 0.121 0.191 0.324 0.325
Fine-tuned Baseline 0.297 0.406 0.130 0.193 0.391 0.367
Three Layers Perceptron 0.435 0.734 0.108 0.313 0.527 0.495
Four Layers Perceptron 0.378 0.470 0.100 0.369 0.491 0.458

Authors

References:

[1] Emile Chapuis,Pierre Colombo, Matthieu Labeau, and Chloé Clavel. Code-switched inspired losses for generic spoken dialog representations. EMNLP 2021

[2] Emile Chapuis,Pierre Colombo, Matteo Manica, Matthieu Labeau, and Chloé Clavel. Hierarchical pre-training for sequence labelling in spoken dialog. Finding of EMNLP 2020

[3]Tanvi Dinkar, Pierre Colombo , Matthieu Labeau, and Chloé Clavel. The importance of fillers for text representations of speech transcripts. EMNLP 2020

[4] Hamid Jalalzai, Pierre Colombo , Chloe Clavel, Eric Gaussier, Giovanna Varni, Emmanuel Vignon, and Anne Sabourin. Heavy-tailed representations, text polarity classification & data augmentation. NeurIPS 2020

[5] Pierre Colombo, Emile Chapuis, Matteo Manica, Emmanuel Vignon, Giovanna Varni, and Chloé Clavel. Guiding attention in sequence-to-sequence models for dialogue act prediction. (oral) AAAI 2020

[6] Alexandre Garcia,Pierre Colombo, Slim Essid, Florence d’Alché-Buc, and Chloé Clavel. From the token to the review: A hierarchical multimodal approach to opinion mining. EMNLP 2020

[7] Pierre Colombo, Wojciech Witon, Ashutosh Modi, James Kennedy, and Mubbasir Kapadia. Affect-driven dialog generation. NAACL 2019

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages