Training multilingual models for sentiment analysis: by fine-tuning pre-trained multilingual language models such as multiBERT, XLMRoberta, etc. on the sentiment prediction task and comparing the performance of such multilingual models with (1) monolingual English model trained on English tweets and test on Arabic tweets translated to English (with pre-trained machine translation or Google Translate) and (2) multilingual model trained on English tweets and test on Arabic tweets (so, zero-shot classification). Do multilingual models benefit from being trained on multilingual data?
-
Notifications
You must be signed in to change notification settings - Fork 3
Training multilingual models for sentiment analysis: by fine-tuning pre-trained multilingual language models such as multiBERT, XLMRoberta, etc. on the sentiment prediction task and comparing the performance of such multilingual models with (1) monolingual English model trained on English tweets and test on Arabic tweets translated to English (w…
License
alt113/CS505-Spring-MultiLingual-Twitter-Classification
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
Training multilingual models for sentiment analysis: by fine-tuning pre-trained multilingual language models such as multiBERT, XLMRoberta, etc. on the sentiment prediction task and comparing the performance of such multilingual models with (1) monolingual English model trained on English tweets and test on Arabic tweets translated to English (w…
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published