XLM-R-Large-Tweet

XLM-R-Large-Tweet is a version of the XLM-R-Large-Tweet-Base*, fine-tuned for sentiment analysis using 5,610 annotated Serbian COVID-19 vaccination-related tweets. Specifically, it is tailored for five-class sentiment analysis to capture finer sentiment nuances in the social media domain using the following scale: very negative, negative, neutral, positive, and very positive.

*XLM-R-Large-Tweet-Base is an additionally pretrained version of the XLM-RoBERTa large-sized model.

How to Use

To use the model, you can load it with the following code:

from transformers import AutoTokenizer, XLMRobertaForSequenceClassification

model_name = "DarijaM/XLM-R-Large-Tweet"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = XLMRobertaForSequenceClassification.from_pretrained(model_name)
Downloads last month
17
Inference API
Unable to determine this model's library. Check the docs .