File size: 3,823 Bytes
af436d9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
---
language: 
- hu
- sk
- pl
- cs
tags:
- emotion-classification
- roberta
- fine-tuned
- multilingual

license: mit
datasets:
- custom

model-index:
- name: Multilingual Fine-tuned RoBERTa for Emotion Classification
  results:
  - task:
      type: text-classification
      name: Multilingual Emotion Classification
    dataset:
      name: Multilingual Custom Dataset (Hungarian, Slovak, Polish, Czech)
      type: text
    metrics:
      - name: Precision (Macro Avg)
        type: precision
        value: 0.86
      - name: Recall (Macro Avg)
        type: recall
        value: 0.86
      - name: F1 Score (Macro Avg)
        type: f1
        value: 0.86
      - name: Accuracy
        type: accuracy
        value: 0.84

---

# Multilingual Fine-tuned RoBERTa Model for Emotion Classification

## Model Description
This model is a multilingual fine-tuned version of the [RoBERTa](https://huggingface.co./roberta-base) model, specifically tailored for emotion classification tasks in Hungarian, Slovak, Polish, and Czech languages. 
The model was trained to classify textual data into six emotional categories (**anger, fear, disgust, sadness, joy,** and **none of them**).

## Intended Use
This model is intended for classifying textual data into emotional categories across multiple languages, including Hungarian, Slovak, Polish, and Czech. 
It can be used in applications such as sentiment analysis, social media monitoring, customer feedback analysis, and similar tasks. 
The model predicts the dominant emotion in a given text among the six predefined categories.

## Metrics

| **Class**       | **Precision (P)** | **Recall (R)** | **F1-Score (F1)** |
|-----------------|-------------------|----------------|-------------------|
| **anger**       | 0.74              | 0.81           | 0.77              |
| **fear**        | 0.98              | 0.98           | 0.98              |
| **disgust**     | 0.94              | 0.95           | 0.95              |
| **sadness**     | 0.87              | 0.87           | 0.87              |
| **joy**         | 0.89              | 0.89           | 0.89              |
| **none of them**| 0.77              | 0.69           | 0.73              |
| **Accuracy**    |                   |                | **0.84**          |
| **Macro Avg**   | 0.86              | 0.86           | 0.86              |
| **Weighted Avg**| 0.84              | 0.84           | 0.84              |

### Overall Performance
- **Accuracy:** 0.84
- **Macro Average Precision:** 0.86
- **Macro Average Recall:** 0.86
- **Macro Average F1-Score:** 0.86

### Class-wise Performance
The model demonstrates strong performance across different emotional categories, with particularly high precision, recall, and F1 scores in the **fear**, **disgust**, and **joy** categories. 
The model performs moderately well in detecting **anger** and **none of them** categories, but still achieves adequate accuracy in these cases.

## Limitations
- **Context Sensitivity:** The model may struggle with recognizing emotions that require deeper contextual understanding.
- **Class Imbalance:** The model's performance on the "none of them" category suggests that further training with more balanced datasets could improve accuracy.
- **Generalization:** The model's performance may vary depending on the text's domain, language style, and length, especially across different languages.

## How to Use
You can use this model directly with the `transformers` library from Hugging Face. Below is an example of how to load and use the model:

```python
from transformers import pipeline

# Load the fine-tuned model
classifier = pipeline("text-classification", model="visegradmedia-emotion/Emotion_RoBERTa_pooled_V4")

# Example usage
result = classifier("Nagyon örömtelinek érzem magam ma!")
print(result)