|
--- |
|
license: cc-by-sa-4.0 |
|
language: |
|
- de |
|
- en |
|
- es |
|
- da |
|
- pl |
|
- sv |
|
- nl |
|
metrics: |
|
- accuracy |
|
pipeline_tag: text-classification |
|
tags: |
|
- partypress |
|
- political science |
|
- parties |
|
- press releases |
|
--- |
|
|
|
|
|
# PARTYPRESS multilingual |
|
|
|
Fine-tuned model in seven languages on texts from nine countries, based on [bert-base-multilingual-cased](https://huggingface.co./bert-base-multilingual-cased). It used in Erfort et al. (2023). |
|
|
|
|
|
## Model description |
|
|
|
tbs |
|
|
|
## Model variations |
|
|
|
tbd (monolingual) |
|
|
|
## Intended uses & limitations |
|
|
|
tbd |
|
|
|
### How to use |
|
|
|
tbd |
|
|
|
### Limitations and bias |
|
|
|
tbd |
|
|
|
## Training data |
|
|
|
For the training data, please refer to [bert-base-multilingual-cased](https://huggingface.co./bert-base-multilingual-cased) |
|
|
|
## Training procedure |
|
|
|
### Preprocessing |
|
|
|
For the preprocessing, please refer to [bert-base-multilingual-cased](https://huggingface.co./bert-base-multilingual-cased) |
|
|
|
### Pretraining |
|
|
|
For the pretraining, please refer to [bert-base-multilingual-cased](https://huggingface.co./bert-base-multilingual-cased) |
|
|
|
## Evaluation results |
|
|
|
Fine-tuned on our downstream task, this model achieves the following results: |
|
|
|
|
|
### BibTeX entry and citation info |
|
|
|
```bibtex |
|
@article{erfort_partypress_2023, |
|
author = {Cornelius Erfort and |
|
Lukas F. Stoetzer and |
|
Heike Klüver}, |
|
title = {The PARTYPRESS Database: A New Comparative Database of Parties’ Press Releases}, |
|
journal = {Research and Politics}, |
|
volume = {forthcoming}, |
|
year = {2023}, |
|
} |
|
``` |
|
|
|
|
|
|
|
|