English-Romanian Translate

v1.0 v2.0 v3.0

Finetune

This model is a finetune of the Helsinki-NLP/opus-mt-en-ro model, on ccmatrix(full) dataset.

How to use

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("BlackKakapo/opus-mt-en-ro")
model = AutoModelForSeq2SeqLM.from_pretrained("BlackKakapo/opus-mt-en-ro")

Benchmarks

Model Test Dataset BLEU Translation Length Reference Length
BlackKakapo/opus-mt-en-ro ("Helsinki-NLP/tatoeba_mt", "eng-ron", "test") 0.53118 40516 40352
Helsinki-NLP/opus-mt-en-ro ("Helsinki-NLP/tatoeba_mt", "eng-ron", "test") 0.45844 39222 40352
Helsinki-NLP/opus-tatoeba-en-ro ("Helsinki-NLP/tatoeba_mt", "eng-ron", "test") 0.46994 39423 40352
Downloads last month
306
Safetensors
Model size
74.7M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train BlackKakapo/opus-mt-en-ro

Space using BlackKakapo/opus-mt-en-ro 1