Samvardhan777's picture
Update README.md
c9f1288 verified
|
raw
history blame
425 Bytes
metadata
Model description:
  Model type: >-
    A 2B parameter GPT-like model finetuned on 100,000 samples consisting of an
    equal proportion of English and German samples.
  Language(s): Bilingual. English and German.
  License: Google Gemma Terms of Use
  Finetuned from model: Samvardhan777/gemma-2b-mt-German-to-English
  Training Precision: bfloat16
  Training Hardware: Free Google Colab
license: mit
pipeline_tag: translation