gpt2-vietnamese / README.md
nhanv's picture
Update README.md
27f0101
|
raw
history blame
1.06 kB
metadata
language: vi
tags:
  - vi
  - vietnamese
  - gpt2
  - text-generation
  - lm
  - nlp
datasets:
  - oscar
widget:
  - text: 生命、宇宙、そして万物についての究極の疑問の答えは

GPT-2

Pretrained model on Vietnamese language using a causal language modeling (CLM) objective. It was introduced in this paper and first released at this page.

How to use the model

from transformers import GPT2Tokenizer, AutoModelForCausalLM

tokenizer = GPT2Tokenizer.from_pretrained("nhanv/vi-gpt2")

model = AutoModelForCausalLM.from_pretrained("nhanv/vi-gpt2")

Model architecture

A 12-layer, 768-hidden-size transformer-based language model.

Training

The model was trained on Vietnamese Oscar dataset (32 GB) to optimize a traditional language modelling objective on v3-8 TPU for around 6 days. It reaches around 13.4 perplexity on a chosen validation set from Oscar.