Edit model card

GPT-2

Fine tune gpt2 model on Urdu news dataset using a causal language modeling (CLM) objective.

How to use

You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we set a seed for reproducibility:

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("Imran1/gpt2-urdu-news")

model = AutoModelForCausalLM.from_pretrained("Imran1/gpt2-urdu-news")

Training data

I fine tune gpt2 for downstream task like text generation, only for 1000 sample so it may not be good. Due to resources limitation.

Evaluation results

training loss 3.042

Downloads last month
30
Safetensors
Model size
175M params
Tensor type
F32
ยท
U8
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Space using Imran1/gpt2-urdu-news 1