Jokers
Collection
Models for jokes generation
•
4 items
•
Updated
This model is a fine-tuned version of igorktech/rugpt3-joker-150k on the baneks dataset for 10 epochs. It achieved 2.0391
overall loss during training.
Model evaluation has not been performed.
The model is a fine-tuned variant of the igorktech/rugpt3-joker-150k architecture with causal language modeling head.
The model should be used for studying abilities of natural language models to generate jokes.
The model is trained on a list of anecdotes pulled from a few vk communities (see baneks dataset for more details).
The following hyperparameters were used during training:
Train Loss | Epoch |
---|---|
2.0391 | 10 |
Base model
igorktech/rugpt3-joker-150k