|
--- |
|
library_name: peft |
|
base_model: google/flan-t5-large |
|
--- |
|
|
|
# Model Card for honicky/t5-short-story-character-extractor |
|
|
|
I trained this model as part of a learning project to build a children's story authoring tool for parents of young children. See http://www.storytime.glass/ |
|
This model takes in a short story and outputs a comma separated list of characters in the story. |
|
|
|
I'm not sure yet how useful this fine-tune is: rather it is for me to learn about the nuts and bolts of fine-tuning. |
|
|
|
|
|
## Model Details |
|
|
|
The model is a fine-tune of a sequence-to-sequence model [Flan T5 Large](https://huggingface.co./google/flan-t5-large), so a different architecture from decoder-only models like GPT*. Maybe this allows it to perform this transformation task (transform a story into a list of characters) using a smaller model? |
|
|
|
* Trained using `transformers.Seq2SeqTrainer` plus the corresponding collator, tokenizer etc |
|
|
|
- **Developed by:** RJ Honicky |
|
- **Model type:** Encoder-Decoder Transformer |
|
- **Language(s):** English (fine tune data set) |
|
- **License:** MIT |
|
- **Finetuned from model:** `google/flan-t5-large` |
|
|
|
### Model Sources |
|
|
|
- **Repository:** https://github.com/honicky/character-extraction |
|
- **Weights and Biases**: https://wandb.ai/honicky/t5_target_finetune_for_character_extraction/runs/mx57gh45 |
|
|
|
## Uses |
|
|
|
Primarily for use in https://github.com/honicky/story-time and for learning. |
|
|