|
--- |
|
license: gemma |
|
datasets: |
|
- sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo |
|
- sam-paech/gutenbergs_1_2_3_antislop-dpo |
|
language: |
|
- en |
|
base_model: |
|
- google/gemma-2-9b-it |
|
library_name: transformers |
|
tags: |
|
- creative-writing |
|
--- |
|
# Oblivion's End |
|
|
|
A merged LoRA for gemma-2-9b-it, trained using DPO datasets for creative writing using [my DPO training notebook](https://github.com/mkturkcan/dpo-model-trainer). |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
|
|
- **Finetuned from model:** google/gemma-2-9b-it |
|
### Model Sources [optional] |
|
|
|
- **Repository:** [GitHub](https://github.com/mkturkcan/dpo-model-trainer/tree/main). |
|
|
|
## Uses |
|
|
|
Made for creative writing. |
|
|
|
## Training Details |
|
|
|
### Training Data |
|
|
|
Check out the model card details. |
|
|
|
### Training Procedure |
|
|
|
Model training performance (margins) are available in the [wandb instance](https://api.wandb.ai/links/mkturkcan/4djkmhwp). |
|
|
|
#### Training Hyperparameters |
|
|
|
- **Training regime:** bf16 on a 1x 80GB A100 node. |
|
|
|
## Model Examination [optional] |
|
|
|
<!-- Relevant interpretability work for the model goes here --> |
|
|
|
[More Information Needed] |
|
|
|
## Environmental Impact |
|
|
|
Total emissions are estimated to be 0.83 kgCO$_2$eq. |
|
|
|
|