File size: 815 Bytes
04e44ed
 
 
 
 
 
 
 
c2f5171
04e44ed
c2f5171
04e44ed
 
 
2616703
c2f5171
 
 
 
 
 
 
 
 
04e44ed
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
---
tags:
- text-generation
library_name: transformers
---

## Model description
Dalio Bot Pre-trained on Principles, fine-tuned on handwritten examples.

Pre-trained model: Jellywibble/dalio-pretrained-book-bs4-seed1 (based-off OPT30B)

Fine-tuning dataset: Jellywibble/dalio_handwritten-conversations

## Model Parameters
- 4xA40 (eff. batch size = 4)
- base_mode_name Jellywibble/dalio-pretrained-book-bs4-seed1
- dataset_name Jellywibble/dalio_handwritten-conversations
- block size 500
- per_device_train_batch_size 1
- gradient_accumulation steps 1
- learning_rate 2e-6
- seed 28
- validation split percentage 20
- hellaswag_sample_size 100

## Metrics
- Hellaswag Perplexity: 29.9
- Eval acc: 57.1%
- Eval loss: 1.971
- wandb: https://wandb.ai/jellywibble/huggingface/runs/12lgyt20?workspace=user-jellywibble