Uploaded better trained version.
Browse files- README.md +3 -1
- adapter_model.safetensors +1 -1
- images/banner.avif +0 -0
README.md
CHANGED
@@ -10,9 +10,11 @@ pipeline_tag: text-generation
|
|
10 |
library_name: peft
|
11 |
---
|
12 |
# GPT4chan 24B QLoRA
|
|
|
|
|
13 |
This model is [mistralai/Mistral-Small-24B-Base-2501](https://huggingface.co/mistralai/Mistral-Small-24B-Base-2501) QLoRA fine-tuned on [v2ray/4chan](https://huggingface.co/datasets/v2ray/4chan) using [QLoRA](https://github.com/LagPixelLOL/qlora).
|
14 |
|
15 |
-
Trained using 8x H100 with global batch size 64, using 2e-4 learning rate, for
|
16 |
## Prompt Format
|
17 |
```
|
18 |
board<|start_header_id|>id<|end_header_id|>content<|start_header_id|>id<|end_header_id|>content...<|start_header_id|>id<|end_header_id|>
|
|
|
10 |
library_name: peft
|
11 |
---
|
12 |
# GPT4chan 24B QLoRA
|
13 |
+

|
14 |
+
|
15 |
This model is [mistralai/Mistral-Small-24B-Base-2501](https://huggingface.co/mistralai/Mistral-Small-24B-Base-2501) QLoRA fine-tuned on [v2ray/4chan](https://huggingface.co/datasets/v2ray/4chan) using [QLoRA](https://github.com/LagPixelLOL/qlora).
|
16 |
|
17 |
+
Trained using 8x H100 with global batch size 64, using 2e-4 learning rate, for 4000 steps, which is approximately 5 epochs.
|
18 |
## Prompt Format
|
19 |
```
|
20 |
board<|start_header_id|>id<|end_header_id|>content<|start_header_id|>id<|end_header_id|>content...<|start_header_id|>id<|end_header_id|>
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3458585544
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:763fa36edb7a6aa96ad98855ae7b89837559a0e7aad7b996512a1e72e8e91f9a
|
3 |
size 3458585544
|
images/banner.avif
ADDED
![]() |