Update README.md
Browse files
README.md
CHANGED
@@ -26,11 +26,11 @@ ctx_len = 1024
|
|
26 |
n_layer = 32
|
27 |
n_embd = 4096
|
28 |
|
|
|
|
|
29 |
Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
|
30 |
* Pile loss 1.8415
|
31 |
* LAMBADA ppl 4.38, acc 67.18%
|
32 |
* PIQA acc 76.06%
|
33 |
* SC2016 acc 73.44%
|
34 |
* Hellaswag acc_norm 65.51%
|
35 |
-
|
36 |
-
I am fine-tuning it to ctx_len 4096.
|
|
|
26 |
n_layer = 32
|
27 |
n_embd = 4096
|
28 |
|
29 |
+
(there are ctx_len 2048 and 4096 models though they might be slightly weaker at generating short contents)
|
30 |
+
|
31 |
Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
|
32 |
* Pile loss 1.8415
|
33 |
* LAMBADA ppl 4.38, acc 67.18%
|
34 |
* PIQA acc 76.06%
|
35 |
* SC2016 acc 73.44%
|
36 |
* Hellaswag acc_norm 65.51%
|
|
|
|