BlinkDL commited on
Commit
ea1f914
1 Parent(s): b97b0b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -24,10 +24,10 @@ ctx_len = 1024
24
  n_layer = 32
25
  n_embd = 4096
26
 
27
- (there are ctx_len 2048 and 4096 models though they might be slightly weaker at generating short contents)
28
 
29
  Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
30
- * Pile loss 1.8415
31
  * LAMBADA ppl 4.38, acc 67.18%
32
  * PIQA acc 76.06%
33
  * SC2016 acc 73.44%
 
24
  n_layer = 32
25
  n_embd = 4096
26
 
27
+ (There are ctx_len 2048 and 4096 models too. Use them only when your ctxlen is long. Might be slightly weaker for short ctxlens.)
28
 
29
  Final checkpoint: RWKV-4-Pile-7B-20221115-8047.pth : Trained on the Pile for 332B tokens.
30
+ * Pile loss 1.8415T
31
  * LAMBADA ppl 4.38, acc 67.18%
32
  * PIQA acc 76.06%
33
  * SC2016 acc 73.44%