orionweller commited on
Commit
51a8cb4
·
verified ·
1 Parent(s): 28159db

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +29 -0
  2. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  4. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  5. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  6. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  7. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds +3 -0
  8. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  9. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  10. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  11. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds +3 -0
  12. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds +3 -0
  13. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  14. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds +3 -0
  15. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds +3 -0
  16. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds +3 -0
  17. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds +3 -0
  18. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds +3 -0
  19. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  20. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds +3 -0
  21. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds +3 -0
  22. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  23. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  24. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds +3 -0
  25. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds +3 -0
  26. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds +3 -0
  27. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds +3 -0
  28. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00045.mds +3 -0
  29. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds +3 -0
  30. train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00047.mds +3 -0
  31. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11399-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11399-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1371-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1371-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16032-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16032-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18793-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18793-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19876-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19876-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23108-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23108-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23118-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23118-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23873-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23873-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23924-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23924-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24262-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24262-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -11825,3 +11825,32 @@ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/
11825
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
11826
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11827
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11825
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
11826
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11827
  train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11828
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds filter=lfs diff=lfs merge=lfs -text
11829
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11830
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11831
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
11832
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds filter=lfs diff=lfs merge=lfs -text
11833
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00047.mds filter=lfs diff=lfs merge=lfs -text
11834
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds filter=lfs diff=lfs merge=lfs -text
11835
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
11836
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds filter=lfs diff=lfs merge=lfs -text
11837
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11838
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text
11839
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
11840
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11841
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11842
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
11843
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11844
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text
11845
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00045.mds filter=lfs diff=lfs merge=lfs -text
11846
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text
11847
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds filter=lfs diff=lfs merge=lfs -text
11848
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text
11849
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text
11850
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds filter=lfs diff=lfs merge=lfs -text
11851
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text
11852
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
11853
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds filter=lfs diff=lfs merge=lfs -text
11854
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
11855
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11856
+ train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c7ae5582a3db54815c76484a82491bf0293294f1f0c86bea2e07c4e1aad7fe7
3
+ size 67108479
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46d9bcc0cbdb830abe785ef7b34a015b3f2fd1addd3306de3183e3651691a4a0
3
+ size 67108333
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d18b3a731d72389985067f0eac2a8f49cd7d170448d17dcf0dccbf6091980bbe
3
+ size 67106948
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a28afc264d043a9b150b7a136bcf4be59fcaeec94f89f68ad4a2a7a6b88e88e0
3
+ size 67107128
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b677237978611768b1d85ef81f3decd7eea1595889dad0173dd31891b14f835f
3
+ size 67106992
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44768999f1ef112479ae383be113c7474f354a77b3e7792a48e51f8a29e9d9f3
3
+ size 67108561
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bad02552a2194656655e549632105111de03f15849d73008dc0936ed8f8fcfd4
3
+ size 67108429
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:671d5fc509a8917d530b467e16735d788e4603d7a0e35eeef0fd90fc19a03b2c
3
+ size 67106871
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97d83e645d752d51c394b497cdbd1e2e3f088570a9e291951874b576118c098e
3
+ size 67108771
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d21bb824d2a458eeffe940a8c45014bec43291b1d6ba37f828d9e8ac334dde2a
3
+ size 67108388
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ca7ef7d566a14252764e6afad8c5e5e16d5c66fa5072f4ac76a7c99ee654faa
3
+ size 67108811
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52f28965f2115a38cde75bc2e422e1c1dd3a60d575a1b23b78baac15bc5e0a1f
3
+ size 67108552
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0eb03e49ba085bac43f7f61abbf09af13b3dcaeede6644b96359f61846ce3a
3
+ size 67107338
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30d4fae6a6a81eb6c8499a30aa4c6fa293d1077329b27696b02e01f1636f0071
3
+ size 67107023
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6255e267c233028d089f82dc607a81669ec7654c7c55ef13b773355fd9c77abe
3
+ size 67107746
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3f36e25671d82d02bea0518389de11e7436a3275e1cfcdfa00090468876f832
3
+ size 67108209
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab4414877c3da4ad58cb90c2131f56f7f2ff93804d0655bd044d95eb3f00c14f
3
+ size 67108230
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00a9efbef7d7e877a25d1ee450c05141afdddb6515a2a491a9462978de0b0f68
3
+ size 67107199
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:661ad65cd9828010bf124bb114780d3c378fc3fc861623d5b9ad5e5702b98eb6
3
+ size 67107832
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72a4d8d49749f73d790596429ba41c7b0c191b4b601105abb20e74ad3c386510
3
+ size 67107825
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f3dca60b2377675113be2e4cf58cefa097ca673eb3f4126ebcaf3714d392ba4
3
+ size 67107402
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6dbe182a5e5e5cd868968bbfe2b97748e7414b595043e354bd0d03ed7371135
3
+ size 67108366
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:408d57b13a15ccbcad08e82d8898473115aa593156c1d23afeb5eb4d433345d3
3
+ size 67108744
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30b935c6544d1df9514a361cd23f44e4e9eefb5183f90bc018a3208d874eaa5a
3
+ size 67108001
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c5ac59c4ad9ced4a200cfe23ded928cca07ef3bb73e2716bccf8a56973ce41
3
+ size 67108040
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d9ec018d2ff241c57b87ac05b256d2960e11502ca09ad006568594724f7c8bf
3
+ size 67107755
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00045.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:352d172d94a54c9aec9ae851f8184e11f4038a15584e2c87a6c837ae8f413942
3
+ size 67108224
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8ca79ca3df75d9ad07d649355f824a07c1424ba19a00427d00b62eeca53a2d1
3
+ size 67107416
train/cc_en_head/cc_en_head_0041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00047.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ee350a71f4a874613d38020694b73ebd7bf39822bacd80898d38781d1caf8e
3
+ size 41784911
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11399-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108812, "hashes": {}}, "samples": 43344, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47550560, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17312904, "hashes": {}}, "samples": 11213, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12302402, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11399-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40522300,
3
+ "num_truncated_tokens": 40488300
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1371-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108540, "hashes": {}}, "samples": 44738, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48049050, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 7983718, "hashes": {}}, "samples": 5329, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5736451, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1371-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 35996805,
3
+ "num_truncated_tokens": 35969692
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16032-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108484, "hashes": {}}, "samples": 44601, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47976281, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9592981, "hashes": {}}, "samples": 6166, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6808211, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16032-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36779635,
3
+ "num_truncated_tokens": 36751178
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18793-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108290, "hashes": {}}, "samples": 43796, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47667812, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14866725, "hashes": {}}, "samples": 9554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10530779, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18793-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39336404,
3
+ "num_truncated_tokens": 39304137
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19876-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107620, "hashes": {}}, "samples": 43225, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48012005, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16626823, "hashes": {}}, "samples": 10796, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11913822, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19876-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40195061,
3
+ "num_truncated_tokens": 40161557
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23108-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108357, "hashes": {}}, "samples": 44266, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47820862, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10754563, "hashes": {}}, "samples": 7165, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7677834, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23108-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37339900,
3
+ "num_truncated_tokens": 37310681
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23118-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108673, "hashes": {}}, "samples": 43661, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47481705, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12969275, "hashes": {}}, "samples": 8425, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9304737, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23118-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38426715,
3
+ "num_truncated_tokens": 38396996
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23873-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108831, "hashes": {}}, "samples": 43558, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47741528, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15271795, "hashes": {}}, "samples": 9935, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10801294, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23873-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39534681,
3
+ "num_truncated_tokens": 39502277
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23924-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108023, "hashes": {}}, "samples": 43671, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47765037, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13222318, "hashes": {}}, "samples": 8771, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9442209, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23924-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38542098,
3
+ "num_truncated_tokens": 38511609
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24262-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108575, "hashes": {}}, "samples": 42901, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47597903, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19154886, "hashes": {}}, "samples": 12345, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13577455, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24262-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41421709,
3
+ "num_truncated_tokens": 41386190
4
+ }