ibraheemmoosa commited on
Commit
2cb68d2
·
1 Parent(s): ebed999

Saving weights and logs of step 1

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "T5WithLMHeadModel"
4
  ],
5
  "d_ff": 2048,
6
  "d_kv": 64,
@@ -52,5 +52,5 @@
52
  },
53
  "transformers_version": "4.9.0.dev0",
54
  "use_cache": true,
55
- "vocab_size": 32128
56
  }
 
1
  {
2
  "architectures": [
3
+ "T5ForConditionalGeneration"
4
  ],
5
  "d_ff": 2048,
6
  "d_kv": 64,
 
52
  },
53
  "transformers_version": "4.9.0.dev0",
54
  "use_cache": true,
55
+ "vocab_size": 32103
56
  }
events.out.tfevents.1626249974.t1v-n-616e0bbc-w-0.583139.3.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd0ec6ac04c2dce6b3639ab982327312bf0830d0ff3e6b06e68869a5a1231ecf
3
+ size 532
events.out.tfevents.1626250221.t1v-n-616e0bbc-w-0.584780.3.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26e027ed7808463b8a057416bb1d3962dd13294f6fc330dead847ceb8e01720e
3
+ size 532
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3aeb17712163b42aeab40c21ee3f1614dd5331e16c42a60686eed4a2651609ba
3
+ size 241981002
t5-train.sh CHANGED
@@ -1,6 +1,6 @@
1
  #!/bin/bash
2
  HF_DATASETS_IN_MEMORY_MAX_SIZE=200000000
3
- MODEL_DIR=checkpoints
4
  TOKENIZER_DIR=tokenizer
5
  ./run_t5_mlm_flax.py \
6
  --output_dir="${MODEL_DIR}" \
 
1
  #!/bin/bash
2
  HF_DATASETS_IN_MEMORY_MAX_SIZE=200000000
3
+ MODEL_DIR=.
4
  TOKENIZER_DIR=tokenizer
5
  ./run_t5_mlm_flax.py \
6
  --output_dir="${MODEL_DIR}" \