thesven commited on
Commit
b927bdb
1 Parent(s): 7ed7192

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -2142,12 +2142,15 @@
2142
  "rms_norm_eps": 1e-05,
2143
  "rope_scaling": {
2144
  "factor": 8.0,
2145
- "type": "linear"
 
 
 
2146
  },
2147
  "rope_theta": 500000.0,
2148
  "tie_word_embeddings": false,
2149
  "torch_dtype": "float16",
2150
- "transformers_version": "4.42.4",
2151
  "use_cache": true,
2152
  "vocab_size": 128256
2153
  }
 
2142
  "rms_norm_eps": 1e-05,
2143
  "rope_scaling": {
2144
  "factor": 8.0,
2145
+ "high_freq_factor": 4.0,
2146
+ "low_freq_factor": 1.0,
2147
+ "original_max_position_embeddings": 8192,
2148
+ "rope_type": "llama3"
2149
  },
2150
  "rope_theta": 500000.0,
2151
  "tie_word_embeddings": false,
2152
  "torch_dtype": "float16",
2153
+ "transformers_version": "4.43.1",
2154
  "use_cache": true,
2155
  "vocab_size": 128256
2156
  }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
- "transformers_version": "4.42.4"
12
  }
 
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.43.1"
12
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67ddba8a0d4956a1f0e929f06403a2185d0dfa5d9ec96afa10d43f326de1b06b
3
  size 4682270360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcacc85ff3f62946ee7f31c5c2df897adcec2da3bf35bd76fd407dcdd54a00b7
3
  size 4682270360