Push model using huggingface_hub.
Browse files- README.md +3 -3
- adapter_config.json +2 -2
- adapter_model.safetensors +1 -1
- config.json +5 -5
- pytorch_model.bin +1 -1
README.md
CHANGED
@@ -25,7 +25,7 @@ You can then generate text as follows:
|
|
25 |
```python
|
26 |
from transformers import pipeline
|
27 |
|
28 |
-
generator = pipeline("text-generation", model="DarshanDeshpande//tmp/
|
29 |
outputs = generator("Hello, my llama is cute")
|
30 |
```
|
31 |
|
@@ -35,8 +35,8 @@ If you want to use the model for training or to obtain the outputs from the valu
|
|
35 |
from transformers import AutoTokenizer
|
36 |
from trl import AutoModelForCausalLMWithValueHead
|
37 |
|
38 |
-
tokenizer = AutoTokenizer.from_pretrained("DarshanDeshpande//tmp/
|
39 |
-
model = AutoModelForCausalLMWithValueHead.from_pretrained("DarshanDeshpande//tmp/
|
40 |
|
41 |
inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
|
42 |
outputs = model(**inputs, labels=inputs["input_ids"])
|
|
|
25 |
```python
|
26 |
from transformers import pipeline
|
27 |
|
28 |
+
generator = pipeline("text-generation", model="DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
|
29 |
outputs = generator("Hello, my llama is cute")
|
30 |
```
|
31 |
|
|
|
35 |
from transformers import AutoTokenizer
|
36 |
from trl import AutoModelForCausalLMWithValueHead
|
37 |
|
38 |
+
tokenizer = AutoTokenizer.from_pretrained("DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
|
39 |
+
model = AutoModelForCausalLMWithValueHead.from_pretrained("DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
|
40 |
|
41 |
inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
|
42 |
outputs = model(**inputs, labels=inputs["input_ids"])
|
adapter_config.json
CHANGED
@@ -19,8 +19,8 @@
|
|
19 |
"rank_pattern": {},
|
20 |
"revision": null,
|
21 |
"target_modules": [
|
22 |
-
"
|
23 |
-
"
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM",
|
26 |
"use_dora": false,
|
|
|
19 |
"rank_pattern": {},
|
20 |
"revision": null,
|
21 |
"target_modules": [
|
22 |
+
"v_proj",
|
23 |
+
"q_proj"
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM",
|
26 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3695848
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:60130f2e757132e31170ed5b40cd440c9ecd5d2351a2accc897956f2951fb4f4
|
3 |
size 3695848
|
config.json
CHANGED
@@ -26,9 +26,9 @@
|
|
26 |
"cliprange": 0.2,
|
27 |
"cliprange_value": 0.2,
|
28 |
"vf_coef": 0.1,
|
29 |
-
"batch_size":
|
30 |
"forward_batch_size": null,
|
31 |
-
"mini_batch_size":
|
32 |
"gradient_accumulation_steps": 1,
|
33 |
"world_size": 1,
|
34 |
"ppo_epochs": 1,
|
@@ -45,7 +45,7 @@
|
|
45 |
"whiten_rewards": false,
|
46 |
"is_encoder_decoder": false,
|
47 |
"is_peft_model": true,
|
48 |
-
"backward_batch_size":
|
49 |
-
"global_backward_batch_size":
|
50 |
-
"global_batch_size":
|
51 |
}
|
|
|
26 |
"cliprange": 0.2,
|
27 |
"cliprange_value": 0.2,
|
28 |
"vf_coef": 0.1,
|
29 |
+
"batch_size": 25,
|
30 |
"forward_batch_size": null,
|
31 |
+
"mini_batch_size": 25,
|
32 |
"gradient_accumulation_steps": 1,
|
33 |
"world_size": 1,
|
34 |
"ppo_epochs": 1,
|
|
|
45 |
"whiten_rewards": false,
|
46 |
"is_encoder_decoder": false,
|
47 |
"is_peft_model": true,
|
48 |
+
"backward_batch_size": 25,
|
49 |
+
"global_backward_batch_size": 25,
|
50 |
+
"global_batch_size": 25
|
51 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 9724
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4848be04eebc4fd14531016e726563a8291df25b605763e41bc037ec0b5e0ac9
|
3 |
size 9724
|