Spaces:
Build error
Build error
llama3 p2 epoch 1 results
Browse files
llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml
CHANGED
@@ -25,7 +25,7 @@ output_dir: saves/llama3-8b/lora/sft_bf16_p1_full
|
|
25 |
logging_steps: 10
|
26 |
save_steps: 175
|
27 |
plot_loss: true
|
28 |
-
|
29 |
|
30 |
### train
|
31 |
per_device_train_batch_size: 16
|
|
|
25 |
logging_steps: 10
|
26 |
save_steps: 175
|
27 |
plot_loss: true
|
28 |
+
overwrite_output_dir: true
|
29 |
|
30 |
### train
|
31 |
per_device_train_batch_size: 16
|
results/llama3-8b_lora_sft_bf16-p2.csv
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
scripts/eval-mgtv-llama3_8b.sh
CHANGED
@@ -17,7 +17,6 @@ grep MemTotal /proc/meminfo
|
|
17 |
#cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
|
18 |
|
19 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
20 |
-
export START_EPOCH=1
|
21 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
22 |
|
23 |
#export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
|
@@ -27,11 +26,13 @@ export MODEL_PREFIX=llama3-8b_lora_sft_bf16
|
|
27 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
|
28 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
|
29 |
export USING_P1_PROMPT_TEMPLATE=true
|
|
|
30 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
31 |
-
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
32 |
|
33 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
|
34 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
|
35 |
export USING_P1_PROMPT_TEMPLATE=false
|
|
|
36 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
37 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
|
|
17 |
#cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
|
18 |
|
19 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
|
|
20 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
21 |
|
22 |
#export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
|
|
|
26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
|
27 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
|
28 |
export USING_P1_PROMPT_TEMPLATE=true
|
29 |
+
export START_EPOCH=3
|
30 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
31 |
+
#python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
32 |
|
33 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
|
34 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
|
35 |
export USING_P1_PROMPT_TEMPLATE=false
|
36 |
+
export START_EPOCH=1
|
37 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
38 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|