inflaton commited on
Commit
4a18d7a
·
1 Parent(s): 127b0e9

llama3 p2 epoch 1 results

Browse files
llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml CHANGED
@@ -25,7 +25,7 @@ output_dir: saves/llama3-8b/lora/sft_bf16_p1_full
25
  logging_steps: 10
26
  save_steps: 175
27
  plot_loss: true
28
- #overwrite_output_dir: true
29
 
30
  ### train
31
  per_device_train_batch_size: 16
 
25
  logging_steps: 10
26
  save_steps: 175
27
  plot_loss: true
28
+ overwrite_output_dir: true
29
 
30
  ### train
31
  per_device_train_batch_size: 16
results/llama3-8b_lora_sft_bf16-p2.csv CHANGED
The diff for this file is too large to render. See raw diff
 
scripts/eval-mgtv-llama3_8b.sh CHANGED
@@ -17,7 +17,6 @@ grep MemTotal /proc/meminfo
17
  #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
18
 
19
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
20
- export START_EPOCH=1
21
  export RESIZE_TOKEN_EMBEDDINGS=true
22
 
23
  #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
@@ -27,11 +26,13 @@ export MODEL_PREFIX=llama3-8b_lora_sft_bf16
27
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
28
  export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
29
  export USING_P1_PROMPT_TEMPLATE=true
 
30
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
31
- python llm_toolkit/eval_logical_reasoning_all_epochs.py
32
 
33
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
34
  export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
35
  export USING_P1_PROMPT_TEMPLATE=false
 
36
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
37
  python llm_toolkit/eval_logical_reasoning_all_epochs.py
 
17
  #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
18
 
19
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
 
20
  export RESIZE_TOKEN_EMBEDDINGS=true
21
 
22
  #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
 
26
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
27
  export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
28
  export USING_P1_PROMPT_TEMPLATE=true
29
+ export START_EPOCH=3
30
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
31
+ #python llm_toolkit/eval_logical_reasoning_all_epochs.py
32
 
33
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
34
  export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
35
  export USING_P1_PROMPT_TEMPLATE=false
36
+ export START_EPOCH=1
37
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
38
  python llm_toolkit/eval_logical_reasoning_all_epochs.py