Spaces:
Build error
Build error
change BATCH_SIZE to 1 for qwen2-72b eval
Browse files
llm_toolkit/eval_logical_reasoning.py
CHANGED
@@ -28,6 +28,7 @@ test_data = os.getenv("TEST_DATA", None)
|
|
28 |
using_llama_factory = os.getenv("USING_LLAMA_FACTORY") == "true"
|
29 |
max_new_tokens = int(os.getenv("MAX_NEW_TOKENS", 16))
|
30 |
repetition_penalty = float(os.getenv("REPETITION_PENALTY", 1.0))
|
|
|
31 |
|
32 |
dtype = (
|
33 |
torch.float32
|
@@ -84,7 +85,7 @@ predictions = eval_model(
|
|
84 |
datasets["test"],
|
85 |
max_new_tokens=max_new_tokens,
|
86 |
repetition_penalty=repetition_penalty,
|
87 |
-
batch_size=
|
88 |
)
|
89 |
|
90 |
gpu_stats = torch.cuda.get_device_properties(0)
|
|
|
28 |
using_llama_factory = os.getenv("USING_LLAMA_FACTORY") == "true"
|
29 |
max_new_tokens = int(os.getenv("MAX_NEW_TOKENS", 16))
|
30 |
repetition_penalty = float(os.getenv("REPETITION_PENALTY", 1.0))
|
31 |
+
batch_size = int(os.getenv("BATCH_SIZE", 2))
|
32 |
|
33 |
dtype = (
|
34 |
torch.float32
|
|
|
85 |
datasets["test"],
|
86 |
max_new_tokens=max_new_tokens,
|
87 |
repetition_penalty=repetition_penalty,
|
88 |
+
batch_size=batch_size,
|
89 |
)
|
90 |
|
91 |
gpu_stats = torch.cuda.get_device_properties(0)
|
notebooks/00_Data Analysis.ipynb
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
scripts/eval-mgtv-qwen2_72b.sh
CHANGED
@@ -19,9 +19,10 @@ export USING_LLAMA_FACTORY=true
|
|
19 |
export USING_P1_PROMPT_TEMPLATE=false
|
20 |
export LOAD_IN_4BIT=true
|
21 |
|
22 |
-
export START_EPOCH=
|
23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
24 |
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
|
|
25 |
|
26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
27 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
|
|
19 |
export USING_P1_PROMPT_TEMPLATE=false
|
20 |
export LOAD_IN_4BIT=true
|
21 |
|
22 |
+
export START_EPOCH=7
|
23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
24 |
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
25 |
+
export BATCH_SIZE=1
|
26 |
|
27 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
28 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|