elsayan123123 commited on
Commit
bc055a7
·
1 Parent(s): 5aafbb1

add model-centric approach in README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -1
README.md CHANGED
@@ -11,4 +11,36 @@ license: apache-2.0
11
  short_description: lab2
12
  ---
13
 
14
- An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  short_description: lab2
12
  ---
13
 
14
+ # Improving Model Performance: Model-Centric Approach
15
+
16
+ ## Model-Centric Approach
17
+ To enhance the performance of our model, we adopted a **model-centric strategy**, specifically tuning the rank of the LoRA (Low-Rank Adaptation) model. The evaluation results below show the model's performance across three datasets: **HellaSwag**, **WinoGrande**, and **Abstract Algebra**. By adjusting the rank (denoted as `r`), we observed meaningful variations in accuracy across tasks, demonstrating that the choice of rank is critical for model optimization.
18
+
19
+ ### Datasets
20
+ 1. **HellaSwag**:
21
+ A benchmark designed to test a model's ability to complete sentences based on commonsense reasoning. It features adversarial examples to ensure that high performance requires deep understanding rather than superficial pattern matching.
22
+
23
+ 2. **WinoGrande**:
24
+ A dataset focused on commonsense pronoun resolution. Each question presents an ambiguous pronoun and requires selecting the correct referent, which is challenging for models due to the subtle reasoning involved.
25
+
26
+ 3. **Abstract Algebra**:
27
+ A dataset that evaluates a model's understanding of advanced mathematical concepts, specifically algebraic structures like groups, rings, and fields.
28
+
29
+ ### Evaluation Results
30
+
31
+ | Rank | HellaSwag | WinoGrande | Abstract Algebra |
32
+ |------|-----------------|-----------------|------------------|
33
+ | r8 | 0.591 | <span style="color:red">0.200</span> | 0.595 |
34
+ | r16 | **<span style="color:green">0.592</span>** | 0.210 | 0.594 |
35
+ | r32 | 0.590 | **<span style="color:green">0.230</span>** | **<span style="color:green">0.597</span>** |
36
+ | r64 | <span style="color:red">0.588</span> | 0.210 | <span style="color:red">0.585</span> |
37
+
38
+ ### Findings
39
+ - On **HellaSwag**, the performance peaked at rank **r16** (0.592) but showed minimal sensitivity to rank changes overall.
40
+ - For **WinoGrande**, performance improved consistently up to rank **r32**, reaching 0.230. This demonstrates that more capacity (via higher rank) benefits tasks requiring subtle reasoning.
41
+ - In **Abstract Algebra**, rank **r32** yielded the best result (0.597), while larger ranks (e.g., r64) slightly degraded performance, likely due to overfitting.
42
+
43
+ **Conclusion**: These findings suggest that while increasing the rank provides more capacity for adaptation and can improve performance, overly large ranks can lead to diminishing returns and overfitting. Careful tuning is essential to balance these effects and achieve optimal results.
44
+
45
+
46
+