Quazim0t0 commited on
Commit
5b6e08a
·
verified ·
1 Parent(s): 2de1730

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -3
README.md CHANGED
@@ -9,14 +9,25 @@ tags:
9
  license: apache-2.0
10
  language:
11
  - en
 
 
 
 
 
 
12
  ---
13
 
14
  # Uploaded model
15
 
16
  - **Developed by:** Quazim0t0
17
- - **License:** apache-2.0
18
  - **Finetuned from model :** unsloth/phi-4-unsloth-bnb-4bit
 
 
 
 
 
 
 
19
 
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
9
  license: apache-2.0
10
  language:
11
  - en
12
+ datasets:
13
+ - bespokelabs/Bespoke-Stratos-17k
14
+ - bespokelabs/Bespoke-Stratos-35k
15
+ - NovaSky-AI/Sky-T1_data_17k
16
+ - Quazim0t0/BenfordsLawReasoningJSON
17
+ - open-thoughts/OpenThoughts-114k
18
  ---
19
 
20
  # Uploaded model
21
 
22
  - **Developed by:** Quazim0t0
 
23
  - **Finetuned from model :** unsloth/phi-4-unsloth-bnb-4bit
24
+ - **GGUF**
25
+ - **Trained for 8 Hours on A800 with the Bespoke Stratos 17k Dataset.**
26
+ - **Trained for 6 Hours on A800 with the Bespoke Stratos 35k Dataset.**
27
+ - **Trained for 2 Hours on A800 with the Benford's Law Reasoning Small 430 Row Dataset, ensuring no overfitting.**
28
+ - **Trained for 4 Hours on A800 with the Sky-T1_data_17k Dataset**
29
+ - **Trained for 2 Hours on A800 with the Openthoughts 114k Dataset.**
30
+ - **15$ Training...I'm actually amazed by the results.**
31
 
 
32
 
33
+ If using this model for Open WebUI here is a simple function to organize the models responses: https://openwebui.com/f/quaz93/phi4_turn_r1_distill_thought_function_v1