ApatheticWithoutTheA commited on
Commit
b7ce96e
·
verified ·
1 Parent(s): 396b43e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -3
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - sequelbox/Raiden-DeepSeek-R1
5
+ language:
6
+ - en
7
+ base_model:
8
+ - google/gemma-2-2b-it
9
+ pipeline_tag: text-generation
10
+ library_name: mlx
11
+ ---
12
+
13
+ ## Model Summary
14
+
15
+ This model is a fine-tuned version of **gemma-2-2b-it**, optimized for instruction-following and reasoning tasks. It was trained using **MLX** and **LoRA** on the **sequelbox/Raiden-DeepSeek-R1** dataset, which consists of **62.9k examples** generated by Deepseek R1. The fine-tuning process ran for **600 iterations** to enhance the model’s ability to generate coherent and contextually relevant responses.
16
+
17
+ ## Model Details
18
+
19
+ - **Base Model**: gemma-2-2b-it
20
+ - **Fine-tuning Method**: MLX + LoRA
21
+ - **Dataset**: [sequelbox/Raiden-DeepSeek-R1](https://huggingface.co/datasets/sequelbox/Raiden-DeepSeek-R1)
22
+ - **Iterations**: 600
23
+
24
+ ## Capabilities
25
+
26
+ This model improves upon **gemma-2-2b-it** with additional instruction-following and reasoning capabilities derived from Deepseek R1-generated examples. The model will answer questions with a straight-forward answer for simple questions, and generate long chain-of-thought reasoning text for more complex problems. It is well-suited for:
27
+
28
+ - Question answering
29
+ - Reasoning-based tasks
30
+ - Coding
31
+ - Running on consumer hardware
32
+
33
+ ## Limitations
34
+ - Sometimes chain-of-thought reasoning is not triggered for more complex problems when it probably should be. You can nudge the model if needed by simply asking it to show its thoughts and it will generate <think> tags and begin reasoning.
35
+ - With harder than average complex reasoning problems, the model can get stuck in long "thinking" thought loops without ever coming to a conclusive answer.