Triangle104 commited on
Commit
53280bc
·
verified ·
1 Parent(s): 3247d75

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +179 -0
README.md CHANGED
@@ -16,6 +16,185 @@ tags:
16
  This model was converted to GGUF format from [`prithivMLmods/Deepthink-Llama-3-8B-Preview`](https://huggingface.co/prithivMLmods/Deepthink-Llama-3-8B-Preview) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/prithivMLmods/Deepthink-Llama-3-8B-Preview) for more details on the model.
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ## Use with llama.cpp
20
  Install llama.cpp through brew (works on Mac and Linux)
21
 
 
16
  This model was converted to GGUF format from [`prithivMLmods/Deepthink-Llama-3-8B-Preview`](https://huggingface.co/prithivMLmods/Deepthink-Llama-3-8B-Preview) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/prithivMLmods/Deepthink-Llama-3-8B-Preview) for more details on the model.
18
 
19
+ ---
20
+ The Deepthink-Llama-3-8B-Preview is a fine-tuned version of the Llama-3.1-8B base model, further enhanced with the Rethinking R1 Dataset Logits
21
+ for superior text generation. This model is designed for advanced
22
+ reasoning, structured problem-solving, and contextually rich outputs,
23
+ making it an excellent choice for applications in education, programming, research, and creative writing.
24
+
25
+
26
+ With its optimized architecture, Deepthink-Llama-3-8B-Preview excels at:
27
+
28
+
29
+ Logical reasoning and step-by-step problem solving
30
+ Mathematical and coding tasks, leveraging specialized expert models
31
+ Generating long-form content (up to 8K tokens) with improved coherence
32
+ Understanding structured data, including tables and JSON outputs
33
+ Instruction following and adapting to diverse system prompts, making it ideal for chatbots and AI assistants
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+ Key Features
42
+
43
+
44
+
45
+
46
+ Supports long-context processing of up to 128K tokens
47
+ Multilingual capabilities for 29+ languages, including English, Chinese, Spanish, French, German, Arabic, and more
48
+ Fine-tuned using Supervised Fine-Tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF)
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+ Model Architecture
57
+
58
+
59
+
60
+
61
+ Deepthink-Llama-3-8B-Preview is built on the optimized transformer architecture of Llama-3.1-8B, integrating enhanced dataset logits from Rethinking R1 for better contextual understanding and output quality.
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+ Use with transformers
70
+
71
+
72
+
73
+
74
+ To run conversational inference using transformers >= 4.43.0, use the pipeline abstraction or leverage the generate() function with the Auto classes.
75
+
76
+
77
+ Ensure your environment is updated with:
78
+
79
+
80
+ pip install --upgrade transformers
81
+
82
+
83
+
84
+
85
+
86
+
87
+
88
+
89
+ Example Usage
90
+
91
+
92
+
93
+
94
+ import torch
95
+ from transformers import pipeline
96
+
97
+ model_id = "prithivMLmods/Deepthink-Llama-3-8B-Preview"
98
+ pipe = pipeline(
99
+ "text-generation",
100
+ model=model_id,
101
+ torch_dtype=torch.bfloat16,
102
+ device_map="auto",
103
+ )
104
+
105
+ messages = [
106
+ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
107
+ {"role": "user", "content": "Who are you?"},
108
+ ]
109
+
110
+ outputs = pipe(
111
+ messages,
112
+ max_new_tokens=256,
113
+ )
114
+ print(outputs[0]["generated_text"][-1])
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+ Intended Use
124
+
125
+
126
+
127
+
128
+ Deepthink-Llama-3-8B-Preview is designed for a wide
129
+ range of applications requiring deep reasoning, structured outputs, and
130
+ logical text generation. It is particularly suited for:
131
+
132
+
133
+ Education & Research: Generating detailed explanations, step-by-step solutions, and structured academic content.
134
+ Programming & Code Generation: Assisting in code writing, debugging, and algorithm explanations with improved logic structuring.
135
+ AI Chatbots & Assistants: Providing context-aware, instruction-following responses for conversational AI applications.
136
+ Creative Writing: Generating high-quality stories, articles, and structured narratives with coherence.
137
+ Data Analysis & Structured Output Generation: Interpreting and generating JSON, tables, and formatted outputs for structured data processing.
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+ Limitations
146
+
147
+
148
+
149
+
150
+ While Deepthink-Llama-3-8B-Preview is optimized for deep reasoning and structured outputs, it has some limitations:
151
+
152
+
153
+ Not a Real-time Knowledge Source
154
+
155
+
156
+ The model is trained on a fixed dataset and does not have real-time
157
+ internet access. It may not provide up-to-date information on rapidly
158
+ evolving topics.
159
+
160
+
161
+ Potential Biases
162
+
163
+
164
+ As with all AI models, responses may reflect biases present in the
165
+ training data. Users should critically evaluate outputs, especially in
166
+ sensitive domains.
167
+
168
+
169
+ Mathematical & Logical Reasoning Constraints
170
+
171
+
172
+ While strong in step-by-step reasoning, it may occasionally produce
173
+ incorrect mathematical calculations or logical inconsistencies. External
174
+ verification is recommended for critical applications.
175
+
176
+
177
+ Handling of Extremely Long Contexts
178
+
179
+
180
+ While it supports up to 128K tokens, efficiency and coherence may degrade when processing very long documents or conversations.
181
+
182
+
183
+ Limited Handling of Ambiguity
184
+
185
+
186
+ The model may struggle with highly ambiguous or context-dependent
187
+ queries, sometimes generating plausible but incorrect responses.
188
+
189
+
190
+ Ethical & Compliance Considerations
191
+
192
+
193
+ Not intended for generating misinformation, automating legal or
194
+ medical decisions, or other high-risk applications without human
195
+ oversight.
196
+
197
+ ---
198
  ## Use with llama.cpp
199
  Install llama.cpp through brew (works on Mac and Linux)
200