Update README.md
Browse files
README.md
CHANGED
@@ -52,4 +52,54 @@ generated_ids = [
|
|
52 |
]
|
53 |
|
54 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
55 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
]
|
53 |
|
54 |
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
55 |
+
```
|
56 |
+
|
57 |
+
# **Intended Use:**
|
58 |
+
|
59 |
+
1. **Multilingual Dialogue Systems:**
|
60 |
+
- Designed for conversational AI applications, capable of handling dialogue across multiple languages.
|
61 |
+
- Useful in customer service, chatbots, and other dialogue-centric use cases.
|
62 |
+
|
63 |
+
2. **Reasoning and QWQ Dataset Applications:**
|
64 |
+
- Optimized for tasks requiring logical reasoning and contextual understanding, particularly in synthetic datasets like QWQ.
|
65 |
+
|
66 |
+
3. **Agentic Retrieval:**
|
67 |
+
- Supports retrieval-augmented generation tasks, helping systems fetch and synthesize information effectively.
|
68 |
+
|
69 |
+
4. **Summarization Tasks:**
|
70 |
+
- Excels in summarizing long or complex text while maintaining coherence and relevance.
|
71 |
+
|
72 |
+
5. **Instruction-Following Tasks:**
|
73 |
+
- Can execute tasks based on specific user instructions due to instruction-tuning during training.
|
74 |
+
|
75 |
+
6. **Language Generation:**
|
76 |
+
- Suitable for generating coherent and contextually relevant text in various domains and styles.
|
77 |
+
|
78 |
+
# **Limitations:**
|
79 |
+
|
80 |
+
1. **Synthetic Dataset Bias:**
|
81 |
+
- Optimization for QWQ and similar datasets may make the model less effective on real-world or less structured data.
|
82 |
+
|
83 |
+
2. **Data Dependency:**
|
84 |
+
- Performance may degrade on tasks or languages not well-represented in the training dataset.
|
85 |
+
|
86 |
+
3. **Computational Requirements:**
|
87 |
+
- The optimized transformer architecture may demand significant computational resources, especially for fine-tuning or large-scale deployments.
|
88 |
+
|
89 |
+
4. **Potential Hallucinations:**
|
90 |
+
- Like most auto-regressive models, it may generate plausible-sounding but factually incorrect or nonsensical outputs.
|
91 |
+
|
92 |
+
5. **RLHF-Specific Biases:**
|
93 |
+
- Reinforcement Learning with Human Feedback (RLHF) can introduce biases based on the preferences of the annotators involved in the feedback process.
|
94 |
+
|
95 |
+
6. **Limited Domain Adaptability:**
|
96 |
+
- While effective in reasoning and dialogue tasks, it may struggle with highly specialized domains or out-of-distribution tasks.
|
97 |
+
|
98 |
+
7. **Multilingual Limitations:**
|
99 |
+
- Although optimized for multilingual use, certain low-resource languages may exhibit poorer performance compared to high-resource ones.
|
100 |
+
|
101 |
+
8. **Ethical Concerns:**
|
102 |
+
- May inadvertently generate inappropriate or harmful content if safeguards are not applied, particularly in sensitive applications.
|
103 |
+
|
104 |
+
9. **Real-Time Usability:**
|
105 |
+
- Latency in inference time could limit its effectiveness in real-time applications or when scaling to large user bases.
|