prithivMLmods commited on
Commit
113a9cc
·
verified ·
1 Parent(s): 4ec39ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -1
README.md CHANGED
@@ -52,4 +52,54 @@ generated_ids = [
52
  ]
53
 
54
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
55
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  ]
53
 
54
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
55
+ ```
56
+
57
+ # **Intended Use:**
58
+
59
+ 1. **Multilingual Dialogue Systems:**
60
+ - Designed for conversational AI applications, capable of handling dialogue across multiple languages.
61
+ - Useful in customer service, chatbots, and other dialogue-centric use cases.
62
+
63
+ 2. **Reasoning and QWQ Dataset Applications:**
64
+ - Optimized for tasks requiring logical reasoning and contextual understanding, particularly in synthetic datasets like QWQ.
65
+
66
+ 3. **Agentic Retrieval:**
67
+ - Supports retrieval-augmented generation tasks, helping systems fetch and synthesize information effectively.
68
+
69
+ 4. **Summarization Tasks:**
70
+ - Excels in summarizing long or complex text while maintaining coherence and relevance.
71
+
72
+ 5. **Instruction-Following Tasks:**
73
+ - Can execute tasks based on specific user instructions due to instruction-tuning during training.
74
+
75
+ 6. **Language Generation:**
76
+ - Suitable for generating coherent and contextually relevant text in various domains and styles.
77
+
78
+ # **Limitations:**
79
+
80
+ 1. **Synthetic Dataset Bias:**
81
+ - Optimization for QWQ and similar datasets may make the model less effective on real-world or less structured data.
82
+
83
+ 2. **Data Dependency:**
84
+ - Performance may degrade on tasks or languages not well-represented in the training dataset.
85
+
86
+ 3. **Computational Requirements:**
87
+ - The optimized transformer architecture may demand significant computational resources, especially for fine-tuning or large-scale deployments.
88
+
89
+ 4. **Potential Hallucinations:**
90
+ - Like most auto-regressive models, it may generate plausible-sounding but factually incorrect or nonsensical outputs.
91
+
92
+ 5. **RLHF-Specific Biases:**
93
+ - Reinforcement Learning with Human Feedback (RLHF) can introduce biases based on the preferences of the annotators involved in the feedback process.
94
+
95
+ 6. **Limited Domain Adaptability:**
96
+ - While effective in reasoning and dialogue tasks, it may struggle with highly specialized domains or out-of-distribution tasks.
97
+
98
+ 7. **Multilingual Limitations:**
99
+ - Although optimized for multilingual use, certain low-resource languages may exhibit poorer performance compared to high-resource ones.
100
+
101
+ 8. **Ethical Concerns:**
102
+ - May inadvertently generate inappropriate or harmful content if safeguards are not applied, particularly in sensitive applications.
103
+
104
+ 9. **Real-Time Usability:**
105
+ - Latency in inference time could limit its effectiveness in real-time applications or when scaling to large user bases.