ShuklaShreyansh commited on
Commit
53ba45b
verified
1 Parent(s): 40013b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +290 -3
README.md CHANGED
@@ -1,3 +1,290 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - unsloth
5
+ - Agriculture
6
+ - QA
7
+ - LLM
8
+ datasets:
9
+ - KisanVaani/agriculture-qa-english-only
10
+ language:
11
+ - en
12
+ base_model:
13
+ - unsloth/Llama-3.2-3B-Instruct
14
+ new_version: ShuklaShreyansh/Agro-QA
15
+ pipeline_tag: question-answering
16
+ library_name: transformers
17
+ ---
18
+
19
+
20
+
21
+ # Model Card for Agro-QA
22
+
23
+ This model is fine-tuned for agricultural question-answering tasks. It leverages the Llama-3.2-3B-Instruct model to address a variety of topics in agriculture, such as crop selection, pest management, irrigation, and farming best practices.
24
+
25
+ ## Model Details
26
+
27
+ ### Model Description
28
+
29
+ - **Developed by:** Shukla Shreyansh
30
+ - **Model type:** Question Answering (QA)
31
+ - **Language(s) (NLP):** English
32
+ - **License:** Apache-2.0
33
+ - **Finetuned from model:** [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct)
34
+
35
+ ---
36
+
37
+ ## Uses
38
+
39
+ ### Direct Use
40
+
41
+ The model is intended for question-answering applications specific to agriculture. It provides insights into farming techniques, crop choices, pest management, and related topics.
42
+
43
+ ### Out-of-Scope Use
44
+
45
+ The model is not designed for non-agriculture-related questions or tasks requiring specialized domain knowledge outside of agriculture.
46
+
47
+ ---
48
+
49
+ ## Training Details
50
+
51
+ ### Training Data
52
+
53
+ The model is fine-tuned on the [KisanVaani/agriculture-qa-english-only](https://huggingface.co/datasets/KisanVaani/agriculture-qa-english-only) dataset, a curated collection of questions and answers focused on agricultural topics.
54
+
55
+ ### Training Procedure
56
+
57
+ - **Training regime:** Mixed precision (FP16)
58
+ - **Batch size:** 2 (per device)
59
+ - **Epochs:** 1
60
+ - **Learning rate:** 2e-4
61
+ - **Optimizer:** AdamW with 8-bit precision
62
+
63
+ ---
64
+
65
+ ## Evaluation
66
+
67
+ ### Testing Data
68
+
69
+ The model is evaluated on a subset of the training dataset to measure its performance in answering agriculture-related questions.
70
+
71
+ ### Metrics
72
+
73
+ - **Accuracy:** [More Information Needed]
74
+ - **F1 Score:** [More Information Needed]
75
+
76
+ ---
77
+
78
+ ## How to Get Started with the Model
79
+
80
+ Use the code below to load and use the model:
81
+
82
+ ```python
83
+ from transformers import AutoTokenizer, AutoModelForCausalLM
84
+
85
+ # Load tokenizer
86
+ tokenizer = AutoTokenizer.from_pretrained("ShuklaShreyansh/Agro-QA")
87
+
88
+ # Load model
89
+ model = AutoModelForCausalLM.from_pretrained("ShuklaShreyansh/Agro-QA").to("cuda")
90
+
91
+ # Example usage
92
+ messages = [{"role": "user", "content": "What are the best rabi crops to grow?"}]
93
+ inputs = tokenizer.apply_chat_template(messages, tokenize=True, return_tensors="pt").to("cuda")
94
+ output = model.generate(input_ids=inputs['input_ids'], max_new_tokens=128)
95
+ print(tokenizer.decode(output[0]))
96
+ ```
97
+ # Model Card for Model ID
98
+
99
+ <!-- Provide a quick summary of what the model is/does. -->
100
+
101
+ This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
102
+
103
+ ## Model Details
104
+
105
+ ### Model Description
106
+
107
+ <!-- Provide a longer summary of what this model is. -->
108
+
109
+
110
+
111
+ - **Developed by:** [More Information Needed]
112
+ - **Funded by [optional]:** [More Information Needed]
113
+ - **Shared by [optional]:** [More Information Needed]
114
+ - **Model type:** [More Information Needed]
115
+ - **Language(s) (NLP):** [More Information Needed]
116
+ - **License:** [More Information Needed]
117
+ - **Finetuned from model [optional]:** [More Information Needed]
118
+
119
+ ### Model Sources [optional]
120
+
121
+ <!-- Provide the basic links for the model. -->
122
+
123
+ - **Repository:** [More Information Needed]
124
+ - **Paper [optional]:** [More Information Needed]
125
+ - **Demo [optional]:** [More Information Needed]
126
+
127
+ ## Uses
128
+
129
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
130
+
131
+ ### Direct Use
132
+
133
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
134
+
135
+ [More Information Needed]
136
+
137
+ ### Downstream Use [optional]
138
+
139
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
140
+
141
+ [More Information Needed]
142
+
143
+ ### Out-of-Scope Use
144
+
145
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
146
+
147
+ [More Information Needed]
148
+
149
+ ## Bias, Risks, and Limitations
150
+
151
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
152
+
153
+ [More Information Needed]
154
+
155
+ ### Recommendations
156
+
157
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
158
+
159
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
160
+
161
+ ## How to Get Started with the Model
162
+
163
+ Use the code below to get started with the model.
164
+
165
+ [More Information Needed]
166
+
167
+ ## Training Details
168
+
169
+ ### Training Data
170
+
171
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
172
+
173
+ [More Information Needed]
174
+
175
+ ### Training Procedure
176
+
177
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
178
+
179
+ #### Preprocessing [optional]
180
+
181
+ [More Information Needed]
182
+
183
+
184
+ #### Training Hyperparameters
185
+
186
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
187
+
188
+ #### Speeds, Sizes, Times [optional]
189
+
190
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## Evaluation
195
+
196
+ <!-- This section describes the evaluation protocols and provides the results. -->
197
+
198
+ ### Testing Data, Factors & Metrics
199
+
200
+ #### Testing Data
201
+
202
+ <!-- This should link to a Dataset Card if possible. -->
203
+
204
+ [More Information Needed]
205
+
206
+ #### Factors
207
+
208
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
209
+
210
+ [More Information Needed]
211
+
212
+ #### Metrics
213
+
214
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
215
+
216
+ [More Information Needed]
217
+
218
+ ### Results
219
+
220
+ [More Information Needed]
221
+
222
+ #### Summary
223
+
224
+
225
+
226
+ ## Model Examination [optional]
227
+
228
+ <!-- Relevant interpretability work for the model goes here -->
229
+
230
+ [More Information Needed]
231
+
232
+ ## Environmental Impact
233
+
234
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
235
+
236
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
237
+
238
+ - **Hardware Type:** [More Information Needed]
239
+ - **Hours used:** [More Information Needed]
240
+ - **Cloud Provider:** [More Information Needed]
241
+ - **Compute Region:** [More Information Needed]
242
+ - **Carbon Emitted:** [More Information Needed]
243
+
244
+ ## Technical Specifications [optional]
245
+
246
+ ### Model Architecture and Objective
247
+
248
+ [More Information Needed]
249
+
250
+ ### Compute Infrastructure
251
+
252
+ [More Information Needed]
253
+
254
+ #### Hardware
255
+
256
+ [More Information Needed]
257
+
258
+ #### Software
259
+
260
+ [More Information Needed]
261
+
262
+ ## Citation [optional]
263
+
264
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
265
+
266
+ **BibTeX:**
267
+
268
+ [More Information Needed]
269
+
270
+ **APA:**
271
+
272
+ [More Information Needed]
273
+
274
+ ## Glossary [optional]
275
+
276
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
277
+
278
+ [More Information Needed]
279
+
280
+ ## More Information [optional]
281
+
282
+ [More Information Needed]
283
+
284
+ ## Model Card Authors [optional]
285
+
286
+ [More Information Needed]
287
+
288
+ ## Model Card Contact
289
+
290
+ [More Information Needed]