yatharth97
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -1,199 +1,86 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
tags: []
|
4 |
---
|
5 |
|
6 |
-
# Model Card for
|
7 |
-
|
8 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
9 |
-
|
10 |
|
|
|
11 |
|
12 |
## Model Details
|
13 |
|
14 |
### Model Description
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
19 |
-
|
20 |
-
- **Developed by:** [More Information Needed]
|
21 |
-
- **Funded by [optional]:** [More Information Needed]
|
22 |
-
- **Shared by [optional]:** [More Information Needed]
|
23 |
-
- **Model type:** [More Information Needed]
|
24 |
-
- **Language(s) (NLP):** [More Information Needed]
|
25 |
-
- **License:** [More Information Needed]
|
26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
27 |
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
- **Repository:** [
|
33 |
-
- **Paper [optional]:** [More Information Needed]
|
34 |
-
- **Demo [optional]:** [More Information Needed]
|
35 |
|
36 |
## Uses
|
37 |
|
38 |
-
|
39 |
-
|
40 |
-
### Direct Use
|
41 |
|
42 |
-
|
43 |
|
44 |
-
|
45 |
|
46 |
-
|
47 |
|
48 |
-
|
49 |
|
50 |
-
|
51 |
|
52 |
### Out-of-Scope Use
|
53 |
|
54 |
-
|
55 |
-
|
56 |
-
[More Information Needed]
|
57 |
|
58 |
## Bias, Risks, and Limitations
|
59 |
|
60 |
-
|
61 |
-
|
62 |
-
[More Information Needed]
|
63 |
|
64 |
### Recommendations
|
65 |
|
66 |
-
|
67 |
-
|
68 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
69 |
|
70 |
## How to Get Started with the Model
|
71 |
|
72 |
-
|
73 |
-
|
74 |
-
[More Information Needed]
|
75 |
|
76 |
## Training Details
|
77 |
|
78 |
### Training Data
|
79 |
|
80 |
-
|
81 |
-
|
82 |
-
[More Information Needed]
|
83 |
|
84 |
### Training Procedure
|
85 |
|
86 |
-
|
87 |
-
|
88 |
-
#### Preprocessing [optional]
|
89 |
-
|
90 |
-
[More Information Needed]
|
91 |
|
|
|
92 |
|
93 |
#### Training Hyperparameters
|
94 |
|
95 |
-
|
96 |
-
|
97 |
-
#### Speeds, Sizes, Times [optional]
|
98 |
-
|
99 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
100 |
-
|
101 |
-
[More Information Needed]
|
102 |
|
103 |
## Evaluation
|
104 |
|
105 |
-
|
106 |
-
|
107 |
-
### Testing Data, Factors & Metrics
|
108 |
-
|
109 |
-
#### Testing Data
|
110 |
-
|
111 |
-
<!-- This should link to a Dataset Card if possible. -->
|
112 |
-
|
113 |
-
[More Information Needed]
|
114 |
-
|
115 |
-
#### Factors
|
116 |
-
|
117 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
118 |
-
|
119 |
-
[More Information Needed]
|
120 |
-
|
121 |
-
#### Metrics
|
122 |
-
|
123 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
124 |
-
|
125 |
-
[More Information Needed]
|
126 |
|
127 |
-
|
128 |
|
129 |
-
|
130 |
|
131 |
-
|
132 |
|
|
|
133 |
|
|
|
134 |
|
135 |
-
|
136 |
-
|
137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
138 |
-
|
139 |
-
[More Information Needed]
|
140 |
-
|
141 |
-
## Environmental Impact
|
142 |
-
|
143 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
144 |
-
|
145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
146 |
-
|
147 |
-
- **Hardware Type:** [More Information Needed]
|
148 |
-
- **Hours used:** [More Information Needed]
|
149 |
-
- **Cloud Provider:** [More Information Needed]
|
150 |
-
- **Compute Region:** [More Information Needed]
|
151 |
-
- **Carbon Emitted:** [More Information Needed]
|
152 |
-
|
153 |
-
## Technical Specifications [optional]
|
154 |
-
|
155 |
-
### Model Architecture and Objective
|
156 |
-
|
157 |
-
[More Information Needed]
|
158 |
|
159 |
### Compute Infrastructure
|
160 |
|
161 |
-
|
162 |
-
|
163 |
-
#### Hardware
|
164 |
-
|
165 |
-
[More Information Needed]
|
166 |
-
|
167 |
-
#### Software
|
168 |
-
|
169 |
-
[More Information Needed]
|
170 |
-
|
171 |
-
## Citation [optional]
|
172 |
-
|
173 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
174 |
-
|
175 |
-
**BibTeX:**
|
176 |
-
|
177 |
-
[More Information Needed]
|
178 |
-
|
179 |
-
**APA:**
|
180 |
-
|
181 |
-
[More Information Needed]
|
182 |
-
|
183 |
-
## Glossary [optional]
|
184 |
-
|
185 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
186 |
-
|
187 |
-
[More Information Needed]
|
188 |
-
|
189 |
-
## More Information [optional]
|
190 |
-
|
191 |
-
[More Information Needed]
|
192 |
-
|
193 |
-
## Model Card Authors [optional]
|
194 |
-
|
195 |
-
[More Information Needed]
|
196 |
-
|
197 |
-
## Model Card Contact
|
198 |
-
|
199 |
-
[More Information Needed]
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
tags: [chatbot, task-oriented, multi-turn-qa, English, fine-tuned, meta-llama2-7b, financial-reports]
|
4 |
---
|
5 |
|
6 |
+
# Model Card for Meta LLAMA2-7B Custom Task-Oriented Chatbot
|
|
|
|
|
|
|
7 |
|
8 |
+
This model is a fine-tuned version of Meta's LLAMA2-7B model, adapted to function as a task-oriented chatbot that processes and answers questions related to financial 10K reports.
|
9 |
|
10 |
## Model Details
|
11 |
|
12 |
### Model Description
|
13 |
|
14 |
+
Developed by Yatharth Mahesh Sant, this model is a causal language model fine-tuned from Meta's LLAMA2-7B to specifically handle queries and tasks associated with 10K financial reports. It is designed to assist financial analysts and stakeholders by providing detailed, accurate answers to inquiries about company performances, financial standings, and other key metrics contained within 10K reports.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
+
- **Developed by:** Yatharth Mahesh Sant
|
17 |
+
- **Model type:** Causal LM
|
18 |
+
- **Language(s) (NLP):** English
|
19 |
+
- **Finetuned from:** meta/llama2-7b
|
20 |
+
- **Repository:** [Meta LLAMA2-7B](https://huggingface.co/meta-llama/Llama-2-7b)
|
|
|
|
|
21 |
|
22 |
## Uses
|
23 |
|
24 |
+
### Intended Use
|
|
|
|
|
25 |
|
26 |
+
This model is meant to be used as a task-oriented bot to interact with users querying about details in 10K financial reports, enhancing the efficiency of financial analysis and decision-making processes.
|
27 |
|
28 |
+
### Direct Use
|
29 |
|
30 |
+
The model can directly answer questions from financial reports, serving as an automated assistant to financial analysts, investors, and regulatory authorities who require quick, reliable interpretations of financial data.
|
31 |
|
32 |
+
### Downstream Use
|
33 |
|
34 |
+
The model can be integrated into financial analysis software, used to power internal data review tools in corporations, or serve as a support system in investor relations departments to automate responses to common shareholder inquiries.
|
35 |
|
36 |
### Out-of-Scope Use
|
37 |
|
38 |
+
This model is not designed for non-financial texts or languages other than English. It may not perform well in informal conversational settings or handle off-topic inquiries effectively.
|
|
|
|
|
39 |
|
40 |
## Bias, Risks, and Limitations
|
41 |
|
42 |
+
The model's performance and responses are based on the data it was trained on, which primarily includes structured financial texts. As such, it may inherit biases from this data or fail to comprehend nuanced questions not directly related to financial reporting.
|
|
|
|
|
43 |
|
44 |
### Recommendations
|
45 |
|
46 |
+
It is recommended that responses generated by this model be reviewed by a qualified financial analyst to confirm their accuracy before being used in critical decision-making processes.
|
|
|
|
|
47 |
|
48 |
## How to Get Started with the Model
|
49 |
|
50 |
+
To get started with this model, please refer to the specific deployment guides and API documentation provided in the repository linked above.
|
|
|
|
|
51 |
|
52 |
## Training Details
|
53 |
|
54 |
### Training Data
|
55 |
|
56 |
+
The model was fine-tuned on a comprehensive dataset comprising several years' worth of 10K reports from companies across various industries, annotated for key financial metrics and queries.
|
|
|
|
|
57 |
|
58 |
### Training Procedure
|
59 |
|
60 |
+
#### Preprocessing
|
|
|
|
|
|
|
|
|
61 |
|
62 |
+
Training data was preprocessed to normalize financial terminology and remove any non-relevant sections of the reports, focusing on the sections most pertinent to common queries.
|
63 |
|
64 |
#### Training Hyperparameters
|
65 |
|
66 |
+
The model was trained using a learning rate of 5e-5 with a batch size of 32 over 4 epochs, employing a transformer-based architecture optimized for natural language understanding tasks.
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
|
68 |
## Evaluation
|
69 |
|
70 |
+
### Testing Data
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
71 |
|
72 |
+
The model was evaluated on a separate validation set consisting of annotated 10K reports not seen during training to ensure it can generalize across different texts and query types.
|
73 |
|
74 |
+
### Metrics
|
75 |
|
76 |
+
Evaluation metrics included accuracy, F1 score, and a custom metric for response relevance to financial queries.
|
77 |
|
78 |
+
## Technical Specifications
|
79 |
|
80 |
+
### Model Architecture
|
81 |
|
82 |
+
The model employs a transformer-based architecture, leveraging attention mechanisms to focus on relevant parts of the text when generating responses.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
83 |
|
84 |
### Compute Infrastructure
|
85 |
|
86 |
+
Training was conducted on cloud-based GPUs with support for high-throughput training sessions.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|