colesmcintosh's picture
Update README.md
c631825 verified
---
library_name: transformers
tags:
- trl
- sft
license: mit
datasets:
- gbharti/finance-alpaca
base_model:
- mistralai/Mistral-7B-v0.1
language:
- en
---
### Model Description
This model is based on the Mistral 7B architecture, PEFT fine-tuned on financial data. It is designed to handle various finance-related NLP tasks such as financial text analysis, sentiment detection, market trend analysis, and more. This model leverages the powerful transformer architecture of Mistral with specialized fine-tuning for financial applications.
- **Developed by:** Cole McIntosh
- **Model type:** Transformer-based large language model (LLM)
- **Language(s) (NLP):** English
- **Finetuned from model:** Mistral 7B
## Uses
### Direct Use
The Mistral 7B Finance Fine-tuned model is designed to assist users with finance-related natural language processing tasks such as:
- Financial report analysis
- Sentiment analysis of financial news
- Forecasting market trends based on textual data
- Analyzing earnings call transcripts
- Extracting structured information from unstructured financial text
### Downstream Use
This model can be fine-tuned further for more specific tasks such as:
- Portfolio analysis based on sentiment scores
- Predictive analysis for stock market movements
- Automated financial report generation
### Out-of-Scope Use
This model should not be used for tasks unrelated to finance or those requiring a high level of factual accuracy in non-financial domains. It is not suitable for:
- Medical or legal document analysis
- General conversational chatbots (as the model may provide misleading financial interpretations)
- Decision-making without human oversight, especially in high-stakes financial operations
### Recommendations
- Carefully review model outputs, especially in critical financial decisions.
- Use up-to-date fine-tuning datasets to ensure relevance.
- Cross-validate the model's predictions or insights with alternative data sources or human expertise.
## How to Get Started with the Model
You can use the Hugging Face `transformers` library to load and use this model. Here’s a basic example:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("colesmcintosh/mistral_7b_finance_finetuned")
model = AutoModelForCausalLM.from_pretrained("colesmcintosh/mistral_7b_finance_finetuned")
# Example usage
inputs = tokenizer("Analyze the financial outlook for Q3 2024.", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0]))
```
## Training Details
### Training Procedure
The fine-tuning process used PEFT to accelerate training on GPUs.
#### Summary
The model performs well in finance-specific tasks like sentiment analysis and entity recognition. It demonstrates strong generalization across different sectors but shows slight performance drops when analyzing non-English financial texts.
### Model Architecture and Objective
The model is based on the Mistral 7B architecture, a highly optimized transformer-based model. Its primary objective is text generation and understanding, with a focus on financial texts.
### Compute Infrastructure
#### Hardware
The model was fine-tuned using:
- 1 NVIDIA A100 GPU (40 GB)
#### Software
- Hugging Face `transformers` library
- PEFT finetuning