reissbaker's picture
Add README
97e01d4
---
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
library_name: peft
---
# Model Card for Model ID
This LoRA adapter was extracted from
[mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated](https://huggingface.co./mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated)
and uses
[meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co./meta-llama/Llama-3.1-70B-Instruct)
as a base.
## Model Details
The model was extracted by running
[mlabonne/harmful_behaviors](https://huggingface.co./datasets/mlabonne/harmful_behaviors)
and the user prompts (but not assistant responses or system messages) from
[Guilherme34/uncensor](https://huggingface.co./datasets/Guilherme34/uncensor)
through the original abliterated model to generate a dataset of
prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl
using FSDP. Since the original abliterated model isn't perfect at avoiding
refusals, the dataset was cleaned to remove the few refusals generated prior to
training.
### Model Description
- **Developed by:** @reissbaker
- **Funded by:** Synthetic Lab
- **License:** Apache 2.0
- **Finetuned from model:** Llama 3.1 70B Instruct
## How to Get Started with the Model
Run the model with one click on [glhf.chat](https://glhf.chat).
#### Training Hyperparameters
* BF16 mixed-precision
* 4e-4 LR
* Linear LR schedule
* Fused AdamW optimizer