|
--- |
|
license: other |
|
--- |
|
<div style="width: 100%;"> |
|
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> |
|
</div> |
|
<div style="display: flex; justify-content: space-between; width: 100%;"> |
|
<div style="display: flex; flex-direction: column; align-items: flex-start;"> |
|
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p> |
|
</div> |
|
<div style="display: flex; flex-direction: column; align-items: flex-end;"> |
|
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p> |
|
</div> |
|
</div> |
|
|
|
# WizardLM: An Instruction-following LLM Using Evol-Instruct |
|
|
|
These files are the result of merging the [delta weights](https://huggingface.co./victor123/WizardLM) with the original Llama7B model. |
|
|
|
The code for merging is provided in the [WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). |
|
|
|
The original WizardLM deltas are in float32, and this results in producing an HF repo that is also float32, and is much larger than a normal 7B Llama model. |
|
|
|
Therefore for this repo I converted the merged model to float16, to produce a standard size 7B model. |
|
|
|
This was achieved by running **`model = model.half()`** prior to saving. |
|
|
|
## WizardLM-7B HF |
|
|
|
This repo contains the full unquantised model files in HF format for GPU inference and as a base for quantisation/conversion. |
|
|
|
## Other repositories available |
|
|
|
* [4bit GGML models for CPU inference](https://huggingface.co./TheBloke/wizardLM-7B-GGML) |
|
* [4bit GPTQ models for GPU inference](https://huggingface.co./TheBloke/wizardLM-7B-GPTQ) |
|
|
|
## Want to support my work? |
|
|
|
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills. |
|
|
|
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects. |
|
|
|
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try. |
|
|
|
* Patreon: coming soon! (just awaiting approval) |
|
* Ko-Fi: https://ko-fi.com/TheBlokeAI |
|
* Discord: https://discord.gg/UBgz4VXf |
|
# Original model info |
|
|
|
## Full details in the model's Github page |
|
|
|
[WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). |
|
|
|
## Overview of Evol-Instruct |
|
|
|
Evol-Instruct is a novel method using LLMs instead of humans to automatically mass-produce open-domain instructions of various difficulty levels and skills range, to improve the performance of LLMs. |
|
|
|
Although on our complexity-balanced test set, WizardLM-7B outperforms ChatGPT in the high-complexity instructions, it still lag behind ChatGPT on the entire test set, and we also consider WizardLM to still be in a baby state. This repository will continue to improve WizardLM, train on larger scales, add more training data, and innovate more advanced large-model training methods. |
|
|
|
![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_overall.png) |
|
![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_running.png) |
|
|