--- license: other ---
TheBlokeAI

Chat & support: my new Discord server

Want to contribute? Patreon coming soon!

# WizardLM: An Instruction-following LLM Using Evol-Instruct These files are the result of merging the [delta weights](https://huggingface.co./victor123/WizardLM) with the original Llama7B model. The code for merging is provided in the [WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). The original WizardLM deltas are in float32, and this results in producing an HF repo that is also float32, and is much larger than a normal 7B Llama model. Therefore for this repo I converted the merged model to float16, to produce a standard size 7B model. This was achieved by running **`model = model.half()`** prior to saving. ## WizardLM-7B HF This repo contains the full unquantised model files in HF format for GPU inference and as a base for quantisation/conversion. ## Other repositories available * [4bit GGML models for CPU inference](https://huggingface.co./TheBloke/wizardLM-7B-GGML) * [4bit GPTQ models for GPU inference](https://huggingface.co./TheBloke/wizardLM-7B-GPTQ) ## Want to support my work? I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills. So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects. Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try. * Patreon: coming soon! (just awaiting approval) * Ko-Fi: https://ko-fi.com/TheBlokeAI * Discord: https://discord.gg/UBgz4VXf # Original model info ## Full details in the model's Github page [WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). ## Overview of Evol-Instruct Evol-Instruct is a novel method using LLMs instead of humans to automatically mass-produce open-domain instructions of various difficulty levels and skills range, to improve the performance of LLMs. Although on our complexity-balanced test set, WizardLM-7B outperforms ChatGPT in the high-complexity instructions, it still lag behind ChatGPT on the entire test set, and we also consider WizardLM to still be in a baby state. This repository will continue to improve WizardLM, train on larger scales, add more training data, and innovate more advanced large-model training methods. ![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_overall.png) ![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_running.png)