|
--- |
|
license: other |
|
--- |
|
|
|
# WizardLM: An Instruction-following LLM Using Evol-Instruct |
|
|
|
These files are the result of merging the [delta weights](https://huggingface.co/victor123/WizardLM) with the original Llama7B model. |
|
|
|
The code for merging is provided in the [WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). |
|
|
|
## WizardLM-7B HF |
|
|
|
This repo contains the full unquantised model files in HF format for GPU inference and as a base for quantisation/conversion. |
|
|
|
## Other repositories available |
|
|
|
* [4bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GGML) |
|
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizardLM-7B-GPTQ) |
|
|
|
# Original model info |
|
|
|
## Full details in the model's Github page |
|
|
|
[WizardLM official Github repo](https://github.com/nlpxucan/WizardLM). |
|
|
|
## Overview of Evol-Instruct |
|
|
|
Evol-Instruct is a novel method using LLMs instead of humans to automatically mass-produce open-domain instructions of various difficulty levels and skills range, to improve the performance of LLMs. |
|
|
|
Although on our complexity-balanced test set, WizardLM-7B outperforms ChatGPT in the high-complexity instructions, it still lag behind ChatGPT on the entire test set, and we also consider WizardLM to still be in a baby state. This repository will continue to improve WizardLM, train on larger scales, add more training data, and innovate more advanced large-model training methods. |
|
|
|
![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_overall.png) |
|
![info](https://github.com/nlpxucan/WizardLM/raw/main/imgs/git_running.png) |