Pankaj Mathur
commited on
Commit
•
7511d16
1
Parent(s):
b02da34
Update README.md
Browse files
README.md
CHANGED
@@ -5,16 +5,16 @@ language:
|
|
5 |
library_name: adapter-transformers
|
6 |
---
|
7 |
# Wizardlm Alpaca Dolly Orca Open_LLaMa_13b
|
8 |
-
An Open_LLaMA-13B model trained on custom explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying
|
9 |
|
10 |
|
11 |
# Dataset
|
12 |
|
13 |
-
We trained [OpenLLaMa-
|
14 |
|
15 |
-
We leverage all of the 15 system instructions provided in
|
16 |
|
17 |
-
This helps student model aka [
|
18 |
|
19 |
Please see below example usage how the **System** prompt is added before each *instruction*.
|
20 |
|
@@ -22,7 +22,7 @@ Please see below example usage how the **System** prompt is added before each *i
|
|
22 |
|
23 |
The training configurations are provided in the table below.
|
24 |
|
25 |
-
The training takes on
|
26 |
|
27 |
We used DeepSpeed with Zero-3 approaches for parallel gpu training by writing our own fine tunning scripts plus leveraging some of the model training code provided by amazing [OpenAlpaca repo](https://github.com/yxuansu/OpenAlpaca)
|
28 |
|
@@ -32,7 +32,7 @@ Here are some of params used during training:
|
|
32 |
|:-------------:|:-------------:|
|
33 |
|*batch_size*|16|
|
34 |
|*train_micro_batch_size_per_gpu*|2|
|
35 |
-
|*gradient_accumulation_steps*|
|
36 |
|*Learning rate*|2e-5|
|
37 |
|*Max length*|1024|
|
38 |
|*Epochs*|3|
|
@@ -41,14 +41,14 @@ Here are some of params used during training:
|
|
41 |
|
42 |
# Example Usage
|
43 |
|
44 |
-
Below shows an example on how to use
|
45 |
|
46 |
```python
|
47 |
import torch
|
48 |
from transformers import LlamaForCausalLM, LlamaTokenizer
|
49 |
|
50 |
-
#
|
51 |
-
model_path = 'psmathur/
|
52 |
tokenizer = LlamaTokenizer.from_pretrained(model_path)
|
53 |
model = LlamaForCausalLM.from_pretrained(
|
54 |
model_path, torch_dtype=torch.float16, device_map='auto',
|
@@ -94,24 +94,23 @@ generate_text(system, instruction, input)
|
|
94 |
**P.S. I am #opentowork and #collaboration, if you can help, please reach out to me at [email protected]**
|
95 |
|
96 |
Next Goals:
|
97 |
-
1) Try more data
|
98 |
-
2) Try
|
99 |
-
3)
|
100 |
-
4) Provide
|
101 |
-
6) Provide 4bit GGML/GPTQ quantized model (may be [TheBloke](https://huggingface.co/TheBloke) can help here)
|
102 |
|
103 |
|
104 |
Reference:
|
105 |
-
If you found
|
106 |
|
107 |
```
|
108 |
-
@misc{
|
109 |
author = {Pankaj Mathur},
|
110 |
-
title = {
|
111 |
year = {2023},
|
112 |
publisher = {GitHub, HuggingFace},
|
113 |
journal = {GitHub repository, HuggingFace repository},
|
114 |
-
howpublished = {\url{https://github.com/pankajarm/
|
115 |
}
|
116 |
```
|
117 |
```
|
|
|
5 |
library_name: adapter-transformers
|
6 |
---
|
7 |
# Wizardlm Alpaca Dolly Orca Open_LLaMa_13b
|
8 |
+
An Open_LLaMA-13B model trained on custom explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.
|
9 |
|
10 |
|
11 |
# Dataset
|
12 |
|
13 |
+
We trained [OpenLLaMa-13B model](https://github.com/openlm-research/open_llama) on custom explain tuned [WizardLM ~70K](https://github.com/nlpxucan/WizardLM), [Alpaca dataset ~52K](https://crfm.stanford.edu/2023/03/13/alpaca.html) & [Dolly-V2 ~15K](https://github.com/databrickslabs/dolly) created using approaches from [Orca Research Paper](https://arxiv.org/abs/2306.02707).
|
14 |
|
15 |
+
We leverage all of the 15 system instructions provided in Orca Research Paper. to generate custom datasets, in contrast to vanilla instruction tuning approaches used by original datasets.
|
16 |
|
17 |
+
This helps student model aka [wizardlm_alpaca_dolly_orca_open_llama_13b](https://huggingface.co/psmathur/wizardlm_alpaca_dolly_orca_open_llama_13b) to learn ***thought*** process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).
|
18 |
|
19 |
Please see below example usage how the **System** prompt is added before each *instruction*.
|
20 |
|
|
|
22 |
|
23 |
The training configurations are provided in the table below.
|
24 |
|
25 |
+
The training takes on 8x A100(80G) GPUs and lasts for around 15 Hours for cost of $180 using [Lambda Labs](https://lambdalabs.com)
|
26 |
|
27 |
We used DeepSpeed with Zero-3 approaches for parallel gpu training by writing our own fine tunning scripts plus leveraging some of the model training code provided by amazing [OpenAlpaca repo](https://github.com/yxuansu/OpenAlpaca)
|
28 |
|
|
|
32 |
|:-------------:|:-------------:|
|
33 |
|*batch_size*|16|
|
34 |
|*train_micro_batch_size_per_gpu*|2|
|
35 |
+
|*gradient_accumulation_steps*|1|
|
36 |
|*Learning rate*|2e-5|
|
37 |
|*Max length*|1024|
|
38 |
|*Epochs*|3|
|
|
|
41 |
|
42 |
# Example Usage
|
43 |
|
44 |
+
Below shows an example on how to use this model
|
45 |
|
46 |
```python
|
47 |
import torch
|
48 |
from transformers import LlamaForCausalLM, LlamaTokenizer
|
49 |
|
50 |
+
# Hugging Face model_path
|
51 |
+
model_path = 'psmathur/wizardlm_alpaca_dolly_orca_open_llama_13b'
|
52 |
tokenizer = LlamaTokenizer.from_pretrained(model_path)
|
53 |
model = LlamaForCausalLM.from_pretrained(
|
54 |
model_path, torch_dtype=torch.float16, device_map='auto',
|
|
|
94 |
**P.S. I am #opentowork and #collaboration, if you can help, please reach out to me at [email protected]**
|
95 |
|
96 |
Next Goals:
|
97 |
+
1) Try more data like actually using FLAN-v2, just like Orka Research Paper (I am open for suggestions)
|
98 |
+
2) Try smaller OpenLLaMA models 7B and 3B
|
99 |
+
3) Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui)
|
100 |
+
4) Provide 4bit GGML/GPTQ quantized model (may be [TheBloke](https://huggingface.co/TheBloke) can help here)
|
|
|
101 |
|
102 |
|
103 |
Reference:
|
104 |
+
If you found wizardlm_alpaca_dolly_orca_open_llama_13b useful in your research or applications, please kindly cite using the following BibTeX:
|
105 |
|
106 |
```
|
107 |
+
@misc{wizardlm_alpaca_dolly_orca_open_llama_13b,
|
108 |
author = {Pankaj Mathur},
|
109 |
+
title = {wizardlm_alpaca_dolly_orca_open_llama_13b: An explain tuned OpenLLaMA-13b model on custom wizardlm, alpaca, & dolly datasets},
|
110 |
year = {2023},
|
111 |
publisher = {GitHub, HuggingFace},
|
112 |
journal = {GitHub repository, HuggingFace repository},
|
113 |
+
howpublished = {\url{https://github.com/pankajarm/wizardlm_alpaca_dolly_orca_open_llama_13b}, \url{https://https://huggingface.co/psmathur/wizardlm_alpaca_dolly_orca_open_llama_13b}},
|
114 |
}
|
115 |
```
|
116 |
```
|