Added training config
Browse files
README.md
CHANGED
@@ -61,7 +61,8 @@ print(tokenizer.decode(generation_output[0]))
|
|
61 |
|
62 |
### Training Data
|
63 |
|
64 |
-
Converted the following datasets to alpaca:instruction format
|
|
|
65 |
1. [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin)
|
66 |
- ORCA style dataset generously created by [Eric Hartford](https://huggingface.co/ehartford)
|
67 |
- Only used the 1 million GPT4 generated instructions file [flan1m-alpaca-uncensored.jsonl](https://huggingface.co/datasets/ehartford/dolphin/blob/main/flan1m-alpaca-uncensored.jsonl).
|
@@ -76,9 +77,90 @@ Converted the following datasets to alpaca:instruction format:
|
|
76 |
|
77 |
### Training Procedure
|
78 |
|
79 |
-
Trained using axolotl QLoRa on RunPod 8x A6000 on Community Cloud for
|
80 |
|
81 |
-
|
|
|
|
|
82 |
```yaml
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
83 |
|
84 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
61 |
|
62 |
### Training Data
|
63 |
|
64 |
+
Converted the following datasets to alpaca:instruction format.
|
65 |
+
|
66 |
1. [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin)
|
67 |
- ORCA style dataset generously created by [Eric Hartford](https://huggingface.co/ehartford)
|
68 |
- Only used the 1 million GPT4 generated instructions file [flan1m-alpaca-uncensored.jsonl](https://huggingface.co/datasets/ehartford/dolphin/blob/main/flan1m-alpaca-uncensored.jsonl).
|
|
|
77 |
|
78 |
### Training Procedure
|
79 |
|
80 |
+
Trained using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) QLoRa on [RunPod](https://www.runpod.io/console/gpu-cloud) 8x A6000 on Community Cloud for 3 epochs (~14 hours - ~$70).
|
81 |
|
82 |
+
<details>
|
83 |
+
<summary>axolotl training config:</summary>
|
84 |
+
|
85 |
```yaml
|
86 |
+
base_model: openlm-research/open_llama_7b_v2
|
87 |
+
base_model_config: openlm-research/open_llama_7b_v2
|
88 |
+
model_type: LlamaForCausalLM
|
89 |
+
tokenizer_type: LlamaTokenizer
|
90 |
+
load_in_8bit: false
|
91 |
+
load_in_4bit: true
|
92 |
+
strict: false
|
93 |
+
|
94 |
+
push_dataset_to_hub:
|
95 |
+
hub_model_id:
|
96 |
+
hf_use_auth_token:
|
97 |
|
98 |
+
datasets:
|
99 |
+
- path: json
|
100 |
+
type: alpaca
|
101 |
+
data_files: /disk/flan1m-alpaca-uncensored.jsonl
|
102 |
+
shards: 8
|
103 |
+
- path: sahil2801/code_instructions_120k
|
104 |
+
type: alpaca
|
105 |
+
- path: LinhDuong/chatdoctor-200k
|
106 |
+
type: alpaca
|
107 |
+
shards: 2
|
108 |
+
- path: kaiokendev/SuperCOT-dataset
|
109 |
+
type: alpaca
|
110 |
+
- path: medalpaca/medical_meadow_mediqa
|
111 |
+
type: alpaca
|
112 |
+
|
113 |
+
dataset_prepared_path: last_run_prepared
|
114 |
+
val_set_size: 0.01
|
115 |
+
adapter: qlora
|
116 |
+
lora_model_dir:
|
117 |
+
sequence_len: 2048
|
118 |
+
max_packed_sequence_len: 2048
|
119 |
+
lora_r: 8
|
120 |
+
lora_alpha: 32
|
121 |
+
lora_dropout: 0.05
|
122 |
+
lora_target_modules:
|
123 |
+
lora_target_linear: true
|
124 |
+
lora_fan_in_fan_out:
|
125 |
+
|
126 |
+
wandb_mode: true
|
127 |
+
wandb_project:
|
128 |
+
wandb_watch:
|
129 |
+
wandb_run_id:
|
130 |
+
wandb_log_model: 'openllama_checkpoint'
|
131 |
+
output_dir: /disk/open_llama_7b_v2_dolphin_qlora
|
132 |
+
gradient_accumulation_steps: 2
|
133 |
+
micro_batch_size: 16
|
134 |
+
num_epochs: 3
|
135 |
+
optimizer: paged_adamw_32bit
|
136 |
+
torchdistx_path:
|
137 |
+
lr_scheduler: cosine
|
138 |
+
learning_rate: 0.0002
|
139 |
+
train_on_inputs: false
|
140 |
+
group_by_length: false
|
141 |
+
bf16: true
|
142 |
+
fp16: false
|
143 |
+
tf32: true
|
144 |
+
gradient_checkpointing: true
|
145 |
+
early_stopping_patience:
|
146 |
+
resume_from_checkpoint:
|
147 |
+
local_rank:
|
148 |
+
logging_steps: 1
|
149 |
+
xformers_attention: true
|
150 |
+
flash_attention:
|
151 |
+
gptq_groupsize:
|
152 |
+
gptq_model_v1:
|
153 |
+
warmup_steps: 1000
|
154 |
+
eval_steps: 5000
|
155 |
+
save_steps:
|
156 |
+
debug:
|
157 |
+
deepspeed:
|
158 |
+
weight_decay: 0.0000001
|
159 |
+
fsdp:
|
160 |
+
fsdp_config:
|
161 |
+
special_tokens:
|
162 |
+
bos_token: "<s>"
|
163 |
+
eos_token: "</s>"
|
164 |
+
unk_token: "<unk>"
|
165 |
+
```
|
166 |
+
</details>
|