yhyhy3 commited on
Commit
4028efa
1 Parent(s): 255a884

Added training config

Browse files
Files changed (1) hide show
  1. README.md +86 -4
README.md CHANGED
@@ -61,7 +61,8 @@ print(tokenizer.decode(generation_output[0]))
61
 
62
  ### Training Data
63
 
64
- Converted the following datasets to alpaca:instruction format:
 
65
  1. [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin)
66
  - ORCA style dataset generously created by [Eric Hartford](https://huggingface.co/ehartford)
67
  - Only used the 1 million GPT4 generated instructions file [flan1m-alpaca-uncensored.jsonl](https://huggingface.co/datasets/ehartford/dolphin/blob/main/flan1m-alpaca-uncensored.jsonl).
@@ -76,9 +77,90 @@ Converted the following datasets to alpaca:instruction format:
76
 
77
  ### Training Procedure
78
 
79
- Trained using axolotl QLoRa on RunPod 8x A6000 on Community Cloud for 2 epochs (~14 hours).
80
 
81
- axolotl training config:
 
 
82
  ```yaml
 
 
 
 
 
 
 
 
 
 
 
83
 
84
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
 
62
  ### Training Data
63
 
64
+ Converted the following datasets to alpaca:instruction format.
65
+
66
  1. [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin)
67
  - ORCA style dataset generously created by [Eric Hartford](https://huggingface.co/ehartford)
68
  - Only used the 1 million GPT4 generated instructions file [flan1m-alpaca-uncensored.jsonl](https://huggingface.co/datasets/ehartford/dolphin/blob/main/flan1m-alpaca-uncensored.jsonl).
 
77
 
78
  ### Training Procedure
79
 
80
+ Trained using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) QLoRa on [RunPod](https://www.runpod.io/console/gpu-cloud) 8x A6000 on Community Cloud for 3 epochs (~14 hours - ~$70).
81
 
82
+ <details>
83
+ <summary>axolotl training config:</summary>
84
+
85
  ```yaml
86
+ base_model: openlm-research/open_llama_7b_v2
87
+ base_model_config: openlm-research/open_llama_7b_v2
88
+ model_type: LlamaForCausalLM
89
+ tokenizer_type: LlamaTokenizer
90
+ load_in_8bit: false
91
+ load_in_4bit: true
92
+ strict: false
93
+
94
+ push_dataset_to_hub:
95
+ hub_model_id:
96
+ hf_use_auth_token:
97
 
98
+ datasets:
99
+ - path: json
100
+ type: alpaca
101
+ data_files: /disk/flan1m-alpaca-uncensored.jsonl
102
+ shards: 8
103
+ - path: sahil2801/code_instructions_120k
104
+ type: alpaca
105
+ - path: LinhDuong/chatdoctor-200k
106
+ type: alpaca
107
+ shards: 2
108
+ - path: kaiokendev/SuperCOT-dataset
109
+ type: alpaca
110
+ - path: medalpaca/medical_meadow_mediqa
111
+ type: alpaca
112
+
113
+ dataset_prepared_path: last_run_prepared
114
+ val_set_size: 0.01
115
+ adapter: qlora
116
+ lora_model_dir:
117
+ sequence_len: 2048
118
+ max_packed_sequence_len: 2048
119
+ lora_r: 8
120
+ lora_alpha: 32
121
+ lora_dropout: 0.05
122
+ lora_target_modules:
123
+ lora_target_linear: true
124
+ lora_fan_in_fan_out:
125
+
126
+ wandb_mode: true
127
+ wandb_project:
128
+ wandb_watch:
129
+ wandb_run_id:
130
+ wandb_log_model: 'openllama_checkpoint'
131
+ output_dir: /disk/open_llama_7b_v2_dolphin_qlora
132
+ gradient_accumulation_steps: 2
133
+ micro_batch_size: 16
134
+ num_epochs: 3
135
+ optimizer: paged_adamw_32bit
136
+ torchdistx_path:
137
+ lr_scheduler: cosine
138
+ learning_rate: 0.0002
139
+ train_on_inputs: false
140
+ group_by_length: false
141
+ bf16: true
142
+ fp16: false
143
+ tf32: true
144
+ gradient_checkpointing: true
145
+ early_stopping_patience:
146
+ resume_from_checkpoint:
147
+ local_rank:
148
+ logging_steps: 1
149
+ xformers_attention: true
150
+ flash_attention:
151
+ gptq_groupsize:
152
+ gptq_model_v1:
153
+ warmup_steps: 1000
154
+ eval_steps: 5000
155
+ save_steps:
156
+ debug:
157
+ deepspeed:
158
+ weight_decay: 0.0000001
159
+ fsdp:
160
+ fsdp_config:
161
+ special_tokens:
162
+ bos_token: "<s>"
163
+ eos_token: "</s>"
164
+ unk_token: "<unk>"
165
+ ```
166
+ </details>