Spaces:
Sleeping
Sleeping
FantasticGNU
commited on
Commit
•
3157e75
1
Parent(s):
79ec1bc
Update model/openllama.py
Browse files- model/openllama.py +11 -10
model/openllama.py
CHANGED
@@ -205,16 +205,17 @@ class OpenLLAMAPEFTModel(nn.Module):
|
|
205 |
target_modules=['q_proj', 'k_proj', 'v_proj', 'o_proj']
|
206 |
)
|
207 |
|
208 |
-
|
209 |
-
|
210 |
-
|
211 |
-
|
212 |
-
|
213 |
-
|
214 |
-
|
215 |
-
|
216 |
-
|
217 |
-
|
|
|
218 |
self.llama_model = get_peft_model(self.llama_model, peft_config)
|
219 |
self.llama_model.print_trainable_parameters()
|
220 |
|
|
|
205 |
target_modules=['q_proj', 'k_proj', 'v_proj', 'o_proj']
|
206 |
)
|
207 |
|
208 |
+
config = AutoConfig.from_pretrained(vicuna_ckpt_path)
|
209 |
+
with init_empty_weights():
|
210 |
+
self.llama_model = AutoModelForCausalLM.from_config(config)
|
211 |
+
self.llama_model = load_checkpoint_and_dispatch(self.llama_model, vicuna_ckpt_path, device_map="auto", no_split_module_classes=["OPTDecoderLayer"], offload_folder="offload", offload_state_dict = True)
|
212 |
+
self.llama_model.to(torch.float16)
|
213 |
+
# try:
|
214 |
+
# self.llama_model = AutoModelForCausalLM.from_pretrained(vicuna_ckpt_path, torch_dtype=torch.float16, device_map='auto', offload_folder="offload", offload_state_dict = True)
|
215 |
+
# except:
|
216 |
+
# pass
|
217 |
+
# finally:
|
218 |
+
# print(self.llama_model.hf_device_map)
|
219 |
self.llama_model = get_peft_model(self.llama_model, peft_config)
|
220 |
self.llama_model.print_trainable_parameters()
|
221 |
|