Update README.md
#2
by
mojanjp
- opened
README.md
CHANGED
@@ -16,6 +16,16 @@ Our model hasn't been fine-tuned through reinforcement learning from human feedb
|
|
16 |
|
17 |
## Intended Uses
|
18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
Phi-2 is intended for research purposes only. Given the nature of the training data, the phi-2 model is best suited for prompts using the QA format, the chat format, and the code format.
|
20 |
|
21 |
#### QA format:
|
|
|
16 |
|
17 |
## Intended Uses
|
18 |
|
19 |
+
Below are example codes to load phi-2, we support two modes of execution for the model:
|
20 |
+
1. loading in fp-16 format with flash-attention support:
|
21 |
+
```python
|
22 |
+
model = AutoModelForCausalLM.from_pretrained('microsoft/phi-2', torch_dtype='auto', flash_attn=True, flash_rotary=True, fused_dense=True, trust_remote_code=True)
|
23 |
+
```
|
24 |
+
2. loading in fp-16 without flash-attention
|
25 |
+
```python
|
26 |
+
model = AutoModelForCausalLM.from_pretrained('microsoft/phi-2', torch_dtype='auto', trust_remote_code=True)
|
27 |
+
```
|
28 |
+
|
29 |
Phi-2 is intended for research purposes only. Given the nature of the training data, the phi-2 model is best suited for prompts using the QA format, the chat format, and the code format.
|
30 |
|
31 |
#### QA format:
|