seonghyeonye
commited on
Commit
•
9b42e78
1
Parent(s):
369669b
Update README.md
Browse files
README.md
CHANGED
@@ -16,14 +16,14 @@ Our overall explanation models along with ablations can be found in our [paper](
|
|
16 |
Here is how to use the model in PyTorch:
|
17 |
```python
|
18 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
19 |
-
tokenizer = AutoTokenizer.from_pretrained("seonghyeonye/
|
20 |
-
model = AutoModelForSeq2SeqLM.from_pretrained("seonghyeonye/
|
21 |
inputs = tokenizer.encode("input: this is the best cast iron skillet you will ever buy\noutput: Positive", return_tensors="pt")
|
22 |
outputs = model.generate(inputs)
|
23 |
print(tokenizer.decode(outputs[0]))
|
24 |
```
|
25 |
If you want to use another checkpoint, please replace the path in `AutoTokenizer` and `AutoModelForSeq2SeqLM`.
|
26 |
-
**Note: the model was trained with
|
27 |
|
28 |
# Training procedure
|
29 |
FLIPPED models are based on [T5](https://huggingface.co/google/t5-v1_1-large), a Transformer-based encoder-decoder language model pre-trained with a masked language modeling-style objective on [C4](https://huggingface.co/datasets/c4).
|
|
|
16 |
Here is how to use the model in PyTorch:
|
17 |
```python
|
18 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
19 |
+
tokenizer = AutoTokenizer.from_pretrained("seonghyeonye/flipped_3B")
|
20 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("seonghyeonye/flipped_3B")
|
21 |
inputs = tokenizer.encode("input: this is the best cast iron skillet you will ever buy\noutput: Positive", return_tensors="pt")
|
22 |
outputs = model.generate(inputs)
|
23 |
print(tokenizer.decode(outputs[0]))
|
24 |
```
|
25 |
If you want to use another checkpoint, please replace the path in `AutoTokenizer` and `AutoModelForSeq2SeqLM`.
|
26 |
+
**Note: the model was trained with fp32 activations. As such, we highly discourage running inference with fp16.**
|
27 |
|
28 |
# Training procedure
|
29 |
FLIPPED models are based on [T5](https://huggingface.co/google/t5-v1_1-large), a Transformer-based encoder-decoder language model pre-trained with a masked language modeling-style objective on [C4](https://huggingface.co/datasets/c4).
|