seonghyeonye commited on
Commit
9b42e78
1 Parent(s): 369669b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -16,14 +16,14 @@ Our overall explanation models along with ablations can be found in our [paper](
16
  Here is how to use the model in PyTorch:
17
  ```python
18
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
19
- tokenizer = AutoTokenizer.from_pretrained("seonghyeonye/flipped_11B")
20
- model = AutoModelForSeq2SeqLM.from_pretrained("seonghyeonye/flipped_11B")
21
  inputs = tokenizer.encode("input: this is the best cast iron skillet you will ever buy\noutput: Positive", return_tensors="pt")
22
  outputs = model.generate(inputs)
23
  print(tokenizer.decode(outputs[0]))
24
  ```
25
  If you want to use another checkpoint, please replace the path in `AutoTokenizer` and `AutoModelForSeq2SeqLM`.
26
- **Note: the model was trained with bf16 activations. As such, we highly discourage running inference with fp16. fp32 or bf16 should be preferred.**
27
 
28
  # Training procedure
29
  FLIPPED models are based on [T5](https://huggingface.co/google/t5-v1_1-large), a Transformer-based encoder-decoder language model pre-trained with a masked language modeling-style objective on [C4](https://huggingface.co/datasets/c4).
 
16
  Here is how to use the model in PyTorch:
17
  ```python
18
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
19
+ tokenizer = AutoTokenizer.from_pretrained("seonghyeonye/flipped_3B")
20
+ model = AutoModelForSeq2SeqLM.from_pretrained("seonghyeonye/flipped_3B")
21
  inputs = tokenizer.encode("input: this is the best cast iron skillet you will ever buy\noutput: Positive", return_tensors="pt")
22
  outputs = model.generate(inputs)
23
  print(tokenizer.decode(outputs[0]))
24
  ```
25
  If you want to use another checkpoint, please replace the path in `AutoTokenizer` and `AutoModelForSeq2SeqLM`.
26
+ **Note: the model was trained with fp32 activations. As such, we highly discourage running inference with fp16.**
27
 
28
  # Training procedure
29
  FLIPPED models are based on [T5](https://huggingface.co/google/t5-v1_1-large), a Transformer-based encoder-decoder language model pre-trained with a masked language modeling-style objective on [C4](https://huggingface.co/datasets/c4).