Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,15 @@ Anna is a 6B parameter autoregressive language model based on the GPT-J architec
|
|
17 |
Anna was trained on German text with a similar methodology to [Boris](https://huggingface.co/Cedille/fr-boris), our French model. We started training from GPT-J, which has been trained on [The Pile](https://pile.eleuther.ai/). As a consequence the model still has good performance in English language. Anna makes use of the unmodified GPT-2 tokenizer.
|
18 |
|
19 |
# How to run
|
20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
|
22 |
## Contact us
|
23 |
For any custom development please contact us at [email protected].
|
|
|
17 |
Anna was trained on German text with a similar methodology to [Boris](https://huggingface.co/Cedille/fr-boris), our French model. We started training from GPT-J, which has been trained on [The Pile](https://pile.eleuther.ai/). As a consequence the model still has good performance in English language. Anna makes use of the unmodified GPT-2 tokenizer.
|
18 |
|
19 |
# How to run
|
20 |
+
|
21 |
+
## Loading the model
|
22 |
+
```
|
23 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
24 |
+
|
25 |
+
tokenizer = AutoTokenizer.from_pretrained("Cedille/de-anna")
|
26 |
+
model = AutoModelForCausalLM.from_pretrained("Cedille/de-anna")
|
27 |
+
```
|
28 |
+
|
29 |
|
30 |
## Contact us
|
31 |
For any custom development please contact us at [email protected].
|