mahyar-najibi
commited on
Commit
•
0033cff
1
Parent(s):
47562d5
Update README.
Browse files
README.md
CHANGED
@@ -46,17 +46,17 @@ We have provided an example function to generate output from OpenELM models load
|
|
46 |
|
47 |
You can try the model by running the following command:
|
48 |
```
|
49 |
-
python generate_openelm.py --model [
|
50 |
```
|
51 |
Please refer to [this link](https://huggingface.co/docs/hub/security-tokens) to obtain your hugging face access token.
|
52 |
|
53 |
Additional arguments to the hugging face generate function can be passed via `generate_kwargs`. As an example, to speedup the inference, you can try [lookup token speculative generation](https://huggingface.co/docs/transformers/generation_strategies) by passing the `prompt_lookup_num_tokens` argument as follows:
|
54 |
```
|
55 |
-
python generate_openelm.py --model [
|
56 |
```
|
57 |
Alternatively, model-wise speculative generation with an [assistive model](https://huggingface.co/blog/assisted-generation) can be also tried by passing a smaller model model through the `assistant_model` argument, for example:
|
58 |
```
|
59 |
-
python generate_openelm.py --model [
|
60 |
```
|
61 |
|
62 |
|
|
|
46 |
|
47 |
You can try the model by running the following command:
|
48 |
```
|
49 |
+
python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2
|
50 |
```
|
51 |
Please refer to [this link](https://huggingface.co/docs/hub/security-tokens) to obtain your hugging face access token.
|
52 |
|
53 |
Additional arguments to the hugging face generate function can be passed via `generate_kwargs`. As an example, to speedup the inference, you can try [lookup token speculative generation](https://huggingface.co/docs/transformers/generation_strategies) by passing the `prompt_lookup_num_tokens` argument as follows:
|
54 |
```
|
55 |
+
python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 prompt_lookup_num_tokens=10
|
56 |
```
|
57 |
Alternatively, model-wise speculative generation with an [assistive model](https://huggingface.co/blog/assisted-generation) can be also tried by passing a smaller model model through the `assistant_model` argument, for example:
|
58 |
```
|
59 |
+
python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 --assistant_model [SMALLER_MODEL_NAME]
|
60 |
```
|
61 |
|
62 |
|