Update README.md
Browse files
README.md
CHANGED
@@ -138,6 +138,8 @@ For improved practical inference speed, we strongly recommend running Jellyfish
|
|
138 |
We provide two simple Python code examples for inference using the Jellyfish model.
|
139 |
|
140 |
#### Using Transformers and Torch Modules
|
|
|
|
|
141 |
```python
|
142 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
143 |
import torch
|
@@ -192,6 +194,7 @@ response = tokenizer.decode(
|
|
192 |
print(response)
|
193 |
|
194 |
```
|
|
|
195 |
|
196 |
#### Using vLLM
|
197 |
```python
|
|
|
138 |
We provide two simple Python code examples for inference using the Jellyfish model.
|
139 |
|
140 |
#### Using Transformers and Torch Modules
|
141 |
+
<div style="height: auto; max-height: 400px; overflow-y: scroll;">
|
142 |
+
|
143 |
```python
|
144 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
145 |
import torch
|
|
|
194 |
print(response)
|
195 |
|
196 |
```
|
197 |
+
</div>
|
198 |
|
199 |
#### Using vLLM
|
200 |
```python
|