HCZhang commited on
Commit
4d96901
1 Parent(s): babbb41

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -138,6 +138,8 @@ For improved practical inference speed, we strongly recommend running Jellyfish
138
  We provide two simple Python code examples for inference using the Jellyfish model.
139
 
140
  #### Using Transformers and Torch Modules
 
 
141
  ```python
142
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
143
  import torch
@@ -192,6 +194,7 @@ response = tokenizer.decode(
192
  print(response)
193
 
194
  ```
 
195
 
196
  #### Using vLLM
197
  ```python
 
138
  We provide two simple Python code examples for inference using the Jellyfish model.
139
 
140
  #### Using Transformers and Torch Modules
141
+ <div style="height: auto; max-height: 400px; overflow-y: scroll;">
142
+
143
  ```python
144
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
145
  import torch
 
194
  print(response)
195
 
196
  ```
197
+ </div>
198
 
199
  #### Using vLLM
200
  ```python