Update README.md
Browse files
README.md
CHANGED
@@ -115,6 +115,7 @@ We used LoRA to speed up the training process, targeting the q_proj and v_proj m
|
|
115 |
|
116 |
## Uses
|
117 |
|
|
|
118 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
119 |
We provide the prompts used for both the model's fine-tuning and inference.
|
120 |
You can structure your data according to these prompts.
|
|
|
115 |
|
116 |
## Uses
|
117 |
|
118 |
+
For improved practical inference speed, we strongly recommend running Jellyfish using [vLLM](https://github.com/vllm-project/vllm).
|
119 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
120 |
We provide the prompts used for both the model's fine-tuning and inference.
|
121 |
You can structure your data according to these prompts.
|