Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ Since the release of [Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai
|
|
19 |
on a wide range of benchmarks. But most of the benchmarks are evaluated on `short context`, and not much has been investigated on its performance on long context tasks.
|
20 |
Then We evaluated `Mistral-7B-Instruct-v0.1` against benchmarks that are specifically designed to assess the capabilities of LLMs in handling longer context.
|
21 |
Although the performance of the models on long context was fairly competitive on long context less than 4096 tokens,
|
22 |
-
there were some limitations on its performance on longer context. Motivated by improving its performance on longer context, we finetuned the Mistral 7B model, and
|
23 |
|
24 |
### [Topic Retrieval](https://lmsys.org/blog/2023-06-29-longchat/) ###
|
25 |
|Model Name|Input length| Input length | Input length| Input length| Input length|
|
|
|
19 |
on a wide range of benchmarks. But most of the benchmarks are evaluated on `short context`, and not much has been investigated on its performance on long context tasks.
|
20 |
Then We evaluated `Mistral-7B-Instruct-v0.1` against benchmarks that are specifically designed to assess the capabilities of LLMs in handling longer context.
|
21 |
Although the performance of the models on long context was fairly competitive on long context less than 4096 tokens,
|
22 |
+
there were some limitations on its performance on longer context. Motivated by improving its performance on longer context, we finetuned the Mistral 7B model, and produced `Mistrallite`. The model managed to `signifantly boost the performance of long context handling` over Mistral-7B-Instruct-v0.1. The detailed `long context evalutaion results` are as below:
|
23 |
|
24 |
### [Topic Retrieval](https://lmsys.org/blog/2023-06-29-longchat/) ###
|
25 |
|Model Name|Input length| Input length | Input length| Input length| Input length|
|