namespace-Pt
commited on
Commit
•
ce11373
1
Parent(s):
b710d82
Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -35,7 +35,7 @@ We evaluate the model on [LongBench](https://arxiv.org/abs/2308.14508) using 32K
|
|
35 |
We evaluate the model on [InfiniteBench](https://arxiv.org/pdf/2402.13718.pdf) using 80K context length and the official prompt template. The results of GPT-4 is copied from the [paper](https://arxiv.org/pdf/2402.13718.pdf). For [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct), we use 8K context length.
|
36 |
|
37 |
|Model|LongBookQA Eng|LongBookSum Eng|
|
38 |
-
|
39 |
|GPT-4|22.22|14.73|
|
40 |
|[meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)|7.00|**16.40**|
|
41 |
|[gradientai/Llama-3-8B-Instruct-262k](https://huggingface.co/NousResearch/Yarn-Mistral-7b-128k)|20.30|10.34|
|
|
|
35 |
We evaluate the model on [InfiniteBench](https://arxiv.org/pdf/2402.13718.pdf) using 80K context length and the official prompt template. The results of GPT-4 is copied from the [paper](https://arxiv.org/pdf/2402.13718.pdf). For [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct), we use 8K context length.
|
36 |
|
37 |
|Model|LongBookQA Eng|LongBookSum Eng|
|
38 |
+
|:-:|:-:|:-:|
|
39 |
|GPT-4|22.22|14.73|
|
40 |
|[meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)|7.00|**16.40**|
|
41 |
|[gradientai/Llama-3-8B-Instruct-262k](https://huggingface.co/NousResearch/Yarn-Mistral-7b-128k)|20.30|10.34|
|