Update README.md
Browse files
README.md
CHANGED
@@ -9,11 +9,11 @@ language:
|
|
9 |
<img src="https://i.imgur.com/d8Bl04i.png" alt="PicToModel" width="330"/>
|
10 |
|
11 |
## Model Details
|
12 |
-
Jellyfish-13B is a large language model with 13 billion parameters
|
13 |
|
14 |
-
We fine-tuned [Open-Orca/OpenOrca-Platypus2-13B](https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B) using the datasets
|
15 |
-
Its performance is competitive,
|
16 |
-
|
17 |
|
18 |
| Task | Dataset | Non-LLM SoTA<sup>1</sup> | GPT-3.5<sup>2</sup> | GPT-4<sup>2</sup> | Jellyfish-13B| Jellyfish-13B-Resoning |
|
19 |
| ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|
|
|
9 |
<img src="https://i.imgur.com/d8Bl04i.png" alt="PicToModel" width="330"/>
|
10 |
|
11 |
## Model Details
|
12 |
+
Jellyfish-13B is a large language model equipped with 13 billion parameters. It's tailored specifically for data preprocessing tasks, including entity matching, data imputation, error detection, and schema matching.
|
13 |
|
14 |
+
We fine-tuned [Open-Orca/OpenOrca-Platypus2-13B](https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B) using the datasets pertinent to data preprocessing tasks.
|
15 |
+
Its performance is competitive, rivaling previous state-of-the-art algorithms and LLMs such as OpenAI's GPT 3.5 and GPT 4, ([as demonstrated in our earlier studies](https://arxiv.org/abs/2308.16361))
|
16 |
+
It is notable that as a 13B model, Jellyfish allows for cost-effective local execution without compromising data security.
|
17 |
|
18 |
| Task | Dataset | Non-LLM SoTA<sup>1</sup> | GPT-3.5<sup>2</sup> | GPT-4<sup>2</sup> | Jellyfish-13B| Jellyfish-13B-Resoning |
|
19 |
| ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|