Update README.md
Browse files
README.md
CHANGED
@@ -4,7 +4,7 @@ license: mit
|
|
4 |
This model is a 7B [Self-RAG](https://selfrag.github.io/) model that generates outputs to diverse user queries as well as *reflection tokens* to call the retrieval system adaptively and criticize its own output and retrieved passages.
|
5 |
|
6 |
Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
|
7 |
-
At inference, we leverage reflection tokens covering diverse
|
8 |
See full descriptions in [our paper](https://akariasai.github.io/files/adaptive_retrieval_augmented_lm_arxiv.pdf).
|
9 |
|
10 |
## Usage
|
@@ -53,12 +53,19 @@ or
|
|
53 |
If you have additional input.
|
54 |
You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
|
55 |
|
|
|
|
|
|
|
|
|
|
|
56 |
## Citation and contact
|
57 |
If you use this model, please cite our work:
|
58 |
```
|
59 |
@article{asai2023selfrag,
|
60 |
author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
|
61 |
-
title = {{Self-RAG}:
|
62 |
year = {2023},
|
|
|
|
|
63 |
}
|
64 |
```
|
|
|
4 |
This model is a 7B [Self-RAG](https://selfrag.github.io/) model that generates outputs to diverse user queries as well as *reflection tokens* to call the retrieval system adaptively and criticize its own output and retrieved passages.
|
5 |
|
6 |
Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
|
7 |
+
At inference, we leverage reflection tokens covering diverse aspects of generations to sample the best output aligning users' preferences.
|
8 |
See full descriptions in [our paper](https://akariasai.github.io/files/adaptive_retrieval_augmented_lm_arxiv.pdf).
|
9 |
|
10 |
## Usage
|
|
|
53 |
If you have additional input.
|
54 |
You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
|
55 |
|
56 |
+
## Training details
|
57 |
+
Our training data is available at the HuggingFace dataset [selfrag_train_data](https://huggingface.co/datasets/selfrag/selfrag_train_data).
|
58 |
+
See our official repository for the training details.
|
59 |
+
We used 8 A100 40GB for training on the Stability HPC server.
|
60 |
+
|
61 |
## Citation and contact
|
62 |
If you use this model, please cite our work:
|
63 |
```
|
64 |
@article{asai2023selfrag,
|
65 |
author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
|
66 |
+
title = {{Self-RAG}: Learning to Retrieve, Generate, and Critique through Self-Reflection},
|
67 |
year = {2023},
|
68 |
+
journal = { arXiv preprint arXiv:2310.11511 },
|
69 |
+
URL = {https://arxiv.org/abs/2310.11511}
|
70 |
}
|
71 |
```
|