File size: 1,973 Bytes
7bf9726 3fcb2a4 7d58553 5295924 7d58553 ba54e9f a31b16e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
---
license: llama2
datasets:
- databricks/databricks-dolly-15k
language:
- en
pipeline_tag: text-generation
---
# Instruct_Llama70B_Dolly15k
Fine-tuned from Llama-2-70B,used Dolly15k for the dataset. 80% for training, 15% validation, 5% test. Trained for 1.5 epochs using QLora. Trained with 1024 context window.
# Model Details
* **Trained by**: trained by [Brillibits](https://www.youtube.com/channel/UCAq9THVHhPK0Zv4Xi-88Jmg).
* **Model type:** **Instruct_Llama70B_Dolly15k** is an auto-regressive language model based on the Llama 2 transformer architecture.
* **Language(s)**: English
* **License for Instruct_Llama70B_Dolly15ks**: llama2 license
# Prompting
## Prompt Template With Context
```
Write a 10-line poem about a given topic
Input:
The topic is about racecars
Output:
```
## Prompt Template Without Context
```
Who was the was the second president of the United States?
Output:
```
## Professional Assistance
This model and other models like it are great, but where LLMs hold the most promise is when they are applied on custom data to automate a wide variety of tasks
If you have a dataset and want to see if you might be able to apply that data to automate some tasks, and you are looking for professional assistance, contact me [here](mailto:[email protected])
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Brillibits__Instruct_Llama70B_Dolly15k)
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | 60.97 |
| ARC (25-shot) | 68.34 |
| HellaSwag (10-shot) | 87.21 |
| MMLU (5-shot) | 69.52 |
| TruthfulQA (0-shot) | 46.46 |
| Winogrande (5-shot) | 84.29 |
| GSM8K (5-shot) | 42.68 |
| DROP (3-shot) | 28.26 |
|