mallorbc's picture
Update README.md
19c2b66 verified
metadata
license: llama2
datasets:
  - databricks/databricks-dolly-15k
language:
  - en
pipeline_tag: text-generation

Instruct_Llama70B_Dolly15k

Fine-tuned from Llama-2-70B,used Dolly15k for the dataset. 80% for training, 15% validation, 5% test. Trained for 1.5 epochs using QLora. Trained with 1024 context window.

Model Details

  • Trained by: trained by Brillibits. See YouTube as well.
  • Model type: Instruct_Llama70B_Dolly15k is an auto-regressive language model based on the Llama 2 transformer architecture.
  • Language(s): English
  • License for Instruct_Llama70B_Dolly15ks: llama2 license

Prompting

Prompt Template With Context

Write a 10-line poem about a given topic

Input:

The topic is about racecars

Output:

Prompt Template Without Context

Who was the was the second president of the United States?

Output:

Professional Assistance

This model and other models like it are great, but where LLMs hold the most promise is when they are applied on custom data to automate a wide variety of tasks

If you have a dataset and want to see if you might be able to apply that data to automate some tasks, and you are looking for professional assistance, contact me here

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 60.97
ARC (25-shot) 68.34
HellaSwag (10-shot) 87.21
MMLU (5-shot) 69.52
TruthfulQA (0-shot) 46.46
Winogrande (5-shot) 84.29
GSM8K (5-shot) 42.68
DROP (3-shot) 28.26