File size: 3,853 Bytes
f6f7ce1 6e925aa b1f00b0 cee96fb d778524 cee96fb 45db21f cee96fb b1f00b0 cee96fb d778524 cee96fb d778524 797084e 6fc7ce0 c512bd2 6fc7ce0 c512bd2 be841c5 6fc7ce0 be841c5 b17924e 6fc7ce0 f27df4b 6fc7ce0 b17924e 85f418b b17924e bbba376 b5ed3b2 f27df4b bbba376 f27df4b 6fc7ce0 23a79dc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 |
---
license: mit
datasets:
- tatsu-lab/alpaca
language:
- en
metrics:
- accuracy
library_name: transformers
pipeline_tag: text-generation
widget:
- text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\nInstruction: What is artificial intelligence?\nResponse: "
example_title: "Knowledge-AI"
- text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\nInstruction: Write a haiku poem on cows\nResponse: "
example_title: "Poem Generation"
- text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\nInstruction: What is the meaning of life?\nResponse: "
example_title: "Philosophy"
- text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\nInstruction: What is the sky blue?\nResponse: "
example_title: "Knowledge-sky"
- text: "Below is an instruction that describes a task. Write a response that appropriately completes the request.\nInstruction: Define electrostatic potential\nResponse: "
example_title: "Knowledge-electricity"
---
# Model Card for Model ID
This is the fine tuned version of OpenAI's GPT-2 with Alpaca dataset.
This has been fine tuned for 20 epochs at batch size of 14.
### Model Description
```
license: mit
dataset: tatsu-lab/alpaca
language: en
library_name: transformers
pipeline_tag: text-generation
base_model: gpt2
```
## Examples
1. Give me tips to be healthy
```
Here are some tips to be healthy:
- Eat a balanced diet that includes plenty of fruits, vegetables, whole grains, lean proteins, and healthy fats.
- Exercise regularly to maintain a healthy weight and reduce the risk of developing any health issues.
- Get enough sleep and stay hydrated by drinking plenty of water.
- Get enough fiber and healthy fats to support your overall health and wellbeing.
- Practice mindful breathing and avoid activities that can increase stress.
- Get enough sleep and maintain a healthy weight.
```
2. Write a poem on cows
```
Cows are one of the most beloved animals in the world,
They live in lush green fields and care for the ground.
Their milk is a sweet beverage,
They live in a world of peace and tranquility.
Though cows may seem small in comparison,
They have a special bond with their owners.
They care for their cows, and they do it right.
The milk they produce is vital,
They have a special place in our hearts.
They are loyal friends, and they will never be defeated.
Cows are a source of joy and joy,
They will never be out of place again.
```
3. Which is the fastest animal in the world
```
The cheetah is the fastest animal in the world.
```
## Bias, Risks, and Limitations
This model has all the biases, risks and limitations of base gpt2 model.
## Recommendation
The ideal format for inference is:
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
Instruction:{question}
Response:
```
Replace {question} with the question of your choice.
The parameters I used for inference are:
```
top_k=20
top_p=.9
temperature = .7
```
## References used
1. GPT2
@article{radford2019language,
title={Language Models are Unsupervised Multitask Learners},
author={Radford, Alec and Wu, Jeff and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya},
year={2019}
}
2. tatsu-lab/alpaca
@misc{alpaca,
author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
title = {Stanford Alpaca: An Instruction-following LLaMA model},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}},
}
|