update
Browse files
README.md
CHANGED
@@ -4,4 +4,56 @@ language:
|
|
4 |
- en
|
5 |
---
|
6 |
|
7 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
- en
|
5 |
---
|
6 |
|
7 |
+
# Model Card: kevin009/lamatama
|
8 |
+
|
9 |
+
## Model Description
|
10 |
+
The `kevin009/lamatama` model is a groundbreaking achievement in the field of language modeling, showcasing the power of leveraging a substantial dataset and state-of-the-art training techniques. This model is designed to push the boundaries of what's possible in natural language understanding and generation.
|
11 |
+
|
12 |
+
### Training Details
|
13 |
+
- **Model Architecture**: The `kevin009/lamatama` model is built upon the architecture and tokenizer of Llama 2, ensuring compatibility and easy integration with various open-source projects.
|
14 |
+
- **Dataset**: It was pretrained on an impressive 3 trillion tokens, a scale that allows for a deep and nuanced understanding of language.
|
15 |
+
- **Training Period**: The training process was carried out over 90 days, utilizing 16 A100-40G GPUs, a testament to the model's efficiency and the team's optimization skills.
|
16 |
+
|
17 |
+
### Fine-tuning
|
18 |
+
This specific version of the model has been fine-tuned to excel in chat-based applications. It builds upon the `TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T` model, incorporating learnings and optimizations from HF's Zephyr's training recipe.
|
19 |
+
|
20 |
+
- **Initial Phase**: The model was first fine-tuned on a variant of the UltraChat dataset, which is rich in synthetic dialogues generated by ChatGPT.
|
21 |
+
- **Further Alignment**: Subsequent alignment was achieved using 🤗 TRL's DPOTrainer with the openbmb/UltraFeedback dataset, comprising 64k prompts and model completions ranked by GPT-4.
|
22 |
+
|
23 |
+
## How to Use
|
24 |
+
Ensure you have `transformers>=4.34`. For detailed instructions and updates, check out the GitHub page for `kevin009/lamatama`.
|
25 |
+
|
26 |
+
### Installation (for versions <= v4.34)
|
27 |
+
```bash
|
28 |
+
pip install git+https://github.com/huggingface/transformers.git
|
29 |
+
pip install accelerate
|
30 |
+
```
|
31 |
+
|
32 |
+
### Example Usage
|
33 |
+
Here's a quick guide on using `kevin009/lamatama` for generating text:
|
34 |
+
|
35 |
+
```python
|
36 |
+
import torch
|
37 |
+
from transformers import pipeline
|
38 |
+
|
39 |
+
# Initialize the pipeline
|
40 |
+
pipe = pipeline("text-generation", model="kevin009/lamatama", torch_dtype=torch.bfloat16, device_map="auto")
|
41 |
+
|
42 |
+
# Sample dialogue with templating
|
43 |
+
messages = [
|
44 |
+
{"role": "system", "content": "You are a friendly chatbot who always responds in the style of a pirate"},
|
45 |
+
{"role": "user", "content": "How many helicopters can a human eat in one sitting?"}
|
46 |
+
]
|
47 |
+
|
48 |
+
# Generate prompt and outputs
|
49 |
+
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
50 |
+
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
51 |
+
print(outputs[0]["generated_text"])
|
52 |
+
```
|
53 |
+
|
54 |
+
## Acknowledgements
|
55 |
+
This model is a product of collaboration and innovative approaches to language modeling. We extend our thanks to all contributors, as well as the creators of the datasets and training methodologies that made `kevin009/lamatama` a reality.
|
56 |
+
|
57 |
+
---
|
58 |
+
|
59 |
+
This model card introduces `kevin009/lamatama`, a versatile and powerful language model fine-tuned for chat applications, demonstrating exceptional understanding and generation capabilities.
|