kevin009 commited on
Commit
bb8349c
1 Parent(s): 8be44d0
Files changed (1) hide show
  1. README.md +53 -1
README.md CHANGED
@@ -4,4 +4,56 @@ language:
4
  - en
5
  ---
6
 
7
- Details will be updated
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - en
5
  ---
6
 
7
+ # Model Card: kevin009/lamatama
8
+
9
+ ## Model Description
10
+ The `kevin009/lamatama` model is a groundbreaking achievement in the field of language modeling, showcasing the power of leveraging a substantial dataset and state-of-the-art training techniques. This model is designed to push the boundaries of what's possible in natural language understanding and generation.
11
+
12
+ ### Training Details
13
+ - **Model Architecture**: The `kevin009/lamatama` model is built upon the architecture and tokenizer of Llama 2, ensuring compatibility and easy integration with various open-source projects.
14
+ - **Dataset**: It was pretrained on an impressive 3 trillion tokens, a scale that allows for a deep and nuanced understanding of language.
15
+ - **Training Period**: The training process was carried out over 90 days, utilizing 16 A100-40G GPUs, a testament to the model's efficiency and the team's optimization skills.
16
+
17
+ ### Fine-tuning
18
+ This specific version of the model has been fine-tuned to excel in chat-based applications. It builds upon the `TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T` model, incorporating learnings and optimizations from HF's Zephyr's training recipe.
19
+
20
+ - **Initial Phase**: The model was first fine-tuned on a variant of the UltraChat dataset, which is rich in synthetic dialogues generated by ChatGPT.
21
+ - **Further Alignment**: Subsequent alignment was achieved using 🤗 TRL's DPOTrainer with the openbmb/UltraFeedback dataset, comprising 64k prompts and model completions ranked by GPT-4.
22
+
23
+ ## How to Use
24
+ Ensure you have `transformers>=4.34`. For detailed instructions and updates, check out the GitHub page for `kevin009/lamatama`.
25
+
26
+ ### Installation (for versions <= v4.34)
27
+ ```bash
28
+ pip install git+https://github.com/huggingface/transformers.git
29
+ pip install accelerate
30
+ ```
31
+
32
+ ### Example Usage
33
+ Here's a quick guide on using `kevin009/lamatama` for generating text:
34
+
35
+ ```python
36
+ import torch
37
+ from transformers import pipeline
38
+
39
+ # Initialize the pipeline
40
+ pipe = pipeline("text-generation", model="kevin009/lamatama", torch_dtype=torch.bfloat16, device_map="auto")
41
+
42
+ # Sample dialogue with templating
43
+ messages = [
44
+ {"role": "system", "content": "You are a friendly chatbot who always responds in the style of a pirate"},
45
+ {"role": "user", "content": "How many helicopters can a human eat in one sitting?"}
46
+ ]
47
+
48
+ # Generate prompt and outputs
49
+ prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
50
+ outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
51
+ print(outputs[0]["generated_text"])
52
+ ```
53
+
54
+ ## Acknowledgements
55
+ This model is a product of collaboration and innovative approaches to language modeling. We extend our thanks to all contributors, as well as the creators of the datasets and training methodologies that made `kevin009/lamatama` a reality.
56
+
57
+ ---
58
+
59
+ This model card introduces `kevin009/lamatama`, a versatile and powerful language model fine-tuned for chat applications, demonstrating exceptional understanding and generation capabilities.