|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
model-index: |
|
- name: lamatama |
|
results: |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: AI2 Reasoning Challenge (25-Shot) |
|
type: ai2_arc |
|
config: ARC-Challenge |
|
split: test |
|
args: |
|
num_few_shot: 25 |
|
metrics: |
|
- type: acc_norm |
|
value: 36.35 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: HellaSwag (10-Shot) |
|
type: hellaswag |
|
split: validation |
|
args: |
|
num_few_shot: 10 |
|
metrics: |
|
- type: acc_norm |
|
value: 61.12 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MMLU (5-Shot) |
|
type: cais/mmlu |
|
config: all |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 24.72 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: TruthfulQA (0-shot) |
|
type: truthful_qa |
|
config: multiple_choice |
|
split: validation |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: mc2 |
|
value: 37.67 |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: Winogrande (5-shot) |
|
type: winogrande |
|
config: winogrande_xl |
|
split: validation |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 60.77 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: GSM8k (5-shot) |
|
type: gsm8k |
|
config: main |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 2.27 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=kevin009/lamatama |
|
name: Open LLM Leaderboard |
|
--- |
|
|
|
# Model Card: kevin009/lamatama |
|
|
|
## Model Description |
|
The `kevin009/lamatama` model is a groundbreaking achievement in the field of language modeling, showcasing the power of leveraging a substantial dataset and state-of-the-art training techniques. This model is designed to push the boundaries of what's possible in natural language understanding and generation. |
|
|
|
### Training Details |
|
- **Model Architecture**: The `kevin009/lamatama` model is built upon the architecture and tokenizer of Llama 2, ensuring compatibility and easy integration with various open-source projects. |
|
- **Dataset**: It was pretrained on an impressive 3 trillion tokens, a scale that allows for a deep and nuanced understanding of language. |
|
- **Training Period**: The training process was carried out over 90 days, utilizing 16 A100-40G GPUs, a testament to the model's efficiency and the team's optimization skills. |
|
|
|
### Fine-tuning |
|
This specific version of the model has been fine-tuned to excel in chat-based applications. It builds upon the `TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T` model, incorporating learnings and optimizations from HF's Zephyr's training recipe. |
|
|
|
- **Initial Phase**: The model was first fine-tuned on a variant of the UltraChat dataset, which is rich in synthetic dialogues generated by ChatGPT. |
|
- **Further Alignment**: Subsequent alignment was achieved using 🤗 TRL's DPOTrainer with the openbmb/UltraFeedback dataset, comprising 64k prompts and model completions ranked by GPT-4. |
|
|
|
## How to Use |
|
Ensure you have `transformers>=4.34`. For detailed instructions and updates, check out the GitHub page for `kevin009/lamatama`. |
|
|
|
### Installation (for versions <= v4.34) |
|
```bash |
|
pip install git+https://github.com/huggingface/transformers.git |
|
pip install accelerate |
|
``` |
|
|
|
### Example Usage |
|
Here's a quick guide on using `kevin009/lamatama` for generating text: |
|
|
|
```python |
|
import torch |
|
from transformers import pipeline |
|
|
|
# Initialize the pipeline |
|
pipe = pipeline("text-generation", model="kevin009/lamatama", torch_dtype=torch.bfloat16, device_map="auto") |
|
|
|
# Sample dialogue with templating |
|
messages = [ |
|
{"role": "system", "content": "You are a friendly chatbot who always responds in the style of a pirate"}, |
|
{"role": "user", "content": "How many helicopters can a human eat in one sitting?"} |
|
] |
|
|
|
# Generate prompt and outputs |
|
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
|
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) |
|
print(outputs[0]["generated_text"]) |
|
``` |
|
|
|
## Acknowledgements |
|
This model is a product of collaboration and innovative approaches to language modeling. We extend our thanks to all contributors, as well as the creators of the datasets and training methodologies that made `kevin009/lamatama` a reality. |
|
|
|
--- |
|
|
|
This model card introduces `kevin009/lamatama`, a versatile and powerful language model fine-tuned for chat applications, demonstrating exceptional understanding and generation capabilities. |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_kevin009__lamatama) |
|
|
|
| Metric |Value| |
|
|---------------------------------|----:| |
|
|Avg. |37.15| |
|
|AI2 Reasoning Challenge (25-Shot)|36.35| |
|
|HellaSwag (10-Shot) |61.12| |
|
|MMLU (5-Shot) |24.72| |
|
|TruthfulQA (0-shot) |37.67| |
|
|Winogrande (5-shot) |60.77| |
|
|GSM8k (5-shot) | 2.27| |
|
|
|
|