llama-2-7b-guanaco / README.md
mlabonne's picture
Update README.md
21f0e70
metadata
license: apache-2.0
datasets:
  - timdettmers/openassistant-guanaco
pipeline_tag: text-generation

Llama-2-7b-guanaco

📝 Article | 💻 Colab | 📄 Script

This is a llama-2-7b-chat-hf model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2 dataset.

🔧 Training

It was trained on a Google Colab notebook with a T4 GPU and high RAM.

💻 Usage

# pip install transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "mlabonne/llama-2-7b-miniguanaco"
prompt = "What is a large language model?"

tokenizer = AutoTokenizer.from_pretrained(model)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

sequences = pipeline(
    f'<s>[INST] {prompt} [/INST]',
    do_sample=True,
    top_k=10,
    num_return_sequences=1,
    eos_token_id=tokenizer.eos_token_id,
    max_length=200,
)
for seq in sequences:
    print(f"Result: {seq['generated_text']}")

Output:

A large language model is a type of artificial intelligence (AI) model that is trained to generate human-like language. The models can be trained on text from a specific genre, such as news articles, or on a large corpus of text, such as the internet. They can then be used to generate text, such as articles, stories or even entire books. These models are often used in applications such as chatbots, language translation and content generation. They have been used to write books such as: "The Last Days of New Paris" by China Miéville.

The large models are also used for many other applications such as:

  • Translation
  • Summarization
  • Sentiment Analysis
  • Text classification
  • Generative writing (creates articles, stories, and more.)
  • Conversational language generation.