mlabonne commited on
Commit
75e7737
1 Parent(s): 32b92e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -6
README.md CHANGED
@@ -4,18 +4,59 @@ datasets:
4
  - timdettmers/openassistant-guanaco
5
  pipeline_tag: text-generation
6
  ---
 
7
 
8
  📝 [Article](https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32) |
9
  💻 [Colab](https://colab.research.google.com/drive/1PEQyJO1-f6j0S_XJ8DV50NkpzasXkrzd?usp=sharing)
10
 
11
- This is a Llama 2-7b model QLoRA fine-tuned (4-bit precision) on the [`mlabonne/guanaco-llama2-1k`](https://huggingface.co/datasets/mlabonne/guanaco-llama2) dataset.
12
 
13
- It was trained on a Google Colab notebook with a T4 GPU and high RAM. It is mainly designed for educational purposes, not for inference.
14
 
15
- You can easily import it using the `AutoModelForCausalLM` class from `transformers`:
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ```
18
- from transformers import AutoModelForCausalLM
19
 
20
- model = AutoModelForCausalLM("mlabonne/llama-2-7b-miniguanaco")
21
- ```
 
 
 
 
 
 
 
 
 
 
4
  - timdettmers/openassistant-guanaco
5
  pipeline_tag: text-generation
6
  ---
7
+ # Llama-2-7b-guanaco
8
 
9
  📝 [Article](https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32) |
10
  💻 [Colab](https://colab.research.google.com/drive/1PEQyJO1-f6j0S_XJ8DV50NkpzasXkrzd?usp=sharing)
11
 
12
+ <center><img src="https://i.imgur.com/C2x7n2a.png" width="300"></center>
13
 
14
+ This is a Llama 2-7b model QLoRA fine-tuned (4-bit precision) on the [`mlabonne/guanaco-llama2`](https://huggingface.co/datasets/mlabonne/guanaco-llama2) dataset.
15
 
16
+ ## 🔧 Training
17
 
18
+ It was trained on a Google Colab notebook with a T4 GPU and high RAM.
19
+
20
+ ## 💻 Usage
21
+
22
+ ``` python
23
+ # pip install transformers accelerate
24
+
25
+ from transformers import AutoTokenizer
26
+ import transformers
27
+ import torch
28
+
29
+ model = "mlabonne/llama-2-7b-miniguanaco"
30
+ prompt = "What is a large language model?"
31
+
32
+ tokenizer = AutoTokenizer.from_pretrained(model)
33
+ pipeline = transformers.pipeline(
34
+ "text-generation",
35
+ model=model,
36
+ torch_dtype=torch.float16,
37
+ device_map="auto",
38
+ )
39
+
40
+ sequences = pipeline(
41
+ f'<s>[INST] {prompt} [/INST]',
42
+ do_sample=True,
43
+ top_k=10,
44
+ num_return_sequences=1,
45
+ eos_token_id=tokenizer.eos_token_id,
46
+ max_length=200,
47
+ )
48
+ for seq in sequences:
49
+ print(f"Result: {seq['generated_text']}")
50
  ```
 
51
 
52
+ Output:
53
+ >A large language model is a type of artificial intelligence (AI) model that is trained to generate human-like language.Љe models can be trained on text from a specific genre, such as news articles, or on a large corpus of text, such as the internet. They can then be used to generate text, such as articles, stories or even entire books. These models are often used in applications such as chatbots, language translation and content generation. They have been used to write books such as: "The Last Days of New Paris" by China Miéville.
54
+ >
55
+ >The large models are also used for many other applications such as:
56
+ >
57
+ >- Translation
58
+ >- Summarization
59
+ >- Sentiment Analysis
60
+ >- Text classification
61
+ >- Generative writing (creates articles, stories, and more.)
62
+ >- Conversational language generation.