arnavgrg commited on
Commit
8b9d038
1 Parent(s): 74a0694

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -1
README.md CHANGED
@@ -3,8 +3,37 @@ library_name: peft
3
  tags:
4
  - text-generation
5
  ---
6
- ## Training procedure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
 
 
 
 
 
 
 
 
8
 
9
  The following `bitsandbytes` quantization config was used during training:
10
  - load_in_8bit: False
@@ -16,6 +45,7 @@ The following `bitsandbytes` quantization config was used during training:
16
  - bnb_4bit_quant_type: nf4
17
  - bnb_4bit_use_double_quant: True
18
  - bnb_4bit_compute_dtype: float16
 
19
  ### Framework versions
20
 
21
 
 
3
  tags:
4
  - text-generation
5
  ---
6
+ ## QLoRA weights using Llama-2-7b for the Code Alpaca Dataset
7
+
8
+ This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
9
+ I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU.
10
+
11
+ Dataset: https://github.com/sahil280114/codealpaca
12
+
13
+ To use these weights:
14
+ ```
15
+ from peft import PeftModel, PeftConfig
16
+ from transformers import AutoModelForCausalLM
17
+
18
+ config = PeftConfig.from_pretrained("arnavgrg/codealpaca-qlora")
19
+ model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf")
20
+ model = PeftModel.from_pretrained(model, "arnavgrg/codealpaca-qlora")
21
+ ```
22
+
23
+ Prompt Template:
24
+ ```
25
+ Below is an instruction that describes a task, paired with an input
26
+ that provides further context. Write a response that appropriately
27
+ completes the request.
28
 
29
+ ### Instruction: {instruction}
30
+
31
+ ### Input: {input}
32
+
33
+ ### Response:
34
+ ```
35
+
36
+ ## Training procedure
37
 
38
  The following `bitsandbytes` quantization config was used during training:
39
  - load_in_8bit: False
 
45
  - bnb_4bit_quant_type: nf4
46
  - bnb_4bit_use_double_quant: True
47
  - bnb_4bit_compute_dtype: float16
48
+
49
  ### Framework versions
50
 
51