arnavgrg commited on
Commit
e1dd7de
1 Parent(s): bbd04f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -24,7 +24,9 @@ inference:
24
  max_new_tokens: 1024
25
 
26
  ---
27
- ## QLoRA weights using Llama-2-7b for the Code Alpaca Dataset
 
 
28
 
29
  This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
30
  I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
@@ -34,6 +36,8 @@ Dataset and training parameters are borrowed from: https://github.com/sahil28011
34
  but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
35
  toolkit for LLMs that Predibase is built on.
36
 
 
 
37
  To use these weights:
38
  ```python
39
  from peft import PeftModel, PeftConfig
 
24
  max_new_tokens: 1024
25
 
26
  ---
27
+ # QLoRA weights using Llama-2-7b for the Code Alpaca Dataset
28
+
29
+ # Fine-Tuning on Predibase
30
 
31
  This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
32
  I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
 
36
  but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
37
  toolkit for LLMs that Predibase is built on.
38
 
39
+ # How To Use The Model
40
+
41
  To use these weights:
42
  ```python
43
  from peft import PeftModel, PeftConfig