Update README.md
Browse files
README.md
CHANGED
@@ -24,7 +24,9 @@ inference:
|
|
24 |
max_new_tokens: 1024
|
25 |
|
26 |
---
|
27 |
-
|
|
|
|
|
28 |
|
29 |
This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
|
30 |
I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
|
@@ -34,6 +36,8 @@ Dataset and training parameters are borrowed from: https://github.com/sahil28011
|
|
34 |
but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
|
35 |
toolkit for LLMs that Predibase is built on.
|
36 |
|
|
|
|
|
37 |
To use these weights:
|
38 |
```python
|
39 |
from peft import PeftModel, PeftConfig
|
|
|
24 |
max_new_tokens: 1024
|
25 |
|
26 |
---
|
27 |
+
# QLoRA weights using Llama-2-7b for the Code Alpaca Dataset
|
28 |
+
|
29 |
+
# Fine-Tuning on Predibase
|
30 |
|
31 |
This model was fine-tuned using [Predibase](https://predibase.com/), the first low-code AI platform for engineers.
|
32 |
I fine-tuned base Llama-2-7b using LoRA with 4 bit quantization on a single T4 GPU, which cost approximately $3 to train
|
|
|
36 |
but all of these parameters including DeepSpeed can be directly used with [Ludwig](https://ludwig.ai/latest/), the open-source
|
37 |
toolkit for LLMs that Predibase is built on.
|
38 |
|
39 |
+
# How To Use The Model
|
40 |
+
|
41 |
To use these weights:
|
42 |
```python
|
43 |
from peft import PeftModel, PeftConfig
|