ABerlanga commited on
Commit
f3e1757
1 Parent(s): 25ed65e

Model save

Browse files
Files changed (1) hide show
  1. README.md +8 -7
README.md CHANGED
@@ -1,12 +1,13 @@
1
  ---
 
2
  library_name: peft
3
  tags:
4
  - trl
5
  - sft
6
  - generated_from_trainer
 
7
  datasets:
8
  - generator
9
- base_model: meta-llama/Llama-2-7b-hf
10
  model-index:
11
  - name: code-llama-7b-text-to-sql
12
  results: []
@@ -17,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # code-llama-7b-text-to-sql
19
 
20
- This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the generator dataset.
21
 
22
  ## Model description
23
 
@@ -37,11 +38,11 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0002
40
- - train_batch_size: 1
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 2
44
- - total_train_batch_size: 2
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: constant
47
  - lr_scheduler_warmup_ratio: 0.03
@@ -53,8 +54,8 @@ The following hyperparameters were used during training:
53
 
54
  ### Framework versions
55
 
56
- - PEFT 0.7.2.dev0
57
  - Transformers 4.36.2
58
- - Pytorch 2.2.0a0+81ea7a4
59
  - Datasets 2.16.1
60
- - Tokenizers 0.15.1
 
1
  ---
2
+ license: llama2
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: codellama/CodeLlama-7b-hf
9
  datasets:
10
  - generator
 
11
  model-index:
12
  - name: code-llama-7b-text-to-sql
13
  results: []
 
18
 
19
  # code-llama-7b-text-to-sql
20
 
21
+ This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on the generator dataset.
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0002
41
+ - train_batch_size: 3
42
  - eval_batch_size: 8
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
+ - total_train_batch_size: 6
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_ratio: 0.03
 
54
 
55
  ### Framework versions
56
 
57
+ - PEFT 0.9.1.dev0
58
  - Transformers 4.36.2
59
+ - Pytorch 2.1.2+cu121
60
  - Datasets 2.16.1
61
+ - Tokenizers 0.15.2