yahma commited on
Commit
e09bb18
1 Parent(s): b9e824e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -7,11 +7,11 @@ This repo contains a low-rank adapter for LLaMA-13b fit on the Cleaned Alpaca da
7
 
8
  This version of the weights was trained with the following hyperparameters:
9
 
10
- Cleaned dataset: Snapshot April 2, 2023
11
- Epochs: 3
12
- Validation set size: 2000
13
  Batch size: 128
14
- Micro batch size: 8
15
  Cutoff length: 512
16
  Learning rate: 3e-4
17
  Lora r: 16
@@ -22,10 +22,10 @@ That is:
22
  python finetune.py \
23
  --base_model='decapoda-research/llama-13b-hf' \
24
  --data_path 'yahma/alpaca-cleaned' \
25
- --num_epochs=3 \
26
  --cutoff_len=512 \
27
  --output_dir='./lora-alpaca' \
28
  --lora_target_modules='[q_proj,k_proj, v_proj, o_proj]' \
29
  --lora_r=16 \
30
- --val_set_size 2000 \
31
- --micro_batch_size=8
 
7
 
8
  This version of the weights was trained with the following hyperparameters:
9
 
10
+ Cleaned dataset: Snapshot April 9, 2023
11
+ Epochs: 4
12
+ Validation set size: 1500
13
  Batch size: 128
14
+ Micro batch size: 4
15
  Cutoff length: 512
16
  Learning rate: 3e-4
17
  Lora r: 16
 
22
  python finetune.py \
23
  --base_model='decapoda-research/llama-13b-hf' \
24
  --data_path 'yahma/alpaca-cleaned' \
25
+ --num_epochs=4 \
26
  --cutoff_len=512 \
27
  --output_dir='./lora-alpaca' \
28
  --lora_target_modules='[q_proj,k_proj, v_proj, o_proj]' \
29
  --lora_r=16 \
30
+ --val_set_size 1500 \
31
+ --micro_batch_size=4