tcarwash commited on
Commit
656b1d8
1 Parent(s): 007bbbd

tcarwash/tinyllama-instruct

Browse files
README.md CHANGED
@@ -5,7 +5,7 @@ tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- base_model: tinyllama/tinyllama-1.1b-chat-v1.0
9
  datasets:
10
  - generator
11
  model-index:
@@ -18,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # tinyllama_instruct
20
 
21
- This model is a fine-tuned version of [tinyllama/tinyllama-1.1b-chat-v1.0](https://huggingface.co/tinyllama/tinyllama-1.1b-chat-v1.0) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.2206
24
 
25
  ## Model description
26
 
@@ -46,15 +46,16 @@ The following hyperparameters were used during training:
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_steps: 0.03
49
- - num_epochs: 2
50
- - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:-----:|:----:|:---------------:|
56
- | 1.3509 | 1.0 | 1069 | 1.2261 |
57
- | 1.3601 | 2.0 | 2138 | 1.2206 |
 
 
58
 
59
 
60
  ### Framework versions
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: tinyllama/tinyllama-1.1b-intermediate-step-1431k-3t
9
  datasets:
10
  - generator
11
  model-index:
 
18
 
19
  # tinyllama_instruct
20
 
21
+ This model is a fine-tuned version of [tinyllama/tinyllama-1.1b-intermediate-step-1431k-3t](https://huggingface.co/tinyllama/tinyllama-1.1b-intermediate-step-1431k-3t) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.3012
24
 
25
  ## Model description
26
 
 
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_steps: 0.03
49
+ - num_epochs: 4
 
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss |
54
+ |:-------------:|:-----:|:-----:|:---------------:|
55
+ | 1.4385 | 1.0 | 4263 | 1.2766 |
56
+ | 1.4772 | 2.0 | 8526 | 1.2743 |
57
+ | 1.0998 | 3.0 | 12789 | 1.2837 |
58
+ | 1.3263 | 4.0 | 17052 | 1.3012 |
59
 
60
 
61
  ### Framework versions
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa83878e9ed23aea2d73dc55616f82325803e26ec7abc6627a79f7ca02cd64ba
3
- size 36073064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb0045517d11e5dc96444a144c9f7e188f081679dbbc1a52764bdf7ffcf551b
3
+ size 36058104
runs/May03_22-17-03_3d3bf14e01d0/events.out.tfevents.1714774626.3d3bf14e01d0.265.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c31b464bc4985979d24f62e32be6a14e720f0490f1a9309ce6b603f77b5e9919
3
+ size 6817
runs/May03_22-17-50_3d3bf14e01d0/events.out.tfevents.1714774673.3d3bf14e01d0.265.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4f4576169022536120fecf08e569c80760734d771e95461d32916ee0343197a
3
+ size 366170
tokenizer_config.json CHANGED
@@ -28,11 +28,10 @@
28
  }
29
  },
30
  "bos_token": "<s>",
31
- "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": false,
35
- "model_max_length": 2048,
36
  "pad_token": "</s>",
37
  "padding_side": "right",
38
  "sp_model_kwargs": {},
 
28
  }
29
  },
30
  "bos_token": "<s>",
 
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
33
  "legacy": false,
34
+ "model_max_length": 1000000000000000019884624838656,
35
  "pad_token": "</s>",
36
  "padding_side": "right",
37
  "sp_model_kwargs": {},
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d92afdb66ecaab798d9aed611224a18b199a1621cd3dc2a1d006ba3466f40d24
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db0cad0f34b276b350f7482750bf5f1c4fe718452bce3cf3888dce72e6f5b0e0
3
  size 4984