ale-bay commited on
Commit
af95ed9
1 Parent(s): 7f2ec99

Model save

Browse files
README.md CHANGED
@@ -2,16 +2,12 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
  - trl
10
  - sft
11
  - generated_from_trainer
12
  base_model: mistralai/Mistral-7B-v0.1
13
  datasets:
14
- - HuggingFaceH4/ultrachat_200k
15
  model-index:
16
  - name: zephyr-7b-sft-qlora
17
  results: []
@@ -22,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  # zephyr-7b-sft-qlora
24
 
25
- This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the HuggingFaceH4/ultrachat_200k dataset.
26
  It achieves the following results on the evaluation set:
27
- - Loss: 0.9499
28
 
29
  ## Model description
30
 
@@ -48,10 +44,10 @@ The following hyperparameters were used during training:
48
  - eval_batch_size: 8
49
  - seed: 42
50
  - distributed_type: multi-GPU
51
- - num_devices: 4
52
  - gradient_accumulation_steps: 2
53
- - total_train_batch_size: 32
54
- - total_eval_batch_size: 32
55
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
  - lr_scheduler_type: cosine
57
  - lr_scheduler_warmup_ratio: 0.1
@@ -61,13 +57,13 @@ The following hyperparameters were used during training:
61
 
62
  | Training Loss | Epoch | Step | Validation Loss |
63
  |:-------------:|:-----:|:----:|:---------------:|
64
- | 0.9462 | 1.0 | 4357 | 0.9499 |
65
 
66
 
67
  ### Framework versions
68
 
69
  - PEFT 0.7.1
70
- - Transformers 4.38.2
71
- - Pytorch 2.1.2
72
- - Datasets 2.14.6
73
  - Tokenizers 0.15.2
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
 
 
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  base_model: mistralai/Mistral-7B-v0.1
9
  datasets:
10
+ - generator
11
  model-index:
12
  - name: zephyr-7b-sft-qlora
13
  results: []
 
18
 
19
  # zephyr-7b-sft-qlora
20
 
21
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.9517
24
 
25
  ## Model description
26
 
 
44
  - eval_batch_size: 8
45
  - seed: 42
46
  - distributed_type: multi-GPU
47
+ - num_devices: 2
48
  - gradient_accumulation_steps: 2
49
+ - total_train_batch_size: 16
50
+ - total_eval_batch_size: 16
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_ratio: 0.1
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:-----:|:----:|:---------------:|
60
+ | 0.9567 | 1.0 | 8714 | 0.9517 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.7.1
66
+ - Transformers 4.39.3
67
+ - Pytorch 2.3.0+cu121
68
+ - Datasets 2.19.1
69
  - Tokenizers 0.15.2
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27d5963f9a1f5993302ffc19b8d9a1b6a8767fc761869a0d9ec7a7d27890d238
3
  size 83946192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:440d938e3bc97bee5a47be212f3d9a03e01eb54c44ec61cbfb7462ae2fae7351
3
  size 83946192
all_results.json CHANGED
@@ -1,13 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.9498828649520874,
4
- "eval_runtime": 985.6149,
5
- "eval_samples": 23110,
6
- "eval_samples_per_second": 15.656,
7
- "eval_steps_per_second": 0.49,
8
- "train_loss": 0.9569843266025059,
9
- "train_runtime": 33545.7523,
10
- "train_samples": 207865,
11
- "train_samples_per_second": 4.156,
12
- "train_steps_per_second": 0.13
13
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.8657062510599203,
4
+ "train_runtime": 31020.7888,
5
+ "train_samples": 207864,
6
+ "train_samples_per_second": 4.495,
7
+ "train_steps_per_second": 0.281
 
 
 
 
 
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.9569843266025059,
4
- "train_runtime": 33545.7523,
5
- "train_samples": 207865,
6
- "train_samples_per_second": 4.156,
7
- "train_steps_per_second": 0.13
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.8657062510599203,
4
+ "train_runtime": 31020.7888,
5
+ "train_samples": 207864,
6
+ "train_samples_per_second": 4.495,
7
+ "train_steps_per_second": 0.281
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff