arshiakarimian1 commited on
Commit
06a14c1
1 Parent(s): 5049d2e

Model save

Browse files
README.md CHANGED
@@ -17,8 +17,6 @@ should probably proofread and complete it, then remove this comment. -->
17
  # spam-llama3.1-8B-teacher
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 1.6686
22
 
23
  ## Model description
24
 
@@ -49,16 +47,6 @@ The following hyperparameters were used during training:
49
  - num_epochs: 1
50
  - mixed_precision_training: Native AMP
51
 
52
- ### Training results
53
-
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:------:|:----:|:---------------:|
56
- | 1.5942 | 0.2010 | 56 | 1.8397 |
57
- | 1.7531 | 0.4020 | 112 | 1.7066 |
58
- | 1.1626 | 0.6030 | 168 | 1.6820 |
59
- | 1.2725 | 0.8039 | 224 | 1.6686 |
60
-
61
-
62
  ### Framework versions
63
 
64
  - PEFT 0.12.0
 
17
  # spam-llama3.1-8B-teacher
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the None dataset.
 
 
20
 
21
  ## Model description
22
 
 
47
  - num_epochs: 1
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
 
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - PEFT 0.12.0
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "q_proj",
 
25
  "o_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "q_proj",
24
+ "v_proj",
25
  "o_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:526532394b5ab803debfa6cefa4a3e0d938f58a86992bf2363c8d603e67e760c
3
  size 176186808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f11430c63d5e64a51e333f694fcc29fa5c294f4bf1d0d3025456909601e8f7
3
  size 176186808
runs/Aug30_15-59-38_d06352ad8bfb/events.out.tfevents.1725033580.d06352ad8bfb.6293.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e321187f77fa654185547545747399c34307ada0e71e95f9ed1877e2b98dd98
3
+ size 6437
runs/Aug30_16-12-42_d06352ad8bfb/events.out.tfevents.1725034364.d06352ad8bfb.10154.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93b040e8fbd739bfe40d4cdf2c0d3c6464b8d9396cdbd59496fd6a744b5b4f3
3
+ size 8300
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca9e0e00416efd7eb3938fd9803d15f02b9fb777f8207eb245b106eca76190d0
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eadffd5349933551cb43ef16891bd78a9eb6c095fd94f7bcaa916cb3695ff167
3
  size 5496