paul-stansifer
commited on
Commit
•
bb3d49f
1
Parent(s):
c8f6395
paul-stansifer/llama3-qwantz-coherent
Browse files
README.md
CHANGED
@@ -9,7 +9,6 @@ metrics:
|
|
9 |
model-index:
|
10 |
- name: llama3-qwantz-coherent
|
11 |
results: []
|
12 |
-
pipeline_tag: text-classification
|
13 |
---
|
14 |
|
15 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -17,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
# llama3-qwantz-coherent
|
19 |
|
20 |
-
This model is a fine-tuned version of [unsloth/llama-3-8b-bnb-4bit](https://huggingface.co/unsloth/llama-3-8b-bnb-4bit) on
|
21 |
It achieves the following results on the evaluation set:
|
22 |
-
- Loss: 0.
|
23 |
-
- Accuracy: 0.
|
24 |
|
25 |
## Model description
|
26 |
|
@@ -51,38 +50,8 @@ The following hyperparameters were used during training:
|
|
51 |
|
52 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
53 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
54 |
-
| 0.
|
55 |
-
|
56 |
-
|
57 |
-
```
|
58 |
-
Can save 90% of coherent strings by discarding 94% of dp strings (cutoff is 75.22732615470886)
|
59 |
-
Can save 95% of coherent strings by discarding 91% of dp strings (cutoff is -64.16597366333008)
|
60 |
-
Can save 98% of coherent strings by discarding 86% of dp strings (cutoff is -93.8580572605133)
|
61 |
-
Can save 99% of coherent strings by discarding 78% of dp strings (cutoff is -99.2882251739502)
|
62 |
-
|
63 |
-
I have constructed a rocket-ship for myself ==> coherent: 99.92%
|
64 |
-
I have constructed a refund for a ==> dp: 98.67%
|
65 |
-
Descartes was a dude who wrote "Cogito ergo sum" which means "I think, therefore I am". PRETTY ==> coherent: 100.00%
|
66 |
-
Descartes was a dude who wrote "Cogito ergo sum" which means "I think, therefore finite lifetimes the ==> dp: 99.99%
|
67 |
-
That's certainly one way of looking at it, right, Dromiceiomimums? ==> coherent: 100.00%
|
68 |
-
That's certainly one way of looking at it, is the ==> dp: 98.15%
|
69 |
-
I'm here to pick up my prescription "Happy New year 2004" glasses! They have a plastic "2" on ==> coherent: 99.79%
|
70 |
-
I'm here to pick up my prescription "Happy New year 2004" glasses! They have come into cartoon stereotypes ==> dp: 99.99%
|
71 |
-
I didn't mean for that to be ==> coherent: 99.13%
|
72 |
-
I didn't mean for the police officer ==> dp: 64.15%
|
73 |
-
You know what would go down if Nintendo came over? ==> coherent: 100.00%
|
74 |
-
You know what would go down if Nintendo i live ==> dp: 100.00%
|
75 |
-
"Aw shucks! I guess it IS true that you're never too ==> coherent: 100.00%
|
76 |
-
"Aw shucks! I guess it IS true that worse the turmeric ==> dp: 100.00%
|
77 |
-
Is it true that the only questions worth asking are those that ==> coherent: 99.98%
|
78 |
-
Is it true that the only questions worth preserving if i've been ==> dp: 99.99%
|
79 |
-
What? No, he was in pieces. His hand even landed in ==> coherent: 99.44%
|
80 |
-
What? No, he was in pieces. His gun that an excellent ==> dp: 99.99%
|
81 |
-
Also, many of the signs are really evocative, so they're easy to ==> coherent: 100.00%
|
82 |
-
Also, many of the signs are really evocative, so they're approved aaargh ==> dp: 100.00%
|
83 |
-
Another beautiful hot day! I look forward to these "dog days" ==> coherent: 99.97%
|
84 |
-
Another beautiful hot day! I look forward to return to make ==> dp: 99.81%
|
85 |
-
```
|
86 |
|
87 |
### Framework versions
|
88 |
|
|
|
9 |
model-index:
|
10 |
- name: llama3-qwantz-coherent
|
11 |
results: []
|
|
|
12 |
---
|
13 |
|
14 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
16 |
|
17 |
# llama3-qwantz-coherent
|
18 |
|
19 |
+
This model is a fine-tuned version of [unsloth/llama-3-8b-bnb-4bit](https://huggingface.co/unsloth/llama-3-8b-bnb-4bit) on an unknown dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.3295
|
22 |
+
- Accuracy: 0.8758
|
23 |
|
24 |
## Model description
|
25 |
|
|
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
52 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
53 |
+
| 0.4482 | 1.0 | 1428 | 0.3295 | 0.8758 |
|
54 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
### Framework versions
|
57 |
|
adapter_config.json
CHANGED
@@ -20,10 +20,10 @@
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
-
"o_proj",
|
24 |
"q_proj",
|
25 |
"v_proj",
|
26 |
-
"k_proj"
|
|
|
27 |
],
|
28 |
"task_type": "SEQ_CLS",
|
29 |
"use_dora": false,
|
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
|
|
23 |
"q_proj",
|
24 |
"v_proj",
|
25 |
+
"k_proj",
|
26 |
+
"o_proj"
|
27 |
],
|
28 |
"task_type": "SEQ_CLS",
|
29 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 54593240
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aafbb760ed8b904e0cf8d1c3726ac2d22e6f65940f9e2b06b486f66dc54de260
|
3 |
size 54593240
|
runs/May03_18-12-38_048cd167e598/events.out.tfevents.1714759959.048cd167e598.211.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:446fb0a0caac3cfa4a41f0b30f0e82fd25cde23bb53e72329bc4884f97d0341c
|
3 |
+
size 17949
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5048
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73f9fa2fbada4e354ca6f78290205d9ec5eecfada10232d0a46295520fe058ef
|
3 |
size 5048
|