zzz99 commited on
Commit
38974bc
1 Parent(s): ae6d589

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -19,14 +19,14 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
- "v_proj",
24
- "lm_head",
25
- "down_proj",
26
  "up_proj",
 
 
27
  "gate_proj",
 
28
  "o_proj",
29
- "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
 
 
 
23
  "up_proj",
24
+ "lm_head",
25
+ "v_proj",
26
  "gate_proj",
27
+ "down_proj",
28
  "o_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e98738d7c7a516fbc2dcae6a1c316482ad8288f876b2879250f64d4b63afbff2
3
  size 992238864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50c28859ca840accc184363cc96e0526bbc6a85a4feb319d60b5c93ee09bd057
3
  size 992238864
tokenizer.json CHANGED
@@ -1,19 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 960,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": "BatchLongest",
11
- "direction": "Right",
12
- "pad_to_multiple_of": null,
13
- "pad_id": 100001,
14
- "pad_type_id": 0,
15
- "pad_token": "<|end▁of▁sentence|>"
16
- },
17
  "added_tokens": [
18
  {
19
  "id": 100000,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 100000,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e54dc6fb7c8c9eabb359fd070b47703a6beaac7e72014e3502d0a89eaee6c31f
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdce42c771ab4c03b8c157048e368f4375c8f7d94191586d17aec8e85913ec1a
3
  size 4856