somya-kr commited on
Commit
0dc1609
1 Parent(s): 298ca42

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "down_proj",
23
- "up_proj",
24
- "gate_proj",
25
  "o_proj",
26
- "v_proj",
27
  "q_proj",
28
- "k_proj"
 
 
 
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "o_proj",
 
23
  "q_proj",
24
+ "k_proj",
25
+ "v_proj",
26
+ "gate_proj",
27
+ "up_proj",
28
+ "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d555c4f610eeeb2405b79ecdad4ae8a2def6c55e9c263faaadd9d380fcdafeb1
3
  size 1803907984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31fd187e03d4f54f1775f5f618036156fb669a333f993a2fba012f0167c05830
3
  size 1803907984
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32016,
3
+ "<|im_start|>": 32017
4
+ }
runs/Oct27_22-46-52_raptor/events.out.tfevents.1730049421.raptor.135062.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d713e1f63795bb3dc49788ae1e9e9fcb7ab03b4943b59fd943ced5e91ec7b80e
3
+ size 5884
runs/Oct27_23-03-42_raptor/events.out.tfevents.1730050459.raptor.138788.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff2d779fc09d24990d73c162bf8704d1958cdb3ec6fbcb86567cec1c49f76863
3
+ size 5114
runs/Oct28_14-26-42_raptor/events.out.tfevents.1730105811.raptor.595411.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18834d5f46112b8c7e9214cb3625b31e8b1adb8805a3c3e73d1df91c7df70f1f
3
+ size 30832
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 3072,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45ccb9c8b6b561889acea59191d66986d314e7cbd6a78abc6e49b139ca91c1e6
3
+ size 500058
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b53eb4d251e5db689026a91e9ad601c28b10ca056b174a416017a0af36c7d6e
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2d83858b87a600254f04f8866e544c89f6f8eca9113f2cf68cdaf775a37dbed
3
  size 4728