beamaia commited on
Commit
9189f08
1 Parent(s): 28e8f6e

Training in progress, epoch 0

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e44ce263e6fd885f50d82ca515b9325375b43ee36ededb75acf161ce88bc2e41
3
- size 48
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:193d05e101073a4ae7cdda8c619f89cf75a5b0ba374cc19eeef19ee184820448
3
+ size 27280152
special_tokens_map.json CHANGED
@@ -18,7 +18,13 @@
18
  "rstrip": false,
19
  "single_word": false
20
  },
21
- "pad_token": "</s>",
 
 
 
 
 
 
22
  "unk_token": {
23
  "content": "<unk>",
24
  "lstrip": false,
 
18
  "rstrip": false,
19
  "single_word": false
20
  },
21
+ "pad_token": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
  "unk_token": {
29
  "content": "<unk>",
30
  "lstrip": false,
tokenizer.json CHANGED
@@ -62,12 +62,6 @@
62
  "id": "A",
63
  "type_id": 0
64
  }
65
- },
66
- {
67
- "SpecialToken": {
68
- "id": "</s>",
69
- "type_id": 0
70
- }
71
  }
72
  ],
73
  "pair": [
@@ -83,12 +77,6 @@
83
  "type_id": 0
84
  }
85
  },
86
- {
87
- "SpecialToken": {
88
- "id": "</s>",
89
- "type_id": 0
90
- }
91
- },
92
  {
93
  "SpecialToken": {
94
  "id": "<s>",
@@ -100,24 +88,9 @@
100
  "id": "B",
101
  "type_id": 1
102
  }
103
- },
104
- {
105
- "SpecialToken": {
106
- "id": "</s>",
107
- "type_id": 1
108
- }
109
  }
110
  ],
111
  "special_tokens": {
112
- "</s>": {
113
- "id": "</s>",
114
- "ids": [
115
- 2
116
- ],
117
- "tokens": [
118
- "</s>"
119
- ]
120
- },
121
  "<s>": {
122
  "id": "<s>",
123
  "ids": [
 
62
  "id": "A",
63
  "type_id": 0
64
  }
 
 
 
 
 
 
65
  }
66
  ],
67
  "pair": [
 
77
  "type_id": 0
78
  }
79
  },
 
 
 
 
 
 
80
  {
81
  "SpecialToken": {
82
  "id": "<s>",
 
88
  "id": "B",
89
  "type_id": 1
90
  }
 
 
 
 
 
 
91
  }
92
  ],
93
  "special_tokens": {
 
 
 
 
 
 
 
 
 
94
  "<s>": {
95
  "id": "<s>",
96
  "ids": [
tokenizer_config.json CHANGED
@@ -35,10 +35,8 @@
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "</s>",
37
  "legacy": true,
38
- "max_lenght": 2048,
39
  "model_max_length": 1000000000000000019884624838656,
40
  "pad_token": "</s>",
41
- "padding": true,
42
  "sp_model_kwargs": {},
43
  "spaces_between_special_tokens": false,
44
  "tokenizer_class": "LlamaTokenizer",
 
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "</s>",
37
  "legacy": true,
 
38
  "model_max_length": 1000000000000000019884624838656,
39
  "pad_token": "</s>",
 
40
  "sp_model_kwargs": {},
41
  "spaces_between_special_tokens": false,
42
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01268df0ff31e93c45f40da3c0a0c587f3b406fbd2c46a470d00cad57303062d
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eba5abb21dec3e0acafd4e7530e2fd84655532ed1b84c020cb7961952a74c43
3
  size 4792