HugoVoxx commited on
Commit
d864d82
1 Parent(s): bb7f30d

HugoVoxx/Gemma-2-2b-it-ag

Browse files
README.md CHANGED
@@ -27,14 +27,14 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/hugovoxx-fpt-university/Fine-tune%20Gemma-2-2b-it%20on%20AlphaGeometry%20Dataset/runs/z1bghq7x)
31
 
32
  This model was trained with SFT.
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.0
37
- - Transformers: 4.46.1
38
  - Pytorch: 2.4.0
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.0
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/hugovoxx-fpt-university/Fine-tune%20Gemma-2-2b-it%20on%20AlphaGeometry%20Dataset/runs/8lpwdqxj)
31
 
32
  This model was trained with SFT.
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.0
37
+ - Transformers: 4.46.2
38
  - Pytorch: 2.4.0
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.0
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
  "down_proj",
25
- "o_proj",
26
- "v_proj",
27
- "up_proj",
28
  "q_proj",
29
- "gate_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "down_proj",
24
+ "gate_proj",
25
+ "k_proj",
 
26
  "q_proj",
27
+ "up_proj",
28
+ "o_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ace8410a30faa2f6d79e67fdbfebbb25e9f281452ef4b228f549a28828982849
3
  size 664584480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:828b2f4a9f9248f70a7f24c0db1ba92387398ff2472917fc1b42497f8286f1f9
3
  size 664584480
special_tokens_map.json CHANGED
@@ -17,13 +17,7 @@
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
- "pad_token": {
21
- "content": "<pad>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
  "unk_token": {
28
  "content": "<unk>",
29
  "lstrip": false,
 
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
+ "pad_token": "<eos>",
 
 
 
 
 
 
21
  "unk_token": {
22
  "content": "<unk>",
23
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6ce83119bb404f7f0a6e621b76759d476357dcd01241a90f9ca136ae2b3c11c
3
  size 34362972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9afc4612a55c9c4b6223058eb59ae051813e1a51cb3f95b0f29ad864d50e4781
3
  size 34362972
tokenizer_config.json CHANGED
@@ -2004,7 +2004,7 @@
2004
  "clean_up_tokenization_spaces": false,
2005
  "eos_token": "<eos>",
2006
  "model_max_length": 1000000000000000019884624838656,
2007
- "pad_token": "<pad>",
2008
  "sp_model_kwargs": {},
2009
  "spaces_between_special_tokens": false,
2010
  "tokenizer_class": "GemmaTokenizer",
 
2004
  "clean_up_tokenization_spaces": false,
2005
  "eos_token": "<eos>",
2006
  "model_max_length": 1000000000000000019884624838656,
2007
+ "pad_token": "<eos>",
2008
  "sp_model_kwargs": {},
2009
  "spaces_between_special_tokens": false,
2010
  "tokenizer_class": "GemmaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:596413e096e180e4d997c1384160426f0264bcbb06d1845112850d4504f35f5c
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5832e276812db679395e1cdc57431215e3be0e358d384f67cebebe545fe2745a
3
  size 5496