moriire commited on
Commit
3775e52
1 Parent(s): c835dd3

health-assistant-v1

Browse files
README.md CHANGED
@@ -9,9 +9,6 @@ base_model: microsoft/phi-2
9
  model-index:
10
  - name: medical
11
  results: []
12
- language:
13
- - en
14
- pipeline_tag: conversational
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -43,8 +40,9 @@ The following hyperparameters were used during training:
43
  - eval_batch_size: 8
44
  - seed: 42
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
- - lr_scheduler_type: linear
47
- - num_epochs: 1
 
48
 
49
  ### Training results
50
 
@@ -53,7 +51,7 @@ The following hyperparameters were used during training:
53
  ### Framework versions
54
 
55
  - PEFT 0.7.2.dev0
56
- - Transformers 4.38.0.dev0
57
- - Pytorch 2.1.0+cu121
58
  - Datasets 2.16.1
59
  - Tokenizers 0.15.1
 
9
  model-index:
10
  - name: medical
11
  results: []
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
+ - lr_scheduler_type: cosine
44
+ - lr_scheduler_warmup_ratio: 0.03
45
+ - training_steps: 500
46
 
47
  ### Training results
48
 
 
51
  ### Framework versions
52
 
53
  - PEFT 0.7.2.dev0
54
+ - Transformers 4.37.1
55
+ - Pytorch 2.1.2+cu121
56
  - Datasets 2.16.1
57
  - Tokenizers 0.15.1
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "microsoft/phi-2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -9,21 +9,18 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 16,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
  "fc2",
23
- "q_proj",
24
- "k_proj",
25
- "v_proj",
26
- "dense",
27
  "fc1"
28
  ],
29
  "task_type": "CAUSAL_LM",
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 64,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 32,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
  "fc2",
23
+ "Wqkv",
 
 
 
24
  "fc1"
25
  ],
26
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c132c6d9ca6e90c576373c0597d6e5ddc8abf3ead848cd5225bf8d2ca2ad9b3
3
- size 94422368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:629ac58a2a288f05c82f9e6abcb9919b1fb3902e226fe86d5f49052e30bb8579
3
+ size 188814192
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "50256": {
@@ -317,8 +319,10 @@
317
  "bos_token": "<|endoftext|>",
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
 
320
  "model_max_length": 2048,
321
  "pad_token": "<|endoftext|>",
 
322
  "tokenizer_class": "CodeGenTokenizer",
323
  "unk_token": "<|endoftext|>"
324
  }
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": true,
4
  "add_prefix_space": false,
5
  "added_tokens_decoder": {
6
  "50256": {
 
319
  "bos_token": "<|endoftext|>",
320
  "clean_up_tokenization_spaces": true,
321
  "eos_token": "<|endoftext|>",
322
+ "errors": "replace",
323
  "model_max_length": 2048,
324
  "pad_token": "<|endoftext|>",
325
+ "padding_side": "left",
326
  "tokenizer_class": "CodeGenTokenizer",
327
  "unk_token": "<|endoftext|>"
328
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b89e647d42e38c3dfff9c2cfc1566acd2fd9be63f465d45ea5d2f4fed7049ac7
3
- size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ff903669e02c7a64a326fa91b78408c6dd0c58c9a68f32fc72fc615b4368d20
3
+ size 4664
vocab.json CHANGED
The diff for this file is too large to render. See raw diff