piotr25691 commited on
Commit
56aa291
1 Parent(s): ed4c338

Name override with rsLoRA(rank=128, alpha=256)

Browse files
README.md CHANGED
@@ -2,6 +2,7 @@
2
  base_model: SicariusSicariiStuff/Impish_LLAMA_3B
3
  datasets:
4
  - KingNish/reasoning-base-20k
 
5
  language:
6
  - en
7
  license: llama3.2
@@ -17,9 +18,9 @@ tags:
17
 
18
  # Model Description
19
 
20
- A work in progress uncensored roleplay reasoning Llama 3.2 3B model trained on reasoning data.
21
 
22
- Since I used different training code, it is unknown whether it generates the same kind of reasoning.
23
  Here is what inference code you should use:
24
  ```py
25
  from transformers import AutoModelForCausalLM, AutoTokenizer
@@ -27,7 +28,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
27
  MAX_REASONING_TOKENS = 1024
28
  MAX_RESPONSE_TOKENS = 512
29
 
30
- model_name = "piotr25691/thea-3b-25r"
31
 
32
  model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")
33
  tokenizer = AutoTokenizer.from_pretrained(model_name)
@@ -43,7 +44,7 @@ reasoning_inputs = tokenizer(reasoning_template, return_tensors="pt").to(model.d
43
  reasoning_ids = model.generate(**reasoning_inputs, max_new_tokens=MAX_REASONING_TOKENS)
44
  reasoning_output = tokenizer.decode(reasoning_ids[0, reasoning_inputs.input_ids.shape[1]:], skip_special_tokens=True)
45
 
46
- # print("REASONING: " + reasoning_output)
47
 
48
  # Generate answer
49
  messages.append({"role": "reasoning", "content": reasoning_output})
 
2
  base_model: SicariusSicariiStuff/Impish_LLAMA_3B
3
  datasets:
4
  - KingNish/reasoning-base-20k
5
+ - piotr25691/thea-name-overrides
6
  language:
7
  - en
8
  license: llama3.2
 
18
 
19
  # Model Description
20
 
21
+ An uncensored roleplay reasoning Llama 3.2 3B model trained on reasoning data.
22
 
23
+ It has been trained using improved training code, and gives an improved performance.
24
  Here is what inference code you should use:
25
  ```py
26
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
28
  MAX_REASONING_TOKENS = 1024
29
  MAX_RESPONSE_TOKENS = 512
30
 
31
+ model_name = "piotr25691/thea-rp-3b-25r"
32
 
33
  model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")
34
  tokenizer = AutoTokenizer.from_pretrained(model_name)
 
44
  reasoning_ids = model.generate(**reasoning_inputs, max_new_tokens=MAX_REASONING_TOKENS)
45
  reasoning_output = tokenizer.decode(reasoning_ids[0, reasoning_inputs.input_ids.shape[1]:], skip_special_tokens=True)
46
 
47
+ print("REASONING: " + reasoning_output)
48
 
49
  # Generate answer
50
  messages.append({"role": "reasoning", "content": reasoning_output})
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7de2f1c5c5b7d1297db469f8be0a0611a70e4dbf44497dae039d8fafdb76119
3
  size 4965799096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0be8107a36536ae24d8636b10d2ae99d262a1d05f966e7aad5140c781a4304e3
3
  size 4965799096
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79507f77145d985d4874a55670c05cc14963998ad09af195d5a9cd4fcae2de80
3
  size 1459729952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b77771ae45d388f52a3fbe1c085f67ce3baccd837ec2565f1025af5ce9c5e2
3
  size 1459729952
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:051830f2f6c06d23b79bfeb1cb00c36ab32a29c2905e80e0b8e22148b654ec8b
3
- size 17210197
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12487b766b0b1584dcc5311824df327d5ea154939524790c643cdf2a3f6adf9f
3
+ size 17209921