Vezora commited on
Commit
abc22f7
1 Parent(s): 4679b79

Correct sequence length.

Browse files

I actually checked the script and realized I used 32k I though I had used 16k no wonder training was a little longer than expected 😅 oh well it just makes a better model.

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -23,7 +23,7 @@ license: apache-2.0
23
  - **Json Mode** I did train this model on answering in JSON and using JSON tools., I have yet to try it, in depth but preliminary test shows it works, including.
24
  - **Agent abilities** I did train this model on agent datasets, that teach it to do real world tasks such as picking up an object, and even navigating a webpage based off HTML.
25
  - **Good Chili Recipe** The model gives a good chili recipe :)
26
- - **16k Sequence Length** This model was trained with a 16k sequence length.
27
 
28
  ### Experimental Nature
29
  Please note that Mistral-22b is still in a WIP. v0.3 has started training now, with a different method than used before, this is to hopefully make the model more round in its internel knowlledge. Through my testing I found V2 to be a significant improvement over v.1.
 
23
  - **Json Mode** I did train this model on answering in JSON and using JSON tools., I have yet to try it, in depth but preliminary test shows it works, including.
24
  - **Agent abilities** I did train this model on agent datasets, that teach it to do real world tasks such as picking up an object, and even navigating a webpage based off HTML.
25
  - **Good Chili Recipe** The model gives a good chili recipe :)
26
+ - **32k Sequence Length** This model was trained with a 32k sequence length.
27
 
28
  ### Experimental Nature
29
  Please note that Mistral-22b is still in a WIP. v0.3 has started training now, with a different method than used before, this is to hopefully make the model more round in its internel knowlledge. Through my testing I found V2 to be a significant improvement over v.1.