adding note about the LlamaTokenizerFast is not included in this build so the Inference API will not work. please use the LlamaTokenizerFast from: Doctor-Shotgun/TinyLlama-1.1B-32k-Instruct to use this model at this time to the README
Browse files
README.md
CHANGED
@@ -12,6 +12,14 @@ This model was merged with the following HuggingFace TinyLlama models using ties
|
|
12 |
- Tensoic/TinyLlama-1.1B-3T-openhermes
|
13 |
- Josephgflowers/TinyLlama-3T-Cinder-v1.3
|
14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
## How do I fine-tune this model?
|
16 |
|
17 |
Please refer to the Unsloth fine-tuning guide for:
|
@@ -20,8 +28,6 @@ Please refer to the Unsloth fine-tuning guide for:
|
|
20 |
|
21 |
## How do I generate my own model merges?
|
22 |
|
23 |
-
Here's [the standalone python script](https://huggingface.co/matlok/tinyllama-cinder-openhermes-32k/blob/main/run-tiny-merge.py) used with logs below:
|
24 |
-
|
25 |
```python3
|
26 |
#!/usr/bin/env python3
|
27 |
|
|
|
12 |
- Tensoic/TinyLlama-1.1B-3T-openhermes
|
13 |
- Josephgflowers/TinyLlama-3T-Cinder-v1.3
|
14 |
|
15 |
+
## Why does the Inference API on HuggingFace not work for this merged model?
|
16 |
+
|
17 |
+
The included [merge python script](https://huggingface.co/matlok/tinyllama-cinder-openhermes-32k/blob/main/run-tiny-merge.py) does not contain the **LlamaTokenizerFast** tokenizer. This means the HuggingFace Inference API will not work. The tokenizer to use with this model is:
|
18 |
+
|
19 |
+
```
|
20 |
+
TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
21 |
+
```
|
22 |
+
|
23 |
## How do I fine-tune this model?
|
24 |
|
25 |
Please refer to the Unsloth fine-tuning guide for:
|
|
|
28 |
|
29 |
## How do I generate my own model merges?
|
30 |
|
|
|
|
|
31 |
```python3
|
32 |
#!/usr/bin/env python3
|
33 |
|