InferenceIllusionist's picture
Update README.md
50307b6 verified
|
raw
history blame
2.36 kB
metadata
tags:
  - gguf
license: cc-by-nc-4.0

Model Card for Fimbulvetr-11B-v2-iMat-GGUF

Looking for Q3/Q4/Q5 quants? See the link in the model card below.

All credits to Sao10K for the original model. This is just a quick test of the new quantization types such as IQ_3S in an attempt to further reduce VRAM requirements.

Quantized from fp16 with love. Importance matrix file Fimbulvetr-11B-v2-imatrix.dat was calculated using Q8_0.

Please Note: Inferencing for newer formats (IQ3_S, IQ4_NL, etc) was tested on llama.cpp. These newer quants may not work with other methods yet (as of 2/25/24).

Original model card details below.


Fox1

Cute girl to catch your attention.

https://huggingface.co/Sao10K/Fimbulvetr-11B-v2-GGUF <------ GGUF

Fimbulvetr-v2 - A Solar-Based Model

Prompt Formats - Alpaca or Vicuna. Either one works fine. Recommended SillyTavern Presets - Universal Light

Alpaca:

### Instruction:
<Prompt>
### Input:
<Insert Context Here>
### Response:

Vicuna:

System: <Prompt>

User: <Input>

Assistant:

Changelogs:

25/2 - repo renamed to remove test, model card redone. Model's officially out.
15/2 - Heavy testing complete. Good feedback.


Rant - Kept For Historical Reasons

Ramble to meet minimum length requirements:

Tbh i wonder if this shit is even worth doing. Like im just some broke guy lmao I've spent so much. And for what? I guess creds. Feels good when a model gets good feedback, but it seems like im invisible sometimes. I should be probably advertising myself and my models on other places but I rarely have the time to. Probably just internal jealousy sparking up here and now. Wahtever I guess.

Anyway cool EMT vocation I'm doing is cool except it pays peanuts, damn bruh 1.1k per month lmao. Government to broke to pay for shit. Pays the bills I suppose.

Anyway cool beans, I'm either going to continue the Solar Train or go to Mixtral / Yi when I get paid.

You still here?