Edit model card

Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF

This model was converted to GGUF format from Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

Not Officially Benchmarked Yet! (Please submit any benchmarking or eval results via the Community tab.)

Uploaded model

  • Developed by: Solshine (Caleb DeLeeuw)
  • License: LLama 3.1 License
  • Finetuned from model : Solshine/reflection-llama-3.1-8B-Solshine-trainround2-16bit

Inspired by and featuring the Reflection Tuning technique pioneered by Matt Shumer (possibly earlier innovated by the team at Anthropic, and Mlabbone's Hermes.)

To the authors' knowledge, this is V3 of the first "reflection tuned" Llama 3.1 8B LLM

As per the inspiring model "mattshumer/Reflection-Llama-3.1-70B" (this model was not used in the training process nor as a foundational model, but only served as inspiration) :


During sampling, the model will start by outputting reasoning inside <thinking> and </thinking> tags, and then once it is satisfied with its reasoning, it will output the final answer inside <output> and </output> tags. Each of these tags are special tokens, trained into the model.

This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.

Inside the <thinking> section, the model may output one or more <reflection> tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.

System Prompt:
The system prompt used for training this model is:

You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.

We recommend using this exact system prompt to get the best results from Reflection Llama-3.1 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.

Chat Format:
As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:

<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>

what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux)

brew install llama.cpp

Invoke the llama.cpp server or the CLI.

CLI:

llama-cli --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -p "The meaning to life and the universe is"

Server:

llama-server --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -c 2048

Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well.

Step 1: Clone llama.cpp from GitHub.

git clone https://github.com/ggerganov/llama.cpp

Step 2: Move into the llama.cpp folder and build it with LLAMA_CURL=1 flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).

cd llama.cpp && LLAMA_CURL=1 make

Step 3: Run inference through the main binary.

./llama-cli --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -p "The meaning to life and the universe is"

or

./llama-server --hf-repo Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF --hf-file reflection-llama-3.1-8b-solshine-trainround3-16bit-q4_k_m.gguf -c 2048
Downloads last month
64
GGUF
Model size
8.03B params
Architecture
llama

4-bit

Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF

Datasets used to train Solshine/reflection-llama-3.1-8B-Solshine-trainround3-16bit-Q4_K_M-GGUF