π«π· Calme-2
Collection
New Calme-2 fine-tuned models
β’
30 items
β’
Updated
β’
4
This model is a fine-tuned version of the dnhkng/RYS-XLarge
, pushing the boundaries of natural language understanding and generation even further. My goal was to create a versatile and robust model that excels across a wide range of benchmarks and real-world applications.
This model is suitable for a wide range of applications, including but not limited to:
Coming soon!
Coming soon!
This model uses ChatML
prompt template:
<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}
# Use a pipeline as a high-level helper
from transformers import pipeline
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.3-rys-78b")
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.3-rys-78b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.3-rys-78b")
As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.