You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip

Model overview

Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip is a fine-tuned version of the Llama 3.1 model, specifically optimized for Albanian language tasks. This model is tailored to perform a variety of natural language processing tasks in Albanian, utilizing a quantized 4-bit precision to maintain efficiency and scalability while supporting extensive inference tasks.

Model Details

  • Model Name: Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip
  • Base Model: Llama 3.1
  • Model Size: 8 billion parameters
  • Quantization: 4-bit precision (bnb)
  • Language: Albanian
  • License: llama3.1

Limitations

  • Representation of Harms & Stereotypes: Potential for biased outputs reflecting real-world societal biases.
  • Inappropriate or Offensive Content: Risk of generating content that may be offensive or inappropriate in certain contexts.
  • Information Reliability: Possibility of producing inaccurate or outdated information.
  • Dataset Size: The Albanian dataset used for fine-tuning was not very large, which may affect the model's performance and coverage.

Intended Use

  • Intended Use Cases: This model is suitable for various NLP tasks in Albanian, including conversational AI, text generation, and language understanding.
  • Out-of-scope Use: This model should not be used in ways that violate laws, regulations, or ethical guidelines. It is also not intended for use in languages other than Albanian unless appropriately fine-tuned.

Responsible AI Considerations

Developers using this model should:

  • Evaluate and mitigate risks related to accuracy, safety, and fairness.
  • Ensure compliance with applicable laws and regulations.
  • Implement additional safeguards for high-risk scenarios and sensitive contexts.
  • Inform end-users that they are interacting with an AI system.
  • Use feedback mechanisms and contextual information grounding techniques (RAG) to enhance output reliability.
!pip3 install -U transformers peft accelerate bitsandbytes

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
import torch

hf_token = "hf_...."

torch.random.manual_seed(0)

model = AutoModelForCausalLM.from_pretrained(
    "Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip",
    device_map="cuda",
    torch_dtype="auto",
    trust_remote_code=True,
    token=hf_token,
)

tokenizer = AutoTokenizer.from_pretrained("Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip", token=hf_token)

messages = [
    {"role": "system", "content": "Je një asistent inteligjent shumë i dobishëm."},
    {"role": "user", "content": "Identifiko emrat e personave në këtë artikull 'Majlinda Kelmendi (lindi më 9 maj 1991), është një xhudiste shqiptare nga Peja, Kosovë.'"},
]

pipe = pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer,
)

generation_args = {
    "max_new_tokens": 2048,
    "return_full_text": False,
    "temperature": 0.9,
    "do_sample": True,
}

prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=False)
output = pipe(prompt, **generation_args)
print(output[0]['generated_text'])

Acknowledgements

This model is built upon the Meta-Llama-3.1-8B-Instruct by leveraging its robust capabilities and further fine-tuning it for Albanian language tasks. Special thanks to the developers and researchers who contributed to the original Llama3.1.

Downloads last month
19
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip

Quantizations
1 model

Dataset used to train Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip

Collection including Kushtrim/Llama-3.1-8B-Instruct-bnb-4bit-shqip