Edit model card

Tamil Large Language Model

Developing a large language model for Tamil is a significant achievement, highlighting the potential for advanced AI technologies to support and enrich Tamil digital content. Created by Mervin Praison, this model stands as a testament to innovation and dedication towards enhancing language processing capabilities for Tamil speakers around the globe. It is designed to understand, interpret, and generate Tamil text, offering invaluable tools for businesses, educators, and researchers. By leveraging this model, we can unlock new opportunities for content creation, automate and improve customer support, and contribute to preserving and promoting the Tamil language in the digital age.

Uploaded model

  • Developed by: mervinpraison
  • Instruction Fine Tuned : alpaca tamil dataset

How to use?

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")
model = AutoModelForCausalLM.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")

query_to_llm = "ஆரோக்கியமாக இருப்பதற்கான இரண்டு வழிகள்"
inputs = tokenizer.encode(query_to_llm, return_tensors="pt")
outputs = model.generate(inputs, max_length=200)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Downloads last month
36
Safetensors
Model size
8.54B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mervinpraison/tamil-large-language-model-7b-v1.0

Base model

google/gemma-7b
Finetuned
(85)
this model