File size: 1,574 Bytes
35da623 860a8a8 35da623 4987a8f 36ccb1a 4987a8f 35da623 4987a8f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- gemma
- trl
- sft
base_model: google/gemma-7b
---
# Tamil Large Language Model
Developing a large language model for Tamil is a significant achievement, highlighting the potential for advanced AI technologies to support and enrich Tamil digital content. Created by Mervin Praison, this model stands as a testament to innovation and dedication towards enhancing language processing capabilities for Tamil speakers around the globe. It is designed to understand, interpret, and generate Tamil text, offering invaluable tools for businesses, educators, and researchers. By leveraging this model, we can unlock new opportunities for content creation, automate and improve customer support, and contribute to preserving and promoting the Tamil language in the digital age.
## Uploaded model
- **Developed by:** mervinpraison
- **Instruction Fine Tuned :** alpaca tamil dataset
## How to use?
```
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")
model = AutoModelForCausalLM.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")
query_to_llm = "ஆரோக்கியமாக இருப்பதற்கான இரண்டு வழிகள்"
inputs = tokenizer.encode(query_to_llm, return_tensors="pt")
outputs = model.generate(inputs, max_length=200)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
|