Llama-Express.1 / README.md
prithivMLmods's picture
Update README.md
f74da0d verified
---
license: llama3.2
datasets:
- AI-MO/NuminaMath-CoT
- prithivMLmods/Math-Solve
- amphora/QwQ-LongCoT-130K
- prithivMLmods/Deepthink-Reasoning
language:
- en
base_model:
- meta-llama/Llama-3.2-1B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- Express
- Llama
- Ollama
- v.1
- text-generation-inference
---
# **Llama-Express.1**
Llama-Express.1 is a 1B model based on Llama 3.2 (1B), fine-tuned on long chain-of-thought datasets. This instruction-tuned, text-only model is optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. It outperforms many of the available open-source and closed chat models.
# **Use with transformers**
Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.
Make sure to update your transformers installation via `pip install --upgrade transformers`.
```python
import torch
from transformers import pipeline
model_id = "prithivMLmods/Llama-Express.1"
pipe = pipeline(
"text-generation",
model=model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
messages = [
{"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
{"role": "user", "content": "Who are you?"},
]
outputs = pipe(
messages,
max_new_tokens=256,
)
print(outputs[0]["generated_text"][-1])
```
# **Intended Use**
1. **Multilingual Dialogue**:
- Designed for high-quality, multilingual conversations, making it suitable for applications requiring natural, fluid dialogue across languages.
2. **Agentic Retrieval**:
- Optimized for retrieval-based tasks where reasoning and contextual chaining are crucial for extracting and summarizing relevant information.
3. **Summarization Tasks**:
- Effective in generating concise and accurate summaries from complex and lengthy texts, suitable for academic, professional, and casual use cases.
4. **Instruction-Following Applications**:
- Fine-tuned for tasks requiring adherence to user-provided instructions, making it ideal for automation workflows, content creation, and virtual assistant integrations.
# **Limitations**
1. **Monomodal Focus**:
- As a text-only model, it cannot process multimodal inputs like images, audio, or videos, limiting its versatility in multimedia applications.
2. **Context Length Constraints**:
- While optimized for long chain-of-thought reasoning, extreme cases with very large contexts may still lead to degraded performance or truncation issues.
3. **Bias and Ethics**:
- The model might reflect biases present in the training datasets, potentially resulting in outputs that could be culturally insensitive or inappropriate.
4. **Performance in Low-Resource Languages**:
- While multilingual, its effectiveness may vary across languages, with possible performance drops in underrepresented or low-resource languages.
5. **Dependency on Input Quality**:
- The model's output is heavily influenced by the clarity and specificity of the input instructions. Ambiguous or vague prompts may lead to suboptimal results.
6. **Lack of Real-Time Internet Access**:
- Without real-time retrieval capabilities, it cannot provide up-to-date information or verify facts against the latest data.