File size: 2,753 Bytes
88d33aa
 
 
 
 
 
 
 
 
363f579
caef1b7
 
 
88d33aa
 
 
451688b
 
 
88d33aa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
db9c3c4
 
 
 
 
 
 
 
88d33aa
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
license: llama3.1
pipeline_tag: text-generation
---

**Llama3.1-Typhoon2-70B**: Thai Large Language Model (Instruct)

**Llama3.1-Typhoon2-70B** is a pretrained only Thai ๐Ÿ‡น๐Ÿ‡ญ large language model with 70 billion parameters, and it is based on Llama3.1-70B.

For technical-report. please see our [arxiv](https://arxiv.org/abs/2412.13702).
*To acknowledge Meta's effort in creating the foundation model and to comply with the license, we explicitly include "llama-3.1" in the model name.

## **Performance**

| Model                   | ThaiExam  | ONET     | IC        | A-Level   | TGAT      | TPAT      | M3Exam    | Math       | Science    | Social     | Thai       |
|--------------------------|-----------|----------|-----------|-----------|-----------|-----------|-----------|------------|------------|------------|------------|
| **Typhoon1.5x 70B instruct**     | 62.96%    | 60.49%   | **71.57%** | 53.54%    | **72.30%** | **56.89%** | **62.54%** | **45.70%** | 62.56%     | 77.73%     | 64.19%     |
| **Llama3.1 70B**         | 60.74%    | 62.34%   | 67.36%    | 53.54%    | 66.15%    | 54.31%    | 60.35%    | 38.91%     | 62.56%     | 76.99%     | 62.96%     |
| **Typhoon2 Llama 3.1 70B base** | **63.39%** | **65.43%** | 69.47%    | **59.84%** | 66.15%    | 56.03%    | 62.33%    | 42.98%     | **63.28%** | **78.60%** | **64.47%** |


## **Model Description**

- **Model type**: A 70B decoder-only model based on Llama architecture.
- **Requirement**: transformers 4.45.0 or newer.
- **Primary Language(s)**: Thai ๐Ÿ‡น๐Ÿ‡ญ and English ๐Ÿ‡ฌ๐Ÿ‡ง
- **License**: [Llama 3.1 Community License](https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE)


## **Intended Uses & Limitations**

This model is a pretrained base model. Thus, it may not be able to follow human instructions without using one/few-shot learning or instruction fine-tuning. The model does not have any moderation mechanisms, and may generate harmful or inappropriate responses.

## **Follow us**

**https://twitter.com/opentyphoon**

## **Support**

**https://discord.gg/CqyBscMFpg**

## **Citation**

- If you find Typhoon2 useful for your work, please cite it using:
```
@misc{typhoon2,
      title={Typhoon 2: A Family of Open Text and Multimodal Thai Large Language Models}, 
      author={Kunat Pipatanakul and Potsawee Manakul and Natapong Nitarach and Warit Sirichotedumrong and Surapon Nonesung and Teetouch Jaknamon and Parinthapat Pengpun and Pittawat Taveekitworachai and Adisai Na-Thalang and Sittipong Sripaisarnmongkol and Krisanapong Jirayoot and Kasima Tharnpipitchai},
      year={2024},
      eprint={2412.13702},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2412.13702}, 
}
```