File size: 914 Bytes
c7042da
 
 
 
b53dbcc
7e69d1e
 
c7042da
 
 
bca8177
14d991c
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
---
license: cc
---
# Bert2Bert (Encoder-Decoder) on Liputan6 100k dataset
Dataset source: https://huggingface.co/datasets/fajrikoto/id_liputan6 <br>
Model used for Fine Tuning (Encoder-Decoder):<br>
https://huggingface.co/cahya/bert-base-indonesian-1.5G <br><br>
Trained on 1x3090 @ 8 epoch (EarlyStopping Callbacks)

Train logs, metrics, and params: https://wandb.ai/willy030125/huggingface/runs/sb2kcuck <br>
https://www.comet.com/willy030125/huggingface/5dd7c19d0c85472abdf4136529f4322c <br>
Eval results and Perplexity: <a href="https://huggingface.co/Willy030125/Bert2Bert_Liputan6_100k_8epoch/blob/main/eval_results.json">eval_results.json</a><br>

Usage:
```python
from transformers import AutoTokenizer, EncoderDecoderModel
tokenizer = AutoTokenizer.from_pretrained("Willy030125/Bert2Bert_Liputan6_100k_8epoch")
model = EncoderDecoderModel.from_pretrained("Willy030125/Bert2Bert_Liputan6_100k_8epoch")
```