|
--- |
|
license: unknown |
|
datasets: |
|
- KorQuAD/squad_kor_v1 |
|
language: |
|
- ko |
|
base_model: |
|
- CurtisJeon/klue-roberta-large-korquad_v1_qa |
|
pipeline_tag: question-answering |
|
--- |
|
|
|
# KLUE RoBERTa Large KorQuAD v1 QA - Fine-tuned |
|
|
|
์ด ๋ชจ๋ธ์ [CurtisJeon/klue-roberta-large-korquad_v1_qa](https://huggingface.co/CurtisJeon/klue-roberta-large-korquad_v1_qa)๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ํ์ฌ ์ถ๊ฐ ๋ฐ์ดํฐ๋ก fine-tuningํ ํ๊ตญ์ด ์ง์์๋ต(QA) ๋ชจ๋ธ์
๋๋ค. |
|
|
|
## ๋ชจ๋ธ ์ ๋ณด |
|
|
|
- ๊ธฐ๋ณธ ๋ชจ๋ธ: KLUE RoBERTa Large |
|
- ํ์คํฌ: ์ง์์๋ต (Question Answering) |
|
- ์ธ์ด: ํ๊ตญ์ด |
|
- ํ๋ จ ๋ฐ์ดํฐ: KorQuAD v1 + [์์ฒด ๋ฐ์ดํฐ] |
|
|
|
## ๋ชจ๋ธ ๊ตฌ์กฐ |
|
|
|
- RobertaForQuestionAnswering ์ํคํ
์ฒ ์ฌ์ฉ + CNN ๋ ์ด์ด(without a dropout) |
|
- 24๊ฐ์ hidden layers |
|
- 1024 hidden size |
|
- 16 attention heads |
|
- ์ด ํ๋ผ๋ฏธํฐ: ์ฝ 355M |
|
|
|
|
|
## ์ฌ์ฉ ๋ฐฉ๋ฒ |
|
|
|
์ด ๋ชจ๋ธ์ Hugging Face Transformers ๋ผ์ด๋ธ๋ฌ๋ฆฌ๋ฅผ ์ฌ์ฉํ์ฌ ์ฝ๊ฒ ๋ก๋ํ๊ณ ์ฌ์ฉํ ์ ์์ต๋๋ค: |
|
|
|
```python |
|
from transformers import AutoModelForQuestionAnswering, AutoTokenizer |
|
|
|
model_name = "HANTAEK/klue-roberta-large-korquad-v1-qa-finetuned" |
|
model = AutoModelForQuestionAnswering.from_pretrained(model_name) |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |