Model Details

Granite-3.1-2B-instruct-KR-Summarization

Granite-3.1-2B-instruct-KR-Summarization is continued pretrained(fully fine-tuned) language model based on Granite-3.1-2B-Instruct.

This model is trained fully with publicily available resource at HuggingFace dataset hub, preprocessed Korean texts.

The train was done on A6000 48GB * 4.

Model developers Dongwook Min (mindw96)

Variations Granite-3.1-2B-instruct-KR-Summarization comes in one size — 2B.

Input Models input text only.

Output Models generate text only.

Model Architecture Granite 3.1 is an auto-regressive language model that uses an optimized transformer architecture.

Model Release Date 02.01.2025.

Capabilities

  • Summarization
Downloads last month
4
Safetensors
Model size
2.53B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mindw96/granite-3.1-2b-instruct-korean-summarization

Finetuned
(2)
this model

Dataset used to train mindw96/granite-3.1-2b-instruct-korean-summarization