GreekDeBERTaV3-base

GreekDeBERTaV3-base is a language model pre-trained specifically for Greek NLP tasks. It is based on the DeBERTaV3 architecture, incorporating improvements from the replaced token detection (RTD) task during pre-training.

Model Overview

  • Model Architecture: DeBERTaV3-base
  • Language: Greek
  • Pre-training Tasks: Replaced Token Detection (RTD)
  • Tokenizer: SentencePiece Model (spm.model)

This model was trained on a diverse corpus of Greek texts and is suitable for tasks like Part-of-Speech tagging, Named Entity Recognition, and Natural Language Inference.

Files

  • config.json: Configuration file for the model.
  • pytorch_model.bin: The PyTorch weights of the model.
  • spm.model: The SentencePiece tokenizer model.
  • vocab.txt: A human-readable vocabulary file that contains the list of tokens used by the model.
  • tokenizer_config.json: Tokenizer configuration file.

How to Use

You can use this model with the Hugging Face transformers library:

from transformers import AutoTokenizer, AutoModelForTokenClassification

tokenizer = AutoTokenizer.from_pretrained("AI-team-UoA/GreekDeBERTaV3-base")
model = AutoModelForTokenClassification.from_pretrained("AI-team-UoA/GreekDeBERTaV3-base")
Downloads last month
8
Inference Examples
Unable to determine this model's library. Check the docs .