AAVE_PoS-Tagger / README.md
DrewG's picture
Update README.md
5e4fefd
metadata
license: apache-2.0
tags:
  - generated_from_trainer
metrics:
  - precision
  - recall
  - f1
  - accuracy
model-index:
  - name: bert-finetuned-ner
    results: []
language:
  - en
pipeline_tag: token-classification

bert-finetuned-AAVE-PoS

This model is a version of bert-base-cased fine-tuned on a dataset of African American Vernacular English (AAVE) which was published alongside Jørgensen et al. 2016. It achieves the following results on the evaluation set:

  • Loss: 0.2582
  • Precision: 0.8632
  • Recall: 0.8730
  • F1: 0.8681
  • Accuracy: 0.9356

Model description

More information needed

Intended uses & limitations

This model is intended to help close the gap in part-of-speech tagging performance between Standard American English (SAE) and African American English (AAVE) which differ liguistically in many well-documented ways. It was fine-tuned on data gathered from Twitter, and is thus ingrained with what linguists call 'register bias'.

Training and evaluation data

Code hosted at GitHub.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 3 (this amount of data overfits on 3+)

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Accuracy
No log 1.0 223 0.2982 0.8196 0.8350 0.8272 0.9216
No log 2.0 446 0.2625 0.8599 0.8680 0.8640 0.9326
0.4647 3.0 669 0.2582 0.8632 0.8730 0.8681 0.9356

Framework versions

  • Transformers 4.29.2
  • Pytorch 1.13.1+cpu
  • Datasets 2.12.0
  • Tokenizers 0.13.3