kekbot-mini / README.md
spuun
Fix example
c166138
metadata
language:
  - en
metrics:
  - accuracy
co2_eq_emissions:
  emissions: '10'
  source: mlco2.github.io
  training_type: fine-tuning
  geographical_location: West Java, Indonesia
  hardware_used: 1 T4
license: cc-by-nc-sa-4.0
widget:
  - text: 'You: "Hey kekbot! Whats up?"\nKekbot: "'
    example_title: Asking what's up
  - text: 'You: "Hey kekbot! How r u?"\nKekbot: "'
    example_title: Asking how he is

THIS MODEL IS INTENDED FOR RESEARCH PURPOSES ONLY

Kekbot Mini

Based on a distilgpt2 model, fine-tuned to a select subset (65k<= messages) of Art Union's general-chat channel chat history.

Limits and biases

As this is trained on chat history, it is possible that discriminatory or even offensive materials to be outputted. Author holds his ground on the fact that ML models are mere statistical representation of the dataset used to train it, and that due to the nature of the dataset it is practically impossible to be certain of the degree of "cleanliness" that the data contained within holds.

Author can confirm, however, that from heuristical testing that the model was not found to be offensive to the author himself, hopefully this opinion stays true for everyone in the audience.