--- license: apache-2.0 language: - en --- A modified GPT2 architecture with 25m non-embedding parameters, no biases, embedding-ln, scaled sin position embeddings, and a modification that makes the model's transformer run over the sequence four times before going to the language modelling head. | model | avg | arc | hellaswag | mmlu | truthfulqa | | --- | --- | --- | --- | --- | --- | | horizon-25m-v0 | 30.625 | 20.22 | 26.23 | 25.9 | 50.15 | | cramp-25m | 30.57 | 21.76 | 27.35 | 25.53 | 47.66 | | gpt2 | 30.06 | 22.1 | 31.6 | 25.86 | 40.67 | | pythia 70m deduped | 30.25 | 21.08 | 27.17 | 25.26 | 47.51 | | pythia 70m | 30.46 | 21.59 | 27.29 | 25.9 | 47.06 | | pythia 160m deduped | 31.16 | 24.06 | 30.34 | 24.95 | 44.34 | | pythia 160m | 30.58 | 22.78 | 30.34 | 24.95 | 44.26 | Dataset (Horizon-v0) | Source | Documents | | --- | --- | | arxiv | 8.78k | | github | 8.82k | | books | 10k | | wiki | 14.67k | | openwebtext v2 | 30.73k