Impossible language models trained from scratch with a GPT-2 Small architecture that lacks positional encodings.
-
mission-impossible-lms/no-shuffle-gpt2-no-pos
0.1B • Updated • 8 -
mission-impossible-lms/nondeterministic-shuffle-gpt2-no-pos
0.1B • Updated • 7 -
mission-impossible-lms/deterministic-shuffle-s21-gpt2-no-pos
0.1B • Updated • 10 -
mission-impossible-lms/deterministic-shuffle-s57-gpt2-no-pos
0.1B • Updated • 9