Suggest an alternative to

hlb-gpt

Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wikitext-103 on a single A100 in <100 seconds. Scales to larger models with one parameter change (feature currently in alpha).

Why do you think that https://github.com/karpathy/randomfun is a good alternative to hlb-gpt

A URL to the alternative repo (e.g. GitHub, GitLab)

Here you can share your experience with the project you are suggesting or its comparison with hlb-gpt. Optional.

A valid email to send you a verification link when necessary or log in.