This project uses a distilledGPT2 encoding and a generic Transformer-based architecture, trained on a portion of a Huggingface database (corbt/all-recipes), to train a 83 M parameter generative model. Use it to find your next great recipe (1% of t…
This project uses a distilledGPT2 encoding and a generic Transformer-based architecture, trained on a portion of a Huggingface database (corbt/all-recipes), to train a 83 M parameter generative model. Use it to find your next great recipe (1% of the time) or your next disaster (99% of the time… hey I don’t have any H100s). Disclaimer: you shouldn’t probably take these recipes seriously unless you’re absolutely sure it’s fine.
agent for boring code, some documentation (models in readme), debugging. but all architectures, examples and core functions are all mine.