no code implementations • 1 Apr 2024 • Achintya Kundu, Fabian Lim, Aaron Chew, Laura Wynter, Penny Chong, Rhui Dih Lee
Supernet training of LLMs is of great interest in industrial applications as it confers the ability to produce a palette of smaller models at constant cost, regardless of the number of models (of different size / latency) produced.
no code implementations • 27 Mar 2023 • Achintya Kundu, Laura Wynter, Rhui Dih Lee, Luis Angel Bathen
Hence, we propose Transfer-Once-For-All (TOFA) for supernet-style training on small data sets with constant computational training cost over any number of edge deployment scenarios.