Search Results for author: Wonkyo Choe

Found 1 papers, 0 papers with code

STI: Turbocharge NLP Inference at the Edge via Elastic Pipelining

no code implementations11 Jul 2022 Liwei Guo, Wonkyo Choe, Felix Xiaozhu Lin

Yet, the unprecedented size of an NLP model stresses both latency and memory, creating a tension between the two key resources of a mobile device.

Management

Cannot find the paper you are looking for? You can Submit a new open access paper.