no code implementations • 9 Jan 2024 • Qinyi Luo, Penghan Wang, Wei zhang, Fan Lai, Jiachen Mao, Xiaohan Wei, Jun Song, Wei-Yu Tsai, Shuai Yang, Yuxi Hu, Xuehai Qian
Huge embedding tables in modern Deep Learning Recommender Models (DLRM) require prohibitively large memory during training and inference.
no code implementations • 11 Jul 2023 • Zixuan Ma, Haojie Wang, Jingze Xing, Liyan Zheng, Chen Zhang, Huanqi Cao, Kezhao Huang, Shizhi Tang, Penghan Wang, Jidong Zhai
To accelerate DNN computation, tensor compilers are proposed to generate efficient code on different domain-specific accelerators.
1 code implementation • 15 Mar 2023 • Canwen Xu, Julian McAuley, Penghan Wang
We present Mirror, an open-source platform for data exploration and analysis powered by large language models.