no code implementations • LT4HALA (LREC) 2022 • Hailin Zhang, Ziyu Yang, Yingwen Fu, Ruoyao Ding
In addition, we perform a series of training strategies based on the provided ancient Chinese pre-trained model to enhance the model performance.
Chinese Word Segmentation Cultural Vocal Bursts Intensity Prediction +5
2 code implementations • 29 Feb 2024 • Penghao Zhao, Hailin Zhang, Qinhan Yu, Zhengren Wang, Yunteng Geng, Fangcheng Fu, Ling Yang, Wentao Zhang, Jie Jiang, Bin Cui
We first classify RAG foundations according to how the retriever augments the generator, distilling the fundamental abstractions of the augmentation methodologies for various retrievers and generators.
1 code implementation • 6 Dec 2023 • Hailin Zhang, Zirui Liu, Boxuan Chen, Yikai Zhao, Tong Zhao, Tong Yang, Bin Cui
Guided by our design philosophy, we further propose a multi-level hash embedding framework to optimize the embedding tables of non-hot features.
1 code implementation • 27 Nov 2023 • Hailin Zhang, Penghao Zhao, Xupeng Miao, Yingxia Shao, Zirui Liu, Tong Yang, Bin Cui
Learnable embedding vector is one of the most important applications in machine learning, and is widely used in various database-related domains.
1 code implementation • NeurIPS 2023 • Hailin Zhang, Yujing Wang, Qi Chen, Ruiheng Chang, Ting Zhang, Ziming Miao, Yingyan Hou, Yang Ding, Xupeng Miao, Haonan Wang, Bochen Pang, Yuefeng Zhan, Hao Sun, Weiwei Deng, Qi Zhang, Fan Yang, Xing Xie, Mao Yang, Bin Cui
We empirically show that our model achieves better performance on the commonly used academic benchmarks MSMARCO Passage and Natural Questions, with comparable serving latency to dense retrieval solutions.
1 code implementation • 11 Jun 2023 • Hailin Zhang, Defang Chen, Can Wang
Multi-Teacher knowledge distillation provides students with additional supervision from multiple pre-trained teachers with diverse information sources.
2 code implementations • 25 Nov 2022 • Xupeng Miao, Yujie Wang, Youhe Jiang, Chunan Shi, Xiaonan Nie, Hailin Zhang, Bin Cui
Transformer models have achieved state-of-the-art performance on various domains of applications and gradually becomes the foundations of the advanced large deep learning (DL) models.
1 code implementation • CVPR 2022 • Defang Chen, Jian-Ping Mei, Hailin Zhang, Can Wang, Yan Feng, Chun Chen
Knowledge distillation aims to compress a powerful yet cumbersome teacher model into a lightweight student model without much sacrifice of performance.
Ranked #3 on Knowledge Distillation on CIFAR-100
1 code implementation • 30 Dec 2021 • Hailin Zhang, Defang Chen, Can Wang
Knowledge distillation is initially introduced to utilize additional supervision from a single teacher model for the student model training.
3 code implementations • 14 Dec 2021 • Xupeng Miao, Hailin Zhang, Yining Shi, Xiaonan Nie, Zhi Yang, Yangyu Tao, Bin Cui
Embedding models have been an effective learning paradigm for high-dimensional data.
no code implementations • 22 Aug 2019 • Man Qi, Niv DeMalach, Tao Sun, Hailin Zhang
Thus, we developed an extension of resource competition theory to investigate partial and total preemption (in the latter, the preemptor is unaffected by species with lower preemption rank).