no code implementations • 8 Jun 2023 • Leilei Wang
Owing to the impressive dot-product attention, the Transformers have been the dominant architectures in various natural language processing (NLP) tasks.
no code implementations • 27 Nov 2022 • Yu Guo, Zhilong Xie, Xingyan Chen, Huangen Chen, Leilei Wang, Huaming Du, Shaopeng Wei, Yu Zhao, Qing Li, Gang Wu
We address the problem by introducing a novel joint method on top of BERT which explicitly models the multiple sub-tokens features after wordpiece tokenization, thereby contributing to the two tasks.