no code implementations • 3 Jun 2024 • Weichao Zhao, Hao Feng, Qi Liu, Jingqun Tang, Shu Wei, Binghong Wu, Lei Liao, YongJie Ye, Hao liu, Houqiang Li, Can Huang
In this mechanism, all the involved diverse visual table understanding (VTU) tasks and multi-source visual embeddings are abstracted as concepts.
1 code implementation • 31 May 2024 • Weichao Zhao, Hezhen Hu, Wengang Zhou, Yunyao Mao, Min Wang, Houqiang Li
To this end, we propose a Motion-Aware masked autoencoder with Semantic Alignment (MASA) that integrates rich motion cues and global semantic information in a self-supervised learning paradigm for SLR.
no code implementations • 8 Aug 2023 • Weichao Zhao, Hezhen Hu, Wengang Zhou, Li Li, Houqiang Li
Reconstructing interacting hands from monocular RGB data is a challenging task, as it involves many interfering factors, e. g. self- and mutual occlusion and similar textures.
no code implementations • 8 May 2023 • Hezhen Hu, Weichao Zhao, Wengang Zhou, Houqiang Li
In our framework, the hand pose is regarded as a visual token, which is derived from an off-the-shelf detector.
Ranked #1 on Sign Language Recognition on WLASL
no code implementations • 10 Feb 2023 • Weichao Zhao, Hezhen Hu, Wengang Zhou, Jiaxin Shi, Houqiang Li
In this work, we are dedicated to leveraging the BERT pre-training success and modeling the domain-specific statistics to fertilize the sign language recognition~(SLR) model.
no code implementations • ICCV 2021 • Hezhen Hu, Weichao Zhao, Wengang Zhou, Yuechen Wang, Houqiang Li
To validate the effectiveness of our method on SLR, we perform extensive experiments on four public benchmark datasets, i. e., NMFs-CSL, SLR500, MSASL and WLASL.
Ranked #1 on Sign Language Recognition on WLASL100 (using extra training data)
no code implementations • CVPR 2021 • Hezhen Hu, Weilun Wang, Wengang Zhou, Weichao Zhao, Houqiang Li
Then, a transformation flow is calculated based on the correspondence of the source and target topology map.