1 code implementation • 14 Dec 2023 • Wenyi Hong, Weihan Wang, Qingsong Lv, Jiazheng Xu, Wenmeng Yu, Junhui Ji, Yan Wang, Zihan Wang, Yuxuan Zhang, Juanzi Li, Bin Xu, Yuxiao Dong, Ming Ding, Jie Tang
People are spending an enormous amount of time on digital devices through graphical user interfaces (GUIs), e. g., computer or smartphone screens.
Ranked #15 on Visual Question Answering on MM-Vet
1 code implementation • 6 Nov 2023 • Weihan Wang, Qingsong Lv, Wenmeng Yu, Wenyi Hong, Ji Qi, Yan Wang, Junhui Ji, Zhuoyi Yang, Lei Zhao, Xixuan Song, Jiazheng Xu, Bin Xu, Juanzi Li, Yuxiao Dong, Ming Ding, Jie Tang
We introduce CogVLM, a powerful open-source visual language foundation model.
Ranked #4 on Visual Question Answering (VQA) on InfiMM-Eval
3 code implementations • ACL 2022 • Huisheng Mao, Ziqi Yuan, Hua Xu, Wenmeng Yu, Yihe Liu, Kai Gao
The platform features a fully modular video sentiment analysis framework consisting of data management, feature extraction, model training, and result analysis modules.
2 code implementations • 9 Feb 2021 • Wenmeng Yu, Hua Xu, Ziqi Yuan, Jiele Wu
On MOSI and MOSEI datasets, our method surpasses the current state-of-the-art methods.
1 code implementation • ACL 2020 • Wenmeng Yu, Hua Xu, Fanyang Meng, Yilin Zhu, Yixiao Ma, Jiele Wu, Jiyun Zou, Kai-Cheng Yang
Previous studies in multimodal sentiment analysis have used limited datasets, which only contain unified multimodal annotations.