1 code implementation • 1 Apr 2024 • Miaoran Zhang, Mingyang Wang, Jesujoba O. Alabi, Dietrich Klakow
This paper presents our system developed for the SemEval-2024 Task 1: Semantic Textual Relatedness for African and Asian Languages.
no code implementations • 31 Mar 2024 • Mingyang Wang, Heike Adel, Lukas Lange, Jannik Strötgen, Hinrich Schütze
Continual learning aims at incrementally acquiring new knowledge while not forgetting existing knowledge.
no code implementations • 15 Mar 2024 • Odin Zhang, Yufei Huang, Shichen Cheng, Mengyao Yu, Xujun Zhang, Haitao Lin, Yundian Zeng, Mingyang Wang, Zhenxing Wu, Huifeng Zhao, Zaixi Zhang, Chenqing Hua, Yu Kang, Sunliang Cui, Peichen Pan, Chang-Yu Hsieh, Tingjun Hou
Most earlier 3D structure-based molecular generation approaches follow an atom-wise paradigm, incrementally adding atoms to a partially built molecular fragment within protein pockets.
no code implementations • 20 Feb 2024 • Miaoran Zhang, Vagrant Gautam, Mingyang Wang, Jesujoba O. Alabi, Xiaoyu Shen, Dietrich Klakow, Marius Mosbach
Compared to work on monolingual (English) in-context learning, multilingual in-context learning is under-explored, and we lack an in-depth understanding of the role of demonstrations in this context.
1 code implementation • 15 Nov 2023 • Yihong Liu, Peiqin Lin, Mingyang Wang, Hinrich Schütze
Instead of pretraining multilingual language models from scratch, a more efficient method is to adapt existing pretrained language models (PLMs) to new languages via vocabulary extension and continued pretraining.
no code implementations • 23 Oct 2023 • Zihao Yan, Fubao Su, Mingyang Wang, Ruizhen Hu, Hao Zhang, Hui Huang
We introduce an active 3D reconstruction method which integrates visual perception, robot-object interaction, and 3D scanning to recover both the exterior and interior, i. e., unexposed, geometries of a target 3D object.
no code implementations • 23 Oct 2023 • Mingyang Wang, Heike Adel, Lukas Lange, Jannik Strötgen, Hinrich Schütze
However, not all languages positively influence each other and it is an open research question how to select the most suitable set of languages for multilingual training and avoid negative interference among languages whose characteristics or data distributions are not compatible.
no code implementations • 29 Apr 2023 • Mingyang Wang, Zhenshan Bing, Xiangtong Yao, Shuai Wang, Hang Su, Chenguang Yang, Kai Huang, Alois Knoll
On MuJoCo and Meta-World benchmarks, MoSS outperforms prior works in terms of asymptotic performance, sample efficiency (3-50x faster), adaptation efficiency, and generalization robustness on broad and diverse task distributions.
no code implementations • 28 Apr 2023 • Mingyang Wang, Heike Adel, Lukas Lange, Jannik Strötgen, Hinrich Schütze
In this work, we propose to leverage language-adaptive and task-adaptive pretraining on African texts and study transfer learning with source language selection on top of an African language-centric pretrained language model.
no code implementations • 4 Jun 2020 • Xiang Li, Mingyang Wang, Yi Fang
Previous researches have extensively studied the problem of height estimation from aerial images based on stereo or multi-view image matching.
no code implementations • 14 Oct 2019 • Xiang Li, Mingyang Wang, Congcong Wen, Lingjing Wang, Nan Zhou, Yi Fang
Based on this convolution module, we further developed a multi-scale fully convolutional neural network with downsampling and upsampling blocks to enable hierarchical point feature learning.