no code implementations • 14 May 2024 • Yashuai Yan, Esteve Valls Mascaro, Tobias Egle, Dongheui Lee
This paper addresses the critical need for refining robot motions that, despite achieving a high visual similarity through human-to-humanoid retargeting methods, fall short of practical execution in the physical realm.
no code implementations • 7 Feb 2024 • Esteve Valls Mascaro, Yashuai Yan, Dongheui Lee
Integrating robots into populated environments is a complex challenge that requires an understanding of human social dynamics.
no code implementations • 28 Sep 2023 • Esteve Valls Mascaro, Daniel Sliwowski, Dongheui Lee
In this paper, we propose a Human-Object Interaction (HOI) anticipation framework for collaborative robots.
Ranked #1 on Human-Object Interaction Anticipation on VidHOI
Human-Object Interaction Anticipation Human-Object Interaction Detection
no code implementations • 11 Sep 2023 • Yashuai Yan, Esteve Valls Mascaro, Dongheui Lee
Additionally, we propose a consistency term to build a common latent space that captures the similarity of the poses with precision while allowing direct robot motion control from the latent space.
no code implementations • 14 Aug 2023 • Esteve Valls Mascaro, Hyemin Ahn, Dongheui Lee
Experimental results show that our model successfully forecasts human motion on the Human3. 6M dataset.
no code implementations • 28 Feb 2023 • Hyemin Ahn, Esteve Valls Mascaro, Dongheui Lee
After many researchers observed fruitfulness from the recent diffusion probabilistic model, its effectiveness in image generation is actively studied these days.
no code implementations • 16 Feb 2023 • Esteve Valls Mascaro, Shuo Ma, Hyemin Ahn, Dongheui Lee
In addition, our model is tested in conditions where the human motion is severely occluded, demonstrating its robustness in reconstructing and predicting 3D human motion in a highly noisy environment.
1 code implementation • 25 Jul 2022 • Esteve Valls Mascaro, Hyemin Ahn, Dongheui Lee
Our framework first extracts two level of human information over the N observed videos human actions through a Hierarchical Multi-task MLP Mixer (H3M).
Ranked #1 on Long Term Action Anticipation on Ego4D