no code implementations • 4 Dec 2023 • Yao-Chih Lee, Zhoutong Zhang, Kevin Blackburn-Matzen, Simon Niklaus, Jianming Zhang, Jia-Bin Huang, Feng Liu
Specifically, we build a global static scene model using an extended plane-based scene representation to synthesize temporally coherent novel video.
no code implementations • 16 Feb 2023 • Ting-Hsuan Liao, Songwei Ge, Yiran Xu, Yao-Chih Lee, Badour AlBahar, Jia-Bin Huang
There has been tremendous progress in large-scale text-to-image synthesis driven by diffusion models enabling versatile downstream applications such as 3D object synthesis from texts, image editing, and customized generation.
no code implementations • CVPR 2023 • Yao-Chih Lee, Ji-Ze Genevieve Jang, Yi-Ting Chen, Elizabeth Qiu, Jia-Bin Huang
Temporal consistency is essential for video editing applications.
1 code implementation • 4 Aug 2022 • Yao-Chih Lee, Kuan-Wei Tseng, Guan-Sheng Chen, Chu-Song Chen
It can improve the robustness of learning-based methods with flow-guided keyframes and well-established depth prior.
1 code implementation • 22 Jun 2021 • Hau Chu, Jia-Hong Lee, Yao-Chih Lee, Ching-Hsien Hsu, Jia-Da Li, Chu-Song Chen
This paper introduces an approach for multi-human 3D pose estimation and tracking based on calibrated multi-view.
Ranked #6 on 3D Multi-Person Pose Estimation on Campus
no code implementations • CVPR 2021 • Yao-Chih Lee, Kuan-Wei Tseng, Yu-Ta Chen, Chien-Cheng Chen, Chu-Song Chen, Yi-Ping Hung
We take advantage of the recent self-supervised framework on jointly learning depth and camera ego-motion estimation on raw videos.