no code implementations • 10 May 2024 • Ethan Elms, Yasir Latif, Tae Ha Park, Tat-Jun Chin
Event sensors offer high temporal resolution visual sensing, which makes them ideal for perceiving fast visual phenomena without suffering from motion blur.
1 code implementation • 8 Mar 2022 • Tae Ha Park, Simone D'Amico
These tasks are all related to detection and pose estimation of a target spacecraft from an image, such as prediction of pre-defined satellite keypoints, direct pose regression, and binary segmentation of the satellite foreground.
2 code implementations • 6 Oct 2021 • Tae Ha Park, Marcus Märtens, Gurvan Lecuyer, Dario Izzo, Simone D'Amico
Autonomous vision-based spaceborne navigation is an enabling technology for future on-orbit servicing and space logistics missions.
no code implementations • 12 Aug 2021 • Tae Ha Park, Juergen Bosse, Simone D'Amico
This work presents the most recent advances of the Robotic Testbed for Rendezvous and Optical Navigation (TRON) at Stanford University - the first robotic testbed capable of validating machine learning algorithms for spaceborne optical navigation.
no code implementations • 5 Nov 2019 • Mate Kisantal, Sumant Sharma, Tae Ha Park, Dario Izzo, Marcus Märtens, Simone D'Amico
Reliable pose estimation of uncooperative satellites is a key technology for enabling future on-orbit servicing and debris removal missions.
2 code implementations • 1 Sep 2019 • Tae Ha Park, Sumant Sharma, Simone D'Amico
It is also shown that when using the texture-randomized spacecraft images during training, regressing 3D bounding box corners leads to better performance on spaceborne images than regressing surface keypoints, as NST inevitably distorts the spacecraft's geometric features to which the surface keypoints have closer relation.