Search Results for author: Kyle Stachowicz

Found 4 papers, 0 papers with code

RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes

no code implementations7 May 2024 Kyle Stachowicz, Sergey Levine

The high-speed off-road driving task represents a particularly challenging instantiation of this problem: a high-return policy should drive as aggressively and as quickly as possible, which often requires getting close to the edge of the set of "safe" states, and therefore places a particular burden on the method to avoid frequent failures.

reinforcement-learning

SELFI: Autonomous Self-Improvement with Reinforcement Learning for Social Navigation

no code implementations1 Mar 2024 Noriaki Hirose, Dhruv Shah, Kyle Stachowicz, Ajay Sridhar, Sergey Levine

Specifically, SELFI stabilizes the online learning process by incorporating the same model-based learning objective from offline pre-training into the Q-values learned with online model-free reinforcement learning.

Collision Avoidance reinforcement-learning +2

ViNT: A Foundation Model for Visual Navigation

no code implementations26 Jun 2023 Dhruv Shah, Ajay Sridhar, Nitish Dashora, Kyle Stachowicz, Kevin Black, Noriaki Hirose, Sergey Levine

In this paper, we describe the Visual Navigation Transformer (ViNT), a foundation model that aims to bring the success of general-purpose pre-trained models to vision-based robotic navigation.

Visual Navigation

FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing

no code implementations19 Apr 2023 Kyle Stachowicz, Dhruv Shah, Arjun Bhorkar, Ilya Kostrikov, Sergey Levine

We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL).

Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.