Reinforcement Learning

GowU: Uncertainty-Guided Tree Search for Hard Exploration

We introduce Go-With-Uncertainty (GowU), a new approach to exploration in reinforcement learning that treats exploration as a particle-based search guided by uncertainty, rather than as learning a policy to maximize an exploration objective. GowU achieves state-of-the-art results on Montezuma’s Revenge, Pitfall!, and Venture, and solves pixel-based MuJoCo Adroit and AntMaze tasks without expert demonstrations.

Decoupling Exploration and Policy Optimization: Uncertainty Guided Tree Search for Hard Exploration

Pre-Print 2026 (New!)

End-to-End Efficient RL for Linear Bellman Complete MDPs with Deterministic Transitions

Pre-Print 2026

Oracle-Efficient Adversarial Reinforcement Learning via Max-Following

ICML 2025 Workshop

Is a Good Foundation Necessary for Efficient Reinforcement Learning? The Computational Role of the Base Model in Exploration

Pre-Print 2025

Beating Adversarial Low-Rank MDPs with Unknown Transition and Bandit Feedback

NeurIPS 2024

The Power of Resets in Online Reinforcement Learning

NeurIPS 2024 (Spotlight)

Sample and Oracle Efficient Reinforcement Learning for MDPs with Linearly-Realizable Value Functions

Pre-Print 2024

Reinforcement Learning under Latent Dynamics: Toward Statistical and Algorithmic Modularity

ICML 2024

Efficient Model-Free Exploration in Low-Rank MDPs

Pre-Print 2023