1Amazon FAR (Frontier AI & Robotics), 2UC Berkeley, 3CMU, 4Stanford University
* Equal contribution. Amazon FAR team co-lead.

Interactive Demo

Explore the PHP policy in a real-time MuJoCo simulation running fully in your browser.

The demo may take a few seconds to load as models and assets are streamed.

💡 Tips:
1️⃣ To climb an obstacle: first align and run toward it, then keep holding W during approach, climbing, on top, and descent, until you’re back on the ground.
2️⃣ When running, tap A / D to align with the next obstacle, or press Y to switch between low and high speed and find your rhythm!
3️⃣ If the robot fails or drifts off course, hit the Reload button or press BACKSPACE to restart a new run.

💡 Keep holding W during approach, climbing, on top, and descent, until you’re back on the ground!

🎮 How to Play

Controls
W Move forward A Turn left D Turn right Climb Autonomous when pressing W Y Toggle speed mode SPACE Pause BACKSPACE Reset run

PHP enables highly agile parkour

Cat Vault + Dash Vault

Speed Vault

1.25m Wall Climbing + Roll

Rolling Down from 1.25m Wall

PHP enables long-horizon skill chaining

Adaptive to Real-Time Obstacle Displacement

0.76m Obstacle Climbing + Stepping

0.58m Obstacle Climbing + Stepping

Continuous Stepping

Abstract

While recent advances in humanoid locomotion have achieved stable walking on varied terrains, capturing the agility and adaptivity of highly dynamic human motions remains an open challenge. In particular, agile parkour in complex environments demands not only low-level robustness, but also human-like motion expressiveness, long-horizon skill composition, and perception-driven decision-making. In this paper, we present Perceptive Humanoid Parkour (PHP), a modular framework that enables humanoid robots to autonomously perform long-horizon, vision-based parkour across challenging obstacle courses. Our approach first leverages motion matching, formulated as nearest-neighbor search in a feature space, to compose retargeted atomic human skills into long-horizon kinematic trajectories. This framework enables the flexible composition and smooth transition of complex skill chains while preserving the elegance and fluidity of dynamic human motions. Next, we train motion-tracking reinforcement learning (RL) expert policies for these composed motions, and distill them into a single depth-based, multi-skill student policy, using a combination of DAgger and RL. Crucially, the combination of perception and skill composition enables autonomous, context-aware decision-making: using only onboard depth sensing and a discrete 2D velocity command, the robot selects and executes whether to step over, climb onto, vault or roll off obstacles of varying geometries and heights. We validate our framework with extensive real-world experiments on a Unitree G1 humanoid robot, demonstrating highly dynamic parkour skills such as climbing tall obstacles up to 1.25m (96% robot height), as well as long-horizon multi-obstacle traversal with closed-loop adaptation to real-time obstacle perturbations.