General Indoor Navigation of Humanoid Robots
Published:
Overview
We are developing a mobile manipulator platform capable of safe, reliable, and generalizable autonomy in human environments. The system will demonstrate full building-scale navigation and interaction—handling tasks such as operating elevators, opening doors, traversing staircases, and moving across multiple floors—without human intervention.
Technical Approach
3D Mapping & Localization:
- High-precision SLAM (fastlio2) and global local planning (3D A*, cmu_local_planner) for robust navigation in dynamic, cluttered spaces.
Perception & Scene Understanding:
- Fusion of LiDAR, RGB-D, and semantic models for reliable obstacle avoidance, spatial reasoning, and environment grounding.
Manipulation Capabilities:
- Dexterous interaction with common infrastructure i.e. button pressing, door pushing/pulling, object placement.
Locomotion Capabilities:
- Reinforcement learning based control for robust and smooth locomotion on plane, rough terrain and stairs
- Capable of push-recovery and fall-recovery
Learning & Adaptation:
- Reinforcement learning and imitation learning in simulation (Isaac Gym, MuJoCo)
- Policy transferred to hardware for real-world tasks.
Agentic Autonomy:
- High-level task planning and skill composition, enabling long-horizon missions such as “navigate from office A to lab B across floors. Then grab me a coke from the kitchen on the second floor and bring it to me”
- Using VLM for general perception (segmenting elevator buttons, interacting with elevator panels)

Repository
Dedicated website coming up soon.
Contributors

