Hello there! I am a second-year PhD student at HKU (the University of Hong Kong), advised by Prof. Ping Luo.

From February 2024 to October 2025, I worked as a research intern with the Humanoid Research team at Shanghai AI Lab, where I was fortunate to work with Dr. Jiangmiao Pang and Dr. Jingbo Wang.

I received both my M.Eng. and B.Eng. degrees from Tsinghua University under the supervision of Prof. Guijin Wang.

Research interests: Whole-body control Reinforcement learning Human-object interaction

If you are interested in any of these topics, or would simply like to chat, feel free to drop me an email.

profile photo

Junli Ren「任峻立」

Projects
SMASH: Mastering Scalable Whole-Body Skills for Humanoid Ping-Pong with Egocentric Vision

Junli Ren†,*, Yinghui Li†,*, Kai Zhang*, Penglin Fu*, Haoran Jiang, Yixuan Pan, Guangjun Zeng, Tao Huang, Weizhong Guo, Peng Lu, Tianyu Li, Jingbo Wang, Li Chen, Hongyang Li, Ping Luo

Preprint

[Project Page] [Paper] [Video]

The first ego-centric humanoid table tennis player, as well as agile smash behaviors.

Humanoid Goalkeeper: Learning from Position Conditioned Task-Motion Constraints

Junli Ren*, Junfeng Long*, Tao Huang, Huayi Wang, Zirui Wang, Feiyu Jia, Wentao Zhang, Jingbo Wang†, Ping Luo†, Jiangmiao Pang†

Preprint

[Project Page] [Paper] [Video] [code]

Humanoid Goalkeeper learns a single end-to-end RL policy, executing agile, human-like motions to intercept flying balls, as well as performing tasks such as escaping a ball using jump and squat motions.

VB-Com: Learning Vision-Blind Composite Humanoid Locomotion Against Deficient Perception

Junli Ren, Tao Huang, Huayi Wang, Zirui Wang, Qingwei Ben, Junfeng Long, Yanchao Yang Jiangmiao Pang†, Ping Luo†

International Conference on Robotics and Automation (ICRA), 2026

[Project Page] [Paper] [Video] [BibTeX]

We propose VB-Com, a composite framework that enables humanoid robots to determine when to rely on the vision policy and when to switch to the blind policy under perceptual deficiency.

AdaMimic: Towards Adaptable Humanoid Control via Adaptive Motion Tracking

Tao Huang, Huayi Wang, Junli Ren, Kangning Yin, Zirui Wang, Xiao Chen, Feiyu Jia, Wentao Zhang, Junfeng Long, Jingbo Wang†, Jiangmiao Pang†

International Conference on Robotics and Automation (ICRA), 2026

[Project Page] [Paper] [Video] [code]

we introduce AdaMimic, a novel motion tracking algorithm that enables adaptable humanoid control from a single reference motion.

PhysHSI: Towards a Real-World Generalizable and Natural Humanoid-Scene Interaction System

Huayi Wang*, Wentao Zhang*, Runyi Yu*, Tao Huang, Junli Ren, Feiyu Jia, Zirui Wang, Xiaojie Niu, Xiao Chen, Jiahe Chen, Qifeng Chen†, Jingbo Wang†, Jiangmiao Pang†

Preprint

[Project Page] [Paper] [Video] [code]

We present a physical-world humanoid-scene interaction system, PhysHSI, that enables humanoids to autonomously perform diverse interaction tasks while maintaining natural and lifelike behaviors.

Learning Humanoid Standing-up Control across Diverse Postures

Tao Huang, Junli Ren, Huayi Wang, Zirui Wang, Qingwei Ben, Muning Wen, Xiao Chen, Jianan Li, Jiangmiao Pang†

Robotics: Science and Systems (RSS 2025)
🏆 Best Paper Finalist

[Project Page] [Paper] [Video] [BibTeX]

we present HoST (Humanoid Standing-up Control), a reinforcement learning framework that learns standing-up control from scratch, enabling robust sim-to-real transfer across diverse postures.

BeamDojo: Learning Agile Humanoid Locomotion on Sparse Footholds

Huayi Wang, Zirui Wang, Junli Ren, Qingwei Ben, Tao Huang, Weinan Zhang, Jiangmiao Pang†

Robotics: Science and Systems (RSS 2025)

[Project Page] [Paper] [Video] [BibTeX]

BeamDojo achieves efficient learning in simulation and enables agile locomotion with precise foot placement on sparse footholds in the real world, maintaining a high success rate even under significant external disturbances.

Learning Humanoid Locomotion with Perceptive Internal Model

Junfeng Long*, Junli Ren*, Moji Shi*, Zirui Wang, Tao Huang, Ping Luo, Jiangmiao Pang†

International Conference on Robotics and Automation (ICRA), 2025

[Project Page] [Paper] [Code] [BibTeX]

We propose the Perceptive Intenal Model (PIM), a method to estimate environmental disturbances with perceptive information, enabling agile and robust locomotion for various humanoid robots on various terrains.

TOP-Nav: Legged Navigation Integrating Terrain, Obstacle and Proprioception Estimation

Junli Ren*, Yikai Liu*, Yingru Dai, Junfeng Long, Guijin Wang†

Conference on Robot Learning (CoRL), 2024

[Project Page] [Paper] [Code] [BibTeX]

We propose TOP-Nav, a novel legged navigation framework that integrates a comprehensive path planner with Terrain awareness, Obstacle avoidance and close-loop Proprioception.


Updated at April. 2026.
Template from Jon Barron.