FLAM: Foundation Model-Based Body Stabilization for Humanoid Locomotion and Manipulation

Xianqi Zhang1, Hongliang Wei1, Wenrui Wang1,

Xingtao Wang*1, Xiaopeng Fan1,2, Senior Member, IEEE, Debin Zhao1,2, Member, IEEE

1Harbin Institute of Technology, 2Peng Cheng Laboratory

Abstract


Humanoid robots have attracted significant attention in recent years. Reinforcement Learning (RL) is one of the main ways to control the whole body of humanoid robots. RL enables agents to complete tasks by learning from environment interactions, guided by task rewards. However, existing RL methods rarely explicitly consider the impact of body stability on humanoid locomotion and manipulation. Achieving high performance in whole-body control remains a challenge for RL methods that rely solely on task rewards. In this paper, we propose a Foundation model-based method for humanoid Locomotion And Manipulation (FLAM for short). FLAM integrates a stabilizing reward function with a basic policy. The stabilizing reward function is designed to encourage the robot to learn stable postures, thereby accelerating the learning process and facilitating task completion. Specifically, the robot pose is first mapped to the 3D virtual human model. Then, the human pose is stabilized and reconstructed through a human motion reconstruction model. Finally, the pose before and after reconstruction is used to compute the stabilizing reward. By combining this stabilizing reward with the task reward, FLAM effectively guides policy learning. Experimental results on a humanoid robot benchmark demonstrate that FLAM outperforms state-of-the-art RL methods, highlighting its effectiveness in improving stability and overall performance.


Framework


Fig.1 (Top left) The framework of FLAM. (Bottom left) The robot-human pose mapping process. Joint mappings are simplified for clarity. (Right) The overview of the stabilizing reward function.


Experiments


The proposed method demonstrated strong performance in humanoid locomotion and manipulation tasks. Many tasks are incompleted, highlighting the challenges of humanoid robot control. Further research is required.

Locomotion

Fig.2 Performance comparison of methods on locomotion tasks. The dashed lines qualitatively indicate task success.

Manipulation

Fig.3 Performance comparison of methods on manipulation tasks. The dashed lines qualitatively indicate task success.


Visualization


Locomotion

Walk

Stand

Run

Reach

Hurdle

Crawl

Maze

Sit-Hard

Balance-Hard

Stair

Slide

Pole

Manipulation

Door

Bookshelf-Simple

Booksehlf-Hard

Spoon

Powerlift

Room

Insert-Small

Insert-Normal


Contact

Thanks for your attention. If you have any questions, feel free to contact Xianqi Zhang [zhangxianqi@stu.hit.edu.cn].

BibTex


@article{zhang2025flam, title={FLAM: Foundation Model-Based Body Stabilization for Humanoid Locomotion and Manipulation}, author={Zhang, Xianqi and Wei, Hongliang and Wang, Wenrui and Wang, Xingtao and Fan, Xiaopeng and Zhao, Debin}, journal={arXiv preprint arXiv:2503.22249}, year={2025} }