← Back to all projects

Augmenting Learned Centroidal Controller with Adaptive Force Control

Timeline: Spring 2025 Role: Reinforcement Learning and Control Engineer Team: 4 members Focus: Robust Quadrupedal Locomotion
Quadruped Locomotion Reinforcement Learning L1 Adaptive Control Robustness
L1 adaptive filter response plot
Mass versus distance comparison chart
Poster for the RL-AFQP project

Research Objective

This work enhances the CAJUN hierarchical reinforcement learning framework for quadrupedal jumping by introducing an L1 adaptive control augmentation to the low-level Quadratic Program (QP) solver. While CAJUN excels at learning centroidal motion policies for continuous jumping, it suffers from degraded performance under dynamic changes and sim-to-sim transfer. Our goal was to improve robustness to model uncertainty and payload variation by integrating adaptive force compensation directly into the control loop.

Key Features Achieved

  • Implemented L1 adaptive law to estimate and compensate unmodeled disturbances in centroidal wrench space.
  • Modified CAJUN’s low-level QP solver to include filtered disturbance feedback for real-time adaptation.
  • Validated robustness under varying payloads (up to 2× nominal mass) and inertia mismatch in MuJoCo simulations.
  • Demonstrated consistent jump trajectories and ~4× performance gain over baseline QP controllers in perturbed setups.
  • Achieved stable low-frequency adaptation with high-frequency control loops through discrete low-pass filtering.

Challenges & Learnings

  • Observed CAJUN’s sensitivity to simulator-specific inertial parameters, requiring careful dynamic calibration.
  • Addressed instability in disturbance feedback by tuning cutoff frequencies of the L1 filter for smooth adaptation.
  • Gained deeper insight into centroidal dynamics modeling and the trade-off between fast adaptation and stability margins.
  • Learned integration of model-free learning with adaptive model-based control for hybrid locomotion frameworks.