Feedback MPC for Torque-Controlled Legged Robots

05/15/2019 ∙ by Ruben Grandia, et al. ∙ ETH Zurich 0

The computational power of mobile robots is currently insufficient to achieve torque level whole-body Model Predictive Control (MPC) at the update rates required for complex dynamic systems such as legged robots. This problem is commonly circumvented by using a fast tracking controller to compensate for model errors between updates. In this work, we show that the feedback policy from a Differential Dynamic Programming (DDP) based MPC algorithm is a viable alternative to bridge the gap between the low MPC update rate and the actuation command rate. We propose to augment the DDP approach with a relaxed barrier function to address inequality constraints arising from the friction cone. A frequency-dependent cost function is used to reduce the sensitivity to high-frequency model errors and actuator bandwidth limits. We demonstrate that our approach can find stable locomotion policies for the torque-controlled quadruped, ANYmal, both in simulation and on hardware.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 5

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Model Predictive Control (MPC) has gained broad interest in the robotics community as a tool for motion control of complex and dynamic systems. The ability to deal with nonlinearities and constraints has popularized the technique for many robotic applications, such as quadrotor control [1], autonomous racing [2], and legged locomotion [3, 4, 5].

MPC strategies typically optimize an open-loop control sequence for a given cost function over a fixed time horizon. The control sequence is then executed until a new control update can be calculated based on the current state estimate. While this strategy assumes that the model is exact and that there are no external disturbances, the repeated optimization provides a feedback mechanism that can correct for modeling errors provided that the control loop can be executed at a sufficiently high rate. However, for high dimensional systems such as legged robots and due to the computational restrictions of mobile platforms, the achievable update rate of the MPC loop is insufficient to effectively deal with model uncertainty and external disturbances.

As a remedy, a separately designed, light-weight motion tracker is often used in practice [6]. The motion tracker runs at a higher rate than the MPC loop and provides feedback correction to the control sequence that was designed by the MPC. For complex systems such as legged robots it is a challenging task to design a controller that tracks arbitrary motions while satisfying the many constraints arising from the locomotion task. The fundamental problem is that such motion trackers do not look ahead in the horizon and therefore cannot anticipate changes in contact configuration. As an alternative, projected (time-varying) Linear Quadratic Regulators (LQR) have been proposed as a framework to automatically design feedback controllers around a given reference trajectory [7, 8]. However, the stabilizing feedback policy is always designed in a secondary stage and often with a different objective function than the one used for computing the optimal trajectories, which leads to inconsistency between the feedback policy and the trajectory optimization.

Fig. 1: ANYmal, the torque controlled-legged robot used in this work.

In this work, we propose a feedback MPC approach for motion control of a legged system and show that the optimized feedback policy can directly be deployed on hardware. We achieve stable locomotion under a very low update rate (15 Hz). Between updates, the optimized feedback policy removes the need for a separate motion controller. Furthermore, the modification of the control inputs is consistent with the MPC and thus produces a continuous signal across update instances.

To be able to directly apply the feedback strategy on a legged system, the optimized policy needs to respect all the constraints of the locomotion task such as friction and unilateral constraints of contact forces. To achieve this, we propose to extend the SLQ (Sequential Linear Quadratic) algorithm [9]. We use SLQ in a real-time iteration MPC scheme [10] where the algorithm optimizes a constrained feedback policy

(1)

with , where and are locally optimal input and state trajectories, respectively. is a time-varying LQR gain matrix which maps the state deviation from to an admissible control correction. We extend the algorithm to problems with inequality constraints using a barrier function method to accurately handle the constraints arising from the friction cone. We further use the frequency-aware MPC approach introduced in [11] to render the resulting feedback policy robust to the bandwidth limitations imposed by real actuators.

We perform experiments in simulation and on a real legged system (Fig. 1) and demonstrate that our approach is able to find robust and stable locomotion polices at MPC update rates as low as 15 Hz, which facilitates onboard execution on constrained mobile platforms.

I-a Related Work

Methods to incorporate robustness explicitly into the MPC methodology have been previously studied in the literature [12]. Min-max MPC [13], for example, optimizes an open-loop control sequence for the worst-case disturbance inside a predefined set. While this formulation appears attractive, it can be overly conservative due to its inability to include the notion of feedback that is inherently present in the receding-horizon implementation of the control [14].

Min-max Feedback MPC was proposed to address this shortcoming by planning over a state-dependent control policy instead of an open-loop feedforward sequence [15]. Unfortunately optimizing the feedback policy for all possible disturbance realizations does not yet scale to the problem dimensions encountered in legged robotics. However, even without considering disturbances, optimizing over the feedback policy has an additional advantage. When the update rate of the MPC loop is low, the feedback policy can provide local correction to the deviation of the real platform from the optimal trajectories. Exploiting this additional aspect of feedback MPC has not yet been fully explored in robotic applications that are subject to path constraints.

The feedback policy that minimizes a cost function for a given dynamical system and path constraints can be computed using the Hamilton-Jacobi-Bellman (HJB) equation [16]. While directly solving this equation for high dimensional systems is prohibitively complex, a variant of the dynamic programming approach known as Differential Dynamic Programming (DDP) [17] has proven to be a powerful tool in many practical applications. The SLQ method that we use in this work is a DDP-based approach which uses a Gauss-Newton approximation. Consequently, it only considers the linearized dynamics instead of a second-order approximation.

Although using the LQR gains derived from a DDP-based approach directly for motion tracking generates promising results in simulation, it dramatically fails on real hardware. This phenomenon has been reported before in other real-world applications of LQR on torque-controlled robots [8, 18]. Focchi et al. [19] have shown that instability can occur if the limitations of the low-level torque controller are neglected in the high-level control design. They have argued that the bandwidth of the low-level controller inversely relates to the achievable impedance of the high-level controller. To this end, to apply the SLQ feedback policy on hardware, we need to encode these bandwidth limitations in our optimization problem. In this work, we use the frequency-aware MPC approach introduced in [11]. This MPC formulation penalizes control actions in the frequency domain and automatically finds a trade-off between the bandwidth limitation of actuators and the stiffness of the high-level feedback policy.

I-B Contributions

We propose a whole-body MPC approach for quadrupedal robots, where the actuation commands are computed directly based on the MPC feedback policy. Specifically we present the following contributions which we empirically validate on the ANYmal platform (Fig. 1) in simulation and on real hardware:

  • We propose to apply feedback MPC for whole-body control of a legged system. To the best of our knowledge, this is the first time that such a control scheme is applied on hardware for motion control of legged robots.

  • The SLQ algorithm is extended to include inequality constraints through a barrier function method, which allows us to formulate friction cone constraints.

  • We show that our feedback MPC algorithm directly designs constraint-satisfactory LQR gains without additional computational cost.

  • A frequency domain design approach is used to incorporate actuation bandwidth limits in the MPC formulation to avoid rendering stiff gains. Thus, the feedback gains can be directly applied to the robot.

  • We show that the feedback MPC algorithm is capable of bridging the gap between low update-rate MPC and high rate execution of torque commands using only an onboard computer with moderate computational power.

Ii Method

Ii-a Problem Definition

Consider the following nonlinear optimal control problem with cost functional

(2)

where is the state and is the input at time . is a time-varying running cost, and is the cost at the terminal state . Our goal is to find an input trajectory that minimizes this cost subject to the following system dynamics, initial condition, and general equality and inequality constraints:

(3)
(4)
(5)
(6)
(7)

The feedback policy which minimizes this problem can be calculated using a DDP-based method. A variant of this method known as SLQ is introduced in [9], where it solves the above optimization problem in the absence of the inequality constraints in equation (7). This method computes a time-varying, state-affine control policy based on a quadratic approximation of the optimal value function in an iterative process. The SLQ approach uses a Lagrangian method to enforce the state-input equality constraints in (5). The pure state constraints in (6) are handled by adding a quadratic penalty term to the cost function.

In SLQ, the simulation (forward pass) and the optimization (backward pass) iterations alternate. Once the backward pass is completed, a forward pass computes a new trajectory based on the improved feedback policy. The local, Linear Quadratic (LQ), approximation of the nonlinear optimal control problem is constructed after each forward pass. The LQ model permits an efficient solution of the approximate problem by solving the Riccati differential equation. The feedback policy is then updated with an appropriate linesearch procedure in the direction of the LQ problem’s solution.

We follow the same SLQ approach and extend the method with inequality constraints through a relaxed barrier function approach.

Ii-B Relaxed Barrier Functions

Using a barrier function is a well know technique to absorb inequality constraint into the cost function. For each constraint in a given set of inequality constraints, a barrier term is added to the cost

(8)

A widely used barrier function is the logarithmic barrier used in interior-point methods. The optimal solution is approached by letting over successive iterations. However, a downside of the log-barrier is that it is only defined over the feasible space, and evaluates to infinity outside. Due to the rollout mechanism in the SLQ approach, one cannot ensure that successive iterations remain inside the feasible region at all time. Furthermore, the Hessian of the log-barrier goes to infinity as one approaches the constraint boundary, which results in an ill-conditioned LQ approximation.

The relaxed barrier functions previously proposed for MPC problems addresses both these issues [20] and is therefore particularly suitable for the SLQ approach. This barrier function is defined as a log-barrier function on the interior of the feasible space, and switched to a different function at a distance from the constraint boundary.

(9)

We use the quadratic extension proposed in [21]:

(10)

This results in a barrier function with continuity. The relaxed barrier function is plotted as a function of the constraint value in Fig. 2. The quadratic extension is defined over all possible constraint violations and continuously extends the log-barrier at . Moreover, the quadratic extension puts an upper bound to the curvature of the barrier function, which prevents ill-conditioning of the LQ approximation.

Fig. 2: Comparison of the log-barrier function and relaxed barrier function as defined in (9) with .

Note that by letting , the standard logarithmic barrier is retrieved. Furthermore, it has been shown that the optimal solution can be obtained for a nonzero value of [22], when the gradient of the penalty term is larger than the Lagrange multiplier of the associated constraint. Optimization with the relaxed barrier function can thus be interpreted as an augmented Lagrangian approach when and as a log-barrier method for .

Ii-C LQ Approximation

With the inequality constraints embedded in the cost function, we obtain the following linearization of the system dynamics in (3) and state-inputs constraints in (5) for a given state trajectory and input trajectory :

(11)
(12)

where , and are deviations from the previous iteration, around which the LQ approximation is made. Note that the time-dependency of the matrices was dropped to shorten the notation. The quadratic approximation of the cost in (8) is given by

(13)
(14)

which requires access to the second-order approximation of the barrier term and inequality constraints.

With the optimal control problem reduced to an equality constrained LQ approximation, the constrained Riccati backward pass in [9] yields the quadratic value function . This value function induces the optimal feedback policy in equation (1) with feedback gains computed as

(15)

where is the right pseudo-inverse of , and

is the identity matrix. Notice how the feedback gains ensure that the equality constraints are satisfied by projecting the first term to the nullspace of the constraints, and by adding the term

to satisfy the constraint when the state deviates from the plan.

Ii-D Frequency Shaping

As discussed in Sect. I-A, it has been proven difficult to use feedback gains from an LQR design on a torque-controlled robot. We propose to use the frequency-dependent cost function introduced in our previous work [11], which was used to render the feedforward solution robust to high frequency disturbances. In this work, we show that it has a similar effect on the feedback structure. We briefly summarize how the problem is adapted and refer to [11] for further details.

A frequency-dependent cost on the inputs can be introduced by evaluating the cost function on auxiliary inputs . The auxiliary inputs are defined by frequency-dependent shaping functions applied to the system inputs such that

(16)

where denotes elements associated to individual inputs, is the signal frequency in , and and

are the Fourier transform of the auxiliary input and system input respectively. Following our previous work, we use high pass filters to achieve increased costs at higher input frequencies:

(17)

The transfer function , with state space realization , is constructed such that . The original system is augmented with an additional filter state, , such that , and optimization is performed w.r.t the auxiliary inputs . The augmented system dynamics and state-input constraints are defined as

(18)

The feedback policy obtained from this augmented system is of the form

After optimization, the original input is retrieved by substituting this policy into the output function of the filter, , resulting in the complete feedback policy

(19)

Iii Implementation

We apply our approach to the kinodynamic model of a quadruped robot, which describes the dynamics of a single free-floating body along with the kinematics for each leg. The Equations of Motion (EoM) are given by

where and are the rotation matrix of the base with respect the global frame and the transformation matrix from angular velocities in the base frame to the Euler angles derivatives in the global frame. is the gravitational acceleration in body frame, and

are the moment of inertia about the CoM and the total mass respectively. The inertia is assumed to be constant and taken at the default configuration of the robot.

is the position of the foot with respect to CoM. is the orientation of the base in Euler angles, is the position of the CoM in world frame, is the angular rate, and is the linear velocity of the CoM.

 is the vector of twelve joint positions. The inputs of the model are the joint velocity commands

and end-effector contact forces .

Iii-a Equality Constraints

The equality constraints depend on the mode of each leg at a certain point in time. We assume that the mode sequence is a predefined function of time. The resulting mode-depended constraints are

where and are respectively the end-effector velocity and contact force in world frame. These constraints ensure that a stance leg remains on the ground and a swing leg follows the predefined curve in the direction of the local surface normal to avoid foot scuffing. Furthermore, the constraints enforce zero contact force at swing legs.

Iii-B Inequality Constraints

Our proposed relaxed barrier method allows to model the friction cone without the commonly used polytope approximation. The cone constraint for each end-effector,

(20)

is defined by the surface normal and friction coefficient . After projecting the contact forces to the local frame of the surface, a canonical second-order cone constraint is found in terms of local contact forces . An effective cone constraint used in conjunction with barrier methods [23] is given by

(21)

However, the gradient of this constraint is not defined at , which causes numerical issues close to the origin. While for interior point methods this problem can be solved by using the squared constraint, , this strategy does not work well together with the relaxed barrier function due to the saddle point it introduces at the origin. Since the relaxed barrier function allows infeasible iterates, the solutions can cross the origin and end up in the negative reflection of the cone, which became a feasible region through the squaring operation. We, therefore, use the perturbed cone

(22)

which is differentiable at the origin, remains infeasible for any negative , and is a conservative lower bound for the original cone (21). It therefore holds that

(23)

In Fig. 3 the level sets of this constraint are compared to the original cone. It can be seen that the constraint is convex and the zero crossing of is strictly inside the feasible region.

Fig. 3: Comparison of the friction cone constraint and the perturbed cone for , . The contour of is shown together with the zero crossing of .

Iii-C Torque Computation

The control inputs consist of contact forces and joint velocities. These commands have to be translated to torques. When using only the feedforward trajectories, desired accelerations and contact forces are extracted and tracked by a hierarchical inverse dynamics controller [24]. When using the feedback policy, we forward simulate the system under the feedback policy for a short time and extract desired accelerations from this rollout. The inverse dynamics is then only used to convert the desired accelerations into torques, without adding additional feedback. This inverse dynamics controller is evaluated at , while the SLQ-MPC algorithm runs asynchronously on a second onboard Intel i7-4600U@2.1GHz dual core processor.

Finally, each individual motor has a local, embedded, control loop. For the stance legs a torque controller is used, and for the swing legs the motors take the commanded torque as a feedforward term and close the loop over the desired position and joint velocities.

Iv Results

We first show the qualitative differences between a feedback policy and a feedforward policy when operating under low update rates and disturbances. Then, the influence of the relaxed barrier function cone constraint on the planned feedback gains is shown. We also examine the structure in the obtained feedback matrices and compare the difference between those obtained with frequency shaping and those without. Finally, we show that the proposed elements, when taken together, lead to a method that can be successfully executed on the onboard hardware of a torque-controlled robot.

We use a diagonal cost on the state and control inputs for all experiments. When frequency shaping is used, we set , for the contact force inputs and , for the joint velocity inputs in (17).

Iv-a Feedback MPC

We first investigate the effect of low update frequencies and the use of feedback policies from the SLQ-MPC in simulation. We introduce model errors to show the performance of the different strategies. The mass of the control model is increased by with respect to the simulation model. Each MPC controller is brought to a stable trot gait and commanded to move forward.

First, feedforward MPC is used with an update rate equal to the control frequency of . Every control loop thus has access to the optimal solution from the current state. The resulting desired linear accelerations are shown in the top of Fig. 4. Discontinuities only arise around a contact switch; the desired accelerations are continuous otherwise. In the middle plot, the desired accelerations are shown when the update rate is restricted to an update frequency of . The updates are clearly visible because every time the feedforward trajectory is updated, the accumulated deviation from the feedforward plan is reset and a new open loop trajectory is tracked. In the bottom plot, we show the performance when the SLQ feedback policy is used. The discontinuities at updates are significantly reduced and smooth trajectories comparable to MPC with high update rate are retrieved.

These experiments show that using the feedback from the MPC can recover some of the performance lost due to lower update rates. By using a policy that is consistent with future MPC updates, the discontinuities at the updates are reduced up to the validity of the linear quadratic approximation.

Fig. 4: Desired linear acceleration for the center of mass, as commanded by the MPC. In the top two plots a feedforward MPC strategy is used with update rates of and respectively. In the bottom plot, feedback MPC is used with an update rate of . The trajectories are for a trotting gait with a command to move ahead send at

Iv-B Feedback Gains Near Inequality Constraints

In the following experiment we prescribe a task where we require to lift the left front leg and simultaneously set the desired body location towards the front left, outside of the support polygon. This task requires the algorithm to coordinate the step with the body movement. The experiment is performed without the frequency shaped cost to allow us to focus on the effect of using the inequality constraints. In the left side of Fig. 5 we show the optimized solution without inequality constraints. Without the cone constraint, the optimal strategy is to produce negative contact forces in the right hind leg, such that the desired body position can be reached as early as possible. Furthermore, we plot the maximum gain in the row associated with each vertical contact force, which shows that the zero force feedback terms for the swing leg are compliant with the corresponding equality constraint.

The resulting solution when adding inequality constraints is shown on the right of Fig. 5. Here, we used fixed barrier parameters , , under which all constraints are strictly satisfied. With the inequality constraints, the contact forces remain positive on the right hind leg. Additionally, where before the feedback gains are about equal for the three remaining stance legs, now, the feedback gains for the right hind leg are significantly reduced. When the contact force approaches zero, the feedback gains go to zero as well, which ensures that the inequality constraint is not violated after applying the feedback policy at a disturbed state. This interaction between the inequality constraints and the feedback gains is a result of the barrier function. As the constraint boundary is approached, the Hessian of the barrier function w.r.t. the contact forces increases. This increases the input costs in (15), and thus reduces the feedback gains.

This gradual decrease in feedback gains cannot be obtained with a clamping strategy, where the gains are unaffected, or with an active set method, where the gains instantaneously decrease to zero when the constraint becomes active [25].

As a secondary effect, we see that the feedback gains on the left front leg increase before lifting the leg. Because the right hind leg is forced to have low feedback gains in the upcoming phase, the body position before lifting the foot is of high importance, which is reflected in the gains.

In addition, we note that the distance to the constraint boundary can be regulated by choosing a different barrier scaling . This means that using a finite , in contrast to decreasing it to zero as done in an interior point method, can be used to trade some optimality for a larger stability margin.

Fig. 5: Optimal vertical contact force trajectory and maximum feedback gain associated with that contact force when planning a simultaneous step and reach task without (left), and with (right) inequality constraints. The time where the left front (LF) is in the air is marked in gray.
(a) (b)
Fig. 6: Feedback matrix for ANYmal in stance configuration without (a) and with (b) frequency shaped cost. The color of each state-input entry represents the magnitude of the feedback. In (b) the matrix is split into four parts with input and states associated with each block marked at the left and bottom. The top left for example shows the gains between system states, , and system inputs.

Iv-C Feedback Structure

We visualize the feedback matrix for ANYmal in a full stance configuration. The gains obtained without using the frequency-dependent cost function are shown in Fig. 6a. The state acting on each column of the matrix is shown above the figure, and the control input affected by each row is shown on the left. The color intensity shows the magnitude of each entry in the feedback matrix, with zero shown as white and the highest gain shown in black.

In the joint velocity part of the feedback matrix, one can see how the equality constraints that require zero velocity at the end-effectors are reflected: The joint velocity commands are highly dependent on the linear and angular velocity of the base to achieve this constraint. This empirically verifies that (15) indeed produces feedback matrices that are consistent with the constraints.

The feedback matrix in Fig 6a can be compared to the feedback matrix obtained when using the frequency shaped cost function, shown in Fig 6b. We split the feedback matrix in Fig 6b into four parts, with the vertical split between system and filter state, and horizontal split between system and auxiliary input, corresponding to the partitioning in (19). First, the left side is inspected. Here we recognize a feedback pattern similar to that in Fig 6a, obtained without the frequency-dependent cost function. Indeed since the shaping functions in (17) have unit DC-gain, the feedback matrix is expected to be approximately equal to the matrix obtained without frequency shaping. Furthermore, we can recognize that the direct gains from state to contact forces are an order of magnitude smaller than before.

For the frequency shaped feedback policy, the main feedback flows from system state to auxiliary inputs. The auxiliary inputs then drive the filter states, , which in turn provide the adaptation of the system contact force inputs.

For the joint velocities, however, the feedback from filter states to joint velocities is zero, as expected. The fact that end-effector velocities are constrained to be zero is still reflected in the feedback matrix in exactly the same way as before. Feedback terms therefore appear in the bottom right partition of Fig 6b, to satisfy the equality constraint, , for the rows associated with joint velocities.

This analysis shows that using the frequency-dependent cost function introduces smoothness and reduced direct gains where possible, but at the same time still respects hard equality constraints on the original system inputs.

Fig. 7: Desired and measured torque signals in the left front knee under a constant disturbance of . The desired torque is the result of both feedforward and feedback signals. Where in the top plot the feedback is provided by the conventional tracking controller, in the bottom plot the feedback from the frequency shaped SLQ algorithm is used.

Iv-D Hardware Experiments: Disturbance Rejection

As seen in the accompanying video111A video of the experiments is available at https://youtu.be/R3YTdcQHdtE, when LQR gains obtained from the SLQ-MPC are used on hardware, the system becomes unstable even in full stance phase. As demonstrated in Sect. IV-C, the frequency shaped formulation reduces the direct gains between state and input, which enables successful deployment on hardware. All hardware experiments are therefore performed with both frequency shaping and inequality constraints active.

We first perform a simple experiment to verify the qualitative difference observed in simulation between using only a feedforward policy with a conventional tracking controller or when using a feedback policy. The robot is put in a standing configuration with the desired position set to the initial position with zero velocity. Afterward, we place a mass ( of the total mass) on top of the robot to induce a constant disturbance. Fig. 7 shows the resulting desired and measured torque trajectories in the left front knee after the system reaches an equilibrium.

In the top plot, the points at which the feedforward policy is updated are clearly visible. At each update, the state reference is reset to the measured reference, which effectively nullifies the feedback of the tracking controller. Since the feedforward control signal does not account for the additional disturbance, the system deviates from the desired trajectory and builds up feedback in the tracking controller until the next update arrives. In the bottom plot, where the feedback policy updates arrive at the same rate as the feedforward case, there are no discontinuous jumps in desired torque. This verifies our earlier observation in simulation.

Fig. 8: Gait pattern for the dynamic walk used in the hardware experiments. Colored areas represent that a leg is in contact.
Fig. 9: Torque and joint velocity trajectories during three cycles of a dynamic walk with ANYmal. The desired (dashed line) and measured (full line) are shown for the HAA (Hip Abduction Adduction), HFE (Hip Flexion Extension), and KFE (Knee Flexion Extension).

Iv-E Hardware Experiments: Dynamic Walking

Finally, we demonstrate that the proposed method achieves stable walking with all computations running on the onboard computers. We use a gait known as dynamic walk. The gait pattern is shown in Fig. 8. It consists of a mixture of underactuated and overactuated contact configurations when two and three feet are on the ground, respectively. The proposed friction cone constraint ensures that the trajectory does not require negative contact forces and thus successfully navigates the intriguing pattern of support polygons.

A receding horizon of was used, for which the MPC reaches an update frequency of approximately . The resulting desired and measured torque and joint velocity trajectories are shown in Fig. 9 for the left front leg. The desired and measured signals are close to each other at all time, showing that the applied feedback policy respects the bandwidth limits of the actuators.

V Conclusion

In this work, we proposed to use feedback MPC as an effective way to handle the slow update rate associated with the computational restrictions of mobile platforms.

We proposed a relaxed barrier function method to extend the SLQ algorithm to optimization problems with inequality constraints. In particular, the friction cone is implemented through a perturbed second-order cone constraint. This formulation adds a convex penalty to the cost function and avoids numerical ill-conditioning at the origin of the cone.

A frequency-aware MPC approach was used to systematically include the bandwidth limit of the actuators in the feedback policy design. This was a key factor to achieve closed-loop stability on hardware without any detuning of the low-level actuator controllers as suggested in [8]. The frequency-aware approach effectively allows to set high gains in the low-frequency spectrum and to attenuate gains in high frequency. It thus increases the robustness of the feedback policy in the presence of high-frequency disturbances.

We showed that the feedback policy is consistent with the constraints of the locomotion task. We empirically confirmed that the MPC policy reduces the feedback gains near the boundaries of the friction cone to respect the inequality constraints. We also demonstrated that the optimized policy sets zero gains on the contact force of the swing legs and encodes the zero end-effector velocity constraint for stance legs to satisfy state-input equality constraints.

References

  • [1] K. Alexis, C. Papachristos, G. Nikolakopoulos, and A. Tzes, “Model predictive quadrotor indoor position control,” in 2011 19th Mediterranean Conference on Control Automation (MED), June 2011, pp. 1247–1252.
  • [2] A. Liniger, A. Domahidi, and M. Morari, “Optimization-based autonomous racing of 1:43 scale rc cars,” Optimal Control Applications and Methods, vol. 36, no. 5, pp. 628–647, 2015.
  • [3] Y. Tassa, T. Erez, and E. Todorov, “Synthesis and stabilization of complex behaviors through online trajectory optimization,” in IROS, 2012, pp. 4906–4913.
  • [4] J. Koenemann, A. D. Prete, Y. Tassa, E. Todorov, O. Stasse, M. Bennewitz, and N. Mansard, “Whole-body model-predictive control applied to the hrp-2 humanoid,” in IROS, 2015, pp. 3346–3351.
  • [5] F. Farshidian, E. Jelavic, A. Satapathy, M. Giftthaler, and J. Buchli, “Real-time motion planning of legged robots: A model predictive control approach,” in Humanoids, 2017, pp. 577–584.
  • [6] R. M. Murray, Optimization-based control.   California Institute of Technology, CA, 2009.
  • [7] M. Posa, S. Kuindersma, and R. Tedrake, “Optimization and stabilization of trajectories for constrained dynamical systems,” in ICRA.   IEEE, 2016, pp. 1366–1373.
  • [8] S. Mason, N. Rotella, S. Schaal, and L. Righetti, “Balancing and walking using full dynamics lqr control with contact constraints,” in 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids).   IEEE, 2016, pp. 63–68.
  • [9] F. Farshidian, M. Neunert, A. W. Winkler, G. Rey, and J. Buchli, “An efficient optimal planning and control framework for quadrupedal locomotion,” in ICRA.   IEEE, 2017, pp. 93–100.
  • [10] M. Diehl, H. G. Bock, and J. P. Schlöder, “A real-time iteration scheme for nonlinear optimization in optimal feedback control,” SIAM Journal on control and optimization, vol. 43, no. 5, pp. 1714–1736, 2005.
  • [11] R. Grandia, F. Farshidian, A. Dosovitskiy, R. Ranftl, and M. Hutter, “Frequency-aware model predictive control,” IEEE Robotics and Automation Letters, vol. 4, no. 2, pp. 1517–1524, 2019.
  • [12] D. Mayne, J. Rawlings, C. Rao, and P. Scokaert, “Constrained model predictive control: Stability and optimality,” Automatica, vol. 36, no. 6, pp. 789 – 814, 2000.
  • [13] A. Bemporad and M. Morari, “Robust model predictive control: A survey,” in Robustness in identification and control, A. Garulli and A. Tesi, Eds.   London: Springer London, 1999, pp. 207–226.
  • [14] J. Lee and Z. Yu, “Worst-case formulations of model predictive control for systems with bounded parameters,” Automatica, vol. 33, no. 5, pp. 763 – 781, 1997.
  • [15] P. O. M. Scokaert and D. Q. Mayne, “Min-max feedback model predictive control for constrained linear systems,” IEEE Transactions on Automatic Control, vol. 43, no. 8, pp. 1136–1142, 1998.
  • [16] D. P. Bertsekas, Dynamic Programming and Optimal Control.   Athena Scientific, 1995.
  • [17] D. Mayne, “A second-order gradient method for determining optimal trajectories of non-linear discrete-time systems,” International Journal of Control, vol. 3, no. 1, pp. 85–95, 1966.
  • [18] S. Mason, L. Righetti, and S. Schaal, “Full dynamics lqr control of a humanoid robot: An experimental study on balancing and squatting,” in Humanoids.   IEEE, 2014, pp. 374–379.
  • [19] M. Focchi, G. A. Medrano-Cerda, T. Boaventura, M. Frigerio, C. Semini, J. Buchli, and D. G. Caldwell, “Robot impedance control and passivity analysis with inner torque and velocity feedback loops,” Control Theory and Technology, vol. 14, no. 2, pp. 97–112, 2016.
  • [20] C. Feller and C. Ebenbauer, “Relaxed logarithmic barrier function based model predictive control of linear systems,” IEEE Transactions on Automatic Control, vol. 62, no. 3, pp. 1223–1238, 2017.
  • [21] J. Hauser and A. Saccon, “A barrier function method for the optimization of trajectory functionals with constraints,” in Proceedings of the 45th IEEE Conference on Decision and Control, 2006, pp. 864–869.
  • [22] A. P. Aguiar, F. A. Bayer, J. Hauser, A. J. Häusler, G. Notarstefano, A. M. Pascoal, A. Rucco, and A. Saccon, Constrained Optimal Motion Planning for Autonomous Vehicles Using PRONTO.   Cham: Springer International Publishing, 2017, pp. 207–226.
  • [23] M. S. Lobo, L. Vandenberghe, S. Boyd, and H. Lebret, “Applications of second-order cone programming,” Linear algebra and its applications, vol. 284, no. 1-3, pp. 193–228, 1998.
  • [24] C. D. Bellicoso, C. Gehring, J. Hwangbo, P. Fankhauser, and M. Hutter, “Perception-less terrain adaptation through whole body control and hierarchical optimization,” in Humanoids, 2016, pp. 558–564.
  • [25] Y. Tassa, N. Mansard, and E. Todorov, “Control-limited differential dynamic programming,” in ICRA.   IEEE, 2014, pp. 1168–1175.