CDS 110b: Receding Horizon Control

From MurrayWiki
Jump to: navigation, search
CDS 110b Schedule Project

This set of lectures builds on the previous three weeks and explores the use of online optimization as a tool for control of nonlinear control. We being with an high-level discussion of optimization-based control, refining some of the concepts initially introduced in Week 1. We then describe the technique of receding horizon control (RHC), including a proof of stability for a particular form of receding horizon control that makes use of a control Lyapunov function as a terminal cost. A detailed implementation example, the Caltech ducted fan, is used to explore some of the computational tradeoffs in optimization-based control.

Prerequisites: Students should be familiar with the concepts of trajectory generation and optimal control as described in Weeks 1-3. For the proof of stability for the receding horizon controller that we use, familiarity with Lyapunov stability analysis at the level given in AM08, Chapter 4 (Dynamic Behavior) is required.

References and Further Reading

Frequently Asked Questions

Q: How do you do trajectory optimization using differential flatness

The basic idea in using flatness for optimal trajectory generation is to rewrite the cost function and constraints in terms of the flat outputs and then parameterize the flat outputs in terms of a set of basis functions:

$$

z(t) = \sum_i \alpha_i \psi_i(t)

$$

Here, $\psi_i$, $>i = 1, \dots, N$ are the basis functions (eg, $\psi_i(t) = t^i$) and $\alpha_i$ are constant coefficients.

Once you have parameterized the flat outputs by $\alpha$, you can convert all expressions involving $z$ into functions involving $\alpha$. This process is described in a more detail in the lectures notes (Section 4).

Q: Is the condition given by Jadbabaei and Hauser and example of a CLF or the definition of a CLF?

I was a bit sloppy defining CLFs in lecture. The formal definition is given in the lectures notes (Section 2.2, Defn 1). Briefly, given a system

$$

\dot x = f(x,u),

$$

we say that a (smooth) function $V(x)$ is a control Lyapunov function (CLF) if

  • $V(x) > 0$ for all $x \neq 0$
  • $V(x) = 0$ if and only if $x = 0$
  • The derivative of $V$ along trajectories of the system satisfies
$$

\min_u \dot V(x) |_{\dot x = f(x,u)} = \min_u \frac{\partial V}{\partial x} f(x, u) < 0

$$

for all $x$.

The condition for stability given in lecture is that there exists a CLF for the system that in addition satisfies the relationship

$$

\min_u (\dot V(x) + L(x, u)) < 0

$$

along the trajectories of the system. Thus we have to have the derivative of $V$ be sufficiently negative definite in order to insure that the terminal cost $V(x)$ provides stability.

Q: Why do receeding horizon trajectories need to go to zero (on slide 4)?

It is common in control problems to assume that the desired trajectory goes to zero as its desired end state. This is implicitly the case whenever you see an integral cost of the form $x^T Q x$ or a terminal cost $x^T(T) P_1 x(T)$, both of which are minimized when $x$ is zero. There are two ways to think about this:

  • If we wish to move to a different (equilibrium) point $x_d$, we can always change the state to $x_{new} = x - x_d$ and then the new state $x_{new}$ has zero as the desired equilibrium point.
  • If we want to track a trajectory $x_d(t)$ (not constant), then we can solve the problem for the error system given by substrating the desired state.

This is explained in more dtail in the lecture notes on LQR control (Section 3).