CDS 110b: Receding Horizon Control
CDS 110b | ← Schedule → | Project | Course Text |
This set of lectures builds on the previous three weeks and explores the use of online optimization as a tool for control of nonlinear control. We being with an high-level discussion of optimization-based control, refining some of the concepts initially introduced in Week 1. We then describe the technique of receding horizon control (RHC), including a proof of stability for a particular form of receding horizon control that makes use of a control Lyapunov function as a terminal cost. A detailed implementation example, the Caltech ducted fan, is used to explore some of the computational tradeoffs in optimization-based control.
Prerequisites: Students should be familiar with the concepts of trajectory generation and optimal control as described in Weeks 1-3. For the proof of stability for the receding horizon controller that we use, familiarity with Lyapunov stability analysis at the level given in AM08, Chapter 4 (Dynamic Behavior) is required.
- Lecture slides on RHC overview (Mon)
- Lecture notes on RHC analysis (Wed) - the notation in these notes is slightly different than the text. In lecture, I will use the textbook notation.
- HW #4 (due 6 Feb 08): problems 3.1, 3.3 and 3.2. Students working on the course project should do problems 3.1 and 3.3 only.
References and Further Reading
- R. M. Murray, Optimization-Based Control. Preprint, 2008: Chapter 3 - Receding Horizon Control
Online Control Customization via Optimization-Based Control, R. M. Murray et al. In Software-Enabled Control: Information Technology for Dynamical Systems, T. Samad and G. Balas (eds.), IEEE Press, 2001. This paper talks about the CLF-based nonlinear RHC approach and its application on the Caltech ducted fan using NTG.
Constrained model predictive control: Stability and optimality, D. Q. Mayne, J. B. Rawlings, C. V. Rao and P. O. M. Scokaert. Automatica, 2000, Vol. 36, No. 6, pp. 789-814. This is one of the most referenced comprehensive survey papers on MPC. Gives a nice overview about its history and explains the most important issues and various approaches.
Frequently Asked Questions
Q: How do you do trajectory optimization using differential flatness
The basic idea in using flatness for optimal trajectory generation is to rewrite the cost function and constraints in terms of the flat outputs and then parameterize the flat outputs in terms of a set of basis functions:
$$ z(t) = \sum_i \alpha_i \psi_i(t)
$$Here, $\psi_i$, $>i = 1, \dots, N$ are the basis functions (eg, $\psi_i(t) = t^i$) and $\alpha_i$ are constant coefficients.
Once you have parameterized the flat outputs by $\alpha$, you can convert all expressions involving $z$ into functions involving $\alpha$. This process is described in a more detail in the lectures notes (Section 4).
Q: Is the condition given by Jadbabaei and Hauser and example of a CLF or the definition of a CLF?
I was a bit sloppy defining CLFs in lecture. The formal definition is given in the lectures notes (Section 2.2, Defn 1). Briefly, given a system
$$ \dot x = f(x,u),
$$we say that a (smooth) function $V(x)$ is a control Lyapunov function (CLF) if
- $V(x) > 0$ for all $x \neq 0$
- $V(x) = 0$ if and only if $x = 0$
- The derivative of $V$ along trajectories of the system satisfies
$$ \min_u \dot V(x) |_{\dot x = f(x,u)} = \min_u \frac{\partial V}{\partial x} f(x, u) < 0
$$for all $x$.
The condition for stability given in lecture is that there exists a CLF for the system that in addition satisfies the relationship
$$ \min_u (\dot V(x) + L(x, u)) < 0
$$along the trajectories of the system. Thus we have to have the derivative of $V$ be sufficiently negative definite in order to insure that the terminal cost $V(x)$ provides stability.
Q: Why do receeding horizon trajectories need to go to zero (on slide 4)?
It is common in control problems to assume that the desired trajectory goes to zero as its desired end state. This is implicitly the case whenever you see an integral cost of the form $x^T Q x$ or a terminal cost $x^T(T) P_1 x(T)$, both of which are minimized when $x$ is zero. There are two ways to think about this:
- If we wish to move to a different (equilibrium) point $x_d$, we can always change the state to $x_{new} = x - x_d$ and then the new state $x_{new}$ has zero as the desired equilibrium point.
- If we want to track a trajectory $x_d(t)$ (not constant), then we can solve the problem for the error system given by substrating the desired state.
This is explained in more dtail in the lecture notes on LQR control (Section 3).