CDS 110b: Optimal Control: Difference between revisions

From Murray Wiki
Jump to navigationJump to search
added link to Lewis and Syrmos
added link to google books
Line 17: Line 17:
== References and Further Reading ==
== References and Further Reading ==
* [http://www.cds.caltech.edu/~macmardg/cds110b/pontryagin.pdf Notes on Pontryagin's Maximum Principle] (courtesy of Doug MacMynowski) - this comes from a book on dynamic programming (DP) and uses a slightly different notation than we used in class.
* [http://www.cds.caltech.edu/~macmardg/cds110b/pontryagin.pdf Notes on Pontryagin's Maximum Principle] (courtesy of Doug MacMynowski) - this comes from a book on dynamic programming (DP) and uses a slightly different notation than we used in class.
* A good recent book on optimal control is [http://books.google.com/books?ie=UTF-8&hl=en&vid=ISBN0471033782&id=jkD37elP6NIC Lewis and Syrmos], which is available for puchase from Amazon
* A good recent book on optimal control is [http://books.google.com/books?ie=UTF-8&hl=en&vid=ISBN0471033782&id=jkD37elP6NIC Lewis and Syrmos], which can be searched via [http://books.google.com Google Books] and purchased online.


== Frequently Asked Questions ==
== Frequently Asked Questions ==

Revision as of 21:20, 2 January 2006

WARNING: This page is for a previous year.
See current course homepage to find most recent page available.
Course Home L7-2: Sensitivity L8-1: Robust Stability L9-1: Robust Perf Schedule

This lecture provides an overview of optimal control theory. Beginning with a review of optimization, we introduce the notion of Lagrange multipliers and provide a summary of the Pontryagin's maximum principle.

Lecture Outline

  1. Introduction: two degree of freedom design and trajectory generation
  2. Review of optimization: necessary conditions for extrema, with and without constraints
  3. Optimal control: Pontryagin Maximum Principle
  4. Examples: bang-bang control and Caltech ducted fan (if time)

Lecture Materials

References and Further Reading

Frequently Asked Questions

Q: What do you mean by penalizing something, from Q>=0 "penalizes" state error?

According to the form of the quadratic cost function J, there are three quadratic terms such as xTQx, uTRu, and x(T)TP1x(T). When Q0 and if Q is relative big, the value of x will have bigger contribution to the value of J. In order to keep J small, x must be relatively small. So selecting a big Q can keep x in small value regions. This is what the "penalizing" means.

So in the optimal control design, the relative values of Q, R, and P1 represent how important X, U, and X(T) are in the designer's concerns.

Zhipu Jin,13 Jan 03