CDS 110b: Linear Quadratic Regulators: Difference between revisions

From Murray Wiki
Jump to navigationJump to search
No edit summary
 
(9 intermediate revisions by the same user not shown)
Line 1: Line 1:
{{cds110b-wi06}}
{{cds110b-wi08 lecture|prev=Optimal Control|next=Receding Horizon Control}}
This lecture provides a brief derivation of the linear quadratic  regulator (LQR) and describes how to design an LQR-based compensator.  The use of integral feedback to eliminate steady state error is also described. __NOTOC__
This lecture provides a brief derivation of the linear quadratic  regulator (LQR) and describes how to design an LQR-based compensator.  The use of integral feedback to eliminate steady state error is also described. __NOTOC__


== Lecture Outline ==
* {{cds110b-wi08 pdfs|L3-1_lqr.pdf|Lecture Presentation}}  
<ol type=I>
* {{cds110b-wi08 pdfs|hw3.pdf|Homework 3}} - due 30 Jan 08
<li> Derivation of the LQR regulator
* {{cds110b-wi08 pdfs|pvtol_lqr.m|pvtol_lqr.m}} - MATLAB script demonstrating LQR design
<li> Choosing LQR weights
<li> Incorporating a reference trajectory
<li> Integral feedback
<li> Design example
</ol>
 
== Lecture Materials ==
* {{cds110b-pdfs|L2-1_LQR.pdf|Lecture Presentation}} ({{cds110b-mp3s|L2-1_LQR.mp3|MP3}})
* {{cds110b-pdfs|lqr.pdf|Lecture notes on LQR control}}
* {{cds110b-pdfs|hw2.pdf|Homework 2}}


== References and Further Reading ==
== References and Further Reading ==
* Friedland, Ch 9 - this is the assigned reading for this lecture. The derivation of the LQR controller is done differently, so it gives an alternate approach.
* R. M. Murray, ''Optimization-Based Control''. Preprint, 2008: {{cds110b-wi08 pdfs|optimal_26Jan08.pdf|Chapter 2 - Optimal Control}}
* Lewis and Syrmos, Section 3.4 - this follows the derivation in the notes above.  I am not putting in a scan of this chapter since the course text is available, but you are free to have a look via [http://books.google.com/books?ie=UTF-8&hl=en&vid=ISBN0471033782&id=jkD37elP6NIC Google Books].
* Lewis and Syrmos, Section 3.4 - this follows the derivation in the notes above.  I am not putting in a scan of this chapter since the course text is available, but you are free to have a look via [http://books.google.com/books?ie=UTF-8&hl=en&vid=ISBN0471033782&id=jkD37elP6NIC Google Books].
* Friedland, Ch 9 - the derivation of the LQR controller is done differently, so it gives an alternate approach.


== Frequently Asked Questions ==
== Frequently Asked Questions ==


'''Q: What do you mean by the terms damping ratio and natural frequency in 1c and 1d on HW #2?'''
'''Q: What do you mean by penalizing something, from <math>Q_x \geq 0</math> "penalizes" state error?'''


<blockquote>
<blockquote>
<p>The position control system is a second-order system. So for any control law, there would be some natural frequency and damping ratio for the closed loop system (Recall from CDS110a that these terms are defined for any second order system; for a revision of these concepts, you can take a look at ocw.mit.edu/NR/rdonlyres/Mathematics/18-03Spring2004/ B76E6F4F-7B05-4DA0-A5A5-03FA4ACCB6B2/0/sup_13.pdf) All you have to do in 1c and 1d is to find out the value of these terms for the control law from 1b.</p>
<p>According to the form of the quadratic cost function <math>J</math>, there are three quadratic terms such
as <math>x^T Q_x x</math>, <math>u^T Q_u u</math>, and <math>x(T)^T P_1 x(T)</math>. When <math>Q_x \geq 0</math> and if <math>Q_x</math> is relative big, the value of <math>x</math> will have bigger contribution to the value of <math>J</math>. In order to keep <math>J</math> small, <math>x</math> must be relatively small. So  selecting a big <math>Q_x</math> can keep <math>x</math> in small value regions. This is what the "penalizing" means.</p>
 
<p>So in the optimal control design, the relative values of <math>Q_x</math>, <math>Q_u</math>, and <math>P_1</math> represent how important  <math>X</math>, <math>U</math>, and <math>X(T)</math> are in the designer's concerns.</p>


<p>Vijay Gupta, 17 Jan 05</p>
<p>Zhipu Jin,13 Jan 03</p>
</blockquote>
</blockquote>

Latest revision as of 03:29, 2 March 2008

CDS 110b Schedule Project Course Text

This lecture provides a brief derivation of the linear quadratic regulator (LQR) and describes how to design an LQR-based compensator. The use of integral feedback to eliminate steady state error is also described.

References and Further Reading

  • R. M. Murray, Optimization-Based Control. Preprint, 2008: Chapter 2 - Optimal Control
  • Lewis and Syrmos, Section 3.4 - this follows the derivation in the notes above. I am not putting in a scan of this chapter since the course text is available, but you are free to have a look via Google Books.
  • Friedland, Ch 9 - the derivation of the LQR controller is done differently, so it gives an alternate approach.

Frequently Asked Questions

Q: What do you mean by penalizing something, from "penalizes" state error?

According to the form of the quadratic cost function , there are three quadratic terms such as , , and . When and if is relative big, the value of will have bigger contribution to the value of . In order to keep small, must be relatively small. So selecting a big can keep in small value regions. This is what the "penalizing" means.

So in the optimal control design, the relative values of , , and represent how important , , and are in the designer's concerns.

Zhipu Jin,13 Jan 03