# Difference between revisions of "CDS 110b: Optimal Control"

(→Frequently Asked Questions) |
(→Frequently Asked Questions) |
||

Line 20: | Line 20: | ||

== Frequently Asked Questions == | == Frequently Asked Questions == | ||

+ | |||

+ | '''Q: In the example on Bang-Bang control discussed in the lecture, how is the control law for <math>u</math> obtained?''' | ||

+ | |||

+ | <blockquote> | ||

+ | <p>Pontryagin's Maximum Principle says that <math>u</math> has to be chosen to minimise the Hamiltonian <math>H(x,u,\lambda)</math> for given values of <math>x</math> and <math>\lambda</math>. In the example, <math>H = 1 + ({\lambda}^TA)x + ({\lambda}^TB)u</math>. At first glance, it seems that the more negative <math>u</math> is the more <math>H</math> will be minimised. And since the most negative value of <math>u</math> allowed is <math>-1</math>, <math>u=-1</math>. However, the co-efficient of <math>u</math> may be of either sign. Therefore, the sign of <math>u</math> has to be chosen such that the sign of the term <math>({\lambda}^TB)u</math> is negative. That's how we come up with <math>u = -sign({\lambda}^TB)</math>.</p> | ||

+ | |||

+ | <p>Shaunak Sen, 12 Jan 06</p> | ||

+ | </blockquote> | ||

'''Q: Notation question for you: In the Lecture notes from Wednesday, I'm assuming that <math>T</math> is the final time and <sup><math>T</math></sup> (superscript T) is a transpose operation. Am I correct in my assumption?''' | '''Q: Notation question for you: In the Lecture notes from Wednesday, I'm assuming that <math>T</math> is the final time and <sup><math>T</math></sup> (superscript T) is a transpose operation. Am I correct in my assumption?''' |

## Revision as of 05:15, 13 January 2006

WARNING: This page is for a previous year.See current course homepage to find most recent page available. |

Course Home | L7-2: Sensitivity | L8-1: Robust Stability | L9-1: Robust Perf | Schedule |

This lecture provides an overview of optimal control theory. Beginning with a review of optimization, we introduce the notion of Lagrange multipliers and provide a summary of the Pontryagin's maximum principle.

## Lecture Outline

- Introduction: two degree of freedom design and trajectory generation
- Review of optimization: necessary conditions for extrema, with and without constraints
- Optimal control: Pontryagin Maximum Principle
- Examples: bang-bang control and Caltech ducted fan (if time)

## Lecture Materials

## References and Further Reading

- Excerpt from LS95 on optimal control - This excerpt is from Lewis and Syrmos, 1995 and gives a derivation of the necessary conditions for optimaliity. A few pages have been left out from the middle that contained some additional examples (which you can find in similar books in the library, if you are interested). Other parts of the book can be searched via Google Books and purchased online.
- Notes on Pontryagin's Maximum Principle (courtesy of Doug MacMynowski) - this comes from a book on dynamic programming (DP) and uses a slightly different notation than we used in class.

## Frequently Asked Questions

**Q: In the example on Bang-Bang control discussed in the lecture, how is the control law for obtained?**

Pontryagin's Maximum Principle says that has to be chosen to minimise the Hamiltonian for given values of and . In the example, . At first glance, it seems that the more negative is the more will be minimised. And since the most negative value of allowed is , . However, the co-efficient of may be of either sign. Therefore, the sign of has to be chosen such that the sign of the term is negative. That's how we come up with .

Shaunak Sen, 12 Jan 06

**Q: Notation question for you: In the Lecture notes from Wednesday, I'm assuming that is the final time and ^{} (superscript T) is a transpose operation. Am I correct in my assumption?**

Yes, you are correct.

Jeremy Gillula, 07 Jan 05

**Q: What do you mean by penalizing something, from Q>=0 "penalizes" state error?**

According to the form of the quadratic cost function , there are three quadratic terms such as , , and . When and if is relative big, the value of will have bigger contribution to the value of . In order to keep small, must be relatively small. So selecting a big can keep in small value regions. This is what the "penalizing" means.

So in the optimal control design, the relative values of , , and represent how important , , and are in the designer's concerns.

Zhipu Jin,13 Jan 03