Finite-Horizon Optimal Control and Stabilization of Time-Scalable Systems
Alex Fax and Richard Murray
2000 Conference on Decision and Control
In this paper, we consider the optimal control of time-scalable systems.
The time-scaling property is shown to convert the PDE associated
with the Hamilton-Jacobi-Bellman (HJB) equation to a purely spatial PDE.
Solution of this PDE yields the value function at a fixed time, and that solutio
n can be
scaled to find the value function at any point in time. Furthermore, in certain
cases the
unscaled control law stabilizes the system, and the unscaled value function acts
as a
Lyapunov function for that system. For the example of the
nonholonomic integrator, this PDE is solved, and the
resulting optimal trajectories coincide with the known solution to that problem.
Conference
Submission (PDF, 234K, 13 pages)
Downloading and printing FAQ
Richard Murray
(murray@cds. caltech.edu)