Finite-Horizon Optimal Control and Stabilization of Time-Scalable Systems

Alex Fax and Richard Murray
2000 Conference on Decision and Control

In this paper, we consider the optimal control of time-scalable systems. The time-scaling property is shown to convert the PDE associated with the Hamilton-Jacobi-Bellman (HJB) equation to a purely spatial PDE. Solution of this PDE yields the value function at a fixed time, and that solutio n can be scaled to find the value function at any point in time. Furthermore, in certain cases the unscaled control law stabilizes the system, and the unscaled value function acts as a Lyapunov function for that system. For the example of the nonholonomic integrator, this PDE is solved, and the resulting optimal trajectories coincide with the known solution to that problem.

Conference Submission (PDF, 234K, 13 pages)
Downloading and printing FAQ


Richard Murray (murray@cds. caltech.edu)