
Continuous Time Dynamical Systems
Optimal control deals with the problem of finding a control law for a given system such that a certain optimality criterion is achieved. An optimal control is a set of differential equations describing the paths of the control variables that minimize the cost functional.
This book, Continuous Time Dynamical Systems: State Estimation and Optimal Control with Orthogonal Functions, considers different...
Optimal control deals with the problem of finding a control law for a given system such that a certain optimality criterion is achieved. An optimal control is a set of differential equations describing the paths of the control variables that minimize the cost functional.
This book, Continuous Time Dynamical Systems: State Estimation and Optimal Control with Orthogonal Functions, considers different...