Search results
Results from the WOW.Com Content Network
Optimal control is a particular control technique in which the control signal optimizes a certain "cost index": for example, in the case of a satellite, the jet thrusts needed to bring it to desired trajectory that consume the least amount of fuel. Two optimal control design methods have been widely used in industrial applications, as it has ...
In control theory, the state-transition matrix is a matrix whose product with the state vector at an initial time gives at a later time . The state-transition matrix can be used to obtain the general solution of linear dynamical systems.
H ∞ (i.e. "H-infinity") methods are used in control theory to synthesize controllers to achieve stabilization with guaranteed performance. To use H ∞ methods, a control designer expresses the control problem as a mathematical optimization problem and then finds the controller that solves this optimization.
Optimal control problem benchmark (Luus) with an integral objective, inequality, and differential constraint. Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. [1]
The internal state variables are the smallest possible subset of system variables that can represent the entire state of the system at any given time. [13] The minimum number of state variables required to represent a given system, , is usually equal to the order of the system's defining differential equation, but not necessarily.
A Carathéodory-π solution can be applied towards the practical stabilization of a control system. [ 6 ] [ 7 ] It has been used to stabilize an inverted pendulum, [ 6 ] control and optimize the motion of robots, [ 7 ] [ 8 ] slew and control the NPSAT1 spacecraft [ 3 ] and produce guidance commands for low-thrust space missions.
One of the main results in the theory is that the solution is provided by the linear–quadratic regulator (LQR), a feedback controller whose equations are given below. LQR controllers possess inherent robustness with guaranteed gain and phase margin , [ 1 ] and they also are part of the solution to the LQG (linear–quadratic–Gaussian) problem .
Inspired by—but distinct from—the Hamiltonian of classical mechanics, the Hamiltonian of optimal control theory was developed by Lev Pontryagin as part of his maximum principle. [2] Pontryagin proved that a necessary condition for solving the optimal control problem is that the control should be chosen so as to optimize the Hamiltonian. [3]