Top Qs
Timeline
Chat
Perspective

Machine learning control

Subfield of machine learning, intelligent control, and control theory From Wikipedia, the free encyclopedia

Remove ads

Machine learning control (MLC) is a subfield of machine learning, intelligent control, and control theory which aims to solve optimal control problems with machine learning methods. Key applications are complex nonlinear systems for which linear control theory methods are not applicable.

Types of problems and tasks

Summarize
Perspective

Four types of problems are commonly encountered:

  • Control parameter identification: MLC translates to a parameter identification[1] if the structure of the control law is given but the parameters are unknown. One example is the genetic algorithm for optimizing coefficients of a PID controller[2] or discrete-time optimal control.[3]
  • Control design as regression problem of the first kind: MLC approximates a general nonlinear mapping from sensor signals to actuation commands, if the sensor signals and the optimal actuation command are known for every state. One example is the computation of sensor feedback from a known full state feedback. Neural networks are commonly used for such tasks.[4]
  • Control design as regression problem of the second kind: MLC may also identify arbitrary nonlinear control laws which minimize the cost function of the plant. In this case, neither a model, the control law structure, nor the optimizing actuation command needs to be known. The optimization is only based on the control performance (cost function) as measured in the plant. Genetic programming is a powerful regression technique for this purpose.[5]
  • Reinforcement learning control: The control law may be continually updated over measured performance changes (rewards) using reinforcement learning.[6][7]
Remove ads

Adaptive Dynamic Programming

Summarize
Perspective

Adaptive Dynamic Programming (ADP), also known as approximate dynamic programming or neuro-dynamic programming, is a machine learning control method that combines reinforcement learning with dynamic programming to solve optimal control problems for complex systems. ADP addresses the "curse of dimensionality" in traditional dynamic programming by approximating value functions or control policies using parametric structures such as neural networks. The core idea revolves around learning a control policy that minimizes a long-term cost function , defined as , where is the system state, is the control input, is the instantaneous reward, and is a discount factor. ADP employs two interacting components: a critic that estimates the value function , and an actor that updates the control policy . The critic and actor are trained iteratively using temporal difference learning or gradient descent to satisfy the Hamilton-Jacobi-Bellman (HJB) equation:  

 

where describes the system dynamics. Key variants include heuristic dynamic programming (HDP), dual heuristic programming (DHP), and globalized dual heuristic programming (GDHP).[7]

ADP has been applied to robotics, power systems, and autonomous vehicles, offering a data-driven framework for near-optimal control without requiring full system models. Challenges remain in ensuring stability guarantees and convergence for general nonlinear systems.  

Remove ads

Applications

MLC has been successfully applied to many nonlinear control problems, exploring unknown and often unexpected actuation mechanisms. Example applications include:

Many more engineering MLC application are summarized in the review article of PJ Fleming & RC Purshouse (2002).[12]

As is the case for all general nonlinear methods, MLC does not guarantee convergence, optimality, or robustness for a range of operating conditions.

See also

References

Loading content...

Further reading

Loading related searches...

Wikiwand - on

Seamless Wikipedia browsing. On steroids.

Remove ads