Submit manuscript...
eISSN: 2574-8092

International Robotics & Automation Journal

Mini Review Volume 3 Issue 2

A survey of adaptive control

Oded Yechiel, Hugo Guterman

Department Electrical and Computer Engineering, Ben Gurion University of the Negev, Israel

Correspondence: Oded Yechiel, Department Electrical and Computer Engineering, Ben-Gurion University of the Negev, Israel, Tel 13851662331, Fax 84891836

Received: August 09, 2017 | Published: October 11, 2017

Citation: Yechiel O, Guterman H. A survey of adaptive control. Int Rob Auto J. 2017;3(2):290-292. DOI: 10.15406/iratj.2017.03.00053

Download PDF

Abstract

An adaptive control system is a system which can cope with changes and uncertainty of the plant. The adaptive controllers are playing a key role in intelligent control systems, which are required for complex systems. This paper surveys different methods in controlling a plant with uncertainties in the plant's model. The survey starts with theoretical adaptive controller schemes and moves to more abstract methodology of machine learning techniques in control.

Keywords: adaptive control, machine learning, neural networks, reinforcement learning, fuzzy logic

Abbreviations

LTI, linear time invariant; NN, neural networks; FLC, fuzzy logic control; MRAC, model reference adaptive control; MRC, model reference control; RLC, reinforcement learning control; ANFIS, adaptive network-based fuzzy inference system.

Introduction

Research on establishing a stable system using control theory has been ongoing for nearly two hundred years. Various methods and techniques have been developed throughout the years in order to control systems in a robust and stable manner. Naturally, as the controlled system is more complex, it is required to design more advanced control schemes. The most basic group of systems, (from here-on noted as a "plant"), which is analyzed and manipulated by a control system, are known linear time invariant (LTI) plants. For these sorts of systems, classic control theory produces remarkable benchmarks and is reviewed in many books and papers.1 Modern control theory, such as optimal attempt to optimize the control law by applying a cost functional on the state and the control variables.2,3 However, these methods rely on knowing the model of the plant or at least have a very good approximation.

In case the model of the plant is unknown or is non-linear, it is essential to generate a control scheme that is adaptive, and can change according to the plant's output. There are various methods for applying adaptive control and it depends on the problem at hand.4 In case the model of the system is known, but, the parameters of the model are unknown constants, Model Predictive Controllers (MRC) has shown remarkable results.5 Machine learning techniques have also shown great promise in the field of adaptive control. Methods, such as Neural Networks,6 reinforcement learning,7 and fuzzy logic,8 basically approximate a nonlinear function and therefore can provide a good representation of the nonlinear unknown plant. Machine learning is mostly used as a model-free controller. The plant is considered as a black box, and the input and output data from the plant is collected and later trained upon. After the training stage, the machine learning system represents the plant’s model and is able to control the plant without any need of a mathematical model. This paper provides a brief review on several adaptive control approaches, including theoretic approaches and machine learning based approaches. For each approach which will be reviewed a use case reference is provided.

Model reference adaptive control

In case the plant's model parameters are unknown, there are several methods in adaptive control and system identification to predict these parameters online and generate a control law for the system. One method that works extremely well and is the most used is Model Reference Adaptive Control (MRAC).5 MRAC changes the control law in a manner that will cause the system to mimic a reference model. This process is known as indirect control. In MRAC, the main idea is to have a reference model y m MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyEa8aadaWgaaqcfasaa8qacaWGTbaapaqabaaaaa@3907@  that has the same order as the original system; for a first order systems with unknown parameters a and b,

x ˙ +ax=bu MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmiEa8aagaGaa8qacqGHRaWkcaWGHbGaamiEaiabg2da9iaa dkgacaWG1baaaa@3D6B@ (1)

Hence, the reference model would be,

x ˙ m + a m x= b m r MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmiEa8aagaGaamaaBaaajuaibaWdbiaad2gaaKqba+aabeaa peGaey4kaSIaamyya8aadaWgaaqcfasaa8qacaWGTbaajuaGpaqaba WdbiaadIhacqGH9aqpcaWGIbWdamaaBaaajuaibaWdbiaad2gaaKqb a+aabeaapeGaamOCaaaa@4370@ (2)

The design objective is to make the tracking error e=x x m MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyzaiabg2da9iaadIhacqGHsislcaWG4bWdamaaBaaajuai baWdbiaad2gaaKqba+aabeaaaaa@3D6E@ converge to 0. By subtracting equation (2) from (1), we get

e ˙ + a m e=b( u a x a r r ) MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmyza8aagaGaa8qacqGHRaWkcaWGHbWdamaaBaaajuaibaWd biaad2gaa8aabeaajuaGpeGaamyzaiabg2da9iaadkgadaqadaWdae aapeGaamyDaiabgkHiTiaadggapaWaaSbaaKqbGeaapeGaamiEaaWd aeqaaKqba+qacqGHsislcaWGHbWdamaaBaaajuaibaWdbiaadkhaaK qba+aabeaapeGaamOCaaGaayjkaiaawMcaaaaa@49C1@  (3)

Where, a x = a a m b MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyya8aadaWgaaqcfasaa8qacaWG4baapaqabaqcfa4dbiab g2da9maalaaapaqaa8qacaWGHbGaeyOeI0Iaamyya8aadaWgaaqcfa saa8qacaWGTbaajuaGpaqabaaabaWdbiaadkgaaaaaaa@407A@ , a r = b m b MathType@MTEF@5@5@+= feaagKart1ev2aqatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyya8aadaWgaaqcfasaa8qacaWGYbaapaqabaqcfa4dbiab g2da9maalaaapaqaa8qacaWGIbWdamaaBaaajuaibaWdbiaad2gaa8 aabeaaaKqbagaapeGaamOyaaaaaaa@3EA2@ .

The control law is given by (4) as,

u= a ^ r r+ a ^ x x MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyDaiabg2da9iqadggagaqca8aadaWgaaqcfasaa8qacaWG Ybaapaqabaqcfa4dbiaadkhacqGHRaWkceWGHbGbaKaapaWaaSbaaK qbGeaapeGaamiEaaqcfa4daeqaa8qacaWG4baaaa@4185@ (4)

Where, it can be shown using the right Lyapunov function that the change in estimation of a ^ r MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmyyayaajaWdamaaBaaajuaibaWdbiaadkhaa8aabeaaaaa@3903@  and a ^ x MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmyyayaajaWdamaaBaaajuaibaWdbiaadIhaaKqba+aabeaa aaa@3997@ can be calculated using (5) and (6).

a ^ r =sgn( b ) γ r er MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmyyayaajaWdamaaBaaajuaibaWdbiaadkhaa8aabeaajuaG peGaeyypa0JaeyOeI0Iaam4CaiaadEgacaWGUbWaaeWaa8aabaWdbi aadkgaaiaawIcacaGLPaaacqaHZoWzpaWaaSbaaKqbGeaapeGaamOC aaWdaeqaaKqba+qacaWGLbGaamOCaaaa@4694@ (5)

a ^ x =sgn( b ) γ x ex MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmyyayaajaWdamaaBaaajuaibaWdbiaadIhaaKqba+aabeaa peGaeyypa0JaeyOeI0Iaam4CaiaadEgacaWGUbWaaeWaa8aabaWdbi aadkgaaiaawIcacaGLPaaacqaHZoWzpaWaaSbaaKqbGeaapeGaamiE aaqcfa4daeqaa8qacaWGLbGaamiEaaaa@46A6@ (6)

Where, γ r MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaeq4SdC2damaaBaaajuaibaWdbiaadkhaa8aabeaaaaa@39B4@  and γ x MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaeq4SdC2damaaBaaajuaibaWdbiaadIhaaKqba+aabeaaaaa@3A48@ are tuning parameters.

A toy example is shown in Figure 1, for the system,

x ˙ =ax+u MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmiEa8aagaGaa8qacqGH9aqpcaWGHbGaamiEaiabgUcaRiaa dwhaaaa@3C83@  (7)

Where, the parameter a is unknown. If we choose a reference model as shown in (8)

x ˙ m =2 x m +r MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGabmiEa8aagaGaamaaBaaajuaibaWdbiaad2gaa8aabeaajuaG peGaeyypa0JaeyOeI0IaaGOmaiaadIhapaWaaSbaaKqbGeaapeGaam yBaaqcfa4daeqaa8qacqGHRaWkcaWGYbaaaa@413E@ (8)

The resulting control law u= a ^ x x+r MathType@MTEF@5@5@+= feaagKart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLn hiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr 4rNCHbGeaGqiVCI8FfYJH8YrFfeuY=Hhbbf9v8qqaqFr0xc9pk0xbb a9q8WqFfeaY=biLkVcLq=JHqpepeea0=as0Fb9pgeaYRXxe9vr0=vr 0=vqpWqaaeaabiGaciaacaqabeaadaqaaqaaaOqaaKqbacbaaaaaaa aapeGaamyDaiabg2da9iqadggagaqca8aadaWgaaqcfasaa8qacaWG 4baajuaGpaqabaWdbiaadIhacqGHRaWkcaWGYbaaaa@3E7D@ with a tuning parameter of 0.5 will produce Figure 1. The main challenges of working with MRAC are that it requires an understanding of how the system actually operates. Providing a false reference model, could react with a devastating and unstable control law. In addition, in practice one should stop the adaptive component of the controller when the system is close to convergence.

Figure 1 Model reference adaptive control example.

Reinforcement learning control

Reinforcement learning control (RLC), which was initially introduced by Sutton et al.,7 and later improved by many others9,10 is a natural approach. With RLC, the control system starts with no information whatsoever and solves the Bellman equation based on trial‑ and‑ error experience. Sutton et al.7 have shown that RLC is direct adaptive optimal control, and that the system will converge to the optimal solution given infinite amount of trials. Hwangbo et al.11 have shown an example of training a controller for stabilizing a quad rotor using RLC. The RLC directly mapped the state of the quad rotor to actuator command, making any predefined control structure obsolete. The trained policy showed outstanding performance, without making any use of the model. The main disadvantage of the RLC is the long convergence time. Since the policy space can be extremely large, it requires a large number of iterations until convergence is reached. In addition, if the plant is unstable, or safety is an issue, using RLC in practice could be challenging.

Neural networks in control

Neural networks (NN) provide model-free learning controllers for a class of nonlinear systems, in the sense that a structural or parameterized model of the system dynamics is not needed.6 Throughout, there are repeatable design algorithms and guarantees of system performance including both small tracking errors and bounded NN weights.12 It has been shown that as uncertainty about the controlled system increases or as one desires to consider human user inputs at higher levels of abstraction, the NN controllers acquire more and more structure, eventually acquiring a hierarchical structure that resembles some of the elegant architectures proposed by computer science engineers using high-level design approaches based on cognitive linguistics, reinforcement learning, psychological theories, adaptive critics, or optimal dynamic programming techniques13 Neural networks have shown great promise in complex non-linear system control.14–16 One of the earlier successes in NN control was backing up a full trailer into a dock.17 It was shown in simulation how the backup of the full trailer into the dock was managed successfully after training several thousand times, thus replacing the need for a controller based on nonlinear mathematical equations. The disadvantage of the neural network control is that it does require a large training data to work with. It is not always simple to generate this data, and many times the data is generated by using a working controller, and extracting the plant’s input‑ output data for the training of the network. However, if this working controller already exists, the necessity of the NN controller is questionable.

Fuzzy logic and ANFIS

The term "fuzzy" refers to the fact that the logic involved can deal with concepts that cannot be expressed as "true" or "false" but rather as "partially true." Fuzzy logic control (FLC) has met tremendous interest in applications over the past decades, especially due to the nonlinearities of the control and the fact that it is designed based more on human experience than on a model.18 The human experience is represented by a series of if then rules that correlates well-defined linguistic variables in the input to well-defined linguistic variables in the output. However, the fact remains that the performance of the resulting controller is difficult to analyze due to its nonlinear effects. It is even difficult to establish general and non-conservative criteria for such an essential, qualitative property as closed loop stability. Despite the drawbacks, there are many systems that utilize the fuzzy logic tool in order to develop complex adaptive systems. One example is for air‑conditioning control where it was shown that using an FLC instead of previous controller designs the FLC reduced power consumption by 24%, and using neuro-fuzzy controllers provides even better energy consumption.19 Another example is the Adaptive Network-based Fuzzy Inference System (a.k.a. ANFIS).20 ANFIS is a fuzzy inference system that can construct an input-output mapping based on both human knowledge and stipulated input-output data pairs. The ANFIS architecture is employed to model nonlinear functions identify nonlinear components in a control system online. The ANFIS architecture has shown impressive results in system identification and in the control of stationary systems.21,22

Conclusion

In this paper we have reviewed several self learning adaptive control schemes and applications. The MRAC scheme provides a simple control law by estimating the model’s parameters; the controller is straightforward and easy for implementation. The main disadvantage of the MRAC approach is that it requires an understanding of the plant’s dynamics. In addition, in practice the adaptive component should be turned off after convergence, and it is not always clear how to do this. The machine-learning controllers perform well even if the system is nonlinear and unknown. The main advantage of the machine learning controllers is that there is no need for a working knowledge of the plant’s mathematical model; the controller acts as a black box, and approximates the model based on the data gathered. However, it is hard to guarantee performance, and time to convergence can be very high.

Acknowledgments

No financial interest exists.

Conflict of interest

Author declares that no conflict of interest exists.

References

  1. Dorf RC, Bishop RH. Modern control systems. USA: Prentice–Hall; 1998. p. 1–831.
  2. Gopal M. Modern control system theory. USA: Halsted Press; 1993. p. 1–644.
  3. Bubnicki Z. Modern control theory. USA: Springer Science & Business Media. 2005.
  4. Aseltine J, Mancini A, Sarture C. A survey of adaptive control systems. IRE Transactions on Automatic Control. 1958;6(1):102–108.
  5. Narendra K, Annaswamy A. Stable adaptive systems. USA: Prentice–Hall; 1986. p. 1–494.
  6. Narendra KS, Parthasarathy K. Adaptive identification and control of dynamical systems using neural networks. Proceedings of the 28th IEEE Conference on Decision and Control. 1989.
  7. Sutton RS, Barto AG, Williams RJ. Reinforcement learning is direct adaptive optimal control. IEEE Control Systems. 1992;12(2):19–22.
  8. CC Lee. Fuzzy logic in control systems: fuzzy logic controller. IEEE Transactions on systems, man, and cybernetics. 1990;20(2):419–435.
  9. Matignon L, Laurent GJ, Piat NLF. Improving Reinforcement Learning Speed for Robot Control. IEEE/RSJ International Conference on Intelligent Robots and Systems. 2006.
  10. Deisenroth MP, Rasmussen CE. Efficient reinforcement learning for motor control. 10th International PhD Workshop on Systems and Control. 2009. p. 1–6.
  11. Hwangbo J, Siegwart R ISa, Hutter M. Control of a Quadrotor With Reinforcement Learning. IEEE Robotics and Automation Letters. 2017;2(4):2096–2103.
  12. Campos J, Lewis FL. Adaptive critic neural network for feed forward compensation. Proceedings of the 1999 American Control Conference. USA. 1999.
  13. Abu–Khalaf M, Lewis FL. Nearly optimal state feedback control of constrained nonlinear systems using a neural networks HJB approach. Annual Reviews in Control. 2004;28(2):239–251.
  14. Tang ZL, Ge SS, Tee KP. Adaptive NN control for uncertain pure–feedback nonlinear systems with state constraints subject to unknown disturbances. 54th IEEE Conference on Decision and Control. 2015.
  15. Zhou Q, Shi P, Tian Y, et al. Approximation–based adaptive tracking control for MIMO nonlinear systems with input saturation. IEEE Trans Cybern. 2015;45(10):2119–2128.
  16. Chen B, Zhang H, Lin C. Observer–Based Adaptive Neural Network Control for Nonlinear Systems in Non strict–Feedback Form. IEEE Trans Neural Netw Learn Syst. 2016;27(1):89–98.
  17. Nguyen DH, Widrow B. Neural networks for self–learning control systems. IEEE Control systems magazine. 1990;10(3):18–23.
  18. Driankov D, Hellendoorn H, Reinfrank M. Introduction to Fuzzy Control. New York: Springer Verlag. 2012.
  19. Arshdeep Kaur, Amrit Kaur. Comparison of Fuzzy Logic and Neuro Fuzzy Algorithms for Air Conditioning System. International Journal of Soft Computing and Engineering. 2012;2(1):417–420.
  20. Jang JS. ANFIS: adaptive–network–based fuzzy inference system. IEEE transactions on systems, man, and cybernetics. 1993;23(3):665–685.
  21. Joelianto E, Anura DC, Priyanto M. ANFIS–hybrid reference control for improving transient response of controlled systems using PID controller. International Journal of Artificial Intelligence. 2003;10(13):S88–S111.
  22. Yang H, Fu YT, Zhang KP, et al. Speed tracking control using an ANFIS model for high–speed electric multiple unit. Control Engineering Practice. 2014;23:57–65.
Creative Commons Attribution License

©2017 Yechiel, et al. This is an open access article distributed under the terms of the, which permits unrestricted use, distribution, and build upon your work non-commercially.