Robust Fuzzy Q-Learning-Based Strictly Negative Imaginary Tracking Controllers for the Uncertain Quadrotor Systems
Abstract
Quadrotors are one of the popular unmanned aerial vehicles (UAVs) due to their versatility and simple design. However, the tuning of gains for quadrotor flight controllers can be laborious, and accurately stable control of trajectories can be difficult to maintain under exogenous disturbances and uncertain system parameters. This paper introduces a novel robust and adaptive control synthesis methodology for a quadrotor robot’s attitude and altitude stabilization. The developed method is based on the fuzzy reinforcement learning and Strictly Negative Imaginary (SNI) property. The first stage of our control approach is to transform a nonlinear quadrotor system into an equivalent Negative-Imaginary (NI) linear model by means of the feedback linearization (FL) technique. The second phase is to design a control scheme that adapts online the Strictly Negative Imaginary (SNI) controller gains via fuzzy Q-learning, inspired by biological learning. The proposed controller does not require any prior training. The performance of the designed controller is compared with that of a fixed-gain SNI controller, a fuzzy-SNI controller, and a conventional PID controller in a series of numerical simulations. Furthermore, the stability of the proposed controller and the adaptive laws are proofed using the NI theorem.
Index Terms:
Strictly Negative Imaginary Controller, Adaptive Fuzzy, Reinforcement learning, Q-Learning, Robust and Adaptive Control, Quadcopter Unmanned Aerial Vehicle, Uncertainties.I Introduction
Robustness and adaptability have been the driving themes behind the development of modern control techniques. Negative Imaginary (NI) systems’ theory provides one such technique that can provide stability and robustness against uncertainties. This has been used beneficially in different applications including large flexible structures [1], multi-link robotic arms [2] and aerial robotics [3, 4, 5].
A stable linear single input single output (SISO) system with a transfer function is said to have the Strictly Negative Imaginary (SNI) property when the following requirement is fulfilled: . This indicates that the Nyquist contour of always lies below the real axis for all positive frequencies [6]. Stability analysis using NI theory only uses the loop gains at zero and infinite frequencies. According to [6, 7, 8], the feedback interconnection between an NI plant with transfer function and an SNI controller with transfer function , obtains robust stability if the DC gain at the zero frequency is less than one: .
In general, most physical systems are inherently nonlinear and time-varying. Hence, it is very difficult to get an accurate mathematical model of the system. Several of the existing techniques like Linear Quadratic Regulator (LQR/LQG) [9], Model Predictive control (MPC) [10], and linear-distributed model predictive control [11] require an accurate model. To use the NI systems theory, one can linearise the given non-linear systems around the desired operating point and then apply the traditional linear NI control methods. Due to the limitations on the accuracy of the mathematical models (non-linear and the linearised version), there is no guarantee of robustness provided by the linear SNI theory. This paper provides a novel solution to handle this problem.
Feedback linearisation [12, 13] is a convenient tool to obtain a linear system representation from a non-linear model. The non-linearity is effectively eliminated using a feedback controller. The equivalent dynamic system has a particular structure, that is used in this paper to design the SNI controller. Although a generalised linear model can be obtained by the use of linear approximations, the system’s main parameters are required to be fixed and well-known. Hence, the accuracy of the model can be undermined when unknown disturbances or unmodelled parameters are present.
Since a broad class of unstructured disturbances and uncertainties are often present in the system dynamics, adaptive controllers have gained importance over the years in handling uncertainties and parameter variations. To enhance the robustness of an SNI controller, the parameters of the SNI controller, consisting of the DC gain and time constant, can be adaptive. This adaptation can be carried out online using various adaptive control methods (e.g., Fuzzy Interference System (FIS) [14, 15] or Artificial Neural Networks (ANN)) [16, 17] to adapt to environmental changes, structural damage, or uncertain parameters in the linear SNI/NI systems, where necessary and sufficient conditions for internal stability are still guaranteed; see [18, 19, 4]. Using adaptive control strategies, one can improve tracking performance and eliminate the necessity for an accurate dynamic model.
Generally, the adaptive control schemes are categorized into two primary techniques: either passive (PT) or active (AT). Unlike the AT control strategies dramatically increasing the computational time due to the design of the additional fault detection and diagnosis observers, PT approaches can reject the faults’ influences faster, accommodating the system recovered from the fault states quicker. Furthermore, many recent AT approaches have just been designed for the SISO case; whereas, many existing dynamic systems in nature are MIMO [16, 20].
Intelligent PT controllers using fuzzy logic have been designed for various nonlinear control applications [21, 22], system identification techniques [23, 24], and active disturbance rejection systems [25]. The design of Fuzzy Interference Systems (FIS) are based on human intuition and hence FIS are termed as ‘Knowledge-Based Systems (KBS).’ They are not only convenient to design, but are also easy to implement and interpret, particularly from the perspective of lay persons (e.g. average drone operators). These systems can also offer computational simplicity and robustness compared to some conventional mathematically-based auto-tuning procedures (e.g. MIT Rules [26]), whose performance entirely depends on the accuracy of mathematical models of the systems. In addition, fuzzy systems can deal with real-time control applications under multiple disturbances (e.g. sensor noise, wind gusts and uncertain dynamics). Furthermore, an adaptive fuzzy controller can be easily designed with adjustable parameters along with an embedded mechanism for adjusting them (e.g., a general fuzzy neural network [27], or a sliding mode control for Takagi-Sugeno fuzzy systems [28]) in order to improve the performance of the controller.
However, fuzzy logic controller design requires expert knowledge, which is often not available. Moreover, controller designers can spend many hours manually tuning complicated control loops. Hence, intelligent embedding approaches have been proposed to address theses issues. Some of the offline techniques to automatically tune the parameters of a fuzzy logic controller include genetic algorithms[29], ant colony optimization [30], particle swarm optimization [31, 32], etc. Due to their computational complexity, these are mainly used in an offline mode.
Q-learning is a model-free reinforcement learning algorithm that allows an agent to select an action at a given state by exploring various actions in a state and updating the desired reward with the actual reward for that action. The agent directly interacts with the environment to generate responses, and makes updates to state action pairs recorded in a Q-table. Hence, it does not require agents to have a model of the environment and an adaption law (e.g., adaptive fuzzy) as required by dynamic programming (DP) [33]. Moreover, it also requires a small amount of computation, compared to other machine learning approaches [34, 35].
Q-learning has gained importance over the last few years as an effective embedding tool for adaptive controllers. A hybrid Ziegler-Nichols (Z-N) fuzzy Q-learning multi agent system (MAS) approach for online adjustment of the gains of a PID controller is proposed in [36]. Using the online adjustment capability of the PID gains, the overshoot is controlled and, simultaneously, problems regarding the dynamic nature of the process are addressed. The three gains of the PID controller are initially set by the Z-N method and then, through the control process, are adjusted online via the MAS. Fuzzy Q-learning has been used in the recently published works in order to supervise a PID controller for a motor speed control application [37]. Additionally, fuzzy Q-learning in MAS has been successfully applied in energy management applications for direct control of the power exchange amongst the units of a micro-grid [38]. Here the fuzzy Q-learning approach is used to tune the parameters of the SNI control technique.
The contributions of this paper for advancing current research boundaries include:
-
1.
The paper presents a novel method to transform a general nonlinear dynamical equations into a MIMO NI linear model.
-
2.
The above transformation is used, through a hybrid approach between an adaptive NI MIMO feed-forward control loop and an adaptive SNI MIMO controller, to propose a control strategy for the second-order NI linear dynamics with double poles at the origin and to improve the quadrotor’s altitude and attitude response and accuracy. The proposed control and conversion technique can deal with time-Varying, and coupling uncertainties, which are one of the main drawbacks of the recently published adaptive SNI SISO control methods [18, 39, 19].
-
3.
A novel adaptive control scheme, called Fuzzy Q-learning SNI (Fuzzy-QL-SNI), is introduced to design the adaptive MIMO SNI controller. The Fuzzy-QL-SNI controllers are capable of adapting the SNI gains online according to the environmental changes without the need for pre-training or the use of an expert’s system experience.
-
4.
Through extensive simulations of a quadrotor system that include the nominal system, external disturbances, and internal parameter variations, the relative merits of our adaptive control system are investigated with respect to a fixed-gain SNI, fuzzy-SNI, and traditional PID control methods.
The structure of this paper is given as follows. Section II presents the main theory of Negative-Imaginary systems, the feedback linearisation technique, fuzzy logic systems, and reinforcement learning. Section III describes a dynamic model of a quadrotor and presents the design of the proposed control strategy, while Section IV analyses the stability of the closed-loop control system. Section V simulates the capabilities and performance of the proposed Fuzzy-QL-SNI control techniques. Finally, Section VI presents conclusions drawn from this work.
For the convenience, Table I summaries a list of major variables used in this paper.
Symbols | Parameters |
---|---|
Maximum eigenvalue | |
Fuzzy set of the input | |
Total number of fuzzy rules | |
Script input vector | |
Output variable vector of the defuzzification process | |
Firing strength of the rule | |
Fuzzy consequent | |
Agent states | |
Allowable actions | |
Reward value | |
Accumulated reward function | |
Discounting factor | |
Action-Value function | |
Actual q-value of the chosen action | |
Rate of change in q values | |
Learning rate | |
Control input vector | |
DC gain of the SNI controller | |
Time constant of the SNI controller | |
Feed Forward gain |
II Preliminaries
II-A Negative-Imaginary systems
In this section, we briefly present the negative imaginary systems theory and its formal definition for linear time invariant (LTI) systems.
Negative imaginary systems theory was introduced by Lanzon and Petersen in [11, 6]. A generalized definition was developed in [7] to accommodate systems with free body motion. We can understand the negative imaginary property by considering the case of single-input single-output (SISO) system. For instance, the negative imaginary property is defined by considering the properties of the imaginary part of the frequency response and requiring the condition to be satisfied for all . Roughly speaking, the NI systems are stable systems having a phase lag between 0 and for all . That is, their Nyquist plot lies below the real axis when the frequency varies in the open interval . This is similar to positive real systems, where the Nyquist plot is constrained to lie in the right half of the complex plane.
Consider an LTI system that has the following state space representation,
(1) | ||||
(2) |
Here, the matrices The transfer function matrix corresponding to this system is given by:
(3) |
For the above LTI system, an NI system is defined as follows;
Definition 1
[7] A square transfer function matrix is NI if the following conditions are satisfied:
-
1.
has no pole in .
-
2.
For all such that is not a pole of ,
(4) -
3.
If with is a pole of , then it is a simple pole and the residue matrix is Hermitian and positive semidefinite.
-
4.
If is a pole of , then for all and is Hermitian and positive semidefinite.
Similarly, the strictly negative imaginary system (SNI) is also defined as follows;
Definition 2
[11] A square transfer function matrix is SNI if the following conditions are satisfied:
-
1.
has no pole in .
-
2.
For all , .

The fundamental result of interconnected negative imaginary systems theory can be summarised as follows;
Lemma II.1
where is the maximum eigenvalue.
II-B Feedback linearisation for nonlinear problems
Feedback linearization is a widely used approach in linearizing nonlinear dynamical systems [40]. The basic idea of feedback linearization is to find a transformation that transforms given nonlinear dynamics into an equivalent linear dynamical system (see Fig. 2).
Consider the following nonlinear system
(6) |
where represents the state vector, represents the input vector, and represents output vector. The main idea is to design the control law as follows;
(7) |
such that the dynamics from the new input to the output are linear.

The negative imaginary property arises in many practical systems. For example, such systems arise when considering the transfer function from a force actuator to a corresponding colocated position sensor (for instance, a piezoelectric sensor) in a lightly damped structure [41, 6, 42]. Another area where the underlying system dynamics are NI, are nano-positioning systems; see e.g., [42, 43, 44]. Similar issues also arise in application to the area of robotics, where position measurements are also widely used.
II-C Sugeno Fuzzy Inference System
A fuzzy controller or model uses fuzzy rules, which are linguistic if-then statements, consisting of fuzzy sets, fuzzy logic, and fuzzy inference, to express complex processes without the use of complicated models [18, 19]. These rules embed expert’s control/modeling knowledge and represent experience in connecting fuzzy variables using linguistic terms [45]. Fig. 3 describes the overall architecture of Fuzzy Logic control. The common structure of a fuzzy rule is:
(8) |
where is a fuzzy set of the input. is the total number of fuzzy rules. While represents the script input vector, denotes the output variable obtained from the defuzzification process, and indicates a fuzzy consequent offered by the expert.

The fundamental operators ‘and/or/not’ combine the conditions of the variable inputs that are satisfied. Based on the degree of match between fuzzy input and the rules, which rules need to be implemented and the firing strength of the rule is defined according to the given input area.
Sugeno (TS) fuzzy inference, which is the most commonly used, works with singleton output membership functions that are either constant or a linear function of the input values. By using a weighted average over all rule outputs to compute the script output, the defuzzification process for the Sugeno system is more computationally efficient than that of a Mamdani system. The final output can be obtained using the Wang-Mendel technique [46]:
(9) |
where is the consequent of rule .
II-D Q-learning
Q-learning is a subclass of reinforcement learning, which is one of the three basic machine learning platforms; along with supervised learning and unsupervised learning. Reinforcement learning represents the learning process in the nature around us. Reinforcement learning relies on a goal-directed problem solving mechanism. Q-learning is a model-free control method, where the control policy generated depends on the interaction or the feedback between the environment and the agent without knowing any information about the system model. Therefore, a Q-reinforcement learning policy is defined as a learning control law for the system to decide on the actions to be performed.
The main components in reinforcement learning in general are the agent states , the agent allowable actions , and the rewards . Here, the agent or the controller observes states and therefore performs an action (control signal) . Hence, the environment (the system under control) updates the new states and produces a reward . The reward is then received by the agent and chooses a new action . This feedback interaction proceeds until a terminal state is reached.
The main aim of the agent (the controller) in reinforcement learning is to find or learn the optimal control policy. This control policy is to maximize a discounted accumulated reward function , which may take the following form:
(10) |
Here, is a discounting factor that defines the significance of the future rewards, .
The action-value function (Q-Function), , indicates the expected return with initial state and initial action and all forward actions controlled by policy :
(11) |
In the Q-learning algorithm, the update rule is governed by the following equation:
(12) |
where the learning rate represents the degree the new information overrides the old one.
Fuzzy Q-Learning is a fuzzy extension of the Q-learning algorithm. In creating an FQL model, one first needs to specify the input states and their corresponding fuzzy sets and then build a Fuzzy Inference System (FIS) to integrate with the Q-Learning algorithm. In Section III-B, a fuzzy Q-Learning algorithm for our Quadrotor system is developed.
III Quadrotor Model and Control Strategy
III-A Negative-Imaginary Quadrotor Model Using the Feedback Linearisation Technique
The general dynamic model of a quadrotor using the Euler-Lagrangian method has been derived in the article [47, 48]. To derive the quadrotor’s motions in 3D space, an inertial frame whose origin is at the defined home location and a body frame whose origin is at the quadrotor’s center of gravity are formed as in Fig. 4.

Let denote the generalized coordinates and let denote the position vector of the air vehicle with respect to an inertial frame. Moreover, the orientation of the quadrotor is represented by the three Euler angles, namely the roll angle around the axis, the pitch angle around the axis, and the yaw angle around the axis, that generate a rotational vector . Based on the Euler angles, coordinates are transferred between inertial and body frames as required.
The overall dynamic model describing the nonlinear quadrotor behavior is summarized as follows [49]:
(13) |
where
(14) |
in which stands for the moments of inertia about the respective axes and represents the mass of the quadrotor. (, ) and (, ) are the translational and rotational drag coefficients. Further, () and () represent () and (), respectively. While is the rotor torque coefficient, is the rotor thrust coefficient. Additionally, stands for the inertia of the propeller and = (-1)i+1 is the total angular speed of propellers, wherein is the speed of the rotor. Moreover, the control signal vector includes the total thrust and the pitch, roll, and yaw moments. Finally, represents the distance from the center of the quadrotor to the actuator axes.
The proposed state-space model of the quadrotor in (13) cannot be converted into an equivalent linear and controllable system by static state feedback due to under-actuation. It is, however, possible to resort to dynamic state feedback to achieve full-state linearisation. After that, starting from the new feedback linearized design, a linear controller can be designed. In this section, the feedback linearization approach for the quadrotor UAV attitude and altitude will be presented.
According to (7), the control input vector can be chosen to eliminate all nonlinear terms in (13), leading to a linear system. Using dynamic inversion, can be obtained by:
(15) |
In our approach, for convenience, drag coefficients are assumed to be zero since drags are negligible at low speed [50]. The attitude and altitude quadrotor dynamic formulas in (13), therefore, can be re-written as:
(17) |
III-B Fuzzy Q-Learning Adaption Law
In the Q Learning method, the values of for the state-action (SA) pairs in the table are continuously computed and updated at each iteration using a Q function. However, if the number of SA pairs is large, the computational load will dramatically increase, which might fail in real-time implementation. For scaling up to large or continuous SA spaces, TS fuzzy function approximation can be used in combination with Q-Learning to generate the Fuzzy Q-Learning control theory [52, 37]. In this new method, the state of the agent is the crisp set of the fuzzy inputs . Hence, the firing strength of each rule determines the degree to which the agent matches a state.
Additionally, the rules do not have fixed consequences, meaning that there are no predefined state-action pairs. They can be increased through the exploration/exploitation algorithm. Thus, the learning agent has to find the best conclusion for each rule via competing actions (e.g., the action with the best Q-value):
(21) |
where is the possible action in rule and is a corresponding Q value. Furthermore, is defined by is and is … and is , in which , j = 1,…,n are TS fuzzy rules.
Therefore, the global action for input vector and the selected rule consequences can be expressed as follows:
(22) |
On the other hand, the corresponding Q-value is calculated by the following equation:
(23) |
where is the actual q-value of the chosen action in the fired rule , using an exploration/exploitation strategy and an action that has the maximum Q value for the fired rule , such that = , .
Based on gradient descent, the rate of change in q values are calculated as follows:
(24) |
where
(25) |
Thus, the q values are updated as:
(26) |
where indicates the learning rate that determines how much the model updates the action value depending on the reward prediction error.
The pseudo code of this algorithm is presented as in in Table II:
Steps | Description |
---|---|
1 | Measure the state |
2 | Choose the actual consequence for each fired rule using |
the exploration/exploitation approach | |
3 | Compute the global consequence and its |
corresponding Q-value | |
4 | Implement the action and observe the new state |
5 | Calculate the reward |
6 | Update q-values using (24). |
III-C Control strategy
The control strategy for the quadrotor plant is shown in Fig. 5. The tracking errors arise from the subtraction of the desired attitude and altitude set points and their actual values. The error signals are considered as the input vector of both SNI tracking controllers and the FQL agents. The SNI tracking control system whose gains are tuned offline by minimizing the tracking error function and then adjusted online by the fuzzy Q-Learning algorithm has the following general representation as follows:
(27) |
where is the DC gain, is the time constant, and is a feed forward gain.

Note that the SNI controller in (27) is in a different form to the one used in [3, 19, 4], since it contains a negative feed forward term . The primary reason for using a feed forward term in the SNI controller is due to stability conditions. This will be explained in the next section.
The proposed SNI gains are tuned online by the fuzzy Q-Learning algorithm. The input signal of the fuzzy Q-Learning agent is bounded within a range of [-2 2]. There are six Gaussian-shaped membership functions (e.g., NB, NS, Z, PS, PB) for the state variables, as depicted in Fig. 6.

The output of each agent is a vector with two output variables , each of them for each gain of the SNI controller. Each output variable involves a group of three singleton fuzzy actions. For the output variable , its singleton fuzzy sets are {-20, 0, 20} and for the output variable , the group of the fuzzy singletons is {-0.002, 0, 0.002}. Each linguistic value, produced at the output variable, denotes the percentage variation of each gain according to its initial value.
As mentioned in Section II. D, the agent communicates with the environment and performs updates to the state-action pairs in the Q-Table. Moreover, there are two primary operations: exploring and exploiting to acquire information through its experience. While the exploiting action uses the available data to make a decision via the max value of those actions, the exploring action enables the agent to explore new states that may not be selected during the exploitation process. To accelerate the learning process at the beginning of the experiment, where the fuzzy Q-Learning agent mainly explores, the agent will search for new knowledge during a certain number of iterations; see Fig 9(c). After that, the agent conducts exploitation for 99% of the time and the remaining 1% of the time is exploration for a given state.
The agent reward is selected the same as in [37]:
(28) |
where and are the tracking errors in the next and current states. The input error signal is limited in range of [-2 2]m. If the tracking error is 0, becomes 1. The obtained reward result can take either positive or negative values. If the reward values are positive, the error in the following state is decreased (reward). Otherwise, if the reward values are negative, the next state’s error is raised (punishment).
The future discounted accumulated reward therefore, is rewritten as:
(29) |
.
IV Stability Analysis
The linear transfer function matrix from the new input vector variable to the output vector is a double integrator as depicted in Eq. (20), which represents a lossless NI system as presented in [51, 7]. In this case, as indicated in [7], the SNI controller must have a negative DC-gain. The reason for this is the poles’ location of the linearized system, which implies that the DC-gain of the linearized system is infinity. The transfer function matrix and at the zero frequency can be expressed as:
(30) |
(31) |
Therefore, for the stability DC-gain condition in Lemma 2.1 to be satisfied, the controller has to have a negative DC-gain at the zero frequency, and hence;
(32) |
V Simulation Results
Since quadrotor systems are highly nonlinear and time-varying, designing a high-performance tracking controller is challenging in the presence of uncertainties and variations in the system dynamics. A hybrid approach, including an adaptive SNI controller based on FQL theory and fixed-gain NI feed-forward gains, is developed to handle extreme (fast and huge variations) pose attitudes and altitude without any prior training. Simulations are conducted to test the capabilities of the proposed adaptive SNI controller to achieve resilient and precise flight control under wind disturbance and pose attitudes recovery under multiple uncertainties.
Fig. 7 shows the whole system implemented in MATLAB/Simulink. The dynamic model of the system (13)-(19), mentioned in Section III. A, is used to test controller design results on a full nonlinear simulation of the quadrotor UAV. The main parameters of the quadrotor UAV used in our experiments can be seen in Table III. Moreover, there are four input signal pairs (desired and actual roll/pitch/yaw and height). The tracking error is treated as the input of the tracking controllers and the adaption laws. The MIMO output of each adaptive strategy using FQL adjusted gains and , respectively. Meanwhile, each SNI tracking controller’s output was connected with the corresponding adaptive NI feed-forward control gain , which is chosen as , in order to enhance their convergence rate and to guarantee system stability as mentioned in (32). After control moment values are produced, they are given to the input of the linear-system generator, where the nonlinear system dynamics are transformed into a fully linearized system. The roll/pitch/yaw moment and total thrust, produced by the linear-system generator, are fed to the quadrotor plant.
Parameter | Description | Values |
---|---|---|
Mass of the quadrotor | 0.65 kg | |
Roll inertia | kg.m2 | |
Pitch inertia | kg.m2 | |
Yaw inertia | kg.m2 | |
Total rotational moment of inertia | kg.m2 | |
around the propeller axis | ||
Drag factor | Nm.s2 | |
Thrust factor | N.s2 | |
Distance to the center of | 0.23 m | |
the quadrotor | ||
Gravitational acceleration | 9.81 m.s-2 |

The sample time is set to 0.01s since the physical VICON Motion Tracker system in our lab published the pose data to the relevant drone at the same frequency. Moreover, Xrotor Pro X9 brushless motors mounted on the quadrotor also normally operates at 0.01s. During the simulation, various values of the setpoints (roll, pitch, yaw, and altitude) are given to the quadrotor system. The roll angle is controlled to follow a sin wave; whereas, the pitch and yaw angles are commanded to track a square wave and a step function, respectively. Moreover, the altitude is also maintained at 1.5m after the 1s flight time.
The simulation will compare the robustness and performance of a conventional PID controller, tuned using the Ziegler-Nichols, a hybrid fixed-gain SNI (SNI) controller, tuned by trial-and-error, and a hybrid fuzzy-SNI (fuzzy-SNI) with those of the proposed method under two scenarios: without and with added multiple uncertainties. It is worth noting that the comparative controllers’ detailed parameters can be found in our previously published works [18, 19]. This comparison aims to highlight the improvement of the SNI and fuzzy-SNI through the fuzzy Q-Learning algorithm’s supervision and its adaptive ability (robustness) against internal and external changes. The learning rate and the discount factor are set to = 0.1 and = 0.7, respectively. The proposed strategy will work in the exploration phase for 0.7s at the beginning of the simulation before being switched into the exploitation phase for the remaining flight time. The selections of optimal exploration time, , parameters will be described in the next subsection V. A.
The evaluation metrics for the simulation tests are as follows; the Root Mean Square error (), the steady-state offset (), and the rise time (). The following equations calculate the and :
(33) |
where T is sample size.
(34) |
V-A Nominal Dynamics
The performance of the proposed adaptive controller is evaluated under a nominal condition. According to the simulation results observed in Fig. 8, it is obvious that most of the proposed controllers are capable of tracking the desired reference values with satisfactory performance, reasonable rise time and low values, as summarized in Table IV. In contrast, although the PID controller tracks well the constant z setpoints and the sin-wave trajectory, it cannot handle the desired extreme square and step trajectories by experiencing a high overshoot of approximately 25%. Thus, it can be seen that the proposed SNI control methods are more robust and have a better tracking capability than the traditional PID controller.
![]() |
(a) Roll Angle |
![]() |
(b) Pitch Angle |
![]() |
(c) Yaw Angle |
![]() |
(d) Z-Axis Position |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.017 | 0.055 | 0.057 | 0.46 | [rad] |
ts | 0.16 | 0.25 | 0.25 | 2.5 | [s] |
RMSE | 0.012 | 0.058 | 0.071 | 0.32 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.004 | 0.08 | 0.12 | 0.23 | [rad] |
ts | 0.16 | 0.16 | 0.23 | 1.4 | [s] |
RMSE | 0.32 | 0.33 | 0.38 | 0.72 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0 | 0 | 0 | 0 | [rad] |
ts | 8.5 | 8.5 | 7.9 | 2.36 | [s] |
RMSE | 0.25 | 0.37 | 0.38 | 0.39 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0 | 0 | 0 | 0 | [m] |
ts | 1.5 | 1.8 | 1.84 | 1.5 | [s] |
RMSE | 0.007 | 0.09 | 0.083 | 0.077 | [m] |
Additionally, the above experiment has been repeated under the different samples of , , and the initial exploration period. Firstly, as shown in Fig. 9(b), increasing the constant discount factor results in the slight variations of the RSME metrics. However, the best improvement is observed as = 0.7. Similarly, learning rate = 0.1 produces the smallest RMSE values among the potential candidates; see Fig. 9(a). Lastly, the obtained results in Fig. 9(c) demonstrate that having an exploration phase at the start time noticeably reduces the tracking errors, especially the vertical direction, due to the faster learning time. Furthermore, the smallest root means square error (RMSE) results are obtained if the beginning exploration phase is implemented in the first 0.7s of the simulation. Consequently, a 0.7s exploration phase is set at the beginning of all simulations.



V-B Disturbances and Uncertainties
For robustness analysis, the performance of the closed-loop control system will be examined against exogenous disturbances (e.g., wind gusts and air turbulence) and parameter changes. The Dryden wind gust model in MATLAB is applied to replicate a realistic wind gust scenario under realistic flight circumstances. The model employs continuous gusts along three axes based on a spatially varying stochastic process with a maximum of 5m/s [53]. In addition to the Dryden gust turbulence, the effects of air stream are also considered in our numerical experiment by the use of the 1-cos air turbulence model in [54]. The combined effects of extreme wind gusts and turbulence applied to the system throughout the simulation time is shown in Fig. 10.

Furthermore, the model uncertainties, consisting of the moment of inertia about each axis and mass, are set in consideration of 15% bias. However, the quadrotor simulation model parameters in the linear-system generator are kept as in Table III. Therefore, the residual nonlinearities and uncertainties acting on the quadrotor must be accommodated by the tracking controllers.
Evidently, the effects of the multiple disturbances alters the performance of all four controllers. Nonetheless, the fuzzy-QL-SNI controller outperforms the fuzzy-SNI, SNI, and PID by stabilizing the system with less drift from the desired pose references, as shown in the plots in Fig. 11. The PID method is unable to achieve the desired tracking performance when tracking the extreme pitch trajectory. Although the tracking performances are better for the yaw and z step trajectories, the flight paths experience the high-overshoot for yaw control and chattering phenomena in the steady-state phase, and chattering phenomena in the steady-state phase, demonstrating a slow recovery from disturbances. The similar results can be seen in Table V, the PID control approach produces large figures of 0.48rad for roll, 0.25rad for pitch, 0.51rad for yaw, and 0.11m for altitude. Moreover, the PID controller has the maximum of 0.31rad for roll, 0.72rad for pitch, 0.34rad for yaw, and 0.09m for altitude while its values for the roll and pitch are high, namely 2.5s and 1.4s. However, the SNI controllers are able to handle the wind turbulence and the system’s uncertain parameters, much faster and with less error than the PID. This is reflected in both columns of Table V and Fig. 11. Furthermore, the Fuzzy-QL-SNI controller’s performance yields much more desirable values than the other ones. Particularly, the proposed adaptive strategy obtains the lowest values of [0.026rad, 0.32rad, 0.25rad, 0.067m] for roll, pitch, yaw, and height, which are approximately 78%, 61%, and 60% lower than those with the PID, SNI and Fuzzy-SNI. The response of the Fuzzy-QL-SNI is noticeably faster with the numbers of [0.35, 0.17, 5.8, 1.5]s, almost twice as fast as that of the other control schemes. Furthermore, there are only slight differences in the evaluation metrics of the two given cases considered for the Fuzzy-QL-SNI. These observations validate the fast adaption capability and the robustness of our closed-loop adaptive control system.
![]() |
(a) Roll Dynamics |
![]() |
(b) Pitch Dynamics |
![]() |
(c) Yaw Dynamics |
![]() |
(s) Z-Axis Position |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.069 | 0.11 | 0.11 | 0.48 | [rad] |
ts | 0.35 | 0.35 | 0.35 | 2.5 | [s] |
RMSE | 0.0461 | 0.0654 | 0.0628 | 0.3179 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.08 | 0.09 | 0.11 | 0.25 | [rad] |
ts | 0.17 | 0.17 | 0.24 | 1.4 | [s] |
RMSE | 0.3297 | 0.3346 | 0.3816 | 0.7229 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.0215 | 0.0236 | 0.0261 | 0.5108 | [rad] |
ts | 5.8 | 8 | 3 | 2.3 | [s] |
RMSE | 0.244 | 0.3316 | 0.3408 | 0.3411 | [rad] |
Metrics | Fuzzy-QL-SNI | Fuzzy-SNI | SNI | PID | Units |
SO | 0.017 | 0.018 | 0.022 | 0.11 | [m] |
ts | 1.5 | 1.8 | 1.82 | 1.5 | [s] |
RMSE | 0.0672 | 0.0835 | 0.0783 | 0.0903 | [m] |
Another strength of the Fuzzy-QL-SNI controller is that it delivers low-frequency and unsaturated output control signals. This is demonstrated in Fig. 12, where the Fuzzy-QL-SNI controller produces smooth moment and thrust control signals, even under external and internal disturbances without high-frequency oscillations, which is a drawback of the other control methods [55, 56].

VI Conclusion
In our study, a novel concept of a hybrid adaptive SNI MIMO controller and a feedback linearisation technique has been presented for the accurate attitude and altitude tracking problem of an uncertain nonlinear quadrotor drone under the various flight schemes. The mathematical model of a quadrotor’s dynamics is transformed into a corresponding NI linear model by means of nonlinear feedback, performing a desirable cancellation of nonlinearities. Nevertheless, model uncertainties create uncertain nonlinear terms. Therefore, the proposed hybrid control system, consisting of a feed-forward NI gain and the adaptive SNI controller using the FQL algorithm, is developed to enhance the robust tracking performance and accommodate uncertainties and variation in plant dynamics. Our simulation results show that the fuzzy-QL-SNI exhibits better trajectory tracking compared to the adaptive fuzzy-SNI, SNI, and PID control schemes based on the achievement of the lowest and values. In addition to being robust and transparent, the proposed control system also demonstrates better robustness against uncertainties, namely errors in modeling and exogenous disturbances (wind gusts and air turbulence) and a superior noise rejection capability with respect to the performance of the other methods.
For future work, such controllers will be executed in real-time applications or extended to control networked multi-robot system consensus.
References
- [1] M. A. Mabrok, A. G. Kallapur, I. R. Petersen, and A. Lanzon. Spectral conditions for negative imaginary systems with applications to nanopositioning. IEEE/ASME Transactions on Mechatronics, 19(3):895–903, 2013.
- [2] O. Skeik and Alexander Lanzon. Robust output consensus of homogeneous multi-agent systems with negative imaginary dynamics. Automatica, 113:108799, 2020.
- [3] V. P. Tran, M. Garratt, and I. R. Petersen. Switching time-invariant formation control of a collaborative multi-agent system using Negative Imaginary systems theory. Control Engineering Practice, 95:104245, 2020.
- [4] V. P. Tran, F. Santoso, M. Garratt, and S. Anavatti. Distributed artificial neural networks-based adaptive strictly negative imaginary formation controllers for unmanned aerial vehicles in time-varying environments. IEEE Transactions on Industrial Informatics, 17(6):3910–3919, 2021.
- [5] V. P. Tran, F. Santoso, and M. A. Garratt. Adaptive trajectory tracking for quadrotor systems in unknown wind environments using particle swarm optimization-based strictly negative imaginary controllers. IEEE Transactions on Aerospace and Electronic Systems, 57(3):1742–1752, 2021.
- [6] I. R. Petersen and A. Lanzon. Feedback control of negative-imaginary systems. IEEE Control Systems Magazine, 30(5):54–72, 2010.
- [7] M. A. Mabrok, A. G. Kallapur, I. R. Petersen, and A. Lanzon. Generalizing negative imaginary systems theory to include free body dynamics: Control of highly resonant structures with free body motion. IEEE Transactions on Automatic Control, 59(10):2692–2707, 2014.
- [8] A. G. Ghallab, M. A. Mabrok, and I. R. Petersen. Lyapunov-based stability of feedback interconnections of negative imaginary systems. IFAC-PapersOnLine, 50(1):3424–3428, 2017.
- [9] H. Jafari, M. Zareh, J. Roshanian, and A. Nikkhah. An optimal guidance law applied to quadrotor using LQR method. Transactions of the Japan Society for Aeronautical and Space Sciences, 53(179):32–39, 2010.
- [10] S. Vazquez, J. Rodriguez, M. Rivera, L. G. Franquelo, and M. Norambuena. Model predictive control for power converters and drives: Advances and trends. IEEE Transactions on Industrial Electronics, 64(2):935–947, 2016.
- [11] A. Lanzon and I. R. Petersen. Stability robustness of a feedback interconnection of systems with negative imaginary frequency response. IEEE Transactions on Automatic Control, 2008.
- [12] H. Voos. Nonlinear control of a quadrotor micro-UAV using feedback-linearization. In 2009 IEEE International Conference on Mechatronics, pages 1–6. IEEE, 2009.
- [13] M. A. Lotufo, L. Colangelo, C. Perez-Montenegro, C. Novara, and E. Canuto. Embedded model control for UAV quadrotor via feedback linearization. IFAC-PapersOnLine, 49(17):266–271, 2016.
- [14] Y. Li and S. Tong. Adaptive fuzzy output constrained control design for multi-input multioutput stochastic nonstrict-feedback nonlinear systems. IEEE Transactions on Cybernetics, 47(12):4086–4095, 2016.
- [15] W. Sun, S. F. Su, Y. Wu, J. Xia, and V. T. Nguyen. Adaptive fuzzy control with high-order barrier Lyapunov functions for high-order uncertain nonlinear systems with full-state constraints. IEEE Transactions on Cybernetics, 2019.
- [16] Y. Li, K. Sun, and S. Tong. Observer-based adaptive fuzzy fault-tolerant optimal control for siso nonlinear systems. IEEE transactions on cybernetics, 49(2):649–661, 2018.
- [17] W. Bai, Q. Zhou, T. Li, and H. Li. Adaptive reinforcement learning neural network control for uncertain nonlinear system with input saturation. IEEE Transactions on Cybernetics, 2019.
- [18] V. P. Tran, F. Santoso, M. A. Garratt, and I. R. Petersen. Adaptive second order Strictly Negative Imaginary controllers based on the interval type-2 fuzzy systems for a hovering quadrotor with uncertainties. IEEE/ASME Transactions on Mechatronics, 25(1):11–20, 2020.
- [19] V. P. Tran, F. Santoso, M. A. Garratt, and I. R. Petersen. Fuzzy self-tuning of strictly negative-imaginary controllers for trajectory tracking of a quadcopter unmanned aerial vehicle. IEEE Transactions on Industrial Electronics, 68(6):5036–5045, 2021.
- [20] Y. Li, Y. Liu, and S. Tong. Observer-based neuro-adaptive optimized control of strict-feedback nonlinear systems with state constraints. IEEE Transactions on Neural Networks and Learning Systems, 2021.
- [21] J. Qiu, K. Sun, T. Wang, and H. Gao. Observer-based fuzzy adaptive event-triggered control for pure-feedback nonlinear systems with prescribed performance. IEEE Transactions on Fuzzy systems, 27(11):2152–2162, 2019.
- [22] X. Zhang, Y. Wang, G. Zhu, X. Chen, Z. Li, C. Wang, and C. Y. Su. Compound adaptive fuzzy quantized control for quadrotor and its experimental verification. IEEE Transactions on Cybernetics, 2020.
- [23] C. W. Chang and C. W. Tao. A novel approach to implement Takagi-Sugeno fuzzy models. IEEE transactions on cybernetics, 47(9):2353–2361, 2017.
- [24] C. M. Lin, T. L. Le, and T. T. Huynh. Self-evolving function-link interval type-2 fuzzy neural network for nonlinear system identification and control. Neurocomputing, 275:2239–2250, 2018.
- [25] J. Fei and T. Wang. Adaptive fuzzy-neural-network based on RBFNN control for active power filter. International Journal of Machine Learning and Cybernetics, 10(5):1139–1150, 2019.
- [26] N. Tariba, A. Bouknadel, A. Haddou, N. Ikken, H. E. Omari, and H.E. Omari. Comparative study of adaptive controller using MIT rules and Lyapunov method for MPPT standalone PV systems. In AIP Conference Proceedings, volume 1801, page 040008. AIP Publishing LLC, 2017.
- [27] P. M. Kebria, A. Khosravi, S. Nahavandi, D. Wu, and F. Bello. Adaptive type-2 fuzzy neural-network control for teleoperation systems with delay and uncertainties. IEEE Transactions on Fuzzy Systems, 2019.
- [28] Y. Wang, H. R. Karimi, H. Shen, Z. Fang, and M. Liu. Fuzzy-model-based sliding mode control of nonlinear descriptor systems. IEEE transactions on cybernetics, 49(9):3409–3419, 2018.
- [29] S. Loussaief and A. Abdelkrim. Convolutional neural network hyper-parameters optimization based on genetic algorithms. International Journal of Advanced Computer Science and Applications, 9(10):252–266, 2018.
- [30] F. Olivas, F. Valdez, O. Castillo, C. I. Gonzalez, G. Martinez, and P. Melin. Ant colony optimization with dynamic parameter adaptation based on interval type-2 fuzzy logic systems. Applied Soft Computing, 53:74–87, 2017.
- [31] F. Aboura. Tuning PID controller using hybrid genetic algorithm particle swarm optimization method for AVR system. In 2019 International Aegean Conference on Electrical Machines and Power Electronics (ACEMP) & 2019 International Conference on Optimization of Electrical and Electronic Equipment (OPTIM), pages 570–574. IEEE, 2019.
- [32] D. D. Chen and Y. C. Lin. A particle swarm optimization-based multi-level processing parameters optimization method for controlling microstructures of an aged superalloy during isothermal forging. Metals and Materials International, 25(5):1246–1257, 2019.
- [33] A. K. Mondal and N. Jamali. A survey of reinforcement learning techniques: strategies, recent development, and future directions. arXiv preprint arXiv:2001.06921, 2020.
- [34] D. Liu, X. Yang, D. Wang, and Q. Wei. Reinforcement-learning-based robust controller design for continuous-time uncertain nonlinear systems subject to input constraints. IEEE transactions on cybernetics, 45(7):1372–1385, 2015.
- [35] H. Van Hasselt, A. Guez, and D. Silver. Deep reinforcement learning with double Q-learning. In Thirtieth AAAI conference on artificial intelligence, 2016.
- [36] I. Carlucho, M. De Paula, S. A. Villar, and G. G. Acosta. Incremental Q-learning strategy for adaptive PID control of mobile robots. Expert Systems with Applications, 80:183–199, 2017.
- [37] P. Kofinas and A. I. Dounis. Fuzzy Q-learning agent for online tuning of PID controller for DC motor speed control. Algorithms, 11(10):148, 2018.
- [38] M. Esmaeili, H. Shayeghi, H. M. Nejad, and A. Younesi. Reinforcement learning based PID controller design for LFC in a microgrid. COMPEL-The international journal for computation and mathematics in electrical and electronic engineering, 2017.
- [39] V. P. Tran, F. Santoso, M. Garratt, and S. Anavatti. Neural network-based self-learning of an adaptive strictly negative imaginary tracking controller for a quadrotor transporting a cable-suspended payload with minimum swing. IEEE Transactions on Industrial Electronics, 68(10):10258–10268, 2021.
- [40] F. Garces, V. M. Becerra, C. Kambhampati, and K. Warwick. Introduction to feedback linearisation. In Strategies for Feedback Linearisation, pages 27–60. Springer, 2003.
- [41] J. L. Fanson and T. K. Caughley. Positive position feedback control for large space structures. AIAA Journal, 28(4):717–724, April 1990.
- [42] B. Bhikkaji and S. Moheimani. Fast scanning using piezoelectric tube nanopositioners: A negative imaginary approach. In Proc. IEEE/ASME Int. Conf. Advanced Intelligent Mechatronics AIM 2009, pages 274–279, Singapore, July 2009.
- [43] A. Sebastian and S. M. Salapaka. Design methodologies for robust nano-positioning. IEEE Transactions on Control Systems Technology, 13(6):868–876, 2005.
- [44] Ivan M. Diaz and Emiliano Pereira mand Paul Reynolds. Integral resonant control scheme for cancelling human-induced vibrations in light-weight pedestrian structures. Structural Control and Health Monitoring, 19:55–69, 2012.
- [45] L. H. Tsoukalas and R. E. Uhrig. Fuzzy and neural approaches in engineering. 1997.
- [46] J. Gou, Z. Fan, C. Wang, W. Luo, and H. Chi. An improved Wang-Mendel method based on the FSFDP clustering algorithm and sample correlation. Journal of Intelligent & Fuzzy Systems, 31(6):2839–2850, 2016.
- [47] M. Chen, S. Xiong, and Q. Wu. Tracking flight control of quadrotor based on disturbance observer. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2019.
- [48] Y. Wu, K. Hu, X. M. Sun, and Y. Ma. Nonlinear control of quadrotor for fault tolerance: A total failure of one actuator. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2019.
- [49] G. Jithu and P. R. Jayasree. Quadrotor modelling and control. In 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), pages 1167–1172. IEEE, 2016.
- [50] S.Musa. Techniques for quadcopter modeling and design: A review. Journal of unmanned system Technology, 5(3):66–75, 2018.
- [51] M. A. Mabrok, A. G. Kallapur, I. R. Petersen, and A. Lanzon. A stability result on the feedback interconnection of negative imaginary systems with poles at the origin. In 2012 2nd Australian Control Conference, pages 98–103. IEEE, 2012.
- [52] H. Van Hasselt. Reinforcement learning in continuous state and action spaces. In Reinforcement learning, pages 207–251. Springer, 2012.
- [53] S. Gage. Creating a unified graphical wind turbulence model from multiple specifications. In AIAA Modeling and simulation technologies conference and exhibit, page 5529, 2003.
- [54] S. I. Adelfang and O. Smith. Gust models for launch vehicle ascent. In 36th AIAA aerospace sciences meeting and exhibit, page 747, 1998.
- [55] N. T. Nguyen. Model-reference adaptive control. In Model-Reference Adaptive Control, pages 83–123. Springer, 2018.
- [56] D. Yao, H. Li, R. Lu, and Y. Shi. Distributed sliding-mode tracking control of second-order nonlinear multiagent systems: an event-triggered approach. IEEE Transactions on Cybernetics, 2020.