SlideShare a Scribd company logo
Hierarchical Algorithms of Quasi-Linear ARX
Neural Networks for Identification of Nonlinear
Systems
Mohammad Abu Jami’in, Member, IAENG, Yuyun, Member, IAENG, Eko Julianto, Member, IAENG
Abstract—A quasi-linear ARX neural network model
(QARXNN) is a nonlinear model built using neural networks
(NN). It has a linear-ARX structure where NN is an embedded
system to give the parameters for the regression vector. There
are two contributions in this paper, 1) Hierarchical Algorithms
is proposed for the training of QARXNN model, 2) an adaptive
learning is implemented to update learning rate in NN training
to ensure global minima. First, the system is estimated by using
LSE algorithm. Second, nonlinear sub-model performed using
NN is trained to refine error of LSE algorithm. The linear
parameters obtained from LSE algorithm is set as bias vector
for the output nodes of NN. Global minima point is reached by
adjusting the learning rate based on Lyapunov stability theory
to ensure convergence of error. The proposed algorithm is
used for the identification and prediction of nonlinear dynamic
systems. Finally, the experiments and numerical simulations
reveal that the proposed method gives satisfactory results.
Index Terms—System Identification, quasi linear-ARX neural
network, hierarchical algorithm, convergence of error, nonlin-
ear system
I. INTRODUCTION
SYSTEM identification is an extensive problem because
it is difficult to propose a model capable of describing
efficiently in every possible nonlinear system. Neural net-
works (NN) and neuro-fuzzy networks are a type of black-
box models with nonparametric or multi-parametric models
[1]. The Black-box models for the identification of nonlinear
systems do not require parametric expressions. The most
useful property of NN is its ability to approximate arbitrary
linear or nonlinear mapping through training. However, there
are three major drawbacks: 1) these methods take the system
as a whole and do not permit separate analysis and identi-
fication of the linear dynamics, 2) it is difficult to use their
parameters for controller design, and 3) the stability analysis
is not easy, and the parameter tuning is generally a time-
consuming process due to its nonlinear and multi-parametric
nature [2], [3].
Some researchers adopted hybrid approach by mixing
linear and nonlinear models to improve accuracy, simplicity
and reducing computational complexity [4], [1], [5], [6].
A hybrid fuzzy time series model is constructed based
on ARIMA and Interval Type-2 Fuzzy Inference System
was proposed to improve forecasting accuracy [4]. In this
Manuscript received 05 April, 2017; revised 19 July, 2017.
Mohammad Abu Jami’in is with the Department of Automation Engi-
neering, Politeknik Perkapalan Negeri Surabaya, Surabaya, Jawa Timur,
Indonesia. e-mail: jammysby@gmail.com.
Yuyun is with the Department of Computer Engineering, STMIK Han-
dayani, Makassar, Sulawesi Selatan, Indonesia.
Eko Julianto is with the Department of Piping Engineering, Politeknik
Perkapalan Negeri Surabaya, Surabaya, Jawa Timur, Indonesia.
paper, we propose a quasi-linear ARX neural network model
(QARXNN). It has an easy-to-use structure shown by its
ARX-like structure. Nonlinear coefficients are implemented
by NN to parameterize the input vector. It is injected to
the QARXNN model as an embedded system [1]. This is
the interesting aspect of the structure that the coefficients
have some meaning to describe the dynamic behavior of the
identified plants. The parameters of system modeling can
be used to estimate the stability of the modeled system and
can be set as controller parameters [8], [9]. The coefficients
consist of linear and nonlinear parts that allow for system
analysis using the approach of linear theory.
In the previous research, the embedded system of the
QARXNN model was modified by using neuro-fuzzy,
wavelet, radial-basis function, and multilayer perceptron
(MLP) networks in order to increase the accuracy of model
for system identification [10]. However, this work does
not discuss the stable and global convergence of system
identification. In this paper, a hierarchical algorithm was
proposed for the training of QARXNN model based on
Lyapunov stability theory. The learning rate is updated every
leaning step in order to guarantee the convergence of error.
Thus, the error will be asymptotically goes to zero by the
number of training time.
Lyapunov stability is a well-known theory for control
application to guarantee closed-loop control stability. A
lyapunov-like analysis was proposed to derive stable learning
law in order to ensure the stability property of the iden-
tification error [11]. Using the Lyapunov stability theorem
for the training algorithm allows for increasing the speed of
the convergence in the learning process, guaranteeing global
convergence of the model, and avoiding local minima and
achieving global minima of cost function [12], [13]. The
Lyapunov function has been used to track convergence of
error; the output of tracking error can then asymptotically
converge to zero [13] and can be used to analyze the stability
of the identification process with stable learning law [14],
[15], [16].
In this paper, the training of QARXNN model is performed
hierarchically between linear and nonlinear sub-models based
on Lyapunov stability theorem. First, the system is estimated
using a linear sub-model under LSE algorithm. From this
step, we have linear part parameters. Second, the residual
errors of LSE algorithm (linear sub-model) are refined by
nonlinear sub-model performed using NN. The parameters
of linear part is set as bias vector for the output nodes of
NN, where the residual errors of linear sub-model is set as
the target output to train NN sub-model. Thus, the errors of
the linear sub-model is refined by the training of nonlinear
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
sub-model. From the training of NN, we have nonlinear
part parameters. Thus, by summing between the linear and
nonlinear parts parameters, we have nonlinear parameters.
If the system is linear then the coefficients converge to
the specific value of the constant, and if the system is
nonlinear then the coefficients will be a function of time [17].
Especially in control applications, this approach allow us to
derive the linear or nonlinear controllers based on the linear
or nonlinear parameters. A simple switching mechanism is
introduced between the linear and nonlinear parts parameters
in order to derive linear and nonlinear controls. [8], [18], [7].
The rest of the paper is organized as follows: Section
2 contains a brief discussion of the QARXNN model. In
Section 3, the proposed hierarchical learning algorithm is
derived with linear and nonlinear sub-models. Convergence
analysis of the proposed algorithm is also described in this
section. Experimental results are presented in Section 4, and
finally conclusions are recorded in Section 5.
II. QUASI-LINEAR ARX NEURAL NETWORK MODEL
Consider a single-input-single-output (SISO) black-box
nonlinear system whose input-output relation is described by
y(t) = g(ϕ(t)) (1)
where, g(·), y(t) ∈ R, and ϕ(t) = [y(t−1) · · · y(t−ny)u(t−
1) · · · u(t−nu)]T
are nonlinear function, system output, and a
regression vector, respectively. The general nonlinear system
expressed in (1) can be represented in a regression which has
been shown in Ref.[1,2]. The unknown nonlinear function
g(ϕ(t)) can be performed using Taylor expansion on a small
region of around ϕ(t) = 0:
y(t) = g(0) + ˙g(0)ϕ(t) +
1
2
ϕT
(t)¨g(0)ϕ(t) + · · · . (2)
As (2) shows, the coefficients are infinite. However, Taylor
expansion is only used as a bridge to transform the nonlinear
system. Applying the Taylor series transformation and the
system dynamics, the nonlinear system can be expressed as
[1]:
y(t) = y(0) + ϕ(t)T
Ω(ϕ(t)). (3)
where, y(0) = g(0) is the value for the initial condition, and
Ω(ϕ(t)) is the coefficient of the regression vector by
Ω(ϕ(t)) = [˙g(0)ϕ(t) +
1
2
ϕT
(t)¨g(0)ϕ(t) + · · ·]T
. (4)
= [a1,t · · · anu,t b0,t · · · bny−1,t] (5)
where, ai,t = ai(ϕ(t)), (i = 1, · · · , ny) is the coefficient of
the regression vector of output and bj,t = bj(ϕ(t)), (j =
0, · · · , nu − 1) is the coefficient of the regression vector of
input. The orders of time delay for the system input-output
are nu and ny.
The Quasi-linear ARX neural network is a flexible non-
linear model that has linear properties [19]. Fig. 1 shows
the structure of quasi-linear ARX neural network model.
Selecting NN as an embedded system, the model can be
written as follows:
y(t, ϕ(t)) = ϕ(t)T
ℵ(ϕ(t), Ω) (6)
= ϕ(t)T
W2ΓW1(ϕ(t)) + ϕ(t)T
θ, (7)
Fig. 1. A quasi-linear ARX neural network model.
where Ω = {W1, W2, θ} sets the network parameters. The
network parameter is Γ, W1, W2, and θ, being the diag-
onal nonlinear operator with identical sigmoidal elements
on hidden nodes, the weights of hidden layer, the weights
of output layer, and the bias vector for the output nodes,
respectively. The coefficients or state dependent parameter
estimation (SDPE), denoted by ℵ(ϕ(t), Ω) is implemented
using NN stated as
ℵ(ϕ(t), Ω) = W2ΓW1(ϕ(t)) + θ (8)
= δ(ϕ(t), W) + θ. (9)
As this demonstrates, when compared to conventional feed-
forward neural networks, the quasi-linear ARX neural net-
work has an easy-to-use structure and introduces efficient
algorithms for the parameter estimation as shown by (7) and
(8). The quasi-linear ARX neural networks model consists
of two parts: the second term of the right side of (7) shows
the linear part, while the first term is the nonlinear part.
The coefficients of the linear and nonlinear parts are fused
to parameterize the input vector. Error in the linear and
nonlinear sub-models is shared. In one step, both linear and
nonlinear parts are updated simultaneously. The estimated
parameter of the linear part changes at every learning step.
Thus, it is difficult to analyze the system identified linearly
using the linear part.
III. HIERARCHICAL ALGORITHM WITH ADAPTIVE
LEARNING FOR QARXNN MODEL
In this paper, we use a different approach to describe the
QARXNN model by separating the linear and nonlinear sub-
models. First, the linear sub-model is used to identify the
system to obtain the linear parameter estimate. The linear
estimator allows for the calculation of the linear parameter
estimation and the residual error of the model. This feature
allow for analyzing the identified plant as a linear system.
Second, the residual error of the linear sub-model is used
to train NN. NN is a compensator to refine the error of
the linear sub-model. The proposed hierarchical algorithm
fixed the linear parameter estimator, which does not change
in any training step of NN. Interestingly, this method can be
used to describe the system linearly and nonlinearly via the
switching mechanism [8]. In order to perform the hierarchical
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
algorithm, the Equation (7) can be rewritten as:
y(t, ϕ(t)) = zl + zn (10)
zl = ϕT
(t)θ (11)
zn = ϕT
(t)δ(ϕ(t), W) (12)
where, zl is the linear estimator (linear sub-model) and zn
is the nonlinear estimator (nonlinear sub-model).
In our main theoretical result, the following assumption
are made:
A1. The pairs of the input-output training data are bounded.
A2. The SDPE ℵ(ϕ(t), Ω) is bounded.
A3. There is an optimal weight of SDPE denoted by
ℵ∗
(ϕ(t), Ω).
By using the QARXNN model, the estimated output of (7)
can be rewritten in the following form:
ˆy(t) = ϕT
(t)ˆℵ(ϕ(t), Ω) (13)
if there exists ℵ∗
(ϕ(t), Ω) such that
y(t) = ϕT
(t)ℵ∗
(ϕ(t), Ω) (14)
Using (13) and (14), the adaptation error e(t) is derived as
follows:
e(t) = y(t) − ˆy(t)
= ϕT
(t)ℵ∗
(ϕ(t), Ω) − ϕT
(t)ˆℵ(ϕ(t), Ω)
= ϕT
(t)Ξ (15)
Ξ = ℵ∗
(ϕ(t), Ω) − ˆℵ(ϕ(t), Ω). (16)
where Ξ is the adaptation error of SDPE.
Suppose that the assumption A2. is valid, then there is a
positive real number α for ∥ e(t) ∥< α. The accuracy
and convergence of error characteristics are then greatly
influenced by the adaptation error Ξ.
The adaptation error is then distributed to the two sub-
models by
el = y(t) − ϕT
(t)ˆθ (17)
en = y(t) − ϕT
(t)ˆδ(ϕ(t), W). (18)
where ˆθ denotes the linear part-parameter estimator and
ˆδ(ϕ(t), W) denotes nonlinear part-parameter estimator. The
residual error of the linear and nonlinear sub-models are el
and en, respectively. The linear part-parameter estimator is
the parameter for the linear system. Therefore, it is possible
to analyze the system in a linear fashion. Hereafter, el can be
set as the target output for the nonlinear sub-model zn. By
introducing two linear and nonlinear sub-models, equation
(15) can be rewritten in the following form:
e(t) = y(t) − ˆy(t)
= y(t) − ϕT
(t)ˆℵ(ϕ(t), Ω)
= y(t) − ϕT
(t)(ˆδ(ϕ(t), W) + ˆθ)
= y(t) − ϕT
(t)ˆδ(ϕ(t), W) − ϕT
(t)ˆθ
= el − zn (19)
= en − zl. (20)
The desired goal of system identification is to make
the adaptation error e(t) reach and remain at zero. Two
algorithms are implemented hierarchically. First, the system
is estimated using the linear sub-model under the LSE
algorithm to obtain linear parameter estimator θ. Second,
the residual error of the linear sub-model is calculated by
(17) and set as the target output for the nonlinear sub-
model performed by a multilayer perceptron neural network
(MLPNN). Because the linear parameter estimator is fixed
and set as a bias vector for the output nodes of NN, sub-
model zn is estimated by (12) until its output converge to el.
By (19), e(t) will converge to zero if the output of nonlinear
sub-model zn converges to el. Thus, the nonlinear sub-model
becomes a booster to refine the estimator of the linear sub-
model.
To train the NN for the nonlinear sub-model, adaptive
learning is proposed for training similar to a BP algorithm.
The fixed learning rate of BP is replaced by the adaptive
learning rate performed based on the Lyapunov stability the-
orem. The selected Lyapunov function is defined as follows:
V =
1
2
e2
(k) (21)
where e(k) = {e1
l − z1
n, e2
l − z2
n, · · · , eN
l − zN
n }T
, N is the
number of training data, and k is the learning sequence.
Performing the derivative of (21) in discrete form, we have
∆V (k) = V (k + 1) − V (k)
=
1
2
e2
(k + 1) − e2
(k) (22)
By introducing the increment of error,
e(k + 1) = e(k) + ∆e(k) (23)
Substituting (23) in (22), we have
∆V (k) =
1
2
(
(e(k) + ∆e(k))2
− e2
(k)
)
=
1
2
(∆e(k))2
+ e(k)∆e(k) (24)
The parameters of weight W are updated using the gradi-
ent descent algorithm that may be defined as
W(k + 1) = W(k) + ∆W
= W(k) + η
(
− ∂V
∂W
)
= W(k) + η
( ∂zn
∂W
)
e(k). (25)
Thus, ∆W can be rewritten as
∆W = η
( ∂zn
∂W
)
e(k) (26)
For learning, the difference of error is due to the change
during learning, which can be expressed as
∆e(k) ≈
∂e
∂W
≈ −
( ∂zn
∂W
)T
∆W (27)
Let ∂zn
∂W = H. By using (26) and (27) in (24), we have
∆V (k) = (−H∆W)2
+ e(k)(−H∆W(k))
= H2
(ηHe(k))2
+ e(k)(−HηHe(k))
= η2
H4
e2
(k) − ηH2
e2
(k) (28)
According to the Lyapunov stability theorem, the condition
of stability is dependent on a negative ∆V (k). Therefore, the
trajectory of ∆V (k) is as follows:
∆V (k) = −λe2
(k) (29)
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
Substituting (28) and (29), we have
λ = ηH2
− η2
H4
(30)
Based on the Lyapunov stability theorem, the stability is
ensured if V (k) ) is positive-definite and if ∆V (k) is
negative-definite. Therefore, the learning rate η is adapted
to guarantee a global convergence condition, which can be
satisfied if λ > 0. Therefore, the learning rate is adapted to
satisfy the condition as follows:
η
( ∂zn
∂W
)2
< 1 (31)
The Lyapunov function is selected with the positive def-
inite function that represents the energy function. In each
learning step, the parameters of the network is updated based
on reducing the energy, and the energy is always dissipated.
The global convergence point of cost function is ensured
if the energy and dissipated energy converge to zero. This
can be shown by both as the error and the gradient of
error converge to zero through training. To guarantee global
convergence under Lyapunov stability theory, the derivative
of the Lyapunov function is ensured in that negative definite
function.
The detailed steps to train a QARXNN model based on
Lyapunov theorem is described as follows:
1) Set y(t) as zl then estimate θ using LSE algorithm of
(11).
2) Calculate el using (17) and set as output guide to train
sub-model zn in (12).
3) Use ˆθ as bias vector for output nodes of MLPNN,
and small initial values of W1(k) and W2(k). Update
W1(k), W2(k) by using adaptive learning based on
Lyapunov function (ALLF). Set k = 1, k is the
sequence of learning number.
4) Select Lyapunov function candidate, the candidate
function is stated as V = f(e), where V = 0 only
if e = 0, V > 0 only if e ̸= 0, e = el − zn.
5) Update the learning rate based on energy function by
˙V < 0. Based on Lyapunov theory, if V > 0 and
˙V < 0, then the error will converge to zero at time
goes to infinity limk→∞ e(k) = 0.
6) Update the weights using (25) and learning condition
satisfy (31)
7) Stop if pre-specified condition is met, otherwise goto
step 3, set k = k + 1
The parameters of the prediction model obtained from
system identification is tested to predict the output of d ahead
prediction as follows:
ˆy(t + d) = ϕT
(t + d)ˆℵ(ϕ(t + d), Ω) (32)
where, ˆℵ(ϕ(t + d), Ω) =
[ˆa(1,t+d · · · ˆa(ny,t+d)
ˆb(1,t+d) · · ·ˆb(nu,t+d)]T
and
ϕ(t + d) = [y(t + d − 1) y(t + d − 2) · · · y(t + d −
ny) u(t + d − 1) u(t + d − 2) · · · u(t + d − nu)]T
are SDPE
and the input vector of d step ahead prediction, respectively.
IV. SIMULATION STUDIES
In order to evaluate the performance of the proposed
method, a number of simulation studies are carried out for
identification and prediction of nonlinear systems. The plant
models are taken from the literature in order to complete a
performance comparison.
Fig. 2. Identification scheme using QARXNN model.
0 50 100 150 200 250 300
−1
−0.5
0
0.5
1
t
u(t)
0 50 100 150 200 250 300
−1
−0.5
0
0.5
1
t
y(t)
Fig. 3. System’s input-output of training data.
A. System Identification Example 1
System identification is used to find the relationship be-
tween the input and output of the system. The structure of
the system identification is shown in Fig. 2. The input of the
system modeling is the input u and output y composed with
time delay, and ˆy is the estimated output of QARXNN model.
The QARXNN model parameters are updated according to
estimation error e. A nonlinear dynamic system taken from
Narendra [20] is used in this example as follows:
y(t) = f(y(t − 1), y(t − 2), y(t − 3), u(t − 1), u(t − 2))
f(x1, x2, x3, x4, x5) =
x1x2x3x5(x3 − 1) + x4
1 + x2
2 + x2
3
, (33)
A number of input-output training data is obtained by testing
the system with a pseudo random binary sequence (PRBS)
function with the amplitude [-1.0;1.0]. The first 300 training
data are shown in Fig. 3. We set the model parameters by
ℵ(5,5,5,1) which demonstrates that the model contains five
elements of input ϕ(t) = [y(t−1)y(t−2)y(t−3)u(t−1)u(t−
2)]T
, five hidden nodes, five output nodes of the embedded
system and one output node. Thus, QARXNN model has 55
parameters.
An index of normalized prediction error (NPE) and root
mean square (RMS) error are introduced in order to measure
the performance of system identification and prediction is
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
0
2
4
6
8
10
−5
0
5
10
0
5
10
15
∆ee
√
e2+∆e2
Fig. 4. Learning pattern of QARXNN model with the proposed method.
0
0.01
0.02
0.03
0.04
0.05
−1
0
1
2
0.045
0.05
0.055
0.06
0.065
0.07
0.075
∆e
e
√
e2+∆e2
Fig. 5. Learning pattern of QARXNN model with original BP method.
stated as follows:
NPE = (
∑N
t=1(y(t) − ˆy(t))2
∑N
t=1(y(t))2
)1/2
x100%. (34)
RMS =
√
∑N
t=1(y(t) − ˆy(t))2
N
(35)
where ˆy(t), y(t), t = 1, 2, · · · , N are the estimated output,
the system output, and time. N denotes the length of training
data.
The learning pattern of the proposed method and the
original BP method have been illustrated in Fig. 4 and Fig.
5. As we can see in Fig. 4 and Fig. 5, the error and gradient
error of learning contribute in training the model in order
to obtain optimal weights, whereas BP only contributes by
learning error. According to the Lyapunov stability theorem,
the global minimum point of cost function can be guaranteed
if the learning error and the derivative of the error converge
to zero.
The model obtained from the system identification is tested
with the input-output of the training data to estimate next step
prediction shown in Fig. 6 and Fig. 7. The training data has
uniform distribution. Therefore, the NPE index prediction
using the proposed method is consistent to the specific value
of 21.7118 for all times. Compared with the BP method
that gives an average NPE=30.6449 with standard deviation
0.0389. Thus, the proposed method is able to ensure a stable
identification error.
0 50 100 150 200 250 300
−1
−0.5
0
0.5
1
t
y(t)andˆy(t)
(a)
0 50 100 150 200 250 300
−1
−0.5
0
0.5
1
t
y(t)andˆy(t)
(b)
Fig. 6. Output of one-step-ahead prediction (a) the proposed method. (b)
QARXNN-BP method.
0 200 400 600 800 1000 1200
21.7118
21.7118
21.7118
t
NPE
0 200 400 600 800 1000 1200
30.4
30.6
30.8
t
NPE
(b)
(a)
Fig. 7. NPE index of one-step-ahead prediction. (a) the proposed method.
(b) QARXNN-BP method.
After training, the test signal u(t) is fed to the model as
follows:
u(t) =



sin(2πt/250), if t ≤ 500
0.8 sin(2πt/250)
+0.2 sin(2πt/25), if t > 500
(36)
Fig. 8, Fig. 9 and Fig. 10 show the results of the simulation
using the tested signal. The accuracy of the proposed method
is 0.0297, whereas the BP method is 0.0792. Compared to
the previous work shown in Tab.I, it can be seen that the
proposed method performs better than the other methods.
TABLE I
SIMULATION RESULTS AND COMPARISON
Model RMS Number
Error of parameters
ARX model 0.0866 6
NN 0.0678 341
WN 3 0.0503 12
Q-ARX-NF 0.0478 85
Q-ARX-WN 0.0367 42
Q-ARX-NN < 0.01 246
Q-ARX-NN-ALLF 0.0297 55
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
0 100 200 300 400 500 600 700 800
−1
−0.5
0
0.5
1
t
y(t)andˆy(t)
(a)
0 100 200 300 400 500 600 700 800
−1
−0.5
0
0.5
1
t
y(t)andˆy(t)
(b)
y(t) ˆy(t)
y(t) ˆy(t)
Fig. 8. Simulation results with the deterministic signal (a) the proposed
method. (b) QARXNN-BP method
0 100 200 300 400 500 600 700 800
−0.3
−0.2
−0.1
0
0.1
0.2
t
y(t)-ˆy(t)
Lyapunov BP
Fig. 9. The error of simulation tested with deterministic signal.
B. System Identification Example 2
This example considers the modeling of the nonlinear
system to be as follows:
y(t) = 0.72y(t − 1) + 0.025y(t − 2)u(t − 2)
+ 0.01u2
(t − 3) + 0.2u(t − 4). (37)
The same mathematical model of dynamic systems is also
used in [21], [22], [23]. The current output of the system
depends on two previous output values and three previous
input values. However, only two values, y(t−1) and u(t−1),
0 100 200 300 400 500 600 700 800
0
0.02
0.04
0.06
0.08
0.1
t
RMS
Lyapunov
BP
Fig. 10. RMS errors of the simulation tested with deterministic signal.
0 20 40 60 80 100 120 140 160 180 200
0
0.05
0.1
0.15
0.2
0.25
epoch
RMS
Fig. 11. RMS values obtained during training.
0 100 200 300 400 500 600 700 800 900 1000
−1
−0.5
0
0.5
1
t
y(t),ˆy(t)ande(t)
y(t) ˆy(t) e(t)
Fig. 12. The results of identification
are fed to the QARXNN model in order to compare it to other
models. The length of input of training data is 900. Half of
the inputs are independently and identically distributed (i.i.d.)
in a uniform sequence over [-2,2], and the remaining are
sinusoid generated by 1.05 sin(πt/45). QARXNN model is
trained for 200 epochs, as shown in Fig. 11. After training,
the following same test signal u(t) is used for testing the
performance of the QARXNN model:
u(t) =



sin(πt/25), if t < 250
1.0, if 250 ≤ t < 500
−1.0, if 500 ≤ t < 750
0.3 sin(πt/25)
+0.1 sin(πt/32)
+0.6 sin(πt/10), if 750 ≤ t < 1000
(38)
The parameters of the model are denoted as ℵ(2,4,2,1),
which shows the model contains two elements of input, four
hidden nodes, two output nodes of the embedded system and
one output node . Thus, QARXNN model has 20 parameters.
Fig. 12 compares the actual output of the system with that of
the QARXNN model. The RMS values of the tested model
are shown in Fig. 13. A detailed performance comparison
of the QARXNN model is given in Tab.II.
TABLE II
COMPARISON OF QARXNN MODEL WITH OTHER MODELS
Models Number RMS RMS
of parameters train test
RSONFIN [24] 49 0.03 0.06
TRFN-S [21] 33 0.0067 0.0313
FWNN [23] 27 0.01973 0.022609
FWNN [23] 43 0.018713 0.020169
AWN [22] 12 0.009368 0.022933
AWN [22] 20 0.009391 0.023259
Q-ARX-NN-ALLF 20 0.01637 0.01999
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
0 100 200 300 400 500 600 700 800 900 1000
0.01
0.015
0.02
0.025
t
RMS
Fig. 13. RMS values of tested model
0 100 200 300 400 500 600 700 800 900 1000
0
5
10
15
20
25
epoch
RMS
Fig. 14. RMS values obtained during training.
C. Prediction of nonlinear system
This example considers a nonlinear second order dynami-
cal model that has been used in [20], [25], [16]. The process
is described by the following equation:
y(t) = 0.3y(t − 1) + 0.6y(t − 2) + f[u(t)] (39)
where
f[u(t)] = u3
(t) + 0.3u2
(t) − 0.4u(t) (40)
u(t) = sin(2πt/250) + sin(2πt/25). (41)
The function ′
f ′
is a polynomial with three degree of current
input signal u(t). The 500 input-output data is generated.
Three hundred data are used in training procedure and the
remaining 200 data-points are for prediction. The network
parameters of QARXNN model used is ℵ(3,3,3,1). The model
contains twenty one parameters with three elements of input,
three hidden nodes, three output nodes of embedded system
and one output node.
The models are trained with 1000 epochs, as shown in
Fig. 14. The results of the system identification in which
the model is tested with training data is shown in Fig. 15,
which also shows that the RMS value is 0.3254. Fig. 16
contains the results of prediction when the adaptation of the
trained model stops after 300 samplings. As we can see, the
proposed model is able to predict a nonlinear system where
the RMS index for prediction is 0.343.
V. DISCUSSION AND CONCLUSION
This paper discusses two issues: 1) a different approach
to the hierarchical algorithm of linear and nonlinear sub-
models of the quasi-linear ARX neural network is introduced.
First, the system is identified using a linear estimator by LSE
algorithm. Second, the residual error of the linear sub-model
is set as a target output performed using NN. 2) In order
to improve the performance of system identification and
0 50 100 150 200 250 300
−40
−20
0
20
40
t
y(t)andˆy(t)
0 50 100 150 200 250 300
−1
0
1
2
t
e(t)
0 50 100 150 200 250 300
0
0.1
0.2
0.3
0.4
0.5
t
RMS
Fig. 15. Testing the model with training data.
0 20 40 60 80 100 120 140 160 180 200
−40
−20
0
20
40
t
y(t)andˆy(t)
0 20 40 60 80 100 120 140 160 180 200
−1
0
1
2
t
e(t)
0 20 40 60 80 100 120 140 160 180 200
0.2
0.4
0.6
0.8
1
t
e(t)
Fig. 16. Prediction output after training model stops at t = 300.
prediction, we propose using the adaptive learning performed
based on Lyapunov stability theorem. The fixed learning rate
of BP is replaced by the adaptive learning rate based on Lya-
punov stability theory. By using a Lyapunov function-based
adaptive learning approach to the system embedded in the
QARXNN model, the proposed method is able to improve
the performance of system identification and prediction.
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________
REFERENCES
[1] J. Hu, K. Kumamaru and K. Hirasawa, ”A Quasi-ARMAX approach
to modelling of nonlinear systems,” Int. J. Control, vol. 74, no. 18, pp
1754-1766, 2001
[2] E. Kamal, A. Aitouche, R. Ghorbani and M. Bayart, ”Robust nonlinear
control of wind energy conversion systems,” Int. J. Electr. Power and
Energy Syst., vol. 44, pp 202-209, 2013
[3] Y. Cheng, L. Wang and J. Hu, ”Identification of Quasi-ARX neuro-
fuzzy model with an SVR and GA approach,” IEICE Trans. Funda-
mentals,vol. E.95-A, no. 5, pp 876-883, 2012
[4] S. Hassan, J. Jaafar, B.S. Brahim and A.J. Tahseen, ”A Hybrid Fuzzy
Time Series Model for Forecasting,” Engineering Letters, vol. 20, no.
1, pp 88-93, 2012
[5] G. P. Zhang, ”Time series forecasting using a hybrid ARMA and neural
network model,” Neurocomputing, vol. 50, pp 159-175, 2003
[6] Kuan-Heng Chen and Khaldoun Khashanah, ”Analysis of Systemic
Risk: A Vine Copula-based ARMA-GARCH Model,” Engineering
Letters, vol. 24, no. 3, pp 268-273, 2016
[7] M.A. Jami’in, I. Sutrisno and J. Hu, ”Nonlinear Adaptive Control for
Wind Energy Conversion Systems Based on Quasi-ARX Neural Net-
works Model,” Lecture Notes in Engineering and Computer Science:
Proceedings of The International MultiConference of Engineers and
Computer Scientists 2014, IMECS 2014, 12-14 March, 2014, Hong
Kong, pp 313-318.
[8] M.A. Jami‘in, I Sutrisno, J. Hu, N.B. Mariun and M.H. Marhaban,
”An adaptive predictive control based on a quasi-ARX neural network
model,” Proc. 13th International Conference on Control, Automation,
Robotics and Vision, pp 253-258, Dec. 2014
[9] M.A. Jami‘in, I Sutrisnoa, J. Hu, N.B. Mariun and M.H. Marhaban,
”Quasi-ARX Neural Network Based Adaptive Predictive Control for
Nonlinear Systems,” IEEJ Transaction on Electrical and Electronic
Engineering, vol. 11, no. 1, pp 83-90, 2016
[10] Y. Cheng, L. Wang and J. Hu, ”Quasi-ARX wavelet network for
SVR based nonlinear system identification,” Nonlinear Theory and
its Applications (NOLTA), IEICE, vol. 2, no. 2, pp 165-179, 2011
[11] W Yu, A.S. Poznyak, X. Li, ”Multilayer dynamic neural networks for
non-linear system on-line identification,” Int. J. Control, vol. 74, no.
18, pp 1858-1864, 2001
[12] L. Bahera, S. Kumar and A. Patnaik, ”On adaptive learning rate that
guarantees convergence in feedforward networks,” IEEE Transaction
on Neural Networks, vol. 17, no. 5, pp 1116-1125, 2006
[13] Z. Man, H.R. Wu, S. Liu and X. Yu, ”A new adaptive backpropagation
algorithm based on lyapunov stability theory for neural networks,”
IEEE Transaction on Neural Networks, vol. 17, no. 6, pp 1580-1591,
2006
[14] V. Becera, F Garces, S. Nasuto and W Holderbaum, ”An efficient
parameterization of dynamic neural networks for nonlinear system
identification,” IEEE Transaction on Neural Networks, vol. 16, no.
4, pp 983-988, 2005
[15] F. Abdollahi, H.A. Talebi and R.V. Patel, ”Stable identification of
nonlinear systems using neural networks: Theory and experiments,”
IEEE/ASME Trans. on Mechatronics, vol. 11, no. 4, pp 488-495, 2006
[16] M.A. Shoorehdeli, M. Teshnehlab and A.K. Sedigh, ”Identification
using ANFIS with intelligent hybrid stable learning algorithm ap-
proaches,” Neural Computing and Applications, vol. 18, pp 157-174,
2009
[17] M.A. Jami’in, I. Sutrisno and J. Hu, ”Deep searching for parameter
estimation of the linear time invariant (LTI) system by using quasi-
ARX neural network,” Proc. IEEE International Joint Conference on
Neural Network, pp 2758-2762, 2013
[18] M.A. Jamiin, J. Hu and E. Julianto, ”A Lyapunov Based Switching
Control to Track Maximum Power Point of WECS,” Proc. IEEE
International Joint Conference on Neural Network, pp 3883-3888,
2016
[19] J. Hu and K. Hirasawa, A method for applying neural networks to
control of nonlinear systems, in book entitled Neural Information
Processing: Research and Development, Springer, Berlin, Germany,
pp 351-369, 2004
[20] K.S. Narendra and K. Parthasarathy, ”Identification and control of
dynamical systems using neural networks,” IEEE Trans. on Neural
Networks, vol. 1, no. 1, pp 4-27, 1990
[21] C.F. Juang, ”A TSK-type recurrent fuzzy network for dynamic sys-
tems processing by neural network and genetic algorithms,” IEEE
Trans.Fuzzy Syst., vol. 10, no. 2, pp 155-170, 2002
[22] Y. Oysal and S Yilmaz, ”An adaptive wavelet network for function
learning,” Neural Computing and Applications, vol. 19, pp 383-392,
2010
[23] R.H. Abiyev, O. Kaynak, ”Fuzzy wavelet neural networks for iden-
tification and control of dynamic plants a novel structure and a
comparative study,” IEEE Trans. Ind. Electron, vol. 55, no. 8, pp 3133-
3140, 2008
[24] C.F. Juang and C.T. Lin, ”A recurrent self-organizing neural fuzy
inference network,” IEEE Trans. Neural Network, vol. 10, no. 4, pp
828-845, 1999
[25] A. Banakar, M.F. Azeem, ”Artificial wavelet neural network and its
application in neuro-fuzzy models,” Appl. Soft Comput., vol. 8, no. 4,
pp 1463-1485, 2008
Mohammad Abu Jami’in received the M.E. degree in Control Engineering
from Institut Teknologi Sepuluh Nopember (ITS) Surabaya, Indonesia in
2008, and the Doctor of Engineering in Neurocomputing from Waseda
University, Japan in 2016. He is currently a Lecturer with the Politeknik
Perkapalan Negeri Surabaya, Indonesia. His research interests include arti-
ficial intelligence and its applications.
Yuyun received the B.E. degree in Computer Engineering STMIK
Dipanegara Makassar and the M.E. degree in Electrical Engineering from
Hasanuddin University Makassar, Sulawesi Selatan, Indonesia in 2010 and
2013. He is currently a lecture in STMIK Handayani Makassar, Indonesia.
His research interests include Computer Engineering.
Eko Julianto received the B.E. degree in Marine Engineering from Institut
Teknologi Sepuluh Nopember (ITS) Surabaya, Indonesia in 1989, and the
M.E degree in Marine Engineering from University Of New Castle Upon
Tyne, UK, 1994. He is currently a Lecturer with the Politeknik Perkapalan
Negeri Surabaya, Indonesia. His research interests include Marine Engineer-
ing.
Engineering Letters, 25:3, EL_25_3_14
(Advance online publication: 23 August 2017)
______________________________________________________________________________________

More Related Content

PDF
Fault detection based on novel fuzzy modelling
PDF
Single Channel Speech De-noising Using Kernel Independent Component Analysis...
PDF
Cone crusher model identification using
PDF
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
PDF
Automatic time series forecasting using nonlinear autoregressive neural netwo...
PDF
Partial stabilization based guidance
PDF
Analysis of intelligent system design by neuro adaptive control
PDF
Analysis of intelligent system design by neuro adaptive control no restriction
Fault detection based on novel fuzzy modelling
Single Channel Speech De-noising Using Kernel Independent Component Analysis...
Cone crusher model identification using
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
Automatic time series forecasting using nonlinear autoregressive neural netwo...
Partial stabilization based guidance
Analysis of intelligent system design by neuro adaptive control
Analysis of intelligent system design by neuro adaptive control no restriction

What's hot (18)

PDF
Ag04606202206
PDF
Performance Analysis of Various Activation Functions in Generalized MLP Archi...
PDF
High performance extreme learning machines a complete toolbox for big data a...
PDF
H046014853
PPT
Decentralized Data Fusion Algorithm using Factor Analysis Model
PDF
40120130406002
PDF
On selection of periodic kernels parameters in time series prediction
PDF
solver (1)
PDF
Modified Projective Synchronization of Chaotic Systems with Noise Disturbance,...
PDF
Full parameterization process
PDF
N41049093
PDF
Tutorial on Markov Random Fields (MRFs) for Computer Vision Applications
PDF
Investigations on Hybrid Learning in ANFIS
PDF
Controller Design Based On Fuzzy Observers For T-S Fuzzy Bilinear Models
PDF
Inference & Learning in Linear Chain Conditional Random Fields (CRFs)
PDF
Adaptive network based fuzzy
PDF
F017533540
PDF
Black-box modeling of nonlinear system using evolutionary neural NARX model
Ag04606202206
Performance Analysis of Various Activation Functions in Generalized MLP Archi...
High performance extreme learning machines a complete toolbox for big data a...
H046014853
Decentralized Data Fusion Algorithm using Factor Analysis Model
40120130406002
On selection of periodic kernels parameters in time series prediction
solver (1)
Modified Projective Synchronization of Chaotic Systems with Noise Disturbance,...
Full parameterization process
N41049093
Tutorial on Markov Random Fields (MRFs) for Computer Vision Applications
Investigations on Hybrid Learning in ANFIS
Controller Design Based On Fuzzy Observers For T-S Fuzzy Bilinear Models
Inference & Learning in Linear Chain Conditional Random Fields (CRFs)
Adaptive network based fuzzy
F017533540
Black-box modeling of nonlinear system using evolutionary neural NARX model
Ad

Similar to Hierarchical algorithms of quasi linear ARX Neural Networks for Identification of Nonlinear Systems (20)

PDF
Cone Crusher Model Identification Using Block-Oriented Systems with Orthonorm...
PDF
Solution of Inverse Kinematics for SCARA Manipulator Using Adaptive Neuro-Fuz...
PDF
Power system transient stability margin estimation using artificial neural ne...
PDF
On Selection of Periodic Kernels Parameters in Time Series Prediction
PDF
ON SELECTION OF PERIODIC KERNELS PARAMETERS IN TIME SERIES PREDICTION
PDF
Investigation of auto-oscilational regimes of the system by dynamic nonlinear...
PDF
Adaptive modified backpropagation algorithm based on differential errors
PDF
DYNAMICS, ADAPTIVE CONTROL AND EXTENDED SYNCHRONIZATION OF HYPERCHAOTIC SYSTE...
PDF
DYNAMICS, ADAPTIVE CONTROL AND EXTENDED SYNCHRONIZATION OF HYPERCHAOTIC SYSTE...
PDF
Direct digital control scheme for controlling hybrid dynamic systems using AN...
PDF
Simulation and stability analysis of neural network based control scheme for ...
PDF
PDF
Ab35157161
PDF
Ab35157161
PDF
Identification and Control of Three-Links Electrically Driven Robot Arm Using...
PDF
State Estimation based Inverse Dynamic Controller for Hybrid system using Art...
PDF
Design of airfoil using backpropagation training with mixed approach
PDF
Design of airfoil using backpropagation training with mixed approach
PDF
Optimal neural network models for wind speed prediction
PDF
Optimal neural network models for wind speed prediction
Cone Crusher Model Identification Using Block-Oriented Systems with Orthonorm...
Solution of Inverse Kinematics for SCARA Manipulator Using Adaptive Neuro-Fuz...
Power system transient stability margin estimation using artificial neural ne...
On Selection of Periodic Kernels Parameters in Time Series Prediction
ON SELECTION OF PERIODIC KERNELS PARAMETERS IN TIME SERIES PREDICTION
Investigation of auto-oscilational regimes of the system by dynamic nonlinear...
Adaptive modified backpropagation algorithm based on differential errors
DYNAMICS, ADAPTIVE CONTROL AND EXTENDED SYNCHRONIZATION OF HYPERCHAOTIC SYSTE...
DYNAMICS, ADAPTIVE CONTROL AND EXTENDED SYNCHRONIZATION OF HYPERCHAOTIC SYSTE...
Direct digital control scheme for controlling hybrid dynamic systems using AN...
Simulation and stability analysis of neural network based control scheme for ...
Ab35157161
Ab35157161
Identification and Control of Three-Links Electrically Driven Robot Arm Using...
State Estimation based Inverse Dynamic Controller for Hybrid system using Art...
Design of airfoil using backpropagation training with mixed approach
Design of airfoil using backpropagation training with mixed approach
Optimal neural network models for wind speed prediction
Optimal neural network models for wind speed prediction
Ad

More from Yuyun Wabula (7)

PDF
DATA MINING UNTUK PENENTUAN MODEL TINGKAT KESUKSESAN KELULUSAN MURID SMA PAD...
PDF
Dynamic Land-Use Map Based on Twitter Data
PDF
Location Based Service in Social Media: An Overview of Application
PDF
Using Social Networking Data to Understand Urban Human Mobility
PDF
Yuyun, Neural Network Algorithm for Determining a Majors for Senior High Scho...
PDF
The synchronization of database at the faculty of engineering and the rectora...
PDF
Fortei Hasanuddin university (unhas) 2011
DATA MINING UNTUK PENENTUAN MODEL TINGKAT KESUKSESAN KELULUSAN MURID SMA PAD...
Dynamic Land-Use Map Based on Twitter Data
Location Based Service in Social Media: An Overview of Application
Using Social Networking Data to Understand Urban Human Mobility
Yuyun, Neural Network Algorithm for Determining a Majors for Senior High Scho...
The synchronization of database at the faculty of engineering and the rectora...
Fortei Hasanuddin university (unhas) 2011

Recently uploaded (20)

PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PPTX
additive manufacturing of ss316l using mig welding
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PPTX
Sustainable Sites - Green Building Construction
PPTX
bas. eng. economics group 4 presentation 1.pptx
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PDF
Digital Logic Computer Design lecture notes
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PPTX
Internet of Things (IOT) - A guide to understanding
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PPTX
Construction Project Organization Group 2.pptx
PPTX
CH1 Production IntroductoryConcepts.pptx
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
Embodied AI: Ushering in the Next Era of Intelligent Systems
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
additive manufacturing of ss316l using mig welding
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
Automation-in-Manufacturing-Chapter-Introduction.pdf
Sustainable Sites - Green Building Construction
bas. eng. economics group 4 presentation 1.pptx
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
Digital Logic Computer Design lecture notes
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Internet of Things (IOT) - A guide to understanding
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
Construction Project Organization Group 2.pptx
CH1 Production IntroductoryConcepts.pptx
Operating System & Kernel Study Guide-1 - converted.pdf
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx

Hierarchical algorithms of quasi linear ARX Neural Networks for Identification of Nonlinear Systems

  • 1. Hierarchical Algorithms of Quasi-Linear ARX Neural Networks for Identification of Nonlinear Systems Mohammad Abu Jami’in, Member, IAENG, Yuyun, Member, IAENG, Eko Julianto, Member, IAENG Abstract—A quasi-linear ARX neural network model (QARXNN) is a nonlinear model built using neural networks (NN). It has a linear-ARX structure where NN is an embedded system to give the parameters for the regression vector. There are two contributions in this paper, 1) Hierarchical Algorithms is proposed for the training of QARXNN model, 2) an adaptive learning is implemented to update learning rate in NN training to ensure global minima. First, the system is estimated by using LSE algorithm. Second, nonlinear sub-model performed using NN is trained to refine error of LSE algorithm. The linear parameters obtained from LSE algorithm is set as bias vector for the output nodes of NN. Global minima point is reached by adjusting the learning rate based on Lyapunov stability theory to ensure convergence of error. The proposed algorithm is used for the identification and prediction of nonlinear dynamic systems. Finally, the experiments and numerical simulations reveal that the proposed method gives satisfactory results. Index Terms—System Identification, quasi linear-ARX neural network, hierarchical algorithm, convergence of error, nonlin- ear system I. INTRODUCTION SYSTEM identification is an extensive problem because it is difficult to propose a model capable of describing efficiently in every possible nonlinear system. Neural net- works (NN) and neuro-fuzzy networks are a type of black- box models with nonparametric or multi-parametric models [1]. The Black-box models for the identification of nonlinear systems do not require parametric expressions. The most useful property of NN is its ability to approximate arbitrary linear or nonlinear mapping through training. However, there are three major drawbacks: 1) these methods take the system as a whole and do not permit separate analysis and identi- fication of the linear dynamics, 2) it is difficult to use their parameters for controller design, and 3) the stability analysis is not easy, and the parameter tuning is generally a time- consuming process due to its nonlinear and multi-parametric nature [2], [3]. Some researchers adopted hybrid approach by mixing linear and nonlinear models to improve accuracy, simplicity and reducing computational complexity [4], [1], [5], [6]. A hybrid fuzzy time series model is constructed based on ARIMA and Interval Type-2 Fuzzy Inference System was proposed to improve forecasting accuracy [4]. In this Manuscript received 05 April, 2017; revised 19 July, 2017. Mohammad Abu Jami’in is with the Department of Automation Engi- neering, Politeknik Perkapalan Negeri Surabaya, Surabaya, Jawa Timur, Indonesia. e-mail: [email protected]. Yuyun is with the Department of Computer Engineering, STMIK Han- dayani, Makassar, Sulawesi Selatan, Indonesia. Eko Julianto is with the Department of Piping Engineering, Politeknik Perkapalan Negeri Surabaya, Surabaya, Jawa Timur, Indonesia. paper, we propose a quasi-linear ARX neural network model (QARXNN). It has an easy-to-use structure shown by its ARX-like structure. Nonlinear coefficients are implemented by NN to parameterize the input vector. It is injected to the QARXNN model as an embedded system [1]. This is the interesting aspect of the structure that the coefficients have some meaning to describe the dynamic behavior of the identified plants. The parameters of system modeling can be used to estimate the stability of the modeled system and can be set as controller parameters [8], [9]. The coefficients consist of linear and nonlinear parts that allow for system analysis using the approach of linear theory. In the previous research, the embedded system of the QARXNN model was modified by using neuro-fuzzy, wavelet, radial-basis function, and multilayer perceptron (MLP) networks in order to increase the accuracy of model for system identification [10]. However, this work does not discuss the stable and global convergence of system identification. In this paper, a hierarchical algorithm was proposed for the training of QARXNN model based on Lyapunov stability theory. The learning rate is updated every leaning step in order to guarantee the convergence of error. Thus, the error will be asymptotically goes to zero by the number of training time. Lyapunov stability is a well-known theory for control application to guarantee closed-loop control stability. A lyapunov-like analysis was proposed to derive stable learning law in order to ensure the stability property of the iden- tification error [11]. Using the Lyapunov stability theorem for the training algorithm allows for increasing the speed of the convergence in the learning process, guaranteeing global convergence of the model, and avoiding local minima and achieving global minima of cost function [12], [13]. The Lyapunov function has been used to track convergence of error; the output of tracking error can then asymptotically converge to zero [13] and can be used to analyze the stability of the identification process with stable learning law [14], [15], [16]. In this paper, the training of QARXNN model is performed hierarchically between linear and nonlinear sub-models based on Lyapunov stability theorem. First, the system is estimated using a linear sub-model under LSE algorithm. From this step, we have linear part parameters. Second, the residual errors of LSE algorithm (linear sub-model) are refined by nonlinear sub-model performed using NN. The parameters of linear part is set as bias vector for the output nodes of NN, where the residual errors of linear sub-model is set as the target output to train NN sub-model. Thus, the errors of the linear sub-model is refined by the training of nonlinear Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 2. sub-model. From the training of NN, we have nonlinear part parameters. Thus, by summing between the linear and nonlinear parts parameters, we have nonlinear parameters. If the system is linear then the coefficients converge to the specific value of the constant, and if the system is nonlinear then the coefficients will be a function of time [17]. Especially in control applications, this approach allow us to derive the linear or nonlinear controllers based on the linear or nonlinear parameters. A simple switching mechanism is introduced between the linear and nonlinear parts parameters in order to derive linear and nonlinear controls. [8], [18], [7]. The rest of the paper is organized as follows: Section 2 contains a brief discussion of the QARXNN model. In Section 3, the proposed hierarchical learning algorithm is derived with linear and nonlinear sub-models. Convergence analysis of the proposed algorithm is also described in this section. Experimental results are presented in Section 4, and finally conclusions are recorded in Section 5. II. QUASI-LINEAR ARX NEURAL NETWORK MODEL Consider a single-input-single-output (SISO) black-box nonlinear system whose input-output relation is described by y(t) = g(ϕ(t)) (1) where, g(·), y(t) ∈ R, and ϕ(t) = [y(t−1) · · · y(t−ny)u(t− 1) · · · u(t−nu)]T are nonlinear function, system output, and a regression vector, respectively. The general nonlinear system expressed in (1) can be represented in a regression which has been shown in Ref.[1,2]. The unknown nonlinear function g(ϕ(t)) can be performed using Taylor expansion on a small region of around ϕ(t) = 0: y(t) = g(0) + ˙g(0)ϕ(t) + 1 2 ϕT (t)¨g(0)ϕ(t) + · · · . (2) As (2) shows, the coefficients are infinite. However, Taylor expansion is only used as a bridge to transform the nonlinear system. Applying the Taylor series transformation and the system dynamics, the nonlinear system can be expressed as [1]: y(t) = y(0) + ϕ(t)T Ω(ϕ(t)). (3) where, y(0) = g(0) is the value for the initial condition, and Ω(ϕ(t)) is the coefficient of the regression vector by Ω(ϕ(t)) = [˙g(0)ϕ(t) + 1 2 ϕT (t)¨g(0)ϕ(t) + · · ·]T . (4) = [a1,t · · · anu,t b0,t · · · bny−1,t] (5) where, ai,t = ai(ϕ(t)), (i = 1, · · · , ny) is the coefficient of the regression vector of output and bj,t = bj(ϕ(t)), (j = 0, · · · , nu − 1) is the coefficient of the regression vector of input. The orders of time delay for the system input-output are nu and ny. The Quasi-linear ARX neural network is a flexible non- linear model that has linear properties [19]. Fig. 1 shows the structure of quasi-linear ARX neural network model. Selecting NN as an embedded system, the model can be written as follows: y(t, ϕ(t)) = ϕ(t)T ℵ(ϕ(t), Ω) (6) = ϕ(t)T W2ΓW1(ϕ(t)) + ϕ(t)T θ, (7) Fig. 1. A quasi-linear ARX neural network model. where Ω = {W1, W2, θ} sets the network parameters. The network parameter is Γ, W1, W2, and θ, being the diag- onal nonlinear operator with identical sigmoidal elements on hidden nodes, the weights of hidden layer, the weights of output layer, and the bias vector for the output nodes, respectively. The coefficients or state dependent parameter estimation (SDPE), denoted by ℵ(ϕ(t), Ω) is implemented using NN stated as ℵ(ϕ(t), Ω) = W2ΓW1(ϕ(t)) + θ (8) = δ(ϕ(t), W) + θ. (9) As this demonstrates, when compared to conventional feed- forward neural networks, the quasi-linear ARX neural net- work has an easy-to-use structure and introduces efficient algorithms for the parameter estimation as shown by (7) and (8). The quasi-linear ARX neural networks model consists of two parts: the second term of the right side of (7) shows the linear part, while the first term is the nonlinear part. The coefficients of the linear and nonlinear parts are fused to parameterize the input vector. Error in the linear and nonlinear sub-models is shared. In one step, both linear and nonlinear parts are updated simultaneously. The estimated parameter of the linear part changes at every learning step. Thus, it is difficult to analyze the system identified linearly using the linear part. III. HIERARCHICAL ALGORITHM WITH ADAPTIVE LEARNING FOR QARXNN MODEL In this paper, we use a different approach to describe the QARXNN model by separating the linear and nonlinear sub- models. First, the linear sub-model is used to identify the system to obtain the linear parameter estimate. The linear estimator allows for the calculation of the linear parameter estimation and the residual error of the model. This feature allow for analyzing the identified plant as a linear system. Second, the residual error of the linear sub-model is used to train NN. NN is a compensator to refine the error of the linear sub-model. The proposed hierarchical algorithm fixed the linear parameter estimator, which does not change in any training step of NN. Interestingly, this method can be used to describe the system linearly and nonlinearly via the switching mechanism [8]. In order to perform the hierarchical Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 3. algorithm, the Equation (7) can be rewritten as: y(t, ϕ(t)) = zl + zn (10) zl = ϕT (t)θ (11) zn = ϕT (t)δ(ϕ(t), W) (12) where, zl is the linear estimator (linear sub-model) and zn is the nonlinear estimator (nonlinear sub-model). In our main theoretical result, the following assumption are made: A1. The pairs of the input-output training data are bounded. A2. The SDPE ℵ(ϕ(t), Ω) is bounded. A3. There is an optimal weight of SDPE denoted by ℵ∗ (ϕ(t), Ω). By using the QARXNN model, the estimated output of (7) can be rewritten in the following form: ˆy(t) = ϕT (t)ˆℵ(ϕ(t), Ω) (13) if there exists ℵ∗ (ϕ(t), Ω) such that y(t) = ϕT (t)ℵ∗ (ϕ(t), Ω) (14) Using (13) and (14), the adaptation error e(t) is derived as follows: e(t) = y(t) − ˆy(t) = ϕT (t)ℵ∗ (ϕ(t), Ω) − ϕT (t)ˆℵ(ϕ(t), Ω) = ϕT (t)Ξ (15) Ξ = ℵ∗ (ϕ(t), Ω) − ˆℵ(ϕ(t), Ω). (16) where Ξ is the adaptation error of SDPE. Suppose that the assumption A2. is valid, then there is a positive real number α for ∥ e(t) ∥< α. The accuracy and convergence of error characteristics are then greatly influenced by the adaptation error Ξ. The adaptation error is then distributed to the two sub- models by el = y(t) − ϕT (t)ˆθ (17) en = y(t) − ϕT (t)ˆδ(ϕ(t), W). (18) where ˆθ denotes the linear part-parameter estimator and ˆδ(ϕ(t), W) denotes nonlinear part-parameter estimator. The residual error of the linear and nonlinear sub-models are el and en, respectively. The linear part-parameter estimator is the parameter for the linear system. Therefore, it is possible to analyze the system in a linear fashion. Hereafter, el can be set as the target output for the nonlinear sub-model zn. By introducing two linear and nonlinear sub-models, equation (15) can be rewritten in the following form: e(t) = y(t) − ˆy(t) = y(t) − ϕT (t)ˆℵ(ϕ(t), Ω) = y(t) − ϕT (t)(ˆδ(ϕ(t), W) + ˆθ) = y(t) − ϕT (t)ˆδ(ϕ(t), W) − ϕT (t)ˆθ = el − zn (19) = en − zl. (20) The desired goal of system identification is to make the adaptation error e(t) reach and remain at zero. Two algorithms are implemented hierarchically. First, the system is estimated using the linear sub-model under the LSE algorithm to obtain linear parameter estimator θ. Second, the residual error of the linear sub-model is calculated by (17) and set as the target output for the nonlinear sub- model performed by a multilayer perceptron neural network (MLPNN). Because the linear parameter estimator is fixed and set as a bias vector for the output nodes of NN, sub- model zn is estimated by (12) until its output converge to el. By (19), e(t) will converge to zero if the output of nonlinear sub-model zn converges to el. Thus, the nonlinear sub-model becomes a booster to refine the estimator of the linear sub- model. To train the NN for the nonlinear sub-model, adaptive learning is proposed for training similar to a BP algorithm. The fixed learning rate of BP is replaced by the adaptive learning rate performed based on the Lyapunov stability the- orem. The selected Lyapunov function is defined as follows: V = 1 2 e2 (k) (21) where e(k) = {e1 l − z1 n, e2 l − z2 n, · · · , eN l − zN n }T , N is the number of training data, and k is the learning sequence. Performing the derivative of (21) in discrete form, we have ∆V (k) = V (k + 1) − V (k) = 1 2 e2 (k + 1) − e2 (k) (22) By introducing the increment of error, e(k + 1) = e(k) + ∆e(k) (23) Substituting (23) in (22), we have ∆V (k) = 1 2 ( (e(k) + ∆e(k))2 − e2 (k) ) = 1 2 (∆e(k))2 + e(k)∆e(k) (24) The parameters of weight W are updated using the gradi- ent descent algorithm that may be defined as W(k + 1) = W(k) + ∆W = W(k) + η ( − ∂V ∂W ) = W(k) + η ( ∂zn ∂W ) e(k). (25) Thus, ∆W can be rewritten as ∆W = η ( ∂zn ∂W ) e(k) (26) For learning, the difference of error is due to the change during learning, which can be expressed as ∆e(k) ≈ ∂e ∂W ≈ − ( ∂zn ∂W )T ∆W (27) Let ∂zn ∂W = H. By using (26) and (27) in (24), we have ∆V (k) = (−H∆W)2 + e(k)(−H∆W(k)) = H2 (ηHe(k))2 + e(k)(−HηHe(k)) = η2 H4 e2 (k) − ηH2 e2 (k) (28) According to the Lyapunov stability theorem, the condition of stability is dependent on a negative ∆V (k). Therefore, the trajectory of ∆V (k) is as follows: ∆V (k) = −λe2 (k) (29) Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 4. Substituting (28) and (29), we have λ = ηH2 − η2 H4 (30) Based on the Lyapunov stability theorem, the stability is ensured if V (k) ) is positive-definite and if ∆V (k) is negative-definite. Therefore, the learning rate η is adapted to guarantee a global convergence condition, which can be satisfied if λ > 0. Therefore, the learning rate is adapted to satisfy the condition as follows: η ( ∂zn ∂W )2 < 1 (31) The Lyapunov function is selected with the positive def- inite function that represents the energy function. In each learning step, the parameters of the network is updated based on reducing the energy, and the energy is always dissipated. The global convergence point of cost function is ensured if the energy and dissipated energy converge to zero. This can be shown by both as the error and the gradient of error converge to zero through training. To guarantee global convergence under Lyapunov stability theory, the derivative of the Lyapunov function is ensured in that negative definite function. The detailed steps to train a QARXNN model based on Lyapunov theorem is described as follows: 1) Set y(t) as zl then estimate θ using LSE algorithm of (11). 2) Calculate el using (17) and set as output guide to train sub-model zn in (12). 3) Use ˆθ as bias vector for output nodes of MLPNN, and small initial values of W1(k) and W2(k). Update W1(k), W2(k) by using adaptive learning based on Lyapunov function (ALLF). Set k = 1, k is the sequence of learning number. 4) Select Lyapunov function candidate, the candidate function is stated as V = f(e), where V = 0 only if e = 0, V > 0 only if e ̸= 0, e = el − zn. 5) Update the learning rate based on energy function by ˙V < 0. Based on Lyapunov theory, if V > 0 and ˙V < 0, then the error will converge to zero at time goes to infinity limk→∞ e(k) = 0. 6) Update the weights using (25) and learning condition satisfy (31) 7) Stop if pre-specified condition is met, otherwise goto step 3, set k = k + 1 The parameters of the prediction model obtained from system identification is tested to predict the output of d ahead prediction as follows: ˆy(t + d) = ϕT (t + d)ˆℵ(ϕ(t + d), Ω) (32) where, ˆℵ(ϕ(t + d), Ω) = [ˆa(1,t+d · · · ˆa(ny,t+d) ˆb(1,t+d) · · ·ˆb(nu,t+d)]T and ϕ(t + d) = [y(t + d − 1) y(t + d − 2) · · · y(t + d − ny) u(t + d − 1) u(t + d − 2) · · · u(t + d − nu)]T are SDPE and the input vector of d step ahead prediction, respectively. IV. SIMULATION STUDIES In order to evaluate the performance of the proposed method, a number of simulation studies are carried out for identification and prediction of nonlinear systems. The plant models are taken from the literature in order to complete a performance comparison. Fig. 2. Identification scheme using QARXNN model. 0 50 100 150 200 250 300 −1 −0.5 0 0.5 1 t u(t) 0 50 100 150 200 250 300 −1 −0.5 0 0.5 1 t y(t) Fig. 3. System’s input-output of training data. A. System Identification Example 1 System identification is used to find the relationship be- tween the input and output of the system. The structure of the system identification is shown in Fig. 2. The input of the system modeling is the input u and output y composed with time delay, and ˆy is the estimated output of QARXNN model. The QARXNN model parameters are updated according to estimation error e. A nonlinear dynamic system taken from Narendra [20] is used in this example as follows: y(t) = f(y(t − 1), y(t − 2), y(t − 3), u(t − 1), u(t − 2)) f(x1, x2, x3, x4, x5) = x1x2x3x5(x3 − 1) + x4 1 + x2 2 + x2 3 , (33) A number of input-output training data is obtained by testing the system with a pseudo random binary sequence (PRBS) function with the amplitude [-1.0;1.0]. The first 300 training data are shown in Fig. 3. We set the model parameters by ℵ(5,5,5,1) which demonstrates that the model contains five elements of input ϕ(t) = [y(t−1)y(t−2)y(t−3)u(t−1)u(t− 2)]T , five hidden nodes, five output nodes of the embedded system and one output node. Thus, QARXNN model has 55 parameters. An index of normalized prediction error (NPE) and root mean square (RMS) error are introduced in order to measure the performance of system identification and prediction is Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 5. 0 2 4 6 8 10 −5 0 5 10 0 5 10 15 ∆ee √ e2+∆e2 Fig. 4. Learning pattern of QARXNN model with the proposed method. 0 0.01 0.02 0.03 0.04 0.05 −1 0 1 2 0.045 0.05 0.055 0.06 0.065 0.07 0.075 ∆e e √ e2+∆e2 Fig. 5. Learning pattern of QARXNN model with original BP method. stated as follows: NPE = ( ∑N t=1(y(t) − ˆy(t))2 ∑N t=1(y(t))2 )1/2 x100%. (34) RMS = √ ∑N t=1(y(t) − ˆy(t))2 N (35) where ˆy(t), y(t), t = 1, 2, · · · , N are the estimated output, the system output, and time. N denotes the length of training data. The learning pattern of the proposed method and the original BP method have been illustrated in Fig. 4 and Fig. 5. As we can see in Fig. 4 and Fig. 5, the error and gradient error of learning contribute in training the model in order to obtain optimal weights, whereas BP only contributes by learning error. According to the Lyapunov stability theorem, the global minimum point of cost function can be guaranteed if the learning error and the derivative of the error converge to zero. The model obtained from the system identification is tested with the input-output of the training data to estimate next step prediction shown in Fig. 6 and Fig. 7. The training data has uniform distribution. Therefore, the NPE index prediction using the proposed method is consistent to the specific value of 21.7118 for all times. Compared with the BP method that gives an average NPE=30.6449 with standard deviation 0.0389. Thus, the proposed method is able to ensure a stable identification error. 0 50 100 150 200 250 300 −1 −0.5 0 0.5 1 t y(t)andˆy(t) (a) 0 50 100 150 200 250 300 −1 −0.5 0 0.5 1 t y(t)andˆy(t) (b) Fig. 6. Output of one-step-ahead prediction (a) the proposed method. (b) QARXNN-BP method. 0 200 400 600 800 1000 1200 21.7118 21.7118 21.7118 t NPE 0 200 400 600 800 1000 1200 30.4 30.6 30.8 t NPE (b) (a) Fig. 7. NPE index of one-step-ahead prediction. (a) the proposed method. (b) QARXNN-BP method. After training, the test signal u(t) is fed to the model as follows: u(t) =    sin(2πt/250), if t ≤ 500 0.8 sin(2πt/250) +0.2 sin(2πt/25), if t > 500 (36) Fig. 8, Fig. 9 and Fig. 10 show the results of the simulation using the tested signal. The accuracy of the proposed method is 0.0297, whereas the BP method is 0.0792. Compared to the previous work shown in Tab.I, it can be seen that the proposed method performs better than the other methods. TABLE I SIMULATION RESULTS AND COMPARISON Model RMS Number Error of parameters ARX model 0.0866 6 NN 0.0678 341 WN 3 0.0503 12 Q-ARX-NF 0.0478 85 Q-ARX-WN 0.0367 42 Q-ARX-NN < 0.01 246 Q-ARX-NN-ALLF 0.0297 55 Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 6. 0 100 200 300 400 500 600 700 800 −1 −0.5 0 0.5 1 t y(t)andˆy(t) (a) 0 100 200 300 400 500 600 700 800 −1 −0.5 0 0.5 1 t y(t)andˆy(t) (b) y(t) ˆy(t) y(t) ˆy(t) Fig. 8. Simulation results with the deterministic signal (a) the proposed method. (b) QARXNN-BP method 0 100 200 300 400 500 600 700 800 −0.3 −0.2 −0.1 0 0.1 0.2 t y(t)-ˆy(t) Lyapunov BP Fig. 9. The error of simulation tested with deterministic signal. B. System Identification Example 2 This example considers the modeling of the nonlinear system to be as follows: y(t) = 0.72y(t − 1) + 0.025y(t − 2)u(t − 2) + 0.01u2 (t − 3) + 0.2u(t − 4). (37) The same mathematical model of dynamic systems is also used in [21], [22], [23]. The current output of the system depends on two previous output values and three previous input values. However, only two values, y(t−1) and u(t−1), 0 100 200 300 400 500 600 700 800 0 0.02 0.04 0.06 0.08 0.1 t RMS Lyapunov BP Fig. 10. RMS errors of the simulation tested with deterministic signal. 0 20 40 60 80 100 120 140 160 180 200 0 0.05 0.1 0.15 0.2 0.25 epoch RMS Fig. 11. RMS values obtained during training. 0 100 200 300 400 500 600 700 800 900 1000 −1 −0.5 0 0.5 1 t y(t),ˆy(t)ande(t) y(t) ˆy(t) e(t) Fig. 12. The results of identification are fed to the QARXNN model in order to compare it to other models. The length of input of training data is 900. Half of the inputs are independently and identically distributed (i.i.d.) in a uniform sequence over [-2,2], and the remaining are sinusoid generated by 1.05 sin(πt/45). QARXNN model is trained for 200 epochs, as shown in Fig. 11. After training, the following same test signal u(t) is used for testing the performance of the QARXNN model: u(t) =    sin(πt/25), if t < 250 1.0, if 250 ≤ t < 500 −1.0, if 500 ≤ t < 750 0.3 sin(πt/25) +0.1 sin(πt/32) +0.6 sin(πt/10), if 750 ≤ t < 1000 (38) The parameters of the model are denoted as ℵ(2,4,2,1), which shows the model contains two elements of input, four hidden nodes, two output nodes of the embedded system and one output node . Thus, QARXNN model has 20 parameters. Fig. 12 compares the actual output of the system with that of the QARXNN model. The RMS values of the tested model are shown in Fig. 13. A detailed performance comparison of the QARXNN model is given in Tab.II. TABLE II COMPARISON OF QARXNN MODEL WITH OTHER MODELS Models Number RMS RMS of parameters train test RSONFIN [24] 49 0.03 0.06 TRFN-S [21] 33 0.0067 0.0313 FWNN [23] 27 0.01973 0.022609 FWNN [23] 43 0.018713 0.020169 AWN [22] 12 0.009368 0.022933 AWN [22] 20 0.009391 0.023259 Q-ARX-NN-ALLF 20 0.01637 0.01999 Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 7. 0 100 200 300 400 500 600 700 800 900 1000 0.01 0.015 0.02 0.025 t RMS Fig. 13. RMS values of tested model 0 100 200 300 400 500 600 700 800 900 1000 0 5 10 15 20 25 epoch RMS Fig. 14. RMS values obtained during training. C. Prediction of nonlinear system This example considers a nonlinear second order dynami- cal model that has been used in [20], [25], [16]. The process is described by the following equation: y(t) = 0.3y(t − 1) + 0.6y(t − 2) + f[u(t)] (39) where f[u(t)] = u3 (t) + 0.3u2 (t) − 0.4u(t) (40) u(t) = sin(2πt/250) + sin(2πt/25). (41) The function ′ f ′ is a polynomial with three degree of current input signal u(t). The 500 input-output data is generated. Three hundred data are used in training procedure and the remaining 200 data-points are for prediction. The network parameters of QARXNN model used is ℵ(3,3,3,1). The model contains twenty one parameters with three elements of input, three hidden nodes, three output nodes of embedded system and one output node. The models are trained with 1000 epochs, as shown in Fig. 14. The results of the system identification in which the model is tested with training data is shown in Fig. 15, which also shows that the RMS value is 0.3254. Fig. 16 contains the results of prediction when the adaptation of the trained model stops after 300 samplings. As we can see, the proposed model is able to predict a nonlinear system where the RMS index for prediction is 0.343. V. DISCUSSION AND CONCLUSION This paper discusses two issues: 1) a different approach to the hierarchical algorithm of linear and nonlinear sub- models of the quasi-linear ARX neural network is introduced. First, the system is identified using a linear estimator by LSE algorithm. Second, the residual error of the linear sub-model is set as a target output performed using NN. 2) In order to improve the performance of system identification and 0 50 100 150 200 250 300 −40 −20 0 20 40 t y(t)andˆy(t) 0 50 100 150 200 250 300 −1 0 1 2 t e(t) 0 50 100 150 200 250 300 0 0.1 0.2 0.3 0.4 0.5 t RMS Fig. 15. Testing the model with training data. 0 20 40 60 80 100 120 140 160 180 200 −40 −20 0 20 40 t y(t)andˆy(t) 0 20 40 60 80 100 120 140 160 180 200 −1 0 1 2 t e(t) 0 20 40 60 80 100 120 140 160 180 200 0.2 0.4 0.6 0.8 1 t e(t) Fig. 16. Prediction output after training model stops at t = 300. prediction, we propose using the adaptive learning performed based on Lyapunov stability theorem. The fixed learning rate of BP is replaced by the adaptive learning rate based on Lya- punov stability theory. By using a Lyapunov function-based adaptive learning approach to the system embedded in the QARXNN model, the proposed method is able to improve the performance of system identification and prediction. Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________
  • 8. REFERENCES [1] J. Hu, K. Kumamaru and K. Hirasawa, ”A Quasi-ARMAX approach to modelling of nonlinear systems,” Int. J. Control, vol. 74, no. 18, pp 1754-1766, 2001 [2] E. Kamal, A. Aitouche, R. Ghorbani and M. Bayart, ”Robust nonlinear control of wind energy conversion systems,” Int. J. Electr. Power and Energy Syst., vol. 44, pp 202-209, 2013 [3] Y. Cheng, L. Wang and J. Hu, ”Identification of Quasi-ARX neuro- fuzzy model with an SVR and GA approach,” IEICE Trans. Funda- mentals,vol. E.95-A, no. 5, pp 876-883, 2012 [4] S. Hassan, J. Jaafar, B.S. Brahim and A.J. Tahseen, ”A Hybrid Fuzzy Time Series Model for Forecasting,” Engineering Letters, vol. 20, no. 1, pp 88-93, 2012 [5] G. P. Zhang, ”Time series forecasting using a hybrid ARMA and neural network model,” Neurocomputing, vol. 50, pp 159-175, 2003 [6] Kuan-Heng Chen and Khaldoun Khashanah, ”Analysis of Systemic Risk: A Vine Copula-based ARMA-GARCH Model,” Engineering Letters, vol. 24, no. 3, pp 268-273, 2016 [7] M.A. Jami’in, I. Sutrisno and J. Hu, ”Nonlinear Adaptive Control for Wind Energy Conversion Systems Based on Quasi-ARX Neural Net- works Model,” Lecture Notes in Engineering and Computer Science: Proceedings of The International MultiConference of Engineers and Computer Scientists 2014, IMECS 2014, 12-14 March, 2014, Hong Kong, pp 313-318. [8] M.A. Jami‘in, I Sutrisno, J. Hu, N.B. Mariun and M.H. Marhaban, ”An adaptive predictive control based on a quasi-ARX neural network model,” Proc. 13th International Conference on Control, Automation, Robotics and Vision, pp 253-258, Dec. 2014 [9] M.A. Jami‘in, I Sutrisnoa, J. Hu, N.B. Mariun and M.H. Marhaban, ”Quasi-ARX Neural Network Based Adaptive Predictive Control for Nonlinear Systems,” IEEJ Transaction on Electrical and Electronic Engineering, vol. 11, no. 1, pp 83-90, 2016 [10] Y. Cheng, L. Wang and J. Hu, ”Quasi-ARX wavelet network for SVR based nonlinear system identification,” Nonlinear Theory and its Applications (NOLTA), IEICE, vol. 2, no. 2, pp 165-179, 2011 [11] W Yu, A.S. Poznyak, X. Li, ”Multilayer dynamic neural networks for non-linear system on-line identification,” Int. J. Control, vol. 74, no. 18, pp 1858-1864, 2001 [12] L. Bahera, S. Kumar and A. Patnaik, ”On adaptive learning rate that guarantees convergence in feedforward networks,” IEEE Transaction on Neural Networks, vol. 17, no. 5, pp 1116-1125, 2006 [13] Z. Man, H.R. Wu, S. Liu and X. Yu, ”A new adaptive backpropagation algorithm based on lyapunov stability theory for neural networks,” IEEE Transaction on Neural Networks, vol. 17, no. 6, pp 1580-1591, 2006 [14] V. Becera, F Garces, S. Nasuto and W Holderbaum, ”An efficient parameterization of dynamic neural networks for nonlinear system identification,” IEEE Transaction on Neural Networks, vol. 16, no. 4, pp 983-988, 2005 [15] F. Abdollahi, H.A. Talebi and R.V. Patel, ”Stable identification of nonlinear systems using neural networks: Theory and experiments,” IEEE/ASME Trans. on Mechatronics, vol. 11, no. 4, pp 488-495, 2006 [16] M.A. Shoorehdeli, M. Teshnehlab and A.K. Sedigh, ”Identification using ANFIS with intelligent hybrid stable learning algorithm ap- proaches,” Neural Computing and Applications, vol. 18, pp 157-174, 2009 [17] M.A. Jami’in, I. Sutrisno and J. Hu, ”Deep searching for parameter estimation of the linear time invariant (LTI) system by using quasi- ARX neural network,” Proc. IEEE International Joint Conference on Neural Network, pp 2758-2762, 2013 [18] M.A. Jamiin, J. Hu and E. Julianto, ”A Lyapunov Based Switching Control to Track Maximum Power Point of WECS,” Proc. IEEE International Joint Conference on Neural Network, pp 3883-3888, 2016 [19] J. Hu and K. Hirasawa, A method for applying neural networks to control of nonlinear systems, in book entitled Neural Information Processing: Research and Development, Springer, Berlin, Germany, pp 351-369, 2004 [20] K.S. Narendra and K. Parthasarathy, ”Identification and control of dynamical systems using neural networks,” IEEE Trans. on Neural Networks, vol. 1, no. 1, pp 4-27, 1990 [21] C.F. Juang, ”A TSK-type recurrent fuzzy network for dynamic sys- tems processing by neural network and genetic algorithms,” IEEE Trans.Fuzzy Syst., vol. 10, no. 2, pp 155-170, 2002 [22] Y. Oysal and S Yilmaz, ”An adaptive wavelet network for function learning,” Neural Computing and Applications, vol. 19, pp 383-392, 2010 [23] R.H. Abiyev, O. Kaynak, ”Fuzzy wavelet neural networks for iden- tification and control of dynamic plants a novel structure and a comparative study,” IEEE Trans. Ind. Electron, vol. 55, no. 8, pp 3133- 3140, 2008 [24] C.F. Juang and C.T. Lin, ”A recurrent self-organizing neural fuzy inference network,” IEEE Trans. Neural Network, vol. 10, no. 4, pp 828-845, 1999 [25] A. Banakar, M.F. Azeem, ”Artificial wavelet neural network and its application in neuro-fuzzy models,” Appl. Soft Comput., vol. 8, no. 4, pp 1463-1485, 2008 Mohammad Abu Jami’in received the M.E. degree in Control Engineering from Institut Teknologi Sepuluh Nopember (ITS) Surabaya, Indonesia in 2008, and the Doctor of Engineering in Neurocomputing from Waseda University, Japan in 2016. He is currently a Lecturer with the Politeknik Perkapalan Negeri Surabaya, Indonesia. His research interests include arti- ficial intelligence and its applications. Yuyun received the B.E. degree in Computer Engineering STMIK Dipanegara Makassar and the M.E. degree in Electrical Engineering from Hasanuddin University Makassar, Sulawesi Selatan, Indonesia in 2010 and 2013. He is currently a lecture in STMIK Handayani Makassar, Indonesia. His research interests include Computer Engineering. Eko Julianto received the B.E. degree in Marine Engineering from Institut Teknologi Sepuluh Nopember (ITS) Surabaya, Indonesia in 1989, and the M.E degree in Marine Engineering from University Of New Castle Upon Tyne, UK, 1994. He is currently a Lecturer with the Politeknik Perkapalan Negeri Surabaya, Indonesia. His research interests include Marine Engineer- ing. Engineering Letters, 25:3, EL_25_3_14 (Advance online publication: 23 August 2017) ______________________________________________________________________________________