ArticlePDF Available

Computationally Efficient Data-Driven Higher Order Optimal Iterative Learning Control

Authors:

Abstract and Figures

Based on a nonlifted iterative dynamic linearization formulation, a novel data-driven higher order optimal iterative learning control (DDHOILC) is proposed for a class of nonlinear repetitive discrete-time systems. By using the historical data, additional tracking errors and control inputs in previous iterations are used to enhance the online control performance. From the online data, additional control inputs of previous time instants within the current iteration are utilized to improve transient response. The data-driven property of the proposed method implies that no model information except for the I/O data is utilized. The computational complexity is reduced by avoiding matrix inverse operation in the proposed DDHOILC approach due to the nonlifted linear formulation of the original model. The asymptotic convergence is proved rigorously. Furthermore, the convergence property is analyzed and evaluated via three performance indexes. By elaborately selecting the higher order factors, the higher order learning control law outperforms the lower order one in terms of convergence performance. Simulation results verify the effectiveness of the proposed approach.
Content may be subject to copyright.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 1
Computationally Efficient Data-Driven Higher
Order Optimal Iterative Learning Control
Ronghu Chi , Zhongsheng Hou, Senior Member, IEEE, Shangtai Jin, and Biao Huang, Fellow, IEEE
Abstract Based on a nonlifted iterative dynamic linearization
formulation, a novel data-driven higher order optimal iterative
learning control (DDHOILC) is proposed for a class of non-
linear repetitive discrete-time systems. By using the historical
data, additional tracking errors and control inputs in previous
iterations are used to enhance the online control performance.
From the online data, additional control inputs of previous time
instants within the current iteration are utilized to improve
transient response. The data-driven property of the proposed
method implies that no model information except for the I/O data
is utilized. The computational complexity is reduced by avoiding
matrix inverse operation in the proposed DDHOILC approach
due to the nonlifted linear formulation of the original model.
The asymptotic convergence is proved rigorously. Furthermore,
the convergence property is analyzed and evaluated via three
performance indexes. By elaborately selecting the higher order
factors, the higher order learning control law outperforms the
lower order one in terms of convergence performance. Simulation
results verify the effectiveness of the proposed approach.
Index Terms—Computational efficiency, convergence evalua-
tion, data driven, higher order learning law, nonlifted iterative
dynamic linearization.
I. INTRODUCTION
IN PRACTICAL industries, many processes repetitively
perform the same task. To improve the tracking accuracy
of such processes, iterative learning control (ILC) [1] was
proposed with the ability of learning from previous executions.
Amann et al. [2] pioneered the optimization-based ILC for
linear systems. Since then, many alternative approaches of
optimal ILC have been explored with successful applica-
tions [3]–[8] because all the tracking errors, as well as the
constraints on the input difference between trials, input effort,
Manuscript received November 20, 2016; revised November 25, 2017 and
February 25, 2018; accepted March 5, 2018. This work was supported in part
by the National Science Foundation of China under Grant 61374102, Grant
61573054, and Grant 61433002 and in part by the Taishan Scholar Program of
Shandong Province of China. (Corresponding author: Ronghu Chi.)
R. Chi is with the School of Automation and Electronic Engineering,
Qingdao University of Science and Technology, Qingdao 266061, China
(e-mail: ronghu_chi@hotmail.com).
Z. Hou and S. Jin are with the Advanced Control Systems Lab-
oratory, School of Electronics and Information Engineering, Beijing
Jiaotong University, Beijing 100044, China (e-mail: zhshhou@bjtu.edu.cn;
shtjin@bjtu.edu.cn).
B. Huang is with the Department of Chemical and Materials Engi-
neering, University of Alberta, Edmonton, AB T6G 2G6, Canada (e-mail:
bhuang@ualberta.ca).
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/TNNLS.2018.2814628
and system output can be easily considered through the objec-
tive function. By properly selecting weight matrices to satisfy
the convergence condition, the optimal ILC can make tracking
error converge asymptotically along the iteration direction,
which is most desired in practical applications. However, two
major problems still remain that hamper the application of
optimal ILC methods to practical problems.
The first is the computation efficiency problem of the
optimal ILC methods. Due to the supervector representation
used in norm-optimal ILCs, the lifted system matrix dimension
is increasing exponentially with the batch length and may have
several million elements in the case of the robotic applica-
tions [9], which turns out to be computationally infeasible.
Therefore, a computationally efficient optimal ILC is more
attractive in real industries. Recently, several works have been
done to address the issue of computation efficiency of the
lifted optimal ILC [10]–[13]. But, all the above optimal ILC
approaches [3]–[6], [10]–[13] are limited to exactly known
linear system models or linearly approximated models. This
is the second problem of optimal ILCs preventing them from
real applications.
Although Volckaert et al. [7] and Axelsson et al. [8] have
discussed the optimal ILC design for nonlinear systems by
adding model estimation, an explicit linearized model is still
required to approximate the original nonlinear plant, whose
mathematical model should be known exactly. Therefore,
a negative influence on the stability and robustness will occur
due to the model mismatch and model complexity.
Moreover, with the increasingly large scale and complexity
in real production processes, it is very difficult to acquire their
accurate mathematical models, no matter linear or nonlinear,
by first principle or system identification [14]. Consequently,
data-driven control [14] becomes an interesting and attractive
topic in recent years, where no explicit model is required for
the controller design and analysis. Recently, several develop-
ments [15], [16] under the term of “data-driven ILC” have
been reported by estimating a system representation using
I/O measurements. However, these approaches are designed
and analyzed in a linear system framework. In [17], a data-
driven optimal ILC is presented for nonlinear systems based on
a lifted representation, where the issue of efficient computation
is still open.
In order to achieve a better control performance, the higher
order ILC was originally proposed in [18] to employ control
information of previous iterations. Subsequently, many alterna-
tive higher order ILC methods [19]–[22] have been proposed.
2162-237X © 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
2IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
The basic idea of the works in [18]–[22] is to utilize more
historical data collected from previous iterations to enhance
the online control performance since the historical data con-
tain system information. However, it is worth pointing out
that online data in the current iteration, compared with the
historical data in the previous iterations, can reflect varying
characteristics of a process and the process disturbances in real
time. Therefore, whenever possible, one should incorporate the
historical as well as online data to achieve a better control
performance. Furthermore, the existing higher order methods
mainly focused on PID-type and optimization-based schemes,
and their design and analysis is also limited to linear systems.
Convergence speed is one of the most important factors to
consider in the ILC. The convergence speed is defined in the
frequency domain in [23] and the tradeoff in normal optimal
ILC in terms of robustness, convergence speed, and steady
state error has been described. In [24], the relationship between
the learning gains and the convergencerate has been discussed.
Son et al. [25] discussed the compromise between robustness
and convergence speed in the robust ILC design.
Higher order ILC [26]–[30] offers the possibility of a higher
error convergence speed by utilizing more control information
from several previous trials than the standard ILC that only
uses the immediate previous one trial. The effects of the mem-
ory length on the error convergence rate have been explored
and evaluated in [26]–[29]; however, up to now, there is little
work done to guarantee in theory that the higher order ILC
outperforms the lower order one with a faster convergence.
In this paper, a novel historical and online data-driven
higher order optimal iterative learning control (DDHOILC) is
proposed based on a nonlifted iterative dynamic linearization
for the nonlinear discrete-time system. The nonlifted iterative
dynamic linearization used in this paper does not require an
exactly known mathematical model of the controlled plant
and results in a model equivalent to the original nonlinear
plant without an approximation. Thus, the proposed approach
is data-driven and no process model is explicitly required.
Owing to the use of additional historical and online data,
that is, more data related to the tracking errors from the
previous iterations and more data from control inputs at
previous time instants within the current iteration, a better
control performance is achieved by applying the proposed
DDHOILC. Rigorous mathematical analysis is provided to
show the asymptotic convergence of the tracking error and to
evaluate the convergence property of the proposed DDHOILC.
It is concluded that under certain conditions, the convergence
speed of the higher order learning law can be faster than that
of the lower order one by selecting the higher order factors
and controller parameters properly. Furthermore, due to the
avoidance of the matrix inverse calculation in the control law,
an efficient computation can be achieved even though the
number of samples in the trial and the trajectory length is
larger. Simulations in this paper verify the derived theoretical
results.
The remainder of this paper is structured as follows.
Section II formulates a nonlifted dynamical linearization for
nonlinear systems. Section III is the controller design of
DDHOILC. Section IV shows the asymptotic convergence
of the proposed method. Section V evaluates convergence
property of the proposed method with rigorous derivations.
Two examples are considered in Section VI to verify the
effectiveness of the proposed approach. Section VII provides
the conclusion.
II. NONLIFTED ITERATIVE DYNAMIC LINEARIZATION
OF REPETITIVE NONLINEAR SYSTEMS
Consider a class of repeatable nonaffine nonlinear discrete-
time systems with unknown orders
yk(t+1)=f(yk(t),...,yk(tny), uk(t),...,uk(tnu))
(1)
where yk(t)and uk(t)are the system output and input,
respectively, and yk(t)=0anduk(t)=0forallt<0,
f(·)is an unknown real nonlinear function and continuously
differentiable, and f(0·0)=0, nyand nuare the orders of
system output and input, respectively, t∈{0,...,N}, with N
being the endpoint of the finite time interval, and kdenotes
the index of iteration.
By following the similar steps in [17], the system output can
be expressed by initial states and the control input series as:
yk(t+1)=gt(yk(0), uk(0),...,uk(t)) (2)
where gt(·),t=0,...,N1, is a proper nonlinear function
and is also continuously differentiable.
In the following discussion, two assumptions [17] are made.
Assumption 1: The initial value yk(0)is unchanged for all
iterations, i.e., yk(0)=c0,k,andc0is a constant.
Assumption 2: Nonlinear function gt(·)is globally
Lipschitz, i.e.,
|gt(x1,u1)gt(x2,u2)|≤Lx|x1x2|+Luu1u2
where Lxand Luare the two positive Lipschitz constants.
Remark 1: Because the original controlled plant (1) is
nonlinear, nonaffine, and completely unknown, the globally
Lipschitz condition (Assumption 2) is required in the follow-
ing analysis as a tradeoff. The way of relaxing Assumption 2
to locally Lipschitz nonlinearity is to transfer the original
nonaffine system into an affine form with respect to control
input by referring the existing works of contraction-mapping-
based ILC with locally Lipschitz nonlinearities [31], [32].
It is not a trivial work to extend the proposed ILC to locally
Lipschitz nonlinear systems because the original controlled
plant considered in this paper is nonlinear and nonaffine.
On the other hand, some recent works have shown that the
contraction-mapping-basedILC approach can deal with locally
Lipschitz nonlinearity, where the nonlinear system considered
is affine to the control input. Therefore, the extension may be
possible by dividing the original plant into two parts: locally
Lipschitz nonlinearity and affine control input.
The iterative dynamic linearization in [17] can be easily
extended to a nonlifted form, summarized as follows.
Lemma 1: For the nonlinear system (2), which is a reformu-
lation of system (1) with respect to system output, initial state,
and control input, satisfying assumptions 1 and 2, according
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
CHI et al.: COMPUTATIONALLY EFFICIENT DDHOILC 3
to the mean value Theorem, there exists an optimal gradient
vector θt
k(t)such that
yk(t+1)=yk1(t+1)+θt
k(t)uk(t)(3)
and θt
k(t)≤Lu,whereuk(t)=uk(t)uk1(t),uk(t)=
[uk(0), uk(1),...,uk(t)]T, whose dimension varies with time
instants, e.g., uk(t1)=[uk(0), uk(1),...,uk(t1)]T;and
θt
k(t)=θt
k(0), θ t
k(1),...,θt
k(t)
=gt(·,·)
uk(0),gt(·,·)
uk(1),...,gt(·,·)
uk(t)
whose dimension also varies with time instant, denotes the
optimal partial derivatives of gt(·,·)with respect to uk(t)in
the interval of [uk(t), uk1(t)].
Readers can refer to the literature [17] for the detail proof.
Note that the above nonlifted iterative dynamic lineariza-
tion (3) is completely equivalent to the original nonlinear
system (2), which is equivalent to (1) in turn.
III. HIGHER ORDER DATA -DRIVEN OPTIMAL ILC
Assume that yd(t),t∈{0,...,N}is a target trajectory and
is bounded for all time instants. The control objective is to
make the tracking error ek(t)=yd(t)yk(t)converge to
zero asymptotically along the iterations. Since θt
k(t)in (3) is
unknown and slowly iteration and time varying, a modified
projection algorithm [17] is utilized for its iterative estimate
ˆ
θt
k(t)=ˆ
θt
k1(t)
+η(yk1(t+1)ˆ
θt
k1(t)uk1(t))uT
k1(t)
µ+uk1(t)2(4)
ˆ
θt
k(t)=ˆ
θt
0(t), if sgnˆ
θt
k(i)= sgnˆ
θt
0(i)or
ˆ
θt
k(t)
ε
(5)
where i=0,...,t,ˆ
θt
k(t)is the estimation of θt
k(t),µ>0
denotes a weighting factor, η(0,2)is a step-size factor, and
εis a small positive scalar. The initial value of ˆ
θt
0(t)should
be selected such that all its elements have the same signs as
that of θt
k(t), whose elements’ signs can be identified by using
the I/O data sampled from the controlled plant.
Remark 2: In this paper, we mainly focus on the controlled
systems with relatively insignificant measurement noise, and
thus, the direction of θt
k(t)can be obtained by the experimental
trials along with other priori knowledge. The cases with
significant random data noise will be addressed in our future
work followingsimilar line as the stochastic ILC methods [33].
However, it would be difficult for most of the existing control
methods to deal with the case that θt
k(t)converges to 0 because
the control action becomes weaker and weaker.
Consider an objective function with higher order factors
J(uk(t), α)=M
m=1
αmekm+1(t+1)2
+λ(uk(t)uk1(t))2(6)
where λ>0 is a weighting factor, and α=[α1,...,α
M]
denotes the higher order factors and M
m=1αm=1,
0
m1, α1+α2M
m=3αmα>0, where Mis
a positive integer.
Remark 3: Since the controlled plant (1) is a nonlinear
unknown process and the system uncertainties are inevitable,
according to [25] theSterm used in the classical optimal ILC
has not been included in the cost function (6) to enhance the
robustness of the designed control system.
Rewrite (3) as
yk(t+1)=yk1(t+1)
+θt
k(t1), θ t
k(t)uT
k(t1), uk(t)T(7)
where θt
k(t1)=[θt
k(0), θ t
k(1),...,θt
k(t1)].
Substituting (7) into (6), yields
J(uk(t), α)
=α1ek1(t+1)
t1
i=0
θt
k(i)uk(i)θt
k(t)uk(t)
+
M
m=2
αmekm+1(t+1)2
+λ|uk(t)uk1(t)|2
(8)
where uk(i)=uk(i)uk1(i),i=0,1,...,t.
Minimizing objective function (6) with respect to uk(t)by
replacing the unknown θt
k(t)with ˆ
θt
k(t), a learning control law
is derived as
uk(t)=uk1(t)ρα2
1ˆ
θt
k(t)t1
i=0ˆ
θt
k(i)uk(i)
λ+α2
1ˆ
θt
k(t)2
+
ρα1ˆ
θt
k(t)α1ek1(t+1)+M
m=2αmekm+1(t+1)
λ+α2
1ˆ
θt
k(t)2
(9)
where ρ>0 is a positive factor.
Remark 4: From the proposed DDHOILC (4), (5), and (9),
it is clearly seen that: 1) there is no matrix inverse calculation
included, so an improved computational efficiency can be
achieved; 2) the unknown parameters in (9) are updated in
terms of (4) and (5) such that the proposed method is more
flexible for the modifications/or expansions of the controlled
plant; and 3) there is no model information used except for
the input and output measurements, so the proposed method
is data driven and suitable to complex nonlinear processes in
practice.
IV. CONVERGENCE ANALYSIS
Before proceeding with the analysis, three lemmas are given
as follows.
Lemma 2 [34]: Let
A=
010··· 0
001··· 0
.
.
..
.
..
.
.....
.
.
000··· 1
a1a2a3··· at
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
4IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
and s(A)is the spectral radius of A.Ift
i=1|ai|<d,0<
d<1, then s(A)<d.
Lemma3[35]:ACn×n,ands(A)is the spectral radius
of A. Then, for any δ>0, there always exists a proper matrix
norm ·
von the normed vector space v, such that
Av<s(A)+δ.
Lemma4[35]:Let Avand Aµbe the proper matrix
norm of Aon the normed vector spaces νand µ, respectively.
Then, there must exist γσ>0suchthat
γAv≥AµσAv,ACm×n.
The boundedness of ˆ
θt
k(t)has been shown in [17], so we
have |ˆ
θt
k(i)|≤bθ,i=0,1,...,t,wherebθis a positive
constant. Then, the convergence property of the proposed
DDHOILC (4), (5), and (9) is summarized in the following
theorem.
Theorem 1: Consider nonlinear system (1) satisfying
assumptions 1 and 2. Applying the proposed DDHOILC (4),
(5), and (9) with proper controller parameters λand ρsuch
that
λ>max ρ2α2
1b2
θN2,γ2ρ2
4,ρ2L2
u
4,γ2ρ2L2
u
4
where γis a positive constant, it is guaranteed that (t1) the
tracking error ek(t+1)converges to zero asymptotically and
iteratively; (t2) the control system is bounded-input-bounded-
output stable.
Remark 5: From Theorem 1, the bounds bθ,Lu,and γ
should be known for a proper selection of λ. For a practical
control process in which the above bounds are known, one can
select a larger λconservatively to guarantee the asymptotical
convergence of the tracking error. The tradeoff is that the
convergence speed may become slower. Alternatively, one
can estimate these bounds by using experiments and then
determine the range of λby using the estimated values.
Proof: The two matrices Au
k(t)and Bu
k(t)are defined in
the equation given at the bottom of this page.
Define Cu=[0,...,0,1]TRt+1,and¯ek1(t+1)=
[ekM+1(t+1), ekM+2(t+1),...,ek1(t+1)]TRM1.
Furthermore, we define an expended vector with fixed (t+1)
dimensions as ¯uk(t)=uk(t)=[uk(0)uk(1)··· uk(t)]T,then
¯uk(t1)=[uk(1)uk(0)··· uk(t1)]T, where the input
signal uk(t)is set as zero for t<0.
Then, according to (9), one has
¯uk(t)=Au
k(t)¯uk(t1)+CuBu
k(t)¯ek1(t+1). (10)
Note that the following inequality holds:
t1
i=0
ρα2
1ˆ
θt
k(t)ˆ
θt
k(i)
λ+α2
1ˆ
θt
k(t)
2
t1
i=0
ρα2
1ˆ
θt
k(t)ˆ
θt
k(i)
2λα1ˆ
θt
k(t)
t1
i=0
ρα1ˆ
θt
k(i)
2λ.(11)
Since tis finite over {0,1,...,N}, by properly select-
ing λand ρ,suchλ>(ρα
1bθN)2, there exists a series
of positive constants at time instant tof the kth iteration
0<d1(k,t)<0.5where0<d1(k,t)<0.5isdenedin
the following equation such that for each fixed time instant t
and fixed iteration number k, the following inequality
holds, i.e.,
0
t1
i=0
ρα2
1ˆ
θt
k(t)ˆ
θt
k(i)
λ+α2
1ˆ
θt
k(t)
2
t1
i=0
ρα1ˆ
θt
k(i)
2λ
ρα1bθN
2λ=d1(k,t)<0.5 (12)
which implies that the spectral radius s(Au
k(t)) < d1(k,t)
holds for all tand kin view of Lemma 2. Then, according
to Lemma 3, there exist a series of arbitrary small positive
constants δ1(k,t)such that
Au
k(t)
vs(Au
k(t)) +δ1(k,t)d1(k,t)+δ1(k,t)
=d2(k,t)d2<0.5 (13)
where Au
k(t)vdenotes a proper matrix norm of Au
k(t)on the
normed vector space vat thetth time instant of the kth iteration,
and 0 <d2(k,t)<0.5 is a series of positive constants.
d2=supk,td2(k,t). It is obvious that 0 <d2<0.5 because
0<d2(k,t)<0.5 holds for all time instants tand iterations k,
and the condition of 0 <d2<0.5 will be used in the following
analysis.
Taking norm on both sides of (10), yields
¯uk(t)v
Au
k(t)
v¯uk(t1)v
+Cuv
Bu
k(t)
v¯ek1(t+1)v.(14)
According to Lemma 4, there exists a constant γsuch that
Bu
k(t)
vγ
Bu
k(t)
2.
Au
k(t)=
01 0··· 0
00 1··· 0
.
.
..
.
..
.
.....
.
.
00 0··· 1
0ρα2
1ˆ
θt
k(t)ˆ
θt
k(0)
λ+α2
1ˆ
θt
k(t)
2ρα2
1ˆ
θt
k(t)ˆ
θt
k(1)
λ+α2
1ˆ
θt
k(t)
2··· −ρα2
1ˆ
θt
k(t)ˆ
θt
k(t1)
λ+α2
1ˆ
θt
k(t)
2
(t+1)×(t+1)
Bu
k(t)=ρα1ˆ
θt
k(tM
λ+α2
1ˆ
θt
k(t)
2
ρα1ˆ
θt
k(tM1
λ+α2
1ˆ
θt
k(t)
2··· ρα1ˆ
θt
k(t3
λ+α2
1ˆ
θt
k(t)
2
ρα1ˆ
θt
k(t)(α1+α2)
λ+α2
1ˆ
θt
k(t)
21×(M1)
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
CHI et al.: COMPUTATIONALLY EFFICIENT DDHOILC 5
So, the following inequality can be derived directly since
M
m=1αm=1, 0
m1:
Bu
k(t)
2
vγ2
Bu
k(t)
2
2
γ2ρα1ˆ
θt
k(t)21+α2)2+α2
3+···+α2
M
4λα2
1ˆ
θt
k(t)2
γ2ρα1ˆ
θt
k(t)21+α2+···+αM)2
4λα2
1ˆ
θt
k(t)2=γ2ρ2
4λ.
(15)
One can select λproperly such that λ>γ
2ρ2/4, then
Bu
k(t)
vγ
Bu
k(t)
2γρ/(2λ) =d3<1 (16)
where 0 <d3<1 is a positive constant.
Hence, one can derive from (13)–(16) that
¯uk(t)vd2¯uk(t1)v+d3¯ek1(t+1)v
···d3
t
i=0
dti
2¯ek1(i+1)v.(17)
By virtue of (3) and (10), one has
ek(t+1)=ek1(t+1)θt
k(t)CuBu
k(t)¯ek1(t+1)
θt
k(t)Au
k(t)¯uk(t1). (18)
The second error term in (18) can be further expanded as
θt
k(t)CuBu
k(t)¯ek1(t+1)
=θt
k(t)Bu
k(t)¯ek1(t+1)
=ρα1θt
k(t)ˆ
θt
k(t)(α1+α2)
λ+α2
1ˆ
θt
k(t)2ek1(t+1)
+
M
m=3
ρα1θt
k(t)ˆ
θt
k(tm
λ+α2
1ˆ
θt
k(t)2ekm+1(t+1). (19)
Let ϑk,1(t)=ρα1θt
k(t)ˆ
θt
k(t)(α1+α2)/(λ +α2
1ˆ
θt
k(t)2),and
ϑk,j(t)=ρα1θt
k(t)ˆ
θt
k(tj+1/(λ +α2
1ˆ
θt
k(t)2),j=2,...,
M1, for notational convenience. According to the reset algo-
rithm (5), it is easy to obtain that θt
k(t)ˆ
θt
k(t)>0. Since
0
m1,m=1,...,M, one can get ϑk,j(t)>0,
j=1,...,M1and
0
k,1(t)ρθt
k(t)
2λand 0
k,j(t)ρθt
k(t)
2λ.(20)
Since θt
k(t)≤Lu, by properly selecting λand ρsuch
that λ>ρ
2L2
u/4, one can guarantee that
0
k,1(t)ρLu
2λ<1and0
k,j(t)ρLu
2λ<1.(21)
Define
Ae
k(t)=
01 0··· 0
00 1··· 0
.
.
..
.
..
.
.....
.
.
00 0··· 1
ϑk,M(t)ϑk,M1(t)ϑk,M2(t)···1ϑk,1(t)
and Ce=[0··· 01]TRM1. In terms of (18), we have
¯ek(t+1)=Ae
k(t)¯ek1(t+1)Ceθopt
k(t)Au
k(t)¯uk(t1).
(22)
From (21)
|ϑk,M(t)|+|ϑk,M1(t)|+···+|ϑk,2(t)|+|1ϑk,1(t)|
=1k,1(t)ϑk,2(t)ϑk,3(t)−···ϑk,M(t))
=1
ρα1α1+α2M
m=3αmθt
k(t)ˆ
θt
k(t)
λ+α2
1ˆ
θt
k(t)2.(23)
According to the condition 0
1+α2M
m=3αm=
¯α<1, as long as λ>ρ
2L2
u/4, the following inequality holds:
0<M1
ρα1α1+α2M
m=3αmθt
k(t)ˆ
θt
k(t)
λ+α2
1ˆ
θt
k(t)2
ρα1¯αθt
k(t)ˆ
θt
k(t)
2α1λˆ
θt
k(t)
<ρLu
2λ=d4<1 (24)
where 0 <M1<1and0<d4<1 are two positive constants.
Therefore, it is obtained from (23) and (24) that
1d4<|ϑk,M(t)|+···+|ϑk,2(t)|+|1ϑk,1(t)|
1M1<1.(25)
In terms of Lemma 2 and inequality (25), one can obtain
that s(Ae
k(t)) < 1M1<1. Consequently, one can find an
arbitrarily small positive constant δ2such that
Ae
k(t)
vsAe
k(t)+δ21M1+δ2d5<1 (26)
where 0 <d5<1 is a positive constant.
By virtue of (13), (17), and (26), one can derive that
¯ek(t+1)
v
Ae
k(t)
v¯ek1(t+1)v
+Cev
θt
k
v
Au
k(t)
v¯uk(t1)v
d5¯ek1(t+1)v+Lud2¯uk(t1)v
d5¯ek1(t+1)v
+Lud3
t
t=0
dti+1
2¯ek1(i+1)v.(27)
Assume that ¯ek+1)v=maxt∈{0,...,N1}
{¯ek(t+1)v}=¯emax
kis attained at time instant τ+1,
where ¯emax
kdenotes the maximum value at the kth iteration.
From (27), we have
¯emax
k=¯ek+1)v
d5+Lud3d2
1d2¯ek1+1)vd6¯emax
k1(28)
where d6=d5+(Lud3d2)/(1d2).Sinceλ > ((γ 2ρ2L2
u)/4),
one has 0 <Lud3<1. Furthermore, 0 <d5<1and0<
d2/(1d2)<1 due to 0 <d2<0.5, so by selecting λ
properly, one can get 0 <d6<1. Then inequality (28) implies
that
¯emax
kd6¯emax
k1···dk
6¯emax
0.(29)
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
6IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
Since the initial value ¯emax
0is bounded, we have
0lim
k→∞ ¯ek(t+1)vlim
k→∞ ¯emax
k=0.(30)
So, the asymptotic convergence of tracking error is proved.
The boundedness of both of yd(t+1)and ek(t+1)means
that yk(t+1)is bounded. It is also clear that
¯uk(t)v
k
j=1¯uj(t)v+¯u0(t)v.(31)
According to (17) and (31), we have
¯uk(t)vd3
k
j=1
t
i=0
dti
2¯ej1(i+1)v+¯u0(t)v.(32)
Since ¯ek(t+1)v≤¯emax
k, one can derive that
¯uk(t)vd31dt
2
1d2
k
j=1¯emax
k+¯u0(t)v.(33)
Then, according to (29)
¯uk(t)vM3dk
6+dk1
6+···+d0
6¯emax
0+¯u0(t)v
M31
1d6¯emax
0+¯u0(t)v(34)
where M3=d3/(1d2)is bounded. Because ¯emax
0and
¯u0(t)vare bounded, then (34) means that the control input
is bounded at all-time instants and iterations.
Remark 6: The spectral radius analysis is used to achieve the
convergence of the proposed method by extending the similar
results in [36] from time domain into iteration domain. On the
other hand, one can also prove the convergence via linear
matrix inequality technique similar to the existing work in
[37], but the selectable range of the controller parameters may
become indistinct.
V. CONVERGENCE PROPERTY EVA L U AT I O N
The convergence speed is of interest for practical applica-
tions. However, the quantification of the convergence speed is
very difficult and little related work can be found in [23]–[25].
In this paper, according to (14) and (27), it is clear that
the convergence performance is affected mainly by the norm
values of Au
k(t)vBu
k(t)v,andAe
k(t)v. Therefore, one can
define three indexes as follows to evaluate the convergence
property of the proposed higher order learning control law:
S1(α)=
Au
k(t)
v,S2(α)=
Bu
k(t)
v,S3(α)=
Ae
k(t)
v.
All the three indices reflect the convergence rate qualita-
tively in the same direction, that is, the smaller the indexes
are, the faster the convergence rate is.
Therefore, in order to guarantee that higher order algorithm
outperforms the lower order one, it is required that Sn(αH)<
Sn(αL),n=1,2,3, where αHdenotes the weighting vector
for the higher order algorithm, and αLis for the lower order
algorithm.
Convergence Index 1: Provided that αL
1
H
1,then
t1
i=0
ραH
12ˆ
θt
k(t)ˆ
θt
k(i)
λ+αH
12ˆ
θt
k(t)2
t1
i=0
ραL
12ˆ
θt
k(t)ˆ
θt
k(i)
λ+αL
12ˆ
θt
k(t)2
.(35)
According to (12) and (13), inequality (35) implies that
S1(αH)<S1(αL). So, one can select the higher order factors
such that αL
1
H
1to make the higher order learning law
have a better performance than the lower order one.
Convergence Index 2: Because αH
1+αH
2+···+αH
H=1,
αL
1+αL
2+···+αL
L=1, α1+α2M
m=3αmα>0, and
H>L, it is obvious that
ραH
1ˆ
θt
k(t)2αH
1+αH
22+αH
32+···+αH
H2
λ+αH
12ˆ
θt
k(t)2
<ραH
1ˆ
θt
k(t)2αL
1+αL
22+αL
32+···+αL
L2
λ+αH
12ˆ
θt
k(t)22.(36)
According to (15), in order to warrant S2(αH)<S2(αL),
it is required that
αH
1
λ+αH
12ˆ
θt
k(t)2<αL
1
λ+αL
12ˆ
θt
k(t)2.(37)
Solving (37) subject to the condition of αL
1
H
1,
we get λ>α
H
1αL
1ˆ
θt
k(t)2. That is, the conditions guaranteeing
S2(αH)<S2(αL)are λ>α
H
1αL
1b2
θand αL
1
H
1.
Convergence Index 3: Define ¯αH= H
1+αH
2
H
h=3αH
h),¯αL=L
1+αL
2L
l=3αL
h).IfαL
1¯αL
H
1¯αH,
then
λαL
1¯αLαH
1¯αH
L
1αH
1αH
1¯αHαL
1¯αLˆ
θopt
k(t)
2(38)
which means
αH
1¯αH
λ+αH
12ˆ
θt
k(t)2>αL
1¯αL
λ+αL
12ˆ
θopt
k(t)
2.(39)
Hence, according to (23)–(26), inequality (39) implies that
S3(αH)<S3(αL).
Note that the condition of αL
1¯αL
H
1¯αHis not applicable
to the cases of first-order and second-order algorithms where
¯α1α2=1 are fixed and cannot be manipulated further.
So, for the first-order and second-order algorithms, one can
only use S1(α)and S2(α)to evaluate the convergence property
of the tracking error in theory.
Remark 7: As a summary of the above three cases, for the
higher order algorithms with more than third orders, one can
select αproperly such that αL
1
H
1,λ>α
H
1αL
1b2
θ,and
αL
1¯αL
H
1¯αHto guarantee an improved performance of the
higher order learning control law.
Remark 8: Note that the controller parameters must be
selected to guarantee the convergence of the proposed method
at first, and then, the convergence property is evaluated in the
sequel. So a faster convergence by selecting the higher order
factors properly does not neglect the tracking performance but
actually enhances it.
Remark 9: In general, one can fix the value of λand
then tune the value of ρby trials such that they satisfy the
condition given in Theorem 1. If the parameters are selected
differently under the given condition, the convergence can still
be guaranteed but the convergence rate may be different. The
larger values of ρand smaller values of λmay generally
produce a faster convergence speed.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
CHI et al.: COMPUTATIONALLY EFFICIENT DDHOILC 7
Fig. 1. Tracking errors with respect to iterations in Example 1.
VI. SIMULATION EXAMPLES
Example 1: For comparison with traditional optimal ILC,
a linear time-varying (LTV) system is adopted from [38].
When applying the traditional lifted optimal ILC, the linear
model is required to be known exactly. However, it is con-
sidered to be unknown and only serves for generating the I/O
data for evaluating the proposed DDHOILC. The LTV system
is shown as follows:
x(t+1)=A(t)x(t)+B(t)u(t)
y(t)=C(t)x(t)(40)
where
A(t)=01
0.5103t0.5103t,B(t)=0
1
and C(t)=[10];t∈{0,...,200}. The control task is to
track the following desired trajectory:
yd(t)=106(t1)3(40.03(t1)), t∈{0,...,200}.
(41)
In the simulation, the initial states in all iterations are set
as 0 and the input of the first iteration is 0.
The controller parameters are selected as: ˆ
θt
0(t)=0.9,
ρ=1, η=1, λ=0.5, and µ=0.1. For comparison, the
first-order, second-order, third-order, and fourth-order forms
of the proposed DDHOILC are applied respectively with the
same simulation conditions except that the higher order factors
are selected differently: α1=[1]for the first order, α2=
[0.9,0.1]for the second-order, and α3=[0.8,0.14,0.06]
for the third-order, and α4=[0.75,0.23,0.01,0.01]for the
fourth order. Note that α1
1
2
1
3
1
4
1and α3
1¯α3=
0.704
4
1¯α4=0.72 satisfy the conditions derived from the
convergence indexes.
It should be noted that in order to compare the proposed
DDHOILC and the traditional OILC under the same simula-
tion conditions, all the learning control laws start from the
third iteration because the calculation of the control input in
the fourth-order DDHOILC uses the error information from
the previous three iterations.
Fig. 2. Computation time in Example 1.
The simulation results are shown in Figs. 1 and 2. Fig. 1
demonstrates the asymptotic convergence of the proposed
nonlifted DDHOILC, where the y-axis is the mean absolute
value of tracking error emean (k)=200
t=1|yd(t)yk(t)|/200.
Obviously, the asymptotic convergence of tracking error can
be guaranteed by the proposed DDHOILC. Meanwhile, it is
demonstrated that a faster convergence can be achieved by
selecting the higher order factors according to the conditions
given in this paper.
Fig. 2 shows the computation time of the proposed
DDHOILC with third-order learning control law as an exam-
ple. The horizontal axis denotes the operating length of the
controlled process within each batch. The unit of the horizontal
axis is 200 instants per batch. The vertical axis denotes the
calculation time and the unit is seconds.
The calculation is performed in MATLAB on a Lenovo
ThinkPad laptop computer with a 2.40 GHz Intel Core i3
process and 2 GB of RAM. The batch length is changed
by selecting different sampling rate. It is verified that the
computation time of the proposed DDHOILC via nonlifted
iterative dynamic linearization increases very slowly with
the increasing size of data sets because no matrix inverse
operation is required for the implementation of the learning
control law.
For a comparison, an optimal ILC law is selected from [13]
Uk+1=(GTQG +R+S)1(GTQG +R)Uk
+(GTQG +R+S)1GTQEk(42)
where Q,R,andSare real-valued symmetric positive definite
matrices. By selecting Q=I,R=0.2I,S=0the simulation
results are also shown in Figs. 1 and 2.
From Fig. 1, it is seen that the control performance of
the proposed DDHOILC is better than that of the traditional
OILC because more historical and online control information
is employed. Only the historical control information in the
immediate previous one iteration is used in the traditional
OILC.
Fig. 2 shows that the computational time of the tradi-
tional OILC approach (42) increases dramatically as the
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
8IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
TAB L E I
PARAMETERS OF A CSTR SYSTEM
trial length increases. The matrix inverse operation in lifted
OILC becomes impractical with increasing dimension and data
points.
Example 2: In chemical industries, the continuously stirred
tank reactor (CSTR) is a highly nonlinear process with a com-
plex dynamic behavior [39]–[41]. In this paper, we consider
the concentration control of the CSTR, and the extensively
used model is shown as follows:
˙
CA=F
V(CA0CA)k0eE/RTRCA+σ1(k)(43)
˙
TR=F
V(TA0TR)+H
ρCpk0eE/RTRCA+Qσ
ρCpV(44)
where CAis the concentration of Ain the reactor (kmol/m3),
TRis the reactor temperature (K),Qσ=2.789 ×104denotes
the removed heat from the reactor (kJ/min),Vdenotes the
reactor volume (m3),k0is a pre-exponential constant (min1),
Eis the activation energy (kJ/kmol),His the reaction
enthalpy (kJ/kmol),Cpis the heat capacity (kJ/kgK),andρ
denotes fluid density (kg/m3). The parameter values are listed
in Table I [39]. 0 <CA0<2kmol/m3is the control input.
The control objective is to drive the system to an unstable
steady state Cas =0.57 kmol/m3where CA0is the control
input. The finite batch time is 3 min and the sampling time is
h=0.01 min.
In practice, a process can be affected by various distur-
bances. So, the measurement error σ1(k)of concentration
CAis considered in the simulation as shown in (43), which
randomly varies with both the time instants and the batch
numbers over the interval of [−0.01 0.01]. Besides, the ini-
tial concentration fluctuation in the inlet flow CA0is also
considered, that is, xk(0)=(0.47 +σ2(k))kmol/m3,where
kdenotes the batch number of the CSTR process, and σ2(k)
also randomly varies over the interval of [−0.01 0.01].
In the simulation, the initial input signal at the first
batch is selected as u0(t)=0.05 for all the time instants
t∈{0,1,...,299}; the controller parameters are chosen as
ρ=2, η=0.00001, λ=0.001, and µ=0.1. The
initial value ˆ
θt
0(t)=0.1. Select the higher order weighting
factors as α1=[1],α2=[0.9,0.1],α3=[0.7,0.2,0.1],
and α4=[0.6,0.38,0.015,0.005]for the corresponding first-
order, second-order, third-order and fourth-order algorithms,
respectively. Apparently, α1
1
2
1
3
1
4
1and α3
1¯α3=
Fig. 3. Tracking errors with respect to iterations in Example 2.
TAB L E I I
VALUES OF THE EVAL U AT E INDEX WITH DIFFERENT
ORDERS OF THE PROPOSED DDHOILC
0.56
4
1¯α4=0.576 also satisfy the conditions derived
in Section V.
The mean tracking error, emean(k)=300
t=1
|Cas CA,k(t)|/300, is shown in Fig. 3 by using the
proposed DDHOILC.
From Fig. 3, the convergence of the tracking error is clearly
seen. Note that the proposed DDHOILC is executed by using
the input and output measurements only, without requiring any
model information of the controlled nonlinear system. In this
sense, it is data driven and is applicable to complex nonlinear
industrial processes.
Furthermore, to evaluate the control performance of higher
order control laws numerically, a numerical index is defined as
JIE =
100
k=1
emean(k)
where emean(k)is the mean tracking error defined earlier.
Applying the above four different order algorithms,
the numerical indices are shown in Table II. It is clear
that the higher order algorithm can achieve a better control
performance than the lower order one.
It should be noted that it is difficult to apply the traditional
OILC [13] to the CSTR process because of its strong nonlin-
earities and strong uncertainties. Comparatively, the proposed
DDHOILC in this paper is data driven and can be directly
applied to such a nonlinear uncertain process.
VII. CONCLUSION
In this paper, a new historical and online DDHOILC is
proposed for a class of nonlinear and nonaffine discrete-time
systems, which repetitively run over a finite time interval.
A nonlifted iterative dynamic linearization is constructed for
the nonlinear plant at first, and then the DDHOILC is proposed
by designing a control objective function with higher order
factors introduced. More historical and online measurements in
previous iterations and in previous time instants of the current
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
CHI et al.: COMPUTATIONALLY EFFICIENT DDHOILC 9
iteration, respectively, are fully utilized to enhance the control
performance. It is shown theoretically that the convergence
property of the higher order learning control law can be better
than that of the lower order one by properly selecting the
higher order factors. Furthermore, the computational complex-
ity of the proposed DDHOILC is reduced significantly since no
matrix inversion is included in the proposed learning control
law. The proposed DDHOILC does not require any explicit
model information except for the I/O data and thus can be
more applicable to complex nonlinear processes in practice.
It is worth pointing out that a basic premise of ILC is that
the desired task should be performed under restricted repetitive
conditions such as identical initial condition and identical trial
length for all iterations. However, these assumptions can be
violated in many practical applications. Therefore, the results
will be further extended to the systems with nonrepetitive
uncertainties in our future work.
REFERENCES
[1] S. Arimoto, S. Kawamura, and F. Miyazaki, “Bettering operation of
robots by learning,” J. Robot. Syst., vol. 1, no. 2, pp. 123–140, 1984.
[2] N. Amann, D. H. Owens, and E. Rogers, “Iterative learning control for
discrete-time systems with exponential rate of convergence,” IEE Proc.-
Control Theory Appl., vol. 143, no. 2, pp. 217–224, Mar. 1996.
[3] N. Amann, D. H. Owens, and E. Rogers, “Predictive optimal iterative
learning control,” Int. J. Control, vol. 69, no. 2, pp. 203–226, 1998.
[4] D. H. Owens, B. Chu, and M. Songjun, “Parameter-optimal iterative
learning control using polynomial representations of the inverse plant,”
Int. J. Control, vol. 85, no. 5, pp. 533–544, 2012.
[5] D. A. Bristow, “Weighting matrix design for robust monotonic conver-
gence in norm optimal iterative learning control,” in Proc. IEEE Amer.
Control Conf., Jun. 2008, pp. 4554–4560.
[6] D. H. Owens, C. T. Freeman, and B. Chu, “An inverse-model approach
to multivariable norm optimal iterative learning control with auxiliary
optimisation,” Int. J. Control, vol. 87, no. 8, pp. 1646–1671, 2014.
[7] M. Volckaert, M. Diehl, and J. Swevers, “Generalization of norm optimal
ILC for nonlinear systems with constraints,” Mech. Syst. Signal Process.,
vol. 39, nos. 1–2, pp. 280–296, 2013.
[8] P. Axelsson, R. Karlsson, and M. Norrlöf, “Estimation-based norm-
optimal iterative learning control,” Syst. Control Lett., vol. 73, pp. 76–80,
Nov. 2014.
[9] W. B. J. Hakvoort, R. G. K. M. Aarts, J. van Dijk, and J. B. Jonker,
“Lifted system iterative learning control applied to an industrial robot,”
Control Eng. Pract., vol. 16, no. 4, pp. 377–391, 2008.
[10] K. L. Barton, D. A. Bristow, and A. G. Alleyne, “A numerical method
for determining monotonicity and convergence rate in iterative learning
control,” Int. J. Control, vol. 83, no. 2, pp. 219–226, 2010.
[11] J. K. Rice, and M. Verhaegen, “A structured matrix approach to efficient
calculation of LQG repetitive learning controllers in the lifted setting,”
Int. J. Control, vol. 83, no. 6, pp. 1265–1276, 2010.
[12] A. Haber, R. Fraanje, and M. Verhaegen, “Linear computational com-
plexity robust ILC for lifted systems,” Automatica, vol. 48, no. 6,
pp. 1102–1110, 2012.
[13] H. Sun and A. G. Alleyne, “A computationally efficient norm opti-
mal iterative learning control approach for LTV systems,Automatica,
vol. 50, pp. 141–148, Jan. 2014.
[14] Z. Hou and S. Jin, Model Free Adaptive Control: Theory and Applica-
tions. Boca Raton, FL, USA: CRC Press, 2013.
[15] K.-S. Kim and Q. Zou, “Model-less inversion-based iterative control for
output tracking: Piezo actuator example,” in Proc. Amer. Control Conf.,
Jun. 2008, pp. 2710–2715.
[16] P. Janssens, G. Pipeleers, and J. Swevers, “A data-driven constrained
norm-optimal iterative learning control framework for LTI systems,”
IEEE Trans. Control Syst. Technol., vol. 21, no. 2, pp. 546–551,
Mar. 2013.
[17] R. H. Chi, Z. S. Hou, B. Huang, and S. T. Jin, “A unified data-driven
design framework of optimality-based generalized iterative learning
control,” Comput. Chem. Eng., vol. 77, pp. 10–23, Jun. 2015.
[18] Z. Bien, and K. M. Huh, “Higher-order iterative learning control
algorithm,” IEE Proc. D-Control Theory Appl., vol. 136, no. 3,
pp. 105–112, May 1989.
[19] Y. Chen, C. Wen, and M. Sun, “A robust high-order P-type iterative
learning controller using current iteration tracking error,” Int. J. Control,
vol. 68, no. 2, pp. 331–342, 1997.
[20] Y. Chen, Z. Gong, and C. Wen, “Analysis of a high-order iterative
learning control algorithm for uncertain nonlinear systems with state
delays,” Automatica, vol. 34, no. 3, pp. 345–353, 1998.
[21] J. Hätönen, D. H. Owens, and K. Feng, “Basis functions and parame-
ter optimisation in high-order iterative learning control,” Automatica,
vol. 42, pp. 287–294, Feb. 2006.
[22] S. Gunnarsson and M. Norrlöf, “On the disturbance properties of
high order iterative learning control algorithms,” Automatica, vol. 42,
pp. 2031–2034, Nov. 2006.
[23] X. Ge, J. L. Stein, and T. Ersal, “A frequency domain approach for
designing filters for norm-optimal iterative learning control and its
fundamental tradeoff between robustness, convergence speed and steady
state error,” in Proc. Amer. Control Conf. (ACC), Boston, MA, USA,
July. 2016, pp. 384–391.
[24] X. Li, Q. Ren, and J.-X. Xu, “Precise speed tracking control of a robotic
fish via iterative learning control,” IEEE Trans. Ind. Electron., vol. 63,
no. 4, pp. 2221–2228, Apr. 2016.
[25] T. D. Son, G. Pipeleers, and J. Swevers, “Robust monotonic convergent
iterative learning control,” IEEE Trans. Automat. Control, vol. 61, no. 4,
pp. 1063–1068, Apr. 2016.
[26] J.-X. Xu and Y. Tan, “Robust optimal design and convergence properties
analysis of iterative learning control approaches,” Automatica, vol. 38,
pp. 1867–1880, Nov. 2002.
[27] R. Schmid, “Comments on ‘robust optimal design and convergence
properties analysis of iterative learning control approaches’ and ‘On
the P-type and Newton-type ILC schemes for dynamic systems
with non-affine input factors,”’ Automatica, vol. 43, pp. 1666–1669,
Sep. 2007.
[28] X. Wang, B. Chu, and E. Rogers, “Repetitive process based higher-
order iterative learning control law design,” in Proc. Amer. Control
Conf. (ACC), Boston, MA, USA, Jul. 2016, pp. 378–383.
[29] X. Wang, B. Chu, and E. Rogers, “New results on higher-order iterative
learning control for discrete linear systems,” in Proc. 10th Int. Workshop
Multidimensional Syst., Sep. 2017, pp. 1–6.
[30] Y.-S. Wei and X.-D. Li, “Robust higher-order ILC for non-linear
discrete-time systems with varying trail lengths and random initial state
shifts,” IET Control Theory Appl., vol. 11, no. 15, pp. 2440–2447,
Oct. 2017.
[31] D. Meng, Y. Jia, J. Du, and J. Zhang, “High-precision formation
control of nonlinear multi-agent systems with switching topologies:
A learning approach,” Int. J. Robust Nonlinear Control, vol. 25, no. 13,
pp. 1993–2018, 2015.
[32] D. Meng and K. L. Moore, “Robust cooperative learning control
for directed networks with nonlinear dynamics,” Automatica, vol. 75,
pp. 172–181, Jan. 2017.
[33] H.-F. Chen and H.-T. Fang, “Output tracking for nonlinear stochastic
systems by iterative learning control,” IEEE Trans. Automatic Control,
vol. 49, no. 4, pp. 583–588, Apr. 2004.
[34] E. I. Jury, Theory and Application of the Z-Transform Method.
New York, NY, USA: Wiley, 1964.
[35] L. Huang, Linear Algebra System and Control Theory. Beijing, China:
Science Press, 1984.
[36] Z. Hou and S. Jin, “A novel data-driven control approach for a class
of discrete-time nonlinear systems,” IEEE Trans. Control Syst. Technol.,
vol. 19, no. 6, pp. 1549–1558, Nov. 2011.
[37] D. Meng, Y. Jia, and J. Du, “Robust consensus tracking control for
multiagent systems with initial state shifts, disturbances, and switching
topologies,” IEEE Trans. Neural Netw. Learn. Syst., vol. 26, no. 4,
pp. 809–824, Apr. 2015.
[38] D.-H. Hwang, Z. Bien, and S.-R. Oh, “Iterative learning control method
for discrete-time dynamic systems,” IEE Proc. D-Control Theory Appl.,
vol. 138, no. 2, pp. 139–144, Mar. 1991.
[39] P. Mhaskara, N. H. El-Farrab, and P. D. Christofides, “Stabilization of
nonlinear systems with state and control constraints using Lyapunov-
based predictive control,” Syst. Control Lett., vol. 55, no. 8, pp. 650–659,
2006.
[40] W.-D. Chang, “Nonlinear CSTR control system design using an artificial
bee colony algorithm,” Simul. Model. Pract. Theory, vol. 31, pp. 1–9,
Feb. 2013.
[41] R. Chi, B. Huang, Z. Hou, and S. Jin, “Data-driven high-order
terminal iterative learning control with a faster convergence speed,”
Int. J. Robust Nonlinear Control, vol. 28, no. 1, pp. 103–119, 2018.
doi: 10.1002/rnc.3861.2017.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
10 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
Ronghu Chi received the Ph.D. degree from
Beijing Jiaotong University, Beijing China, in 2007.
From 2011 to 2012, he was a Visiting Scholar
with Nanyang Technological University, Singapore.
From 2014 to 2015, he was a Visiting Professor with
the University of Alberta, Edmonton, AB, Canada.
In 2007, he joined the Qingdao University of Science
and Technology, Qingdao, China, where he is cur-
rently a Full Professor with the School of Automa-
tion and Electronic Engineering. He has published
over 100 papers in important international journals
and conference proceedings. His current research interests include iterative
learning control, data-driven control, intelligent transportation systems, and
so on.
Dr. Chi has also served as a Council Member of the Shandong Institute
of Automation and a Committee Member of the Data-Driven Control, Learn-
ing and Optimization Professional Committee, and so on. He received the
Taishan Scholarship in 2016. He served as various positions in international
conferences. He was an Invited Guest Editor of the International Journal of
Automation and Computing.
Zhongsheng Hou (SM’13) received the bachelor’s
and master’s degrees from the Jilin University of
Technology, Changchun, China, in 1983 and 1988,
respectively, and the Ph.D. degree from Northeastern
University, Shenyang, China, in 1994.
From 1995 to 1997, he was a Post-Doctoral Fellow
with the Harbin Institute of Technology, Harbin,
China. From 2002 to 2003, he was a Visiting
Scholar with Yale University, New Haven, CT, USA.
In 1997, he joined Beijing Jiaotong University,
Beijing, China, where he is currently a Distinguished
Professor and the Founding Director of the Advanced Control Systems
Laboratory, and the Head of the Department of Automatic Control. He is also
the Founding Director of the Technical Committee on Data-Driven Control,
Learning and Optimization, Chinese Association of Automation. His current
research interests include data-driven control, model-free adaptive control,
learning control, and intelligent transportation systems.
Dr. Hou is an IFAC Technical Committee Member on Adaptive and Learning
Systems and Transportation Systems. His original contribution in model-free
adaptive control has been recognized by over 152 different field applications.
He has also done a lot pioneering work in Data-Driven Control and Learning
Control and organized many special issues on data-driven control in the
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
in 2011, the IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS in 2016,
and so on.
Shangtai Jin received the bachelor’s, master’s, and
Ph.D. degrees from Beijing Jiaotong University,
Beijing, China, in 1999, 2004, and 2009,
respectively.
He is currently a Lecturer with Beijing Jiaotong
University. His current research interests include
model-free adaptive control, data-driven control,
learning control, and intelligent transportation
systems.
Biao Huang (M’97–SM’11–F’18) received the
B.Sc. and M.Sc. degrees in automatic control from
the Beijing University of Aeronautics and Astronau-
tics, Beijing, China, in 1983 and 1986, respectively,
and the Ph.D. degree in process control from the
University of Alberta, Edmonton, AB, Canada, in
1997.
In 1997, he joined the Department of Chemical
and Materials Engineering, University of Alberta,
as an Assistant Professor, where he is currently a
Professor. He has applied his expertise extensively in
industrial practice. His current research interests include process control, data
analytics, system identification, control performance assessment, Bayesian
methods, and state estimation.
Dr. Huang is a fellow of the Canadian Academy of Engineering and the
Chemical Institute of Canada. He is the Industrial Research Chair in control
of oil sands processes with the Natural Sciences and Engineering Research
Council of Canada and the Industry Chair of process control with Alberta
Innovates. He is a recipient of a number of awards, including the Alexander
von Humboldt Research Fellowship from Germany, the Best Paper Award
from the IFAC Journal of Process Control, the APEGA Summit Award in
Research Excellence, and the Bantrel Award in Design and Industrial Practice,
and so on.
... priori model information and/or manual parameter tuning. Real-world applicability also poses a strong requirement w.r.t. the state of the art, because methods of RL (Heess et al., 2017;Tassa, Doron, Muldal, Erez, Li, Casas, Budden, Abdolmaleki, Merel, Lefrancq, Lillicrap, & Riedmiller, 2018;Tsounis, Alge, Lee, Farshidian, & Hutter, 2020), datadriven ILC (Chi, Hou, Jin, & Huang, 2018;Huo, Freeman, & Liu, 2019;Yu, Hou, Bu, & Yu, 2020) and data-driven MPC (Berberich, Köhler, Müller, & Allgöwer, 2020 are predominantly validated by simulation and not real-world experiments. Nonetheless, some methods of RL have been validated in real-world experiments (Kober & Peters, 2008;Kormushev, Calinon, & Caldwell, 2013;Peters & Schaal, 2008), but these methods require system-and task-specific prior knowledge in the form of good initial policies and are, hence, not plug-and-play applicable. ...
Article
The project “Multiscale Investigation Of Switchable Metallicity In Tunable Conductance Devices” en- compasses a comprehensive web application designed to revolutionize the way businesses handle MnS2 components, such as pressure switches, variable resistors, memory devices, batteries, and supercapacitors, through a series of interconnected modules. At its core, this application streamlines client interactions, optimizes MnS2 material processing, and ensures seamless application integration, all while leveraging advanced porosity analysis techniques. A significant enhancement in the Porosity Analysis module is the incorporation of the Decision Tree Regressor, a machine learning algorithm, to accurately evaluate the conductivity of materials. This addition not only boosts the precision of porosity and conductivity assessments but also empowers the system to predict material performance with higher accuracy, thereby enabling more tailored solutions for client requirements.
Chapter
ILC is a useful method for enhancing system efficiency by taking into account prior execution.
Article
The data-driven high-order pseudo-partial derivative-based model-free adaptive iterative learning control (HOPPD-MFAILC) is always slow to converge and difficult to have excellent tracking results. To address the problem, an improved high-order pseudo-partial derivative-based model-free adaptive iterative learning control (iHOPPD-MFAILC) with fast convergence is proposed. First, to reduce the impact of the initial value of the pseudo-partial derivative (PPD) on the convergence speed of the algorithm, the initial PPD is corrected by introducing the high-order model estimation error. Second, to reduce the influence of system noise on the control performance, the original HOPPD-MFAILC control law is improved by introducing time-varying iterative proportional and time-varying iterative integral terms. Then, the convergence of the proposed improved control algorithm is demonstrated by theoretical analysis. Finally, simulations and experiments on the ball screw motion system show that the proposed iHOPPD-MFAILC can track the desired trajectory better. In addition, iHOPPD-MFAILC has better robustness in the noisy environment and achieves better convergence as well as trajectory tracking performance under different initial PPD conditions. The proposed control scheme has excellent application potential in precision motion control.
Article
Batch processes are typically nonlinear systems with constraints. Model predictive control (MPC) and iterative learning control (ILC) are effective methods for controlling batch processes. By combining batch-wise ILC and time-wise MPC, this article proposes a multirate control scheme for constrained nonlinear systems. Two-dimensional (2-D) framework is used to combine historical batch data with current measurements. The ILC part uses run-to-run control with previous iteration data, and the MPC part uses real-time control with current sampled measurements. Real-time feedback-based MPC in the time axis and run-to-run ILC in the batch axis are combined to optimize the current inputs based on previous batch input–output data and real-time system measurements. Rather than achieving control objectives in a single batch, our design allows multiple batches to be executed successively. To establish the stability of the combined scheme, rigorous theoretical analysis is presented next. The combined scheme with improved performance is then validated through two illustrative numerical examples.
Article
This article concentrates on the challenging adaptive tracking problem of multi‐input and multi‐output (MIMO) nonlinear systems with unknown nonlinear dynamics, for which a novel optimization‐based data‐driven adaptive control (ODDAC) equipped with an extended dynamic linearization method is developed. Through considering MIMO nonlinear systems in the fully‐actuated case and over‐actuated case separately, the ODDAC is proposed consisting of a parameter updating algorithm and an adaptive control law. A new design of parameter updating algorithm is presented such that the estimation can be guaranteed to be bounded by a strict contraction process. By leveraging properties of the nonnegative matrix, a grouping‐based contraction mapping (GCM) analysis method is proposed for the convergence of tracking error. Notably, the GCM does not require the contraction mapping condition to hold at all time instants. The proposed ODDAC is data‐based, avoiding reliance on model information, and its validity is verified through simulations.
Article
Full-text available
In this paper, a novel high-order optimal terminal iterative learning control (high-order OTILC) is proposed via a data-driven approach for nonlinear discrete-time systems with unknown orders in the input and output. The objective is to track the desired values at the endpoint of the operation cycle. The terminal tracking errors over more than one previous iterations are used to enhance the high-order OTILC's performance with faster convergence. From rigor of the analysis, the monotonic convergence of the terminal tracking error is proved along the iteration direction. More importantly, the condition for a high-order OTILC to outperform the low-order ones is first established by this work. The learning gain is not fixed but iteratively updated by using the input and output (I/O) data, which enhances the flexibility of the proposed controller for modifications and expansions. The proposed method is data-driven in which no explicit models are used except for the input and output data. The applications to a highly nonlinear continuous stirred tank reactor and a highly nonlinear fed-batch fermentater demonstrate the effectiveness of the proposed high-order OTILC design.
Article
Full-text available
This paper presents an approach to deal with model uncertainty in iterative learning control (ILC). Model uncertainty generally degrades the performance of conventional learning algorithms. To deal with this problem, a robust worst-case norm-optimal ILC design is introduced. The design problem is reformulated as a convex optimization problem, which can be solved efficiently. The paper also shows that the proposed robust ILC is equivalent to conventional norm-optimal ILC with trial-varying parameters; accordingly, the design trade-off between robustness and convergence speed is analyzed.
Book
Model Free Adaptive Control: Theory and Applications summarizes theory and applications of model-free adaptive control (MFAC). MFAC is a novel adaptive control method for the unknown discrete-time nonlinear systems with time-varying parameters and time-varying structure, and the design and analysis of MFAC merely depend on the measured input and output data of the controlled plant, which makes it more applicable for many practical plants. This book covers new concepts, including pseudo partial derivative, pseudo gradient, pseudo Jacobian matrix, and generalized Lipschitz conditions, etc.; dynamic linearization approaches for nonlinear systems, such as compact-form dynamic linearization, partial-form dynamic linearization, and full-form dynamic linearization; a series of control system design methods, including MFAC prototype, model-free adaptive predictive control, model-free adaptive iterative learning control, and the corresponding stability analysis and typical applications in practice. In addition, some other important issues related to MFAC are also discussed. They are the MFAC for complex connected systems, the modularized controller designs between MFAC and other control methods, the robustness of MFAC, and the symmetric similarity for adaptive control system design. The book is written for researchers who are interested in control theory and control engineering, senior undergraduates and graduated students in engineering and applied sciences, as well as professional engineers in process control.
Thesis
Iterative learning control has been developed for processes or systems that complete the same finite duration task over and over again. The mode of operation is that after each execution is complete the system resets to the starting location, the next execution is completed and so on. Each execution is known as a trial and its duration is termed the trial length. Once each trial is complete the information generated is available for use in computing the control input for next trial. This thesis uses the repetitive process setting to develop new results on the design of higher-order ILC control laws. The basic idea of higher-order ILC is to use information from a finite number of previous trials, as opposed to just the previous trial, to update the control input to be applied on next trial, with the basic objective of improving the error convergence performance. The first set of new results in this thesis develops theory that shows how this improvement can be achieved together with a measure of the improvement available over a non-higher order law. The repetitive process setting for analysis is known to require attenuation of the frequency content of the previous trial error from trial-to-trial over the complete spectrum. However, in many cases performance specifications will only be required over finite frequency ranges. Hence the possibility that the performance specifications could be too stringent. The second set of new results in this thesis develop design algorithms that allow different frequency specifications over finite frequency ranges. As in other areas, model uncertainties arise in applications. This motivates the development of a robust control theory and associated design algorithms. These constitute the third set of new results. Unlike alternatives, the repetitive process setting avoids the appearance of product terms between matrices of the nominal system dynamics statespace model and those used to describe the uncertainty set. Finally, detailed simulation results support the new designs, based on one axis of a gantry robot executing a pick and place operation to which iterative learning control is especially suited.
Conference Paper
Iterative learning control is applicable to systems that make sweeps or passes through dynamics defined over a finite duration. Once each pass is complete all information generated as its dynamics evolve are available for use in designing the control action to be applied on the next sweep. The design problem is to construct a sequence of control inputs to enforce convergence to a specified reference of the sequence formed from the output produced on each pass and in this form of control the input is that used on the previous pass plus a correction term computed using previous pass output. A critical feature is the ability to use information that would be non-causal in the standard setting provided it is generated on a previous pass. Higher order iterative learning control uses information from more than the previous pass and is the subject of this paper where the generalized KalmanYakubovich-Popov lemma is used to develop new designs
Article
This study addresses a robust iterative learning control (ILC) scheme for non-linear discrete-time systems in which both the trail lengths and the initial state shifts could be randomly variant in iteration domain. The proposed higher-order ILC law guarantees that as the iteration number goes to infinity, the ILC tracking errors at the desired output trail period are bounded in mathematical expectation, and the bound of tracking errors is proportional to the random initial state shifts. Specifically, the ILC tracking errors in mathematical expectation can be driven to zero as the expectation of initial state shifts is zero. Two numerical examples are carried out to demonstrate the effectiveness of the proposed higher-order ILC law.
Article
This paper studies a class of robust cooperative learning control problems for directed networks of agents (a) with nonidentical nonlinear dynamics that do not satisfy a global Lipschitz condition and (b) in the presence of switching topologies, initial state shifts and external disturbances. All uncertainties are not only time-varying but also iteration-varying. It is shown that the relative formation of nonlinear agents achieved via cooperative learning can be guaranteed to converge to the desired formation exponentially fast as the number of iterations increases. A necessary and sufficient condition for exponential convergence of the cooperative learning process is that at each time step, the network topology graph of nonlinear agents can be rendered quasi-strongly connected through switching along the iteration axis. Simulation tests illustrate the effectiveness of our proposed cooperative learning results in refining arbitrary high precision relative formation of nonlinear agents.
Conference Paper
This paper focuses on Norm-Optimal Iterative Learning Control (NO-ILC) framework for Single-Input-Single-Output (SISO) Linear Time Invariant (LTI) systems and considers the filter design problem in frequency domain. Modeling uncertainty, in general, degrades the performance of NO-ILC. Hence, ensuring Robust Monotonic Convergence (RMC) against modeling uncertainty is important, but the state-of-art filter design techniques lead to a conservative performance. To ensure a more aggressive performance, a new modeling uncertainty formulation is proposed in this paper and a frequency dependent filter design is proposed. Through a frequency domain analysis, an equation characterizing the fundamental trade-off of NO-ILC with respect to robustness, convergence speed and steady state error at each frequency is presented, which allows the design of NO-ILC to achieve different performance requirements at different frequencies. Simulation examples are given to confirm the analysis and demonstrate the utility of the developed filter design technique.
Article
In this paper, we present a novel work in which an iterative learning control (ILC) method is applied to a two-link Carangiform robotic fish in real time and achieves precise speed tracking performance. By virtue of the Lagrangian mechanics method, we establish a mathematical model for the robotic fish. The robotic fish model is highly nonlinear and nonaffine in control input, which hinders the applicability of most control methods that require affine-in-input. ILC is suitable because it works for such circumstances. A P-type ILC algorithm is adopted for speed tracking tasks of the robotic fish. The rigorous convergence analysis is derived based on composite energy function (CEF). In practice, the precise model of robotic fish is difficult to be obtained due to many uncertain factors. By employing ILC, the speed tracking control performance can be improved significantly without using the perfect model. Both simulations and experiments are conducted to illustrate the effectiveness of ILC, and excellent speed tracking is achieved for the robotic fish.