90 4 Learning-based Control
The existance of a Lyapunov function assures stability as given by the
following theorem.
Theorem 1: If V is a Lyapunov function of (4.30) in some neighborhood
of an equilibrium state X = 0, then X = 0 is a stable equilibrium.
If in addition − ∆V is positive definite with respect to X = 0, then the
origin is asymptotically stable.
So far, the definition of stability and asymptotical stability are in terms
of perturbations of initial conditions. If the model error e is small, one hopes
that at least qualitatively, the behavior of the original system and that of the
perturbed one will be similar. For the exact relation, stable under perturba-
tionsneeds to be defined [94].
Definition 3: Let X ( X
0
, t ) denote the solution of (4.30) with the initial
condition X
0
= X ( X
0
, 0). The origin X = 0 is said to be stable under pertur-
bations if for all > 0 there exists δ
1
( ) and δ
2
( ) such that ||X
0
|| < δ
1
and
||e ( t, X ) || < δ
2
for all k > 0 imply X ( X
0
, t ) < for all t ≤ 0.
If in addition, for all there is an r and a T ( ) such that ||X
0
|| < r and
||e ( t, X ) || < δ
2
( ) for all t > 0 imply ||X ( X
0
, t ) || < for all t > T ( ), the
origin is said to be strongly stable under perturbations(SSUP).
Strongly stable under perturbations (SSUP) means that the equilibrium is
stable, and that states started in B
r
⊂ Ω actually converge to the error bound
at limited time. Ω is called a domain of attraction of the solution (while the
domain of attraction refers to the largest such region, i.e., to the set of all
points such that trajectories initialed at these points eventually converge to
the error bound.
With this in mind the following theorem [61], [94] can be stated:
Theorem 2: If f is Lipschitz continuous in a neighborhood of the equilib-
rium, then the system (4.30) is strongly stable under perturbations iff it is
asymptotically stable.
In this paper, Support Vector Machine (SVM) will be considered as a
neural network structure to learn the human expert control process. In the
next section, a rough introduction to the SVM learner that we will use, will
be provided.
Convergence Analysis
There are many kernels that satisfy the Mercer’s condition as described in
[27]. In this paper, we take a simple polynomial kernel in Equation (4.19):
K ( X
i
, X ) = ((X
i
· X ) + 1)
d
, (4.32)
where d is user defined (Taken from [108]).
After the off-line training process, we obtain the support values ( α and
α
∗
) and the corresponding support vectors. Let X
i
be sample data of X .
By expanding Equation (4.19) according to Equation (4.32), Let
ˆ
f ( X )=