File loading please wait...
Citation preview
=
STOCHASTIC SYSTEMS "Theory and Applications
Digitized by the Internet Archive in 2023 with funding from Kahle/Austin Foundation
https ://archive.org/details/stochasticsystem0000puga
STOCHASTIC SYSTEMS V. S. Pugachev I. N. Sinitsyn Russian Academy of Sciencies
Translated by
|. V. Sinitsyna
Ve World Scientific NewJersey
«Londons Singapore e Hong Kong
Published by World Scientific Publishing Co. Pte. Ltd. P O Box 128, Farrer Road, Singapore 912805 USA office: Suite 1B, 1060 Main Street, River Edge, NJ 07661
UK office: 57 Shelton Street, Covent Garden, London WC2H 9HE
British Library Cataloguing-in-Publication Data A catalogue record for this book is available from the British Library.
STOCHASTIC
SYSTEMS
Theory and Applications
Copyright © 2001 by World Scientific Publishing Co. Pte. Ltd. All rights reserved. This book, or parts thereof, may not be reproduced in any form or by any means, electronic or mechanical, including photocopying, recording or any information storage and retrieval system now known or to be invented, without written permission from the Publisher.
For photocopying of material in this volume, please pay a copying fee through the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to
photocopy is not required from the publisher.
ISBN
981-02-4742-7
Printed in Singapore by World Scientific Printers (S) Pte Ltd
PREFACE
The monograph is dedicated to the systematic presentation of the applied advanced theory and analytical methods for the stochastic systems, i.e. the dynamical systems described by the finite- and infinitedimensional stochastic differential, difference, integral, integrodifferential etc. equations. It is based on the results of the fundamental research performed in the Institute for Informatic Problems of the Russian Academy of Sciences in the context of the scientific program “Stochastic Systems” and the lecture courses-delivered by the authors in the domestic and the foreign Technical Universities. The book may be used as the textbook for the applied mathematics faculties of the universities. The unified procedure, thorough selection of the examples and the problems (over 500) and the large applications make this book useful for the graduates, the post-graduates and the lecturers. The book is of considerable interest also for the mathematicians who deal with the stochastic equations and their applications. The construction of general stochastic systems theory is based on the equations for the multi-dimensional characteristic functions and functionals. The stochastic differential equations with the arbitrary processes with the independent increments are studied. The equations with the Wiener and the Poisson processes are considered as a special case. The methods of the parametrization of the multi-dimensional distributions in the nonlinear stochastic systems based on the moments, the semiinvariants, the quasimoments and on the consistent orthogonal expansions are stated systematically. Special attention is paid to linear and reducible to linear stochastic systems theory based on the canonical representations (canonical expansions and integral canonical representations). Most attention has been concentrated on the structural theory of the composed stochastic systems on the grounds of the conditional probability measures. Chapter 1 is devoted to the mathematical models of the dynamic systems under conditions of the random disturbances and their characteristics. The linear and the nonlinear continuous, discrete and continuousdiscrete systems described by the stochastic equations in the finite- and the infinite-dimensional spaces are considered. Special attention is paid to the composed stochastic systems. The main notions of probability distributions theory of the random variables, the random processes and the random functions are stated in Chapter 2. After determining the probability spaces, the conditional
vl
Preface
probabilities, probabilities in the finite and the infinite products of the spaces are considered. The conditions of the existence of the regular probabilities are established. Further the different probability measures of the random functions and the probabilities of events connected with the random functions are studied in detail. The last sections are devoted to the distributions of the separable random functions, to the derivation of the criteria of the continuity and the differentiability of the random functions. Chapter 3 contains the account of the basic notions connected with the expectation of the random variables in Banach and topological linear spaces, the properties of the operators of the moments of the second and higher orders, the properties of the characteristic functionals and the correspondence between the probability measures and the characteristic functionals are established. The conditional operators of the moments and their properties are studied in detail. Special section is devoted to the elements of the probabilities distribution in the infinite-dimensional spaces. The properties of the normal distribution on the product of two spaces, the conditional normal distributions and the normal distributions in Hilbert spaces are considered. Approximate representation of the oneand the multi-dimensional distributions based on the consistent orthogonal expansions is treated. Special attention is paid to the semiinvariant and the quasimoments. A systemical presentation of general theory of the canonical expansions is given. Practical ways of the canonical expansions constructing are described. The joint canonical expansions are considered. Chapter 4 is devoted to the selected topics of stochastic calculus. The theory of the convergence, the continuity and the differentiability is considered. The integrability of the random functions over nonrandom and random measures is discussed. The theory of the stochastic Ito integrals, the symmetrized (Stratonovich) integral and the 6-integral is outlined systematically. The stochastic integrals over the Wiener and the Poisson measures are studied. The connection between the processes with the independent increments and the It6 integral is established.
The elements of the theory of the generalized random functions are stated. A systematic account to the integral canonical representations is given. The connection of the integral canonical representations with the spectral one is established. Practical ways of constructing the integral canonical representations and, in particular, the methods of the shaping
filters are considered. studied.
The joint integral canonical representations are
~~
Preface
vil
The general theory of the finite- and the infinite-dimensional stochastic systems is given in Chapter 5. At the beginning the differentiation of the Ito process is set out in detail. The Ito formulae for the differentiation of the composite functions in the cases of the Wiener, the Poisson and the general random process with the independent increments have a significant place. The theory of the stochastic integral, differential and integrodifferential Ito equations including the methods of reducing the state equations to the stochastic differential equations, change of variables, sufficient conditions of the existence and the uniqueness of the mean square solutions are considered. The central problem of stochastic system theory —he problem of calculation of the multidimensional distributions is stated. The elements of numerical integration of the stochastic differential equations are given. The theory of the one- and the multi-dimensional distributions in the continuous, discrete and continuous-discrete systems in the finite-dimensional spaces are stated in detail. For such systems the known equations (Fokker—Plank— Kolmogorov, Feller-Kolmogorov, Pugachev and their generalizations) are derived. The problems of one-dimensional and multi-dimensional distributions theory on the basis of the equations for the correspondent characteristic functionals in composite stochastic systems including the stochastic systems in the Hilbert spaces are considered. The main statements of the structural theory of the composed stochastic systems on the basis of conditionally probability measures are outlined. The last section contains the applications to optimal online filtering problems. Simple derivation of the Kushner—Stratonovich and the Zakai equations for the normal white noises is presented. Corresponding equations for the composite stochastic systems are given in Problems. Chapter 6 contains the account of the main methods of linear stochastic systems theory in connection with continuous, discrete and continuous-discrete systems. The spectral and correlation methods, the methods of general theory of linear stochastic systems are outlined. The correlation methods based on the canonical representations are considered with great attention. The general theory of the canonical representations of the linear transformations is given. The separate subsections are devoted to the solution of the linear operator equations by the canonical representations methods. The applications to the offline spectral and correlation analysis and the shaping linear filters design are given. Software for the analytical modeling problems is described. The Kalman-Bucy and the Liptser—-Shiryaev online filters are considered.
vill
Preface
Chapter 7 is devoted to the approximate methods of nonlinear finite- and infinite-dimensional stochastic systems theory based on the parametrization of the distributions. The normal approximation, statistical and the equivalent linearization methods are considered. The Genenormalization of the composite stochastic systems is treated. ral problem of parametrization of the multi-dimensional distributions is stated. The methods of the moments, the semiinvariants, the quasimoments and the orthogonal expansions are given: The practical problems of reducing the number of equations for the automatic derivation of the equations and the corresponding software for distribution parameters are considered. The solution of the structural parametrization problem of the multi-dimensional distributions is outlined on the basis of the ellipsoidal approximation method. The canonical representations methods for the nonlinear stochastic systems are considered. The ba-
sic methods are given for stochastic systems with the normal white noises. Corresponding methods for the composite stochastic systems are given in Problems. The applications of the given methods to the nonlinear stochastic systems offline and online analysis, modeling and filtering problems in mechanics, physics, biology, chemistry, finance and control are stated. Specialized software is described. Chapters 1, 5-7 give efficient toolkit for the analytical modeling and the design of the stochastic systems. To facilitate the mastering of the stated methods about 500 thoroughly selected examples and problems are given in the correspondent subsections, sections and chapters of this book. Appendices 1-7 contain auxiliary materials which include the information about tables of stochastic It6 differentials for the typical composite functions, the Hermite polynomials of a vector argument, the polynomials orthogonal to gamma and y?-distributions, the tables of the coefficients of statistical linearization for the typical composite functions, integrands evaluations in the equations of the basic methods of distribution parametrization, the tables of a basic characteristics and transformation the structural rules of the linear systems and the tables of some definite integrals and special functions. For information about the notions and theorems from vatious parts
of mathematics used in the book we advise (Korn and Korn 1968).
Bibliographical notes and References are given at the end of the book. The authors do not pretend in any way to provide a complete bibliography. Only those sources are given in the list which are cited in the text. Basic notations are presentented before Index.
~
Preface
ix
The book is divided into chapters, sections and subsections. The formulae (examples) are numerated with indication of the number of the chapter and the section, for instance, (1.2.3) means that formula (3) belongs to section 2 of chapter 1. The problems are numerated only with indication of the number of chapter. At each Appendix the separate numeration is accepted. For the convenience of the readers the formulations of all basic results and statements are given in the separate theorems and the corollaries or in the text italics. The beginnings and the ends of the evaluations, the proofs and the discussions which lead to certain results are indicated by triangular indices p and Let u(t, a ) be the solution of the homogeneous equation
= au
(1.2.2)
satisfying the initial condition u(r ie ) = I, i.e. the fundamental matrix of solutions
of Eq. (1.2.2). By the change of variables
7 = u(t, to)v the first Eq. of (1.2.1) is
reduced to the form
uv + uv = auv
+.a,z2+ a9
or by virtue of Eq. (1.2.2)
uv =ayr+ao. Remark.
It is known from differential equations theory that the determinant
of the matrix u(t, to) is expressed by t
A = |u(t,to)| = exp
J vra(ryar
(1:2:3)
to
where tr a(T) is the trace of the matrix alr): To derive this formula it is sufficient to notice that according to the differentiation rule of a determinant and the properties
of the determinants
A =
condition A(to) =
|u(to, to)| = 1 is determined by formula (1.2.3).
from this formula that
Atr a(r i The solution of this equation with the initial
It follows
A # 0 (we take for granted that the integral in (122.3)/is
finite at all t, 9 owing to which the exponential function does not vanish). Hence, taking into account that the matrix u is always invertible we get
v=u(aye+ ao).
1.2. Linear Differential Systems
25
Integrating this equation with the initial condition v(to) = u(to, to)~+2z(to) = 74 gives t
v(t) = zo + it u(r, to) | [ai(r)a(r) + ao(r) Jdr,
(1.2.4)
and
z(t) = u(t, to) v(t) = u(t, to)zo 1.2.5
+ u(t,to) fu(r, to)* [ar(z)e(r) +ao(r)|dr.< to : Let us prove that at any to, 7, t, tp
(
)
Putting u = u(t, to),
v= u(t, to).= u(t,to)4 , we have at any tp andt
> tp uv!
=
I.
(1.2.11)
Differentiating this formula we get:
uyve + uve = 0. So yielding by Eqs. (1.2.2), (1.2.11) we have
uv; = —auv’ = —a ort? = —u“la.
(i. 2612)
Hence by virtue of Eq (1.2.11) we obtain the differential equation for UV: vr = —-y!a,
Now we notice that according to (1.2.6) and (1.2.11) u(t, 7) = u(t, to)v(7, to)’. Differentiating this formula with respect to 7 and taking into account that according
to (1.2.12) 0;(7, to)? = —v(r7, to)? a(r), we find i,
(t,T) = u(t, to)vr(7, to)? = —u(t, to)v(7, to)? a(r) = —u(t, r)a(r).
Thus u(t, 7) as a function of T at fixed t is determined at T < t by the equation
u,(t,7) = —u(t, T)a(r)
leah)
with the initial condition u(t,t) = J att =t. 4 Transposing Eq. (1.2.13) we obtain
UAT)
=
alr) ulter)? «
Cle2ei4}
This equation is the adjoint equation to Eq. (1.2.2). Therefore u(t, 7)? as a function of T at fixed t is determined at 7 < t¢ by Eq. (1.2.14) adjoint to Eq. (1.2.2) with the condition u(t,t)” = I at the end. So u(t, 7) may be found by the integration of the adjoint Eq. (1.2.14) in reversed time
with the initial condition u(t,t) = I. 1.2.3.
Transformation of Linear System Equations into Cauchy Form
If the input-output relation of a system is represented in the form of a differential equation of higher than the first order then it usually contains not only the derivatives of the output but the derivatives of the input as well. Therefore, the problem arises to transform the equation
28
Chapter
1 Dynamical Systems and Their Characteristics
of a linear system into a set of equations of the first order which do not contain the derivatives of the input. Now we shall show how to do this. Let us consider a nonanticipative system whose behaviour is described by the linear differential equation of the form
See
Ybe,
K=O
k=0
(1.2.15)
where the coefficients ao, a1, ..., Qn, bo, b1, ... , bm depend in the general case on time t, and m < n.. It is evidently sufficient to consider the case of m = n. The case of m < n is the special case where Onis
Ort
=
());
> Let us introduce new variables
Z=Y—
QoL,
241
where qo, 41, ---, Yn—1
= 2% — get (K=1,...,n-—1),
(1.2.16)
are some functions which we shall determine from the
condition that the resulting equations of the first order contain no derivatives of the
input Z.
From Eqs.
(1.2.16) using the known formula for the derivatives of the
product of two functions
7
Uv Che SS en PuP)y Ae CrCee =
me
»
|
ee
1.2.17 ve
p\(r — p)!
(
we find
Y¥=%1+ 402, 8
Soon
yo) = Zs41 + Ge)
= 2540
em
r=0
Pg
Pal)
(1.2.19)
r=0,p=0
(SP
oe
eae
After changing the order of summation we get
NT NG +
Lorde? ((Sad
teen 1). » «(1-3.20)
Differentiating formula (1.2.19) corresponding to s = n — 1 gives
mee y\" ” ain
#9CES—r— 12)yet) =z, +
(of ("),t)
(1.2.21) Ahh ait 5mqd, + al Sor? p=1
r=p
1.2. Linear Differential Systems
29
Substituting the expressions (1.2.18), (1.2.20) and (1.2.21) into Eq. (1.2.15) we obtain
nin
+
> a-izi+e
AP
ip!
Ueeoee Da
ail
s=0
=
a >x) eos CPa,q’,”) >, = al?) , Sy
(1.2.22)
i)
Equating the coefficients at the corresponding derivatives of the input & in the leftand right-hand sides of Eq. (1.2.22) we obtain the equations for the functions gy:
Mee ce “r)=b, = (p=1,...,n).
Gres)
s=pfr=p
The last of these equations corresponding to p = 7 gives
Gh= an'b,.
(1.2.24)
For solving the other Eqs. (1.2.23) notice that
YS s=p
Lead? reppe Se r=p
s=p
i009
pI
1=0
n—pn—p—l
!)
.ye yg Pi teds--p-1 =) = 1=0 s=p+l
(=O
n—-pn—p—h
Fa
(1)
Dang pti aptlthTh =O
n—p—In—p—-—h
C.Pap erEn gy? = Gn n—p T yO h=0)
Y=0
h=0
>s Coeitptlthdy [=0
Substituting this expression into (1.2.23) and putting n — p= k-1k-h
=:
Onqk + Dey ny) Chaepian—ethtiq,. h=0
()
@) ip
(1.2.25) k we get
=
= on-k (K=1,-..,2—1).
7=0
(122.20) From
(1.2.26) we obtain the recursive formula for determining the functions q;,
+) 9n-1: k-1k—-h
h=0 [=0
(1.2.27)
30
Chapter 1 Dynamical
Systems and Their Characteristics
After determining in such a way the functions qo, 41,---,Qn—1
we transform
Eq. (1.2.22) into the form n
n
n-1
8
pee
Onin + 0 az = { ty Sod ree t=
l=1
sa)
Solving this equation for Zp, gives together with Eqs.
k =1,...,m-—1
hae)
1=0
(1.2.16) corresponding
to
the equation forming the set of the first order equations Ze = Ze41
+ Mee
Zn =
(kerk,
Ss Wy od
=
n
-Soa;
(1.2.29) Qj=121
+ nk ,
where
heathy
a) Ome ya, (1.2.30)
s—0
l=i
n-1
n
s
= an1 |bo— ) fangs — Soas Dain =
I
h=0
$=1
Si
Noticing that n
s
Sal
l=1
)
non
(a) I
n
n—l
n-1n-h
dias Doar = DOD oases=DOD ansiah? = DO DTansiah? (i)
gel!
i=h=)
(=O
a
(1.2.31)
we transform the formula (1.2.30) for gn into the form n—-1ln-h
Grit= Gawlibax My: Ds ansiqe?
:
(162.32))
h=0M=0
It is evident that this formula coincides with (1.2.27) at k = n. Thus all the functions
qi, --+ 5 Un are determined by (1.2.27). The
=---=
case
of ™
» Aay=
Bere
5
Qqn-1
ace eGOne
In
(1.2.35)
In the case of m = n the output y depends explicitly on the input zx. But this is a rare case since m < n practically always yields qg = 0. The
expression (1.2.34) of the output y does not contain z in this case. +,
In the special case of a stationary system all the coefficients aj, 2n; 00 01, +5 Unare Constant. ald thus Gp, d1,---,dn Gevermined
successively by formulae (1.2.24) and (1.2.27) are constant. qi? Oe Ut ree Ly ee considerably simpler in this case:
Therefore
an loniitiac (1.2.21) become
k-1
Go
Ge Dene Te te
(50-1-
So tm-tena) i ae reared Oe
Formulae (1.2.33) in this case take the form w=
7a
di =
(5.—
>a h =n—m
=n
me
oat
— 04
Gree
a
On
(k =n—m+1,...,
mn). (1.2.38)
32
Chapter 1 Dynamical
Systems and Their Characteristics
Notice that all previous calculations and consequently formulae (1.2.27), (1.2.33), (1.2.37) and (1.2.38) for go, 41, ---, Qn are valid both for the scalar and the vector input x and the output y. In the latter case ao, Gy js » Ope. B05 O15 1284 Ony Weds a> > Gea te plenent, matrices of the corresponding dimensions. Example
1.2.2. In order to transform the equation
a3 ¥ +aoy + ayy + aoy = bet + bit + bo
(1)
into a set of the first order equations we put according to (1.2.16) and (1 .2.33) OS
Ai,
22 = 21 — q12, 23 = 2 — Gov. Then by (1.2.33) we find
qo=0,
q1 = a3'b2, go = a3 '(b1 — aoqi — 20191),
q3 = 43 ‘(bo — 4191 — 4241 — 4391 — 4292 — 4342) As a result we obtain Eqs. (1.2.29) which has in this case the form 3 Ai—Zo7+Ne
,
‘ 232—23 4-090
,
: SI 23=—azg (aoz1
+4122
+4223)+
932 ,
Y
en =21.
(11) 1.2.4. Inverse Systems The system inverse to a system A is such a system whose output is equal to the input of the system A when its input is equal to the output of A (under the corresponding initial conditions). Thus the inverse system performs the transformation of signals inverse to that performed by a given system A. It is evident that if the system B is inverse to the system A then the system A is inverse to the system B. In other words, A and B represent two reciprocal or mutually inverse systems. The amplifier with the gain k and the amplifier with the gain 1/k, the differentiator and the integrator may serve as examples of pairs of mutually inverse systems.
If a system with the input «# and the output y is described by Eq.
(1.2.15) then the inverse system is described by the same differential equation in which y serves as its input and z serves as its output. If m = n, then the reciprocal systems described by differential equa-
tion (1.2.15) are of the same type (i.e. both differential systems). After transforming the equation into a set of the first order equations, as we have already seen in Subsection 1.2.3, the output of each of these two
1.2. Linear Differential Systems
33
systems will contain its input with some coefficient and the remaining part of the output will be determined by the corresponding set. of the first order differential equations. If m = 0 then the output of the inverse system is given by
2 = b5* > apy).
(1.2.39)
It goes without saying that for the existence of the inverse system in this case it is necessary that the inequality bo # 0 be valid at all t in the case of an one-dimensional system and that the matrix bo be invertible at all ¢ in the case of a finite-dimensional system. Under this condition the inverse system performs the linear differential operation
Lig) aD
ya Fa dte
(1.2.40)
k=0
> If 0 < m
< n then the operations which are performed by the inverse system
include the (n = m)-tiple differentiation of the input. As a result the output of the inverse system contains a linear combination of the input and its derivatives up to
the (n = m)*h order. In order to determine the inverse system it is necessary in this case to find the remaining part of its output. For this purpose we put nm—m
p=zt+nazt Do ay
(1.2.41)
1=0
and determine the coefficients Cg, Cj , ...
, Cn—m
In such a way that the variable Z
contains no linear combination of the input Yy and its derivatives.
Using the known formula (1.2.17) of the derivatives of the product of two func-
tions we find from (1.2.41)
Bie aom
ie Se ssCree y(t 1=0
r=0
=a? Ne Se Cre!(s—r) yltr) a= r=0)1=0
4
Cre os yl)
(1.2.42)
After this evident transformation of the double sum it is expedient to change again
the order of summation in order to obtain the coefficient at y6*) in the explicit form.
34
Chapter 1 Dynamical
Systems and Their Characteristics
For determining the limits of the inner sum we notice that0 k —n +m and
min(k,s)
ve
k=0
the order of summation
are possible we come to
Crea y(®),
Pe
s=0
0
min(k, s). As a result we obtain
or eltSyl) ,
r=max(0,k—n+m)
and
(1.2.43)
taking into account
that
k
0 weget
5Hb,x) = snbp z6*) s=0
+
3
k=0
min(k,s)
ha
DyPnnidesu CBee
k=0 s=max(0,k—n+m)
Eyes
masmenaeaay
r=max(0,k—n+m)
Let us substitute this expression into Eq.
(1.2.15) and equate the coefficients at
y™) | wens 5 y”) in the left- and the right-hand sides of the equality obtained.
taking into account that min(k, s) = s at s m
s
3
)
SF
Gthicaaaz
(b= sit pederp ny)
s=max(0,k—n+m) r=max(0,k—n+m)
In order to solve these equations for Cg, Cj , ... , Cn—m
order of summation.
m
Se
Then
m we get
(1.2.45) we change once more the
Remembering that r < 5 we get
m
Y Clb clbansea, (b= 9) oo)
(1.2.46)
r=max(0,k—n+m) s=r
or putting
s
=r+h
m
3. r=max(0,k—n+m)
Mm—Tf,
h=0
Clinbrgnce?,=a, (k= m,...,n).
(1.2.47)
1.2. Linear Differential Systems
At k =
n the sum
35
contains only one item corresponding
Therefore, we have at
k = n the relation bmCn—m
to
r =
m,
Gee eeb aa Separating in the item corresponding to
only one value
beCh om
h = 0 when r =
(1.2.48)
T = Mm and taking into account that h has
m—r
ys
De Cotte h=0
ol flWem
(1.2.49)
7 < ™), these equations can be solved for Cx_m. m—1
Cela = 07,
toages
>k—m
for all
As a result we obtain
m—r
ds
ec
r=max(0,k—n+m)
After determining Cyn—m
areal Cae be
.
As the sum here contains only Cc; with | greater than k — m (k —r items since
we
emp.
h=0
eal
(1.2.50)
by (1.2.48) we may find successively Cp—m—1,---,
Co using formula (1.2.50).
0.
™ we transform Eqs. (1.2.45), k > be z'*) =
»» a,y”) ;
k=0
k=0
(1.2.51)
where
m—1
ide
min(k,s)
3
Sy
‘ea ane.
(1.2.52)
s=max(0,k—n+m) r=max(0,k—n+m)
it 2.0piiynenotan— 1): Changing the order of summation here and taking into account that max min(k, s) s
=katk Giese”
r=max(0,k—n+m) s=r
(1.2.53)
Systems and Their Characteristics
Chapter 1 Dynamical
36
or
k
m—r
r=max(0,k—n-+-m)
h=0
Ge in), —
Sy
SiC hidrthle alk SU gleg cnc Mm),
(1.2.54)
Formula (1.2.41) shows that the system inverse to the system described by Eq. (e215)
at 0
|o)
lie ee
2 lrg56 ala let Hy |
y; (S )
7:
6)
—>
hs
yHy,ty,-);
Ys;
y)
— > y=y,ty,ty, y;
Fig. 1.3.3 A block diagram is called a graphical image of the system structure or its blocks. Here the structure is regarded the aggregate of the blocks and the connections representing the loops. The elements of a block
diagram are (Fig.1.3.3): e Linear dynamic links presented by the blocks with the operators describing the functions filled in the blocks or in the description applied (Fig.1.3.3a). e Linear and nonlinear functional generators (Fig.1.3.3b). e The summators presented by the circles divided into the sectors. Generally the summators realize the algebraic summation. Some times we use + signs in circles (Fig.1.3.3c). e Couplings which transmit the directed action without changing (Fig.1.3.3d). e Branching points on the coupling lines (Fig.1.3.3e). Equally with the block diagrams the oriented graphs are often used (Fig.1.3.4a,b,c). The main elements of an oriented graph are edge, node and arrow. The oriented graphs do not need special notations for summators and couplings as in the block diagrams.
1.3. Conneétions of Systems
45
qg)
Fig. 1.3.4 Transformation rules for block diagrams and graphs are given in Subsection 1.3.4 and Appendix 6. 1.3.2.
Weighting Functions of Linear Systems Typical Connections
Consider at first a series connection of two linear systems whose weighting functions gi(t,7) and go(t,7) are given (Fig.1.3.5). Find the weighting function g(t, 7) of a composite system.
> For this purpose we find its response on the unit impulse. As a result of the action of the unit impulse on the input of the first system its output will represent its weighting function 91 (t, af). Consequently, the input of the second system is equal
a(t
gityn).
(17371)
Here the input of the second system according to general formula (1.1.8) is expressed by formula [oe)
foe}
yo(t) = J s2lt,o)22(0) a0 = J 92lt,0)ni(0,7) do. —00
—oo
(1:3:2)
46
Chapter 1 Dynamical
Systems and Their Characteristics
But the output of the second system represents also in a given case the output
of the whole series connection of the considered systems, i.e. the weighting function obtained as a result of the connection of a system
(1.3.3)
yo(t) = g(t,7).
Comparing this formula with (1.3.2) we obtain the following formula for the weighting function of a series connection of two linear systems
gts
J salt-o)au(er) do. 4
(1.3.4)
Thus the weighting function of two connected in series linear systems is determined by formula (1.3.4). In the special case of nonanticipative systems we have g;(c,7) = 0 at o < 7 and go(t,0) = 0 at o > t. Consequently, for nonanticipative systems an integrand in (1.3.4) differs from zero only in the limits +
For determining the weighting function g(t, ai) of a composite system we shall find its response on the unit impulse 6(t = a) Accounting that while delivering the unit impulse 6(¢ = T) on the inputs of connected systems the outputs are equal to
the corresponding weighting functions g1 (t, T) and ga(t, T) we obtain
g(t, 7) = gi(t, 7) + ga(t, 7). 4
(1.3.7)
Thus at the parallel connection of the linear systems thetr weighting functions are added. For studying the systems with a feedback we use the notion of the
reciprocal (mutually inverse) systems (Subsection 1.2.4). Consider a series connection of two reciprocal linear systems. The weighting function of the linear system inverse relatively to the system with the weigh-
ting function g(t,7) we shall denote by the symbol g~ (t, rT). The weighting function of a series connection of two reciprocal systems is equal to 6(t—7). On the other hand the weighting function of a series connection may be calculated using formula (1.3.4). As a result we obtain [oe)
ifg(t, 0)g(0,7) do = 6(t — 7). —oo
(1.3.8)
48
Chapter 1 Dynamical
Systems and Their Characteristics
This relation represents the main property of the weighting functions of the reciprocal linear systems. Changing the order of the connection of the systems we have another relation co
/ g(t,0)9 (0,7) do = 6(t — 7).
(1.3.9)
—0o
Thus the wetghting functions of two reciprocaldinear systems satisfy two
integral relations (1.3.8) and (1.3.9) which in general do not coincide. The weighting function g(t, 7) of the system with a feedback mapped in Fig.1.3.7 may be defined as in the previous cases after observing the passage the unit impulse. But this way is rather difficult and leads to
an integral equation relative to the sought weighting function g(t, 7). It is rather easy to find the weighting function of an inverse system (Subsection 1.2.4).
Bigs 13.7
Fig. 1.3.8
> It is evident that the system which is inverse relative to the system with a feedback is the parallel connection of the system relative to the system in the direct loop and the system in the feedback loop (Fig.1.3.8).
Really, denoting the input
of the considered system by Z, the output by y and the output of the system in a feedback loop by Z we see that © — Z will be the input of the system in the direct loop and y will be its output.
While supplying the output
y on the input of the
parallel connection of the systems with the weighting functions g; and g2 we shall obtain on the output of the first system the signal 2 — Z, and on the output of the second system the signal z. Adding these signals will give on the output of a parallel connection the signal X (Fig.1.3.8).
Thus a series connection of the systems shown
in Fig.1.3.7 and 1.3.8 give an ideal servomechanism.
are reciprocal.
Consequently,
these systems
Applying formula (1.3.7) for the weighting function of the parallel
connection of the linear systems we find
g(t,7)= 9; (t,7) + ga(t,7).4
(1.3.10)
1.3. Connections of Systems
49
This formula determines the weighting function of the linear system which is inverse relative to the system with a feedback shown on Fig.1.3.7.
1.3.3.
Connections of Stationary Linear Systems
At first let us consider a series connection of two stationary linear
systems with transfer functions ®;(s) and ®2(s) (Fig.1.3.9).
Suppose
that on the input of this series connection a signal representing an exponential function of time e*’ acts indefinitely for a long time. This signal while passing through the first system is multiplied by the transfer func-
tion ®;(s). Thus on the output of the first system and on the input of the second system will be the function ®;(s)e**. By virtue of the linearity of the second system its output will be equal to the product of its response on the exponential signal e** by the constant multiplier ®2(s),
i.e.
©1(s)®2(s)e*'.
Thus a series connection of the stationary linear
systems gives the stationary linear system whose transfer function ®(s) as equal to the product of the transfer functions of connected systems
&(s) = 1 (s)2(s).
Sao
(1.3.11)
oo
(s) = ®, (5) ®, (S) Fig. 1.3.9 This formula is easily spread to a series connection of any number of the stationary linear systems. Namely, if the transfer functions of
n connected in a series systems are equal to ®(s),...,®n(s) then the transfer function of the connection ®(s) is determined by formula
@(s) = ©1(s)b2(s)---@,(s) = ]] O:(s).
(1.3.12)
p=
Thus the result of a series connection of the stationary linear systems does not depend on the order of their connection.
50
Chapter 1 Dynamical
Systems and Their Characteristics
As the modulus of the product of the complex numbers is equal to the product of the moduluses of the factors, and the argument is equal
to the sum of the arguments of the factors then from (1.3.12) follow the formulae for the gain- and phase-frequency characteristics of a series connection nm
|&(iw)| = [] ee (iw)], k=
(1.3.13)
\
Ig [B(iw)| = Solg|®x(iw)|, arg @(iw) = So arg, (iw). Formulae (1.3.14) show that at a sertes connection
(1.3.14)
of the statio-
nary linear systems their dectbel-log and log phase-frequency characteristics are summarized.
®, (s) es
e
®,(s) oa.
D(s)e
:
®, (S) e.
(s) = ®, (5) + ®,(S) Fig. 1.3.10 For determining the transfer function of the parallel connection of the stationary linear systems with the weighting functions ®,(s) and 2(s) (Fig.1.3.10) we suppose that on the input of this connection the signal representing the exponential time function e*’ acts indefinitely for a long time. As a result the outputs of the connected systems will be equal to ®;(s)e%’ and ®(s)e*' correspondingly. The output of the connection will be equal to the sum of these expressions. Consequently, the transfer function of a parallel connection of the stationary linear system is equal to the sum of the transfer functions of connected systems:
B(s) = O;(s) + 9(s).
(1.3.15)
1.3. Conriéctions of Systems
51
Evidently, this formula is also spread to any number of parallelly connected systems:
(s) = 5 ;(s).
(1.3.16)
Let us pass to the determination of the transfer functions of the stationary linear systems closed by a feedback. At first we consider the system consisting of the stationary linear system with the transfer function ®;(s) closed by the nonnegative feedback containing a statio-
nary linear system with the transfer function ®2(s) (Fig.1.3.11).
For
determining the transfer function (s) of this system let us consider an
inverse system which has the transfer function 1/®(s) (Subsection 1.3.2). This inverse system represents a parallel connection of the system with
the transfer function 1/;(s) and the system with the transfer function ®5(s). Therefore using formula (1.3.16) we get
(5) °@ (s)'® +1 (s)'®
= (s)®
Fig. 1.3.11
mane = ee + @2(s) = i+ ®1(s)P2(s)
®(s)
®i(s)
#1 (s)
From here we find the transfer function of the system with a feed-
back:
Aiea
RAS,
1+ ©1(s)®o(s)
(1.3.17)
At a rigid negative feedback it is necessary to put ®(s) = 1 in formula (3.17): Thus at any connections of the stationary linear systems we always obtain the stationary linear systems, whose transfer functions and frequency response characteristics are determined by means of elementary
52
Chapter
1 Dynamical Systems and Their Characteristics
algebraic operations on given transfer functions (on frequency response characteristics respectively) of connected systems. At all possible connections of the stationary linear systems described by ordinary differential equations as a result we always obtain stationary linear systems whose behaviour is also described by ordinary differential equations. Really, in this case the transfer functions of all connecting systems represent the rational functions of s. Therefore all derived for-
mulae for the transfer functions of the connections determine the transfer function of the connection ®(s) in the form of the rational function of s. After presenting this function in the form of the ratio of two polynomials and replacing s by the operator of the differentiation with respect
to time D we shall find the operators F(D) and H(D) and Eq. (1.2.60) for the input — output relation. 1.3.4. Structural Transformation Rules
It is expedient to transform the block diagram into more convenient form by means of the partition of the composite systems into more simple ones (decomposition), the connection of the simple systems into one (aggregation) and another transformations for the sake of simplicity of the structure on retention of the number and the order of included systems. The main principle of structural transformations consists in the fact that all the transformations of the block diagram of the linear sys-
tem (Subsection 1.3.1) should be performed in such a way that all the inputs and the outputs of each transforming section of the block diagram will remain invariant. With regard to Subsections 1.3.2 and 1.3.3 from this principle follows the rules of the structural transformations of linear systems given in Appendix 6.
Fig. 1.3.12 Example
1.3.1.
For transforming
the flexible feedback
by the rigid
one (Fig.1.3.12a) we interchange the branching point and A» system in a feedback (Fig.1.3.12b).
~
1.3. Connections of Systems
53
Fig. 1.3.13 Example
1.3.2.
Consider the translation of the system Ag along the
connected in a series stationary linear systems Aj, ...
, An (Fig.1.3.13a).
As the
stationary linear systems connected in a series may be arbitrary interchanged then the mutually inverse (reciprocal) systems introduced by the translation of the branching point and the summator via the one and the same system are mutually compensated. So for the translation of the parallel loop from Ay to Aj or vica versa it is necessary to consider the translation of the branching point via 4; and summator through Ap
(Fig.1.3.13b).
b)
Fig. 1.3.14 Example
1.3.3.
For the system (Fig.1.3.14a) analogously we conclude
that for the translation of Ag in a feedback from A, to Ap and vica versa along the connected in a series stationary linear systems Az, ...
, Ap it is necessary to
54
Chapter 1 Dynamical
consider only the summator
Systems and Their Characteristics
translation via A, and the branching point via An
(Fig.1.3.14b). Example
1.3.4. Block diagram of the linear system with two inputs and
one output (Fig.1.3.15a) may be transformed into the block diagram on (Fig.1.3.15b).
Fig. 1.3.15 Example
input and two outputs
1.3.5.
For the block diagram of the linear system with one
(Fig.1.3.16a) the transformed
block diagram is given on
Fig.1.3.16b.
b)
Fig. 1.3.16 Example
1.3.6.
Fig.1.3.17a presents the block diagram of the linear
system with a parallel loop and the additional branching point in one of the loops. Fig.1.3.17b and Fig.1.3.17c show the sequential stages of the summator translation. Example
1.3.7.
Block diagram of the composed linear system with a
feedback loop having the internal summator is shown on Fig.1.3.18a. The sequential
stages of the summator translation are given on Fig.1.3.18b and 1.3.18c.
1.4. Stochastic Differential Systems
Fig. 1.3.17
55
Fig. 1.3.18
1.4. Stochastic Differential Systems 1.4.1. General Form of Equations of Stochastic Differential Systems Stochastic models of systems involve the action of various random factors. While using models described by differential equations the inclusion of random factors lead to the equations which contain random functions, i.e. such functions whose values at given values of the argu-
ments are random variables (Chapter 2). Eqs. (1.1.4) for a stochastic system must be replaced in the general case by the equations
GAP
( ZEAyeyves GZ),
(1.4.1)
where F'(z,a,t) and G(z,t) are random functions of the p-dimensional vector z, n-dimensional vector x and time t (as a rule G is independent
56
Chapter 1 Dynamical
Systems and Their Characteristics
of x). In consequence of the randomness of the right-hand sides of Eqs. (1.4.1) and also perhaps of the initial value of the state vector 7) = Z(to) the state vector of the system Z and the output Y represent the random variables at any fixed moment t. This is the reason to denote them by capital letters as well as the random functions in the right-hand sides of
Eqs. (1.4.1). The state vector of the system Z(t) and its output Y(t) considered as functions of time ¢ represent random functions of time t (in the general case vector random functions)» In every specific trial
the random functions F'(z,«,t) and G(z,t) are realized in the form of some deterministic functions f(z,z,t)°and g(z,t) and these realizations determine the corresponding realizations z(t), y(t) of the state vector
Z(t) and the output Y(t) satisfying the differential equations (which are the realizations of Eqs. (1.4.1.): z = f(z,a,t), y = g(z,t). Thus we come to the necessity to study the differential equations with random functions in the right-hand sides. In practical problems the randomness of the right-hand sides of the differential equations arises usually from the fact that they represent known functions some of whose arguments are considered as random variables or as random functions of time t and perhaps of the state and the output of the system. But in the latter case these functions are usually replaced by the random functions of time which are only obtained by assuming that their arguments Z and Y are known functions of time corresponding to the nominal regime of system functioning. In practical problems such an assumption usually provides sufficient accuracy. Thus we may restrict ourselves to the case where all uncertain variables in the right-hand sides of differential equations may be considered
as random functions of time. Then Eqs. (1.4.1) may be written in the form
Z = f(Z,x,Ni(t),t), Y = g(Z, No(t),t),
(1.4.2)
where f and g are known functions whose arguments include random functions of time N;(t) and N2(t). The initial state vector of the system Zo in practical problems is always a random variable independent of the
random functions N;(t) and N2(t) (independent of random disturbances acting on the system). Every realization [m1 (t)? no(t)? ie of the random function [Nj (t)? N2(t)"]” determines the corresponding realizations f(z, «, ni(t),t), g(z, no(t), t) of the functions f(z, x, Ni(t),t), g(z, No(t),t), and in accordance with this Eqs. (1.4.2) determine respective realizations z(t) and y(t) of the state vector of the system Z(t) and its output Y(t).
~~
1.4. Stochastic Differential Systems
57
Example 1.4.1. The equations of motion in Examples 1.1.4 and 1.1.5 describe the stochastic models of motion of various mechanical systems if the forces and the moments
in the right-hand sides of these equations represent random functions
of time.
1.4.2. Stochastic Linear Differential Systems The differential equations of a linear stochastic system differ from Eqs. (1.2.1) of a linear deterministic system by additional random items: Z=aZ+ayxr+ag+
ay Ni(t),
Y =bZ+b).+
bi No(t) ,
(1.4.3)
In these equations N;(t) and No(t) are random functions of time (in general they are vector random functions). Introducing the composite vector
random function W(t) = [ Ni(t)? No(t)? |” and block-matrices a4 = [a20], bj = [06,] where 0 is the matrix whose elements are all equal to zero we represent the
random items of Eqs.(1.4.3) in the form a2Ni(t) = a,N(t), 6: No(t) = b},.N(t). Therefore without loss of generality we may omit the indices at random functions and write Eqs. (1.4.3) in the form Z=aZ+ajxr+anp+a2,N(t),
Y=bZ+b0+b:N(t).
(1.4.4)
In practical problems the deviations of a nonlinear system from the required regime may sometimes be assumed as sufficiently small. In such a case the equations of the system may often be linearized relative to random deviations from the required regime and relative to random disturbances acting on the system. In such cases the nonlinear equations describing the behaviour of the system are replaced by approximate linear equations in the deviations. It gives the opportunity to study the nominal regime of system functioning using the deterministic nonlinear
model (1.1.4) and then to study random deviations from the nominal regime using the much simpler linear stochastic models (1.4.4). 1.4.3. Linear Systems with Parametric Notses In practical problems we have to deal with linear systems in which the noises depend linearly on the state vector of the system. In such cases we have to use for the description of the behaviour of a system linear differential equations with fluctuating coefficients. The fluctuations of the coefficients of the differential equations of a linear system are usually
58
Chapter 1 Dynamical
Systems and Their Characteristics
called parametric noises. For such systems the matrices az and 6, in Eqs. (1.4.4) depend not only on time, but are also linear functions of the state
vector of a system. Thus Eqs. (1.4.4) in the case of parametric noises are replaced by the equations P Z=aZ + ajz + ao + (000+ Sanz] N(t),
ia Y =bZ+bo+ (i+ SZ) N(t).
(1.4.5)
P
(p=)
Example
1.4.2.
The motion of a pendulum with vibrating suspension
point in viscous medium is described by the equation
Ag+ Bg + mgl(1 + No)sing + mglN, cosy = 0,
(1)
where ¢ is the deviation angle of a pendulum from the vertical, A is the centroidal
moment
of inertia, B is the coefficient of moment
is the static moment, gN; vertical components
of forces of viscous friction, mgl
= gN, (t) and gN2 = gN2(t) are the horizontal and
of the acceleration vector of the suspension point representing
random functions of time.
At small angles Y putting in (I) sin p & », cos y & 1 and assuming as state variables
p =
Z}, yg =
Zy we obtain a linear stochastic differential system with
additive and parametric noises whose equation has the form
lela alla; ofl S]4) Lm(1) 2, 2
|..0,
ot
Z
2a
ia
0
0
N,
where we = mgl/A, 26 = B/A.
1.4.4. Stochastic Bilinear Differential Systems
The systems in which the vectors Z, Y, Ni, N2 in Eqs. allowing the decomposition of the form ae
|zz
vu
a a
lev tee
aT?
aN
(1.4.2)
[Ni2"0| ,
are called so, with the components Z’, Y' and Z”, Y” satisfying the linear equation of the form Eq. (1.4.3) and the bilinear equations, respectively:
~
1.5. Stochastic Discrete Systems
Z =aZ' +ayx+ao+a.N{, :
Z" =|
P
Act) AZ|
59
Y'=b2Z' + bo +b,N5, P
Z", Y" =|
Bot >> BZ, |2”.
jvesl
(1.4.6)
ivemt!
Here a, aj, a2, Ap, Ax, bo, b1, Bo, By are the matrices, ao, bo are the vectors of the suitable dimension. The bilinear stochastic systems described by Eqs. (1.4.6) are an important special case of the stochastic systems with random parameters,
Z = f(Z,2, Ni(t),U,tY, Y = 9(Z, No(t), U,2),
(1.4.7)
U being a vector of random parameters. Example
1.4.3. The equations of a body motion around a fixed axis in
viscous media have the form
d*o _
Ady
(1)
di? bral ate Here I is an axial moment
of inertia, ) is an angular velocity, A /T is specific coeffi-
cient of viscous friction which is a random time function satisfying to linear differential equation
A=-a\+hN, where @ and hare some positive coefficients,
A=A/T,
(1)
NV= N (t) is a random time function.
Denoting Z; = ~, Z2 = A we reduce the Eqs.
(I) and (II) to the following two-
dimensional bilinear stochastic differential equations:
EVAL
IN
(II)
1.5. Stochastic Discrete Systems 1.8.1. Discrete Systems
It is known that for a discrete insertion of the inputs into a system it is necessary to modulate some parameters of the input impulses in accordance with their values. A device forming the sequence of the impulses which depends on an input is called an impulse element. Usually the impulses form remains invariable at the modulation. We may distinguish three main types of the impulses modulation:
60
Chapter 1 Dynamical Systems and Their Characteristics
e pulse-amplitude modulation (PAM) at which the amplitude of the impulses a; depends linearly (with the gain pam) on the input value at the moment of the impulse action t;, a, = az (tx) = Epama(te); e pulse-width modulation (PWM) at which the duration of the impulse 7; linearly depends (with the gain pwm) on the input value at the moment of the impulse action, T;(t,) = €pwm2(tx);
e time-impulse modulation (TIM) at which the time shift 7 of an impulse linearly depends (with the gain tm) on the input value at a definite time moment, Ts = €yimz(tx). At PAM and PWM the modulating signal changes the area (i.e. the intensity) of the impulses, and at TIM the area of the impulse remains constant. The dependence of modulating parameter of the impulses produced by an impulse element on the correspondent discrete input values is called a characteristic of an impulse element. The latter may be linear or nonlinear. The impulse element with linear characteristic is linear, and with a nonlinear characteristic is nonlinear. The impulse elements vary also in the form and the character of the impulses modulation, in impulses frequency and their relative duration. Usually the impulse elements are periodically generate one impulse at each period. The period of the impulses succession TJ; is called a period of
repetition (a tact) of a discrete system. resents a repetition rate. called a relative duration
The variable w, = 22/T, rep-
The parameters y, y = T;/T, and 1 — y are and impulses ratio.
1.5.2. Characteristics of Discrete Linear Systems Let us denote g,(t) the response of a linear discrete system on a short-time input signal equal to 1 and acting only during the k'® impulse action. Then its response on a short-time signal equal to x(t;) and which acts only during the time of the k*® impulse action will be equal to g(t )a(t~) on the basis of the superposition principle. The response of a linear discrete system on the whole sequence of the impulses modulated by the input disturbance «(t;) on the basis of the superposition principle
(Subsection 1.1.2) will be determined by the formula co
y(t)= S > ge(t)e(te)-
(1.5.1)
k=-00
The functions g;,(t) characterize completely a linear discrete system as knowing these functions we may calculate the response of a discrete
1.5. Stochastic Discrete Systems
61
linear system on any input disturbance x(t). The functions g,(t) determine the part of the input values acting at different time moments t, in the formation of the output at any time moment t. That is why the functions g;(t) are called the wetghting coefficients of a linear discrete system.
Any nonanticipative (physically realizable) discrete system (1.5.1) may response at a given moment t only on the signal acting on it at the previous instant. Therefore for any nonanticipative discrete system
ge(t) =0 at t> ge(ti)a(te). k=—0o
(1.5.10)
64
Chapter 1 Dynamical
Systems and Their Characteristics
For brevity setting 2, = x(tz), yr = (ty), ge(t;) = gie we can rewrite formula (1.5.10) finally in the form
Teng
i OME
ERPS gS
(1.5.11)
k=-00
In the case when the instants of an output fixing coincide with the in-
stants of the impulses action, t; = t; (J = 0,+1,+2,...), formula (1.5.11) at t = t; will give the following expression for the values of an output of a nonanticipative linear discrete system: if
yt =p einGinte fle
a 2 ath he
(1.5.12)
Roa)
Formula (1.5.12) and the analogous formulae for the discrete systems with many inputs and outputs represent, in particular, a digital computer as a discrete linear ystem because the results of the calculations depend linearly on the initial data. In this case the initial data introduced at each step of the calculations are the input signals and the results of the calculations are the outputs signals. The weighting coefficients gj), determine the program of the calculations. At the sufficiently large impulses repetition rate any discrete system may be considered approximately as a continuous one. Really, the response of any system on a short-term unit signal acting during one period of the impulses repetition TJ, at infinitesimal 7}. is an infinitesimal of the order T;,. Therefore the function g,(t) at small T, = At may be expressed by formula
ge(t) © g(t, te)T> = g(t, ty) At.
(1.5.13)
Here g(t,t,) is some function independent of T,. So, for instance, in the case of a series connection of an impulse element and a continuous linear system with the weighting function g;(t,7) we can express the dependence of the impulses of a given form on the period of the repetition
of the impulses assuming the function n(t) in the form n(t) = A(t/T;), where h(€) is a function different from zero only at the interval 0 < €
[v1 yo)", =
[0, Q |
is the input,
0) B, C are constant coefficients deter-
mining inertial, dissipative and position forces respectively. Show that the matrix of
fundamental solutions u(t, wf) and the matrix of transfer functions ®(s) have the form
u(t,7) = e~*-7) cos W_(t — T) + ns sinw,(t — T) x
oo sinw,(t — 7)
2
4
se sinw,(t — T)
&(s) =
cosw,(t — T) — ee sinw,(t — T)
(II)
1 ee
As’ + Bs+C | (9¢ = B/A,w? = w? — €?,w? = C/A).
As°+Bs+C
(111) 1.6. Show that for a linear nonstationary system
i =y2,
Yo=—cet
"yi+2,
t>0,
the elements of the matrix of fundamental solutions u(t, T) ab 2S
(I) Ky |1 = Ac|,
Cal /4 are determined by the formulae:
uii(t,7) = z ds
ja+ 27) (=) — (1-27) (=) |;
watt = Spe [Clon | rt)i/2
Wai(b; Fl ale(< 2 () |,
usatts)= @ (5) a
)-a-2(9)'].
a
~
Problems
91
Find u(t, 7) at c > 1/4. 1.7. Prove that for a nonstationary linear system
Y1 =
UPS
Y=
—t-2y,
—ttyta,
t>
0,
(I)
the matrix of fundamental solutions u(t, T) is equal to
alter)
cos In=t
—711sin In =
: 7 sin Int=
(I)
¢os In at
-z
1.8. Prove that the weighting function g(t, T) of a nonstationary linear system
ytt'yt+(l—n7t-*)y=a2,
y=z,
(1)
is determined by:
g(t,7) = 5[In(7)Na(t) — Nn(t)In(t)]7
(11)
where Jn, (7) and N, (t) are the Bessel functions of the first and the second kind respectively. 1.9. For a stationary linear system ¢
Peaseor, y= wile = 242983
a
hl, B= |2296s
fu
|G
(I)
where a is a constant square matrix, b a scalar constant, y the output, show that the elements of the matrix of transfer fucntions (s) =
=
—(a => sI))~'b at A(s)
la = sl| are determined by the formulae
®11(s) = —b |s? — s(azq + a33) + 422433 — d23432| /A(s) ,
®12(s) = b(—say2 + a12433 — a32013)/A(s) , $13(s) = —b(sa13 + a12423 — 413422)/A(s), o1(s) = 6(—sag1 + 421433 — 431423)/A(s) , ®o0(s) = —b[ s* — s(a11 + a33) + 411433 — 413431|/A(s),
o3(s) = b(—saz3 + 411423 — 421413)/A(s), $31(s) = —b(sag1 + a21432 — 431422)/A(s) , 30(s) = b(—sag2 + 411432 — 412431)/A(s) ,
@33(s) = —b [s? — s(ay1 + 22) + a11422 — ai2a21 |/A(s).
(II)
92
Chapter 1 Dynamical Systems and Their Characteristics
1.10. Show that for the stationary system of Problem 1.9 at 4d}; = @31 = 413
="a39 = 0) 419 = Go3 = lye
—w?, a92 = —2€, 433 = —Q
the elements of
the matrix of transfer functions are determined by
$11(s) =—b[s* + (26 + a)s + 2ae] /A(s),
12(s) = —b(s + a)/A(s), P13(s) = —b/A(s), o1(s) = bwo(s + a)/A(s), O22(s) = —bs(s +a)/A(s), Do3(s )= —bs/A(s), 3) = O35 => 0, $33(s) ) = —b(s? + 2es +.w2)/A(s), A(s) = —(s +a) [we +
(2e + s)s] ,
(1)
Check that the elements of the matrix of fundamental solutions u(t, tp) are given by uia(t, T) =
(- 1/w? ) [uri(t, T) + a/wouar(t, T) a
31 = U32 = 0, u3a(t, T) =
uza(t, ®) iy
exp {—a(t a T)} )
ug3(t, 7) = (—1/w?) {uai(t,7)(1 —
ea /w?) — w[usi(t, 7) — usa(t, T)]} ;
(II) the functions U;1, U12, U21, U22 being determined in Problem 1.5. 1.11. Show that for the stationary linear mechanical system with n degrees of
freedom the differential equations of motion have the following form:
(a) In Lagrangian variables
Ag + (B+ B’)g-+ (C+ C)g=Q;, where
¢ =
og On
[q edn
e is the vector of generalized coordinates,
is the vector of generalized forces. In this case the vector
the output, and the vector
(I) and Q =
[Qi
y = q represents
£ = Q represents the input, A is the symmetric matrix of
the inertia coefficients, B is the symmetric matrix of the dissipative or acceleration forces, B’ is the antisymmetric matrix of gyro forces, C’ is the symmetric matrix of the conservative position forces, and C’ ' is the antisymmetric matrix of nonconservative position forces. (b) In canonical variables
sla feet beh eT ledealaane®
where g and p =
Ag are the vector of generalized coordinates and the vector of
generalized impulses respectively.
In this case [q? pt]? represents the state vector
of the system, the input and output being the same as in the case of the Lagrangian variables.
Problems
93
1.12. Show that for the stable stationary system of Problem 1.11 at n = 2
a Ai2
Ai2 es i A22 | | Qe a
Bi, Biz
Bi Boe
4
0 —Byan»
Ci1 Cie | OC 12 (|Cho, C22 dare i2 O
Bip 0
q1 q2
=
qi q2 Q1 Qe |
(1)
the elements of the matrix of transfer functions ®(s) are determined by the formulae
11(s) = {Aj ({C| + C42) — s[Cii(C22 +5)
— (Biz + By2)(C12 — Ch) /A(s), 12(s) = -{Aj,(|C] + C'ts) — s[C11(Biz — Bip) — (Bi + s)(Ci2 — Cia) }/A(s),
o1(s) =—{-Aj,(|C] + C42) + s[(Ci2 + C{.)(Baa + 8)
(II)
— C2o(b12 + Biy)|}/A(s), o0(s) ={AZ(|C] + Cq2) + s[(C12 — Clp)( Biz — Bip) — C2(Bi1 + s)]}/A(s) where A(s) = s*4 +1159 + 1728? + 173s +14,
ry = Byy Az, + BorAj, — 2B12 Aj, ,
ro = [BiAg — (Bia + Biz) Ate] |Bo2Ary — (B12 — Bip) AyD] + (C1149 + C22Ay, — 2C12Ajp) , r3 = [By Aj — (Bio + Biz) Aja] [(Bi2 — Biz) Ad, — B22 AQ | a [ BooAy
(Biz wy By») At | [Cir Ax x (Ci2 im Cia) Ain
“+ [(Biz - Bi) Are — B22 Ajp| [CiAi
a (Cie te Cya)An
+ {[ Bi Az, + (Bio + Bia) Ary] + [(Bia + Bi) AT — Br Ax] } x [C2245 — (C12 — Ci2) AQ] , ra = (IC] + Ale Aj; =
Agen
|A|
(4,9 =
|We 3
|A|
=
Aj1A22—A?,
IC | =
(111)
CiiCo9—C-,
1.13. Show that if the matrices A~! B and A7~ 1C in Problem 1.11 are commute and B’ = 0, C’ = 0, then there exists always a linear transformation of the output
y = L¢ satisfying the conditions
TWA fe dele Blas diag (266s 5... 2En),
TE OP
diag {uen rasu2 Js,
I
94
Chapter 1 Dynamical
Systems and Their Characteristics
reducing the set of equations of motion to n independent linear differential secondorder equations
Cy + 2enCn
where
Up
=
up
,
(h7= 1)
ane
un(t) is the A*® element of the column-matrix
corresponding equations at 1.14.
+2 Cn =
B = Meys\. ei
Consider a mechanical
(II)
EO}
Write the
nC :
system with n degrees of freedom
described by
Lagrange equations
d
(OL
OL
a (i) — 3 = 2 where q =
-
[q1 soli ‘ieis the vector of generalized coordinates,
L = T'— II is the
Lagrange function, 7’ = T(q, q, t) is the kinetic energy, I] = II(q) is the potential
energy; Q =
[Qi MGS
servative forces.
ie Q.=
Show that Eq.
Q(4; q; t) is the vector of generalized noncon-
(I) may be transformed into the set of Hamilton
equations in canonical variables gq and
G==-, Here
p = OT /Oq
p=->-+Q.
H = H(q, p, t) being the Hamilton function,
(II)
H = pq -T.
1.15. The vertical oscillations of the body of a car while moving on the rough road are described by the equations
My + 20(y — ¢) + 2v(y- 9) =0, where M/ is the car mass,
w(y = q) and ply = q) are the nonlinear functions
determining damping and restoring forces respectively,
g =
q(c) is the function
characterizing the microprofile of the road in the vertical plane,
7 =
vt, VU being
the car velocity. We assume that both pairs of wheels of the car undergo the same vertical displacement at every instant due to the roughness of the road (i.e. the road
profile in the vertical plane has a very long correlation interval).
Find the transfer
function of the car at equalities w(y = q) = b(y = q) and ply = q) = c(y = q); b and c being constant coefficients, assuming q as an input and Y as an output. 1.16.The equations of motion of the body with one degree of freedom equipped
with the nonlinear dynamic damper have the form
pa + ba + cx + pa(@ — Za) + vale — vq) = F(t),
Hata — pa(@ — £a) — Ya(x@ — 2a) = 0,
“wn
Problems
95
where & and £q are the displacements of the body and of the damper, pi and [lq their
and C the coefficients of viscous friction and of restoring forces respectively,
masses,
wa(x = La) and a(x = La) the functions characterizing the forces of viscous friction and the restoring forces, and F(t) the disturbing force. —%q=
a(x
ba(« _
La) and pa(r = La) =
In the special case where
cae. — La), with bg, cq being
constant coefficients find the transfer function of the system considering the force F(t) as an input, and the displacements © and &q of the body and damper as the components of the output.
1.17. Show that the nonlinear equation n—-1
Oe
Nea
See yy
fee, et)
k=n-—m
m
=D da(ysy’, 0. yO", 1), (m 0 that P(Q) = 1.
If a nonnegative finite or o-finite measure v is determined in the measurable space (Q,S) then the probability P = P(A) may be presented in the form of the sum of the v-continuous and v-singular measures, and the v-continuous part represents an integral over the measure v with
respect to some function f(w)
P(A) = | tev@s) + P(A).
(2.1.1)
A When the probability P(A) is v-continuous, P;(A) = 0 and
P(A) = :;Jobin
(2.1.2)
A
As it is known from the functional analysis in this case the function f(w) represents Radon-Nikodym derivative of the probability P with respect to the measure vy and is called a probability density with respect
to measure v in the space (2S, P). Any subset N of the set No € S for which P(No) = 0 is called a zero
set.
If zero set N belongs to the o-algebra S of the probability space
(Q,S, P) then, evidently P(N) = 0. But zero set does not need belonging to S (if No C S then the subsets of the set No do not necessarily belong
to S). On such zero sets the probability P(N) is not determined. The probability space (Q,S, P) is called a complete one if the o-algebra S contains all zero sets of this space. If the probability space (Q,S, P) is not complete then it may be always supplemented after adding in S all zero sets and all the unions of zero sets with the sets from S and determining the probability P on the supplemented sets.
102
Chapter 2 Random
2.2. Random
Variables
Variables, Processes and Functions
2.2.1. Definitions
In elementary probability theory a random variable is a variable which assumes as a result of a trial one of possible values and it is im-
possible (before a trial) to know what value it will assume namely. Here it is supposed that in the space of possible values of each random variable there exists a probability distribution. In order to come to an exact definition of a random variable notion we shall refer to the examples from elementary probability theory and random functions theory. Example
2.2.1. The number of the occurrences of the event A (in one
trial) in Examples 2.1.1 and 2.1.8 represents a random variable which assumes the value 1 when A appears and the value () when A does not appear. It is easy to notice that this random variable represents a function of a point in the probability space
(Q, S:, iP) of Example 2.1.8 equal to 1 at w = W, and 0 at WwW = Wo. Example
2.2.2.
nm triales in Bernoulli 0,1,...,m.
scheme
This random
The
number
represents
of the occurrences
a random
variable
of the event
A at
with possible values
variable is a point function in the probability
space
(Q, Ss) P) of Example 2.1.9 equal to.m at all points W to which correspond the se-
quences
A,,...,
=O
lees8):
Example sidered
An
containing
letters
A
and
n
—
m
letters
A
(m
2.2.3. The finite-dimensional random vector X may be con-
as the point function
Example 2.1.10.
‘m
r(w)
=
wW in the probability
space
(Q, On P) of
Each component X, of the random vector X represents also the
point function r,(w) in the probability space (Q, S, Vane Example
2.2.4.
Infinite-dimensional random
vector X
(the sequence
{xX x} of the scalar random variables) may be also considered as the point function z(w) = w in the space QQ of all numerical sequences
{xz}. It is evident that any
component of an infinite-dimensional random vector is the function of a point in the space 2 (Lm at any fixed M
is the function of the sequence
{7x }): The same is
true, in particular, for the sequence of the events in Example 2.1.6.
The considered examples turn us to the following definition of a random variable. A random variable with values in a measurable space
(X,A) is called (S,A)-measurable function of a point in the probability space (Q,S, P) mapping Q into X and determined everywhere in Q besides may be the set of zero probability P. The value of a random variable at any given point w of the space (2 (i.e. the value which it assumes when as a result of a trial the elementary event w appears) is called a realization of this random variable.
2.2. Random Variables
103
It is evident that any random variable in (X, A) is at the same time a random variable in (X,.A’) where A’ is any o-algebra in X containing APA’ eA: We shall often denote a random variable by the same letter as the space of its values. While considering the set of the random variables with the values in one and the same space the denotations of these variables will be supplemented with the indexes. So, for instance, X, Xq represent the random variables with the values in the space X. If X, Xq are the random variables in (X,.A) then according to the definition X
= z(w), Xq = Lq(w) where x(w), w(w) are (S,A)-measurable functions mapping 2 into X. a 2.2.2. Probability Measure
By virtue of (A, S)-measurability of a random variable
X = 2(w)
in (X, A) the probability of the fact that it will take the value from the set A is determined by the following formula:
P(X 6A),= Plar(A))sucAce A, where as usual z~1(A) is an inverse set A in the space 2. Formula
f,(A) = P(X € A) = P(27"(A)),
AEA,
(2.2.1)
determines the probability uz, induced by the random variable X = z(w) in the space of its values. The triple (X, A, zz) represent the probability
space induced into (X, A) by the random variable X = a(w). The space X of the values of the random variable X = a(w) is called its phase space or the space of the realizations. The probability pz in the phase space is called the probability measure or the distrubution of the random variable X. The probability space (X,A, uz) is called the phase probability space of the random variable X = x(w). It is evident that any probability space may be considered as a phase space of some
random variable. If the random variable
X = «(w) is considered in its
phase probability set then the argument w in its denotion naturally is not indicated. In such cases we shall denote it simply by X and its realizations by z. If the nonnegative finite or o-finite measure v is determined in the
space (X,.A) then the probability measure pz of the random variable X may be presented by formula
entAye= /f(a)v(dz)+ps(A), A
AEA,
(2.2.2)
104
Chapter
2 Random
Variables, Processes and Functions
where (A) is the v-singular measure. In the special case when the measure jl, is v-continuous, 5(A) = 0, we have
Ay= [semwees
“Aten
(2.2.3)
Here f(x) is Radon—Nikodym derivative of the probability measure pz with respect to the measure vy which is called a probability density of the random variable X with respect. to the measure v. Remark.
Inelementary probability theory the probability Aensity of
variable is always considered in a finite-dimensional phase space (Example
arandom 2.1.10).
In problems which are connected with one random variable it is always convenient to assume its phase probability space as the probability space.
But in the problems
connected with the set of the random variables (possibly with different phase spaces) it is expedient to study all of them in one probability space. Example
2.2.5.
The two-points set {0, Hi serves as the phase space
of the random variable in Example 2.2.1, and the probability measure is completely
determined by its value in one of these two points: a0
Hae({1}) = 0), Hr ({0}) =),
lle
Example
2.2.6.
The sett {0,are
as n} serves as the phase space of
the number of the occurrences of an event in 7 trials in Example 2.2.2, and the probability measure [/, is completely determined by its values in these points: Ha ({m})
SOR
pt ote
T(t = nl) ml i — a)
Example
2.2.7.
In Examples
sO:
cehatt
2.2.3 and 2.2.4 the space of elementary
events () itself serves as the phase space of the complex (correspondingly real) random
vector X, and the complex plane (correspondingly real axis) serves as the common phase space of its components.
2.2.3. Equivalence of Random
Variables
Let X = z(w) be a random variable with its phase space X. The inverse mapping z~'(A) of the o-algebra A represents the o-algebra which is a part of the o-algebra S in the space of elementary events
, «1(A) Cc S. In the general case 2~!(A) does not coincide with S. Thus the variable X = x(w) induces in the space 2 some o-algebra
S; = a '(A), as a rule more poor than the o-algebra S. Therefore the random variable mined
on the same
Y = y(w) in the measurable space (Y, B) deterprobability space
(Q,S, P)will be not obligatory
(Sz, B)-measurable (from the fact that y~!(B) € S does not follow that y~'(B) € Sz). So if the random variable Y in (Y, B) whose phase space
a“
2.2. Random
Variables
105
Y being a separable B-space is (S,,B)-measurable then it is (A, B)measurable function of the random variable X. The random variables which are equal almost surely, i.e. with the probability 1, are called equivalent. Thus two statements are valid.
Theorem 2.2.1. If the random variable Y = y(w) in (Y,B) whose phase space Y being a separable B-space is (S,,B)-measurable then it ts equivalent to some (A, B)-measurable function of the random variable
aX eel Ga) Theorem 2.2.2. Jf (A, B)-measurable function valent to such a function) there are none restrictions
the random variable Y = y(w) in (Y,B) ts of the random variable X = x(w) (or ts equithen it is (S,,B)-measurable. In this case on the phase space of the random variable Y.
It is evident that all equivalent random variables have one and the same probability measure.
If the random variable Y = y(w) in (Y,B) is (A, B)-measurable function of the random variable X = x(w), Y = y(X) then it is arandom variable in the probability space (X,A,p,).
Therefore the probability
measure fly of this random variable may be expressed in terms of the probability measure pz of the random variable X by formula (2.2.1)
Hy(B) =ps(p*(B)),
BeB.
(2.2.4)
The same formula determines a probability measure of any random va-
riable Y equivalent to (A, B)-measurable function p(X) of the random variable X. 2.2.4. Two Types of Convergence of Random Sequences The sequence of the random variables {X;}, X; = %,(w) with the values in the topological space is called a.s. convergent (almost surely, with probability 1) to the random variable X = z(w) if the sequence of
the functions {xz,(w)} converges to «(w) almost everywhere on {2 relatively to the measure P. The sequence of the random variables {X;}, X, = xx(w) with the values in the separable B-space X is called convergent in probability to the random variable X = x(w) if at any € > 0
lim P(\| Xe — X [I> €) = Jim P({w : |]24(w) — (w) ||> €}) = 0. The following statements are valid.
106
Chapter 2 Random
Variables, Processes and Functions
Theorem 2.2.3. For the a.s. (respectively in probability) convergence of the sequence of the random variables {X;,} with the values in the separable B-space X it is necessary and sufficient that this sequence will be fundamental a.s. (respectively in probability). Theorem 2.2.4. Any a.s. convergent sequence of the random variables {X;,} with the values in the separable B-space ts also convergent in probability. The inverse is in general case not true. But from any sequence of random variables convergent in probability we may separate such sequence which a.s. converges. Example
2.2.8.
Let Xn be the random variable with two possible va-
lues 0 and n whose probabilities are equal to 1 — n-? and n~? respectively.
The
sequence of the random variables {Xn} corresponding ton = 1,2, ... , converges in probability to 0 as at any € > 0
P(|Xnl
1
26) = P(Xn =n) ==Z70
at
noo.
n
This sequence also a.s. converges as at any €,0 < € €
ifevenatone
n>p)
p)
PUXal
Let (Xe A) be an arbitrary measurable space, r(w) bea
(ST, A) measu-
rable function in infinite product of the spaces Q? with the values in X. Suppose that
there exists a countable class of the sets { Fi,} generating the o-algebra A. io
PR
Then
will be a countable class of the sets in Q7 generating the o-algebra
te (A) & S™ induced in QT by the function z(w). There exists such sequence
of the countable subsets {Ly} of the set 7’, that zo We put L = ULx. Then we shall have a—1(F,)
such countable set
L C T' that the o-algebra S
(k == pe Pe a As Zor foe (A) e So
(F;.) e SU
(k = lade. ne
E S” at all k. Thus there exists
contains all the sets a= 1( Pi)
A) is the minimal 0-algebra containing all ae
Oy) then
Consequently, the function r(w) is (S™ A)-measurable. Thus any
measurable function z(w) in Q7 with the values in the measurable space (X, A) in which the o-algebra A is generated by the countable class of the sets is (S¥ . A). measurable at some countable
D C ‘Ts i.e. depends on no more than a countable set
2.5. Probabilities in Infinite Product of Spaces
TE
of the coordinates of the point W in OF eat particular, any measurable functional in ee depends on no more than a countable set of the coordinates of the point W in
Ors Now
we pass to the construction of the probability in OD Suppose that at (Q¢, Sz) the probability P,( Az) and the regular conditional probabilities Paster els (At lw, fonds wi, ) correspondent to all hy, meal ad. (n each space
Sly dives .) are given. ‘Then applying formula (2.4.18) sequentially we find the probabilities in all possible finite products of the spaces {);:
Patt, Peso
As) = ijPtg,t, (At, lw) Pr, (der) ,
th lad XtvenXaAen)
=
Nae
At, XX At, x Ps...
oe
ane
(2.5.1) eer
Ee
he
ER
| Wr—1)
1
tn-1
(dw, ».G VOR
iegee ent
DS dwn-1)
a Space fice oeeng
After determining the probability 1 ,tz on the measurable rectangles Os, x oF by
the first formula (2.5.1) at first we shall extend it on the o-algebra St, Xx Sia Then using the second formula on the measurable
(2.5.1) at
m =
3 we detemine the probability Psy ,ta,t3
rectangles Qs, x Q:, x Qt, and extend it on the 0-algebra
Si, x St, x S; ,- Continuing this process we determine consequently the probabilities Pay peeytn atn=
4,5,....
It is easy to understand that the conditional prob-
abilities F eeks ae, td (Azur, tnd ies) (t, Ls
Gh
ote
i Seatise :) should be
consistent in such a way that they will give one and the same probability tae eeeat ss in the product of the spaces OQ, DK ores XK ey
(n Be 2, ose: Ps in any ways.
Then
the obtained set of the probabilities in all finite products of the spaces (Q4, S:) will satisfy the conditions of the consistence and the symmetry.
Let us assign now the function of the set P on the set of all measurable rectangles of the infinite product of the spaces Qr by formula Pe
where
Ae
x
Pax
Ay),
(2.5.2)
Ft is the rectangle with the sides At, seuss. 5 At, which le in the spaces
MtHe cme
Wem correspondingly.
set of the rectangles may
This function is additive as the bases of any finite
be considered
as belonging to one and the same
product of the spaces, for instance Sire X:++:& is additive in (Q4, XX
finite
G2 me and the probability jb sete
Ue ; St, X-22X St, MesFurthermore if we restrict ourselves
128
Chapter 2 Random
Variables, Processes and Functions
to measurable rectangles whose all bases belong to one and the same finite product of the spaces Q:, KISLEX
Q:, then the function P(R) is O-additive by virtue of
the @-additivity of the probability lady ,.:,tn+
Consequently, the function P(R) is
the function on the set of all measurable rectangles of the space (a>, Sf) which have the basis in one and the same finite product of the spaces Q:, Xoo
X GPAs
According to the theorem from functional analysis about the measure extension it may
be uniquely extended over the d-algebra St, x
cylinders of the space Q? with the bases in OF > ea
aX
St, of all measurable
Chey.4 A
It is evident that the
extension is determined by formula
PCR
re
Sa
eee
Bye)
(2.5.3)
where Cr, ,.1, tn is the cylinder in Q7 with the measurable base Bi, Oo. eke (Q4,
KPO
Seen SE):
Let us consider now the problem of the possibility of the function P (R) exten-
sion determined on the class of the measurable rectangles of the space (QT ; Sy by formula (2.5.2) on the whole o-algebra SE. The class C of all measurable rectangles is a semi-algebra and satisfies the conditions of the known from functional analysis theorem about the measure extension. In order that it may be possible to use the theorem about the measure extension for determining the probability on the whole o-algebra ST it remains only to prove
the -additivity of the function of the set P on the class of all measurable rectangles
which is determined by formula (2.5.2).
For this purpose it is sufficient to show
that P is continuous in zero, i.e. that for any decreasing sequence of the rectangles eae gsRp a) Ry at p < q with the empty intersection lim PCr
= 0. Hence it
is sufficient to consider the case when the set of the spaces {Qa} in which lie the sides of the rectangles Ry, is infinite (the case of a finite set of the spaces {Qz, } we have already considered).
We denote by AM*) the side of the rectangle Ry, in the
space Oe and put BS) = AbE+) KATE
Alt) where 7, is the number of the
sides of the rectangle R,,. Applying formula (2.4.18) we may present formula (2.5.2) at
R=
R, in the form
P(Rn) = / P(BO\or)Pa (der) = f1400 es)P(Blen) Pa, (der), AY) where
1
P (Bo Nw)
is
Saye o Se Dy au St, y I
0 that Pia)
> € atalln.
As
Pe thts) = | then it follows that there exists such point @] in Q;, that f(@1) > €. It is evident that ©, E€ AW at all n as f(#1) = 0 if wy does not belong to AW
beginning with some
{P(BS
mn.
But the sequence of the conditional probabilities
jz) is nonincreasing and lim P(BY
—
a, =
f(@1).
Therefore from
ds) ==
f(@1) > € follows P(Bs No) > € at all n. Further on the basis of (2.4.20) we have
P(BY ay) = i P(BY) |, w2) Pes ,19(dw2|@1) ,
(2.5.4)
A?) where
She
P(BY
ay, we) is a conditional probability in (Q45 X-++X
St, ). From P(BY?
Here Wy E Awe
St, :
a) > € follows the existence of such point W2 in Qy,
that P(By? lo, We) > €atalln andW2 we shall get the sequence
Qe,
€ Al?) at all n. Continuing this process
of the points {@;, } in the space Q4, (k aan
Np
he 4 3)
Consequently, all the points W of the infinite product of the space
QT? which have the coordinates Carn
ays Wp in the spaces ay Pee
to all rectangles Rt, whose sides lie in the spaces Oe Rota
Qe, belong
Q4,. Hence it follows
that the points of the space Q? which have the coordinates W. k in all the spaces
8
(k ct
.) belong to all the rectangles R,,, i.e. the intersection of all the
rectangles R,, is not empty. So at lim PCRs) of 0 the intersection of all the rectangles R,, cannot be emply. Consequently,
lim P (Rn)
=
0 if NR,
=
@
what proves the continuity of the
function P in zero and at the same time its 0-additivity on the set of all measurable rectangles of the space Q?.
Thus formula (2.5.2) determines the probability in the
space QT on the set of all measurable rectangles C. Hence there exists an unique extension of the probability P determined by formula (2.5.2) on the 0-algebra st
generated by the set C of all measurable rectangles. It follows from formula (2.5.2) that the probability es ,.:,tn projection of the probability P on the subspace (Q:, X-°:X
represents the
eae Phas ae St, )
of the space (OF. Siu. Hence it follows that the probability Py, in the subspace (Stee ars) of the space vine So}, LC of the probability P on (QO% Soh
T determined in such a way is the projection
130
Chapter
2 Random
Variables, Processes and Functions
The requirement of the regularity of all conditional probabilities in each space
is essential as formula (2.5.4) and all analogous formulae which are necessary for unbounded extension of the process in given proof of the continuity of P in zero have the sense only if all conditional probabilities (Qe, S:) are regular. Now let us consider the random vector X = x(w) with the countable set of the components X;
=
Lp (w) (k i
Me
a oD Suppose that on the 0-algebra of
Borel sets A of the phase space of this random variable
X its conditional distribution
pF(Alw)
relative to some O-algebra F C S of the probability space (Q, Se P) is
assigned.
The space (X ; A) may be considered as the countable product of one-
dimensional spaces Ce 5 Be):
x=R?=][
ke,
A=B”
where
Ry
= R
axis (k sl
is the real axis; By
=J] Br, k=1
pel
=
Wap ae Ns In this case we may
B is the o-algebra of Borel sets of this ensure that any countable-dimensional
random variables have conditional distributions. In other words there exist the regular conditional probabilities determined on the correspondent ¢-algebras of Borel sets in any countable-dimensional probability spaces.
Finally let us consider the random surable space
variable X
(Xx; A) and its conditional distribution
=
z(w) in arbitrary meaF(Alw)
relative to some
g-algebra F C S of the sets of the probability space (Q, S, Py; Suppose that there
2.7. Probability Measure of Random
Function
133
exists such finite- or countable-dimensional random variable (Q, A, ft) in the probability space and that the o-algebra A,, = ua? (BN) induced by it coincides with
A. According to (2.2.4) formula
v¢(Blw) determines
=pr(u-'(B)lw),
Be B",
the conditional distribution of the variable
UV =
u(w).
(2.6.4) In conformity
with the proved theorems there exists the regular conditional distribution v'-( Blw) of the variable UV. Naturally the question arises: whether is it possible to determine
the regular conditional distribution Hr (Alw) of the variable
X by formula (2.6.4).
As by supposition any set A € A is thé inverse of some set B € i
then formula
p'p(Alw) =ve(Blw), A=u-1(B),
determines p'-(Alw) for all sets
A € S .
determines Ue (Alw) uniquely not for all
u—!(B)
(2.6.5)
But in the general case this formula
A € S. For the existence of the regular
conditional distributions of the random variable X it is sufficient that such a finite-
or countable-dimensional function u(x) in (X Ai He) be measurable relative to A and the corresponding ¢-algebra of Borel sets B” will exist such that its range A, will be Borel set and the o-algebra A,
=
uw? (BE) induced by it will coincide
with A. As any probability space may be considered as the phase probability space of
some random yariable then for the existence of the regular conditional probabilities
in the probability space (Q, S, i it is sufficient the existence of finite- or countabledimensional random variable having Borel range and inducing the o-algebra S.
p-1 dkyk, Where the coefficients equations
d,...,d,
Ms Kepnasedp ah (eygt)
satisfy the linear
(ease... 1),
k=1
It is easy to check that formulae (2.7.1) give the same normal multidimensional distributions of the random process X;. Consequently, the random process X; is normally distributed and its probability measure is determined by formula (2.7.12). Remark.
These results are easily generalized on the random
processes
with any consistent set of the multi-dimensional probability densities f, (a1 sitar
t;,...,tn)
(or
with
any
consistent
set
of
the
conditional
frl@i thoi, so, GagSiss+ 4, Sn). end my thas care ai(ayes
=
fr(eisti) fr @o5t2|21;th) oA Gran (nite les, cap tn_15
gs
distributions
re thao
158.
a2 1))-
The probability measure of a random process of the set of all measurable cylinders in the functional space (Xxf : Al) is determined in the general case by formula
(2.7.16) By the formula of such a type we may also approximate with any degree of the accu-
racy the value of the probability measure of the random process on any measurable
set of the space (Xe ; ALY.
2.7.7. Streams of Events Let us consider some notions of the theory of the stochastic streams of the events which are important for the stochastic systems of queueing.
A stream
of the homogeneous events may be characterized by the distributions
of the random instants of the events occurrences.
Let us assume JQ as the initial in-
stant and suppose that the probability measure 1’; (A) of the moment 7} of the occurrence of the first event and the conditional probability measures l,, (Alt, sess ini)
of the instants of the occurrences T;, (n = 2, 3, . . .) of all following events are known. All these conditional distributions are regular as they are determined on the d-algebra B of Borel sets of the real axis R (Section 2.4). Considering T;, as a random function
of an integer argument NM we may determine its multi-dimensional distributions by formula
~
2.7. Probability Measure of Random
Pe ates,
ON
“ Un
d(T,
Function
2111, --
143
5 ine)
An-2
x ‘h Bie
fiat
net nies I) [Ti so*
the o-algebra B™
D4
Bx.) we may extend uniquely this measure on
of the space R™.
Let us consider a special case of a homogeneous
stream in conformity to the
recurrent stream of the events. Recall that a stream of the events is called recurrent if all the intervals between adjacent events are the independent random variables. Below (A = T) means that as usual a set of the numbers obtained by the subtraction of T from all the numbers
contained in A.
Then for such a stream of the events
with one and the same distribution lV of all the intervals between adjacent events we
have v1(A) = v(A), Vn (Alty, i. ti 2 40) = Uy(A — boa) (n eee
.) and
formula (2.7.17) for 1/1...» will take the form
Vi nar = [van Ai
CAR|
/ v(dt2)... A2-T1
/
Vi An — TmO1)v(dtme1)-
An-1-Tn-2
(2.7.20)
If there exists the probability density f(t) of the time interval between adjacent events
the
(A= ffet, A
then formula (2.7.20) may be rewritten in the form
".n(Arx:-*xXAn)=
eer
[-f(n)dn | f(t2—m1)dr::.
eee
| f(tm—tm-1)dt.
Cmpay
144
Chapter
2 Random
Variables, Processes and Functions
It is convenient to connect the stream of the events with the random function
X (t) which represents the number of the events occuring at the time interval [to stl. Evidently, all the realizations of this random function are the step functions which increase by unit jump at each instant of the occurrence of the next event (by 7 if the
moments of the occurrence of n events coincide). Let us find the multi-dimensional distributions [l4,,...,t
of the random function X (t). As at any moment
n
ft the ran-
dom function X (t) may have only integer non-negative values then for determining Lt,...,tn it is sufficient to find the probabilities
‘
n
Pty,..utn(™M1,---,™Mn) = P( () {X (te) = me})
(2.7.22)
k=1
for all m;, = 0 Aleem)
(= 9ny ans
gS
But these probabilities are easily determined by means
ae |, of the known
ee distributions
Vky,...,bn Of the instants of the events occurrences found before. It is easy to see that j EL ermal
eC
CDSs)
Mn
)—
i
x(ti5 te |(t2, ta|) 6
ey
RO
els MD
(tant
ag!
5 ¥ Gesip ty ]
ene CO)
271-20)
For the recurrent stream of the events formula (2.7.23) takes the form ty
.
ti-T1
ti-Tm,-1
Playtest) = f(dn) ff v(dn)... farms) to
0
t2-Tm,
x
t2-Tm,+41
iA (dtm, +1)
‘) (dT m,+2)
ti-Tmy
0
tn-Tmy—y
Dee
eS
0
tn-Tmy—y
ii
VAG fe
mare
ae
/
tn-1-Tm,_1
(oe)
VA diay)
0
In the special case of the stationary Poisson stream
/ tn-Tmn
of events
Claes
e
(2.7.24) f(t) =
\ew*#
where A is the intensity of a stream (the mean number of the events in an unit of
time) formula (2.7.24) gives Pty,..,tn(™1,
AM (tr —
ito
2 eens Mn)
I
(tare ty) eee alt A4 eee =! “Reh sts) m,!(mz —m})!...(™Mn — Mn-1)!
oe
OC
OO
n
:
(2.7.25)
~
2.7. Probability Measure of Random
Function
145
(0< mi) e(h)}) < 6(h)
(2.9.13)
that at any h € (0,7)
at allt,t+7 €T,|| 7 ||=A (The Kolmogorov—Loéve criterion). > For proving that at the fulfillment of conditions (2.9.12) and (2.9.13) almost all realizations of the random function w(t w) are continuous it is sufficient to show that almost at all w
sup
|| z(t’,w)—a(t,w) || 70 at h-0.
t,t’Er
|| t/—-4 ||e})
= 0at7 50.
Therefore any countable set L dense in T is a separant for a(t, w). In particular, the set L of all the points of the region T' with the dyadic-rational coordinates may serve as a separant, i.e. with the coordinates of the type k2~”, where k is any integer,
and p is any natural number.
Thus we may restrict oneselves by the values “ACs w)
on the set I, on the points 7 with dyatic-rational coordinates. Let us introduce the variables
eee lO
spe
SES
Pn
el et
gee Kec
Fy ae
BO)
hn wt 8 eee i)
|
(ea
where the upper bound is taken over all the points | =
(E12
—:
Ke2e* ) ET
and over all q from | till n. It is evident that Zp 2 e(27P) if and only if when even one of the variables
SOE Oo mE ie is larger or equal e(27?).
kg el) eee, KDA)
ie
WEAR
pes
ics instar)
a AE on
Nall
For calculating the number of the variables ae
we denote by 7}, Lebesgue measure of the projection of the set 7’ on the axis m
(m FS
ek n). Then the number of different values of the m*? coordinate of
the type k,,27? in the region 7’ does not exceed 2?T,,.
(ki2—*, aloe E22)
So the number of points
in the region T' does not exceed 2"?7; ...T7;,.
such point corresponds 7 variables ed the random variables Yes ;,,
As to each
oe (q — 1) ree n) then the number of
does not exceed n2"PT, ...T;,. So on the basis of
(2.9.13) we find
Pepe eOPt)) Shy PIV as
beatal2 4) < per PGT, 622). nr
Let us introduce the events
Ap = {Zp = €(2°?)} = {w : Hw) 2 (2°? )},
164
Chapter
2 Random
Variables, Processes and Functions
B, = U 4p, N=limB,= () By. p—=4q
quik
According to the proved above PAG) co
PLB, ys SD EPCAD
1 then by virtue of (2.9.12) P(N) = (). Thus the set NV co
is zero. Ifw € N thenas N = UB, then W € By at any g. But B, =
() Ap.
p=4q Therefore W belongs to all Ap at p > q. It means that Zp = Zp (w) < efr?) at any point WEN
at all sufficiently large p.
Let us take now ie ial odbr
3 | v-l
two arbitrary points I, l! with dyadic-rational
coordinates,
||< 271 then there exists such point lp) with the coordinates
k127-4,..., kp, 274% that all the components of the vectors of the differences | — Ip and I’ — Ig is smaller than 2~%.
Here the difference of the mth coordinates of the
points | and Ig will be expressed by the dyadic number of the type Im
a ys eyaeeS el Rea) B p=qtl where each of the numbers Tp
is equal to 0 and 1. If we perform the transition from
the point /p into | by the sequential steps of the length Tmp Dak (p = qt+l1, q4+2,.. .)
at first on one axis further on the other and so on, on the n*® axis and use repeatedly the inequality of the triangular then we get
ICD En
ouk eae m=1
seme p=q+1
ae nO: p=qtl
The similar inequality we obtain for | z(l’,w) = t(Ip,w)
|. Consequently,
|| (0) —aw) || e}) = 0.
(2.10.2)
The random function X(t) = «z(t,w) is called differentiable in probability on
T if it is differentiable in probability at all t € 7’.
The random function X(t) = a(t,w) is called a.s.
differentiable
(differentiable with probability 1) at the point t € T if there exists such
random function Xj,(t) = x/,(t,w) that
P({w: ah (t,w)
a! (t,w)
at
h—-O})=1.
The random function X(t) = a(t,w) is called a.s.
(2.10.3) differentiable
(differentiable with probability 1) on T if it is a.s. differentiable at all teT.
The random function X(t) = x(t, w) is called differentiable if almost all its realizations are differentiable: P({w:2(t,w) Remark.
differentiable on
T}) =1.
(2.10.4)
It follows from the general properties of the convergence
in
probability and a.s. (Section 2.2) that the random function a.s. differentiable is also differentiable in probability.
The inverse in the general case is not valid.
It is also
evident that the differentiable random function is differentiable a.s., and consequently, in probability.
But it does not follow from the a.s.
differentiability of a random
function that it is differentiable as the uncountable union of zero sets Ny =
{w ;
oP (, w) ee oh (ie w)} at h — 0 even at one m}, t € T may be not zero set.
168
Chapter 2 Random
Example rentiable.
2.10.1.
Variables, Processes and Functions
The random process X (t) of Example 2.7.2 is diffe-
Consequently, it is differentiable a.s.
and in probability.
The same refers
to the random process X(t) of Example 2.7.4. Example
2.10.2.
The random process Y(t) of Example
2.7.3 is a.s.
differentiable, and consequently, in probability at any ft “2 Q. Therefore it is a.s. differentiable and in probability at any interval 7' which does not contain the origin.
But it is not differentiable on any interval (to, t1), to > 0, t) > 2to as none of its realizations corresponding to |w|
e}) =0.
(2.10.5)
As it is easily seen this condition imposes the restriction only on the
three-dimensional distribution of the random function X(t) = x(t,w). Knowing its three-dimensional distribution s4;, 4,4, we may present condition (2.10.5) in the form 5
lim
h,l0
%2—-
1
He,tbembstbemt (4(15 #258) : |Se
h
T3—-— ZL
Seni tans|2 -}) =
(2.10.6) As a result it is always easy to determine whether a given random function is differentiable in probability or not. As regards to a.s. differentiability and the differentiability of a random function then these properties cannot be established directly by means of its multi- dimensional distributions. But there exist sufficient conditions of the differentiability of a separable random function which are analogous to conditions
(2.9.12) and (2.9.13). 2.10.2. Sufficient Conditions of Differentiability
Suppose the separable random function X(t) = a(t,w) with the values in the separable B-space determined in the finite region T' of the
n-dimensional space satisfies conditions (2.9.12), (2.9.13) and besides that at any h, 0 e1(h)}) < 61(h) (m,r =1,...,n).
(2.10.7)
Here €;(h) and 6;(h) are the nondecreasing positive continuous functions such that co
a e272) < 00, p—1
,
82d (DEP I oo.
(2.10.8)
p=
We shall prove that almost all realizations of the random function
X(t) = x(t,w) have continuous partial derivatives over all the components of the vector t. For this purpose we construct for each m = 1,...,n the sequence of the finite-valued functions (p = 1, 2,...): Ue,@) = 2) elke
at
ee, 208
k2? nder(2-?)}, (oe)
oe)
Di,= P=4YOR, Nm =f) Dt. p=1 Analogously as in Subsection that for any
2.9.4 we ensure in the fact that PON a) =
0 and
WE N,, we have uP (w) < n2? sup te) ays?) e1(h)}) < &(h) In the special case of the scalar
(m,r=1,...,n).
(2.10.13)
t we have m = r = 1 and condition
(2.10.13) takes the form
Meith t+an({(@1, 2, @3) :|| e1+a3—2x2 ||> €1(h)}) < 61(A). (2.10.14) Among the functions ¢,(h) and 6;(h) satisfying condition (2.10.8) it is expedient to choose the following functions:
ea = an. Example
2.10.3.
bre) = vA ths aaa
i = 0s
(2.10.15)
Let us consider the centered real random
xX(t) == z(t, w) with the normal four-dimensional distribution.
function
In this case all
random variables
Yinr = X(t + emh+ eph) —X(t+emh) —X(t+e-h)+ X(t)
(1)
are normally distributed, their expectations are equal to zero, and the covariances are determined by formula
on p(t, h) = K(t+ emh +eph,t + emh + eh)
+K(t + emh,t+ emh) + K(t+ e-h,t+e,h) + K(t,t) +2K (t+ emh + erh,t) + 2K(t + emh,t + eh) —2K (t+ emh + erh,t+ emh) —2K(t+ emh + epy,t + e,h)
—2K (t+ emh,t) — 2K(t + e,-h,t).
(iI)
2.10. Differentiable Random
Functions
We)
After determining the variances of the variables Y,,, we get for the probabilities in
(2.10.7) the expressions
P(| Ym |> €1(h)) = 1- 20( er).
(111)
Hence similarly as in Example 2.9.3 we obtain that the condition
Fn O
Wiehe
inir
Alnus,1)
(IV)
at some C, Y > 0) is sufficient for the fact that almost all realizations of the random function X (t) will have continuous first derivatives with respect to all components
of the vector ¢. In particular, the covariance function k(r) = _ y sin w|r|) satisfies condition
(IV) at
vy=
Liz only at
y =
De~*!7l(cos WT a/w.
At other
values of 7 it does not satisfy (IV)). Thus almost all realizations of the normal random function X (t) with such covariance function at
Y = @ /W have continuous
derivatives. It goes without saying that the obtained result does not give grounds for
the conclusion that at Y = a /W the random function X (t) cannot have almost all differentiable realizations.
2.10.8.
Trace Probabilities
The space X7 on which the probability measure of the random func-
tion X(t) is determined by its multi-dimensional distributions represents a space of all functions of the variable t, ¢ € 7’, with the values in the space X. Hence any measurable set of the functions includes not only
the realizations of the random function X(t) but a set of other functions with very irregular behaviour as any set from the o-algebra A” restricts the function only in countable set of the points t. Therefore the space X7 is too large. It is always convenient to consider the random variables in the minimal phase space. In particular, as the phase space of a random function it is expedient to assume the set of all its possible realizations
or the minimal class of the functions which contains all its realizations. Namely on this phase space it is necessary to determine the probability measure of a random function. So, for instance, for a separable random function it is natural to determine the probability measure in the space of the functions satisfying the condition of separability (2.8.1). Let (Q, S, P) be the probability space, 2’ C Q be a set on which the probability P should be transferred. In the general case the set 12’ does not belong to the o-algebra S in consequence of which the probability P is not determined on it. We denote by S’ the class of the subsets of the
174
Chapter 2 Random
Variables, Processes and Functions
set 2/ which represents the intersection of the sets of the o-algebra S with OQ’, S’' ={B:B=0'A, AES}. It is easy to see that S’ represents the o-algebra of the sets of the space 2’. The class S’ contains the empty set @ and space 2’. In parallel with any set B = 0’ A, A € S, S’ contains its complement Q\B in Q! as O/\B = 0/\0'A = 0'A and AE S. Finally, for any sequence of the sets {B,}, By € S’ we have By = 2’ A;z, Apr
ES
and (J By = 2’ U Ax € S' as Ag € S. The o-algebra S’ of the sets of the space 2’ is called the trace of the o-algebra S on 2’. It is easy to understand that the transfer of the probability P on the o-algebra S’ is possible in principle only in the case when P(C) = 1 for any set CES
which contains 2’, C D 0’. In this case evidently P(D) = 0 for any set D € S which does not intersect with 2’. If this condition is fulfilled then the probability P’ on the o-algebra S’ is determined uniquely by formula
PLBY= PYM ApS P(A),
> Really, for any sets Ay, Ag € S such that
A;\A2 and A2\Ai —
P(A2\A1)
B =
Q’A,
=
0’ Ag the sets
do not intersect with Q/ in consequence of which P(A;\A2)
= Oand P(A)
= P(A).
It proves the uniqueness of the definition
of P’. Further on it is evident that Pi a) ) = DAG
(2.10.16)
Meare Ss
0 and PY)
=
Finally for any mutually exclusive sets By = Q’ Ax, ArpES
P(C) SHG (k =
15203. .)
the sets Aj, may be also assumed as mutually exclusive as otherwise their intersections do not intersect with 22'and they may be rejected. Then we shall have
Ply) Be) SP (SAS
Fe Ag) Sheree
(2.10.17)
Equalities (2.10.17) prove the 0-additivity of P'. Thus formula (2.10.16) determines uniquely the probability P’ on the 7 -algebra S'. 4
The probability P’ on the measurable space ((2’, S’) determined in such a way is called the trace of the probability P on this space or the trace probability. Usually the trace probability P’ is determined by means of the notion of the outer measure which is known from functional analysis. The outer probability Pp is determined for any set B C Q by formula PB) = pant. tal(Gan (2.10.18) The trace probability P’ represents the contraction of the outer proba-
bility Po defined by (2.10.18) on the class of the sets S’. > Really, for any set
B € S’ we have
B= 0/AC
A,AE€ESin consequence
of which Po(B) < P'(A). On the other hand, for any set
CE S,
CD
B=M'A
melon Differentiable Random
we have P(A) < P(C) at P(A\C) Consequently,
P‘(A)
the equality Po(B) a
< Po(B). Po(Q'A)
=
Functions
1745)
P(C) as A\C do not intersect with (’.
From the obtained opposite inequalities follows =
P‘(A) for any B € S’. The condition of the
existence of the trace probability may be written by means of the outer probability in the form Po(’) ==, Lene
Now let fz be a probability measure of the random function X(t) determined in the space (X7,.A’); Y(t) be a separable random function
equivalent to X(t). We denote by H the set of all realizations of the separable random function Y(t). Suppose that the set C € A? contains H. The set H restricts involved functions only in the countable set of the points for instance Lc C T. At each point ! € Le only zero
set of the realizations of the random function X(t) does not coincide with the correspondent realizations of the random function Y(t). The countable union of these zero sets correspondent to all 1 € Lc is also
a zero set.
Thus almost all realizations of the random function X(t)
coincide with the correspondent realizations of the random function Y(t) on Le. Consequently, any set C € A? which contains all the realizations
of the random function Y(t) contains also almost all the realizations of the random function X(t). Therefore u(C) = 1 for any set of the functions C which contains the space H of all the realizations of the
random function Y(t).
It means that the transfer of the probability
measure fiz on the space H is possible. We denote by B the o-algebra of the sets of the space H which represents the trace of the o-algebra AT on H, B={B:B=HA,A€ fr Bee(t)X°0 (tz) (twe Thus
Kz f =
K, f represents a vector func-
k=)
tion from the space X with the components CHO
seFe fr EX2(t)X9 (te) (tx).
Hence it is clear that a covariance operator of the vector eure function X (t) is determined by a matrix of the covariance and cross-covariance functions of its com-
ponents
K(t,t!) =
AN OPMRBeey Perma Kon (ts?! Kon (Bytial the Kok(et!)
7
Redgltnttys Haph( eke! Momathhien(t 4a) where
Hy, (iat ho)
20 oO) 1pga
RD
1D)
Similarly we convince in the fact that an operator of the second order moment of
the vector random function X (t) is determined by the matrix I’, (ds t’) of the initial second order moments of its components. Formula (3.3.17) gives in this case the known relation between
the covariance function, the initial second order moments
3.3.
and the expectations.
Moments
195
This relation has the matrix form K,(t,t’)
=
T, (t, t')
— mz (t)m,(t')*, where mz(t)* = [[email protected]. 0], mp(t) = EXp(t) (p 4 Peet n). Analogously as in Example 3.3.2 if the matrix functions K,, Ch toy and T, (t, ts) exist then the domain of the operator Ky and I‘, contains Fy.
Let X = a(w) and Y = y(w) be the random variables in the spaces X and Y correspondingly, determined in the probability space (Q,S, P). If each of the spaces X and Y represent a B-space or a weakly topological linear space, and F and G are the correspondent. adjoint spaces of the linear functionals then formula
a
BXGY = | e@)ave}P (ae)
(3.3.19)
determines the operator ,y mapping G into X. The operator [zy deter-
mined by formula (3.3.19) is called a cross-operator of the second order moment of the random variables X and Y. A cross-operator of the second order moment of the centered random variables \X°
X —m
and Y° =Y—
my is called a cross-covariance
operator of the random variables X and Y. It follows from (3.3.19) that the cross-covariance operator Kz, of the random variables X and Y is determined by formula
Rey
PX GVO = flee) —mz]g[y(w) — my, ]P(dw).
(3.3.20)
It is easily derived the relation between Kzy, Pry, mz and my from
(3.3.19) and (3.3.20) analogous to (3.3.17) Keyg = Ceyg mmegmy..
(320.24)
Hence it follows the following statement. Theorem 3.3.2. For the existence of the operator Kzy tt is necessary and sufficient the existence of mz, my and Tzy. The domains of the operators Tyy and Key coincide. If X and Y are the random variables with the finite or countable
set of the realizations {x,, yx} then Pryg = S > te IVEPE , k
Keyg = Popes — Is
Mx) 9(Ys — My )Pk 5
(3.3.22)
Chapter 3 Moments,
196
Characteristic Functions
and Functionals
where pz is the probability of the occurrence of the kt pair of the realizations (x.y.) (k = 1,2,...) of the variables X, Y. In particular, it
follows from (3.3.18) and (3.3.22) that for the nonrandom vectors X = c, Y = d we have I, = cfe, Pege= cgd, Kz = Kay = 0. It follows also
from (3.3.20) that Kz, = 0 if even one of the vectors X and Y be not random. The random variables X and Y are called uncorrelated if Kryf =0
for all f € F (ie. Kry = 0). If Kayf # 0 even for one f € F then the variables X and Y are called correlated. > Suppose that the random variables X and Y are independent. the probability maesure /lz of the composite random variable L= to the product
of the probability
measures
In this case
(x F Y) is equal
[lz, [ly of the variables
NXGwY
wee leet
p(x) be an arbitrary scalar [/,-integrable function in the space X, and w be an arbitrary {ly-integrable function in the space Y with the values in some B-space or
in weakly complete topological linear space.
In this case according to the definition
of an integral in functional analysis the product p(x) p(y) is integrable over the product jz of the measures [/, and [ly, and on the basis of the Fubini theorem we have
Be XW) = |ole Wma de x dy)
= [ote) [vedas aa) nolae) = f (2)nolae)- f¥adua(a) = Ey(X)EV(Y). 4
(3.3.23)
Thus we proved the following theorem. Theorem 3.3.3. If the random variable U represents a function of the random variable X, and the random variable V represents a function of the random variable Y and one of the variables U, V is scalar then in the case of independent variables X and Y an expectation of the product of the variables U and V is equal to the product of their expectations.
Putting, in particular, p(X) = X° = X — mz, ¥(Y) = gY° = gY —gmy we get Kayg = EX°gY® = EX° Egy =0. Thus the independent random variables are always uncorrelated. The inverse as it is known from elementary probability theory is not always true. Bex
aameprliemsis4e
lf Xe =
X(t), t € J} is a scalar random function
whose all the realizations belong to the linear space X, and Y = Lis) s€Sisa scalar random function whose all the realizations belong to the linear space Y then
3'3- Moments
197
putting in X and Y the weak topologies of Example 3.3.2 by means of the linear F’, and G, similarly as in Example 3.3.2 we come to the conclusion
functionals
Dey and Key are determined by the mixed initial moment
that the operators
the second order Poy (t, s) =
Key (5) = EX°(t)Y%(s) of the functions Hgxeacmup
tise
#3:3.5501t 16 —
random function, and Y
=
X and Y.
oh); t € Ty is the n-dimensional vector
Vis), s € Sis the m-dimensional random function
then putting in the functional phase spaces we come
of
EX(t)Y(s) and by the cross- covariance function
X and Y the topologies of Example 3.3.3
to the conclusion that the operators [ cy and Key are determined by the
matrices Ie (es s) and Key (t, s) of the mixed initial moments of the second order and cross-covariance functions of the components of the vector functions X (t) and
Y(s). 3.3.2. Properties of Second Order Moments
In elementary probability theory the following properties of the second order moments for the random variables X and Y are known:
e [* =T, (hermitivity); e the matrices
[, and K, are nonnegatively determined
(i.e. for
any complex vector u) u?T,u@ > 0 and u? K,u > 0; e Tyc =TZ, and Kyz = Kjy. In elementary theory of random functions for the random functions
X(t) and Y(t), t € T; the analogous properties take place: e
Ty (ty 9 to) —
baits ; to)*;
e for any t,,...,¢y N
Lg ulTp(tp,ty)tg
>Oand
€ J) and any complex vectors uj,...,uNn N
J) up Ko(ty,tg)ttg > 0; p,q=
Poi=, e
Dya(ta,ti)
=
De9(¢i,
2)" and
Eyxitt2.01)
=
Key(tinte)
Let us study the main properties of the operators of the second order moment and the covariance operators. (i) From property (ii) of the expectations follows that operators Ts, Kz, Vcy and Key are the adjoint linear operators. (ii) From property (iii) of the expectations for any continuous linear functionalf follows
fl f= E(fX)(FX) = EB |fX|?>0.
(3.3.24)
Thus the operator of the second order moment and the covariance operator are positive [, > 0, Kz > 0. (iii) From property (iii) of the expectations follows also that for any fas fo (e Pe
mk
filsfo= B(AX)(fpX) = felch.
(3.3.25)
Chapter 3 Moments, Characteristic Functions
198
and Functionals
Thus the operator of the second order moment and covariance operator possess the symmetry property (3.3.25). Analogously we have for any functionals f € F., 9 € Ge
(3.3.26)
flryg = E(FX)GY) = glayf-
Y = aX, U (iv) According to property (i) of the expectations if SSA
|a|” lL
ae
|a|? Fee dpe Cho coukh gay OCIS ae.
(v) From property (1i) of the expectations follows that if
Vow) doy Moan? =e baton vol.
(3.3.27)
v=1
then n
Ty=
0
wali,
Ky=
ar ane
)
LpsRe
=
Pes
DL Gyan
-'Nos —.
V;p=1
wapKryz,
(3.3.28)
Mae
Dot Groene v,p=1
Here T,, and Ky, are a cross-operator of the second order moment and a cross-covariance operator of the random variables X, and X, (v,4=1,...,n) correspondingly.
(vi) From property (iii) of the permutability of an operator of an expectation with a linear operator follows the relations between the operators of the second order moment and the covariance operators of the random variables connected by the linear transformations. Let T be a linear operator mapping a phase space of the random variable X into the linear space Y, and S be a linear operator mapping a phase space of the random variable Z into the linear space U. Suppose that there exist the adjoint operators 7J* and S*. We introduce the random variables Y = TX and U = SZ. From property (iii) it follows that afX 7s a random variable with a finite set of the realizations 21,...,@nN € Dp then Y = TX is a random variable with the realizations y) = Tx1,..., yn = Tan, and the operators Ty, Tyz, Tay, Ky, Kyc and Kzy are expresed by the formulae:
hay piss
Ky =TK,T",
.=
6 Diese spe Te
beret stpag tat Kye =TkKe2 pou hago
\
(3.3.29)
The domain of the operators Ty, Try, Ky and Kzy serves the domain Dry~ of the operator T*, and the operators [yc and Ky, similarly as T,
3.3. Moments
199
and K,, are determined on the whole space F. Really, denoting by G the space of the linear functionals in Y adjoint with Y we obtain for any
g€ Dr CG
Tyg = EY gY = ETX(gT)X
=TEX(T*g)X =TTeT*g.
(3.3.30)
Analogously other formulae (3.3.29) are proved. In exactly the same
way if X and Z are random
variables
a finite set of the pairs of the realizations (x1,21),...,(@N,ZN),
with
71;
...,2nN € Dr, 2,...,2%n € Dg then a cross-operator of the random variables Y = TX and U = SZ are ezpressed by the formulae
Pee
Ss
eK aT ee
Sw
(3.3.31)
and its domain serves Ds-. If X and Y are B-spaces or weakly complete topological linear spaces and the operator T is continuous then follows that if there exist T, and Ky then the operators Ty, Vycz, Vey, Ky, Kyx and Kgy exist also and as the domain of the operators Ty, and Ky, serves the subspace Dp,, and as the domain of the operators Ty, Vry, Ky and Kzy serves the subspace {g : 9 € Dr-,T*g
€ Dp,}.
Analogously, if X, Y, Z and U are B-spaces or weakly complete topological linear spaces, and the operators T and S are continuous then from the existence
of the operators Tz,
and Ky, follows the ex-
istence of the operators Ty, Kyy whose domain serves the subspace {k > k € Ds+,S*k € Dr,, } of the space K adjoint with U. Let X and Y be the separable B-spaces, T' be a closed linear operator and A, C Dr then from the existence of the operators T, and Tyr, Ky and Ky, with the common domain D and nonempty intersection of the domain Ar» of the operator T* with D follows the existence of the operators Ty, Vay, Ky and Kyy with the domain {g : g € Dr-,T*g € D} and formulae (3.3.29). Really, in this case for any f € D there exist the expectations [,f = EXfX, ean = EY fX = ETXfX, and consequently, [gf € Dr and lyzf = TT,f. Similarly, for any g € Dre, T*g € D there exist the expectations
Vee
beer
2 gi
yor BYoYS
A(T 9 ral 9
ee
BEAT? GX;
(3.3.32)
and here [,7*g € Dp and Tyg =TTcT"g. Analogously, if X, Y, Z and U are separable B-spaces, T’ and S are closed linear operators, Ag C Dr, Az C Ds then from the existence
200
Chapter 3 Moments,
Characteristic Functions
and Functionals
of the operators Tz, Tyz, Krz, Kyz and nonempty
intersection of the
domain Ag+ of the operator S* with D = Drp,, = Dx,, follows the existence of the operators Tyu, Kyu with the domain {k : k € Dg+,S*k € D} and formulae (3.3.31) also. Suppose that X and Y are weakly complete topological linear spaces and the operatorT is closed in the topologies of these spaces. We denote by V ={X,TX} a random variable whose phase space serves the graph of the operatorT then from the existence of the operators Tyr, Kur and nonempty intersection of the domain Ar of the adjoint operator T* with the domain D of the operator Ty, follows the existence of the operators T,, Dyc, Ke, Kyx with the domain D and the operatorsTy, Vy, Ky, Kay with the domain {g : g € Dr~,T*g € D}, and formulae (3.3.29) also. Really, in this case for any f € D there exist the expectations of the form (3.3.32) and [zf € Dp and lycf = TT,f. And in exactly the same way for any g € Drp+, T*g € D there exist the expectations
Ppp f OX OVS EX (Pg) anet sl
ge Dr and-tyg=
MHP Teg)
byg = EVO SEI
Oty),
TT pay.
Analogously, if X, Y, Z and U are weakly complete linear spaces, the operator T' ts closed in weak topological spaces X, Y and the operator S is closed in weak topologies of the spaces Z, U, and A, C Dr, A; C Dg then from the existence of the operators T,,, Kyz and nonempty intersection As» with D = Dp,, follows the existence of the operators Pez, Cyz, Kez, Kyz with the domain D and the operators Tru, Tyu, Keu,
Kyu with the domain {k : k € Ds+, S*k € D}, and relations (3.3.31) alSo. (vii) From property (iv) of the expectations follows that the operators and K of the equivalent random variables coincide, lp = Ty, Ky, = Ky, lez
=Tyu, Kez = Kyu if the variable X is equivalent to Y
and Z is equivalent to U. (vill) From property
(v) of the
expectations
follows
that
if
E || X ||’< 00 then the operators T, and Ky exist and bounded (and consequently, continuous). Really, as ||XfX ||_> Yp whose elements are the random variables Yp in B-space converges a.s. and in probability to the random variable X and all its finite segments are majored on the norm by the random variable U, EU? < oo then there exist the operators lpg, Kpq (p,q¢ = 1,2,...), Tx and Ke, Peper ey = Ye. determined everywhere on F, and
fi
al
eels
p,q=1
Here the series converge at any f € Fy.
p,q=1
Kaif
(3.3.34)
204
Chapter 3 Moments,
Characteristic Functions
and Functionals
In exactly the same way if the series )\Y, whose terms are the random variables Y, in weakly complete topological linear space weakly converges a.s. or in probability to the random variable X, if there exist all the operators Ing, Kpq, Vx and Kz and for any continuous linear functional f € F. there exists such random variable Uf, EU} < co that > fYp|
< Uy at alln then formulae (3.3.34) in which the series weakly
p=1
converge tol, f and K,f are valid. Example
3.3.8.
If X is a finite-dimensional random vector then from
the properties (ii) and (iii) of the operators quadratic form correspondent
[; and Ky follows the positiveness of
to the matrices
I‘, and K,, and the symmetry
of
these matrices Ygp = Ypq> Kap = ke. From formulae (3.3.29) we may derive the correspondent
formulae of the transformation
of the matrices
of the second order
moments and the covariance matrices at linear transformation of the random vector X.
In particular,
the random vector
we
obtain the following formulae
Y =
AX
for the covariance matrix
and a cross-covariance matrix of the vectors
Xx: Ky = Ak, A, Kye = AK.
Similarly relations (3.3.31) give the following
formula for the cross- covariance matrix of the random vectors
= BZ: Ky, = AKy;B*. Example
3.3.9.
of
Y and
In the.case where X
=
Y =
AX
and U
X(t), t € Tj is the n-
dimensional vector random function from properties (ii) and (iii) of the operators I‘, and K, follow the positive definiteness and the symmetry
Ppt
land KE
i
sia
iad Pes Bee
P,q= 1
=.
S19
nce
58 fe Ret MIG
10
P,I=
for any I, any vectors f;,..., fj andt,, 1K (ee:i
of the matrix functions
I
... ,t; € T, Te(t;t') =
Deltsee
(esc e. Analogously in the case of two vector random functions
.e— X(t) and Y = Y(s) formula (3.3.26) gives the following relation between the matrices Key (a s) and ast):
Epes; t) = Veale hie
Let the random function X (t) = of the integrability are fulfilled. function
z(t, w) be measurable and the conditions
We introduce the ™m-dimensional vector random
vee J ols) x@otae)
(1)
Ty
In this case relations (3.3.29) give the following formulae for the covariance function of the random function (6) and the cross- covariance function of the random functions
Y(s) and X (t):
Kale cies /)9(s,t)Ke(t,t')g(s",t/)"o(dt)o(dt’), Rey
(IH)
3.3.“Moments
Kyo(8,t) =
205
a(s,t')K,(t’, t)o(dt’).
(III)
Ty
Notice that Tats, s’) and ives: t) are determined by the similar formulae. In the special case when
2 = m = | from these formulae follow the correspondent formulae
for the covariance function of the scalar random functions. Analogously from (3.3.31) follows the formula for the cross-covariance function of the random functions Y(s)
and U(I) = f A( I, »)Z(v)r(de):
Kyals,!) =| f 9(s,0)Kes(t,»)h(l,»)*o(at)r(do.. at
3.8.8.
(IV)
Vi
White Noise
As it is known
from elementary theory of random
processes
the
random process X(t) with zero expectation and the covariance function which contains as a multiplier the 6-function,.mz(t) = 0, Kz(ti,t2) = v(t,)6(t; — te) is called a white notse in the broad sense. Taking into the consideration that 6(t; — t2) = 0 at t; # tz the multiplier v(t,) may be replaced by the multiplier v(t2) or by the symmetric multiplier
\/v(t1)v(t2). The multiplier v(t) at the 6-function is called an intensity of the white noise X(t). The intensity of a scalar white noise is essentially positive. The intensity of a vector white noise represents nonnegatively determined symmetrical matrix. It is evident that the variance of a white noise is infinite, and its values in two quite close points are noncorrelated. For solving practical problems it is expedient to substitute a random process by a white noise. We may do it only in that case when the least interval between the values of an argument at which the values of an random process are practically uncorrelated is sufficiently small. This interval which is called a correlation interval . If the variable
|K.(t1,t2)| /Ke(ti,t1) for the scalar random process X(t) may be assumed practically equal to zero at |t; —t2| > 7% and the variable 7; is
sufficiently small then the random process X(t) may be assumed as a nonstationary white noise with the intensity equal to
oe iArelearn:
(3.3.35)
Characteristic Functions
Chapter 3 Moments,
206
and Functionals
The correlation interval 7; for the scalar random process X(t) is determined by formula
{ ene Ke(t,t +7) ee 1 eld wee max | Rody)
i
3 3.36 ( )
The vector random process may be assumed as a white noise if all its components may be assumed as white noises. 3.3.4. Moments
of Higher Orders
In elementary probability theory for more detailed characteristic of real scalar and vector random variables the following formulae for initial and central moments of higher orders are used:
ee OCGA Ap
ie ett), GPM Ca = Op,,.
(3.3.37)
ry == EXD.r Xo,Tn
OID
Se
y eg he Sp Crar
(3.3.38)
oa Ee?
trp boomy
The vector variable r = (r1,..., 7m) in formula (3.3.38) is called a multi-indez. ‘The following formulae of the connection have place for the scalar random variables between initial and the centred moments:
Nip a
aya; )
=
: dXCe [pm pete Ce»
=
r! Gaze!
ina WyGT
a)
br = (m1) TPCRapmyr? p=0
(3.3.39) oO.
(3.3.40)
Analogously the higher moments for the real scalar random func-
tions X(t), t € T; are determined in elementary theory of random functions Qtr
= Prof er —oo
— oo
S00
adir =
tefelery ccs
DAG
zeithy
eon
oy
ty)
teen
dey,
(3.3.41)
3.3¢ Moments
207
prltpgt obite) SEX? (t)... XL) 2 ov [ler=mmetta))
Burts
oe,
fer ~ melt
elie as y Un ORG © ALpl,
(3.3.42)
The mixed initial and central moments of the order ry ( [rl ‘+--+ Tp) for the real scalar random functions X;(t), ..., Xn(t) are calculated by formulae
dag
alatl
sot
(),
= BY Xi({)... X10)...
ryt)
seve
9 t))
XA OP).. xis
kd order moment
of the n- dimensional vector random
function is determined by the polylinear form of the moments and cross moments of the r*} order of all its components.
3.4. Conditional Moments” 3.4.1. Definitions In elementary probability theory the different conditional characteristics of the random variables and vectors are used. So a conditio-
nal expectation
of a given function y(X) of a random scalar or vector
variable X at a given value of a random scalar or vector variable Y is determined by formula
B[y(X)ly] = / (2) fo(aly)de,
(3.4.1)
where f2(z|y) is a conditional density of the variable X at a given value y of the random variable Y. As a special case from (3.4.1) follows a formula for a conditional expectation of a random variable at a given value of the variable X: co
E[X|y] = | ehelyae.
(3.4.2)
—oco
The denotion f2(z|y) is used for the density of the random variable X dependent on the parameter y also in the case when y is not the value
of some random variable Y. Here formulae (3.4.1) and (3.4.2) determine the expectations y(X) and X as the functions of the parameter y. The expectation of the random variable X as the function of the parameter y on which the distribution X depends is called a regression ofX on y. In the special case when the parameter y is a possible value of some random
210
Chapter 3 Moments,
Characteristic Functions
and Functionals
variable Y the regression X on y represents a conditional expectation X ater nay Knowing a conditional expectation we may determine all conditional moments of the random variables.
A conditional expectation of the random variable Z = y(X) considered as the function of the random variable Y, EF[p(X)|Y] is called a conditional expectation of the random variable Z = p(X) relatively to Y. As the conditional expectations and the conditional moments of different orders of the random variables relatively to random variable Y are the random variables themselves then in its turn we may determine the expectations and the moments of different orders. Finally, we present the formula for total expectation which is often used in elementary probability theory
Bo XY
= Ee
A
al
le
(3.4.3)
Hence at y(X,Y) = X we find
EX =E[E[X|Y]].
(3.4.4)
Now we suppose that on the probability space (Q,S, P) the random variable X = x(w) whose phase space serves a measurable linear space
(X, A) is given. Let us introduce similarly as in Subsection 3.2.1 a weak topology in X by means of sufficiently complete set of the linear functionals F, and admit that X is complete in this topology (i.e. is weakly complete topological linear space). Suppose that the conditional prob-
ability Pr(Alw),
A € S relatively to some o-algebra F C S is regular.
A conditional expectation of a random variables X = x(w) relatively to a-algebra F is called a weak integral
Bp(X\w) = /oe ao Kes
(3.4.5)
In the special case when X is a separable B-space, and F, is a space
of continuous linear functionals on X integral (3.4.5) may be considered as a strong integral (if «(w’) is Pr(Alw)-integrable at a given value w EQ). The measure Pr(A|w) at all A € S is deteremined almost at all w relatively to the measure P and is a measurable function w relatively F. On the basis of the theorems about the measurable functions
fEr(X|w) for any functional f € F. represents the function w measurable relatively to F determined almost at all w relatively to the measure P. Consequently, the inverse images of the sets {x : |fix — fi2xo|
3.4. Coriditional Moments
2h
0 determined by the function
Ey(X|w) and together with them the inverse images of all the sets of the c-algebra A’ generated by this class of the sets belong to the o-
algebra F. It means that the function E¢(X|w) is (F, A’)-measurable, i.e. Is a random variable in (X, A’) determined on the probability space (Q,F, Pr) where Pr is the contraction of the probability P on F. If the o-algebra F is induced by some random variable Y = y(w) then F = Sy, i.e. represents the inverse image of the correspondent o-algebra B of the sets of the phase space of the variable Y. Then formula (3.4.5) determines a conditional expectation of the random variable X relatively to the random variable Y. In this case as it was shown in
Section 2.3 P¢(A|w) is a measurable function of the random variable Y = y(w) relatively to B.
Consequently,
a conditional expectation of
the random variable X relatively to Y represents the (B, A’)-measurable function of the random variable Y. Therefore we shall denote a conditional expectation of the variable X relatively to Y by asymbol E(X|Y) or shortly,
mx |y. By the change of variables
x = x(w) with account of
formula (2.3.7) for a conditional probability measure formula (3.4.5) in this case is reduced to the form
mxly = E(X|Y) = | eus(ael¥).
(3.4.6)
This formula determines a conditional expectation of the random variable X relatively to Y in the case of the existence of a regular condition-
ally distribution ,(Aly). Substituting in (3.4.6) the random variable Y by its realization y we obtain conditional expectation mx, of the variable X at a given value y of the variable Y.
= E(X|y)
3.4.2. Properties of Conditional Moments The conditional expectations possess all the properties of the expectations which are established in Section 3.1 with the only difference that each of the presented statements is valid only almost for all w (or y) relatively to the measure P (correspondingly jy) and may not be valid for some zero set of the values w (correspondingly y). Alongside with these general properties the conditional expectations possess a number of specific properties. At first we prove the following statement.
22
Chapter 3 Moments,
Characteristic Functions
and Functionals
Theorem 3.4.1.If a conditional expectation EX |w) represents a weakly Pr-integrable function and the random variable X has an expectation then for any set BE F
| Po(X\w)Pr (da) ze[Pads). B
(3.4.7)
B
> For the proof we rewrite this formula in the form
‘iPr(dw) /Au peas
eye /OVP (any
(3.4.8)
B
It is easy to see that formula (3.4.8) is valid in the special case when r(w) represents an indicator of any set A E S, z(w) =e 1,(w). Really, in a given case
|Pr(ae) |1ao')Pe(de'le) = |Pe(Alw)Pr (de). B
B
But the latter integral is equal to P(AB), and consequently,
iPz(dw) i1,(w’)Pr(dw'|w) = P(AB) fri 14(w)P(dw). B
From the validity of formula (3.4.8) for the indicators of all measurable sets follows its validity for any simple function r(w).
sis about a monotonous
By the theorem from functional analy-
convergence follows the validity of formula (3.4.8) for any
nonnegative function r(w) (independently of the fact whether the integrals in it are finite or infinite). As any complex function z(w) may be presented in the form z(w) = r1(w) = L(w) + 143(W) _ ix4(w) where 21, £2, 3
and 4
are nonnegative
real-valued functions then formula (3.4.8) is also valid for any complex function r(w)
and from the existence of one integral in it follows the existence of another. Now let r(w) be the function with the values in any weakly complete topological linear space
X for which there exist weak integrals in formulae (3.4.5) and (3.4.8). For any continuous linear functional f € F, formula (3.4.8) is valid also for the complex function fx (w):
|Pr(aw) ffe(u'\Pe(dulle) = ffo(e)P(dw). B
B
ae
3.4. Conditional
From the weak
Moments
213
Py(Alw)-integrability of the function z(w’), the Pr-integrability
of the function Er(X|w)
and the weak P-integrability of the function z(w) follow
three equalities:
/Pr(dw) |Sotpaty =f /Pr(das) f ef!) Px(du' be), [t2)Ptaw) =F fewyP(de), B
B
f |Pr(dw) |2(u/)Px(dullw) =F fe(w)P(de). « B
B
In consequence of the fact that the measure Pr represents the contraction of the measure P on the o-algebra F, and coincides with P on all the sets in F, from the Pz-integrability of the conditional expec-
tation E¢(X|w) follows its P-integrability and the coincidence of both integrals. Therefore formula (3.4.7) may be rewritten in the form
| Be(x\u)P(de) = [ern), B
Bef.
(3.4.9)
B
In particular, putting B = { and taking into the consideration that the integrals at B = Q represent the expectations of the correspondent random variables we get
E[By(X|w)] = BX.
(3.4.10)
This formula is called a formula of a total expectation.
If o-algebra F
is induced by the random variable Y then E-(X|w) = E(X|Y) and formula (3.4.10) takes the form
E[E(X|Y)]= EX.
(3.4.11)
Now we consider a special case when the function z(w) is measurable relatively to the o-algebra F, i.e.
(F,A) is measurable.
In this case
integral (3.4.5) coincides with the integral over the function «(w’) with respect to contracted on the o-algebra F measure Pr(Alw). As for any set B € F we have P-(B|w) = 1p(w) almost at all w then
Br(X\w).= /ee
Na
Chapter 3 Moments,
214
almost at all w.
Characteristic Functions
and Functionals
Thus if the function z(w) is (F,A)-measurable then
almost at all w there exists an expectation of the random variable X
= 2(w) relatively to F which coincides with this random variable: Erx(X|w) = 2(w)
almost at all
w.
(3.4.12)
We shall prove also the following theorem.
Theorem 3.4.2. If a scalar random variable Z = z(w) 1s measurable relatively to F then for any random
variable
X = x(w) for which
there exists E¢(ZX|w), almost at allw
Ex(ZX|w) = z(w)Er(X|w) = ZEr(X|w).
(3.4.13)
> Firstly we consider a case when z(w) represents an indicator of the set
B € F,
74(6) ms 1p(w). Then
Ex(ZX |w) = fto(o')e(w!)Pr(do' le) = J ee) Pe(de'le). B As Pr(Blw)
= 1, (w) almost at all w then the latter integral coincides with
/AEN ame Saha almost at all w € B and is equal to 0 almost at all w € B, ice.
i#() Pe(ae!
())SAge)Be Oro)
almost at all W ,what proves formula (3.4.13). Consequently, formula (3.4.13) is also valid for all simple functions Z (w) relatively to F. Now let z(w) be an arbitrary nonnegative measurable function relatively to F.
After presenting it as the limit of the increasing sequence of the simple functions
{z” (w)} we get the random variable ZX
as the weak limit of the consequence of
the random variables {Z” X } which a.s. converges and for any functional
f € F,
the sequence {Z eX } is modulus bounded of the random variable Z |f.X | which
has a finite conditional expectation.
By property (xiii) of the expectations (Sub-
section 3.1.3) follows that E¢(ZX |w) = lim(weak)
F(Z" X |w). And as formula
(3.4.13) is valid for the simple functions Z” = 2” (w) then almost at all W we get lim (weak) Ex(Z" X|w) = lim ( weak ) Z” E-¢(X|w) = ZE F(X |w).
3.4. Coriditional Moments
This proves formula
DANS
(3.4.13) for the nonnegative random
variables
Z.
Accoun-
ting that any complex random variables may be expressed in terms of four nonnega-
tive real random variables we ensure in the validity of formula (3.4.9) in the general case.
0
a
0
5
1ah
it
o< [Ey uf
r€(a,b), _
: at
sin f(z + a) — sin f(x — b)
at
lyrate
orviay ric iby
uenje (afb)
a
eee
< [tw < 7 TU 0
then by virtue of (3.5.28) the integrand in (3.5.27) enclosed in parentheses at all F is absolute-values bounded by the [4,-integrable function p(x) =
2. Therefore
the limit transition in (3.5.27) may be fulfilled under the sign of the integral. Then accounting that the measure [ly of the perimeter of the rectangle R is equal to zero
we obtain
im. p= J tr(e)ne(de) = ee tuys
(3.5.29)
what proves formula (3.5.24).
Any rectangle may be represented as the limit of a monotonous sequence of the rectangles of the considered type and the measure [lz is continuous.
Therefore a
characteristic functional of a random vector X determines uniformly its probability measure
/l, on the set of all the rectangles of the n-dimensional space X.
In the
consequence of the d-additivity of the measure /l, it may be uniquely extend of over
the 0-algebra which is generated by the set of the rectangles, i.e. on the 0-algebra of Borel sets of the phase space of the random vector X.
When
the characteristic
function 9x(f) of the random vector X is integrable (over Lebesgue) because of the boundedness of the second multiplier under the integral (3.5.24) the whole integrand
in (3.5.24) is integrable. Therefore formula (3.5.24) may be rewritten in the form
*
PM
omifnar— o-ifnbe
po(R)= fo f [pe
telfis
Salt -th
228
Chapter 3 Moments, Characteristic Functions
= aay faetna
and Functionals
feta.
(3.5.30)
R
In consequence of the boundedness of the function 9u(f)e~*F ” and its integrability over f it is an integrable function of two vectors f and Z. Therefore we may change
the order of the integration in formula (3.5.30). Then we obtain
fia) = La’ je /ef? g,(f)df .
(27)"
R
Hence it is clear that in a given case there exists the probability density f, (x) of the random vector
X which is determined by formula
1
aah fz vy) (2)= ence
ete
Gx(f df
:
+0. (3.5.31)
Then formula (3.5.4) takes the form
92(f) = [eo tele)ae
(3.5.32)
And vice versa if there exists the probability density ts (x) of the random vector
X over the Lebesgue measure then its characteristic functional gx(f) is determined by formula (3.5.32). On the basis of the known Fourier theorem Tee) is expressed in terms of 9n(f) by formula (3.5.31) and the function 9x(f) is integrable.
Let now X be a random variable in an arbitrary real topological linear space,
and Gade) f € F be its characteristic functional.
According to property (vi) of a
characteristic functional (Section 3.5.2) the expression
n
Gla PEIN, a=
oF
y arf
(3.5.33)
Kao
at any fixed fj ,---)
dn € F represents a characteristic function of a random vector
with the components Z; = ff, X,...Z, it determines
= fn X. In accordance with Theorem 3.5.1
completely and uniquely the probability measure
Lt of this random
vector on the 0-algebra of Borel sets of the correspondent n-dimensional space, and in particular, on the rectangles {z : 21 € Ay, .+. SRALE An}: But the random
3.5. Characteristic Functions
and Functionals
229
vector Z represents a function of the random variable X and as the prototype of the rectangle iz : 27
Ay,
X serves the set {x wiz
298A
E An} in a phase space of the random variable
Ga;
wee
inte
An}: Consequently, on the basis of
formula (2.2.4) we have
Hy({2:24 CIA), «.. , Zn
Ant) = poe:
fiz EA
Hence it is clear that the characteristic functional Gt)
,. ., In@ © An}).
(3.5.34)
together with the measure
[lz of the vector Z determines also uniquely the probability measure [a of the random
variable
X on the sets of the type {x PUTS
5 Pe
ne
Piead Dae
oeON
ey
Al ,c.ec
Ink & An} at given 1,
te are arbitrary then a characteristic functional
determines a probability measure of the random variable X on such sets which are
correspondent to all n, f} ,...
, fn € F and to all Borel sets Ay , ... , An. After
determining the measure [lz on the class of such sets we may extend it uniquely on the o-algebra generated by this class of the sets
As it is known lues in the B-space
=
from the theory of the measurable functions with the va-
there exists a sequence
eo) ae Ye lpn (z) (n ia.
of the elementary
Fy)uniformly convergent
functions
to y(z).
y” (x)
On the basis
k=
of property (xii) of an integral with respect to the bounded function over a finite measure for all p the inequalities are valid
|[ere
fvdny SLO
[vane fv Let us take an arbitrary € >
y(x) || ,
a sup |[y"(2) — y(z) ||.
(0) and choose 7 in such a way
that the inequality
sup, lly” (x) = y(«)|| < €/3 will be fulfilled. After that we choose Po in such a way that at all p > po the inequalities will be fulfilled
tp(D) — u(DE)| < See :
3Namaxlvel |Eee
3.6. Sequences of Probability Measures and Characteristic Functionals
231
Then we shall have at all p > po
|[van — fva] < |[vee +| [vray
3.6.2.
fat any
fvray| +] frau
fava Let F,oe) be a distribution function correspondent to the measure [lp (Sec-
tion 2.2).
Show that we may single out the subsequence {Fo, (x)} convergent to
some distribution function {F,(x)}.
(x) at all the points of its continuity from the sequence
For this purpose we take any countable set D,
for instance, a set of all the points with rational coordinates.
at all p then we may
single out from the sequence
=
4 Pina, dense in X, As 0 < Fy (xy) < 1
1H, (2)} such a subsequence
{Foi (x)} that the numerical sequence {Ft (a4 )} will be convergent.
Analogously
we may single out from the sequence {Foi (x)} such a subsequence { Fn2(ax)} that the numerical sequence
{ F,2(2)} will be convergent.
Continuing this process at
any T we single out from the sequence eee el (x)} such a subsequence defor (tt) that the numerical sequence {Poe
Set will be convergent.
The diagonal sequence
WEES EY, will converge at all the points of the set D, as the sequence {F(@;)} at any Tf is a subsequence of the sequence {ieee} y at p > Tr. Suppose that the function Fp (x) represents the function F'mp (x) from the initial sequence {Fm(«)}
(p = 1, 2,...). We denote the limit function Fp,(«) = a and define the function F(a) =
Lm
C,ps+o
Fim, (2) at z € D,
Fp,(#r), where {x,y }is the sequence of the
232
Chapter 3 Moments,
Characteristic Functions
and Functionals
points of the set D, convergent below to the point Z, Z; < Z. We show that the se-
quence {Fim (2) } converges to F(a) at all the points of the continuity F(a). From the inequality Fin, (2’) 2/(aé). After this accounting that u(B) == ait Lm, (B) < C we choose pz in such a way that at all p > pp would be Lm, (B) . NG:A bo) v=1
where z(t 4) is some function which belongs to the space X.
Consequently, in this
case ® = X a positively determined functional 9(f) on f serves as a characteristic functional of some random function X (t). Exam
p1e
the scalar functions,
3.6.3. Consider a case when X represents the B-space Lp of
1 < p < 00. We assume as F’ the space Dg, q= p/(p a 1) of
continuous linear functionals on Lp dual with Lp . The space Lp is reflexive in such
a sense that the set of all continuous on Iq linear functionals coincides with Lp. But
3.6. Sequences of Probability Measures and Characteristic Functionals
241
the space of all (not only continuous) linear functionals Ly does not coincide with Dy.
Therefore not every continuous positively determined functional 9(f) on Lg
determines the probability measure on Lp :
3.6.5. Probability Distributions in Banach Spaces. Mourier Theorem Now let us consider the case when X represents the B-space. It is natural to assume as the set F of the linear functionals on X topologically dual with X. In the general case the space © of all linear functionals on F, does not coincide with X. According to Theorem 3.6.6 not every continuous positively determined functional g(f) on F, determines the probability measure on X. In order the functional would determine the probability measure in X it must satisfy some additional conditions. Notice that at any fixed f,,..., f, € F the function
IA eeleieaian)
my
(so)
(3.6.12)
=a
of the real variables a;,...,@,
determines not only the n-dimensional
distribution py,.¢, of the random function Z(f) (Subsection 3.6.4) but also the distribution
e 12 spoudo2) 6 By) PfaessifAF = Hhs,.ufn L2(f) : (2(fi), 3 -,2(fn)) S B})
on the o-algebra A;,,5,
of the sets{x : (fiz,...,fnx)
(3.6.13)
€ B} of the
space X correspondent to all the Borel sets B of the n-dimensional space R”. The aggregate of all such sets correspondent to all finite sets of the vectors f;,...,fn € F as it is easy to see forms the algebra Ao of the sets of the space X. Formula (3.6.13) at all n, fi,..., fn € F determines on this algebra Ap the nonnegative additive function of the set w,. If the function pz is g-additive on Ao then according to the theorem about the extension of the measure we may extend it till the measure on the
o-algebra A generated by the class of all sets {x : (f,z,..., fnv) € B}. But from the theorem of Subsection
3.6.4 we may conclude that the
function p, in the general case will be not o-additive on Ag. It will be o-additive only in that case when the distribution py in the space (®, F)
is completely concentrated on the subspace (X,.A), A= XF. The arguments given above show that for finding the conditions at which a continuous positively determined functional g(f), f € F
Chapter 3 Moments, Characteristic Functions and Functionals
242
determines the probability measure on X it is sufficient to know at what conditons the function of the set jz is the o-additive on the algebra Ap.
For short notation we determine the sets Ah, = {x : |fpx| < N} and put
By =() Aw= 12: 8UDpey pela) p=
UNG Nee eee)
(a0 da)
z
All these sets belong to the algebra Ao and the function pz is determined on it. Let us prove the following statement.
Theorem 3.6.7 (The Mourier theorem). In the case when X 1s the B-space and dual with X the B-space F, is separable (in this case X is also separable) for the o-additivity of the function pr, on Ao tt ts necessary and sufficient the existence of such a countable set {fr} which as dense on the unit sphere of the space F, that at anyé > 0, at all natural n and all sufficiently large natural N the condition will be fulfilled
Ha lie. Sup |fpe| > IV})
For proving the necessity of condition
(3.6.15) we suppose
that [ly is the
measure on Ag. Then we may extend it uniquely till the measure on the & -algebra
A generated by a weak
topology of the space X.
But in the separable B-space
o-algebra generated by a weak topology coincides with the 0-algebra of Borel sets generated by a strong (i.e. natural metric) topology.
Consequently,
the sets By
= HED : \|x|| =< N} belong to the o-algebra A and the function fz is determined for them.
As {Bn} is a decreasing sequence of the sets and lim Bn
then by virtue of the continuity of the function {ly we
= 1) Bn
have lim Lz (By)
Hence at any € > (0 there exists such N, that Lz (Bn)
ClAgy
Now we suppose that (An
=
‘then of the
©CO we
may
Cin iG As... tbat
@. Then it will be
and if lim ps (Cp) = 0) then at any € > 0 there exists such nN, that
n,.
After choosing nt. thatieny
nt
we get Me(An) < Ur(Cn) + én < € at alln > nz, nt. Hence it will follow that lim Ha( AR) =
0. Thus it is sufficient to prove the continuity of [ae in zero for the
sequences of weakly closed sets {Gye Let {Cn} be an arbitrary decreasing sequence of weakly closed sets, C, € Apo, with empty intersection Grbts n —
OOo.
GG
() Ch
=
@.
Prove that Ha (Cn) —
0 at
For this purpose we suppose that [Uy C,) > € at all n for some € > 0
and show that in this case the intersection of the sets C;, cannot be empty.
At first
we show that at any sufficiently large NV the intersection of any set C', with the set
Bn = tn : ||| < N} cannot be empty. As Cy
&
9i,--+;9m
C,Bn
=
€
@
Agvthen. Fe
Cy
=
ee
: (giz, seat ; Gene) (S Beir} at some
and the closed set F,, of the m-dimensional
space
R™.
™, If
then
reamintalte||eaN"
(3.6.16)
TEC,
In fact, in consequence of the fact that the linear functionals g1,...,9m
€ Fc are
continuous in the strong topology of the space X generated by a norm the relations 21 = 91@,..-.,;2%m
= Jm¥@ establish a continuous mapping of X on R™.
Therefore
the set C’, as the inverse set of the closed set F, at this mapping is closed in a strong topology of the space X.
Consequently, in C, there exists the point Zo in which
the continuous nonnegative function p(x) = ||| achieves its accurate lower bound r and if r < N then the intersection C,Byn
cannot be empty.
Now we consider the function (21, ess rn) which is inf ||2|| on the set {x : Jit = 21,-+--5JIme
=
Oi
Pm}
aitpel Betz
pW
inf
|||.
(3.6.17)
91 T=21,---)Jml=2Zm
As this function is continuous p(azi, Feds 5 OZ; ) = ay(z1, hd; at) at any a > 0 and y(0, Jee
0) = 0 then the set {21, Pom
aes
p(21, Pra taA Erni) < c} represents
244
Chapter 3 Moments,
Characteristic Functions
and Functionals
a simply connected region of the space R’™ which contains the origin. Therefore it follows from (3.6.16) that if C, By
=
@
then the set F, does not intersect with
the set {z1, eterna
p(21, «te ayn) < r} (i.e. is disposed outside of this set).
The set Do = {2}, A:
Bay: p(21, ae
5 ia) 0) is convex.
let {oka teed host E€ De, {iz1's aL eA function p(21, Hei
€ D,.
Really,
From definition (3.6.17) of the
5 2m) follows that at any € > 0 there exist such z’, 2” € X that
gee elope! kmvaii(k aid pi alpm)and: |e fez] po Ze) Rep ||e44| < p(2i,...,2,) +e. Taking € < c— max{p(z4,..-52m); P(21,---> 2m)} we have \|a’|I, \|ar’’|| N|hil|.
Really, by virtue of the completeness of the B-
spaceX there exists the vector 2g in X for which ||zo|| SA
chize = ||Ai|I. Putting
at an arbitrary 6 > 0 x5 = (b; +6)x9/||Aill, we shall have \|5]| = (b; +6)/||hal|
and hyxs = 6} + 6. Consequently, the point 25 belongs to the set iz shye. > by} and if b;) < N||hil| then by virtue of the arbitrariness 6 > 0) \|s]| < N. But if b; > N||hil| then N|\hil|
N.
Introducing the linear functionals hi =
245
hi/||Ail|,
belonging to the unit sphere of the space F’, we obtain from (3.6.18) and from the fact that b; > N||hil| the following relations:
Cn C U {x : hjx > bi/|lAil|}
C U {a : |Ajz| > N},
l=1
(3.6.19)
f=
tte(Cn) < fhe (Ut : |afe| > vy) | As according to the condition the set ee the space
(3.6.20)
is dense on the unit sphere of
F’, then for each hj there exists convergent
to it the sequence
. {fo}, fir = hf ge hi at r — co (l = 1,...,58). continuity of the functional 9(f) at T — ©O we have
{fir }
Here by virtue of the
s
Gini
(Wises
40s)
=
9
Safir
Nesil
s
9
So arht
= 9h',..,h(@1,---,@s).
(3.6.21)
[fees
But
een ar AO
istic function
inverse
of some
See Qs) at any
S-dimensional
Levi theorem it follows from
ty measures converges
correspondent
vector.
F. represents
by virtue of the
(3.6.21) that the sequence
of the probabili-
correspondent
ae hasan re)
to the characteristic function
Qs). Consequently, at any N > 0)
ee (Ute : |Aja| > v}) = im Ite (U420s figatl r=" At any f we have
a character-
Therefore
to the characteristic functions gerne
to the probability measure
Lf eee A (a1, oe
Sia eifen fe €
random
f
vy) ais
(3.,.6:22)
=|
s — U {x : fir | > N} (© BN: where 7, is the largest of the num(eal
bers of the functionals fj, =
dais tee Pe
Foote n=
max{pir, Ad
Deas
And as Teak ES) < € at all n thenwe get from (3.6.22)
Te (Ut Seo
vy) Zon Bet)
ee
(3.6.23)
246
Chapter
3 Moments,
Characteristic Functions
and Functionals
Accounting (3.6.23) from (3.6.20) and (3.6.22) we get fle(Cn) < € what contradicts the supposition that fis('Gy)
> €.
Thus the intersection By Cy cannot
be empty if there exists such € > 0 that jee Cia) > € at all n. a centered system of closed subsets of the set By
X.
And the set By
But {BnC,, } is
in a weak topology of the space
is bounded and consequently is compact in a weak topology.
So the intersection () BynCy,
cannot be empty,
and it means
that the intersection
n
() C, also cannot be empty.
Thus we proved that if there exists such € > 0 that
een)
> € at all nm then () Cr e
if () Cr
=
@
@ . Hence it follows that lim filC,)
=)
what proves the sufficiency of condition (3.6.15) for the continuity
of [ly in zero on Apo, and consequently, for the o-additivity of Zz on Apo. Thus the
sufficiency (3.6.15) is proved.
Really, on the basis of formula (3.5.9) a characteristic function of the random vector Z is determined by the formula n
GeO a> gy Ons.) Se Oe
ied p=1
3.7. Normal Distributions in Linear Spaces
J
n
247
n
1
= exp ‘ ) Op fp Mz — 5
y
orb
keh.
P,q=1
p=!
But fp Mg, represents an expectation of the random variable
7, = fo X , consequent-
ly, 9z(a1, nied Qn) is a characteristic function of a normally
distributed random
vector, what proves Theorem 3.7.1. Further, a characteristic function of the variable Z in Theorem 3.7.2 is determined by the formula
; 1 4 D, > = exp § iam, . 1, bp od gz(a) = exp 4 tam, — 9% — re On the other hand, 9z(@) =
Be'%4
=
oer
Bet*i% | Hence it is clear that a charac-
teristic functional of the random variable X in a given point f € F' is equal
GAf) Se! ifx
G41)
exp
ifs
— Pg 5fKof
As it is valid at any f € F then 9a(f) at any f € F is expressed by formula
(3.7.1).
Really, for any f},fo € F, fi -
= fo) =
BX (Fp fa)X*
(3.7.6)
fe, Kefi
=
Kafe
we have Kabfs
0 ond (fie fz) Kah; — fo) = Eh
= fo) X°|? = 0. Thus (fi = fo) X° = 0 with probability 1 and Kya(hi = fo) —
BY
Chi — f2)X°
uniqueness
=
(0). Hence we find Kychi
of the mapping
X —>
Y determined
=
Kyo fo what proves
by Eqs.
(3.7.5).
of this mapping follows from the adjoint linearity of the operators
the
The linearity A,
and Kye
(Section 3.2). Now
we show that the operator L, determined by Eqs.
For this purpose we notice that at any
for all 29
=
Kyfo
and
© =
|fxy = fox1| 0 \gLx = gLxo|
then Ffn& —
Characteristic Functions and Functionals
fxo at all f € F. But %, = Kifependas
fxo at any
~igla'=
Sieh
@ E A K, and lim f,@, evidently, represents the result
of the action of some linear functional fp on %.
Consequently, the sequence of the
functionals thn} converges to fo in such a sense that fnt —
fox atallze
ie.
If fo € F then Ky fo€ Ak, and as Ky fo = £0 by virtue of the continuity of the operator Ky, then £0 € jG The continuity A, in the weak topologies of the spaces F’ and X follows from the fact that by virtue of the equality f; K, f =
or ; \fix’ = fiz| O and f,f;
— fev K cE}, “=
If fo€F for some fundamental sequences
sept
Tela
€ F
of the point ‘i
C Ax,
then extending the
space F’ on account of the inclusion into it all such fp and extending the operator ie on the continuity we get Yo € Auge for all fundamental sequences Zo
=
limz,.
Thus
Age
may
Keyg e Ax, at any g € G. Otherwise for the functional does not belong to Ae
Latved oa Ni,
be always considered as a closed set.
But
then
g € G for which Keyg
we may find such linear functional f € F' that fx = 0 at
any 2 € Nig. and S Keyg a 0. But it is impossible as
fKeygl? =|EGX oY DP REY
XP ElgY|? =f. fe giao
and from fx = Oat x € Ax, will follow fK f = 0 and Kryg = 0). The obtained contradiction proves that 1
= Kayg E Ak, at any g € G. But then there exists
such linear functional f; € F that x1 = Kf,
and
|\fei — fori| = |fKefi — fokefi| = |fikef — fikefo| = |fie — fizol. Consequently,
|gLa = gLzxo| 0,g E€G, x
€ Ak, and all Z in
iS re \fix = fi Zo| Let us consider the linear mapping S'
uu =
of the space measurable
Z =
(x, y) tame
X
X Y
as the operator
=p
=
into itself.
torus
y! =
G—
Le
This mapping evidently is continuous and
1 according
to the proved is continuous
and for any
h=(f,9)€H hv = fe+ gy = fze+ gy— gle =(f —L"9)e + gy = h'u, where h’ =
(f = L*g, 9) € H. As a result of this the set {v : (hiv, sete hnv)
S B} € C has the inverse set {u 3 ( iu, : HO fa kan) Sc B} € C. Therefore the random variable
Zins® (ar, (0) in the space
Z = X XY
Sar
oY
is normally distributed. We find a cross covariance operator
Kytz of the variables Y’ and X. Putting
ef SEY
POSEY
But according to (3.3.29) Kw, it follows from Eq.
eYeaaiee LK"
W = LX° we find
Gea) =
LKg
(3.7.6) that Kyte =
(LX
GX") = Kye or hue
therefore Kyte =
Kye — LK,.
Hence
0. Thus the random variables Y’ and X
are uncorrelated, and consequently, by virtue of the normality of the distribution of the variable Z’ they are also independent.
Therefore the probability measure of the
random variable Z’ = (X, vs |is determined by the formula (Section 2.4)
Hz1(C) = f ue(de)
f10(e, nay (de), Cie ©,
(3.7.7)
where /lg and /ly’ are the probability measures of the random variables X and Y’. Now we notice that the mapping S is convertible as to a given value
V =
corresponds the unique value U =
Therefore on
Gea
eo y}, y =
y + La.
oe y }
the basis of formula (2.2.4) from (3.7.7) it follows the expression for the values of the
254
Chapter 3 Moments,
Characteristic Functions and Functionals
probability measure of the random variable Z on the measurable rectangles Ax
B,
AEA, BEB:
uz(Ax B) = uz({(2,y): = p({(z,y'/): 2 € A,y’ €
2 € A,y € B}) B—m, — L(x — mz)})
= fy (B= my ~ E(e— me) ad). A
Hence we conclude that there exists a conditional distribution of the random variables
Y relatively to X connected with probability measure of the random variable Y’ by
formula
Hy( Ble) = py
(B-—m, —L(e#—mz)),
BEB.
(3.7.8)
This formula shows that Hy (Bx) at any fixed value & of the random variable X represents a measure
determined on the d-algebra B.
Consequently,
a conditional
distribution of the random variable Y relatively to X is regular (Section 2.3). Fi-
nally, it follows from (3.7.8) that a conditional distribution of the random variable Y relatively to _X is normal as the random variable Y’ is normally distributed.
By the
symmetry we conclude that in the considered case there exists a regular conditional distribution of the random variable Y relatively to X and this distribution is normal.
|||. Hence on the basis of
p=1 property (ix) of the expectations follows nm
n
BYZ} = EY (X", pp)? — El|X°|?. foal
pl
As E(X®, Pp)” = Oe Pp; Lp) = Ap and the series SS Ap converges then [e.@)
E\|X|? = lim > E(X°, gp)? = p=1 Hence it follows that E\|X°||? =
Theorem
3.7.13.
[o-e)
YE: p=l
|ma||? + E||X°||? 0) we may write
pa({ssu(s) >€}) alr],
E(X, Yp)?,
p=1
(3:7.21)
p=1
where X is a random variable for which the contraction of the function [tz on the
o-algebra Aine.
serves
as the probability
distribution.
For determining
the
expectations in this formula we shall find a characteristic function of a joint probability distribution of the random variables Z,; =
(xX, £1); 1,0 Gn eS (X, Yn).
According to formula (3.5.9) this characteristic function is obtained from (3.7.17) at n
i=
De Ap Pp (Section 2.4): p=1 n
UE
ey
ae
wis oo)
=
exp
a )
Ap (Mz; Pp)
p=1 1
=
n
nm
7 CrgOa) os pig Ka
p== exp
P,q=l
Se ims, en) — 5aAp p=l
Hence it is clear that the random variables Z,,..., Zp, are independent,
ly distributed with expectations
normal-
(mz, 1), ns, Gite. Yn), and the variances 1,
, An. Therefore E(x, ye) = EZ = (mz, Pp)” =F Ap- Substituting this expression into (3.7.21) and taking into account that by virtue of the Bessel inequality n
(Ms; ¥p)” < ||mell?, we get p=1
to
(42: (ee)? > N2b |< 5 [Imell? + ap n
Z
1
2
p=1
2
nm
2
3.7. Normal Distributions in Linear Spaces
263
In consequence of the convergence of the series a Ap we conclude that at any
€ > 0 there exists such N, that at all n and N SN n
fag
li
tees
(3.7.22)
p=1
Now we notice that the set of all finite linear combinations of the eigenvectors Yp of the operator Ky with the rational coefficients is countable and dense in X and therefore the set of the vectors
pe
Fn =
AnpPp,
ashe
Anp = LI Hree
p=1 correspondent
'np
TTT LTE
Cie
&
te aN.
to all possible finite sets of the rational numbers
Tn1,..-.,TnN,,
is
countable and dense on the unit sphere of the space X. Here in consequence of the Nn fact that >. a = 1 for any z € X the following inequality holds: p=l1
2
Nn
(2, Jn) =
Sones,
Pp)
Na
Sot (x :Pp)*
p=1
pl
Hence it follows that
and at any narutal N we have
{e:sup (2, f)|>N}C fe: D> (yp)? > NPP p=1 Using inequality (3.7.22) we obtain
tell Saceupille i fy)| > N}) 0, where {%p } is
p=1 n
an arbitrary sequence of the vectors of H-space
X then (Ke ip f) —
a3 Ap Ge ~p)? p=1
and formula (3.7.17) gives
golf) =exp }i(rme,f)— >Yolen)? p-
(1)
In this case for any vector f orthogonal to the subspace G' generated by the vectors
Y1,---,;Pn variable
we have (Fe £1) =
Z =
Ch Yn) =
0). Consequently,
the random
(Xx; f) has the expectation (ite, f) and zero variance.
It means
that the distribution of the random variable X is completely concentrated on the
hyperplane Mm, + G.
Example
3.7.4.
The functional g(f) =
mal distribution on any O-algebra Axl
e7 1/211 FI? determines a nor-
AF. of the sets of the type {x : ae, fi),
oon (ar, ta)) (S B} at the fixed linearly independent
f),..., fn. Really, putting
n i =
SS Op tp we obtain p=
gf) =exp §=5 D>apa Sys Sa) p
(1)
p,q=l
The right-hand side of this formula represents a characteristic function of some normally distributed n-dimensional random vector Z = {Z;, ster Lae
The expecta-
tion of such vector is equal to 0, and the elements of its covariance matrix K, are
the scalar products (Ses fg As f1,..., fn are linearly independent then the determinant of the matrix K, is strictly positive in consequence of which the distribution
of the vector Z is not singular. Let us consider the random variable
XK Spare Sul
be a lic:
(II)
Qe
where koeq are the elements of the matrix
oa 1
This is the normally distributed
random variable in H-space X whose all realizations belong to the n-dimensional
subspace G formed by the vectors f,,..., fn. A characteristic functional of the variable
X is determined by the evident formula
gx(f) = Bef.
n
= E exp
n
aya q—
koa, fr) |24 lap
3.7. Normal Distributions in Linear Spaces
=o.
(Soi f fp):
265
COD konlf, ia) p=1
=o
{=} t 3 (sede)rye ey 2
q=1
(111)
Sal
But SS (psf) piI=1
3
Rep ks, akae fr )(F; fs)
Tsai
=> (f, fre)(f, fs)HF Leh
(IV)
ns= Accounting that Us P ‘PS are the elements of the matrix K,, and ke are the elements .
°
ow
of the inverse matrix K’>*
Oe
n
.
in consequence of which
=
ee
2 keq( fas fp) = 5sp we get q=
(Fp, Ia) pee Kepksalds Ir )\Ipte)
p,q=1
=
(seal
py aa
urs
Cie ee
(V)
Tesi
The latter expression represents a square of the norm of the projection of the vector ij
on the subspace G.. Therefore the obtained expression for the characteristic functional of the random variable _X may be rewritten in the form gz (f) = e~1/2\|PfIl
2
, where
P is an operator of the projection on the subspace G. In particular, for any vector
fsaG
;
ge(f). sien
-1/2
=a).
(VI)
The contraction g(f) on any finite-dimensional subspace of the space X simi-
larly as the contraction of a characteristic functional of a normally distributed random variable determines a normal distribution on this finite-dimensional subspace.
If we take the orthonormal vectors f,,..., fn then the random variables 27, ., Zn will be independent and their variances will be equal to 1. Therefore
ae
He ({:; Nec)
n
< v}) < be (A{x : |(z, fp)| < N}
= [[ vole :I(e, fo) < N}) = [28(9)]". pal
(VII)
266
Chapter 3 Moments,
Characteristic Functions
and Functionals
For Laplace function 0(N) at any N we have 0 < 20(N)
< 1. So at n —
00 we
get from (VII)
Te ({>: Neila fn ES vs} — 0. ep) co
Thus as ys (gz: eae = \|z||? we conclude that [lz cannot be a probability distribupail
tion on X.
3.8. Approximate Representation of Distributions 3.8.1. Pearson Curves System In applications it is often required to represent a density or a distribution function of a random variable approximately by a suitable analytical expression. At practice any simplification of an analytical representation of a density leads inevitably to the parametrization. The system of the distribution curves of Pearson (Pearson 1894, 1895, 1902) is widely used for an approximate analytical representation of the densities of scalar random variables. Pearson noticed that majority of the densities y = f(a) satisfies the following ordinary differential equation: dy _
z+a
dx
bo + by2 + box?
where a, bo, b;, bz are some constants. In particular, normal, exponential, Student, 6-, y-, y- and y?- distributions satisfy this equation. Varying the parameters a, bo, b;, b2 one may obtain a great variety of distribution curves. The most general way of approximate representation of distributions is their representation in the form of linear combinations of some given functions, in particular, in the form of various truncated series.
3.8.2. Orthogonal Expansion of Density Let w(x) be some density in the r-dimensional space R" for which all the moments exist. A system of pairs of polynomials p,(zx), q(x) (v = 0,1,2,...) is called a biorthonormal system with the weight w(x) if co
ilw(2)p,(x)q,(a)de = 6, = {{ 2 ;? z: —oco
(3.8.1)
3.8. Approximate Representation of Distributions
267
A system of pairs of polynomials p,(x), q(x) (v = 0,1,2,...) is called a brorthogonal system with the weight w if condition (3.8.1) is fulfilled only at p # v. Any biorthogonal system of pairs of polynomials {p, (zx), q,(x)} may be reduced to a biorthonormal system by dividing the polynomials Pv(£), qv(“) by numbers a,, 3, respectively whose product is equal to
the integral in (3.8.1) at the corresponding v and p = v. It is evident that at every v one of the numbers a,, (3, may be chosen arbitrarily.
In the special case where q,(x) = p,(x) (v = 0,1,2,...) condition (3.8.1) takes the form
/ w(a)p €e)py(#)de = by,
(3.8.2)
In this case the system of polynomials {p,(x)} is orthonormal if it satisfies condition (3.8.2) at all v, and orthogonal if it satisfies the condition (3.8.2) only at uw # v. Any orthogonal system of polynomials {p,(x)} may be normalized by dividing p,(x) by the square root of the integral in (3.8.2) at the corresponding v and p= v. It is evident that the existence of all the moments for the density
w(x) is necessary and sufficient for the existence of all the integrals in (3.8.1) and (3.8.2). It is expedient to use multi-index numeration
of the polynomials
pv(z), q(x) in such a way that the sum of the coordinates
|vy| =
+.--.+ vy, of a multi-index subscript v = [1...v,] be equal to the degree of the polynomials p,(z) and q,(x). Then the number of linearly independent polynomials of a given degree m = |v| will be equal to the number of independent monomials of degree m, 1.e. to Cy,_1
= (r+m-—1)!/[m\(r — 1)!]. It is evident that a multi-index numeration may always be replaced by the usual one. > Let f(x) be the density of some 7-dimensional random vector X for which the moments
of all orders exist.
Let us try to represent
the density
f(z) by an
expansion of the form [oe)
Sy
f(z) = w(x) Vi,
soy
Cy Dy (& ) .
(328.3)
a)
To determine the coefficients C, of expansion (3.8.3) we multiply (3.8.3) termwise by qdrat) and integrate over the whole r-dimensional space R". As a result taking into
268
Chapter 3 Moments,
Characteristic Functions
and Functionals
account (3.8.1) we shall have
iff(e)a(a)de=
D> goon)
/ SS ATTY ee
Leas)
Consequently,
y= f Hle)a(@)de =Bas(X) = a(a),
(3.8.5)
—co
where qv(a) represents a linear combination of the moments of the random variable
X which is obtained by replacing all the monomials ah eee ake in qv(x) by the corresponding moments
Q@/;, ...@,-
Thus all the coefficients C, of expansion (3.8.3)
are expressed in terms of the moments of the random vector X. Here copo(2x) ="
always since by (3.8.1), po(x) and go(x) are mutually inverse constants (the poly-
nomials of zero degree). To construct expansion (3.8.3) it is convenient to choose the density w(x) in such a way that all its moments
corresponding moments
of the first and the second order coincide with the
of the density f(z) of the random vector X.
the polynomials qv(2) of the first and the second degree ( |v | =
Then for all
Vi pape oop
Ye
=a 2) we shall have by virtue of (3.8.1)
cy = :f(2)qv(x)de = iw(x)q,(2)dx
syed!
/ uebiesia4h G)de ws
In consequence of this expansion (3.8.3) takes the form
fe). =0(2) |ty) k=3
eps (a)
(3.8.6)
[v=
In some cases it is convenient to express the coefficients C, in (3.8.6) not in terms of the moments of the random vector X as in (3.8.5) but in terms of its characteristic function. Minding that any moment of the vector X is obtained by the differentiation
of its characteristic function g(A) with respect to the corresponding coordinates of
3.8. Approximate Repfesentation of Distributions
269
the vector 7 followed by equalizing A to zero (Subsection 3.5.3) we may represent
formula (3.8.5) in the form
cy = [qr(0/t0A)9(A)],~5 - 4
Remark.
(3.8.7)
Formula (3.8.6) as well as formula (3.8.3) gives the expansion
of the function f(x)/V w(x) in terms of the unit vectors { Jw(@}pr(«)} of the function space L2(R’), which form a biorthonormal system of the vectors together
with the vectors { w(x) gy (x) i.On the basis of the general theorems of functional
analysis expansion (3.8.3) of the function f(a)/4 / w(«
[
in case
PO) 6 < oo
w(z)
—0oo
m.s.
converges
to f(x)//w(2).
complete in D2(R").
The sequence of functions
{
w(x)py (x) is
The sequences of Hermite functions (Appendix 2) may serve
as examples of complete systems of functions in the corresponding spaces Lo( Py
Formula (3.8.6) determines the orthogonal expansion of the density f(z). Partial sums of the series (3.8.6) may be used to approximate the distribution defined by f(x) with any degree of accuracy also in the case where the random variables X has no moments. It is sufficient in this case to replace the distribution f(x) by the corresponding truncated distribution ip (e) = f(@j\ipte)/ | seae,
(3.8.8)
D
approximating f(x) with sufficient accuracy and then to approximate fp(x) by a partial sum of series (3.8.6). A finite segment of expansion (3.8.6) may be a function which assumes small negative values at some x and nevertheless may give a good approximation to f(z). Retaining in (3.8.6) only the polynomials not higher than N*» degree we obtain the approximate formula for the density f(z): N
f(x) © ft(z) =
w(2)}1+ >> SY) epr(z) | . k=3 |v|=k
(3.8.9)
270
Chapter 3 Moments,
Characteristic Functions
and Functionals
The function f*(x) approximating the density f(x) by virtue of (3.8.5) is completely determined by the moments of a random variable of up to the NV order. In this case the moments of the function f*(z) up to the N* order coincide with the corresponding random variable X.
moments
of the
> Really, multiplying the last equality in (3.8.9) by dp (2) and using (3.8.1) and (3.8.5) we get
/ Pf(e)qu(e)de = ¢, = dua) at
|p| < N.
Thus
the expectations of all the polynomials
(3.8.10) q(X)
of not higher
than Nth degree calculated by means of the approximating function fe (x) coincide
with the expectations of the corresponding polynomials qu(X ) calculated by means of the true density fla):
But the polynomials
q(x)
of any given degree k are
linearly independent and their number coincides with the number of the moments
of
the k'® order. Therefore from the coincidence of the expectations of the polynomials
qv(X) it follows also the coincidence of the moments of the function f* (x) and of the density f(z) up to the N*® order. Denoting the moments of the approximating function f* (x) by oe piece
and Hy sk,
WE May write the obtained result in
the form * Oat peveilbex
=
ON
eben
3
Leon,
LORE
Sy Lr
seed os
at
ky
+---+k,
yo Eo) N
k=3
|v| =k
Vy 4a
(3.8.24)
v;!
Formula (3.8.5) determining the coefficients of this expansion takes the form
C=
/ f(x)G,(x — m)dz = EG,(X —m)=G,(p), —oo
(3.8.25)
274
Chapter
3 Moments,
Characteristic Functions
and Functionals
where G_() is the linear combination of central moments of X obtained as a result of replacement in G,(« — m) of every monomial of the form (X1 - m1)" ...(Xp — m,)"r by the corresponding moment fn, as, Bee
According to the result of Subsection 3.8.1 all the moments (both initial and central) of the function f*(«) approximating the density f(a) up to the N* order coincide with the corresponding moments of the density f(x), and the moments of higher orders of the function f*(z) are expressed in terms of the moments of up to the N“ order from the relations
Grkyet) = Oat.
lz) as
(3.8.26)
where by the asterisk the moments of the function f*(«z) are indicated.
The coefficients c, of H,(a—m)/(™!...v,!) in the Hermite polynomial expansion of the density f(z) are called the quasimoments
random variable X. The number
of the
|v| = 1; +---+ , is called the order
of the quasimoment c,. According to (3.8.25) the quasimoment of the k*> order represents a linear combination of central moments up to the kt order. 3.8.4. Relations between Quastmoments and Semiinvariants
For expressing the quasimoments of a random variable in terms of its semiinvariants we may certainly substitute in formula (3.8.25) the expressions of the centered moments in terms of the semiinvariants. In this case following formulae are used:
Igy
an
Sey eee
eae
eee
Mt eee
Poe Ing)
|
ea
cond tty
(3.8.27)
Rte q.yaar),
RS (3.8.28)
es
(nt othn)/2] (_yyp 1 Oslusgano nde
=
dons ae
Pp
p=2
P
tid ilsllpee Vigo Viuit...+Unp
[(hit--+hp)/2] Hai
be
(3.8.29)
Uae
k=1
OL, Cote p=r
hal ashy! a oa © | B-
3.8. Approximate Répresentation of Distributions
x
DS.
2
ies
7%
Rg
PS
IlSime a ee
ey See
Ned l
gk 04,2) bad frcrith > 4).
(3.8.30)
The integral part of the correspondent number is marked here by square brackets. But in the general case the result is achieved by means of very complicated and cumbersome calculations. Therefore we shall try to express directly the quasimoments c, in terms of the semiinvariants fey > Let X be the r-dimensional random vector with an expectation M, a covari-
ance matrix K and a density f(x)< We put u = tA in formula (2) of Appendix 2 for the generating function of Hermite polynomials Ge (x) corresponding to this ran-
dom vector and replace the variable
ZTby
© — m.
Then taking into account that
(u? _ m?)\ = AT (2 == m), we shall have
exp {iaT(e —m)+ st Ka} = ye,
(x —m),
(3.8.31) where =
the summation
extends
to all nonnegative
values
of 1j,...,V,,
and
V
[14 Se PS le After multiplying this equality by the density Tie) of the random
vector X and integrating with respect to Z we shall get by (3.8.25) g
aN
tN
EX]25((A) pa = A ies se nieae
(3.8.32)
LyA k=3
|v|=
-\T where g()
is the characteristic
function
of X.
Thus
eta
T m+"
KX/2 g(A) as
a function of 7X serves as a generating function for the quasimoments as well as g(A) serves as a generating function for the initial moments, € =e ™g(A) serves as
a generating function for the central moments,
and In g(A) serves as a generating
function for the semiinvariants. Substituting into (3.8.32) the expression of g(A) in terms of the semiinvariants,
g(A)= en 9A) = exp
5m ye ve. 5 ae 5
h=1 |s|=h
BY
HG31663)
and remembering that the semiinvariants of the first and the second order coincide
with the expectations and the elements of the covariance matrix of X respectively we may write
Pp >i) h=3) |s|=
vee,
ie
Sony
cope
(alles, a hh Lasers 1
276
Chapter 3 Moments,
Characteristic Functions
and Functionals
or
3
4h
Deg p=
(iz) as (OR
De de,
h=3 |s|=h
Stl ose
ee tis
=e
k=3 |v| =k
(id)? Pye
...(trp)”" yel
ae
Hence it is directly clear that the quasimoments C, of the third, the fourth and the fifth order coincide with the corresponding semiinvariants Cy = Ky ( |v | 2S, 45 5).
If follows from the fact that the expansion in brackets in the left-hand side begins with the terms of the third degree in consequence of which only the item of the first
degree p = | in the left-hand side may give the terms of the third, the fourth and the fifth degrees.
As regards the terms of the sixth and higher degrees in the left-hand
side for which \v| > 6 they are present in the items corresponding to p < [ \v| Ao i Therefore, after raising the expression in the brackets to the corresponding powers and collecting the terms of the same degrees we obtain
$
Cll /3] 5
V —* = pat SSPp: rea SO VA a 5 ol Yaw
where qj Ze = [qi1---
ir] ah yee
”
“
CRW MCP B® Iv] =6,7,...), (3.8.34) Gil-:---Qpr:
Wp ais = ioage aoe 4h as well as V == [71 «ot Vel
are 7-dimensional multi-index subscripts and the inner sum is extended over all the
values of
1, --- , Qp, lqi| asa,
[14 a: AR
The first term in the right-hand side of formula (3.8.34) corresponding
ldp| > 3, whose sum is equal to the vector V=
to p = | is equal to Ky.
V2),
26,
(P=0,1,...,0),
|
which m.s.
converge at each t € 7}. Similarly all the derivatives of the covariance
function 0? +4 K, ite t') /OtP ot'! (p,q = 0,1, ..., ) will be expressed in terms of the canonical expansions
P+
ike
=
/
Se)
DoDePaG),
e=01,....9),
ail
convergent at all t, t’ € 7}. In the case of a vector argument the operation 0/Ot may
be substituted by any differential operation in a space of the argument @, in
particular, by the operation of the differentiation over any component
of the vec-
tor t. Example
3.9.10.
Find a canonical expansion
of a scalar white noise
(Subsection 3.3.3). Let X(t) be a white noise of the intensity v(t). According to
296
Chapter 3 Moments,
Characteristic Functions
the results of Example 3.9.6 as an orthogonal basis in
and Functionals
Hz we may take any complete
sequence of the functions { Fr (t)} which are orthogonal on a given interval T}:
(fatten
= | tol?
fn(tat
= Dnébnm .-
(I)
For constructing such sequence by means of formulae (3.9.13) and (3.9.14) we may
take any complete sequence of the continuous functions {Pn (t)}. Formula (3.9.16) for the coordinate functions in this case gives Zn (t) = BD V(t) fn (t). The canonical expansion of a white noise will be m.s.
convergent,in such sense that any stochas-
tic integral which contains this white noise will be expressed in terms of the series convergent in mean square:
frotxma==
[OOROd, fQeHe.
Hence in particular it follows that the Wiener process
W(t) at any finite interval
[0, IB ] is expressed in terms of a canonical expansion:
W(t)= YoVaan(t), an(t)=p ffalar,
(III)
where aie (7 )} is a complete sequence of the orthogonal functions at the interval
[0, i ik The random variables V,, are determined as the stochastic integrals of the form
Dy
Vaz ikfn(t)dW(t) .
(IV)
0
So the canonical expansions of the stationary white noise V(t) = W(t) and its co-
variance function K’,(t, t’) = v6(t—t’) at the time interval [0, 7) ]are determined by the following m.s. convergent series:
1 V(t)
=
T,
[o-0) MS,
Vier"
Kel)
=
=
33
eiAn(t-t
(V)
)
where A, = 2an/T); Do = D[Va] = HVE. The canonical expansions of the nonstationary white noise V(t) and its covari-
ance function K, (t, t’) =a)4 (t)4 (t')8(t — t’) have the following form: t
V(t)= a
=
5
ea aie (Oh So Vile n=—0o
ae
/
co
:
.
SE ein(t—t!) n=—oo
(VI)
3.9. Canonical Expansions
297
where A, = 2an/T); Digs D[Va] — aA Exam
p1e
3.9.11. Under the conditions of Example 3.9.3 at
D =
1 and
the time interval (0, T\] the canonical expansion takes place:
Dace peerfret eee nC = sin si | (:= 2) + =| et where D,
=
D [Vex ]=
transcendental equation:
An
tgw7|
+ An); W 1, W2,... are the positive roots of the
=
—2aw/(a?
= w?); An are the eigenvalues of
the integral equation
T;
/ent Example
-
;
7
Wee evel
(I!)
3.9.12. It is not difficult to make sure that for the covariance
function K,(t, t’) = v(t)v(t’)e* |t-#"| at the time interval [0,7] the canonical expansion has the following form:
where An, Wy, Dn are determined in the previous example.
Examp1e
3.9.13. Consider the random function
X with the expectation
Mie = Vie (t) and the covariance function
Kp(t,t') =
Reine
qi(t)qa(t’)
ata
at
ae
(1)
t >t’,
where qi = 41 (t) and qz = qo(t) are such that the relation q2/1 is a monotonous increasing function. Find its canonical expansion at the interval Ty —T,
0) that
Oy z ot
GulT aie Fa\Galt).\ ana Le . qi(T2 — T1)q2(T2) qo(T2 — T1)qi(t) Liles a qo(T> — T1)qi(T2)
oY
The variable € is monotonously increasing function of t and to the interval TT, —T,
< t < T> corresponds the interval 0 < € < 7. We determine the function Q(E) by the equality Q?(E) = qi(t)qe (t). Then the change of the variable t by & gives the
298
Chapter 3 Moments,
Characteristic Functions
opportunity to present ke (tt) in the form HptAt)
and Functionals
=
Q(E\Q(E')e
* |é-¢ |J
Using the results of the previous example we obtain
X(t) = m,(t) + 5 oYni EE GROEN) sy |wno(t) + a aeciti) n=l
where A, = 2a/(a? + w2), I Dger
1
D[Va] = An(n + pvea
go(t)
/a1(T2 —T1)q1(T2)
Oy n@) Vq(T —Tas)
and Wry, are the positive roots of the equation tg nw =
—2aw
(Iv) /(a? = w?),
3.9.4. Inverse Theorem about Canonical Expansions
Theorem 3.9.2 (The inverse Pugachev theorem).
Jf none of the
vectors x, belongs to the closed linear span of the other vectors z,, 1.e. to the subspace X, closed in the weak topology of the space X formed by the sequence {Cu} vey then to canonical expansion (3.9.18) of the covariance operator K, corresponds canonical expansion (3.9.11) of the random
variable X.
> Really, if x,€X,
at none of
then at each V there exists the linear func-
tional f, € F, whose zero subspace contains X, but does not contain Z,.
This
functional may be always chosen in such a way that f,£, = 1 (for this purpose it is sufficient to assume Te = Jy,
where gy is an arbitrary linear functional with zero
subspace which contains X, but not XL,). Thus the condition of Theorem is sufficient for the existence of the sequence of the linear functions ae: € F, satisfying
the condition f, x —
by p- To this sequence corresponds the orthogonal sequence
of the vectors of the space H,.
by canonical expansion i.2)
Keo
=
In fact if the covariance operator Ky, is expressed
(3.9.18) then from fy, €
Fo, fvEp
=
ov follows that
ee
we Dy EMGven)
= D, «xy. Hence it is clear that the elements is of the
N=
space F’, may be replaced by the correspondent classes of the equivalent elements, i.e. by the vectors of the space H, and then uke fi i= Dp SinPy sal DI Sup: On the
basis of Theorem 3.9.1 the orthogonal sequence of the vectors fut C Hy
generates
canonical expansion (3.9.11) of the random variable X with the coordinate vectors
gy.
3.9. Canonical Expansions
299
~~
3.9.5.
Way to Construct Canonical Expansions
To find the systems of the linear functionals < Lent C F, satisfying the condi-
tion of the biorthogonality ful p = vp together with the coordinate vectors 2, of
canonical expansion (3.9.18) we may use the following method which is known from functional analysis. At first we take an arbitrary sequence {Yn } C F, for which at
any 7 the following condition is fulfilled,
$101
pot.
.-.
Yn*1
Sere | cs
ae
apa
Pee
P2%n
---
Pn&n
Pitn
(3.9.19)
|
The mentioned functionals exist always as for any nonzero vector
£ € X in FP, there
exists such a functional ~ that px # 0, and consequently, it may be always chosen in such a way that px =
1. After determining the sequence {Yn} C F, we assume
that V1 St Psy
Wy =
Ayivit:
: +ay
y-1Py-1t
Pr
(v =
2, 3, oa sss (3.9.20)
and find @,, sequentially from the conditions Wy Cy = 0 at
u< V. As a result we
obtain p-1
Ay. = —Pv%1,
| 29 Area
Ap = —Pvl yp —Dandaes
(
=
Nae |
8
)
A=1
and
Wy By Had
a= (atl jl
(3.9.21)
(v Sal eat .) Really, from (3.9.20) with the account that Prly
A we find v-1
ye, = Ay1¥it1+ Prt,
Ply = So avyataey + Yy Zp , AS
pl
Oy
tee SS ovataty ai
yr ee
Specter Esa.(pbate colt
weil ae We
A=A
Putting Dy fy, = 0 at fp < V and excluding the unknown Qj,
...
, @y,y—1 from
the obtained l/ equations we shall have
W124
Caan
a
ee
Wi Ly-1 Wily
0
Bes,
0
Py Xt
0
Py @2
vars
Vy—12y-1-
Pvfy-1
a
.Py-12y
PyLy
irs
anes
pe prky
-
300
Chapter 3 Moments,
Characteristic Functions
and Functionals
We replace here zeros by the correspondent expressions yp\,, and subtract from the (v = 1) column all the previous columns multiplied by @)—1,1,.-.
, @y—1,v-2
correspondingly and then subtract from the (v = ae column all the previous ones multiplied by @y—21,---
, @y—2,v-3 correspondingly and so on, from the second
column the first one multiplied by 0/21. By virtue of (3.9.20) and (3.9.19) we get
Coe REO) ek
Putting here sequentially we find W,2,
y =
at all v.
Oy
H1
O12
PIe
Gite
Prt7
+.»
Pyb2|
City
Poty
©.
Opty
_ 4.
2,3,... and accounting that @121.=
After this we derive formulae
912;
=
1
(3.9.21) from the obtained
equations. Finally, after determining the sequence {vn} we put that [oe
ey Sag
fi Se Be
(3.9.22)
aay,
A=v
and calculate Bu sequentially from the condition f, Cy = Oy i As a result we get p-1
ae (ee apanston (3
Bw =1, Bu =— )Perdren
3.9.6. Canonical Expansion of Random Theorem
3.9.3.
If the random
iy (3.9.23)
Variable in H-space variable X in H-space possesses
such a property that E || X° ||?< co (or what is the same E || X ||? < co) then tt may
be presented by a canonical expansion in powers of
the ergenvectors of tts covariance > Let X
be random
operator a.s and m.s.
convergent.
variable whose phase space is H-space
X.
The space
of the continuous linear functionals dual with X may be identified with X as any
continuous linear functional on X is representable in the form fx = (a f) where f € X. Assuming F’, = X we express a covariance operator of the random variable X by formula
KepeS Ee (pae es Suppose that
| x
RSS co.
(3.9.24)
In this case as it was proved (Section 3.3) the
covariance operator Ky is the positively self-adjoint completely continuous
tor.
opera-
Therefore it has a finite or a countable set of the eigenvalues {An} and the
correspondent orthonormal sequence of the eigenvectors {Pn }:
Kon
= An
Pes
(Ore
or n-
(3.9.25)
3.9. Canonical Expansions
301
,
Here for any vector i € X take place the expansions
f=fot (fi > yv)ev,
Kefo=0,
(3.9.26)
v=1
Kef = fot > o(f, ev)or -
(3.9.27)
ie
Hence it is clear that the range A K,, is the separable H-space with the orthonormal basis of the operator K',. Introducing into Ak, another scalar product
Cho)t teh et en het
Ao
fx)
and supplying A K, by the limits of all fundamental sequences in the metric generated by this scalar product we obtain an. auxiliary-
H-space H,.
In essence the latter
space does not differ from the space H, constructed in Subsection 3.9.2.
Really,
the characteristic property is only in the fact that each element in a new space Hy represents not a class of all the elements X equivalent relatively to Kz, but only one
element of this which belongs to Ax,. The orthonormal basis {Yn} in A K,, formed by the eigenvectors of the operator K, appears to be an orthogonal basis in H, as (gv, Pu)i =
(KeGv, Pp)
= AG ey, Pu) = Ay by . Therefore according to the direct Pugachev theorem the sequence {Pn} determines a canonical expansion of the random variable X
X=me+)
[e.e)
Xe,
(3.9.28)
VK
where X, = (a (2) ) are the uncorrelated random variables with zero expectations and the variances which are equal to the correspondent
eigenvalues
A,.
Formula
(3.9.27) gives here the correspondent canonical expansion of the covariance operator of the random variable X. Expansion (3.9.28) according to the proved theorem weakly m.s.
converges to
X and expansion (3.9.27) weakly converges to Ky f. But according to the theorem about the spectral expansion of completely continuous self-adjoint operator which is
known from functional analysis series (3.9.27) converges strongly in the metric of the
space X. It turns out that in this case canonical expansion (3.9.28) of the random variable
X also strongly a.s.
and m.s.
converges in the metric of the space X.
In
fact almost all the realizations of the centered random variable X° belong to the space Ax,
and this means
that they belong also to the space Hz.
Consequently,
almost all the realizations «° of the random variable X M may be presented by the expansions in powers of the basis {Yn }:
2° =)
(2°, or) yr.
(3.9.29)
302
Chapter 3 Moments, Characteristic Functions and Functionals
Hence it is clear that the sequence of the random variables
=
Xie == STUB
(Sta)
(3.9.30)
vot
a.s.
converges to the random
variable X |
This proves the strong a.s.
gence of canonical expansion (3.9.28). As all random Mig
X”
conver-
are a.s. bounded [o-e)
on the norm
of the random
variable day
| DEE
les 2 |X, |?
viii
Cer)er. 4F- ow(h ever i—3) (3.9.42)
v=.
From formulae (3.9.42) and the convergence of expansions (3.9.35), (3.9.36) in the metric of the space H, follows weak m.s.
convergence of canonical expansion
(3.9.40) of the random variable Y . Finally, from canonical expansion (3.9.40) of the random variable Y we may determine the correspondent canonical expansion of its covariance operator:
Ky ¥y =)
\ywuy("tr). 4
3.9.43
y=)
Theorem 3.9.4. The random variables X and Y in a separable Hz space may be represented by canonical expansions (3.9.36) and (3.9.40) with the same coordinate vectors if the spectrum of the operator A=Kyz'Ky consists only of the countable set of the eigenvalues. > The condition of the proved theorem is fulfilled for instance, when there exists such number Yo
>
0 that the operator
of the projection on the subspace
continuous.
B =
A — YoP where P is the operator
A 4 of the values of the operator A is completely
Really, in this case the spectrum of the operator B consists only of the
eigenvalues from a finite or a countable set {An} to which the orthonormal sequence of the eigenvectors {yn} corresponds:
USOSC
Se 9 A
=m
ty key
(3.9.44)
Here for any vector f € Hz we have
f= DOU bee + (F)vt)],
(3.9.45)
i
BYS Se) or Val
(3.9.46)
3.9. Canonical Expansions
where {yi,} is an orthonormal
307
sequence of the eigenvectors correspondent to zero
eigenvalue of the operator B (the space H, supposed to be separable).
After dis-
posing the eigenvectros WY, and wi, in one sequence {Yn } we notice that all of them serve as eigenvectors of the operator A. In this case the eigenvalues of the operator A equal to the number oy iat Ay respond to the eigenvectors Y,y belonging to the subspace A g of the values of the operator B; the eigenvalue of the operator A equal
to Yo corresponds to the eigenvectors v), belonging to zero subspace VB erator B but not belonging to zero subspace
of the op-
N 4 of the operator A; zero eigenvalue
of the operator A responds to the eigenvectors wi, belonging to N4.
The equation
Ay = Y¥ may be rewritten in the form:
Kyp =7KepConsequently,
the problem
(3.9.47)
of finding the eigenvalues and the eigenvectors of the
operator A is reduced to the calculation of the values 7 at which Eq.
(3.9.47) has
the solutions different from zero and to the definition of the correspondent solutions.
Assuming in (3.9.47)
7 = Yv
# 0, ~ = Yy we may rewrite formula (3.9.38) for the
coordinate vectors U, in the form
ig= apy, Myo,
Remark. number
Yo
>
HY = 12). ee
(3.9.48)
The conditions of Theorem 3.9.4 are fulfilled if there exists such
0 that the operator
the projection on the subspace Ay
B =
A —
oP
where
P is the operator of
of the values of the operator A is completely
continuous,
Example
3.9.16.
Let us consider the random vectors
n-dimensional Euclidean space.
UE > Kef = 0} G
X and Y in the
Let Ky and Ky be their covariance matrices and
oi, ; Kyf = 0}.
If Kz has the rank m
< n then Hy
represents the ™-dimensional factor-space of Example 3.9.1. Eq. (3.9.47) in this case
has M eigenvalues Y1 , ..- , Ym different from zero and the correspondent orthogonal eigenvectors Uj,
...
, Um in H,,. Consequently, the vectors
X and Y are expressed
by the joint canonical expansions: m™m
X=met+ > Wu,
Yom
m
+>) Wu.
Dieaue
(I)
v=
The variances of the random variables V; , ... , Vim are equal to 1, and the variances
of the variables W, , ...
, Wm
are equal to 71, ...,
Ym correspondingly.
The
308
Chapter 3 Moments,
Characteristic Functions
and Functionals
covariance matrices of the random vectors X and Y are expressed by the canonical expansions
m
KS
m
Ne
we
Se ttt, :
Hall
E x am
p1e
(II)
|
3.9.17. Consider two m.s. continuous finite-dimensional vector
random functions X (t) and bal alt t € T;. As F, we assume
linear functionals as in Subsection 3.9.3.
the same set of the
Then we obtain the same
H-space Hz.
If zero subspace of the operator Kg is contained in zero subspace of the operator
Ky and the operator
A = kk Ky in H, has the spectrum which consists only of
the countable set of the eigenvalues then these eigenvalues 7, and the correspondent eigenfunctions pr(t) are determined by the integral equation
i)Ky (t,t’)e(t’)dt! = 7 /K,(t,t')e(t)dt’.
(I)
According to the proved theorem the random functions X (¢) and Y (t) are expressed in a given case by the joint canonical expansions:
X(t) =me(t)+ )>Vw(t),
Y(t)=my(t)+ 5 >Wru(t),
(ID)
where
ere /K,(t,t)p,@)at! = = ‘Ky(t,t)p,(@)at’, and the second expression is valid only at Y,
(AI)
x 0. The variances of the random
variables V, are equal to 1, and the variances of the variables W,, are equal to the correspondent eigenvalues 7. Here the covariance functions of the random functions XE (t) and ie) are expressed by the canonical expansions
Katto
yoany (tag(s
Keg (tet
We yy tty(up
(t)ael
(BV)
p=1
Canonical expansions (II) m.s. converge at each t€ 7}, and expansions (IV) at all be t"€ Wie
Example
3.9.18.
If X(t) and Y(t) are two finite-dimensional vector
random functions which have m.s.
derivatives till the order n inclusively, zero sub-
space of the operator
K
the same
that in Example
FP’, and
H,
A l= KS a AG y in H,
is contained in zero subspace
of the operator
3.9.8) and the spectrum
Ky
(at
of the operator
consists only of the countable set of the eigenvalues 7, then
3.9. Canonical Expansions
309
after determining 7, and the correspondent eigenfunctions (~, (t) from Eq. (I) of Ex-
ample 3.9.17 we get expansions (II) and (III) which may be termwise differentiated n times with respect to t and n times with respect to t’.
Example
3.9.19.
Let X(t) and Y(t) under the conditions of Exam-
ple 3.9.17 represent the scalar random functions of the scalar variable t, t € [0,7; ]
with the exponential covariance functions: = Ae~Altl r=t+tke
Ky, (t, t) =
De-@!"l
Ty
Ti
A iecb lt= ti p(r)dr = yD iCe
0
p(r)dr .
(1)
0
-s
It is easy to verify by direct calculations that the operator space H,
ana Ky (it)
(1) has the form
A =
K a Ky
in the
(a+ Behl" 4 Pr $e Ah) f(r)dr.
(I)
is determined by formula
Af =Kz'K,f =4 A Be
=K-'K,f=
t
Ti
el 2aD 0
In fact for any function f(t) € H, we have Ty
Ti
UG RS A fe? —" Bat
= Keg
0
a
eth g(r)dr, 0
where g(t) is the right-hand side of formula (II). It is clear from (II) that the operator B=A-
yol at Yo =
BA/aD
is completely continuous and as a result of this
the spectrum of the operator A consists only of the countable set of the eigenvalues
17, }2 It is easy to check out by the substitution that the eigenvalues 7, at which Eq. (1) has the solutions different from zero are determined by formula
YW
phPA a? + w? AsaD— ae ats B? +w?2
(v mea »2y eee be),
ry. (IV)
where W,, are the positive roots of the transcendental equation
tgwT}
_
_
(a+ Bw? — af)
(w? 71 ap)? +
(a + B)2w?
)
(V)
310
Chapter
3 Moments,
Characteristic Functions
and Functionals
and the correspondent solutions of Eq. (I) are the orthonormalized eigenfunctions
pr(t) = oy [(a + Bw,(t)+ (1) + w(Ti)6E-—T)],
(VD)
where
hoe
VS
DE +o) (B+ Gren}
W V
Finally,
the
(w+ B)(w2 +08) >”
=
cos W iz
coordinate
(
sInW tact’)
; B)
functions
of the joint
W V
canonical
ga oe:
il
expansion
V
(IV) of the
random
functions X(t) and Y (t) have the form Ti
tt) = Dy
py (T)dr
0
AC?
Seilt—s
= = fe Blt
2 ala
ly (r)dr = Dey
+
ee
Plt).
(IX)
0 Similarly as in Subsection 3.9.3 it is sufficient to find the orthogonal eigenvectors
(,
of the operator At
]sea
which
are not obligatory
normalized.
Then the variances of the random variables V,, in expansion (3.9.37) will be equal to correspondent numbers
D,
= | Ypy [=
~y Key,
the variances
of the ran-
dom variables W,, in expansion (3.9.37) will be equal to the correspondent numbers A=
9
a
le Ky (Y,, and the coordinate vectors are determined by formula
ge ee
a eee
(X)
Here instead of (3.9.39) and (3.9.43) for the covariance operators Kz and Ky we get formulae
Ke)
pyr euy vol
ie ae v=k
ay
(XI)
Problems
Sli
Problems 3.1. Show that a scalar white noise may be obtained by the limit transition from
any random function _X (t) whose covariance function Ky (t, t') decreases sufficiently quickly with increasing |¢ — rn | Namely if ele t’)| < a(t)o(t')e~olt-#
then
the random function Y(t) whose covariance function is determined by the formula
t+t’
Ky(tt) = AK, ( at the limit
h—
+h
t—t’
t+t’
sola
BRA
) noo
00 turns into a white noise. Give the generalization in the case of
the random functions in the finite-dimensional spaces. 3.2. Prove that for the reduction of the scalar random function Y (t) = bo (t)
+ by (t) X(t) to the stationary by linear transformation it is necessary and sufficient
that its covariance function will depend on the difference of the arguments
Kaltt = Ky (t2)/4/D,
GQ)Dy) =, 8= 2).
Give the generalization on the case of the functions in the finite-dimensional and infinite-dimensional spaces. 3.3.
=
Prove
that for the reduction
to the stationary
random
function
Ys)
bo(t) + bi (t)X (y(t)) by linear transformation with the simultaneous change
of an argument it is necessary and sufficient that the ratio € of the partial derivatives of its correlation function may be expressed in the form of the ratio of the values of some function
=
wv) taken with inverse sign at the values of the argument
equal to t and t’ correspondingly:
_ DRylt t/t
€= ——_
=
VO
og gy
OR,(t,t)/av ~~ We)’
Kale
etl Bie.Si en
VD, Dye)
Give the generalization on the case of the random functions in the finite-dimensional
spaces. 3.4. Using the result of Problem 3.3 prove that the random function ‘a Ga with the covariance function
Kyat) Se expla or
Nar
is reduced to the stationary one at p(t) =at+
Get)"[a beet?) bt?.
3.5. Using the result of Problem 3.3 prove that the random function Y (t) with the covariance function
K,(t,t’) = Dexp{u(t +t’) — alt -t’|}
Si
Chapter 3 Moments,
Characteristic Functions
and Functionals
is reduced to the stationary by the transformation X(s) = e_ BY xX (s) is the stationary random
function with the exponential
(0), s = t? where covariance
function
R,(o) = De~*l¢l,¢ = 5 —s!. 3.6.
Let the function K(epey
t, t' € Ty; be the covariance function of a
random process and R(u) be a polynomial with the positive coefficients. Prove that
the function Ky(t,t’)
=
R(K(t,t’)) is also a covariance function of a random
process. 3.7.
Show
that an expectation and the second initial moment
of a random
function
Y(t) sat SE ant Xn(T)X1(7)dr 0
where X (t) = whose
components
[Xy (t) ma
iat
(t) |? is a normally distributed random
have the expectations
process
™] (t) and the covariance and cross co-
variance functions Kj; (ty . ta), are determined by the formulae n
Wig (t) = i Ss anil ma(T)mi(7) + Kai(r, 7) ]d7, 9
(1)
ftl=1 t1
te
ra
Py(tist) = ff S> *anrdyg[ Kui(n1,71)Kpq(t1, 72) oO
0
h,lp,q=1
+Knp (11, 72) Kig(71, 72) + Kng (1, 72) Kip (11, 72)
+Kni(t1, 71) Mp(71)1M4 (72) + Knap (11, T2) m1 (71)Mg (72)
+Kng(t1,; T2)1M1(71) Mp(72) + Kip (71, T2) Mn (71)Mg (72) + Kig(71,; T2)Mn (71) Mp (72) + Kpg(T2,72)mMn(t1)mi(71) Jdridro.
(ID)
3.8. Show that an expectation and a variance of the random function Malis a)
which expresses the relative time during which the random process X (t) exceeds the
given function a(t) at the time interval (to, t),
Y(t,a) = a | ux — a(r) ar,
(1)
are determined by the formulae
ane, t tg @ Mie ry far f fileinae, oA) 46
(II)
Problems t
1 Dy (t) =
rage}
t
co
(@—t)? tH)? ahJanan
sii:
co
/
/ [ fo(@1,
2; 71, T2)
a(71) a(T2)
—fi (x15 71) fi (#1; 72) Jda dae where
f} (a; z), fo(a1 )Cosel, T2) are the one-dimensional
densities of the random process
(III) and two-dimensional
(t).
3.9. It is known that the number of the intersections of the scalar process X (t)
by the given curve a(t) at the time interval (to, t) is determined by formula
Z(ta) = f|X(ry= a(r)|8(X(r) = a()) dr,
(1)
Prove that the moment a@, (7 = 0, 1, 2,...) of the number of the intersections Z(t) is determined by the formula
t a=
t
fan... fdr, to
x fo
0 where f, (x1, Ula
to
fancied lavas
say
Ty eee ted
«dy
(11)
0 ce Cr US
laee
5 ff) is the r-dimensional density of the vector
random function [ X(r) Yq) ies Vin) = X(T). 3.10. It is known that the number of the stationary points of the process X(t)
at the interval (to, t), i.e. the points S where X(s) = 0 is determined by the formula t
vl) = f (XC) Olar
(1)
to
Derive the formula for the moment
a,
(r =
0,1,2,...)
of the number
of the
stationary points. Give the generalization on the case of the vector process. 3.11. It is known that the number of the exceeding by the scalar process X(t)
of the curve a(t) at the time interval (to, t) is determined by the formula
¥(t,a) = /[X(r) — a(r)|1(X(r) — a(7))6(X(r) — a(7))dr.
(1)
314
Chapter 3 Moments,
Characteristic Functions
and Functionals
Prove that an expectation and a variance of the random function Y (t) are calculated by the formulae
my(t) = fdr fnfi(a(r),a(r) +ns7)¢n, Dy(t) ii heb from
(u)
lotn)v dn) + ms at), ate
+2; 71,72) —fi(a(tr), a(71) +15 71) fr (a(T2), @(72)+1025 72) Jdmidne (IIL) where filz, reas we T2) and fo(#1, G5 dias Coo Tt. T2) are the one-dimensional and
two-dimensional densities of the vector process [ X(t) X (t) ‘ie 3.12. Show that under the conditions of Problems 3.9 for the stationary process AC (t) and constant level a(t) = d an approximate expression for the density of the
variable Z(t, a) has the form
f(z) = fol) fila
1+ Dem]
(I)
OF 6 en §(z —m)
(II)
Here Pr (Z) are the orthonormal polynomials with the weight fo(z), po(z)
=e
r
Dia)
tee R=0
C=
Hhf(z)pr(z)dz = aro + So are =e
(Far 2pses Ie)
(11)
k=1
where @1, @2, ..-. , Mp are the moments of Z(t, a). 3.13. Under the conditions of Problem 3.12 restricting ourselves by the moments no higher than the second order show that
Pra |) =
_ (aatay Tag)(L+2on) da?
ave
3
Ogee
Capa
oben [i+ a2 — a, — a? a2
OF
2 = rm], m=0,1,2,.00
Problems
315
3.14. Generalize the results of Problem 3.12 on nonstationary case and show that
the expression for the distribution function of the moment 7' of the first intersection has the form
Ftp]
P(P A, Yr (t). Show that one of its canonical expansions will be the following (Pez!
ROH pe Vee),
(I)
where the uncorrelated random coefficients V, and the coordinate functions Wy (t) are determined from equations r—1
A=V
A= So tree os (22, SS
(II)
p=)
by(t)=pr(t)+
>> apgr(t) (V=1,...,n-1), r=v4+l1
(11)
bn(t) = Yn(t). Here d,, and the variances D), of the random variables V, are expressed in terms of the covariance moments avi
kyi/Dy ,
a
hyp —
k, p of the random variables Ap:
-1
. Yo dyrxGyra Dy Pail
f-D eulthn fly
15 BVI,
(IV)
Problems
317 v-1
DEST
Dike koe
59 lay.|
2,
Dx.
(V)
A=1.
3.20. Under the conditions of Problem 3.20 show that if the scalar centered ran-
dom function X (t) is represented by means of interpolational Lagrange polynomial:
X=
LMGart
p(t) = (t— t1)(t —te)... (t- th),
then in formulae (III)—(V) we shall have SS
AE
i
Gt)
Pa
y |2,
ee
tw
qe
Wat
=D),
mfoperrit)
oh
in a)
3.21.
Let X be a normally distributed random vector,
its covariance matrix.
We
divide the vector X
Mm its expectation, K
into two blocks X;
and X2, X
a
=
[xP xe | , and represent in the corresponding block form its expectation ™
=
[m? mr l , covariance matrix K and the inverse matrix C' = Tae
Kir
Ki
| Ko
Koo
Ke
ei Ona |
se tel . C21
I
C22
()
Show that conditional density of the vector X9 at a given value 21 of the vector xX,
and conditional moments are given by formulae
Wee
fo(z2 |21) =[(20 1
X exp {-3 [uz Coot
=
de
iF ud C22(C yy C211) aE (Cy
Cau)?
Conus
+ (Cyy'C21u1)” C22CzyCam] }= [(2n)"|K|/| Ku ly? 1
X exp {-3 [uz + (Ca
Gag
|C2
=
[us + Cy
=[(2n)" |K|/| Kul?
Co141 |}
318
Chapter 3 Moments,
1
F
Characteristic Functions
and Functionals
2!
i.
io
xX exp {-3 [u3, _ (Ka Kj'u)* | C22 [us _ Ka Kjtu]},
(II)
Kay. = Cy = Koo — Ka Kj} Kin,
(III) (IV)
|Koj| = [Keo — Ka Ky’ Ki2| = IKI /|Kul .
(V)
Maj. = M2 + Ku Ky (a1 — m1),
~
CHAPTER 4 STOCHASTIC INTEGRALS, SPECTRAL AND INTEGRAL CANONICAL REPRESENTATIONS
The main goal of this chapter is to give essentials of stochastic integrals theory and its applications to spectral and integral canonical representations based on stochastic measures theory. Section 4.1 is devoted to theory of m.s. limits. In Section 4.2 theory of m.s. continuity is considered. Special subsection is dedicated to stochastic processes with uncorrelated increments. Theory of m.s. differentiable stochastic functions in finite- and infinite-dimensional spaces is studied in Section 4.3. Sections 4.4 and 4.5 are dedicated to the integration theory over the nonrandom and random measures. Section 4.6 contains the stochastic Ito, Stratonovich and 6-integrals theory. Special attention is paid to the Wiener and Poisson processes and the corresponding measures. Processes with independent increments and the corresponding white noises are outlined. Section 4.7 is devoted to the theory of generalized stochastic functions. Detailed theory of the integral canonical and spectral representations is given in the last Section 4.8. 4.1. Mean
Square Limits
4.1.1. Definitions
In accordance with the general notions of the convergence almost everywhere and in measure in Subsection 2.2.4 two types of the convergence of the random variables were defined: almost sure (a.s.) and in probability. Let us present the necessary definitions for the case of the random variables with the values on B-space.
Let {X,}, X, = z,(w) be an arbitrary set of the random variables with the values in B-space dependent on the parameter r which assumes the values from some set R. In the accordance to the definitions of Subsection 2.2.4 the random variable X, converges almost sure (a.s.) to
the random variable
X = x(w) at r— ro if
P(X, — X) = P({w : «,(w) > «(w)}) = 1. The random variable X, converges in probability
(4.1.1)
to X if at any € > 0
lim P(X; — X ||> €) = Jim P({w : |]ee(w) — 2(v) |]> €}) = 0. (4.1.2)
320
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
For random variables X, with the values in a linear space with a weak topology given by means of the set of the linear functionals F, two types of a weak convergence of the random variables are determined: a.s. and in probability. The random variable X, weakly a.s. converges to X if fX, a.s. converges to fX at any f € F,. The random variable X, weakly converges to X in probability if fX, converges in probability to fX at any f € Fy. Similarly in accordance with the notion of the convergence of the sequence of the functions from L, in p-meah the convergence of the random variables in p-mean is determined. Consider a set of the random
variables {X,}, X,; = t,-(w) with the values in B-space X. that X, converges
in mean
of the order
p or shortly
It is said
in p-mean
to
the random variable X = x(w) at r — ro, if there exist the absolute moments E]||X ||’, E|| X; ||P and
E\|X,-X ||P 30 In particular,
X,
converges
in mean
at
r—oro.
square (m.s.
(4.1.3) converges)
to the
random variable X, X, — X if E||X ||?, E| X; |]? < co and
E\|X,-X|/?+0 at
r—ro.
(4.1.4)
Hence the definition of a weak convergence in p-mean for the ran-
dom variables {X,}, X, = z,(w) with the values in a weak complete topological linear space which a weak topology is determined by means of the set of the linear functionals F, follows. They say that X, weakly converges in mean of the orderp (in p-mean) to the random variable X
if at any f € F, there exist E| fX |?, E| fX, |? and
E| fX,—fX P40
at
r—ro.
(4.1.5)
In particular, at p = 2 we come to the definition of weak convergence in mean square: the random variable X, weakly converges in mean square ( weakly m.s. converges) to the random variable X, X, “> X if at
f € F, there exist E| fX |?, E| fX, |? and
E|fxX,-fX|)?}>0
at
r—oro.
(4.1.6)
From the general theorems of functional analysis follows that any sequence of the random variables convergent in p-mean also converges to the same limit in probability and we may extract from it the subsequence
4.1. Mean,Square Limits
aa
convergent almost sure. Besides that if the sequence of the random variables {X;}, Xz = tz(w), El] Xz ||P < 00 (or El]fXz |? < oo at all f € F.) a.s. converges (weakly converges-or in probability) to the
random variable X = z(w) and is majorized on the norm by the random variable U = u(w) > 0, EU? < oo, ||a,(w)|| < u(w) almost at all w (correspondingly by the random variable Us = uy(w) > 0, EU} < ©, |fax(w) |< us (w) almost at all w) then E]| X ||P < co (correspondingly E| fX |P < oo at all f € F.) and the sequence {X;,} converges (weakly converges) in p-mean to the random variable X. Remark.
The sequence of the random variables convergent in P-mean may be
not a.s. convergent and vice versa the sequence of the random variables convergent in probability or a.s. may be not convérgent in P-mean. Example
4.1.1. Let Xp be a random variable with two possible values
0 and n whose probabilities are equal to | — n-? andn~? correspondingly. We saw in Example 2.2.10 that the sequece TAK
converges to () in probability and almost
sure. But it does not m.s. converge to 0 as at all n
E|X, —0|? =0-(1—-1/n?) +n? -1/n? = 1. Example
4.1.2. Let X, be a random variable with two possible values
0 and 1 whose probabilities are equal to 1 — n-! and n7! respectively.
We saw
in Example 2.2.11 that this sequence converges in probability to 0) but does not a.s.
converges.
But it m.s. converges to () as at
nN—
CO
E|X, —0|?=0-(1—1/n)+1-1/n=1/n—>0. While studying m.s. convergence of the random variables we naturally suppose that there exist the second order moments at all considered random variables. Therefore we restrict ourselves to the random variables which have the operators of the second order moments. Later on it will be necessary to use the following lemma.
Lemma
4.1.1 (Loéve).
Let {X,} and {Y;},
re R, s € S be two
sets of the scalar (tn the general case complex-valued) random vartables. If X, m.s. converges to X at r — ro, and Y,; m.s. converges to Y at $s — 80 then there exists limEX,Y, = EXY independent of the fact how the potnt (r,s) in the product of the spaces Rx S tends to the pont (ro, So).
> For proving it is sufficient to notice that
BX
VAS EX VS EX,Y, ae
DH E(Xn = X)Y,
322
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
|EX,(¥, -Y) |? < B|X, PE|Y,—Y |, E| X, |? < 2E|X |? + 2E| X, —X [’, |BOX Xe X Se MARL, Therefore EX,Y, > EXY if E| X, — X |?
0, E|Y,
—Y |?
0.4
Corollary 4.1.1. If X, = X then EX, ~ EX and X° = Xo. > This follows immediately from the formula
EA)EXpte
prs
\
ASP da DEPome |? hg
Corollary 4.1 2. IfX, —3 X, Y, 5 Y then the cross-covariance moment of the variables X, and Y, converges to the cross-covariance moment of the random variables X and Y independently of the fact how
(r,s) tends to (ro, So). > Really, according to Corollary 4.1.1 xe ——
EX°Y® = EX°Y°. 4
4.1.2. Theorems about m.s.
Seu ve —=;
Y° therefore
Convergence
Let {X,,r € R} be a set of the random variables in a weakly complete topological linear space X whose a weak topology is given by means of the space of the linear functionals F. and [,, be a cross-operator of the second order moment of the variables X, and Xj.
Theorem 4.1.1. Let the space ® of all linear functionals on F, coincides with X then for a weak m.s. convergence f X, to some random variable X in the same space it is necessary and sufficient that at any f € F. would exist the imit lim fI,;f = y(f). This condition implies T,S—-To
that at any f,g € F, there exists the hmit
lim
fl,s9 = flg where T
Tr,STO
ts an operator of the second order moment
of the variable X.
> In fact according to the Loéve lemma it follows from f any f € PF, andatall
X, i
fX that at
f,ge Fy
fUrsg = E(fXr)(gXs) > E(fX)(gX) = fTg. In particular at
g =
fd follows the necessity of the condition of the theorem and
o(f) = FTF. i flrsf= E(FX,)(FXe) > Of) at any f € Fe then E| fX, — fX, |? = B| FX, |? + Bl fXs |?
4.1.
Mean Square Limits
323
—E(fX,)(fXs) — E(fXs)(fX-r) > 0.
(4.1.7)
But the random variables f X; represent the functions fx, (w) of the variable W belonging to H-space D2(Q, S, PF) (Section 3.2), and | Tin \Pe = Bl oe
ea There-
fore from (4.1.7) in consequence of the completeness of H-space it follows the existence of m.s._ a random
limit Us
=
linear functional
im. fX, on F',.
at each f € Fy.
As according to the condition
linear functional Y on FP’, presentable in the form pf then there exists a random f
©
F..
Thus
X;
function
weakly
m.s.
This limit evidently is
X
in the space
converges
proves the sufficiency of the condition.
= X
to some
® =
fx at some
X
any
x € X
such
that U;
=
fx.
random
variable
X
what
The condition of the coincidence of ® and
xX (the reflexibility of the space X relative to F,) is- always fulfilled in important
special case when some
space
the space
Y, and
PF, =
X
itself represents a set of all linear functionals
Ej : Ee
=
ARI)
ES Y}
(ie.
=
in this case the set ® of all linear functions on F', is the space X. sis of Corollaries 4.1.1 and 4.1.2 of the Loéve lemma
X,
aang
X
Ey).
on
Really,
On the baif and only if
EX, +> EX and fKy.g = E(fXp)(gX9) > E(fX°)(gX°) = fKeg at all f,g € FP; (ie. when EX, X, and K,,
converges to K
linear operators mapping
{L ¢ | filgi | x(t1) at allt; € Th. And vice versa if z,(t) — a(t) at all t € 7, then for any functional f of the form (4.1.8) fz, — fx. It is evident that in a given case the set of all linear functionals on F, may be identified with the space of the realizations of the random functions X,(t) as any linear functional on
F, is expressed by formula (4.1.8).
Corollary 4.1.3. For the m.s. convergence of X,(t) to the random function X(t) it is necessary and sufficient the existence of the limit
bm. s(t =, Jim 9 Xft) X40) r,s—To
T,S—To
at each fired t € T; and here also exists the lamit
lima Pog(t, 22) =n (DE
a,
Fe
Hele
r,s—-To
Corollary 4.1.4.
For the m.s.
convergence
of X,(t) to X(t) tt
is necessary and sufficient the convergence of the expectations mz, (t) to m,(t) for allt € T; and the convergence of the cross-covariance functions
Ko(tdi) t6cke(Cd sor, alliytele.
Now let {X;,(t)} be a set of n-dimensional vector random functions, t € 7); F. be a set of linear functionals of the type
N
ee en.
(4.1.9)
kT
correspondent to all natural N, to all n-dimensional vectors f,,..., fn with the complex components and to all t1,...,t~ € T;. Analogously as in the previous case we ensure that a weak convergence in the space of the realizations of the random functions X, represents a component-
wise convergence of the vector functions at all t € T; and the set ® of all linear functionals on F, may be identified with the space X of the
realizations of the random functions X,(t).
Corollary 4.1.5. For the m.s. convergence X,(t) to some vector random function X(t) it 1s necessary and sufficient the existence of the lamit
lim
T,S—To
*trt'-s(,2t) =
im
T,8S—1To
tre
X(t) XAt)
et 675.
4.1,
Mean Square Limits
325
Here there exist the limit
Li
PH (pth) SD)
X(t) iS D819 5 wtj tS Tye
r,s—-To
Corollary 4.1.6. For the m.s. convergence X,(t) to the vector random function X(t) the convergence of the expectations m;z,(t) to mz(t) and the covariance and cross-covariance functions K,s(t,t') to the covariance function K,(t,t'), t, t' € T; is necessary and sufficient. Example expectations
Be Dye
and
4.1.3. Consider a set of the random functions X, (t) with zero with
covariance
and
cross-covariance
functions
K,, th t )
abt eh Dey= Wes] (Gol Mi haKre(tt)= Ker ltt).
But according to Theorem
4.1.5 X, (t) m.s.
X (t) if and only if there exists the limit
converges to some random function a =
lima,.
Here mz (t) ="
()Pand.
K,(t,t')= a/2e7*l*- #1, Example
4.1.4. Under the conditions of the previous example if lima,
= 00 then X, (t) m.s. converges to none of the random functions. But as we shall see in Section 4.7, X; (t) m.s.
converges to the generalized random function X(t)
which represents a white noise (Subsection 3.2.3).
4.1.3. Theorems about m.s. Space
Convergence of Random Function in Linear
Now we pass to the random functions with the values in the arbitrary linear space L with a weak topology given by means of the set of the linear functionals A. As usual we suppose that A represents a linear space and it is sufficiently rich for the fact that it will follow / = 0 from Al = 0 at all A € A. The cross-second moment I,,(t,t’) of the random functions X,(t) and X,(t’) in this case is a cross-operator of the second order moment of their values at any t,t’ € 7;. In the same way their
cross-covariance function K,;(t,t’) represents a covariance-operator of their values at any t,t’ € 7). Example
4.1.5. A scalar or a finite-dimensional vector random function
of two variables Z(s, t), s € S,t € 7; may be considered as a random function of the variable t with the values in the space L of the scalar or the finite-dimensional
vector functions of the variable s € S. As a set of the linear functionals on L we may assume a set of all the functionals of the type Al = ict Ak I(s% .: It is expedient to determine the set F', of the linear functionals on the space of the realizations of such
random functions as the set of all functionals of the form (4.1.8) where f,..., fn
326
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
are the arbitrary functions from A; f,,...,
fy € A. We are frequently deal with,
for instance, the product of the form g(s, t) X(t) or G(s, t) X(t) under the sign
of the integral over t where g(s, t) is a determinate function, and G(s, t) is a random function of the variables 5, t. In this case it is expedient to consider Z(s, t)
= g(s,t)X(t) or Z(s,t) = G(s,t)X(t) as a random function of the variable t with the values in the space of the functions of the variable s.
From Corollaries 4.1.3 and 4.1.4 about the convergence of the scalar random functions we have the following statement. \
Corollary 4.1.7. The sequence X,(t) weakly m.s. converges to some random function X(t) with the values in L if and only if at any A EA there exists the limit
lim
AT,,(¢,t)A =
for each fixed t € T,. the limit
lim AL, s(t,’
r,s
lim
E[AX,(t)][AX;(t)]
Under this condition for all X,X' € A there exists
= E[AX(t)][VX@)] = AT (et),
Corollary 4.1.8.
For a-weak m.s.
tT.
convergence of X,(t) to X(t)
there are necessary and sufficient a weak convergence of the expectations
Mz,(t) to mz(t) and at each A € A the existence of the limit
lim AK;s(t,t)A =
r,S—To
4.2. Mean
lim E[AXP(t)][AX2()].
T,S—To
Square Continuity of Random
Function
4.2.1. Definitions
The random function X(t) with the values in B-space defined on the topological space 7} is called continuous in mean square (m.s. continuous) at the point t € T; if at any € > 0 there exists such vicinity V;(e) of the point ¢ that
E\| X(t’) — X(t) ||?
The necessity of the condition follows directly from the Loéve lemma:
from
X(t!) = X(t) follows T',(t’, t”) = EX(t') X(t”) > E| X(t) |? = Te (t,t) at t’,t'’ —
t.
For proving the sufficiency of the condition we notice that from
[Uae .t)— y(t) |< 2/3 at allt” © V4(e) we have
E|X(¢)—X(@)t=T.(t,1)—Ts(t,t)—
SA
elea) Leila teat cae
Hence by virtue of the arbitrariness
€ > 0 it follows m.s.
continuity X (t) at the
point t. If X(t) is m.s. continuous on 7} then from the Love lemma it follows that
ee}
—
Tz (t1,t2) at t’ >
th, t!” —
to, ie. the continuity Tre
| at all
points of TJ, x 7. From the relation
E| X(t’) — X(t) |? =| me(t’) — me(t) |?+ B| XO) — X°(t) |? and just proved statement we conclude that for m.s.
continuity of the random func-
tions X (t) at the point t (on 7}) the continuity of its expectation at the point t (on T}) and the continuity of its covariance function K; (t, t') at the point (& t) of
the product of the spaces 7, X J}, then K,(t, t’) is continuous on 7} X JT} are necessary and sufficient.
to and with its increment
on the interval [t,to) taken with the
inverse sign if t < to. Later on we shall consider only such processes with the uncorrelated increments. Any process with the uncorrelated increments X(t) is reduced to such a process. Really, the random process Y(t) = X(t)—X¢, represents a process with the uncorrelated increments which possesses the required property Y;, = 0. Thus if X;, = 0 then the value X; of the process X(t) at any instant ¢ is uncorrelated with its future increments on the intervals which follow the instant to and
4.2. Mean Square Continuity of Random
Function
~
329
with its previous increments on the intervals which precede the instant
bo: BXP(XP, XP)
= 0 atit' to orta< te By Theorem 4.1.2 of the m.s. —
©o.
of the Ito integral Y.
convergence the limit
if and only if there exists the limit HY, ved independent m
(4.6.28)
-
y =
l.i.m.Yp
exists
of the way in which n,
By this theorem the necessity of the condition follows immediately from
the equality
Bly, =-y E|xaMy| sa) - 4] 2
To derive this equality we note that
B Wal =So ExW,) KEM) [we?) — WE») | Ife
jes
x |wl”) — xi") ] and by independence of wit) — we, ) on sate )), xt *) and wt)
= wt?)
at any k > l,
BX (HX) (WEP) -WER» ][wel — war) | = Bx) x@@) [wWe™)-wa] B[wal”) -WE] =o
B|x@@of |wee)- wey],
360
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
(nm) _ rcp). (n) |? # |wee?)-wee,) =6|x¢e®,)|
(n)
Further, by independence of W(t.)
B|waP)|
fad)
on W (t;, )
|”
oye (adores gp oe) yielding
=2|we | +2 |wEr)- wary] 2
A
2
Consequently,
B\we)- we) =2 wey] -2 |wE»|
2
EP RG) 2 > To prove the sufficiency of the condition we unite two partitions P,, Py, of
the interval (a, b}. Let P be the resulting partition, P to
=a,tg=
: (a, b] =
Q Dr, (ts-1 ; cae
b where each point ts coincides with some 4”? or with some aia
Then we obtain
Q ea
X(r{) [k(ts) — k(ts-1)]
ator lacimdgian tmfieSotyly at A(t, Hentai enhl By virtue of the m.s.
a ree, cei
continuity of X (#), lim EY,, Yn coincides with the limit
lim wae DP atts)=kta]
n,m—oo
b
= [EXP ak(o.« a
The definition of the It6 stochastic integral is easily extended to the
case of a matrix process X(t) and a vector process W(t). Theorem 4.6.7. The vector It6 integral (4.6.27) exists if and only af all the scalar Ito integrals involved in (4.6.27) exist. The necessary and sufficient condition of the existence of the vector integral (4.6.27) is the existence of the Riemann-Stteltjes integral b
Hf,EX(t)dk(t)X(t)* = BYY*, a
4.6. Stochastic Integral ~
361
which ts equal to the covariance matrix of the vector Ité integral Y. The existence of the integrals b
[ BX. (t) Xn (t)dkjn(t) a
is an immediate consequence
of the existence of integral (4.6.28) for all
the elements of the matrix function X(t).
Let Y(t) be a random process determined by the formula 7 ya
Y(t) = Y¥(to)+/Xi(r)dr + /X(r)dW(r),
(4.6.29)
where the first integral represents a m.s. integral of a random function
X(t), and the second integral represents the Ito integral of a random function X2(t) with respect to a process with independent increments
W(t). The random functions X;(t) and X(t) are supposed satisfying the conditions under which both integrals exist at any ¢t in some interval
T.
Let X(t) be a random function.
The Ito integral of the random
function X(t) with respect to the process Y(t) in the interval defined as the sum of the integrals
(a,)] is
/X(r)d¥(ry= fx x(x eigde /X(r)Xo(r)dW(r),
(4.6.30)
where the first integral represents a m.s. integral and the second integral represents the Ito integral. Here, certainly, the random functions
X(t)X1(t) and X(t)X2(t) must satisfy the conditions under which both integrals in the right-hand side of formula (4.6.30) exist. 4.6.8. Stochastic Integrals over Wiener and Poisson Measures
As it is known (Subsection 4.6.5) the stochastic measure of the Wiener process W(t) is defined by Z((t1,t2]) = W(t2) — W(t). So the It6 stochastic integral may be calculated by formula (4.6.27) because
condition (4.6.28) is fulfilled.
362
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
Example
4.6.3. Find the variance of the stochastic integral
= /X(r)dW(r),
where X(t) is a random function with the expectation Mz (t) = ct and the covariance function Ky (ty ; ta) = Deb(titt2)—o |ti-tal , and W(t) is a standard Wiener process. According to formula (4.6.28) and taking into account that v(t) = ‘
E'|X(t)|? = |ms(t)|’ + Dz (t) = c7t? + De? | 4B
TP
Dive ee [ra - Df rar = eT? [3+ D(e7#? — 1)/2je. 0
Let P°(A, A) be a centered Poisson measure, Y(t, u) be a real m.s. continuous random function of two variables ¢t and u with second order moment independent of P((t;,t2],A) at any t < t € Ri = R2\{0}, A € Bo. Let {P,} be a sequence of partitions interval (a, b] and of set 0 < |u| < R,
: (a,0] = U I, }, M20,
=5,
scalar finite < to, of the
(4.6.31)
{u:0< |u| < Ry= Ua”, A\”) €Bo, fe) such that max(t)
2
Me
1) 70, max
sup
|u’—u|
~0asn
oo.
u weal)
M.s. limit 7” of the sequence of integral sums
Sa. Ue) tena) Poe vant =i
Mbeya Rate tA) 9(4.6539)
ysl
if it exists, is called a stochastic integral of the random function Y(t, uv) with respect to the Poisson measure P(A, A) extended over the interval
(a,b] and the set {u : 0< |u| < Ry}, b
fe =)
/
a 0f elt, Maple, d)*a9().
(4.28.39)
Besides we notice that condition (4.8.35) may be written in the equivalent form N
Df volt dd)ap(r, A)? = 16(t — 7),
(4.8.40)
where J is the unit matrix. In particular, when A is a finite-dimensional space and the measures
Hp(t, B) and o,(B) are absolutely continuous relatively to the Lebesgue measure, formulae (4.8.38), (4.8.40) and (4.8.34) may be presented in the form 1 ot A) say |Katt EX Ja,(t’, r)dt’ Nae. ,
(4.8.41)
N
> fatt A)ap(7, A)?dd = 16(t — 7),
(4.8.42)
i)as
(4.8.43)
ee ees
where vp(A) and a,(t, A) are the Radon—Nikodym derivatives of the measures 0,(B) and p,(t,B) with respect to the Lebesgue measure. The
386
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
stochastic measures lated white noises
Z, will be in this case generated by the noncorre-
V(A) = iixc r)T X°(t)dt ,
(4.8.44)
whose intensities are equal to the correspondent functions v,(A). Integral canonical representation (4.8.36) of the random function X(t) will take the form
. N
X(t) = melt) +)» |Vo(A)aplt, 4A, _
(4.8.45)
and the correspondent representation of the covariance function (4.8.39) will be N
K,(t,t') =~ /Yp(d)zp(t, Nxp(t’, AAA.
(4.8.46)
p=l Thus for constructing an integral canonical representation of the
random function X(t) it is sufficient to find the functions «p(t, A), ap(t, )
and v,(A) (p= 1,..., N) satisfying conditions (4.8.41)-(4.8.43). Here the random function V,(A) determined by formula (4.8.44) will be uncorrelated white noises with the intensities vp(A) and the random function X(t) will be expressed by the integral canonical representation (4.8.45). Example
4.8.1. Let the random function NS
X(t) is defined by the
linear differential equation 1 (t)X + do(t)X = V with the white noise V of unit intensity, and X (to) =
0 and is equal to zero at t =
to. While integrating this
equation we find
X(t) = a(6) iSoy no =e {-|. ar}. ay(T)q1(7) a;(T) 0
Formula (I) gives an integral canonical representation of the random function X at
time interval tg
< t < to + 7}, T; > 0. The coordinate functions of this integral
canonical representation are determined by formula
heges oti venscull) seal ooh
(II)
4.8. Integral Canonical and Spectral Representations
387
After rewriting formula (1) in the form totTi
vine / [ay(A)6"(A— t)+a(A)6(A—4)] X(t)at
(HIN)
to
and comparing (III) with formula to+Ti
V(\) = 4 a(t, \)X°(t)dt , to
(Iv)
7 ea
we see that in a given case a(t, A) = ay(A)6"(A — t) b ao(A)d(A ~— ye
4.8.4. Integral Canonical Representations and Shaping Filters
In applications the pairs of the matrix functions z(t, A), a(t, A) satisfying conditions (4.8.42) and (4.8.43) are often known. So, for instance, the weighting functions of two any reciprocal systems with N inputs and N outputs w(t, 7) and w(t, 7) always satisfy the conditions
(Subsection 1.2.4)
/sc(6, Meta. eld Meals
ea
‘ianette aNee TOR Dy,
(4.8.47) (4.8.48)
Comparing these formulae with (4.8.42) and (4.8.43) we see that if we determine the matrices-columns z,(t, A) and ap(t,A) (p=1,..., N) as
the correspondent columns of the matrices x(t, A) = w(t, A) and a(t, \) = w~(A,t)? then conditions (4.8.42) and (4.8.43) will be satisfied. Thus for finding an integral canonical representation of a random function tt is sufficient to choose among all possible pairs of the weighting functions of the reciprocal systems such a patr which satisfies condition (4.8.41) at some functions vp(X). In the general case this problem remains still rather difficult and not always has the solution. We notice that the construction of an integral canonical representation of a random function is equivalent to finding such a linear system which transforms the vector random function X(t) into the vector white noise V(t) of the same dimension with uncorrelated components. The inverse system will form the random function X(t) from the white noise
388
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
V(t) in consequence of which it is usually called a shaping filter of the random function X(t). Thus a mathematical problem of finding an integral canonical representation of a random function ts equivalent to the problem of finding a shaping filter ofa given random function. For many standard random functions (for instance, stationary) the problem of constructing the integral canonical representations is sufficiently simple (see Sections 6.4 and 6.5).
4.8.5. ables
Joint Integral Canonical Representation
of Two Random
Vari-
We return to the problem of Subsection 3.4.7 and consider two random variables X and Y with one and the same phase space and the covariance operators K, and Ky. Suppose that zero subspace No of the operator K, is fully contained in zero subspace of the operator Ky. It was shown in Subsection 3.4.7 that in this case there exists the operator A = K;!Ky which represents a positive self-adjoint operator on Hy. As it is known from functional analysis each positive self-adjoint operator may be presented by a spectral representation co
A = fre(aa),
(4.8.49)
where €(B) is a representation of the unit on Borel o-algebra B of the
real axis €(B) = 0 for any Borel set B of a negative semiaxis by virtue of the positiveness of A generated by the operator A. Let exists such a sequence of the vectors {y,} C H, that the set of the linear combinations of the vectors €(B)y, correspondent to all B € Bandy =1,2,..., is
dense in H, and the vectors €(B)y, and €(B’)y, are orthogonal for ue #v and any B, B’ € B. We choose such functions v,(A) (v = 1,2...) that for any vector f € H, the equality has the place
padtacleh pop ecregeruen
= ON
Then the random variable X and its covariance operator pressed in terms of the integral canonical representations
(4.8.50) Kz are ex-
X =m, + > f wana),
(4.8.51)
Kes fay [0)] oy (dd).
(4.8.52)
4.8. Integral Canonical and Spectral Representations
389
Here Z,(B) = €(B)y,X° are the uncorrelated stochastic measures with the covariance functions o,(B) = (€(B)y,Ww) =|| E(B) ||? and
ways px,
(4.8.53)
Theorem 4.8.3. Let X and Y be two random varaibles with one and the same phase space and covariance operators Kz and Ky. Let the operator A = Kz!Ay exists and defines the canonical representations
(4.8.51), (4.8.52). Then the random variable Y and its covariance operator Ky are expressed in terms of the integral canonical representatios
Y =m, + git ,
(4.8.54)
K, ==>[mo
(4.8.55)
Here U,(B) = E(B)yY°
) |Aov(ad).
are the uncorrelated stochastic measures with
the covariance functions
7,(B) = (AE(B)w,wv)= /avery
(4.8.56)
B
Formulae (4.8.51-(4.8.56) define the joint integral canonical representation of two random variables X and Y.
> Really, as EU,U,(B’) = E(B) KyE(B’Yn = E(B)WyKr AE(B')7p =
(Bye,
AE(B')y,); and any self-adjoint operator is permutable with its ex-
pansion of the unit then
BU, (B)Tn(BY = (E(B), Arn) = fMECBB') > Em) = f ME@) 10)=5. | Aov(ad). BB!
BBE
Hence it follows that U, (B) = E(B) py Y © are the uncorrelated stochastic measures with the covariance functions T;, (B). Further as for any f € Hy 2
E \fy°- »alfv,(A)U, (dd)
390
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
=(ar-ayo feee@rns-Y [focrelayn )= 0 co
by virtue of the fact that f =
SS ‘ifey (A)E(dA) Ww then the random variable Y v=1
is expressed by formula (4.8.54) . Finally from the direct theorem about the integral canonical representations it follows that
Ke S [0 [= O) (ad). 4
Remark1.
In particular if there exists such vector
Y € Hy that the set of
the linear combinations of the vectors E(B)y correspondent to all Borel sets B of
the real axis is dense in H
then sums (4.8.51), (4.8.52), (4.8.54) and (4.8.55) contain
one item and these formulae take the form
X= og + froz@), ie fo |x) |o(da), (4.8.57) Y =m, + froua@), oe fo [e) |Ae(ar) (4.8.58) Here Z(B)
= E(B)yX°®, U(B) = E(B)yY°
are the stochastic measures with
the covariance functions o(B), and
AES ieie
(4.8.59)
B
By virtue of (4.8.53) the function v(A) is equal to
dZ v(A)
=
BX*——
dU
(A) = EY°—()). dr
(4.8.60)
In this case when the spectrum of the operator A consists only of the countable
set {7,} of the eigenvalues 7, we have E(B) =
ya (., Pv) Pv where (, is the WEB
eigenvector correspondent to the eigenvalue 7,
E(B)py = 3
py
at
v€& B;
at EB
and the set of the vectors € (B Voy is dense in H,. Integral canonical representations
(4.8.51), (4.8.52), (4.8.54) and (4.8.55) of the random variables
X and Y and their
4.8. Integral Canonical and Spectral Representations
covariance
operators
turn into the canonical
= E(B)p, X°, W.= Remark
expansions
391
(Subsection 3.4.7) at V,
E(Byo,¥*, any Borel B, y, € B and u, = v(yw).
2. Let us consider the self-adjoint operator
A=
Aa
Ky which is
determined as an operator of the equation
/Read at a given function f(t) € Da
representation
(4.8.49).
/Ky(t, r)f(r)dr
C Hg. We know that A is expressed by a spectral
If there exists such a sequence
of the functions
{Vv (t)}
C H, that the set of the linear combinations of their projections E(B) py at all Borel sets B of the real axis is dense in Hy and E(B)y,
and. ECB
Ys are orthogonal at
Ll a y and any B, B’ € B then according to the proved Theorem 4.8.3 the random
functions X (t) and Y (t) are jointly expressed in terms of the integral canonical representations
X(t) =m, (t) + byfr, Z,(ad), be
¥(t) = my(t) + > fvv(t, UL(dd).
(4.8.61)
vel
And the following relations take place:
Katt) =) A) Bitrate AGA GA) ,
i,
(4.8.62)
Ky (tt) = 50) Ort)
vy (C A) Aos (a),
Hemet
where
re
ae w(t.) = BX) soe S20)
(v= 1,2,...).
(4.8.63)
In the special case when there exists the unique function y(t) for which a set of the linear combinations of the functions E(B)y correspondent to all Borel sets B of
the real axis is dense in H, the sums in (4.8.61) and (4.8.62) contain only one term in consequence of which formulae (4.8.61) and (4.8.62) have the form
X(t) = malt) + fo(t,2)2(0), Y(t) = my(t) + fv(t, )U(@X), (4.8.64) K,(t,t’) = /v(t, A)v(t’, A)"o(dd), Ky(t,t) = iv(t, A)v, (t’, A)"Ao(dd), (4.8.65)
392
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
where
Te
Bx.
Zy.
(4.8.66)
4.8.6. Spectral and Integral Canonical Representations of Scalar Random Functions Theorem 4.8.4. The spectral representation of a scalar m.s. continuous stationary random function with vector argument is one of its possible integral canonical representations. > Let X (t) be a m.s.
continuous
scalar stationary random
function of the
n-dimensional vector t, t € R”; kee) be its covariance functions;
X be a linear
space of all the functions of the variable t; F’, be a set of the linear functionals N
{eS
yy Tea UA) correspondent to all NV, f;,..., fn, ti, ..., tn. In this p=l
case H, represents the range of an isometric mapping of a subspace in the H-space
D4 (Cres pe a) formed by the compliteness set of all functions (A) = ip So eid" te : As the space D2(R”, B”, co) is separable then Hz is separable too. We determine in H,, the expansion of unit E€ by formula
E(B)f = wae [Ota
fsje
rar
B
+7 ‘dd. ae1 | ta )e ear
8. (4.8.67)
= —— | 1p(A)p(A)e™
We assume Y = 6(t). Then by virtue of (4.8.67) we have
chit | 190) €(B)y(t) ub = mate aay |Brine ta) dN eds= fo aa
en atheiA7t +e
B
As a set of the simple functions is dense in D2(R”, ‘igs co) so the set of linear combinations > GL (Bi) ; ¢; € Cis also dense in Hy. Consequently, for obtaining a
an integral canonical representation of a random function we may restrict ourselves to one expansion of unit € and one vector 7. Formula
Z(B) == €(B)yX°
pee
ae: = Ca [x 0 (dt [tare
=
\"tdd
(4.8.68)
4.8. Integral Canonical and Spectral Representations
393
determines a stochastic measure with the covariance function o( BB’), and for any f € Hg we have
(fF, E(B) = [eQ)taQ)o(ar) = fotar fsme
at.
B -\T
By virtue of (4.8.9) u(A) = vlf(t)e
-\T
‘dt, and a(X) = rt
Consequently, the
random function X (t) is determined by the integral canonical representation of the
form (4.8.20):
X(o=imyg + fe Z(a0),
(4.8.69)
Integral canonical representation (4.8,21) of a covariance operator in this case gives the integral representation of the covariance function
kale ) (the known
Bochner
theorem)
ky(ty—ty) = [erersa(aa), d Formula (4.8.69) gives a spectral representation
(4.8.70) of the stationary
function X(t) with vector argument. According to this fact the measure o in (4.8.70) is called a spectral measure of the stationary random function X(t). In particular, when the part of the measure o which is singular relatively to Lebesgue measure in R” is concentrated on the countable set of the points we may determine a spectral density of the
stationary random function X(t):
se(A) =
"S,(A) _ 8"a((—00,A) Ory NES = GAGA OAL
(4.8.71)
which may contain a linear combination of the 6-functions. The stationary random functions encountered in the problems of practice have always the spectral densities in this sense. Then spectral expansion
(4.8.69) may be rewritten in the form
XO) = ms + fvaye** an, where
(4.8.72)
V(A) is a white noise whose intensity is equal to the spectral
density s;(A) of the random function X(t). As the spectral density S_(A) is integrable (by the finiteness of the spectral measure o) then it is expressed by the Fourier integral
Se(A) = oon nr far fsa (meu -\? dp.
394
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
But according to (4.8.70) we have
[eese(wydy = /eH"? (dys) = ke(r). Consequently, the spectral density of a scalar stationary random function with vector argument is expressed in terms of its covariance function by formula ie = oa (onyr | ie(r)e-?” . dr glee:
Example
4.8.2.
Find an approximate canonical expansion of a scalar
stationary random function of the scalar argument t, spectral density Sz (A) at the time interval interval of the frequencies.
X =
We partition this interval into 2N + 1 equal segments We enumerate the segments of the real
axis obtained by such a way by the integers from —N
segment.
X(t) possessing the
|¢| —aclneinet (-Q, Q) be an arbitrary
and denote the length of each of them by 2.
mean of the ®
AOir (4.8.73)
‘
to N and denote by A, the
Substituting the integral in the infinite limits in integral
canonical expansion (4.8.72) by the integral in the limits (-Q, Q) we present the obtained integral in the form of the sum
of 2NV + 1 integrals extended
over the
segment 2@ length. In each integral instead of the multiplier e’*'tAt we substitute its value in the middle of the interval etrvt Then we obtain the random function N
Xw(t) = mz + yap Vee
(I)
where Av+ta
ee
sf V(A)dA (v = 0,+1,42,..., +N).
(11)
Ap-—a
The variances D,, of the random variables V,, by virtue of the spectral representation of the variance
X are equal to Av+a
Dos
i Sa(A dA @ = OFEL SEN’); Av-a
As V(A) is a white noise and the integration intervals in (I) at the correspondent values of the index Y do not overlap then the random variables V, are uncorrelated.
We estimate the degree of the proximity of X y (t) to the function X(t). We have
x()-Xw(=
-
[ee)
f VO) art
[VQ)N)e* dr
—co
a
4.8. Integral Canonical and Spectral Representations N
395
Av+ta
+),
/ V(A) (ec? — ef**) dd.
aN
(111)
mit
As the expectation of the difference X — Xy is equal to zero and the isolated items
in the right-hand side of (III) are uncorrelated then the variance of the difference X — Xj equal to the expectation of the modulus square of this difference is equal to
the sum of the variances of the integrals entering into (III). Each of these integrals represents a random function expressed by integral canonical representation (III). Therefore we may use formula (II). Consequently, —2
fore)
E |\X(t) —Xn(t)|? = / so(A)d+ f sa(X)d —0oo
N
2
Avpta
4 se
/ $z(A) te woe
bi
(IV)
geet 5 ars
Supposing that the interval
|t| < T is fixed we choose the frequency
{2 and the
number @ in such a way that the first two integrals in (IV) would be smaller than the arbitrarily given number €: -2
co
/ s2(A)dd 1 we get from (V)-(VII)
pw(Bn) < c2-"-)/(1 — 278+!) vn,
(VIII)
uw(N) < pw(Bn) < 277-9),
(IX)
where c’ = e/(1 = dap 1), As inequality (IX) is valid for all n then btw (N) =+(). -_
Leta
ee)
=
-
_
U By.
_
Any function x(t) € N belongs to set B,, and conse-
n=1
quently, to all sets Ap, Die iso.
‘calls 4h Pat(m
12" y= atm2-?) (Al
(X)
for all sufficiently large p. Let s, s' € S be arbitrary binary-rational numbers, qd be
the least natural number satisfying the condition | s'—s
I< 2-4+1.
In this case
there exists a point $9 = m27~?1 between the points s and s’ for some ™m and i
Sal chemi epeeah
(XI)
p=
where the numbers point
Sg
to
the
k; are equal to zero or unit. point
S
sequentially
by
If we realize the transfer from the
the
steps
of the
length
boot
§
; kids i" then due to (X) we get
oe (44
Wieiayke aster (a-P)a < AQ- Beswie p=1
|2(s') -2(8) Ik
=
0 at s’ > 5, s, s’ € S C (0, a]. This proves that all the functions z(t) es
excluding the set NV of zero measure [hw
are continuous on the set S of binary-rational numbers for any finite interval (0, a]. Now let us take an arbitrary t € (0, a) and the sequence
{3,.} of binary-rational
numbers converging to t and |Sp — t I< 2~?. Define the sets
Ap = {2(t') :| a(t) — a(sp) |= A2-P*}, lee)
co
Bris "| eAs
Pao
p=n
fae
Poy)
N=
Analogously we prove that pw (N) =
0 and that the sequence
A2Se))
con-
verges to “(t) beyond the set N.Thus at any fixed t € (0, a) almost all relative to the measure
for any @. Alor
{yw
functions from
the space
XT
are continuous
at the point t
This result make possible to extend the measure [yw on the 0-algebra {B
“DS
AC, CG e€ i
of the subspace C’ C XT of continuous func-
tions from the space of all functions NG putting pw (B) = Lw (A), Be
AC:
As it is known any function from the space XT? without changing the values
of measure on rectangle sets from the space X T and also on all sets of an induced o-algebra may be replaced by some so called separable function. almostly defined for all t by its values on some
Such a function is
countable set called a separant.
our case separant is the set of all binary-rational numbers
In
and separable function is
the set of all continuous functions. Notice that the Lebesgue extention of the measure [/w initially defined on the
sets of the rectangles of all functions X F may be determined on the space C’ of all continuous functions.
4.12.
Prove that the 0-algebra A‘TC in Problem 4.11 coincides with the o-
algebra C of the space C' = C’ (0,oo) of continuous functions with supnorm induced by the set of all balls.
4.13. Prove that almost all relative to the Wiener measure Jw functions of the space Fe
pig are equal to zero at
ft=
0.
4.14. Let us consider the function f(x) — Or (s)x?(t1) + ~2(s)x?(t2) where Y1 (s), p2(s) are the continuous functions on the bounded closed interval S, i.e. the elements of the B-space C(S), and a(t) is a scalar function belonging to the space
XT? with the o-algebra A
Oe
(0, 00) of Problem 4.8. Prove that the function
f(z) maps the measurable space (xt F A’) into the separable B-space C(S).
Solution.
This function is (Ar, B)-measurable as the inverse image of
any ball of the space C(S)
f~'(Sr()) = {2(t) : Y(s) — r < g1(s) 2? (1) + G2(s) 2? (te)
Problems
< ¥(s)+7r, may
be presented as a countable
dimensional bases w(s) tk
415
Vs € S}
intersection of the cylinders in XT
with two-
Sieh) (s) ze + ~2(s) 25 )ye(s) 2? (te), ha
mapping the space XT? of Problem 4.8 in the separable B-space C(S) is proved.
4.15. Prove the measurability of the function b
F(a) = fols) = f(s, 2°@ at, a
where y(s, t) is a continuous function of s, t, s € S, S is a bounded closed set. Solution.
tEef
=
This function maps the space of all functions A? of the variable
[0, oe) ) into the separable B-space C(S). In accordance with the result
of Problem 2.3.8 the function f(z) is determined almost everywhere in XT relative to the Wiener measure [/yy.
Therefore almost for all functions x(t) € XT
there
exists the limit
:
¢ witpahigth=5
3
f(t) = lim fa() = lim —— 2, Pls, te) 2 (te), th =at+k
(b = a)/(n = ty But all the functions fy, (x) are CA. B)-measurable
according to the result of Problem 4.14. Consequently, f(x) represents the limit almost everywhere relative to the measure [lw of the convergent sequence of (AR; B)-
measurable functions.
So the function F(z) is (Al. B)-measurable where Aly is
the complement of the d-algebra AT? relative to the measure Hw, i.e. [Ly is measurable.
4.16.
Calculate
+ ~2(s)z?(t2)
LLw
the
integral
of the
function
f(z)
over the whole space of the functions XT
=
(s)x(t1)
on the Wiener measure
(Problem 4.8) if p14 (s) and p2(s) are continuous functions on bounded closed
interval S' of the real axis. Solution.
In a given case the argument
of the function f(z) is the
function x(t) from the space X° of scalar functions of t € S on which the Wiener measure /ly
is determined.
The space of the values of the function f(x) is the
B-space of continuous functions C(S). As the function f(z) depends only on the
416
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
values of the function x(t) in two fixed points t;, tg then for calculating the integral f f(x) pw (dz) we may use the expression of Problem 4.8 of the measure rectangular sets at
yw
on
2 = 2. As a result we receive
| #@nw (a2) = RE —
/3[vr(s)e
+2(s)x3 | Jexp{-— ,ee
|
(1)
For calculating the integrals over 21, 22 we use formulae
per
cx l2de =
2m 32/20
Ve
/ —ex?/24
2a
;
5S)
ee Toe)
foe)
[oe
/ ze
/dz = 0,
/ wre
/2dz = de|iss €
So we find
(11)
C
J fo) (a2) = or(sts + g2(0)t2-
cu)
4.17. Using Problem 4.16 calculate the integral of the function
fies S> ve(s)2?(te),
Cee Clos)
Ceeaibe 2, re)
ke!
over the whole space of functions X S with the Wiener measure Lew. 4.18. Show that the integral of the function
f(z) = >) ve(s)2™ (te) K=1
over the whole space of functions X T yelative to the Wiener measure is equal to zero at odd m
= 2r4+ 1. At evenm = 2r
J$@)nw (ae)=(2 - MD SENOLS pil
Préblems
A17
4.19. Calculate the integral relative to the Wiener measure of the function
f(x) = 91(s)e(ti) + Ya2(s)a(t2), over the rectangle set of space XT
v1, 2 € C(S),
with the base
A =
(a1, bi) Xx (do, bs) in the
product of the spaces Xt, x Xt, é
4.20. Show that the following formula is valid
[seni (da)= ef © euls)eu(s) mine} kl=1
for the integral over the whole space X T relative to the Wiener measure function
of the
1
fa
exp{ >)pe(s)e(te) }.
(II)
neil
4.21. Calculate the integral b
/f(x)uw(de), f(2) = yp(s, t)x?(t) dt.
(1)
a
In Problem 4.15 it was evaluated that the function f(x) is (Abs B)-measurable, and in Problem 4.16 was calculated the integral for the function obtained from F(a) by the substitution of the integral by the sum.
On the basis of these results, putting
as in Problem 4.15
ful) =
b—a n
92 ols,te)2"te),
(nt)
k=0
we get
/fa(2)uw (dz) =
h=
n—-1
7> (5,tate,
(111)
n—
k=0
b
sim,ffale\uw (de) = | o(s,2) dt.
(IV)
It remains to prove the possibility of the limit passage under the integral sign. For
this purpose we notice that n—1
I fo(2) I< gn(z) = 2 > sup ols,te)la%Cte). u
k=
ses
—(V)
418
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical
Representations
Putting b
a(e)=lim an(2)=fsup lo(s,t)l2%(0)at, nm—-oco
(VI)
a
we obtain similarly as earlier in (III)—(VI) b
lim
| oa(oayw ds
n—-oo
[ov |p(s, t)|t dt.
(VII)
sESs
a
But in accordance with the Fatou lemma
/9(z)Hw (de) < im /gn (x) uw (dz).
(VIII)
Thus the function g(«) is [Ly -integrable and the functions fn are bounded in the
measure by //w-integrable function g(a) +cat
some c > 0. As fr (x) =m f(z)
almost everywhere relative to the measure [lw we get b
/f(a)uw (de) = tim /ge 4.22.
Using the procedure
of Problem
Gas fened, 4.21 prove
x)
that the integral of the
function
F(a) = fo(s,t)2""(@ at
(1)
over the Wiener measure [iw extended on the whole space XT is determined by b
/f(«)uw(dz) = (2r — 1)! iv(s,t)t” dt.
(II)
4.23. Prove that the integral of the function b
Fle) = exn{|o(s,)2( ar}
(1)
a
over the Wiener measure /lw extended on the whole space XT? is determined by
| #@)nw (a2)besole SGP idadeamcuaien its}. (I)
Problems
Solution.
419
Introducing the functions
fav) = exp ict y rita} k=0
and using the result of Problem 4.16 we find similarly as in Problem 4.21 bb
im
fn(x) uw (dz) = exp{||p(s, t1) p(s, t2)min(t1, tz) dty tts}.
(IIT) It remains to prove the possibility of the limit passage under the integral sign. We notice for this purpose that according to the Fatou lemma which may be used as a
result of nonnegativity of all the functions fie) at each s € S,
if(a) yw(dx) = tlim fn (2) w(da) < lim fkfals)uw(de). From here and from (III) it is clear that the function f(z) is Ly -integrable and its norm is equal to b
| (2) [I= sup exp{icon ae a
And as the sequence SP (x)} converges to iz) almost everywhere relative to the measure [lw then
| fa(2) [I= sup exp aes Salva] < Cll Fe) I se
at some
C' >
k=0
1. Thus the sequence of the functions {ine}
convergent almost
everywhere relative to 4w is bounded in the norm by {lw -integrable function. Hence
according to the Lebesgue theorem it follows that
J fw (ae) = jim, [fale)pr (do), and we obtain the required formula (II). The formulae obtained as a result of solution of Problems 4.22 and 4.23 and the conditions of the existence of the Bochner integral show that the functions [o.e)
fia) = fos,t)x* (tat 0
420
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
fo(x) = cxof fo(s.0e00 it} 0
are {Ly -integrable if and only if the improper Riemann integrals co
[ov lp(s, t)|t dt,
0
ses
// sup |y(s, t1) p(s, t2)| min(t1, tz) dt, dte sES
0 0
converge. Notice that in Problem 4.16 and in Problems 4.17—4.20, 4.22 and 4.23 the conditions of the existence of the Bochner integral (the integrability of the norm of
the functions Fle) considered as the functions of s, s € S from the space C(S)) are fulfilled. 4.24. Calculate the integrals relative to the Wiener measure //y extended over the whole space X T of the following functions:
fila) =sin f9(s,t) (at
ful) = cos Jo(s,t)(8) at pinpoint fos.) x(t) dt ey
ek fos.) n(t) dt,
fs(z)= ae
t) a(t) dt sin fpals t) x(t) dt,
fe(x) = exp}Joilst) a(t) dt cos Jpal(s,t) a(t) dt
Hfal)e =-af fora ¥(2) ae near}, ve) fr (= exp y1(s,t) (x)
yi(s
x(t) oaths
oit) x(t) dt,
t) dt cosh [pa (s,t) a(t) dt,
=i si(s,t1)...(9(8,tm)e(t1)...2(tm) dty ... dtm,
fro(2) = ¥(2)sin fols, 8) 2( at fus(2) = W(2) cos f(5,2) 2(t) at
Problems
421
fio(z) = (x) sh Ta) a(t) dt,
fis(x) = (2) ch ae Instruction.
a(t) dt.
For calculating
the integrals of fo—f13 replace in
the formulae of Problem 4.23 and in the formulae for the integrals of the functions
fi,..-, £4 the function p(s, t) by the function ay(s, t), differentiate the obtained formulae ™ times with respect to @ and put aw = | after that.
4.25. Calculate the integral relative to the Wiener measure Jw
extended over
the whole space X T of the function
b
b
“
.
f(a) = ff elestas.--stm)eltr)--2(tm) dtr. te If the functions ( in all previous formulae do not depend on § then all the integrals in Problem 4.16 and in Problems 4.17—4.20, 4.23 and 4.24 represent abstract
Lebesgue integrals. 4.26.
Consider the spaces
Dax
Aa, bw).
The measure
extended to the space of the continuous functions C' C CAt
Setting in the space C’ an ordinary norm || z l= sup
space C(T).
[Ly
may
be
XT? with the o-algebra | a(t) |we get the
Show that the o-algebra in the space C(T) induced by a set of all
open balls enters into the 0-algebra CAT. Solution.
For this purpose we notice that any open ball in Cr} represents
the limit of decreasing sequence of the rectangles:
{a(t) :| a(t)-— y(t) |< r WteCcT} = lim {2(t) : | 2( 8%.) — Ose) |
f(x)
of Problem
| and whether
{fn (@)} converges to f(z) in p-mean?
4.23
belongs
the sequence
to the space
of the functions
422
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
4.28.
Prove that the value of the measure
[yw
on any set CeE
Alts.co) is
expressed by the formula
pw (C) = pee fexrt-Zdefrolevvaldvia)
©
4
at any fixed t;.
At all possible values t; € [0,0o) formula (I) determines the
measure {Ly on all the sets Ce
A
4.29. Let (xe ; Ay) be an infinite product of,the measurable spaces
Ae ey Aaa tEeT
where B™
Ae
ee
Ace
ea
ee
teT
is the o-algebra of Borel sets of the space R™.
It is evident that Xe
represents a space of all m-dimensional vector functions with the domain J’. We shall
determine on the semialgebra of the measurable rectangles of the space (x? ; A’) the measure by formula
[WAU xr
1
Grr viKi
eX Ae is ae
J,
| ef exp{-e7K-*2/2} dz, dx
X At3:t1,t2(de3} v1, oye : tN; ota hone
OSa; Z1,---, oo
(I)
where ie
Ly
pales
L2
SO wise,
K(t1, t2)
K (to, t1)
K (te, te)
K(s1 ; S2) is a continuous matrix function of the dimension ™
X mM satisfying the
following conditions:
(i) K(s2, 1) = K(s1, 2)"; N
(ii) oe ul K(s;, Sj) Uj > 0 at any N, s1,..., $y and m-dimensional vec-
t,j=1 tors Uj,...,UN;
(iii) tro(ty, t2) < Blt, _ to|7 at some B, y > 0), where o(ti, t2) =
K (ti, t1) =
K (ti, tz) =
Aig :tetoctesa (Be DiAlsisey Creag) (k bE in the correspondent
spaces
K (te, t1) + K (to, ta),
n) are finite nonnegative measures
(Xt, ; Ai, ) representing
at each
B
€
vA rs (At,
Kos X Ais cet ; B)-measurable functions Z1,...,£@;~—1. The measure / is 0-additive on the semialgebra C of rectangles of the space X T and may be uniquely extended on the d-algebra Ar obtained by the supplement of A? relative to LL.
Problems
423
Prove that the measure /£ in (I) is finite. Instruction.
Apply the formula
u({a(t) : |wn(te) — ee(ti)| > n} = 17
st 20H ORE tL, U2) 0" (ee Lyre em), where Orn(ty . t2) are diagonal elements of the matrix o(ty ; to), an inequality of (1) from Problem 4.11 and take into account that m
Je(t) — a(t) = .]D lea(te) — we(tr)]? < D0 lee (te) — ee (ta)h; ks!
k=1
as a result
{x(t) : |e(t2) — x(t1)| > ©} C LJ {2@) : |ee (ta) — ae (t1)| > €/m}, k=1
and take @ of Problem 4.11 in the interval (0, y/2). 4.30. Prove that the integral over the measure /l of Problem 4.29 of the function
FOLDS Grr mea abe Rist where m
(pI
(A, 1 ll
Callas n) are continuous
(1)
matrix functions of the dimension
X m with the values in a separable B-space extended to the whole space C’ is
determined by formula
pee yin) H(t ng ti):
[|t@u(ae) Sg ait
(II)
bil=1
4.31. Under the conditions of Problem 4.29 determine the measure
Aq
wee
Xe xX Ax
1 =a!
Oo anere tt, (Olas oi,
fatale,
tr
fe
—2" K-'z/2\ dz
at,
(don) 015, n—1);
(I)
424
Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations
where v1
ps
x2
3
ee
;
v4
Kitesty)
K (ty, ta)
K (ti, ts)
K (ta, t1)
K (tg, ta)
K (te, ts)
KGazta)
K (to, ta)
K (ts, t1)
K (ts, ta)
K (ts, tz)
K (ts, t4)
K (ta, ti)
K (ta, ta)
K (ta, tz)
K (ta, ta)
:
Kk (ty ; to) is a continuous together with its derivatives till the second order inclusively
the matrix function of the dimension ™
X ™
possessing properties
(i), (ii) from
\
Problem 4.29 and the additional property
tr o(ty, to, ts, t4) < Br’
at some B > 0, y > 2,7 = max; j \t; — bale
o(ti, to, t3,t4) = K(t1, ti) + K(to, te) + K(ts,t3) + K (ta, ta) —K(t1, te) = K (te, t1) = K (#1, ts) = K (tg, t1) + K (ti, ta) + K (t4, t1)
+K(to,t3) + K(ts,t2) — K (to, ta) — K(ta,t2) — K(ta, ta) — K(ta, ts); the measures NG
ae
toon (k Soest
n) are the same as in Problem 3.8.6. Prove
that almost all the functions of the space X T yelative to the measure p/ are continuous and have continuous first derivatives. Instruction.
prc certt aee (one JE
Use formula
Vier y
_7T K7~*2/2} g-1 exp{—ax* dx; dry dxr3 dzr4
|Cak —23k—Cort+@ir|>
= 1 — 20(n/./orn(t ta,ts,r, ta)), where Ork(t1, to, t3,t4) are the diagonal elements of the matrix o(ti, ta, ts, ta), use from inequality (I) Problem 4.11, take q@ in the interval 7 qi ap
~~
GENERAL
CHAPTER 5 THEORY OF STOCHASTIC AND ITS APPLICATIONS
SYSTEMS
Chapter 5 is dedicated to the general theory of stochastic systems described by integral, differential, integrodifferential, difference and etc stochastic equations_in the finite- and infinite-dimensional spaces. Section 5.1 is devoted to the stochastic Ito differentials, the Stratonovich differentials and the 6-differentials. The It6 formula and its various generalizations are proved. Basic definitions concerning the stochastic differential equations, the change of variables and sufficient conditions of the existence and the uniqueness of the solutions are given in Sections 5.2 and 5.3. Practical problems connected with the transformation of the system equations into the stochastic differential equations based on the shaping filters are considered in Section 5.4. Numerical methods and algorithms for the finite-dimensional stochastic differential equations solution are presented in Section 5.5. Section 5.6 contains the general theory of the multi-dimensional distributions in the continuous stochastic systems described by the stochastic integral and the differential equations. The multi-dimensional distributions in the stochastic discrete and continuous-discrete systems are considered in Section 5.7. Structural theory of the distributions in the stochastic composed systems is given in Section 5.8. Section 5.9 is devoted to the theory of the multi-dimensional distributions in the stochastic infinite-dimensional systems. The last Section 5.10 contains applications to filtering based on the online data processing.
5.1. Stochastic Differentials 5.1.1. Stochastic Differential of Ito Process
A random process Z(t) is called the [t6 process if it is representable in the form t
Z(t) = Z(to) + [ x(nar+ [owe to
A |/Vee to Ré
crada
(5.1.1)
426
Chapter 5 General Theory of Stochastic Systems
Here to > 0, W(t) is the Wiener process, P°(A,A) is the independent of W(t) centered Poisson measure, the first integral is the m.s. integral, the second integral is the It6 stochastic integral, the third integral is the stochastic integral with respect to the Poisson measure P(A, A), X(t), Yi(t), Yo(t, u) are random functions satisfying the conditions of the existence of the respective integrals. If Z(t) and W(t) are the vector random processes, then X(t) and Y(t, u) represent the vector random
functions of the same dimensions as Z(t) and Y;(t) represent the matrix random function. The formal expression
dZ = X(t)dt + ¥;(t)dW(t) + HiY(t, u)P°(dt, du).
(5.1.2)
Ro
is called the stochastic differential of the Ito process Z(t). Remark.
It should be emphasized that expression (5.1.2) has no usual sense
of differential since it involves the infinitesimal quantities of different orders. Really, replacing the integrals by the respective integral sums,
the increment of Z (t) ona
small interval (t,t + At] may be expressed as
AZ = X(t)At + ¥i(t)AW(t)4 /¥o(t,u)P%((t,t + Ad},du), (5.1.3) RG where AW
=
Wt
+ At) — W(t). Since t+At
B(AW)? = B[W(t + At) - W(t)? = / Br\dr oe Olt). t
the second item in the right-hand side of (5.1.3) has the order V At.
In the special case where Y2(t, u) = 0 and Y(t) = Y(t) with probability 1 the Ito stochastic differential of the It6 process t
Z(t) = Z (to)+/X(r)dt + fiY(r)dW(r)
(5.1.4)
to
is given by
dZ = Xdt+ YdW, where the argument t is omitted for brevity.
(5.1.5)
Bedy Stochastic Differentials
427
5.1.2. Ito Formula
Theorem 5.1.1. Let y(z,t) be a scalar real function, Z(t) be an Ito process given by Eq. (5.1.4). If the scalar function y = y(z,t) is continuous and bounded together with its first and second derivatives, its second derivatives satisfy the Lipschitz condition
Oop(2",i)" 07 ol, t) Oz, 02,
-
> [plz + Ye"*,t)—o(z,t)] w= 9 (s+ yor ue, 7 cs
Sei
—p(z + Yy7,t) + Spel(z,t)7 Yq" yp’ = o(up’) = o(h). SE
Its probability is equal to jure. 0 @ —Hp
=sal sefea=
o(h » The values of V;
corresponding to all other bad Psks are bounded and their probabilities are equal My
5
to I] hosts up ,m,+---+
my,
> 1. So we have the following estimate
ae!
for the expectation of lV;| :
EWV|1
s=1
434
Chapter 5 General Theory of Stochastic Systems
My,
Mi
= o(h) +e(1 — exp {dont} -S ya exp {yh) = o(h)p Hence
a
EIS) [g(Z* + AZ}, te-1) — o(Z*, te-1) ] k=1
N
-Sol¢
Py Wyte)
= o(Z*,th-1) a Re(2"
ten)
Yo") uP
k=
N
,
—S> [e(Z* + Yo, te-1)
‘
— o(Z*, te-1)] P™*| < No(h) = No(N7").
K=1
This proves together with the estimates of Subsection 5.1.2 that (5.1.17), may
be
rewritten as N
Dey (20)+) (AZt + az) — (4 (to), to) k=1
N
X(te-1) = S"{e.(Z* + AZE, te-1) + Ge(Z* + AZE, ty_1)? k=1
?
il waite [¥22(Z* +AZ5, tp 1) Yi (te 1)V (te 1) Yates 1) )T ]}h
N +) >9.(Z* + AZ* te_1)7 ¥i(te-1) AW* k=
N M,
+ OD Le? + YF, te-1)
— (2? te-1) — pe(Z* tea) VE" | we
N
+3090
k [9(Z" ik + VP,k te-1) — 9(Z*,te-1)] Ph t+0(h).
(5.1.18)
gail Sail
Finally we notice that due to the boundedness
of the function v(z, t) and its first
and second derivatives and to (5.1.6) the first factors in the first two sums in (5.1.18) differ from their values at AzZé = 0 by infinitesimals o(h). Therefore AZ
sums may be omitted giving N
~ (2 ag S\(AZt + az. as ~(Z(to), to) Koil
in these
5.1.
Stochastic Differentials
435
“a
N = So {e(Z*, te-1) + 0,(Z*,te-1)? X(te-1) neal
+5 tr [ pz2(Z", th—1) Yi (te—-1)v(t Yie-1) (te-1)? |}h N
+e
ae Pali )AWF+
M,
SS) [ [o(Z* + YF*, th-1) Ra's
—9(Z* ,te-1) — p2(Z*,te-1Yo" )? |we N
+ 5D [e(* + Yo te-1) — (Z*, te-1)] P+ O(A). Passing to the limit when h and maximal diameter of the sets A; tend to zero we find
y(Z(t), t) =
~(Z(to), to) 43 [Molen
a ~2(Z,, 7)" X(r)
t
450 PUA
AGUA Gmttra /gs Fi 2) Vitra (2) to
+
up(dr, du) (Zr, tT)" Yo(r, u)|
,7)—P2 nn u),7)—P(4r oN Ugcee
+
[e(Z,
v
+ Yo(r, u), 7) —
9(Z,,7)]P°(dr,
du) ,
to 0 For proving the existence of m.s.
sequence of the random functions for
solution of Eqs.
(5.2.14) we construct the
p = 0,1,2...
2(ar). Yp4i(t) = O(t) + falolr).rt)dr + f Yp(0),7) to
to
(5.2.18)
5.2. Stochastic Differential Equations
451
We estimate the differences t
Y;(t) — Yo(t) = ®(t) + ia(Yo(r), 7,t)dr + ib(Xo(r), 7, t)Z(dr),
(5.2.19)
Yosi(t) —Yo() =O) + fLalp(r),78) ~ a(¥%-1(7), 7,1)]ar t
7 /[b(Yo(r), 758) — B(%p_a(r)y7,t)]Z(dr).
(5.2.20)
to
As
a(Yo0(r), 7, t) is m.s.
continuous
||a(Yo(r), 7, t) || We shall cite the proofin connection with the random functions with the values
in B-space Ly (Q, Pople i) of n-dimensional vector functions of the variable W. As it is known from functional analysis the norm of the value of the random function
Alt) = z(t,w) at each ¢ is determined by formula
IXOIE = fle o)IPPde) = BIXEOP.
6.39)
In other words, the square of the norm \|x (t)]]? of the value X (t) at a given t in B-space D2(Q, On, F*) represents an expectation of the square of its Euclidean
norm ||X (t)||. Then Eqs. (5.3.1), (5.3.3) and (5.3.7) may be considered as the equations in B-space D2(Q, ny al Be) and condition (5.3.8) will be written in the form of the Lipchitz condition
|Fi(¥i(7), 7,t) — F(Y(r),7,t)|| < k|l¥i(7) -— Y(r)II.
(5.3.10)
Therefore for the completion of the proof it is sufficient to use the known from the theory of the ordinary differential equations the Picard method
of the successive
approximations. We
determine
the
sequence
of m.s.
continuous
random
functions
for p
a Ube Py pee t
Xo(t) = (t), Xp4i(t) = O(2) + fP%s(r), ner
(5.3.11)
to
and estimate the norms of the differences t
:
X(t) — Xo(t) = J For), ntar,
(5.3.12)
to
Xp4i(t) — Xp(t) = [[email protected]
— F(Xp-1(7), 7, t) |dr.~
(5.3.13)
As according to the condition the random function F(Xo(r), Py t) of the variables 7, t is m.s. order moment
continuous in consequence of m.s. I|F(Xo(7), T,t)||
2
=
continuity of O(r ) then its second 2
E
\|F(Xo(7), T5 t)|| is continuous and finite
at all T, t, and consequently, achieves its upper bound on the finite closed interval
458
Chapter 5 General Theory of Stochastic Systems
r € to, t],t€ [to,to + T]. Therefore ||F(Xo(r),7;t)||’ < ¢ at some ¢ < 00. Hence from (5.3.12) we find
1X1 (t) — Xo(t)|) < [ieco), Pare tie
Oat
After that from (5.3.13) and (5.3.10) by the induction we get
Koes(0)ae—Xp(OI
F(X((r), 7,4), LF (Xp(7), 7 le < NF (Xo(7), 7, lle
+ TIF (Xq(7), 7,1) — P(Xq-1(7),7,8)I] < cet), Thus
at all 7 the sequence
integrable function of
F(X(r), a; #) is bounded
(6.3.17)
on the norm
by the
7. Consequently, on the basis of Lebesgue theorem which is
known from functional analysis the function F’ (X (r Vs Tas t) is integrable over T and
we may pass in (5.3.11) to the limit at p —
OO under the sign of the integral.
It
means that the limit function X(t) satisfies Eq. (5.3.7) what proves the existence of the m.s. solution of Eq. (5.3.7).
Let Y (t) be any solution of Eq. (5.3.7). After substituting it into (5.3.7) and subtracting it from (5.3.11) we obtain t
bead ee ‘iF(Y(r),7,t)dr, to
Aprit)—Y
y= /[F(Xp(r),7,t) — F(Y((7), 7, t)] dr,
5.4. Transformation of System Equations
_
into Stochastic Equations
459
_kPt1(t — to)P+?
|Xpai(t) — Y(t)|| = iene
tb
Hence passing to the limit at p — 00 we shall have ||X (t) _ Y (t)|| = 0. It proves the uniqueness of the m.s.
solution of Eq. (5.3.7), and consequently, also of Eq. (5.3.1) (with the accuracy till the equivalence). to. The system which is described by Eq. (5.4.6) for the normally distributed white noise V = V(t) with zero expectation EV = 0 and with the matrix of the intensities vy= v(t) we shall call the normal stochastic differential system. For some composite stochastic differential systems in the finitedimensional space the stochastic differential équatin of the form is used
dY = a(Y,t)dt + b(Y,t)dWo + :c(Y,t, u)P(t, du).
(5.4.7)
Ro Here a = a(y,t) and b = b(y,t) are the known (p x 1)-dimensional and (p x m)-dimensional functions of the vector Y and time t; Wo = Wo(t) is a m-dimensional Wiener random process of the intensity vp = vo(t);
c(y,t, u) is (px 1)-dimensional function of y, ¢ and of the auxiliary (q x 1)dimensional parameter u; [ dP°(t, A) is a centered Poisson measure: A
parte) = fare) A
A
[ret aya,
(5.4.8)
A
where { dP(t, A) is a number of the jumps of a Poisson process at time A
interval A; vp(t, A) is the intensity of the Poisson process P(t, A); A is some Borel set of the space Rf with the pricked origin of the coordinates. The integral in Eq. (5.4.7) is extended on the whole space R{ with the pricked origin of the coordinates. The initial value Yo of the vector Y represents a random variable which does not depend on the increments of Wiener process Wo(t) and of the Poisson process P(t, A) at time intervals A = (t1,t2] which follows to, to < t; < tg for any set A. Remark. In the case when the integrand c(y, Us u) in Eq. (5.4.7) admits the presentation c(y,t,u) = b(y, t)c’(u) Eq. (5.4.7) is reduced to the form of (5.4.6) if we assume
W(t) = Wo(t) + fewrre. du). RG
5.5. Numerical Integration-of Stochastic Differential Equation
463
In some practical cases to stochastic differential Eq. (5.4.7) may be reduced the stochastic integrodifferential equations of the form of Eq. (5.4.9)
dY, = a(Yf., t)dt + B(Y#, t)dWo + /o(¥i,t,u)P%t,du).
(5.4.9)
Ro Here a = a(yj,,t), 6 = b(y},,t), ¢ = c(yj,,t,u) at each instant depend not only on the current value of the process Y(t) but on all its values Y(r) at the interval [to,t), i.e. the components of the vector functions a, c and the elements of the matrix 6 are the functionals of the process
Y(r), 7 € [to, t). In problems of practice for calculating the probabilities of the events connected with the random functions it is sufficient to know the multi-
dimensional distributions (Section 2.7). Therefore the central problem of the theory of the continuous stochastic systems (Subsection 1.6.5) is the problem of probability analysis of the multi-dimensional distributions of the processes which satisfy the stochastic differen-
tial equations of the form Eq. (5.4.6) or Eq. (5.4.7) at the correspondent initial conditions. In the theory of continuous stochastic systems two principally different approaches to distributions calculation are recognized. The first general approach is based on the statistical simulation, 1.e. on the direct
numerical solution of stochastic differential Eqs.
(5.4.6), (5.4.7) with
the corresponding statistical data processing (Section 5.5). The second general approach is based on the theory of continuous Markov processes and supposes the analytical simulation, i.e. the solution of equations in the functional spaces for the one-dimensional and the multi-dimensional
distributions (Section 5.6). Along with the general methods of the stochastic systems theory the special methods are distinguished which are oriented on the linear stochastic systems (Chapter 6) and on the nonlinear stochastic systems
(Chapter 7). 5.5. Numerical Integration of Stochastic Differential Equation 5.5.1. Introductory Remarks Integration of stochastic differential equations has some peculiarities. The fact is that all numerical integration methods of the ordinary differential equations besides
464
Chapter 5 General Theory of Stochastic Systems
the simplest Euler method
are based on the calculation of the increments
sought functions on each step by means mean value.
of the
of using the integral theorem about the
In the accordance with this fact the right-hand sides of the equations
(the derivatives of the sought functions) are taken in the mean points of the intervals. Various methods of the numerical integration differ in essence from one another only by the way of approximate finding the mean
values of the right-hand sides of the
equations.
The theorem about the mean value is not applicable to the stochastic integrals. But some
analog of the theorem about the mean
integrals with respect to the nonrandom functions.
value is valid for the stochastic
It consists in the fact that the
best approximation of the stochastic integral with respect to the continuous nonran-
dom function gives the product of the value of a given function by the increment of the process over which the integration is performed.
Therefore, for instance, all the
methods of the numerical integration of the ordinary differential equations may be
used for the stochastic differential equations (5.4.6) whose coefficient at the white
noise is a determinate time functions f, i.e. b(Y, t) = b(t). If the function b(Y, t) in Eq. (5.4.6) depends on Y then the numerical integration method of such equations is required to be chosen depending on the fact in what sense the stochastic integral is assumed.
In the case of the It6 equation the increment of the process Y(t)
at the interval [t,¢ + At) should be determined by formula AY
=
a(Y;,t)At
= b(Y%, t) AW what corresponds to the Euler method of the numerical integration. Thus the It6 equations may be integrated by Euler method.
But for better accuracy
of the calculation of the first item we may take the value of the function a(Y;, t) in
some mean point of the integration interval Lt. t+ At). Here we may use any method of the numerical integration, for instance, the widely known Adams method and the
Runge-Kutta method. But the value of the function b(Y; ; t) should be always taken at the initial point of the integration interval.
In the case of the equation in the
0-differential the value of this equation should be taken at the point t + OAt. But for its approximate finding we have to elaborate new methods of the numerical integration.
While simulating the stochastic differential equations by means of the analog computers for the functions bY, t) which depends on Y it is necessary beforehand to
reduce the It6 equations to the symmetrized (Stratonovich) form. It may be explained by the fact that at such simulating a white noise should be substituted by a process with a small correlation interval but such one which is different from zero as a white noise is physically nonrealizable.
In this case a simulating process will be close to the
solution of a stochastic differential equation if and only if a given equation is assumed
as a symmetrized stochastic differential equation.
While simulating the processes
which are determined by Eq. (5.4.6) it is necessary to simulate the random variables or random processes. At numerical integration of the stochastic differential equations
5.5. Numerical Integration*of Stochastic Differential Equation
on digital computers it is required to simulate the increment AW
465
of the random
process W(t) at each step. While simulating by means of the analog computer it is
necessary to generate the wide-band random processes.
5.5.2. Existence of Exact Difference Equation Stochastic Differential Equation
Correspondent to Given
Let us consider the It6 stochastic differential equation (5.4.7). Suppose that it is required to substitute this equation by a difference equation for the values of the
process Y(t) in a given discrete series of the equidistant points {tk i} ty = kh where h is an interval between adjacent points t ry
a
3k+1 —tx. In principal the problem
is solved exactly as the values of Markov process Y(t) at the points t, form Markov
random sequence {Y;}, Yi, = Y (kh), and any Markov sequence is determined by some stochastic difference equation. by means
But to compose this exact difference equation
of a given differential equation is practically impossible.
First of all for
its composing it is necessary to find a transient distribution of Markov process Y(t) determined by Eq.
(5.4.7) and then we may compose a difference equation for the
sequence {Yn} by means of the found transient distribution. But the exact evaluation of the transient distribution of the process Y (t) is possible only in some specials cases. In the general case we are satisfied with an approximate evaluation of a transient
distribution of the process (2):
As a result by this transient distribution we may
obtain only an approximate difference equation.
The latter makes it expedient the
application of very complicated algorithm derivation of an exact difference equation for the sequence
‘nee
As we
are satisfied only with an approximate
difference
equation even at the application of a given algorithm then it is expedient to make it by more simple ways.
At first we substitute an integral over the variable U in Eq. correspondent integral sum.
(5.4.7) by the
As a result Eq. (5.4.7) will be replaced by the following: N
dY = a(Y,t)dt + 0(Y,t)dWo + 5 c(Y¥,t)dP?,
(5.5.1)
cil
where Celyy; t) are the p-dimensional vector functions which represent the values of the function cy, Ue u) in some of the partition of g-dimensional et
Pas ae gs My and P(t)
are
mean
point U; correspondent
to the elements
sphere of sufficiently large radius, the
centered
simple
Poisson
Uj;
processes
€
A;
Aj (2
(Subsec-
tion 4.6.6):
Poy = Po(0/4) Ai) — w([0,4), Al), 7 = 1,2, 2a.
(5.5.2)
The intensities of these processes are determined in terms of the expectation p(A, A)
of Poisson measure P (A, A) by formula
y(t) = du ([0,t), Ai)/dt.
(5.5.3)
466
Chapter 5 General Theory of Stochastic Systems
The replacement of Eq. (5.5.1) by the correspondent difference equation is performed by the Euler method.
5.5.3.
Euler Method
The
simplest
way
of
the
replacement
rence equation consists in the replacement
of
Eq.
(5.5.1)
by
a
diffe
of all the differentials by the elements
of the integral sums:
Y((n + 1h) — Y(nh) = a(Y (nh), nh)h + 1(Y (nh, nh))[ Wo((n + 1)h) N
—Wo(nh)] +) ei(¥(mh), nh)[ P2((n + 1)h) ~ P2(nh)]. Putting
Yn =Y(nh),
Yn(Yn)=Y(nh) + a(Y(nh), nh), vin(Yn)=b(Y (nh), nh),
Win(Ya) = c-1(¥ (nh), nh), Vin = Wo((n + 1h) —Wo(nh), Vie iP, 1
(5.5.4)
+ 1h) Fe(Wh), tS 2 Nee
we get a stochastic difference equation
Viutd =
ON
OE) a oy Wil Valin: t=]
Introducing the p X (m oe N) block matrix
Yn(Fa) = [Yin(Fa)---Pavstyn(Fa)]
(5.5.5)
and (m +N )-dimensional random vector
Vi, Sieve
reyes
(5.5.6)
we may write the obtained difference equation in short form:
Yn¢i = Gn(Yn) + Ya(¥n)Vai
(5.5.7)
As the Wiener and the Poisson processes are the processes with the independent increments (Subsection 4.6.3) then the random vectors V,, form the sequence of the
5.5. Numerical Integration*6f Stochastic Differential Equation
467
independent random vectors {Vn } and the blocks Vj, of the vectors V;, have the
normal distribution V (0, Gis) where (n+1)h
oF
il valde = aalhh:
(5.5.8)
nh
Vo(t) is the intensity
of the Wiener
process
Wo(t),
the scalar blocks
Vj,
(7
= 2,...,.N +1) have the Poisson distributions with the parameters (n+1)h
PS
d V;(7)dr = 4(nh)h.
(5.5.9)
nh
These distributions determine completely the distributions of the random vectors V,,.
The covariance matrix Gy, of the vector V, represents a block-diagonal
matrix
+
Gael
TRANG
OP
abe
0
DE tia
MOO )esE (2 0s
0
0
0
=
(4)
(5.5.10)
HN+1,n
Eq. (5.5.7) determines Herr at a given ¥:, with the accuracy till fh in the determinate item Yn, (Yn) and with the accuracy till Vh in the random item Vn (Yolen
Remark.
The stated method of the substitution of the stochastic differential
equation by a difference one in essence does not differ from the Euler method of the
numerical integration of the ordinary differential equations.
5.5.4. More Exact Equations For deriving more accurate difference equations than Eq. (5.5.7) we substitute (5.5.1) by the correspondent integral equation
(n+1)h
(n+1)h
A¥n= fave rdr+ f W(¥%;,r)aWar) nh
nh
N
(n+1)h
+> felts naredr). SUE
is
(6.5.11)
468
Chapter 5 General Theory of Stochastic Systems
For the purpose of the approximate calculation of the integrals we determine Y, by means of the linear interpolation of the random function Y(t) at the interval
(nh, (n ap 1)h). Then setting as usual AY, = Y((n i 1)h) ae Y (nh) we have
a(¥,7) = a(Ya +
r—nh
A¥»,1), (V7)
= (Fa+
7T—nh
A¥a,7),
(5.5.12) e:(¥;, 7) 2c; (¥.+
7 —nh
scaly
eT Ne
For calculating the right-hand sides in these formulae we use the generalized It6
formula (5.1.24). So we have
o(Y + dY,t + dt) = 9(Y,t) + {¢:(Y,t) + oy (Y, t)7a(Y, t) N
+e, t):o0(Y,t) + Sey jell
+ ¢;(Y,t),t) — 9[Y,t)
—py(Y,t)” ci(Y,t)]us(t) }dt + vy(Y, t)7O(X, t)dWo N
+S “[e(¥ + c(¥,t), t) — 9(¥, t)]dP?,
(5.5.13)
where ;(y,t) is a partial derivative of the function y(y, t) over time ¢; Py (y, t) is a matrix whose rows represent the partial derivatives of the matrix-row ply, he with respect to the components of the vector Y; Pyy (y, t) 5 o(y, t) is a vector whose components are the traces of the products of the matrices of the second derivatives
of the correspondent components of the vector function ply, t) over the components
of the vector y on the matrix o(y, t):
[Yyy(y,t) : o(y, t)le= trl Geyy(y, t)o(y, t)], o(y,t) = b(y, t)vo(t)b(y, t)’.
(5.5.14) (5.5.15)
It is convenient to transform formula (5.5.13) in such a way that namely the Poisson processes P; (t) would enter into it directly, not the centered processes P(t) :
Taking into account that
P(t) = P,(t) - /Ore
(5.5.16)
5.5. Numerical Integration’ of Stochastic Differential Equation
469
we may rewrite formula (5.5.13) in the form
oY + dY,t + dt) = o(¥,t) + (oi(¥,t) + oy (Vt) al, 2) N
—ral¥ deal] + set) so, t)}ae+ gy(¥." y
N
x(Y,t)dWo + So[e(¥ + ei(¥,t),t)— oY, t)aPi. For extending formulae is necessary
to modify
(5.5.17)
(5.5.13) and (5.5.17) over the matrix functions
the notation of some.its members
in such a way
it
that the
expressions Dy (y, t) in the context of the matrices algebra will have also the sense for the matrix fucntion Y. Taking into account that in the case of the scalar or the
vector function
for
any
p-dimensional
vector
U
we
may
substitute
the
items
y es an
x a(Y,t), Gy(¥,t)" W(Y,t)dWo, vy(Y,t) c:(¥,t) by the items a(Y,t)”
x (8/dy)? (Y,t), WEY, t)(O/dy)e(Y,t), c«(¥,t)" (8/dy)e(¥, t) correspondingly.
Therefore formulae
(5.5.13) and (5.5.17) will be also valid for the
matrix functions Y as ul (d /Oy) represents a scalar differential operator for any p-dimensional vector U and its application to the vector or the matrix function 9
means
its application to all components
of the vector Y or to all elements of the
matrix (9. The variable Pyy (Y,t) : o(Y,t) in the case of the matrix function ~ represents a matrix whose elements are the traces of the products by the matrix 7 (y, t) of the matrices of the second derivatives of the correspondent elements of the matrix
ply, t) with respect to the components of the vector y:
[Yuy(y,t) : o(y,t)Jer= tl Peryy(y, t)o(y, t)].
(5.5.18)
Using formula (5.5.17) and its modification suitable for the matrix function ~
and accounting also (5.5.15) and the fact that according to (5.4.7) and (5.5.16)
7T—nh
4
AYn =
N
+0 t=1
T—nh
tT—nh
5
=
ays, nh)h +
|=
T—mnh
b(Yn, nh) AW,,
—¢i(¥n nh)(APin — Vinh),
(5.5.19)
470
Chapter 5 General Theory of Stochastic Systems
(where as earlier MWS
Wo((n + 1)h) —
Wo(nh), NP
Se
P;((n + 1)h)
— P; (nh)) we shall have with the accuracy till infinitesimal of higher order relatively
to h the following equalities:
rT—nh
x
a—
¥
1
> evin
T—nh:>
A¥,,7] =a+a;(7 — nh)+
(r—nhy\?
his
)
h
(dyy : o)h+
T—nh h
ay LAW,
N
+0 E ula sees nh)= a|A Pie 7—nh
AY] = b+b;(7 — nh) +
N
res
Dy. [(¥,ES me es mh)-
=)
|
Me
&
S
=}
3
Dol] Re
ole
_
=a +
N
aes
Ne
ad) (Ciyy
7—nh
h
AIPEY Ano 21)
C; (%+ dee Mt AY) = co + ci¢(7 — nh)+
x
(5.5.20)
—nh wis
h
“iy
— nh :o)h+ of
bAWn
oN
7—nh
F
cj,nh
= «|Py:
(5.5.22)
Here the arguments Vis nh at all functions which depend on are omitted them for brevity.
For
increasing
the
accuracy
of the
calculations,
in particular,
variance matrix of a normally distributed random vector AW,
[T Wo(nh)and
the parameters
of Poisson
distributions
=
of the
co-
Wo((n + 1)h)
of the random
variables
[IPR = 12k ((n + 1)h)— P;(nh) we may take the values of the intensities 9 (t) of
5.5. Numerical Integrationof Stochastic Differential Equation
the process Wo (t) in formula (5.5.15) and
471
(t) of Poisson streams which generate
the processes P; (t) at the mean point nh + h/2 at the interval (nh, (n + 1)h):
o(Yn, nh) = b(Yn, nh) (mn+ ;)ee nh)",
h vin=
(5.5.23)
(mh+ 3), 2
1
ae.
To facilitate the further calculations we find the increments of the functions a, b, c; in the sums of equation (5.5.19) by means of the linear interpolation at small
interval (nh, (n = 1)h): -%
a(%,+e E ses mh)—a& 7T—nh
=
h
Nba
=
4 =
Ae
i(%+ E ses, mh)—b
T—nh
Ce (% + 7anth)
7—nh
—¢
rn
= —,Aitins
|
(5.5.24)
Asan = a(Y, + ;,nh) —a(Yn, nh), Aibn = 0(Y¥, + cx, nh) — b(Yn, nh), AjCin = ci(Yn + c;, nh) — 7 Ome nh).
(5.5.25)
Using formulae (5.5.20)—(5.5.24) we find the following approximate expressions
of the integrals in (5.5.11) (n+1)h
1
a(Y,,7)dr = ah+ 5
1
N
+ 34yy | h?
a, + ay (2—Soamm
nh
i=1 1
N
Te es (x bAW,, + Aiea]: h,
(n+1)h
(5.5.26)
N
b(Y,,7)dWo(r) = DAW, +
3
by + G - yi] “a h
“ih
iil
N
(n+1)h
h
+AweiT os+ atara| i a" dWo(r) Y
t=1
ph
(n+1)h
+5(byy :o)
h
eA nh
:
9
) dWo(r),
(5.5.27)
472
Chapter 5 General Theory of Stochastic Systems
(n+1)h
N
c:(Y¥;,7)dP?(r) = G¢APin +
Cit + chy a
Sm
ws
| A
jal
(n+1)h
N
+eFiy DAWn n + D> Ajcin 7 Cin AP; jn
BED h pes i nh
gal
,
(n+1)h
h
+5(ciyy 12)
2
Se: ) dP9(r),
(5.5.28)
nh
.
(n+1)h
N
1
ci(Y,,7)u%i(r)dr = {ns +15 |. + chy (= pI “| Ha
nh
1
1
+3 Ciyy : | +
3
-
cybAWn, + ye
ee
hi
Vin.
(5.5.29)
j=l Further on substituting the obtained approximate expressions of the integrals
in to (5.5.11) and taking into considerations that AY, = Y((n + 1)h) — Y;(nh) 0)
fra | Y,;, we come to the difference equation N+1
Fat =Pn(¥n) + Yo [Pin(Fn)Vin + Vin(Yas Vin Vin + Vin (Yn) Vin| fie
(5.5.30)
Here the following denotions are introduced:
—_
_—
=—
Onl Yn) = Yon |an, Bh)
WY
=_—
deci (ins nh)vjn | h+ j=l
N = Ss Tis {n(nh) — >pate nh)Vjn + tin nh)? 1
S,
j=
=
N girl Yn, hn)? Yen
N a(Yn, nh) —S\og (Yn, hn)vjnt |
ial
Geil
N Y AG: dgg (Vinh)
tery y, apgh us j=l
y ts}
2
(5.5.31)
5.5. Numerical Integration”of Stochastic Differential Equation
473
N
tin(Yn) = (Zn nh) + 5 dy(Yna,nh)” — S° ej (Yn, hn)? vn |b(Yn, nh), j=1
(5.5.32) N Vin(Yn) = Sa
nh) + | A;_ia,
— ST Si-cintn| h,
(5.5.33)
pict
Pin(¥n Va) = {, nh) + |a(¥n,nh)* N is fag = wy cj(Yn, hn)? vjn°|—b(Yn, nh) = Oy iil 0 N +V,L b(Yn, nh)? —b(Yn, nh} h+ Sa AjbaVi41n, oy
Y.\
ns V0)
=
(5.5.34)
gs
nh)?
{eta nh) ae Goren
»
--
a(Yn, nh) — beg GY a ini]
h
aol
N
+c;-1,y(Yn, nh)" (Ya, nh)Vin + >, Ajej-inVj4i,0;
(5.5.35)
j=l
: eghee
C
1
1 A glbuy (Yn Rh)
o( Yn, Teh,
*
4
Sea Nabi glci-1.yy(Yn, nh) Sai
Ynett):
Vin = AW, = Wo((n + 1)h) — Wo(nh),
Vin = AP;-1n = Pi-1((n +1)h) —Py-a(nh),
nh
nh
i=2,...,N,
(5.5.36) (5.5.37) (5.5.38)
(6.5.39)
474
Chapter 5 General Theory of Stochastic Systems
(n+1)h
Vi es ii SRS ins: Fe
(5.5.41)
nh
(n+1)h
9
T—nh a
dP?_,(r);
Introducing the vector V, vi Me=
(Ke Venmae
/
Vv, =
(5.5.42)
Ie
C= 2g,
nh
Wii
Et the block matrix
Yn (Yn VE?) = [din(Yn) Pin(Yns VA?) Yin(Yn) : .wnsi,n(Yn)
VNeialin)
VD)
Vrann
Ya) |
and the block random vector
Te
Vn =
a
[Vin Vin
Vin
AP ew HGOE iy BUTE
Von Von
Von
793
/
"
a8
» Vn-+1,n Vu-+ijn Vivr41,n | )
(5.5.43)
we may write Eq. (5.5.30) in short form
Vota
olin)
The variables which enter into Eqs.
Un Yas VO Va.
(5.5.44)
(5.5.44) are determined by formulae (5.5.31)—
(5.5.43). 6.8.8. Distribution of Random
Vector Vp,
We find a distribution of the random vector V,,. It is clear that the expectations of the pas
tor Rasa
variables Vin Vie are equal to zero and that the scalar vec-
Vin
T
Pe
Ae?
f
:
] has the normal distribution, and the scalar variables Von,
: Hae have the Poisson distributions with the parameters (n+1)h
a=
i Ver
GT = ea (nn
5m)[et
2 oe
INE de
nh
It is also evident VN4in
7
; VN41,n
(5.5.45) that the random Tr,
Wee ae
Vi
ee. aii
5
[ are independent by virtue of the independence of the pro-
cesses Wo (t), P, (t), a pendent.
vectors
a Py(t) and that at different n the variables V,, are inde-
But at any given 2, n the vectors V;,, Vee in) Vin are dependent.
5.5. Numerical Integrationof Stochastic Differential Equation
475
For complete definition of a distribution of the random vector V,, in Eq. (5.5.44) it is sufficient to find the covariance matrix of the normally distributed random vec-
tor [V2 Vin a yee ize Using the known formulae for the covariance and crosscovariance matrices of the stochastic integrals we find the blocks of the covariance matrix Ky, of the random vector [Vv ye
Ve 13
(n+1)h
Kinit = BVia Ve =
/ Vo(r)dr = vo(nh + h/2)h, nh
(n+1)h 7
Finio = Va
T—nh
pio
A Vo(t)dr = svo(nh +h/2)h, ———
nh (n+1)h
Ree Eve
‘i
/ (=") vo(r)dr & sv0(nh +h/2)h nh Kan
gv= Kingoy
Kigist= Kan, 135
(n+1)h Kano = EV inVin ees
:
/ (=")
Vo(t)dt = Zvo(nh +h/2)h,
nh Kinj31 = Kinz,
Kin,32 = Kin,23;
(n+1)h
=n) vo(r)dr © =vo(nh 1 / (=") + h/2)h
Ki,3s2 BYE
nh
(5.5.46)
In practice it is expedient to approximate the stochastic integrals Vie Vn in
(5.5.41), (5.5.42) (¢ = 2,..., N+1) by means of the following analog of the integral theorem about the mean for the stochastic integrals: (n+1)h Rh Vin =
/
a
ts
adh, h
‘ 0 LT
~w
nh
(24)
V;,~1(T) dt
n (n+1)h
f°
%-1(r)dr
nh
1 XA Piha 8 5APi-1n
= 51 Vin,
(5.5.47)
476
Chapter 5 General Theory of Stochastic Systems
(n+1)h (n+1)h
7]
S
Vin
(
i
‘i
er
0 h
~
) dP;_,(7)
y,_4(r)dr
(th)?
n (n+1)h
=.
ui yj,-1(7)dr
nh
1 RAR dn A
1 sem
ore EL
(5.5.48)
5.5.6. Analysis of More Precise Equations At the statistical simulation of a system by means of Eq. (5.5.44) it is also not
difficult to simulate the random variables distributed according to the normal and
the Poisson laws.
It is easy to see that the right-hand side of difference equation
(5.5.44) is determined with the accuracy till h? in the determinate (at a given vay item Yn Oy) and with the accuracy till h3/2 in the random item Pal Ya VP) Vp. While deriving Eq. (5.5.44) two small errors took place. Firstly, while replacing ye by the variable ys + (r = nh)AY, /h we substituted the random functions b(¥;; T) and Cit)
independent
of dWo(r)
and dP?(r)
functions depending on the random parameter AY,
by the nonrandom
which depends on the values
dW (rT) and d POG) at the interval (nh, (n + 1)h). Secondly, at ci(y,t) ze 0 even at one 2 the realizations of the random process Y(t) have the discontinuities of the first kind in the random points despite of m.s.
continuity of rte): Therefore
strictly speaking we cannot perform the linear interpolation of a given process. The first error may be removed by two ways.
The first one consists in the replacement
of the interpolation of the process Wah) by its extrapolation what is equivalent to
the substitution of AY, by the variable AY,,_1 in the obtained expression for Y;. But this way will lead to the appearance of the variables Youn and Age in the right-hand side of a difference equation,
i.e.
to the substitution
of the first order by the difference equation of the second order.
of the equation
The second way
consists in the abandonment of the interpolation of the process Y (t) at the interval
and in the direct expression of the increments of the functions a(Y;,, uF), b(Y,, iT),
G (Wen T) at small interval (nh, (n =e 1)h) according to the generalized Ité formula with the substitution the differentials by the increments. At this method the second error is also removed.
But the difference equation obtained in such a way will be
more complicated. The random variables which represent two-fold integrals over the components of the Wiener process W(t) =
Wo (t) and over the Poisson processes
will be included into this equation:
(n+1)h +
/ /dW;(o)aW;(r),
nh
nh
(n41)h 7
/ hsdP;(0)dP9(r),
nh
nh
5.5. Numerical Integration of Stochastic Differential Equation (n+1)h
f.
(n+1)h
/ /dP? (o)dW;(r), nh
7
u /dP?(c)dW,(r).
nh
nh
477
(5.5.49)
nh
It is very difficult to find the distributions of these random variables and only the
first two of them are easily calculated at 7 = 12:
(n+1)h +
famcoyarn(ny= Abin =val +b/ayh ,
2
a
*
:
nh
nh
(n+1)h
nh
,
7
/BEN a cay ree
(5.5.50)
nh
As regards to the second error it cannot essentially influence on the result as the probability of the occurrence of a jump of the Poisson process on sufficiently small
interval (nh, (n + 1)A) is infinitesimal. The accuracy of the approximation of a stochastic differential equation by a difference one may
be increased later on.
In particular, in one of the ways it is
sufficient to express a(¥ry Fy b(Y,, im) ey
T) at the interval (nh, (n + 1)h)
by the It6 integral formula: T
aly, , n= a(Faynh) + fay(¥s,8) + a2(Ye,3)" a(Y;,5) nh
N
1
di
— Soci(¥o, w(s)) + 5ayy (Yor) : 01%, 8)ds+ fay(¥.,3)" iz}
nh N
cit
xb(Ye,8)dWals)-+ ¥>[La(¥s+ex(¥e,8), 8) -a(¥s,8) ]4PM(). (5.5.51) dag
Formulae for b(Y;, T), ron eas T) are analogous.
To the integrals obtained we may
then to apply the same method which was used for the calculation of the integral in (5.5.11).
As a result we shall obtain the right-hand side of the difference equation
with the accuracy till h® in the determinate (at a given 5) item and h®/? in the random item.
The process of the more precise definition of the difference equation
correspondent to a given stochastic differential equation may be continued. But each new correction requires the existence of the derivatives of the functions a, b, c; of more higher orders.
478
Chapter 5 General Theory of Stochastic Systems
We may use another method for precising the difference equation. Namely, the
integrands in (5.5.51) and in analogous formulae for b(Y;, Fy) GLYs; T) we may express by means
of the generalized Ité formula substituting in it the differentials
by the increments.
Here the three-fold integrals over the components of the Wiener
process Wo (t) and the Poisson processes P; (t) will enter into the difference equation. For further correction of the approximation the stochastic differential equation by the
difference one in this case the integrands in (5.5.51) and in the correspondent formulae for b(Y, . T) Ci ee 5 T) in its turn should be presented by the integral It6 formula and
later on may be used the differential It6 formula with the change of the differentials by the increments.
This process may
be continued further and as a result it will
lead to the presentation of the process Y (t) at the interval (nh, (n + 1)h) by the stochastic analogs of Taylor formula.
In this case the multiple stochastic integrals
over the Wiener process Wo (t) and over the Poisson processes P; (t) will enter into the difference equation. insuperable difficulty. same
component
The finding of a distribution of these integrals represents And only the integrals of any multiplicity over one and the
of the Wiener process
Wo (t) or over one and the same
Poisson
process P; (t) are calculated very simply.
In order to avoid the calculations of the derivatives of the functions aes fh), Oe
a) G; (Ya ; T) while using two stated ways of the approximation of a stochastic
differential equation by a difference one, we may recommend
to substitute them by
the retios of the finite increments, for instance, at the interval (nh, (n + 1)h) over
the variable tf and at the intervals (Yi ; Ynk ++ Gn
Yee nh)h) over the components
of the vector y.
5.8.7. Strong and Weak Approximations rential Equations
of Nonlinear Stochastic Diffe-
The obtained in Subsections 5.5.2—5.5.6 difference equations may be used both
for the theoretical research and for the numerical integration of the stochastic differential equations. In both cases we must know the distribution of all random variables which enter into the difference equations. In a given case the difference equations will represent the so-called a strong approximation
of the stochastic differential equations.
While numerical integrating such an approximation is necessary if is required to obtain the realizations of the process NACE
But often there is no need to obtain the
realizations of the process but it is sufficient only to have the estimates of the moments or of the expectations of some functions of the value at some instant of the
random process Y(t). In such cases we may give up the use of the exact distributions entering into the difference equations of the random variables and substitute them by some
more simple distributions with the same moment
characteristics.
At the
5.6. Multi-Dimensional Distribufions in Stochastic Differential Systems
substitution of the random
479
variables by the variables with other distributions the
difference equation will represent
a weak approximation
of a stochastic differential
equation. For the stationary
and nonstationary
stochastic
linear differential equations
standard numerical methods of the ordinary differential equations may be used.
Systematic presentation of the modern
numerical methods
for the nonlinear
stochastic differential equations integration based on a strong and a weak approximation and different classes of the numerical schemes are considered for, example, in
(Artemiev and Averina 1997, Kloeden and Platen 1994, 1997, 1999).
5.6. Multi-Dimensional Distributions in Stochastic Differential Systems 5.6.1.
One-Dimensional
Characteristic Function
Let us consider the system whose state vector (in the general case the extended state vector) is described by the Ito stochastic differential
Eq. (5.4.6) where V is a white noise in the strict sense (Subsection 4.6.4). The problem is to find all the multi-dimensional distributions of the state vector of the system Y(t), supposing that the one-dimensional distribution (and consequently, all the multi-dimensional distributions (Subsection 4.6.3) of the process with the independent increments
W(t) = W(to)+/V(r)dr
(5.6.1)
is known. > In order to find the one-dimensional characteristic function of the random
process Y (t) let us consider two time instants t and t+ At. According to the definition of a characteristic function (Subsection 3.5.1) the values of the one- dimensional characteristic fucntion gj (A; t) of the process Y (t) at the instants t and t + At are expressed by
gi
Xt} Se OO Gis te At) = el OA), ante
Ca waehg
(5.6.2)
Subtracting the first formula (5.6.2) from the second one termwise we shall have
gi(A;t + At) — 91(A;t) eA
| ef Y(eFapn ae,
E Ei ane. + 1 eT Y(t).
(5.6.3)
480
Chapter 5 General Theory of Stochastic Systems
But it follows from Eq. (5.4.6) and Eq. (5.6.1) that with accuracy up to the infinite-
simals of higher order AY (t)
=Y(t+ A) — Y(t) = a(Y(t), t)At+ b((Y(t), t)AW, = (5.6.4)
or omitting the argument t of the function Y(t) in (5.6.4) we get AY (t) = a(Y,t)At+b(Y,t)AW,
AW=W(t+A)-—W(t).
(5.6.5)
Substituting (5.6.5) into (5.6.3) we find
nist + At) — 9102) = B Le?EONarvonnaw] _ 1)7Y om Bie
ee
_ eiAT (Yt) AW
eid" WY t)AW sl joyee or with accuracy up to the infinitesimals of higher order relative to At
gi(A;t + At) — gi(A;t) = RCL
ian
P57 {er Maw
ENT aly, t)At 4 eiATH(Y AW _ i}ees
BS
a 1) iseid (YY,t)AW _ i
(5.6.6) For calculating the expectation in (5.6.6) let us use the formula of total expec-
tation (Subsection 3.4.2) taking first the conditional expectation at fixed value y of the random
variable
Y , and then taking the expectation
randomness of Y. As in the Ité Eq.
AW
Y =
of the
Y(t) and
= Wit + At) — W(t) are independent, the conditional expectation of any
function of the random variable AW
at a given y does not depend on Y and is equal
to the unconditional expectation of this function.
Be
with the account
(5.4.6) the random variables
HAW
Consequently,
Tay, t)eӴ hog [ema
S neta
SH {ixTa(¥, t)e"Y B eae eA
Waa
|y |
|y|}
nely [ia 16 $a Aged camel eee
(5.6.7)
and similarly
Ee UYt)AW4ia7Y
_ pp [arora
og
AeA) )
(5.6.8)
5.6. Multi-Dimensional Distributions in Stochastic Differential Systems
the expectation in brackets being taken at fixed Y. characteristic function of the increment
AW
ay But He'# AW
of the process
481
represents the
W(t) which we shall
denote by h(p; t,ti+ At):
Belt4W — h(yt t+ At). Therefore
(5.6.9)
ee
Bei MOO4W — (bY, 1) Aste
Ady
(5.6.10)
and formulae (5.6.7) and (5.6.8) take the form
Be HAW AT ALY,the Y = BidT a, t)h(O(Y, t)? A;t, t+ Ate’”, Bed YNAW4HATY _ PAY, t)T A;t,t + At)e?’”.
(5.6.11) (5.6.12)
Substituting expressions (5.6.11) and (5.6.12) into (5.6.6) we obtain
gi (A;t + At) — gi(A;t) = E{ir7 h(b(Y,t)7 A; t,t + At)At
AO t) Ate
Ale We*.
(5.6.13)
Now we use formula (4.6.10) determining the characteristic function of the increment of a process with independent increments in terms of its one-dimensional characteristic function.
Then denoting by hy (p; t) the one-dimensional characteristic function
of the process W(t) we shall have
h(u;t At) — hi (ust) A TA Dnteg UTA: ed oc es le+ At) eeea eehi(pyt at +Ee ee EiAD hi (pt) hi (p51) ( Supposing that h, (yu; t) has the continuous time derivative and using the Lagrange
finite increments formula we find from (5.6.14)
_ Ohi (457) where T € (igt + At). Consequently, (5.6.14) is equal to
h(w;3t+ At)—1l=
1
Ohi (u57)
hi(u;t)
ot
=
or with the accuracy up to the infinitesimals of higher orders relative to At
ag hA(ust+ At)—1
ee
1
CAT
Ohy (p15 t)
a DEL
At,
6.1 (5.6.16)
482
Chapter 5 General Theory of Stochastic Systems
Putting
Ah
-t)=
1
1(4;t —_
5.6.17
we find for (5.6.16)
h(p;t + At) —1 = x(p;t)At.
(5.6.18)
Substituting (5.6.18) into (5.6.13) we get with the accuracy up to the infinitesimals
of higher order relative to At
gi(A;t + At) — 91 (A;t) = EB {irT a(Y, t) + x(8(Y, be t)} eV
Ag.
(5.6.19)
Finally, dividing both parts of this formula by At and passing to the limit at At —
0
we obtain the equation which determines the one-dimensional characteristic function
gi(A; t) of the state vector Y = Y(t):
cee = B{irTal¥,t) + x(W(¥,t)7 Ast}
Ya
(6.6.20)
The expectation in the right-hand side of (5.6.20) is determined by the one-dimensional distribution of the process Y(t) which in turn is completely determined by its one-dimensional characteristic function
gi(A;t). Consequently, the right-hand side of formula (5.6.20) at a given t represents a functional of the characteristic function g;(A;t). Therefore
(5.6.20) represents the equation determining g1();t). Let Yo = Y (to) be the initial value of the state vector of the system at the instant to representing a random variable independent of the values of the white noise V(t) at t > to. Denote the characteristic function
of the random variable Yo by go(A). Then the initial condition for Eq. (5.6.20) will be 91(A; to) = go(A). (5.6.21) Eq. (5.6.20) and the inittal condition (5.6.21) determine completely and uniquely the one-dimensional characteristic function gi(A;t) of the state vector of the system (5.4.6) at any time instant t > to. Eq. (5.6.20) was first obtained at the begining of the 1940s by Pugachev (Pugachev 1944) who apparently was the first who studied stochastic differential equations with an arbitrary process with the independent increments. Bernstein, It6 and Gukhman studied only the stochastic differential equations with the Wiener processes (Bernstein
1934, 1938, Ito 1951a, b, Gikhman 1947, 1950a, b).
5.6. Multi-Dimensional
Remark
1.
Distributions in Stochastic Differential Systems
It goes without saying that Eq.
483
(5.6.20) is valid only if the
expectation in the right-hand side exists. It follows from our derivation of Eq. (5.6.20)
that in this case g1(A; t) is differentiable with respect to t and satisfies Eq. (5.6.20). Some sufficient conditions of the existence of the one-dimensional distributions are
given in Subsection 5.6.12.
Remark It6 equation.
2. We emphasize that Eq. (5.6.20) is valid only if Eq. (5.4.6) is an Only under this condition the instantaneous value of the state vector
of the system Y (t) is independent of the increment AW
= W(t + At) _ W(t) of oT
the process Wit), and the conditional expectation of the random variable e’4 relative to Y coincides with its unconditional expectation.
deriving formula (5.6.13).
5.6.2.
ca
We used this fact while
- Pa
Multi-Dimenstonal
Characteristic Functions
Analogously the equations determining other multi-dimensional characteristic functions of the state vector of a system are derived. > Let
ga (ia
Renae erik
depen ta.) =e
i> MY (te)
.
(5.6.22)
lk
be the n-dimensional characteristic function of the state vector Y of
asystem. Taking
two values t,, and t,, + At of the last argument t,, at fixed values of all preceding arguments tj
CPM’s p14 and j1p of the stochastic systems A and B connected in parallel
(Fig.1.6.1) determine a conditional probability measure in the direct product of the spaces Y X Z of the outputs of these systems.
This measure is expressed by formula
Pp Ole ii}pa(dy |2)up(dz |2) = /yn(Ey |#)pa(dy |2), €
where € € Bx
C, fy =
y
{z
(5.8.1)
: (y, 2) E E} is the section of the set E at the
point y, B, C are o-algebras of the sets in the spaces Y and Z correspondingly.
The
integration over Y in the latter integral is performed practically over the projection E on the set Y as gy # @
only in the case when y belongs to this projection. Analogously
CMP’s
14,
Bp
of the systems
Systems
511
A, B connected
sequentially
(Fig.1.6.2) are determined. The conditional probability measure is determined in the direct product
Y X Z. This measure is given by formula
Haan |x)= ec ss
|a,y)ma(dy|z),
EE BxXC.
(5.8.2)
A special case of the sequential connection (Fig.1.6.2) presented on Fig.1.6.3 is also
stochastic system (we denote it by C’) whose CPM —
is obtained from (5.8.3) at €
Seale 5 Fee: 7
no(F|2)= fua(Pleyualdyle),
FEC. 2
(5.83)
Ye
Theorem 5.8.2. The sequential connection of the stochastic systems A and B represents the stochastic system C whose CPM wc 1s
determined by formula
(5.8.3).
> Let us consider the closure [A] of the stochastic system by elementary feedback
(Fig.1.6.4a,b).
Let pa(G
|Ab. y), G € B be CPM
of the system A. Suppose that
there exists such measurable mapping Ty : Y — Y
of the space Y into itself that
the measure [J 4 (FF 1G |x, y) does not depend on y. This latter measure if it exists we shall denote by pu(G |r) then the identity is valid
MG
caja
Gled),
Geb,
(@yyeX
x.
(5.8.4)
We put here y = yo. If Ty, is a convertible transformation then
H(Ty,G |x) = pa(G| 2,40),
GEB,
where {14(G | 2, yo) is a conditional probability measure of the output Y of the system [A] at a given
Z € X. Therefore CPM
of the system [A] may be naturally
determined by formula
pal(G |c)=(TysGd
2), Ge B.
(5.8.5)
This formula is reduced also to the form
mpay(G |) = pa(T,'Ty,G|2,y),
GEB,
(5.8.6)
512
Chapter 5 General Theory of Stochastic Systems
in consequence of (5.8.4). On the basis of formula
pa(Typ Ghay)= pa(Ly,'G |x, yo)
(5.8.7)
pa(G |2,9) = val), TG 2, yo),
(5.8.8)
or equivalently
which follows from (5.8.4) the right-hand side in (5.8.6) does not depend on the choice of yo EY.
4
Theorem 5.8.3. The closure of the stochastic system A by an elementary feedback represents the stochastic system [A] whose CPM pa} is determined by formula (5.8.6). The stated theorems establish the completeness of the class of the stochastic systems relatively to all possible typical connections (Sub-
section 1.6.2). Formulae (5.8.1), (5.8.2) and (5.8.6) give the principal opportunity to find CPM of any connections of the stochastic systems by given CPM’s of the connected systems. The established properties of a class of the stochastic systems permit to use the structural methods for study of the composed stochastic systems. After partitioning a composite system into the subsystems we may study the characteristics of the subsystems and further to study the behaviour of the system as a whole. The structural methods for the linear and the nonlinear stochastic systems will be considered in Chapters 6 and 7. For the stochastic composed systems reducible to the systems considered in Sections 5.6 and 5.7 their CPM’s are determined by the multi-dimensional distributions.
§.8.2. Distributions in Stochastic Systems with Random Structure Let us consider a system which is described by stochastic differential equation:
Y =a(Y, S(t),t)+0(Y, S(t), t)V (k=1,..., N), where
S =
(5.8.9)
S (t) be some step random process S (t) which describes the random
changes of a system structure.
Let 51, ...,
SN
be possible values of S(t).
At
§ (t) = S, the system is described by the It6 equation
Y =a(Y, sg, t) + 0(Y, sz,t)V (k=1, ..2, .N). We
consider
a case when
the transitions
(5.8.10)
of a system from one structure
to
another form the Poisson stream of the events. We denote the Poisson process of the
transitions into the k*® structure by Py (t) and its intensity by lz (Y, Si t) taking
5.8. Distributions in Stochastic Composed Systems
into account
that it depends on the random
the system state at the transition moment.
513
process S and may
also depend on
It easy to see that the process S(t) is
determined by the following It6 stochastic differential equation:
dS = \°(sx — S)dP, .
(5.8.11)
Supposing that while changing the structure of the system its state vector may receive a random increment by a jump we write the sought stochastic differential equation
of a system with randomly changed structure in the form “
N
dY =a(Y,S,t)dt + 0(Y,S,t)}dW +5
~dQz,
(5.8.12)
k=1
where Q
(t) is the general Poisson process with variable distribution of the jumps
generated by simple Poisson process Py (t) (k =s+lensd
§
N). The distribution of
each jump of the process Q, (t) may depend not only on time but on the state vector of the system Y and its structure at the transition moment
into the k*® structure.
Thus after adding to the state vector of the system Y the value of the process S determining its structure we shall get a system of It6 stochastic differential equa-
tions which describes the behaviour of a system with randomly changed structure. Naturally, the vector Y= The
equations
[ee ee \" will be its state vector.
of a system
with randomly
changing
structure
(5.8.11) at Poisson stream of the changes of the structure may
(5.8.10) and
be written in the
general form of the It6 equations
dY =a(Y,t)dt+b(Y,t)dW. =
Here
4 i!
Y =
[Ss ay
(5.8.13)
we
is the state vector of a system; W(t) is a process with indeTe
pendent increments which consists of NV + 1 independent blocks W(t), [P, OF | ;
wets
Tape
Wis
[ Pv Qh, | ; a(y, t), b(y, t) are block matrices equal to
a(j,t) = |Hoan || : = 47
(y, t)
b(y, s,t)
| sy; —s
0
| sv—s
0
|
I
Pty 0
ie
0
Using Eq. (5.8.12) we set up Eq. (5.6.7) for an one- dimensional characteristic
function of the state vector for the system with the randomly changing structure.
514
Chapter 5 General Theory of Stochastic Systems
For this purpose we find at first the function Xf; t) correspondent to the process W(t), ti —
[ fy... LN ee In accordance with the mentioned
above in Subsec-
tion 5.6.3 the function Y(/i; t) in this case represents the sum of the functions X correspondent to the independent blocks Wit), [Pi (t)Qi(t)7 |. 3a [Pn (t)Qn (ee which compose the process W(t): N
X(H; Y,S,t) = x(ust) +) xe(Hes Y,5,1),
(5.8.14)
k=,
where x(y;t), Xe (He; i Lahey t) are the functions Ox(t) Ele correspondingly (k gl
Lee
eee N); —
x of the processes. W(t), [Pi.(t) [ue?at ae pt," is the parti-
tion of the vector [Z into the correspondent blocks; /l, =
[ko pee is the expansion
of the vector [1x into the blocks correspondent to the processes Pj, (t) and Qx(t). Here we took into consideration that the intensity of the streams of the jumps of the processes Py (t) and the distributions of the jumps of the processes Qx(t) may
depend on the state vector of the system yi= [S bed es Using formula (5.6.8) for the functions VY and the formula for the characteristic function of the increments of the Poisson process and of the general Poisson process which is generated by it at
infinitesimal interval (t, s] according to which hy (jax; Y, S,t)
= 1+ [e'**9 9,(un; Y,S,t) — 1] (Y,S,t)(s —t) + o(s—t).
(5.8.15)
We find from (5.8.15)
niu(ns ¥o(Syt) = [Le ox(ase
Gib
Ly] WEG St)5
(5.8.16)
where gx (MK3 Yy, S, t) is a characteristic function of the jumps of the process Q, (t)at
the instant t. Thus (5.8.14) takes the form N
X(H; Y,S,t) =x(ujt)+
>> [et**°gn(ue; Y,S,t) — 1] 2 (Y, S,t). (5.8.17) k=1
Therefore by virtue of (5.8.17) we have
LOW, St) AVY Sity =v (00, Se) Ne) N
Pye |elm Seg, (d; vigays 1 v(Y, S,t), iil
(5.8.18)
5.8. Distributions in Stochastic Composed Systems
where
\ =
515
[Xo dite is the partition of the vector A into the blocks correspondent
to the blocks S, Y of the state vector of the system Y.
Substituting (5.8.18) into
(5.6.7) we obtain the equation for the one-dimensional characteristic function of the
process in the system with the random changes of the structure:
Ogi (A; t) =
B{AinTaly, S,t) + x(0(Y, S,t)? d;t)
N
+50 iene darren S,t)- 1 »Z(Y, guy k=i
Basing on Eqs. may
derive
ciatbiad .
(5.819)
-a
(5.8.19) for the one-dimensional
the equations
for the conditional
characteristic function we
characteristic
functions
ferent structures and for the probabilities of these structures.
in the dif-
Taking into account
that S(t) is a discrete random variable with possible values $1 , ...
, SN
and the
probabilities of these values p;(t), ... , p(t) at each t we find by formula of the total expectation N
Get) = BeoS(e) Ha" ¥(#) — Spe
tg (Vt |1),
(5.8.20)
Rl
where 91 (A; t |S1) is a conditional characteristic function of the state vector of the system Y in the [th structure.
Analogously an expectation in the right-hand of Eq.
(5.8.19) for 91 (A; t) is calculated. As a result the equation will take the form
N Sefton Opigi(A;t |81) 1=1
N = » opie 2
ot
[{aray, sj,t) + x(O(Y, si, t)" A; t)} ed ¥ |si |
f=
N
N
Le:
+50 peo" yay |efalen=") l=1 k=1 N
—- Soo l=1
ge (2;Y, 51, t)ve(Y, si, te" ell si|
N
seu VEY, Si, fe
¥ |sre}.
(5.8.21)
k=!
Here the index (1) at the sums over k shows that the item correspondent to k =
does not enter into the sum.
|
It does not enter into the sum as the processes P} (t)
516
Chapter 5 General Theory of Stochastic Systems
and Q (t) maintain the constant values at the absence of the change of the structure
as a result of which gi(A; Y} Sig t) = 1. After changing the order of the summation
in the first double sum (5.8.21) we get N
N
4
-\T
De
Spe |es, sr, tualY,s1,te™"Y |si].
Loi
(fs
(5.8.22)
After replacing in (5.8.22) the indexes k and | (the,double sum will not change in this case) and substituting the obtained expression into Eq. (5.8.21) we shall have
N s eirosi Opigi(A;t |s1) Ot
feat
N
= > pie
EB[{at acy, si,t) + x(O(Y, si, t)" A;t)} ei¥ |si|
N N +o peo SpE a
[ms ¥, se, t)a(Y, se, tye¥ |sx|
k= N
~ So peo e [u(¥, t)e"Y |si |.
(5.8.23)
=i
N
where for brevity is assumed i(Y, t) =
LOWY,
Sint): Comparing the coeffi-
[geil
cients at similar exponential functions and denoting by fily;t
| S1) a conditional
one-dimensional density of the process Y(t) in the [* structure we set up the system
of the equations from Eq. (5.8.23)
0 eS r a e
. ii;{id? a(y, s1,t)
+x(b(y, 81,t)? A; the" }fa(yst |sr)dy N
+) ka
co
pe / H(A;Y,Sk, t)M(y, se, te
—pI / u(y, te —oo
“ fi(yst |se)dy
=
Y fr(y;t bspldgh (i= dead) ME
(5.8.24)
5.8. Distributions in Stochastic Composed Systems
Putting in Eq. =
(5.8.24)
4 =
517
°0 and taking into consideration that gi(0;t
| 51)
Is x (0; t) = 0 we get the equations for the probabilities of the structures
=1,...,
(1
N) of the system at the moment t:
N
CO
Cpl
co
1)
py / Vi(y, Skt) fi(yst |se )dy — pr / ily, t)fr(yst; |si)dy.
p=>
;
k=1
—0o
—0o
:
(5.8.25)
Using Eqs. (5.8.24) we derive the equations for the conditional densities from the equations for the conditional characteristic functions in the case of the Wiener process W(t).
For this purpose following Subsection 5.6.6 we change the denotion
of the integration variable y by 7, multiply the equation for pigi(A;t
| s}) by
eid" y /(20 yp and integrate it over \. The first integral is transformed similarly as in Subsection 5.6.6 and gives as a result an item which is analogous to the right-hand
side of Eq. (5.6.42). The second integral as a result of formula
jad
.
a (2x)P | (Asn, ses the "dA =a(y—nin.set),
(5.8.26)
where gi(Y; 1), Sk, t) is the density of the jumps of the process (;,(t) at given 1), Sk and t is transformed to the form foe)
/ qi(y — 737, sk, t)(n, sk, t)fi(n;t | sz )dn.
(5.8.27)
—oo
The
third
integral
in Eq.
(5.8.23)
ny, t)fi (y;t |s1)3 As a result for!
as
a result
of (5.6.37)
takes
the
form
= 1,..., N we have by virtue of (5.8.27)
Opifily;t OF [aly sr HACE | sr) ] _ |sr) = —Pr5y ie
+e tr laa [b(y, s1, t)u(t)b(y, sx, t)” frly;t |s)|} N
#0? si
co
fay — 131, 8k, t)™i(n, sx, t)fi(n;t| se)dn — pira(y;t) fi(yst| 51). ne
(5.8.28)
518
Chapter 5 General Theory of Stochastic Systems
In important
special case where at the transition from one structure
other the state vector of the system Y does not receive a random items with dQr in the equation for Y are absent (Qk (t) =
to an-
increment
the
0).Here in the equa-
tion for the one-dimensional characteristic function gi(A; t) and the equations for the conditional one-dimensional
characteristic functions gi(A;t
| S1) at different
structures we have gil A; Y, x,t) = 1. Correspondingly in the equations for the conditional one-dimensional densities fj (y;t |$1) at the different structures we get
gi(y—n; 7, sk,t) = 6(y—7) and the equations for Firlost:| ay)
is hea)
take the form
a) e it ar Lalu, sist Se ae) = —prg, ) fa(yst |81)] T
+5 tr las [d(y, si, t)v(t)b(y, 81, t)? fa(y;t | =|}
N +5 WOu(y, se,t) [pefilyst |se) — prfi(yst |1)] (=1,..., N).
=
(5.8.29)
5.9. Distributions in Stochastic Infinite-Dimensional Systems 5.9.1. Distributions in Stochastic Continuous Hilbert Systems
At first let us consider the random process
Y = Y(t) with the
values in separable H-space Y which is determined by the It6 stochastic differential equation of the form:
dY =a(Y,t)dt + b(Y,t)dW, Y(to) = Yo.
(5.9.1)
Here W = W(t) is a random process with the independent increments in some other H-space from W to Y, a(Y,t) is an operator mapping Y x R into Y, b(Y,t) is an operator-valued function mapping Y x R into the space of the linear operators which act from W to Y, Yo is a random variable in the space Y independent of the future increments of the process W. The central problem of stochastic systems theory consists in finding the multi-dimensional functionals of the process Y: n
Gio iis os
Pen tig
= os
git
=e
OX
04,74)
ie
(n 21, Ios.)
(5.9.2)
5.9. Distributions in Stochastic Infinite-Dimensional Systems
519
Here 41, ..., An are the elements of the adjoint space )*, (Ax, Yt,) is a scalar product of the elements A, and Y;, in H-space Y. The multidimensional functionals gn = gn(A1,..., Anjti, ..., tn) completely de-
termine at the known conditions a distribution of a process (Section 3.6). The stated problem is easily solved while performing the following conditions. (i) The space Y of the values of the process Y(t) at each ¢ is a separable H-space.
(ii) The random process W(t) represents the Wiener process with the values in the separable H-space W whose covariance operator of the value at each given t is determined by formula t
FSi tt
fear,
(5.9.3)
to
where v(t) is the intensity of the process W(t) which represents at each t the trace-class adjoint operator.
(iii) The function a(Y,t) with the values in the H-space Y is determined by formula
a(Y,t) =a,Y + a'(Y,t).
(5.9.4)
Here a, being a closed linear operator which is an infinitesimally generating operator of strongly continuous semigroup of the restricted oper-
ators {u(t)} satisfying the condition ||u(t)|| < ce~7 at some c, y > 0 and equation
dujat= au,
uo) = 1;
(5.9.5)
where J is an unit operator, a/(Y,t) is the strongly differentiable function over Y and ¢ with the values in the H-space J.
(iv) The operator-valued function b(Y,t¢) at each Y and ¢ represents such compact linear operator that the self-adjoint operator
o(Y,t,7)
= u(é—r)o(Y,t)u(t—7)",
of ¥,t) = 0(Y,t)v(@)0(Y, 1)", (5.9.6) is trace-class and at each t the following conditions is fulfilled:
t / tral
Y,t,e)dr-< co.
(5.9.7)
to
The representability of the function a(Y, t) in Eq. (5.9.1) in the form of (5.9.4) with the restrictions on aj, a’(Y,t) and condition (5.9.7) are
520
Chapter 5 General Theory of Stochastic Systems
sufficient for the existence of the solution of the Ito stochastic integral equation
t
t
Y(t) = u(t)Yo + ju —r)a'(Y(r),7)dr+ ju —T)b(Y(r),7)dW, to
to
(5.9.8) and for the existence and the integrability of the operator of the second order moment. The property of the operators v(t), o(Y,t), v(Y,t, 7) to be trace-class is necessary and sufficient for v(t)At, o(Y,t)At, v(Y,t, r)Ar be the covariance operators of the normally distributed random variab-
les AW, b(Y,t)AW(t) and u(t — 7)b(Y,7)AW(r) for the infinitesimals At and Ar.
Let us show that while performing conditions (ii)-(iv) the multidimensional characteristic functionals (5.9.2) are determined by equations Ogn(A1
>) ans
9 Agel;
relied
ta)
=
Bfivan a(¥
ta)
ate
1
—5 (ny (Yinstn)) An)exp 2
ie
iS Ok, Yi) \Gah
cD ged)
k=
(5.9.9) and the initial conditions
91(A1;0) = go(A1), gn(A1, nue Pe A
es ERE,
(5.9.10)
24) = gn—1(A1, sae An—-1+An3 #1, i
a Cet)
:
(5.9.11) where go(A1) is the characteristic functional of the random variable Yo. > In fact it is sufficient to use Ité formula (5.1.15) at a given 7 for the calculation n
of the stochastic differential of the composite function exo]i520, Yi) which k=)
is considered as the function of a scalar random process
Ni Site: ne ei i Nee 1
ee
Yaw
Oss Y(tn)) at the fixed
and to use the formula of the transformation of
a covariance operator in linear spaces (Subsection 3.3.2).
0. Example
5.9.1.
Let us consider
a stochastic system described
by a
nonlinear wave stochastic differential equation
@U
o?U
GU Ts nla?
re = c7(1 =8 Vio + 2h; reel) — gh (3)
with the boundary conditions U(0,t) = U (2154) '==10. oHere =
Ce, + Vo
(I)
es U(x, t); Vi,2
Vi ,2(2, t) are the independent normally distributed white noises with the inten-
sities
Yj 2 = V19(X1 , 29, t); (et hi 2 afe constant coefficients. After putting
¥,=U, ¥,=au/at,
Y=[Yi Yo] ,
A
©
ARCA Sardar
enme |
0
ah
0
ie in |ezary, /da? | we reduce Eq.
(I) to the form of Eq. (5.9.1). Eq.
(5.9.9) for the one-dimensional
characteristic functional gj = gj (Ay , AQ} t) will take the following form:
Ogi(A1,
ee
A23t
=
:
BYvas, ¥e)
+8
;
502%)
(22.0%Tae) ae
2 + 2h, Yo —- zhi?
15(da,v2d2)] Run Vi VatiOa’ ay \" )tiQr2, s , oY o (00,(Sa at) (52) Aa) = ie)
-5
(IIT)
where a scalar product is given by the formula (y, w) = i)p(x)(x)dx : 0
Along with Eq. (5.9.1) there are the stochastic differential equations in the separable H-space Y of the following form:
dY =a(Y,t)dt + b(Y,t)dWo + /c(Y,t,u)P%(dt,du).
(5.9.12)
Ro Here Wo = Wo(t) is the Wiener process with the values in the H-space Wo, P°(A,€) is the centered Poisson measure, A = (t1,¢2) is some
522
Chapter 5 General Theory of Stochastic Systems
time interval, € is some Borel set of the space Rj representing the qdimensional space with pricked origin, a = a(Y,t) and c = c(Y,t, wu) are
the functions mapping Y x R and Y x Rx Ré into J, b = 0(Y,t) is an operator-valued function mapping Y x R into the space of the linear operators which act from Wo to Y. We shall assume that the initial value Yo is an element of the space Y and does not depend on the increments of the Wiener process Wp and the values of the Poisson measure P°(A, €) at the time interval A = (1, ta], to < ti < te for every E. In this case similarly as in the previous case it is shown that multidimensional characteristic functionals (5.9.2) satisfy the equations AG aCAyh, eas
toied bra eareee)
at, +x(An, Yt,j3tn)] exp Fon %) k=1
= BYLin,al ¥i.5 ta) \(ni BQyees)
(5.9.13)
with initial conditions (5.9.10) and (5.9.11) where 1
x(A, 95 t) = —5Q, ool, 4) + {efvely.tu) Lae uA, c(y,t,u)) yp(t, du),
(5.9.14)
R?
oo(y,t) = b(y, t)vo(t)b(y, t)*, vo is the intensity of the process Wo which represents the trace-class operator, b(y,t)* is an operator adjoint with b(y,t), vp(t, du) is the intensity of the simple Poisson process. The obtained results may be generalized on the stochastic conti-
nuous systems in B-spaces with a basis (Pugachev and Sinitsyn 2000). 5.9.2. Distributions in Markov Stochastic Discrete Systems
Let us obtain the equations of the discrete Markov systems which are important in applications. For this purpose we consider the random
sequence {Y;,} with the values in some measurable space (Y,B) which satisfies the following stochastic difference equation: Yead
=e
Yin We)
bis
he
ee
)
(5.9.15)
where {V,} is the sequence of the independent random variables with the values in other measurable space (V,€); we(yz, vz) are (B x E, B)-
5.10. Applications.
Optimal Online Filtering
523
measurable functions mapping Y x V into J. Let the initial value Y; be
independent of the sequence {Vi}; 41(B), B € B is the distribution of Yi; vx(£), E € E is the distribution of V;. In this case {Y;,} is a Markov
sequence. Consequently (Subsection 2.7.9), its multi-dimensional distributions are completely determined by the initial distribution p(B),
B €B
and by the transition distribution y,%41(B |Y¥,), B € B and
wesa(B |y) = ve(we*(y, BY),
(5.9.16)
where we (y, B) € V is the inverse image of the set B € Y which is determined by the function w,(y,v) at the fixed y. Recurrent Eq. (5.9.15)
at k = 1,2,... and given initial distribution 1(B) permit to find the distributions of the sequence {Y;}. . 5.10.
Applications.
Optimal Online Filtering
5.10.1. Introductory Remarks Problems of determining the state of a system using the results of some measurements often arise in practice.
As the measurements
are always accompanied by
random errors one should speak not about the determination of the system state but about its estimation (filtering, extrapolation, interpolation) by statistical processing
of the results of measurements. At practice a filtering problem is technically solved by means
of passing the
measured signal Paks: through the device which is called a filter for “filtering off’ the noise and for obtaining a signal at the output which reproduces the required process Xx (t) with possibly great accuracy.
Let us consider the continuous stochastic system described by Eq. (5.4.6)
X = 9(X,t) + V(X, t)V where
X
is the n-dimensional
state vector of the system,
(5.10.1) V is the r-dimensio-
nal vector white normal noise, and y(z, ae v(a, t) are known functions of the system
state and time. The values of the function p(x, t) are evidently the n-dimensional vectors, and the values of the function v(a, t) are the N X T matrices. If the state vector
of the system
X
is continuously
measured
then
the n-
dimensional random process Y (t) =—¢ (t) +U (t) will be the result of measurements where UV (t) is the measurement error which represents usually a random function of
time.
However, it is not the components of the state vector, but some functions of
the state vector are usually measured among which may be some of the components
524
Chapter 5 General Theory of Stochastic Systems
of the state vector. The result of the measurements is determined in the general case by the formula
Y =Y(t) = 0(X,U,t)
(5.10.2)
where Y is the 1-dimensional vector, U is the measurement
error representing a
vector random function of time of dimension 7 > 1, and Yo(Z, u, t) is the known function of the system state, measurement
error and time.
In the general case the
function (9 is nonlinear both relative to the state vector and to the measurement error. The dimension of the error vector U in problems of practice cannot be smaller than the number of measured variables 7; as each measurement
is accompanied by
an error and the errors of various instruments are usually independent
(or at least
contain independent items). In some cases the measurements are performed by instruments possessing inertia or the results of measurements may be filtered by various filters. So the general model of measurements which are performed in a system may be described by the differential equation
Y = 91(Y,X,U,t). The result of measurements
(5.10.3)
represent the random process Y (@): We come
the problem of filtering of the state vector of
asystem X at eachinstant
thus to
t > to using
the results of continuous measurement of the process Y determined by Eq. (5.10.3)
in the time interval [to, t]. Remark.
We emphasize that such a statement of the filtering problem is
adequate only if online data processing, i.e. data processing in the real time basis is possible.
If online data processing is impossible and one has to be content with
post data processing when the measurement is finished, then the above statement of
the filtering problem has no sense.
The estimates will always be better if the whole
measurement is used to estimate the state vector of a system at any instant t during the experiment.
Therefore, other techniques for solving filtering problem must be
used in such cases (Subsection 1.6.5). Thus we come to the exact statement of the filtering problem. A vector random process
sare
ie 1s determined
by the Ité stochastic dif-
ferential equations
dY = 9,(Y,X,t)dt + di (Y, X,t)dW, dX = 9(Y, X,t)dt + ¥(Y, X,t)dW, where Y is a N1-dimensional
random
process,
X
is a n-dimensional
(5.10.4) (5.10.5) process,
W
is a N-dimensional process, v1(y, azyt) and ply, x,t) are known vector functions mapping
the space R™!
x R”
x R into the spaces R™!
and R”™ respectively,
ply, AD t) and vy, ape t) are the known matriz functions mapping R"!
and
xXR” x R
5.10. Applications.
into R™"
and R""
respectively.
of the system) at any instantt
the process
5.10.2.
Optimal Online Filtering
525
It is required to filter the process (the state vector
> to using the results of continuous measurement
of
Y in the time interval [to, t].
General Formula for Optimal Filtering
It is natural to require that the filtering of the process some sense.
Y(t) be optimal in
The mean square error (m.s.e.) criterion B|X, — X; l=
min serves
as a natural criterion of optimality in many problems of mathematical statistics.
If
we assume this criterion, then the general solution of the filtering problem follows directly from the known property of the second order moments:
second order moments
of a scalar random variable is its variance.
the smallest of all
Hence it follows
that the best approximation of a random variable by a nonrandom variable from the m.s.e.
criterion point of view is given by its expectation.
In particular,
the best
approximation of a random variable using the results of measurement is given by the conditional expectation relative to the results of measurements.
We denote by Ye the aggregate of the values of the measured process in the time interval [to, t], Y;, = {Y(r) TE [to.t}}. Then the optimal estimate of the
vector
X, = X (u), which gives the solution of problem at u = t is determined by
the formula
ba
1S
eleeB
(5.10.6)
This formula determines the optimal estimate of the value X,, of any random function
xX (u) using the results of measurement of other random function Y(t) in the interval [to, ae It is also valid for the case of a vector argument ¢ and the measurement
of
the random function Y (t) on any set J’ of values of tf.
In the general case it is necessary for practical application of formula (5.10.6) to find the conditional distribution of X,.
the general case is not yet solved.
This problem is very difficult and in
In the particular case of the processes Y(t) and
X (t) determined by Eqs. (5.10.4), (5.10.65) it may be solved under some additional restrictions. But even in these cases the practical application of formula (5.10.6) is a matter of great and often insuperable difficulty. The reason is that the determination of the conditional distribution always requires very cumbersome and time-consuming
calculations
which
can be performed
only after measurement.
The
statement
of
problems in Subsection 5.10.1 implies that the estimates must be calculated in the real time while the results of measurement appear. But the difficulties of application of formula
(5.10.6) do not reduce the importance of the optimal estimation.
This
theory is necessary for studying the potential accuracy of the estimates, i.e. maximal achievable accuracy of the estimation.
526
Chapter 5 General Theory of Stochastic Systems
§.10.3. Auxiliary Problem The general formula for the stochastic differential of the optimal estimate of a given function of the state vector of a system underlies the optimal filtration theory.
Let f(X 5 t) be some scalar function of the n-dimensional state vector of a system and of time.
Its optimal estimate using the results of observation ¥? according to
(5.10.6) is determined by the formula
(5.10.7)
FOS E[FALOTRY:
This estimate represents a functional of the random process Y(t) in the variable
time interval [to, t], and consequently, is itself a random function of t. We put the auxiliary mathematical problem:
to find the It6 stochastic differential of this random
process. This problem may be solved under the condition that W(t) in Eqs.
(5.10.4),
(5.10.5) represents the Wiener process whose dimension 7 is not smaller than the dimension
1; of the measurement
process ¥(¢), and that the furiction Yj in Eq.
(5.10.4) does not depend on X. Eqs. (5.10.4), (5.10.5) in this case have the form
dY = ,(Y,X,t)dt + U1 (Y,t)dW, dX = 9(Y,X,t)dt + (Y, X, t)dW. Lemma
(5.10.8) (5.10.9)
5.10.1. If o1(y,t) = ily, t)v((t)dr(y,t)? where v(t) being
the intensity of the Wiener process then Eqs. (5.10.8), (5.10.9) may
be transformed
to the following form:
dY = 91 (Y,X,t)dt + Ui (Y, t)dWa,
AX = oY, X,t)dt + Hi(Y,X,1)dW, + WHY, X)dWy. Here W, = W, (t) and W(t) first,
Oe hOal 20-10)
are two independent Wiener processes formed by the
= T—N 1 components and by the last components of the process W’ = W'(t)
with independent components: t
w(t) = /Ghai ueDGMD 0
where yp, V1, vy be some matrices and Q(y, t) be any orthogonal real matrix possibly
depending on Y and time t. > As it is known the power of a symmetrical nonnegative definite matrix VY
is determined by the formula y* =
ANSAS,
where
A =
diag{\i,...,
Aq}
5.10. Applications.
Optimal Online Filtering
52m
is the diagonal matrix to which V is reduced by the orthogonal transformation A,
and A* =
diag fag RSS
As}. Then the random process defined by (5.10.11)
represents a Wiener process with independent
components.
Really, the covariance
matrix of the value of the process W’(t) at a given ¢ is determined by the formula
EW'(t)w'(t)" = /Q(y, ry V?(r)v(r v4 2(r)A(y, 7) dr = Ht, 0
since Q(y, T)Q(y, r)t = I by virtue of the orthogonality of the matrix Q(y, T). This formula shows that the components of the process W'(t) are uncorrelated, and consequently, are independent as a Wiener process is normally distributed. Since the intensity of the process W'(t) is equal to the identity matrix, each component of the
process W(t) represents a standard Wiener process. Let us pass from the process
W(t) to the process
W'(t) in Eqs.
(5.10.8),
(5.10.9). Then the differential dW will be replaced by y/2(4)Q(y, t)P dw’. Now we choose the orthogonal matrix (2(y, t) in such a way that the matrix b;(y,t) be reduced to a block form [0 vi (y, t)] where the first block represents the N1 X (r—ny) matrix whose
elements are all equal to zero, and the second block represents an
m4 X nN, matrix. It is possible under some conditions, as the orthogonal Tr X Tr matrix Q has r(r =
Hi? arbitrary elements which must satisfy ni(r =
and rr — 1)/2 = r?/4 abi Ti
ny) conditions,
2; ni(r _ ny) < r?/4. As a result of such a
transformation
vase [= [very v1(Y, t)
[eroacoraw
ei) ee
1/24
PF
A CAV ALS OC
iart
:
=|Seater[aM =
1(Y,t)
7 hagee Y, X,t) Fea? 8
/
V1 (Y, t)dWe
i |W'(Y, X,t)dW, + b"(Y, X,t)dWy |: To find the matrices 1’, wy" and vy and establish the conditions under which the transformation of Eqs.
(5.10.8), (5.10.9) into the form of Eqs. (5.10.10) is possible
we evaluate the conditional covariance matrix of the random vector
Wien 18”=[warxy vary |" relative to the random variables Y, =
X where AW
= Wit AP At) — W(t), AW'
W'(t + At) - W'(t). Taking into account that AW
and AW’
are independent
528
Chapter 5 General Theory of Stochastic Systems
of Y;, Xz and their covariance matrices are equal to y(t)At + o(At) and [At respectively, using the formula of transformation of the covariance matrix at a linear transformation of a random vector and omitting for brevity the arguments of the functions w, 1, yw, yw", wy we have
oe emmy ewe ~ yp Fay paceman 1
Mews pl en»
aglt
| pl
wed at :
yt
yt
Hence, dividing both parts by At we obtain as At + 0)
vivdt pdt i ie ce yvyt yyy? ppt wp?
+ py?
| ’
f= byl, Wt! = wed, Ve +e = doy. (5.10.11) It remains to solve these equations successively for v1, wy” and x’. The matrix wy determined by formula (5.10.11) is independent of X, as 1 is independent of X. The second Eq. (5.10.11) has the solution if and only if the matrix vi (y,t ) is invertible at all y, t. It is necessary for this that the oe
py vt
invertible at all y, t since on the basis of the first Eq. (5.10.11)
|vivyf |.
In this case the second Eq. (5.10.11) gives ae
Wil =
be
pul (pi! ee After that the
third Eq. (5.10.11) is reduced to the form
>
wee? = dvd? — pod (YP (Wh ed*, or taking into account that by virtue of the first Eq.
(5.10.11) (oy rT (Ont )
= (pivyt)-
vy = wd? — dol (dvd) dy. The right-hand side of this equation represents
a symmetrical matrix.
Let us prove
that it is nonnegative definite. For this purpose we notice that the matrix
es ;
,
pvp? pup, ‘
pivyt ppt
|?
F
being proportional to the covariance matrix of the random vector
j
:
ke
,
is nonnegative definite in consequence of which at any vector U =
Ie
[yr yr | dw
ae
[ut us l
ul Au = ThPabyvyt Ui + UyPFpyyt uy + uy;Pb uu.+ UsPvt
us >> 0.
5.10. Applications. Optimal Online Filtering
Putting here uj = —(~1 yr) fk
aL
529
divdT u2 we shall have
T\-1
lB
a
is
us Pryy (divdy )~ pivy? ue — ug pot (dive) diveduy —uZ byt (divdt)“ dvd? us + ul dy up.> 0, or after cancellations
us [yvy? — prot (divet)
pvp? |us > 0.
The validity of this inequality for all n-dimensional vectors U2 proves our statement.
After this the solution of the last Eq. —
ia [v _ yor (v4 ypr)- tyr
ces vy and
(5.10.11) is given by the explicit formula y’
|any
It is-evident
that multiplying the matri-
on the right by arbitrary orthogonal matrices of the corresponding
dimensions we obtain another solution of Eqs. (5.10.11). Thus Eqs.
(5.10.11) have an infinite set of solutions, and consequently, Eqs.
(5.10.8), (5.10.9) are reducible to the form of Eqs.
(5.10.10) if the matrix 01 (y,t)
= vi(y, t)v(t)vr(y, ie is invertible at all y, t. < Lemma
random
5.10.2.
The stochastic
differential of the optimal estimate
of the
variable f (Xz, t) for Eqs. (5.10.4), (5.10.5) is given by formula:
df = Elfi(X,t) + fe(X,t)" 9(Y, X,t)
+ hte{fee(X (Wy? UY,XH} IVAIdt + BLK, 1)(ei, X,)% oF +fo(X,t)? (budt (Y, X,t) |Yilavyr)"(¥, (d¥ —Grdt), (5.10.12) where
(bub (dur
pr(z)
)(y, z,t) = p(y, 2, t)v(t)yv(y, z; ae )(y, ev; t) = vy, v, t)v(t)di(y, yh
(divbl)—*(y,t) = Wily, ty(dr(y, t)7)77,
(5.10.13)
~i= / yip:(x)dx= Elyi(X1,¥%,t |Y;,),
(5.10.14)
being the conditional
density
of Xz relative
to Ye; the derivatives
fy, Tse
fer and all the conditional expectations in its right-hand side are supposed to exist.
p> Let us consider two close instants t and ¢t + At, At > 0, and write
Af(t) =f(t+ At)—f(t)
530
Chapter 5 General Theory of Stochastic Systems
= E[f(Xipar,t+ At) |Vit**] -
[F(X0)1 YA |
= E[ f(Xtzar,t+ At) — f(Xt,t) 1Yi]
+E [f(Xisat,t + At) |¥i*4"]—F [f(Xtpar,.t
+A) |Yi]. (6-10.15)
We evaluate every item in the right-hand side separately.
Using the differentiation
formula of a composite function (5.1.7) by virtue of the second Eqs. (5.10.10) we get omitting for brevity the arguments of the functions
f(Xtzar,t + At) — f(Xt,t)
= {(Xt) + fo(X1,t)"9+ tr (teal Xe 0100" |vs|) mi
+ fo(X1t)7 (WAM + YAW) = {f(t 1 oe
+5 [fea(Xe(yy + ¥'b"™)] Late f( Xe)" (WAM +W"AM). Substituting
=
here the expression
pl" (AY =
of AW
from
the first Eqs.
(5.10.10),
AW,
yi At), taking the conditional expectation of the obtained ex-
pression relative to Ye: using the third Eqs. independence of AW;,
AW2,
(5.10.11) and taking into account the
AY we obtain at fixed At
E| f(Xtzar,t + At) —f(Xs,t) |Ys
E [f(X00) ++fe(Xe,t)? p+ str{foo(X+, ty" } |a At
B |fo(X1,t) wd" (AY — gi At) |¥s] . Further we find from the first two equations of Eqs. (5.10.11) -1
eawi dt ), yO
-1
=e
oh
=
(beet).
Substituting these expressions into the previous formula and taking into account that w 1 does not depend on X; we shall have
E |f(Xtzar + At) — f(X1,t) |Ve| = E | fi(Ys,t) + fo(Xt,t) e+ tt {fee(Xz,t)puy"} || At
5.10. Applications.-Optimal Online Filtering
531
+E [fo(Xe,t) buys (vivdT)- (AY — pi At) |Yi].
(5.10.16)
For evaluating the second item in (5.10.5) we shall find the conditional density of the random vector X;4, relative to viper
the random vector X; relative to Y,, a(n random variable AY
Let pr(z) be the conditional density of
| Y, =) the conditional density of the
at given values Y, & of the random vectors Y;, X; at a fixed
value Z of the random vector X;. As the first equation of (5.10.10) determines the Markov process Y;, the density qu(n | Y, x) coincides with the conditional density
of the random variable AY at given values ie and @ of the random variables Ye and X;.
Consequently,
on the basis of the multiplication theorem of the densities
the conditional density p(x) of the random vector X;} relative to Voe
or, what
is the same, relative to vee and AY js determined by the formula
pr(@)qr(AY | Yi,2)
p(x) =
(5.10.17)
ofpr(a)qe(AY | %,a)da Here the numerator represents the joint conditional density of the random variables
X; and AY
relative to ys and the integral in the denominator represents the
conditional density of the random variable AY
relative to ae
For determining
qi(Ay |x; zr) we notice that by the first equation of (5.10.10) at AY=
X = x
yil%t, z,t)At aF wy (¥1, t) AW2.
Hence it is clear that at given values of Y; and X; = & the variable AY
represents
the linear function of the normally distributed random variable AW 2. Consequently, the conditional distribution of AY
for determining q(A
at given values of Y; and X; = @ is normal and
| Ys, a) it is sufficient to find the corresponding conditional
expectation and covariance matrix of the random formulae
for the expectation
and the covariance
vector AY. matrix
Using the known
of a linear function of a
random vector, taking into account that the covariance matrix of the random vector AW
is equal to At, and omitting the arguments of the functions ~; and wy we
obtain
PLAY
IY 52 Par,
Ky
=
tk
At,
or, by virtue of the first Eq. (5.10.11)
E[AY |¥i,2]=giAt, Kay = vivyfAt Consequently,
a(AY | Yin)= coxp{-SR (AY?—pFAt) (dive) "(ay pid},
532
Chapter 5 General Theory of Stochastic Systems
c being the normalizing constant.
Taking into account that the quantities At and
AY are infinitesimal we may represent the exponential function by the Taylor formula
with the residual of order o(At). Then we obtain
O(
Vouliptaw Goze cexp{— —— 5 :AG AYE(ahd NE AX
tof (divdf TAY — Sef hvu) oath = cexp {=A
vf ty tay|f+ofand) AY
Z
1
Ss
5¥1 (Wid) peAtt sor (vivdt )'AYAY® (vivdy
:
gi].
Retaining only the expectation of the random variable NY UMY -
EAYAY?™= gp?At?+ of b//At = piv yt At + o(At), we obtain q(AY
|Y;, xz)
1 = cexp{ AY" avd tay} [14 oF duvet) tay], and, as {1 is independent of 2, co
| pr(z)qe(AY | ¥;, x)dx —oo
= cexp {-aaphYT Wit) tay} [1467 (vyivdt) AY], where
$j
g(AY
is defined
by
(5.10.14).
Substituting
the
obtained
expressions
of
| Ys; zr) and of the integral into (5.10.17) we shall have after the can-
cellation
! =~ 1+¢f (ivy? )- LAY p,() = pi(a“Ty aT WORDLE hydED TAY" AY Representing 1/[1 + pT (vy yp? )-} NG
by the Taylor formula with the residual
of order o(At), performing the multiplication and retaining only the expectation of
the random variable
AY AY”
as before we get
Pi(z) = p(x) + ot (divdt)*AY][1— OT Wivdt) AY
5.10. Applications.“Optimal Online Filtering
033
+01 (YivdT) AYAY? (hivyt)-* 41] = pr(){1+ (¢t — OT )(divdt) TAY — AYAY? (pivyf)-'eT]} = p(x)[1+ (9T — $1 )(divdt)“*(AY — @ At)]. Now
we may
evaluate the second item in the right-hand side of formula
(5.10.5).
Taking into account that according to (5.1.7)
f(Xt+at,t + At)
= (Xt) + {fi %e,t) + fo(Xe te + 5ttlfeo(Xe,t)oryT At 4+fa(X+,t)? = f(X1,t) +
(b' AW, + 4 AWs)
Att fo(Xe,t)?(W'AW; + yp"AW2)
where w’ and w”’ depend only on Y;, Xz, t, we find
E[ f(Xipar,t+ At) |Yt°] — B[ f(Xipar,t+ At) |Yi]
AW, +o” AW2) | (Wi)? = i [f(w,t) + fidt + fo(x,t 12.9)
Its PAE Hie /[F(@,t) + Att fo(z,t)" (WAM, +" AWe)\(et — 61) (vivdt (AY — Gi At)pe(x) de. Using the first Eq.
(5.10.10) and retaining here only the items of order not higher
than At we get
E[ f(Xtpat,t + At) |Yh] — B[ f(Xegar,t+ At) |Yi]
= f sat) + (FF — ef )e@)de (vv) MaY - p.d9) a FEfo(x,t)? (b' AW
AWE yi" + bp" AW2AW?
yp''")
x(vivPT) (er — G1)pr(w)de = Elf (Xe, tr — $1) 1Yi
534
Chapter 5 General Theory of Stochastic Systems
x (div?) (AY — gidt)+/ fo(a,t)? (W AW AW? wT =o:
+43" AW AWe pl" )\(bivd7 )~!(o1 — G1) pi (2) dz. Finally, retaining only the expectations of the random
AW,AW?
taking into consideration that EAW,
variables AW,
AW?
=
(5.10.18) AW?
and
0 in consequence of
the independence of W2 and Wj and the second Eq.\ (5.10.11) ppt
is ouyr
we obtain from (5.10.18)
E[f(Xtpat,t+ At) |YO" ]—
E[f(Xisar,t+At) LY]
= E[f(X+,t)(¢t — $f) |Vs] bivd fT) (AY — 61 At) (5.10.19) (vivdt )*(y1 — 61) |¥,] At. wot ,t) +E |fe(Xe Substituting expressions (5.10.16) and (5.10.19) into (5.10.15) and replacing the increments by the differentials we obtain the final formula (5.10.12).
Agw(t, Tr),
(I)
(a=ail
where JV is the random number of impulses acting on the network in the time in-
terval (to, t); T,,..., I Aj,...,An
the random
the random instants of action of the impulses, and
magnitudes
of the impulses acting on the input of the
circuit. The simplest way to find the n-dimensional
distribution of the random func-
tion X (t) is to evaluate the joint characteristic function of its values at the ins-
tants t;,...,%¢,.
Denoting by ¢ the largest of the numbers
t),..., tn, t
= max(ty abies teas and taking into account that w(t, Th) = Opatt) during the
time interval (to, t). Then the probability of the events B,, will be determined by the formula t
P(Bm) = Te"? 0 = [u@er (aves eareeetSe
Sera)
to
To evaluate the conditional expectation relative to the event B,, in (II) we shall use again the formula of total expectation taking first the conditional expectation relative
to the random variables Tj , ... , Tj. Then taking into account the independence of the random variables A; , ...
Ty
PRT
, Am and their independence of the random variables
we’ obtain
exp
ya Dn w(to, te) > Bm
m
= Eexp
Nini 5 Sh w(tp, Tk) inal m
=
FE|E| exp
bit, Ay a1
p=
n
Apto (tps
op |Thy
vob
p=l
- eT] >» w(tp,Tr) | ,
(IV)
Problems
545
where ga(A) is the characteristic function of the random variable A, (k —eleD2t !): For evaluating the last expectation we shall find the joint conditional distribution of the instants of action of the impulses 7) , ... , Tj relative to the event B,,.
us take arbitrary instants of time Tj ,...,
Tm, T1
.
(VIII)
p=1
to
Thus the multi-dimensional characteristic functions of the random function X (t) are
determined by Oty
MABE
nyt,
y song ten)
= exp
> Apw(tp, tT) | —Dl
ajdt > =. Ax)
pt (he
5.2.
An)
\
Ga to
(Al a
aa
Derive the formulae for the stochastic Ité differentials of the functions of
the standard Wiener process given in Table A.1.1. Obtain analogous formulae for the stochastic 0-differentials.
Problems
5.3.
547
Derive the formulae for the stochastic It6 differentials of the functions of
a vector Wiener process given in Table A.1.2.
Obtain analogous formulae for the
stochastic 0-differentials. 5.4. Prove that the stochastic It6 differential of a scalar function
UV= y(P) ofa
Poisson process P is determined by the formula dU = [p(P+ 1) = y(P) |dP. Evaluate
by
this
formula
Table A.1.1.
the
stochastic
differentials
of
the
functions
given
in
:
5.5. Show that the scalar stochastic differential equations of the first order with
the 6-differential (i) es= —e(1 + V2)Y + kV,
GY = (14 Va)el¥) + kV, (iii) Y = go(Y,t) + o gi(Y,t)Vi may be transformed into ay naeam It6 equations:
(i) Y = —e(1 — 20ev22 + Vo)Y + k(—20ev21 + Vi), (ii) Y = [1+ 2evo2! i%) +V2|y(Y)+ bial (Y)+Vi], (iii) Y = yo(Y,t) + 0 S Vpg Pp(Y, tp, (¥, t) + oYpi(Y,t)Vi, P,I=
where y’ LE = 0y/ Ody, and
V =
noise with the intensity ’.
Under what conditions to the It6 equations coincide in
[Vi Van
ies is the aan
distributed white
this case with the initial equations? 5.6.
Let Vit) be such a m.s.
n — | times differentiable random process that
its (n — 1)*> m.s. derivative has the Ité differential
dy" -1)
sy ve
vier) de aw,
ny OT ta,
(1)
This relation is called the stochastic Ité differential equation of the n>} order relative to the process NAAN This equation may also be written in the form
Veer where
V =
YG
Vee
OOo
ee
wy at)
dW /dt is the white noise in the strict sense.
em (LL)
Prove that Eq.
(II) is
transformed into a standard form of the stochastic It6 differential equation for some n-dimensional vector random process. In the case of the Wiener process W(t) derive
the formulae of the transition from the It6 equation of the n*) order to the equation in the 0-differential of the N*» order and vice versa. 5.7.
Show that the scalar stochastic differential equations of the second order
with the 0-differential:
:
(i) Y + 2wo(1 + V3)Y = we (1 => V2)Y = kV
(i) ¥Y= go(¥,Y,t)+ 0 oY, Y,4)Vi f=
548
Chapter 5 General Theory of Stochastic Systems
correspond to the following It6 equations: (i) Y+
2£wo(1 — 20133£wWo + V3)Y + we(1 = 20V23EWo + Vo)Y
= k(—20€113W0 + V1), Gy
:
= va ly).
n
t) ae 0
in en
et
;
Pe
Ula
;
ts
Lh=1 n
+>
°
i
’
91 (Y, Y, t)Vi, where Yi, GY > YAu bli. OCne (as
5
t)/OY, and V
(=0
=
[Vi serail
12 is the normally distributed white noise of the intensity VY. Find
the conditions under which the It6 equations coincide with the initial equations.
5.8. Show that the n-dimensional characteristic function 9, and the distribution density f,, at the random initial condition Yo and at the absence of the noises in Eq.
(5.4.6) when b(Y, t) = 0 are determined by the following formula:
Oni
Oa LAl eos
Ceol
foe)
a
/ eT P(t)
+
FEAR O(EnN) fo (im)din (n = iif 2, i. =)
(1)
—oo
f2 RG
ee
een)
= fo(y™*(t1, yn) ey (t1, 91) 16(yo — v(t2, p~*(t1, (y1)))
(Yn — elms P(t (i) (n= 2,3,--9),
(11)
where Ya) = p(t, Yo) is the solution of a system which represents the function of the time and the random initial value Yo.
5.9.
Show
that for the one-dimensional nonlinear system Y =
—yy'tr
(r
= —1) with the random initial value Yo its n-dimensional density is determined at
ly] < (ryti)~ 1/r by the following formula:
fn(yi; o-
55
OA
ile Ae
Besa
=
fo(ywi(1
xT] 6m — (1 = ryt yf) OP)
me rytryt)/"
(1 -
rytyyy)~Atn/n)
$ rtiys(1= rytayy)“OF9 PO",
(=
where fo(yo) is the density of the random initial value Yo. 5.10. Find the one-dimensional distribution of a stationary process in the conservative mechanical system:
Problems
with the random initial conditions where ralized coordinates;
p = Aq =
[pi ep
549
q = [q1 a odllLa Es is the vector of the geneyy ‘i is the vector of the generalized im-
pulses; A is the generalized mass (constant); H is the Hamilton function (the total
energy of asystem) equal to
H = H(q, Pp) = p! A~+p/2+1(q) where II(q) is the
potential energy of a system. Show that at II(q) = q Cq/2 the stationary distribu-
tions of the coordinates and the impulses are determined by the densities (1 (q? Cq) and ~2 (p? A-!p) where (); and $2 to a great extent are the arbitrary functions (more exactly they both are expressed in terms of one arbitrary function). Show that
the distributions of the doubled potential energy Yj = peo q and the doubled kinetic energy Y2 =
TA
tp are determined by f} (y1) = Ci yr
2g, (y1)1(y1),
fo(y2) = coy? !? G(yo)1(y2), where C, and C2 are the normalized constants. In the special case of the exponential functions ~; and $2, yi(z) = p2(zr) = e72/2 , the distributions of thedouble potential and the kinetic energy present
y?-distribution with n degrees of freedom.
Find the distribution of the total energy
yi = (Y1 + Y2)/2. 5.11. Check that for the nonlinear system described by the second order equa-
tion
Ys 2eY +¢'(VY)=V
(e> 0),
(I)
the one-dimensional density and the one-dimensional characteristic function of the strictly stationary process are determined by the formulae:
fily, y) = cexp{—h?[ p(y) + 97/2]} (Ah? = 2ev), Hives) = A co
e970) dy,
(II) (IIT)
where C is the normalizing constant, V is the stationary normally distributed white
noise of the intensity V, p(y) is the nonlinear function, y'(y)
Ov(y)/Oy.
Consi-
der the special cases: (a) y/(y) = aiy + azy"; (b) y’(y) = asgny; (c) ’(y) = @ailt Y: 5.12.
Show that the one-dimensional density of the strictly stationary process
in the nonlinear system
Y =¥(Y)—2eaY +bV,
e>0,
(1)
is determined by the formula:
fily) = Vor/me"Y, a = 4e/v.
(II)
Here Y(y) = [i(y) --. Pn(y)]", Ova(y)/Oye = 0 (k = 1,...7), y" v(y) =
yy ViVi k=1
Vay
=
[Vi , ean
Ven ips is the normally
distributed statio-
550
Chapter 5 General Theory of Stochastic Systems
nary white noise with the independent
components
of the same
intensity V, b is
the constant N X N-matrix, d@ is the constant N X M-matrix satisfying the condition
a+a! = 2d07. 5.13.
Show that the one-dimensional density of the strictly stationary process
in the system
Y =ay'(Y) +0, is determined
by the formula:
fily)
=
(I)
ce2¥(y)/Y
constant, ply) is the nonlinear function, y'(y) = (rectangular in general) matrix; a+
at
=
Cc being
the normalization
Op(y)/dy,
b is the constant
@ is the constant matrix satisfying the condition
2bb7 " V is the stationary’normally
distributed white noise with the
independent components of the same intensity V.
5.14. Show that the one-dimensional distribution of the strictly stationary pro-
cess Y(z) = [Y1 (t) Y2 (t) [# in the nonlinear stochastic system
Y, = asgn(Y2— BYi), Y2=—-Nni+V,
(1)
filyr, yo) = cexp{—B(yz + 2a|y2 — Byr|)/v},
(II)
is determined by:
where C is the normalizing constant V, is the stationary normally distributed white noise of the intensity Y. Find the corresponding characteristic function.
5.15. Show that the one-dimensional distribution of the stationary process in the
nonlinear mechanical system with n degrees of freedom whose behaviour is described by the equations in canonical variables
q= OH /Op, p = —0H/0q
is determined by formula:
— 2ept+ b(q)V,
b(q) = A(q)\/?
(I)
1985):
fi(g,p)=ce"*",
a = 4e/v,
(II)
where C is the normalization constant. Here q =
[q1 Soll ie is the vector of gene-
ralized coordinates,
ie
p=
A(q)q =
[pi -++DPn |
pulses, A(q) is the generalized mass
is the vector of generalized im-
which represents symmetric positive definite
matrix depending in general on g, H is Hamiltonian function (the total energy of the
system) which is determined
by the formula
H =
H(q,p)
=
p? A(q)~1p/2
+ II(q), II(q) being the potential energy of the system, 2€ is a specific coefficient
of viscous friction, components
V =
[Vi prev i7, is the vector of generalized forces whose
represent independent
normally distributed stationary white noises of
intensity /. Find the known Gibbs formula (Gibbs 1928) at b(q) vie
Problems
551
5.16. Show that the function f; = f; (y; t) will be the solution of Eq. (5.6.56) with the normal white noise V if and only if the vector function
@ = a(y, t) assumes
such a presentation
a(y,t) = ay(y,t) + a3(y, t),
(1)
that the function fj is the density of the invariant measure of the ordinary differential equation
;
¥ —a,(),1);
(II)
i.e. satisfies the condition
ee
I
and the correspondent function as = ak(y, t) is determined by formula T
SR
5.17. Suppose that for Eq. (5.4.6) at a normal white noise V the presentation
(1) of Problem 5.16 a density {44 =
H1(y, t) of integral invariant (II) are known.
Besides let
(i) the matrix @ be invertible, det lem me U) (ii) the vector function Y; = 71(y, t) where
lyst) = 078) ab —
ited
|esZin] ‘
satisfies the condition of the curl absence: rot 7; = 0 or
O71: /OY; => O71; /OYi
(t,9 =
“ siie\ce =P),
(iii) the scalar function
Fy(y, t) >
[tw t)dy,
is the first integral of Eq. (II), (iv) the condition of the normalization is fulfilled
exc
exp Fi(y,t)dy = 1.
552
Chapter 5 General Theory of Stochastic Systems
Show that there exists the solution of Eq.
(5.4.6)
Y = Y (t) for which the
one-dimensional density fj (y; t) is i determined by formula
fi(yst) = p(y; t) exp Fi(y, t). 5.18.
(I)
Show that the function 1i(y, t) of Problem 5.17 is equal to fily; t) if
and only if there exist such a matrix function Aj = Ay (y, t) that
apy = (GA1/Oy)", where as =
Ar + Ay = op,
(1)
(0) = al. Find the correspondent formulae for the transition function
and the multi-dimensional densities.
5.19.
Consider Problem 5.17 for the system with the singular total matrix 7
= byb?
YSQYoY't),
VS Py owl et)
bol 2,0),
(I)
where Y’ and Y” are the s-dimensional vectors, 25 = p, Vo is the r-dimensional normal white noise of the intensity V9 = Vo (t), p is the dimension of the vector Y,
Co.
bo(Y’, NA t)Vo (t)bo(Y’, Ne it is the nonsingular matrix. Show that the
one-dimensional distribution is given by the following formulae:
fly syotha aly
&) expt lye),
(II)
where
Kiya .tv= /7} dy' dy",
ny yt) =o9 Po, Oy
Oni/Oy; = Oy; /Oy:,
or
Ge + ay (Om)
(IIT)
(IV)
or + gyi (Pie)
1_ 90 Oln fi Be
=
0,
a Pi + Py.
(V)
(VI)
5.20. Show that the solution of Problem 5.18 for the system with the singular total matrix 0 exist if and only if the conditions are fulfilled:
P31 = (OB, /dy")", Bi +BY = oop,
(I)
sil Pia= P—p} 1) 41 is the density of the finite integral invariant of the equations
=Q,Y"= P}. Show that fi(y’, y”,t)
= wi(y’, y”, y).
Problems
5.21.
553
Show that for any vectors g, p and vector functions Q(q,P), P(q, Pp) of
the same dimensions as q, Pp if there exists a function N(q) satisfying the condition gr
or
9g NO?)
+ Ci
ee. q) = 0,
then the one-dimensional distribution of the solution stochastic system
i= Qa.) P= Pla”) 2(Hal + 4(Q)V, may be presented by formula:
s al
filq,p) =cN(q)e PY),
W(H) = fean,
pi4/y
aD
Here b(q) is any rectangular matrix, a(q) is any square matrix satisfying the condi-
tion a(q) a a(q)? = 2b(q)b(q)?, Hus H(q,p) is the first integral of Eq. (I) at a(q) = b(q) =f, 2e(H) is coefficient of viscous friction,
V = V(t) is the normal
white noise with independent components and equal intensities VY. Find
in special
case
€ =
filq,p) = cN(q)e"*#4?),
const
the following
generalized
Gibbs
formula:
a = 4e/v.
5.22. Show that for the scalar nonstationary stochastic system
Y = ao(t) +a(t)Y +0(Y,t)V, (Y,t) = Vc(t)Y, Y(to)=Yo, the solution of Eq.
(1)
(5.6.65) for the one-dimensional characteristic function has the
form
gi(A;t) = go(v(t, 4)~+) exp
if v7, e(t, A)ao(r)ar
Here V is the normal white noise of the intensity v(t),
ae
t
re = |a(ryar t
=
Palas to
- feceyio ar,
:
(II)
504
Chapter 5 General Theory of Stochastic Systems
v(1,9(t,)) =
—
-
to
:
1/2 [ v(71)¢(11) exp -{ a(a)do | dry + y(t,Tr) to
At
the constant
fucntions,
to
dg,
@, Cc,
Y =
1 and
the initial condition
go(A)
= (1 — 2iDo)~1/? we obtain
gi(A;t) = [exp(at) — icd(exp(at) — 1)/2a]-***°
x [1 — 2éDor(t)]7!/? exp (21)
(III)
where r(t) = A[exp(at) — icd(exp(at) — 1)/2a]~*. Give the generalization for the case of vector Eq. (I) when b(y,t) =
diag [bi(y,t)... by (y, ae be(y, t)
= oj(t)" y. =
ae
5.23.
Show
that under
the conditions
of the previous
problem
the multi-
dimensional characteristic function is determined by the recurrent formula:
Grill X15 eas pistes | rte )e=Gaer(Ans
+07 (ta; Antal tite
oe)
Anza)
pexp [/ W(7, (tn, An))ao(r)dr | ,
(I)
tn-1
where 1) = 2,5,2.
ve(t) [f(y — yeest) — fi(y;t)] . 5.25. Using formula (5.6.39) and Eq. (5.6.32) show that in the case of the general Poisson process W(t) and unit matrix b(y,t) = I the one-dimensional density is defined by the following integrodifferential Kolmogorov—Feller equation:
ei: ee—FT Latust)Aaluiel +v(t)[(F*fid(y,t)-A(yst)],
(
where f * f; being the convolution of the density f(y) of the jumps magnitude of the Poisson process with the density f; (y; t):
(FeAl) = f fy- Miran,
(nt)
Write a similar equation for fn(y1,---, Ynjti,---., tn). Obtain from Eq. the corresponding equation for the case where
(1)
W(t) represents the product of a
constant h by the simple Poisson process.
5.26. Using formula (5.6.41) for the function x (p; t) show that Eq. (5.6.32) is reduced to the following integrodinfferential generalized Kolmogoroy—Feller:
Of es (y; = — oF we 9, [ey Ofi¥ 4] Wh
ts re la [H(ost)ua(dben 0" ACwe)]} N
+ > x(t) k=.
where
/ ge(y— 750, t)films t)dn — filyst) | , —
{; (E, n, t) is the
gu (ch O(y, DEX), i.e.
0O
density
corresponding
to
the
characteristic
function
the density of the random vector representing the product
of the matrix (d(y, t) ck) by the random vector Y,; which is the random jump magnitude of the general Poisson process JPR (t).
556
Chapter 5 General Theory of Stochastic Systems
5.27.
Show that in the case where x(p; t) is given by (5.6.47) Eq.
(5.6.32) is
reduced to
AED Fat iti + [{ituo t)c(a);t)— Ro
zee(OF [ou too(toeu." nnso))} fi(y;t)+ a [b(y, t)e(x) f(y bret a)dz.
CHAPTER 6 METHODS OF LINEAR STOCHASTIC SYSTEMS THEORY AND THEIR APPLICATIONS
Chapter 6 is devoted to the methods of the linear stochastic systems described by the linear stochastic integral, differential, difference and other operator equations in the finite-dimensional and infinite-dimensional spaces. In Section 6.1 the spectral and correlation methods are stated. Section 6.2 contains the account of the methods of general stochastic systems theory based on the multi-dimensional distributions. In Sections 6.3 and 6.4 methods based the canonical representations are considered. Sections 6.5 and 6.6 present applications of linear stochastic systems theory to the linear shaping filters design, analytical modeling and optimal linear filtering based the online data processing.
6.1. Spectral and Correlation Methods 6.1.1. Representation of State Vector
Let us consider a linear system. form
In this case Eq.
(5.4.6) has the
Y =aY t+ao+0V,
(6.1.1)
where a = a(t), ap = ao(t), b = b(t) may in the general case be the functions of time t, and V is a white noise whose intensity function of time t. Remark.
vymay be a
For finding the first and the second order moments of the random
process Ve) determined by linear differential Eq. (6.1.1) there is no necessity that the white noise V be the white noise in the strict sense, i.e. the derivative of a process with the independent increments.
It is sufficient that it be simply the white noise
(i.e. the derivative of the process with the uncorrelated increments) as well as in the
theory of Subsection 3.2.3. Therefore we shall consider V as an arbitrary white noise in this section. In other sections studying the multi-dimensional distributions of the
process age we shall always consider V as the white noise in the strict sense.
Using formula (1.2.8) determining the solution of Eq. (6.1.1) we get representation on the system state vector Y by the formula t
WAG
ultsto)¥o+ fult, VV to
t
(r)ar + fultsr)a0(r)ar, to
(6.1.2)
508
Chapter 6 Methods of Linear Stochastic Systems Theory
where u(t,7) ts the matric determined as a function of t by the homo-
geneous differential equation du/dt = Gears ra &
a(t)u and the initial condition
6.1.2. Representations of First and Second Order Moments
Taking into account that the expectation of a white noise is equal to zero we find by virtue of (6.1.2) the following formula for the expectation
of the system state vector Y(t)
‘
m(t) = u(t, to)mo + ic T)ao(r)dr.
(6.1.3)
where mg is the expectation of the initial value Yo of the state vector Y. The covariance function of the state vector Y is determined by the formula K(ty ; ta) =
u(ty ; to) Kou(te, to)*+
min(t1,t2)
+
/
u(t, T)b(r)v(r)b(r) 7 u(te, ars
(6.1.4)
to
where Ko is the covariance matrix of the initial value Yo of the state vector Y. While deriving this formula we remembered that the initial state of the system Yo is independent of the white noise V(t) at ¢ > to and that
u(t,7)
= 0 at r >t. Due to this the upper limit of integration is equal
to min(t;,t2). Besides that we took into account that the coefficients of the equations of real systems are always real while the elements of the
matrix u(t,7) may be complex in this case. After determining the expectation and the covariance function of the state vector Y we may find its second order moment which in this case of real-valued functions m(t) and K(t,,t2) gives
T'(t1,t2) = K (ti, t2)+ m(t1)m(te)?.
(6.1.5)
6.1.3. Differential Equation for Expectation
In practical problems it is usually sufficient to find probabilistic characteristics of the state vector of a system Y at every given instant
6.1. Spectralkkand Correlation Methods
559
t (determined by the one-dimensional distribution), i.e. only the values K(t,t) = K(t) and [(t,t) = I(t) of the covariance function and the second order moment. In other words, it is sufficient to find the expectation, the covariance matrix and the second order moment of the state
vector Y at every instant ¢t. It goes without saying the all these vari-
ables for a linear system may be determined by formulae (6.1.3)—(6.1.5) at t) = t2 = ¢t. But in the case of a linear system they may be calculated more simply, ‘namely by the integration of the corresponding linear differential equations. > In order to derive the differential equation for the expectation of the vector
Y we differentiate formula (6.1.3)z t
m(t) = ur(t, to)mo + [ut T)ao(T)dr + ao(t) to t
= a(t) | u(t, to)mo + Ju(t, r)ao(r)ar)
+ ao(t).
to
But
the expression in the brackets
is equal to mM
=
m(t) by virtue of (6.1.3).
Consequently,
m=am+ay.4
(6.1.6)
Thus we obtained the differential equation for the expectation of the vector Y. Integrating Eq. (6.1.6) with the initial condition m(to) = mo we may calculate the expectation of the random vector Y tn stochastic linear system (6.1.1). 6.1.4. Differential Equation for Covariance Matriz > In order to derive the equation for the covariance matrix K(t) of the vector
Y we put in (6.1.4) t) = tz =T: K(t) =
K(t,t) =
u(t, to) Kou(t, to)*
t
a /u(t, r)b(r)v(r)b(r)? u(t, 7) to
ate
(6.1.7)
560
Chapter 6 Methods of Linear Stochastic Systems Theory
Differentiating this formula with respect to t we obtain
K(t) = u(t, to)Kou(t, to)” + [ee
+ fe
ome
to)” pneryPult, 7)" ar + u(t, to) Kour(t,
poryu (or
ult, rar 4 b(t)v(t)b(t)” ,
to
or as U;(t, T) = a(t)u(t, 7), ur(t, 7)* = u(t, T)*a(t)? (Subsection 1.2.2),
K(t) = a(t) u(tto) out to) + fut, 2)0(7)0(7)0()P u(t, 7)*ar|+ [ultto)
to)
+ /u(t, T)b(r)v(r)b(r)7 u(t, rar] a(t)? + b(t)v(t)b(t)? . to
But the expression in the brackets by virtue of (6.1.7) is equal to
K =
K(t).
Consequently,
K =aK+Ka™+bvb". 4
(6.1.8)
Thus we obtained the differential equation for the covariance matrix of the value of the random vector Y at a given t. Integrating Eq. (6.1.8)
with the initial condition K(to)
= Ko we may calculate the covariance
matrix of the random vector Y in stochastic linear system (6.1.1). 6.1.5. Differential Equation for Second Order Moment > For deriving the differential equation for the initial second order moment
I(t) of the vector Y we put in (6.1.5) t} =
tp =
t. Then we shall have I(t)
= K(t) ar m(t)m(t)? or, omitting the argument f
T=K+mm’'.
(6.1.9)
6.1. Spectral and Correlation Methods
561
Differentiating this formula we find
T=K+mm? + mm? . Substituting here the expressions
™ and K from Eqs. (6.1.6), (6.1.8) we obtain
T = aK + Ka? + byb? + amm? + apm? + mm? a? + mai , or using formula (6.1.9)
T= al +Ta? + bvb? + agm? + ma? . 4
(6.1.10)
- va
Integrating Eq.
(6.1.10) after Eq.
(6.1.6) which determines the
expectation m with the initial condition T(to) may calculate the initial second order moment
= To = Ko + mom we of the random vector Y
in stochastic linear system (6.1.1). Remark. (6.1.8) or Eqs.
It is evident that the direct numerical integration of Eqs. (6.1.6), (6.1.6), (6.1.10) is much simpler than the calculation of m
and K
or m and I by formulae (6.1.3)-(6.1.5) with preceding multiple integration of the homogeneous equation U = aU with the initial condition u(r, im) = I at various T necessary for finding u(t, Ts) as a function of two variables.
6.1.6. Differential Equation for Covariance Function > Starting from formula (6.1.4) we shall derive the equation for the covariance function iG is ; to) of the random process a(t) considered as a function tg at any
fixed t;. For this purpose we rewrite formula (6.1.4) for the case tj < tg:
K(t1,t2) = u(ti, to) Kou(te, to)* + [ee
DoH
ult, edie
to
Differentiating the is formula with respect to tg we find
OK e (t,t e 2
= u(t1, to) Kout, (te , to)i
+ /ss(tah BCE) BT)? ty,(tgpNod = 1(#) 440) Kotte, to)"a7(42) to
562
Chapter 6 Methods of Linear Stochastic Systems Theory
+ /u(t, 7)b(r)v(r)b(T)? u(te, r)*a! (t2)dr,
-
OK (t1, te)/Ote =
K (ty, t2)a(t2)? Ue
UO
(6.1.11)
The initial condition for this equation has the form K(ty , t) = K(ty is J
Integrating Eq.
(6.1.11) at various
values
of t; we
number of sections of the covariance function K(t,,t2) obtain K(t,t2) at tz t.
Kips) = ate
a
To
(6.1.12)
Summarizing the results in Subsections 6.1.1—6.1.6 we obtain.
Theorem 6.1.1. Eqs. (6.1.6), (6.1.8), (6.1.10)-(6.1.12) with the corresponding initial conditions determine completely the expectation, the second order moment and the covariance function of the process Y(t) in stochastic linear differential system defined by Eq. (6.1.1). Remark.
Theorem 6.1.1 is, basic for different correlation methods
stochastic differential systems.
The white noise in Eq.
of linear
(6.1.1) is understood in the
wie sense (Subsection 3.2.3).
Example
6.1.1. For the random process Y(t) determined by the equa-
tion
Y+aY
=V2DaV,
(I)
V being a white noise of unit intensity, Eqs. (6.1.6), (6.1.8), (6.1.10), (6.1.11) have the form
My =—am,, Dy =—20Dy+2Da, Ky = Dy, :
ly
—2aTy
+ 2Da,
The condition Ky, (ty ; Ti) =
OKy(ti, ta) /Ote =
—aKy(t1,
I t2) c
(
)
1), (ti) will serve as the initial condition for the last
equation.
Example
6.1.2. For the vector process which is determined by the dif-
ferential equations
Yi =Yo2+qV, Yo =—b?Y; — 2aY2 + qV,
(I)
6.1. Spectral and Correlation Methods
563
where V is a white noise of unit intensity, Eqs. (6.1.6), (6.1.8), (6.1.10) and (6.1.11) have the form my oe
er
0
=
1
mg, 7
Mo
bees Pater:
7
=
—b?m, 0
a
2am,
—b?
)
71
is Zoo |+ | | teeed
0
1
0
—b?
q
OK a (th, ta) _— (ie aa oR) = K (t,t) 9 ae
Example
where
(IIT)
te[ Sk)? Zh [s]ieor om s
tion
(II)
(Vv)
6.1.3. For the random process Y (t) determined by the equa-
Y = (w+ 2at)Y — (4 2at)(a + bt) + b+ be V , V is a white noise of the intensity
7 =
2¢ Eqs.
(1)
(6.1.6), (6.1.8), (6.1.10),
(6.1.11) have the form
My = (w+ 2at)my, — (w+ 2at)(a + bt) +b,
(II)
Dat a 2at)Dy 2pkesg Ky = Dy, (IIT) Ty =2(u+ 2at)Dy + 2b°te?4* — 2(u + at)(a + bt)my + 2bm, , (IV) OK, (ty, te)/Ote = (4 + 2at) Ky(t1, te) (V)
6.1.7. Stationary Processes in Stationary Linear Systems
We consider a stable stationary linear system (6.1.1) under the influence of a stationary white noise with constant intensity v. In this case a, do, b are constant, the function u(t, 7) depends only on the difference
of the arguments, u(t,7) = w(t — 7), and formulae (6.1.3), (6.1.7) and (6.1.4) at to + —oo take the form (oe)
n= J e@a0d,
(6.1.13)
0 lore)
K= [wees w(eyrae,
(6.1.14)
0
ka eels
(tit tint = [re + 1)bvb? w(é)*dé (r > 0). (6.1.15)
564
Chapter 6 Methods of Linear Stochastic Systems Theory
It is clear from these formulae that under the influence of a stationary white noise the wide sense stationary process settles in course of time in a stable stationary linear differential system. It is also clear from these formulae with the account of (1.1.14) that the condition of stability of a system is not only sufficient but also necessary for the existence of a stationary process. As m and K are constant in a steady state we get linear algebraic equations for m and K,
am+a)=0, aK + Ka" +bvb™ =0
(6.1.16)
putting in Eqs. (6.1.6), (6.1.8) mn =0, K =0. If the initial values mp and Ko and K satisfy these equations, then
Eqs. (6.1.6), (6.1.8) have the evident solution: m = mo, K = Ko. In this case at any to the process Y(t) is wide sense stationary. To find the covariance function k(r) of the stationary process Y(t) in a linear system we transpose Eq. (6.1.11) after changing the order of the variables ¢; and ft, in it.
Then remembering that the matrix
a(t) = a is constant and that K(t2,t,)’ = K(t:,t2) we shall have at SS
UG: OKz(t1, te) /Ote =
aK (ty, to).
Putting
here
K (ty, t2) =
k(r),
7 =t, —t2, tg =t and performing the change of variables t} = t +7 we obtain the equation
dkla) (das ak(t)igete0..
(6.1.17)
This ordinary linear differential equation with the initial condition k(0) = K determines k(r) at r > 0. At r < 0 according to the property of the covariance function of a stationary random process it is determined
by the formula k(r) = k(—r)?. Corollary 6.1.1. In asymptotically stable linear stationary system defined by Eq. (6.1.1) under the influence of the stationary white noise the wide sense stationary process settles with the expectation, the covariance matriz and the covariance function given by Eqs. (6.1.16), (6.1.17). Example
6.1.4. Under the conditions of Example 6.1.1 the variance Dy
and the covariance function ky (r ) of the stationary process Y (t) are determined by the equations
—aD,-eD =,
7
0.
For the existence of a stationary process in the system the condition
@ >
necessary and sufficient.
dky(7)/ar —
cb, (t),
&,(0) =
(Q) is
6.1. Spectral and Correlation Methods
Example
K
6.1.5.
For Example
6.1.2 at a@ >
565
(0 the covariance matrix
and the covariance function k(r) of the stationary process are defined by the
equations
Dl
AP
one
tle:
11
a
[ie a |R+K[Y Dn |+E |lead =o. dk(r)
_
dr
“A
0
1
tsb? | —2a
[k@), 7 >0, KO)=k.
Notice that the obtained results are easily extended to nonstation-
ary linear systems (6.1.1) at constant a, b and vy and an arbitrary time
function a(t). In this case the equations obtained for K and k(r) remain valid, and m represents a function of time determined by Eq.
(6.1.6).
The process Y(t) is covariance stationary in the system for which m is determined by Eq. (6.1.6) at any initial condition, and K and k(r) are determined by the method outlined.
6.1.8. Spectral Methods
Let us consider a stable stationary linear system with the transfer functions ®(s) whose input is a stationary random process X(t) with the spectral density s,(w). We shall prove that the output Y(t) of this system is a steady state, i.e. under the infinitely long action of the input represents a stationary random function. > We represent the random process X (t) by the spectral representation (Subsection 4.8.6)
fore)
X(t) = mz (t) + / e*V(w)dw .
(6.1.18)
On the basis of the superposition principle to this spectral representation it corre-
sponds the spectral representation of the output
Y(t) = my(t) + / e”' B(iw)V (w)dw .
We find the covariance function of the random function Y (t):
roo)
Ky fists) = / eft ciwia(iw)s,(w)B(iw)*dw —oo
(6.1.19)
566
Chapter 6 Methods of Linear Stochastic Systems Theory [o.0)
= / ei (t1-t2) 6 (iw) s_(w)B(iw)*dw .
(6.1.20)
—OO
Hence it is clear that the output of the system in a steady state represents a sta-
tionary random function of time. Let us find its spectral density. Subsection 4.8.8 at (o)
ky)
w =
Using formulae of
A:
oo
J slwyelerae,
Sy(w) = = i k.(aje
2dr,
(6.1.21)
we come to the conclusion that the spectral density of the output of the system Y (t) is determined by
Sy(w) = B(tw)s,(w)P(iw)*.
(6.1.22)
Now we shall find the cross-covariance function of the input and the
output of the system. According to (6.1.8), (6.1.9) we find Kt)
= / etw(ti-ta) (Ww) @(iw)* dw .
(6.1.23)
—oco
Hence it is clear that the input and the output in a steady state are crossstationary and the cross-spectral density of the input and the output of the system is determined by
Sry (W) = Sa(WlOw) Remark.
(6.1.24)
Formulae (6.1.22)—(6.1.24) are valid only for stable stationary
systems functioning in a steady state.
Practically the formulae are applicable when
the time of input action exceeds the time of the transient process. In should be noticed that if a system is described by a differential equation and consequently, its transfer
function is rational, then the output may be also a stationary random function for any time of system functioning with special initial conditions.
initial value of the output Y (to) =
Namely, the random
Yo should be chosen in such a way that the
covariance matrix of the random vector Z(t) =
aus e@ea hl ie be independent
of t. For this it is sufficient to take the random initial value Yo whose covariance matrix and cross-covariance matrix are equal to co
Ky, = &0) = / D(tw)s_(w)P(iw)* dw , —0oco
Ke
(Ska
fos
Oe / Sz (w)@(iw)* dw .
(6.1.25)
6.1. Spectral and Correlation Methods
567
But the appearance of these initial conditions in real systems is practically negligible.
So we proved the following proposition.
Theorem 6.1.2. The output Y(t) of asympotically stable stationary linear system with transfer function ®(s) under the influence of the wide sense stationary input X(t) in a steady state ts the wide sense stationary process with the spectral and correlation (6.1.21)-(6.1.24).
characteristics given by Eqs.
> For calculating the variances and the covariances of the output components of a stable stationary system functioning in a steady state under the action of a
stationary input at every given instant f it is sufficient to put in formula (6.1.20)
t; = tg = t. Then we get the following formula for the covariance matrix of the value of the output at the instant f: co
Fie be) = / Bia lee oan \ides
(6.1.26)
—oo
In practical problems, to find the variances and the covariances of the output
components by formula (6.1.26) we have to calculate the integrals of rational functions. For calculating such integrals it is convenient to use the formula
i)bo (iw)??? + by (iw)?"-3 + --- + bon—1 (iw) + ban—2 4
|ag(iw)” + ay(iw)?-1 + +--+ an_1(iw) + ap?
SI flee
6.1.27 (6.1.27)
——D
Here
Naess
(oil
“Gie
“Gao
Ara
Coie
COD)
ewe
Cn
Cni
Cn2
+--+
Cnn
Dn
=
bo
OW
oon
ae
bo
COD
anne
COR
bon—2
Cn2
+++
Cnn
where Cypg = G2p—g at 0 < 2p— q < Nj Cypg = Vat Ip—q< Vorlp—q>n. For using formula (6.1.27) it is necessary to represent the numerator in (6.1.26) in the form of a polynomial relative to tw and the denominator as the square of the modulus of the polynomial relative to 2W with positive coefficients dg, @1 , ... , Gn. We may also use formula (6.1.27) for calculating the covariances of the input components with the output components according to the formula
Toy = eae0
/ S$,(w)®(iw)* dw. < 12,6;
(6.1.28)
568
Chapter 6 Methods of Linear Stochastic Systems Theory
Remark.
The solution of the linear differential equation for the covariance
function k(r ) of a stable system represents a linear combination of functions containing damped exponents.
To each such function it corresponds a rational spectral
density. Thus the stationary process in the stationary linear system (6.1.1) possesses the rational spectral density. We may also come to this conclusion if we use formula
(1.2.57) at s = iw: ®(iw) = —(a — wwI)~'d. Example
6.1.6. Find the variance of a random function whose spectral
density is determined by the formula Sy (w) =
Da/n(a?
+ w?).
According to
formula (6.1.26) we find
Dele xpi eeelt eeeBie Doe ee y oT / a2 + w2 1 (64
(G9)
up eee =ad
a
(a9)
/ iw + a|’
—0oOo
In this case
2 = 1, G9 =
h we have K(I,h)
SK (hae). In the case of stationary discrete linear system (6.1.29) the matrices b; and the vector ag; do not depend on the discrete time I: a; = a, b; = 6, ao} = a and Eq. (6.1.29) takes the form Yah
=
aY;
+ a9 +
bY, .
i
(6.1.33)
The variables V; have one and the same distribution, and consequently one and the same covariance matrix G. If there exists the stationary
discrete random process {Y;} in Eq. = m; = mj41
(6.1.33) then its expectation m
and covariance matrix K = K; = Kj4, are constant.
By
virtue of Eqs. (6.1.30) and (6.1.31) they satisfy the algebraic equations:
m=am+a),
K=aKa! +6G0' .
(6.1.34)
It is evident that these equations have the unique solution if the matrix
(I — a) is convertible. Thus we proved the following statements.
Theorem 6.1.3. Recurrent Eqs. (6.1.30)-(6.1.32) with the corresponding initial conditions determine completely the expectation, the covariance matriz and the covariance function of the process Y; in stochas-
tic discrete linear system (6.1.29). Corollary 6.1.3. In asympotically stable linear stationary system defined by Eq. (6.1.33) under the influence of the stationary white noise the wide sense stationary process settles with the expectation, the covariance matriz and the covariance function given by Eqs. (6.1.34).
Now we consider the more general equation: A(V)¥,
=c+ B(V)V. ,
(6.1.35)
where A(V) and B(V) are the matrix polynomials over the shift operator V: pri
AVY STVe Saya r=0
P
BV y= Ope s=—N
6.1. Spectral and Correlation Methods
yal
(the polynomials B(V) contain also the negative degrees of V if N > 0). According to the spectral theory and under the conditions of asymptotically stable linear system the spectral density s,(w) for the stationary sequence {Y;} is expressed in terms of the spectral density s,(w) of the
sequence {V;} and the transfer function ®(z) of Eq. (6.1.35):
Sy(w) = B(e™)sy(w)B(e™)* .
(6.1.36)
Consequently, 20
ies }syed Preteen wyeteteyt da
(6.1.37)
0
As the variables V; are independent then the spectral density s,(w) of the sequence {V;} is constant: s,(w) = G/2z where G is the covariance matrix of the variables V;. Substituting this expression into (6.1.37) we get 20
ss az| Be
Gaede
(6.1.38)
0 After the change of the variables z = e formula (6.1.38) will take the form ; “
as K =~
z P9 U(2)GH(e7)*=, *\* —
(6.1.39) 1.
where the unit circle |z| = 1 with the curve counterlock-wise serves as the integration counter. Later on it is expedient to pass to a new variable for the calculation z—1 Ww
eaten
tg—.
©
Then the unit circle of the plane of the variable z turns into the real axis
of the variable A and formula (6.1.39) will take the form
eaters, Ls Bay “asked k=> f o(5) co (+) ae
(6.1.40)
—oo
As the transfer function ®(z) of Eq. (6.1.35) is the rational function of z then the integrand function in (6.1.40) will be the rational function
5t2
Chapter 6 Methods of Linear Stochastic Systems Theory
d. This fact gives the opportunity to use for the calculations of the covariance matrix K the known formulae for the integrals over the rational
functions (Subsection 6.1.8). In some cases it is also expedient to find the covariance function
k(1) = E(¥, — mp) (Vidi — mh 41)
(6.1.41)
of the stationary sequence {Y;}. For this purpose we use the formulae of the connection between the spectral-correlation characteristics (Subsec-
tion 4.8.6): 20
,
k(l) = ae| Be Goer ue
ds,
(6.1.42)
0
Then the correspondent formulae with the variables of the integration z and A have the form 20
ky (I) = 5 ib(e*”)GH(e)*e"™ du ,
(6.1.43)
0
pai f Dunes (5) or(eaettalh Ge (5) 1+4éd\ aya (144d)?! n(v=—
144)
Formula (6.1.44) is more convenient for the calculations than formulae (6.1.42) and (6.1.43) as the integrand in it is rational in consequence of which we may use the known formula (6.1.27) for the integrals over the rational functions for the calculation of the elements of the matrix k/(I). Remark.
As regards to the constant expectation of the stationary sequence
Algats it is defined by the trivial equation: A(1)my the unique solution My
the polynomial
= c. This equation has always
= A(daic as 1 cannot be the root of the determinant of
|A(z)| in the case of the asymptotically stable system (6.1.35).
Thus we have the statement.
Theorem 6.1.4. In asympotically stable discrete linear system (6.1.35) under the influence of the stationary white noise the wide sense stationary process settles with the spectral and the correlation characte-
ristics defined by Eqs. (6.1.36), (6.1.38)-(6.1.40), (6.1.42)-(6.1.44).
6.1. Spectral and Correlation Methods
6.1.10.
573
Correlation Theory of Linear Transformations
Let the expectation m,(t) and the covariance function K,(t,t’) of the input X(t) be given. Let us find my(s) and K,(s, s’) of the output Y(s) for the linear transformation
Y(s) = AX(t),
(6.1.45)
where A is an arbitrary linear operator. This problem is easily solved if we assume that the operation of the expectation and the operator A are commutative. It is often valid for linear operators in the problems of practice. ; According to the properties of the first and the second order moments we have
EY(s) = E[AX(t)] = AE(X(t),
BL Y°(s)¥%(s') |= B |AX°(t) AX) |
(6.1.46)
—E |AvAeeX°(t) XD |Aer |x°(t)X°@) | m,(s) = Am,(t),
"Ky(8,8 ) = AvAy Kolt,t) = AvAiKe(t,t).
(6.1.47)
(6.1.48)
Here the index at the operator A indicates that this operator acts upon the function of a given argument at the fixed values of all other variables.
It is evident that formula (6.1.46) is also valid for the initial moment of the second order: Ty(s, 5’) = Ay AT c(t, t’) = Ay Atl e(t,t’).
Thus formulae (6.1.44) and (6.1.48) are at the basis of the correlation theory of the linear transformations of random functions. Hence the correspondent formulae of Subsection 6.1.1-6.1.9 are obtained as the special case for the concrete operators A. The general solution of this problem is given in Sections 6.3 and 6.4. 6.1.11. Spectral and Correlation Methods for Stochastic Infinite-Dimensional Linear Systems
Let us consider a linear stochastic differential system in a separable
H-space Y (Subsection 5.9.1): dY = (aY + ao)at + bdW.
(6.1.49)
574
Chapter 6 Methods of Linear Stochastic Systems Theory
Using formulae
of Subsection
6.1.10 we
get for expectation
m,
m
= EY°®, covariance operator K, KA = E[Y(A,Y?)], \ € Y where Y,° = Y; —m of the process Y; = Y(t) at a given instant ¢ the following equations:
m=am-+ao,
KX =aKi+ K)a* +A
(6.1.50)
(c = bvd*).
For covariance operator K(t;,t2)A2 = E DAs (Aa, ae
(6.1.51) at t; and te
we have
OK i(t1, to)Ao =e Pa Ao a) ate 2 ts
99)(621.52)
2
The following conditions serve as the initial conditions for m(t), K(t)A and K(t1,¢2)A2 in operator equations (6.1.50)—(6.1.51):
mo = m(to), K(to)A= E[YLA;Y2)] , K(t,t)A2 = 2 [YP (2, YP)] (6.1.53) For determining the processes stationary in the wide sense in (6.1.49) at a, ao, b and o independent of time it is sufficient to put in Eqs.
(6.1.50) and (6.1.51) rn = 0 and K =0. As a result we obtain the equations:
am+a,9=0,
aKA+ KiXa*+oA=0.
(6.1.54)
For the covariance operator of the stationary process
K(t1,t2)A2 = E[Y,2(A2, ¥2)] =k(r)A2 (ti — te = 7) Eq. (6.1.52) takes the form
dk(r) dr
= ak(r).
(6.1.55)
In the correspondence with the formulae of Subsection 4.8.6 the following equation for the spectral operator sy(w) holds:
Sy(w)Az = B(iw)s,(w)P(iw)* ,
(6.1.56)
where sy = sy(w) is a spectral operator of the white noise V, ®(iw) = —(a—iwl)~'b. Thus we get the following propositions.
6.2. Methods of General Linear Stochastic Systems Theory
Theorem
6.1.5.
Eqs.
575
(6.1.50)-(6.1.52) with initial conditions
(6.1.53) determine completely the expectation and covariance operators in stochastic linear differential system defined by Bq. (6.1.40). Corolary 6.1.5. In asymptotically stable stochastic linear differential system (6.1.49) under the influence stationary white noise the wide sense stationary process settles with the spectral and the correlation operators defined by Eqs. (6.1.54)-(6.1.56). Example
6.1.8.
Under the conditions of Example 5.9.1 Eqs.
and (6.1.51) at constant c, h, v for mj; =
m;(x1, £2, t), Ki; =
(6.1.50)
Kal 2i, £20)
have the following form:
OK
pe
Si
Om,
ete
Koi + Ki2,
age
OK
—
Ot
Om,
= Kapa?
seit = ¢?
6.2. Methods
OK 12
ak
,0?m,
WO
Ko4+¢?
Ohad
One
07 Ky,
Oa?
+hkKi2,
+2hKo1,
0?K.
1
(1)
a
2
+ 4hKoo+v.
(II)
of General Linear Stochastic Systems Theory
6.2.1. Equations for Characteristic Functions
In the case of a linear system (6.1.1) where V being the white noise (in the strict sense) Eqs. (5.6.48), (5.6.49) as a(0/i0X,t) = —ia(t)O/OA + ao(t) take the form
0 0 oa - a
+f [irT ao + x(b7 d; t)|1,
=f = Malta) (>> aun) aos [iA7ao+ x(b7 A;t)] g.
(6.2.4)
kth
The standard algorithm for integrating equations of the first order in partial derivatives linear relative to the derivatives consists of the following operations. (i) The integration of the corresponding set of ordinary differential equations which has in this case the form
dt Sais addy 1 P
- Yann (=i
ee
a AdX» P
ean
dg ee iA? bT dst
AP UE
AD
ees) (
eal
This set of equations is obtained from the original equation in partial derivatives in the following way: the differential of each of the variables t, A, ,---, Ap is divided by the coefficient of the derivative of the function g with respect to this variable, the differential of the function is divided by the right-hand side of the equation and the ratios thus obtained are equalized to one another. (ii) The general solution of Eqs. (6.2.5) is represented in the form
CHEATS = Apso) = CE (k= eee Poe ie
(6.2.6)
where c;, ... , Cp, Cp4i1 are arbitrary constants (i.e. in the form of p+ 1 independent first integrals of these equations). (ili) An arbitrary function f(x1, ..., zp) of p variables is taken and the equation is written Gn+i(G AT ese; Ap,9)
6.2. Methods of General Linear Stochastic Systems Theory
nif
(qoy (tpg tar ahr py) (>
og,
577
(ty AgiyrenyAghg))%
(6.2.7)
The solution of this equation relative to g depending on the arbitrary function f represents a complete integral of the equation in partial derivatives. > Let us apply this algorithm for the solution of Eq.
(6.2.1).
In this case
the coefficients of the derivatives are linear function of the variables A; , ... independent
, Ap
of the unknown function g, and the right-hand side of the equation is
linear relative to g. Rewriting the set of Eqs. (6.2.5) in the Cauchy form we shall have
dX
=
sa
meg
he
hy cy).
(6.2.8)
d
= = [iTap + x(07A;t)] 9.
(6.2.9)
The first p of these equations form a set of linear homogeneous to Ay hes
equations relative
ibe Ap whose matrix of coefficients represents the transposed matrix of the
coefficients of the original Eq. (6.1.1) with the sign reversed. In matrix form this set of equations is
afd
oe
(6.2.10)
In order to find the general solution of this equation we denote by u(t, T )the solution of the homogeneous equation du/dt =
au with the initial condition u(r, T) =
ls
t > T > to. Then the solution v(t, T) of the adjoint equation dv/dt = —aly with
the initial condition v(T, 7) = I is determined by formula Ot, TF) = us; mut (Subsection 1.2.2). Consequently, the general solution of Eq. (6.2.10) is determined by formula
Nea(tw):ey ‘
;
A
‘
where C is an arbitrary constant p-dimensional vector,
(6.2.11) C =
[cy x Goo
ae
F
, and @ is
an arbitrary time instant within the interval [to, 1): Substituting expression (6.2.11) into Eq. (6.2.9) we transform it to the form
dg/dt = [ic’ u(t, ~)~'ao(t) + x(b(t)? u(t, a)1%;t)] g.
(6.2.12)
The general solution of Eq. (6.2.12) has the form
9 = Cp41 XP
/ [ic? u(r, 2)*ao(7) + x(b(r) u(r, a)""Fe;7)]dro, Qa
578
Chapter 6 Methods
of Linear Stochastic Systems
where Cp41 is an arbitrary scalar constant.
pression of ¢ from (6.2.11) to Cp41
Theory
Substituting into this equation the ex-
¢c= u(t, a)? ), and solving the obtained equation relative
we get t
Cp+i
i
i= 9 eXP4i—
[iAT u(t, w)u(r, «)~*a0(7)
/
+ x(b(r)P u(r, a)
(6.2.13)
ule, a)Ter)|dr :
But according to Subsection 1.2.2 for anya
s(, Ge
44: Sa
As
(6.2231)
s=1
Analogously we find a difference equation of the n-dimensional characteristic function: Jiypntae
="exD (11, doe. | BE; Anion, As solving Eq.
toils
g) Wee Ko
kok
Ave)
es Ant, deen) | 102-38)
(6.2.38) at the initial conditions (5.3.78) we get the
recurrent formula for the mu!ti-dimensional distributions:
Di ipech a Ab
iesaeesssa)
sa ae Oleg A Bierce, id sierotok cde
T ieT 980 T boot emt)
x exp {id aoz, }he, (bh An)
Kiel x I] exp {id7 ag, ...ds4140s}hs(br ay,
.. ae) :
(6.2.39)
si
Thus we have the following statements.
Theorem
6.2.3.
Finite formula (6.2.37) and recurrent formula
(6.2.39) give the explicit expressions for all multi-dimensional characteristic functions of discrete linear system (6.1.29). Corollary
6.2.3.
At the normal distributions of the variables Vy,
and Y; Eqs. (6.1.37)—(6.1.39) completely determine all multi-dimensional normal distributions of the sequence {Y;}.
6.2. Methods of General Linear Stochastic Systems Theory
591
Theorem 6.2.4. If there exists a stationary process in a stationary discrete system then it is stationary in the strict sense. The one-dimensional distribution is determined by Eqs. (5.3.76) at gi(A;t) =
gi(A).
6.2.7. Theory
For
Methods
the
of Stochastic
stochastic
linear
Infinite-Dimensional
differential
system
Linear
in the
Systems
separable
H-space y:
dY = (aY + ao)dt + bdW
(6.2.40)
equation for the one-dimensional characteristic functional of Y = Y;
g = 91(\;t) = Eexp [2A, Y(t)]
(6.2.41)
may be presented in the following form:
= eriotir a eco 500,03)|Re
(6.2.42)
Recall that in (6.2.40)-(6.2.42) A is an element of the space Y, W = W(t) is the Wiener process with the values in the separable H-space W, the covariance operator whose values at each ¢ are determined by formula (5.9.3) and its intensity v at each t is the trace-class self-adjoint operator; a = A+f, where A is the closed linear operator which satisfies conditions (iii) of Subsection 5.9.1, f; = f1(t) is the bounded linear operator at each t; aj = ao(t) is some element of the H-space J; (A, ao)
is the scalar product of the elements A and ap from J; b = Ob(t) is the operator mapping W into Y; o = by b* is the operator which satisfies con-
ditions (iv) of Subsection 5.9.1. Condition (5.9.10) serves as the initial condition for Eq. (6.2.42). In the case when the operators ao, a; and o do not depend on time
supposing 0g/0t = 0 we come to the following equation for determining the stationary process in Eq. (6.2.40): 1
(A, aDyg) + | t(A, ao) — 3 (As CAG
0"
(6.2.43)
Theorem 6.2.4. Under the conditions (i)—(iv) of Subsection 5.9.1 the one-dimensional characteristic functional (6.2.41) of the process Y; in stochastic linear differential system (6.2.40) ts defined by Eq. (6.2.42).
592
Chapter 6 Methods of Linear Stochastic Systems Theory
Corollary 6.2.4. Under conditions of Theorem 6.2.4 in statwonary system (6.2.40) the one-dimensional characteristic functional of the strictly stationary process is defined by Eq. (6.2.43). Example
6.2.5.
Let us consider a linear stochastic differential system
which is described by the following stochastic wave equation relative to
0°U See on
pou ee ~ ° Ox?
with the boundary conditions U(0, t) =
present
Eq.
(6.2.42)
I
Ot
U2)
()
ter 0), "Bere.
V =
for the
V(z,t) is a
Vy ==p (£1; r2,t); c, h are the
Yj = U, Yo = (0/dt)U, Y =
< Yo ws Poh we
U a t):
OU ee)
normally distributed white noise of the intensity constant coefficients. After putting
U=
[Yi Y2 i,
=¥ 0 ro fal
one-dimensional
characteristic
a fucntional
g
= g(A1, A2; t) >. ¢0){i(Ai ; Y;) + i(A2, Y2)} in the form Og
OE =
1
:
07Y,
—902 vr2)g + in|Jo,¥%) + (, ott)
+2h(a, Ya)SON
ain}
(II)
2l
where the scalar product is given by the formula (9, wv) = f p(x)p(x)dz. 0
As it follows from Eq. (5.9.9) the multi-dimensional characteristic functionals g, which are determined by (5.9.2) (n = 2,3,...) satisfy Eq.
(6.2.42). Here (5.9.10) and (5.9.11) will be the initial conditions. 6.3. Methods of Linear Stochastic Canonical Expansions
Systems
Theory based on
6.3.1. Introductory Remarks In elementary theory of the random processes the method of canonical expansions is used generally for the correlation analysis and simulation of the nonstationary processes which are obtained as a result of linear (integral, differential, discrete and etc) transformations. Hence we
distinguish two methods.
6.3. Methods based on Canonical Expansions
593
According to the first one the canonical expansions of the random input and weighting (transfer) functions of a system are supposed to be
given. Let the random input X(t) and its covariance function K,(t, t’) of an one-dimensional system be given by a canonical expansion of the
form (3.9.6) and (3.9.7):
X(t) = m,(t) + veViay(t), Ke(t,t') = 54Dyay(t)a,(t').
(6.3.1)
For linear integral transformation (1.1.13) -s
t
Y(t) = ‘;remapeeun
(6.3.2)
to
where g(t, 7) is a weighting function of a system the output Y(t) and its covariance function will be expressed in terms of the canonical expansions:
Y(t) =m,(t)+ )Vow(t), Ky(t,t’)= >> Diw(t)y(),
(6.3.3)
Va.
and
t
t
wry J str)me(nyar, wee [ot.neu(rjar resi els ) to
the second method sions of a random which describe the the correspondent pectations and the Example
to
(6.3.4) supposes the knowledge about the canonical expaninput (6.3.1) and the differential difference equation system. In this case it is necessary to set up and solve linear differential or difference equations for the excoordinate functions and later to use formulae (6.3.3). 6.3.1. Find the variance of the output in the system
EVE VereXs
(I)
on whose the white noise input acts with constant intensity / beginning with to
= 0. The system was at rest till the instant tg = 0. The weighting function of the
594
Chapter 6 Methods of Linear Stochastic Systems Theory
system g(t, 1E) was determined in Example 1.1.6.
Therefore we have the following
exact solution for the variance:
Tal AVteney f feOC D;RS @=7
FER 6(r asr)drdr!imalia = = (1 pa €
9 is
0 0
(II)
Hence at t — ©O we find DES (0) = vy[2T. For applying the first method we use a canonical expansion of a white noise at
the interval 7} obtained in Example 3.9.10.
Therefore we have for the coordinate
functions according to formula (6.3.4): t
yr (t)
=
1
.
TT, /e —(t-7)/T+iwyt gyT
=
e BWyt
e~t/T
(IIT)
T%(1 oe iTwy,) ;
Substituting this expression of the coordinate functions into formula (6.3.3) at t = t’ and accounting that D, = v7
we get oo
eivnt _ o-t/T
The second methods gives for the coordinate functions a linear differential equation
Bie TYn + Yn =
While integrating Eq.
me
(V)
(V) at zero initial condition we shall obtain formula (III).
For the estimate of the accuracy of the presentation of the variance Y by a finite
segment of series (IV) we notice that at
>> T' formula (IV) may be approximately
substituted by the following one: co
D,(t) # Dy (co) = = Me rea}
ss
(VI)
V
=—{142 oe
at
Y T? aE = + elias 49227?
In practice we may use formula (VI) at t > 37°. In particular, at Tj /To = 4 while restricting ourselves to the first six terms of the series in formula (VI) we find
y
6
(|
8
1,94
6.3. Methods based on Canonical Expansions
595
Thus at restricting oneselves to the first six terms in (VI) what corresponds to the restriction in the canonical expansion of a white noise by the first 13 terms (i.e. by 6 harmonics) we assume an error about 3%.
Remark.
Example 6.3.1 convinces us in the fact that for the analysis of linear
dynamic systems we may practically use the canonical expansions of input random
disturbances restricting ourselves by moderate number of the terms even in those cases when it is necessary to take very large number of the terms of these canonical
expansions for the presentation of the disturbances themselves with sufficient accuracy. Usually knowing the cut frequency (Subsection 1.1.7) it is possible to estimate
approximately what coordinate functions in the canonical expansion of the output
variable should be known and what functions can be negleced.
6.3.2.
Theory of General Linear Transfomations
Theorem 6.3.1. Let T be a linear operator mapping the phase space X of the random variable X into some linear space Y, conttnuous in the weak topologies of the spaces X and Y determined by the sets of the linear functionals F and G respectively and measurable relatively to the o-algebras of A and B generated by the classes of the S6iSe1 yt (sie awe, fal) Ct and ye: Oe td. Gn Y) Cea at allfRie SES hia o. oe Inne he Owens dn CG and Borel seis A, B of the correspondent spaces. Then the operator T will be (A, B)-measurable > For proving the (A, B)-measurability of the operator J’ we notice that by
virtue of its continuity the mapping g7' at any g € G represents a continuous linear functional on X.
As in the weak topology only those linear functionals by which
this topology is determined are continuous then g7' € F'.Consequently, in this case there exists the adjoint operator 7
G into F, T*g
determined on the whole space G which maps
= gT € F-. Therefore the set {y : (gy, ra
serves as the inverse image of the set {x ; ((T* 91)x pote
Jn) Ee B} EB
& (T*gip)2) fc B} EA.
As the inverse set Ap of the o-algebra B is the minimal o-algebra which contains all the sets Nes : ((7* oi )zt oe
om (T* gn )a) E B} then Ap
C A what proves
(A, B)-measurability of the operator J’.
Ww.
(6.3.9)
v=1
After obtaining canonical expansion (6.3.5) of the random variable Y we may find its covariance operator by formula (3.9.18). Theorem 6.3.2. If the linear operator T is continuous in the weak topologies of the spaces X and Y then canonical expansion (6.3.9) of the
6.3. Methods
random variable space Y.
based.on
Canonical
Y = TX +b m.s.
Expansions
597
converges in the weak topology of the
> According to (6.3.3) for any g € G the formula is valid n
E
2
oY a ie Woyw|
nm
=£
Dak os aE as
y=)
law |?
v=1
Sila
x
n
ams ode
OT 25 lie
Ves
By virtue of the continuity of the operator J’ there exists the continuous
adjoint
operator 7™* determined on the whole space G and gJ'x = (T*g)x for any
© € X
and here 7*g € F’. Thus using again formula (6.3.7) we get a
2
B \g¥°-"Vigw|
n
= 8 \(T*9)X°| — >> Dy |\(T*9)2r)? yal
y=]
mn
2
= EB \(T*g)X°-)_V(T*g)av|
.
vill
The latter expression tends to zero at 2 —
OO because of the weak m.s. convergence
of the canonical expansion of the random variable X what proves the weak m.s.
convergence of obtained canonical expansion (6.3.7) of the random variable Y. < Remark.
The proved theorem with some modifications is extended over
more large class of the linear operators. Namely as it is easy to see all our arguments remain valid for any linear operator J’ which has the adjoint operator 7™* except
that the canonical expansion of the random variables gY will converge not for any
functional Dy
g € G but for those functionals
g € G which belong to the domain
of the adjoint operator 7* and are mapped by the operator 7™ at the points
of the space F’. Thus canonical expansion (6.3.9) of the random variable Y in this case m.s.
converges only in more weak topology which is determined in the space Y
by the set of the linear functionals G’ belonging to the intersection of the space G' with such part of the domain of the adjoint operator 7* which is mapped by this
operator into F’. Here the operator J’ is measurable relatively to the o-algebra A of the space X and the o-algebra B’ of the space Y generated by the class of the
sets {y : (91y, 505
gnY) (E B} at all finite sets 9} ,
sets B of the correspondent finite-dimensional spaces.
... , Jn € G’ and all Borel It is proved similarly as the
(A, B)-measurability of the continuous operator J' if we account that T*g € F for any g € G’.
It goes without saying that this generalization has the sense only in
598
Chapter 6 Methods of Linear Stochastic Systems Theory
that case when the space G’ is sufficiently complete for the fact that from gy = 0
for any g € G’ will follow y = 0. Example
6.3.2. Let X bea space of the finite-dimensional vector func-
tions of the scalar variable t, £ € J
which are continuous
together with
their
derivatives till the order N inclusively, F’ be a space of the linear functionals on X of the form
fo =)
Nm
o> free (th).
Let us consider the operator of the differentiation
D =
(I) d/ dt.
The operator Dy
maps the space X into the space Y of the functions which are continuous together
with their derivatives till the order
N — | inclusively ( Salteaisrse JV ): We assume
the set of the linear functionals on
Y of the form N-I
m
gv = >) > gy (te) q=0
as the space G. in X.
(II)
k=1
Now we consider the random variable X (t) with the realizations
The operator D! transforms it into the random function Y(t) =
D'X(t)
= x”) (t) with the realizations in the space Y. As N-I
m
gDin= dd joe (te),
(111)
¢=0Rh=1
then the functional gD! at any g € G belongs to F’. Consequently, D! is continuous in the weak topologies of the spaces X Theorem
6.3.2.
the operator
and Y, and we may use
The adjoint operator (D')* in a given case is determined on the
whole space G and maps G on the subspace F” of the space F’: N
SEAS OF RO Eel)
m
NyEAL
(IV)
p=l+1k=1 Therefore after expressing the random function X (t) by any canonical expansion we shall obtain for its derivative of the order | the canonical expansion
XOW) = mY) +
UV2P@), tek. Vat
(V)
6.3. Methods based-on Canonical Expansions
Ex
a mp
le
599
6.3.3.4et X(t), t € 7, be a vector random function of the
finite-dimensional vector t, X be its phase space, F’ be a set of all linear functionals on X representable in the form
fe= fp(y?o(eyat for almost
all the functions
of the random
=
x(t) € X
function X (t)).
(1)
(relatively to the probability measure
Le
Let us consider the linear integral operator
Lx
J U(s, t)x(t)dt which maps X into the space Y of the m-dimensional vector
functions of the finite-dimensional vector s, s € S. Here we shall assume that at any fixed
s € S each row of the matrix |(s, t) determines some linear Mena Rog from F’,
We determine on Y a set of the linear functionals G of the form gy = > Ik2 y(sx). k=
Then for any g € G we shall have gLxz = i" > gf U(sx, t)x(t)dt. Putting b= N
FEED YV Se, 2) 85
(II)
we reduce this formula to the form (I). Thus in a given case the operator L is continuous in the weak topologies of the spaces X and Y formed by the sets of the linear functionals F' and G correspondingly To any functional g € G,, i.e. to any finite set of the pairs {sx, gk} the adjoint operator L* sets up the functional f € F' which is determined by formulae (I) and (II). Thus to any canonical expansion of the random function X (t) the corresponds canonical expansion
¥ (s)= my(s) + D> Vw(s)
(111)
v=!l
of the random function Vis) =
(oe Expansion (III) m.s.
LX(t
fom
=
I(s,t) $ )X (t)dt and here
Gish [ronenoae,
(IV)
converges at all s € S. After obtaining canonical expansion
(III) of the random function Y(s) we find its covariance function Fonsi ; S2). Example where
X (t) represents
6.3.4. Let us consider a special case of the previous example a white noise of the intensity v(t).
In this case we
take
the space of the main functions as F' and use the canonical expansion of the white
600
Chapter 6 Methods of Linear Stochastic Systems Theory
noise of Example 3.9.10.
As it was shown in this case any orthonormal system of
the functions {Yn (t)} with the weight v(t), fv(t)ep(t)? Gy (t)dt = bnq gives oo the canonical expansion of the white noise X(t) = Dy Vp Pp (t) which weakly p=1 m.s. aca aes in the sense that for any main function f(t) € F the series f X => Vp J ‘B f( (Ge Pp(t)dt m.s.
p=1 usual we assume equal to zero.
converges.
The expectation
Let us consider the random function
of a white noise as
‘
Y(s) = LX(t) aahI(s, t)X (t)dt, where each row of the matrix |(s, t) at any fixed from the space of the main functions.
(I)
s € S' forms the vector function
In this case the operator L is continuous in
the weak topologies of the spaces _X and Y and the stated method gives a canonical expansion of the random function ous Jr
Y(s)= yf (s, t)pp(t)dt , which m.s. converges at any Example
6.3.5.
(II)
$s€ S. Let X (t) be a finite-dimensional vector random func-
tion which represents a random variable with the phase space Lo Ger Le HL), po
We assume as F’ the adjoint space Iq (Ti on, Ht), q= p/(p = dy Let us consider the linear integral operator
us) = La = |Us, t)x(t)m(ae),
(1)
where |(e t) satisfies the following conditions: (i) at eachs
€ S I(s, t) belongs to the space Lq (Th, Win HL);
(ii) at each t € T} I(s, t) belongs to the space of the finite-dimensional vector
functions Lit? S, v) m>
1; (iii) for any function z(t) € bp( Ii, J, HL) the function y(s) = i I(s, t)ax(t) (dt) belongs to the space Lm(S, S,v);
=
(iv) for any function g(s) which belongs to ee space Ly, (S, So, V), n m/(m — 1) the function f (S)n= eS t)? g(s )v(ds) belongs to the space
prey ay
In this case for any z(t) €E Dea sini ie HL) and any g(s) € |be
y=
owiouc)
= [96s)
S, v)
(as) f(3,t)e(e)u(at)
6.3. Methods based. on Canonical Expansions
601
= [{[actes.ontas)} eoucan = [sPeOucay. It is clear from (I) and (II) that gl € F’, and consequently, the operator L from Ly (Ti ie jt) into Lm (S, 5 v) is continuous in the weak topologies determined by the correspondent dual spaces /’ = Das
Sy 7) and-G =
Eats, aN V). Conse-
quently, any canonical expansion of the random function X (t) which is weakly m.s.
convergent in Ly GEG J : Ht) gives canonical expansion (III) of Example 6.3.4 of the random function which weakly m.s. converges in Ly, (S; S; V),
Hi(s) as /I(s, t)X(t)u(dt)
(IIT)
where My(s) = f I(s,t)m(t)u(dt), yy(s) = f U(s,t)x,(t)u(dt). 6.3.3. Assignment of Probability Measures by Canonical Expansions As we have already seen in Subsection 3.9.2 to each vector f from the H-space H, connected with a given random variable X corresponds a scalar random variable (more exactly a class of the equivalent random variables) fX. The measurability of the function fx of the variable x for f belonging to the factor-space F'/N is evident. The measurability of fx for f € Hz, fEF/N follows from the fact that fx as the limit of the sequence {fnz}, fn € F/N of the space L2(X,A, uz) also belongs to L2(X,A, pz) by virtue of the completeness of Lo(X,A, 2). There-
fore the probability measure V = {V,} of the coefficients V, = f,X° of canonical expansion (6.3.6) of the random variable X is completely determined by the probability measure py of the random vector V on the cylindrical sets of its phase space,
(lio: Qa. (Aes
Ses
ty.) PER
Bt} eS Sein
(6.3.10)
Knowing the measure pl, on the cylindrical sets we may continue it uniquely by the theorem about the measure extension over the o-algebra B of the phase set V. Here the o-algebra of Borel sets of the complex plane is designated as B, and as N is the set of positive integers.. Our interest is in the inverse problem: knowing the probability measure [ly of the random vector V find the probability measure pz of the random variable X.
602
Chapter 6 Methods
of Linear Stochastic Systems Theory
> For solving this problem at first we establish the measurability of the linear co
functional pfu
=
se) Ui
Ter) relatively to the 0-algebra BN
at any f € Hyg.
v=1
But the measurability follows immediately from m.s. convergence of the series
Ey Wied ag EG
PO
(6.3.11)
v=)
at any f € H, (Subsection 3.9.3).
In reality from m.s. convergence of this series
follows the existence of such subsequence
Apt
that the sequence
of the random
Np
variables {pv} where gV
=
Se ViCfen;
(p ole
2) 2 .) a.s.
converges
v=1
to YF V.
In other words,
the sequence
of measurable
linear functionals
{v4uv}
converges almost everywhere (relatively to the measure /l,) to the linear functional pfu.
Consequently, the functional YU
is measurable.
Formula (6.3.6) which determines a canonical expansion of the random variable X establishes the mapping of the phase space of the random vector V into the space X. 91,
As the inverse image of the set {x
; (giz, BER
na,
Se B} at any
--», 9n € Hg and any Borel set B of the n-dimensional space at this mapping
by virtue of (6.3.11) serves the set {v : (giMmz Gal 5 pers e B}. This set belongs to the 0-algebra B
On tat t 9,V)
because of the measurability of the
linear functionals (yg, ,..., g,,- Knowing the probability measure /l, of the random vector V we may determine the probability measure [lz of the random variable X
on all the sets of the type {x : (Giese bona Ga®) Se B}:
pe({ee
(gies. & gheyesBy)
= po({u = (grime + gi¥, ---) InMs + ¥y,v) € B}).
(6.3.12)
The measure /lz is uniquely expended over the o-algebra A generated by the class of all such sets
= Be \=1
hehehe)
=e
\a
ee
Kai astaierad p=
-s
After substituting expression (I) for gz into (II) and taking into consideration that
fy Kalu = iis Fis) = Dy by p (Subsection 3.9.3) we get
1
5
(111)
=exp 4-5 | Dyhi ¢ go(h) pal
The formula shows that the random vector V is normally distributed and its components are independent. Example
6.3.7. Suppose that the random vector V of the coefficients of
canonical expansion (6.3.5) of the random variables consequently, its components
are independent).
X is normally distributed (and
Its probability measure jl, on the
measurable cylindrical sets is expressed by formula
pall coal upetiat Py, bee }) Sl (Om) n Dy deter in |—1/2
BoP
is % 3D
za
RPyid tat
(I) On the sets of the type {v ; (ADy } Bh ae:; AM)v) ‘e B} at any linearly indepen-
dent h{) Ns
taal hi) € H the measure fly is determined by formula
po({v i {AMv, , aes nro} & B})
=a (Qa) %elG| ew
{ster dz,
(II)
604
Chapter 6 Methods of Linear Stochastic Systems Theory
eo) ie |
where C’ is a matrix with the elements Coq =
AP) ALY (p, ie
pra:
Se n).
a1
Using formula (6.3.12) we find the probability measure of the random variable Xx:
Ho({e : (912, ---, nt) € BY) n
(2m
-1/2
1 ts feat exp 4 — 52 D2
ar,
(IIT)
B-a
where @ is a vector with the components dp =
gpMz
(p a=
matrix whose elements with the account that (Yo,) vy =
0° > Di (Gotu) Gav
formula dyg =
(p,q al
n); Disa
Jp,
are determined by
Weare Go n). After determining the
vol
measure [ly on the sets of such a type it is uniquely extended over the 0-algebra |
generated by these sets. Remark.
In a given case the measure
[ly may be determined in another
way. Namely, knowing the characteristic functional of the random vector V we may find the characteristic functional of the random variable
X : (oe)
28 Onlf)atbe:
SfX
eo
Afni
xe
ie
Lapis
a ee
. Py Vay)
exp
are
Be
a |
The latter expectation represents the value of the characteristic functional of the
random vector V at hy = fx, (v wal
9x(f) =e
aes Ay As a result we obtain
exp
uedy Drie)
:
(V)
[ims |
Example
6.3.8. Let X(t), t € T) be the r-dimensional normally dis-
tributed vector random
function,
F' be a set of all linear functionals of the form
N
fre
Ss fi z(t); t;,...,
tw € Tj. According to the results obtained in the
[il
previous example the multi-dimensional distributions of the random function X (t)
are determined by formula LOERe
GEG
= an
+12)
Un
a
: (a5) vey
tn) € B})
1
(Di | Nae / exp {-3e7 De dz.
(1)
B-a
Here a is the nr-dimensional
x (t{(p—1)/r]41) (p Se)
| ec
vector with the components
Ap =
Mp_-[(p-1)/r]r
nr), D is the matrix whose elements are expressed
6.3. Methods based.on Canonical Expansions
605
in terms of the coordinate functions ait) of the canonical expansion of the random
functions X (t) by formula
dpg =) /Dvahas ,a) =2y,-[(s-1)/r}r (tf(e-vy/r}41) (S=1, -.., nr), p=1 (II) Formula (V) of Example 6.3.7 gives the following expression of the characteristic functional of the random function X (t):
go(f) =expfi ff(t)? mo(t)dt-7 [[sotentoae] aay ya
Putting here, in particular, f(t) =
nor = Se. d(i-1)r4mem6(t — t;), where €1, i=ain=
., @, are the unit vectors of coordinate axes of the r-dimensional space we det the n-dimensional characteristic function of the random function X (ay.
6.3.4. Solution of Equations by Canonical Expansions The canonical expansions allow to find the solutions of the linear operator equations of the type
LK, =®,
(6.3.13)
where L is an unknown linear operator, Kz is a covariance operator of some random variable X, ® is known linearly-adjoint operator. In particular, the problem of finding the regression for the random variables X and Y with the normal joint distribution is reduced to this equation. In the latter case ® = Kyz. Let X be a phase space of the random variable X, F' is a space
of the linear functionals by means of which the weak topology and the o-algebra A in the space X are determined, Y is a space in which the operator ® maps F, G is a space of the linear functionals by means of which the weak topology and the o-algebra B in the space Y are determined. Usually in practice it is required to find not any solution of Eq. (6.3.13) (it may have not unique solution) but only such solution L that the random variable gLX will have the finite second order moment at any g € G. Besides that it is usually required that the operator L would be permutable with the operator of an expectation.
Theorem 6.3.4. If the auxiliary H-space H, (Subsection 3.9.2) connected with the random variable X is separable then any canonical
606
Chapter 6 Methods of Linear Stochastic Systems Theory
expansion of the variable X used for the representation of the solution
Eq. (6.3.13) in all the cases when tt exists. > Firstly, we define the necessary conditions for the existance of the solution
Eq. (6.3.13). Let D be a required solution of Eq. (6.3.13). It represents the (A, B)measurable
linear operator as only in this case L_X
(Ys B). Let {fv} be the orthogonal basis in Hz, variable X
will be a random
variable in
|| te [t= D,. Then the random
may be presented by canonical expansion (6.3.6) where
V, =
| PP.§ °
are uncorrelated random variables with zero expectation and the variances which are equal to the correspondent numbers D),,, and @, are the coordinate vectors satisfying
the conditions (Subsection 3.9.3):
Sn
De Tee
Sit
Oat
(6.3.14)
From the fact that the random variable g_X has the finite second order moment at any g € G and the operator L is permutable with the operator of an expectation
follows the convergence of the series [e.e)
SO Deigin
(6.3.15)
esi |
at any g € G. As by virtue of the first formula of (6.3.14)
n
:
E |gLX° — 5 V,(gLa,)| Vy = gLEX°V, — Du(gLt,) pI
=gLEX(f,X°) = D,(gLbz,)=9L Ke fy — Duglz,=0 (v=1,..-/n), n
then 1 jonx" =
2
Dy Vo(ube.)
n
=
it; ae
=
ae. IDs, \gLa,|°. Hence
Vii
Vet)
it follows that series (6.3.15) converges at any
g € G. Further, it follows from the
canonical expansion of the covariance operator:
eS
[e-e)
=
bs. py By
fey ), y € Hg,
vs co
and from the convergence of series (6.3.15) that LK, f =
3 D, (Le;
———w
(fez);
v=
where the series weakly converges at any f € Hz, in the weak topology of the space co
Y.
Substituting this expression into (6.3.13) we get y) Di(iey)( fev) =ubf; v=1
f © H,. From here and from the convergence of series (6.3.15) it follows that Eq. (6.3.15) may
have the required solution only in that case when the operator @® is
representable by the expansion
GEN VDI WP i
|
Tpy
(6.3.16)
6.3. Methods based-on Canonical Expansions
607
where Y, are the vectors of the space Y which satisfy the condition
Ss Ds low | < oo
atany
gEG.
(6.3.17)
v=)
For determining the coefficients y, we replace the index of the summation VY in (6.3.16) by the index
and apply the operator ® to the vector f,. Then by virtue
of (6.3.14) we obtain D, y, = ®f,. Consequently,
et) ge 0 fra Veet 2
red
(6.3.18)
-s
> The operator L in (6.3.13) is determined only at the domain Ak, covariance
operator
K,.
Therefore
the variable
L_X
has the sense
E IN ie... In this case there exists such a vector fo €
Hy
that mg
of the
only at Mz =
Ia foe
And as {fi} is the basis in Hy then fo = iD. Cut and by virtue of (6.3.14) Mz co
=
je Dz Cy, and this series converges in the weak topology of the space ACs lea H=1 order to obtain the explicit expression for the coefficients C, we apply to the previous equality for the functional f,. Then taking into accounting the second relation of (6.3.14) we get Dice = Jyms, (v A
ie
3) and consequently,
co
Mo = Devise
:
(6.3.19)
jr
Finally, noticing that on the basis of the first relation of (6.3.14) fymz = fy Ke fo Tear
= foKefy
¢
iS
= Dv foxy and accounting the convergence of the series 3 DF
at any f € Hz, we come
fz,
2
to the conclusion that the following condition must be
filfilled:
SO
yee
a
(6.3.20)
v=1
Theorem
6.3.5.
For the existence of the linear operator L which
satisfies Eq. (6.3.1), is permutable with the operator of the expectation and gives the finite second order moment of the random variable gLX,
conditions
(6.3.16), (6.3.17), (6.3.19, (6.3.20) are necessary.
We prove that these conditions are also sufficient for the existence
of the solution of Eq. (6.3.13).
608
Chapter 6 Methods of Linear Stochastic Systems Theory
> For this purpose we shall find the operator L in the form of the series
ie enn where for
7), are yet unknown
any
f
€
Hg
we
vectors. get
Then
(OY Goss 6 =.
(6.3.21)
taking into the consideration
0°
(6.3.7)
00
oe Nd
hal
=
pas Nd
pH
Kath
p=1
co.
=
my Dalton):
In order the operator L would satisfy Eq. (6.3.13) it is neces-
r=
sary gteet at
any
f
=> Dunu(fep)5 Dahl oole= = we 4
€
4H,
the
following
After comparing
equality
this expansion
b>DuYu( fen) at any f €
Ny = UW oss riydeh Ae
will
with
Hy.
be fulfilled:
(6.3.16) we
Putting here
Df
obtain f =
fy,
Substituting this expression into (6.3.21) and
accounting (6.3.18) we obtain a fomal solution of Eq. (6.3.13) in the form of series
b= = Yow) f, =Lely: 1(@f,) fy.
(6.3.22)
v=
Let us take a look in what sense expansion (6.3.22) determines the solution of Eq. (6.3.13) which possesses thedis eae prapertics. For this purpose we consider the
random variable
weakly m.s.
Y = LX = DSUp jA=
3 Up Ch Mz + Vz i This expansion
converges as it follows from (6.3.77) and (6.3.20) that at any € > 0,
g € G, for all p and at sufficiently large n 2
n+p
E
So (gw) (frre +V,)|
2
n+p
=
vr=n
S\(9w)( fume) v=n
n+p
:
n+p
ar py D, lay |? vr=n
n+p
n+p
vrn
vr=n
) Dy lawl’ - >> Dr" fmol? + 5 Dy lawl? D5V, rfKye(s,7) f-(r)dr. In the majority operator
of the problems
Ly to the random
function X Oe).
function
of the practice it is necessary
evan)
to applythe
X(t) but not only to the centered random
As we have already seen for the existence of such an operator it
is necessary and sufficient the fulfilment of conditiofs (6.3.19) and (6.3.20) besides
conditions (6.3.16) and (6.3.17). In a given case these conditions take the form
Mz(t) =oestt) [oePmelrir
py | fo"
metry] < OO.
(VIII) Therefore formula (III) determines the required solution of Eq. (I) and the series LX
=
SS. yp (s) pe f(t)? mz(t)dt + VN m.s.
converges at each
s € S to the
PSik
random variable with the finite matrix of the second order moments.
Example
6.3.10. Inthe special case of the finite set 7) = {t, rick stn}
all the functions f(t) € H,
are the linear combinations of the 6-functions with the
peculiarities at the points t; , ...
, t47, the canonical expansion of the random func-
tion represents the finite sum which contains no more than nN items (Example 3.9.8), and the formulae of the previous example give the solution of Eq. (I) of the previous example in the form of the finite sum.
large N the solutionof Eq.
The operator 1; in this case is a matrix.
At
(I) of the previous example which is presented by the
formulae of the previous example is considerably simpler for the calculation than the
solution which is obtained by the inversion of the block matrix Tie (us ; ip Example
the spaces X
6.3.11.
Example 6.3.9 is easily generalized on the case when
and Y represent the spaces of the functions with the values in any
linear spaces with the weak topologies which are determined by the sets of the linear functionals A, and Ay respectively.
In this case in all previous formulae f(t)’ is
a function with the values in A, at any t and the variables 9} v. ©, gn are the vectors of the space Ay. The convergence of expansion (III) of Example 6.3.9 at the
replacement of x(t) by X(t) or X(t) in a given case should be considered as m.s. convergence at each
§ € S in the weak topology of the space Y of the values of the
functions y(s) EY.
The stated methods may be also used for the calculation of the Radon-Nikodym derivatives of the output measure jy by input measure
Le, dpty/du, (Pugachev and Sinitsyn 2000).
6.4. Methods based on Integral Canonical Representations
613
6.4. Methods of Linear Stochastic Systems Theory based on Integral Canonical Representations 6.4.1. Introductory Remarks
The integral canonical expansions (Section 4.8) along with the canonical expansions (Sections 3.9, 6.3) are widely used for the analysis and the simulation of the random processes in the dynamic systems with random disturbances. The methods of the application of the integral canonical representations are analogous stated in Subsection 6.3.1. According to the first method if an one-dimensional linear stochastic system is given by a linear integral transformation (6.3.2), and an input and its covariance function are determined by some integral canonical
representation (Section 4.8):
X(t)=mett+>0 |¥ (A)z-(t, )dd, c— LAL
(6.4.1)
Kee) )=¥ [uw a(t, A)a,(t,A)dd, — ans
then the output Y(t) by virtue of (6.3.2) will be expressed by the integral canonical representation:
=m, (1)4°
fv r)yr(t,A)d
ral,
y(t) AC =
(6.4.2)
fr)Jur(t,A)yr(t!,A)dA r= 1A.
and
F
my(t)=f9(t,7)ma(r)dr to
ee
(6.4.3)
[ote rievtn Nye pees eos sana.
The formulae for the finite-dimensional linear system and the vector inputs are written analogously.
614
Chapter 6 Methods of Linear Stochastic Systems Theory
Now we consider the second method. Let the equations be given which describe the behaviour of a linear stochastic system. ‘Then for the correlation analysis we may use the first method after determining before-hand the weighting functions of a system. For obtaining the equations which for the expectations all random inputs and outputs should be substituted by their correspondent expectations in the equations of the considered system. While deriving the equations which determine the coordinate functions it is necessary to replace all random inputs and outputs by the correspondent coordinate functions in the equations of the system. Example
6.4.1. Find a variance of the output of the Spatianarss oscilla-
ting link
4
Y+a1Y =
for MN,
0, k(t)
=
D,e7
+aoY
217 -AtN=
(I)
=X
Li \ieweuhave
z(t,w) =
eiwt uel
y(t,w) + ar y(t,w) + aoy(t,w) = e*, y(0,w) = 0.
(11)
Hence after integrating we find
(65) we
OD
eet
(Az — tw)e*1* — (A, — iw)e>??
—————————
do tayiw—w?
©
ere
(A, —Az)(ao + ayiw —w?)
(11) ’
where 1,2 = —2 +17 are the roots of the characteristic equation part ajA+ ag = 0. Consequently,
aD,
Dy (t) = —T
/(cos wt— bee sin yt—e—** cos yt)? +(sin wt —cian sin yt)?
x | —-—_
F
(a? + w?) Jag + ayiw — w?/?
—0o
—
l
(IV) At t —
©o for the asymptotically stable system
(B >
0) we find the stationary
value of the variance from (IV)
D
(eo)
=
=
D,(a + a,)
aoa;(a? + aya + ao)
We may come to the same result if we account that [o.e)
Dy(00) = fsy()du, s4(w) = [®(iw)[? salu), 0
(V)
6.4. Methods based on Integral Canonical Representations
De
a
’
s2(w) = =, TW
at+w
615
i)
0(iw) = ——____, do + ayww — Ww
6.4.2. General Theory of Integral Canonical Representations of Linear Transformations > After expressing the random variable
X by the integral canonical represen-
tation (Section 4.8)
Rie sare: 55i;2,(A)Z-(d2),
(6.4.4)
we obtain for the random variable (6.3.5) the analogous canonical representation
d= prrtgn | 3} Yr (A)Z, (dd) , where My = Tm,+6), Yr (A) = TanlA} (A CAS is continuous in the weak topologies of the spaces
(6.4.5)
= 1.2.45 Ae If the operatorT'
X and Y thenrepresentation (6.4.5)
similarly as in (6.4.4) represents weakly m.s. convergent series of the weak stochastic
integrals.
In other words, for any linear functional
gY = gIX
is expressed by m.s.
g € G the random
variable
convergent series of the stochastic integrals:
gY¥ =gmy+)_
gr (A)Z, (dd).
(6.4.6)
After obtaining canonical representation (6.4.5) of the random variable Y we may find its covariance operator by formulae of Subsection 4.8.2. [2OGHZO),
(11)
r=1
where a(t, A) is the I*® derivative of the function z(t, A) over t. Example
6.4.3. In Example 6.3.3 after expressing the random function
Pe (t) by representation (I) we get for the random function Y(s) an analogous integral canonical representation
¥(s) =my(s) +90 fun(s,2)Z-(aX), where
(1)
q
yrs fA) [iedeelt rat. After obtaining representation
(II)
(I) of the random function Y(s) we may
find its
covariance function K, visi , $2) by means of the known formulae. Example
6.4.4.
In the special case when
X (t) is a stationary scalar
random function of the finite-dimensional argument ¢ it may be expressed by spectral
presentation (Subsection 4.8.6):
OTe
i:eZiT t(dd).
(I)
Then we get for Y(s) the integral canonical representation
¥(s) =m,(s) + fu(s,2)2(@d),
(11)
where it is denoted
My (s) = [iedmedt, y(s,A) = [iene
ae,
(IIT)
6.4. Methods based on Integral Canonical Representations
617
In particular, when t = § is time and Y(@) represents an output of asympto-
tically stable stationary linear system which performs in the stationary regime then y(t, X) = (id)e?,
where (id) is a frequency characteristic of a system. In this
case the formula for the covariance function of the stationary random function Yate)
gives an integral representation co
ky(r) = i eT B(id)@(iA)*o(dd) , —
(IV)
OS
where ¢ is the spectral measure of the random function X (t). Finally when there exists the spectral density of the random function X (t), and Y(t) is the scalar random function then the obtained formula gives the known expression for the spectral
density of the random function Y(t): Sy (\) = Example
6.4.5.
|6(id)|? Bela):
In the other special case when
X(t) is a stationary
m-dimensional vector random function it may be expressed in terms of the integral canonical representation (Subsection 4.8.7):
— 1 X()=met+ >> /ap(A)e*”*Z, (dd).
(1)
p=1
Then we get for the random function ¥ (3) the integral canonical representation
m
¥ (8)= my(s)+> fups,2)Zp(@d),
(it)
pal
where My (s) is determined by the formula as in the previous example and
Up (8, A) = /I(s,t)ap(A)e* ‘dt.
(IIT)
Ift = S is time and Y(s) represents the output of the asymptotically stationary linear system which performs in stationary regime then Yp (t, A) ces (1A) ap Gene where (7d)
is the frequency characteristic of the system.
In this case for the co-
variance function of the stationary random function Y(t) the following integral representation takes place:
ky(7) = > fe” 8G A)ap()ap(A)" HA)" 9(4A), (AV) p=!
618
Chapter 6 Methods of Linear Stochastic Systems Theory
where 0 p is a covariance function of the stochastic measure
Zp: According to the
formula for the matrix 7 (B) of the spectral and cross-spectral measures of the com-
ponents of the vector random function X(t) (Subsection 4.8.3) we have m
Ye | ae()ap(2)" (AA) = (8),
(v)
p=1B
and consequently, we may present formula for holt a in the form
eee /eOTS(iNo(ANO(A.
(VI)
In the case when there exists the matrix of the spectral and the cross-spectral densities
of the components of the vector random function X (t) the obtained formula gives the following expression for the matrix of the spectral and the cross-spectral densities of the components of the vector random function Y (é):
sy(A) = ®(id)se(A)B(iA)* Example
(VII)
6.4.6. Interpreting the random function X(t) of Example 6.3.5
as representation (I) of Example 6.4.2, we obtain for the function Ys) representation
(1) of Example 6.4.3 where
un(s,) = fUs, )zr(t, Adu(dt). The
correspondent
integral
canonical
representation
of the
covariance
function
Ky (s1 ; $2) of the random function Y(s) is found by the formulae of Subsection 4.8.2.
6.4.3. Solution Method Analogously
of Equations
we may
solve Eq.
by Integral Canonical
(6.3.13) by means
representation of the random variable X.
Representations
of the integral canonical
As we have already seen in Section 4.8
any sequence of the expansion of unit {E,} and the orthogonal vectors 477, ¥ such that the set of the vectors of all ENB) 7; is dense in Hy give the following integral
canonical representation of the random variable X ,
2 He, /r,(X)Ze(dd).
(6.4.7)
6.4. Methods based on Integral Canonical Representations
Here Zea
619
Ey Vy X° are uncorrelated stochastic measures, and the functions tp (Xi)
satisfy on the basis of Subsection 4.8.2 the following conditions:
J %Q)or(@) Stn
ee
re
(6.4.8)
B
Eel Bye A) = or, lBr) for any set B (B CA,BeE measure
(6.4.9)
S) where 0, is a covariance function of the stochastic
Z,.
> If L is a solution of Eq. (6.3.1 3) which possesses the required properties then
by virtue of (6.4.8) for any
E
g €
Gand
BES
gLX°- > /gLa,(A)Z, (dd)
Z,(B)
= gL EX" Z,(B) = [ ghay(d)ou(d2) B
= gLEX° [E,(Byy.X°|— 9h KeE,(B)y.=0,
(v=1,..., 7).
L
Consequently, according to (6.4.8) we have the equality
2 EB
gLX°-S~ /gL, (A)Z,(dd)
= E |gLXx°|’ — > [oter(aye |@LX°Z,(dd)| rool
= B |gLX°|’ - > fot-0)8 |SEK E- (a) | r=)
= E |gLX°|” - yi \gLx,(A)|° o-(dA). r=1
Hence it follows that the condition is fulfilled
D /[lgLe,(X)|? o-(dX) < 00.
(6.4.10)
620
Chapter 6 Methods of Linear Stochastic Systems Theory
cb oe it follows from the canonical representation of the covariance operator Kz f =
shLal A)f2,(A)o,(dd) iB that at any f € H, by virtue of condition (6.4.10)
‘hie:series
EK f= >. (Lx, (A) fa, (A)or(dd) converges in the weak topology of the space Y . It follows from the obtained expression LKgg
that Eq. (6.3.13) may have the required solution only in that case when the
operator ® is representable by
o= Y fey [-2-0)| 2408,9
peamnd
Cxa
and if we account (6.4.10) then for any g € G the following condition takes place:
(6.4.12)
Sef lgyr(A)|° o(dA) < 00.
For determining the functions y, (A) we substitute in (6.4.11) the index of the summation Y by p/ and apply the operator ® to the vector €,(B)y--
consideration (6.4.9) we get BE,(B)y, :
Then taking into
= f yr (A)o, (dA). Hence it follows that B
for any g € G the function gy, (A) represents the Radon—Nikodym derivative of the scalar measure gE,(B Vr over the measure 0,. This result may be written in the form
dO€,7;
He yeas
Os
(6.4.13)
If the operator [ should be applied to the random variable X but not only to the centered random variable X° then one more condition is necessary, namely My
KE Ak, which on the basis of Subsection 4.8.2 with account that u, = fz, (A)
at any Si € Hy gives
itn = Keo Ke Yofft )Ep(dr)¥p= Y |Fae, E, (dX) Yr00 Hence taking into the consideration
(6.4.8) we get Mz
=
Sa J tr (A) fotr(A) r=1
X Or (dd). Therefore the expectation must be presentable by the expansion
me =)~ /Eiicaonexdy,
(6.4.14)
6.4. Methods based on Integral Canonical Representations
621
and even
Se / I O)P ar(aa) = ei
Sf[fore(2)[? o¢(dd) = foKefo We shall find the solution of Eq. (6.3.13) in the form
_— Sf nAveoryr
(6.4.16)
ra
Then we have for any f € H,
EKef =>r=1 |mE Ref r==>fnO)FRELDV, or accounting (6.4.8)
LES f= Se/tr (A) far(A)or (d). r=1
In order the operator L will satisfy Eq. (6.3.13) it is necessary at any f € Hz the equality Of
=
‘3 f nr (A) far (A)o, (dA) holds.
After comparing
this equality
ral
with (6.4.11) we shall have at any fee HZ
Df maFeeOou(ar) => fwe) FeeP)op(ad). a
Putting f =
[Trapt
bal Baa
by virtue of (6.4.9) we get J nr(A)or (dd) = B
x o,(dd). As this equality is valid for any set
f yr (A) B
B € S then Nr (A) = Yr (A) almost
622
Chapter 6 Methods of Linear Stochastic Systems Theory
at all \ relatively to the measure 0,. Substituting the latter integral expression into
(6.4.16) and with account of (6.4.13) we find that
ic: ob a TOE eae >"fZe 0) £.(d\ye.4 (6.4.17) Similarly as under conditions of Subsection ment is proved.
6.3.1 the following state-
Theorem 6.4.3. In the case of the fulfilment of conditions (6.4.11)
and (6.4.12) the random variable Y° = LX° has the covariance operator Ky determined on the whole space G and the operator L is permutable with the operator of the expectation. And if besides (6.4.11) and (6.4.12)
conditions (6.4.14) and (6.4.15) are also fulfilled then there exists the random variable Y = LX which has the expectation my = Lm, and the operator of the second order moment determined on the whole space G.
Let be exist the measures pp with the values in H,, the nonnegative
measures o, and the functions z,(A) with the values in X satisfying the following conditions (Subsection 4.8.4) in some measurable space (A, S):
Keftp(B) = [eoQdon(aa), Bes,
(6.4.18)
B pol Beg) = booq lala) Then
(6.4.19)
N
>, J f%p(A)up(ddrA) = f at any f © Hz and we may take &)(B) p=
a
=
¥ S400)
(A) ]Hp(4A), Yr =
er(A)
(r =
1,
., NV). In this case amas (6.4.= takes the form N
N
b= Sf w(r)no(d) =o fFEEup(ad). Example
(6.4.20)
6.4.7. For solving the equation of Example 6.3.9 by the integral
canonical representations method it is required to find some number WN of the vector
measures [lp (b, B ) (of the same dimension that the vector random function X (t)), N nonnegative measures
op(B) and N sets of the vector functions oe
A) (of
the same dimension that X (t)) determined on some measurable space (A, S ) and
6.4. Methods based on Integral Canonical Representations
satisfying the conditions of Subsection 4.8.5.
623
Then formula (6.4.20) will give the
following expression of the linear operator Ly which satisfies Eq. (1) of Example 6.3.9: N
ie
Y /tsdawae,
(1)
where
I(s,t) ->f Here in order the operator
tation and the random
[eena akcat dr >jtp(t, dr)? .
L; will be permutable
function
Ys)
=
(11)
with the operator of an expec-
EXC)
will have the finite covari-
ance function at all s € S' it is necessary and sufficient that the function p(s, Ti) N
may be represented as p(s, T) = > ‘jUels, Aloe tis r)*op(dd), where Yp (s, A) =Sh MOSET) T) Sus(X) BO ze dr (eE =i
; N) and at all s € S thecondition of the
form (6.4.12) will be fulfilled: > f l¥p(s, A)? op(dA) < co. p=l1
In the special case of the finite-dimensional space A and absolutely continuous
measures /lp , B) and op(B) formula (II) takes the form
I(s,t) = i ray) {felssr)apte Mar} ap(t,ayaa, where
dpi, A) and
Vp(A) are the Radon—Nikodym
derivatives
(IIT)
of the measures
Lp (t, B) and op(B) over the Lebesgue measure respectively. Example when
the random
6.4.8. Let us consider a special case of the previous example function X (t) is real and its forming filter is known
(Subsec-
tion 4.8.4). Let w(t, T) be a weighting function of a forming filter, w™(t, T)be a weighting function of an inverse system. Then we may assume as JV the number of the components of the random function X (t) and as the matrices-columns
Lp (t, ) and Gp (t, A) (p ai
egies
N) the correspondent columns of the matrices
x(t, A) = w(t, A) and a(t, A) = w-(A, AE
Considering equal to unit the inten-
sities of the components of a white noise which is transformed by a forming filter into
the random function X(t) we get from (III) the following formula:
Ks,t) =f {[osu Q.ar} webdr.
av)
From this result as the special cases the Wiener formula known from the theory of
optimal linear systems and its generalizations on the finite-dimensional linear systems are derived.
624
Chapter 6 Methods of Linear Stochastic Systems Theory
6.5. Applications.
Shaping Filters Design and Modeling
6.5.1. Shaping Filters for Scalar Stationary Processes Let us show how we may design the shaping filter for a scalar stationary random function X (t) with rational spectral density. > Suppose that the spectral density of a real scalar stationary random function hg (t) with zero expectation is determined by the formula
S2(w) = P(w)/QW), where P(w) and Q(w) are the polynomials. and Q(w) contain only even degrees of W.
-
(6.5.1)
As Sz (w) is an even function, P(w)
In practical problems the degree 2™ of
the numerator P (w) is always smaller then the degree 27 of the denominator Q (w), as only in this case the variance of the random function X (t) equal to the integral of the spectral density is finite. Besides that it is necessary for the convergence of the integral of the spectral density that the denominator Q(w) of the spectral density be
different from zero at any real value of W. Finally, since a spectral density is essentially positive at all (real) values of the frequency W,all the coefficients of the polynomials 12 (w) and Q(w) are real. It follows from these properties of the polynomials P (w) and Q(w) that the roots of each of them are pair-wise conjugate complex numbers and to each root corresponds the root of the same polynomial of opposite sign.
In
other words, the roots of the polynomials P(w) and Q(w) are disposed on the plane of the complex variable
W symmetrically relative to the real and imaginary axes.
Let us factorize the polynomials P (w) and Q(w) and select the factors corresponding to the roots disposed in the upper half-plane of the complex variables Ww.
ee (w) has real roots then all these roots are multiple roots of even multiplicity as each of such roots is the point where the plot of the spectral density is tangent to
the real axis being disposed wholly above it. The half of the factors corresponding to each of such roots should be considered as belonging to the upper half-plane and the
other half as belonging to the lower half-plane. Taking all the factors in the factorized polynomial P (w), corresponding to the upper half-plane with the arithmetic square root of the coefficient Pam
of w?™
in this polynomial and 7™ as additional factors
we shall obtain the polynomial of degree
™ in W whose all the roots are disposed
in the upper half-plane and on the real axis. We denote this polynomial by H (iw). It is easy to see that all the coefficients of the polynomial
H (iw) considered as a
polynomial in iw are positive. Now let us take some root of the polynomial H (iw).
Let this root be a +28,
8 > 0. Ifa
2 0, then according to the property of
the polynomials P (w) and Q(w) proved above the polynomial H (iw) has also the root —@ + 1. Consequently, the polynomial H (iw) represents the product of the
6.5. Applications.
Shaping Filters Design and Modeling
positive number ,/P2mby the factors of the form i(w sa
625
co 1B)i(w +a—
i) and
i(w — 12), where 3 > 0. But
i(w — a —if)i(w + a — iB) = (iw)? + 26(iw) + a? + 6? ,
i(w — ip) = iw + 8, Consequently, the product of any number of such factors represents the polynomial in 1wW with positive coefficients.
The remaining ™ factors in the factorized polynomial P (w) corresponding to the roots disposed in the lower half-plane multiplied by VP2m and by (-i)™ form the polynomial H (—iw).
which is obtained from H (iw) by the change of sign at 1W, i.e.
Therefore, each root & +78 of the polynomial P (w) in the upper half-
plane corresponds to the root @ — 7 in the lower half-plane. Therefore, each factor
of the form i(w — a — iB)i(w +a — iB) = (iw)? + 2B(iw) + a? + B? or 1(w _ i) = iw + B of the polynomial H (tw) corresponds to the remaining factor
of the form —i(w — a +%)(—i)(w+a+if) = (—iw)? +26(-iw) +a? + 6? or respectively —i(w + i) = Thus
-—iw+ 16
selecting in the factorized polynomial
P (w) the factors corresponding
to the roots disposed in the upper half-plane and adding the factor 1”
./pom we
represent the polynomial P (w) at real values of W in the form
P(w) = H(iw)H(—iw) = |H(iw)|’ , where
H (tw) is a polynomial in 2W with positive coefficients.
(6.5.2) In exactly the same
way selecting in the factorized polynomial Q(w) the factors corresponding to the roots disposed in the upper half-plane, adding to them the factor 1” Va2n, dan being
the coefficient of w2” in the polynomial Q(w), and denoting the resulting polynomial
by F (tw) we represent Q (w) at real values of W in the form
Q(w) = F(iw) F(—-iw) = |F(iw)|? , where F’ (iw) is the polynomial in 2W with positive coefficients.
(6.5.3) Representing the
numerator and the denominator of the spectral density Sz (w) by formulae (6.5.2),
(6.5.3) we obtain the following expression for the spectral density Sz (w):
sz (w) = |H(iw)/F(iw)|” .
(6.5.4)
Such a representation of the spectral density is called factorization. Notice now that the function
®(s) = H(s)/F(s)
(6.5.5)
626
Chapter 6 Methods of Linear Stochastic Systems Theory
represents the transfer function of some stationary linear system.
Since the multipli-
cation of a complex number by the imaginary unit 7 represents the counter-clock wise rotation of the vector representing this number by the angle 7 /2 the left half-plane of the complex variables variable W.
$ = iW corresponds to the upper half-plane of the complex
Consequently, all the roots of the polynomials H (s) and F’ (s) are dis-
posed in the left half-plane of the variables S, and the function O( Ss) determined by formula (6.5.5) represents the transfer function of a stable stationary linear system.
Recalling that the stationary random function with constant spectral density Sq represents a white noise of intensity 277S9 we come to the conclusion that the stationary random function X (t) with rational spectral density Sz, (w) may be considered as the response of the stable stationary linear system with the transfer function ®(s) determined by formula (6.5.5) to the input representing a white noise with unit spectral density.
Consequently, the system with the transfer function ®(s) Ss H(s)/F(s)
represents the shaping filter for the random function X (t). On the basis of the results of Subsections 1.1.6 and 1.2.5 the stationary linear system with the transfer function ®(s) =
differential equation with the operator
H(s)/F(s)
F(D),
a
is described by the linear
d/dt, in the left-hand side
(acting on the output) and the operator H(D) in the right-hand side (acting on the input). Taking into consideration that the input of the shaping filter is the white noise
V(t) with unit spectral density whose intensity is equal to 277, and the output of the shaping filter is the random function X (t), we obtain the differential equation of the shaping filter:
F(D)X
=
H(D)V
the polynomial F(s) by az (k cd) polynomial H(s) by by (k =i
gi a Lie}
or denoting the coefficient
m),
Sa x®) = k=0
of s* in
Sa n), and the coefficient of s* in the
mV, k=0
Thus the shaping filter for the random function X (t) represents a system described by a linear stochastic differential equation of the
th
Nn” order. Replacing this equation
by the corresponding set of differential equations of the first order (Subsection 1.2.3) we obtain the equations of the shaping filter in the form Mp
Seer
(k=
Aik. = Ai pst
OV
1
a1);
( OT
aed)
aL)
(6.5.6)
n
Xin = an" So ap-1 Xie + gaV, k=1
and A(t) = O(Xi(b),t) = ire Ae
ee
eee
here
6.5. Applications.
Shaping Filters Design and Modeling
627
~
Qn—m)---+,Qn
are constant coefficients determined by formulae: -1 In—m
=
an,
bmn )
k-1
qbes a,
bn—k —
yi
An—k+hQh
Stig
nt
1
cn.
h=n-—m
i)
(6.5.7)
Ps
Remark.
In some
cases it is desirable to form a random function X (t)
of a white noise of unit intensity.
equal to 1/2r. dom from
The
spectral density of such a white noise is
Consequently, representing the spectral density Sz (w) of the ran-
function X (t) in the form {6.5.4) it is necessary |H(iw)/F(iw)|?
polynomial
the factor 1/2r.
H (iw) the additional
We may
in such cases
attain it by including into the
factor Jor or by including
mial F (iw) the additional factor 1/
2m.
to separate
into the polyno-
As a result instead of (6.5.4) we get
Se(w) = |H(iw)/F(iw)|? (1/27). While applying formula (6.5.4) or more general formulae (6.5.8) two conditions must be satisfied:
the condition of stability of a system and the condition of sta-
tionarity of its output.
The first condition is satisfied by virtue of our method of
constructing the shaping filter. This filter is stable since all the poles of its transfer
function (6.5.5) are disposed in the left half-plane of the variables s. condition may be satisfied by the choice of initial conditions for Eqs.
The second (6.5.6).
This
is possible because the shaping filter is a mere theoretical tool for reducing system differential equations to stochastic equations due to which the initial conditions for the equations of the shaping filter may be taken arbitrarily. We shall now show how we may do it
Sy(w = B(iw)s_(w)P(iw)* .
(6.5.8)
> As it is known, a stationary random process has a constant covariance matrix (the variance in the case of a scalar process).
In other words, the covariance ma-
trix of the value of the stationary random process X
(t) at any time moment f is
independent of ¢ and is equal to the value of the covariance function ke, (T)of this process at T =
0, ke40); co
kei (0) = / Sz, (w)dw .
(6.5.9)
—oo
Therefore, for providing the stationarity of the process Xj (t) at the output of the shaping filter one should take the random initial value Xi (to) = X09 at the instant
628
Chapter 6 Methods of Linear Stochastic Systems Theory
to with the covariance matrix | (0) determined by formula (6.5.9). It remains to find the spectral density Sz, (w) of the process X4 (t) determined by Eq. (6.5.6).
Using formula (6.5.8) we find
Sz,(W) = By (iw), (—iw)? so ,
(6.5.10)
where Sq is the spectral density of the white noise V equal to 1 if sz (w) is repre-
sented by formula (6.5.4) or 1/2x ALES 7s (w) is represented by formula (6.5.8).
For
determining the transfer function ®; (s) of that part of the shaping filter which gives the vector random process X 4 (t) =
[Xu (t) seen Ores (t) ie at the output we notice
that the transfer function of the whole shaping filter whose output represents the pro-
cess X(t) = X11 (t) was found earlier. It is determined by formula (6.5.5), 13(s) =
®(s) =o (s)y)F (s). Therefore, the remaining elements of the matrix-column
®,(s) may be found from the first n — 1 by replacing V by the function ere Xik by the function 14(s)e** (k =e
ee n), and the operator d/dt by the variable
S (Subsection 1.2.5). As a result we obtain the equations
s®14(s) = Py x41(s)
(K=1,...,n—m-—1),
s®1;,(s) = Oy 44i(s) tan (K=n-—m,...,n—1). Solving these = ®(s) =
equations
in succession
and putting in the first of them
®11(s)
H(s)/F(s) we find the elements of the matrix-column ®, (s):
Dip(s)Se"r *O(sy Sa"
A (s)/P(s)"
(Se
nye
k-1—n+m
®1,(s) = s*~" G(s) —
Dae
qz—1-15!
I=0 k—1—n+m
ls 1H (s)/F(s) —
dX i peovates
Reducing the expressions of 1% (s) (k Snr
(EF =a — In ly oe
lee..26 n) to a common
denominator and using formula (6.5.7) we obtain after simple but rather cumbersome evaluations
#11(s) = H(s)/F(s),
G14(s) = Az(s)/F(s) (k=2,..., 7). (6.5.11)
Here
HIG)
oe
GUS
ee aay,
(6.5.12)
Hi(s) =) cers” (kan—m+1,...,),
(6.5.13)
n—1
r=0
6.5. Applications.
Shaping Filters Design and Modeling
629
where k-1
C=
yi
Q—k+4r411
r=)
Fm
=n
k— 1),
Qj—k+r411
(r=m—n+k,...,k—-—2)
l=k-1-r
k-1 Cha
SS l=n-—m
(only ifn —m
> 2), k-—l+n-r
ep
Ze
aid
AP oe
1)
(6.5.14)
i=-
For calculating the elements of the matrix kz, (0) determined by formula (6.5.9) we may use formula (6.1.27). < Example
6.5.1. Design the shaping filter for a stationary random func-
tion X(t) with the spectral density
S3(@) =
1)
To
a
tw
(I)
In this case the numerator P (w) represents a constant and the denominator has
two pure imaginary roots +i@. F(iw) = 1w+ad.
Consequently, we may assume H (iw) = 4) Da/n,
Then we obtain
sole) = [®(iu))?, (8) =~. Thus
(IN)
the shaping filter represents in this case the first order filter with the time
constant J’ =
1/a and the gain k =
4/ D/ra.
The differential equation of this
shaping filter has the form
X+aX =/Da/rV, V being the white noise of the intensity
(IIT)
vy= 27.
Putting H (iw) sin/ 2Day F(iw) = iw + a, we have
so(u) = 5 |®()I?, (8) =
V2Da
sta
(Iv)
630
Chapter 6 Methods of Linear Stochastic Systems Theory
The differential equation of the shaping filter will in this case take the form
X +aX = V2Dev ,
(V)
V being a white noise of unit intensity. Taking in both cases the random initial value Ki = X (to) with the variance co
Ds. =
Da
/ Sz (w)dw =
[o-e)
dw
/ aa?
= ID
(VI)
we shall obtain at the output of the shaping filter the stationary random process
X(t) with given spectral density Sz, (w). Example
6.5.2. Design the shaping filter for a stationary random func-
tion with spectral density
mj)
so(#)
(a+SR ywo)d*+(aCe) Dnm ee EEN wow? ee b4+42(a2 —w?)w? + w4
In this case the numerator
9 et _ 42 =
vf
P(w) has two roots WwW
=
9
I
.
+764,
(1) b? =
b?(a
ate ywo)(a = ywo)71, and the denominator Q(w) has four roots +(wo + ia). Selecting the roots disposed in the upper half-plane according to the outlined method
H (iw) = V/2D(a — yuo) i(w — iby) = /2D(a — yuo) (iw + 1), F (iw) = i(w — wo — ia) i(w + Wo — ia) = (iw)? + 2a(iw) + b?. Then we shall have
vl
ee
so(w) = 5— |®(t)|"
_ A(s) _
, O(s) = 110
V2D(a— ywo)(s + 61)
ee
(II)
The differential equation of this shaping filter has the form
X + 2aX +0?X = /2D(a— ywo)(V + 61V),
(IIT)
V being a white noise of unit intensity. To this equation corresponds the set of two
first-order
Eqs.
(6.5.6)
for the
vector
random
function
Xj (t)
=
[X(t)
KG) —nVG)| Xi = MatonV
Xia =
0
Re aX
EV,
(IV)
6.5. Applications.
Shaping Filters Design and Modeling
where according to formulae (6.5.7) qj = Xx (by =
2D(a — ywo), —
631 /2D(a
= ywo)
2a).
For determining the covariance matrix of the initial value X}9 =
X,1 (to) we
find by formula (6.5.11) the elements of the matrix-column ,(s):
928 — qb?
qi(s+ 61)
Pee ee aE deeb cE heen opera a Ou After this we find by formulae (6.5.10), (6.5.4), (6.5.8)
years
wee
Yo
ee
ke,,(0) = 24 / ee de Sis) an J \(w)? + 2a(iw) + 6?|
(VI)
—oo
Biel @) ete fight LeMert 2 a) dw 2m
)? W) |(iw)? + 2a(iw) + 62
bf
ec2t ie
= —D(a— (a
wo)
ywo), ((VII )
\2 4.bac?
ke(0) =2agoffate |(tw)? + 2aiw + b?| —oo
=D G — by/a® — y2w? + 2a(a — 10) : Taking the random initial value X;9 =
(VII)
Xj, (to) with the covariance matrix whose
elements are determined by the derived formulae we shall get at the first output of the shaping filter the stationary random function X (t) with given spectral density
Sz(w). 6.5.2. Shaping Filter for Vector Stationary Process Analogously the problem of finding the shaping filter for a stationary vector random function X(t) with a rational spectral density Sz (w) is solved. In this case
all the elements of the matrix Sz (w) represent the ratios of the polynomials in W. In accordance with (6.1.22) for finding the shaping filter for X(t) it is sufficient to represent the spectral density Sz (w) by the formula
S_(w) = F(iw)~1H (iw) H (iw)* F(iw)7**,
(6.5.15)
where F' (s) and H (s) are the matrices whose elements are all polynomials in 5, and all the roots of the determinants
of these matrices
posed in the left half-plane of the complex variable s.
F(s) and H(s) are disThe representation of the
632
Chapter 6 Methods of Linear Stochastic Systems Theory
matrix of spectral and cross-spectral densities Sz (w) of the components tionary vector random function in such a form is called a factorization trix Sy (w). After factorizing the spectral density Sz (w) we may
of a staof the ma-
write the differ-
ential equation of the shaping filter, transform it into a set of the first order differential Eqs.
(6.5.6) and find the initial conditions for them providing the statio-
narity of the process at the output of the filter in the same way as in Subsection 6.5.1. Example
6.5.3. The spectral density of two-dimensional stationary vec-
tor random function X (t) is determined by the formula
ki = aj tw
S,(w) =
ky ky me QQ — (a1 — a)iw +w 4 k
kik2 102+ (a) — a2)iw +w 1
as + uy
.
a) I
Find the shaping filter for X(t): In this case
ale) Wei x |
Od
(2 ry |[ a
0
(11)
(a2 — iw)7} |
aril i
ao+s
athe lire
(111
and the set of differential equations of the shaping filter has the form X,+0,X;
=kV,
X_o+a2X2
= keV,
(IV)
V being a white noise of the intensity 27. Taking the random initial value of the vector Xg = X (to) with the covariance matrix [o-e)
Poy o| )
hig / so(w)
me
k?/ax [Oy WK, longi
oaaet
2akyko/(ay 1%2 1k
1
+
/og
2 a2) :
(V)
—0o
we
shall obtain at the output of the filter the two-dimensional
stationary vector
random process with given spectral density S_ (w).
6.5.3. Shaping Filter for Process Reductble to Stationary One The method
outlined in Subsections
6.5.1 and 6.5.2 may
sign filters for random processes reducible to stationary ones.
also be used to de-
The case of the pro-
cess reducible to stationary by the transformation of the process itself is trivial.
If
6.5. Applications.
Shaping Filters Design and Modeling
633
X (t) = W(X, (t), t) where Xj(t) is a stationary random function, then substituting this expression into differential Eq. (5.4.2) we obtain an equation with a stationary random function in the right-hand side. Therefore it remains to consider the case of a process X (t) reducible to stationary by the transformation of the argument.
D Leek
(
ty v(X1(~(t)),t) where Xj(s) is a stationary random func-
tion with zero expectation and a rational spectral density, and
monotonous increasing differentiable function.
s =
y(t) being a
After constructing the shaping filter
for the random function X 1(s) we write the differential equation of this filter and
the corresponding initial condition in the form dX; = aX,ds a bdW, , where
a and 0 are constant
matrices
X1((to))
=X
of the corresponding
(6.5.16)
dimensions,
and
Wj
=W, (s) is a process with uncorrelated increments with the covariance function
Kw, (s1, $2) = ki(min(s1, s2) , ki(s) = ki(so)
+11(s— 80),
(6.5.17)
VY; being the constant intensity of the stationary white noise Vj (s) = Wi (s), and
i) = p(to). Let us consider the random process W(t) = Wi(¢(t)). It is evident that W(t) is a process with uncorrelated increments as the monotonous increasing function
§ = p(t) maps any nonintersecting intervals on the S-axis on nonintersect-
ing intervals of the t-axis. The covariance function of the process W(t) is determined by the evident formula
Kw (ti, t2) = Kw, (p(t1), p(t2)) = kmin(y(ti), p(t2)),
(6.5.18)
where t
k(t) = ki(y(t)) = ki(y(to)) + rile(t) — p(to)] = k(to) + [veer to
Hence, it is clear that the intensity of the white noise V(t) = v(t) = 1 9(t). The change of variables
W(t) is equal to
s = p(t) in Eq. (6.5.16) yields
1X,(p(t)) = aplt)Xy(olt))at + bdW, Xi(vlto)) = Xo. Introducing the random process X92 (t) = XG. (y(t)), we represent Eq.
(6.5.19) (6.5.19) in
the form dX»
=
ap(t)Xodt ae bdW ,
Xo(to) =
Xo .
(6.5.20)
634
Chapter 6 Methods of Linear Stochastic Systems Theory
Thus the shaping filter in this case is described by Eq.
(6.5.20) and by the formula
X(t) = 4(Xo(t),2). 4
Note the fact that after writing the equation of the shaping filter for the random function X 1 (s) in the form of an ordinary differential equation with a white noise in
the right-hand side and performing the change of variables
s = p(t) in this equation
we shall get in the coefficient in front of the white noise in Eq. (6.5.20) the additional factor p(t). But such a derivation of Eq.
(6.5.20) is false. This illustrates the fact
that ordinary rules of the change of independent variable are not applicable when
there is a white noise in a differential equation.
It is necessary to remember
this
when we deal with stochastic differential equations. The
covariance
function
of the
initial
value
of the
process
X1 (s),
Xo
Xj (y(to)) is determined by Eqs. (6.5.9)—(6.5.11). Example
=
Co
6.5.4. The random function X(t) with the expectation Mz (t)
+ cit and the covariance function:
Kz (ti, ta) =
reduced to a stationary one by the transformations:
Debts +t2)—@ |#1-23|
is
X (t) =co+tectt+ e#* Xy (s),
s = t?. Here X 1 (s) is a stationary random function with zero expectation and the covariance function ka. (c) =
De-«lel , 0 =
81 — Sg.
This covariance function
corresponds to the spectral density Sz, (w) ss Da/n(a? te Papas Using the method of Subsection 6.5.1 we find the differential equation of the random function Xj (s):
dX,/ds
=
=a
+vV 2DaV,
)
(1)
V; being a white noise of unit intensity. Passing to the variable t we replace Eq. (I) by Eq. (6.5.20) for the random function X92 (t) = AG (t?) which has in this case the form
Xo = —2atX,+ V2DaV,
(II)
where V is a white noise of the intensity 2¢. The initial condition for this equation in accordance with Eq. (6.5.20) has the form: X9 (to) = Xo, where XQ is a random variable with zero expectation and the variance
ir
Dey = f 52,(w)dw =
Dat =
odee
(III)
6.5.4. Shaping Filters for Stationary Functions of Vector Argument Many problems of practice lead to the partial differential equations which contain stationary or reducible to stationary random functions of a vector argument.
6.5. Applications.
Shaping Filters Design and Modeling
635
For the application of the methods of the linear stochastic systems theory to such problems it is expedient to present the random functions entering into the equations
as the solutions of the linear stochastic differential equations. Here we give a general method
of finding the linear stochastic differential equations in the correspondent
H-spaces for a wide class of the stationary functions of a vector argument.
Let X (é, zr) bea
scalar stationary random function of the scalar variable f which
we shall assume as the time and V-dimensional vector variable 2; k(ty —to,%,- £2) be its covariance function which is equal to k(t, — to, 0) alec k(0, |
—
co
—
somal
r2) at t; = tg = t. We shall consider here normally distributed random
functions X (t, zr) satisfying two conditions: (i) at a given & the random function Xz (t) has a rational spectral density independent of ©
cE
s(w) = = / k(z, 0)e*”"dr = P(o?)’
(6.5.21)
where P(w?) and Q(w?) are the polynomials and Pw?) has no zeros on the real axis;
(ii) at a given f the covariance function of the random function Xi(z) at some C > 0 anda
>
0 satisfies the inequality
|k(0, 21 — t2)| < Ce~%!*1-721 For obtaining a stochastic ee
differential equation for the random
(6.5.22) function
X
(t) considered as a random function of the time ft at a given © we may apply
the same method
of shaping filters which was used for the scalar random process
(Subsection 6.5.1). According to this method it is necessary to present the spectral density Sz (w) in the form
Vo
$_(w) == =Qn
H (iw) |? F(iw)
(G2
:
where F(s) and H(s) are the polynoms whose roots at
(
)
$ = 1W be in the left half-
plane; lp is the constant which is usually obtained for the sake of convenience to write
the equation for X(t, ay Formula (6.5.23) allows to treat Xz (t) as the result of the passage of the white noise of the intensity /o through a stationary linear system with
the transfer function H(s)/F(s). differential equation
where
V
=
This fact will give for X, (t) a linear stochastic
F(0/dt)X = H(a/atyV,
(6.5.24)
V(t, r) ——Ve (t) is the stationary normally distributed white noise
which depends on time f at a given © whose intensity is equal to lp.
636
Chapter 6 Methods of Linear Stochastic Systems Theory
If we have n
F(s) = ) Saget Pres l= S > be s* Sanrcons
(6.5.25)
k=O
k=0
then Eq. (6.5.24) represents a stochastic linear differential equation of the n‘} order
which should be assumed as the Ité standard equation:
dZ = AZdt + qdW. Here Z is the vector with the components
trix with the elements =i
ie,
aj; =
65,641) Grae
Z; =
(6.5.26)
X, Zo MA wt 9 Zn, A is the ma-
erg
ee , N, 7 is the vector with the components Gj =
=
Lid
—ana;-1,
--* = Q(n—m-1
1
= 0, Qn—m
see k-1
Gd=4,
b,—-E—
Doge An—k+hQh
ee 2s
a
O
eee
h=n-—m
We=
Wt, r) is the scalar Wiener process with the values in the space of the
functions of the variable £ whose weak m.s. derivative is the white noise
V=
V, (t)
of the intensity V:
ween
t
/Tate
(6.5.27)
0
Eq. (6.5.2) is reduced to Eq. (6.5.26) by a formal change of the variables (6.5.6):
Geek Ze
OZ Ot
Dee, See
OZ =——aV,
he aan
in
de
kon—m,...,n-1,
(6.5.28)
where gx are determined by formulae (6.5.7) and quq = qq. v. We shall consider the random function X (gf:zr) as a random process with the values X+(-) in the H-space L2(D) of the functions of the variable x,
D C Roe
Then as a covariance operator of the value of this process at a given ft will be an
integral operator with the square-integrable kernel kz (0, r—€ ):
Kap = |ke(0,2—€)o(6)dé D
(6.5.29)
6.5. Applications.
Shaping Filters Design and Modeling
637
The analogous formula takes the place for the covariance operator KX , , of the vector process
Zt (x):
pines ikz,(2— Op(E)ae , D
where the kernel Rats
= r2) =
kz(0, |
r2) is a covariance function of the
process Z1(2x) at a given t. For finding the kernel v(x = £) of an operator of the intensity of the Wieter process W
we use an equation which connects the covariance
matrix k,, of the value of the process
Z = Z(t, x) at a given t with the intensity
V of the Wiener process W. This equation has the form
Ok, Ot As the process
s
: “i= Aky, ike
Abe (qh Vc
Z is stationary then the covariance matrix
(6.5.30) ee of its value at a
given ¢ does not depend on f and is expressed in terms of its covariance function
kz (ty —t2,21— r2) by the formula k,, = k,(0, Li- r2). Therefore Eq. (6.5.30) has in a given case the form
0 = Ak, (0,21 — 22) + kz (0,21 — 22)A? + qq? v(21— 22).
(6.5.31)
Matrix Eq. (6.5.31) represents a system of n(n + 1)/2 scalar linear algebraic equations with n(n + 1)/2 unknowns, namely n(n + 1)/2 — ments of the matrix k,(0, i of the matrix k,(0, i
ZAr(t, r) =
r2) and the unknown
£2) is known:
Kzaoe\U; Cie
1 of the unknown
ele-
y(avy = Zo). One element r2) =
kz(0, iy = r2) as
X(t, 2). After excluding from (6.5.29) the unknown elements of the
matrix k,(0, lees r2) we shall find the unknown function v(2x4 = to):
v(@; — £2) = ckz(0, x1 — 22), where C is some constant.
At 21 =
2
=
(6.5.32)
Z@ this formula gives the intensity Vp of
the process Wt, zr) at a given ©: Vo = ck, (0, 0).
From (5.9.3) it follows the formula for the kernels of the operators: t
kw(t, 21 — £2) = fre — x2)dr = ckz (0,21 — a2). 0 Here k, is the kernel of the operator
Ky (t) in (5.9.3) at a given f.
(6.5.33)
Condition
(6.5.22) provides the trace-type of the operators ky and V as in this case by virtue
638
Chapter 6 Methods of Linear Stochastic Systems Theory
of the square integralability of the kernel k. for any orthonormal basis {pr (x)} in
TAD):
adé (E)d) (x by, ize — €)pepr k=1pp co
=c)) x2(0, 2 ~ pn(a)pe (€)dadé Let X(t, x) =
o( Xi (p(t), x), t, x) where X(s, x) is a stationary random
function, p(t) is monotonously increasing differentiable function; (0) =
0. After
writing Eq. (6.5.26) for the random function X 1 (s, x),
dZ = aZds + bdWy, where Z\(s, zr) =X
(6.5.35)
(s, 2), and Wi(s, 2) is a Wiener process with the kernel of
the covariance operator V(x = L2)s and after changing the variables
5 = p(t) we
obtain a linear stochastic differentiable equation in the H-space Ly (D):
dZ = ApZdt + qdW.
(6.5.36)
Here W = Wit, xr) =W, (s, zr) is a Wiener process with the kernel of the covari-
ance operator of its value at a given t equal to k(t, 1—-
£2) — yt (a1 — r2)p(t)
6.5. Applications.
Shaping Filters Design and Modeling
639
where 1’; (x4 = £2) is the kernel of an operator of the intensity of the Wiener process Wi(s).
dq
6.5.5. Software for Analytical Modeling Problems At practice analytical modeling problems for the linear stochastic systems are
usually solved by means the dialog packages.
of the universal or the specialized computer libraries and
The corresponding algorithms are based on equations given in
Sections 6.1-6.3, 6.5.
”
Let us consider some
lized dialog packages
examples
of analytical modeling solved on the specia-
“StS-Analysis”
(Versions 1.0 and 2.0) based on equations of
Sections 6.1 and 6.5. Working with any version of the “StS-Analysis”
package, an user must
only
type on the display with the aid of the keyboard the original stochastic equations in the natural mathematical form without any special programmers tricks, introduce the corresponding initial conditions and indicate which characteristics must be calculated. Once the original stochastic equations are introduced the user may
correct them,
i.e. to remove some equations, to replace some of them by others, to correct errors
and so on.
When
computations.
all the data are introduced and checked the user has to start
Then the package “StS-Analysis” automatically derives the equations
for the corresponding distribution parameters and solves them. The user may observe visually the behaviour of a system under study in the process of calculations by tracing the graphs or the tables of current values of various characteristics on the display and
get the copies of these graphs and tables from the file after the calculations finished.
So the use of any of the versions of the package
“StS-Analysis”
are
does not
require that the user know mathematics and programming. Any version of the “StS-Analysis”
package automatically derives the equations
for the distribution parameters from the original stochastic equations introduced by the user in the natural mathematical
form and solves these equations.
It should
be remarked that the “StS-Analysis” package does not simulate a stochastic system and does not use the statistical simulation (Monte-Carlo) method but performs the theoretical calculation of the necessary characteristics of a system behaviour by the linear stochastic systems theory methods.
Example
6.5.5.
For the stochastic two-dimensional linear discrete sys-
tem
Y2(1 +1) = —bY} (I) + a¥o(1)
(11)
(a, b being known constants) the exact equations for the expectations, the variances and the covariances (Subsection 6.1.9) have the following form: my(I + 1) =am, (1) + bm2(1),
(IIT)
640
Chapter 6 Methods of Linear Stochastic Systems Theory
m2(l + 1) = —bm, (I) + am2(l),
(IV)
D,(1+ 1) = a? Dy(I) + 6? Do(l) + 2abKy2(1) + 1,
(V)
D,(I + 1) = b? D,(1) + a” D2(1) =— 2abK12(1),
(VI)
K12(1+ 1) = —abD, (I) + abDo(1) + (a? — 6?) Ky2(I). Graphs of m,(l), m2(l), D,()), D2(l) and Ky2(1) at initial conditions: =F
EY;,(0) = JONG (0) = DY2(0) = EY;(0)Y2(0)
a) a=1.0; b=0.25
(VI) EY2(0)
= 0 are given on Fig.6.5.1.
b) a=1.0; b=0.25
ma2(I)
Fig. 6.5.1
6.6. Applications.
6.6. Applications.
Optimal Online Linear Filtering
641
Optimal Online Linear Filtering
6.6.1. Equations of Linear Filtering The optimal filtering problem may be solved completely in the case of linear
Eqs. (5.10.4) and (5.10.5)
dY = (bY +.b,X + bo)dt + yidW,
(6.6.1)
dX = (aY +a,X + ao)dt + ydW,
(6.6.2)
where the coefficients @, @1, do, 6, b1, bo, w and 1 in the general case depend on
time t. In this case the distribution of the process LYat)2 X(t)2 |i is normal if the distribution of its initial value
ben XK
ae
is normal.
Consequently, the conditional
distribution of the state vector of the system relative to the process fe is also normal
and for its determining it is sufficient to find the conditional expectation X and the conditional covariance matrix
R.
For this purpose
we use formulae
(5.10.31) for the stochastic differentials of the random variables
(5.10.35) and
X and FR.
> Substituting into (5.10.31) and (5.10.35) the expressions p(y, #,t) =
ay
+ a,x + ao, yi(y, 2, t) = by + b1x + bo we obtain
dX = (aY +a1X + ao)dt + {E[X(X? — X7) |Yf]of + put }(vivbT)1[dY — (bY +b, X 4 bo)dé],
(6.6.3)
dR = E|(X — X)(YTa? +. XTal + a?) + (aY + a,X + ap)(X7 — XT) + dy" | Yi Jt
— B[(X — X)(¥PH + XT OT + 6g) + but |Yi] x (pvt )- 1E[(bY +b,
X + bo)(X7 — XT) 4 divyt | Yi]dt
+ OLD an bla — Fa) = XYXP = XP) [VE l=1
hs
+ B[X — X |¥i]67 + hiE[X? — X7 |vay) x
dY, —
(3binYR + Yo 1p Xp + mna hail
where @j,
(6.6.4)
p=1
are the elements of the matrix
@ =
bT (yy vyt)-}
and (3, is the pth
column of the matrix vyt (v1 vy? )-). Taking into account that
B[X(X? — RT) |¥f] = B[(X — X)X7 |YZ] eee spo) Mlk,
642
Chapter 6 Methods of Linear Stochastic Systems Theory
BX — Xi Yj) =—Xi— X=,
E[(X — X)Y? |Y,,] = E[X —X |¥,,]¥* =0 and that the conditional central third order moments are equal to zero in consequence
of the normality of the conditional distribution we transform formulae (6.6.3), (6.6.4) into the form
dX = (aY + a,X + ao)dt
+ (ROP + pv )(divd?)—1[dY —(bY + b1X 4 bo)dt], (6.6.5) dR =[a,R+ Ral + pvy7
— (ROT + vd (vd) (bR+ viv" )ldt.
to. This optimal estimate Xs unbiased at all
t > to, as by virtue of the formula of total expectation and formula (5.10.6)
B(X; — Xt) = B{E(X; — X1) |Y¥i]} = E(X: — Xi) = 0. Eqs.
(6.6.6)—(6.6.8) solve completely and exactly the problem of optimal linear fil-
tering of the state of a linear system defined by Eqs. (6.6.1), (6.6.2).
6.6.2. Kalman-Bucy Filter Eqs. the case of
(6.6.6) and (6.6.8) were fifst obtained by Kalman and Bucy in 1960s in a = b = 0. This case corresponds to the practical problem of filtering of
the signal X determined by (6.6.2) at a = 0 when the signal b; X + bo is measured with an additive disturbance representing a white noise.
In this case Eq.
(6.6.1),
(6.6.2) may be rewritten in the form
X =a,X +agt+yV, Z=uX+b4+urV.
(6.6.10)
V being a white noise, i.e. the derivative of a Wiener process Wit). The equation
of the optimal filter (6.6.8) has in this case the form
5 Ts PO RT Eq.
(6.6.11) determines
EN Se HE
the structure of the optimal filter.
(6.6.11) Namely,
the optimal
filter may be obtained from the given system (the system generating the signal X; Fig.6.6.1a by the installation of the amplifier with the gain G before its input and by applying negative feedback containing the amplifier with the gain b; to the system obtained (Fig.6.6.1b). Receiving the measured process Z at the input with the function of time bo, subtracted from it the filter obtained in this way will give the
optimal estimate
X of the vector
X at the output.
a)
Fig. 6.6.1
644
Chapter 6 Methods of Linear Stochastic Systems Theory
As the process V(t) representing the integral of the measured process is not
used
Xy =
EX
in the optimal
expectation
Z (t)
of the vector
Xo,
should be assumed as the initial value of the estimate x at b=
and the conditional
—
filter, the conditional
EXo)(X¢
covariance
ae EXé)
matrix
of the random
vector Xg,
Ro
=
Lo;
E(Xo
should be assumed as the initial value of the matrix
R
while integrating Eq. (6.6.6) and calculating the gain (3. But these conditional characteristics are usually unknown.
This compels one to take arbitrary initial of K and
R. Certainly, x will not be the optimal estimate of the vector X in this case and
may be only an asymptotically optimal estimate if only the first equation of (6.6.10) and Eq.
(6.6.6) describe a stable process (i.e. if the system described by the first equation of (6.6.10) and Eq. (6.6.6) is stable). It goes without saying that the gains ( and b; represent in the general case
the matrices.
So we consider here the amplifiers of vector signals performing linear
transformations of the vectors received at the input which are determined
by the
corresponding matrices, i.e. by matrix gains. Optimal linear filters constructed by the method
outlined are usually called
Kalman-Bucy filters. For the linear stochastic systems the problems of Kalman—Bucy filter design are usually solved by means
of the universal or the specialized computer libraries and
dialog packages. Example
6.6.1.
Find the optimal filter for filtering a sinusoid signal of
a given frequency Wg which is measured with an additive normally distributed white noise independent of the signal.
The sinusoid signal X ; (t) and its derivative X 9 (t) = xe (t) may be considered as the components of the state vector of a system described by the differential equation
Eu 5 dt |Xo}
ed ak es |
—w2
0
Bes Xo]
(I)
The measured process is determined by the formula
Y=Z=X,4+V=fl0X+V. Consequently, in this cased
= dg)
= b= bo = yp =0, db) = [1 0], Dy =i,
Og
1
Eq. (6.6.6) represents the set of equations
Ri; = 2Rip — vy R?,
Rio= we Ri + Reo — v1 Ri Rio,
(II)
6.6. Applications.
Optimal Online Linear Filtering
645
Rog = —2w2 Riz — v-1 3.
(II)
After determining 11, Rj2 and Roo by the integration of these equations with the initial conditions R41 (to) = we find m.s.e.
EX??, Ri2(to) 22 CRY
of the filtering of the signal X;, R11 =
Moa Ro2(to)
El(X,
= Aa)
—
EX
|Y;7] and
of its derivative X2, Roo = El(X2 = a)" |Yi} and the gain
pauai RH awnt |Fh ea | sospalesesl = els (IV) Eq. (6.6.8) for optimal estimates Hew x has in this case the form
]-[2: Renee. 0 77
The block diagram of the optimal filter found is shown in Fig.6.6.2.
Fig. 6.6.2
6.6.3. Case of Equations Linear in State Vector Let us now consider the more general case of Eqs.
(5.10.9) linear only in the
state vector of the system
dY = [bi(Y,t)X + bo(Y, t)]at + vi (¥, t)aW, dX =[a,(Y,t)X + ao(Y, t)]dt + ¥(Y, t)dW.
(6.6.12) (6.6.13)
In this case for any realization y(t) of the measurement process Y (t) Eq. (6.6.13) is linear in X. Therefore, we may make the conclusion that the conditional distribution
of the random vector X; is normal for any realization Vi = {y,;7 € [to, t]} of the measurements Mes at all t > to if the initial distribution of X is normal.
646
Chapter 6 Methods of Linear Stochastic Systems Theory
> Assuming
the conditional distribution of X; relative to Le as normal we
write for this case formulae (5.10.31) and (5.10.35) for the stochastic differentials of the optimal estimate X and the covariance matrix R of the state vector of the
system. Substituting into (5.10.31) and (5.10.35) the expressions gly, x,t) == ai(y,t)r+ao(y,t),
yily, 2,t) = bi(y,t)a+bo(y, t), (6.6.14)
taking into account the independence of y on X and recalling that all the central moments of the third order are equal to zero for a normal distribution, we shall obtain in the same way as in Subsection 6.6.1
dX = [a,(Y,t)X + ao(Y, t)]dt + [Rb (Y,t)? + (v7 )(Y, t)] x (WyvyT)—1(Y, t){dY — [bi (Y,t)X + bo(Y,t)]dt}, (6.6.15) dR = {a,(Y,t)R+ Ra,(Y,t)? + (pvyp7)(Y, t) — [Rbi (Y, t)?
+ (bud )(Y M@ivdt)(VY NY HR + (buy )(¥,t)]}at. 4 (6.6.16) These equations represent a closed set of equations determining B¢ and R as
well in the case of linear filtering.
Therefore, the optimal estimate X oh the state
vector of the system X and its conditional covariance matrix FR characterizing the accuracy of the optimal estimate x , may be calculated by the integration of the set
Egs. (6.6.15), (6.6.16) while obtaining the results of measurements. Notice that contrary to linear filtering it is impossible in this case to calculate R beforehand when the results of observations are not yet received, as the coefficients of
Eq. (6.6.16) depend on the results of measurements.
Therefore the optimal filter in
this case must perform the simultaneous integration of both Eqs. (6.6.15), (6.6.16). It leads to the essential increase of the order of the optimal filter.
> Now we shall prove that the conditional distribution of the state vector of the system is normal in this case.
For this purpose it is sufficient to show that Eq.
(5.10.21) for the conditional characteristic function ge(A) of the vector X; has the solution
;
gi(A) = exp {arg _ 7 Ra} where
(6.6.17)
and R are determined by Egs. (6.6.15), (6.6.16). Substituting into Eq. (5.10.31) expressions (6.6.14) of the functions Y and
we reduce it to the form
dge(d) = El{id? (aX + a0) — 5?" Jei*™* |v, at 1
5
6.6. Applications.
Optimal Online Linear Filtering
647
+E[{(XP—X?OP ar? pot pe * | VE] (dived?) [d¥ —(b1 X+50)at] (6.6.18) where
the arguments
of the functions
@1, ao, 61, bo, wy and yj are omitted for
brevity. Taking into account that
Elixe?’* |yt] =
0gt(A) OX
(6.6.19)
we get from Eq. (6.6.18)
dg:(A) = [xP “8 +, (aPa a A" wv" 1) |dt ge(A) bypr +(X*ws by = OT aN aly — id* ag 4 dy; fb )gt(A) = i
x(dvyt)-! [ay Ly bpXp bo)at|
(6.6.20)
We evaluate separately the left- and right-hand sides of this equation for the function
gr(A) determined by formula (6.6.17).
Using differentiation formula (5.1.15) of a
composite function in the case of a Wiener process
W, taking into consideration
that the stochastic differential of the process X(t) is determined by formula (6.6.15) and that (Rb? ar wyt
)\(bivet)-
dy plays the role of the matrix Y in (5.1.15)
in this case, we find the stochastic differential of the process gi(A) determined by
formula (6.6.17):
dg:(A)-= n(A){ anak = strlAN? (RO + pvt)
(Hie) ied avd MR + divdT ah — 57 dRda(>) = w(a){ ia?(ax + ao)
EAT (OE+ WE)
ALR avy )A]
-5\ aR + Raf + pvp" — (Rot + wrt) x(vivel) (bi R+ bavsTy)a hat+ gi(A)irT (ROT
Chapter 6 Methods of Linear Stochastic Systems Theory
648
+yudt (dvd)dY — (b1X + bode] 1
;
= g:(X) Euce + ao) — 5) (ak + Raf + wera] dt
+9:(A)iA7 (REP + oud (dived)? |ay — (bX + boat]. (6.6.21) As ae
= May Rd, we get from (6.6.21)
dgi(A) = gt(A Ec
i
1
+ ao) — A? a, RA — Swit
|dt
+9i(A)iXT (ROT + pw (davdT)? |a¥ —(b1X + bo)at]. (6.6.22) For evaluating the right-hand side of Eq. (6.6.20) we notice that (6.6.17) yields
092(A)
ay
eT gi(A)(iX — RA).
Using this formula we find
[Pa eu + (Pa = SAT yey a)0) |dt
Li[ieee Pxate = AT wat Jada) |
x(wivdt7} |aY — (6X + bo)at|
re Eucr. hp es +g(A)(iA? RO? + id Eq.
Swi
|dt
v7 )(div dT) [dY — (b1X + bo)dt]. (6.6.23)
(6.6.23) is identical to (6.6.22).
Consequently,
the normal conditional cha-
racteristic function determined by formula (6.6.17) with the parameters X and R
determined by Eqs. (6.6.15), (6.6.16) satisfies Eq. (6.6.20). This proves the normality of the conditional distribution of the random vector Xz and the validity of Eqs.
(6.6.15), (6.6.16) which determine the parameters NG and R; of this distribution if the initial values of Xi and FR at the moment tg represent the conditional expectation
and the conditional covariance matrix of the random vector Xg = Yo= Vee respectively.
to using the measurements
Find the optimal esti-
results of the process Y in time
interval [to : t]. Replacing the vector 0 by the random process O(t) which is determined by the differential equation dO(t)
=
0, and assuming
O as the state vector X
the corresponding system we obtain the differential equation of form Eqs.
of
(6.6.12),
(6.6.13) at a1(Y,t) = 0, ao(Y,t) = 0, ¥(Y,t) = 0. Eqs. (6.6.15), (6.6.16) take in this case the form
dO = Rb, (Y,t)? (divdT)-1(Y,2){ay — [bi (Y, 1) + bo(¥,t)]ae} , (1) dR = —Rb,(Y,t)? (divd)-1(¥, tbr (Y, t)Rat.
(III)
As the parameter @ is unknown and may be nonrandom the initial values of O and R are taken arbitrarily. The equations obtained will give in this case the estimate O which will be optimal assuming that the parameter @ is random and has the normal distribution which the expectation Og and the covariance matrix Ro.
Problems 6.1. De
Show that for the stationary random processes with covariance functions
“210 os a|r|) and De-4l"|
(1+ a|r| + a7?) the intensities of the equi-
valent white noises are equal to 4D/ Q@ and 16D/ 3@ respectively.
Calculate the
correlation intervals. 6.2.
Show
that for the process reducible to a stationary one determined by
formula Y (t) = bo(t) + bi(t) X(t), where X(t) being the stationary function the intensity of the equivalent white noise and the correlation interval are determined by
650
Chapter 6 Methods of Linear Stochastic Systems Theory
a Calculate
v(t) and
T, for the process
determined
by formula
Yb)
=p
(t)
+ bi(t)X(¢(t)), where p(t) being monotone increasing function. 6.3. Show that for the two-dimensional stationary system of Problem 1.3 with
uncorrelated inputs Xj (t) and X9 (t) with spectral densities $1 (w) and S(w) the elements of the spectral density matrix Sy (w) of the process Y(t) =
[Y (t) Yo (t) a
are determined by
:
$11(w) = $1(w)|®11 (iw) |?+ s2(w)|®12(tw)/’, $12(w)
=
81 (W)®11 (tw) Bo) (tw) + $2(wW)®12(tw)H29(iw),
$91(w) = 81(W)®11 (iw) Bo) (iw) + 82 (W)®12(tw) Do0(tw),
822(w) = $1 (w)|®2i (tw)|?+ 82(w)|22(iw)|?. Find the cross-spectral density of the processes X(t) =
[X1 (t)X2 aie and Y(t).
6.4. Check that for the two-dimensional stationary system of Problem 1.4 with
a stationary input X(t) = [X1 (t)X2 (t)|7 with spectral density s(w) the elements of the matrices Sy (w) and Sy (w) are determined by the formulae
Syiyy = 811(w) Yi)|?+ 522(w) |Y2w)|? + 512(w)[¥1(w) oo) + Y2(w)vi(w)];
Syiyo = + Syayx = Syay2 =
—$11(w) pi(~)b2(w) + 522(w)p2(w) div) 812(w)(|¥i(w)|? — Iv2)]’), —$11(w)P2(w)P2(w)+ 812(w)([41(w)|? — |v2)]*), 811(w) |Yo(w)|° + s22(w) |Yr(w)|? — si2(v) [ti(w)42(w)
ae V1 (w)po(w)],
Se,y, = $11(w)¥i(w) + $12(w)p2(w),
Saiy2 = —$11(W)b2(w) + si2(w)yi(w), Seay, = $12(wW)v1(w) + $22(w)y2(w),
Swoy2 = —$12(w)P2(w) + 522(w)y1(w),
Problems
651
~~
where
Yi(w) = [e(w? +e? + w?) + iw(w? — e? — w?)] c?(w),
Po(w) = [(w? +e? —w?) — 2eiw|c?(w)w, 9
6.5.
~3(w) = (w? +e? — w?)? 4 4e7w?). Show
that
=—Zel,b=J,
for the system
of the third
order
of Problem
1.9 at a
s. (w) = sl, I being the 3 X 3 unit matrix, the spectral density
of the output is equal to Sy (w) = sI/(w? aF 4e?). 6.6. Show that under the conditions of Problem 1.10 at X =
[0 0V2Da Ey,
V being a white noise of unit intensity, the spectral density of the output is equal to |®13(iw)|?
Bie ola iaa*
6.7.
a
®13(t1w)O23(tw)
®13(iw)O33(tw)
iw) D13(tw) Do3(tw)Pi3(tw)
iw)|? —— |®a3(tw)|
iw) 33 D3 (tw) 33 (iw)(tw)
O33 (iw) D13(iw)
O33 (iw) @o3(iw)
|©33(iw) |?
For the system with two degrees of freedom of Problem
1.11 with the
vector of the generalized forces representing a stationary process with spectral density
SQ (w) show that the elements of the matrices Sq(w) and SQq (w) are determined by the formulae
Sanq(W) = $11(W) Bar (tw) By (tw) + $29(W) Pp2(iw) O72 (iw) +s12(w)[ Dp,1 (tw) Dio (iw) + Dp2(iw) Py (iw) ],
SQiqn () = $11(H) Par(iw) + $12(H) Pra(iw), SQaqn(W) = 812(W)Pni (tw) + 822(w)Pr2(tw) (h = 1, 2). 6.8.
Find the spectral density Sy (w) and the covariance matrix
Ky of the
stationary vector process Ye) in the stationary system
Y
Yp0)!
VAR Y4
Oter
thd wpe pas 0
Yi
0
fede)
0
ohfestigert
ae
KOH
Fhe
TAOS
1016
Yo
(1)
ay
pe Pe
0
0
—w?
—2€
Y4
q2
where V is a white noise of unit intensity, and ¢1, 42, Wo, W1, €, €1, (€,€1 > 0) are some constant coefficients. 6.9.
Find the spectral density Sy (w) and the variance Dy of the vertical os-
cillations of the body of a car in linear approximation (Problem 1.15) assuming that
the spectral density of the road microprofile is known, $] (w) =2 s(p)/v, p=
Qn/c.
652
Chapter 6 Methods of Linear Stochastic Systems Theory
6.10.
Find the spectral density Sz (w) and the variance D), of the stationary
random oscillations of an object with dynamic damper of oscillations in Problem 1.16 under the action of the random disturbing force with the spectral density sp (w). 6.11.
For the stationary linear system of Problem 1.12 with three degrees of
freedom at A = I, C = wi], Bo =)2el,| B! = C* =.0,4Qg(w) = (27)?sl (I being the unit 3 X 3-matrix) derive the formulae for the spectral density Sy (w) of the stationary process aah) = [qi(t) q2 (t) q3(t) qi(t) ga(t) s(t) (he Find also the covariance matrix Ky. 6.12. Derive the formulae for the spectral and the cross-spectral densities Sz (w) and Spy (w) for the system of Problem 1.13 at
B =
2€A and B =
nC assuming
that the spectral density Sz (w) of the input is known.
6.13. ®(s) =
Show
that for the linear stationary system with the transfer function
(dis + do)(c3s° + C28 +ces+
cola. the input being a white noise
v(codj+e2ds) with the intensity l/, the variance of the output is equal to Dy _TDaaleies 2eteay” Calculate also the covariance key:
6.14. Show that the covariance function ky (r) and the spectral density Sy (w) of the random function Y(t) components
=P
Gl (t)X2 (t), where X, (t) and X9 (t) are the
of the two-dimensional real normally distributed random process with
the known covariance function and spectral density
m= [EG tan| = [te are calculated by the formulae
ky (7) = ki(r)ko(7) + k12(7) kai(7) + miki (7) + m5ki (7),
Hee / Ge Sven Ode / somos (Out +m? so(w) +ms1 (w). Give the generalization to the case of Y (2) = xe (t)X2 (t) where Xj (t) and X9 (t) are the n-dimensional vectors. 6.15.
Find the variances and the covariances of Y; and Y2 in Problem 1.6 at
the input X being the white noise with the intensity V.
6.16. Find the variance of Y; in Problem 1.7 for X being the white noise with the intensity V.
6.17.
Find the variance and the covariance function of Y; in Problem 1.7 for
X being the white noise with intensity V.
Problems
653
6.18. Deduce formulae for the expectations of the number of intersections
Y of
a constant level @ and of the number of stationary points Y; during the time J' for a stationary normally distributed random process X (t) with the expectation Mz and the covariance function ke(r):
_T | k¥(0)
Bie
,
ee
ereatiy eee 1) Fe
(0)
tesa rec
6.19. Show that for the system
Yj=3,
iG
(I)
ap + V,
Pa a
where
do is the constant,
V is the white noise of the constant
intensity
V, the
expectations, the variance, the covariance and the cross-covariance functions of the
processes Y (t) and Y> (t) at zero initial conditions are determined by the formulae my,
=
agt? /2,
K41(t1,
m2:
te) —
Koi (ti, te) =
=
dot,
bes
ky
=
vt? /3,
min(ty, t2),
vt2/2 min(ty,
tz),
ky =
Ky0(t1, ta) = Ko2(t1,
vt?/2,
kao =
vt, /2 min(ty,
vt,
(II)
ta),
to) => vymin(t1, tz).
(III)
6.20. Prove that for the Langevin equations Y, =
Y,
Y> —
—2eY> a5 V;
(I)
where € > (is the constant, V is the white noise of constant intensity l’, the variances and the covariance of the processes Yj (t) and Y (t) at zero initial conditions are determined by the formulae:
ky, = v(4et — 3 + 4e—7** — e~ ***) /16e%, kia = vl — 2677? 6.21.
e748") [8e*, »kop = (1 e5**")/4e.-
(ID)
Show that under the conditions of Problem 1.5, the generalized force Q
being the white noise of intensity /, Eqs. (6.1.8) and (6.1.11) have the form
kiy = 2ki2/A, ki = —Chi — (B/A)k12 + (1/A)ko2,
koo = —2Ck12 — (2B/A)koo + v,
(I)
OK 11(t1, t2)/Ote = Kio(ti, t2)/A,
aK i (ts 1s) /ots = —CK is(t1, 4)= BRia(hayt)/A) OKo1(t1, t2)/Ote = Koo(ti, t2)/A,
Raa (ti, fly 2 CRA (Hs) = BRaa(ti,t2)/A.
(I)
654
Chapter 6 Methods of Linear Stochastic Systems Theory
Write the explicit formulae for the solution of Eqs. (I), (IJ). 6.22. Prove that the covariance and cross-covariance functions of the stationary
and cross-stationary processes Yj (t) and Y2(t) of Example 1.1.9, the input X (t) being a white noise of the intensity
, are determined by the formulae
Bris = rar e~S“ol7l(cosw.7 + ysinw,|r]), 0 koo(T) = ies e~S¥el7l (cos WeT — ysinw,|T|),
Vy kyo(T) = Seebes
: a sinwelTke(y =¢/V1—¢7,
we = wov/1 — C?).
6.23. Show that under the conditions of Problem 6.21 the variance of the coordinate and of the impulse in the regime of stationary random oscillations, i.e. when the
oscillations represent a two-dimensional stationary random process, are determined by the formulae kj; = Vi ZB:
kia = 0, ko5 = Av /2B. Show that in the case of
heavy damping (B 2A AG ), the process of stationary oscillations settling consists
of two stages: first, the oscillations of the impulse settle, then follows the slow settling of the oscillations of the coordinate.
6.24. = we
Show that the set of Eqs.
(6.1.8) for the system of Problem 1.3 at Zp
briV; (h iin 2s where V =
[Vi te
if is a white noise of intensity
V, and bp; are the coefficients depending on time has the form nm
kyy = 2(a41k11 + ai2ki2) + > Vrrbirbin, ri h=1
kia = aoiky + (411 + ao2)ki2 + ai2ke2 + a Vrproirban, ryh=l n
ko = 2(d2ik12 + ao2ko2) + ‘se Urnbar ban. pha
Obtain from these equations the corresponding equations for the system of Prob-
lem 1.4 at X;
=
Vij, Xp
=
Vp».
Find the stationary solutions in the case of
constant intensity Y. Consider the process of settling of the stationary solution.
6.25. Show that for the linear system with one degree of freedom of Problem 1.5 in the case where the generalized force with the spectral density Sz (w) =
@ represents a stationary random process
Da[1(a?
+ w?) the set of Eq. (6.1.8) for the
components Y; = q, Y2 = p, Ys = Q/A of the state vector has the form
kiy =2ky2,
kip = ko3 — oki,
Problems
‘
655
-
;
kia = keg — woki1 — 2eki2 + h13, ko = —2(woki2 + 2eko2 — kos), ;
:
:
kez = —woki3 — (a + 2€)ko3 + k33, k33 = —2a(k33 — D). 6.26.
Under the conditions of Problem 1.15 show that if a car is moving with
the speed U and the covariance function of the road profile is exponential, =
De-l1,
then the state variables Y;
= Y, Yo =
Y, Y3 =
ky(o )
q (components of
the expended state vector) satisfy Eq. (6.1.1)
0
a=
| -—w2
1
0
-2e
w?
0-
—av
0
with the white noise of unit intensity.
0
|,
a=0,
b=]
be b3
Verify the validity of the following formulae
for the variances and the covariances in the stationary regime: D kx
=
koe =
wW6
2 [deav +
2Dav
(14 $4)a)
pt
(< SF it) )
2E/
€
k33 == DD.
kia. =),
ko3 =
kg
avk13
=
=
2D;
avpD
we + Bea aa we + 2eav + av? } © 6.27. Show that for the system with two degrees of freedom of Problem 1.12 at
OSV
= [Vi V2 ih where V is a white noise of the intensity
/ Eqs. (6.1.8) have
the form
ki = 2(Ajok13 — Ajok14), ki2 = —Ajgki3 + Ayyki4 + Agoko3 — Ajokoa,
ki3 = —Cirkir — (Cio + Cha) ki2 — [B11 Ady — (Biz + By2 Aja) Jhis
—[(Bi2 + Byy) Ay, — Bir Ajy )kia + Aggks3 — Ajoksa, ki4= —(Ci2 — Cio) k11 — Cooki2 — [(Bi2 — By2)Ag_ — Bo2 Aj )kis
—[ Bog Ay, — (Biz + By) Aye |hia + Aggksa — Ajgkaa, kog = 2(—Ajgkos + Aji koa), ko3 = —Cirki2 — (C12 + C{2)k22 — [ Bir Ady — (Biz + Biz) Ajo |kos
—[(Bi2 + Biz) Ay — Bir Aj |hoa — Ajokaa + Ajykaa, koa = —(Ciz — C}y)k12 — Co2ko2 — [(Bi2 — Bia) Az, — Baz Ajy kos
656
Chapter 6 Methods of Linear Stochastic Systems Theory
—[Bo2Ay, — (Bia + Bia) Ajy koa — Atokaa + Aq haa,
ka3 = —2Ci1k1a — 2(C12 + Cjp)koa — 2[ B11 Ag — (Biz + Biz) Ajy Jka
—2[ (Bio + Byy)Ay, — Bi AjeJksa + 111, k34 = —Ciikia — (Cig — Cla)kis — Caokos — (C12 + Cia) ke
—[Bi1A2, — Boo Ay, — 2Bi2Ajp Jksa — [(Bi2 + Bia) Aj, — Bir Ajo |haa —[ (Biz — By2)Ag, — Bar Ajy Jkas + “12, kag = —2(Ci2 — Cia) k14 — 2C22k24 — 2[ (Biz — Bi2)Az, — Bor Ajy Jka4 +2[ Boo Ay, — (Biz — Biz) Ajo )kaa + 22, where A;; being the elements of the inverse matrix A~ My 6.28.
Write the equations for the variances and the covariances of the coordi-
nates of the linear system with the dynamic damper of Problem 1.16. Show that at ba—n05 (ca =
0) variance of the process X (t) in the regime of stationary random
oscillations is determined by: D, = v[ prc + (Hat p)b2]/2c7bap2. 6.29. Using the equations of Problem 6.27 write the equations for the variances and the covariance of the processes Yj (t) and Y2(t) at Ay; = Agog = A; Aj2 = 0; Bi,
=
Boo
=
B; Bio =
0; Bi
=
H; 41
=
¥92
= V; M12
=
0. Verify that
the solutions of these equations with zero initial conditions are determined by the formulae:
(a) At
C1, = Cig = Coo = Clip = 0 and
A= v/(H? + B?):
kit hoo = Al, skys = bog = A/2, “Fig = ka, = 0 ki4
6=
(b) At Cy, = BH —AK:
=
—ko3
=
HA/2B,
Cis =.Goy-
=
k33
0, Cio =
kii = keg = Bw/2K,
= kag
K(BH
=
V/2AB;
> AK)
ands
v/6,
33 = kag = KH/A,
kyo = kig = koa = kag =0,
keyg = —ko3 = K/2.
Using formulae (5.1.15), (5.1.16), (5.1.24) calculate the derivative of
H = (p" A~'p+q"Cq)/2. Show at the stationary random oscillation that the mean power FH damping (B’ fe 0) and position nonconservative (C - 0) forces.
is wasted on
Problems
6.30.
Show that Eqs.
657
(6.1.8) for stationaty linear system of Problem
Q = IV where | is a constant m X n-matrix,
1.11 at
V is an n-dimensional white noise of
intensity l/ represents the set of equations:
Ky, = A7* Koi + Ky A“,
Kg = A-'K22 — Kui(C — C') — Kin A71(B — B’), Rete Ae
Rog
(CaO iy
(GG) Kia
(Bt BA
Kai( C=C) (84-8)
Koy,
Ar a=
Ky A-'(B — B) +P; where Ky; and K92 are the covariance matrices of the vectors of canonical variables Yi =
q, Y2 = Pp, and the blocks Ky
of Y; and Y2.
Cee =), RC. 6.31.
and Ko
Consider the special cases:
Write Eqs.
(a)
(Cl 20, 8? =A,
are the cross-covariance matrices
C =
Cen
Boats
As (b)
(6.1.11) for the blocks of the covariance and the cross-
covariance functions of the canonical variables of the system of Problem
6.30 for
the regime of stationary random oscillations. 6.32.
Show
that under the conditions of Problem
6.19 the one-dimensional
characteristic function of the process Y(t) = [Yi (t) Y> (t) iia is determined by the formula
Gis
2b)
Gola,A } expyiagt(X" + X’'t/2)
t
nar.
+/ Re Are 6.33.
Show
that under the conditions of Problem 6.20 for a stationary white
noise with any function x( jt) the one-dimensional stationary distribution of the process Y9 (t) is detemined by the characteristic function:
Z:
gi(A) = exp
»
DP
xa
x | ul dy. 0
6.34.
Under
the conditions
of Problem
6.25 write the formula for the one-
dimensional characteristic function of the strictly stationary process Yj (t) in the
case of an arbitrary function x(H). Consider also the case of x(H) = =i, vp/2.
658
Chapter 6 Methods of Linear Stochastic Systems Theory
6.35. The three-dimensional Brownian motion of a particle under the action of the disturbing force representing the stationary white noise is described by the vector
Langevin equation
Y +2cY = y,,
(I)
Y being the state vector of the particle, V is the stationary white noise with the
known function x( HL). Show that the one-dimensional characteristic function of the velocity vector
U = Y of the particle in the stationary regime is determined by the
formula
Re) ee
i!xe
"dr } SMD).
0
Consider the case of the normally distributed white noise. 6.36. Show that the multi-dimensional distributions of the random process Y (t) determined by the differential equation
Y =ao+aY
+ bX (t),
(1)
where X (t) is the random function representing the solution of the stochastic differential equation
X=aot+aX
+ BV
(II)
with independent initial values Yo and XQ are determined by the formula: n Gri
seoeins ticle
00
n
S— uri(te, to)” A
9
So ura(te, to)? A
peal n
Xx €xXp
tk
iso k=)
n
== k=1
it
[ent
tk
rao(r)ar
to
te
/x
kod
+
fra(te,r)a0(r)ar to
n
BU
Ys tod)it
|r
a
Ag)
l=k
Ii where % (p), 9o(c) are the characteristic functions of the initial values Yo, - 2
the process Y (t), X (t); U11 (t, T) is the fundamental matrix of solutions of the homogeneous equation duj,/dt = auj4, ui1(T, T) =F
ur2(t, T) is the solution
of the equation duj2/dt = auU12+ bu22 with zero initial condition, u12(T, T) — J)
_Problems
659
u2a(t, T) is the fundamental matrix of solutions of the equation duy2/dt = U9} u22(T, T) — rf (Ig is the unit matrix of dimension q). 6.37. Show that the stationary variance Dy of the output of the system consist-
ing of the impulse device which generates the rectangular impulses of the duration Ty (with the period repetition 7;) and the linear link with the transfer function ®(s) under the action of the impulse white noise with the variance D) of each impulse are
determined by the formulae:
1) Dy = kD/(k + 2) at ®(s) = k/s,
yw ee EDpl21) at ®(s) = k/(Ts + 1). Here p(z1) = 2?(z)’ — 1)?/(1 — z?), 21 = e-7*/7, y = Ta/Th. 6.38.
Find the expressions for the spectral densities and the variances of the
output for typical linear systems with distributed parameters (Table A.5.4).
6.39. Consider Example 6.3.1 for the case when 7' = 7}, mz (t) =a-+t
Kotter Solution.
bto,
eet. According to the first method if we use the result of Example 3.9.12
at y(t) = VDe®*, choosing the coordinate functions in the form
zn(t) = —
sin E (:-) - | (ene
Oi)
we obtain respectively t
my = af VP (a4 balan = a+ Ht 7) —(a- We!" () 1
—
VolEh= :,Method 8aiD Aural pata le sin(wnt + yn)—e-/T sin Pn| :
(T+ An)
J+ BT)? + w2T?
n [oP sin(wnt +n) —e/Tsingn] (Tr an)[ tet tort]
Ne Dy _
nt
tT
Bp Nl
ae
aoe
(IIT)
”
CIE
Saye
Here the exact solution of te variance DS (t) is determined by formula t
t
Dy*(t) = - /fire
rae
yy
660
Chapter 6 Methods of Linear Stochastic Systems Theory
D (2A a4 ex Plt)
2D (eft _ 1)
(1+. 67) [P(@nep)r]
pen? 4 eam
(V)
(2: = 8-—a+1/T). While using the second method the following equations are solved
Tring+ my
b1, = 0-4
Tin
+ Un
=
2MEO De sin
T
[wn (¢- 7) +=],
(v1) their integrals which vanish at t = 0) are expressed by formulae (I) and (III).
CHAPTER 7 METHODS OF NONLINEAR STOCHASTIC SYSTEMS THEORY AND THEIR APPLICATIONS
The last Chapter is dedicated to the basic methods of the nonlinear stochastic systems described by the stochastic integral, differential, integrodifferential, difference and other nonlinear operator equations. Section 7.1 contains the derivation of the formulae for the derivatives with respect to the moments of different orders. Special attention is paid to
the derivation of the infinite system of the equations for the moments. The equations for the moments in the stochastic linear systems with the parametric noise are composed. In Section 7.2 the methods of the normal approximation and the statistical linearization for determining the oneand the multi-dimensional distributions are stated. The methods of the equivalent linearization and the general methods of the normalization which form the basis of the structural methods of nonlinear stochastic systems theory have a significant place in Section 7.2. Section 7.3 contains the brief account of the methods of general nonlinear stochastic systems and the statement of the general problem of the parametrization of the one- and the multi-dimensional distributions. Methods of the reduction of the number of the equations for the distributions parameters are considered. The moments methods, the semiinvariants methods, and the moment-semiinvariants methods for defining the one- and the multi-dimensional distributions are stated in Section 7.4. Section 7.5 is dedicated to the methods of the orthogonal expansions and the quasimoments. Method of structural parametrization based on the ellipsoidal approximation of the one- and the multi-dimensional distributions is given in Section 7.6. Section 7.7 is devoted to the methods based on the canonical expansions and the integral canonical representations. Sections 7.8 and 7.9 contain the applications of the nonlinear stochastic systems theory methods to the analysis, modeling and conditionally optimal filtering.
7.1. Moments
of State Vector
7.1.1. Formula for Derivative of Expectation > The simplest way to find the formula for the time derivative of the expectation of the state vector of a system is to pass to the expectations directly in Eq. (5.4.6)
for the normal white noise with the matrix intensity
Vy=
v(t). Then taking into
662
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
consideration that in the It6 equation the value of the process Y(t) at any ft is independent
of the value of the white noise V at the same instant we obtain the
following formula for the time derivative of the expectation m(t) = EY (t) of the
process Y(t): i=
Remark.
Ea(Y,t) sii,
Cisled)
This formula is not a closed equation which
determines
the
expectation of the process Yi(G)y as the expectation in the right-hand side depends on the unknown one-dimensional distribution of the process AGE).
7.1.2. Formula for Derivative of Second Order Moment > Let us denote by Yxj = EY; (EN (t) the components of the matrix of mo-
ments of the second order [ = EY(t)Y(t).
Using formula (5.1.7), we evaluate a
stochastic differential of the product Y;z (t)Y; (t). As a result we obtain the following
expression for a stochastic differential of the product:
dY,(t)Y; (t) = \%az(Y, t) + Y; a; ‘ee t) + 533 De (=)
os),
Dre. w}dt
Sb
m
+ S—[Yeber(¥,t) + ¥jbjr(¥, t)]dW,.
(7.1.2)
(Pal
While evaluating the expectation of both parts of (7.1.2) and accounting that [ =
lye; | (kj 3 leateg Pp) we come to the sought formula for the derivative of the
second order moment
De By of, 4)
Vue Viet)! bie
2 HY +
(7.1.3)
Here
o (Yt Remark.
Formulae
)b=.b(¥id)
ple )o04)7 ab
(7.1.4)
(7.1.1) and (7.1.3) are not closed equations for m
and I’ in the general case, as their right-hand sides depend on the one-dimensional distribution of the process Ya) i.e. not only on m and I’. And only in some special
cases the right-hand sides of formulae (7.1.1) and (7.1.3) may be the functions of m
and I’ independent
of other characteristics of the one-dimensional
distribution
of the process NAGAY In such cases (7.1.1) and (7.1.3) will be ordinary differential equations which determine the moments
m and I’. So, for instance, in the case of
7.1. Moments of State Vector
663
linear functions @ and b formulae (7.1.1) and (7.1.3) represent the equations which determine the expectation ™ and the second order moment separately.
I" of the process Y(t)
These equations certainly coincide with equations
of Subsection
6.1.1
obtained for the more general case.
7.1.3. Formula for Derivative of Covariance Matriz The formula for the time derivative of the covariance matrix K of
the vector Y is easily derived from formulae (7.1.1) and (7.1.3). > In order to obtain this formula it is sufficient to differentiate the relation KC =T—mm!
with respect to t and substitute into the obtained formula the expres-
sions of m and I’ from (7.1.1) and (7.1.3). As a result we get
K = E {a(Y,t)(Y? — m7) + (Y — m)a(¥, 2)" + o(¥,t)}
a
(7.15)
So the results obtained in Subsections 7.1.1 and 7.1.2 may be summarized in the following statement.
Theorem 7.1.1. [f the random process Y(t) defined by Eq. (5.4.6) with the normal white noise V = V(t) has the finite expectation m, the second order moment T and the covariance matriz K then formulae
FAECES)
a(t
Srareeaids
For formulae for the derivatives of expectation and second order mo-
ments in the case of Eqs. (5.4.7) and (5.4.6) for V being the nongaussian white noise see Problems 7.1 and 7.2. Example
7.1.1. For the system described by the equation
Y= Ye Ve
(I)
formulae (7.1.1), (7.1.3) and (7.1.5) have the form
m= l| —EY?, a2 = —-2EY* + vaz,
:
D = —2EY* + MEY? + v(m? + D). In this case the process Y(t) is scalar, and consequently, I‘ and tively the initial second order moment
(II) K represent respec-
a(t) and the variance D(t) of the value
of the process Y(t) at the instant t. The right-hand sides of formulae (II) depend on the third and the fourth order moments
of the process Y (t), and consequently,
formulae (II) are not the equations determining ™,
2
and D.
The attempts to
664
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
close these equations by adding the equations for the third and the fourth order moments will lead to the appearance of the moments of the fifth and sixth orders, and
so on.
Consequently, such attempts will not lead to a closed set of finite number of
equations.
Example
7.1.2, For the system
Y =aY +ap+bVYV
(I)
formulae (7.1.1), (7.1.3) and (7.1.6) give m=am-+ao,
&2 =2(aa2z+aom)+b?vym,
D=2aD+b?vm.
These ordinary differential equations with given initial conditions:
(II)
m(to) =
Ne
a2(to) = 20, D(to) = Do = a0 — m? determine consecutively the moments mM, A
and D.
7.1.4. Formulae for Derivatives of Second Order Moment ance Function
and Covari-
Later on we shall also need the formulae for the partial derivatives of
the second order moment I'(t1,t2) and the covariance function K (ty, t2) of the process Y(t) defined by Eqs. (5.4.6) with respect to the second argument fg at t1 < to. > Using
d:,V(t1,t2) = EY (t1)dY(t2)’, d:,K(t1,t2) = E[Y(t1) — m(t1)]dY (tz)?
(7.1.6)
and using Eqs. (5.4.6) we shall get at t} < tg the following formulae:
OV (ti, t2)/Ot2 = EY;,a(Y;i,,t2)",
(FAM
OK (ti, te)/Ote = E(V:, — m:, )a(%,,t2)? . «
(7.1.8)
Theorem 7.1.2. If the random process Y(t) defined by Eq. (5.4.6) with the normal white noise V = V(t) has the finite first and the second order moments for instants t; and tz then the matriz of the initial
moments and the covariance matrix satisfy formulae (7.1.7) and (7.1.8). Example
7.1.3.
For the system of Example 7.1.1 formulae (7.1.7) and
(7.1.8) have the form
AV (ti, t2)/Ot2 = -EY,,Y2, OK (t1,t2)/Ot2 = —E(¥:, — m:,)¥3.
7.1. Moments of State Vector
Example
7.1.4,
665
For the system of Example 7.1.2 formulae (7.1.7) and
(7.1.8) have the form OT (t1, t2)/Ot2
=
a(t2)T'(t1, te) ,
OK (ti, t2)/Ote
=
a(t.) K (ti, te) ;
(1)
These formulae represent the ordinary differential equations at any ;. Together with
the initial conditions:
P(t, . t) = ao(ty ); K(ty ; t1) = D(t1) they determine the
second order moment P(t, ; to) and the covariance function K(ty ; to) of the process Y(t) atrbje —nn h=1
b=k
Opening in Eq. (7.1.16) the brackets we get
b Sule
Rad
agmiie mat + bovb2
(7.1.16)
668
Chapter 7 Methods of Nonlinear Stochastic Systems Theory P
P
h=1
Ayla
+S \(b,vb) + bovb, )EYn + D> bavb7 EYaYi = al +Ta™ + apm? + ma? + bovbd
+S
Pp
P
2(bavbg + bovb;)mna + > bavbt yn.
sal
(110)
hil=ti
Eq. (7.1.17) does not contain any characteristics of the random vector Y except in its
expectation and second moments (the elements of the matrix I’). Consequently, after
integrating Eq.
(7.1.15) which determines the expectation ™ of the vector Y, Eq. (7.1.17) with the initial condition I'(to) = Tp (and consequently, Yai (to) = 7)
completely determines the second order moment
I(t) of the vector Y¥():
In exactly the same way we reduce Eq. (7.1.5) to the form
P K =aK + Kal? + bovbt + So (bnvbg + bovb? mp
as P + $5 bavb? (mam + kn), K = [kn],
(7.1.18)
pal
where
kp] is the covariance
of the components
Y;, and Y) of the vector
Y
(A, l
=1,..., p). Eq. (7.1.18) with the initial conditions K(to) = Ko (kin = kp,) completely determines the covariance matrix K(t) of the vector Y(t) at any time
instant ¢ after finding its expectation Mm. < Thus if we confine ourselves to the moments
of the first and the second
order
of the state vector and of the output while studying a linear system with parametric
white noises (7.1.14) then these moments
may
be exactly determined
by subsequent
wntegration of Eqs. (7.1.15). (7.1.17) or Eqs. (7.1.15), (7.1.18) as well as in the case of a hnear system.
For a linear system with parametric white noises (7.1.14) formula (7.1.8) gives the following equation for the covariance function of the process YA(E) at to > ti:
OK (ti, te) /Ote =
K (t1, t2)a(t2)? ‘
(7.1.19)
The initial condition for this equation has the form K(ty ; ty) = K(ty ,: At last it is easy to conclude that the moments (7.1.14)
are determined
of the state vector
by the infinite set of Eqs.
(7.1.13) which
Y of a system in this case
is
7.1. Momefits of State Vector
669
decomposed into independent sets of equations for the moments
(Karsoitese, pr ss Oatley Sct. P
P
a, = Sk
Gr o&k-e, + ete
Ga! t
ines
gai P
P
4 9 Dal Es er 1) (ornote-e, tt; aia pel
p pal
>4
Fonceteadthtegtes-2e
q Uu= 1
5
a
Pp
=
Ak+e,—e,
q=1
P a
of each given order k
al| = kychine thy = 1y2,....):
Daya
n=
kinks (crsocnevne
asi
Pp
Ors,egtk+eg—er—es q=1
Fir ») cra pielcnriyiA—t,—«,
:
qual
(7.1.20)
Here as well as in Eqs. (7.1.13) Qs is equal to zero if at least one of the components of the multi-index S is negative, and is equal to unity if all the components of the multi-index S$ are equal to zero.
Example
7.1.7. In the case of scalar Y and V Eq. (7.1.14) has the form Y=aY+aot(bo
+h
Y)V.
(1)
Eq. (7.1.15) has in this case the same form as in the general case of vector Y,V.
(7.1.17), (7.1.18) for the second initial moment @2 = T and the variance the process
Eqs.
D = K of
Y take the form
Gg = (2a + 6?)az + 2(ao + bobiv)m + bev, ;
D = (2a — b2v)D + 2bobiym + bivm? 4 bev.
II
(i)
In the case of a normally distributed white noise V we may also obtain the exact equations for the moments
of higher orders. These equations have the form
1
api—-ki|.are atk —1)b?v |a, +k [a0 + (k — 1)bobiv Jagi
i + 5 h(k Sl b2vag > “(k= ap4apey)s
(IIT)
Let us consider the stationary linear system (7.1.14) with stationary parametric white noises.
In this case @, dg, bo, bp, and vy are constant.
Therefore, putting in
670
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Eqs.
(7.1.15), (7.1.18) m =
(0 and K
= 0 we obtain algebraic equations for the
expectation and the elements of the covariance matrix of the value of the stationary
process in the system at any instant tf:
am+aj=0,
(Aa 21)
Pp
Pp
b=
i
aK +Ka™ +bovbh +> (brvb5 +bovb; mnt Y_ bavbs (mami+kni) = 0. (hak 22) For finding the covariance function of the stationary process bz ), T=1t,-To, we write the general equation for the covariance function
Hence,
using
OK (ty, t2)/Ot2 = K(ti,t2)a? at te
>t).
the
function
property
of
the
covariance
(7.1.23) K (ty p to)
= K (te, t,)? and transposing Eq. (7.1.23) we shall have OK (ta, t1)/Ote = ak (to, t1) ateoy Q. At
T < 0 covariance function is determined
by
k(r) = k(—r)?.
We may also determine the moments
of higher orders of the one-dimensional
distribution of the stationary process in the system. For this purpose one should put
a, = 0 (ki, ee Remark.
al aa bs epay et |k| = hy Pies
By = Be nee
(7.1.26)
It should be noticed that the wide sense stationary process in
a linear system with parametric white noises may physically exist only in the case
where the systems described by deterministic Eq. (7.1.15), (7.1.18) are stable. But this condition is generally not sufficient for the existence of the strictly stationary process in the system.
7.1. Moments of State Vector
671
7.1.7. Formulae for Moments in Stochastic Discrete Nonlinear Systems The formulae for the first and the second order moments in conformity to dis-
crete system (5.7.1) are obtained on the basis of Eq. (5.7.7) and have the following form
mis, = Eun (Yi, Vi), (71220) Key = E{wi(¥i, Vi) — migi] [wr(¥i, Vi)" — mf, ] . (7.1.28) Analogously on the basis of Eqs. (5.7.14), (5.7.15) and (5.7.17) the formulae for the moments
are derived also for the case of continuously-discrete systems (5.7.10) and
(5.7.11).
For the nonlinear systems formulae (7.1.27) and (7.1.28) are not closed
equations.
7.1.8. Equations for Moments of Discrete Linear System with Parametric Noises As an example of direct derivation of the equations of Subsection 7.1.7 for the first and the second order moments let us consider the following discrete linear system with the parametric noises: P
Yiga = a Y) + ao7+
bor +
) bi; Yi;
V,,
(han 29)
j=l where bo), bi, --.
, byt are the matrices of the same dimension as the matrix do]
ands Yyqyisey Yip are the components of the vector Y7.
> After taking the expectations of both parts of Eq. the independence of Y; and Vj and also the equality
(7.1.29) and accounting
EV; = 0 on the basis of the
theorem of the expectations multiplication for the independent variables we get
Mmi41 = aym; + aor.
(7.1.30)
Similarly as for an ordinary linear system while deriving the equations for the covariance matrix K; =
Eq.
E(Y; = mi)(Yi = m)
(7.1.29) term-wise Eq.
of the process {Y;} we subtract from
(7.1.30) and multiply the result by the correspondent
equality for the transpose matrix.
After taking the expectation of both parts of the
obtained equality we come to the equation
EV
aa mug1)(Yit1 a Mi41) = a E(Y; — mi)(¥;" a mj )a;
672
Chapter 7 Methods of Nonlinear Stochastic Systems Theory P
P
gal
Gia
(7.1.31)
+E | bor +S ds1¥jr |ViVi" |bor + D0 ORY | -
But by virtue of the independence of Y; and Vj and according to formula of the total expectation we have
P EB bor + >> br Yiy viv,
p bor + > bY iy
Gia
jal
Pp
= E {bn + >> by %y | BLVV
P
.
[Yi] |b+ Do onY,
ah
j7=1 P
P
Yo
j=1
=E |bo + >5by ¥y |B[VVT] | bo + D0 ORY Pp
P
j=l
Gil
= E | bor + 5) by ¥ig |Gr |OG + Don P = boiGib2; + ‘Ss (bo1G1b}, + bj1Gibby) mu; j=l
p.P + SS biG (rag min + keayn)
(7.1.32)
jJ=1 h=1
where ™j; are the components of the vector ™; ki;h are the elements of the matrix kj; G} is the covariance matrix of the variables V). Substituting expression (7.1.32)
into Eq. (7.1.31) we get the sought difference equation for the matrix K7:
P Kiga = ay Kyat + borGibhy + Y>(borGibh,
j=l
P
+bj1G1bg)) mj + YS bj1Grbhi (maj min + kin). 4 =
h=1
(7.1133)
7.2. Methods
of Normal
Approximation
and Statistical Linearization
Thus Kq. (7.1.30) with the initial condition my
the expectation {mi} Ly
=
E(Y,
=
of the process {Yi}, and Eq.
mi)(Y4
=
m,)?
determines
= HY,
673
determines
completely
(7.1.32) with the initial value
the covariance
matrix
{Ki}
of the
process {Yi}. For finding the covariance function Kj, h) of the process {Yi} we
by means
of Eqs.
(7.1.30) and (7.1.33) E(Y; — ms )(Yru _ a)
Accounting the independence we
evaluate
ae
Oe
of Yp, Y; and V; atj < hand the inequality EV, = 0
obtain
K(j,h+1)=K(j,h)az . Eq.
(7.1.34) with the initial condition K(j, J) =
(7.1.34) K; determines
completely
K(j,h) ath >j. Ath Approximating the one-dimensional distribution of the random process YZ)
in normal stochastic differential system (5.4.6) by a normal one we shall have
gi(A;t) & exp {iat
it
37 Ka};
CS:
fily;t) © [(2m)? [K|]7"/? exp {-30" —m™)K-(y - m)}_€7.2.2) where ™
and
K
are the unknown
vector of the system Y.
expectation and covariance matrix of the state
674
Chapter 7 Methods
of Nonlinear Stochastic Systems Theory
After calculating the expectations in (7.1.1) and (7.1.5) for the normal distribution V (m, K ) we obtain the ordinary differential equations approximately determining ™ and K:
m= i(m, K,t),
m(to) = mo,
(7.2.3)
K = 92(m,K,t),
K(to) = Ko,
(7.2.4)
where
gi(m, K,t) = Ena(Y,t), p2(m,
K a) =
poi(m,
(7.2.5)
K,t) + poi(m,
Katy.
+ ~22(m,
yoi(m, K,t) = Ena(¥,t)(Y? —m’), o(Y,t) = W(Y,t)v(t)b(Y, t)”, 2o(m, K,t) = Ewo(Y,t) = Enb(Y, t)v(t)0(Y, t)?
Ket)
£026)
(7.2.7) (7.2.8) (7.2.9)
and the subscript V denotes that the expectation is calculated for the normal distri-
bution N(m, K) of the random variable Y :
Ex()= ae f ew [507 mi")Kym] The number of equations for m
(NAM) is equal to QNam
dy. (7.2.10
and K by the normal approximation method
= P(p + 3)/ 2. 4
Thus, the NAM for the one-dimensional distributions in the normal
stochastic differential system is based on Eqs. (7.2.3), (7.2.4) at (7.2.5)(7.2.10) and corresponding initial conditions. 7.2.2. Statistical Linearization Method
In the special case of the unit matrix b(y,t), b(y,t) = I, g22(m, K,t) = v(t). In this case NAM gives the same equations as the Kazakov statistical linearization method (SLM). > Really, SLM is based on the approximate formula
O(Y,t) ¥ yo +ki(Y —m), where (9 and ky are determined by minimizing the mean assumption
of normality
of the distribution of Y.
(7.2.11) square error under the
Here Yo and ky are given by
7.2. Methods of Normal Approximation
the formulae
~9
=
and Statistical Linearization
yi(m, fagteaA and ae) =
ki(m, IK, t) =
675
poi(m, Nyt). Hex,
Replacing the function a(Y, t) by the obtained linear function of Y we reduce Eq.
(5.4.6) in the case of b(Y, t) = I toa
linear stochastic differential equation
Y =¢o4+hi(Y —m)+V,
(7,212)
Using methods of Section 6.1 we get in this case the following approximate equations for mand
K:
m= 0,
(2213)
K=khK+Kk? +v.
(7.2.14)
Substituting here the obtained expressions of (9 and ky, we see that these equations
coincide with Eqs. (7.2.3), (7.2.4) at 22(m, iG t) = VV @
For SLM practical application the tables of formulae were compiled
for Yo and k; = [0/dm)yo iF for typical scalar and vector composite functions. In Appendix 3 the formulae for go are given for some typical composite functions. These formulae may be used for the determination
of the function goi(m, K,t) = kK = [(d/dm) yf ]* K, while setting up Eqs. (7.2.3), (7.2.4) of the normal approximation method. We may also use these formulae for determining the function ~22(m, K,t), as
according to (7.2.9) it represents the first item in the formula of the form of Eq. (7.2.11) for the statistically linearized function b(Y, t)v(t)0(Y, t)?. For NAM equations in case of Eqs. (5.4.7) and (5.4.6) at the nongaussian white noise see Problems 7.9 and 7.10. Example
7.2.1. Let us consider the system described by the stochastic
differential equation of Example 7.1.1. In this case a(y, t) = — 43> b(y, t) == i) axel
formulae (7.2.5), (7.2.7), (7.2.9) give
ew ,D,t)=i(m,D,t)=-— | y Pe 1 yoi(m, D,t)
es
,_D,t) =
;
if ify(y —
27D
V
y
Wom) 2D ty.
()I
prea: m)e
(y
2
fl eee= 1 2)
) 12D dy,
dy,
(II)
lll
676
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Taking into account
y? =m? + Im(y— m) + (y—m)’*, y? = m3 + 3m?(y — m) + 3m(y — m)? + (y—m)’, that for the normal distribution the central moments to zero, and the central fourth order moment
(1)-(1V)
(IV)
of the odd orders are equal
is equal to 3D? we obtain from Eqs.
|
~1(m, D,t) = —m(m? + 3D), yai(m, D,t) = —3D(m? + D), 22(m,
D,t) =
v(m?
+ D),
.
yo(m, D,t) = —6D(m? + D) + v(m? + D) = (v —6D)(m? + D). Consequently, Eqs. (7.2.3), (7.2.4) have in this case the form
tn = —m(m? +3D), D = (v—6D)(m? + D).
(v)
After integrating Eqs. (V) with the initial conditions m(to) = hii D(to) = D5 we determine
completely the approximate normal one-dimensional
the process Y(t).
The obtained equations may
equations of Example
distribution
of
also be derived from the first two
7.1.5 by substituting into them the expressions of the third
and the fourth moments in terms of the expectation and the variance for the normal distribution. Example
7.2.2. For the system described by the equation
where p(y) is the nonlinear characteristic of the limiter shown in Fig.7.2.1a formulae (7.2.5), (7.2.7), (7.2.9) give
Ys
@(y)
b)
Pig iecel!
0 (=I sgny
7.2. Methods of Normal Approximation
1
yi(m, D,t) =—
V 2nD 20
and Statistical Linearization
—l
-!i e~(y-m)
l
: Pays
: 12D
fyom)
J
Hf eeminras]
677
dy
J
; (p22(m, D,t) =
vy
(II)
1 -
gar(m,
'
peat
=
D,t)= —e| tf(y mye _
—l
a
I
—(y-m)?/2D
4
,
fee)
+ [uly = myer
Pay 40 f(y= men
a4
?P dy (i
i
For performing the integration we notice that the integrals of a normal density representing the probabilities of the occurrence of a normally distributed random vari-
able in the corresponding intervals are expressed in terms of the Laplace function D(z) = —1/¥V 20 gien" /2dy by the formula 0
The other integrals are reduced to the integrals p
B
p
pte’ Pau i
/ude"
/? =
(seat? _ ae~*"/?)
p
+ fwd a
=
(Be-#°/? - ae~*"/?) + V2 [6(8) — B(a)}.
(VI)
678
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Using formulae (IV)-(VI) and taking into account that @(—z)
=
—(2), (00)
= 1/2 we find
yi(m, D,t) = — [d+ mye SS a
a
mina -ole(3B)+0( 8) P
Egs. (7.2.3), (7.2.4) have in this case the form
l+m
en
6
——
} —(l-—m)®
L+ =)
(‘- =)
l—-m\
| ——
vf fon[- £58] ae [-R")}, on [c+me (SE)
D=—2D.|®\
——
| (VD
ie,
| +2 |——
VD
(ss)
] | +.
X
cS
Analogously in the case when p(y) = [sgn y (Fig.7.2.1b), we obtain
WES
lem ato okie pr(m,D), v1 Duala =216(7)
XI (X1)
m
D = —2Dk,(m,D) + v, ky(m, D) = Sissre-. (22 Kom .
2
2
ORETY)
us
Example
Eq.
7.2.3. For the system with the two-dimensional state-space by
(I) of Example 7.1.6 the matrices a(Y, t) and b(Y, t) are determined by the
formulae
a Vite sods 0 a(¥,t)=-| ; A ; wv.= [2],
(1)
and formulae (7.2.5), (7.2.7), (7.2.9) give
bG0 A Org ees ki2 palm, K,t) == ~By | : a =olee pas(m,
gaa(m,
¢
Y,
Y;
K,t) = ~Ex |is y [sm
K,t)
(II)
Yo — mz |
tn |Kaye Vay, eae CC)Aum) 2) Co 0 v0 h}= i Onera = By |)|
ae (IV)
Taking into account that for the normal distribution all the central moments of the
odd orders are equal to zero we find from (I)-(IV)
7.2. Methods of Normal Approximation
and Statistical Linearization
679
EnYi¥2Y, = Ey {mym2YP + m7 YP + moY,? + YP} = mki2+ mek,
(V)
EnY1Y2Yy = miko. + m2ki2, EnY2Y? = ki2, EnYoY¥2 = kao, eR
Kye
| moki, + myki2
meky2+ miko |
aky2
p2(m, tole t) —
ed | —2(m2ki13 + myky2) = a)kyo—
(VI)
ako
yoi(m, 1G t) = yai(m, K, te a2 22(m,
—(m2
(VI)
ist t)
~=—(me2 + a)ki2 — mi ko2 |
my koo
vh2 —
2ako2
; Thus Eqs. (7.2.3), (7.2.4) using (V)—(VIII) have in this case the form
(VIII)
my, = —m,mMz — ky2, m2 = —amz, ky, = —2(moki1 + mki2), ko2 =
vh? =
(IX)
kig = —(m2 + a)ki2 — my k22,
2ako.
(X)
0,483
0,108
0,479
0,104
0,477
0,102
0,475
0,100
0,473
0,098
0,471
0,096
0,469 0,467
0,094 0,092
0,481
0,106
0,465
0,090
0,1
0,3
0,5
0,7
0,9
Fig. 7.2.2
Pig. 7.2.3
The results of calculations of the solution curves Eqs. in Fig.7.2.2 and 7.2.3 at my,(0) =
m2(0)
=A):
k11(0)
(IX), (X) are shown se
of) be Kio s=
(0);
ko2(0) = 1. The exact distribution curve is given for comparison. So the expectation coincides with the exact one. The accuracy of NAM
3%.
for 7 = VD, D = kj; is about
680
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
7.2.38. Normal Approximation Method. In exactly the same
Multi-Dimensional Distributions
way the normal approximations of all other
multi-dimensional distributions of the process Y(t) may be found.
As
the normal distribution is completely determined by the moments of the first and the second orders, and the latter are determined by the twodimensional distribution, it is sufficient to find the normal approximation of the two-dimensional characteristic function g2(A1, A2;t1,t2). For this purpose it is sufficient to find the covariance function K(t,,t2) of the
process Y(t). > In order to obtain an equation for K(t, ; to) in the case of system (5.4.6) it is sufficient to evaluate approximately the expectation in (7.1.8) replacing the unknown two-dimensional
distribution of the process Y (t) by a normal one.
As a result we
obtain
OK (ty, t2)/Ot2 = En [Y:, — m(ti)] a(¥,, ta)?
(7.2.15)
where the subscript JV at the expectation means that it is calculated for a normal joint distribution of the random variables Yi, and Yeas For calculating this expectation we
use the formula of total expectation:
first, we evaluate the conditional expectation
relative to em and after that we evaluate the expectation of the obtained function of
the random variable Ve . Then using the known formula for the conditional expectation of the part of the components of a normally distributed random vector relative
to the remaining components we get En|Y1, = m(ty)]a(Yi,, te)” > En{En|[Yi, lY2.] rz m(t2)}a(Yr,, to)”
= K(t;,t2)K(t2)~'En[Yi,—m(t2)Ja(¥i,,t2)” -
(7.2.16)
But on the basis of (7.2.7)
En[Y¥i, — m(t2)]a(Y%r,,t2)” = Goi(m(t2), K(t2), ta)? .
(7.2.17)
Consequently, from (7.2.10) and (7.2.17) we get
En[¥:, — m(t1 ))a(¥i,,t2)” = K (tr, t2)K7"(t2)eo1(m(t2), K (te), ta)” -
(7.2.18)
Substituting expression (7.2.18) into (7.2.15) we obtain the approximate equation for the covariance function of the process Y (t):
OK (ti, t2)/Ot2 = K(t1,t2)K(t2)7 '~oi(m(t2), K(t2), te)”.
(7.2.19)
This equation with the initial condition K(ty ; ty) = K(t1) represents at any fixed t; an ordinary linear differential equation determining
of t2,t1 < tg. 4
(tsts tz) as the function
7.2. Methods of Normal Approximation and Statistical Linearization
681
Thus Eqs. (7.2.3), (7.2.4), (7.2.19) define consecutively the expectation m(t), the covariance matric K(t,t) = K(t) and the covariance function K(t,,t2) of the process Y(t) according to NAM in the normal stochastic differential system described by. After this we may determine all the multi-dimensional distributions (approximately normal) of the process Y(t) by formulae of (Subsection 2.7.5):
ieee’
0.6;
las
| An)
=
exp
{XPin
a
5X" RX}
(n =
oe
« ee
(7.2.20) m|
N= [ATAZ An] z =|
Tp
¥
> Hn = [my (ta)? my (to)? ...My(ta)7T?
Kat AE (itats
Kati)
K(ta,ta)
ot
...
C14)
K(tayta)
(host )lk (it) bod cacti
(in be)
fn(Yt yi m1 Ynj try +-+ tn) rs
1
e
rw [(27)"|K,]71/? exp {-5(a — m2) Ko (Gn — itn)} (p= (722-21) where jn = [yl yZ ...yZ]’. Example the form
7.2.4. Under the conditions of Example 7.2.1 Eq. (7.2.19) has
OK (ti, t2)/Ot2 = —3 |m?(t2) + D(t2) ]K(t1, ta).
The solution of this equation with the initial condition K(ty ; ty) =
(1) D(t1) is ex-
pressed by the formula
K(t,tz) = D(t1) exp
=3 f{m?(r) + D(x)} dr
(II)
According to this formula we may calculate the covariance function of the process Y (t) after the integration of the equations determining its expectation m(t) and
variance Die). Example
7.2.5. Under the conditions of Example 7.2.2 (Fig.7.2.1a) Eq.
(7.2.19) has the form OK (ti, ta) the [+ m(to) — m(t2) oR tte) c(Se + (SEE |Kania
iien'(D
682
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
The solution of this equation with the initial condition K(ty , to) =
D(t1) is ex-
pressed by the formula
K (ti, t2) = D(ti) exp
-
[fs(Sz) +0
(=z)
:
dt
(ul)
For the special case (Fig.7.2.1b) in formulae (XI) and (XII) of Example 7.2.1 we need to put
ki(m(t), D(t)) =1 Example
7.2.6.
m(t)?
aby
To obtain by NAM
-5o |Ae)
= hh
the equation for the covariance
function
rar iyi of the process ato) =
Kii(t1,te) Koi(ti,te)
Kie(ti, ta) | Keo(ti, te)
[Yi (t)Y> (t) |" under the conditions of Example 7.2.3 we
represent the formula obtained for $21 (m, iG t) in the form
yoi(m, K,t) te &
ac
i
Kk.
Substituting this expression at tf = ¢9 into Eq. (7.2.19) we obtain OK
(ty, t2) :
Sec
tae
M2(t2)
K (ti, ta) | Fy
0
or in scalar form
OK 41 (ti,to)/Ot2
= —me2(t2) Kir (ty, t2) — mi (te) K12(t1, ta) ,
OK y2(ti ; to) /Ot2 —- —aKyo(ti : ta) ; OK o1(t1, t2)/Ote
=
—mM2(t2)Ko1(t1,
OK 20(t1, t2)/Ot2 = t2) —
—aKo2(ti ; to) .
m4 (t2) Koo(t1, ta) ;
(I) The initial conditions for Eqs. =
(I) have the form Kai(ts, t1) =
fy 2). The second and the fourth equations are easily integrated.
kai (ti) (h,l Their solu-
tions are expressed by the formulae
Kyo(ti, te) = kio(ty)e~ 202-*1) pn
tapes ba:
Koo(t1,to) = kaa(tyew"-™),
to > th.
Substituting these expressions into the first and the third Eqs. find Kii(ti ; to) and Ko (ty ; to).
(11) (I) respectively we
7.2. Methods of Normal Approximation
and Statistical Linearization
683
~~
7.2.4. Approximate Determination of Stationary Processes NAM may also be used to determine approximately the characteristics of the strictly stationary process in the normal nonlinear stochastic differential system with the stationary normal white noise V. In this case a(y,t) = a(y), b(y,t) = b(y), v(t) = v do not depend on time t, and consequently, the functions y; and y determined by formulae (7.2.5)
and (7.2.6) are also independent oft. For finding the expectation and the covariance matrix of the value of the stationary process at any ¢ we should put in Eqs. (7.2.3), (7.2.4) m= 0, K =0. As a result we obtain the equations
yi(m, K) =0,
Ly)
y2(m, K) = 0.
(7.2.23)
If the constant vector m and the nonnegative definite matrix K exist sat-
isfying these equations, and this solution of Eqs. (7.2.3), (7.2.4) is stable according to Liapunov then we may suppose that this solution characterizes the stationary process in the system. In this case for determining the covariance function k(r) of the stationary process we should use the same method as in Subsection 7.2.3. As a result we obtain
dk(r)/dr = yai(m, K)K—'k(r).
(7.2.24)
Thus Eq. (7.2.24) with the initial condition k(0) = K determines the covariance function of the stationary process in the system att > 0. At
7 h, where Eni(-) =
i / (-)fni(y)nni(v)dydv; fri(y) and nni(v) being the
normal densities of random values Y; and Vj.
7.2.6. Normal Approzimation stonal Nonlinear Systems
Method for Stochastic Infinite-Dimen-
In the general case the exact solution of the equations of Section 5.9 for the multi-dimensional functionals g, of the normal stochastic systems is impossible. After taking the derivatives of the first and the second order with respect to both parts of Eq. (5.9.1) at nm = 1 we obtain the formulae of the time derivatives of an operator of the expectation m and the covariance operator K of the process Y;:
m = Ed(Y,t), Kd = Efa(Y,t)(A, (Y — m)) + (Y — m)(A, a(Y, 2) + 0(Y,t)A], o(Y¥,t) = 0(Y,t)v(t)0(Y, t)*.
(7232) (7.2.33)
686
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
If we calculate here the expectations replacing the unknown true one-
dimensional distribution by a normal one (3.7.00): gn (A;t) = exp | 1(A, m) — (KA,)| = exp (a, — s(A x» |:
(7.2.34)
then Eq. (7.2.32), (7.2.33) at H = Ey together with the correspondent initial conditions will represent a closed system of the ordinary differential equations in a separable H-space Y which approximately determines the operators m and Kk. Similarly after taking the mixed second strong derivative over j and Az with respect to both parts of Eq. (5.9.9) at n = 2 multiplied
by [—2(A1, mz, )— i(A1, mez,)]and evaluating the expectation for the twodimensional normal distribution we get OK (ty, ta) OL
v2 =
En(%1,
= m+, )(A2, a(Yt, _ m+, )) Shon
Te,
(7.2.35)
2
for the covariance operator K (#1, t2)A2
=
En ((Y1, —
M+, ); A2(Y,
—
Mt, )) p
(7.2.36)
Eq. (7.2.35) at the fixed t; is an ordinary differential equation in a H-space ) for the operator K(t;,t2) as the function tz at the initial condition
K(t1,t1) = K(t1).
(7.2.37)
For approximate finding the stationary process in (5.4.10) at the stationary white noise V = W we may also use the normal approximation
method. Accounting that in this case a(Y,t) = a(Y), b(y,t) = b(Y), v(t) = v, o(¥,t) = o(¥) do not depend on time and putting in Eqs. (7.2.32) and (7.2.33) m =0 and K =0, E = En we shall have
Eya(Y) =0, (7.2.38) En [a(¥)(d, (Y — m)) + (Y — m)(A, a(Y)) + o(¥)A] = 0.(7.2.39) The obtained results may be summarized as follows. Theorem
7.2.3.
If the random process Y; in the normal stochastic
differential system defined by Eq. (5.4.9) in a separable H-space has the bounded the first order operator moment m and the trace-type covariance operator K then its one-dimensional characteristic functional may be
7.2. Methods of Normal Approximation
approximated
and Statistical Linearization
by normal ones (7.2.34) with m
and K satisfing
(712583). (7.2.33). Corollary
7.2.3.
687
Eqs.
If the stationary (in the strict sense) process Y;
in stattonary normal stochastic differential system (5.4.9) has the tracetype covariance operator then its approximate characteristic functional ts given by (7.2.34) with m and K defined by Eqs . (7.2.38), (7.2.39). Example
7.2.9.
Under the conditions of Example
5.9.1 Eqs.
(7.2.32),
(7.2.33) for m; (x,t) and Ky; = Ki; (21, £2, 2), 1,j = 1,2 give
dm,
—_—_
=m
at
OK Ot
dm,
40°m,
a
LT
Rar?
eee
+ 2him2
OK
Ki
OK
2 >
ee Ot
=
zhama(ms + 3K 22) 4
0? K
Koo +c?
Ox?
(1)
J
= + h(x2,t)Ki2,
07K =
=
K22 ait ce? a
se
aK
att
rere (pa,i a's 4
rah
= t 2h(a1,t)
ie
av
adel3 4 sa
0*Ki4
“s 07m,
dy
ag, lonoies ae
OzronennsOn?
where h(a, t) =h,—h2
Koi )
—
ansdx2
he
[m3 (a, t) ap Ko2(a, ae t) 1,
7.2.7. Methods of Equivalent Linearization and Structural Methods
Let f(u,y) be the joint density of the random vectors U and Y, my and Ky be the expectation and the covariance matrix of the random
vector Y, and |K,| # 0. From the known equations of the m.s. linear regression (Section 3.4): CT
IT
or
gly = Puy
rm
(7.2.40)
(7.2.41)
we get CO
Ee Kage
8000
= / fo —mu)(y — iy) Ky f(t, y)dudy — 00 —0o
688
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
= f (maa) = ma} y= my)" Ky Aaladdy,
(7.2.42)
where f,(y) is the density of the random vector Y. This formula together with the approximate formula my,(Y) & my, + g(Y — my)
(7.2.43)
yields the Kazakov statistical linearization of the regression m,(Y ) (Subsection 7.2.2). Analogously, in case-of (7.2.41) we have
—
Lat.
s
uy™ TZ}f(u, y)dudy = i mu (y)yT Py fi(y)dy.
a
cca
—oo
(7.2.44)
Formula (7.2.44) together with the approximate formula
my(Y) + gY
(7.2.45)
yields the Booton statistical linearization of the regression m,(Y). The regression m,(y) represents deterministic regression model for U = ¢(Y). To obtain a stochastic regression model it is sufficient to
represent U = U(Y) in the form U = my(y) + U’ where U’ being some random variable. For finding a deterministic linear stochastic model it is sufficient to know the expectations m,, my and the covariance matrices Ky, Kuy. But for finding a stochastic linear model it is also necessary to know the distribution of the variable U for any y or at least expectation
mux(y) (the regression) and the covariance matrix Ky(y) (coinciding with the covariance matrix Ky/(y) of the random variable U’). A more general problem of the best approximation of the regression by the finite linear combination of given functions y;(y), ..., gn (y) is reduced to the problem of the best approximation of the regression, as any linear combination of the functions yi(y), ... , gyw(y) represents a linear function of the variables u; = yi(y),..., un = yn(y). Example case of scalar
7.2.10. Design the optimal polynomial regression model in the
Y and U. Putting
U1
=
Y;
io
Et
bien sathNee als
u=([luju...uy]’,
g=[agn...gn]
(I)
7.2. Methods of Normal Appreximation
and Statistical Linearization
689
we reduce the problem to the finding model
u(y) = g(u).
(II)
Eq. (7.2.41) in this case takes the form
[agi...gn] =
1 a + QN
where @p =
Mie” a 3
Bo SCR ra Att
AN41
rile the pth moment
moment of [Y U]”
..-
={ muyrsenyn]
(111)
QIN
of Y, and Yp =
EUY?P
is the (p iP re
containing U in the first degree. Eqs. (II), (III) determine the
polynomial m.s. regression of U on Y of the degree N which serves as the optimal polynomial regression model.
The equations of the composed nonlinear stochastic systems after the equivalent linearization of the nonlinearities entering into the equations may be studied by the methods of Chapter 6 and first of all by the structural ones. Hence usually for the description of the composed stochastic systems the different describing functions are taken, for instance, the equivalent weighting functions g = g(t,7;m,K) and the transfer function ®(iw, m, K). The implementation of the spectral and the correlation methods
(Subsection 6.1.8) for Eq. (5.4.6) assumed as symmetrized Stratonovich equation allows to substitute the joint solution of Eqs. (7.2.22)—(7.2.24) by the solution of Eq. (7.2.22) and the corresponding the equations for the matrix sy of the spectral densities and the covariance matrix K: Sy = Sy(w;m, K) = ®(tw;m, K)sy(w)®(tw;m, K)*,
(7.2.46)
Ee ae J slwim, Kydes
(7.2.47)
—0o
Here s,(w) = v/27 is the spectral density of the stationary white noise V; ®(iw;m, Kk) is the equivalent transfer function of the statistically
linearized system for Y°,
Ye oem KY.
y =Y —m,
s=d/dt,
(7.2.48)
equal to
®(s;m, K) = —[a(m, K) — sI]~' b(m, K),
(7.2.49)
690
Chapter 7 Methods
where
@(m,K)
=
of Nonlinear Stochastic Systems Theory
Eya(Y,t)Y°T
=
oi(m,K)K7};
b(m,K)
= Ey0(Y,t).
7.2.8. General Problem of Stochastic Systems Normalization In the class of the stochastic systems there are the simplest systems which permit comparatively simple mathematical description and at the connections form the systems of the same type. These are the normal systems. A normal system we shall call such a stochastic system whose the joint distribution of the input and the output is normal at the normal distribution of the input. It goes without saying that at the distribution of the input different from the normal one the joint distribution of the input and the output may be nonnormal even a normal system. The opportunity of applying the well developed methods of the linear systems theory (Chapter 6) to the normal systems and the relative simplicity of the mathematical description of the normal system cause the natural quest for extending the methods of the research of the normal systems as the approximate ones over sufficiently the large class of the stochastic systems different from normal. Thus the problem of the normalization of the stochastic nonlinear systems arises, i.e. the problem of finding a suitable normal model for a given nonlinear stochastic system.
Let us prove the following theorem. Theorem 7.2.4. For the normalization of the given stochastic nonlinear system it is sufficient to linearize the regression of the output on the input and to average over all possible realizations of the input the conditional covariance operator of the output at the given input. > In order to linearize the regression we may try to approximate it by the linear dependence on the criterion of the minimum of the mean square error (m.s.e.) at any
instant (to use the method of least squares).
Then the problem will be reduced to
the definition of the linear operator I, and the variable a from the condition
Bla+ LX — E[Y|X]|? = min,
(7.2.50)
where / [Y |x | is a conditional expectation of the output Y at the given realization & of the input
X (Section 3.4). We apply to the random variable
Z =
a+ LX —Y
the known relation:
E\Z)? = E\E[Z|X]|? + ED[ Z|X],
(7.2.51)
7.3. Methods
of General Nenlinear Stochastic Systems Theory
which is valid for any random variables
691
X , Z where Di Z |x |is a conditional variance
of the variable Z at a given realization & of the variable X. As a result we obtain
from (7.2.50)
Bla+ LX -Y? =Bla+LX —ELY|X]/?+ED[Y|X].
(7.2.52)
Hence it is clear that the problem of the optimal linear approximation of the regression of the random variablé Y on _X is equivalent to the problem of the optimal linear
approximation of the random variable Y itself on the m.s.e. criterion.
While deriving the equations for J, we calculate the stochastic differential
of the product [Yi (t) —mM, (t)]" Fa hYS(#) — Mp (t) pe by means of It6 formula. But at first we notice that formulae (7.1.1) and (7.1.5) at approximation (7.4.1) may be written in the following form:
i=
$1,0(m, K,t) + se SS ~i,v(m, K,t)qv(@)
;
(7.4.6)
l=. |v
N
K= ~2,0(m, K,t) + os 3
~2,v(m, K,t)q,(a@),
(7.4.7)
[=8F 1
where
(7.4.8)
yiy(m, K,t) = fistartunGhostoas,
pay(m, K,t) = /{a(y,t)(y7 — m?) + (y— m)a(y,t)? + o(y,t)}wi(y)pr(y)dy,
t)”. o(y,t) = b(y,t)v(t)b(y,
(7.4.9)
(7.4.10)
Here the coefficients ~1,0(m, Kk, t) and 2,0(m, 1¢ t) are determined by formulae (7.4.8) and (7.4.9) at py = po(y) = 1 with the change v by 0.
698
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Evaluating directly by formula (5.1.7) the stochastic differential of the product [Yi(t) - m,(t)]" ane [Y,(t) — Mp (t)|"? we get with the account of (7.4.8) and
(7.4.9) N
jin = Gr 0(m, Kit) +) D| Pro(m, K,t)av(a) i=S
Pp
SISO fee
|v| =!
ACT
Sh te SEN
(7.4.11)
s=1
where the functions py (m, KB, 0) are determined according to (7.4.4) by means of the change of the products Yj ...Yp by (y1 — m1) aoa (Yp _ Mp), and €, is a vector whose all the components are equal to () besides the s*
component equal to 1.
YS) ga(m,K, the, 9.
(7.5.9)
i—sa yi
Now let derive the equations for the coefficients c, in (7.5.1). >: We use formula (3.8.7) changing index VY on K:
c= E (=) acsd|
(7.5.10)
Differentiating this formula with respect to time t and taking into consideration that the polynomial qx (y) depends on the expectation M and the covariance matrix K of the random vector Y; which are the functions of time we find
x= fu) 2B} (BY + tr {|af (x) nt]
A=0
'
so] (7.5.11)
where qn’ (y) is the matrix-column of the derivatives of the polynomial qx (y) with respect to the components
of the vector m,
and a
is the square matrix of the
derivatives of the polynomial dx (y) with respect to the elements of the matrix K.
7.5. Methods based on Orthogonal Expansions and Quasimoments
721
Substituting into this formula the expression 0g;(A;t)/0t from Eq.
(5.6.20) and
replacing the density f (y; t) by its truncated orthogonal expansion (7.5.1) we obtain
én = |{ooea [iAP ay, t) + x(b(y,t)" Ast)] ors
A=0
N
x}
1+ x > cypy(y) | wi(y)dy+q i=3
(a)m-+ tr [ak (a) K] 7 UGo.12)
|v |=!
where x (us; t) for the normal white noise is determined by (5.6.37), qe’ (a) and qk (a) represent as usual the resultOf replacing the monomials v4 whee Yp? in the expressions of the polynomials qn (y) and ons (y) by the corresponding moments Qp, ,
.-,@r,. Putting
gxo(m, K,t) = [{ a (aa) x [id? a(y, t) + x(b(y, t)? A;pera
wi(y)dy,
(7.5.13) nv (dads, ts
is dn eaea [iA E (ax)
a(y, t)
+ xbnOT rider) —po(adan(adey
(7.5.14)
and substituting the expressions of m and K from (7.5.8), (7.5.9) we obtain the equations
Ce = Gno(m, K,t) + yio(m, K,t)? g(a) + tr [po0(m, K, t)g* (a)| N
+> So {eev(m, K,t)+e10(m, K,t) ae(a)+ tr [par (m, K, tae (a)] i=se|pl=t
N
FP l=3
i2
8.6)
ol
(7.5.15)
|v) =I
Here the moments @,, ,...,
@p, in qn’ (a) and qi (a) must be replaced by their
expressions in terms of the coefficients Cy in accordance with (3.8.14).
ty and Eq. (7.5.24) with the initial condition K(ti,t1) = K(ti) we determine approzimately the covariance function K(t1,t2) of the process Y(t) and all the coefficients c,,,, of truncated expansion (7.5.23) of the two-dimensional density fo(yi, y2;t1,t2) which remained unknown. Further, subsequently integrating Eq. (7.5.30) with initial conditions (7.5.33) att) N at least n—N subscripts are equal to zero for any of the coefficients cy,,..,v,
tm (7.5.23).
As a result all the multi-dimensional
distributions of the process Y(t) will be approximately determined. 7.6.4. Consistent Hermite Polynomial Expansions
In the same way as in Subsection 7.5.2 it is proved that while using the consistent Hermite polynomial expansions the elements of the
7.5. Methods based on Orthggonal Expansions and Quasimoments
matrices qf)
729
x, (@), qe _.,&,(&) are proportional to the correspond-
ing quasimoments ne oben
ho) Sher otk
eee her
(hela tesailerl pare reba Gah gamers)
(Py 8
Anes
= mthrins
SA ioc, AN).
ong. pep Sey, -
ee ail oe og MOdey
(7.5.34)
} Kn —Cs
(7.5.35)
tse ects [Kyl atO eas. ,LV),
1 oe ec)
== Ta kine(Kner -
io ag er)
Dera gta
hae
er )
(7.5.36)
= —KnrknsCry,...,%n—er—es
(rv, $ el 5.32 Dee
(7.5.37)
THlkifehe+ leelieont AGN),
|i] +--+ [Kn] =3; Oat [ra] ee Ger... ee(%) ==Oat + +--+ [en] =3and |r| +--++ [en] =4; B
(7.5.38)
In the case of the normal white noise V in Eq. (5.4.6) we have
{Mee Gastric oraz aa) #x(b(un sta) Anitndle™e } ONO)
=
(Gs, Yt
—
Mt,
+++)
Yn-1
—
mz,
9/40An)[irp
(Yn, tn)
+x(btn)” (tAnitn) n } EXPLIAR(Yn— Mn)}} 4, to), c20(ty ; to),
c13(ty 3 to) of the two-dimensional distribution of the process Y are determined by the equations: OK (ty, t2)/Ote =
—3[m?(t2) = D(t2)|K (t1, ta)
—3m/(t2)c12(t1, t2) — c1a(t1, te),
(1)
ci (ti, t2)/Ot2 = —3m(t2) K?(t1, te) — 3[m?(t2) + D(t2)]c21(t1, te) —6K (ty, t2)c12(t1, t2)— 3m(t2)c22(th, ta),
(11)
Oci2(t1, t2)/Ote = 2[vm(te) — 8m(t2) D(t2) — 6c3(t2)| K (t1, te)
+[v — 6m?(t2) — 12D(t2)]e12(t1, t2) — 6m(t2)c13(t1, ta),
(III)
desi (t1,t2)/Ot2 = —6K3(ty,te)— 18m(t2) K(t1, t2)ca1(t1, t2)
—3[m?(t2) + D(t2)]es1(t1,t2)— 9K (t1, te) c20(t1,t2)+ ¢3(t1)ea(t2), (IV) Oc29(t1, t2)/Ote = 2[v — 6D(te)| K7(t1, t2) + 2[vm(t2)
~8m(t2) D(t2) + ¢3(t2)]ca1(t1,t2)— 24m(t2) K(t1, t2)c12(t1, t2) +[v—6m?(t2)—12D(t2)]c22(t1, t2) -12K (t1, t2)c1a(t1, t2) +6 D(t1)D?(t2), Vv 0c13(ti } to) /Ot2 =
3{ D(t2)[2v+3D(t2)]
+3[2vm(t2)
—6m(t2)c3(t2)
=F 12m(t2)D(t2)
+3[v =
3m?(t2) =
—3c4(t2)}
+ c3(t2)}c12(t1,
9 D(t2)}e13(t1,
K (ty Oe
ta)
t2)
(VI)
with the initial conditions
K(ty,t2)
= Di);
c3i(ty, ty) =
coi(41, 41) = cia(t1, t1) = ca(ty),
€29(t1, t1) =
€13(ty, t1) =
c4(ty).
(VI)
7.5. Methods based on Orthogonal Expansions and Quasimoments
The
quasimoments
¢111(t1,¢2,¢3),
C211(ti,t2,t3),
(au
c121(t1, ta,ts),
€112(t1, ta, t3) of the three-dimensional distribution are determined by the equations
e111(t1, to, t3)/Otz = —3[m?(tz) + D(ts)]e111(t1, ta, ts) —3m(t3)c112(t1, te, t3) — 6m(t3)K (t1, ts) K (ta, ts) —3K (ty, t3)e12(ta, ts) — 3K(to, t3)c12(t1, ts),
(VIII)
c211(t1, te, t3)/Ot3 = —12m(ts) K(t1, t3)c111(t1, ta, ts) —3[m?(t3) + D(ts)]e211(t1, te,t3) — 6K (t1, t3)¢119(t1, ta, ts) —6K?(t1,t3)K(ta, ts) + c3(t3)co1(t1, ta) —6m(t3) K (te, t3)co1(t1, t3)— 3.4K (te, t3)c22(t1, ta),
(IX)
0cj21(t1, te, t3)/Otz = —12m/(tz)K (ta, t3)e111(t1, ta, ts)
—3[m?(ts) + D(ts)]c121(t1, te,ts) — 6K (t1, t3)¢112(t1, ta, ts)
—6K (ti, t3)K?(te, t3) — 6m(t3) K(t1, ts)c21 (to, ts) +c3(t3)c12(ti,ta) — 3K (t1,t3)c22(te, ts),
(X)
0c112(t1, te, t3)/Ot3 = 2[vm(t3) — 6m(tz) D(tz) + c3(tz)]e111(t1, te, ts)
+[v — 6m?(t3) — 12D(ts)]e112(t1, te, t3)
+2[v — 6 D(ts)] K(t1, t3)K (to,ts) — 12m/(t3)[K (to,t3)e12(t1, ts) ic (pata ote(torte iene| We(egrets tite ick K(taste cis ta,ta ex) with
the
initial
conditions:
= c22(t1,t2); C121 (t1,
e111 (ty yo7ts)
b=
Cro bas to);
co11(ti UD t3)
t2,t3) = C112(ti, te, t3) = c13(t1, tz). The only quasi-
moment of the four-dimensional distribution C11 11(t1 SUD UEY t4) depending on four
variables t;, tg, t3, t4 is determined by the equation
0¢1111(t1, tz, t3, t4)/Ota = —3[m? (ta) + D(ta)]er111(t1, ta,ts, ta)
—6K (ti, t4)K (ta, ta) K (ts, ta) + e3(ta)er11(t1, ta, ts) —6m(t4)[K (ts, ta)c111(t1, to, ta)+ K (to,ta)er11(t1, ts, ta) +K (ty, t4)ce111(to, ts, t4)] — 3[K (ta, ta)e112(t1, te, ta) +K (te, ta)ce112(ti, ts,ta) + K (tr, ta)c112(t2, ts, ta)] with the initial condition e1111(t1 , to, ts, t4) = €112(t1 UB ts).
(XII)
Coe
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
7.5.8. Approximate Determination
of Stationary Processes
For the approximate determination of the one-dimensional distribution of the strictly stationary process in stochastic stationary nonlinear differential system (5.4.6) by the methods based on the orthogonal ex-
pansions one should put in Eqs. (7.5.8), (7.5.9), (7.5.15)
m=0,
K=0,
& =0.
(7.5.40)
If the equations obtained in such a way have the solution which may serve as the parameters of the corresponding truncated orthogonal expansion of the one-dimensional distribution, then we may suppose that the strictly stationary process exists in a system. In this case for determining other multi-dimensional distributions of this stationary process one should replace the derivatives with respect to t, in Eqs. (7.5.24), (7.5.30) by the derivatives with respect to T,-1 = tn —t, and write the
initial conditions (7.5.33) in the form Croeana (11 yrrey
Tn-2) Treo) Goi
tee 5 eal
(1 Tecanape h) Tn—2) .
(7.5.41) We summarize the results obtained in Subsections 7.5.1—7.5.5 as the following statements. Theorem 7.5.1. If the random process Y(t) in normal stochastic differential system (5.4.6) has the finite coefficients of the consistent orthogonal expansions then its one- and multi-dimensional distributions may be approximated by expansions (7.5.1), (7.5.23) with m(t), K(t), Cr(t)s Fe(tito) d Gay Alone (E> “tut, En) defined by Eqs. (£9578) labo 9),
(7.5.15), (7.5.24), (7.5.30) at initial conditions (7.5.16), (7.5.33).
Corollary 7.5.1. If the stationary (in the strict sense) process in stationary normal stochastic differential system (5.4.6) exists and has the finite coefficients of the consistent orthogonal expansions then tts one- and multi-dimensional distributions may be approximately expressed by (7.5.1), (70:23) nth im “hep PACr) sepa Gee, Ta; Tan) defined by Eqs. (7.5.40), (7.5.41).
The equations of the orthogonal expansions method for Eqs. (5.4.6) at the nongaussian white noise and Eqs. (5.4.7) are given in Problems 7.34-7.35.
7.5. Methods based on Orthogonal Expansions and Quasimoments
733
7.5.6. Reducing Number of Equations
The main difficulty of practical use of the stated methods especially for the multi-dimensional systems is a fast increase the number of the equations for the coefficients of the segments of the orthogonal expansions with the increase of the dimension p of the state vector y and the maximal order N of coefficients used. The dependence of the equations number for the moments cited in Tables 7.3.1 and 7.3.2 remains valid for the coefficients of orthogonal expansions, too.
While comparing the ciphers v, [v7 .. vty in Tables 7.3.1 and 7.3.2 we see that the number of the equations of the orthogonal expansion may be considerably reduced by means of the approximations of the distributions which include the mixed coefficients of the second order only. Such an approximation of the distribution may be obtained by different ways. For instance, we may assume that all mixed coefficients of the orthogonal expansions of higher than the second order are equal to zero. Another way consists of the use of the known expressions of the mixed central moments in terms of the variances and the covariances of the components of the random vector for the normal distributions (7.4.35). Both these approaches give the opportunity to use any approximations of the distributions and the correspondent equations for the coefficients of the orthogonal expansions on which the approximating function f(y;0) or f*(y1, .-- , Yn} Un) depends in Eqs. (5.4.6). Thus while using the quasimoments method it is natural to neglect all the terms of the expansions containing the coefficients c,, cy, ,...,y, with
more than one component of the multi-indexes v, [v7 ...v7]? different from zero. The approximation of a distribution (Subsection 7.4.4) may be obtained in various ways. For instance, one may assume that all mixed semiinvariants of orders higher than the second are equal to zero. Then
one obtains the recursive formula (7.4.37) determining the mixed moments of the third, the fourth and higher orders as the function of the moments of the components of the random vector and the mixed second order moments (Kashkarova and Shin 1986). Another way is to use the known expressions for the mixed central moments in terms of the variances and the covariances of the components of the random vector for
the normal distributions Satna
[Pea Unescenstei
Pry ,...,Ts
9mm!
:
_ ) ie Paka
a ()} ifri+---+r,
ga
=2m+1
if ry
+--+;
(iP
i
Adis,
(7.5.42)
734
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Both these approaches allow one to use any of the distributions of Sections 7.4 and 7.5 and to write the corresponding equations for moments, the semiinvariants and the coefficients of orthogonal expansions on which the approximating density depends. If the method of orthogonal expansions (in particular the method of quasimoments) is used, then it is natural to neglect all the terms of the expansions containing coefficients Cy, Cy,,...,v,
v,[v?...v?P]°
n
With more than one component of the multi-indexes
C2, different from zero.
Example
7.5.4. For the system of Examples 7.4.1 and 7.4.3 the method
of this subsection using formula (7.5.42) gives
H12(t1,t2) = ar(ti,t2)
,
=0, pis(ti, te) = 3D(t2)K (ti, te),
J22(t1,te) = D(t1)D(t2) + 2K* (tr,
te),
(I)
Hai(t1, t2) = 3D(t1) K(t1, ta) ,
Prii(t1, t2,t3) = 0, Hi1a(t1, te, t3) =
D(t3)K (t1, ta) + 2K (t1,t3) K (te, t3) ;
}121(t1, ta, ts) = D(tz)K (ti, ts) i 2K (ti, tz) K (ta, ts) F
(i)
faii(t1, ta, t3) = D(ty) K (ta, t3) + 2K (ti, to) K (ti : t3) .
#1111(#1, ta, t3, ta) = K(ti, to) K(ts, ta) + K(t1, t3)K (te, ta) + K (ti, ta) K (te, t3) :
(III)
Substituting the obtained expressions of Hio(ti ‘ ta) and His(ti ; to) into the equation of Example 7.4.3 for the covariance function we get
OK (t1, t2)/Ot2 = —3 [m?(t2) + D(tz)] K(t1,t2), te > th.
(IV)
Integrating this approximate equation with the initial condition K (ti ‘ to) = D(t1) we find all the mixed fourth order moments using the above formulae.
Then all the
multi-dimensional distributions of the process Y (t) will be determined in the same way as in Examples 7.4.1 and 7.4.3.
The same equation for the covariance function yields the method of quasimo-
ments if in formulae (I) and (II) of Example 7.4.3 we put c, = 0 if more than one component of the multi-index VY is different from zero.
Example
7.5.5. For the system of Example 7.4.2 formulae (7.4.37) give
M12 = far = 0, pis = 3kigke2,
poe = kiike2 + 2k?,, p31 = 3kirk12,
(I)
7.5. Methods based on Orthogonal Expansions and Quasimoments
139
and the equations of Example 7.4.2 for the moments of the one-dimensional distribution are reduced to:
my
=
—™M 1M
—
ky,
my
=
—amyg,
(II)
ki = —2(m2k11 + myki2), kia = —(m2 + a)ki2 — miko. ,
ko = —2ako9 + h2v, 30
(II)
= —3(M2f30 + 2ki1k12), fos = —3apX03,
[t40 = —4(3k1 24430 + Moplag + 8myki1k12), fos = —4apiog + 6h*vk 29.
(IV)
Equalizing to zero the coefficients C12, C21, C13, C22, C31 in the method of quasimoments
yields the above equations for ™ 1, M2,
k11, k12, k22, but the equations
for {403, 430, [40 and /94 are replaced by
€30 = —3(2k11k12 + meoczo), ¢o3 = —3acoz , C49 = —4(3k12¢30 + M2c40),
Coa = —4aco4.
(V)
As for the multi-dimensional distributions, all the approximations of this subsection yield the equations
OK 11(t1, t2)/Ot2 OK 12(t1,te)/Ot2 OK (tr, tz) /Ote OK 29(t1, t2)/Ot2
—mo(t2)Ky1(t1,t2) — my (t2)K12(t1, ta) , —akj(t1, ta),
—m2(t2)Ko1(t1, t2) — mi (te) Kea(ti, te) ,
(VI)
—aKo9(t1, te) , py > Ula
for the covariance function of the process Y(t) =
[Y4 (t1)Y (te ie
Integrating
at first the equations for the moments or quasimoments of the one-dimensional distribution, and then the above equations for the covariance functions with the ini-
tial conditions:
Kyi (ti k t;) =
Koo(ti 4 t,) =
koo(t1) we may find all the multi-dimensional distributions of the
kii(t1), Ky2(t1, t1) =
Kai (ti < ty) =
ky2(t1),
process Y (t) = [Y (t1)Y (te) 1, using the corresponding fragments of the Hermite polynomial expansions or of the Edgeworth series.
Examples 7.5.4 and 7.5.5 illustrate essential simplifications which give the approximations of distributions considered in this subsection. In the problem of Example 7.4.2 the number of equations for the complete set of moments of the first four orders amounts to 115 instead of 13 obtained in Example 7.5.5.
736
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
7.5.7. Methods based on Orthogonal Expansions for Discrete Nonlinear Systems
As applied to stochastic discrete nonlinear systems (5.7.1) we obtain the following equations of the orthogonal expansions method: film
a
Yn;
0)
=
w(y1 prety
N
x
Yn y™, K)
.
¢14+ Se
SD
el Hieale ae eeen F , doespe
osSad Mock)
h=3 |r| +--+ |Yal =h
0 oh
=
Eq
(wi(%, Vi) =
:
E (55) E exp {iP wi(Mi, v)} | A=0
cli rlntl SRG ee
)
(7.5.44)
liebe {1,...,vn +being the biorthogonal polynomials with (oe)
fi — 00
Kite
co
fei ees tal
RP
cose
ee
ta)
—oco
uh
BA
dui
rg adyua
Obs eee rong
(7.5.48)
CCK Mead LZ ET ht Rb er aaa 50Pearl2ponsSer ae ly mh Wales ors Paht sade deg) Mery mbes Lyetlen| =r id Ha Tckielig l= 1, yw.
Thus Eqs. (7.2.29), (7.2.30), (7.5.44), (7.2.31), (7.5.45) atn = 2 and Eq.
(7.5.45) at n > 2 determine recurrently all the multi-dimensional
distributions {Y;} in stochastic discrete nonlinear system
(5.7.1).
7.6. Structural Parametrization of Distributions.
EAM
Wot
For the continuous-discrete stochastic systems the corresponding equations are given in (Pugachev and Sinitsyn 2000) 7.5.9.
Other Methods of Distributions Parametrization
The distributions parametrization methods stated above are based on the consistent biorthogonal expansions of the densities and lead to an infinite system ofthe nonlinear ordinary differential equations for the coefficients of the expansion. For obtaining the approximate solutions this system is substituted by a finite one. The estimating of the influence of the rejected terms in the general case encounters with difficulties. The
Fourier series method (Moschuk 1994) is based only on the orthogonal expansions of the densities and is reduced to the solution of an infinite system of the linear ordinary differential equations for the coefficients of the expansion. Here “the linearity” of the reduced system allows to estimate the accuracy of the Fourier series method. The main advantage of a given method over the known ones is in the availability of the accuracy estimates and in the possibility to construct analytical expressions for the coefficients of the expansion of an one-dimensional density and transfer densities. The corresponding algorithms are simpler than in the moments methods, the semiinvariants and quasimoments methods. The method is easily generalized on the stochastic discrete and continuousdiscrete nonlinear systems and on the stochastic composite systems as well. The method of the orthogonal expansions may be successfully used in those cases when the additional restrictions over the structure of a
distribution are known (Section 7.6). 7.6.
Structural Parametrization
of Distributions.
Ellipsoidal Approximation Method 7.6.1. Introductory Remarks
For structural parametrization and/or approximation of the probability densities of the random vectors we shall apply the ellipsoidal densities, i.e. the densities whose planes of the levels of equal probability are similar concentric ellipsoids (the ellipses for two-dimensional vectors, the ellipsoids for three-dimensional vectors, the hyperellipsoids for the vectors of the dimension more than three). In particular, a normal distribution in any finite-dimensional space has an ellipsoidal structure. The distinctive characteristics of such distributions consists in the fact
738
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
that their probability densities are the functions of positively determined quadratic form u = u(y) = (y? —m™ )C(y—m) where m is an expectation of the random vector Y, C' is some positively determined matrix. > Let us find ellipsoidal approximation
(EA) for the density of 7-dimensional
random vector by means of the truncated expansion based on biorthogonal polynomials {pr,v(u(y)), dr,v(u(y))}, depending only on the quadratic form U =
u(y)
for which some probability density of the ellipsoidal structure w(u(y)) serves as the
weight: co
/ w(u(y))Prr(u(y))aru(u(y))dy =6.
(7.6.1)
—oco
The indexes VY and [U at the polynomials mean their degrees relatively to the variable u. The concrete form and the properties of the polynomials are determined further.
But without the loss of generality we may Then the probability density of the vector
assume
that qr,o(u)
=
Pr,o(u)
=
all.
Y may be approximately presented by the
expression of the form N
f(y) © f*(u) = w(u) > cry Pry (u).
(7.6.2)
yv=0
For determining the coefficients Cry in (7.6.2) we multiply term-wise equality (7.6.2) by Yr,yu(u) and integrate over the whole space 2”. Then by virtue of condition
(7.6.1) for pp = 1,..., N we obtain CO
[oe)
N
/ NPM CaLTES ihw(u) )enyPrv(t)Gryul)dy = Cry. Thus the coefficients C, ,,are determined by formula
loo)
cov = f Ho)ten(wldy = Baew(U), VEIN).
(7638)
—oo
As Pr,o(t) and qr,0(U) are reciprocal constants then always C, 0Pr9
(the polynomials
of zero degree)
= 1 and we come to the result: N
fly) & f*(u) = w(u) 1+ So Cry Prv(u) v=2
4
(7.6.4)
7.6. Structural Parametrization of Distributions.
EAM
739
Formula (7.6.4) expresses the essence of the ellipsoidal approximatron of the probability density of the random vector Y. > For constructing the system of the polynomials
{pr, v(u u), qr, u(u) } satisfy-
ing the condition of the biorthonormality relatively to the quadratic form u ==(y
= m)?C(y = m) we transform condition (7.6.1). The change of variables =m
te Ca 1/2¢ gives U = (To==
© =
y
I¢|? and
|e0oredanalwdy =lor? fw lCP ypeo(lCPannICP4G a
(7.6.5) aiVJu,..
CaSsia
Hence “s take
We make the change of variables
(; =
into account that C Spee
(To = uU and achemrecem nb ay S00
sSe C=
ae Cea
After the transformations (7.6.5) is reduced to the following form:
1/2_r/2
= a
P
/Wee Dian (14) lash alee aes) 0
For the applications the case when
the normal
distribution is chosen as the
distribution w(u) is of great importance
w(u) = w(z7? Cz) = accounting that C' = K—!
1
exp(—z? K~12/2);
Jen ikl
(7.6.7)
we reduce the condition of the biorthonormality (7.6.6)
to the form
1 SR
Spy
i |Pov (e)arald
r/2-1,—u/2 é
thU SnOe ft
(7.6.8 )
0
Thus the problem of the choosing of the polynomials system {p,,,(u), dr,u(u)} which ts used at the ellipsoidal approrimation of the densities
(7.6.4) and (7.6.3) ts reduced to finding a biorthonormal system of the polynomials for which the x?-distribution with r degrees of the freedom serves as the weight.
740
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Taking into consideration that y?-distribution represents a special
case of the gamma-distribution (Appendices 2, 7) koi
w(u) = eae
one
(7.6.9)
we consider a system of the polynomials S?(u), d’
By (uy = (Fi ean te
(ut? Cay
NG > Alar
OY
1076.10)
At a=r/2—1 and k = 1/2 we obtain a system of the polynomials which are orthogonal relatively to y?-distribution with r degrees of the freedom:
(r+ 2v — 2)!! see oe ) = Seta)hcae Vv
A
Sie
(pe
|)
T—0)
The main properties of the polynomials S?(u) and S,,(u) are given in Appendix 2. Between the polynomials S,,(u) and the system of the
polynomials {p,,,(u), dr,.(u)} the following relations exist: Pr,v(u) = Sey (th),
(r— 2)!
Gry(u) = Ga Qv aDann)»
(7.6.12)
r>2.
(7.6.13)
Now we consider the expansion of the densities of nonnegative ran-
dom variables with respect to the polynomials S%(u) and S,,(u). Let L2([0,00)) be a H-space of the functions integrable with the modulus square. After test we make sure that the functions
Oya) = V/s),(4) = if eae
Seu)
(7.6.14)
form the orthogonal basis in L2([0,00)). Therefore any function F(u) € L2([0,0o)) may be presented as m.s. convergent (in the norm of the space) expansion F(u) = Vw(u) rc, s, (a),
(7.6.15)
7.6. Structural Parametrization
of Distributions.
EAM
741
whose coefficients are determined by the relation
Rac itF(u)S2(u)du /Sea
(7.6.16)
In particular any probability density y(u) of the nonnegative ran-
dom variable satisfying the condition {o(u)/Vo(u)} € L2([0, co)), ss2*(u) cat) co
0
2
oo
du
u
=
T(o+1)
a,u/2,.2 / era (u)du < ov, 0
(7.6.17)
may be presented by the expansion
p(u) = w(u) >> cy Se(u).
(7.6.18)
v=0
Now we consider the H-space L2(R") and the orthogonal system of the functions in them where the polynomials S,,(u) are given by formula
(7.6.11), and w(w) is a normal distribution of the r-dimensional random vector (7.6.7). It is evident that this system is not complete in L2(R"). But the expansion of the probability density f(u) = f((y? — m?)C(y —m)) of the random vector Y which has an ellipsoidal structure over the polynomials p,,(u) = S,,,(u),
f(u) = w(u) pe Cru Prinl o),
(7.6.19)
v=0
m.s. converges to the function f(u) itself. The coefficients of the expansion in this case are determined by relation (7.6.16) which by virtue of
(7.6.12) and (7.6.13) will take the form
cow =f flu)pen(udy /PPEOE
(7.6.20)
Thus the system of the functions {,/w(u)S,,(u)} forms the basis in the subspace of the space Lo(R") generated by the functions f(u) of
the quadratic form u=(y—m)? C(y—™m).
742
Chapter 7 Methods
Remark.
of Nonlinear Stochastic Systems Theory
In practice the weak convergence of the probability measures ge-
nerated by the segments of the density expansion to the probability measure generated by the density itself is more important than m.s. convergence of the segments of the density expansion over the polynomials Sry (u) to the density namely, N
feoyd A uniformly relative to
Cry Pry (u)dy a
ams A at N —
/f(u)dy
(7.6.21)
A oo on the o-algebra of Borel sets of the space R’.
Thus the partial sums of series (7.6.19) give the approximation of the distribution, i.e. the probability of any event A determined by the density f(u) with any degree of the accuracy.
The finite segment
of this expansion may be practically used for
an approximate presentation of f(u) with any degree of the accuracy even in those
cases when f(u)/V w(u) does not belong to D2(R").
In this case it is sufficient
to substitute f(u) by the truncated density (Subsection 3.8.1). Expansion (7.6.19) is valid only for the densities which have the ellipsoidal structure.
It is impossible in
principal to approximate with any degree of the accuracy by means of the ellipsoidal approximation (7.6.4) the densities which arbitrarily depend on the vector y.
It is not difficult to make sure in the fact that at the erpansion over the polynomial S,,(u) the probability densities of the random vector Y and all tts possible projections are consistent. In other words, at integrating the expansions over the polynomials S;,4;,(u), h+1 = 1, of the probability densities of the r-dimensional vector Y ,
N 1
fly) = Van]
—u
V1.4 Da Coin Shou)
u=(y-m)K(y-m),
;
y=[y7y7f,
(7.6.22)
on all the components of the /-dimensional vector y’’ we obtain the expansion over the polynomials S;,,(u1) of the probability density of the h-dimensional vector Y’ with the same coefficients,
f(y’) =
ip
1 Set?
VJ(27) | Kai
(yn ie
(he
N
1+
d
Ch, Shy (us
:
cen eh ta
where K,. 1s a covariance matrix of the vector Y’.
,
(7.6.23)
7.6.
Structural Parametrization
But approximation
of Distributions.
(7.6.23) the probability
BAM
743
density of h-dimen-
sional random vector Y‘ obtained by the integration of expansion (7.6.22) the density of (h + 1)-dimensional vector is not optimal ellipsoidal approximation of the density. For the random r-dimensional vector with an arbitrary distribution the ellipsoidal approximation (7.6.4) of its distribution determines exactly the moments till the N‘ order inclusively of the quadratic form
U =(Y —m)? K-1(Y —m), ie.
i HO gis Hecate BhateraenTees p
(7.6.24)
(E”4 stands for expectation relative to ellipsoidally approximated dis-
tribution). The ellipsoidal approximation of any distribution determines exactly the expectation of the polynomials p,o(u), gr,o(u),---,Pr,w(U), dr,n(U). In this case the initial moments of the order s, s = s; +---+ s, of the random vector Y at the approximation (7.6.4) determined by formula ess
co
cc
— Sir —- As
N
— =
Sj BY;
y=2
ae
Ms
[o@}
ms / yi. yp"w(u)dy + ss Cry | yi) Yous ae
s oe
pry (u)w(u)dy!
(736.25)
Lae
Thus at the ellipsoidal approrimation of the distribution of the random vector its moments are combined as the sums of the correspondent moments of the normal distribution and the expectations of the products of the polynomials p,,(u) by the degrees of the components of the vector
Y at the normal density w(u). 7.6.2. Ellipsoidal Approrimation of One-Dimensional Distribution We use the ellipsoidal approximation method (EAM) for finding the one-dimensional probability density f;(y;t) of the p-dimensional random process Y(t) which is determined by the Eq. (5.4.6) at the normally distributed white noise.
Suppose that we know a distribution of the
initial value Yo = Y (to) of the random process Y(t). Following the idea of EAM (Subsection 7.6.1) we present the one-dimensional density in the form of a segment of the orthogonal expansion (7.6.4) in terms of the polynomials dependent on the quadratic form u = (y? — m?)C(y — m)
744
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
where m and K = C7! are the expectation and the covariance matrix of the random process Y(t): N
fi(yst)= fPAM (u) = wi(u)
1+ eno)
(7.6.26)
Here w(u) is the normal density of the p-dimensional random vector which is chosen in correspondence with the requirement cp; = 0. The optimal coefficients of the expansion cp, are determined by relation (7.6.3) which in our case will take the form
=
/ RENCTA: Pen CD)
eae Oe 0 aan 7
Re MPa 18 TA elO°)
O°)
The set of the polynomials {pp ,(u), gp,,(u)} is constructed on the base of the orthogonal set of the polynomials {5, ,(u)} (Subsection 7.6.1) according to the following rule which provides the biorthonormality of
system (7.6.1) at p> 2:
Poult) =Spult)s toolt)=7 EN Je Se nom rG os) where the polynomial S,,(u) is given by formula (7.6.11).
Thus the
solution of the problem of finding the one-dimensional probability density by EAM is reduced to finding the expectation m and the covariance matrix K of the state vector of normal system (5.4.6) and the coefficients of the correspondent expansion Cp, also. > For deriving the equations for the expectation
covariance matrix K(t) = we use Ité formula
(5.1.15).
m(t)
—
EY (t) and the
E(Y (t) a m(t))(Y (t) _ m(t))? for normal system As a result after the evaluation
of the expectations
accounting (7.6.26) we get the equations N
m = pio(m, K,t) + > cpvyir(m, K,t),
(7.6.29)
v=2
N K=
yoo(m, se, t) ai y Cp,v Pav(m, ive), v=2
(7.6.30)
7.6. Structural Parametrization of Distributions.
EAM
745
where the following indications are introduced
yio(m, K,t) = J eta,then(u)dy,
(7.6.31)
Pin(miK3t)is J e.tpp (wer u)dy,
(7.6.32)
20(m, K,t) = [lawn —m")+(y—m)a(y,t)? +o(y, t)}wi(u)dy,
ve
(7.6.33)
pov(m, K,t) = / [a(y, t)(y7 ~
7a) +(y- m)a(y, t)”
+o(y,t) |pp,y(u)wi(u)dy,
(7.6.34)
a(y,t) = b(y, t)vo(t)b(y,t)*.
epytev(m, K,t),
«=2,...,N,
(7.6.36)
(7)
where
beo(m, K,t) = J lp ew- m)" K~'a(y,t) + tr K~*o(y,t)) —0oo
746
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
+2qh(u)(y — m)* K~*o(y,t)(y — m),
(7.6.37)
Pru (m, K,t) = / [4p ,(u)(2(y — m)? K~*a(y,t)+ tr K~*o(y,t)] +2q/(u)(y — m)? K-o(y,t)K—!(y — m).
(7.6.38)
Eqs. (7.6.29), (7.6.30), (7.6.36) at the initial conditions m(to) =m,
K (to) as Ko,
epie(ta) =
ee
(kK = De SAI N)
determine m, K, cCp2,.-.,¢p,n as time functions.
(7.6.39)
For finding the vari-
ables cp ,, the density of the initial value Yo of the state vector of a system should be approximated by formula (7.6.26). > In fact in Ité formula =
(5.1.15) for the nonlinear
function
(Y(t),t)
dp,k(U) the items which enter in it sequentially have the following form.
first item is
Ogp t(U gl¥,8) = EEO OSoU
The
(7.6.40)
where
OU ork i —m’*)C(Y T\C ian —2(Y° rr —m' bY )Cm+(Y* —m) If we differentiate the relation KC
KC+KC=0,
=
(7.6.41)
I and find from it Gi
C=-K7!KC=-CKC,
(7.6.42)
then (7.6.40) gives
O4p,.(U) = Ion U)(-2(¥ —m)" Crn—(Y—m)?CKC(Y ey —m)). (7.6.43) The expression Py (% t) in the second item of the Ité formula with the account of the value of the derivative OU/OY
= 2C(Y = m) will take the form
yy (Y,t) = 2¢, .(U)C(Y — m).
(7.6.44)
Therefore the second item will be completely equal to
gy(Y,t)"a= 2q/, .(U)(Y — m)? Ca(Y,t).
(7.6.45)
7.6. Structural Parametrization of Distributions.
EAM
TAT
“~~
For calculating the third item we shall find at first the expression yy
¥; t):
os(¥t) =On,OPH (Yet)LIE _~ 20g, 20a (U)\(¥ (UIP ==m)FC oY
oY
OY — m)? (7p OU = 205nw nV) ae(¥ — m)"C + Tint ie ores
(7.6.46)
= 495 (U)C(Y — m)(¥ — m)?C + 2Gp,«(UJC.
The third item with the account of the indication of (7.6.35) will be written as
stl euu(¥, thevoe™]= See{4g (U)C(Y = m)(¥ = my? 1
au
1
Wf
+24p,.(U) ]o(Y,t)}.
T
(7.6.47)
Finally after denoting 6 = b(y, t) we have
gy (Y,t)"bdW = 21, .(U)(Y — m)"CbdW.
(7.6.48)
At the following stage we shall find the expectations of the obtained items. For the first item of (7.6.43) we have
Bq, (UY(V? =m
C= gy(u(y? — mB CAPA (w)ay,
We evaluate the expectation in this formula. While introducing a new vector variable
oi CHAa(y — mM) and noticing that u = (y? = mm? )C'l4zCR2(y —m)=
\y|?
we obtain [e.e)
Eq, .(U)(¥7 —m?)C= faPyk (ll?) ¥701" FFA(Ig)?)|C|!/7dy =0, —0o
(7.6.49)
as the integrand is odd relatively to each of the coordinates of the vector y and the integral with respect to odd function in symmetrical limits is always equal to zero.
For evaluating the expectation in the second item of (7.6.43) we use formula
ET AE = tr[ AEE? ] which is valid for any vectors € and the quadratic matrices A. We obtain
(Y? —m™)CKC(Y — m) =tr[KC(Y —m)(Y—m)'C.
(7.6.50)
748
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Thus it is necessary to find an expectation
Eq, .(U)C(Y — m)(Y - m)'C = Ew,q),.(U)C(Y — m)(Y — m)'C N
+>) cp,» Eur %,n(U)Pp»(U)C(Y — m)(Y — m)*C. yv=2
Here and further on Ew, means an expectation at the density wi(u). It is evident
that the finding of this expectation is reduced to the evaluation of the expectations of the form Pou
Ga. = m)(Y = m)? CO} as Ip, (U) and Pp,v (U) are
the polynomials of the variable u. = Cag
BylU
After performing the change of the variables y
= m) = C!/2z in the integral which expresses F,,, we may write
CRXP QS ES UECLLY
CU NaC VEGA UT YY
At first we calculate the diagonal elements of the matrix Bi. UFYY7. bution of the random vector =.=
eer
are identical.
|i.e.
all the diagonal elements of the matrix
Therefore we may put Beg {U FY 4} =
2 and taking into account that
= hi U
As the distri-
Y = C’ 1/2X is spherically symmetric then Ew, U Ry?
U =
bLal
& ps 2
3S ve =
YTY
qa.
Bey
ve
After summarizing over 5
we have Ea Ary
ae
yy
t=1
= pa.
As far as the nondiagonal elements of this matrix concerned to be zero for the same reason that in formula
(7.6.49).
they turn out
As a result we obtain
Ci? Ba hV Ys Chex: (Bil!ie/p)G. Gonaeruently, (EU "OX X* GC = (Ew, U**t1/p)C. Thus 1 Eq,,.(U)CXX7C N
+0 v=2
Cc
Pv
=
pom {4p,.(U)UC}
Ba {ah «(U) Pp,» (U)UC}.
(7.6.51)
B
For calculating the expectations in (7.6.51) it is sufficient to know the probability the density w(u) of the variable U correspondent to the normal probability density W 1 (u):
w(u)
she
1
DTT (p/ay" p/2-1,—-u/2 €
(7.6.52)
Now we find 1
co
Eq,qo xl(U)CXX7C = seroinn ) 2P/2T (p/2)p |pn 0
(u)u
pide igHi?
du
7.6. Structural Parametrization of Distributions.
749
co
N
+
EAM
eis1 hdDp Deo» oot FTG
|eh
v=
0
Pre( uu p/2,-u/2 e du>C..
/
(7.6.53)
E
For the sake of the convenience we introduce the denotions
Upeetn a dantoararrrs wg ule de, ? 0 1
[oe
= Acwereeta
a ae 9
9p/27
eo
(p/2)p
—u/2 tte PeWop,kcolth)Bp. P, v( Seal dens
(7.6.55)
J
/
acs that in the consequence of the orthogonality of Po, v(u) to all the functions u> at r > V the variable Y,, vanishes at Y > K. Therefore
kK Eq,,.(U)CXXTC
=
|.+
Ss ote
G
YEP
We evaluate the variables ¥,9 and Y,,, determined by formulae (7.6.54) and (7.6.55). At first we find y,,,. But virtue of the orthogonality of u to the polynomial
Pp,n(u) at AX< kK Yen =
1 TR
(p+2K—2)!"(2K)! Ga
fe—a/9 Don ujur'e du.
0 (7.6.56) For the same reason formula for Y,,
may
be written in accordance with formula
(7.6.3) and Appendix 7 in the form
ck
be TE FOR al EN IES
“edu
= —,
(7.6.57) Now we evaluate 7x, for arbitrary
Y < K. While integrating by parts we obtain
il
Yer = P70 (p/2) /In (¥)Pp,v(u)ur?e“/?du 0
1 S fae mat SOPi,
p/2,,—u/2}0o
750
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Ef tentU)Pp,v (u yur teel2du— fdyn Pp y(wurl2em "du 0
0
(7.6.58)
Pp (wale?
+5 |dm 0
It is evident that the first item is equal to zero. The first integral in (7.6.58) at V < K vanishes by virtue of the condition of the biorthonormality (7.6.8). The equality to zero of the second integral is shown by means of the same way as in (7.6.57).
Analogously the third integral is equal to zero at
vy Yo...2...0 + i=l
+
."
r
3:
a
i=1 j=i41
ye
eS)
GoW 710 ¢:.610.5:50 t
3
hae ere
)
(7.6.117)
where Vege
=
BX
1Ne
? mits Sh
=
EX}? X? /4)j.
(7.6.118)
For the second item in (7.6.110) we have
Ew,Pi,2(Ui)X} = 2? - 2'Ry 2= 8Ra2 = 24D?, as
(7.6.119)
Ra2= DY) Chupad_, = Cimgag) = wy = 3Di, Jh|=4
where p’ and a{_, are the central and the initial moments of the normal distribution N(0, K’) respectively.
7.6. Structural Paramefrization of Distributions.
EAM
765
Thus the ellipsoidal approximation of the r-dimensional distribution gives for the fourth moment of the first and the i* coordinates
EpaX? = 3D? + 24c,2D?, EpaX? =3D?+424c,9D?.
—(7.6.120)
By virtue of (7.6.103) and (7.6.104) we have
Oia hee
ee pak a Bk
(7.6.121)
S34" ae, 0)ee 1(4,0)-..50) =6:(4,0,...,0)
7.6.122 (7.6.122)
The fourth order moments of the vector may be also given by the
expressions of the type EX?X?, EX; X3, EX;X;X,X.
We evaluate
BX? X he The remained two types of the moments are determined analogously. After choosing, for example, 7 = 1 and j = 2, we write
—
j[tif@da
N
Sel v=2
ies
a, = / oe | situ (wdin a,
ae [Fee
wuluddés «die.
(7.6.123)
ES
Let us integrate over £3,...,Z, using (7.6.23). As a result we obtain co
Ne | #BF@dé. a= _
—
v=2
| | attiu(us)aes diy —o6O
Dew f lc ey
oC
=—©6cOo
Uz) w(uz)dz; dro,
(7.6.124)
eS
where ug = £7/\; + 23/2, w(u2) is two-dimensional normal density. We rewrite (7.6.124) in the form of the expectations and substitute the values of the normal moments
PraXixo= Hoey SMEn X? X2 po, (U2). v=2
(7.6.125)
766
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
We evaluate the expectation in (7.6.125):
Bw, P2,v(U2)X2X3 = Ewp2,2(U2)X7XF
= 2? .2!Roo2 = 8Re2,2 = 8DiD2
(7.6.126)
by virtue of the fact that at y = 2 it turns out that Fuso
Uz) Kade Ean
and, ath: < 7; ok
R2,2,2 =
»
hi
ah2,,w
w
Cy° Cy nip Din ale)
SND
Tay, Wee
tee
C203 19 2%.0 ee D, Do.
|h|=4
According to (7.6.126) the ellipsoidal approximation (7.6.125) of the mixed fourth order moment will take the form
EX? X2\=.D Do + 8c; 9W1 Do.
(7.6.127)
Therefore finally we get:
6: (Qr2h
One 9 Oiceeb de
Mot BAX,
(7.6.128)
6(2,2,0,...,0) = 61(2,2,0,...,0)/BX2X2.
(7.6.129)
Example 7.6.1. Under conditions of Example 7.5.5 Eqs. (7.6.30), (7.6.36) have the form:
Mm, =—m
m2 —k12,
mz = —amo;
(7.6.29),
(I)
ki = —2(moki1 +m, k12), kia = —2(mz + a)ky2 — myko2,
koo = —2akoo + hv; €9,2 = 4h?vC29 + €2,2{(m + a)(6 + 8Ci2k12) + 8m
(II) Cy2k12 — 3h?2vCo9},
(II)
62,3 = —24h7?vC2243c2,2{(m2+a)(Ci2k12—5) +m Cy2k22 +4, 5h2v C29} +3c2,3{6(ms + a)Cyok12 + 6m, Ciek12 — 3h?vC22},
(IV)
€2,4 = 192h?vCo2+24h7?VO 29,2 +4c0,3{(m2+a)(Ci2k12—7) +m Cy 2k22 +6, 5h?vC29} +2c¢2,4{(me +a)(14+ 16Cy2ki2)+ 16m, Ci2k22 —Th?vCo},
(V)
7.6. Structural Parametrization of Distributions.
C25 = =1920h2Cy5
EAM
767
_ 240h?vC2gc,2 + 40h?vCo2¢9,3
+5c2,4{(m2 + a)(Ci2k12 — 9) + mi Ciako2 — 8, 5h?vC29} +5c2 5{(m2 + a)(9 + Ci2k12) + 10m, Cy2ko2 —- 4, 5h2vC29}.
(VI)
Here OF are the elements of the matrix C; C = K ee, The peculiarity of these equations is the fact that Eq.
(I) for the expectation
and Eq. (II) for the covariance matrix are separated from Eqs. (III)—(VI) for the coefficients of the expansion. Here Eq. (I), (II) coincide with the correspondent equations
of NAM
(Example 7.2.3). If we wish to restrict our consideration to the moments
till the fourth order we may put C2.3 = C2,4 = C25 = 0 in the obtained equations. By the same procedure the equations-in the case of accounting the moments till the
sixth or the eighth order are obtained.
On Figs. 7.6.1—7.6.5 the accuracy of EAM while calculating the intial moments Qc
Sm,
LDs=
ki1, @2, @4, &g, Ag in accordance with time ¢ at the interval
[0, 1]: a = 5, A = 1 and at the initial conditions: m,(0) = 0,5, m2(0) = 0,5, k11(0)
=
D,(0)
=
ad
k22(0)
=
D2(0)
at
Re
ky2(0)
=
0 is shown
The accuracy of @4, @g, Q@g ranges about — 2%, 8%, 20% relative to the exact solution respectively. Hence we see that EAM gives a good accuracy similar as MSM
(Subsection 7.4.7). In this case EAM with the account of the moments will the 10th order gives 9 equations instend of 65 ones at MSM.
0,485
a=m,
0,481 0,477
: + ES; EAMio; MSMy2
0,473 4
—
0,469 + 0,465 0,1
0,3
0,5
Fig. 7.6.1
0,7
t 0,9
Fig. 7.6.2
.
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
768
re 0,26
4
0,25 +
EAM\o i
ES
0,24
EAM,
0,23
0,22
MSM
0,21
=
0,20 +
0,1
t
0,3
0,5
0,7
0,9
Fig. 7.6.3
1,0 Og
0,9 0,8
0,7
EAM,
iF
ES
0.6
BS
EAM, ~|
0,4
MSM,,
0,3 0,2
+
1
2
3
4
t
5
Fig. 7.6.5
The method
of the ellipsoidal approximation
in the case of Eqs.
(5.4.6) with the nongaussian white noise and (5.4.7) is given in (Pugachev and Sinitsyn 2000). 7.7. Methods based on Canonical Representations 7.7.1. Linearization by Canonical Expansions
The linearization of the operator equations from the point of view
of general theory of the stochastic systems transformation (Section 6.5) may be used in two variants. Firstly, a given dependence between random functions may be directly linearized and by such a way the nonlinear equations connecting the random functions may be substituted by linear equations. Secondly, the canonical expansions method which leads
7.7. Methods based onmCanonical Representations
769
to the replacement of the operations over the random functions by the operations over the ordinary random variables may be used. After that we may use an ordinary in the applications method of linearization of the functional dependences between random variables. The method of direct linearization of the nonlinear transformation of random functions consists in the replacement of all given equations connecting the random functions by approximate linear equations which reflect adequately the real dependence between the random functions in the domain practically possible realizations of the random functions. As the expectations of the random variables are the mean values near which their possible realizations are considered then in practice it is more convenient to perform the linearization of the relations between the random functions relatively their deviations from the expectations, i.e. the centered random functions. In this case all the functions which enter into the given equations should be expanded into the Taylor series over the centered random functions and rejected the terms of these series higher than the first degree. The degree of the accuracy of the approximation obtained in such a way may be estimated by maximal possible value of the rejected terms in the region of practically possible realizations of random functions. After substituting the given equations connecting the random functions by the approximate linear equations we may use the methods of Section 6.3 for the approximate determination of the expectations and the covariance functions found as a result of the considered nonlinear transformation. Let us pass to the application of the canonical expansions method to the approximate analysis of the nonlinear transformations of the random functions. Let the scalar random function Y(t) be obtained as a result of the transformation of the scalar random functions X(t) by means of the nonlinear operator A:
Y(t) = AX(t).
(ain
Substituting here instead of the random function X(t) any its canonical expansion we find
Y(t)=A {meloSy veo} =wia(VanVoetent)ei
anuntl:T2)
Equality (7.7.2) represents the random function Y(t) as some function, in general case, nonlinear of the random variables V, into which the argument t enters as a parameter. While linearizing this function and
770
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
taking into consideration that EV, = 0, v = 1,2,... we obtain the following approximate expansion of the function Y(t) and its covariance function K,(t, t’):
Y(t) © Ame(t)+ S>Wwi(t), Ky(t.t!) © >> Diw(t)y(t).
(7.7.3)
Here
w(g= |ee
sora {mets Dvseuto}] , Lowy
plafinans7vtid)
where zero below indicates that after the calculation of the derivative all the random variables should be equated to zero, V, = 0. We apply the stated arguments to the nonlinear differential system
RS
Cs
eR
a
cetera OS
aaa
ere,
(Talo)
Here f; are the known functions and the random disturbances = X,(t,Y1,.-., Ym) allow the joint canonical expansion
Nprmamich yy Vota (3 adgewl mn),
X,
(7.7.6)
V
where Wi. = (EY es LR lye eye brl is seek mi Vy ale one uncorrelated random coefficients with zero expectations and the vari-
ances D,. If we substitute (7.7.6) into (7.7.5) then the right-hand sides of Eqs. (7.7.5) will depend on the random coefficients V,. Here they will be known functions of the variables t, Y;,..., Ym and V,. As a result
of the integration of these equations we obtain Y; = yz (t, Vi, Vo,.-..). If we consider vy, as the functions of V, sufficiently close to the linear ones
in the region of practically possible values of V, then we obtain in the linear approximation
¥, *[%]+
OY;
5 VY |a | v
(7.7.7)
an!
where by zero below is indicated that after the calculation all the derivatives should be taken at V, = 0.
Hence using the formulae for the canonical expansions of the vector random functions (Subsection 3.9.3) we find the approximate canonical expansion Y; and the covariance functions: Yh y
my +SoVWwe,
v
7.7. Methods based onCanonical Representations
KY, (t,t')
real
2, Dodon|(t)yq(t) (p,q,k =1,...,m).
(7.7.8)
where mi © [Ye Jo, we(t) = [OY /OV, ]). From (7.7.5) after the differentiation m{, yy, over time and accounting that
& (SE) = 5 ShMe
: Of Oe yo
OV, )]~ 4+ OY, OV, OX gis
". Ims AY,
ay, + Loy
p=1
2
we find
-
+ OX, OV,
. ., OLyy OY,
ay, t Lay ay,” Bh p=l
4
my = = Ye lo =[ filo »
(7.7.9)
: ahproyy Of, Om= imt)=5 || = Ofk oa a“s Ome ehale dt |OV, J, »» omy Oms OY,
Ofr
4
Consequently,
=]
q
(7.7.9) and (7.7.10) lead to the following ordinary
differential equations for m% and yx:
mieafe
miami
y!| ieqimP) |
Wk = > Axp(t)wp + > beq(t)evq(t), p=1
(Fe FNY)
(7.7.12)
q=1
where
im? = ni (& MiG. ard (3
=
—
((k,p=1 »P
~
me beg(t =
“. —— Of, Om — Om= Om
ame
(keel
Cy At = ipa (emt, Oe
2
ae
Le
eG
o, mi)
yin
)
matt);
(797.13)
Sole, )
qu)
el aas
12)
Thus after presenting the vector random function by canonical
expansion
(7.7.6) we get a system
X =[X1...Xn es
of ordinary nonlinear
772
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
differential Eqs. (7.7.11) for an approrimate determination of the ezpectation of the random function Y = [Y1..-Ym i* and a system of the linear ordinary differential equations with the one and the same matriz at unknown set of Eqs. (7.7.12), for determination of the coordinate functions. After integrating Eqs. (7.7.11) and (7.7.12) we shall find by formulae (7.7.8) the approzimate canonical expansion Y and its covart-
ance function K¥(t,t’). It remains to solve the problem about the initial conditions for Eqs. (7.7.11) and (7.7.12). If the initial conditions are nonrandom, i.e. t = to, Yx = yro (kK =1,..., m) then it is necessary to integrate the equations at the initial conditions t = to, my = yko, Yk = 0. For the random initial
conditions t= to, Y¥r= ype (k = es ay eas. (11) atid (i722) are integrated at the initial conditions t = to, mZ = my, wr = yro- In the latter case mo
and y,x0 are found from the joint canonical expansion
of (n + m)-dimensional random function [ X7 YJ lee Analogously Eqs. (7.7.1) and (7.7.12) are written for the equations unsolved relatively to the derivatives, F'(Y, Y, TaD CEG: HAE ot: payin). bE) 02 Na Ya (a el ore: Example
7.7.1. Consider the scalar nonlinear system
Vos where
X =
oe
X (t) is a stationary random
me (Oi ts
(I)
process with zero expectation
and the
known spectral density Sz (w). We use a segment of the canonical expansion of the
stationary random function X (t) at the interval of the duration 27’ (Example 4.8.5): N
Me
We PaO
ame
(II)
v=—N Wy+ta
Here the variances V, are equal to Div, ] =) DF
=
f
Sz (w)dw, Tea
Wits ot
Wy-a +2,..., where 2q@ is the length of each of (2N + 1) segments of the interval of the frequences (-Q, Q). In a given case Eq. (7.7.11) has the form My =a _ mi, My (0) = 1. It allows the integral my (t) = 1. Eq. (7.7.12) approximately determining the coordinate functions py = Yy,(t,w,) with account My (t) = 1 has the form yy = —2y, + et
v(t,uv)
U5 (tape,
Its integral vanishing at t = 0 is equal t
2t
Je2r+iwvt dr = 0
=
Bee
ae! ——____
Ill (1m
7.7. Methods based oneCanonical Representations
773
Consequently,
N
Jeivet 43. gat?
Do
Dy(t)= 9 Dy vr=—N
TS
Example disturbance X
1 = 2e-2" cost
=bien syioee
=
7.7.2.
y=)
+ e7*
(IV)
Solve Example 7.7.1 supposing that the input random
X(t, Y) is a real stationary random function with the spectral
density Sz (w, A). It is evident that similarly as in Example 7.7.1, My (t) = 1. For determining the variances of the random function Y we use a segment of its canonical expansion in the region 0) < t < 27’, lY |
we get Vive
Differentiating Eq. (II) over
—2Yy
v2
a
Wy Ye
.
(IIT)
V,, and putting after this V, = 0 we get
; Yyvev3
=
1 =
—2Y,vevs
i
3 (Ys YWv3 FP W2Yrivs
i YWsYvive) o
As by the condition Y (0) = | then the integrals of Eqs. vanish, grip, (O) =a) Ae
(IV)
(III) and (IV) at t = 0
veil) = 0. After substituting expression (III) of Exam-
ple 7.7.1 for Y, into (II) and integrating this equation we find Yi V2 (t). Analogously
substituting Y, and Yy,p, (t) into (IV) we find Yy, v.13 (t). By virtue of the fact that the random function X (t) is normally distributed all V, are normally distributed, and consequently, V, are not only uncorrelated but
are independent as well.
Therefore all their third order moments
are equal to zero
7.7. Methods based gn Canonical Representations
and between indexes
the fourth order moments
coincide pairwise:
== D,, Ds,
[lyyyy
=
783
only those are different from zero EV?
—
a.
Myzrivev2
=
EV?
whose EV}
Consequently, after calculating with the stipulated accuracy finally we
get
my(t)
=1+4+
»
Dy yr
v=—oo0
=e 2
op Vy=—0CO
(t),
Dy (
Mey
D,y3(t)
v=—oco
Se
(V)
Dy, D,.. [ue, (t) + 3Yv, (Peer)
5
V2=>—-00
Expressing the random function Y(s) in terms of V, we have the principal possibility of finding the n-dimensional distributions of the random function at any n. In the general case the determination of the distributions of the random functions by means of the canonical expansions is connected with great calculations difficulties. But for some classes of the nonlinear stochastic systems the canonical expansions method may be sucessfully used together with the distributions parametrization methods stated in Sections 7.2—7.6. Hence the essential advantages connected with the calculation of the integrals with respect to the nonlinear functions in the equations for the parameters of the nongaussian distributions V, appear. Example form
7.7.6. Let us consider the nonlinear differential system of the
F(t, DY =9(V) +X), XH=mOH+
U2),
where F(t, D) is the polynomial relatively to the differentiation operator
D = d/dt
with the coefficients dependent on time t; p = ~(Y) is the nonlinear, for instance, the discontinuous function, and (0) =,(); X(t) is the normally distributed random
function. Performing by the formulae of Subsection 7.2.2 the statistical linearization of the nonlinear function ( we substitute initial Eqs. (I) by linear one:
F(t, D)Y = ko(my, Dy)my + ki(my, Dy)(Y — my) + X(t).
(II)
The functions My (t) and yy (t) which determine the canonical expansion Y(t) and Ky tt, t’) in the form
Y=m,(t)+ > W(t), Ky(t,t') = >) Diw(t)y(t),
(IIT)
784
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
are defined by the following equations, see (II):
F(t, D)my = ko(tmy, Dy)my + Mz ; FU, D)y, = bil, Dy)tp 42
(IV)
;
where Dy = iS Dy lyv (ol The joint integration of nonlinear ordinary Eqs. (IV)
vy allows to find My (t), Yy (t) and Ky (t, ne In the case when the function X(t) has the nongaussian distribution Eqs. (IV) will have the following form:
FG, D)my = ko, ,D™))my + Mz ;
ex
EG
D)yy
=
kiy(™m,, 1B
(V)
wy + By,
where the coefficients ko(my, D{*)) and kj, (my j D&)) will depend on the random variables Viu- Here the coefficients kj, will be different for various V,.
The canoncial expansions method forms the basis of: e the methods of equivalent disturbances (Dostupov 1970),
e the methods of the stochastic functional series (Adamian 1983), e the statistical linearization and the normalization methods based on the canonical expansions (Sinitsyn 1974, Pugachev and Sinitsyn 1978), e the modifications of the methods of Sections 7.4-7.6 based on the canonical expansions. 7.8. Applications. Analysis of Stochastic Systems and Analytical Modeling 7.8.1. Accuracy Analysis of Single Loop Nonlinear Control Systems At first let us consider the one-dimensional stationary control system (Fig.7.8.1) consisting of the linear stationary system with the transfer function ®(s) and the nonlinear element whose input and output variables are connected by the single-valued nonlinear dependence
U =
g(Y). The input X we shall assume as the stationary
random function with the constant expectation M, and the spectral density Sz, (w).
Using the statistical linerization of the nonlinear element in the form
US PY YS Coligny)
bility, Dali) 2h
Yo Aimy
181)
we shall obtain the following dependences between the expectations M, and My and the centred random functions
U° =
U — My
and y°®:
T= ko(Mgy Dy mee U =
Day,
(7.8.2)
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
785
where
“ko(my, Dy) = Yo(my, Dy)my" From the relations (Fig.7.8.1) My
=
©(0)(mz
— mw), Yo
@(s)(X° = Oe)
accounting (7.8.1), (7.8.2) we have the following statistically linearized operator equations: My
—
y? = =
(0 B(°)(0;
p() G(s; :
my, Dy)mz
my,
pe y)X-;
D
,
B{)(0; my, Dy) a
o(1) y
(8;} My, Dy)1)=
Tae
yrry
oe) RieeD
a
oa
(7.8.3)
So from Eq. (7.8.3) we get the corresponding equation for the variance
|__*) _
7
Dy)®(iw) e my, e 1+ ki( Seal L Solving numerically Eqs.
(7.8.4)
Sp (w)dw .
(7.83) and (7.8.4) jointly we find the expectation My
and the variance Dy Z
Example
7.8.1. Find the expectation and the variance of the output of
the system which consists of the stationary oscillatory link, with the transfer function
®(s) =
mies aS
$ao), and the negative feedback across the relay element
arranged in Fig.7.8.1. Assume that the input X is the stationary random process,
Mi =~const,
bk2(7T) = Dre
lia
Ds
a
Sz(w) = Sita
(I)
From Eqs. (7.8.3) and (7.8.4) we have Me
thes ao + ko(my, Dy)’
(i)
=f T
anes eets 2
a
ws. (a; +w?) lap —w? + ayiw + ky(my, Dy)|
TT |)
786
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
D,(a +41)
lat hitms Dig
Se
We complete the solution of Eqs.
aa+ a eee HEGRD NT
otIll
(I), (II) using Table A.3.1 at the following
values for the parameters of the system and the random input:
@p = 1, o? =0.2, a =1/a, l=1, eg = 1.
(IV)
Substituting values (IV) into Eq. (II) at
Eq. (II) becomes
A)
rachis)
(VI)
It is clear that Eq. (VI) is equivalent to the following one:
in,
m
= 1 —2h
Z
Vil
From Eq. (III) at values (IV) and relations
we get implicit equation for Dy
2D, Dy =
(IX)
+? (ae) |22+ ae (Se) The results of the numerical solution of Eqs. (VI), (IX) are given in Table 7.8.1. Table 7.8.1.
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
787
For the nonlinear control system illustrated on Fig.7.8.2 we have the following expressions of the transfer functions of the statistically linearized system:
ko(m,, Dy) ®(s)
CAL PR My, AS 5 (8,
Nl Dy)ey Pakgtmg,
ad D5) G(s) ’
(7.8.5)
Dy)®(s) 60) —ki (my, DAs)’ y (sm (s4ng, 4)DJ) = ie eke 1
(sera
De ee
Mein,
0
(7.8.6)
Eb)
CHS By
1+ ko(m,, Dy) ®(s) ’
(78.2)
:
(7.8.8)
1+ ki(my, Dy) ®(s) ©
Consequently,
ko(my, Dy )®(0) eg ~~ a1+ko(m,, La TY Dy)NL (0) a
a.ka (at,
De
| ee aon,
-
1
My kale, D. (0) a (he
(iw)
2
D.) |Sz (w) TEAGa, Deo Gn) dw,
_ ff TERR __ se(w)dw) Dy = AOATES Solving numerically Eqs.
(7.8.10)
(7.8.9) and (7.8.10) jointly, we find at first mM, and D,,
and further My and Dy ‘
Example
7.8.2. Find the accuracy characteristics of the control system
(Fig.7.8.2) consisting of the linear drive with the transfer function (s) =e a/s(T's + 1); the device for comparing
the input and the output,
and the relay element
emitting drive control signals of the constant absolute magnitude.
Analogously to
the previous example we have at first
Oe PES y
and My
= Mz.
T?2s?+s+ koa’
Then (1) Ses
®;
s(Ts + 1)
T?s2+5+ kya’
(I)
kya fe = elem cil i toa ® T?7s24+s+ ka’
am)
788
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
ne
aD,k?a?
=e
us
—00
r ues
aD,
ple
/
dw
(a? + w?) |kya — Tw? + iw]?
Drkya(1 + aT)
:
IV
ae
a(1+aT)+ kya’
/
w?(T2w? + 1)dw
(a? + w?) |kya ~ Tw? + iw|?
"a(l+ aT)+hia 1+
aT
©
+ k,aT
Here ko and ky are defined by formulae (V), (VIII) of Example 7.8.1. The results of
numerical solution of Eqs. (IV), (V) at! = T =
1, a = 0.5, a = 5 are given in
Table 7.8.2. Table 7.8.2.
Table 7.8.2 demonstrates
the “choking”
effect of the relay element by large
fluctuations.
7.8.2. Solid Dynamics in Stochastic Media Let us consider the dynamics of the solid with one fixed point (Example 1.1.5) in the stochastic media using the following models of the stochastic media:
1°
SoA, = —Gol ewe + V2, My = —Golywy + Vy,
=
I
—polzwz
(7.8.11)
SIF ee
ys Vy , V; being the normal white noises with the equal intensities, vi = = Vv, =
V0, Yo being the viscous coefficient;
Vy
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
2°
Mz = —Yolewe + Vy", My = —Polywy + Vy’, Mz =—o0I1,w,
789
(7.8.12)
+ ve
ee Ve Vi being the normal white noises with intensities satisfying the conditions: 1 ==
Yoles VyLi= Voly, VY,ti =ro Volz;
=
Mz = —$ilgwe + Vz",
My = —Y2lywy + Vy", M, = —p3l,w,
ie
(7.8.13)
ar Vore
Une yes being the independent normal white noises with the intensities y///’ vw?
ae Fya $1, ~2, 3
being the viscous coefficients.
In case 1° the Ité—Euler dynamical equations have the following form:
IpWe + (Iz — Ty )wywz = —Yolewe + Vz,
Tywy + Ie — Iz )w2w2 = —polywy + Vy,
(7.8.14)
T,0, + (ly — Ip )woWy = —Polezw, + Vy. Eqs. (7.8.14) have the stationary regime of fluctuations with the following onedimensional density:
fi(Wr, Wy, wz) = cexp[—(Y0n/M)], where
(7.8.15)
n= I2ug + Iywy + 1207, ¢ = [po/(mv0)]” Lely le. More than that. In the nonstationary regime of fluctuations Eqs. (7.8.14) may
be transformed into the following set of the independent linear equations
dL; dt
= —yoli+ V; (i = 1,2,3).
(7.8.16)
From Eq. (7.8.16) using the results of Subsection 6.2.3 we get the explicit formula for the multi-dimensional characteristic fucntions for the variables D; (2 saul, 2, 3); n
AneGi
gn(A1 a)
tes
y ee) =
ft
eee k=
x exp
==>
n
n
yD tA [tr —th| — e~ volts +t) pape
hstt=1
UG o1 0)
790
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
where Jo (A) being the characteristic function of the initial conditions: L; (to). Hence
the density of the variable 7) = L ae Le ap Rs > 0 and its first and second moments are
fn(n) = (¥0/0)0(3/2)./n exp(—¥on/Vo),
(7.8.18)
(IC being the gamma-function, see Appendix 7),
En = 3v9/(2¢0), In case 2° Eqs.
Dn = 3v0/(4¢0).
(7.8.19)
(7.8.13) have the stationary regime of fluctuations with the
following one-dimensinal density:
2
fi(We, Wy, wz) = cexp (0
);
(7.8.20)
where
3/2
27 = Taw, Pilggbh iews, -¢
Vy
(#) TVo
Using It6 formula (5.1.7) we get the first order nonlinear stochastic differential equation for the variable T,,
T= 205) valet V2 1AV
(7.8.21)
Here V being the normmal white noise of the intensity 9.
The equations for the
multi-dimensional characteristic function g/,(A1,.-., Anjti,.--, tn) for T(t) defined by Eq. (7.8.21) are as follows:
Te OgSS + Siv0Angha. 29m = An(=2p0 + iYoAn) At, /
The solution of Eq. (7.8.22) forn =
(7.8.22)
1,2,... according to Subsection 6.2.2 at the
intial conditions
9;(A5to) = 9o(A); 9, (1 , te
tNnit xd otere siniitnaizilnew)
= In—101
yo
(7.8.23)
An=2)An=1 me cere a echt
tact)
may be presented in the following explicit form:
gi (Ast) egg (Ae 7b 00,t— 16) 9/2(\,t — 0), y (7.8.24) In(A1;
SEC
Anti:
Cg OHR
rin)
=
GreTtnr
sey
NG,
Nees
+
Ane
Colin
—tn—1)
7.8. Applications.
Analysis of Stechastic Systems and Analytical Modeling
=A
=
bit Ongtwert ten) iMesietadWF (Anite where
791
tea) (n= 2, 3).&)j (7.8.25)
(A, 7) = 1— idvo(1 — €7 207) /(2H0).
(7.8.26)
In special case go(A) = (1 = ir6)~3/? from formula (7.8.24) we have
LY) g1( 434) = fa — i
+ 4X (
20
~ 5)ec
t
gi(A; co) =(1 — idAvp/2~)°!” :
eelo.20)
(7.8.28)
Formula (7.8.28) corresponds to the x?-distrubution with three degree of freedom. From Eq. (7.8.21) we have the linear equations for the expectation i=
BT:
and the variance D’/ = DT’,
dm! _
a¥o|
dD" |
;
;
In case 3° we have the following It6—Euler dynamical equations We = Cj WyWz — PiWe + Vi, Wy = €2QWzWe — PoWy + Vo, WwW, = where
EgWeWy
LD Espn €1
=
iT
9
Ven
ar
— P3Wz
te V3,
@8 Aalewih 2
ie
ts sla1, ose
Ved
diate
(7.8.30)
—
ie
)
ie
V3
Ti
From Eqs. (7.8.30) using the moments method (Subsectio 7.4.1) we get the infinite set of equations for the initial moments Aj,m; = Ewr WyPk
Anmk = —(ngitmye+kys)anmk+Ne1 On—1,m41,k+1 +ME2An41,m—1,k+1 1 +ke30n41,.m41,k-1
1
a guin(n 1 1) 0, 23 mak ae gvzm(m 7
+5usk(h ey
L)Gaee.k
(7.8.31)
792
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
With the accuracy up to the moments of the fourth order we get from Eq. (7.8.31):
100 = %010 = @110 = %101 = %11 = 9, 200
11 /(291), 020 = ¥2/(2"2), M002 = v3/(2¥3).
Remark1.
(7.8.32)
The nonlinear stochatic differential equations for the solid in the
stochastic media admit the exact normal stationary solution
f(w) = cexp(—w7 Aw).
(7.8.33)
Here
w = [wewyws]’,
A=AT+ pl = Ivy, ¢ = (2n)-9/? |Ap”?, peau a0 leO le pie |
LS
v="
Ve O 0.
pve WO"
aie
0 COe 34,
Say
1 O10 KO gaa Odes O Os 1
pr io Cte ONS O20 naGa
where A and }L are some real coefficients.
7.8.3. Stochastic Hereditary Systems Many problems of the theory ofhereditary systemms (for instance, the systems which contain the elements described by hereditary mechanics, physics, bilogy, chemistry) under the conditions of random disturbances lead to the study of the following
stochastic integrodifferential system: t
X =a(X,t)+ fax
nber
to t
Ub EX:t)+ /bi(X(r),7,t)dr |W, X(to) = Xo,
(7.8.34)
to
which is considered in the It6 sense. Here X € R? is the state vector; W € R14 is the vector white noise of the intensity
Wis
V = V (t) independent of the initial condition Xo;
W(t) is the random second order process with the independent increments;
a=a(X,t),a:
Rix
a; = a;(X(r),7,t),a, b=a(X,t),b:
R= RP; : RPx Rx
R—
RP;
R—
RP?!
Rix R— R?!;
by = bi(X(r),7,t),b)
: RPx Rx
7.8. Applications,
Analysis of Stechastic Systems and Analytical Modeling
793
In applications the functions @; and b; are represented in the form N
a, = A(t,7)p(X(r),7), 61 = >> Ba(t,7)va(X((7),7),
(7.8.35)
h=1
where hereditary kernels A(t, 7) and Bp (t, 7) are (p x p)- and (p X q)-martix functions, ~p (X(7), T) is the (q x q)-matrix function, y( X(T), c) is the p-dimensional
vector fucntion. Let us suppose that the hereditary kernelsaes \(t,7) = {Aj; (t, Ty;
oe
Op and Bill)
Bultst),
0 = 1N,t
= 1,p, 7 =
14g satistly
the conditions of the nonanticipativeness and the conditions of dying hereditary (or
assymptotical stability) 7a
A)
(7.8.36)
es)
7)| dr < 00, i |Ai;(¢,
/ [Braj(t,7)| dr< co:
(7.8.37)
In the case where the hereditary kernels satisfy the conditions
attor paar (OM Bir) SB
OP CS ts
(7.8.38)
one say about the invariance of the stationary of the hereditary. The important class
of hereditary kernels are singular kernels
,7) Bus (t,7) = BE; (OBu; (7). = AZMAG(r), Aj;(t
.(7.8.39)
The simplest exmaples of the functions satisfying conditions (7.8.36)—(7.8.39) are the
functions exp(—a@ |€| )1(€), exp(—a |€| )[cosw€+~ysinw |£|]1(€), 1(€) is the unit step function. One of the most effective ways of obtaining the statistical dynamics equations
for Eq. (7.8.34) is their reduction to the stochastic differential equations by means of the approximation of the real hereditary kernels by the functions satisfying some ordinary differential equations.
Let us consider Eq. (7.8.34) under condition (7.8.35) t
Few aye /rCeatieetararie to
rare Arey /By(t,7)a(X(7),r)dr |W, X(to)= Xo~ (7.8140) h=1 7,
794
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
We shall determine a class of admissible kernels by conditions (7.8.36), (7.8.37) and by the order and the form of linear ordinary equations to which the kernels as the functions of t and T must satisfy.
Let the kernels Aj; it. T) and Brij (, T) in (7.8.40) satisfy (7.8.36), (7.8.37) and at fixed T be the solutions of linear differential equations
Fe Ay)(i'r) OOS wien gS Typ, ne
(7.8.41)
S > Biri Brig(t, 7) = Wir O(t = B)ermdypreps deg Vil
at a fixed t the kernels are determined by the formulae
Ay; ( (tsT)= —
SH?
Al, (t,7),
Us = dg &
(7.8.42) Bnij((t, T) =
Yo YES Blt
Tce
pal
i Us j= ied,
and the functions At, (i, rg and Bia; (t, T) entering into (7.8.42) at a fixedt are the solutions of the linear differential equations
DoF A (t,r) = 5,16(t— Tr), lr =Tp,
z
(7.8.43)
Sy
hiePais(oT) = OO
7), Fal,
l= ig:
esi
Here Fy
=
Ft D), Bee
=
Bait D), Ong
=
GOpaleaO
ens
a iis (t, D) are some linear differential operators, the order of the operators H,. and Wp,
denoted by m,
the order of F,; and Vj,
is denoted by n,n
> m,
the superscript t at the operator indicates that the operator acts on the function
considered as the function of ¢ at a fixed T. As it is known
from
the linear differential systems
theory
(Section 1.2) the
integral terms which enter into (7.8.40) t
¥i= fSAal,o(X(),Den, t= Tp, to
P
Ie
a = [Xo Basn(t,r)dnes(X(0), Der,i=Typ, 7=1,4,
7.8. Applications.
Analysis of St6chastic Systems and Analytical Modeling
795
represent the solutions of the following equations:
P P Sure = Pree,
i ie
_
k=}
SESS OU (=k 2
pes Lae
a
r=A
’
(7.8.45)
atS Yodo Ware dna, 7 = Tp. (=e al
of their kernels satisfy (7.8.41)—(7.8.43). It is clear that here the nonlinear functions ~, =
p(X, t) and Wnei(X, t)
must be differentiable not less thanjn times. It is evident that if we introduce the vectors Z’ and €! of the dimensions p(1
=5 qN) x 1 and (p =F q?N) x 1 and the vectors U and € of dimensions pqN x 1 and q?N x 1 assuming
, gayi...)
,e=[b?...05]" N
.
(7.8.46)
Then the scalar Eqs. (7.8.45) may be written in the form
F'Z!' = H'e'.
(7.8.47)
Here the matrix differential operators F’ ‘and H’' are
F! = {Fi,(t,D)} =)
a,D*, r,1=1,p(1+ qN), =
(7.8.48)
k=0
H S4{HGG, Dyy= SeRD Srl hehe, k=0
where WE
= a,(t) and J, = Bx (t) are the (p(1 ap qN) x p(l AP qN)) and (p+ q’N) x (p+ q’N) matrix coefficients. Further we shall replace (7.8.47) by the set of equations in the Cauchy for according to the known rules of (Subsection 1.2.3): n
Gf =Big tae’, k= Tal, Zl = — oop ta 12}! + ane’ f=1
(7.8.49)
796
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
Here Zh are the p(l + qN )-dimensional vectors, h = I,n, dk = Ik (t) being the
p(l ap qN) x q?N matrix determined by the formulae k-1k—-h
Gk = a? in badSe 5
Cab son
nana ,
k=1,n—-1,
1 p=—t0) (0)
n-ln-h
—
Oe
c — pas a
ont
j
(7.8.50)
hk=0'1=0
Here the dependence between Z’ and €’ will exist. So
Z = Zi + gee’, 90 = Op Pn-
(7.8.51)
By virtue of (7.8.46) and (7.8.51), integral terms (7.8.44) take the form t
Y= fae T)p(X(r),7)dr = a’(Zy’ + qoe’), to
N Si
NS N = Sif Balt, 7)va(X(r), 7)dr = (>: s)(Z1' + qe’),
al
h=1},
pil
(7.8.52) where the following notations are introduces:
a’ = [150], 0) SeAy
1
[OLalle Aaah Or = Tom. foe Bgl.
(7s)
n
and Ip is the (p x Pp) unit matrix.
Consequently, Eq.
(7.8.40) together with Eqs.
(7.8.49), (7.8.52) for the extended state vector of the dimension np(1 + qN),
Za(xXtZ]", zits |Z... 2a") is reduced to
:
Al
;
Z = c(Z,t)+\(Z,t)W, Z(to) = Zo,
where
_ | (Xt) +a'[Z7/ + go(t)e'(X, t)]
oat) = ee yee) AZ) Seas
N
& u,[Z1' + go(t)e"(X,t)] 0
(7.8.54)
7.8. Applications.
Analysis of Sto¢hastic Systems and Analytical Modeling
7197
Zo = ([XT0]", y(t) =diag [ai(t)...qn(t)], ft) -=
I 0
0 I
Bt Stes
0 0
Saye
sae
Rieke
Keer
0
0
veo
a
—azta;
apes
—a7lan-1
—az lag
»
T=
Ip(i4qn)-
:
7.8.55
The matrices a’, bi, which enter into (7.8.51) are determined by (7.8.53); we = ~
are determined by (7.8.50), (7.8.51); €’ is determined by (7.8.46). The dimensions of c(Z, t) and I(Z, t) are equal to np(2 = qN) and np(2 + qN) xX q. Eqs. (7.8.50) are linear in Z” and nonlinear in AD Let us consider now the class of kernels satisfying the conditions (7.8.36) and the condition of sationarity (7.8.38).
As it is known from the theory of linear sta-
tionary systems (Subsection 1.2.5) in this case it is sufficient to suppose instead of (7.8.37) that the Laplace transforms of hereditary kernels A(é ) and ie (€) are rational functions of the scalar variable S, i.e. admit the representation of the form
| A@etae = F(s)-!H(s), | Balee-‘tag = ,(s)-! W(s). ; Here =
;
the
order
of the
matrix
polynomials
(7.8.56) H(s)
=
{H,i(s)}
{Wnri(s)} is equal to Mm and the order of the polynomials F(s) =
®;,(s) = {Prri(s)} is equal ton,
and
W(s)
{ F,i(s)} and
Nn> M. So Eq. (7.8.40) by means of extending
the state vector up to dimension np(2 + qN) may be reduced to Eq. (7.8.54). In conclusion we consider the class of the hereditary kernels satisfying the con-
ditions (7.8.36), (7.8.37) and (7.8.39). In such a case the order of the operators Fj and ®;,,; is equal to one, n =
1, the order of the operators Ay; and Whej is equal
to 0, m = 0. In this special case Eqs. (7.8.44), (7.8.45) take the following form:
[AG nex),
ar AY
Ye
ae rYl :
(7.8.57)
to Tt
J Baal X). rar ee DGU
|e
gaping zl Ural, shlley
)
to
Y=A-Y, Y(to)=0, Un = B, Un, Un(to) = 0.
(7.8.58)
Here At =
{Az (t)}and Ske = {Bi, (} are (p X p)- and (p X q)-matrices,
Meandsliz
=
Lage
eek
Une le are the p-dimensional vectors.
As a result
798
Eq.
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
(7.8.40) is reduced to Eq.
state vector
=
Z =
(7.8.54) for the extended p(2 at qN )-dimensional
ee yt Ut as if we
introduce
(pqN )-dimensional vector
U
[ee a Oe les the p(2 aP qN )-dimensional vector function c(Z, t), the (p(2
+qN) x q)-dimensional matrix-function I(Z, t) and the initial p(2+qN )-dimensio-
nal vector Zp assuming
a(X,t)+ ATY
c(Z, t) =
A-Y B-U
, (Z,t)=
(X,t)+ BTU
0 0
B® = [Be...BE], Zod XE0iOo tue oS wi(F-82539) The functions c(Z,t) and [(Z,t) are linear in Y and U and nonlinear in X. We emphasize that in the case of singular kernels the differentiability of ( and Wn is not obligatory.
So Eq.
(7.8.40) by means
of extending the state vector up to the
p(2 + qN)-dimension may be reduced to (7.8.54) under condition (7.8.59). Example
7.8.3. The scalar linear stochastic integrodifferential system t
oa,
+ Ge iSOGOU ld OW
SO Ie (1)
ON (do, 41, A, b, @ are constant) is reduced to the following linear two-dimensional stochastic differential system:
X = ap
+a,X + (A/a)¥
Example
+bW, Y
=a(X-Y), Xo=Yo=0.
(Il)
7.8.4. The scalar system with the parametric noise t
X =ap+a,X + |b+ Bf PC-Ox(r)ar
W, Xo=0,.
(1)
0
(do, a1, b, B, B are constant) is reduced to the following two-dimensional stochastic differential with the parametric noise:
X =ayp+a,;X +[b+(B/B)Y]W, Y = B(X -Y), Xo = Yo =0. (II) Example
7.8.5. The scalar linear nonlinear stochastic system t
SS Ga
a
feoeecxmyar +bW, Xo =0, 0
(I)
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
799
(ao, 41, @, 6 are constant) is reduced to the following nonlinear two-dimensional stochastic differential system:
X =ap+a,X+(1/a)¥ +bW, Y =a[y(X)—Y],
Xo = Yo =0. (II)
7.8.4. Normalization of Risk and Profit Financial Transactions In stochastic finance as it is known (Shiryaev 1999) there are widely used continuous and discrete stochastic models defined by the stochastic differential and difference equations.
A
Normalization of such stochastic models by
NAM gives the possibility to design
the specialized algorithms for various problem of the financial analysis and modeling.
The results of the statistical simulation given in (Artemiev et al 1996) for the following types of the stochastic differential equation
¥ =a0@)(Y"@ -—Y
] tea@xev
(7.8.60)
¥1 = aoi(t) [¥i"(¢) — Yi] + Bor (t)¥y" Vi, Yo = cvoa(t) [¥3(t) — Yo] + Boa(t)¥y7¥s"*Va,
(7.8.61)
Y3 = ao3(t)(Y3 — Ya) + Bos(t)¥3"° Vs, Y4 = aoa(t)(Y3 — Ya)
(a(t), Aoi (t), Boi (i), Y;* (t) being some known time functions) confirm practically good NAM
accuracy (5-10%) at 71, 62, Y2, ¥3 < 1.
7.8.5. Normalization of Queueing Systems An important class of the nonlinear stochastic systems is the class of the queueing systems. For normalization of the queueing systems it is expedient to decompose the queueing system into the typical subsystems and use the normalization method
to each subsystem separately. Because of the completeness of the class of the normal systems relatively to the connections the normal model of the queueing system is obtained as the result. As it is known any queueing system may be considered as a system consisting
of the following typical elements: the queueing channels, the waiting places in queue, the distributor D which determines
the direction of the motion of every client in
queue depending on the availability of the free channels or the places in queue with the account of the priority. The stream of the clients which is directed into the given
800
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
channel or at the given place in queue serves as the input of each channel and each place in queue.
The stream of clients served by the channel and the signal fixing
the channel (is it free or occupied) are the outputs of the channel.
The streams of
the clients and the signals of the feedback from the channels and the places in queue carrying the information about the state of each channel and each place in queue serve as the inputs of the distributor. The streams of the clients directed into the channels and in queue and the stream of the clients who received the refusals in service will be the outputs of the distributor D. Example of scheme of the queueing system with n
channels and without queue is illustrated on Fig.7.8.3.
Fig.7.8.3
Fig.7.8.4
We show how we may normalize the service channel, main element of the queue-
ing system (Fig.7.8.4).
Let y(c) be a probability density of the queueing time TJ’.
The input of a channel, i.e. the stream of the clients entering into the given channel, may be presented in the form of the sequence of unit impulses correspondent to the instants of the clients arrivals.
It is also convenient to present the output stream of
the clients in the form of the sequence of unit impulses which appear at the instants
when the clients leave the channel.
The other output of the channel which carries
the information about its state may be presented as the binary signal equal to zero in the channel is free and is equal to 1 if the channel is occupied. At such agreement
the channel will response on unit input impulse 6(t == ) which acts at the definite instant T by unit impulse é(t ==
T) at the random instant (7 + T) on one input
and by the signal equal to 1 inside of the random interval (rTee
T) and equal to
zero outside this interval on another output:
a(t) = 6(t—7), Vi(t)=6t-—7-T),
Yo(t) =1(t-7)1(T +7 -1).
(7.8.62)
Taking into consideration that the response of a linear system on unit impulse
represents its weighting function we shall consider queueing channel as a linear system with the following random weighting functions:
Gi(t,r)
=6(t-7-T),
Go(t,r)
=1t-—7)1(T+7 -t).
(7.8.63)
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
801
Then its response on any determinate input z(t) will be expressed by formulae t
t
¥i(t) = u)Si aT) e(n)det¥a(¢i=/1(T-+17—1)x(r)dr. (7.8.64) to
to
Now for the linearization of the regression it is sufficient to find the conditional
expectations of the outputs Yj (t) and Y (t) at a given realization x(t) of the random signal X (t). As a result we obtain t
E[Y,(t)|2£] = [e@ar
6(t— +r — 0) y(c)do = [ve — 7)x(r)dr, aS)
(7.8.65)
E[Y2(t)|z] = /x(r)dr [1(o+7 —t)y(c)do = [ve — 7)a(r)dr,
i
oe) where w(t) = (p p(a)do 9
:
(7.8.66) F(t) is the probability of the fact that queueing
t
time 7’ will exceed t. Formulae (7.8.65) and (7.8.66) show that the queuing channel may be considered as a linear system with the weighting functions p(¢ ) and wH(¢ ) and with the additive noises on its outputs (Fig.7.8.5).
Y,( Y(y
Y (0
yy
Y,(0
Y,()
Fig. 7.8.5 At the exponential WC ) =
e
distribution
#6 the channel may
of queuing
be considered
time 7’ when
p(¢)
=
peas ,
as the parallel connection
of two
aperiodic links with one and the same time constant |y)pt and with the gains | and 1 fpt with the additive noises on the outputs.
In case of determinate =
queueing
time 7' when
p(¢)
=
6(¢ =
Tey, ¥(¢)
1(¢ )U(T —¢ ) the channel may be considered as the parallel connection of the
802
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
ideal lagged link and the integrator with the finite memory
T on whose outputs
additive noises act.
Now it is sufficient for the normalization of the channel to average on all possible realizations a(t) of the normally distributed input X (t) the covariance function of
the vector output at the given a(t). From (7.8.65) and (7.8.66) after averaging we obtain t
Reh u(y tsa [ oloninge,¢) —r)Ce(7,7 + |t—7'|)dr to
-| | to
—r)y(t! —7')Te(7,7')drdr’,
Go
147)
Kyyyalo(tst’) = ff ot-nlae—t to
|
(7.8.67)
to
ve -7) Pela vara
to
(7.8.68)
tate
me
A
"[(otonaxt —7;t'—7'})—v(t—7T)¥o(' — 7) Pn (7, 7’)dr, to to
‘
(7.8.69)
where I’, (t, ts) is the second initial moment
of the input.
Now we show how we may construct the normal model of the distributor the one-dimensional queueing system without queue.
D in
In this case the distributor D
represents a determinate system with two inputs and two outputs. The stream of the clients enters on the first input, and on the second input — a binary signal about the
state of a channel.
If the second input is equal to zero then the distributor directs
the stream of the clients in queueing channel.
In this case the first output of the
distributor coincides with the first input, and the second one is equal to zero. If the second input is equal to 1 then the distributor gives the clients a refusal and directs
an input stream from the system.
In this case the first output of the distributor is
equal to zero, and the second one coincides with the first input. Thus the dependence
between the inputs 21 (€); LQ (t) and the outputs Yj (t), Y2 (t) of the distributor with the one-channel queueing system without quence has the form
yi(t) = a1 (t)[1— xo(t)],
yo(t) = x1 (t)ro(t).
These formulae show that the distributor
(7.8.70)
D represents the parallel connection
of two elementary determinate nonlinear inertialess links.
For the normalization of
the distributor it is sufficient to linearize dependences (7.8.70).
We may do it, for
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
803
example, by the Taylor formula expansion of Yj (t) and Y92 (t) in the vicinity of the
expectation of the vector of random inputs. As a result we get
Y(t) = [1 — me, (t)]X1(4) — me, (1)X20), Y2(t) = mz, (t)X1i(t) + me, (t)X9(t). Example
7.8.6.
(7.8.71)
For testing the accuracy of the approximation of the
queueing systems by the normal models we use the stated method to the simplest
one-channel
queueing system without
queue.
Supposing the distribution of time
queueing as an exponential one we obtain a structural scheme
of the considered
system (Fig.7.8.6). Here D is an distributor which determines the movement direction of each next client; jl /(u + s) and | /(Wise s) are the transfer functions describing
the distribution of time queueing; Y (t) and Y> (t) are the input and the output. The equations which describe the behaviour of the system have the form:
ve + (mz + wy) = Me_My, — PM UN + eS My,)X,
Y= Y{+ Yi", Yo = (my,/p)X + (me/p)(Yi — my.) + meV",
(I)
(II)
Fig. 7.8.6 where U" and ee are the normal noises with zero expectations.
Hence we obtain
the equations for the expectations of the signals:
ty, + (me + H)My, = WMz,
My, = (My, /pL)Me.
(IIT)
The stationary (steady state) values of the expectations of the outputs are determined by formulae
My, = pm, /(Ms + f) ;
My. = mz /(me + p).
(IV)
804
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
In the special case of the Poisson input stream the expectation M,
is equal
to the intensity of the stream \ and the obtained formulae give the known
exact
solution. For the determinate time queueing J’ the structural scheme of a system is rep-
resented in Fig.7.8.7. Consequently, the equations of the system have the form:
Yaviy,
V@=]Z6=T),
t
U(t) = i inde
Z=0— ms — ma",
(V)
ee
t-T
Fig. 7.8.7
Hence we get the equations for the expectations of the signals
negr(t) Tag ty
mgt Poms
J
= (lig
ts!
(V1)
Mz (EJOT METily, = Thy lta
In the steady state Eqs. (VI) give the opportunity to derive the following expressions of the constants My,, My,, Mz
y=
Mg = tin fT
and My:
arin (le ig) a tee re
Le)
VL)
Solution (VII) also coincides with exact one in the case of the Poisson input stream. Now we notice that at the exponential distribution of time queueing the mean time
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
805
Y,()
Fig. 7.8.8
queueing is equal to 1mipt. Therefore the formulae obtained for the case of the determinate time queueing coincide with the formulae at the exponential distribution of time queueing. Example
7.8.7. Let us estimate the mean number of the served clients
who received the refusal and the variance of the clients number on the outputs of the one-channel queuing system without queue (Fig.7.8.8). Here D is the distributor which determines
the direction of the movement
of each next client in dependence
of the fact whether the channel is free or occupied; p(t — T) and 4)(t — T) are the weighting functions connected with the probability density plo ) of time queueing 7' ioe)
by the relation (co) = if p(¢)d¢; X (t) is the input stream of the clients into the
a system; Yj (t) is the stream of the served clients; Y> (t) is the stream of the clients who received the refusal and J denote the integrators whose outputs represent the
numbers of the served clients and the clients who received refusals beginning with the instant fo till the instant f; Y{'(t) and U"(t) are the inner noises of the system
which are determined by formulae (7.8.67)-(7.8.69) at
X = Z:
min{t,t’}
Kyr(t,t’) =
/
y(min{t, t’} — r)T (7,7 + |t —t'|)dr
to
i |/p(t — r)p(t! — r')P.(r,7/)drdr’, to to
(I)
806
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
he
Kyunltst) = ff et—niae—t -1+7)— 0 to
Feed’
to
(II)
t
v(t! —7’) | 2(7, 7’) drdr’. quiet) = J f[(omaxte-r t!—7})—Y(t—7T) to
to
(111)
Thus the set of the equations which describe the performance of the queueing system has the following form:
Z(t) =[1— mu(t)] X(t) — me (t)U° (A),
(IV)
Yo(t) = my(t)X(t)+ mz (t)U°(t),
(V)
nit)= [et=n)a(ndr+ Hi",
(v1)
U(t) = /W(t = 7) Z(r)dr + U"(2),
(VII)
where U zie) is a centered random function; Mz (t) and My (t) are the expectations of X (t) and U(t). Accounting that My”
= My
= 0, we get from Eqs. (IV)-(VII)
the set of the equations for the expectations:
m,(t)
=[1—mu(t)]me(t),
t
nae pER
my,(t)= mu(t)me(t), t
EO
/sbtaSaaseeoee) ovisNAT)
to
to
Hence in the steady state for the Poisson stream Mz (t) = X putting tg = —oo we find
t
Mz
Oita
Asner
= X / b(t—r)dr,6Tt Sa ees —00
Taking into consideration the relations
t
oe)
/ y(t —7r)dr = | e(o)do ='1, —00
0
(IX)
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
807
|w(t — r)dr = | oyae ™ [o~ W(c)]do =jaceiape
vio)= f eon
(X)
from (IX) finally we obtain
mz = my, =u/(A+p), my =YO+H), my, =2/A4+n). (OM) Thus the normalization method‘for the steady state gives the exact values for the expectations of the outputs at any distribution of time queueing plo iN
At last we give an approximate the steady state of the performance.
estimate of the variances of the outputs for In this case it is convenient
to present Eqs.
(IV)-(VII) for Y; (t) and Y2(t) in the following form:
NO=NO+VO-VN),
HOH=V~O)+ Vo"),
Yi) = F(D)X(), Yi") = neat Yo(t) = P(D)X(t),
— (XM)
cam
Yo"(t) = Q(D)U"(é).
Here 1
weed
My
®,
D
r~ ey
My,
a
fiietak
j
Q(D) ~~
_ My + mz2(D) BU) where D =
socnet
We m, ®2(D)
-
d/dt; 0, (D) and ®,(D)
D
1+ oD My
XIV (
)
T+ Mm, ®2(D) :
are the transfer functions correspondent to
the weighting functions p(t = T) and w(t = T) respectively. It follows from the properties of a normal stochastic system that the input X(t)
is uncorrelated with the inner noises Y{"(t) and U"(t) and it means that Y/(t) is also uncorrelated with ¥(@) and with VK)! Analogously Y(t) is uncorrelated with Ys ey
Consequently,
Kyi ty
By Gt') + Ky?)
+ Kyu (t, t’) — Kyityy (t,t’) — Kyity! (t,¢’),
(XV)
Ky, (t,t’) = Ky (t,t!) + Kyr(t,t”).
(XVI)
808
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
The variances of the total numbers of the clients Vj (t) and V2(t) are determined by formulae
Die)= //lees Ca to
eee
(XVII)
to
In the special case when X(t) is a white noise of the intensity A, Ky, (t, t’) =
Ad(t — t) applying the formulae of the linear theory of the stationary systems
(Subsection 6.1.8) to Eqs. (XIII) we obtain
hu
'
OO) = apeaiarn eu r3
AF
(20 + 3p)
(XVIII)
_
Bx
GIR)
;
(XXI)
Kyl) gal Gy 0) + eg Ot CaP IG etryAbame mua aeenye Ap
Kan
Kyu
(te)l
dp?
Bed
iM
=
ee pe 204
! —(A+p)|t-t
Ae i
—-—
De por
———
a
Ait
,
—(A+p)|t-t'|
!
;
(XXII)
OR a owe
(A+ 4)? 4(X + p)5 = ATH? tle" QtHMIt-t1 4+(A+uy2) jé—t— te
Kyyi(t,t’)
|
j
se SS
eee rate meatih jt bead —t' le leAPS
Kea
ay
dp?
Op?
————
Dd K, m(t
as
=
Ale "404
Nee
eelray a lt TO
,
Ph 2
es t|
Atal
—t’ leee
’
(XXIII)
:
(XXIV)
Accounting (XV), (XVI) and (XVIII)-(XXIV) after evaluations we find from (XVII)
D,,(t) =
a) ATH yo, (A+ p)4
Ani oy, (He.3 + Ant Dies — d*) (A+ y)°
7.8. Applications.
Analysis of Stochastic Systems and Analytical Modeling
809
¢ r? pw?(2A + 14A? 0+ 19Ap? + 10p3)
2(A + ps)? TANS _A pr (4A* + TAu + 4p*)
tm
2(A + p)®
AP?22/93 (23 + 14020 + 19? + 10u*) |-atnye
2(A +p)? My?
Dy, (t) = SP
0g
2+ We
(A+ p)° (tu
Ad? p(4r? + 13d? + 14Ap? + 83)
200+ ny"
N38(403 + Se 132 0+ 14Ap? + 83) mere rT! + ee A+ py" ee e From (XXV) at t —
(XXV)
AB(A4 + GAB + 137py?+ 12Ap3 + 5y*)
(A+ #)
AMPA + 2H)
}
(XXVI)
oo for Dy, (t) we obtain D,, (t) = BEY (t) = My? /(A+p)4,
i.e. the variances of the number those clients who are served during time unit and those who are received the refusal are equal and do not depend on time.
7.8.6. Software for Nonlinear Stochastic Systems Analytical Modeling We see that the theory of stochastic systems given in Chapters 1, 5—7 is rather
complicated and requires considerable mathematical background for studying it. To solve any practical problem of stochastic systems analysis one has to derive the respec-
tive equations determining distribution parameters from original stochastic equations and then to produce a program for solving these equations. This is a tedious and time-
consuming job which can be done only by those who have an extensive knowledge of mathematics and some skill in programming.
At practice two main approaches to the
software design for the distributions parametrization methods in nonlinear stochastic systems are usually used:
@ the creation of the specialized libraries for different programming languages; e@ the creation of the specialized dialogue software packages. The library approach is effective for the analytical modeling and the automatic
derivation of the equations for the distributions parameters and for the statistical simulation of linear stochastic systems also.
The choice of one or another method
is determined by available information about numerical model of a real system and
random disturbances, required accuracy of the calculations of system efficiency and by the information technologies of research also. For the nonlinear stochastic systems the derivation of equations for the distribution parameters is rather cumbersome and
810
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
accounts about 90% of the whole process.
In this situation it is necessary to create
the dialogue packages which realize the main operations of the nonlinear analysis.
In recent years actual prerequisites for further improvement
of the software
appeared on the basis of using modern highly effective software of data processing
and information storage. One of the directions of the development is connected with the implementation of the mathematical packages which possess the wide possibilities
of analytical derivation such as “Mathematica” used functions.
what allows to expand the classes of
The other direction of the development
is the implementation
of
the modern object oriented control systems of the\data storage of equations of the
intermediate results. The package “StS—Analysis, version 1.0” is designed for studying the stochastic
differential systems described by Eq. (5.4.6) with the standard Wiener process W(t) by NAM
for the polynomial in y function a(y, t) of not higher than the third degree
and the function b(y, t) independent of y or linear in y. The package “StS—Analysis, version 1.0” has been widely used in Russia since 1987. The package “StS—Analysis, version 2.0” is destined for studying by NAM
the stochastic systems described by the
difference equations.
Working with any version of the “StS—Analysis”
package, the user must
only
type on the display with the aid of the keyboard the original stochastic equations in the natural mathematical
form without
any special programmers
tricks, intro-
duce the corresponding initial conditions and fix the statistical characteristics must
be calculated.
Once the original stochastic equations are introduced the user may
correct them, i.e. to remove some equations, to replace some of them by others, to
correct errors and so on. When all the data are introduced and checked the user has to start computations. NAM
Then the package “StS—Analysis”
automatically derives the
equations for the corresponding distribution parameters and solves them. The
user may observe visually the behaviour of a system under study in the process of
calculations by tracing the graphs or the tables of current values of various statistical characteristics on the display and get the copies of these graphs and tables from the file after the calculations are finished. So the use of any of the versions of the package “StS—Analysis” does not require that the user know mathematics and programming. Any
version of the
“StS—-Analysis”
package
automatically
derives
the NAM
equations for the distribution parameters from original stochastic equations introduced by the user in the natural mathematical form and solves these equations.
It should be remarked that the “StS—Analysis”
package does not simulate a
stochastic system and does not use the statistical simulation (Monte-Carlo) method but performs the theoretical determination of necessary probability characteristics of a system behaviour by the normal approximation method.
7.8. Applications.
Example
Analysis of Stochastic Systems and Analytical Modeling
811
7.8.8. Let us consider the scalar stochastic nonlinear discrete
system
Yi (0+ 1) = —aYP(l) + ¥i(0) + dV, (I).
(I)
Formulae for the expectation and the variace have the form
my (I+ 1) = —aBY (1) + mi(2),
(II)
Dy (+1) = @(BY(I) — BYA(NEY2(0) —2a( EY} (1) — m, (NEY 3(I)) + Dy (1) + 82.
(IIT)
The higher moments approximately are expressed in terms of the first and the second moments:
BY}(I)= Di(l) + mi(0),
(IV)
EBY3(l) = mi (I EY?(l) + 2mi(1)D, (0),
(V)
BY (0) = mi(NEYA() + 3D, ()BY2(),
(V1)
BY?(l) = mi(IEYA()) + 4Di (QEYS()), EY£(l) = m (EY?(1!)+ 5 Di (NEY; (0. On Fig 7.8.9 graphs of m; and D, for a = 0.05, at the initial conditions m,(0) =
(VII) (VIII)
b = 0.05 anda
= 0.1, b = 0.05
lke D,(0) = 0.5 obtained by NAM
a) a=0.05; b=0.05
are given.
b) a=0.1; b=0.05
1
7
0,5
0,5
m,(l)
m,(1) DY,
0
0
D,(I)
=
100
200
300
400
|
0
0
50
100
=
150
200
250
|
Fig. 7.8.9
Example
7.8.9. Stochastic three-dimensional nonlinear discrete system
is described by the following equatons:
¥i(14+1) = Vi(1) + a¥i (0),
(I)
Y2(1 + 1) = Yo(1) + 2aY3(I),
(II)
Y3(1+ 1) = Y3(1) — aYo(1) — 2aY3(1) — aY3(l) + 3a¥ (1)Vi (0)
(III)
812
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
at the initial conditions
EY,(0) = EY2(0) = 0.5, EY3(0) = 0.8
DY;(0) = DY2(0) =1, EBY;(0)¥2(0) = 0.4. Other variances and covariances are assumed equal to zero.
tions and the variances obtained by NAM
ie
Graphs of the expecta-
are given on Fig.7.8.10.
a) a=0.01; b=0.02; c=0.03
b) a=0.01; b=0.02; c=0.03
D3(I) 0,5 4
D,(1) ; Da(!)
Fig. 7.8.10
Example
7.8.10.
Let. us consider the stochastic nonlinear differential
equations of the chemical reactor:
Y; =14+ Y/Y. — (14 Y3)%1 + 0.5v, Yoo Vosy (Yack Ve, Y3 =
—-Y; Y3 == C,
(I) (II) (IIT)
where V; and V2 are the normal mutually independent white noises with the unit
intensities,
C being some constant. The graphs of the expectations, the variances and
the covariances at
c=
1.3 and the initial conditions
EY,(0) = 0.5, EY2(0) = 3.3, EY3(0) = 3.0, DY; (0) = 0.01, DY2(0) = 0.5. DY3(0) = 0 obtained by NAM
are given on Fig.7.8.11.
(IV)
7.9. Online Conditionally
Optimal Nonlinear Filtering
813
Fig. 7.8.11
7.9. Online Conditionally Optimal Nonlinear Filtering 7.9.1. Preliminary Remarks The simplicity of the Kalman and the Kalman-Bucy filters (Subsection 6.6.2) leads to the idea of conditionally optimal filtermg. This idea is to reject the absolute (unconditional) optimality and to restrict oneself with finding the optimal estimate in some bounded class of admissible estimates satisfying some rather simple (differential
or difference) equations which may be solved online while receiving the results of the measurements.
The main problem of applying the conditionally
optimal filtering
theory is to define a class of admissible filters. It is recommended usually to define the class of admissible filters as the set of all filters governed by the differential or the difference equations of the prescribed form with some undetermined coefficients. The problem of optimization in this case is reduced to finding optimal values of all undetermined coefficients which appear generally time functions.
814
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
The first characteristic feature of the nonlinear conditionally optimal estima-
tion problem is that it is essentially the multi-criteria problem since m.s.e.
E |X (t)
_ X (t)|? is desired to be minimal at any ¢ in some interval. So there is an infinite (noncountable in the case of continuous time) set of criteria in this problem.
For-
mula (5.10.6) gives the absolutely optimal estimate which is optimal at any t. But
restricting the class of admissible estimates we cannot hope to find an estimate which would be optimal for any t. Only in the linear problems this is possible.
Thus one
has to restrict oneself only by the Pareto optimal decision. The second characteristic feature of the conditionally optimal filtering is that the optimal values of the coefficients in the equations of a filter are determined only by prior data without using measurements.
So the optimal filter may be designed beforehand quite in the same
way as the Kalman or the Kalman-—Bucy filters. The measurements need to be used
only in the process of filtering while solving the equations of the filter. Let {X i} be a state sequence of random variables.
Another measurement
se-
quence of random variables {Yi} is defined by the equation
Y; = w!(X1, Vi),
(7.9.1)
{Vi} being a sequence of independent random variables with known distributions.
In
problems of practice the random variables V; cause the errors. This is the reason why
they are usually called measurement
errors.
The problem is to find an estimate X;
at each | of the random variable Xj] using the measurements of the previous random variables Y; , ... , Yj—-1 in the class of admissible estimates defined by the formula Xx = AU,, A being some matrix of the rank n, and the difference equation
Vig = 01001, Vi) + 1
(7.9.2)
with a given sequence of the vector structural functions Ci(u, y) and all possible values of the matrices of coefficients 6; and the coefficients vectors yi.
Any given
sequence of the structural functions ¢] (u, y) determines the class of admissible filters. Any choice of {6}, {yi} and the given functions C(u, y) determines the admissible filter. The dimension 71 of the vectors Y; is supposed less than the dimension 7 of
the vectors X;, 1
(0) DX (0) DX,(0)
= = = =
EX;(0) = 4000, EX2(0) = —20, 160000, DX, = 10000, DX2(0)=1
(VIN)
are given on Figs.7.9.1-7.9.3. The expections, the variances and the covariances of the processes X; (I) (2 ale 2) are given on Fig.7.9.1. The variances and the covariances of X;(I) are practicaly identical. Graphs of a11(I), a12(1), ai(l), a22(I), Bx(0), B2(1), ya (1), y2(0) are shown on Fig.7.9.2. The results of statistical simulation the processes AGAl), X;(l) (i Nh
m.s.e.
is
does not exceed 1%. For the linear Kalman filter (at @22 = 0) maximal m.s.e.
is
equal to 5%.
2) are presented on Fig.7.9.3.
Maximal
824
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
0
4000
8
16
-20
3000
-50 4
EX, 130J
950000 4 750000 +
550000 + 350000 + 150000
32
at
24
32
ft
Problems
825
Problems 7.1. Using formula (5.1.16) show that in the case (5.4.7) the function o(Y, t) in formulae (7.1.3), (7.1.5) must be replaced by ((Y, t),
dbAieaateau /RNa Rs
nee du),
(1)
Chapter 7 Methods
826
of Nonlinear Stochastic Systems Theory
o(Y,t) = b(Y,t)vo(t)b(Y, t)”. 7.2.
Show
that for Eq.
(II)
(5.4.6) in the case of an arbitrary process with the
independent uncrements W(t) formulae (7.1.3) and (7.1.5) are valid if we assume
0? x(A3t) Ht) =bald) =|aa t) =[v,,(t)]
=
|——— =
()
|
I
where x(A;t) = 0/0t\n hy (A; t), hi(A;t) be the one-dimensional characteristic functions of W(t). 7.3.
Show that the infinite set of equations for the moments in the case Eq.
(5.4.6) where a(y,t) is defined by (7.4.1), b(y,t) = b(t), V =
W, W
being an
arbitrary process with the independent increments has the following form: p
N
Qk, ..,kp = S k, pS
y Pry nce
ky
OF het paglig ha bhi tees lcs bes eeig SEs 5S)
ky
+ > “i = CE 0 hy=0
hp=0
(ki,
php
where
kp =
=O, 1 R236. 3 kak rn
Nhvac a
h
(I)
Ca Xbuy Jp Ubi —hi yey by—hy
Qhrs
Be
pee sae
152,
ys
(het)
eo
|.0
Me
Also show that Eqs. (I) may be written in the more compact form using multi-
indexes k = [ei
8 ke) pr Chik dhs |, Cu en P
a
(pail
(eat.
N
SS Ke
) Kp
eogoK kp
k
Ar hOht+k—e, + > Chxnoe-n h=0
(IIT)
h=0
0, eis
|k| =
kj +----- k, aby
ay
where €, is the vector whose components are equal to zero except the rth component a
G
which is equal to the unity, €, =
[Mere rr
Ok = Wey... bys Gh = Gry
ae ?
skp
XS he
hes
(IV)
Xn = Oat |h| =hy+---+hy < 2 and yp = Ys at h = e, + €,. In Eqs. (III) Qs is equal to zero if at least one of the components of the multi-index $ is negative,
Problems
827
and is equal to unity if all the components
of the multi-index s are equal to zero,
ao
=
il
7.4. For the linear system (1.4.5) with the parametric noises N(t) considered as the system with the 6-differentials write out the correspondent It6 equations assuming
that the white noise is normally distributed. 7.5. Consider the linear system with parametric noises
Ag+ (1+ Va)(B + BY)q+(1t+Va\(C+C')g=(1+NU)Q*, where g =
[q eer if and.
=
(1)
Kae ye es (ii are the n-dimensional vectors;
A, B, C are the symmetric matrices; B’, C’ are the antisymmetric n X n-matrices; a—
[Vi V2 V3 ‘ea is the normally distributed white noise of the intensity ”. Show
that to Eq. (I) corresponds the following Ité equation:
Ag+ (1+ V3)(B+ B’)g+ (1+ Ve)\(C+C')q = (14+ Vi)Q* + 6[-1,3Q*A7'(B — B’)
+023(C + C’)gA~*(B — B’) + v33(B + B')gA~"(B—B')]. 7.6. Considering the equations of Problem 1.14 at IT
vw = ¥(q,t),
(I)
= ~V, Q = pV where
¢ = (4,9, t) are matrix functions of the stated variables, V is the
normally distributed vector white noise of the intensity l/ as the equations with the
0-differential, reduce them to the Ité equations for the canonical variables. 7.7.
Show
that for the linear stochastic system with the parametric noise of
Problem 5.5 (i) equations for the expectation, the variance and the covariance function have the form vis
—e(1 =
20€V22)m
=
2€0kr42,
D = —2e(1 — 20€v22)D + kh? + €7¥22(m? + D) —2ekmry2, OK
(ti, t2)/Ote a
—e(1 =
20€V22) K (ty, te).
(1)
(II) (I)
Find the stationary values of m and D and the conditions of the stationary regime existence. 7.8. Show that for the linear stochastic system with parametric noises of Prob-
lem 5.7 (i) equations for ™ and K have the form .
Mm,
.
=
m2,
—2m2Cwo(1
2,
M2=
—m,w6(1
ca 20CwoV33)
oe 20CwoV23)
=
20kC€wo9133
(1)
ki; = 2k12, kyo = —w2(1 — 20Cwove3)k11 — 2¢wo(1 — 20Cwov33)k12 + k22,
828
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
hea = —Ww(1 = 20CwoV23)k12 _ 4Cwo(1 — 20¢woV33)k22 + k?v44 —2m,kwi42
- 4m2Cwokv43
+ w6V22(m?
+ ky)
+4Cwevo3(mi me + ki2) + 4¢7w2 nugs(ms + ko2). 7.9.
Using formula (5.1.16) show that Eqs.
(5.4.7) at functions 1, (2, 21
(II)
(7.2.3), (7.2.4) are valid for Eq.
defined by (7.2.5)—(7.2.7), at Yo2 = Eno and o
and ¢ defined by formula (I) of Problem 7.1. 7.10.
Sow that NAM
equations for stochastic differential system (5.4.6) in the
case of an arbitrary process with the independent increments W
coincide with the
equations of Subsection 7.2.1 under condition (I) of Problem 7.2. 7.11. Derive NAM
7.12. Derive
equations for nonlinear stochastic discrete system (527-2):
NAM equations for the continuous-discrete system defined by Eqs.
(5.7.10), (5.7.11). 7.13. For the typical composite functions derive formulae for (Yo for NAM
given
in Appendix 3. 7.14.
Show
that
NAM
gives
the
exponential
covariance
function
k(r )
= De~°!"l for the stationary process in the system of Example 7.2.1 where (i) De
(v/6b3)'/?, a=
(i) D =
(3/2vb3)!/?
(V2mv/8b2)'/?,
wa =
at ply) =
b3y°;
(8/m)'/2(V/2rv/8be)*/3 at v(y)
= boy” sgny. 7.15. Using the equations of Problems 7.14 in the case (i) deduce the formula for the standard deviation 7.16.
0 = VD for an arbitrary instant.
Show that the expectation
the system
Mm and the variance D of the state vector of
TY +Y_= Mea’),
26g,
(V being the stationary white noise of the intensity ) obtained by NAM
(I) are deter-
mined by the equations
Tin = bmg — m[1 + ab(m? + 3D)],
T?D = bv — 2DT[1+4+ 3ab(m},)].
(II) Find the moments method values of m and D for the regime of stationary oscillations. Using MM
with the account
of the moments
whether or not this method rejects the obtained
7.17.
Using NAM
of up to the fourth order verify
m and D as a superfluous solution.
show that the variances and the covariance of the state
variables in the stochastic nonlinear system
TY, + Yi = bsgn(Y2—-Yi),
Yo+aY2 = V2DaV,
(I)
(the expectations of the initial values of Yj and Y2 being equal to zero, V the white
noise of the unit intensity) are determined by the equations
Thy, = 26ky2 + 2(14 A)ki1,
Tkie = Bkoo — (1+ aT + B)ki2,
Problems
829
koo = 2a(D— kao), 8 =b[ (ki: + boo — 2ky2)e/2]-2/2. Find k11, k12, ke for the stationary regime. moments
Using MM
(I)
with the account of the
of up to the fourth order verify whether or not this method
rejects the
obtained variance kj for the stationary regime as a superfluous solution. Solve this problem also in the case of the nonzero expectations of the initial values of Y; and
Y>. 7.18. Using NAM
show that for the system
TY +Y =bsgn(X -Y), X +aX = V2DaV,
(I)
V being the normal white noise of the unit intensity, NAM equations for the variances and the covariances of the random variables Y;
= Y, Yo = Y, Y3 = X have the
form
kyy =2k12,
Thig = Tho — k12 + B(kis — k11),
ki3 = keg —ak13, Tko2 = 2keo + 28(ko3 — k12),
k33 = 2a(D — kgs),
Tko3 = —(1 + aT )ko3 + B(k33 — k13),
6 = b[ (ki1 + kag — 2k13)/2]-*/?, if the expectations of the initial values of Y;, Y2, Y3 are equal to zero.
(II) Find the
stationary values of the variances and the covariances. 7.19. For the stochastic nonlinear oscillator
Y +0(Y,Y)+w2¥.=V,
(I)
where V is the white noise of the intensity / show that the variances and the covariances of the random variables Yj} = NAM
Y, Yo = Y are determined by the following
equations:
kay = 2kia, ki2 = ko — (we + ki)kia — kokie, hoo =
—2(w2
+ ki)k12 _
2koko2 =e VY,
(11)
where k; and kg are the statistical linearization coefficients of the functions p(y ; y2); the expectations of the initial values of Yj, Y2 being equal to zero.
7.20.
Using NAM
and the results of Problem 7.19 verify the validity of the
following approximate formulae for the variances and the covariances of the stationary process in the systems:
(i) the oscillator with dry friction 9(y1, y2) = bo sgn yo, ki
=
mv? /8w2b?,
ki2 =
0,
koo —
we kit;
830
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
(ii) the Rayleigh oscillator Y(y1, y2) = —biy2 + bays, ki =
(1/6w2b3)(by
-F ,/? + 6vb3),
ki2 =
0,
ko = w2ki1;
(iii) the Van der Paul oscillator p(y ‘ y2) = b(y? = 1)y2,
ky = (1+
4/1 + 2v/w2b)/2, big =0, kag = wk.
7.21. Using formula (5.1.16) show that the equations for the initial moments in the case of Eq.
(5.4.7) may be presented in the form (7.4.3) if we add to function
r,v defined by (7.4.4) the following item: (oe)
Agr = [ f{tm+aqenr sel Yptep(y, tu) )*
singe 8
P ees » Oe
TOs ook: pretuty)}
s=1
xup(t, du)wi(y)pr (y)dy.
(I)
7.22. Show that the equations for the initial moments in the case of Eq. (5.4.6) for the nongaussian white noise have the form (7.4.3) at
vi girl Aaa a / aan -. O(iAp)”®
x [47 a(y,t) + x(b(y,4)? A;2)] es wildy, v3
hod
(I)
glrl
‘ |ata —oo
ie
Pe,
ot
x [id a(y, t) + x(b(y, t)” A; t)Je! "| Py(yjwi(y)dy.
(II)
A=0
7.23. Using formula (5.1.16) show that the equations for the central moments in
the case of Eq. (5.4.7) have the form (7.4.6), (7.4.7), (7.4.11), where 1,v is defined by (7.4.8), 2,v is defined by (7.4.9) at 0, substituted by O of Problem 7.1 and Priv is defined in Problem 7.21.
Problems
831
7.24. Show that the equations for the central moments in the case of Eq. (5.4.6)
with the nongaussian white noise have the form N
™41,0,h = ¥1,0,n(m, K,t)+)>
wa Ove RU, 1g) Qi
teel
ca
D)
k=3 |v| =k
(I)
P
Hr = Pr,o(m, K,t) = > ThP1,0,n(™, K,t)br—ey =
N
P
a8ne I=3
en(m, K,t) —S> TrPivva(m, K,t)Mr—e, |qv(a) ||v|=l
(ID)
h=1
CAS |S
ae
a
eeeae
where
BS stan,Korey / Ree nere
(111)
penktn, it) = i an(y,t)pv(v)wr (y)dy ,
(IV)
Yr,o(m,
Bat)
Me
/ onal fRtrcnaer
ona
ae
a(y, t)
+ x(H(yTAsnjer™e-m | wi(y)dy,
(V)
A=—0
Veet, K,tet)
/ at Ae a(y,t) (At onaemcceas
H (HUNT AsnIEeo-—} OURO
aaah
A=0
7.25. Derive the equations of Subsections 7.4.1, 7.4.2 for the initial and central moments in conformity to stochastic discrete system (5.7.2). 7.26. Show that the equations for the multi-dimensional initial moments in the
case of Eq.
(5.4.7) have the form (7.4.15) with Yq, |...
kajvi,... apelin’ Ke, tn)
containing the following additional term: AGr, ,...,&ni¥1 spit yam
aiin, tn)
832
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
(oe)
(oe) Vi1
Vip
Vn—-1,1
aie of fatness —0Co
or
noe
*Yn—1,1
Re
4{am ot c1(Yn; tn, u) lw:
[Yap rh Cp(Ynytnyt@)3” = Inks Une? fi > UnsYni x eles
ROACUT ON 7.27.
byeltndu) 2 Yap? Cs(Yns tn, uw)
GaP ey aca 5 hls Ue JaUL
1a
(I)
Show that the equations for the multi-dimensional moments
in the case
of Eq. (5.4.6) with the nongaussian white noise have the form co
Gr
aireltisstn\lOn=
Te
[o.e)
ff
ee COM
girl Lae
aS
eel ete | > 6)
xy? ey
ie
es5 wi : ye1
Init eee
Se Onidgr
ag, © #4)
7.28. Derive the equations for the multi-dimensional moments in the case of Eq.
(5.4.6) for the nongaussian white noise. 7.29. Derive the equation for the multi-dimensional central moments in the case
of Eq. (5.4.7). 7.30.
For the system of Example 1.3.3 in the case when the vector of random
disturbances N(t) =
[Ni (t) No (t) ig is the normal two-dimensional white noise
with the matrix of the intensity / show that the moments
@,;
of the process y
= [Y1 Yo ie are determined by an infinite set of equations for the moments: : Ops
= =
TOpr—1,s41
2 — SWoAr41,s—1
—
4 2ESQy,s + ViQwor(r =
Papa
1 +5wo8(s ae 1)(1107,5~2 - 21120r41,5—2 ae V220r42,s—2)
abet Vi ipa eset fd)
es
pee
7.31. Under condition of Example 7.2.0 at y(y) = y? show that the equations for the first three semiinvariants m, D and K3 have the form
m=—Kk3—3mD—m',
(1)
Problems
833
D= =6mk3 — 6m?D — 6D? +1,
k3 = —9k3m? — 18m3D — 27Dkz3.
(II)
7.32. Prove that for the stochastic two-dimensional system
Yj = ag—a¥, +hiVy,
Yo
SHY
ove
(I)
(ao, a, hy, hy > 0, V; and V2 are the independent normal white noises with the unit intensities) the equations of the semiinvariant method with the accuracy till the
first four semiinvariants Kj; (t, 7 = 0,1, 2,3, 4) have the form
Ko1 = —4K19
+49, K10 = Kioko1,
Ko1 = —2ak29 + hi,
(1)
K11 = —@K11 — Ko1k20 — K10K11 — Ka1, k20 = —2(K11K01 + KioKo2 + K12) + h3, K30 = —3ak30, Kao = —4aKk40, Ko1 = —2ak21 — Ko1k30 — 2K20K11 — K10K21 — K31 ,
(II)
k12 = —ak12 — 2(Ko1K21 + K20K02 + K7, + K10K12 + K22) ;
Ko3 = —3(Ko1k12 + 2K11K02 + K10K03 + 3k13),
(IH)
k31 = —3ak31 — Ko1Kk40 — 3K30K11 — 3K20K21 — K10K31, K22 = —2ak22 — 2(Ko1K31 + K30K02 + 3k21K11 + 2K20K12 + Ki0K22),
K13 = —@k 13 — 3(Ko1K22 + 2K21K02 + K20K03 + 3k11K12 + K10K13) , koa = —4(Ko1k13 + 3k12K02 + K10K04 + 3K11K03) Show that at dg >
(4) Kg
(IV)
hia disappears the difference between the values Ko), Ki),
calculated with the accuracy till the semiinvariants of the second fourth orders. 7.33.
Derive
the equations for the first two moments
of Problems
5.24. Show that the solution of these equations at the initial conditions Mp
5.23 and =
EYo,
Do = DY? has the form
a(t) = a10 + @11 exp(—at), ao(t) = a29 + a1 exp(—at) + a22 exp(—2at) , where @19 =
= @o/@; Q11
=
ayi(1 + 2aoa1)/aaj; 92
Mo — Q10; G29 =
(1 + 2apaj)ao/2aia*;
a1
= Do — G20 — M21. Find the stationary solutions.
7.34. Show that the equations for the coefficients of the orthogonal expansions in the case of Eq. (5.4.7) have the form of Eqs. (7.5.8), (7.5.9) at co
y20(m, K,t) = /[a(y,t)(y” —m™) + (y—m)a(y, t)” + a(y, t)]wi(y)dy ,
tie
(1)
Chapter 7 Methods of Nonlinear Stochastic Systems Theory
834
, +2(u,Olen (oo), maly 1)? paul, K,t)= ffaly)ly? =m") +(y— (II) Deanne st) = [jae Tan (y) aly yet
J,
=
E a qx(y)o(y,t )
+ f[tated+ tats) act) - Peta, tyap]
HD
Rg
x ve(ls du) booa(ydpe(w)ay (Iel Ml = 3, and further,
c= OulaK
Deen ad
tr [wm, K,1)K|
N
+S° De cy [ Pav (m, K,t) + ¥™ (m, K,t)? m] i=Saly |i
+50 |vk(dF Pio 4lie send otpen
(IV)
Here Yxv(m, K, t) is defined by (III), and
vB Cm, Kt) = faeyun(y)po(uay (lel el = 3,-6.9), (V) cSie a8 = ‘Age (y)wi(y)pr(y)dy (In|, Vv] =3,-.., N) (VD functions Puv(M, 1K , t), wr (Mm, K at); pko(m, K, t) are defined by formulae of
Appendix 4 at py (y)= po(y)= 1. 7.35.
Show
that the equations
for the multi-dimensional
coefficients
of the
orthogonal expansions in the case of Eq. (5.4.7) are as follows:
On, ’ ae
+R
ea Arora)
tL Obn Es
ys
| erie
a,0(tin Knstn)*rin(tn) + tr [BEM eg,0(Mns Kustn)K (tn)|
Problems
835
Se ee N
ar )
)
BPE
A
es Fetnibal
ss, Bob)
1=3 |vi] +--+ [val =! 1
+ tr
Matas Sore
ete.
Madey vp (Mn; Fostn
cere
)
1, (17m, Rasta)
" +S
Cher
pies
[ve ar
en
atta)
ee
acr
ny
K (tn) |
CR (tat Bins te)
= lhe.Neer | peer
oe)
||
")
ike] eumaxrntt. ... NV),
where
Dee or aa VA eh
-
CUT?
=f
OE
olGr 55 fin; tn)
eG BO aE)
eeeOYn ae
a(Yn stn)
Fo tesssesne(ths To prove formula(6) we multiply term-wise the equalities
on
1 a -1u>0.
(18)
Sy (u) are connected with the generalized Laguerre polynomials
L°(ax) (Appendix 7) by the relation
Si (u) = (—k)-’v LE (ku) . 7
After differentialting in (18) we get
S2(u) = (-k)> » aEee
ae
(19)
As a generating function of these polynomials serves eus/(1+s/k) p(s)
= (14 (1 a s/k)ott
= ~
seu).
The system of the functions Old ea PEI 209 (41) (v a Oy Leen. .) is complete in the space Lo ({0, 00)). Consequently, any function f(u), in particular, any density of a random variable satisfying the condition
f?(u)
yo] 2e—kul2 du
l1.4 Property
3: uo
l
[csz 0
(wa = sug tie
S74) (uo).
(25)
Relation (25) gives as new property of incomplete gamma-function:
y Wr.9) = peteta
(26)
0 We shall formulate it as one of the properties of the polynomials af (u). Property
4: For the incomplete gamma-functions and any values of the para-
meters @ and k the following equality is true
ea eet
—k)-’~°
Pt (Gesot ka al) eae wares
—1)*-
=
(8
ee
FSS
+
CREE 1k
Cl
(27)
> After substituting (19) into (25) we shall have
i
i
fea e7 ,—ku4 Siaa (u)du = —(k) = rf ure J )
—ku
ul
(
1)
S > CH atyut (—ku)4du na T(a+p+t1)
i =< a+v te ( ) one 1) aCe Fo
tesdu.
— (28)
“Appendices
849
Let us make the change of variable t = ku in (28) then uo
ure *" $%(u)du oO
kuo
T(a+v+1) =F See eid) e aONS » (OeSee Pe aeET a Wevieth re
4
dt ae
=
G20)
Applying formula (26) to the right-hand side of (29) we find Uo~
[cet
se(udu
0
va
55”
ir 8 1l-a
Set
=1)#-"
2
)
Hs
Cr
T(a+v+1)
OS, ine
d
das
1, kuo)
But according to (25) we have
prv-i-a ey
a
plier 1) y(a + w+ h1, akuo) ”*T(a+p+t1)
1 = ou
1 e —k MENGE 1 (uo).
After multiplying the both parts of the latter equality by k we come to (27). < Property
5.
The system of the polynomials
{o5; (u)} at a given @
>
—1
represents an orthogonal system of the polynomials for which the gamma-distribution
w(u) = kor foe) hot
yo
ent
Ta SF 1) serves as the weight:
:
a
1
eee 2) ey ene k?*T(a@ +1)
eeei
fiern
> For proving the orthogonality of the polynomials
at V = Lb, at
a y”
V = pl.
(30)
So (u) and Dy (u) it is
sufficient to show the orthogonality of So (u) to the degrees ur, A < Vv. We have rs kort
a
fee J T'(a+1)
bw
du
850
Appendices
bbyYoopt nae
ili
”“T(a+pt
ipa
1)
iba
T'(a+1)
Let us partition this sum into two ones taking into account that
Ch
Cl MCEise. Oe
DA Oe re Co ry
en
—1 nee
Then pe kotly@
‘
5
i FY. S%(u)(ku)*du Thy EY joes 0
vr
v-1
ce Patyt DMatrAtut)) T(a+p+1)[(a+1)
ae
enetict EE aoe NC Sar eee +(—k)’ bat T(a+ p+ 2)P(a +1) Here jt — 1 is taken as ps in the second sum what involves the change of the lower limit of the summation by zero, and the upper limit by the number / — 1. Thus we
have
I Te
T(a+1)
v-1
=(Sh)
Sachs p=0
Stu) (uy py
\U
U
U
T(at+v+1)(a+A4+p4+1) aa T(a+p+1)0(a+1)
a+p+1
=i (oat) alae > a 9 in Oc T(at+v4+1)l(a+A+p4+1) T(a+p+1)l(a+1) p=0 The obtained sum is Sed that C’ Se =
C,ae
hied to the initial sum but has one less item. Accounting
+ C,iese we partition once more
the sum into two ones and
substitute 1 — 1 eeLi, then
iS Liat mea
U
\(k U rd U
= (-k)-”A( = 1) Sicnetog_ er
Appendices
851
After performing this transformation A times we shall have koetlye
Tatty?
Sv (u(kuy\du
T(a+v+1)
( 1 relative to 3
But the polynomial of the pes degree
orthogonal to all w, A = 0,1,2,..., v—1 is determined uniquely with the accuracy till a constant multiplier.
Consequently,
the integral in (32) may
differ from the
polynomial Sh,v (u1) only by a constant multiplier which as it is easy to notice is equal to 1. 4 Let us indicate one more property of Peromd Ta) which follows from the proper-
ties of the generalized Laguerre polynomials.
Accounting the relations between the
polynomials Lo (u) and De Lt), So(u) = (—k)’v! Le (ku) we write
ee
I ee
Qv+a—1—ku
Sk
=
Property 3. Relation (34) in the case of the x?-distribution takes the following form:
S,v(u) =
—(4v+r—4—u)S,~1(u)
—2(v—1)(r—442v)S,,~-2(u).
3. Statistical Linearization of Typical Composite Functions 3.1. Scalar Argument Functions (Table A.3.1)
p(X) & vo(m, D) + ki(m, D)X°, where
X°=X—m,
m=EX,
D=E|X?|’,
ki(m, D) = 0¢0(m, D)/Om, i!
O(z) = i
/
2
dt,
1
(z)= aod
2
hh
0
C=m/VD,
¢t=(m+a)/VD,
(7 =(m—-a)/VD.
(35)
854
Appendices
Table A.3.1
TNS,
os)
Qn = Man-1 + (n— 1)Dan-2, ag lias =m
Seed ec fe = (m+ eD)Ba $ (=) DBs yf = e~ © P/2(m cos am — aD sin am)
fae eee 16. ketoaX (= 2. 1,
Jl EOC
Oat
oe 3)
Soa
oC OG)
|X| 1
10. Xp Xpcosa?X
V1, .-.-,
ear 2e5
=
P
eal?
=e~?
sin a?
Vm are nonnega-
T
+
|v| it
sh
£ pis
ale A
Ka/2(m, cosa’m + Kpasina?
Jp
kpAYy—e, a Mp Yy—e» §
s
) Urkpr Vy—ey—er = kp Vy —2¢y r=1
tive integers
— + =r
=f S. Vr kyr By —ey—ey —
Sea
Aghaaen
n
m)
856
Appendices
Table A.3.2 (continued)
12:EX45 Up XE e cosa
ff
Xx:
s Vr == MpVy-e, — Kp@%y—e,
:
n
V4, ..., Vn — nonnegative integers
c
c
Ar oy Ur kor Vy —ep—e, =e kppYy-26,
\v| > 1
tr {a(K + mm7)}
[1 — 2®(n) —2k1 2m, ®'(9) //k11(kitk22 — kf) ®(nz) + [1+ 2®(n)
14. sgn (x — asgn X2)
+2k12m,
(7) /s/k11(ki1k22 _ Ki5)}
x®(nz ) + 2ki2'(n)[®'(n0 ) — 9!(nt)\/ hii (kt k22 — kis) nt =(m + a)/Vki1, na = (m, +)/Vki1. (Approximate formula for small l,
l= ky2a/./kii(ki1tk22 — k?,)) 4, Integrands Evaluation in Equations Parametrization of Distributions
of Methods
based on
4.1. Moments Methods Owing to the known expressions of the semiinvariants in terms of the characteristic function (Section 3.8)
r Kay
its
i=
Als! x(u5 t) O(ipis)* ...
Kio “nea ld) (S36 ice he Sos
.0)
O(iptg)*
ae
ae eee
a(t)»
(1)
Ther ogc; ls higeee2. 7 = ae, N) being the semialee 0 when |s| = 1 since these semiin-
invariants of the process W(t) (Ks, nec
variants represent the expectations of the components of the process W(t) which are all equal to zero. Hence, by the Taylor formula
x(p;t)= > oDoer SS fan
k=2 |s|=
(2)
Appendices
857
where Py is the residual term. Substituting here the expression /l = b(y, EIR and collecting the items of the same degrees in A, , ...
, Ap we obtain
(oust) ast) = 0 my ns
lust) +m, (3)
k=2 |h|= where Why,
..,hy(¥,
t)
K sth me t vedere) Modsaahnce
= HEL aefan Slew we ee he eyes
PP ips eso Pais
Pie
Pipt:Pqp=hp
Pirt--+Pq, =h1
ee ay&
Function D(s
ee
oS 5. fie
se
a>0
aeoa
oo eRe
b
wae a“s
Jes
hae
t=8=9
261.2
= 1+ Ver—-1
b?6"(t—r) +2cb6!(t—r)+4(t—T)
b2 52 Yebs+1
Appendices
865
Transfer Functions and Frequency Characteristics for Some Continuous Linear Systems Table A.5.2. Gain-Frequency
Phase-Frequency
Characteristic A(w)
Characteristic w(w)
OatwI1/a b G a’s- + 2eas +1
9. b?s? + 2ebs +1
et Ba b aes S| (tie ara)? + 4aze2w?
(1 — bw)? + 4e?b2w?
ee: 2aEw arcs eee
arctg ; abe
Some Typicl Models of Stochastic Discrete Systems Table A.5.3.
1. Nonlinear Autoregression (AR): e | order
years = pr (Yr a Rata t5 Yeqi—1)+
+0, (YE, ---) Yeri—i)Vi e first order
Yeti = pe(Ye) + Ve (Ye)Vi
866
Appendices
Table A.5.3 (continued)
2. Linear AR:
11 e | order
VRS")
> arp Sere TV
EVE
r=0
Yeon
cpYp + ae Vi is Yq => Gpr Ea t
@ first order e | order with
r=0
I
Moving Average (ARMA)
+A, + ys bskVi+s $=0
3. Linear Integrated AR (IAR) I-1
e | order
ara
=
arp Na Vere + VeVi nO)
I-1
e | order with
Nea
=
bseVi s=0
4. Nonlinear ARIMA
Yui =
SS Ark Petr (Ye+r) r=0
I
of | order
=F > bskVe4+s s=0
5. Stationary linear
A(V)Y;
=—C+
B(V)Vk 1-1
ARIMA of I order
A(V) = IV? —
> Op r=0
BW
I
iv
s=—N
Vug = Urqi
Y= OH)+WO
Appendices
867
Operators, Weighting (Green) and Transfer Functions for Some Typical Linear One-Dimensional Systems with the Distributed Parameters Table A.5.4. Weighting (Green)
Transfer
Function g(€,€1,t, 7) , |Function ®(€, €1, s)
1. y(t) = e(t—A) Visi =0,NG