Stochastic Systems: Theory and Applications 9810247427, 9789810247423

This book presents the general theory and basic methods of linear and nonlinear stochastic systems (StS) i.e. dynamical

147 7 35MB

English Pages 928 [936] Year 2002

Report DMCA / Copyright

DOWNLOAD PDF FILE

Recommend Papers

Stochastic Systems: Theory and Applications
 9810247427, 9789810247423

  • 0 0 0
  • Like this paper and download? You can publish your own PDF file online for free in a few minutes! Sign Up
File loading please wait...
Citation preview

=

STOCHASTIC SYSTEMS "Theory and Applications

Digitized by the Internet Archive in 2023 with funding from Kahle/Austin Foundation

https ://archive.org/details/stochasticsystem0000puga

STOCHASTIC SYSTEMS V. S. Pugachev I. N. Sinitsyn Russian Academy of Sciencies

Translated by

|. V. Sinitsyna

Ve World Scientific NewJersey

«Londons Singapore e Hong Kong

Published by World Scientific Publishing Co. Pte. Ltd. P O Box 128, Farrer Road, Singapore 912805 USA office: Suite 1B, 1060 Main Street, River Edge, NJ 07661

UK office: 57 Shelton Street, Covent Garden, London WC2H 9HE

British Library Cataloguing-in-Publication Data A catalogue record for this book is available from the British Library.

STOCHASTIC

SYSTEMS

Theory and Applications

Copyright © 2001 by World Scientific Publishing Co. Pte. Ltd. All rights reserved. This book, or parts thereof, may not be reproduced in any form or by any means, electronic or mechanical, including photocopying, recording or any information storage and retrieval system now known or to be invented, without written permission from the Publisher.

For photocopying of material in this volume, please pay a copying fee through the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to

photocopy is not required from the publisher.

ISBN

981-02-4742-7

Printed in Singapore by World Scientific Printers (S) Pte Ltd

PREFACE

The monograph is dedicated to the systematic presentation of the applied advanced theory and analytical methods for the stochastic systems, i.e. the dynamical systems described by the finite- and infinitedimensional stochastic differential, difference, integral, integrodifferential etc. equations. It is based on the results of the fundamental research performed in the Institute for Informatic Problems of the Russian Academy of Sciences in the context of the scientific program “Stochastic Systems” and the lecture courses-delivered by the authors in the domestic and the foreign Technical Universities. The book may be used as the textbook for the applied mathematics faculties of the universities. The unified procedure, thorough selection of the examples and the problems (over 500) and the large applications make this book useful for the graduates, the post-graduates and the lecturers. The book is of considerable interest also for the mathematicians who deal with the stochastic equations and their applications. The construction of general stochastic systems theory is based on the equations for the multi-dimensional characteristic functions and functionals. The stochastic differential equations with the arbitrary processes with the independent increments are studied. The equations with the Wiener and the Poisson processes are considered as a special case. The methods of the parametrization of the multi-dimensional distributions in the nonlinear stochastic systems based on the moments, the semiinvariants, the quasimoments and on the consistent orthogonal expansions are stated systematically. Special attention is paid to linear and reducible to linear stochastic systems theory based on the canonical representations (canonical expansions and integral canonical representations). Most attention has been concentrated on the structural theory of the composed stochastic systems on the grounds of the conditional probability measures. Chapter 1 is devoted to the mathematical models of the dynamic systems under conditions of the random disturbances and their characteristics. The linear and the nonlinear continuous, discrete and continuousdiscrete systems described by the stochastic equations in the finite- and the infinite-dimensional spaces are considered. Special attention is paid to the composed stochastic systems. The main notions of probability distributions theory of the random variables, the random processes and the random functions are stated in Chapter 2. After determining the probability spaces, the conditional

vl

Preface

probabilities, probabilities in the finite and the infinite products of the spaces are considered. The conditions of the existence of the regular probabilities are established. Further the different probability measures of the random functions and the probabilities of events connected with the random functions are studied in detail. The last sections are devoted to the distributions of the separable random functions, to the derivation of the criteria of the continuity and the differentiability of the random functions. Chapter 3 contains the account of the basic notions connected with the expectation of the random variables in Banach and topological linear spaces, the properties of the operators of the moments of the second and higher orders, the properties of the characteristic functionals and the correspondence between the probability measures and the characteristic functionals are established. The conditional operators of the moments and their properties are studied in detail. Special section is devoted to the elements of the probabilities distribution in the infinite-dimensional spaces. The properties of the normal distribution on the product of two spaces, the conditional normal distributions and the normal distributions in Hilbert spaces are considered. Approximate representation of the oneand the multi-dimensional distributions based on the consistent orthogonal expansions is treated. Special attention is paid to the semiinvariant and the quasimoments. A systemical presentation of general theory of the canonical expansions is given. Practical ways of the canonical expansions constructing are described. The joint canonical expansions are considered. Chapter 4 is devoted to the selected topics of stochastic calculus. The theory of the convergence, the continuity and the differentiability is considered. The integrability of the random functions over nonrandom and random measures is discussed. The theory of the stochastic Ito integrals, the symmetrized (Stratonovich) integral and the 6-integral is outlined systematically. The stochastic integrals over the Wiener and the Poisson measures are studied. The connection between the processes with the independent increments and the It6 integral is established.

The elements of the theory of the generalized random functions are stated. A systematic account to the integral canonical representations is given. The connection of the integral canonical representations with the spectral one is established. Practical ways of constructing the integral canonical representations and, in particular, the methods of the shaping

filters are considered. studied.

The joint integral canonical representations are

~~

Preface

vil

The general theory of the finite- and the infinite-dimensional stochastic systems is given in Chapter 5. At the beginning the differentiation of the Ito process is set out in detail. The Ito formulae for the differentiation of the composite functions in the cases of the Wiener, the Poisson and the general random process with the independent increments have a significant place. The theory of the stochastic integral, differential and integrodifferential Ito equations including the methods of reducing the state equations to the stochastic differential equations, change of variables, sufficient conditions of the existence and the uniqueness of the mean square solutions are considered. The central problem of stochastic system theory —he problem of calculation of the multidimensional distributions is stated. The elements of numerical integration of the stochastic differential equations are given. The theory of the one- and the multi-dimensional distributions in the continuous, discrete and continuous-discrete systems in the finite-dimensional spaces are stated in detail. For such systems the known equations (Fokker—Plank— Kolmogorov, Feller-Kolmogorov, Pugachev and their generalizations) are derived. The problems of one-dimensional and multi-dimensional distributions theory on the basis of the equations for the correspondent characteristic functionals in composite stochastic systems including the stochastic systems in the Hilbert spaces are considered. The main statements of the structural theory of the composed stochastic systems on the basis of conditionally probability measures are outlined. The last section contains the applications to optimal online filtering problems. Simple derivation of the Kushner—Stratonovich and the Zakai equations for the normal white noises is presented. Corresponding equations for the composite stochastic systems are given in Problems. Chapter 6 contains the account of the main methods of linear stochastic systems theory in connection with continuous, discrete and continuous-discrete systems. The spectral and correlation methods, the methods of general theory of linear stochastic systems are outlined. The correlation methods based on the canonical representations are considered with great attention. The general theory of the canonical representations of the linear transformations is given. The separate subsections are devoted to the solution of the linear operator equations by the canonical representations methods. The applications to the offline spectral and correlation analysis and the shaping linear filters design are given. Software for the analytical modeling problems is described. The Kalman-Bucy and the Liptser—-Shiryaev online filters are considered.

vill

Preface

Chapter 7 is devoted to the approximate methods of nonlinear finite- and infinite-dimensional stochastic systems theory based on the parametrization of the distributions. The normal approximation, statistical and the equivalent linearization methods are considered. The Genenormalization of the composite stochastic systems is treated. ral problem of parametrization of the multi-dimensional distributions is stated. The methods of the moments, the semiinvariants, the quasimoments and the orthogonal expansions are given: The practical problems of reducing the number of equations for the automatic derivation of the equations and the corresponding software for distribution parameters are considered. The solution of the structural parametrization problem of the multi-dimensional distributions is outlined on the basis of the ellipsoidal approximation method. The canonical representations methods for the nonlinear stochastic systems are considered. The ba-

sic methods are given for stochastic systems with the normal white noises. Corresponding methods for the composite stochastic systems are given in Problems. The applications of the given methods to the nonlinear stochastic systems offline and online analysis, modeling and filtering problems in mechanics, physics, biology, chemistry, finance and control are stated. Specialized software is described. Chapters 1, 5-7 give efficient toolkit for the analytical modeling and the design of the stochastic systems. To facilitate the mastering of the stated methods about 500 thoroughly selected examples and problems are given in the correspondent subsections, sections and chapters of this book. Appendices 1-7 contain auxiliary materials which include the information about tables of stochastic It6 differentials for the typical composite functions, the Hermite polynomials of a vector argument, the polynomials orthogonal to gamma and y?-distributions, the tables of the coefficients of statistical linearization for the typical composite functions, integrands evaluations in the equations of the basic methods of distribution parametrization, the tables of a basic characteristics and transformation the structural rules of the linear systems and the tables of some definite integrals and special functions. For information about the notions and theorems from vatious parts

of mathematics used in the book we advise (Korn and Korn 1968).

Bibliographical notes and References are given at the end of the book. The authors do not pretend in any way to provide a complete bibliography. Only those sources are given in the list which are cited in the text. Basic notations are presentented before Index.

~

Preface

ix

The book is divided into chapters, sections and subsections. The formulae (examples) are numerated with indication of the number of the chapter and the section, for instance, (1.2.3) means that formula (3) belongs to section 2 of chapter 1. The problems are numerated only with indication of the number of chapter. At each Appendix the separate numeration is accepted. For the convenience of the readers the formulations of all basic results and statements are given in the separate theorems and the corollaries or in the text italics. The beginnings and the ends of the evaluations, the proofs and the discussions which lead to certain results are indicated by triangular indices p and Let u(t, a ) be the solution of the homogeneous equation

= au

(1.2.2)

satisfying the initial condition u(r ie ) = I, i.e. the fundamental matrix of solutions

of Eq. (1.2.2). By the change of variables

7 = u(t, to)v the first Eq. of (1.2.1) is

reduced to the form

uv + uv = auv

+.a,z2+ a9

or by virtue of Eq. (1.2.2)

uv =ayr+ao. Remark.

It is known from differential equations theory that the determinant

of the matrix u(t, to) is expressed by t

A = |u(t,to)| = exp

J vra(ryar

(1:2:3)

to

where tr a(T) is the trace of the matrix alr): To derive this formula it is sufficient to notice that according to the differentiation rule of a determinant and the properties

of the determinants

A =

condition A(to) =

|u(to, to)| = 1 is determined by formula (1.2.3).

from this formula that

Atr a(r i The solution of this equation with the initial

It follows

A # 0 (we take for granted that the integral in (122.3)/is

finite at all t, 9 owing to which the exponential function does not vanish). Hence, taking into account that the matrix u is always invertible we get

v=u(aye+ ao).

1.2. Linear Differential Systems

25

Integrating this equation with the initial condition v(to) = u(to, to)~+2z(to) = 74 gives t

v(t) = zo + it u(r, to) | [ai(r)a(r) + ao(r) Jdr,

(1.2.4)

and

z(t) = u(t, to) v(t) = u(t, to)zo 1.2.5

+ u(t,to) fu(r, to)* [ar(z)e(r) +ao(r)|dr.< to : Let us prove that at any to, 7, t, tp

(

)

Putting u = u(t, to),

v= u(t, to).= u(t,to)4 , we have at any tp andt

> tp uv!

=

I.

(1.2.11)

Differentiating this formula we get:

uyve + uve = 0. So yielding by Eqs. (1.2.2), (1.2.11) we have

uv; = —auv’ = —a ort? = —u“la.

(i. 2612)

Hence by virtue of Eq (1.2.11) we obtain the differential equation for UV: vr = —-y!a,

Now we notice that according to (1.2.6) and (1.2.11) u(t, 7) = u(t, to)v(7, to)’. Differentiating this formula with respect to 7 and taking into account that according

to (1.2.12) 0;(7, to)? = —v(r7, to)? a(r), we find i,

(t,T) = u(t, to)vr(7, to)? = —u(t, to)v(7, to)? a(r) = —u(t, r)a(r).

Thus u(t, 7) as a function of T at fixed t is determined at T < t by the equation

u,(t,7) = —u(t, T)a(r)

leah)

with the initial condition u(t,t) = J att =t. 4 Transposing Eq. (1.2.13) we obtain

UAT)

=

alr) ulter)? «

Cle2ei4}

This equation is the adjoint equation to Eq. (1.2.2). Therefore u(t, 7)? as a function of T at fixed t is determined at 7 < t¢ by Eq. (1.2.14) adjoint to Eq. (1.2.2) with the condition u(t,t)” = I at the end. So u(t, 7) may be found by the integration of the adjoint Eq. (1.2.14) in reversed time

with the initial condition u(t,t) = I. 1.2.3.

Transformation of Linear System Equations into Cauchy Form

If the input-output relation of a system is represented in the form of a differential equation of higher than the first order then it usually contains not only the derivatives of the output but the derivatives of the input as well. Therefore, the problem arises to transform the equation

28

Chapter

1 Dynamical Systems and Their Characteristics

of a linear system into a set of equations of the first order which do not contain the derivatives of the input. Now we shall show how to do this. Let us consider a nonanticipative system whose behaviour is described by the linear differential equation of the form

See

Ybe,

K=O

k=0

(1.2.15)

where the coefficients ao, a1, ..., Qn, bo, b1, ... , bm depend in the general case on time t, and m < n.. It is evidently sufficient to consider the case of m = n. The case of m < n is the special case where Onis

Ort

=

());

> Let us introduce new variables

Z=Y—

QoL,

241

where qo, 41, ---, Yn—1

= 2% — get (K=1,...,n-—1),

(1.2.16)

are some functions which we shall determine from the

condition that the resulting equations of the first order contain no derivatives of the

input Z.

From Eqs.

(1.2.16) using the known formula for the derivatives of the

product of two functions

7

Uv Che SS en PuP)y Ae CrCee =

me

»

|

ee

1.2.17 ve

p\(r — p)!

(

we find

Y¥=%1+ 402, 8

Soon

yo) = Zs41 + Ge)

= 2540

em

r=0

Pg

Pal)

(1.2.19)

r=0,p=0

(SP

oe

eae

After changing the order of summation we get

NT NG +

Lorde? ((Sad

teen 1). » «(1-3.20)

Differentiating formula (1.2.19) corresponding to s = n — 1 gives

mee y\" ” ain

#9CES—r— 12)yet) =z, +

(of ("),t)

(1.2.21) Ahh ait 5mqd, + al Sor? p=1

r=p

1.2. Linear Differential Systems

29

Substituting the expressions (1.2.18), (1.2.20) and (1.2.21) into Eq. (1.2.15) we obtain

nin

+

> a-izi+e

AP

ip!

Ueeoee Da

ail

s=0

=

a >x) eos CPa,q’,”) >, = al?) , Sy

(1.2.22)

i)

Equating the coefficients at the corresponding derivatives of the input & in the leftand right-hand sides of Eq. (1.2.22) we obtain the equations for the functions gy:

Mee ce “r)=b, = (p=1,...,n).

Gres)

s=pfr=p

The last of these equations corresponding to p = 7 gives

Gh= an'b,.

(1.2.24)

For solving the other Eqs. (1.2.23) notice that

YS s=p

Lead? reppe Se r=p

s=p

i009

pI

1=0

n—pn—p—l

!)

.ye yg Pi teds--p-1 =) = 1=0 s=p+l

(=O

n—-pn—p—h

Fa

(1)

Dang pti aptlthTh =O

n—p—In—p—-—h

C.Pap erEn gy? = Gn n—p T yO h=0)

Y=0

h=0

>s Coeitptlthdy [=0

Substituting this expression into (1.2.23) and putting n — p= k-1k-h

=:

Onqk + Dey ny) Chaepian—ethtiq,. h=0

()

@) ip

(1.2.25) k we get

=

= on-k (K=1,-..,2—1).

7=0

(122.20) From

(1.2.26) we obtain the recursive formula for determining the functions q;,

+) 9n-1: k-1k—-h

h=0 [=0

(1.2.27)

30

Chapter 1 Dynamical

Systems and Their Characteristics

After determining in such a way the functions qo, 41,---,Qn—1

we transform

Eq. (1.2.22) into the form n

n

n-1

8

pee

Onin + 0 az = { ty Sod ree t=

l=1

sa)

Solving this equation for Zp, gives together with Eqs.

k =1,...,m-—1

hae)

1=0

(1.2.16) corresponding

to

the equation forming the set of the first order equations Ze = Ze41

+ Mee

Zn =

(kerk,

Ss Wy od

=

n

-Soa;

(1.2.29) Qj=121

+ nk ,

where

heathy

a) Ome ya, (1.2.30)

s—0

l=i

n-1

n

s

= an1 |bo— ) fangs — Soas Dain =

I

h=0

$=1

Si

Noticing that n

s

Sal

l=1

)

non

(a) I

n

n—l

n-1n-h

dias Doar = DOD oases=DOD ansiah? = DO DTansiah? (i)

gel!

i=h=)

(=O

a

(1.2.31)

we transform the formula (1.2.30) for gn into the form n—-1ln-h

Grit= Gawlibax My: Ds ansiqe?

:

(162.32))

h=0M=0

It is evident that this formula coincides with (1.2.27) at k = n. Thus all the functions

qi, --+ 5 Un are determined by (1.2.27). The

=---=

case

of ™


» Aay=

Bere

5

Qqn-1

ace eGOne

In

(1.2.35)

In the case of m = n the output y depends explicitly on the input zx. But this is a rare case since m < n practically always yields qg = 0. The

expression (1.2.34) of the output y does not contain z in this case. +,

In the special case of a stationary system all the coefficients aj, 2n; 00 01, +5 Unare Constant. ald thus Gp, d1,---,dn Gevermined

successively by formulae (1.2.24) and (1.2.27) are constant. qi? Oe Ut ree Ly ee considerably simpler in this case:

Therefore

an loniitiac (1.2.21) become

k-1

Go

Ge Dene Te te

(50-1-

So tm-tena) i ae reared Oe

Formulae (1.2.33) in this case take the form w=

7a

di =

(5.—

>a h =n—m

=n

me

oat

— 04

Gree

a

On

(k =n—m+1,...,

mn). (1.2.38)

32

Chapter 1 Dynamical

Systems and Their Characteristics

Notice that all previous calculations and consequently formulae (1.2.27), (1.2.33), (1.2.37) and (1.2.38) for go, 41, ---, Qn are valid both for the scalar and the vector input x and the output y. In the latter case ao, Gy js » Ope. B05 O15 1284 Ony Weds a> > Gea te plenent, matrices of the corresponding dimensions. Example

1.2.2. In order to transform the equation

a3 ¥ +aoy + ayy + aoy = bet + bit + bo

(1)

into a set of the first order equations we put according to (1.2.16) and (1 .2.33) OS

Ai,

22 = 21 — q12, 23 = 2 — Gov. Then by (1.2.33) we find

qo=0,

q1 = a3'b2, go = a3 '(b1 — aoqi — 20191),

q3 = 43 ‘(bo — 4191 — 4241 — 4391 — 4292 — 4342) As a result we obtain Eqs. (1.2.29) which has in this case the form 3 Ai—Zo7+Ne

,

‘ 232—23 4-090

,

: SI 23=—azg (aoz1

+4122

+4223)+

932 ,

Y

en =21.

(11) 1.2.4. Inverse Systems The system inverse to a system A is such a system whose output is equal to the input of the system A when its input is equal to the output of A (under the corresponding initial conditions). Thus the inverse system performs the transformation of signals inverse to that performed by a given system A. It is evident that if the system B is inverse to the system A then the system A is inverse to the system B. In other words, A and B represent two reciprocal or mutually inverse systems. The amplifier with the gain k and the amplifier with the gain 1/k, the differentiator and the integrator may serve as examples of pairs of mutually inverse systems.

If a system with the input «# and the output y is described by Eq.

(1.2.15) then the inverse system is described by the same differential equation in which y serves as its input and z serves as its output. If m = n, then the reciprocal systems described by differential equa-

tion (1.2.15) are of the same type (i.e. both differential systems). After transforming the equation into a set of the first order equations, as we have already seen in Subsection 1.2.3, the output of each of these two

1.2. Linear Differential Systems

33

systems will contain its input with some coefficient and the remaining part of the output will be determined by the corresponding set. of the first order differential equations. If m = 0 then the output of the inverse system is given by

2 = b5* > apy).

(1.2.39)

It goes without saying that for the existence of the inverse system in this case it is necessary that the inequality bo # 0 be valid at all t in the case of an one-dimensional system and that the matrix bo be invertible at all ¢ in the case of a finite-dimensional system. Under this condition the inverse system performs the linear differential operation

Lig) aD

ya Fa dte

(1.2.40)

k=0

> If 0 < m

< n then the operations which are performed by the inverse system

include the (n = m)-tiple differentiation of the input. As a result the output of the inverse system contains a linear combination of the input and its derivatives up to

the (n = m)*h order. In order to determine the inverse system it is necessary in this case to find the remaining part of its output. For this purpose we put nm—m

p=zt+nazt Do ay

(1.2.41)

1=0

and determine the coefficients Cg, Cj , ...

, Cn—m

In such a way that the variable Z

contains no linear combination of the input Yy and its derivatives.

Using the known formula (1.2.17) of the derivatives of the product of two func-

tions we find from (1.2.41)

Bie aom

ie Se ssCree y(t 1=0

r=0

=a? Ne Se Cre!(s—r) yltr) a= r=0)1=0

4

Cre os yl)

(1.2.42)

After this evident transformation of the double sum it is expedient to change again

the order of summation in order to obtain the coefficient at y6*) in the explicit form.

34

Chapter 1 Dynamical

Systems and Their Characteristics

For determining the limits of the inner sum we notice that0 k —n +m and

min(k,s)

ve

k=0

the order of summation

are possible we come to

Crea y(®),

Pe

s=0

0

min(k, s). As a result we obtain

or eltSyl) ,

r=max(0,k—n+m)

and

(1.2.43)

taking into account

that

k
0 weget

5Hb,x) = snbp z6*) s=0

+

3

k=0

min(k,s)

ha

DyPnnidesu CBee

k=0 s=max(0,k—n+m)

Eyes

masmenaeaay

r=max(0,k—n+m)

Let us substitute this expression into Eq.

(1.2.15) and equate the coefficients at

y™) | wens 5 y”) in the left- and the right-hand sides of the equality obtained.

taking into account that min(k, s) = s at s m

s

3

)

SF

Gthicaaaz

(b= sit pederp ny)

s=max(0,k—n+m) r=max(0,k—n+m)

In order to solve these equations for Cg, Cj , ... , Cn—m

order of summation.

m

Se

Then

m we get

(1.2.45) we change once more the

Remembering that r < 5 we get

m

Y Clb clbansea, (b= 9) oo)

(1.2.46)

r=max(0,k—n+m) s=r

or putting

s

=r+h

m

3. r=max(0,k—n+m)

Mm—Tf,

h=0

Clinbrgnce?,=a, (k= m,...,n).

(1.2.47)

1.2. Linear Differential Systems

At k =

n the sum

35

contains only one item corresponding

Therefore, we have at

k = n the relation bmCn—m

to

r =

m,

Gee eeb aa Separating in the item corresponding to

only one value

beCh om

h = 0 when r =

(1.2.48)

T = Mm and taking into account that h has

m—r

ys

De Cotte h=0

ol flWem

(1.2.49)

7 < ™), these equations can be solved for Cx_m. m—1

Cela = 07,

toages

>k—m

for all

As a result we obtain

m—r

ds

ec

r=max(0,k—n+m)

After determining Cyn—m

areal Cae be

.

As the sum here contains only Cc; with | greater than k — m (k —r items since

we

emp.

h=0

eal

(1.2.50)

by (1.2.48) we may find successively Cp—m—1,---,

Co using formula (1.2.50).

0.

™ we transform Eqs. (1.2.45), k > be z'*) =

»» a,y”) ;

k=0

k=0

(1.2.51)

where

m—1

ide

min(k,s)

3

Sy

‘ea ane.

(1.2.52)

s=max(0,k—n+m) r=max(0,k—n+m)

it 2.0piiynenotan— 1): Changing the order of summation here and taking into account that max min(k, s) s

=katk Giese”

r=max(0,k—n+m) s=r

(1.2.53)

Systems and Their Characteristics

Chapter 1 Dynamical

36

or

k

m—r

r=max(0,k—n-+-m)

h=0

Ge in), —

Sy

SiC hidrthle alk SU gleg cnc Mm),

(1.2.54)

Formula (1.2.41) shows that the system inverse to the system described by Eq. (e215)

at 0
|o)

lie ee

2 lrg56 ala let Hy |

y; (S )

7:

6)

—>

hs

yHy,ty,-);

Ys;

y)

— > y=y,ty,ty, y;

Fig. 1.3.3 A block diagram is called a graphical image of the system structure or its blocks. Here the structure is regarded the aggregate of the blocks and the connections representing the loops. The elements of a block

diagram are (Fig.1.3.3): e Linear dynamic links presented by the blocks with the operators describing the functions filled in the blocks or in the description applied (Fig.1.3.3a). e Linear and nonlinear functional generators (Fig.1.3.3b). e The summators presented by the circles divided into the sectors. Generally the summators realize the algebraic summation. Some times we use + signs in circles (Fig.1.3.3c). e Couplings which transmit the directed action without changing (Fig.1.3.3d). e Branching points on the coupling lines (Fig.1.3.3e). Equally with the block diagrams the oriented graphs are often used (Fig.1.3.4a,b,c). The main elements of an oriented graph are edge, node and arrow. The oriented graphs do not need special notations for summators and couplings as in the block diagrams.

1.3. Conneétions of Systems

45

qg)

Fig. 1.3.4 Transformation rules for block diagrams and graphs are given in Subsection 1.3.4 and Appendix 6. 1.3.2.

Weighting Functions of Linear Systems Typical Connections

Consider at first a series connection of two linear systems whose weighting functions gi(t,7) and go(t,7) are given (Fig.1.3.5). Find the weighting function g(t, 7) of a composite system.

> For this purpose we find its response on the unit impulse. As a result of the action of the unit impulse on the input of the first system its output will represent its weighting function 91 (t, af). Consequently, the input of the second system is equal

a(t

gityn).

(17371)

Here the input of the second system according to general formula (1.1.8) is expressed by formula [oe)

foe}

yo(t) = J s2lt,o)22(0) a0 = J 92lt,0)ni(0,7) do. —00

—oo

(1:3:2)

46

Chapter 1 Dynamical

Systems and Their Characteristics

But the output of the second system represents also in a given case the output

of the whole series connection of the considered systems, i.e. the weighting function obtained as a result of the connection of a system

(1.3.3)

yo(t) = g(t,7).

Comparing this formula with (1.3.2) we obtain the following formula for the weighting function of a series connection of two linear systems

gts

J salt-o)au(er) do. 4

(1.3.4)

Thus the weighting function of two connected in series linear systems is determined by formula (1.3.4). In the special case of nonanticipative systems we have g;(c,7) = 0 at o < 7 and go(t,0) = 0 at o > t. Consequently, for nonanticipative systems an integrand in (1.3.4) differs from zero only in the limits +

For determining the weighting function g(t, ai) of a composite system we shall find its response on the unit impulse 6(t = a) Accounting that while delivering the unit impulse 6(¢ = T) on the inputs of connected systems the outputs are equal to

the corresponding weighting functions g1 (t, T) and ga(t, T) we obtain

g(t, 7) = gi(t, 7) + ga(t, 7). 4

(1.3.7)

Thus at the parallel connection of the linear systems thetr weighting functions are added. For studying the systems with a feedback we use the notion of the

reciprocal (mutually inverse) systems (Subsection 1.2.4). Consider a series connection of two reciprocal linear systems. The weighting function of the linear system inverse relatively to the system with the weigh-

ting function g(t,7) we shall denote by the symbol g~ (t, rT). The weighting function of a series connection of two reciprocal systems is equal to 6(t—7). On the other hand the weighting function of a series connection may be calculated using formula (1.3.4). As a result we obtain [oe)

ifg(t, 0)g(0,7) do = 6(t — 7). —oo

(1.3.8)

48

Chapter 1 Dynamical

Systems and Their Characteristics

This relation represents the main property of the weighting functions of the reciprocal linear systems. Changing the order of the connection of the systems we have another relation co

/ g(t,0)9 (0,7) do = 6(t — 7).

(1.3.9)

—0o

Thus the wetghting functions of two reciprocaldinear systems satisfy two

integral relations (1.3.8) and (1.3.9) which in general do not coincide. The weighting function g(t, 7) of the system with a feedback mapped in Fig.1.3.7 may be defined as in the previous cases after observing the passage the unit impulse. But this way is rather difficult and leads to

an integral equation relative to the sought weighting function g(t, 7). It is rather easy to find the weighting function of an inverse system (Subsection 1.2.4).

Bigs 13.7

Fig. 1.3.8

> It is evident that the system which is inverse relative to the system with a feedback is the parallel connection of the system relative to the system in the direct loop and the system in the feedback loop (Fig.1.3.8).

Really, denoting the input

of the considered system by Z, the output by y and the output of the system in a feedback loop by Z we see that © — Z will be the input of the system in the direct loop and y will be its output.

While supplying the output

y on the input of the

parallel connection of the systems with the weighting functions g; and g2 we shall obtain on the output of the first system the signal 2 — Z, and on the output of the second system the signal z. Adding these signals will give on the output of a parallel connection the signal X (Fig.1.3.8).

Thus a series connection of the systems shown

in Fig.1.3.7 and 1.3.8 give an ideal servomechanism.

are reciprocal.

Consequently,

these systems

Applying formula (1.3.7) for the weighting function of the parallel

connection of the linear systems we find

g(t,7)= 9; (t,7) + ga(t,7).4

(1.3.10)

1.3. Connections of Systems

49

This formula determines the weighting function of the linear system which is inverse relative to the system with a feedback shown on Fig.1.3.7.

1.3.3.

Connections of Stationary Linear Systems

At first let us consider a series connection of two stationary linear

systems with transfer functions ®;(s) and ®2(s) (Fig.1.3.9).

Suppose

that on the input of this series connection a signal representing an exponential function of time e*’ acts indefinitely for a long time. This signal while passing through the first system is multiplied by the transfer func-

tion ®;(s). Thus on the output of the first system and on the input of the second system will be the function ®;(s)e**. By virtue of the linearity of the second system its output will be equal to the product of its response on the exponential signal e** by the constant multiplier ®2(s),

i.e.

©1(s)®2(s)e*'.

Thus a series connection of the stationary linear

systems gives the stationary linear system whose transfer function ®(s) as equal to the product of the transfer functions of connected systems

&(s) = 1 (s)2(s).

Sao

(1.3.11)

oo

(s) = ®, (5) ®, (S) Fig. 1.3.9 This formula is easily spread to a series connection of any number of the stationary linear systems. Namely, if the transfer functions of

n connected in a series systems are equal to ®(s),...,®n(s) then the transfer function of the connection ®(s) is determined by formula

@(s) = ©1(s)b2(s)---@,(s) = ]] O:(s).

(1.3.12)

p=

Thus the result of a series connection of the stationary linear systems does not depend on the order of their connection.

50

Chapter 1 Dynamical

Systems and Their Characteristics

As the modulus of the product of the complex numbers is equal to the product of the moduluses of the factors, and the argument is equal

to the sum of the arguments of the factors then from (1.3.12) follow the formulae for the gain- and phase-frequency characteristics of a series connection nm

|&(iw)| = [] ee (iw)], k=

(1.3.13)

\

Ig [B(iw)| = Solg|®x(iw)|, arg @(iw) = So arg, (iw). Formulae (1.3.14) show that at a sertes connection

(1.3.14)

of the statio-

nary linear systems their dectbel-log and log phase-frequency characteristics are summarized.

®, (s) es

e

®,(s) oa.

D(s)e

:

®, (S) e.

(s) = ®, (5) + ®,(S) Fig. 1.3.10 For determining the transfer function of the parallel connection of the stationary linear systems with the weighting functions ®,(s) and 2(s) (Fig.1.3.10) we suppose that on the input of this connection the signal representing the exponential time function e*’ acts indefinitely for a long time. As a result the outputs of the connected systems will be equal to ®;(s)e%’ and ®(s)e*' correspondingly. The output of the connection will be equal to the sum of these expressions. Consequently, the transfer function of a parallel connection of the stationary linear system is equal to the sum of the transfer functions of connected systems:

B(s) = O;(s) + 9(s).

(1.3.15)

1.3. Conriéctions of Systems

51

Evidently, this formula is also spread to any number of parallelly connected systems:

(s) = 5 ;(s).

(1.3.16)

Let us pass to the determination of the transfer functions of the stationary linear systems closed by a feedback. At first we consider the system consisting of the stationary linear system with the transfer function ®;(s) closed by the nonnegative feedback containing a statio-

nary linear system with the transfer function ®2(s) (Fig.1.3.11).

For

determining the transfer function (s) of this system let us consider an

inverse system which has the transfer function 1/®(s) (Subsection 1.3.2). This inverse system represents a parallel connection of the system with

the transfer function 1/;(s) and the system with the transfer function ®5(s). Therefore using formula (1.3.16) we get

(5) °@ (s)'® +1 (s)'®

= (s)®

Fig. 1.3.11

mane = ee + @2(s) = i+ ®1(s)P2(s)

®(s)

®i(s)

#1 (s)

From here we find the transfer function of the system with a feed-

back:

Aiea

RAS,

1+ ©1(s)®o(s)

(1.3.17)

At a rigid negative feedback it is necessary to put ®(s) = 1 in formula (3.17): Thus at any connections of the stationary linear systems we always obtain the stationary linear systems, whose transfer functions and frequency response characteristics are determined by means of elementary

52

Chapter

1 Dynamical Systems and Their Characteristics

algebraic operations on given transfer functions (on frequency response characteristics respectively) of connected systems. At all possible connections of the stationary linear systems described by ordinary differential equations as a result we always obtain stationary linear systems whose behaviour is also described by ordinary differential equations. Really, in this case the transfer functions of all connecting systems represent the rational functions of s. Therefore all derived for-

mulae for the transfer functions of the connections determine the transfer function of the connection ®(s) in the form of the rational function of s. After presenting this function in the form of the ratio of two polynomials and replacing s by the operator of the differentiation with respect

to time D we shall find the operators F(D) and H(D) and Eq. (1.2.60) for the input — output relation. 1.3.4. Structural Transformation Rules

It is expedient to transform the block diagram into more convenient form by means of the partition of the composite systems into more simple ones (decomposition), the connection of the simple systems into one (aggregation) and another transformations for the sake of simplicity of the structure on retention of the number and the order of included systems. The main principle of structural transformations consists in the fact that all the transformations of the block diagram of the linear sys-

tem (Subsection 1.3.1) should be performed in such a way that all the inputs and the outputs of each transforming section of the block diagram will remain invariant. With regard to Subsections 1.3.2 and 1.3.3 from this principle follows the rules of the structural transformations of linear systems given in Appendix 6.

Fig. 1.3.12 Example

1.3.1.

For transforming

the flexible feedback

by the rigid

one (Fig.1.3.12a) we interchange the branching point and A» system in a feedback (Fig.1.3.12b).

~

1.3. Connections of Systems

53

Fig. 1.3.13 Example

1.3.2.

Consider the translation of the system Ag along the

connected in a series stationary linear systems Aj, ...

, An (Fig.1.3.13a).

As the

stationary linear systems connected in a series may be arbitrary interchanged then the mutually inverse (reciprocal) systems introduced by the translation of the branching point and the summator via the one and the same system are mutually compensated. So for the translation of the parallel loop from Ay to Aj or vica versa it is necessary to consider the translation of the branching point via 4; and summator through Ap

(Fig.1.3.13b).

b)

Fig. 1.3.14 Example

1.3.3.

For the system (Fig.1.3.14a) analogously we conclude

that for the translation of Ag in a feedback from A, to Ap and vica versa along the connected in a series stationary linear systems Az, ...

, Ap it is necessary to

54

Chapter 1 Dynamical

consider only the summator

Systems and Their Characteristics

translation via A, and the branching point via An

(Fig.1.3.14b). Example

1.3.4. Block diagram of the linear system with two inputs and

one output (Fig.1.3.15a) may be transformed into the block diagram on (Fig.1.3.15b).

Fig. 1.3.15 Example

input and two outputs

1.3.5.

For the block diagram of the linear system with one

(Fig.1.3.16a) the transformed

block diagram is given on

Fig.1.3.16b.

b)

Fig. 1.3.16 Example

1.3.6.

Fig.1.3.17a presents the block diagram of the linear

system with a parallel loop and the additional branching point in one of the loops. Fig.1.3.17b and Fig.1.3.17c show the sequential stages of the summator translation. Example

1.3.7.

Block diagram of the composed linear system with a

feedback loop having the internal summator is shown on Fig.1.3.18a. The sequential

stages of the summator translation are given on Fig.1.3.18b and 1.3.18c.

1.4. Stochastic Differential Systems

Fig. 1.3.17

55

Fig. 1.3.18

1.4. Stochastic Differential Systems 1.4.1. General Form of Equations of Stochastic Differential Systems Stochastic models of systems involve the action of various random factors. While using models described by differential equations the inclusion of random factors lead to the equations which contain random functions, i.e. such functions whose values at given values of the argu-

ments are random variables (Chapter 2). Eqs. (1.1.4) for a stochastic system must be replaced in the general case by the equations

GAP

( ZEAyeyves GZ),

(1.4.1)

where F'(z,a,t) and G(z,t) are random functions of the p-dimensional vector z, n-dimensional vector x and time t (as a rule G is independent

56

Chapter 1 Dynamical

Systems and Their Characteristics

of x). In consequence of the randomness of the right-hand sides of Eqs. (1.4.1) and also perhaps of the initial value of the state vector 7) = Z(to) the state vector of the system Z and the output Y represent the random variables at any fixed moment t. This is the reason to denote them by capital letters as well as the random functions in the right-hand sides of

Eqs. (1.4.1). The state vector of the system Z(t) and its output Y(t) considered as functions of time ¢ represent random functions of time t (in the general case vector random functions)» In every specific trial

the random functions F'(z,«,t) and G(z,t) are realized in the form of some deterministic functions f(z,z,t)°and g(z,t) and these realizations determine the corresponding realizations z(t), y(t) of the state vector

Z(t) and the output Y(t) satisfying the differential equations (which are the realizations of Eqs. (1.4.1.): z = f(z,a,t), y = g(z,t). Thus we come to the necessity to study the differential equations with random functions in the right-hand sides. In practical problems the randomness of the right-hand sides of the differential equations arises usually from the fact that they represent known functions some of whose arguments are considered as random variables or as random functions of time t and perhaps of the state and the output of the system. But in the latter case these functions are usually replaced by the random functions of time which are only obtained by assuming that their arguments Z and Y are known functions of time corresponding to the nominal regime of system functioning. In practical problems such an assumption usually provides sufficient accuracy. Thus we may restrict ourselves to the case where all uncertain variables in the right-hand sides of differential equations may be considered

as random functions of time. Then Eqs. (1.4.1) may be written in the form

Z = f(Z,x,Ni(t),t), Y = g(Z, No(t),t),

(1.4.2)

where f and g are known functions whose arguments include random functions of time N;(t) and N2(t). The initial state vector of the system Zo in practical problems is always a random variable independent of the

random functions N;(t) and N2(t) (independent of random disturbances acting on the system). Every realization [m1 (t)? no(t)? ie of the random function [Nj (t)? N2(t)"]” determines the corresponding realizations f(z, «, ni(t),t), g(z, no(t), t) of the functions f(z, x, Ni(t),t), g(z, No(t),t), and in accordance with this Eqs. (1.4.2) determine respective realizations z(t) and y(t) of the state vector of the system Z(t) and its output Y(t).

~~

1.4. Stochastic Differential Systems

57

Example 1.4.1. The equations of motion in Examples 1.1.4 and 1.1.5 describe the stochastic models of motion of various mechanical systems if the forces and the moments

in the right-hand sides of these equations represent random functions

of time.

1.4.2. Stochastic Linear Differential Systems The differential equations of a linear stochastic system differ from Eqs. (1.2.1) of a linear deterministic system by additional random items: Z=aZ+ayxr+ag+

ay Ni(t),

Y =bZ+b).+

bi No(t) ,

(1.4.3)

In these equations N;(t) and No(t) are random functions of time (in general they are vector random functions). Introducing the composite vector

random function W(t) = [ Ni(t)? No(t)? |” and block-matrices a4 = [a20], bj = [06,] where 0 is the matrix whose elements are all equal to zero we represent the

random items of Eqs.(1.4.3) in the form a2Ni(t) = a,N(t), 6: No(t) = b},.N(t). Therefore without loss of generality we may omit the indices at random functions and write Eqs. (1.4.3) in the form Z=aZ+ajxr+anp+a2,N(t),

Y=bZ+b0+b:N(t).

(1.4.4)

In practical problems the deviations of a nonlinear system from the required regime may sometimes be assumed as sufficiently small. In such a case the equations of the system may often be linearized relative to random deviations from the required regime and relative to random disturbances acting on the system. In such cases the nonlinear equations describing the behaviour of the system are replaced by approximate linear equations in the deviations. It gives the opportunity to study the nominal regime of system functioning using the deterministic nonlinear

model (1.1.4) and then to study random deviations from the nominal regime using the much simpler linear stochastic models (1.4.4). 1.4.3. Linear Systems with Parametric Notses In practical problems we have to deal with linear systems in which the noises depend linearly on the state vector of the system. In such cases we have to use for the description of the behaviour of a system linear differential equations with fluctuating coefficients. The fluctuations of the coefficients of the differential equations of a linear system are usually

58

Chapter 1 Dynamical

Systems and Their Characteristics

called parametric noises. For such systems the matrices az and 6, in Eqs. (1.4.4) depend not only on time, but are also linear functions of the state

vector of a system. Thus Eqs. (1.4.4) in the case of parametric noises are replaced by the equations P Z=aZ + ajz + ao + (000+ Sanz] N(t),

ia Y =bZ+bo+ (i+ SZ) N(t).

(1.4.5)

P

(p=)

Example

1.4.2.

The motion of a pendulum with vibrating suspension

point in viscous medium is described by the equation

Ag+ Bg + mgl(1 + No)sing + mglN, cosy = 0,

(1)

where ¢ is the deviation angle of a pendulum from the vertical, A is the centroidal

moment

of inertia, B is the coefficient of moment

is the static moment, gN; vertical components

of forces of viscous friction, mgl

= gN, (t) and gN2 = gN2(t) are the horizontal and

of the acceleration vector of the suspension point representing

random functions of time.

At small angles Y putting in (I) sin p & », cos y & 1 and assuming as state variables

p =

Z}, yg =

Zy we obtain a linear stochastic differential system with

additive and parametric noises whose equation has the form

lela alla; ofl S]4) Lm(1) 2, 2

|..0,

ot

Z

2a

ia

0

0

N,

where we = mgl/A, 26 = B/A.

1.4.4. Stochastic Bilinear Differential Systems

The systems in which the vectors Z, Y, Ni, N2 in Eqs. allowing the decomposition of the form ae

|zz

vu

a a

lev tee

aT?

aN

(1.4.2)

[Ni2"0| ,

are called so, with the components Z’, Y' and Z”, Y” satisfying the linear equation of the form Eq. (1.4.3) and the bilinear equations, respectively:

~

1.5. Stochastic Discrete Systems

Z =aZ' +ayx+ao+a.N{, :

Z" =|

P

Act) AZ|

59

Y'=b2Z' + bo +b,N5, P

Z", Y" =|

Bot >> BZ, |2”.

jvesl

(1.4.6)

ivemt!

Here a, aj, a2, Ap, Ax, bo, b1, Bo, By are the matrices, ao, bo are the vectors of the suitable dimension. The bilinear stochastic systems described by Eqs. (1.4.6) are an important special case of the stochastic systems with random parameters,

Z = f(Z,2, Ni(t),U,tY, Y = 9(Z, No(t), U,2),

(1.4.7)

U being a vector of random parameters. Example

1.4.3. The equations of a body motion around a fixed axis in

viscous media have the form

d*o _

Ady

(1)

di? bral ate Here I is an axial moment

of inertia, ) is an angular velocity, A /T is specific coeffi-

cient of viscous friction which is a random time function satisfying to linear differential equation

A=-a\+hN, where @ and hare some positive coefficients,

A=A/T,

(1)

NV= N (t) is a random time function.

Denoting Z; = ~, Z2 = A we reduce the Eqs.

(I) and (II) to the following two-

dimensional bilinear stochastic differential equations:

EVAL

IN

(II)

1.5. Stochastic Discrete Systems 1.8.1. Discrete Systems

It is known that for a discrete insertion of the inputs into a system it is necessary to modulate some parameters of the input impulses in accordance with their values. A device forming the sequence of the impulses which depends on an input is called an impulse element. Usually the impulses form remains invariable at the modulation. We may distinguish three main types of the impulses modulation:

60

Chapter 1 Dynamical Systems and Their Characteristics

e pulse-amplitude modulation (PAM) at which the amplitude of the impulses a; depends linearly (with the gain pam) on the input value at the moment of the impulse action t;, a, = az (tx) = Epama(te); e pulse-width modulation (PWM) at which the duration of the impulse 7; linearly depends (with the gain pwm) on the input value at the moment of the impulse action, T;(t,) = €pwm2(tx);

e time-impulse modulation (TIM) at which the time shift 7 of an impulse linearly depends (with the gain tm) on the input value at a definite time moment, Ts = €yimz(tx). At PAM and PWM the modulating signal changes the area (i.e. the intensity) of the impulses, and at TIM the area of the impulse remains constant. The dependence of modulating parameter of the impulses produced by an impulse element on the correspondent discrete input values is called a characteristic of an impulse element. The latter may be linear or nonlinear. The impulse element with linear characteristic is linear, and with a nonlinear characteristic is nonlinear. The impulse elements vary also in the form and the character of the impulses modulation, in impulses frequency and their relative duration. Usually the impulse elements are periodically generate one impulse at each period. The period of the impulses succession TJ; is called a period of

repetition (a tact) of a discrete system. resents a repetition rate. called a relative duration

The variable w, = 22/T, rep-

The parameters y, y = T;/T, and 1 — y are and impulses ratio.

1.5.2. Characteristics of Discrete Linear Systems Let us denote g,(t) the response of a linear discrete system on a short-time input signal equal to 1 and acting only during the k'® impulse action. Then its response on a short-time signal equal to x(t;) and which acts only during the time of the k*® impulse action will be equal to g(t )a(t~) on the basis of the superposition principle. The response of a linear discrete system on the whole sequence of the impulses modulated by the input disturbance «(t;) on the basis of the superposition principle

(Subsection 1.1.2) will be determined by the formula co

y(t)= S > ge(t)e(te)-

(1.5.1)

k=-00

The functions g;,(t) characterize completely a linear discrete system as knowing these functions we may calculate the response of a discrete

1.5. Stochastic Discrete Systems

61

linear system on any input disturbance x(t). The functions g,(t) determine the part of the input values acting at different time moments t, in the formation of the output at any time moment t. That is why the functions g;(t) are called the wetghting coefficients of a linear discrete system.

Any nonanticipative (physically realizable) discrete system (1.5.1) may response at a given moment t only on the signal acting on it at the previous instant. Therefore for any nonanticipative discrete system

ge(t) =0 at t> ge(ti)a(te). k=—0o

(1.5.10)

64

Chapter 1 Dynamical

Systems and Their Characteristics

For brevity setting 2, = x(tz), yr = (ty), ge(t;) = gie we can rewrite formula (1.5.10) finally in the form

Teng

i OME

ERPS gS

(1.5.11)

k=-00

In the case when the instants of an output fixing coincide with the in-

stants of the impulses action, t; = t; (J = 0,+1,+2,...), formula (1.5.11) at t = t; will give the following expression for the values of an output of a nonanticipative linear discrete system: if

yt =p einGinte fle

a 2 ath he

(1.5.12)

Roa)

Formula (1.5.12) and the analogous formulae for the discrete systems with many inputs and outputs represent, in particular, a digital computer as a discrete linear ystem because the results of the calculations depend linearly on the initial data. In this case the initial data introduced at each step of the calculations are the input signals and the results of the calculations are the outputs signals. The weighting coefficients gj), determine the program of the calculations. At the sufficiently large impulses repetition rate any discrete system may be considered approximately as a continuous one. Really, the response of any system on a short-term unit signal acting during one period of the impulses repetition TJ, at infinitesimal 7}. is an infinitesimal of the order T;,. Therefore the function g,(t) at small T, = At may be expressed by formula

ge(t) © g(t, te)T> = g(t, ty) At.

(1.5.13)

Here g(t,t,) is some function independent of T,. So, for instance, in the case of a series connection of an impulse element and a continuous linear system with the weighting function g;(t,7) we can express the dependence of the impulses of a given form on the period of the repetition

of the impulses assuming the function n(t) in the form n(t) = A(t/T;), where h(€) is a function different from zero only at the interval 0 < €

[v1 yo)", =

[0, Q |

is the input,

0) B, C are constant coefficients deter-

mining inertial, dissipative and position forces respectively. Show that the matrix of

fundamental solutions u(t, wf) and the matrix of transfer functions ®(s) have the form

u(t,7) = e~*-7) cos W_(t — T) + ns sinw,(t — T) x

oo sinw,(t — 7)

2

4

se sinw,(t — T)

&(s) =

cosw,(t — T) — ee sinw,(t — T)

(II)

1 ee

As’ + Bs+C | (9¢ = B/A,w? = w? — €?,w? = C/A).

As°+Bs+C

(111) 1.6. Show that for a linear nonstationary system

i =y2,

Yo=—cet

"yi+2,

t>0,

the elements of the matrix of fundamental solutions u(t, T) ab 2S

(I) Ky |1 = Ac|,

Cal /4 are determined by the formulae:

uii(t,7) = z ds

ja+ 27) (=) — (1-27) (=) |;

watt = Spe [Clon | rt)i/2

Wai(b; Fl ale(< 2 () |,

usatts)= @ (5) a

)-a-2(9)'].

a

~

Problems

91

Find u(t, 7) at c > 1/4. 1.7. Prove that for a nonstationary linear system

Y1 =

UPS

Y=

—t-2y,

—ttyta,

t>

0,

(I)

the matrix of fundamental solutions u(t, T) is equal to

alter)

cos In=t

—711sin In =

: 7 sin Int=

(I)

¢os In at

-z

1.8. Prove that the weighting function g(t, T) of a nonstationary linear system

ytt'yt+(l—n7t-*)y=a2,

y=z,

(1)

is determined by:

g(t,7) = 5[In(7)Na(t) — Nn(t)In(t)]7

(11)

where Jn, (7) and N, (t) are the Bessel functions of the first and the second kind respectively. 1.9. For a stationary linear system ¢

Peaseor, y= wile = 242983

a

hl, B= |2296s

fu

|G

(I)

where a is a constant square matrix, b a scalar constant, y the output, show that the elements of the matrix of transfer fucntions (s) =

=

—(a => sI))~'b at A(s)

la = sl| are determined by the formulae

®11(s) = —b |s? — s(azq + a33) + 422433 — d23432| /A(s) ,

®12(s) = b(—say2 + a12433 — a32013)/A(s) , $13(s) = —b(sa13 + a12423 — 413422)/A(s), o1(s) = 6(—sag1 + 421433 — 431423)/A(s) , ®o0(s) = —b[ s* — s(a11 + a33) + 411433 — 413431|/A(s),

o3(s) = b(—saz3 + 411423 — 421413)/A(s), $31(s) = —b(sag1 + a21432 — 431422)/A(s) , 30(s) = b(—sag2 + 411432 — 412431)/A(s) ,

@33(s) = —b [s? — s(ay1 + 22) + a11422 — ai2a21 |/A(s).

(II)

92

Chapter 1 Dynamical Systems and Their Characteristics

1.10. Show that for the stationary system of Problem 1.9 at 4d}; = @31 = 413

="a39 = 0) 419 = Go3 = lye

—w?, a92 = —2€, 433 = —Q

the elements of

the matrix of transfer functions are determined by

$11(s) =—b[s* + (26 + a)s + 2ae] /A(s),

12(s) = —b(s + a)/A(s), P13(s) = —b/A(s), o1(s) = bwo(s + a)/A(s), O22(s) = —bs(s +a)/A(s), Do3(s )= —bs/A(s), 3) = O35 => 0, $33(s) ) = —b(s? + 2es +.w2)/A(s), A(s) = —(s +a) [we +

(2e + s)s] ,

(1)

Check that the elements of the matrix of fundamental solutions u(t, tp) are given by uia(t, T) =

(- 1/w? ) [uri(t, T) + a/wouar(t, T) a

31 = U32 = 0, u3a(t, T) =

uza(t, ®) iy

exp {—a(t a T)} )

ug3(t, 7) = (—1/w?) {uai(t,7)(1 —

ea /w?) — w[usi(t, 7) — usa(t, T)]} ;

(II) the functions U;1, U12, U21, U22 being determined in Problem 1.5. 1.11. Show that for the stationary linear mechanical system with n degrees of

freedom the differential equations of motion have the following form:

(a) In Lagrangian variables

Ag + (B+ B’)g-+ (C+ C)g=Q;, where

¢ =

og On

[q edn

e is the vector of generalized coordinates,

is the vector of generalized forces. In this case the vector

the output, and the vector

(I) and Q =

[Qi

y = q represents

£ = Q represents the input, A is the symmetric matrix of

the inertia coefficients, B is the symmetric matrix of the dissipative or acceleration forces, B’ is the antisymmetric matrix of gyro forces, C’ is the symmetric matrix of the conservative position forces, and C’ ' is the antisymmetric matrix of nonconservative position forces. (b) In canonical variables

sla feet beh eT ledealaane®

where g and p =

Ag are the vector of generalized coordinates and the vector of

generalized impulses respectively.

In this case [q? pt]? represents the state vector

of the system, the input and output being the same as in the case of the Lagrangian variables.

Problems

93

1.12. Show that for the stable stationary system of Problem 1.11 at n = 2

a Ai2

Ai2 es i A22 | | Qe a

Bi, Biz

Bi Boe

4

0 —Byan»

Ci1 Cie | OC 12 (|Cho, C22 dare i2 O

Bip 0

q1 q2

=

qi q2 Q1 Qe |

(1)

the elements of the matrix of transfer functions ®(s) are determined by the formulae

11(s) = {Aj ({C| + C42) — s[Cii(C22 +5)

— (Biz + By2)(C12 — Ch) /A(s), 12(s) = -{Aj,(|C] + C'ts) — s[C11(Biz — Bip) — (Bi + s)(Ci2 — Cia) }/A(s),

o1(s) =—{-Aj,(|C] + C42) + s[(Ci2 + C{.)(Baa + 8)

(II)

— C2o(b12 + Biy)|}/A(s), o0(s) ={AZ(|C] + Cq2) + s[(C12 — Clp)( Biz — Bip) — C2(Bi1 + s)]}/A(s) where A(s) = s*4 +1159 + 1728? + 173s +14,

ry = Byy Az, + BorAj, — 2B12 Aj, ,

ro = [BiAg — (Bia + Biz) Ate] |Bo2Ary — (B12 — Bip) AyD] + (C1149 + C22Ay, — 2C12Ajp) , r3 = [By Aj — (Bio + Biz) Aja] [(Bi2 — Biz) Ad, — B22 AQ | a [ BooAy

(Biz wy By») At | [Cir Ax x (Ci2 im Cia) Ain

“+ [(Biz - Bi) Are — B22 Ajp| [CiAi

a (Cie te Cya)An

+ {[ Bi Az, + (Bio + Bia) Ary] + [(Bia + Bi) AT — Br Ax] } x [C2245 — (C12 — Ci2) AQ] , ra = (IC] + Ale Aj; =

Agen

|A|

(4,9 =

|We 3

|A|

=

Aj1A22—A?,

IC | =

(111)

CiiCo9—C-,

1.13. Show that if the matrices A~! B and A7~ 1C in Problem 1.11 are commute and B’ = 0, C’ = 0, then there exists always a linear transformation of the output

y = L¢ satisfying the conditions

TWA fe dele Blas diag (266s 5... 2En),

TE OP

diag {uen rasu2 Js,

I

94

Chapter 1 Dynamical

Systems and Their Characteristics

reducing the set of equations of motion to n independent linear differential secondorder equations

Cy + 2enCn

where

Up

=

up

,

(h7= 1)

ane

un(t) is the A*® element of the column-matrix

corresponding equations at 1.14.

+2 Cn =

B = Meys\. ei

Consider a mechanical

(II)

EO}

Write the

nC :

system with n degrees of freedom

described by

Lagrange equations

d

(OL

OL

a (i) — 3 = 2 where q =

-

[q1 soli ‘ieis the vector of generalized coordinates,

L = T'— II is the

Lagrange function, 7’ = T(q, q, t) is the kinetic energy, I] = II(q) is the potential

energy; Q =

[Qi MGS

servative forces.

ie Q.=

Show that Eq.

Q(4; q; t) is the vector of generalized noncon-

(I) may be transformed into the set of Hamilton

equations in canonical variables gq and

G==-, Here

p = OT /Oq

p=->-+Q.

H = H(q, p, t) being the Hamilton function,

(II)

H = pq -T.

1.15. The vertical oscillations of the body of a car while moving on the rough road are described by the equations

My + 20(y — ¢) + 2v(y- 9) =0, where M/ is the car mass,

w(y = q) and ply = q) are the nonlinear functions

determining damping and restoring forces respectively,

g =

q(c) is the function

characterizing the microprofile of the road in the vertical plane,

7 =

vt, VU being

the car velocity. We assume that both pairs of wheels of the car undergo the same vertical displacement at every instant due to the roughness of the road (i.e. the road

profile in the vertical plane has a very long correlation interval).

Find the transfer

function of the car at equalities w(y = q) = b(y = q) and ply = q) = c(y = q); b and c being constant coefficients, assuming q as an input and Y as an output. 1.16.The equations of motion of the body with one degree of freedom equipped

with the nonlinear dynamic damper have the form

pa + ba + cx + pa(@ — Za) + vale — vq) = F(t),

Hata — pa(@ — £a) — Ya(x@ — 2a) = 0,

“wn

Problems

95

where & and £q are the displacements of the body and of the damper, pi and [lq their

and C the coefficients of viscous friction and of restoring forces respectively,

masses,

wa(x = La) and a(x = La) the functions characterizing the forces of viscous friction and the restoring forces, and F(t) the disturbing force. —%q=

a(x

ba(« _

La) and pa(r = La) =

In the special case where

cae. — La), with bg, cq being

constant coefficients find the transfer function of the system considering the force F(t) as an input, and the displacements © and &q of the body and damper as the components of the output.

1.17. Show that the nonlinear equation n—-1

Oe

Nea

See yy

fee, et)

k=n-—m

m

=D da(ysy’, 0. yO", 1), (m 0 that P(Q) = 1.

If a nonnegative finite or o-finite measure v is determined in the measurable space (Q,S) then the probability P = P(A) may be presented in the form of the sum of the v-continuous and v-singular measures, and the v-continuous part represents an integral over the measure v with

respect to some function f(w)

P(A) = | tev@s) + P(A).

(2.1.1)

A When the probability P(A) is v-continuous, P;(A) = 0 and

P(A) = :;Jobin

(2.1.2)

A

As it is known from the functional analysis in this case the function f(w) represents Radon-Nikodym derivative of the probability P with respect to the measure vy and is called a probability density with respect

to measure v in the space (2S, P). Any subset N of the set No € S for which P(No) = 0 is called a zero

set.

If zero set N belongs to the o-algebra S of the probability space

(Q,S, P) then, evidently P(N) = 0. But zero set does not need belonging to S (if No C S then the subsets of the set No do not necessarily belong

to S). On such zero sets the probability P(N) is not determined. The probability space (Q,S, P) is called a complete one if the o-algebra S contains all zero sets of this space. If the probability space (Q,S, P) is not complete then it may be always supplemented after adding in S all zero sets and all the unions of zero sets with the sets from S and determining the probability P on the supplemented sets.

102

Chapter 2 Random

2.2. Random

Variables

Variables, Processes and Functions

2.2.1. Definitions

In elementary probability theory a random variable is a variable which assumes as a result of a trial one of possible values and it is im-

possible (before a trial) to know what value it will assume namely. Here it is supposed that in the space of possible values of each random variable there exists a probability distribution. In order to come to an exact definition of a random variable notion we shall refer to the examples from elementary probability theory and random functions theory. Example

2.2.1. The number of the occurrences of the event A (in one

trial) in Examples 2.1.1 and 2.1.8 represents a random variable which assumes the value 1 when A appears and the value () when A does not appear. It is easy to notice that this random variable represents a function of a point in the probability space

(Q, S:, iP) of Example 2.1.8 equal to 1 at w = W, and 0 at WwW = Wo. Example

2.2.2.

nm triales in Bernoulli 0,1,...,m.

scheme

This random

The

number

represents

of the occurrences

a random

variable

of the event

A at

with possible values

variable is a point function in the probability

space

(Q, Ss) P) of Example 2.1.9 equal to.m at all points W to which correspond the se-

quences

A,,...,

=O

lees8):

Example sidered

An

containing

letters

A

and

n



m

letters

A

(m

2.2.3. The finite-dimensional random vector X may be con-

as the point function

Example 2.1.10.

‘m

r(w)

=

wW in the probability

space

(Q, On P) of

Each component X, of the random vector X represents also the

point function r,(w) in the probability space (Q, S, Vane Example

2.2.4.

Infinite-dimensional random

vector X

(the sequence

{xX x} of the scalar random variables) may be also considered as the point function z(w) = w in the space QQ of all numerical sequences

{xz}. It is evident that any

component of an infinite-dimensional random vector is the function of a point in the space 2 (Lm at any fixed M

is the function of the sequence

{7x }): The same is

true, in particular, for the sequence of the events in Example 2.1.6.

The considered examples turn us to the following definition of a random variable. A random variable with values in a measurable space

(X,A) is called (S,A)-measurable function of a point in the probability space (Q,S, P) mapping Q into X and determined everywhere in Q besides may be the set of zero probability P. The value of a random variable at any given point w of the space (2 (i.e. the value which it assumes when as a result of a trial the elementary event w appears) is called a realization of this random variable.

2.2. Random Variables

103

It is evident that any random variable in (X, A) is at the same time a random variable in (X,.A’) where A’ is any o-algebra in X containing APA’ eA: We shall often denote a random variable by the same letter as the space of its values. While considering the set of the random variables with the values in one and the same space the denotations of these variables will be supplemented with the indexes. So, for instance, X, Xq represent the random variables with the values in the space X. If X, Xq are the random variables in (X,.A) then according to the definition X

= z(w), Xq = Lq(w) where x(w), w(w) are (S,A)-measurable functions mapping 2 into X. a 2.2.2. Probability Measure

By virtue of (A, S)-measurability of a random variable

X = 2(w)

in (X, A) the probability of the fact that it will take the value from the set A is determined by the following formula:

P(X 6A),= Plar(A))sucAce A, where as usual z~1(A) is an inverse set A in the space 2. Formula

f,(A) = P(X € A) = P(27"(A)),

AEA,

(2.2.1)

determines the probability uz, induced by the random variable X = z(w) in the space of its values. The triple (X, A, zz) represent the probability

space induced into (X, A) by the random variable X = a(w). The space X of the values of the random variable X = a(w) is called its phase space or the space of the realizations. The probability pz in the phase space is called the probability measure or the distrubution of the random variable X. The probability space (X,A, uz) is called the phase probability space of the random variable X = x(w). It is evident that any probability space may be considered as a phase space of some

random variable. If the random variable

X = «(w) is considered in its

phase probability set then the argument w in its denotion naturally is not indicated. In such cases we shall denote it simply by X and its realizations by z. If the nonnegative finite or o-finite measure v is determined in the

space (X,.A) then the probability measure pz of the random variable X may be presented by formula

entAye= /f(a)v(dz)+ps(A), A

AEA,

(2.2.2)

104

Chapter

2 Random

Variables, Processes and Functions

where (A) is the v-singular measure. In the special case when the measure jl, is v-continuous, 5(A) = 0, we have

Ay= [semwees

“Aten

(2.2.3)

Here f(x) is Radon—Nikodym derivative of the probability measure pz with respect to the measure vy which is called a probability density of the random variable X with respect. to the measure v. Remark.

Inelementary probability theory the probability Aensity of

variable is always considered in a finite-dimensional phase space (Example

arandom 2.1.10).

In problems which are connected with one random variable it is always convenient to assume its phase probability space as the probability space.

But in the problems

connected with the set of the random variables (possibly with different phase spaces) it is expedient to study all of them in one probability space. Example

2.2.5.

The two-points set {0, Hi serves as the phase space

of the random variable in Example 2.2.1, and the probability measure is completely

determined by its value in one of these two points: a0

Hae({1}) = 0), Hr ({0}) =),

lle

Example

2.2.6.

The sett {0,are

as n} serves as the phase space of

the number of the occurrences of an event in 7 trials in Example 2.2.2, and the probability measure [/, is completely determined by its values in these points: Ha ({m})

SOR

pt ote

T(t = nl) ml i — a)

Example

2.2.7.

In Examples

sO:

cehatt

2.2.3 and 2.2.4 the space of elementary

events () itself serves as the phase space of the complex (correspondingly real) random

vector X, and the complex plane (correspondingly real axis) serves as the common phase space of its components.

2.2.3. Equivalence of Random

Variables

Let X = z(w) be a random variable with its phase space X. The inverse mapping z~'(A) of the o-algebra A represents the o-algebra which is a part of the o-algebra S in the space of elementary events

, «1(A) Cc S. In the general case 2~!(A) does not coincide with S. Thus the variable X = x(w) induces in the space 2 some o-algebra

S; = a '(A), as a rule more poor than the o-algebra S. Therefore the random variable mined

on the same

Y = y(w) in the measurable space (Y, B) deterprobability space

(Q,S, P)will be not obligatory

(Sz, B)-measurable (from the fact that y~!(B) € S does not follow that y~'(B) € Sz). So if the random variable Y in (Y, B) whose phase space

a“

2.2. Random

Variables

105

Y being a separable B-space is (S,,B)-measurable then it is (A, B)measurable function of the random variable X. The random variables which are equal almost surely, i.e. with the probability 1, are called equivalent. Thus two statements are valid.

Theorem 2.2.1. If the random variable Y = y(w) in (Y,B) whose phase space Y being a separable B-space is (S,,B)-measurable then it ts equivalent to some (A, B)-measurable function of the random variable

aX eel Ga) Theorem 2.2.2. Jf (A, B)-measurable function valent to such a function) there are none restrictions

the random variable Y = y(w) in (Y,B) ts of the random variable X = x(w) (or ts equithen it is (S,,B)-measurable. In this case on the phase space of the random variable Y.

It is evident that all equivalent random variables have one and the same probability measure.

If the random variable Y = y(w) in (Y,B) is (A, B)-measurable function of the random variable X = x(w), Y = y(X) then it is arandom variable in the probability space (X,A,p,).

Therefore the probability

measure fly of this random variable may be expressed in terms of the probability measure pz of the random variable X by formula (2.2.1)

Hy(B) =ps(p*(B)),

BeB.

(2.2.4)

The same formula determines a probability measure of any random va-

riable Y equivalent to (A, B)-measurable function p(X) of the random variable X. 2.2.4. Two Types of Convergence of Random Sequences The sequence of the random variables {X;}, X; = %,(w) with the values in the topological space is called a.s. convergent (almost surely, with probability 1) to the random variable X = z(w) if the sequence of

the functions {xz,(w)} converges to «(w) almost everywhere on {2 relatively to the measure P. The sequence of the random variables {X;}, X, = xx(w) with the values in the separable B-space X is called convergent in probability to the random variable X = x(w) if at any € > 0

lim P(\| Xe — X [I> €) = Jim P({w : |]24(w) — (w) ||> €}) = 0. The following statements are valid.

106

Chapter 2 Random

Variables, Processes and Functions

Theorem 2.2.3. For the a.s. (respectively in probability) convergence of the sequence of the random variables {X;,} with the values in the separable B-space X it is necessary and sufficient that this sequence will be fundamental a.s. (respectively in probability). Theorem 2.2.4. Any a.s. convergent sequence of the random variables {X;,} with the values in the separable B-space ts also convergent in probability. The inverse is in general case not true. But from any sequence of random variables convergent in probability we may separate such sequence which a.s. converges. Example

2.2.8.

Let Xn be the random variable with two possible va-

lues 0 and n whose probabilities are equal to 1 — n-? and n~? respectively.

The

sequence of the random variables {Xn} corresponding ton = 1,2, ... , converges in probability to 0 as at any € > 0

P(|Xnl

1

26) = P(Xn =n) ==Z70

at

noo.

n

This sequence also a.s. converges as at any €,0 < € €

ifevenatone

n>p)
p)

PUXal

Let (Xe A) be an arbitrary measurable space, r(w) bea

(ST, A) measu-

rable function in infinite product of the spaces Q? with the values in X. Suppose that

there exists a countable class of the sets { Fi,} generating the o-algebra A. io

PR

Then

will be a countable class of the sets in Q7 generating the o-algebra

te (A) & S™ induced in QT by the function z(w). There exists such sequence

of the countable subsets {Ly} of the set 7’, that zo We put L = ULx. Then we shall have a—1(F,)

such countable set

L C T' that the o-algebra S

(k == pe Pe a As Zor foe (A) e So

(F;.) e SU

(k = lade. ne

E S” at all k. Thus there exists

contains all the sets a= 1( Pi)

A) is the minimal 0-algebra containing all ae

Oy) then

Consequently, the function r(w) is (S™ A)-measurable. Thus any

measurable function z(w) in Q7 with the values in the measurable space (X, A) in which the o-algebra A is generated by the countable class of the sets is (S¥ . A). measurable at some countable

D C ‘Ts i.e. depends on no more than a countable set

2.5. Probabilities in Infinite Product of Spaces

TE

of the coordinates of the point W in OF eat particular, any measurable functional in ee depends on no more than a countable set of the coordinates of the point W in

Ors Now

we pass to the construction of the probability in OD Suppose that at (Q¢, Sz) the probability P,( Az) and the regular conditional probabilities Paster els (At lw, fonds wi, ) correspondent to all hy, meal ad. (n each space

Sly dives .) are given. ‘Then applying formula (2.4.18) sequentially we find the probabilities in all possible finite products of the spaces {);:

Patt, Peso

As) = ijPtg,t, (At, lw) Pr, (der) ,

th lad XtvenXaAen)

=

Nae

At, XX At, x Ps...

oe

ane

(2.5.1) eer

Ee

he

ER

| Wr—1)

1

tn-1

(dw, ».G VOR

iegee ent

DS dwn-1)

a Space fice oeeng

After determining the probability 1 ,tz on the measurable rectangles Os, x oF by

the first formula (2.5.1) at first we shall extend it on the o-algebra St, Xx Sia Then using the second formula on the measurable

(2.5.1) at

m =

3 we detemine the probability Psy ,ta,t3

rectangles Qs, x Q:, x Qt, and extend it on the 0-algebra

Si, x St, x S; ,- Continuing this process we determine consequently the probabilities Pay peeytn atn=

4,5,....

It is easy to understand that the conditional prob-

abilities F eeks ae, td (Azur, tnd ies) (t, Ls

Gh

ote

i Seatise :) should be

consistent in such a way that they will give one and the same probability tae eeeat ss in the product of the spaces OQ, DK ores XK ey

(n Be 2, ose: Ps in any ways.

Then

the obtained set of the probabilities in all finite products of the spaces (Q4, S:) will satisfy the conditions of the consistence and the symmetry.

Let us assign now the function of the set P on the set of all measurable rectangles of the infinite product of the spaces Qr by formula Pe

where

Ae

x

Pax

Ay),

(2.5.2)

Ft is the rectangle with the sides At, seuss. 5 At, which le in the spaces

MtHe cme

Wem correspondingly.

set of the rectangles may

This function is additive as the bases of any finite

be considered

as belonging to one and the same

product of the spaces, for instance Sire X:++:& is additive in (Q4, XX

finite

G2 me and the probability jb sete

Ue ; St, X-22X St, MesFurthermore if we restrict ourselves

128

Chapter 2 Random

Variables, Processes and Functions

to measurable rectangles whose all bases belong to one and the same finite product of the spaces Q:, KISLEX

Q:, then the function P(R) is O-additive by virtue of

the @-additivity of the probability lady ,.:,tn+

Consequently, the function P(R) is

the function on the set of all measurable rectangles of the space (a>, Sf) which have the basis in one and the same finite product of the spaces Q:, Xoo

X GPAs

According to the theorem from functional analysis about the measure extension it may

be uniquely extended over the d-algebra St, x

cylinders of the space Q? with the bases in OF > ea

aX

St, of all measurable

Chey.4 A

It is evident that the

extension is determined by formula

PCR

re

Sa

eee

Bye)

(2.5.3)

where Cr, ,.1, tn is the cylinder in Q7 with the measurable base Bi, Oo. eke (Q4,

KPO

Seen SE):

Let us consider now the problem of the possibility of the function P (R) exten-

sion determined on the class of the measurable rectangles of the space (QT ; Sy by formula (2.5.2) on the whole o-algebra SE. The class C of all measurable rectangles is a semi-algebra and satisfies the conditions of the known from functional analysis theorem about the measure extension. In order that it may be possible to use the theorem about the measure extension for determining the probability on the whole o-algebra ST it remains only to prove

the -additivity of the function of the set P on the class of all measurable rectangles

which is determined by formula (2.5.2).

For this purpose it is sufficient to show

that P is continuous in zero, i.e. that for any decreasing sequence of the rectangles eae gsRp a) Ry at p < q with the empty intersection lim PCr

= 0. Hence it

is sufficient to consider the case when the set of the spaces {Qa} in which lie the sides of the rectangles Ry, is infinite (the case of a finite set of the spaces {Qz, } we have already considered).

We denote by AM*) the side of the rectangle Ry, in the

space Oe and put BS) = AbE+) KATE

Alt) where 7, is the number of the

sides of the rectangle R,,. Applying formula (2.4.18) we may present formula (2.5.2) at

R=

R, in the form

P(Rn) = / P(BO\or)Pa (der) = f1400 es)P(Blen) Pa, (der), AY) where

1

P (Bo Nw)

is

Saye o Se Dy au St, y I
0 that Pia)

> € atalln.

As

Pe thts) = | then it follows that there exists such point @] in Q;, that f(@1) > €. It is evident that ©, E€ AW at all n as f(#1) = 0 if wy does not belong to AW

beginning with some

{P(BS

mn.

But the sequence of the conditional probabilities

jz) is nonincreasing and lim P(BY



a, =

f(@1).

Therefore from

ds) ==

f(@1) > € follows P(Bs No) > € at all n. Further on the basis of (2.4.20) we have

P(BY ay) = i P(BY) |, w2) Pes ,19(dw2|@1) ,

(2.5.4)

A?) where

She

P(BY

ay, we) is a conditional probability in (Q45 X-++X

St, ). From P(BY?

Here Wy E Awe

St, :

a) > € follows the existence of such point W2 in Qy,

that P(By? lo, We) > €atalln andW2 we shall get the sequence

Qe,

€ Al?) at all n. Continuing this process

of the points {@;, } in the space Q4, (k aan

Np

he 4 3)

Consequently, all the points W of the infinite product of the space

QT? which have the coordinates Carn

ays Wp in the spaces ay Pee

to all rectangles Rt, whose sides lie in the spaces Oe Rota

Qe, belong

Q4,. Hence it follows

that the points of the space Q? which have the coordinates W. k in all the spaces

8

(k ct

.) belong to all the rectangles R,,, i.e. the intersection of all the

rectangles R,, is not empty. So at lim PCRs) of 0 the intersection of all the rectangles R,, cannot be emply. Consequently,

lim P (Rn)

=

0 if NR,

=

@

what proves the continuity of the

function P in zero and at the same time its 0-additivity on the set of all measurable rectangles of the space Q?.

Thus formula (2.5.2) determines the probability in the

space QT on the set of all measurable rectangles C. Hence there exists an unique extension of the probability P determined by formula (2.5.2) on the 0-algebra st

generated by the set C of all measurable rectangles. It follows from formula (2.5.2) that the probability es ,.:,tn projection of the probability P on the subspace (Q:, X-°:X

represents the

eae Phas ae St, )

of the space (OF. Siu. Hence it follows that the probability Py, in the subspace (Stee ars) of the space vine So}, LC of the probability P on (QO% Soh

T determined in such a way is the projection

130

Chapter

2 Random

Variables, Processes and Functions

The requirement of the regularity of all conditional probabilities in each space

is essential as formula (2.5.4) and all analogous formulae which are necessary for unbounded extension of the process in given proof of the continuity of P in zero have the sense only if all conditional probabilities (Qe, S:) are regular. Now let us consider the random vector X = x(w) with the countable set of the components X;

=

Lp (w) (k i

Me

a oD Suppose that on the 0-algebra of

Borel sets A of the phase space of this random variable

X its conditional distribution

pF(Alw)

relative to some O-algebra F C S of the probability space (Q, Se P) is

assigned.

The space (X ; A) may be considered as the countable product of one-

dimensional spaces Ce 5 Be):

x=R?=][

ke,

A=B”

where

Ry

= R

axis (k sl

is the real axis; By

=J] Br, k=1

pel

=

Wap ae Ns In this case we may

B is the o-algebra of Borel sets of this ensure that any countable-dimensional

random variables have conditional distributions. In other words there exist the regular conditional probabilities determined on the correspondent ¢-algebras of Borel sets in any countable-dimensional probability spaces.
Finally let us consider the random surable space

variable X

(Xx; A) and its conditional distribution

=

z(w) in arbitrary meaF(Alw)

relative to some

g-algebra F C S of the sets of the probability space (Q, S, Py; Suppose that there

2.7. Probability Measure of Random

Function

133

exists such finite- or countable-dimensional random variable (Q, A, ft) in the probability space and that the o-algebra A,, = ua? (BN) induced by it coincides with

A. According to (2.2.4) formula

v¢(Blw) determines

=pr(u-'(B)lw),

Be B",

the conditional distribution of the variable

UV =

u(w).

(2.6.4) In conformity

with the proved theorems there exists the regular conditional distribution v'-( Blw) of the variable UV. Naturally the question arises: whether is it possible to determine

the regular conditional distribution Hr (Alw) of the variable

X by formula (2.6.4).

As by supposition any set A € A is thé inverse of some set B € i

then formula

p'p(Alw) =ve(Blw), A=u-1(B),

determines p'-(Alw) for all sets

A € S .

determines Ue (Alw) uniquely not for all

u—!(B)

(2.6.5)

But in the general case this formula

A € S. For the existence of the regular

conditional distributions of the random variable X it is sufficient that such a finite-

or countable-dimensional function u(x) in (X Ai He) be measurable relative to A and the corresponding ¢-algebra of Borel sets B” will exist such that its range A, will be Borel set and the o-algebra A,

=

uw? (BE) induced by it will coincide

with A. As any probability space may be considered as the phase probability space of

some random yariable then for the existence of the regular conditional probabilities

in the probability space (Q, S, i it is sufficient the existence of finite- or countabledimensional random variable having Borel range and inducing the o-algebra S.
p-1 dkyk, Where the coefficients equations

d,...,d,

Ms Kepnasedp ah (eygt)

satisfy the linear

(ease... 1),

k=1

It is easy to check that formulae (2.7.1) give the same normal multidimensional distributions of the random process X;. Consequently, the random process X; is normally distributed and its probability measure is determined by formula (2.7.12). Remark.

These results are easily generalized on the random

processes

with any consistent set of the multi-dimensional probability densities f, (a1 sitar

t;,...,tn)

(or

with

any

consistent

set

of

the

conditional

frl@i thoi, so, GagSiss+ 4, Sn). end my thas care ai(ayes

=

fr(eisti) fr @o5t2|21;th) oA Gran (nite les, cap tn_15

gs

distributions

re thao

158.

a2 1))-

The probability measure of a random process of the set of all measurable cylinders in the functional space (Xxf : Al) is determined in the general case by formula

(2.7.16) By the formula of such a type we may also approximate with any degree of the accu-

racy the value of the probability measure of the random process on any measurable

set of the space (Xe ; ALY.

2.7.7. Streams of Events Let us consider some notions of the theory of the stochastic streams of the events which are important for the stochastic systems of queueing.

A stream

of the homogeneous events may be characterized by the distributions

of the random instants of the events occurrences.

Let us assume JQ as the initial in-

stant and suppose that the probability measure 1’; (A) of the moment 7} of the occurrence of the first event and the conditional probability measures l,, (Alt, sess ini)

of the instants of the occurrences T;, (n = 2, 3, . . .) of all following events are known. All these conditional distributions are regular as they are determined on the d-algebra B of Borel sets of the real axis R (Section 2.4). Considering T;, as a random function

of an integer argument NM we may determine its multi-dimensional distributions by formula

~

2.7. Probability Measure of Random

Pe ates,

ON

“ Un

d(T,

Function

2111, --

143

5 ine)

An-2

x ‘h Bie

fiat

net nies I) [Ti so*

the o-algebra B™

D4

Bx.) we may extend uniquely this measure on

of the space R™.

Let us consider a special case of a homogeneous

stream in conformity to the

recurrent stream of the events. Recall that a stream of the events is called recurrent if all the intervals between adjacent events are the independent random variables. Below (A = T) means that as usual a set of the numbers obtained by the subtraction of T from all the numbers

contained in A.

Then for such a stream of the events

with one and the same distribution lV of all the intervals between adjacent events we

have v1(A) = v(A), Vn (Alty, i. ti 2 40) = Uy(A — boa) (n eee

.) and

formula (2.7.17) for 1/1...» will take the form

Vi nar = [van Ai

CAR|

/ v(dt2)... A2-T1

/

Vi An — TmO1)v(dtme1)-

An-1-Tn-2

(2.7.20)

If there exists the probability density f(t) of the time interval between adjacent events

the

(A= ffet, A

then formula (2.7.20) may be rewritten in the form

".n(Arx:-*xXAn)=

eer

[-f(n)dn | f(t2—m1)dr::.

eee

| f(tm—tm-1)dt.

Cmpay

144

Chapter

2 Random

Variables, Processes and Functions

It is convenient to connect the stream of the events with the random function

X (t) which represents the number of the events occuring at the time interval [to stl. Evidently, all the realizations of this random function are the step functions which increase by unit jump at each instant of the occurrence of the next event (by 7 if the

moments of the occurrence of n events coincide). Let us find the multi-dimensional distributions [l4,,...,t

of the random function X (t). As at any moment

n

ft the ran-

dom function X (t) may have only integer non-negative values then for determining Lt,...,tn it is sufficient to find the probabilities



n

Pty,..utn(™M1,---,™Mn) = P( () {X (te) = me})

(2.7.22)

k=1

for all m;, = 0 Aleem)

(= 9ny ans

gS

But these probabilities are easily determined by means

ae |, of the known

ee distributions

Vky,...,bn Of the instants of the events occurrences found before. It is easy to see that j EL ermal

eC

CDSs)

Mn

)—

i

x(ti5 te |(t2, ta|) 6

ey

RO

els MD

(tant

ag!

5 ¥ Gesip ty ]

ene CO)

271-20)

For the recurrent stream of the events formula (2.7.23) takes the form ty

.

ti-T1

ti-Tm,-1

Playtest) = f(dn) ff v(dn)... farms) to

0

t2-Tm,

x

t2-Tm,+41

iA (dtm, +1)

‘) (dT m,+2)

ti-Tmy

0

tn-Tmy—y

Dee

eS

0

tn-Tmy—y

ii

VAG fe

mare

ae

/

tn-1-Tm,_1

(oe)

VA diay)

0

In the special case of the stationary Poisson stream

/ tn-Tmn

of events

Claes

e

(2.7.24) f(t) =

\ew*#

where A is the intensity of a stream (the mean number of the events in an unit of

time) formula (2.7.24) gives Pty,..,tn(™1,

AM (tr —

ito

2 eens Mn)

I

(tare ty) eee alt A4 eee =! “Reh sts) m,!(mz —m})!...(™Mn — Mn-1)!

oe

OC

OO

n

:

(2.7.25)

~

2.7. Probability Measure of Random

Function

145

(0< mi) e(h)}) < 6(h)

(2.9.13)

that at any h € (0,7)

at allt,t+7 €T,|| 7 ||=A (The Kolmogorov—Loéve criterion). > For proving that at the fulfillment of conditions (2.9.12) and (2.9.13) almost all realizations of the random function w(t w) are continuous it is sufficient to show that almost at all w

sup

|| z(t’,w)—a(t,w) || 70 at h-0.

t,t’Er

|| t/—-4 ||e})

= 0at7 50.

Therefore any countable set L dense in T is a separant for a(t, w). In particular, the set L of all the points of the region T' with the dyadic-rational coordinates may serve as a separant, i.e. with the coordinates of the type k2~”, where k is any integer,

and p is any natural number.

Thus we may restrict oneselves by the values “ACs w)

on the set I, on the points 7 with dyatic-rational coordinates. Let us introduce the variables

eee lO

spe

SES

Pn

el et

gee Kec

Fy ae

BO)

hn wt 8 eee i)

|

(ea

where the upper bound is taken over all the points | =

(E12

—:

Ke2e* ) ET

and over all q from | till n. It is evident that Zp 2 e(27P) if and only if when even one of the variables

SOE Oo mE ie is larger or equal e(27?).

kg el) eee, KDA)

ie

WEAR

pes

ics instar)

a AE on

Nall

For calculating the number of the variables ae

we denote by 7}, Lebesgue measure of the projection of the set 7’ on the axis m

(m FS

ek n). Then the number of different values of the m*? coordinate of

the type k,,27? in the region 7’ does not exceed 2?T,,.

(ki2—*, aloe E22)

So the number of points

in the region T' does not exceed 2"?7; ...T7;,.

such point corresponds 7 variables ed the random variables Yes ;,,

As to each

oe (q — 1) ree n) then the number of

does not exceed n2"PT, ...T;,. So on the basis of

(2.9.13) we find

Pepe eOPt)) Shy PIV as

beatal2 4) < per PGT, 622). nr

Let us introduce the events

Ap = {Zp = €(2°?)} = {w : Hw) 2 (2°? )},

164

Chapter

2 Random

Variables, Processes and Functions

B, = U 4p, N=limB,= () By. p—=4q

quik

According to the proved above PAG) co

PLB, ys SD EPCAD

1 then by virtue of (2.9.12) P(N) = (). Thus the set NV co

is zero. Ifw € N thenas N = UB, then W € By at any g. But B, =

() Ap.

p=4q Therefore W belongs to all Ap at p > q. It means that Zp = Zp (w) < efr?) at any point WEN

at all sufficiently large p.

Let us take now ie ial odbr

3 | v-l

two arbitrary points I, l! with dyadic-rational

coordinates,

||< 271 then there exists such point lp) with the coordinates

k127-4,..., kp, 274% that all the components of the vectors of the differences | — Ip and I’ — Ig is smaller than 2~%.

Here the difference of the mth coordinates of the

points | and Ig will be expressed by the dyadic number of the type Im

a ys eyaeeS el Rea) B p=qtl where each of the numbers Tp

is equal to 0 and 1. If we perform the transition from

the point /p into | by the sequential steps of the length Tmp Dak (p = qt+l1, q4+2,.. .)

at first on one axis further on the other and so on, on the n*® axis and use repeatedly the inequality of the triangular then we get

ICD En

ouk eae m=1

seme p=q+1

ae nO: p=qtl

The similar inequality we obtain for | z(l’,w) = t(Ip,w)

|. Consequently,

|| (0) —aw) || e}) = 0.

(2.10.2)

The random function X(t) = «z(t,w) is called differentiable in probability on

T if it is differentiable in probability at all t € 7’.

The random function X(t) = a(t,w) is called a.s.

differentiable

(differentiable with probability 1) at the point t € T if there exists such

random function Xj,(t) = x/,(t,w) that

P({w: ah (t,w)

a! (t,w)

at

h—-O})=1.

The random function X(t) = a(t,w) is called a.s.

(2.10.3) differentiable

(differentiable with probability 1) on T if it is a.s. differentiable at all teT.

The random function X(t) = x(t, w) is called differentiable if almost all its realizations are differentiable: P({w:2(t,w) Remark.

differentiable on

T}) =1.

(2.10.4)

It follows from the general properties of the convergence

in

probability and a.s. (Section 2.2) that the random function a.s. differentiable is also differentiable in probability.

The inverse in the general case is not valid.

It is also

evident that the differentiable random function is differentiable a.s., and consequently, in probability.

But it does not follow from the a.s.

differentiability of a random

function that it is differentiable as the uncountable union of zero sets Ny =

{w ;

oP (, w) ee oh (ie w)} at h — 0 even at one m}, t € T may be not zero set.

168

Chapter 2 Random

Example rentiable.

2.10.1.

Variables, Processes and Functions

The random process X (t) of Example 2.7.2 is diffe-

Consequently, it is differentiable a.s.

and in probability.

The same refers

to the random process X(t) of Example 2.7.4. Example

2.10.2.

The random process Y(t) of Example

2.7.3 is a.s.

differentiable, and consequently, in probability at any ft “2 Q. Therefore it is a.s. differentiable and in probability at any interval 7' which does not contain the origin.

But it is not differentiable on any interval (to, t1), to > 0, t) > 2to as none of its realizations corresponding to |w|

  • e}) =0.

    (2.10.5)

    As it is easily seen this condition imposes the restriction only on the

    three-dimensional distribution of the random function X(t) = x(t,w). Knowing its three-dimensional distribution s4;, 4,4, we may present condition (2.10.5) in the form 5

    lim

    h,l0

    %2—-

    1

    He,tbembstbemt (4(15 #258) : |Se

    h

    T3—-— ZL

    Seni tans|2 -}) =

    (2.10.6) As a result it is always easy to determine whether a given random function is differentiable in probability or not. As regards to a.s. differentiability and the differentiability of a random function then these properties cannot be established directly by means of its multi- dimensional distributions. But there exist sufficient conditions of the differentiability of a separable random function which are analogous to conditions

    (2.9.12) and (2.9.13). 2.10.2. Sufficient Conditions of Differentiability

    Suppose the separable random function X(t) = a(t,w) with the values in the separable B-space determined in the finite region T' of the

    n-dimensional space satisfies conditions (2.9.12), (2.9.13) and besides that at any h, 0 e1(h)}) < 61(h) (m,r =1,...,n).

    (2.10.7)

    Here €;(h) and 6;(h) are the nondecreasing positive continuous functions such that co

    a e272) < 00, p—1

    ,

    82d (DEP I oo.

    (2.10.8)

    p=

    We shall prove that almost all realizations of the random function

    X(t) = x(t,w) have continuous partial derivatives over all the components of the vector t. For this purpose we construct for each m = 1,...,n the sequence of the finite-valued functions (p = 1, 2,...): Ue,@) = 2) elke

    at

    ee, 208

    k2? nder(2-?)}, (oe)

    oe)

    Di,= P=4YOR, Nm =f) Dt. p=1 Analogously as in Subsection that for any

    2.9.4 we ensure in the fact that PON a) =

    0 and

    WE N,, we have uP (w) < n2? sup te) ays?) e1(h)}) < &(h) In the special case of the scalar

    (m,r=1,...,n).

    (2.10.13)

    t we have m = r = 1 and condition

    (2.10.13) takes the form

    Meith t+an({(@1, 2, @3) :|| e1+a3—2x2 ||> €1(h)}) < 61(A). (2.10.14) Among the functions ¢,(h) and 6;(h) satisfying condition (2.10.8) it is expedient to choose the following functions:

    ea = an. Example

    2.10.3.

    bre) = vA ths aaa

    i = 0s

    (2.10.15)

    Let us consider the centered real random

    xX(t) == z(t, w) with the normal four-dimensional distribution.

    function

    In this case all

    random variables

    Yinr = X(t + emh+ eph) —X(t+emh) —X(t+e-h)+ X(t)

    (1)

    are normally distributed, their expectations are equal to zero, and the covariances are determined by formula

    on p(t, h) = K(t+ emh +eph,t + emh + eh)

    +K(t + emh,t+ emh) + K(t+ e-h,t+e,h) + K(t,t) +2K (t+ emh + erh,t) + 2K(t + emh,t + eh) —2K (t+ emh + erh,t+ emh) —2K(t+ emh + epy,t + e,h)

    —2K (t+ emh,t) — 2K(t + e,-h,t).

    (iI)

    2.10. Differentiable Random

    Functions

    We)

    After determining the variances of the variables Y,,, we get for the probabilities in

    (2.10.7) the expressions

    P(| Ym |> €1(h)) = 1- 20( er).

    (111)

    Hence similarly as in Example 2.9.3 we obtain that the condition

    Fn O

    Wiehe

    inir

    Alnus,1)

    (IV)

    at some C, Y > 0) is sufficient for the fact that almost all realizations of the random function X (t) will have continuous first derivatives with respect to all components

    of the vector ¢. In particular, the covariance function k(r) = _ y sin w|r|) satisfies condition

    (IV) at

    vy=

    Liz only at

    y =

    De~*!7l(cos WT a/w.

    At other

    values of 7 it does not satisfy (IV)). Thus almost all realizations of the normal random function X (t) with such covariance function at

    Y = @ /W have continuous

    derivatives. It goes without saying that the obtained result does not give grounds for

    the conclusion that at Y = a /W the random function X (t) cannot have almost all differentiable realizations.

    2.10.8.

    Trace Probabilities

    The space X7 on which the probability measure of the random func-

    tion X(t) is determined by its multi-dimensional distributions represents a space of all functions of the variable t, ¢ € 7’, with the values in the space X. Hence any measurable set of the functions includes not only

    the realizations of the random function X(t) but a set of other functions with very irregular behaviour as any set from the o-algebra A” restricts the function only in countable set of the points t. Therefore the space X7 is too large. It is always convenient to consider the random variables in the minimal phase space. In particular, as the phase space of a random function it is expedient to assume the set of all its possible realizations

    or the minimal class of the functions which contains all its realizations. Namely on this phase space it is necessary to determine the probability measure of a random function. So, for instance, for a separable random function it is natural to determine the probability measure in the space of the functions satisfying the condition of separability (2.8.1). Let (Q, S, P) be the probability space, 2’ C Q be a set on which the probability P should be transferred. In the general case the set 12’ does not belong to the o-algebra S in consequence of which the probability P is not determined on it. We denote by S’ the class of the subsets of the

    174

    Chapter 2 Random

    Variables, Processes and Functions

    set 2/ which represents the intersection of the sets of the o-algebra S with OQ’, S’' ={B:B=0'A, AES}. It is easy to see that S’ represents the o-algebra of the sets of the space 2’. The class S’ contains the empty set @ and space 2’. In parallel with any set B = 0’ A, A € S, S’ contains its complement Q\B in Q! as O/\B = 0/\0'A = 0'A and AE S. Finally, for any sequence of the sets {B,}, By € S’ we have By = 2’ A;z, Apr

    ES

    and (J By = 2’ U Ax € S' as Ag € S. The o-algebra S’ of the sets of the space 2’ is called the trace of the o-algebra S on 2’. It is easy to understand that the transfer of the probability P on the o-algebra S’ is possible in principle only in the case when P(C) = 1 for any set CES

    which contains 2’, C D 0’. In this case evidently P(D) = 0 for any set D € S which does not intersect with 2’. If this condition is fulfilled then the probability P’ on the o-algebra S’ is determined uniquely by formula

    PLBY= PYM ApS P(A),

    > Really, for any sets Ay, Ag € S such that

    A;\A2 and A2\Ai —

    P(A2\A1)

    B =

    Q’A,

    =

    0’ Ag the sets

    do not intersect with Q/ in consequence of which P(A;\A2)

    = Oand P(A)

    = P(A).

    It proves the uniqueness of the definition

    of P’. Further on it is evident that Pi a) ) = DAG

    (2.10.16)

    Meare Ss

    0 and PY)

    =

    Finally for any mutually exclusive sets By = Q’ Ax, ArpES

    P(C) SHG (k =

    15203. .)

    the sets Aj, may be also assumed as mutually exclusive as otherwise their intersections do not intersect with 22'and they may be rejected. Then we shall have

    Ply) Be) SP (SAS

    Fe Ag) Sheree

    (2.10.17)

    Equalities (2.10.17) prove the 0-additivity of P'. Thus formula (2.10.16) determines uniquely the probability P’ on the 7 -algebra S'. 4

    The probability P’ on the measurable space ((2’, S’) determined in such a way is called the trace of the probability P on this space or the trace probability. Usually the trace probability P’ is determined by means of the notion of the outer measure which is known from functional analysis. The outer probability Pp is determined for any set B C Q by formula PB) = pant. tal(Gan (2.10.18) The trace probability P’ represents the contraction of the outer proba-

    bility Po defined by (2.10.18) on the class of the sets S’. > Really, for any set

    B € S’ we have

    B= 0/AC

    A,AE€ESin consequence

    of which Po(B) < P'(A). On the other hand, for any set

    CE S,

    CD

    B=M'A

    melon Differentiable Random

    we have P(A) < P(C) at P(A\C) Consequently,

    P‘(A)

    the equality Po(B) a

    < Po(B). Po(Q'A)

    =

    Functions

    1745)

    P(C) as A\C do not intersect with (’.

    From the obtained opposite inequalities follows =

    P‘(A) for any B € S’. The condition of the

    existence of the trace probability may be written by means of the outer probability in the form Po(’) ==, Lene

    Now let fz be a probability measure of the random function X(t) determined in the space (X7,.A’); Y(t) be a separable random function

    equivalent to X(t). We denote by H the set of all realizations of the separable random function Y(t). Suppose that the set C € A? contains H. The set H restricts involved functions only in the countable set of the points for instance Lc C T. At each point ! € Le only zero

    set of the realizations of the random function X(t) does not coincide with the correspondent realizations of the random function Y(t). The countable union of these zero sets correspondent to all 1 € Lc is also

    a zero set.

    Thus almost all realizations of the random function X(t)

    coincide with the correspondent realizations of the random function Y(t) on Le. Consequently, any set C € A? which contains all the realizations

    of the random function Y(t) contains also almost all the realizations of the random function X(t). Therefore u(C) = 1 for any set of the functions C which contains the space H of all the realizations of the

    random function Y(t).

    It means that the transfer of the probability

    measure fiz on the space H is possible. We denote by B the o-algebra of the sets of the space H which represents the trace of the o-algebra AT on H, B={B:B=HA,A€ fr Bee(t)X°0 (tz) (twe Thus

    Kz f =

    K, f represents a vector func-

    k=)

    tion from the space X with the components CHO

    seFe fr EX2(t)X9 (te) (tx).

    Hence it is clear that a covariance operator of the vector eure function X (t) is determined by a matrix of the covariance and cross-covariance functions of its com-

    ponents

    K(t,t!) =

    AN OPMRBeey Perma Kon (ts?! Kon (Bytial the Kok(et!)

    7

    Redgltnttys Haph( eke! Momathhien(t 4a) where

    Hy, (iat ho)

    20 oO) 1pga

    RD

    1D)

    Similarly we convince in the fact that an operator of the second order moment of

    the vector random function X (t) is determined by the matrix I’, (ds t’) of the initial second order moments of its components. Formula (3.3.17) gives in this case the known relation between

    the covariance function, the initial second order moments

    3.3.

    and the expectations.

    Moments

    195

    This relation has the matrix form K,(t,t’)

    =

    T, (t, t')

    — mz (t)m,(t')*, where mz(t)* = [[email protected]. 0], mp(t) = EXp(t) (p 4 Peet n). Analogously as in Example 3.3.2 if the matrix functions K,, Ch toy and T, (t, ts) exist then the domain of the operator Ky and I‘, contains Fy.

    Let X = a(w) and Y = y(w) be the random variables in the spaces X and Y correspondingly, determined in the probability space (Q,S, P). If each of the spaces X and Y represent a B-space or a weakly topological linear space, and F and G are the correspondent. adjoint spaces of the linear functionals then formula

    a

    BXGY = | e@)ave}P (ae)

    (3.3.19)

    determines the operator ,y mapping G into X. The operator [zy deter-

    mined by formula (3.3.19) is called a cross-operator of the second order moment of the random variables X and Y. A cross-operator of the second order moment of the centered random variables \X°

    X —m

    and Y° =Y—

    my is called a cross-covariance

    operator of the random variables X and Y. It follows from (3.3.19) that the cross-covariance operator Kz, of the random variables X and Y is determined by formula

    Rey

    PX GVO = flee) —mz]g[y(w) — my, ]P(dw).

    (3.3.20)

    It is easily derived the relation between Kzy, Pry, mz and my from

    (3.3.19) and (3.3.20) analogous to (3.3.17) Keyg = Ceyg mmegmy..

    (320.24)

    Hence it follows the following statement. Theorem 3.3.2. For the existence of the operator Kzy tt is necessary and sufficient the existence of mz, my and Tzy. The domains of the operators Tyy and Key coincide. If X and Y are the random variables with the finite or countable

    set of the realizations {x,, yx} then Pryg = S > te IVEPE , k

    Keyg = Popes — Is

    Mx) 9(Ys — My )Pk 5

    (3.3.22)

    Chapter 3 Moments,

    196

    Characteristic Functions

    and Functionals

    where pz is the probability of the occurrence of the kt pair of the realizations (x.y.) (k = 1,2,...) of the variables X, Y. In particular, it

    follows from (3.3.18) and (3.3.22) that for the nonrandom vectors X = c, Y = d we have I, = cfe, Pege= cgd, Kz = Kay = 0. It follows also

    from (3.3.20) that Kz, = 0 if even one of the vectors X and Y be not random. The random variables X and Y are called uncorrelated if Kryf =0

    for all f € F (ie. Kry = 0). If Kayf # 0 even for one f € F then the variables X and Y are called correlated. > Suppose that the random variables X and Y are independent. the probability maesure /lz of the composite random variable L= to the product

    of the probability

    measures

    In this case

    (x F Y) is equal

    [lz, [ly of the variables

    NXGwY

    wee leet

    p(x) be an arbitrary scalar [/,-integrable function in the space X, and w be an arbitrary {ly-integrable function in the space Y with the values in some B-space or

    in weakly complete topological linear space.

    In this case according to the definition

    of an integral in functional analysis the product p(x) p(y) is integrable over the product jz of the measures [/, and [ly, and on the basis of the Fubini theorem we have

    Be XW) = |ole Wma de x dy)

    = [ote) [vedas aa) nolae) = f (2)nolae)- f¥adua(a) = Ey(X)EV(Y). 4

    (3.3.23)

    Thus we proved the following theorem. Theorem 3.3.3. If the random variable U represents a function of the random variable X, and the random variable V represents a function of the random variable Y and one of the variables U, V is scalar then in the case of independent variables X and Y an expectation of the product of the variables U and V is equal to the product of their expectations.

    Putting, in particular, p(X) = X° = X — mz, ¥(Y) = gY° = gY —gmy we get Kayg = EX°gY® = EX° Egy =0. Thus the independent random variables are always uncorrelated. The inverse as it is known from elementary probability theory is not always true. Bex

    aameprliemsis4e

    lf Xe =

    X(t), t € J} is a scalar random function

    whose all the realizations belong to the linear space X, and Y = Lis) s€Sisa scalar random function whose all the realizations belong to the linear space Y then

    3'3- Moments

    197

    putting in X and Y the weak topologies of Example 3.3.2 by means of the linear F’, and G, similarly as in Example 3.3.2 we come to the conclusion

    functionals

    Dey and Key are determined by the mixed initial moment

    that the operators

    the second order Poy (t, s) =

    Key (5) = EX°(t)Y%(s) of the functions Hgxeacmup

    tise

    #3:3.5501t 16 —

    random function, and Y

    =

    X and Y.

    oh); t € Ty is the n-dimensional vector

    Vis), s € Sis the m-dimensional random function

    then putting in the functional phase spaces we come

    of

    EX(t)Y(s) and by the cross- covariance function

    X and Y the topologies of Example 3.3.3

    to the conclusion that the operators [ cy and Key are determined by the

    matrices Ie (es s) and Key (t, s) of the mixed initial moments of the second order and cross-covariance functions of the components of the vector functions X (t) and

    Y(s). 3.3.2. Properties of Second Order Moments

    In elementary probability theory the following properties of the second order moments for the random variables X and Y are known:

    e [* =T, (hermitivity); e the matrices

    [, and K, are nonnegatively determined

    (i.e. for

    any complex vector u) u?T,u@ > 0 and u? K,u > 0; e Tyc =TZ, and Kyz = Kjy. In elementary theory of random functions for the random functions

    X(t) and Y(t), t € T; the analogous properties take place: e

    Ty (ty 9 to) —

    baits ; to)*;

    e for any t,,...,¢y N

    Lg ulTp(tp,ty)tg

    >Oand

    € J) and any complex vectors uj,...,uNn N

    J) up Ko(ty,tg)ttg > 0; p,q=

    Poi=, e

    Dya(ta,ti)

    =

    De9(¢i,

    2)" and

    Eyxitt2.01)

    =

    Key(tinte)

    Let us study the main properties of the operators of the second order moment and the covariance operators. (i) From property (ii) of the expectations follows that operators Ts, Kz, Vcy and Key are the adjoint linear operators. (ii) From property (iii) of the expectations for any continuous linear functionalf follows

    fl f= E(fX)(FX) = EB |fX|?>0.

    (3.3.24)

    Thus the operator of the second order moment and the covariance operator are positive [, > 0, Kz > 0. (iii) From property (iii) of the expectations follows also that for any fas fo (e Pe

    mk

    filsfo= B(AX)(fpX) = felch.

    (3.3.25)

    Chapter 3 Moments, Characteristic Functions

    198

    and Functionals

    Thus the operator of the second order moment and covariance operator possess the symmetry property (3.3.25). Analogously we have for any functionals f € F., 9 € Ge

    (3.3.26)

    flryg = E(FX)GY) = glayf-

    Y = aX, U (iv) According to property (i) of the expectations if SSA

    |a|” lL

    ae

    |a|? Fee dpe Cho coukh gay OCIS ae.

    (v) From property (1i) of the expectations follows that if

    Vow) doy Moan? =e baton vol.

    (3.3.27)

    v=1

    then n

    Ty=

    0

    wali,

    Ky=

    ar ane

    )

    LpsRe

    =

    Pes

    DL Gyan

    -'Nos —.

    V;p=1

    wapKryz,

    (3.3.28)

    Mae

    Dot Groene v,p=1

    Here T,, and Ky, are a cross-operator of the second order moment and a cross-covariance operator of the random variables X, and X, (v,4=1,...,n) correspondingly.

    (vi) From property (iii) of the permutability of an operator of an expectation with a linear operator follows the relations between the operators of the second order moment and the covariance operators of the random variables connected by the linear transformations. Let T be a linear operator mapping a phase space of the random variable X into the linear space Y, and S be a linear operator mapping a phase space of the random variable Z into the linear space U. Suppose that there exist the adjoint operators 7J* and S*. We introduce the random variables Y = TX and U = SZ. From property (iii) it follows that afX 7s a random variable with a finite set of the realizations 21,...,@nN € Dp then Y = TX is a random variable with the realizations y) = Tx1,..., yn = Tan, and the operators Ty, Tyz, Tay, Ky, Kyc and Kzy are expresed by the formulae:

    hay piss

    Ky =TK,T",

    .=

    6 Diese spe Te

    beret stpag tat Kye =TkKe2 pou hago

    \

    (3.3.29)

    The domain of the operators Ty, Try, Ky and Kzy serves the domain Dry~ of the operator T*, and the operators [yc and Ky, similarly as T,

    3.3. Moments

    199

    and K,, are determined on the whole space F. Really, denoting by G the space of the linear functionals in Y adjoint with Y we obtain for any

    g€ Dr CG

    Tyg = EY gY = ETX(gT)X

    =TEX(T*g)X =TTeT*g.

    (3.3.30)

    Analogously other formulae (3.3.29) are proved. In exactly the same

    way if X and Z are random

    variables

    a finite set of the pairs of the realizations (x1,21),...,(@N,ZN),

    with

    71;

    ...,2nN € Dr, 2,...,2%n € Dg then a cross-operator of the random variables Y = TX and U = SZ are ezpressed by the formulae

    Pee

    Ss

    eK aT ee

    Sw

    (3.3.31)

    and its domain serves Ds-. If X and Y are B-spaces or weakly complete topological linear spaces and the operator T is continuous then follows that if there exist T, and Ky then the operators Ty, Vycz, Vey, Ky, Kyx and Kgy exist also and as the domain of the operators Ty, and Ky, serves the subspace Dp,, and as the domain of the operators Ty, Vry, Ky and Kzy serves the subspace {g : 9 € Dr-,T*g

    € Dp,}.

    Analogously, if X, Y, Z and U are B-spaces or weakly complete topological linear spaces, and the operators T and S are continuous then from the existence

    of the operators Tz,

    and Ky, follows the ex-

    istence of the operators Ty, Kyy whose domain serves the subspace {k > k € Ds+,S*k € Dr,, } of the space K adjoint with U. Let X and Y be the separable B-spaces, T' be a closed linear operator and A, C Dr then from the existence of the operators T, and Tyr, Ky and Ky, with the common domain D and nonempty intersection of the domain Ar» of the operator T* with D follows the existence of the operators Ty, Vay, Ky and Kyy with the domain {g : g € Dr-,T*g € D} and formulae (3.3.29). Really, in this case for any f € D there exist the expectations [,f = EXfX, ean = EY fX = ETXfX, and consequently, [gf € Dr and lyzf = TT,f. Similarly, for any g € Dre, T*g € D there exist the expectations

    Vee

    beer

    2 gi

    yor BYoYS

    A(T 9 ral 9

    ee

    BEAT? GX;

    (3.3.32)

    and here [,7*g € Dp and Tyg =TTcT"g. Analogously, if X, Y, Z and U are separable B-spaces, T’ and S are closed linear operators, Ag C Dr, Az C Ds then from the existence

    200

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    of the operators Tz, Tyz, Krz, Kyz and nonempty

    intersection of the

    domain Ag+ of the operator S* with D = Drp,, = Dx,, follows the existence of the operators Tyu, Kyu with the domain {k : k € Dg+,S*k € D} and formulae (3.3.31) also. Suppose that X and Y are weakly complete topological linear spaces and the operatorT is closed in the topologies of these spaces. We denote by V ={X,TX} a random variable whose phase space serves the graph of the operatorT then from the existence of the operators Tyr, Kur and nonempty intersection of the domain Ar of the adjoint operator T* with the domain D of the operator Ty, follows the existence of the operators T,, Dyc, Ke, Kyx with the domain D and the operatorsTy, Vy, Ky, Kay with the domain {g : g € Dr~,T*g € D}, and formulae (3.3.29) also. Really, in this case for any f € D there exist the expectations of the form (3.3.32) and [zf € Dp and lycf = TT,f. And in exactly the same way for any g € Drp+, T*g € D there exist the expectations

    Ppp f OX OVS EX (Pg) anet sl

    ge Dr and-tyg=

    MHP Teg)

    byg = EVO SEI

    Oty),

    TT pay.

    Analogously, if X, Y, Z and U are weakly complete linear spaces, the operator T' ts closed in weak topological spaces X, Y and the operator S is closed in weak topologies of the spaces Z, U, and A, C Dr, A; C Dg then from the existence of the operators T,,, Kyz and nonempty intersection As» with D = Dp,, follows the existence of the operators Pez, Cyz, Kez, Kyz with the domain D and the operators Tru, Tyu, Keu,

    Kyu with the domain {k : k € Ds+, S*k € D}, and relations (3.3.31) alSo. (vii) From property (iv) of the expectations follows that the operators and K of the equivalent random variables coincide, lp = Ty, Ky, = Ky, lez

    =Tyu, Kez = Kyu if the variable X is equivalent to Y

    and Z is equivalent to U. (vill) From property

    (v) of the

    expectations

    follows

    that

    if

    E || X ||’< 00 then the operators T, and Ky exist and bounded (and consequently, continuous). Really, as ||XfX ||_> Yp whose elements are the random variables Yp in B-space converges a.s. and in probability to the random variable X and all its finite segments are majored on the norm by the random variable U, EU? < oo then there exist the operators lpg, Kpq (p,q¢ = 1,2,...), Tx and Ke, Peper ey = Ye. determined everywhere on F, and

    fi

    al

    eels

    p,q=1

    Here the series converge at any f € Fy.

    p,q=1

    Kaif

    (3.3.34)

    204

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    In exactly the same way if the series )\Y, whose terms are the random variables Y, in weakly complete topological linear space weakly converges a.s. or in probability to the random variable X, if there exist all the operators Ing, Kpq, Vx and Kz and for any continuous linear functional f € F. there exists such random variable Uf, EU} < co that > fYp|

    < Uy at alln then formulae (3.3.34) in which the series weakly

    p=1

    converge tol, f and K,f are valid. Example

    3.3.8.

    If X is a finite-dimensional random vector then from

    the properties (ii) and (iii) of the operators quadratic form correspondent

    [; and Ky follows the positiveness of

    to the matrices

    I‘, and K,, and the symmetry

    of

    these matrices Ygp = Ypq> Kap = ke. From formulae (3.3.29) we may derive the correspondent

    formulae of the transformation

    of the matrices

    of the second order

    moments and the covariance matrices at linear transformation of the random vector X.

    In particular,

    the random vector

    we

    obtain the following formulae

    Y =

    AX

    for the covariance matrix

    and a cross-covariance matrix of the vectors

    Xx: Ky = Ak, A, Kye = AK.

    Similarly relations (3.3.31) give the following

    formula for the cross- covariance matrix of the random vectors

    = BZ: Ky, = AKy;B*. Example

    3.3.9.

    of

    Y and

    In the.case where X

    =

    Y =

    AX

    and U

    X(t), t € Tj is the n-

    dimensional vector random function from properties (ii) and (iii) of the operators I‘, and K, follow the positive definiteness and the symmetry

    Ppt

    land KE

    i

    sia

    iad Pes Bee

    P,q= 1

    =.

    S19

    nce

    58 fe Ret MIG

    10

    P,I=

    for any I, any vectors f;,..., fj andt,, 1K (ee:i

    of the matrix functions

    I

    ... ,t; € T, Te(t;t') =

    Deltsee

    (esc e. Analogously in the case of two vector random functions

    .e— X(t) and Y = Y(s) formula (3.3.26) gives the following relation between the matrices Key (a s) and ast):

    Epes; t) = Veale hie

    Let the random function X (t) = of the integrability are fulfilled. function

    z(t, w) be measurable and the conditions

    We introduce the ™m-dimensional vector random

    vee J ols) x@otae)

    (1)

    Ty

    In this case relations (3.3.29) give the following formulae for the covariance function of the random function (6) and the cross- covariance function of the random functions

    Y(s) and X (t):

    Kale cies /)9(s,t)Ke(t,t')g(s",t/)"o(dt)o(dt’), Rey

    (IH)

    3.3.“Moments

    Kyo(8,t) =

    205

    a(s,t')K,(t’, t)o(dt’).

    (III)

    Ty

    Notice that Tats, s’) and ives: t) are determined by the similar formulae. In the special case when

    2 = m = | from these formulae follow the correspondent formulae

    for the covariance function of the scalar random functions. Analogously from (3.3.31) follows the formula for the cross-covariance function of the random functions Y(s)

    and U(I) = f A( I, »)Z(v)r(de):

    Kyals,!) =| f 9(s,0)Kes(t,»)h(l,»)*o(at)r(do.. at

    3.8.8.

    (IV)

    Vi

    White Noise

    As it is known

    from elementary theory of random

    processes

    the

    random process X(t) with zero expectation and the covariance function which contains as a multiplier the 6-function,.mz(t) = 0, Kz(ti,t2) = v(t,)6(t; — te) is called a white notse in the broad sense. Taking into the consideration that 6(t; — t2) = 0 at t; # tz the multiplier v(t,) may be replaced by the multiplier v(t2) or by the symmetric multiplier

    \/v(t1)v(t2). The multiplier v(t) at the 6-function is called an intensity of the white noise X(t). The intensity of a scalar white noise is essentially positive. The intensity of a vector white noise represents nonnegatively determined symmetrical matrix. It is evident that the variance of a white noise is infinite, and its values in two quite close points are noncorrelated. For solving practical problems it is expedient to substitute a random process by a white noise. We may do it only in that case when the least interval between the values of an argument at which the values of an random process are practically uncorrelated is sufficiently small. This interval which is called a correlation interval . If the variable

    |K.(t1,t2)| /Ke(ti,t1) for the scalar random process X(t) may be assumed practically equal to zero at |t; —t2| > 7% and the variable 7; is

    sufficiently small then the random process X(t) may be assumed as a nonstationary white noise with the intensity equal to

    oe iArelearn:

    (3.3.35)

    Characteristic Functions

    Chapter 3 Moments,

    206

    and Functionals

    The correlation interval 7; for the scalar random process X(t) is determined by formula

    { ene Ke(t,t +7) ee 1 eld wee max | Rody)

    i

    3 3.36 ( )

    The vector random process may be assumed as a white noise if all its components may be assumed as white noises. 3.3.4. Moments

    of Higher Orders

    In elementary probability theory for more detailed characteristic of real scalar and vector random variables the following formulae for initial and central moments of higher orders are used:

    ee OCGA Ap

    ie ett), GPM Ca = Op,,.

    (3.3.37)

    ry == EXD.r Xo,Tn

    OID

    Se

    y eg he Sp Crar

    (3.3.38)

    oa Ee?

    trp boomy

    The vector variable r = (r1,..., 7m) in formula (3.3.38) is called a multi-indez. ‘The following formulae of the connection have place for the scalar random variables between initial and the centred moments:

    Nip a

    aya; )

    =

    : dXCe [pm pete Ce»

    =

    r! Gaze!

    ina WyGT

    a)

    br = (m1) TPCRapmyr? p=0

    (3.3.39) oO.

    (3.3.40)

    Analogously the higher moments for the real scalar random func-

    tions X(t), t € T; are determined in elementary theory of random functions Qtr

    = Prof er —oo

    — oo

    S00

    adir =

    tefelery ccs

    DAG

    zeithy

    eon

    oy

    ty)

    teen

    dey,

    (3.3.41)

    3.3¢ Moments

    207

    prltpgt obite) SEX? (t)... XL) 2 ov [ler=mmetta))

    Burts

    oe,

    fer ~ melt

    elie as y Un ORG © ALpl,

    (3.3.42)

    The mixed initial and central moments of the order ry ( [rl ‘+--+ Tp) for the real scalar random functions X;(t), ..., Xn(t) are calculated by formulae

    dag

    alatl

    sot

    (),

    = BY Xi({)... X10)...

    ryt)

    seve

    9 t))

    XA OP).. xis

    kd order moment

    of the n- dimensional vector random

    function is determined by the polylinear form of the moments and cross moments of the r*} order of all its components.

    3.4. Conditional Moments” 3.4.1. Definitions In elementary probability theory the different conditional characteristics of the random variables and vectors are used. So a conditio-

    nal expectation

    of a given function y(X) of a random scalar or vector

    variable X at a given value of a random scalar or vector variable Y is determined by formula

    B[y(X)ly] = / (2) fo(aly)de,

    (3.4.1)

    where f2(z|y) is a conditional density of the variable X at a given value y of the random variable Y. As a special case from (3.4.1) follows a formula for a conditional expectation of a random variable at a given value of the variable X: co

    E[X|y] = | ehelyae.

    (3.4.2)

    —oco

    The denotion f2(z|y) is used for the density of the random variable X dependent on the parameter y also in the case when y is not the value

    of some random variable Y. Here formulae (3.4.1) and (3.4.2) determine the expectations y(X) and X as the functions of the parameter y. The expectation of the random variable X as the function of the parameter y on which the distribution X depends is called a regression ofX on y. In the special case when the parameter y is a possible value of some random

    210

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    variable Y the regression X on y represents a conditional expectation X ater nay Knowing a conditional expectation we may determine all conditional moments of the random variables.

    A conditional expectation of the random variable Z = y(X) considered as the function of the random variable Y, EF[p(X)|Y] is called a conditional expectation of the random variable Z = p(X) relatively to Y. As the conditional expectations and the conditional moments of different orders of the random variables relatively to random variable Y are the random variables themselves then in its turn we may determine the expectations and the moments of different orders. Finally, we present the formula for total expectation which is often used in elementary probability theory

    Bo XY

    = Ee

    A

    al

    le

    (3.4.3)

    Hence at y(X,Y) = X we find

    EX =E[E[X|Y]].

    (3.4.4)

    Now we suppose that on the probability space (Q,S, P) the random variable X = x(w) whose phase space serves a measurable linear space

    (X, A) is given. Let us introduce similarly as in Subsection 3.2.1 a weak topology in X by means of sufficiently complete set of the linear functionals F, and admit that X is complete in this topology (i.e. is weakly complete topological linear space). Suppose that the conditional prob-

    ability Pr(Alw),

    A € S relatively to some o-algebra F C S is regular.

    A conditional expectation of a random variables X = x(w) relatively to a-algebra F is called a weak integral

    Bp(X\w) = /oe ao Kes

    (3.4.5)

    In the special case when X is a separable B-space, and F, is a space

    of continuous linear functionals on X integral (3.4.5) may be considered as a strong integral (if «(w’) is Pr(Alw)-integrable at a given value w EQ). The measure Pr(A|w) at all A € S is deteremined almost at all w relatively to the measure P and is a measurable function w relatively F. On the basis of the theorems about the measurable functions

    fEr(X|w) for any functional f € F. represents the function w measurable relatively to F determined almost at all w relatively to the measure P. Consequently, the inverse images of the sets {x : |fix — fi2xo|

    3.4. Coriditional Moments

    2h

    0 determined by the function

    Ey(X|w) and together with them the inverse images of all the sets of the c-algebra A’ generated by this class of the sets belong to the o-

    algebra F. It means that the function E¢(X|w) is (F, A’)-measurable, i.e. Is a random variable in (X, A’) determined on the probability space (Q,F, Pr) where Pr is the contraction of the probability P on F. If the o-algebra F is induced by some random variable Y = y(w) then F = Sy, i.e. represents the inverse image of the correspondent o-algebra B of the sets of the phase space of the variable Y. Then formula (3.4.5) determines a conditional expectation of the random variable X relatively to the random variable Y. In this case as it was shown in

    Section 2.3 P¢(A|w) is a measurable function of the random variable Y = y(w) relatively to B.

    Consequently,

    a conditional expectation of

    the random variable X relatively to Y represents the (B, A’)-measurable function of the random variable Y. Therefore we shall denote a conditional expectation of the variable X relatively to Y by asymbol E(X|Y) or shortly,

    mx |y. By the change of variables

    x = x(w) with account of

    formula (2.3.7) for a conditional probability measure formula (3.4.5) in this case is reduced to the form

    mxly = E(X|Y) = | eus(ael¥).

    (3.4.6)

    This formula determines a conditional expectation of the random variable X relatively to Y in the case of the existence of a regular condition-

    ally distribution ,(Aly). Substituting in (3.4.6) the random variable Y by its realization y we obtain conditional expectation mx, of the variable X at a given value y of the variable Y.

    = E(X|y)

    3.4.2. Properties of Conditional Moments The conditional expectations possess all the properties of the expectations which are established in Section 3.1 with the only difference that each of the presented statements is valid only almost for all w (or y) relatively to the measure P (correspondingly jy) and may not be valid for some zero set of the values w (correspondingly y). Alongside with these general properties the conditional expectations possess a number of specific properties. At first we prove the following statement.

    22

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    Theorem 3.4.1.If a conditional expectation EX |w) represents a weakly Pr-integrable function and the random variable X has an expectation then for any set BE F

    | Po(X\w)Pr (da) ze[Pads). B

    (3.4.7)

    B

    > For the proof we rewrite this formula in the form

    ‘iPr(dw) /Au peas

    eye /OVP (any

    (3.4.8)

    B

    It is easy to see that formula (3.4.8) is valid in the special case when r(w) represents an indicator of any set A E S, z(w) =e 1,(w). Really, in a given case

    |Pr(ae) |1ao')Pe(de'le) = |Pe(Alw)Pr (de). B

    B

    But the latter integral is equal to P(AB), and consequently,

    iPz(dw) i1,(w’)Pr(dw'|w) = P(AB) fri 14(w)P(dw). B

    From the validity of formula (3.4.8) for the indicators of all measurable sets follows its validity for any simple function r(w).

    sis about a monotonous

    By the theorem from functional analy-

    convergence follows the validity of formula (3.4.8) for any

    nonnegative function r(w) (independently of the fact whether the integrals in it are finite or infinite). As any complex function z(w) may be presented in the form z(w) = r1(w) = L(w) + 143(W) _ ix4(w) where 21, £2, 3

    and 4

    are nonnegative

    real-valued functions then formula (3.4.8) is also valid for any complex function r(w)

    and from the existence of one integral in it follows the existence of another. Now let r(w) be the function with the values in any weakly complete topological linear space

    X for which there exist weak integrals in formulae (3.4.5) and (3.4.8). For any continuous linear functional f € F, formula (3.4.8) is valid also for the complex function fx (w):

    |Pr(aw) ffe(u'\Pe(dulle) = ffo(e)P(dw). B

    B

    ae

    3.4. Conditional

    From the weak

    Moments

    213

    Py(Alw)-integrability of the function z(w’), the Pr-integrability

    of the function Er(X|w)

    and the weak P-integrability of the function z(w) follow

    three equalities:

    /Pr(dw) |Sotpaty =f /Pr(das) f ef!) Px(du' be), [t2)Ptaw) =F fewyP(de), B

    B

    f |Pr(dw) |2(u/)Px(dullw) =F fe(w)P(de). « B

    B

    In consequence of the fact that the measure Pr represents the contraction of the measure P on the o-algebra F, and coincides with P on all the sets in F, from the Pz-integrability of the conditional expec-

    tation E¢(X|w) follows its P-integrability and the coincidence of both integrals. Therefore formula (3.4.7) may be rewritten in the form

    | Be(x\u)P(de) = [ern), B

    Bef.

    (3.4.9)

    B

    In particular, putting B = { and taking into the consideration that the integrals at B = Q represent the expectations of the correspondent random variables we get

    E[By(X|w)] = BX.

    (3.4.10)

    This formula is called a formula of a total expectation.

    If o-algebra F

    is induced by the random variable Y then E-(X|w) = E(X|Y) and formula (3.4.10) takes the form

    E[E(X|Y)]= EX.

    (3.4.11)

    Now we consider a special case when the function z(w) is measurable relatively to the o-algebra F, i.e.

    (F,A) is measurable.

    In this case

    integral (3.4.5) coincides with the integral over the function «(w’) with respect to contracted on the o-algebra F measure Pr(Alw). As for any set B € F we have P-(B|w) = 1p(w) almost at all w then

    Br(X\w).= /ee

    Na

    Chapter 3 Moments,

    214

    almost at all w.

    Characteristic Functions

    and Functionals

    Thus if the function z(w) is (F,A)-measurable then

    almost at all w there exists an expectation of the random variable X

    = 2(w) relatively to F which coincides with this random variable: Erx(X|w) = 2(w)

    almost at all

    w.

    (3.4.12)

    We shall prove also the following theorem.

    Theorem 3.4.2. If a scalar random variable Z = z(w) 1s measurable relatively to F then for any random

    variable

    X = x(w) for which

    there exists E¢(ZX|w), almost at allw

    Ex(ZX|w) = z(w)Er(X|w) = ZEr(X|w).

    (3.4.13)

    > Firstly we consider a case when z(w) represents an indicator of the set

    B € F,

    74(6) ms 1p(w). Then

    Ex(ZX |w) = fto(o')e(w!)Pr(do' le) = J ee) Pe(de'le). B As Pr(Blw)

    = 1, (w) almost at all w then the latter integral coincides with

    /AEN ame Saha almost at all w € B and is equal to 0 almost at all w € B, ice.

    i#() Pe(ae!

    ())SAge)Be Oro)

    almost at all W ,what proves formula (3.4.13). Consequently, formula (3.4.13) is also valid for all simple functions Z (w) relatively to F. Now let z(w) be an arbitrary nonnegative measurable function relatively to F.

    After presenting it as the limit of the increasing sequence of the simple functions

    {z” (w)} we get the random variable ZX

    as the weak limit of the consequence of

    the random variables {Z” X } which a.s. converges and for any functional

    f € F,

    the sequence {Z eX } is modulus bounded of the random variable Z |f.X | which

    has a finite conditional expectation.

    By property (xiii) of the expectations (Sub-

    section 3.1.3) follows that E¢(ZX |w) = lim(weak)

    F(Z" X |w). And as formula

    (3.4.13) is valid for the simple functions Z” = 2” (w) then almost at all W we get lim (weak) Ex(Z" X|w) = lim ( weak ) Z” E-¢(X|w) = ZE F(X |w).

    3.4. Coriditional Moments

    This proves formula

    DANS

    (3.4.13) for the nonnegative random

    variables

    Z.

    Accoun-

    ting that any complex random variables may be expressed in terms of four nonnega-

    tive real random variables we ensure in the validity of formula (3.4.9) in the general case.


    0

    a

    0

    5

    1ah

    it

    o< [Ey uf

    r€(a,b), _

    : at

    sin f(z + a) — sin f(x — b)

    at

    lyrate

    orviay ric iby

    uenje (afb)

    a

    eee

    < [tw < 7 TU 0

    then by virtue of (3.5.28) the integrand in (3.5.27) enclosed in parentheses at all F is absolute-values bounded by the [4,-integrable function p(x) =

    2. Therefore

    the limit transition in (3.5.27) may be fulfilled under the sign of the integral. Then accounting that the measure [ly of the perimeter of the rectangle R is equal to zero

    we obtain

    im. p= J tr(e)ne(de) = ee tuys

    (3.5.29)

    what proves formula (3.5.24).
    Any rectangle may be represented as the limit of a monotonous sequence of the rectangles of the considered type and the measure [lz is continuous.

    Therefore a

    characteristic functional of a random vector X determines uniformly its probability measure

    /l, on the set of all the rectangles of the n-dimensional space X.

    In the

    consequence of the d-additivity of the measure /l, it may be uniquely extend of over

    the 0-algebra which is generated by the set of the rectangles, i.e. on the 0-algebra of Borel sets of the phase space of the random vector X.

    When

    the characteristic

    function 9x(f) of the random vector X is integrable (over Lebesgue) because of the boundedness of the second multiplier under the integral (3.5.24) the whole integrand

    in (3.5.24) is integrable. Therefore formula (3.5.24) may be rewritten in the form

    *

    PM

    omifnar— o-ifnbe

    po(R)= fo f [pe

    telfis

    Salt -th

    228

    Chapter 3 Moments, Characteristic Functions

    = aay faetna

    and Functionals

    feta.

    (3.5.30)

    R

    In consequence of the boundedness of the function 9u(f)e~*F ” and its integrability over f it is an integrable function of two vectors f and Z. Therefore we may change

    the order of the integration in formula (3.5.30). Then we obtain

    fia) = La’ je /ef? g,(f)df .

    (27)"

    R

    Hence it is clear that in a given case there exists the probability density f, (x) of the random vector

    X which is determined by formula

    1

    aah fz vy) (2)= ence

    ete

    Gx(f df

    :

    +0. (3.5.31)

    Then formula (3.5.4) takes the form

    92(f) = [eo tele)ae

    (3.5.32)

    And vice versa if there exists the probability density ts (x) of the random vector

    X over the Lebesgue measure then its characteristic functional gx(f) is determined by formula (3.5.32). On the basis of the known Fourier theorem Tee) is expressed in terms of 9n(f) by formula (3.5.31) and the function 9x(f) is integrable.
    Let now X be a random variable in an arbitrary real topological linear space,

    and Gade) f € F be its characteristic functional.

    According to property (vi) of a

    characteristic functional (Section 3.5.2) the expression

    n

    Gla PEIN, a=

    oF

    y arf

    (3.5.33)

    Kao

    at any fixed fj ,---)

    dn € F represents a characteristic function of a random vector

    with the components Z; = ff, X,...Z, it determines

    = fn X. In accordance with Theorem 3.5.1

    completely and uniquely the probability measure

    Lt of this random

    vector on the 0-algebra of Borel sets of the correspondent n-dimensional space, and in particular, on the rectangles {z : 21 € Ay, .+. SRALE An}: But the random

    3.5. Characteristic Functions

    and Functionals

    229

    vector Z represents a function of the random variable X and as the prototype of the rectangle iz : 27

    Ay,

    X serves the set {x wiz

    298A

    E An} in a phase space of the random variable

    Ga;

    wee

    inte

    An}: Consequently, on the basis of

    formula (2.2.4) we have

    Hy({2:24 CIA), «.. , Zn

    Ant) = poe:

    fiz EA

    Hence it is clear that the characteristic functional Gt)

    ,. ., In@ © An}).

    (3.5.34)

    together with the measure

    [lz of the vector Z determines also uniquely the probability measure [a of the random

    variable

    X on the sets of the type {x PUTS

    5 Pe

    ne

    Piead Dae

    oeON

    ey

    Al ,c.ec

    Ink & An} at given 1,

    te are arbitrary then a characteristic functional

    determines a probability measure of the random variable X on such sets which are

    correspondent to all n, f} ,...

    , fn € F and to all Borel sets Ay , ... , An. After

    determining the measure [lz on the class of such sets we may extend it uniquely on the o-algebra generated by this class of the sets
    As it is known lues in the B-space

    =

    from the theory of the measurable functions with the va-

    there exists a sequence

    eo) ae Ye lpn (z) (n ia.

    of the elementary

    Fy)uniformly convergent

    functions

    to y(z).

    y” (x)

    On the basis

    k=

    of property (xii) of an integral with respect to the bounded function over a finite measure for all p the inequalities are valid

    |[ere

    fvdny SLO

    [vane fv Let us take an arbitrary € >

    y(x) || ,

    a sup |[y"(2) — y(z) ||.

    (0) and choose 7 in such a way

    that the inequality

    sup, lly” (x) = y(«)|| < €/3 will be fulfilled. After that we choose Po in such a way that at all p > po the inequalities will be fulfilled

    tp(D) — u(DE)| < See :

    3Namaxlvel |Eee

    3.6. Sequences of Probability Measures and Characteristic Functionals

    231

    Then we shall have at all p > po

    |[van — fva] < |[vee +| [vray

    3.6.2.

    fat any

    fvray| +] frau

    fava Let F,oe) be a distribution function correspondent to the measure [lp (Sec-

    tion 2.2).

    Show that we may single out the subsequence {Fo, (x)} convergent to

    some distribution function {F,(x)}.

    (x) at all the points of its continuity from the sequence

    For this purpose we take any countable set D,

    for instance, a set of all the points with rational coordinates.

    at all p then we may

    single out from the sequence

    =

    4 Pina, dense in X, As 0 < Fy (xy) < 1

    1H, (2)} such a subsequence

    {Foi (x)} that the numerical sequence {Ft (a4 )} will be convergent.

    Analogously

    we may single out from the sequence {Foi (x)} such a subsequence { Fn2(ax)} that the numerical sequence

    { F,2(2)} will be convergent.

    Continuing this process at

    any T we single out from the sequence eee el (x)} such a subsequence defor (tt) that the numerical sequence {Poe

    Set will be convergent.

    The diagonal sequence

    WEES EY, will converge at all the points of the set D, as the sequence {F(@;)} at any Tf is a subsequence of the sequence {ieee} y at p > Tr. Suppose that the function Fp (x) represents the function F'mp (x) from the initial sequence {Fm(«)}

    (p = 1, 2,...). We denote the limit function Fp,(«) = a and define the function F(a) =

    Lm

    C,ps+o

    Fim, (2) at z € D,

    Fp,(#r), where {x,y }is the sequence of the

    232

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    points of the set D, convergent below to the point Z, Z; < Z. We show that the se-

    quence {Fim (2) } converges to F(a) at all the points of the continuity F(a). From the inequality Fin, (2’) 2/(aé). After this accounting that u(B) == ait Lm, (B) < C we choose pz in such a way that at all p > pp would be Lm, (B) . NG:A bo) v=1

    where z(t 4) is some function which belongs to the space X.

    Consequently, in this

    case ® = X a positively determined functional 9(f) on f serves as a characteristic functional of some random function X (t). Exam

    p1e

    the scalar functions,

    3.6.3. Consider a case when X represents the B-space Lp of

    1 < p < 00. We assume as F’ the space Dg, q= p/(p a 1) of

    continuous linear functionals on Lp dual with Lp . The space Lp is reflexive in such

    a sense that the set of all continuous on Iq linear functionals coincides with Lp. But

    3.6. Sequences of Probability Measures and Characteristic Functionals

    241

    the space of all (not only continuous) linear functionals Ly does not coincide with Dy.

    Therefore not every continuous positively determined functional 9(f) on Lg

    determines the probability measure on Lp :

    3.6.5. Probability Distributions in Banach Spaces. Mourier Theorem Now let us consider the case when X represents the B-space. It is natural to assume as the set F of the linear functionals on X topologically dual with X. In the general case the space © of all linear functionals on F, does not coincide with X. According to Theorem 3.6.6 not every continuous positively determined functional g(f) on F, determines the probability measure on X. In order the functional would determine the probability measure in X it must satisfy some additional conditions. Notice that at any fixed f,,..., f, € F the function

    IA eeleieaian)

    my

    (so)

    (3.6.12)

    =a

    of the real variables a;,...,@,

    determines not only the n-dimensional

    distribution py,.¢, of the random function Z(f) (Subsection 3.6.4) but also the distribution

    e 12 spoudo2) 6 By) PfaessifAF = Hhs,.ufn L2(f) : (2(fi), 3 -,2(fn)) S B})

    on the o-algebra A;,,5,

    of the sets{x : (fiz,...,fnx)

    (3.6.13)

    € B} of the

    space X correspondent to all the Borel sets B of the n-dimensional space R”. The aggregate of all such sets correspondent to all finite sets of the vectors f;,...,fn € F as it is easy to see forms the algebra Ao of the sets of the space X. Formula (3.6.13) at all n, fi,..., fn € F determines on this algebra Ap the nonnegative additive function of the set w,. If the function pz is g-additive on Ao then according to the theorem about the extension of the measure we may extend it till the measure on the

    o-algebra A generated by the class of all sets {x : (f,z,..., fnv) € B}. But from the theorem of Subsection

    3.6.4 we may conclude that the

    function p, in the general case will be not o-additive on Ag. It will be o-additive only in that case when the distribution py in the space (®, F)

    is completely concentrated on the subspace (X,.A), A= XF. The arguments given above show that for finding the conditions at which a continuous positively determined functional g(f), f € F

    Chapter 3 Moments, Characteristic Functions and Functionals

    242

    determines the probability measure on X it is sufficient to know at what conditons the function of the set jz is the o-additive on the algebra Ap.

    For short notation we determine the sets Ah, = {x : |fpx| < N} and put

    By =() Aw= 12: 8UDpey pela) p=

    UNG Nee eee)

    (a0 da)

    z

    All these sets belong to the algebra Ao and the function pz is determined on it. Let us prove the following statement.

    Theorem 3.6.7 (The Mourier theorem). In the case when X 1s the B-space and dual with X the B-space F, is separable (in this case X is also separable) for the o-additivity of the function pr, on Ao tt ts necessary and sufficient the existence of such a countable set {fr} which as dense on the unit sphere of the space F, that at anyé > 0, at all natural n and all sufficiently large natural N the condition will be fulfilled

    Ha lie. Sup |fpe| > IV})
    For proving the necessity of condition

    (3.6.15) we suppose

    that [ly is the

    measure on Ag. Then we may extend it uniquely till the measure on the & -algebra

    A generated by a weak

    topology of the space X.

    But in the separable B-space

    o-algebra generated by a weak topology coincides with the 0-algebra of Borel sets generated by a strong (i.e. natural metric) topology.

    Consequently,

    the sets By

    = HED : \|x|| =< N} belong to the o-algebra A and the function fz is determined for them.

    As {Bn} is a decreasing sequence of the sets and lim Bn

    then by virtue of the continuity of the function {ly we

    = 1) Bn

    have lim Lz (By)

    Hence at any € > (0 there exists such N, that Lz (Bn)




    ClAgy

    Now we suppose that (An

    =

    ‘then of the

    ©CO we

    may

    Cin iG As... tbat

    @. Then it will be

    and if lim ps (Cp) = 0) then at any € > 0 there exists such nN, that

    n,.

    After choosing nt. thatieny
    nt

    we get Me(An) < Ur(Cn) + én < € at alln > nz, nt. Hence it will follow that lim Ha( AR) =

    0. Thus it is sufficient to prove the continuity of [ae in zero for the

    sequences of weakly closed sets {Gye Let {Cn} be an arbitrary decreasing sequence of weakly closed sets, C, € Apo, with empty intersection Grbts n —

    OOo.

    GG

    () Ch

    =

    @.

    Prove that Ha (Cn) —

    0 at

    For this purpose we suppose that [Uy C,) > € at all n for some € > 0

    and show that in this case the intersection of the sets C;, cannot be empty.

    At first

    we show that at any sufficiently large NV the intersection of any set C', with the set

    Bn = tn : ||| < N} cannot be empty. As Cy

    &

    9i,--+;9m

    C,Bn

    =



    @

    Agvthen. Fe

    Cy

    =

    ee

    : (giz, seat ; Gene) (S Beir} at some

    and the closed set F,, of the m-dimensional

    space

    R™.

    ™, If

    then

    reamintalte||eaN"

    (3.6.16)

    TEC,

    In fact, in consequence of the fact that the linear functionals g1,...,9m

    € Fc are

    continuous in the strong topology of the space X generated by a norm the relations 21 = 91@,..-.,;2%m

    = Jm¥@ establish a continuous mapping of X on R™.

    Therefore

    the set C’, as the inverse set of the closed set F, at this mapping is closed in a strong topology of the space X.

    Consequently, in C, there exists the point Zo in which

    the continuous nonnegative function p(x) = ||| achieves its accurate lower bound r and if r < N then the intersection C,Byn

    cannot be empty.

    Now we consider the function (21, ess rn) which is inf ||2|| on the set {x : Jit = 21,-+--5JIme

    =

    Oi

    Pm}

    aitpel Betz

    pW

    inf

    |||.

    (3.6.17)

    91 T=21,---)Jml=2Zm

    As this function is continuous p(azi, Feds 5 OZ; ) = ay(z1, hd; at) at any a > 0 and y(0, Jee

    0) = 0 then the set {21, Pom

    aes

    p(21, Pra taA Erni) < c} represents

    244

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    a simply connected region of the space R’™ which contains the origin. Therefore it follows from (3.6.16) that if C, By

    =

    @

    then the set F, does not intersect with

    the set {z1, eterna

    p(21, «te ayn) < r} (i.e. is disposed outside of this set).

    The set Do = {2}, A:

    Bay: p(21, ae

    5 ia) 0) is convex.

    let {oka teed host E€ De, {iz1's aL eA function p(21, Hei

    € D,.

    Really,

    From definition (3.6.17) of the

    5 2m) follows that at any € > 0 there exist such z’, 2” € X that

    gee elope! kmvaii(k aid pi alpm)and: |e fez] po Ze) Rep ||e44| < p(2i,...,2,) +e. Taking € < c— max{p(z4,..-52m); P(21,---> 2m)} we have \|a’|I, \|ar’’|| N|hil|.

    Really, by virtue of the completeness of the B-

    spaceX there exists the vector 2g in X for which ||zo|| SA

    chize = ||Ai|I. Putting

    at an arbitrary 6 > 0 x5 = (b; +6)x9/||Aill, we shall have \|5]| = (b; +6)/||hal|

    and hyxs = 6} + 6. Consequently, the point 25 belongs to the set iz shye. > by} and if b;) < N||hil| then by virtue of the arbitrariness 6 > 0) \|s]| < N. But if b; > N||hil| then N|\hil|

    N.

    Introducing the linear functionals hi =

    245

    hi/||Ail|,

    belonging to the unit sphere of the space F’, we obtain from (3.6.18) and from the fact that b; > N||hil| the following relations:

    Cn C U {x : hjx > bi/|lAil|}

    C U {a : |Ajz| > N},

    l=1

    (3.6.19)

    f=

    tte(Cn) < fhe (Ut : |afe| > vy) | As according to the condition the set ee the space

    (3.6.20)

    is dense on the unit sphere of

    F’, then for each hj there exists convergent

    to it the sequence

    . {fo}, fir = hf ge hi at r — co (l = 1,...,58). continuity of the functional 9(f) at T — ©O we have

    {fir }

    Here by virtue of the

    s

    Gini

    (Wises

    40s)

    =

    9

    Safir

    Nesil

    s

    9

    So arht

    = 9h',..,h(@1,---,@s).

    (3.6.21)

    [fees

    But

    een ar AO

    istic function

    inverse

    of some

    See Qs) at any

    S-dimensional

    Levi theorem it follows from

    ty measures converges

    correspondent

    vector.

    F. represents

    by virtue of the

    (3.6.21) that the sequence

    of the probabili-

    correspondent

    ae hasan re)

    to the characteristic function

    Qs). Consequently, at any N > 0)

    ee (Ute : |Aja| > v}) = im Ite (U420s figatl r=" At any f we have

    a character-

    Therefore

    to the characteristic functions gerne

    to the probability measure

    Lf eee A (a1, oe

    Sia eifen fe €

    random

    f

    vy) ais

    (3.,.6:22)

    =|

    s — U {x : fir | > N} (© BN: where 7, is the largest of the num(eal

    bers of the functionals fj, =

    dais tee Pe

    Foote n=

    max{pir, Ad

    Deas

    And as Teak ES) < € at all n thenwe get from (3.6.22)

    Te (Ut Seo

    vy) Zon Bet)

    ee

    (3.6.23)

    246

    Chapter

    3 Moments,

    Characteristic Functions

    and Functionals

    Accounting (3.6.23) from (3.6.20) and (3.6.22) we get fle(Cn) < € what contradicts the supposition that fis('Gy)

    > €.

    Thus the intersection By Cy cannot

    be empty if there exists such € > 0 that jee Cia) > € at all n. a centered system of closed subsets of the set By

    X.

    And the set By

    But {BnC,, } is

    in a weak topology of the space

    is bounded and consequently is compact in a weak topology.

    So the intersection () BynCy,

    cannot be empty,

    and it means

    that the intersection

    n

    () C, also cannot be empty.

    Thus we proved that if there exists such € > 0 that

    een)

    > € at all nm then () Cr e

    if () Cr

    =

    @

    @ . Hence it follows that lim filC,)

    =)

    what proves the sufficiency of condition (3.6.15) for the continuity

    of [ly in zero on Apo, and consequently, for the o-additivity of Zz on Apo. Thus the

    sufficiency (3.6.15) is proved.
    Really, on the basis of formula (3.5.9) a characteristic function of the random vector Z is determined by the formula n

    GeO a> gy Ons.) Se Oe

    ied p=1

    3.7. Normal Distributions in Linear Spaces

    J

    n

    247

    n

    1

    = exp ‘ ) Op fp Mz — 5

    y

    orb

    keh.

    P,q=1

    p=!

    But fp Mg, represents an expectation of the random variable

    7, = fo X , consequent-

    ly, 9z(a1, nied Qn) is a characteristic function of a normally

    distributed random

    vector, what proves Theorem 3.7.1. Further, a characteristic function of the variable Z in Theorem 3.7.2 is determined by the formula

    ; 1 4 D, > = exp § iam, . 1, bp od gz(a) = exp 4 tam, — 9% — re On the other hand, 9z(@) =

    Be'%4

    =

    oer

    Bet*i% | Hence it is clear that a charac-

    teristic functional of the random variable X in a given point f € F' is equal

    GAf) Se! ifx

    G41)

    exp

    ifs

    — Pg 5fKof

    As it is valid at any f € F then 9a(f) at any f € F is expressed by formula

    (3.7.1).
    Really, for any f},fo € F, fi -

    = fo) =

    BX (Fp fa)X*

    (3.7.6)

    fe, Kefi

    =

    Kafe

    we have Kabfs

    0 ond (fie fz) Kah; — fo) = Eh

    = fo) X°|? = 0. Thus (fi = fo) X° = 0 with probability 1 and Kya(hi = fo) —

    BY

    Chi — f2)X°

    uniqueness

    =

    (0). Hence we find Kychi

    of the mapping

    X —>

    Y determined

    =

    Kyo fo what proves

    by Eqs.

    (3.7.5).

    of this mapping follows from the adjoint linearity of the operators

    the

    The linearity A,

    and Kye

    (Section 3.2). Now

    we show that the operator L, determined by Eqs.

    For this purpose we notice that at any

    for all 29

    =

    Kyfo

    and

    © =

    |fxy = fox1| 0 \gLx = gLxo|

    then Ffn& —

    Characteristic Functions and Functionals

    fxo at all f € F. But %, = Kifependas

    fxo at any

    ~igla'=

    Sieh

    @ E A K, and lim f,@, evidently, represents the result

    of the action of some linear functional fp on %.

    Consequently, the sequence of the

    functionals thn} converges to fo in such a sense that fnt —

    fox atallze

    ie.

    If fo € F then Ky fo€ Ak, and as Ky fo = £0 by virtue of the continuity of the operator Ky, then £0 € jG The continuity A, in the weak topologies of the spaces F’ and X follows from the fact that by virtue of the equality f; K, f =

    or ; \fix’ = fiz| O and f,f;

    — fev K cE}, “=

    If fo€F for some fundamental sequences

    sept

    Tela

    € F

    of the point ‘i

    C Ax,

    then extending the

    space F’ on account of the inclusion into it all such fp and extending the operator ie on the continuity we get Yo € Auge for all fundamental sequences Zo

    =

    limz,.

    Thus

    Age

    may

    Keyg e Ax, at any g € G. Otherwise for the functional does not belong to Ae

    Latved oa Ni,

    be always considered as a closed set.

    But

    then

    g € G for which Keyg

    we may find such linear functional f € F' that fx = 0 at

    any 2 € Nig. and S Keyg a 0. But it is impossible as

    fKeygl? =|EGX oY DP REY

    XP ElgY|? =f. fe giao

    and from fx = Oat x € Ax, will follow fK f = 0 and Kryg = 0). The obtained contradiction proves that 1

    = Kayg E Ak, at any g € G. But then there exists

    such linear functional f; € F that x1 = Kf,

    and

    |\fei — fori| = |fKefi — fokefi| = |fikef — fikefo| = |fie — fizol. Consequently,

    |gLa = gLzxo| 0,g E€G, x

    € Ak, and all Z in

    iS re \fix = fi Zo| Let us consider the linear mapping S'

    uu =

    of the space measurable

    Z =

    (x, y) tame

    X

    X Y

    as the operator

    =p

    =

    into itself.

    torus

    y! =

    G—

    Le

    This mapping evidently is continuous and

    1 according

    to the proved is continuous

    and for any

    h=(f,9)€H hv = fe+ gy = fze+ gy— gle =(f —L"9)e + gy = h'u, where h’ =

    (f = L*g, 9) € H. As a result of this the set {v : (hiv, sete hnv)

    S B} € C has the inverse set {u 3 ( iu, : HO fa kan) Sc B} € C. Therefore the random variable

    Zins® (ar, (0) in the space

    Z = X XY

    Sar

    oY

    is normally distributed. We find a cross covariance operator

    Kytz of the variables Y’ and X. Putting

    ef SEY

    POSEY

    But according to (3.3.29) Kw, it follows from Eq.

    eYeaaiee LK"

    W = LX° we find

    Gea) =

    LKg

    (3.7.6) that Kyte =

    (LX

    GX") = Kye or hue

    therefore Kyte =

    Kye — LK,.

    Hence

    0. Thus the random variables Y’ and X

    are uncorrelated, and consequently, by virtue of the normality of the distribution of the variable Z’ they are also independent.

    Therefore the probability measure of the

    random variable Z’ = (X, vs |is determined by the formula (Section 2.4)

    Hz1(C) = f ue(de)

    f10(e, nay (de), Cie ©,

    (3.7.7)

    where /lg and /ly’ are the probability measures of the random variables X and Y’. Now we notice that the mapping S is convertible as to a given value

    V =

    corresponds the unique value U =

    Therefore on

    Gea

    eo y}, y =

    y + La.

    oe y }

    the basis of formula (2.2.4) from (3.7.7) it follows the expression for the values of the

    254

    Chapter 3 Moments,

    Characteristic Functions and Functionals

    probability measure of the random variable Z on the measurable rectangles Ax

    B,

    AEA, BEB:

    uz(Ax B) = uz({(2,y): = p({(z,y'/): 2 € A,y’ €

    2 € A,y € B}) B—m, — L(x — mz)})

    = fy (B= my ~ E(e— me) ad). A

    Hence we conclude that there exists a conditional distribution of the random variables

    Y relatively to X connected with probability measure of the random variable Y’ by

    formula

    Hy( Ble) = py

    (B-—m, —L(e#—mz)),

    BEB.

    (3.7.8)

    This formula shows that Hy (Bx) at any fixed value & of the random variable X represents a measure

    determined on the d-algebra B.

    Consequently,

    a conditional

    distribution of the random variable Y relatively to X is regular (Section 2.3). Fi-

    nally, it follows from (3.7.8) that a conditional distribution of the random variable Y relatively to _X is normal as the random variable Y’ is normally distributed.

    By the

    symmetry we conclude that in the considered case there exists a regular conditional distribution of the random variable Y relatively to X and this distribution is normal.


    |||. Hence on the basis of

    p=1 property (ix) of the expectations follows nm

    n

    BYZ} = EY (X", pp)? — El|X°|?. foal

    pl

    As E(X®, Pp)” = Oe Pp; Lp) = Ap and the series SS Ap converges then [e.@)

    E\|X|? = lim > E(X°, gp)? = p=1 Hence it follows that E\|X°||? =

    Theorem

    3.7.13.

    [o-e)

    YE: p=l

    |ma||? + E||X°||? 0) we may write

    pa({ssu(s) >€}) alr],

    E(X, Yp)?,

    p=1

    (3:7.21)

    p=1

    where X is a random variable for which the contraction of the function [tz on the

    o-algebra Aine.

    serves

    as the probability

    distribution.

    For determining

    the

    expectations in this formula we shall find a characteristic function of a joint probability distribution of the random variables Z,; =

    (xX, £1); 1,0 Gn eS (X, Yn).

    According to formula (3.5.9) this characteristic function is obtained from (3.7.17) at n

    i=

    De Ap Pp (Section 2.4): p=1 n

    UE

    ey

    ae

    wis oo)

    =

    exp

    a )

    Ap (Mz; Pp)

    p=1 1

    =

    n

    nm

    7 CrgOa) os pig Ka

    p== exp

    P,q=l

    Se ims, en) — 5aAp p=l

    Hence it is clear that the random variables Z,,..., Zp, are independent,

    ly distributed with expectations

    normal-

    (mz, 1), ns, Gite. Yn), and the variances 1,

    , An. Therefore E(x, ye) = EZ = (mz, Pp)” =F Ap- Substituting this expression into (3.7.21) and taking into account that by virtue of the Bessel inequality n

    (Ms; ¥p)” < ||mell?, we get p=1

    to

    (42: (ee)? > N2b |< 5 [Imell? + ap n

    Z

    1

    2

    p=1

    2

    nm

    2

    3.7. Normal Distributions in Linear Spaces

    263

    In consequence of the convergence of the series a Ap we conclude that at any

    € > 0 there exists such N, that at all n and N SN n

    fag

    li

    tees

    (3.7.22)

    p=1

    Now we notice that the set of all finite linear combinations of the eigenvectors Yp of the operator Ky with the rational coefficients is countable and dense in X and therefore the set of the vectors

    pe

    Fn =

    AnpPp,

    ashe

    Anp = LI Hree

    p=1 correspondent

    'np

    TTT LTE

    Cie

    &

    te aN.

    to all possible finite sets of the rational numbers

    Tn1,..-.,TnN,,

    is

    countable and dense on the unit sphere of the space X. Here in consequence of the Nn fact that >. a = 1 for any z € X the following inequality holds: p=l1

    2

    Nn

    (2, Jn) =

    Sones,

    Pp)

    Na

    Sot (x :Pp)*

    p=1

    pl

    Hence it follows that

    and at any narutal N we have

    {e:sup (2, f)|>N}C fe: D> (yp)? > NPP p=1 Using inequality (3.7.22) we obtain

    tell Saceupille i fy)| > N}) 0, where {%p } is

    p=1 n

    an arbitrary sequence of the vectors of H-space

    X then (Ke ip f) —

    a3 Ap Ge ~p)? p=1

    and formula (3.7.17) gives

    golf) =exp }i(rme,f)— >Yolen)? p-

    (1)

    In this case for any vector f orthogonal to the subspace G' generated by the vectors

    Y1,---,;Pn variable

    we have (Fe £1) =

    Z =

    Ch Yn) =

    0). Consequently,

    the random

    (Xx; f) has the expectation (ite, f) and zero variance.

    It means

    that the distribution of the random variable X is completely concentrated on the

    hyperplane Mm, + G.

    Example

    3.7.4.

    The functional g(f) =

    mal distribution on any O-algebra Axl

    e7 1/211 FI? determines a nor-

    AF. of the sets of the type {x : ae, fi),

    oon (ar, ta)) (S B} at the fixed linearly independent

    f),..., fn. Really, putting

    n i =

    SS Op tp we obtain p=

    gf) =exp §=5 D>apa Sys Sa) p

    (1)

    p,q=l

    The right-hand side of this formula represents a characteristic function of some normally distributed n-dimensional random vector Z = {Z;, ster Lae

    The expecta-

    tion of such vector is equal to 0, and the elements of its covariance matrix K, are

    the scalar products (Ses fg As f1,..., fn are linearly independent then the determinant of the matrix K, is strictly positive in consequence of which the distribution

    of the vector Z is not singular. Let us consider the random variable

    XK Spare Sul

    be a lic:

    (II)

    Qe

    where koeq are the elements of the matrix

    oa 1

    This is the normally distributed

    random variable in H-space X whose all realizations belong to the n-dimensional

    subspace G formed by the vectors f,,..., fn. A characteristic functional of the variable

    X is determined by the evident formula

    gx(f) = Bef.

    n

    = E exp

    n

    aya q—

    koa, fr) |24 lap

    3.7. Normal Distributions in Linear Spaces

    =o.

    (Soi f fp):

    265

    COD konlf, ia) p=1

    =o

    {=} t 3 (sede)rye ey 2

    q=1

    (111)

    Sal

    But SS (psf) piI=1

    3

    Rep ks, akae fr )(F; fs)

    Tsai

    => (f, fre)(f, fs)HF Leh

    (IV)

    ns= Accounting that Us P ‘PS are the elements of the matrix K,, and ke are the elements .

    °

    ow

    of the inverse matrix K’>*

    Oe

    n

    .

    in consequence of which

    =

    ee

    2 keq( fas fp) = 5sp we get q=

    (Fp, Ia) pee Kepksalds Ir )\Ipte)

    p,q=1

    =

    (seal

    py aa

    urs

    Cie ee

    (V)

    Tesi

    The latter expression represents a square of the norm of the projection of the vector ij

    on the subspace G.. Therefore the obtained expression for the characteristic functional of the random variable _X may be rewritten in the form gz (f) = e~1/2\|PfIl

    2

    , where

    P is an operator of the projection on the subspace G. In particular, for any vector

    fsaG

    ;

    ge(f). sien

    -1/2

    =a).

    (VI)

    The contraction g(f) on any finite-dimensional subspace of the space X simi-

    larly as the contraction of a characteristic functional of a normally distributed random variable determines a normal distribution on this finite-dimensional subspace.

    If we take the orthonormal vectors f,,..., fn then the random variables 27, ., Zn will be independent and their variances will be equal to 1. Therefore

    ae

    He ({:; Nec)

    n

    < v}) < be (A{x : |(z, fp)| < N}

    = [[ vole :I(e, fo) < N}) = [28(9)]". pal

    (VII)

    266

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    For Laplace function 0(N) at any N we have 0 < 20(N)

    < 1. So at n —

    00 we

    get from (VII)

    Te ({>: Neila fn ES vs} — 0. ep) co

    Thus as ys (gz: eae = \|z||? we conclude that [lz cannot be a probability distribupail

    tion on X.

    3.8. Approximate Representation of Distributions 3.8.1. Pearson Curves System In applications it is often required to represent a density or a distribution function of a random variable approximately by a suitable analytical expression. At practice any simplification of an analytical representation of a density leads inevitably to the parametrization. The system of the distribution curves of Pearson (Pearson 1894, 1895, 1902) is widely used for an approximate analytical representation of the densities of scalar random variables. Pearson noticed that majority of the densities y = f(a) satisfies the following ordinary differential equation: dy _

    z+a

    dx

    bo + by2 + box?

    where a, bo, b;, bz are some constants. In particular, normal, exponential, Student, 6-, y-, y- and y?- distributions satisfy this equation. Varying the parameters a, bo, b;, b2 one may obtain a great variety of distribution curves. The most general way of approximate representation of distributions is their representation in the form of linear combinations of some given functions, in particular, in the form of various truncated series.

    3.8.2. Orthogonal Expansion of Density Let w(x) be some density in the r-dimensional space R" for which all the moments exist. A system of pairs of polynomials p,(zx), q(x) (v = 0,1,2,...) is called a biorthonormal system with the weight w(x) if co

    ilw(2)p,(x)q,(a)de = 6, = {{ 2 ;? z: —oco

    (3.8.1)

    3.8. Approximate Representation of Distributions

    267

    A system of pairs of polynomials p,(x), q(x) (v = 0,1,2,...) is called a brorthogonal system with the weight w if condition (3.8.1) is fulfilled only at p # v. Any biorthogonal system of pairs of polynomials {p, (zx), q,(x)} may be reduced to a biorthonormal system by dividing the polynomials Pv(£), qv(“) by numbers a,, 3, respectively whose product is equal to

    the integral in (3.8.1) at the corresponding v and p = v. It is evident that at every v one of the numbers a,, (3, may be chosen arbitrarily.

    In the special case where q,(x) = p,(x) (v = 0,1,2,...) condition (3.8.1) takes the form

    / w(a)p €e)py(#)de = by,

    (3.8.2)

    In this case the system of polynomials {p,(x)} is orthonormal if it satisfies condition (3.8.2) at all v, and orthogonal if it satisfies the condition (3.8.2) only at uw # v. Any orthogonal system of polynomials {p,(x)} may be normalized by dividing p,(x) by the square root of the integral in (3.8.2) at the corresponding v and p= v. It is evident that the existence of all the moments for the density

    w(x) is necessary and sufficient for the existence of all the integrals in (3.8.1) and (3.8.2). It is expedient to use multi-index numeration

    of the polynomials

    pv(z), q(x) in such a way that the sum of the coordinates

    |vy| =

    +.--.+ vy, of a multi-index subscript v = [1...v,] be equal to the degree of the polynomials p,(z) and q,(x). Then the number of linearly independent polynomials of a given degree m = |v| will be equal to the number of independent monomials of degree m, 1.e. to Cy,_1

    = (r+m-—1)!/[m\(r — 1)!]. It is evident that a multi-index numeration may always be replaced by the usual one. > Let f(x) be the density of some 7-dimensional random vector X for which the moments

    of all orders exist.

    Let us try to represent

    the density

    f(z) by an

    expansion of the form [oe)

    Sy

    f(z) = w(x) Vi,

    soy

    Cy Dy (& ) .

    (328.3)

    a)

    To determine the coefficients C, of expansion (3.8.3) we multiply (3.8.3) termwise by qdrat) and integrate over the whole r-dimensional space R". As a result taking into

    268

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    account (3.8.1) we shall have

    iff(e)a(a)de=

    D> goon)

    / SS ATTY ee

    Leas)

    Consequently,

    y= f Hle)a(@)de =Bas(X) = a(a),

    (3.8.5)

    —co

    where qv(a) represents a linear combination of the moments of the random variable

    X which is obtained by replacing all the monomials ah eee ake in qv(x) by the corresponding moments

    Q@/;, ...@,-

    Thus all the coefficients C, of expansion (3.8.3)

    are expressed in terms of the moments of the random vector X. Here copo(2x) ="

    always since by (3.8.1), po(x) and go(x) are mutually inverse constants (the poly-

    nomials of zero degree). To construct expansion (3.8.3) it is convenient to choose the density w(x) in such a way that all its moments

    corresponding moments

    of the first and the second order coincide with the

    of the density f(z) of the random vector X.

    the polynomials qv(2) of the first and the second degree ( |v | =

    Then for all

    Vi pape oop

    Ye

    =a 2) we shall have by virtue of (3.8.1)

    cy = :f(2)qv(x)de = iw(x)q,(2)dx

    syed!

    / uebiesia4h G)de ws

    In consequence of this expansion (3.8.3) takes the form

    fe). =0(2) |ty) k=3

    eps (a)

    (3.8.6)

    [v=

    In some cases it is convenient to express the coefficients C, in (3.8.6) not in terms of the moments of the random vector X as in (3.8.5) but in terms of its characteristic function. Minding that any moment of the vector X is obtained by the differentiation

    of its characteristic function g(A) with respect to the corresponding coordinates of

    3.8. Approximate Repfesentation of Distributions

    269

    the vector 7 followed by equalizing A to zero (Subsection 3.5.3) we may represent

    formula (3.8.5) in the form

    cy = [qr(0/t0A)9(A)],~5 - 4

    Remark.

    (3.8.7)

    Formula (3.8.6) as well as formula (3.8.3) gives the expansion

    of the function f(x)/V w(x) in terms of the unit vectors { Jw(@}pr(«)} of the function space L2(R’), which form a biorthonormal system of the vectors together

    with the vectors { w(x) gy (x) i.On the basis of the general theorems of functional

    analysis expansion (3.8.3) of the function f(a)/4 / w(«

    [

    in case

    PO) 6 < oo

    w(z)

    —0oo

    m.s.

    converges

    to f(x)//w(2).

    complete in D2(R").

    The sequence of functions

    {

    w(x)py (x) is

    The sequences of Hermite functions (Appendix 2) may serve

    as examples of complete systems of functions in the corresponding spaces Lo( Py

    Formula (3.8.6) determines the orthogonal expansion of the density f(z). Partial sums of the series (3.8.6) may be used to approximate the distribution defined by f(x) with any degree of accuracy also in the case where the random variables X has no moments. It is sufficient in this case to replace the distribution f(x) by the corresponding truncated distribution ip (e) = f(@j\ipte)/ | seae,

    (3.8.8)

    D

    approximating f(x) with sufficient accuracy and then to approximate fp(x) by a partial sum of series (3.8.6). A finite segment of expansion (3.8.6) may be a function which assumes small negative values at some x and nevertheless may give a good approximation to f(z). Retaining in (3.8.6) only the polynomials not higher than N*» degree we obtain the approximate formula for the density f(z): N

    f(x) © ft(z) =

    w(2)}1+ >> SY) epr(z) | . k=3 |v|=k

    (3.8.9)

    270

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    The function f*(x) approximating the density f(x) by virtue of (3.8.5) is completely determined by the moments of a random variable of up to the NV order. In this case the moments of the function f*(z) up to the N* order coincide with the corresponding random variable X.

    moments

    of the

    > Really, multiplying the last equality in (3.8.9) by dp (2) and using (3.8.1) and (3.8.5) we get

    / Pf(e)qu(e)de = ¢, = dua) at

    |p| < N.

    Thus

    the expectations of all the polynomials

    (3.8.10) q(X)

    of not higher

    than Nth degree calculated by means of the approximating function fe (x) coincide

    with the expectations of the corresponding polynomials qu(X ) calculated by means of the true density fla):

    But the polynomials

    q(x)

    of any given degree k are

    linearly independent and their number coincides with the number of the moments

    of

    the k'® order. Therefore from the coincidence of the expectations of the polynomials

    qv(X) it follows also the coincidence of the moments of the function f* (x) and of the density f(z) up to the N*® order. Denoting the moments of the approximating function f* (x) by oe piece

    and Hy sk,

    WE May write the obtained result in

    the form * Oat peveilbex

    =

    ON

    eben

    3

    Leon,

    LORE

    Sy Lr

    seed os

    at

    ky

    +---+k,


    yo Eo) N

    k=3

    |v| =k

    Vy 4a

    (3.8.24)

    v;!

    Formula (3.8.5) determining the coefficients of this expansion takes the form

    C=

    / f(x)G,(x — m)dz = EG,(X —m)=G,(p), —oo

    (3.8.25)

    274

    Chapter

    3 Moments,

    Characteristic Functions

    and Functionals

    where G_() is the linear combination of central moments of X obtained as a result of replacement in G,(« — m) of every monomial of the form (X1 - m1)" ...(Xp — m,)"r by the corresponding moment fn, as, Bee

    According to the result of Subsection 3.8.1 all the moments (both initial and central) of the function f*(«) approximating the density f(a) up to the N* order coincide with the corresponding moments of the density f(x), and the moments of higher orders of the function f*(z) are expressed in terms of the moments of up to the N“ order from the relations

    Grkyet) = Oat.

    lz) as

    (3.8.26)

    where by the asterisk the moments of the function f*(«z) are indicated.

    The coefficients c, of H,(a—m)/(™!...v,!) in the Hermite polynomial expansion of the density f(z) are called the quasimoments

    random variable X. The number

    of the

    |v| = 1; +---+ , is called the order

    of the quasimoment c,. According to (3.8.25) the quasimoment of the k*> order represents a linear combination of central moments up to the kt order. 3.8.4. Relations between Quastmoments and Semiinvariants

    For expressing the quasimoments of a random variable in terms of its semiinvariants we may certainly substitute in formula (3.8.25) the expressions of the centered moments in terms of the semiinvariants. In this case following formulae are used:

    Igy

    an

    Sey eee

    eae

    eee

    Mt eee

    Poe Ing)

    |

    ea

    cond tty

    (3.8.27)

    Rte q.yaar),

    RS (3.8.28)

    es

    (nt othn)/2] (_yyp 1 Oslusgano nde

    =

    dons ae

    Pp

    p=2

    P

    tid ilsllpee Vigo Viuit...+Unp

    [(hit--+hp)/2] Hai

    be

    (3.8.29)

    Uae

    k=1

    OL, Cote p=r

    hal ashy! a oa © | B-

    3.8. Approximate Répresentation of Distributions

    x

    DS.

    2

    ies

    7%

    Rg

    PS

    IlSime a ee

    ey See

    Ned l

    gk 04,2) bad frcrith > 4).

    (3.8.30)

    The integral part of the correspondent number is marked here by square brackets. But in the general case the result is achieved by means of very complicated and cumbersome calculations. Therefore we shall try to express directly the quasimoments c, in terms of the semiinvariants fey > Let X be the r-dimensional random vector with an expectation M, a covari-

    ance matrix K and a density f(x)< We put u = tA in formula (2) of Appendix 2 for the generating function of Hermite polynomials Ge (x) corresponding to this ran-

    dom vector and replace the variable

    ZTby

    © — m.

    Then taking into account that

    (u? _ m?)\ = AT (2 == m), we shall have

    exp {iaT(e —m)+ st Ka} = ye,

    (x —m),

    (3.8.31) where =

    the summation

    extends

    to all nonnegative

    values

    of 1j,...,V,,

    and

    V

    [14 Se PS le After multiplying this equality by the density Tie) of the random

    vector X and integrating with respect to Z we shall get by (3.8.25) g

    aN

    tN

    EX]25((A) pa = A ies se nieae

    (3.8.32)

    LyA k=3

    |v|=

    -\T where g()

    is the characteristic

    function

    of X.

    Thus

    eta

    T m+"

    KX/2 g(A) as

    a function of 7X serves as a generating function for the quasimoments as well as g(A) serves as a generating function for the initial moments, € =e ™g(A) serves as

    a generating function for the central moments,

    and In g(A) serves as a generating

    function for the semiinvariants. Substituting into (3.8.32) the expression of g(A) in terms of the semiinvariants,

    g(A)= en 9A) = exp

    5m ye ve. 5 ae 5

    h=1 |s|=h

    BY

    HG31663)

    and remembering that the semiinvariants of the first and the second order coincide

    with the expectations and the elements of the covariance matrix of X respectively we may write

    Pp >i) h=3) |s|=

    vee,

    ie

    Sony

    cope

    (alles, a hh Lasers 1

    276

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    or

    3

    4h

    Deg p=

    (iz) as (OR

    De de,

    h=3 |s|=h

    Stl ose

    ee tis

    =e

    k=3 |v| =k

    (id)? Pye

    ...(trp)”" yel

    ae

    Hence it is directly clear that the quasimoments C, of the third, the fourth and the fifth order coincide with the corresponding semiinvariants Cy = Ky ( |v | 2S, 45 5).

    If follows from the fact that the expansion in brackets in the left-hand side begins with the terms of the third degree in consequence of which only the item of the first

    degree p = | in the left-hand side may give the terms of the third, the fourth and the fifth degrees.

    As regards the terms of the sixth and higher degrees in the left-hand

    side for which \v| > 6 they are present in the items corresponding to p < [ \v| Ao i Therefore, after raising the expression in the brackets to the corresponding powers and collecting the terms of the same degrees we obtain

    $

    Cll /3] 5

    V —* = pat SSPp: rea SO VA a 5 ol Yaw

    where qj Ze = [qi1---

    ir] ah yee





    CRW MCP B® Iv] =6,7,...), (3.8.34) Gil-:---Qpr:

    Wp ais = ioage aoe 4h as well as V == [71 «ot Vel

    are 7-dimensional multi-index subscripts and the inner sum is extended over all the

    values of

    1, --- , Qp, lqi| asa,

    [14 a: AR

    The first term in the right-hand side of formula (3.8.34) corresponding

    ldp| > 3, whose sum is equal to the vector V=

    to p = | is equal to Ky.
    V2),

    26,

    (P=0,1,...,0),

    |

    which m.s.

    converge at each t € 7}. Similarly all the derivatives of the covariance

    function 0? +4 K, ite t') /OtP ot'! (p,q = 0,1, ..., ) will be expressed in terms of the canonical expansions

    P+

    ike

    =

    /

    Se)

    DoDePaG),

    e=01,....9),

    ail

    convergent at all t, t’ € 7}. In the case of a vector argument the operation 0/Ot may

    be substituted by any differential operation in a space of the argument @, in

    particular, by the operation of the differentiation over any component

    of the vec-

    tor t. Example

    3.9.10.

    Find a canonical expansion

    of a scalar white noise

    (Subsection 3.3.3). Let X(t) be a white noise of the intensity v(t). According to

    296

    Chapter 3 Moments,

    Characteristic Functions

    the results of Example 3.9.6 as an orthogonal basis in

    and Functionals

    Hz we may take any complete

    sequence of the functions { Fr (t)} which are orthogonal on a given interval T}:

    (fatten

    = | tol?

    fn(tat

    = Dnébnm .-

    (I)

    For constructing such sequence by means of formulae (3.9.13) and (3.9.14) we may

    take any complete sequence of the continuous functions {Pn (t)}. Formula (3.9.16) for the coordinate functions in this case gives Zn (t) = BD V(t) fn (t). The canonical expansion of a white noise will be m.s.

    convergent,in such sense that any stochas-

    tic integral which contains this white noise will be expressed in terms of the series convergent in mean square:

    frotxma==

    [OOROd, fQeHe.

    Hence in particular it follows that the Wiener process

    W(t) at any finite interval

    [0, IB ] is expressed in terms of a canonical expansion:

    W(t)= YoVaan(t), an(t)=p ffalar,

    (III)

    where aie (7 )} is a complete sequence of the orthogonal functions at the interval

    [0, i ik The random variables V,, are determined as the stochastic integrals of the form

    Dy

    Vaz ikfn(t)dW(t) .

    (IV)

    0

    So the canonical expansions of the stationary white noise V(t) = W(t) and its co-

    variance function K’,(t, t’) = v6(t—t’) at the time interval [0, 7) ]are determined by the following m.s. convergent series:

    1 V(t)

    =

    T,

    [o-0) MS,

    Vier"

    Kel)

    =

    =

    33

    eiAn(t-t

    (V)

    )

    where A, = 2an/T); Do = D[Va] = HVE. The canonical expansions of the nonstationary white noise V(t) and its covari-

    ance function K, (t, t’) =a)4 (t)4 (t')8(t — t’) have the following form: t

    V(t)= a

    =

    5

    ea aie (Oh So Vile n=—0o

    ae

    /

    co

    :

    .

    SE ein(t—t!) n=—oo

    (VI)

    3.9. Canonical Expansions

    297

    where A, = 2an/T); Digs D[Va] — aA Exam

    p1e

    3.9.11. Under the conditions of Example 3.9.3 at

    D =

    1 and

    the time interval (0, T\] the canonical expansion takes place:

    Dace peerfret eee nC = sin si | (:= 2) + =| et where D,

    =

    D [Vex ]=

    transcendental equation:

    An

    tgw7|

    + An); W 1, W2,... are the positive roots of the

    =

    —2aw/(a?

    = w?); An are the eigenvalues of

    the integral equation

    T;

    /ent Example

    -

    ;

    7

    Wee evel

    (I!)

    3.9.12. It is not difficult to make sure that for the covariance

    function K,(t, t’) = v(t)v(t’)e* |t-#"| at the time interval [0,7] the canonical expansion has the following form:

    where An, Wy, Dn are determined in the previous example.

    Examp1e

    3.9.13. Consider the random function

    X with the expectation

    Mie = Vie (t) and the covariance function

    Kp(t,t') =

    Reine

    qi(t)qa(t’)

    ata

    at

    ae

    (1)

    t >t’,

    where qi = 41 (t) and qz = qo(t) are such that the relation q2/1 is a monotonous increasing function. Find its canonical expansion at the interval Ty —T,

    0) that

    Oy z ot

    GulT aie Fa\Galt).\ ana Le . qi(T2 — T1)q2(T2) qo(T2 — T1)qi(t) Liles a qo(T> — T1)qi(T2)

    oY

    The variable € is monotonously increasing function of t and to the interval TT, —T,

    < t < T> corresponds the interval 0 < € < 7. We determine the function Q(E) by the equality Q?(E) = qi(t)qe (t). Then the change of the variable t by & gives the

    298

    Chapter 3 Moments,

    Characteristic Functions

    opportunity to present ke (tt) in the form HptAt)

    and Functionals

    =

    Q(E\Q(E')e

    * |é-¢ |J

    Using the results of the previous example we obtain

    X(t) = m,(t) + 5 oYni EE GROEN) sy |wno(t) + a aeciti) n=l

    where A, = 2a/(a? + w2), I Dger

    1

    D[Va] = An(n + pvea

    go(t)

    /a1(T2 —T1)q1(T2)

    Oy n@) Vq(T —Tas)

    and Wry, are the positive roots of the equation tg nw =

    —2aw

    (Iv) /(a? = w?),

    3.9.4. Inverse Theorem about Canonical Expansions

    Theorem 3.9.2 (The inverse Pugachev theorem).

    Jf none of the

    vectors x, belongs to the closed linear span of the other vectors z,, 1.e. to the subspace X, closed in the weak topology of the space X formed by the sequence {Cu} vey then to canonical expansion (3.9.18) of the covariance operator K, corresponds canonical expansion (3.9.11) of the random

    variable X.

    > Really, if x,€X,

    at none of

    then at each V there exists the linear func-

    tional f, € F, whose zero subspace contains X, but does not contain Z,.

    This

    functional may be always chosen in such a way that f,£, = 1 (for this purpose it is sufficient to assume Te = Jy,

    where gy is an arbitrary linear functional with zero

    subspace which contains X, but not XL,). Thus the condition of Theorem is sufficient for the existence of the sequence of the linear functions ae: € F, satisfying

    the condition f, x —

    by p- To this sequence corresponds the orthogonal sequence

    of the vectors of the space H,.

    by canonical expansion i.2)

    Keo

    =

    In fact if the covariance operator Ky, is expressed

    (3.9.18) then from fy, €

    Fo, fvEp

    =

    ov follows that

    ee

    we Dy EMGven)

    = D, «xy. Hence it is clear that the elements is of the

    N=

    space F’, may be replaced by the correspondent classes of the equivalent elements, i.e. by the vectors of the space H, and then uke fi i= Dp SinPy sal DI Sup: On the

    basis of Theorem 3.9.1 the orthogonal sequence of the vectors fut C Hy

    generates

    canonical expansion (3.9.11) of the random variable X with the coordinate vectors

    gy.

    3.9. Canonical Expansions

    299

    ~~

    3.9.5.

    Way to Construct Canonical Expansions

    To find the systems of the linear functionals < Lent C F, satisfying the condi-

    tion of the biorthogonality ful p = vp together with the coordinate vectors 2, of

    canonical expansion (3.9.18) we may use the following method which is known from functional analysis. At first we take an arbitrary sequence {Yn } C F, for which at

    any 7 the following condition is fulfilled,

    $101

    pot.

    .-.

    Yn*1

    Sere | cs

    ae

    apa

    Pee

    P2%n

    ---

    Pn&n

    Pitn

    (3.9.19)

    |

    The mentioned functionals exist always as for any nonzero vector

    £ € X in FP, there

    exists such a functional ~ that px # 0, and consequently, it may be always chosen in such a way that px =

    1. After determining the sequence {Yn} C F, we assume

    that V1 St Psy

    Wy =

    Ayivit:

    : +ay

    y-1Py-1t

    Pr

    (v =

    2, 3, oa sss (3.9.20)

    and find @,, sequentially from the conditions Wy Cy = 0 at

    u< V. As a result we

    obtain p-1

    Ay. = —Pv%1,

    | 29 Area

    Ap = —Pvl yp —Dandaes

    (

    =

    Nae |

    8

    )

    A=1

    and

    Wy By Had

    a= (atl jl

    (3.9.21)

    (v Sal eat .) Really, from (3.9.20) with the account that Prly

    A we find v-1

    ye, = Ay1¥it1+ Prt,

    Ply = So avyataey + Yy Zp , AS

    pl

    Oy

    tee SS ovataty ai

    yr ee

    Specter Esa.(pbate colt

    weil ae We

    A=A

    Putting Dy fy, = 0 at fp < V and excluding the unknown Qj,

    ...

    , @y,y—1 from

    the obtained l/ equations we shall have

    W124

    Caan

    a

    ee

    Wi Ly-1 Wily

    0

    Bes,

    0

    Py Xt

    0

    Py @2

    vars

    Vy—12y-1-

    Pvfy-1

    a

    .Py-12y

    PyLy

    irs

    anes

    pe prky

    -

    300

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    We replace here zeros by the correspondent expressions yp\,, and subtract from the (v = 1) column all the previous columns multiplied by @)—1,1,.-.

    , @y—1,v-2

    correspondingly and then subtract from the (v = ae column all the previous ones multiplied by @y—21,---

    , @y—2,v-3 correspondingly and so on, from the second

    column the first one multiplied by 0/21. By virtue of (3.9.20) and (3.9.19) we get

    Coe REO) ek

    Putting here sequentially we find W,2,

    y =

    at all v.

    Oy

    H1

    O12

    PIe

    Gite

    Prt7

    +.»

    Pyb2|

    City

    Poty

    ©.

    Opty

    _ 4.

    2,3,... and accounting that @121.=

    After this we derive formulae

    912;

    =

    1

    (3.9.21) from the obtained

    equations. Finally, after determining the sequence {vn} we put that [oe

    ey Sag

    fi Se Be

    (3.9.22)

    aay,

    A=v

    and calculate Bu sequentially from the condition f, Cy = Oy i As a result we get p-1

    ae (ee apanston (3

    Bw =1, Bu =— )Perdren

    3.9.6. Canonical Expansion of Random Theorem

    3.9.3.

    If the random

    iy (3.9.23)

    Variable in H-space variable X in H-space possesses

    such a property that E || X° ||?< co (or what is the same E || X ||? < co) then tt may

    be presented by a canonical expansion in powers of

    the ergenvectors of tts covariance > Let X

    be random

    operator a.s and m.s.

    convergent.

    variable whose phase space is H-space

    X.

    The space

    of the continuous linear functionals dual with X may be identified with X as any

    continuous linear functional on X is representable in the form fx = (a f) where f € X. Assuming F’, = X we express a covariance operator of the random variable X by formula

    KepeS Ee (pae es Suppose that

    | x

    RSS co.

    (3.9.24)

    In this case as it was proved (Section 3.3) the

    covariance operator Ky is the positively self-adjoint completely continuous

    tor.

    opera-

    Therefore it has a finite or a countable set of the eigenvalues {An} and the

    correspondent orthonormal sequence of the eigenvectors {Pn }:

    Kon

    = An

    Pes

    (Ore

    or n-

    (3.9.25)

    3.9. Canonical Expansions

    301

    ,

    Here for any vector i € X take place the expansions

    f=fot (fi > yv)ev,

    Kefo=0,

    (3.9.26)

    v=1

    Kef = fot > o(f, ev)or -

    (3.9.27)

    ie

    Hence it is clear that the range A K,, is the separable H-space with the orthonormal basis of the operator K',. Introducing into Ak, another scalar product

    Cho)t teh et en het

    Ao

    fx)

    and supplying A K, by the limits of all fundamental sequences in the metric generated by this scalar product we obtain an. auxiliary-

    H-space H,.

    In essence the latter

    space does not differ from the space H, constructed in Subsection 3.9.2.

    Really,

    the characteristic property is only in the fact that each element in a new space Hy represents not a class of all the elements X equivalent relatively to Kz, but only one

    element of this which belongs to Ax,. The orthonormal basis {Yn} in A K,, formed by the eigenvectors of the operator K, appears to be an orthogonal basis in H, as (gv, Pu)i =

    (KeGv, Pp)

    = AG ey, Pu) = Ay by . Therefore according to the direct Pugachev theorem the sequence {Pn} determines a canonical expansion of the random variable X

    X=me+)

    [e.e)

    Xe,

    (3.9.28)

    VK

    where X, = (a (2) ) are the uncorrelated random variables with zero expectations and the variances which are equal to the correspondent

    eigenvalues

    A,.

    Formula

    (3.9.27) gives here the correspondent canonical expansion of the covariance operator of the random variable X. Expansion (3.9.28) according to the proved theorem weakly m.s.

    converges to

    X and expansion (3.9.27) weakly converges to Ky f. But according to the theorem about the spectral expansion of completely continuous self-adjoint operator which is

    known from functional analysis series (3.9.27) converges strongly in the metric of the

    space X. It turns out that in this case canonical expansion (3.9.28) of the random variable

    X also strongly a.s.

    and m.s.

    converges in the metric of the space X.

    In

    fact almost all the realizations of the centered random variable X° belong to the space Ax,

    and this means

    that they belong also to the space Hz.

    Consequently,

    almost all the realizations «° of the random variable X M may be presented by the expansions in powers of the basis {Yn }:

    2° =)

    (2°, or) yr.

    (3.9.29)

    302

    Chapter 3 Moments, Characteristic Functions and Functionals

    Hence it is clear that the sequence of the random variables

    =

    Xie == STUB

    (Sta)

    (3.9.30)

    vot

    a.s.

    converges to the random

    variable X |

    This proves the strong a.s.

    gence of canonical expansion (3.9.28). As all random Mig

    X”

    conver-

    are a.s. bounded [o-e)

    on the norm

    of the random

    variable day

    | DEE

    les 2 |X, |?


    viii

    Cer)er. 4F- ow(h ever i—3) (3.9.42)

    v=.

    From formulae (3.9.42) and the convergence of expansions (3.9.35), (3.9.36) in the metric of the space H, follows weak m.s.

    convergence of canonical expansion

    (3.9.40) of the random variable Y . Finally, from canonical expansion (3.9.40) of the random variable Y we may determine the correspondent canonical expansion of its covariance operator:

    Ky ¥y =)

    \ywuy("tr). 4

    3.9.43

    y=)

    Theorem 3.9.4. The random variables X and Y in a separable Hz space may be represented by canonical expansions (3.9.36) and (3.9.40) with the same coordinate vectors if the spectrum of the operator A=Kyz'Ky consists only of the countable set of the eigenvalues. > The condition of the proved theorem is fulfilled for instance, when there exists such number Yo

    >

    0 that the operator

    of the projection on the subspace

    continuous.

    B =

    A — YoP where P is the operator

    A 4 of the values of the operator A is completely

    Really, in this case the spectrum of the operator B consists only of the

    eigenvalues from a finite or a countable set {An} to which the orthonormal sequence of the eigenvectors {yn} corresponds:

    USOSC

    Se 9 A

    =m

    ty key

    (3.9.44)

    Here for any vector f € Hz we have

    f= DOU bee + (F)vt)],

    (3.9.45)

    i

    BYS Se) or Val

    (3.9.46)

    3.9. Canonical Expansions

    where {yi,} is an orthonormal

    307

    sequence of the eigenvectors correspondent to zero

    eigenvalue of the operator B (the space H, supposed to be separable).

    After dis-

    posing the eigenvectros WY, and wi, in one sequence {Yn } we notice that all of them serve as eigenvectors of the operator A. In this case the eigenvalues of the operator A equal to the number oy iat Ay respond to the eigenvectors Y,y belonging to the subspace A g of the values of the operator B; the eigenvalue of the operator A equal

    to Yo corresponds to the eigenvectors v), belonging to zero subspace VB erator B but not belonging to zero subspace

    of the op-

    N 4 of the operator A; zero eigenvalue

    of the operator A responds to the eigenvectors wi, belonging to N4.

    The equation

    Ay = Y¥ may be rewritten in the form:

    Kyp =7KepConsequently,

    the problem

    (3.9.47)

    of finding the eigenvalues and the eigenvectors of the

    operator A is reduced to the calculation of the values 7 at which Eq.

    (3.9.47) has

    the solutions different from zero and to the definition of the correspondent solutions.

    Assuming in (3.9.47)

    7 = Yv

    # 0, ~ = Yy we may rewrite formula (3.9.38) for the

    coordinate vectors U, in the form

    ig= apy, Myo,

    Remark. number

    Yo

    >

    HY = 12). ee

    (3.9.48)

    The conditions of Theorem 3.9.4 are fulfilled if there exists such

    0 that the operator

    the projection on the subspace Ay

    B =

    A —

    oP

    where

    P is the operator of

    of the values of the operator A is completely

    continuous,

    Example

    3.9.16.

    Let us consider the random vectors

    n-dimensional Euclidean space.

    UE > Kef = 0} G

    X and Y in the

    Let Ky and Ky be their covariance matrices and

    oi, ; Kyf = 0}.

    If Kz has the rank m

    < n then Hy

    represents the ™-dimensional factor-space of Example 3.9.1. Eq. (3.9.47) in this case

    has M eigenvalues Y1 , ..- , Ym different from zero and the correspondent orthogonal eigenvectors Uj,

    ...

    , Um in H,,. Consequently, the vectors

    X and Y are expressed

    by the joint canonical expansions: m™m

    X=met+ > Wu,

    Yom

    m

    +>) Wu.

    Dieaue

    (I)

    v=

    The variances of the random variables V; , ... , Vim are equal to 1, and the variances

    of the variables W, , ...

    , Wm

    are equal to 71, ...,

    Ym correspondingly.

    The

    308

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    covariance matrices of the random vectors X and Y are expressed by the canonical expansions

    m

    KS

    m

    Ne

    we

    Se ttt, :

    Hall

    E x am

    p1e

    (II)

    |

    3.9.17. Consider two m.s. continuous finite-dimensional vector

    random functions X (t) and bal alt t € T;. As F, we assume

    linear functionals as in Subsection 3.9.3.

    the same set of the

    Then we obtain the same

    H-space Hz.

    If zero subspace of the operator Kg is contained in zero subspace of the operator

    Ky and the operator

    A = kk Ky in H, has the spectrum which consists only of

    the countable set of the eigenvalues then these eigenvalues 7, and the correspondent eigenfunctions pr(t) are determined by the integral equation

    i)Ky (t,t’)e(t’)dt! = 7 /K,(t,t')e(t)dt’.

    (I)

    According to the proved theorem the random functions X (¢) and Y (t) are expressed in a given case by the joint canonical expansions:

    X(t) =me(t)+ )>Vw(t),

    Y(t)=my(t)+ 5 >Wru(t),

    (ID)

    where

    ere /K,(t,t)p,@)at! = = ‘Ky(t,t)p,(@)at’, and the second expression is valid only at Y,

    (AI)

    x 0. The variances of the random

    variables V, are equal to 1, and the variances of the variables W,, are equal to the correspondent eigenvalues 7. Here the covariance functions of the random functions XE (t) and ie) are expressed by the canonical expansions

    Katto

    yoany (tag(s

    Keg (tet

    We yy tty(up

    (t)ael

    (BV)

    p=1

    Canonical expansions (II) m.s. converge at each t€ 7}, and expansions (IV) at all be t"€ Wie

    Example

    3.9.18.

    If X(t) and Y(t) are two finite-dimensional vector

    random functions which have m.s.

    derivatives till the order n inclusively, zero sub-

    space of the operator

    K

    the same

    that in Example

    FP’, and

    H,

    A l= KS a AG y in H,

    is contained in zero subspace

    of the operator

    3.9.8) and the spectrum

    Ky

    (at

    of the operator

    consists only of the countable set of the eigenvalues 7, then

    3.9. Canonical Expansions

    309

    after determining 7, and the correspondent eigenfunctions (~, (t) from Eq. (I) of Ex-

    ample 3.9.17 we get expansions (II) and (III) which may be termwise differentiated n times with respect to t and n times with respect to t’.

    Example

    3.9.19.

    Let X(t) and Y(t) under the conditions of Exam-

    ple 3.9.17 represent the scalar random functions of the scalar variable t, t € [0,7; ]

    with the exponential covariance functions: = Ae~Altl r=t+tke

    Ky, (t, t) =

    De-@!"l

    Ty

    Ti

    A iecb lt= ti p(r)dr = yD iCe

    0

    p(r)dr .

    (1)

    0

    -s

    It is easy to verify by direct calculations that the operator space H,

    ana Ky (it)

    (1) has the form

    A =

    K a Ky

    in the

    (a+ Behl" 4 Pr $e Ah) f(r)dr.

    (I)

    is determined by formula

    Af =Kz'K,f =4 A Be

    =K-'K,f=

    t

    Ti

    el 2aD 0

    In fact for any function f(t) € H, we have Ty

    Ti

    UG RS A fe? —" Bat

    = Keg

    0

    a

    eth g(r)dr, 0

    where g(t) is the right-hand side of formula (II). It is clear from (II) that the operator B=A-

    yol at Yo =

    BA/aD

    is completely continuous and as a result of this

    the spectrum of the operator A consists only of the countable set of the eigenvalues

    17, }2 It is easy to check out by the substitution that the eigenvalues 7, at which Eq. (1) has the solutions different from zero are determined by formula

    YW

    phPA a? + w? AsaD— ae ats B? +w?2

    (v mea »2y eee be),

    ry. (IV)

    where W,, are the positive roots of the transcendental equation

    tgwT}

    _

    _

    (a+ Bw? — af)

    (w? 71 ap)? +

    (a + B)2w?

    )

    (V)

    310

    Chapter

    3 Moments,

    Characteristic Functions

    and Functionals

    and the correspondent solutions of Eq. (I) are the orthonormalized eigenfunctions

    pr(t) = oy [(a + Bw,(t)+ (1) + w(Ti)6E-—T)],

    (VD)

    where

    hoe

    VS

    DE +o) (B+ Gren}

    W V

    Finally,

    the

    (w+ B)(w2 +08) >”

    =

    cos W iz

    coordinate

    (

    sInW tact’)

    ; B)

    functions

    of the joint

    W V

    canonical

    ga oe:

    il

    expansion

    V

    (IV) of the

    random

    functions X(t) and Y (t) have the form Ti

    tt) = Dy

    py (T)dr

    0

    AC?

    Seilt—s

    = = fe Blt

    2 ala

    ly (r)dr = Dey

    +

    ee

    Plt).

    (IX)

    0 Similarly as in Subsection 3.9.3 it is sufficient to find the orthogonal eigenvectors

    (,

    of the operator At

    ]sea

    which

    are not obligatory

    normalized.

    Then the variances of the random variables V,, in expansion (3.9.37) will be equal to correspondent numbers

    D,

    = | Ypy [=

    ~y Key,

    the variances

    of the ran-

    dom variables W,, in expansion (3.9.37) will be equal to the correspondent numbers A=

    9

    a

    le Ky (Y,, and the coordinate vectors are determined by formula

    ge ee

    a eee

    (X)

    Here instead of (3.9.39) and (3.9.43) for the covariance operators Kz and Ky we get formulae

    Ke)

    pyr euy vol

    ie ae v=k

    ay

    (XI)

    Problems

    Sli

    Problems 3.1. Show that a scalar white noise may be obtained by the limit transition from

    any random function _X (t) whose covariance function Ky (t, t') decreases sufficiently quickly with increasing |¢ — rn | Namely if ele t’)| < a(t)o(t')e~olt-#

    then

    the random function Y(t) whose covariance function is determined by the formula

    t+t’

    Ky(tt) = AK, ( at the limit

    h—

    +h

    t—t’

    t+t’

    sola

    BRA

    ) noo

    00 turns into a white noise. Give the generalization in the case of

    the random functions in the finite-dimensional spaces. 3.2. Prove that for the reduction of the scalar random function Y (t) = bo (t)

    + by (t) X(t) to the stationary by linear transformation it is necessary and sufficient

    that its covariance function will depend on the difference of the arguments

    Kaltt = Ky (t2)/4/D,

    GQ)Dy) =, 8= 2).

    Give the generalization on the case of the functions in the finite-dimensional and infinite-dimensional spaces. 3.3.

    =

    Prove

    that for the reduction

    to the stationary

    random

    function

    Ys)

    bo(t) + bi (t)X (y(t)) by linear transformation with the simultaneous change

    of an argument it is necessary and sufficient that the ratio € of the partial derivatives of its correlation function may be expressed in the form of the ratio of the values of some function

    =

    wv) taken with inverse sign at the values of the argument

    equal to t and t’ correspondingly:

    _ DRylt t/t

    €= ——_

    =

    VO

    og gy

    OR,(t,t)/av ~~ We)’

    Kale

    etl Bie.Si en

    VD, Dye)

    Give the generalization on the case of the random functions in the finite-dimensional

    spaces. 3.4. Using the result of Problem 3.3 prove that the random function ‘a Ga with the covariance function

    Kyat) Se expla or

    Nar

    is reduced to the stationary one at p(t) =at+

    Get)"[a beet?) bt?.

    3.5. Using the result of Problem 3.3 prove that the random function Y (t) with the covariance function

    K,(t,t’) = Dexp{u(t +t’) — alt -t’|}

    Si

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    is reduced to the stationary by the transformation X(s) = e_ BY xX (s) is the stationary random

    function with the exponential

    (0), s = t? where covariance

    function

    R,(o) = De~*l¢l,¢ = 5 —s!. 3.6.

    Let the function K(epey

    t, t' € Ty; be the covariance function of a

    random process and R(u) be a polynomial with the positive coefficients. Prove that

    the function Ky(t,t’)

    =

    R(K(t,t’)) is also a covariance function of a random

    process. 3.7.

    Show

    that an expectation and the second initial moment

    of a random

    function

    Y(t) sat SE ant Xn(T)X1(7)dr 0

    where X (t) = whose

    components

    [Xy (t) ma

    iat

    (t) |? is a normally distributed random

    have the expectations

    process

    ™] (t) and the covariance and cross co-

    variance functions Kj; (ty . ta), are determined by the formulae n

    Wig (t) = i Ss anil ma(T)mi(7) + Kai(r, 7) ]d7, 9

    (1)

    ftl=1 t1

    te

    ra

    Py(tist) = ff S> *anrdyg[ Kui(n1,71)Kpq(t1, 72) oO

    0

    h,lp,q=1

    +Knp (11, 72) Kig(71, 72) + Kng (1, 72) Kip (11, 72)

    +Kni(t1, 71) Mp(71)1M4 (72) + Knap (11, T2) m1 (71)Mg (72)

    +Kng(t1,; T2)1M1(71) Mp(72) + Kip (71, T2) Mn (71)Mg (72) + Kig(71,; T2)Mn (71) Mp (72) + Kpg(T2,72)mMn(t1)mi(71) Jdridro.

    (ID)

    3.8. Show that an expectation and a variance of the random function Malis a)

    which expresses the relative time during which the random process X (t) exceeds the

    given function a(t) at the time interval (to, t),

    Y(t,a) = a | ux — a(r) ar,

    (1)

    are determined by the formulae

    ane, t tg @ Mie ry far f fileinae, oA) 46

    (II)

    Problems t

    1 Dy (t) =

    rage}

    t

    co

    (@—t)? tH)? ahJanan

    sii:

    co

    /

    / [ fo(@1,

    2; 71, T2)

    a(71) a(T2)

    —fi (x15 71) fi (#1; 72) Jda dae where

    f} (a; z), fo(a1 )Cosel, T2) are the one-dimensional

    densities of the random process

    (III) and two-dimensional

    (t).

    3.9. It is known that the number of the intersections of the scalar process X (t)

    by the given curve a(t) at the time interval (to, t) is determined by formula

    Z(ta) = f|X(ry= a(r)|8(X(r) = a()) dr,

    (1)

    Prove that the moment a@, (7 = 0, 1, 2,...) of the number of the intersections Z(t) is determined by the formula

    t a=

    t

    fan... fdr, to

    x fo

    0 where f, (x1, Ula

    to

    fancied lavas

    say

    Ty eee ted

    «dy

    (11)

    0 ce Cr US

    laee

    5 ff) is the r-dimensional density of the vector

    random function [ X(r) Yq) ies Vin) = X(T). 3.10. It is known that the number of the stationary points of the process X(t)

    at the interval (to, t), i.e. the points S where X(s) = 0 is determined by the formula t

    vl) = f (XC) Olar

    (1)

    to

    Derive the formula for the moment

    a,

    (r =

    0,1,2,...)

    of the number

    of the

    stationary points. Give the generalization on the case of the vector process. 3.11. It is known that the number of the exceeding by the scalar process X(t)

    of the curve a(t) at the time interval (to, t) is determined by the formula

    ¥(t,a) = /[X(r) — a(r)|1(X(r) — a(7))6(X(r) — a(7))dr.

    (1)

    314

    Chapter 3 Moments,

    Characteristic Functions

    and Functionals

    Prove that an expectation and a variance of the random function Y (t) are calculated by the formulae

    my(t) = fdr fnfi(a(r),a(r) +ns7)¢n, Dy(t) ii heb from

    (u)

    lotn)v dn) + ms at), ate

    +2; 71,72) —fi(a(tr), a(71) +15 71) fr (a(T2), @(72)+1025 72) Jdmidne (IIL) where filz, reas we T2) and fo(#1, G5 dias Coo Tt. T2) are the one-dimensional and

    two-dimensional densities of the vector process [ X(t) X (t) ‘ie 3.12. Show that under the conditions of Problems 3.9 for the stationary process AC (t) and constant level a(t) = d an approximate expression for the density of the

    variable Z(t, a) has the form

    f(z) = fol) fila

    1+ Dem]

    (I)

    OF 6 en §(z —m)

    (II)

    Here Pr (Z) are the orthonormal polynomials with the weight fo(z), po(z)

    =e

    r

    Dia)

    tee R=0

    C=

    Hhf(z)pr(z)dz = aro + So are =e

    (Far 2pses Ie)

    (11)

    k=1

    where @1, @2, ..-. , Mp are the moments of Z(t, a). 3.13. Under the conditions of Problem 3.12 restricting ourselves by the moments no higher than the second order show that

    Pra |) =

    _ (aatay Tag)(L+2on) da?

    ave

    3

    Ogee

    Capa

    oben [i+ a2 — a, — a? a2

    OF

    2 = rm], m=0,1,2,.00

    Problems

    315

    3.14. Generalize the results of Problem 3.12 on nonstationary case and show that

    the expression for the distribution function of the moment 7' of the first intersection has the form

    Ftp]

    P(P A, Yr (t). Show that one of its canonical expansions will be the following (Pez!

    ROH pe Vee),

    (I)

    where the uncorrelated random coefficients V, and the coordinate functions Wy (t) are determined from equations r—1

    A=V

    A= So tree os (22, SS

    (II)

    p=)

    by(t)=pr(t)+

    >> apgr(t) (V=1,...,n-1), r=v4+l1

    (11)

    bn(t) = Yn(t). Here d,, and the variances D), of the random variables V, are expressed in terms of the covariance moments avi

    kyi/Dy ,

    a

    hyp —

    k, p of the random variables Ap:

    -1

    . Yo dyrxGyra Dy Pail

    f-D eulthn fly

    15 BVI,

    (IV)

    Problems

    317 v-1

    DEST

    Dike koe

    59 lay.|

    2,

    Dx.

    (V)

    A=1.

    3.20. Under the conditions of Problem 3.20 show that if the scalar centered ran-

    dom function X (t) is represented by means of interpolational Lagrange polynomial:

    X=

    LMGart

    p(t) = (t— t1)(t —te)... (t- th),

    then in formulae (III)—(V) we shall have SS

    AE

    i

    Gt)

    Pa

    y |2,

    ee

    tw

    qe

    Wat

    =D),

    mfoperrit)

    oh

    in a)

    3.21.

    Let X be a normally distributed random vector,

    its covariance matrix.

    We

    divide the vector X

    Mm its expectation, K

    into two blocks X;

    and X2, X

    a

    =

    [xP xe | , and represent in the corresponding block form its expectation ™

    =

    [m? mr l , covariance matrix K and the inverse matrix C' = Tae

    Kir

    Ki

    | Ko

    Koo

    Ke

    ei Ona |

    se tel . C21

    I

    C22

    ()

    Show that conditional density of the vector X9 at a given value 21 of the vector xX,

    and conditional moments are given by formulae

    Wee

    fo(z2 |21) =[(20 1

    X exp {-3 [uz Coot

    =

    de

    iF ud C22(C yy C211) aE (Cy

    Cau)?

    Conus

    + (Cyy'C21u1)” C22CzyCam] }= [(2n)"|K|/| Ku ly? 1

    X exp {-3 [uz + (Ca

    Gag

    |C2

    =

    [us + Cy

    =[(2n)" |K|/| Kul?

    Co141 |}

    318

    Chapter 3 Moments,

    1

    F

    Characteristic Functions

    and Functionals

    2!

    i.

    io

    xX exp {-3 [u3, _ (Ka Kj'u)* | C22 [us _ Ka Kjtu]},

    (II)

    Kay. = Cy = Koo — Ka Kj} Kin,

    (III) (IV)

    |Koj| = [Keo — Ka Ky’ Ki2| = IKI /|Kul .

    (V)

    Maj. = M2 + Ku Ky (a1 — m1),

    ~

    CHAPTER 4 STOCHASTIC INTEGRALS, SPECTRAL AND INTEGRAL CANONICAL REPRESENTATIONS

    The main goal of this chapter is to give essentials of stochastic integrals theory and its applications to spectral and integral canonical representations based on stochastic measures theory. Section 4.1 is devoted to theory of m.s. limits. In Section 4.2 theory of m.s. continuity is considered. Special subsection is dedicated to stochastic processes with uncorrelated increments. Theory of m.s. differentiable stochastic functions in finite- and infinite-dimensional spaces is studied in Section 4.3. Sections 4.4 and 4.5 are dedicated to the integration theory over the nonrandom and random measures. Section 4.6 contains the stochastic Ito, Stratonovich and 6-integrals theory. Special attention is paid to the Wiener and Poisson processes and the corresponding measures. Processes with independent increments and the corresponding white noises are outlined. Section 4.7 is devoted to the theory of generalized stochastic functions. Detailed theory of the integral canonical and spectral representations is given in the last Section 4.8. 4.1. Mean

    Square Limits

    4.1.1. Definitions

    In accordance with the general notions of the convergence almost everywhere and in measure in Subsection 2.2.4 two types of the convergence of the random variables were defined: almost sure (a.s.) and in probability. Let us present the necessary definitions for the case of the random variables with the values on B-space.

    Let {X,}, X, = z,(w) be an arbitrary set of the random variables with the values in B-space dependent on the parameter r which assumes the values from some set R. In the accordance to the definitions of Subsection 2.2.4 the random variable X, converges almost sure (a.s.) to

    the random variable

    X = x(w) at r— ro if

    P(X, — X) = P({w : «,(w) > «(w)}) = 1. The random variable X, converges in probability

    (4.1.1)

    to X if at any € > 0

    lim P(X; — X ||> €) = Jim P({w : |]ee(w) — 2(v) |]> €}) = 0. (4.1.2)

    320

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    For random variables X, with the values in a linear space with a weak topology given by means of the set of the linear functionals F, two types of a weak convergence of the random variables are determined: a.s. and in probability. The random variable X, weakly a.s. converges to X if fX, a.s. converges to fX at any f € F,. The random variable X, weakly converges to X in probability if fX, converges in probability to fX at any f € Fy. Similarly in accordance with the notion of the convergence of the sequence of the functions from L, in p-meah the convergence of the random variables in p-mean is determined. Consider a set of the random

    variables {X,}, X,; = t,-(w) with the values in B-space X. that X, converges

    in mean

    of the order

    p or shortly

    It is said

    in p-mean

    to

    the random variable X = x(w) at r — ro, if there exist the absolute moments E]||X ||’, E|| X; ||P and

    E\|X,-X ||P 30 In particular,

    X,

    converges

    in mean

    at

    r—oro.

    square (m.s.

    (4.1.3) converges)

    to the

    random variable X, X, — X if E||X ||?, E| X; |]? < co and

    E\|X,-X|/?+0 at

    r—ro.

    (4.1.4)

    Hence the definition of a weak convergence in p-mean for the ran-

    dom variables {X,}, X, = z,(w) with the values in a weak complete topological linear space which a weak topology is determined by means of the set of the linear functionals F, follows. They say that X, weakly converges in mean of the orderp (in p-mean) to the random variable X

    if at any f € F, there exist E| fX |?, E| fX, |? and

    E| fX,—fX P40

    at

    r—ro.

    (4.1.5)

    In particular, at p = 2 we come to the definition of weak convergence in mean square: the random variable X, weakly converges in mean square ( weakly m.s. converges) to the random variable X, X, “> X if at

    f € F, there exist E| fX |?, E| fX, |? and

    E|fxX,-fX|)?}>0

    at

    r—oro.

    (4.1.6)

    From the general theorems of functional analysis follows that any sequence of the random variables convergent in p-mean also converges to the same limit in probability and we may extract from it the subsequence

    4.1. Mean,Square Limits

    aa

    convergent almost sure. Besides that if the sequence of the random variables {X;}, Xz = tz(w), El] Xz ||P < 00 (or El]fXz |? < oo at all f € F.) a.s. converges (weakly converges-or in probability) to the

    random variable X = z(w) and is majorized on the norm by the random variable U = u(w) > 0, EU? < oo, ||a,(w)|| < u(w) almost at all w (correspondingly by the random variable Us = uy(w) > 0, EU} < ©, |fax(w) |< us (w) almost at all w) then E]| X ||P < co (correspondingly E| fX |P < oo at all f € F.) and the sequence {X;,} converges (weakly converges) in p-mean to the random variable X. Remark.

    The sequence of the random variables convergent in P-mean may be

    not a.s. convergent and vice versa the sequence of the random variables convergent in probability or a.s. may be not convérgent in P-mean. Example

    4.1.1. Let Xp be a random variable with two possible values

    0 and n whose probabilities are equal to | — n-? andn~? correspondingly. We saw in Example 2.2.10 that the sequece TAK

    converges to () in probability and almost

    sure. But it does not m.s. converge to 0 as at all n

    E|X, —0|? =0-(1—-1/n?) +n? -1/n? = 1. Example

    4.1.2. Let X, be a random variable with two possible values

    0 and 1 whose probabilities are equal to 1 — n-! and n7! respectively.

    We saw

    in Example 2.2.11 that this sequence converges in probability to 0) but does not a.s.

    converges.

    But it m.s. converges to () as at

    nN—

    CO

    E|X, —0|?=0-(1—1/n)+1-1/n=1/n—>0. While studying m.s. convergence of the random variables we naturally suppose that there exist the second order moments at all considered random variables. Therefore we restrict ourselves to the random variables which have the operators of the second order moments. Later on it will be necessary to use the following lemma.

    Lemma

    4.1.1 (Loéve).

    Let {X,} and {Y;},

    re R, s € S be two

    sets of the scalar (tn the general case complex-valued) random vartables. If X, m.s. converges to X at r — ro, and Y,; m.s. converges to Y at $s — 80 then there exists limEX,Y, = EXY independent of the fact how the potnt (r,s) in the product of the spaces Rx S tends to the pont (ro, So).

    > For proving it is sufficient to notice that

    BX

    VAS EX VS EX,Y, ae

    DH E(Xn = X)Y,

    322

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    |EX,(¥, -Y) |? < B|X, PE|Y,—Y |, E| X, |? < 2E|X |? + 2E| X, —X [’, |BOX Xe X Se MARL, Therefore EX,Y, > EXY if E| X, — X |?

    0, E|Y,

    —Y |?

    0.4

    Corollary 4.1.1. If X, = X then EX, ~ EX and X° = Xo. > This follows immediately from the formula

    EA)EXpte

    prs

    \

    ASP da DEPome |? hg

    Corollary 4.1 2. IfX, —3 X, Y, 5 Y then the cross-covariance moment of the variables X, and Y, converges to the cross-covariance moment of the random variables X and Y independently of the fact how

    (r,s) tends to (ro, So). > Really, according to Corollary 4.1.1 xe ——

    EX°Y® = EX°Y°. 4

    4.1.2. Theorems about m.s.

    Seu ve —=;

    Y° therefore

    Convergence

    Let {X,,r € R} be a set of the random variables in a weakly complete topological linear space X whose a weak topology is given by means of the space of the linear functionals F. and [,, be a cross-operator of the second order moment of the variables X, and Xj.

    Theorem 4.1.1. Let the space ® of all linear functionals on F, coincides with X then for a weak m.s. convergence f X, to some random variable X in the same space it is necessary and sufficient that at any f € F. would exist the imit lim fI,;f = y(f). This condition implies T,S—-To

    that at any f,g € F, there exists the hmit

    lim

    fl,s9 = flg where T

    Tr,STO

    ts an operator of the second order moment

    of the variable X.

    > In fact according to the Loéve lemma it follows from f any f € PF, andatall

    X, i

    fX that at

    f,ge Fy

    fUrsg = E(fXr)(gXs) > E(fX)(gX) = fTg. In particular at

    g =

    fd follows the necessity of the condition of the theorem and

    o(f) = FTF. i flrsf= E(FX,)(FXe) > Of) at any f € Fe then E| fX, — fX, |? = B| FX, |? + Bl fXs |?

    4.1.

    Mean Square Limits

    323

    —E(fX,)(fXs) — E(fXs)(fX-r) > 0.

    (4.1.7)

    But the random variables f X; represent the functions fx, (w) of the variable W belonging to H-space D2(Q, S, PF) (Section 3.2), and | Tin \Pe = Bl oe

    ea There-

    fore from (4.1.7) in consequence of the completeness of H-space it follows the existence of m.s._ a random

    limit Us

    =

    linear functional

    im. fX, on F',.

    at each f € Fy.

    As according to the condition

    linear functional Y on FP’, presentable in the form pf then there exists a random f

    ©

    F..

    Thus

    X;

    function

    weakly

    m.s.

    This limit evidently is

    X

    in the space

    converges

    proves the sufficiency of the condition.

    = X

    to some

    ® =

    fx at some

    X

    any

    x € X

    such

    that U;

    =

    fx.

    random

    variable

    X

    what

    The condition of the coincidence of ® and

    xX (the reflexibility of the space X relative to F,) is- always fulfilled in important

    special case when some

    space

    the space

    Y, and

    PF, =

    X

    itself represents a set of all linear functionals

    Ej : Ee

    =

    ARI)

    ES Y}

    (ie.

    =

    in this case the set ® of all linear functions on F', is the space X. sis of Corollaries 4.1.1 and 4.1.2 of the Loéve lemma

    X,

    aang

    X

    Ey).

    on

    Really,

    On the baif and only if

    EX, +> EX and fKy.g = E(fXp)(gX9) > E(fX°)(gX°) = fKeg at all f,g € FP; (ie. when EX, X, and K,,

    converges to K

    linear operators mapping

    {L ¢ | filgi | x(t1) at allt; € Th. And vice versa if z,(t) — a(t) at all t € 7, then for any functional f of the form (4.1.8) fz, — fx. It is evident that in a given case the set of all linear functionals on F, may be identified with the space of the realizations of the random functions X,(t) as any linear functional on

    F, is expressed by formula (4.1.8).

    Corollary 4.1.3. For the m.s. convergence of X,(t) to the random function X(t) it is necessary and sufficient the existence of the limit

    bm. s(t =, Jim 9 Xft) X40) r,s—To

    T,S—To

    at each fired t € T; and here also exists the lamit

    lima Pog(t, 22) =n (DE

    a,

    Fe

    Hele

    r,s—-To

    Corollary 4.1.4.

    For the m.s.

    convergence

    of X,(t) to X(t) tt

    is necessary and sufficient the convergence of the expectations mz, (t) to m,(t) for allt € T; and the convergence of the cross-covariance functions

    Ko(tdi) t6cke(Cd sor, alliytele.

    Now let {X;,(t)} be a set of n-dimensional vector random functions, t € 7); F. be a set of linear functionals of the type

    N

    ee en.

    (4.1.9)

    kT

    correspondent to all natural N, to all n-dimensional vectors f,,..., fn with the complex components and to all t1,...,t~ € T;. Analogously as in the previous case we ensure that a weak convergence in the space of the realizations of the random functions X, represents a component-

    wise convergence of the vector functions at all t € T; and the set ® of all linear functionals on F, may be identified with the space X of the

    realizations of the random functions X,(t).

    Corollary 4.1.5. For the m.s. convergence X,(t) to some vector random function X(t) it 1s necessary and sufficient the existence of the lamit

    lim

    T,S—To

    *trt'-s(,2t) =

    im

    T,8S—1To

    tre

    X(t) XAt)

    et 675.

    4.1,

    Mean Square Limits

    325

    Here there exist the limit

    Li

    PH (pth) SD)

    X(t) iS D819 5 wtj tS Tye

    r,s—-To

    Corollary 4.1.6. For the m.s. convergence X,(t) to the vector random function X(t) the convergence of the expectations m;z,(t) to mz(t) and the covariance and cross-covariance functions K,s(t,t') to the covariance function K,(t,t'), t, t' € T; is necessary and sufficient. Example expectations

    Be Dye

    and

    4.1.3. Consider a set of the random functions X, (t) with zero with

    covariance

    and

    cross-covariance

    functions

    K,, th t )

    abt eh Dey= Wes] (Gol Mi haKre(tt)= Ker ltt).

    But according to Theorem

    4.1.5 X, (t) m.s.

    X (t) if and only if there exists the limit

    converges to some random function a =

    lima,.

    Here mz (t) ="

    ()Pand.

    K,(t,t')= a/2e7*l*- #1, Example

    4.1.4. Under the conditions of the previous example if lima,

    = 00 then X, (t) m.s. converges to none of the random functions. But as we shall see in Section 4.7, X; (t) m.s.

    converges to the generalized random function X(t)

    which represents a white noise (Subsection 3.2.3).

    4.1.3. Theorems about m.s. Space

    Convergence of Random Function in Linear

    Now we pass to the random functions with the values in the arbitrary linear space L with a weak topology given by means of the set of the linear functionals A. As usual we suppose that A represents a linear space and it is sufficiently rich for the fact that it will follow / = 0 from Al = 0 at all A € A. The cross-second moment I,,(t,t’) of the random functions X,(t) and X,(t’) in this case is a cross-operator of the second order moment of their values at any t,t’ € 7;. In the same way their

    cross-covariance function K,;(t,t’) represents a covariance-operator of their values at any t,t’ € 7). Example

    4.1.5. A scalar or a finite-dimensional vector random function

    of two variables Z(s, t), s € S,t € 7; may be considered as a random function of the variable t with the values in the space L of the scalar or the finite-dimensional

    vector functions of the variable s € S. As a set of the linear functionals on L we may assume a set of all the functionals of the type Al = ict Ak I(s% .: It is expedient to determine the set F', of the linear functionals on the space of the realizations of such

    random functions as the set of all functionals of the form (4.1.8) where f,..., fn

    326

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    are the arbitrary functions from A; f,,...,

    fy € A. We are frequently deal with,

    for instance, the product of the form g(s, t) X(t) or G(s, t) X(t) under the sign

    of the integral over t where g(s, t) is a determinate function, and G(s, t) is a random function of the variables 5, t. In this case it is expedient to consider Z(s, t)

    = g(s,t)X(t) or Z(s,t) = G(s,t)X(t) as a random function of the variable t with the values in the space of the functions of the variable s.

    From Corollaries 4.1.3 and 4.1.4 about the convergence of the scalar random functions we have the following statement. \

    Corollary 4.1.7. The sequence X,(t) weakly m.s. converges to some random function X(t) with the values in L if and only if at any A EA there exists the limit

    lim

    AT,,(¢,t)A =

    for each fixed t € T,. the limit

    lim AL, s(t,’

    r,s

    lim

    E[AX,(t)][AX;(t)]

    Under this condition for all X,X' € A there exists

    = E[AX(t)][VX@)] = AT (et),

    Corollary 4.1.8.

    For a-weak m.s.

    tT.

    convergence of X,(t) to X(t)

    there are necessary and sufficient a weak convergence of the expectations

    Mz,(t) to mz(t) and at each A € A the existence of the limit

    lim AK;s(t,t)A =

    r,S—To

    4.2. Mean

    lim E[AXP(t)][AX2()].

    T,S—To

    Square Continuity of Random

    Function

    4.2.1. Definitions

    The random function X(t) with the values in B-space defined on the topological space 7} is called continuous in mean square (m.s. continuous) at the point t € T; if at any € > 0 there exists such vicinity V;(e) of the point ¢ that

    E\| X(t’) — X(t) ||?

    The necessity of the condition follows directly from the Loéve lemma:

    from

    X(t!) = X(t) follows T',(t’, t”) = EX(t') X(t”) > E| X(t) |? = Te (t,t) at t’,t'’ —

    t.

    For proving the sufficiency of the condition we notice that from

    [Uae .t)— y(t) |< 2/3 at allt” © V4(e) we have

    E|X(¢)—X(@)t=T.(t,1)—Ts(t,t)—

    SA

    elea) Leila teat cae

    Hence by virtue of the arbitrariness

    € > 0 it follows m.s.

    continuity X (t) at the

    point t. If X(t) is m.s. continuous on 7} then from the Love lemma it follows that

    ee}



    Tz (t1,t2) at t’ >

    th, t!” —

    to, ie. the continuity Tre

    | at all

    points of TJ, x 7. From the relation

    E| X(t’) — X(t) |? =| me(t’) — me(t) |?+ B| XO) — X°(t) |? and just proved statement we conclude that for m.s.

    continuity of the random func-

    tions X (t) at the point t (on 7}) the continuity of its expectation at the point t (on T}) and the continuity of its covariance function K; (t, t') at the point (& t) of

    the product of the spaces 7, X J}, then K,(t, t’) is continuous on 7} X JT} are necessary and sufficient.
    to and with its increment

    on the interval [t,to) taken with the

    inverse sign if t < to. Later on we shall consider only such processes with the uncorrelated increments. Any process with the uncorrelated increments X(t) is reduced to such a process. Really, the random process Y(t) = X(t)—X¢, represents a process with the uncorrelated increments which possesses the required property Y;, = 0. Thus if X;, = 0 then the value X; of the process X(t) at any instant ¢ is uncorrelated with its future increments on the intervals which follow the instant to and

    4.2. Mean Square Continuity of Random

    Function

    ~

    329

    with its previous increments on the intervals which precede the instant

    bo: BXP(XP, XP)

    = 0 atit' to orta< te By Theorem 4.1.2 of the m.s. —

    ©o.

    of the Ito integral Y.

    convergence the limit

    if and only if there exists the limit HY, ved independent m

    (4.6.28)

    -

    y =

    l.i.m.Yp

    exists

    of the way in which n,

    By this theorem the necessity of the condition follows immediately from

    the equality

    Bly, =-y E|xaMy| sa) - 4] 2

    To derive this equality we note that

    B Wal =So ExW,) KEM) [we?) — WE») | Ife

    jes

    x |wl”) — xi") ] and by independence of wit) — we, ) on sate )), xt *) and wt)

    = wt?)

    at any k > l,

    BX (HX) (WEP) -WER» ][wel — war) | = Bx) x@@) [wWe™)-wa] B[wal”) -WE] =o

    B|x@@of |wee)- wey],

    360

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    (nm) _ rcp). (n) |? # |wee?)-wee,) =6|x¢e®,)|

    (n)

    Further, by independence of W(t.)

    B|waP)|

    fad)

    on W (t;, )

    |”

    oye (adores gp oe) yielding

    =2|we | +2 |wEr)- wary] 2

    A

    2

    Consequently,

    B\we)- we) =2 wey] -2 |wE»|

    2

    EP RG) 2 > To prove the sufficiency of the condition we unite two partitions P,, Py, of

    the interval (a, b}. Let P be the resulting partition, P to

    =a,tg=

    : (a, b] =

    Q Dr, (ts-1 ; cae

    b where each point ts coincides with some 4”? or with some aia

    Then we obtain

    Q ea

    X(r{) [k(ts) — k(ts-1)]

    ator lacimdgian tmfieSotyly at A(t, Hentai enhl By virtue of the m.s.

    a ree, cei

    continuity of X (#), lim EY,, Yn coincides with the limit

    lim wae DP atts)=kta]

    n,m—oo

    b

    = [EXP ak(o.« a

    The definition of the It6 stochastic integral is easily extended to the

    case of a matrix process X(t) and a vector process W(t). Theorem 4.6.7. The vector It6 integral (4.6.27) exists if and only af all the scalar Ito integrals involved in (4.6.27) exist. The necessary and sufficient condition of the existence of the vector integral (4.6.27) is the existence of the Riemann-Stteltjes integral b

    Hf,EX(t)dk(t)X(t)* = BYY*, a

    4.6. Stochastic Integral ~

    361

    which ts equal to the covariance matrix of the vector Ité integral Y. The existence of the integrals b

    [ BX. (t) Xn (t)dkjn(t) a

    is an immediate consequence

    of the existence of integral (4.6.28) for all

    the elements of the matrix function X(t).

    Let Y(t) be a random process determined by the formula 7 ya

    Y(t) = Y¥(to)+/Xi(r)dr + /X(r)dW(r),

    (4.6.29)

    where the first integral represents a m.s. integral of a random function

    X(t), and the second integral represents the Ito integral of a random function X2(t) with respect to a process with independent increments

    W(t). The random functions X;(t) and X(t) are supposed satisfying the conditions under which both integrals exist at any ¢t in some interval

    T.

    Let X(t) be a random function.

    The Ito integral of the random

    function X(t) with respect to the process Y(t) in the interval defined as the sum of the integrals

    (a,)] is

    /X(r)d¥(ry= fx x(x eigde /X(r)Xo(r)dW(r),

    (4.6.30)

    where the first integral represents a m.s. integral and the second integral represents the Ito integral. Here, certainly, the random functions

    X(t)X1(t) and X(t)X2(t) must satisfy the conditions under which both integrals in the right-hand side of formula (4.6.30) exist. 4.6.8. Stochastic Integrals over Wiener and Poisson Measures

    As it is known (Subsection 4.6.5) the stochastic measure of the Wiener process W(t) is defined by Z((t1,t2]) = W(t2) — W(t). So the It6 stochastic integral may be calculated by formula (4.6.27) because

    condition (4.6.28) is fulfilled.

    362

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    Example

    4.6.3. Find the variance of the stochastic integral

    = /X(r)dW(r),

    where X(t) is a random function with the expectation Mz (t) = ct and the covariance function Ky (ty ; ta) = Deb(titt2)—o |ti-tal , and W(t) is a standard Wiener process. According to formula (4.6.28) and taking into account that v(t) = ‘

    E'|X(t)|? = |ms(t)|’ + Dz (t) = c7t? + De? | 4B

    TP

    Dive ee [ra - Df rar = eT? [3+ D(e7#? — 1)/2je. 0

    Let P°(A, A) be a centered Poisson measure, Y(t, u) be a real m.s. continuous random function of two variables ¢t and u with second order moment independent of P((t;,t2],A) at any t < t € Ri = R2\{0}, A € Bo. Let {P,} be a sequence of partitions interval (a, b] and of set 0 < |u| < R,

    : (a,0] = U I, }, M20,

    =5,

    scalar finite < to, of the

    (4.6.31)

    {u:0< |u| < Ry= Ua”, A\”) €Bo, fe) such that max(t)

    2

    Me

    1) 70, max

    sup

    |u’—u|

    ~0asn

    oo.

    u weal)

    M.s. limit 7” of the sequence of integral sums

    Sa. Ue) tena) Poe vant =i

    Mbeya Rate tA) 9(4.6539)

    ysl

    if it exists, is called a stochastic integral of the random function Y(t, uv) with respect to the Poisson measure P(A, A) extended over the interval

    (a,b] and the set {u : 0< |u| < Ry}, b

    fe =)

    /

    a 0f elt, Maple, d)*a9().

    (4.28.39)

    Besides we notice that condition (4.8.35) may be written in the equivalent form N

    Df volt dd)ap(r, A)? = 16(t — 7),

    (4.8.40)

    where J is the unit matrix. In particular, when A is a finite-dimensional space and the measures

    Hp(t, B) and o,(B) are absolutely continuous relatively to the Lebesgue measure, formulae (4.8.38), (4.8.40) and (4.8.34) may be presented in the form 1 ot A) say |Katt EX Ja,(t’, r)dt’ Nae. ,

    (4.8.41)

    N

    > fatt A)ap(7, A)?dd = 16(t — 7),

    (4.8.42)

    i)as

    (4.8.43)

    ee ees

    where vp(A) and a,(t, A) are the Radon—Nikodym derivatives of the measures 0,(B) and p,(t,B) with respect to the Lebesgue measure. The

    386

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    stochastic measures lated white noises

    Z, will be in this case generated by the noncorre-

    V(A) = iixc r)T X°(t)dt ,

    (4.8.44)

    whose intensities are equal to the correspondent functions v,(A). Integral canonical representation (4.8.36) of the random function X(t) will take the form

    . N

    X(t) = melt) +)» |Vo(A)aplt, 4A, _

    (4.8.45)

    and the correspondent representation of the covariance function (4.8.39) will be N

    K,(t,t') =~ /Yp(d)zp(t, Nxp(t’, AAA.

    (4.8.46)

    p=l Thus for constructing an integral canonical representation of the

    random function X(t) it is sufficient to find the functions «p(t, A), ap(t, )

    and v,(A) (p= 1,..., N) satisfying conditions (4.8.41)-(4.8.43). Here the random function V,(A) determined by formula (4.8.44) will be uncorrelated white noises with the intensities vp(A) and the random function X(t) will be expressed by the integral canonical representation (4.8.45). Example

    4.8.1. Let the random function NS

    X(t) is defined by the

    linear differential equation 1 (t)X + do(t)X = V with the white noise V of unit intensity, and X (to) =

    0 and is equal to zero at t =

    to. While integrating this

    equation we find

    X(t) = a(6) iSoy no =e {-|. ar}. ay(T)q1(7) a;(T) 0

    Formula (I) gives an integral canonical representation of the random function X at

    time interval tg

    < t < to + 7}, T; > 0. The coordinate functions of this integral

    canonical representation are determined by formula

    heges oti venscull) seal ooh

    (II)

    4.8. Integral Canonical and Spectral Representations

    387

    After rewriting formula (1) in the form totTi

    vine / [ay(A)6"(A— t)+a(A)6(A—4)] X(t)at

    (HIN)

    to

    and comparing (III) with formula to+Ti

    V(\) = 4 a(t, \)X°(t)dt , to

    (Iv)

    7 ea

    we see that in a given case a(t, A) = ay(A)6"(A — t) b ao(A)d(A ~— ye

    4.8.4. Integral Canonical Representations and Shaping Filters

    In applications the pairs of the matrix functions z(t, A), a(t, A) satisfying conditions (4.8.42) and (4.8.43) are often known. So, for instance, the weighting functions of two any reciprocal systems with N inputs and N outputs w(t, 7) and w(t, 7) always satisfy the conditions

    (Subsection 1.2.4)

    /sc(6, Meta. eld Meals

    ea

    ‘ianette aNee TOR Dy,

    (4.8.47) (4.8.48)

    Comparing these formulae with (4.8.42) and (4.8.43) we see that if we determine the matrices-columns z,(t, A) and ap(t,A) (p=1,..., N) as

    the correspondent columns of the matrices x(t, A) = w(t, A) and a(t, \) = w~(A,t)? then conditions (4.8.42) and (4.8.43) will be satisfied. Thus for finding an integral canonical representation of a random function tt is sufficient to choose among all possible pairs of the weighting functions of the reciprocal systems such a patr which satisfies condition (4.8.41) at some functions vp(X). In the general case this problem remains still rather difficult and not always has the solution. We notice that the construction of an integral canonical representation of a random function is equivalent to finding such a linear system which transforms the vector random function X(t) into the vector white noise V(t) of the same dimension with uncorrelated components. The inverse system will form the random function X(t) from the white noise

    388

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    V(t) in consequence of which it is usually called a shaping filter of the random function X(t). Thus a mathematical problem of finding an integral canonical representation of a random function ts equivalent to the problem of finding a shaping filter ofa given random function. For many standard random functions (for instance, stationary) the problem of constructing the integral canonical representations is sufficiently simple (see Sections 6.4 and 6.5).

    4.8.5. ables

    Joint Integral Canonical Representation

    of Two Random

    Vari-

    We return to the problem of Subsection 3.4.7 and consider two random variables X and Y with one and the same phase space and the covariance operators K, and Ky. Suppose that zero subspace No of the operator K, is fully contained in zero subspace of the operator Ky. It was shown in Subsection 3.4.7 that in this case there exists the operator A = K;!Ky which represents a positive self-adjoint operator on Hy. As it is known from functional analysis each positive self-adjoint operator may be presented by a spectral representation co

    A = fre(aa),

    (4.8.49)

    where €(B) is a representation of the unit on Borel o-algebra B of the

    real axis €(B) = 0 for any Borel set B of a negative semiaxis by virtue of the positiveness of A generated by the operator A. Let exists such a sequence of the vectors {y,} C H, that the set of the linear combinations of the vectors €(B)y, correspondent to all B € Bandy =1,2,..., is

    dense in H, and the vectors €(B)y, and €(B’)y, are orthogonal for ue #v and any B, B’ € B. We choose such functions v,(A) (v = 1,2...) that for any vector f € H, the equality has the place

    padtacleh pop ecregeruen

    = ON

    Then the random variable X and its covariance operator pressed in terms of the integral canonical representations

    (4.8.50) Kz are ex-

    X =m, + > f wana),

    (4.8.51)

    Kes fay [0)] oy (dd).

    (4.8.52)

    4.8. Integral Canonical and Spectral Representations

    389

    Here Z,(B) = €(B)y,X° are the uncorrelated stochastic measures with the covariance functions o,(B) = (€(B)y,Ww) =|| E(B) ||? and

    ways px,

    (4.8.53)

    Theorem 4.8.3. Let X and Y be two random varaibles with one and the same phase space and covariance operators Kz and Ky. Let the operator A = Kz!Ay exists and defines the canonical representations

    (4.8.51), (4.8.52). Then the random variable Y and its covariance operator Ky are expressed in terms of the integral canonical representatios

    Y =m, + git ,

    (4.8.54)

    K, ==>[mo

    (4.8.55)

    Here U,(B) = E(B)yY°

    ) |Aov(ad).

    are the uncorrelated stochastic measures with

    the covariance functions

    7,(B) = (AE(B)w,wv)= /avery

    (4.8.56)

    B

    Formulae (4.8.51-(4.8.56) define the joint integral canonical representation of two random variables X and Y.

    > Really, as EU,U,(B’) = E(B) KyE(B’Yn = E(B)WyKr AE(B')7p =

    (Bye,

    AE(B')y,); and any self-adjoint operator is permutable with its ex-

    pansion of the unit then

    BU, (B)Tn(BY = (E(B), Arn) = fMECBB') > Em) = f ME@) 10)=5. | Aov(ad). BB!

    BBE

    Hence it follows that U, (B) = E(B) py Y © are the uncorrelated stochastic measures with the covariance functions T;, (B). Further as for any f € Hy 2

    E \fy°- »alfv,(A)U, (dd)

    390

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    =(ar-ayo feee@rns-Y [focrelayn )= 0 co

    by virtue of the fact that f =

    SS ‘ifey (A)E(dA) Ww then the random variable Y v=1

    is expressed by formula (4.8.54) . Finally from the direct theorem about the integral canonical representations it follows that

    Ke S [0 [= O) (ad). 4

    Remark1.

    In particular if there exists such vector

    Y € Hy that the set of

    the linear combinations of the vectors E(B)y correspondent to all Borel sets B of

    the real axis is dense in H

    then sums (4.8.51), (4.8.52), (4.8.54) and (4.8.55) contain

    one item and these formulae take the form

    X= og + froz@), ie fo |x) |o(da), (4.8.57) Y =m, + froua@), oe fo [e) |Ae(ar) (4.8.58) Here Z(B)

    = E(B)yX°®, U(B) = E(B)yY°

    are the stochastic measures with

    the covariance functions o(B), and

    AES ieie

    (4.8.59)

    B

    By virtue of (4.8.53) the function v(A) is equal to

    dZ v(A)

    =

    BX*——

    dU

    (A) = EY°—()). dr

    (4.8.60)

    In this case when the spectrum of the operator A consists only of the countable

    set {7,} of the eigenvalues 7, we have E(B) =

    ya (., Pv) Pv where (, is the WEB

    eigenvector correspondent to the eigenvalue 7,

    E(B)py = 3

    py

    at

    v€& B;

    at EB

    and the set of the vectors € (B Voy is dense in H,. Integral canonical representations

    (4.8.51), (4.8.52), (4.8.54) and (4.8.55) of the random variables

    X and Y and their

    4.8. Integral Canonical and Spectral Representations

    covariance

    operators

    turn into the canonical

    = E(B)p, X°, W.= Remark

    expansions

    391

    (Subsection 3.4.7) at V,

    E(Byo,¥*, any Borel B, y, € B and u, = v(yw).

    2. Let us consider the self-adjoint operator

    A=

    Aa

    Ky which is

    determined as an operator of the equation

    /Read at a given function f(t) € Da

    representation

    (4.8.49).

    /Ky(t, r)f(r)dr

    C Hg. We know that A is expressed by a spectral

    If there exists such a sequence

    of the functions

    {Vv (t)}

    C H, that the set of the linear combinations of their projections E(B) py at all Borel sets B of the real axis is dense in Hy and E(B)y,

    and. ECB

    Ys are orthogonal at

    Ll a y and any B, B’ € B then according to the proved Theorem 4.8.3 the random

    functions X (t) and Y (t) are jointly expressed in terms of the integral canonical representations

    X(t) =m, (t) + byfr, Z,(ad), be

    ¥(t) = my(t) + > fvv(t, UL(dd).

    (4.8.61)

    vel

    And the following relations take place:

    Katt) =) A) Bitrate AGA GA) ,

    i,

    (4.8.62)

    Ky (tt) = 50) Ort)

    vy (C A) Aos (a),

    Hemet

    where

    re

    ae w(t.) = BX) soe S20)

    (v= 1,2,...).

    (4.8.63)

    In the special case when there exists the unique function y(t) for which a set of the linear combinations of the functions E(B)y correspondent to all Borel sets B of

    the real axis is dense in H, the sums in (4.8.61) and (4.8.62) contain only one term in consequence of which formulae (4.8.61) and (4.8.62) have the form

    X(t) = malt) + fo(t,2)2(0), Y(t) = my(t) + fv(t, )U(@X), (4.8.64) K,(t,t’) = /v(t, A)v(t’, A)"o(dd), Ky(t,t) = iv(t, A)v, (t’, A)"Ao(dd), (4.8.65)

    392

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    where

    Te

    Bx.

    Zy.

    (4.8.66)

    4.8.6. Spectral and Integral Canonical Representations of Scalar Random Functions Theorem 4.8.4. The spectral representation of a scalar m.s. continuous stationary random function with vector argument is one of its possible integral canonical representations. > Let X (t) be a m.s.

    continuous

    scalar stationary random

    function of the

    n-dimensional vector t, t € R”; kee) be its covariance functions;

    X be a linear

    space of all the functions of the variable t; F’, be a set of the linear functionals N

    {eS

    yy Tea UA) correspondent to all NV, f;,..., fn, ti, ..., tn. In this p=l

    case H, represents the range of an isometric mapping of a subspace in the H-space

    D4 (Cres pe a) formed by the compliteness set of all functions (A) = ip So eid" te : As the space D2(R”, B”, co) is separable then Hz is separable too. We determine in H,, the expansion of unit E€ by formula

    E(B)f = wae [Ota

    fsje

    rar

    B

    +7 ‘dd. ae1 | ta )e ear

    8. (4.8.67)

    = —— | 1p(A)p(A)e™

    We assume Y = 6(t). Then by virtue of (4.8.67) we have

    chit | 190) €(B)y(t) ub = mate aay |Brine ta) dN eds= fo aa

    en atheiA7t +e

    B

    As a set of the simple functions is dense in D2(R”, ‘igs co) so the set of linear combinations > GL (Bi) ; ¢; € Cis also dense in Hy. Consequently, for obtaining a

    an integral canonical representation of a random function we may restrict ourselves to one expansion of unit € and one vector 7. Formula

    Z(B) == €(B)yX°

    pee

    ae: = Ca [x 0 (dt [tare

    =

    \"tdd

    (4.8.68)

    4.8. Integral Canonical and Spectral Representations

    393

    determines a stochastic measure with the covariance function o( BB’), and for any f € Hg we have

    (fF, E(B) = [eQ)taQ)o(ar) = fotar fsme

    at.

    B -\T

    By virtue of (4.8.9) u(A) = vlf(t)e

    -\T

    ‘dt, and a(X) = rt

    Consequently, the

    random function X (t) is determined by the integral canonical representation of the

    form (4.8.20):

    X(o=imyg + fe Z(a0),

    (4.8.69)

    Integral canonical representation (4.8,21) of a covariance operator in this case gives the integral representation of the covariance function

    kale ) (the known

    Bochner

    theorem)

    ky(ty—ty) = [erersa(aa), d Formula (4.8.69) gives a spectral representation

    (4.8.70) of the stationary

    function X(t) with vector argument. According to this fact the measure o in (4.8.70) is called a spectral measure of the stationary random function X(t). In particular, when the part of the measure o which is singular relatively to Lebesgue measure in R” is concentrated on the countable set of the points we may determine a spectral density of the

    stationary random function X(t):

    se(A) =

    "S,(A) _ 8"a((—00,A) Ory NES = GAGA OAL

    (4.8.71)

    which may contain a linear combination of the 6-functions. The stationary random functions encountered in the problems of practice have always the spectral densities in this sense. Then spectral expansion

    (4.8.69) may be rewritten in the form

    XO) = ms + fvaye** an, where

    (4.8.72)

    V(A) is a white noise whose intensity is equal to the spectral

    density s;(A) of the random function X(t). As the spectral density S_(A) is integrable (by the finiteness of the spectral measure o) then it is expressed by the Fourier integral

    Se(A) = oon nr far fsa (meu -\? dp.

    394

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    But according to (4.8.70) we have

    [eese(wydy = /eH"? (dys) = ke(r). Consequently, the spectral density of a scalar stationary random function with vector argument is expressed in terms of its covariance function by formula ie = oa (onyr | ie(r)e-?” . dr glee:

    Example

    4.8.2.

    Find an approximate canonical expansion of a scalar

    stationary random function of the scalar argument t, spectral density Sz (A) at the time interval interval of the frequencies.

    X =

    We partition this interval into 2N + 1 equal segments We enumerate the segments of the real

    axis obtained by such a way by the integers from —N

    segment.

    X(t) possessing the

    |¢| —aclneinet (-Q, Q) be an arbitrary

    and denote the length of each of them by 2.

    mean of the ®

    AOir (4.8.73)



    to N and denote by A, the

    Substituting the integral in the infinite limits in integral

    canonical expansion (4.8.72) by the integral in the limits (-Q, Q) we present the obtained integral in the form of the sum

    of 2NV + 1 integrals extended

    over the

    segment 2@ length. In each integral instead of the multiplier e’*'tAt we substitute its value in the middle of the interval etrvt Then we obtain the random function N

    Xw(t) = mz + yap Vee

    (I)

    where Av+ta

    ee

    sf V(A)dA (v = 0,+1,42,..., +N).

    (11)

    Ap-—a

    The variances D,, of the random variables V,, by virtue of the spectral representation of the variance

    X are equal to Av+a

    Dos

    i Sa(A dA @ = OFEL SEN’); Av-a

    As V(A) is a white noise and the integration intervals in (I) at the correspondent values of the index Y do not overlap then the random variables V, are uncorrelated.

    We estimate the degree of the proximity of X y (t) to the function X(t). We have

    x()-Xw(=

    -

    [ee)

    f VO) art

    [VQ)N)e* dr

    —co

    a

    4.8. Integral Canonical and Spectral Representations N

    395

    Av+ta

    +),

    / V(A) (ec? — ef**) dd.

    aN

    (111)

    mit

    As the expectation of the difference X — Xy is equal to zero and the isolated items

    in the right-hand side of (III) are uncorrelated then the variance of the difference X — Xj equal to the expectation of the modulus square of this difference is equal to

    the sum of the variances of the integrals entering into (III). Each of these integrals represents a random function expressed by integral canonical representation (III). Therefore we may use formula (II). Consequently, —2

    fore)

    E |\X(t) —Xn(t)|? = / so(A)d+ f sa(X)d —0oo

    N

    2

    Avpta

    4 se

    / $z(A) te woe

    bi

    (IV)

    geet 5 ars

    Supposing that the interval

    |t| < T is fixed we choose the frequency

    {2 and the

    number @ in such a way that the first two integrals in (IV) would be smaller than the arbitrarily given number €: -2

    co

    / s2(A)dd 1 we get from (V)-(VII)

    pw(Bn) < c2-"-)/(1 — 278+!) vn,

    (VIII)

    uw(N) < pw(Bn) < 277-9),

    (IX)

    where c’ = e/(1 = dap 1), As inequality (IX) is valid for all n then btw (N) =+(). -_

    Leta

    ee)

    =

    -

    _

    U By.

    _

    Any function x(t) € N belongs to set B,, and conse-

    n=1

    quently, to all sets Ap, Die iso.

    ‘calls 4h Pat(m

    12" y= atm2-?) (Al

    (X)

    for all sufficiently large p. Let s, s' € S be arbitrary binary-rational numbers, qd be

    the least natural number satisfying the condition | s'—s

    I< 2-4+1.

    In this case

    there exists a point $9 = m27~?1 between the points s and s’ for some ™m and i

    Sal chemi epeeah

    (XI)

    p=

    where the numbers point

    Sg

    to

    the

    k; are equal to zero or unit. point

    S

    sequentially

    by

    If we realize the transfer from the

    the

    steps

    of the

    length

    boot

    §

    ; kids i" then due to (X) we get

    oe (44

    Wieiayke aster (a-P)a < AQ- Beswie p=1

    |2(s') -2(8) Ik

    =
    0 at s’ > 5, s, s’ € S C (0, a]. This proves that all the functions z(t) es

    excluding the set NV of zero measure [hw

    are continuous on the set S of binary-rational numbers for any finite interval (0, a]. Now let us take an arbitrary t € (0, a) and the sequence

    {3,.} of binary-rational

    numbers converging to t and |Sp — t I< 2~?. Define the sets

    Ap = {2(t') :| a(t) — a(sp) |= A2-P*}, lee)

    co

    Bris "| eAs

    Pao

    p=n

    fae

    Poy)

    N=

    Analogously we prove that pw (N) =

    0 and that the sequence

    A2Se))

    con-

    verges to “(t) beyond the set N.Thus at any fixed t € (0, a) almost all relative to the measure

    for any @. Alor

    {yw

    functions from

    the space

    XT

    are continuous

    at the point t

    This result make possible to extend the measure [yw on the 0-algebra {B

    “DS

    AC, CG e€ i

    of the subspace C’ C XT of continuous func-

    tions from the space of all functions NG putting pw (B) = Lw (A), Be

    AC:

    As it is known any function from the space XT? without changing the values

    of measure on rectangle sets from the space X T and also on all sets of an induced o-algebra may be replaced by some so called separable function. almostly defined for all t by its values on some

    Such a function is

    countable set called a separant.

    our case separant is the set of all binary-rational numbers

    In

    and separable function is

    the set of all continuous functions. Notice that the Lebesgue extention of the measure [/w initially defined on the

    sets of the rectangles of all functions X F may be determined on the space C’ of all continuous functions.

    4.12.

    Prove that the 0-algebra A‘TC in Problem 4.11 coincides with the o-

    algebra C of the space C' = C’ (0,oo) of continuous functions with supnorm induced by the set of all balls.

    4.13. Prove that almost all relative to the Wiener measure Jw functions of the space Fe

    pig are equal to zero at

    ft=

    0.

    4.14. Let us consider the function f(x) — Or (s)x?(t1) + ~2(s)x?(t2) where Y1 (s), p2(s) are the continuous functions on the bounded closed interval S, i.e. the elements of the B-space C(S), and a(t) is a scalar function belonging to the space

    XT? with the o-algebra A

    Oe

    (0, 00) of Problem 4.8. Prove that the function

    f(z) maps the measurable space (xt F A’) into the separable B-space C(S).

    Solution.

    This function is (Ar, B)-measurable as the inverse image of

    any ball of the space C(S)

    f~'(Sr()) = {2(t) : Y(s) — r < g1(s) 2? (1) + G2(s) 2? (te)

    Problems

    < ¥(s)+7r, may

    be presented as a countable

    dimensional bases w(s) tk

    415

    Vs € S}

    intersection of the cylinders in XT

    with two-

    Sieh) (s) ze + ~2(s) 25 )ye(s) 2? (te), ha

    mapping the space XT? of Problem 4.8 in the separable B-space C(S) is proved.

    4.15. Prove the measurability of the function b

    F(a) = fols) = f(s, 2°@ at, a

    where y(s, t) is a continuous function of s, t, s € S, S is a bounded closed set. Solution.

    tEef

    =

    This function maps the space of all functions A? of the variable

    [0, oe) ) into the separable B-space C(S). In accordance with the result

    of Problem 2.3.8 the function f(z) is determined almost everywhere in XT relative to the Wiener measure [/yy.

    Therefore almost for all functions x(t) € XT

    there

    exists the limit

    :

    ¢ witpahigth=5

    3

    f(t) = lim fa() = lim —— 2, Pls, te) 2 (te), th =at+k

    (b = a)/(n = ty But all the functions fy, (x) are CA. B)-measurable

    according to the result of Problem 4.14. Consequently, f(x) represents the limit almost everywhere relative to the measure [lw of the convergent sequence of (AR; B)-

    measurable functions.

    So the function F(z) is (Al. B)-measurable where Aly is

    the complement of the d-algebra AT? relative to the measure Hw, i.e. [Ly is measurable.

    4.16.

    Calculate

    + ~2(s)z?(t2)

    LLw

    the

    integral

    of the

    function

    f(z)

    over the whole space of the functions XT

    =

    (s)x(t1)

    on the Wiener measure

    (Problem 4.8) if p14 (s) and p2(s) are continuous functions on bounded closed

    interval S' of the real axis. Solution.

    In a given case the argument

    of the function f(z) is the

    function x(t) from the space X° of scalar functions of t € S on which the Wiener measure /ly

    is determined.

    The space of the values of the function f(x) is the

    B-space of continuous functions C(S). As the function f(z) depends only on the

    416

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    values of the function x(t) in two fixed points t;, tg then for calculating the integral f f(x) pw (dz) we may use the expression of Problem 4.8 of the measure rectangular sets at

    yw

    on

    2 = 2. As a result we receive

    | #@nw (a2) = RE —

    /3[vr(s)e

    +2(s)x3 | Jexp{-— ,ee

    |

    (1)

    For calculating the integrals over 21, 22 we use formulae

    per

    cx l2de =

    2m 32/20

    Ve

    / —ex?/24

    2a

    ;

    5S)

    ee Toe)

    foe)

    [oe

    / ze

    /dz = 0,

    / wre

    /2dz = de|iss €

    So we find

    (11)

    C

    J fo) (a2) = or(sts + g2(0)t2-

    cu)

    4.17. Using Problem 4.16 calculate the integral of the function

    fies S> ve(s)2?(te),

    Cee Clos)

    Ceeaibe 2, re)

    ke!

    over the whole space of functions X S with the Wiener measure Lew. 4.18. Show that the integral of the function

    f(z) = >) ve(s)2™ (te) K=1

    over the whole space of functions X T yelative to the Wiener measure is equal to zero at odd m

    = 2r4+ 1. At evenm = 2r

    J$@)nw (ae)=(2 - MD SENOLS pil

    Préblems

    A17

    4.19. Calculate the integral relative to the Wiener measure of the function

    f(x) = 91(s)e(ti) + Ya2(s)a(t2), over the rectangle set of space XT

    v1, 2 € C(S),

    with the base

    A =

    (a1, bi) Xx (do, bs) in the

    product of the spaces Xt, x Xt, é

    4.20. Show that the following formula is valid

    [seni (da)= ef © euls)eu(s) mine} kl=1

    for the integral over the whole space X T relative to the Wiener measure function

    of the

    1

    fa

    exp{ >)pe(s)e(te) }.

    (II)

    neil

    4.21. Calculate the integral b

    /f(x)uw(de), f(2) = yp(s, t)x?(t) dt.

    (1)

    a

    In Problem 4.15 it was evaluated that the function f(x) is (Abs B)-measurable, and in Problem 4.16 was calculated the integral for the function obtained from F(a) by the substitution of the integral by the sum.

    On the basis of these results, putting

    as in Problem 4.15

    ful) =

    b—a n

    92 ols,te)2"te),

    (nt)

    k=0

    we get

    /fa(2)uw (dz) =

    h=

    n—-1

    7> (5,tate,

    (111)

    n—

    k=0

    b

    sim,ffale\uw (de) = | o(s,2) dt.

    (IV)

    It remains to prove the possibility of the limit passage under the integral sign. For

    this purpose we notice that n—1

    I fo(2) I< gn(z) = 2 > sup ols,te)la%Cte). u

    k=

    ses

    —(V)

    418

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical

    Representations

    Putting b

    a(e)=lim an(2)=fsup lo(s,t)l2%(0)at, nm—-oco

    (VI)

    a

    we obtain similarly as earlier in (III)—(VI) b

    lim

    | oa(oayw ds

    n—-oo

    [ov |p(s, t)|t dt.

    (VII)

    sESs

    a

    But in accordance with the Fatou lemma

    /9(z)Hw (de) < im /gn (x) uw (dz).

    (VIII)

    Thus the function g(«) is [Ly -integrable and the functions fn are bounded in the

    measure by //w-integrable function g(a) +cat

    some c > 0. As fr (x) =m f(z)

    almost everywhere relative to the measure [lw we get b

    /f(a)uw (de) = tim /ge 4.22.

    Using the procedure

    of Problem

    Gas fened, 4.21 prove

    x)

    that the integral of the

    function

    F(a) = fo(s,t)2""(@ at

    (1)

    over the Wiener measure [iw extended on the whole space XT is determined by b

    /f(«)uw(dz) = (2r — 1)! iv(s,t)t” dt.

    (II)

    4.23. Prove that the integral of the function b

    Fle) = exn{|o(s,)2( ar}

    (1)

    a

    over the Wiener measure /lw extended on the whole space XT? is determined by

    | #@)nw (a2)besole SGP idadeamcuaien its}. (I)

    Problems

    Solution.

    419

    Introducing the functions

    fav) = exp ict y rita} k=0

    and using the result of Problem 4.16 we find similarly as in Problem 4.21 bb

    im

    fn(x) uw (dz) = exp{||p(s, t1) p(s, t2)min(t1, tz) dty tts}.

    (IIT) It remains to prove the possibility of the limit passage under the integral sign. We notice for this purpose that according to the Fatou lemma which may be used as a

    result of nonnegativity of all the functions fie) at each s € S,

    if(a) yw(dx) = tlim fn (2) w(da) < lim fkfals)uw(de). From here and from (III) it is clear that the function f(z) is Ly -integrable and its norm is equal to b

    | (2) [I= sup exp{icon ae a

    And as the sequence SP (x)} converges to iz) almost everywhere relative to the measure [lw then

    | fa(2) [I= sup exp aes Salva] < Cll Fe) I se

    at some

    C' >

    k=0

    1. Thus the sequence of the functions {ine}

    convergent almost

    everywhere relative to 4w is bounded in the norm by {lw -integrable function. Hence

    according to the Lebesgue theorem it follows that

    J fw (ae) = jim, [fale)pr (do), and we obtain the required formula (II). The formulae obtained as a result of solution of Problems 4.22 and 4.23 and the conditions of the existence of the Bochner integral show that the functions [o.e)

    fia) = fos,t)x* (tat 0

    420

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    fo(x) = cxof fo(s.0e00 it} 0

    are {Ly -integrable if and only if the improper Riemann integrals co

    [ov lp(s, t)|t dt,

    0

    ses

    // sup |y(s, t1) p(s, t2)| min(t1, tz) dt, dte sES

    0 0

    converge. Notice that in Problem 4.16 and in Problems 4.17—4.20, 4.22 and 4.23 the conditions of the existence of the Bochner integral (the integrability of the norm of

    the functions Fle) considered as the functions of s, s € S from the space C(S)) are fulfilled. 4.24. Calculate the integrals relative to the Wiener measure //y extended over the whole space X T of the following functions:

    fila) =sin f9(s,t) (at

    ful) = cos Jo(s,t)(8) at pinpoint fos.) x(t) dt ey

    ek fos.) n(t) dt,

    fs(z)= ae

    t) a(t) dt sin fpals t) x(t) dt,

    fe(x) = exp}Joilst) a(t) dt cos Jpal(s,t) a(t) dt

    Hfal)e =-af fora ¥(2) ae near}, ve) fr (= exp y1(s,t) (x)

    yi(s

    x(t) oaths

    oit) x(t) dt,

    t) dt cosh [pa (s,t) a(t) dt,

    =i si(s,t1)...(9(8,tm)e(t1)...2(tm) dty ... dtm,

    fro(2) = ¥(2)sin fols, 8) 2( at fus(2) = W(2) cos f(5,2) 2(t) at

    Problems

    421

    fio(z) = (x) sh Ta) a(t) dt,

    fis(x) = (2) ch ae Instruction.

    a(t) dt.

    For calculating

    the integrals of fo—f13 replace in

    the formulae of Problem 4.23 and in the formulae for the integrals of the functions

    fi,..-, £4 the function p(s, t) by the function ay(s, t), differentiate the obtained formulae ™ times with respect to @ and put aw = | after that.

    4.25. Calculate the integral relative to the Wiener measure Jw

    extended over

    the whole space X T of the function

    b

    b



    .

    f(a) = ff elestas.--stm)eltr)--2(tm) dtr. te If the functions ( in all previous formulae do not depend on § then all the integrals in Problem 4.16 and in Problems 4.17—4.20, 4.23 and 4.24 represent abstract

    Lebesgue integrals. 4.26.

    Consider the spaces

    Dax

    Aa, bw).

    The measure

    extended to the space of the continuous functions C' C CAt

    Setting in the space C’ an ordinary norm || z l= sup

    space C(T).

    [Ly

    may

    be

    XT? with the o-algebra | a(t) |we get the

    Show that the o-algebra in the space C(T) induced by a set of all

    open balls enters into the 0-algebra CAT. Solution.

    For this purpose we notice that any open ball in Cr} represents

    the limit of decreasing sequence of the rectangles:

    {a(t) :| a(t)-— y(t) |< r WteCcT} = lim {2(t) : | 2( 8%.) — Ose) |


    f(x)

    of Problem

    | and whether

    {fn (@)} converges to f(z) in p-mean?

    4.23

    belongs

    the sequence

    to the space

    of the functions

    422

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    4.28.

    Prove that the value of the measure

    [yw

    on any set CeE

    Alts.co) is

    expressed by the formula

    pw (C) = pee fexrt-Zdefrolevvaldvia)

    ©

    4

    at any fixed t;.

    At all possible values t; € [0,0o) formula (I) determines the

    measure {Ly on all the sets Ce

    A

    4.29. Let (xe ; Ay) be an infinite product of,the measurable spaces

    Ae ey Aaa tEeT

    where B™

    Ae

    ee

    Ace

    ea

    ee

    teT

    is the o-algebra of Borel sets of the space R™.

    It is evident that Xe

    represents a space of all m-dimensional vector functions with the domain J’. We shall

    determine on the semialgebra of the measurable rectangles of the space (x? ; A’) the measure by formula

    [WAU xr

    1

    Grr viKi

    eX Ae is ae

    J,

    | ef exp{-e7K-*2/2} dz, dx

    X At3:t1,t2(de3} v1, oye : tN; ota hone

    OSa; Z1,---, oo

    (I)

    where ie

    Ly

    pales

    L2

    SO wise,

    K(t1, t2)

    K (to, t1)

    K (te, te)

    K(s1 ; S2) is a continuous matrix function of the dimension ™

    X mM satisfying the

    following conditions:

    (i) K(s2, 1) = K(s1, 2)"; N

    (ii) oe ul K(s;, Sj) Uj > 0 at any N, s1,..., $y and m-dimensional vec-

    t,j=1 tors Uj,...,UN;

    (iii) tro(ty, t2) < Blt, _ to|7 at some B, y > 0), where o(ti, t2) =

    K (ti, t1) =

    K (ti, tz) =

    Aig :tetoctesa (Be DiAlsisey Creag) (k bE in the correspondent

    spaces

    K (te, t1) + K (to, ta),

    n) are finite nonnegative measures

    (Xt, ; Ai, ) representing

    at each

    B



    vA rs (At,

    Kos X Ais cet ; B)-measurable functions Z1,...,£@;~—1. The measure / is 0-additive on the semialgebra C of rectangles of the space X T and may be uniquely extended on the d-algebra Ar obtained by the supplement of A? relative to LL.

    Problems

    423

    Prove that the measure /£ in (I) is finite. Instruction.

    Apply the formula

    u({a(t) : |wn(te) — ee(ti)| > n} = 17

    st 20H ORE tL, U2) 0" (ee Lyre em), where Orn(ty . t2) are diagonal elements of the matrix o(ty ; to), an inequality of (1) from Problem 4.11 and take into account that m

    Je(t) — a(t) = .]D lea(te) — we(tr)]? < D0 lee (te) — ee (ta)h; ks!

    k=1

    as a result

    {x(t) : |e(t2) — x(t1)| > ©} C LJ {2@) : |ee (ta) — ae (t1)| > €/m}, k=1

    and take @ of Problem 4.11 in the interval (0, y/2). 4.30. Prove that the integral over the measure /l of Problem 4.29 of the function

    FOLDS Grr mea abe Rist where m

    (pI

    (A, 1 ll

    Callas n) are continuous

    (1)

    matrix functions of the dimension

    X m with the values in a separable B-space extended to the whole space C’ is

    determined by formula

    pee yin) H(t ng ti):

    [|t@u(ae) Sg ait

    (II)

    bil=1

    4.31. Under the conditions of Problem 4.29 determine the measure

    Aq

    wee

    Xe xX Ax

    1 =a!

    Oo anere tt, (Olas oi,

    fatale,

    tr

    fe

    —2" K-'z/2\ dz

    at,

    (don) 015, n—1);

    (I)

    424

    Chapter 4 Stochastic Integrals, Spectral and Integral Canonical Representations

    where v1

    ps

    x2

    3

    ee

    ;

    v4

    Kitesty)

    K (ty, ta)

    K (ti, ts)

    K (ta, t1)

    K (tg, ta)

    K (te, ts)

    KGazta)

    K (to, ta)

    K (ts, t1)

    K (ts, ta)

    K (ts, tz)

    K (ts, t4)

    K (ta, ti)

    K (ta, ta)

    K (ta, tz)

    K (ta, ta)

    :

    Kk (ty ; to) is a continuous together with its derivatives till the second order inclusively

    the matrix function of the dimension ™

    X ™

    possessing properties

    (i), (ii) from

    \

    Problem 4.29 and the additional property

    tr o(ty, to, ts, t4) < Br’

    at some B > 0, y > 2,7 = max; j \t; — bale

    o(ti, to, t3,t4) = K(t1, ti) + K(to, te) + K(ts,t3) + K (ta, ta) —K(t1, te) = K (te, t1) = K (#1, ts) = K (tg, t1) + K (ti, ta) + K (t4, t1)

    +K(to,t3) + K(ts,t2) — K (to, ta) — K(ta,t2) — K(ta, ta) — K(ta, ts); the measures NG

    ae

    toon (k Soest

    n) are the same as in Problem 3.8.6. Prove

    that almost all the functions of the space X T yelative to the measure p/ are continuous and have continuous first derivatives. Instruction.

    prc certt aee (one JE

    Use formula

    Vier y

    _7T K7~*2/2} g-1 exp{—ax* dx; dry dxr3 dzr4

    |Cak —23k—Cort+@ir|>

    = 1 — 20(n/./orn(t ta,ts,r, ta)), where Ork(t1, to, t3,t4) are the diagonal elements of the matrix o(ti, ta, ts, ta), use from inequality (I) Problem 4.11, take q@ in the interval 7 qi ap

    ~~

    GENERAL

    CHAPTER 5 THEORY OF STOCHASTIC AND ITS APPLICATIONS

    SYSTEMS

    Chapter 5 is dedicated to the general theory of stochastic systems described by integral, differential, integrodifferential, difference and etc stochastic equations_in the finite- and infinite-dimensional spaces. Section 5.1 is devoted to the stochastic Ito differentials, the Stratonovich differentials and the 6-differentials. The It6 formula and its various generalizations are proved. Basic definitions concerning the stochastic differential equations, the change of variables and sufficient conditions of the existence and the uniqueness of the solutions are given in Sections 5.2 and 5.3. Practical problems connected with the transformation of the system equations into the stochastic differential equations based on the shaping filters are considered in Section 5.4. Numerical methods and algorithms for the finite-dimensional stochastic differential equations solution are presented in Section 5.5. Section 5.6 contains the general theory of the multi-dimensional distributions in the continuous stochastic systems described by the stochastic integral and the differential equations. The multi-dimensional distributions in the stochastic discrete and continuous-discrete systems are considered in Section 5.7. Structural theory of the distributions in the stochastic composed systems is given in Section 5.8. Section 5.9 is devoted to the theory of the multi-dimensional distributions in the stochastic infinite-dimensional systems. The last Section 5.10 contains applications to filtering based on the online data processing.

    5.1. Stochastic Differentials 5.1.1. Stochastic Differential of Ito Process

    A random process Z(t) is called the [t6 process if it is representable in the form t

    Z(t) = Z(to) + [ x(nar+ [owe to

    A |/Vee to Ré

    crada

    (5.1.1)

    426

    Chapter 5 General Theory of Stochastic Systems

    Here to > 0, W(t) is the Wiener process, P°(A,A) is the independent of W(t) centered Poisson measure, the first integral is the m.s. integral, the second integral is the It6 stochastic integral, the third integral is the stochastic integral with respect to the Poisson measure P(A, A), X(t), Yi(t), Yo(t, u) are random functions satisfying the conditions of the existence of the respective integrals. If Z(t) and W(t) are the vector random processes, then X(t) and Y(t, u) represent the vector random

    functions of the same dimensions as Z(t) and Y;(t) represent the matrix random function. The formal expression

    dZ = X(t)dt + ¥;(t)dW(t) + HiY(t, u)P°(dt, du).

    (5.1.2)

    Ro

    is called the stochastic differential of the Ito process Z(t). Remark.

    It should be emphasized that expression (5.1.2) has no usual sense

    of differential since it involves the infinitesimal quantities of different orders. Really, replacing the integrals by the respective integral sums,

    the increment of Z (t) ona

    small interval (t,t + At] may be expressed as

    AZ = X(t)At + ¥i(t)AW(t)4 /¥o(t,u)P%((t,t + Ad},du), (5.1.3) RG where AW

    =

    Wt

    + At) — W(t). Since t+At

    B(AW)? = B[W(t + At) - W(t)? = / Br\dr oe Olt). t

    the second item in the right-hand side of (5.1.3) has the order V At.

    In the special case where Y2(t, u) = 0 and Y(t) = Y(t) with probability 1 the Ito stochastic differential of the It6 process t

    Z(t) = Z (to)+/X(r)dt + fiY(r)dW(r)

    (5.1.4)

    to

    is given by

    dZ = Xdt+ YdW, where the argument t is omitted for brevity.

    (5.1.5)

    Bedy Stochastic Differentials

    427

    5.1.2. Ito Formula

    Theorem 5.1.1. Let y(z,t) be a scalar real function, Z(t) be an Ito process given by Eq. (5.1.4). If the scalar function y = y(z,t) is continuous and bounded together with its first and second derivatives, its second derivatives satisfy the Lipschitz condition

    Oop(2",i)" 07 ol, t) Oz, 02,

    -

    > [plz + Ye"*,t)—o(z,t)] w= 9 (s+ yor ue, 7 cs

    Sei

    —p(z + Yy7,t) + Spel(z,t)7 Yq" yp’ = o(up’) = o(h). SE

    Its probability is equal to jure. 0 @ —Hp

    =sal sefea=

    o(h » The values of V;

    corresponding to all other bad Psks are bounded and their probabilities are equal My

    5

    to I] hosts up ,m,+---+

    my,

    > 1. So we have the following estimate

    ae!

    for the expectation of lV;| :

    EWV|1

    s=1

    434

    Chapter 5 General Theory of Stochastic Systems

    My,

    Mi

    = o(h) +e(1 — exp {dont} -S ya exp {yh) = o(h)p Hence

    a

    EIS) [g(Z* + AZ}, te-1) — o(Z*, te-1) ] k=1

    N

    -Sol¢

    Py Wyte)

    = o(Z*,th-1) a Re(2"

    ten)

    Yo") uP

    k=

    N

    ,

    —S> [e(Z* + Yo, te-1)



    — o(Z*, te-1)] P™*| < No(h) = No(N7").

    K=1

    This proves together with the estimates of Subsection 5.1.2 that (5.1.17), may

    be

    rewritten as N

    Dey (20)+) (AZt + az) — (4 (to), to) k=1

    N

    X(te-1) = S"{e.(Z* + AZE, te-1) + Ge(Z* + AZE, ty_1)? k=1

    ?

    il waite [¥22(Z* +AZ5, tp 1) Yi (te 1)V (te 1) Yates 1) )T ]}h

    N +) >9.(Z* + AZ* te_1)7 ¥i(te-1) AW* k=

    N M,

    + OD Le? + YF, te-1)

    — (2? te-1) — pe(Z* tea) VE" | we

    N

    +3090

    k [9(Z" ik + VP,k te-1) — 9(Z*,te-1)] Ph t+0(h).

    (5.1.18)

    gail Sail

    Finally we notice that due to the boundedness

    of the function v(z, t) and its first

    and second derivatives and to (5.1.6) the first factors in the first two sums in (5.1.18) differ from their values at AzZé = 0 by infinitesimals o(h). Therefore AZ

    sums may be omitted giving N

    ~ (2 ag S\(AZt + az. as ~(Z(to), to) Koil

    in these

    5.1.

    Stochastic Differentials

    435

    “a

    N = So {e(Z*, te-1) + 0,(Z*,te-1)? X(te-1) neal

    +5 tr [ pz2(Z", th—1) Yi (te—-1)v(t Yie-1) (te-1)? |}h N

    +e

    ae Pali )AWF+

    M,

    SS) [ [o(Z* + YF*, th-1) Ra's

    —9(Z* ,te-1) — p2(Z*,te-1Yo" )? |we N

    + 5D [e(* + Yo te-1) — (Z*, te-1)] P+ O(A). Passing to the limit when h and maximal diameter of the sets A; tend to zero we find

    y(Z(t), t) =

    ~(Z(to), to) 43 [Molen

    a ~2(Z,, 7)" X(r)

    t

    450 PUA

    AGUA Gmttra /gs Fi 2) Vitra (2) to

    +

    up(dr, du) (Zr, tT)" Yo(r, u)|

    ,7)—P2 nn u),7)—P(4r oN Ugcee

    +

    [e(Z,

    v

    + Yo(r, u), 7) —

    9(Z,,7)]P°(dr,

    du) ,

    to 0 For proving the existence of m.s.

    sequence of the random functions for

    solution of Eqs.

    (5.2.14) we construct the

    p = 0,1,2...

    2(ar). Yp4i(t) = O(t) + falolr).rt)dr + f Yp(0),7) to

    to

    (5.2.18)

    5.2. Stochastic Differential Equations

    451

    We estimate the differences t

    Y;(t) — Yo(t) = ®(t) + ia(Yo(r), 7,t)dr + ib(Xo(r), 7, t)Z(dr),

    (5.2.19)

    Yosi(t) —Yo() =O) + fLalp(r),78) ~ a(¥%-1(7), 7,1)]ar t

    7 /[b(Yo(r), 758) — B(%p_a(r)y7,t)]Z(dr).

    (5.2.20)

    to

    As

    a(Yo0(r), 7, t) is m.s.

    continuous

    ||a(Yo(r), 7, t) || We shall cite the proofin connection with the random functions with the values

    in B-space Ly (Q, Pople i) of n-dimensional vector functions of the variable W. As it is known from functional analysis the norm of the value of the random function

    Alt) = z(t,w) at each ¢ is determined by formula

    IXOIE = fle o)IPPde) = BIXEOP.

    6.39)

    In other words, the square of the norm \|x (t)]]? of the value X (t) at a given t in B-space D2(Q, On, F*) represents an expectation of the square of its Euclidean

    norm ||X (t)||. Then Eqs. (5.3.1), (5.3.3) and (5.3.7) may be considered as the equations in B-space D2(Q, ny al Be) and condition (5.3.8) will be written in the form of the Lipchitz condition

    |Fi(¥i(7), 7,t) — F(Y(r),7,t)|| < k|l¥i(7) -— Y(r)II.

    (5.3.10)

    Therefore for the completion of the proof it is sufficient to use the known from the theory of the ordinary differential equations the Picard method

    of the successive

    approximations. We

    determine

    the

    sequence

    of m.s.

    continuous

    random

    functions

    for p

    a Ube Py pee t

    Xo(t) = (t), Xp4i(t) = O(2) + fP%s(r), ner

    (5.3.11)

    to

    and estimate the norms of the differences t

    :

    X(t) — Xo(t) = J For), ntar,

    (5.3.12)

    to

    Xp4i(t) — Xp(t) = [[email protected]

    — F(Xp-1(7), 7, t) |dr.~

    (5.3.13)

    As according to the condition the random function F(Xo(r), Py t) of the variables 7, t is m.s. order moment

    continuous in consequence of m.s. I|F(Xo(7), T,t)||

    2

    =

    continuity of O(r ) then its second 2

    E

    \|F(Xo(7), T5 t)|| is continuous and finite

    at all T, t, and consequently, achieves its upper bound on the finite closed interval

    458

    Chapter 5 General Theory of Stochastic Systems

    r € to, t],t€ [to,to + T]. Therefore ||F(Xo(r),7;t)||’ < ¢ at some ¢ < 00. Hence from (5.3.12) we find

    1X1 (t) — Xo(t)|) < [ieco), Pare tie

    Oat

    After that from (5.3.13) and (5.3.10) by the induction we get

    Koes(0)ae—Xp(OI
    F(X((r), 7,4), LF (Xp(7), 7 le < NF (Xo(7), 7, lle

    + TIF (Xq(7), 7,1) — P(Xq-1(7),7,8)I] < cet), Thus

    at all 7 the sequence

    integrable function of

    F(X(r), a; #) is bounded

    (6.3.17)

    on the norm

    by the

    7. Consequently, on the basis of Lebesgue theorem which is

    known from functional analysis the function F’ (X (r Vs Tas t) is integrable over T and

    we may pass in (5.3.11) to the limit at p —

    OO under the sign of the integral.

    It

    means that the limit function X(t) satisfies Eq. (5.3.7) what proves the existence of the m.s. solution of Eq. (5.3.7).
    Let Y (t) be any solution of Eq. (5.3.7). After substituting it into (5.3.7) and subtracting it from (5.3.11) we obtain t

    bead ee ‘iF(Y(r),7,t)dr, to

    Aprit)—Y

    y= /[F(Xp(r),7,t) — F(Y((7), 7, t)] dr,

    5.4. Transformation of System Equations

    _

    into Stochastic Equations

    459

    _kPt1(t — to)P+?

    |Xpai(t) — Y(t)|| = iene

    tb

    Hence passing to the limit at p — 00 we shall have ||X (t) _ Y (t)|| = 0. It proves the uniqueness of the m.s.

    solution of Eq. (5.3.7), and consequently, also of Eq. (5.3.1) (with the accuracy till the equivalence). to. The system which is described by Eq. (5.4.6) for the normally distributed white noise V = V(t) with zero expectation EV = 0 and with the matrix of the intensities vy= v(t) we shall call the normal stochastic differential system. For some composite stochastic differential systems in the finitedimensional space the stochastic differential équatin of the form is used

    dY = a(Y,t)dt + b(Y,t)dWo + :c(Y,t, u)P(t, du).

    (5.4.7)

    Ro Here a = a(y,t) and b = b(y,t) are the known (p x 1)-dimensional and (p x m)-dimensional functions of the vector Y and time t; Wo = Wo(t) is a m-dimensional Wiener random process of the intensity vp = vo(t);

    c(y,t, u) is (px 1)-dimensional function of y, ¢ and of the auxiliary (q x 1)dimensional parameter u; [ dP°(t, A) is a centered Poisson measure: A

    parte) = fare) A

    A

    [ret aya,

    (5.4.8)

    A

    where { dP(t, A) is a number of the jumps of a Poisson process at time A

    interval A; vp(t, A) is the intensity of the Poisson process P(t, A); A is some Borel set of the space Rf with the pricked origin of the coordinates. The integral in Eq. (5.4.7) is extended on the whole space R{ with the pricked origin of the coordinates. The initial value Yo of the vector Y represents a random variable which does not depend on the increments of Wiener process Wo(t) and of the Poisson process P(t, A) at time intervals A = (t1,t2] which follows to, to < t; < tg for any set A. Remark. In the case when the integrand c(y, Us u) in Eq. (5.4.7) admits the presentation c(y,t,u) = b(y, t)c’(u) Eq. (5.4.7) is reduced to the form of (5.4.6) if we assume

    W(t) = Wo(t) + fewrre. du). RG

    5.5. Numerical Integration-of Stochastic Differential Equation

    463

    In some practical cases to stochastic differential Eq. (5.4.7) may be reduced the stochastic integrodifferential equations of the form of Eq. (5.4.9)

    dY, = a(Yf., t)dt + B(Y#, t)dWo + /o(¥i,t,u)P%t,du).

    (5.4.9)

    Ro Here a = a(yj,,t), 6 = b(y},,t), ¢ = c(yj,,t,u) at each instant depend not only on the current value of the process Y(t) but on all its values Y(r) at the interval [to,t), i.e. the components of the vector functions a, c and the elements of the matrix 6 are the functionals of the process

    Y(r), 7 € [to, t). In problems of practice for calculating the probabilities of the events connected with the random functions it is sufficient to know the multi-

    dimensional distributions (Section 2.7). Therefore the central problem of the theory of the continuous stochastic systems (Subsection 1.6.5) is the problem of probability analysis of the multi-dimensional distributions of the processes which satisfy the stochastic differen-

    tial equations of the form Eq. (5.4.6) or Eq. (5.4.7) at the correspondent initial conditions. In the theory of continuous stochastic systems two principally different approaches to distributions calculation are recognized. The first general approach is based on the statistical simulation, 1.e. on the direct

    numerical solution of stochastic differential Eqs.

    (5.4.6), (5.4.7) with

    the corresponding statistical data processing (Section 5.5). The second general approach is based on the theory of continuous Markov processes and supposes the analytical simulation, i.e. the solution of equations in the functional spaces for the one-dimensional and the multi-dimensional

    distributions (Section 5.6). Along with the general methods of the stochastic systems theory the special methods are distinguished which are oriented on the linear stochastic systems (Chapter 6) and on the nonlinear stochastic systems

    (Chapter 7). 5.5. Numerical Integration of Stochastic Differential Equation 5.5.1. Introductory Remarks Integration of stochastic differential equations has some peculiarities. The fact is that all numerical integration methods of the ordinary differential equations besides

    464

    Chapter 5 General Theory of Stochastic Systems

    the simplest Euler method

    are based on the calculation of the increments

    sought functions on each step by means mean value.

    of the

    of using the integral theorem about the

    In the accordance with this fact the right-hand sides of the equations

    (the derivatives of the sought functions) are taken in the mean points of the intervals. Various methods of the numerical integration differ in essence from one another only by the way of approximate finding the mean

    values of the right-hand sides of the

    equations.

    The theorem about the mean value is not applicable to the stochastic integrals. But some

    analog of the theorem about the mean

    integrals with respect to the nonrandom functions.

    value is valid for the stochastic

    It consists in the fact that the

    best approximation of the stochastic integral with respect to the continuous nonran-

    dom function gives the product of the value of a given function by the increment of the process over which the integration is performed.

    Therefore, for instance, all the

    methods of the numerical integration of the ordinary differential equations may be

    used for the stochastic differential equations (5.4.6) whose coefficient at the white

    noise is a determinate time functions f, i.e. b(Y, t) = b(t). If the function b(Y, t) in Eq. (5.4.6) depends on Y then the numerical integration method of such equations is required to be chosen depending on the fact in what sense the stochastic integral is assumed.

    In the case of the It6 equation the increment of the process Y(t)

    at the interval [t,¢ + At) should be determined by formula AY

    =

    a(Y;,t)At

    = b(Y%, t) AW what corresponds to the Euler method of the numerical integration. Thus the It6 equations may be integrated by Euler method.

    But for better accuracy

    of the calculation of the first item we may take the value of the function a(Y;, t) in

    some mean point of the integration interval Lt. t+ At). Here we may use any method of the numerical integration, for instance, the widely known Adams method and the

    Runge-Kutta method. But the value of the function b(Y; ; t) should be always taken at the initial point of the integration interval.

    In the case of the equation in the

    0-differential the value of this equation should be taken at the point t + OAt. But for its approximate finding we have to elaborate new methods of the numerical integration.

    While simulating the stochastic differential equations by means of the analog computers for the functions bY, t) which depends on Y it is necessary beforehand to

    reduce the It6 equations to the symmetrized (Stratonovich) form. It may be explained by the fact that at such simulating a white noise should be substituted by a process with a small correlation interval but such one which is different from zero as a white noise is physically nonrealizable.

    In this case a simulating process will be close to the

    solution of a stochastic differential equation if and only if a given equation is assumed

    as a symmetrized stochastic differential equation.

    While simulating the processes

    which are determined by Eq. (5.4.6) it is necessary to simulate the random variables or random processes. At numerical integration of the stochastic differential equations

    5.5. Numerical Integration*of Stochastic Differential Equation

    on digital computers it is required to simulate the increment AW

    465

    of the random

    process W(t) at each step. While simulating by means of the analog computer it is

    necessary to generate the wide-band random processes.

    5.5.2. Existence of Exact Difference Equation Stochastic Differential Equation

    Correspondent to Given

    Let us consider the It6 stochastic differential equation (5.4.7). Suppose that it is required to substitute this equation by a difference equation for the values of the

    process Y(t) in a given discrete series of the equidistant points {tk i} ty = kh where h is an interval between adjacent points t ry

    a

    3k+1 —tx. In principal the problem

    is solved exactly as the values of Markov process Y(t) at the points t, form Markov

    random sequence {Y;}, Yi, = Y (kh), and any Markov sequence is determined by some stochastic difference equation. by means

    But to compose this exact difference equation

    of a given differential equation is practically impossible.

    First of all for

    its composing it is necessary to find a transient distribution of Markov process Y(t) determined by Eq.

    (5.4.7) and then we may compose a difference equation for the

    sequence {Yn} by means of the found transient distribution. But the exact evaluation of the transient distribution of the process Y (t) is possible only in some specials cases. In the general case we are satisfied with an approximate evaluation of a transient

    distribution of the process (2):

    As a result by this transient distribution we may

    obtain only an approximate difference equation.

    The latter makes it expedient the

    application of very complicated algorithm derivation of an exact difference equation for the sequence

    ‘nee

    As we

    are satisfied only with an approximate

    difference

    equation even at the application of a given algorithm then it is expedient to make it by more simple ways.

    At first we substitute an integral over the variable U in Eq. correspondent integral sum.

    (5.4.7) by the

    As a result Eq. (5.4.7) will be replaced by the following: N

    dY = a(Y,t)dt + 0(Y,t)dWo + 5 c(Y¥,t)dP?,

    (5.5.1)

    cil

    where Celyy; t) are the p-dimensional vector functions which represent the values of the function cy, Ue u) in some of the partition of g-dimensional et

    Pas ae gs My and P(t)

    are

    mean

    point U; correspondent

    to the elements

    sphere of sufficiently large radius, the

    centered

    simple

    Poisson

    Uj;

    processes



    A;

    Aj (2

    (Subsec-

    tion 4.6.6):

    Poy = Po(0/4) Ai) — w([0,4), Al), 7 = 1,2, 2a.

    (5.5.2)

    The intensities of these processes are determined in terms of the expectation p(A, A)

    of Poisson measure P (A, A) by formula

    y(t) = du ([0,t), Ai)/dt.

    (5.5.3)

    466

    Chapter 5 General Theory of Stochastic Systems

    The replacement of Eq. (5.5.1) by the correspondent difference equation is performed by the Euler method.

    5.5.3.

    Euler Method

    The

    simplest

    way

    of

    the

    replacement

    rence equation consists in the replacement

    of

    Eq.

    (5.5.1)

    by

    a

    diffe

    of all the differentials by the elements

    of the integral sums:

    Y((n + 1h) — Y(nh) = a(Y (nh), nh)h + 1(Y (nh, nh))[ Wo((n + 1)h) N

    —Wo(nh)] +) ei(¥(mh), nh)[ P2((n + 1)h) ~ P2(nh)]. Putting

    Yn =Y(nh),

    Yn(Yn)=Y(nh) + a(Y(nh), nh), vin(Yn)=b(Y (nh), nh),

    Win(Ya) = c-1(¥ (nh), nh), Vin = Wo((n + 1h) —Wo(nh), Vie iP, 1

    (5.5.4)

    + 1h) Fe(Wh), tS 2 Nee

    we get a stochastic difference equation

    Viutd =

    ON

    OE) a oy Wil Valin: t=]

    Introducing the p X (m oe N) block matrix

    Yn(Fa) = [Yin(Fa)---Pavstyn(Fa)]

    (5.5.5)

    and (m +N )-dimensional random vector

    Vi, Sieve

    reyes

    (5.5.6)

    we may write the obtained difference equation in short form:

    Yn¢i = Gn(Yn) + Ya(¥n)Vai

    (5.5.7)

    As the Wiener and the Poisson processes are the processes with the independent increments (Subsection 4.6.3) then the random vectors V,, form the sequence of the

    5.5. Numerical Integration*6f Stochastic Differential Equation

    467

    independent random vectors {Vn } and the blocks Vj, of the vectors V;, have the

    normal distribution V (0, Gis) where (n+1)h

    oF

    il valde = aalhh:

    (5.5.8)

    nh

    Vo(t) is the intensity

    of the Wiener

    process

    Wo(t),

    the scalar blocks

    Vj,

    (7

    = 2,...,.N +1) have the Poisson distributions with the parameters (n+1)h

    PS

    d V;(7)dr = 4(nh)h.

    (5.5.9)

    nh

    These distributions determine completely the distributions of the random vectors V,,.

    The covariance matrix Gy, of the vector V, represents a block-diagonal

    matrix

    +

    Gael

    TRANG

    OP

    abe

    0

    DE tia

    MOO )esE (2 0s

    0

    0

    0

    =

    (4)

    (5.5.10)

    HN+1,n

    Eq. (5.5.7) determines Herr at a given ¥:, with the accuracy till fh in the determinate item Yn, (Yn) and with the accuracy till Vh in the random item Vn (Yolen

    Remark.

    The stated method of the substitution of the stochastic differential

    equation by a difference one in essence does not differ from the Euler method of the

    numerical integration of the ordinary differential equations.

    5.5.4. More Exact Equations For deriving more accurate difference equations than Eq. (5.5.7) we substitute (5.5.1) by the correspondent integral equation

    (n+1)h

    (n+1)h

    A¥n= fave rdr+ f W(¥%;,r)aWar) nh

    nh

    N

    (n+1)h

    +> felts naredr). SUE

    is

    (6.5.11)

    468

    Chapter 5 General Theory of Stochastic Systems

    For the purpose of the approximate calculation of the integrals we determine Y, by means of the linear interpolation of the random function Y(t) at the interval

    (nh, (n ap 1)h). Then setting as usual AY, = Y((n i 1)h) ae Y (nh) we have

    a(¥,7) = a(Ya +

    r—nh

    A¥»,1), (V7)

    = (Fa+

    7T—nh

    A¥a,7),

    (5.5.12) e:(¥;, 7) 2c; (¥.+

    7 —nh

    scaly

    eT Ne

    For calculating the right-hand sides in these formulae we use the generalized It6

    formula (5.1.24). So we have

    o(Y + dY,t + dt) = 9(Y,t) + {¢:(Y,t) + oy (Y, t)7a(Y, t) N

    +e, t):o0(Y,t) + Sey jell

    + ¢;(Y,t),t) — 9[Y,t)

    —py(Y,t)” ci(Y,t)]us(t) }dt + vy(Y, t)7O(X, t)dWo N

    +S “[e(¥ + c(¥,t), t) — 9(¥, t)]dP?,

    (5.5.13)

    where ;(y,t) is a partial derivative of the function y(y, t) over time ¢; Py (y, t) is a matrix whose rows represent the partial derivatives of the matrix-row ply, he with respect to the components of the vector Y; Pyy (y, t) 5 o(y, t) is a vector whose components are the traces of the products of the matrices of the second derivatives

    of the correspondent components of the vector function ply, t) over the components

    of the vector y on the matrix o(y, t):

    [Yyy(y,t) : o(y, t)le= trl Geyy(y, t)o(y, t)], o(y,t) = b(y, t)vo(t)b(y, t)’.

    (5.5.14) (5.5.15)

    It is convenient to transform formula (5.5.13) in such a way that namely the Poisson processes P; (t) would enter into it directly, not the centered processes P(t) :

    Taking into account that

    P(t) = P,(t) - /Ore

    (5.5.16)

    5.5. Numerical Integration’ of Stochastic Differential Equation

    469

    we may rewrite formula (5.5.13) in the form

    oY + dY,t + dt) = o(¥,t) + (oi(¥,t) + oy (Vt) al, 2) N

    —ral¥ deal] + set) so, t)}ae+ gy(¥." y

    N

    x(Y,t)dWo + So[e(¥ + ei(¥,t),t)— oY, t)aPi. For extending formulae is necessary

    to modify

    (5.5.17)

    (5.5.13) and (5.5.17) over the matrix functions

    the notation of some.its members

    in such a way

    it

    that the

    expressions Dy (y, t) in the context of the matrices algebra will have also the sense for the matrix fucntion Y. Taking into account that in the case of the scalar or the

    vector function

    for

    any

    p-dimensional

    vector

    U

    we

    may

    substitute

    the

    items

    y es an

    x a(Y,t), Gy(¥,t)" W(Y,t)dWo, vy(Y,t) c:(¥,t) by the items a(Y,t)”

    x (8/dy)? (Y,t), WEY, t)(O/dy)e(Y,t), c«(¥,t)" (8/dy)e(¥, t) correspondingly.

    Therefore formulae

    (5.5.13) and (5.5.17) will be also valid for the

    matrix functions Y as ul (d /Oy) represents a scalar differential operator for any p-dimensional vector U and its application to the vector or the matrix function 9

    means

    its application to all components

    of the vector Y or to all elements of the

    matrix (9. The variable Pyy (Y,t) : o(Y,t) in the case of the matrix function ~ represents a matrix whose elements are the traces of the products by the matrix 7 (y, t) of the matrices of the second derivatives of the correspondent elements of the matrix

    ply, t) with respect to the components of the vector y:

    [Yuy(y,t) : o(y,t)Jer= tl Peryy(y, t)o(y, t)].

    (5.5.18)

    Using formula (5.5.17) and its modification suitable for the matrix function ~

    and accounting also (5.5.15) and the fact that according to (5.4.7) and (5.5.16)

    7T—nh

    4

    AYn =

    N

    +0 t=1

    T—nh

    tT—nh

    5

    =

    ays, nh)h +

    |=

    T—mnh

    b(Yn, nh) AW,,

    —¢i(¥n nh)(APin — Vinh),

    (5.5.19)

    470

    Chapter 5 General Theory of Stochastic Systems

    (where as earlier MWS

    Wo((n + 1)h) —

    Wo(nh), NP

    Se

    P;((n + 1)h)

    — P; (nh)) we shall have with the accuracy till infinitesimal of higher order relatively

    to h the following equalities:

    rT—nh

    x

    a—

    ¥

    1

    > evin

    T—nh:>

    A¥,,7] =a+a;(7 — nh)+

    (r—nhy\?

    his

    )

    h

    (dyy : o)h+

    T—nh h

    ay LAW,

    N

    +0 E ula sees nh)= a|A Pie 7—nh

    AY] = b+b;(7 — nh) +

    N

    res

    Dy. [(¥,ES me es mh)-

    =)

    |

    Me

    &

    S

    =}

    3

    Dol] Re

    ole

    _

    =a +

    N

    aes

    Ne

    ad) (Ciyy

    7—nh

    h

    AIPEY Ano 21)

    C; (%+ dee Mt AY) = co + ci¢(7 — nh)+

    x

    (5.5.20)

    —nh wis

    h

    “iy

    — nh :o)h+ of

    bAWn

    oN

    7—nh

    F

    cj,nh

    = «|Py:

    (5.5.22)

    Here the arguments Vis nh at all functions which depend on are omitted them for brevity.

    For

    increasing

    the

    accuracy

    of the

    calculations,

    in particular,

    variance matrix of a normally distributed random vector AW,

    [T Wo(nh)and

    the parameters

    of Poisson

    distributions

    =

    of the

    co-

    Wo((n + 1)h)

    of the random

    variables

    [IPR = 12k ((n + 1)h)— P;(nh) we may take the values of the intensities 9 (t) of

    5.5. Numerical Integrationof Stochastic Differential Equation

    the process Wo (t) in formula (5.5.15) and

    471

    (t) of Poisson streams which generate

    the processes P; (t) at the mean point nh + h/2 at the interval (nh, (n + 1)h):

    o(Yn, nh) = b(Yn, nh) (mn+ ;)ee nh)",

    h vin=

    (5.5.23)

    (mh+ 3), 2

    1

    ae.

    To facilitate the further calculations we find the increments of the functions a, b, c; in the sums of equation (5.5.19) by means of the linear interpolation at small

    interval (nh, (n = 1)h): -%

    a(%,+e E ses mh)—a& 7T—nh

    =

    h

    Nba

    =

    4 =

    Ae

    i(%+ E ses, mh)—b

    T—nh

    Ce (% + 7anth)

    7—nh

    —¢

    rn

    = —,Aitins

    |

    (5.5.24)

    Asan = a(Y, + ;,nh) —a(Yn, nh), Aibn = 0(Y¥, + cx, nh) — b(Yn, nh), AjCin = ci(Yn + c;, nh) — 7 Ome nh).

    (5.5.25)

    Using formulae (5.5.20)—(5.5.24) we find the following approximate expressions

    of the integrals in (5.5.11) (n+1)h

    1

    a(Y,,7)dr = ah+ 5

    1

    N

    + 34yy | h?

    a, + ay (2—Soamm

    nh

    i=1 1

    N

    Te es (x bAW,, + Aiea]: h,

    (n+1)h

    (5.5.26)

    N

    b(Y,,7)dWo(r) = DAW, +

    3

    by + G - yi] “a h

    “ih

    iil

    N

    (n+1)h

    h

    +AweiT os+ atara| i a" dWo(r) Y

    t=1

    ph

    (n+1)h

    +5(byy :o)

    h

    eA nh

    :

    9

    ) dWo(r),

    (5.5.27)

    472

    Chapter 5 General Theory of Stochastic Systems

    (n+1)h

    N

    c:(Y¥;,7)dP?(r) = G¢APin +

    Cit + chy a

    Sm

    ws

    | A

    jal

    (n+1)h

    N

    +eFiy DAWn n + D> Ajcin 7 Cin AP; jn

    BED h pes i nh

    gal

    ,

    (n+1)h

    h

    +5(ciyy 12)

    2

    Se: ) dP9(r),

    (5.5.28)

    nh

    .

    (n+1)h

    N

    1

    ci(Y,,7)u%i(r)dr = {ns +15 |. + chy (= pI “| Ha

    nh

    1

    1

    +3 Ciyy : | +

    3

    -

    cybAWn, + ye

    ee

    hi

    Vin.

    (5.5.29)

    j=l Further on substituting the obtained approximate expressions of the integrals

    in to (5.5.11) and taking into considerations that AY, = Y((n + 1)h) — Y;(nh) 0)

    fra | Y,;, we come to the difference equation N+1

    Fat =Pn(¥n) + Yo [Pin(Fn)Vin + Vin(Yas Vin Vin + Vin (Yn) Vin| fie

    (5.5.30)

    Here the following denotions are introduced:

    —_

    _—

    =—

    Onl Yn) = Yon |an, Bh)

    WY

    =_—

    deci (ins nh)vjn | h+ j=l

    N = Ss Tis {n(nh) — >pate nh)Vjn + tin nh)? 1

    S,

    j=

    =

    N girl Yn, hn)? Yen

    N a(Yn, nh) —S\og (Yn, hn)vjnt |

    ial

    Geil

    N Y AG: dgg (Vinh)

    tery y, apgh us j=l

    y ts}

    2

    (5.5.31)

    5.5. Numerical Integration”of Stochastic Differential Equation

    473

    N

    tin(Yn) = (Zn nh) + 5 dy(Yna,nh)” — S° ej (Yn, hn)? vn |b(Yn, nh), j=1

    (5.5.32) N Vin(Yn) = Sa

    nh) + | A;_ia,

    — ST Si-cintn| h,

    (5.5.33)

    pict

    Pin(¥n Va) = {, nh) + |a(¥n,nh)* N is fag = wy cj(Yn, hn)? vjn°|—b(Yn, nh) = Oy iil 0 N +V,L b(Yn, nh)? —b(Yn, nh} h+ Sa AjbaVi41n, oy

    Y.\

    ns V0)

    =

    (5.5.34)

    gs

    nh)?

    {eta nh) ae Goren

    »

    --

    a(Yn, nh) — beg GY a ini]

    h

    aol

    N

    +c;-1,y(Yn, nh)" (Ya, nh)Vin + >, Ajej-inVj4i,0;

    (5.5.35)

    j=l

    : eghee

    C

    1

    1 A glbuy (Yn Rh)

    o( Yn, Teh,

    *

    4

    Sea Nabi glci-1.yy(Yn, nh) Sai

    Ynett):

    Vin = AW, = Wo((n + 1)h) — Wo(nh),

    Vin = AP;-1n = Pi-1((n +1)h) —Py-a(nh),

    nh

    nh

    i=2,...,N,

    (5.5.36) (5.5.37) (5.5.38)

    (6.5.39)

    474

    Chapter 5 General Theory of Stochastic Systems

    (n+1)h

    Vi es ii SRS ins: Fe

    (5.5.41)

    nh

    (n+1)h

    9

    T—nh a

    dP?_,(r);

    Introducing the vector V, vi Me=

    (Ke Venmae

    /

    Vv, =

    (5.5.42)

    Ie

    C= 2g,

    nh

    Wii

    Et the block matrix

    Yn (Yn VE?) = [din(Yn) Pin(Yns VA?) Yin(Yn) : .wnsi,n(Yn)

    VNeialin)

    VD)

    Vrann

    Ya) |

    and the block random vector

    Te

    Vn =

    a

    [Vin Vin

    Vin

    AP ew HGOE iy BUTE

    Von Von

    Von

    793

    /

    "

    a8

    » Vn-+1,n Vu-+ijn Vivr41,n | )

    (5.5.43)

    we may write Eq. (5.5.30) in short form

    Vota

    olin)

    The variables which enter into Eqs.

    Un Yas VO Va.

    (5.5.44)

    (5.5.44) are determined by formulae (5.5.31)—

    (5.5.43). 6.8.8. Distribution of Random

    Vector Vp,

    We find a distribution of the random vector V,,. It is clear that the expectations of the pas

    tor Rasa

    variables Vin Vie are equal to zero and that the scalar vec-

    Vin

    T

    Pe

    Ae?

    f

    :

    ] has the normal distribution, and the scalar variables Von,

    : Hae have the Poisson distributions with the parameters (n+1)h

    a=

    i Ver

    GT = ea (nn

    5m)[et

    2 oe

    INE de

    nh

    It is also evident VN4in

    7

    ; VN41,n

    (5.5.45) that the random Tr,

    Wee ae

    Vi

    ee. aii

    5

    [ are independent by virtue of the independence of the pro-

    cesses Wo (t), P, (t), a pendent.

    vectors

    a Py(t) and that at different n the variables V,, are inde-

    But at any given 2, n the vectors V;,, Vee in) Vin are dependent.

    5.5. Numerical Integrationof Stochastic Differential Equation

    475

    For complete definition of a distribution of the random vector V,, in Eq. (5.5.44) it is sufficient to find the covariance matrix of the normally distributed random vec-

    tor [V2 Vin a yee ize Using the known formulae for the covariance and crosscovariance matrices of the stochastic integrals we find the blocks of the covariance matrix Ky, of the random vector [Vv ye

    Ve 13

    (n+1)h

    Kinit = BVia Ve =

    / Vo(r)dr = vo(nh + h/2)h, nh

    (n+1)h 7

    Finio = Va

    T—nh

    pio

    A Vo(t)dr = svo(nh +h/2)h, ———

    nh (n+1)h

    Ree Eve

    ‘i

    / (=") vo(r)dr & sv0(nh +h/2)h nh Kan

    gv= Kingoy

    Kigist= Kan, 135

    (n+1)h Kano = EV inVin ees

    :

    / (=")

    Vo(t)dt = Zvo(nh +h/2)h,

    nh Kinj31 = Kinz,

    Kin,32 = Kin,23;

    (n+1)h

    =n) vo(r)dr © =vo(nh 1 / (=") + h/2)h

    Ki,3s2 BYE

    nh

    (5.5.46)

    In practice it is expedient to approximate the stochastic integrals Vie Vn in

    (5.5.41), (5.5.42) (¢ = 2,..., N+1) by means of the following analog of the integral theorem about the mean for the stochastic integrals: (n+1)h Rh Vin =

    /

    a

    ts

    adh, h

    ‘ 0 LT

    ~w

    nh

    (24)

    V;,~1(T) dt

    n (n+1)h



    %-1(r)dr

    nh

    1 XA Piha 8 5APi-1n

    = 51 Vin,

    (5.5.47)

    476

    Chapter 5 General Theory of Stochastic Systems

    (n+1)h (n+1)h

    7]

    S

    Vin

    (

    i

    ‘i

    er

    0 h

    ~

    ) dP;_,(7)

    y,_4(r)dr

    (th)?

    n (n+1)h

    =.

    ui yj,-1(7)dr

    nh

    1 RAR dn A

    1 sem

    ore EL

    (5.5.48)

    5.5.6. Analysis of More Precise Equations At the statistical simulation of a system by means of Eq. (5.5.44) it is also not

    difficult to simulate the random variables distributed according to the normal and

    the Poisson laws.

    It is easy to see that the right-hand side of difference equation

    (5.5.44) is determined with the accuracy till h? in the determinate (at a given vay item Yn Oy) and with the accuracy till h3/2 in the random item Pal Ya VP) Vp. While deriving Eq. (5.5.44) two small errors took place. Firstly, while replacing ye by the variable ys + (r = nh)AY, /h we substituted the random functions b(¥;; T) and Cit)

    independent

    of dWo(r)

    and dP?(r)

    functions depending on the random parameter AY,

    by the nonrandom

    which depends on the values

    dW (rT) and d POG) at the interval (nh, (n + 1)h). Secondly, at ci(y,t) ze 0 even at one 2 the realizations of the random process Y(t) have the discontinuities of the first kind in the random points despite of m.s.

    continuity of rte): Therefore

    strictly speaking we cannot perform the linear interpolation of a given process. The first error may be removed by two ways.

    The first one consists in the replacement

    of the interpolation of the process Wah) by its extrapolation what is equivalent to

    the substitution of AY, by the variable AY,,_1 in the obtained expression for Y;. But this way will lead to the appearance of the variables Youn and Age in the right-hand side of a difference equation,

    i.e.

    to the substitution

    of the first order by the difference equation of the second order.

    of the equation

    The second way

    consists in the abandonment of the interpolation of the process Y (t) at the interval

    and in the direct expression of the increments of the functions a(Y;,, uF), b(Y,, iT),

    G (Wen T) at small interval (nh, (n =e 1)h) according to the generalized Ité formula with the substitution the differentials by the increments. At this method the second error is also removed.

    But the difference equation obtained in such a way will be

    more complicated. The random variables which represent two-fold integrals over the components of the Wiener process W(t) =

    Wo (t) and over the Poisson processes

    will be included into this equation:

    (n+1)h +

    / /dW;(o)aW;(r),

    nh

    nh

    (n41)h 7

    / hsdP;(0)dP9(r),

    nh

    nh

    5.5. Numerical Integration of Stochastic Differential Equation (n+1)h

    f.

    (n+1)h

    / /dP? (o)dW;(r), nh

    7

    u /dP?(c)dW,(r).

    nh

    nh

    477

    (5.5.49)

    nh

    It is very difficult to find the distributions of these random variables and only the

    first two of them are easily calculated at 7 = 12:

    (n+1)h +

    famcoyarn(ny= Abin =val +b/ayh ,

    2

    a

    *

    :

    nh

    nh

    (n+1)h

    nh

    ,

    7

    /BEN a cay ree

    (5.5.50)

    nh

    As regards to the second error it cannot essentially influence on the result as the probability of the occurrence of a jump of the Poisson process on sufficiently small

    interval (nh, (n + 1)A) is infinitesimal. The accuracy of the approximation of a stochastic differential equation by a difference one may

    be increased later on.

    In particular, in one of the ways it is

    sufficient to express a(¥ry Fy b(Y,, im) ey

    T) at the interval (nh, (n + 1)h)

    by the It6 integral formula: T

    aly, , n= a(Faynh) + fay(¥s,8) + a2(Ye,3)" a(Y;,5) nh

    N

    1

    di

    — Soci(¥o, w(s)) + 5ayy (Yor) : 01%, 8)ds+ fay(¥.,3)" iz}

    nh N

    cit

    xb(Ye,8)dWals)-+ ¥>[La(¥s+ex(¥e,8), 8) -a(¥s,8) ]4PM(). (5.5.51) dag

    Formulae for b(Y;, T), ron eas T) are analogous.

    To the integrals obtained we may

    then to apply the same method which was used for the calculation of the integral in (5.5.11).

    As a result we shall obtain the right-hand side of the difference equation

    with the accuracy till h® in the determinate (at a given 5) item and h®/? in the random item.

    The process of the more precise definition of the difference equation

    correspondent to a given stochastic differential equation may be continued. But each new correction requires the existence of the derivatives of the functions a, b, c; of more higher orders.

    478

    Chapter 5 General Theory of Stochastic Systems

    We may use another method for precising the difference equation. Namely, the

    integrands in (5.5.51) and in analogous formulae for b(Y;, Fy) GLYs; T) we may express by means

    of the generalized Ité formula substituting in it the differentials

    by the increments.

    Here the three-fold integrals over the components of the Wiener

    process Wo (t) and the Poisson processes P; (t) will enter into the difference equation. For further correction of the approximation the stochastic differential equation by the

    difference one in this case the integrands in (5.5.51) and in the correspondent formulae for b(Y, . T) Ci ee 5 T) in its turn should be presented by the integral It6 formula and

    later on may be used the differential It6 formula with the change of the differentials by the increments.

    This process may

    be continued further and as a result it will

    lead to the presentation of the process Y (t) at the interval (nh, (n + 1)h) by the stochastic analogs of Taylor formula.

    In this case the multiple stochastic integrals

    over the Wiener process Wo (t) and over the Poisson processes P; (t) will enter into the difference equation. insuperable difficulty. same

    component

    The finding of a distribution of these integrals represents And only the integrals of any multiplicity over one and the

    of the Wiener process

    Wo (t) or over one and the same

    Poisson

    process P; (t) are calculated very simply.

    In order to avoid the calculations of the derivatives of the functions aes fh), Oe

    a) G; (Ya ; T) while using two stated ways of the approximation of a stochastic

    differential equation by a difference one, we may recommend

    to substitute them by

    the retios of the finite increments, for instance, at the interval (nh, (n + 1)h) over

    the variable tf and at the intervals (Yi ; Ynk ++ Gn

    Yee nh)h) over the components

    of the vector y.

    5.8.7. Strong and Weak Approximations rential Equations

    of Nonlinear Stochastic Diffe-

    The obtained in Subsections 5.5.2—5.5.6 difference equations may be used both

    for the theoretical research and for the numerical integration of the stochastic differential equations. In both cases we must know the distribution of all random variables which enter into the difference equations. In a given case the difference equations will represent the so-called a strong approximation

    of the stochastic differential equations.

    While numerical integrating such an approximation is necessary if is required to obtain the realizations of the process NACE

    But often there is no need to obtain the

    realizations of the process but it is sufficient only to have the estimates of the moments or of the expectations of some functions of the value at some instant of the

    random process Y(t). In such cases we may give up the use of the exact distributions entering into the difference equations of the random variables and substitute them by some

    more simple distributions with the same moment

    characteristics.

    At the

    5.6. Multi-Dimensional Distribufions in Stochastic Differential Systems

    substitution of the random

    479

    variables by the variables with other distributions the

    difference equation will represent

    a weak approximation

    of a stochastic differential

    equation. For the stationary

    and nonstationary

    stochastic

    linear differential equations

    standard numerical methods of the ordinary differential equations may be used.

    Systematic presentation of the modern

    numerical methods

    for the nonlinear

    stochastic differential equations integration based on a strong and a weak approximation and different classes of the numerical schemes are considered for, example, in

    (Artemiev and Averina 1997, Kloeden and Platen 1994, 1997, 1999).

    5.6. Multi-Dimensional Distributions in Stochastic Differential Systems 5.6.1.

    One-Dimensional

    Characteristic Function

    Let us consider the system whose state vector (in the general case the extended state vector) is described by the Ito stochastic differential

    Eq. (5.4.6) where V is a white noise in the strict sense (Subsection 4.6.4). The problem is to find all the multi-dimensional distributions of the state vector of the system Y(t), supposing that the one-dimensional distribution (and consequently, all the multi-dimensional distributions (Subsection 4.6.3) of the process with the independent increments

    W(t) = W(to)+/V(r)dr

    (5.6.1)

    is known. > In order to find the one-dimensional characteristic function of the random

    process Y (t) let us consider two time instants t and t+ At. According to the definition of a characteristic function (Subsection 3.5.1) the values of the one- dimensional characteristic fucntion gj (A; t) of the process Y (t) at the instants t and t + At are expressed by

    gi

    Xt} Se OO Gis te At) = el OA), ante

    Ca waehg

    (5.6.2)

    Subtracting the first formula (5.6.2) from the second one termwise we shall have

    gi(A;t + At) — 91(A;t) eA

    | ef Y(eFapn ae,

    E Ei ane. + 1 eT Y(t).

    (5.6.3)

    480

    Chapter 5 General Theory of Stochastic Systems

    But it follows from Eq. (5.4.6) and Eq. (5.6.1) that with accuracy up to the infinite-

    simals of higher order AY (t)

    =Y(t+ A) — Y(t) = a(Y(t), t)At+ b((Y(t), t)AW, = (5.6.4)

    or omitting the argument t of the function Y(t) in (5.6.4) we get AY (t) = a(Y,t)At+b(Y,t)AW,

    AW=W(t+A)-—W(t).

    (5.6.5)

    Substituting (5.6.5) into (5.6.3) we find

    nist + At) — 9102) = B Le?EONarvonnaw] _ 1)7Y om Bie

    ee

    _ eiAT (Yt) AW

    eid" WY t)AW sl joyee or with accuracy up to the infinitesimals of higher order relative to At

    gi(A;t + At) — gi(A;t) = RCL

    ian

    P57 {er Maw

    ENT aly, t)At 4 eiATH(Y AW _ i}ees

    BS

    a 1) iseid (YY,t)AW _ i

    (5.6.6) For calculating the expectation in (5.6.6) let us use the formula of total expec-

    tation (Subsection 3.4.2) taking first the conditional expectation at fixed value y of the random

    variable

    Y , and then taking the expectation

    randomness of Y. As in the Ité Eq.

    AW

    Y =

    of the

    Y(t) and

    = Wit + At) — W(t) are independent, the conditional expectation of any

    function of the random variable AW

    at a given y does not depend on Y and is equal

    to the unconditional expectation of this function.

    Be

    with the account

    (5.4.6) the random variables

    HAW

    Consequently,

    Tay, t)eӴ hog [ema

    S neta

    SH {ixTa(¥, t)e"Y B eae eA

    Waa

    |y |

    |y|}

    nely [ia 16 $a Aged camel eee

    (5.6.7)

    and similarly

    Ee UYt)AW4ia7Y

    _ pp [arora

    og

    AeA) )

    (5.6.8)

    5.6. Multi-Dimensional Distributions in Stochastic Differential Systems

    the expectation in brackets being taken at fixed Y. characteristic function of the increment

    AW

    ay But He'# AW

    of the process

    481

    represents the

    W(t) which we shall

    denote by h(p; t,ti+ At):

    Belt4W — h(yt t+ At). Therefore

    (5.6.9)

    ee

    Bei MOO4W — (bY, 1) Aste

    Ady

    (5.6.10)

    and formulae (5.6.7) and (5.6.8) take the form

    Be HAW AT ALY,the Y = BidT a, t)h(O(Y, t)? A;t, t+ Ate’”, Bed YNAW4HATY _ PAY, t)T A;t,t + At)e?’”.

    (5.6.11) (5.6.12)

    Substituting expressions (5.6.11) and (5.6.12) into (5.6.6) we obtain

    gi (A;t + At) — gi(A;t) = E{ir7 h(b(Y,t)7 A; t,t + At)At

    AO t) Ate

    Ale We*.

    (5.6.13)

    Now we use formula (4.6.10) determining the characteristic function of the increment of a process with independent increments in terms of its one-dimensional characteristic function.

    Then denoting by hy (p; t) the one-dimensional characteristic function

    of the process W(t) we shall have

    h(u;t At) — hi (ust) A TA Dnteg UTA: ed oc es le+ At) eeea eehi(pyt at +Ee ee EiAD hi (pt) hi (p51) ( Supposing that h, (yu; t) has the continuous time derivative and using the Lagrange

    finite increments formula we find from (5.6.14)

    _ Ohi (457) where T € (igt + At). Consequently, (5.6.14) is equal to

    h(w;3t+ At)—1l=

    1

    Ohi (u57)

    hi(u;t)

    ot

    =

    or with the accuracy up to the infinitesimals of higher orders relative to At

    ag hA(ust+ At)—1

    ee

    1

    CAT

    Ohy (p15 t)

    a DEL

    At,

    6.1 (5.6.16)

    482

    Chapter 5 General Theory of Stochastic Systems

    Putting

    Ah

    -t)=

    1

    1(4;t —_

    5.6.17

    we find for (5.6.16)

    h(p;t + At) —1 = x(p;t)At.

    (5.6.18)

    Substituting (5.6.18) into (5.6.13) we get with the accuracy up to the infinitesimals

    of higher order relative to At

    gi(A;t + At) — 91 (A;t) = EB {irT a(Y, t) + x(8(Y, be t)} eV

    Ag.

    (5.6.19)

    Finally, dividing both parts of this formula by At and passing to the limit at At —

    0

    we obtain the equation which determines the one-dimensional characteristic function

    gi(A; t) of the state vector Y = Y(t):

    cee = B{irTal¥,t) + x(W(¥,t)7 Ast}

    Ya

    (6.6.20)

    The expectation in the right-hand side of (5.6.20) is determined by the one-dimensional distribution of the process Y(t) which in turn is completely determined by its one-dimensional characteristic function

    gi(A;t). Consequently, the right-hand side of formula (5.6.20) at a given t represents a functional of the characteristic function g;(A;t). Therefore

    (5.6.20) represents the equation determining g1();t). Let Yo = Y (to) be the initial value of the state vector of the system at the instant to representing a random variable independent of the values of the white noise V(t) at t > to. Denote the characteristic function

    of the random variable Yo by go(A). Then the initial condition for Eq. (5.6.20) will be 91(A; to) = go(A). (5.6.21) Eq. (5.6.20) and the inittal condition (5.6.21) determine completely and uniquely the one-dimensional characteristic function gi(A;t) of the state vector of the system (5.4.6) at any time instant t > to. Eq. (5.6.20) was first obtained at the begining of the 1940s by Pugachev (Pugachev 1944) who apparently was the first who studied stochastic differential equations with an arbitrary process with the independent increments. Bernstein, It6 and Gukhman studied only the stochastic differential equations with the Wiener processes (Bernstein

    1934, 1938, Ito 1951a, b, Gikhman 1947, 1950a, b).

    5.6. Multi-Dimensional

    Remark

    1.

    Distributions in Stochastic Differential Systems

    It goes without saying that Eq.

    483

    (5.6.20) is valid only if the

    expectation in the right-hand side exists. It follows from our derivation of Eq. (5.6.20)

    that in this case g1(A; t) is differentiable with respect to t and satisfies Eq. (5.6.20). Some sufficient conditions of the existence of the one-dimensional distributions are

    given in Subsection 5.6.12.

    Remark It6 equation.

    2. We emphasize that Eq. (5.6.20) is valid only if Eq. (5.4.6) is an Only under this condition the instantaneous value of the state vector

    of the system Y (t) is independent of the increment AW

    = W(t + At) _ W(t) of oT

    the process Wit), and the conditional expectation of the random variable e’4 relative to Y coincides with its unconditional expectation.

    deriving formula (5.6.13).

    5.6.2.

    ca

    We used this fact while

    - Pa

    Multi-Dimenstonal

    Characteristic Functions

    Analogously the equations determining other multi-dimensional characteristic functions of the state vector of a system are derived. > Let

    ga (ia

    Renae erik

    depen ta.) =e

    i> MY (te)

    .

    (5.6.22)

    lk

    be the n-dimensional characteristic function of the state vector Y of

    asystem. Taking

    two values t,, and t,, + At of the last argument t,, at fixed values of all preceding arguments tj
    CPM’s p14 and j1p of the stochastic systems A and B connected in parallel

    (Fig.1.6.1) determine a conditional probability measure in the direct product of the spaces Y X Z of the outputs of these systems.

    This measure is expressed by formula

    Pp Ole ii}pa(dy |2)up(dz |2) = /yn(Ey |#)pa(dy |2), €

    where € € Bx

    C, fy =

    y

    {z

    (5.8.1)

    : (y, 2) E E} is the section of the set E at the

    point y, B, C are o-algebras of the sets in the spaces Y and Z correspondingly.

    The

    integration over Y in the latter integral is performed practically over the projection E on the set Y as gy # @

    only in the case when y belongs to this projection. Analogously

    CMP’s

    14,

    Bp

    of the systems

    Systems

    511

    A, B connected

    sequentially

    (Fig.1.6.2) are determined. The conditional probability measure is determined in the direct product

    Y X Z. This measure is given by formula

    Haan |x)= ec ss

    |a,y)ma(dy|z),

    EE BxXC.

    (5.8.2)

    A special case of the sequential connection (Fig.1.6.2) presented on Fig.1.6.3 is also

    stochastic system (we denote it by C’) whose CPM —

    is obtained from (5.8.3) at €

    Seale 5 Fee: 7

    no(F|2)= fua(Pleyualdyle),

    FEC. 2

    (5.83)

    Ye

    Theorem 5.8.2. The sequential connection of the stochastic systems A and B represents the stochastic system C whose CPM wc 1s

    determined by formula

    (5.8.3).

    > Let us consider the closure [A] of the stochastic system by elementary feedback

    (Fig.1.6.4a,b).

    Let pa(G

    |Ab. y), G € B be CPM

    of the system A. Suppose that

    there exists such measurable mapping Ty : Y — Y

    of the space Y into itself that

    the measure [J 4 (FF 1G |x, y) does not depend on y. This latter measure if it exists we shall denote by pu(G |r) then the identity is valid

    MG

    caja

    Gled),

    Geb,

    (@yyeX

    x.

    (5.8.4)

    We put here y = yo. If Ty, is a convertible transformation then

    H(Ty,G |x) = pa(G| 2,40),

    GEB,

    where {14(G | 2, yo) is a conditional probability measure of the output Y of the system [A] at a given

    Z € X. Therefore CPM

    of the system [A] may be naturally

    determined by formula

    pal(G |c)=(TysGd

    2), Ge B.

    (5.8.5)

    This formula is reduced also to the form

    mpay(G |) = pa(T,'Ty,G|2,y),

    GEB,

    (5.8.6)

    512

    Chapter 5 General Theory of Stochastic Systems

    in consequence of (5.8.4). On the basis of formula

    pa(Typ Ghay)= pa(Ly,'G |x, yo)

    (5.8.7)

    pa(G |2,9) = val), TG 2, yo),

    (5.8.8)

    or equivalently

    which follows from (5.8.4) the right-hand side in (5.8.6) does not depend on the choice of yo EY.

    4

    Theorem 5.8.3. The closure of the stochastic system A by an elementary feedback represents the stochastic system [A] whose CPM pa} is determined by formula (5.8.6). The stated theorems establish the completeness of the class of the stochastic systems relatively to all possible typical connections (Sub-

    section 1.6.2). Formulae (5.8.1), (5.8.2) and (5.8.6) give the principal opportunity to find CPM of any connections of the stochastic systems by given CPM’s of the connected systems. The established properties of a class of the stochastic systems permit to use the structural methods for study of the composed stochastic systems. After partitioning a composite system into the subsystems we may study the characteristics of the subsystems and further to study the behaviour of the system as a whole. The structural methods for the linear and the nonlinear stochastic systems will be considered in Chapters 6 and 7. For the stochastic composed systems reducible to the systems considered in Sections 5.6 and 5.7 their CPM’s are determined by the multi-dimensional distributions.

    §.8.2. Distributions in Stochastic Systems with Random Structure Let us consider a system which is described by stochastic differential equation:

    Y =a(Y, S(t),t)+0(Y, S(t), t)V (k=1,..., N), where

    S =

    (5.8.9)

    S (t) be some step random process S (t) which describes the random

    changes of a system structure.

    Let 51, ...,

    SN

    be possible values of S(t).

    At

    § (t) = S, the system is described by the It6 equation

    Y =a(Y, sg, t) + 0(Y, sz,t)V (k=1, ..2, .N). We

    consider

    a case when

    the transitions

    (5.8.10)

    of a system from one structure

    to

    another form the Poisson stream of the events. We denote the Poisson process of the

    transitions into the k*® structure by Py (t) and its intensity by lz (Y, Si t) taking

    5.8. Distributions in Stochastic Composed Systems

    into account

    that it depends on the random

    the system state at the transition moment.

    513

    process S and may

    also depend on

    It easy to see that the process S(t) is

    determined by the following It6 stochastic differential equation:

    dS = \°(sx — S)dP, .

    (5.8.11)

    Supposing that while changing the structure of the system its state vector may receive a random increment by a jump we write the sought stochastic differential equation

    of a system with randomly changed structure in the form “

    N

    dY =a(Y,S,t)dt + 0(Y,S,t)}dW +5

    ~dQz,

    (5.8.12)

    k=1

    where Q

    (t) is the general Poisson process with variable distribution of the jumps

    generated by simple Poisson process Py (t) (k =s+lensd

    §

    N). The distribution of

    each jump of the process Q, (t) may depend not only on time but on the state vector of the system Y and its structure at the transition moment

    into the k*® structure.

    Thus after adding to the state vector of the system Y the value of the process S determining its structure we shall get a system of It6 stochastic differential equa-

    tions which describes the behaviour of a system with randomly changed structure. Naturally, the vector Y= The

    equations

    [ee ee \" will be its state vector.

    of a system

    with randomly

    changing

    structure

    (5.8.11) at Poisson stream of the changes of the structure may

    (5.8.10) and

    be written in the

    general form of the It6 equations

    dY =a(Y,t)dt+b(Y,t)dW. =

    Here

    4 i!

    Y =

    [Ss ay

    (5.8.13)

    we

    is the state vector of a system; W(t) is a process with indeTe

    pendent increments which consists of NV + 1 independent blocks W(t), [P, OF | ;

    wets

    Tape

    Wis

    [ Pv Qh, | ; a(y, t), b(y, t) are block matrices equal to

    a(j,t) = |Hoan || : = 47

    (y, t)

    b(y, s,t)

    | sy; —s

    0

    | sv—s

    0

    |

    I

    Pty 0

    ie

    0

    Using Eq. (5.8.12) we set up Eq. (5.6.7) for an one- dimensional characteristic

    function of the state vector for the system with the randomly changing structure.

    514

    Chapter 5 General Theory of Stochastic Systems

    For this purpose we find at first the function Xf; t) correspondent to the process W(t), ti —

    [ fy... LN ee In accordance with the mentioned

    above in Subsec-

    tion 5.6.3 the function Y(/i; t) in this case represents the sum of the functions X correspondent to the independent blocks Wit), [Pi (t)Qi(t)7 |. 3a [Pn (t)Qn (ee which compose the process W(t): N

    X(H; Y,S,t) = x(ust) +) xe(Hes Y,5,1),

    (5.8.14)

    k=,

    where x(y;t), Xe (He; i Lahey t) are the functions Ox(t) Ele correspondingly (k gl

    Lee

    eee N); —

    x of the processes. W(t), [Pi.(t) [ue?at ae pt," is the parti-

    tion of the vector [Z into the correspondent blocks; /l, =

    [ko pee is the expansion

    of the vector [1x into the blocks correspondent to the processes Pj, (t) and Qx(t). Here we took into consideration that the intensity of the streams of the jumps of the processes Py (t) and the distributions of the jumps of the processes Qx(t) may

    depend on the state vector of the system yi= [S bed es Using formula (5.6.8) for the functions VY and the formula for the characteristic function of the increments of the Poisson process and of the general Poisson process which is generated by it at

    infinitesimal interval (t, s] according to which hy (jax; Y, S,t)

    = 1+ [e'**9 9,(un; Y,S,t) — 1] (Y,S,t)(s —t) + o(s—t).

    (5.8.15)

    We find from (5.8.15)

    niu(ns ¥o(Syt) = [Le ox(ase

    Gib

    Ly] WEG St)5

    (5.8.16)

    where gx (MK3 Yy, S, t) is a characteristic function of the jumps of the process Q, (t)at

    the instant t. Thus (5.8.14) takes the form N

    X(H; Y,S,t) =x(ujt)+

    >> [et**°gn(ue; Y,S,t) — 1] 2 (Y, S,t). (5.8.17) k=1

    Therefore by virtue of (5.8.17) we have

    LOW, St) AVY Sity =v (00, Se) Ne) N

    Pye |elm Seg, (d; vigays 1 v(Y, S,t), iil

    (5.8.18)

    5.8. Distributions in Stochastic Composed Systems

    where

    \ =

    515

    [Xo dite is the partition of the vector A into the blocks correspondent

    to the blocks S, Y of the state vector of the system Y.

    Substituting (5.8.18) into

    (5.6.7) we obtain the equation for the one-dimensional characteristic function of the

    process in the system with the random changes of the structure:

    Ogi (A; t) =

    B{AinTaly, S,t) + x(0(Y, S,t)? d;t)

    N

    +50 iene darren S,t)- 1 »Z(Y, guy k=i

    Basing on Eqs. may

    derive

    ciatbiad .

    (5.819)

    -a

    (5.8.19) for the one-dimensional

    the equations

    for the conditional

    characteristic function we

    characteristic

    functions

    ferent structures and for the probabilities of these structures.

    in the dif-

    Taking into account

    that S(t) is a discrete random variable with possible values $1 , ...

    , SN

    and the

    probabilities of these values p;(t), ... , p(t) at each t we find by formula of the total expectation N

    Get) = BeoS(e) Ha" ¥(#) — Spe

    tg (Vt |1),

    (5.8.20)

    Rl

    where 91 (A; t |S1) is a conditional characteristic function of the state vector of the system Y in the [th structure.

    Analogously an expectation in the right-hand of Eq.

    (5.8.19) for 91 (A; t) is calculated. As a result the equation will take the form

    N Sefton Opigi(A;t |81) 1=1

    N = » opie 2

    ot

    [{aray, sj,t) + x(O(Y, si, t)" A; t)} ed ¥ |si |

    f=

    N

    N

    Le:

    +50 peo" yay |efalen=") l=1 k=1 N

    —- Soo l=1

    ge (2;Y, 51, t)ve(Y, si, te" ell si|

    N

    seu VEY, Si, fe

    ¥ |sre}.

    (5.8.21)

    k=!

    Here the index (1) at the sums over k shows that the item correspondent to k =

    does not enter into the sum.

    |

    It does not enter into the sum as the processes P} (t)

    516

    Chapter 5 General Theory of Stochastic Systems

    and Q (t) maintain the constant values at the absence of the change of the structure

    as a result of which gi(A; Y} Sig t) = 1. After changing the order of the summation

    in the first double sum (5.8.21) we get N

    N

    4

    -\T

    De

    Spe |es, sr, tualY,s1,te™"Y |si].

    Loi

    (fs

    (5.8.22)

    After replacing in (5.8.22) the indexes k and | (the,double sum will not change in this case) and substituting the obtained expression into Eq. (5.8.21) we shall have

    N s eirosi Opigi(A;t |s1) Ot

    feat

    N

    = > pie

    EB[{at acy, si,t) + x(O(Y, si, t)" A;t)} ei¥ |si|

    N N +o peo SpE a

    [ms ¥, se, t)a(Y, se, tye¥ |sx|

    k= N

    ~ So peo e [u(¥, t)e"Y |si |.

    (5.8.23)

    =i

    N

    where for brevity is assumed i(Y, t) =

    LOWY,

    Sint): Comparing the coeffi-

    [geil

    cients at similar exponential functions and denoting by fily;t

    | S1) a conditional

    one-dimensional density of the process Y(t) in the [* structure we set up the system

    of the equations from Eq. (5.8.23)

    0 eS r a e

    . ii;{id? a(y, s1,t)

    +x(b(y, 81,t)? A; the" }fa(yst |sr)dy N

    +) ka

    co

    pe / H(A;Y,Sk, t)M(y, se, te

    —pI / u(y, te —oo

    “ fi(yst |se)dy

    =

    Y fr(y;t bspldgh (i= dead) ME

    (5.8.24)

    5.8. Distributions in Stochastic Composed Systems

    Putting in Eq. =

    (5.8.24)

    4 =

    517

    °0 and taking into consideration that gi(0;t

    | 51)

    Is x (0; t) = 0 we get the equations for the probabilities of the structures

    =1,...,

    (1

    N) of the system at the moment t:

    N

    CO

    Cpl

    co

    1)

    py / Vi(y, Skt) fi(yst |se )dy — pr / ily, t)fr(yst; |si)dy.

    p=>

    ;

    k=1

    —0o

    —0o

    :

    (5.8.25)

    Using Eqs. (5.8.24) we derive the equations for the conditional densities from the equations for the conditional characteristic functions in the case of the Wiener process W(t).

    For this purpose following Subsection 5.6.6 we change the denotion

    of the integration variable y by 7, multiply the equation for pigi(A;t

    | s}) by

    eid" y /(20 yp and integrate it over \. The first integral is transformed similarly as in Subsection 5.6.6 and gives as a result an item which is analogous to the right-hand

    side of Eq. (5.6.42). The second integral as a result of formula

    jad

    .

    a (2x)P | (Asn, ses the "dA =a(y—nin.set),

    (5.8.26)

    where gi(Y; 1), Sk, t) is the density of the jumps of the process (;,(t) at given 1), Sk and t is transformed to the form foe)

    / qi(y — 737, sk, t)(n, sk, t)fi(n;t | sz )dn.

    (5.8.27)

    —oo

    The

    third

    integral

    in Eq.

    (5.8.23)

    ny, t)fi (y;t |s1)3 As a result for!

    as

    a result

    of (5.6.37)

    takes

    the

    form

    = 1,..., N we have by virtue of (5.8.27)

    Opifily;t OF [aly sr HACE | sr) ] _ |sr) = —Pr5y ie

    +e tr laa [b(y, s1, t)u(t)b(y, sx, t)” frly;t |s)|} N

    #0? si

    co

    fay — 131, 8k, t)™i(n, sx, t)fi(n;t| se)dn — pira(y;t) fi(yst| 51). ne

    (5.8.28)

    518

    Chapter 5 General Theory of Stochastic Systems

    In important

    special case where at the transition from one structure

    other the state vector of the system Y does not receive a random items with dQr in the equation for Y are absent (Qk (t) =

    to an-

    increment

    the

    0).Here in the equa-

    tion for the one-dimensional characteristic function gi(A; t) and the equations for the conditional one-dimensional

    characteristic functions gi(A;t

    | S1) at different

    structures we have gil A; Y, x,t) = 1. Correspondingly in the equations for the conditional one-dimensional densities fj (y;t |$1) at the different structures we get

    gi(y—n; 7, sk,t) = 6(y—7) and the equations for Firlost:| ay)

    is hea)

    take the form

    a) e it ar Lalu, sist Se ae) = —prg, ) fa(yst |81)] T

    +5 tr las [d(y, si, t)v(t)b(y, 81, t)? fa(y;t | =|}

    N +5 WOu(y, se,t) [pefilyst |se) — prfi(yst |1)] (=1,..., N).

    =

    (5.8.29)

    5.9. Distributions in Stochastic Infinite-Dimensional Systems 5.9.1. Distributions in Stochastic Continuous Hilbert Systems

    At first let us consider the random process

    Y = Y(t) with the

    values in separable H-space Y which is determined by the It6 stochastic differential equation of the form:

    dY =a(Y,t)dt + b(Y,t)dW, Y(to) = Yo.

    (5.9.1)

    Here W = W(t) is a random process with the independent increments in some other H-space from W to Y, a(Y,t) is an operator mapping Y x R into Y, b(Y,t) is an operator-valued function mapping Y x R into the space of the linear operators which act from W to Y, Yo is a random variable in the space Y independent of the future increments of the process W. The central problem of stochastic systems theory consists in finding the multi-dimensional functionals of the process Y: n

    Gio iis os

    Pen tig

    = os

    git

    =e

    OX

    04,74)

    ie

    (n 21, Ios.)

    (5.9.2)

    5.9. Distributions in Stochastic Infinite-Dimensional Systems

    519

    Here 41, ..., An are the elements of the adjoint space )*, (Ax, Yt,) is a scalar product of the elements A, and Y;, in H-space Y. The multidimensional functionals gn = gn(A1,..., Anjti, ..., tn) completely de-

    termine at the known conditions a distribution of a process (Section 3.6). The stated problem is easily solved while performing the following conditions. (i) The space Y of the values of the process Y(t) at each ¢ is a separable H-space.

    (ii) The random process W(t) represents the Wiener process with the values in the separable H-space W whose covariance operator of the value at each given t is determined by formula t

    FSi tt

    fear,

    (5.9.3)

    to

    where v(t) is the intensity of the process W(t) which represents at each t the trace-class adjoint operator.

    (iii) The function a(Y,t) with the values in the H-space Y is determined by formula

    a(Y,t) =a,Y + a'(Y,t).

    (5.9.4)

    Here a, being a closed linear operator which is an infinitesimally generating operator of strongly continuous semigroup of the restricted oper-

    ators {u(t)} satisfying the condition ||u(t)|| < ce~7 at some c, y > 0 and equation

    dujat= au,

    uo) = 1;

    (5.9.5)

    where J is an unit operator, a/(Y,t) is the strongly differentiable function over Y and ¢ with the values in the H-space J.

    (iv) The operator-valued function b(Y,t¢) at each Y and ¢ represents such compact linear operator that the self-adjoint operator

    o(Y,t,7)

    = u(é—r)o(Y,t)u(t—7)",

    of ¥,t) = 0(Y,t)v(@)0(Y, 1)", (5.9.6) is trace-class and at each t the following conditions is fulfilled:

    t / tral

    Y,t,e)dr-< co.

    (5.9.7)

    to

    The representability of the function a(Y, t) in Eq. (5.9.1) in the form of (5.9.4) with the restrictions on aj, a’(Y,t) and condition (5.9.7) are

    520

    Chapter 5 General Theory of Stochastic Systems

    sufficient for the existence of the solution of the Ito stochastic integral equation

    t

    t

    Y(t) = u(t)Yo + ju —r)a'(Y(r),7)dr+ ju —T)b(Y(r),7)dW, to

    to

    (5.9.8) and for the existence and the integrability of the operator of the second order moment. The property of the operators v(t), o(Y,t), v(Y,t, 7) to be trace-class is necessary and sufficient for v(t)At, o(Y,t)At, v(Y,t, r)Ar be the covariance operators of the normally distributed random variab-

    les AW, b(Y,t)AW(t) and u(t — 7)b(Y,7)AW(r) for the infinitesimals At and Ar.

    Let us show that while performing conditions (ii)-(iv) the multidimensional characteristic functionals (5.9.2) are determined by equations Ogn(A1

    >) ans

    9 Agel;

    relied

    ta)

    =

    Bfivan a(¥

    ta)

    ate

    1

    —5 (ny (Yinstn)) An)exp 2

    ie

    iS Ok, Yi) \Gah

    cD ged)

    k=

    (5.9.9) and the initial conditions

    91(A1;0) = go(A1), gn(A1, nue Pe A

    es ERE,

    (5.9.10)

    24) = gn—1(A1, sae An—-1+An3 #1, i

    a Cet)

    :

    (5.9.11) where go(A1) is the characteristic functional of the random variable Yo. > In fact it is sufficient to use Ité formula (5.1.15) at a given 7 for the calculation n

    of the stochastic differential of the composite function exo]i520, Yi) which k=)

    is considered as the function of a scalar random process

    Ni Site: ne ei i Nee 1

    ee

    Yaw

    Oss Y(tn)) at the fixed

    and to use the formula of the transformation of

    a covariance operator in linear spaces (Subsection 3.3.2).
    0. Example

    5.9.1.

    Let us consider

    a stochastic system described

    by a

    nonlinear wave stochastic differential equation

    @U

    o?U

    GU Ts nla?

    re = c7(1 =8 Vio + 2h; reel) — gh (3)

    with the boundary conditions U(0,t) = U (2154) '==10. oHere =

    Ce, + Vo

    (I)

    es U(x, t); Vi,2

    Vi ,2(2, t) are the independent normally distributed white noises with the inten-

    sities

    Yj 2 = V19(X1 , 29, t); (et hi 2 afe constant coefficients. After putting

    ¥,=U, ¥,=au/at,

    Y=[Yi Yo] ,

    A

    ©

    ARCA Sardar

    enme |

    0

    ah

    0

    ie in |ezary, /da? | we reduce Eq.

    (I) to the form of Eq. (5.9.1). Eq.

    (5.9.9) for the one-dimensional

    characteristic functional gj = gj (Ay , AQ} t) will take the following form:

    Ogi(A1,

    ee

    A23t

    =

    :

    BYvas, ¥e)

    +8

    ;

    502%)

    (22.0%Tae) ae

    2 + 2h, Yo —- zhi?

    15(da,v2d2)] Run Vi VatiOa’ ay \" )tiQr2, s , oY o (00,(Sa at) (52) Aa) = ie)

    -5

    (IIT)

    where a scalar product is given by the formula (y, w) = i)p(x)(x)dx : 0

    Along with Eq. (5.9.1) there are the stochastic differential equations in the separable H-space Y of the following form:

    dY =a(Y,t)dt + b(Y,t)dWo + /c(Y,t,u)P%(dt,du).

    (5.9.12)

    Ro Here Wo = Wo(t) is the Wiener process with the values in the H-space Wo, P°(A,€) is the centered Poisson measure, A = (t1,¢2) is some

    522

    Chapter 5 General Theory of Stochastic Systems

    time interval, € is some Borel set of the space Rj representing the qdimensional space with pricked origin, a = a(Y,t) and c = c(Y,t, wu) are

    the functions mapping Y x R and Y x Rx Ré into J, b = 0(Y,t) is an operator-valued function mapping Y x R into the space of the linear operators which act from Wo to Y. We shall assume that the initial value Yo is an element of the space Y and does not depend on the increments of the Wiener process Wp and the values of the Poisson measure P°(A, €) at the time interval A = (1, ta], to < ti < te for every E. In this case similarly as in the previous case it is shown that multidimensional characteristic functionals (5.9.2) satisfy the equations AG aCAyh, eas

    toied bra eareee)

    at, +x(An, Yt,j3tn)] exp Fon %) k=1

    = BYLin,al ¥i.5 ta) \(ni BQyees)

    (5.9.13)

    with initial conditions (5.9.10) and (5.9.11) where 1

    x(A, 95 t) = —5Q, ool, 4) + {efvely.tu) Lae uA, c(y,t,u)) yp(t, du),

    (5.9.14)

    R?

    oo(y,t) = b(y, t)vo(t)b(y, t)*, vo is the intensity of the process Wo which represents the trace-class operator, b(y,t)* is an operator adjoint with b(y,t), vp(t, du) is the intensity of the simple Poisson process. The obtained results may be generalized on the stochastic conti-

    nuous systems in B-spaces with a basis (Pugachev and Sinitsyn 2000). 5.9.2. Distributions in Markov Stochastic Discrete Systems

    Let us obtain the equations of the discrete Markov systems which are important in applications. For this purpose we consider the random

    sequence {Y;,} with the values in some measurable space (Y,B) which satisfies the following stochastic difference equation: Yead

    =e

    Yin We)

    bis

    he

    ee

    )

    (5.9.15)

    where {V,} is the sequence of the independent random variables with the values in other measurable space (V,€); we(yz, vz) are (B x E, B)-

    5.10. Applications.

    Optimal Online Filtering

    523

    measurable functions mapping Y x V into J. Let the initial value Y; be

    independent of the sequence {Vi}; 41(B), B € B is the distribution of Yi; vx(£), E € E is the distribution of V;. In this case {Y;,} is a Markov

    sequence. Consequently (Subsection 2.7.9), its multi-dimensional distributions are completely determined by the initial distribution p(B),

    B €B

    and by the transition distribution y,%41(B |Y¥,), B € B and

    wesa(B |y) = ve(we*(y, BY),

    (5.9.16)

    where we (y, B) € V is the inverse image of the set B € Y which is determined by the function w,(y,v) at the fixed y. Recurrent Eq. (5.9.15)

    at k = 1,2,... and given initial distribution 1(B) permit to find the distributions of the sequence {Y;}. . 5.10.

    Applications.

    Optimal Online Filtering

    5.10.1. Introductory Remarks Problems of determining the state of a system using the results of some measurements often arise in practice.

    As the measurements

    are always accompanied by

    random errors one should speak not about the determination of the system state but about its estimation (filtering, extrapolation, interpolation) by statistical processing

    of the results of measurements. At practice a filtering problem is technically solved by means

    of passing the

    measured signal Paks: through the device which is called a filter for “filtering off’ the noise and for obtaining a signal at the output which reproduces the required process Xx (t) with possibly great accuracy.

    Let us consider the continuous stochastic system described by Eq. (5.4.6)

    X = 9(X,t) + V(X, t)V where

    X

    is the n-dimensional

    state vector of the system,

    (5.10.1) V is the r-dimensio-

    nal vector white normal noise, and y(z, ae v(a, t) are known functions of the system

    state and time. The values of the function p(x, t) are evidently the n-dimensional vectors, and the values of the function v(a, t) are the N X T matrices. If the state vector

    of the system

    X

    is continuously

    measured

    then

    the n-

    dimensional random process Y (t) =—¢ (t) +U (t) will be the result of measurements where UV (t) is the measurement error which represents usually a random function of

    time.

    However, it is not the components of the state vector, but some functions of

    the state vector are usually measured among which may be some of the components

    524

    Chapter 5 General Theory of Stochastic Systems

    of the state vector. The result of the measurements is determined in the general case by the formula

    Y =Y(t) = 0(X,U,t)

    (5.10.2)

    where Y is the 1-dimensional vector, U is the measurement

    error representing a

    vector random function of time of dimension 7 > 1, and Yo(Z, u, t) is the known function of the system state, measurement

    error and time.

    In the general case the

    function (9 is nonlinear both relative to the state vector and to the measurement error. The dimension of the error vector U in problems of practice cannot be smaller than the number of measured variables 7; as each measurement

    is accompanied by

    an error and the errors of various instruments are usually independent

    (or at least

    contain independent items). In some cases the measurements are performed by instruments possessing inertia or the results of measurements may be filtered by various filters. So the general model of measurements which are performed in a system may be described by the differential equation

    Y = 91(Y,X,U,t). The result of measurements

    (5.10.3)

    represent the random process Y (@): We come

    the problem of filtering of the state vector of

    asystem X at eachinstant

    thus to

    t > to using

    the results of continuous measurement of the process Y determined by Eq. (5.10.3)

    in the time interval [to, t]. Remark.

    We emphasize that such a statement of the filtering problem is

    adequate only if online data processing, i.e. data processing in the real time basis is possible.

    If online data processing is impossible and one has to be content with

    post data processing when the measurement is finished, then the above statement of

    the filtering problem has no sense.

    The estimates will always be better if the whole

    measurement is used to estimate the state vector of a system at any instant t during the experiment.

    Therefore, other techniques for solving filtering problem must be

    used in such cases (Subsection 1.6.5). Thus we come to the exact statement of the filtering problem. A vector random process

    sare

    ie 1s determined

    by the Ité stochastic dif-

    ferential equations

    dY = 9,(Y,X,t)dt + di (Y, X,t)dW, dX = 9(Y, X,t)dt + ¥(Y, X,t)dW, where Y is a N1-dimensional

    random

    process,

    X

    is a n-dimensional

    (5.10.4) (5.10.5) process,

    W

    is a N-dimensional process, v1(y, azyt) and ply, x,t) are known vector functions mapping

    the space R™!

    x R”

    x R into the spaces R™!

    and R”™ respectively,

    ply, AD t) and vy, ape t) are the known matriz functions mapping R"!

    and

    xXR” x R

    5.10. Applications.

    into R™"

    and R""

    respectively.

    of the system) at any instantt

    the process

    5.10.2.

    Optimal Online Filtering

    525

    It is required to filter the process (the state vector

    > to using the results of continuous measurement

    of

    Y in the time interval [to, t].

    General Formula for Optimal Filtering

    It is natural to require that the filtering of the process some sense.

    Y(t) be optimal in

    The mean square error (m.s.e.) criterion B|X, — X; l=

    min serves

    as a natural criterion of optimality in many problems of mathematical statistics.

    If

    we assume this criterion, then the general solution of the filtering problem follows directly from the known property of the second order moments:

    second order moments

    of a scalar random variable is its variance.

    the smallest of all

    Hence it follows

    that the best approximation of a random variable by a nonrandom variable from the m.s.e.

    criterion point of view is given by its expectation.

    In particular,

    the best

    approximation of a random variable using the results of measurement is given by the conditional expectation relative to the results of measurements.

    We denote by Ye the aggregate of the values of the measured process in the time interval [to, t], Y;, = {Y(r) TE [to.t}}. Then the optimal estimate of the

    vector

    X, = X (u), which gives the solution of problem at u = t is determined by

    the formula

    ba

    1S

    eleeB

    (5.10.6)

    This formula determines the optimal estimate of the value X,, of any random function

    xX (u) using the results of measurement of other random function Y(t) in the interval [to, ae It is also valid for the case of a vector argument ¢ and the measurement

    of

    the random function Y (t) on any set J’ of values of tf.

    In the general case it is necessary for practical application of formula (5.10.6) to find the conditional distribution of X,.

    the general case is not yet solved.

    This problem is very difficult and in

    In the particular case of the processes Y(t) and

    X (t) determined by Eqs. (5.10.4), (5.10.65) it may be solved under some additional restrictions. But even in these cases the practical application of formula (5.10.6) is a matter of great and often insuperable difficulty. The reason is that the determination of the conditional distribution always requires very cumbersome and time-consuming

    calculations

    which

    can be performed

    only after measurement.

    The

    statement

    of

    problems in Subsection 5.10.1 implies that the estimates must be calculated in the real time while the results of measurement appear. But the difficulties of application of formula

    (5.10.6) do not reduce the importance of the optimal estimation.

    This

    theory is necessary for studying the potential accuracy of the estimates, i.e. maximal achievable accuracy of the estimation.

    526

    Chapter 5 General Theory of Stochastic Systems

    §.10.3. Auxiliary Problem The general formula for the stochastic differential of the optimal estimate of a given function of the state vector of a system underlies the optimal filtration theory.

    Let f(X 5 t) be some scalar function of the n-dimensional state vector of a system and of time.

    Its optimal estimate using the results of observation ¥? according to

    (5.10.6) is determined by the formula

    (5.10.7)

    FOS E[FALOTRY:

    This estimate represents a functional of the random process Y(t) in the variable

    time interval [to, t], and consequently, is itself a random function of t. We put the auxiliary mathematical problem:

    to find the It6 stochastic differential of this random

    process. This problem may be solved under the condition that W(t) in Eqs.

    (5.10.4),

    (5.10.5) represents the Wiener process whose dimension 7 is not smaller than the dimension

    1; of the measurement

    process ¥(¢), and that the furiction Yj in Eq.

    (5.10.4) does not depend on X. Eqs. (5.10.4), (5.10.5) in this case have the form

    dY = ,(Y,X,t)dt + U1 (Y,t)dW, dX = 9(Y,X,t)dt + (Y, X, t)dW. Lemma

    (5.10.8) (5.10.9)

    5.10.1. If o1(y,t) = ily, t)v((t)dr(y,t)? where v(t) being

    the intensity of the Wiener process then Eqs. (5.10.8), (5.10.9) may

    be transformed

    to the following form:

    dY = 91 (Y,X,t)dt + Ui (Y, t)dWa,

    AX = oY, X,t)dt + Hi(Y,X,1)dW, + WHY, X)dWy. Here W, = W, (t) and W(t) first,

    Oe hOal 20-10)

    are two independent Wiener processes formed by the

    = T—N 1 components and by the last components of the process W’ = W'(t)

    with independent components: t

    w(t) = /Ghai ueDGMD 0

    where yp, V1, vy be some matrices and Q(y, t) be any orthogonal real matrix possibly

    depending on Y and time t. > As it is known the power of a symmetrical nonnegative definite matrix VY

    is determined by the formula y* =

    ANSAS,

    where

    A =

    diag{\i,...,

    Aq}

    5.10. Applications.

    Optimal Online Filtering

    52m

    is the diagonal matrix to which V is reduced by the orthogonal transformation A,

    and A* =

    diag fag RSS

    As}. Then the random process defined by (5.10.11)

    represents a Wiener process with independent

    components.

    Really, the covariance

    matrix of the value of the process W’(t) at a given ¢ is determined by the formula

    EW'(t)w'(t)" = /Q(y, ry V?(r)v(r v4 2(r)A(y, 7) dr = Ht, 0

    since Q(y, T)Q(y, r)t = I by virtue of the orthogonality of the matrix Q(y, T). This formula shows that the components of the process W'(t) are uncorrelated, and consequently, are independent as a Wiener process is normally distributed. Since the intensity of the process W'(t) is equal to the identity matrix, each component of the

    process W(t) represents a standard Wiener process. Let us pass from the process

    W(t) to the process

    W'(t) in Eqs.

    (5.10.8),

    (5.10.9). Then the differential dW will be replaced by y/2(4)Q(y, t)P dw’. Now we choose the orthogonal matrix (2(y, t) in such a way that the matrix b;(y,t) be reduced to a block form [0 vi (y, t)] where the first block represents the N1 X (r—ny) matrix whose

    elements are all equal to zero, and the second block represents an

    m4 X nN, matrix. It is possible under some conditions, as the orthogonal Tr X Tr matrix Q has r(r =

    Hi? arbitrary elements which must satisfy ni(r =

    and rr — 1)/2 = r?/4 abi Ti

    ny) conditions,

    2; ni(r _ ny) < r?/4. As a result of such a

    transformation

    vase [= [very v1(Y, t)

    [eroacoraw

    ei) ee

    1/24

    PF

    A CAV ALS OC

    iart

    :

    =|Seater[aM =

    1(Y,t)

    7 hagee Y, X,t) Fea? 8

    /

    V1 (Y, t)dWe

    i |W'(Y, X,t)dW, + b"(Y, X,t)dWy |: To find the matrices 1’, wy" and vy and establish the conditions under which the transformation of Eqs.

    (5.10.8), (5.10.9) into the form of Eqs. (5.10.10) is possible

    we evaluate the conditional covariance matrix of the random vector

    Wien 18”=[warxy vary |" relative to the random variables Y, =

    X where AW

    = Wit AP At) — W(t), AW'

    W'(t + At) - W'(t). Taking into account that AW

    and AW’

    are independent

    528

    Chapter 5 General Theory of Stochastic Systems

    of Y;, Xz and their covariance matrices are equal to y(t)At + o(At) and [At respectively, using the formula of transformation of the covariance matrix at a linear transformation of a random vector and omitting for brevity the arguments of the functions w, 1, yw, yw", wy we have

    oe emmy ewe ~ yp Fay paceman 1

    Mews pl en»

    aglt

    | pl

    wed at :

    yt

    yt

    Hence, dividing both parts by At we obtain as At + 0)

    vivdt pdt i ie ce yvyt yyy? ppt wp?

    + py?

    | ’

    f= byl, Wt! = wed, Ve +e = doy. (5.10.11) It remains to solve these equations successively for v1, wy” and x’. The matrix wy determined by formula (5.10.11) is independent of X, as 1 is independent of X. The second Eq. (5.10.11) has the solution if and only if the matrix vi (y,t ) is invertible at all y, t. It is necessary for this that the oe

    py vt

    invertible at all y, t since on the basis of the first Eq. (5.10.11)

    |vivyf |.

    In this case the second Eq. (5.10.11) gives ae

    Wil =

    be

    pul (pi! ee After that the

    third Eq. (5.10.11) is reduced to the form

    >

    wee? = dvd? — pod (YP (Wh ed*, or taking into account that by virtue of the first Eq.

    (5.10.11) (oy rT (Ont )

    = (pivyt)-

    vy = wd? — dol (dvd) dy. The right-hand side of this equation represents

    a symmetrical matrix.

    Let us prove

    that it is nonnegative definite. For this purpose we notice that the matrix

    es ;

    ,

    pvp? pup, ‘

    pivyt ppt

    |?

    F

    being proportional to the covariance matrix of the random vector

    j

    :

    ke

    ,

    is nonnegative definite in consequence of which at any vector U =

    Ie

    [yr yr | dw

    ae

    [ut us l

    ul Au = ThPabyvyt Ui + UyPFpyyt uy + uy;Pb uu.+ UsPvt

    us >> 0.

    5.10. Applications. Optimal Online Filtering

    Putting here uj = —(~1 yr) fk

    aL

    529

    divdT u2 we shall have

    T\-1

    lB

    a

    is

    us Pryy (divdy )~ pivy? ue — ug pot (dive) diveduy —uZ byt (divdt)“ dvd? us + ul dy up.> 0, or after cancellations

    us [yvy? — prot (divet)

    pvp? |us > 0.

    The validity of this inequality for all n-dimensional vectors U2 proves our statement.

    After this the solution of the last Eq. —

    ia [v _ yor (v4 ypr)- tyr

    ces vy and

    (5.10.11) is given by the explicit formula y’

    |any

    It is-evident

    that multiplying the matri-

    on the right by arbitrary orthogonal matrices of the corresponding

    dimensions we obtain another solution of Eqs. (5.10.11). Thus Eqs.

    (5.10.11) have an infinite set of solutions, and consequently, Eqs.

    (5.10.8), (5.10.9) are reducible to the form of Eqs.

    (5.10.10) if the matrix 01 (y,t)

    = vi(y, t)v(t)vr(y, ie is invertible at all y, t. < Lemma

    random

    5.10.2.

    The stochastic

    differential of the optimal estimate

    of the

    variable f (Xz, t) for Eqs. (5.10.4), (5.10.5) is given by formula:

    df = Elfi(X,t) + fe(X,t)" 9(Y, X,t)

    + hte{fee(X (Wy? UY,XH} IVAIdt + BLK, 1)(ei, X,)% oF +fo(X,t)? (budt (Y, X,t) |Yilavyr)"(¥, (d¥ —Grdt), (5.10.12) where

    (bub (dur

    pr(z)

    )(y, z,t) = p(y, 2, t)v(t)yv(y, z; ae )(y, ev; t) = vy, v, t)v(t)di(y, yh

    (divbl)—*(y,t) = Wily, ty(dr(y, t)7)77,

    (5.10.13)

    ~i= / yip:(x)dx= Elyi(X1,¥%,t |Y;,),

    (5.10.14)

    being the conditional

    density

    of Xz relative

    to Ye; the derivatives

    fy, Tse

    fer and all the conditional expectations in its right-hand side are supposed to exist.

    p> Let us consider two close instants t and ¢t + At, At > 0, and write

    Af(t) =f(t+ At)—f(t)

    530

    Chapter 5 General Theory of Stochastic Systems

    = E[f(Xipar,t+ At) |Vit**] -

    [F(X0)1 YA |

    = E[ f(Xtzar,t+ At) — f(Xt,t) 1Yi]

    +E [f(Xisat,t + At) |¥i*4"]—F [f(Xtpar,.t

    +A) |Yi]. (6-10.15)

    We evaluate every item in the right-hand side separately.

    Using the differentiation

    formula of a composite function (5.1.7) by virtue of the second Eqs. (5.10.10) we get omitting for brevity the arguments of the functions

    f(Xtzar,t + At) — f(Xt,t)

    = {(Xt) + fo(X1,t)"9+ tr (teal Xe 0100" |vs|) mi

    + fo(X1t)7 (WAM + YAW) = {f(t 1 oe

    +5 [fea(Xe(yy + ¥'b"™)] Late f( Xe)" (WAM +W"AM). Substituting

    =

    here the expression

    pl" (AY =

    of AW

    from

    the first Eqs.

    (5.10.10),

    AW,

    yi At), taking the conditional expectation of the obtained ex-

    pression relative to Ye: using the third Eqs. independence of AW;,

    AW2,

    (5.10.11) and taking into account the

    AY we obtain at fixed At

    E| f(Xtzar,t + At) —f(Xs,t) |Ys

    E [f(X00) ++fe(Xe,t)? p+ str{foo(X+, ty" } |a At

    B |fo(X1,t) wd" (AY — gi At) |¥s] . Further we find from the first two equations of Eqs. (5.10.11) -1

    eawi dt ), yO

    -1

    =e

    oh

    =

    (beet).

    Substituting these expressions into the previous formula and taking into account that w 1 does not depend on X; we shall have

    E |f(Xtzar + At) — f(X1,t) |Ve| = E | fi(Ys,t) + fo(Xt,t) e+ tt {fee(Xz,t)puy"} || At

    5.10. Applications.-Optimal Online Filtering

    531

    +E [fo(Xe,t) buys (vivdT)- (AY — pi At) |Yi].

    (5.10.16)

    For evaluating the second item in (5.10.5) we shall find the conditional density of the random vector X;4, relative to viper

    the random vector X; relative to Y,, a(n random variable AY

    Let pr(z) be the conditional density of

    | Y, =) the conditional density of the

    at given values Y, & of the random vectors Y;, X; at a fixed

    value Z of the random vector X;. As the first equation of (5.10.10) determines the Markov process Y;, the density qu(n | Y, x) coincides with the conditional density

    of the random variable AY at given values ie and @ of the random variables Ye and X;.

    Consequently,

    on the basis of the multiplication theorem of the densities

    the conditional density p(x) of the random vector X;} relative to Voe

    or, what

    is the same, relative to vee and AY js determined by the formula

    pr(@)qr(AY | Yi,2)

    p(x) =

    (5.10.17)

    ofpr(a)qe(AY | %,a)da Here the numerator represents the joint conditional density of the random variables

    X; and AY

    relative to ys and the integral in the denominator represents the

    conditional density of the random variable AY

    relative to ae

    For determining

    qi(Ay |x; zr) we notice that by the first equation of (5.10.10) at AY=

    X = x

    yil%t, z,t)At aF wy (¥1, t) AW2.

    Hence it is clear that at given values of Y; and X; = & the variable AY

    represents

    the linear function of the normally distributed random variable AW 2. Consequently, the conditional distribution of AY

    for determining q(A

    at given values of Y; and X; = @ is normal and

    | Ys, a) it is sufficient to find the corresponding conditional

    expectation and covariance matrix of the random formulae

    for the expectation

    and the covariance

    vector AY. matrix

    Using the known

    of a linear function of a

    random vector, taking into account that the covariance matrix of the random vector AW

    is equal to At, and omitting the arguments of the functions ~; and wy we

    obtain

    PLAY

    IY 52 Par,

    Ky

    =

    tk

    At,

    or, by virtue of the first Eq. (5.10.11)

    E[AY |¥i,2]=giAt, Kay = vivyfAt Consequently,

    a(AY | Yin)= coxp{-SR (AY?—pFAt) (dive) "(ay pid},

    532

    Chapter 5 General Theory of Stochastic Systems

    c being the normalizing constant.

    Taking into account that the quantities At and

    AY are infinitesimal we may represent the exponential function by the Taylor formula

    with the residual of order o(At). Then we obtain

    O(

    Vouliptaw Goze cexp{— —— 5 :AG AYE(ahd NE AX

    tof (divdf TAY — Sef hvu) oath = cexp {=A

    vf ty tay|f+ofand) AY

    Z

    1

    Ss

    5¥1 (Wid) peAtt sor (vivdt )'AYAY® (vivdy

    :

    gi].

    Retaining only the expectation of the random variable NY UMY -

    EAYAY?™= gp?At?+ of b//At = piv yt At + o(At), we obtain q(AY

    |Y;, xz)

    1 = cexp{ AY" avd tay} [14 oF duvet) tay], and, as {1 is independent of 2, co

    | pr(z)qe(AY | ¥;, x)dx —oo

    = cexp {-aaphYT Wit) tay} [1467 (vyivdt) AY], where

    $j

    g(AY

    is defined

    by

    (5.10.14).

    Substituting

    the

    obtained

    expressions

    of

    | Ys; zr) and of the integral into (5.10.17) we shall have after the can-

    cellation

    ! =~ 1+¢f (ivy? )- LAY p,() = pi(a“Ty aT WORDLE hydED TAY" AY Representing 1/[1 + pT (vy yp? )-} NG

    by the Taylor formula with the residual

    of order o(At), performing the multiplication and retaining only the expectation of

    the random variable

    AY AY”

    as before we get

    Pi(z) = p(x) + ot (divdt)*AY][1— OT Wivdt) AY

    5.10. Applications.“Optimal Online Filtering

    033

    +01 (YivdT) AYAY? (hivyt)-* 41] = pr(){1+ (¢t — OT )(divdt) TAY — AYAY? (pivyf)-'eT]} = p(x)[1+ (9T — $1 )(divdt)“*(AY — @ At)]. Now

    we may

    evaluate the second item in the right-hand side of formula

    (5.10.5).

    Taking into account that according to (5.1.7)

    f(Xt+at,t + At)

    = (Xt) + {fi %e,t) + fo(Xe te + 5ttlfeo(Xe,t)oryT At 4+fa(X+,t)? = f(X1,t) +

    (b' AW, + 4 AWs)

    Att fo(Xe,t)?(W'AW; + yp"AW2)

    where w’ and w”’ depend only on Y;, Xz, t, we find

    E[ f(Xipar,t+ At) |Yt°] — B[ f(Xipar,t+ At) |Yi]

    AW, +o” AW2) | (Wi)? = i [f(w,t) + fidt + fo(x,t 12.9)

    Its PAE Hie /[F(@,t) + Att fo(z,t)" (WAM, +" AWe)\(et — 61) (vivdt (AY — Gi At)pe(x) de. Using the first Eq.

    (5.10.10) and retaining here only the items of order not higher

    than At we get

    E[ f(Xtpat,t + At) |Yh] — B[ f(Xegar,t+ At) |Yi]

    = f sat) + (FF — ef )e@)de (vv) MaY - p.d9) a FEfo(x,t)? (b' AW

    AWE yi" + bp" AW2AW?

    yp''")

    x(vivPT) (er — G1)pr(w)de = Elf (Xe, tr — $1) 1Yi

    534

    Chapter 5 General Theory of Stochastic Systems

    x (div?) (AY — gidt)+/ fo(a,t)? (W AW AW? wT =o:

    +43" AW AWe pl" )\(bivd7 )~!(o1 — G1) pi (2) dz. Finally, retaining only the expectations of the random

    AW,AW?

    taking into consideration that EAW,

    variables AW,

    AW?

    =

    (5.10.18) AW?

    and

    0 in consequence of

    the independence of W2 and Wj and the second Eq.\ (5.10.11) ppt

    is ouyr

    we obtain from (5.10.18)

    E[f(Xtpat,t+ At) |YO" ]—

    E[f(Xisar,t+At) LY]

    = E[f(X+,t)(¢t — $f) |Vs] bivd fT) (AY — 61 At) (5.10.19) (vivdt )*(y1 — 61) |¥,] At. wot ,t) +E |fe(Xe Substituting expressions (5.10.16) and (5.10.19) into (5.10.15) and replacing the increments by the differentials we obtain the final formula (5.10.12).
    Agw(t, Tr),

    (I)

    (a=ail

    where JV is the random number of impulses acting on the network in the time in-

    terval (to, t); T,,..., I Aj,...,An

    the random

    the random instants of action of the impulses, and

    magnitudes

    of the impulses acting on the input of the

    circuit. The simplest way to find the n-dimensional

    distribution of the random func-

    tion X (t) is to evaluate the joint characteristic function of its values at the ins-

    tants t;,...,%¢,.

    Denoting by ¢ the largest of the numbers

    t),..., tn, t

    = max(ty abies teas and taking into account that w(t, Th) = Opatt) during the

    time interval (to, t). Then the probability of the events B,, will be determined by the formula t

    P(Bm) = Te"? 0 = [u@er (aves eareeetSe

    Sera)

    to

    To evaluate the conditional expectation relative to the event B,, in (II) we shall use again the formula of total expectation taking first the conditional expectation relative

    to the random variables Tj , ... , Tj. Then taking into account the independence of the random variables A; , ...

    Ty

    PRT

    , Am and their independence of the random variables

    we’ obtain

    exp

    ya Dn w(to, te) > Bm

    m

    = Eexp

    Nini 5 Sh w(tp, Tk) inal m

    =

    FE|E| exp

    bit, Ay a1

    p=

    n

    Apto (tps

    op |Thy

    vob

    p=l

    - eT] >» w(tp,Tr) | ,

    (IV)

    Problems

    545

    where ga(A) is the characteristic function of the random variable A, (k —eleD2t !): For evaluating the last expectation we shall find the joint conditional distribution of the instants of action of the impulses 7) , ... , Tj relative to the event B,,.

    us take arbitrary instants of time Tj ,...,

    Tm, T1


    .

    (VIII)

    p=1

    to

    Thus the multi-dimensional characteristic functions of the random function X (t) are

    determined by Oty

    MABE

    nyt,

    y song ten)

    = exp

    > Apw(tp, tT) | —Dl

    ajdt > =. Ax)

    pt (he

    5.2.

    An)

    \

    Ga to

    (Al a

    aa

    Derive the formulae for the stochastic Ité differentials of the functions of

    the standard Wiener process given in Table A.1.1. Obtain analogous formulae for the stochastic 0-differentials.

    Problems

    5.3.

    547

    Derive the formulae for the stochastic It6 differentials of the functions of

    a vector Wiener process given in Table A.1.2.

    Obtain analogous formulae for the

    stochastic 0-differentials. 5.4. Prove that the stochastic It6 differential of a scalar function

    UV= y(P) ofa

    Poisson process P is determined by the formula dU = [p(P+ 1) = y(P) |dP. Evaluate

    by

    this

    formula

    Table A.1.1.

    the

    stochastic

    differentials

    of

    the

    functions

    given

    in

    :

    5.5. Show that the scalar stochastic differential equations of the first order with

    the 6-differential (i) es= —e(1 + V2)Y + kV,

    GY = (14 Va)el¥) + kV, (iii) Y = go(Y,t) + o gi(Y,t)Vi may be transformed into ay naeam It6 equations:

    (i) Y = —e(1 — 20ev22 + Vo)Y + k(—20ev21 + Vi), (ii) Y = [1+ 2evo2! i%) +V2|y(Y)+ bial (Y)+Vi], (iii) Y = yo(Y,t) + 0 S Vpg Pp(Y, tp, (¥, t) + oYpi(Y,t)Vi, P,I=

    where y’ LE = 0y/ Ody, and

    V =

    noise with the intensity ’.

    Under what conditions to the It6 equations coincide in

    [Vi Van

    ies is the aan

    distributed white

    this case with the initial equations? 5.6.

    Let Vit) be such a m.s.

    n — | times differentiable random process that

    its (n — 1)*> m.s. derivative has the Ité differential

    dy" -1)

    sy ve

    vier) de aw,

    ny OT ta,

    (1)

    This relation is called the stochastic Ité differential equation of the n>} order relative to the process NAAN This equation may also be written in the form

    Veer where

    V =

    YG

    Vee

    OOo

    ee

    wy at)

    dW /dt is the white noise in the strict sense.

    em (LL)

    Prove that Eq.

    (II) is

    transformed into a standard form of the stochastic It6 differential equation for some n-dimensional vector random process. In the case of the Wiener process W(t) derive

    the formulae of the transition from the It6 equation of the n*) order to the equation in the 0-differential of the N*» order and vice versa. 5.7.

    Show that the scalar stochastic differential equations of the second order

    with the 0-differential:

    :

    (i) Y + 2wo(1 + V3)Y = we (1 => V2)Y = kV

    (i) ¥Y= go(¥,Y,t)+ 0 oY, Y,4)Vi f=

    548

    Chapter 5 General Theory of Stochastic Systems

    correspond to the following It6 equations: (i) Y+

    2£wo(1 — 20133£wWo + V3)Y + we(1 = 20V23EWo + Vo)Y

    = k(—20€113W0 + V1), Gy

    :

    = va ly).

    n

    t) ae 0

    in en

    et

    ;

    Pe

    Ula

    ;

    ts

    Lh=1 n

    +>

    °

    i



    91 (Y, Y, t)Vi, where Yi, GY > YAu bli. OCne (as

    5

    t)/OY, and V

    (=0

    =

    [Vi serail

    12 is the normally distributed white noise of the intensity VY. Find

    the conditions under which the It6 equations coincide with the initial equations.

    5.8. Show that the n-dimensional characteristic function 9, and the distribution density f,, at the random initial condition Yo and at the absence of the noises in Eq.

    (5.4.6) when b(Y, t) = 0 are determined by the following formula:

    Oni

    Oa LAl eos

    Ceol

    foe)

    a

    / eT P(t)

    +

    FEAR O(EnN) fo (im)din (n = iif 2, i. =)

    (1)

    —oo

    f2 RG

    ee

    een)

    = fo(y™*(t1, yn) ey (t1, 91) 16(yo — v(t2, p~*(t1, (y1)))

    (Yn — elms P(t (i) (n= 2,3,--9),

    (11)

    where Ya) = p(t, Yo) is the solution of a system which represents the function of the time and the random initial value Yo.

    5.9.

    Show

    that for the one-dimensional nonlinear system Y =

    —yy'tr

    (r

    = —1) with the random initial value Yo its n-dimensional density is determined at

    ly] < (ryti)~ 1/r by the following formula:

    fn(yi; o-

    55

    OA

    ile Ae

    Besa

    =

    fo(ywi(1

    xT] 6m — (1 = ryt yf) OP)

    me rytryt)/"

    (1 -

    rytyyy)~Atn/n)

    $ rtiys(1= rytayy)“OF9 PO",

    (=

    where fo(yo) is the density of the random initial value Yo. 5.10. Find the one-dimensional distribution of a stationary process in the conservative mechanical system:

    Problems

    with the random initial conditions where ralized coordinates;

    p = Aq =

    [pi ep

    549

    q = [q1 a odllLa Es is the vector of the geneyy ‘i is the vector of the generalized im-

    pulses; A is the generalized mass (constant); H is the Hamilton function (the total

    energy of asystem) equal to

    H = H(q, Pp) = p! A~+p/2+1(q) where II(q) is the

    potential energy of a system. Show that at II(q) = q Cq/2 the stationary distribu-

    tions of the coordinates and the impulses are determined by the densities (1 (q? Cq) and ~2 (p? A-!p) where (); and $2 to a great extent are the arbitrary functions (more exactly they both are expressed in terms of one arbitrary function). Show that

    the distributions of the doubled potential energy Yj = peo q and the doubled kinetic energy Y2 =

    TA

    tp are determined by f} (y1) = Ci yr

    2g, (y1)1(y1),

    fo(y2) = coy? !? G(yo)1(y2), where C, and C2 are the normalized constants. In the special case of the exponential functions ~; and $2, yi(z) = p2(zr) = e72/2 , the distributions of thedouble potential and the kinetic energy present

    y?-distribution with n degrees of freedom.

    Find the distribution of the total energy

    yi = (Y1 + Y2)/2. 5.11. Check that for the nonlinear system described by the second order equa-

    tion

    Ys 2eY +¢'(VY)=V

    (e> 0),

    (I)

    the one-dimensional density and the one-dimensional characteristic function of the strictly stationary process are determined by the formulae:

    fily, y) = cexp{—h?[ p(y) + 97/2]} (Ah? = 2ev), Hives) = A co

    e970) dy,

    (II) (IIT)

    where C is the normalizing constant, V is the stationary normally distributed white

    noise of the intensity V, p(y) is the nonlinear function, y'(y)

    Ov(y)/Oy.

    Consi-

    der the special cases: (a) y/(y) = aiy + azy"; (b) y’(y) = asgny; (c) ’(y) = @ailt Y: 5.12.

    Show that the one-dimensional density of the strictly stationary process

    in the nonlinear system

    Y =¥(Y)—2eaY +bV,

    e>0,

    (1)

    is determined by the formula:

    fily) = Vor/me"Y, a = 4e/v.

    (II)

    Here Y(y) = [i(y) --. Pn(y)]", Ova(y)/Oye = 0 (k = 1,...7), y" v(y) =

    yy ViVi k=1

    Vay

    =

    [Vi , ean

    Ven ips is the normally

    distributed statio-

    550

    Chapter 5 General Theory of Stochastic Systems

    nary white noise with the independent

    components

    of the same

    intensity V, b is

    the constant N X N-matrix, d@ is the constant N X M-matrix satisfying the condition

    a+a! = 2d07. 5.13.

    Show that the one-dimensional density of the strictly stationary process

    in the system

    Y =ay'(Y) +0, is determined

    by the formula:

    fily)

    =

    (I)

    ce2¥(y)/Y

    constant, ply) is the nonlinear function, y'(y) = (rectangular in general) matrix; a+

    at

    =

    Cc being

    the normalization

    Op(y)/dy,

    b is the constant

    @ is the constant matrix satisfying the condition

    2bb7 " V is the stationary’normally

    distributed white noise with the

    independent components of the same intensity V.

    5.14. Show that the one-dimensional distribution of the strictly stationary pro-

    cess Y(z) = [Y1 (t) Y2 (t) [# in the nonlinear stochastic system

    Y, = asgn(Y2— BYi), Y2=—-Nni+V,

    (1)

    filyr, yo) = cexp{—B(yz + 2a|y2 — Byr|)/v},

    (II)

    is determined by:

    where C is the normalizing constant V, is the stationary normally distributed white noise of the intensity Y. Find the corresponding characteristic function.

    5.15. Show that the one-dimensional distribution of the stationary process in the

    nonlinear mechanical system with n degrees of freedom whose behaviour is described by the equations in canonical variables

    q= OH /Op, p = —0H/0q

    is determined by formula:

    — 2ept+ b(q)V,

    b(q) = A(q)\/?

    (I)

    1985):

    fi(g,p)=ce"*",

    a = 4e/v,

    (II)

    where C is the normalization constant. Here q =

    [q1 Soll ie is the vector of gene-

    ralized coordinates,

    ie

    p=

    A(q)q =

    [pi -++DPn |

    pulses, A(q) is the generalized mass

    is the vector of generalized im-

    which represents symmetric positive definite

    matrix depending in general on g, H is Hamiltonian function (the total energy of the

    system) which is determined

    by the formula

    H =

    H(q,p)

    =

    p? A(q)~1p/2

    + II(q), II(q) being the potential energy of the system, 2€ is a specific coefficient

    of viscous friction, components

    V =

    [Vi prev i7, is the vector of generalized forces whose

    represent independent

    normally distributed stationary white noises of

    intensity /. Find the known Gibbs formula (Gibbs 1928) at b(q) vie

    Problems

    551

    5.16. Show that the function f; = f; (y; t) will be the solution of Eq. (5.6.56) with the normal white noise V if and only if the vector function

    @ = a(y, t) assumes

    such a presentation

    a(y,t) = ay(y,t) + a3(y, t),

    (1)

    that the function fj is the density of the invariant measure of the ordinary differential equation

    ;

    ¥ —a,(),1);

    (II)

    i.e. satisfies the condition

    ee

    I

    and the correspondent function as = ak(y, t) is determined by formula T

    SR

    5.17. Suppose that for Eq. (5.4.6) at a normal white noise V the presentation

    (1) of Problem 5.16 a density {44 =

    H1(y, t) of integral invariant (II) are known.

    Besides let

    (i) the matrix @ be invertible, det lem me U) (ii) the vector function Y; = 71(y, t) where

    lyst) = 078) ab —

    ited

    |esZin] ‘

    satisfies the condition of the curl absence: rot 7; = 0 or

    O71: /OY; => O71; /OYi

    (t,9 =

    “ siie\ce =P),

    (iii) the scalar function

    Fy(y, t) >

    [tw t)dy,

    is the first integral of Eq. (II), (iv) the condition of the normalization is fulfilled

    exc

    exp Fi(y,t)dy = 1.

    552

    Chapter 5 General Theory of Stochastic Systems

    Show that there exists the solution of Eq.

    (5.4.6)

    Y = Y (t) for which the

    one-dimensional density fj (y; t) is i determined by formula

    fi(yst) = p(y; t) exp Fi(y, t). 5.18.

    (I)

    Show that the function 1i(y, t) of Problem 5.17 is equal to fily; t) if

    and only if there exist such a matrix function Aj = Ay (y, t) that

    apy = (GA1/Oy)", where as =

    Ar + Ay = op,

    (1)

    (0) = al. Find the correspondent formulae for the transition function

    and the multi-dimensional densities.

    5.19.

    Consider Problem 5.17 for the system with the singular total matrix 7

    = byb?

    YSQYoY't),

    VS Py owl et)

    bol 2,0),

    (I)

    where Y’ and Y” are the s-dimensional vectors, 25 = p, Vo is the r-dimensional normal white noise of the intensity V9 = Vo (t), p is the dimension of the vector Y,

    Co.

    bo(Y’, NA t)Vo (t)bo(Y’, Ne it is the nonsingular matrix. Show that the

    one-dimensional distribution is given by the following formulae:

    fly syotha aly

    &) expt lye),

    (II)

    where

    Kiya .tv= /7} dy' dy",

    ny yt) =o9 Po, Oy

    Oni/Oy; = Oy; /Oy:,

    or

    Ge + ay (Om)

    (IIT)

    (IV)

    or + gyi (Pie)

    1_ 90 Oln fi Be

    =

    0,

    a Pi + Py.

    (V)

    (VI)

    5.20. Show that the solution of Problem 5.18 for the system with the singular total matrix 0 exist if and only if the conditions are fulfilled:

    P31 = (OB, /dy")", Bi +BY = oop,

    (I)

    sil Pia= P—p} 1) 41 is the density of the finite integral invariant of the equations

    =Q,Y"= P}. Show that fi(y’, y”,t)

    = wi(y’, y”, y).

    Problems

    5.21.

    553

    Show that for any vectors g, p and vector functions Q(q,P), P(q, Pp) of

    the same dimensions as q, Pp if there exists a function N(q) satisfying the condition gr

    or

    9g NO?)

    + Ci

    ee. q) = 0,

    then the one-dimensional distribution of the solution stochastic system

    i= Qa.) P= Pla”) 2(Hal + 4(Q)V, may be presented by formula:

    s al

    filq,p) =cN(q)e PY),

    W(H) = fean,

    pi4/y

    aD

    Here b(q) is any rectangular matrix, a(q) is any square matrix satisfying the condi-

    tion a(q) a a(q)? = 2b(q)b(q)?, Hus H(q,p) is the first integral of Eq. (I) at a(q) = b(q) =f, 2e(H) is coefficient of viscous friction,

    V = V(t) is the normal

    white noise with independent components and equal intensities VY. Find

    in special

    case

    € =

    filq,p) = cN(q)e"*#4?),

    const

    the following

    generalized

    Gibbs

    formula:

    a = 4e/v.

    5.22. Show that for the scalar nonstationary stochastic system

    Y = ao(t) +a(t)Y +0(Y,t)V, (Y,t) = Vc(t)Y, Y(to)=Yo, the solution of Eq.

    (1)

    (5.6.65) for the one-dimensional characteristic function has the

    form

    gi(A;t) = go(v(t, 4)~+) exp

    if v7, e(t, A)ao(r)ar

    Here V is the normal white noise of the intensity v(t),

    ae

    t

    re = |a(ryar t

    =

    Palas to

    - feceyio ar,

    :

    (II)

    504

    Chapter 5 General Theory of Stochastic Systems

    v(1,9(t,)) =



    -

    to

    :

    1/2 [ v(71)¢(11) exp -{ a(a)do | dry + y(t,Tr) to

    At

    the constant

    fucntions,

    to

    dg,

    @, Cc,

    Y =

    1 and

    the initial condition

    go(A)

    = (1 — 2iDo)~1/? we obtain

    gi(A;t) = [exp(at) — icd(exp(at) — 1)/2a]-***°

    x [1 — 2éDor(t)]7!/? exp (21)

    (III)

    where r(t) = A[exp(at) — icd(exp(at) — 1)/2a]~*. Give the generalization for the case of vector Eq. (I) when b(y,t) =

    diag [bi(y,t)... by (y, ae be(y, t)

    = oj(t)" y. =

    ae

    5.23.

    Show

    that under

    the conditions

    of the previous

    problem

    the multi-

    dimensional characteristic function is determined by the recurrent formula:

    Grill X15 eas pistes | rte )e=Gaer(Ans

    +07 (ta; Antal tite

    oe)

    Anza)

    pexp [/ W(7, (tn, An))ao(r)dr | ,

    (I)

    tn-1

    where 1) = 2,5,2.
    ve(t) [f(y — yeest) — fi(y;t)] . 5.25. Using formula (5.6.39) and Eq. (5.6.32) show that in the case of the general Poisson process W(t) and unit matrix b(y,t) = I the one-dimensional density is defined by the following integrodifferential Kolmogorov—Feller equation:

    ei: ee—FT Latust)Aaluiel +v(t)[(F*fid(y,t)-A(yst)],

    (

    where f * f; being the convolution of the density f(y) of the jumps magnitude of the Poisson process with the density f; (y; t):

    (FeAl) = f fy- Miran,

    (nt)

    Write a similar equation for fn(y1,---, Ynjti,---., tn). Obtain from Eq. the corresponding equation for the case where

    (1)

    W(t) represents the product of a

    constant h by the simple Poisson process.

    5.26. Using formula (5.6.41) for the function x (p; t) show that Eq. (5.6.32) is reduced to the following integrodinfferential generalized Kolmogoroy—Feller:

    Of es (y; = — oF we 9, [ey Ofi¥ 4] Wh

    ts re la [H(ost)ua(dben 0" ACwe)]} N

    + > x(t) k=.

    where

    / ge(y— 750, t)films t)dn — filyst) | , —

    {; (E, n, t) is the

    gu (ch O(y, DEX), i.e.

    0O

    density

    corresponding

    to

    the

    characteristic

    function

    the density of the random vector representing the product

    of the matrix (d(y, t) ck) by the random vector Y,; which is the random jump magnitude of the general Poisson process JPR (t).

    556

    Chapter 5 General Theory of Stochastic Systems

    5.27.

    Show that in the case where x(p; t) is given by (5.6.47) Eq.

    (5.6.32) is

    reduced to

    AED Fat iti + [{ituo t)c(a);t)— Ro

    zee(OF [ou too(toeu." nnso))} fi(y;t)+ a [b(y, t)e(x) f(y bret a)dz.

    CHAPTER 6 METHODS OF LINEAR STOCHASTIC SYSTEMS THEORY AND THEIR APPLICATIONS

    Chapter 6 is devoted to the methods of the linear stochastic systems described by the linear stochastic integral, differential, difference and other operator equations in the finite-dimensional and infinite-dimensional spaces. In Section 6.1 the spectral and correlation methods are stated. Section 6.2 contains the account of the methods of general stochastic systems theory based on the multi-dimensional distributions. In Sections 6.3 and 6.4 methods based the canonical representations are considered. Sections 6.5 and 6.6 present applications of linear stochastic systems theory to the linear shaping filters design, analytical modeling and optimal linear filtering based the online data processing.

    6.1. Spectral and Correlation Methods 6.1.1. Representation of State Vector

    Let us consider a linear system. form

    In this case Eq.

    (5.4.6) has the

    Y =aY t+ao+0V,

    (6.1.1)

    where a = a(t), ap = ao(t), b = b(t) may in the general case be the functions of time t, and V is a white noise whose intensity function of time t. Remark.

    vymay be a

    For finding the first and the second order moments of the random

    process Ve) determined by linear differential Eq. (6.1.1) there is no necessity that the white noise V be the white noise in the strict sense, i.e. the derivative of a process with the independent increments.

    It is sufficient that it be simply the white noise

    (i.e. the derivative of the process with the uncorrelated increments) as well as in the

    theory of Subsection 3.2.3. Therefore we shall consider V as an arbitrary white noise in this section. In other sections studying the multi-dimensional distributions of the

    process age we shall always consider V as the white noise in the strict sense.

    Using formula (1.2.8) determining the solution of Eq. (6.1.1) we get representation on the system state vector Y by the formula t

    WAG

    ultsto)¥o+ fult, VV to

    t

    (r)ar + fultsr)a0(r)ar, to

    (6.1.2)

    508

    Chapter 6 Methods of Linear Stochastic Systems Theory

    where u(t,7) ts the matric determined as a function of t by the homo-

    geneous differential equation du/dt = Gears ra &

    a(t)u and the initial condition

    6.1.2. Representations of First and Second Order Moments

    Taking into account that the expectation of a white noise is equal to zero we find by virtue of (6.1.2) the following formula for the expectation

    of the system state vector Y(t)



    m(t) = u(t, to)mo + ic T)ao(r)dr.

    (6.1.3)

    where mg is the expectation of the initial value Yo of the state vector Y. The covariance function of the state vector Y is determined by the formula K(ty ; ta) =

    u(ty ; to) Kou(te, to)*+

    min(t1,t2)

    +

    /

    u(t, T)b(r)v(r)b(r) 7 u(te, ars

    (6.1.4)

    to

    where Ko is the covariance matrix of the initial value Yo of the state vector Y. While deriving this formula we remembered that the initial state of the system Yo is independent of the white noise V(t) at ¢ > to and that

    u(t,7)

    = 0 at r >t. Due to this the upper limit of integration is equal

    to min(t;,t2). Besides that we took into account that the coefficients of the equations of real systems are always real while the elements of the

    matrix u(t,7) may be complex in this case. After determining the expectation and the covariance function of the state vector Y we may find its second order moment which in this case of real-valued functions m(t) and K(t,,t2) gives

    T'(t1,t2) = K (ti, t2)+ m(t1)m(te)?.

    (6.1.5)

    6.1.3. Differential Equation for Expectation

    In practical problems it is usually sufficient to find probabilistic characteristics of the state vector of a system Y at every given instant

    6.1. Spectralkkand Correlation Methods

    559

    t (determined by the one-dimensional distribution), i.e. only the values K(t,t) = K(t) and [(t,t) = I(t) of the covariance function and the second order moment. In other words, it is sufficient to find the expectation, the covariance matrix and the second order moment of the state

    vector Y at every instant ¢t. It goes without saying the all these vari-

    ables for a linear system may be determined by formulae (6.1.3)—(6.1.5) at t) = t2 = ¢t. But in the case of a linear system they may be calculated more simply, ‘namely by the integration of the corresponding linear differential equations. > In order to derive the differential equation for the expectation of the vector

    Y we differentiate formula (6.1.3)z t

    m(t) = ur(t, to)mo + [ut T)ao(T)dr + ao(t) to t

    = a(t) | u(t, to)mo + Ju(t, r)ao(r)ar)

    + ao(t).

    to

    But

    the expression in the brackets

    is equal to mM

    =

    m(t) by virtue of (6.1.3).

    Consequently,

    m=am+ay.4

    (6.1.6)

    Thus we obtained the differential equation for the expectation of the vector Y. Integrating Eq. (6.1.6) with the initial condition m(to) = mo we may calculate the expectation of the random vector Y tn stochastic linear system (6.1.1). 6.1.4. Differential Equation for Covariance Matriz > In order to derive the equation for the covariance matrix K(t) of the vector

    Y we put in (6.1.4) t) = tz =T: K(t) =

    K(t,t) =

    u(t, to) Kou(t, to)*

    t

    a /u(t, r)b(r)v(r)b(r)? u(t, 7) to

    ate

    (6.1.7)

    560

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Differentiating this formula with respect to t we obtain

    K(t) = u(t, to)Kou(t, to)” + [ee

    + fe

    ome

    to)” pneryPult, 7)" ar + u(t, to) Kour(t,

    poryu (or

    ult, rar 4 b(t)v(t)b(t)” ,

    to

    or as U;(t, T) = a(t)u(t, 7), ur(t, 7)* = u(t, T)*a(t)? (Subsection 1.2.2),

    K(t) = a(t) u(tto) out to) + fut, 2)0(7)0(7)0()P u(t, 7)*ar|+ [ultto)

    to)

    + /u(t, T)b(r)v(r)b(r)7 u(t, rar] a(t)? + b(t)v(t)b(t)? . to

    But the expression in the brackets by virtue of (6.1.7) is equal to

    K =

    K(t).

    Consequently,

    K =aK+Ka™+bvb". 4

    (6.1.8)

    Thus we obtained the differential equation for the covariance matrix of the value of the random vector Y at a given t. Integrating Eq. (6.1.8)

    with the initial condition K(to)

    = Ko we may calculate the covariance

    matrix of the random vector Y in stochastic linear system (6.1.1). 6.1.5. Differential Equation for Second Order Moment > For deriving the differential equation for the initial second order moment

    I(t) of the vector Y we put in (6.1.5) t} =

    tp =

    t. Then we shall have I(t)

    = K(t) ar m(t)m(t)? or, omitting the argument f

    T=K+mm’'.

    (6.1.9)

    6.1. Spectral and Correlation Methods

    561

    Differentiating this formula we find

    T=K+mm? + mm? . Substituting here the expressions

    ™ and K from Eqs. (6.1.6), (6.1.8) we obtain

    T = aK + Ka? + byb? + amm? + apm? + mm? a? + mai , or using formula (6.1.9)

    T= al +Ta? + bvb? + agm? + ma? . 4

    (6.1.10)

    - va

    Integrating Eq.

    (6.1.10) after Eq.

    (6.1.6) which determines the

    expectation m with the initial condition T(to) may calculate the initial second order moment

    = To = Ko + mom we of the random vector Y

    in stochastic linear system (6.1.1). Remark. (6.1.8) or Eqs.

    It is evident that the direct numerical integration of Eqs. (6.1.6), (6.1.6), (6.1.10) is much simpler than the calculation of m

    and K

    or m and I by formulae (6.1.3)-(6.1.5) with preceding multiple integration of the homogeneous equation U = aU with the initial condition u(r, im) = I at various T necessary for finding u(t, Ts) as a function of two variables.

    6.1.6. Differential Equation for Covariance Function > Starting from formula (6.1.4) we shall derive the equation for the covariance function iG is ; to) of the random process a(t) considered as a function tg at any

    fixed t;. For this purpose we rewrite formula (6.1.4) for the case tj < tg:

    K(t1,t2) = u(ti, to) Kou(te, to)* + [ee

    DoH

    ult, edie

    to

    Differentiating the is formula with respect to tg we find

    OK e (t,t e 2

    = u(t1, to) Kout, (te , to)i

    + /ss(tah BCE) BT)? ty,(tgpNod = 1(#) 440) Kotte, to)"a7(42) to

    562

    Chapter 6 Methods of Linear Stochastic Systems Theory

    + /u(t, 7)b(r)v(r)b(T)? u(te, r)*a! (t2)dr,

    -

    OK (t1, te)/Ote =

    K (ty, t2)a(t2)? Ue

    UO

    (6.1.11)

    The initial condition for this equation has the form K(ty , t) = K(ty is J

    Integrating Eq.

    (6.1.11) at various

    values

    of t; we

    number of sections of the covariance function K(t,,t2) obtain K(t,t2) at tz t.

    Kips) = ate

    a

    To

    (6.1.12)

    Summarizing the results in Subsections 6.1.1—6.1.6 we obtain.

    Theorem 6.1.1. Eqs. (6.1.6), (6.1.8), (6.1.10)-(6.1.12) with the corresponding initial conditions determine completely the expectation, the second order moment and the covariance function of the process Y(t) in stochastic linear differential system defined by Eq. (6.1.1). Remark.

    Theorem 6.1.1 is, basic for different correlation methods

    stochastic differential systems.

    The white noise in Eq.

    of linear

    (6.1.1) is understood in the

    wie sense (Subsection 3.2.3).

    Example

    6.1.1. For the random process Y(t) determined by the equa-

    tion

    Y+aY

    =V2DaV,

    (I)

    V being a white noise of unit intensity, Eqs. (6.1.6), (6.1.8), (6.1.10), (6.1.11) have the form

    My =—am,, Dy =—20Dy+2Da, Ky = Dy, :

    ly

    —2aTy

    + 2Da,

    The condition Ky, (ty ; Ti) =

    OKy(ti, ta) /Ote =

    —aKy(t1,

    I t2) c

    (

    )

    1), (ti) will serve as the initial condition for the last

    equation.

    Example

    6.1.2. For the vector process which is determined by the dif-

    ferential equations

    Yi =Yo2+qV, Yo =—b?Y; — 2aY2 + qV,

    (I)

    6.1. Spectral and Correlation Methods

    563

    where V is a white noise of unit intensity, Eqs. (6.1.6), (6.1.8), (6.1.10) and (6.1.11) have the form my oe

    er

    0

    =

    1

    mg, 7

    Mo

    bees Pater:

    7

    =

    —b?m, 0

    a

    2am,

    —b?

    )

    71

    is Zoo |+ | | teeed

    0

    1

    0

    —b?

    q

    OK a (th, ta) _— (ie aa oR) = K (t,t) 9 ae

    Example

    where

    (IIT)

    te[ Sk)? Zh [s]ieor om s

    tion

    (II)

    (Vv)

    6.1.3. For the random process Y (t) determined by the equa-

    Y = (w+ 2at)Y — (4 2at)(a + bt) + b+ be V , V is a white noise of the intensity

    7 =

    2¢ Eqs.

    (1)

    (6.1.6), (6.1.8), (6.1.10),

    (6.1.11) have the form

    My = (w+ 2at)my, — (w+ 2at)(a + bt) +b,

    (II)

    Dat a 2at)Dy 2pkesg Ky = Dy, (IIT) Ty =2(u+ 2at)Dy + 2b°te?4* — 2(u + at)(a + bt)my + 2bm, , (IV) OK, (ty, te)/Ote = (4 + 2at) Ky(t1, te) (V)

    6.1.7. Stationary Processes in Stationary Linear Systems

    We consider a stable stationary linear system (6.1.1) under the influence of a stationary white noise with constant intensity v. In this case a, do, b are constant, the function u(t, 7) depends only on the difference

    of the arguments, u(t,7) = w(t — 7), and formulae (6.1.3), (6.1.7) and (6.1.4) at to + —oo take the form (oe)

    n= J e@a0d,

    (6.1.13)

    0 lore)

    K= [wees w(eyrae,

    (6.1.14)

    0

    ka eels

    (tit tint = [re + 1)bvb? w(é)*dé (r > 0). (6.1.15)

    564

    Chapter 6 Methods of Linear Stochastic Systems Theory

    It is clear from these formulae that under the influence of a stationary white noise the wide sense stationary process settles in course of time in a stable stationary linear differential system. It is also clear from these formulae with the account of (1.1.14) that the condition of stability of a system is not only sufficient but also necessary for the existence of a stationary process. As m and K are constant in a steady state we get linear algebraic equations for m and K,

    am+a)=0, aK + Ka" +bvb™ =0

    (6.1.16)

    putting in Eqs. (6.1.6), (6.1.8) mn =0, K =0. If the initial values mp and Ko and K satisfy these equations, then

    Eqs. (6.1.6), (6.1.8) have the evident solution: m = mo, K = Ko. In this case at any to the process Y(t) is wide sense stationary. To find the covariance function k(r) of the stationary process Y(t) in a linear system we transpose Eq. (6.1.11) after changing the order of the variables ¢; and ft, in it.

    Then remembering that the matrix

    a(t) = a is constant and that K(t2,t,)’ = K(t:,t2) we shall have at SS

    UG: OKz(t1, te) /Ote =

    aK (ty, to).

    Putting

    here

    K (ty, t2) =

    k(r),

    7 =t, —t2, tg =t and performing the change of variables t} = t +7 we obtain the equation

    dkla) (das ak(t)igete0..

    (6.1.17)

    This ordinary linear differential equation with the initial condition k(0) = K determines k(r) at r > 0. At r < 0 according to the property of the covariance function of a stationary random process it is determined

    by the formula k(r) = k(—r)?. Corollary 6.1.1. In asymptotically stable linear stationary system defined by Eq. (6.1.1) under the influence of the stationary white noise the wide sense stationary process settles with the expectation, the covariance matriz and the covariance function given by Eqs. (6.1.16), (6.1.17). Example

    6.1.4. Under the conditions of Example 6.1.1 the variance Dy

    and the covariance function ky (r ) of the stationary process Y (t) are determined by the equations

    —aD,-eD =,

    7

    0.

    For the existence of a stationary process in the system the condition

    @ >

    necessary and sufficient.

    dky(7)/ar —

    cb, (t),

    &,(0) =

    (Q) is

    6.1. Spectral and Correlation Methods

    Example

    K

    6.1.5.

    For Example

    6.1.2 at a@ >

    565

    (0 the covariance matrix

    and the covariance function k(r) of the stationary process are defined by the

    equations

    Dl

    AP

    one

    tle:

    11

    a

    [ie a |R+K[Y Dn |+E |lead =o. dk(r)

    _

    dr

    “A

    0

    1

    tsb? | —2a

    [k@), 7 >0, KO)=k.

    Notice that the obtained results are easily extended to nonstation-

    ary linear systems (6.1.1) at constant a, b and vy and an arbitrary time

    function a(t). In this case the equations obtained for K and k(r) remain valid, and m represents a function of time determined by Eq.

    (6.1.6).

    The process Y(t) is covariance stationary in the system for which m is determined by Eq. (6.1.6) at any initial condition, and K and k(r) are determined by the method outlined.

    6.1.8. Spectral Methods

    Let us consider a stable stationary linear system with the transfer functions ®(s) whose input is a stationary random process X(t) with the spectral density s,(w). We shall prove that the output Y(t) of this system is a steady state, i.e. under the infinitely long action of the input represents a stationary random function. > We represent the random process X (t) by the spectral representation (Subsection 4.8.6)

    fore)

    X(t) = mz (t) + / e*V(w)dw .

    (6.1.18)

    On the basis of the superposition principle to this spectral representation it corre-

    sponds the spectral representation of the output

    Y(t) = my(t) + / e”' B(iw)V (w)dw .

    We find the covariance function of the random function Y (t):

    roo)

    Ky fists) = / eft ciwia(iw)s,(w)B(iw)*dw —oo

    (6.1.19)

    566

    Chapter 6 Methods of Linear Stochastic Systems Theory [o.0)

    = / ei (t1-t2) 6 (iw) s_(w)B(iw)*dw .

    (6.1.20)

    —OO

    Hence it is clear that the output of the system in a steady state represents a sta-

    tionary random function of time. Let us find its spectral density. Subsection 4.8.8 at (o)

    ky)

    w =

    Using formulae of

    A:

    oo

    J slwyelerae,

    Sy(w) = = i k.(aje

    2dr,

    (6.1.21)

    we come to the conclusion that the spectral density of the output of the system Y (t) is determined by

    Sy(w) = B(tw)s,(w)P(iw)*.

    (6.1.22)

    Now we shall find the cross-covariance function of the input and the

    output of the system. According to (6.1.8), (6.1.9) we find Kt)

    = / etw(ti-ta) (Ww) @(iw)* dw .

    (6.1.23)

    —oco

    Hence it is clear that the input and the output in a steady state are crossstationary and the cross-spectral density of the input and the output of the system is determined by

    Sry (W) = Sa(WlOw) Remark.

    (6.1.24)

    Formulae (6.1.22)—(6.1.24) are valid only for stable stationary

    systems functioning in a steady state.

    Practically the formulae are applicable when

    the time of input action exceeds the time of the transient process. In should be noticed that if a system is described by a differential equation and consequently, its transfer

    function is rational, then the output may be also a stationary random function for any time of system functioning with special initial conditions.

    initial value of the output Y (to) =

    Namely, the random

    Yo should be chosen in such a way that the

    covariance matrix of the random vector Z(t) =

    aus e@ea hl ie be independent

    of t. For this it is sufficient to take the random initial value Yo whose covariance matrix and cross-covariance matrix are equal to co

    Ky, = &0) = / D(tw)s_(w)P(iw)* dw , —0oco

    Ke

    (Ska

    fos

    Oe / Sz (w)@(iw)* dw .

    (6.1.25)

    6.1. Spectral and Correlation Methods

    567

    But the appearance of these initial conditions in real systems is practically negligible.

    So we proved the following proposition.

    Theorem 6.1.2. The output Y(t) of asympotically stable stationary linear system with transfer function ®(s) under the influence of the wide sense stationary input X(t) in a steady state ts the wide sense stationary process with the spectral and correlation (6.1.21)-(6.1.24).

    characteristics given by Eqs.

    > For calculating the variances and the covariances of the output components of a stable stationary system functioning in a steady state under the action of a

    stationary input at every given instant f it is sufficient to put in formula (6.1.20)

    t; = tg = t. Then we get the following formula for the covariance matrix of the value of the output at the instant f: co

    Fie be) = / Bia lee oan \ides

    (6.1.26)

    —oo

    In practical problems, to find the variances and the covariances of the output

    components by formula (6.1.26) we have to calculate the integrals of rational functions. For calculating such integrals it is convenient to use the formula

    i)bo (iw)??? + by (iw)?"-3 + --- + bon—1 (iw) + ban—2 4

    |ag(iw)” + ay(iw)?-1 + +--+ an_1(iw) + ap?

    SI flee

    6.1.27 (6.1.27)

    ——D

    Here

    Naess

    (oil

    “Gie

    “Gao

    Ara

    Coie

    COD)

    ewe

    Cn

    Cni

    Cn2

    +--+

    Cnn

    Dn

    =

    bo

    OW

    oon

    ae

    bo

    COD

    anne

    COR

    bon—2

    Cn2

    +++

    Cnn

    where Cypg = G2p—g at 0 < 2p— q < Nj Cypg = Vat Ip—q< Vorlp—q>n. For using formula (6.1.27) it is necessary to represent the numerator in (6.1.26) in the form of a polynomial relative to tw and the denominator as the square of the modulus of the polynomial relative to 2W with positive coefficients dg, @1 , ... , Gn. We may also use formula (6.1.27) for calculating the covariances of the input components with the output components according to the formula

    Toy = eae0

    / S$,(w)®(iw)* dw. < 12,6;

    (6.1.28)

    568

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Remark.

    The solution of the linear differential equation for the covariance

    function k(r ) of a stable system represents a linear combination of functions containing damped exponents.

    To each such function it corresponds a rational spectral

    density. Thus the stationary process in the stationary linear system (6.1.1) possesses the rational spectral density. We may also come to this conclusion if we use formula

    (1.2.57) at s = iw: ®(iw) = —(a — wwI)~'d. Example

    6.1.6. Find the variance of a random function whose spectral

    density is determined by the formula Sy (w) =

    Da/n(a?

    + w?).

    According to

    formula (6.1.26) we find

    Dele xpi eeelt eeeBie Doe ee y oT / a2 + w2 1 (64

    (G9)

    up eee =ad

    a

    (a9)

    / iw + a|’

    —0oOo

    In this case

    2 = 1, G9 =

    h we have K(I,h)

    SK (hae). In the case of stationary discrete linear system (6.1.29) the matrices b; and the vector ag; do not depend on the discrete time I: a; = a, b; = 6, ao} = a and Eq. (6.1.29) takes the form Yah

    =

    aY;

    + a9 +

    bY, .

    i

    (6.1.33)

    The variables V; have one and the same distribution, and consequently one and the same covariance matrix G. If there exists the stationary

    discrete random process {Y;} in Eq. = m; = mj41

    (6.1.33) then its expectation m

    and covariance matrix K = K; = Kj4, are constant.

    By

    virtue of Eqs. (6.1.30) and (6.1.31) they satisfy the algebraic equations:

    m=am+a),

    K=aKa! +6G0' .

    (6.1.34)

    It is evident that these equations have the unique solution if the matrix

    (I — a) is convertible. Thus we proved the following statements.

    Theorem 6.1.3. Recurrent Eqs. (6.1.30)-(6.1.32) with the corresponding initial conditions determine completely the expectation, the covariance matriz and the covariance function of the process Y; in stochas-

    tic discrete linear system (6.1.29). Corollary 6.1.3. In asympotically stable linear stationary system defined by Eq. (6.1.33) under the influence of the stationary white noise the wide sense stationary process settles with the expectation, the covariance matriz and the covariance function given by Eqs. (6.1.34).

    Now we consider the more general equation: A(V)¥,

    =c+ B(V)V. ,

    (6.1.35)

    where A(V) and B(V) are the matrix polynomials over the shift operator V: pri

    AVY STVe Saya r=0

    P

    BV y= Ope s=—N

    6.1. Spectral and Correlation Methods

    yal

    (the polynomials B(V) contain also the negative degrees of V if N > 0). According to the spectral theory and under the conditions of asymptotically stable linear system the spectral density s,(w) for the stationary sequence {Y;} is expressed in terms of the spectral density s,(w) of the

    sequence {V;} and the transfer function ®(z) of Eq. (6.1.35):

    Sy(w) = B(e™)sy(w)B(e™)* .

    (6.1.36)

    Consequently, 20

    ies }syed Preteen wyeteteyt da

    (6.1.37)

    0

    As the variables V; are independent then the spectral density s,(w) of the sequence {V;} is constant: s,(w) = G/2z where G is the covariance matrix of the variables V;. Substituting this expression into (6.1.37) we get 20

    ss az| Be

    Gaede

    (6.1.38)

    0 After the change of the variables z = e formula (6.1.38) will take the form ; “

    as K =~

    z P9 U(2)GH(e7)*=, *\* —

    (6.1.39) 1.

    where the unit circle |z| = 1 with the curve counterlock-wise serves as the integration counter. Later on it is expedient to pass to a new variable for the calculation z—1 Ww

    eaten

    tg—.

    ©

    Then the unit circle of the plane of the variable z turns into the real axis

    of the variable A and formula (6.1.39) will take the form

    eaters, Ls Bay “asked k=> f o(5) co (+) ae

    (6.1.40)

    —oo

    As the transfer function ®(z) of Eq. (6.1.35) is the rational function of z then the integrand function in (6.1.40) will be the rational function

    5t2

    Chapter 6 Methods of Linear Stochastic Systems Theory

    d. This fact gives the opportunity to use for the calculations of the covariance matrix K the known formulae for the integrals over the rational

    functions (Subsection 6.1.8). In some cases it is also expedient to find the covariance function

    k(1) = E(¥, — mp) (Vidi — mh 41)

    (6.1.41)

    of the stationary sequence {Y;}. For this purpose we use the formulae of the connection between the spectral-correlation characteristics (Subsec-

    tion 4.8.6): 20

    ,

    k(l) = ae| Be Goer ue

    ds,

    (6.1.42)

    0

    Then the correspondent formulae with the variables of the integration z and A have the form 20

    ky (I) = 5 ib(e*”)GH(e)*e"™ du ,

    (6.1.43)

    0

    pai f Dunes (5) or(eaettalh Ge (5) 1+4éd\ aya (144d)?! n(v=—

    144)

    Formula (6.1.44) is more convenient for the calculations than formulae (6.1.42) and (6.1.43) as the integrand in it is rational in consequence of which we may use the known formula (6.1.27) for the integrals over the rational functions for the calculation of the elements of the matrix k/(I). Remark.

    As regards to the constant expectation of the stationary sequence

    Algats it is defined by the trivial equation: A(1)my the unique solution My

    the polynomial

    = c. This equation has always

    = A(daic as 1 cannot be the root of the determinant of

    |A(z)| in the case of the asymptotically stable system (6.1.35).

    Thus we have the statement.

    Theorem 6.1.4. In asympotically stable discrete linear system (6.1.35) under the influence of the stationary white noise the wide sense stationary process settles with the spectral and the correlation characte-

    ristics defined by Eqs. (6.1.36), (6.1.38)-(6.1.40), (6.1.42)-(6.1.44).

    6.1. Spectral and Correlation Methods

    6.1.10.

    573

    Correlation Theory of Linear Transformations

    Let the expectation m,(t) and the covariance function K,(t,t’) of the input X(t) be given. Let us find my(s) and K,(s, s’) of the output Y(s) for the linear transformation

    Y(s) = AX(t),

    (6.1.45)

    where A is an arbitrary linear operator. This problem is easily solved if we assume that the operation of the expectation and the operator A are commutative. It is often valid for linear operators in the problems of practice. ; According to the properties of the first and the second order moments we have

    EY(s) = E[AX(t)] = AE(X(t),

    BL Y°(s)¥%(s') |= B |AX°(t) AX) |

    (6.1.46)

    —E |AvAeeX°(t) XD |Aer |x°(t)X°@) | m,(s) = Am,(t),

    "Ky(8,8 ) = AvAy Kolt,t) = AvAiKe(t,t).

    (6.1.47)

    (6.1.48)

    Here the index at the operator A indicates that this operator acts upon the function of a given argument at the fixed values of all other variables.

    It is evident that formula (6.1.46) is also valid for the initial moment of the second order: Ty(s, 5’) = Ay AT c(t, t’) = Ay Atl e(t,t’).

    Thus formulae (6.1.44) and (6.1.48) are at the basis of the correlation theory of the linear transformations of random functions. Hence the correspondent formulae of Subsection 6.1.1-6.1.9 are obtained as the special case for the concrete operators A. The general solution of this problem is given in Sections 6.3 and 6.4. 6.1.11. Spectral and Correlation Methods for Stochastic Infinite-Dimensional Linear Systems

    Let us consider a linear stochastic differential system in a separable

    H-space Y (Subsection 5.9.1): dY = (aY + ao)at + bdW.

    (6.1.49)

    574

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Using formulae

    of Subsection

    6.1.10 we

    get for expectation

    m,

    m

    = EY°®, covariance operator K, KA = E[Y(A,Y?)], \ € Y where Y,° = Y; —m of the process Y; = Y(t) at a given instant ¢ the following equations:

    m=am-+ao,

    KX =aKi+ K)a* +A

    (6.1.50)

    (c = bvd*).

    For covariance operator K(t;,t2)A2 = E DAs (Aa, ae

    (6.1.51) at t; and te

    we have

    OK i(t1, to)Ao =e Pa Ao a) ate 2 ts

    99)(621.52)

    2

    The following conditions serve as the initial conditions for m(t), K(t)A and K(t1,¢2)A2 in operator equations (6.1.50)—(6.1.51):

    mo = m(to), K(to)A= E[YLA;Y2)] , K(t,t)A2 = 2 [YP (2, YP)] (6.1.53) For determining the processes stationary in the wide sense in (6.1.49) at a, ao, b and o independent of time it is sufficient to put in Eqs.

    (6.1.50) and (6.1.51) rn = 0 and K =0. As a result we obtain the equations:

    am+a,9=0,

    aKA+ KiXa*+oA=0.

    (6.1.54)

    For the covariance operator of the stationary process

    K(t1,t2)A2 = E[Y,2(A2, ¥2)] =k(r)A2 (ti — te = 7) Eq. (6.1.52) takes the form

    dk(r) dr

    = ak(r).

    (6.1.55)

    In the correspondence with the formulae of Subsection 4.8.6 the following equation for the spectral operator sy(w) holds:

    Sy(w)Az = B(iw)s,(w)P(iw)* ,

    (6.1.56)

    where sy = sy(w) is a spectral operator of the white noise V, ®(iw) = —(a—iwl)~'b. Thus we get the following propositions.

    6.2. Methods of General Linear Stochastic Systems Theory

    Theorem

    6.1.5.

    Eqs.

    575

    (6.1.50)-(6.1.52) with initial conditions

    (6.1.53) determine completely the expectation and covariance operators in stochastic linear differential system defined by Bq. (6.1.40). Corolary 6.1.5. In asymptotically stable stochastic linear differential system (6.1.49) under the influence stationary white noise the wide sense stationary process settles with the spectral and the correlation operators defined by Eqs. (6.1.54)-(6.1.56). Example

    6.1.8.

    Under the conditions of Example 5.9.1 Eqs.

    and (6.1.51) at constant c, h, v for mj; =

    m;(x1, £2, t), Ki; =

    (6.1.50)

    Kal 2i, £20)

    have the following form:

    OK

    pe

    Si

    Om,

    ete

    Koi + Ki2,

    age

    OK



    Ot

    Om,

    = Kapa?

    seit = ¢?

    6.2. Methods

    OK 12

    ak

    ,0?m,

    WO

    Ko4+¢?

    Ohad

    One

    07 Ky,

    Oa?

    +hkKi2,

    +2hKo1,

    0?K.

    1

    (1)

    a

    2

    + 4hKoo+v.

    (II)

    of General Linear Stochastic Systems Theory

    6.2.1. Equations for Characteristic Functions

    In the case of a linear system (6.1.1) where V being the white noise (in the strict sense) Eqs. (5.6.48), (5.6.49) as a(0/i0X,t) = —ia(t)O/OA + ao(t) take the form

    0 0 oa - a

    +f [irT ao + x(b7 d; t)|1,

    =f = Malta) (>> aun) aos [iA7ao+ x(b7 A;t)] g.

    (6.2.4)

    kth

    The standard algorithm for integrating equations of the first order in partial derivatives linear relative to the derivatives consists of the following operations. (i) The integration of the corresponding set of ordinary differential equations which has in this case the form

    dt Sais addy 1 P

    - Yann (=i

    ee

    a AdX» P

    ean

    dg ee iA? bT dst

    AP UE

    AD

    ees) (

    eal

    This set of equations is obtained from the original equation in partial derivatives in the following way: the differential of each of the variables t, A, ,---, Ap is divided by the coefficient of the derivative of the function g with respect to this variable, the differential of the function is divided by the right-hand side of the equation and the ratios thus obtained are equalized to one another. (ii) The general solution of Eqs. (6.2.5) is represented in the form

    CHEATS = Apso) = CE (k= eee Poe ie

    (6.2.6)

    where c;, ... , Cp, Cp4i1 are arbitrary constants (i.e. in the form of p+ 1 independent first integrals of these equations). (ili) An arbitrary function f(x1, ..., zp) of p variables is taken and the equation is written Gn+i(G AT ese; Ap,9)

    6.2. Methods of General Linear Stochastic Systems Theory

    nif

    (qoy (tpg tar ahr py) (>

    og,

    577

    (ty AgiyrenyAghg))%

    (6.2.7)

    The solution of this equation relative to g depending on the arbitrary function f represents a complete integral of the equation in partial derivatives. > Let us apply this algorithm for the solution of Eq.

    (6.2.1).

    In this case

    the coefficients of the derivatives are linear function of the variables A; , ... independent

    , Ap

    of the unknown function g, and the right-hand side of the equation is

    linear relative to g. Rewriting the set of Eqs. (6.2.5) in the Cauchy form we shall have

    dX

    =

    sa

    meg

    he

    hy cy).

    (6.2.8)

    d

    = = [iTap + x(07A;t)] 9.

    (6.2.9)

    The first p of these equations form a set of linear homogeneous to Ay hes

    equations relative

    ibe Ap whose matrix of coefficients represents the transposed matrix of the

    coefficients of the original Eq. (6.1.1) with the sign reversed. In matrix form this set of equations is

    afd

    oe

    (6.2.10)

    In order to find the general solution of this equation we denote by u(t, T )the solution of the homogeneous equation du/dt =

    au with the initial condition u(r, T) =

    ls

    t > T > to. Then the solution v(t, T) of the adjoint equation dv/dt = —aly with

    the initial condition v(T, 7) = I is determined by formula Ot, TF) = us; mut (Subsection 1.2.2). Consequently, the general solution of Eq. (6.2.10) is determined by formula

    Nea(tw):ey ‘

    ;

    A



    where C is an arbitrary constant p-dimensional vector,

    (6.2.11) C =

    [cy x Goo

    ae

    F

    , and @ is

    an arbitrary time instant within the interval [to, 1): Substituting expression (6.2.11) into Eq. (6.2.9) we transform it to the form

    dg/dt = [ic’ u(t, ~)~'ao(t) + x(b(t)? u(t, a)1%;t)] g.

    (6.2.12)

    The general solution of Eq. (6.2.12) has the form

    9 = Cp41 XP

    / [ic? u(r, 2)*ao(7) + x(b(r) u(r, a)""Fe;7)]dro, Qa

    578

    Chapter 6 Methods

    of Linear Stochastic Systems

    where Cp41 is an arbitrary scalar constant.

    pression of ¢ from (6.2.11) to Cp41

    Theory

    Substituting into this equation the ex-

    ¢c= u(t, a)? ), and solving the obtained equation relative

    we get t

    Cp+i

    i

    i= 9 eXP4i—

    [iAT u(t, w)u(r, «)~*a0(7)

    /

    + x(b(r)P u(r, a)

    (6.2.13)

    ule, a)Ter)|dr :

    But according to Subsection 1.2.2 for anya

    s(, Ge

    44: Sa

    As

    (6.2231)

    s=1

    Analogously we find a difference equation of the n-dimensional characteristic function: Jiypntae

    ="exD (11, doe. | BE; Anion, As solving Eq.

    toils

    g) Wee Ko

    kok

    Ave)

    es Ant, deen) | 102-38)

    (6.2.38) at the initial conditions (5.3.78) we get the

    recurrent formula for the mu!ti-dimensional distributions:

    Di ipech a Ab

    iesaeesssa)

    sa ae Oleg A Bierce, id sierotok cde

    T ieT 980 T boot emt)

    x exp {id aoz, }he, (bh An)

    Kiel x I] exp {id7 ag, ...ds4140s}hs(br ay,

    .. ae) :

    (6.2.39)

    si

    Thus we have the following statements.

    Theorem

    6.2.3.

    Finite formula (6.2.37) and recurrent formula

    (6.2.39) give the explicit expressions for all multi-dimensional characteristic functions of discrete linear system (6.1.29). Corollary

    6.2.3.

    At the normal distributions of the variables Vy,

    and Y; Eqs. (6.1.37)—(6.1.39) completely determine all multi-dimensional normal distributions of the sequence {Y;}.

    6.2. Methods of General Linear Stochastic Systems Theory

    591

    Theorem 6.2.4. If there exists a stationary process in a stationary discrete system then it is stationary in the strict sense. The one-dimensional distribution is determined by Eqs. (5.3.76) at gi(A;t) =

    gi(A).

    6.2.7. Theory

    For

    Methods

    the

    of Stochastic

    stochastic

    linear

    Infinite-Dimensional

    differential

    system

    Linear

    in the

    Systems

    separable

    H-space y:

    dY = (aY + ao)dt + bdW

    (6.2.40)

    equation for the one-dimensional characteristic functional of Y = Y;

    g = 91(\;t) = Eexp [2A, Y(t)]

    (6.2.41)

    may be presented in the following form:

    = eriotir a eco 500,03)|Re

    (6.2.42)

    Recall that in (6.2.40)-(6.2.42) A is an element of the space Y, W = W(t) is the Wiener process with the values in the separable H-space W, the covariance operator whose values at each ¢ are determined by formula (5.9.3) and its intensity v at each t is the trace-class self-adjoint operator; a = A+f, where A is the closed linear operator which satisfies conditions (iii) of Subsection 5.9.1, f; = f1(t) is the bounded linear operator at each t; aj = ao(t) is some element of the H-space J; (A, ao)

    is the scalar product of the elements A and ap from J; b = Ob(t) is the operator mapping W into Y; o = by b* is the operator which satisfies con-

    ditions (iv) of Subsection 5.9.1. Condition (5.9.10) serves as the initial condition for Eq. (6.2.42). In the case when the operators ao, a; and o do not depend on time

    supposing 0g/0t = 0 we come to the following equation for determining the stationary process in Eq. (6.2.40): 1

    (A, aDyg) + | t(A, ao) — 3 (As CAG

    0"

    (6.2.43)

    Theorem 6.2.4. Under the conditions (i)—(iv) of Subsection 5.9.1 the one-dimensional characteristic functional (6.2.41) of the process Y; in stochastic linear differential system (6.2.40) ts defined by Eq. (6.2.42).

    592

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Corollary 6.2.4. Under conditions of Theorem 6.2.4 in statwonary system (6.2.40) the one-dimensional characteristic functional of the strictly stationary process is defined by Eq. (6.2.43). Example

    6.2.5.

    Let us consider a linear stochastic differential system

    which is described by the following stochastic wave equation relative to

    0°U See on

    pou ee ~ ° Ox?

    with the boundary conditions U(0, t) =

    present

    Eq.

    (6.2.42)

    I

    Ot

    U2)

    ()

    ter 0), "Bere.

    V =

    for the

    V(z,t) is a

    Vy ==p (£1; r2,t); c, h are the

    Yj = U, Yo = (0/dt)U, Y =

    < Yo ws Poh we

    U a t):

    OU ee)

    normally distributed white noise of the intensity constant coefficients. After putting

    U=

    [Yi Y2 i,

    =¥ 0 ro fal

    one-dimensional

    characteristic

    a fucntional

    g

    = g(A1, A2; t) >. ¢0){i(Ai ; Y;) + i(A2, Y2)} in the form Og

    OE =

    1

    :

    07Y,

    —902 vr2)g + in|Jo,¥%) + (, ott)

    +2h(a, Ya)SON

    ain}

    (II)

    2l

    where the scalar product is given by the formula (9, wv) = f p(x)p(x)dz. 0

    As it follows from Eq. (5.9.9) the multi-dimensional characteristic functionals g, which are determined by (5.9.2) (n = 2,3,...) satisfy Eq.

    (6.2.42). Here (5.9.10) and (5.9.11) will be the initial conditions. 6.3. Methods of Linear Stochastic Canonical Expansions

    Systems

    Theory based on

    6.3.1. Introductory Remarks In elementary theory of the random processes the method of canonical expansions is used generally for the correlation analysis and simulation of the nonstationary processes which are obtained as a result of linear (integral, differential, discrete and etc) transformations. Hence we

    distinguish two methods.

    6.3. Methods based on Canonical Expansions

    593

    According to the first one the canonical expansions of the random input and weighting (transfer) functions of a system are supposed to be

    given. Let the random input X(t) and its covariance function K,(t, t’) of an one-dimensional system be given by a canonical expansion of the

    form (3.9.6) and (3.9.7):

    X(t) = m,(t) + veViay(t), Ke(t,t') = 54Dyay(t)a,(t').

    (6.3.1)

    For linear integral transformation (1.1.13) -s

    t

    Y(t) = ‘;remapeeun

    (6.3.2)

    to

    where g(t, 7) is a weighting function of a system the output Y(t) and its covariance function will be expressed in terms of the canonical expansions:

    Y(t) =m,(t)+ )Vow(t), Ky(t,t’)= >> Diw(t)y(),

    (6.3.3)

    Va.

    and

    t

    t

    wry J str)me(nyar, wee [ot.neu(rjar resi els ) to

    the second method sions of a random which describe the the correspondent pectations and the Example

    to

    (6.3.4) supposes the knowledge about the canonical expaninput (6.3.1) and the differential difference equation system. In this case it is necessary to set up and solve linear differential or difference equations for the excoordinate functions and later to use formulae (6.3.3). 6.3.1. Find the variance of the output in the system

    EVE VereXs

    (I)

    on whose the white noise input acts with constant intensity / beginning with to

    = 0. The system was at rest till the instant tg = 0. The weighting function of the

    594

    Chapter 6 Methods of Linear Stochastic Systems Theory

    system g(t, 1E) was determined in Example 1.1.6.

    Therefore we have the following

    exact solution for the variance:

    Tal AVteney f feOC D;RS @=7

    FER 6(r asr)drdr!imalia = = (1 pa €

    9 is

    0 0

    (II)

    Hence at t — ©O we find DES (0) = vy[2T. For applying the first method we use a canonical expansion of a white noise at

    the interval 7} obtained in Example 3.9.10.

    Therefore we have for the coordinate

    functions according to formula (6.3.4): t

    yr (t)

    =

    1

    .

    TT, /e —(t-7)/T+iwyt gyT

    =

    e BWyt

    e~t/T

    (IIT)

    T%(1 oe iTwy,) ;

    Substituting this expression of the coordinate functions into formula (6.3.3) at t = t’ and accounting that D, = v7

    we get oo

    eivnt _ o-t/T

    The second methods gives for the coordinate functions a linear differential equation

    Bie TYn + Yn =

    While integrating Eq.

    me

    (V)

    (V) at zero initial condition we shall obtain formula (III).

    For the estimate of the accuracy of the presentation of the variance Y by a finite

    segment of series (IV) we notice that at

    >> T' formula (IV) may be approximately

    substituted by the following one: co

    D,(t) # Dy (co) = = Me rea}

    ss

    (VI)

    V

    =—{142 oe

    at

    Y T? aE = + elias 49227?

    In practice we may use formula (VI) at t > 37°. In particular, at Tj /To = 4 while restricting ourselves to the first six terms of the series in formula (VI) we find

    y

    6

    (|

    8

    1,94

    6.3. Methods based on Canonical Expansions

    595

    Thus at restricting oneselves to the first six terms in (VI) what corresponds to the restriction in the canonical expansion of a white noise by the first 13 terms (i.e. by 6 harmonics) we assume an error about 3%.

    Remark.

    Example 6.3.1 convinces us in the fact that for the analysis of linear

    dynamic systems we may practically use the canonical expansions of input random

    disturbances restricting ourselves by moderate number of the terms even in those cases when it is necessary to take very large number of the terms of these canonical

    expansions for the presentation of the disturbances themselves with sufficient accuracy. Usually knowing the cut frequency (Subsection 1.1.7) it is possible to estimate

    approximately what coordinate functions in the canonical expansion of the output

    variable should be known and what functions can be negleced.

    6.3.2.

    Theory of General Linear Transfomations

    Theorem 6.3.1. Let T be a linear operator mapping the phase space X of the random variable X into some linear space Y, conttnuous in the weak topologies of the spaces X and Y determined by the sets of the linear functionals F and G respectively and measurable relatively to the o-algebras of A and B generated by the classes of the S6iSe1 yt (sie awe, fal) Ct and ye: Oe td. Gn Y) Cea at allfRie SES hia o. oe Inne he Owens dn CG and Borel seis A, B of the correspondent spaces. Then the operator T will be (A, B)-measurable > For proving the (A, B)-measurability of the operator J’ we notice that by

    virtue of its continuity the mapping g7' at any g € G represents a continuous linear functional on X.

    As in the weak topology only those linear functionals by which

    this topology is determined are continuous then g7' € F'.Consequently, in this case there exists the adjoint operator 7

    G into F, T*g

    determined on the whole space G which maps

    = gT € F-. Therefore the set {y : (gy, ra

    serves as the inverse image of the set {x ; ((T* 91)x pote

    Jn) Ee B} EB

    & (T*gip)2) fc B} EA.

    As the inverse set Ap of the o-algebra B is the minimal o-algebra which contains all the sets Nes : ((7* oi )zt oe

    om (T* gn )a) E B} then Ap

    C A what proves

    (A, B)-measurability of the operator J’.


    Ww.

    (6.3.9)

    v=1

    After obtaining canonical expansion (6.3.5) of the random variable Y we may find its covariance operator by formula (3.9.18). Theorem 6.3.2. If the linear operator T is continuous in the weak topologies of the spaces X and Y then canonical expansion (6.3.9) of the

    6.3. Methods

    random variable space Y.

    based.on

    Canonical

    Y = TX +b m.s.

    Expansions

    597

    converges in the weak topology of the

    > According to (6.3.3) for any g € G the formula is valid n

    E

    2

    oY a ie Woyw|

    nm



    Dak os aE as

    y=)

    law |?

    v=1

    Sila

    x

    n

    ams ode

    OT 25 lie

    Ves

    By virtue of the continuity of the operator J’ there exists the continuous

    adjoint

    operator 7™* determined on the whole space G and gJ'x = (T*g)x for any

    © € X

    and here 7*g € F’. Thus using again formula (6.3.7) we get a

    2

    B \g¥°-"Vigw|

    n

    = 8 \(T*9)X°| — >> Dy |\(T*9)2r)? yal

    y=]

    mn

    2

    = EB \(T*g)X°-)_V(T*g)av|

    .

    vill

    The latter expression tends to zero at 2 —

    OO because of the weak m.s. convergence

    of the canonical expansion of the random variable X what proves the weak m.s.

    convergence of obtained canonical expansion (6.3.7) of the random variable Y. < Remark.

    The proved theorem with some modifications is extended over

    more large class of the linear operators. Namely as it is easy to see all our arguments remain valid for any linear operator J’ which has the adjoint operator 7™* except

    that the canonical expansion of the random variables gY will converge not for any

    functional Dy

    g € G but for those functionals

    g € G which belong to the domain

    of the adjoint operator 7* and are mapped by the operator 7™ at the points

    of the space F’. Thus canonical expansion (6.3.9) of the random variable Y in this case m.s.

    converges only in more weak topology which is determined in the space Y

    by the set of the linear functionals G’ belonging to the intersection of the space G' with such part of the domain of the adjoint operator 7* which is mapped by this

    operator into F’. Here the operator J’ is measurable relatively to the o-algebra A of the space X and the o-algebra B’ of the space Y generated by the class of the

    sets {y : (91y, 505

    gnY) (E B} at all finite sets 9} ,

    sets B of the correspondent finite-dimensional spaces.

    ... , Jn € G’ and all Borel It is proved similarly as the

    (A, B)-measurability of the continuous operator J' if we account that T*g € F for any g € G’.

    It goes without saying that this generalization has the sense only in

    598

    Chapter 6 Methods of Linear Stochastic Systems Theory

    that case when the space G’ is sufficiently complete for the fact that from gy = 0

    for any g € G’ will follow y = 0. Example

    6.3.2. Let X bea space of the finite-dimensional vector func-

    tions of the scalar variable t, £ € J

    which are continuous

    together with

    their

    derivatives till the order N inclusively, F’ be a space of the linear functionals on X of the form

    fo =)

    Nm

    o> free (th).

    Let us consider the operator of the differentiation

    D =

    (I) d/ dt.

    The operator Dy

    maps the space X into the space Y of the functions which are continuous together

    with their derivatives till the order

    N — | inclusively ( Salteaisrse JV ): We assume

    the set of the linear functionals on

    Y of the form N-I

    m

    gv = >) > gy (te) q=0

    as the space G. in X.

    (II)

    k=1

    Now we consider the random variable X (t) with the realizations

    The operator D! transforms it into the random function Y(t) =

    D'X(t)

    = x”) (t) with the realizations in the space Y. As N-I

    m

    gDin= dd joe (te),

    (111)

    ¢=0Rh=1

    then the functional gD! at any g € G belongs to F’. Consequently, D! is continuous in the weak topologies of the spaces X Theorem

    6.3.2.

    the operator

    and Y, and we may use

    The adjoint operator (D')* in a given case is determined on the

    whole space G and maps G on the subspace F” of the space F’: N

    SEAS OF RO Eel)

    m

    NyEAL

    (IV)

    p=l+1k=1 Therefore after expressing the random function X (t) by any canonical expansion we shall obtain for its derivative of the order | the canonical expansion

    XOW) = mY) +

    UV2P@), tek. Vat

    (V)

    6.3. Methods based-on Canonical Expansions

    Ex

    a mp

    le

    599

    6.3.3.4et X(t), t € 7, be a vector random function of the

    finite-dimensional vector t, X be its phase space, F’ be a set of all linear functionals on X representable in the form

    fe= fp(y?o(eyat for almost

    all the functions

    of the random

    =

    x(t) € X

    function X (t)).

    (1)

    (relatively to the probability measure

    Le

    Let us consider the linear integral operator

    Lx

    J U(s, t)x(t)dt which maps X into the space Y of the m-dimensional vector

    functions of the finite-dimensional vector s, s € S. Here we shall assume that at any fixed

    s € S each row of the matrix |(s, t) determines some linear Mena Rog from F’,

    We determine on Y a set of the linear functionals G of the form gy = > Ik2 y(sx). k=

    Then for any g € G we shall have gLxz = i" > gf U(sx, t)x(t)dt. Putting b= N

    FEED YV Se, 2) 85

    (II)

    we reduce this formula to the form (I). Thus in a given case the operator L is continuous in the weak topologies of the spaces X and Y formed by the sets of the linear functionals F' and G correspondingly To any functional g € G,, i.e. to any finite set of the pairs {sx, gk} the adjoint operator L* sets up the functional f € F' which is determined by formulae (I) and (II). Thus to any canonical expansion of the random function X (t) the corresponds canonical expansion

    ¥ (s)= my(s) + D> Vw(s)

    (111)

    v=!l

    of the random function Vis) =

    (oe Expansion (III) m.s.

    LX(t

    fom

    =

    I(s,t) $ )X (t)dt and here

    Gish [ronenoae,

    (IV)

    converges at all s € S. After obtaining canonical expansion

    (III) of the random function Y(s) we find its covariance function Fonsi ; S2). Example where

    X (t) represents

    6.3.4. Let us consider a special case of the previous example a white noise of the intensity v(t).

    In this case we

    take

    the space of the main functions as F' and use the canonical expansion of the white

    600

    Chapter 6 Methods of Linear Stochastic Systems Theory

    noise of Example 3.9.10.

    As it was shown in this case any orthonormal system of

    the functions {Yn (t)} with the weight v(t), fv(t)ep(t)? Gy (t)dt = bnq gives oo the canonical expansion of the white noise X(t) = Dy Vp Pp (t) which weakly p=1 m.s. aca aes in the sense that for any main function f(t) € F the series f X => Vp J ‘B f( (Ge Pp(t)dt m.s.

    p=1 usual we assume equal to zero.

    converges.

    The expectation

    Let us consider the random function

    of a white noise as



    Y(s) = LX(t) aahI(s, t)X (t)dt, where each row of the matrix |(s, t) at any fixed from the space of the main functions.

    (I)

    s € S' forms the vector function

    In this case the operator L is continuous in

    the weak topologies of the spaces _X and Y and the stated method gives a canonical expansion of the random function ous Jr

    Y(s)= yf (s, t)pp(t)dt , which m.s. converges at any Example

    6.3.5.

    (II)

    $s€ S. Let X (t) be a finite-dimensional vector random func-

    tion which represents a random variable with the phase space Lo Ger Le HL), po

    We assume as F’ the adjoint space Iq (Ti on, Ht), q= p/(p = dy Let us consider the linear integral operator

    us) = La = |Us, t)x(t)m(ae),

    (1)

    where |(e t) satisfies the following conditions: (i) at eachs

    € S I(s, t) belongs to the space Lq (Th, Win HL);

    (ii) at each t € T} I(s, t) belongs to the space of the finite-dimensional vector

    functions Lit? S, v) m>

    1; (iii) for any function z(t) € bp( Ii, J, HL) the function y(s) = i I(s, t)ax(t) (dt) belongs to the space Lm(S, S,v);

    =

    (iv) for any function g(s) which belongs to ee space Ly, (S, So, V), n m/(m — 1) the function f (S)n= eS t)? g(s )v(ds) belongs to the space

    prey ay

    In this case for any z(t) €E Dea sini ie HL) and any g(s) € |be

    y=

    owiouc)

    = [96s)

    S, v)

    (as) f(3,t)e(e)u(at)

    6.3. Methods based. on Canonical Expansions

    601

    = [{[actes.ontas)} eoucan = [sPeOucay. It is clear from (I) and (II) that gl € F’, and consequently, the operator L from Ly (Ti ie jt) into Lm (S, 5 v) is continuous in the weak topologies determined by the correspondent dual spaces /’ = Das

    Sy 7) and-G =

    Eats, aN V). Conse-

    quently, any canonical expansion of the random function X (t) which is weakly m.s.

    convergent in Ly GEG J : Ht) gives canonical expansion (III) of Example 6.3.4 of the random function which weakly m.s. converges in Ly, (S; S; V),

    Hi(s) as /I(s, t)X(t)u(dt)

    (IIT)

    where My(s) = f I(s,t)m(t)u(dt), yy(s) = f U(s,t)x,(t)u(dt). 6.3.3. Assignment of Probability Measures by Canonical Expansions As we have already seen in Subsection 3.9.2 to each vector f from the H-space H, connected with a given random variable X corresponds a scalar random variable (more exactly a class of the equivalent random variables) fX. The measurability of the function fx of the variable x for f belonging to the factor-space F'/N is evident. The measurability of fx for f € Hz, fEF/N follows from the fact that fx as the limit of the sequence {fnz}, fn € F/N of the space L2(X,A, uz) also belongs to L2(X,A, pz) by virtue of the completeness of Lo(X,A, 2). There-

    fore the probability measure V = {V,} of the coefficients V, = f,X° of canonical expansion (6.3.6) of the random variable X is completely determined by the probability measure py of the random vector V on the cylindrical sets of its phase space,

    (lio: Qa. (Aes

    Ses

    ty.) PER

    Bt} eS Sein

    (6.3.10)

    Knowing the measure pl, on the cylindrical sets we may continue it uniquely by the theorem about the measure extension over the o-algebra B of the phase set V. Here the o-algebra of Borel sets of the complex plane is designated as B, and as N is the set of positive integers.. Our interest is in the inverse problem: knowing the probability measure [ly of the random vector V find the probability measure pz of the random variable X.

    602

    Chapter 6 Methods

    of Linear Stochastic Systems Theory

    > For solving this problem at first we establish the measurability of the linear co

    functional pfu

    =

    se) Ui

    Ter) relatively to the 0-algebra BN

    at any f € Hyg.

    v=1

    But the measurability follows immediately from m.s. convergence of the series

    Ey Wied ag EG

    PO

    (6.3.11)

    v=)

    at any f € H, (Subsection 3.9.3).

    In reality from m.s. convergence of this series

    follows the existence of such subsequence

    Apt

    that the sequence

    of the random

    Np

    variables {pv} where gV

    =

    Se ViCfen;

    (p ole

    2) 2 .) a.s.

    converges

    v=1

    to YF V.

    In other words,

    the sequence

    of measurable

    linear functionals

    {v4uv}

    converges almost everywhere (relatively to the measure /l,) to the linear functional pfu.

    Consequently, the functional YU

    is measurable.


    Formula (6.3.6) which determines a canonical expansion of the random variable X establishes the mapping of the phase space of the random vector V into the space X. 91,

    As the inverse image of the set {x

    ; (giz, BER

    na,

    Se B} at any

    --», 9n € Hg and any Borel set B of the n-dimensional space at this mapping

    by virtue of (6.3.11) serves the set {v : (giMmz Gal 5 pers e B}. This set belongs to the 0-algebra B

    On tat t 9,V)

    because of the measurability of the

    linear functionals (yg, ,..., g,,- Knowing the probability measure /l, of the random vector V we may determine the probability measure [lz of the random variable X

    on all the sets of the type {x : (Giese bona Ga®) Se B}:

    pe({ee

    (gies. & gheyesBy)

    = po({u = (grime + gi¥, ---) InMs + ¥y,v) € B}).

    (6.3.12)

    The measure /lz is uniquely expended over the o-algebra A generated by the class of all such sets


    = Be \=1

    hehehe)

    =e

    \a

    ee

    Kai astaierad p=

    -s

    After substituting expression (I) for gz into (II) and taking into consideration that

    fy Kalu = iis Fis) = Dy by p (Subsection 3.9.3) we get

    1

    5

    (111)

    =exp 4-5 | Dyhi ¢ go(h) pal

    The formula shows that the random vector V is normally distributed and its components are independent. Example

    6.3.7. Suppose that the random vector V of the coefficients of

    canonical expansion (6.3.5) of the random variables consequently, its components

    are independent).

    X is normally distributed (and

    Its probability measure jl, on the

    measurable cylindrical sets is expressed by formula

    pall coal upetiat Py, bee }) Sl (Om) n Dy deter in |—1/2

    BoP

    is % 3D

    za

    RPyid tat

    (I) On the sets of the type {v ; (ADy } Bh ae:; AM)v) ‘e B} at any linearly indepen-

    dent h{) Ns

    taal hi) € H the measure fly is determined by formula

    po({v i {AMv, , aes nro} & B})

    =a (Qa) %elG| ew

    {ster dz,

    (II)

    604

    Chapter 6 Methods of Linear Stochastic Systems Theory

    eo) ie |

    where C’ is a matrix with the elements Coq =

    AP) ALY (p, ie

    pra:

    Se n).

    a1

    Using formula (6.3.12) we find the probability measure of the random variable Xx:

    Ho({e : (912, ---, nt) € BY) n

    (2m

    -1/2

    1 ts feat exp 4 — 52 D2

    ar,

    (IIT)

    B-a

    where @ is a vector with the components dp =

    gpMz

    (p a=

    matrix whose elements with the account that (Yo,) vy =

    0° > Di (Gotu) Gav

    formula dyg =

    (p,q al

    n); Disa

    Jp,

    are determined by

    Weare Go n). After determining the

    vol

    measure [ly on the sets of such a type it is uniquely extended over the 0-algebra |

    generated by these sets. Remark.

    In a given case the measure

    [ly may be determined in another

    way. Namely, knowing the characteristic functional of the random vector V we may find the characteristic functional of the random variable

    X : (oe)

    28 Onlf)atbe:

    SfX

    eo

    Afni

    xe

    ie

    Lapis

    a ee

    . Py Vay)

    exp

    are

    Be

    a |

    The latter expectation represents the value of the characteristic functional of the

    random vector V at hy = fx, (v wal

    9x(f) =e

    aes Ay As a result we obtain

    exp

    uedy Drie)

    :

    (V)

    [ims |

    Example

    6.3.8. Let X(t), t € T) be the r-dimensional normally dis-

    tributed vector random

    function,

    F' be a set of all linear functionals of the form

    N

    fre

    Ss fi z(t); t;,...,

    tw € Tj. According to the results obtained in the

    [il

    previous example the multi-dimensional distributions of the random function X (t)

    are determined by formula LOERe

    GEG

    = an

    +12)

    Un

    a

    : (a5) vey

    tn) € B})

    1

    (Di | Nae / exp {-3e7 De dz.

    (1)

    B-a

    Here a is the nr-dimensional

    x (t{(p—1)/r]41) (p Se)

    | ec

    vector with the components

    Ap =

    Mp_-[(p-1)/r]r

    nr), D is the matrix whose elements are expressed

    6.3. Methods based.on Canonical Expansions

    605

    in terms of the coordinate functions ait) of the canonical expansion of the random

    functions X (t) by formula

    dpg =) /Dvahas ,a) =2y,-[(s-1)/r}r (tf(e-vy/r}41) (S=1, -.., nr), p=1 (II) Formula (V) of Example 6.3.7 gives the following expression of the characteristic functional of the random function X (t):

    go(f) =expfi ff(t)? mo(t)dt-7 [[sotentoae] aay ya

    Putting here, in particular, f(t) =

    nor = Se. d(i-1)r4mem6(t — t;), where €1, i=ain=

    ., @, are the unit vectors of coordinate axes of the r-dimensional space we det the n-dimensional characteristic function of the random function X (ay.

    6.3.4. Solution of Equations by Canonical Expansions The canonical expansions allow to find the solutions of the linear operator equations of the type

    LK, =®,

    (6.3.13)

    where L is an unknown linear operator, Kz is a covariance operator of some random variable X, ® is known linearly-adjoint operator. In particular, the problem of finding the regression for the random variables X and Y with the normal joint distribution is reduced to this equation. In the latter case ® = Kyz. Let X be a phase space of the random variable X, F' is a space

    of the linear functionals by means of which the weak topology and the o-algebra A in the space X are determined, Y is a space in which the operator ® maps F, G is a space of the linear functionals by means of which the weak topology and the o-algebra B in the space Y are determined. Usually in practice it is required to find not any solution of Eq. (6.3.13) (it may have not unique solution) but only such solution L that the random variable gLX will have the finite second order moment at any g € G. Besides that it is usually required that the operator L would be permutable with the operator of an expectation.

    Theorem 6.3.4. If the auxiliary H-space H, (Subsection 3.9.2) connected with the random variable X is separable then any canonical

    606

    Chapter 6 Methods of Linear Stochastic Systems Theory

    expansion of the variable X used for the representation of the solution

    Eq. (6.3.13) in all the cases when tt exists. > Firstly, we define the necessary conditions for the existance of the solution

    Eq. (6.3.13). Let D be a required solution of Eq. (6.3.13). It represents the (A, B)measurable

    linear operator as only in this case L_X

    (Ys B). Let {fv} be the orthogonal basis in Hz, variable X

    will be a random

    variable in

    || te [t= D,. Then the random

    may be presented by canonical expansion (6.3.6) where

    V, =

    | PP.§ °

    are uncorrelated random variables with zero expectation and the variances which are equal to the correspondent numbers D),,, and @, are the coordinate vectors satisfying

    the conditions (Subsection 3.9.3):

    Sn

    De Tee

    Sit

    Oat

    (6.3.14)

    From the fact that the random variable g_X has the finite second order moment at any g € G and the operator L is permutable with the operator of an expectation

    follows the convergence of the series [e.e)

    SO Deigin

    (6.3.15)

    esi |

    at any g € G. As by virtue of the first formula of (6.3.14)

    n

    :

    E |gLX° — 5 V,(gLa,)| Vy = gLEX°V, — Du(gLt,) pI

    =gLEX(f,X°) = D,(gLbz,)=9L Ke fy — Duglz,=0 (v=1,..-/n), n

    then 1 jonx" =

    2

    Dy Vo(ube.)

    n

    =

    it; ae

    =

    ae. IDs, \gLa,|°. Hence

    Vii

    Vet)

    it follows that series (6.3.15) converges at any

    g € G. Further, it follows from the

    canonical expansion of the covariance operator:

    eS

    [e-e)

    =

    bs. py By

    fey ), y € Hg,

    vs co

    and from the convergence of series (6.3.15) that LK, f =

    3 D, (Le;

    ———w

    (fez);

    v=

    where the series weakly converges at any f € Hz, in the weak topology of the space co

    Y.

    Substituting this expression into (6.3.13) we get y) Di(iey)( fev) =ubf; v=1

    f © H,. From here and from the convergence of series (6.3.15) it follows that Eq. (6.3.15) may

    have the required solution only in that case when the operator @® is

    representable by the expansion

    GEN VDI WP i

    |

    Tpy

    (6.3.16)

    6.3. Methods based-on Canonical Expansions

    607

    where Y, are the vectors of the space Y which satisfy the condition

    Ss Ds low | < oo

    atany

    gEG.

    (6.3.17)

    v=)

    For determining the coefficients y, we replace the index of the summation VY in (6.3.16) by the index

    and apply the operator ® to the vector f,. Then by virtue

    of (6.3.14) we obtain D, y, = ®f,. Consequently,

    et) ge 0 fra Veet 2

    red

    (6.3.18)

    -s

    > The operator L in (6.3.13) is determined only at the domain Ak, covariance

    operator

    K,.

    Therefore

    the variable

    L_X

    has the sense

    E IN ie... In this case there exists such a vector fo €

    Hy

    that mg

    of the

    only at Mz =

    Ia foe

    And as {fi} is the basis in Hy then fo = iD. Cut and by virtue of (6.3.14) Mz co

    =

    je Dz Cy, and this series converges in the weak topology of the space ACs lea H=1 order to obtain the explicit expression for the coefficients C, we apply to the previous equality for the functional f,. Then taking into accounting the second relation of (6.3.14) we get Dice = Jyms, (v A

    ie

    3) and consequently,

    co

    Mo = Devise

    :

    (6.3.19)

    jr

    Finally, noticing that on the basis of the first relation of (6.3.14) fymz = fy Ke fo Tear

    = foKefy

    ¢

    iS

    = Dv foxy and accounting the convergence of the series 3 DF

    at any f € Hz, we come

    fz,

    2

    to the conclusion that the following condition must be

    filfilled:

    SO

    yee

    a

    (6.3.20)

    v=1

    Theorem

    6.3.5.

    For the existence of the linear operator L which

    satisfies Eq. (6.3.1), is permutable with the operator of the expectation and gives the finite second order moment of the random variable gLX,

    conditions

    (6.3.16), (6.3.17), (6.3.19, (6.3.20) are necessary.

    We prove that these conditions are also sufficient for the existence

    of the solution of Eq. (6.3.13).

    608

    Chapter 6 Methods of Linear Stochastic Systems Theory

    > For this purpose we shall find the operator L in the form of the series

    ie enn where for

    7), are yet unknown

    any

    f



    Hg

    we

    vectors. get

    Then

    (OY Goss 6 =.

    (6.3.21)

    taking into the consideration



    (6.3.7)

    00

    oe Nd

    hal

    =

    pas Nd

    pH

    Kath

    p=1

    co.

    =

    my Dalton):

    In order the operator L would satisfy Eq. (6.3.13) it is neces-

    r=

    sary gteet at

    any

    f

    => Dunu(fep)5 Dahl oole= = we 4



    4H,

    the

    following

    After comparing

    equality

    this expansion

    b>DuYu( fen) at any f €

    Ny = UW oss riydeh Ae

    will

    with

    Hy.

    be fulfilled:

    (6.3.16) we

    Putting here

    Df

    obtain f =

    fy,

    Substituting this expression into (6.3.21) and

    accounting (6.3.18) we obtain a fomal solution of Eq. (6.3.13) in the form of series

    b= = Yow) f, =Lely: 1(@f,) fy.

    (6.3.22)

    v=

    Let us take a look in what sense expansion (6.3.22) determines the solution of Eq. (6.3.13) which possesses thedis eae prapertics. For this purpose we consider the

    random variable

    weakly m.s.

    Y = LX = DSUp jA=

    3 Up Ch Mz + Vz i This expansion

    converges as it follows from (6.3.77) and (6.3.20) that at any € > 0,

    g € G, for all p and at sufficiently large n 2

    n+p

    E

    So (gw) (frre +V,)|

    2

    n+p

    =

    vr=n

    S\(9w)( fume) v=n

    n+p

    :

    n+p

    ar py D, lay |? vr=n

    n+p

    n+p

    vrn

    vr=n

    ) Dy lawl’ - >> Dr" fmol? + 5 Dy lawl? D5V, rfKye(s,7) f-(r)dr. In the majority operator

    of the problems

    Ly to the random

    function X Oe).

    function

    of the practice it is necessary

    evan)

    to applythe

    X(t) but not only to the centered random

    As we have already seen for the existence of such an operator it

    is necessary and sufficient the fulfilment of conditiofs (6.3.19) and (6.3.20) besides

    conditions (6.3.16) and (6.3.17). In a given case these conditions take the form

    Mz(t) =oestt) [oePmelrir

    py | fo"

    metry] < OO.

    (VIII) Therefore formula (III) determines the required solution of Eq. (I) and the series LX

    =

    SS. yp (s) pe f(t)? mz(t)dt + VN m.s.

    converges at each

    s € S to the

    PSik

    random variable with the finite matrix of the second order moments.

    Example

    6.3.10. Inthe special case of the finite set 7) = {t, rick stn}

    all the functions f(t) € H,

    are the linear combinations of the 6-functions with the

    peculiarities at the points t; , ...

    , t47, the canonical expansion of the random func-

    tion represents the finite sum which contains no more than nN items (Example 3.9.8), and the formulae of the previous example give the solution of Eq. (I) of the previous example in the form of the finite sum.

    large N the solutionof Eq.

    The operator 1; in this case is a matrix.

    At

    (I) of the previous example which is presented by the

    formulae of the previous example is considerably simpler for the calculation than the

    solution which is obtained by the inversion of the block matrix Tie (us ; ip Example

    the spaces X

    6.3.11.

    Example 6.3.9 is easily generalized on the case when

    and Y represent the spaces of the functions with the values in any

    linear spaces with the weak topologies which are determined by the sets of the linear functionals A, and Ay respectively.

    In this case in all previous formulae f(t)’ is

    a function with the values in A, at any t and the variables 9} v. ©, gn are the vectors of the space Ay. The convergence of expansion (III) of Example 6.3.9 at the

    replacement of x(t) by X(t) or X(t) in a given case should be considered as m.s. convergence at each

    § € S in the weak topology of the space Y of the values of the

    functions y(s) EY.

    The stated methods may be also used for the calculation of the Radon-Nikodym derivatives of the output measure jy by input measure

    Le, dpty/du, (Pugachev and Sinitsyn 2000).

    6.4. Methods based on Integral Canonical Representations

    613

    6.4. Methods of Linear Stochastic Systems Theory based on Integral Canonical Representations 6.4.1. Introductory Remarks

    The integral canonical expansions (Section 4.8) along with the canonical expansions (Sections 3.9, 6.3) are widely used for the analysis and the simulation of the random processes in the dynamic systems with random disturbances. The methods of the application of the integral canonical representations are analogous stated in Subsection 6.3.1. According to the first method if an one-dimensional linear stochastic system is given by a linear integral transformation (6.3.2), and an input and its covariance function are determined by some integral canonical

    representation (Section 4.8):

    X(t)=mett+>0 |¥ (A)z-(t, )dd, c— LAL

    (6.4.1)

    Kee) )=¥ [uw a(t, A)a,(t,A)dd, — ans

    then the output Y(t) by virtue of (6.3.2) will be expressed by the integral canonical representation:

    =m, (1)4°

    fv r)yr(t,A)d

    ral,

    y(t) AC =

    (6.4.2)

    fr)Jur(t,A)yr(t!,A)dA r= 1A.

    and

    F

    my(t)=f9(t,7)ma(r)dr to

    ee

    (6.4.3)

    [ote rievtn Nye pees eos sana.

    The formulae for the finite-dimensional linear system and the vector inputs are written analogously.

    614

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Now we consider the second method. Let the equations be given which describe the behaviour of a linear stochastic system. ‘Then for the correlation analysis we may use the first method after determining before-hand the weighting functions of a system. For obtaining the equations which for the expectations all random inputs and outputs should be substituted by their correspondent expectations in the equations of the considered system. While deriving the equations which determine the coordinate functions it is necessary to replace all random inputs and outputs by the correspondent coordinate functions in the equations of the system. Example

    6.4.1. Find a variance of the output of the Spatianarss oscilla-

    ting link

    4

    Y+a1Y =

    for MN,

    0, k(t)

    =

    D,e7

    +aoY

    217 -AtN=

    (I)

    =X

    Li \ieweuhave

    z(t,w) =

    eiwt uel

    y(t,w) + ar y(t,w) + aoy(t,w) = e*, y(0,w) = 0.

    (11)

    Hence after integrating we find

    (65) we

    OD

    eet

    (Az — tw)e*1* — (A, — iw)e>??

    —————————

    do tayiw—w?

    ©

    ere

    (A, —Az)(ao + ayiw —w?)

    (11) ’

    where 1,2 = —2 +17 are the roots of the characteristic equation part ajA+ ag = 0. Consequently,

    aD,

    Dy (t) = —T

    /(cos wt— bee sin yt—e—** cos yt)? +(sin wt —cian sin yt)?

    x | —-—_

    F

    (a? + w?) Jag + ayiw — w?/?

    —0o



    l

    (IV) At t —

    ©o for the asymptotically stable system

    (B >

    0) we find the stationary

    value of the variance from (IV)

    D

    (eo)

    =

    =

    D,(a + a,)

    aoa;(a? + aya + ao)

    We may come to the same result if we account that [o.e)

    Dy(00) = fsy()du, s4(w) = [®(iw)[? salu), 0

    (V)

    6.4. Methods based on Integral Canonical Representations

    De

    a



    s2(w) = =, TW

    at+w

    615

    i)

    0(iw) = ——____, do + ayww — Ww

    6.4.2. General Theory of Integral Canonical Representations of Linear Transformations > After expressing the random variable

    X by the integral canonical represen-

    tation (Section 4.8)

    Rie sare: 55i;2,(A)Z-(d2),

    (6.4.4)

    we obtain for the random variable (6.3.5) the analogous canonical representation

    d= prrtgn | 3} Yr (A)Z, (dd) , where My = Tm,+6), Yr (A) = TanlA} (A CAS is continuous in the weak topologies of the spaces

    (6.4.5)

    = 1.2.45 Ae If the operatorT'

    X and Y thenrepresentation (6.4.5)

    similarly as in (6.4.4) represents weakly m.s. convergent series of the weak stochastic

    integrals.

    In other words, for any linear functional

    gY = gIX

    is expressed by m.s.

    g € G the random

    variable

    convergent series of the stochastic integrals:

    gY¥ =gmy+)_

    gr (A)Z, (dd).

    (6.4.6)

    After obtaining canonical representation (6.4.5) of the random variable Y we may find its covariance operator by formulae of Subsection 4.8.2. [2OGHZO),

    (11)

    r=1

    where a(t, A) is the I*® derivative of the function z(t, A) over t. Example

    6.4.3. In Example 6.3.3 after expressing the random function

    Pe (t) by representation (I) we get for the random function Y(s) an analogous integral canonical representation

    ¥(s) =my(s) +90 fun(s,2)Z-(aX), where

    (1)

    q

    yrs fA) [iedeelt rat. After obtaining representation

    (II)

    (I) of the random function Y(s) we may

    find its

    covariance function K, visi , $2) by means of the known formulae. Example

    6.4.4.

    In the special case when

    X (t) is a stationary scalar

    random function of the finite-dimensional argument ¢ it may be expressed by spectral

    presentation (Subsection 4.8.6):

    OTe

    i:eZiT t(dd).

    (I)

    Then we get for Y(s) the integral canonical representation

    ¥(s) =m,(s) + fu(s,2)2(@d),

    (11)

    where it is denoted

    My (s) = [iedmedt, y(s,A) = [iene

    ae,

    (IIT)

    6.4. Methods based on Integral Canonical Representations

    617

    In particular, when t = § is time and Y(@) represents an output of asympto-

    tically stable stationary linear system which performs in the stationary regime then y(t, X) = (id)e?,

    where (id) is a frequency characteristic of a system. In this

    case the formula for the covariance function of the stationary random function Yate)

    gives an integral representation co

    ky(r) = i eT B(id)@(iA)*o(dd) , —

    (IV)

    OS

    where ¢ is the spectral measure of the random function X (t). Finally when there exists the spectral density of the random function X (t), and Y(t) is the scalar random function then the obtained formula gives the known expression for the spectral

    density of the random function Y(t): Sy (\) = Example

    6.4.5.

    |6(id)|? Bela):

    In the other special case when

    X(t) is a stationary

    m-dimensional vector random function it may be expressed in terms of the integral canonical representation (Subsection 4.8.7):

    — 1 X()=met+ >> /ap(A)e*”*Z, (dd).

    (1)

    p=1

    Then we get for the random function ¥ (3) the integral canonical representation

    m

    ¥ (8)= my(s)+> fups,2)Zp(@d),

    (it)

    pal

    where My (s) is determined by the formula as in the previous example and

    Up (8, A) = /I(s,t)ap(A)e* ‘dt.

    (IIT)

    Ift = S is time and Y(s) represents the output of the asymptotically stationary linear system which performs in stationary regime then Yp (t, A) ces (1A) ap Gene where (7d)

    is the frequency characteristic of the system.

    In this case for the co-

    variance function of the stationary random function Y(t) the following integral representation takes place:

    ky(7) = > fe” 8G A)ap()ap(A)" HA)" 9(4A), (AV) p=!

    618

    Chapter 6 Methods of Linear Stochastic Systems Theory

    where 0 p is a covariance function of the stochastic measure

    Zp: According to the

    formula for the matrix 7 (B) of the spectral and cross-spectral measures of the com-

    ponents of the vector random function X(t) (Subsection 4.8.3) we have m

    Ye | ae()ap(2)" (AA) = (8),

    (v)

    p=1B

    and consequently, we may present formula for holt a in the form

    eee /eOTS(iNo(ANO(A.

    (VI)

    In the case when there exists the matrix of the spectral and the cross-spectral densities

    of the components of the vector random function X (t) the obtained formula gives the following expression for the matrix of the spectral and the cross-spectral densities of the components of the vector random function Y (é):

    sy(A) = ®(id)se(A)B(iA)* Example

    (VII)

    6.4.6. Interpreting the random function X(t) of Example 6.3.5

    as representation (I) of Example 6.4.2, we obtain for the function Ys) representation

    (1) of Example 6.4.3 where

    un(s,) = fUs, )zr(t, Adu(dt). The

    correspondent

    integral

    canonical

    representation

    of the

    covariance

    function

    Ky (s1 ; $2) of the random function Y(s) is found by the formulae of Subsection 4.8.2.

    6.4.3. Solution Method Analogously

    of Equations

    we may

    solve Eq.

    by Integral Canonical

    (6.3.13) by means

    representation of the random variable X.

    Representations

    of the integral canonical

    As we have already seen in Section 4.8

    any sequence of the expansion of unit {E,} and the orthogonal vectors 477, ¥ such that the set of the vectors of all ENB) 7; is dense in Hy give the following integral

    canonical representation of the random variable X ,

    2 He, /r,(X)Ze(dd).

    (6.4.7)

    6.4. Methods based on Integral Canonical Representations

    Here Zea

    619

    Ey Vy X° are uncorrelated stochastic measures, and the functions tp (Xi)

    satisfy on the basis of Subsection 4.8.2 the following conditions:

    J %Q)or(@) Stn

    ee

    re

    (6.4.8)

    B

    Eel Bye A) = or, lBr) for any set B (B CA,BeE measure

    (6.4.9)

    S) where 0, is a covariance function of the stochastic

    Z,.

    > If L is a solution of Eq. (6.3.1 3) which possesses the required properties then

    by virtue of (6.4.8) for any

    E

    g €

    Gand

    BES

    gLX°- > /gLa,(A)Z, (dd)

    Z,(B)

    = gL EX" Z,(B) = [ ghay(d)ou(d2) B

    = gLEX° [E,(Byy.X°|— 9h KeE,(B)y.=0,

    (v=1,..., 7).

    L

    Consequently, according to (6.4.8) we have the equality

    2 EB

    gLX°-S~ /gL, (A)Z,(dd)

    = E |gLXx°|’ — > [oter(aye |@LX°Z,(dd)| rool

    = B |gLX°|’ - > fot-0)8 |SEK E- (a) | r=)

    = E |gLX°|” - yi \gLx,(A)|° o-(dA). r=1

    Hence it follows that the condition is fulfilled

    D /[lgLe,(X)|? o-(dX) < 00.

    (6.4.10)

    620

    Chapter 6 Methods of Linear Stochastic Systems Theory

    cb oe it follows from the canonical representation of the covariance operator Kz f =

    shLal A)f2,(A)o,(dd) iB that at any f € H, by virtue of condition (6.4.10)

    ‘hie:series

    EK f= >. (Lx, (A) fa, (A)or(dd) converges in the weak topology of the space Y . It follows from the obtained expression LKgg

    that Eq. (6.3.13) may have the required solution only in that case when the

    operator ® is representable by

    o= Y fey [-2-0)| 2408,9

    peamnd

    Cxa

    and if we account (6.4.10) then for any g € G the following condition takes place:

    (6.4.12)

    Sef lgyr(A)|° o(dA) < 00.

    For determining the functions y, (A) we substitute in (6.4.11) the index of the summation Y by p/ and apply the operator ® to the vector €,(B)y--

    consideration (6.4.9) we get BE,(B)y, :

    Then taking into

    = f yr (A)o, (dA). Hence it follows that B

    for any g € G the function gy, (A) represents the Radon—Nikodym derivative of the scalar measure gE,(B Vr over the measure 0,. This result may be written in the form

    dO€,7;

    He yeas

    Os

    (6.4.13)

    If the operator [ should be applied to the random variable X but not only to the centered random variable X° then one more condition is necessary, namely My

    KE Ak, which on the basis of Subsection 4.8.2 with account that u, = fz, (A)

    at any Si € Hy gives

    itn = Keo Ke Yofft )Ep(dr)¥p= Y |Fae, E, (dX) Yr00 Hence taking into the consideration

    (6.4.8) we get Mz

    =

    Sa J tr (A) fotr(A) r=1

    X Or (dd). Therefore the expectation must be presentable by the expansion

    me =)~ /Eiicaonexdy,

    (6.4.14)

    6.4. Methods based on Integral Canonical Representations

    621

    and even

    Se / I O)P ar(aa) = ei

    Sf[fore(2)[? o¢(dd) = foKefo We shall find the solution of Eq. (6.3.13) in the form

    _— Sf nAveoryr

    (6.4.16)

    ra

    Then we have for any f € H,

    EKef =>r=1 |mE Ref r==>fnO)FRELDV, or accounting (6.4.8)

    LES f= Se/tr (A) far(A)or (d). r=1

    In order the operator L will satisfy Eq. (6.3.13) it is necessary at any f € Hz the equality Of

    =

    ‘3 f nr (A) far (A)o, (dA) holds.

    After comparing

    this equality

    ral

    with (6.4.11) we shall have at any fee HZ

    Df maFeeOou(ar) => fwe) FeeP)op(ad). a

    Putting f =

    [Trapt

    bal Baa

    by virtue of (6.4.9) we get J nr(A)or (dd) = B

    x o,(dd). As this equality is valid for any set

    f yr (A) B

    B € S then Nr (A) = Yr (A) almost

    622

    Chapter 6 Methods of Linear Stochastic Systems Theory

    at all \ relatively to the measure 0,. Substituting the latter integral expression into

    (6.4.16) and with account of (6.4.13) we find that

    ic: ob a TOE eae >"fZe 0) £.(d\ye.4 (6.4.17) Similarly as under conditions of Subsection ment is proved.

    6.3.1 the following state-

    Theorem 6.4.3. In the case of the fulfilment of conditions (6.4.11)

    and (6.4.12) the random variable Y° = LX° has the covariance operator Ky determined on the whole space G and the operator L is permutable with the operator of the expectation. And if besides (6.4.11) and (6.4.12)

    conditions (6.4.14) and (6.4.15) are also fulfilled then there exists the random variable Y = LX which has the expectation my = Lm, and the operator of the second order moment determined on the whole space G.

    Let be exist the measures pp with the values in H,, the nonnegative

    measures o, and the functions z,(A) with the values in X satisfying the following conditions (Subsection 4.8.4) in some measurable space (A, S):

    Keftp(B) = [eoQdon(aa), Bes,

    (6.4.18)

    B pol Beg) = booq lala) Then

    (6.4.19)

    N

    >, J f%p(A)up(ddrA) = f at any f © Hz and we may take &)(B) p=

    a

    =

    ¥ S400)

    (A) ]Hp(4A), Yr =

    er(A)

    (r =

    1,

    ., NV). In this case amas (6.4.= takes the form N

    N

    b= Sf w(r)no(d) =o fFEEup(ad). Example

    (6.4.20)

    6.4.7. For solving the equation of Example 6.3.9 by the integral

    canonical representations method it is required to find some number WN of the vector

    measures [lp (b, B ) (of the same dimension that the vector random function X (t)), N nonnegative measures

    op(B) and N sets of the vector functions oe

    A) (of

    the same dimension that X (t)) determined on some measurable space (A, S ) and

    6.4. Methods based on Integral Canonical Representations

    satisfying the conditions of Subsection 4.8.5.

    623

    Then formula (6.4.20) will give the

    following expression of the linear operator Ly which satisfies Eq. (1) of Example 6.3.9: N

    ie

    Y /tsdawae,

    (1)

    where

    I(s,t) ->f Here in order the operator

    tation and the random

    [eena akcat dr >jtp(t, dr)? .

    L; will be permutable

    function

    Ys)

    =

    (11)

    with the operator of an expec-

    EXC)

    will have the finite covari-

    ance function at all s € S' it is necessary and sufficient that the function p(s, Ti) N

    may be represented as p(s, T) = > ‘jUels, Aloe tis r)*op(dd), where Yp (s, A) =Sh MOSET) T) Sus(X) BO ze dr (eE =i

    ; N) and at all s € S thecondition of the

    form (6.4.12) will be fulfilled: > f l¥p(s, A)? op(dA) < co. p=l1

    In the special case of the finite-dimensional space A and absolutely continuous

    measures /lp , B) and op(B) formula (II) takes the form

    I(s,t) = i ray) {felssr)apte Mar} ap(t,ayaa, where

    dpi, A) and

    Vp(A) are the Radon—Nikodym

    derivatives

    (IIT)

    of the measures

    Lp (t, B) and op(B) over the Lebesgue measure respectively. Example when

    the random

    6.4.8. Let us consider a special case of the previous example function X (t) is real and its forming filter is known

    (Subsec-

    tion 4.8.4). Let w(t, T) be a weighting function of a forming filter, w™(t, T)be a weighting function of an inverse system. Then we may assume as JV the number of the components of the random function X (t) and as the matrices-columns

    Lp (t, ) and Gp (t, A) (p ai

    egies

    N) the correspondent columns of the matrices

    x(t, A) = w(t, A) and a(t, A) = w-(A, AE

    Considering equal to unit the inten-

    sities of the components of a white noise which is transformed by a forming filter into

    the random function X(t) we get from (III) the following formula:

    Ks,t) =f {[osu Q.ar} webdr.

    av)

    From this result as the special cases the Wiener formula known from the theory of

    optimal linear systems and its generalizations on the finite-dimensional linear systems are derived.

    624

    Chapter 6 Methods of Linear Stochastic Systems Theory

    6.5. Applications.

    Shaping Filters Design and Modeling

    6.5.1. Shaping Filters for Scalar Stationary Processes Let us show how we may design the shaping filter for a scalar stationary random function X (t) with rational spectral density. > Suppose that the spectral density of a real scalar stationary random function hg (t) with zero expectation is determined by the formula

    S2(w) = P(w)/QW), where P(w) and Q(w) are the polynomials. and Q(w) contain only even degrees of W.

    -

    (6.5.1)

    As Sz (w) is an even function, P(w)

    In practical problems the degree 2™ of

    the numerator P (w) is always smaller then the degree 27 of the denominator Q (w), as only in this case the variance of the random function X (t) equal to the integral of the spectral density is finite. Besides that it is necessary for the convergence of the integral of the spectral density that the denominator Q(w) of the spectral density be

    different from zero at any real value of W. Finally, since a spectral density is essentially positive at all (real) values of the frequency W,all the coefficients of the polynomials 12 (w) and Q(w) are real. It follows from these properties of the polynomials P (w) and Q(w) that the roots of each of them are pair-wise conjugate complex numbers and to each root corresponds the root of the same polynomial of opposite sign.

    In

    other words, the roots of the polynomials P(w) and Q(w) are disposed on the plane of the complex variable

    W symmetrically relative to the real and imaginary axes.

    Let us factorize the polynomials P (w) and Q(w) and select the factors corresponding to the roots disposed in the upper half-plane of the complex variables Ww.

    ee (w) has real roots then all these roots are multiple roots of even multiplicity as each of such roots is the point where the plot of the spectral density is tangent to

    the real axis being disposed wholly above it. The half of the factors corresponding to each of such roots should be considered as belonging to the upper half-plane and the

    other half as belonging to the lower half-plane. Taking all the factors in the factorized polynomial P (w), corresponding to the upper half-plane with the arithmetic square root of the coefficient Pam

    of w?™

    in this polynomial and 7™ as additional factors

    we shall obtain the polynomial of degree

    ™ in W whose all the roots are disposed

    in the upper half-plane and on the real axis. We denote this polynomial by H (iw). It is easy to see that all the coefficients of the polynomial

    H (iw) considered as a

    polynomial in iw are positive. Now let us take some root of the polynomial H (iw).

    Let this root be a +28,

    8 > 0. Ifa

    2 0, then according to the property of

    the polynomials P (w) and Q(w) proved above the polynomial H (iw) has also the root —@ + 1. Consequently, the polynomial H (iw) represents the product of the

    6.5. Applications.

    Shaping Filters Design and Modeling

    positive number ,/P2mby the factors of the form i(w sa

    625

    co 1B)i(w +a—

    i) and

    i(w — 12), where 3 > 0. But

    i(w — a —if)i(w + a — iB) = (iw)? + 26(iw) + a? + 6? ,

    i(w — ip) = iw + 8, Consequently, the product of any number of such factors represents the polynomial in 1wW with positive coefficients.

    The remaining ™ factors in the factorized polynomial P (w) corresponding to the roots disposed in the lower half-plane multiplied by VP2m and by (-i)™ form the polynomial H (—iw).

    which is obtained from H (iw) by the change of sign at 1W, i.e.

    Therefore, each root & +78 of the polynomial P (w) in the upper half-

    plane corresponds to the root @ — 7 in the lower half-plane. Therefore, each factor

    of the form i(w — a — iB)i(w +a — iB) = (iw)? + 2B(iw) + a? + B? or 1(w _ i) = iw + B of the polynomial H (tw) corresponds to the remaining factor

    of the form —i(w — a +%)(—i)(w+a+if) = (—iw)? +26(-iw) +a? + 6? or respectively —i(w + i) = Thus

    -—iw+ 16

    selecting in the factorized polynomial

    P (w) the factors corresponding

    to the roots disposed in the upper half-plane and adding the factor 1”

    ./pom we

    represent the polynomial P (w) at real values of W in the form

    P(w) = H(iw)H(—iw) = |H(iw)|’ , where

    H (tw) is a polynomial in 2W with positive coefficients.

    (6.5.2) In exactly the same

    way selecting in the factorized polynomial Q(w) the factors corresponding to the roots disposed in the upper half-plane, adding to them the factor 1” Va2n, dan being

    the coefficient of w2” in the polynomial Q(w), and denoting the resulting polynomial

    by F (tw) we represent Q (w) at real values of W in the form

    Q(w) = F(iw) F(—-iw) = |F(iw)|? , where F’ (iw) is the polynomial in 2W with positive coefficients.

    (6.5.3) Representing the

    numerator and the denominator of the spectral density Sz (w) by formulae (6.5.2),

    (6.5.3) we obtain the following expression for the spectral density Sz (w):

    sz (w) = |H(iw)/F(iw)|” .

    (6.5.4)

    Such a representation of the spectral density is called factorization. Notice now that the function

    ®(s) = H(s)/F(s)

    (6.5.5)

    626

    Chapter 6 Methods of Linear Stochastic Systems Theory

    represents the transfer function of some stationary linear system.

    Since the multipli-

    cation of a complex number by the imaginary unit 7 represents the counter-clock wise rotation of the vector representing this number by the angle 7 /2 the left half-plane of the complex variables variable W.

    $ = iW corresponds to the upper half-plane of the complex

    Consequently, all the roots of the polynomials H (s) and F’ (s) are dis-

    posed in the left half-plane of the variables S, and the function O( Ss) determined by formula (6.5.5) represents the transfer function of a stable stationary linear system.

    Recalling that the stationary random function with constant spectral density Sq represents a white noise of intensity 277S9 we come to the conclusion that the stationary random function X (t) with rational spectral density Sz, (w) may be considered as the response of the stable stationary linear system with the transfer function ®(s) determined by formula (6.5.5) to the input representing a white noise with unit spectral density.

    Consequently, the system with the transfer function ®(s) Ss H(s)/F(s)

    represents the shaping filter for the random function X (t). On the basis of the results of Subsections 1.1.6 and 1.2.5 the stationary linear system with the transfer function ®(s) =

    differential equation with the operator

    H(s)/F(s)

    F(D),

    a

    is described by the linear

    d/dt, in the left-hand side

    (acting on the output) and the operator H(D) in the right-hand side (acting on the input). Taking into consideration that the input of the shaping filter is the white noise

    V(t) with unit spectral density whose intensity is equal to 277, and the output of the shaping filter is the random function X (t), we obtain the differential equation of the shaping filter:

    F(D)X

    =

    H(D)V

    the polynomial F(s) by az (k cd) polynomial H(s) by by (k =i

    gi a Lie}

    or denoting the coefficient

    m),

    Sa x®) = k=0

    of s* in

    Sa n), and the coefficient of s* in the

    mV, k=0

    Thus the shaping filter for the random function X (t) represents a system described by a linear stochastic differential equation of the

    th

    Nn” order. Replacing this equation

    by the corresponding set of differential equations of the first order (Subsection 1.2.3) we obtain the equations of the shaping filter in the form Mp

    Seer

    (k=

    Aik. = Ai pst

    OV

    1

    a1);

    ( OT

    aed)

    aL)

    (6.5.6)

    n

    Xin = an" So ap-1 Xie + gaV, k=1

    and A(t) = O(Xi(b),t) = ire Ae

    ee

    eee

    here

    6.5. Applications.

    Shaping Filters Design and Modeling

    627

    ~

    Qn—m)---+,Qn

    are constant coefficients determined by formulae: -1 In—m

    =

    an,

    bmn )

    k-1

    qbes a,

    bn—k —

    yi

    An—k+hQh

    Stig

    nt

    1

    cn.

    h=n-—m

    i)

    (6.5.7)

    Ps

    Remark.

    In some

    cases it is desirable to form a random function X (t)

    of a white noise of unit intensity.

    equal to 1/2r. dom from

    The

    spectral density of such a white noise is

    Consequently, representing the spectral density Sz (w) of the ran-

    function X (t) in the form {6.5.4) it is necessary |H(iw)/F(iw)|?

    polynomial

    the factor 1/2r.

    H (iw) the additional

    We may

    in such cases

    attain it by including into the

    factor Jor or by including

    mial F (iw) the additional factor 1/

    2m.

    to separate

    into the polyno-

    As a result instead of (6.5.4) we get

    Se(w) = |H(iw)/F(iw)|? (1/27). While applying formula (6.5.4) or more general formulae (6.5.8) two conditions must be satisfied:

    the condition of stability of a system and the condition of sta-

    tionarity of its output.

    The first condition is satisfied by virtue of our method of

    constructing the shaping filter. This filter is stable since all the poles of its transfer

    function (6.5.5) are disposed in the left half-plane of the variables s. condition may be satisfied by the choice of initial conditions for Eqs.

    The second (6.5.6).

    This

    is possible because the shaping filter is a mere theoretical tool for reducing system differential equations to stochastic equations due to which the initial conditions for the equations of the shaping filter may be taken arbitrarily. We shall now show how we may do it

    Sy(w = B(iw)s_(w)P(iw)* .

    (6.5.8)

    > As it is known, a stationary random process has a constant covariance matrix (the variance in the case of a scalar process).

    In other words, the covariance ma-

    trix of the value of the stationary random process X

    (t) at any time moment f is

    independent of ¢ and is equal to the value of the covariance function ke, (T)of this process at T =

    0, ke40); co

    kei (0) = / Sz, (w)dw .

    (6.5.9)

    —oo

    Therefore, for providing the stationarity of the process Xj (t) at the output of the shaping filter one should take the random initial value Xi (to) = X09 at the instant

    628

    Chapter 6 Methods of Linear Stochastic Systems Theory

    to with the covariance matrix | (0) determined by formula (6.5.9). It remains to find the spectral density Sz, (w) of the process X4 (t) determined by Eq. (6.5.6).

    Using formula (6.5.8) we find

    Sz,(W) = By (iw), (—iw)? so ,

    (6.5.10)

    where Sq is the spectral density of the white noise V equal to 1 if sz (w) is repre-

    sented by formula (6.5.4) or 1/2x ALES 7s (w) is represented by formula (6.5.8).

    For

    determining the transfer function ®; (s) of that part of the shaping filter which gives the vector random process X 4 (t) =

    [Xu (t) seen Ores (t) ie at the output we notice

    that the transfer function of the whole shaping filter whose output represents the pro-

    cess X(t) = X11 (t) was found earlier. It is determined by formula (6.5.5), 13(s) =

    ®(s) =o (s)y)F (s). Therefore, the remaining elements of the matrix-column

    ®,(s) may be found from the first n — 1 by replacing V by the function ere Xik by the function 14(s)e** (k =e

    ee n), and the operator d/dt by the variable

    S (Subsection 1.2.5). As a result we obtain the equations

    s®14(s) = Py x41(s)

    (K=1,...,n—m-—1),

    s®1;,(s) = Oy 44i(s) tan (K=n-—m,...,n—1). Solving these = ®(s) =

    equations

    in succession

    and putting in the first of them

    ®11(s)

    H(s)/F(s) we find the elements of the matrix-column ®, (s):

    Dip(s)Se"r *O(sy Sa"

    A (s)/P(s)"

    (Se

    nye

    k-1—n+m

    ®1,(s) = s*~" G(s) —

    Dae

    qz—1-15!

    I=0 k—1—n+m

    ls 1H (s)/F(s) —

    dX i peovates

    Reducing the expressions of 1% (s) (k Snr

    (EF =a — In ly oe

    lee..26 n) to a common

    denominator and using formula (6.5.7) we obtain after simple but rather cumbersome evaluations

    #11(s) = H(s)/F(s),

    G14(s) = Az(s)/F(s) (k=2,..., 7). (6.5.11)

    Here

    HIG)

    oe

    GUS

    ee aay,

    (6.5.12)

    Hi(s) =) cers” (kan—m+1,...,),

    (6.5.13)

    n—1

    r=0

    6.5. Applications.

    Shaping Filters Design and Modeling

    629

    where k-1

    C=

    yi

    Q—k+4r411

    r=)

    Fm

    =n

    k— 1),

    Qj—k+r411

    (r=m—n+k,...,k—-—2)

    l=k-1-r

    k-1 Cha

    SS l=n-—m

    (only ifn —m

    > 2), k-—l+n-r

    ep

    Ze

    aid

    AP oe

    1)

    (6.5.14)

    i=-

    For calculating the elements of the matrix kz, (0) determined by formula (6.5.9) we may use formula (6.1.27). < Example

    6.5.1. Design the shaping filter for a stationary random func-

    tion X(t) with the spectral density

    S3(@) =

    1)

    To

    a

    tw

    (I)

    In this case the numerator P (w) represents a constant and the denominator has

    two pure imaginary roots +i@. F(iw) = 1w+ad.

    Consequently, we may assume H (iw) = 4) Da/n,

    Then we obtain

    sole) = [®(iu))?, (8) =~. Thus

    (IN)

    the shaping filter represents in this case the first order filter with the time

    constant J’ =

    1/a and the gain k =

    4/ D/ra.

    The differential equation of this

    shaping filter has the form

    X+aX =/Da/rV, V being the white noise of the intensity

    (IIT)

    vy= 27.

    Putting H (iw) sin/ 2Day F(iw) = iw + a, we have

    so(u) = 5 |®()I?, (8) =

    V2Da

    sta

    (Iv)

    630

    Chapter 6 Methods of Linear Stochastic Systems Theory

    The differential equation of the shaping filter will in this case take the form

    X +aX = V2Dev ,

    (V)

    V being a white noise of unit intensity. Taking in both cases the random initial value Ki = X (to) with the variance co

    Ds. =

    Da

    / Sz (w)dw =

    [o-e)

    dw

    / aa?

    = ID

    (VI)

    we shall obtain at the output of the shaping filter the stationary random process

    X(t) with given spectral density Sz, (w). Example

    6.5.2. Design the shaping filter for a stationary random func-

    tion with spectral density

    mj)

    so(#)

    (a+SR ywo)d*+(aCe) Dnm ee EEN wow? ee b4+42(a2 —w?)w? + w4

    In this case the numerator

    9 et _ 42 =

    vf

    P(w) has two roots WwW

    =

    9

    I

    .

    +764,

    (1) b? =

    b?(a

    ate ywo)(a = ywo)71, and the denominator Q(w) has four roots +(wo + ia). Selecting the roots disposed in the upper half-plane according to the outlined method

    H (iw) = V/2D(a — yuo) i(w — iby) = /2D(a — yuo) (iw + 1), F (iw) = i(w — wo — ia) i(w + Wo — ia) = (iw)? + 2a(iw) + b?. Then we shall have

    vl

    ee

    so(w) = 5— |®(t)|"

    _ A(s) _

    , O(s) = 110

    V2D(a— ywo)(s + 61)

    ee

    (II)

    The differential equation of this shaping filter has the form

    X + 2aX +0?X = /2D(a— ywo)(V + 61V),

    (IIT)

    V being a white noise of unit intensity. To this equation corresponds the set of two

    first-order

    Eqs.

    (6.5.6)

    for the

    vector

    random

    function

    Xj (t)

    =

    [X(t)

    KG) —nVG)| Xi = MatonV

    Xia =

    0

    Re aX

    EV,

    (IV)

    6.5. Applications.

    Shaping Filters Design and Modeling

    where according to formulae (6.5.7) qj = Xx (by =

    2D(a — ywo), —

    631 /2D(a

    = ywo)

    2a).

    For determining the covariance matrix of the initial value X}9 =

    X,1 (to) we

    find by formula (6.5.11) the elements of the matrix-column ,(s):

    928 — qb?

    qi(s+ 61)

    Pee ee aE deeb cE heen opera a Ou After this we find by formulae (6.5.10), (6.5.4), (6.5.8)

    years

    wee

    Yo

    ee

    ke,,(0) = 24 / ee de Sis) an J \(w)? + 2a(iw) + 6?|

    (VI)

    —oo

    Biel @) ete fight LeMert 2 a) dw 2m

    )? W) |(iw)? + 2a(iw) + 62

    bf

    ec2t ie

    = —D(a— (a

    wo)

    ywo), ((VII )

    \2 4.bac?

    ke(0) =2agoffate |(tw)? + 2aiw + b?| —oo

    =D G — by/a® — y2w? + 2a(a — 10) : Taking the random initial value X;9 =

    (VII)

    Xj, (to) with the covariance matrix whose

    elements are determined by the derived formulae we shall get at the first output of the shaping filter the stationary random function X (t) with given spectral density

    Sz(w). 6.5.2. Shaping Filter for Vector Stationary Process Analogously the problem of finding the shaping filter for a stationary vector random function X(t) with a rational spectral density Sz (w) is solved. In this case

    all the elements of the matrix Sz (w) represent the ratios of the polynomials in W. In accordance with (6.1.22) for finding the shaping filter for X(t) it is sufficient to represent the spectral density Sz (w) by the formula

    S_(w) = F(iw)~1H (iw) H (iw)* F(iw)7**,

    (6.5.15)

    where F' (s) and H (s) are the matrices whose elements are all polynomials in 5, and all the roots of the determinants

    of these matrices

    posed in the left half-plane of the complex variable s.

    F(s) and H(s) are disThe representation of the

    632

    Chapter 6 Methods of Linear Stochastic Systems Theory

    matrix of spectral and cross-spectral densities Sz (w) of the components tionary vector random function in such a form is called a factorization trix Sy (w). After factorizing the spectral density Sz (w) we may

    of a staof the ma-

    write the differ-

    ential equation of the shaping filter, transform it into a set of the first order differential Eqs.

    (6.5.6) and find the initial conditions for them providing the statio-

    narity of the process at the output of the filter in the same way as in Subsection 6.5.1. Example

    6.5.3. The spectral density of two-dimensional stationary vec-

    tor random function X (t) is determined by the formula

    ki = aj tw

    S,(w) =

    ky ky me QQ — (a1 — a)iw +w 4 k

    kik2 102+ (a) — a2)iw +w 1

    as + uy

    .

    a) I

    Find the shaping filter for X(t): In this case

    ale) Wei x |

    Od

    (2 ry |[ a

    0

    (11)

    (a2 — iw)7} |

    aril i

    ao+s

    athe lire

    (111

    and the set of differential equations of the shaping filter has the form X,+0,X;

    =kV,

    X_o+a2X2

    = keV,

    (IV)

    V being a white noise of the intensity 27. Taking the random initial value of the vector Xg = X (to) with the covariance matrix [o-e)

    Poy o| )

    hig / so(w)

    me

    k?/ax [Oy WK, longi

    oaaet

    2akyko/(ay 1%2 1k

    1

    +

    /og

    2 a2) :

    (V)

    —0o

    we

    shall obtain at the output of the filter the two-dimensional

    stationary vector

    random process with given spectral density S_ (w).

    6.5.3. Shaping Filter for Process Reductble to Stationary One The method

    outlined in Subsections

    6.5.1 and 6.5.2 may

    sign filters for random processes reducible to stationary ones.

    also be used to de-

    The case of the pro-

    cess reducible to stationary by the transformation of the process itself is trivial.

    If

    6.5. Applications.

    Shaping Filters Design and Modeling

    633

    X (t) = W(X, (t), t) where Xj(t) is a stationary random function, then substituting this expression into differential Eq. (5.4.2) we obtain an equation with a stationary random function in the right-hand side. Therefore it remains to consider the case of a process X (t) reducible to stationary by the transformation of the argument.

    D Leek

    (

    ty v(X1(~(t)),t) where Xj(s) is a stationary random func-

    tion with zero expectation and a rational spectral density, and

    monotonous increasing differentiable function.

    s =

    y(t) being a

    After constructing the shaping filter

    for the random function X 1(s) we write the differential equation of this filter and

    the corresponding initial condition in the form dX; = aX,ds a bdW, , where

    a and 0 are constant

    matrices

    X1((to))

    =X

    of the corresponding

    (6.5.16)

    dimensions,

    and

    Wj

    =W, (s) is a process with uncorrelated increments with the covariance function

    Kw, (s1, $2) = ki(min(s1, s2) , ki(s) = ki(so)

    +11(s— 80),

    (6.5.17)

    VY; being the constant intensity of the stationary white noise Vj (s) = Wi (s), and

    i) = p(to). Let us consider the random process W(t) = Wi(¢(t)). It is evident that W(t) is a process with uncorrelated increments as the monotonous increasing function

    § = p(t) maps any nonintersecting intervals on the S-axis on nonintersect-

    ing intervals of the t-axis. The covariance function of the process W(t) is determined by the evident formula

    Kw (ti, t2) = Kw, (p(t1), p(t2)) = kmin(y(ti), p(t2)),

    (6.5.18)

    where t

    k(t) = ki(y(t)) = ki(y(to)) + rile(t) — p(to)] = k(to) + [veer to

    Hence, it is clear that the intensity of the white noise V(t) = v(t) = 1 9(t). The change of variables

    W(t) is equal to

    s = p(t) in Eq. (6.5.16) yields

    1X,(p(t)) = aplt)Xy(olt))at + bdW, Xi(vlto)) = Xo. Introducing the random process X92 (t) = XG. (y(t)), we represent Eq.

    (6.5.19) (6.5.19) in

    the form dX»

    =

    ap(t)Xodt ae bdW ,

    Xo(to) =

    Xo .

    (6.5.20)

    634

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Thus the shaping filter in this case is described by Eq.

    (6.5.20) and by the formula

    X(t) = 4(Xo(t),2). 4

    Note the fact that after writing the equation of the shaping filter for the random function X 1 (s) in the form of an ordinary differential equation with a white noise in

    the right-hand side and performing the change of variables

    s = p(t) in this equation

    we shall get in the coefficient in front of the white noise in Eq. (6.5.20) the additional factor p(t). But such a derivation of Eq.

    (6.5.20) is false. This illustrates the fact

    that ordinary rules of the change of independent variable are not applicable when

    there is a white noise in a differential equation.

    It is necessary to remember

    this

    when we deal with stochastic differential equations. The

    covariance

    function

    of the

    initial

    value

    of the

    process

    X1 (s),

    Xo

    Xj (y(to)) is determined by Eqs. (6.5.9)—(6.5.11). Example

    =

    Co

    6.5.4. The random function X(t) with the expectation Mz (t)

    + cit and the covariance function:

    Kz (ti, ta) =

    reduced to a stationary one by the transformations:

    Debts +t2)—@ |#1-23|

    is

    X (t) =co+tectt+ e#* Xy (s),

    s = t?. Here X 1 (s) is a stationary random function with zero expectation and the covariance function ka. (c) =

    De-«lel , 0 =

    81 — Sg.

    This covariance function

    corresponds to the spectral density Sz, (w) ss Da/n(a? te Papas Using the method of Subsection 6.5.1 we find the differential equation of the random function Xj (s):

    dX,/ds

    =

    =a

    +vV 2DaV,

    )

    (1)

    V; being a white noise of unit intensity. Passing to the variable t we replace Eq. (I) by Eq. (6.5.20) for the random function X92 (t) = AG (t?) which has in this case the form

    Xo = —2atX,+ V2DaV,

    (II)

    where V is a white noise of the intensity 2¢. The initial condition for this equation in accordance with Eq. (6.5.20) has the form: X9 (to) = Xo, where XQ is a random variable with zero expectation and the variance

    ir

    Dey = f 52,(w)dw =

    Dat =

    odee

    (III)

    6.5.4. Shaping Filters for Stationary Functions of Vector Argument Many problems of practice lead to the partial differential equations which contain stationary or reducible to stationary random functions of a vector argument.

    6.5. Applications.

    Shaping Filters Design and Modeling

    635

    For the application of the methods of the linear stochastic systems theory to such problems it is expedient to present the random functions entering into the equations

    as the solutions of the linear stochastic differential equations. Here we give a general method

    of finding the linear stochastic differential equations in the correspondent

    H-spaces for a wide class of the stationary functions of a vector argument.

    Let X (é, zr) bea

    scalar stationary random function of the scalar variable f which

    we shall assume as the time and V-dimensional vector variable 2; k(ty —to,%,- £2) be its covariance function which is equal to k(t, — to, 0) alec k(0, |



    co



    somal

    r2) at t; = tg = t. We shall consider here normally distributed random

    functions X (t, zr) satisfying two conditions: (i) at a given & the random function Xz (t) has a rational spectral density independent of ©

    cE

    s(w) = = / k(z, 0)e*”"dr = P(o?)’

    (6.5.21)

    where P(w?) and Q(w?) are the polynomials and Pw?) has no zeros on the real axis;

    (ii) at a given f the covariance function of the random function Xi(z) at some C > 0 anda

    >

    0 satisfies the inequality

    |k(0, 21 — t2)| < Ce~%!*1-721 For obtaining a stochastic ee

    differential equation for the random

    (6.5.22) function

    X

    (t) considered as a random function of the time ft at a given © we may apply

    the same method

    of shaping filters which was used for the scalar random process

    (Subsection 6.5.1). According to this method it is necessary to present the spectral density Sz (w) in the form

    Vo

    $_(w) == =Qn

    H (iw) |? F(iw)

    (G2

    :

    where F(s) and H(s) are the polynoms whose roots at

    (

    )

    $ = 1W be in the left half-

    plane; lp is the constant which is usually obtained for the sake of convenience to write

    the equation for X(t, ay Formula (6.5.23) allows to treat Xz (t) as the result of the passage of the white noise of the intensity /o through a stationary linear system with

    the transfer function H(s)/F(s). differential equation

    where

    V

    =

    This fact will give for X, (t) a linear stochastic

    F(0/dt)X = H(a/atyV,

    (6.5.24)

    V(t, r) ——Ve (t) is the stationary normally distributed white noise

    which depends on time f at a given © whose intensity is equal to lp.

    636

    Chapter 6 Methods of Linear Stochastic Systems Theory

    If we have n

    F(s) = ) Saget Pres l= S > be s* Sanrcons

    (6.5.25)

    k=O

    k=0

    then Eq. (6.5.24) represents a stochastic linear differential equation of the n‘} order

    which should be assumed as the Ité standard equation:

    dZ = AZdt + qdW. Here Z is the vector with the components

    trix with the elements =i

    ie,

    aj; =

    65,641) Grae

    Z; =

    (6.5.26)

    X, Zo MA wt 9 Zn, A is the ma-

    erg

    ee , N, 7 is the vector with the components Gj =

    =

    Lid

    —ana;-1,

    --* = Q(n—m-1

    1

    = 0, Qn—m

    see k-1

    Gd=4,

    b,—-E—

    Doge An—k+hQh

    ee 2s

    a

    O

    eee

    h=n-—m

    We=

    Wt, r) is the scalar Wiener process with the values in the space of the

    functions of the variable £ whose weak m.s. derivative is the white noise

    V=

    V, (t)

    of the intensity V:

    ween

    t

    /Tate

    (6.5.27)

    0

    Eq. (6.5.2) is reduced to Eq. (6.5.26) by a formal change of the variables (6.5.6):

    Geek Ze

    OZ Ot

    Dee, See

    OZ =——aV,

    he aan

    in

    de

    kon—m,...,n-1,

    (6.5.28)

    where gx are determined by formulae (6.5.7) and quq = qq. v. We shall consider the random function X (gf:zr) as a random process with the values X+(-) in the H-space L2(D) of the functions of the variable x,

    D C Roe

    Then as a covariance operator of the value of this process at a given ft will be an

    integral operator with the square-integrable kernel kz (0, r—€ ):

    Kap = |ke(0,2—€)o(6)dé D

    (6.5.29)

    6.5. Applications.

    Shaping Filters Design and Modeling

    637

    The analogous formula takes the place for the covariance operator KX , , of the vector process

    Zt (x):

    pines ikz,(2— Op(E)ae , D

    where the kernel Rats

    = r2) =

    kz(0, |

    r2) is a covariance function of the

    process Z1(2x) at a given t. For finding the kernel v(x = £) of an operator of the intensity of the Wieter process W

    we use an equation which connects the covariance

    matrix k,, of the value of the process

    Z = Z(t, x) at a given t with the intensity

    V of the Wiener process W. This equation has the form

    Ok, Ot As the process

    s

    : “i= Aky, ike

    Abe (qh Vc

    Z is stationary then the covariance matrix

    (6.5.30) ee of its value at a

    given ¢ does not depend on f and is expressed in terms of its covariance function

    kz (ty —t2,21— r2) by the formula k,, = k,(0, Li- r2). Therefore Eq. (6.5.30) has in a given case the form

    0 = Ak, (0,21 — 22) + kz (0,21 — 22)A? + qq? v(21— 22).

    (6.5.31)

    Matrix Eq. (6.5.31) represents a system of n(n + 1)/2 scalar linear algebraic equations with n(n + 1)/2 unknowns, namely n(n + 1)/2 — ments of the matrix k,(0, i of the matrix k,(0, i

    ZAr(t, r) =

    r2) and the unknown

    £2) is known:

    Kzaoe\U; Cie

    1 of the unknown

    ele-

    y(avy = Zo). One element r2) =

    kz(0, iy = r2) as

    X(t, 2). After excluding from (6.5.29) the unknown elements of the

    matrix k,(0, lees r2) we shall find the unknown function v(2x4 = to):

    v(@; — £2) = ckz(0, x1 — 22), where C is some constant.

    At 21 =

    2

    =

    (6.5.32)

    Z@ this formula gives the intensity Vp of

    the process Wt, zr) at a given ©: Vo = ck, (0, 0).

    From (5.9.3) it follows the formula for the kernels of the operators: t

    kw(t, 21 — £2) = fre — x2)dr = ckz (0,21 — a2). 0 Here k, is the kernel of the operator

    Ky (t) in (5.9.3) at a given f.

    (6.5.33)

    Condition

    (6.5.22) provides the trace-type of the operators ky and V as in this case by virtue

    638

    Chapter 6 Methods of Linear Stochastic Systems Theory

    of the square integralability of the kernel k. for any orthonormal basis {pr (x)} in

    TAD):

    adé (E)d) (x by, ize — €)pepr k=1pp co

    =c)) x2(0, 2 ~ pn(a)pe (€)dadé Let X(t, x) =

    o( Xi (p(t), x), t, x) where X(s, x) is a stationary random

    function, p(t) is monotonously increasing differentiable function; (0) =

    0. After

    writing Eq. (6.5.26) for the random function X 1 (s, x),

    dZ = aZds + bdWy, where Z\(s, zr) =X

    (6.5.35)

    (s, 2), and Wi(s, 2) is a Wiener process with the kernel of

    the covariance operator V(x = L2)s and after changing the variables

    5 = p(t) we

    obtain a linear stochastic differentiable equation in the H-space Ly (D):

    dZ = ApZdt + qdW.

    (6.5.36)

    Here W = Wit, xr) =W, (s, zr) is a Wiener process with the kernel of the covari-

    ance operator of its value at a given t equal to k(t, 1—-

    £2) — yt (a1 — r2)p(t)

    6.5. Applications.

    Shaping Filters Design and Modeling

    639

    where 1’; (x4 = £2) is the kernel of an operator of the intensity of the Wiener process Wi(s).

    dq

    6.5.5. Software for Analytical Modeling Problems At practice analytical modeling problems for the linear stochastic systems are

    usually solved by means the dialog packages.

    of the universal or the specialized computer libraries and

    The corresponding algorithms are based on equations given in

    Sections 6.1-6.3, 6.5.



    Let us consider some

    lized dialog packages

    examples

    of analytical modeling solved on the specia-

    “StS-Analysis”

    (Versions 1.0 and 2.0) based on equations of

    Sections 6.1 and 6.5. Working with any version of the “StS-Analysis”

    package, an user must

    only

    type on the display with the aid of the keyboard the original stochastic equations in the natural mathematical form without any special programmers tricks, introduce the corresponding initial conditions and indicate which characteristics must be calculated. Once the original stochastic equations are introduced the user may

    correct them,

    i.e. to remove some equations, to replace some of them by others, to correct errors

    and so on.

    When

    computations.

    all the data are introduced and checked the user has to start

    Then the package “StS-Analysis” automatically derives the equations

    for the corresponding distribution parameters and solves them. The user may observe visually the behaviour of a system under study in the process of calculations by tracing the graphs or the tables of current values of various characteristics on the display and

    get the copies of these graphs and tables from the file after the calculations finished.

    So the use of any of the versions of the package

    “StS-Analysis”

    are

    does not

    require that the user know mathematics and programming. Any version of the “StS-Analysis”

    package automatically derives the equations

    for the distribution parameters from the original stochastic equations introduced by the user in the natural mathematical

    form and solves these equations.

    It should

    be remarked that the “StS-Analysis” package does not simulate a stochastic system and does not use the statistical simulation (Monte-Carlo) method but performs the theoretical calculation of the necessary characteristics of a system behaviour by the linear stochastic systems theory methods.

    Example

    6.5.5.

    For the stochastic two-dimensional linear discrete sys-

    tem

    Y2(1 +1) = —bY} (I) + a¥o(1)

    (11)

    (a, b being known constants) the exact equations for the expectations, the variances and the covariances (Subsection 6.1.9) have the following form: my(I + 1) =am, (1) + bm2(1),

    (IIT)

    640

    Chapter 6 Methods of Linear Stochastic Systems Theory

    m2(l + 1) = —bm, (I) + am2(l),

    (IV)

    D,(1+ 1) = a? Dy(I) + 6? Do(l) + 2abKy2(1) + 1,

    (V)

    D,(I + 1) = b? D,(1) + a” D2(1) =— 2abK12(1),

    (VI)

    K12(1+ 1) = —abD, (I) + abDo(1) + (a? — 6?) Ky2(I). Graphs of m,(l), m2(l), D,()), D2(l) and Ky2(1) at initial conditions: =F

    EY;,(0) = JONG (0) = DY2(0) = EY;(0)Y2(0)

    a) a=1.0; b=0.25

    (VI) EY2(0)

    = 0 are given on Fig.6.5.1.

    b) a=1.0; b=0.25

    ma2(I)

    Fig. 6.5.1

    6.6. Applications.

    6.6. Applications.

    Optimal Online Linear Filtering

    641

    Optimal Online Linear Filtering

    6.6.1. Equations of Linear Filtering The optimal filtering problem may be solved completely in the case of linear

    Eqs. (5.10.4) and (5.10.5)

    dY = (bY +.b,X + bo)dt + yidW,

    (6.6.1)

    dX = (aY +a,X + ao)dt + ydW,

    (6.6.2)

    where the coefficients @, @1, do, 6, b1, bo, w and 1 in the general case depend on

    time t. In this case the distribution of the process LYat)2 X(t)2 |i is normal if the distribution of its initial value

    ben XK

    ae

    is normal.

    Consequently, the conditional

    distribution of the state vector of the system relative to the process fe is also normal

    and for its determining it is sufficient to find the conditional expectation X and the conditional covariance matrix

    R.

    For this purpose

    we use formulae

    (5.10.31) for the stochastic differentials of the random variables

    (5.10.35) and

    X and FR.

    > Substituting into (5.10.31) and (5.10.35) the expressions p(y, #,t) =

    ay

    + a,x + ao, yi(y, 2, t) = by + b1x + bo we obtain

    dX = (aY +a1X + ao)dt + {E[X(X? — X7) |Yf]of + put }(vivbT)1[dY — (bY +b, X 4 bo)dé],

    (6.6.3)

    dR = E|(X — X)(YTa? +. XTal + a?) + (aY + a,X + ap)(X7 — XT) + dy" | Yi Jt

    — B[(X — X)(¥PH + XT OT + 6g) + but |Yi] x (pvt )- 1E[(bY +b,

    X + bo)(X7 — XT) 4 divyt | Yi]dt

    + OLD an bla — Fa) = XYXP = XP) [VE l=1

    hs

    + B[X — X |¥i]67 + hiE[X? — X7 |vay) x

    dY, —

    (3binYR + Yo 1p Xp + mna hail

    where @j,

    (6.6.4)

    p=1

    are the elements of the matrix

    @ =

    bT (yy vyt)-}

    and (3, is the pth

    column of the matrix vyt (v1 vy? )-). Taking into account that

    B[X(X? — RT) |¥f] = B[(X — X)X7 |YZ] eee spo) Mlk,

    642

    Chapter 6 Methods of Linear Stochastic Systems Theory

    BX — Xi Yj) =—Xi— X=,

    E[(X — X)Y? |Y,,] = E[X —X |¥,,]¥* =0 and that the conditional central third order moments are equal to zero in consequence

    of the normality of the conditional distribution we transform formulae (6.6.3), (6.6.4) into the form

    dX = (aY + a,X + ao)dt

    + (ROP + pv )(divd?)—1[dY —(bY + b1X 4 bo)dt], (6.6.5) dR =[a,R+ Ral + pvy7

    — (ROT + vd (vd) (bR+ viv" )ldt.
    to. This optimal estimate Xs unbiased at all

    t > to, as by virtue of the formula of total expectation and formula (5.10.6)

    B(X; — Xt) = B{E(X; — X1) |Y¥i]} = E(X: — Xi) = 0. Eqs.

    (6.6.6)—(6.6.8) solve completely and exactly the problem of optimal linear fil-

    tering of the state of a linear system defined by Eqs. (6.6.1), (6.6.2).

    6.6.2. Kalman-Bucy Filter Eqs. the case of

    (6.6.6) and (6.6.8) were fifst obtained by Kalman and Bucy in 1960s in a = b = 0. This case corresponds to the practical problem of filtering of

    the signal X determined by (6.6.2) at a = 0 when the signal b; X + bo is measured with an additive disturbance representing a white noise.

    In this case Eq.

    (6.6.1),

    (6.6.2) may be rewritten in the form

    X =a,X +agt+yV, Z=uX+b4+urV.

    (6.6.10)

    V being a white noise, i.e. the derivative of a Wiener process Wit). The equation

    of the optimal filter (6.6.8) has in this case the form

    5 Ts PO RT Eq.

    (6.6.11) determines

    EN Se HE

    the structure of the optimal filter.

    (6.6.11) Namely,

    the optimal

    filter may be obtained from the given system (the system generating the signal X; Fig.6.6.1a by the installation of the amplifier with the gain G before its input and by applying negative feedback containing the amplifier with the gain b; to the system obtained (Fig.6.6.1b). Receiving the measured process Z at the input with the function of time bo, subtracted from it the filter obtained in this way will give the

    optimal estimate

    X of the vector

    X at the output.

    a)

    Fig. 6.6.1

    644

    Chapter 6 Methods of Linear Stochastic Systems Theory

    As the process V(t) representing the integral of the measured process is not

    used

    Xy =

    EX

    in the optimal

    expectation

    Z (t)

    of the vector

    Xo,

    should be assumed as the initial value of the estimate x at b=

    and the conditional



    filter, the conditional

    EXo)(X¢

    covariance

    ae EXé)

    matrix

    of the random

    vector Xg,

    Ro

    =

    Lo;

    E(Xo

    should be assumed as the initial value of the matrix

    R

    while integrating Eq. (6.6.6) and calculating the gain (3. But these conditional characteristics are usually unknown.

    This compels one to take arbitrary initial of K and

    R. Certainly, x will not be the optimal estimate of the vector X in this case and

    may be only an asymptotically optimal estimate if only the first equation of (6.6.10) and Eq.

    (6.6.6) describe a stable process (i.e. if the system described by the first equation of (6.6.10) and Eq. (6.6.6) is stable). It goes without saying that the gains ( and b; represent in the general case

    the matrices.

    So we consider here the amplifiers of vector signals performing linear

    transformations of the vectors received at the input which are determined

    by the

    corresponding matrices, i.e. by matrix gains. Optimal linear filters constructed by the method

    outlined are usually called

    Kalman-Bucy filters. For the linear stochastic systems the problems of Kalman—Bucy filter design are usually solved by means

    of the universal or the specialized computer libraries and

    dialog packages. Example

    6.6.1.

    Find the optimal filter for filtering a sinusoid signal of

    a given frequency Wg which is measured with an additive normally distributed white noise independent of the signal.

    The sinusoid signal X ; (t) and its derivative X 9 (t) = xe (t) may be considered as the components of the state vector of a system described by the differential equation

    Eu 5 dt |Xo}

    ed ak es |

    —w2

    0

    Bes Xo]

    (I)

    The measured process is determined by the formula

    Y=Z=X,4+V=fl0X+V. Consequently, in this cased

    = dg)

    = b= bo = yp =0, db) = [1 0], Dy =i,

    Og

    1

    Eq. (6.6.6) represents the set of equations

    Ri; = 2Rip — vy R?,

    Rio= we Ri + Reo — v1 Ri Rio,

    (II)

    6.6. Applications.

    Optimal Online Linear Filtering

    645

    Rog = —2w2 Riz — v-1 3.

    (II)

    After determining 11, Rj2 and Roo by the integration of these equations with the initial conditions R41 (to) = we find m.s.e.

    EX??, Ri2(to) 22 CRY

    of the filtering of the signal X;, R11 =

    Moa Ro2(to)

    El(X,

    = Aa)



    EX

    |Y;7] and

    of its derivative X2, Roo = El(X2 = a)" |Yi} and the gain

    pauai RH awnt |Fh ea | sospalesesl = els (IV) Eq. (6.6.8) for optimal estimates Hew x has in this case the form

    ]-[2: Renee. 0 77

    The block diagram of the optimal filter found is shown in Fig.6.6.2.

    Fig. 6.6.2

    6.6.3. Case of Equations Linear in State Vector Let us now consider the more general case of Eqs.

    (5.10.9) linear only in the

    state vector of the system

    dY = [bi(Y,t)X + bo(Y, t)]at + vi (¥, t)aW, dX =[a,(Y,t)X + ao(Y, t)]dt + ¥(Y, t)dW.

    (6.6.12) (6.6.13)

    In this case for any realization y(t) of the measurement process Y (t) Eq. (6.6.13) is linear in X. Therefore, we may make the conclusion that the conditional distribution

    of the random vector X; is normal for any realization Vi = {y,;7 € [to, t]} of the measurements Mes at all t > to if the initial distribution of X is normal.

    646

    Chapter 6 Methods of Linear Stochastic Systems Theory

    > Assuming

    the conditional distribution of X; relative to Le as normal we

    write for this case formulae (5.10.31) and (5.10.35) for the stochastic differentials of the optimal estimate X and the covariance matrix R of the state vector of the

    system. Substituting into (5.10.31) and (5.10.35) the expressions gly, x,t) == ai(y,t)r+ao(y,t),

    yily, 2,t) = bi(y,t)a+bo(y, t), (6.6.14)

    taking into account the independence of y on X and recalling that all the central moments of the third order are equal to zero for a normal distribution, we shall obtain in the same way as in Subsection 6.6.1

    dX = [a,(Y,t)X + ao(Y, t)]dt + [Rb (Y,t)? + (v7 )(Y, t)] x (WyvyT)—1(Y, t){dY — [bi (Y,t)X + bo(Y,t)]dt}, (6.6.15) dR = {a,(Y,t)R+ Ra,(Y,t)? + (pvyp7)(Y, t) — [Rbi (Y, t)?

    + (bud )(Y M@ivdt)(VY NY HR + (buy )(¥,t)]}at. 4 (6.6.16) These equations represent a closed set of equations determining B¢ and R as

    well in the case of linear filtering.

    Therefore, the optimal estimate X oh the state

    vector of the system X and its conditional covariance matrix FR characterizing the accuracy of the optimal estimate x , may be calculated by the integration of the set

    Egs. (6.6.15), (6.6.16) while obtaining the results of measurements. Notice that contrary to linear filtering it is impossible in this case to calculate R beforehand when the results of observations are not yet received, as the coefficients of

    Eq. (6.6.16) depend on the results of measurements.

    Therefore the optimal filter in

    this case must perform the simultaneous integration of both Eqs. (6.6.15), (6.6.16). It leads to the essential increase of the order of the optimal filter.

    > Now we shall prove that the conditional distribution of the state vector of the system is normal in this case.

    For this purpose it is sufficient to show that Eq.

    (5.10.21) for the conditional characteristic function ge(A) of the vector X; has the solution

    ;

    gi(A) = exp {arg _ 7 Ra} where

    (6.6.17)

    and R are determined by Egs. (6.6.15), (6.6.16). Substituting into Eq. (5.10.31) expressions (6.6.14) of the functions Y and

    we reduce it to the form

    dge(d) = El{id? (aX + a0) — 5?" Jei*™* |v, at 1

    5

    6.6. Applications.

    Optimal Online Linear Filtering

    647

    +E[{(XP—X?OP ar? pot pe * | VE] (dived?) [d¥ —(b1 X+50)at] (6.6.18) where

    the arguments

    of the functions

    @1, ao, 61, bo, wy and yj are omitted for

    brevity. Taking into account that

    Elixe?’* |yt] =

    0gt(A) OX

    (6.6.19)

    we get from Eq. (6.6.18)

    dg:(A) = [xP “8 +, (aPa a A" wv" 1) |dt ge(A) bypr +(X*ws by = OT aN aly — id* ag 4 dy; fb )gt(A) = i

    x(dvyt)-! [ay Ly bpXp bo)at|

    (6.6.20)

    We evaluate separately the left- and right-hand sides of this equation for the function

    gr(A) determined by formula (6.6.17).

    Using differentiation formula (5.1.15) of a

    composite function in the case of a Wiener process

    W, taking into consideration

    that the stochastic differential of the process X(t) is determined by formula (6.6.15) and that (Rb? ar wyt

    )\(bivet)-

    dy plays the role of the matrix Y in (5.1.15)

    in this case, we find the stochastic differential of the process gi(A) determined by

    formula (6.6.17):

    dg:(A)-= n(A){ anak = strlAN? (RO + pvt)

    (Hie) ied avd MR + divdT ah — 57 dRda(>) = w(a){ ia?(ax + ao)

    EAT (OE+ WE)

    ALR avy )A]

    -5\ aR + Raf + pvp" — (Rot + wrt) x(vivel) (bi R+ bavsTy)a hat+ gi(A)irT (ROT

    Chapter 6 Methods of Linear Stochastic Systems Theory

    648

    +yudt (dvd)dY — (b1X + bode] 1

    ;

    = g:(X) Euce + ao) — 5) (ak + Raf + wera] dt

    +9:(A)iA7 (REP + oud (dived)? |ay — (bX + boat]. (6.6.21) As ae

    = May Rd, we get from (6.6.21)

    dgi(A) = gt(A Ec

    i

    1

    + ao) — A? a, RA — Swit

    |dt

    +9i(A)iXT (ROT + pw (davdT)? |a¥ —(b1X + bo)at]. (6.6.22) For evaluating the right-hand side of Eq. (6.6.20) we notice that (6.6.17) yields

    092(A)

    ay

    eT gi(A)(iX — RA).

    Using this formula we find

    [Pa eu + (Pa = SAT yey a)0) |dt

    Li[ieee Pxate = AT wat Jada) |

    x(wivdt7} |aY — (6X + bo)at|

    re Eucr. hp es +g(A)(iA? RO? + id Eq.

    Swi

    |dt

    v7 )(div dT) [dY — (b1X + bo)dt]. (6.6.23)

    (6.6.23) is identical to (6.6.22).

    Consequently,

    the normal conditional cha-

    racteristic function determined by formula (6.6.17) with the parameters X and R

    determined by Eqs. (6.6.15), (6.6.16) satisfies Eq. (6.6.20). This proves the normality of the conditional distribution of the random vector Xz and the validity of Eqs.

    (6.6.15), (6.6.16) which determine the parameters NG and R; of this distribution if the initial values of Xi and FR at the moment tg represent the conditional expectation

    and the conditional covariance matrix of the random vector Xg = Yo= Vee respectively.
    to using the measurements

    Find the optimal esti-

    results of the process Y in time

    interval [to : t]. Replacing the vector 0 by the random process O(t) which is determined by the differential equation dO(t)

    =

    0, and assuming

    O as the state vector X

    the corresponding system we obtain the differential equation of form Eqs.

    of

    (6.6.12),

    (6.6.13) at a1(Y,t) = 0, ao(Y,t) = 0, ¥(Y,t) = 0. Eqs. (6.6.15), (6.6.16) take in this case the form

    dO = Rb, (Y,t)? (divdT)-1(Y,2){ay — [bi (Y, 1) + bo(¥,t)]ae} , (1) dR = —Rb,(Y,t)? (divd)-1(¥, tbr (Y, t)Rat.

    (III)

    As the parameter @ is unknown and may be nonrandom the initial values of O and R are taken arbitrarily. The equations obtained will give in this case the estimate O which will be optimal assuming that the parameter @ is random and has the normal distribution which the expectation Og and the covariance matrix Ro.

    Problems 6.1. De

    Show that for the stationary random processes with covariance functions

    “210 os a|r|) and De-4l"|

    (1+ a|r| + a7?) the intensities of the equi-

    valent white noises are equal to 4D/ Q@ and 16D/ 3@ respectively.

    Calculate the

    correlation intervals. 6.2.

    Show

    that for the process reducible to a stationary one determined by

    formula Y (t) = bo(t) + bi(t) X(t), where X(t) being the stationary function the intensity of the equivalent white noise and the correlation interval are determined by

    650

    Chapter 6 Methods of Linear Stochastic Systems Theory

    a Calculate

    v(t) and

    T, for the process

    determined

    by formula

    Yb)

    =p

    (t)

    + bi(t)X(¢(t)), where p(t) being monotone increasing function. 6.3. Show that for the two-dimensional stationary system of Problem 1.3 with

    uncorrelated inputs Xj (t) and X9 (t) with spectral densities $1 (w) and S(w) the elements of the spectral density matrix Sy (w) of the process Y(t) =

    [Y (t) Yo (t) a

    are determined by

    :

    $11(w) = $1(w)|®11 (iw) |?+ s2(w)|®12(tw)/’, $12(w)

    =

    81 (W)®11 (tw) Bo) (tw) + $2(wW)®12(tw)H29(iw),

    $91(w) = 81(W)®11 (iw) Bo) (iw) + 82 (W)®12(tw) Do0(tw),

    822(w) = $1 (w)|®2i (tw)|?+ 82(w)|22(iw)|?. Find the cross-spectral density of the processes X(t) =

    [X1 (t)X2 aie and Y(t).

    6.4. Check that for the two-dimensional stationary system of Problem 1.4 with

    a stationary input X(t) = [X1 (t)X2 (t)|7 with spectral density s(w) the elements of the matrices Sy (w) and Sy (w) are determined by the formulae

    Syiyy = 811(w) Yi)|?+ 522(w) |Y2w)|? + 512(w)[¥1(w) oo) + Y2(w)vi(w)];

    Syiyo = + Syayx = Syay2 =

    —$11(w) pi(~)b2(w) + 522(w)p2(w) div) 812(w)(|¥i(w)|? — Iv2)]’), —$11(w)P2(w)P2(w)+ 812(w)([41(w)|? — |v2)]*), 811(w) |Yo(w)|° + s22(w) |Yr(w)|? — si2(v) [ti(w)42(w)

    ae V1 (w)po(w)],

    Se,y, = $11(w)¥i(w) + $12(w)p2(w),

    Saiy2 = —$11(W)b2(w) + si2(w)yi(w), Seay, = $12(wW)v1(w) + $22(w)y2(w),

    Swoy2 = —$12(w)P2(w) + 522(w)y1(w),

    Problems

    651

    ~~

    where

    Yi(w) = [e(w? +e? + w?) + iw(w? — e? — w?)] c?(w),

    Po(w) = [(w? +e? —w?) — 2eiw|c?(w)w, 9

    6.5.

    ~3(w) = (w? +e? — w?)? 4 4e7w?). Show

    that

    =—Zel,b=J,

    for the system

    of the third

    order

    of Problem

    1.9 at a

    s. (w) = sl, I being the 3 X 3 unit matrix, the spectral density

    of the output is equal to Sy (w) = sI/(w? aF 4e?). 6.6. Show that under the conditions of Problem 1.10 at X =

    [0 0V2Da Ey,

    V being a white noise of unit intensity, the spectral density of the output is equal to |®13(iw)|?

    Bie ola iaa*

    6.7.

    a

    ®13(t1w)O23(tw)

    ®13(iw)O33(tw)

    iw) D13(tw) Do3(tw)Pi3(tw)

    iw)|? —— |®a3(tw)|

    iw) 33 D3 (tw) 33 (iw)(tw)

    O33 (iw) D13(iw)

    O33 (iw) @o3(iw)

    |©33(iw) |?

    For the system with two degrees of freedom of Problem

    1.11 with the

    vector of the generalized forces representing a stationary process with spectral density

    SQ (w) show that the elements of the matrices Sq(w) and SQq (w) are determined by the formulae

    Sanq(W) = $11(W) Bar (tw) By (tw) + $29(W) Pp2(iw) O72 (iw) +s12(w)[ Dp,1 (tw) Dio (iw) + Dp2(iw) Py (iw) ],

    SQiqn () = $11(H) Par(iw) + $12(H) Pra(iw), SQaqn(W) = 812(W)Pni (tw) + 822(w)Pr2(tw) (h = 1, 2). 6.8.

    Find the spectral density Sy (w) and the covariance matrix

    Ky of the

    stationary vector process Ye) in the stationary system

    Y

    Yp0)!

    VAR Y4

    Oter

    thd wpe pas 0

    Yi

    0

    fede)

    0

    ohfestigert

    ae

    KOH

    Fhe

    TAOS

    1016

    Yo

    (1)

    ay

    pe Pe

    0

    0

    —w?

    —2€

    Y4

    q2

    where V is a white noise of unit intensity, and ¢1, 42, Wo, W1, €, €1, (€,€1 > 0) are some constant coefficients. 6.9.

    Find the spectral density Sy (w) and the variance Dy of the vertical os-

    cillations of the body of a car in linear approximation (Problem 1.15) assuming that

    the spectral density of the road microprofile is known, $] (w) =2 s(p)/v, p=

    Qn/c.

    652

    Chapter 6 Methods of Linear Stochastic Systems Theory

    6.10.

    Find the spectral density Sz (w) and the variance D), of the stationary

    random oscillations of an object with dynamic damper of oscillations in Problem 1.16 under the action of the random disturbing force with the spectral density sp (w). 6.11.

    For the stationary linear system of Problem 1.12 with three degrees of

    freedom at A = I, C = wi], Bo =)2el,| B! = C* =.0,4Qg(w) = (27)?sl (I being the unit 3 X 3-matrix) derive the formulae for the spectral density Sy (w) of the stationary process aah) = [qi(t) q2 (t) q3(t) qi(t) ga(t) s(t) (he Find also the covariance matrix Ky. 6.12. Derive the formulae for the spectral and the cross-spectral densities Sz (w) and Spy (w) for the system of Problem 1.13 at

    B =

    2€A and B =

    nC assuming

    that the spectral density Sz (w) of the input is known.

    6.13. ®(s) =

    Show

    that for the linear stationary system with the transfer function

    (dis + do)(c3s° + C28 +ces+

    cola. the input being a white noise

    v(codj+e2ds) with the intensity l/, the variance of the output is equal to Dy _TDaaleies 2eteay” Calculate also the covariance key:

    6.14. Show that the covariance function ky (r) and the spectral density Sy (w) of the random function Y(t) components

    =P

    Gl (t)X2 (t), where X, (t) and X9 (t) are the

    of the two-dimensional real normally distributed random process with

    the known covariance function and spectral density

    m= [EG tan| = [te are calculated by the formulae

    ky (7) = ki(r)ko(7) + k12(7) kai(7) + miki (7) + m5ki (7),

    Hee / Ge Sven Ode / somos (Out +m? so(w) +ms1 (w). Give the generalization to the case of Y (2) = xe (t)X2 (t) where Xj (t) and X9 (t) are the n-dimensional vectors. 6.15.

    Find the variances and the covariances of Y; and Y2 in Problem 1.6 at

    the input X being the white noise with the intensity V.

    6.16. Find the variance of Y; in Problem 1.7 for X being the white noise with the intensity V.

    6.17.

    Find the variance and the covariance function of Y; in Problem 1.7 for

    X being the white noise with intensity V.

    Problems

    653

    6.18. Deduce formulae for the expectations of the number of intersections

    Y of

    a constant level @ and of the number of stationary points Y; during the time J' for a stationary normally distributed random process X (t) with the expectation Mz and the covariance function ke(r):

    _T | k¥(0)

    Bie

    ,

    ee

    ereatiy eee 1) Fe

    (0)

    tesa rec

    6.19. Show that for the system

    Yj=3,

    iG

    (I)

    ap + V,

    Pa a

    where

    do is the constant,

    V is the white noise of the constant

    intensity

    V, the

    expectations, the variance, the covariance and the cross-covariance functions of the

    processes Y (t) and Y> (t) at zero initial conditions are determined by the formulae my,

    =

    agt? /2,

    K41(t1,

    m2:

    te) —

    Koi (ti, te) =

    =

    dot,

    bes

    ky

    =

    vt? /3,

    min(ty, t2),

    vt2/2 min(ty,

    tz),

    ky =

    Ky0(t1, ta) = Ko2(t1,

    vt?/2,

    kao =

    vt, /2 min(ty,

    vt,

    (II)

    ta),

    to) => vymin(t1, tz).

    (III)

    6.20. Prove that for the Langevin equations Y, =

    Y,

    Y> —

    —2eY> a5 V;

    (I)

    where € > (is the constant, V is the white noise of constant intensity l’, the variances and the covariance of the processes Yj (t) and Y (t) at zero initial conditions are determined by the formulae:

    ky, = v(4et — 3 + 4e—7** — e~ ***) /16e%, kia = vl — 2677? 6.21.

    e748") [8e*, »kop = (1 e5**")/4e.-

    (ID)

    Show that under the conditions of Problem 1.5, the generalized force Q

    being the white noise of intensity /, Eqs. (6.1.8) and (6.1.11) have the form

    kiy = 2ki2/A, ki = —Chi — (B/A)k12 + (1/A)ko2,

    koo = —2Ck12 — (2B/A)koo + v,

    (I)

    OK 11(t1, t2)/Ote = Kio(ti, t2)/A,

    aK i (ts 1s) /ots = —CK is(t1, 4)= BRia(hayt)/A) OKo1(t1, t2)/Ote = Koo(ti, t2)/A,

    Raa (ti, fly 2 CRA (Hs) = BRaa(ti,t2)/A.

    (I)

    654

    Chapter 6 Methods of Linear Stochastic Systems Theory

    Write the explicit formulae for the solution of Eqs. (I), (IJ). 6.22. Prove that the covariance and cross-covariance functions of the stationary

    and cross-stationary processes Yj (t) and Y2(t) of Example 1.1.9, the input X (t) being a white noise of the intensity

    , are determined by the formulae

    Bris = rar e~S“ol7l(cosw.7 + ysinw,|r]), 0 koo(T) = ies e~S¥el7l (cos WeT — ysinw,|T|),

    Vy kyo(T) = Seebes

    : a sinwelTke(y =¢/V1—¢7,

    we = wov/1 — C?).

    6.23. Show that under the conditions of Problem 6.21 the variance of the coordinate and of the impulse in the regime of stationary random oscillations, i.e. when the

    oscillations represent a two-dimensional stationary random process, are determined by the formulae kj; = Vi ZB:

    kia = 0, ko5 = Av /2B. Show that in the case of

    heavy damping (B 2A AG ), the process of stationary oscillations settling consists

    of two stages: first, the oscillations of the impulse settle, then follows the slow settling of the oscillations of the coordinate.

    6.24. = we

    Show that the set of Eqs.

    (6.1.8) for the system of Problem 1.3 at Zp

    briV; (h iin 2s where V =

    [Vi te

    if is a white noise of intensity

    V, and bp; are the coefficients depending on time has the form nm

    kyy = 2(a41k11 + ai2ki2) + > Vrrbirbin, ri h=1

    kia = aoiky + (411 + ao2)ki2 + ai2ke2 + a Vrproirban, ryh=l n

    ko = 2(d2ik12 + ao2ko2) + ‘se Urnbar ban. pha

    Obtain from these equations the corresponding equations for the system of Prob-

    lem 1.4 at X;

    =

    Vij, Xp

    =

    Vp».

    Find the stationary solutions in the case of

    constant intensity Y. Consider the process of settling of the stationary solution.

    6.25. Show that for the linear system with one degree of freedom of Problem 1.5 in the case where the generalized force with the spectral density Sz (w) =

    @ represents a stationary random process

    Da[1(a?

    + w?) the set of Eq. (6.1.8) for the

    components Y; = q, Y2 = p, Ys = Q/A of the state vector has the form

    kiy =2ky2,

    kip = ko3 — oki,

    Problems



    655

    -

    ;

    kia = keg — woki1 — 2eki2 + h13, ko = —2(woki2 + 2eko2 — kos), ;

    :

    :

    kez = —woki3 — (a + 2€)ko3 + k33, k33 = —2a(k33 — D). 6.26.

    Under the conditions of Problem 1.15 show that if a car is moving with

    the speed U and the covariance function of the road profile is exponential, =

    De-l1,

    then the state variables Y;

    = Y, Yo =

    Y, Y3 =

    ky(o )

    q (components of

    the expended state vector) satisfy Eq. (6.1.1)

    0

    a=

    | -—w2

    1

    0

    -2e

    w?

    0-

    —av

    0

    with the white noise of unit intensity.

    0

    |,

    a=0,

    b=]

    be b3

    Verify the validity of the following formulae

    for the variances and the covariances in the stationary regime: D kx

    =

    koe =

    wW6

    2 [deav +

    2Dav

    (14 $4)a)

    pt

    (< SF it) )

    2E/



    k33 == DD.

    kia. =),

    ko3 =

    kg

    avk13

    =

    =

    2D;

    avpD

    we + Bea aa we + 2eav + av? } © 6.27. Show that for the system with two degrees of freedom of Problem 1.12 at

    OSV

    = [Vi V2 ih where V is a white noise of the intensity

    / Eqs. (6.1.8) have

    the form

    ki = 2(Ajok13 — Ajok14), ki2 = —Ajgki3 + Ayyki4 + Agoko3 — Ajokoa,

    ki3 = —Cirkir — (Cio + Cha) ki2 — [B11 Ady — (Biz + By2 Aja) Jhis

    —[(Bi2 + Byy) Ay, — Bir Ajy )kia + Aggks3 — Ajoksa, ki4= —(Ci2 — Cio) k11 — Cooki2 — [(Bi2 — By2)Ag_ — Bo2 Aj )kis

    —[ Bog Ay, — (Biz + By) Aye |hia + Aggksa — Ajgkaa, kog = 2(—Ajgkos + Aji koa), ko3 = —Cirki2 — (C12 + C{2)k22 — [ Bir Ady — (Biz + Biz) Ajo |kos

    —[(Bi2 + Biz) Ay — Bir Aj |hoa — Ajokaa + Ajykaa, koa = —(Ciz — C}y)k12 — Co2ko2 — [(Bi2 — Bia) Az, — Baz Ajy kos

    656

    Chapter 6 Methods of Linear Stochastic Systems Theory

    —[Bo2Ay, — (Bia + Bia) Ajy koa — Atokaa + Aq haa,

    ka3 = —2Ci1k1a — 2(C12 + Cjp)koa — 2[ B11 Ag — (Biz + Biz) Ajy Jka

    —2[ (Bio + Byy)Ay, — Bi AjeJksa + 111, k34 = —Ciikia — (Cig — Cla)kis — Caokos — (C12 + Cia) ke

    —[Bi1A2, — Boo Ay, — 2Bi2Ajp Jksa — [(Bi2 + Bia) Aj, — Bir Ajo |haa —[ (Biz — By2)Ag, — Bar Ajy Jkas + “12, kag = —2(Ci2 — Cia) k14 — 2C22k24 — 2[ (Biz — Bi2)Az, — Bor Ajy Jka4 +2[ Boo Ay, — (Biz — Biz) Ajo )kaa + 22, where A;; being the elements of the inverse matrix A~ My 6.28.

    Write the equations for the variances and the covariances of the coordi-

    nates of the linear system with the dynamic damper of Problem 1.16. Show that at ba—n05 (ca =

    0) variance of the process X (t) in the regime of stationary random

    oscillations is determined by: D, = v[ prc + (Hat p)b2]/2c7bap2. 6.29. Using the equations of Problem 6.27 write the equations for the variances and the covariance of the processes Yj (t) and Y2(t) at Ay; = Agog = A; Aj2 = 0; Bi,

    =

    Boo

    =

    B; Bio =

    0; Bi

    =

    H; 41

    =

    ¥92

    = V; M12

    =

    0. Verify that

    the solutions of these equations with zero initial conditions are determined by the formulae:

    (a) At

    C1, = Cig = Coo = Clip = 0 and

    A= v/(H? + B?):

    kit hoo = Al, skys = bog = A/2, “Fig = ka, = 0 ki4

    6=

    (b) At Cy, = BH —AK:

    =

    —ko3

    =

    HA/2B,

    Cis =.Goy-

    =

    k33

    0, Cio =

    kii = keg = Bw/2K,

    = kag

    K(BH

    =

    V/2AB;

    > AK)

    ands

    v/6,

    33 = kag = KH/A,

    kyo = kig = koa = kag =0,

    keyg = —ko3 = K/2.

    Using formulae (5.1.15), (5.1.16), (5.1.24) calculate the derivative of

    H = (p" A~'p+q"Cq)/2. Show at the stationary random oscillation that the mean power FH damping (B’ fe 0) and position nonconservative (C - 0) forces.

    is wasted on

    Problems

    6.30.

    Show that Eqs.

    657

    (6.1.8) for stationaty linear system of Problem

    Q = IV where | is a constant m X n-matrix,

    1.11 at

    V is an n-dimensional white noise of

    intensity l/ represents the set of equations:

    Ky, = A7* Koi + Ky A“,

    Kg = A-'K22 — Kui(C — C') — Kin A71(B — B’), Rete Ae

    Rog

    (CaO iy

    (GG) Kia

    (Bt BA

    Kai( C=C) (84-8)

    Koy,

    Ar a=

    Ky A-'(B — B) +P; where Ky; and K92 are the covariance matrices of the vectors of canonical variables Yi =

    q, Y2 = Pp, and the blocks Ky

    of Y; and Y2.

    Cee =), RC. 6.31.

    and Ko

    Consider the special cases:

    Write Eqs.

    (a)

    (Cl 20, 8? =A,

    are the cross-covariance matrices

    C =

    Cen

    Boats

    As (b)

    (6.1.11) for the blocks of the covariance and the cross-

    covariance functions of the canonical variables of the system of Problem

    6.30 for

    the regime of stationary random oscillations. 6.32.

    Show

    that under the conditions of Problem

    6.19 the one-dimensional

    characteristic function of the process Y(t) = [Yi (t) Y> (t) iia is determined by the formula

    Gis

    2b)

    Gola,A } expyiagt(X" + X’'t/2)

    t

    nar.

    +/ Re Are 6.33.

    Show

    that under the conditions of Problem 6.20 for a stationary white

    noise with any function x( jt) the one-dimensional stationary distribution of the process Y9 (t) is detemined by the characteristic function:

    Z:

    gi(A) = exp

    »

    DP

    xa

    x | ul dy. 0

    6.34.

    Under

    the conditions

    of Problem

    6.25 write the formula for the one-

    dimensional characteristic function of the strictly stationary process Yj (t) in the

    case of an arbitrary function x(H). Consider also the case of x(H) = =i, vp/2.

    658

    Chapter 6 Methods of Linear Stochastic Systems Theory

    6.35. The three-dimensional Brownian motion of a particle under the action of the disturbing force representing the stationary white noise is described by the vector

    Langevin equation

    Y +2cY = y,,

    (I)

    Y being the state vector of the particle, V is the stationary white noise with the

    known function x( HL). Show that the one-dimensional characteristic function of the velocity vector

    U = Y of the particle in the stationary regime is determined by the

    formula

    Re) ee

    i!xe

    "dr } SMD).

    0

    Consider the case of the normally distributed white noise. 6.36. Show that the multi-dimensional distributions of the random process Y (t) determined by the differential equation

    Y =ao+aY

    + bX (t),

    (1)

    where X (t) is the random function representing the solution of the stochastic differential equation

    X=aot+aX

    + BV

    (II)

    with independent initial values Yo and XQ are determined by the formula: n Gri

    seoeins ticle

    00

    n

    S— uri(te, to)” A

    9

    So ura(te, to)? A

    peal n

    Xx €xXp

    tk

    iso k=)

    n

    == k=1

    it

    [ent

    tk

    rao(r)ar

    to

    te

    /x

    kod

    +

    fra(te,r)a0(r)ar to

    n

    BU

    Ys tod)it

    |r

    a

    Ag)

    l=k

    Ii where % (p), 9o(c) are the characteristic functions of the initial values Yo, - 2

    the process Y (t), X (t); U11 (t, T) is the fundamental matrix of solutions of the homogeneous equation duj,/dt = auj4, ui1(T, T) =F

    ur2(t, T) is the solution

    of the equation duj2/dt = auU12+ bu22 with zero initial condition, u12(T, T) — J)

    _Problems

    659

    u2a(t, T) is the fundamental matrix of solutions of the equation duy2/dt = U9} u22(T, T) — rf (Ig is the unit matrix of dimension q). 6.37. Show that the stationary variance Dy of the output of the system consist-

    ing of the impulse device which generates the rectangular impulses of the duration Ty (with the period repetition 7;) and the linear link with the transfer function ®(s) under the action of the impulse white noise with the variance D) of each impulse are

    determined by the formulae:

    1) Dy = kD/(k + 2) at ®(s) = k/s,

    yw ee EDpl21) at ®(s) = k/(Ts + 1). Here p(z1) = 2?(z)’ — 1)?/(1 — z?), 21 = e-7*/7, y = Ta/Th. 6.38.

    Find the expressions for the spectral densities and the variances of the

    output for typical linear systems with distributed parameters (Table A.5.4).

    6.39. Consider Example 6.3.1 for the case when 7' = 7}, mz (t) =a-+t

    Kotter Solution.

    bto,

    eet. According to the first method if we use the result of Example 3.9.12

    at y(t) = VDe®*, choosing the coordinate functions in the form

    zn(t) = —

    sin E (:-) - | (ene

    Oi)

    we obtain respectively t

    my = af VP (a4 balan = a+ Ht 7) —(a- We!" () 1



    VolEh= :,Method 8aiD Aural pata le sin(wnt + yn)—e-/T sin Pn| :

    (T+ An)

    J+ BT)? + w2T?

    n [oP sin(wnt +n) —e/Tsingn] (Tr an)[ tet tort]

    Ne Dy _

    nt

    tT

    Bp Nl

    ae

    aoe

    (IIT)



    CIE

    Saye

    Here the exact solution of te variance DS (t) is determined by formula t

    t

    Dy*(t) = - /fire

    rae

    yy

    660

    Chapter 6 Methods of Linear Stochastic Systems Theory

    D (2A a4 ex Plt)

    2D (eft _ 1)

    (1+. 67) [P(@nep)r]

    pen? 4 eam

    (V)

    (2: = 8-—a+1/T). While using the second method the following equations are solved

    Tring+ my

    b1, = 0-4

    Tin

    + Un

    =

    2MEO De sin

    T

    [wn (¢- 7) +=],

    (v1) their integrals which vanish at t = 0) are expressed by formulae (I) and (III).

    CHAPTER 7 METHODS OF NONLINEAR STOCHASTIC SYSTEMS THEORY AND THEIR APPLICATIONS

    The last Chapter is dedicated to the basic methods of the nonlinear stochastic systems described by the stochastic integral, differential, integrodifferential, difference and other nonlinear operator equations. Section 7.1 contains the derivation of the formulae for the derivatives with respect to the moments of different orders. Special attention is paid to

    the derivation of the infinite system of the equations for the moments. The equations for the moments in the stochastic linear systems with the parametric noise are composed. In Section 7.2 the methods of the normal approximation and the statistical linearization for determining the oneand the multi-dimensional distributions are stated. The methods of the equivalent linearization and the general methods of the normalization which form the basis of the structural methods of nonlinear stochastic systems theory have a significant place in Section 7.2. Section 7.3 contains the brief account of the methods of general nonlinear stochastic systems and the statement of the general problem of the parametrization of the one- and the multi-dimensional distributions. Methods of the reduction of the number of the equations for the distributions parameters are considered. The moments methods, the semiinvariants methods, and the moment-semiinvariants methods for defining the one- and the multi-dimensional distributions are stated in Section 7.4. Section 7.5 is dedicated to the methods of the orthogonal expansions and the quasimoments. Method of structural parametrization based on the ellipsoidal approximation of the one- and the multi-dimensional distributions is given in Section 7.6. Section 7.7 is devoted to the methods based on the canonical expansions and the integral canonical representations. Sections 7.8 and 7.9 contain the applications of the nonlinear stochastic systems theory methods to the analysis, modeling and conditionally optimal filtering.

    7.1. Moments

    of State Vector

    7.1.1. Formula for Derivative of Expectation > The simplest way to find the formula for the time derivative of the expectation of the state vector of a system is to pass to the expectations directly in Eq. (5.4.6)

    for the normal white noise with the matrix intensity

    Vy=

    v(t). Then taking into

    662

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    consideration that in the It6 equation the value of the process Y(t) at any ft is independent

    of the value of the white noise V at the same instant we obtain the

    following formula for the time derivative of the expectation m(t) = EY (t) of the

    process Y(t): i=

    Remark.

    Ea(Y,t) sii,

    Cisled)

    This formula is not a closed equation which

    determines

    the

    expectation of the process Yi(G)y as the expectation in the right-hand side depends on the unknown one-dimensional distribution of the process AGE).

    7.1.2. Formula for Derivative of Second Order Moment > Let us denote by Yxj = EY; (EN (t) the components of the matrix of mo-

    ments of the second order [ = EY(t)Y(t).

    Using formula (5.1.7), we evaluate a

    stochastic differential of the product Y;z (t)Y; (t). As a result we obtain the following

    expression for a stochastic differential of the product:

    dY,(t)Y; (t) = \%az(Y, t) + Y; a; ‘ee t) + 533 De (=)

    os),

    Dre. w}dt

    Sb

    m

    + S—[Yeber(¥,t) + ¥jbjr(¥, t)]dW,.

    (7.1.2)

    (Pal

    While evaluating the expectation of both parts of (7.1.2) and accounting that [ =

    lye; | (kj 3 leateg Pp) we come to the sought formula for the derivative of the

    second order moment

    De By of, 4)

    Vue Viet)! bie

    2 HY +

    (7.1.3)

    Here

    o (Yt Remark.

    Formulae

    )b=.b(¥id)

    ple )o04)7 ab

    (7.1.4)

    (7.1.1) and (7.1.3) are not closed equations for m

    and I’ in the general case, as their right-hand sides depend on the one-dimensional distribution of the process Ya) i.e. not only on m and I’. And only in some special

    cases the right-hand sides of formulae (7.1.1) and (7.1.3) may be the functions of m

    and I’ independent

    of other characteristics of the one-dimensional

    distribution

    of the process NAGAY In such cases (7.1.1) and (7.1.3) will be ordinary differential equations which determine the moments

    m and I’. So, for instance, in the case of

    7.1. Moments of State Vector

    663

    linear functions @ and b formulae (7.1.1) and (7.1.3) represent the equations which determine the expectation ™ and the second order moment separately.

    I" of the process Y(t)

    These equations certainly coincide with equations

    of Subsection

    6.1.1

    obtained for the more general case.

    7.1.3. Formula for Derivative of Covariance Matriz The formula for the time derivative of the covariance matrix K of

    the vector Y is easily derived from formulae (7.1.1) and (7.1.3). > In order to obtain this formula it is sufficient to differentiate the relation KC =T—mm!

    with respect to t and substitute into the obtained formula the expres-

    sions of m and I’ from (7.1.1) and (7.1.3). As a result we get

    K = E {a(Y,t)(Y? — m7) + (Y — m)a(¥, 2)" + o(¥,t)}

    a

    (7.15)

    So the results obtained in Subsections 7.1.1 and 7.1.2 may be summarized in the following statement.

    Theorem 7.1.1. [f the random process Y(t) defined by Eq. (5.4.6) with the normal white noise V = V(t) has the finite expectation m, the second order moment T and the covariance matriz K then formulae

    FAECES)

    a(t

    Srareeaids

    For formulae for the derivatives of expectation and second order mo-

    ments in the case of Eqs. (5.4.7) and (5.4.6) for V being the nongaussian white noise see Problems 7.1 and 7.2. Example

    7.1.1. For the system described by the equation

    Y= Ye Ve

    (I)

    formulae (7.1.1), (7.1.3) and (7.1.5) have the form

    m= l| —EY?, a2 = —-2EY* + vaz,

    :

    D = —2EY* + MEY? + v(m? + D). In this case the process Y(t) is scalar, and consequently, I‘ and tively the initial second order moment

    (II) K represent respec-

    a(t) and the variance D(t) of the value

    of the process Y(t) at the instant t. The right-hand sides of formulae (II) depend on the third and the fourth order moments

    of the process Y (t), and consequently,

    formulae (II) are not the equations determining ™,

    2

    and D.

    The attempts to

    664

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    close these equations by adding the equations for the third and the fourth order moments will lead to the appearance of the moments of the fifth and sixth orders, and

    so on.

    Consequently, such attempts will not lead to a closed set of finite number of

    equations.

    Example

    7.1.2, For the system

    Y =aY +ap+bVYV

    (I)

    formulae (7.1.1), (7.1.3) and (7.1.6) give m=am-+ao,

    &2 =2(aa2z+aom)+b?vym,

    D=2aD+b?vm.

    These ordinary differential equations with given initial conditions:

    (II)

    m(to) =

    Ne

    a2(to) = 20, D(to) = Do = a0 — m? determine consecutively the moments mM, A

    and D.

    7.1.4. Formulae for Derivatives of Second Order Moment ance Function

    and Covari-

    Later on we shall also need the formulae for the partial derivatives of

    the second order moment I'(t1,t2) and the covariance function K (ty, t2) of the process Y(t) defined by Eqs. (5.4.6) with respect to the second argument fg at t1 < to. > Using

    d:,V(t1,t2) = EY (t1)dY(t2)’, d:,K(t1,t2) = E[Y(t1) — m(t1)]dY (tz)?

    (7.1.6)

    and using Eqs. (5.4.6) we shall get at t} < tg the following formulae:

    OV (ti, t2)/Ot2 = EY;,a(Y;i,,t2)",

    (FAM

    OK (ti, te)/Ote = E(V:, — m:, )a(%,,t2)? . «

    (7.1.8)

    Theorem 7.1.2. If the random process Y(t) defined by Eq. (5.4.6) with the normal white noise V = V(t) has the finite first and the second order moments for instants t; and tz then the matriz of the initial

    moments and the covariance matrix satisfy formulae (7.1.7) and (7.1.8). Example

    7.1.3.

    For the system of Example 7.1.1 formulae (7.1.7) and

    (7.1.8) have the form

    AV (ti, t2)/Ot2 = -EY,,Y2, OK (t1,t2)/Ot2 = —E(¥:, — m:,)¥3.

    7.1. Moments of State Vector

    Example

    7.1.4,

    665

    For the system of Example 7.1.2 formulae (7.1.7) and

    (7.1.8) have the form OT (t1, t2)/Ot2

    =

    a(t2)T'(t1, te) ,

    OK (ti, t2)/Ote

    =

    a(t.) K (ti, te) ;

    (1)

    These formulae represent the ordinary differential equations at any ;. Together with

    the initial conditions:

    P(t, . t) = ao(ty ); K(ty ; t1) = D(t1) they determine the

    second order moment P(t, ; to) and the covariance function K(ty ; to) of the process Y(t) atrbje —nn h=1

    b=k

    Opening in Eq. (7.1.16) the brackets we get

    b Sule

    Rad

    agmiie mat + bovb2

    (7.1.16)

    668

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory P

    P

    h=1

    Ayla

    +S \(b,vb) + bovb, )EYn + D> bavb7 EYaYi = al +Ta™ + apm? + ma? + bovbd

    +S

    Pp

    P

    2(bavbg + bovb;)mna + > bavbt yn.

    sal

    (110)

    hil=ti

    Eq. (7.1.17) does not contain any characteristics of the random vector Y except in its

    expectation and second moments (the elements of the matrix I’). Consequently, after

    integrating Eq.

    (7.1.15) which determines the expectation ™ of the vector Y, Eq. (7.1.17) with the initial condition I'(to) = Tp (and consequently, Yai (to) = 7)

    completely determines the second order moment

    I(t) of the vector Y¥():

    In exactly the same way we reduce Eq. (7.1.5) to the form

    P K =aK + Kal? + bovbt + So (bnvbg + bovb? mp

    as P + $5 bavb? (mam + kn), K = [kn],

    (7.1.18)

    pal

    where

    kp] is the covariance

    of the components

    Y;, and Y) of the vector

    Y

    (A, l

    =1,..., p). Eq. (7.1.18) with the initial conditions K(to) = Ko (kin = kp,) completely determines the covariance matrix K(t) of the vector Y(t) at any time

    instant ¢ after finding its expectation Mm. < Thus if we confine ourselves to the moments

    of the first and the second

    order

    of the state vector and of the output while studying a linear system with parametric

    white noises (7.1.14) then these moments

    may

    be exactly determined

    by subsequent

    wntegration of Eqs. (7.1.15). (7.1.17) or Eqs. (7.1.15), (7.1.18) as well as in the case of a hnear system.

    For a linear system with parametric white noises (7.1.14) formula (7.1.8) gives the following equation for the covariance function of the process YA(E) at to > ti:

    OK (ti, te) /Ote =

    K (t1, t2)a(t2)? ‘

    (7.1.19)

    The initial condition for this equation has the form K(ty ; ty) = K(ty ,: At last it is easy to conclude that the moments (7.1.14)

    are determined

    of the state vector

    by the infinite set of Eqs.

    (7.1.13) which

    Y of a system in this case

    is

    7.1. Momefits of State Vector

    669

    decomposed into independent sets of equations for the moments

    (Karsoitese, pr ss Oatley Sct. P

    P

    a, = Sk

    Gr o&k-e, + ete

    Ga! t

    ines

    gai P

    P

    4 9 Dal Es er 1) (ornote-e, tt; aia pel

    p pal

    >4

    Fonceteadthtegtes-2e

    q Uu= 1

    5

    a

    Pp

    =

    Ak+e,—e,

    q=1

    P a

    of each given order k

    al| = kychine thy = 1y2,....):

    Daya

    n=

    kinks (crsocnevne

    asi

    Pp

    Ors,egtk+eg—er—es q=1

    Fir ») cra pielcnriyiA—t,—«,

    :

    qual

    (7.1.20)

    Here as well as in Eqs. (7.1.13) Qs is equal to zero if at least one of the components of the multi-index S is negative, and is equal to unity if all the components of the multi-index S$ are equal to zero.

    Example

    7.1.7. In the case of scalar Y and V Eq. (7.1.14) has the form Y=aY+aot(bo

    +h

    Y)V.

    (1)

    Eq. (7.1.15) has in this case the same form as in the general case of vector Y,V.

    (7.1.17), (7.1.18) for the second initial moment @2 = T and the variance the process

    Eqs.

    D = K of

    Y take the form

    Gg = (2a + 6?)az + 2(ao + bobiv)m + bev, ;

    D = (2a — b2v)D + 2bobiym + bivm? 4 bev.

    II

    (i)

    In the case of a normally distributed white noise V we may also obtain the exact equations for the moments

    of higher orders. These equations have the form

    1

    api—-ki|.are atk —1)b?v |a, +k [a0 + (k — 1)bobiv Jagi

    i + 5 h(k Sl b2vag > “(k= ap4apey)s

    (IIT)

    Let us consider the stationary linear system (7.1.14) with stationary parametric white noises.

    In this case @, dg, bo, bp, and vy are constant.

    Therefore, putting in

    670

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Eqs.

    (7.1.15), (7.1.18) m =

    (0 and K

    = 0 we obtain algebraic equations for the

    expectation and the elements of the covariance matrix of the value of the stationary

    process in the system at any instant tf:

    am+aj=0,

    (Aa 21)

    Pp

    Pp

    b=

    i

    aK +Ka™ +bovbh +> (brvb5 +bovb; mnt Y_ bavbs (mami+kni) = 0. (hak 22) For finding the covariance function of the stationary process bz ), T=1t,-To, we write the general equation for the covariance function

    Hence,

    using

    OK (ty, t2)/Ot2 = K(ti,t2)a? at te

    >t).

    the

    function

    property

    of

    the

    covariance

    (7.1.23) K (ty p to)

    = K (te, t,)? and transposing Eq. (7.1.23) we shall have OK (ta, t1)/Ote = ak (to, t1) ateoy Q. At

    T < 0 covariance function is determined

    by

    k(r) = k(—r)?.

    We may also determine the moments

    of higher orders of the one-dimensional

    distribution of the stationary process in the system. For this purpose one should put

    a, = 0 (ki, ee Remark.

    al aa bs epay et |k| = hy Pies

    By = Be nee

    (7.1.26)

    It should be noticed that the wide sense stationary process in

    a linear system with parametric white noises may physically exist only in the case

    where the systems described by deterministic Eq. (7.1.15), (7.1.18) are stable. But this condition is generally not sufficient for the existence of the strictly stationary process in the system.

    7.1. Moments of State Vector

    671

    7.1.7. Formulae for Moments in Stochastic Discrete Nonlinear Systems The formulae for the first and the second order moments in conformity to dis-

    crete system (5.7.1) are obtained on the basis of Eq. (5.7.7) and have the following form

    mis, = Eun (Yi, Vi), (71220) Key = E{wi(¥i, Vi) — migi] [wr(¥i, Vi)" — mf, ] . (7.1.28) Analogously on the basis of Eqs. (5.7.14), (5.7.15) and (5.7.17) the formulae for the moments

    are derived also for the case of continuously-discrete systems (5.7.10) and

    (5.7.11).

    For the nonlinear systems formulae (7.1.27) and (7.1.28) are not closed

    equations.

    7.1.8. Equations for Moments of Discrete Linear System with Parametric Noises As an example of direct derivation of the equations of Subsection 7.1.7 for the first and the second order moments let us consider the following discrete linear system with the parametric noises: P

    Yiga = a Y) + ao7+

    bor +

    ) bi; Yi;

    V,,

    (han 29)

    j=l where bo), bi, --.

    , byt are the matrices of the same dimension as the matrix do]

    ands Yyqyisey Yip are the components of the vector Y7.

    > After taking the expectations of both parts of Eq. the independence of Y; and Vj and also the equality

    (7.1.29) and accounting

    EV; = 0 on the basis of the

    theorem of the expectations multiplication for the independent variables we get

    Mmi41 = aym; + aor.

    (7.1.30)

    Similarly as for an ordinary linear system while deriving the equations for the covariance matrix K; =

    Eq.

    E(Y; = mi)(Yi = m)

    (7.1.29) term-wise Eq.

    of the process {Y;} we subtract from

    (7.1.30) and multiply the result by the correspondent

    equality for the transpose matrix.

    After taking the expectation of both parts of the

    obtained equality we come to the equation

    EV

    aa mug1)(Yit1 a Mi41) = a E(Y; — mi)(¥;" a mj )a;

    672

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory P

    P

    gal

    Gia

    (7.1.31)

    +E | bor +S ds1¥jr |ViVi" |bor + D0 ORY | -

    But by virtue of the independence of Y; and Vj and according to formula of the total expectation we have

    P EB bor + >> br Yiy viv,

    p bor + > bY iy

    Gia

    jal

    Pp

    = E {bn + >> by %y | BLVV

    P

    .

    [Yi] |b+ Do onY,

    ah

    j7=1 P

    P

    Yo

    j=1

    =E |bo + >5by ¥y |B[VVT] | bo + D0 ORY Pp

    P

    j=l

    Gil

    = E | bor + 5) by ¥ig |Gr |OG + Don P = boiGib2; + ‘Ss (bo1G1b}, + bj1Gibby) mu; j=l

    p.P + SS biG (rag min + keayn)

    (7.1.32)

    jJ=1 h=1

    where ™j; are the components of the vector ™; ki;h are the elements of the matrix kj; G} is the covariance matrix of the variables V). Substituting expression (7.1.32)

    into Eq. (7.1.31) we get the sought difference equation for the matrix K7:

    P Kiga = ay Kyat + borGibhy + Y>(borGibh,

    j=l

    P

    +bj1G1bg)) mj + YS bj1Grbhi (maj min + kin). 4 =

    h=1

    (7.1133)

    7.2. Methods

    of Normal

    Approximation

    and Statistical Linearization

    Thus Kq. (7.1.30) with the initial condition my

    the expectation {mi} Ly

    =

    E(Y,

    =

    of the process {Yi}, and Eq.

    mi)(Y4

    =

    m,)?

    determines

    = HY,

    673

    determines

    completely

    (7.1.32) with the initial value

    the covariance

    matrix

    {Ki}

    of the

    process {Yi}. For finding the covariance function Kj, h) of the process {Yi} we

    by means

    of Eqs.

    (7.1.30) and (7.1.33) E(Y; — ms )(Yru _ a)

    Accounting the independence we

    evaluate

    ae

    Oe

    of Yp, Y; and V; atj < hand the inequality EV, = 0

    obtain

    K(j,h+1)=K(j,h)az . Eq.

    (7.1.34) with the initial condition K(j, J) =

    (7.1.34) K; determines

    completely

    K(j,h) ath >j. Ath Approximating the one-dimensional distribution of the random process YZ)

    in normal stochastic differential system (5.4.6) by a normal one we shall have

    gi(A;t) & exp {iat

    it

    37 Ka};

    CS:

    fily;t) © [(2m)? [K|]7"/? exp {-30" —m™)K-(y - m)}_€7.2.2) where ™

    and

    K

    are the unknown

    vector of the system Y.

    expectation and covariance matrix of the state

    674

    Chapter 7 Methods

    of Nonlinear Stochastic Systems Theory

    After calculating the expectations in (7.1.1) and (7.1.5) for the normal distribution V (m, K ) we obtain the ordinary differential equations approximately determining ™ and K:

    m= i(m, K,t),

    m(to) = mo,

    (7.2.3)

    K = 92(m,K,t),

    K(to) = Ko,

    (7.2.4)

    where

    gi(m, K,t) = Ena(Y,t), p2(m,

    K a) =

    poi(m,

    (7.2.5)

    K,t) + poi(m,

    Katy.

    + ~22(m,

    yoi(m, K,t) = Ena(¥,t)(Y? —m’), o(Y,t) = W(Y,t)v(t)b(Y, t)”, 2o(m, K,t) = Ewo(Y,t) = Enb(Y, t)v(t)0(Y, t)?

    Ket)

    £026)

    (7.2.7) (7.2.8) (7.2.9)

    and the subscript V denotes that the expectation is calculated for the normal distri-

    bution N(m, K) of the random variable Y :

    Ex()= ae f ew [507 mi")Kym] The number of equations for m

    (NAM) is equal to QNam

    dy. (7.2.10

    and K by the normal approximation method

    = P(p + 3)/ 2. 4

    Thus, the NAM for the one-dimensional distributions in the normal

    stochastic differential system is based on Eqs. (7.2.3), (7.2.4) at (7.2.5)(7.2.10) and corresponding initial conditions. 7.2.2. Statistical Linearization Method

    In the special case of the unit matrix b(y,t), b(y,t) = I, g22(m, K,t) = v(t). In this case NAM gives the same equations as the Kazakov statistical linearization method (SLM). > Really, SLM is based on the approximate formula

    O(Y,t) ¥ yo +ki(Y —m), where (9 and ky are determined by minimizing the mean assumption

    of normality

    of the distribution of Y.

    (7.2.11) square error under the

    Here Yo and ky are given by

    7.2. Methods of Normal Approximation

    the formulae

    ~9

    =

    and Statistical Linearization

    yi(m, fagteaA and ae) =

    ki(m, IK, t) =

    675

    poi(m, Nyt). Hex,

    Replacing the function a(Y, t) by the obtained linear function of Y we reduce Eq.

    (5.4.6) in the case of b(Y, t) = I toa

    linear stochastic differential equation

    Y =¢o4+hi(Y —m)+V,

    (7,212)

    Using methods of Section 6.1 we get in this case the following approximate equations for mand

    K:

    m= 0,

    (2213)

    K=khK+Kk? +v.

    (7.2.14)

    Substituting here the obtained expressions of (9 and ky, we see that these equations

    coincide with Eqs. (7.2.3), (7.2.4) at 22(m, iG t) = VV @

    For SLM practical application the tables of formulae were compiled

    for Yo and k; = [0/dm)yo iF for typical scalar and vector composite functions. In Appendix 3 the formulae for go are given for some typical composite functions. These formulae may be used for the determination

    of the function goi(m, K,t) = kK = [(d/dm) yf ]* K, while setting up Eqs. (7.2.3), (7.2.4) of the normal approximation method. We may also use these formulae for determining the function ~22(m, K,t), as

    according to (7.2.9) it represents the first item in the formula of the form of Eq. (7.2.11) for the statistically linearized function b(Y, t)v(t)0(Y, t)?. For NAM equations in case of Eqs. (5.4.7) and (5.4.6) at the nongaussian white noise see Problems 7.9 and 7.10. Example

    7.2.1. Let us consider the system described by the stochastic

    differential equation of Example 7.1.1. In this case a(y, t) = — 43> b(y, t) == i) axel

    formulae (7.2.5), (7.2.7), (7.2.9) give

    ew ,D,t)=i(m,D,t)=-— | y Pe 1 yoi(m, D,t)

    es

    ,_D,t) =

    ;

    if ify(y —

    27D

    V

    y

    Wom) 2D ty.

    ()I

    prea: m)e

    (y

    2

    fl eee= 1 2)

    ) 12D dy,

    dy,

    (II)

    lll

    676

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Taking into account

    y? =m? + Im(y— m) + (y—m)’*, y? = m3 + 3m?(y — m) + 3m(y — m)? + (y—m)’, that for the normal distribution the central moments to zero, and the central fourth order moment

    (1)-(1V)

    (IV)

    of the odd orders are equal

    is equal to 3D? we obtain from Eqs.

    |

    ~1(m, D,t) = —m(m? + 3D), yai(m, D,t) = —3D(m? + D), 22(m,

    D,t) =

    v(m?

    + D),

    .

    yo(m, D,t) = —6D(m? + D) + v(m? + D) = (v —6D)(m? + D). Consequently, Eqs. (7.2.3), (7.2.4) have in this case the form

    tn = —m(m? +3D), D = (v—6D)(m? + D).

    (v)

    After integrating Eqs. (V) with the initial conditions m(to) = hii D(to) = D5 we determine

    completely the approximate normal one-dimensional

    the process Y(t).

    The obtained equations may

    equations of Example

    distribution

    of

    also be derived from the first two

    7.1.5 by substituting into them the expressions of the third

    and the fourth moments in terms of the expectation and the variance for the normal distribution. Example

    7.2.2. For the system described by the equation

    where p(y) is the nonlinear characteristic of the limiter shown in Fig.7.2.1a formulae (7.2.5), (7.2.7), (7.2.9) give

    Ys

    @(y)

    b)

    Pig iecel!

    0 (=I sgny

    7.2. Methods of Normal Approximation

    1

    yi(m, D,t) =—

    V 2nD 20

    and Statistical Linearization

    —l

    -!i e~(y-m)

    l

    : Pays

    : 12D

    fyom)

    J

    Hf eeminras]

    677

    dy

    J

    ; (p22(m, D,t) =

    vy

    (II)

    1 -

    gar(m,

    '

    peat

    =

    D,t)= —e| tf(y mye _

    —l

    a

    I

    —(y-m)?/2D

    4

    ,

    fee)

    + [uly = myer

    Pay 40 f(y= men

    a4

    ?P dy (i

    i

    For performing the integration we notice that the integrals of a normal density representing the probabilities of the occurrence of a normally distributed random vari-

    able in the corresponding intervals are expressed in terms of the Laplace function D(z) = —1/¥V 20 gien" /2dy by the formula 0

    The other integrals are reduced to the integrals p

    B

    p

    pte’ Pau i

    /ude"

    /? =

    (seat? _ ae~*"/?)

    p

    + fwd a

    =

    (Be-#°/? - ae~*"/?) + V2 [6(8) — B(a)}.

    (VI)

    678

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Using formulae (IV)-(VI) and taking into account that @(—z)

    =

    —(2), (00)

    = 1/2 we find

    yi(m, D,t) = — [d+ mye SS a

    a

    mina -ole(3B)+0( 8) P

    Egs. (7.2.3), (7.2.4) have in this case the form

    l+m

    en

    6

    ——

    } —(l-—m)®

    L+ =)

    (‘- =)

    l—-m\

    | ——

    vf fon[- £58] ae [-R")}, on [c+me (SE)

    D=—2D.|®\

    ——

    | (VD

    ie,

    | +2 |——

    VD

    (ss)

    ] | +.

    X

    cS

    Analogously in the case when p(y) = [sgn y (Fig.7.2.1b), we obtain

    WES

    lem ato okie pr(m,D), v1 Duala =216(7)

    XI (X1)

    m

    D = —2Dk,(m,D) + v, ky(m, D) = Sissre-. (22 Kom .

    2

    2

    ORETY)

    us

    Example

    Eq.

    7.2.3. For the system with the two-dimensional state-space by

    (I) of Example 7.1.6 the matrices a(Y, t) and b(Y, t) are determined by the

    formulae

    a Vite sods 0 a(¥,t)=-| ; A ; wv.= [2],

    (1)

    and formulae (7.2.5), (7.2.7), (7.2.9) give

    bG0 A Org ees ki2 palm, K,t) == ~By | : a =olee pas(m,

    gaa(m,

    ¢

    Y,

    Y;

    K,t) = ~Ex |is y [sm

    K,t)

    (II)

    Yo — mz |

    tn |Kaye Vay, eae CC)Aum) 2) Co 0 v0 h}= i Onera = By |)|

    ae (IV)

    Taking into account that for the normal distribution all the central moments of the

    odd orders are equal to zero we find from (I)-(IV)

    7.2. Methods of Normal Approximation

    and Statistical Linearization

    679

    EnYi¥2Y, = Ey {mym2YP + m7 YP + moY,? + YP} = mki2+ mek,

    (V)

    EnY1Y2Yy = miko. + m2ki2, EnY2Y? = ki2, EnYoY¥2 = kao, eR

    Kye

    | moki, + myki2

    meky2+ miko |

    aky2

    p2(m, tole t) —

    ed | —2(m2ki13 + myky2) = a)kyo—

    (VI)

    ako

    yoi(m, 1G t) = yai(m, K, te a2 22(m,

    —(m2

    (VI)

    ist t)

    ~=—(me2 + a)ki2 — mi ko2 |

    my koo

    vh2 —

    2ako2

    ; Thus Eqs. (7.2.3), (7.2.4) using (V)—(VIII) have in this case the form

    (VIII)

    my, = —m,mMz — ky2, m2 = —amz, ky, = —2(moki1 + mki2), ko2 =

    vh? =

    (IX)

    kig = —(m2 + a)ki2 — my k22,

    2ako.

    (X)

    0,483

    0,108

    0,479

    0,104

    0,477

    0,102

    0,475

    0,100

    0,473

    0,098

    0,471

    0,096

    0,469 0,467

    0,094 0,092

    0,481

    0,106

    0,465

    0,090

    0,1

    0,3

    0,5

    0,7

    0,9

    Fig. 7.2.2

    Pig. 7.2.3

    The results of calculations of the solution curves Eqs. in Fig.7.2.2 and 7.2.3 at my,(0) =

    m2(0)

    =A):

    k11(0)

    (IX), (X) are shown se

    of) be Kio s=

    (0);

    ko2(0) = 1. The exact distribution curve is given for comparison. So the expectation coincides with the exact one. The accuracy of NAM

    3%.

    for 7 = VD, D = kj; is about

    680

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    7.2.38. Normal Approximation Method. In exactly the same

    Multi-Dimensional Distributions

    way the normal approximations of all other

    multi-dimensional distributions of the process Y(t) may be found.

    As

    the normal distribution is completely determined by the moments of the first and the second orders, and the latter are determined by the twodimensional distribution, it is sufficient to find the normal approximation of the two-dimensional characteristic function g2(A1, A2;t1,t2). For this purpose it is sufficient to find the covariance function K(t,,t2) of the

    process Y(t). > In order to obtain an equation for K(t, ; to) in the case of system (5.4.6) it is sufficient to evaluate approximately the expectation in (7.1.8) replacing the unknown two-dimensional

    distribution of the process Y (t) by a normal one.

    As a result we

    obtain

    OK (ty, t2)/Ot2 = En [Y:, — m(ti)] a(¥,, ta)?

    (7.2.15)

    where the subscript JV at the expectation means that it is calculated for a normal joint distribution of the random variables Yi, and Yeas For calculating this expectation we

    use the formula of total expectation:

    first, we evaluate the conditional expectation

    relative to em and after that we evaluate the expectation of the obtained function of

    the random variable Ve . Then using the known formula for the conditional expectation of the part of the components of a normally distributed random vector relative

    to the remaining components we get En|Y1, = m(ty)]a(Yi,, te)” > En{En|[Yi, lY2.] rz m(t2)}a(Yr,, to)”

    = K(t;,t2)K(t2)~'En[Yi,—m(t2)Ja(¥i,,t2)” -

    (7.2.16)

    But on the basis of (7.2.7)

    En[Y¥i, — m(t2)]a(Y%r,,t2)” = Goi(m(t2), K(t2), ta)? .

    (7.2.17)

    Consequently, from (7.2.10) and (7.2.17) we get

    En[¥:, — m(t1 ))a(¥i,,t2)” = K (tr, t2)K7"(t2)eo1(m(t2), K (te), ta)” -

    (7.2.18)

    Substituting expression (7.2.18) into (7.2.15) we obtain the approximate equation for the covariance function of the process Y (t):

    OK (ti, t2)/Ot2 = K(t1,t2)K(t2)7 '~oi(m(t2), K(t2), te)”.

    (7.2.19)

    This equation with the initial condition K(ty ; ty) = K(t1) represents at any fixed t; an ordinary linear differential equation determining

    of t2,t1 < tg. 4

    (tsts tz) as the function

    7.2. Methods of Normal Approximation and Statistical Linearization

    681

    Thus Eqs. (7.2.3), (7.2.4), (7.2.19) define consecutively the expectation m(t), the covariance matric K(t,t) = K(t) and the covariance function K(t,,t2) of the process Y(t) according to NAM in the normal stochastic differential system described by. After this we may determine all the multi-dimensional distributions (approximately normal) of the process Y(t) by formulae of (Subsection 2.7.5):

    ieee’

    0.6;

    las

    | An)

    =

    exp

    {XPin

    a

    5X" RX}

    (n =

    oe

    « ee

    (7.2.20) m|

    N= [ATAZ An] z =|

    Tp

    ¥

    > Hn = [my (ta)? my (to)? ...My(ta)7T?

    Kat AE (itats

    Kati)

    K(ta,ta)

    ot

    ...

    C14)

    K(tayta)

    (host )lk (it) bod cacti

    (in be)

    fn(Yt yi m1 Ynj try +-+ tn) rs

    1

    e

    rw [(27)"|K,]71/? exp {-5(a — m2) Ko (Gn — itn)} (p= (722-21) where jn = [yl yZ ...yZ]’. Example the form

    7.2.4. Under the conditions of Example 7.2.1 Eq. (7.2.19) has

    OK (ti, t2)/Ot2 = —3 |m?(t2) + D(t2) ]K(t1, ta).

    The solution of this equation with the initial condition K(ty ; ty) =

    (1) D(t1) is ex-

    pressed by the formula

    K(t,tz) = D(t1) exp

    =3 f{m?(r) + D(x)} dr

    (II)

    According to this formula we may calculate the covariance function of the process Y (t) after the integration of the equations determining its expectation m(t) and

    variance Die). Example

    7.2.5. Under the conditions of Example 7.2.2 (Fig.7.2.1a) Eq.

    (7.2.19) has the form OK (ti, ta) the [+ m(to) — m(t2) oR tte) c(Se + (SEE |Kania

    iien'(D

    682

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    The solution of this equation with the initial condition K(ty , to) =

    D(t1) is ex-

    pressed by the formula

    K (ti, t2) = D(ti) exp

    -

    [fs(Sz) +0

    (=z)

    :

    dt

    (ul)

    For the special case (Fig.7.2.1b) in formulae (XI) and (XII) of Example 7.2.1 we need to put

    ki(m(t), D(t)) =1 Example

    7.2.6.

    m(t)?

    aby

    To obtain by NAM

    -5o |Ae)

    = hh

    the equation for the covariance

    function

    rar iyi of the process ato) =

    Kii(t1,te) Koi(ti,te)

    Kie(ti, ta) | Keo(ti, te)

    [Yi (t)Y> (t) |" under the conditions of Example 7.2.3 we

    represent the formula obtained for $21 (m, iG t) in the form

    yoi(m, K,t) te &

    ac

    i

    Kk.

    Substituting this expression at tf = ¢9 into Eq. (7.2.19) we obtain OK

    (ty, t2) :

    Sec

    tae

    M2(t2)

    K (ti, ta) | Fy

    0

    or in scalar form

    OK 41 (ti,to)/Ot2

    = —me2(t2) Kir (ty, t2) — mi (te) K12(t1, ta) ,

    OK y2(ti ; to) /Ot2 —- —aKyo(ti : ta) ; OK o1(t1, t2)/Ote

    =

    —mM2(t2)Ko1(t1,

    OK 20(t1, t2)/Ot2 = t2) —

    —aKo2(ti ; to) .

    m4 (t2) Koo(t1, ta) ;

    (I) The initial conditions for Eqs. =

    (I) have the form Kai(ts, t1) =

    fy 2). The second and the fourth equations are easily integrated.

    kai (ti) (h,l Their solu-

    tions are expressed by the formulae

    Kyo(ti, te) = kio(ty)e~ 202-*1) pn

    tapes ba:

    Koo(t1,to) = kaa(tyew"-™),

    to > th.

    Substituting these expressions into the first and the third Eqs. find Kii(ti ; to) and Ko (ty ; to).

    (11) (I) respectively we

    7.2. Methods of Normal Approximation

    and Statistical Linearization

    683

    ~~

    7.2.4. Approximate Determination of Stationary Processes NAM may also be used to determine approximately the characteristics of the strictly stationary process in the normal nonlinear stochastic differential system with the stationary normal white noise V. In this case a(y,t) = a(y), b(y,t) = b(y), v(t) = v do not depend on time t, and consequently, the functions y; and y determined by formulae (7.2.5)

    and (7.2.6) are also independent oft. For finding the expectation and the covariance matrix of the value of the stationary process at any ¢ we should put in Eqs. (7.2.3), (7.2.4) m= 0, K =0. As a result we obtain the equations

    yi(m, K) =0,

    Ly)

    y2(m, K) = 0.

    (7.2.23)

    If the constant vector m and the nonnegative definite matrix K exist sat-

    isfying these equations, and this solution of Eqs. (7.2.3), (7.2.4) is stable according to Liapunov then we may suppose that this solution characterizes the stationary process in the system. In this case for determining the covariance function k(r) of the stationary process we should use the same method as in Subsection 7.2.3. As a result we obtain

    dk(r)/dr = yai(m, K)K—'k(r).

    (7.2.24)

    Thus Eq. (7.2.24) with the initial condition k(0) = K determines the covariance function of the stationary process in the system att > 0. At

    7 h, where Eni(-) =

    i / (-)fni(y)nni(v)dydv; fri(y) and nni(v) being the

    normal densities of random values Y; and Vj.

    7.2.6. Normal Approzimation stonal Nonlinear Systems

    Method for Stochastic Infinite-Dimen-

    In the general case the exact solution of the equations of Section 5.9 for the multi-dimensional functionals g, of the normal stochastic systems is impossible. After taking the derivatives of the first and the second order with respect to both parts of Eq. (5.9.1) at nm = 1 we obtain the formulae of the time derivatives of an operator of the expectation m and the covariance operator K of the process Y;:

    m = Ed(Y,t), Kd = Efa(Y,t)(A, (Y — m)) + (Y — m)(A, a(Y, 2) + 0(Y,t)A], o(Y¥,t) = 0(Y,t)v(t)0(Y, t)*.

    (7232) (7.2.33)

    686

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    If we calculate here the expectations replacing the unknown true one-

    dimensional distribution by a normal one (3.7.00): gn (A;t) = exp | 1(A, m) — (KA,)| = exp (a, — s(A x» |:

    (7.2.34)

    then Eq. (7.2.32), (7.2.33) at H = Ey together with the correspondent initial conditions will represent a closed system of the ordinary differential equations in a separable H-space Y which approximately determines the operators m and Kk. Similarly after taking the mixed second strong derivative over j and Az with respect to both parts of Eq. (5.9.9) at n = 2 multiplied

    by [—2(A1, mz, )— i(A1, mez,)]and evaluating the expectation for the twodimensional normal distribution we get OK (ty, ta) OL

    v2 =

    En(%1,

    = m+, )(A2, a(Yt, _ m+, )) Shon

    Te,

    (7.2.35)

    2

    for the covariance operator K (#1, t2)A2

    =

    En ((Y1, —

    M+, ); A2(Y,



    Mt, )) p

    (7.2.36)

    Eq. (7.2.35) at the fixed t; is an ordinary differential equation in a H-space ) for the operator K(t;,t2) as the function tz at the initial condition

    K(t1,t1) = K(t1).

    (7.2.37)

    For approximate finding the stationary process in (5.4.10) at the stationary white noise V = W we may also use the normal approximation

    method. Accounting that in this case a(Y,t) = a(Y), b(y,t) = b(Y), v(t) = v, o(¥,t) = o(¥) do not depend on time and putting in Eqs. (7.2.32) and (7.2.33) m =0 and K =0, E = En we shall have

    Eya(Y) =0, (7.2.38) En [a(¥)(d, (Y — m)) + (Y — m)(A, a(Y)) + o(¥)A] = 0.(7.2.39) The obtained results may be summarized as follows. Theorem

    7.2.3.

    If the random process Y; in the normal stochastic

    differential system defined by Eq. (5.4.9) in a separable H-space has the bounded the first order operator moment m and the trace-type covariance operator K then its one-dimensional characteristic functional may be

    7.2. Methods of Normal Approximation

    approximated

    and Statistical Linearization

    by normal ones (7.2.34) with m

    and K satisfing

    (712583). (7.2.33). Corollary

    7.2.3.

    687

    Eqs.

    If the stationary (in the strict sense) process Y;

    in stattonary normal stochastic differential system (5.4.9) has the tracetype covariance operator then its approximate characteristic functional ts given by (7.2.34) with m and K defined by Eqs . (7.2.38), (7.2.39). Example

    7.2.9.

    Under the conditions of Example

    5.9.1 Eqs.

    (7.2.32),

    (7.2.33) for m; (x,t) and Ky; = Ki; (21, £2, 2), 1,j = 1,2 give

    dm,

    —_—_

    =m

    at

    OK Ot

    dm,

    40°m,

    a

    LT

    Rar?

    eee

    + 2him2

    OK

    Ki

    OK

    2 >

    ee Ot

    =

    zhama(ms + 3K 22) 4

    0? K

    Koo +c?

    Ox?

    (1)

    J

    = + h(x2,t)Ki2,

    07K =

    =

    K22 ait ce? a

    se

    aK

    att

    rere (pa,i a's 4

    rah

    = t 2h(a1,t)

    ie

    av

    adel3 4 sa

    0*Ki4

    “s 07m,

    dy

    ag, lonoies ae

    OzronennsOn?

    where h(a, t) =h,—h2

    Koi )



    ansdx2

    he

    [m3 (a, t) ap Ko2(a, ae t) 1,

    7.2.7. Methods of Equivalent Linearization and Structural Methods

    Let f(u,y) be the joint density of the random vectors U and Y, my and Ky be the expectation and the covariance matrix of the random

    vector Y, and |K,| # 0. From the known equations of the m.s. linear regression (Section 3.4): CT

    IT

    or

    gly = Puy

    rm

    (7.2.40)

    (7.2.41)

    we get CO

    Ee Kage

    8000

    = / fo —mu)(y — iy) Ky f(t, y)dudy — 00 —0o

    688

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    = f (maa) = ma} y= my)" Ky Aaladdy,

    (7.2.42)

    where f,(y) is the density of the random vector Y. This formula together with the approximate formula my,(Y) & my, + g(Y — my)

    (7.2.43)

    yields the Kazakov statistical linearization of the regression m,(Y ) (Subsection 7.2.2). Analogously, in case-of (7.2.41) we have



    Lat.

    s

    uy™ TZ}f(u, y)dudy = i mu (y)yT Py fi(y)dy.

    a

    cca

    —oo

    (7.2.44)

    Formula (7.2.44) together with the approximate formula

    my(Y) + gY

    (7.2.45)

    yields the Booton statistical linearization of the regression m,(Y). The regression m,(y) represents deterministic regression model for U = ¢(Y). To obtain a stochastic regression model it is sufficient to

    represent U = U(Y) in the form U = my(y) + U’ where U’ being some random variable. For finding a deterministic linear stochastic model it is sufficient to know the expectations m,, my and the covariance matrices Ky, Kuy. But for finding a stochastic linear model it is also necessary to know the distribution of the variable U for any y or at least expectation

    mux(y) (the regression) and the covariance matrix Ky(y) (coinciding with the covariance matrix Ky/(y) of the random variable U’). A more general problem of the best approximation of the regression by the finite linear combination of given functions y;(y), ..., gn (y) is reduced to the problem of the best approximation of the regression, as any linear combination of the functions yi(y), ... , gyw(y) represents a linear function of the variables u; = yi(y),..., un = yn(y). Example case of scalar

    7.2.10. Design the optimal polynomial regression model in the

    Y and U. Putting

    U1

    =

    Y;

    io

    Et

    bien sathNee als

    u=([luju...uy]’,

    g=[agn...gn]

    (I)

    7.2. Methods of Normal Appreximation

    and Statistical Linearization

    689

    we reduce the problem to the finding model

    u(y) = g(u).

    (II)

    Eq. (7.2.41) in this case takes the form

    [agi...gn] =

    1 a + QN

    where @p =

    Mie” a 3

    Bo SCR ra Att

    AN41

    rile the pth moment

    moment of [Y U]”

    ..-

    ={ muyrsenyn]

    (111)

    QIN

    of Y, and Yp =

    EUY?P

    is the (p iP re

    containing U in the first degree. Eqs. (II), (III) determine the

    polynomial m.s. regression of U on Y of the degree N which serves as the optimal polynomial regression model.

    The equations of the composed nonlinear stochastic systems after the equivalent linearization of the nonlinearities entering into the equations may be studied by the methods of Chapter 6 and first of all by the structural ones. Hence usually for the description of the composed stochastic systems the different describing functions are taken, for instance, the equivalent weighting functions g = g(t,7;m,K) and the transfer function ®(iw, m, K). The implementation of the spectral and the correlation methods

    (Subsection 6.1.8) for Eq. (5.4.6) assumed as symmetrized Stratonovich equation allows to substitute the joint solution of Eqs. (7.2.22)—(7.2.24) by the solution of Eq. (7.2.22) and the corresponding the equations for the matrix sy of the spectral densities and the covariance matrix K: Sy = Sy(w;m, K) = ®(tw;m, K)sy(w)®(tw;m, K)*,

    (7.2.46)

    Ee ae J slwim, Kydes

    (7.2.47)

    —0o

    Here s,(w) = v/27 is the spectral density of the stationary white noise V; ®(iw;m, Kk) is the equivalent transfer function of the statistically

    linearized system for Y°,

    Ye oem KY.

    y =Y —m,

    s=d/dt,

    (7.2.48)

    equal to

    ®(s;m, K) = —[a(m, K) — sI]~' b(m, K),

    (7.2.49)

    690

    Chapter 7 Methods

    where

    @(m,K)

    =

    of Nonlinear Stochastic Systems Theory

    Eya(Y,t)Y°T

    =

    oi(m,K)K7};

    b(m,K)

    = Ey0(Y,t).

    7.2.8. General Problem of Stochastic Systems Normalization In the class of the stochastic systems there are the simplest systems which permit comparatively simple mathematical description and at the connections form the systems of the same type. These are the normal systems. A normal system we shall call such a stochastic system whose the joint distribution of the input and the output is normal at the normal distribution of the input. It goes without saying that at the distribution of the input different from the normal one the joint distribution of the input and the output may be nonnormal even a normal system. The opportunity of applying the well developed methods of the linear systems theory (Chapter 6) to the normal systems and the relative simplicity of the mathematical description of the normal system cause the natural quest for extending the methods of the research of the normal systems as the approximate ones over sufficiently the large class of the stochastic systems different from normal. Thus the problem of the normalization of the stochastic nonlinear systems arises, i.e. the problem of finding a suitable normal model for a given nonlinear stochastic system.

    Let us prove the following theorem. Theorem 7.2.4. For the normalization of the given stochastic nonlinear system it is sufficient to linearize the regression of the output on the input and to average over all possible realizations of the input the conditional covariance operator of the output at the given input. > In order to linearize the regression we may try to approximate it by the linear dependence on the criterion of the minimum of the mean square error (m.s.e.) at any

    instant (to use the method of least squares).

    Then the problem will be reduced to

    the definition of the linear operator I, and the variable a from the condition

    Bla+ LX — E[Y|X]|? = min,

    (7.2.50)

    where / [Y |x | is a conditional expectation of the output Y at the given realization & of the input

    X (Section 3.4). We apply to the random variable

    Z =

    a+ LX —Y

    the known relation:

    E\Z)? = E\E[Z|X]|? + ED[ Z|X],

    (7.2.51)

    7.3. Methods

    of General Nenlinear Stochastic Systems Theory

    which is valid for any random variables

    691

    X , Z where Di Z |x |is a conditional variance

    of the variable Z at a given realization & of the variable X. As a result we obtain

    from (7.2.50)

    Bla+ LX -Y? =Bla+LX —ELY|X]/?+ED[Y|X].

    (7.2.52)

    Hence it is clear that the problem of the optimal linear approximation of the regression of the random variablé Y on _X is equivalent to the problem of the optimal linear

    approximation of the random variable Y itself on the m.s.e. criterion.
    While deriving the equations for J, we calculate the stochastic differential

    of the product [Yi (t) —mM, (t)]" Fa hYS(#) — Mp (t) pe by means of It6 formula. But at first we notice that formulae (7.1.1) and (7.1.5) at approximation (7.4.1) may be written in the following form:

    i=

    $1,0(m, K,t) + se SS ~i,v(m, K,t)qv(@)

    ;

    (7.4.6)

    l=. |v

    N

    K= ~2,0(m, K,t) + os 3

    ~2,v(m, K,t)q,(a@),

    (7.4.7)

    [=8F 1

    where

    (7.4.8)

    yiy(m, K,t) = fistartunGhostoas,

    pay(m, K,t) = /{a(y,t)(y7 — m?) + (y— m)a(y,t)? + o(y,t)}wi(y)pr(y)dy,

    t)”. o(y,t) = b(y,t)v(t)b(y,

    (7.4.9)

    (7.4.10)

    Here the coefficients ~1,0(m, Kk, t) and 2,0(m, 1¢ t) are determined by formulae (7.4.8) and (7.4.9) at py = po(y) = 1 with the change v by 0.

    698

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Evaluating directly by formula (5.1.7) the stochastic differential of the product [Yi(t) - m,(t)]" ane [Y,(t) — Mp (t)|"? we get with the account of (7.4.8) and

    (7.4.9) N

    jin = Gr 0(m, Kit) +) D| Pro(m, K,t)av(a) i=S

    Pp

    SISO fee

    |v| =!

    ACT

    Sh te SEN

    (7.4.11)

    s=1

    where the functions py (m, KB, 0) are determined according to (7.4.4) by means of the change of the products Yj ...Yp by (y1 — m1) aoa (Yp _ Mp), and €, is a vector whose all the components are equal to () besides the s*

    component equal to 1.
    YS) ga(m,K, the, 9.

    (7.5.9)

    i—sa yi

    Now let derive the equations for the coefficients c, in (7.5.1). >: We use formula (3.8.7) changing index VY on K:

    c= E (=) acsd|

    (7.5.10)

    Differentiating this formula with respect to time t and taking into consideration that the polynomial qx (y) depends on the expectation M and the covariance matrix K of the random vector Y; which are the functions of time we find

    x= fu) 2B} (BY + tr {|af (x) nt]

    A=0

    '

    so] (7.5.11)

    where qn’ (y) is the matrix-column of the derivatives of the polynomial qx (y) with respect to the components

    of the vector m,

    and a

    is the square matrix of the

    derivatives of the polynomial dx (y) with respect to the elements of the matrix K.

    7.5. Methods based on Orthogonal Expansions and Quasimoments

    721

    Substituting into this formula the expression 0g;(A;t)/0t from Eq.

    (5.6.20) and

    replacing the density f (y; t) by its truncated orthogonal expansion (7.5.1) we obtain

    én = |{ooea [iAP ay, t) + x(b(y,t)" Ast)] ors

    A=0

    N

    x}

    1+ x > cypy(y) | wi(y)dy+q i=3

    (a)m-+ tr [ak (a) K] 7 UGo.12)

    |v |=!

    where x (us; t) for the normal white noise is determined by (5.6.37), qe’ (a) and qk (a) represent as usual the resultOf replacing the monomials v4 whee Yp? in the expressions of the polynomials qn (y) and ons (y) by the corresponding moments Qp, ,

    .-,@r,. Putting

    gxo(m, K,t) = [{ a (aa) x [id? a(y, t) + x(b(y, t)? A;pera

    wi(y)dy,

    (7.5.13) nv (dads, ts

    is dn eaea [iA E (ax)

    a(y, t)

    + xbnOT rider) —po(adan(adey

    (7.5.14)

    and substituting the expressions of m and K from (7.5.8), (7.5.9) we obtain the equations

    Ce = Gno(m, K,t) + yio(m, K,t)? g(a) + tr [po0(m, K, t)g* (a)| N

    +> So {eev(m, K,t)+e10(m, K,t) ae(a)+ tr [par (m, K, tae (a)] i=se|pl=t

    N

    FP l=3

    i2

    8.6)

    ol

    (7.5.15)

    |v) =I

    Here the moments @,, ,...,

    @p, in qn’ (a) and qi (a) must be replaced by their

    expressions in terms of the coefficients Cy in accordance with (3.8.14).
    ty and Eq. (7.5.24) with the initial condition K(ti,t1) = K(ti) we determine approzimately the covariance function K(t1,t2) of the process Y(t) and all the coefficients c,,,, of truncated expansion (7.5.23) of the two-dimensional density fo(yi, y2;t1,t2) which remained unknown. Further, subsequently integrating Eq. (7.5.30) with initial conditions (7.5.33) att) N at least n—N subscripts are equal to zero for any of the coefficients cy,,..,v,

    tm (7.5.23).

    As a result all the multi-dimensional

    distributions of the process Y(t) will be approximately determined. 7.6.4. Consistent Hermite Polynomial Expansions

    In the same way as in Subsection 7.5.2 it is proved that while using the consistent Hermite polynomial expansions the elements of the

    7.5. Methods based on Orthggonal Expansions and Quasimoments

    matrices qf)

    729

    x, (@), qe _.,&,(&) are proportional to the correspond-

    ing quasimoments ne oben

    ho) Sher otk

    eee her

    (hela tesailerl pare reba Gah gamers)

    (Py 8

    Anes

    = mthrins

    SA ioc, AN).

    ong. pep Sey, -

    ee ail oe og MOdey

    (7.5.34)

    } Kn —Cs

    (7.5.35)

    tse ects [Kyl atO eas. ,LV),

    1 oe ec)

    == Ta kine(Kner -

    io ag er)

    Dera gta

    hae

    er )

    (7.5.36)

    = —KnrknsCry,...,%n—er—es

    (rv, $ el 5.32 Dee

    (7.5.37)

    THlkifehe+ leelieont AGN),

    |i] +--+ [Kn] =3; Oat [ra] ee Ger... ee(%) ==Oat + +--+ [en] =3and |r| +--++ [en] =4; B

    (7.5.38)

    In the case of the normal white noise V in Eq. (5.4.6) we have

    {Mee Gastric oraz aa) #x(b(un sta) Anitndle™e } ONO)

    =

    (Gs, Yt



    Mt,

    +++)

    Yn-1



    mz,

    9/40An)[irp

    (Yn, tn)

    +x(btn)” (tAnitn) n } EXPLIAR(Yn— Mn)}} 4, to), c20(ty ; to),

    c13(ty 3 to) of the two-dimensional distribution of the process Y are determined by the equations: OK (ty, t2)/Ote =

    —3[m?(t2) = D(t2)|K (t1, ta)

    —3m/(t2)c12(t1, t2) — c1a(t1, te),

    (1)

    ci (ti, t2)/Ot2 = —3m(t2) K?(t1, te) — 3[m?(t2) + D(t2)]c21(t1, te) —6K (ty, t2)c12(t1, t2)— 3m(t2)c22(th, ta),

    (11)

    Oci2(t1, t2)/Ote = 2[vm(te) — 8m(t2) D(t2) — 6c3(t2)| K (t1, te)

    +[v — 6m?(t2) — 12D(t2)]e12(t1, t2) — 6m(t2)c13(t1, ta),

    (III)

    desi (t1,t2)/Ot2 = —6K3(ty,te)— 18m(t2) K(t1, t2)ca1(t1, t2)

    —3[m?(t2) + D(t2)]es1(t1,t2)— 9K (t1, te) c20(t1,t2)+ ¢3(t1)ea(t2), (IV) Oc29(t1, t2)/Ote = 2[v — 6D(te)| K7(t1, t2) + 2[vm(t2)

    ~8m(t2) D(t2) + ¢3(t2)]ca1(t1,t2)— 24m(t2) K(t1, t2)c12(t1, t2) +[v—6m?(t2)—12D(t2)]c22(t1, t2) -12K (t1, t2)c1a(t1, t2) +6 D(t1)D?(t2), Vv 0c13(ti } to) /Ot2 =

    3{ D(t2)[2v+3D(t2)]

    +3[2vm(t2)

    —6m(t2)c3(t2)

    =F 12m(t2)D(t2)

    +3[v =

    3m?(t2) =

    —3c4(t2)}

    + c3(t2)}c12(t1,

    9 D(t2)}e13(t1,

    K (ty Oe

    ta)

    t2)

    (VI)

    with the initial conditions

    K(ty,t2)

    = Di);

    c3i(ty, ty) =

    coi(41, 41) = cia(t1, t1) = ca(ty),

    €29(t1, t1) =

    €13(ty, t1) =

    c4(ty).

    (VI)

    7.5. Methods based on Orthogonal Expansions and Quasimoments

    The

    quasimoments

    ¢111(t1,¢2,¢3),

    C211(ti,t2,t3),

    (au

    c121(t1, ta,ts),

    €112(t1, ta, t3) of the three-dimensional distribution are determined by the equations

    e111(t1, to, t3)/Otz = —3[m?(tz) + D(ts)]e111(t1, ta, ts) —3m(t3)c112(t1, te, t3) — 6m(t3)K (t1, ts) K (ta, ts) —3K (ty, t3)e12(ta, ts) — 3K(to, t3)c12(t1, ts),

    (VIII)

    c211(t1, te, t3)/Ot3 = —12m(ts) K(t1, t3)c111(t1, ta, ts) —3[m?(t3) + D(ts)]e211(t1, te,t3) — 6K (t1, t3)¢119(t1, ta, ts) —6K?(t1,t3)K(ta, ts) + c3(t3)co1(t1, ta) —6m(t3) K (te, t3)co1(t1, t3)— 3.4K (te, t3)c22(t1, ta),

    (IX)

    0cj21(t1, te, t3)/Otz = —12m/(tz)K (ta, t3)e111(t1, ta, ts)

    —3[m?(ts) + D(ts)]c121(t1, te,ts) — 6K (t1, t3)¢112(t1, ta, ts)

    —6K (ti, t3)K?(te, t3) — 6m(t3) K(t1, ts)c21 (to, ts) +c3(t3)c12(ti,ta) — 3K (t1,t3)c22(te, ts),

    (X)

    0c112(t1, te, t3)/Ot3 = 2[vm(t3) — 6m(tz) D(tz) + c3(tz)]e111(t1, te, ts)

    +[v — 6m?(t3) — 12D(ts)]e112(t1, te, t3)

    +2[v — 6 D(ts)] K(t1, t3)K (to,ts) — 12m/(t3)[K (to,t3)e12(t1, ts) ic (pata ote(torte iene| We(egrets tite ick K(taste cis ta,ta ex) with

    the

    initial

    conditions:

    = c22(t1,t2); C121 (t1,

    e111 (ty yo7ts)

    b=

    Cro bas to);

    co11(ti UD t3)

    t2,t3) = C112(ti, te, t3) = c13(t1, tz). The only quasi-

    moment of the four-dimensional distribution C11 11(t1 SUD UEY t4) depending on four

    variables t;, tg, t3, t4 is determined by the equation

    0¢1111(t1, tz, t3, t4)/Ota = —3[m? (ta) + D(ta)]er111(t1, ta,ts, ta)

    —6K (ti, t4)K (ta, ta) K (ts, ta) + e3(ta)er11(t1, ta, ts) —6m(t4)[K (ts, ta)c111(t1, to, ta)+ K (to,ta)er11(t1, ts, ta) +K (ty, t4)ce111(to, ts, t4)] — 3[K (ta, ta)e112(t1, te, ta) +K (te, ta)ce112(ti, ts,ta) + K (tr, ta)c112(t2, ts, ta)] with the initial condition e1111(t1 , to, ts, t4) = €112(t1 UB ts).

    (XII)

    Coe

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    7.5.8. Approximate Determination

    of Stationary Processes

    For the approximate determination of the one-dimensional distribution of the strictly stationary process in stochastic stationary nonlinear differential system (5.4.6) by the methods based on the orthogonal ex-

    pansions one should put in Eqs. (7.5.8), (7.5.9), (7.5.15)

    m=0,

    K=0,

    & =0.

    (7.5.40)

    If the equations obtained in such a way have the solution which may serve as the parameters of the corresponding truncated orthogonal expansion of the one-dimensional distribution, then we may suppose that the strictly stationary process exists in a system. In this case for determining other multi-dimensional distributions of this stationary process one should replace the derivatives with respect to t, in Eqs. (7.5.24), (7.5.30) by the derivatives with respect to T,-1 = tn —t, and write the

    initial conditions (7.5.33) in the form Croeana (11 yrrey

    Tn-2) Treo) Goi

    tee 5 eal

    (1 Tecanape h) Tn—2) .

    (7.5.41) We summarize the results obtained in Subsections 7.5.1—7.5.5 as the following statements. Theorem 7.5.1. If the random process Y(t) in normal stochastic differential system (5.4.6) has the finite coefficients of the consistent orthogonal expansions then its one- and multi-dimensional distributions may be approximated by expansions (7.5.1), (7.5.23) with m(t), K(t), Cr(t)s Fe(tito) d Gay Alone (E> “tut, En) defined by Eqs. (£9578) labo 9),

    (7.5.15), (7.5.24), (7.5.30) at initial conditions (7.5.16), (7.5.33).

    Corollary 7.5.1. If the stationary (in the strict sense) process in stationary normal stochastic differential system (5.4.6) exists and has the finite coefficients of the consistent orthogonal expansions then tts one- and multi-dimensional distributions may be approximately expressed by (7.5.1), (70:23) nth im “hep PACr) sepa Gee, Ta; Tan) defined by Eqs. (7.5.40), (7.5.41).

    The equations of the orthogonal expansions method for Eqs. (5.4.6) at the nongaussian white noise and Eqs. (5.4.7) are given in Problems 7.34-7.35.

    7.5. Methods based on Orthogonal Expansions and Quasimoments

    733

    7.5.6. Reducing Number of Equations

    The main difficulty of practical use of the stated methods especially for the multi-dimensional systems is a fast increase the number of the equations for the coefficients of the segments of the orthogonal expansions with the increase of the dimension p of the state vector y and the maximal order N of coefficients used. The dependence of the equations number for the moments cited in Tables 7.3.1 and 7.3.2 remains valid for the coefficients of orthogonal expansions, too.

    While comparing the ciphers v, [v7 .. vty in Tables 7.3.1 and 7.3.2 we see that the number of the equations of the orthogonal expansion may be considerably reduced by means of the approximations of the distributions which include the mixed coefficients of the second order only. Such an approximation of the distribution may be obtained by different ways. For instance, we may assume that all mixed coefficients of the orthogonal expansions of higher than the second order are equal to zero. Another way consists of the use of the known expressions of the mixed central moments in terms of the variances and the covariances of the components of the random vector for the normal distributions (7.4.35). Both these approaches give the opportunity to use any approximations of the distributions and the correspondent equations for the coefficients of the orthogonal expansions on which the approximating function f(y;0) or f*(y1, .-- , Yn} Un) depends in Eqs. (5.4.6). Thus while using the quasimoments method it is natural to neglect all the terms of the expansions containing the coefficients c,, cy, ,...,y, with

    more than one component of the multi-indexes v, [v7 ...v7]? different from zero. The approximation of a distribution (Subsection 7.4.4) may be obtained in various ways. For instance, one may assume that all mixed semiinvariants of orders higher than the second are equal to zero. Then

    one obtains the recursive formula (7.4.37) determining the mixed moments of the third, the fourth and higher orders as the function of the moments of the components of the random vector and the mixed second order moments (Kashkarova and Shin 1986). Another way is to use the known expressions for the mixed central moments in terms of the variances and the covariances of the components of the random vector for

    the normal distributions Satna

    [Pea Unescenstei

    Pry ,...,Ts

    9mm!

    :

    _ ) ie Paka

    a ()} ifri+---+r,

    ga

    =2m+1

    if ry

    +--+;

    (iP

    i

    Adis,

    (7.5.42)

    734

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Both these approaches allow one to use any of the distributions of Sections 7.4 and 7.5 and to write the corresponding equations for moments, the semiinvariants and the coefficients of orthogonal expansions on which the approximating density depends. If the method of orthogonal expansions (in particular the method of quasimoments) is used, then it is natural to neglect all the terms of the expansions containing coefficients Cy, Cy,,...,v,

    v,[v?...v?P]°

    n

    With more than one component of the multi-indexes

    C2, different from zero.

    Example

    7.5.4. For the system of Examples 7.4.1 and 7.4.3 the method

    of this subsection using formula (7.5.42) gives

    H12(t1,t2) = ar(ti,t2)

    ,

    =0, pis(ti, te) = 3D(t2)K (ti, te),

    J22(t1,te) = D(t1)D(t2) + 2K* (tr,

    te),

    (I)

    Hai(t1, t2) = 3D(t1) K(t1, ta) ,

    Prii(t1, t2,t3) = 0, Hi1a(t1, te, t3) =

    D(t3)K (t1, ta) + 2K (t1,t3) K (te, t3) ;

    }121(t1, ta, ts) = D(tz)K (ti, ts) i 2K (ti, tz) K (ta, ts) F

    (i)

    faii(t1, ta, t3) = D(ty) K (ta, t3) + 2K (ti, to) K (ti : t3) .

    #1111(#1, ta, t3, ta) = K(ti, to) K(ts, ta) + K(t1, t3)K (te, ta) + K (ti, ta) K (te, t3) :

    (III)

    Substituting the obtained expressions of Hio(ti ‘ ta) and His(ti ; to) into the equation of Example 7.4.3 for the covariance function we get

    OK (t1, t2)/Ot2 = —3 [m?(t2) + D(tz)] K(t1,t2), te > th.

    (IV)

    Integrating this approximate equation with the initial condition K (ti ‘ to) = D(t1) we find all the mixed fourth order moments using the above formulae.

    Then all the

    multi-dimensional distributions of the process Y (t) will be determined in the same way as in Examples 7.4.1 and 7.4.3.

    The same equation for the covariance function yields the method of quasimo-

    ments if in formulae (I) and (II) of Example 7.4.3 we put c, = 0 if more than one component of the multi-index VY is different from zero.

    Example

    7.5.5. For the system of Example 7.4.2 formulae (7.4.37) give

    M12 = far = 0, pis = 3kigke2,

    poe = kiike2 + 2k?,, p31 = 3kirk12,

    (I)

    7.5. Methods based on Orthogonal Expansions and Quasimoments

    139

    and the equations of Example 7.4.2 for the moments of the one-dimensional distribution are reduced to:

    my

    =

    —™M 1M



    ky,

    my

    =

    —amyg,

    (II)

    ki = —2(m2k11 + myki2), kia = —(m2 + a)ki2 — miko. ,

    ko = —2ako9 + h2v, 30

    (II)

    = —3(M2f30 + 2ki1k12), fos = —3apX03,

    [t40 = —4(3k1 24430 + Moplag + 8myki1k12), fos = —4apiog + 6h*vk 29.

    (IV)

    Equalizing to zero the coefficients C12, C21, C13, C22, C31 in the method of quasimoments

    yields the above equations for ™ 1, M2,

    k11, k12, k22, but the equations

    for {403, 430, [40 and /94 are replaced by

    €30 = —3(2k11k12 + meoczo), ¢o3 = —3acoz , C49 = —4(3k12¢30 + M2c40),

    Coa = —4aco4.

    (V)

    As for the multi-dimensional distributions, all the approximations of this subsection yield the equations

    OK 11(t1, t2)/Ot2 OK 12(t1,te)/Ot2 OK (tr, tz) /Ote OK 29(t1, t2)/Ot2

    —mo(t2)Ky1(t1,t2) — my (t2)K12(t1, ta) , —akj(t1, ta),

    —m2(t2)Ko1(t1, t2) — mi (te) Kea(ti, te) ,

    (VI)

    —aKo9(t1, te) , py > Ula

    for the covariance function of the process Y(t) =

    [Y4 (t1)Y (te ie

    Integrating

    at first the equations for the moments or quasimoments of the one-dimensional distribution, and then the above equations for the covariance functions with the ini-

    tial conditions:

    Kyi (ti k t;) =

    Koo(ti 4 t,) =

    koo(t1) we may find all the multi-dimensional distributions of the

    kii(t1), Ky2(t1, t1) =

    Kai (ti < ty) =

    ky2(t1),

    process Y (t) = [Y (t1)Y (te) 1, using the corresponding fragments of the Hermite polynomial expansions or of the Edgeworth series.

    Examples 7.5.4 and 7.5.5 illustrate essential simplifications which give the approximations of distributions considered in this subsection. In the problem of Example 7.4.2 the number of equations for the complete set of moments of the first four orders amounts to 115 instead of 13 obtained in Example 7.5.5.

    736

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    7.5.7. Methods based on Orthogonal Expansions for Discrete Nonlinear Systems

    As applied to stochastic discrete nonlinear systems (5.7.1) we obtain the following equations of the orthogonal expansions method: film

    a

    Yn;

    0)

    =

    w(y1 prety

    N

    x

    Yn y™, K)

    .

    ¢14+ Se

    SD

    el Hieale ae eeen F , doespe

    osSad Mock)

    h=3 |r| +--+ |Yal =h

    0 oh

    =

    Eq

    (wi(%, Vi) =

    :

    E (55) E exp {iP wi(Mi, v)} | A=0

    cli rlntl SRG ee

    )

    (7.5.44)

    liebe {1,...,vn +being the biorthogonal polynomials with (oe)

    fi — 00

    Kite

    co

    fei ees tal

    RP

    cose

    ee

    ta)

    —oco

    uh

    BA

    dui

    rg adyua

    Obs eee rong

    (7.5.48)

    CCK Mead LZ ET ht Rb er aaa 50Pearl2ponsSer ae ly mh Wales ors Paht sade deg) Mery mbes Lyetlen| =r id Ha Tckielig l= 1, yw.

    Thus Eqs. (7.2.29), (7.2.30), (7.5.44), (7.2.31), (7.5.45) atn = 2 and Eq.

    (7.5.45) at n > 2 determine recurrently all the multi-dimensional

    distributions {Y;} in stochastic discrete nonlinear system

    (5.7.1).

    7.6. Structural Parametrization of Distributions.

    EAM

    Wot

    For the continuous-discrete stochastic systems the corresponding equations are given in (Pugachev and Sinitsyn 2000) 7.5.9.

    Other Methods of Distributions Parametrization

    The distributions parametrization methods stated above are based on the consistent biorthogonal expansions of the densities and lead to an infinite system ofthe nonlinear ordinary differential equations for the coefficients of the expansion. For obtaining the approximate solutions this system is substituted by a finite one. The estimating of the influence of the rejected terms in the general case encounters with difficulties. The

    Fourier series method (Moschuk 1994) is based only on the orthogonal expansions of the densities and is reduced to the solution of an infinite system of the linear ordinary differential equations for the coefficients of the expansion. Here “the linearity” of the reduced system allows to estimate the accuracy of the Fourier series method. The main advantage of a given method over the known ones is in the availability of the accuracy estimates and in the possibility to construct analytical expressions for the coefficients of the expansion of an one-dimensional density and transfer densities. The corresponding algorithms are simpler than in the moments methods, the semiinvariants and quasimoments methods. The method is easily generalized on the stochastic discrete and continuousdiscrete nonlinear systems and on the stochastic composite systems as well. The method of the orthogonal expansions may be successfully used in those cases when the additional restrictions over the structure of a

    distribution are known (Section 7.6). 7.6.

    Structural Parametrization

    of Distributions.

    Ellipsoidal Approximation Method 7.6.1. Introductory Remarks

    For structural parametrization and/or approximation of the probability densities of the random vectors we shall apply the ellipsoidal densities, i.e. the densities whose planes of the levels of equal probability are similar concentric ellipsoids (the ellipses for two-dimensional vectors, the ellipsoids for three-dimensional vectors, the hyperellipsoids for the vectors of the dimension more than three). In particular, a normal distribution in any finite-dimensional space has an ellipsoidal structure. The distinctive characteristics of such distributions consists in the fact

    738

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    that their probability densities are the functions of positively determined quadratic form u = u(y) = (y? —m™ )C(y—m) where m is an expectation of the random vector Y, C' is some positively determined matrix. > Let us find ellipsoidal approximation

    (EA) for the density of 7-dimensional

    random vector by means of the truncated expansion based on biorthogonal polynomials {pr,v(u(y)), dr,v(u(y))}, depending only on the quadratic form U =

    u(y)

    for which some probability density of the ellipsoidal structure w(u(y)) serves as the

    weight: co

    / w(u(y))Prr(u(y))aru(u(y))dy =6.

    (7.6.1)

    —oco

    The indexes VY and [U at the polynomials mean their degrees relatively to the variable u. The concrete form and the properties of the polynomials are determined further.

    But without the loss of generality we may Then the probability density of the vector

    assume

    that qr,o(u)

    =

    Pr,o(u)

    =

    all.

    Y may be approximately presented by the

    expression of the form N

    f(y) © f*(u) = w(u) > cry Pry (u).

    (7.6.2)

    yv=0

    For determining the coefficients Cry in (7.6.2) we multiply term-wise equality (7.6.2) by Yr,yu(u) and integrate over the whole space 2”. Then by virtue of condition

    (7.6.1) for pp = 1,..., N we obtain CO

    [oe)

    N

    / NPM CaLTES ihw(u) )enyPrv(t)Gryul)dy = Cry. Thus the coefficients C, ,,are determined by formula

    loo)

    cov = f Ho)ten(wldy = Baew(U), VEIN).

    (7638)

    —oo

    As Pr,o(t) and qr,0(U) are reciprocal constants then always C, 0Pr9

    (the polynomials

    of zero degree)

    = 1 and we come to the result: N

    fly) & f*(u) = w(u) 1+ So Cry Prv(u) v=2

    4

    (7.6.4)

    7.6. Structural Parametrization of Distributions.

    EAM

    739

    Formula (7.6.4) expresses the essence of the ellipsoidal approximatron of the probability density of the random vector Y. > For constructing the system of the polynomials

    {pr, v(u u), qr, u(u) } satisfy-

    ing the condition of the biorthonormality relatively to the quadratic form u ==(y

    = m)?C(y = m) we transform condition (7.6.1). The change of variables =m

    te Ca 1/2¢ gives U = (To==

    © =

    y

    I¢|? and

    |e0oredanalwdy =lor? fw lCP ypeo(lCPannICP4G a

    (7.6.5) aiVJu,..

    CaSsia

    Hence “s take

    We make the change of variables

    (; =

    into account that C Spee

    (To = uU and achemrecem nb ay S00

    sSe C=

    ae Cea

    After the transformations (7.6.5) is reduced to the following form:

    1/2_r/2

    = a

    P

    /Wee Dian (14) lash alee aes) 0

    For the applications the case when

    the normal

    distribution is chosen as the

    distribution w(u) is of great importance

    w(u) = w(z7? Cz) = accounting that C' = K—!

    1

    exp(—z? K~12/2);

    Jen ikl

    (7.6.7)

    we reduce the condition of the biorthonormality (7.6.6)

    to the form

    1 SR

    Spy

    i |Pov (e)arald

    r/2-1,—u/2 é

    thU SnOe ft

    (7.6.8 )

    0

    Thus the problem of the choosing of the polynomials system {p,,,(u), dr,u(u)} which ts used at the ellipsoidal approrimation of the densities

    (7.6.4) and (7.6.3) ts reduced to finding a biorthonormal system of the polynomials for which the x?-distribution with r degrees of the freedom serves as the weight.

    740

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Taking into consideration that y?-distribution represents a special

    case of the gamma-distribution (Appendices 2, 7) koi

    w(u) = eae

    one

    (7.6.9)

    we consider a system of the polynomials S?(u), d’

    By (uy = (Fi ean te

    (ut? Cay

    NG > Alar

    OY

    1076.10)

    At a=r/2—1 and k = 1/2 we obtain a system of the polynomials which are orthogonal relatively to y?-distribution with r degrees of the freedom:

    (r+ 2v — 2)!! see oe ) = Seta)hcae Vv

    A

    Sie

    (pe

    |)

    T—0)

    The main properties of the polynomials S?(u) and S,,(u) are given in Appendix 2. Between the polynomials S,,(u) and the system of the

    polynomials {p,,,(u), dr,.(u)} the following relations exist: Pr,v(u) = Sey (th),

    (r— 2)!

    Gry(u) = Ga Qv aDann)»

    (7.6.12)

    r>2.

    (7.6.13)

    Now we consider the expansion of the densities of nonnegative ran-

    dom variables with respect to the polynomials S%(u) and S,,(u). Let L2([0,00)) be a H-space of the functions integrable with the modulus square. After test we make sure that the functions

    Oya) = V/s),(4) = if eae

    Seu)

    (7.6.14)

    form the orthogonal basis in L2([0,00)). Therefore any function F(u) € L2([0,0o)) may be presented as m.s. convergent (in the norm of the space) expansion F(u) = Vw(u) rc, s, (a),

    (7.6.15)

    7.6. Structural Parametrization

    of Distributions.

    EAM

    741

    whose coefficients are determined by the relation

    Rac itF(u)S2(u)du /Sea

    (7.6.16)

    In particular any probability density y(u) of the nonnegative ran-

    dom variable satisfying the condition {o(u)/Vo(u)} € L2([0, co)), ss2*(u) cat) co

    0

    2

    oo

    du

    u

    =

    T(o+1)

    a,u/2,.2 / era (u)du < ov, 0

    (7.6.17)

    may be presented by the expansion

    p(u) = w(u) >> cy Se(u).

    (7.6.18)

    v=0

    Now we consider the H-space L2(R") and the orthogonal system of the functions in them where the polynomials S,,(u) are given by formula

    (7.6.11), and w(w) is a normal distribution of the r-dimensional random vector (7.6.7). It is evident that this system is not complete in L2(R"). But the expansion of the probability density f(u) = f((y? — m?)C(y —m)) of the random vector Y which has an ellipsoidal structure over the polynomials p,,(u) = S,,,(u),

    f(u) = w(u) pe Cru Prinl o),

    (7.6.19)

    v=0

    m.s. converges to the function f(u) itself. The coefficients of the expansion in this case are determined by relation (7.6.16) which by virtue of

    (7.6.12) and (7.6.13) will take the form

    cow =f flu)pen(udy /PPEOE

    (7.6.20)

    Thus the system of the functions {,/w(u)S,,(u)} forms the basis in the subspace of the space Lo(R") generated by the functions f(u) of

    the quadratic form u=(y—m)? C(y—™m).

    742

    Chapter 7 Methods

    Remark.

    of Nonlinear Stochastic Systems Theory

    In practice the weak convergence of the probability measures ge-

    nerated by the segments of the density expansion to the probability measure generated by the density itself is more important than m.s. convergence of the segments of the density expansion over the polynomials Sry (u) to the density namely, N

    feoyd A uniformly relative to

    Cry Pry (u)dy a

    ams A at N —

    /f(u)dy

    (7.6.21)

    A oo on the o-algebra of Borel sets of the space R’.

    Thus the partial sums of series (7.6.19) give the approximation of the distribution, i.e. the probability of any event A determined by the density f(u) with any degree of the accuracy.

    The finite segment

    of this expansion may be practically used for

    an approximate presentation of f(u) with any degree of the accuracy even in those

    cases when f(u)/V w(u) does not belong to D2(R").

    In this case it is sufficient

    to substitute f(u) by the truncated density (Subsection 3.8.1). Expansion (7.6.19) is valid only for the densities which have the ellipsoidal structure.

    It is impossible in

    principal to approximate with any degree of the accuracy by means of the ellipsoidal approximation (7.6.4) the densities which arbitrarily depend on the vector y.

    It is not difficult to make sure in the fact that at the erpansion over the polynomial S,,(u) the probability densities of the random vector Y and all tts possible projections are consistent. In other words, at integrating the expansions over the polynomials S;,4;,(u), h+1 = 1, of the probability densities of the r-dimensional vector Y ,

    N 1

    fly) = Van]

    —u

    V1.4 Da Coin Shou)

    u=(y-m)K(y-m),

    ;

    y=[y7y7f,

    (7.6.22)

    on all the components of the /-dimensional vector y’’ we obtain the expansion over the polynomials S;,,(u1) of the probability density of the h-dimensional vector Y’ with the same coefficients,

    f(y’) =

    ip

    1 Set?

    VJ(27) | Kai

    (yn ie

    (he

    N

    1+

    d

    Ch, Shy (us

    :

    cen eh ta

    where K,. 1s a covariance matrix of the vector Y’.

    ,

    (7.6.23)

    7.6.

    Structural Parametrization

    But approximation

    of Distributions.

    (7.6.23) the probability

    BAM

    743

    density of h-dimen-

    sional random vector Y‘ obtained by the integration of expansion (7.6.22) the density of (h + 1)-dimensional vector is not optimal ellipsoidal approximation of the density. For the random r-dimensional vector with an arbitrary distribution the ellipsoidal approximation (7.6.4) of its distribution determines exactly the moments till the N‘ order inclusively of the quadratic form

    U =(Y —m)? K-1(Y —m), ie.

    i HO gis Hecate BhateraenTees p

    (7.6.24)

    (E”4 stands for expectation relative to ellipsoidally approximated dis-

    tribution). The ellipsoidal approximation of any distribution determines exactly the expectation of the polynomials p,o(u), gr,o(u),---,Pr,w(U), dr,n(U). In this case the initial moments of the order s, s = s; +---+ s, of the random vector Y at the approximation (7.6.4) determined by formula ess

    co

    cc

    — Sir —- As

    N

    — =

    Sj BY;

    y=2

    ae

    Ms

    [o@}

    ms / yi. yp"w(u)dy + ss Cry | yi) Yous ae

    s oe

    pry (u)w(u)dy!

    (736.25)

    Lae

    Thus at the ellipsoidal approrimation of the distribution of the random vector its moments are combined as the sums of the correspondent moments of the normal distribution and the expectations of the products of the polynomials p,,(u) by the degrees of the components of the vector

    Y at the normal density w(u). 7.6.2. Ellipsoidal Approrimation of One-Dimensional Distribution We use the ellipsoidal approximation method (EAM) for finding the one-dimensional probability density f;(y;t) of the p-dimensional random process Y(t) which is determined by the Eq. (5.4.6) at the normally distributed white noise.

    Suppose that we know a distribution of the

    initial value Yo = Y (to) of the random process Y(t). Following the idea of EAM (Subsection 7.6.1) we present the one-dimensional density in the form of a segment of the orthogonal expansion (7.6.4) in terms of the polynomials dependent on the quadratic form u = (y? — m?)C(y — m)

    744

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    where m and K = C7! are the expectation and the covariance matrix of the random process Y(t): N

    fi(yst)= fPAM (u) = wi(u)

    1+ eno)

    (7.6.26)

    Here w(u) is the normal density of the p-dimensional random vector which is chosen in correspondence with the requirement cp; = 0. The optimal coefficients of the expansion cp, are determined by relation (7.6.3) which in our case will take the form

    =

    / RENCTA: Pen CD)

    eae Oe 0 aan 7

    Re MPa 18 TA elO°)

    O°)

    The set of the polynomials {pp ,(u), gp,,(u)} is constructed on the base of the orthogonal set of the polynomials {5, ,(u)} (Subsection 7.6.1) according to the following rule which provides the biorthonormality of

    system (7.6.1) at p> 2:

    Poult) =Spult)s toolt)=7 EN Je Se nom rG os) where the polynomial S,,(u) is given by formula (7.6.11).

    Thus the

    solution of the problem of finding the one-dimensional probability density by EAM is reduced to finding the expectation m and the covariance matrix K of the state vector of normal system (5.4.6) and the coefficients of the correspondent expansion Cp, also. > For deriving the equations for the expectation

    covariance matrix K(t) = we use Ité formula

    (5.1.15).

    m(t)



    EY (t) and the

    E(Y (t) a m(t))(Y (t) _ m(t))? for normal system As a result after the evaluation

    of the expectations

    accounting (7.6.26) we get the equations N

    m = pio(m, K,t) + > cpvyir(m, K,t),

    (7.6.29)

    v=2

    N K=

    yoo(m, se, t) ai y Cp,v Pav(m, ive), v=2

    (7.6.30)

    7.6. Structural Parametrization of Distributions.

    EAM

    745

    where the following indications are introduced

    yio(m, K,t) = J eta,then(u)dy,

    (7.6.31)

    Pin(miK3t)is J e.tpp (wer u)dy,

    (7.6.32)

    20(m, K,t) = [lawn —m")+(y—m)a(y,t)? +o(y, t)}wi(u)dy,

    ve

    (7.6.33)

    pov(m, K,t) = / [a(y, t)(y7 ~

    7a) +(y- m)a(y, t)”

    +o(y,t) |pp,y(u)wi(u)dy,

    (7.6.34)

    a(y,t) = b(y, t)vo(t)b(y,t)*.
    epytev(m, K,t),

    «=2,...,N,

    (7.6.36)

    (7)

    where

    beo(m, K,t) = J lp ew- m)" K~'a(y,t) + tr K~*o(y,t)) —0oo

    746

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    +2qh(u)(y — m)* K~*o(y,t)(y — m),

    (7.6.37)

    Pru (m, K,t) = / [4p ,(u)(2(y — m)? K~*a(y,t)+ tr K~*o(y,t)] +2q/(u)(y — m)? K-o(y,t)K—!(y — m).

    (7.6.38)

    Eqs. (7.6.29), (7.6.30), (7.6.36) at the initial conditions m(to) =m,

    K (to) as Ko,

    epie(ta) =

    ee

    (kK = De SAI N)

    determine m, K, cCp2,.-.,¢p,n as time functions.

    (7.6.39)

    For finding the vari-

    ables cp ,, the density of the initial value Yo of the state vector of a system should be approximated by formula (7.6.26). > In fact in Ité formula =

    (5.1.15) for the nonlinear

    function

    (Y(t),t)

    dp,k(U) the items which enter in it sequentially have the following form.

    first item is

    Ogp t(U gl¥,8) = EEO OSoU

    The

    (7.6.40)

    where

    OU ork i —m’*)C(Y T\C ian —2(Y° rr —m' bY )Cm+(Y* —m) If we differentiate the relation KC

    KC+KC=0,

    =

    (7.6.41)

    I and find from it Gi

    C=-K7!KC=-CKC,

    (7.6.42)

    then (7.6.40) gives

    O4p,.(U) = Ion U)(-2(¥ —m)" Crn—(Y—m)?CKC(Y ey —m)). (7.6.43) The expression Py (% t) in the second item of the Ité formula with the account of the value of the derivative OU/OY

    = 2C(Y = m) will take the form

    yy (Y,t) = 2¢, .(U)C(Y — m).

    (7.6.44)

    Therefore the second item will be completely equal to

    gy(Y,t)"a= 2q/, .(U)(Y — m)? Ca(Y,t).

    (7.6.45)

    7.6. Structural Parametrization of Distributions.

    EAM

    TAT

    “~~

    For calculating the third item we shall find at first the expression yy

    ¥; t):

    os(¥t) =On,OPH (Yet)LIE _~ 20g, 20a (U)\(¥ (UIP ==m)FC oY

    oY

    OY — m)? (7p OU = 205nw nV) ae(¥ — m)"C + Tint ie ores

    (7.6.46)

    = 495 (U)C(Y — m)(¥ — m)?C + 2Gp,«(UJC.

    The third item with the account of the indication of (7.6.35) will be written as

    stl euu(¥, thevoe™]= See{4g (U)C(Y = m)(¥ = my? 1

    au

    1

    Wf

    +24p,.(U) ]o(Y,t)}.

    T

    (7.6.47)

    Finally after denoting 6 = b(y, t) we have

    gy (Y,t)"bdW = 21, .(U)(Y — m)"CbdW.

    (7.6.48)

    At the following stage we shall find the expectations of the obtained items. For the first item of (7.6.43) we have

    Bq, (UY(V? =m

    C= gy(u(y? — mB CAPA (w)ay,

    We evaluate the expectation in this formula. While introducing a new vector variable

    oi CHAa(y — mM) and noticing that u = (y? = mm? )C'l4zCR2(y —m)=

    \y|?

    we obtain [e.e)

    Eq, .(U)(¥7 —m?)C= faPyk (ll?) ¥701" FFA(Ig)?)|C|!/7dy =0, —0o

    (7.6.49)

    as the integrand is odd relatively to each of the coordinates of the vector y and the integral with respect to odd function in symmetrical limits is always equal to zero.

    For evaluating the expectation in the second item of (7.6.43) we use formula

    ET AE = tr[ AEE? ] which is valid for any vectors € and the quadratic matrices A. We obtain

    (Y? —m™)CKC(Y — m) =tr[KC(Y —m)(Y—m)'C.

    (7.6.50)

    748

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Thus it is necessary to find an expectation

    Eq, .(U)C(Y — m)(Y - m)'C = Ew,q),.(U)C(Y — m)(Y — m)'C N

    +>) cp,» Eur %,n(U)Pp»(U)C(Y — m)(Y — m)*C. yv=2

    Here and further on Ew, means an expectation at the density wi(u). It is evident

    that the finding of this expectation is reduced to the evaluation of the expectations of the form Pou

    Ga. = m)(Y = m)? CO} as Ip, (U) and Pp,v (U) are

    the polynomials of the variable u. = Cag

    BylU

    After performing the change of the variables y

    = m) = C!/2z in the integral which expresses F,,, we may write

    CRXP QS ES UECLLY

    CU NaC VEGA UT YY

    At first we calculate the diagonal elements of the matrix Bi. UFYY7. bution of the random vector =.=

    eer

    are identical.

    |i.e.

    all the diagonal elements of the matrix

    Therefore we may put Beg {U FY 4} =

    2 and taking into account that

    = hi U

    As the distri-

    Y = C’ 1/2X is spherically symmetric then Ew, U Ry?

    U =

    bLal

    & ps 2

    3S ve =

    YTY

    qa.

    Bey

    ve

    After summarizing over 5

    we have Ea Ary

    ae

    yy

    t=1

    = pa.

    As far as the nondiagonal elements of this matrix concerned to be zero for the same reason that in formula

    (7.6.49).

    they turn out

    As a result we obtain

    Ci? Ba hV Ys Chex: (Bil!ie/p)G. Gonaeruently, (EU "OX X* GC = (Ew, U**t1/p)C. Thus 1 Eq,,.(U)CXX7C N

    +0 v=2

    Cc

    Pv

    =

    pom {4p,.(U)UC}

    Ba {ah «(U) Pp,» (U)UC}.

    (7.6.51)

    B

    For calculating the expectations in (7.6.51) it is sufficient to know the probability the density w(u) of the variable U correspondent to the normal probability density W 1 (u):

    w(u)

    she

    1

    DTT (p/ay" p/2-1,—-u/2 €

    (7.6.52)

    Now we find 1

    co

    Eq,qo xl(U)CXX7C = seroinn ) 2P/2T (p/2)p |pn 0

    (u)u

    pide igHi?

    du

    7.6. Structural Parametrization of Distributions.

    749

    co

    N

    +

    EAM

    eis1 hdDp Deo» oot FTG

    |eh

    v=

    0

    Pre( uu p/2,-u/2 e du>C..

    /

    (7.6.53)

    E

    For the sake of the convenience we introduce the denotions

    Upeetn a dantoararrrs wg ule de, ? 0 1

    [oe

    = Acwereeta

    a ae 9

    9p/27

    eo

    (p/2)p

    —u/2 tte PeWop,kcolth)Bp. P, v( Seal dens

    (7.6.55)

    J

    /

    acs that in the consequence of the orthogonality of Po, v(u) to all the functions u> at r > V the variable Y,, vanishes at Y > K. Therefore

    kK Eq,,.(U)CXXTC

    =

    |.+

    Ss ote

    G

    YEP

    We evaluate the variables ¥,9 and Y,,, determined by formulae (7.6.54) and (7.6.55). At first we find y,,,. But virtue of the orthogonality of u to the polynomial

    Pp,n(u) at AX< kK Yen =

    1 TR

    (p+2K—2)!"(2K)! Ga

    fe—a/9 Don ujur'e du.

    0 (7.6.56) For the same reason formula for Y,,

    may

    be written in accordance with formula

    (7.6.3) and Appendix 7 in the form

    ck

    be TE FOR al EN IES

    “edu

    = —,

    (7.6.57) Now we evaluate 7x, for arbitrary

    Y < K. While integrating by parts we obtain

    il

    Yer = P70 (p/2) /In (¥)Pp,v(u)ur?e“/?du 0

    1 S fae mat SOPi,

    p/2,,—u/2}0o

    750

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Ef tentU)Pp,v (u yur teel2du— fdyn Pp y(wurl2em "du 0

    0

    (7.6.58)

    Pp (wale?

    +5 |dm 0

    It is evident that the first item is equal to zero. The first integral in (7.6.58) at V < K vanishes by virtue of the condition of the biorthonormality (7.6.8). The equality to zero of the second integral is shown by means of the same way as in (7.6.57).

    Analogously the third integral is equal to zero at

    vy Yo...2...0 + i=l

    +

    ."

    r

    3:

    a

    i=1 j=i41

    ye

    eS)

    GoW 710 ¢:.610.5:50 t

    3

    hae ere

    )

    (7.6.117)

    where Vege

    =

    BX

    1Ne

    ? mits Sh

    =

    EX}? X? /4)j.

    (7.6.118)

    For the second item in (7.6.110) we have

    Ew,Pi,2(Ui)X} = 2? - 2'Ry 2= 8Ra2 = 24D?, as

    (7.6.119)

    Ra2= DY) Chupad_, = Cimgag) = wy = 3Di, Jh|=4

    where p’ and a{_, are the central and the initial moments of the normal distribution N(0, K’) respectively.

    7.6. Structural Paramefrization of Distributions.

    EAM

    765

    Thus the ellipsoidal approximation of the r-dimensional distribution gives for the fourth moment of the first and the i* coordinates

    EpaX? = 3D? + 24c,2D?, EpaX? =3D?+424c,9D?.

    —(7.6.120)

    By virtue of (7.6.103) and (7.6.104) we have

    Oia hee

    ee pak a Bk

    (7.6.121)

    S34" ae, 0)ee 1(4,0)-..50) =6:(4,0,...,0)

    7.6.122 (7.6.122)

    The fourth order moments of the vector may be also given by the

    expressions of the type EX?X?, EX; X3, EX;X;X,X.

    We evaluate

    BX? X he The remained two types of the moments are determined analogously. After choosing, for example, 7 = 1 and j = 2, we write



    j[tif@da

    N

    Sel v=2

    ies

    a, = / oe | situ (wdin a,

    ae [Fee

    wuluddés «die.

    (7.6.123)

    ES

    Let us integrate over £3,...,Z, using (7.6.23). As a result we obtain co

    Ne | #BF@dé. a= _



    v=2

    | | attiu(us)aes diy —o6O

    Dew f lc ey

    oC

    =—©6cOo

    Uz) w(uz)dz; dro,

    (7.6.124)

    eS

    where ug = £7/\; + 23/2, w(u2) is two-dimensional normal density. We rewrite (7.6.124) in the form of the expectations and substitute the values of the normal moments

    PraXixo= Hoey SMEn X? X2 po, (U2). v=2

    (7.6.125)

    766

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    We evaluate the expectation in (7.6.125):

    Bw, P2,v(U2)X2X3 = Ewp2,2(U2)X7XF

    = 2? .2!Roo2 = 8Re2,2 = 8DiD2

    (7.6.126)

    by virtue of the fact that at y = 2 it turns out that Fuso

    Uz) Kade Ean

    and, ath: < 7; ok

    R2,2,2 =

    »

    hi

    ah2,,w

    w

    Cy° Cy nip Din ale)

    SND

    Tay, Wee

    tee

    C203 19 2%.0 ee D, Do.

    |h|=4

    According to (7.6.126) the ellipsoidal approximation (7.6.125) of the mixed fourth order moment will take the form

    EX? X2\=.D Do + 8c; 9W1 Do.

    (7.6.127)

    Therefore finally we get:

    6: (Qr2h

    One 9 Oiceeb de

    Mot BAX,

    (7.6.128)

    6(2,2,0,...,0) = 61(2,2,0,...,0)/BX2X2.

    (7.6.129)

    Example 7.6.1. Under conditions of Example 7.5.5 Eqs. (7.6.30), (7.6.36) have the form:

    Mm, =—m

    m2 —k12,

    mz = —amo;

    (7.6.29),

    (I)

    ki = —2(moki1 +m, k12), kia = —2(mz + a)ky2 — myko2,

    koo = —2akoo + hv; €9,2 = 4h?vC29 + €2,2{(m + a)(6 + 8Ci2k12) + 8m

    (II) Cy2k12 — 3h?2vCo9},

    (II)

    62,3 = —24h7?vC2243c2,2{(m2+a)(Ci2k12—5) +m Cy2k22 +4, 5h2v C29} +3c2,3{6(ms + a)Cyok12 + 6m, Ciek12 — 3h?vC22},

    (IV)

    €2,4 = 192h?vCo2+24h7?VO 29,2 +4c0,3{(m2+a)(Ci2k12—7) +m Cy 2k22 +6, 5h?vC29} +2c¢2,4{(me +a)(14+ 16Cy2ki2)+ 16m, Ci2k22 —Th?vCo},

    (V)

    7.6. Structural Parametrization of Distributions.

    C25 = =1920h2Cy5

    EAM

    767

    _ 240h?vC2gc,2 + 40h?vCo2¢9,3

    +5c2,4{(m2 + a)(Ci2k12 — 9) + mi Ciako2 — 8, 5h?vC29} +5c2 5{(m2 + a)(9 + Ci2k12) + 10m, Cy2ko2 —- 4, 5h2vC29}.

    (VI)

    Here OF are the elements of the matrix C; C = K ee, The peculiarity of these equations is the fact that Eq.

    (I) for the expectation

    and Eq. (II) for the covariance matrix are separated from Eqs. (III)—(VI) for the coefficients of the expansion. Here Eq. (I), (II) coincide with the correspondent equations

    of NAM

    (Example 7.2.3). If we wish to restrict our consideration to the moments

    till the fourth order we may put C2.3 = C2,4 = C25 = 0 in the obtained equations. By the same procedure the equations-in the case of accounting the moments till the

    sixth or the eighth order are obtained.

    On Figs. 7.6.1—7.6.5 the accuracy of EAM while calculating the intial moments Qc

    Sm,

    LDs=

    ki1, @2, @4, &g, Ag in accordance with time ¢ at the interval

    [0, 1]: a = 5, A = 1 and at the initial conditions: m,(0) = 0,5, m2(0) = 0,5, k11(0)

    =

    D,(0)

    =

    ad

    k22(0)

    =

    D2(0)

    at

    Re

    ky2(0)

    =

    0 is shown

    The accuracy of @4, @g, Q@g ranges about — 2%, 8%, 20% relative to the exact solution respectively. Hence we see that EAM gives a good accuracy similar as MSM

    (Subsection 7.4.7). In this case EAM with the account of the moments will the 10th order gives 9 equations instend of 65 ones at MSM.

    0,485

    a=m,

    0,481 0,477

    : + ES; EAMio; MSMy2

    0,473 4



    0,469 + 0,465 0,1

    0,3

    0,5

    Fig. 7.6.1

    0,7

    t 0,9

    Fig. 7.6.2

    .

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    768

    re 0,26

    4

    0,25 +

    EAM\o i

    ES

    0,24

    EAM,

    0,23

    0,22

    MSM

    0,21

    =

    0,20 +

    0,1

    t

    0,3

    0,5

    0,7

    0,9

    Fig. 7.6.3

    1,0 Og

    0,9 0,8

    0,7

    EAM,

    iF

    ES

    0.6

    BS

    EAM, ~|

    0,4

    MSM,,

    0,3 0,2

    +

    1

    2

    3

    4

    t

    5

    Fig. 7.6.5

    The method

    of the ellipsoidal approximation

    in the case of Eqs.

    (5.4.6) with the nongaussian white noise and (5.4.7) is given in (Pugachev and Sinitsyn 2000). 7.7. Methods based on Canonical Representations 7.7.1. Linearization by Canonical Expansions

    The linearization of the operator equations from the point of view

    of general theory of the stochastic systems transformation (Section 6.5) may be used in two variants. Firstly, a given dependence between random functions may be directly linearized and by such a way the nonlinear equations connecting the random functions may be substituted by linear equations. Secondly, the canonical expansions method which leads

    7.7. Methods based onmCanonical Representations

    769

    to the replacement of the operations over the random functions by the operations over the ordinary random variables may be used. After that we may use an ordinary in the applications method of linearization of the functional dependences between random variables. The method of direct linearization of the nonlinear transformation of random functions consists in the replacement of all given equations connecting the random functions by approximate linear equations which reflect adequately the real dependence between the random functions in the domain practically possible realizations of the random functions. As the expectations of the random variables are the mean values near which their possible realizations are considered then in practice it is more convenient to perform the linearization of the relations between the random functions relatively their deviations from the expectations, i.e. the centered random functions. In this case all the functions which enter into the given equations should be expanded into the Taylor series over the centered random functions and rejected the terms of these series higher than the first degree. The degree of the accuracy of the approximation obtained in such a way may be estimated by maximal possible value of the rejected terms in the region of practically possible realizations of random functions. After substituting the given equations connecting the random functions by the approximate linear equations we may use the methods of Section 6.3 for the approximate determination of the expectations and the covariance functions found as a result of the considered nonlinear transformation. Let us pass to the application of the canonical expansions method to the approximate analysis of the nonlinear transformations of the random functions. Let the scalar random function Y(t) be obtained as a result of the transformation of the scalar random functions X(t) by means of the nonlinear operator A:

    Y(t) = AX(t).

    (ain

    Substituting here instead of the random function X(t) any its canonical expansion we find

    Y(t)=A {meloSy veo} =wia(VanVoetent)ei

    anuntl:T2)

    Equality (7.7.2) represents the random function Y(t) as some function, in general case, nonlinear of the random variables V, into which the argument t enters as a parameter. While linearizing this function and

    770

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    taking into consideration that EV, = 0, v = 1,2,... we obtain the following approximate expansion of the function Y(t) and its covariance function K,(t, t’):

    Y(t) © Ame(t)+ S>Wwi(t), Ky(t.t!) © >> Diw(t)y(t).

    (7.7.3)

    Here

    w(g= |ee

    sora {mets Dvseuto}] , Lowy

    plafinans7vtid)

    where zero below indicates that after the calculation of the derivative all the random variables should be equated to zero, V, = 0. We apply the stated arguments to the nonlinear differential system

    RS

    Cs

    eR

    a

    cetera OS

    aaa

    ere,

    (Talo)

    Here f; are the known functions and the random disturbances = X,(t,Y1,.-., Ym) allow the joint canonical expansion

    Nprmamich yy Vota (3 adgewl mn),

    X,

    (7.7.6)

    V

    where Wi. = (EY es LR lye eye brl is seek mi Vy ale one uncorrelated random coefficients with zero expectations and the vari-

    ances D,. If we substitute (7.7.6) into (7.7.5) then the right-hand sides of Eqs. (7.7.5) will depend on the random coefficients V,. Here they will be known functions of the variables t, Y;,..., Ym and V,. As a result

    of the integration of these equations we obtain Y; = yz (t, Vi, Vo,.-..). If we consider vy, as the functions of V, sufficiently close to the linear ones

    in the region of practically possible values of V, then we obtain in the linear approximation

    ¥, *[%]+

    OY;

    5 VY |a | v

    (7.7.7)

    an!

    where by zero below is indicated that after the calculation all the derivatives should be taken at V, = 0.

    Hence using the formulae for the canonical expansions of the vector random functions (Subsection 3.9.3) we find the approximate canonical expansion Y; and the covariance functions: Yh y

    my +SoVWwe,

    v

    7.7. Methods based onCanonical Representations

    KY, (t,t')

    real

    2, Dodon|(t)yq(t) (p,q,k =1,...,m).

    (7.7.8)

    where mi © [Ye Jo, we(t) = [OY /OV, ]). From (7.7.5) after the differentiation m{, yy, over time and accounting that

    & (SE) = 5 ShMe

    : Of Oe yo

    OV, )]~ 4+ OY, OV, OX gis

    ". Ims AY,

    ay, + Loy

    p=1

    2

    we find

    -

    + OX, OV,

    . ., OLyy OY,

    ay, t Lay ay,” Bh p=l

    4

    my = = Ye lo =[ filo »

    (7.7.9)

    : ahproyy Of, Om= imt)=5 || = Ofk oa a“s Ome ehale dt |OV, J, »» omy Oms OY,

    Ofr

    4

    Consequently,

    =]

    q

    (7.7.9) and (7.7.10) lead to the following ordinary

    differential equations for m% and yx:

    mieafe

    miami

    y!| ieqimP) |

    Wk = > Axp(t)wp + > beq(t)evq(t), p=1

    (Fe FNY)

    (7.7.12)

    q=1

    where

    im? = ni (& MiG. ard (3

    =



    ((k,p=1 »P

    ~

    me beg(t =

    “. —— Of, Om — Om= Om

    ame

    (keel

    Cy At = ipa (emt, Oe

    2

    ae

    Le

    eG

    o, mi)

    yin

    )

    matt);

    (797.13)

    Sole, )

    qu)

    el aas

    12)

    Thus after presenting the vector random function by canonical

    expansion

    (7.7.6) we get a system

    X =[X1...Xn es

    of ordinary nonlinear

    772

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    differential Eqs. (7.7.11) for an approrimate determination of the ezpectation of the random function Y = [Y1..-Ym i* and a system of the linear ordinary differential equations with the one and the same matriz at unknown set of Eqs. (7.7.12), for determination of the coordinate functions. After integrating Eqs. (7.7.11) and (7.7.12) we shall find by formulae (7.7.8) the approzimate canonical expansion Y and its covart-

    ance function K¥(t,t’). It remains to solve the problem about the initial conditions for Eqs. (7.7.11) and (7.7.12). If the initial conditions are nonrandom, i.e. t = to, Yx = yro (kK =1,..., m) then it is necessary to integrate the equations at the initial conditions t = to, my = yko, Yk = 0. For the random initial

    conditions t= to, Y¥r= ype (k = es ay eas. (11) atid (i722) are integrated at the initial conditions t = to, mZ = my, wr = yro- In the latter case mo

    and y,x0 are found from the joint canonical expansion

    of (n + m)-dimensional random function [ X7 YJ lee Analogously Eqs. (7.7.1) and (7.7.12) are written for the equations unsolved relatively to the derivatives, F'(Y, Y, TaD CEG: HAE ot: payin). bE) 02 Na Ya (a el ore: Example

    7.7.1. Consider the scalar nonlinear system

    Vos where

    X =

    oe

    X (t) is a stationary random

    me (Oi ts

    (I)

    process with zero expectation

    and the

    known spectral density Sz (w). We use a segment of the canonical expansion of the

    stationary random function X (t) at the interval of the duration 27’ (Example 4.8.5): N

    Me

    We PaO

    ame

    (II)

    v=—N Wy+ta

    Here the variances V, are equal to Div, ] =) DF

    =

    f

    Sz (w)dw, Tea

    Wits ot

    Wy-a +2,..., where 2q@ is the length of each of (2N + 1) segments of the interval of the frequences (-Q, Q). In a given case Eq. (7.7.11) has the form My =a _ mi, My (0) = 1. It allows the integral my (t) = 1. Eq. (7.7.12) approximately determining the coordinate functions py = Yy,(t,w,) with account My (t) = 1 has the form yy = —2y, + et

    v(t,uv)

    U5 (tape,

    Its integral vanishing at t = 0 is equal t

    2t

    Je2r+iwvt dr = 0

    =

    Bee

    ae! ——____

    Ill (1m

    7.7. Methods based oneCanonical Representations

    773

    Consequently,

    N

    Jeivet 43. gat?

    Do

    Dy(t)= 9 Dy vr=—N

    TS

    Example disturbance X

    1 = 2e-2" cost

    =bien syioee

    =

    7.7.2.

    y=)

    + e7*

    (IV)

    Solve Example 7.7.1 supposing that the input random

    X(t, Y) is a real stationary random function with the spectral

    density Sz (w, A). It is evident that similarly as in Example 7.7.1, My (t) = 1. For determining the variances of the random function Y we use a segment of its canonical expansion in the region 0) < t < 27’, lY |

    we get Vive

    Differentiating Eq. (II) over

    —2Yy

    v2

    a

    Wy Ye

    .

    (IIT)

    V,, and putting after this V, = 0 we get

    ; Yyvev3

    =

    1 =

    —2Y,vevs

    i

    3 (Ys YWv3 FP W2Yrivs

    i YWsYvive) o

    As by the condition Y (0) = | then the integrals of Eqs. vanish, grip, (O) =a) Ae

    (IV)

    (III) and (IV) at t = 0

    veil) = 0. After substituting expression (III) of Exam-

    ple 7.7.1 for Y, into (II) and integrating this equation we find Yi V2 (t). Analogously

    substituting Y, and Yy,p, (t) into (IV) we find Yy, v.13 (t). By virtue of the fact that the random function X (t) is normally distributed all V, are normally distributed, and consequently, V, are not only uncorrelated but

    are independent as well.

    Therefore all their third order moments

    are equal to zero

    7.7. Methods based gn Canonical Representations

    and between indexes

    the fourth order moments

    coincide pairwise:

    == D,, Ds,

    [lyyyy

    =

    783

    only those are different from zero EV?



    a.

    Myzrivev2

    =

    EV?

    whose EV}

    Consequently, after calculating with the stipulated accuracy finally we

    get

    my(t)

    =1+4+

    »

    Dy yr

    v=—oo0

    =e 2

    op Vy=—0CO

    (t),

    Dy (

    Mey

    D,y3(t)

    v=—oco

    Se

    (V)

    Dy, D,.. [ue, (t) + 3Yv, (Peer)

    5

    V2=>—-00

    Expressing the random function Y(s) in terms of V, we have the principal possibility of finding the n-dimensional distributions of the random function at any n. In the general case the determination of the distributions of the random functions by means of the canonical expansions is connected with great calculations difficulties. But for some classes of the nonlinear stochastic systems the canonical expansions method may be sucessfully used together with the distributions parametrization methods stated in Sections 7.2—7.6. Hence the essential advantages connected with the calculation of the integrals with respect to the nonlinear functions in the equations for the parameters of the nongaussian distributions V, appear. Example form

    7.7.6. Let us consider the nonlinear differential system of the

    F(t, DY =9(V) +X), XH=mOH+

    U2),

    where F(t, D) is the polynomial relatively to the differentiation operator

    D = d/dt

    with the coefficients dependent on time t; p = ~(Y) is the nonlinear, for instance, the discontinuous function, and (0) =,(); X(t) is the normally distributed random

    function. Performing by the formulae of Subsection 7.2.2 the statistical linearization of the nonlinear function ( we substitute initial Eqs. (I) by linear one:

    F(t, D)Y = ko(my, Dy)my + ki(my, Dy)(Y — my) + X(t).

    (II)

    The functions My (t) and yy (t) which determine the canonical expansion Y(t) and Ky tt, t’) in the form

    Y=m,(t)+ > W(t), Ky(t,t') = >) Diw(t)y(t),

    (IIT)

    784

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    are defined by the following equations, see (II):

    F(t, D)my = ko(tmy, Dy)my + Mz ; FU, D)y, = bil, Dy)tp 42

    (IV)

    ;

    where Dy = iS Dy lyv (ol The joint integration of nonlinear ordinary Eqs. (IV)

    vy allows to find My (t), Yy (t) and Ky (t, ne In the case when the function X(t) has the nongaussian distribution Eqs. (IV) will have the following form:

    FG, D)my = ko, ,D™))my + Mz ;

    ex

    EG

    D)yy

    =

    kiy(™m,, 1B

    (V)

    wy + By,

    where the coefficients ko(my, D{*)) and kj, (my j D&)) will depend on the random variables Viu- Here the coefficients kj, will be different for various V,.

    The canoncial expansions method forms the basis of: e the methods of equivalent disturbances (Dostupov 1970),

    e the methods of the stochastic functional series (Adamian 1983), e the statistical linearization and the normalization methods based on the canonical expansions (Sinitsyn 1974, Pugachev and Sinitsyn 1978), e the modifications of the methods of Sections 7.4-7.6 based on the canonical expansions. 7.8. Applications. Analysis of Stochastic Systems and Analytical Modeling 7.8.1. Accuracy Analysis of Single Loop Nonlinear Control Systems At first let us consider the one-dimensional stationary control system (Fig.7.8.1) consisting of the linear stationary system with the transfer function ®(s) and the nonlinear element whose input and output variables are connected by the single-valued nonlinear dependence

    U =

    g(Y). The input X we shall assume as the stationary

    random function with the constant expectation M, and the spectral density Sz, (w).

    Using the statistical linerization of the nonlinear element in the form

    US PY YS Coligny)

    bility, Dali) 2h

    Yo Aimy

    181)

    we shall obtain the following dependences between the expectations M, and My and the centred random functions

    U° =

    U — My

    and y°®:

    T= ko(Mgy Dy mee U =

    Day,

    (7.8.2)

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    785

    where

    “ko(my, Dy) = Yo(my, Dy)my" From the relations (Fig.7.8.1) My

    =

    ©(0)(mz

    — mw), Yo

    @(s)(X° = Oe)

    accounting (7.8.1), (7.8.2) we have the following statistically linearized operator equations: My



    y? = =

    (0 B(°)(0;

    p() G(s; :

    my, Dy)mz

    my,

    pe y)X-;

    D

    ,

    B{)(0; my, Dy) a

    o(1) y

    (8;} My, Dy)1)=

    Tae

    yrry

    oe) RieeD

    a

    oa

    (7.8.3)

    So from Eq. (7.8.3) we get the corresponding equation for the variance

    |__*) _

    7

    Dy)®(iw) e my, e 1+ ki( Seal L Solving numerically Eqs.

    (7.8.4)

    Sp (w)dw .

    (7.83) and (7.8.4) jointly we find the expectation My

    and the variance Dy Z

    Example

    7.8.1. Find the expectation and the variance of the output of

    the system which consists of the stationary oscillatory link, with the transfer function

    ®(s) =

    mies aS

    $ao), and the negative feedback across the relay element

    arranged in Fig.7.8.1. Assume that the input X is the stationary random process,

    Mi =~const,

    bk2(7T) = Dre

    lia

    Ds

    a

    Sz(w) = Sita

    (I)

    From Eqs. (7.8.3) and (7.8.4) we have Me

    thes ao + ko(my, Dy)’

    (i)

    =f T

    anes eets 2

    a

    ws. (a; +w?) lap —w? + ayiw + ky(my, Dy)|

    TT |)

    786

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    D,(a +41)

    lat hitms Dig

    Se

    We complete the solution of Eqs.

    aa+ a eee HEGRD NT

    otIll

    (I), (II) using Table A.3.1 at the following

    values for the parameters of the system and the random input:

    @p = 1, o? =0.2, a =1/a, l=1, eg = 1.

    (IV)

    Substituting values (IV) into Eq. (II) at

    Eq. (II) becomes

    A)

    rachis)

    (VI)

    It is clear that Eq. (VI) is equivalent to the following one:

    in,

    m

    = 1 —2h

    Z

    Vil

    From Eq. (III) at values (IV) and relations

    we get implicit equation for Dy

    2D, Dy =

    (IX)

    +? (ae) |22+ ae (Se) The results of the numerical solution of Eqs. (VI), (IX) are given in Table 7.8.1. Table 7.8.1.

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    787

    For the nonlinear control system illustrated on Fig.7.8.2 we have the following expressions of the transfer functions of the statistically linearized system:

    ko(m,, Dy) ®(s)

    CAL PR My, AS 5 (8,

    Nl Dy)ey Pakgtmg,

    ad D5) G(s) ’

    (7.8.5)

    Dy)®(s) 60) —ki (my, DAs)’ y (sm (s4ng, 4)DJ) = ie eke 1

    (sera

    De ee

    Mein,

    0

    (7.8.6)

    Eb)

    CHS By

    1+ ko(m,, Dy) ®(s) ’

    (78.2)

    :

    (7.8.8)

    1+ ki(my, Dy) ®(s) ©

    Consequently,

    ko(my, Dy )®(0) eg ~~ a1+ko(m,, La TY Dy)NL (0) a

    a.ka (at,

    De

    | ee aon,

    -

    1

    My kale, D. (0) a (he

    (iw)

    2

    D.) |Sz (w) TEAGa, Deo Gn) dw,

    _ ff TERR __ se(w)dw) Dy = AOATES Solving numerically Eqs.

    (7.8.10)

    (7.8.9) and (7.8.10) jointly, we find at first mM, and D,,

    and further My and Dy ‘

    Example

    7.8.2. Find the accuracy characteristics of the control system

    (Fig.7.8.2) consisting of the linear drive with the transfer function (s) =e a/s(T's + 1); the device for comparing

    the input and the output,

    and the relay element

    emitting drive control signals of the constant absolute magnitude.

    Analogously to

    the previous example we have at first

    Oe PES y

    and My

    = Mz.

    T?2s?+s+ koa’

    Then (1) Ses

    ®;

    s(Ts + 1)

    T?s2+5+ kya’

    (I)

    kya fe = elem cil i toa ® T?7s24+s+ ka’

    am)

    788

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    ne

    aD,k?a?

    =e

    us

    —00

    r ues

    aD,

    ple

    /

    dw

    (a? + w?) |kya — Tw? + iw]?

    Drkya(1 + aT)

    :

    IV

    ae

    a(1+aT)+ kya’

    /

    w?(T2w? + 1)dw

    (a? + w?) |kya ~ Tw? + iw|?

    "a(l+ aT)+hia 1+

    aT

    ©

    + k,aT

    Here ko and ky are defined by formulae (V), (VIII) of Example 7.8.1. The results of

    numerical solution of Eqs. (IV), (V) at! = T =

    1, a = 0.5, a = 5 are given in

    Table 7.8.2. Table 7.8.2.

    Table 7.8.2 demonstrates

    the “choking”

    effect of the relay element by large

    fluctuations.

    7.8.2. Solid Dynamics in Stochastic Media Let us consider the dynamics of the solid with one fixed point (Example 1.1.5) in the stochastic media using the following models of the stochastic media:



    SoA, = —Gol ewe + V2, My = —Golywy + Vy,

    =

    I

    —polzwz

    (7.8.11)

    SIF ee

    ys Vy , V; being the normal white noises with the equal intensities, vi = = Vv, =

    V0, Yo being the viscous coefficient;

    Vy

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling



    Mz = —Yolewe + Vy", My = —Polywy + Vy’, Mz =—o0I1,w,

    789

    (7.8.12)

    + ve

    ee Ve Vi being the normal white noises with intensities satisfying the conditions: 1 ==

    Yoles VyLi= Voly, VY,ti =ro Volz;

    =

    Mz = —$ilgwe + Vz",

    My = —Y2lywy + Vy", M, = —p3l,w,

    ie

    (7.8.13)

    ar Vore

    Une yes being the independent normal white noises with the intensities y///’ vw?

    ae Fya $1, ~2, 3

    being the viscous coefficients.

    In case 1° the Ité—Euler dynamical equations have the following form:

    IpWe + (Iz — Ty )wywz = —Yolewe + Vz,

    Tywy + Ie — Iz )w2w2 = —polywy + Vy,

    (7.8.14)

    T,0, + (ly — Ip )woWy = —Polezw, + Vy. Eqs. (7.8.14) have the stationary regime of fluctuations with the following onedimensional density:

    fi(Wr, Wy, wz) = cexp[—(Y0n/M)], where

    (7.8.15)

    n= I2ug + Iywy + 1207, ¢ = [po/(mv0)]” Lely le. More than that. In the nonstationary regime of fluctuations Eqs. (7.8.14) may

    be transformed into the following set of the independent linear equations

    dL; dt

    = —yoli+ V; (i = 1,2,3).

    (7.8.16)

    From Eq. (7.8.16) using the results of Subsection 6.2.3 we get the explicit formula for the multi-dimensional characteristic fucntions for the variables D; (2 saul, 2, 3); n

    AneGi

    gn(A1 a)

    tes

    y ee) =

    ft

    eee k=

    x exp

    ==>

    n

    n

    yD tA [tr —th| — e~ volts +t) pape

    hstt=1

    UG o1 0)

    790

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    where Jo (A) being the characteristic function of the initial conditions: L; (to). Hence

    the density of the variable 7) = L ae Le ap Rs > 0 and its first and second moments are

    fn(n) = (¥0/0)0(3/2)./n exp(—¥on/Vo),

    (7.8.18)

    (IC being the gamma-function, see Appendix 7),

    En = 3v9/(2¢0), In case 2° Eqs.

    Dn = 3v0/(4¢0).

    (7.8.19)

    (7.8.13) have the stationary regime of fluctuations with the

    following one-dimensinal density:

    2

    fi(We, Wy, wz) = cexp (0

    );

    (7.8.20)

    where

    3/2

    27 = Taw, Pilggbh iews, -¢

    Vy

    (#) TVo

    Using It6 formula (5.1.7) we get the first order nonlinear stochastic differential equation for the variable T,,

    T= 205) valet V2 1AV

    (7.8.21)

    Here V being the normmal white noise of the intensity 9.

    The equations for the

    multi-dimensional characteristic function g/,(A1,.-., Anjti,.--, tn) for T(t) defined by Eq. (7.8.21) are as follows:

    Te OgSS + Siv0Angha. 29m = An(=2p0 + iYoAn) At, /

    The solution of Eq. (7.8.22) forn =

    (7.8.22)

    1,2,... according to Subsection 6.2.2 at the

    intial conditions

    9;(A5to) = 9o(A); 9, (1 , te

    tNnit xd otere siniitnaizilnew)

    = In—101

    yo

    (7.8.23)

    An=2)An=1 me cere a echt

    tact)

    may be presented in the following explicit form:

    gi (Ast) egg (Ae 7b 00,t— 16) 9/2(\,t — 0), y (7.8.24) In(A1;

    SEC

    Anti:

    Cg OHR

    rin)

    =

    GreTtnr

    sey

    NG,

    Nees

    +

    Ane

    Colin

    —tn—1)

    7.8. Applications.

    Analysis of Stechastic Systems and Analytical Modeling

    =A

    =

    bit Ongtwert ten) iMesietadWF (Anite where

    791

    tea) (n= 2, 3).&)j (7.8.25)

    (A, 7) = 1— idvo(1 — €7 207) /(2H0).

    (7.8.26)

    In special case go(A) = (1 = ir6)~3/? from formula (7.8.24) we have

    LY) g1( 434) = fa — i

    + 4X (

    20

    ~ 5)ec

    t

    gi(A; co) =(1 — idAvp/2~)°!” :

    eelo.20)

    (7.8.28)

    Formula (7.8.28) corresponds to the x?-distrubution with three degree of freedom. From Eq. (7.8.21) we have the linear equations for the expectation i=

    BT:

    and the variance D’/ = DT’,

    dm! _

    a¥o|

    dD" |

    ;

    ;

    In case 3° we have the following It6—Euler dynamical equations We = Cj WyWz — PiWe + Vi, Wy = €2QWzWe — PoWy + Vo, WwW, = where

    EgWeWy

    LD Espn €1

    =

    iT

    9

    Ven

    ar

    — P3Wz

    te V3,

    @8 Aalewih 2

    ie

    ts sla1, ose

    Ved

    diate

    (7.8.30)



    ie

    )

    ie

    V3

    Ti

    From Eqs. (7.8.30) using the moments method (Subsectio 7.4.1) we get the infinite set of equations for the initial moments Aj,m; = Ewr WyPk

    Anmk = —(ngitmye+kys)anmk+Ne1 On—1,m41,k+1 +ME2An41,m—1,k+1 1 +ke30n41,.m41,k-1

    1

    a guin(n 1 1) 0, 23 mak ae gvzm(m 7

    +5usk(h ey

    L)Gaee.k

    (7.8.31)

    792

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    With the accuracy up to the moments of the fourth order we get from Eq. (7.8.31):

    100 = %010 = @110 = %101 = %11 = 9, 200

    11 /(291), 020 = ¥2/(2"2), M002 = v3/(2¥3).

    Remark1.

    (7.8.32)

    The nonlinear stochatic differential equations for the solid in the

    stochastic media admit the exact normal stationary solution

    f(w) = cexp(—w7 Aw).

    (7.8.33)

    Here

    w = [wewyws]’,

    A=AT+ pl = Ivy, ¢ = (2n)-9/? |Ap”?, peau a0 leO le pie |

    LS

    v="

    Ve O 0.

    pve WO"

    aie

    0 COe 34,

    Say

    1 O10 KO gaa Odes O Os 1

    pr io Cte ONS O20 naGa

    where A and }L are some real coefficients.

    7.8.3. Stochastic Hereditary Systems Many problems of the theory ofhereditary systemms (for instance, the systems which contain the elements described by hereditary mechanics, physics, bilogy, chemistry) under the conditions of random disturbances lead to the study of the following

    stochastic integrodifferential system: t

    X =a(X,t)+ fax

    nber

    to t

    Ub EX:t)+ /bi(X(r),7,t)dr |W, X(to) = Xo,

    (7.8.34)

    to

    which is considered in the It6 sense. Here X € R? is the state vector; W € R14 is the vector white noise of the intensity

    Wis

    V = V (t) independent of the initial condition Xo;

    W(t) is the random second order process with the independent increments;

    a=a(X,t),a:

    Rix

    a; = a;(X(r),7,t),a, b=a(X,t),b:

    R= RP; : RPx Rx

    R—

    RP;

    R—

    RP?!

    Rix R— R?!;

    by = bi(X(r),7,t),b)

    : RPx Rx

    7.8. Applications,

    Analysis of Stechastic Systems and Analytical Modeling

    793

    In applications the functions @; and b; are represented in the form N

    a, = A(t,7)p(X(r),7), 61 = >> Ba(t,7)va(X((7),7),

    (7.8.35)

    h=1

    where hereditary kernels A(t, 7) and Bp (t, 7) are (p x p)- and (p X q)-martix functions, ~p (X(7), T) is the (q x q)-matrix function, y( X(T), c) is the p-dimensional

    vector fucntion. Let us suppose that the hereditary kernelsaes \(t,7) = {Aj; (t, Ty;

    oe

    Op and Bill)

    Bultst),

    0 = 1N,t

    = 1,p, 7 =

    14g satistly

    the conditions of the nonanticipativeness and the conditions of dying hereditary (or

    assymptotical stability) 7a

    A)

    (7.8.36)

    es)

    7)| dr < 00, i |Ai;(¢,

    / [Braj(t,7)| dr< co:

    (7.8.37)

    In the case where the hereditary kernels satisfy the conditions

    attor paar (OM Bir) SB

    OP CS ts

    (7.8.38)

    one say about the invariance of the stationary of the hereditary. The important class

    of hereditary kernels are singular kernels

    ,7) Bus (t,7) = BE; (OBu; (7). = AZMAG(r), Aj;(t

    .(7.8.39)

    The simplest exmaples of the functions satisfying conditions (7.8.36)—(7.8.39) are the

    functions exp(—a@ |€| )1(€), exp(—a |€| )[cosw€+~ysinw |£|]1(€), 1(€) is the unit step function. One of the most effective ways of obtaining the statistical dynamics equations

    for Eq. (7.8.34) is their reduction to the stochastic differential equations by means of the approximation of the real hereditary kernels by the functions satisfying some ordinary differential equations.

    Let us consider Eq. (7.8.34) under condition (7.8.35) t

    Few aye /rCeatieetararie to

    rare Arey /By(t,7)a(X(7),r)dr |W, X(to)= Xo~ (7.8140) h=1 7,

    794

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    We shall determine a class of admissible kernels by conditions (7.8.36), (7.8.37) and by the order and the form of linear ordinary equations to which the kernels as the functions of t and T must satisfy.

    Let the kernels Aj; it. T) and Brij (, T) in (7.8.40) satisfy (7.8.36), (7.8.37) and at fixed T be the solutions of linear differential equations

    Fe Ay)(i'r) OOS wien gS Typ, ne

    (7.8.41)

    S > Biri Brig(t, 7) = Wir O(t = B)ermdypreps deg Vil

    at a fixed t the kernels are determined by the formulae

    Ay; ( (tsT)= —

    SH?

    Al, (t,7),

    Us = dg &

    (7.8.42) Bnij((t, T) =

    Yo YES Blt

    Tce

    pal

    i Us j= ied,

    and the functions At, (i, rg and Bia; (t, T) entering into (7.8.42) at a fixedt are the solutions of the linear differential equations

    DoF A (t,r) = 5,16(t— Tr), lr =Tp,

    z

    (7.8.43)

    Sy

    hiePais(oT) = OO

    7), Fal,

    l= ig:

    esi

    Here Fy

    =

    Ft D), Bee

    =

    Bait D), Ong

    =

    GOpaleaO

    ens

    a iis (t, D) are some linear differential operators, the order of the operators H,. and Wp,

    denoted by m,

    the order of F,; and Vj,

    is denoted by n,n

    > m,

    the superscript t at the operator indicates that the operator acts on the function

    considered as the function of ¢ at a fixed T. As it is known

    from

    the linear differential systems

    theory

    (Section 1.2) the

    integral terms which enter into (7.8.40) t

    ¥i= fSAal,o(X(),Den, t= Tp, to

    P

    Ie

    a = [Xo Basn(t,r)dnes(X(0), Der,i=Typ, 7=1,4,

    7.8. Applications.

    Analysis of St6chastic Systems and Analytical Modeling

    795

    represent the solutions of the following equations:

    P P Sure = Pree,

    i ie

    _

    k=}

    SESS OU (=k 2

    pes Lae

    a

    r=A



    (7.8.45)

    atS Yodo Ware dna, 7 = Tp. (=e al

    of their kernels satisfy (7.8.41)—(7.8.43). It is clear that here the nonlinear functions ~, =

    p(X, t) and Wnei(X, t)

    must be differentiable not less thanjn times. It is evident that if we introduce the vectors Z’ and €! of the dimensions p(1

    =5 qN) x 1 and (p =F q?N) x 1 and the vectors U and € of dimensions pqN x 1 and q?N x 1 assuming

    , gayi...)

    ,e=[b?...05]" N

    .

    (7.8.46)

    Then the scalar Eqs. (7.8.45) may be written in the form

    F'Z!' = H'e'.

    (7.8.47)

    Here the matrix differential operators F’ ‘and H’' are

    F! = {Fi,(t,D)} =)

    a,D*, r,1=1,p(1+ qN), =

    (7.8.48)

    k=0

    H S4{HGG, Dyy= SeRD Srl hehe, k=0

    where WE

    = a,(t) and J, = Bx (t) are the (p(1 ap qN) x p(l AP qN)) and (p+ q’N) x (p+ q’N) matrix coefficients. Further we shall replace (7.8.47) by the set of equations in the Cauchy for according to the known rules of (Subsection 1.2.3): n

    Gf =Big tae’, k= Tal, Zl = — oop ta 12}! + ane’ f=1

    (7.8.49)

    796

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    Here Zh are the p(l + qN )-dimensional vectors, h = I,n, dk = Ik (t) being the

    p(l ap qN) x q?N matrix determined by the formulae k-1k—-h

    Gk = a? in badSe 5

    Cab son

    nana ,

    k=1,n—-1,

    1 p=—t0) (0)

    n-ln-h



    Oe

    c — pas a

    ont

    j

    (7.8.50)

    hk=0'1=0

    Here the dependence between Z’ and €’ will exist. So

    Z = Zi + gee’, 90 = Op Pn-

    (7.8.51)

    By virtue of (7.8.46) and (7.8.51), integral terms (7.8.44) take the form t

    Y= fae T)p(X(r),7)dr = a’(Zy’ + qoe’), to

    N Si

    NS N = Sif Balt, 7)va(X(r), 7)dr = (>: s)(Z1' + qe’),

    al

    h=1},

    pil

    (7.8.52) where the following notations are introduces:

    a’ = [150], 0) SeAy

    1

    [OLalle Aaah Or = Tom. foe Bgl.

    (7s)

    n

    and Ip is the (p x Pp) unit matrix.

    Consequently, Eq.

    (7.8.40) together with Eqs.

    (7.8.49), (7.8.52) for the extended state vector of the dimension np(1 + qN),

    Za(xXtZ]", zits |Z... 2a") is reduced to

    :

    Al

    ;

    Z = c(Z,t)+\(Z,t)W, Z(to) = Zo,

    where

    _ | (Xt) +a'[Z7/ + go(t)e'(X, t)]

    oat) = ee yee) AZ) Seas

    N

    & u,[Z1' + go(t)e"(X,t)] 0

    (7.8.54)

    7.8. Applications.

    Analysis of Sto¢hastic Systems and Analytical Modeling

    7197

    Zo = ([XT0]", y(t) =diag [ai(t)...qn(t)], ft) -=

    I 0

    0 I

    Bt Stes

    0 0

    Saye

    sae

    Rieke

    Keer

    0

    0

    veo

    a

    —azta;

    apes

    —a7lan-1

    —az lag

    »

    T=

    Ip(i4qn)-

    :

    7.8.55

    The matrices a’, bi, which enter into (7.8.51) are determined by (7.8.53); we = ~

    are determined by (7.8.50), (7.8.51); €’ is determined by (7.8.46). The dimensions of c(Z, t) and I(Z, t) are equal to np(2 = qN) and np(2 + qN) xX q. Eqs. (7.8.50) are linear in Z” and nonlinear in AD Let us consider now the class of kernels satisfying the conditions (7.8.36) and the condition of sationarity (7.8.38).

    As it is known from the theory of linear sta-

    tionary systems (Subsection 1.2.5) in this case it is sufficient to suppose instead of (7.8.37) that the Laplace transforms of hereditary kernels A(é ) and ie (€) are rational functions of the scalar variable S, i.e. admit the representation of the form

    | A@etae = F(s)-!H(s), | Balee-‘tag = ,(s)-! W(s). ; Here =

    ;

    the

    order

    of the

    matrix

    polynomials

    (7.8.56) H(s)

    =

    {H,i(s)}

    {Wnri(s)} is equal to Mm and the order of the polynomials F(s) =

    ®;,(s) = {Prri(s)} is equal ton,

    and

    W(s)

    { F,i(s)} and

    Nn> M. So Eq. (7.8.40) by means of extending

    the state vector up to dimension np(2 + qN) may be reduced to Eq. (7.8.54). In conclusion we consider the class of the hereditary kernels satisfying the con-

    ditions (7.8.36), (7.8.37) and (7.8.39). In such a case the order of the operators Fj and ®;,,; is equal to one, n =

    1, the order of the operators Ay; and Whej is equal

    to 0, m = 0. In this special case Eqs. (7.8.44), (7.8.45) take the following form:

    [AG nex),

    ar AY

    Ye

    ae rYl :

    (7.8.57)

    to Tt

    J Baal X). rar ee DGU

    |e

    gaping zl Ural, shlley

    )

    to

    Y=A-Y, Y(to)=0, Un = B, Un, Un(to) = 0.

    (7.8.58)

    Here At =

    {Az (t)}and Ske = {Bi, (} are (p X p)- and (p X q)-matrices,

    Meandsliz

    =

    Lage

    eek

    Une le are the p-dimensional vectors.

    As a result

    798

    Eq.

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    (7.8.40) is reduced to Eq.

    state vector

    =

    Z =

    (7.8.54) for the extended p(2 at qN )-dimensional

    ee yt Ut as if we

    introduce

    (pqN )-dimensional vector

    U

    [ee a Oe les the p(2 aP qN )-dimensional vector function c(Z, t), the (p(2

    +qN) x q)-dimensional matrix-function I(Z, t) and the initial p(2+qN )-dimensio-

    nal vector Zp assuming

    a(X,t)+ ATY

    c(Z, t) =

    A-Y B-U

    , (Z,t)=

    (X,t)+ BTU

    0 0

    B® = [Be...BE], Zod XE0iOo tue oS wi(F-82539) The functions c(Z,t) and [(Z,t) are linear in Y and U and nonlinear in X. We emphasize that in the case of singular kernels the differentiability of ( and Wn is not obligatory.

    So Eq.

    (7.8.40) by means

    of extending the state vector up to the

    p(2 + qN)-dimension may be reduced to (7.8.54) under condition (7.8.59). Example

    7.8.3. The scalar linear stochastic integrodifferential system t

    oa,

    + Ge iSOGOU ld OW

    SO Ie (1)

    ON (do, 41, A, b, @ are constant) is reduced to the following linear two-dimensional stochastic differential system:

    X = ap

    +a,X + (A/a)¥

    Example

    +bW, Y

    =a(X-Y), Xo=Yo=0.

    (Il)

    7.8.4. The scalar system with the parametric noise t

    X =ap+a,X + |b+ Bf PC-Ox(r)ar

    W, Xo=0,.

    (1)

    0

    (do, a1, b, B, B are constant) is reduced to the following two-dimensional stochastic differential with the parametric noise:

    X =ayp+a,;X +[b+(B/B)Y]W, Y = B(X -Y), Xo = Yo =0. (II) Example

    7.8.5. The scalar linear nonlinear stochastic system t

    SS Ga

    a

    feoeecxmyar +bW, Xo =0, 0

    (I)

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    799

    (ao, 41, @, 6 are constant) is reduced to the following nonlinear two-dimensional stochastic differential system:

    X =ap+a,X+(1/a)¥ +bW, Y =a[y(X)—Y],

    Xo = Yo =0. (II)

    7.8.4. Normalization of Risk and Profit Financial Transactions In stochastic finance as it is known (Shiryaev 1999) there are widely used continuous and discrete stochastic models defined by the stochastic differential and difference equations.

    A

    Normalization of such stochastic models by

    NAM gives the possibility to design

    the specialized algorithms for various problem of the financial analysis and modeling.

    The results of the statistical simulation given in (Artemiev et al 1996) for the following types of the stochastic differential equation

    ¥ =a0@)(Y"@ -—Y

    ] tea@xev

    (7.8.60)

    ¥1 = aoi(t) [¥i"(¢) — Yi] + Bor (t)¥y" Vi, Yo = cvoa(t) [¥3(t) — Yo] + Boa(t)¥y7¥s"*Va,

    (7.8.61)

    Y3 = ao3(t)(Y3 — Ya) + Bos(t)¥3"° Vs, Y4 = aoa(t)(Y3 — Ya)

    (a(t), Aoi (t), Boi (i), Y;* (t) being some known time functions) confirm practically good NAM

    accuracy (5-10%) at 71, 62, Y2, ¥3 < 1.

    7.8.5. Normalization of Queueing Systems An important class of the nonlinear stochastic systems is the class of the queueing systems. For normalization of the queueing systems it is expedient to decompose the queueing system into the typical subsystems and use the normalization method

    to each subsystem separately. Because of the completeness of the class of the normal systems relatively to the connections the normal model of the queueing system is obtained as the result. As it is known any queueing system may be considered as a system consisting

    of the following typical elements: the queueing channels, the waiting places in queue, the distributor D which determines

    the direction of the motion of every client in

    queue depending on the availability of the free channels or the places in queue with the account of the priority. The stream of the clients which is directed into the given

    800

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    channel or at the given place in queue serves as the input of each channel and each place in queue.

    The stream of clients served by the channel and the signal fixing

    the channel (is it free or occupied) are the outputs of the channel.

    The streams of

    the clients and the signals of the feedback from the channels and the places in queue carrying the information about the state of each channel and each place in queue serve as the inputs of the distributor. The streams of the clients directed into the channels and in queue and the stream of the clients who received the refusals in service will be the outputs of the distributor D. Example of scheme of the queueing system with n

    channels and without queue is illustrated on Fig.7.8.3.

    Fig.7.8.3

    Fig.7.8.4

    We show how we may normalize the service channel, main element of the queue-

    ing system (Fig.7.8.4).

    Let y(c) be a probability density of the queueing time TJ’.

    The input of a channel, i.e. the stream of the clients entering into the given channel, may be presented in the form of the sequence of unit impulses correspondent to the instants of the clients arrivals.

    It is also convenient to present the output stream of

    the clients in the form of the sequence of unit impulses which appear at the instants

    when the clients leave the channel.

    The other output of the channel which carries

    the information about its state may be presented as the binary signal equal to zero in the channel is free and is equal to 1 if the channel is occupied. At such agreement

    the channel will response on unit input impulse 6(t == ) which acts at the definite instant T by unit impulse é(t ==

    T) at the random instant (7 + T) on one input

    and by the signal equal to 1 inside of the random interval (rTee

    T) and equal to

    zero outside this interval on another output:

    a(t) = 6(t—7), Vi(t)=6t-—7-T),

    Yo(t) =1(t-7)1(T +7 -1).

    (7.8.62)

    Taking into consideration that the response of a linear system on unit impulse

    represents its weighting function we shall consider queueing channel as a linear system with the following random weighting functions:

    Gi(t,r)

    =6(t-7-T),

    Go(t,r)

    =1t-—7)1(T+7 -t).

    (7.8.63)

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    801

    Then its response on any determinate input z(t) will be expressed by formulae t

    t

    ¥i(t) = u)Si aT) e(n)det¥a(¢i=/1(T-+17—1)x(r)dr. (7.8.64) to

    to

    Now for the linearization of the regression it is sufficient to find the conditional

    expectations of the outputs Yj (t) and Y (t) at a given realization x(t) of the random signal X (t). As a result we obtain t

    E[Y,(t)|2£] = [e@ar

    6(t— +r — 0) y(c)do = [ve — 7)x(r)dr, aS)

    (7.8.65)

    E[Y2(t)|z] = /x(r)dr [1(o+7 —t)y(c)do = [ve — 7)a(r)dr,

    i

    oe) where w(t) = (p p(a)do 9

    :

    (7.8.66) F(t) is the probability of the fact that queueing

    t

    time 7’ will exceed t. Formulae (7.8.65) and (7.8.66) show that the queuing channel may be considered as a linear system with the weighting functions p(¢ ) and wH(¢ ) and with the additive noises on its outputs (Fig.7.8.5).

    Y,( Y(y

    Y (0

    yy

    Y,(0

    Y,()

    Fig. 7.8.5 At the exponential WC ) =

    e

    distribution

    #6 the channel may

    of queuing

    be considered

    time 7’ when

    p(¢)

    =

    peas ,

    as the parallel connection

    of two

    aperiodic links with one and the same time constant |y)pt and with the gains | and 1 fpt with the additive noises on the outputs.

    In case of determinate =

    queueing

    time 7' when

    p(¢)

    =

    6(¢ =

    Tey, ¥(¢)

    1(¢ )U(T —¢ ) the channel may be considered as the parallel connection of the

    802

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    ideal lagged link and the integrator with the finite memory

    T on whose outputs

    additive noises act.

    Now it is sufficient for the normalization of the channel to average on all possible realizations a(t) of the normally distributed input X (t) the covariance function of

    the vector output at the given a(t). From (7.8.65) and (7.8.66) after averaging we obtain t

    Reh u(y tsa [ oloninge,¢) —r)Ce(7,7 + |t—7'|)dr to

    -| | to

    —r)y(t! —7')Te(7,7')drdr’,

    Go

    147)

    Kyyyalo(tst’) = ff ot-nlae—t to

    |

    (7.8.67)

    to

    ve -7) Pela vara

    to

    (7.8.68)

    tate

    me

    A

    "[(otonaxt —7;t'—7'})—v(t—7T)¥o(' — 7) Pn (7, 7’)dr, to to



    (7.8.69)

    where I’, (t, ts) is the second initial moment

    of the input.

    Now we show how we may construct the normal model of the distributor the one-dimensional queueing system without queue.

    D in

    In this case the distributor D

    represents a determinate system with two inputs and two outputs. The stream of the clients enters on the first input, and on the second input — a binary signal about the

    state of a channel.

    If the second input is equal to zero then the distributor directs

    the stream of the clients in queueing channel.

    In this case the first output of the

    distributor coincides with the first input, and the second one is equal to zero. If the second input is equal to 1 then the distributor gives the clients a refusal and directs

    an input stream from the system.

    In this case the first output of the distributor is

    equal to zero, and the second one coincides with the first input. Thus the dependence

    between the inputs 21 (€); LQ (t) and the outputs Yj (t), Y2 (t) of the distributor with the one-channel queueing system without quence has the form

    yi(t) = a1 (t)[1— xo(t)],

    yo(t) = x1 (t)ro(t).

    These formulae show that the distributor

    (7.8.70)

    D represents the parallel connection

    of two elementary determinate nonlinear inertialess links.

    For the normalization of

    the distributor it is sufficient to linearize dependences (7.8.70).

    We may do it, for

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    803

    example, by the Taylor formula expansion of Yj (t) and Y92 (t) in the vicinity of the

    expectation of the vector of random inputs. As a result we get

    Y(t) = [1 — me, (t)]X1(4) — me, (1)X20), Y2(t) = mz, (t)X1i(t) + me, (t)X9(t). Example

    7.8.6.

    (7.8.71)

    For testing the accuracy of the approximation of the

    queueing systems by the normal models we use the stated method to the simplest

    one-channel

    queueing system without

    queue.

    Supposing the distribution of time

    queueing as an exponential one we obtain a structural scheme

    of the considered

    system (Fig.7.8.6). Here D is an distributor which determines the movement direction of each next client; jl /(u + s) and | /(Wise s) are the transfer functions describing

    the distribution of time queueing; Y (t) and Y> (t) are the input and the output. The equations which describe the behaviour of the system have the form:

    ve + (mz + wy) = Me_My, — PM UN + eS My,)X,

    Y= Y{+ Yi", Yo = (my,/p)X + (me/p)(Yi — my.) + meV",

    (I)

    (II)

    Fig. 7.8.6 where U" and ee are the normal noises with zero expectations.

    Hence we obtain

    the equations for the expectations of the signals:

    ty, + (me + H)My, = WMz,

    My, = (My, /pL)Me.

    (IIT)

    The stationary (steady state) values of the expectations of the outputs are determined by formulae

    My, = pm, /(Ms + f) ;

    My. = mz /(me + p).

    (IV)

    804

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    In the special case of the Poisson input stream the expectation M,

    is equal

    to the intensity of the stream \ and the obtained formulae give the known

    exact

    solution. For the determinate time queueing J’ the structural scheme of a system is rep-

    resented in Fig.7.8.7. Consequently, the equations of the system have the form:

    Yaviy,

    V@=]Z6=T),

    t

    U(t) = i inde

    Z=0— ms — ma",

    (V)

    ee

    t-T

    Fig. 7.8.7

    Hence we get the equations for the expectations of the signals

    negr(t) Tag ty

    mgt Poms

    J

    = (lig

    ts!

    (V1)

    Mz (EJOT METily, = Thy lta

    In the steady state Eqs. (VI) give the opportunity to derive the following expressions of the constants My,, My,, Mz

    y=

    Mg = tin fT

    and My:

    arin (le ig) a tee re

    Le)

    VL)

    Solution (VII) also coincides with exact one in the case of the Poisson input stream. Now we notice that at the exponential distribution of time queueing the mean time

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    805

    Y,()

    Fig. 7.8.8

    queueing is equal to 1mipt. Therefore the formulae obtained for the case of the determinate time queueing coincide with the formulae at the exponential distribution of time queueing. Example

    7.8.7. Let us estimate the mean number of the served clients

    who received the refusal and the variance of the clients number on the outputs of the one-channel queuing system without queue (Fig.7.8.8). Here D is the distributor which determines

    the direction of the movement

    of each next client in dependence

    of the fact whether the channel is free or occupied; p(t — T) and 4)(t — T) are the weighting functions connected with the probability density plo ) of time queueing 7' ioe)

    by the relation (co) = if p(¢)d¢; X (t) is the input stream of the clients into the

    a system; Yj (t) is the stream of the served clients; Y> (t) is the stream of the clients who received the refusal and J denote the integrators whose outputs represent the

    numbers of the served clients and the clients who received refusals beginning with the instant fo till the instant f; Y{'(t) and U"(t) are the inner noises of the system

    which are determined by formulae (7.8.67)-(7.8.69) at

    X = Z:

    min{t,t’}

    Kyr(t,t’) =

    /

    y(min{t, t’} — r)T (7,7 + |t —t'|)dr

    to

    i |/p(t — r)p(t! — r')P.(r,7/)drdr’, to to

    (I)

    806

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    he

    Kyunltst) = ff et—niae—t -1+7)— 0 to

    Feed’

    to

    (II)

    t

    v(t! —7’) | 2(7, 7’) drdr’. quiet) = J f[(omaxte-r t!—7})—Y(t—7T) to

    to

    (111)

    Thus the set of the equations which describe the performance of the queueing system has the following form:

    Z(t) =[1— mu(t)] X(t) — me (t)U° (A),

    (IV)

    Yo(t) = my(t)X(t)+ mz (t)U°(t),

    (V)

    nit)= [et=n)a(ndr+ Hi",

    (v1)

    U(t) = /W(t = 7) Z(r)dr + U"(2),

    (VII)

    where U zie) is a centered random function; Mz (t) and My (t) are the expectations of X (t) and U(t). Accounting that My”

    = My

    = 0, we get from Eqs. (IV)-(VII)

    the set of the equations for the expectations:

    m,(t)

    =[1—mu(t)]me(t),

    t

    nae pER

    my,(t)= mu(t)me(t), t

    EO

    /sbtaSaaseeoee) ovisNAT)

    to

    to

    Hence in the steady state for the Poisson stream Mz (t) = X putting tg = —oo we find

    t

    Mz

    Oita

    Asner

    = X / b(t—r)dr,6Tt Sa ees —00

    Taking into consideration the relations

    t

    oe)

    / y(t —7r)dr = | e(o)do ='1, —00

    0

    (IX)

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    807

    |w(t — r)dr = | oyae ™ [o~ W(c)]do =jaceiape

    vio)= f eon

    (X)

    from (IX) finally we obtain

    mz = my, =u/(A+p), my =YO+H), my, =2/A4+n). (OM) Thus the normalization method‘for the steady state gives the exact values for the expectations of the outputs at any distribution of time queueing plo iN

    At last we give an approximate the steady state of the performance.

    estimate of the variances of the outputs for In this case it is convenient

    to present Eqs.

    (IV)-(VII) for Y; (t) and Y2(t) in the following form:

    NO=NO+VO-VN),

    HOH=V~O)+ Vo"),

    Yi) = F(D)X(), Yi") = neat Yo(t) = P(D)X(t),

    — (XM)

    cam

    Yo"(t) = Q(D)U"(é).

    Here 1

    weed

    My

    ®,

    D

    r~ ey

    My,

    a

    fiietak

    j

    Q(D) ~~

    _ My + mz2(D) BU) where D =

    socnet

    We m, ®2(D)

    -

    d/dt; 0, (D) and ®,(D)

    D

    1+ oD My

    XIV (

    )

    T+ Mm, ®2(D) :

    are the transfer functions correspondent to

    the weighting functions p(t = T) and w(t = T) respectively. It follows from the properties of a normal stochastic system that the input X(t)

    is uncorrelated with the inner noises Y{"(t) and U"(t) and it means that Y/(t) is also uncorrelated with ¥(@) and with VK)! Analogously Y(t) is uncorrelated with Ys ey

    Consequently,

    Kyi ty

    By Gt') + Ky?)

    + Kyu (t, t’) — Kyityy (t,t’) — Kyity! (t,¢’),

    (XV)

    Ky, (t,t’) = Ky (t,t!) + Kyr(t,t”).

    (XVI)

    808

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    The variances of the total numbers of the clients Vj (t) and V2(t) are determined by formulae

    Die)= //lees Ca to

    eee

    (XVII)

    to

    In the special case when X(t) is a white noise of the intensity A, Ky, (t, t’) =

    Ad(t — t) applying the formulae of the linear theory of the stationary systems

    (Subsection 6.1.8) to Eqs. (XIII) we obtain

    hu

    '

    OO) = apeaiarn eu r3

    AF

    (20 + 3p)

    (XVIII)

    _

    Bx

    GIR)

    ;

    (XXI)

    Kyl) gal Gy 0) + eg Ot CaP IG etryAbame mua aeenye Ap

    Kan

    Kyu

    (te)l

    dp?

    Bed

    iM

    =

    ee pe 204

    ! —(A+p)|t-t

    Ae i

    —-—

    De por

    ———

    a

    Ait

    ,

    —(A+p)|t-t'|

    !

    ;

    (XXII)

    OR a owe

    (A+ 4)? 4(X + p)5 = ATH? tle" QtHMIt-t1 4+(A+uy2) jé—t— te

    Kyyi(t,t’)

    |

    j

    se SS

    eee rate meatih jt bead —t' le leAPS

    Kea

    ay

    dp?

    Op?

    ————

    Dd K, m(t

    as

    =

    Ale "404

    Nee

    eelray a lt TO

    ,

    Ph 2

    es t|

    Atal

    —t’ leee



    (XXIII)

    :

    (XXIV)

    Accounting (XV), (XVI) and (XVIII)-(XXIV) after evaluations we find from (XVII)

    D,,(t) =

    a) ATH yo, (A+ p)4

    Ani oy, (He.3 + Ant Dies — d*) (A+ y)°

    7.8. Applications.

    Analysis of Stochastic Systems and Analytical Modeling

    809

    ¢ r? pw?(2A + 14A? 0+ 19Ap? + 10p3)

    2(A + ps)? TANS _A pr (4A* + TAu + 4p*)

    tm

    2(A + p)®

    AP?22/93 (23 + 14020 + 19? + 10u*) |-atnye

    2(A +p)? My?

    Dy, (t) = SP

    0g

    2+ We

    (A+ p)° (tu

    Ad? p(4r? + 13d? + 14Ap? + 83)

    200+ ny"

    N38(403 + Se 132 0+ 14Ap? + 83) mere rT! + ee A+ py" ee e From (XXV) at t —

    (XXV)

    AB(A4 + GAB + 137py?+ 12Ap3 + 5y*)

    (A+ #)

    AMPA + 2H)

    }

    (XXVI)

    oo for Dy, (t) we obtain D,, (t) = BEY (t) = My? /(A+p)4,

    i.e. the variances of the number those clients who are served during time unit and those who are received the refusal are equal and do not depend on time.

    7.8.6. Software for Nonlinear Stochastic Systems Analytical Modeling We see that the theory of stochastic systems given in Chapters 1, 5—7 is rather

    complicated and requires considerable mathematical background for studying it. To solve any practical problem of stochastic systems analysis one has to derive the respec-

    tive equations determining distribution parameters from original stochastic equations and then to produce a program for solving these equations. This is a tedious and time-

    consuming job which can be done only by those who have an extensive knowledge of mathematics and some skill in programming.

    At practice two main approaches to the

    software design for the distributions parametrization methods in nonlinear stochastic systems are usually used:

    @ the creation of the specialized libraries for different programming languages; e@ the creation of the specialized dialogue software packages. The library approach is effective for the analytical modeling and the automatic

    derivation of the equations for the distributions parameters and for the statistical simulation of linear stochastic systems also.

    The choice of one or another method

    is determined by available information about numerical model of a real system and

    random disturbances, required accuracy of the calculations of system efficiency and by the information technologies of research also. For the nonlinear stochastic systems the derivation of equations for the distribution parameters is rather cumbersome and

    810

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    accounts about 90% of the whole process.

    In this situation it is necessary to create

    the dialogue packages which realize the main operations of the nonlinear analysis.

    In recent years actual prerequisites for further improvement

    of the software

    appeared on the basis of using modern highly effective software of data processing

    and information storage. One of the directions of the development is connected with the implementation of the mathematical packages which possess the wide possibilities

    of analytical derivation such as “Mathematica” used functions.

    what allows to expand the classes of

    The other direction of the development

    is the implementation

    of

    the modern object oriented control systems of the\data storage of equations of the

    intermediate results. The package “StS—Analysis, version 1.0” is designed for studying the stochastic

    differential systems described by Eq. (5.4.6) with the standard Wiener process W(t) by NAM

    for the polynomial in y function a(y, t) of not higher than the third degree

    and the function b(y, t) independent of y or linear in y. The package “StS—Analysis, version 1.0” has been widely used in Russia since 1987. The package “StS—Analysis, version 2.0” is destined for studying by NAM

    the stochastic systems described by the

    difference equations.

    Working with any version of the “StS—Analysis”

    package, the user must

    only

    type on the display with the aid of the keyboard the original stochastic equations in the natural mathematical

    form without

    any special programmers

    tricks, intro-

    duce the corresponding initial conditions and fix the statistical characteristics must

    be calculated.

    Once the original stochastic equations are introduced the user may

    correct them, i.e. to remove some equations, to replace some of them by others, to

    correct errors and so on. When all the data are introduced and checked the user has to start computations. NAM

    Then the package “StS—Analysis”

    automatically derives the

    equations for the corresponding distribution parameters and solves them. The

    user may observe visually the behaviour of a system under study in the process of

    calculations by tracing the graphs or the tables of current values of various statistical characteristics on the display and get the copies of these graphs and tables from the file after the calculations are finished. So the use of any of the versions of the package “StS—Analysis” does not require that the user know mathematics and programming. Any

    version of the

    “StS—-Analysis”

    package

    automatically

    derives

    the NAM

    equations for the distribution parameters from original stochastic equations introduced by the user in the natural mathematical form and solves these equations.

    It should be remarked that the “StS—Analysis”

    package does not simulate a

    stochastic system and does not use the statistical simulation (Monte-Carlo) method but performs the theoretical determination of necessary probability characteristics of a system behaviour by the normal approximation method.

    7.8. Applications.

    Example

    Analysis of Stochastic Systems and Analytical Modeling

    811

    7.8.8. Let us consider the scalar stochastic nonlinear discrete

    system

    Yi (0+ 1) = —aYP(l) + ¥i(0) + dV, (I).

    (I)

    Formulae for the expectation and the variace have the form

    my (I+ 1) = —aBY (1) + mi(2),

    (II)

    Dy (+1) = @(BY(I) — BYA(NEY2(0) —2a( EY} (1) — m, (NEY 3(I)) + Dy (1) + 82.

    (IIT)

    The higher moments approximately are expressed in terms of the first and the second moments:

    BY}(I)= Di(l) + mi(0),

    (IV)

    EBY3(l) = mi (I EY?(l) + 2mi(1)D, (0),

    (V)

    BY (0) = mi(NEYA() + 3D, ()BY2(),

    (V1)

    BY?(l) = mi(IEYA()) + 4Di (QEYS()), EY£(l) = m (EY?(1!)+ 5 Di (NEY; (0. On Fig 7.8.9 graphs of m; and D, for a = 0.05, at the initial conditions m,(0) =

    (VII) (VIII)

    b = 0.05 anda

    = 0.1, b = 0.05

    lke D,(0) = 0.5 obtained by NAM

    a) a=0.05; b=0.05

    are given.

    b) a=0.1; b=0.05

    1

    7

    0,5

    0,5

    m,(l)

    m,(1) DY,

    0

    0

    D,(I)

    =

    100

    200

    300

    400

    |

    0

    0

    50

    100

    =

    150

    200

    250

    |

    Fig. 7.8.9

    Example

    7.8.9. Stochastic three-dimensional nonlinear discrete system

    is described by the following equatons:

    ¥i(14+1) = Vi(1) + a¥i (0),

    (I)

    Y2(1 + 1) = Yo(1) + 2aY3(I),

    (II)

    Y3(1+ 1) = Y3(1) — aYo(1) — 2aY3(1) — aY3(l) + 3a¥ (1)Vi (0)

    (III)

    812

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    at the initial conditions

    EY,(0) = EY2(0) = 0.5, EY3(0) = 0.8

    DY;(0) = DY2(0) =1, EBY;(0)¥2(0) = 0.4. Other variances and covariances are assumed equal to zero.

    tions and the variances obtained by NAM

    ie

    Graphs of the expecta-

    are given on Fig.7.8.10.

    a) a=0.01; b=0.02; c=0.03

    b) a=0.01; b=0.02; c=0.03

    D3(I) 0,5 4

    D,(1) ; Da(!)

    Fig. 7.8.10

    Example

    7.8.10.

    Let. us consider the stochastic nonlinear differential

    equations of the chemical reactor:

    Y; =14+ Y/Y. — (14 Y3)%1 + 0.5v, Yoo Vosy (Yack Ve, Y3 =

    —-Y; Y3 == C,

    (I) (II) (IIT)

    where V; and V2 are the normal mutually independent white noises with the unit

    intensities,

    C being some constant. The graphs of the expectations, the variances and

    the covariances at

    c=

    1.3 and the initial conditions

    EY,(0) = 0.5, EY2(0) = 3.3, EY3(0) = 3.0, DY; (0) = 0.01, DY2(0) = 0.5. DY3(0) = 0 obtained by NAM

    are given on Fig.7.8.11.

    (IV)

    7.9. Online Conditionally

    Optimal Nonlinear Filtering

    813

    Fig. 7.8.11

    7.9. Online Conditionally Optimal Nonlinear Filtering 7.9.1. Preliminary Remarks The simplicity of the Kalman and the Kalman-Bucy filters (Subsection 6.6.2) leads to the idea of conditionally optimal filtermg. This idea is to reject the absolute (unconditional) optimality and to restrict oneself with finding the optimal estimate in some bounded class of admissible estimates satisfying some rather simple (differential

    or difference) equations which may be solved online while receiving the results of the measurements.

    The main problem of applying the conditionally

    optimal filtering

    theory is to define a class of admissible filters. It is recommended usually to define the class of admissible filters as the set of all filters governed by the differential or the difference equations of the prescribed form with some undetermined coefficients. The problem of optimization in this case is reduced to finding optimal values of all undetermined coefficients which appear generally time functions.

    814

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    The first characteristic feature of the nonlinear conditionally optimal estima-

    tion problem is that it is essentially the multi-criteria problem since m.s.e.

    E |X (t)

    _ X (t)|? is desired to be minimal at any ¢ in some interval. So there is an infinite (noncountable in the case of continuous time) set of criteria in this problem.

    For-

    mula (5.10.6) gives the absolutely optimal estimate which is optimal at any t. But

    restricting the class of admissible estimates we cannot hope to find an estimate which would be optimal for any t. Only in the linear problems this is possible.

    Thus one

    has to restrict oneself only by the Pareto optimal decision. The second characteristic feature of the conditionally optimal filtering is that the optimal values of the coefficients in the equations of a filter are determined only by prior data without using measurements.

    So the optimal filter may be designed beforehand quite in the same

    way as the Kalman or the Kalman-—Bucy filters. The measurements need to be used

    only in the process of filtering while solving the equations of the filter. Let {X i} be a state sequence of random variables.

    Another measurement

    se-

    quence of random variables {Yi} is defined by the equation

    Y; = w!(X1, Vi),

    (7.9.1)

    {Vi} being a sequence of independent random variables with known distributions.

    In

    problems of practice the random variables V; cause the errors. This is the reason why

    they are usually called measurement

    errors.

    The problem is to find an estimate X;

    at each | of the random variable Xj] using the measurements of the previous random variables Y; , ... , Yj—-1 in the class of admissible estimates defined by the formula Xx = AU,, A being some matrix of the rank n, and the difference equation

    Vig = 01001, Vi) + 1

    (7.9.2)

    with a given sequence of the vector structural functions Ci(u, y) and all possible values of the matrices of coefficients 6; and the coefficients vectors yi.

    Any given

    sequence of the structural functions ¢] (u, y) determines the class of admissible filters. Any choice of {6}, {yi} and the given functions C(u, y) determines the admissible filter. The dimension 71 of the vectors Y; is supposed less than the dimension 7 of

    the vectors X;, 1

    (0) DX (0) DX,(0)

    = = = =

    EX;(0) = 4000, EX2(0) = —20, 160000, DX, = 10000, DX2(0)=1

    (VIN)

    are given on Figs.7.9.1-7.9.3. The expections, the variances and the covariances of the processes X; (I) (2 ale 2) are given on Fig.7.9.1. The variances and the covariances of X;(I) are practicaly identical. Graphs of a11(I), a12(1), ai(l), a22(I), Bx(0), B2(1), ya (1), y2(0) are shown on Fig.7.9.2. The results of statistical simulation the processes AGAl), X;(l) (i Nh

    m.s.e.

    is

    does not exceed 1%. For the linear Kalman filter (at @22 = 0) maximal m.s.e.

    is

    equal to 5%.

    2) are presented on Fig.7.9.3.

    Maximal

    824

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    0

    4000

    8

    16

    -20

    3000

    -50 4

    EX, 130J

    950000 4 750000 +

    550000 + 350000 + 150000

    32

    at

    24

    32

    ft

    Problems

    825

    Problems 7.1. Using formula (5.1.16) show that in the case (5.4.7) the function o(Y, t) in formulae (7.1.3), (7.1.5) must be replaced by ((Y, t),

    dbAieaateau /RNa Rs

    nee du),

    (1)

    Chapter 7 Methods

    826

    of Nonlinear Stochastic Systems Theory

    o(Y,t) = b(Y,t)vo(t)b(Y, t)”. 7.2.

    Show

    that for Eq.

    (II)

    (5.4.6) in the case of an arbitrary process with the

    independent uncrements W(t) formulae (7.1.3) and (7.1.5) are valid if we assume

    0? x(A3t) Ht) =bald) =|aa t) =[v,,(t)]

    =

    |——— =

    ()

    |

    I

    where x(A;t) = 0/0t\n hy (A; t), hi(A;t) be the one-dimensional characteristic functions of W(t). 7.3.

    Show that the infinite set of equations for the moments in the case Eq.

    (5.4.6) where a(y,t) is defined by (7.4.1), b(y,t) = b(t), V =

    W, W

    being an

    arbitrary process with the independent increments has the following form: p

    N

    Qk, ..,kp = S k, pS

    y Pry nce

    ky

    OF het paglig ha bhi tees lcs bes eeig SEs 5S)

    ky

    + > “i = CE 0 hy=0

    hp=0

    (ki,

    php

    where

    kp =

    =O, 1 R236. 3 kak rn

    Nhvac a

    h

    (I)

    Ca Xbuy Jp Ubi —hi yey by—hy

    Qhrs

    Be

    pee sae

    152,

    ys

    (het)

    eo

    |.0

    Me

    Also show that Eqs. (I) may be written in the more compact form using multi-

    indexes k = [ei

    8 ke) pr Chik dhs |, Cu en P

    a

    (pail

    (eat.

    N

    SS Ke

    ) Kp

    eogoK kp

    k

    Ar hOht+k—e, + > Chxnoe-n h=0

    (IIT)

    h=0

    0, eis

    |k| =

    kj +----- k, aby

    ay

    where €, is the vector whose components are equal to zero except the rth component a

    G

    which is equal to the unity, €, =

    [Mere rr

    Ok = Wey... bys Gh = Gry

    ae ?

    skp

    XS he

    hes

    (IV)

    Xn = Oat |h| =hy+---+hy < 2 and yp = Ys at h = e, + €,. In Eqs. (III) Qs is equal to zero if at least one of the components of the multi-index $ is negative,

    Problems

    827

    and is equal to unity if all the components

    of the multi-index s are equal to zero,

    ao

    =

    il

    7.4. For the linear system (1.4.5) with the parametric noises N(t) considered as the system with the 6-differentials write out the correspondent It6 equations assuming

    that the white noise is normally distributed. 7.5. Consider the linear system with parametric noises

    Ag+ (1+ Va)(B + BY)q+(1t+Va\(C+C')g=(1+NU)Q*, where g =

    [q eer if and.

    =

    (1)

    Kae ye es (ii are the n-dimensional vectors;

    A, B, C are the symmetric matrices; B’, C’ are the antisymmetric n X n-matrices; a—

    [Vi V2 V3 ‘ea is the normally distributed white noise of the intensity ”. Show

    that to Eq. (I) corresponds the following Ité equation:

    Ag+ (1+ V3)(B+ B’)g+ (1+ Ve)\(C+C')q = (14+ Vi)Q* + 6[-1,3Q*A7'(B — B’)

    +023(C + C’)gA~*(B — B’) + v33(B + B')gA~"(B—B')]. 7.6. Considering the equations of Problem 1.14 at IT

    vw = ¥(q,t),

    (I)

    = ~V, Q = pV where

    ¢ = (4,9, t) are matrix functions of the stated variables, V is the

    normally distributed vector white noise of the intensity l/ as the equations with the

    0-differential, reduce them to the Ité equations for the canonical variables. 7.7.

    Show

    that for the linear stochastic system with the parametric noise of

    Problem 5.5 (i) equations for the expectation, the variance and the covariance function have the form vis

    —e(1 =

    20€V22)m

    =

    2€0kr42,

    D = —2e(1 — 20€v22)D + kh? + €7¥22(m? + D) —2ekmry2, OK

    (ti, t2)/Ote a

    —e(1 =

    20€V22) K (ty, te).

    (1)

    (II) (I)

    Find the stationary values of m and D and the conditions of the stationary regime existence. 7.8. Show that for the linear stochastic system with parametric noises of Prob-

    lem 5.7 (i) equations for ™ and K have the form .

    Mm,

    .

    =

    m2,

    —2m2Cwo(1

    2,

    M2=

    —m,w6(1

    ca 20CwoV33)

    oe 20CwoV23)

    =

    20kC€wo9133

    (1)

    ki; = 2k12, kyo = —w2(1 — 20Cwove3)k11 — 2¢wo(1 — 20Cwov33)k12 + k22,

    828

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    hea = —Ww(1 = 20CwoV23)k12 _ 4Cwo(1 — 20¢woV33)k22 + k?v44 —2m,kwi42

    - 4m2Cwokv43

    + w6V22(m?

    + ky)

    +4Cwevo3(mi me + ki2) + 4¢7w2 nugs(ms + ko2). 7.9.

    Using formula (5.1.16) show that Eqs.

    (5.4.7) at functions 1, (2, 21

    (II)

    (7.2.3), (7.2.4) are valid for Eq.

    defined by (7.2.5)—(7.2.7), at Yo2 = Eno and o

    and ¢ defined by formula (I) of Problem 7.1. 7.10.

    Sow that NAM

    equations for stochastic differential system (5.4.6) in the

    case of an arbitrary process with the independent increments W

    coincide with the

    equations of Subsection 7.2.1 under condition (I) of Problem 7.2. 7.11. Derive NAM

    7.12. Derive

    equations for nonlinear stochastic discrete system (527-2):

    NAM equations for the continuous-discrete system defined by Eqs.

    (5.7.10), (5.7.11). 7.13. For the typical composite functions derive formulae for (Yo for NAM

    given

    in Appendix 3. 7.14.

    Show

    that

    NAM

    gives

    the

    exponential

    covariance

    function

    k(r )

    = De~°!"l for the stationary process in the system of Example 7.2.1 where (i) De

    (v/6b3)'/?, a=

    (i) D =

    (3/2vb3)!/?

    (V2mv/8b2)'/?,

    wa =

    at ply) =

    b3y°;

    (8/m)'/2(V/2rv/8be)*/3 at v(y)

    = boy” sgny. 7.15. Using the equations of Problems 7.14 in the case (i) deduce the formula for the standard deviation 7.16.

    0 = VD for an arbitrary instant.

    Show that the expectation

    the system

    Mm and the variance D of the state vector of

    TY +Y_= Mea’),

    26g,

    (V being the stationary white noise of the intensity ) obtained by NAM

    (I) are deter-

    mined by the equations

    Tin = bmg — m[1 + ab(m? + 3D)],

    T?D = bv — 2DT[1+4+ 3ab(m},)].

    (II) Find the moments method values of m and D for the regime of stationary oscillations. Using MM

    with the account

    of the moments

    whether or not this method rejects the obtained

    7.17.

    Using NAM

    of up to the fourth order verify

    m and D as a superfluous solution.

    show that the variances and the covariance of the state

    variables in the stochastic nonlinear system

    TY, + Yi = bsgn(Y2—-Yi),

    Yo+aY2 = V2DaV,

    (I)

    (the expectations of the initial values of Yj and Y2 being equal to zero, V the white

    noise of the unit intensity) are determined by the equations

    Thy, = 26ky2 + 2(14 A)ki1,

    Tkie = Bkoo — (1+ aT + B)ki2,

    Problems

    829

    koo = 2a(D— kao), 8 =b[ (ki: + boo — 2ky2)e/2]-2/2. Find k11, k12, ke for the stationary regime. moments

    Using MM

    (I)

    with the account of the

    of up to the fourth order verify whether or not this method

    rejects the

    obtained variance kj for the stationary regime as a superfluous solution. Solve this problem also in the case of the nonzero expectations of the initial values of Y; and

    Y>. 7.18. Using NAM

    show that for the system

    TY +Y =bsgn(X -Y), X +aX = V2DaV,

    (I)

    V being the normal white noise of the unit intensity, NAM equations for the variances and the covariances of the random variables Y;

    = Y, Yo = Y, Y3 = X have the

    form

    kyy =2k12,

    Thig = Tho — k12 + B(kis — k11),

    ki3 = keg —ak13, Tko2 = 2keo + 28(ko3 — k12),

    k33 = 2a(D — kgs),

    Tko3 = —(1 + aT )ko3 + B(k33 — k13),

    6 = b[ (ki1 + kag — 2k13)/2]-*/?, if the expectations of the initial values of Y;, Y2, Y3 are equal to zero.

    (II) Find the

    stationary values of the variances and the covariances. 7.19. For the stochastic nonlinear oscillator

    Y +0(Y,Y)+w2¥.=V,

    (I)

    where V is the white noise of the intensity / show that the variances and the covariances of the random variables Yj} = NAM

    Y, Yo = Y are determined by the following

    equations:

    kay = 2kia, ki2 = ko — (we + ki)kia — kokie, hoo =

    —2(w2

    + ki)k12 _

    2koko2 =e VY,

    (11)

    where k; and kg are the statistical linearization coefficients of the functions p(y ; y2); the expectations of the initial values of Yj, Y2 being equal to zero.

    7.20.

    Using NAM

    and the results of Problem 7.19 verify the validity of the

    following approximate formulae for the variances and the covariances of the stationary process in the systems:

    (i) the oscillator with dry friction 9(y1, y2) = bo sgn yo, ki

    =

    mv? /8w2b?,

    ki2 =

    0,

    koo —

    we kit;

    830

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    (ii) the Rayleigh oscillator Y(y1, y2) = —biy2 + bays, ki =

    (1/6w2b3)(by

    -F ,/? + 6vb3),

    ki2 =

    0,

    ko = w2ki1;

    (iii) the Van der Paul oscillator p(y ‘ y2) = b(y? = 1)y2,

    ky = (1+

    4/1 + 2v/w2b)/2, big =0, kag = wk.

    7.21. Using formula (5.1.16) show that the equations for the initial moments in the case of Eq.

    (5.4.7) may be presented in the form (7.4.3) if we add to function

    r,v defined by (7.4.4) the following item: (oe)

    Agr = [ f{tm+aqenr sel Yptep(y, tu) )*

    singe 8

    P ees » Oe

    TOs ook: pretuty)}

    s=1

    xup(t, du)wi(y)pr (y)dy.

    (I)

    7.22. Show that the equations for the initial moments in the case of Eq. (5.4.6) for the nongaussian white noise have the form (7.4.3) at

    vi girl Aaa a / aan -. O(iAp)”®

    x [47 a(y,t) + x(b(y,4)? A;2)] es wildy, v3

    hod

    (I)

    glrl

    ‘ |ata —oo

    ie

    Pe,

    ot

    x [id a(y, t) + x(b(y, t)” A; t)Je! "| Py(yjwi(y)dy.

    (II)

    A=0

    7.23. Using formula (5.1.16) show that the equations for the central moments in

    the case of Eq. (5.4.7) have the form (7.4.6), (7.4.7), (7.4.11), where 1,v is defined by (7.4.8), 2,v is defined by (7.4.9) at 0, substituted by O of Problem 7.1 and Priv is defined in Problem 7.21.

    Problems

    831

    7.24. Show that the equations for the central moments in the case of Eq. (5.4.6)

    with the nongaussian white noise have the form N

    ™41,0,h = ¥1,0,n(m, K,t)+)>

    wa Ove RU, 1g) Qi

    teel

    ca

    D)

    k=3 |v| =k

    (I)

    P

    Hr = Pr,o(m, K,t) = > ThP1,0,n(™, K,t)br—ey =

    N

    P

    a8ne I=3

    en(m, K,t) —S> TrPivva(m, K,t)Mr—e, |qv(a) ||v|=l

    (ID)

    h=1

    CAS |S

    ae

    a

    eeeae

    where

    BS stan,Korey / Ree nere

    (111)

    penktn, it) = i an(y,t)pv(v)wr (y)dy ,

    (IV)

    Yr,o(m,

    Bat)

    Me

    / onal fRtrcnaer

    ona

    ae

    a(y, t)

    + x(H(yTAsnjer™e-m | wi(y)dy,

    (V)

    A=—0

    Veet, K,tet)

    / at Ae a(y,t) (At onaemcceas

    H (HUNT AsnIEeo-—} OURO

    aaah

    A=0

    7.25. Derive the equations of Subsections 7.4.1, 7.4.2 for the initial and central moments in conformity to stochastic discrete system (5.7.2). 7.26. Show that the equations for the multi-dimensional initial moments in the

    case of Eq.

    (5.4.7) have the form (7.4.15) with Yq, |...

    kajvi,... apelin’ Ke, tn)

    containing the following additional term: AGr, ,...,&ni¥1 spit yam

    aiin, tn)

    832

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    (oe)

    (oe) Vi1

    Vip

    Vn—-1,1

    aie of fatness —0Co

    or

    noe

    *Yn—1,1

    Re

    4{am ot c1(Yn; tn, u) lw:

    [Yap rh Cp(Ynytnyt@)3” = Inks Une? fi > UnsYni x eles

    ROACUT ON 7.27.

    byeltndu) 2 Yap? Cs(Yns tn, uw)

    GaP ey aca 5 hls Ue JaUL

    1a

    (I)

    Show that the equations for the multi-dimensional moments

    in the case

    of Eq. (5.4.6) with the nongaussian white noise have the form co

    Gr

    aireltisstn\lOn=

    Te

    [o.e)

    ff

    ee COM

    girl Lae

    aS

    eel ete | > 6)

    xy? ey

    ie

    es5 wi : ye1

    Init eee

    Se Onidgr

    ag, © #4)

    7.28. Derive the equations for the multi-dimensional moments in the case of Eq.

    (5.4.6) for the nongaussian white noise. 7.29. Derive the equation for the multi-dimensional central moments in the case

    of Eq. (5.4.7). 7.30.

    For the system of Example 1.3.3 in the case when the vector of random

    disturbances N(t) =

    [Ni (t) No (t) ig is the normal two-dimensional white noise

    with the matrix of the intensity / show that the moments

    @,;

    of the process y

    = [Y1 Yo ie are determined by an infinite set of equations for the moments: : Ops

    = =

    TOpr—1,s41

    2 — SWoAr41,s—1



    4 2ESQy,s + ViQwor(r =

    Papa

    1 +5wo8(s ae 1)(1107,5~2 - 21120r41,5—2 ae V220r42,s—2)

    abet Vi ipa eset fd)

    es

    pee

    7.31. Under condition of Example 7.2.0 at y(y) = y? show that the equations for the first three semiinvariants m, D and K3 have the form

    m=—Kk3—3mD—m',

    (1)

    Problems

    833

    D= =6mk3 — 6m?D — 6D? +1,

    k3 = —9k3m? — 18m3D — 27Dkz3.

    (II)

    7.32. Prove that for the stochastic two-dimensional system

    Yj = ag—a¥, +hiVy,

    Yo

    SHY

    ove

    (I)

    (ao, a, hy, hy > 0, V; and V2 are the independent normal white noises with the unit intensities) the equations of the semiinvariant method with the accuracy till the

    first four semiinvariants Kj; (t, 7 = 0,1, 2,3, 4) have the form

    Ko1 = —4K19

    +49, K10 = Kioko1,

    Ko1 = —2ak29 + hi,

    (1)

    K11 = —@K11 — Ko1k20 — K10K11 — Ka1, k20 = —2(K11K01 + KioKo2 + K12) + h3, K30 = —3ak30, Kao = —4aKk40, Ko1 = —2ak21 — Ko1k30 — 2K20K11 — K10K21 — K31 ,

    (II)

    k12 = —ak12 — 2(Ko1K21 + K20K02 + K7, + K10K12 + K22) ;

    Ko3 = —3(Ko1k12 + 2K11K02 + K10K03 + 3k13),

    (IH)

    k31 = —3ak31 — Ko1Kk40 — 3K30K11 — 3K20K21 — K10K31, K22 = —2ak22 — 2(Ko1K31 + K30K02 + 3k21K11 + 2K20K12 + Ki0K22),

    K13 = —@k 13 — 3(Ko1K22 + 2K21K02 + K20K03 + 3k11K12 + K10K13) , koa = —4(Ko1k13 + 3k12K02 + K10K04 + 3K11K03) Show that at dg >

    (4) Kg

    (IV)

    hia disappears the difference between the values Ko), Ki),

    calculated with the accuracy till the semiinvariants of the second fourth orders. 7.33.

    Derive

    the equations for the first two moments

    of Problems

    5.24. Show that the solution of these equations at the initial conditions Mp

    5.23 and =

    EYo,

    Do = DY? has the form

    a(t) = a10 + @11 exp(—at), ao(t) = a29 + a1 exp(—at) + a22 exp(—2at) , where @19 =

    = @o/@; Q11

    =

    ayi(1 + 2aoa1)/aaj; 92

    Mo — Q10; G29 =

    (1 + 2apaj)ao/2aia*;

    a1

    = Do — G20 — M21. Find the stationary solutions.

    7.34. Show that the equations for the coefficients of the orthogonal expansions in the case of Eq. (5.4.7) have the form of Eqs. (7.5.8), (7.5.9) at co

    y20(m, K,t) = /[a(y,t)(y” —m™) + (y—m)a(y, t)” + a(y, t)]wi(y)dy ,

    tie

    (1)

    Chapter 7 Methods of Nonlinear Stochastic Systems Theory

    834

    , +2(u,Olen (oo), maly 1)? paul, K,t)= ffaly)ly? =m") +(y— (II) Deanne st) = [jae Tan (y) aly yet

    J,

    =

    E a qx(y)o(y,t )

    + f[tated+ tats) act) - Peta, tyap]

    HD

    Rg

    x ve(ls du) booa(ydpe(w)ay (Iel Ml = 3, and further,

    c= OulaK

    Deen ad

    tr [wm, K,1)K|

    N

    +S° De cy [ Pav (m, K,t) + ¥™ (m, K,t)? m] i=Saly |i

    +50 |vk(dF Pio 4lie send otpen

    (IV)

    Here Yxv(m, K, t) is defined by (III), and

    vB Cm, Kt) = faeyun(y)po(uay (lel el = 3,-6.9), (V) cSie a8 = ‘Age (y)wi(y)pr(y)dy (In|, Vv] =3,-.., N) (VD functions Puv(M, 1K , t), wr (Mm, K at); pko(m, K, t) are defined by formulae of

    Appendix 4 at py (y)= po(y)= 1. 7.35.

    Show

    that the equations

    for the multi-dimensional

    coefficients

    of the

    orthogonal expansions in the case of Eq. (5.4.7) are as follows:

    On, ’ ae

    +R

    ea Arora)

    tL Obn Es

    ys

    | erie

    a,0(tin Knstn)*rin(tn) + tr [BEM eg,0(Mns Kustn)K (tn)|

    Problems

    835

    Se ee N

    ar )

    )

    BPE

    A

    es Fetnibal

    ss, Bob)

    1=3 |vi] +--+ [val =! 1

    + tr

    Matas Sore

    ete.

    Madey vp (Mn; Fostn

    cere

    )

    1, (17m, Rasta)

    " +S

    Cher

    pies

    [ve ar

    en

    atta)

    ee

    acr

    ny

    K (tn) |

    CR (tat Bins te)

    = lhe.Neer | peer

    oe)

    ||

    ")

    ike] eumaxrntt. ... NV),

    where

    Dee or aa VA eh

    -

    CUT?

    =f

    OE

    olGr 55 fin; tn)

    eG BO aE)

    eeeOYn ae

    a(Yn stn)

    Fo tesssesne(ths To prove formula(6) we multiply term-wise the equalities

    on

    1 a -1u>0.

    (18)

    Sy (u) are connected with the generalized Laguerre polynomials

    L°(ax) (Appendix 7) by the relation

    Si (u) = (—k)-’v LE (ku) . 7

    After differentialting in (18) we get

    S2(u) = (-k)> » aEee

    ae

    (19)

    As a generating function of these polynomials serves eus/(1+s/k) p(s)

    = (14 (1 a s/k)ott

    = ~

    seu).

    The system of the functions Old ea PEI 209 (41) (v a Oy Leen. .) is complete in the space Lo ({0, 00)). Consequently, any function f(u), in particular, any density of a random variable satisfying the condition

    f?(u)

    yo] 2e—kul2 du

    l1.4 Property

    3: uo

    l

    [csz 0

    (wa = sug tie

    S74) (uo).

    (25)

    Relation (25) gives as new property of incomplete gamma-function:

    y Wr.9) = peteta

    (26)

    0 We shall formulate it as one of the properties of the polynomials af (u). Property

    4: For the incomplete gamma-functions and any values of the para-

    meters @ and k the following equality is true

    ea eet

    —k)-’~°

    Pt (Gesot ka al) eae wares

    —1)*-

    =

    (8

    ee

    FSS

    +

    CREE 1k

    Cl

    (27)

    > After substituting (19) into (25) we shall have

    i

    i

    fea e7 ,—ku4 Siaa (u)du = —(k) = rf ure J )

    —ku

    ul

    (

    1)

    S > CH atyut (—ku)4du na T(a+p+t1)

    i =< a+v te ( ) one 1) aCe Fo

    tesdu.

    — (28)

    “Appendices

    849

    Let us make the change of variable t = ku in (28) then uo

    ure *" $%(u)du oO

    kuo

    T(a+v+1) =F See eid) e aONS » (OeSee Pe aeET a Wevieth re

    4

    dt ae

    =

    G20)

    Applying formula (26) to the right-hand side of (29) we find Uo~

    [cet

    se(udu

    0

    va

    55”

    ir 8 1l-a

    Set

    =1)#-"

    2

    )

    Hs

    Cr

    T(a+v+1)

    OS, ine

    d

    das

    1, kuo)

    But according to (25) we have

    prv-i-a ey

    a

    plier 1) y(a + w+ h1, akuo) ”*T(a+p+t1)

    1 = ou

    1 e —k MENGE 1 (uo).

    After multiplying the both parts of the latter equality by k we come to (27). < Property

    5.

    The system of the polynomials

    {o5; (u)} at a given @

    >

    —1

    represents an orthogonal system of the polynomials for which the gamma-distribution

    w(u) = kor foe) hot

    yo

    ent

    Ta SF 1) serves as the weight:

    :

    a

    1

    eee 2) ey ene k?*T(a@ +1)

    eeei

    fiern

    > For proving the orthogonality of the polynomials

    at V = Lb, at

    a y”

    V = pl.

    (30)

    So (u) and Dy (u) it is

    sufficient to show the orthogonality of So (u) to the degrees ur, A < Vv. We have rs kort

    a

    fee J T'(a+1)

    bw

    du

    850

    Appendices

    bbyYoopt nae

    ili

    ”“T(a+pt

    ipa

    1)

    iba

    T'(a+1)

    Let us partition this sum into two ones taking into account that

    Ch

    Cl MCEise. Oe

    DA Oe re Co ry

    en

    —1 nee

    Then pe kotly@



    5

    i FY. S%(u)(ku)*du Thy EY joes 0

    vr

    v-1

    ce Patyt DMatrAtut)) T(a+p+1)[(a+1)

    ae

    enetict EE aoe NC Sar eee +(—k)’ bat T(a+ p+ 2)P(a +1) Here jt — 1 is taken as ps in the second sum what involves the change of the lower limit of the summation by zero, and the upper limit by the number / — 1. Thus we

    have

    I Te

    T(a+1)

    v-1

    =(Sh)

    Sachs p=0

    Stu) (uy py

    \U

    U

    U

    T(at+v+1)(a+A4+p4+1) aa T(a+p+1)0(a+1)

    a+p+1

    =i (oat) alae > a 9 in Oc T(at+v4+1)l(a+A+p4+1) T(a+p+1)l(a+1) p=0 The obtained sum is Sed that C’ Se =

    C,ae

    hied to the initial sum but has one less item. Accounting

    + C,iese we partition once more

    the sum into two ones and

    substitute 1 — 1 eeLi, then

    iS Liat mea

    U

    \(k U rd U

    = (-k)-”A( = 1) Sicnetog_ er

    Appendices

    851

    After performing this transformation A times we shall have koetlye

    Tatty?

    Sv (u(kuy\du

    T(a+v+1)

    ( 1 relative to 3

    But the polynomial of the pes degree

    orthogonal to all w, A = 0,1,2,..., v—1 is determined uniquely with the accuracy till a constant multiplier.

    Consequently,

    the integral in (32) may

    differ from the

    polynomial Sh,v (u1) only by a constant multiplier which as it is easy to notice is equal to 1. 4 Let us indicate one more property of Peromd Ta) which follows from the proper-

    ties of the generalized Laguerre polynomials.

    Accounting the relations between the

    polynomials Lo (u) and De Lt), So(u) = (—k)’v! Le (ku) we write

    ee

    I ee

    Qv+a—1—ku

    Sk

    =

    Property 3. Relation (34) in the case of the x?-distribution takes the following form:

    S,v(u) =

    —(4v+r—4—u)S,~1(u)

    —2(v—1)(r—442v)S,,~-2(u).

    3. Statistical Linearization of Typical Composite Functions 3.1. Scalar Argument Functions (Table A.3.1)

    p(X) & vo(m, D) + ki(m, D)X°, where

    X°=X—m,

    m=EX,

    D=E|X?|’,

    ki(m, D) = 0¢0(m, D)/Om, i!

    O(z) = i

    /

    2

    dt,

    1

    (z)= aod

    2

    hh

    0

    C=m/VD,

    ¢t=(m+a)/VD,

    (7 =(m—-a)/VD.

    (35)

    854

    Appendices

    Table A.3.1

    TNS,

    os)

    Qn = Man-1 + (n— 1)Dan-2, ag lias =m

    Seed ec fe = (m+ eD)Ba $ (=) DBs yf = e~ © P/2(m cos am — aD sin am)

    fae eee 16. ketoaX (= 2. 1,

    Jl EOC

    Oat

    oe 3)

    Soa

    oC OG)

    |X| 1

    10. Xp Xpcosa?X

    V1, .-.-,

    ear 2e5

    =

    P

    eal?

    =e~?

    sin a?

    Vm are nonnega-

    T

    +

    |v| it

    sh

    £ pis

    ale A

    Ka/2(m, cosa’m + Kpasina?

    Jp

    kpAYy—e, a Mp Yy—e» §

    s

    ) Urkpr Vy—ey—er = kp Vy —2¢y r=1

    tive integers

    — + =r

    =f S. Vr kyr By —ey—ey —

    Sea

    Aghaaen

    n

    m)

    856

    Appendices

    Table A.3.2 (continued)

    12:EX45 Up XE e cosa

    ff

    Xx:

    s Vr == MpVy-e, — Kp@%y—e,

    :

    n

    V4, ..., Vn — nonnegative integers

    c

    c

    Ar oy Ur kor Vy —ep—e, =e kppYy-26,

    \v| > 1

    tr {a(K + mm7)}

    [1 — 2®(n) —2k1 2m, ®'(9) //k11(kitk22 — kf) ®(nz) + [1+ 2®(n)

    14. sgn (x — asgn X2)

    +2k12m,

    (7) /s/k11(ki1k22 _ Ki5)}

    x®(nz ) + 2ki2'(n)[®'(n0 ) — 9!(nt)\/ hii (kt k22 — kis) nt =(m + a)/Vki1, na = (m, +)/Vki1. (Approximate formula for small l,

    l= ky2a/./kii(ki1tk22 — k?,)) 4, Integrands Evaluation in Equations Parametrization of Distributions

    of Methods

    based on

    4.1. Moments Methods Owing to the known expressions of the semiinvariants in terms of the characteristic function (Section 3.8)

    r Kay

    its

    i=

    Als! x(u5 t) O(ipis)* ...

    Kio “nea ld) (S36 ice he Sos

    .0)

    O(iptg)*

    ae

    ae eee

    a(t)»

    (1)

    Ther ogc; ls higeee2. 7 = ae, N) being the semialee 0 when |s| = 1 since these semiin-

    invariants of the process W(t) (Ks, nec

    variants represent the expectations of the components of the process W(t) which are all equal to zero. Hence, by the Taylor formula

    x(p;t)= > oDoer SS fan

    k=2 |s|=

    (2)

    Appendices

    857

    where Py is the residual term. Substituting here the expression /l = b(y, EIR and collecting the items of the same degrees in A, , ...

    , Ap we obtain

    (oust) ast) = 0 my ns

    lust) +m, (3)

    k=2 |h|= where Why,

    ..,hy(¥,

    t)

    K sth me t vedere) Modsaahnce

    = HEL aefan Slew we ee he eyes

    PP ips eso Pais

    Pie

    Pipt:Pqp=hp

    Pirt--+Pq, =h1

    ee ay&

    Function D(s

    ee

    oS 5. fie

    se

    a>0

    aeoa

    oo eRe

    b

    wae a“s

    Jes

    hae

    t=8=9

    261.2

    = 1+ Ver—-1

    b?6"(t—r) +2cb6!(t—r)+4(t—T)

    b2 52 Yebs+1

    Appendices

    865

    Transfer Functions and Frequency Characteristics for Some Continuous Linear Systems Table A.5.2. Gain-Frequency

    Phase-Frequency

    Characteristic A(w)

    Characteristic w(w)

    OatwI1/a b G a’s- + 2eas +1

    9. b?s? + 2ebs +1

    et Ba b aes S| (tie ara)? + 4aze2w?

    (1 — bw)? + 4e?b2w?

    ee: 2aEw arcs eee

    arctg ; abe

    Some Typicl Models of Stochastic Discrete Systems Table A.5.3.

    1. Nonlinear Autoregression (AR): e | order

    years = pr (Yr a Rata t5 Yeqi—1)+

    +0, (YE, ---) Yeri—i)Vi e first order

    Yeti = pe(Ye) + Ve (Ye)Vi

    866

    Appendices

    Table A.5.3 (continued)

    2. Linear AR:

    11 e | order

    VRS")

    > arp Sere TV

    EVE

    r=0

    Yeon

    cpYp + ae Vi is Yq => Gpr Ea t

    @ first order e | order with

    r=0

    I

    Moving Average (ARMA)

    +A, + ys bskVi+s $=0

    3. Linear Integrated AR (IAR) I-1

    e | order

    ara

    =

    arp Na Vere + VeVi nO)

    I-1

    e | order with

    Nea

    =

    bseVi s=0

    4. Nonlinear ARIMA

    Yui =

    SS Ark Petr (Ye+r) r=0

    I

    of | order

    =F > bskVe4+s s=0

    5. Stationary linear

    A(V)Y;

    =—C+

    B(V)Vk 1-1

    ARIMA of I order

    A(V) = IV? —

    > Op r=0

    BW

    I

    iv

    s=—N

    Vug = Urqi

    Y= OH)+WO

    Appendices

    867

    Operators, Weighting (Green) and Transfer Functions for Some Typical Linear One-Dimensional Systems with the Distributed Parameters Table A.5.4. Weighting (Green)

    Transfer

    Function g(€,€1,t, 7) , |Function ®(€, €1, s)

    1. y(t) = e(t—A) Visi =0,NG