beamer lecture

Upload: j-jesus-villanueva-garcia

Post on 04-Apr-2018

260 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/29/2019 Beamer Lecture

    1/43

    Stochastic Signal Processing

    Arnab Ghoshal

    Spoken Language SystemsSaarland University

    November 09, 2009

    Arnab Ghoshal Stochastic Signal Processing

    http://find/
  • 7/29/2019 Beamer Lecture

    2/43

    Outline

    1 Introduction to probability

    2 Random variables, probability distributions

    3 Moments of random variables

    4

    Random vectors5 Random sequences

    6 Random processes

    Coverage: Stark & Woods: Probability and Random Processeswith Applications to Signal Processing, Chapters 1-7

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    3/43

    Sigma fields

    For a set , a subset F of the power-set2 forms a -field (or-algebra) if it satisfies the following properties:

    1 Non-empty: F, and F

    2 Closed under complement: Ec

    F,

    E

    F

    3 Closed under countable unions and intersections:

    i=1Ei F and i=1Ei F, Ei F

    For example: If is the sample space of an experiment thenthe subsets of are called events and form a -field.

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    4/43

    Axiomatic definition of probability

    Given a sample space , and a -field F of events defined on, we define probability Pr[] as a measure on each event

    E F, such that:1 Pr[E] 02 Pr[] = 13 Pr[E F] = Pr[E] + Pr[F], if E F =

    The triplet P= (,F, Pr) is called a probability space.

    It follows from the axiomatic definition

    Pr[E F] = Pr[E] + Pr[F] Pr[E F]

    Pr Nn=1En = Nn=1 Pr[En], if Ei Ej = for all i = jThis can be extended to the case of countable additivity:

    Pr

    n=1

    En

    =

    n=1

    Pr[En], if i = j, Ei Ej = .

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    5/43

    Conditional probability

    Given a probability space P= (,F, Pr), and events

    E, F F, we define:

    1 Joint Probability: Pr[E, F] = Pr[E F]

    2 Conditional probability: Pr[E|F] = Pr[E,F]Pr[F]

    3

    Events E and F are independentiff:

    Pr[E, F] = Pr[E] Pr[F] Pr[E|F] = Pr[E] Pr[F|E] = Pr[F]

    4 If E1, . . . ,En are exhaustive (i.e. ni=1Ei = ) and mutually

    exclusive (i.e. Ei Ej = , i = j), then:

    Pr[F] = Pr[F|E1]Pr[E1] + . . . + Pr[F|En]Pr[En]

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    6/43

    Bayes Theorem

    Given a probability space P= (,F, Pr), a set of disjointexhaustive events E1,E2, . . . ,En F, and any event F Fwith Pr[F] > 0, the probability Pr[Ei|F] is calculated as:

    Pr[Ei|F] =Pr[F|Ei] Pr[Ei]

    Pr[F|E1]Pr[E1] + . . . + Pr[F|En] Pr[En]

    Proof:

    Pr[Ei|F] =Pr[F,Ei]

    Pr[F]

    (1)

    =Pr[F|Ei]Pr[Ei]

    Pr[F]

    (2)=

    Pr[F|Ei]Pr[Ei]

    Pr[F|E1]Pr[E1] + . . . + Pr[F|En]Pr[En]

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    7/43

    Random Variable

    A real Random Variable X() is a mapping from to the realline (i.e. X : ), which assigns a number X() to everyoutcome , and satisfies the following properties:

    1 For every Borel set B making up the -field B on ,the set EB { : X() B} is an event

    2 Pr[X = ] = Pr[X = +] = 0

    Note that:

    An RV can be complex: Z() = X() + jY(), where X andY are real RVs.

    Property implies that the set { : X() x} is an eventfor every x.

    An RV can also be thought of as a mapping between

    probability spaces, i.e. X : (,F,Pr) (,B,PrX)

    Arnab Ghoshal Stochastic Signal Processing

    P b bili Di ib i F i

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    8/43

    Probability Distribution Function

    The distribution function F() of the RV X is defined as:

    FX(x) = Pr[{ : X() x}]

    Properties:

    1 FX() = 1, FX() = 0

    2

    FX(x) is a non-decreasing function of x, i.e. if x1 x2 thenFX(x1) FX(x2)

    3 FX(x) is continuous from the right, i.e.

    FX(x) = lim0

    FX(x + )

    4 If FX(x0) = 0 then FX(x) = 0 for every x x05 Pr[X > x] = 1 FX(x)

    6 Pr[x1 < X x2] = FX(x2) FX(x1)

    Arnab Ghoshal Stochastic Signal Processing

    P b bilit D it F ti ( df)

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    9/43

    Probability Density Function (pdf)

    If FX(x) is continuous and differentiable, then the pdf isdefined as

    fX(x) dFX(x)

    dx= lim

    x0

    Pr[x X x + x]

    x

    If X is discrete and Pr[X = xi] = pi, then

    fX(x) =

    i

    pi(x xi), where (x) =

    1, for x = 00, otherwise

    Properties:

    1 fX(x) 02 FX(x) =

    x

    fX(y)dy

    3fX(y)dy = FX() FX() = 1

    4 Pr[x1 < X x2] = FX(x2) FX(x1) = x2x1 fX(y)dyArnab Ghoshal Stochastic Signal Processing

    C diti l di t ib ti d d iti

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    10/43

    Conditional distributions and densities

    The conditional distribution functionof X : given theevent B is defined as:

    FX(x|B) Pr[X x,B]

    Pr[B],

    where Pr[X x,B] = Pr[{ B : X() x}].

    Similarly, the conditional densityf(x|B) is defined as:

    f(x|B) dFX(x|B)

    dx

    For exhaustive and mutually exclusive B1, . . . ,Bn,

    FX(x) = FX(x|B1) Pr[B1] + FX(x|B2) Pr[B2] + . . . + FX(x|Bn) Pr[Bn]

    fX(x) = fX(x|B1) Pr[B1] + fX(x|B2)Pr[B2] + . . . + fX(x|Bn)Pr[Bn]

    Arnab Ghoshal Stochastic Signal Processing

    B th d t t l b bilit f df

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    11/43

    Bayes theorem and total probability for pdfs

    From definition of conditional probability, it follows

    Pr[B|x1< X x2] =Pr[x1< X x2|B]Pr[x1< X x2]

    Pr[B] =FX(x2|B) FX(x1|B)

    FX(x2) FX(x1)Pr[B]

    We can compute Pr[B|X = x] as the limit:

    Pr[B|X = x] = limx0

    Pr[B|x < X x + x] =fX(x|B)

    fX(x)Pr[B]

    Total probability formula:

    Pr[B] =

    Pr[B|X = x] fX(x) dx

    Arnab Ghoshal Stochastic Signal Processing

    Joint distributions and densities

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    12/43

    Joint distributions and densities

    The joint distribution of two RVs X and Y is defined as:

    FXY(x,y) Pr[X x, Y y].

    For continuous and differentiable FXY(x,y), the joint pdf is:

    fXY(x,y) =2

    x y

    FXY(x,y)

    FXY(x,y) =

    x

    y

    fXY(, ) d d

    Statistics for each of the RVs are called marginals:

    FX(x) = FXY(x, ) FY(y) = FXY(,y)

    fX(x) =

    fXY(x,y) dy fY(y) =

    fXY(x,y) dx

    Arnab Ghoshal Stochastic Signal Processing

    Functions of random variables

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    13/43

    Functions of random variables

    For a real RV X and a function g() defined on the reals, the

    expression Y = g(X) is a new RV, defined as follows:For every , X() is a real number, and g(X()) Y()is another real number assigned to by the RV Y.

    The events { : Y() y} { : g(X()) y}, and

    FY(y) = Pr[Y y] = Pr[g(X) y]

    pdf of Y = g(X): If y = g(x) has n real roots x1, . . . ,xn, then,

    fY(y) =

    ni=1

    fX(xi)|g(xi)|

    g(xi) = 0.

    Arnab Ghoshal Stochastic Signal Processing

    Expected value of an RV

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    14/43

    Expected value of an RV

    The expected value (or mean) of an RV is defined as:

    E[X]

    x fX(x) dx = X

    The expected value of Y = g(X) is:

    E[g(X)] =

    g(x) fX(x) dx,

    The linearityof expectation follows from its definition:

    E

    n

    i=1

    i gi(X)

    =

    ni=1

    i E[gi(X)]

    Arnab Ghoshal Stochastic Signal Processing

    Variance and Covariance

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    15/43

    Variance and Covariance

    The variance of an RV X is defined as:

    2

    X

    (x X

    )2fX

    (x) dx

    For two RVs X and Y, the covariance is defined as:

    Cov[X, Y] E[(X X)(Y Y)] = E[XY] XY,

    where E[XY] is called the correlationof X and Y.

    Cauchy-Schwarz inequality:

    Cov[X, Y] E[(X X)2]E[(Y Y)2]Expectation and variance may not exist; e.g Cauchy pdf:

    fX(x) =1

    (x2 + 1)

    Arnab Ghoshal Stochastic Signal Processing

    Independence and correlation

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    16/43

    Independence and correlation

    Two random variables X and Y are said to be independentif their joint distributions and densities can be factorized as

    the product of the marginals:

    FXY(x,y) = FX(x)FY(y) fXY(x,y) = fX(x)fY(y)

    RVs X & Y are said to be uncorrelatedif

    Cov[X, Y] = 0 E[XY] = E[X]E[Y]

    Independence uncorrelated, but not vice versa

    For uncorrelated X & Y: 2X+Y = 2

    X + 2Y

    Arnab Ghoshal Stochastic Signal Processing

    Random vectors

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    17/43

    Random vectors

    A random vector is a vector X = [X1, . . . ,Xn]T, whose

    components Xi are random variables

    The probability distribution function of X is the jointdistributionof the RVs Xi:

    FX(x) Pr[X x] Pr[{X1 x1, . . . ,Xn xn}].

    The probability density function of X is similarly:

    fX(x) nFX(x)

    x1 . . . xn.

    For two random vectors X = [X1, . . . ,Xn]

    T

    andY = [Y1, . . . , Ym]T:

    FXY Pr[X x,Y y] fXY(x, y) (n+m)FXY(x, y)

    x1 . . . xn y1 . . . ym.

    Arnab Ghoshal Stochastic Signal Processing

    Random Vectors: Expectation vector

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    18/43

    Random Vectors: Expectation vector

    The expectation of X is a vector whose elements are:

    i

    . . .

    xi fX(x1, . . . ,xn) dx1 . . . dxn

    =

    xi

    fXi (

    xi)

    dxi

    where fXi (xi) is the marginal of the i-th component:

    fXi (xi)

    . . .

    fX(x1, . . . ,xn) dx1 . . . dxi1dxi+1 . . . dxn

    Arnab Ghoshal Stochastic Signal Processing

    Random Vectors: Covariance matrix

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    19/43

    Random Vectors: Covariance matrix

    For a realrandom vector X, the covariance matrix is:

    K E[(X X)(X X)T],

    while for a complexrandom vector Z it is:

    K E[(Z Z)(Z Z)H],

    where ZH is the conjugate transpose or Hermitian

    transposeof a complex-valued vector (or matrix) Z.

    Similarly, the correlation matrix is defined as:

    R

    E[XXT] = K+ T, for real random vector X

    E[ZZH] = K + H, for complex random vector Z

    Arnab Ghoshal Stochastic Signal Processing

    Properties of covariance matrix

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    20/43

    Properties of covariance matrix

    1 K is symmetric(i.e. K = KT) for real RV, and Hermitian

    (i.e. K = KH

    ) for complex RV.

    Kij E[(Xi i)(Xj j)]

    = E[(Xj j)(Xi i)]

    = Kji i,j = 1, . . . , n

    2 K is positive semi-definite(p.s.d) i.e. yTKy 0, y

    yTKy = yTE[(X )(X )T] y

    = E[yT(X)(X

    )Ty]

    = E[(yT(X ))2] 0

    3 If K is full-rank, then it is positive definite

    Arnab Ghoshal Stochastic Signal Processing

    Diagonalization of two covariance matrices

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    21/43

    Diagonalization of two covariance matrices

    Let P and Q be n n real symmetric matrices, and P be positivedefinite. Then,

    P = UDUT, where UTU = I and D = diag(1, . . . , n), i > 0.

    The diagonal matrix Z = diag(1/21 , . . . ,

    1/2n ) exists and is real.

    ZTUTPUZ = ZTDZ = I.

    Now, A ZTUTQUZ is real symmetric, and so it can be

    factorized as A = WWT, where WTW = I and = diag(1, . . . , n).

    The transform V UZW simultaneously diagonalizes P and Q.

    VTPV = WTZTUTPUZW = WTIW = I

    VTQV = WTZTUTQUZW = WTAW =

    QV = (VT)1 = PV

    Arnab Ghoshal Stochastic Signal Processing

    Random sequences

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    22/43

    Random sequences

    A random sequence is a sequence of RVs. Formally, a random

    sequence X[n, ] is a mapping of the sample space into the

    space of real (or complex) valued sequences, such thatX

    [n

    , ]is a random variable for each integer n.

    A random sequence X[n] is statistically specifiedby its N-thorder distribution functions for all N 1, and for all times n:

    FX(xn,xn+1, . . . ,xn+N1; n, n + 1, . . . , n + N 1)= Pr{X[n] xn,X[n + 1] xn+1, . . . ,X[n + N 1] xn+N1}

    Similarly, the N-th order densities are obtained as:

    fX(

    xn,

    xn+1, . . . ,

    xn+N1;

    n,

    n+ 1, . . . ,

    n+

    N 1)

    =NFX(xn,xn+1, . . . ,xn+N1; n, n + 1, . . . , n + N 1)

    xnxn+1 . . . xn+N1

    Note that, for each order N, we need to specify an infinite

    number of PDFs for all times < n < +.Arnab Ghoshal Stochastic Signal Processing

    Moments of Random sequences

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    23/43

    q

    A random sequence can also be specified (in a weaker sense)

    by its moments:

    The mean functionof a random sequence at time n is:

    X[n] E{X[n]} =

    xn fX(xn; n) dx

    The (auto)correlation functionfor times m and n is:

    RXX[m, n] E{X[m]X[n]} =

    xmxn fX(xm,xn; m, n) dxm dxn

    Similarly, (auto)covariance KXX[m, n] is defined as thecorrelation of the centeredsequence Xc[n] = X[n] X[n]

    KXX[m, n] E{(X[m] X[m])(X[n] X[n])}

    Arnab Ghoshal Stochastic Signal Processing

    Properties of correlation functions

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    24/43

    p

    RXX[m, n] is Hermitian symmetric, i.e. RXX[m, n] = RXX[n, m]

    RXX[m, n] is Positive semidefinite, i.e. for any a1, . . . , aN,

    Nn=1

    Nm=1

    ana

    mRXX[m, n] 0, for all N > 0.

    The average powerof X[n] is given by RXX[n, n] = E{|X[n]|2}

    Diagonal dominance: |RXX[m, n]| RXX[m, m]RXX[n, n],which follows from the Cauchy-Schwarz inequality.

    Arnab Ghoshal Stochastic Signal Processing

    Stationary sequences

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    25/43

    y q

    A random sequence X[n] is said to be stationary if for all ordersN 1 and all shifts k,

    FX(xn,xn+1, . . . ,xn+N1; n, n + 1, . . . , n + N 1)

    = FX(xn,xn+1, . . . ,xn+N1; n + k, n + 1 + k, . . . , n + N 1 + k)

    A random sequence is called wide-sense stationary(WSS) if

    (1) The mean function is constant for all n, i.e X[n] = X[0](2) For all times m and n and all shifts k, the autocorrelation

    (autocovariance) function does not depend on k

    RXX[m, n] = RXX[m + k, n + k] KXX[m, n] = KXX[m + k, n + k]

    Properties of WSS sequences:

    1 Hermitian symmetric: RXX[m] = R

    XX[m]

    2 m, 0 |RXX[m]| RXX[0], and |RXY[m]|

    RXX[0]RYY[0]

    Arnab Ghoshal Stochastic Signal Processing

    WSS Random sequences and LTI systems

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    26/43

    q y

    For an LTI system L() and a WSS random sequence X[n], themean of the output random sequence Y[n] is

    E{Y[n]} = E{h[n] X[n]} = h[n] E{X[n]} = H(z)|z=1X,

    if the impulse response h[n] of L() is absolutely summable.

    The output cross-correlation function RXY[m] is given by:

    RXY[m] E{X[m]Y[0]}

    =

    k=h[k]E{X[m]X[k]}

    =

    k=

    h[k]RXX[m k]

    = h[m] RXX[m]

    Arnab Ghoshal Stochastic Signal Processing

    WSS Random sequences and LTI systems (contd.)

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    27/43

    q y

    The output auto-correlation function RYY[m] is given by:

    RYY[m]

    E{Y[m]Y

    [0]}

    =

    k=

    h[k]E{X[m k]Y[0]}

    =

    k=

    h[k]RXY[m k]

    = h[m] RXY[m]

    = h[m] (h[m] RXX[m])

    = (h[m] h[m]) RXX[m]

    = g[m] RXX[m]

    where g[m] h[m] h[m] is called the autocorrelation impulseresponse(AIR)

    Arnab Ghoshal Stochastic Signal Processing

    Power Spectral Density

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    28/43

    p y

    For a WSS random sequence X[n], the power spectral density(psd) is defined as the Fourier transform of its autocorrelation

    function:

    SXX()

    m=

    RXX[m]ejm, for +.

    Hence the autocorrelation can be computed as the IFT:

    RXX[m] =

    1

    2 S

    XX()ejm

    d.

    Cross-power spectral density between two jointly WSS random

    sequences X[n] and Y[n] is:

    SXY()

    m=

    RXY[m]ejm

    , for +.

    For an LTI system Y[n] = h[n] X[n] with WSS input X[n]:

    SYY = G()SXX() = |H()|2SXX().

    Arnab Ghoshal Stochastic Signal Processing

    Properties of power spectral density

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    29/43

    SXX()

    m=

    RXX[m]ejm, for +.

    1 SXX() is real valued for any X[n] since RXX[m] is Hermitiansymmetric

    2 SXX() is an even function of if X[n] is real-valued

    3 For any X[n], SXX() 0 for every 4 If RXX[m] has finite support(i.e. RXX[m] = 0, |m| > M for

    some finite M > 0) then SXX() is an analytic function in .5 Average power in X[n] can be computed as:

    E{|X[n]|2} = RXX[0] = 12

    SXX() d.

    Average power in a narrow band [0 , 0 + ] centered at0 is approximately SXX(0)

    Arnab Ghoshal Stochastic Signal Processing

    Vector random sequences

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    30/43

    A vector random sequence is a sequence of random vectors,such that for each event we have a random vector X[n, ].Let X[n] and Y[n] be the input and output of a first-orderLCCDE:

    Y[n] = AY[n 1] + BX[n]

    With zero initial-conditions, the response to X[n] is:

    Y[n] =n

    k=0AnkBX[k] h[n] X[n].

    where h[n] = AnB u[n] is the vector impulse response.

    Arnab Ghoshal Stochastic Signal Processing

    Vector random sequences (contd.)

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    31/43

    For WSS X[n] the cross-correlation matrices are obtained as:

    RXY[m] E{X[m]YH[0]} RYX[m] E{Y[m]X

    H[0]}

    = RXX[m] hH[m]. = h[m] RXX[m].

    The output autocorrelation matrix is:

    RYY[m] = h[m] RXX[m] hH[m].

    Taking the matrix Fourier transform, the psd is:

    SYY() = H()SXX()HH().

    Arnab Ghoshal Stochastic Signal Processing

    Convergence of series

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    32/43

    Convergence A sequence of numbers xn converges to x

    if > 0, N() such that n > N() |xn x| < .

    Cauchy criterion A sequence xn converges to a limit iff

    > 0, N() such that n, m > N() |xn xm| < .

    Uniform convergence A sequence of functionsfn : D converges uniformly to the function f(x) iff > 0, N() s.t. n > N() |fn(x) f(x)| < , x D.

    Pointwise convergence A sequence of functions fn(x)

    defined over the same domain D, converges pointwise tothe function f(x) iff > 0, and x D, N(,x) such thatn > N(,x) |fn(x) f(x)| < .

    Arnab Ghoshal Stochastic Signal Processing

    Convergence of random sequences

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    33/43

    Sure convergence The random sequence X[n]

    converges to the RV X if the sequence of functions X[n, ]converges to the function X() for every .

    Almost-sure convergence The random sequence X[n, ]converges almost surely (or with probability-1) to X() if

    limnX[n, ] = X(), A and Pr[A] = 1.

    Pr[{ : limn

    X[n, ] = X()}] = 1

    Mean-square convergence A random sequence X[n]converges to the RV X in mean-square sense if

    limnE{|X[n] X|2} = 0

    Arnab Ghoshal Stochastic Signal Processing

    Convergence of random sequences (contd.)

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    34/43

    Convergence in probability The probability

    Pr[|X[n] X| > ] is a sequence of numbers which depend

    on . If this sequence has a limit of 0 i.e.

    limn

    Pr[|X[n] X| > ] = 0,

    then the random sequence X[n, ] is said to converge to

    the RV X() in probability.

    Convergence in distribution A random sequence X[n]with probability distribution function Fn(x) converges indistribution to the random variable X with probability

    distribution F(x) if

    limn

    Fn(x) = F(x),

    at all x for which F is continuous.

    Arnab Ghoshal Stochastic Signal Processing

    Random processes

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    35/43

    A random process X(t, ) is a mapping from the sample space to the space of continuous time functions, such that X(t, ) isa random variable for each t (, +).

    X(t) E[X(t)], for all < t < +

    RXX(t1, t2) E[X(t1)X(t2)], for all < t1, t2 < +

    KXX(t1, t2) E[(X(t1) X(t1))(X(t2)

    X(t2))]

    KXX(t1, t2) = RXX(t1, t2) X(t1)X(t2)

    Variance function: 2X(t) KXX(t, t) and average powerfunction: E[|X(t)|2] = RXX(t, t)

    Hermitian symmetry: RXX(t1, t2) = R

    XX(t2, t1)

    Positive semi-definite: for all N > 0 and t1 < t2 < . . . < tN,

    and for any a1, . . . , aN:N

    i=1

    Nj=1 aia

    j RXX(ti, tj) 0.

    Diagonal dominance: |RXX(t, s)| RXX(t, t)RXX(s, s)Arnab Ghoshal Stochastic Signal Processing

    Random processes (contd.)

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    36/43

    X(t) and Y(t) are uncorrelatedif RXY(t1, t2) = X(t1)Y(t2)for all t1 and t2.

    X(t) and Y(t) are orthogonal if RXY(t1, t2) = 0 for all t1, t2.

    X(t) and Y(t) are independent if for all N > 0

    FXY(x1,y1,x2,y2, . . . ,xN,yN; t1, t2, . . . , tN)

    = FX(x1, . . . ,xN; t1, . . . , tN)FY(y1, . . . ,yN; t1, . . . , tN)

    X(t) is stationary if for all N > 0 and all times T,

    FX(x1, . . . ,xN; t1, . . . , tN) = FX(x1, . . . ,xN; t1 + T, . . . , tN + T)

    X(t) is called wide-sense stationary(WSS) if: E[X(t)] = X, a constant; and E[X(t+ )X

    (t)] = RXX()for all < < +, independent of the time parameter t.

    Arnab Ghoshal Stochastic Signal Processing

    WSS Random processes and LTI systems

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    37/43

    For an LTI system L() with impulse response h(t) and a WSSrandom process X(t), the mean of the output process Y(t) is

    E[Y(t)] =

    h()X(t ) d = XH(0),

    where H() is the frequency response of L.

    The output cross-correlation functions are given by:

    RXY() = h() RXX()

    RYX() = h() RXX()

    The output autocorrelation function is given by:

    RYY() = h() h() RXX()

    = g() RXX()

    where g() h() h() is the autocorrelation impulseresponse.

    Arnab Ghoshal Stochastic Signal Processing

    Power Spectral Density

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    38/43

    For a WSS random process X(t), the power spectral density(psd) is defined as the Fourier transform of its autocorrelation

    function:SXX()

    RXX()ej d

    The autocorrelation is the inverse Fourier transform of psd:

    RXX[m] = 12

    SXX()ej d.

    Cross-power spectral density between two jointly WSS random

    processes X(t) and Y(t) is:

    SXY()

    RXY()ej d.

    Arnab Ghoshal Stochastic Signal Processing

    Properties of power spectral density

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    39/43

    SXX()

    RXX()ej d

    1 SXX() is real valued since RXX() is Hermitian symmetric.

    2 SXX() is an even function of if X(t) is real-valued

    3 SXX() 0 for every

    4 Average power in X(t) is computed as:

    E{|X[n]|2} = RXX[0] =1

    2

    SXX() d.

    5

    Power in the frequency band (1, 2) is1

    221 SXX() d.

    6 For every real, non-negative and integrable function F(),there exists a stationary random process with power

    spectral density S() = F().

    Arnab Ghoshal Stochastic Signal Processing

    Periodic and cyclostationary processes

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    40/43

    A random process X(t) is wide-sense periodic if thereexists a T > 0 such that:

    X(t) = X(t+ T) for all t

    KXX(t1, t2) = KXX(t1 + T, t2) = KXX(t1, t2 + T) for all t1, t2.

    The smallest such T is called the period

    A random process X(t) is wide-sense cyclostationary ifthere exists a T > 0 such that:

    X(t) = X(t+ T) for all t

    KXX(t1, t2) = KXX(t1 + T, t2 + T) for all t1, t2.

    Arnab Ghoshal Stochastic Signal Processing

    Vector processes

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    41/43

    Let X(t) [X1(t),X2(t)]T

    , and Y(t) [Y1(t), Y2(t)]T

    be the inputand output of a general two-channel LTI system denoted by:

    h(t)

    h11(t) h12(t)h21(t) h22(t)

    .

    Then Y(t) = h(t) X(t) where vector convolution is defined as:

    (h(t) X(t))i N

    j=1hij(t) Xj(t).

    Arnab Ghoshal Stochastic Signal Processing

    Vector processes

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    42/43

    Defining the input and output correlation matrices, we get:

    RXX()

    RX1X1() RX1X2()RX2X1() RX2X2()

    RYY()

    RY1Y1() RY1Y2()RY2Y1() RY2Y2()

    ,

    RYY() = h() RXX() hH().

    Taking the matrix Fourier transform:

    SYY() = H()SXX()HH().

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/
  • 7/29/2019 Beamer Lecture

    43/43

    Arnab Ghoshal Stochastic Signal Processing

    http://find/http://goback/