Upload
priyanshu-agarwal
View
217
Download
0
Embed Size (px)
Citation preview
8/3/2019 HW1_Group_3
1/24
MAE 600: Homework 1Due on February 2nd, 2012
Instructor: Professor P. Singla
Priyanshu Agarwal (Contribution: 50 %)
Suren Kumar (Contribution: 50 %)
8/3/2019 HW1_Group_3
2/24
Contents
Problem 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
Problem 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
Problem 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Problem 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
(a) E[x] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11(b) E
(x x) (x x)T
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
(c)E(y y) (y y)
T . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14Problem 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
Problem 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
1
8/3/2019 HW1_Group_3
3/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1
Problem 1
Prove that Gaussian distribution function p(x) = 12
e(x)2
22 is a probability density function.
Solution 1: To prove that p(x) is a probability density function we need to prove,
1. p(x) 02.
p(x)dx = 1
Proving (1) is trivial because exponential raised to any power will result in a positive quantity and variance
is a positive quantity.
p(x)dx =
1
2e
(x)2
22 dx (1)
=1
2
e(x)2
22 dx (2)
Let y = x2
, which implies dy = dx2
. Substituting y and dy in Equation 2, we get
p(x)dx =1
ey2
dy (3)
Integral in Equation 3 is in form of a gaussian integral. Consider e
y2dy,
ey2
dy2
=
ey2
dy
ez2
dz (4)
=
e(y2+z2)dzdy (5)
=
e(y2+z2)dA (6)
where z is a dummy variable and dA = dzdy is the area of a small element in Y Z plane. PhysicallyEquation 6 refers to area of complete Y Z plane. This area could also be obtained in terms of polarcoordinates (r, ) when r [0, ), [0, 2]. Original coordinates (y, z) can be expressed in terms of polarcoordinates as
y = r cos()
z = r sin()
y2 + z2 = r2
Also in terms of radial coordinates, area of a small strip can be written as dA = rdrd. Equation 6 can nowbe written as
ey2
dy2
=
20
0
rer2
drd
=
20
12
er20
d
Problem 1 continued on next page. . . Page 2 of 23
8/3/2019 HW1_Group_3
4/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 1 (continued)
=
20
1
2d
=1
2|20
=
This implies that ey2dy = . By substituting this result in Equation 3, we get,
p(x)dx = 1
Hence, Guassian density function is a valid probability density function.
Page 3 of 23
8/3/2019 HW1_Group_3
5/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 1
Problem 2
The number =1
3E(x )3 is called the skewness of random variable, x. Show that Poisson distribution
has =1
.
Solution: 2 The probability mass function for Poisson distribution of a discrete random variable is givenby
p(x, ) =xe
x!(7)
where x Z (set of non-negative integers) and is a positive real number.The mean of x is given by
= E[x]
=
x=0
xxe
x!
= ex=0
x x
x!
= e
0 + + 22
2!+ 3
3
3!+ ...
= e
1 +
1!+
2
2!+ ...
From Taylor series expansion of e
e = 1 +
1!+
2
2!+ ...
So,
= ee
= (8)
Now, the variance of x is given by
2 = E
(x )2
= Ex2 + 2 2x
= Ex2 + E2 E[2x] ( E[x + y] = E[x] + E[y])= Ex2 + 2 2E[x] ( E[c] = c, E[cx] = cE[x], where c is a constant)= Ex2 + 2 22 ( E[x] = )
2 = Ex2 2 (9)From Eqn.s 8 and 9, we get
2 = Ex2 2 (10)Problem 2 continued on next page. . . Page 4 of 23
8/3/2019 HW1_Group_3
6/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 2 (continued)
Now by definition E[x2] is given by
E[x2] =
x=0x2p(x, )
=x=0
x2xe
x!
= ex=0
x2x
x!
= e
0 + 12 + 222
2!+ 32
3
3!+ ...
E[x2] = e
1 + 2
1!+ 3
2
2!+ ...
(11)
Now, using the Taylor series expansion of e, we get
e = 1 +
1!+
2
2!+ ...
Multiplying both the sides by , we get
e = +2
1!+
3
2!+
4
3!+ ...
Differentiating both sides w.r.t , we get
e
+ e
= 1 + 2
1! + 3
2
2! + 4
3
3! + ...
( + 1) e = 1 + 2
1!+ 3
2
2!+ 4
3
3!+ ... (12)
From Eqn.s 11 and 12, we get
E[x2] = e ( + 1) eE[x2] = ( + 1) (13)
From Eqn. 10 and 13, we get
2 = ( + 1) 22 =
=
(14)
Now E[(x )3] is given by
Problem 2 continued on next page. . . Page 5 of 23
8/3/2019 HW1_Group_3
7/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 2 (continued)
E
(x )3
= Ex3 3 3x (x )= Ex3 E3 3Ex2 + E3x2
E(x )3
= Ex3 3 3Ex2 + 33 ( E[x] = , E[c] = c, E[cx] = cE[x], where c is a constant)(15)
From Eqn.s 8, 13, and 15, we get
E
(x )3
= Ex3 3 3Ex2 + 33 (16)Now by definition E[x3] is given by
Ex3 = x=0
x3p(x, )
=x=0
x3 xe
x!
= ex=0
x3x
x!
= e
0 + 13 + 232
2!+ 33
3
3!+ ...
E[x3] = e
1 + 22
1!+ 32
2
2!+ ...
(17)
Now, to evaluate the series in Eqn. 17, multiply Eqn. 12 on both sides by
( + 1) e = + 22
1!+ 3
3
2!+ 4
4
3!+ ..
Differentiating both sides w.r.t , we get
( + 1) e +
e + e + e
= + 22
1!+ 3
3
2!+ 4
4
3!+ ..
2 + 3 + 1
e = + 2
2
1!+ 3
3
2!+ 4
4
3!+ .. (18)
From Eqn.s 17 and 18, we get
E[x3] = ee 2 + 3 + 1E[x3] = 3 + 32 + (19)
Now, from Eqn.s 13, 19 and 16, we get
Problem 2 continued on next page. . . Page 6 of 23
8/3/2019 HW1_Group_3
8/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 2 (continued)
E
(x )3
= 3 + 32 + 3 3.. ( + 1) + 33
= 3 + 32 + 3 33 32 + 33
E(x )3
= (20)Now, skewness of a random variable is defined as
=1
3E(x )3 (21)
So, from Eqn.s 14, 20 and 21, we get
=
32
=1
(22)
From Eqn. 8 and 22, we get
=1
Hence, skewness of random variable with Poisson distribution is =1
.
Page 7 of 23
8/3/2019 HW1_Group_3
9/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 2
Problem 3
Show that the equal probability contours for a multi-variate Gaussian density function will be hyper-ellipsoid
surfaces. Also, draw equal probability contours (without using contour command in MATLAB) for a nor-
mally distributed 2-dimensional random vector with zero mean and following covariance matrix:
P =
4 1
1 4
Solution 3: N-dimensional gaussian density function is given by
p(x) =1
(2)n
2 || 12 exp(1
2[x ]T1[x ])
where is NN covariance matrix and is n-dimensional mean vector. Equal probability density contourscan be obtained by finding xp(x) = k, where k is a constant. To prove that equal probability densitycontours for a multi-variate gaussian is hyper-ellipsoid surface,
k =1
(2)n
2 || 12 exp(1
2 [x ]T1[x ])
k(2)n
2 || 12 = exp(12
[x ]T1[x ])
Taking log on both sides, we get
ln(k(2)n
2 || 12 ) = (12
[x ]T1[x ]) (23)2ln(k(2)n2 || 12 ) = [x ]T1[x ] (24)
LHS of Equation 24 is a constant and futhermore let it be equal to k1.
k1
= [x ]T
1
[x ] (25)Equation 25 is equation of a general hyper-ellipsoid in N-dimensional space given that is a positive
definite matrix. It can be further demonstrated using the eigen decomposition of the positive definite matrix
1 = VDVT, where D is the diagonal matrix containing eigenvalues and V is the orthonormal matrix
containing eigenvectors on its columns. Also the diagonal entries of D are positive. V is in form of a rotation
matrix. Let Y be a new vector defined as Y = VT[x ] . Substituting this result in Equation 25,
k1 = YTDY (26)
1 =i=Ni=1
i
k1yi2 (27)
Since each eigenvalue of a positive definited symmetric matrix is greater than zero, Equation 27 is an hyper-ellipsoid in n-dimensional space which is rotated with respect to original space. So it will be hyperellipsoid
in the original space as well.
Problem 3 continued on next page. . . Page 8 of 23
8/3/2019 HW1_Group_3
10/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 3 (continued)
For the current case after substituting and in Equation 25, we get
k1 =1
15
x y
4 11 4
x
y
(28)
15k1 = 4x2 2xy + 4y2 (29)
Equation 27 can be compared with equation of a general ellipse centered at ( x, y) = (0, 0), Ax2 + Bxy + Cy2,to get the angle of inclination with x-axis as tan 2 = B
CA . In current case the angle with x-axis is 450
degrees. Applying the rotation and getting the ellipse equation in rotated coordinates ( x1, y1),we get
x = x1 cos() y1 sin()x =
x12
y12
y = x1 sin() + y1 cos()
y =x1
2+
y12
15k1 = 3x21 + 5y21
Equation of ellipse in (x1, y1) coordinates in parametric form can be written as (x1, y1) = (5k1 cos(), 3k1 sin())as is varied from 0 to 2 Listing 1 shows a MATLAB script that was used to generate Figure 1.
6 4 2 0 2 4 66
4
2
0
2
4
6
X axis
Ya
xis
Equal Probability density Contours for Guassian Pdf
p(x) = 0.001
p(x) = 0.005
p(x) = 0.01
Figure 1: The plot shows that for large values of k the function f(k) converges to the value of 2
Listing 1: MATLAB code Used to Generate Figure 1
% Drawing a hyperellipsoid using equal probability contour of density
c l c ;c le ar a l l ; c lo se a l l ;
Problem 3 continued on next page. . . Page 9 of 23
8/3/2019 HW1_Group_3
11/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 3 (continued)
% Equal Probability density contour value
k = [0.001 0.005 0.01];
5 types = {-rs,--bo,:k+};
detE = 15;
fo r i = 1 :length(k)
k1 = -2*lo g(k(i)*(2*pi )*(detE(0.5)))*detE;
% Getting 360 datapoints to plot a ellipse10 data = zeros(360,2);
counter = 0;
fo r psi=0:pi /360:2*pi
x1 = sqrt(k1/3)*s in (psi);
y1 = sqrt(k1/5)*co s(psi);
15 counter = counter+1;
data(counter,:)=[(x1-y1)/sqrt(2),(x1+y1)/sqrt(2)];
end
%Plot the data
plot(data(:,1),data(:,2),types{i});
20 hold on;
end
xlabel(X axis \rightarrow);ylabel(Y axis \rightarrow);
t i t l e (Equal Probability density Contours for Guassian Pdf);
25 legend(p(x) = 0.001,p(x) = 0.005,p(x) = 0.01);
Page 10 of 23
8/3/2019 HW1_Group_3
12/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 3
Problem 4
Consider the following linear model:
y = Hx + (30)
where, y Rm, x Rn and is a Gaussian white noise vector with zero mean and covariance R. Giventhat x =
HTR1H
1HTR1y. Derive the relations for the following:
(a) E[x]
Solution: 4(a)
Using the given expression for x, we get
E[x] = E
HTR1H
1
HTR1y
=
HTR1H1 HTR1E[y] ( E[cx] = cE[x], where c is a constant)
Substituting the expression for y from the given linear model (Eqn. 30)
E[x] = HTR1H1 HTR1E[Hx + ]=
HTR1H1
HTR1 (HE[x] + E[])=
HTR1H1
HTR1HE[x] ( noise is zero mean, so E[] = 0)=
HTR1H1
HTR1Hx ( x is the true value which is treated as a constant, so E[x] = x)E[x] = x H
TR1H1
HTR1H = I (31)
Problem 4 continued on next page. . . Page 11 of 23
8/3/2019 HW1_Group_3
13/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 4 [(b) E
(x x) (x x)T
]
(b) E
(x x) (x x)T
Solution: 4(b)
E(x x) (x x)T
= E(x x) xT xT
= ExxT xx xxT + xxT= ExxT ExxT ExxT + ExxT= xxT xExT E[x] xT + ExxT (32)( x is the true value which is treated as a constant, so E[x] = x)
also, E[xxT] = E[x]xT, E[xxT] = xE[xT]
Now from Eqn.s 31 and 36, we get
E
(x x) (x x)T
= xxT xxT xxT + E
xxT
E(x x) (x x)T = xxT + ExxT (33)Now, let us consider ExxT and use the expression given for x.
ExxT = EHTR1H1 HTR1y HTR1H1 HTR1yT
ExxT = EHTR1H1 HTR1yyTRTH HTRTHTExxT = HTR1H1 HTR1EyyTRTH HTRTHT (34)
Now, let us consider EyyT and use the linear model Eqn. 30.EyyT = E(Hx + ) (Hx + )T
= E(Hx + ) xTHT + T= EHxxTHT + HxT + xTHT + T= HxxTHT + HxET + E[] xTHT + ET( x is the true value which is treated as a constant, so E[x] = x)
also, E[HxxTHT] = HxxTHT, EHxT = HxET , E[xxT] = xE[xT], ExTHT = E[] xTHTEyy
T
= HxxTHT + R ( noise is zero mean, so E[] = 0) (35)
Now, from Eqn.s 34 and 35, we get
ExxT = HTR1H1 HTR1 HxxTHT + RRTH HTRTHT
Problem 4 [(b) E
(x x) (x x)T
] continued on next page. . . Page 12 of 23
8/3/2019 HW1_Group_3
14/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 4
ExxT = HTR1H1 HTR1HxxTHTRTH HTRTHT +HTR1H
1HTR1RRTH
HTRTH
T= xxT +
HTR1H
1HTRTH
HTRTH
T
HTR1H1
HTR1H = I
= xxT + HTR1H1 HTR1H HTR1H1 RT = RExxT = xxT + HTR1H1 (36)
Now, from Eqn.s 33 and 36, we get
E
(x x) (x x)T
= xxT + xxT + HTR1H1E
(x x) (x x)T
=
HTR1H1
(37)
Problem 4 continued on next page. . . Page 13 of 23
8/3/2019 HW1_Group_3
15/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 4 [(c) E
(y y) (y y)T
]
(c) E
(y y) (y y)T
Solution: 4(c)
Here y = Hx
E
(y y) (y y)T
= E(y y) yT yT= EyyT yyT yyT + yyT= EyyT EyyT EyyT + EyyT (38)
Now, from Eqn.s 35, 38 and the given linear model (Eqn. 30), we get
E
(y y) (y y)T
= EyyT EyyT EyyT + EyyT= HxxTHT + R E(Hx + ) x
THT EHx (Hx + )T
+ EHxxTHT
= HxxTHT + R EHxxTHT + xTHT EHxxTHT + HxT + HExxTHTE
(y y) (y y)T
= HxxTHT + R HxExTHT ExTHT HE[x] xTHT HExT + HExxTHT(39)
( x is the true value which is treated as a constant, so E[x] = x)
Now, from Eqn.s 31, 34 and 39 and the given expression for x, we get
E
(y y) (y y)T
= HxxTHT + R HxExTHT ExTHT HE[x] xTHT HExT + HExxTHT= HxxTHT + R
HxxTHT
EyTRTH HTR1HTHT Hxx
THT
HE
HTR1H
1HTR1yT
+ HxxTHT + H
HTR1H
1HT
= HxxTHT + R HxxTHT EyTRTH HTR1HT HT HxxTHTH
HTR1H1
HTR1EyT + HxxTHT + H HTR1H1 HT(40)
Now, again using the given linear model (Eqn. 30), we get
E
(y y) (y y)T
= HxxTHT + R HxxTHT E
(Hx + )
T
RTH
HTR1H
T
HT HxxTHT
H
HTR1H1 HTR1E(Hx + ) T + HxxTHT + H HTR1H1 HT
= R ExTHT + TRTH HTR1HT HTH
HTR1H1
HTR1EHxT + T + H HTR1H1 HT= R E[] xTHT ETRTH HTR1HT HT
H
HTR1H1
HTR1
HxET + ET + H HTR1H1 HT
Problem 4 [(c) E(y y) (y y)T
] continued on next page. . . Page 14 of 23
8/3/2019 HW1_Group_3
16/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 4
= R RRTH HTR1HT HT H HTR1H1 HTR1R + H HTR1H1 HT( noise is zero mean, so E[] = 0)
= R
RRTH HTR1HT HT
H HTR1H1 HT + H HTR1H1 HT( noise is zero mean, so E[] = 0)
= R RRTH HTR1HT HT RT = R1
= R H HTR1HT HT RT = R1
= R H HTR1H1 HT= R HTHTR1HH11= 0
Page 15 of 23
8/3/2019 HW1_Group_3
17/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 4
Problem 5
Let us consider the problem of choosing an appropriate radar system for precise position tracking of an
airplane. You are supposed to estimate the range of an airplane which must have an error variance of less
than or equal to 50m2 using two radar stations measurements:
y1 = r + 1
y2 = r + 2
where r is the true range of the aircraft. It is given that one radar station has been built such that 1 is
zero-mean Gaussian noise with variance of 100m2. The other Radar station has not yet been built, and 2can be assumed to be zero-mean Gaussian noise with variance of Rm2, where R is a design parameter. Since
accurate radar station cost money, hence your objective is to find the maximum value of R that is acceptable.
Furthermore assume that a priori knowledge of r indicates that it is well modeled as zero-mean Gaussian
and have a variance of 1km2. Once again, find the maximum value of R that is acceptable and discuss how
the a priori knowledge affect your solution for R. With that R find the equations for an estimator that
incorporates both y1 and y2 simultaneously. Also find the equations for an estimator that incorporates y1
and y2 recursively. Show that these are the same estimates with the same variance.
Solution : Consider a combined model for two observationy1y2
=
1
1
r +
12
(41)
Y = Hr + , x = r (42)
For this problem, there is no prior information available and hence we choose maximum likelihood solution as
the best estimate, which is given by x = (HTR1H)1
HTR1y. R is covariance matrix of the system which
is given by E(T) because has zero mean. Also both the observations and their errors are uncorrelated.Hence R can be determined as
R =
E(1T1 ) 00 E(2T2 )
R =
100 0
0 R
Error covariance (which will be variance in current case because their is only one true quantity range ) is
given by
E((x x)(x x)T) = E((x (HTR1H)1HTR1Y)(x (HTR1H)1HTR1Y)T)
Substituting Y from Equation 42, and noting that (HT
R1H)1
(HT
R1H) = I, we get
E((x x)(x x)T) = (HTR1H)1HTR1E(T)(HTR1)T(HTR1H)T
E(T) = RE((x x)(x x)T) = (HTR1H)1
E((x x)(x x)T) = (1 1 1100 00 1
R
1
1
)1
Problem 5 continued on next page. . . Page 16 of 23
8/3/2019 HW1_Group_3
18/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 5 (continued)
E((x x)(x x)T) = ( 1100
+1
R)1
As a design requirement we are given that its error variance should be less than or equal to 50 m2.
(1
100+
1
R)1 50
1
100+
1
R 1
501
R 1
100
R 100
Hence the maximum acceptable value of R is 100m2.
In case of apriori information, we can use the maximum a posteriori solution , in which case the esti-
mate is given by x = (HTR1H + Q1)1
(HTR1Y + Q1xa). In current case, the prior is given as zero
mean (xa = 0) and with variance of 1km2. Error covariance (which will be variance in current case because
their is only one true quantity range ) is given by
E((x x)(x x)T) = E((x (HTR1H + Q1)1HTR1Y)(x (HTR1H + Q1)1HTR1Y)T)(x x) = x (HTR1H + Q1)1(HTR1Y)
= x (HTR1H + Q1)1(HTR1Hx + HTR1)= x (HTR1H + Q1)1((HTR1H + Q1)x Q1x + HTR1)= (HTR1H + Q1)
1Q1x (HTR1H + Q1)1HTR1)
E((x x)(x x)T) = E((HTR1H + Q1)1Q1x (HTR1H + Q1)1HTR1)(HTR1H + Q1)
1Q1x (HTR1H + Q1)1HTR1)T)
Noting that x, are uncorrelated we can simplify this expression further by using the linear property ofexpectation operator.
E((x x)(x x)T) = (HTR1H + Q1)1E(Q1xxTQT + HTR1TRTH)(HTR1H + Q1)T
Noting that E(xxT) = Q (as given by the prior with zero mean and Q covariance), E()T = R (by definition)and x, are uncorrelated, we can further simplify this expression as
E((x x)(x x)T) = (HTR1H + Q1)1(QT + HTRTH)(HTR1H + Q1)T
= (HTR1H + Q1)1
= (1 1 1100 0
01R 1
1 + 1
106
)1
Problem 5 continued on next page. . . Page 17 of 23
8/3/2019 HW1_Group_3
19/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 5 (continued)
As a design requirement we are given that final error variance should be less than or equal to 50 m2.
(1
100+
1
R+
1
106)1 50
1
100+
1
R+
1
106 1
50
1R
9999106
R 100.010001
Hence the maximum acceptable value of R is 100.01m2. We can see from the results that even when prior
has comparatively high variance, error variance after including it is reduced and this allows us to choose a
greater R design value. So unless the variance of prior information is , it always improves the covarianceof estimate. In this case, the prior information allowed us to choose a bigger value of variance for one of our
observation to result in same error variance.
To design an estimator incorporating y1, y2 simultaneously, lets assume the form of estimator as x = MY +n.
For this estimator to be unbiased (
E(x) = x), we get the constraints as MH = I, n = 0. Designing the
estimation to have minimum variance, we get
E((x x)(x x)T) = E(xxT) E(xxT)E(T) = RE(xT) = E(Tx) = 0 x, are uncorrelatedE(xxT) = MHE(xxT)HTMT + MRMT
= E(xxT) + MRMTE((x x)(x x)T) = MRMT
This is a constrained optimization problem which can be solved using lagrange multipliers. Loss function
can be formed as
J =1
2T r(MRMT) + T r((I MH))
MJ = MR THT = 0M = THTR1
J = I MH = 0T = (HTR1H)1
M = (HTR1H)1HTR1
Error variance of this estimator is given by
E((x
x)(x
x)T) = MRMT = (HTR1H)1. Substituting
Problem 5 continued on next page. . . Page 18 of 23
8/3/2019 HW1_Group_3
20/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 5 (continued)
values for current case
H =
1
1
R = 100 0
0 RE((x x)(x x)T) = (HTR1H)1
= (1
R+
1
100)1
= 50m2
Hence error variance of sequential estimator is 50m2.
To design an estimator that incorporates y1, y2 sequentially, consider the information given by first
observation as prior for second observation. The form of our system in now Y = r + 2 with H matrix being
equal to 1 and R matrix being equal to
E(22
T). A priori estimate of true state x is available as xa = x + w
and associated a priori error covariance matrix is given by E(wwT) = Q. Since in first observation equation1 has zero mean ,w also has zero mean. Also 1 and 2 are uncorrelated which implies that E(wT2 ) = 0Consider the form of estimator as x = MY + Nxa + n. For perfect observation (2 = 0), we have our
observation model as Y = Hx. If we also assume perfect a priori estimates, (xa = x, w = 0), our estimator
should give us perfect results. Using these conditions, we get
x = MHx + Nx + n
x = (MH + N)x + n
This gives us two constraints MH + N = I and n = 0. Thus our desired estimator has the form x =
MY + Nxa. Error in the estimate is given by
x x = MY + Nxa x= MHx + M + Nx + Nw x= (MH + N)x + M + Nw x= M + Nw
Our aim is to minimize the variance of error given by
E((x x)(x x)T) = E((M + Nw)(M + Nw)T)= ME(T)MT + NE(wwT)NT= MRMT + NQNT
Variance of error needs to be minimized and the final problem can be formed as a lagrange multiplier based
Problem 5 continued on next page. . . Page 19 of 23
8/3/2019 HW1_Group_3
21/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 5 (continued)
minimization due to presence of constraints. Loss function can be formulated as
J =1
2T r(MRMT + NQNT) + T r((I MH + N))
MJ = MR THT = 0M = THTR1
NJ = NQ T = 0N = TQ1
J = I MH + N = 0T = (HTR1H + Q1)1
M = (HTR1H + Q1)1HTR1
N = (HTR1H + Q1)1Q1
Error variance estimate is given by E((x x)(x x)T) = MRMT + NQNT. Substituting values for currentcase where H = 1, R = R, Q = 100m2,we get
M = (1
R+
1
100)1
1
R
N = (1
R+
1
100)1
1
100
MRMT + NQNT = (1
R+
1
100)2(
1
R+
1
100)
= (1
R+
1
100)1
R = 100
E((x x)(x x)T) = 50m2
This implies that error variance of both sequential and combined estimator is same.
Page 20 of 23
8/3/2019 HW1_Group_3
22/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 5
Problem 6
Let us consider the problem of estimation of a constant vector x Rn based upon m measurements givenby the following model:
y = Hx + (43)
Let x and be modeled as jointly Gaussian, zero-mean vectors with
E[xxT] = X, E[T] = R, E[xT] = S (44)
Your task is to find the conditional mean E[x|y] and corresponding conditional error covariance.
Solution: 6
x and are given to be jointly normal with zero-mean. Since, any linear combination of jointly Gaussian
random variables is also jointly Gaussian, x and y are also jointly Gaussian.
The mean of y using the given model (Eqn. 43) is given by
E[y] = E[Hx + ]= HE[x] + E[]= 0 ( E[x] = 0, E[] = 0)
So, the joint distribution of x and y can be written as
p(x, y) =1
(2)m+n2 det(P)
12
e1
2[xT yT]P1
x
y
(45)
where
P =
E[xxT] E[xyT]E[yxT] E[yyT]
, E[xyT] = E[yxT]T
Now, let us evaluate E[xyT] using the given model (Eqn. 43)
E[xyT] = E[x (Hx + )T]= ExxTHT + xT= ExxTHT + ExT= XHT + S (46)
Now, evaluate E[yyT] again using the given model (Eqn. 43)
E[yyT] = E[(Hx + ) (Hx + )T]= E[HxxTHT + HxT + xTHT + T]= HE[xxT]HT + HE[xT] + E[xT]HT + E[T]= HXHT + HS + STHT + R (47)
Problem 6 continued on next page. . . Page 21 of 23
8/3/2019 HW1_Group_3
23/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 6 (continued)
So, P is given by
P =
X XHT + S
XHT + S
T
HXHT + HS + STHT + R
,
Now, let P is given by
P =
Px Pxy
PxyT Py
,
So, P1 using the block matrix inverse from matrix cookbook can be written as
P1 =
Px PxyPy1PxyT
1
Px PxyPy1PxyT1
PxyPy1
Py1PxyT
Px PxyPy1PxyT1
Py1 + Py
1PxyT
Px PxyPy1PxyT1
PxyPy1
,
Now,
[xT yT]
Px PxyPy1PxyT
1
Px PxyPy1PxyT1
PxyPy1
Py1PxyT
Px PxyPy1PxyT1
Py1 + Py
1PxyT
Px PxyPy1PxyT1
PxyPy1
x
y
= [xT
Px PxyPy1PxyT1
yTPy1PxyT
Px PxyPy1PxyT1
xT Px PxyPy1PTxy1 PxyPx1 + yT
Py1 + Py
1PxyT
Px PxyPy1PxyT1
PxyPy1
]
x
y
= xT
Px PxyPy1PxyT
1
x yTPy1PxyT
Px PxyPy1PxyT1
x
xT Px PxyPy1PTxy1 PxyPx1y + yTPy1 + Py1PxyT Px PxyPy1PxyT1 PxyPy1y= yTPy
1y + xT
Px PxyPy1PxyT1
x yTPy1PxyT
Px PxyPy1PxyT1
x
xT Px PxyPy1PTxy1 PxyPx1y + yTPy1PxyT Px PxyPy1PxyT1 PxyPy1y= yTPy
1y + xT
Px PxyPy1PxyT1
x PxyPy1y
yTPy1Pxy
T
Px PxyPy1PxyT1
x PxyPy1y
= yTPy1y +
xT yTPy1PTxy
Px PxyPy1PxyT1
x PxyPy1y
= y
T
Py1
y +
x PxyPy1
yT
Px PxyPy1
PxyT1 x PxyPy1y
So, the joint distribution of x and y becomes
p(x, y) =1
(2)m+n2 det(P)
12
e1
2
yTPy
1y +
x PxyPy1yT
Px PxyPy1PxyT1
x PxyPy1y
Problem 6 continued on next page. . . Page 22 of 23
8/3/2019 HW1_Group_3
24/24
Priyanshu Agarwal, Suren Kumar MAE 600 : Homework 1 Problem 6 (continued)
det(P) = det(Py)det(Px PxyPy1PxyT) (48)
p(x, y) =1
(2)m+n2 det(Py)
12 det(Px PxyPy1PxyT) 12
e1
2
yTPy
1y +
x PxyPy1yT
Px PxyPy1PxyT1
x PxyPy1y
(49)
Now,
p(x|y) = p(x, y)p(y)
= 1
(2)n
2 det(Px PxyPy1PxyT) 12e
1
2
x PxyPy1
yT
Px PxyPy1
PxyT1
x PxyPy1
y
(50)
Hence, the mean and covariance of the conditional distribution is given by
E[x|y] = PxyPy1yE[xxT|y] = Px PxyPy1PxyT
where
Px = XPxy = XH
T + S
Py = HXHT + HS + STHT + R