Machine Learning Chapter 4 Solutions Problems Show That The Set Equations Has Unique Solution And

subject Type Homework Help
subject Pages 11
subject Words 3558
subject Authors Sergios Theodoridis

Unlock document.

This document is partially blurred.
Unlock all pages and 1 million more documents.
Get Access
page-pf1
1
Solutions To Problems of Chapter 4
4.1. Show that the set of equations
Σθ=p
has a unique solution if Σ > 0 and infinite many if Σis singular.
Solution: a) Let Σ > 0. Then the linear system of equations has a unique
solution. The converse is also true. Let the linear system has a unique
4.2. Show that the set of equations
Σθ=p
has always a solution.
Solution: The existence of solution when Σ > 0 is obvious. Let Σbe singu-
lar. Then, in order to guarantee a solution, we have to show that plies in
the range space of Σ. For this, it suffices to show that pa,a N (Σ).
page-pf2
2
4.3. Show that the shape of the isovalue contours of the mean-square error
(J(θ)) surface,
J(θ) = J(θ)+(θθ)TΣ(θθ),
are ellipses whose axes depend on the eigenstructure of Σ.
Hint: Assume that Σhas discrete eigenvalues.
Solution: Since Σis symmetric, we know that it can be diagonalized,
Σ=QΛQT,
4.4. Prove that if the true relation between the input xand the true output y
is linear, i.e.,
y = θT
ox+ v,θoRl
where v is independent of x, then the optimal MSE θsatisfies
θ=θo.
Solution: The optimal parameter vector is given by
Σθ=E[xy] = E[xxTθo+ v]
=Σθo,
page-pf3
3
4.5. Show that if
y = θT
ox+ v,θoRk
where v is independent of x, then the optimal MSE θRl, l < k is equal
to the top lcomponents of θo, if the components of xare uncorrelated.
Solution: Let
θo=θ1
o
θ2
o,θ1
oRl,θ2
oRkl.
4.6. Derive the normal equations by minimizing the cost in (4.15).
Hint: Express the cost in terms of the real part θrand its imaginary part
θiof θand optimize with respect to θr,θi.
Solution: The cost function is
J(θ) := E|yθHx|2.
page-pf4
4
where
θ=θr
θi,x=xr
xi,˜
x=xi
xr.(3)
The cost in (2) can now be written as
4.7. Consider the multichannel filtering task
ˆ
y=ˆyr
ˆyi= Θ xr
xi.
Estimate Θ so that to minimize the error norm:
E[||yˆ
y||2].
Solution: The error norm is equal to
J(Θ) = Eyrˆyr2+Eyiˆyi2
=EyrθT
11xrθT
12xi2+EyiθT
21xrθT
22xi2,
page-pf5
5
4.8. Show that (4.34) is the same as (4.25).
Solution: By the respective definitions, we have
ˆyr+jˆyi= (θT
rjθT
i)(xr+jxi) +
4.9. Show that the MSE achieved by a linear complex-valued estimator is al-
ways larger than that obtained by a widely linear one. Equality is achieved
only under the circularity conditions.
Solution: The minimum MSE for the linear filter is
MSEl=E[(y θH
x)(yxHθ)]
=E[y2] + θH
E[xxH]θθH
E[xy]E[yxH]θ
page-pf6
6
4.10. Show that under the second order circularity assumption, the conditions
in (4.39) hold true.
Solution: By the second order circularity condition we have
E[xxT]=0,
4.11. Show that if
f:CR,
then the Cauchy-Riemann conditions are violated.
Proof: Let
f(x+jy) = u(x, y)R.
Then by assumption, the imaginary part v(x, y) is identically zero. Hence
the Cauchy-Riemann conditions, i.e.,
4.12. Derive the optimality condition in (4.45).
Solution: We will show that any other filter, hi, i Z, results in a larger
MSE compared to the filter wi, i Z, which satisfies the condition. In-
deed, we have that
A:= E(dnX
i
hiuni)2=E(dnX
i
(hiwi+wi)uni)2.
j
i
page-pf7
7
4.13. Show Equations (4.50) and (4.51).
Solution:
a) Eq. (4.51): We know from Chapter 2 that the power spectrum of the
output process, y(n, m) is related to that of the input, d(n, m), by
b) Eq. (4.50): By the respective definition we have that,
rdu(k, l) = Ed(n, m)u(nk, m l),
and also
+
X
+
X
i0=−∞
j0=−∞
page-pf8
8
4.14. Derive the normal equations for Example 4.2.
Solution: We have
E[unun] = E0.5sn+ sn1+ηn0.5sn+ sn1
+ηn
= 0.25rs(0) + rs(0) + rη(0)
Similarly,
E[undn] = Eunsn1
4.15. The input to the channel is a white noise sequence snof variance σ2
s. The
output of the channel is the AR processes
yn=a1yn1+ sn.(5)
The channel also adds white noise ηnof variance σ2
η. Design an optimal
equalizer of order two, which at its output recovers an approximation of
snL. Sometimes, this equalization task is also known as whitening, since
in this case the action of the equalizer is to “whiten” the AR process.
Solution: The input to the equalizer is
un= yn+ηn.
page-pf9
9
Thus, the elements of the input covariance/autocorrelation matrix are
given by
ru(0) = E[unun] = E[(yn+ηn)(yn+ηn)]
s
1a2
1
1a2
1
η
4.16. Show that the forward and backward MSE optimal predictors are conju-
gate reverse of each other.
Solution: By the respective definitions we have
am=Σ1
mr
m
page-pfa
10
Since Σmis a Hermitian Toeplitz matrix, it is easily checked out that
4.17. Show that the MSE prediction errors (αf
m=αb
m) are updated according
to the recursion
αb
m=αb
m1(1 − |κm1|2).
Solution: By the respective definition we have
αb
m=r(0) rH
mJmΣ1
mJmrm=r(0) rH
mJmbm=r(0) rH
ma
m
=r(0) rH
m1r(m)(a
m1
0+b
m1
1κ
m1)
4.18. Derive the BLUE in the Gauss-Markov theorem.
Solution: The optimization task is
H:= arg minHtrace{HΣηHT},
.
.
hT
l
page-pfb
11
Observe that
trace{HΣηHT}=
l
X
i=1
hT
iΣηhi.(6)
page-pfc
12
4.19. Show that the mean-square error (which in this case coincides with the
variance of the estimator) of any linear unbiased estimator is higher than
that associated with the BLUE.
Solution: The mean-square error of any His given by
MSE(H) = trace{HΣηHT}.
However,
Hence,
HΣηHT
=HΣηΣ1
ηX(XTΣ1
ηX)1= (XTΣ1
ηX)1.
page-pfd
13
4.20. Show that if Σηis positive definite, then XTΣ1
ηXis also positive definite
if Xis full rank.
Solution: Recall from linear algebra that if Xis full rank, then XTX
is positive definite and vice versa. Indeed, assume that this is not the
case. Then, there will be a6=0, such that
XTXa=0,
4.21. Derive a MSE optimal linearly constrained widely linear beamformer.
Solution: The output of the widely linear beamformer is given by
ˆs(t) = wHu(t) + vHu(t)
page-pfe
14
The error signal is
Let us write the two constraints in a more compact form to facilitate the
optimization. To this end, we have that
XH˜
we=1
0,
where
page-pff
15
and finally
4.22. Prove that the Kalman gain that minimizes the error variance matrix
Pn|n=E[(xnˆ
xn|n)(xnˆ
xn|n)T],
is given by
Kn=Pn|n1HH
n(Rn+HnPn|n1HT
n)1.
Hint: Use the following formulas
trace{AB}
A =BT(AB a square matrix)
trace{ACAT}
A = 2AC, (C=CT).
Solution: We know that
ˆ
xn|n=ˆ
xn|n1+Kn(ynHnˆ
xn|n1).
Thus,
Pn|n=E(xnˆ
xn|n)(xnˆ
xn|n)T
page-pf10
16
Hence, we have that
4.23. Show that in Kalman filtering, the prior and posterior error covariance
matrices are related as
Pn|n=Pn|n1KnHnPn|n1.
which results in the desired update.
4.24. Derive the Kalman algorithm in terms of the inverse state-error covariance
matrices, P1
n|n. In statistics, the inverse error covariance matrix is related
to Fisher’s information matrix, hence the name of the scheme.
Solution: To build the Kalman algorithm around the inverse state-error
covariance matrices P1
n|n, P 1
n|n1, we need to apply the following matrix
inversion Lemmas,
page-pf11
17
which gives

Trusted by Thousands of
Students

Here are what students say about us.

Copyright ©2022 All rights reserved. | CoursePaper is not sponsored or endorsed by any college or university.