Professional Documents
Culture Documents
Where
(1)
Since, the real life signal has also noise in it, hence a more appropriate system model is given by
(2)
and
is no-longer as
as
i.e.
min
Minimization of length of residual vector or square of length are equivalent in the sense, the minimum is
reached at same value of
. However, working with squares eliminates the problem of square roots in
calculating derivatives. Hence, a least square problem is defined as
(6)
The scalar
is introduced only for the sake of convenience. Again it does not affect the optimal
i.e. min
i.e. min
Thus,
as
i.e.
min
Minimization of length of residual vector or square of length are equivalent in the sense, the minimum is
reached at same value of
. However, working with squares eliminates the problem of square roots in
calculating derivatives. Hence, a least square problem is defined as
(6)
The scalar
is introduced only for the sake of convenience. Again it does not affect the optimal
min
i.e. min
i.e. min
Thus,
with
and
Hence,
and
The number of unknowns (n) is less than number of knowns (m) i.e. redundancy
b is a
vector, A is a
matrix. We know from the basics of the optimization theory that at the
minimum or maximum, the gradient of the objective function is zero. The gradient of the objective
function
is a
vector given by
Setting this gradient to zero amounts to solving the following linear system of equations.
(7)
The solution of the above system of equations gives the optimal
definiteness property of
that
(8)
Example : 1
We now repeat the previous example with a 3-sample window. The analogous results to example 1 are
shown in table 1. It can be seen that as a consequence of using a larger data window, the accuracy of
voltage estimation improves significantly.
Table 1 : 3 - Sample Estimation
Randn multiplier(E)
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
2.1
2.2
2.3
2.4
2.5
2.6
2.7
2.8
2.9
3.0
Mean
10.0061
10.0130
10.0207
10.0293
10.0387
10.0490
10.0601
10.0719
10.0846
10.0982
10.1125
10.1277
10.1436
10.1604
10.1780
10.1964
10.2156
10.2356
10.2564
10.2781
10.3005
10.3238
10.3479
10.3727
10.3985
10.4250
10.4524
10.4807
10.5099
10.5400
Standard deviation
0.0927
0.1855
0.2783
0.3712
0.4641
0.5570
0.6499
0.7428
0.8358
0.9287
1.0216
1.1144
1.2072
1.3000
1.3927
1.4853
1.5778
1.6702
1.7625
1.8547
1.9468
2.0386
2.1303
2.2218
2.3130
2.4040
2.4947
2.5849
2.6747
2.7638
standard deviation reduces to 0.0927. Similarly with E = 3, standard deviation with 2-sample approaches
4.3830, which reduces to 2.7638 with 3-sample window. This shows that redundancy helps in filtering out
noise. However estimation spans a large data window.
Review Questions
1.
2.
3.
Now consider
where A is a full-column, rank matrix. Show that
for all
is positive definite.
Show that at the minimum of a function f(x), the matrix of second derivatives
is at least
positive semi
definite. Also, show that if H is positive definite, then we have a strict local minimum.
4.
5.
A symmetric real
for all
Hence, show
is positive semi definite.
that if A is not at all full-column rank matrix, then
Show that a real symmetric
matrix is positive definite, if and only if all its Eigen values are real and
greater than
zero. Hence, comment on the positive definiteness of following matrices.
a) I n (
- identity matrix).
b)
c)
d)
Recap