没有合适的资源?快使用搜索试试~ 我知道了~
首页模式识别(第四版)(希腊)西奥多里蒂斯 习题解答pdf
模式识别(第四版)(希腊)西奥多里蒂斯 习题解答pdf
3星 · 超过75%的资源 需积分: 48 162 下载量 82 浏览量
更新于2023-03-16
4
收藏 1.3MB PDF 举报
模式识别(第4版) (希腊)西奥多里蒂斯 习题解答pdf下载;注意:是最新版本的答案。
资源详情
资源推荐
Notes and Solutions for: Pattern Recognition by
Sergios Theodoridis and Konstantinos Koutroumbas.
John L. Weatherwax
∗
October 17, 2015
∗
wax@alum.mit.edu
1
Text copyright
c
2015 John L. Weatherwax
All Right s Reserved
Please Do Not Redistribute Without Permission from the Author
2
Introduction
Here you’ll find some notes that I wrote up as I worked through this excellent book. I’ve
worked hard to make these notes as good as I can, but I have no illusions tha t they are perfect.
If you feel that that there is a better way to accomplish or explain an exercise or derivation
presented in these notes; or that one or more of the explanations is unclear, incomplete,
or misleading, please tell me. If yo u find an error of any kind – technical, grammatical,
typographical, whatever – please tell me that, too. I’ll gladly add to the acknowledgments
in later printings the name of the first person to bring each problem to my attention.
Acknowledgments
Special thanks to (most recent comments are listed first): Ronald Santos, Karonny F., and
Mohammad Heshajin for helping improve these notes and solutions.
All comments (no matt er how small) are much appreciated. In fact, if you find these notes
useful I wo uld appreciate a contribution in the form of a solution to a problem that I did
not work, a ma t hematical derivation of a statement or comment made in the book that was
unclear, a piece of code that implements one of the algorithms discussed, or a correction to
a typo (spelling, grammar, etc) about these notes. Sort of a “take a penny, leave a penny”
type o f approach. Remember: pay it forward.
3
Classifiers Based on Baye s Decision Theory
Notes on the text
Minimizing the average risk
The symbol r
k
is the expected risk associated with observing an object from class k. This
risk is divided up into parts that depend on what we then do when an object from class k
with feature vector x is observed. Now we only o bserve the feature vector x and not the true
class label k. Since we must still perform an action when we observe x let λ
ki
represent the
loss associated with the event t ha t the object is truly from class k and we decided that it is
from class i. D efine r
k
as the expected loss when an object of type k is presented to us. Then
r
k
=
M
X
i=1
λ
ki
P (we classify this object as a member of class i)
=
M
X
i=1
λ
ki
Z
R
i
p(x|ω
k
)dx ,
which is the books equation 2.14. Thus the total risk r is the expected value of the class
dependent risks r
k
taking into account how likely each class is or
r =
M
X
k=1
r
k
P (ω
k
)
=
M
X
k=1
M
X
i=1
λ
ki
Z
R
i
p(x|ω
k
)P (ω
k
)dx
=
M
X
i=1
Z
R
i
M
X
k=1
λ
ki
p(x|ω
k
)P (ω
k
)
!
dx . (1)
The decision rule that leads to the smallest total risk is obtained by selecting R
i
to be the
region of feature space in which the integrand above is as small as possible. That is, R
i
should be defined as the values of x such that for that value of i we have
M
X
k=1
λ
ki
p(x|ω
k
)P (ω
k
) <
M
X
k=1
λ
kj
p(x|ω
k
)P (ω
k
) ∀j .
In words the index i, when put in the sum above gives the sma lles t value when compared to
all other possible choices. For these values of x we should select class ω
i
as our classification
decision.
4
Bayesian classification with normal distributions
When the covariance matrices f or two classes are the same a nd diagonal i.e. Σ
i
= Σ
j
= σ
2
I
then the discrimination functions g
ij
(x) are given by
g
ij
(x) = w
T
(x −x
0
) = (µ
i
− µ
j
)
T
(x − x
0
) , (2)
since the vector w is w = µ
i
− µ
j
in this case. Note that the point x
0
is on the decision
hyperplane i.e. satisfies g
ij
(x) = 0 since g
ij
(x
0
) = w
T
(x
0
− x
0
) = 0. Let x be another point
on the decision hyperplane, then x − x
0
is a vector in the decision hyperplane. Since x is a
point on the decision hyperplane it also must satisfy g
ij
(x) = 0 from the functional form for
g
ij
(·) and the definition of w is this means that
w
T
(x − x
0
) = (µ
i
− µ
j
)
T
(x − x
0
) = 0 .
This is the statement that the line connecting µ
i
and µ
j
is orthogonal to the decision hy-
perplane. In t he same way, when the covariance matrices of each class are not diagonal but
are nevertheless the Σ
i
= Σ
j
= Σ the same logic that we used above states that the decision
hyperplane is a gain orthogonal to the vector w which in this case is Σ
−1
(µ
i
−µ
j
).
The magnitude o f P (ω
i
) relative to P (ω
j
) influences how close t he decision hyperplane is
to t he respective class means µ
i
or µ
j
, in the sense tha t the class with the larger a priori
probability will have a “larger” region of R
l
assigned to it for classification. For example, if
P (ω
i
) < P (ω
j
) then ln
P (ω
i
)
P (ω
j
)
< 0 so the point x
0
which in the case Σ
i
= Σ
j
= Σ is given
by
x
0
=
1
2
(µ
i
+ µ
j
) − ln
P (ω
i
)
P (ω
j
)
µ
i
− µ
j
||µ
i
− µ
j
||
2
Σ
−1
, (3)
we can write as
x
0
=
1
2
(µ
i
+ µ
j
) + α(µ
i
− µ
j
) ,
with the value of α > 0. Since µ
i
− µ
j
is a vector fr om µ
j
to µ
i
the expression f or x
0
above
starts at the midpoint
1
2
(µ
i
+ µ
j
) a nd moves closer to µ
i
. Meaning that the amo unt of R
l
assigned to class ω
j
is “larg er” than the amount assigned to class ω
i
. This is expected since
the prior probability of class ω
j
is larger than that of ω
i
.
Notes on Example 2.2
To see the final lengths of the principal axes we start with the transformed equation of
constant Mahalanobis distance of d
m
=
√
2.952 or
(x
′
1
− µ
′
11
)
2
λ
1
+
(x
′
2
−µ
′
12
)
2
λ
2
= (
√
2.952)
2
= 2.952 .
Since we want the principal axis about (0, 0) we have µ
′
11
= µ
′
12
= 0 and λ
1
and λ
2
are the
eigenvalues given by solving |Σ − λI| = 0. In this case, we get λ
1
= 1 (in direction v
1
) and
λ
2
= 2 (in direction v
2
). Then the above becomes in “ standa r d form” for a conic section
(x
′
1
)
2
2.952λ
1
+
(x
′
2
)
2
2.952λ
2
= 1 .
5
剩余211页未读,继续阅读
雪国一枝花
- 粉丝: 0
- 资源: 3
上传资源 快速赚钱
- 我的内容管理 收起
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
会员权益专享
最新资源
- BSC关键绩效财务与客户指标详解
- 绘制企业战略地图:从财务到客户价值的六步法
- BSC关键绩效指标详解:财务与运营效率评估
- 手持移动数据终端:常见问题与WIFI设置指南
- 平衡计分卡(BSC):绩效管理与战略实施工具
- ESP8266智能家居控制系统设计与实现
- ESP8266在智能家居中的应用——网络家电控制系统
- BSC:平衡计分卡在绩效管理与信息技术中的应用
- 手持移动数据终端:常见问题与解决办法
- BSC模板:四大领域关键绩效指标详解(财务、客户、运营与成长)
- BSC:从绩效考核到计算机网络的关键概念
- BSC模板:四大维度关键绩效指标详解与预算达成分析
- 平衡计分卡(BSC):绩效考核与战略实施工具
- K-means聚类算法详解及其优缺点
- 平衡计分卡(BSC):从绩效考核到战略实施
- BSC:平衡计分卡与计算机网络中的应用
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功