没有合适的资源?快使用搜索试试~ 我知道了~
首页扰动LDA:揭示经验均值与期望差异提升人脸识别性能
扰动LDA:揭示经验均值与期望差异提升人脸识别性能
0 下载量 144 浏览量
更新于2024-07-15
收藏 826KB PDF 举报
扰动LDA(Perturbation LDA,简称P-LDA)是一种改进的Fisher线性判别分析方法,针对生物识别和模式识别领域中的问题,特别是在处理实际数据时,尤其关注类经验均值与期望值之间的潜在差异。Fisher LDA的基本原理是基于一个假设,即类别的经验均值等于其期望值,但在实际应用中,这个假设可能并不总是成立,特别是在样本量有限或者存在噪声的情况下。 传统的Fisher准则依赖于精确的均值估计,而P-LDA通过引入扰动随机向量来处理这种不确定性。在P-LDA中,类内和类间的协方差矩阵被扩展到考虑扰动因素,这使得模型能够适应数据中的扰动分布,从而提高分类性能。为了估计扰动随机向量的协方差矩阵,研究者提出了一种有效的方法,确保在实际应用中可以准确地估计这些参数。 该论文在2009年的《模式识别》期刊(PatternRecognition)上发表,讨论了P-LDA在合成数据集和真实面部图像数据集上的评估。实验结果显示,在欠采样(即样本数量不足的情况下)的情形下,P-LDA相较于基于Fisher LDA的传统方法表现更优,因为其能够更好地处理数据集中存在的偏差和噪声,提高了分类的鲁棒性和准确性。 P-LDA的主要贡献在于将统计学中的扰动分析概念引入到机器学习的模式识别任务中,特别是对于对数据质量有高要求的应用,如人脸识别,它能够提供更加稳健和有效的特征提取和降维策略。通过比较和分析,P-LDA不仅展示了其理论优势,也展示了在实际场景中的实用价值,为未来的相关研究提供了新的视角和方法。
资源详情
资源推荐
W.-S. Zheng et al. / Pattern Recognition 42 (2009) 764 -- 779 767
between the expectation values of f
b
(W, n)andf
w
(W, n)withrespect
to
n such that the uncertainty is considered to be over the domain
of
n.Thatis:
˜
W
opt
= arg max
W
E
n
[f
b
(W, n)]/E
n
[f
w
(W, n)]
= arg max
W
f
b
(W)/f
w
(W)
It can be verified that
f
b
(W) = E
n
[f
b
(W, n)] = trace(W
T
˜
S
b
W) (15)
f
w
(W) = E
n
[f
w
(W, n)] = trace(W
T
˜
S
w
W) (16)
So, it is exactly the optimization model formulated in Eq. (12). This
gives an more intuitive understanding of the effects of covariance
matrices
˜
S
w
and
˜
S
b
.ThoughinP-LDA
ˆ
S
w
and
ˆ
S
b
are perturbated
by S
w
and S
b
, respectively, however in Section 5 we will show
˜
S
w
and
˜
S
b
will converge to the precise within-class and between-class
covariance matrices, respectively. This will show the rationality of
P-LDA, since the class empirical mean is almost its expectation value
when sample size is large enough and then the perturbation effect
could be ignored.
2.2. P-LDA under mixture of Gaussian distribution
This section extends Theorem 1 by altering the class distribu-
tion from single Gaussian to mixture of Gaussians [3]. Therefore, the
probability density function of a sample x in class C
k
is:
p(x|C
k
) =
I
k
i=1
P( i|k )N(x|u
i
k
, N
i
k
), (17)
where u
i
k
is the expectation of x in the i th Gaussian component
(GC) N(x|u
i
k
, N
i
k
)ofclassC
k
, N
i
k
is its covariance matrix and P(i|k)is
the prior probability of the ith GC of class C
k
. Such density function
indicates that any sample x in class C
k
mainly distributes in one
of the GC. Therefore, Theorem 1 under single Gaussian distribution
can be extended to learning perturbation in each GC. To do so, the
clusters within each class should be first determined such that data
in each cluster are approximately normally distributed. Then those
clusters are labeled as subclasses, respectively. Finally P-LDA is used
to learn the discriminant information of all those subclasses. It is
similar to the idea of Zhu and Martinez [26] who extended classical
Fisher's LDA to the mixture of Gaussian distribution case.
In details, suppose there are I
k
GCs (clusters) in class C
k
and N
i
k
out of all N samples are in the ith GC of class C
k
.LetC
i
k
denote the ith
GC of class C
k
.Ifwedenotex
k
i,s
as the sth sample of C
i
k
, s = 1, ...,N
i
k
,
then a perturbation random vector
n
k
i,s
for x
k
i,s
can be modeled, where
n
k
i,s
∼ N(0, X
C
i
k
), X
C
i
k
∈
n×n
, so that
˜
x
k
i,s
= x
k
i,s
+ n
k
i,s
is a random
vector stochastically describes the expectation o f subclass C
i
k
, i.e.,
u
i
k
. Then P-LDA can be extended to the mixture of Gaussians case
by classifying the subclasses {C
i
k
}
k=1,...,L
i=1,
...,I
k
. Thus we get the following
theorem
3
a straightforward extension of Theorem 1 and the proof
is omitted.
Theorem 2. Under the Gaussian mixture distribution of data within
each class, the projection matrix of perturbation LDA (P-LDA),
˜
W
opt
,
can be found as follows:
˜
W
opt
= arg max
W
trace(W
T
˜
S
b
W)
trace(W
T
˜
S
w
W)
= arg max
W
trace(W
T
(
ˆ
S
b
+ S
b
)W)
trace(W
T
(
ˆ
S
w
+ S
w
)W)
(18)
where
˜
S
b
= E
n
[
1
2
L
k=1
L
j=1
I
k
i=1
I
j
s=1
N
i
k
N
×
N
s
j
N
(
˜
u
i
k
−
˜
u
s
j
)(
˜
u
i
k
−
˜
u
s
j
)
T
] =
ˆ
S
b
+ S
b
,
S
b
=
L
k=1
I
k
i=1
(N−N
i
k
)
2
N
3
X
C
i
k
+
L
k=1
I
k
i=1
N
i
k
N
3
L
j=1
I
j
s=1,(j,s) (k,i)
(N
s
j
X
C
s
j
),
ˆ
S
b
=
1
2
L
k=1
L
j=1
I
k
i=1
I
j
s=1
N
i
k
N
×
N
s
j
N
(
ˆ
u
i
k
−
ˆ
u
s
j
)(
ˆ
u
i
k
−
ˆ
u
s
j
)
T
,
˜
S
w
=
L
k=1
I
k
i=1
N
i
k
N
˜
S
i
k
=
ˆ
S
w
+ S
w
,
˜
S
i
k
= E
n
k,i
[
N
i
k
s=1
1
N
i
k
(x
k
i,s
−
˜
u
i
k
)(x
k
i,s
−
˜
u
i
k
)
T
],
S
w
=
1
N
L
k=1
I
k
i=1
X
C
i
k
,
ˆ
S
w
=
1
N
L
k=1
I
k
i=1
N
i
k
s=1
(x
k
i,s
−
ˆ
u
i
k
)(x
k
i,s
−
ˆ
u
i
k
)
T
,
ˆ
u
i
k
=
1
N
i
k
N
i
k
s=1
x
k
i,s
,
˜
u
i
k
=
ˆ
u
i
k
+
1
N
i
k
N
i
k
s=1
n
k
i,s
, i = 1, ...,I
k
, k = 1, ...,L,
n
k,i
={n
k
i,1
, ...,n
k
i,N
i
k
}, n
={n
1,1
, ...,n
1,I
1
, ...,n
L,1
, ...,n
L,I
L
}.
3. Estimation of perturbation covariance matrices
For implementation of P-LDA, we need to properly estimate two
perturbation covariance matrices S
b
and S
w
. Parameter estimation
is challenging, since it is always ill-posed [3,23] due to limited sam-
ple size and the curse of high dimensionality. A more robust and
tractable way to overcome this problem is to perform some regu-
larized estimation. It is indeed the motivation here. A method will
be suggested to implement P-LDA with parameter estimation in an
entire PCA subspace without discarding any nonzero principal com-
ponent. Unlike the covariance matrix estimation on sample data, we
will introduce an indirect way for estimation of the covariance ma-
trices of perturbation random vectors, since the observation values
of the perturbation random vectors are hard to be found directly.
For derivation, parameter estimation would focus on P-LDA un-
der single Gaussian distribution, and it could be easily generalized
to the Gaussian mixture distribution case by Theorem 2. This section
3
The designs of
˜
S
b
and
˜
S
w
in the criterion are not restricted to the presented
forms. The goal here is just to present a way how to generalize the analysis under
single Gaussian case.
剩余15页未读,继续阅读
weixin_38519619
- 粉丝: 2
- 资源: 905
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
- 高效办公必备:可易文件夹批量生成器
- 吉林大学图形学与人机交互课程作业解析
- 8086与8255打造简易乒乓球游戏机教程
- Win10下C++开发工具包:Bongo Cat Mver、GLEW、GLFW
- Bootstrap前端开发:六页果蔬展示页面
- MacOS兼容版VSCode 1.85.1:最后支持10.13.x版本
- 掌握cpp2uml工具及其使用方法指南
- C51单片机星形流水灯设计与Proteus仿真教程
- 深度远程启动管理器使用教程与工具包
- SAAS云建站平台,一台服务器支持数万独立网站
- Java开发的博客API系统:完整功能与接口文档
- 掌握SecureCRT:打造高效SSH超级终端
- JAVA飞机大战游戏实现与源码分享
- SSM框架开发的在线考试系统设计与实现
- MEMS捷联惯导解算与MATLAB仿真指南
- Java实现的学生考试系统开发实战教程
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功