没有合适的资源?快使用搜索试试~ 我知道了~
首页联合概率分布适应(JPDA)在迁移学习中的应用
联合概率分布适应(JPDA)在迁移学习中的应用
需积分: 9 1 下载量 140 浏览量
更新于2024-09-05
收藏 560KB PDF 举报
"本文介绍了转移学习(Transfer Learning)领域的一个新方法——联合概率分布适应(Joint Probability Distribution Adaptation, JPDA)。JPDA旨在解决现有转移学习方法中存在的问题,这些问题通常依赖于联合概率分布度量,该度量是边缘分布和条件分布的加权和,但在优化时独立处理这两者,忽略了它们之间的内在依赖性。 文章指出,传统的转移学习方法在解决源域和目标域之间任务转换时,可能无法充分考虑两个域之间的关联性和类别的区分度。为此,作者提出了一种新的JPDA方法,它在分布适应过程中,通过最小化对应类别的联合概率差异来增强源域和目标域之间的转移性,同时通过最大化不同类别的联合概率差异来提高分类的辨别力。 JPDA的创新之处在于其能够同时优化转移性和辨别性,这在实验中得到了验证。论文在六个图像分类数据集上进行了实验,结果表明JPDA的表现优于几个基于度量的最先进的转移学习方法。这些数据集可能包括如MNIST、CIFAR-10、CIFAR-100等具有挑战性的图像识别任务,证明了JPDA在实际应用中的有效性。 论文的发表日期为2019年12月1日,归档于arXiv,分类为计算机科学的机器学习领域(cs.LG)。作者Wen Zhang和Dongrui Wu通过JPDA的提出,为深度学习和迁移学习领域的研究提供了新的视角和工具,对于解决跨域学习问题,尤其是数据不足或目标域与源域存在显著差异的情况,具有重要的理论和实践意义。"
资源详情
资源推荐
arXiv:1912.00320v1 [cs.LG] 1 Dec 2019
Transferability versus Discriminability:
Joint Probability Distribution Adap tation (JPDA)
Wen Zhang
1
and Dongrui Wu
2
Abstract. Transfer learning makes use of data or knowledge in
one task to help solve a different, yet related, task. Many existing
TL approaches are based on a joint probability distribution metric,
which is a weighted sum of the marginal distribution and the condi-
tional distribution; however, they optimize the two distributions in-
dependently, and ignore their intrinsic dependency. This paper pro-
poses a novel and frustratingly easy Joint Probability Distribution
Adaptation (JPDA) approach, to replace the frequently-used joint
maximum mean discrepancy metric in transfer learning. During the
distribution adaptation, JPDA improves the transferability between
the source and the target domains by minimizing the joint prob-
ability discrepancy of the corresponding class, and also increases
the discriminability between different classes by maximizing their
joint probability discrepancy. Experiments on six image classifica-
tion datasets demonstrated that JPDA outperforms several state-of-
the-art metric-based transfer learning approaches.
1 INTRODUCTION
A basic assumption in statistical machine learning is that the train-
ing and the test data come from the same distri bution. However, this
assumption does not hold in many real-world applications. For exam-
ple, in image recognition, the distributions in training and testing can
be different due to varying scene, lighting, view angle, image r eso-
lution, etc. Annotating data for a new domain is often expensive and
time-consuming, thus t here are application scenarios that we have
plenty of data, but none or a very small amount of them are labeled
[25]. Transfer learning (TL) has shown promising performance in
handling such a challenge, by transferring knowl edge from a labeled
source domain to a new (unlabeled) target domain [24, 25]. In the
last decade, it has been widely used in image recognition [6, 10, 19],
emotion recognition [23], brain-comput er interfaces [14, 29], and so
on [15, 16, 26].
Typical TL approaches can be categorized into parameter-based
transfers [25], instance-based transfers, and feature transformation
based transfers. Parameter-based transfers need some labeled data,
whereas this paper focuses on unsupervised domain adaptation, in
which the target domain does not have any labeled data at all.
Instance-based transfers assume that the source and the target do-
mains share the same conditional distri bution [25, 30], which usually
does not hold in practice. Feature transformation based transfers re-
lax this assumption, and only assume that there exists a common
subspace, in which the source and t he target domains have similar
1
School of Artificial Intelligence and Automation, Huazhong University of
Science and Technology, Wuhan, China. Email: wenz@hust.edu.cn
2
School of Artificial Intelligence and Automation, Huazhong University of
Science and Technology, Wuhan, China. Email: drwu@hust.edu.cn
distributions. This paper considers feature transformation based T L.
According to Pan and Yang [25], TL can be applied when the
source and the target domains have different feature spaces, label
spaces, marginal probability distributions, and/or conditional prob-
ability distributions. Existing feature transformation based TL ap-
proaches mainly focus on minimizing the distribution divergence
between the source and the target domains by a distribution met-
ric. Frequently used such metrics include maximum mean discrep-
ancy (MMD) [11], KullbackLeibler divergence [27], Wasserstein dis-
tance [17], etc. MMD on marginal and/or conditional distribution is
probably the most popular metric in TL. Existing MMD based distri-
bution adaptation approaches consider either the marginal distribu-
tion only [24], or both the marginal and the conditional distributions
with equal weight [5, 19, 20] or different weights [28], even in deep
learning [8, 18] and adversarial learning [7, 21].
Among them, joint distribution adaptation (JDA) [19] is the most
widely used baseline in TL, whose idea is to measure the distribu-
tion shift between two domains using the marginal and the condi-
tional MMD. Some works extended JDA by adding a regularization
term [20], structural consistency [13], source domain discriminabil-
ity [30], etc. For JDA based approaches, the marginal and conditional
distributions are often treated equally, which may not be optimal; so,
balanced distribution adaptation (BDA) [28] was proposed to give
them different weights. However, both JDA and BDA consider the
marginal and conditional distributions separately, ignoring the intrin-
sic dependency between them. The performance may be improved if
this dependency can be taken into consideration.
Two measures need to be considered during feature transformation
to facilitate domain adaptation [4]. One is transferability, which mea-
sures how capable the feature r epresentation can minimize the cross-
domain discrepancies. The other is discriminability, which measures
how easily different classes can be distinguished by a supervised
classifier. Traditional distribution adaptation approaches usually seek
to achieve high transferability [3,5,19], so that the knowledge learned
from the source domain can be effectively transferred to the target do-
main; however, the feature discriminability has been l argely ignored.
This paper considers the scenario that the source and the target
domains share the same feature and label spaces, which is the most
common assumption in TL. Different from joint MMD based ap-
proaches, we do not use (weighted) sum of the marginal and con-
ditional MMDs to estimate the distribution discrepancy; instead, we
use the joint probability distribution directly, which in theory can
better leverage the relationship between different distri butions. To
consider both transferability and discriminability simultaneously, we
propose joint probability MMD for distribution adaptation, which
minimizes the distribution discrepancy of the same class between dif-
ferent domains and maximizes the distribution discrepancy between
下载后可阅读完整内容,剩余6页未读,立即下载
amazingftt
- 粉丝: 0
- 资源: 1
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
- 李兴华Java基础教程:从入门到精通
- U盘与硬盘启动安装教程:从菜鸟到专家
- C++面试宝典:动态内存管理与继承解析
- C++ STL源码深度解析:专家级剖析与关键技术
- C/C++调用DOS命令实战指南
- 神经网络补偿的多传感器航迹融合技术
- GIS中的大地坐标系与椭球体解析
- 海思Hi3515 H.264编解码处理器用户手册
- Oracle基础练习题与解答
- 谷歌地球3D建筑筛选新流程详解
- CFO与CIO携手:数据管理与企业增值的战略
- Eclipse IDE基础教程:从入门到精通
- Shell脚本专家宝典:全面学习与资源指南
- Tomcat安装指南:附带JDK配置步骤
- NA3003A电子水准仪数据格式解析与转换研究
- 自动化专业英语词汇精华:必备术语集锦
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功