没有合适的资源?快使用搜索试试~ 我知道了~
首页元学习中的可微分凸优化:提升 Few-Shot 识别性能
元学习中的可微分凸优化:提升 Few-Shot 识别性能
需积分: 6 0 下载量 150 浏览量
更新于2024-08-11
收藏 2.64MB PDF 举报
"Meta-Learning with Differentiable Convex Optimization" 是一篇2019年的研究论文,主要关注元学习(meta-learning)在小样本学习(few-shot learning)中的应用。传统上,许多元学习方法倾向于使用简单的基础学习器,如最近邻分类器。然而,该研究发现,在小样本环境下,有监督训练的线性预测器能够提供更好的泛化能力。作者提出了将这些线性预测器作为元学习的基础学习者,以提升小样本学习的表示学习性能。 研究的核心在于利用线性分类器的两大特性:一是凸优化问题的最优解隐式微分,二是优化问题的对偶公式。这些特性使得作者能够通过设计名为MetaOptNet的方法,有效地在保持高维嵌入的同时提高泛化能力,同时相对控制住计算成本。MetaOptNet在诸如miniImageNet、tieredImageNet、CIFAR-FS和FC100等小样本识别基准上展现出了顶尖的性能,这表明了在小样本场景下,线性预测器的使用不仅能提供良好的特征表示,而且在特征大小与性能之间找到了一个有效的平衡点。 这篇论文的贡献在于,它不仅提出了一种新的元学习策略,还展示了如何结合凸优化的理论优势,优化模型以适应小样本学习任务。这种方法对于那些追求高效和准确的小样本学习系统来说,具有重要的理论和实践价值。通过这种方式,研究者们可以更好地理解并改进元学习模型,使其在实际应用场景中更加有效。"
资源详情
资源推荐
Meta-Learning with Differentiable Convex Optimization
Kwonjoon Lee
2
Subhransu Maji
1,3
Avinash Ravichandran
1
Stefano Soatto
1,4
1
Amazon Web Services
2
UC San Diego
3
UMass Amherst
4
UCLA
kwl042@ucsd.edu {smmaji,ravinash,soattos}@amazon.com
Abstract
Many meta-learning approaches for few-shot learning
rely on simple base learners such as nearest-neighbor clas-
sifiers. However, even in the few-shot regime, discrimina-
tively trained linear predictors can offer better generaliza-
tion. We propose to use these predictors as base learners to
learn representations for few-shot learning and show they
offer better tradeoffs between feature size and performance
across a range of few-shot recognition benchmarks. Our
objective is to learn feature embeddings that generalize well
under a linear classification rule for novel categories. To
efficiently solve the objective, we exploit two properties of
linear classifiers: implicit differentiation of the optimality
conditions of the convex problem and the dual formulation
of the optimization problem. This allows us to use high-
dimensional embeddings with improved generalization at a
modest increase in computational overhead. Our approach,
named MetaOptNet, achieves state-of-the-art performance
on miniImageNet, tieredImageNet, CIFAR-FS, and FC100
few-shot learning benchmarks. Our code is available on-
line
1
.
1. Introduction
The ability to learn from a few examples is a hallmark
of human intelligence, yet it remains a challenge for mod-
ern machine learning systems. This problem has received
significant attention from the machine learning community
recently where few-shot learning is cast as a meta-learning
problem (e.g., [22, 8, 33, 28]). The goal is to minimize gen-
eralization error across a distribution of tasks with few train-
ing examples. Typically, these approaches are composed of
an embedding model that maps the input domain into a fea-
ture space and a base learner that maps the feature space
to task variables. The meta-learning objective is to learn
an embedding model such that the base learner generalizes
well across tasks.
While many choices for base learners exist, nearest-
neighbor classifiers and their variants (e.g., [28, 33]) are
1
https://github.com/kjunelee/MetaOptNet
popular as the classification rule is simple and the approach
scales well in the low-data regime. However, discrimina-
tively trained linear classifiers often outperform nearest-
neighbor classifiers (e.g., [4, 16]) in the low-data regime
as they can exploit the negative examples which are often
more abundant to learn better class boundaries. Moreover,
they can effectively use high dimensional feature embed-
dings as model capacity can be controlled by appropriate
regularization such as weight sparsity or norm.
Hence, in this paper, we investigate linear classifiers as
the base learner for a meta-learning based approach for few-
shot learning. The approach is illustrated in Figure 1 where
a linear support vector machine (SVM) is used to learn a
classifier given a set of labeled training examples and the
generalization error is computed on a novel set of examples
from the same task. The key challenge is computational
since the meta-learning objective of minimizing the gener-
alization error across tasks requires training a linear classi-
fier in the inner loop of optimization (see Section 3). How-
ever, the objective of linear models is convex and can be
solved efficiently. We observe that two additional properties
arising from the convex nature that allows efficient meta-
learning: implicit differentiation of the optimization [2, 11]
and the low-rank nature of the classifier in the few-shot set-
ting. The first property allows the use of off-the-shelf con-
vex optimizers to estimate the optima and implicitly differ-
entiate the optimality or Karush-Kuhn-Tucker (KKT) con-
ditions to train embedding model. The second property
means that the number of optimization variables in the dual
formation is far smaller than the feature dimension for few-
shot learning.
To this end, we have incorporated a differentiable
quadratic programming (QP) solver [1] which allows end-
to-end learning of the embedding model with various linear
classifiers, e.g., multiclass support vector machines (SVMs)
[5] or linear regression, for few-shot classification tasks.
Making use of these properties, we show that our method
is practical and offers substantial gains over nearest neigh-
bor classifiers at a modest increase in computational costs
(see Table 3). Our method achieves state-of-the-art perfor-
mance on 5-way 1-shot and 5-shot classification for popu-
arXiv:1904.03758v1 [cs.CV] 7 Apr 2019
下载后可阅读完整内容,剩余8页未读,立即下载
liz_lee
- 粉丝: 68
- 资源: 36
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
- WebLogic集群配置与管理实战指南
- AIX5.3上安装Weblogic 9.2详细步骤
- 面向对象编程模拟试题详解与解析
- Flex+FMS2.0中文教程:开发流媒体应用的实践指南
- PID调节深入解析:从入门到精通
- 数字水印技术:保护版权的新防线
- 8位数码管显示24小时制数字电子钟程序设计
- Mhdd免费版详细使用教程:硬盘检测与坏道屏蔽
- 操作系统期末复习指南:进程、线程与系统调用详解
- Cognos8性能优化指南:软件参数与报表设计调优
- Cognos8开发入门:从Transformer到ReportStudio
- Cisco 6509交换机配置全面指南
- C#入门:XML基础教程与实例解析
- Matlab振动分析详解:从单自由度到6自由度模型
- Eclipse JDT中的ASTParser详解与核心类介绍
- Java程序员必备资源网站大全
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功