Taskonomy = task taxonomy (分类学、分类法)
<<<<<<< HEAD
- Self-supervised learning leverage the inherent relationships between tasks to learn a desired expensive one (e.g. object detection) via a cheap surrogate (e.g. colorization) =======
- Self-supervised learning leverage the inherent relationships between tasks to learn a desired expensive one (e.g. object detection) via a cheap surrogate (e.g. colorization)
bb2b8d8fff03308190420d4f4476947be5fb198a 利用task内在的关系,通过一个简单的代替品来学习一个复杂的任务
- Unsupervised learning is concerned with the redundancies in the input domain and leveraging them for forming compact representations, which are usually agnostic to the downstream task
利用输入冗余来形成一个紧凑的表达,类似于downstream task
- Meta-learning generally seeks performing the learning at a level higher than where conventional learning occurs
在更高的层次上学习
- Multi-task learning targets developing systems that can provide multiple outputs for an input in one run
对于一个输出有多个输出
- Domain adaption seeks to render a function that is developed on a certain domain applicable to another
在一个domian上学习的函数可以适用于其他domain
- problem: maximize the collective performance on a set of tasks
$\mathcal T={t_1,...,t_n}$ , subject to the constraint that we have a limited supervision budget$\gamma$ - task dictionary:
$\mathcal V=\mathcal T\cup\mathcal S$
$\mathcal T$ 是目标任务,$\mathcal S$ 是训练任务,$\mathcal T-\mathcal T\cap\mathcal S$ 是target-only任务,$\mathcal S-\mathcal T\cap\mathcal S$ 是source-only任务
- four step process
- a task-specific network for each task in
$\mathcal S$ is trained - all feasible transfers between sources and targets are trained
higher-order transfers: multiple inputs task to transfer to one target
- the task affinities acquired from transfer function performances are normalized
- we synthesize a hypergraph which can predict the performance of any transfer policy and optimize for the optimal one
- a task-specific network for each task in
train a fully supervised task-specific network for each task in
每一个task都有encoder和decoder
-
$t\in \mathcal T, s\in \mathcal S$ , a transfer network learns a small readout function for$t$ given a statistic computed for$s$ - 对一幅图像$I\in\mathcal D$的表达是$s$的encoder$E_s(I)$。 readout function
$D_{s\to t}$ 有参数$\theta_{s\to t}$, 他需要最小化损失$L_t$$$D_{s\to t}=\arg\min_\theta\mathbb E_{I\in\mathcal D}[L_t(D_\theta(E_s(I)),f_t(I))]$$
$f_t(I)$ 是$t$任务、$I$图片的ground truth。$E_s(I)$可能不足以表达$t$任务,因此$D_{s\to t}$的performance可以作为任务相似度的度量(metric as task affinity)
have an affinity matrix of transferabilities across tasks
找到任务之间的关系、相似性,每一对任务的关系用一个标量表示,组成一个矩阵
devise a global transfer policy which maximizes collective performance across all tasks, while minimizing the used supervision.
找到最好的transfor的方式
