RUN AWAY FROM YOUR TEACHER: A NEW SELF SUPERVISED APPROACH SOLVING THE PUZZLE OF BYOL

motivation

 representation collapse 

Wang & Isola (2020) has empirically demonstrated that the balance of the alignment loss and the uniformity loss is necessary when learning representations through contrastive method.

experiment

1.We examine the speculation that the performance drop is caused by the representation collapse

2.We first replace the complex structure with linear mapping qw(·) = W(·). This replacement provides a naive solution to representation collapse: W = I.initialize W = I, while it never converges to this apparent collapse.

derivation

BYOL LOSS UPPER BOUNDING

 

let λ = β/α

Understanding why BYOL works without collapse is approximately equivalent to understanding how minimizing Lcross-model(qw ◦ fθ, fξ) effectively regularizes the alignment loss.

 

 works as well

 

Removing the predictor,although minimizing  fails to yield better representation than the random baseline.it prevents the overly-optimized alignment loss.

1.why linear predictor improve the quality of feature

architecture

 

UNDERSTANDING BYOL VIA RAFT

proof:

.

(1)

(2)

 

(3)

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值