【论文阅读】Semi-Supervised Deep Regression with Uncertainty Consistency and Variational Model Ensembling

论文下载
GitHub
bib:

@INPROCEEDINGS{DaiLi2023Semi,
  title		= {Semi-Supervised Deep Regression with Uncertainty Consistency and Variational Model Ensembling via Bayesian Neural Networks},
  author	= {Weihang Dai and Xiaomeng Li and Kwang-Ting Cheng},
  booktitle	= {AAAI},
  year		= {2023},
  pages     = {1--10}
}

1. 摘要

Deep regression is an important problem with numerous applications.

These range from computer vision tasks such as age estimation from photographs, to medical tasks such as ejection fraction estimation from echocardiograms for disease tracking.

提出半监督回归的应用场景,年龄估计和医学任务,后续的实验也是按照这样进行的。

Semi-supervised approaches for deep regression are notably under-explored compared to classification and segmentation tasks, however.

说得太对了,半监督兴,半监督回归王。

Unlike classification tasks, which rely on thresholding functions for generating class pseudo-labels, regression tasks use real number target predictions directly as pseudo-labels, making them more sensitive to prediction quality.

半监督分类任务中的伪标签通过阈值来获取,而回归任务中的伪标签是一个实数,对于伪标签的质量更加严苛。

In this work, we propose a novel approach to semi-supervised regression, namely UncertaintyConsistent Variational Model Ensembling (UCVME), which improves training by generating high-quality pseudo-labels and uncertainty estimates for heteroscedastic regression.

这段话中的关键字很多:

  • UncertaintyConsistent:不确定性一致性
  • Variational Model:变分模型
  • Ensembling:集成
  • heteroscedastic regression:异方差回归

Given that aleatoric uncertainty is only dependent on input data by definition and should be equal for the same inputs, we present a novel uncertainty consistency loss for co-trained models.

aleatoric uncertainty: 随机不确定性,指的是数据中的不确定性,与模型不相干。

Our consistency loss significantly improves uncertainty estimates and allows higher quality pseudo-labels to be assigned greater importance under heteroscedastic regression.

Furthermore, we introduce a novel variational model ensembling approach to reduce prediction noise and generate more robust pseudo-labels.

这里简单来说就是用两个模型的平均值来打伪标签,也叫集成(ensemble)。

We analytically show our method generates higher quality targets for unlabeled data and further improves training.

Experiments show that our method outperforms state-of-the-art alternatives on different tasks and can be competitive with supervised methods that use full labels.

2. 算法描述

2.1. Bayesian Neural Networks(BNN)

由于这个只是其中的方法,我在这里不会详细的介绍,只是大概了解它是什么,能做什么。

Bayesian Neural Networks从名字能知道也是一种神经网络,可以理解为一种神经网络的变体。其中最大的不同是,普通的神经网络的参数是一个常数,是一个确定的值,其输出当然也是一个确定的值。BNN中的参数则不是,它是一个变量,其输出也是一个变量。看到这里,就会有疑问,都是变量,那网络的前向过程怎么计算呢,这不是玩儿呢?这都是不用怀疑,在实际的操作中,我们是假设变量服从某一种分布来简化问题,其中正态分布比较常见。

具体的来说,在只有猫狗的数据集中,对于一张猫的图片来说,普通的神经网络会输出[0.8, 0.2],表示有0.8的概率表示猫,有0.2的概率表示狗。而在BNN中,对于结果是分布的形式[ N ( 0.7 , 0. 1 2 ) \mathcal{N}(0.7, 0.1^2) N(0.7,0.12), N ( 0.2 , 0.0 1 2 ) \mathcal{N}(0.2, 0.01^2) N(0.2,0.012)],其中方差表示对于预测结果的不确定性。

2.1. UCVME

符号意义
D : = { ( x i , y i ) } i = 1 N D := \{(x_i, y_i)\}^{N}_{i=1} D:={(xi,yi)}i=1N有标记数据
D ′ = { x i ′ ′ } i ′ = 1 N ′ D' = \{x^{'}_{i^{'}}\}^{N^{'}}_{i^{'}=1} D={xi}i=1N无标记数据
f m  where  m ∈ { a , b } f_m \text{ where } m \in \{a, b\} fm where m{a,b}two BNNs using Monte Carlo dropout
y ^ i , m \hat{y}_{i,m} y^i,mprediction of model f m f_m fmfor target label y i y_i yi
z ^ i , m \hat{z}_{i,m} z^i,mlog-uncertainty prediction log ⁡ ( σ 2 ) \log(\sigma^2) log(σ2) of model f m f_m fmfor target label y i y_i yi

我们将 σ 2 \sigma^2 σ2表示为任意不确定性,但在实践中预测对数不确定性 log ⁡ ( σ 2 ) \log(\sigma^2) log(σ2),这通常是为了避免获得对方差的负面预测。

强行理解一波的话,就是缩小预测不确定性的值域,有点类似于标准化,值域小了,预测的准确度就会高一些。

UCVME is based on two novel ideas: enforcing aleatoric uncertainty consistency to improve uncertainty-based loss weighting, and variational model ensembling for generating high-quality pseudo-labels.

Novel ideas:

  1. aleatoric uncertainty consistency
  2. variational model ensembling

两者都是为了一个目标generating high-quality pseudo-labels

heteroscedastic regression loss:
L r e g = 1 N ∑ i = 1 N ( y i − y ^ i ) 2 2 σ i 2 + ln ⁡ ( σ i 2 ) 2 (1) \mathcal{L}_{reg} = \frac{1}{N}\sum_{i=1}^{N}\frac{(y_i-\hat{y}_i)^2}{2\sigma_i^2}+\frac{\ln(\sigma_i^2)}{2}\tag{1} Lreg=N1i=1N2σi2(yiy^i)2+2ln(σi2)(1)
值得注意的是,这个loss的表达是来自于已有的工作12
最大似然:
max ⁡ θ log ⁡ p ( y ∣ x , θ ) = max ⁡ θ ∑ i = 1 N log ⁡ p ( y i ∣ y ^ i ( x i , θ ) , σ i 2 ( x i , θ ) ) = max ⁡ θ ∑ i = 1 N log ⁡ N ( y ^ i , σ i 2 ) = max ⁡ θ ∑ i = 1 N log ⁡ 1 2 π σ i 2 exp ⁡ ( − ∥ y i − y ^ i ∥ 2 2 σ i 2 ) = max ⁡ θ ∑ i = 1 N { − ∥ y i − y ^ i ∥ 2 2 σ i 2 − log ⁡ σ i 2 2 − log ⁡ 2 π 2 } \begin{aligned} & \max _\theta \log p(y \mid x, \theta) \\ & =\max _\theta \sum_{i=1}^N \log p\left(y_i \mid \hat{y}_i\left(x_i, \theta\right), \sigma_i^2\left(x_i, \theta\right)\right) \\ & =\max _\theta \sum_{i=1}^N \log \mathcal{N}\left(\hat{y}_i, \sigma_i^2\right) \\ & =\max _\theta \sum_{i=1}^N \log \frac{1}{\sqrt{2 \pi \sigma_i^2}} \exp \left(-\frac{\left\|y_i-\hat{y}_i\right\|^2}{2 \sigma_i^2}\right) \\ & =\max _\theta \sum_{i=1}^N\left\{-\frac{\left\|y_i-\hat{y}_i\right\|^2}{2 \sigma_i^2}-\frac{\log \sigma_i^2}{2}-\frac{\log 2 \pi}{2}\right\} \end{aligned} θmaxlogp(yx,θ)=θmaxi=1Nlogp(yiy^i(xi,θ),σi2(xi,θ))=θmaxi=1NlogN(y^i,σi2)=θmaxi=1Nlog2πσi2 1exp(2σi2yiy^i2)=θmaxi=1N{2σi2yiy^i22logσi22log2π}

  • labeled inputs:

    • uncertainty consistent
      L u n c l b = 1 N ∑ i = 1 N ( z ^ i , a − z ^ i , b ) 2 \mathcal{L}_{unc}^{lb} = \frac{1}{N}\sum_{i=1}^{N}(\hat{z}_{i,a} - \hat{z}_{i,b})^2 Lunclb=N1i=1N(z^i,az^i,b)2
    • Heteroscedastic regression loss
      L r e g l b = 1 N ∑ m = a , b ∑ i = 1 N ( ( y i , m − y ^ i ) 2 2 exp ⁡ ( z ^ i , m ) + z ^ i , m 2 2 ) \mathcal{L}_{reg}^{lb} = \frac{1}{N}\sum_{m=a, b}\sum_{i=1}^{N}(\frac{(y_{i,m}-\hat{y}_i)^2}{2\exp(\hat{z}_{i, m})}+\frac{\hat{z}_{i,m}^2}{2}) Lreglb=N1m=a,bi=1N(2exp(z^i,m)(yi,my^i)2+2z^i,m2)
  • unlabeled inputs:

    • uncertainty consistent
      L u n c u l b = 1 N ′ ∑ m = a , b ∑ i = 1 N ′ ( z ^ i , m − z ~ i , m ) 2 \mathcal{L}_{unc}^{ulb} = \frac{1}{N'}\sum_{m=a, b}\sum_{i=1}^{N'}(\hat{z}_{i,m} - \widetilde{z}_{i,m})^2 Lunculb=N1m=a,bi=1N(z^i,mz i,m)2
      where y ~ i = 1 T ∑ t = 1 T y ^ i , a t + y ^ i , b t 2 \widetilde{y}_i = \frac{1}{T}\sum_{t=1}^{T}\frac{\hat{y}_{i,a}^t+ \hat{y}_{i,b}^t}{2} y i=T1t=1T2y^i,at+y^i,bt.
      值得注意的是这里只是将两个模型的预测值做了平均(essenble)。
    • Heteroscedastic regression loss
      L r e g u l b = 1 N ′ ∑ m = a , b ∑ i = 1 N ′ ( ( y ^ i , m − y ~ i ) 2 2 exp ⁡ ( z ~ i , m ) + z ~ i , m 2 2 ) \mathcal{L}_{reg}^{ulb} = \frac{1}{N'}\sum_{m=a, b}\sum_{i=1}^{N'}(\frac{(\hat{y}_{i,m}-\widetilde{y}_i)^2}{2\exp(\widetilde{z}_{i, m})}+\frac{\widetilde{z}_{i,m}^2}{2}) Lregulb=N1m=a,bi=1N(2exp(z i,m)(y^i,my i)2+2z i,m2)
      where z ~ i = 1 T ∑ t = 1 T z ^ i , a t + z ^ i , b t 2 \widetilde{z}_i = \frac{1}{T}\sum_{t=1}^{T}\frac{\hat{z}_{i,a}^t+ \hat{z}_{i,b}^t}{2} z i=T1t=1T2z^i,at+z^i,bt.
      值得注意的是这里只是将两个模型的预测值做了平均(essenble)。作者从biasvariance decomposition的角度证明了有效性,对标thresholding function for smoothing

Total Loss:
L = L r e g l b + L u n c l b + ω u l b ( L r e g u l b + L u n c u l b ) \mathcal{L} = \mathcal{L}_{reg}^{lb} + \mathcal{L}_{unc}^{lb} + \omega_{ulb}(\mathcal{L}_{reg}^{ulb}+\mathcal{L}_{unc}^{ulb}) L=Lreglb+Lunclb+ωulb(Lregulb+Lunculb)

3. 实验

3.1. Age Estimation from Photographs

3.2. Ejection Fraction Estimation from Echocardiogram Videos


  1. Kendall A, Gal Y. What uncertainties do we need in bayesian deep learning for computer vision?[J]. Advances in neural information processing systems, 2017, 30. ↩︎

  2. https://zhuanlan.zhihu.com/p/568912284 ↩︎

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 1
    评论
近年来,半监督深度面部表情识别成为了人们关注的热点问题之一。在这个领域,研究人员一直致力于利用少量有标签的数据和大量无标签的数据来提高面部表情识别的准确性和鲁棒性。Adaptive是一种有效的半监督学习方法,它能够自适应地利用标签和无标签数据,使得深度学习模型在应用于面部表情识别时更加有效。 半监督学习是一种机器学习方法,利用少量有标签的数据和大量无标签的数据来训练模型。在面部表情识别中,往往很难获取大量有标签的数据,而无标签数据却很容易获取,因此半监督学习成为了一种有吸引力的解决方案。通过利用Adaptive方法,研究人员可以更好地利用无标签数据,提高模型的泛化能力和鲁棒性,从而提升面部表情识别的准确性。 Adaptive方法还可以帮助模型在数据分布变化时自适应地调整,使得模型更具灵活性和稳健性。在面部表情识别任务中,由于不同环境和条件下的面部表情具有差异性,Adaptive方法能够使模型更好地适应这种差异,提高识别的鲁棒性。 总之,半监督深度面部表情识别与Adaptive方法的结合,有望提高面部表情识别的准确性和鲁棒性,为人们提供更加高效和可靠的面部表情识别技术。相信随着更多研究和实践的开展,半监督深度面部表情识别将迎来更加广阔的发展前景。

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

来日可期1314

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值