Swin Transformer, SwinIR, SwinFIR

Swin Transformer

痛点:尺度变化 – 提出Hieracical

attention的变化: sliding window attention,减少了attention的计算复杂度,同时通过sliding增强connection,实现全局attention

和ViT的区别:ViT16*下采样,尺寸单一;Swin Transformer多尺度

[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-W37DSRfP-1667876270378)(1103 Swin Transformer.assets/image-20221023201931256.png)]

local的思维:在一个小范围算attention是基本够用的,全局算self-attention有浪费。(基于假设:属性相似的目标距离是相对接近的)

Patch Partition: 相当于打碎成block

Patch Merging: MaxPooling,提供多尺度特征

[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-ofrHCWXf-1667876270379)(1103 Swin Transformer.assets/image-20221023203559144.png)]

从H * W * C变换为H/2 * W/2 * 2C

总backbone:

[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-E63nNxJC-1667876270379)(1103 Swin Transformer.assets/image-20221023204154777.png)]

回顾一下Attention is all you need: 从CNN, RNN变为Transformer

​ CNN:Transformer借鉴了CNN的多通道以提取不同特性的特征,另外CNN的金字塔结构可以将相隔较远的信息归纳在一起
​ RNN:认为RNN结构并行化差,序列化执行效率低下 纳在一起 ​ RNN:认为RNN结构并行化差,序列化执行效率低下
​ 总体来说就是借鉴了优点、改进缺陷

SwinIR: Swin Transformer for Image Restoration

Related Work(IR的方法)
  1. Traditional model-based

  2. CNN-based (SRCNN…):

    a flurry of CNN-based models have been proposed to improve model representation ability by using more elaborate neural network architecture designs, such as residual block, dense block and others . Some of them have exploited the attention mechanism inside the CNN framework, such as channel attention, non-local attention and adaptive patch aggregation.

  3. Vision Transformer

SwinIR Model

[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-2TQqra03-1667876348924)(1103 SwinIR SwinFIR.assets/image-20221025083143664.png)]

SwinIR consists of three modules:
shallow feature extraction, deep feature extraction and high-quality (HQ) image reconstruction modules.

1.Shallow feature extraction

Shallow Features: we use a 3*3 convolutional layer HSF() to extract shallow feature as

  • 2
    点赞
  • 2
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值