Image Captioning 经典论文合辑

Image Caption: Automatically describing the content of an image

domain:CV+NLP

Category:(by myself, you can read the survey for detail.)

  • CNN+RNN, with attention mechanisms
  • Reinforcement Learning
  • GAN
  • Compositional Architecture: Review Network, Guiding Network, RFNet (fusion)
  • Cross-domain: objects not present in training captions (out-of-domain/cross domain captioning) Peter Anderson  has some publications about this.  (Show, Adapt and Tell, Constrained Beam Search)
  • Stylized Caption: Senticap, StyleNet, SemStyle
  • Novel Object-based Image Captioning: NOC(Captioning Images wuth Diverse Objects), Neural Baby Talk
  • Diversity(句子的多样性问题)
  • Dense Caption: DenseCap
  • Image Paragraph: (Maybe another research area, but I still place it here. More difficult than image caption)

数据集:Flickr8k,Flickr30k,MSCOCO,Visual Genome,Conceptual Captions(ACL 2018谷歌收集的数据集)

评测指标:BLEU,METEOR,CIDEr,ROUGE,SPICE

A Comprehensive Survey of Deep Learning for Image Captioning (ACM Computing Surveys, October 2018.)

这是一篇比较新的关于Image Caption的综述文章(Survey).

 

Learning to Evaluate Image Captioning(CVPR 2018)

learning based evaluation metric. (rule-based metrics)

BLEU, METEOR, ROUGE, CIDEr: mainly measure the world overlap between generated and reference captions.

SPICE: measures the similarity of scene graphs constructed from the candidate and reference sentence, correlates well with human judgments, but fails to capture the syntactic structure of a sentence.

It is worth nothing that all above mentioned metrics rely solely on similarity between candidate and reference captions, without taking the image into consideration.

Propose a novel learning based discriminative evaluation metric that is directly trained to distinguish between human and machine-generated captions.

Critique: a binary classifier that makes a Turing Test type judgment in which it differentiates between human-written and machine-generated captions.

 

 

Show and Tell: A Neural Image Caption Generator(CVPR 2015)

directly maximize the probability of the correct description given the image by using the following formulation:

θ are the parameters of our model, I is an image, and S its correct transcription

Encoder:Inception-V2

Decoder:LSTM
Inference:BeamSearch

 

Attention Mechanism:

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention (ICML 2015)

Highlight:Attention Mechnism(Soft&Hard)

 "Soft" attention:different parts,different subregions

"Hard" attention:only one subregion.Random choice

 

Sumary:

1.Attention involves focus of certain parts of input

2.Soft Attention is Deterministic.Hard attention is Stochastic.

3.Attention is used in NMT, AttnGAN, teaching machines to read.

Image Captioning with Semantic Attention(CVPR 2016)

(Related work)divided Image Caption into two categories:top-down and bottom-up

Bottom-up: the classical ones(templated-based), start with visual concepts, objects, attributes, words and phrases, and combine them into sentences using language models

Top-down: the modern ones(end-to-end), translate from a visual representation to a language counterpart. The visual representation comes from a convolutional neural network which is often pretrained for image classification on large-scale datasets. Translation is accomplished through recurrent neural networks based on language models.

 multi-label classifier

SCA-CNN: Spatial and Channel-Wise Attention in Convolutional Networks for Image Captioning(CVPR 2017)

Highlight:Spatial and Channel-Wise Attention

Knowing When to Look: Adaptive Attention via a Visual Sentinel for Image Captioning(CVPR 2017)

Hightlight:Adaptive Attention

Semantic Compositional Networks for Visual Captioning(CVPR 2017)

 

Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering(CVPR 2018)

In the human visual system, attention can be focused volitionally by top-down signals determined by the current task(e.g.,looking for something), and automatically by bottom-up signals associated with unexpected, novel or salient stimuli.

top-down:attention mechanisms driven by non-visual or task-specific context; feature weights;

bottom-up:purely visual feed-forward attention mechanisms;based on Faster-RCNN proposes image regions (feature vector);

 

 

 

Stylized Caption: SentiCap, StyleNet, SemStyle

SentiCap:Generating Image Descriptions with Sentiments (AAAI 2016)

 

StyleNet: Generating Attractive Visual Captions with Styles (CVPR 2017)

produce attractive visual captions with styles only using monolingual stylized language corpus (without images) and standard factual image/video-caption pairs.

 SemStyle: Learning to Generate Stylised Image Captions using Unaligned Text (CVPR 2018)

 

First this model maps the image into a semantic term representation via the term generator, then the language generator uses these terms to generate a caption in the target style.

The term generator takes an image as input, extracts features using a CNN and then generates an ordered term sequence summarising the image semantics.

The language generator takes the term sequences as input, encodes it with an RNN and then using an attention based RNN decodes it into natural language with a specific style.

Two-stage learning strategy: learning the term generator network using only a standard image caption dataset(MSCOCO), learning the language generator network on styled text data(romantic novels)

 "Factual" or "Emotional": Stylized Image Captioning with Adaptive Learning and Attention (ECCV 2018)

Style-factual LSTM

 

GAN:

Show, Adapt and Tell: Adversarial Training of Cross-domain Image Captioner (ICCV 2017)

Highlight: Cross-domian

Two New Dataset: CUB-200, TGIF

a "cross-domain" captioner which is trained in  a "source" domain with paired data and generalized to other "target" domains with very little cost (e.g., no paired data required)

The domain critic assesses whether the generated captions are indistinguishable from sentences in the target domain.

The multi-modal critic assesses whether an image and its generated caption is a valid pair.

During training, the critics and captioner act as adversaries——captioner aims to generate indistinguishable captions, whereas critics aim at distinguishing them.

Exposure bias: It happens when a model is trained to maximize the likelihood given ground truth words but follow its own predictions during test inference. (curriculum learning, professor forcing, policy gradient)

 

The first componet is a standard CNN-RNN-based captioner.

 

The second component consists of two critics to provide reward. One critic assesses the similarity between y and in style. The other critic assesses the relevancy between x and y, given paired data P in the source domain as example pairs.

compute a reward for each generated sentence y. Both the captioner and two critics are iteratively trained using a novel adversarial training procedure.

Captioner as an Agent

Critics

For cross-domain image captioning, a good caption needs to satisfy two criteria: (1).the generated sentence resembles the sentence drawn from the target domain. (2).the generated sentence is relevant to the input image.

Domain Critic(DC) to classify sentences as "source" domain, "target" domain, or "generated" ones. The DC model consists of an encoder(CNN) and a classifier(a fully connected layer and a softmax layer).

Multi-modal Critic to check the relevance between a sentence y and an image x, MC to classify (x,y) as "paired", "unpaired", "generated" data.

Adversarial Training

 

 

 Beyond Narrative Description: Generating Poetry from Images by Multi-Adversarial Training (ACM MM 2018)

这篇论文是ACM MM 2018 best paper. 

The framework of poetry generation with multi-adversarial training. A deep coupled visual-poetic model(e) is trained by human annotated image-poem pairs(a). The image features (b) are poetic multi-CNN features obtained by fine-tuning CNNs with the extracted poetic symbols (e.g, objects, scenes and sentiments) by a POS parser from poems. The sentence features(d) of poems are extracted from a skip-thought model(c) trained on the largest public poem corpus (UniM-Poem). A RNN-based sentence generator(f) is trained as agent and two discriminators considering multi-modal (g) and poem-style (h) critics of a generated poem to a given image provide rewards to policy gradient(i). POS parser extracts Part-Of-Speech words from poems.

 

The framework consists of several parts: (1) a deep coupled visual-poetic embedding model to learn poetic representations from images, and (2) a multi-adversarial training procedure optimized by policy gradient.

image embedding :

poem embedding: 

 

The image and poem are embedded together by minimizing a pairwise ranking loss with dot-product similarity:

mis a contrastive(irrelevant unpaired) poem for image embedding x, and vice-versa with xk . denotes the constrastive margin.

 

 

 RL:

Self-critical Sequence Training for Image Captioning (CVPR 2017)

FC models

    

    

   

Attention Model (Att2in)

Attention Model (Att2all)

 

Reinforcement Learning

Sequence Generation as an RL problem

agent, environment, action, state, reward

the goal of training is to minimize the negtive expected reward:

typical:

Policy Gradient with Reinforce

In practice the expected gradient can be approximated using a single Monto-Carlo sample

Reinforce with a Baseline

The baseline b can be arbitrary function, as long as it does not depend on the "action" ws

The baseline does not change the expected gradient, but importantly, it can reduce the variance of the gradient esimate.

Final Gradient Expression

 

Self-critical sequence training (SCST)

The central idea of the SCST approach is to baseline the REINFORCE algorithm with the reward obtained by the current model under the inference algorithm used at test time.

 

is the reward obtained by the current model under the inference algorithm used at test time.

 

Deep Reinforcement Learning-based Image Captioning with Embedding Reward (CVPR 2017)

A decision-making framework for image captioning.

A "policy network" and a "value network" to collaboratively generate captions.

 

Compositional Architecture:

Review Networks for Caption Generation (NIPS 2016)

Reviewer

Tr: a hyperparameter that specifies the number of review steps

The intuition behind the reviewer module is to review all the information encoded by the encoder and learn thought vectors that are a more compact, abstractive, and global representation than the original encoder hidden states.

The reviewer performs Tr review steps on the encoder hidden states H and outputs a thought vector ft after each step. 

Attentive Input Reviewer(common setting in RNNs)

 Attentive Output Reviewer()

 

Weight Tying

 

Recurrent Fusion Network for Image Captioning (ECCV 2018)

All existing models employ only one encoder, so the performance heavily depends on the expressive ability of the deployed CNN.

To exploit complementary information from multiple encoders, propose a Recurrent Fusion Network(RFNet) for image captioning.

Ensemble and Fusion: The combining process can occur at the input, intermediate, and output stage of the target model. RFNet can be regarded as a kind of intermediate fusion methods.

Multiple CNNs are employed as the encoders and a recurrent fusion procedure is inserted after the encoders to form better representations for the decoder. The fusion procedure consists of two stages.

The first stage exploits interactions among the representations from multiple CNNs to generate multiple sets of thought vectors.

The second stage performs multi-attention on the sets of thought vectors from the first stage, and generates a new set of thought vectors for the decoder.

M(上图M=3) CNNs serve as the encoders. The numbers of steps in stages I and II are denoted as T1(T1=2) abd T2(T2=3)

The hidden states from stage I and II are regarded as the thought vectors.

The fusion stage I contains M review componets. The input of each review component at each time step is the concatenation of hidden states from all components at the previous time step. Each review component outputs the hidden state as  a thought vector.

The fusion stage II is a review component that performs the multi-attention mechanism on the multiple sets of thought vectors from the fusion stage I.

Fusion Stage I

At time step t, the transition of the m-th review compoent is computed as

is the attention model for the m-th review component at time step t.

Stage I can be regarded as a grid LSTM with independent attention mechanisms. M x T1 LSTMs are used in fusion stage I.

The set  of thought vectors generated from the m-th component is denoted as :

 

Fusion Stage II

 

The hidden states of fusion stage II are collected to form the thought vector set:

 

Learning to Guide Decoding for Image Captioning (AAAI 2018)

这篇和之前的RFNet是同一个作者, 文章的风格也极其相似.

 Convolutional Image Captioning (CVPR 2018)

As the figure illustrates, our technique contains three main components similar to the RNN technique. The first and the last components are word embeddings in both cases. Masked convolutions are employed in CNN-based approach. This component, unlike the RNN, is feed-forward without any recurrent function.

Inference. a simple feed-forward deep net,fw for modeling Prediction of a word yi relies on past words y<i or their representations:

To disallow convolution operations from using information of future word tokens, we use masked convolutional layers that operate omly on 'past' data.

 

 

 

 

A Neural Compositional Paradigm for Image Captioning (NIPS 2018)

 

 

转载于:https://www.cnblogs.com/czhwust/p/imagecaption.html

  • 0
    点赞
  • 9
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值