abstract
Referring video object segmentation (R-VOS) is an emerging cross-modal task that aims to segment the target object referred by a language expression in all video frames.
In this work, we propose a simple and unified framework built upon Transformer, termed ReferFormer.
It views the language as queries and directly attends to the most relevant regions in the video frames.
Concretely, we introduce a small set of object queries conditioned on the language as the input to the Transformer.
In this manner, all the queries are obligated to find the referred objects only.
They are eventually transformed into dynamic kernels which capture the crucial object-level information, and play the role of convolution filters to generate the segment