Jifeng Dai, Kaiming He, Jian Sun. Convolutional feature masking for joint object and stuff segmentation. CVPR, 2015.

Dai et al. Propose Convolutional Feature Masking (CFM) for instance segmentation. A CFM layer tackles the following two problem: instead of applying masks or bounding boxes on the image content, thereby creating artifacts and requiring multiple forward passes, the convolutional feature maps are computed first and then masked within the CFM layer. In their work, the object proposal masks are obtained from a state-of-the art object proposal generator such as Selective Search (SS) [22] or Multiscale Combinatorial Grouping (MCG) [1]. This difference is illustrated in Figure 1 and Figure 2 shows the idea of the CFM layer.

Figure 1: Illustration of the proposed approach in comparison to state-of-the-art approaches. The Convolutional Feature Maskign (CFM) layer allows to first compute all feature maps and then use proposals to mask the features.

They propose two different ways to integrate the CFM layer with an SPP layer [11] as shown in Figure 3. In experiments, they show that both approaches perform similarly and therefore propose to use Design B (right in Figure 3) as it requires less computation.

Figure 2: Illustration of the CFM layer.

Figure 3: Illustration of the two considered designs (A and B) to integrate the CFM layer with a SPP layer.

What is your opinion on this article? Let me know your thoughts on Twitter @davidstutz92 or LinkedIn in/davidstutz92.