Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Reading Circle (Self-supervised Equivariant Attention Mechanism for Weakly Supervised Semantic Segmentation)

pyman
July 29, 2020

Reading Circle (Self-supervised Equivariant Attention Mechanism for Weakly Supervised Semantic Segmentation)

Explanation of Self-supervised Equivariant Attention Mechanism for Weakly Supervised Semantic Segmentation [Wang+, CVPR20]

pyman

July 29, 2020
Tweet

More Decks by pyman

Other Decks in Research

Transcript

  1. Semantic Segmentation Assign a semantic category to each pixel. 2

    Supervised Dataset : image & pixel-level class label ✗ huge annotation cost ↓ Weakly-supervised Dataset : image & image-level class label
  2. Weakly-Supervised SS Methods Three steps to train on the image-level

    label. 1. predict an initial category-wise response map to localize the object 2. refine the initial response as the pseudo GT 3. train the segmentation network based on pseudo labels 3 1 2 3
  3. Weakly-Supervised SS Methods Three steps to train on the image-level

    label. 1. predict an initial category-wise response map to localize the object 2. refine the initial response as the pseudo GT 3. train the segmentation network based on pseudo labels 4 1
  4. What’s New Introduce a self-supervised equivariant attention mechanism (SEAM). -

    Narrow the supervision gap between fully and weakly supervised semantic segmentation 5
  5. What’s New Introduce a self-supervised equivariant attention mechanism (SEAM). -

    Focus on affine transformation 7 Previous Proposed CAM varies depending on the size of the input image
  6. What’s New Introduce a self-supervised equivariant attention mechanism (SEAM). -

    Focus on affine transformation 8 Previous Proposed Consistent CAM regardless of the size of the input image
  7. Pixel Correlation Module (PCM) Modify CAM by self attention mechanism.

    12 “Non-local neural networks” [Wang+, CVPR18]
  8. Self Attention [Wang+, CVPR18] Non-local mean operation x : input

    signal y : output signal g : representation function f : similarity function (scalar) 13 Gaussian Embedded Gaussian Dot product Concatenation
  9. Pixel Correlation Module (PCM) Modify CAM by self attention mechanism.

    15 “Non-local neural networks” [Wang+, CVPR18]
  10. Loss Design of SEAM 2. Equivariant Regularization (ER) Loss 20

    Consistency between before and after affine transformation
  11. Loss Design of SEAM 2. Equivariant Cross Regularization (ECR) Loss

    22 to further improve the ability of network for equivariance learning
  12. Dataset PASCAL VOC 2012 semantic segmentation benchmark 21 categories one

    or multiple object class 1,464 images in training set 1,449 images in validation set 1,456 images in test set 24
  13. Conclusion Weakly-supervised Learning 1. generate pseudo GT label 2. apply

    supervised learning Weakly-supervised Semantic Segmentation Self-supervised Equivariant Attention Mechanism ↓ appropriate CAM ↓ better pseudo GT label 30