TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/GSoP-Net

GSoP-Net

Global second-order pooling convolutional networks

GeneralIntroduced 20001 papers
Source Paper

Description

A Gsop block has a squeeze module and an excitation module, and uses a second-order pooling to model high-order statistics while gathering global information. In the squeeze module, a GSoP block firstly reduces the number of channels from ccc to c′c'c′ (c′<cc' < cc′<c) using a 1×11 \times 11×1 convolution, then computes a c′×c′c' \times c'c′×c′ covariance matrix for the different channels to obtain their correlation. Next, row-wise normalization is performed on the covariance matrix. Each (i,j)(i, j)(i,j) in the normalized covariance matrix explicitly relates channel iii to channel jjj.

In the excitation module, a GSoP block performs row-wise convolution to maintain structural information and output a vector. Then a fully-connected layer and a sigmoid function are applied to get a ccc-dimensional attention vector. Finally, it multiplies the input features by the attention vector, as in an SE block. A GSoP block can be formulated as: \begin{align} s = F_\text{gsop}(X, \theta) & = \sigma (W \text{RC}(\text{Cov}(\text{Conv}(X)))) \end{align} \begin{align} Y & = s X \end{align} Here, Conv(⋅)\text{Conv}(\cdot)Conv(⋅) reduces the number of channels, Cov(⋅)\text{Cov}(\cdot)Cov(⋅) computes the covariance matrix and RC(⋅)\text{RC}(\cdot)RC(⋅) means row-wise convolution.

Papers Using This Method

Global Second-order Pooling Convolutional Networks2018-11-29