TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Generalized Mean Pooling

Generalized Mean Pooling

Computer VisionIntroduced 20006 papers

Description

Generalized Mean Pooling (GeM) computes the generalized mean of each channel in a tensor. Formally:

e=[(1∣Ω∣∑_u∈Ωxp_cu)1p]_c=1,⋯ ,C\textbf{e} = \left[\left(\frac{1}{|\Omega|}\sum\_{u\in{\Omega}}x^{p}\_{cu}\right)^{\frac{1}{p}}\right]\_{c=1,\cdots,C}e=[(∣Ω∣1​∑_u∈Ωxp_cu)p1​]_c=1,⋯,C

where p>0p > 0p>0 is a parameter. Setting this exponent as p>1p > 1p>1 increases the contrast of the pooled feature map and focuses on the salient features of the image. GeM is a generalization of the average pooling commonly used in classification networks (p=1p = 1p=1) and of spatial max-pooling layer (p=∞p = \inftyp=∞).

Source: MultiGrain

Image Source: Eva Mohedano

Papers Using This Method

Efficient Probabilistic Modeling of Crystallization at Mesoscopic Scale2024-05-26MinkUNeXt: Point Cloud-based Large-scale Place Recognition using 3D Sparse Convolutions2024-03-12GaitMM: Multi-Granularity Motion Sequence Learning for Gait Recognition2022-09-18Deep Learning Based Image Retrieval in the JPEG Compressed Domain2021-07-08Unifying Deep Local and Global Features for Image Search2020-01-14MultiGrain: a unified image embedding for classes and instances2019-02-14