TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/ShuffleNet V2 Block

ShuffleNet V2 Block

Computer VisionIntroduced 200021 papers
Source Paper

Description

ShuffleNet V2 Block is an image model block used in the ShuffleNet V2 architecture, where speed is the metric optimized for (instead of indirect ones like FLOPs). It utilizes a simple operator called channel split. At the beginning of each unit, the input of ccc feature channels are split into two branches with c−c′c - c'c−c′ and c′c'c′ channels, respectively. Following G3, one branch remains as identity. The other branch consists of three convolutions with the same input and output channels to satisfy G1. The two 1×11\times11×1 convolutions are no longer group-wise, unlike the original ShuffleNet. This is partially to follow G2, and partially because the split operation already produces two groups. After convolution, the two branches are concatenated. So, the number of channels keeps the same (G1). The same “channel shuffle” operation as in ShuffleNet is then used to enable information communication between the two branches.

The motivation behind channel split is that alternative architectures, where pointwise group convolutions and bottleneck structures are used, lead to increased memory access cost. Additionally more network fragmentation with group convolutions reduces parallelism (less friendly for GPU), and the element-wise addition operation, while they have low FLOPs, have high memory access cost. Channel split is an alternative where we can maintain a large number of equally wide channels (equally wide minimizes memory access cost) without having dense convolutions or too many groups.

Papers Using This Method

Rethinking Information Loss in Medical Image Segmentation with Various-sized Targets2024-03-28Fragility, Robustness and Antifragility in Deep Learning2023-12-15A Non-monotonic Smooth Activation Function2023-10-16Real Time Egocentric Segmentation for Video-self Avatar in Mixed Reality2022-07-04SMU: smooth activation function for deep networks using smoothing maximum technique2021-11-08SAU: Smooth activation function using convolution with approximate identities2021-09-27ErfAct and Pserf: Non-monotonic Smooth Trainable Activation Functions2021-09-09Rethinking Image Deraining via Rain Streaks and Vapors2020-08-03Egocentric Human Segmentation for Mixed Reality2020-05-25DyNet: Dynamic Convolution for Accelerating Convolutional Neural Networks2020-04-22CNN-CASS: CNN for Classification of Coronary Artery Stenosis Score in MPR Images2020-01-23Depth-wise Decomposition for Accelerating Separable Convolutions in Efficient Convolutional Neural Networks2019-10-21ThunderNet: Towards Real-Time Generic Object Detection on Mobile Devices2019-10-01Mish: A Self Regularized Non-Monotonic Activation Function2019-08-23DiCENet: Dimension-wise Convolutions for Efficient Networks2019-06-08Butterfly Transform: An Efficient FFT Based Neural Architecture Design2019-06-05ThunderNet: Towards Real-time Generic Object Detection2019-03-28DetNAS: Backbone Search for Object Detection2019-03-26ShuffleNASNets: Efficient CNN models through modified Efficient Neural Architecture Search2018-12-07DSNet for Real-Time Driving Scene Semantic Segmentation2018-12-06