TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/DropConnect

DropConnect

GeneralIntroduced 200084 papers
Source Paper

Description

DropConnect generalizes Dropout by randomly dropping the weights rather than the activations with probability 1−p1-p1−p. DropConnect is similar to Dropout as it introduces dynamic sparsity within the model, but differs in that the sparsity is on the weights WWW, rather than the output vectors of a layer. In other words, the fully connected layer with DropConnect becomes a sparsely connected layer in which the connections are chosen at random during the training stage. Note that this is not equivalent to setting WWW to be a fixed sparse matrix during training.

For a DropConnect layer, the output is given as:

r=a((M∗W)v) r = a \left(\left(M * W\right){v}\right)r=a((M∗W)v)

Here rrr is the output of a layer, vvv is the input to a layer, WWW are weight parameters, and MMM is a binary matrix encoding the connection information where M_ij∼Bernoulli(p)M\_{ij} \sim \text{Bernoulli}\left(p\right)M_ij∼Bernoulli(p). Each element of the mask MMM is drawn independently for each example during training, essentially instantiating a different connectivity for each example seen. Additionally, the biases are also masked out during training.

Papers Using This Method

Dynamic DropConnect: Enhancing Neural Network Robustness through Adaptive Edge Dropping Strategies2025-02-27Advanced Deep Learning Techniques for Analyzing Earnings Call Transcripts: Methodologies and Applications2025-02-27No Argument Left Behind: Overlapping Chunks for Faster Processing of Arbitrarily Long Legal Texts2024-10-24Terrain Classification Enhanced with Uncertainty for Space Exploration Robots from Proprioceptive Data2024-07-03RICo: Reddit ideological communities2024-06-05Information-Theoretic Generalization Bounds for Deep Neural Networks2024-04-04Exploring Multi-Level Threats in Telegram Data with AI-Human Annotation: A Preliminary Study2023-12-15Illicit Darkweb Classification via Natural-language Processing: Classifying Illicit Content of Webpages based on Textual Information2023-12-08Bayesian posterior approximation with stochastic ensembles2022-12-15Disentangled Uncertainty and Out of Distribution Detection in Medical Generative Models2022-11-11Explainable and High-Performance Hate and Offensive Speech Detection2022-06-26SoftDropConnect (SDC) -- Effective and Efficient Quantification of the Network Uncertainty in Deep MR Image Analysis2022-01-20IIITT@Dravidian-CodeMix-FIRE2021: Transliterate or translate? Sentiment analysis of code-mixed text in Dravidian languages2021-11-15Entropy optimized semi-supervised decomposed vector-quantized variational autoencoder model based on transfer learning for multiclass text classification and generation2021-11-10Offensive Language Identification in Low-resourced Code-mixed Dravidian languages using Pseudo-labeling2021-08-27Towards Offensive Language Identification for Tamil Code-Mixed YouTube Comments and Posts2021-08-24Learning ULMFiT and Self-Distillation with Calibration for Medical Dialogue System2021-07-20Structured DropConnect for Uncertainty Inference in Image Classification2021-06-16The Regularizing Effect of Different Output Layer Designs in Deep Neural Networks2021-05-21WHOSe Heritage: Classification of UNESCO World Heritage "Outstanding Universal Value" Documents with Soft Labels2021-04-12