TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Dropout

Dropout

GeneralIntroduced 200027477 papers
Source Paper

Description

Dropout is a regularization technique for neural networks that drops a unit (along with connections) at training time with a specified probability ppp (a common value is p=0.5p=0.5p=0.5). At test time, all units are present, but with weights scaled by ppp (i.e. www becomes pwpwpw).

The idea is to prevent co-adaptation, where the neural network becomes too reliant on particular connections, as this could be symptomatic of overfitting. Intuitively, dropout can be thought of as creating an implicit ensemble of neural networks.

Papers Using This Method

Making Language Model a Hierarchical Classifier and Generator2025-07-17DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Developing Visual Augmented Q&A System using Scalable Vision Embedding Retrieval & Late Interaction Re-ranker2025-07-16Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows2025-07-16DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Addressing Data Imbalance in Transformer-Based Multi-Label Emotion Detection with Weighted Loss2025-07-15HANS-Net: Hyperbolic Convolution and Adaptive Temporal Attention for Accurate and Generalizable Liver and Tumor Segmentation in CT Imaging2025-07-15Langevin Flows for Modeling Neural Latent Dynamics2025-07-15Generative Click-through Rate Prediction with Applications to Search Advertising2025-07-15Biological Processing Units: Leveraging an Insect Connectome to Pioneer Biofidelic Neural Architectures2025-07-15KV-Latent: Dimensional-level KV Cache Reduction with Frequency-aware Rotary Positional Embedding2025-07-15Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15LiLM-RDB-SFC: Lightweight Language Model with Relational Database-Guided DRL for Optimized SFC Provisioning2025-07-15Overcoming catastrophic forgetting in neural networks2025-07-14A Simple Approximate Bayesian Inference Neural Surrogate for Stochastic Petri Net Models2025-07-14Efficient Federated Learning with Heterogeneous Data and Adaptive Dropout2025-07-14SentiDrop: A Multi Modal Machine Learning model for Predicting Dropout in Distance Learning2025-07-14Leveraging RAG-LLMs for Urban Mobility Simulation and Analysis2025-07-14Token Compression Meets Compact Vision Transformers: A Survey and Comparative Evaluation for Edge AI2025-07-13Learning from Synthetic Labs: Language Models as Auction Participants2025-07-12