TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/AWD-LSTM

AWD-LSTM

ASGD Weight-Dropped LSTM

SequentialIntroduced 200052 papers
Source Paper

Description

ASGD Weight-Dropped LSTM, or AWD-LSTM, is a type of recurrent neural network that employs DropConnect for regularization, as well as NT-ASGD for optimization - non-monotonically triggered averaged SGD - which returns an average of last iterations of weights. Additional regularization techniques employed include variable length backpropagation sequences, variational dropout, embedding dropout, weight tying, independent embedding/hidden size, activation regularization and temporal activation regularization.

Papers Using This Method

Advanced Deep Learning Techniques for Analyzing Earnings Call Transcripts: Methodologies and Applications2025-02-27No Argument Left Behind: Overlapping Chunks for Faster Processing of Arbitrarily Long Legal Texts2024-10-24RICo: Reddit ideological communities2024-06-05Exploring Multi-Level Threats in Telegram Data with AI-Human Annotation: A Preliminary Study2023-12-15Illicit Darkweb Classification via Natural-language Processing: Classifying Illicit Content of Webpages based on Textual Information2023-12-08Explainable and High-Performance Hate and Offensive Speech Detection2022-06-26IIITT@Dravidian-CodeMix-FIRE2021: Transliterate or translate? Sentiment analysis of code-mixed text in Dravidian languages2021-11-15Offensive Language Identification in Low-resourced Code-mixed Dravidian languages using Pseudo-labeling2021-08-27Towards Offensive Language Identification for Tamil Code-Mixed YouTube Comments and Posts2021-08-24Learning ULMFiT and Self-Distillation with Calibration for Medical Dialogue System2021-07-20WHOSe Heritage: Classification of UNESCO World Heritage "Outstanding Universal Value" Documents with Soft Labels2021-04-12L3CubeMahaSent: A Marathi Tweet-based Sentiment Analysis Dataset2021-03-21indicnlp@kgp at DravidianLangTech-EACL2021: Offensive Language Identification in Dravidian Languages2021-02-14indicnlp@ kgp at DravidianLangTech-EACL2021: Offensive Language Identification in Dravidian Languages2021-02-14Train your classifier first: Cascade Neural Networks Training from upper layers to lower layers2021-02-09Experimental Evaluation of Deep Learning models for Marathi Text Classification2021-01-13LaDiff ULMFiT: A Layer Differentiated training approach for ULMFiT2021-01-13Post-Training Weighted Quantization of Neural Networks for Language Models2021-01-01HinglishNLP at SemEval-2020 Task 9: Fine-tuned Language Models for Hinglish Sentiment Detection2020-12-01Smash at SemEval-2020 Task 7: Optimizing the Hyperparameters of ERNIE 2.0 for Humor Ranking and Rating2020-12-01