TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Deep Independently Recurrent Neural Network (IndRNN)

Deep Independently Recurrent Neural Network (IndRNN)

Shuai Li, Wanqing Li, Chris Cook, Yanbo Gao

2019-10-11Skeleton Based Action RecognitionSequential Image ClassificationLanguage Modelling
PaperPDFCode(official)

Abstract

Recurrent neural networks (RNNs) are known to be difficult to train due to the gradient vanishing and exploding problems and thus difficult to learn long-term patterns and construct deep networks. To address these problems, this paper proposes a new type of RNNs with the recurrent connection formulated as Hadamard product, referred to as independently recurrent neural network (IndRNN), where neurons in the same layer are independent of each other and connected across layers. Due to the better behaved gradient backpropagation, IndRNN with regulated recurrent weights effectively addresses the gradient vanishing and exploding problems and thus long-term dependencies can be learned. Moreover, an IndRNN can work with non-saturated activation functions such as ReLU (rectified linear unit) and be still trained robustly. Different deeper IndRNN architectures, including the basic stacked IndRNN, residual IndRNN and densely connected IndRNN, have been investigated, all of which can be much deeper than the existing RNNs. Furthermore, IndRNN reduces the computation at each time step and can be over 10 times faster than the commonly used Long short-term memory (LSTM). Experimental results have shown that the proposed IndRNN is able to process very long sequences and construct very deep networks. Better performance has been achieved on various tasks with IndRNNs compared with the traditional RNN, LSTM and the popular Transformer.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)86.7Dense IndRNN
VideoNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Zero-Shot LearningNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Zero-Shot LearningNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Activity RecognitionNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Activity RecognitionNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Language ModellingPenn Treebank (Word Level)Test perplexity50.97Dense IndRNN+dynamic eval
Language ModellingPenn Treebank (Word Level)Test perplexity56.37Dense IndRNN
Language ModellingPenn Treebank (Character Level)Bit per Character (BPC)1.18Dense IndRNN
Action LocalizationNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Action LocalizationNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Action DetectionNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Action DetectionNTU RGB+DAccuracy (CV)93.97Dense IndRNN
3D Action RecognitionNTU RGB+DAccuracy (CS)86.7Dense IndRNN
3D Action RecognitionNTU RGB+DAccuracy (CV)93.97Dense IndRNN
Action RecognitionNTU RGB+DAccuracy (CS)86.7Dense IndRNN
Action RecognitionNTU RGB+DAccuracy (CV)93.97Dense IndRNN

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Assay2Mol: large language model-based drug design using BioAssay context2025-07-16Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing2025-07-16