TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/mRNN

mRNN

Multiplicative RNN

SequentialIntroduced 20111 papers

Description

A Multiplicative RNN (mRNN) is a type of recurrent neural network with multiplicative connections. In a standard RNN, the current input x_tx\_{t}x_t is first transformed via the visible-to-hidden weight matrix W_hxW\_{hx}W_hx and then contributes additively to the input for the current hidden state. An mRNN allows the current input (a character in the original example) to affect the hidden state dynamics by determining the entire hidden-to-hidden matrix (which defines the non-linear dynamics) in addition to providing an additive bias.

To achieve this goal, the authors modify the RNN so that its hidden-to-hidden weight matrix is a (learned) function of the current input x_tx\_{t}x_t:

h_t=tanh⁡(W_hxx_t+W_hh(x_y)h_t−1+b_h) h\_{t} = \tanh\left(W\_{hx}x\_{t} + W\_{hh}^{\left(x\_{y}\right)}h\_{t-1} + b\_{h}\right)h_t=tanh(W_hxx_t+W_hh(x_y)h_t−1+b_h)

o_t=W_ohh_t+b_oo\_{t} = W\_{oh}h\_{t} + b\_{o}o_t=W_ohh_t+b_o

This is the same as the equations for a standard RNN, except that W_hhW\_{hh}W_hh is replaced with W(xt)_hhW^{(xt)}\_{hh}W(xt)_hh. allowing each input (character) to specify a different hidden-to-hidden weight matrix.

Papers Using This Method

Data-driven Preference Learning Methods for Sorting Problems with Multiple Temporal Criteria2023-09-22