Arraymancer icon indicating copy to clipboard operation
Arraymancer copied to clipboard

Implement orthogonal weight initialisation

Open mratsim opened this issue 5 years ago • 0 comments

Seems to be key for RNNs to learn long-term dependencies.

Foundational paper: Exact solutions to the nonlinear dynamics of learning in deep linear neural networks 2013, Saxe et al, https://arxiv.org/abs/1312.6120

Papers

  • Neural Photo Editing with Introspective Adversarial Networks, Brock et al, https://arxiv.org/abs/1609.07093
  • Tunable Efficient Unitary Neural Networks (EUNN) and their application to RNNs, Jing et al (Lecun supervision), https://arxiv.org/abs/1612.05231
  • On orthogonality and learning recurrent networks with long term dependencies, Vorontsov et al, https://arxiv.org/abs/1702.00071
  • Recurrent Orthogonal Networks and Long-Memory Tasks, Henaff et al (Lecun supervision), https://arxiv.org/abs/1602.06662
  • Regularizing RNNs by Stabilizing Activations, Krueger et al, https://arxiv.org/abs/1511.08400
  • On the difficulty of training Recurrent Neural Networks, Pascanu et al (Bengio supervision), https://arxiv.org/abs/1211.5063
  • Initialization Matters: Orthogonal Predictive State Recurrent Neural Networks, Choromanski et al, https://openreview.net/pdf?id=HJJ23bW0b

mratsim avatar Dec 17 '18 23:12 mratsim