learning-to-learn-by-pytorch icon indicating copy to clipboard operation
learning-to-learn-by-pytorch copied to clipboard

Input dimension of LSTM differs from the original paper

Open BigWZhu opened this issue 5 years ago • 0 comments

As I read the original paper and the repo from deepmind, it seems to me that the LSTM optimizer should only take 1 variable as input to optimize and save the LSTM state for each variables. In other words, with an arbitrary number of parameters, it just updates one after another. While in this implementation, the dimension of the optimizer is fixed as the number of the optimizee.

BigWZhu avatar Feb 13 '20 18:02 BigWZhu