keras-language-modeling icon indicating copy to clipboard operation
keras-language-modeling copied to clipboard

:book: Some language modeling tools for Keras

Results 29 keras-language-modeling issues
Sort by recently updated
recently updated
newest added

The blog post that you mention in your README does not exist anymore. The link gives `404` error.

I could not replicate the results with embedding + maxpooling layer. I ran 3000 epochs. Could you please suggest your parameters for producing following results. Embedding + Max Pooling: -...

when i train CNN model,it epochs 2 twice ,the loss value became nan. it seems getting in local minmize. ![image](https://cloud.githubusercontent.com/assets/8564700/19342657/48cc11d2-9165-11e6-9abe-320f24da3541.png)

I am having a bit of trouble understanding how to incorporate the AttentionLSTM layer into my code. In your blog you have said that *"The attentional component can be tacked...

@codekansas Can you share your parameter settings for the results shared in `results.notes`?

Hi, Could I get attention weight from "K.function([model.layers[0].input], [model.layers['my_attention_lstm_layer'].output])" of attention model?

Train on 16686 samples, validate on 1854 samples Epoch 1/1 16686/16686 [==============================] - 1s - loss: 0.0060 - val_loss: 0.0340Fitting epoch 2000 2016-10-28 07:51:05 -- Epoch 1999 Loss = 0.0060,...

This statement (https://github.com/codekansas/keras-language-modeling/blob/master/attention_lstm.py#L106) overwrites the trainable_weights of the inner LSTM layer. It should add to the weights of the LSTM Layer instead.

I noticed that you run the attention through a sigmoid because you were having numerical problems: https://github.com/codekansas/keras-language-modeling/blob/master/attention_lstm.py#L54 This may work, but I think that should actually be a softmax. In...

if I interrupt the training process, can I continue it later?