vikram-gupta
vikram-gupta
Thanks @jamesweb1 for that. I am more or less on same timelines ! Do you think that using multiple GPUs would help us in bringing the training time down, considerably....
@macournoyer I think we could take a pass on the dataset (only the lines used for the training) to count the frequencies of the words and then keep removing the...
I am also getting the same responses when training with the following params - th train.lua --cuda --hiddenSize 1000 --numLayers 2 --dataset 0 --batchSize 5 Ran one more experiment with...
Thanks @chenb67 I trained the models with the following params. Note that, i used --seqLSTM flag because the code was crashing during evaluation as we are converting the input to...
@CamilleMaurice not yet !
Hi @visonpon were you able to figure out where the intermediate loss is being used?
For Protonet, the test accuracy with and without augmentation comes out as following for me: ``` With augmentation: 600 Test Acc = 59.63% +- 0.70% Without augmentation: 600 Test Acc...