paraphrase-id-tensorflow icon indicating copy to clipboard operation
paraphrase-id-tensorflow copied to clipboard

Various models and code (Manhattan LSTM, Siamese LSTM + Matching Layer, BiMPM) for the paraphrase identification task, specifically with the Quora Question Pairs dataset.

Results 13 paraphrase-id-tensorflow issues
Sort by recently updated
recently updated
newest added

I met this error when trying to run the code (only change Glove to 840B.300d but remain filename as 6B.300d) Does anybody know how to fix this? File "scripts/run_model/run_bimpm.py", line...

Hi, I appreciate your sharing this project! It is a very thoughtful work and friendly to newers! I have some questions when reading the code. 1. In dataset.py Instead of...

Hi, i am running the biMPM model to predict, getting the following result: ``` Traceback (most recent call last): File "run_bimpm.py", line 267, in 66%|███████████████████████████████████████████████████████████████▊ | 2345735/3563475 [09:22

Looks like the train time is pretty long on AWS instances with K80s. Adding MultiGPU data parallelism would be a good way to mitigate this (as done in https://www.tensorflow.org/tutorials/using_gpu#using_multiple_gpus)

enhancement
help wanted

Traceback (most recent call last): File "scripts/run_model/run_siamese_matching_bilstm.py", line 16, in from duplicate_questions.data.instances.sts_instance import STSInstance File "scripts/run_model/../../duplicate_questions/data/instances/sts_instance.py", line 13, in class STSInstance(TextInstance): File "scripts/run_model/../../duplicate_questions/data/instances/sts_instance.py", line 69, in STSInstance @overrides File "/home//f.local/lib/python3.6/site-packages/overrides/overrides.py",...

in most cases, skip-gram works better than GLOVE and using skip-gram vectors can improve the performance. Also, there are much more open source libraries to work with skip-gram.

This would make our model training a lot faster.

enhancement
help wanted

right now, the data pipeline will tokenize the input into both words / characters, even if you only want words. This is fine for now since character tokenization isn't that...

enhancement
help wanted

Right now, the model can "train" (training on train data / periodically measure validation accuracy / loss) and it can "predict" (given an unlabeled test set, make predictions). It would...

enhancement
help wanted

SwitchableDropoutWrapper currently has to run the LSTM cell twice, one to get the dropped out inputs and one to get the un-dropped out inputs (and then use `tf.cond` to output...

enhancement
help wanted