meshed-memory-transformer
meshed-memory-transformer copied to clipboard
Meshed-Memory Transformer for Image Captioning. CVPR 2020
Meshed-Memory Transformer Evaluation Evaluation: 0%| Evaluation: 0%| | 0/500 [00:00
Meshed-Memory Transformer Evaluation Evaluation: 0%| | 0/500 [00:00
Traceback (most recent call last): File "test.py", line 77, in scores = predict_captions(model, dict_dataloader_test, text_field) File "test.py", line 26, in predict_captions out, _ = model.beam_search(images, 20, text_field.vocab.stoi[''], 5) File "/home/mingjie/meshed-memory-transformer/models/captioning_model.py",...
I use https://github.com/peteanderson80/bottom-up-attention/ for feature extraction on my own images, and then run the image caption model, but the result caption is incomplete. e.g.  caption: "a view of a...
Hello, Many thanks for releasing the repo. I am trying to train a model on a custom variation of MSCOCO though I keep the train/test/valid sizes equal to the Karpathy...
I tried to test the model, but I got this problem. Does anybody know how to solve it? I used the environment provided here. python test.py --features_path data/coco_detections.hdf5 --annotation_folder annotations...
Hi,dear author: Thanks your great job,I am rencently working on RL training in another task. I take your scst training code to reference, and I got a terrible result. My...
Hi, Thanks for providing the implementation. I checked the paper and observe that only performance with reinforce optimization is reported. I wonder if you can report the performance of your...
Hi. When I run the test.py to evaluate, the model just generates without any other tokens.
Hello! Thanks for this great code! I find there is the visualization of attention states for captions, [like this](https://github.com/aimagelab/meshed-memory-transformer/blob/master/images/results.png). I want to reproduce the visualization results, but I cannot find...