a-PyTorch-Tutorial-to-Image-Captioning
a-PyTorch-Tutorial-to-Image-Captioning copied to clipboard
Show, Attend, and Tell | a PyTorch Tutorial to Image Captioning
Hello, can this model detect and recognize text in images containing text
The model was trained on flicker8k, but the results achieved only half the BLEU-4 score mentioned by the authors (about 0.14-0.15). I have not modified any parameters in train.py. May...
How to slove this problem in eval.py?
I encounter with a bug which is "**ValueError: max() arg is an empty sequence**" when I run **caption.py**. I find that the parameter **complete_inds** and **complete_seqs_scores** is null during the...
Traceback (most recent call last): File "D:\majority_design\image_caption\ic_train.py", line 322, in main() File "D:\majority_design\image_caption\ic_train.py", line 116, in main train(train_loader=train_loader, File "D:\majority_design\image_caption\ic_train.py", line 180, in train loss = criterion(scores, targets) File "D:\anaconda3\envs\pytorch\lib\site-packages\torch\nn\modules\module.py",...
Silly question. I'm not following how to implement this model? Is there a step by step example notebook anywhere in which I can review?
I have tried to download the dataset but the link does not redirect anywhere. Maybe you have the dataset in another place or can you upload it to another place...
When I train a new model in flickr8k and flickr30k dataset in my environment, I find that the **trianing loss is too high**(about 10) and the **bleu-4 is too low**(about...
File "train.py", line 329, in main() File "train.py", line 116, in main epoch=epoch) File "train.py", line 184, in train loss += alpha_c * ((1. - alphas.sum(dim=1)) ** 2).mean() UnboundLocalError: local...