a-PyTorch-Tutorial-to-Image-Captioning
a-PyTorch-Tutorial-to-Image-Captioning copied to clipboard
what is the lowest loss we can achieve with this architecture?
I tried to overfitting the model but it seems so hard to achieve
I've minimize the data into 100 image with 1 captions per image and the lowest loss I can achieve when training in 500 epochs is around 0.9xxx and the loss value is stagnant till 1000 epoch
what is the lowest loss you've achieve @sgrvinod ?
Hey, sorry, I've been really busy and I'll respond to your email (and here) in the next day or two.
I don't remember the loss values, especially the training loss - but I also don't recall noticing that it was stagnant or slowing down. I wouldn't think that the model couldn't overfit.
Is this on (a small part of) MSCOCO or another dataset?
the data is from flickr8k images and karpathy's json file but I cut it to 100 training image and 20 validation image here are the json in my gist
Hi,
When train on coco, mine minimum is about 3.x. Is it normal? My generated captions seems very bad also...