GraphTERN icon indicating copy to clipboard operation
GraphTERN copied to clipboard

Fail to reproduce result

Open Pranav-chib opened this issue 2 years ago • 4 comments

Thankyou for such a wonderful paper, I like it alot but Fail to reproduce results. (0.42 / 0.58) after running the through read me i am getting (ADE: 0.95480105, FDE: 1.42633025) reproduced output. can you please specify some other setting why this is happening. thankyou

Pranav-chib avatar Jul 26 '23 11:07 Pranav-chib

Thanks for your interest in my work! In my experience, when training on the ETH set, early stopping between 16 and 32 epochs significantly improves performance.

Even so, FDE 1.4 is a value I've never seen before. Actually, I have not run training since early 2021 and only evaluated with pre-trained weights. I will check it out with the published code and let you know.

InhwanBae avatar Aug 07 '23 02:08 InhwanBae

Hi @Pranav-chib,

I'll share the results of my tests from the past week and some interesting findings with you. First, I was able to reproduce the table in the paper using the GitHub code I uploaded. What's interesting is that using the latest PyTorch release significantly reduces both the training speed and final performance. It even caused errors that interrupted the training. I recommend using version 1.6 or an older version as much as possible. (up to 1.11 seems to converge well)

Secondly, I have good news for you. I've made some updates to the source code a bit. Using the new codebase, you should be able to get better results!

InhwanBae avatar Aug 14 '23 15:08 InhwanBae

Hello, esteemed author. I have recently embarked upon replicating your code and encountered similar circumstances. Instead of creating a separate conda environment based on the setup you provided, I attempted to reproduce it using my existing environment. Now, I am facing a few issues. Firstly, when training on five datasets simultaneously with a single GPU, the training process is intermittently interrupted and does not reach the designated number of epochs. Additionally, there is a significant deviation in performance metrics. I am now prepared to attempt the replication using your recommended configuration.

LOCKE98 avatar Aug 15 '23 11:08 LOCKE98

Hi @LOCKE98, Changing or commenting out the seed may help solve the interruptions. Since I also ran 40 model training on an 8GPU server, running multiple experiments on a single GPU might not be the cause of the interruption.

InhwanBae avatar Aug 17 '23 06:08 InhwanBae