MagicDrive icon indicating copy to clipboard operation
MagicDrive copied to clipboard

Generated Data for Augmentation

Open mrabiabrn opened this issue 1 year ago • 14 comments

Hi,

I noticed the code for validation set generations but didn't find any code for training data augmentation. Should we follow the same procedure for generating training data? Could you provide more details on this or share a link to the generated data for BEV perception? That would be greatly appreciated.

mrabiabrn avatar Oct 01 '24 11:10 mrabiabrn

Yes, you can refer to the procedure for validation set. They are similar.

flymin avatar Oct 01 '24 11:10 flymin

Hi, I tried to generate new samples from training data using the provided validation set generation script. However, I realized that for training instances, generations are not diverse and quite similar to the original data (color of the vehicles, shape of the road, background, etc.) This is not the case for validation samples. I can see diverse generations for the same bounding boxes. I added examples from training and validation generation results below. What do you think could be the reason for this?

Validation original vs generated image

Training original vs generated image

mrabiabrn avatar Oct 07 '24 12:10 mrabiabrn

In some cases, it may happen. However, using such data to augment the original training set leads to improvements in downstream tasks.

If it is severe in your case, you can try editing the scene condition to generate different data for augmentation.

flymin avatar Oct 09 '24 10:10 flymin

This issue is stale because it has been open for 7 days with no activity. If you do not have any follow-ups, the issue will be closed soon.

github-actions[bot] avatar Oct 16 '24 16:10 github-actions[bot]

Training generations are generally like this in my case. Augmenting training data with this doesn't improve CVT performance, it even hurts it. I can try editing the scene and text condition, but to reproduce your results, it would be great if you could share your training and validation set generations so I can identify any discrepancies.

mrabiabrn avatar Oct 17 '24 07:10 mrabiabrn

We already released the model weights. You can sample our model and see it.

I cannot share the data. However, I have to admit that our cases on the training set are similar to yours. We did not modify the code for perception models; we only added more generated data as described in our paper. Maybe you can also try bevfusion and see.

flymin avatar Oct 17 '24 09:10 flymin

This issue is stale because it has been open for 7 days with no activity. If you do not have any follow-ups, the issue will be closed soon.

github-actions[bot] avatar Oct 25 '24 16:10 github-actions[bot]

@flymin Hello author, the road segmentation performance of CVT in Table 1 is 61, which is confirmed to be 59.3 in Table 4. We also reproduced 59.x. How was 61 obtained? Or what is the difference between these two data?

RYHSmmc avatar Oct 31 '24 06:10 RYHSmmc

which is confirmed to be 59.3 in Table 4

This is not true. Please also see Figure 7. I think the problem lies in $M={0}$.

flymin avatar Oct 31 '24 08:10 flymin

This issue is stale because it has been open for 7 days with no activity. If you do not have any follow-ups, the issue will be closed soon.

github-actions[bot] avatar Nov 08 '24 16:11 github-actions[bot]

To confirm, the results in Table 1 were generated with use_zero_map_as_unconditional = True and a guidance scale of 2. Is this correct?

mrabiabrn avatar Nov 12 '24 06:11 mrabiabrn

Yes. And sorry for the late reply.

flymin avatar Nov 18 '24 08:11 flymin

No problem at all, and thanks for clarifying!

mrabiabrn avatar Nov 18 '24 08:11 mrabiabrn

This issue is stale because it has been open for 7 days with no activity. If you do not have any follow-ups, the issue will be closed soon.

github-actions[bot] avatar Nov 26 '24 16:11 github-actions[bot]