self-supervised-depth-completion icon indicating copy to clipboard operation
self-supervised-depth-completion copied to clipboard

training with vlp-16 dataset

Open tzuchiehTT opened this issue 5 years ago • 7 comments

Hi Fangchang,

Our lab is currently working on a project which requires generating depth maps from our vlp-16 lidar and camera setting. Your work looks great as the depth map solution. Since we got different size images as input, I think what we need to do to use this network is (1) read in our own calibration information (K) and (2) crop input images as (width, high) both multiples of 16 (since we got errors when going through decode layers with some other sizes), is that right?

We've tested with a rather small dataset (only ~700 frames) and got results like the figure showing below. We are wondering if the dataset is too small or the depth info from vlp-16 is too sparse since the results remain clear projected lines. It would be great if you have any suggestions, thanks!

comparison_best

tzuchiehTT avatar Apr 08 '19 19:04 tzuchiehTT

For cameras with different intrinsic parameters than the KITTI cameras, there is a slightly more complicated image resizing process than simple cropping. In particular, we need to resize the image in a way such that the resized image has the same effective intrinsics as KITTI.

fangchangma avatar Apr 13 '19 04:04 fangchangma

Hi Fangchang, I have the same question. Why can't we just use our own intrinsic parameters instead of resizing images? Is it related to some hyperparameters needed to tune?

XiaotaoGuo avatar Aug 05 '19 07:08 XiaotaoGuo

@XiaotaoGuo It might or might not work well with a different set of intrinsics, but there is simply no guarantee that the trained network would transfer directly to this new set of intrinsics (and image sizes). My suggestion is to keep the test images as close to the train images as possible.

fangchangma avatar Aug 05 '19 19:08 fangchangma

Thanks! What if we use our own dataset to train the network and test with it?

XiaotaoGuo avatar Aug 06 '19 02:08 XiaotaoGuo

What if we use our own dataset to train the network and test with it?

Then there is no need for any resizing

fangchangma avatar Aug 06 '19 23:08 fangchangma

Hi @MelvinTT, Did you manage to work this out? I am fighting similar problems with my own dataset, generated with a VLP32 and a Zed camera.

christian-lanius avatar Sep 10 '19 13:09 christian-lanius

What if we use our own dataset to train the network and test with it?

Then there is no need for any resizing

Hi, When we use 'resize' operation, how to deal with the sparse depth? Could it be correct to just resize the sparse depth input like the rgb images, although it is sparse? Thanks!

longyangqi avatar Apr 05 '20 04:04 longyangqi