RGBD_Semantic_Segmentation_PyTorch icon indicating copy to clipboard operation
RGBD_Semantic_Segmentation_PyTorch copied to clipboard

Where's DeepLabv3+ model Decoder?

Open Albert-learner opened this issue 3 years ago • 3 comments

Hello, I'm trying to study about your code. But when I analysis your code, I can't find DeepLabV3+ Decoder in your code. Could you please tell me where the DeepLabV3+ model's Decoder?

Albert-learner avatar Apr 06 '22 08:04 Albert-learner

And I have another question. You're using one of Semantic Segmentation model, DeepLab V3+. But I know that the result of Semantic Segmentation result is [height, width, channels], and apply argmax so that I could get one channel array. When I analyze the result of Semantic Segmentation ath NYU Depth V2 dataset, the result array(inference DeepLab V3+) consists of different min cost and max cost at each channel array. I'd like to know that why result array of each category's range is different.

Albert-learner avatar Apr 07 '22 07:04 Albert-learner

Hello, I'm trying to study about your code. But when I analysis your code, I can't find DeepLabV3+ Decoder in your code. Could you please tell me where the DeepLabV3+ model's Decoder?

Hi, DeepLabV3+ decoder is defined here.

charlesCXK avatar May 06 '22 17:05 charlesCXK

And I have another question. You're using one of Semantic Segmentation model, DeepLab V3+. But I know that the result of Semantic Segmentation result is [height, width, channels], and apply argmax so that I could get one channel array. When I analyze the result of Semantic Segmentation ath NYU Depth V2 dataset, the result array(inference DeepLab V3+) consists of different min cost and max cost at each channel array. I'd like to know that why result array of each category's range is different.

The absolute values of the result array do not make much sense.

  • During training, the result array is supervised by the CrossEntropy loss, which pushes the confidence of the correct category to be higher than other categories. In the meanwhile, the SoftMax operation in CrossEntropy loss hopes that the confidence of the correct category is significantly different from the other categories. For other categories, we don't care what their confidence numbers are, just make them as low as possible.
  • During inference, we simply select the category with the maximum confidence as the prediction, which is consistent with the training.

charlesCXK avatar May 06 '22 17:05 charlesCXK