kaggle-carvana-image-masking-challenge icon indicating copy to clipboard operation
kaggle-carvana-image-masking-challenge copied to clipboard

Top 15% ranked solution to the Carvana Image Masking Challenge on Kaggle

This repository contains my personal code for the Kaggle competition - Carvana Image Masking Challenge.

NBViewer Link Description Train set loss / dice coeff Validation set loss / dice coeff Kaggle Public LB Dice coeff
000-general-viz-explore.ipynb Exploration - - -
01-unet-128x128.ipynb U-Net; 128x128; Contracting path filter sizes: 64, 128, 256, 512, 1024 0.0191; 0.9899 0.0130; 0.9939 0.9908
02-unet-256x256.ipynb U-Net; 256x256; Contracting path filter sizes: 32, 64, 128, 256, 512, 1024 0.0122; 0.9936 0.0100; 0.9951 0.9942
03-unet-512x512.ipynb U-Net; 512x512; Contracting path filter sizes: 16, 32, 64, 128, 256, 512, 1024 0.0095; 0.9950 0.0089; 0.9954 0.9958
04-unet-1024x1024.ipynb U-Net; 1024x1024; Contracting path filter sizes: 8, 16, 32, 64, 128, 256, 512, 1024 0.0084; 0.9956 0.0085; 0.9956 0.9961
07-threshold-analysis.ipynb Check to see if tuning the threshold value makes any difference to the score. - - -
08-bounding-boxes.ipynb Using the model from Notebook 2, bounding boxes are first visualized, and then calculated for both train and test sets. - - -
09-full-res-model-vertical-cut-and-bbox.ipynb U-Net; 1024x1024; Contracting path filter sizes: 8, 16, 32, 64, 128, 256, 512, 1024; Vertical height of the original image is first reduced to 1024 from 1280, using bounding boxes. The resulting 1918x1024 image is cut vertically (with some overlap) to two images, each of size 1024x1024. 0.0071; 0.9967 0.0077; 0.9966 0.9967
10-model-predictions-viz.ipynb Models from notebooks 3,4, 9 are compared. Also worst predictions from Notebook 9's model are visualized - - -
11-ensemble.ipynb Ensemble of Notebook 3 and 9 models, by taking average of predictions from both models. - - 0.9967
12-tiramisu-256.ipynb One Hundred Layers Tiramisu; 256x256; Didn't fully train. Each epoch was taking way too long time to complete even for 256x256 resolution. - - -
13-bounding-boxes-revisit.ipynb Boundary boxes are analysed further - - -
14-full-res-model-1920-1080.ipynb U-Net; 1920x1280; Contracting path filter sizes: 8, 16, 32, 64, 128, 256, 512, 1024; The output from Notebook 3's model is added as an additional depth layer, after resizing to full resolution. 0.0030; 0.9986 0.0139; 0.9958 -
16-unet-inception.ipynb UNet's VGG style Conv-BN-Act layers are replaced by InceptionV3 style blocks 0.0184; 0.9905 0.0163; 0.9915 -
17-full-res-model-angle-09-bbox.ipynb U-Net; 1280x896; Contracting path filter sizes: 8, 16, 32, 64, 128, 256, 512, 1024; Original images are cut to 1280x896 from 1918x1280, using bounding boxes. 0.0072; 0.9969 0.0120; 0.9957 -
18-full-res-model-angle-09-bbox-no-maxpool.ipynb Replaced max pooling from the above model, with 2x2 Convolution layer with 2x2 stride 0.0283; 0.9880 0.0417; 0.9844 -
19-full-res-model-all-angles-vertical-cut-no-bbox.ipynb The original 1918x1280 image is cut into two halves vertically (with some overlap), each with size 1024x1280. 0.0064; 0.9968 0.0069; 0.9966 0.9967
20-full-res-model-all-angles-horizontal-cut-no-bbox.ipynb The original 1918x1280 image is cut into two halves horizontally (with some overlap), each with size 1920x768. - - 0.9967
21-ensemble.ipynb Ensemble of models from Notebooks 19 and 20 (averaged the predictions) - - 0.9968
22-unet-different-initial-filters.ipynb U-net architecture with more filters in the first layer - - -

The Private Leaderboard score for the Ensemble model was 0.9965.