linnan wang
linnan wang
Hello @rwightman , Thank you so much for submitting this bug report; we're working on fixing the license issue and sincerely apologize for not doing it right in the first...
> @linnanwang thanks for updating the copyright/license info > > The issue with padding is that the 3x3 convs in the EdgeResidual (FusedMBConv) layers have 0 padding, for typical PyTorch...
@rwightman Also I really appreciate your advices in the naming convention. Yes, our stem and head is opposite to the conventions in timm and the community and we will revise...
@rwightman Thanks Ross. It will be great to see GPUNet into Timm's library. I'm talking to the team to see our bandwidth and make a properly plan ahead. But I'm...
Hello @rwightman, we have changed the naming and padding of GPUNet and the following is the **new** network structure from get_configs(batch=1, latency="0.65ms", gpuType="GV100") on our side. Could you please kindly...
Thank you @rwightman. The activation in stem/head are all RELU now. Another thing I want to confirm with you is our customized InvertedResidual. We search for its internal structure and...
Thank you @rwightman. I'm pasting all the model definitions here for you to double check with Timm's compatibility. Please note re-train these models are very expensive, so let's try to...
get_configs(batch=1, latency="1.75ms", gpuType="GV100") ``` GPUNet( (network): Sequential( (stem: 2): Prologue( (net): Sequential( (0): Conv2d(3, 32, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), bias=False) (1): BatchNorm2d(32, eps=0.001, momentum=0.1, affine=True, track_running_stats=True) (2): ReLU(inplace=True)...
get_configs(batch=1, latency="0.5ms-D", gpuType="GV100") ``` GPUNet( (network): Sequential( (stem: 2): Prologue( (net): Sequential( (0): Conv2d(3, 32, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), bias=False) (1): BatchNorm2d(32, eps=0.001, momentum=0.1, affine=True, track_running_stats=True) (2): ReLU(inplace=True)...
get_configs(batch=1, latency="0.8ms-D", gpuType="GV100") ``` GPUNet( (network): Sequential( (stem: 2): Prologue( (net): Sequential( (0): Conv2d(3, 32, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), bias=False) (1): BatchNorm2d(32, eps=0.001, momentum=0.1, affine=True, track_running_stats=True) (2): ReLU(inplace=True)...