I am implementing the mobilenet v3 paper and face the error.

In the paper, the figure 10 shows that after 16 stride they use the average pooling

`

import torch

import torch.nn as nn

import torch.nn.functional as F

class _LRASPP(nn.Module):

“”“Lite R-ASPP”""`def __init__(self, in_channels, norm_layer, **kwargs): super(_LRASPP, self).__init__() out_channels = 128 self.b0 = nn.Sequential( nn.Conv2d(in_channels, out_channels, 1, bias=False), norm_layer(out_channels), nn.ReLU(True) ) self.b1 = nn.Sequential( nn.AvgPool2d(kernel_size=(49, 49), stride=(16, 20)), # check it nn.Conv2d(in_channels, out_channels, 1, bias=False), nn.Sigmoid(), ) def forward(self, x): size = x.size()[2:] feat1 = self.b0(x) feat2 = self.b1(x) feat2 = F.interpolate(feat2, size, mode='bilinear', align_corners=True) x = feat1 * feat2 # check it return x`

if

name== ‘main’:

net = _LRASPP(in_channels=576,norm_layer=nn.BatchNorm2d)`input_size=(1, 576,64, 32) net=net.cuda() x = torch.randn(input_size) x=x.cuda() net.eval() for i in range(10): out = net(x) print(out.shape)`

`

However I face the error about

RuntimeError: Given input size: (576x64x32). Calculated output size: (576x1x0). Output size is too small

In the paper, they use the 1024 x 512 size input. Which mean, their final output is 64 32 after 16 stride.

My question is how can we apply the 49 x 49 average pooling?

If I use the stride 8, of course I can apply it but the paper said they use 16 strides or 32 strides with or without LR ASPP. So It does not make any sense. Also no extra information is exists. Any idea for this?

I tested the pytorch 1.0.0 and this code is working but 1.6 verstion this code has error.

Still pooling operation has issus about it.