class SizeBlock(nn.Module): def __init__(self, conv): super(SizeBlock, self).__init__() self.conv, inc = nc2dc(conv) self.glob = nn.Sequential( nn.Linear(2, 64), nn.ReLU(inplace=True), nn.Linear(64, 32) ) self.local = nn.Sequential( nn.Conv2d(inc, 32, 3, padding=1), nn.ReLU(inplace=True), nn.Conv2d(32, 32, 3, padding=1) ) self.fuse = nn.Sequential( nn.Conv2d(64, 32, 3, padding=1), nn.ReLU(inplace=True), nn.Conv2d(32, 3 * 3 * 2, 3, padding=1) ) self.relu = nn.ReLU() def forward(self, x, bsize): b, c, h, w = x.shape g_offset = self.glob(bsize) g_offset = g_offset.view(b, -1, 1, 1).repeat(1, 1, h, w).contiguous() l_offset = self.local(x) offset = self.fuse(torch.cat((g_offset, l_offset), dim=1)) fea = self.conv(x, offset) return self.relu(fea)和class ResBase(nn.Module): def __init__(self, res_name): super(ResBase, self).__init__() # model_resnet = res_dict[res_name](pretrained=False, norm_layer=BN_2D) model_resnet = res_dict[res_name](pretrained=True) self.sizeblock = SizeBlock self.conv1 = model_resnet.conv1 self.bn1 = model_resnet.bn1 self.relu = model_resnet.relu self.maxpool = model_resnet.maxpool self.layer1 = model_resnet.layer1 self.layer2 = model_resnet.layer2 self.layer3 = model_resnet.layer3 self.layer4 = model_resnet.layer4 self.avgpool = model_resnet.avgpool self.in_features = model_resnet.fc.in_features def forward(self, x, msize): print(x.shape) # torch.Size([8, 3, 384, 384]) x = self.sizeblock(x, msize) x = self.conv1(x) print(x.shape) # torch.Size([8, 64, 192, 192]) x = self.bn1(x) x = self.relu(x) # x = self.self.selist[1](x, msize) x = self.maxpool(x) print(x.shape) # torch.Size([8, 64, 96, 96]) x = self.layer1(x) print(x.shape) # torch.Size([8, 256, 96, 96]) # x = self.self.selist[2](x, msize) x = self.layer2(x) print(x.shape) # torch.Size([8, 512, 48, 48]) # x = self.self.selist[3](x, msize) x = self.layer3(x) # print(x.shape) # torch.Size([8, 1024, 24, 24]) x = self.layer4(x) # print(x.shape) # torch.Size([8, 2048, 12, 12]) x = self.avgpool(x) print(x.shape) # torch.Size([8, 2048, 1, 1]) x = x.view(x.size(0), -1) print(x.shape) # torch.Size([8, 2048]) a = input() return x,如何使用SizeBlock的forward函数
时间: 2023-05-31 22:07:23 浏览: 214
浅析PyTorch中nn.Module的使用
5星 · 资源好评率100%
这是两个PyTorch的神经网络模块的定义代码。第一个模块是一个尺寸块,包含一个全局特征模块和一个本地特征模块,同时还有一个融合模块。该模块的输入是一组图像和一个块大小,并通过卷积层返回处理后的特征图像。第二个模块是一个残差基础模块,继承了PyTorch的nn.Module类。
阅读全文