pytorch异常——RuntimeError:Given groups=1, weight of size..., expected of...

文章目录

省流

  • nn.Conv2d 需要的输入张量格式为 (batch_size, channels, height, width),但您的示例输入张量 x 是 (batch_size, height, width, channels)。因此,需要对输入张量进行转置。

  • 注意,TensorFlow使用"NHWC"(批次、高度、宽度、通道)格式,而PyTorch使用"NCHW"(批次、通道、高度、宽度)格式

异常报错

bash 复制代码
RuntimeError: Given groups=1, weight of size [16, 3, 2, 3], 
expected input[8, 65, 66, 3] to have 3 channels, 
but got 65 channels instead

异常截图

异常代码

python 复制代码
def down_shifted_conv2d(x , num_filters , filters_size = [2,3],stride = 1, **kwargs):
    batch_size,H,W,channels = x.shape

    padding = (0,0,
        int(((filters_size[1]) - 1) / 2 ) , int((int(filters_size[1]) - 1) / 2),
        int(filters_size[0]) - 1 , 0,
        0,0)
    x_paded = nn.functional.pad(x, padding)
    print(x_paded.shape)
    conv_layer = nn.Conv2d(in_channels=channels, out_channels=num_filters, 
                           kernel_size=filters_size,
                           stride=stride, **kwargs)
    
    return conv_layer(x_paded)
# Example usage
x = torch.randn(8, 64, 64, 3)  # Example input with batch size 8, height and width 64, and 3 channels
num_filters = 16
output = down_shifted_conv2d(x, num_filters)
print(output.shape)

原因解释

  • 在pytorch中,"nn.Conv2d"需要输入的张量格式为(batch_size,channels,height,width),原图输入的x的格式是(batch_size,height ,weight,channel)所以需要对tensor进行转置。

  • 矩阵交换维度的函数permute,按照编号,将新的顺序填好即可

python 复制代码
def down_shifted_conv2d(x , num_filters , filters_size = [2,3], stride = 1, **kwargs):
    batch_size, H, W, channels = x.shape
    
    # Transpose the input tensor to (batch_size, channels, height, width)
    x = x.permute(0, 3, 1, 2)
    
    # Padding
    padding = (int((filters_size[1] - 1) / 2), int((filters_size[1] - 1) / 2),
               filters_size[0] - 1, 0)
    
    x_paded = F.pad(x, padding)

修正代码

python 复制代码
def down_shifted_conv2d(x , num_filters , filters_size = [2,3],stride = 1, **kwargs):
    batch_size,H,W,channels = x.shape
    # 按照顺序对4个维度分别进行填充
    padding = (0,0,
        int(((filters_size[1]) - 1) / 2 ) , int((int(filters_size[1]) - 1) / 2),
        int(filters_size[0]) - 1 , 0,
        0,0)
    x_paded = nn.functional.pad(x, padding)
    x_paded = x_paded.permute(0,3,1,2)
    # 进行卷积
    conv_layer = nn.Conv2d(in_channels=channels, out_channels=num_filters, 
                           kernel_size=filters_size,
                           stride=stride, **kwargs)
    
    return conv_layer(x_paded)
# Example usage
x = torch.randn(8, 64, 64, 3)  
num_filters = 16
output = down_shifted_conv2d(x, num_filters)
print(output.shape)

执行结果

相关推荐
Hoper.J8 小时前
PyTorch 模型保存与加载的三种常用方式
人工智能·pytorch·python
没有余地 EliasJie9 小时前
Windows Ubuntu下搭建深度学习Pytorch训练框架与转换环境TensorRT
pytorch·windows·深度学习·ubuntu·pycharm·conda·tensorflow
被制作时长两年半的个人练习生10 小时前
【pytorch】权重为0的情况
人工智能·pytorch·深度学习
GarryLau14 小时前
使用pytorch进行迁移学习的两个步骤
pytorch·迁移学习·torchvision
醒了就刷牙19 小时前
56 门控循环单元(GRU)_by《李沐:动手学深度学习v2》pytorch版
pytorch·深度学习·gru
橙子小哥的代码世界19 小时前
【深度学习】05-RNN循环神经网络-02- RNN循环神经网络的发展历史与演化趋势/LSTM/GRU/Transformer
人工智能·pytorch·rnn·深度学习·神经网络·lstm·transformer
最近好楠啊1 天前
Pytorch实现RNN实验
人工智能·pytorch·rnn
IFTICing1 天前
【文献阅读】Attention Bottlenecks for Multimodal Fusion
人工智能·pytorch·python·神经网络·学习·模态融合
醒了就刷牙1 天前
67 自注意力_by《李沐:动手学深度学习v2》pytorch版
人工智能·pytorch·深度学习
涩即是Null1 天前
PyTorch实现卷积神经网络CNN
人工智能·pytorch·深度学习·神经网络·卷积神经网络