loaderr

Traceback (most recent call last):

File "/ai/zhdata/lyp/multiyolov5_point_608_736/train_608_736.py", line 718, in <module>

train(hyp, opt, device, tb_writer)

File "/ai/zhdata/lyp/multiyolov5_point_608_736/train_608_736.py", line 166, in train

ema.ema.load_state_dict(ckpt['ema'].float().state_dict())

File "/ai/zhdata/lyp/conda/anaconda3/envs/mmd3.0/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1482, in load_state_dict

raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(

RuntimeError: Error(s) in loading state_dict for Model:

Missing key(s) in state_dict: "model.24.m.0.cv1.conv.weight", "model.24.m.0.cv1.bn.weight", "model.24.m.0.cv1.bn.bias", "model.24.m.0.cv1.bn.running_mean", "model.24.m.0.cv1.bn.running_var", "model.24.m.0.cv2.conv.weight", "model.24.m.0.cv2.bn.weight", "model.24.m.0.cv2.bn.bias", "model.24.m.0.cv2.bn.running_mean", "model.24.m.0.cv2.bn.running_var", "model.24.m.0.cv3.conv.weight", "model.24.m.0.cv3.bn.weight", "model.24.m.0.cv3.bn.bias", "model.24.m.0.cv3.bn.running_mean", "model.24.m.0.cv3.bn.running_var", "model.24.m.0.m.0.cv1.conv.weight", "model.24.m.0.m.0.cv1.bn.weight", "model.24.m.0.m.0.cv1.bn.bias", "model.24.m.0.m.0.cv1.bn.running_mean", "model.24.m.0.m.0.cv1.bn.running_var", "model.24.m.0.m.0.cv2.conv.weight", "model.24.m.0.m.0.cv2.bn.weight", "model.24.m.0.m.0.cv2.bn.bias", "model.24.m.0.m.0.cv2.bn.running_mean", "model.24.m.0.m.0.cv2.bn.running_var", "model.24.m.2.m.0.cv1.conv.weight", "model.24.m.2.m.0.cv1.bn.weight", "model.24.m.2.m.0.cv1.bn.bias", "model.24.m.2.m.0.cv1.bn.running_mean", "model.24.m.2.m.0.cv1.bn.running_var", "model.24.m.2.m.0.cv2.conv.weight", "model.24.m.2.m.0.cv2.bn.weight", "model.24.m.2.m.0.cv2.bn.bias", "model.24.m.2.m.0.cv2.bn.running_mean", "model.24.m.2.m.0.cv2.bn.running_var", "model.24.m.4.cv1.conv.weight", "model.24.m.4.cv1.bn.weight", "model.24.m.4.cv1.bn.bias", "model.24.m.4.cv1.bn.running_mean", "model.24.m.4.cv1.bn.running_var", "model.24.m.4.cv2.conv.weight", "model.24.m.4.cv2.bn.weight", "model.24.m.4.cv2.bn.bias", "model.24.m.4.cv2.bn.running_mean", "model.24.m.4.cv2.bn.running_var", "model.24.m.4.cv3.conv.weight", "model.24.m.4.cv3.bn.weight", "model.24.m.4.cv3.bn.bias", "model.24.m.4.cv3.bn.running_mean", "model.24.m.4.cv3.bn.running_var", "model.24.m.4.m.0.cv1.conv.weight", "model.24.m.4.m.0.cv1.bn.weight", "model.24.m.4.m.0.cv1.bn.bias", "model.24.m.4.m.0.cv1.bn.running_mean", "model.24.m.4.m.0.cv1.bn.running_var", "model.24.m.4.m.0.cv2.conv.weight", "model.24.m.4.m.0.cv2.bn.weight", "model.24.m.4.m.0.cv2.bn.bias", "model.24.m.4.m.0.cv2.bn.running_mean", "model.24.m.4.m.0.cv2.bn.running_var", "model.24.m.5.weight", "model.24.m.5.bias", "model.24.decoder1.3.cv1.conv.weight", "model.24.decoder1.3.cv1.bn.weight", "model.24.decoder1.3.cv1.bn.bias", "model.24.decoder1.3.cv1.bn.running_mean", "model.24.decoder1.3.cv1.bn.running_var", "model.24.decoder1.3.cv2.conv.weight", "model.24.decoder1.3.cv2.bn.weight", "model.24.decoder1.3.cv2.bn.bias", "model.24.decoder1.3.cv2.bn.running_mean", "model.24.decoder1.3.cv2.bn.running_var", "model.24.decoder1.3.cv3.conv.weight", "model.24.decoder1.3.cv3.bn.weight", "model.24.decoder1.3.cv3.bn.bias", "model.24.decoder1.3.cv3.bn.running_mean", "model.24.decoder1.3.cv3.bn.running_var", "model.24.decoder1.3.m.0.cv1.conv.weight", "model.24.decoder1.3.m.0.cv1.bn.weight", "model.24.decoder1.3.m.0.cv1.bn.bias", "model.24.decoder1.3.m.0.cv1.bn.running_mean", "model.24.decoder1.3.m.0.cv1.bn.running_var", "model.24.decoder1.3.m.0.cv2.conv.weight", "model.24.decoder1.3.m.0.cv2.bn.weight", "model.24.decoder1.3.m.0.cv2.bn.bias", "model.24.decoder1.3.m.0.cv2.bn.running_mean", "model.24.decoder1.3.m.0.cv2.bn.running_var", "model.24.decoder1.5.cv1.conv.weight", "model.24.decoder1.5.cv1.bn.weight", "model.24.decoder1.5.cv1.bn.bias", "model.24.decoder1.5.cv1.bn.running_mean", "model.24.decoder1.5.cv1.bn.running_var", "model.24.decoder1.5.cv2.conv.weight", "model.24.decoder1.5.cv2.bn.weight", "model.24.decoder1.5.cv2.bn.bias", "model.24.decoder1.5.cv2.bn.running_mean", "model.24.decoder1.5.cv2.bn.running_var", "model.24.decoder1.5.cv3.conv.weight", "model.24.decoder1.5.cv3.bn.weight", "model.24.decoder1.5.cv3.bn.bias", "model.24.decoder1.5.cv3.bn.running_mean", "model.24.decoder1.5.cv3.bn.running_var", "model.24.decoder1.5.m.0.cv1.conv.weight", "model.24.decoder1.5.m.0.cv1.bn.weight", "model.24.decoder1.5.m.0.cv1.bn.bias", "model.24.decoder1.5.m.0.cv1.bn.running_mean", "model.24.decoder1.5.m.0.cv1.bn.running_var", "model.24.decoder1.5.m.0.cv2.conv.weight", "model.24.decoder1.5.m.0.cv2.bn.weight", "model.24.decoder1.5.m.0.cv2.bn.bias", "model.24.decoder1.5.m.0.cv2.bn.running_mean", "model.24.decoder1.5.m.0.cv2.bn.running_var", "model.24.decoder1.6.weight", "model.24.decoder1.6.bias", "model.24.m32.2.cv1.conv.weight", "model.24.m32.2.cv1.bn.weight", "model.24.m32.2.cv1.bn.bias", "model.24.m32.2.cv1.bn.running_mean", "model.24.m32.2.cv1.bn.running_var", "model.24.m32.2.cv2.conv.weight", "model.24.m32.2.cv2.bn.weight", "model.24.m32.2.cv2.bn.bias", "model.24.m32.2.cv2.bn.running_mean", "model.24.m32.2.cv2.bn.running_var", "model.24.m32.2.cv3.conv.weight", "model.24.m32.2.cv3.bn.weight", "model.24.m32.2.cv3.bn.bias", "model.24.m32.2.cv3.bn.running_mean", "model.24.m32.2.cv3.bn.running_var", "model.24.m32.2.m.0.cv1.conv.weight", "model.24.m32.2.m.0.cv1.bn.weight", "model.24.m32.2.m.0.cv1.bn.bias", "model.24.m32.2.m.0.cv1.bn.running_mean", "model.24.m32.2.m.0.cv1.bn.running_var", "model.24.m32.2.m.0.cv2.conv.weight", "model.24.m32.2.m.0.cv2.bn.weight", "model.24.m32.2.m.0.cv2.bn.bias", "model.24.m32.2.m.0.cv2.bn.running_mean", "model.24.m32.2.m.0.cv2.bn.running_var", "model.24.m16.0.conv.weight", "model.24.m16.0.bn.weight", "model.24.m16.0.bn.bias", "model.24.m16.0.bn.running_mean", "model.24.m16.0.bn.running_var".

Unexpected key(s) in state_dict: "model.24.m.0.conv.weight", "model.24.m.0.bn.weight", "model.24.m.0.bn.bias", "model.24.m.0.bn.running_mean", "model.24.m.0.bn.running_var", "model.24.m.0.bn.num_batches_tracked", "model.24.m.1.cv1.conv.weight", "model.24.m.1.cv1.bn.weight", "model.24.m.1.cv1.bn.bias", "model.24.m.1.cv1.bn.running_mean", "model.24.m.1.cv1.bn.running_var", "model.24.m.1.cv1.bn.num_batches_tracked", "model.24.m.1.cv2.conv.weight", "model.24.m.1.cv2.bn.weight", "model.24.m.1.cv2.bn.bias", "model.24.m.1.cv2.bn.running_mean", "model.24.m.1.cv2.bn.running_var", "model.24.m.1.cv2.bn.num_batches_tracked", "model.24.m.1.cv3.conv.weight", "model.24.m.1.cv3.bn.weight", "model.24.m.1.cv3.bn.bias", "model.24.m.1.cv3.bn.running_mean", "model.24.m.1.cv3.bn.running_var", "model.24.m.1.cv3.bn.num_batches_tracked", "model.24.m.1.m.0.cv1.conv.weight", "model.24.m.1.m.0.cv1.bn.weight", "model.24.m.1.m.0.cv1.bn.bias", "model.24.m.1.m.0.cv1.bn.running_mean", "model.24.m.1.m.0.cv1.bn.running_var", "model.24.m.1.m.0.cv1.bn.num_batches_tracked", "model.24.m.1.m.0.cv2.conv.weight", "model.24.m.1.m.0.cv2.bn.weight", "model.24.m.1.m.0.cv2.bn.bias", "model.24.m.1.m.0.cv2.bn.running_mean", "model.24.m.1.m.0.cv2.bn.running_var", "model.24.m.1.m.0.cv2.bn.num_batches_tracked", "model.24.m.2.m.cv1.conv.weight", "model.24.m.2.m.cv1.bn.weight", "model.24.m.2.m.cv1.bn.bias", "model.24.m.2.m.cv1.bn.running_mean", "model.24.m.2.m.cv1.bn.running_var", "model.24.m.2.m.cv1.bn.num_batches_tracked", "model.24.m.2.m.cv2.conv.weight", "model.24.m.2.m.cv2.bn.weight", "model.24.m.2.m.cv2.bn.bias", "model.24.m.2.m.cv2.bn.running_mean", "model.24.m.2.m.cv2.bn.running_var", "model.24.m.2.m.cv2.bn.num_batches_tracked", "model.24.m.3.weight", "model.24.m.3.bias", "model.24.decoder1.2.cv1.conv.weight", "model.24.decoder1.2.cv1.bn.weight", "model.24.decoder1.2.cv1.bn.bias", "model.24.decoder1.2.cv1.bn.running_mean", "model.24.decoder1.2.cv1.bn.running_var", "model.24.decoder1.2.cv1.bn.num_batches_tracked", "model.24.decoder1.2.cv2.conv.weight", "model.24.decoder1.2.cv2.bn.weight", "model.24.decoder1.2.cv2.bn.bias", "model.24.decoder1.2.cv2.bn.running_mean", "model.24.decoder1.2.cv2.bn.running_var", "model.24.decoder1.2.cv2.bn.num_batches_tracked", "model.24.decoder1.2.cv3.conv.weight", "model.24.decoder1.2.cv3.bn.weight", "model.24.decoder1.2.cv3.bn.bias", "model.24.decoder1.2.cv3.bn.running_mean", "model.24.decoder1.2.cv3.bn.running_var", "model.24.decoder1.2.cv3.bn.num_batches_tracked", "model.24.decoder1.2.m.cv1.conv.weight", "model.24.decoder1.2.m.cv1.bn.weight", "model.24.decoder1.2.m.cv1.bn.bias", "model.24.decoder1.2.m.cv1.bn.running_mean", "model.24.decoder1.2.m.cv1.bn.running_var", "model.24.decoder1.2.m.cv1.bn.num_batches_tracked", "model.24.decoder1.2.m.cv2.conv.weight", "model.24.decoder1.2.m.cv2.bn.weight", "model.24.decoder1.2.m.cv2.bn.bias", "model.24.decoder1.2.m.cv2.bn.running_mean", "model.24.decoder1.2.m.cv2.bn.running_var", "model.24.decoder1.2.m.cv2.bn.num_batches_tracked", "model.24.decoder1.3.weight", "model.24.decoder1.3.bias", "model.24.m8.1.cv1.conv.weight", "model.24.m8.1.cv1.bn.weight", "model.24.m8.1.cv1.bn.bias", "model.24.m8.1.cv1.bn.running_mean", "model.24.m8.1.cv1.bn.running_var", "model.24.m8.1.cv1.bn.num_batches_tracked", "model.24.m8.1.cv2.conv.weight", "model.24.m8.1.cv2.bn.weight", "model.24.m8.1.cv2.bn.bias", "model.24.m8.1.cv2.bn.running_mean", "model.24.m8.1.cv2.bn.running_var", "model.24.m8.1.cv2.bn.num_batches_tracked", "model.24.m8.1.cv3.conv.weight", "model.24.m8.1.cv3.bn.weight", "model.24.m8.1.cv3.bn.bias", "model.24.m8.1.cv3.bn.running_mean", "model.24.m8.1.cv3.bn.running_var", "model.24.m8.1.cv3.bn.num_batches_tracked", "model.24.m8.1.m.0.cv1.conv.weight", "model.24.m8.1.m.0.cv1.bn.weight", "model.24.m8.1.m.0.cv1.bn.bias", "model.24.m8.1.m.0.cv1.bn.running_mean", "model.24.m8.1.m.0.cv1.bn.running_var", "model.24.m8.1.m.0.cv1.bn.num_batches_tracked", "model.24.m8.1.m.0.cv2.conv.weight", "model.24.m8.1.m.0.cv2.bn.weight", "model.24.m8.1.m.0.cv2.bn.bias", "model.24.m8.1.m.0.cv2.bn.running_mean", "model.24.m8.1.m.0.cv2.bn.running_var", "model.24.m8.1.m.0.cv2.bn.num_batches_tracked", "model.24.m32.1.cv1.conv.weight", "model.24.m32.1.cv1.bn.weight", "model.24.m32.1.cv1.bn.bias", "model.24.m32.1.cv1.bn.running_mean", "model.24.m32.1.cv1.bn.running_var", "model.24.m32.1.cv1.bn.num_batches_tracked", "model.24.m32.1.cv2.conv.weight", "model.24.m32.1.cv2.bn.weight", "model.24.m32.1.cv2.bn.bias", "model.24.m32.1.cv2.bn.running_mean", "model.24.m32.1.cv2.bn.running_var", "model.24.m32.1.cv2.bn.num_batches_tracked", "model.24.m32.1.cv3.conv.weight", "model.24.m32.1.cv3.bn.weight", "model.24.m32.1.cv3.bn.bias", "model.24.m32.1.cv3.bn.running_mean", "model.24.m32.1.cv3.bn.running_var", "model.24.m32.1.cv3.bn.num_batches_tracked", "model.24.m32.1.m.0.cv1.conv.weight", "model.24.m32.1.m.0.cv1.bn.weight", "model.24.m32.1.m.0.cv1.bn.bias", "model.24.m32.1.m.0.cv1.bn.running_mean", "model.24.m32.1.m.0.cv1.bn.running_var", "model.24.m32.1.m.0.cv1.bn.num_batches_tracked", "model.24.m32.1.m.0.cv2.conv.weight", "model.24.m32.1.m.0.cv2.bn.weight", "model.24.m32.1.m.0.cv2.bn.bias", "model.24.m32.1.m.0.cv2.bn.running_mean", "model.24.m32.1.m.0.cv2.bn.running_var", "model.24.m32.1.m.0.cv2.bn.num_batches_tracked", "model.24.m16.0.cv1.conv.weight", "model.24.m16.0.cv1.bn.weight", "model.24.m16.0.cv1.bn.bias", "model.24.m16.0.cv1.bn.running_mean", "model.24.m16.0.cv1.bn.running_var", "model.24.m16.0.cv1.bn.num_batches_tracked", "model.24.m16.0.cv2.conv.weight", "model.24.m16.0.cv2.bn.weight", "model.24.m16.0.cv2.bn.bias", "model.24.m16.0.cv2.bn.running_mean", "model.24.m16.0.cv2.bn.running_var", "model.24.m16.0.cv2.bn.num_batches_tracked", "model.24.m16.0.cv3.conv.weight", "model.24.m16.0.cv3.bn.weight", "model.24.m16.0.cv3.bn.bias", "model.24.m16.0.cv3.bn.running_mean", "model.24.m16.0.cv3.bn.running_var", "model.24.m16.0.cv3.bn.num_batches_tracked", "model.24.m16.0.m.0.cv1.conv.weight", "model.24.m16.0.m.0.cv1.bn.weight", "model.24.m16.0.m.0.cv1.bn.bias", "model.24.m16.0.m.0.cv1.bn.running_mean", "model.24.m16.0.m.0.cv1.bn.running_var", "model.24.m16.0.m.0.cv1.bn.num_batches_tracked", "model.24.m16.0.m.0.cv2.conv.weight", "model.24.m16.0.m.0.cv2.bn.weight", "model.24.m16.0.m.0.cv2.bn.bias", "model.24.m16.0.m.0.cv2.bn.running_mean", "model.24.m16.0.m.0.cv2.bn.running_var", "model.24.m16.0.m.0.cv2.bn.num_batches_tracked".

size mismatch for model.24.m.2.cv1.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([32, 128, 1, 1]).

size mismatch for model.24.m.2.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([32, 128, 1, 1]).

size mismatch for model.24.m.2.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv3.conv.weight: copying a param with shape torch.Size([256, 320, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 64, 1, 1]).

size mismatch for model.24.m.2.cv3.bn.weight: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.bias: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.running_mean: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.running_var: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 256, 1, 1]).

size mismatch for model.24.decoder1.1.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 256, 1, 1]).

size mismatch for model.24.decoder1.1.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv3.conv.weight: copying a param with shape torch.Size([256, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([128, 128, 1, 1]).

size mismatch for model.24.decoder1.1.cv3.bn.weight: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.bias: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.running_mean: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.running_var: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.m.0.cv1.conv.weight: copying a param with shape torch.Size([128, 128, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 64, 1, 1]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.conv.weight: copying a param with shape torch.Size([128, 128, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 64, 3, 3]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

相关推荐
PersistJiao14 分钟前
Codex、Claude Code、gstack三者的关系
人工智能
数智工坊32 分钟前
【Mask2Former论文阅读】:基于掩码注意力的通用分割Transformer,大一统全景/实例/语义分割
论文阅读·深度学习·transformer
一切皆是因缘际会39 分钟前
AI数字分身的底层原理:破解意识、自我与人格复刻的核心难题
大数据·人工智能·ai·架构
翔云12345643 分钟前
vLLM全解析:定义、用途与竞品对比
人工智能·ai·大模型
ASKED_20191 小时前
KDD Cup 2026 腾讯算法广告大赛赛题解读: UNI-REC (统一序列建模与特征交叉)
人工智能
fpcc1 小时前
AI和大模型——Fine-tuning
人工智能·深度学习
爱问的艾文1 小时前
八周带你手搓AI应用-Day4-赋予你的AI“记忆力”
人工智能
ACP广源盛139246256732 小时前
IX8024与科学大模型的碰撞@ACP#筑牢科研 AI 算力高速枢纽分享
运维·服务器·网络·数据库·人工智能·嵌入式硬件·电脑
向量引擎2 小时前
向量引擎接入 GPT Image 2 和 deepseek v4:一个 api key 把热门模型串起来,开发者终于不用深夜修接口了
人工智能·gpt·计算机视觉·aigc·api·ai编程·key
努力努力再努力FFF2 小时前
医生对AI辅助诊断感兴趣,作为临床人员该怎么了解和学习?
人工智能·学习