loaderr

Traceback (most recent call last):

File "/ai/zhdata/lyp/multiyolov5_point_608_736/train_608_736.py", line 718, in <module>

train(hyp, opt, device, tb_writer)

File "/ai/zhdata/lyp/multiyolov5_point_608_736/train_608_736.py", line 166, in train

ema.ema.load_state_dict(ckpt['ema'].float().state_dict())

File "/ai/zhdata/lyp/conda/anaconda3/envs/mmd3.0/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1482, in load_state_dict

raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(

RuntimeError: Error(s) in loading state_dict for Model:

Missing key(s) in state_dict: "model.24.m.0.cv1.conv.weight", "model.24.m.0.cv1.bn.weight", "model.24.m.0.cv1.bn.bias", "model.24.m.0.cv1.bn.running_mean", "model.24.m.0.cv1.bn.running_var", "model.24.m.0.cv2.conv.weight", "model.24.m.0.cv2.bn.weight", "model.24.m.0.cv2.bn.bias", "model.24.m.0.cv2.bn.running_mean", "model.24.m.0.cv2.bn.running_var", "model.24.m.0.cv3.conv.weight", "model.24.m.0.cv3.bn.weight", "model.24.m.0.cv3.bn.bias", "model.24.m.0.cv3.bn.running_mean", "model.24.m.0.cv3.bn.running_var", "model.24.m.0.m.0.cv1.conv.weight", "model.24.m.0.m.0.cv1.bn.weight", "model.24.m.0.m.0.cv1.bn.bias", "model.24.m.0.m.0.cv1.bn.running_mean", "model.24.m.0.m.0.cv1.bn.running_var", "model.24.m.0.m.0.cv2.conv.weight", "model.24.m.0.m.0.cv2.bn.weight", "model.24.m.0.m.0.cv2.bn.bias", "model.24.m.0.m.0.cv2.bn.running_mean", "model.24.m.0.m.0.cv2.bn.running_var", "model.24.m.2.m.0.cv1.conv.weight", "model.24.m.2.m.0.cv1.bn.weight", "model.24.m.2.m.0.cv1.bn.bias", "model.24.m.2.m.0.cv1.bn.running_mean", "model.24.m.2.m.0.cv1.bn.running_var", "model.24.m.2.m.0.cv2.conv.weight", "model.24.m.2.m.0.cv2.bn.weight", "model.24.m.2.m.0.cv2.bn.bias", "model.24.m.2.m.0.cv2.bn.running_mean", "model.24.m.2.m.0.cv2.bn.running_var", "model.24.m.4.cv1.conv.weight", "model.24.m.4.cv1.bn.weight", "model.24.m.4.cv1.bn.bias", "model.24.m.4.cv1.bn.running_mean", "model.24.m.4.cv1.bn.running_var", "model.24.m.4.cv2.conv.weight", "model.24.m.4.cv2.bn.weight", "model.24.m.4.cv2.bn.bias", "model.24.m.4.cv2.bn.running_mean", "model.24.m.4.cv2.bn.running_var", "model.24.m.4.cv3.conv.weight", "model.24.m.4.cv3.bn.weight", "model.24.m.4.cv3.bn.bias", "model.24.m.4.cv3.bn.running_mean", "model.24.m.4.cv3.bn.running_var", "model.24.m.4.m.0.cv1.conv.weight", "model.24.m.4.m.0.cv1.bn.weight", "model.24.m.4.m.0.cv1.bn.bias", "model.24.m.4.m.0.cv1.bn.running_mean", "model.24.m.4.m.0.cv1.bn.running_var", "model.24.m.4.m.0.cv2.conv.weight", "model.24.m.4.m.0.cv2.bn.weight", "model.24.m.4.m.0.cv2.bn.bias", "model.24.m.4.m.0.cv2.bn.running_mean", "model.24.m.4.m.0.cv2.bn.running_var", "model.24.m.5.weight", "model.24.m.5.bias", "model.24.decoder1.3.cv1.conv.weight", "model.24.decoder1.3.cv1.bn.weight", "model.24.decoder1.3.cv1.bn.bias", "model.24.decoder1.3.cv1.bn.running_mean", "model.24.decoder1.3.cv1.bn.running_var", "model.24.decoder1.3.cv2.conv.weight", "model.24.decoder1.3.cv2.bn.weight", "model.24.decoder1.3.cv2.bn.bias", "model.24.decoder1.3.cv2.bn.running_mean", "model.24.decoder1.3.cv2.bn.running_var", "model.24.decoder1.3.cv3.conv.weight", "model.24.decoder1.3.cv3.bn.weight", "model.24.decoder1.3.cv3.bn.bias", "model.24.decoder1.3.cv3.bn.running_mean", "model.24.decoder1.3.cv3.bn.running_var", "model.24.decoder1.3.m.0.cv1.conv.weight", "model.24.decoder1.3.m.0.cv1.bn.weight", "model.24.decoder1.3.m.0.cv1.bn.bias", "model.24.decoder1.3.m.0.cv1.bn.running_mean", "model.24.decoder1.3.m.0.cv1.bn.running_var", "model.24.decoder1.3.m.0.cv2.conv.weight", "model.24.decoder1.3.m.0.cv2.bn.weight", "model.24.decoder1.3.m.0.cv2.bn.bias", "model.24.decoder1.3.m.0.cv2.bn.running_mean", "model.24.decoder1.3.m.0.cv2.bn.running_var", "model.24.decoder1.5.cv1.conv.weight", "model.24.decoder1.5.cv1.bn.weight", "model.24.decoder1.5.cv1.bn.bias", "model.24.decoder1.5.cv1.bn.running_mean", "model.24.decoder1.5.cv1.bn.running_var", "model.24.decoder1.5.cv2.conv.weight", "model.24.decoder1.5.cv2.bn.weight", "model.24.decoder1.5.cv2.bn.bias", "model.24.decoder1.5.cv2.bn.running_mean", "model.24.decoder1.5.cv2.bn.running_var", "model.24.decoder1.5.cv3.conv.weight", "model.24.decoder1.5.cv3.bn.weight", "model.24.decoder1.5.cv3.bn.bias", "model.24.decoder1.5.cv3.bn.running_mean", "model.24.decoder1.5.cv3.bn.running_var", "model.24.decoder1.5.m.0.cv1.conv.weight", "model.24.decoder1.5.m.0.cv1.bn.weight", "model.24.decoder1.5.m.0.cv1.bn.bias", "model.24.decoder1.5.m.0.cv1.bn.running_mean", "model.24.decoder1.5.m.0.cv1.bn.running_var", "model.24.decoder1.5.m.0.cv2.conv.weight", "model.24.decoder1.5.m.0.cv2.bn.weight", "model.24.decoder1.5.m.0.cv2.bn.bias", "model.24.decoder1.5.m.0.cv2.bn.running_mean", "model.24.decoder1.5.m.0.cv2.bn.running_var", "model.24.decoder1.6.weight", "model.24.decoder1.6.bias", "model.24.m32.2.cv1.conv.weight", "model.24.m32.2.cv1.bn.weight", "model.24.m32.2.cv1.bn.bias", "model.24.m32.2.cv1.bn.running_mean", "model.24.m32.2.cv1.bn.running_var", "model.24.m32.2.cv2.conv.weight", "model.24.m32.2.cv2.bn.weight", "model.24.m32.2.cv2.bn.bias", "model.24.m32.2.cv2.bn.running_mean", "model.24.m32.2.cv2.bn.running_var", "model.24.m32.2.cv3.conv.weight", "model.24.m32.2.cv3.bn.weight", "model.24.m32.2.cv3.bn.bias", "model.24.m32.2.cv3.bn.running_mean", "model.24.m32.2.cv3.bn.running_var", "model.24.m32.2.m.0.cv1.conv.weight", "model.24.m32.2.m.0.cv1.bn.weight", "model.24.m32.2.m.0.cv1.bn.bias", "model.24.m32.2.m.0.cv1.bn.running_mean", "model.24.m32.2.m.0.cv1.bn.running_var", "model.24.m32.2.m.0.cv2.conv.weight", "model.24.m32.2.m.0.cv2.bn.weight", "model.24.m32.2.m.0.cv2.bn.bias", "model.24.m32.2.m.0.cv2.bn.running_mean", "model.24.m32.2.m.0.cv2.bn.running_var", "model.24.m16.0.conv.weight", "model.24.m16.0.bn.weight", "model.24.m16.0.bn.bias", "model.24.m16.0.bn.running_mean", "model.24.m16.0.bn.running_var".

Unexpected key(s) in state_dict: "model.24.m.0.conv.weight", "model.24.m.0.bn.weight", "model.24.m.0.bn.bias", "model.24.m.0.bn.running_mean", "model.24.m.0.bn.running_var", "model.24.m.0.bn.num_batches_tracked", "model.24.m.1.cv1.conv.weight", "model.24.m.1.cv1.bn.weight", "model.24.m.1.cv1.bn.bias", "model.24.m.1.cv1.bn.running_mean", "model.24.m.1.cv1.bn.running_var", "model.24.m.1.cv1.bn.num_batches_tracked", "model.24.m.1.cv2.conv.weight", "model.24.m.1.cv2.bn.weight", "model.24.m.1.cv2.bn.bias", "model.24.m.1.cv2.bn.running_mean", "model.24.m.1.cv2.bn.running_var", "model.24.m.1.cv2.bn.num_batches_tracked", "model.24.m.1.cv3.conv.weight", "model.24.m.1.cv3.bn.weight", "model.24.m.1.cv3.bn.bias", "model.24.m.1.cv3.bn.running_mean", "model.24.m.1.cv3.bn.running_var", "model.24.m.1.cv3.bn.num_batches_tracked", "model.24.m.1.m.0.cv1.conv.weight", "model.24.m.1.m.0.cv1.bn.weight", "model.24.m.1.m.0.cv1.bn.bias", "model.24.m.1.m.0.cv1.bn.running_mean", "model.24.m.1.m.0.cv1.bn.running_var", "model.24.m.1.m.0.cv1.bn.num_batches_tracked", "model.24.m.1.m.0.cv2.conv.weight", "model.24.m.1.m.0.cv2.bn.weight", "model.24.m.1.m.0.cv2.bn.bias", "model.24.m.1.m.0.cv2.bn.running_mean", "model.24.m.1.m.0.cv2.bn.running_var", "model.24.m.1.m.0.cv2.bn.num_batches_tracked", "model.24.m.2.m.cv1.conv.weight", "model.24.m.2.m.cv1.bn.weight", "model.24.m.2.m.cv1.bn.bias", "model.24.m.2.m.cv1.bn.running_mean", "model.24.m.2.m.cv1.bn.running_var", "model.24.m.2.m.cv1.bn.num_batches_tracked", "model.24.m.2.m.cv2.conv.weight", "model.24.m.2.m.cv2.bn.weight", "model.24.m.2.m.cv2.bn.bias", "model.24.m.2.m.cv2.bn.running_mean", "model.24.m.2.m.cv2.bn.running_var", "model.24.m.2.m.cv2.bn.num_batches_tracked", "model.24.m.3.weight", "model.24.m.3.bias", "model.24.decoder1.2.cv1.conv.weight", "model.24.decoder1.2.cv1.bn.weight", "model.24.decoder1.2.cv1.bn.bias", "model.24.decoder1.2.cv1.bn.running_mean", "model.24.decoder1.2.cv1.bn.running_var", "model.24.decoder1.2.cv1.bn.num_batches_tracked", "model.24.decoder1.2.cv2.conv.weight", "model.24.decoder1.2.cv2.bn.weight", "model.24.decoder1.2.cv2.bn.bias", "model.24.decoder1.2.cv2.bn.running_mean", "model.24.decoder1.2.cv2.bn.running_var", "model.24.decoder1.2.cv2.bn.num_batches_tracked", "model.24.decoder1.2.cv3.conv.weight", "model.24.decoder1.2.cv3.bn.weight", "model.24.decoder1.2.cv3.bn.bias", "model.24.decoder1.2.cv3.bn.running_mean", "model.24.decoder1.2.cv3.bn.running_var", "model.24.decoder1.2.cv3.bn.num_batches_tracked", "model.24.decoder1.2.m.cv1.conv.weight", "model.24.decoder1.2.m.cv1.bn.weight", "model.24.decoder1.2.m.cv1.bn.bias", "model.24.decoder1.2.m.cv1.bn.running_mean", "model.24.decoder1.2.m.cv1.bn.running_var", "model.24.decoder1.2.m.cv1.bn.num_batches_tracked", "model.24.decoder1.2.m.cv2.conv.weight", "model.24.decoder1.2.m.cv2.bn.weight", "model.24.decoder1.2.m.cv2.bn.bias", "model.24.decoder1.2.m.cv2.bn.running_mean", "model.24.decoder1.2.m.cv2.bn.running_var", "model.24.decoder1.2.m.cv2.bn.num_batches_tracked", "model.24.decoder1.3.weight", "model.24.decoder1.3.bias", "model.24.m8.1.cv1.conv.weight", "model.24.m8.1.cv1.bn.weight", "model.24.m8.1.cv1.bn.bias", "model.24.m8.1.cv1.bn.running_mean", "model.24.m8.1.cv1.bn.running_var", "model.24.m8.1.cv1.bn.num_batches_tracked", "model.24.m8.1.cv2.conv.weight", "model.24.m8.1.cv2.bn.weight", "model.24.m8.1.cv2.bn.bias", "model.24.m8.1.cv2.bn.running_mean", "model.24.m8.1.cv2.bn.running_var", "model.24.m8.1.cv2.bn.num_batches_tracked", "model.24.m8.1.cv3.conv.weight", "model.24.m8.1.cv3.bn.weight", "model.24.m8.1.cv3.bn.bias", "model.24.m8.1.cv3.bn.running_mean", "model.24.m8.1.cv3.bn.running_var", "model.24.m8.1.cv3.bn.num_batches_tracked", "model.24.m8.1.m.0.cv1.conv.weight", "model.24.m8.1.m.0.cv1.bn.weight", "model.24.m8.1.m.0.cv1.bn.bias", "model.24.m8.1.m.0.cv1.bn.running_mean", "model.24.m8.1.m.0.cv1.bn.running_var", "model.24.m8.1.m.0.cv1.bn.num_batches_tracked", "model.24.m8.1.m.0.cv2.conv.weight", "model.24.m8.1.m.0.cv2.bn.weight", "model.24.m8.1.m.0.cv2.bn.bias", "model.24.m8.1.m.0.cv2.bn.running_mean", "model.24.m8.1.m.0.cv2.bn.running_var", "model.24.m8.1.m.0.cv2.bn.num_batches_tracked", "model.24.m32.1.cv1.conv.weight", "model.24.m32.1.cv1.bn.weight", "model.24.m32.1.cv1.bn.bias", "model.24.m32.1.cv1.bn.running_mean", "model.24.m32.1.cv1.bn.running_var", "model.24.m32.1.cv1.bn.num_batches_tracked", "model.24.m32.1.cv2.conv.weight", "model.24.m32.1.cv2.bn.weight", "model.24.m32.1.cv2.bn.bias", "model.24.m32.1.cv2.bn.running_mean", "model.24.m32.1.cv2.bn.running_var", "model.24.m32.1.cv2.bn.num_batches_tracked", "model.24.m32.1.cv3.conv.weight", "model.24.m32.1.cv3.bn.weight", "model.24.m32.1.cv3.bn.bias", "model.24.m32.1.cv3.bn.running_mean", "model.24.m32.1.cv3.bn.running_var", "model.24.m32.1.cv3.bn.num_batches_tracked", "model.24.m32.1.m.0.cv1.conv.weight", "model.24.m32.1.m.0.cv1.bn.weight", "model.24.m32.1.m.0.cv1.bn.bias", "model.24.m32.1.m.0.cv1.bn.running_mean", "model.24.m32.1.m.0.cv1.bn.running_var", "model.24.m32.1.m.0.cv1.bn.num_batches_tracked", "model.24.m32.1.m.0.cv2.conv.weight", "model.24.m32.1.m.0.cv2.bn.weight", "model.24.m32.1.m.0.cv2.bn.bias", "model.24.m32.1.m.0.cv2.bn.running_mean", "model.24.m32.1.m.0.cv2.bn.running_var", "model.24.m32.1.m.0.cv2.bn.num_batches_tracked", "model.24.m16.0.cv1.conv.weight", "model.24.m16.0.cv1.bn.weight", "model.24.m16.0.cv1.bn.bias", "model.24.m16.0.cv1.bn.running_mean", "model.24.m16.0.cv1.bn.running_var", "model.24.m16.0.cv1.bn.num_batches_tracked", "model.24.m16.0.cv2.conv.weight", "model.24.m16.0.cv2.bn.weight", "model.24.m16.0.cv2.bn.bias", "model.24.m16.0.cv2.bn.running_mean", "model.24.m16.0.cv2.bn.running_var", "model.24.m16.0.cv2.bn.num_batches_tracked", "model.24.m16.0.cv3.conv.weight", "model.24.m16.0.cv3.bn.weight", "model.24.m16.0.cv3.bn.bias", "model.24.m16.0.cv3.bn.running_mean", "model.24.m16.0.cv3.bn.running_var", "model.24.m16.0.cv3.bn.num_batches_tracked", "model.24.m16.0.m.0.cv1.conv.weight", "model.24.m16.0.m.0.cv1.bn.weight", "model.24.m16.0.m.0.cv1.bn.bias", "model.24.m16.0.m.0.cv1.bn.running_mean", "model.24.m16.0.m.0.cv1.bn.running_var", "model.24.m16.0.m.0.cv1.bn.num_batches_tracked", "model.24.m16.0.m.0.cv2.conv.weight", "model.24.m16.0.m.0.cv2.bn.weight", "model.24.m16.0.m.0.cv2.bn.bias", "model.24.m16.0.m.0.cv2.bn.running_mean", "model.24.m16.0.m.0.cv2.bn.running_var", "model.24.m16.0.m.0.cv2.bn.num_batches_tracked".

size mismatch for model.24.m.2.cv1.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([32, 128, 1, 1]).

size mismatch for model.24.m.2.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([32, 128, 1, 1]).

size mismatch for model.24.m.2.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([32]).

size mismatch for model.24.m.2.cv3.conv.weight: copying a param with shape torch.Size([256, 320, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 64, 1, 1]).

size mismatch for model.24.m.2.cv3.bn.weight: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.bias: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.running_mean: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.m.2.cv3.bn.running_var: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 256, 1, 1]).

size mismatch for model.24.decoder1.1.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.conv.weight: copying a param with shape torch.Size([128, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 256, 1, 1]).

size mismatch for model.24.decoder1.1.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.cv3.conv.weight: copying a param with shape torch.Size([256, 256, 1, 1]) from checkpoint, the shape in current model is torch.Size([128, 128, 1, 1]).

size mismatch for model.24.decoder1.1.cv3.bn.weight: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.bias: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.running_mean: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.cv3.bn.running_var: copying a param with shape torch.Size([256]) from checkpoint, the shape in current model is torch.Size([128]).

size mismatch for model.24.decoder1.1.m.0.cv1.conv.weight: copying a param with shape torch.Size([128, 128, 1, 1]) from checkpoint, the shape in current model is torch.Size([64, 64, 1, 1]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv1.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.conv.weight: copying a param with shape torch.Size([128, 128, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 64, 3, 3]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.weight: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.running_mean: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

size mismatch for model.24.decoder1.1.m.0.cv2.bn.running_var: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([64]).

相关推荐
HIT_Weston7 小时前
45、【Agent】【OpenCode】本地代理分析(请求&接收回调)
人工智能·agent·opencode
逻辑君8 小时前
认知神经科学研究报告【20260010】
人工智能·深度学习·神经网络·机器学习
星河耀银海8 小时前
远控体验分享:安全与实用性参考
人工智能·安全·微服务
企业架构师老王8 小时前
2026企业架构演进:科普Agent(龙虾)如何从“极客玩具”走向实在Agent规模化落地?
人工智能·ai·架构
GreenTea8 小时前
一文搞懂Harness Engineering与Meta-Harness
前端·人工智能·后端
鬼先生_sir8 小时前
Spring AI Alibaba 1.1.2.2 完整知识点库
人工智能·ai·agent·源码解析·springai
深念Y9 小时前
豆包AI能力集成方案:基于会话管理的API网关设计
人工智能
龙文浩_9 小时前
Attention Mechanism: From Theory to Code
人工智能·深度学习·神经网络·学习·自然语言处理
ulimate_9 小时前
八卡算力、三个Baseline算法(WALLOSS、pi0、DreamZero)
人工智能
深小乐9 小时前
AI 周刊【2026.04.06-04.12】:Anthropic 藏起最强模型、AI 社会矛盾激化、"欢乐马"登顶
人工智能