报错处理:RuntimeError: Input type torch.FloatTensor and weight type -torch.cuda.FloatTensor should...

1. 错误名称2. 错误原因3. 修复方法4. mnist数据集测试的案例参考资料

1. 错误名称

return F.conv2d(input, weight, bias, self.stride,

RuntimeError: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same

2. 错误原因

根据stackoverflow的问答,这个错误产生的原因是: You get this error because your model is on the GPU, but your data is on the CPU. So, you need to send your input tensors to the GPU. 意思是输入数据和模型不在一个地方,模型在GPU上,数据在CPU上,应该把数据输入到GPU上面。

3. 修复方法

Stackoverflow给出了几种建议方法:

第一种是添加代码把数据输入到GPU

# You get this error because your model is on the GPU, but your data is on the CPU. So, you need to send your input tensors to the GPU.

inputs, labels = data # this is what you had

inputs, labels = inputs.cuda(), labels.cuda() # add this line

或者这种,把数据tensor和标签修改为模型所输入的位置

x = x.to(device, dtype=torch.float32)

y = y.to(device, dtype=torch.float32)

亲测可用的是下面这个:

model.to(dev)

data = data.to(dev)

如果数据的不会修改,建议model.to(dev)去掉,但是这样的话,测试了似乎是优先选择cpu模式了

4. mnist数据集测试的案例

其实,现在问题症结已经找到了,就是因为用的pytorch的DataLoader,与前面这些不一致,需要通过iter函数遍历数据,所以需要每次都手动把数据喂给GPU

# mnist数据加载

transform = transforms.Compose([transforms.ToTensor(), transforms.Normalize(mean=(0.5,), std=(0.5,))])

train_set = datasets.MNIST(root='./data', train=True, transform=transform, download=True)

test_set = datasets.MNIST(root='./data', train=False, transform=transform, download=True)

batch_size = 32

num_workers = 1

train_loader = data.DataLoader(train_set, batch_size=batch_size, shuffle=True, num_workers=num_workers)

test_loader = data.DataLoader(test_set, batch_size=batch_size, shuffle=False, num_workers=num_workers)

loaders = {'train':train_loader,

'test':test_loader

}

#############################################################################################################

# 模型训练

from torch.autograd import Variable

num_epochs = 10

def train(num_epochs, cnn, loaders):

cnn.train()

# Train the model

total_step = len(loaders['train'])

for epoch in range(num_epochs):

for i, (images, labels) in enumerate(train_loader):

images, labels = images.to(device), labels.to(device)

# gives batch data, normalize x when iterate train_loader

# print(f"images.shape:{images.shape};bx shape:{Variable(images).shape}; by shape:{Variable(labels).shape}; output shape:{model(Variable(images) ).shape}")

b_x = Variable(images) # batch x

b_y = Variable(labels) # batch y

output = model(b_x)

loss = loss_func(output, b_y)

# clear gradients for this training step

optimizer.zero_grad()

# backpropagation, compute gradients

loss.backward()

# apply gradients

optimizer.step()

if (i + 1) % 100 == 0:

print('Epoch [{}/{}], Step [{}/{}], Loss: {:.4f}'

.format(epoch + 1, num_epochs, i + 1, total_step, loss.item()))

pass

pass

pass

用gpu的测试结果

Epoch [1/10], Step [100/1875], Loss: 1.1673

Epoch [1/10], Step [200/1875], Loss: 0.4501

Epoch [1/10], Step [300/1875], Loss: 0.6198

Epoch [1/10], Step [400/1875], Loss: 0.3469

Epoch [1/10], Step [500/1875], Loss: 0.3964

Epoch [1/10], Step [600/1875], Loss: 0.4239

Epoch [1/10], Step [700/1875], Loss: 0.4740

Epoch [1/10], Step [800/1875], Loss: 0.3085

Epoch [1/10], Step [900/1875], Loss: 0.6178

Epoch [1/10], Step [1000/1875], Loss: 0.6015

Epoch [1/10], Step [1100/1875], Loss: 0.1161

Epoch [1/10], Step [1200/1875], Loss: 0.2946

Epoch [1/10], Step [1300/1875], Loss: 0.2689

Epoch [1/10], Step [1400/1875], Loss: 0.3746

Epoch [1/10], Step [1500/1875], Loss: 0.2356

Epoch [1/10], Step [1600/1875], Loss: 0.0904

Epoch [1/10], Step [1700/1875], Loss: 0.2226

参考资料

【1】stackoverflow 问答:RuntimeError: Input type (torch.FloatTensor) and…

好文链接

评论可见,请评论后查看内容,谢谢!!!评论后请刷新页面。