Changes

Jump to: navigation, search

DPS921/PyTorch: Convolutional Neural Networks

1,175 bytes added, 14:12, 30 November 2020
Single-Machine Model
The Single-Machine Model is a parallel model that is done on one machine, with multiple GPUs. It is more convenient than a distributed training model with multiple machines.
 
==== Basic Usage ====
 
Here is a toy model that contains two linear layers. Each linear layer is designed to run a separate GPU.
 
import torch
import torch.nn as nn
import torch.optim as optim
 
 
class ToyModel(nn.Module):
def __init__(self):
super(ToyModel, self).__init__()
self.net1 = torch.nn.Linear(10, 10).to('cuda:0')
self.relu = torch.nn.ReLU()
self.net2 = torch.nn.Linear(10, 5).to('cuda:1')
 
def forward(self, x):
x = self.relu(self.net1(x.to('cuda:0')))
return self.net2(x.to('cuda:1'))
 
The code is very similar to a single GPU implementation, except for the ''.to('cuda:x')'' calls, where ''cuda:0'' and ''cuda:1'' are each their own GPU.
 
model = ToyModel()
loss_fn = nn.MSELoss()
optimizer = optim.SGD(model.parameters(), lr=0.001)
 
optimizer.zero_grad()
outputs = model(torch.randn(20, 10))
labels = torch.randn(20, 5).to('cuda:1')
loss_fn(outputs, labels).backward()
optimizer.step()
 
The backward() and torch.optim will automatically take care of gradients as if the model is on one GPU. You only need to make sure that the labels are on the same device as the outputs when calling the loss function.
== Getting Started With Jupyter ==
56
edits

Navigation menu