Test pytorch code with minimal computational overhead.
The computational overhead of neural networks discourages thorough testing during development and within CI/CD pipelines.
Torchmocks replaces common building blocks (such as torch.nn.Conv2d) with replicas that only keep track of tensor shapes and device location. This is often the only information that we need to check to ensure proper function of pytorch code.
pip install torchmocks
import torch
import torchmocks
from torchvision.models import resnet152
def test_mock_resnet():
net = resnet152()
torchmocks.mock(net)
image_batch = torch.zeros(4, 3, 255, 255)
output = net(image_batch)
assert output.shape == (4, 1000)
You can exercise most of your training code with torchmocks and the run_fast_dev option for Trainer. See full example here.
def test_training():
dataset = MockDataset()
train_loader = torch.utils.data.DataLoader(dataset, batch_size=3)
val_loader = torch.utils.data.DataLoader(dataset, batch_size=3)
model = ExamplePytorchLightningModule()
mock(model, debug=True)
trainer = pytorch_lightning.Trainer(fast_dev_run=2)
trainer.fit(model, train_loader, val_loader)
This is a work in progress and only a handful of torch modules have been mocked. Modules that have not been mocked will run their normal computation during the forward pass. I'm also exploring other ways to do shape inference in order to mock operations that don't appear in the module tree. Let me know if you have any ideas.