Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Summary: Our tests have been red for a while due to failing memory bechmarks. ## Issue When benchmarking opacus we run the training script multiple times within one process: ``` for i in range(args.num_runs): run_layer_benchmark( ... ) ``` We use built-in pytorch tools to check memory stats. Crucially, we verify that `torch.cuda.memory_allocated()` is 0 before the run starts. Normally, it should be 0, as all previous tensors are out of scope and should have been collected. It all worked fine until something changed and some GPU memory stayed allocated between runs. No idea why, but explicit cache clearing or object deletion didn't help. So I gave up and disabled memory benchmarking, since it seems like it's not a complicated thing to do due to some PyTorch update Pull Request resolved: #589 Reviewed By: JohnlNguyen Differential Revision: D45691684 Pulled By: karthikprasad fbshipit-source-id: 82006e503240532840d3fb6dc0314f2202780973
- Loading branch information