You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
parameters and buffers across all `mp_ranks` and moves them to the CPU.
50
52
51
53
#### Add `trace_memory_usage` option to `smp.DistributedModel` to measure memory usage per module
52
54
53
55
Adds `trace_memory_usage` option to `smp.DistributedModel`. This attempts to measure memory usage per module during
54
-
tracing. If this is disabled, memory usage will be estimated through the sizes of tensors returned from the module.
56
+
tracing. If this is disabled, memory usage is estimated through the sizes of tensors returned from the module.
55
57
This option is disabled by default.
56
58
57
59
## Bug Fixes
@@ -60,9 +62,9 @@ This option is disabled by default.
60
62
61
63
-`torch.nn.Sequential`: Fix a bug with `torch.nn.Sequential` which causes a failure with the error message : `shouldnt go less than 0, there is a bug` when the inputs to the first module don't require grads.
62
64
63
-
-`smp.DistributedModel` - Fix a bug with `DistributedModel` execution when a module has multiple parents. The bug surfaces with the error message: `actual_parent should be different than module_execution_stack parent only for torch.nn.ModuleList`
65
+
-`smp.DistributedModel`: Fix a bug with `DistributedModel` execution when a module has multiple parents. The bug surfaces with the error message: `actual_parent should be different than module_execution_stack parent only for torch.nn.ModuleList`
64
66
65
-
-`apex.optimizers.FusedNovoGrad` - Fix a bug with `apex.optimizers.FusedNovoGrad` which surfaces with the error message: `KeyError: 'exp_avg_sq'`
67
+
-`apex.optimizers.FusedNovoGrad`: Fix a bug with `apex.optimizers.FusedNovoGrad` which surfaces with the error message: `KeyError: 'exp_avg_sq'`
0 commit comments