Skip to content

Commit c08bdca

Browse files
author
Talia Chopra
committed
documentation: small typo fixes
1 parent 79c5fa3 commit c08bdca

File tree

2 files changed

+9
-7
lines changed

2 files changed

+9
-7
lines changed

doc/api/training/smd_model_parallel.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@ How to Use this Guide
2424
=====================
2525

2626
The library contains a Common API that is shared across frameworks, as well as APIs
27-
that are specific to supported frameworks, TensorFlow and PyTroch. To use the library, reference the
28-
**Common API** documentation alongside framework specific API documentation.
27+
that are specific to supported frameworks, TensorFlow and PyTorch. To use the library, reference the
28+
**Common API** documentation alongside the framework specific API documentation.
2929

3030
.. toctree::
3131
:maxdepth: 1

doc/api/training/smd_model_parallel_release_notes/smd_model_parallel_change_log.md

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ The following sections describe new feature releases that are common across fram
1212

1313
### Common across frameworks
1414

15-
#### Custom slicing support (smp_slice method) for objects passed to smp.step decorated functions
15+
#### Custom slicing support (`smp_slice` method) for objects passed to `smp.step` decorated functions
1616

1717
To pass an object to `smp.step` that contains tensors that needs to be split across
1818
microbatches and is not an instance of list, dict, tuple or set, you should implement `smp_slice` method for the object.
@@ -46,12 +46,14 @@ def step(custom_obj):
4646

4747
#### Add support for smp.DistributedModel.cpu()
4848

49-
`smp.DistributedModel.cpu()` allgathers parameters and buffers across all `mp_ranks` and moves them to the CPU.
49+
`smp.DistributedModel.cpu()`
50+
[allgather](https://sagemaker.readthedocs.io/en/stable/api/training/smd_model_parallel_common_api.html#smp.allgather)s
51+
parameters and buffers across all `mp_ranks` and moves them to the CPU.
5052

5153
#### Add `trace_memory_usage` option to `smp.DistributedModel` to measure memory usage per module
5254

5355
Adds `trace_memory_usage` option to `smp.DistributedModel`. This attempts to measure memory usage per module during
54-
tracing. If this is disabled, memory usage will be estimated through the sizes of tensors returned from the module.
56+
tracing. If this is disabled, memory usage is estimated through the sizes of tensors returned from the module.
5557
This option is disabled by default.
5658

5759
## Bug Fixes
@@ -60,9 +62,9 @@ This option is disabled by default.
6062

6163
- `torch.nn.Sequential`: Fix a bug with `torch.nn.Sequential` which causes a failure with the error message : `shouldnt go less than 0, there is a bug` when the inputs to the first module don't require grads.
6264

63-
- `smp.DistributedModel` - Fix a bug with `DistributedModel` execution when a module has multiple parents. The bug surfaces with the error message: `actual_parent should be different than module_execution_stack parent only for torch.nn.ModuleList`
65+
- `smp.DistributedModel`: Fix a bug with `DistributedModel` execution when a module has multiple parents. The bug surfaces with the error message: `actual_parent should be different than module_execution_stack parent only for torch.nn.ModuleList`
6466

65-
- `apex.optimizers.FusedNovoGrad` - Fix a bug with `apex.optimizers.FusedNovoGrad` which surfaces with the error message: `KeyError: 'exp_avg_sq'`
67+
- `apex.optimizers.FusedNovoGrad`: Fix a bug with `apex.optimizers.FusedNovoGrad` which surfaces with the error message: `KeyError: 'exp_avg_sq'`
6668

6769
## Improvements
6870

0 commit comments

Comments
 (0)