Skip to content

Commit 89a3f4d

Browse files
authored
Merge branch 'master' into 1.8.1_PT_containers
2 parents 7ad725b + 85321d3 commit 89a3f4d

File tree

4 files changed

+45
-6
lines changed

4 files changed

+45
-6
lines changed

CHANGELOG.md

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,26 @@
11
# Changelog
22

3+
## v2.34.0 (2021-04-12)
4+
5+
### Features
6+
7+
* Add support for accelerator in Clarify
8+
9+
### Bug Fixes and Other Changes
10+
11+
* add Documentation for how to use
12+
* enable local mode tests that were skipped
13+
* add integ test for HuggingFace with TensorFlow
14+
15+
### Documentation Changes
16+
17+
* release notes for smdistributed.dataparallel v1.1.1
18+
* fixing the SageMaker distributed version references
19+
20+
### Testing and Release Infrastructure
21+
22+
* pin version for ducutils
23+
324
## v2.33.0 (2021-04-05)
425

526
### Features

VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.33.1.dev0
1+
2.34.1.dev0

doc/api/training/sdp_versions/latest.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11

2-
Version 1.1.0 (Latest)
2+
Version 1.1.1 (Latest)
33
======================
44

55
.. toctree::

doc/api/training/smd_data_parallel_release_notes/smd_data_parallel_change_log.md

Lines changed: 22 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,41 @@
1+
# Sagemaker Distributed Data Parallel 1.1.1 Release Notes
2+
3+
* New Features
4+
* Bug Fixes
5+
* Known Issues
6+
7+
*New Features:*
8+
9+
* Adds support for PyTorch 1.8.1
10+
11+
*Bug Fixes:*
12+
13+
* Fixes a bug that was causing gradients from one of the worker nodes to be added twice resulting in incorrect `all_reduce` results under some conditions.
14+
15+
*Known Issues:*
16+
17+
* SageMaker distributed data parallel still is not efficient when run using a single node. For the best performance, use multi-node distributed training with `smdistributed.dataparallel`. Use a single node only for experimental runs while preparing your training pipeline.
18+
119
# Sagemaker Distributed Data Parallel 1.1.0 Release Notes
220

321
* New Features
422
* Bug Fixes
523
* Improvements
624
* Known Issues
725

8-
New Features:
26+
*New Features:*
927

1028
* Adds support for PyTorch 1.8.0 with CUDA 11.1 and CUDNN 8
1129

12-
Bug Fixes:
30+
*Bug Fixes:*
1331

1432
* Fixes crash issue when importing `smdataparallel` before PyTorch
1533

16-
Improvements:
34+
*Improvements:*
1735

1836
* Update `smdataparallel` name in python packages, descriptions, and log outputs
1937

20-
Known Issues:
38+
*Known Issues:*
2139

2240
* SageMaker DataParallel is not efficient when run using a single node. For the best performance, use multi-node distributed training with `smdataparallel`. Use a single node only for experimental runs while preparing your training pipeline.
2341

0 commit comments

Comments
 (0)