Skip to content

change: move script mode branch to master #234

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 87 commits into from
Sep 19, 2019

Conversation

mvsusp
Copy link
Contributor

@mvsusp mvsusp commented Sep 19, 2019

Description of changes:

  • move script mode branch to master

By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

icywang86rui and others added 30 commits September 27, 2018 10:23
* Scriptmode with cpu docker py2 and py3 docker file

* Migrate to sagemaker-containers 2.1
* Remove serving related packages and code from container
* Add py3 container
* Add integ and unit tests for script mode
* Remove non-asci characters from  README

* Changes based on pr comments

* Move conftest to test root dir

* Add default values for test args

* add docker-compose to test requirement
* Add tox.ini and configure coverage and flake runs

* Add more unit tests
* Configure unit tests to run with both py2 and py3
* Add flake checks
* Fix broken integ tests

* Add import style check

* Add .flake8

* Add source module in coverage command

* Add newlines
* Add mnist sagemaker tests

* Use account-id instead of ecr-image

* Merge gpu and cpu sagemaker tests

* remove _run_mnist_training
* Add Script Mode example
* Add benchmarking script
* edited tf script mode notebook
* Implement distributed support

* Launch parameter server if user set sagemaker_parameter_server_enabled to be True
* Add integ tests
* Add unit tests
* Add distributed sagemaker integ test
* Add 1.11.0 and modify Dockerfile to reduce image size
* Add CI configuration files
* Setting S3 environment variables before training starts

* Remove S3 environment variable setting in test training script

* Add unit tests
* Force framework libraries to re-install
* Update sagemaker containers
* Unset CUDA_VISIBLE_DEVICES for worker processes

* Add comments
The tests all passed not sure why the sagemaker tests are not reporting success.
* Add Keras support
* Create parameter server in different thread
* Fixing some integ tests
This test is only configured to run with 'local'. Change it to use the correct instance type accordingly.
Need this change for container release. Change is only to disable tests
* Add S3 plugin tests

TensorFlow's S3 plugin doesn't work well with S3's eventual consistency model so
we have seen training job failing due to checkpoint or model exporting to S3.
Recently we have released our prod containers with a S3 plugin patch. This
should reduce or eliminate such errors.

The Test added writes a checkpoint to S3 after every training step. It fails
with vanilla TensorFlow.

* Remove distributed_mnist.py

* Fix line too long
This test shouldn't save checkpoints since the two hosts are justing running
training jobs independently. The checkpoints interfere with each other. Changing
the test to use the Keras mnist script here.

This change also changed the saved model path to /opt/ml/opt so we can just use
the estimator.model_data path to assert the model exists.
* Use the test argement framework_version in all tests

* Make flake8 happy
@sagemaker-bot
Copy link
Collaborator

AWS CodeBuild CI Report

  • Result: FAILED
  • Build Logs (available for 30 days)

Powered by github-codebuild-logs, available on the AWS Serverless Application Repository

@mvsusp mvsusp changed the title [WIP] change: move script mode branch to master change: move script mode branch to master Sep 19, 2019
@mvsusp mvsusp requested a review from chuyang-deng September 19, 2019 18:05
@mvsusp mvsusp merged commit 12fd7ef into aws:master Sep 19, 2019
@mvsusp mvsusp deleted the mvs-script-mode-to-master branch September 19, 2019 18:11
# If the training job is part of the multiple training jobs for tuning, we need to append the training job name to
# model_dir in case they read from/write to the same object
if '_tuning_objective_metric' in hyperparameters:
model_dir = _model_dir_with_training_job(hyperparameters.get('model_dir'), env.job_name)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

it seems that hyperparameters does not have 'model_dir' while running my tuning job. Should hyperparameters here be user_hyperparameters instead? @mvsusp

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants