2020/02/26 - Amazon SageMaker Service - 5 updated api methods
Changes Update sagemaker client to latest version
{'OutputConfig': {'TargetDevice': {'jetson_xavier', 'amba_cv22', 'imx8qm'}}}
Starts a model compilation job. After the model has been compiled, Amazon SageMaker saves the resulting model artifacts to an Amazon Simple Storage Service (Amazon S3) bucket that you specify.
If you choose to host your model using Amazon SageMaker hosting services, you can use the resulting model artifacts as part of the model. You can also use the artifacts with AWS IoT Greengrass. In that case, deploy them as an ML resource.
In the request body, you provide the following:
A name for the compilation job
Information about the input model artifacts
The output location for the compiled model and the device (target) that the model runs on
The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker assumes to perform the model compilation job
You can also provide a Tag to track the model compilation job's resource use and costs. The response body contains the CompilationJobArn for the compiled job.
To stop a model compilation job, use StopCompilationJob. To get information about a particular model compilation job, use DescribeCompilationJob. To get information about multiple model compilation jobs, use ListCompilationJobs.
See also: AWS API Documentation
Request Syntax
client.create_compilation_job( CompilationJobName='string', RoleArn='string', InputConfig={ 'S3Uri': 'string', 'DataInputConfig': 'string', 'Framework': 'TENSORFLOW'|'KERAS'|'MXNET'|'ONNX'|'PYTORCH'|'XGBOOST' }, OutputConfig={ 'S3OutputLocation': 'string', 'TargetDevice': 'lambda'|'ml_m4'|'ml_m5'|'ml_c4'|'ml_c5'|'ml_p2'|'ml_p3'|'ml_inf1'|'jetson_tx1'|'jetson_tx2'|'jetson_nano'|'jetson_xavier'|'rasp3b'|'imx8qm'|'deeplens'|'rk3399'|'rk3288'|'aisage'|'sbe_c'|'qcs605'|'qcs603'|'amba_cv22' }, StoppingCondition={ 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 } )
string
[REQUIRED]
A name for the model compilation job. The name must be unique within the AWS Region and within your AWS account.
string
[REQUIRED]
The Amazon Resource Name (ARN) of an IAM role that enables Amazon SageMaker to perform tasks on your behalf.
During model compilation, Amazon SageMaker needs your permission to:
Read input data from an S3 bucket
Write model artifacts to an S3 bucket
Write logs to Amazon CloudWatch Logs
Publish metrics to Amazon CloudWatch
You grant permissions for all of these tasks to an IAM role. To pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission. For more information, see Amazon SageMaker Roles.
dict
[REQUIRED]
Provides information about the location of input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.
S3Uri (string) -- [REQUIRED]
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
DataInputConfig (string) -- [REQUIRED]
Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are InputConfig$Framework specific.
TensorFlow : You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input":[1,1024,1024,3]}
If using the CLI, {\"input\":[1,1024,1024,3]}
Examples for two inputs:
If using the console, {"data1": [1,28,28,1], "data2":[1,28,28,1]}
If using the CLI, {\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]}
KERAS : You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, DataInputConfig should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input_1":[1,3,224,224]}
If using the CLI, {\"input_1\":[1,3,224,224]}
Examples for two inputs:
If using the console, {"input_1": [1,3,224,224], "input_2":[1,3,224,224]}
If using the CLI, {\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]}
MXNET/ONNX : You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"data":[1,3,1024,1024]}
If using the CLI, {\"data\":[1,3,1024,1024]}
Examples for two inputs:
If using the console, {"var1": [1,1,28,28], "var2":[1,1,28,28]}
If using the CLI, {\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]}
PyTorch : You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.
Examples for one input in dictionary format:
If using the console, {"input0":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224]}
Example for one input in list format: [[1,3,224,224]]
Examples for two inputs in dictionary format:
If using the console, {"input0":[1,3,224,224], "input1":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]}
Example for two inputs in list format: [[1,3,224,224], [1,3,224,224]]
XGBOOST : input data name and shape are not needed.
Framework (string) -- [REQUIRED]
Identifies the framework in which the model was trained. For example: TENSORFLOW.
dict
[REQUIRED]
Provides information about the output location for the compiled model and the target device the model runs on.
S3OutputLocation (string) -- [REQUIRED]
Identifies the S3 path where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
TargetDevice (string) -- [REQUIRED]
Identifies the device that you want to run your model on after it has been compiled. For example: ml_c5.
dict
[REQUIRED]
Specifies a limit to how long a model compilation job can run. When the job reaches the time limit, Amazon SageMaker ends the compilation job. Use this API to cap model training costs.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that the training or compilation job can run. If job does not complete during this time, Amazon SageMaker ends the job. If value is not specified, default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, how long you are willing to wait for a managed spot training job to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the training job runs. It must be equal to or greater than MaxRuntimeInSeconds .
dict
Response Syntax
{ 'CompilationJobArn': 'string' }
Response Structure
(dict) --
CompilationJobArn (string) --
If the action is successful, the service sends back an HTTP 200 response. Amazon SageMaker returns the following data in JSON format:
CompilationJobArn : The Amazon Resource Name (ARN) of the compiled job.
{'OutputConfig': {'TargetDevice': {'jetson_xavier', 'amba_cv22', 'imx8qm'}}}
Returns information about a model compilation job.
To create a model compilation job, use CreateCompilationJob. To get information about multiple model compilation jobs, use ListCompilationJobs.
See also: AWS API Documentation
Request Syntax
client.describe_compilation_job( CompilationJobName='string' )
string
[REQUIRED]
The name of the model compilation job that you want information about.
dict
Response Syntax
{ 'CompilationJobName': 'string', 'CompilationJobArn': 'string', 'CompilationJobStatus': 'INPROGRESS'|'COMPLETED'|'FAILED'|'STARTING'|'STOPPING'|'STOPPED', 'CompilationStartTime': datetime(2015, 1, 1), 'CompilationEndTime': datetime(2015, 1, 1), 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 }, 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'ModelArtifacts': { 'S3ModelArtifacts': 'string' }, 'RoleArn': 'string', 'InputConfig': { 'S3Uri': 'string', 'DataInputConfig': 'string', 'Framework': 'TENSORFLOW'|'KERAS'|'MXNET'|'ONNX'|'PYTORCH'|'XGBOOST' }, 'OutputConfig': { 'S3OutputLocation': 'string', 'TargetDevice': 'lambda'|'ml_m4'|'ml_m5'|'ml_c4'|'ml_c5'|'ml_p2'|'ml_p3'|'ml_inf1'|'jetson_tx1'|'jetson_tx2'|'jetson_nano'|'jetson_xavier'|'rasp3b'|'imx8qm'|'deeplens'|'rk3399'|'rk3288'|'aisage'|'sbe_c'|'qcs605'|'qcs603'|'amba_cv22' } }
Response Structure
(dict) --
CompilationJobName (string) --
The name of the model compilation job.
CompilationJobArn (string) --
The Amazon Resource Name (ARN) of an IAM role that Amazon SageMaker assumes to perform the model compilation job.
CompilationJobStatus (string) --
The status of the model compilation job.
CompilationStartTime (datetime) --
The time when the model compilation job started the CompilationJob instances.
You are billed for the time between this timestamp and the timestamp in the DescribeCompilationJobResponse$CompilationEndTime field. In Amazon CloudWatch Logs, the start time might be later than this time. That's because it takes time to download the compilation job, which depends on the size of the compilation job container.
CompilationEndTime (datetime) --
The time when the model compilation job on a compilation job instance ended. For a successful or stopped job, this is when the job's model artifacts have finished uploading. For a failed job, this is when Amazon SageMaker detected that the job failed.
StoppingCondition (dict) --
Specifies a limit to how long a model compilation job can run. When the job reaches the time limit, Amazon SageMaker ends the compilation job. Use this API to cap model training costs.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that the training or compilation job can run. If job does not complete during this time, Amazon SageMaker ends the job. If value is not specified, default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, how long you are willing to wait for a managed spot training job to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the training job runs. It must be equal to or greater than MaxRuntimeInSeconds .
CreationTime (datetime) --
The time that the model compilation job was created.
LastModifiedTime (datetime) --
The time that the status of the model compilation job was last modified.
FailureReason (string) --
If a model compilation job failed, the reason it failed.
ModelArtifacts (dict) --
Information about the location in Amazon S3 that has been configured for storing the model artifacts used in the compilation job.
S3ModelArtifacts (string) --
The path of the S3 object that contains the model artifacts. For example, s3://bucket-name/keynameprefix/model.tar.gz .
RoleArn (string) --
The Amazon Resource Name (ARN) of the model compilation job.
InputConfig (dict) --
Information about the location in Amazon S3 of the input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.
S3Uri (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
DataInputConfig (string) --
Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are InputConfig$Framework specific.
TensorFlow : You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input":[1,1024,1024,3]}
If using the CLI, {\"input\":[1,1024,1024,3]}
Examples for two inputs:
If using the console, {"data1": [1,28,28,1], "data2":[1,28,28,1]}
If using the CLI, {\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]}
KERAS : You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, DataInputConfig should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input_1":[1,3,224,224]}
If using the CLI, {\"input_1\":[1,3,224,224]}
Examples for two inputs:
If using the console, {"input_1": [1,3,224,224], "input_2":[1,3,224,224]}
If using the CLI, {\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]}
MXNET/ONNX : You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"data":[1,3,1024,1024]}
If using the CLI, {\"data\":[1,3,1024,1024]}
Examples for two inputs:
If using the console, {"var1": [1,1,28,28], "var2":[1,1,28,28]}
If using the CLI, {\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]}
PyTorch : You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.
Examples for one input in dictionary format:
If using the console, {"input0":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224]}
Example for one input in list format: [[1,3,224,224]]
Examples for two inputs in dictionary format:
If using the console, {"input0":[1,3,224,224], "input1":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]}
Example for two inputs in list format: [[1,3,224,224], [1,3,224,224]]
XGBOOST : input data name and shape are not needed.
Framework (string) --
Identifies the framework in which the model was trained. For example: TENSORFLOW.
OutputConfig (dict) --
Information about the output location for the compiled model and the target device that the model runs on.
S3OutputLocation (string) --
Identifies the S3 path where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
TargetDevice (string) --
Identifies the device that you want to run your model on after it has been compiled. For example: ml_c5.
{'CompilationJobSummaries': {'CompilationTargetDevice': {'amba_cv22', 'imx8qm', 'jetson_xavier'}}}
Lists model compilation jobs that satisfy various filters.
To create a model compilation job, use CreateCompilationJob. To get information about a particular model compilation job you have created, use DescribeCompilationJob.
See also: AWS API Documentation
Request Syntax
client.list_compilation_jobs( NextToken='string', MaxResults=123, CreationTimeAfter=datetime(2015, 1, 1), CreationTimeBefore=datetime(2015, 1, 1), LastModifiedTimeAfter=datetime(2015, 1, 1), LastModifiedTimeBefore=datetime(2015, 1, 1), NameContains='string', StatusEquals='INPROGRESS'|'COMPLETED'|'FAILED'|'STARTING'|'STOPPING'|'STOPPED', SortBy='Name'|'CreationTime'|'Status', SortOrder='Ascending'|'Descending' )
string
If the result of the previous ListCompilationJobs request was truncated, the response includes a NextToken . To retrieve the next set of model compilation jobs, use the token in the next request.
integer
The maximum number of model compilation jobs to return in the response.
datetime
A filter that returns the model compilation jobs that were created after a specified time.
datetime
A filter that returns the model compilation jobs that were created before a specified time.
datetime
A filter that returns the model compilation jobs that were modified after a specified time.
datetime
A filter that returns the model compilation jobs that were modified before a specified time.
string
A filter that returns the model compilation jobs whose name contains a specified string.
string
A filter that retrieves model compilation jobs with a specific DescribeCompilationJobResponse$CompilationJobStatus status.
string
The field by which to sort results. The default is CreationTime .
string
The sort order for results. The default is Ascending .
dict
Response Syntax
{ 'CompilationJobSummaries': [ { 'CompilationJobName': 'string', 'CompilationJobArn': 'string', 'CreationTime': datetime(2015, 1, 1), 'CompilationStartTime': datetime(2015, 1, 1), 'CompilationEndTime': datetime(2015, 1, 1), 'CompilationTargetDevice': 'lambda'|'ml_m4'|'ml_m5'|'ml_c4'|'ml_c5'|'ml_p2'|'ml_p3'|'ml_inf1'|'jetson_tx1'|'jetson_tx2'|'jetson_nano'|'jetson_xavier'|'rasp3b'|'imx8qm'|'deeplens'|'rk3399'|'rk3288'|'aisage'|'sbe_c'|'qcs605'|'qcs603'|'amba_cv22', 'LastModifiedTime': datetime(2015, 1, 1), 'CompilationJobStatus': 'INPROGRESS'|'COMPLETED'|'FAILED'|'STARTING'|'STOPPING'|'STOPPED' }, ], 'NextToken': 'string' }
Response Structure
(dict) --
CompilationJobSummaries (list) --
An array of CompilationJobSummary objects, each describing a model compilation job.
(dict) --
A summary of a model compilation job.
CompilationJobName (string) --
The name of the model compilation job that you want a summary for.
CompilationJobArn (string) --
The Amazon Resource Name (ARN) of the model compilation job.
CreationTime (datetime) --
The time when the model compilation job was created.
CompilationStartTime (datetime) --
The time when the model compilation job started.
CompilationEndTime (datetime) --
The time when the model compilation job completed.
CompilationTargetDevice (string) --
The type of device that the model will run on after compilation has completed.
LastModifiedTime (datetime) --
The time when the model compilation job was last modified.
CompilationJobStatus (string) --
The status of the model compilation job.
NextToken (string) --
If the response is truncated, Amazon SageMaker returns this NextToken . To retrieve the next set of model compilation jobs, use this token in the next request.
{'TrialComponentName': 'string'}
Lists the trials in your account. Specify an experiment name to limit the list to the trials that are part of that experiment. Specify a trial component name to limit the list to the trials that associated with that trial component. The list can be filtered to show only trials that were created in a specific time range. The list can be sorted by trial name or creation time.
See also: AWS API Documentation
Request Syntax
client.list_trials( ExperimentName='string', TrialComponentName='string', CreatedAfter=datetime(2015, 1, 1), CreatedBefore=datetime(2015, 1, 1), SortBy='Name'|'CreationTime', SortOrder='Ascending'|'Descending', MaxResults=123, NextToken='string' )
string
A filter that returns only trials that are part of the specified experiment.
string
A filter that returns only trials that are associated with the specified trial component.
datetime
A filter that returns only trials created after the specified time.
datetime
A filter that returns only trials created before the specified time.
string
The property used to sort results. The default value is CreationTime .
string
The sort order. The default value is Descending .
integer
The maximum number of trials to return in the response. The default value is 10.
string
If the previous call to ListTrials didn't return the full set of trials, the call returns a token for getting the next set of trials.
dict
Response Syntax
{ 'TrialSummaries': [ { 'TrialArn': 'string', 'TrialName': 'string', 'DisplayName': 'string', 'TrialSource': { 'SourceArn': 'string', 'SourceType': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1) }, ], 'NextToken': 'string' }
Response Structure
(dict) --
TrialSummaries (list) --
A list of the summaries of your trials.
(dict) --
A summary of the properties of a trial. To get the complete set of properties, call the DescribeTrial API and provide the TrialName .
TrialArn (string) --
The Amazon Resource Name (ARN) of the trial.
TrialName (string) --
The name of the trial.
DisplayName (string) --
The name of the trial as displayed. If DisplayName isn't specified, TrialName is displayed.
TrialSource (dict) --
The source of the trial.
SourceArn (string) --
The Amazon Resource Name (ARN) of the source.
SourceType (string) --
The source job type.
CreationTime (datetime) --
When the trial was created.
LastModifiedTime (datetime) --
When the trial was last modified.
NextToken (string) --
A token for getting the next set of trials, if there are any.
{'ExcludeRetainedVariantProperties': [{'VariantPropertyType': 'DesiredInstanceCount ' '| DesiredWeight ' '| ' 'DataCaptureConfig'}], 'RetainAllVariantProperties': 'boolean'}
Deploys the new EndpointConfig specified in the request, switches to using newly created endpoint, and then deletes resources provisioned for the endpoint using the previous EndpointConfig (there is no availability loss).
When Amazon SageMaker receives the request, it sets the endpoint status to Updating . After updating the endpoint, it sets the status to InService . To check the status of an endpoint, use the DescribeEndpoint API.
Note
You must not delete an EndpointConfig in use by an endpoint that is live or while the UpdateEndpoint or CreateEndpoint operations are being performed on the endpoint. To update an endpoint, you must create a new EndpointConfig .
See also: AWS API Documentation
Request Syntax
client.update_endpoint( EndpointName='string', EndpointConfigName='string', RetainAllVariantProperties=True|False, ExcludeRetainedVariantProperties=[ { 'VariantPropertyType': 'DesiredInstanceCount'|'DesiredWeight'|'DataCaptureConfig' }, ] )
string
[REQUIRED]
The name of the endpoint whose configuration you want to update.
string
[REQUIRED]
The name of the new endpoint configuration.
boolean
When updating endpoint resources, enables or disables the retention of variant properties, such as the instance count or the variant weight. To retain the variant properties of an endpoint when updating it, set RetainAllVariantProperties to true . To use the variant properties specified in a new EndpointConfig call when updating an endpoint, set RetainAllVariantProperties to false .
list
When you are updating endpoint resources with RetainAllVariantProperties, whose value is set to true , ExcludeRetainedVariantProperties specifies the list of type VariantProperty to override with the values provided by EndpointConfig . If you don't specify a value for ExcludeAllVariantProperties , no variant properties are overridden.
(dict) --
Specifies a production variant property type for an Endpoint.
If you are updating an endpoint with the RetainAllVariantProperties option set to true , the VariantProperty objects listed in ExcludeRetainedVariantProperties override the existing variant properties of the endpoint.
VariantPropertyType (string) -- [REQUIRED]
The type of variant property. The supported values are:
DesiredInstanceCount : Overrides the existing variant instance counts using the InitialInstanceCount values in the ProductionVariants.
DesiredWeight : Overrides the existing variant weights using the InitialVariantWeight values in the ProductionVariants.
DataCaptureConfig : (Not currently supported.)
dict
Response Syntax
{ 'EndpointArn': 'string' }
Response Structure
(dict) --
EndpointArn (string) --
The Amazon Resource Name (ARN) of the endpoint.