2021/12/01 - Amazon SageMaker Service - 9 new19 updated api methods
Changes This release enables - 1/ Inference endpoint configuration recommendations and ability to run custom load tests to meet performance needs. 2/ Deploy serverless inference endpoints. 3/ Query, filter and retrieve end-to-end ML lineage graph, and incorporate model quality/bias detection in ML workflow.
Stops an Inference Recommender job.
See also: AWS API Documentation
Request Syntax
client.stop_inference_recommendations_job( JobName='string' )
string
[REQUIRED]
The name of the job you want to stop.
None
A list of lineage groups shared with your Amazon Web Services account. For more information, see Cross-Account Lineage Tracking in the Amazon SageMaker Developer Guide.
See also: AWS API Documentation
Request Syntax
client.list_lineage_groups( CreatedAfter=datetime(2015, 1, 1), CreatedBefore=datetime(2015, 1, 1), SortBy='Name'|'CreationTime', SortOrder='Ascending'|'Descending', NextToken='string', MaxResults=123 )
datetime
A timestamp to filter against lineage groups created after a certain point in time.
datetime
A timestamp to filter against lineage groups created before a certain point in time.
string
The parameter by which to sort the results. The default is CreationTime.
string
The sort order for the results. The default is Ascending.
string
If the response is truncated, SageMaker returns this token. To retrieve the next set of algorithms, use it in the subsequent request.
integer
The maximum number of endpoints to return in the response. This value defaults to 10.
dict
Response Syntax
{ 'LineageGroupSummaries': [ { 'LineageGroupArn': 'string', 'LineageGroupName': 'string', 'DisplayName': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1) }, ], 'NextToken': 'string' }
Response Structure
(dict) --
LineageGroupSummaries (list) --
A list of lineage groups and their properties.
(dict) --
Lists a summary of the properties of a lineage group. A lineage group provides a group of shareable lineage entity resources.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group resource.
LineageGroupName (string) --
The name or Amazon Resource Name (ARN) of the lineage group.
DisplayName (string) --
The display name of the lineage group summary.
CreationTime (datetime) --
The creation time of the lineage group summary.
LastModifiedTime (datetime) --
The last modified time of the lineage group summary.
NextToken (string) --
If the response is truncated, SageMaker returns this token. To retrieve the next set of algorithms, use it in the subsequent request.
Starts a recommendation job. You can create either an instance recommendation or load test job.
See also: AWS API Documentation
Request Syntax
client.create_inference_recommendations_job( JobName='string', JobType='Default'|'Advanced', RoleArn='string', InputConfig={ 'ModelPackageVersionArn': 'string', 'JobDurationInSeconds': 123, 'TrafficPattern': { 'TrafficType': 'PHASES', 'Phases': [ { 'InitialNumberOfUsers': 123, 'SpawnRate': 123, 'DurationInSeconds': 123 }, ] }, 'ResourceLimit': { 'MaxNumberOfTests': 123, 'MaxParallelOfTests': 123 }, 'EndpointConfigurations': [ { 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'InferenceSpecificationName': 'string', 'EnvironmentParameterRanges': { 'CategoricalParameterRanges': [ { 'Name': 'string', 'Value': [ 'string', ] }, ] } }, ] }, JobDescription='string', StoppingConditions={ 'MaxInvocations': 123, 'ModelLatencyThresholds': [ { 'Percentile': 'string', 'ValueInMilliseconds': 123 }, ] }, Tags=[ { 'Key': 'string', 'Value': 'string' }, ] )
string
[REQUIRED]
A name for the recommendation job. The name must be unique within the Amazon Web Services Region and within your Amazon Web Services account.
string
[REQUIRED]
Defines the type of recommendation job. Specify Default to initiate an instance recommendation and Advanced to initiate a load test. If left unspecified, Amazon SageMaker Inference Recommender will run an instance recommendation ( DEFAULT) job.
string
[REQUIRED]
The Amazon Resource Name (ARN) of an IAM role that enables Amazon SageMaker to perform tasks on your behalf.
dict
[REQUIRED]
Provides information about the versioned model package Amazon Resource Name (ARN), the traffic pattern, and endpoint configurations.
ModelPackageVersionArn (string) -- [REQUIRED]
The Amazon Resource Name (ARN) of a versioned model package.
JobDurationInSeconds (integer) --
Specifies the maximum duration of the job, in seconds.>
TrafficPattern (dict) --
Specifies the traffic pattern of the job.
TrafficType (string) --
Defines the traffic patterns.
Phases (list) --
Defines the phases traffic specification.
(dict) --
Defines the traffic pattern.
InitialNumberOfUsers (integer) --
Specifies how many concurrent users to start with.
SpawnRate (integer) --
Specified how many new users to spawn in a minute.
DurationInSeconds (integer) --
Specifies how long traffic phase should be.
ResourceLimit (dict) --
Defines the resource limit of the job.
MaxNumberOfTests (integer) --
Defines the maximum number of load tests.
MaxParallelOfTests (integer) --
Defines the maximum number of parallel load tests.
EndpointConfigurations (list) --
Specifies the endpoint configuration to use for a job.
(dict) --
The endpoint configuration for the load test.
InstanceType (string) -- [REQUIRED]
The instance types to use for the load test.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
EnvironmentParameterRanges (dict) --
The parameter you want to benchmark against.
CategoricalParameterRanges (list) --
Specified a list of parameters for each category.
(dict) --
Environment parameters you want to benchmark your load test against.
Name (string) -- [REQUIRED]
The Name of the environment variable.
Value (list) -- [REQUIRED]
The list of values you can pass.
(string) --
string
Description of the recommendation job.
dict
A set of conditions for stopping a recommendation job. If any of the conditions are met, the job is automatically stopped.
MaxInvocations (integer) --
The maximum number of requests per minute expected for the endpoint.
ModelLatencyThresholds (list) --
The interval of time taken by a model to respond as viewed from SageMaker. The interval includes the local communication time taken to send the request and to fetch the response from the container of a model and the time taken to complete the inference in the container.
(dict) --
The model latency threshold.
Percentile (string) --
The model latency percentile threshold.
ValueInMilliseconds (integer) --
The model latency percentile value in milliseconds.
list
The metadata that you apply to Amazon Web Services resources to help you categorize and organize them. Each tag consists of a key and a value, both of which you define. For more information, see Tagging Amazon Web Services Resources in the Amazon Web Services General Reference.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
dict
Response Syntax
{ 'JobArn': 'string' }
Response Structure
(dict) --
JobArn (string) --
The Amazon Resource Name (ARN) of the recommendation job.
Provides a list of properties for the requested lineage group. For more information, see Cross-Account Lineage Tracking in the Amazon SageMaker Developer Guide.
See also: AWS API Documentation
Request Syntax
client.describe_lineage_group( LineageGroupName='string' )
string
[REQUIRED]
The name of the lineage group.
dict
Response Syntax
{ 'LineageGroupName': 'string', 'LineageGroupArn': 'string', 'DisplayName': 'string', 'Description': 'string', 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' } }
Response Structure
(dict) --
LineageGroupName (string) --
The name of the lineage group.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
DisplayName (string) --
The display name of the lineage group.
Description (string) --
The description of the lineage group.
CreationTime (datetime) --
The creation time of lineage group.
CreatedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LastModifiedTime (datetime) --
The last modified time of the lineage group.
LastModifiedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
Lists recommendation jobs that satisfy various filters.
See also: AWS API Documentation
Request Syntax
client.list_inference_recommendations_jobs( CreationTimeAfter=datetime(2015, 1, 1), CreationTimeBefore=datetime(2015, 1, 1), LastModifiedTimeAfter=datetime(2015, 1, 1), LastModifiedTimeBefore=datetime(2015, 1, 1), NameContains='string', StatusEquals='PENDING'|'IN_PROGRESS'|'COMPLETED'|'FAILED'|'STOPPING'|'STOPPED', SortBy='Name'|'CreationTime'|'Status', SortOrder='Ascending'|'Descending', NextToken='string', MaxResults=123 )
datetime
A filter that returns only jobs created after the specified time (timestamp).
datetime
A filter that returns only jobs created before the specified time (timestamp).
datetime
A filter that returns only jobs that were last modified after the specified time (timestamp).
datetime
A filter that returns only jobs that were last modified before the specified time (timestamp).
string
A string in the job name. This filter returns only recommendations whose name contains the specified string.
string
A filter that retrieves only inference recommendations jobs with a specific status.
string
The parameter by which to sort the results.
string
The sort order for the results.
string
If the response to a previous ListInferenceRecommendationsJobsRequest request was truncated, the response includes a NextToken. To retrieve the next set of recommendations, use the token in the next request.
integer
The maximum number of recommendations to return in the response.
dict
Response Syntax
{ 'InferenceRecommendationsJobs': [ { 'JobName': 'string', 'JobDescription': 'string', 'JobType': 'Default'|'Advanced', 'JobArn': 'string', 'Status': 'PENDING'|'IN_PROGRESS'|'COMPLETED'|'FAILED'|'STOPPING'|'STOPPED', 'CreationTime': datetime(2015, 1, 1), 'CompletionTime': datetime(2015, 1, 1), 'RoleArn': 'string', 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string' }, ], 'NextToken': 'string' }
Response Structure
(dict) --
InferenceRecommendationsJobs (list) --
The recommendations created from the Amazon SageMaker Inference Recommender job.
(dict) --
A structure that contains a list of recommendation jobs.
JobName (string) --
The name of the job.
JobDescription (string) --
The job description.
JobType (string) --
The recommendation job type.
JobArn (string) --
The Amazon Resource Name (ARN) of the recommendation job.
Status (string) --
The status of the job.
CreationTime (datetime) --
A timestamp that shows when the job was created.
CompletionTime (datetime) --
A timestamp that shows when the job completed.
RoleArn (string) --
The Amazon Resource Name (ARN) of an IAM role that enables Amazon SageMaker to perform tasks on your behalf.
LastModifiedTime (datetime) --
A timestamp that shows when the job was last modified.
FailureReason (string) --
If the job fails, provides information why the job failed.
NextToken (string) --
A token for getting the next set of recommendations, if there are any.
The resource policy for the lineage group.
See also: AWS API Documentation
Request Syntax
client.get_lineage_group_policy( LineageGroupName='string' )
string
[REQUIRED]
The name or Amazon Resource Name (ARN) of the lineage group.
dict
Response Syntax
{ 'LineageGroupArn': 'string', 'ResourcePolicy': 'string' }
Response Structure
(dict) --
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
ResourcePolicy (string) --
The resource policy that gives access to the lineage group in another account.
Lists the domain, framework, task, and model name of standard machine learning models found in common model zoos.
See also: AWS API Documentation
Request Syntax
client.list_model_metadata( SearchExpression={ 'Filters': [ { 'Name': 'Domain'|'Framework'|'Task'|'FrameworkVersion', 'Value': 'string' }, ] }, NextToken='string', MaxResults=123 )
dict
One or more filters that searches for the specified resource or resources in a search. All resource objects that satisfy the expression's condition are included in the search results. Specify the Framework, FrameworkVersion, Domain or Task to filter supported. Filter names and values are case-sensitive.
Filters (list) --
A list of filter objects.
(dict) --
Part of the search expression. You can specify the name and value (domain, task, framework, framework version, task, and model).
Name (string) -- [REQUIRED]
The name of the of the model to filter by.
Value (string) -- [REQUIRED]
The value to filter the model metadata.
string
If the response to a previous ListModelMetadataResponse request was truncated, the response includes a NextToken. To retrieve the next set of model metadata, use the token in the next request.
integer
The maximum number of models to return in the response.
dict
Response Syntax
{ 'ModelMetadataSummaries': [ { 'Domain': 'string', 'Framework': 'string', 'Task': 'string', 'Model': 'string', 'FrameworkVersion': 'string' }, ], 'NextToken': 'string' }
Response Structure
(dict) --
ModelMetadataSummaries (list) --
A structure that holds model metadata.
(dict) --
A summary of the model metadata.
Domain (string) --
The machine learning domain of the model.
Framework (string) --
The machine learning framework of the model.
Task (string) --
The machine learning task of the model.
Model (string) --
The name of the model.
FrameworkVersion (string) --
The framework version of the model.
NextToken (string) --
A token for getting the next set of recommendations, if there are any.
Provides the results of the Inference Recommender job. One or more recommendation jobs are returned.
See also: AWS API Documentation
Request Syntax
client.describe_inference_recommendations_job( JobName='string' )
string
[REQUIRED]
The name of the job. The name must be unique within an Amazon Web Services Region in the Amazon Web Services account.
dict
Response Syntax
{ 'JobName': 'string', 'JobDescription': 'string', 'JobType': 'Default'|'Advanced', 'JobArn': 'string', 'RoleArn': 'string', 'Status': 'PENDING'|'IN_PROGRESS'|'COMPLETED'|'FAILED'|'STOPPING'|'STOPPED', 'CreationTime': datetime(2015, 1, 1), 'CompletionTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'InputConfig': { 'ModelPackageVersionArn': 'string', 'JobDurationInSeconds': 123, 'TrafficPattern': { 'TrafficType': 'PHASES', 'Phases': [ { 'InitialNumberOfUsers': 123, 'SpawnRate': 123, 'DurationInSeconds': 123 }, ] }, 'ResourceLimit': { 'MaxNumberOfTests': 123, 'MaxParallelOfTests': 123 }, 'EndpointConfigurations': [ { 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'InferenceSpecificationName': 'string', 'EnvironmentParameterRanges': { 'CategoricalParameterRanges': [ { 'Name': 'string', 'Value': [ 'string', ] }, ] } }, ] }, 'StoppingConditions': { 'MaxInvocations': 123, 'ModelLatencyThresholds': [ { 'Percentile': 'string', 'ValueInMilliseconds': 123 }, ] }, 'InferenceRecommendations': [ { 'Metrics': { 'CostPerHour': ..., 'CostPerInference': ..., 'MaxInvocations': 123, 'ModelLatency': 123 }, 'EndpointConfiguration': { 'EndpointName': 'string', 'VariantName': 'string', 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'InitialInstanceCount': 123 }, 'ModelConfiguration': { 'InferenceSpecificationName': 'string', 'EnvironmentParameters': [ { 'Key': 'string', 'ValueType': 'string', 'Value': 'string' }, ] } }, ] }
Response Structure
(dict) --
JobName (string) --
The name of the job. The name must be unique within an Amazon Web Services Region in the Amazon Web Services account.
JobDescription (string) --
The job description that you provided when you initiated the job.
JobType (string) --
The job type that you provided when you initiated the job.
JobArn (string) --
The Amazon Resource Name (ARN) of the job.
RoleArn (string) --
The Amazon Resource Name (ARN) of the Amazon Web Services Identity and Access Management (IAM) role you provided when you initiated the job.
Status (string) --
The status of the job.
CreationTime (datetime) --
A timestamp that shows when the job was created.
CompletionTime (datetime) --
A timestamp that shows when the job completed.
LastModifiedTime (datetime) --
A timestamp that shows when the job was last modified.
FailureReason (string) --
If the job fails, provides information why the job failed.
InputConfig (dict) --
Returns information about the versioned model package Amazon Resource Name (ARN), the traffic pattern, and endpoint configurations you provided when you initiated the job.
ModelPackageVersionArn (string) --
The Amazon Resource Name (ARN) of a versioned model package.
JobDurationInSeconds (integer) --
Specifies the maximum duration of the job, in seconds.>
TrafficPattern (dict) --
Specifies the traffic pattern of the job.
TrafficType (string) --
Defines the traffic patterns.
Phases (list) --
Defines the phases traffic specification.
(dict) --
Defines the traffic pattern.
InitialNumberOfUsers (integer) --
Specifies how many concurrent users to start with.
SpawnRate (integer) --
Specified how many new users to spawn in a minute.
DurationInSeconds (integer) --
Specifies how long traffic phase should be.
ResourceLimit (dict) --
Defines the resource limit of the job.
MaxNumberOfTests (integer) --
Defines the maximum number of load tests.
MaxParallelOfTests (integer) --
Defines the maximum number of parallel load tests.
EndpointConfigurations (list) --
Specifies the endpoint configuration to use for a job.
(dict) --
The endpoint configuration for the load test.
InstanceType (string) --
The instance types to use for the load test.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
EnvironmentParameterRanges (dict) --
The parameter you want to benchmark against.
CategoricalParameterRanges (list) --
Specified a list of parameters for each category.
(dict) --
Environment parameters you want to benchmark your load test against.
Name (string) --
The Name of the environment variable.
Value (list) --
The list of values you can pass.
(string) --
StoppingConditions (dict) --
The stopping conditions that you provided when you initiated the job.
MaxInvocations (integer) --
The maximum number of requests per minute expected for the endpoint.
ModelLatencyThresholds (list) --
The interval of time taken by a model to respond as viewed from SageMaker. The interval includes the local communication time taken to send the request and to fetch the response from the container of a model and the time taken to complete the inference in the container.
(dict) --
The model latency threshold.
Percentile (string) --
The model latency percentile threshold.
ValueInMilliseconds (integer) --
The model latency percentile value in milliseconds.
InferenceRecommendations (list) --
The recommendations made by Inference Recommender.
(dict) --
A list of recommendations made by Amazon SageMaker Inference Recommender.
Metrics (dict) --
The metrics used to decide what recommendation to make.
CostPerHour (float) --
Defines the cost per hour for the instance.
CostPerInference (float) --
Defines the cost per inference for the instance .
MaxInvocations (integer) --
The expected maximum number of requests per minute for the instance.
ModelLatency (integer) --
The expected model latency at maximum invocation per minute for the instance.
EndpointConfiguration (dict) --
Defines the endpoint configuration parameters.
EndpointName (string) --
The name of the endpoint made during a recommendation job.
VariantName (string) --
The name of the production variant (deployed model) made during a recommendation job.
InstanceType (string) --
The instance type recommended by Amazon SageMaker Inference Recommender.
InitialInstanceCount (integer) --
The number of instances recommended to launch initially.
ModelConfiguration (dict) --
Defines the model configuration.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
EnvironmentParameters (list) --
Defines the environment parameters that includes key, value types, and values.
(dict) --
A list of environment parameters suggested by the Amazon SageMaker Inference Recommender.
Key (string) --
The environment key suggested by the Amazon SageMaker Inference Recommender.
ValueType (string) --
The value type suggested by the Amazon SageMaker Inference Recommender.
Value (string) --
The value suggested by the Amazon SageMaker Inference Recommender.
Use this action to inspect your lineage and discover relationships between entities. For more information, see Querying Lineage Entities in the Amazon SageMaker Developer Guide.
See also: AWS API Documentation
Request Syntax
client.query_lineage( StartArns=[ 'string', ], Direction='Both'|'Ascendants'|'Descendants', IncludeEdges=True|False, Filters={ 'Types': [ 'string', ], 'LineageTypes': [ 'TrialComponent'|'Artifact'|'Context'|'Action', ], 'CreatedBefore': datetime(2015, 1, 1), 'CreatedAfter': datetime(2015, 1, 1), 'ModifiedBefore': datetime(2015, 1, 1), 'ModifiedAfter': datetime(2015, 1, 1), 'Properties': { 'string': 'string' } }, MaxDepth=123, MaxResults=123, NextToken='string' )
list
[REQUIRED]
A list of resource Amazon Resource Name (ARN) that represent the starting point for your lineage query.
(string) --
string
Associations between lineage entities are directed. This parameter determines the direction from the StartArn(s) the query will look.
boolean
Setting this value to True will retrieve not only the entities of interest but also the Associations and lineage entities on the path. Set to False to only return lineage entities that match your query.
dict
A set of filtering parameters that allow you to specify which entities should be returned.
Properties - Key-value pairs to match on the lineage entities' properties.
LineageTypes - A set of lineage entity types to match on. For example: TrialComponent, Artifact, or Context.
CreatedBefore - Filter entities created before this date.
ModifiedBefore - Filter entities modified before this date.
ModifiedAfter - Filter entities modified after this date.
Types (list) --
Filter the lineage entities connected to the StartArn by type. For example: DataSet, Model, Endpoint, or ModelDeployment.
(string) --
LineageTypes (list) --
Filter the lineage entities connected to the ``StartArn``(s) by the type of the lineage entity.
(string) --
CreatedBefore (datetime) --
Filter the lineage entities connected to the ``StartArn``(s) by created date.
CreatedAfter (datetime) --
Filter the lineage entities connected to the ``StartArn``(s) after the create date.
ModifiedBefore (datetime) --
Filter the lineage entities connected to the ``StartArn``(s) before the last modified date.
ModifiedAfter (datetime) --
Filter the lineage entities connected to the ``StartArn``(s) after the last modified date.
Properties (dict) --
Filter the lineage entities connected to the ``StartArn``(s) by a set if property key value pairs. If multiple pairs are provided, an entity will be included in the results if it matches any of the provided pairs.
(string) --
(string) --
integer
The maximum depth in lineage relationships from the StartArns that will be traversed. Depth is a measure of the number of Associations from the StartArn entity to the matched results.
integer
Limits the number of vertices in the results. Use the NextToken in a response to to retrieve the next page of results.
string
Limits the number of vertices in the request. Use the NextToken in a response to to retrieve the next page of results.
dict
Response Syntax
{ 'Vertices': [ { 'Arn': 'string', 'Type': 'string', 'LineageType': 'TrialComponent'|'Artifact'|'Context'|'Action' }, ], 'Edges': [ { 'SourceArn': 'string', 'DestinationArn': 'string', 'AssociationType': 'ContributedTo'|'AssociatedWith'|'DerivedFrom'|'Produced' }, ], 'NextToken': 'string' }
Response Structure
(dict) --
Vertices (list) --
A list of vertices connected to the start entity(ies) in the lineage graph.
(dict) --
A lineage entity connected to the starting entity(ies).
Arn (string) --
The Amazon Resource Name (ARN) of the lineage entity resource.
Type (string) --
The type of the lineage entity resource. For example: DataSet, Model, Endpoint, etc...
LineageType (string) --
The type of resource of the lineage entity.
Edges (list) --
A list of edges that connect vertices in the response.
(dict) --
A directed edge connecting two lineage entities.
SourceArn (string) --
The Amazon Resource Name (ARN) of the source lineage entity of the directed edge.
DestinationArn (string) --
The Amazon Resource Name (ARN) of the destination lineage entity of the directed edge.
AssociationType (string) --
The type of the Association(Edge) between the source and destination. For example ContributedTo, Produced, or DerivedFrom.
NextToken (string) --
Limits the number of vertices in the response. Use the NextToken in a response to to retrieve the next page of results.
{'ModelPackageSummaries': {'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}}}
This action batch describes a list of versioned model packages
See also: AWS API Documentation
Request Syntax
client.batch_describe_model_package( ModelPackageArnList=[ 'string', ] )
list
[REQUIRED]
The list of Amazon Resource Name (ARN) of the model package groups.
(string) --
dict
Response Syntax
{ 'ModelPackageSummaries': { 'string': { 'ModelPackageGroupName': 'string', 'ModelPackageVersion': 123, 'ModelPackageArn': 'string', 'ModelPackageDescription': 'string', 'CreationTime': datetime(2015, 1, 1), 'InferenceSpecification': { 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, 'ModelPackageStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting', 'ModelApprovalStatus': 'Approved'|'Rejected'|'PendingManualApproval' } }, 'BatchDescribeModelPackageErrorMap': { 'string': { 'ErrorCode': 'string', 'ErrorResponse': 'string' } } }
Response Structure
(dict) --
ModelPackageSummaries (dict) --
The summaries for the model package versions
(string) --
(dict) --
Provides summary information about the model package.
ModelPackageGroupName (string) --
The group name for the model package
ModelPackageVersion (integer) --
The version number of a versioned model.
ModelPackageArn (string) --
The Amazon Resource Name (ARN) of the model package.
ModelPackageDescription (string) --
The description of the model package.
CreationTime (datetime) --
The creation time of the mortgage package summary.
InferenceSpecification (dict) --
Defines how to perform inference generation after a training job is run.
Containers (list) --
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) --
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) --
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
ModelPackageStatus (string) --
The status of the mortgage package.
ModelApprovalStatus (string) --
The approval status of the model.
BatchDescribeModelPackageErrorMap (dict) --
A map of the resource and BatchDescribeModelPackageError objects reporting the error associated with describing the model package.
(string) --
(dict) --
The error code and error description associated with the resource.
ErrorCode (string) --
ErrorResponse (string) --
{'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}}
Create a machine learning algorithm that you can use in Amazon SageMaker and list in the Amazon Web Services Marketplace.
See also: AWS API Documentation
Request Syntax
client.create_algorithm( AlgorithmName='string', AlgorithmDescription='string', TrainingSpecification={ 'TrainingImage': 'string', 'TrainingImageDigest': 'string', 'SupportedHyperParameters': [ { 'Name': 'string', 'Description': 'string', 'Type': 'Integer'|'Continuous'|'Categorical'|'FreeText', 'Range': { 'IntegerParameterRangeSpecification': { 'MinValue': 'string', 'MaxValue': 'string' }, 'ContinuousParameterRangeSpecification': { 'MinValue': 'string', 'MaxValue': 'string' }, 'CategoricalParameterRangeSpecification': { 'Values': [ 'string', ] } }, 'IsTunable': True|False, 'IsRequired': True|False, 'DefaultValue': 'string' }, ], 'SupportedTrainingInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', ], 'SupportsDistributedTraining': True|False, 'MetricDefinitions': [ { 'Name': 'string', 'Regex': 'string' }, ], 'TrainingChannels': [ { 'Name': 'string', 'Description': 'string', 'IsRequired': True|False, 'SupportedContentTypes': [ 'string', ], 'SupportedCompressionTypes': [ 'None'|'Gzip', ], 'SupportedInputModes': [ 'Pipe'|'File'|'FastFile', ] }, ], 'SupportedTuningJobObjectiveMetrics': [ { 'Type': 'Maximize'|'Minimize', 'MetricName': 'string' }, ] }, InferenceSpecification={ 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ValidationSpecification={ 'ValidationRole': 'string', 'ValidationProfiles': [ { 'ProfileName': 'string', 'TrainingJobDefinition': { 'TrainingInputMode': 'Pipe'|'File'|'FastFile', 'HyperParameters': { 'string': 'string' }, 'InputDataConfig': [ { 'ChannelName': 'string', 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'AttributeNames': [ 'string', ] }, 'FileSystemDataSource': { 'FileSystemId': 'string', 'FileSystemAccessMode': 'rw'|'ro', 'FileSystemType': 'EFS'|'FSxLustre', 'DirectoryPath': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'RecordWrapperType': 'None'|'RecordIO', 'InputMode': 'Pipe'|'File'|'FastFile', 'ShuffleConfig': { 'Seed': 123 } }, ], 'OutputDataConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, 'ResourceConfig': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', 'InstanceCount': 123, 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 } }, 'TransformJobDefinition': { 'MaxConcurrentTransforms': 123, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' } } }, ] }, CertifyForMarketplace=True|False, Tags=[ { 'Key': 'string', 'Value': 'string' }, ] )
string
[REQUIRED]
The name of the algorithm.
string
A description of the algorithm.
dict
[REQUIRED]
Specifies details about training jobs run by this algorithm, including the following:
The Amazon ECR path of the container and the version digest of the algorithm.
The hyperparameters that the algorithm supports.
The instance types that the algorithm supports for training.
Whether the algorithm supports distributed training.
The metrics that the algorithm emits to Amazon CloudWatch.
Which metrics that the algorithm emits can be used as the objective metric for hyperparameter tuning jobs.
The input channels that the algorithm supports for training data. For example, an algorithm might support train, validation, and test channels.
TrainingImage (string) -- [REQUIRED]
The Amazon ECR registry path of the Docker image that contains the training algorithm.
TrainingImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
SupportedHyperParameters (list) --
A list of the HyperParameterSpecification objects, that define the supported hyperparameters. This is required if the algorithm supports automatic model tuning.>
(dict) --
Defines a hyperparameter to be used by an algorithm.
Name (string) -- [REQUIRED]
The name of this hyperparameter. The name must be unique.
Description (string) --
A brief description of the hyperparameter.
Type (string) -- [REQUIRED]
The type of this hyperparameter. The valid types are Integer, Continuous, Categorical, and FreeText.
Range (dict) --
The allowed range for this hyperparameter.
IntegerParameterRangeSpecification (dict) --
A IntegerParameterRangeSpecification object that defines the possible values for an integer hyperparameter.
MinValue (string) -- [REQUIRED]
The minimum integer value allowed.
MaxValue (string) -- [REQUIRED]
The maximum integer value allowed.
ContinuousParameterRangeSpecification (dict) --
A ContinuousParameterRangeSpecification object that defines the possible values for a continuous hyperparameter.
MinValue (string) -- [REQUIRED]
The minimum floating-point value allowed.
MaxValue (string) -- [REQUIRED]
The maximum floating-point value allowed.
CategoricalParameterRangeSpecification (dict) --
A CategoricalParameterRangeSpecification object that defines the possible values for a categorical hyperparameter.
Values (list) -- [REQUIRED]
The allowed categories for the hyperparameter.
(string) --
IsTunable (boolean) --
Indicates whether this hyperparameter is tunable in a hyperparameter tuning job.
IsRequired (boolean) --
Indicates whether this hyperparameter is required.
DefaultValue (string) --
The default value for this hyperparameter. If a default value is specified, a hyperparameter cannot be required.
SupportedTrainingInstanceTypes (list) -- [REQUIRED]
A list of the instance types that this algorithm can use for training.
(string) --
SupportsDistributedTraining (boolean) --
Indicates whether the algorithm supports distributed training. If set to false, buyers can't request more than one instance during training.
MetricDefinitions (list) --
A list of MetricDefinition objects, which are used for parsing metrics generated by the algorithm.
(dict) --
Specifies a metric that the training algorithm writes to stderr or stdout. Amazon SageMakerhyperparameter tuning captures all defined metrics. You specify one metric that a hyperparameter tuning job uses as its objective metric to choose the best training job.
Name (string) -- [REQUIRED]
The name of the metric.
Regex (string) -- [REQUIRED]
A regular expression that searches the output of a training job and gets the value of the metric. For more information about using regular expressions to define metrics, see Defining Objective Metrics.
TrainingChannels (list) -- [REQUIRED]
A list of ChannelSpecification objects, which specify the input sources to be used by the algorithm.
(dict) --
Defines a named input source, called a channel, to be used by an algorithm.
Name (string) -- [REQUIRED]
The name of the channel.
Description (string) --
A brief description of the channel.
IsRequired (boolean) --
Indicates whether the channel is required by the algorithm.
SupportedContentTypes (list) -- [REQUIRED]
The supported MIME types for the data.
(string) --
SupportedCompressionTypes (list) --
The allowed compression types, if data compression is used.
(string) --
SupportedInputModes (list) -- [REQUIRED]
The allowed input mode, either FILE or PIPE.
In FILE mode, Amazon SageMaker copies the data from the input source onto the local Amazon Elastic Block Store (Amazon EBS) volumes before starting your training algorithm. This is the most commonly used input mode.
In PIPE mode, Amazon SageMaker streams input data from the source directly to your algorithm without using the EBS volume.
(string) --
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
SupportedTuningJobObjectiveMetrics (list) --
A list of the metrics that the algorithm emits that can be used as the objective metric in a hyperparameter tuning job.
(dict) --
Defines the objective metric for a hyperparameter tuning job. Hyperparameter tuning uses the value of this metric to evaluate the training jobs it launches, and returns the training job that results in either the highest or lowest value for this metric, depending on the value you specify for the Type parameter.
Type (string) -- [REQUIRED]
Whether to minimize or maximize the objective metric.
MetricName (string) -- [REQUIRED]
The name of the metric to use for the objective metric.
dict
Specifies details about inference jobs that the algorithm runs, including the following:
The Amazon ECR paths of containers that contain the inference code and model artifacts.
The instance types that the algorithm supports for transform jobs and real-time endpoints used for inference.
The input and output content formats that the algorithm supports for inference.
Containers (list) -- [REQUIRED]
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) -- [REQUIRED]
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) -- [REQUIRED]
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedContentTypes (list) -- [REQUIRED]
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) -- [REQUIRED]
The supported MIME types for the output data.
(string) --
dict
Specifies configurations for one or more training jobs and that Amazon SageMaker runs to test the algorithm's training code and, optionally, one or more batch transform jobs that Amazon SageMaker runs to test the algorithm's inference code.
ValidationRole (string) -- [REQUIRED]
The IAM roles that Amazon SageMaker uses to run the training jobs.
ValidationProfiles (list) -- [REQUIRED]
An array of AlgorithmValidationProfile objects, each of which specifies a training job and batch transform job that Amazon SageMaker runs to validate your algorithm.
(dict) --
Defines a training job and a batch transform job that Amazon SageMaker runs to validate your algorithm.
The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.
ProfileName (string) -- [REQUIRED]
The name of the profile for the algorithm. The name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).
TrainingJobDefinition (dict) -- [REQUIRED]
The TrainingJobDefinition object that describes the training job that Amazon SageMaker runs to validate your algorithm.
TrainingInputMode (string) -- [REQUIRED]
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
HyperParameters (dict) --
The hyperparameters used for the training job.
(string) --
(string) --
InputDataConfig (list) -- [REQUIRED]
An array of Channel objects, each of which specifies an input source.
(dict) --
A channel is a named input source that training algorithms can consume.
ChannelName (string) -- [REQUIRED]
The name of the channel.
DataSource (dict) -- [REQUIRED]
The location of the channel data.
S3DataSource (dict) --
The S3 location of the data source that is associated with a channel.
S3DataType (string) -- [REQUIRED]
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects that match the specified key name prefix for model training.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for model training.
If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile can only be used if the Channel's input mode is Pipe.
S3Uri (string) -- [REQUIRED]
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix
A manifest might look like this: s3://bucketname/example.manifest A manifest is an S3 object which is a JSON file consisting of an array of elements. The first element is a prefix which is followed by one or more suffixes. SageMaker appends the suffix elements to the prefix to get a full set of S3Uri. Note that the prefix must be a valid non-empty S3Uri that precludes users from specifying a manifest whose individual S3Uri is sourced from different S3 buckets. The following code example shows a valid manifest format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] This JSON is equivalent to the following S3Uri list: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uri in this manifest is the input data for the channel for this data source. The object that each S3Uri points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
S3DataDistributionType (string) --
If you want Amazon SageMaker to replicate the entire dataset on each ML compute instance that is launched for model training, specify FullyReplicated.
If you want Amazon SageMaker to replicate a subset of data on each ML compute instance that is launched for model training, specify ShardedByS3Key. If there are n ML compute instances launched for a training job, each instance gets approximately 1/n of the number of S3 objects. In this case, model training on each machine uses only the subset of training data.
Don't choose more ML compute instances for training than available S3 objects. If you do, some nodes won't get any data and you will pay for nodes that aren't getting any training data. This applies in both File and Pipe modes. Keep this in mind when developing algorithms.
In distributed training, where you use multiple ML compute EC2 instances, you might choose ShardedByS3Key. If the algorithm requires copying training data to the ML storage volume (when TrainingInputMode is set to File), this copies 1/n of the number of objects.
AttributeNames (list) --
A list of one or more attribute names to use that are found in a specified augmented manifest file.
(string) --
FileSystemDataSource (dict) --
The file system that is associated with a channel.
FileSystemId (string) -- [REQUIRED]
The file system id.
FileSystemAccessMode (string) -- [REQUIRED]
The access mode of the mount of the directory associated with the channel. A directory can be mounted either in ro (read-only) or rw (read-write) mode.
FileSystemType (string) -- [REQUIRED]
The file system type.
DirectoryPath (string) -- [REQUIRED]
The full path to the directory to associate with the channel.
ContentType (string) --
The MIME type of the data.
CompressionType (string) --
If training data is compressed, the compression type. The default value is None. CompressionType is used only in Pipe input mode. In File mode, leave this field unset or set it to None.
RecordWrapperType (string) --
Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format. In this case, Amazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.
In File mode, leave this field unset or set it to None.
InputMode (string) --
(Optional) The input mode to use for the data channel in a training job. If you don't set a value for InputMode, Amazon SageMaker uses the value set for TrainingInputMode. Use this parameter to override the TrainingInputMode setting in a AlgorithmSpecification request when you have a channel that needs a different input mode from the training job's general setting. To download the data from Amazon Simple Storage Service (Amazon S3) to the provisioned ML storage volume, and mount the directory to a Docker volume, use File input mode. To stream data directly from Amazon S3 to the container, choose Pipe input mode.
To use a model for incremental training, choose File input model.
ShuffleConfig (dict) --
A configuration for a shuffle option for input data in a channel. If you use S3Prefix for S3DataType, this shuffles the results of the S3 key prefix matches. If you use ManifestFile, the order of the S3 object references in the ManifestFile is shuffled. If you use AugmentedManifestFile, the order of the JSON lines in the AugmentedManifestFile is shuffled. The shuffling order is determined using the Seed value.
For Pipe input mode, shuffling is done at the start of every epoch. With large datasets this ensures that the order of the training data is different for each epoch, it helps reduce bias and possible overfitting. In a multi-node training job when ShuffleConfig is combined with S3DataDistributionType of ShardedByS3Key, the data is shuffled across nodes so that the content sent to a particular node on the first epoch might be sent to a different node on the second epoch.
Seed (integer) -- [REQUIRED]
Determines the shuffling order in ShuffleConfig value.
OutputDataConfig (dict) -- [REQUIRED]
the path to the S3 bucket where you want to store model artifacts. Amazon SageMaker creates subfolders for the artifacts.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
// KMS Key Alias "alias/ExampleAlias"
// Amazon Resource Name (ARN) of a KMS Key Alias "arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"
If you use a KMS key ID or an alias of your KMS key, the Amazon SageMaker execution role must include permissions to call kms:Encrypt. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. Amazon SageMaker uses server-side encryption with KMS-managed keys for OutputDataConfig. If you use a bucket policy with an s3:PutObject permission that only allows objects with server-side encryption, set the condition key of s3:x-amz-server-side-encryption to "aws:kms". For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateTrainingJob, CreateTransformJob, or CreateHyperParameterTuningJob requests. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
S3OutputPath (string) -- [REQUIRED]
Identifies the S3 path where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
ResourceConfig (dict) -- [REQUIRED]
The resources, including the ML compute instances and ML storage volumes, to use for model training.
InstanceType (string) -- [REQUIRED]
The ML compute instance type.
InstanceCount (integer) -- [REQUIRED]
The number of ML compute instances to use. For distributed training, provide a value greater than 1.
VolumeSizeInGB (integer) -- [REQUIRED]
The size of the ML storage volume that you want to provision.
ML storage volumes store model artifacts and incremental states. Training algorithms might also use the ML storage volume for scratch space. If you want to store the training data in the ML storage volume, choose File as the TrainingInputMode in the algorithm specification.
You must specify sufficient ML storage for your scenario.
VolumeKmsKeyId (string) --
The Amazon Web Services KMS key that Amazon SageMaker uses to encrypt data on the storage volume attached to the ML compute instance(s) that run the training job.
The VolumeKmsKeyId can be in any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
StoppingCondition (dict) -- [REQUIRED]
Specifies a limit to how long a model training job can run. It also specifies how long a managed Spot training job has to complete. When the job reaches the time limit, Amazon SageMaker ends the training job. Use this API to cap model training costs.
To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms can use this 120-second window to save the model artifacts.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that a training or compilation job can run.
For compilation jobs, if the job does not complete during this time, you will receive a TimeOut error. We recommend starting with 900 seconds and increase as necessary based on your model.
For all other jobs, if the job does not complete during this time, Amazon SageMaker ends the job. When RetryStrategy is specified in the job request, MaxRuntimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt. The default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, that a managed Spot training job has to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the job can run. It must be equal to or greater than MaxRuntimeInSeconds. If the job does not complete during this time, Amazon SageMaker ends the job.
When RetryStrategy is specified in the job request, MaxWaitTimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt.
TransformJobDefinition (dict) --
The TransformJobDefinition object that describes the transform job that Amazon SageMaker runs to validate your algorithm.
MaxConcurrentTransforms (integer) --
The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.
MaxPayloadInMB (integer) --
The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).
BatchStrategy (string) --
A string that determines the number of records included in a single mini-batch.
SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.
Environment (dict) --
The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.
(string) --
(string) --
TransformInput (dict) -- [REQUIRED]
A description of the input source and the way the transform job consumes it.
DataSource (dict) -- [REQUIRED]
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
S3DataSource (dict) -- [REQUIRED]
The S3 location of the data source that is associated with a channel.
S3DataType (string) -- [REQUIRED]
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.
The following values are compatible: ManifestFile, S3Prefix
The following value is not compatible: AugmentedManifestFile
S3Uri (string) -- [REQUIRED]
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix.
A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
ContentType (string) --
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
CompressionType (string) --
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.
SplitType (string) --
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.
TransformOutput (dict) -- [REQUIRED]
Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.
S3OutputPath (string) -- [REQUIRED]
The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.
For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.
Accept (string) --
The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.
AssembleWith (string) --
Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
TransformResources (dict) -- [REQUIRED]
Identifies the ML compute instances for the transform job.
InstanceType (string) -- [REQUIRED]
The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.
InstanceCount (integer) -- [REQUIRED]
The number of ML compute instances to use in the transform job. For distributed transform jobs, specify a value greater than 1. The default value is 1.
VolumeKmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.
The VolumeKmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
boolean
Whether to certify the algorithm so that it can be listed in Amazon Web Services Marketplace.
list
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
dict
Response Syntax
{ 'AlgorithmArn': 'string' }
Response Structure
(dict) --
AlgorithmArn (string) --
The Amazon Resource Name (ARN) of the new algorithm.
{'ModelPackageVersionArn': 'string'}
Starts a model compilation job. After the model has been compiled, Amazon SageMaker saves the resulting model artifacts to an Amazon Simple Storage Service (Amazon S3) bucket that you specify.
If you choose to host your model using Amazon SageMaker hosting services, you can use the resulting model artifacts as part of the model. You can also use the artifacts with Amazon Web Services IoT Greengrass. In that case, deploy them as an ML resource.
In the request body, you provide the following:
A name for the compilation job
Information about the input model artifacts
The output location for the compiled model and the device (target) that the model runs on
The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker assumes to perform the model compilation job.
You can also provide a Tag to track the model compilation job's resource use and costs. The response body contains the CompilationJobArn for the compiled job.
To stop a model compilation job, use StopCompilationJob. To get information about a particular model compilation job, use DescribeCompilationJob. To get information about multiple model compilation jobs, use ListCompilationJobs.
See also: AWS API Documentation
Request Syntax
client.create_compilation_job( CompilationJobName='string', RoleArn='string', ModelPackageVersionArn='string', InputConfig={ 'S3Uri': 'string', 'DataInputConfig': 'string', 'Framework': 'TENSORFLOW'|'KERAS'|'MXNET'|'ONNX'|'PYTORCH'|'XGBOOST'|'TFLITE'|'DARKNET'|'SKLEARN', 'FrameworkVersion': 'string' }, OutputConfig={ 'S3OutputLocation': 'string', 'TargetDevice': 'lambda'|'ml_m4'|'ml_m5'|'ml_c4'|'ml_c5'|'ml_p2'|'ml_p3'|'ml_g4dn'|'ml_inf1'|'ml_eia2'|'jetson_tx1'|'jetson_tx2'|'jetson_nano'|'jetson_xavier'|'rasp3b'|'imx8qm'|'deeplens'|'rk3399'|'rk3288'|'aisage'|'sbe_c'|'qcs605'|'qcs603'|'sitara_am57x'|'amba_cv22'|'amba_cv25'|'x86_win32'|'x86_win64'|'coreml'|'jacinto_tda4vm'|'imx8mplus', 'TargetPlatform': { 'Os': 'ANDROID'|'LINUX', 'Arch': 'X86_64'|'X86'|'ARM64'|'ARM_EABI'|'ARM_EABIHF', 'Accelerator': 'INTEL_GRAPHICS'|'MALI'|'NVIDIA' }, 'CompilerOptions': 'string', 'KmsKeyId': 'string' }, VpcConfig={ 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] }, StoppingCondition={ 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 }, Tags=[ { 'Key': 'string', 'Value': 'string' }, ] )
string
[REQUIRED]
A name for the model compilation job. The name must be unique within the Amazon Web Services Region and within your Amazon Web Services account.
string
[REQUIRED]
The Amazon Resource Name (ARN) of an IAM role that enables Amazon SageMaker to perform tasks on your behalf.
During model compilation, Amazon SageMaker needs your permission to:
Read input data from an S3 bucket
Write model artifacts to an S3 bucket
Write logs to Amazon CloudWatch Logs
Publish metrics to Amazon CloudWatch
You grant permissions for all of these tasks to an IAM role. To pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission. For more information, see Amazon SageMaker Roles.
string
The Amazon Resource Name (ARN) of a versioned model package. Provide either a ModelPackageVersionArn or an InputConfig object in the request syntax. The presence of both objects in the CreateCompilationJob request will return an exception.
dict
Provides information about the location of input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.
S3Uri (string) -- [REQUIRED]
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
DataInputConfig (string) -- [REQUIRED]
Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are InputConfig$Framework specific.
TensorFlow: You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input":[1,1024,1024,3]}
If using the CLI, {\"input\":[1,1024,1024,3]}
Examples for two inputs:
If using the console, {"data1": [1,28,28,1], "data2":[1,28,28,1]}
If using the CLI, {\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]}
KERAS: You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, DataInputConfig should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input_1":[1,3,224,224]}
If using the CLI, {\"input_1\":[1,3,224,224]}
Examples for two inputs:
If using the console, {"input_1": [1,3,224,224], "input_2":[1,3,224,224]}
If using the CLI, {\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]}
MXNET/ONNX/DARKNET: You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"data":[1,3,1024,1024]}
If using the CLI, {\"data\":[1,3,1024,1024]}
Examples for two inputs:
If using the console, {"var1": [1,1,28,28], "var2":[1,1,28,28]}
If using the CLI, {\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]}
PyTorch: You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.
Examples for one input in dictionary format:
If using the console, {"input0":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224]}
Example for one input in list format: [[1,3,224,224]]
Examples for two inputs in dictionary format:
If using the console, {"input0":[1,3,224,224], "input1":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]}
Example for two inputs in list format: [[1,3,224,224], [1,3,224,224]]
XGBOOST: input data name and shape are not needed.
DataInputConfig supports the following parameters for CoreML OutputConfig$TargetDevice (ML Model format):
shape: Input shape, for example {"input_1": {"shape": [1,224,224,3]}}. In addition to static input shapes, CoreML converter supports Flexible input shapes:
Range Dimension. You can use the Range Dimension feature if you know the input shape will be within some specific interval in that dimension, for example: {"input_1": {"shape": ["1..10", 224, 224, 3]}}
Enumerated shapes. Sometimes, the models are trained to work only on a select set of inputs. You can enumerate all supported input shapes, for example: {"input_1": {"shape": [[1, 224, 224, 3], [1, 160, 160, 3]]}}
default_shape: Default input shape. You can set a default shape during conversion for both Range Dimension and Enumerated Shapes. For example {"input_1": {"shape": ["1..10", 224, 224, 3], "default_shape": [1, 224, 224, 3]}}
type: Input type. Allowed values: Image and Tensor. By default, the converter generates an ML Model with inputs of type Tensor (MultiArray). User can set input type to be Image. Image input type requires additional input parameters such as bias and scale.
bias: If the input type is an Image, you need to provide the bias vector.
scale: If the input type is an Image, you need to provide a scale factor.
CoreML ClassifierConfig parameters can be specified using OutputConfig$CompilerOptions. CoreML converter supports Tensorflow and PyTorch models. CoreML conversion examples:
Tensor type input:
"DataInputConfig": {"input_1": {"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3]}}
Tensor type input without input name (PyTorch):
"DataInputConfig": [{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224]}]
Image type input:
"DataInputConfig": {"input_1": {"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255}}
"CompilerOptions": {"class_labels": "imagenet_labels_1000.txt"}
Image type input without input name (PyTorch):
"DataInputConfig": [{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255}]
"CompilerOptions": {"class_labels": "imagenet_labels_1000.txt"}
Depending on the model format, DataInputConfig requires the following parameters for ml_eia2 OutputConfig:TargetDevice.
For TensorFlow models saved in the SavedModel format, specify the input names from signature_def_key and the input model shapes for DataInputConfig. Specify the signature_def_key in OutputConfig:CompilerOptions if the model does not use TensorFlow's default signature def key. For example:
"DataInputConfig": {"inputs": [1, 224, 224, 3]}
"CompilerOptions": {"signature_def_key": "serving_custom"}
For TensorFlow models saved as a frozen graph, specify the input tensor names and shapes in DataInputConfig and the output tensor names for output_names in OutputConfig:CompilerOptions. For example:
"DataInputConfig": {"input_tensor:0": [1, 224, 224, 3]}
"CompilerOptions": {"output_names": ["output_tensor:0"]}
Framework (string) -- [REQUIRED]
Identifies the framework in which the model was trained. For example: TENSORFLOW.
FrameworkVersion (string) --
Specifies the framework version to use.
This API field is only supported for PyTorch framework versions 1.4, 1.5, and 1.6 for cloud instance target devices: ml_c4, ml_c5, ml_m4, ml_m5, ml_p2, ml_p3, and ml_g4dn.
dict
[REQUIRED]
Provides information about the output location for the compiled model and the target device the model runs on.
S3OutputLocation (string) -- [REQUIRED]
Identifies the S3 bucket where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
TargetDevice (string) --
Identifies the target device or the machine learning instance that you want to run your model on after the compilation has completed. Alternatively, you can specify OS, architecture, and accelerator using TargetPlatform fields. It can be used instead of TargetPlatform.
TargetPlatform (dict) --
Contains information about a target platform that you want your model to run on, such as OS, architecture, and accelerators. It is an alternative of TargetDevice.
The following examples show how to configure the TargetPlatform and CompilerOptions JSON strings for popular target platforms:
Raspberry Pi 3 Model B+ "TargetPlatform": {"Os": "LINUX", "Arch": "ARM_EABIHF"}, "CompilerOptions": {'mattr': ['+neon']}
Jetson TX2 "TargetPlatform": {"Os": "LINUX", "Arch": "ARM64", "Accelerator": "NVIDIA"}, "CompilerOptions": {'gpu-code': 'sm_62', 'trt-ver': '6.0.1', 'cuda-ver': '10.0'}
EC2 m5.2xlarge instance OS "TargetPlatform": {"Os": "LINUX", "Arch": "X86_64", "Accelerator": "NVIDIA"}, "CompilerOptions": {'mcpu': 'skylake-avx512'}
RK3399 "TargetPlatform": {"Os": "LINUX", "Arch": "ARM64", "Accelerator": "MALI"}
ARMv7 phone (CPU) "TargetPlatform": {"Os": "ANDROID", "Arch": "ARM_EABI"}, "CompilerOptions": {'ANDROID_PLATFORM': 25, 'mattr': ['+neon']}
ARMv8 phone (CPU) "TargetPlatform": {"Os": "ANDROID", "Arch": "ARM64"}, "CompilerOptions": {'ANDROID_PLATFORM': 29}
Os (string) -- [REQUIRED]
Specifies a target platform OS.
LINUX: Linux-based operating systems.
ANDROID: Android operating systems. Android API level can be specified using the ANDROID_PLATFORM compiler option. For example, "CompilerOptions": {'ANDROID_PLATFORM': 28}
Arch (string) -- [REQUIRED]
Specifies a target platform architecture.
X86_64: 64-bit version of the x86 instruction set.
X86: 32-bit version of the x86 instruction set.
ARM64: ARMv8 64-bit CPU.
ARM_EABIHF: ARMv7 32-bit, Hard Float.
ARM_EABI: ARMv7 32-bit, Soft Float. Used by Android 32-bit ARM platform.
Accelerator (string) --
Specifies a target platform accelerator (optional).
NVIDIA: Nvidia graphics processing unit. It also requires gpu-code, trt-ver, cuda-ver compiler options
MALI: ARM Mali graphics processor
INTEL_GRAPHICS: Integrated Intel graphics
CompilerOptions (string) --
Specifies additional parameters for compiler options in JSON format. The compiler options are TargetPlatform specific. It is required for NVIDIA accelerators and highly recommended for CPU compilations. For any other cases, it is optional to specify CompilerOptions.
DTYPE: Specifies the data type for the input. When compiling for ml_* (except for ml_inf) instances using PyTorch framework, provide the data type (dtype) of the model's input. "float32" is used if "DTYPE" is not specified. Options for data type are:
float32: Use either "float" or "float32".
int64: Use either "int64" or "long".
For example, {"dtype" : "float32"}.
CPU: Compilation for CPU supports the following compiler options.
mcpu: CPU micro-architecture. For example, {'mcpu': 'skylake-avx512'}
mattr: CPU flags. For example, {'mattr': ['+neon', '+vfpv4']}
ARM: Details of ARM CPU compilations.
NEON: NEON is an implementation of the Advanced SIMD extension used in ARMv7 processors. For example, add {'mattr': ['+neon']} to the compiler options if compiling for ARM 32-bit platform with the NEON support.
NVIDIA: Compilation for NVIDIA GPU supports the following compiler options.
gpu_code: Specifies the targeted architecture.
trt-ver: Specifies the TensorRT versions in x.y.z. format.
cuda-ver: Specifies the CUDA version in x.y format.
For example, {'gpu-code': 'sm_72', 'trt-ver': '6.0.1', 'cuda-ver': '10.1'}
ANDROID: Compilation for the Android OS supports the following compiler options:
ANDROID_PLATFORM: Specifies the Android API levels. Available levels range from 21 to 29. For example, {'ANDROID_PLATFORM': 28}.
mattr: Add {'mattr': ['+neon']} to compiler options if compiling for ARM 32-bit platform with NEON support.
INFERENTIA: Compilation for target ml_inf1 uses compiler options passed in as a JSON string. For example, "CompilerOptions": "\"--verbose 1 --num-neuroncores 2 -O2\"". For information about supported compiler options, see Neuron Compiler CLI.
CoreML: Compilation for the CoreML OutputConfig$TargetDevice supports the following compiler options:
class_labels: Specifies the classification labels file name inside input tar.gz file. For example, {"class_labels": "imagenet_labels_1000.txt"}. Labels inside the txt file should be separated by newlines.
EIA: Compilation for the Elastic Inference Accelerator supports the following compiler options:
precision_mode: Specifies the precision of compiled artifacts. Supported values are "FP16" and "FP32". Default is "FP32".
signature_def_key: Specifies the signature to use for models in SavedModel format. Defaults is TensorFlow's default signature def key.
output_names: Specifies a list of output tensor names for models in FrozenGraph format. Set at most one API field, either: signature_def_key or output_names.
For example: {"precision_mode": "FP32", "output_names": ["output:0"]}
KmsKeyId (string) --
The Amazon Web Services Key Management Service key (Amazon Web Services KMS) that Amazon SageMaker uses to encrypt your output models with Amazon S3 server-side encryption after compilation job. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
dict
A VpcConfig object that specifies the VPC that you want your compilation job to connect to. Control access to your models by configuring the VPC. For more information, see Protect Compilation Jobs by Using an Amazon Virtual Private Cloud.
SecurityGroupIds (list) -- [REQUIRED]
The VPC security group IDs. IDs have the form of sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) -- [REQUIRED]
The ID of the subnets in the VPC that you want to connect the compilation job to for accessing the model in Amazon S3.
(string) --
dict
[REQUIRED]
Specifies a limit to how long a model compilation job can run. When the job reaches the time limit, Amazon SageMaker ends the compilation job. Use this API to cap model training costs.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that a training or compilation job can run.
For compilation jobs, if the job does not complete during this time, you will receive a TimeOut error. We recommend starting with 900 seconds and increase as necessary based on your model.
For all other jobs, if the job does not complete during this time, Amazon SageMaker ends the job. When RetryStrategy is specified in the job request, MaxRuntimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt. The default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, that a managed Spot training job has to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the job can run. It must be equal to or greater than MaxRuntimeInSeconds. If the job does not complete during this time, Amazon SageMaker ends the job.
When RetryStrategy is specified in the job request, MaxWaitTimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt.
list
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
dict
Response Syntax
{ 'CompilationJobArn': 'string' }
Response Structure
(dict) --
CompilationJobArn (string) --
If the action is successful, the service sends back an HTTP 200 response. Amazon SageMaker returns the following data in JSON format:
CompilationJobArn: The Amazon Resource Name (ARN) of the compiled job.
{'ProductionVariants': {'ServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}}}
Creates an endpoint configuration that Amazon SageMaker hosting services uses to deploy models. In the configuration, you identify one or more models, created using the CreateModel API, to deploy and the resources that you want Amazon SageMaker to provision. Then you call the CreateEndpoint API.
In the request, you define a ProductionVariant, for each model that you want to deploy. Each ProductionVariant parameter also describes the resources that you want Amazon SageMaker to provision. This includes the number and type of ML compute instances to deploy.
If you are hosting multiple models, you also assign a VariantWeight to specify how much traffic you want to allocate to each model. For example, suppose that you want to host two models, A and B, and you assign traffic weight 2 for model A and 1 for model B. Amazon SageMaker distributes two-thirds of the traffic to Model A, and one-third to model B.
See also: AWS API Documentation
Request Syntax
client.create_endpoint_config( EndpointConfigName='string', ProductionVariants=[ { 'VariantName': 'string', 'ModelName': 'string', 'InitialInstanceCount': 123, 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'InitialVariantWeight': ..., 'AcceleratorType': 'ml.eia1.medium'|'ml.eia1.large'|'ml.eia1.xlarge'|'ml.eia2.medium'|'ml.eia2.large'|'ml.eia2.xlarge', 'CoreDumpConfig': { 'DestinationS3Uri': 'string', 'KmsKeyId': 'string' }, 'ServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 } }, ], DataCaptureConfig={ 'EnableCapture': True|False, 'InitialSamplingPercentage': 123, 'DestinationS3Uri': 'string', 'KmsKeyId': 'string', 'CaptureOptions': [ { 'CaptureMode': 'Input'|'Output' }, ], 'CaptureContentTypeHeader': { 'CsvContentTypes': [ 'string', ], 'JsonContentTypes': [ 'string', ] } }, Tags=[ { 'Key': 'string', 'Value': 'string' }, ], KmsKeyId='string', AsyncInferenceConfig={ 'ClientConfig': { 'MaxConcurrentInvocationsPerInstance': 123 }, 'OutputConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string', 'NotificationConfig': { 'SuccessTopic': 'string', 'ErrorTopic': 'string' } } } )
string
[REQUIRED]
The name of the endpoint configuration. You specify this name in a CreateEndpoint request.
list
[REQUIRED]
An list of ProductionVariant objects, one for each model that you want to host at this endpoint.
(dict) --
Identifies a model that you want to host and the resources chosen to deploy for hosting it. If you are deploying multiple models, tell Amazon SageMaker how to distribute traffic among the models by specifying variant weights.
VariantName (string) -- [REQUIRED]
The name of the production variant.
ModelName (string) -- [REQUIRED]
The name of the model that you want to host. This is the name that you specified when creating the model.
InitialInstanceCount (integer) --
Number of instances to launch initially.
InstanceType (string) --
The ML compute instance type.
InitialVariantWeight (float) --
Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. The traffic to a production variant is determined by the ratio of the VariantWeight to the sum of all VariantWeight values across all ProductionVariants. If unspecified, it defaults to 1.0.
AcceleratorType (string) --
The size of the Elastic Inference (EI) instance to use for the production variant. EI instances provide on-demand GPU computing for inference. For more information, see Using Elastic Inference in Amazon SageMaker.
CoreDumpConfig (dict) --
Specifies configuration for a core dump from the model container when the process crashes.
DestinationS3Uri (string) -- [REQUIRED]
The Amazon S3 bucket to send the core dump to.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the core dump data at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
// KMS Key Alias "alias/ExampleAlias"
// Amazon Resource Name (ARN) of a KMS Key Alias "arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"
If you use a KMS key ID or an alias of your KMS key, the Amazon SageMaker execution role must include permissions to call kms:Encrypt. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. Amazon SageMaker uses server-side encryption with KMS-managed keys for OutputDataConfig. If you use a bucket policy with an s3:PutObject permission that only allows objects with server-side encryption, set the condition key of s3:x-amz-server-side-encryption to "aws:kms". For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateEndpoint and UpdateEndpoint requests. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
ServerlessConfig (dict) --
The serverless configuration for an endpoint. Specifies a serverless endpoint configuration instead of an instance-based endpoint configuration.
MemorySizeInMB (integer) -- [REQUIRED]
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) -- [REQUIRED]
The maximum number of concurrent invocations your serverless endpoint can process.
dict
EnableCapture (boolean) --
InitialSamplingPercentage (integer) -- [REQUIRED]
DestinationS3Uri (string) -- [REQUIRED]
KmsKeyId (string) --
CaptureOptions (list) -- [REQUIRED]
(dict) --
CaptureMode (string) -- [REQUIRED]
CaptureContentTypeHeader (dict) --
CsvContentTypes (list) --
(string) --
JsonContentTypes (list) --
(string) --
list
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
string
The Amazon Resource Name (ARN) of a Amazon Web Services Key Management Service key that Amazon SageMaker uses to encrypt data on the storage volume attached to the ML compute instance that hosts the endpoint.
The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
The KMS key policy must grant permission to the IAM role that you specify in your CreateEndpoint, UpdateEndpoint requests. For more information, refer to the Amazon Web Services Key Management Service section Using Key Policies in Amazon Web Services KMS
dict
Specifies configuration for how an endpoint performs asynchronous inference. This is a required field in order for your Endpoint to be invoked using InvokeEndpointAsync.
ClientConfig (dict) --
Configures the behavior of the client used by Amazon SageMaker to interact with the model container during asynchronous inference.
MaxConcurrentInvocationsPerInstance (integer) --
The maximum number of concurrent requests sent by the SageMaker client to the model container. If no value is provided, Amazon SageMaker will choose an optimal value for you.
OutputConfig (dict) -- [REQUIRED]
Specifies the configuration for asynchronous inference invocation outputs.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the asynchronous inference output in Amazon S3.
S3OutputPath (string) -- [REQUIRED]
The Amazon S3 location to upload inference responses to.
NotificationConfig (dict) --
Specifies the configuration for notifications of inference results for asynchronous inference.
SuccessTopic (string) --
Amazon SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
ErrorTopic (string) --
Amazon SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
dict
Response Syntax
{ 'EndpointConfigArn': 'string' }
Response Structure
(dict) --
EndpointConfigArn (string) --
The Amazon Resource Name (ARN) of the endpoint configuration.
{'Containers': {'InferenceSpecificationName': 'string'}, 'PrimaryContainer': {'InferenceSpecificationName': 'string'}}
Creates a model in Amazon SageMaker. In the request, you name the model and describe a primary container. For the primary container, you specify the Docker image that contains inference code, artifacts (from prior training), and a custom environment map that the inference code uses when you deploy the model for predictions.
Use this API to create a model if you want to use Amazon SageMaker hosting services or run a batch transform job.
To host your model, you create an endpoint configuration with the CreateEndpointConfig API, and then create an endpoint with the CreateEndpoint API. Amazon SageMaker then deploys all of the containers that you defined for the model in the hosting environment.
For an example that calls this method when deploying a model to Amazon SageMaker hosting services, see Deploy the Model to Amazon SageMaker Hosting Services (Amazon Web Services SDK for Python (Boto 3)).
To run a batch transform using your model, you start a job with the CreateTransformJob API. Amazon SageMaker uses your model and your dataset to get inferences which are then saved to a specified S3 location.
In the CreateModel request, you must define a container with the PrimaryContainer parameter.
In the request, you also provide an IAM role that Amazon SageMaker can assume to access model artifacts and docker image for deployment on ML compute hosting instances or for batch transform jobs. In addition, you also use the IAM role to manage permissions the inference code needs. For example, if the inference code access any other Amazon Web Services resources, you grant necessary permissions via this role.
See also: AWS API Documentation
Request Syntax
client.create_model( ModelName='string', PrimaryContainer={ 'ContainerHostname': 'string', 'Image': 'string', 'ImageConfig': { 'RepositoryAccessMode': 'Platform'|'Vpc', 'RepositoryAuthConfig': { 'RepositoryCredentialsProviderArn': 'string' } }, 'Mode': 'SingleModel'|'MultiModel', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' }, 'ModelPackageName': 'string', 'InferenceSpecificationName': 'string', 'MultiModelConfig': { 'ModelCacheSetting': 'Enabled'|'Disabled' } }, Containers=[ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageConfig': { 'RepositoryAccessMode': 'Platform'|'Vpc', 'RepositoryAuthConfig': { 'RepositoryCredentialsProviderArn': 'string' } }, 'Mode': 'SingleModel'|'MultiModel', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' }, 'ModelPackageName': 'string', 'InferenceSpecificationName': 'string', 'MultiModelConfig': { 'ModelCacheSetting': 'Enabled'|'Disabled' } }, ], InferenceExecutionConfig={ 'Mode': 'Serial'|'Direct' }, ExecutionRoleArn='string', Tags=[ { 'Key': 'string', 'Value': 'string' }, ], VpcConfig={ 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] }, EnableNetworkIsolation=True|False )
string
[REQUIRED]
The name of the new model.
dict
The location of the primary docker image containing inference code, associated artifacts, and custom environment map that the inference code uses when the model is deployed for predictions.
ContainerHostname (string) --
This parameter is ignored for models that contain only a PrimaryContainer.
When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.
Image (string) --
The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker
ImageConfig (dict) --
Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers
RepositoryAccessMode (string) -- [REQUIRED]
Set this to one of the following values:
Platform - The model image is hosted in Amazon ECR.
Vpc - The model image is hosted in a private Docker registry in your VPC.
RepositoryAuthConfig (dict) --
(Optional) Specifies an authentication configuration for the private docker registry where your model image is hosted. Specify a value for this property only if you specified Vpc as the value for the RepositoryAccessMode field, and the private Docker registry where the model image is hosted requires authentication.
RepositoryCredentialsProviderArn (string) -- [REQUIRED]
The Amazon Resource Name (ARN) of an Amazon Web Services Lambda function that provides credentials to authenticate to the private Docker registry where your model image is hosted. For information about how to create an Amazon Web Services Lambda function, see Create a Lambda function with the console in the Amazon Web Services Lambda Developer Guide.
Mode (string) --
Whether the container hosts a single model or multiple models.
ModelDataUrl (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.
If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelPackageName (string) --
The name or Amazon Resource Name (ARN) of the model package to use to create the model.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
MultiModelConfig (dict) --
Specifies additional configuration for multi-model endpoints.
ModelCacheSetting (string) --
Whether to cache models for a multi-model endpoint. By default, multi-model endpoints cache models so that a model does not have to be loaded into memory each time it is invoked. Some use cases do not benefit from model caching. For example, if an endpoint hosts a large number of models that are each invoked infrequently, the endpoint might perform better if you disable model caching. To disable model caching, set the value of this parameter to Disabled.
list
Specifies the containers in the inference pipeline.
(dict) --
Describes the container, as part of model definition.
ContainerHostname (string) --
This parameter is ignored for models that contain only a PrimaryContainer.
When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.
Image (string) --
The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker
ImageConfig (dict) --
Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers
RepositoryAccessMode (string) -- [REQUIRED]
Set this to one of the following values:
Platform - The model image is hosted in Amazon ECR.
Vpc - The model image is hosted in a private Docker registry in your VPC.
RepositoryAuthConfig (dict) --
(Optional) Specifies an authentication configuration for the private docker registry where your model image is hosted. Specify a value for this property only if you specified Vpc as the value for the RepositoryAccessMode field, and the private Docker registry where the model image is hosted requires authentication.
RepositoryCredentialsProviderArn (string) -- [REQUIRED]
The Amazon Resource Name (ARN) of an Amazon Web Services Lambda function that provides credentials to authenticate to the private Docker registry where your model image is hosted. For information about how to create an Amazon Web Services Lambda function, see Create a Lambda function with the console in the Amazon Web Services Lambda Developer Guide.
Mode (string) --
Whether the container hosts a single model or multiple models.
ModelDataUrl (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.
If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelPackageName (string) --
The name or Amazon Resource Name (ARN) of the model package to use to create the model.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
MultiModelConfig (dict) --
Specifies additional configuration for multi-model endpoints.
ModelCacheSetting (string) --
Whether to cache models for a multi-model endpoint. By default, multi-model endpoints cache models so that a model does not have to be loaded into memory each time it is invoked. Some use cases do not benefit from model caching. For example, if an endpoint hosts a large number of models that are each invoked infrequently, the endpoint might perform better if you disable model caching. To disable model caching, set the value of this parameter to Disabled.
dict
Specifies details of how containers in a multi-container endpoint are called.
Mode (string) -- [REQUIRED]
How containers in a multi-container are run. The following values are valid.
SERIAL - Containers run as a serial pipeline.
DIRECT - Only the individual container that you specify is run.
string
[REQUIRED]
The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances or for batch transform jobs. Deploying on ML compute instances is part of model hosting. For more information, see Amazon SageMaker Roles.
list
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
dict
A VpcConfig object that specifies the VPC that you want your model to connect to. Control access to and from your model container by configuring the VPC. VpcConfig is used in hosting services and in batch transform. For more information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch Transform Jobs by Using an Amazon Virtual Private Cloud.
SecurityGroupIds (list) -- [REQUIRED]
The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) -- [REQUIRED]
The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.
(string) --
boolean
Isolates the model container. No inbound or outbound network calls can be made to or from the model container.
dict
Response Syntax
{ 'ModelArn': 'string' }
Response Structure
(dict) --
ModelArn (string) --
The ARN of the model created in Amazon SageMaker.
{'AdditionalInferenceSpecifications': [{'Containers': [{'ContainerHostname': 'string', 'Environment': {'string': 'string'}, 'Framework': 'string', 'FrameworkVersion': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string', 'ProductId': 'string'}], 'Description': 'string', 'Name': 'string', 'SupportedContentTypes': ['string'], 'SupportedRealtimeInferenceInstanceTypes': ['ml.t2.medium ' '| ' 'ml.t2.large ' '| ' 'ml.t2.xlarge ' '| ' 'ml.t2.2xlarge ' '| ' 'ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.m5d.large ' '| ' 'ml.m5d.xlarge ' '| ' 'ml.m5d.2xlarge ' '| ' 'ml.m5d.4xlarge ' '| ' 'ml.m5d.12xlarge ' '| ' 'ml.m5d.24xlarge ' '| ' 'ml.c4.large ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.large ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.c5d.large ' '| ' 'ml.c5d.xlarge ' '| ' 'ml.c5d.2xlarge ' '| ' 'ml.c5d.4xlarge ' '| ' 'ml.c5d.9xlarge ' '| ' 'ml.c5d.18xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge ' '| ' 'ml.r5.large ' '| ' 'ml.r5.xlarge ' '| ' 'ml.r5.2xlarge ' '| ' 'ml.r5.4xlarge ' '| ' 'ml.r5.12xlarge ' '| ' 'ml.r5.24xlarge ' '| ' 'ml.r5d.large ' '| ' 'ml.r5d.xlarge ' '| ' 'ml.r5d.2xlarge ' '| ' 'ml.r5d.4xlarge ' '| ' 'ml.r5d.12xlarge ' '| ' 'ml.r5d.24xlarge ' '| ' 'ml.inf1.xlarge ' '| ' 'ml.inf1.2xlarge ' '| ' 'ml.inf1.6xlarge ' '| ' 'ml.inf1.24xlarge'], 'SupportedResponseMIMETypes': ['string'], 'SupportedTransformInstanceTypes': ['ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge']}], 'Domain': 'string', 'DriftCheckBaselines': {'Bias': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PostTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'Explainability': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelDataQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}, 'ModelMetrics': {'Bias': {'PostTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'SamplePayloadUrl': 'string', 'Task': 'string'}
Creates a model package that you can use to create Amazon SageMaker models or list on Amazon Web Services Marketplace, or a versioned model that is part of a model group. Buyers can subscribe to model packages listed on Amazon Web Services Marketplace to create models in Amazon SageMaker.
To create a model package by specifying a Docker container that contains your inference code and the Amazon S3 location of your model artifacts, provide values for InferenceSpecification. To create a model from an algorithm resource that you created or subscribed to in Amazon Web Services Marketplace, provide a value for SourceAlgorithmSpecification.
See also: AWS API Documentation
Request Syntax
client.create_model_package( ModelPackageName='string', ModelPackageGroupName='string', ModelPackageDescription='string', InferenceSpecification={ 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ValidationSpecification={ 'ValidationRole': 'string', 'ValidationProfiles': [ { 'ProfileName': 'string', 'TransformJobDefinition': { 'MaxConcurrentTransforms': 123, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' } } }, ] }, SourceAlgorithmSpecification={ 'SourceAlgorithms': [ { 'ModelDataUrl': 'string', 'AlgorithmName': 'string' }, ] }, CertifyForMarketplace=True|False, Tags=[ { 'Key': 'string', 'Value': 'string' }, ], ModelApprovalStatus='Approved'|'Rejected'|'PendingManualApproval', MetadataProperties={ 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, ModelMetrics={ 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Bias': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } }, ClientToken='string', CustomerMetadataProperties={ 'string': 'string' }, DriftCheckBaselines={ 'Bias': { 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } }, Domain='string', Task='string', SamplePayloadUrl='string', AdditionalInferenceSpecifications=[ { 'Name': 'string', 'Description': 'string', 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ] )
string
The name of the model package. The name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).
This parameter is required for unversioned models. It is not applicable to versioned models.
string
The name or Amazon Resource Name (ARN) of the model package group that this model version belongs to.
This parameter is required for versioned models, and does not apply to unversioned models.
string
A description of the model package.
dict
Specifies details about inference jobs that can be run with models based on this model package, including the following:
The Amazon ECR paths of containers that contain the inference code and model artifacts.
The instance types that the model package supports for transform jobs and real-time endpoints used for inference.
The input and output content formats that the model package supports for inference.
Containers (list) -- [REQUIRED]
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) -- [REQUIRED]
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) -- [REQUIRED]
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedContentTypes (list) -- [REQUIRED]
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) -- [REQUIRED]
The supported MIME types for the output data.
(string) --
dict
Specifies configurations for one or more transform jobs that Amazon SageMaker runs to test the model package.
ValidationRole (string) -- [REQUIRED]
The IAM roles to be used for the validation of the model package.
ValidationProfiles (list) -- [REQUIRED]
An array of ModelPackageValidationProfile objects, each of which specifies a batch transform job that Amazon SageMaker runs to validate your model package.
(dict) --
Contains data, such as the inputs and targeted instance types that are used in the process of validating the model package.
The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.
ProfileName (string) -- [REQUIRED]
The name of the profile for the model package.
TransformJobDefinition (dict) -- [REQUIRED]
The TransformJobDefinition object that describes the transform job used for the validation of the model package.
MaxConcurrentTransforms (integer) --
The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.
MaxPayloadInMB (integer) --
The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).
BatchStrategy (string) --
A string that determines the number of records included in a single mini-batch.
SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.
Environment (dict) --
The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.
(string) --
(string) --
TransformInput (dict) -- [REQUIRED]
A description of the input source and the way the transform job consumes it.
DataSource (dict) -- [REQUIRED]
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
S3DataSource (dict) -- [REQUIRED]
The S3 location of the data source that is associated with a channel.
S3DataType (string) -- [REQUIRED]
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.
The following values are compatible: ManifestFile, S3Prefix
The following value is not compatible: AugmentedManifestFile
S3Uri (string) -- [REQUIRED]
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix.
A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
ContentType (string) --
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
CompressionType (string) --
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.
SplitType (string) --
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.
TransformOutput (dict) -- [REQUIRED]
Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.
S3OutputPath (string) -- [REQUIRED]
The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.
For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.
Accept (string) --
The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.
AssembleWith (string) --
Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
TransformResources (dict) -- [REQUIRED]
Identifies the ML compute instances for the transform job.
InstanceType (string) -- [REQUIRED]
The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.
InstanceCount (integer) -- [REQUIRED]
The number of ML compute instances to use in the transform job. For distributed transform jobs, specify a value greater than 1. The default value is 1.
VolumeKmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.
The VolumeKmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
dict
Details about the algorithm that was used to create the model package.
SourceAlgorithms (list) -- [REQUIRED]
A list of the algorithms that were used to create a model package.
(dict) --
Specifies an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your Amazon SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
AlgorithmName (string) -- [REQUIRED]
The name of an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your Amazon SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.
boolean
Whether to certify the model package for listing on Amazon Web Services Marketplace.
This parameter is optional for unversioned models, and does not apply to versioned models.
list
A list of key value pairs associated with the model. For more information, see Tagging Amazon Web Services resources in the Amazon Web Services General Reference Guide.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
string
Whether the model is approved for deployment.
This parameter is optional for versioned models, and does not apply to unversioned models.
For versioned models, the value of this parameter must be set to Approved to deploy the model.
dict
Metadata properties of the tracking entity, trial, or trial component.
CommitId (string) --
The commit ID.
Repository (string) --
The repository.
GeneratedBy (string) --
The entity this entity was generated by.
ProjectId (string) --
The project ID.
dict
A structure that contains model metrics reports.
ModelQuality (dict) --
Metrics that measure the quality of a model.
Statistics (dict) --
Model quality statistics.
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Constraints (dict) --
Model quality constraints.
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
ModelDataQuality (dict) --
Metrics that measure the quality of the input data for a model.
Statistics (dict) --
Data quality statistics for a model.
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Constraints (dict) --
Data quality constraints for a model.
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Bias (dict) --
Metrics that measure bais in a model.
Report (dict) --
The bias report for a model
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
PreTrainingReport (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
PostTrainingReport (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Explainability (dict) --
Metrics that help explain a model.
Report (dict) --
The explainability report for a model.
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
string
A unique token that guarantees that the call to this API is idempotent.
This field is autopopulated if not provided.
dict
The metadata properties associated with the model package versions.
(string) --
(string) --
dict
Represents the drift check baselines that can be used when the model monitor is set using the model package. For more information, see the topic on Drift Detection against Previous Baselines in SageMaker Pipelines in the Amazon SageMaker Developer Guide.
Bias (dict) --
Represents the drift check bias baselines that can be used when the model monitor is set using the model package.
ConfigFile (dict) --
The bias config file for a model.
ContentType (string) --
The type of content stored in the file source.
ContentDigest (string) --
The digest of the file source.
S3Uri (string) -- [REQUIRED]
The Amazon S3 URI for the file source.
PreTrainingConstraints (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
PostTrainingConstraints (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Explainability (dict) --
Represents the drift check explainability baselines that can be used when the model monitor is set using the model package.
Constraints (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
ConfigFile (dict) --
The explainability config file for the model.
ContentType (string) --
The type of content stored in the file source.
ContentDigest (string) --
The digest of the file source.
S3Uri (string) -- [REQUIRED]
The Amazon S3 URI for the file source.
ModelQuality (dict) --
Represents the drift check model quality baselines that can be used when the model monitor is set using the model package.
Statistics (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Constraints (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
ModelDataQuality (dict) --
Represents the drift check model data quality baselines that can be used when the model monitor is set using the model package.
Statistics (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
Constraints (dict) --
ContentType (string) -- [REQUIRED]
ContentDigest (string) --
S3Uri (string) -- [REQUIRED]
string
The machine learning domain of your model package and its components. Common machine learning domains include computer vision and natural language processing.
string
The machine learning task your model package accomplishes. Common machine learning tasks include object detection and image classification.
string
The Amazon Simple Storage Service (Amazon S3) path where the sample payload are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
list
An array of additional Inference Specification objects. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.
(dict) --
A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package
Name (string) -- [REQUIRED]
A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.
Description (string) --
A description of the additional Inference specification
Containers (list) -- [REQUIRED]
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) -- [REQUIRED]
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) -- [REQUIRED]
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
dict
Response Syntax
{ 'ModelPackageArn': 'string' }
Response Structure
(dict) --
ModelPackageArn (string) --
The Amazon Resource Name (ARN) of the new model package.
{'LineageGroupArn': 'string'}
Describes an action.
See also: AWS API Documentation
Request Syntax
client.describe_action( ActionName='string' )
string
[REQUIRED]
The name of the action to describe.
dict
Response Syntax
{ 'ActionName': 'string', 'ActionArn': 'string', 'Source': { 'SourceUri': 'string', 'SourceType': 'string', 'SourceId': 'string' }, 'ActionType': 'string', 'Description': 'string', 'Status': 'Unknown'|'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'Properties': { 'string': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'LineageGroupArn': 'string' }
Response Structure
(dict) --
ActionName (string) --
The name of the action.
ActionArn (string) --
The Amazon Resource Name (ARN) of the action.
Source (dict) --
The source of the action.
SourceUri (string) --
The URI of the source.
SourceType (string) --
The type of the source.
SourceId (string) --
The ID of the source.
ActionType (string) --
The type of the action.
Description (string) --
The description of the action.
Status (string) --
The status of the action.
Properties (dict) --
A list of the action's properties.
(string) --
(string) --
CreationTime (datetime) --
When the action was created.
CreatedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LastModifiedTime (datetime) --
When the action was last modified.
LastModifiedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
MetadataProperties (dict) --
Metadata properties of the tracking entity, trial, or trial component.
CommitId (string) --
The commit ID.
Repository (string) --
The repository.
GeneratedBy (string) --
The entity this entity was generated by.
ProjectId (string) --
The project ID.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
{'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}}
Returns a description of the specified algorithm that is in your account.
See also: AWS API Documentation
Request Syntax
client.describe_algorithm( AlgorithmName='string' )
string
[REQUIRED]
The name of the algorithm to describe.
dict
Response Syntax
{ 'AlgorithmName': 'string', 'AlgorithmArn': 'string', 'AlgorithmDescription': 'string', 'CreationTime': datetime(2015, 1, 1), 'TrainingSpecification': { 'TrainingImage': 'string', 'TrainingImageDigest': 'string', 'SupportedHyperParameters': [ { 'Name': 'string', 'Description': 'string', 'Type': 'Integer'|'Continuous'|'Categorical'|'FreeText', 'Range': { 'IntegerParameterRangeSpecification': { 'MinValue': 'string', 'MaxValue': 'string' }, 'ContinuousParameterRangeSpecification': { 'MinValue': 'string', 'MaxValue': 'string' }, 'CategoricalParameterRangeSpecification': { 'Values': [ 'string', ] } }, 'IsTunable': True|False, 'IsRequired': True|False, 'DefaultValue': 'string' }, ], 'SupportedTrainingInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', ], 'SupportsDistributedTraining': True|False, 'MetricDefinitions': [ { 'Name': 'string', 'Regex': 'string' }, ], 'TrainingChannels': [ { 'Name': 'string', 'Description': 'string', 'IsRequired': True|False, 'SupportedContentTypes': [ 'string', ], 'SupportedCompressionTypes': [ 'None'|'Gzip', ], 'SupportedInputModes': [ 'Pipe'|'File'|'FastFile', ] }, ], 'SupportedTuningJobObjectiveMetrics': [ { 'Type': 'Maximize'|'Minimize', 'MetricName': 'string' }, ] }, 'InferenceSpecification': { 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, 'ValidationSpecification': { 'ValidationRole': 'string', 'ValidationProfiles': [ { 'ProfileName': 'string', 'TrainingJobDefinition': { 'TrainingInputMode': 'Pipe'|'File'|'FastFile', 'HyperParameters': { 'string': 'string' }, 'InputDataConfig': [ { 'ChannelName': 'string', 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'AttributeNames': [ 'string', ] }, 'FileSystemDataSource': { 'FileSystemId': 'string', 'FileSystemAccessMode': 'rw'|'ro', 'FileSystemType': 'EFS'|'FSxLustre', 'DirectoryPath': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'RecordWrapperType': 'None'|'RecordIO', 'InputMode': 'Pipe'|'File'|'FastFile', 'ShuffleConfig': { 'Seed': 123 } }, ], 'OutputDataConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, 'ResourceConfig': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', 'InstanceCount': 123, 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 } }, 'TransformJobDefinition': { 'MaxConcurrentTransforms': 123, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' } } }, ] }, 'AlgorithmStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting', 'AlgorithmStatusDetails': { 'ValidationStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ], 'ImageScanStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ] }, 'ProductId': 'string', 'CertifyForMarketplace': True|False }
Response Structure
(dict) --
AlgorithmName (string) --
The name of the algorithm being described.
AlgorithmArn (string) --
The Amazon Resource Name (ARN) of the algorithm.
AlgorithmDescription (string) --
A brief summary about the algorithm.
CreationTime (datetime) --
A timestamp specifying when the algorithm was created.
TrainingSpecification (dict) --
Details about training jobs run by this algorithm.
TrainingImage (string) --
The Amazon ECR registry path of the Docker image that contains the training algorithm.
TrainingImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
SupportedHyperParameters (list) --
A list of the HyperParameterSpecification objects, that define the supported hyperparameters. This is required if the algorithm supports automatic model tuning.>
(dict) --
Defines a hyperparameter to be used by an algorithm.
Name (string) --
The name of this hyperparameter. The name must be unique.
Description (string) --
A brief description of the hyperparameter.
Type (string) --
The type of this hyperparameter. The valid types are Integer, Continuous, Categorical, and FreeText.
Range (dict) --
The allowed range for this hyperparameter.
IntegerParameterRangeSpecification (dict) --
A IntegerParameterRangeSpecification object that defines the possible values for an integer hyperparameter.
MinValue (string) --
The minimum integer value allowed.
MaxValue (string) --
The maximum integer value allowed.
ContinuousParameterRangeSpecification (dict) --
A ContinuousParameterRangeSpecification object that defines the possible values for a continuous hyperparameter.
MinValue (string) --
The minimum floating-point value allowed.
MaxValue (string) --
The maximum floating-point value allowed.
CategoricalParameterRangeSpecification (dict) --
A CategoricalParameterRangeSpecification object that defines the possible values for a categorical hyperparameter.
Values (list) --
The allowed categories for the hyperparameter.
(string) --
IsTunable (boolean) --
Indicates whether this hyperparameter is tunable in a hyperparameter tuning job.
IsRequired (boolean) --
Indicates whether this hyperparameter is required.
DefaultValue (string) --
The default value for this hyperparameter. If a default value is specified, a hyperparameter cannot be required.
SupportedTrainingInstanceTypes (list) --
A list of the instance types that this algorithm can use for training.
(string) --
SupportsDistributedTraining (boolean) --
Indicates whether the algorithm supports distributed training. If set to false, buyers can't request more than one instance during training.
MetricDefinitions (list) --
A list of MetricDefinition objects, which are used for parsing metrics generated by the algorithm.
(dict) --
Specifies a metric that the training algorithm writes to stderr or stdout. Amazon SageMakerhyperparameter tuning captures all defined metrics. You specify one metric that a hyperparameter tuning job uses as its objective metric to choose the best training job.
Name (string) --
The name of the metric.
Regex (string) --
A regular expression that searches the output of a training job and gets the value of the metric. For more information about using regular expressions to define metrics, see Defining Objective Metrics.
TrainingChannels (list) --
A list of ChannelSpecification objects, which specify the input sources to be used by the algorithm.
(dict) --
Defines a named input source, called a channel, to be used by an algorithm.
Name (string) --
The name of the channel.
Description (string) --
A brief description of the channel.
IsRequired (boolean) --
Indicates whether the channel is required by the algorithm.
SupportedContentTypes (list) --
The supported MIME types for the data.
(string) --
SupportedCompressionTypes (list) --
The allowed compression types, if data compression is used.
(string) --
SupportedInputModes (list) --
The allowed input mode, either FILE or PIPE.
In FILE mode, Amazon SageMaker copies the data from the input source onto the local Amazon Elastic Block Store (Amazon EBS) volumes before starting your training algorithm. This is the most commonly used input mode.
In PIPE mode, Amazon SageMaker streams input data from the source directly to your algorithm without using the EBS volume.
(string) --
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
SupportedTuningJobObjectiveMetrics (list) --
A list of the metrics that the algorithm emits that can be used as the objective metric in a hyperparameter tuning job.
(dict) --
Defines the objective metric for a hyperparameter tuning job. Hyperparameter tuning uses the value of this metric to evaluate the training jobs it launches, and returns the training job that results in either the highest or lowest value for this metric, depending on the value you specify for the Type parameter.
Type (string) --
Whether to minimize or maximize the objective metric.
MetricName (string) --
The name of the metric to use for the objective metric.
InferenceSpecification (dict) --
Details about inference jobs that the algorithm runs.
Containers (list) --
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) --
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) --
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
ValidationSpecification (dict) --
Details about configurations for one or more training jobs that Amazon SageMaker runs to test the algorithm.
ValidationRole (string) --
The IAM roles that Amazon SageMaker uses to run the training jobs.
ValidationProfiles (list) --
An array of AlgorithmValidationProfile objects, each of which specifies a training job and batch transform job that Amazon SageMaker runs to validate your algorithm.
(dict) --
Defines a training job and a batch transform job that Amazon SageMaker runs to validate your algorithm.
The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.
ProfileName (string) --
The name of the profile for the algorithm. The name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).
TrainingJobDefinition (dict) --
The TrainingJobDefinition object that describes the training job that Amazon SageMaker runs to validate your algorithm.
TrainingInputMode (string) --
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
HyperParameters (dict) --
The hyperparameters used for the training job.
(string) --
(string) --
InputDataConfig (list) --
An array of Channel objects, each of which specifies an input source.
(dict) --
A channel is a named input source that training algorithms can consume.
ChannelName (string) --
The name of the channel.
DataSource (dict) --
The location of the channel data.
S3DataSource (dict) --
The S3 location of the data source that is associated with a channel.
S3DataType (string) --
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects that match the specified key name prefix for model training.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for model training.
If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile can only be used if the Channel's input mode is Pipe.
S3Uri (string) --
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix
A manifest might look like this: s3://bucketname/example.manifest A manifest is an S3 object which is a JSON file consisting of an array of elements. The first element is a prefix which is followed by one or more suffixes. SageMaker appends the suffix elements to the prefix to get a full set of S3Uri. Note that the prefix must be a valid non-empty S3Uri that precludes users from specifying a manifest whose individual S3Uri is sourced from different S3 buckets. The following code example shows a valid manifest format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] This JSON is equivalent to the following S3Uri list: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uri in this manifest is the input data for the channel for this data source. The object that each S3Uri points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
S3DataDistributionType (string) --
If you want Amazon SageMaker to replicate the entire dataset on each ML compute instance that is launched for model training, specify FullyReplicated.
If you want Amazon SageMaker to replicate a subset of data on each ML compute instance that is launched for model training, specify ShardedByS3Key. If there are n ML compute instances launched for a training job, each instance gets approximately 1/n of the number of S3 objects. In this case, model training on each machine uses only the subset of training data.
Don't choose more ML compute instances for training than available S3 objects. If you do, some nodes won't get any data and you will pay for nodes that aren't getting any training data. This applies in both File and Pipe modes. Keep this in mind when developing algorithms.
In distributed training, where you use multiple ML compute EC2 instances, you might choose ShardedByS3Key. If the algorithm requires copying training data to the ML storage volume (when TrainingInputMode is set to File), this copies 1/n of the number of objects.
AttributeNames (list) --
A list of one or more attribute names to use that are found in a specified augmented manifest file.
(string) --
FileSystemDataSource (dict) --
The file system that is associated with a channel.
FileSystemId (string) --
The file system id.
FileSystemAccessMode (string) --
The access mode of the mount of the directory associated with the channel. A directory can be mounted either in ro (read-only) or rw (read-write) mode.
FileSystemType (string) --
The file system type.
DirectoryPath (string) --
The full path to the directory to associate with the channel.
ContentType (string) --
The MIME type of the data.
CompressionType (string) --
If training data is compressed, the compression type. The default value is None. CompressionType is used only in Pipe input mode. In File mode, leave this field unset or set it to None.
RecordWrapperType (string) --
Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format. In this case, Amazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.
In File mode, leave this field unset or set it to None.
InputMode (string) --
(Optional) The input mode to use for the data channel in a training job. If you don't set a value for InputMode, Amazon SageMaker uses the value set for TrainingInputMode. Use this parameter to override the TrainingInputMode setting in a AlgorithmSpecification request when you have a channel that needs a different input mode from the training job's general setting. To download the data from Amazon Simple Storage Service (Amazon S3) to the provisioned ML storage volume, and mount the directory to a Docker volume, use File input mode. To stream data directly from Amazon S3 to the container, choose Pipe input mode.
To use a model for incremental training, choose File input model.
ShuffleConfig (dict) --
A configuration for a shuffle option for input data in a channel. If you use S3Prefix for S3DataType, this shuffles the results of the S3 key prefix matches. If you use ManifestFile, the order of the S3 object references in the ManifestFile is shuffled. If you use AugmentedManifestFile, the order of the JSON lines in the AugmentedManifestFile is shuffled. The shuffling order is determined using the Seed value.
For Pipe input mode, shuffling is done at the start of every epoch. With large datasets this ensures that the order of the training data is different for each epoch, it helps reduce bias and possible overfitting. In a multi-node training job when ShuffleConfig is combined with S3DataDistributionType of ShardedByS3Key, the data is shuffled across nodes so that the content sent to a particular node on the first epoch might be sent to a different node on the second epoch.
Seed (integer) --
Determines the shuffling order in ShuffleConfig value.
OutputDataConfig (dict) --
the path to the S3 bucket where you want to store model artifacts. Amazon SageMaker creates subfolders for the artifacts.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
// KMS Key Alias "alias/ExampleAlias"
// Amazon Resource Name (ARN) of a KMS Key Alias "arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"
If you use a KMS key ID or an alias of your KMS key, the Amazon SageMaker execution role must include permissions to call kms:Encrypt. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. Amazon SageMaker uses server-side encryption with KMS-managed keys for OutputDataConfig. If you use a bucket policy with an s3:PutObject permission that only allows objects with server-side encryption, set the condition key of s3:x-amz-server-side-encryption to "aws:kms". For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateTrainingJob, CreateTransformJob, or CreateHyperParameterTuningJob requests. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
S3OutputPath (string) --
Identifies the S3 path where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
ResourceConfig (dict) --
The resources, including the ML compute instances and ML storage volumes, to use for model training.
InstanceType (string) --
The ML compute instance type.
InstanceCount (integer) --
The number of ML compute instances to use. For distributed training, provide a value greater than 1.
VolumeSizeInGB (integer) --
The size of the ML storage volume that you want to provision.
ML storage volumes store model artifacts and incremental states. Training algorithms might also use the ML storage volume for scratch space. If you want to store the training data in the ML storage volume, choose File as the TrainingInputMode in the algorithm specification.
You must specify sufficient ML storage for your scenario.
VolumeKmsKeyId (string) --
The Amazon Web Services KMS key that Amazon SageMaker uses to encrypt data on the storage volume attached to the ML compute instance(s) that run the training job.
The VolumeKmsKeyId can be in any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
StoppingCondition (dict) --
Specifies a limit to how long a model training job can run. It also specifies how long a managed Spot training job has to complete. When the job reaches the time limit, Amazon SageMaker ends the training job. Use this API to cap model training costs.
To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms can use this 120-second window to save the model artifacts.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that a training or compilation job can run.
For compilation jobs, if the job does not complete during this time, you will receive a TimeOut error. We recommend starting with 900 seconds and increase as necessary based on your model.
For all other jobs, if the job does not complete during this time, Amazon SageMaker ends the job. When RetryStrategy is specified in the job request, MaxRuntimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt. The default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, that a managed Spot training job has to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the job can run. It must be equal to or greater than MaxRuntimeInSeconds. If the job does not complete during this time, Amazon SageMaker ends the job.
When RetryStrategy is specified in the job request, MaxWaitTimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt.
TransformJobDefinition (dict) --
The TransformJobDefinition object that describes the transform job that Amazon SageMaker runs to validate your algorithm.
MaxConcurrentTransforms (integer) --
The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.
MaxPayloadInMB (integer) --
The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).
BatchStrategy (string) --
A string that determines the number of records included in a single mini-batch.
SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.
Environment (dict) --
The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.
(string) --
(string) --
TransformInput (dict) --
A description of the input source and the way the transform job consumes it.
DataSource (dict) --
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
S3DataSource (dict) --
The S3 location of the data source that is associated with a channel.
S3DataType (string) --
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.
The following values are compatible: ManifestFile, S3Prefix
The following value is not compatible: AugmentedManifestFile
S3Uri (string) --
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix.
A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
ContentType (string) --
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
CompressionType (string) --
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.
SplitType (string) --
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.
TransformOutput (dict) --
Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.
S3OutputPath (string) --
The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.
For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.
Accept (string) --
The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.
AssembleWith (string) --
Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
TransformResources (dict) --
Identifies the ML compute instances for the transform job.
InstanceType (string) --
The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.
InstanceCount (integer) --
The number of ML compute instances to use in the transform job. For distributed transform jobs, specify a value greater than 1. The default value is 1.
VolumeKmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.
The VolumeKmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
AlgorithmStatus (string) --
The current status of the algorithm.
AlgorithmStatusDetails (dict) --
Details about the current status of the algorithm.
ValidationStatuses (list) --
The status of algorithm validation.
(dict) --
Represents the overall status of an algorithm.
Name (string) --
The name of the algorithm for which the overall status is being reported.
Status (string) --
The current status.
FailureReason (string) --
if the overall status is Failed, the reason for the failure.
ImageScanStatuses (list) --
The status of the scan of the algorithm's Docker image container.
(dict) --
Represents the overall status of an algorithm.
Name (string) --
The name of the algorithm for which the overall status is being reported.
Status (string) --
The current status.
FailureReason (string) --
if the overall status is Failed, the reason for the failure.
ProductId (string) --
The product identifier of the algorithm.
CertifyForMarketplace (boolean) --
Whether the algorithm is certified to be listed in Amazon Web Services Marketplace.
{'LineageGroupArn': 'string'}
Describes an artifact.
See also: AWS API Documentation
Request Syntax
client.describe_artifact( ArtifactArn='string' )
string
[REQUIRED]
The Amazon Resource Name (ARN) of the artifact to describe.
dict
Response Syntax
{ 'ArtifactName': 'string', 'ArtifactArn': 'string', 'Source': { 'SourceUri': 'string', 'SourceTypes': [ { 'SourceIdType': 'MD5Hash'|'S3ETag'|'S3Version'|'Custom', 'Value': 'string' }, ] }, 'ArtifactType': 'string', 'Properties': { 'string': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'LineageGroupArn': 'string' }
Response Structure
(dict) --
ArtifactName (string) --
The name of the artifact.
ArtifactArn (string) --
The Amazon Resource Name (ARN) of the artifact.
Source (dict) --
The source of the artifact.
SourceUri (string) --
The URI of the source.
SourceTypes (list) --
A list of source types.
(dict) --
The ID and ID type of an artifact source.
SourceIdType (string) --
The type of ID.
Value (string) --
The ID.
ArtifactType (string) --
The type of the artifact.
Properties (dict) --
A list of the artifact's properties.
(string) --
(string) --
CreationTime (datetime) --
When the artifact was created.
CreatedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LastModifiedTime (datetime) --
When the artifact was last modified.
LastModifiedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
MetadataProperties (dict) --
Metadata properties of the tracking entity, trial, or trial component.
CommitId (string) --
The commit ID.
Repository (string) --
The repository.
GeneratedBy (string) --
The entity this entity was generated by.
ProjectId (string) --
The project ID.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
{'ModelPackageVersionArn': 'string'}
Returns information about a model compilation job.
To create a model compilation job, use CreateCompilationJob. To get information about multiple model compilation jobs, use ListCompilationJobs.
See also: AWS API Documentation
Request Syntax
client.describe_compilation_job( CompilationJobName='string' )
string
[REQUIRED]
The name of the model compilation job that you want information about.
dict
Response Syntax
{ 'CompilationJobName': 'string', 'CompilationJobArn': 'string', 'CompilationJobStatus': 'INPROGRESS'|'COMPLETED'|'FAILED'|'STARTING'|'STOPPING'|'STOPPED', 'CompilationStartTime': datetime(2015, 1, 1), 'CompilationEndTime': datetime(2015, 1, 1), 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 }, 'InferenceImage': 'string', 'ModelPackageVersionArn': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'ModelArtifacts': { 'S3ModelArtifacts': 'string' }, 'ModelDigests': { 'ArtifactDigest': 'string' }, 'RoleArn': 'string', 'InputConfig': { 'S3Uri': 'string', 'DataInputConfig': 'string', 'Framework': 'TENSORFLOW'|'KERAS'|'MXNET'|'ONNX'|'PYTORCH'|'XGBOOST'|'TFLITE'|'DARKNET'|'SKLEARN', 'FrameworkVersion': 'string' }, 'OutputConfig': { 'S3OutputLocation': 'string', 'TargetDevice': 'lambda'|'ml_m4'|'ml_m5'|'ml_c4'|'ml_c5'|'ml_p2'|'ml_p3'|'ml_g4dn'|'ml_inf1'|'ml_eia2'|'jetson_tx1'|'jetson_tx2'|'jetson_nano'|'jetson_xavier'|'rasp3b'|'imx8qm'|'deeplens'|'rk3399'|'rk3288'|'aisage'|'sbe_c'|'qcs605'|'qcs603'|'sitara_am57x'|'amba_cv22'|'amba_cv25'|'x86_win32'|'x86_win64'|'coreml'|'jacinto_tda4vm'|'imx8mplus', 'TargetPlatform': { 'Os': 'ANDROID'|'LINUX', 'Arch': 'X86_64'|'X86'|'ARM64'|'ARM_EABI'|'ARM_EABIHF', 'Accelerator': 'INTEL_GRAPHICS'|'MALI'|'NVIDIA' }, 'CompilerOptions': 'string', 'KmsKeyId': 'string' }, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] } }
Response Structure
(dict) --
CompilationJobName (string) --
The name of the model compilation job.
CompilationJobArn (string) --
The Amazon Resource Name (ARN) of the model compilation job.
CompilationJobStatus (string) --
The status of the model compilation job.
CompilationStartTime (datetime) --
The time when the model compilation job started the CompilationJob instances.
You are billed for the time between this timestamp and the timestamp in the DescribeCompilationJobResponse$CompilationEndTime field. In Amazon CloudWatch Logs, the start time might be later than this time. That's because it takes time to download the compilation job, which depends on the size of the compilation job container.
CompilationEndTime (datetime) --
The time when the model compilation job on a compilation job instance ended. For a successful or stopped job, this is when the job's model artifacts have finished uploading. For a failed job, this is when Amazon SageMaker detected that the job failed.
StoppingCondition (dict) --
Specifies a limit to how long a model compilation job can run. When the job reaches the time limit, Amazon SageMaker ends the compilation job. Use this API to cap model training costs.
MaxRuntimeInSeconds (integer) --
The maximum length of time, in seconds, that a training or compilation job can run.
For compilation jobs, if the job does not complete during this time, you will receive a TimeOut error. We recommend starting with 900 seconds and increase as necessary based on your model.
For all other jobs, if the job does not complete during this time, Amazon SageMaker ends the job. When RetryStrategy is specified in the job request, MaxRuntimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt. The default value is 1 day. The maximum value is 28 days.
MaxWaitTimeInSeconds (integer) --
The maximum length of time, in seconds, that a managed Spot training job has to complete. It is the amount of time spent waiting for Spot capacity plus the amount of time the job can run. It must be equal to or greater than MaxRuntimeInSeconds. If the job does not complete during this time, Amazon SageMaker ends the job.
When RetryStrategy is specified in the job request, MaxWaitTimeInSeconds specifies the maximum time for all of the attempts in total, not each individual attempt.
InferenceImage (string) --
The inference image to use when compiling a model. Specify an image only if the target device is a cloud instance.
ModelPackageVersionArn (string) --
The Amazon Resource Name (ARN) of the versioned model package that was provided to SageMaker Neo when you initiated a compilation job.
CreationTime (datetime) --
The time that the model compilation job was created.
LastModifiedTime (datetime) --
The time that the status of the model compilation job was last modified.
FailureReason (string) --
If a model compilation job failed, the reason it failed.
ModelArtifacts (dict) --
Information about the location in Amazon S3 that has been configured for storing the model artifacts used in the compilation job.
S3ModelArtifacts (string) --
The path of the S3 object that contains the model artifacts. For example, s3://bucket-name/keynameprefix/model.tar.gz.
ModelDigests (dict) --
Provides a BLAKE2 hash value that identifies the compiled model artifacts in Amazon S3.
ArtifactDigest (string) --
Provides a hash value that uniquely identifies the stored model artifacts.
RoleArn (string) --
The Amazon Resource Name (ARN) of an IAM role that Amazon SageMaker assumes to perform the model compilation job.
InputConfig (dict) --
Information about the location in Amazon S3 of the input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.
S3Uri (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).
DataInputConfig (string) --
Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are InputConfig$Framework specific.
TensorFlow: You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input":[1,1024,1024,3]}
If using the CLI, {\"input\":[1,1024,1024,3]}
Examples for two inputs:
If using the console, {"data1": [1,28,28,1], "data2":[1,28,28,1]}
If using the CLI, {\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]}
KERAS: You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, DataInputConfig should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"input_1":[1,3,224,224]}
If using the CLI, {\"input_1\":[1,3,224,224]}
Examples for two inputs:
If using the console, {"input_1": [1,3,224,224], "input_2":[1,3,224,224]}
If using the CLI, {\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]}
MXNET/ONNX/DARKNET: You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.
Examples for one input:
If using the console, {"data":[1,3,1024,1024]}
If using the CLI, {\"data\":[1,3,1024,1024]}
Examples for two inputs:
If using the console, {"var1": [1,1,28,28], "var2":[1,1,28,28]}
If using the CLI, {\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]}
PyTorch: You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.
Examples for one input in dictionary format:
If using the console, {"input0":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224]}
Example for one input in list format: [[1,3,224,224]]
Examples for two inputs in dictionary format:
If using the console, {"input0":[1,3,224,224], "input1":[1,3,224,224]}
If using the CLI, {\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]}
Example for two inputs in list format: [[1,3,224,224], [1,3,224,224]]
XGBOOST: input data name and shape are not needed.
DataInputConfig supports the following parameters for CoreML OutputConfig$TargetDevice (ML Model format):
shape: Input shape, for example {"input_1": {"shape": [1,224,224,3]}}. In addition to static input shapes, CoreML converter supports Flexible input shapes:
Range Dimension. You can use the Range Dimension feature if you know the input shape will be within some specific interval in that dimension, for example: {"input_1": {"shape": ["1..10", 224, 224, 3]}}
Enumerated shapes. Sometimes, the models are trained to work only on a select set of inputs. You can enumerate all supported input shapes, for example: {"input_1": {"shape": [[1, 224, 224, 3], [1, 160, 160, 3]]}}
default_shape: Default input shape. You can set a default shape during conversion for both Range Dimension and Enumerated Shapes. For example {"input_1": {"shape": ["1..10", 224, 224, 3], "default_shape": [1, 224, 224, 3]}}
type: Input type. Allowed values: Image and Tensor. By default, the converter generates an ML Model with inputs of type Tensor (MultiArray). User can set input type to be Image. Image input type requires additional input parameters such as bias and scale.
bias: If the input type is an Image, you need to provide the bias vector.
scale: If the input type is an Image, you need to provide a scale factor.
CoreML ClassifierConfig parameters can be specified using OutputConfig$CompilerOptions. CoreML converter supports Tensorflow and PyTorch models. CoreML conversion examples:
Tensor type input:
"DataInputConfig": {"input_1": {"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3]}}
Tensor type input without input name (PyTorch):
"DataInputConfig": [{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224]}]
Image type input:
"DataInputConfig": {"input_1": {"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255}}
"CompilerOptions": {"class_labels": "imagenet_labels_1000.txt"}
Image type input without input name (PyTorch):
"DataInputConfig": [{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255}]
"CompilerOptions": {"class_labels": "imagenet_labels_1000.txt"}
Depending on the model format, DataInputConfig requires the following parameters for ml_eia2 OutputConfig:TargetDevice.
For TensorFlow models saved in the SavedModel format, specify the input names from signature_def_key and the input model shapes for DataInputConfig. Specify the signature_def_key in OutputConfig:CompilerOptions if the model does not use TensorFlow's default signature def key. For example:
"DataInputConfig": {"inputs": [1, 224, 224, 3]}
"CompilerOptions": {"signature_def_key": "serving_custom"}
For TensorFlow models saved as a frozen graph, specify the input tensor names and shapes in DataInputConfig and the output tensor names for output_names in OutputConfig:CompilerOptions. For example:
"DataInputConfig": {"input_tensor:0": [1, 224, 224, 3]}
"CompilerOptions": {"output_names": ["output_tensor:0"]}
Framework (string) --
Identifies the framework in which the model was trained. For example: TENSORFLOW.
FrameworkVersion (string) --
Specifies the framework version to use.
This API field is only supported for PyTorch framework versions 1.4, 1.5, and 1.6 for cloud instance target devices: ml_c4, ml_c5, ml_m4, ml_m5, ml_p2, ml_p3, and ml_g4dn.
OutputConfig (dict) --
Information about the output location for the compiled model and the target device that the model runs on.
S3OutputLocation (string) --
Identifies the S3 bucket where you want Amazon SageMaker to store the model artifacts. For example, s3://bucket-name/key-name-prefix.
TargetDevice (string) --
Identifies the target device or the machine learning instance that you want to run your model on after the compilation has completed. Alternatively, you can specify OS, architecture, and accelerator using TargetPlatform fields. It can be used instead of TargetPlatform.
TargetPlatform (dict) --
Contains information about a target platform that you want your model to run on, such as OS, architecture, and accelerators. It is an alternative of TargetDevice.
The following examples show how to configure the TargetPlatform and CompilerOptions JSON strings for popular target platforms:
Raspberry Pi 3 Model B+ "TargetPlatform": {"Os": "LINUX", "Arch": "ARM_EABIHF"}, "CompilerOptions": {'mattr': ['+neon']}
Jetson TX2 "TargetPlatform": {"Os": "LINUX", "Arch": "ARM64", "Accelerator": "NVIDIA"}, "CompilerOptions": {'gpu-code': 'sm_62', 'trt-ver': '6.0.1', 'cuda-ver': '10.0'}
EC2 m5.2xlarge instance OS "TargetPlatform": {"Os": "LINUX", "Arch": "X86_64", "Accelerator": "NVIDIA"}, "CompilerOptions": {'mcpu': 'skylake-avx512'}
RK3399 "TargetPlatform": {"Os": "LINUX", "Arch": "ARM64", "Accelerator": "MALI"}
ARMv7 phone (CPU) "TargetPlatform": {"Os": "ANDROID", "Arch": "ARM_EABI"}, "CompilerOptions": {'ANDROID_PLATFORM': 25, 'mattr': ['+neon']}
ARMv8 phone (CPU) "TargetPlatform": {"Os": "ANDROID", "Arch": "ARM64"}, "CompilerOptions": {'ANDROID_PLATFORM': 29}
Os (string) --
Specifies a target platform OS.
LINUX: Linux-based operating systems.
ANDROID: Android operating systems. Android API level can be specified using the ANDROID_PLATFORM compiler option. For example, "CompilerOptions": {'ANDROID_PLATFORM': 28}
Arch (string) --
Specifies a target platform architecture.
X86_64: 64-bit version of the x86 instruction set.
X86: 32-bit version of the x86 instruction set.
ARM64: ARMv8 64-bit CPU.
ARM_EABIHF: ARMv7 32-bit, Hard Float.
ARM_EABI: ARMv7 32-bit, Soft Float. Used by Android 32-bit ARM platform.
Accelerator (string) --
Specifies a target platform accelerator (optional).
NVIDIA: Nvidia graphics processing unit. It also requires gpu-code, trt-ver, cuda-ver compiler options
MALI: ARM Mali graphics processor
INTEL_GRAPHICS: Integrated Intel graphics
CompilerOptions (string) --
Specifies additional parameters for compiler options in JSON format. The compiler options are TargetPlatform specific. It is required for NVIDIA accelerators and highly recommended for CPU compilations. For any other cases, it is optional to specify CompilerOptions.
DTYPE: Specifies the data type for the input. When compiling for ml_* (except for ml_inf) instances using PyTorch framework, provide the data type (dtype) of the model's input. "float32" is used if "DTYPE" is not specified. Options for data type are:
float32: Use either "float" or "float32".
int64: Use either "int64" or "long".
For example, {"dtype" : "float32"}.
CPU: Compilation for CPU supports the following compiler options.
mcpu: CPU micro-architecture. For example, {'mcpu': 'skylake-avx512'}
mattr: CPU flags. For example, {'mattr': ['+neon', '+vfpv4']}
ARM: Details of ARM CPU compilations.
NEON: NEON is an implementation of the Advanced SIMD extension used in ARMv7 processors. For example, add {'mattr': ['+neon']} to the compiler options if compiling for ARM 32-bit platform with the NEON support.
NVIDIA: Compilation for NVIDIA GPU supports the following compiler options.
gpu_code: Specifies the targeted architecture.
trt-ver: Specifies the TensorRT versions in x.y.z. format.
cuda-ver: Specifies the CUDA version in x.y format.
For example, {'gpu-code': 'sm_72', 'trt-ver': '6.0.1', 'cuda-ver': '10.1'}
ANDROID: Compilation for the Android OS supports the following compiler options:
ANDROID_PLATFORM: Specifies the Android API levels. Available levels range from 21 to 29. For example, {'ANDROID_PLATFORM': 28}.
mattr: Add {'mattr': ['+neon']} to compiler options if compiling for ARM 32-bit platform with NEON support.
INFERENTIA: Compilation for target ml_inf1 uses compiler options passed in as a JSON string. For example, "CompilerOptions": "\"--verbose 1 --num-neuroncores 2 -O2\"". For information about supported compiler options, see Neuron Compiler CLI.
CoreML: Compilation for the CoreML OutputConfig$TargetDevice supports the following compiler options:
class_labels: Specifies the classification labels file name inside input tar.gz file. For example, {"class_labels": "imagenet_labels_1000.txt"}. Labels inside the txt file should be separated by newlines.
EIA: Compilation for the Elastic Inference Accelerator supports the following compiler options:
precision_mode: Specifies the precision of compiled artifacts. Supported values are "FP16" and "FP32". Default is "FP32".
signature_def_key: Specifies the signature to use for models in SavedModel format. Defaults is TensorFlow's default signature def key.
output_names: Specifies a list of output tensor names for models in FrozenGraph format. Set at most one API field, either: signature_def_key or output_names.
For example: {"precision_mode": "FP32", "output_names": ["output:0"]}
KmsKeyId (string) --
The Amazon Web Services Key Management Service key (Amazon Web Services KMS) that Amazon SageMaker uses to encrypt your output models with Amazon S3 server-side encryption after compilation job. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
VpcConfig (dict) --
A VpcConfig object that specifies the VPC that you want your compilation job to connect to. Control access to your models by configuring the VPC. For more information, see Protect Compilation Jobs by Using an Amazon Virtual Private Cloud.
SecurityGroupIds (list) --
The VPC security group IDs. IDs have the form of sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) --
The ID of the subnets in the VPC that you want to connect the compilation job to for accessing the model in Amazon S3.
(string) --
{'LineageGroupArn': 'string'}
Describes a context.
See also: AWS API Documentation
Request Syntax
client.describe_context( ContextName='string' )
string
[REQUIRED]
The name of the context to describe.
dict
Response Syntax
{ 'ContextName': 'string', 'ContextArn': 'string', 'Source': { 'SourceUri': 'string', 'SourceType': 'string', 'SourceId': 'string' }, 'ContextType': 'string', 'Description': 'string', 'Properties': { 'string': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LineageGroupArn': 'string' }
Response Structure
(dict) --
ContextName (string) --
The name of the context.
ContextArn (string) --
The Amazon Resource Name (ARN) of the context.
Source (dict) --
The source of the context.
SourceUri (string) --
The URI of the source.
SourceType (string) --
The type of the source.
SourceId (string) --
The ID of the source.
ContextType (string) --
The type of the context.
Description (string) --
The description of the context.
Properties (dict) --
A list of the context's properties.
(string) --
(string) --
CreationTime (datetime) --
When the context was created.
CreatedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LastModifiedTime (datetime) --
When the context was last modified.
LastModifiedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
{'PendingDeploymentSummary': {'ProductionVariants': {'CurrentServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}, 'DesiredServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}}}, 'ProductionVariants': {'CurrentServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}, 'DesiredServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}}}
Returns the description of an endpoint.
See also: AWS API Documentation
Request Syntax
client.describe_endpoint( EndpointName='string' )
string
[REQUIRED]
The name of the endpoint.
dict
Response Syntax
{ 'EndpointName': 'string', 'EndpointArn': 'string', 'EndpointConfigName': 'string', 'ProductionVariants': [ { 'VariantName': 'string', 'DeployedImages': [ { 'SpecifiedImage': 'string', 'ResolvedImage': 'string', 'ResolutionTime': datetime(2015, 1, 1) }, ], 'CurrentWeight': ..., 'DesiredWeight': ..., 'CurrentInstanceCount': 123, 'DesiredInstanceCount': 123, 'VariantStatus': [ { 'Status': 'Creating'|'Updating'|'Deleting'|'ActivatingTraffic'|'Baking', 'StatusMessage': 'string', 'StartTime': datetime(2015, 1, 1) }, ], 'CurrentServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 }, 'DesiredServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 } }, ], 'DataCaptureConfig': { 'EnableCapture': True|False, 'CaptureStatus': 'Started'|'Stopped', 'CurrentSamplingPercentage': 123, 'DestinationS3Uri': 'string', 'KmsKeyId': 'string' }, 'EndpointStatus': 'OutOfService'|'Creating'|'Updating'|'SystemUpdating'|'RollingBack'|'InService'|'Deleting'|'Failed', 'FailureReason': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'LastDeploymentConfig': { 'BlueGreenUpdatePolicy': { 'TrafficRoutingConfiguration': { 'Type': 'ALL_AT_ONCE'|'CANARY'|'LINEAR', 'WaitIntervalInSeconds': 123, 'CanarySize': { 'Type': 'INSTANCE_COUNT'|'CAPACITY_PERCENT', 'Value': 123 }, 'LinearStepSize': { 'Type': 'INSTANCE_COUNT'|'CAPACITY_PERCENT', 'Value': 123 } }, 'TerminationWaitInSeconds': 123, 'MaximumExecutionTimeoutInSeconds': 123 }, 'AutoRollbackConfiguration': { 'Alarms': [ { 'AlarmName': 'string' }, ] } }, 'AsyncInferenceConfig': { 'ClientConfig': { 'MaxConcurrentInvocationsPerInstance': 123 }, 'OutputConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string', 'NotificationConfig': { 'SuccessTopic': 'string', 'ErrorTopic': 'string' } } }, 'PendingDeploymentSummary': { 'EndpointConfigName': 'string', 'ProductionVariants': [ { 'VariantName': 'string', 'DeployedImages': [ { 'SpecifiedImage': 'string', 'ResolvedImage': 'string', 'ResolutionTime': datetime(2015, 1, 1) }, ], 'CurrentWeight': ..., 'DesiredWeight': ..., 'CurrentInstanceCount': 123, 'DesiredInstanceCount': 123, 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'AcceleratorType': 'ml.eia1.medium'|'ml.eia1.large'|'ml.eia1.xlarge'|'ml.eia2.medium'|'ml.eia2.large'|'ml.eia2.xlarge', 'VariantStatus': [ { 'Status': 'Creating'|'Updating'|'Deleting'|'ActivatingTraffic'|'Baking', 'StatusMessage': 'string', 'StartTime': datetime(2015, 1, 1) }, ], 'CurrentServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 }, 'DesiredServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 } }, ], 'StartTime': datetime(2015, 1, 1) } }
Response Structure
(dict) --
EndpointName (string) --
Name of the endpoint.
EndpointArn (string) --
The Amazon Resource Name (ARN) of the endpoint.
EndpointConfigName (string) --
The name of the endpoint configuration associated with this endpoint.
ProductionVariants (list) --
An array of ProductionVariantSummary objects, one for each model hosted behind this endpoint.
(dict) --
Describes weight and capacities for a production variant associated with an endpoint. If you sent a request to the UpdateEndpointWeightsAndCapacities API and the endpoint status is Updating, you get different desired and current values.
VariantName (string) --
The name of the variant.
DeployedImages (list) --
An array of DeployedImage objects that specify the Amazon EC2 Container Registry paths of the inference images deployed on instances of this ProductionVariant.
(dict) --
Gets the Amazon EC2 Container Registry path of the docker image of the model that is hosted in this ProductionVariant.
If you used the registry/repository[:tag] form to specify the image path of the primary container when you created the model hosted in this ProductionVariant, the path resolves to a path of the form registry/repository[@digest]. A digest is a hash value that identifies a specific version of an image. For information about Amazon ECR paths, see Pulling an Image in the Amazon ECR User Guide.
SpecifiedImage (string) --
The image path you specified when you created the model.
ResolvedImage (string) --
The specific digest path of the image hosted in this ProductionVariant.
ResolutionTime (datetime) --
The date and time when the image path for the model resolved to the ResolvedImage
CurrentWeight (float) --
The weight associated with the variant.
DesiredWeight (float) --
The requested weight, as specified in the UpdateEndpointWeightsAndCapacities request.
CurrentInstanceCount (integer) --
The number of instances associated with the variant.
DesiredInstanceCount (integer) --
The number of instances requested in the UpdateEndpointWeightsAndCapacities request.
VariantStatus (list) --
The endpoint variant status which describes the current deployment stage status or operational status.
(dict) --
Describes the status of the production variant.
Status (string) --
The endpoint variant status which describes the current deployment stage status or operational status.
Creating: Creating inference resources for the production variant.
Deleting: Terminating inference resources for the production variant.
Updating: Updating capacity for the production variant.
ActivatingTraffic: Turning on traffic for the production variant.
Baking: Waiting period to monitor the CloudWatch alarms in the automatic rollback configuration.
StatusMessage (string) --
A message that describes the status of the production variant.
StartTime (datetime) --
The start time of the current status change.
CurrentServerlessConfig (dict) --
The serverless configuration for the endpoint.
MemorySizeInMB (integer) --
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) --
The maximum number of concurrent invocations your serverless endpoint can process.
DesiredServerlessConfig (dict) --
The serverless configuration requested for the endpoint update.
MemorySizeInMB (integer) --
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) --
The maximum number of concurrent invocations your serverless endpoint can process.
DataCaptureConfig (dict) --
EnableCapture (boolean) --
CaptureStatus (string) --
CurrentSamplingPercentage (integer) --
DestinationS3Uri (string) --
KmsKeyId (string) --
EndpointStatus (string) --
The status of the endpoint.
OutOfService: Endpoint is not available to take incoming requests.
Creating: CreateEndpoint is executing.
Updating: UpdateEndpoint or UpdateEndpointWeightsAndCapacities is executing.
SystemUpdating: Endpoint is undergoing maintenance and cannot be updated or deleted or re-scaled until it has completed. This maintenance operation does not change any customer-specified values such as VPC config, KMS encryption, model, instance type, or instance count.
RollingBack: Endpoint fails to scale up or down or change its variant weight and is in the process of rolling back to its previous configuration. Once the rollback completes, endpoint returns to an InService status. This transitional status only applies to an endpoint that has autoscaling enabled and is undergoing variant weight or capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called explicitly.
InService: Endpoint is available to process incoming requests.
Deleting: DeleteEndpoint is executing.
Failed: Endpoint could not be created, updated, or re-scaled. Use DescribeEndpointOutput$FailureReason for information about the failure. DeleteEndpoint is the only operation that can be performed on a failed endpoint.
FailureReason (string) --
If the status of the endpoint is Failed, the reason why it failed.
CreationTime (datetime) --
A timestamp that shows when the endpoint was created.
LastModifiedTime (datetime) --
A timestamp that shows when the endpoint was last modified.
LastDeploymentConfig (dict) --
The most recent deployment configuration for the endpoint.
BlueGreenUpdatePolicy (dict) --
Update policy for a blue/green deployment. If this update policy is specified, SageMaker creates a new fleet during the deployment while maintaining the old fleet. SageMaker flips traffic to the new fleet according to the specified traffic routing configuration. Only one update policy should be used in the deployment configuration. If no update policy is specified, SageMaker uses a blue/green deployment strategy with all at once traffic shifting by default.
TrafficRoutingConfiguration (dict) --
Defines the traffic routing strategy to shift traffic from the old fleet to the new fleet during an endpoint deployment.
Type (string) --
Traffic routing strategy type.
ALL_AT_ONCE: Endpoint traffic shifts to the new fleet in a single step.
CANARY: Endpoint traffic shifts to the new fleet in two steps. The first step is the canary, which is a small portion of the traffic. The second step is the remainder of the traffic.
LINEAR: Endpoint traffic shifts to the new fleet in n steps of a configurable size.
WaitIntervalInSeconds (integer) --
The waiting time (in seconds) between incremental steps to turn on traffic on the new endpoint fleet.
CanarySize (dict) --
Batch size for the first step to turn on traffic on the new endpoint fleet. Value must be less than or equal to 50% of the variant's total instance count.
Type (string) --
Specifies the endpoint capacity type.
INSTANCE_COUNT: The endpoint activates based on the number of instances.
CAPACITY_PERCENT: The endpoint activates based on the specified percentage of capacity.
Value (integer) --
Defines the capacity size, either as a number of instances or a capacity percentage.
LinearStepSize (dict) --
Batch size for each step to turn on traffic on the new endpoint fleet. Value must be 10-50% of the variant's total instance count.
Type (string) --
Specifies the endpoint capacity type.
INSTANCE_COUNT: The endpoint activates based on the number of instances.
CAPACITY_PERCENT: The endpoint activates based on the specified percentage of capacity.
Value (integer) --
Defines the capacity size, either as a number of instances or a capacity percentage.
TerminationWaitInSeconds (integer) --
Additional waiting time in seconds after the completion of an endpoint deployment before terminating the old endpoint fleet. Default is 0.
MaximumExecutionTimeoutInSeconds (integer) --
Maximum execution timeout for the deployment. Note that the timeout value should be larger than the total waiting time specified in TerminationWaitInSeconds and WaitIntervalInSeconds.
AutoRollbackConfiguration (dict) --
Automatic rollback configuration for handling endpoint deployment failures and recovery.
Alarms (list) --
List of CloudWatch alarms in your account that are configured to monitor metrics on an endpoint. If any alarms are tripped during a deployment, SageMaker rolls back the deployment.
(dict) --
An Amazon CloudWatch alarm configured to monitor metrics on an endpoint.
AlarmName (string) --
The name of a CloudWatch alarm in your account.
AsyncInferenceConfig (dict) --
Returns the description of an endpoint configuration created using the CreateEndpointConfig API.
ClientConfig (dict) --
Configures the behavior of the client used by Amazon SageMaker to interact with the model container during asynchronous inference.
MaxConcurrentInvocationsPerInstance (integer) --
The maximum number of concurrent requests sent by the SageMaker client to the model container. If no value is provided, Amazon SageMaker will choose an optimal value for you.
OutputConfig (dict) --
Specifies the configuration for asynchronous inference invocation outputs.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the asynchronous inference output in Amazon S3.
S3OutputPath (string) --
The Amazon S3 location to upload inference responses to.
NotificationConfig (dict) --
Specifies the configuration for notifications of inference results for asynchronous inference.
SuccessTopic (string) --
Amazon SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
ErrorTopic (string) --
Amazon SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
PendingDeploymentSummary (dict) --
Returns the summary of an in-progress deployment. This field is only returned when the endpoint is creating or updating with a new endpoint configuration.
EndpointConfigName (string) --
The name of the endpoint configuration used in the deployment.
ProductionVariants (list) --
List of PendingProductionVariantSummary objects.
(dict) --
The production variant summary for a deployment when an endpoint is creating or updating with the CreateEndpoint or UpdateEndpoint operations. Describes the ``VariantStatus ``, weight and capacity for a production variant associated with an endpoint.
VariantName (string) --
The name of the variant.
DeployedImages (list) --
An array of DeployedImage objects that specify the Amazon EC2 Container Registry paths of the inference images deployed on instances of this ProductionVariant.
(dict) --
Gets the Amazon EC2 Container Registry path of the docker image of the model that is hosted in this ProductionVariant.
If you used the registry/repository[:tag] form to specify the image path of the primary container when you created the model hosted in this ProductionVariant, the path resolves to a path of the form registry/repository[@digest]. A digest is a hash value that identifies a specific version of an image. For information about Amazon ECR paths, see Pulling an Image in the Amazon ECR User Guide.
SpecifiedImage (string) --
The image path you specified when you created the model.
ResolvedImage (string) --
The specific digest path of the image hosted in this ProductionVariant.
ResolutionTime (datetime) --
The date and time when the image path for the model resolved to the ResolvedImage
CurrentWeight (float) --
The weight associated with the variant.
DesiredWeight (float) --
The requested weight for the variant in this deployment, as specified in the endpoint configuration for the endpoint. The value is taken from the request to the CreateEndpointConfig operation.
CurrentInstanceCount (integer) --
The number of instances associated with the variant.
DesiredInstanceCount (integer) --
The number of instances requested in this deployment, as specified in the endpoint configuration for the endpoint. The value is taken from the request to the CreateEndpointConfig operation.
InstanceType (string) --
The type of instances associated with the variant.
AcceleratorType (string) --
The size of the Elastic Inference (EI) instance to use for the production variant. EI instances provide on-demand GPU computing for inference. For more information, see Using Elastic Inference in Amazon SageMaker.
VariantStatus (list) --
The endpoint variant status which describes the current deployment stage status or operational status.
(dict) --
Describes the status of the production variant.
Status (string) --
The endpoint variant status which describes the current deployment stage status or operational status.
Creating: Creating inference resources for the production variant.
Deleting: Terminating inference resources for the production variant.
Updating: Updating capacity for the production variant.
ActivatingTraffic: Turning on traffic for the production variant.
Baking: Waiting period to monitor the CloudWatch alarms in the automatic rollback configuration.
StatusMessage (string) --
A message that describes the status of the production variant.
StartTime (datetime) --
The start time of the current status change.
CurrentServerlessConfig (dict) --
The serverless configuration for the endpoint.
MemorySizeInMB (integer) --
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) --
The maximum number of concurrent invocations your serverless endpoint can process.
DesiredServerlessConfig (dict) --
The serverless configuration requested for this deployment, as specified in the endpoint configuration for the endpoint.
MemorySizeInMB (integer) --
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) --
The maximum number of concurrent invocations your serverless endpoint can process.
StartTime (datetime) --
The start time of the deployment.
{'ProductionVariants': {'ServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}}}
Returns the description of an endpoint configuration created using the CreateEndpointConfig API.
See also: AWS API Documentation
Request Syntax
client.describe_endpoint_config( EndpointConfigName='string' )
string
[REQUIRED]
The name of the endpoint configuration.
dict
Response Syntax
{ 'EndpointConfigName': 'string', 'EndpointConfigArn': 'string', 'ProductionVariants': [ { 'VariantName': 'string', 'ModelName': 'string', 'InitialInstanceCount': 123, 'InstanceType': 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', 'InitialVariantWeight': ..., 'AcceleratorType': 'ml.eia1.medium'|'ml.eia1.large'|'ml.eia1.xlarge'|'ml.eia2.medium'|'ml.eia2.large'|'ml.eia2.xlarge', 'CoreDumpConfig': { 'DestinationS3Uri': 'string', 'KmsKeyId': 'string' }, 'ServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 } }, ], 'DataCaptureConfig': { 'EnableCapture': True|False, 'InitialSamplingPercentage': 123, 'DestinationS3Uri': 'string', 'KmsKeyId': 'string', 'CaptureOptions': [ { 'CaptureMode': 'Input'|'Output' }, ], 'CaptureContentTypeHeader': { 'CsvContentTypes': [ 'string', ], 'JsonContentTypes': [ 'string', ] } }, 'KmsKeyId': 'string', 'CreationTime': datetime(2015, 1, 1), 'AsyncInferenceConfig': { 'ClientConfig': { 'MaxConcurrentInvocationsPerInstance': 123 }, 'OutputConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string', 'NotificationConfig': { 'SuccessTopic': 'string', 'ErrorTopic': 'string' } } } }
Response Structure
(dict) --
EndpointConfigName (string) --
Name of the Amazon SageMaker endpoint configuration.
EndpointConfigArn (string) --
The Amazon Resource Name (ARN) of the endpoint configuration.
ProductionVariants (list) --
An array of ProductionVariant objects, one for each model that you want to host at this endpoint.
(dict) --
Identifies a model that you want to host and the resources chosen to deploy for hosting it. If you are deploying multiple models, tell Amazon SageMaker how to distribute traffic among the models by specifying variant weights.
VariantName (string) --
The name of the production variant.
ModelName (string) --
The name of the model that you want to host. This is the name that you specified when creating the model.
InitialInstanceCount (integer) --
Number of instances to launch initially.
InstanceType (string) --
The ML compute instance type.
InitialVariantWeight (float) --
Determines initial traffic distribution among all of the models that you specify in the endpoint configuration. The traffic to a production variant is determined by the ratio of the VariantWeight to the sum of all VariantWeight values across all ProductionVariants. If unspecified, it defaults to 1.0.
AcceleratorType (string) --
The size of the Elastic Inference (EI) instance to use for the production variant. EI instances provide on-demand GPU computing for inference. For more information, see Using Elastic Inference in Amazon SageMaker.
CoreDumpConfig (dict) --
Specifies configuration for a core dump from the model container when the process crashes.
DestinationS3Uri (string) --
The Amazon S3 bucket to send the core dump to.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the core dump data at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
// KMS Key ID "1234abcd-12ab-34cd-56ef-1234567890ab"
// Amazon Resource Name (ARN) of a KMS Key "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"
// KMS Key Alias "alias/ExampleAlias"
// Amazon Resource Name (ARN) of a KMS Key Alias "arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"
If you use a KMS key ID or an alias of your KMS key, the Amazon SageMaker execution role must include permissions to call kms:Encrypt. If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. Amazon SageMaker uses server-side encryption with KMS-managed keys for OutputDataConfig. If you use a bucket policy with an s3:PutObject permission that only allows objects with server-side encryption, set the condition key of s3:x-amz-server-side-encryption to "aws:kms". For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateEndpoint and UpdateEndpoint requests. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
ServerlessConfig (dict) --
The serverless configuration for an endpoint. Specifies a serverless endpoint configuration instead of an instance-based endpoint configuration.
MemorySizeInMB (integer) --
The memory size of your serverless endpoint. Valid values are in 1 GB increments: 1024 MB, 2048 MB, 3072 MB, 4096 MB, 5120 MB, or 6144 MB.
MaxConcurrency (integer) --
The maximum number of concurrent invocations your serverless endpoint can process.
DataCaptureConfig (dict) --
EnableCapture (boolean) --
InitialSamplingPercentage (integer) --
DestinationS3Uri (string) --
KmsKeyId (string) --
CaptureOptions (list) --
(dict) --
CaptureMode (string) --
CaptureContentTypeHeader (dict) --
CsvContentTypes (list) --
(string) --
JsonContentTypes (list) --
(string) --
KmsKeyId (string) --
Amazon Web Services KMS key ID Amazon SageMaker uses to encrypt data when storing it on the ML storage volume attached to the instance.
CreationTime (datetime) --
A timestamp that shows when the endpoint configuration was created.
AsyncInferenceConfig (dict) --
Returns the description of an endpoint configuration created using the CreateEndpointConfig API.
ClientConfig (dict) --
Configures the behavior of the client used by Amazon SageMaker to interact with the model container during asynchronous inference.
MaxConcurrentInvocationsPerInstance (integer) --
The maximum number of concurrent requests sent by the SageMaker client to the model container. If no value is provided, Amazon SageMaker will choose an optimal value for you.
OutputConfig (dict) --
Specifies the configuration for asynchronous inference invocation outputs.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the asynchronous inference output in Amazon S3.
S3OutputPath (string) --
The Amazon S3 location to upload inference responses to.
NotificationConfig (dict) --
Specifies the configuration for notifications of inference results for asynchronous inference.
SuccessTopic (string) --
Amazon SNS topic to post a notification to when inference completes successfully. If no topic is provided, no notification is sent on success.
ErrorTopic (string) --
Amazon SNS topic to post a notification to when inference fails. If no topic is provided, no notification is sent on failure.
{'Containers': {'InferenceSpecificationName': 'string'}, 'PrimaryContainer': {'InferenceSpecificationName': 'string'}}
Describes a model that you created using the CreateModel API.
See also: AWS API Documentation
Request Syntax
client.describe_model( ModelName='string' )
string
[REQUIRED]
The name of the model.
dict
Response Syntax
{ 'ModelName': 'string', 'PrimaryContainer': { 'ContainerHostname': 'string', 'Image': 'string', 'ImageConfig': { 'RepositoryAccessMode': 'Platform'|'Vpc', 'RepositoryAuthConfig': { 'RepositoryCredentialsProviderArn': 'string' } }, 'Mode': 'SingleModel'|'MultiModel', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' }, 'ModelPackageName': 'string', 'InferenceSpecificationName': 'string', 'MultiModelConfig': { 'ModelCacheSetting': 'Enabled'|'Disabled' } }, 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageConfig': { 'RepositoryAccessMode': 'Platform'|'Vpc', 'RepositoryAuthConfig': { 'RepositoryCredentialsProviderArn': 'string' } }, 'Mode': 'SingleModel'|'MultiModel', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' }, 'ModelPackageName': 'string', 'InferenceSpecificationName': 'string', 'MultiModelConfig': { 'ModelCacheSetting': 'Enabled'|'Disabled' } }, ], 'InferenceExecutionConfig': { 'Mode': 'Serial'|'Direct' }, 'ExecutionRoleArn': 'string', 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] }, 'CreationTime': datetime(2015, 1, 1), 'ModelArn': 'string', 'EnableNetworkIsolation': True|False }
Response Structure
(dict) --
ModelName (string) --
Name of the Amazon SageMaker model.
PrimaryContainer (dict) --
The location of the primary inference code, associated artifacts, and custom environment map that the inference code uses when it is deployed in production.
ContainerHostname (string) --
This parameter is ignored for models that contain only a PrimaryContainer.
When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.
Image (string) --
The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker
ImageConfig (dict) --
Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers
RepositoryAccessMode (string) --
Set this to one of the following values:
Platform - The model image is hosted in Amazon ECR.
Vpc - The model image is hosted in a private Docker registry in your VPC.
RepositoryAuthConfig (dict) --
(Optional) Specifies an authentication configuration for the private docker registry where your model image is hosted. Specify a value for this property only if you specified Vpc as the value for the RepositoryAccessMode field, and the private Docker registry where the model image is hosted requires authentication.
RepositoryCredentialsProviderArn (string) --
The Amazon Resource Name (ARN) of an Amazon Web Services Lambda function that provides credentials to authenticate to the private Docker registry where your model image is hosted. For information about how to create an Amazon Web Services Lambda function, see Create a Lambda function with the console in the Amazon Web Services Lambda Developer Guide.
Mode (string) --
Whether the container hosts a single model or multiple models.
ModelDataUrl (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.
If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelPackageName (string) --
The name or Amazon Resource Name (ARN) of the model package to use to create the model.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
MultiModelConfig (dict) --
Specifies additional configuration for multi-model endpoints.
ModelCacheSetting (string) --
Whether to cache models for a multi-model endpoint. By default, multi-model endpoints cache models so that a model does not have to be loaded into memory each time it is invoked. Some use cases do not benefit from model caching. For example, if an endpoint hosts a large number of models that are each invoked infrequently, the endpoint might perform better if you disable model caching. To disable model caching, set the value of this parameter to Disabled.
Containers (list) --
The containers in the inference pipeline.
(dict) --
Describes the container, as part of model definition.
ContainerHostname (string) --
This parameter is ignored for models that contain only a PrimaryContainer.
When a ContainerDefinition is part of an inference pipeline, the value of the parameter uniquely identifies the container for the purposes of logging and metrics. For information, see Use Logs and Metrics to Monitor an Inference Pipeline. If you don't specify a value for this parameter for a ContainerDefinition that is part of an inference pipeline, a unique name is automatically assigned based on the position of the ContainerDefinition in the pipeline. If you specify a value for the ContainerHostName for any ContainerDefinition that is part of an inference pipeline, you must specify a value for the ContainerHostName parameter of every ContainerDefinition in that pipeline.
Image (string) --
The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a Docker registry that is accessible from the same VPC that you configure for your endpoint. If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker
ImageConfig (dict) --
Specifies whether the model container is in Amazon ECR or a private Docker registry accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a private Docker registry, see Use a Private Docker Registry for Real-Time Inference Containers
RepositoryAccessMode (string) --
Set this to one of the following values:
Platform - The model image is hosted in Amazon ECR.
Vpc - The model image is hosted in a private Docker registry in your VPC.
RepositoryAuthConfig (dict) --
(Optional) Specifies an authentication configuration for the private docker registry where your model image is hosted. Specify a value for this property only if you specified Vpc as the value for the RepositoryAccessMode field, and the private Docker registry where the model image is hosted requires authentication.
RepositoryCredentialsProviderArn (string) --
The Amazon Resource Name (ARN) of an Amazon Web Services Lambda function that provides credentials to authenticate to the private Docker registry where your model image is hosted. For information about how to create an Amazon Web Services Lambda function, see Create a Lambda function with the console in the Amazon Web Services Lambda Developer Guide.
Mode (string) --
Whether the container hosts a single model or multiple models.
ModelDataUrl (string) --
The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). The S3 path is required for Amazon SageMaker built-in algorithms, but not if you use your own algorithms. For more information on built-in algorithms, see Common Parameters.
If you provide a value for this parameter, Amazon SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the S3 path you provide. Amazon Web Services STS is activated in your IAM user account by default. If you previously deactivated Amazon Web Services STS for a region, you need to reactivate Amazon Web Services STS for that region. For more information, see Activating and Deactivating Amazon Web Services STS in an Amazon Web Services Region in the Amazon Web Services Identity and Access Management User Guide.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelPackageName (string) --
The name or Amazon Resource Name (ARN) of the model package to use to create the model.
InferenceSpecificationName (string) --
The inference specification name in the model package version.
MultiModelConfig (dict) --
Specifies additional configuration for multi-model endpoints.
ModelCacheSetting (string) --
Whether to cache models for a multi-model endpoint. By default, multi-model endpoints cache models so that a model does not have to be loaded into memory each time it is invoked. Some use cases do not benefit from model caching. For example, if an endpoint hosts a large number of models that are each invoked infrequently, the endpoint might perform better if you disable model caching. To disable model caching, set the value of this parameter to Disabled.
InferenceExecutionConfig (dict) --
Specifies details of how containers in a multi-container endpoint are called.
Mode (string) --
How containers in a multi-container are run. The following values are valid.
SERIAL - Containers run as a serial pipeline.
DIRECT - Only the individual container that you specify is run.
ExecutionRoleArn (string) --
The Amazon Resource Name (ARN) of the IAM role that you specified for the model.
VpcConfig (dict) --
A VpcConfig object that specifies the VPC that this model has access to. For more information, see Protect Endpoints by Using an Amazon Virtual Private Cloud
SecurityGroupIds (list) --
The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) --
The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.
(string) --
CreationTime (datetime) --
A timestamp that shows when the model was created.
ModelArn (string) --
The Amazon Resource Name (ARN) of the model.
EnableNetworkIsolation (boolean) --
If True, no inbound or outbound network calls can be made to or from the model container.
{'AdditionalInferenceSpecifications': [{'Containers': [{'ContainerHostname': 'string', 'Environment': {'string': 'string'}, 'Framework': 'string', 'FrameworkVersion': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string', 'ProductId': 'string'}], 'Description': 'string', 'Name': 'string', 'SupportedContentTypes': ['string'], 'SupportedRealtimeInferenceInstanceTypes': ['ml.t2.medium ' '| ' 'ml.t2.large ' '| ' 'ml.t2.xlarge ' '| ' 'ml.t2.2xlarge ' '| ' 'ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.m5d.large ' '| ' 'ml.m5d.xlarge ' '| ' 'ml.m5d.2xlarge ' '| ' 'ml.m5d.4xlarge ' '| ' 'ml.m5d.12xlarge ' '| ' 'ml.m5d.24xlarge ' '| ' 'ml.c4.large ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.large ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.c5d.large ' '| ' 'ml.c5d.xlarge ' '| ' 'ml.c5d.2xlarge ' '| ' 'ml.c5d.4xlarge ' '| ' 'ml.c5d.9xlarge ' '| ' 'ml.c5d.18xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge ' '| ' 'ml.r5.large ' '| ' 'ml.r5.xlarge ' '| ' 'ml.r5.2xlarge ' '| ' 'ml.r5.4xlarge ' '| ' 'ml.r5.12xlarge ' '| ' 'ml.r5.24xlarge ' '| ' 'ml.r5d.large ' '| ' 'ml.r5d.xlarge ' '| ' 'ml.r5d.2xlarge ' '| ' 'ml.r5d.4xlarge ' '| ' 'ml.r5d.12xlarge ' '| ' 'ml.r5d.24xlarge ' '| ' 'ml.inf1.xlarge ' '| ' 'ml.inf1.2xlarge ' '| ' 'ml.inf1.6xlarge ' '| ' 'ml.inf1.24xlarge'], 'SupportedResponseMIMETypes': ['string'], 'SupportedTransformInstanceTypes': ['ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge']}], 'Domain': 'string', 'DriftCheckBaselines': {'Bias': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PostTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'Explainability': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelDataQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}, 'ModelMetrics': {'Bias': {'PostTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'SamplePayloadUrl': 'string', 'Task': 'string'}
Returns a description of the specified model package, which is used to create SageMaker models or list them on Amazon Web Services Marketplace.
To create models in SageMaker, buyers can subscribe to model packages listed on Amazon Web Services Marketplace.
See also: AWS API Documentation
Request Syntax
client.describe_model_package( ModelPackageName='string' )
string
[REQUIRED]
The name or Amazon Resource Name (ARN) of the model package to describe.
When you specify a name, the name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).
dict
Response Syntax
{ 'ModelPackageName': 'string', 'ModelPackageGroupName': 'string', 'ModelPackageVersion': 123, 'ModelPackageArn': 'string', 'ModelPackageDescription': 'string', 'CreationTime': datetime(2015, 1, 1), 'InferenceSpecification': { 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, 'SourceAlgorithmSpecification': { 'SourceAlgorithms': [ { 'ModelDataUrl': 'string', 'AlgorithmName': 'string' }, ] }, 'ValidationSpecification': { 'ValidationRole': 'string', 'ValidationProfiles': [ { 'ProfileName': 'string', 'TransformJobDefinition': { 'MaxConcurrentTransforms': 123, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' } } }, ] }, 'ModelPackageStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting', 'ModelPackageStatusDetails': { 'ValidationStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ], 'ImageScanStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ] }, 'CertifyForMarketplace': True|False, 'ModelApprovalStatus': 'Approved'|'Rejected'|'PendingManualApproval', 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'ModelMetrics': { 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Bias': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'ApprovalDescription': 'string', 'CustomerMetadataProperties': { 'string': 'string' }, 'DriftCheckBaselines': { 'Bias': { 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } }, 'Domain': 'string', 'Task': 'string', 'SamplePayloadUrl': 'string', 'AdditionalInferenceSpecifications': [ { 'Name': 'string', 'Description': 'string', 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ] }
Response Structure
(dict) --
ModelPackageName (string) --
The name of the model package being described.
ModelPackageGroupName (string) --
If the model is a versioned model, the name of the model group that the versioned model belongs to.
ModelPackageVersion (integer) --
The version of the model package.
ModelPackageArn (string) --
The Amazon Resource Name (ARN) of the model package.
ModelPackageDescription (string) --
A brief summary of the model package.
CreationTime (datetime) --
A timestamp specifying when the model package was created.
InferenceSpecification (dict) --
Details about inference jobs that can be run with models based on this model package.
Containers (list) --
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) --
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) --
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
This parameter is required for unversioned models, and optional for versioned models.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
SourceAlgorithmSpecification (dict) --
Details about the algorithm that was used to create the model package.
SourceAlgorithms (list) --
A list of the algorithms that were used to create a model package.
(dict) --
Specifies an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your Amazon SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
AlgorithmName (string) --
The name of an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your Amazon SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.
ValidationSpecification (dict) --
Configurations for one or more transform jobs that SageMaker runs to test the model package.
ValidationRole (string) --
The IAM roles to be used for the validation of the model package.
ValidationProfiles (list) --
An array of ModelPackageValidationProfile objects, each of which specifies a batch transform job that Amazon SageMaker runs to validate your model package.
(dict) --
Contains data, such as the inputs and targeted instance types that are used in the process of validating the model package.
The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.
ProfileName (string) --
The name of the profile for the model package.
TransformJobDefinition (dict) --
The TransformJobDefinition object that describes the transform job used for the validation of the model package.
MaxConcurrentTransforms (integer) --
The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.
MaxPayloadInMB (integer) --
The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).
BatchStrategy (string) --
A string that determines the number of records included in a single mini-batch.
SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.
Environment (dict) --
The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.
(string) --
(string) --
TransformInput (dict) --
A description of the input source and the way the transform job consumes it.
DataSource (dict) --
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
S3DataSource (dict) --
The S3 location of the data source that is associated with a channel.
S3DataType (string) --
If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.
If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.
The following values are compatible: ManifestFile, S3Prefix
The following value is not compatible: AugmentedManifestFile
S3Uri (string) --
Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:
A key name prefix might look like this: s3://bucketname/exampleprefix.
A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.
ContentType (string) --
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
CompressionType (string) --
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.
SplitType (string) --
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.
TransformOutput (dict) --
Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.
S3OutputPath (string) --
The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.
For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.
Accept (string) --
The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.
AssembleWith (string) --
Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.
KmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.
The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.
TransformResources (dict) --
Identifies the ML compute instances for the transform job.
InstanceType (string) --
The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.
InstanceCount (integer) --
The number of ML compute instances to use in the transform job. For distributed transform jobs, specify a value greater than 1. The default value is 1.
VolumeKmsKeyId (string) --
The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.
The VolumeKmsKeyId can be any of the following formats:
Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
Alias name: alias/ExampleAlias
Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
ModelPackageStatus (string) --
The current status of the model package.
ModelPackageStatusDetails (dict) --
Details about the current status of the model package.
ValidationStatuses (list) --
The validation status of the model package.
(dict) --
Represents the overall status of a model package.
Name (string) --
The name of the model package for which the overall status is being reported.
Status (string) --
The current status.
FailureReason (string) --
if the overall status is Failed, the reason for the failure.
ImageScanStatuses (list) --
The status of the scan of the Docker image container for the model package.
(dict) --
Represents the overall status of a model package.
Name (string) --
The name of the model package for which the overall status is being reported.
Status (string) --
The current status.
FailureReason (string) --
if the overall status is Failed, the reason for the failure.
CertifyForMarketplace (boolean) --
Whether the model package is certified for listing on Amazon Web Services Marketplace.
ModelApprovalStatus (string) --
The approval status of the model package.
CreatedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
MetadataProperties (dict) --
Metadata properties of the tracking entity, trial, or trial component.
CommitId (string) --
The commit ID.
Repository (string) --
The repository.
GeneratedBy (string) --
The entity this entity was generated by.
ProjectId (string) --
The project ID.
ModelMetrics (dict) --
Metrics for the model.
ModelQuality (dict) --
Metrics that measure the quality of a model.
Statistics (dict) --
Model quality statistics.
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Constraints (dict) --
Model quality constraints.
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
ModelDataQuality (dict) --
Metrics that measure the quality of the input data for a model.
Statistics (dict) --
Data quality statistics for a model.
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Constraints (dict) --
Data quality constraints for a model.
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Bias (dict) --
Metrics that measure bais in a model.
Report (dict) --
The bias report for a model
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
PreTrainingReport (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
PostTrainingReport (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Explainability (dict) --
Metrics that help explain a model.
Report (dict) --
The explainability report for a model.
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
LastModifiedTime (datetime) --
The last time the model package was modified.
LastModifiedBy (dict) --
Information about the user who created or modified an experiment, trial, trial component, lineage group, or project.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
ApprovalDescription (string) --
A description provided for the model approval.
CustomerMetadataProperties (dict) --
The metadata properties associated with the model package versions.
(string) --
(string) --
DriftCheckBaselines (dict) --
Represents the drift check baselines that can be used when the model monitor is set using the model package. For more information, see the topic on Drift Detection against Previous Baselines in SageMaker Pipelines in the Amazon SageMaker Developer Guide.
Bias (dict) --
Represents the drift check bias baselines that can be used when the model monitor is set using the model package.
ConfigFile (dict) --
The bias config file for a model.
ContentType (string) --
The type of content stored in the file source.
ContentDigest (string) --
The digest of the file source.
S3Uri (string) --
The Amazon S3 URI for the file source.
PreTrainingConstraints (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
PostTrainingConstraints (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Explainability (dict) --
Represents the drift check explainability baselines that can be used when the model monitor is set using the model package.
Constraints (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
ConfigFile (dict) --
The explainability config file for the model.
ContentType (string) --
The type of content stored in the file source.
ContentDigest (string) --
The digest of the file source.
S3Uri (string) --
The Amazon S3 URI for the file source.
ModelQuality (dict) --
Represents the drift check model quality baselines that can be used when the model monitor is set using the model package.
Statistics (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Constraints (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
ModelDataQuality (dict) --
Represents the drift check model data quality baselines that can be used when the model monitor is set using the model package.
Statistics (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Constraints (dict) --
ContentType (string) --
ContentDigest (string) --
S3Uri (string) --
Domain (string) --
The machine learning domain of the model package you specified. Common machine learning domains include computer vision and natural language processing.
Task (string) --
The machine learning task you specified that your model package accomplishes. Common machine learning tasks include object detection and image classification.
SamplePayloadUrl (string) --
The Amazon Simple Storage Service (Amazon S3) path where the sample payload are stored. This path points to a single gzip compressed tar archive (.tar.gz suffix).
AdditionalInferenceSpecifications (list) --
An array of additional Inference Specification objects. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.
(dict) --
A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package
Name (string) --
A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.
Description (string) --
A description of the additional Inference specification
Containers (list) --
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) --
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) --
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
{'LineageGroupArn': 'string'}
Provides a list of a trials component's properties.
See also: AWS API Documentation
Request Syntax
client.describe_trial_component( TrialComponentName='string' )
string
[REQUIRED]
The name of the trial component to describe.
dict
Response Syntax
{ 'TrialComponentName': 'string', 'TrialComponentArn': 'string', 'DisplayName': 'string', 'Source': { 'SourceArn': 'string', 'SourceType': 'string' }, 'Status': { 'PrimaryStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'Message': 'string' }, 'StartTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'Parameters': { 'string': { 'StringValue': 'string', 'NumberValue': 123.0 } }, 'InputArtifacts': { 'string': { 'MediaType': 'string', 'Value': 'string' } }, 'OutputArtifacts': { 'string': { 'MediaType': 'string', 'Value': 'string' } }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'Metrics': [ { 'MetricName': 'string', 'SourceArn': 'string', 'TimeStamp': datetime(2015, 1, 1), 'Max': 123.0, 'Min': 123.0, 'Last': 123.0, 'Count': 123, 'Avg': 123.0, 'StdDev': 123.0 }, ], 'LineageGroupArn': 'string' }
Response Structure
(dict) --
TrialComponentName (string) --
The name of the trial component.
TrialComponentArn (string) --
The Amazon Resource Name (ARN) of the trial component.
DisplayName (string) --
The name of the component as displayed. If DisplayName isn't specified, TrialComponentName is displayed.
Source (dict) --
The Amazon Resource Name (ARN) of the source and, optionally, the job type.
SourceArn (string) --
The source ARN.
SourceType (string) --
The source job type.
Status (dict) --
The status of the component. States include:
InProgress
Completed
Failed
PrimaryStatus (string) --
The status of the trial component.
Message (string) --
If the component failed, a message describing why.
StartTime (datetime) --
When the component started.
EndTime (datetime) --
When the component ended.
CreationTime (datetime) --
When the component was created.
CreatedBy (dict) --
Who created the trial component.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
LastModifiedTime (datetime) --
When the component was last modified.
LastModifiedBy (dict) --
Who last modified the component.
UserProfileArn (string) --
The Amazon Resource Name (ARN) of the user's profile.
UserProfileName (string) --
The name of the user's profile.
DomainId (string) --
The domain associated with the user.
Parameters (dict) --
The hyperparameters of the component.
(string) --
(dict) --
The value of a hyperparameter. Only one of NumberValue or StringValue can be specified.
This object is specified in the CreateTrialComponent request.
StringValue (string) --
The string value of a categorical hyperparameter. If you specify a value for this parameter, you can't specify the NumberValue parameter.
NumberValue (float) --
The numeric value of a numeric hyperparameter. If you specify a value for this parameter, you can't specify the StringValue parameter.
InputArtifacts (dict) --
The input artifacts of the component.
(string) --
(dict) --
Represents an input or output artifact of a trial component. You specify TrialComponentArtifact as part of the InputArtifacts and OutputArtifacts parameters in the CreateTrialComponent request.
Examples of input artifacts are datasets, algorithms, hyperparameters, source code, and instance types. Examples of output artifacts are metrics, snapshots, logs, and images.
MediaType (string) --
The media type of the artifact, which indicates the type of data in the artifact file. The media type consists of a type and a subtype concatenated with a slash (/) character, for example, text/csv, image/jpeg, and s3/uri. The type specifies the category of the media. The subtype specifies the kind of data.
Value (string) --
The location of the artifact.
OutputArtifacts (dict) --
The output artifacts of the component.
(string) --
(dict) --
Represents an input or output artifact of a trial component. You specify TrialComponentArtifact as part of the InputArtifacts and OutputArtifacts parameters in the CreateTrialComponent request.
Examples of input artifacts are datasets, algorithms, hyperparameters, source code, and instance types. Examples of output artifacts are metrics, snapshots, logs, and images.
MediaType (string) --
The media type of the artifact, which indicates the type of data in the artifact file. The media type consists of a type and a subtype concatenated with a slash (/) character, for example, text/csv, image/jpeg, and s3/uri. The type specifies the category of the media. The subtype specifies the kind of data.
Value (string) --
The location of the artifact.
MetadataProperties (dict) --
Metadata properties of the tracking entity, trial, or trial component.
CommitId (string) --
The commit ID.
Repository (string) --
The repository.
GeneratedBy (string) --
The entity this entity was generated by.
ProjectId (string) --
The project ID.
Metrics (list) --
The metrics for the component.
(dict) --
A summary of the metrics of a trial component.
MetricName (string) --
The name of the metric.
SourceArn (string) --
The Amazon Resource Name (ARN) of the source.
TimeStamp (datetime) --
When the metric was last updated.
Max (float) --
The maximum value of the metric.
Min (float) --
The minimum value of the metric.
Last (float) --
The most recent value of the metric.
Count (integer) --
The number of samples used to generate the metric.
Avg (float) --
The average value of the metric.
StdDev (float) --
The standard deviation of the metric.
LineageGroupArn (string) --
The Amazon Resource Name (ARN) of the lineage group.
{'PipelineExecutionSteps': {'Metadata': {'ClarifyCheck': {'BaselineUsedForDriftCheckConstraints': 'string', 'CalculatedBaselineConstraints': 'string', 'CheckJobArn': 'string', 'CheckType': 'string', 'ModelPackageGroupName': 'string', 'RegisterNewBaseline': 'boolean', 'SkipCheck': 'boolean', 'ViolationReport': 'string'}, 'QualityCheck': {'BaselineUsedForDriftCheckConstraints': 'string', 'BaselineUsedForDriftCheckStatistics': 'string', 'CalculatedBaselineConstraints': 'string', 'CalculatedBaselineStatistics': 'string', 'CheckJobArn': 'string', 'CheckType': 'string', 'ModelPackageGroupName': 'string', 'RegisterNewBaseline': 'boolean', 'SkipCheck': 'boolean', 'ViolationReport': 'string'}}}}
Gets a list of PipeLineExecutionStep objects.
See also: AWS API Documentation
Request Syntax
client.list_pipeline_execution_steps( PipelineExecutionArn='string', NextToken='string', MaxResults=123, SortOrder='Ascending'|'Descending' )
string
The Amazon Resource Name (ARN) of the pipeline execution.
string
If the result of the previous ListPipelineExecutionSteps request was truncated, the response includes a NextToken. To retrieve the next set of pipeline execution steps, use the token in the next request.
integer
The maximum number of pipeline execution steps to return in the response.
string
The field by which to sort results. The default is CreatedTime.
dict
Response Syntax
{ 'PipelineExecutionSteps': [ { 'StepName': 'string', 'StartTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'StepStatus': 'Starting'|'Executing'|'Stopping'|'Stopped'|'Failed'|'Succeeded', 'CacheHitResult': { 'SourcePipelineExecutionArn': 'string' }, 'FailureReason': 'string', 'Metadata': { 'TrainingJob': { 'Arn': 'string' }, 'ProcessingJob': { 'Arn': 'string' }, 'TransformJob': { 'Arn': 'string' }, 'TuningJob': { 'Arn': 'string' }, 'Model': { 'Arn': 'string' }, 'RegisterModel': { 'Arn': 'string' }, 'Condition': { 'Outcome': 'True'|'False' }, 'Callback': { 'CallbackToken': 'string', 'SqsQueueUrl': 'string', 'OutputParameters': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Lambda': { 'Arn': 'string', 'OutputParameters': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'QualityCheck': { 'CheckType': 'string', 'BaselineUsedForDriftCheckStatistics': 'string', 'BaselineUsedForDriftCheckConstraints': 'string', 'CalculatedBaselineStatistics': 'string', 'CalculatedBaselineConstraints': 'string', 'ModelPackageGroupName': 'string', 'ViolationReport': 'string', 'CheckJobArn': 'string', 'SkipCheck': True|False, 'RegisterNewBaseline': True|False }, 'ClarifyCheck': { 'CheckType': 'string', 'BaselineUsedForDriftCheckConstraints': 'string', 'CalculatedBaselineConstraints': 'string', 'ModelPackageGroupName': 'string', 'ViolationReport': 'string', 'CheckJobArn': 'string', 'SkipCheck': True|False, 'RegisterNewBaseline': True|False } } }, ], 'NextToken': 'string' }
Response Structure
(dict) --
PipelineExecutionSteps (list) --
A list of PipeLineExecutionStep objects. Each PipeLineExecutionStep consists of StepName, StartTime, EndTime, StepStatus, and Metadata. Metadata is an object with properties for each job that contains relevant information about the job created by the step.
(dict) --
An execution of a step in a pipeline.
StepName (string) --
The name of the step that is executed.
StartTime (datetime) --
The time that the step started executing.
EndTime (datetime) --
The time that the step stopped executing.
StepStatus (string) --
The status of the step execution.
CacheHitResult (dict) --
If this pipeline execution step was cached, details on the cache hit.
SourcePipelineExecutionArn (string) --
The Amazon Resource Name (ARN) of the pipeline execution.
FailureReason (string) --
The reason why the step failed execution. This is only returned if the step failed its execution.
Metadata (dict) --
Metadata for the step execution.
TrainingJob (dict) --
The Amazon Resource Name (ARN) of the training job that was run by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the training job that was run by this step execution.
ProcessingJob (dict) --
The Amazon Resource Name (ARN) of the processing job that was run by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the processing job.
TransformJob (dict) --
The Amazon Resource Name (ARN) of the transform job that was run by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the transform job that was run by this step execution.
TuningJob (dict) --
The Amazon Resource Name (ARN) of the tuning job that was run by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the tuning job that was run by this step execution.
Model (dict) --
The Amazon Resource Name (ARN) of the model that was created by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the created model.
RegisterModel (dict) --
The Amazon Resource Name (ARN) of the model package the model was registered to by this step execution.
Arn (string) --
The Amazon Resource Name (ARN) of the model package.
Condition (dict) --
The outcome of the condition evaluation that was run by this step execution.
Outcome (string) --
The outcome of the Condition step evaluation.
Callback (dict) --
The URL of the Amazon SQS queue used by this step execution, the pipeline generated token, and a list of output parameters.
CallbackToken (string) --
The pipeline generated token from the Amazon SQS queue.
SqsQueueUrl (string) --
The URL of the Amazon Simple Queue Service (Amazon SQS) queue used by the callback step.
OutputParameters (list) --
A list of the output parameters of the callback step.
(dict) --
An output parameter of a pipeline step.
Name (string) --
The name of the output parameter.
Value (string) --
The value of the output parameter.
Lambda (dict) --
The Amazon Resource Name (ARN) of the Lambda function that was run by this step execution and a list of output parameters.
Arn (string) --
The Amazon Resource Name (ARN) of the Lambda function that was run by this step execution.
OutputParameters (list) --
A list of the output parameters of the Lambda step.
(dict) --
An output parameter of a pipeline step.
Name (string) --
The name of the output parameter.
Value (string) --
The value of the output parameter.
QualityCheck (dict) --
The configurations and outcomes of the check step execution. This includes:
The type of the check conducted,
The Amazon S3 URIs of baseline constraints and statistics files to be used for the drift check.
The Amazon S3 URIs of newly calculated baseline constraints and statistics.
The model package group name provided.
The Amazon S3 URI of the violation report if violations detected.
The Amazon Resource Name (ARN) of check processing job initiated by the step execution.
The boolean flags indicating if the drift check is skipped.
If step property BaselineUsedForDriftCheck is set the same as CalculatedBaseline.
CheckType (string) --
The type of the Quality check step.
BaselineUsedForDriftCheckStatistics (string) --
The Amazon S3 URI of the baseline statistics file used for the drift check.
BaselineUsedForDriftCheckConstraints (string) --
The Amazon S3 URI of the baseline constraints file used for the drift check.
CalculatedBaselineStatistics (string) --
The Amazon S3 URI of the newly calculated baseline statistics file.
CalculatedBaselineConstraints (string) --
The Amazon S3 URI of the newly calculated baseline constraints file.
ModelPackageGroupName (string) --
The model package group name.
ViolationReport (string) --
The Amazon S3 URI of violation report if violations are detected.
CheckJobArn (string) --
The Amazon Resource Name (ARN) of the Quality check processing job that was run by this step execution.
SkipCheck (boolean) --
This flag indicates if the drift check against the previous baseline will be skipped or not. If it is set to False, the previous baseline of the configured check type must be available.
RegisterNewBaseline (boolean) --
This flag indicates if a newly calculated baseline can be accessed through step properties BaselineUsedForDriftCheckConstraints and BaselineUsedForDriftCheckStatistics. If it is set to False, the previous baseline of the configured check type must also be available. These can be accessed through the BaselineUsedForDriftCheckConstraints and BaselineUsedForDriftCheckStatistics properties.
ClarifyCheck (dict) --
Container for the metadata for a Clarify check step. The configurations and outcomes of the check step execution. This includes:
The type of the check conducted,
The Amazon S3 URIs of baseline constraints and statistics files to be used for the drift check.
The Amazon S3 URIs of newly calculated baseline constraints and statistics.
The model package group name provided.
The Amazon S3 URI of the violation report if violations detected.
The Amazon Resource Name (ARN) of check processing job initiated by the step execution.
The boolean flags indicating if the drift check is skipped.
If step property BaselineUsedForDriftCheck is set the same as CalculatedBaseline.
CheckType (string) --
The type of the Clarify Check step
BaselineUsedForDriftCheckConstraints (string) --
The Amazon S3 URI of baseline constraints file to be used for the drift check.
CalculatedBaselineConstraints (string) --
The Amazon S3 URI of the newly calculated baseline constraints file.
ModelPackageGroupName (string) --
The model package group name.
ViolationReport (string) --
The Amazon S3 URI of the violation report if violations are detected.
CheckJobArn (string) --
The Amazon Resource Name (ARN) of the check processing job that was run by this step's execution.
SkipCheck (boolean) --
This flag indicates if the drift check against the previous baseline will be skipped or not. If it is set to False, the previous baseline of the configured check type must be available.
RegisterNewBaseline (boolean) --
This flag indicates if a newly calculated baseline can be accessed through step properties BaselineUsedForDriftCheckConstraints and BaselineUsedForDriftCheckStatistics. If it is set to False, the previous baseline of the configured check type must also be available. These can be accessed through the BaselineUsedForDriftCheckConstraints property.
NextToken (string) --
If the result of the previous ListPipelineExecutionSteps request was truncated, the response includes a NextToken. To retrieve the next set of pipeline execution steps, use the token in the next request.
{'Results': {'Endpoint': {'ProductionVariants': {'CurrentServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}, 'DesiredServerlessConfig': {'MaxConcurrency': 'integer', 'MemorySizeInMB': 'integer'}}}, 'ModelPackage': {'AdditionalInferenceSpecifications': [{'Containers': [{'ContainerHostname': 'string', 'Environment': {'string': 'string'}, 'Framework': 'string', 'FrameworkVersion': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string', 'ProductId': 'string'}], 'Description': 'string', 'Name': 'string', 'SupportedContentTypes': ['string'], 'SupportedRealtimeInferenceInstanceTypes': ['ml.t2.medium ' '| ' 'ml.t2.large ' '| ' 'ml.t2.xlarge ' '| ' 'ml.t2.2xlarge ' '| ' 'ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.m5d.large ' '| ' 'ml.m5d.xlarge ' '| ' 'ml.m5d.2xlarge ' '| ' 'ml.m5d.4xlarge ' '| ' 'ml.m5d.12xlarge ' '| ' 'ml.m5d.24xlarge ' '| ' 'ml.c4.large ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.large ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.c5d.large ' '| ' 'ml.c5d.xlarge ' '| ' 'ml.c5d.2xlarge ' '| ' 'ml.c5d.4xlarge ' '| ' 'ml.c5d.9xlarge ' '| ' 'ml.c5d.18xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge ' '| ' 'ml.r5.large ' '| ' 'ml.r5.xlarge ' '| ' 'ml.r5.2xlarge ' '| ' 'ml.r5.4xlarge ' '| ' 'ml.r5.12xlarge ' '| ' 'ml.r5.24xlarge ' '| ' 'ml.r5d.large ' '| ' 'ml.r5d.xlarge ' '| ' 'ml.r5d.2xlarge ' '| ' 'ml.r5d.4xlarge ' '| ' 'ml.r5d.12xlarge ' '| ' 'ml.r5d.24xlarge ' '| ' 'ml.inf1.xlarge ' '| ' 'ml.inf1.2xlarge ' '| ' 'ml.inf1.6xlarge ' '| ' 'ml.inf1.24xlarge'], 'SupportedResponseMIMETypes': ['string'], 'SupportedTransformInstanceTypes': ['ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge']}], 'Domain': 'string', 'DriftCheckBaselines': {'Bias': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PostTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingConstraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'Explainability': {'ConfigFile': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelDataQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}, 'ModelQuality': {'Constraints': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'Statistics': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'InferenceSpecification': {'Containers': {'Framework': 'string', 'FrameworkVersion': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string'}}, 'ModelMetrics': {'Bias': {'PostTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}, 'PreTrainingReport': {'ContentDigest': 'string', 'ContentType': 'string', 'S3Uri': 'string'}}}, 'SamplePayloadUrl': 'string', 'Task': 'string'}, 'TrialComponent': {'LineageGroupArn': 'string'}}}
Finds Amazon SageMaker resources that match a search query. Matching resources are returned as a list of SearchRecord objects in the response. You can sort the search results by any resource property in a ascending or descending order.
You can query against the following value types: numeric, text, Boolean, and timestamp.
See also: AWS API Documentation
Request Syntax
client.search( Resource='TrainingJob'|'Experiment'|'ExperimentTrial'|'ExperimentTrialComponent'|'Endpoint'|'ModelPackage'|'ModelPackageGroup'|'Pipeline'|'PipelineExecution'|'FeatureGroup'|'Project', SearchExpression={ 'Filters': [ { 'Name': 'string', 'Operator': 'Equals'|'NotEquals'|'GreaterThan'|'GreaterThanOrEqualTo'|'LessThan'|'LessThanOrEqualTo'|'Contains'|'Exists'|'NotExists'|'In', 'Value': 'string' }, ], 'NestedFilters': [ { 'NestedPropertyName': 'string', 'Filters': [ { 'Name': 'string', 'Operator': 'Equals'|'NotEquals'|'GreaterThan'|'GreaterThanOrEqualTo'|'LessThan'|'LessThanOrEqualTo'|'Contains'|'Exists'|'NotExists'|'In', 'Value': 'string' }, ] }, ], 'SubExpressions': [ {'... recursive ...'}, ], 'Operator': 'And'|'Or' }, SortBy='string', SortOrder='Ascending'|'Descending', NextToken='string', MaxResults=123 )
string
[REQUIRED]
The name of the Amazon SageMaker resource to search for.
dict
A Boolean conditional statement. Resources must satisfy this condition to be included in search results. You must provide at least one subexpression, filter, or nested filter. The maximum number of recursive SubExpressions, NestedFilters, and Filters that can be included in a SearchExpression object is 50.
Filters (list) --
A list of filter objects.
(dict) --
A conditional statement for a search expression that includes a resource property, a Boolean operator, and a value. Resources that match the statement are returned in the results from the Search API.
If you specify a Value, but not an Operator, Amazon SageMaker uses the equals operator.
In search, there are several property types:
Metrics
To define a metric filter, enter a value using the form "Metrics.<name>", where <name> is a metric name. For example, the following filter searches for training jobs with an "accuracy" metric greater than "0.9":
{
"Name": "Metrics.accuracy",
"Operator": "GreaterThan",
"Value": "0.9"
}
HyperParameters
To define a hyperparameter filter, enter a value with the form "HyperParameters.<name>". Decimal hyperparameter values are treated as a decimal in a comparison if the specified Value is also a decimal value. If the specified Value is an integer, the decimal hyperparameter values are treated as integers. For example, the following filter is satisfied by training jobs with a "learning_rate" hyperparameter that is less than "0.5":
{
"Name": "HyperParameters.learning_rate",
"Operator": "LessThan",
"Value": "0.5"
}
Tags
To define a tag filter, enter a value with the form Tags.<key>.
Name (string) -- [REQUIRED]
A resource property name. For example, TrainingJobName. For valid property names, see SearchRecord. You must specify a valid property for the resource.
Operator (string) --
A Boolean binary operator that is used to evaluate the filter. The operator field contains one of the following values:
Equals
The value of Name equals Value.
NotEquals
The value of Name doesn't equal Value.
Exists
The Name property exists.
NotExists
The Name property does not exist.
GreaterThan
The value of Name is greater than Value. Not supported for text properties.
GreaterThanOrEqualTo
The value of Name is greater than or equal to Value. Not supported for text properties.
LessThan
The value of Name is less than Value. Not supported for text properties.
LessThanOrEqualTo
The value of Name is less than or equal to Value. Not supported for text properties.
In
The value of Name is one of the comma delimited strings in Value. Only supported for text properties.
Contains
The value of Name contains the string Value. Only supported for text properties.
A SearchExpression can include the Contains operator multiple times when the value of Name is one of the following:
Experiment.DisplayName
Experiment.ExperimentName
Experiment.Tags
Trial.DisplayName
Trial.TrialName
Trial.Tags
TrialComponent.DisplayName
TrialComponent.TrialComponentName
TrialComponent.Tags
TrialComponent.InputArtifacts
TrialComponent.OutputArtifacts
A SearchExpression can include only one Contains operator for all other values of Name. In these cases, if you include multiple Contains operators in the SearchExpression, the result is the following error message: " 'CONTAINS' operator usage limit of 1 exceeded."
Value (string) --
A value used with Name and Operator to determine which resources satisfy the filter's condition. For numerical properties, Value must be an integer or floating-point decimal. For timestamp properties, Value must be an ISO 8601 date-time string of the following format: YYYY-mm-dd'T'HH:MM:SS.
NestedFilters (list) --
A list of nested filter objects.
(dict) --
A list of nested Filter objects. A resource must satisfy the conditions of all filters to be included in the results returned from the Search API.
For example, to filter on a training job's InputDataConfig property with a specific channel name and S3Uri prefix, define the following filters:
'{Name:"InputDataConfig.ChannelName", "Operator":"Equals", "Value":"train"}',
'{Name:"InputDataConfig.DataSource.S3DataSource.S3Uri", "Operator":"Contains", "Value":"mybucket/catdata"}'
NestedPropertyName (string) -- [REQUIRED]
The name of the property to use in the nested filters. The value must match a listed property name, such as InputDataConfig.
Filters (list) -- [REQUIRED]
A list of filters. Each filter acts on a property. Filters must contain at least one Filters value. For example, a NestedFilters call might include a filter on the PropertyName parameter of the InputDataConfig property: InputDataConfig.DataSource.S3DataSource.S3Uri.
(dict) --
A conditional statement for a search expression that includes a resource property, a Boolean operator, and a value. Resources that match the statement are returned in the results from the Search API.
If you specify a Value, but not an Operator, Amazon SageMaker uses the equals operator.
In search, there are several property types:
Metrics
To define a metric filter, enter a value using the form "Metrics.<name>", where <name> is a metric name. For example, the following filter searches for training jobs with an "accuracy" metric greater than "0.9":
{
"Name": "Metrics.accuracy",
"Operator": "GreaterThan",
"Value": "0.9"
}
HyperParameters
To define a hyperparameter filter, enter a value with the form "HyperParameters.<name>". Decimal hyperparameter values are treated as a decimal in a comparison if the specified Value is also a decimal value. If the specified Value is an integer, the decimal hyperparameter values are treated as integers. For example, the following filter is satisfied by training jobs with a "learning_rate" hyperparameter that is less than "0.5":
{
"Name": "HyperParameters.learning_rate",
"Operator": "LessThan",
"Value": "0.5"
}
Tags
To define a tag filter, enter a value with the form Tags.<key>.
Name (string) -- [REQUIRED]
A resource property name. For example, TrainingJobName. For valid property names, see SearchRecord. You must specify a valid property for the resource.
Operator (string) --
A Boolean binary operator that is used to evaluate the filter. The operator field contains one of the following values:
Equals
The value of Name equals Value.
NotEquals
The value of Name doesn't equal Value.
Exists
The Name property exists.
NotExists
The Name property does not exist.
GreaterThan
The value of Name is greater than Value. Not supported for text properties.
GreaterThanOrEqualTo
The value of Name is greater than or equal to Value. Not supported for text properties.
LessThan
The value of Name is less than Value. Not supported for text properties.
LessThanOrEqualTo
The value of Name is less than or equal to Value. Not supported for text properties.
In
The value of Name is one of the comma delimited strings in Value. Only supported for text properties.
Contains
The value of Name contains the string Value. Only supported for text properties.
A SearchExpression can include the Contains operator multiple times when the value of Name is one of the following:
Experiment.DisplayName
Experiment.ExperimentName
Experiment.Tags
Trial.DisplayName
Trial.TrialName
Trial.Tags
TrialComponent.DisplayName
TrialComponent.TrialComponentName
TrialComponent.Tags
TrialComponent.InputArtifacts
TrialComponent.OutputArtifacts
A SearchExpression can include only one Contains operator for all other values of Name. In these cases, if you include multiple Contains operators in the SearchExpression, the result is the following error message: " 'CONTAINS' operator usage limit of 1 exceeded."
Value (string) --
A value used with Name and Operator to determine which resources satisfy the filter's condition. For numerical properties, Value must be an integer or floating-point decimal. For timestamp properties, Value must be an ISO 8601 date-time string of the following format: YYYY-mm-dd'T'HH:MM:SS.
SubExpressions (list) --
A list of search expression objects.
(dict) --
A multi-expression that searches for the specified resource or resources in a search. All resource objects that satisfy the expression's condition are included in the search results. You must specify at least one subexpression, filter, or nested filter. A SearchExpression can contain up to twenty elements.
A SearchExpression contains the following components:
A list of Filter objects. Each filter defines a simple Boolean expression comprised of a resource property name, Boolean operator, and value.
A list of NestedFilter objects. Each nested filter defines a list of Boolean expressions using a list of resource properties. A nested filter is satisfied if a single object in the list satisfies all Boolean expressions.
A list of SearchExpression objects. A search expression object can be nested in a list of search expression objects.
A Boolean operator: And or Or.
Operator (string) --
A Boolean operator used to evaluate the search expression. If you want every conditional statement in all lists to be satisfied for the entire search expression to be true, specify And. If only a single conditional statement needs to be true for the entire search expression to be true, specify Or. The default value is And.
string
The name of the resource property used to sort the SearchResults. The default is LastModifiedTime.
string
How SearchResults are ordered. Valid values are Ascending or Descending. The default is Descending.
string
If more than MaxResults resources match the specified SearchExpression, the response includes a NextToken. The NextToken can be passed to the next SearchRequest to continue retrieving results.
integer
The maximum number of results to return.
dict
Response Syntax
{ 'Results': [ { 'TrainingJob': { 'TrainingJobName': 'string', 'TrainingJobArn': 'string', 'TuningJobArn': 'string', 'LabelingJobArn': 'string', 'AutoMLJobArn': 'string', 'ModelArtifacts': { 'S3ModelArtifacts': 'string' }, 'TrainingJobStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'SecondaryStatus': 'Starting'|'LaunchingMLInstances'|'PreparingTrainingStack'|'Downloading'|'DownloadingTrainingImage'|'Training'|'Uploading'|'Stopping'|'Stopped'|'MaxRuntimeExceeded'|'Completed'|'Failed'|'Interrupted'|'MaxWaitTimeExceeded'|'Updating'|'Restarting', 'FailureReason': 'string', 'HyperParameters': { 'string': 'string' }, 'AlgorithmSpecification': { 'TrainingImage': 'string', 'AlgorithmName': 'string', 'TrainingInputMode': 'Pipe'|'File'|'FastFile', 'MetricDefinitions': [ { 'Name': 'string', 'Regex': 'string' }, ], 'EnableSageMakerMetricsTimeSeries': True|False }, 'RoleArn': 'string', 'InputDataConfig': [ { 'ChannelName': 'string', 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'AttributeNames': [ 'string', ] }, 'FileSystemDataSource': { 'FileSystemId': 'string', 'FileSystemAccessMode': 'rw'|'ro', 'FileSystemType': 'EFS'|'FSxLustre', 'DirectoryPath': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'RecordWrapperType': 'None'|'RecordIO', 'InputMode': 'Pipe'|'File'|'FastFile', 'ShuffleConfig': { 'Seed': 123 } }, ], 'OutputDataConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, 'ResourceConfig': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', 'InstanceCount': 123, 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' }, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 }, 'CreationTime': datetime(2015, 1, 1), 'TrainingStartTime': datetime(2015, 1, 1), 'TrainingEndTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'SecondaryStatusTransitions': [ { 'Status': 'Starting'|'LaunchingMLInstances'|'PreparingTrainingStack'|'Downloading'|'DownloadingTrainingImage'|'Training'|'Uploading'|'Stopping'|'Stopped'|'MaxRuntimeExceeded'|'Completed'|'Failed'|'Interrupted'|'MaxWaitTimeExceeded'|'Updating'|'Restarting', 'StartTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'StatusMessage': 'string' }, ], 'FinalMetricDataList': [ { 'MetricName': 'string', 'Value': ..., 'Timestamp': datetime(2015, 1, 1) }, ], 'EnableNetworkIsolation': True|False, 'EnableInterContainerTrafficEncryption': True|False, 'EnableManagedSpotTraining': True|False, 'CheckpointConfig': { 'S3Uri': 'string', 'LocalPath': 'string' }, 'TrainingTimeInSeconds': 123, 'BillableTimeInSeconds': 123, 'DebugHookConfig': { 'LocalPath': 'string', 'S3OutputPath': 'string', 'HookParameters': { 'string': 'string' }, 'CollectionConfigurations': [ { 'CollectionName': 'string', 'CollectionParameters': { 'string': 'string' } }, ] }, 'ExperimentConfig': { 'ExperimentName': 'string', 'TrialName': 'string', 'TrialComponentDisplayName': 'string' }, 'DebugRuleConfigurations': [ { 'RuleConfigurationName': 'string', 'LocalPath': 'string', 'S3OutputPath': 'string', 'RuleEvaluatorImage': 'string', 'InstanceType': 'ml.t3.medium'|'ml.t3.large'|'ml.t3.xlarge'|'ml.t3.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.8xlarge'|'ml.r5.12xlarge'|'ml.r5.16xlarge'|'ml.r5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'VolumeSizeInGB': 123, 'RuleParameters': { 'string': 'string' } }, ], 'TensorBoardOutputConfig': { 'LocalPath': 'string', 'S3OutputPath': 'string' }, 'DebugRuleEvaluationStatuses': [ { 'RuleConfigurationName': 'string', 'RuleEvaluationJobArn': 'string', 'RuleEvaluationStatus': 'InProgress'|'NoIssuesFound'|'IssuesFound'|'Error'|'Stopping'|'Stopped', 'StatusDetails': 'string', 'LastModifiedTime': datetime(2015, 1, 1) }, ], 'Environment': { 'string': 'string' }, 'RetryStrategy': { 'MaximumRetryAttempts': 123 }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'Experiment': { 'ExperimentName': 'string', 'ExperimentArn': 'string', 'DisplayName': 'string', 'Source': { 'SourceArn': 'string', 'SourceType': 'string' }, 'Description': 'string', 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'Trial': { 'TrialName': 'string', 'TrialArn': 'string', 'DisplayName': 'string', 'ExperimentName': 'string', 'Source': { 'SourceArn': 'string', 'SourceType': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ], 'TrialComponentSummaries': [ { 'TrialComponentName': 'string', 'TrialComponentArn': 'string', 'TrialComponentSource': { 'SourceArn': 'string', 'SourceType': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' } }, ] }, 'TrialComponent': { 'TrialComponentName': 'string', 'DisplayName': 'string', 'TrialComponentArn': 'string', 'Source': { 'SourceArn': 'string', 'SourceType': 'string' }, 'Status': { 'PrimaryStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'Message': 'string' }, 'StartTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'Parameters': { 'string': { 'StringValue': 'string', 'NumberValue': 123.0 } }, 'InputArtifacts': { 'string': { 'MediaType': 'string', 'Value': 'string' } }, 'OutputArtifacts': { 'string': { 'MediaType': 'string', 'Value': 'string' } }, 'Metrics': [ { 'MetricName': 'string', 'SourceArn': 'string', 'TimeStamp': datetime(2015, 1, 1), 'Max': 123.0, 'Min': 123.0, 'Last': 123.0, 'Count': 123, 'Avg': 123.0, 'StdDev': 123.0 }, ], 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'SourceDetail': { 'SourceArn': 'string', 'TrainingJob': { 'TrainingJobName': 'string', 'TrainingJobArn': 'string', 'TuningJobArn': 'string', 'LabelingJobArn': 'string', 'AutoMLJobArn': 'string', 'ModelArtifacts': { 'S3ModelArtifacts': 'string' }, 'TrainingJobStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'SecondaryStatus': 'Starting'|'LaunchingMLInstances'|'PreparingTrainingStack'|'Downloading'|'DownloadingTrainingImage'|'Training'|'Uploading'|'Stopping'|'Stopped'|'MaxRuntimeExceeded'|'Completed'|'Failed'|'Interrupted'|'MaxWaitTimeExceeded'|'Updating'|'Restarting', 'FailureReason': 'string', 'HyperParameters': { 'string': 'string' }, 'AlgorithmSpecification': { 'TrainingImage': 'string', 'AlgorithmName': 'string', 'TrainingInputMode': 'Pipe'|'File'|'FastFile', 'MetricDefinitions': [ { 'Name': 'string', 'Regex': 'string' }, ], 'EnableSageMakerMetricsTimeSeries': True|False }, 'RoleArn': 'string', 'InputDataConfig': [ { 'ChannelName': 'string', 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'AttributeNames': [ 'string', ] }, 'FileSystemDataSource': { 'FileSystemId': 'string', 'FileSystemAccessMode': 'rw'|'ro', 'FileSystemType': 'EFS'|'FSxLustre', 'DirectoryPath': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'RecordWrapperType': 'None'|'RecordIO', 'InputMode': 'Pipe'|'File'|'FastFile', 'ShuffleConfig': { 'Seed': 123 } }, ], 'OutputDataConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, 'ResourceConfig': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.p3dn.24xlarge'|'ml.p4d.24xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5n.xlarge'|'ml.c5n.2xlarge'|'ml.c5n.4xlarge'|'ml.c5n.9xlarge'|'ml.c5n.18xlarge', 'InstanceCount': 123, 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' }, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123, 'MaxWaitTimeInSeconds': 123 }, 'CreationTime': datetime(2015, 1, 1), 'TrainingStartTime': datetime(2015, 1, 1), 'TrainingEndTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'SecondaryStatusTransitions': [ { 'Status': 'Starting'|'LaunchingMLInstances'|'PreparingTrainingStack'|'Downloading'|'DownloadingTrainingImage'|'Training'|'Uploading'|'Stopping'|'Stopped'|'MaxRuntimeExceeded'|'Completed'|'Failed'|'Interrupted'|'MaxWaitTimeExceeded'|'Updating'|'Restarting', 'StartTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'StatusMessage': 'string' }, ], 'FinalMetricDataList': [ { 'MetricName': 'string', 'Value': ..., 'Timestamp': datetime(2015, 1, 1) }, ], 'EnableNetworkIsolation': True|False, 'EnableInterContainerTrafficEncryption': True|False, 'EnableManagedSpotTraining': True|False, 'CheckpointConfig': { 'S3Uri': 'string', 'LocalPath': 'string' }, 'TrainingTimeInSeconds': 123, 'BillableTimeInSeconds': 123, 'DebugHookConfig': { 'LocalPath': 'string', 'S3OutputPath': 'string', 'HookParameters': { 'string': 'string' }, 'CollectionConfigurations': [ { 'CollectionName': 'string', 'CollectionParameters': { 'string': 'string' } }, ] }, 'ExperimentConfig': { 'ExperimentName': 'string', 'TrialName': 'string', 'TrialComponentDisplayName': 'string' }, 'DebugRuleConfigurations': [ { 'RuleConfigurationName': 'string', 'LocalPath': 'string', 'S3OutputPath': 'string', 'RuleEvaluatorImage': 'string', 'InstanceType': 'ml.t3.medium'|'ml.t3.large'|'ml.t3.xlarge'|'ml.t3.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.8xlarge'|'ml.r5.12xlarge'|'ml.r5.16xlarge'|'ml.r5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'VolumeSizeInGB': 123, 'RuleParameters': { 'string': 'string' } }, ], 'TensorBoardOutputConfig': { 'LocalPath': 'string', 'S3OutputPath': 'string' }, 'DebugRuleEvaluationStatuses': [ { 'RuleConfigurationName': 'string', 'RuleEvaluationJobArn': 'string', 'RuleEvaluationStatus': 'InProgress'|'NoIssuesFound'|'IssuesFound'|'Error'|'Stopping'|'Stopped', 'StatusDetails': 'string', 'LastModifiedTime': datetime(2015, 1, 1) }, ], 'Environment': { 'string': 'string' }, 'RetryStrategy': { 'MaximumRetryAttempts': 123 }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'ProcessingJob': { 'ProcessingInputs': [ { 'InputName': 'string', 'AppManaged': True|False, 'S3Input': { 'S3Uri': 'string', 'LocalPath': 'string', 'S3DataType': 'ManifestFile'|'S3Prefix', 'S3InputMode': 'Pipe'|'File', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'S3CompressionType': 'None'|'Gzip' }, 'DatasetDefinition': { 'AthenaDatasetDefinition': { 'Catalog': 'string', 'Database': 'string', 'QueryString': 'string', 'WorkGroup': 'string', 'OutputS3Uri': 'string', 'KmsKeyId': 'string', 'OutputFormat': 'PARQUET'|'ORC'|'AVRO'|'JSON'|'TEXTFILE', 'OutputCompression': 'GZIP'|'SNAPPY'|'ZLIB' }, 'RedshiftDatasetDefinition': { 'ClusterId': 'string', 'Database': 'string', 'DbUser': 'string', 'QueryString': 'string', 'ClusterRoleArn': 'string', 'OutputS3Uri': 'string', 'KmsKeyId': 'string', 'OutputFormat': 'PARQUET'|'CSV', 'OutputCompression': 'None'|'GZIP'|'BZIP2'|'ZSTD'|'SNAPPY' }, 'LocalPath': 'string', 'DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'InputMode': 'Pipe'|'File' } }, ], 'ProcessingOutputConfig': { 'Outputs': [ { 'OutputName': 'string', 'S3Output': { 'S3Uri': 'string', 'LocalPath': 'string', 'S3UploadMode': 'Continuous'|'EndOfJob' }, 'FeatureStoreOutput': { 'FeatureGroupName': 'string' }, 'AppManaged': True|False }, ], 'KmsKeyId': 'string' }, 'ProcessingJobName': 'string', 'ProcessingResources': { 'ClusterConfig': { 'InstanceCount': 123, 'InstanceType': 'ml.t3.medium'|'ml.t3.large'|'ml.t3.xlarge'|'ml.t3.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.8xlarge'|'ml.r5.12xlarge'|'ml.r5.16xlarge'|'ml.r5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' } }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123 }, 'AppSpecification': { 'ImageUri': 'string', 'ContainerEntrypoint': [ 'string', ], 'ContainerArguments': [ 'string', ] }, 'Environment': { 'string': 'string' }, 'NetworkConfig': { 'EnableInterContainerTrafficEncryption': True|False, 'EnableNetworkIsolation': True|False, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] } }, 'RoleArn': 'string', 'ExperimentConfig': { 'ExperimentName': 'string', 'TrialName': 'string', 'TrialComponentDisplayName': 'string' }, 'ProcessingJobArn': 'string', 'ProcessingJobStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'ExitMessage': 'string', 'FailureReason': 'string', 'ProcessingEndTime': datetime(2015, 1, 1), 'ProcessingStartTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'CreationTime': datetime(2015, 1, 1), 'MonitoringScheduleArn': 'string', 'AutoMLJobArn': 'string', 'TrainingJobArn': 'string', 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'TransformJob': { 'TransformJobName': 'string', 'TransformJobArn': 'string', 'TransformJobStatus': 'InProgress'|'Completed'|'Failed'|'Stopping'|'Stopped', 'FailureReason': 'string', 'ModelName': 'string', 'MaxConcurrentTransforms': 123, 'ModelClientConfig': { 'InvocationsTimeoutInSeconds': 123, 'InvocationsMaxRetries': 123 }, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'TransformStartTime': datetime(2015, 1, 1), 'TransformEndTime': datetime(2015, 1, 1), 'LabelingJobArn': 'string', 'AutoMLJobArn': 'string', 'DataProcessing': { 'InputFilter': 'string', 'OutputFilter': 'string', 'JoinSource': 'Input'|'None' }, 'ExperimentConfig': { 'ExperimentName': 'string', 'TrialName': 'string', 'TrialComponentDisplayName': 'string' }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] } }, 'LineageGroupArn': 'string', 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ], 'Parents': [ { 'TrialName': 'string', 'ExperimentName': 'string' }, ] }, 'Endpoint': { 'EndpointName': 'string', 'EndpointArn': 'string', 'EndpointConfigName': 'string', 'ProductionVariants': [ { 'VariantName': 'string', 'DeployedImages': [ { 'SpecifiedImage': 'string', 'ResolvedImage': 'string', 'ResolutionTime': datetime(2015, 1, 1) }, ], 'CurrentWeight': ..., 'DesiredWeight': ..., 'CurrentInstanceCount': 123, 'DesiredInstanceCount': 123, 'VariantStatus': [ { 'Status': 'Creating'|'Updating'|'Deleting'|'ActivatingTraffic'|'Baking', 'StatusMessage': 'string', 'StartTime': datetime(2015, 1, 1) }, ], 'CurrentServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 }, 'DesiredServerlessConfig': { 'MemorySizeInMB': 123, 'MaxConcurrency': 123 } }, ], 'DataCaptureConfig': { 'EnableCapture': True|False, 'CaptureStatus': 'Started'|'Stopped', 'CurrentSamplingPercentage': 123, 'DestinationS3Uri': 'string', 'KmsKeyId': 'string' }, 'EndpointStatus': 'OutOfService'|'Creating'|'Updating'|'SystemUpdating'|'RollingBack'|'InService'|'Deleting'|'Failed', 'FailureReason': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'MonitoringSchedules': [ { 'MonitoringScheduleArn': 'string', 'MonitoringScheduleName': 'string', 'MonitoringScheduleStatus': 'Pending'|'Failed'|'Scheduled'|'Stopped', 'MonitoringType': 'DataQuality'|'ModelQuality'|'ModelBias'|'ModelExplainability', 'FailureReason': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'MonitoringScheduleConfig': { 'ScheduleConfig': { 'ScheduleExpression': 'string' }, 'MonitoringJobDefinition': { 'BaselineConfig': { 'BaseliningJobName': 'string', 'ConstraintsResource': { 'S3Uri': 'string' }, 'StatisticsResource': { 'S3Uri': 'string' } }, 'MonitoringInputs': [ { 'EndpointInput': { 'EndpointName': 'string', 'LocalPath': 'string', 'S3InputMode': 'Pipe'|'File', 'S3DataDistributionType': 'FullyReplicated'|'ShardedByS3Key', 'FeaturesAttribute': 'string', 'InferenceAttribute': 'string', 'ProbabilityAttribute': 'string', 'ProbabilityThresholdAttribute': 123.0, 'StartTimeOffset': 'string', 'EndTimeOffset': 'string' } }, ], 'MonitoringOutputConfig': { 'MonitoringOutputs': [ { 'S3Output': { 'S3Uri': 'string', 'LocalPath': 'string', 'S3UploadMode': 'Continuous'|'EndOfJob' } }, ], 'KmsKeyId': 'string' }, 'MonitoringResources': { 'ClusterConfig': { 'InstanceCount': 123, 'InstanceType': 'ml.t3.medium'|'ml.t3.large'|'ml.t3.xlarge'|'ml.t3.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.8xlarge'|'ml.r5.12xlarge'|'ml.r5.16xlarge'|'ml.r5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'VolumeSizeInGB': 123, 'VolumeKmsKeyId': 'string' } }, 'MonitoringAppSpecification': { 'ImageUri': 'string', 'ContainerEntrypoint': [ 'string', ], 'ContainerArguments': [ 'string', ], 'RecordPreprocessorSourceUri': 'string', 'PostAnalyticsProcessorSourceUri': 'string' }, 'StoppingCondition': { 'MaxRuntimeInSeconds': 123 }, 'Environment': { 'string': 'string' }, 'NetworkConfig': { 'EnableInterContainerTrafficEncryption': True|False, 'EnableNetworkIsolation': True|False, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] } }, 'RoleArn': 'string' }, 'MonitoringJobDefinitionName': 'string', 'MonitoringType': 'DataQuality'|'ModelQuality'|'ModelBias'|'ModelExplainability' }, 'EndpointName': 'string', 'LastMonitoringExecutionSummary': { 'MonitoringScheduleName': 'string', 'ScheduledTime': datetime(2015, 1, 1), 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'MonitoringExecutionStatus': 'Pending'|'Completed'|'CompletedWithViolations'|'InProgress'|'Failed'|'Stopping'|'Stopped', 'ProcessingJobArn': 'string', 'EndpointName': 'string', 'FailureReason': 'string', 'MonitoringJobDefinitionName': 'string', 'MonitoringType': 'DataQuality'|'ModelQuality'|'ModelBias'|'ModelExplainability' }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, ], 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'ModelPackage': { 'ModelPackageName': 'string', 'ModelPackageGroupName': 'string', 'ModelPackageVersion': 123, 'ModelPackageArn': 'string', 'ModelPackageDescription': 'string', 'CreationTime': datetime(2015, 1, 1), 'InferenceSpecification': { 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, 'SourceAlgorithmSpecification': { 'SourceAlgorithms': [ { 'ModelDataUrl': 'string', 'AlgorithmName': 'string' }, ] }, 'ValidationSpecification': { 'ValidationRole': 'string', 'ValidationProfiles': [ { 'ProfileName': 'string', 'TransformJobDefinition': { 'MaxConcurrentTransforms': 123, 'MaxPayloadInMB': 123, 'BatchStrategy': 'MultiRecord'|'SingleRecord', 'Environment': { 'string': 'string' }, 'TransformInput': { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile', 'S3Uri': 'string' } }, 'ContentType': 'string', 'CompressionType': 'None'|'Gzip', 'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord' }, 'TransformOutput': { 'S3OutputPath': 'string', 'Accept': 'string', 'AssembleWith': 'None'|'Line', 'KmsKeyId': 'string' }, 'TransformResources': { 'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', 'InstanceCount': 123, 'VolumeKmsKeyId': 'string' } } }, ] }, 'ModelPackageStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting', 'ModelPackageStatusDetails': { 'ValidationStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ], 'ImageScanStatuses': [ { 'Name': 'string', 'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed', 'FailureReason': 'string' }, ] }, 'CertifyForMarketplace': True|False, 'ModelApprovalStatus': 'Approved'|'Rejected'|'PendingManualApproval', 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'MetadataProperties': { 'CommitId': 'string', 'Repository': 'string', 'GeneratedBy': 'string', 'ProjectId': 'string' }, 'ModelMetrics': { 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Bias': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingReport': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Report': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } }, 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'ApprovalDescription': 'string', 'Domain': 'string', 'Task': 'string', 'SamplePayloadUrl': 'string', 'AdditionalInferenceSpecifications': [ { 'Name': 'string', 'Description': 'string', 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ], 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ], 'CustomerMetadataProperties': { 'string': 'string' }, 'DriftCheckBaselines': { 'Bias': { 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PreTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'PostTrainingConstraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'Explainability': { 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'ConfigFile': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } }, 'ModelDataQuality': { 'Statistics': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' }, 'Constraints': { 'ContentType': 'string', 'ContentDigest': 'string', 'S3Uri': 'string' } } } }, 'ModelPackageGroup': { 'ModelPackageGroupName': 'string', 'ModelPackageGroupArn': 'string', 'ModelPackageGroupDescription': 'string', 'CreationTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'ModelPackageGroupStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting'|'DeleteFailed', 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'Pipeline': { 'PipelineArn': 'string', 'PipelineName': 'string', 'PipelineDisplayName': 'string', 'PipelineDescription': 'string', 'RoleArn': 'string', 'PipelineStatus': 'Active', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'LastRunTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'PipelineExecution': { 'PipelineArn': 'string', 'PipelineExecutionArn': 'string', 'PipelineExecutionDisplayName': 'string', 'PipelineExecutionStatus': 'Executing'|'Stopping'|'Stopped'|'Failed'|'Succeeded', 'PipelineExecutionDescription': 'string', 'PipelineExperimentConfig': { 'ExperimentName': 'string', 'TrialName': 'string' }, 'FailureReason': 'string', 'CreationTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'PipelineParameters': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'FeatureGroup': { 'FeatureGroupArn': 'string', 'FeatureGroupName': 'string', 'RecordIdentifierFeatureName': 'string', 'EventTimeFeatureName': 'string', 'FeatureDefinitions': [ { 'FeatureName': 'string', 'FeatureType': 'Integral'|'Fractional'|'String' }, ], 'CreationTime': datetime(2015, 1, 1), 'OnlineStoreConfig': { 'SecurityConfig': { 'KmsKeyId': 'string' }, 'EnableOnlineStore': True|False }, 'OfflineStoreConfig': { 'S3StorageConfig': { 'S3Uri': 'string', 'KmsKeyId': 'string', 'ResolvedOutputS3Uri': 'string' }, 'DisableGlueTableCreation': True|False, 'DataCatalogConfig': { 'TableName': 'string', 'Catalog': 'string', 'Database': 'string' } }, 'RoleArn': 'string', 'FeatureGroupStatus': 'Creating'|'Created'|'CreateFailed'|'Deleting'|'DeleteFailed', 'OfflineStoreStatus': { 'Status': 'Active'|'Blocked'|'Disabled', 'BlockedReason': 'string' }, 'FailureReason': 'string', 'Description': 'string', 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'Project': { 'ProjectArn': 'string', 'ProjectName': 'string', 'ProjectId': 'string', 'ProjectDescription': 'string', 'ServiceCatalogProvisioningDetails': { 'ProductId': 'string', 'ProvisioningArtifactId': 'string', 'PathId': 'string', 'ProvisioningParameters': [ { 'Key': 'string', 'Value': 'string' }, ] }, 'ServiceCatalogProvisionedProductDetails': { 'ProvisionedProductId': 'string', 'ProvisionedProductStatusMessage': 'string' }, 'ProjectStatus': 'Pending'|'CreateInProgress'|'CreateCompleted'|'CreateFailed'|'DeleteInProgress'|'DeleteFailed'|'DeleteCompleted'|'UpdateInProgress'|'UpdateCompleted'|'UpdateFailed', 'CreatedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' }, 'CreationTime': datetime(2015, 1, 1), 'Tags': [ { 'Key': 'string', 'Value': 'string' }, ], 'LastModifiedTime': datetime(2015, 1, 1), 'LastModifiedBy': { 'UserProfileArn': 'string', 'UserProfileName': 'string', 'DomainId': 'string' } } }, ], 'NextToken': 'string' } **Response Structure** :: # This section is too large to render. # Please see the AWS API Documentation linked below. `AWS API Documentation <https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/Search>`_
{'AdditionalInferenceSpecificationsToAdd': [{'Containers': [{'ContainerHostname': 'string', 'Environment': {'string': 'string'}, 'Framework': 'string', 'FrameworkVersion': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ModelInput': {'DataInputConfig': 'string'}, 'NearestModelName': 'string', 'ProductId': 'string'}], 'Description': 'string', 'Name': 'string', 'SupportedContentTypes': ['string'], 'SupportedRealtimeInferenceInstanceTypes': ['ml.t2.medium ' '| ' 'ml.t2.large ' '| ' 'ml.t2.xlarge ' '| ' 'ml.t2.2xlarge ' '| ' 'ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.m5d.large ' '| ' 'ml.m5d.xlarge ' '| ' 'ml.m5d.2xlarge ' '| ' 'ml.m5d.4xlarge ' '| ' 'ml.m5d.12xlarge ' '| ' 'ml.m5d.24xlarge ' '| ' 'ml.c4.large ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.large ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.c5d.large ' '| ' 'ml.c5d.xlarge ' '| ' 'ml.c5d.2xlarge ' '| ' 'ml.c5d.4xlarge ' '| ' 'ml.c5d.9xlarge ' '| ' 'ml.c5d.18xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge ' '| ' 'ml.r5.large ' '| ' 'ml.r5.xlarge ' '| ' 'ml.r5.2xlarge ' '| ' 'ml.r5.4xlarge ' '| ' 'ml.r5.12xlarge ' '| ' 'ml.r5.24xlarge ' '| ' 'ml.r5d.large ' '| ' 'ml.r5d.xlarge ' '| ' 'ml.r5d.2xlarge ' '| ' 'ml.r5d.4xlarge ' '| ' 'ml.r5d.12xlarge ' '| ' 'ml.r5d.24xlarge ' '| ' 'ml.inf1.xlarge ' '| ' 'ml.inf1.2xlarge ' '| ' 'ml.inf1.6xlarge ' '| ' 'ml.inf1.24xlarge'], 'SupportedResponseMIMETypes': ['string'], 'SupportedTransformInstanceTypes': ['ml.m4.xlarge ' '| ' 'ml.m4.2xlarge ' '| ' 'ml.m4.4xlarge ' '| ' 'ml.m4.10xlarge ' '| ' 'ml.m4.16xlarge ' '| ' 'ml.c4.xlarge ' '| ' 'ml.c4.2xlarge ' '| ' 'ml.c4.4xlarge ' '| ' 'ml.c4.8xlarge ' '| ' 'ml.p2.xlarge ' '| ' 'ml.p2.8xlarge ' '| ' 'ml.p2.16xlarge ' '| ' 'ml.p3.2xlarge ' '| ' 'ml.p3.8xlarge ' '| ' 'ml.p3.16xlarge ' '| ' 'ml.c5.xlarge ' '| ' 'ml.c5.2xlarge ' '| ' 'ml.c5.4xlarge ' '| ' 'ml.c5.9xlarge ' '| ' 'ml.c5.18xlarge ' '| ' 'ml.m5.large ' '| ' 'ml.m5.xlarge ' '| ' 'ml.m5.2xlarge ' '| ' 'ml.m5.4xlarge ' '| ' 'ml.m5.12xlarge ' '| ' 'ml.m5.24xlarge ' '| ' 'ml.g4dn.xlarge ' '| ' 'ml.g4dn.2xlarge ' '| ' 'ml.g4dn.4xlarge ' '| ' 'ml.g4dn.8xlarge ' '| ' 'ml.g4dn.12xlarge ' '| ' 'ml.g4dn.16xlarge']}]}
Updates a versioned model.
See also: AWS API Documentation
Request Syntax
client.update_model_package( ModelPackageArn='string', ModelApprovalStatus='Approved'|'Rejected'|'PendingManualApproval', ApprovalDescription='string', CustomerMetadataProperties={ 'string': 'string' }, CustomerMetadataPropertiesToRemove=[ 'string', ], AdditionalInferenceSpecificationsToAdd=[ { 'Name': 'string', 'Description': 'string', 'Containers': [ { 'ContainerHostname': 'string', 'Image': 'string', 'ImageDigest': 'string', 'ModelDataUrl': 'string', 'ProductId': 'string', 'Environment': { 'string': 'string' }, 'ModelInput': { 'DataInputConfig': 'string' }, 'Framework': 'string', 'FrameworkVersion': 'string', 'NearestModelName': 'string' }, ], 'SupportedTransformInstanceTypes': [ 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge', ], 'SupportedRealtimeInferenceInstanceTypes': [ 'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge', ], 'SupportedContentTypes': [ 'string', ], 'SupportedResponseMIMETypes': [ 'string', ] }, ] )
string
[REQUIRED]
The Amazon Resource Name (ARN) of the model package.
string
The approval status of the model.
string
A description for the approval status of the model.
dict
The metadata properties associated with the model package versions.
(string) --
(string) --
list
The metadata properties associated with the model package versions to remove.
(string) --
list
An array of additional Inference Specification objects to be added to the existing array additional Inference Specification. Total number of additional Inference Specifications can not exceed 15. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.
(dict) --
A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package
Name (string) -- [REQUIRED]
A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.
Description (string) --
A description of the additional Inference specification
Containers (list) -- [REQUIRED]
The Amazon ECR registry path of the Docker image that contains the inference code.
(dict) --
Describes the Docker container for the model package.
ContainerHostname (string) --
The DNS host name for the Docker container.
Image (string) -- [REQUIRED]
The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.
If you are using your own custom algorithm instead of an algorithm provided by Amazon SageMaker, the inference code must meet Amazon SageMaker requirements. Amazon SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.
ImageDigest (string) --
An MD5 hash of the training algorithm that identifies the Docker image used for training.
ModelDataUrl (string) --
The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).
ProductId (string) --
The Amazon Web Services Marketplace product ID of the model package.
Environment (dict) --
The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.
(string) --
(string) --
ModelInput (dict) --
A structure with Model Input details.
DataInputConfig (string) -- [REQUIRED]
The input configuration object for the model.
Framework (string) --
The machine learning framework of the model package container image.
FrameworkVersion (string) --
The framework version of the Model Package Container Image.
NearestModelName (string) --
The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.
SupportedTransformInstanceTypes (list) --
A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.
(string) --
SupportedRealtimeInferenceInstanceTypes (list) --
A list of the instance types that are used to generate inferences in real-time.
(string) --
SupportedContentTypes (list) --
The supported MIME types for the input data.
(string) --
SupportedResponseMIMETypes (list) --
The supported MIME types for the output data.
(string) --
dict
Response Syntax
{ 'ModelPackageArn': 'string' }
Response Structure
(dict) --
ModelPackageArn (string) --
The Amazon Resource Name (ARN) of the model.