Amazon SageMaker Service

2024/05/30 - Amazon SageMaker Service - 8 updated api methods

Changes  Adds Model Card information as a new component to Model Package. Autopilot launches algorithm selection for TimeSeries modality to generate AutoML candidates per algorithm.

CreateAutoMLJob (updated) Link ¶
Changes (request)
{'AutoMLJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': {'AutoMLAlgorithms': {'arima',
                                                                                             'cnn-qr',
                                                                                             'deepar',
                                                                                             'ets',
                                                                                             'npts',
                                                                                             'prophet'}}}}}

Creates an Autopilot job also referred to as Autopilot experiment or AutoML job.

You can find the best-performing model after you run an AutoML job by calling DescribeAutoMLJobV2 (recommended) or DescribeAutoMLJob.

See also: AWS API Documentation

Request Syntax

client.create_auto_ml_job(
    AutoMLJobName='string',
    InputDataConfig=[
        {
            'DataSource': {
                'S3DataSource': {
                    'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                    'S3Uri': 'string'
                }
            },
            'CompressionType': 'None'|'Gzip',
            'TargetAttributeName': 'string',
            'ContentType': 'string',
            'ChannelType': 'training'|'validation',
            'SampleWeightAttributeName': 'string'
        },
    ],
    OutputDataConfig={
        'KmsKeyId': 'string',
        'S3OutputPath': 'string'
    },
    ProblemType='BinaryClassification'|'MulticlassClassification'|'Regression',
    AutoMLJobObjective={
        'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
    },
    AutoMLJobConfig={
        'CompletionCriteria': {
            'MaxCandidates': 123,
            'MaxRuntimePerTrainingJobInSeconds': 123,
            'MaxAutoMLJobRuntimeInSeconds': 123
        },
        'SecurityConfig': {
            'VolumeKmsKeyId': 'string',
            'EnableInterContainerTrafficEncryption': True|False,
            'VpcConfig': {
                'SecurityGroupIds': [
                    'string',
                ],
                'Subnets': [
                    'string',
                ]
            }
        },
        'CandidateGenerationConfig': {
            'FeatureSpecificationS3Uri': 'string',
            'AlgorithmsConfig': [
                {
                    'AutoMLAlgorithms': [
                        'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                    ]
                },
            ]
        },
        'DataSplitConfig': {
            'ValidationFraction': ...
        },
        'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING'
    },
    RoleArn='string',
    GenerateCandidateDefinitionsOnly=True|False,
    Tags=[
        {
            'Key': 'string',
            'Value': 'string'
        },
    ],
    ModelDeployConfig={
        'AutoGenerateEndpointName': True|False,
        'EndpointName': 'string'
    }
)
type AutoMLJobName:

string

param AutoMLJobName:

[REQUIRED]

Identifies an Autopilot job. The name must be unique to your account and is case insensitive.

type InputDataConfig:

list

param InputDataConfig:

[REQUIRED]

An array of channel objects that describes the input data and its location. Each channel is a named input source. Similar to InputDataConfig supported by HyperParameterTrainingJobDefinition. Format(s) supported: CSV, Parquet. A minimum of 500 rows is required for the training dataset. There is not a minimum number of rows required for the validation dataset.

  • (dict) --

    A channel is a named input source that training algorithms can consume. The validation dataset size is limited to less than 2 GB. The training dataset size must be less than 100 GB. For more information, see Channel.

    • DataSource (dict) --

      The data source for an AutoML channel.

      • S3DataSource (dict) -- [REQUIRED]

        The Amazon S3 location of the input data.

        • S3DataType (string) -- [REQUIRED]

          The data type.

          • If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix for model training. The S3Prefix should have the following format: s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE

          • If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want SageMaker to use for model training. A ManifestFile should have the format shown below: [ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"}, "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1", "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2", ... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]

          • If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile is available for V2 API jobs only (for example, for jobs created by calling CreateAutoMLJobV2). Here is a minimal, single-record example of an AugmentedManifestFile: {"source-ref": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/cats/cat.jpg", "label-metadata": {"class-name": "cat" } For more information on AugmentedManifestFile, see Provide Dataset Metadata to Training Jobs with an Augmented Manifest File.

        • S3Uri (string) -- [REQUIRED]

          The URL to the Amazon S3 data source. The Uri refers to the Amazon S3 prefix or ManifestFile depending on the data type.

    • CompressionType (string) --

      You can use Gzip or None. The default value is None.

    • TargetAttributeName (string) -- [REQUIRED]

      The name of the target variable in supervised learning, usually represented by 'y'.

    • ContentType (string) --

      The content type of the data from the input source. You can use text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

    • ChannelType (string) --

      The channel type (optional) is an enum string. The default value is training. Channels for training and validation must share the same ContentType and TargetAttributeName. For information on specifying training and validation channel types, see How to specify training and validation datasets.

    • SampleWeightAttributeName (string) --

      If specified, this column name indicates which column of the dataset should be treated as sample weights for use by the objective metric during the training, evaluation, and the selection of the best model. This column is not considered as a predictive feature. For more information on Autopilot metrics, see Metrics and validation.

      Sample weights should be numeric, non-negative, with larger values indicating which rows are more important than others. Data points that have invalid or no weight value are excluded.

      Support for sample weights is available in Ensembling mode only.

type OutputDataConfig:

dict

param OutputDataConfig:

[REQUIRED]

Provides information about encryption and the Amazon S3 output path needed to store artifacts from an AutoML job. Format(s) supported: CSV.

  • KmsKeyId (string) --

    The Key Management Service encryption key ID.

  • S3OutputPath (string) -- [REQUIRED]

    The Amazon S3 output path. Must be 128 characters or less.

type ProblemType:

string

param ProblemType:

Defines the type of supervised learning problem available for the candidates. For more information, see SageMaker Autopilot problem types.

type AutoMLJobObjective:

dict

param AutoMLJobObjective:

Specifies a metric to minimize or maximize as the objective of a job. If not specified, the default objective metric depends on the problem type. See AutoMLJobObjective for the default values.

  • MetricName (string) -- [REQUIRED]

    The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

    The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

    • For tabular problem types:

      • List of available metrics:

        • Regression: MAE, MSE, R2, RMSE

        • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

        • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

      For a description of each metric, see Autopilot metrics for classification and regression.

      • Default objective metrics:

        • Regression: MSE.

        • Binary classification: F1.

        • Multiclass classification: Accuracy.

    • For image or text classification problem types:

    • For time-series forecasting problem types:

    • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

type AutoMLJobConfig:

dict

param AutoMLJobConfig:

A collection of settings used to configure an AutoML job.

  • CompletionCriteria (dict) --

    How long an AutoML job is allowed to run, or how many candidates a job is allowed to generate.

    • MaxCandidates (integer) --

      The maximum number of times a training job is allowed to run.

      For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

    • MaxRuntimePerTrainingJobInSeconds (integer) --

      The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

      For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

      For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

    • MaxAutoMLJobRuntimeInSeconds (integer) --

      The maximum runtime, in seconds, an AutoML job has to complete.

      If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

  • SecurityConfig (dict) --

    The security configuration for traffic encryption or Amazon VPC settings.

    • VolumeKmsKeyId (string) --

      The key used to encrypt stored data.

    • EnableInterContainerTrafficEncryption (boolean) --

      Whether to use traffic encryption between the container layers.

    • VpcConfig (dict) --

      The VPC configuration.

      • SecurityGroupIds (list) -- [REQUIRED]

        The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.

        • (string) --

      • Subnets (list) -- [REQUIRED]

        The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.

        • (string) --

  • CandidateGenerationConfig (dict) --

    The configuration for generating a candidate for an AutoML job (optional).

    • FeatureSpecificationS3Uri (string) --

      A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job. You can input FeatureAttributeNames (optional) in JSON format as shown below:

      { "FeatureAttributeNames":["col1", "col2", ...] }.

      You can also specify the data type of the feature (optional) in the format shown below:

      { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

      In ensembling mode, Autopilot only supports the following data types: numeric, categorical, text, and datetime. In HPO mode, Autopilot can support numeric, categorical, text, datetime, and sequence.

      If only FeatureDataTypes is provided, the column keys ( col1, col2,..) should be a subset of the column names in the input data.

      If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames.

      The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.

    • AlgorithmsConfig (list) --

      Stores the configuration information for the selection of algorithms trained on tabular data.

      The list of available algorithms to choose from depends on the training mode set in TabularJobConfig.Mode.

      • AlgorithmsConfig should not be set if the training mode is set on AUTO.

      • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

      • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

      For the list of all algorithms per problem type and training mode, see AutoMLAlgorithmConfig.

      For more information on each algorithm, see the Algorithm support section in Autopilot developer guide.

      • (dict) --

        The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

        • AutoMLAlgorithms (list) -- [REQUIRED]

          The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

          • For the tabular problem type TabularJobConfig:

          • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

            • Choose your algorithms from this list.

              • "cnn-qr"

              • "deepar"

              • "prophet"

              • "arima"

              • "npts"

              • "ets"

          • (string) --

  • DataSplitConfig (dict) --

    The configuration for splitting the input training dataset.

    Type: AutoMLDataSplitConfig

    • ValidationFraction (float) --

      The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.

  • Mode (string) --

    The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO. In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.

    The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.

    The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.

type RoleArn:

string

param RoleArn:

[REQUIRED]

The ARN of the role that is used to access the data.

type GenerateCandidateDefinitionsOnly:

boolean

param GenerateCandidateDefinitionsOnly:

Generates possible candidates without training the models. A candidate is a combination of data preprocessors, algorithms, and algorithm parameter settings.

type Tags:

list

param Tags:

An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web ServicesResources. Tag keys must be unique per resource.

  • (dict) --

    A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.

    You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.

    For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.

    • Key (string) -- [REQUIRED]

      The tag key. Tag keys must be unique per resource.

    • Value (string) -- [REQUIRED]

      The tag value.

type ModelDeployConfig:

dict

param ModelDeployConfig:

Specifies how to generate the endpoint name for an automatic one-click Autopilot model deployment.

  • AutoGenerateEndpointName (boolean) --

    Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False.

  • EndpointName (string) --

    Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.

rtype:

dict

returns:

Response Syntax

{
    'AutoMLJobArn': 'string'
}

Response Structure

  • (dict) --

    • AutoMLJobArn (string) --

      The unique ARN assigned to the AutoML job when it is created.

CreateAutoMLJobV2 (updated) Link ¶
Changes (request)
{'AutoMLProblemTypeConfig': {'TabularJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': {'AutoMLAlgorithms': {'arima',
                                                                                                                          'cnn-qr',
                                                                                                                          'deepar',
                                                                                                                          'ets',
                                                                                                                          'npts',
                                                                                                                          'prophet'}}}},
                             'TimeSeriesForecastingJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': [{'AutoMLAlgorithms': ['xgboost '
                                                                                                                                         '| '
                                                                                                                                         'linear-learner '
                                                                                                                                         '| '
                                                                                                                                         'mlp '
                                                                                                                                         '| '
                                                                                                                                         'lightgbm '
                                                                                                                                         '| '
                                                                                                                                         'catboost '
                                                                                                                                         '| '
                                                                                                                                         'randomforest '
                                                                                                                                         '| '
                                                                                                                                         'extra-trees '
                                                                                                                                         '| '
                                                                                                                                         'nn-torch '
                                                                                                                                         '| '
                                                                                                                                         'fastai '
                                                                                                                                         '| '
                                                                                                                                         'cnn-qr '
                                                                                                                                         '| '
                                                                                                                                         'deepar '
                                                                                                                                         '| '
                                                                                                                                         'prophet '
                                                                                                                                         '| '
                                                                                                                                         'npts '
                                                                                                                                         '| '
                                                                                                                                         'arima '
                                                                                                                                         '| '
                                                                                                                                         'ets']}]}}}}

Creates an Autopilot job also referred to as Autopilot experiment or AutoML job V2.

For the list of available problem types supported by CreateAutoMLJobV2, see AutoMLProblemTypeConfig.

You can find the best-performing model after you run an AutoML job V2 by calling DescribeAutoMLJobV2.

See also: AWS API Documentation

Request Syntax

client.create_auto_ml_job_v2(
    AutoMLJobName='string',
    AutoMLJobInputDataConfig=[
        {
            'ChannelType': 'training'|'validation',
            'ContentType': 'string',
            'CompressionType': 'None'|'Gzip',
            'DataSource': {
                'S3DataSource': {
                    'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                    'S3Uri': 'string'
                }
            }
        },
    ],
    OutputDataConfig={
        'KmsKeyId': 'string',
        'S3OutputPath': 'string'
    },
    AutoMLProblemTypeConfig={
        'ImageClassificationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            }
        },
        'TextClassificationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'ContentColumn': 'string',
            'TargetLabelColumn': 'string'
        },
        'TimeSeriesForecastingJobConfig': {
            'FeatureSpecificationS3Uri': 'string',
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'ForecastFrequency': 'string',
            'ForecastHorizon': 123,
            'ForecastQuantiles': [
                'string',
            ],
            'Transformations': {
                'Filling': {
                    'string': {
                        'string': 'string'
                    }
                },
                'Aggregation': {
                    'string': 'sum'|'avg'|'first'|'min'|'max'
                }
            },
            'TimeSeriesConfig': {
                'TargetAttributeName': 'string',
                'TimestampAttributeName': 'string',
                'ItemIdentifierAttributeName': 'string',
                'GroupingAttributeNames': [
                    'string',
                ]
            },
            'HolidayConfig': [
                {
                    'CountryCode': 'string'
                },
            ],
            'CandidateGenerationConfig': {
                'AlgorithmsConfig': [
                    {
                        'AutoMLAlgorithms': [
                            'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                        ]
                    },
                ]
            }
        },
        'TabularJobConfig': {
            'CandidateGenerationConfig': {
                'AlgorithmsConfig': [
                    {
                        'AutoMLAlgorithms': [
                            'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                        ]
                    },
                ]
            },
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'FeatureSpecificationS3Uri': 'string',
            'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING',
            'GenerateCandidateDefinitionsOnly': True|False,
            'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression',
            'TargetAttributeName': 'string',
            'SampleWeightAttributeName': 'string'
        },
        'TextGenerationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'BaseModelName': 'string',
            'TextGenerationHyperParameters': {
                'string': 'string'
            },
            'ModelAccessConfig': {
                'AcceptEula': True|False
            }
        }
    },
    RoleArn='string',
    Tags=[
        {
            'Key': 'string',
            'Value': 'string'
        },
    ],
    SecurityConfig={
        'VolumeKmsKeyId': 'string',
        'EnableInterContainerTrafficEncryption': True|False,
        'VpcConfig': {
            'SecurityGroupIds': [
                'string',
            ],
            'Subnets': [
                'string',
            ]
        }
    },
    AutoMLJobObjective={
        'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
    },
    ModelDeployConfig={
        'AutoGenerateEndpointName': True|False,
        'EndpointName': 'string'
    },
    DataSplitConfig={
        'ValidationFraction': ...
    }
)
type AutoMLJobName:

string

param AutoMLJobName:

[REQUIRED]

Identifies an Autopilot job. The name must be unique to your account and is case insensitive.

type AutoMLJobInputDataConfig:

list

param AutoMLJobInputDataConfig:

[REQUIRED]

An array of channel objects describing the input data and their location. Each channel is a named input source. Similar to the InputDataConfig attribute in the CreateAutoMLJob input parameters. The supported formats depend on the problem type:

  • For tabular problem types: S3Prefix, ManifestFile.

  • For image classification: S3Prefix, ManifestFile, AugmentedManifestFile.

  • For text classification: S3Prefix.

  • For time-series forecasting: S3Prefix.

  • For text generation (LLMs fine-tuning): S3Prefix.

  • (dict) --

    A channel is a named input source that training algorithms can consume. This channel is used for AutoML jobs V2 (jobs created by calling CreateAutoMLJobV2).

    • ChannelType (string) --

      The type of channel. Defines whether the data are used for training or validation. The default value is training. Channels for training and validation must share the same ContentType

    • ContentType (string) --

      The content type of the data from the input source. The following are the allowed content types for different problems:

      • For tabular problem types: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

      • For image classification: image/png, image/jpeg, or image/*. The default value is image/*.

      • For text classification: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

      • For time-series forecasting: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

      • For text generation (LLMs fine-tuning): text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

    • CompressionType (string) --

      The allowed compression types depend on the input format and problem type. We allow the compression type Gzip for S3Prefix inputs on tabular data only. For all other inputs, the compression type should be None. If no compression type is provided, we default to None.

    • DataSource (dict) --

      The data source for an AutoML channel (Required).

      • S3DataSource (dict) -- [REQUIRED]

        The Amazon S3 location of the input data.

        • S3DataType (string) -- [REQUIRED]

          The data type.

          • If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix for model training. The S3Prefix should have the following format: s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE

          • If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want SageMaker to use for model training. A ManifestFile should have the format shown below: [ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"}, "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1", "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2", ... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]

          • If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile is available for V2 API jobs only (for example, for jobs created by calling CreateAutoMLJobV2). Here is a minimal, single-record example of an AugmentedManifestFile: {"source-ref": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/cats/cat.jpg", "label-metadata": {"class-name": "cat" } For more information on AugmentedManifestFile, see Provide Dataset Metadata to Training Jobs with an Augmented Manifest File.

        • S3Uri (string) -- [REQUIRED]

          The URL to the Amazon S3 data source. The Uri refers to the Amazon S3 prefix or ManifestFile depending on the data type.

type OutputDataConfig:

dict

param OutputDataConfig:

[REQUIRED]

Provides information about encryption and the Amazon S3 output path needed to store artifacts from an AutoML job.

  • KmsKeyId (string) --

    The Key Management Service encryption key ID.

  • S3OutputPath (string) -- [REQUIRED]

    The Amazon S3 output path. Must be 128 characters or less.

type AutoMLProblemTypeConfig:

dict

param AutoMLProblemTypeConfig:

[REQUIRED]

Defines the configuration settings of one of the supported problem types.

  • ImageClassificationJobConfig (dict) --

    Settings used to configure an AutoML job V2 for the image classification problem type.

    • CompletionCriteria (dict) --

      How long a job is allowed to run, or how many candidates a job is allowed to generate.

      • MaxCandidates (integer) --

        The maximum number of times a training job is allowed to run.

        For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

      • MaxRuntimePerTrainingJobInSeconds (integer) --

        The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

        For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

        For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

      • MaxAutoMLJobRuntimeInSeconds (integer) --

        The maximum runtime, in seconds, an AutoML job has to complete.

        If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

  • TextClassificationJobConfig (dict) --

    Settings used to configure an AutoML job V2 for the text classification problem type.

    • CompletionCriteria (dict) --

      How long a job is allowed to run, or how many candidates a job is allowed to generate.

      • MaxCandidates (integer) --

        The maximum number of times a training job is allowed to run.

        For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

      • MaxRuntimePerTrainingJobInSeconds (integer) --

        The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

        For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

        For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

      • MaxAutoMLJobRuntimeInSeconds (integer) --

        The maximum runtime, in seconds, an AutoML job has to complete.

        If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

    • ContentColumn (string) -- [REQUIRED]

      The name of the column used to provide the sentences to be classified. It should not be the same as the target column.

    • TargetLabelColumn (string) -- [REQUIRED]

      The name of the column used to provide the class labels. It should not be same as the content column.

  • TimeSeriesForecastingJobConfig (dict) --

    Settings used to configure an AutoML job V2 for the time-series forecasting problem type.

    • FeatureSpecificationS3Uri (string) --

      A URL to the Amazon S3 data source containing additional selected features that complement the target, itemID, timestamp, and grouped columns set in TimeSeriesConfig. When not provided, the AutoML job V2 includes all the columns from the original dataset that are not already declared in TimeSeriesConfig. If provided, the AutoML job V2 only considers these additional columns as a complement to the ones declared in TimeSeriesConfig.

      You can input FeatureAttributeNames (optional) in JSON format as shown below:

      { "FeatureAttributeNames":["col1", "col2", ...] }.

      You can also specify the data type of the feature (optional) in the format shown below:

      { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

      Autopilot supports the following data types: numeric, categorical, text, and datetime.

    • CompletionCriteria (dict) --

      How long a job is allowed to run, or how many candidates a job is allowed to generate.

      • MaxCandidates (integer) --

        The maximum number of times a training job is allowed to run.

        For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

      • MaxRuntimePerTrainingJobInSeconds (integer) --

        The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

        For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

        For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

      • MaxAutoMLJobRuntimeInSeconds (integer) --

        The maximum runtime, in seconds, an AutoML job has to complete.

        If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

    • ForecastFrequency (string) -- [REQUIRED]

      The frequency of predictions in a forecast.

      Valid intervals are an integer followed by Y (Year), M (Month), W (Week), D (Day), H (Hour), and min (Minute). For example, 1D indicates every day and 15min indicates every 15 minutes. The value of a frequency must not overlap with the next larger frequency. For example, you must use a frequency of 1H instead of 60min.

      The valid values for each frequency are the following:

      • Minute - 1-59

      • Hour - 1-23

      • Day - 1-6

      • Week - 1-4

      • Month - 1-11

      • Year - 1

    • ForecastHorizon (integer) -- [REQUIRED]

      The number of time-steps that the model predicts. The forecast horizon is also called the prediction length. The maximum forecast horizon is the lesser of 500 time-steps or 1/4 of the time-steps in the dataset.

    • ForecastQuantiles (list) --

      The quantiles used to train the model for forecasts at a specified quantile. You can specify quantiles from 0.01 (p1) to 0.99 (p99), by increments of 0.01 or higher. Up to five forecast quantiles can be specified. When ForecastQuantiles is not provided, the AutoML job uses the quantiles p10, p50, and p90 as default.

      • (string) --

    • Transformations (dict) --

      The transformations modifying specific attributes of the time-series, such as filling strategies for missing values.

      • Filling (dict) --

        A key value pair defining the filling method for a column, where the key is the column name and the value is an object which defines the filling logic. You can specify multiple filling methods for a single column.

        The supported filling methods and their corresponding options are:

        • frontfill: none (Supported only for target column)

        • middlefill: zero, value, median, mean, min, max

        • backfill: zero, value, median, mean, min, max

        • futurefill: zero, value, median, mean, min, max

        To set a filling method to a specific value, set the fill parameter to the chosen filling method value (for example "backfill" : "value"), and define the filling value in an additional parameter prefixed with "_value". For example, to set backfill to a value of 2, you must include two parameters: "backfill": "value" and "backfill_value":"2".

        • (string) --

          • (dict) --

            • (string) --

              • (string) --

      • Aggregation (dict) --

        A key value pair defining the aggregation method for a column, where the key is the column name and the value is the aggregation method.

        The supported aggregation methods are sum (default), avg, first, min, max.

        • (string) --

          • (string) --

    • TimeSeriesConfig (dict) -- [REQUIRED]

      The collection of components that defines the time-series.

      • TargetAttributeName (string) -- [REQUIRED]

        The name of the column representing the target variable that you want to predict for each item in your dataset. The data type of the target variable must be numerical.

      • TimestampAttributeName (string) -- [REQUIRED]

        The name of the column indicating a point in time at which the target value of a given item is recorded.

      • ItemIdentifierAttributeName (string) -- [REQUIRED]

        The name of the column that represents the set of item identifiers for which you want to predict the target value.

      • GroupingAttributeNames (list) --

        A set of columns names that can be grouped with the item identifier column to create a composite key for which a target value is predicted.

        • (string) --

    • HolidayConfig (list) --

      The collection of holiday featurization attributes used to incorporate national holiday information into your forecasting model.

      • (dict) --

        Stores the holiday featurization attributes applicable to each item of time-series datasets during the training of a forecasting model. This allows the model to identify patterns associated with specific holidays.

        • CountryCode (string) --

          The country code for the holiday calendar.

          For the list of public holiday calendars supported by AutoML job V2, see Country Codes. Use the country code corresponding to the country of your choice.

    • CandidateGenerationConfig (dict) --

      Stores the configuration information for how model candidates are generated using an AutoML job V2.

      • AlgorithmsConfig (list) --

        Your Autopilot job trains a default set of algorithms on your dataset. For tabular and time-series data, you can customize the algorithm list by selecting a subset of algorithms for your problem type.

        AlgorithmsConfig stores the customized selection of algorithms to train on your data.

        • For the tabular problem type TabularJobConfig, the list of available algorithms to choose from depends on the training mode set in AutoMLJobConfig.Mode.

          • AlgorithmsConfig should not be set when the training mode AutoMLJobConfig.Mode is set to AUTO.

          • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

          • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

        For the list of all algorithms per training mode, see AlgorithmConfig.

        For more information on each algorithm, see the Algorithm support section in the Autopilot developer guide.

        • For the time-series forecasting problem type TimeSeriesForecastingJobConfig, choose your algorithms from the list provided in AlgorithmConfig. For more information on each algorithm, see the Algorithms support for time-series forecasting section in the Autopilot developer guide.

          • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

          • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

        • (dict) --

          The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

          • AutoMLAlgorithms (list) -- [REQUIRED]

            The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

            • For the tabular problem type TabularJobConfig:

            • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

              • Choose your algorithms from this list.

                • "cnn-qr"

                • "deepar"

                • "prophet"

                • "arima"

                • "npts"

                • "ets"

            • (string) --

  • TabularJobConfig (dict) --

    Settings used to configure an AutoML job V2 for the tabular problem type (regression, classification).

    • CandidateGenerationConfig (dict) --

      The configuration information of how model candidates are generated.

      • AlgorithmsConfig (list) --

        Your Autopilot job trains a default set of algorithms on your dataset. For tabular and time-series data, you can customize the algorithm list by selecting a subset of algorithms for your problem type.

        AlgorithmsConfig stores the customized selection of algorithms to train on your data.

        • For the tabular problem type TabularJobConfig, the list of available algorithms to choose from depends on the training mode set in AutoMLJobConfig.Mode.

          • AlgorithmsConfig should not be set when the training mode AutoMLJobConfig.Mode is set to AUTO.

          • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

          • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

        For the list of all algorithms per training mode, see AlgorithmConfig.

        For more information on each algorithm, see the Algorithm support section in the Autopilot developer guide.

        • For the time-series forecasting problem type TimeSeriesForecastingJobConfig, choose your algorithms from the list provided in AlgorithmConfig. For more information on each algorithm, see the Algorithms support for time-series forecasting section in the Autopilot developer guide.

          • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

          • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

        • (dict) --

          The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

          • AutoMLAlgorithms (list) -- [REQUIRED]

            The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

            • For the tabular problem type TabularJobConfig:

            • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

              • Choose your algorithms from this list.

                • "cnn-qr"

                • "deepar"

                • "prophet"

                • "arima"

                • "npts"

                • "ets"

            • (string) --

    • CompletionCriteria (dict) --

      How long a job is allowed to run, or how many candidates a job is allowed to generate.

      • MaxCandidates (integer) --

        The maximum number of times a training job is allowed to run.

        For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

      • MaxRuntimePerTrainingJobInSeconds (integer) --

        The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

        For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

        For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

      • MaxAutoMLJobRuntimeInSeconds (integer) --

        The maximum runtime, in seconds, an AutoML job has to complete.

        If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

    • FeatureSpecificationS3Uri (string) --

      A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job V2. You can input FeatureAttributeNames (optional) in JSON format as shown below:

      { "FeatureAttributeNames":["col1", "col2", ...] }.

      You can also specify the data type of the feature (optional) in the format shown below:

      { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

      In ensembling mode, Autopilot only supports the following data types: numeric, categorical, text, and datetime. In HPO mode, Autopilot can support numeric, categorical, text, datetime, and sequence.

      If only FeatureDataTypes is provided, the column keys ( col1, col2,..) should be a subset of the column names in the input data.

      If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames.

      The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.

    • Mode (string) --

      The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO. In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.

      The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.

      The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.

    • GenerateCandidateDefinitionsOnly (boolean) --

      Generates possible candidates without training the models. A model candidate is a combination of data preprocessors, algorithms, and algorithm parameter settings.

    • ProblemType (string) --

      The type of supervised learning problem available for the model candidates of the AutoML job V2. For more information, see SageMaker Autopilot problem types.

    • TargetAttributeName (string) -- [REQUIRED]

      The name of the target variable in supervised learning, usually represented by 'y'.

    • SampleWeightAttributeName (string) --

      If specified, this column name indicates which column of the dataset should be treated as sample weights for use by the objective metric during the training, evaluation, and the selection of the best model. This column is not considered as a predictive feature. For more information on Autopilot metrics, see Metrics and validation.

      Sample weights should be numeric, non-negative, with larger values indicating which rows are more important than others. Data points that have invalid or no weight value are excluded.

      Support for sample weights is available in Ensembling mode only.

  • TextGenerationJobConfig (dict) --

    Settings used to configure an AutoML job V2 for the text generation (LLMs fine-tuning) problem type.

    • CompletionCriteria (dict) --

      How long a fine-tuning job is allowed to run. For TextGenerationJobConfig problem types, the MaxRuntimePerTrainingJobInSeconds attribute of AutoMLJobCompletionCriteria defaults to 72h (259200s).

      • MaxCandidates (integer) --

        The maximum number of times a training job is allowed to run.

        For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

      • MaxRuntimePerTrainingJobInSeconds (integer) --

        The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

        For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

        For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

      • MaxAutoMLJobRuntimeInSeconds (integer) --

        The maximum runtime, in seconds, an AutoML job has to complete.

        If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

    • BaseModelName (string) --

      The name of the base model to fine-tune. Autopilot supports fine-tuning a variety of large language models. For information on the list of supported models, see Text generation models supporting fine-tuning in Autopilot. If no BaseModelName is provided, the default model used is Falcon7BInstruct.

    • TextGenerationHyperParameters (dict) --

      The hyperparameters used to configure and optimize the learning process of the base model. You can set any combination of the following hyperparameters for all base models. For more information on each supported hyperparameter, see Optimize the learning process of your text generation models with hyperparameters.

      • "epochCount": The number of times the model goes through the entire training dataset. Its value should be a string containing an integer value within the range of "1" to "10".

      • "batchSize": The number of data samples used in each iteration of training. Its value should be a string containing an integer value within the range of "1" to "64".

      • "learningRate": The step size at which a model's parameters are updated during training. Its value should be a string containing a floating-point value within the range of "0" to "1".

      • "learningRateWarmupSteps": The number of training steps during which the learning rate gradually increases before reaching its target or maximum value. Its value should be a string containing an integer value within the range of "0" to "250".

      Here is an example where all four hyperparameters are configured.

      { "epochCount":"5", "learningRate":"0.5", "batchSize": "32", "learningRateWarmupSteps": "10" }

      • (string) --

        • (string) --

    • ModelAccessConfig (dict) --

      The access configuration file to control access to the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig.

      • AcceptEula (boolean) -- [REQUIRED]

        Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

type RoleArn:

string

param RoleArn:

[REQUIRED]

The ARN of the role that is used to access the data.

type Tags:

list

param Tags:

An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, such as by purpose, owner, or environment. For more information, see Tagging Amazon Web ServicesResources. Tag keys must be unique per resource.

  • (dict) --

    A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.

    You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.

    For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.

    • Key (string) -- [REQUIRED]

      The tag key. Tag keys must be unique per resource.

    • Value (string) -- [REQUIRED]

      The tag value.

type SecurityConfig:

dict

param SecurityConfig:

The security configuration for traffic encryption or Amazon VPC settings.

  • VolumeKmsKeyId (string) --

    The key used to encrypt stored data.

  • EnableInterContainerTrafficEncryption (boolean) --

    Whether to use traffic encryption between the container layers.

  • VpcConfig (dict) --

    The VPC configuration.

    • SecurityGroupIds (list) -- [REQUIRED]

      The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.

      • (string) --

    • Subnets (list) -- [REQUIRED]

      The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.

      • (string) --

type AutoMLJobObjective:

dict

param AutoMLJobObjective:

Specifies a metric to minimize or maximize as the objective of a job. If not specified, the default objective metric depends on the problem type. For the list of default values per problem type, see AutoMLJobObjective.

  • MetricName (string) -- [REQUIRED]

    The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

    The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

    • For tabular problem types:

      • List of available metrics:

        • Regression: MAE, MSE, R2, RMSE

        • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

        • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

      For a description of each metric, see Autopilot metrics for classification and regression.

      • Default objective metrics:

        • Regression: MSE.

        • Binary classification: F1.

        • Multiclass classification: Accuracy.

    • For image or text classification problem types:

    • For time-series forecasting problem types:

    • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

type ModelDeployConfig:

dict

param ModelDeployConfig:

Specifies how to generate the endpoint name for an automatic one-click Autopilot model deployment.

  • AutoGenerateEndpointName (boolean) --

    Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False.

  • EndpointName (string) --

    Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.

type DataSplitConfig:

dict

param DataSplitConfig:

This structure specifies how to split the data into train and validation datasets.

The validation and training datasets must contain the same headers. For jobs created by calling CreateAutoMLJob, the validation dataset must be less than 2 GB in size.

  • ValidationFraction (float) --

    The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.

rtype:

dict

returns:

Response Syntax

{
    'AutoMLJobArn': 'string'
}

Response Structure

  • (dict) --

    • AutoMLJobArn (string) --

      The unique ARN assigned to the AutoMLJob when it is created.

CreateModelPackage (updated) Link ¶
Changes (request)
{'ModelCard': {'ModelCardContent': 'string',
               'ModelCardStatus': 'Draft | PendingReview | Approved | '
                                  'Archived'},
 'SecurityConfig': {'KmsKeyId': 'string'}}

Creates a model package that you can use to create SageMaker models or list on Amazon Web Services Marketplace, or a versioned model that is part of a model group. Buyers can subscribe to model packages listed on Amazon Web Services Marketplace to create models in SageMaker.

To create a model package by specifying a Docker container that contains your inference code and the Amazon S3 location of your model artifacts, provide values for InferenceSpecification. To create a model from an algorithm resource that you created or subscribed to in Amazon Web Services Marketplace, provide a value for SourceAlgorithmSpecification.

See also: AWS API Documentation

Request Syntax

client.create_model_package(
    ModelPackageName='string',
    ModelPackageGroupName='string',
    ModelPackageDescription='string',
    InferenceSpecification={
        'Containers': [
            {
                'ContainerHostname': 'string',
                'Image': 'string',
                'ImageDigest': 'string',
                'ModelDataUrl': 'string',
                'ModelDataSource': {
                    'S3DataSource': {
                        'S3Uri': 'string',
                        'S3DataType': 'S3Prefix'|'S3Object',
                        'CompressionType': 'None'|'Gzip',
                        'ModelAccessConfig': {
                            'AcceptEula': True|False
                        }
                    }
                },
                'ProductId': 'string',
                'Environment': {
                    'string': 'string'
                },
                'ModelInput': {
                    'DataInputConfig': 'string'
                },
                'Framework': 'string',
                'FrameworkVersion': 'string',
                'NearestModelName': 'string',
                'AdditionalS3DataSource': {
                    'S3DataType': 'S3Object'|'S3Prefix',
                    'S3Uri': 'string',
                    'CompressionType': 'None'|'Gzip'
                }
            },
        ],
        'SupportedTransformInstanceTypes': [
            'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
        ],
        'SupportedRealtimeInferenceInstanceTypes': [
            'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
        ],
        'SupportedContentTypes': [
            'string',
        ],
        'SupportedResponseMIMETypes': [
            'string',
        ]
    },
    ValidationSpecification={
        'ValidationRole': 'string',
        'ValidationProfiles': [
            {
                'ProfileName': 'string',
                'TransformJobDefinition': {
                    'MaxConcurrentTransforms': 123,
                    'MaxPayloadInMB': 123,
                    'BatchStrategy': 'MultiRecord'|'SingleRecord',
                    'Environment': {
                        'string': 'string'
                    },
                    'TransformInput': {
                        'DataSource': {
                            'S3DataSource': {
                                'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                                'S3Uri': 'string'
                            }
                        },
                        'ContentType': 'string',
                        'CompressionType': 'None'|'Gzip',
                        'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord'
                    },
                    'TransformOutput': {
                        'S3OutputPath': 'string',
                        'Accept': 'string',
                        'AssembleWith': 'None'|'Line',
                        'KmsKeyId': 'string'
                    },
                    'TransformResources': {
                        'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
                        'InstanceCount': 123,
                        'VolumeKmsKeyId': 'string'
                    }
                }
            },
        ]
    },
    SourceAlgorithmSpecification={
        'SourceAlgorithms': [
            {
                'ModelDataUrl': 'string',
                'ModelDataSource': {
                    'S3DataSource': {
                        'S3Uri': 'string',
                        'S3DataType': 'S3Prefix'|'S3Object',
                        'CompressionType': 'None'|'Gzip',
                        'ModelAccessConfig': {
                            'AcceptEula': True|False
                        }
                    }
                },
                'AlgorithmName': 'string'
            },
        ]
    },
    CertifyForMarketplace=True|False,
    Tags=[
        {
            'Key': 'string',
            'Value': 'string'
        },
    ],
    ModelApprovalStatus='Approved'|'Rejected'|'PendingManualApproval',
    MetadataProperties={
        'CommitId': 'string',
        'Repository': 'string',
        'GeneratedBy': 'string',
        'ProjectId': 'string'
    },
    ModelMetrics={
        'ModelQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelDataQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Bias': {
            'Report': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PreTrainingReport': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PostTrainingReport': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Explainability': {
            'Report': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        }
    },
    ClientToken='string',
    Domain='string',
    Task='string',
    SamplePayloadUrl='string',
    CustomerMetadataProperties={
        'string': 'string'
    },
    DriftCheckBaselines={
        'Bias': {
            'ConfigFile': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PreTrainingConstraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PostTrainingConstraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Explainability': {
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'ConfigFile': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelDataQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        }
    },
    AdditionalInferenceSpecifications=[
        {
            'Name': 'string',
            'Description': 'string',
            'Containers': [
                {
                    'ContainerHostname': 'string',
                    'Image': 'string',
                    'ImageDigest': 'string',
                    'ModelDataUrl': 'string',
                    'ModelDataSource': {
                        'S3DataSource': {
                            'S3Uri': 'string',
                            'S3DataType': 'S3Prefix'|'S3Object',
                            'CompressionType': 'None'|'Gzip',
                            'ModelAccessConfig': {
                                'AcceptEula': True|False
                            }
                        }
                    },
                    'ProductId': 'string',
                    'Environment': {
                        'string': 'string'
                    },
                    'ModelInput': {
                        'DataInputConfig': 'string'
                    },
                    'Framework': 'string',
                    'FrameworkVersion': 'string',
                    'NearestModelName': 'string',
                    'AdditionalS3DataSource': {
                        'S3DataType': 'S3Object'|'S3Prefix',
                        'S3Uri': 'string',
                        'CompressionType': 'None'|'Gzip'
                    }
                },
            ],
            'SupportedTransformInstanceTypes': [
                'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
            ],
            'SupportedRealtimeInferenceInstanceTypes': [
                'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
            ],
            'SupportedContentTypes': [
                'string',
            ],
            'SupportedResponseMIMETypes': [
                'string',
            ]
        },
    ],
    SkipModelValidation='All'|'None',
    SourceUri='string',
    SecurityConfig={
        'KmsKeyId': 'string'
    },
    ModelCard={
        'ModelCardContent': 'string',
        'ModelCardStatus': 'Draft'|'PendingReview'|'Approved'|'Archived'
    }
)
type ModelPackageName:

string

param ModelPackageName:

The name of the model package. The name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).

This parameter is required for unversioned models. It is not applicable to versioned models.

type ModelPackageGroupName:

string

param ModelPackageGroupName:

The name or Amazon Resource Name (ARN) of the model package group that this model version belongs to.

This parameter is required for versioned models, and does not apply to unversioned models.

type ModelPackageDescription:

string

param ModelPackageDescription:

A description of the model package.

type InferenceSpecification:

dict

param InferenceSpecification:

Specifies details about inference jobs that you can run with models based on this model package, including the following information:

  • The Amazon ECR paths of containers that contain the inference code and model artifacts.

  • The instance types that the model package supports for transform jobs and real-time endpoints used for inference.

  • The input and output content formats that the model package supports for inference.

  • Containers (list) -- [REQUIRED]

    The Amazon ECR registry path of the Docker image that contains the inference code.

    • (dict) --

      Describes the Docker container for the model package.

      • ContainerHostname (string) --

        The DNS host name for the Docker container.

      • Image (string) -- [REQUIRED]

        The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

        If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

      • ImageDigest (string) --

        An MD5 hash of the training algorithm that identifies the Docker image used for training.

      • ModelDataUrl (string) --

        The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

      • ModelDataSource (dict) --

        Specifies the location of ML model data to deploy during endpoint creation.

        • S3DataSource (dict) --

          Specifies the S3 location of ML model data to deploy.

          • S3Uri (string) -- [REQUIRED]

            Specifies the S3 path of ML model data to deploy.

          • S3DataType (string) -- [REQUIRED]

            Specifies the type of ML model data to deploy.

            If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

            If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

          • CompressionType (string) -- [REQUIRED]

            Specifies how the ML model data is prepared.

            If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

            If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

            If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

            If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

            • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

            • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

            • Do not use any of the following as file names or directory names:

              • An empty or blank string

              • A string which contains null bytes

              • A string longer than 255 bytes

              • A single dot ( .)

              • A double dot ( ..)

            • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

            • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

          • ModelAccessConfig (dict) --

            Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

            • AcceptEula (boolean) -- [REQUIRED]

              Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

      • ProductId (string) --

        The Amazon Web Services Marketplace product ID of the model package.

      • Environment (dict) --

        The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

        • (string) --

          • (string) --

      • ModelInput (dict) --

        A structure with Model Input details.

        • DataInputConfig (string) -- [REQUIRED]

          The input configuration object for the model.

      • Framework (string) --

        The machine learning framework of the model package container image.

      • FrameworkVersion (string) --

        The framework version of the Model Package Container Image.

      • NearestModelName (string) --

        The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

      • AdditionalS3DataSource (dict) --

        The additional data source that is used during inference in the Docker container for your model package.

        • S3DataType (string) -- [REQUIRED]

          The data type of the additional data source that you specify for use in inference or training.

        • S3Uri (string) -- [REQUIRED]

          The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

        • CompressionType (string) --

          The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

  • SupportedTransformInstanceTypes (list) --

    A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

    This parameter is required for unversioned models, and optional for versioned models.

    • (string) --

  • SupportedRealtimeInferenceInstanceTypes (list) --

    A list of the instance types that are used to generate inferences in real-time.

    This parameter is required for unversioned models, and optional for versioned models.

    • (string) --

  • SupportedContentTypes (list) --

    The supported MIME types for the input data.

    • (string) --

  • SupportedResponseMIMETypes (list) --

    The supported MIME types for the output data.

    • (string) --

type ValidationSpecification:

dict

param ValidationSpecification:

Specifies configurations for one or more transform jobs that SageMaker runs to test the model package.

  • ValidationRole (string) -- [REQUIRED]

    The IAM roles to be used for the validation of the model package.

  • ValidationProfiles (list) -- [REQUIRED]

    An array of ModelPackageValidationProfile objects, each of which specifies a batch transform job that SageMaker runs to validate your model package.

    • (dict) --

      Contains data, such as the inputs and targeted instance types that are used in the process of validating the model package.

      The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.

      • ProfileName (string) -- [REQUIRED]

        The name of the profile for the model package.

      • TransformJobDefinition (dict) -- [REQUIRED]

        The TransformJobDefinition object that describes the transform job used for the validation of the model package.

        • MaxConcurrentTransforms (integer) --

          The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.

        • MaxPayloadInMB (integer) --

          The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).

        • BatchStrategy (string) --

          A string that determines the number of records included in a single mini-batch.

          SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.

        • Environment (dict) --

          The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.

          • (string) --

            • (string) --

        • TransformInput (dict) -- [REQUIRED]

          A description of the input source and the way the transform job consumes it.

          • DataSource (dict) -- [REQUIRED]

            Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.

            • S3DataSource (dict) -- [REQUIRED]

              The S3 location of the data source that is associated with a channel.

              • S3DataType (string) -- [REQUIRED]

                If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.

                If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.

                The following values are compatible: ManifestFile, S3Prefix

                The following value is not compatible: AugmentedManifestFile

              • S3Uri (string) -- [REQUIRED]

                Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:

                • A key name prefix might look like this: s3://bucketname/exampleprefix/.

                • A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.

          • ContentType (string) --

            The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.

          • CompressionType (string) --

            If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.

          • SplitType (string) --

            The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:

            • RecordIO

            • TFRecord

            When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.

        • TransformOutput (dict) -- [REQUIRED]

          Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.

          • S3OutputPath (string) -- [REQUIRED]

            The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.

            For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.

          • Accept (string) --

            The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.

          • AssembleWith (string) --

            Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.

          • KmsKeyId (string) --

            The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:

            • Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab

            • Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab

            • Alias name: alias/ExampleAlias

            • Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias

            If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.

            The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.

        • TransformResources (dict) -- [REQUIRED]

          Identifies the ML compute instances for the transform job.

          • InstanceType (string) -- [REQUIRED]

            The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.

          • InstanceCount (integer) -- [REQUIRED]

            The number of ML compute instances to use in the transform job. The default value is 1, and the maximum is 100. For distributed transform jobs, specify a value greater than 1.

          • VolumeKmsKeyId (string) --

            The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.

            The VolumeKmsKeyId can be any of the following formats:

            • Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab

            • Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab

            • Alias name: alias/ExampleAlias

            • Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias

type SourceAlgorithmSpecification:

dict

param SourceAlgorithmSpecification:

Details about the algorithm that was used to create the model package.

  • SourceAlgorithms (list) -- [REQUIRED]

    A list of the algorithms that were used to create a model package.

    • (dict) --

      Specifies an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.

      • ModelDataUrl (string) --

        The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

      • ModelDataSource (dict) --

        Specifies the location of ML model data to deploy during endpoint creation.

        • S3DataSource (dict) --

          Specifies the S3 location of ML model data to deploy.

          • S3Uri (string) -- [REQUIRED]

            Specifies the S3 path of ML model data to deploy.

          • S3DataType (string) -- [REQUIRED]

            Specifies the type of ML model data to deploy.

            If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

            If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

          • CompressionType (string) -- [REQUIRED]

            Specifies how the ML model data is prepared.

            If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

            If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

            If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

            If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

            • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

            • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

            • Do not use any of the following as file names or directory names:

              • An empty or blank string

              • A string which contains null bytes

              • A string longer than 255 bytes

              • A single dot ( .)

              • A double dot ( ..)

            • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

            • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

          • ModelAccessConfig (dict) --

            Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

            • AcceptEula (boolean) -- [REQUIRED]

              Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

      • AlgorithmName (string) -- [REQUIRED]

        The name of an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.

type CertifyForMarketplace:

boolean

param CertifyForMarketplace:

Whether to certify the model package for listing on Amazon Web Services Marketplace.

This parameter is optional for unversioned models, and does not apply to versioned models.

type Tags:

list

param Tags:

A list of key value pairs associated with the model. For more information, see Tagging Amazon Web Services resources in the Amazon Web Services General Reference Guide.

If you supply ModelPackageGroupName, your model package belongs to the model group you specify and uses the tags associated with the model group. In this case, you cannot supply a tag argument.

  • (dict) --

    A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.

    You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags.

    For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources. For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy.

    • Key (string) -- [REQUIRED]

      The tag key. Tag keys must be unique per resource.

    • Value (string) -- [REQUIRED]

      The tag value.

type ModelApprovalStatus:

string

param ModelApprovalStatus:

Whether the model is approved for deployment.

This parameter is optional for versioned models, and does not apply to unversioned models.

For versioned models, the value of this parameter must be set to Approved to deploy the model.

type MetadataProperties:

dict

param MetadataProperties:

Metadata properties of the tracking entity, trial, or trial component.

  • CommitId (string) --

    The commit ID.

  • Repository (string) --

    The repository.

  • GeneratedBy (string) --

    The entity this entity was generated by.

  • ProjectId (string) --

    The project ID.

type ModelMetrics:

dict

param ModelMetrics:

A structure that contains model metrics reports.

  • ModelQuality (dict) --

    Metrics that measure the quality of a model.

    • Statistics (dict) --

      Model quality statistics.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • Constraints (dict) --

      Model quality constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

  • ModelDataQuality (dict) --

    Metrics that measure the quality of the input data for a model.

    • Statistics (dict) --

      Data quality statistics for a model.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • Constraints (dict) --

      Data quality constraints for a model.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

  • Bias (dict) --

    Metrics that measure bias in a model.

    • Report (dict) --

      The bias report for a model

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • PreTrainingReport (dict) --

      The pre-training bias report for a model.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • PostTrainingReport (dict) --

      The post-training bias report for a model.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

  • Explainability (dict) --

    Metrics that help explain a model.

    • Report (dict) --

      The explainability report for a model.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

type ClientToken:

string

param ClientToken:

A unique token that guarantees that the call to this API is idempotent.

This field is autopopulated if not provided.

type Domain:

string

param Domain:

The machine learning domain of your model package and its components. Common machine learning domains include computer vision and natural language processing.

type Task:

string

param Task:

The machine learning task your model package accomplishes. Common machine learning tasks include object detection and image classification. The following tasks are supported by Inference Recommender: "IMAGE_CLASSIFICATION" | "OBJECT_DETECTION" | "TEXT_GENERATION" | "IMAGE_SEGMENTATION" | "FILL_MASK" | "CLASSIFICATION" | "REGRESSION" | "OTHER".

Specify "OTHER" if none of the tasks listed fit your use case.

type SamplePayloadUrl:

string

param SamplePayloadUrl:

The Amazon Simple Storage Service (Amazon S3) path where the sample payload is stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix). This archive can hold multiple files that are all equally used in the load test. Each file in the archive must satisfy the size constraints of the InvokeEndpoint call.

type CustomerMetadataProperties:

dict

param CustomerMetadataProperties:

The metadata properties associated with the model package versions.

  • (string) --

    • (string) --

type DriftCheckBaselines:

dict

param DriftCheckBaselines:

Represents the drift check baselines that can be used when the model monitor is set using the model package. For more information, see the topic on Drift Detection against Previous Baselines in SageMaker Pipelines in the Amazon SageMaker Developer Guide.

  • Bias (dict) --

    Represents the drift check bias baselines that can be used when the model monitor is set using the model package.

    • ConfigFile (dict) --

      The bias config file for a model.

      • ContentType (string) --

        The type of content stored in the file source.

      • ContentDigest (string) --

        The digest of the file source.

      • S3Uri (string) -- [REQUIRED]

        The Amazon S3 URI for the file source.

    • PreTrainingConstraints (dict) --

      The pre-training constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • PostTrainingConstraints (dict) --

      The post-training constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

  • Explainability (dict) --

    Represents the drift check explainability baselines that can be used when the model monitor is set using the model package.

    • Constraints (dict) --

      The drift check explainability constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • ConfigFile (dict) --

      The explainability config file for the model.

      • ContentType (string) --

        The type of content stored in the file source.

      • ContentDigest (string) --

        The digest of the file source.

      • S3Uri (string) -- [REQUIRED]

        The Amazon S3 URI for the file source.

  • ModelQuality (dict) --

    Represents the drift check model quality baselines that can be used when the model monitor is set using the model package.

    • Statistics (dict) --

      The drift check model quality statistics.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • Constraints (dict) --

      The drift check model quality constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

  • ModelDataQuality (dict) --

    Represents the drift check model data quality baselines that can be used when the model monitor is set using the model package.

    • Statistics (dict) --

      The drift check model data quality statistics.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

    • Constraints (dict) --

      The drift check model data quality constraints.

      • ContentType (string) -- [REQUIRED]

        The metric source content type.

      • ContentDigest (string) --

        The hash key used for the metrics source.

      • S3Uri (string) -- [REQUIRED]

        The S3 URI for the metrics source.

type AdditionalInferenceSpecifications:

list

param AdditionalInferenceSpecifications:

An array of additional Inference Specification objects. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.

  • (dict) --

    A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package

    • Name (string) -- [REQUIRED]

      A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.

    • Description (string) --

      A description of the additional Inference specification

    • Containers (list) -- [REQUIRED]

      The Amazon ECR registry path of the Docker image that contains the inference code.

      • (dict) --

        Describes the Docker container for the model package.

        • ContainerHostname (string) --

          The DNS host name for the Docker container.

        • Image (string) -- [REQUIRED]

          The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

          If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

        • ImageDigest (string) --

          An MD5 hash of the training algorithm that identifies the Docker image used for training.

        • ModelDataUrl (string) --

          The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

        • ModelDataSource (dict) --

          Specifies the location of ML model data to deploy during endpoint creation.

          • S3DataSource (dict) --

            Specifies the S3 location of ML model data to deploy.

            • S3Uri (string) -- [REQUIRED]

              Specifies the S3 path of ML model data to deploy.

            • S3DataType (string) -- [REQUIRED]

              Specifies the type of ML model data to deploy.

              If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

              If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

            • CompressionType (string) -- [REQUIRED]

              Specifies how the ML model data is prepared.

              If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

              If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

              If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

              If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

              • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

              • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

              • Do not use any of the following as file names or directory names:

                • An empty or blank string

                • A string which contains null bytes

                • A string longer than 255 bytes

                • A single dot ( .)

                • A double dot ( ..)

              • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

              • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

            • ModelAccessConfig (dict) --

              Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

              • AcceptEula (boolean) -- [REQUIRED]

                Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

        • ProductId (string) --

          The Amazon Web Services Marketplace product ID of the model package.

        • Environment (dict) --

          The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

          • (string) --

            • (string) --

        • ModelInput (dict) --

          A structure with Model Input details.

          • DataInputConfig (string) -- [REQUIRED]

            The input configuration object for the model.

        • Framework (string) --

          The machine learning framework of the model package container image.

        • FrameworkVersion (string) --

          The framework version of the Model Package Container Image.

        • NearestModelName (string) --

          The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

        • AdditionalS3DataSource (dict) --

          The additional data source that is used during inference in the Docker container for your model package.

          • S3DataType (string) -- [REQUIRED]

            The data type of the additional data source that you specify for use in inference or training.

          • S3Uri (string) -- [REQUIRED]

            The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

          • CompressionType (string) --

            The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

    • SupportedTransformInstanceTypes (list) --

      A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

      • (string) --

    • SupportedRealtimeInferenceInstanceTypes (list) --

      A list of the instance types that are used to generate inferences in real-time.

      • (string) --

    • SupportedContentTypes (list) --

      The supported MIME types for the input data.

      • (string) --

    • SupportedResponseMIMETypes (list) --

      The supported MIME types for the output data.

      • (string) --

type SkipModelValidation:

string

param SkipModelValidation:

Indicates if you want to skip model validation.

type SourceUri:

string

param SourceUri:

The URI of the source for the model package. If you want to clone a model package, set it to the model package Amazon Resource Name (ARN). If you want to register a model, set it to the model ARN.

type SecurityConfig:

dict

param SecurityConfig:

The KMS Key ID ( KMSKeyId) used for encryption of model package information.

  • KmsKeyId (string) -- [REQUIRED]

    The KMS Key ID ( KMSKeyId) used for encryption of model package information.

type ModelCard:

dict

param ModelCard:

The model card associated with the model package. Since ModelPackageModelCard is tied to a model package, it is a specific usage of a model card and its schema is simplified compared to the schema of ModelCard. The ModelPackageModelCard schema does not include model_package_details, and model_overview is composed of the model_creator and model_artifact properties. For more information about the model card associated with the model package, see View the Details of a Model Version.

  • ModelCardContent (string) --

    The content of the model card.

  • ModelCardStatus (string) --

    The approval status of the model card within your organization. Different organizations might have different criteria for model card review and approval.

    • Draft: The model card is a work in progress.

    • PendingReview: The model card is pending review.

    • Approved: The model card is approved.

    • Archived: The model card is archived. No more updates can be made to the model card content. If you try to update the model card content, you will receive the message Model Card is in Archived state.

rtype:

dict

returns:

Response Syntax

{
    'ModelPackageArn': 'string'
}

Response Structure

  • (dict) --

    • ModelPackageArn (string) --

      The Amazon Resource Name (ARN) of the new model package.

DescribeAutoMLJob (updated) Link ¶
Changes (response)
{'AutoMLJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': {'AutoMLAlgorithms': {'arima',
                                                                                             'cnn-qr',
                                                                                             'deepar',
                                                                                             'ets',
                                                                                             'npts',
                                                                                             'prophet'}}}}}

Returns information about an AutoML job created by calling CreateAutoMLJob.

See also: AWS API Documentation

Request Syntax

client.describe_auto_ml_job(
    AutoMLJobName='string'
)
type AutoMLJobName:

string

param AutoMLJobName:

[REQUIRED]

Requests information about an AutoML job using its unique name.

rtype:

dict

returns:

Response Syntax

{
    'AutoMLJobName': 'string',
    'AutoMLJobArn': 'string',
    'InputDataConfig': [
        {
            'DataSource': {
                'S3DataSource': {
                    'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                    'S3Uri': 'string'
                }
            },
            'CompressionType': 'None'|'Gzip',
            'TargetAttributeName': 'string',
            'ContentType': 'string',
            'ChannelType': 'training'|'validation',
            'SampleWeightAttributeName': 'string'
        },
    ],
    'OutputDataConfig': {
        'KmsKeyId': 'string',
        'S3OutputPath': 'string'
    },
    'RoleArn': 'string',
    'AutoMLJobObjective': {
        'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
    },
    'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression',
    'AutoMLJobConfig': {
        'CompletionCriteria': {
            'MaxCandidates': 123,
            'MaxRuntimePerTrainingJobInSeconds': 123,
            'MaxAutoMLJobRuntimeInSeconds': 123
        },
        'SecurityConfig': {
            'VolumeKmsKeyId': 'string',
            'EnableInterContainerTrafficEncryption': True|False,
            'VpcConfig': {
                'SecurityGroupIds': [
                    'string',
                ],
                'Subnets': [
                    'string',
                ]
            }
        },
        'CandidateGenerationConfig': {
            'FeatureSpecificationS3Uri': 'string',
            'AlgorithmsConfig': [
                {
                    'AutoMLAlgorithms': [
                        'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                    ]
                },
            ]
        },
        'DataSplitConfig': {
            'ValidationFraction': ...
        },
        'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING'
    },
    'CreationTime': datetime(2015, 1, 1),
    'EndTime': datetime(2015, 1, 1),
    'LastModifiedTime': datetime(2015, 1, 1),
    'FailureReason': 'string',
    'PartialFailureReasons': [
        {
            'PartialFailureMessage': 'string'
        },
    ],
    'BestCandidate': {
        'CandidateName': 'string',
        'FinalAutoMLJobObjectiveMetric': {
            'Type': 'Maximize'|'Minimize',
            'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss',
            'Value': ...,
            'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
        },
        'ObjectiveStatus': 'Succeeded'|'Pending'|'Failed',
        'CandidateSteps': [
            {
                'CandidateStepType': 'AWS::SageMaker::TrainingJob'|'AWS::SageMaker::TransformJob'|'AWS::SageMaker::ProcessingJob',
                'CandidateStepArn': 'string',
                'CandidateStepName': 'string'
            },
        ],
        'CandidateStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping',
        'InferenceContainers': [
            {
                'Image': 'string',
                'ModelDataUrl': 'string',
                'Environment': {
                    'string': 'string'
                }
            },
        ],
        'CreationTime': datetime(2015, 1, 1),
        'EndTime': datetime(2015, 1, 1),
        'LastModifiedTime': datetime(2015, 1, 1),
        'FailureReason': 'string',
        'CandidateProperties': {
            'CandidateArtifactLocations': {
                'Explainability': 'string',
                'ModelInsights': 'string',
                'BacktestResults': 'string'
            },
            'CandidateMetrics': [
                {
                    'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss',
                    'Value': ...,
                    'Set': 'Train'|'Validation'|'Test',
                    'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro'|'LogLoss'|'InferenceLatency'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'|'Rouge1'|'Rouge2'|'RougeL'|'RougeLSum'|'Perplexity'|'ValidationLoss'|'TrainingLoss'
                },
            ]
        },
        'InferenceContainerDefinitions': {
            'string': [
                {
                    'Image': 'string',
                    'ModelDataUrl': 'string',
                    'Environment': {
                        'string': 'string'
                    }
                },
            ]
        }
    },
    'AutoMLJobStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping',
    'AutoMLJobSecondaryStatus': 'Starting'|'MaxCandidatesReached'|'Failed'|'Stopped'|'MaxAutoMLJobRuntimeReached'|'Stopping'|'CandidateDefinitionsGenerated'|'Completed'|'ExplainabilityError'|'DeployingModel'|'ModelDeploymentError'|'GeneratingModelInsightsReport'|'ModelInsightsError'|'AnalyzingData'|'FeatureEngineering'|'ModelTuning'|'GeneratingExplainabilityReport'|'TrainingModels'|'PreTraining',
    'GenerateCandidateDefinitionsOnly': True|False,
    'AutoMLJobArtifacts': {
        'CandidateDefinitionNotebookLocation': 'string',
        'DataExplorationNotebookLocation': 'string'
    },
    'ResolvedAttributes': {
        'AutoMLJobObjective': {
            'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
        },
        'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression',
        'CompletionCriteria': {
            'MaxCandidates': 123,
            'MaxRuntimePerTrainingJobInSeconds': 123,
            'MaxAutoMLJobRuntimeInSeconds': 123
        }
    },
    'ModelDeployConfig': {
        'AutoGenerateEndpointName': True|False,
        'EndpointName': 'string'
    },
    'ModelDeployResult': {
        'EndpointName': 'string'
    }
}

Response Structure

  • (dict) --

    • AutoMLJobName (string) --

      Returns the name of the AutoML job.

    • AutoMLJobArn (string) --

      Returns the ARN of the AutoML job.

    • InputDataConfig (list) --

      Returns the input data configuration for the AutoML job.

      • (dict) --

        A channel is a named input source that training algorithms can consume. The validation dataset size is limited to less than 2 GB. The training dataset size must be less than 100 GB. For more information, see Channel.

        • DataSource (dict) --

          The data source for an AutoML channel.

          • S3DataSource (dict) --

            The Amazon S3 location of the input data.

            • S3DataType (string) --

              The data type.

              • If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix for model training. The S3Prefix should have the following format: s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE

              • If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want SageMaker to use for model training. A ManifestFile should have the format shown below: [ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"}, "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1", "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2", ... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]

              • If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile is available for V2 API jobs only (for example, for jobs created by calling CreateAutoMLJobV2). Here is a minimal, single-record example of an AugmentedManifestFile: {"source-ref": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/cats/cat.jpg", "label-metadata": {"class-name": "cat" } For more information on AugmentedManifestFile, see Provide Dataset Metadata to Training Jobs with an Augmented Manifest File.

            • S3Uri (string) --

              The URL to the Amazon S3 data source. The Uri refers to the Amazon S3 prefix or ManifestFile depending on the data type.

        • CompressionType (string) --

          You can use Gzip or None. The default value is None.

        • TargetAttributeName (string) --

          The name of the target variable in supervised learning, usually represented by 'y'.

        • ContentType (string) --

          The content type of the data from the input source. You can use text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

        • ChannelType (string) --

          The channel type (optional) is an enum string. The default value is training. Channels for training and validation must share the same ContentType and TargetAttributeName. For information on specifying training and validation channel types, see How to specify training and validation datasets.

        • SampleWeightAttributeName (string) --

          If specified, this column name indicates which column of the dataset should be treated as sample weights for use by the objective metric during the training, evaluation, and the selection of the best model. This column is not considered as a predictive feature. For more information on Autopilot metrics, see Metrics and validation.

          Sample weights should be numeric, non-negative, with larger values indicating which rows are more important than others. Data points that have invalid or no weight value are excluded.

          Support for sample weights is available in Ensembling mode only.

    • OutputDataConfig (dict) --

      Returns the job's output data config.

      • KmsKeyId (string) --

        The Key Management Service encryption key ID.

      • S3OutputPath (string) --

        The Amazon S3 output path. Must be 128 characters or less.

    • RoleArn (string) --

      The ARN of the IAM role that has read permission to the input data location and write permission to the output data location in Amazon S3.

    • AutoMLJobObjective (dict) --

      Returns the job's objective.

      • MetricName (string) --

        The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

        The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

        • For tabular problem types:

          • List of available metrics:

            • Regression: MAE, MSE, R2, RMSE

            • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

            • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

          For a description of each metric, see Autopilot metrics for classification and regression.

          • Default objective metrics:

            • Regression: MSE.

            • Binary classification: F1.

            • Multiclass classification: Accuracy.

        • For image or text classification problem types:

        • For time-series forecasting problem types:

        • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

    • ProblemType (string) --

      Returns the job's problem type.

    • AutoMLJobConfig (dict) --

      Returns the configuration for the AutoML job.

      • CompletionCriteria (dict) --

        How long an AutoML job is allowed to run, or how many candidates a job is allowed to generate.

        • MaxCandidates (integer) --

          The maximum number of times a training job is allowed to run.

          For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

        • MaxRuntimePerTrainingJobInSeconds (integer) --

          The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

          For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

          For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

        • MaxAutoMLJobRuntimeInSeconds (integer) --

          The maximum runtime, in seconds, an AutoML job has to complete.

          If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

      • SecurityConfig (dict) --

        The security configuration for traffic encryption or Amazon VPC settings.

        • VolumeKmsKeyId (string) --

          The key used to encrypt stored data.

        • EnableInterContainerTrafficEncryption (boolean) --

          Whether to use traffic encryption between the container layers.

        • VpcConfig (dict) --

          The VPC configuration.

          • SecurityGroupIds (list) --

            The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.

            • (string) --

          • Subnets (list) --

            The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.

            • (string) --

      • CandidateGenerationConfig (dict) --

        The configuration for generating a candidate for an AutoML job (optional).

        • FeatureSpecificationS3Uri (string) --

          A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job. You can input FeatureAttributeNames (optional) in JSON format as shown below:

          { "FeatureAttributeNames":["col1", "col2", ...] }.

          You can also specify the data type of the feature (optional) in the format shown below:

          { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

          In ensembling mode, Autopilot only supports the following data types: numeric, categorical, text, and datetime. In HPO mode, Autopilot can support numeric, categorical, text, datetime, and sequence.

          If only FeatureDataTypes is provided, the column keys ( col1, col2,..) should be a subset of the column names in the input data.

          If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames.

          The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.

        • AlgorithmsConfig (list) --

          Stores the configuration information for the selection of algorithms trained on tabular data.

          The list of available algorithms to choose from depends on the training mode set in TabularJobConfig.Mode.

          • AlgorithmsConfig should not be set if the training mode is set on AUTO.

          • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

          • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

          For the list of all algorithms per problem type and training mode, see AutoMLAlgorithmConfig.

          For more information on each algorithm, see the Algorithm support section in Autopilot developer guide.

          • (dict) --

            The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

            • AutoMLAlgorithms (list) --

              The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

              • For the tabular problem type TabularJobConfig:

              • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

                • Choose your algorithms from this list.

                  • "cnn-qr"

                  • "deepar"

                  • "prophet"

                  • "arima"

                  • "npts"

                  • "ets"

              • (string) --

      • DataSplitConfig (dict) --

        The configuration for splitting the input training dataset.

        Type: AutoMLDataSplitConfig

        • ValidationFraction (float) --

          The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.

      • Mode (string) --

        The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO. In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.

        The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.

        The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.

    • CreationTime (datetime) --

      Returns the creation time of the AutoML job.

    • EndTime (datetime) --

      Returns the end time of the AutoML job.

    • LastModifiedTime (datetime) --

      Returns the job's last modified time.

    • FailureReason (string) --

      Returns the failure reason for an AutoML job, when applicable.

    • PartialFailureReasons (list) --

      Returns a list of reasons for partial failures within an AutoML job.

      • (dict) --

        The reason for a partial failure of an AutoML job.

        • PartialFailureMessage (string) --

          The message containing the reason for a partial failure of an AutoML job.

    • BestCandidate (dict) --

      The best model candidate selected by SageMaker Autopilot using both the best objective metric and lowest InferenceLatency for an experiment.

      • CandidateName (string) --

        The name of the candidate.

      • FinalAutoMLJobObjectiveMetric (dict) --

        The best candidate result from an AutoML training job.

        • Type (string) --

          The type of metric with the best result.

        • MetricName (string) --

          The name of the metric with the best result. For a description of the possible objective metrics, see AutoMLJobObjective$MetricName.

        • Value (float) --

          The value of the metric with the best result.

        • StandardMetricName (string) --

          The name of the standard metric. For a description of the standard metrics, see Autopilot candidate metrics.

      • ObjectiveStatus (string) --

        The objective's status.

      • CandidateSteps (list) --

        Information about the candidate's steps.

        • (dict) --

          Information about the steps for a candidate and what step it is working on.

          • CandidateStepType (string) --

            Whether the candidate is at the transform, training, or processing step.

          • CandidateStepArn (string) --

            The ARN for the candidate's step.

          • CandidateStepName (string) --

            The name for the candidate's step.

      • CandidateStatus (string) --

        The candidate's status.

      • InferenceContainers (list) --

        Information about the recommended inference container definitions.

        • (dict) --

          A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see ContainerDefinition.

          • Image (string) --

            The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see ContainerDefinition.

          • ModelDataUrl (string) --

            The location of the model artifacts. For more information, see ContainerDefinition.

          • Environment (dict) --

            The environment variables to set in the container. For more information, see ContainerDefinition.

            • (string) --

              • (string) --

      • CreationTime (datetime) --

        The creation time.

      • EndTime (datetime) --

        The end time.

      • LastModifiedTime (datetime) --

        The last modified time.

      • FailureReason (string) --

        The failure reason.

      • CandidateProperties (dict) --

        The properties of an AutoML candidate job.

        • CandidateArtifactLocations (dict) --

          The Amazon S3 prefix to the artifacts generated for an AutoML candidate.

          • Explainability (string) --

            The Amazon S3 prefix to the explainability artifacts generated for the AutoML candidate.

          • ModelInsights (string) --

            The Amazon S3 prefix to the model insight artifacts generated for the AutoML candidate.

          • BacktestResults (string) --

            The Amazon S3 prefix to the accuracy metrics and the inference results observed over the testing window. Available only for the time-series forecasting problem type.

        • CandidateMetrics (list) --

          Information about the candidate metrics for an AutoML job.

          • (dict) --

            Information about the metric for a candidate produced by an AutoML job.

            • MetricName (string) --

              The name of the metric.

            • Value (float) --

              The value of the metric.

            • Set (string) --

              The dataset split from which the AutoML job produced the metric.

            • StandardMetricName (string) --

              The name of the standard metric.

      • InferenceContainerDefinitions (dict) --

        The mapping of all supported processing unit (CPU, GPU, etc...) to inference container definitions for the candidate. This field is populated for the AutoML jobs V2 (for example, for jobs created by calling CreateAutoMLJobV2) related to image or text classification problem types only.

        • (string) --

          Processing unit for an inference container. Currently Autopilot only supports CPU or GPU.

          • (list) --

            Information about the recommended inference container definitions.

            • (dict) --

              A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see ContainerDefinition.

              • Image (string) --

                The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see ContainerDefinition.

              • ModelDataUrl (string) --

                The location of the model artifacts. For more information, see ContainerDefinition.

              • Environment (dict) --

                The environment variables to set in the container. For more information, see ContainerDefinition.

                • (string) --

                  • (string) --

    • AutoMLJobStatus (string) --

      Returns the status of the AutoML job.

    • AutoMLJobSecondaryStatus (string) --

      Returns the secondary status of the AutoML job.

    • GenerateCandidateDefinitionsOnly (boolean) --

      Indicates whether the output for an AutoML job generates candidate definitions only.

    • AutoMLJobArtifacts (dict) --

      Returns information on the job's artifacts found in AutoMLJobArtifacts.

      • CandidateDefinitionNotebookLocation (string) --

        The URL of the notebook location.

      • DataExplorationNotebookLocation (string) --

        The URL of the notebook location.

    • ResolvedAttributes (dict) --

      Contains ProblemType, AutoMLJobObjective, and CompletionCriteria. If you do not provide these values, they are inferred.

      • AutoMLJobObjective (dict) --

        Specifies a metric to minimize or maximize as the objective of an AutoML job.

        • MetricName (string) --

          The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

          The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

          • For tabular problem types:

            • List of available metrics:

              • Regression: MAE, MSE, R2, RMSE

              • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

              • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

            For a description of each metric, see Autopilot metrics for classification and regression.

            • Default objective metrics:

              • Regression: MSE.

              • Binary classification: F1.

              • Multiclass classification: Accuracy.

          • For image or text classification problem types:

          • For time-series forecasting problem types:

          • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

      • ProblemType (string) --

        The problem type.

      • CompletionCriteria (dict) --

        How long a job is allowed to run, or how many candidates a job is allowed to generate.

        • MaxCandidates (integer) --

          The maximum number of times a training job is allowed to run.

          For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

        • MaxRuntimePerTrainingJobInSeconds (integer) --

          The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

          For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

          For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

        • MaxAutoMLJobRuntimeInSeconds (integer) --

          The maximum runtime, in seconds, an AutoML job has to complete.

          If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

    • ModelDeployConfig (dict) --

      Indicates whether the model was deployed automatically to an endpoint and the name of that endpoint if deployed automatically.

      • AutoGenerateEndpointName (boolean) --

        Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False.

      • EndpointName (string) --

        Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.

    • ModelDeployResult (dict) --

      Provides information about endpoint for the model deployment.

      • EndpointName (string) --

        The name of the endpoint to which the model has been deployed.

DescribeAutoMLJobV2 (updated) Link ¶
Changes (response)
{'AutoMLProblemTypeConfig': {'TabularJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': {'AutoMLAlgorithms': {'arima',
                                                                                                                          'cnn-qr',
                                                                                                                          'deepar',
                                                                                                                          'ets',
                                                                                                                          'npts',
                                                                                                                          'prophet'}}}},
                             'TimeSeriesForecastingJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': [{'AutoMLAlgorithms': ['xgboost '
                                                                                                                                         '| '
                                                                                                                                         'linear-learner '
                                                                                                                                         '| '
                                                                                                                                         'mlp '
                                                                                                                                         '| '
                                                                                                                                         'lightgbm '
                                                                                                                                         '| '
                                                                                                                                         'catboost '
                                                                                                                                         '| '
                                                                                                                                         'randomforest '
                                                                                                                                         '| '
                                                                                                                                         'extra-trees '
                                                                                                                                         '| '
                                                                                                                                         'nn-torch '
                                                                                                                                         '| '
                                                                                                                                         'fastai '
                                                                                                                                         '| '
                                                                                                                                         'cnn-qr '
                                                                                                                                         '| '
                                                                                                                                         'deepar '
                                                                                                                                         '| '
                                                                                                                                         'prophet '
                                                                                                                                         '| '
                                                                                                                                         'npts '
                                                                                                                                         '| '
                                                                                                                                         'arima '
                                                                                                                                         '| '
                                                                                                                                         'ets']}]}}}}

Returns information about an AutoML job created by calling CreateAutoMLJobV2 or CreateAutoMLJob.

See also: AWS API Documentation

Request Syntax

client.describe_auto_ml_job_v2(
    AutoMLJobName='string'
)
type AutoMLJobName:

string

param AutoMLJobName:

[REQUIRED]

Requests information about an AutoML job V2 using its unique name.

rtype:

dict

returns:

Response Syntax

{
    'AutoMLJobName': 'string',
    'AutoMLJobArn': 'string',
    'AutoMLJobInputDataConfig': [
        {
            'ChannelType': 'training'|'validation',
            'ContentType': 'string',
            'CompressionType': 'None'|'Gzip',
            'DataSource': {
                'S3DataSource': {
                    'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                    'S3Uri': 'string'
                }
            }
        },
    ],
    'OutputDataConfig': {
        'KmsKeyId': 'string',
        'S3OutputPath': 'string'
    },
    'RoleArn': 'string',
    'AutoMLJobObjective': {
        'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
    },
    'AutoMLProblemTypeConfig': {
        'ImageClassificationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            }
        },
        'TextClassificationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'ContentColumn': 'string',
            'TargetLabelColumn': 'string'
        },
        'TimeSeriesForecastingJobConfig': {
            'FeatureSpecificationS3Uri': 'string',
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'ForecastFrequency': 'string',
            'ForecastHorizon': 123,
            'ForecastQuantiles': [
                'string',
            ],
            'Transformations': {
                'Filling': {
                    'string': {
                        'string': 'string'
                    }
                },
                'Aggregation': {
                    'string': 'sum'|'avg'|'first'|'min'|'max'
                }
            },
            'TimeSeriesConfig': {
                'TargetAttributeName': 'string',
                'TimestampAttributeName': 'string',
                'ItemIdentifierAttributeName': 'string',
                'GroupingAttributeNames': [
                    'string',
                ]
            },
            'HolidayConfig': [
                {
                    'CountryCode': 'string'
                },
            ],
            'CandidateGenerationConfig': {
                'AlgorithmsConfig': [
                    {
                        'AutoMLAlgorithms': [
                            'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                        ]
                    },
                ]
            }
        },
        'TabularJobConfig': {
            'CandidateGenerationConfig': {
                'AlgorithmsConfig': [
                    {
                        'AutoMLAlgorithms': [
                            'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai'|'cnn-qr'|'deepar'|'prophet'|'npts'|'arima'|'ets',
                        ]
                    },
                ]
            },
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'FeatureSpecificationS3Uri': 'string',
            'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING',
            'GenerateCandidateDefinitionsOnly': True|False,
            'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression',
            'TargetAttributeName': 'string',
            'SampleWeightAttributeName': 'string'
        },
        'TextGenerationJobConfig': {
            'CompletionCriteria': {
                'MaxCandidates': 123,
                'MaxRuntimePerTrainingJobInSeconds': 123,
                'MaxAutoMLJobRuntimeInSeconds': 123
            },
            'BaseModelName': 'string',
            'TextGenerationHyperParameters': {
                'string': 'string'
            },
            'ModelAccessConfig': {
                'AcceptEula': True|False
            }
        }
    },
    'AutoMLProblemTypeConfigName': 'ImageClassification'|'TextClassification'|'TimeSeriesForecasting'|'Tabular'|'TextGeneration',
    'CreationTime': datetime(2015, 1, 1),
    'EndTime': datetime(2015, 1, 1),
    'LastModifiedTime': datetime(2015, 1, 1),
    'FailureReason': 'string',
    'PartialFailureReasons': [
        {
            'PartialFailureMessage': 'string'
        },
    ],
    'BestCandidate': {
        'CandidateName': 'string',
        'FinalAutoMLJobObjectiveMetric': {
            'Type': 'Maximize'|'Minimize',
            'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss',
            'Value': ...,
            'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
        },
        'ObjectiveStatus': 'Succeeded'|'Pending'|'Failed',
        'CandidateSteps': [
            {
                'CandidateStepType': 'AWS::SageMaker::TrainingJob'|'AWS::SageMaker::TransformJob'|'AWS::SageMaker::ProcessingJob',
                'CandidateStepArn': 'string',
                'CandidateStepName': 'string'
            },
        ],
        'CandidateStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping',
        'InferenceContainers': [
            {
                'Image': 'string',
                'ModelDataUrl': 'string',
                'Environment': {
                    'string': 'string'
                }
            },
        ],
        'CreationTime': datetime(2015, 1, 1),
        'EndTime': datetime(2015, 1, 1),
        'LastModifiedTime': datetime(2015, 1, 1),
        'FailureReason': 'string',
        'CandidateProperties': {
            'CandidateArtifactLocations': {
                'Explainability': 'string',
                'ModelInsights': 'string',
                'BacktestResults': 'string'
            },
            'CandidateMetrics': [
                {
                    'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss',
                    'Value': ...,
                    'Set': 'Train'|'Validation'|'Test',
                    'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro'|'LogLoss'|'InferenceLatency'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'|'Rouge1'|'Rouge2'|'RougeL'|'RougeLSum'|'Perplexity'|'ValidationLoss'|'TrainingLoss'
                },
            ]
        },
        'InferenceContainerDefinitions': {
            'string': [
                {
                    'Image': 'string',
                    'ModelDataUrl': 'string',
                    'Environment': {
                        'string': 'string'
                    }
                },
            ]
        }
    },
    'AutoMLJobStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping',
    'AutoMLJobSecondaryStatus': 'Starting'|'MaxCandidatesReached'|'Failed'|'Stopped'|'MaxAutoMLJobRuntimeReached'|'Stopping'|'CandidateDefinitionsGenerated'|'Completed'|'ExplainabilityError'|'DeployingModel'|'ModelDeploymentError'|'GeneratingModelInsightsReport'|'ModelInsightsError'|'AnalyzingData'|'FeatureEngineering'|'ModelTuning'|'GeneratingExplainabilityReport'|'TrainingModels'|'PreTraining',
    'AutoMLJobArtifacts': {
        'CandidateDefinitionNotebookLocation': 'string',
        'DataExplorationNotebookLocation': 'string'
    },
    'ResolvedAttributes': {
        'AutoMLJobObjective': {
            'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'BalancedAccuracy'|'R2'|'Recall'|'RecallMacro'|'Precision'|'PrecisionMacro'|'MAE'|'MAPE'|'MASE'|'WAPE'|'AverageWeightedQuantileLoss'
        },
        'CompletionCriteria': {
            'MaxCandidates': 123,
            'MaxRuntimePerTrainingJobInSeconds': 123,
            'MaxAutoMLJobRuntimeInSeconds': 123
        },
        'AutoMLProblemTypeResolvedAttributes': {
            'TabularResolvedAttributes': {
                'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression'
            },
            'TextGenerationResolvedAttributes': {
                'BaseModelName': 'string'
            }
        }
    },
    'ModelDeployConfig': {
        'AutoGenerateEndpointName': True|False,
        'EndpointName': 'string'
    },
    'ModelDeployResult': {
        'EndpointName': 'string'
    },
    'DataSplitConfig': {
        'ValidationFraction': ...
    },
    'SecurityConfig': {
        'VolumeKmsKeyId': 'string',
        'EnableInterContainerTrafficEncryption': True|False,
        'VpcConfig': {
            'SecurityGroupIds': [
                'string',
            ],
            'Subnets': [
                'string',
            ]
        }
    }
}

Response Structure

  • (dict) --

    • AutoMLJobName (string) --

      Returns the name of the AutoML job V2.

    • AutoMLJobArn (string) --

      Returns the Amazon Resource Name (ARN) of the AutoML job V2.

    • AutoMLJobInputDataConfig (list) --

      Returns an array of channel objects describing the input data and their location.

      • (dict) --

        A channel is a named input source that training algorithms can consume. This channel is used for AutoML jobs V2 (jobs created by calling CreateAutoMLJobV2).

        • ChannelType (string) --

          The type of channel. Defines whether the data are used for training or validation. The default value is training. Channels for training and validation must share the same ContentType

        • ContentType (string) --

          The content type of the data from the input source. The following are the allowed content types for different problems:

          • For tabular problem types: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

          • For image classification: image/png, image/jpeg, or image/*. The default value is image/*.

          • For text classification: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

          • For time-series forecasting: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

          • For text generation (LLMs fine-tuning): text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.

        • CompressionType (string) --

          The allowed compression types depend on the input format and problem type. We allow the compression type Gzip for S3Prefix inputs on tabular data only. For all other inputs, the compression type should be None. If no compression type is provided, we default to None.

        • DataSource (dict) --

          The data source for an AutoML channel (Required).

          • S3DataSource (dict) --

            The Amazon S3 location of the input data.

            • S3DataType (string) --

              The data type.

              • If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix for model training. The S3Prefix should have the following format: s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE

              • If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want SageMaker to use for model training. A ManifestFile should have the format shown below: [ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"}, "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1", "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2", ... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]

              • If you choose AugmentedManifestFile, S3Uri identifies an object that is an augmented manifest file in JSON lines format. This file contains the data you want to use for model training. AugmentedManifestFile is available for V2 API jobs only (for example, for jobs created by calling CreateAutoMLJobV2). Here is a minimal, single-record example of an AugmentedManifestFile: {"source-ref": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/cats/cat.jpg", "label-metadata": {"class-name": "cat" } For more information on AugmentedManifestFile, see Provide Dataset Metadata to Training Jobs with an Augmented Manifest File.

            • S3Uri (string) --

              The URL to the Amazon S3 data source. The Uri refers to the Amazon S3 prefix or ManifestFile depending on the data type.

    • OutputDataConfig (dict) --

      Returns the job's output data config.

      • KmsKeyId (string) --

        The Key Management Service encryption key ID.

      • S3OutputPath (string) --

        The Amazon S3 output path. Must be 128 characters or less.

    • RoleArn (string) --

      The ARN of the IAM role that has read permission to the input data location and write permission to the output data location in Amazon S3.

    • AutoMLJobObjective (dict) --

      Returns the job's objective.

      • MetricName (string) --

        The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

        The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

        • For tabular problem types:

          • List of available metrics:

            • Regression: MAE, MSE, R2, RMSE

            • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

            • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

          For a description of each metric, see Autopilot metrics for classification and regression.

          • Default objective metrics:

            • Regression: MSE.

            • Binary classification: F1.

            • Multiclass classification: Accuracy.

        • For image or text classification problem types:

        • For time-series forecasting problem types:

        • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

    • AutoMLProblemTypeConfig (dict) --

      Returns the configuration settings of the problem type set for the AutoML job V2.

      • ImageClassificationJobConfig (dict) --

        Settings used to configure an AutoML job V2 for the image classification problem type.

        • CompletionCriteria (dict) --

          How long a job is allowed to run, or how many candidates a job is allowed to generate.

          • MaxCandidates (integer) --

            The maximum number of times a training job is allowed to run.

            For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

          • MaxRuntimePerTrainingJobInSeconds (integer) --

            The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

            For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

            For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

          • MaxAutoMLJobRuntimeInSeconds (integer) --

            The maximum runtime, in seconds, an AutoML job has to complete.

            If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

      • TextClassificationJobConfig (dict) --

        Settings used to configure an AutoML job V2 for the text classification problem type.

        • CompletionCriteria (dict) --

          How long a job is allowed to run, or how many candidates a job is allowed to generate.

          • MaxCandidates (integer) --

            The maximum number of times a training job is allowed to run.

            For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

          • MaxRuntimePerTrainingJobInSeconds (integer) --

            The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

            For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

            For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

          • MaxAutoMLJobRuntimeInSeconds (integer) --

            The maximum runtime, in seconds, an AutoML job has to complete.

            If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

        • ContentColumn (string) --

          The name of the column used to provide the sentences to be classified. It should not be the same as the target column.

        • TargetLabelColumn (string) --

          The name of the column used to provide the class labels. It should not be same as the content column.

      • TimeSeriesForecastingJobConfig (dict) --

        Settings used to configure an AutoML job V2 for the time-series forecasting problem type.

        • FeatureSpecificationS3Uri (string) --

          A URL to the Amazon S3 data source containing additional selected features that complement the target, itemID, timestamp, and grouped columns set in TimeSeriesConfig. When not provided, the AutoML job V2 includes all the columns from the original dataset that are not already declared in TimeSeriesConfig. If provided, the AutoML job V2 only considers these additional columns as a complement to the ones declared in TimeSeriesConfig.

          You can input FeatureAttributeNames (optional) in JSON format as shown below:

          { "FeatureAttributeNames":["col1", "col2", ...] }.

          You can also specify the data type of the feature (optional) in the format shown below:

          { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

          Autopilot supports the following data types: numeric, categorical, text, and datetime.

        • CompletionCriteria (dict) --

          How long a job is allowed to run, or how many candidates a job is allowed to generate.

          • MaxCandidates (integer) --

            The maximum number of times a training job is allowed to run.

            For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

          • MaxRuntimePerTrainingJobInSeconds (integer) --

            The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

            For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

            For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

          • MaxAutoMLJobRuntimeInSeconds (integer) --

            The maximum runtime, in seconds, an AutoML job has to complete.

            If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

        • ForecastFrequency (string) --

          The frequency of predictions in a forecast.

          Valid intervals are an integer followed by Y (Year), M (Month), W (Week), D (Day), H (Hour), and min (Minute). For example, 1D indicates every day and 15min indicates every 15 minutes. The value of a frequency must not overlap with the next larger frequency. For example, you must use a frequency of 1H instead of 60min.

          The valid values for each frequency are the following:

          • Minute - 1-59

          • Hour - 1-23

          • Day - 1-6

          • Week - 1-4

          • Month - 1-11

          • Year - 1

        • ForecastHorizon (integer) --

          The number of time-steps that the model predicts. The forecast horizon is also called the prediction length. The maximum forecast horizon is the lesser of 500 time-steps or 1/4 of the time-steps in the dataset.

        • ForecastQuantiles (list) --

          The quantiles used to train the model for forecasts at a specified quantile. You can specify quantiles from 0.01 (p1) to 0.99 (p99), by increments of 0.01 or higher. Up to five forecast quantiles can be specified. When ForecastQuantiles is not provided, the AutoML job uses the quantiles p10, p50, and p90 as default.

          • (string) --

        • Transformations (dict) --

          The transformations modifying specific attributes of the time-series, such as filling strategies for missing values.

          • Filling (dict) --

            A key value pair defining the filling method for a column, where the key is the column name and the value is an object which defines the filling logic. You can specify multiple filling methods for a single column.

            The supported filling methods and their corresponding options are:

            • frontfill: none (Supported only for target column)

            • middlefill: zero, value, median, mean, min, max

            • backfill: zero, value, median, mean, min, max

            • futurefill: zero, value, median, mean, min, max

            To set a filling method to a specific value, set the fill parameter to the chosen filling method value (for example "backfill" : "value"), and define the filling value in an additional parameter prefixed with "_value". For example, to set backfill to a value of 2, you must include two parameters: "backfill": "value" and "backfill_value":"2".

            • (string) --

              • (dict) --

                • (string) --

                  • (string) --

          • Aggregation (dict) --

            A key value pair defining the aggregation method for a column, where the key is the column name and the value is the aggregation method.

            The supported aggregation methods are sum (default), avg, first, min, max.

            • (string) --

              • (string) --

        • TimeSeriesConfig (dict) --

          The collection of components that defines the time-series.

          • TargetAttributeName (string) --

            The name of the column representing the target variable that you want to predict for each item in your dataset. The data type of the target variable must be numerical.

          • TimestampAttributeName (string) --

            The name of the column indicating a point in time at which the target value of a given item is recorded.

          • ItemIdentifierAttributeName (string) --

            The name of the column that represents the set of item identifiers for which you want to predict the target value.

          • GroupingAttributeNames (list) --

            A set of columns names that can be grouped with the item identifier column to create a composite key for which a target value is predicted.

            • (string) --

        • HolidayConfig (list) --

          The collection of holiday featurization attributes used to incorporate national holiday information into your forecasting model.

          • (dict) --

            Stores the holiday featurization attributes applicable to each item of time-series datasets during the training of a forecasting model. This allows the model to identify patterns associated with specific holidays.

            • CountryCode (string) --

              The country code for the holiday calendar.

              For the list of public holiday calendars supported by AutoML job V2, see Country Codes. Use the country code corresponding to the country of your choice.

        • CandidateGenerationConfig (dict) --

          Stores the configuration information for how model candidates are generated using an AutoML job V2.

          • AlgorithmsConfig (list) --

            Your Autopilot job trains a default set of algorithms on your dataset. For tabular and time-series data, you can customize the algorithm list by selecting a subset of algorithms for your problem type.

            AlgorithmsConfig stores the customized selection of algorithms to train on your data.

            • For the tabular problem type TabularJobConfig, the list of available algorithms to choose from depends on the training mode set in AutoMLJobConfig.Mode.

              • AlgorithmsConfig should not be set when the training mode AutoMLJobConfig.Mode is set to AUTO.

              • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

              • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

            For the list of all algorithms per training mode, see AlgorithmConfig.

            For more information on each algorithm, see the Algorithm support section in the Autopilot developer guide.

            • For the time-series forecasting problem type TimeSeriesForecastingJobConfig, choose your algorithms from the list provided in AlgorithmConfig. For more information on each algorithm, see the Algorithms support for time-series forecasting section in the Autopilot developer guide.

              • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

              • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

            • (dict) --

              The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

              • AutoMLAlgorithms (list) --

                The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

                • For the tabular problem type TabularJobConfig:

                • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

                  • Choose your algorithms from this list.

                    • "cnn-qr"

                    • "deepar"

                    • "prophet"

                    • "arima"

                    • "npts"

                    • "ets"

                • (string) --

      • TabularJobConfig (dict) --

        Settings used to configure an AutoML job V2 for the tabular problem type (regression, classification).

        • CandidateGenerationConfig (dict) --

          The configuration information of how model candidates are generated.

          • AlgorithmsConfig (list) --

            Your Autopilot job trains a default set of algorithms on your dataset. For tabular and time-series data, you can customize the algorithm list by selecting a subset of algorithms for your problem type.

            AlgorithmsConfig stores the customized selection of algorithms to train on your data.

            • For the tabular problem type TabularJobConfig, the list of available algorithms to choose from depends on the training mode set in AutoMLJobConfig.Mode.

              • AlgorithmsConfig should not be set when the training mode AutoMLJobConfig.Mode is set to AUTO.

              • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

              • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for the given training mode.

            For the list of all algorithms per training mode, see AlgorithmConfig.

            For more information on each algorithm, see the Algorithm support section in the Autopilot developer guide.

            • For the time-series forecasting problem type TimeSeriesForecastingJobConfig, choose your algorithms from the list provided in AlgorithmConfig. For more information on each algorithm, see the Algorithms support for time-series forecasting section in the Autopilot developer guide.

              • When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

              • When AlgorithmsConfig is not provided, CandidateGenerationConfig uses the full set of algorithms for time-series forecasting.

            • (dict) --

              The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

              • AutoMLAlgorithms (list) --

                The selection of algorithms trained on your dataset to generate the model candidates for an Autopilot job.

                • For the tabular problem type TabularJobConfig:

                • For the time-series forecasting problem type TimeSeriesForecastingJobConfig:

                  • Choose your algorithms from this list.

                    • "cnn-qr"

                    • "deepar"

                    • "prophet"

                    • "arima"

                    • "npts"

                    • "ets"

                • (string) --

        • CompletionCriteria (dict) --

          How long a job is allowed to run, or how many candidates a job is allowed to generate.

          • MaxCandidates (integer) --

            The maximum number of times a training job is allowed to run.

            For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

          • MaxRuntimePerTrainingJobInSeconds (integer) --

            The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

            For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

            For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

          • MaxAutoMLJobRuntimeInSeconds (integer) --

            The maximum runtime, in seconds, an AutoML job has to complete.

            If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

        • FeatureSpecificationS3Uri (string) --

          A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job V2. You can input FeatureAttributeNames (optional) in JSON format as shown below:

          { "FeatureAttributeNames":["col1", "col2", ...] }.

          You can also specify the data type of the feature (optional) in the format shown below:

          { "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }

          In ensembling mode, Autopilot only supports the following data types: numeric, categorical, text, and datetime. In HPO mode, Autopilot can support numeric, categorical, text, datetime, and sequence.

          If only FeatureDataTypes is provided, the column keys ( col1, col2,..) should be a subset of the column names in the input data.

          If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames.

          The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.

        • Mode (string) --

          The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO. In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.

          The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.

          The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.

        • GenerateCandidateDefinitionsOnly (boolean) --

          Generates possible candidates without training the models. A model candidate is a combination of data preprocessors, algorithms, and algorithm parameter settings.

        • ProblemType (string) --

          The type of supervised learning problem available for the model candidates of the AutoML job V2. For more information, see SageMaker Autopilot problem types.

        • TargetAttributeName (string) --

          The name of the target variable in supervised learning, usually represented by 'y'.

        • SampleWeightAttributeName (string) --

          If specified, this column name indicates which column of the dataset should be treated as sample weights for use by the objective metric during the training, evaluation, and the selection of the best model. This column is not considered as a predictive feature. For more information on Autopilot metrics, see Metrics and validation.

          Sample weights should be numeric, non-negative, with larger values indicating which rows are more important than others. Data points that have invalid or no weight value are excluded.

          Support for sample weights is available in Ensembling mode only.

      • TextGenerationJobConfig (dict) --

        Settings used to configure an AutoML job V2 for the text generation (LLMs fine-tuning) problem type.

        • CompletionCriteria (dict) --

          How long a fine-tuning job is allowed to run. For TextGenerationJobConfig problem types, the MaxRuntimePerTrainingJobInSeconds attribute of AutoMLJobCompletionCriteria defaults to 72h (259200s).

          • MaxCandidates (integer) --

            The maximum number of times a training job is allowed to run.

            For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

          • MaxRuntimePerTrainingJobInSeconds (integer) --

            The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

            For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

            For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

          • MaxAutoMLJobRuntimeInSeconds (integer) --

            The maximum runtime, in seconds, an AutoML job has to complete.

            If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

        • BaseModelName (string) --

          The name of the base model to fine-tune. Autopilot supports fine-tuning a variety of large language models. For information on the list of supported models, see Text generation models supporting fine-tuning in Autopilot. If no BaseModelName is provided, the default model used is Falcon7BInstruct.

        • TextGenerationHyperParameters (dict) --

          The hyperparameters used to configure and optimize the learning process of the base model. You can set any combination of the following hyperparameters for all base models. For more information on each supported hyperparameter, see Optimize the learning process of your text generation models with hyperparameters.

          • "epochCount": The number of times the model goes through the entire training dataset. Its value should be a string containing an integer value within the range of "1" to "10".

          • "batchSize": The number of data samples used in each iteration of training. Its value should be a string containing an integer value within the range of "1" to "64".

          • "learningRate": The step size at which a model's parameters are updated during training. Its value should be a string containing a floating-point value within the range of "0" to "1".

          • "learningRateWarmupSteps": The number of training steps during which the learning rate gradually increases before reaching its target or maximum value. Its value should be a string containing an integer value within the range of "0" to "250".

          Here is an example where all four hyperparameters are configured.

          { "epochCount":"5", "learningRate":"0.5", "batchSize": "32", "learningRateWarmupSteps": "10" }

          • (string) --

            • (string) --

        • ModelAccessConfig (dict) --

          The access configuration file to control access to the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig.

          • AcceptEula (boolean) --

            Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

    • AutoMLProblemTypeConfigName (string) --

      Returns the name of the problem type configuration set for the AutoML job V2.

    • CreationTime (datetime) --

      Returns the creation time of the AutoML job V2.

    • EndTime (datetime) --

      Returns the end time of the AutoML job V2.

    • LastModifiedTime (datetime) --

      Returns the job's last modified time.

    • FailureReason (string) --

      Returns the reason for the failure of the AutoML job V2, when applicable.

    • PartialFailureReasons (list) --

      Returns a list of reasons for partial failures within an AutoML job V2.

      • (dict) --

        The reason for a partial failure of an AutoML job.

        • PartialFailureMessage (string) --

          The message containing the reason for a partial failure of an AutoML job.

    • BestCandidate (dict) --

      Information about the candidate produced by an AutoML training job V2, including its status, steps, and other properties.

      • CandidateName (string) --

        The name of the candidate.

      • FinalAutoMLJobObjectiveMetric (dict) --

        The best candidate result from an AutoML training job.

        • Type (string) --

          The type of metric with the best result.

        • MetricName (string) --

          The name of the metric with the best result. For a description of the possible objective metrics, see AutoMLJobObjective$MetricName.

        • Value (float) --

          The value of the metric with the best result.

        • StandardMetricName (string) --

          The name of the standard metric. For a description of the standard metrics, see Autopilot candidate metrics.

      • ObjectiveStatus (string) --

        The objective's status.

      • CandidateSteps (list) --

        Information about the candidate's steps.

        • (dict) --

          Information about the steps for a candidate and what step it is working on.

          • CandidateStepType (string) --

            Whether the candidate is at the transform, training, or processing step.

          • CandidateStepArn (string) --

            The ARN for the candidate's step.

          • CandidateStepName (string) --

            The name for the candidate's step.

      • CandidateStatus (string) --

        The candidate's status.

      • InferenceContainers (list) --

        Information about the recommended inference container definitions.

        • (dict) --

          A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see ContainerDefinition.

          • Image (string) --

            The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see ContainerDefinition.

          • ModelDataUrl (string) --

            The location of the model artifacts. For more information, see ContainerDefinition.

          • Environment (dict) --

            The environment variables to set in the container. For more information, see ContainerDefinition.

            • (string) --

              • (string) --

      • CreationTime (datetime) --

        The creation time.

      • EndTime (datetime) --

        The end time.

      • LastModifiedTime (datetime) --

        The last modified time.

      • FailureReason (string) --

        The failure reason.

      • CandidateProperties (dict) --

        The properties of an AutoML candidate job.

        • CandidateArtifactLocations (dict) --

          The Amazon S3 prefix to the artifacts generated for an AutoML candidate.

          • Explainability (string) --

            The Amazon S3 prefix to the explainability artifacts generated for the AutoML candidate.

          • ModelInsights (string) --

            The Amazon S3 prefix to the model insight artifacts generated for the AutoML candidate.

          • BacktestResults (string) --

            The Amazon S3 prefix to the accuracy metrics and the inference results observed over the testing window. Available only for the time-series forecasting problem type.

        • CandidateMetrics (list) --

          Information about the candidate metrics for an AutoML job.

          • (dict) --

            Information about the metric for a candidate produced by an AutoML job.

            • MetricName (string) --

              The name of the metric.

            • Value (float) --

              The value of the metric.

            • Set (string) --

              The dataset split from which the AutoML job produced the metric.

            • StandardMetricName (string) --

              The name of the standard metric.

      • InferenceContainerDefinitions (dict) --

        The mapping of all supported processing unit (CPU, GPU, etc...) to inference container definitions for the candidate. This field is populated for the AutoML jobs V2 (for example, for jobs created by calling CreateAutoMLJobV2) related to image or text classification problem types only.

        • (string) --

          Processing unit for an inference container. Currently Autopilot only supports CPU or GPU.

          • (list) --

            Information about the recommended inference container definitions.

            • (dict) --

              A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see ContainerDefinition.

              • Image (string) --

                The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see ContainerDefinition.

              • ModelDataUrl (string) --

                The location of the model artifacts. For more information, see ContainerDefinition.

              • Environment (dict) --

                The environment variables to set in the container. For more information, see ContainerDefinition.

                • (string) --

                  • (string) --

    • AutoMLJobStatus (string) --

      Returns the status of the AutoML job V2.

    • AutoMLJobSecondaryStatus (string) --

      Returns the secondary status of the AutoML job V2.

    • AutoMLJobArtifacts (dict) --

      The artifacts that are generated during an AutoML job.

      • CandidateDefinitionNotebookLocation (string) --

        The URL of the notebook location.

      • DataExplorationNotebookLocation (string) --

        The URL of the notebook location.

    • ResolvedAttributes (dict) --

      Returns the resolved attributes used by the AutoML job V2.

      • AutoMLJobObjective (dict) --

        Specifies a metric to minimize or maximize as the objective of an AutoML job.

        • MetricName (string) --

          The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset.

          The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type.

          • For tabular problem types:

            • List of available metrics:

              • Regression: MAE, MSE, R2, RMSE

              • Binary classification: Accuracy, AUC, BalancedAccuracy, F1, Precision, Recall

              • Multiclass classification: Accuracy, BalancedAccuracy, F1macro, PrecisionMacro, RecallMacro

            For a description of each metric, see Autopilot metrics for classification and regression.

            • Default objective metrics:

              • Regression: MSE.

              • Binary classification: F1.

              • Multiclass classification: Accuracy.

          • For image or text classification problem types:

          • For time-series forecasting problem types:

          • For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.

      • CompletionCriteria (dict) --

        How long a job is allowed to run, or how many candidates a job is allowed to generate.

        • MaxCandidates (integer) --

          The maximum number of times a training job is allowed to run.

          For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.

        • MaxRuntimePerTrainingJobInSeconds (integer) --

          The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the StoppingCondition used by the CreateHyperParameterTuningJob action.

          For job V2s (jobs created by calling CreateAutoMLJobV2), this field controls the runtime of the job candidate.

          For TextGenerationJobConfig problem types, the maximum time defaults to 72 hours (259200 seconds).

        • MaxAutoMLJobRuntimeInSeconds (integer) --

          The maximum runtime, in seconds, an AutoML job has to complete.

          If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.

      • AutoMLProblemTypeResolvedAttributes (dict) --

        Defines the resolved attributes specific to a problem type.

        • TabularResolvedAttributes (dict) --

          The resolved attributes for the tabular problem type.

          • ProblemType (string) --

            The type of supervised learning problem available for the model candidates of the AutoML job V2 (Binary Classification, Multiclass Classification, Regression). For more information, see SageMaker Autopilot problem types.

        • TextGenerationResolvedAttributes (dict) --

          The resolved attributes for the text generation problem type.

          • BaseModelName (string) --

            The name of the base model to fine-tune.

    • ModelDeployConfig (dict) --

      Indicates whether the model was deployed automatically to an endpoint and the name of that endpoint if deployed automatically.

      • AutoGenerateEndpointName (boolean) --

        Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False.

      • EndpointName (string) --

        Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.

    • ModelDeployResult (dict) --

      Provides information about endpoint for the model deployment.

      • EndpointName (string) --

        The name of the endpoint to which the model has been deployed.

    • DataSplitConfig (dict) --

      Returns the configuration settings of how the data are split into train and validation datasets.

      • ValidationFraction (float) --

        The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.

    • SecurityConfig (dict) --

      Returns the security configuration for traffic encryption or Amazon VPC settings.

      • VolumeKmsKeyId (string) --

        The key used to encrypt stored data.

      • EnableInterContainerTrafficEncryption (boolean) --

        Whether to use traffic encryption between the container layers.

      • VpcConfig (dict) --

        The VPC configuration.

        • SecurityGroupIds (list) --

          The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.

          • (string) --

        • Subnets (list) --

          The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones.

          • (string) --

DescribeModelPackage (updated) Link ¶
Changes (response)
{'ModelCard': {'ModelCardContent': 'string',
               'ModelCardStatus': 'Draft | PendingReview | Approved | '
                                  'Archived'},
 'SecurityConfig': {'KmsKeyId': 'string'}}

Returns a description of the specified model package, which is used to create SageMaker models or list them on Amazon Web Services Marketplace.

To create models in SageMaker, buyers can subscribe to model packages listed on Amazon Web Services Marketplace.

See also: AWS API Documentation

Request Syntax

client.describe_model_package(
    ModelPackageName='string'
)
type ModelPackageName:

string

param ModelPackageName:

[REQUIRED]

The name or Amazon Resource Name (ARN) of the model package to describe.

When you specify a name, the name must have 1 to 63 characters. Valid characters are a-z, A-Z, 0-9, and - (hyphen).

rtype:

dict

returns:

Response Syntax

{
    'ModelPackageName': 'string',
    'ModelPackageGroupName': 'string',
    'ModelPackageVersion': 123,
    'ModelPackageArn': 'string',
    'ModelPackageDescription': 'string',
    'CreationTime': datetime(2015, 1, 1),
    'InferenceSpecification': {
        'Containers': [
            {
                'ContainerHostname': 'string',
                'Image': 'string',
                'ImageDigest': 'string',
                'ModelDataUrl': 'string',
                'ModelDataSource': {
                    'S3DataSource': {
                        'S3Uri': 'string',
                        'S3DataType': 'S3Prefix'|'S3Object',
                        'CompressionType': 'None'|'Gzip',
                        'ModelAccessConfig': {
                            'AcceptEula': True|False
                        }
                    }
                },
                'ProductId': 'string',
                'Environment': {
                    'string': 'string'
                },
                'ModelInput': {
                    'DataInputConfig': 'string'
                },
                'Framework': 'string',
                'FrameworkVersion': 'string',
                'NearestModelName': 'string',
                'AdditionalS3DataSource': {
                    'S3DataType': 'S3Object'|'S3Prefix',
                    'S3Uri': 'string',
                    'CompressionType': 'None'|'Gzip'
                }
            },
        ],
        'SupportedTransformInstanceTypes': [
            'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
        ],
        'SupportedRealtimeInferenceInstanceTypes': [
            'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
        ],
        'SupportedContentTypes': [
            'string',
        ],
        'SupportedResponseMIMETypes': [
            'string',
        ]
    },
    'SourceAlgorithmSpecification': {
        'SourceAlgorithms': [
            {
                'ModelDataUrl': 'string',
                'ModelDataSource': {
                    'S3DataSource': {
                        'S3Uri': 'string',
                        'S3DataType': 'S3Prefix'|'S3Object',
                        'CompressionType': 'None'|'Gzip',
                        'ModelAccessConfig': {
                            'AcceptEula': True|False
                        }
                    }
                },
                'AlgorithmName': 'string'
            },
        ]
    },
    'ValidationSpecification': {
        'ValidationRole': 'string',
        'ValidationProfiles': [
            {
                'ProfileName': 'string',
                'TransformJobDefinition': {
                    'MaxConcurrentTransforms': 123,
                    'MaxPayloadInMB': 123,
                    'BatchStrategy': 'MultiRecord'|'SingleRecord',
                    'Environment': {
                        'string': 'string'
                    },
                    'TransformInput': {
                        'DataSource': {
                            'S3DataSource': {
                                'S3DataType': 'ManifestFile'|'S3Prefix'|'AugmentedManifestFile',
                                'S3Uri': 'string'
                            }
                        },
                        'ContentType': 'string',
                        'CompressionType': 'None'|'Gzip',
                        'SplitType': 'None'|'Line'|'RecordIO'|'TFRecord'
                    },
                    'TransformOutput': {
                        'S3OutputPath': 'string',
                        'Accept': 'string',
                        'AssembleWith': 'None'|'Line',
                        'KmsKeyId': 'string'
                    },
                    'TransformResources': {
                        'InstanceType': 'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
                        'InstanceCount': 123,
                        'VolumeKmsKeyId': 'string'
                    }
                }
            },
        ]
    },
    'ModelPackageStatus': 'Pending'|'InProgress'|'Completed'|'Failed'|'Deleting',
    'ModelPackageStatusDetails': {
        'ValidationStatuses': [
            {
                'Name': 'string',
                'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed',
                'FailureReason': 'string'
            },
        ],
        'ImageScanStatuses': [
            {
                'Name': 'string',
                'Status': 'NotStarted'|'InProgress'|'Completed'|'Failed',
                'FailureReason': 'string'
            },
        ]
    },
    'CertifyForMarketplace': True|False,
    'ModelApprovalStatus': 'Approved'|'Rejected'|'PendingManualApproval',
    'CreatedBy': {
        'UserProfileArn': 'string',
        'UserProfileName': 'string',
        'DomainId': 'string',
        'IamIdentity': {
            'Arn': 'string',
            'PrincipalId': 'string',
            'SourceIdentity': 'string'
        }
    },
    'MetadataProperties': {
        'CommitId': 'string',
        'Repository': 'string',
        'GeneratedBy': 'string',
        'ProjectId': 'string'
    },
    'ModelMetrics': {
        'ModelQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelDataQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Bias': {
            'Report': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PreTrainingReport': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PostTrainingReport': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Explainability': {
            'Report': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        }
    },
    'LastModifiedTime': datetime(2015, 1, 1),
    'LastModifiedBy': {
        'UserProfileArn': 'string',
        'UserProfileName': 'string',
        'DomainId': 'string',
        'IamIdentity': {
            'Arn': 'string',
            'PrincipalId': 'string',
            'SourceIdentity': 'string'
        }
    },
    'ApprovalDescription': 'string',
    'Domain': 'string',
    'Task': 'string',
    'SamplePayloadUrl': 'string',
    'CustomerMetadataProperties': {
        'string': 'string'
    },
    'DriftCheckBaselines': {
        'Bias': {
            'ConfigFile': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PreTrainingConstraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'PostTrainingConstraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'Explainability': {
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'ConfigFile': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        },
        'ModelDataQuality': {
            'Statistics': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            },
            'Constraints': {
                'ContentType': 'string',
                'ContentDigest': 'string',
                'S3Uri': 'string'
            }
        }
    },
    'AdditionalInferenceSpecifications': [
        {
            'Name': 'string',
            'Description': 'string',
            'Containers': [
                {
                    'ContainerHostname': 'string',
                    'Image': 'string',
                    'ImageDigest': 'string',
                    'ModelDataUrl': 'string',
                    'ModelDataSource': {
                        'S3DataSource': {
                            'S3Uri': 'string',
                            'S3DataType': 'S3Prefix'|'S3Object',
                            'CompressionType': 'None'|'Gzip',
                            'ModelAccessConfig': {
                                'AcceptEula': True|False
                            }
                        }
                    },
                    'ProductId': 'string',
                    'Environment': {
                        'string': 'string'
                    },
                    'ModelInput': {
                        'DataInputConfig': 'string'
                    },
                    'Framework': 'string',
                    'FrameworkVersion': 'string',
                    'NearestModelName': 'string',
                    'AdditionalS3DataSource': {
                        'S3DataType': 'S3Object'|'S3Prefix',
                        'S3Uri': 'string',
                        'CompressionType': 'None'|'Gzip'
                    }
                },
            ],
            'SupportedTransformInstanceTypes': [
                'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
            ],
            'SupportedRealtimeInferenceInstanceTypes': [
                'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
            ],
            'SupportedContentTypes': [
                'string',
            ],
            'SupportedResponseMIMETypes': [
                'string',
            ]
        },
    ],
    'SkipModelValidation': 'All'|'None',
    'SourceUri': 'string',
    'SecurityConfig': {
        'KmsKeyId': 'string'
    },
    'ModelCard': {
        'ModelCardContent': 'string',
        'ModelCardStatus': 'Draft'|'PendingReview'|'Approved'|'Archived'
    }
}

Response Structure

  • (dict) --

    • ModelPackageName (string) --

      The name of the model package being described.

    • ModelPackageGroupName (string) --

      If the model is a versioned model, the name of the model group that the versioned model belongs to.

    • ModelPackageVersion (integer) --

      The version of the model package.

    • ModelPackageArn (string) --

      The Amazon Resource Name (ARN) of the model package.

    • ModelPackageDescription (string) --

      A brief summary of the model package.

    • CreationTime (datetime) --

      A timestamp specifying when the model package was created.

    • InferenceSpecification (dict) --

      Details about inference jobs that you can run with models based on this model package.

      • Containers (list) --

        The Amazon ECR registry path of the Docker image that contains the inference code.

        • (dict) --

          Describes the Docker container for the model package.

          • ContainerHostname (string) --

            The DNS host name for the Docker container.

          • Image (string) --

            The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

            If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

          • ImageDigest (string) --

            An MD5 hash of the training algorithm that identifies the Docker image used for training.

          • ModelDataUrl (string) --

            The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

          • ModelDataSource (dict) --

            Specifies the location of ML model data to deploy during endpoint creation.

            • S3DataSource (dict) --

              Specifies the S3 location of ML model data to deploy.

              • S3Uri (string) --

                Specifies the S3 path of ML model data to deploy.

              • S3DataType (string) --

                Specifies the type of ML model data to deploy.

                If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

                If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

              • CompressionType (string) --

                Specifies how the ML model data is prepared.

                If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

                If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

                If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

                If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

                • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

                • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

                • Do not use any of the following as file names or directory names:

                  • An empty or blank string

                  • A string which contains null bytes

                  • A string longer than 255 bytes

                  • A single dot ( .)

                  • A double dot ( ..)

                • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

                • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

              • ModelAccessConfig (dict) --

                Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

                • AcceptEula (boolean) --

                  Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

          • ProductId (string) --

            The Amazon Web Services Marketplace product ID of the model package.

          • Environment (dict) --

            The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

            • (string) --

              • (string) --

          • ModelInput (dict) --

            A structure with Model Input details.

            • DataInputConfig (string) --

              The input configuration object for the model.

          • Framework (string) --

            The machine learning framework of the model package container image.

          • FrameworkVersion (string) --

            The framework version of the Model Package Container Image.

          • NearestModelName (string) --

            The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

          • AdditionalS3DataSource (dict) --

            The additional data source that is used during inference in the Docker container for your model package.

            • S3DataType (string) --

              The data type of the additional data source that you specify for use in inference or training.

            • S3Uri (string) --

              The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

            • CompressionType (string) --

              The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

      • SupportedTransformInstanceTypes (list) --

        A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

        This parameter is required for unversioned models, and optional for versioned models.

        • (string) --

      • SupportedRealtimeInferenceInstanceTypes (list) --

        A list of the instance types that are used to generate inferences in real-time.

        This parameter is required for unversioned models, and optional for versioned models.

        • (string) --

      • SupportedContentTypes (list) --

        The supported MIME types for the input data.

        • (string) --

      • SupportedResponseMIMETypes (list) --

        The supported MIME types for the output data.

        • (string) --

    • SourceAlgorithmSpecification (dict) --

      Details about the algorithm that was used to create the model package.

      • SourceAlgorithms (list) --

        A list of the algorithms that were used to create a model package.

        • (dict) --

          Specifies an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.

          • ModelDataUrl (string) --

            The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

          • ModelDataSource (dict) --

            Specifies the location of ML model data to deploy during endpoint creation.

            • S3DataSource (dict) --

              Specifies the S3 location of ML model data to deploy.

              • S3Uri (string) --

                Specifies the S3 path of ML model data to deploy.

              • S3DataType (string) --

                Specifies the type of ML model data to deploy.

                If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

                If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

              • CompressionType (string) --

                Specifies how the ML model data is prepared.

                If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

                If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

                If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

                If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

                • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

                • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

                • Do not use any of the following as file names or directory names:

                  • An empty or blank string

                  • A string which contains null bytes

                  • A string longer than 255 bytes

                  • A single dot ( .)

                  • A double dot ( ..)

                • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

                • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

              • ModelAccessConfig (dict) --

                Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

                • AcceptEula (boolean) --

                  Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

          • AlgorithmName (string) --

            The name of an algorithm that was used to create the model package. The algorithm must be either an algorithm resource in your SageMaker account or an algorithm in Amazon Web Services Marketplace that you are subscribed to.

    • ValidationSpecification (dict) --

      Configurations for one or more transform jobs that SageMaker runs to test the model package.

      • ValidationRole (string) --

        The IAM roles to be used for the validation of the model package.

      • ValidationProfiles (list) --

        An array of ModelPackageValidationProfile objects, each of which specifies a batch transform job that SageMaker runs to validate your model package.

        • (dict) --

          Contains data, such as the inputs and targeted instance types that are used in the process of validating the model package.

          The data provided in the validation profile is made available to your buyers on Amazon Web Services Marketplace.

          • ProfileName (string) --

            The name of the profile for the model package.

          • TransformJobDefinition (dict) --

            The TransformJobDefinition object that describes the transform job used for the validation of the model package.

            • MaxConcurrentTransforms (integer) --

              The maximum number of parallel requests that can be sent to each instance in a transform job. The default value is 1.

            • MaxPayloadInMB (integer) --

              The maximum payload size allowed, in MB. A payload is the data portion of a record (without metadata).

            • BatchStrategy (string) --

              A string that determines the number of records included in a single mini-batch.

              SingleRecord means only one record is used per mini-batch. MultiRecord means a mini-batch is set to contain as many records that can fit within the MaxPayloadInMB limit.

            • Environment (dict) --

              The environment variables to set in the Docker container. We support up to 16 key and values entries in the map.

              • (string) --

                • (string) --

            • TransformInput (dict) --

              A description of the input source and the way the transform job consumes it.

              • DataSource (dict) --

                Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.

                • S3DataSource (dict) --

                  The S3 location of the data source that is associated with a channel.

                  • S3DataType (string) --

                    If you choose S3Prefix, S3Uri identifies a key name prefix. Amazon SageMaker uses all objects with the specified key name prefix for batch transform.

                    If you choose ManifestFile, S3Uri identifies an object that is a manifest file containing a list of object keys that you want Amazon SageMaker to use for batch transform.

                    The following values are compatible: ManifestFile, S3Prefix

                    The following value is not compatible: AugmentedManifestFile

                  • S3Uri (string) --

                    Depending on the value specified for the S3DataType, identifies either a key name prefix or a manifest. For example:

                    • A key name prefix might look like this: s3://bucketname/exampleprefix/.

                    • A manifest might look like this: s3://bucketname/example.manifest The manifest is an S3 object which is a JSON file with the following format: [ {"prefix": "s3://customer_bucket/some/prefix/"}, "relative/path/to/custdata-1", "relative/path/custdata-2", ... "relative/path/custdata-N" ] The preceding JSON matches the following S3Uris: s3://customer_bucket/some/prefix/relative/path/to/custdata-1 s3://customer_bucket/some/prefix/relative/path/custdata-2 ... s3://customer_bucket/some/prefix/relative/path/custdata-N The complete set of S3Uris in this manifest constitutes the input data for the channel for this datasource. The object that each S3Uris points to must be readable by the IAM role that Amazon SageMaker uses to perform tasks on your behalf.

              • ContentType (string) --

                The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.

              • CompressionType (string) --

                If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses the data for the transform job accordingly. The default value is None.

              • SplitType (string) --

                The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the total size of each object is too large to fit in a single request. You can also use data splitting to improve performance by processing multiple concurrent mini-batches. The default value for SplitType is None, which indicates that input data files are not split, and request payloads contain the entire contents of an input object. Set the value of this parameter to Line to split records on a newline character boundary. SplitType also supports a number of record-oriented binary data formats. Currently, the supported record formats are:

                • RecordIO

                • TFRecord

                When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each request.

            • TransformOutput (dict) --

              Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the transform job.

              • S3OutputPath (string) --

                The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example, s3://bucket-name/key-name-prefix.

                For every S3 object used as input for the transform job, batch transform stores the transformed data with an . out suffix in a corresponding subfolder in the location in the output prefix. For example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it creates an . out file only if the transform job succeeds on the entire file. When the input contains multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for successfully processed objects. If any object fails in the transform job batch transform marks the job as failed to prompt investigation.

              • Accept (string) --

                The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to transfer data from the transform job.

              • AssembleWith (string) --

                Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most convenient to you. To concatenate the results in binary format, specify None. To add a newline character at the end of every transformed record, specify Line.

              • KmsKeyId (string) --

                The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any of the following formats:

                • Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab

                • Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab

                • Alias name: alias/ExampleAlias

                • Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias

                If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in the Amazon Simple Storage Service Developer Guide.

                The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. For more information, see Using Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide.

            • TransformResources (dict) --

              Identifies the ML compute instances for the transform job.

              • InstanceType (string) --

                The ML compute instance type for the transform job. If you are using built-in algorithms to transform moderately sized datasets, we recommend using ml.m4.xlarge or ``ml.m5.large``instance types.

              • InstanceCount (integer) --

                The number of ML compute instances to use in the transform job. The default value is 1, and the maximum is 100. For distributed transform jobs, specify a value greater than 1.

              • VolumeKmsKeyId (string) --

                The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to encrypt model data on the storage volume attached to the ML compute instance(s) that run the batch transform job.

                The VolumeKmsKeyId can be any of the following formats:

                • Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab

                • Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab

                • Alias name: alias/ExampleAlias

                • Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias

    • ModelPackageStatus (string) --

      The current status of the model package.

    • ModelPackageStatusDetails (dict) --

      Details about the current status of the model package.

      • ValidationStatuses (list) --

        The validation status of the model package.

        • (dict) --

          Represents the overall status of a model package.

          • Name (string) --

            The name of the model package for which the overall status is being reported.

          • Status (string) --

            The current status.

          • FailureReason (string) --

            if the overall status is Failed, the reason for the failure.

      • ImageScanStatuses (list) --

        The status of the scan of the Docker image container for the model package.

        • (dict) --

          Represents the overall status of a model package.

          • Name (string) --

            The name of the model package for which the overall status is being reported.

          • Status (string) --

            The current status.

          • FailureReason (string) --

            if the overall status is Failed, the reason for the failure.

    • CertifyForMarketplace (boolean) --

      Whether the model package is certified for listing on Amazon Web Services Marketplace.

    • ModelApprovalStatus (string) --

      The approval status of the model package.

    • CreatedBy (dict) --

      Information about the user who created or modified an experiment, trial, trial component, lineage group, project, or model card.

      • UserProfileArn (string) --

        The Amazon Resource Name (ARN) of the user's profile.

      • UserProfileName (string) --

        The name of the user's profile.

      • DomainId (string) --

        The domain associated with the user.

      • IamIdentity (dict) --

        The IAM Identity details associated with the user. These details are associated with model package groups, model packages, and project entities only.

        • Arn (string) --

          The Amazon Resource Name (ARN) of the IAM identity.

        • PrincipalId (string) --

          The ID of the principal that assumes the IAM identity.

        • SourceIdentity (string) --

          The person or application which assumes the IAM identity.

    • MetadataProperties (dict) --

      Metadata properties of the tracking entity, trial, or trial component.

      • CommitId (string) --

        The commit ID.

      • Repository (string) --

        The repository.

      • GeneratedBy (string) --

        The entity this entity was generated by.

      • ProjectId (string) --

        The project ID.

    • ModelMetrics (dict) --

      Metrics for the model.

      • ModelQuality (dict) --

        Metrics that measure the quality of a model.

        • Statistics (dict) --

          Model quality statistics.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • Constraints (dict) --

          Model quality constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

      • ModelDataQuality (dict) --

        Metrics that measure the quality of the input data for a model.

        • Statistics (dict) --

          Data quality statistics for a model.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • Constraints (dict) --

          Data quality constraints for a model.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

      • Bias (dict) --

        Metrics that measure bias in a model.

        • Report (dict) --

          The bias report for a model

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • PreTrainingReport (dict) --

          The pre-training bias report for a model.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • PostTrainingReport (dict) --

          The post-training bias report for a model.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

      • Explainability (dict) --

        Metrics that help explain a model.

        • Report (dict) --

          The explainability report for a model.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

    • LastModifiedTime (datetime) --

      The last time that the model package was modified.

    • LastModifiedBy (dict) --

      Information about the user who created or modified an experiment, trial, trial component, lineage group, project, or model card.

      • UserProfileArn (string) --

        The Amazon Resource Name (ARN) of the user's profile.

      • UserProfileName (string) --

        The name of the user's profile.

      • DomainId (string) --

        The domain associated with the user.

      • IamIdentity (dict) --

        The IAM Identity details associated with the user. These details are associated with model package groups, model packages, and project entities only.

        • Arn (string) --

          The Amazon Resource Name (ARN) of the IAM identity.

        • PrincipalId (string) --

          The ID of the principal that assumes the IAM identity.

        • SourceIdentity (string) --

          The person or application which assumes the IAM identity.

    • ApprovalDescription (string) --

      A description provided for the model approval.

    • Domain (string) --

      The machine learning domain of the model package you specified. Common machine learning domains include computer vision and natural language processing.

    • Task (string) --

      The machine learning task you specified that your model package accomplishes. Common machine learning tasks include object detection and image classification.

    • SamplePayloadUrl (string) --

      The Amazon Simple Storage Service (Amazon S3) path where the sample payload are stored. This path points to a single gzip compressed tar archive (.tar.gz suffix).

    • CustomerMetadataProperties (dict) --

      The metadata properties associated with the model package versions.

      • (string) --

        • (string) --

    • DriftCheckBaselines (dict) --

      Represents the drift check baselines that can be used when the model monitor is set using the model package. For more information, see the topic on Drift Detection against Previous Baselines in SageMaker Pipelines in the Amazon SageMaker Developer Guide.

      • Bias (dict) --

        Represents the drift check bias baselines that can be used when the model monitor is set using the model package.

        • ConfigFile (dict) --

          The bias config file for a model.

          • ContentType (string) --

            The type of content stored in the file source.

          • ContentDigest (string) --

            The digest of the file source.

          • S3Uri (string) --

            The Amazon S3 URI for the file source.

        • PreTrainingConstraints (dict) --

          The pre-training constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • PostTrainingConstraints (dict) --

          The post-training constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

      • Explainability (dict) --

        Represents the drift check explainability baselines that can be used when the model monitor is set using the model package.

        • Constraints (dict) --

          The drift check explainability constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • ConfigFile (dict) --

          The explainability config file for the model.

          • ContentType (string) --

            The type of content stored in the file source.

          • ContentDigest (string) --

            The digest of the file source.

          • S3Uri (string) --

            The Amazon S3 URI for the file source.

      • ModelQuality (dict) --

        Represents the drift check model quality baselines that can be used when the model monitor is set using the model package.

        • Statistics (dict) --

          The drift check model quality statistics.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • Constraints (dict) --

          The drift check model quality constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

      • ModelDataQuality (dict) --

        Represents the drift check model data quality baselines that can be used when the model monitor is set using the model package.

        • Statistics (dict) --

          The drift check model data quality statistics.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

        • Constraints (dict) --

          The drift check model data quality constraints.

          • ContentType (string) --

            The metric source content type.

          • ContentDigest (string) --

            The hash key used for the metrics source.

          • S3Uri (string) --

            The S3 URI for the metrics source.

    • AdditionalInferenceSpecifications (list) --

      An array of additional Inference Specification objects. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.

      • (dict) --

        A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package

        • Name (string) --

          A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.

        • Description (string) --

          A description of the additional Inference specification

        • Containers (list) --

          The Amazon ECR registry path of the Docker image that contains the inference code.

          • (dict) --

            Describes the Docker container for the model package.

            • ContainerHostname (string) --

              The DNS host name for the Docker container.

            • Image (string) --

              The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

              If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

            • ImageDigest (string) --

              An MD5 hash of the training algorithm that identifies the Docker image used for training.

            • ModelDataUrl (string) --

              The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

            • ModelDataSource (dict) --

              Specifies the location of ML model data to deploy during endpoint creation.

              • S3DataSource (dict) --

                Specifies the S3 location of ML model data to deploy.

                • S3Uri (string) --

                  Specifies the S3 path of ML model data to deploy.

                • S3DataType (string) --

                  Specifies the type of ML model data to deploy.

                  If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

                  If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

                • CompressionType (string) --

                  Specifies how the ML model data is prepared.

                  If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

                  If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

                  If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

                  If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

                  • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

                  • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

                  • Do not use any of the following as file names or directory names:

                    • An empty or blank string

                    • A string which contains null bytes

                    • A string longer than 255 bytes

                    • A single dot ( .)

                    • A double dot ( ..)

                  • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

                  • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

                • ModelAccessConfig (dict) --

                  Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

                  • AcceptEula (boolean) --

                    Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

            • ProductId (string) --

              The Amazon Web Services Marketplace product ID of the model package.

            • Environment (dict) --

              The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

              • (string) --

                • (string) --

            • ModelInput (dict) --

              A structure with Model Input details.

              • DataInputConfig (string) --

                The input configuration object for the model.

            • Framework (string) --

              The machine learning framework of the model package container image.

            • FrameworkVersion (string) --

              The framework version of the Model Package Container Image.

            • NearestModelName (string) --

              The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

            • AdditionalS3DataSource (dict) --

              The additional data source that is used during inference in the Docker container for your model package.

              • S3DataType (string) --

                The data type of the additional data source that you specify for use in inference or training.

              • S3Uri (string) --

                The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

              • CompressionType (string) --

                The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

        • SupportedTransformInstanceTypes (list) --

          A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

          • (string) --

        • SupportedRealtimeInferenceInstanceTypes (list) --

          A list of the instance types that are used to generate inferences in real-time.

          • (string) --

        • SupportedContentTypes (list) --

          The supported MIME types for the input data.

          • (string) --

        • SupportedResponseMIMETypes (list) --

          The supported MIME types for the output data.

          • (string) --

    • SkipModelValidation (string) --

      Indicates if you want to skip model validation.

    • SourceUri (string) --

      The URI of the source for the model package.

    • SecurityConfig (dict) --

      The KMS Key ID ( KMSKeyId) used for encryption of model package information.

      • KmsKeyId (string) --

        The KMS Key ID ( KMSKeyId) used for encryption of model package information.

    • ModelCard (dict) --

      The model card associated with the model package. Since ModelPackageModelCard is tied to a model package, it is a specific usage of a model card and its schema is simplified compared to the schema of ModelCard. The ModelPackageModelCard schema does not include model_package_details, and model_overview is composed of the model_creator and model_artifact properties. For more information about the model card associated with the model package, see View the Details of a Model Version.

      • ModelCardContent (string) --

        The content of the model card.

      • ModelCardStatus (string) --

        The approval status of the model card within your organization. Different organizations might have different criteria for model card review and approval.

        • Draft: The model card is a work in progress.

        • PendingReview: The model card is pending review.

        • Approved: The model card is approved.

        • Archived: The model card is archived. No more updates can be made to the model card content. If you try to update the model card content, you will receive the message Model Card is in Archived state.

UpdateModelPackage (updated) Link ¶
Changes (request)
{'ModelCard': {'ModelCardContent': 'string',
               'ModelCardStatus': 'Draft | PendingReview | Approved | '
                                  'Archived'}}

Updates a versioned model.

See also: AWS API Documentation

Request Syntax

client.update_model_package(
    ModelPackageArn='string',
    ModelApprovalStatus='Approved'|'Rejected'|'PendingManualApproval',
    ApprovalDescription='string',
    CustomerMetadataProperties={
        'string': 'string'
    },
    CustomerMetadataPropertiesToRemove=[
        'string',
    ],
    AdditionalInferenceSpecificationsToAdd=[
        {
            'Name': 'string',
            'Description': 'string',
            'Containers': [
                {
                    'ContainerHostname': 'string',
                    'Image': 'string',
                    'ImageDigest': 'string',
                    'ModelDataUrl': 'string',
                    'ModelDataSource': {
                        'S3DataSource': {
                            'S3Uri': 'string',
                            'S3DataType': 'S3Prefix'|'S3Object',
                            'CompressionType': 'None'|'Gzip',
                            'ModelAccessConfig': {
                                'AcceptEula': True|False
                            }
                        }
                    },
                    'ProductId': 'string',
                    'Environment': {
                        'string': 'string'
                    },
                    'ModelInput': {
                        'DataInputConfig': 'string'
                    },
                    'Framework': 'string',
                    'FrameworkVersion': 'string',
                    'NearestModelName': 'string',
                    'AdditionalS3DataSource': {
                        'S3DataType': 'S3Object'|'S3Prefix',
                        'S3Uri': 'string',
                        'CompressionType': 'None'|'Gzip'
                    }
                },
            ],
            'SupportedTransformInstanceTypes': [
                'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
            ],
            'SupportedRealtimeInferenceInstanceTypes': [
                'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
            ],
            'SupportedContentTypes': [
                'string',
            ],
            'SupportedResponseMIMETypes': [
                'string',
            ]
        },
    ],
    InferenceSpecification={
        'Containers': [
            {
                'ContainerHostname': 'string',
                'Image': 'string',
                'ImageDigest': 'string',
                'ModelDataUrl': 'string',
                'ModelDataSource': {
                    'S3DataSource': {
                        'S3Uri': 'string',
                        'S3DataType': 'S3Prefix'|'S3Object',
                        'CompressionType': 'None'|'Gzip',
                        'ModelAccessConfig': {
                            'AcceptEula': True|False
                        }
                    }
                },
                'ProductId': 'string',
                'Environment': {
                    'string': 'string'
                },
                'ModelInput': {
                    'DataInputConfig': 'string'
                },
                'Framework': 'string',
                'FrameworkVersion': 'string',
                'NearestModelName': 'string',
                'AdditionalS3DataSource': {
                    'S3DataType': 'S3Object'|'S3Prefix',
                    'S3Uri': 'string',
                    'CompressionType': 'None'|'Gzip'
                }
            },
        ],
        'SupportedTransformInstanceTypes': [
            'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m6i.large'|'ml.m6i.xlarge'|'ml.m6i.2xlarge'|'ml.m6i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.12xlarge'|'ml.m6i.16xlarge'|'ml.m6i.24xlarge'|'ml.m6i.32xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.r6i.large'|'ml.r6i.xlarge'|'ml.r6i.2xlarge'|'ml.r6i.4xlarge'|'ml.r6i.8xlarge'|'ml.r6i.12xlarge'|'ml.r6i.16xlarge'|'ml.r6i.24xlarge'|'ml.r6i.32xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge',
        ],
        'SupportedRealtimeInferenceInstanceTypes': [
            'ml.t2.medium'|'ml.t2.large'|'ml.t2.xlarge'|'ml.t2.2xlarge'|'ml.m4.xlarge'|'ml.m4.2xlarge'|'ml.m4.4xlarge'|'ml.m4.10xlarge'|'ml.m4.16xlarge'|'ml.m5.large'|'ml.m5.xlarge'|'ml.m5.2xlarge'|'ml.m5.4xlarge'|'ml.m5.12xlarge'|'ml.m5.24xlarge'|'ml.m5d.large'|'ml.m5d.xlarge'|'ml.m5d.2xlarge'|'ml.m5d.4xlarge'|'ml.m5d.12xlarge'|'ml.m5d.24xlarge'|'ml.c4.large'|'ml.c4.xlarge'|'ml.c4.2xlarge'|'ml.c4.4xlarge'|'ml.c4.8xlarge'|'ml.p2.xlarge'|'ml.p2.8xlarge'|'ml.p2.16xlarge'|'ml.p3.2xlarge'|'ml.p3.8xlarge'|'ml.p3.16xlarge'|'ml.c5.large'|'ml.c5.xlarge'|'ml.c5.2xlarge'|'ml.c5.4xlarge'|'ml.c5.9xlarge'|'ml.c5.18xlarge'|'ml.c5d.large'|'ml.c5d.xlarge'|'ml.c5d.2xlarge'|'ml.c5d.4xlarge'|'ml.c5d.9xlarge'|'ml.c5d.18xlarge'|'ml.g4dn.xlarge'|'ml.g4dn.2xlarge'|'ml.g4dn.4xlarge'|'ml.g4dn.8xlarge'|'ml.g4dn.12xlarge'|'ml.g4dn.16xlarge'|'ml.r5.large'|'ml.r5.xlarge'|'ml.r5.2xlarge'|'ml.r5.4xlarge'|'ml.r5.12xlarge'|'ml.r5.24xlarge'|'ml.r5d.large'|'ml.r5d.xlarge'|'ml.r5d.2xlarge'|'ml.r5d.4xlarge'|'ml.r5d.12xlarge'|'ml.r5d.24xlarge'|'ml.inf1.xlarge'|'ml.inf1.2xlarge'|'ml.inf1.6xlarge'|'ml.inf1.24xlarge'|'ml.dl1.24xlarge'|'ml.c6i.large'|'ml.c6i.xlarge'|'ml.c6i.2xlarge'|'ml.c6i.4xlarge'|'ml.c6i.8xlarge'|'ml.c6i.12xlarge'|'ml.c6i.16xlarge'|'ml.c6i.24xlarge'|'ml.c6i.32xlarge'|'ml.g5.xlarge'|'ml.g5.2xlarge'|'ml.g5.4xlarge'|'ml.g5.8xlarge'|'ml.g5.12xlarge'|'ml.g5.16xlarge'|'ml.g5.24xlarge'|'ml.g5.48xlarge'|'ml.g6.xlarge'|'ml.g6.2xlarge'|'ml.g6.4xlarge'|'ml.g6.8xlarge'|'ml.g6.12xlarge'|'ml.g6.16xlarge'|'ml.g6.24xlarge'|'ml.g6.48xlarge'|'ml.p4d.24xlarge'|'ml.c7g.large'|'ml.c7g.xlarge'|'ml.c7g.2xlarge'|'ml.c7g.4xlarge'|'ml.c7g.8xlarge'|'ml.c7g.12xlarge'|'ml.c7g.16xlarge'|'ml.m6g.large'|'ml.m6g.xlarge'|'ml.m6g.2xlarge'|'ml.m6g.4xlarge'|'ml.m6g.8xlarge'|'ml.m6g.12xlarge'|'ml.m6g.16xlarge'|'ml.m6gd.large'|'ml.m6gd.xlarge'|'ml.m6gd.2xlarge'|'ml.m6gd.4xlarge'|'ml.m6gd.8xlarge'|'ml.m6gd.12xlarge'|'ml.m6gd.16xlarge'|'ml.c6g.large'|'ml.c6g.xlarge'|'ml.c6g.2xlarge'|'ml.c6g.4xlarge'|'ml.c6g.8xlarge'|'ml.c6g.12xlarge'|'ml.c6g.16xlarge'|'ml.c6gd.large'|'ml.c6gd.xlarge'|'ml.c6gd.2xlarge'|'ml.c6gd.4xlarge'|'ml.c6gd.8xlarge'|'ml.c6gd.12xlarge'|'ml.c6gd.16xlarge'|'ml.c6gn.large'|'ml.c6gn.xlarge'|'ml.c6gn.2xlarge'|'ml.c6gn.4xlarge'|'ml.c6gn.8xlarge'|'ml.c6gn.12xlarge'|'ml.c6gn.16xlarge'|'ml.r6g.large'|'ml.r6g.xlarge'|'ml.r6g.2xlarge'|'ml.r6g.4xlarge'|'ml.r6g.8xlarge'|'ml.r6g.12xlarge'|'ml.r6g.16xlarge'|'ml.r6gd.large'|'ml.r6gd.xlarge'|'ml.r6gd.2xlarge'|'ml.r6gd.4xlarge'|'ml.r6gd.8xlarge'|'ml.r6gd.12xlarge'|'ml.r6gd.16xlarge'|'ml.p4de.24xlarge'|'ml.trn1.2xlarge'|'ml.trn1.32xlarge'|'ml.trn1n.32xlarge'|'ml.inf2.xlarge'|'ml.inf2.8xlarge'|'ml.inf2.24xlarge'|'ml.inf2.48xlarge'|'ml.p5.48xlarge'|'ml.m7i.large'|'ml.m7i.xlarge'|'ml.m7i.2xlarge'|'ml.m7i.4xlarge'|'ml.m7i.8xlarge'|'ml.m7i.12xlarge'|'ml.m7i.16xlarge'|'ml.m7i.24xlarge'|'ml.m7i.48xlarge'|'ml.c7i.large'|'ml.c7i.xlarge'|'ml.c7i.2xlarge'|'ml.c7i.4xlarge'|'ml.c7i.8xlarge'|'ml.c7i.12xlarge'|'ml.c7i.16xlarge'|'ml.c7i.24xlarge'|'ml.c7i.48xlarge'|'ml.r7i.large'|'ml.r7i.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.4xlarge'|'ml.r7i.8xlarge'|'ml.r7i.12xlarge'|'ml.r7i.16xlarge'|'ml.r7i.24xlarge'|'ml.r7i.48xlarge',
        ],
        'SupportedContentTypes': [
            'string',
        ],
        'SupportedResponseMIMETypes': [
            'string',
        ]
    },
    SourceUri='string',
    ModelCard={
        'ModelCardContent': 'string',
        'ModelCardStatus': 'Draft'|'PendingReview'|'Approved'|'Archived'
    }
)
type ModelPackageArn:

string

param ModelPackageArn:

[REQUIRED]

The Amazon Resource Name (ARN) of the model package.

type ModelApprovalStatus:

string

param ModelApprovalStatus:

The approval status of the model.

type ApprovalDescription:

string

param ApprovalDescription:

A description for the approval status of the model.

type CustomerMetadataProperties:

dict

param CustomerMetadataProperties:

The metadata properties associated with the model package versions.

  • (string) --

    • (string) --

type CustomerMetadataPropertiesToRemove:

list

param CustomerMetadataPropertiesToRemove:

The metadata properties associated with the model package versions to remove.

  • (string) --

type AdditionalInferenceSpecificationsToAdd:

list

param AdditionalInferenceSpecificationsToAdd:

An array of additional Inference Specification objects to be added to the existing array additional Inference Specification. Total number of additional Inference Specifications can not exceed 15. Each additional Inference Specification specifies artifacts based on this model package that can be used on inference endpoints. Generally used with SageMaker Neo to store the compiled artifacts.

  • (dict) --

    A structure of additional Inference Specification. Additional Inference Specification specifies details about inference jobs that can be run with models based on this model package

    • Name (string) -- [REQUIRED]

      A unique name to identify the additional inference specification. The name must be unique within the list of your additional inference specifications for a particular model package.

    • Description (string) --

      A description of the additional Inference specification

    • Containers (list) -- [REQUIRED]

      The Amazon ECR registry path of the Docker image that contains the inference code.

      • (dict) --

        Describes the Docker container for the model package.

        • ContainerHostname (string) --

          The DNS host name for the Docker container.

        • Image (string) -- [REQUIRED]

          The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

          If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

        • ImageDigest (string) --

          An MD5 hash of the training algorithm that identifies the Docker image used for training.

        • ModelDataUrl (string) --

          The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

        • ModelDataSource (dict) --

          Specifies the location of ML model data to deploy during endpoint creation.

          • S3DataSource (dict) --

            Specifies the S3 location of ML model data to deploy.

            • S3Uri (string) -- [REQUIRED]

              Specifies the S3 path of ML model data to deploy.

            • S3DataType (string) -- [REQUIRED]

              Specifies the type of ML model data to deploy.

              If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

              If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

            • CompressionType (string) -- [REQUIRED]

              Specifies how the ML model data is prepared.

              If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

              If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

              If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

              If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

              • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

              • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

              • Do not use any of the following as file names or directory names:

                • An empty or blank string

                • A string which contains null bytes

                • A string longer than 255 bytes

                • A single dot ( .)

                • A double dot ( ..)

              • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

              • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

            • ModelAccessConfig (dict) --

              Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

              • AcceptEula (boolean) -- [REQUIRED]

                Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

        • ProductId (string) --

          The Amazon Web Services Marketplace product ID of the model package.

        • Environment (dict) --

          The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

          • (string) --

            • (string) --

        • ModelInput (dict) --

          A structure with Model Input details.

          • DataInputConfig (string) -- [REQUIRED]

            The input configuration object for the model.

        • Framework (string) --

          The machine learning framework of the model package container image.

        • FrameworkVersion (string) --

          The framework version of the Model Package Container Image.

        • NearestModelName (string) --

          The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

        • AdditionalS3DataSource (dict) --

          The additional data source that is used during inference in the Docker container for your model package.

          • S3DataType (string) -- [REQUIRED]

            The data type of the additional data source that you specify for use in inference or training.

          • S3Uri (string) -- [REQUIRED]

            The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

          • CompressionType (string) --

            The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

    • SupportedTransformInstanceTypes (list) --

      A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

      • (string) --

    • SupportedRealtimeInferenceInstanceTypes (list) --

      A list of the instance types that are used to generate inferences in real-time.

      • (string) --

    • SupportedContentTypes (list) --

      The supported MIME types for the input data.

      • (string) --

    • SupportedResponseMIMETypes (list) --

      The supported MIME types for the output data.

      • (string) --

type InferenceSpecification:

dict

param InferenceSpecification:

Specifies details about inference jobs that you can run with models based on this model package, including the following information:

  • The Amazon ECR paths of containers that contain the inference code and model artifacts.

  • The instance types that the model package supports for transform jobs and real-time endpoints used for inference.

  • The input and output content formats that the model package supports for inference.

  • Containers (list) -- [REQUIRED]

    The Amazon ECR registry path of the Docker image that contains the inference code.

    • (dict) --

      Describes the Docker container for the model package.

      • ContainerHostname (string) --

        The DNS host name for the Docker container.

      • Image (string) -- [REQUIRED]

        The Amazon EC2 Container Registry (Amazon ECR) path where inference code is stored.

        If you are using your own custom algorithm instead of an algorithm provided by SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both registry/repository[:tag] and registry/repository[@digest] image path formats. For more information, see Using Your Own Algorithms with Amazon SageMaker.

      • ImageDigest (string) --

        An MD5 hash of the training algorithm that identifies the Docker image used for training.

      • ModelDataUrl (string) --

        The Amazon S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive ( .tar.gz suffix).

      • ModelDataSource (dict) --

        Specifies the location of ML model data to deploy during endpoint creation.

        • S3DataSource (dict) --

          Specifies the S3 location of ML model data to deploy.

          • S3Uri (string) -- [REQUIRED]

            Specifies the S3 path of ML model data to deploy.

          • S3DataType (string) -- [REQUIRED]

            Specifies the type of ML model data to deploy.

            If you choose S3Prefix, S3Uri identifies a key name prefix. SageMaker uses all objects that match the specified key name prefix as part of the ML model data to deploy. A valid key name prefix identified by S3Uri always ends with a forward slash (/).

            If you choose S3Object, S3Uri identifies an object that is the ML model data to deploy.

          • CompressionType (string) -- [REQUIRED]

            Specifies how the ML model data is prepared.

            If you choose Gzip and choose S3Object as the value of S3DataType, S3Uri identifies an object that is a gzip-compressed TAR archive. SageMaker will attempt to decompress and untar the object during model deployment.

            If you choose None and chooose S3Object as the value of S3DataType, S3Uri identifies an object that represents an uncompressed ML model to deploy.

            If you choose None and choose S3Prefix as the value of S3DataType, S3Uri identifies a key name prefix, under which all objects represents the uncompressed ML model to deploy.

            If you choose None, then SageMaker will follow rules below when creating model data files under /opt/ml/model directory for use by your inference code:

            • If you choose S3Object as the value of S3DataType, then SageMaker will split the key of the S3 object referenced by S3Uri by slash (/), and use the last part as the filename of the file holding the content of the S3 object.

            • If you choose S3Prefix as the value of S3DataType, then for each S3 object under the key name pefix referenced by S3Uri, SageMaker will trim its key by the prefix, and use the remainder as the path (relative to /opt/ml/model) of the file holding the content of the S3 object. SageMaker will split the remainder by slash (/), using intermediate parts as directory names and the last part as filename of the file holding the content of the S3 object.

            • Do not use any of the following as file names or directory names:

              • An empty or blank string

              • A string which contains null bytes

              • A string longer than 255 bytes

              • A single dot ( .)

              • A double dot ( ..)

            • Ambiguous file names will result in model deployment failure. For example, if your uncompressed ML model consists of two S3 objects s3://mybucket/model/weights and s3://mybucket/model/weights/part1 and you specify s3://mybucket/model/ as the value of S3Uri and S3Prefix as the value of S3DataType, then it will result in name clash between /opt/ml/model/weights (a regular file) and /opt/ml/model/weights/ (a directory).

            • Do not organize the model artifacts in S3 console using folders. When you create a folder in S3 console, S3 creates a 0-byte object with a key set to the folder name you provide. They key of the 0-byte object ends with a slash (/) which violates SageMaker restrictions on model artifact file names, leading to model deployment failure.

          • ModelAccessConfig (dict) --

            Specifies the access configuration file for the ML model. You can explicitly accept the model end-user license agreement (EULA) within the ModelAccessConfig. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

            • AcceptEula (boolean) -- [REQUIRED]

              Specifies agreement to the model end-user license agreement (EULA). The AcceptEula value must be explicitly defined as True in order to accept the EULA that this model requires. You are responsible for reviewing and complying with any applicable license terms and making sure they are acceptable for your use case before downloading or using a model.

      • ProductId (string) --

        The Amazon Web Services Marketplace product ID of the model package.

      • Environment (dict) --

        The environment variables to set in the Docker container. Each key and value in the Environment string to string map can have length of up to 1024. We support up to 16 entries in the map.

        • (string) --

          • (string) --

      • ModelInput (dict) --

        A structure with Model Input details.

        • DataInputConfig (string) -- [REQUIRED]

          The input configuration object for the model.

      • Framework (string) --

        The machine learning framework of the model package container image.

      • FrameworkVersion (string) --

        The framework version of the Model Package Container Image.

      • NearestModelName (string) --

        The name of a pre-trained machine learning benchmarked by Amazon SageMaker Inference Recommender model that matches your model. You can find a list of benchmarked models by calling ListModelMetadata.

      • AdditionalS3DataSource (dict) --

        The additional data source that is used during inference in the Docker container for your model package.

        • S3DataType (string) -- [REQUIRED]

          The data type of the additional data source that you specify for use in inference or training.

        • S3Uri (string) -- [REQUIRED]

          The uniform resource identifier (URI) used to identify an additional data source used in inference or training.

        • CompressionType (string) --

          The type of compression used for an additional data source used in inference or training. Specify None if your additional data source is not compressed.

  • SupportedTransformInstanceTypes (list) --

    A list of the instance types on which a transformation job can be run or on which an endpoint can be deployed.

    This parameter is required for unversioned models, and optional for versioned models.

    • (string) --

  • SupportedRealtimeInferenceInstanceTypes (list) --

    A list of the instance types that are used to generate inferences in real-time.

    This parameter is required for unversioned models, and optional for versioned models.

    • (string) --

  • SupportedContentTypes (list) --

    The supported MIME types for the input data.

    • (string) --

  • SupportedResponseMIMETypes (list) --

    The supported MIME types for the output data.

    • (string) --

type SourceUri:

string

param SourceUri:

The URI of the source for the model package.

type ModelCard:

dict

param ModelCard:

The model card associated with the model package. Since ModelPackageModelCard is tied to a model package, it is a specific usage of a model card and its schema is simplified compared to the schema of ModelCard. The ModelPackageModelCard schema does not include model_package_details, and model_overview is composed of the model_creator and model_artifact properties. For more information about the model card associated with the model package, see View the Details of a Model Version.

  • ModelCardContent (string) --

    The content of the model card.

  • ModelCardStatus (string) --

    The approval status of the model card within your organization. Different organizations might have different criteria for model card review and approval.

    • Draft: The model card is a work in progress.

    • PendingReview: The model card is pending review.

    • Approved: The model card is approved.

    • Archived: The model card is archived. No more updates can be made to the model card content. If you try to update the model card content, you will receive the message Model Card is in Archived state.

rtype:

dict

returns:

Response Syntax

{
    'ModelPackageArn': 'string'
}

Response Structure

  • (dict) --

    • ModelPackageArn (string) --

      The Amazon Resource Name (ARN) of the model.