Amazon EMR Containers

2023/02/09 - Amazon EMR Containers - 3 updated api methods

Changes  EMR on EKS allows configuring retry policies for job runs through the StartJobRun API. Using retry policies, a job cause a driver pod to be restarted automatically if it fails or is deleted. The job's status can be seen in the DescribeJobRun and ListJobRun APIs and monitored using CloudWatch events.

DescribeJobRun (updated) Link ¶
Changes (response)
{'jobRun': {'retryPolicyConfiguration': {'maxAttempts': 'integer'},
            'retryPolicyExecution': {'currentAttemptCount': 'integer'}}}

Displays detailed information about a job run. A job run is a unit of work, such as a Spark jar, PySpark script, or SparkSQL query, that you submit to Amazon EMR on EKS.

See also: AWS API Documentation

Request Syntax

client.describe_job_run(
    id='string',
    virtualClusterId='string'
)
type id:

string

param id:

[REQUIRED]

The ID of the job run request.

type virtualClusterId:

string

param virtualClusterId:

[REQUIRED]

The ID of the virtual cluster for which the job run is submitted.

rtype:

dict

returns:

Response Syntax

{
    'jobRun': {
        'id': 'string',
        'name': 'string',
        'virtualClusterId': 'string',
        'arn': 'string',
        'state': 'PENDING'|'SUBMITTED'|'RUNNING'|'FAILED'|'CANCELLED'|'CANCEL_PENDING'|'COMPLETED',
        'clientToken': 'string',
        'executionRoleArn': 'string',
        'releaseLabel': 'string',
        'configurationOverrides': {
            'applicationConfiguration': [
                {
                    'classification': 'string',
                    'properties': {
                        'string': 'string'
                    },
                    'configurations': {'... recursive ...'}
                },
            ],
            'monitoringConfiguration': {
                'persistentAppUI': 'ENABLED'|'DISABLED',
                'cloudWatchMonitoringConfiguration': {
                    'logGroupName': 'string',
                    'logStreamNamePrefix': 'string'
                },
                's3MonitoringConfiguration': {
                    'logUri': 'string'
                }
            }
        },
        'jobDriver': {
            'sparkSubmitJobDriver': {
                'entryPoint': 'string',
                'entryPointArguments': [
                    'string',
                ],
                'sparkSubmitParameters': 'string'
            },
            'sparkSqlJobDriver': {
                'entryPoint': 'string',
                'sparkSqlParameters': 'string'
            }
        },
        'createdAt': datetime(2015, 1, 1),
        'createdBy': 'string',
        'finishedAt': datetime(2015, 1, 1),
        'stateDetails': 'string',
        'failureReason': 'INTERNAL_ERROR'|'USER_ERROR'|'VALIDATION_ERROR'|'CLUSTER_UNAVAILABLE',
        'tags': {
            'string': 'string'
        },
        'retryPolicyConfiguration': {
            'maxAttempts': 123
        },
        'retryPolicyExecution': {
            'currentAttemptCount': 123
        }
    }
}

Response Structure

  • (dict) --

    • jobRun (dict) --

      The output displays information about a job run.

      • id (string) --

        The ID of the job run.

      • name (string) --

        The name of the job run.

      • virtualClusterId (string) --

        The ID of the job run's virtual cluster.

      • arn (string) --

        The ARN of job run.

      • state (string) --

        The state of the job run.

      • clientToken (string) --

        The client token used to start a job run.

      • executionRoleArn (string) --

        The execution role ARN of the job run.

      • releaseLabel (string) --

        The release version of Amazon EMR.

      • configurationOverrides (dict) --

        The configuration settings that are used to override default configuration.

        • applicationConfiguration (list) --

          The configurations for the application running by the job run.

          • (dict) --

            A configuration specification to be used when provisioning virtual clusters, which can include configurations for applications and software bundled with Amazon EMR on EKS. A configuration consists of a classification, properties, and optional nested configurations. A classification refers to an application-specific configuration file. Properties are the settings you want to change in that file.

            • classification (string) --

              The classification within a configuration.

            • properties (dict) --

              A set of properties specified within a configuration classification.

              • (string) --

                • (string) --

            • configurations (list) --

              A list of additional configurations to apply within a configuration object.

        • monitoringConfiguration (dict) --

          The configurations for monitoring.

          • persistentAppUI (string) --

            Monitoring configurations for the persistent application UI.

          • cloudWatchMonitoringConfiguration (dict) --

            Monitoring configurations for CloudWatch.

            • logGroupName (string) --

              The name of the log group for log publishing.

            • logStreamNamePrefix (string) --

              The specified name prefix for log streams.

          • s3MonitoringConfiguration (dict) --

            Amazon S3 configuration for monitoring log publishing.

            • logUri (string) --

              Amazon S3 destination URI for log publishing.

      • jobDriver (dict) --

        Parameters of job driver for the job run.

        • sparkSubmitJobDriver (dict) --

          The job driver parameters specified for spark submit.

          • entryPoint (string) --

            The entry point of job application.

          • entryPointArguments (list) --

            The arguments for job application.

            • (string) --

          • sparkSubmitParameters (string) --

            The Spark submit parameters that are used for job runs.

        • sparkSqlJobDriver (dict) --

          The job driver for job type.

          • entryPoint (string) --

            The SQL file to be executed.

          • sparkSqlParameters (string) --

            The Spark parameters to be included in the Spark SQL command.

      • createdAt (datetime) --

        The date and time when the job run was created.

      • createdBy (string) --

        The user who created the job run.

      • finishedAt (datetime) --

        The date and time when the job run has finished.

      • stateDetails (string) --

        Additional details of the job run state.

      • failureReason (string) --

        The reasons why the job run has failed.

      • tags (dict) --

        The assigned tags of the job run.

        • (string) --

          • (string) --

      • retryPolicyConfiguration (dict) --

        The configuration of the retry policy that the job runs on.

        • maxAttempts (integer) --

          The maximum number of attempts on the job's driver.

      • retryPolicyExecution (dict) --

        The current status of the retry policy executed on the job.

        • currentAttemptCount (integer) --

          The current number of attempts made on the driver of the job.

ListJobRuns (updated) Link ¶
Changes (response)
{'jobRuns': {'retryPolicyConfiguration': {'maxAttempts': 'integer'},
             'retryPolicyExecution': {'currentAttemptCount': 'integer'}}}

Lists job runs based on a set of parameters. A job run is a unit of work, such as a Spark jar, PySpark script, or SparkSQL query, that you submit to Amazon EMR on EKS.

See also: AWS API Documentation

Request Syntax

client.list_job_runs(
    virtualClusterId='string',
    createdBefore=datetime(2015, 1, 1),
    createdAfter=datetime(2015, 1, 1),
    name='string',
    states=[
        'PENDING'|'SUBMITTED'|'RUNNING'|'FAILED'|'CANCELLED'|'CANCEL_PENDING'|'COMPLETED',
    ],
    maxResults=123,
    nextToken='string'
)
type virtualClusterId:

string

param virtualClusterId:

[REQUIRED]

The ID of the virtual cluster for which to list the job run.

type createdBefore:

datetime

param createdBefore:

The date and time before which the job runs were submitted.

type createdAfter:

datetime

param createdAfter:

The date and time after which the job runs were submitted.

type name:

string

param name:

The name of the job run.

type states:

list

param states:

The states of the job run.

  • (string) --

type maxResults:

integer

param maxResults:

The maximum number of job runs that can be listed.

type nextToken:

string

param nextToken:

The token for the next set of job runs to return.

rtype:

dict

returns:

Response Syntax

{
    'jobRuns': [
        {
            'id': 'string',
            'name': 'string',
            'virtualClusterId': 'string',
            'arn': 'string',
            'state': 'PENDING'|'SUBMITTED'|'RUNNING'|'FAILED'|'CANCELLED'|'CANCEL_PENDING'|'COMPLETED',
            'clientToken': 'string',
            'executionRoleArn': 'string',
            'releaseLabel': 'string',
            'configurationOverrides': {
                'applicationConfiguration': [
                    {
                        'classification': 'string',
                        'properties': {
                            'string': 'string'
                        },
                        'configurations': {'... recursive ...'}
                    },
                ],
                'monitoringConfiguration': {
                    'persistentAppUI': 'ENABLED'|'DISABLED',
                    'cloudWatchMonitoringConfiguration': {
                        'logGroupName': 'string',
                        'logStreamNamePrefix': 'string'
                    },
                    's3MonitoringConfiguration': {
                        'logUri': 'string'
                    }
                }
            },
            'jobDriver': {
                'sparkSubmitJobDriver': {
                    'entryPoint': 'string',
                    'entryPointArguments': [
                        'string',
                    ],
                    'sparkSubmitParameters': 'string'
                },
                'sparkSqlJobDriver': {
                    'entryPoint': 'string',
                    'sparkSqlParameters': 'string'
                }
            },
            'createdAt': datetime(2015, 1, 1),
            'createdBy': 'string',
            'finishedAt': datetime(2015, 1, 1),
            'stateDetails': 'string',
            'failureReason': 'INTERNAL_ERROR'|'USER_ERROR'|'VALIDATION_ERROR'|'CLUSTER_UNAVAILABLE',
            'tags': {
                'string': 'string'
            },
            'retryPolicyConfiguration': {
                'maxAttempts': 123
            },
            'retryPolicyExecution': {
                'currentAttemptCount': 123
            }
        },
    ],
    'nextToken': 'string'
}

Response Structure

  • (dict) --

    • jobRuns (list) --

      This output lists information about the specified job runs.

      • (dict) --

        This entity describes a job run. A job run is a unit of work, such as a Spark jar, PySpark script, or SparkSQL query, that you submit to Amazon EMR on EKS.

        • id (string) --

          The ID of the job run.

        • name (string) --

          The name of the job run.

        • virtualClusterId (string) --

          The ID of the job run's virtual cluster.

        • arn (string) --

          The ARN of job run.

        • state (string) --

          The state of the job run.

        • clientToken (string) --

          The client token used to start a job run.

        • executionRoleArn (string) --

          The execution role ARN of the job run.

        • releaseLabel (string) --

          The release version of Amazon EMR.

        • configurationOverrides (dict) --

          The configuration settings that are used to override default configuration.

          • applicationConfiguration (list) --

            The configurations for the application running by the job run.

            • (dict) --

              A configuration specification to be used when provisioning virtual clusters, which can include configurations for applications and software bundled with Amazon EMR on EKS. A configuration consists of a classification, properties, and optional nested configurations. A classification refers to an application-specific configuration file. Properties are the settings you want to change in that file.

              • classification (string) --

                The classification within a configuration.

              • properties (dict) --

                A set of properties specified within a configuration classification.

                • (string) --

                  • (string) --

              • configurations (list) --

                A list of additional configurations to apply within a configuration object.

          • monitoringConfiguration (dict) --

            The configurations for monitoring.

            • persistentAppUI (string) --

              Monitoring configurations for the persistent application UI.

            • cloudWatchMonitoringConfiguration (dict) --

              Monitoring configurations for CloudWatch.

              • logGroupName (string) --

                The name of the log group for log publishing.

              • logStreamNamePrefix (string) --

                The specified name prefix for log streams.

            • s3MonitoringConfiguration (dict) --

              Amazon S3 configuration for monitoring log publishing.

              • logUri (string) --

                Amazon S3 destination URI for log publishing.

        • jobDriver (dict) --

          Parameters of job driver for the job run.

          • sparkSubmitJobDriver (dict) --

            The job driver parameters specified for spark submit.

            • entryPoint (string) --

              The entry point of job application.

            • entryPointArguments (list) --

              The arguments for job application.

              • (string) --

            • sparkSubmitParameters (string) --

              The Spark submit parameters that are used for job runs.

          • sparkSqlJobDriver (dict) --

            The job driver for job type.

            • entryPoint (string) --

              The SQL file to be executed.

            • sparkSqlParameters (string) --

              The Spark parameters to be included in the Spark SQL command.

        • createdAt (datetime) --

          The date and time when the job run was created.

        • createdBy (string) --

          The user who created the job run.

        • finishedAt (datetime) --

          The date and time when the job run has finished.

        • stateDetails (string) --

          Additional details of the job run state.

        • failureReason (string) --

          The reasons why the job run has failed.

        • tags (dict) --

          The assigned tags of the job run.

          • (string) --

            • (string) --

        • retryPolicyConfiguration (dict) --

          The configuration of the retry policy that the job runs on.

          • maxAttempts (integer) --

            The maximum number of attempts on the job's driver.

        • retryPolicyExecution (dict) --

          The current status of the retry policy executed on the job.

          • currentAttemptCount (integer) --

            The current number of attempts made on the driver of the job.

    • nextToken (string) --

      This output displays the token for the next set of job runs.

StartJobRun (updated) Link ¶
Changes (request)
{'retryPolicyConfiguration': {'maxAttempts': 'integer'}}

Starts a job run. A job run is a unit of work, such as a Spark jar, PySpark script, or SparkSQL query, that you submit to Amazon EMR on EKS.

See also: AWS API Documentation

Request Syntax

client.start_job_run(
    name='string',
    virtualClusterId='string',
    clientToken='string',
    executionRoleArn='string',
    releaseLabel='string',
    jobDriver={
        'sparkSubmitJobDriver': {
            'entryPoint': 'string',
            'entryPointArguments': [
                'string',
            ],
            'sparkSubmitParameters': 'string'
        },
        'sparkSqlJobDriver': {
            'entryPoint': 'string',
            'sparkSqlParameters': 'string'
        }
    },
    configurationOverrides={
        'applicationConfiguration': [
            {
                'classification': 'string',
                'properties': {
                    'string': 'string'
                },
                'configurations': {'... recursive ...'}
            },
        ],
        'monitoringConfiguration': {
            'persistentAppUI': 'ENABLED'|'DISABLED',
            'cloudWatchMonitoringConfiguration': {
                'logGroupName': 'string',
                'logStreamNamePrefix': 'string'
            },
            's3MonitoringConfiguration': {
                'logUri': 'string'
            }
        }
    },
    tags={
        'string': 'string'
    },
    jobTemplateId='string',
    jobTemplateParameters={
        'string': 'string'
    },
    retryPolicyConfiguration={
        'maxAttempts': 123
    }
)
type name:

string

param name:

The name of the job run.

type virtualClusterId:

string

param virtualClusterId:

[REQUIRED]

The virtual cluster ID for which the job run request is submitted.

type clientToken:

string

param clientToken:

[REQUIRED]

The client idempotency token of the job run request.

This field is autopopulated if not provided.

type executionRoleArn:

string

param executionRoleArn:

The execution role ARN for the job run.

type releaseLabel:

string

param releaseLabel:

The Amazon EMR release version to use for the job run.

type jobDriver:

dict

param jobDriver:

The job driver for the job run.

  • sparkSubmitJobDriver (dict) --

    The job driver parameters specified for spark submit.

    • entryPoint (string) -- [REQUIRED]

      The entry point of job application.

    • entryPointArguments (list) --

      The arguments for job application.

      • (string) --

    • sparkSubmitParameters (string) --

      The Spark submit parameters that are used for job runs.

  • sparkSqlJobDriver (dict) --

    The job driver for job type.

    • entryPoint (string) --

      The SQL file to be executed.

    • sparkSqlParameters (string) --

      The Spark parameters to be included in the Spark SQL command.

type configurationOverrides:

dict

param configurationOverrides:

The configuration overrides for the job run.

  • applicationConfiguration (list) --

    The configurations for the application running by the job run.

    • (dict) --

      A configuration specification to be used when provisioning virtual clusters, which can include configurations for applications and software bundled with Amazon EMR on EKS. A configuration consists of a classification, properties, and optional nested configurations. A classification refers to an application-specific configuration file. Properties are the settings you want to change in that file.

      • classification (string) -- [REQUIRED]

        The classification within a configuration.

      • properties (dict) --

        A set of properties specified within a configuration classification.

        • (string) --

          • (string) --

      • configurations (list) --

        A list of additional configurations to apply within a configuration object.

  • monitoringConfiguration (dict) --

    The configurations for monitoring.

    • persistentAppUI (string) --

      Monitoring configurations for the persistent application UI.

    • cloudWatchMonitoringConfiguration (dict) --

      Monitoring configurations for CloudWatch.

      • logGroupName (string) -- [REQUIRED]

        The name of the log group for log publishing.

      • logStreamNamePrefix (string) --

        The specified name prefix for log streams.

    • s3MonitoringConfiguration (dict) --

      Amazon S3 configuration for monitoring log publishing.

      • logUri (string) -- [REQUIRED]

        Amazon S3 destination URI for log publishing.

type tags:

dict

param tags:

The tags assigned to job runs.

  • (string) --

    • (string) --

type jobTemplateId:

string

param jobTemplateId:

The job template ID to be used to start the job run.

type jobTemplateParameters:

dict

param jobTemplateParameters:

The values of job template parameters to start a job run.

  • (string) --

    • (string) --

type retryPolicyConfiguration:

dict

param retryPolicyConfiguration:

The retry policy configuration for the job run.

  • maxAttempts (integer) -- [REQUIRED]

    The maximum number of attempts on the job's driver.

rtype:

dict

returns:

Response Syntax

{
    'id': 'string',
    'name': 'string',
    'arn': 'string',
    'virtualClusterId': 'string'
}

Response Structure

  • (dict) --

    • id (string) --

      This output displays the started job run ID.

    • name (string) --

      This output displays the name of the started job run.

    • arn (string) --

      This output lists the ARN of job run.

    • virtualClusterId (string) --

      This output displays the virtual cluster ID for which the job run was submitted.