- Original Author(s): @natalie-white-aws, @mjanardhan @parag-shah-aws
- Tracking Issue:
- API Bar Raiser: @TheRealAmazonKendra Link to RFC Issue
AWS Glue is a serverless data integration service that makes it easier to discover, prepare, move, and integrate data from multiple sources for analytics, machine learning (ML), and application development.
Wihout an L2 construct, developers define Glue data sources, connections, jobs, and workflows for their data and ETL solutions via the AWS console, the AWS CLI, and Infrastructure as Code tools like CloudFormation and the CDK. However, there are several challenges to defining Glue resources at scale that an L2 construct can resolve. First, developers must reference documentation to determine the valid combinations of job type, Glue version, worker type, language versions, and other parameters that are required for specific job types. Additionally, developers must already know or look up the networking constraints for data source connections, and there is ambiguity around how to securely store secrets for JDBC connections. Finally, developers want prescriptive guidance via best practice defaults for throughput parameters like number of workers and batching.
The Glue L2 construct has convenience methods working backwards from common use cases and sets required parameters to defaults that align with recommended best practices for each job type. It also provides customers with a balance between flexibility via optional parameter overrides, and opinionated interfaces that discouraging anti-patterns, resulting in reduced time to develop and deploy new resources.
- Glue Launch Announcement
- Glue Documentation
- Glue L1 (CloudFormation) Constructs
- Prior version of the @aws-cdk/aws-glue-alpha module
A Job encapsulates a script that connects to data sources, processes them, and then writes output to a data target. There are four types of Glue Jobs: Spark (ETL and Streaming), Python Shell, Ray, and Flex Jobs. Most of the required parameters for these jobs are common across all types, but there are a few differences depending on the languages supported and features provided by each type. For all job types, the L2 defaults to AWS best practice recommendations, such as:
- Use of Secrets Manager for Connection JDBC strings
- Glue job autoscaling
- Default parameter values for Glue job creation
This iteration of the L2 construct introduces breaking changes to the existing glue-alpha-module, but these changes streamline the developer experience, introduce new constants for defaults, and replacing synth-time validations with interface contracts for enforcement of the parameter combinations that Glue supports. As an opinionated construct, the Glue L2 construct does not allow developers to create resources that use non-current versions of Glue or deprecated language dependencies (e.g. deprecated versions of Python). As always, L1s allow you to specify a wider range of parameters if you need or want to use alternative configurations.
Optional and required parameters for each job are enforced via interface rather than validation; see Glue's public documentation for more granular details.
- ETL Jobs
ETL jobs support pySpark and Scala languages, for which there are separate but
similar constructors. ETL jobs default to the G2 worker type, but you can
override this default with other supported worker type values (G1, G2, G4
and G8). ETL jobs defaults to Glue version 4.0, which you can override to 3.0.
The following ETL features are enabled by default:
—enable-metrics, —enable-spark-ui, —enable-continuous-cloudwatch-log.
You can find more details about version, worker type and other features in
Glue's public documentation.
glue.ScalaSparkEtlJob(this, 'ScalaSparkEtlJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-jar'),
className: 'com.example.HelloWorld',
role: iam.IRole,
});
glue.pySparkEtlJob(this, 'pySparkEtlJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
role: iam.IRole,
});
Optional override examples:
glue.ScalaSparkEtlJob(this, 'ScalaSparkEtlJob', {
glueVersion: glue.GlueVersion.V3_0,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-script'),
className: 'com.example.HelloWorld',
extraJars: [glue.Code.fromBucket('bucket-name', 'path-to-extra-jars'),],
description: 'an example Scala Spark ETL job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
glue.pySparkEtlJob(this, 'pySparkEtlJob', {
jobType: glue.JobType.ETL,
glueVersion: glue.GlueVersion.V3_0,
pythonVersion: glue.PythonVersion.3_9,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
description: 'an example pySpark ETL job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
Scala Spark ETL Job Property Interface:
ScalaSparkEtlJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue Job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* Class name (required for Scala)
* Package and class name for the entry point of Glue job execution for
* Java scripts
* */
className: string;
/**
* Extra Jars S3 URL (optional)
* S3 URL where additional jar dependencies are located
*/
extraJars?: string[];
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue performs
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0 for ETL
* */
glueVersion?: glue.GlueVersion;
}
pySpark ETL Job Property Interface:
pySparkEtlJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) takes during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Extra Jars S3 URL (optional)
* S3 URL where additional jar dependencies are located
*/
extraJars?: string[];
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue performs
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0 for ETL
* */
glueVersion?: glue.GlueVersion;
}
- Streaming Jobs
Streaming jobs are similar to ETL jobs, except that they perform ETL on data
streams using the Apache Spark Structured Streaming framework. Some Spark
job features are not available to Streaming ETL jobs. They support Scala
and pySpark languages. PySpark streaming jobs default Python 3.9,
which you can override with any non-deprecated version of Python. It
defaults to the G2 worker type and Glue 4.0, both of which you can override.
The following best practice features are enabled by default:
—enable-metrics, —enable-spark-ui, —enable-continuous-cloudwatch-log
.
new glue.pySparkStreamingJob(this, 'pySparkStreamingJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
role: iam.IRole,
});
new glue.ScalaSparkStreamingJob(this, 'ScalaSparkStreamingJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-jar'),
className: 'com.example.HelloWorld',
role: iam.IRole,
});
Optional override examples:
new glue.pySparkStreamingJob(this, 'pySparkStreamingJob', {
glueVersion: glue.GlueVersion.V3_0,
pythonVersion: glue.PythonVersion.3_9,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
description: 'an example Python Streaming job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
new glue.ScalaSparkStreamingJob(this, 'ScalaSparkStreamingJob', {
glueVersion: glue.GlueVersion.V3_0,
pythonVersion: glue.PythonVersion.3_9,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-script'),
extraJars: [glue.Code.fromBucket('bucket-name', 'path-to-extra-jars'),],
className: 'com.example.HelloWorld',
description: 'an example Python Streaming job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
Scala Spark Streaming Job Property Interface:
ScalaSparkStreamingJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* Class name (required for Scala scripts)
* Package and class name for the entry point of Glue job execution for
* Java scripts
* */
className: string;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Extra Jars S3 URL (optional)
* S3 URL where additional jar dependencies are located
*/
extraJars?: string[];
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue performs
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0
* */
glueVersion?: glue.GlueVersion;
}
pySpark Streaming Job Property Interface:
pySparkStreamingJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue perform
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0
* */
glueVersion?: glue.GlueVersion;
}
- Flex Jobs
The flexible execution class is appropriate for non-urgent jobs such as
pre-production jobs, testing, and one-time data loads. Flexible jobs default
to Glue version 3.0 and worker type G_2X
. The following best practice
features are enabled by default:
—enable-metrics, —enable-spark-ui, —enable-continuous-cloudwatch-log
glue.ScalaSparkFlexEtlJob(this, 'ScalaSparkFlexEtlJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-jar'),
className: 'com.example.HelloWorld',
role: iam.IRole,
});
glue.pySparkFlexEtlJob(this, 'pySparkFlexEtlJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
role: iam.IRole,
});
Optional override examples:
glue.ScalaSparkFlexEtlJob(this, 'ScalaSparkFlexEtlJob', {
glueVersion: glue.GlueVersion.V3_0,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-scala-script'),
className: 'com.example.HelloWorld',
extraJars: [glue.Code.fromBucket('bucket-name', 'path-to-extra-jars')],
description: 'an example pySpark ETL job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
new glue.pySparkFlexEtlJob(this, 'pySparkFlexEtlJob', {
glueVersion: glue.GlueVersion.V3_0,
pythonVersion: glue.PythonVersion.3_9,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
description: 'an example Flex job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
Scala Spark Flex Job Property Interface:
ScalaSparkFlexJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* Class name (required for Scala scripts)
* Package and class name for the entry point of Glue job execution for
* Java scripts
* */
className: string;
/**
* Extra Jars S3 URL (optional)
* S3 URL where additional jar dependencies are located
*/
extraJars?: string[];
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue perform
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0
* */
glueVersion?: glue.GlueVersion;
}
pySpark Flex Job Property Interface:
PySparkFlexJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue perform
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0
* */
glueVersion?: glue.GlueVersion;
}
Python shell jobs support a Python version that depends on the AWS Glue
version you use. These can be used to schedule and run tasks that don't
require an Apache Spark environment. Python shell jobs default to
Python 3.9 and a MaxCapacity of 0.0625
. Python 3.9 supports pre-loaded
analytics libraries using the library-set=analytics
flag, which is
enabled by default.
new glue.PythonShellJob(this, 'PythonShellJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
role: iam.IRole,
});
Optional override examples:
new glue.PythonShellJob(this, 'PythonShellJob', {
glueVersion: glue.GlueVersion.V1_0,
pythonVersion: glue.PythonVersion.3_9,
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
description: 'an example Python Shell job',
numberOfWorkers: 20,
workerType: glue.WorkerType.G8X,
timeout: cdk.Duration.minutes(15),
role: iam.IRole,
});
Python Shell Job Property Interface:
PythonShellJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default G_2X
* */
workerType?: glue.WorkerType;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue perform
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 3.0 for ETL
* */
glueVersion?: glue.GlueVersion;
}
Glue Ray jobs use worker type Z.2X and Glue version 4.0. These are not overrideable since these are the only configuration that Glue Ray jobs currently support. The runtime defaults to Ray2.4 and min workers defaults to 3.
new glue.GlueRayJob(this, 'GlueRayJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
role: iam.IRole,
});
Optional override example:
new glue.GlueRayJob(this, 'GlueRayJob', {
script: glue.Code.fromBucket('bucket-name', 's3prefix/path-to-python-script'),
numberOfWorkers: 50,
minWorkers: 25,
role: iam.IRole,
});
Ray Job Property Interface:
RayJobProps{
/**
* Script Code Location (required)
* Script to run when the Glue job executes. Can be uploaded
* from the local directory structure using fromAsset
* or referenced via S3 location using fromBucket
* */
script: glue.Code;
/**
* IAM Role (required)
* IAM Role to use for Glue job execution
* Must be specified by the developer because the L2 doesn't have visibility
* into the actions the script(s) take during the job execution
* */
role: iam.IRole;
/**
* Name of the Glue job (optional)
* Developer-specified name of the Glue job
* */
name?: string;
/**
* Description (optional)
* Developer-specified description of the Glue job
* */
description?: string;
/**
* Number of Workers (optional)
* Number of workers for Glue to use during job execution
* @default 10
* */
numberOrWorkers?: int;
/**
* Worker Type (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Standard, G_1X, G_2X, G_025X. G_4X, G_8X, Z_2X
* @default Z_2X
* */
workerType?: glue.WorkerType;
/**
* Runtime (optional)
* Type of Worker for Glue to use during job execution
* Enum options: Ray2_2, Ray 2_3, Ray2_4
* @default Ray2_4
* */
runtime?: glue.RayRuntime;
/**
* Max Concurrent Runs (optional)
* The maximum number of runs this Glue job can concurrently run
* @default 1
* */
maxConcurrentRuns?: int;
/**
* Default Arguments (optional)
* The default arguments for every run of this Glue job,
* specified as name-value pairs.
* */
defaultArguments?: {[key: string], string }[];
/**
* Connections (optional)
* List of connections to use for this Glue job
* */
connections?: IConnection[];
/**
* Max Retries (optional)
* Maximum number of retry attempts Glue perform
* if the job fails
* @default 0
* */
maxRetries?: int;
/**
* Timeout (optional)
* Timeout for the Glue job, specified in minutes
* @default 2880 (2 days for non-streaming)
* */
timeout?: cdk.Duration;
/**
* Security Configuration (optional)
* Defines the encryption options for the Glue job
* */
securityConfiguration?: ISecurityConfiguration;
/**
* Tags (optional)
* A list of key:value pairs of tags to apply to this Glue job resource
* */
tags?: {[key: string], string }[];
/**
* Glue Version
* The version of Glue to use to execute this job
* @default 4.0
* */
glueVersion?: glue.GlueVersion;
}
Similar to other L2 constructs, the Glue L2 automates uploading / updating scripts to S3 via an optional fromAsset parameter pointing to a script in the local file structure. You provide the existing S3 bucket and path to which you'd like the script to be uploaded.
glue.ScalaSparkEtlJob(this, 'ScalaSparkEtlJob', {
script: glue.Code.fromAsset('bucket-name', 'local/path/to/scala-jar'),
className: 'com.example.HelloWorld',
});
You can use Glue workflows to create and visualize complex extract, transform, and load (ETL) activities involving multiple crawlers, jobs, and triggers. Standalone triggers are an anti-pattern, so you must create triggers from within a workflow using the L2 construct.
Within a workflow object, there are functions to create different types of triggers with actions and predicates. You then add those triggers to jobs.
StartOnCreation defaults to true for all trigger types, but you can override it if you prefer for your trigger not to start on creation.
- On-Demand Triggers
On-demand triggers can start glue jobs or crawlers. This construct provides convenience functions to create on-demand crawler or job triggers. The constructor takes an optional description parameter, but abstracts the requirement of an actions list using the job or crawler objects using conditional types.
myWorkflow = new glue.Workflow(this, "GlueWorkflow", {
name: "MyWorkflow";
description: "New Workflow";
properties: {'key', 'value'};
});
myWorkflow.onDemandTrigger(this, 'TriggerJobOnDemand', {
description: 'On demand run for ' + glue.JobExecutable.name,
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
});
- Scheduled Triggers
You can create scheduled triggers using cron expressions. This construct provides daily, weekly, and monthly convenience functions, as well as a custom function that allows you to create your own custom timing using the existing event Schedule class without having to build your own cron expressions. The L2 extracts the expression that Glue requires from the Schedule object. The constructor takes an optional description and a list of jobs or crawlers as actions.
// Create Daily Schedule at 00 UTC
myWorkflow.dailyScheduleTrigger(this, 'TriggerCrawlerOnDailySchedule', {
description: 'Scheduled run for ' + glue.JobExecutable.name,
actions: [ jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
});
// Create Weekly schedule at 00 UTC on Sunday
myWorkflow.weeklyScheduleTrigger(this, 'TriggerJobOnWeeklySchedule', {
description: 'Scheduled run for ' + glue.JobExecutable.name,
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
});
// Create Custom schedule, e.g. Monthly on the 7th day at 15:30 UTC
myWorkflow.customScheduleJobTrigger(this, 'TriggerCrawlerOnCustomSchedule', {
description: 'Scheduled run for ' + glue.JobExecutable.name,
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
schedule: events.Schedule.cron(day: '7', hour: '15', minute: '30')
});
There are two types of notify event triggers: batching and non-batching.
For batching triggers, you must specify BatchSize
. For non-batching
triggers, BatchSize
defaults to 1. For both triggers, BatchWindow
defaults to 900 seconds, but you can override the window to align with
your workload's requirements.
myWorkflow.notifyEventTrigger(this, 'MyNotifyTriggerBatching', {
batchSize: int,
jobActions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...],
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ... ]
});
myWorkflow.notifyEventTrigger(this, 'MyNotifyTriggerNonBatching', {
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
});
Conditional triggers have a predicate and actions associated with them. The trigger actions are executed when the predicateCondition is true.
// Triggers on Job and Crawler status
myWorkflow.conditionalTrigger(this, 'conditionalTrigger', {
description: 'Conditional trigger for ' + myGlueJob.name,
actions: [jobOrCrawler: glue.JobExecutable | cdk.CfnCrawler?, ...]
predicateCondition: glue.TriggerPredicateCondition.AND,
jobPredicates: [{'job': JobExecutable, 'state': glue.JobState.FAILED},
{'job': JobExecutable, 'state' : glue.JobState.SUCCEEDED}]
});
A Connection
allows Glue jobs, crawlers and development endpoints to access
certain types of data stores.
***Secrets Management **You must specify JDBC connection credentials in Secrets Manager and provide the Secrets Manager Key name as a property to the job connection.
- **Networking - the CDK determines the best fit subnet for Glue connection configuration **The prior version of the glue-alpha-module requires the developer to specify the subnet of the Connection when it’s defined. Now, you can still specify the specific subnet you want to use, but are no longer required to. You are only required to provide a VPC and either a public or private subnet selection. Without a specific subnet provided, the L2 leverages the existing EC2 Subnet Selection library to make the best choice selection for the subnet.
We’re launching new features to an AWS CDK Glue L2 Construct to provide best-practice defaults and convenience methods to create Glue Jobs, Connections, Triggers, Workflows, and the underlying permissions and configuration.
Developers should use this Construct to reduce the amount of boilerplate code and complexity each individual has to navigate, and make it easier to create best-practice Glue resources.
Glue Crawlers and other resources that are now managed by the AWS LakeFormation team are not in scope for this effort. Developers should use existing methods to create these resources, and the new Glue L2 construct assumes they already exist as inputs. While best practice is for application and infrastructure code to be as close as possible for teams using fully-implemented DevOps mechanisms, in practice these ETL scripts are likely managed by a data science team who know Python or Scala and don’t necessarily own or manage their own infrastructure deployments. We want to meet developers where they are, and not assume that all of the code resides in the same repository, Developers can automate this themselves via the CDK, however, if they do own both.
Validating Glue version and feature use per AWS region at synth time is also not in scope. AWS’ intention is for all features to eventually be propagated to all Global regions, so the complexity involved in creating and updating region- specific configuration to match shifting feature sets does not out-weigh the likelihood that a developer will use this construct to deploy resources to a region without a particular new feature to a region that doesn’t yet support it without researching or manually attempting to use that feature before developing it via IaC. The developer will, of course, still get feedback from the underlying Glue APIs as CloudFormation deploys the resources similar to the current CDK L1 Glue experience.