AwsGlueJobHook

Amazon

Interact with AWS Glue - create job, trigger, crawler

View Source

Last Updated: Apr. 1, 2021

Access Instructions

Install the Amazon provider package into your Airflow environment.

Import the module into your DAG file and instantiate it with your desired params.

Parameters

s3_bucketOptional[str]S3 bucket where logs and local etl script will be uploaded
job_nameOptional[str]unique job name per AWS account
descOptional[str]job description
concurrent_run_limitintThe maximum number of concurrent runs allowed for a job
script_locationOptional[str]path to etl script on s3
retry_limitintMaximum number of times to retry this job if it fails
num_of_dpusintNumber of AWS Glue DPUs to allocate to this Job
region_nameOptional[str]aws region name (example: us-east-1)
iam_role_nameOptional[str]AWS IAM Role for Glue Job

Documentation

Interact with AWS Glue - create job, trigger, crawler

Example DAGs

Improve this module by creating an example DAG.

View Source
  1. Add an `example_dags` directory to the top-level source of the provider package with an empty `__init__.py` file.
  2. Add your DAG to this directory. Be sure to include a well-written and descriptive docstring
  3. Create a pull request against the source code. Once the package gets released, your DAG will show up on the Registry.

Was this page helpful?