Submits a job to a cluster.

View Source

Last Updated: May. 7, 2021

Access Instructions

Install the Google provider package into your Airflow environment.

Import the module into your DAG file and instantiate it with your desired params.


project_idstrRequired. The ID of the Google Cloud project that the job belongs to.
locationstrRequired. The Cloud Dataproc region in which to handle the request.
jobDictRequired. The job resource. If a dict is provided, it must be of the same form as the protobuf message :class:``No role entry for "class" in module "docutils.parsers.rst.languages.en". Trying "class" as canonical role name.Unknown interpreted text role "class".
request_idstrOptional. A unique id used to identify the request. If the server receives two SubmitJobRequest requests with the same id, then the second request will be ignored and the first Job created and stored in the backend is returned. It is recommended to always set this value to a UUID.
retrygoogle.api_core.retry.RetryA retry object used to retry requests. If None is specified, requests will not be retried.
timeoutfloatThe amount of time, in seconds, to wait for the request to complete. Note that if retry is specified, the timeout applies to each individual attempt.
metadataSequence[Tuple[str, str]]Additional metadata that is provided to the method.
impersonation_chainUnion[str, Sequence[str]]Optional service account to impersonate using short-term credentials, or chained list of accounts required to get the access_token of the last account in the list, which will be impersonated in the request. If set as a string, the account must grant the originating account the Service Account Token Creator IAM role. If set as a sequence, the identities from the list must grant Service Account Token Creator IAM role to the directly preceding identity, with first account from the list granting this role to the originating account (templated).
asynchronousboolFlag to return after submitting the job to the Dataproc API. This is useful for submitting long running jobs and waiting on them asynchronously using the DataprocJobSensor
cancel_on_killboolFlag which indicates whether cancel the hook's job or not, when on_kill is called
wait_timeoutintHow many seconds wait for job to be ready. Used only if asynchronous is False


Submits a job to a cluster.

Was this page helpful?