DataprocListBatchesOperator

Google

Lists batch workloads.

View on GitHub

Last Updated: Feb. 25, 2023

Access Instructions

Install the Google provider package into your Airflow environment.

Import the module into your DAG file and instantiate it with your desired params.

Parameters

regionRequiredRequired. The Cloud Dataproc region in which to handle the request.
project_idOptional. The ID of the Google Cloud project that the cluster belongs to.
page_sizeOptional. The maximum number of batches to return in each response. The service may return fewer than this value. The default page size is 20; the maximum page size is 1000.
page_tokenOptional. A page token received from a previous ListBatches call. Provide this token to retrieve the subsequent page.
retryOptional, a retry object used to retry requests. If None is specified, requests will not be retried.
timeoutOptional, the amount of time, in seconds, to wait for the request to complete. Note that if retry is specified, the timeout applies to each individual attempt.
metadataOptional, additional metadata that is provided to the method.
gcp_conn_idOptional, the connection ID used to connect to Google Cloud Platform.
impersonation_chainOptional service account to impersonate using short-term credentials, or chained list of accounts required to get the access_token of the last account in the list, which will be impersonated in the request. If set as a string, the account must grant the originating account the Service Account Token Creator IAM role. If set as a sequence, the identities from the list must grant Service Account Token Creator IAM role to the directly preceding identity, with first account from the list granting this role to the originating account (templated).

Documentation

Lists batch workloads.

Was this page helpful?