Runs Hive job in Data Proc cluster.

View on GitHub

Last Updated: Oct. 23, 2022

Access Instructions

Install the Yandex provider package into your Airflow environment.

Import the module into your DAG file and instantiate it with your desired params.


queryHive query.
query_file_uriURI of the script that contains Hive queries. Can be placed in HDFS or S3.
propertiesA mapping of property names to values, used to configure Hive.
script_variablesMapping of query variable names to values.
continue_on_failureWhether to continue executing queries if a query fails.
nameName of the job. Used for labeling.
cluster_idID of the cluster to run job in. Will try to take the ID from Dataproc Hook object if it’s specified. (templated)
connection_idID of the Yandex.Cloud Airflow connection.


Runs Hive job in Data Proc cluster.

Was this page helpful?