SalesforceToGcsOperator

Google

Submits Salesforce query and uploads results to Google Cloud Storage

View on GitHub

Last Updated: Oct. 23, 2022

Access Instructions

Install the Google provider package into your Airflow environment.

Import the module into your DAG file and instantiate it with your desired params.

Parameters

queryRequiredThe query to make to Salesforce.
bucket_nameRequiredThe bucket to upload to.
object_nameRequiredThe object name to set when uploading the file.
salesforce_conn_idRequiredthe name of the connection that has the parameters we need to connect to Salesforce.
include_deletedTrue if the query should include deleted records.
query_paramsAdditional optional arguments
export_formatDesired format of files to be exported.
coerce_to_timestampTrue if you want all datetime fields to be converted into Unix timestamps. False if you want them to be left in the same format as they were in Salesforce. Leaving the value as False will result in datetimes being strings. Default: False
record_time_addedTrue if you want to add a Unix timestamp field to the resulting data that marks when the data was fetched from Salesforce. Default: False
gzipOption to compress local file or file data for upload
gcp_conn_idthe name of the connection that has the parameters we need to connect to GCS.

Documentation

Submits Salesforce query and uploads results to Google Cloud Storage

See also

For more information on how to use this operator, take a look at the guide: SalesforceToGcsOperator

Was this page helpful?