airflow.providers.microsoft.azure.transfers.oracle_to_azure_data_lake¶
Classes¶
Runs the query against Oracle and stores the file locally before loading it into Azure Data Lake. |
Module Contents¶
- class airflow.providers.microsoft.azure.transfers.oracle_to_azure_data_lake.OracleToAzureDataLakeOperator(*, filename, azure_data_lake_conn_id, azure_data_lake_path, oracle_conn_id, sql, sql_params=None, delimiter=',', encoding='utf-8', quotechar='"', quoting=csv.QUOTE_MINIMAL, **kwargs)[source]¶
Bases:
airflow.providers.common.compat.sdk.BaseOperatorRuns the query against Oracle and stores the file locally before loading it into Azure Data Lake.
- Parameters:
filename (str) – file name to be used by the csv file.
azure_data_lake_conn_id (str) – destination azure data lake connection.
azure_data_lake_path (str) – destination path in azure data lake to put the file.
oracle_conn_id (str) – Source Oracle connection.
sql (str) – SQL query to execute against the Oracle database. (templated)
sql_params (dict | None) – Parameters to use in sql query. (templated)
delimiter (str) – field delimiter in the file.
encoding (str) – encoding type for the file.
quotechar (str) – Character to use in quoting.
quoting (Literal[0, 1, 2, 3]) – Quoting strategy. See csv library for more information. It can take on any of the csv.QUOTE_* constants.
- template_fields: collections.abc.Sequence[str] = ('filename', 'sql', 'sql_params')[source]¶