airflow.providers.amazon.aws.transfers.hive_to_dynamodb
¶
This module contains operator to move data from Hive to DynamoDB.
Module Contents¶
Classes¶
Moves data from Hive to DynamoDB. |
- class airflow.providers.amazon.aws.transfers.hive_to_dynamodb.HiveToDynamoDBOperator(*, sql, table_name, table_keys, pre_process=None, pre_process_args=None, pre_process_kwargs=None, region_name=None, schema='default', hiveserver2_conn_id='hiveserver2_default', aws_conn_id='aws_default', **kwargs)[source]¶
Bases:
airflow.models.BaseOperator
Moves data from Hive to DynamoDB.
Note that for now the data is loaded into memory before being pushed to DynamoDB, so this operator should be used for smallish amount of data.
See also
For more information on how to use this operator, take a look at the guide: Apache Hive to Amazon DynamoDB transfer operator
- Parameters
sql (str) – SQL query to execute against the hive database. (templated)
table_name (str) – target DynamoDB table
table_keys (list) – partition key and sort key
pre_process (Callable | None) – implement pre-processing of source data
pre_process_args (list | None) – list of pre_process function arguments
pre_process_kwargs (list | None) – dict of pre_process function arguments
region_name (str | None) – aws region name (example: us-east-1)
schema (str) – hive database schema
hiveserver2_conn_id (str) – Reference to the :ref: Hive Server2 thrift service connection id <howto/connection:hiveserver2>.
aws_conn_id (str | None) – aws connection
- template_fields: collections.abc.Sequence[str] = ('sql',)[source]¶
- template_ext: collections.abc.Sequence[str] = ('.sql',)[source]¶