PowerDataHub / airflow_redshift_plugin

Move data from Amazon Redshift to other sources like Amazon S3, Apache Druid and more

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Maintained by Powerdatahub.com Tested with Apache Airflow 1.10.3

Airflow Plugin - Redshift

Move data from Amazon Redshift to other sources like Amazon S3, Apache Druid and more

Operators

RedshiftToDruidOperator

Executes an UNLOAD command to s3 and load into Apache Druid

:param s3_bucket: reference to a specific S3 bucket
:type s3_bucket: str
:param s3_key: reference to a specific S3 key
:type s3_key: str
:param schema: reference to a specific schema in redshift database
:type schema: str
:param table: reference to a specific table in redshift database
:type table: str
:param sql: Custom SQL statement to execute
:type sql: str
:param druid_ingest_spec: druid ingestion json spec
:type druid_ingest_spec: json
:param unload_options: reference to a list of UNLOAD options
:type unload_options: list
:param include_header: Should include headers in the final file?
:type include_header: bool
:param autocommit: if True perform autocommit
:type autocommit: bool
:param aws_conn_id: reference to a specific S3 connection
:type aws_conn_id: str
:param redshift_conn_id: reference to a specific redshift database
:type redshift_conn_id: str
:param druid_conn_id: reference to a specific Druid overlord connection
:type druid_conn_id: str

RedshiftToS3CustomOperador

Executes an UNLOAD command to s3 as a CSV with headers

:param schema: reference to a specific schema in redshift database
:type schema: str
:param table: reference to a specific table in redshift database
:type table: str
:param s3_bucket: reference to a specific S3 bucket
:type s3_bucket: str
:param s3_key: reference to a specific S3 key
:type s3_key: str
:param sql: Custom SQL statement to execute
:type sql: str
:param redshift_conn_id: reference to a specific redshift database
:type redshift_conn_id: str
:param aws_conn_id: reference to a specific S3 connection
:type aws_conn_id: str
:param verify: Whether or not to verify SSL certificates for S3 connection.
    By default SSL certificates are verified.
    You can provide the following values:

    - ``False``: do not validate SSL certificates. SSL will still be used
             (unless use_ssl is False), but SSL certificates will not be
             verified.
    - ``path/to/cert/bundle.pem``: A filename of the CA cert bundle to uses.
             You can specify this argument if you want to use a different
             CA cert bundle than the one used by botocore.
:type verify: bool or str
:param unload_options: reference to a list of UNLOAD options
:type unload_options: list

S3ToRedshiftOperator

Executes an COPY command to load a S3 file into Redshift


forthebadge forthebadge forthebadge

About

Move data from Amazon Redshift to other sources like Amazon S3, Apache Druid and more

License:Apache License 2.0


Languages

Language:Python 100.0%