S3 Parquet Export
To write a Parquet file to S3, the httpfs extension is required. This can be installed using the INSTALL SQL command. This only needs to be run once.
INSTALL httpfs;
To load the httpfs extension for usage, use the LOAD SQL command:
LOAD httpfs;
After loading the httpfs extension, set up the credentials to write data. Note that the region parameter should match the region of the bucket you want to access.
CREATE SECRET (
TYPE S3,
KEY_ID 'AKIAIOSFODNN7EXAMPLE',
SECRET 'wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY',
REGION 'us-east-1'
); Tip If you get an IO Error (
Connection error for HTTP HEAD), configure the endpoint explicitly viaENDPOINT 's3.⟨your-region⟩.amazonaws.com'.
Alternatively, use the aws extension to retrieve the credentials automatically:
CREATE SECRET (
TYPE S3,
PROVIDER CREDENTIAL_CHAIN
); After the httpfs extension is set up and the S3 credentials are correctly configured, Parquet files can be written to S3 using the following command:
COPY ⟨table_name⟩ TO 's3://bucket/file.parquet';
Similarly, Google Cloud Storage (GCS) is supported through the Interoperability API. You need to create HMAC keys and provide the credentials as follows:
CREATE SECRET (
TYPE GCS,
KEY_ID 'AKIAIOSFODNN7EXAMPLE',
SECRET 'wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY'
); After setting up the GCS credentials, you can export using:
COPY ⟨table_name⟩ TO 'gs://gcs_bucket/file.parquet';
© Copyright 2018–2024 Stichting DuckDB Foundation
Licensed under the MIT License.
https://duckdb.org/docs/guides/network_cloud_storage/s3_export.html