You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| 1.1.2 | 2024-04-04 |[#36846](https://github.com/airbytehq/airbyte/pull/36846)| (incompatible with CDK, do not use) Remove duplicate S3 Region |
348
349
| 1.1.1 | 2024-01-03 |[#33924](https://github.com/airbytehq/airbyte/pull/33924)| (incompatible with CDK, do not use) Add new ap-southeast-3 AWS region |
349
350
| 1.1.0 | 2023-06-02 |[\#26942](https://github.com/airbytehq/airbyte/pull/26942)| Support schema evolution |
350
351
| 1.0.2 | 2023-04-20 |[\#25366](https://github.com/airbytehq/airbyte/pull/25366)| Fix default catalog to be `hive_metastore`|
Requires an S3 bucket and credentials. Data is copied into S3 as multiple files with a manifest file.
32
+
Requires an S3 bucket and credentials. Data is copied into S3 as multiple files with a manifest
33
+
file.
33
34
34
35
Airbyte automatically picks an approach depending on the given configuration - if S3 configuration
35
36
is present, Airbyte will use the COPY strategy and vice versa.
@@ -69,11 +70,14 @@ Optional parameters:
69
70
(`ab_id`, `data`, `emitted_at`). Normally these files are deleted after the `COPY` command
70
71
completes; if you want to keep them for other purposes, set `purge_staging_data` to `false`.
71
72
-**File Buffer Count**
72
-
- Number of file buffers allocated for writing data. Increasing this number is beneficial for connections using Change Data Capture (CDC) and up to the number of streams within a connection. Increasing the number of file buffers past the maximum number of streams has deteriorating effects.
73
+
- Number of file buffers allocated for writing data. Increasing this number is beneficial for
74
+
connections using Change Data Capture (CDC) and up to the number of streams within a connection.
75
+
Increasing the number of file buffers past the maximum number of streams has deteriorating
76
+
effects.
73
77
74
-
NOTE: S3 staging does not use the SSH Tunnel option for copying data, if configured. SSH Tunnel supports the SQL
75
-
connection only. S3 is secured through public HTTPS access only. Subsequent typing and deduping queries on final table
76
-
are executed over using provided SSH Tunnel configuration.
78
+
NOTE: S3 staging does not use the SSH Tunnel option for copying data, if configured. SSH Tunnel
79
+
supports the SQL connection only. S3 is secured through public HTTPS access only. Subsequent typing
80
+
and deduping queries on final table are executed over using provided SSH Tunnel configuration.
77
81
78
82
## Step 1: Set up Redshift
79
83
@@ -92,14 +96,16 @@ are executed over using provided SSH Tunnel configuration.
92
96
staging S3 bucket \(for the COPY strategy\).
93
97
94
98
### Permissions in Redshift
95
-
Airbyte writes data into two schemas, whichever schema you want your data to land in, e.g. `my_schema`
96
-
and a "Raw Data" schema that Airbyte uses to improve ELT reliability. By default, this raw data schema
97
-
is `airbyte_internal` but this can be overridden in the Redshift Destination's advanced settings.
98
-
Airbyte also needs to query Redshift's
99
+
100
+
Airbyte writes data into two schemas, whichever schema you want your data to land in, e.g.
101
+
`my_schema` and a "Raw Data" schema that Airbyte uses to improve ELT reliability. By default, this
102
+
raw data schema is `airbyte_internal` but this can be overridden in the Redshift Destination's
103
+
advanced settings. Airbyte also needs to query Redshift's
99
104
[SVV_TABLE_INFO](https://docs.aws.amazon.com/redshift/latest/dg/r_SVV_TABLE_INFO.html) table for
100
105
metadata about the tables airbyte manages.
101
106
102
107
To ensure the `airbyte_user` has the correction permissions to:
108
+
103
109
- create schemas in your database
104
110
- grant usage to any existing schemas you want Airbyte to use
105
111
- grant select to the `svv_table_info` table
@@ -187,14 +193,19 @@ characters.
187
193
### Data Size Limitations
188
194
189
195
Redshift specifies a maximum limit of 16MB (and 65535 bytes for any VARCHAR fields within the JSON
190
-
record) to store the raw JSON record data. Thus, when a row is too big to fit, the destination connector will
191
-
do one of the following.
192
-
1. Null the value if the varchar size > 65535, The corresponding key information is added to `_airbyte_meta`.
193
-
2. Null the whole record while trying to preserve the Primary Keys and cursor field declared as part of your stream configuration, if the total record size is > 16MB.
194
-
* For DEDUPE sync mode, if we do not find Primary key(s), we fail the sync.
195
-
* For OVERWRITE and APPEND mode, syncs will succeed with empty records emitted, if we fail to find Primary key(s).
196
+
record) to store the raw JSON record data. Thus, when a row is too big to fit, the destination
197
+
connector will do one of the following.
198
+
199
+
1. Null the value if the varchar size > 65535, The corresponding key information is added to
200
+
`_airbyte_meta`.
201
+
2. Null the whole record while trying to preserve the Primary Keys and cursor field declared as part
202
+
of your stream configuration, if the total record size is > 16MB.
203
+
- For DEDUPE sync mode, if we do not find Primary key(s), we fail the sync.
204
+
- For OVERWRITE and APPEND mode, syncs will succeed with empty records emitted, if we fail to
205
+
find Primary key(s).
196
206
197
-
See AWS docs for [SUPER](https://docs.aws.amazon.com/redshift/latest/dg/r_SUPER_type.html) and [SUPER limitations](https://docs.aws.amazon.com/redshift/latest/dg/limitations-super.html).
207
+
See AWS docs for [SUPER](https://docs.aws.amazon.com/redshift/latest/dg/r_SUPER_type.html) and
| 0.3.53 | 2023-01-03 |[\#17273](https://github.com/airbytehq/airbyte/pull/17273)| Flatten JSON arrays to fix maximum size check for SUPER field |
300
-
| 0.3.52 | 2022-12-30 |[\#20879](https://github.com/airbytehq/airbyte/pull/20879)| Added configurable parameter for number of file buffers (⛔ this version has a bug and will not work; use `0.3.56` instead) |
312
+
| 0.3.52 | 2022-12-30 |[\#20879](https://github.com/airbytehq/airbyte/pull/20879)| Added configurable parameter for number of file buffers (⛔ this version has a bug and will not work; use `0.3.56` instead) |
| 0.3.50 | 2022-09-14 |[\#15668](https://github.com/airbytehq/airbyte/pull/15668)| Wrap logs in AirbyteLogMessage |
303
315
| 0.3.49 | 2022-09-01 |[\#16243](https://github.com/airbytehq/airbyte/pull/16243)| Fix Json to Avro conversion when there is field name clash from combined restrictions (`anyOf`, `oneOf`, `allOf` fields) |
0 commit comments