Skip to content

Commit 4342182

Browse files
authored
Remove duplicate S3 regions from destination specs (#36846)
1 parent 9d53013 commit 4342182

File tree

9 files changed

+38
-27
lines changed

9 files changed

+38
-27
lines changed

airbyte-integrations/connectors/destination-databricks/metadata.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@ data:
22
connectorSubtype: database
33
connectorType: destination
44
definitionId: 072d5540-f236-4294-ba7c-ade8fd918496
5-
dockerImageTag: 1.1.1
5+
dockerImageTag: 1.1.2
66
dockerRepository: airbyte/destination-databricks
77
githubIssueLabel: destination-databricks
88
icon: databricks.svg

airbyte-integrations/connectors/destination-databricks/src/main/resources/spec.json

-1
Original file line numberDiff line numberDiff line change
@@ -156,7 +156,6 @@
156156
"me-central-1",
157157
"me-south-1",
158158
"sa-east-1",
159-
"sa-east-1",
160159
"us-east-1",
161160
"us-east-2",
162161
"us-gov-east-1",

airbyte-integrations/connectors/destination-iceberg/metadata.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@ data:
22
connectorSubtype: database
33
connectorType: destination
44
definitionId: df65a8f3-9908-451b-aa9b-445462803560
5-
dockerImageTag: 0.1.5
5+
dockerImageTag: 0.1.6
66
dockerRepository: airbyte/destination-iceberg
77
githubIssueLabel: destination-iceberg
88
license: MIT

airbyte-integrations/connectors/destination-iceberg/src/main/resources/spec.json

-1
Original file line numberDiff line numberDiff line change
@@ -241,7 +241,6 @@
241241
"me-central-1",
242242
"me-south-1",
243243
"sa-east-1",
244-
"sa-east-1",
245244
"us-east-1",
246245
"us-east-2",
247246
"us-gov-east-1",

airbyte-integrations/connectors/destination-redshift/metadata.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ data:
55
connectorSubtype: database
66
connectorType: destination
77
definitionId: f7a7d195-377f-cf5b-70a5-be6b819019dc
8-
dockerImageTag: 2.4.0
8+
dockerImageTag: 2.4.1
99
dockerRepository: airbyte/destination-redshift
1010
documentationUrl: https://docs.airbyte.com/integrations/destinations/redshift
1111
githubIssueLabel: destination-redshift

airbyte-integrations/connectors/destination-redshift/src/main/resources/spec.json

-1
Original file line numberDiff line numberDiff line change
@@ -138,7 +138,6 @@
138138
"me-central-1",
139139
"me-south-1",
140140
"sa-east-1",
141-
"sa-east-1",
142141
"us-east-1",
143142
"us-east-2",
144143
"us-gov-east-1",

docs/integrations/destinations/databricks.md

+1
Original file line numberDiff line numberDiff line change
@@ -345,6 +345,7 @@ Delta Lake tables are created. You may want to consult the tutorial on
345345

346346
| Version | Date | Pull Request | Subject |
347347
| :------ | :--------- | :------------------------------------------------------------------------------------------------------------------ | :----------------------------------------------------------------------------------------------------------------------- |
348+
| 1.1.2 | 2024-04-04 | [#36846](https://github.com/airbytehq/airbyte/pull/36846) | (incompatible with CDK, do not use) Remove duplicate S3 Region |
348349
| 1.1.1 | 2024-01-03 | [#33924](https://github.com/airbytehq/airbyte/pull/33924) | (incompatible with CDK, do not use) Add new ap-southeast-3 AWS region |
349350
| 1.1.0 | 2023-06-02 | [\#26942](https://github.com/airbytehq/airbyte/pull/26942) | Support schema evolution |
350351
| 1.0.2 | 2023-04-20 | [\#25366](https://github.com/airbytehq/airbyte/pull/25366) | Fix default catalog to be `hive_metastore` |

docs/integrations/destinations/iceberg.md

+1
Original file line numberDiff line numberDiff line change
@@ -61,6 +61,7 @@ specify the target size of compacted Iceberg data file.
6161

6262
| Version | Date | Pull Request | Subject |
6363
| :------ | :--------- | :-------------------------------------------------------- | :--------------------------------------------------------- |
64+
| 0.1.6 | 2024-04-04 | [#36846](https://github.com/airbytehq/airbyte/pull/36846) | Remove duplicate S3 Region |
6465
| 0.1.5 | 2024-01-03 | [#33924](https://github.com/airbytehq/airbyte/pull/33924) | Add new ap-southeast-3 AWS region |
6566
| 0.1.4 | 2023-07-20 | [28506](https://github.com/airbytehq/airbyte/pull/28506) | Support server-managed storage config |
6667
| 0.1.3 | 2023-07-12 | [28158](https://github.com/airbytehq/airbyte/pull/28158) | Bump Iceberg library to 1.3.0 and add REST catalog support |

docs/integrations/destinations/redshift.md

+33-21
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,8 @@ For INSERT strategy:
2929
2. COPY: Replicates data by first uploading data to an S3 bucket and issuing a COPY command. This is
3030
the recommended loading approach described by Redshift
3131
[best practices](https://docs.aws.amazon.com/redshift/latest/dg/c_best-practices-single-copy-command.html).
32-
Requires an S3 bucket and credentials. Data is copied into S3 as multiple files with a manifest file.
32+
Requires an S3 bucket and credentials. Data is copied into S3 as multiple files with a manifest
33+
file.
3334

3435
Airbyte automatically picks an approach depending on the given configuration - if S3 configuration
3536
is present, Airbyte will use the COPY strategy and vice versa.
@@ -69,11 +70,14 @@ Optional parameters:
6970
(`ab_id`, `data`, `emitted_at`). Normally these files are deleted after the `COPY` command
7071
completes; if you want to keep them for other purposes, set `purge_staging_data` to `false`.
7172
- **File Buffer Count**
72-
- Number of file buffers allocated for writing data. Increasing this number is beneficial for connections using Change Data Capture (CDC) and up to the number of streams within a connection. Increasing the number of file buffers past the maximum number of streams has deteriorating effects.
73+
- Number of file buffers allocated for writing data. Increasing this number is beneficial for
74+
connections using Change Data Capture (CDC) and up to the number of streams within a connection.
75+
Increasing the number of file buffers past the maximum number of streams has deteriorating
76+
effects.
7377

74-
NOTE: S3 staging does not use the SSH Tunnel option for copying data, if configured. SSH Tunnel supports the SQL
75-
connection only. S3 is secured through public HTTPS access only. Subsequent typing and deduping queries on final table
76-
are executed over using provided SSH Tunnel configuration.
78+
NOTE: S3 staging does not use the SSH Tunnel option for copying data, if configured. SSH Tunnel
79+
supports the SQL connection only. S3 is secured through public HTTPS access only. Subsequent typing
80+
and deduping queries on final table are executed over using provided SSH Tunnel configuration.
7781

7882
## Step 1: Set up Redshift
7983

@@ -92,14 +96,16 @@ are executed over using provided SSH Tunnel configuration.
9296
staging S3 bucket \(for the COPY strategy\).
9397

9498
### Permissions in Redshift
95-
Airbyte writes data into two schemas, whichever schema you want your data to land in, e.g. `my_schema`
96-
and a "Raw Data" schema that Airbyte uses to improve ELT reliability. By default, this raw data schema
97-
is `airbyte_internal` but this can be overridden in the Redshift Destination's advanced settings.
98-
Airbyte also needs to query Redshift's
99+
100+
Airbyte writes data into two schemas, whichever schema you want your data to land in, e.g.
101+
`my_schema` and a "Raw Data" schema that Airbyte uses to improve ELT reliability. By default, this
102+
raw data schema is `airbyte_internal` but this can be overridden in the Redshift Destination's
103+
advanced settings. Airbyte also needs to query Redshift's
99104
[SVV_TABLE_INFO](https://docs.aws.amazon.com/redshift/latest/dg/r_SVV_TABLE_INFO.html) table for
100105
metadata about the tables airbyte manages.
101106

102107
To ensure the `airbyte_user` has the correction permissions to:
108+
103109
- create schemas in your database
104110
- grant usage to any existing schemas you want Airbyte to use
105111
- grant select to the `svv_table_info` table
@@ -187,14 +193,19 @@ characters.
187193
### Data Size Limitations
188194

189195
Redshift specifies a maximum limit of 16MB (and 65535 bytes for any VARCHAR fields within the JSON
190-
record) to store the raw JSON record data. Thus, when a row is too big to fit, the destination connector will
191-
do one of the following.
192-
1. Null the value if the varchar size > 65535, The corresponding key information is added to `_airbyte_meta`.
193-
2. Null the whole record while trying to preserve the Primary Keys and cursor field declared as part of your stream configuration, if the total record size is > 16MB.
194-
* For DEDUPE sync mode, if we do not find Primary key(s), we fail the sync.
195-
* For OVERWRITE and APPEND mode, syncs will succeed with empty records emitted, if we fail to find Primary key(s).
196+
record) to store the raw JSON record data. Thus, when a row is too big to fit, the destination
197+
connector will do one of the following.
198+
199+
1. Null the value if the varchar size > 65535, The corresponding key information is added to
200+
`_airbyte_meta`.
201+
2. Null the whole record while trying to preserve the Primary Keys and cursor field declared as part
202+
of your stream configuration, if the total record size is > 16MB.
203+
- For DEDUPE sync mode, if we do not find Primary key(s), we fail the sync.
204+
- For OVERWRITE and APPEND mode, syncs will succeed with empty records emitted, if we fail to
205+
find Primary key(s).
196206

197-
See AWS docs for [SUPER](https://docs.aws.amazon.com/redshift/latest/dg/r_SUPER_type.html) and [SUPER limitations](https://docs.aws.amazon.com/redshift/latest/dg/limitations-super.html).
207+
See AWS docs for [SUPER](https://docs.aws.amazon.com/redshift/latest/dg/r_SUPER_type.html) and
208+
[SUPER limitations](https://docs.aws.amazon.com/redshift/latest/dg/limitations-super.html).
198209

199210
### Encryption
200211

@@ -208,15 +219,15 @@ Each stream will be output into its own raw table in Redshift. Each table will c
208219
Redshift is `VARCHAR`.
209220
- `_airbyte_extracted_at`: a timestamp representing when the event was pulled from the data source.
210221
The column type in Redshift is `TIMESTAMP WITH TIME ZONE`.
211-
- `_airbyte_loaded_at`: a timestamp representing when the row was processed into final table.
212-
The column type in Redshift is `TIMESTAMP WITH TIME ZONE`.
222+
- `_airbyte_loaded_at`: a timestamp representing when the row was processed into final table. The
223+
column type in Redshift is `TIMESTAMP WITH TIME ZONE`.
213224
- `_airbyte_data`: a json blob representing with the event data. The column type in Redshift is
214225
`SUPER`.
215226

216227
## Data type map
217228

218229
| Airbyte type | Redshift type |
219-
|:------------------------------------|:---------------------------------------|
230+
| :---------------------------------- | :------------------------------------- |
220231
| STRING | VARCHAR |
221232
| STRING (BASE64) | VARCHAR |
222233
| STRING (BIG_NUMBER) | VARCHAR |
@@ -235,7 +246,8 @@ Each stream will be output into its own raw table in Redshift. Each table will c
235246
## Changelog
236247

237248
| Version | Date | Pull Request | Subject |
238-
|:--------|:-----------|:-----------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
249+
| :------ | :--------- | :--------------------------------------------------------- | :--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
250+
| 2.4.1 | 2024-04-04 | [#36846](https://github.com/airbytehq/airbyte/pull/36846) | Remove duplicate S3 Region |
239251
| 2.4.0 | 2024-03-21 | [\#36589](https://github.com/airbytehq/airbyte/pull/36589) | Adapt to Kotlin cdk 0.28.19 |
240252
| 2.3.2 | 2024-03-21 | [\#36374](https://github.com/airbytehq/airbyte/pull/36374) | Supress Jooq DataAccessException error message in logs |
241253
| 2.3.1 | 2024-03-18 | [\#36255](https://github.com/airbytehq/airbyte/pull/36255) | Mark as Certified-GA |
@@ -297,7 +309,7 @@ Each stream will be output into its own raw table in Redshift. Each table will c
297309
| 0.3.55 | 2023-01-26 | [\#20631](https://github.com/airbytehq/airbyte/pull/20631) | Added support for destination checkpointing with staging |
298310
| 0.3.54 | 2023-01-18 | [\#21087](https://github.com/airbytehq/airbyte/pull/21087) | Wrap Authentication Errors as Config Exceptions |
299311
| 0.3.53 | 2023-01-03 | [\#17273](https://github.com/airbytehq/airbyte/pull/17273) | Flatten JSON arrays to fix maximum size check for SUPER field |
300-
| 0.3.52 | 2022-12-30 | [\#20879](https://github.com/airbytehq/airbyte/pull/20879) | Added configurable parameter for number of file buffers (⛔ this version has a bug and will not work; use `0.3.56` instead) |
312+
| 0.3.52 | 2022-12-30 | [\#20879](https://github.com/airbytehq/airbyte/pull/20879) | Added configurable parameter for number of file buffers (⛔ this version has a bug and will not work; use `0.3.56` instead) |
301313
| 0.3.51 | 2022-10-26 | [\#18434](https://github.com/airbytehq/airbyte/pull/18434) | Fix empty S3 bucket path handling |
302314
| 0.3.50 | 2022-09-14 | [\#15668](https://github.com/airbytehq/airbyte/pull/15668) | Wrap logs in AirbyteLogMessage |
303315
| 0.3.49 | 2022-09-01 | [\#16243](https://github.com/airbytehq/airbyte/pull/16243) | Fix Json to Avro conversion when there is field name clash from combined restrictions (`anyOf`, `oneOf`, `allOf` fields) |

0 commit comments

Comments
 (0)