Skip to content

Source Azure Blob Storage: add integration tests #36542

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
39 commits
Select commit Hold shift + click to select a range
29c6782
Source Azure Blob Storage: mirgate to poetry
artem1205 Mar 26, 2024
33a27ac
Source Azure Blob Storage: update docs
artem1205 Mar 26, 2024
c22fc81
Source Azure Blob Storage: update poetry
artem1205 Mar 26, 2024
f9421c4
Source Azure Blob Storage: update poetry
artem1205 Mar 26, 2024
8d5ce2c
Use DefaultFileBasedCursor instead of default (concurrent) cursor
clnoll Mar 26, 2024
7b24449
Source Azure Blob Storage: update spec
artem1205 Mar 26, 2024
abd16b6
Merge remote-tracking branch 'origin/master' into artem1205/source-az…
artem1205 Mar 27, 2024
d33d95b
Source Azure Blob Storage: update deps
artem1205 Mar 27, 2024
dab3661
Source Azure Blob Storage: add conftest + utils
artem1205 Mar 27, 2024
56d1a5e
Pin CDK version to last
lazebnyi Mar 27, 2024
1a3068f
Source Azure Blob Storage: test csv
artem1205 Mar 28, 2024
30d4bf7
Merge remote-tracking branch 'origin/master' into artem1205/source-az…
artem1205 Mar 28, 2024
0fb39b8
Source Azure Blob Storage: bump version
artem1205 Mar 28, 2024
5f10c5c
Source Azure Blob Storage: bump version
artem1205 Mar 28, 2024
6b98d38
Source Azure Blob Storage: add docker cli
artem1205 Mar 28, 2024
59cfa4d
Source Azure Blob Storage: add jq
artem1205 Mar 28, 2024
d07756e
Merge remote-tracking branch 'origin/master' into artem1205/source-az…
artem1205 Mar 28, 2024
1800601
Source Azure Blob Storage: update docker version
artem1205 Mar 29, 2024
10252d6
Source Azure Blob Storage: fix docker script
artem1205 Mar 29, 2024
defbeea
Source Azure Blob Storage: update poetry
artem1205 Mar 29, 2024
c3f2136
Source Azure Blob Storage: update conftest
artem1205 Mar 29, 2024
8bf120f
Source Azure Blob Storage: update test configuration
artem1205 Mar 29, 2024
d0fb85e
Source Azure Blob Storage: ref
artem1205 Mar 29, 2024
2ff13c6
Source Azure Blob Storage: ref
artem1205 Mar 29, 2024
c959bc9
Source Azure Blob Storage: remove unused
artem1205 Apr 1, 2024
467ebd0
Source Azure Blob Storage: add test for jsonl
artem1205 Apr 1, 2024
d89e5ec
Merge branch 'master' into artem1205/source-azure-blob-storage-certif…
artem1205 Apr 1, 2024
42ff0ee
Source Azure Blob Storage: fix tests scopes
artem1205 Apr 1, 2024
5d681d3
Source Azure Blob Storage: ref tests
artem1205 Apr 1, 2024
32cc7a2
Source Azure Blob Storage: bump CDK
artem1205 Apr 1, 2024
3497a6b
Source Azure Blob Storage: fix test
artem1205 Apr 1, 2024
1f944af
Source Azure Blob Storage: add test for parquet file
artem1205 Apr 2, 2024
ea5b189
Source Azure Blob Storage: add products properties
artem1205 Apr 2, 2024
a095425
Source Azure Blob Storage: add test for avro
artem1205 Apr 2, 2024
c9643fb
Merge remote-tracking branch 'origin/master' into artem1205/source-az…
artem1205 Apr 2, 2024
8c08137
Source Azure Blob Storage: add test for avro
artem1205 Apr 2, 2024
3e35c10
Source Azure Blob Storage: add .coveragerc
artem1205 Apr 2, 2024
f60e01f
Source Azure Blob Storage: fix schema type
artem1205 Apr 2, 2024
f8d76d4
Merge branch 'master' into artem1205/source-azure-blob-storage-certif…
artem1205 Apr 2, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
[run]
omit =
source_azure_blob_storage/run.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,23 @@
# Copyright (c) 2024 Airbyte, Inc., all rights reserved.

from __future__ import annotations

from typing import TYPE_CHECKING

if TYPE_CHECKING:
from dagger import Container


async def pre_connector_install(base_image_container: Container) -> Container:
"""
Docker compose is required to run the integration tests so we install Docker on top of the base image.
"""
return (
base_image_container.with_exec(["sh", "-c", "apt-get update && apt-get install -y curl jq"])
# Download install-docker.sh script
.with_exec(["curl", "-fsSL", "https://get.docker.com", "-o", "/tmp/install-docker.sh"])
# Run the install-docker.sh script with a pinned Docker version
.with_exec(["sh", "/tmp/install-docker.sh", "--version", "25.0"])
# Remove the install-docker.sh script
.with_exec(["rm", "/tmp/install-docker.sh"])
)
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
{
"azure_blob_storage_endpoint": "http://localhost:10000/account1",
"azure_blob_storage_account_name": "account1",
"azure_blob_storage_account_key": "key1",
"azure_blob_storage_container_name": "testcontainer",
"streams": [
{
"name": "users",
"file_type": "avro",
"globs": ["**/test_avro_users*.avro"],
"validation_policy": "Emit Record",
"format": {
"filetype": "avro"
}
},
{
"name": "purchases",
"file_type": "avro",
"globs": ["**/test_avro_purchases*.avro"],
"validation_policy": "Emit Record",
"format": {
"filetype": "avro"
}
},
{
"name": "products",
"file_type": "avro",
"globs": ["**/test_avro_products*.avro"],
"validation_policy": "Emit Record",
"format": {
"filetype": "avro"
}
}
]
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,128 @@
{
"azure_blob_storage_endpoint": "http://localhost:10000/account1",
"azure_blob_storage_account_name": "account1",
"azure_blob_storage_account_key": "key1",
"azure_blob_storage_container_name": "testcontainer",
"streams": [
{
"name": "users",
"file_type": "csv",
"globs": ["**/test_csv_users*.csv"],
"legacy_prefix": "",
"validation_policy": "Emit Record",
"format": {
"filetype": "csv",
"delimiter": ",",
"quote_char": "\"",
"double_quote": true,
"null_values": [
"",
"#N/A",
"#N/A N/A",
"#NA",
"-1.#IND",
"-1.#QNAN",
"-NaN",
"-nan",
"1.#IND",
"1.#QNAN",
"N/A",
"NA",
"NULL",
"NaN",
"n/a",
"nan",
"null"
],
"true_values": ["1", "True", "TRUE", "true"],
"false_values": ["0", "False", "FALSE", "false"],
"inference_type": "Primitive Types Only",
"strings_can_be_null": false,
"encoding": "utf8",
"header_definition": {
"header_definition_type": "From CSV"
}
}
},
{
"name": "purchases",
"file_type": "csv",
"globs": ["**/test_csv_purchases*.csv"],
"legacy_prefix": "",
"validation_policy": "Emit Record",
"format": {
"filetype": "csv",
"delimiter": ",",
"quote_char": "\"",
"double_quote": true,
"null_values": [
"",
"#N/A",
"#N/A N/A",
"#NA",
"-1.#IND",
"-1.#QNAN",
"-NaN",
"-nan",
"1.#IND",
"1.#QNAN",
"N/A",
"NA",
"NULL",
"NaN",
"n/a",
"nan",
"null"
],
"true_values": ["1", "True", "TRUE", "true"],
"false_values": ["0", "False", "FALSE", "false"],
"inference_type": "Primitive Types Only",
"strings_can_be_null": false,
"encoding": "utf8",
"header_definition": {
"header_definition_type": "From CSV"
}
}
},
{
"name": "products",
"file_type": "csv",
"globs": ["**/test_csv_products*.csv"],
"legacy_prefix": "",
"validation_policy": "Emit Record",
"format": {
"filetype": "csv",
"delimiter": ",",
"quote_char": "\"",
"double_quote": true,
"null_values": [
"",
"#N/A",
"#N/A N/A",
"#NA",
"-1.#IND",
"-1.#QNAN",
"-NaN",
"-nan",
"1.#IND",
"1.#QNAN",
"N/A",
"NA",
"NULL",
"NaN",
"n/a",
"nan",
"null"
],
"true_values": ["1", "True", "TRUE", "true"],
"false_values": ["0", "False", "FALSE", "false"],
"inference_type": "Primitive Types Only",
"strings_can_be_null": false,
"encoding": "utf8",
"header_definition": {
"header_definition_type": "From CSV"
}
}
}
]
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,41 @@
{
"azure_blob_storage_endpoint": "http://localhost:10000/account1",
"azure_blob_storage_account_name": "account1",
"azure_blob_storage_account_key": "key1",
"azure_blob_storage_container_name": "testcontainer",
"streams": [
{
"name": "users",
"file_type": "jsonl",
"globs": ["**/test_jsonl_users*.jsonl"],
"legacy_prefix": "",
"newlines_in_values": true,
"validation_policy": "Emit Record",
"format": {
"filetype": "jsonl"
}
},
{
"name": "purchases",
"file_type": "jsonl",
"globs": ["**/test_jsonl_purchases*.jsonl"],
"legacy_prefix": "",
"newlines_in_values": true,
"validation_policy": "Emit Record",
"format": {
"filetype": "jsonl"
}
},
{
"name": "products",
"file_type": "jsonl",
"globs": ["**/test_jsonl_products*.jsonl"],
"legacy_prefix": "",
"newlines_in_values": true,
"validation_policy": "Emit Record",
"format": {
"filetype": "jsonl"
}
}
]
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
{
"azure_blob_storage_endpoint": "http://localhost:10000/account1",
"azure_blob_storage_account_name": "account1",
"azure_blob_storage_account_key": "key1",
"azure_blob_storage_container_name": "testcontainer",
"streams": [
{
"name": "users",
"file_type": "jsonl",
"globs": ["**/test_parquet_users*.parquet"],
"validation_policy": "Emit Record",
"format": {
"filetype": "parquet"
}
},
{
"name": "purchases",
"file_type": "jsonl",
"globs": ["**/test_parquet_purchases*.parquet"],
"validation_policy": "Emit Record",
"format": {
"filetype": "parquet"
}
},
{
"name": "products",
"file_type": "jsonl",
"globs": ["**/test_parquet_products*.parquet"],
"validation_policy": "Emit Record",
"format": {
"filetype": "parquet"
}
}
]
}
Loading