|
1 |
| -# Apify Dataset Source |
| 1 | +# Apify-Dataset source connector |
2 | 2 |
|
3 |
| -This is the repository for the Apify Dataset configuration based source connector. |
| 3 | + |
| 4 | +This is the repository for the Apify-Dataset source connector, written in Python. |
4 | 5 | For information about how to use this connector within Airbyte, see [the documentation](https://docs.airbyte.com/integrations/sources/apify-dataset).
|
5 | 6 |
|
6 | 7 | ## Local development
|
7 | 8 |
|
8 |
| -#### Building via Python |
9 |
| - |
10 |
| -Create a Python virtual environment |
11 |
| - |
12 |
| -``` |
13 |
| -virtualenv --python $(which python3.10) .venv |
14 |
| -``` |
| 9 | +### Prerequisites |
| 10 | +* Python (~=3.9) |
| 11 | +* Poetry (~=1.7) - installation instructions [here](https://python-poetry.org/docs/#installation) |
15 | 12 |
|
16 |
| -Source it |
17 | 13 |
|
18 |
| -``` |
19 |
| -source .venv/bin/activate |
| 14 | +### Installing the connector |
| 15 | +From this connector directory, run: |
| 16 | +```bash |
| 17 | +poetry install --with dev |
20 | 18 | ```
|
21 | 19 |
|
22 |
| -Check connector specifications/definition |
23 | 20 |
|
24 |
| -``` |
25 |
| -python main.py spec |
26 |
| -``` |
| 21 | +### Create credentials |
| 22 | +**If you are a community contributor**, follow the instructions in the [documentation](https://docs.airbyte.com/integrations/sources/apify-dataset) |
| 23 | +to generate the necessary credentials. Then create a file `secrets/config.json` conforming to the `source_apify_dataset/spec.yaml` file. |
| 24 | +Note that any directory named `secrets` is gitignored across the entire Airbyte repo, so there is no danger of accidentally checking in sensitive information. |
| 25 | +See `sample_files/sample_config.json` for a sample config file. |
27 | 26 |
|
28 |
| -Basic check - check connection to the API |
29 | 27 |
|
| 28 | +### Locally running the connector |
30 | 29 | ```
|
31 |
| -python main.py check --config secrets/config.json |
| 30 | +poetry run source-apify-dataset spec |
| 31 | +poetry run source-apify-dataset check --config secrets/config.json |
| 32 | +poetry run source-apify-dataset discover --config secrets/config.json |
| 33 | +poetry run source-apify-dataset read --config secrets/config.json --catalog sample_files/configured_catalog.json |
32 | 34 | ```
|
33 | 35 |
|
34 |
| -Integration tests - read operation from the API |
35 |
| - |
| 36 | +### Running unit tests |
| 37 | +To run unit tests locally, from the connector directory run: |
36 | 38 | ```
|
37 |
| -python main.py read --config secrets/config.json --catalog integration_tests/configured_catalog.json |
| 39 | +poetry run pytest unit_tests |
38 | 40 | ```
|
39 | 41 |
|
40 |
| -#### Create credentials |
41 |
| - |
42 |
| -**If you are a community contributor**, follow the instructions in the [documentation](https://docs.airbyte.com/integrations/sources/apify-dataset) |
43 |
| -to generate the necessary credentials. Then create a file `secrets/config.json` conforming to the `source_apify_dataset/spec.yaml` file. |
44 |
| -Note that any directory named `secrets` is gitignored across the entire Airbyte repo, so there is no danger of accidentally checking in sensitive information. |
45 |
| -See `integration_tests/sample_config.json` for a sample config file. |
46 |
| - |
47 |
| -**If you are an Airbyte core member**, copy the credentials in Lastpass under the secret name `source apify-dataset test creds` |
48 |
| -and place them into `secrets/config.json`. |
49 |
| - |
50 |
| -### Locally running the connector docker image |
51 |
| - |
52 |
| - |
53 |
| -#### Build |
54 |
| -**Via [`airbyte-ci`](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/README.md) (recommended):** |
| 42 | +### Building the docker image |
| 43 | +1. Install [`airbyte-ci`](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/README.md) |
| 44 | +2. Run the following command to build the docker image: |
55 | 45 | ```bash
|
56 | 46 | airbyte-ci connectors --name=source-apify-dataset build
|
57 | 47 | ```
|
58 | 48 |
|
59 |
| -An image will be built with the tag `airbyte/source-apify-dataset:dev`. |
| 49 | +An image will be available on your host with the tag `airbyte/source-apify-dataset:dev`. |
60 | 50 |
|
61 |
| -**Via `docker build`:** |
62 |
| -```bash |
63 |
| -docker build -t airbyte/source-apify-dataset:dev . |
64 |
| -``` |
65 |
| - |
66 |
| -#### Run |
67 | 51 |
|
| 52 | +### Running as a docker container |
68 | 53 | Then run any of the connector commands as follows:
|
69 |
| - |
70 | 54 | ```
|
71 | 55 | docker run --rm airbyte/source-apify-dataset:dev spec
|
72 | 56 | docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-apify-dataset:dev check --config /secrets/config.json
|
73 | 57 | docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-apify-dataset:dev discover --config /secrets/config.json
|
74 | 58 | docker run --rm -v $(pwd)/secrets:/secrets -v $(pwd)/integration_tests:/integration_tests airbyte/source-apify-dataset:dev read --config /secrets/config.json --catalog /integration_tests/configured_catalog.json
|
75 | 59 | ```
|
76 | 60 |
|
77 |
| - |
78 |
| -## Testing |
| 61 | +### Running our CI test suite |
79 | 62 | You can run our full test suite locally using [`airbyte-ci`](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/README.md):
|
80 | 63 | ```bash
|
81 | 64 | airbyte-ci connectors --name=source-apify-dataset test
|
82 | 65 | ```
|
83 | 66 |
|
84 | 67 | ### Customizing acceptance Tests
|
85 |
| -Customize `acceptance-test-config.yml` file to configure tests. See [Connector Acceptance Tests](https://docs.airbyte.com/connector-development/testing-connectors/connector-acceptance-tests-reference) for more information. |
| 68 | +Customize `acceptance-test-config.yml` file to configure acceptance tests. See [Connector Acceptance Tests](https://docs.airbyte.com/connector-development/testing-connectors/connector-acceptance-tests-reference) for more information. |
86 | 69 | If your connector requires to create or destroy resources for use during acceptance tests create fixtures for it and place them inside integration_tests/acceptance.py.
|
87 | 70 |
|
88 |
| -## Dependency Management |
89 |
| - |
90 |
| -All of your dependencies should go in `setup.py`, NOT `requirements.txt`. The requirements file is only used to connect internal Airbyte dependencies in the monorepo for local development. |
91 |
| -We split dependencies between two groups, dependencies that are: |
| 71 | +### Dependency Management |
| 72 | +All of your dependencies should be managed via Poetry. |
| 73 | +To add a new dependency, run: |
| 74 | +```bash |
| 75 | +poetry add <package-name> |
| 76 | +``` |
92 | 77 |
|
93 |
| -- required for your connector to work need to go to `MAIN_REQUIREMENTS` list. |
94 |
| -- required for the testing need to go to `TEST_REQUIREMENTS` list |
| 78 | +Please commit the changes to `pyproject.toml` and `poetry.lock` files. |
95 | 79 |
|
96 |
| -### Publishing a new version of the connector |
| 80 | +## Publishing a new version of the connector |
97 | 81 | You've checked out the repo, implemented a million dollar feature, and you're ready to share your changes with the world. Now what?
|
98 | 82 | 1. Make sure your changes are passing our test suite: `airbyte-ci connectors --name=source-apify-dataset test`
|
99 |
| -2. Bump the connector version in `metadata.yaml`: increment the `dockerImageTag` value. Please follow [semantic versioning for connectors](https://docs.airbyte.com/contributing-to-airbyte/resources/pull-requests-handbook/#semantic-versioning-for-connectors). |
| 83 | +2. Bump the connector version (please follow [semantic versioning for connectors](https://docs.airbyte.com/contributing-to-airbyte/resources/pull-requests-handbook/#semantic-versioning-for-connectors)): |
| 84 | + - bump the `dockerImageTag` value in in `metadata.yaml` |
| 85 | + - bump the `version` value in `pyproject.toml` |
100 | 86 | 3. Make sure the `metadata.yaml` content is up to date.
|
101 |
| -4. Make the connector documentation and its changelog is up to date (`docs/integrations/sources/apify-dataset.md`). |
| 87 | +4. Make sure the connector documentation and its changelog is up to date (`docs/integrations/sources/apify-dataset.md`). |
102 | 88 | 5. Create a Pull Request: use [our PR naming conventions](https://docs.airbyte.com/contributing-to-airbyte/resources/pull-requests-handbook/#pull-request-title-convention).
|
103 | 89 | 6. Pat yourself on the back for being an awesome contributor.
|
104 | 90 | 7. Someone from Airbyte will take a look at your PR and iterate with you to merge it into master.
|
105 |
| - |
| 91 | +8. Once your PR is merged, the new version of the connector will be automatically published to Docker Hub and our connector registry. |
0 commit comments