dilan silva
05/22/2025, 8:55 AMversion: 6.48.15
type: DeclarativeSource
check:
type: CheckStream
stream_names:
- datasets
definitions:
streams:
datasets:
type: DeclarativeStream
name: datasets
retriever:
type: SimpleRetriever
requester:
$ref: "#/definitions/base_requester"
path: /api/export
http_method: GET
record_selector:
type: RecordSelector
extractor:
type: DpathExtractor
field_path: []
decoder:
type: JsonlDecoder
schema_loader:
type: InlineSchemaLoader
schema:
$ref: "#/schemas/datasets"
base_requester:
type: HttpRequester
url_base: >-
<https://03100670-8969-4472-a593-d7a8cef4488b-00-15uejhh5ki8b9.janeway.replit.dev>
streams:
- $ref: "#/definitions/streams/datasets"
spec:
type: Spec
connection_specification:
type: object
$schema: <http://json-schema.org/draft-07/schema#>
required: []
properties: {}
additionalProperties: true
schemas:
datasets:
type: object
$schema: <http://json-schema.org/draft-07/schema#>
additionalProperties: true
properties: {}
Now when I try to locally do this work in my local connector, it does not load the data, but it says Sync success with 0 bytes. In the log I can see these messages, (Not pasting full log here)
2025-05-22 13:54:15 source ERROR Marking stream records as STARTED
2025-05-22 13:54:15 source ERROR Syncing stream instance: records
2025-05-22 13:54:15 source ERROR Setting state of SourceNexusDatasets stream to {}
2025-05-22 13:54:15 source ERROR Syncing stream: records
2025-05-22 13:54:15 source ERROR Making outbound API request
2025-05-22 13:54:15 source INFO Starting syncing SourceNexusDatasets
2025-05-22 13:54:15 source INFO Marking stream records as STARTED
2025-05-22 13:54:15 source INFO Malformed non-Airbyte record (connectionId = 034b7256-278b-4830-bba5-953c366de232): {"type": "DEBUG", "message": "Syncing stream instance: records", "data": {"message": "Syncing stream instance: records", "cursor_field": "[]", "primary_key": "None"}}
2025-05-22 13:54:15 source INFO Setting state of SourceNexusDatasets stream to {}
2025-05-22 13:54:15 source INFO Syncing stream: records
2025-05-22 13:54:15 source INFO Malformed non-Airbyte record (connectionId = 034b7256-278b-4830-bba5-953c366de232): {"type": "DEBUG", "message": "Making outbound API request", "data": {"request_body": "None", "headers": "{'User-Agent': 'python-requests/2.32.3', 'Accept-Encoding': 'gzip, deflate', 'Accept': '*/*', 'Connection': 'keep-alive'}", "url": "<https://03100670-8969-4472-a593-d7a8cef4488b-00-15uejhh5ki8b9.janeway.replit.dev/api/export>", "message": "Making outbound API request"}}
2025-05-22 13:54:15 replication-orchestrator INFO Stream status TRACE received of status: STARTED for stream records
2025-05-22 13:54:15 replication-orchestrator INFO Sending update for records - null -> RUNNING
2025-05-22 13:54:15 replication-orchestrator INFO Stream Status Update Received: records - RUNNING
2025-05-22 13:54:15 replication-orchestrator INFO Creating status: records - RUNNING
2025-05-22 13:54:17 source ERROR Receiving response
2025-05-22 13:54:17 source INFO Malformed non-Airbyte record (connectionId = 034b7256-278b-4830-bba5-953c366de232): {"type": "DEBUG", "message": "Receiving response", "data": {"headers": "{'Content-Disposition': 'attachment; filename=\"ExportData.json\"', 'Content-Type': 'application/json', 'Date': 'Thu, 22 May 2025 08:24:19 GMT', 'Replit-Cluster': 'janeway', 'X-Powered-By': 'Express', 'X-Robots-Tag': 'none, noindex, noarchive, nofollow, nositelinkssearchbox, noimageindex, none, noindex, noarchive, nofollow, nositelinkssearchbox, noimageindex', 'Transfer-Encoding': 'chunked'}", "body": "{\"id\": 1, \"name\": \"Example Item 1\", \"category\": \"Category A\", \"price\": 19.99}\n{\"id\": 2, \"name\": \"Example Item 2\", \"category\": \"Category B\", \"price\": 29.99}\n{\"id\": 3, \"name\": \"Example Item 3\", \"category\": \"Category A\", \"price\": 15.50}\n{\"id\": 4, \"name\": \"Example Item 4\", \"category\": \"Category C\", \"price\": 45.00}\n{\"id\": 5, \"name\": \"Example Item 5\", \"category\": \"Category B\", \"price\": 35.25}", "message": "Receiving response", "status": "200"}}
2025-05-22 13:54:17 source ERROR Read 0 records from records stream
2025-05-22 13:54:17 source ERROR Marking stream records as STOPPED
2025-05-22 13:54:17 source ERROR Finished syncing records
2025-05-22 13:54:17 source ERROR SourceNexusDatasets runtimes:
2025-05-22 13:54:17 source ERROR Syncing stream records 0:00:01.715473
2025-05-22 13:54:17 source ERROR Finished syncing SourceNexusDatasets
I have the same manifest configuration in the local connector except the version
,
version: 0.90.0
type: DeclarativeSource
check:
type: CheckStream
stream_names:
- "records"
definitions:
streams:
records:
type: DeclarativeStream
name: records
retriever:
type: SimpleRetriever
requester:
type: HttpRequester
url_base: <https://03100670-8969-4472-a593-d7a8cef4488b-00-15uejhh5ki8b9.janeway.replit.dev>
path: /api/export
http_method: GET
record_selector:
type: RecordSelector
extractor:
type: DpathExtractor
field_path: []
decoder:
type: JsonlDecoder
schema_loader:
type: InlineSchemaLoader
schema:
$ref: "#/schemas/datasets"
streams:
- "#/definitions/streams/records"
schemas:
datasets:
type: object
$schema: <http://json-schema.org/draft-07/schema#>
additionalProperties: true
properties: {}
Can someone please help on this ? I tried to change the version but it give an error,
jsonschema.exceptions.ValidationError: The manifest version 6.48.15 is greater than the airbyte-cdk package version (0.90.0). Your manifest may contain features
that are not in the current CDK version..
Mathieu Dumoulin
05/22/2025, 5:53 PMHadrien Lepousé
05/24/2025, 7:51 PMpoetry run pytest unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records
Result:
======================================== short test summary info =========================================
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[tickets_web_analytics-tickets-ticket-parent_stream_associations0] - ValueError: Invalid number of matches for `HttpRequestMatcher(request_to_match=ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/objects/ticket', params='', query='archived=false&associations=contacts&associations=deals&associations=companies&limit=100&properties=closed_date,createdate', fragment='') with headers {} and body None), minimum_number_of_expected_match=1, actual_number_of_matches=0)`
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[deals_web_analytics-deals-deal-parent_stream_associations1] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[companies_web_analytics-companies-company-parent_stream_associations2] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[contacts_web_analytics-contacts-contact-parent_stream_associations3] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[engagements_calls_web_analytics-engagements_calls-calls-parent_stream_associations4] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[engagements_emails_web_analytics-engagements_emails-emails-parent_stream_associations5] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[engagements_meetings_web_analytics-engagements_meetings-meetings-parent_stream_associations6] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[engagements_notes_web_analytics-engagements_notes-notes-parent_stream_associations7] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
FAILED unit_tests/integrations/test_web_analytics_streams.py::TestCRMWebAnalyticsStream::test_given_two_pages_when_read_then_return_records[engagements_tasks_web_analytics-engagements_tasks-tasks-parent_stream_associations8] - ValueError: Request ParseResult(scheme='https', netloc='<http://api.hubapi.com|api.hubapi.com>', path='/crm/v3/schemas', params='', query='', fragment='') with headers {} and body None) already mocked
Giulliano Bueno
06/02/2025, 8:41 AMgradlew
build process. Can anyone point me to tutorials or resources on contributing to this project or setting up Databricks as a source without creating a new connector?Erin Yener
06/02/2025, 11:23 AMInputs
section, but would love confirmation on how best to do this.
• If it’s best to modify the YAML directly, would anyone be able to advise me on what section of the YAML to add the API budget to? The docs have some examples, but I’m not clear on how to add this to a forked connector.
• Are there existing connector examples that have this parameter as an optional input so that I can see how it ‘fits’?
Thanks for the help!Jens Mostaert
06/04/2025, 12:53 PMNick Zombolas
06/04/2025, 8:06 PMspec
job. I've been looking into how to implement this steam in yaml instead of calling my python class, but not sure the best way to get this done. any advice? thanks!Mert Ors
06/05/2025, 9:05 AMPaul
06/05/2025, 1:35 PMairbyte-ci connectors --name destination-azure-blob-storage build
But sadly after about 30 seconds it fails. The build output html is empty. The dagger.log file is also empty. This is what i get in the window log below - any ideas what ive missed? ive updated the secrets as per the readme, but no luck.
[23:33:09] INFO root: Setting working directory to /home/pladmin/airbyte/airbyte ensure_repo_root.py:58
[23:33:10] INFO root: Setting working directory to /home/pladmin/airbyte/airbyte ensure_repo_root.py:58
INFO pipelines: airbyte-ci is up to date. Installed version: 5.2.5. Latest version: 5.2.5 auto_update.py:89
INFO pipelines: Called with dagger run: False airbyte_ci.py:127
INFO pipelines.cli.dagger_run: Running command: ['/home/pladmin/bin/dagger', '--silent', 'run', 'airbyte-ci', 'connectors', '--name', 'destination-azure-blob-storage', 'build'] dagger_run.py:120
[23:33:18] INFO root: Setting working directory to /home/pladmin/airbyte/airbyte ensure_repo_root.py:58
[23:33:19] INFO root: Setting working directory to /home/pladmin/airbyte/airbyte ensure_repo_root.py:58
INFO pipelines: airbyte-ci is up to date. Installed version: 5.2.5. Latest version: 5.2.5 auto_update.py:89
INFO pipelines: Called with dagger run: True airbyte_ci.py:127
[23:33:27] INFO pipelines: Will run on the following 1 connectors: destination-azure-blob-storage. commands.py:32
INFO pipelines: Running Dagger Command build... dagger_pipeline_command.py:32
INFO pipelines: If you're running this command for the first time the Dagger engine image will be pulled, it can take a short minute... dagger_pipeline_command.py:33
INFO pipelines: Saving dagger logs to: dagger_pipeline_command.py:43
/home/pladmin/airbyte/airbyte/airbyte-ci/connectors/pipelines/pipeline_reports/airbyte-ci/connectors/build/manual/master/1749094400/b2ffb0185be442ddf72677067d3a8
243fbba770f/dagger.log
INFO pipelines: Building connectors for ['linux/amd64'], use --architecture to change this. commands.py:46
INFO Build connector destination-azure-blob-storage: Should send status check: False pipeline_context.py:222
[23:33:29] INFO root: Using storage driver: fuse-overlayfs docker.py:85
[23:33:56] INFO Build connector destination-azure-blob-storage: Caching the latest CDK version... pipeline_context.py:284
INFO Build connector destination-azure-blob-storage: Should send status check: False pipeline_context.py:222
INFO Build connector destination-azure-blob-storage - Build connector tar: 🚀 Start Build connector tar steps.py:303
ERROR Build connector destination-azure-blob-storage: An error got handled by the ConnectorContext context.py:253
╭───────────────────────────────────────────────────────────────────── Traceback (most recent call last) ──────────────────────────────────────────────────────────────────────╮
│ in run_connector_build_pipeline:49 │
│ │
│ in run_connector_build:33 │
│ │
│ in run_connector_build:60 │
│ │
│ in run:307 │
│ │
│ in __aexit__:772 │
╰──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
ERROR Build connector destination-azure-blob-storage: No test report was provided. This is probably due to an upstream error context.py:255
╭──────────────────────────────────────────────────────────────────────────────────── DESTINATION-AZURE-BLOB-STORAGE - REPORT ────────────────────────────────────────────────────────────────────────────────────╮
│ Steps results │
│ ┏━━━━━━┳━━━━━━━━┳━━━━━━━━━━┓ │
│ ┃ Step ┃ Result ┃ Duration ┃ │
│ ┡━━━━━━╇━━━━━━━━╇━━━━━━━━━━┩ │
│ └──────┴────────┴──────────┘ │
│ ℹ️ You can find more details with step executions logs in the saved HTML report. │
╰───────────────────────────────────────────────────────────────────── ⏲️ Total pipeline duration for destination-azure-blob-storage: 0.36s ────────────────────────────────────────────────────────────────────
Aphonso Henrique do Amaral Rafael
06/05/2025, 5:10 PMJuliette Duizabo
06/09/2025, 4:39 PMConfiguration check failed
'Encountered an error while checking availability of stream sources. Error: Request URL: <https://api.airbyte.com/v1/applications/token>, Response Code: 500, Response Text: {"message":"Internal Server Error","_links":{"self":{"href":"/api/public/v1/applications/token","templated":false}},"_embedded":{"errors":[{"message":"Internal Server Error: class org.jboss.resteasy.client.jaxrs.engines.ManualClosingApacheHttpClient43Engine$3 cannot be cast to class io.micronaut.jaxrs.common.JaxRsMutableResponse (org.jboss.resteasy.client.jaxrs.engines.ManualClosingApacheHttpClient43Engine$3 and io.micronaut.jaxrs.common.JaxRsMutableResponse are in unnamed module of loader \'app\')","_links":{},"_embedded":{}}]}}'
It looks like the issue is on Airbyte's side. Has any of you managed to set up the import of Airbyte metadata to have the observability in their warehouse?Gergely Imreh
06/10/2025, 12:58 PMid
and have values of id1
, id2
, etc)
• The child stream would need a payload to query that puts those id
s into a list in the request body such as:
{"input": [{"id": id1}, {"id": id2}, ....]}
and sends of that query (it's a batch one by default)
Is this possible to configure (with a parent substream like this)?
Or do I have to just run a sequential list of queries with
{"input": [{"id": id1}]}
then
{"input": [{"id": id2}]}
....
This would likely work, though probably hit rate limits, and takes longer time than the one that would run things in one go.
Any suggestions? 🤔Mike Moyer
06/10/2025, 9:34 PMAnthony Smart
06/11/2025, 1:40 PMAlbert Le
06/11/2025, 6:57 PM/studies
.
There is another subendpoint called /studies/{id}/availability
, where {id} is a single study id.
Does the new builder Tool have an automated way of allowing me to call /studies endpoint andget a list of study_ids, and use that response as the input query parameter for the subendpoint?
What i tried: Searched through documentation, but couldn't find anything for my use-case.Jeroen de Vries
06/12/2025, 10:26 AMFailed to build connector image for platform linux/arm64: resolve: process "poetry check" did not complete successfully: exit code: 1
But I want to exclude the use of Poetry.Matthew Wagaman
06/12/2025, 7:17 PMRashi Bhave
06/12/2025, 7:40 PMTheo Marchal
06/16/2025, 9:35 AMAlejandro De La Cruz López
06/16/2025, 10:51 AMClient error : 400 Bad Request {"exceptionStack":"Traceback (most recent call last):\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/connector_builder/connector_builder_handler.py\", line 83, in read_stream\n stream_read = test_read_handler.run_test_read(\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/connector_builder/test_reader/reader.py\", line 128, in run_test_read\n schema, log_messages = self._get_infered_schema(\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/connector_builder/test_reader/reader.py\", line 328, in _get_infered_schema\n schema = schema_inferrer.get_stream_schema(configured_stream.stream.name)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 266, in get_stream_schema\n self._clean(self.stream_to_builder[stream_name].to_schema())\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 154, in _clean\n self._clean_properties(node)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 131, in _clean_properties\n self._clean(value)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 157, in _clean\n self._clean(node[_ITEMS])\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 154, in _clean\n self._clean_properties(node)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 131, in _clean_properties\n self._clean(value)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 160, in _clean\n self._ensure_null_type_on_top(node)\n File \"/home/airbyte/.pyenv/versions/3.10.17/lib/python3.10/site-packages/airbyte_cdk/utils/schema_inferrer.py\", line 134, in _ensure_null_type_on_top\n if isinstance(node[_TYPE], list):\nKeyError: 'type'\n","exceptionClassName":"io.airbyte.protocol.models.v0.AirbyteTraceMessage
Cadu Magalhães
06/16/2025, 4:48 PMArthur Dev
06/17/2025, 2:42 PMKailash Bisht
06/19/2025, 1:05 PMSushmita Sen
06/20/2025, 7:16 AMForrest Hicks
06/20/2025, 5:22 PMAnanta Patil
06/23/2025, 10:50 AMJens Mostaert
06/23/2025, 1:43 PMVivien Morlet
06/23/2025, 5:32 PMKailash Bisht
06/24/2025, 8:19 AMEyþór Helgason
06/24/2025, 9:59 AMpoe
instead of airbyte-ci
? (should be depreciated according to comment above)
I am working on adding metafield definition streams and when I run poe test-integration-tests
in directory /airbyte-integrations/connectors/source-shopify
I get the following error:
Poe => set -eu # Ensure we return non-zero exit code upon failure
if ls integration_tests/test_*.py >/dev/null 2>&1; then
poetry run pytest --junitxml=build/test-results/pytest-integration-tests-junit.xml integration_tests
else
echo "No 'integration_tests' directory found; skipping integration tests."
fi
The currently activated Python version 3.12.8 is not supported by the project (^3.10,<3.12).
Trying to find and use a compatible version.
Using python3.11 (3.11.11)
No 'integration_tests' directory found; skipping integration tests.
Even though integration_tests is defined with the following files:
integration_tests
__init__.py
abnormal_state.json
acceptance.py
configured_catalog.json
expected_records.jsonl
expected_records_transactions_with_user_id.jsonl
invalid_config.json
invalid_config_old.json
invalid_oauth_config.json
state.json
Any help would be greatly appreciated :)