https://linen.dev logo
Join Slack
Powered by
# ask-ai
  • h

    Hari Haran R

    09/04/2024, 3:17 AM
    Copy code
    "date_start_session": 1725349667,
    what is the datetime format for this datetime, getting error thrown during incremental sync
    k
    • 2
    • 9
  • g

    Graham Mackintosh

    09/04/2024, 4:49 AM
    @kapa.ai What authentication / validation methods are available to verify the authenticity of Airbyte success and failure webhook notifications?
    k
    • 2
    • 4
  • p

    Piotr Stankiewicz

    09/04/2024, 4:55 AM
    Can I remove prefix for and existing connection which was already synced?
    k
    • 2
    • 1
  • m

    Maximilian Hein

    09/04/2024, 6:35 AM
    @kapa.ai I have been trying to get airflow and airbyte to work together. Right now the provider for Airbyte is installed and the connection in Airflow is created. I got an example Connection with working Source and Destination in Airbyte and copyed the connection-id from the browser. When I use the the provider everything works fine, but the capabilities of the provider are not enough for my use-case. For my solution I need to access the Airbyte API, for example to change an Airbyte Source. Can you tell me how to do that?
    k
    • 2
    • 1
  • s

    Slackbot

    09/04/2024, 7:03 AM
    This message was deleted.
    k
    • 2
    • 1
  • s

    Slackbot

    09/04/2024, 7:06 AM
    This message was deleted.
    k
    • 2
    • 1
  • k

    kapa.ai

    09/04/2024, 7:24 AM
    Setting the
    JAVA_HOME
    variable in your environment is crucial for many Java applications. It should point to the location of your Java installation directory. Here's a general way to set `JAVA_HOME`: 1. Open your terminal. 2. Check if
    JAVA_HOME
    is already set by typing
    echo $JAVA_HOME
    . If it returns a directory path,
    JAVA_HOME
    is set. If it returns empty, you need to set it. 3. Find where Java is installed. You can do this by executing the command
    update-alternatives --config java
    . 4. Once you have the installation path, open your environment variables file. This could be
    ~/.bashrc
    ,
    ~/.bash_profile
    , or
    ~/.zshrc
    depending on your shell. 5. Add the following line to the file:
    export JAVA_HOME="your/java/installation/path"
    . Replace
    "your/java/installation/path"
    with your actual Java installation path. 6. Save the file and exit. 7. Run
    source ~/.bashrc
    or
    source ~/.bash_profile
    or
    source ~/.zshrc
    to reload the environment variables. 8. Verify that
    JAVA_HOME
    is set correctly by running
    echo $JAVA_HOME
    again. Please note that the exact steps may vary depending on your operating system and shell. If you're using Docker, you might need to set the environment variable in the Dockerfile or docker-compose file. In the context of Airbyte, an issue was reported where the
    JAVA_HOME
    was set to an invalid directory. The issue was discussed in this [Github issue](https://github.com/airbytehq/airbyte/issues/32435). The user was advised to set the
    JAVA_HOME
    variable in their environment to match the location of their Java installation.
  • x

    xinghui huang

    09/04/2024, 7:25 AM
    @kapa.ai ERROR: JAVA_HOME is set to an invalid directory
    k
    • 2
    • 1
  • d

    Durga Gurajala

    09/04/2024, 8:19 AM
    How do i start the connector in schema recovery mode? for the error - Error processing row in table, internal schema size 11, but row size 1 , restart connector with schema recovery mode.
    k
    • 2
    • 1
  • i

    Iulian

    09/04/2024, 8:38 AM
    Failure in source: Saved offset is not valid. Please reset the connection, and then increase oplog retention and/or increase sync frequency to prevent his from happening in the future. See https://docs.airbyte.com/integrations/sources/mongodb-v2#mongodb-oplog-and-change-streams for more details
    k
    • 2
    • 6
  • s

    Slackbot

    09/04/2024, 9:01 AM
    This message was deleted.
    k
    • 2
    • 1
  • h

    Hari Haran R

    09/04/2024, 9:04 AM
    how to convert "date_end_session": 1725436707 this Unix to a proper datetime format @kapa.ai
    k
    • 2
    • 4
  • m

    Moti Zamir

    09/04/2024, 9:15 AM
    Hi guys, I want to understand if there is an option to add custom data masking to our connectors, when BigQuery is our datawarehouse destination. We use the OSS AirByte version and we want a pre-ingestion layer that will mask our data as part of the ingestion process. Also we want to establish policy and roles to allow admin to see the full data and role for developers to see masked data
    k
    • 2
    • 1
  • i

    Ilya Dashkovsky

    09/04/2024, 9:28 AM
    @kapa.ai How to fix this Shopify sync error and what might cause it?
    Copy code
    2024-09-04 09:21:51 source > During the sync, the following streams did not sync successfully: products: AirbyteTracedException("'message'")
    
    2024-09-04 09:21:51 source > None
    
    Traceback (most recent call last):
    
    File "/airbyte/integration_code/main.py", line 9, in <module>
    
    run()
    
    File "/airbyte/integration_code/source_shopify/run.py", line 19, in run
    
    launch(source, sys.argv[1:])
    
    File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/entrypoint.py", line 235, in launch
    
    for message in source_entrypoint.run(parsed_args):
    
    File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/entrypoint.py", line 122, in run
    
    yield from map(AirbyteEntrypoint.airbyte_message_to_string, self.read(source_spec, config, config_catalog, state))
    
    File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/entrypoint.py", line 166, in read
    
    for message in self.source.read(self.logger, config, catalog, state):
    
    File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/abstract_source.py", line 185, in read
    
    raise AirbyteTracedException(message=error_message, failure_type=FailureType.config_error)
    
    airbyte_cdk.utils.traced_exception.AirbyteTracedException: None
    Copy code
    2024-09-04 09:01:03 replication-orchestrator > Attempt 2 to update stream status incomplete null:products error: io.airbyte.api.client.invoker.generated.ApiException: updateStreamStatus call failed with: 400 - {"message":"Incomplete run cause must be set for runs that stopped in an incomplete state.","exceptionClassName":"io.airbyte.server.apis.StreamStatusesApiController$Validations$2","exceptionStack":["io.airbyte.server.apis.StreamStatusesApiController$Validations$2: Incomplete run cause must be set for runs that stopped in an incomplete state.","\tat io.airbyte.server.apis.StreamStatusesApiController$Validations.validate(StreamStatusesApiController.java:100)","\tat io.airbyte.server.apis.StreamStatusesApiController.updateStreamStatus(StreamStatusesApiController.java:58)","\tat io.airbyte.server.apis.$StreamStatusesApiController$Definition$Exec.dispatch(Unknown Source)","\tat io.micronaut.context.AbstractExecutableMethodsDefinition$DispatchedExecutableMethod.invokeUnsafe(AbstractExecutableMethodsDefinition.java:461)","\tat io.micronaut.context.DefaultBeanContext$BeanContextUnsafeExecutionHandle.invokeUnsafe(DefaultBeanContext.java:4276)","\tat io.micronaut.web.router.AbstractRouteMatch.execute(AbstractRouteMatch.java:271)","\tat io.micronaut.http.server.RouteExecutor.executeRouteAndConvertBody(RouteExecutor.java:488)","\tat io.micronaut.http.server.RouteExecutor.lambda$callRoute$6(RouteExecutor.java:465)","\tat io.micronaut.core.execution.ExecutionFlow.lambda$async$1(ExecutionFlow.java:87)","\tat io.micronaut.core.propagation.PropagatedContext.lambda$wrap$3(PropagatedContext.java:211)","\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)","\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)","\tat java.base/java.lang.Thread.run(Thread.java:1583)"]}
    k
    • 2
    • 1
  • h

    Hari Haran R

    09/04/2024, 9:30 AM
    in my records i'm getting "date_start_session": 1725436642 but when converting this Timestamp in Transformation UI, i'm getting with Datetime format, but while airbyte is syncing i'm getting an error thrown as year format is not supported @kapa.ai
    k
    • 2
    • 1
  • h

    Hari Haran R

    09/04/2024, 10:57 AM
    convert unix timestamp to date from field in record my Record Data is date_start_session": 1725436642 @kapa.ai
    k
    • 2
    • 5
  • j

    Julie Choong

    09/04/2024, 11:24 AM
    Just set up in K8s. Syncs take forever to run, always stuck at workload claimed.
    k
    • 2
    • 12
  • c

    Christopher Daniel

    09/04/2024, 11:55 AM
    @kapa.ai, Does open source airbyte support cdc
    k
    • 2
    • 1
  • v

    Vlad Ivanyk

    09/04/2024, 12:00 PM
    give me an example value of lastUpdatedAt of Job
    k
    • 2
    • 6
  • c

    Cosmin Gabriel Petrenciuc

    09/04/2024, 12:30 PM
    @kapa.ai How can I force a retry of the last HTTP request from the parse_response method?
    k
    • 2
    • 1
  • j

    Justin Javis

    09/04/2024, 12:33 PM
    @kapa.ai We keep getting this error setting up a SharePoint source. Configuration check failed 'No streams are available for source SourceMicrosoftSharePoint. This is probably an issue with the connector. Please verify that your configuration provides permissions to list and read files from the source. Contact support if you are unable to resolve this issue.' 2024-09-04 122043 INFO i.a.w.l.p.s.m.Stage(apply):39 - APPLY Stage: CLAIM — (workloadId = 59353119-f0f2-4e5a-a8ba-15d887bc34f6_83c38ba9-4ae4-4645-aca4-f249bda5bb17_0_check) — (dataplaneId = local) 2024-09-04 122055 INFO i.a.c.i.LineGobbler(voidCall):166 - 2024-09-04 122055 INFO i.a.c.i.LineGobbler(voidCall):166 - ----- START CHECK ----- 2024-09-04 122055 INFO i.a.c.i.LineGobbler(voidCall):166 - 2024-09-04 122058 INFO i.a.c.ConnectorWatcher(run):87 - Connector exited, processing output 2024-09-04 122058 INFO i.a.c.ConnectorWatcher(run):90 - Output file jobOutput.json found 2024-09-04 122058 INFO i.a.c.ConnectorWatcher(run):96 - Connector exited with 0 2024-09-04 122058 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):189 - Reading messages from protocol version 0.2.0 2024-09-04 122058 WARN i.a.m.l.MetricClientFactory(getMetricClient):43 - MetricClient has not been initialized. Must call MetricClientFactory.CreateMetricClient before using MetricClient. Using a dummy client for now. Ignore this if Airbyte is configured to not publish any metrics. 2024-09-04 122058 ERROR i.a.w.i.VersionedAirbyteStreamFactory(internalLog):304 - Check failed 2024-09-04 122058 INFO i.a.c.ConnectorMessageProcessor(updateConfigFromControlMessage):231 - Checking for optional control message... 2024-09-04 122058 INFO i.a.c.ConnectorMessageProcessor(updateConfigFromControlMessage):253 - Optional control message not found. Skipping... 2024-09-04 122058 INFO i.a.c.ConnectorWatcher(run):134 - Writing output of 59353119-f0f2-4e5a-a8ba-15d887bc34f6_83c38ba9-4ae4-4645-aca4-f249bda5bb17_0_check to the doc store 2024-09-04 122059 INFO i.a.c.ConnectorWatcher(run):136 - Marking workload as successful 2024-09-04 122059 INFO i.a.c.ConnectorWatcher(exitProperly):189 - Deliberately exiting process with code 0. 2024-09-04 122059 INFO i.a.c.i.LineGobbler(voidCall):166 - 2024-09-04 122059 INFO i.a.c.i.LineGobbler(voidCall):166 - ----- END CHECK -----
    k
    • 2
    • 7
  • d

    Danton Bertuol

    09/04/2024, 2:49 PM
    @kapa.ai my discover schema has a timeout problem. Can I increase the timeout?
    k
    • 2
    • 10
  • v

    Viddu Devigere

    09/04/2024, 3:15 PM
    hello @kapa.ai I have a particular job that fails to launch, (other jobs have no issues). The job uses a MySQL source connector to Snowflake destination connector. App version 0.63.8 The orchestrator job seems to fail to launch. The error I receive is
    Copy code
    2024-09-04 15:12:09 WARN i.t.i.a.ActivityTaskExecutors$BaseActivityTaskExecutor(execute):126 - Activity failure. ActivityId=0659758b-91c3-3c96-aee6-2c1dd945fe7e, activityType=ReplicateV2, attempt=1
    java.lang.RuntimeException: io.temporal.serviceclient.CheckedExceptionWrapper: io.airbyte.workers.exception.WorkerException: Running the launcher replication-orchestrator failed
    	at io.airbyte.commons.temporal.HeartbeatUtils.withBackgroundHeartbeat(HeartbeatUtils.java:62) ~[io.airbyte-airbyte-commons-temporal-core-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.sync.ReplicationActivityImpl.replicateV2(ReplicationActivityImpl.java:174) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103) ~[?:?]
    	at java.base/java.lang.reflect.Method.invoke(Method.java:580) ~[?:?]
    	at io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:64) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:43) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:107) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:124) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:278) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:243) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:216) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.internal.worker.PollTaskExecutor.lambda$process$0(PollTaskExecutor.java:105) ~[temporal-sdk-1.22.3.jar:?]
    	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) ~[?:?]
    	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) ~[?:?]
    	at java.base/java.lang.Thread.run(Thread.java:1583) [?:?]
    Caused by: io.temporal.serviceclient.CheckedExceptionWrapper: io.airbyte.workers.exception.WorkerException: Running the launcher replication-orchestrator failed
    	at io.temporal.serviceclient.CheckedExceptionWrapper.wrap(CheckedExceptionWrapper.java:57) ~[temporal-serviceclient-1.22.3.jar:?]
    	at io.temporal.internal.sync.WorkflowInternal.wrap(WorkflowInternal.java:539) ~[temporal-sdk-1.22.3.jar:?]
    	at io.temporal.activity.Activity.wrap(Activity.java:52) ~[temporal-sdk-1.22.3.jar:?]
    	at io.airbyte.workers.temporal.TemporalAttemptExecution.get(TemporalAttemptExecution.java:143) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.sync.ReplicationActivityImpl.lambda$replicateV2$3(ReplicationActivityImpl.java:207) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.commons.temporal.HeartbeatUtils.withBackgroundHeartbeat(HeartbeatUtils.java:57) ~[io.airbyte-airbyte-commons-temporal-core-0.63.8.jar:?]
    	... 14 more
    Caused by: io.airbyte.workers.exception.WorkerException: Running the launcher replication-orchestrator failed
    	at io.airbyte.workers.sync.LauncherWorker.run(LauncherWorker.java:262) ~[io.airbyte-airbyte-commons-worker-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.TemporalAttemptExecution.get(TemporalAttemptExecution.java:138) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.sync.ReplicationActivityImpl.lambda$replicateV2$3(ReplicationActivityImpl.java:207) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.commons.temporal.HeartbeatUtils.withBackgroundHeartbeat(HeartbeatUtils.java:57) ~[io.airbyte-airbyte-commons-temporal-core-0.63.8.jar:?]
    	... 14 more
    Caused by: io.airbyte.workers.exception.WorkerException: Failed to create pod orchestrator-repl-job-362-attempt-0, pre-existing pod exists which didn't advance out of the NOT_STARTED state.
    	at io.airbyte.workers.sync.LauncherWorker.run(LauncherWorker.java:226) ~[io.airbyte-airbyte-commons-worker-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.TemporalAttemptExecution.get(TemporalAttemptExecution.java:138) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.sync.ReplicationActivityImpl.lambda$replicateV2$3(ReplicationActivityImpl.java:207) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.commons.temporal.HeartbeatUtils.withBackgroundHeartbeat(HeartbeatUtils.java:57) ~[io.airbyte-airbyte-commons-temporal-core-0.63.8.jar:?]
    	... 14 more
    Caused by: io.fabric8.kubernetes.client.KubernetesClientTimeoutException: Timed out waiting for [300000] milliseconds for [Pod] with name:[orchestrator-repl-job-362-attempt-0] in namespace [default].
    	at io.fabric8.kubernetes.client.dsl.internal.BaseOperation.waitUntilCondition(BaseOperation.java:946) ~[kubernetes-client-6.12.1.jar:?]
    	at io.fabric8.kubernetes.client.dsl.internal.BaseOperation.waitUntilCondition(BaseOperation.java:98) ~[kubernetes-client-6.12.1.jar:?]
    	at io.airbyte.workers.process.AsyncOrchestratorPodProcess.create(AsyncOrchestratorPodProcess.java:571) ~[io.airbyte-airbyte-commons-worker-0.63.8.jar:?]
    	at io.airbyte.workers.sync.LauncherWorker.run(LauncherWorker.java:217) ~[io.airbyte-airbyte-commons-worker-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.TemporalAttemptExecution.get(TemporalAttemptExecution.java:138) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.workers.temporal.sync.ReplicationActivityImpl.lambda$replicateV2$3(ReplicationActivityImpl.java:207) ~[io.airbyte-airbyte-workers-0.63.8.jar:?]
    	at io.airbyte.commons.temporal.HeartbeatUtils.withBackgroundHeartbeat(HeartbeatUtils.java:57) ~[io.airbyte-airbyte-commons-temporal-core-0.63.8.jar:?]
    	... 14 more
    k
    • 2
    • 2
  • d

    Dhinesh Balakrishnan

    09/04/2024, 3:41 PM
    i have loading data from redis to mssql, all columns are correctly mapped but my output table '_airbyte_data_' is not populated
    k
    • 2
    • 2
  • d

    Dhinesh Balakrishnan

    09/04/2024, 3:49 PM
    2024-09-04 153132 destination > INFO main i.a.i.d.m.MSSQLDestination(main):132 starting destination: class io.airbyte.integrations.destination.mssql.MSSQLDestination 2024-09-04 153132 destination > INFO main i.a.c.i.b.IntegrationCliParser$Companion(parseOptions):146 integration args: {catalog=destination_catalog.json, write=null, config=destination_config.json} 2024-09-04 153132 destination > INFO main i.a.c.i.b.IntegrationRunner(runInternal):123 Running integration: io.airbyte.cdk.integrations.base.ssh.SshWrappedDestination 2024-09-04 153132 destination > INFO main i.a.c.i.b.IntegrationRunner(runInternal):124 Command: WRITE 2024-09-04 153132 destination > INFO main i.a.c.i.b.IntegrationRunner(runInternal):125 Integration config: IntegrationConfig{command=WRITE, configPath='destination_config.json', catalogPath='destination_catalog.json', statePath='null'} 2024-09-04 153132 destination > WARN main c.n.s.JsonMetaSchema(newValidator):278 Unknown keyword order - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2024-09-04 153132 destination > WARN main c.n.s.JsonMetaSchema(newValidator):278 Unknown keyword airbyte_secret - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2024-09-04 153132 destination > INFO main i.a.c.i.b.s.SshWrappedDestination(getSerializedMessageConsumer):129 No SSH connection options found, using defaults 2024-09-04 153132 destination > INFO main i.a.c.i.b.s.SshTunnel$Companion(getInstance):433 Starting connection with method: NO_TUNNEL 2024-09-04 153133 destination > INFO main c.z.h.HikariDataSource(<init>):79 HikariPool-1 - Starting... 2024-09-04 153133 destination > INFO main c.z.h.HikariDataSource(<init>):81 HikariPool-1 - Start completed. 2024-09-04 153133 destination > INFO main i.a.c.i.d.j.JdbcBufferedConsumerFactory(toWriteConfig$lambda$0):149 Write config: WriteConfig{streamName=srcInvoiceRetent_newred, namespace=dbo, outputSchemaName=airbyte_internal, tmpTableName=_airbyte_tmp_sju_dbo_raw__stream_srcInvoiceRetent_newred, outputTableName=dbo_raw__stream_srcInvoiceRetent_newred, syncMode=append} 2024-09-04 153133 destination > INFO main i.a.c.i.d.a.b.BufferManager(<init>):43 Max 'memory' available for buffer allocation 296 MB 2024-09-04 153133 destination > INFO main i.a.c.i.b.IntegrationRunner$Companion(consumeWriteStream$io_airbyte_airbyte_cdk_java_airbyte_cdk_airbyte_cdk_core):381 Starting buffered read of input stream 2024-09-04 153133 destination > INFO main i.a.c.i.d.a.FlushWorkers(start):74 Start async buffer supervisor 2024-09-04 153133 destination > INFO pool-2-thread-1 i.a.c.i.d.a.b.BufferManager(printQueueInfo):89 [ASYNC QUEUE INFO] Global: max: 296.96 MB, allocated: 10 MB (10.0 MB), %% used: 0.03367428551701215 | State Manager memory usage: Allocated: 10 MB, Used: 0 bytes, percentage Used 0.0 2024-09-04 153133 destination > INFO main i.a.c.i.d.a.AsyncStreamConsumer(start):180 class io.airbyte.cdk.integrations.destination.async.AsyncStreamConsumer started. 2024-09-04 153133 destination > INFO main i.a.c.i.d.j.JdbcBufferedConsumerFactory(onStartFunction$lambda$1):199 Preparing raw tables in destination started for 1 streams 2024-09-04 153133 destination > INFO main i.a.c.i.d.j.JdbcBufferedConsumerFactory(onStartFunction$lambda$1):207 Preparing raw table in destination started for stream srcInvoiceRetent_newred. schema: airbyte_internal, table name: dbo_raw__stream_srcInvoiceRetent_newred 2024-09-04 153133 destination > INFO pool-5-thread-1 i.a.c.i.d.a.FlushWorkers(printWorkerInfo):128 [ASYNC WORKER INFO] Pool queue size: 0, Active threads: 0 2024-09-04 153134 destination > INFO main i.a.c.i.d.j.JdbcBufferedConsumerFactory(onStartFunction$lambda$1):229 Preparing raw tables in destination completed. 2024-09-04 153134 destination > INFO main i.a.c.i.b.IntegrationRunner$Companion(consumeWriteStream$io_airbyte_airbyte_cdk_java_airbyte_cdk_airbyte_cdk_core):403 Finished buffered read of input stream 2024-09-04 153134 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):205 Closing flush workers -- waiting for all buffers to flush 2024-09-04 153134 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):249 REMAINING_BUFFERS_INFO Namespace: dbo Stream: srcInvoiceRetent_newred -- remaining records: 1 2024-09-04 153134 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):250 Waiting for all streams to flush. 2024-09-04 153135 destination > INFO pool-4-thread-1 i.a.c.i.d.a.DetectStreamToFlush(getNextStreamToFlush):110 flushing: trigger info: dbo - srcInvoiceRetent_newred, time trigger: false , size trigger: true current threshold b: 0 bytes, queue size b: 198 bytes, penalty b: 0 bytes, after penalty b: 198 bytes 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.FlushWorkers(flush$lambda$6):136 Flush Worker (c62a9) -- Worker picked up work. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.FlushWorkers(flush$lambda$6):142 Flush Worker (c62a9) -- Attempting to read from queue namespace: dbo, stream: srcInvoiceRetent_newred. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.GlobalMemoryManager(free):78 Freeing 10485562 bytes.. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.FlushWorkers(flush$lambda$6):164 Flush Worker (c62a9) -- Batch contains: 1 records, 198 bytes bytes. 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):253 Closing flush workers -- all buffers flushed 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.s.GlobalAsyncStateManager(flushStates):155 Flushing states 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.s.GlobalAsyncStateManager(flushStates):209 Flushing states complete 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.GlobalMemoryManager(free):78 Freeing 0 bytes.. 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):261 Closing flush workers -- supervisor shut down 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):263 Closing flush workers -- Starting worker pool shutdown.. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.s.GlobalAsyncStateManager(flushStates):155 Flushing states 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.s.GlobalAsyncStateManager(flushStates):209 Flushing states complete 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.GlobalMemoryManager(free):78 Freeing 301 bytes.. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.GlobalMemoryManager(free):78 Freeing 198 bytes.. 2024-09-04 153135 destination > INFO pool-3-thread-1 i.a.c.i.d.a.FlushWorkers(flush$lambda$6):182 Flush Worker (c62a9) -- Worker finished flushing. Current queue size: 0 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.FlushWorkers(close):268 Closing flush workers -- workers shut down 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.b.BufferManager(close):68 Buffers cleared.. 2024-09-04 153135 destination > INFO main i.a.c.i.d.a.AsyncStreamConsumer(close):246 class io.airbyte.cdk.integrations.destination.async.AsyncStreamConsumer closed this is my destination log can you check any error will occur
    k
    • 2
    • 2
  • a

    Aldo Orozco

    09/04/2024, 4:51 PM
    @kapa.ai is there a way to change the initial email I used to log in to the airbyte UI? I tried to change it from Settings > Account Settings and saved the email, but that email doesn't work to log in
    k
    u
    b
    • 4
    • 17
  • a

    Aldo Orozco

    09/04/2024, 4:57 PM
    @kapa.ai is there a way to apply a custom SQL query on a source and use the result to upsert into a destination? Assume these are 2 relational databases (e.g., bigquery to postgres)
    k
    • 2
    • 1
  • k

    Karl Jose Buena

    09/04/2024, 5:11 PM
    @kapa.ai what are available wehhooks I can use in OSS version
    k
    • 2
    • 5
  • k

    kapa.ai

    09/04/2024, 6:30 PM
    Hello! How can I assist you with Airbyte today?
  • a

    Aditya Gupta

    09/04/2024, 6:31 PM
    @kapa.ai how can i create a OAuth flow for my OSS Airbyte Instace?
    k
    • 2
    • 7
1...323334...48Latest