Salvador Pardiñas
01/29/2025, 4:53 PMwith info_1 as (select distinct spm."CODE" as "CODE", spm."ID" as ID
from d_object_info spm
where spm.SIGNUP_DATE is not null and spm.STATUS = 'ACTIVE'),
info_2 as (select *
from "d_object_metric"
where "CATEGORY" = 'SOME CATEGORY'
and TIME_EXTRACT(__time, 'year') = 2024
and TIME_EXTRACT(__time, 'month') = 06)
select spm.CODE,
(select coalesce(sum(METRIC), 0)
from info_2 vma
where vma.CODE = spm.CODE) as total_metric
from info_1 spm
Which was working correctly on Druid 28, we've updated to Druid 31 and all of a sudden we get the following error:
Calcite assertion violated: [Cannot add expression of different type to set: set type is RecordType(VARCHAR COD, VARCHAR ID, DOUBLE EXPR$0) NOT NULL expression type is RecordType(VARCHAR CODE, VARCHAR ID, DOUBLE NOT NULL EXPR$0) NOT NULL
I've looked through the release notes and can't find any breaking changes that might be affecting this query. Any ideas?Salvador Pardiñas
01/29/2025, 6:08 PMinfo Advisionary
01/30/2025, 11:09 AMLionel Mena
01/30/2025, 2:39 PM2025-01-30T14:10:21,505 ERROR [task-runner-0-priority-0] org.apache.druid.indexing.seekablestream.SeekableStreamIndexTaskRunner - Encountered exception in run() before persisting.
java.lang.InterruptedException: null
at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1640) ~[?:?]
at org.apache.druid.indexing.seekablestream.SeekableStreamIndexTaskRunner.possiblyPause(SeekableStreamIndexTaskRunner.java:1356) ~[druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.seekablestream.SeekableStreamIndexTaskRunner.runInternal(SeekableStreamIndexTaskRunner.java:595) [druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.seekablestream.SeekableStreamIndexTaskRunner.run(SeekableStreamIndexTaskRunner.java:277) [druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.seekablestream.SeekableStreamIndexTask.runTask(SeekableStreamIndexTask.java:153) [druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.common.task.AbstractTask.run(AbstractTask.java:179) [druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.overlord.SingleTaskBackgroundRunner$SingleTaskBackgroundRunnerCallable.call(SingleTaskBackgroundRunner.java:478) [druid-indexing-service-30.0.1.jar:30.0.1]
at org.apache.druid.indexing.overlord.SingleTaskBackgroundRunner$SingleTaskBackgroundRunnerCallable.call(SingleTaskBackgroundRunner.java:450) [druid-indexing-service-30.0.1.jar:30.0.1]
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:131) [guava-32.0.1-jre.jar:?]
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:75) [guava-32.0.1-jre.jar:?]
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:82) [guava-32.0.1-jre.jar:?]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) [?:?]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) [?:?]
at java.base/java.lang.Thread.run(Thread.java:840) [?:?]
Overlord error message:
2025-01-30T14:10:34,804 ERROR [TaskQueue-OnComplete-0] org.apache.druid.indexing.overlord.TaskQueue - Ignoring notification for already-complete task: {class=org.apache.druid.indexing.overlord.TaskQueue, task=index_kafka_sdk-americas-realtime_f26b8ca42fa849f_pfaamibg}
You can find attached the:
• Task's full logs
• Associated Overlord and MiddleManager logs.
• Supervisor spec.
• Datasource auto compaction specs.Maytas Monsereenusorn
01/31/2025, 1:24 AMMiguel Vieira Colombo
02/03/2025, 4:04 PM-server
-Xms12g
-Xmx12g
-XX:+ExitOnOutOfMemoryError
-XX:+UseG1GC
-XX:+UseStringDeduplication
-XX:ParallelGCThreads=5
-XX:ConcGCThreads=2
-Duser.timezone=UTC
-Dfile.encoding=UTF-8
-Djava.io.tmpdir=var/tmp
-Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager
-Dderby.stream.error.file=var/druid/derby.log
-Daws.region=us-east-1
-Djute.maxbuffer=15728640
druid.service=druid/coordinator
druid.plaintextPort=8081
druid.coordinator.startDelay=PT10S
druid.coordinator.period=PT5S
druid.manager.segments.pollDuration=PT5S
druid.indexer.logs.kill.enabled=true
druid.indexer.logs.kill.delay=14400000
druid.indexer.logs.kill.durationToRetain=86400000
# Run the overlord service in the coordinator process
druid.coordinator.asOverlord.enabled=true
druid.coordinator.asOverlord.overlordService=druid/overlord
druid.indexer.storage.recentlyFinishedThreshold=PT12H
druid.manager.config.pollDuration=PT10M
druid.indexer.runner.maxZnodeBytes=15728640
druid.metadata.storage.connector.createTables=false
druid.indexer.queue.startDelay=PT5S
druid.indexer.storage.type=metadata
And suggestion, idea or help is really appreciated
ThanksVenugopal Vupparaboina
02/05/2025, 8:43 AMjava.lang.OutOfMemoryError: GC overhead limit exceeded
exception while running an MSQ with broadcast joins:
org.apache.druid.java.util.common.ISE: worker sketch fetch failed
at org.apache.druid.msq.exec.ControllerImpl$RunQueryUntilDone.checkForErrorsInSketchFetcher(ControllerImpl.java:2747)
at org.apache.druid.msq.exec.ControllerImpl$RunQueryUntilDone.run(ControllerImpl.java:2730)
at org.apache.druid.msq.exec.ControllerImpl$RunQueryUntilDone.access$000(ControllerImpl.java:2681)
at org.apache.druid.msq.exec.ControllerImpl.runTask(ControllerImpl.java:433)
at org.apache.druid.msq.exec.ControllerImpl.run(ControllerImpl.java:372)
at org.apache.druid.msq.indexing.MSQControllerTask.runTask(MSQControllerTask.java:258)
at org.apache.druid.indexing.common.task.AbstractTask.run(AbstractTask.java:179)
at org.apache.druid.indexing.overlord.SingleTaskBackgroundRunner$SingleTaskBackgroundRunnerCallable.call(SingleTaskBackgroundRunner.java:478)
at org.apache.druid.indexing.overlord.SingleTaskBackgroundRunner$SingleTaskBackgroundRunnerCallable.call(SingleTaskBackgroundRunner.java:450)
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:131)
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:75)
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:82)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: org.apache.druid.msq.indexing.error.MSQException: WorkerRpcFailed: RPC call to task failed unrecoverably: [query-f0eee42a-a191-4b75-af72-a59254c00669-worker0_0]
at org.apache.druid.msq.exec.ExceptionWrappingWorkerClient$1.onFailure(ExceptionWrappingWorkerClient.java:160)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1119)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1286)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1055)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:807)
at com.google.common.util.concurrent.AbstractTransformFuture.run(AbstractTransformFuture.java:127)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1286)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1055)
at com.google.common.util.concurrent.AbstractFuture.set(AbstractFuture.java:782)
at com.google.common.util.concurrent.SettableFuture.set(SettableFuture.java:49)
at org.apache.druid.rpc.ServiceClientImpl$1.handleResultValue(ServiceClientImpl.java:277)
at org.apache.druid.rpc.ServiceClientImpl$1.onSuccess(ServiceClientImpl.java:190)
at org.apache.druid.rpc.ServiceClientImpl$1.onSuccess(ServiceClientImpl.java:183)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1133)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
... 3 more
Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded
at com.fasterxml.jackson.core.util.ByteArrayBuilder.toByteArray(ByteArrayBuilder.java:163)
at com.fasterxml.jackson.core.json.UTF8StreamJsonParser._decodeBase64(UTF8StreamJsonParser.java:3648)
at com.fasterxml.jackson.core.json.UTF8StreamJsonParser.getBinaryValue(UTF8StreamJsonParser.java:526)
at com.fasterxml.jackson.databind.deser.std.PrimitiveArrayDeserializers$ByteDeser.deserialize(PrimitiveArrayDeserializers.java:469)
at com.fasterxml.jackson.databind.deser.std.PrimitiveArrayDeserializers$ByteDeser.deserialize(PrimitiveArrayDeserializers.java:432)
at com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromString(BeanDeserializerBase.java:1488)
at com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeOther(BeanDeserializer.java:208)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:198)
at com.fasterxml.jackson.databind.deser.SettableBeanProperty.deserialize(SettableBeanProperty.java:542)
at com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeWithErrorWrapping(BeanDeserializer.java:566)
at com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased(BeanDeserializer.java:450)
at com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault(BeanDeserializerBase.java:1405)
Anyone else faced this outofmemory issue ?Mahesha Subrahamanya
02/05/2025, 4:54 PMAnimesh Gupta
02/06/2025, 4:26 AMSuraj Goel
02/10/2025, 2:16 PMReceived pause command, pausing ingestion until resumed.
Updating status of task [index_kafka_task] to [TaskStatus{id=index_kafka_task, status=FAILED, duration=-1, errorMsg=An exception occurred while waiting for task [index_kafka_task...}].
Shutdown [index_kafka_task] because: [An exception occurred while waiting for task [index_kafka_task] to pause: [org.apache.druid.rpc.HttpResponseException: Server error [409 Conflict]; body: Can't pause, task is not in a pausable state (state: [PAUSED])]]
Got shutdown request for task[index_kafka_task]. Asking worker[indexer] to kill it.
Stopping thread for task: index_kafka_task
Shutdown [index_kafka_task] because: [shut down request via HTTP endpoint]
Sent shutdown message to worker: indexer, status 200 OK, response: {"task":"index_kafka_task"}
Task [index_kafka_task] status changed to [SUCCESS].
Setting task[index_kafka_task] work item state from [RUNNING] to [COMPLETE].
Exception:
java.lang.RuntimeException: Stacktrace...
at org.apache.druid.indexing.overlord.hrtr.HttpRemoteTaskRunner$HttpRemoteTaskRunnerWorkItem.setStateUnconditionally(HttpRemoteTaskRunner.java:1913)
at org.apache.druid.indexing.overlord.hrtr.HttpRemoteTaskRunner$HttpRemoteTaskRunnerWorkItem.setState(HttpRemoteTaskRunner.java:1894)
at org.apache.druid.indexing.overlord.hrtr.HttpRemoteTaskRunner$HttpRemoteTaskRunnerWorkItem.setResult(HttpRemoteTaskRunner.java:1881)
at org.apache.druid.indexing.overlord.hrtr.HttpRemoteTaskRunner.taskComplete(HttpRemoteTaskRunner.java:506)
at org.apache.druid.indexing.overlord.hrtr.HttpRemoteTaskRunner.taskAddedOrUpdated(HttpRemoteTaskRunner.java:1695)
at org.apache.druid.indexing.overlord.hrtr.WorkerHolder$2.notifyListener(WorkerHolder.java:432)
at org.apache.druid.indexing.overlord.hrtr.WorkerHolder$2.deltaSync(WorkerHolder.java:425)
at org.apache.druid.server.coordination.ChangeRequestHttpSyncer$1.onSuccess(ChangeRequestHttpSyncer.java:293)
at org.apache.druid.server.coordination.ChangeRequestHttpSyncer$1.onSuccess(ChangeRequestHttpSyncer.java:259)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1133)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:840)
• Is this a known issue ?
• What can be the problem here ?
• Any recent change that is causing this as this issue was not there in Druid-25 ?Kiarash Norouzi
02/11/2025, 5:33 AMDinesh
02/12/2025, 7:01 PMDinesh
02/13/2025, 9:20 AMMateusz Kalinowski
02/13/2025, 3:58 PMreplaceMissingValueWith
?
Example configuration:
{
"type": "map",
"replaceMissingValueWith": "2025-01-01 00:00:00",
"map": {
"1": "1000-01-01 00:00:00",
"2": "1000-01-01 00:00:00"
}
}
I know there is a way to change the SQL to have the 3rd parameter as described here, but I want to change this behaviour just for the dev environment:
The LOOKUP function also accepts a third argument called replaceMissingValueWith as a constant string. If the lookup does not contain a value for the provided key, then the LOOKUP function returns this replaceMissingValueWith value rather than NULL, just like COALESCE. For example, LOOKUP(store, 'store_to_country', 'NA') is equivalent to COALESCE(LOOKUP(store, 'store_to_country'), 'NA').
Jimbo Slice
02/13/2025, 9:30 PMJulian Reyes
02/14/2025, 4:36 PM"transformSpec": {
"filter": {
"type": "not",
"field": {
"type": "in",
"dimension": "user_id",
"value": [
"1234"
]
}
},
"transforms": [
{
"type": "expression",
"name": "isSubRequest",
"expression": "if(\"batch_id\" > 0, 1, 0)"
}
]
}
however I am getting
Failed to submit supervisor: Cannot construct instance of `org.apache.druid.query.filter.InDimFilter`,
problem: values cannot be null at [Source: (org.eclipse.jetty.server.HttpInputOverHTTP); line: 1, column: 2038]
(through reference chain: org.apache.druid.indexing.kinesis.supervisor.KinesisSupervisorSpec["spec"]
->org.apache.druid.indexing.kinesis.supervisor.KinesisSupervisorIngestionSpec["dataSchema"]
->org.apache.druid.segment.indexing.DataSchema["transformSpec"]
->org.apache.druid.segment.transform.TransformSpec["filter"]
->org.apache.druid.query.filter.NotDimFilter["field"])
Not sure if I am missing something or the filter is bad configuredDinesh
02/17/2025, 4:40 AMCarlos M
02/17/2025, 5:51 PMorg.apache.druid.server.metrics.TaskSlotCountStatsMonitor
and org.apache.druid.server.metrics.TaskCountStatsMonitor
in Druid 30.x and above? in both cases the middle managers would refuse to start with Exception in thread "main" java.lang.RuntimeException: com.google.inject.CreationException: Unable to create injector, see the following errors
Sarang Vadali
02/17/2025, 11:32 PM{ "type": "aws-rds-token", "user": "USER", "host": "HOST", "port": PORT, "region": "AWS_REGION" }
What druid property should these values be passed into? An example for this would be very helpful!massimo zorer
02/18/2025, 6:26 PMdruid-protobuf-extensions
fails to flatten the messages if they contain repeated fields. My goal is to flatten the message if it contains repeated fields. So I cloned the git repo of the current extension and modified the parser, the reader, the InputFormat etc. to make it so that given a message that contains repeated fields the Cartesian product over the repeated fields is returned. Also test are fine in my new exstension.
In addition, I also want the fields that are not set in the message but are declared in the protobuf descriptor to be returned.
I loaded my extension into the compose docker but when I try to use it from the ui I get an error on `org.apache.druid.indexing.overlord.sampler.InputSourceSampler.Sample`:
Size of rawColumnsList([[{...}]]) does not correspond to size of inputRows([[{}]])
Do you kindly have any suggestions?Rushil
02/19/2025, 1:37 PMtania manhas
02/20/2025, 1:28 AMorg.apache.druid.server.metrics.QueryCountStatsMonitor
to capture query metrics in Druid. However, the logs indicate that query/count
remains at 0
, However, our queries are being processed. I did research online and it seems this was a bug in earlier versions which has been fixed. We are currently using verison 30.0.1. Can anyone please help me with this? Thank youCarlos M
02/21/2025, 6:56 PMdruid-operator
)? the <https://charts.datainfra>[.]io
seems to be pointing to some parking site:Dinesh
02/26/2025, 9:04 AMDinesh
02/26/2025, 9:07 AMHagen Rother
02/28/2025, 4:54 PMJimbo Slice
03/02/2025, 11:45 AMDinesh
03/03/2025, 7:03 AMChetan Patidar
03/03/2025, 8:32 AMKeith Byrd
03/03/2025, 2:45 PM