Harish Bohara
03/25/2023, 10:54 AMDavid Honghao Shao
03/26/2023, 9:57 PMYeongju Kang
03/27/2023, 2:47 AMgable heng
03/27/2023, 4:00 AMjava.lang.IllegalStateException: Forward index disabled column name must have a dictionary
i do see in the metadata.properties that this column does have hasDictionary = true
though, any idea of where else i can look at to fix this issue?Bharath
03/27/2023, 5:59 AMSid
03/27/2023, 9:41 AMDurgesh Choudhary
03/27/2023, 11:33 AMjavax.management.InstanceAlreadyExistsException: kafka.consumer:type=app-info,id=PartitionGroupMetadataFetcher-
Fizza Abid
03/27/2023, 1:01 PMKun
03/27/2023, 3:32 PM000000000000
) and offset (49611548018071132552656657372575220481442610790216499202
)
◦ Currently the kinesis stream has 2 shards 000000000001
AND 000000000002
◦ and the most recent offset is greater than that (49611548018071132552656657372575220481442610790216499202
)
• I tried LATEST
config for shardIterator But no luck.
◦ Also tried "stream.kinesis.consumer.prop.auto.offset.reset": "largest",
but It does not work too.
• Seems consumer is failure to get shard information
Here are few messages regard to the consumer.
Cannot fetch latest stream offset for clientId {KINESIS_STREAM_NAME}-0 and partitionGroupId 0 with maxWaitTime 5000
…
Starting consumption loop start offset {“shardId-000000000000”:“49611548018071132552656657372575220481442610790216499202"}, finalOffset null
…
Consumed 0 events from (rate:0.0/s), currentOffset={“shardId-000000000000”:“49611548018071132552656657372575220481442610790216499202"}, numRowsConsumedSoFar=0, numRowsIndexedSoFar=0Question Q1. Is there any way to reset kinesis stream offset? Q2. The problematic kinesis stream was created at 2020-09-01. Can’t I use a Kinesis Stream created a long time ago with Pinot? Is there anything I can refer? Thanks.
abhinav wagle
03/27/2023, 10:12 PMStream Error
io.grpc.netty.shaded.io.netty.handler.codec.http2.Http2Exception$StreamException: Stream closed before write could take place
at io.grpc.netty.shaded.io.netty.handler.codec.http2.Http2Exception.streamError(Http2Exception.java:167)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController$FlowState.cancel(DefaultHttp2RemoteFlowController.java:481)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController$1.onStreamClosed(DefaultHttp2RemoteFlowController.java:105)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2Connection.notifyClosed(DefaultHttp2Connection.java:356)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2Connection$ActiveStreams.removeFromActiveStreams(DefaultHttp2Connection.java:1000)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2Connection$ActiveStreams.deactivate(DefaultHttp2Connection.java:956)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2Connection$DefaultStream.close(DefaultHttp2Connection.java:512)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.DefaultHttp2Connection.close(DefaultHttp2Connection.java:152)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.Http2ConnectionHandler$BaseDecoder.channelInactive(Http2ConnectionHandler.java:209)
at io.grpc.netty.shaded.io.netty.handler.codec.http2.Http2ConnectionHandler.channelInactive(Http2ConnectionHandler.java:417)
at io.grpc.netty.shaded.io.grpc.netty.NettyServerHandler.channelInactive(NettyServerHandler.java:601)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:241)
at io.grpc.netty.shaded.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1405)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:262)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:248)
at io.grpc.netty.shaded.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:901)
at io.grpc.netty.shaded.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:819)
at io.grpc.netty.shaded.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
at io.grpc.netty.shaded.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
at io.grpc.netty.shaded.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:384)
at io.grpc.netty.shaded.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
at io.grpc.netty.shaded.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.grpc.netty.shaded.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:829)
Rajat Yadav
03/28/2023, 6:32 AMAsyncLogger error handling event seq=0, value='null': Exception in thread "HttpServer-0" java.lang.OutOfMemoryError: Java heap space
java.lang.OutOfMemoryError: Java heap space
Does Hi team, yesterday we got this error in our pinot-server while running query. Is there anyone configuration to clear heap space without restarting server and how can we configure in helm charts?Danko Andruszkiw
03/28/2023, 12:54 PMDanko Andruszkiw
03/28/2023, 12:55 PMDanko Andruszkiw
03/28/2023, 12:56 PM<http://my-server-ptctlr01:9000/segments/location_density> -> /data/pinot/data/controller/location_density
-rw-r--r--. 1 pinot pinot 4098 Feb 13 13:05 location_density__0__111__20230212T1305Z
[root@my-server-ptctlr02 location_density]$ ll location_density__0__111__20230212T1305Z
-rw-r--r--. 1 pinot pinot 4098 Feb 13 13:05 location_density__0__111__20230212T1305Z
[root@my-server-ptsvr01 location_density_REALTIME]$ pwd
/data/pinot/data/index/location_density_REALTIME
[root@my-server-ptsvr01 location_density_REALTIME]$ ls -l location_density__0__111__20230212T1305Z/*
-rw-r--r--. 1 pinot pinot 0 Mar 27 09:33 location_density__0__111__20230212T1305Z/location_timestamp.bloom.inprogress
location_density__0__111__20230212T1305Z/v3:
total 20
-rw-r--r--. 1 pinot pinot 4534 Mar 27 09:33 columns.psf
-rw-r--r--. 1 pinot pinot 16 Mar 27 09:33 creation.meta
-rw-r--r--. 1 pinot pinot 1260 Mar 27 09:33 index_map
-rw-r--r--. 1 pinot pinot 3906 Mar 27 09:33 metadata.properties
[root@my-server-ptsvr02 location_density_REALTIME]$ ls -l location_density__0__111__20230212T1305Z/*
-rw-r--r--. 1 pinot pinot 0 Mar 27 08:41 location_density__0__111__20230212T1305Z/location_timestamp.bloom.inprogress
location_density__0__111__20230212T1305Z/v3:
total 20
-rw-r--r--. 1 pinot pinot 4534 Mar 27 08:41 columns.psf
-rw-r--r--. 1 pinot pinot 16 Mar 27 08:41 creation.meta
-rw-r--r--. 1 pinot pinot 1260 Mar 27 08:41 index_map
-rw-r--r--. 1 pinot pinot 3906 Mar 27 08:41 metadata.properties
[root@my-server-ptsvr03 location_density_REALTIME]$ ls -l location_density__0__111__20230212T1305Z/*
-rw-r--r--. 1 pinot pinot 0 Mar 27 09:32 location_density__0__111__20230212T1305Z/location_timestamp.bloom.inprogress
location_density__0__111__20230212T1305Z/v3:
total 20
-rw-r--r--. 1 pinot pinot 4534 Mar 27 09:32 columns.psf
-rw-r--r--. 1 pinot pinot 16 Mar 27 09:32 creation.meta
-rw-r--r--. 1 pinot pinot 1260 Mar 27 09:32 index_map
-rw-r--r--. 1 pinot pinot 3906 Mar 27 09:32 metadata.properties
Dugi Sarma
03/28/2023, 8:42 PMfilterConfig
but not quite sure why its not really working out. Attaching the schema and filter function.Damon
03/28/2023, 8:43 PMDamon
03/28/2023, 8:53 PManukin
03/28/2023, 10:35 PMUploadSegment
and every time I upload a new segment, instead of adding to the existing segments, it replaces it with what I provide it. Is there a way I can specify it to append to existing segments. I see an option existing in batched ingestions but cannot find one for ad hoc processes like this.
This is the exact command bin/pinot-admin.sh UploadSegment -controllerHost localhost -controllerPort 9000 -segmentDir ./outputDir
Thank you in advance for helping out with this noob problem. 🙂Pranathi A
03/29/2023, 1:08 PMController response {"streamPartitionMsgOffset":"100000","buildTimeSec":126,"isSplitCommitType":true,"controllerVipUrl":"<http://pinot-na1-controller-1.pinot-na1-controller-headless.revenue-reporting.svc.cluster.local:9000>","statu │
│ Sending request: <http://pinot-na1-controller-1.pinot-na1-controller-headless.revenue-reporting.svc.cluster.local:9000/segmentCommitStart?segmentSizeBytes=85637857&buildTimeMillis=3509&streamPartitionMsgOffset=100000&instance=Serv> │
│ Controller response {"streamPartitionMsgOffset":null,"buildTimeSec":-1,"isSplitCommitType":false,"status":"COMMIT_CONTINUE","offset":-1} for <http://pinot-na1-controller-1.pinot-na1-controller-headless.revenue-reporting.svc.cluste> │
│ Sending request: <http://pinot-na1-controller-1.pinot-na1-controller-headless.revenue-reporting.svc.cluster.local:9000/segmentUpload?segmentSizeBytes=85637857&buildTimeMillis=3509&streamPartitionMsgOffset=100000&instance=Server_pi> │
│ Controller response {"streamPartitionMsgOffset":null,"buildTimeSec":-1,"isSplitCommitType":false,"status":"FAILED","offset":-1} for <http://pinot-na1-controller-1.pinot-na1-controller-headless.revenue-reporting.svc.cluster.local:9> │
│ Could not commit segment. Retrying after hold
Steven Hall
03/29/2023, 7:27 PMabhinav wagle
03/30/2023, 1:39 AM2023-03-30T01:36:32.639Z ERROR SplitRunner-138-61 io.trino.execution.executor.TaskExecutor Error processing Split 20230330_013630_00009_paj2g.4.0.0-9 PinotSplit{splitType=SEGMENT, segments=[<redacted>__8__2333__20230328T2149Z], segmentHost=Optional[<redacted>]} (start = 3.1565860028796E7, wall = 988 ms, cpu = 0 ms, wait = 0 ms, calls = 1)
io.grpc.StatusRuntimeException: UNAVAILABLE: io exception
at io.grpc.Status.asRuntimeException(Status.java:535)
at io.grpc.stub.ClientCalls$BlockingResponseStream.hasNext(ClientCalls.java:648)
at io.trino.plugin.pinot.client.PinotGrpcDataFetcher$PinotGrpcServerQueryClient$ResponseIterator.computeNext(PinotGrpcDataFetcher.java:274)
at io.trino.plugin.pinot.client.PinotGrpcDataFetcher$PinotGrpcServerQueryClient$ResponseIterator.computeNext(PinotGrpcDataFetcher.java:261)
at com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:146)
at com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:141)
at io.trino.plugin.pinot.client.PinotGrpcDataFetcher.endOfData(PinotGrpcDataFetcher.java:85)
at io.trino.plugin.pinot.PinotSegmentPageSource.getNextPage(PinotSegmentPageSource.java:124)
at io.trino.operator.ScanFilterAndProjectOperator$ConnectorPageSourceToPages.process(ScanFilterAndProjectOperator.java:400)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils.getNextState(WorkProcessorUtils.java:240)
at io.trino.operator.WorkProcessorUtils$YieldingProcess.process(WorkProcessorUtils.java:182)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:325)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:325)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:325)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:325)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils.lambda$flatten$7(WorkProcessorUtils.java:296)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:338)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils$3.process(WorkProcessorUtils.java:325)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils.getNextState(WorkProcessorUtils.java:240)
at io.trino.operator.WorkProcessorUtils.lambda$processStateMonitor$3(WorkProcessorUtils.java:219)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorUtils.getNextState(WorkProcessorUtils.java:240)
at io.trino.operator.WorkProcessorUtils.lambda$finishWhen$4(WorkProcessorUtils.java:234)
at io.trino.operator.WorkProcessorUtils$ProcessWorkProcessor.process(WorkProcessorUtils.java:391)
at io.trino.operator.WorkProcessorSourceOperatorAdapter.getOutput(WorkProcessorSourceOperatorAdapter.java:150)
at io.trino.operator.Driver.processInternal(Driver.java:411)
at io.trino.operator.Driver.lambda$process$10(Driver.java:314)
at io.trino.operator.Driver.tryWithLock(Driver.java:706)
at io.trino.operator.Driver.process(Driver.java:306)
at io.trino.operator.Driver.processForDuration(Driver.java:277)
at io.trino.execution.SqlTaskExecution$DriverSplitRunner.processFor(SqlTaskExecution.java:736)
at io.trino.execution.executor.PrioritizedSplitRunner.process(PrioritizedSplitRunner.java:164)
at io.trino.execution.executor.TaskExecutor$TaskRunner.run(TaskExecutor.java:515)
at io.trino.$gen.Trino_400____20230330_004502_2.run(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: io.grpc.netty.shaded.io.netty.channel.unix.Errors$NativeIoException: readAddress(..) failed: Connection reset by peer
Prashant Korade
03/30/2023, 3:56 AM{
"positionid" : 123,
"securityId" : 345,
"Results" :
{
delta : [123, 456, 789],
gamma : [0.0, 1.2, 4.5]
}
}
2. How can we achieve ARRAY JOIN feature provided in clickhouse in pinot.Lee Wei Hern Jason
03/30/2023, 5:39 AM[
{
"errorCode": 200,
"message": "QueryExecutionError:\nio.grpc.StatusRuntimeException: UNAVAILABLE: io exception\n\tat io.grpc.stub.ClientCalls.toStatusRuntimeException(ClientCalls.java:262)\n\tat io.grpc.stub.ClientCalls.getUnchecked(ClientCalls.java:243)\n\tat io.grpc.stub.ClientCalls.blockingUnaryCall(ClientCalls.java:156)\n\tat org.apache.pinot.common.proto.PinotQueryWorkerGrpc$PinotQueryWorkerBlockingStub.submit(PinotQueryWorkerGrpc.java:166)\n...\nCaused by: io.grpc.netty.shaded.io.netty.channel.ConnectTimeoutException: connection timed out: ip-10-110-219-144.ap-southeast-1.compute.internal/10.110.219.144:8421\n\tat io.grpc.netty.shaded.io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:261)\n\tat io.grpc.netty.shaded.io.netty.util.concurrent.PromiseTask.runTask(PromiseTask.java:98)\n\tat io.grpc.netty.shaded.io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:170)\n\tat io.grpc.netty.shaded.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)"
}
]
Shreeram Goyal
03/31/2023, 4:33 AMselect * from table where id=xyz;
Result: No record(s) found.
This doesn't happen for a PK which resides in an online segment. Now when I remove the segmentPrunerType= partition, I start getting result for the consuming segment as well. Can somebody enlighten why am I facing this with the consuming segments only?Lvszn Peng
03/31/2023, 7:10 AMPhil Sheets
04/01/2023, 11:31 PMErich
04/02/2023, 2:33 PMmaster
branch) and I'm getting a MojoFailureException
when trying to build pinot-broker
(Error details below):
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.0:testCompile (default-testCompile) on project pinot-broker: Compilation failure -> [Help 1]
org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.0:testCompile (default-testCompile) on project pinot-broker: Compilation failure
at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:306)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:211)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:165)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:157)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:121)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56)
at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:127)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:294)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192)
at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105)
at org.apache.maven.cli.MavenCli.execute (MavenCli.java:960)
at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:293)
at org.apache.maven.cli.MavenCli.main (MavenCli.java:196)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke (Method.java:568)
at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:282)
at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:225)
at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:406)
at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:347)
Caused by: org.apache.maven.plugin.compiler.CompilationFailureException: Compilation failure
at org.apache.maven.plugin.compiler.AbstractCompilerMojo.execute (AbstractCompilerMojo.java:1219)
at org.apache.maven.plugin.compiler.TestCompilerMojo.execute (TestCompilerMojo.java:181)
at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)
at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:301)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:211)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:165)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:157)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:121)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56)
at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:127)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:294)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192)
at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105)
at org.apache.maven.cli.MavenCli.execute (MavenCli.java:960)
at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:293)
at org.apache.maven.cli.MavenCli.main (MavenCli.java:196)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke (Method.java:568)
at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:282)
at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:225)
at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:406)
at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:347)
[ERROR]
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] <http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException>
My current Java environment is:
➜ pinot git:(master) javac -version
javac 17.0.6
➜ pinot git:(master) java -version
openjdk version "17.0.6" 2023-01-17
OpenJDK Runtime Environment (Red_Hat-17.0.6.0.10-1.fc37) (build 17.0.6+10)
OpenJDK 64-Bit Server VM (Red_Hat-17.0.6.0.10-1.fc37) (build 17.0.6+10, mixed mode, sharing)
Tanmay Varun
04/02/2023, 7:39 PMapache-pinot-0.9.0-bin/bin/pinot-admin.sh LaunchDataIngestionJob -jobSpecFile jobspec.yml
i specified some outputURI in spec that it is not able to create.. is any extra param required in this class or is that output Dir supposed to be specified at time of cluster creation..
error
Caused by: java.io.IOException: Unable to create directory /var/pinot/server/data/segment
at org.apache.commons.io.FileUtils.forceMkdir(FileUtils.java:2384)
at org.apache.pinot.spi.filesystem.LocalPinotFS.mkdir(LocalPinotFS.java:50)
at org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner.init(SegmentGenerationJobRunner.java:126)
Tanmay Varun
04/02/2023, 7:39 PMexecutionFrameworkSpec:
name: 'standalone'
segmentGenerationJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner'
segmentTarPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentTarPushJobRunner'
segmentUriPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentUriPushJobRunner'
jobType: SegmentCreationAndTarPush
inputDirURI: '.'
includeFileNamePattern: 'glob:**/export_json.json'
outputDirURI: '/var/pinot/server/data/segment'
overwriteOutput: true
pinotFSSpecs:
- scheme: hdfs
className: org.apache.pinot.plugin.filesystem.HadoopPinotFS
- scheme: file
className: org.apache.pinot.spi.filesystem.LocalPinotFS
recordReaderSpec:
dataFormat: 'json'
className: 'org.apache.pinot.plugin.inputformat.json.JSONRecordReader'
configClassName: 'org.apache.pinot.plugin.inputformat.json.JSONRecordReaderConfig'
tableSpec:
tableName: 'reporting_benchmark_REALTIME'
pinotClusterSpecs:
- controllerURI: '<http://localhost:9000>'