yelim yu
10/21/2021, 3:15 PMPhilippe Dooze
10/22/2021, 1:23 PMGrant Sherrick
10/22/2021, 3:09 PMCharles
10/25/2021, 3:03 AMSadim Nadeem
10/25/2021, 5:11 AMTiger Zhao
10/25/2021, 8:33 PMjava.lang.IllegalStateException: PinotFS for scheme: s3 has not been initialized
. Is there a way to configure the minions to be able to read from S3? I couldn't find anything in the docs. Thanks!bc Wong
10/25/2021, 10:23 PMAddTable
. Loaded data from Oct 1 via ImportData
.
2. Query select count(1) from tbl where ds = '2021-10-01'
ran successfully.
3. Added REALTIME table via web ui. Kafka ingested a bunch of data for ds = '2021-10-03'
. Query shows new data.
4. But the query from #2 now returns no row. I have to query against tbl_OFFLINE
to see the offline records.
Many thanks!Girish Patel
10/26/2021, 10:21 AMTamás Nádudvari
10/26/2021, 1:33 PMDiogo Baeder
10/26/2021, 2:00 PMsuraj kamath
10/27/2021, 6:18 AM"message": "QueryExecutionError:\norg.apache.pinot.spi.exception.BadQueryRequestException: Caught exception while initializing transform function: lookup\n\tat org.apache.pinot.core.operator.transform.function.TransformFunctionFactory.get(TransformFunctionFactory.java:207)\n\tat org.apache.pinot.core.operator.transform.TransformOperator.<init>(TransformOperator.java:56)\n\tat org.apache.pinot.core.plan.TransformPlanNode.run(TransformPlanNode.java:56)\n\tat org.apache.pinot.core.plan.SelectionPlanNode.run(SelectionPlanNode.java:83)\n\tat org.apache.pinot.core.plan.CombinePlanNode.run(CombinePlanNode.java:100)\n\tat org.apache.pinot.core.plan.InstanceResponsePlanNode.run(InstanceResponsePlanNode.java:33)\n\tat org.apache.pinot.core.plan.GlobalPlanImplV0.execute(GlobalPlanImplV0.java:45)\n\tat org.apache.pinot.core.query.executor.ServerQueryExecutorV1Impl.processQuery(ServerQueryExecutorV1Impl.java:296)\n\tat org.apache.pinot.core.query.executor.ServerQueryExecutorV1Impl.processQuery(ServerQueryExecutorV1Impl.java:216)\n\tat org.apache.pinot.core.query.executor.QueryExecutor.processQuery(QueryExecutor.java:60)\n\tat org.apache.pinot.core.query.scheduler.QueryScheduler.processQueryAndSerialize(QueryScheduler.java:155)\n\tat org.apache.pinot.core.query.scheduler.QueryScheduler.lambda$createQueryFutureTask$0(QueryScheduler.java:139)\n\tat java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)\n\tat java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)",
"errorCode": 200
},
From the error message, it's not very clear to me what's the issue here. Can anyone help ?Mahesh babu
10/27/2021, 6:31 AMsuraj kamath
10/27/2021, 9:22 AMSELECT column1
,column2
,lookup('dimTable', 'username', 'orgId', orgId, 'userId', userId) AS username
FROM tableA
WHERE column2 IN ('Good')
AND column1 != 'Unknown'
AND username = 'user'
limit 100
This returns me the result,
However, the moment I add either a count(*) in the select statement or a "group by" the query keeps on loading.
Eg:
SELECT column1
,column2
,orgId
,lookup('dimTable', 'username', 'orgId', orgId, 'userId', userId) AS username
FROM tableA
WHERE column2 IN ('Good')
AND column1 != 'Unknown'
AND username = 'user'
GROUP BY column1, column2, orgId, username
limit 100
Lars-Kristian Svenøy
10/27/2021, 10:00 AMLuis Fernandez
10/27/2021, 7:13 PMAbhishek Saini
10/28/2021, 5:59 PMhardik
10/29/2021, 6:36 AMYeongju Kang
10/29/2021, 7:31 AMSadim Nadeem
10/29/2021, 3:20 PMElon
10/29/2021, 9:00 PMTony Requist
10/30/2021, 1:08 AM"realtime.segment.flush.threshold.rows": "0",
"realtime.segment.flush.threshold.time": "4h",
"realtime.segment.flush.threshold.segment.size": "40M",
and I am storing segment files in S3. I am seeing a huge number of files in S3 like
TABLE__0__0__20211029T1835Z.tmp.420158c9-1742-4bd2-bbae-5a59d2205cd2
that are all much smaller than 40M. There are several thousand files. What are these?Sadim Nadeem
10/31/2021, 4:29 AMkubectl delete pods <pod>
Sandeep R
10/31/2021, 11:30 PM), currentOffset=1939499, numRowsConsumedSoFar=1, numRowsIndexedSoFar=1
2021/10/31 23:08:26.026 ERROR [LLRealtimeSegmentDataManager_pnrevents__0__0__20211031T2306Z] [pnrevents__0__0__20211031T2306Z] Caught exception while transforming the record: {
"fieldToValueMap" : {
"pt" : null,
"osia" : null,
"excludePII" : 0,
"pcc" : null,
"lname" : null,
"mel" : null,
"tkt" : null,
"kafkaProps" : null,
"timestamp" : null,
"ver" : null,
"dts" : null,
"proxyUrl" : "<http://xyz.com:8080|xyz.com:8080>",
"docid" : null,
"rlc" : null,
"rcode" : null,
"message" : {
"jsver" : "1",
"core" : "1g",
"pcc" : "ABCD",
"notif" : "abc",
"ver" : "5",
"dts" : "20211022",
"lname" : [ "JONES" ],
"pt" : "2021-10-22T02:22:48.196",
"mel" : [ "MULTIPAX.V6@XYZ.COM" ],
"docid" : "1a0343cdrdc455",
"rlc" : "1234F"
},
"itina" : null,
"gname" : null,
"url" : "<https://xyz.com/pir/abc>",
"agencyName" : "MTT",
"securityToken" : "xxxxxxxxxxxxx",
"jsver" : null,
"core" : null,
"notif" : null,
"nama" : null,
"npnr" : null,
"kafka" : null,
"emd" : null,
"phonea" : null
},
"nullValueFields" : [ ]
}
java.lang.RuntimeException: Caught exception while transforming data type for column: message
at org.apache.pinot.segment.local.recordtransformer.DataTypeTransformer.transform(DataTypeTransformer.java:120) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.recordtransformer.CompositeTransformer.transform(CompositeTransformer.java:82) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager.processStreamEvents(LLRealtimeSegmentDataManager.java:510) [pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager.consumeLoop(LLRealtimeSegmentDataManager.java:417) [pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager$PartitionConsumer.run(LLRealtimeSegmentDataManager.java:560) [pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at java.lang.Thread.run(Thread.java:834) [?:?]
Caused by: java.lang.IllegalStateException: Cannot read single-value from Collection: [1, 1g, ABCD, abc, 5, 20211022, [Ljava.lang.Object;@4049e608, 2021-10-22T02:22:48.196, [Ljava.lang.Object;@3eb034ea, 1, 1234F] for column: message
at shaded.com.google.common.base.Preconditions.checkState(Preconditions.java:721) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.recordtransformer.DataTypeTransformer.standardizeCollection(DataTypeTransformer.java:199) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.recordtransformer.DataTypeTransformer.standardize(DataTypeTransformer.java:144) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.recordtransformer.DataTypeTransformer.transform(DataTypeTransformer.java:90) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
Kamal Chavda
11/01/2021, 6:38 PMMap
11/02/2021, 3:59 PMreplication
and replicasPerPartition
? Currently we only have replicasPerPartition
set 2 but in the segment builds, their configs still show num of replicas as 1Stuart Coleman
11/02/2021, 8:55 PMbin/pinot-admin.sh AvroSchemaToPinotSchema -timeColumnName fields.hoursSinceEpoch -avroSchemaFile /tmp/test.avsc -pinotSchemaName myTable -outputDir /tmp/test -fieldsToUnnest entries
with the schema in the pr (https://github.com/yupeng9/incubator-pinot/blob/660a70831cf0f7fc5a63c2f2c902c9c1f9[…]pinot-avro-base/src/test/resources/fake_avro_nested_schema.avsc) i get an exception below - any idea what i am doing wrong?
Exception caught:
java.lang.RuntimeException: Caught exception while extracting data type from field: entries
at org.apache.pinot.plugin.inputformat.avro.AvroUtils.extractFieldDataType(AvroUtils.java:252) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.inputformat.avro.AvroUtils.getPinotSchemaFromAvroSchema(AvroUtils.java:69) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.inputformat.avro.AvroUtils.getPinotSchemaFromAvroSchemaFile(AvroUtils.java:148) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.tools.admin.command.AvroSchemaToPinotSchema.execute(AvroSchemaToPinotSchema.java:99) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.tools.admin.PinotAdministrator.execute(PinotAdministrator.java:166) [pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.tools.admin.PinotAdministrator.main(PinotAdministrator.java:186) [pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
Caused by: java.lang.IllegalStateException: Not one field in the RECORD schema
at shaded.com.google.common.base.Preconditions.checkState(Preconditions.java:444) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.inputformat.avro.AvroUtils.extractSupportedSchema(AvroUtils.java:280) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.inputformat.avro.AvroUtils.extractFieldDataType(AvroUtils.java:247) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
... 5 more
Anish Nair
11/03/2021, 11:15 AM2021/11/03 08:29:22.109 INFO [SegmentFetcherFactory] [HelixTaskExecutor-message_handle_thread] Segment fetcher is not configured for protocol: s3, using default
2021/11/03 08:29:22.109 WARN [PinotFSSegmentFetcher] [HelixTaskExecutor-message_handle_thread] Caught exception while fetching segment from: <s3://pinot-db/pinot-ingestion/mytable/mytable_OFFLINE_2021091800_2021091800_0.tar.gz> to: /tmp/data/pinotSegments/mytable_OFFLINE/tmp-mytable_OFFLINE_2021091800_2021091800_0-90b8d75e-b2e8-4e4f-b115-36e5528c37cf/mytable_OFFLINE_2021091800_2021091800_0.enc
java.lang.IllegalStateException: PinotFS for scheme: s3 has not been initialized
at shaded.com.google.common.base.Preconditions.checkState(Preconditions.java:518) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.spi.filesystem.PinotFSFactory.create(PinotFSFactory.java:78) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
Following are our conf:
Server conf:
pinot.server.instance.enable.split.commit=true
pinot.server.storage.factory.class.s3=org.apache.pinot.plugin.filesystem.S3PinotFS
pinot.server.storage.factory.s3.region=us-east-1
pinot.server.segment.fetcher.protocols=s3
pinot.server.segment.fetcher.s3.class=org.apache.pinot.common.utils.fetcher.PinotFSSegmentFetcher
Controller conf:
controller.data.dir=s3://pinot-db/
controller.local.temp.dir=/tmp/pinot/
controller.enable.split.commit=true
pinot.controller.storage.factory.class.s3=org.apache.pinot.plugin.filesystem.S3PinotFS
pinot.controller.storage.factory.s3.region=us-east-1
pinot.controller.segment.fetcher.protocols=file,http,s3
pinot.controller.segment.fetcher.s3.class=org.apache.pinot.common.utils.fetcher.PinotFSSegmentFetcherLuis Fernandez
11/03/2021, 3:36 PMMap
11/04/2021, 5:54 PMtable1
with fields X
in upsert
mode. When a new field Y
is added to the schema, a simple query
select * from table1 limit 10
in the Pinot explorer will return the following error:
[
{
"message": "MergeResponseError:\nData schema mismatch between merged block: [X(DOUBLE)] and block to merge: [X(DOUBLE),Y(DOUBLE)], drop block to merge",
"errorCode": 500
}
]
However, the following query would work as expected
select * from table1 limit 10 option (skipUpsert=True)
Has anyone seen this before?Yeongju Kang
11/05/2021, 7:23 AM