Facundo Bianco
05/11/2022, 4:11 PMpushJobSpec:
pushParallelism: 20
pushAttempts: 2
segmentUriPrefix: "<s3://bucket-foo>"
segmentUriSuffix : ""
And got this error message:
2022/05/11 142850.531 ERROR [BaseTableDataManager] [HelixTaskExecutor-message_handle_thread] Attempts exceeded when downloading segment: foo_OFFLINE_2022-05-03_2022-05-03_11 for table: foo_OFFLINE from: s3://bucket-foo/data/output/foo_OFFLINE_2022-05-03_2022-05-03_11.tar.gznull to: /tmp/pinot-tmp/server/index/foo_OFFLINE/tmp/tmp-foo_OFFLINE_2022-05-03_2022-05-03_11-b2f3a97c-9c14-4b4c-9874-fb028597a237/foo_OFFLINE_2022-05-03_2022-05-03_11.tar.gz ava:72) [pinot-all-0.10.0-SNAPSHOT-jar-with-dependencies.jar:0.10.0-SNAPSHOT-fIt adds 'null' at end of file's URI. Any idea how to resolve? Thanks in advance.
Saumya Upadhyay
05/12/2022, 4:00 AMSaumya Upadhyay
05/12/2022, 6:54 AMLuy
05/12/2022, 10:33 AMdocker exec -it manual-pinot-controller bin/pinot-admin.sh LaunchDataIngestionJob -jobSpecFile /data/docker-job-spec.yml -exec
But it gives an error.
2022/05/12 09:51:12.633 ERROR [PinotAdministrator] [main] Exception caught:
picocli.CommandLine$UnmatchedArgumentException: Unknown option: '-exec'
Anyone can help me with this?Saumya Upadhyay
05/12/2022, 11:19 AM{
"header": {
"nnTransId": "9003",
"qid": 1,
"timestamp": 1234567890123
},
"status": "N200_SUCCESS"
}
"ingestionConfig": {
"transformConfigs": [
{
"columnName": "header_js",
"transformFunction": "jsonFormat(header)"
},
{
"columnName": "header_nnTransId",
"transformFunction": "JSONPATHSTRING(header, '$.nnTransId')"
} .....
Luis Fernandez
05/12/2022, 2:12 PMAli Atıl
05/12/2022, 2:16 PMStuart Millholland
05/12/2022, 7:49 PMDeepak Mishra
05/13/2022, 3:33 AMSaumya Upadhyay
05/13/2022, 12:37 PMProcessed requestId=6212,table=tSCalibrationAttempt_REALTIME,segments(queried/processed/matched/consuming)=4/2/0/2,schedulerWaitMs=1,reqDeserMs=0,totalExecMs=0,resSerMs=0,totalTimeMs=1,minConsumingFreshnessMs=1652072636087,broker=Broker_pinot-broker-0.pinot-broker-headless.pinot.svc.cluster.local_8099,numDocsScanned=0,scanInFilter=0,scanPostFilter=0,sched=FCFS,threadCpuTimeNs(total/thread/sysActivity/resSer)=0/0/0/0
Map
05/13/2022, 7:03 PM"dateTimeFieldSpecs": [
{
"name": "MSGDATETIME",
"dataType": "STRING",
"format": "1:MILLISECONDS:SIMPLE_DATE_FORMAT:yyyy-MM-dd'T'HH:mm:ss.fff'Z'",
"granularity": "1:MILLISECONDS"
}
Pinot reports that],
invalid datetime format: 1MILLISECONDSSIMPLE_DATE_FORMATyyyy MM dd’T’HHmm:ss.fff’Z’”What would be the right way to write this schema?
Seunghyun
05/14/2022, 12:13 AMnet.jpountz.lz4.LZ4Exception: Malformed input at 13
at net.jpountz.lz4.LZ4JavaUnsafeSafeDecompressor.decompress(LZ4JavaUnsafeSafeDecompressor.java:180) ~[lz4-java-1.7.1.jar:?]
at net.jpountz.lz4.LZ4SafeDecompressor.decompress(LZ4SafeDecompressor.java:145) ~[lz4-java-1.7.1.jar:?]
at org.apache.pinot.segment.local.io.compression.LZ4Decompressor.decompress(LZ4Decompressor.java:42) ~[pinot-segment-local-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f0
0e324e2dc]
at org.apache.pinot.segment.local.segment.index.readers.forward.BaseChunkSVForwardIndexReader.decompressChunk(BaseChunkSVForwardIndexReader.java:137) ~[pinot-segment-local-0.10.0-dev-471.j
ar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.segment.local.segment.index.readers.forward.BaseChunkSVForwardIndexReader.getChunkBuffer(BaseChunkSVForwardIndexReader.java:118) ~[pinot-segment-local-0.10.0-dev-471.ja
r:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.segment.local.segment.index.readers.forward.VarByteChunkSVForwardIndexReader.getStringCompressed(VarByteChunkSVForwardIndexReader.java:72) ~[pinot-segment-local-0.10.0-
dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.segment.local.segment.index.readers.forward.VarByteChunkSVForwardIndexReader.getString(VarByteChunkSVForwardIndexReader.java:61) ~[pinot-segment-local-0.10.0-dev-471.ja
r:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.segment.local.segment.index.readers.forward.VarByteChunkSVForwardIndexReader.getString(VarByteChunkSVForwardIndexReader.java:35) ~[pinot-segment-local-0.10.0-dev-471.ja
r:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.dociditerators.SVScanDocIdIterator$StringMatcher.doesValueMatch(SVScanDocIdIterator.java:176) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4
bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.dociditerators.SVScanDocIdIterator.applyAnd(SVScanDocIdIterator.java:88) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e3
24e2dc]
at org.apache.pinot.core.operator.docidsets.AndDocIdSet.iterator(AndDocIdSet.java:128) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.DocIdSetOperator.getNextBlock(DocIdSetOperator.java:67) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.DocIdSetOperator.getNextBlock(DocIdSetOperator.java:38) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.BaseOperator.nextBlock(BaseOperator.java:49) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.ProjectionOperator.getNextBlock(ProjectionOperator.java:61) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.ProjectionOperator.getNextBlock(ProjectionOperator.java:33) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.BaseOperator.nextBlock(BaseOperator.java:49) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.transform.PassThroughTransformOperator.getNextBlock(PassThroughTransformOperator.java:48) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3f
ecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.transform.PassThroughTransformOperator.getNextBlock(PassThroughTransformOperator.java:31) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3f
ecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.BaseOperator.nextBlock(BaseOperator.java:49) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.query.AggregationGroupByOrderByOperator.getNextBlock(AggregationGroupByOrderByOperator.java:107) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.query.AggregationGroupByOrderByOperator.getNextBlock(AggregationGroupByOrderByOperator.java:46) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.BaseOperator.nextBlock(BaseOperator.java:49) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.combine.GroupByOrderByCombineOperator.processSegments(GroupByOrderByCombineOperator.java:137) ~[pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.operator.combine.BaseCombineOperator$1.runJob(BaseCombineOperator.java:100) [pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at org.apache.pinot.core.util.trace.TraceRunnable.run(TraceRunnable.java:40) [pinot-core-0.10.0-dev-471.jar:0.10.0-dev-471-91c2ebbf297c4bf3fecb5f98413e9f00e324e2dc]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) [?:?]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) [guava-30.1.1-jre.jar:?]
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69) [guava-30.1.1-jre.jar:?]
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) [guava-30.1.1-jre.jar:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
Chengxuan Wang
05/16/2022, 3:35 AMRangeIndexBasedFilterOperator
is used.
"rangeIndexColumns": [
"some_column"
],
btw, we are using pinot 0.8Xiang Fu
Alice
05/16/2022, 8:25 AMDan DC
05/16/2022, 9:14 AMAnish Nair
05/16/2022, 10:31 AM{
"schemaName": "test_dim_tags",
"dimensionFieldSpecs": [
{
"name": "id",
"dataType": "INT"
},
{
"name": "tag_name",
"dataType": "STRING",
"singleValueField": false
}
],
"primaryKeyColumns": [
"id"
]
}
Now when we use this table in lookup with Fact Table, query is returning no data or throwing NullPointerExpection.
We wanted to use pinot's array explode functionality along with lookup.
can someone please help to understand?Stuart Millholland
05/16/2022, 6:54 PMAli Atıl
05/17/2022, 11:03 AMFizza Abid
05/17/2022, 12:01 PMLars-Kristian Svenøy
05/17/2022, 4:12 PMLars-Kristian Svenøy
05/18/2022, 10:44 AMLuy
05/18/2022, 11:41 AM# Please put the mock data source as the first in this configuration.
dataSourceConfigs:
- className: org.apache.pinot.thirdeye.datasource.pinot.PinotThirdEyeDataSource
properties:
zookeeperUrl: 'localhost:2181'
clusterName: 'PinotCluster'
controllerConnectionScheme: 'http'
controllerHost: '127.0.0.1'
controllerPort: 9000
cacheLoaderClassName: org.apache.pinot.thirdeye.datasource.pinot.PinotControllerResponseCacheLoader
metadataSourceConfigs:
- className: org.apache.pinot.thirdeye.auto.onboard.AutoOnboardPinotMetadataSource
And I got an error below.
2022-05-17 18:26:33.104 [main] INFO org.apache.pinot.thirdeye.datalayer.util.DaoProviderUtil - Using existing database at 'jdbc:mysql:///thirdeye?autoReconnect=true'
May 17, 2022 6:26:37 PM org.apache.tomcat.jdbc.pool.ConnectionPool init
SEVERE: Unable to create initial connections of pool.
com.mysql.jdbc.exceptions.jdbc4.MySQLNonTransientConnectionException: Could not create connection to database server. Attempted reconnect 3 times. Giving up.
I asked about this in other channels, but didn't get any answer.
Any guidance will be really helpful.Kriti
05/18/2022, 2:35 PMFailed to start a Pinot [CONTROLLER] at 7.722 since launch
java.lang.NullPointerException: Cannot invoke "java.lang.reflect.Method.invoke(Object, Object[])" because "com.sun.xml.bind.v2.runtime.reflect.opt.Injector.defineClass" is null
Peter Pringle
05/19/2022, 3:56 AMIlleglArgumentExceptionL watermarksMs in RealtimeToOfflineSegmentsTask metadata {"tableNameWithType":"myTable_REALTIME", "watermarksMs":16175808000000} does not match windowStartMs: %d in task configs for table: ....
Peter Pringle
05/19/2022, 3:57 AMAbhijeet Kushe
05/19/2022, 6:02 PMoption (skipUpsert=true)
but those records are not returned with option (skipUpsert=false)
I added additional primary keys 10 days ago and the segment was created 2 days back.Any ideas what the issue is ? @Neha PawarNikhil
05/20/2022, 9:03 PMmvn clean install -DskipTests -Pbin-dist -T 4 -Djdk.version=8
getting this error
Caused by: shaded.com.fasterxml.jackson.databind.JsonMappingException: Incompatible Jackson version: 2.10.0
at shaded.com.fasterxml.jackson.module.scala.JacksonModule$class.setupModule(JacksonModule.scala:64)
at shaded.com.fasterxml.jackson.module.scala.DefaultScalaModule.setupModule(DefaultScalaModule.scala:19)
at shaded.com.fasterxml.jackson.databind.ObjectMapper.registerModule(ObjectMapper.java:808)
at org.apache.spark.util.JsonProtocol$.<init>(JsonProtocol.scala:59)
at org.apache.spark.util.JsonProtocol$.<clinit>(JsonProtocol.scala)
... 32 more
job spec is here
executionFrameworkSpec:
name: 'spark'
segmentGenerationJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.spark.SparkSegmentGenerationJobRunner'
extraConfigs:
stagingDir: <s3://nikhil-dw-dev/pinot/staging/>
dependencyJarDir: '<s3://nikhil-dw-dev/pinot/apache-pinot-incubating-0.7.1-bin/plugins>'
jobType: SegmentCreation
inputDirURI: '<s3://nikhil-dw-dev/pinot/pinot_input/>'
includeFileNamePattern: 'glob:**/*.parquet'
outputDirURI: '<s3://nikhil-dw-dev/pinot/pinot_output3/>'
overwriteOutput: true
pinotFSSpecs:
-
className: org.apache.pinot.plugin.filesystem.S3PinotFS
scheme: s3
configs:
region: us-east-1
recordReaderSpec:
dataFormat: 'parquet'
className: 'org.apache.pinot.plugin.inputformat.parquet.ParquetRecordReader'
tableSpec:
tableName: 'students'
schemaURI: '<s3://nikhil-dw-dev/pinot/students_schema.json>'
tableConfigURI: '<s3://nikhil-dw-dev/pinot/students_table.json>'
Anish Nair
05/23/2022, 3:10 PMLuis Fernandez
05/23/2022, 4:52 PM[
{
"message": "java.net.UnknownHostException: pinot-server-1.pinot-server-headless.pinot.svc.cluster.local: Name or service not known\n\tat java.base/java.net.Inet4AddressImpl.lookupAllHostAddr(Native Method)\n\tat java.base/java.net.InetAddress$PlatformNameService.lookupAllHostAddr(InetAddress.java:929)\n\tat java.base/java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1519)\n\tat java.base/java.net.InetAddress$NameServiceAddresses.get(InetAddress.java:848)",
"errorCode": 425
},
{
"message": "1 servers [pinot-server-1_O] not responded",
"errorCode": 427
}
]