Tanmay Movva
10/05/2021, 2:11 PMNo valid brokers found for backendentityview'
We got to know it is because, the trino-pinot connector doesn’t support mixed case table name. Is anything planned to support mixed case table names in the connector?Kamal Chavda
10/05/2021, 7:18 PMjava.lang.IllegalStateException: Cannot read single-value from Collection: [AQEAACDmEAAA5no2BviTXcB1T2ijhAxBQA==, 4326] for column: point
at shaded.com.google.common.base.Preconditions.checkState(Preconditions.java:721) ~[pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.recordtransformer.DataTypeTransformer.standardizeCollection(DataTypeTransformer.java:193) ~[pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.recordtransformer.DataTypeTransformer.standardize(DataTypeTransformer.java:138) ~[pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.recordtransformer.DataTypeTransformer.transform(DataTypeTransformer.java:88) ~[pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.recordtransformer.CompositeTransformer.transform(CompositeTransformer.java:82) ~[pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager.processStreamEvents(LLRealtimeSegmentDataManager.java:491) [pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager.consumeLoop(LLRealtimeSegmentDataManager.java:402) [pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at org.apache.pinot.core.data.manager.realtime.LLRealtimeSegmentDataManager$PartitionConsumer.run(LLRealtimeSegmentDataManager.java:538) [pinot-all-0.7.1-jar-with-dependencies.jar:0.7.1-afa4b252ab1c424ddd6c859bb305b2aa342b66ed]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_282]
The point column has this as value:
"point" : {
"wkb" : "AQEAACDmEAAA5no2BviTXcB1T2ijhAxBQA==",
"srid" : 4326
},
Any suggestions on how to resolve? I have the column as string in the Pinot table schema.beerus
10/06/2021, 10:29 AMDeepak Mishra
10/06/2021, 2:07 PMdate -v-1d +%F
where ' date is set under includeFilePattern parameter -includeFileNamePattern: 'glob:**/{date}/*.avro'. While executing spark submit job with this command -
$SPARK_HOME/bin/spark-submit --class org.apache.pinot.tools.admin.command.LaunchDataIngestionJobCommand --master "local[2]" \
--deploy-mode client --conf "spark.driver.extraJavaOptions=-Dplugins.dir=${PINOT_DISTRIBUTION_DIR}/plugins \
-Dlog4j2.configurationFile=${PINOT_DISTRIBUTION_DIR}/conf/pinot-ingestion-job-log4j2.xml" \
--conf "spark.driver.extraClassPath=${PINOT_DISTRIBUTION_DIR}/lib/pinot-all-${PINOT_VERSION}-jar-with-dependencies.jar/Users/deemish2/apache pinot 0.8.0 bin/plugins/pinot batch ingestion/pinot batch ingestion spark/pinot batch ingestion spark 0.8.0 shaded.jar/Users/deemish2/apache-pinot-0.8.0-bin/plugins/pinot-file-system/pinot-hdfs/pinot-hdfs-0.8.0-shaded.jar" \
local://${PINOT_DISTRIBUTION_DIR}/lib/pinot-all-${PINOT_VERSION}-jar-with-dependencies.jar \
-jobSpecFile ${PINOT_DIR}/SparkingestionJobSpec.yaml -values date=date -v-1d +%F
It gives error like - Caused by: java.lang.IllegalArgumentException: Positive number of partitions required.
It looks like this argument - -values date=date -v-1d +%F
. works only with bin/pinot-ingestion.sh.
Please help to execute this spark ingestion job to push previous date data in pinotLuis Fernandez
10/06/2021, 7:27 PMWill Gan
10/06/2021, 7:49 PMSadim Nadeem
10/07/2021, 6:08 AMManish Soni
10/07/2021, 7:04 AM2021-10-07 06:41:20.000 WARN [BaseBrokerRequestHandler] [jersey-server-managed-async-executor-15] Failed to find time boundary info for hybrid table:
beerus
10/07/2021, 9:55 AMArpita Bajpai
10/07/2021, 10:16 AMIlya Yatsishin
10/07/2021, 11:03 AMTrying to create instance for class org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner
Initializing PinotFS for scheme file, classname org.apache.pinot.spi.filesystem.LocalPinotFS
Creating an executor service with 1 threads(Job parallelism: 1, available cores: 80.)
Trying to create instance for class org.apache.pinot.plugin.ingestion.batch.standalone.SegmentTarPushJobRunner
Initializing PinotFS for scheme file, classname org.apache.pinot.spi.filesystem.LocalPinotFS
Start pushing segments: []... to locations: [org.apache.pinot.spi.ingestion.batch.spec.PinotClusterSpec@78de58ea] for table trips_OFFLINE
Luis Fernandez
10/08/2021, 1:14 PMselect * from ads_metrics where user_id=x and serve_time >= 1633651200
when I use this query like this the numEntriesScannedInFilter
shoots up quite considerably if I don’t use serve_time I get 0 anyone knows why that may be? I currenly have a rangeindex in the serve_time
column and an invertedIndex + partitioning on the user_idBowen Wan
10/09/2021, 12:43 AMnumDocsScanned
remains the same.
My index config and query are like follow:
"starTreeIndexConfigs": [
{
"dimensionsSplitOrder": [
"id",
"A",
"B",
"C",
"D"
],
"functionColumnPairs": [
"DISTINCT_COUNT_HLL__id"
],
"maxLeafRecords": 10000
}
]
Query:
SELECT DISTINCTCOUNTHLL(id), A FROM MyTable WHERE B = 'a' GROUP BY A ORDER BY DISTINCTCOUNTHLL(id) DESC LIMIT 20
Zsolt Takacs
10/11/2021, 7:54 AMDeepak Mishra
10/11/2021, 9:08 AMsuraj kamath
10/12/2021, 9:35 AMDunith Dhanushka
Failed to generate Pinot segment for file - file:/Users/dunith/Projects/streamlit/rawdata/uber-raw-data-sep14.csv
java.lang.IllegalArgumentException: Invalid format: "null"
at org.joda.time.format.DateTimeParserBucket.doParseMillis(DateTimeParserBucket.java:187) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.joda.time.format.DateTimeFormatter.parseMillis(DateTimeFormatter.java:826) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.segment.creator.impl.SegmentColumnarIndexCreator.writeMetadata(SegmentColumnarIndexCreator.java:552) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.segment.creator.impl.SegmentColumnarIndexCreator.seal(SegmentColumnarIndexCreator.java:512) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.segment.creator.impl.SegmentIndexCreationDriverImpl.handlePostCreation(SegmentIndexCreationDriverImpl.java:284) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.segment.local.segment.creator.impl.SegmentIndexCreationDriverImpl.build(SegmentIndexCreationDriverImpl.java:257) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.ingestion.batch.common.SegmentGenerationTaskRunner.run(SegmentGenerationTaskRunner.java:111) ~[pinot-all-0.8.0-jar-with-dependencies.jar:0.8.0-c4ceff06d21fc1c1b88469a8dbae742a4b609808]
at org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner.lambda$submitSegmentGenTask$1(SegmentGenerationJobRunner.java:263) ~[pinot-batch-ingestion-standalone-0.8.0-shaded.jar:0.8.0-9a0f41bc24243ff74315723b0153b534c2596e30]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
Dunith Dhanushka
Kamal Chavda
10/14/2021, 8:51 PMMetrics aggregation and upsert cannot be enabled together
when creating a realtime table? Will add log and schema in thread.Deepak Mishra
10/15/2021, 9:50 AMTony Requist
10/15/2021, 2:16 PMpinot.controller.storage.factory.s3.serverSideEncryption=aws:kms
pinot.controller.storage.factory.s3.ssekmsKeyId=KEY
and I get the rather odd error message
Unknown value 'aws:kms' for S3PinotFS config: 'serverSideEncryption'. Supported values are: [AES256, aws:kms]
Kamal Chavda
10/15/2021, 5:12 PMOct 15 17:11:23 ip-10-0-7-125 superset[1849861]: ERROR:root:'timestamp'
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: Traceback (most recent call last):
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/flask_appbuilder/api/__init__.py", line 84, in wraps
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return f(self, *args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/views/base_api.py", line 80, in wraps
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: duration, response = time_function(f, self, *args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/utils/core.py", line 1368, in time_function
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: response = func(*args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/utils/log.py", line 224, in wrapper
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: value = f(*args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/datasets/api.py", line 236, in post
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: new_model = CreateDatasetCommand(g.user, item).run()
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/datasets/commands/create.py", line 47, in run
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: self.validate()
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/datasets/commands/create.py", line 87, in validate
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: if database and not DatasetDAO.validate_table_exists(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/datasets/dao.py", line 81, in validate_table_exists
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: database.get_table(table_name, schema=schema)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/superset/superset/models/core.py", line 603, in get_table
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return Table(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "<string>", line 2, in __new__
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/util/deprecations.py", line 139, in warned
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return fn(*args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/sql/schema.py", line 560, in __new__
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: metadata._remove_table(name, schema)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/util/langhelpers.py", line 68, in __exit__
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: compat.raise_(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 182, in raise_
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: raise exception
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/sql/schema.py", line 555, in __new__
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: table._init(name, metadata, *args, **kw)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/sql/schema.py", line 644, in _init
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: self._autoload(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/sql/schema.py", line 667, in _autoload
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: autoload_with.run_callable(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 2212, in run_callable
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return conn.run_callable(callable_, *args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 1653, in run_callable
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return callable_(self, *args, **kwargs)
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 469, in reflecttable
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return insp.reflecttable(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/engine/reflection.py", line 664, in reflecttable
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: for col_d in self.get_columns(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/sqlalchemy/engine/reflection.py", line 390, in get_columns
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: col_defs = self.dialect.get_columns(
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/pinotdb-0.3.6-py3.8.egg/pinotdb/sqlalchemy.py", line 390, in get_columns
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: columns = [
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/pinotdb-0.3.6-py3.8.egg/pinotdb/sqlalchemy.py", line 393, in <listcomp>
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: "type": get_type(spec["dataType"], spec.get("fieldSize")),
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: File "/app/superset/lib/python3.8/site-packages/pinotdb-0.3.6-py3.8.egg/pinotdb/sqlalchemy.py", line 458, in get_type
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: return type_map[data_type.lower()]
Oct 15 17:11:23 ip-10-0-7-125 superset[1849861]: KeyError: 'timestamp'
I've upgraded to latest pinotdb driver.Vibhor Jain
10/18/2021, 1:58 PMDeepak Mishra
10/19/2021, 4:30 AMMahesh babu
10/20/2021, 8:40 AMManish Soni
10/20/2021, 10:19 AMSaad Khan
10/20/2021, 6:30 PMPiyush Chauhan
10/21/2021, 6:54 AMsuraj kamath
10/21/2021, 7:25 AMselect lookup('tableB', 'username', 'orgId', orgId, 'userId', userId) as username from tableA where username is not null limit 10
But I see an error with
Unsupported predicate type: IS_NOT_NULL
Full Error screenshot attachedeywek
10/21/2021, 2:04 PM