Sucheth Shivakumar
05/11/2023, 3:22 AMAbhinav Ittekot
05/11/2023, 5:46 AMMichael Kreis
05/11/2023, 11:06 AM....is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on
According to this documentation an additional permission to deployments/finalizers
is needed, which is not mentioned in you documentation. Is that just missing on the documentation or am I doing something wrong?Amenreet Singh Sodhi
05/11/2023, 11:20 AMKirill Lyashko
05/11/2023, 11:33 AMlast-state
. However when I changed the parallelism job master has failed once during startup, because it couldnāt reschedule state from latest checkpoint and then the job has been started without any state by operator. And I donāt see anything in logs what would explain such behaviour.
Does anyone faced anything similar?
I would be curious to know if itās a feature or bug. And could be the behaviour configured to fail job in such cases?Adam Fleishaker
05/11/2023, 1:52 PMEl Houssine Talab
05/11/2023, 2:35 PMStreamPhysicalOverAggregate doesn't support consuming update and delete changes which is produced by node Deduplicate(keep=[LastRow], key=[home, room, sn], order=[ROWTIME])
Any thoughts? Thanks.
WITH t1 AS (
/* DEDUPLICATION */
SELECT
home,
room,
sn,
temp,
event_ts,
arrival_ts
FROM (
SELECT
*,
ROW_NUMBER() OVER (PARTITION BY home, room, sn ORDER BY event_ts DESC) AS row_num
FROM telemetry
WHERE
event_ts >= CURRENT_WATERMARK(event_ts)
AND ABS(TIMESTAMPDIFF(SECOND, event_ts, arrival_ts)) <= 60
)
WHERE row_num = 1
)
/* OVER AGGREGATION */
SELECT
home,
room,
sn,
event_ts,
arrival_ts,
AVG(temp) OVER (
PARTITION BY home, room
ORDER BY event_ts
RANGE BETWEEN INTERVAL '1' MINUTE PRECEDING AND CURRENT ROW
) as avg_temp
FROM t1
Sucheth Shivakumar
05/11/2023, 2:56 PMConor McGovern
05/11/2023, 3:02 PMDo not upgrade Flink and the Kafka Connector version at the same time.
What is the problem with upgrading both at the same time? Thanks for your help.
(Link to the docs in question: https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/datastream/kafka/#upgrading-to-the-latest-connector-version)Mingliang Liu
05/11/2023, 11:09 PMJob Manager -> Configuration
page shows different configurations under different sections. This is good for categorizing them. But the height of each section is too short, and we need to scroll too much to find a specific configurations. Also searching is not showing all of the matches unless we hit Enter
multiple times or scroll.
Is it a good idea to put different sections of JM configuration to different tabs, similar to Running Jobs -> Checkpoints
?Adam Richardson
05/12/2023, 4:45 AMEl Houssine Talab
05/12/2023, 8:06 AMDheeraj Panangat
05/12/2023, 10:47 AMHash collision on user-specified ID "uid_split_monitor_tableRoot". Most likely cause is a non-unique ID. Please check that all IDs specified via `uid(String)` are unique.
at org.apache.flink.streaming.api.graph.StreamGraphHasherV2.generateNodeHash(StreamGraphHasherV2.java:185)
at org.apache.flink.streaming.api.graph.StreamGraphHasherV2.traverseStreamGraphAndGenerateHashes(StreamGraphHasherV2.java:110)
Our Eg:
table1 = tableEnv.sqlQuery("");
table2 = tableEnv.sqlQuery("");
tableRoot = tableEnv.sqlQuery("");
table3= table1.join(tableRoot);
table4= table2.join(tableRoot)
Has anyone faces this ? Anything we are doing wrong here ?
Appreciate any help.
ThanksSumit Singh
05/12/2023, 11:11 AMaswin jose
05/12/2023, 11:59 AMKaushalya Samarasekera
05/12/2023, 2:06 PMtransactions
(1 partition, 1 replica) exists in kafka, running on a wurstmeister/kafka docker container.
CREATE TABLE transactions
(
`account_id` BIGINT,
`amount` BIGINT
) WITH (
'connector' = 'kafka',
'topic' = 'transactions',
'properties.bootstrap.servers' = 'localhost:59092',
'properties.group.id' = 'testGroup',
'scan.startup.mode' = 'earliest-offset',
'format' = 'json');
Then I run
insert into transactions values (1, 23);
After the above insert, I expected the record to appear as an event in the transactions
topic but it does not.
If I run select * from transactions, I can see the record in the table. Just not in kafka.
Any pointers would be much appreciated.Simon Dahlbacka
05/12/2023, 2:13 PMto_date
, docs say TO_DATE(string1[, string2])
, I am calling it using
TO_DATE(CAST(`OAORDT` as string), 'yyyyMMdd') AS `orderDate`,
Invalid function call:\nTO_DATE(STRING, STRING NOT NULL) Expected signatures are:\nTO_DATE(*). Any Ideas what is wrong? flink 1.16.1Nizar Hejazi
05/14/2023, 6:38 AMWITH ('connector'='kafka', 'format' = 'debezium-json')
supports:
āŖļø insertion, update before, update after, and delete
⦠What does the community use as an external system for dynamic tables?
⢠There are query restrictions (state size for continuous queries evaluated on unbounded streams that need to update perv. emitted results.)
⦠Does Flink has to maintain this state even when dynamic table is written to external system?
⦠Any benchmark for state size for different queries?
ThanksNicholas Erasmus
05/15/2023, 8:11 AMBarisa Obradovic
05/15/2023, 9:21 AMRon Ben Arosh
05/15/2023, 10:47 AMCannot rescale the given pointwise partitioner.\nDid you change the partitioner to forward or rescale?
The filnk job currently contain only āforwardā arrows between operators. This job contain no broadcast and no explicit shuffles.
How this issue can be solved?
or more broadly: how can flink be both autoscaled and contain stable checkpoints?Abhishek Joshi
05/15/2023, 11:27 AMRashmin Patel
05/15/2023, 12:09 PMē°ęå
05/15/2023, 12:11 PMē°ęå
05/15/2023, 12:14 PMJean-Baptiste PIN
05/15/2023, 1:42 PMJean-Baptiste PIN
05/15/2023, 1:42 PMDavid Wisecup
05/15/2023, 3:07 PMCaused by: java.lang.NoClassDefFoundError: org/apache/flink/api/common/serialization/DeserializationSchema
When I change the flink deps as a normal scope I get further, but get this runtime error:
Caused by: java.lang.reflect.InaccessibleObjectException: Unable to make field private final java.lang.Object[] java.util.Arrays$ArrayList.a accessible: module java.base does not "opens java.util" to unnamed module @6366ebe0
Jiahao Wang Zhuo
05/15/2023, 4:12 PMFileSink
using the parquet BulkFormat (on 1.16). I wanted to create a custom rolling policy based on the file size and processing time.
I see on FLINK-13027 that this was supported for StreamingFileSink
since Flink-1.10. However, when I try to extend the CheckpointRollingPolicy
, it still rolls only on checkpoint times or org.apache.flink.util.SerializedThrowable: java.lang.UnsupportedOperationException: Bulk Part Writers do not support "pause and resume" operations.
(if I set shouldRollOnCheckpoint
to false). Would I be correct to say that custom rolling policies are not supported for FileSink
but only for StreamingFileSink
instead (which is marked as deprecated from the docs)? If so, do you have any plan on supporting that? Is there any blocker/challenge on that being possible?Trystan
05/15/2023, 5:29 PMSplit
as well as a SplitState
? if i understand correctly, the SplitState
would be the thing maintaining state as the SourceReader
reads through the split⦠but if weāre in Batch execution mode, would it matter? if the source is not checkpointed (batch mode), wouldnāt the SourceReader
need to resume from zero and reread the entire split?
i feel like iām missing something.
concrete example: a dynamo scan. segments map to splits, but within a segment the results may be paginated. i could store that pagination key⦠but does it matter? if the task fails, wouldnāt all data in the immediate downstream persistence layer be wiped anyway? and if so iād need to resume from zero, not the partially paginated result. the connection between batch recovery and the persisted intermediate storage isnāt super clear to me in this mode