Guruguha Marur Sreenivasa
04/04/2023, 4:20 AMGyula Fóra
04/04/2023, 10:48 AMGuruguha Marur Sreenivasa
04/04/2023, 5:56 PMGuruguha Marur Sreenivasa
04/04/2023, 5:59 PMGuruguha Marur Sreenivasa
04/04/2023, 6:00 PMjobManager:
nodeSelector:
<http://topology.kubernetes.io/zone|topology.kubernetes.io/zone>: us-east-1a
taskManager:
nodeSelector:
<http://topology.kubernetes.io/zone|topology.kubernetes.io/zone>: us-east-1a
David Christle
04/07/2023, 10:40 PMpodTemplate
and then the second podTemplate
under the taskManager
. You’d just add in your node selectors in those parts.Guruguha Marur Sreenivasa
04/08/2023, 9:42 PMKafkaSource
use those AZs and connect to Kafka brokers that are on that AZ so that we can have task managers to be AZ local.
3. This way, we don't incur data transfer costs.David Christle
04/10/2023, 8:04 AMJobManager
in HA mode. Otherwise, since the JM usually runs on just one pod, if it crashes, your job goes down. Maybe pod topology spread constraints could help ensure the two JM pods (the main one and the backup) are scheduled in different zones.
I’m not sure how to make the broker dynamically selected based on zone within Flink.
Even if you could select the brokers in the same zone as the current TM pod, this still has a problem, though - there’s no guarantee that the TaskManager pods are scheduled all in one zone. So, the traffic from Kafka to the individual TMs would stay in the same zone. But the Flink application probably has some shuffle operations, so intermediate data would still cross zones and incur the cost.Guruguha Marur Sreenivasa
04/10/2023, 3:10 PM