https://pinot.apache.org/ logo
Join Slack
Powered by
# pinot-s3
  • k

    Kishore G

    01/11/2021, 1:42 AM
    What do you mean fail over purpose
  • m

    Matt

    01/11/2021, 5:18 PM
    At the moment I am having only single disk per pino in an AZ. If that disk fails I will lose all segments for that pino node
  • m

    Matt

    01/11/2021, 5:20 PM
    I assume the normal solution will be to add disk in multiple AZs and replicate segments. However I am also thinking on the possibility of having segment copy in S3 as deep storage.
  • k

    Kishore G

    01/11/2021, 5:21 PM
    yes, that happens by default if you configure s3 as the deep storage
  • k

    Kishore G

    01/11/2021, 5:21 PM
    you have only one replica
  • m

    Matt

    01/11/2021, 5:21 PM
    Ahh great, so I assume sync back will happen normally during a disk failure?
  • m

    Matt

    01/11/2021, 5:22 PM
    I mean happen automatically
  • k

    Kishore G

    01/11/2021, 5:23 PM
    yes
  • m

    Matt

    01/11/2021, 5:23 PM
    Trying to understand the recovery time
  • k

    Kishore G

    01/11/2021, 5:24 PM
    its the time for k8s to launch a new container + time to pull the segments from s3
  • m

    Matt

    01/11/2021, 5:24 PM
    As this approach is cost effective
  • k

    Kishore G

    01/11/2021, 5:24 PM
    you can use replication factor=2
  • m

    Matt

    01/11/2021, 5:24 PM
    That means 2 copies on same disk
  • k

    Kishore G

    01/11/2021, 5:24 PM
    then there will be two copies within the same AZ
  • k

    Kishore G

    01/11/2021, 5:25 PM
    across multiple nodes
  • m

    Matt

    01/11/2021, 5:30 PM
    ok great , I think I have to do that. Will try with deep storage first and see the recovery time meets RPO
  • k

    Kishore G

    01/11/2021, 5:33 PM
    note all of these can be done on demand
  • k

    Kishore G

    01/11/2021, 5:33 PM
    just change the replication factor in table config to 2 and invoke rebalance
  • k

    Kishore G

    01/11/2021, 5:35 PM
    btw, we should have this conversation in troubleshooting so others can also benefit
  • m

    Matt

    01/11/2021, 5:44 PM
    Sure Thanks for the details Kishore, You been very helpful as always..!!
  • a

    Ankit Kumar

    08/27/2021, 7:10 PM
    Hello guys, Good noon
  • k

    Kishore G

    08/27/2021, 7:12 PM
    hello
  • a

    Ankit Kumar

    08/27/2021, 7:15 PM
    Hi Kishore, Hope you are doing well
  • a

    Ankit Kumar

    08/27/2021, 7:16 PM
    I have recently started using pinot and I was trying to integrate deep storage for historical data, I am able to store and see the data for controller but it doesn’t work for server data
  • a

    Ankit Kumar

    08/27/2021, 7:18 PM
    I have added configurations like class, accessKey, secretKey, endpoint , protocol region and S3PinotFS as the s3, factory classes
  • a

    Ankit Kumar

    08/27/2021, 7:19 PM
    My controller config also has controller.data.dir = s3://<bucket>/path controller.local.temp.dir= local path
  • a

    Ankit Kumar

    08/27/2021, 7:21 PM
    And server has pinot.server.instance.dataDir= local path Pinot.server.instance.segmentTarDir= local path/segment
  • a

    Ankit Kumar

    08/27/2021, 7:21 PM
    Not sure if I have to configure bucket details as well for server
  • k

    Kishore G

    08/27/2021, 7:27 PM
    how are you pushing the segments to pinot
  • a

    Ankit Kumar

    08/27/2021, 8:40 PM
    segments are being pushed via Kafka streaming