Presto: multiple coordinator support

Created on 9 Nov 2015  路  8Comments  路  Source: prestodb/presto

Hi. I'm novie of presto.

Do you, guys have plan to support multiple coodinator?
I have known discovery.uri have only one host. I hope to support multiple coodinators of discovery.uri.

If I wrong, please take a example. It is very helpful for me.

stale

Most helpful comment

Did you use the config at the link I provided? That should do an Active / Backup style HAProxy which means the second coordinator should not be used unless the first goes down. The issue you face might be if the first coordinator fails, then the second is contacted to learn the node addresses, then the first coordinator recovers at which point the first and second coordinator both request node heartbeats. One thing to try might be to restart the second coordinator once the first one comes back up to see if that clears its node info (unless its stored to disk? - it's been a while since I worked on this so I can't remember the behaviour).

The other alternative is to use heartbeat / pacemaker traditional active/standby clustering to keep one coordinator shut down until the other one fails, at which point the cluster tech would start up the other coordinator and fail over the VIP to that server so that worker nodes would transparently contact the second coordinator at the same IP address.

I would definitely like to see better native support for coordinator HA, including documentation, rather than admins having to come up with their own solutions.

All 8 comments

The same issue reported here (on a later date but better explained).

That describes multiple coordinators for performance scaling clusters.

It would be nice if presto supported coordinator HA to start with.

One could use a load balancer for coordinator high availability, I have HAProxy config for this and other related Big Bata and NoSQL tech here:

https://github.com/HariSekhon/haproxy-configs

That describes multiple coordinators for performance scaling clusters.

It would be nice if presto supported coordinator HA to start with.

One could use a load balancer for coordinator high availability, I have HAProxy config for this and other related big data and nosql tech here:

https://github.com/HariSekhon/haproxy-configs

Hi Hari, sorry to disturb you. I tried to deploy a presto cluster with multiple active coordinator nodes, and use haproxy to achieve high availability. But I got a lot of warnings in SqlTaskManager, talking about the node is switching coordinator affinity from one to other. Is that ok?
Thanks a lot

Hi @ZJULothar I am not sure about the HAProxy but as of now, PrestoDB doesn't support having multiple coordinators and this is what happens if you try to configure multiple coordinators, the nodes infinitly switch between those coordinators whenever each coordinator ask for a heartbeat.

@sbmaggarwal OK, thanks for your reply

Did you use the config at the link I provided? That should do an Active / Backup style HAProxy which means the second coordinator should not be used unless the first goes down. The issue you face might be if the first coordinator fails, then the second is contacted to learn the node addresses, then the first coordinator recovers at which point the first and second coordinator both request node heartbeats. One thing to try might be to restart the second coordinator once the first one comes back up to see if that clears its node info (unless its stored to disk? - it's been a while since I worked on this so I can't remember the behaviour).

The other alternative is to use heartbeat / pacemaker traditional active/standby clustering to keep one coordinator shut down until the other one fails, at which point the cluster tech would start up the other coordinator and fail over the VIP to that server so that worker nodes would transparently contact the second coordinator at the same IP address.

I would definitely like to see better native support for coordinator HA, including documentation, rather than admins having to come up with their own solutions.

@HariSekhon I got it, thanks buddy

This issue has been automatically marked as stale because it has not had any activity in the last 2 years. If you feel that this issue is important, just comment and the stale tag will be removed; otherwise it will be closed in 7 days. This is an attempt to ensure that our open issues remain valuable and relevant so that we can keep track of what needs to be done and prioritize the right things.

Was this page helpful?
0 / 5 - 0 ratings