Retrying requests to Riak
grourk at dropcam.com
Thu Apr 7 23:27:12 EDT 2011
I don't want to have a single load balancer because I want to avoid a single point of failure. And we'll be pushing enough data that it would be a huge bottleneck.
A failed node will not receive new requests, but when the requests that were sent to it fail I'd like to retry those automatically instead of having errors bubble up to our application.
On Thursday, April 7, 2011 at 8:17 PM, Wilson MacGyver wrote:
> use haproxy to ping each of the riak node. and remove nodes that
> aren't responding is
> what we did.
> we use a single haproxy instead of 1 per application server.
> On Thu, Apr 7, 2011 at 9:47 PM, Greg Nelson <grourk at dropcam.com> wrote:
> > Hello,
> > I have a simple three node cluster that I have been using for testing and
> > benchmarking Riak. Lately I've been simulating various failure scenarios --
> > like a node going down, disk going bad, etc.
> > My application talks to Riak through an haproxy instance running locally on
> > each application server. It's configured to round-robin over the nodes in
> > the cluster for both HTTP and PBC interfaces, and uses the HTTP /ping health
> > check. I assume this is a rather typical setup.
> Omnem crede diem tibi diluxisse supremum.
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the riak-users