riak failure on concurrent writes
emcclanahan at basho.com
Wed Oct 3 12:08:31 EDT 2012
For fastest loading, you might try hitting more than one node, just
round-robining around the cluster should work.
Additionally, make sure that you've done basic tunings to the cluster,
and try adjusting max_open_files in the eleveldb config section to a
higher value (but not too high, as you don't have a lot of memory on
those nodes). Once you get past the first level in leveldb, you can
run into file handle contention at low values.
Basic sysctl tuning values:
Also try changing your scheduler to deadline or noop (noop is for SSDs, mostly):
On Wed, Oct 3, 2012 at 5:42 AM, Venki Yedidha
<venkatesh.yedidha at gmail.com> wrote:
> Hi All,
> Now that I have a 5 node riak cluster of which all are running on
> eleveldb backend(4 GB ram for each node), I am trying to insert appr: 34000
> objects on one of the riak node in the cluster through riak-erlang client
> asynchronously...Its failing after some time to accept inserts...Do I need
> to change any settings in the config files so that riak runs stable to
> handle concurrent writes upto 35k with out failure?
> Please help me on the above..
> riak-users mailing list
> riak-users at lists.basho.com
More information about the riak-users