riak failure on concurrent writes
siculars at gmail.com
Wed Oct 3 13:38:06 EDT 2012
I have also used the unix command "split" to split large csv files for easier management and parallel loading.
On Oct 3, 2012, at 12:08 PM, Evan Vigil-McClanahan wrote:
> For fastest loading, you might try hitting more than one node, just
> round-robining around the cluster should work.
> Additionally, make sure that you've done basic tunings to the cluster,
> and try adjusting max_open_files in the eleveldb config section to a
> higher value (but not too high, as you don't have a lot of memory on
> those nodes). Once you get past the first level in leveldb, you can
> run into file handle contention at low values.
> Basic sysctl tuning values:
> Also try changing your scheduler to deadline or noop (noop is for SSDs, mostly):
> On Wed, Oct 3, 2012 at 5:42 AM, Venki Yedidha
> <venkatesh.yedidha at gmail.com> wrote:
>> Hi All,
>> Now that I have a 5 node riak cluster of which all are running on
>> eleveldb backend(4 GB ram for each node), I am trying to insert appr: 34000
>> objects on one of the riak node in the cluster through riak-erlang client
>> asynchronously...Its failing after some time to accept inserts...Do I need
>> to change any settings in the config files so that riak runs stable to
>> handle concurrent writes upto 35k with out failure?
>> Please help me on the above..
>> riak-users mailing list
>> riak-users at lists.basho.com
> riak-users mailing list
> riak-users at lists.basho.com
More information about the riak-users