Write_lock error has occurred after inserting 12M data

Alexander Sicular siculars at gmail.com
Fri Jul 30 17:06:24 EDT 2010


This may be another max file handler type of error. Or maybe even an oom thing if the key length is large.


On Jul 30, 2010, at 4:59 PM, Grant Schofield wrote:

> I am not sure if you hit an already fixed bug in Bitcask or not. What version of Riak are you running on currently?
> 
> Grant Schofield
> Developer Advocate
> Basho Technologies
> 
> On Jul 30, 2010, at 1:28 PM, Ken Matsumoto wrote:
> 
>> Hi all,
>> 
>> I just tried to insert 1Billion data records.
>> But I got the "write_lock" error after 12Million data.
>> What is the reason and how should I avoid this?
>> I use bitcask (default) backend and no parameters changed in config file.
>> 1 record is just 70B text data.
>> 
>> Regards,
>> 
>> Ken.
>> 
>> -- 
>> Ken Matsumoto
>> VP / Research & Development
>> Nomura Research Institute America, Inc.
>> NRI Pacific
>> 1400 Fashion Island Blvd., Suite 1010
>> San Mateo, CA 94404, U.S.A.
>> 
>> PLEASE READ:This e-mail is confidential and intended for the named recipient only. If you are not an intended recipient, please notify the sender and delete this e-mail.
>> 
>> 
>> _______________________________________________
>> riak-users mailing list
>> riak-users at lists.basho.com
>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
> 
> _______________________________________________
> riak-users mailing list
> riak-users at lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com





More information about the riak-users mailing list