Error while importing data

Nitish Sharma sharmanitishdutt at
Sat Nov 19 07:22:10 EST 2011

To give my Riak setup a good stress testing, I decided to import a large
dataset (consisting of around 160 million records). But before importing
the whole thing, I tested the import python script (using protocol buffers)
using 1 million records, which was successful with ~2200 writes/sec. The
script, essentially, puts the data into a queue and couple of threads gets
the data from the queue and store it in Riak.
When started with full dataset, after storing several million objects, I
get thread exception with timeout errors.
Following is the traceback:

  File "/usr/lib/python2.7/", line 552, in __bootstrap_inner
  File "/usr/lib/python2.7/", line 505, in run
    self.__target(*self.__args, **self.__kwargs)
  File "", line 23, in worker
  File "/usr/local/lib/python2.7/dist-packages/riak-1.3.0-py2.7.egg/riak/",
line 296, in store
    Result = t.put(self, w, dw, return_body)
  File "/usr/local/lib/python2.7/dist-packages/riak-1.3.0-py2.7.egg/riak/transports/",
line 188, in put
    msg_code, resp = self.recv_msg()
  File "/usr/local/lib/python2.7/dist-packages/riak-1.3.0-py2.7.egg/riak/transports/",
line 370, in recv_msg
    raise Exception(msg.errmsg)
Exception: timeout

The cluster consists of 3 nodes (Ubuntu 10.04). The nodes have enough disk
space; number of file handles used (~2500) are also within limit (32768);
number of concurrent ports 32768. I cant figure out what else could be the
possible reason for the exceptions.

Any Suggestions?

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <>

More information about the riak-users mailing list