Delete bucket?

Sean Cribbs sean at basho.com
Thu Jul 28 17:20:28 EDT 2011


Not sure if it's a problem in PHP (and should go without saying), but make
sure to open a new connection when performing additional client requests
(like a delete) while receiving a streamed response. Otherwise you might run
into a case of "accidental concurrency" like I did (see also
http://seancribbs.com/tech/2011/03/08/how-ripple-uses-fibers/). If your
client has a proper connection pool/stack this should not be a problem.

On Thu, Jul 28, 2011 at 5:11 PM, Jonathan Langevin <
jlangevin at loomlearning.com> wrote:

> Ah, nice. Checked the js client, which mentioned that ?keys=stream is a
> valid option to result in streamed key listings.
> I'll get that implemented in my PHP client, thanks for the pointer :-)
>
> *
>
>  <http://www.loomlearning.com/>
>  Jonathan Langevin
> Systems Administrator
> Loom Inc.
> Wilmington, NC: (910) 241-0433 - jlangevin at loomlearning.com -
> www.loomlearning.com - Skype: intel352
> *
>
>
> On Thu, Jul 28, 2011 at 4:42 PM, Jeremiah Peschka <
> jeremiah.peschka at gmail.com> wrote:
>
>> Depending on the client you're using, you can perform results streaming to
>> pull back results and process them in chunks rather than waiting for a
>> buffer to get filled.
>>
>> It's easy enough to write something like this using Ripple or
>> CorrugatedIron. I'm guessing it's possible with other clients.
>> ---
>> Jeremiah Peschka
>> Founder, Brent Ozar PLF, LLC
>>
>> On Jul 28, 2011, at 1:40 PM, Jonathan Langevin wrote:
>>
>> > I've read on the wiki that to delete a bucket, the only method is to
>> manually delete all keys within the bucket.
>> > So then what is the recommended process for deleting all keys within a
>> bucket, manually?
>> >
>> > I was initially just listing all keys within a bucket, and then
>> iterating the keys to send delete requests, but I hit a wall when I had too
>> many keys to get back in a list request (received header too large errors).
>> >
>> > So I assume the alternative would be to run a mapreduce to pull keys
>> from the bucket with a specified limit, to then execute the deletes?
>> > While that's fine for an "active record" style environment (where there
>> may be cleanup actions that must occur per object being deleted), is there
>> another method for deleting all keys within a bucket, massively? (Maybe via
>> a map call?)
>> >
>> >
>> > Jonathan Langevin
>> > Systems Administrator
>> > Loom Inc.
>> > Wilmington, NC: (910) 241-0433 - jlangevin at loomlearning.com -
>> www.loomlearning.com - Skype: intel352
>> >
>> > _______________________________________________
>> > riak-users mailing list
>> > riak-users at lists.basho.com
>> > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>
>>
>
> _______________________________________________
> riak-users mailing list
> riak-users at lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>


-- 
Sean Cribbs <sean at basho.com>
Developer Advocate
Basho Technologies, Inc.
http://www.basho.com/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.basho.com/pipermail/riak-users_lists.basho.com/attachments/20110728/f6acf126/attachment.html>


More information about the riak-users mailing list