[pybsddb] Bulk Data load into a DB
Jon Kerr Nilsen
j.k.nilsen at fys.uio.no
Mon Mar 29 21:25:59 CEST 2010
On 29. mars 2010, at 19.54, Jesus Cea wrote:
> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA1
>
> On 03/26/2010 11:38 PM, Joepaul Puthenangadi wrote:
>> Was reading through the documentation and found something on the oracle
>> website about the possibility of loading bulk key and values into a hash
>> or btree db using the flag = |DB_MULTIPLE. Not trying to be picky or
>> anything just wanted to know whether that is something yet to be
>> implemented to |bsddb3 module.?
>
> pybsddb is fairly incomplete, yet. It is a work in progress.
>
> DB_MULTIPLE is not supported, currently. In fact using it will crash the
> application, corrupt memory, etc.
>
> I will support "bulk" interface sooner or later, depending of "popular
> demand" :). Are you demanding it? :))
>
> The details are tricky because I want to support arbitrary iterators,
> and I don't want to iterate twice over the data. And I need to iterate
> twice to know the lenght of the data and allocate appropiate memory
> size, trying to avoid "realloc()'s" in the process.
>
> I have had this interface in mind for years. Implementation will be
> "faster" if somebody demands it. I don't have personal need for it.
Hi!
We're working on this storage system Chelonia[1] and we use pybsddb for the replicated meta-database (and hey, it works :)). Anyway, we access the database over WAN, which is quite high in latency, so a bulk interface would be really cool. We practically demand it :)
BTW, great product this pybsddb!
cheers,
Jon
[1] http://www.nordugrid.org/chelonia/
More information about the pybsddb
mailing list