--- /dev/null
+use Module::Build;
+
+use strict;
+
+my $build = Module::Build->new(
+ module_name => 'DBM::Deep',
+ license => 'perl',
+ requires => {
+# 'Carp' => '0.01',
+ 'Digest::MD5' => '1.00',
+ 'Scalar::Util' => '1.18',
+ },
+ optional => {
+ },
+ build_requires => {
+ 'Test::More' => '0.47',
+ 'Test::Exception' => '0.21',
+ },
+ create_makefile_pl => 'traditional',
+ add_to_cleanup => [
+ 'META.yml', '*.bak', '*.gz', 'Makefile.PL', 't/test*.db',
+ ],
+);
+
+$build->create_build_script;
+
--- /dev/null
+Revision history for DBM::Deep.
+
+0.97 ??? ?? ??:??:?? 2006 Pacific
+ - Reorganization of distribution
+ - Migration to Module::Build with EU::MM backwards compatibility
+ - Test coverage improved to ??%
+
+0.96 Oct 14 09:55:00 2005 Pacific
+ - Fixed build (OS X hidden files killed it)
+ - You can now pass in an optional filehandle to the constructor
+
+0.95 Oct 12 13:58:00 2005 Pacific
+ - Added optional autobless flag to preserve and restore blessed hashes
+ - Fixed bug where 0 could not be fetched using get_next_key
+ - Fixed bug where tie() constructor didn't accept a hash ref for args
+ - optimize() now preserves user/group/permissions
+ - Errors are now FATAL (meaning it calls die()), unless you set debug flag
+
+0.94 Apr 13 19:00:26 2004 Pacific
+ - Fixed bug reported by John Cardenas (corruption at key level when
+ replace of less data was done on bucket)
+
+0.93 Feb 15 19:53:17 2004 Pacific
+ - Fixed optmize() on Win32 where orig file couldn't be overwritten unless
+ filehandle was closed first. This change introduces a potential race
+ condition when using locking and optmize() on Win32, but it can be
+ fixed in the future using a soft copy instead of Perl's rename().
+
+0.92 Feb 12 19:10:22 2004 Pacific
+ - Fixed bug where passing a reference to a different DBM::Deep object
+ would still result in an internal reference.
+ - Added export() method for recursively extracting hashes/arrays into
+ standard in-memory Perl structures.
+ - Added import() method for recursively importing existing Perl hash/
+ array structures
+ - Fixed bug where optimize() wouldn't work if base level of DB was
+ an array instead of a hash.
+
+0.91 Feb 12 02:30:22 2004 Pacific
+ - Fixed bug with splice() when length of removed section was 0
+ - Updated POD re: circular refs and optimize()
+ - Had to jump version numbers to 0.91 because in previous releases
+ I used only a single digit after the decimal which was confusing
+ the CPAN indexer.
+
+0.10 Feb 11 08:58:35 2004 Pacific
+ - Fixed bug where default file mode was CLEARING files (Thanks Rich!)
+ - Added experimental support for circular references
+ - Fixed bugs in shift(), unshift() and splice() where nested objects
+ in array would be recursively re-stored as basic hashes/arrays
+ - Fixed typos in POD docs
+
+0.9 Feb 10 03:25:48 2004 Pacific
+ - Added Filters for storing/fetching keys/values
+ - Added hook for supplying own hashing algorithm
+ - FIxed some typos in POD docs, added new sections
+
+0.8 Feb 8 02:38:22 2004 Pacific
+ - Renamed to DBM::Deep for CPAN
+ - Added optimize() method for rekindling unused space
+ - Now returning hybrid tie()/OO object from new()
+ - Basic error handling introduced
+ - Added debug mode for printing errors to STDERR
+ - Added TYPE_HASH and TYPE_ARRAY constants for "type" param
+ - Added clone() method for safe copying of objects
+ - Wrote POD documentation
+ - Added set_pack() function for manipulating LONG_SIZE / LONG_PACK
+ - Added aliases for most tied functions for public use
+ - Now setting binmode() on FileHandle for Win32
+ - Added 45 unit tests
+
+0.7 Jan 4 11:31:50 2003 UTC
+ - Renamed to DeepDB
+ - Changed file signature to DPDB (not compatible with older versions)
+ - Converted array length to packed long instead of sprintf()ed string
+
+0.6 Dec 31 15:12:03 2002 UTC
+ - Some misc optimizations for speed
+
+0.5 Oct 18 08:55:29 2002 UTC
+ - support for force_return_next parameter in traverse_index() method for
+ ultra-fast combined key search/removal
+
+0.4 Oct 15 20:07:47 2002 UTC
+ - now making sure filehandle is open for all DB calls
+
+0.3 Oct 3 19:04:13 2002 UTC
+ - fixed bug that could cause corrupted data when using locking
+
+0.2 Aug 6 16:37:32 2002 UTC
+ - Removed base index caching, as it can cause problems when two processes
+ are populating the db at the same time (even with locking)
+
+0.1 Jun 3 08:06:26 2002 UTC
+ - initial release
--- /dev/null
+Build.PL
+Changes
+README
+Makefile.PL
+MANIFEST
+META.yml
+lib/DBM/Deep.pm
+t/01basic.t
+t/02hash.t
+t/03bighash.t
+t/04array.t
+t/05bigarray.t
+t/07error.t
+t/08locking.t
+t/09deephash.t
+t/10deeparray.t
+t/11largekeys.t
+t/12optimize.t
+t/13clone.t
+t/14setpack.t
+t/15filter.t
+t/16digest.t
+t/17circular.t
+t/18import.t
+t/19export.t
+t/20crossref.t
+t/21_tie.t
+t/22_tie_access.t
+t/23_stupidities.t
+t/24_internal_copy.t
--- /dev/null
+^_build
+^Build$
+^blib
+~$
+\.bak$
+^MANIFEST\.SKIP$
+CVS
+\.svn
+cover_db
+\..*\.sw.?$
+^Makefile$
+^pm_to_blib$
+^MakeMaker-\d
+^blibdirs$
+\.old$
+^#.*#$
+^\.#
--- /dev/null
+NAME
+ DBM::Deep - A pure perl multi-level hash/array DBM
+
+SYNOPSIS
+ use DBM::Deep;
+ my $db = new DBM::Deep "foo.db";
+
+ $db->{key} = 'value'; # tie() style
+ print $db->{key};
+
+ $db->put('key', 'value'); # OO style
+ print $db->get('key');
+
+ # true multi-level support
+ $db->{my_complex} = [
+ 'hello', { perl => 'rules' },
+ 42, 99 ];
+
+DESCRIPTION
+ A unique flat-file database module, written in pure perl. True
+ multi-level hash/array support (unlike MLDBM, which is faked), hybrid OO
+ / tie() interface, cross-platform FTPable files, and quite fast. Can
+ handle millions of keys and unlimited hash levels without significant
+ slow-down. Written from the ground-up in pure perl -- this is NOT a
+ wrapper around a C-based DBM. Out-of-the-box compatibility with Unix,
+ Mac OS X and Windows.
+
+INSTALLATION
+ Hopefully you are using CPAN's excellent Perl module, which will
+ download and install the module for you. If not, get the tarball, and
+ run these commands:
+
+ tar zxf DBM-Deep-*
+ cd DBM-Deep-*
+ perl Makefile.PL
+ make
+ make test
+ make install
+
+SETUP
+ Construction can be done OO-style (which is the recommended way), or
+ using Perl's tie() function. Both are examined here.
+
+ OO CONSTRUCTION
+ The recommended way to construct a DBM::Deep object is to use the new()
+ method, which gets you a blessed, tied hash or array reference.
+
+ my $db = new DBM::Deep "foo.db";
+
+ This opens a new database handle, mapped to the file "foo.db". If this
+ file does not exist, it will automatically be created. DB files are
+ opened in "r+" (read/write) mode, and the type of object returned is a
+ hash, unless otherwise specified (see OPTIONS below).
+
+ You can pass a number of options to the constructor to specify things
+ like locking, autoflush, etc. This is done by passing an inline hash:
+
+ my $db = new DBM::Deep(
+ file => "foo.db",
+ locking => 1,
+ autoflush => 1
+ );
+
+ Notice that the filename is now specified *inside* the hash with the
+ "file" parameter, as opposed to being the sole argument to the
+ constructor. This is required if any options are specified. See OPTIONS
+ below for the complete list.
+
+ You can also start with an array instead of a hash. For this, you must
+ specify the "type" parameter:
+
+ my $db = new DBM::Deep(
+ file => "foo.db",
+ type => DBM::Deep::TYPE_ARRAY
+ );
+
+ Note: Specifing the "type" parameter only takes effect when beginning a
+ new DB file. If you create a DBM::Deep object with an existing file, the
+ "type" will be loaded from the file header, and ignored if it is passed
+ to the constructor.
+
+ TIE CONSTRUCTION
+ Alternatively, you can create a DBM::Deep handle by using Perl's
+ built-in tie() function. This is not ideal, because you get only a
+ basic, tied hash (or array) which is not blessed, so you can't call any
+ functions on it.
+
+ my %hash;
+ tie %hash, "DBM::Deep", "foo.db";
+
+ my @array;
+ tie @array, "DBM::Deep", "bar.db";
+
+ As with the OO constructor, you can replace the DB filename parameter
+ with a hash containing one or more options (see OPTIONS just below for
+ the complete list).
+
+ tie %hash, "DBM::Deep", {
+ file => "foo.db",
+ locking => 1,
+ autoflush => 1
+ };
+
+ OPTIONS
+ There are a number of options that can be passed in when constructing
+ your DBM::Deep objects. These apply to both the OO- and tie- based
+ approaches.
+
+ * file
+ Filename of the DB file to link the handle to. You can pass a full
+ absolute filesystem path, partial path, or a plain filename if the
+ file is in the current working directory. This is a required
+ parameter.
+
+ * mode
+ File open mode (read-only, read-write, etc.) string passed to Perl's
+ FileHandle module. This is an optional parameter, and defaults to
+ "r+" (read/write). Note: If the default (r+) mode is selected, the
+ file will also be auto- created if it doesn't exist.
+
+ * type
+ This parameter specifies what type of object to create, a hash or
+ array. Use one of these two constants: "DBM::Deep::TYPE_HASH" or
+ "DBM::Deep::TYPE_ARRAY". This only takes effect when beginning a new
+ file. This is an optional parameter, and defaults to hash.
+
+ * locking
+ Specifies whether locking is to be enabled. DBM::Deep uses Perl's
+ Fnctl flock() function to lock the database in exclusive mode for
+ writes, and shared mode for reads. Pass any true value to enable.
+ This affects the base DB handle *and any child hashes or arrays*
+ that use the same DB file. This is an optional parameter, and
+ defaults to 0 (disabled). See LOCKING below for more.
+
+ * autoflush
+ Specifies whether autoflush is to be enabled on the underlying
+ FileHandle. This obviously slows down write operations, but is
+ required if you may have multiple processes accessing the same DB
+ file (also consider enable *locking* or at least *volatile*). Pass
+ any true value to enable. This is an optional parameter, and
+ defaults to 0 (disabled).
+
+ * volatile
+ If *volatile* mode is enabled, DBM::Deep will stat() the DB file
+ before each STORE() operation. This is required if an outside force
+ may change the size of the file between transactions. Locking also
+ implicitly enables volatile. This is useful if you want to use a
+ different locking system or write your own. Pass any true value to
+ enable. This is an optional parameter, and defaults to 0 (disabled).
+
+ * autobless
+ If *autobless* mode is enabled, DBM::Deep will preserve blessed
+ hashes, and restore them when fetched. This is an experimental
+ feature, and does have side-effects. Basically, when hashes are
+ re-blessed into their original classes, they are no longer blessed
+ into the DBM::Deep class! So you won't be able to call any DBM::Deep
+ methods on them. You have been warned. This is an optional
+ parameter, and defaults to 0 (disabled).
+
+ * filter_*
+ See FILTERS below.
+
+ * debug
+ Setting *debug* mode will make all errors non-fatal, dump them out
+ to STDERR, and continue on. This is for debugging purposes only, and
+ probably not what you want. This is an optional parameter, and
+ defaults to 0 (disabled).
+
+TIE INTERFACE
+ With DBM::Deep you can access your databases using Perl's standard
+ hash/array syntax. Because all Deep objects are *tied* to hashes or
+ arrays, you can treat them as such. Deep will intercept all reads/writes
+ and direct them to the right place -- the DB file. This has nothing to
+ do with the "TIE CONSTRUCTION" section above. This simply tells you how
+ to use DBM::Deep using regular hashes and arrays, rather than calling
+ functions like "get()" and "put()" (although those work too). It is
+ entirely up to you how to want to access your databases.
+
+ HASHES
+ You can treat any DBM::Deep object like a normal Perl hash reference.
+ Add keys, or even nested hashes (or arrays) using standard Perl syntax:
+
+ my $db = new DBM::Deep "foo.db";
+
+ $db->{mykey} = "myvalue";
+ $db->{myhash} = {};
+ $db->{myhash}->{subkey} = "subvalue";
+
+ print $db->{myhash}->{subkey} . "\n";
+
+ You can even step through hash keys using the normal Perl "keys()"
+ function:
+
+ foreach my $key (keys %$db) {
+ print "$key: " . $db->{$key} . "\n";
+ }
+
+ Remember that Perl's "keys()" function extracts *every* key from the
+ hash and pushes them onto an array, all before the loop even begins. If
+ you have an extra large hash, this may exhaust Perl's memory. Instead,
+ consider using Perl's "each()" function, which pulls keys/values one at
+ a time, using very little memory:
+
+ while (my ($key, $value) = each %$db) {
+ print "$key: $value\n";
+ }
+
+ Please note that when using "each()", you should always pass a direct
+ hash reference, not a lookup. Meaning, you should never do this:
+
+ # NEVER DO THIS
+ while (my ($key, $value) = each %{$db->{foo}}) { # BAD
+
+ This causes an infinite loop, because for each iteration, Perl is
+ calling FETCH() on the $db handle, resulting in a "new" hash for foo
+ every time, so it effectively keeps returning the first key over and
+ over again. Instead, assign a temporary variable to "$db-"{foo}>, then
+ pass that to each().
+
+ ARRAYS
+ As with hashes, you can treat any DBM::Deep object like a normal Perl
+ array reference. This includes inserting, removing and manipulating
+ elements, and the "push()", "pop()", "shift()", "unshift()" and
+ "splice()" functions. The object must have first been created using type
+ "DBM::Deep::TYPE_ARRAY", or simply be a nested array reference inside a
+ hash. Example:
+
+ my $db = new DBM::Deep(
+ file => "foo-array.db",
+ type => DBM::Deep::TYPE_ARRAY
+ );
+
+ $db->[0] = "foo";
+ push @$db, "bar", "baz";
+ unshift @$db, "bah";
+
+ my $last_elem = pop @$db; # baz
+ my $first_elem = shift @$db; # bah
+ my $second_elem = $db->[1]; # bar
+
+ my $num_elements = scalar @$db;
+
+OO INTERFACE
+ In addition to the *tie()* interface, you can also use a standard OO
+ interface to manipulate all aspects of DBM::Deep databases. Each type of
+ object (hash or array) has its own methods, but both types share the
+ following common methods: "put()", "get()", "exists()", "delete()" and
+ "clear()".
+
+ * put()
+ Stores a new hash key/value pair, or sets an array element value.
+ Takes two arguments, the hash key or array index, and the new value.
+ The value can be a scalar, hash ref or array ref. Returns true on
+ success, false on failure.
+
+ $db->put("foo", "bar"); # for hashes
+ $db->put(1, "bar"); # for arrays
+
+ * get()
+ Fetches the value of a hash key or array element. Takes one
+ argument: the hash key or array index. Returns a scalar, hash ref or
+ array ref, depending on the data type stored.
+
+ my $value = $db->get("foo"); # for hashes
+ my $value = $db->get(1); # for arrays
+
+ * exists()
+ Checks if a hash key or array index exists. Takes one argument: the
+ hash key or array index. Returns true if it exists, false if not.
+
+ if ($db->exists("foo")) { print "yay!\n"; } # for hashes
+ if ($db->exists(1)) { print "yay!\n"; } # for arrays
+
+ * delete()
+ Deletes one hash key/value pair or array element. Takes one
+ argument: the hash key or array index. Returns true on success,
+ false if not found. For arrays, the remaining elements located after
+ the deleted element are NOT moved over. The deleted element is
+ essentially just undefined, which is exactly how Perl's internal
+ arrays work. Please note that the space occupied by the deleted
+ key/value or element is not reused again -- see "UNUSED SPACE
+ RECOVERY" below for details and workarounds.
+
+ $db->delete("foo"); # for hashes
+ $db->delete(1); # for arrays
+
+ * clear()
+ Deletes all hash keys or array elements. Takes no arguments. No
+ return value. Please note that the space occupied by the deleted
+ keys/values or elements is not reused again -- see "UNUSED SPACE
+ RECOVERY" below for details and workarounds.
+
+ $db->clear(); # hashes or arrays
+
+ HASHES
+ For hashes, DBM::Deep supports all the common methods described above,
+ and the following additional methods: "first_key()" and "next_key()".
+
+ * first_key()
+ Returns the "first" key in the hash. As with built-in Perl hashes,
+ keys are fetched in an undefined order (which appears random). Takes
+ no arguments, returns the key as a scalar value.
+
+ my $key = $db->first_key();
+
+ * next_key()
+ Returns the "next" key in the hash, given the previous one as the
+ sole argument. Returns undef if there are no more keys to be
+ fetched.
+
+ $key = $db->next_key($key);
+
+ Here are some examples of using hashes:
+
+ my $db = new DBM::Deep "foo.db";
+
+ $db->put("foo", "bar");
+ print "foo: " . $db->get("foo") . "\n";
+
+ $db->put("baz", {}); # new child hash ref
+ $db->get("baz")->put("buz", "biz");
+ print "buz: " . $db->get("baz")->get("buz") . "\n";
+
+ my $key = $db->first_key();
+ while ($key) {
+ print "$key: " . $db->get($key) . "\n";
+ $key = $db->next_key($key);
+ }
+
+ if ($db->exists("foo")) { $db->delete("foo"); }
+
+ ARRAYS
+ For arrays, DBM::Deep supports all the common methods described above,
+ and the following additional methods: "length()", "push()", "pop()",
+ "shift()", "unshift()" and "splice()".
+
+ * length()
+ Returns the number of elements in the array. Takes no arguments.
+
+ my $len = $db->length();
+
+ * push()
+ Adds one or more elements onto the end of the array. Accepts
+ scalars, hash refs or array refs. No return value.
+
+ $db->push("foo", "bar", {});
+
+ * pop()
+ Fetches the last element in the array, and deletes it. Takes no
+ arguments. Returns undef if array is empty. Returns the element
+ value.
+
+ my $elem = $db->pop();
+
+ * shift()
+ Fetches the first element in the array, deletes it, then shifts all
+ the remaining elements over to take up the space. Returns the
+ element value. This method is not recommended with large arrays --
+ see "LARGE ARRAYS" below for details.
+
+ my $elem = $db->shift();
+
+ * unshift()
+ Inserts one or more elements onto the beginning of the array,
+ shifting all existing elements over to make room. Accepts scalars,
+ hash refs or array refs. No return value. This method is not
+ recommended with large arrays -- see <LARGE ARRAYS> below for
+ details.
+
+ $db->unshift("foo", "bar", {});
+
+ * splice()
+ Performs exactly like Perl's built-in function of the same name. See
+ "perldoc -f splice" for usage -- it is too complicated to document
+ here. This method is not recommended with large arrays -- see "LARGE
+ ARRAYS" below for details.
+
+ Here are some examples of using arrays:
+
+ my $db = new DBM::Deep(
+ file => "foo.db",
+ type => DBM::Deep::TYPE_ARRAY
+ );
+
+ $db->push("bar", "baz");
+ $db->unshift("foo");
+ $db->put(3, "buz");
+
+ my $len = $db->length();
+ print "length: $len\n"; # 4
+
+ for (my $k=0; $k<$len; $k++) {
+ print "$k: " . $db->get($k) . "\n";
+ }
+
+ $db->splice(1, 2, "biz", "baf");
+
+ while (my $elem = shift @$db) {
+ print "shifted: $elem\n";
+ }
+
+LOCKING
+ Enable automatic file locking by passing a true value to the "locking"
+ parameter when constructing your DBM::Deep object (see SETUP above).
+
+ my $db = new DBM::Deep(
+ file => "foo.db",
+ locking => 1
+ );
+
+ This causes Deep to "flock()" the underlying FileHandle object with
+ exclusive mode for writes, and shared mode for reads. This is required
+ if you have multiple processes accessing the same database file, to
+ avoid file corruption. Please note that "flock()" does NOT work for
+ files over NFS. See "DB OVER NFS" below for more.
+
+ EXPLICIT LOCKING
+ You can explicitly lock a database, so it remains locked for multiple
+ transactions. This is done by calling the "lock()" method, and passing
+ an optional lock mode argument (defaults to exclusive mode). This is
+ particularly useful for things like counters, where the current value
+ needs to be fetched, then incremented, then stored again.
+
+ $db->lock();
+ my $counter = $db->get("counter");
+ $counter++;
+ $db->put("counter", $counter);
+ $db->unlock();
+
+ # or...
+
+ $db->lock();
+ $db->{counter}++;
+ $db->unlock();
+
+ You can pass "lock()" an optional argument, which specifies which mode
+ to use (exclusive or shared). Use one of these two constants:
+ "DBM::Deep::LOCK_EX" or "DBM::Deep::LOCK_SH". These are passed directly
+ to "flock()", and are the same as the constants defined in Perl's
+ "Fcntl" module.
+
+ $db->lock( DBM::Deep::LOCK_SH );
+ # something here
+ $db->unlock();
+
+ If you want to implement your own file locking scheme, be sure to create
+ your DBM::Deep objects setting the "volatile" option to true. This hints
+ to Deep that the DB file may change between transactions. See "LOW-LEVEL
+ ACCESS" below for more.
+
+IMPORTING/EXPORTING
+ You can import existing complex structures by calling the "import()"
+ method, and export an entire database into an in-memory structure using
+ the "export()" method. Both are examined here.
+
+ IMPORTING
+ Say you have an existing hash with nested hashes/arrays inside it.
+ Instead of walking the structure and adding keys/elements to the
+ database as you go, simply pass a reference to the "import()" method.
+ This recursively adds everything to an existing DBM::Deep object for
+ you. Here is an example:
+
+ my $struct = {
+ key1 => "value1",
+ key2 => "value2",
+ array1 => [ "elem0", "elem1", "elem2" ],
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2"
+ }
+ };
+
+ my $db = new DBM::Deep "foo.db";
+ $db->import( $struct );
+
+ print $db->{key1} . "\n"; # prints "value1"
+
+ This recursively imports the entire $struct object into $db, including
+ all nested hashes and arrays. If the DBM::Deep object contains exsiting
+ data, keys are merged with the existing ones, replacing if they already
+ exist. The "import()" method can be called on any database level (not
+ just the base level), and works with both hash and array DB types.
+
+ Note: Make sure your existing structure has no circular references in
+ it. These will cause an infinite loop when importing.
+
+ EXPORTING
+ Calling the "export()" method on an existing DBM::Deep object will
+ return a reference to a new in-memory copy of the database. The export
+ is done recursively, so all nested hashes/arrays are all exported to
+ standard Perl objects. Here is an example:
+
+ my $db = new DBM::Deep "foo.db";
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ $db->{hash1} = {};
+ $db->{hash1}->{subkey1} = "subvalue1";
+ $db->{hash1}->{subkey2} = "subvalue2";
+
+ my $struct = $db->export();
+
+ print $struct->{key1} . "\n"; # prints "value1"
+
+ This makes a complete copy of the database in memory, and returns a
+ reference to it. The "export()" method can be called on any database
+ level (not just the base level), and works with both hash and array DB
+ types. Be careful of large databases -- you can store a lot more data in
+ a DBM::Deep object than an in-memory Perl structure.
+
+ Note: Make sure your database has no circular references in it. These
+ will cause an infinite loop when exporting.
+
+FILTERS
+ DBM::Deep has a number of hooks where you can specify your own Perl
+ function to perform filtering on incoming or outgoing data. This is a
+ perfect way to extend the engine, and implement things like real-time
+ compression or encryption. Filtering applies to the base DB level, and
+ all child hashes / arrays. Filter hooks can be specified when your
+ DBM::Deep object is first constructed, or by calling the "set_filter()"
+ method at any time. There are four available filter hooks, described
+ below:
+
+ * filter_store_key
+ This filter is called whenever a hash key is stored. It is passed
+ the incoming key, and expected to return a transformed key.
+
+ * filter_store_value
+ This filter is called whenever a hash key or array element is
+ stored. It is passed the incoming value, and expected to return a
+ transformed value.
+
+ * filter_fetch_key
+ This filter is called whenever a hash key is fetched (i.e. via
+ "first_key()" or "next_key()"). It is passed the transformed key,
+ and expected to return the plain key.
+
+ * filter_fetch_value
+ This filter is called whenever a hash key or array element is
+ fetched. It is passed the transformed value, and expected to return
+ the plain value.
+
+ Here are the two ways to setup a filter hook:
+
+ my $db = new DBM::Deep(
+ file => "foo.db",
+ filter_store_value => \&my_filter_store,
+ filter_fetch_value => \&my_filter_fetch
+ );
+
+ # or...
+
+ $db->set_filter( "filter_store_value", \&my_filter_store );
+ $db->set_filter( "filter_fetch_value", \&my_filter_fetch );
+
+ Your filter function will be called only when dealing with SCALAR keys
+ or values. When nested hashes and arrays are being stored/fetched,
+ filtering is bypassed. Filters are called as static functions, passed a
+ single SCALAR argument, and expected to return a single SCALAR value. If
+ you want to remove a filter, set the function reference to "undef":
+
+ $db->set_filter( "filter_store_value", undef );
+
+ REAL-TIME ENCRYPTION EXAMPLE
+ Here is a working example that uses the *Crypt::Blowfish* module to do
+ real-time encryption / decryption of keys & values with DBM::Deep
+ Filters. Please visit
+ <http://search.cpan.org/search?module=Crypt::Blowfish> for more on
+ *Crypt::Blowfish*. You'll also need the *Crypt::CBC* module.
+
+ use DBM::Deep;
+ use Crypt::Blowfish;
+ use Crypt::CBC;
+
+ my $cipher = new Crypt::CBC({
+ 'key' => 'my secret key',
+ 'cipher' => 'Blowfish',
+ 'iv' => '$KJh#(}q',
+ 'regenerate_key' => 0,
+ 'padding' => 'space',
+ 'prepend_iv' => 0
+ });
+
+ my $db = new DBM::Deep(
+ file => "foo-encrypt.db",
+ filter_store_key => \&my_encrypt,
+ filter_store_value => \&my_encrypt,
+ filter_fetch_key => \&my_decrypt,
+ filter_fetch_value => \&my_decrypt,
+ );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_encrypt {
+ return $cipher->encrypt( $_[0] );
+ }
+ sub my_decrypt {
+ return $cipher->decrypt( $_[0] );
+ }
+
+ REAL-TIME COMPRESSION EXAMPLE
+ Here is a working example that uses the *Compress::Zlib* module to do
+ real-time compression / decompression of keys & values with DBM::Deep
+ Filters. Please visit
+ <http://search.cpan.org/search?module=Compress::Zlib> for more on
+ *Compress::Zlib*.
+
+ use DBM::Deep;
+ use Compress::Zlib;
+
+ my $db = new DBM::Deep(
+ file => "foo-compress.db",
+ filter_store_key => \&my_compress,
+ filter_store_value => \&my_compress,
+ filter_fetch_key => \&my_decompress,
+ filter_fetch_value => \&my_decompress,
+ );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_compress {
+ return Compress::Zlib::memGzip( $_[0] ) ;
+ }
+ sub my_decompress {
+ return Compress::Zlib::memGunzip( $_[0] ) ;
+ }
+
+ Note: Filtering of keys only applies to hashes. Array "keys" are
+ actually numerical index numbers, and are not filtered.
+
+ERROR HANDLING
+ Most DBM::Deep methods return a true value for success, and call die()
+ on failure. You can wrap calls in an eval block to catch the die. Also,
+ the actual error message is stored in an internal scalar, which can be
+ fetched by calling the "error()" method.
+
+ my $db = new DBM::Deep "foo.db"; # create hash
+ eval { $db->push("foo"); }; # ILLEGAL -- push is array-only call
+
+ print $db->error(); # prints error message
+
+ You can then call "clear_error()" to clear the current error state.
+
+ $db->clear_error();
+
+ If you set the "debug" option to true when creating your DBM::Deep
+ object, all errors are considered NON-FATAL, and dumped to STDERR. This
+ is only for debugging purposes.
+
+LARGEFILE SUPPORT
+ If you have a 64-bit system, and your Perl is compiled with both
+ LARGEFILE and 64-bit support, you *may* be able to create databases
+ larger than 2 GB. DBM::Deep by default uses 32-bit file offset tags, but
+ these can be changed by calling the static "set_pack()" method before
+ you do anything else.
+
+ DBM::Deep::set_pack(8, 'Q');
+
+ This tells DBM::Deep to pack all file offsets with 8-byte (64-bit) quad
+ words instead of 32-bit longs. After setting these values your DB files
+ have a theoretical maximum size of 16 XB (exabytes).
+
+ Note: Changing these values will NOT work for existing database files.
+ Only change this for new files, and make sure it stays set consistently
+ throughout the file's life. If you do set these values, you can no
+ longer access 32-bit DB files. You can, however, call "set_pack(4, 'N')"
+ to change back to 32-bit mode.
+
+ Note: I have not personally tested files > 2 GB -- all my systems have
+ only a 32-bit Perl. However, I have received user reports that this does
+ indeed work!
+
+LOW-LEVEL ACCESS
+ If you require low-level access to the underlying FileHandle that Deep
+ uses, you can call the "fh()" method, which returns the handle:
+
+ my $fh = $db->fh();
+
+ This method can be called on the root level of the datbase, or any child
+ hashes or arrays. All levels share a *root* structure, which contains
+ things like the FileHandle, a reference counter, and all your options
+ you specified when you created the object. You can get access to this
+ root structure by calling the "root()" method.
+
+ my $root = $db->root();
+
+ This is useful for changing options after the object has already been
+ created, such as enabling/disabling locking, volatile or debug modes.
+ You can also store your own temporary user data in this structure (be
+ wary of name collision), which is then accessible from any child hash or
+ array.
+
+CUSTOM DIGEST ALGORITHM
+ DBM::Deep by default uses the *Message Digest 5* (MD5) algorithm for
+ hashing keys. However you can override this, and use another algorithm
+ (such as SHA-256) or even write your own. But please note that Deep
+ currently expects zero collisions, so your algorithm has to be
+ *perfect*, so to speak. Collision detection may be introduced in a later
+ version.
+
+ You can specify a custom digest algorithm by calling the static
+ "set_digest()" function, passing a reference to a subroutine, and the
+ length of the algorithm's hashes (in bytes). This is a global static
+ function, which affects ALL Deep objects. Here is a working example that
+ uses a 256-bit hash from the *Digest::SHA256* module. Please see
+ <http://search.cpan.org/search?module=Digest::SHA256> for more.
+
+ use DBM::Deep;
+ use Digest::SHA256;
+
+ my $context = Digest::SHA256::new(256);
+
+ DBM::Deep::set_digest( \&my_digest, 32 );
+
+ my $db = new DBM::Deep "foo-sha.db";
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_digest {
+ return substr( $context->hash($_[0]), 0, 32 );
+ }
+
+ Note: Your returned digest strings must be EXACTLY the number of bytes
+ you specify in the "set_digest()" function (in this case 32).
+
+CIRCULAR REFERENCES
+ DBM::Deep has experimental support for circular references. Meaning you
+ can have a nested hash key or array element that points to a parent
+ object. This relationship is stored in the DB file, and is preserved
+ between sessions. Here is an example:
+
+ my $db = new DBM::Deep "foo.db";
+
+ $db->{foo} = "bar";
+ $db->{circle} = $db; # ref to self
+
+ print $db->{foo} . "\n"; # prints "foo"
+ print $db->{circle}->{foo} . "\n"; # prints "foo" again
+
+ One catch is, passing the object to a function that recursively walks
+ the object tree (such as *Data::Dumper* or even the built-in
+ "optimize()" or "export()" methods) will result in an infinite loop. The
+ other catch is, if you fetch the *key* of a circular reference (i.e.
+ using the "first_key()" or "next_key()" methods), you will get the
+ *target object's key*, not the ref's key. This gets even more
+ interesting with the above example, where the *circle* key points to the
+ base DB object, which technically doesn't have a key. So I made
+ DBM::Deep return "[base]" as the key name in that special case.
+
+CAVEATS / ISSUES / BUGS
+ This section describes all the known issues with DBM::Deep. It you have
+ found something that is not listed here, please send e-mail to
+ jhuckaby@cpan.org.
+
+ UNUSED SPACE RECOVERY
+ One major caveat with Deep is that space occupied by existing keys and
+ values is not recovered when they are deleted. Meaning if you keep
+ deleting and adding new keys, your file will continuously grow. I am
+ working on this, but in the meantime you can call the built-in
+ "optimize()" method from time to time (perhaps in a crontab or
+ something) to recover all your unused space.
+
+ $db->optimize(); # returns true on success
+
+ This rebuilds the ENTIRE database into a new file, then moves it on top
+ of the original. The new file will have no unused space, thus it will
+ take up as little disk space as possible. Please note that this
+ operation can take a long time for large files, and you need enough disk
+ space to temporarily hold 2 copies of your DB file. The temporary file
+ is created in the same directory as the original, named with a ".tmp"
+ extension, and is deleted when the operation completes. Oh, and if
+ locking is enabled, the DB is automatically locked for the entire
+ duration of the copy.
+
+ WARNING: Only call optimize() on the top-level node of the database, and
+ make sure there are no child references lying around. Deep keeps a
+ reference counter, and if it is greater than 1, optimize() will abort
+ and return undef.
+
+ AUTOVIVIFICATION
+ Unfortunately, autovivification doesn't work with tied hashes. This
+ appears to be a bug in Perl's tie() system, as *Jakob Schmidt*
+ encountered the very same issue with his *DWH_FIle* module (see
+ <http://search.cpan.org/search?module=DWH_File>), and it is also
+ mentioned in the BUGS section for the *MLDBM* module <see
+ <http://search.cpan.org/search?module=MLDBM>). Basically, on a new db
+ file, this does not work:
+
+ $db->{foo}->{bar} = "hello";
+
+ Since "foo" doesn't exist, you cannot add "bar" to it. You end up with
+ "foo" being an empty hash. Try this instead, which works fine:
+
+ $db->{foo} = { bar => "hello" };
+
+ As of Perl 5.8.7, this bug still exists. I have walked very carefully
+ through the execution path, and Perl indeed passes an empty hash to the
+ STORE() method. Probably a bug in Perl.
+
+ FILE CORRUPTION
+ The current level of error handling in Deep is minimal. Files *are*
+ checked for a 32-bit signature on open(), but other corruption in files
+ can cause segmentation faults. Deep may try to seek() past the end of a
+ file, or get stuck in an infinite loop depending on the level of
+ corruption. File write operations are not checked for failure (for
+ speed), so if you happen to run out of disk space, Deep will probably
+ fail in a bad way. These things will be addressed in a later version of
+ DBM::Deep.
+
+ DB OVER NFS
+ Beware of using DB files over NFS. Deep uses flock(), which works well
+ on local filesystems, but will NOT protect you from file corruption over
+ NFS. I've heard about setting up your NFS server with a locking daemon,
+ then using lockf() to lock your files, but your milage may vary there as
+ well. From what I understand, there is no real way to do it. However, if
+ you need access to the underlying FileHandle in Deep for using some
+ other kind of locking scheme like lockf(), see the "LOW-LEVEL ACCESS"
+ section above.
+
+ COPYING OBJECTS
+ Beware of copying tied objects in Perl. Very strange things can happen.
+ Instead, use Deep's "clone()" method which safely copies the object and
+ returns a new, blessed, tied hash or array to the same level in the DB.
+
+ my $copy = $db->clone();
+
+ LARGE ARRAYS
+ Beware of using "shift()", "unshift()" or "splice()" with large arrays.
+ These functions cause every element in the array to move, which can be
+ murder on DBM::Deep, as every element has to be fetched from disk, then
+ stored again in a different location. This may be addressed in a later
+ version.
+
+PERFORMANCE
+ This section discusses DBM::Deep's speed and memory usage.
+
+ SPEED
+ Obviously, DBM::Deep isn't going to be as fast as some C-based DBMs,
+ such as the almighty *BerkeleyDB*. But it makes up for it in features
+ like true multi-level hash/array support, and cross-platform FTPable
+ files. Even so, DBM::Deep is still pretty fast, and the speed stays
+ fairly consistent, even with huge databases. Here is some test data:
+
+ Adding 1,000,000 keys to new DB file...
+
+ At 100 keys, avg. speed is 2,703 keys/sec
+ At 200 keys, avg. speed is 2,642 keys/sec
+ At 300 keys, avg. speed is 2,598 keys/sec
+ At 400 keys, avg. speed is 2,578 keys/sec
+ At 500 keys, avg. speed is 2,722 keys/sec
+ At 600 keys, avg. speed is 2,628 keys/sec
+ At 700 keys, avg. speed is 2,700 keys/sec
+ At 800 keys, avg. speed is 2,607 keys/sec
+ At 900 keys, avg. speed is 2,190 keys/sec
+ At 1,000 keys, avg. speed is 2,570 keys/sec
+ At 2,000 keys, avg. speed is 2,417 keys/sec
+ At 3,000 keys, avg. speed is 1,982 keys/sec
+ At 4,000 keys, avg. speed is 1,568 keys/sec
+ At 5,000 keys, avg. speed is 1,533 keys/sec
+ At 6,000 keys, avg. speed is 1,787 keys/sec
+ At 7,000 keys, avg. speed is 1,977 keys/sec
+ At 8,000 keys, avg. speed is 2,028 keys/sec
+ At 9,000 keys, avg. speed is 2,077 keys/sec
+ At 10,000 keys, avg. speed is 2,031 keys/sec
+ At 20,000 keys, avg. speed is 1,970 keys/sec
+ At 30,000 keys, avg. speed is 2,050 keys/sec
+ At 40,000 keys, avg. speed is 2,073 keys/sec
+ At 50,000 keys, avg. speed is 1,973 keys/sec
+ At 60,000 keys, avg. speed is 1,914 keys/sec
+ At 70,000 keys, avg. speed is 2,091 keys/sec
+ At 80,000 keys, avg. speed is 2,103 keys/sec
+ At 90,000 keys, avg. speed is 1,886 keys/sec
+ At 100,000 keys, avg. speed is 1,970 keys/sec
+ At 200,000 keys, avg. speed is 2,053 keys/sec
+ At 300,000 keys, avg. speed is 1,697 keys/sec
+ At 400,000 keys, avg. speed is 1,838 keys/sec
+ At 500,000 keys, avg. speed is 1,941 keys/sec
+ At 600,000 keys, avg. speed is 1,930 keys/sec
+ At 700,000 keys, avg. speed is 1,735 keys/sec
+ At 800,000 keys, avg. speed is 1,795 keys/sec
+ At 900,000 keys, avg. speed is 1,221 keys/sec
+ At 1,000,000 keys, avg. speed is 1,077 keys/sec
+
+ This test was performed on a PowerMac G4 1gHz running Mac OS X 10.3.2 &
+ Perl 5.8.1, with an 80GB Ultra ATA/100 HD spinning at 7200RPM. The hash
+ keys and values were between 6 - 12 chars in length. The DB file ended
+ up at 210MB. Run time was 12 min 3 sec.
+
+ MEMORY USAGE
+ One of the great things about DBM::Deep is that it uses very little
+ memory. Even with huge databases (1,000,000+ keys) you will not see much
+ increased memory on your process. Deep relies solely on the filesystem
+ for storing and fetching data. Here is output from */usr/bin/top* before
+ even opening a database handle:
+
+ PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
+ 22831 root 11 0 2716 2716 1296 R 0.0 0.2 0:07 perl
+
+ Basically the process is taking 2,716K of memory. And here is the same
+ process after storing and fetching 1,000,000 keys:
+
+ PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
+ 22831 root 14 0 2772 2772 1328 R 0.0 0.2 13:32 perl
+
+ Notice the memory usage increased by only 56K. Test was performed on a
+ 700mHz x86 box running Linux RedHat 7.2 & Perl 5.6.1.
+
+DB FILE FORMAT
+ In case you were interested in the underlying DB file format, it is
+ documented here in this section. You don't need to know this to use the
+ module, it's just included for reference.
+
+ SIGNATURE
+ DBM::Deep files always start with a 32-bit signature to identify the
+ file type. This is at offset 0. The signature is "DPDB" in network byte
+ order. This is checked upon each file open().
+
+ TAG
+ The DBM::Deep file is in a *tagged format*, meaning each section of the
+ file has a standard header containing the type of data, the length of
+ data, and then the data itself. The type is a single character (1 byte),
+ the length is a 32-bit unsigned long in network byte order, and the data
+ is, well, the data. Here is how it unfolds:
+
+ MASTER INDEX
+ Immediately after the 32-bit file signature is the *Master Index*
+ record. This is a standard tag header followed by 1024 bytes (in 32-bit
+ mode) or 2048 bytes (in 64-bit mode) of data. The type is *H* for hash
+ or *A* for array, depending on how the DBM::Deep object was constructed.
+
+ The index works by looking at a *MD5 Hash* of the hash key (or array
+ index number). The first 8-bit char of the MD5 signature is the offset
+ into the index, multipled by 4 in 32-bit mode, or 8 in 64-bit mode. The
+ value of the index element is a file offset of the next tag for the
+ key/element in question, which is usually a *Bucket List* tag (see
+ below).
+
+ The next tag *could* be another index, depending on how many
+ keys/elements exist. See RE-INDEXING below for details.
+
+ BUCKET LIST
+ A *Bucket List* is a collection of 16 MD5 hashes for keys/elements, plus
+ file offsets to where the actual data is stored. It starts with a
+ standard tag header, with type *B*, and a data size of 320 bytes in
+ 32-bit mode, or 384 bytes in 64-bit mode. Each MD5 hash is stored in
+ full (16 bytes), plus the 32-bit or 64-bit file offset for the *Bucket*
+ containing the actual data. When the list fills up, a *Re-Index*
+ operation is performed (See RE-INDEXING below).
+
+ BUCKET
+ A *Bucket* is a tag containing a key/value pair (in hash mode), or a
+ index/value pair (in array mode). It starts with a standard tag header
+ with type *D* for scalar data (string, binary, etc.), or it could be a
+ nested hash (type *H*) or array (type *A*). The value comes just after
+ the tag header. The size reported in the tag header is only for the
+ value, but then, just after the value is another size (32-bit unsigned
+ long) and then the plain key itself. Since the value is likely to be
+ fetched more often than the plain key, I figured it would be *slightly*
+ faster to store the value first.
+
+ If the type is *H* (hash) or *A* (array), the value is another *Master
+ Index* record for the nested structure, where the process begins all
+ over again.
+
+ RE-INDEXING
+ After a *Bucket List* grows to 16 records, its allocated space in the
+ file is exhausted. Then, when another key/element comes in, the list is
+ converted to a new index record. However, this index will look at the
+ next char in the MD5 hash, and arrange new Bucket List pointers
+ accordingly. This process is called *Re-Indexing*. Basically, a new
+ index tag is created at the file EOF, and all 17 (16 + new one)
+ keys/elements are removed from the old Bucket List and inserted into the
+ new index. Several new Bucket Lists are created in the process, as a new
+ MD5 char from the key is being examined (it is unlikely that the keys
+ will all share the same next char of their MD5s).
+
+ Because of the way the *MD5* algorithm works, it is impossible to tell
+ exactly when the Bucket Lists will turn into indexes, but the first
+ round tends to happen right around 4,000 keys. You will see a *slight*
+ decrease in performance here, but it picks back up pretty quick (see
+ SPEED above). Then it takes a lot more keys to exhaust the next level of
+ Bucket Lists. It's right around 900,000 keys. This process can continue
+ nearly indefinitely -- right up until the point the *MD5* signatures
+ start colliding with each other, and this is EXTREMELY rare -- like
+ winning the lottery 5 times in a row AND getting struck by lightning
+ while you are walking to cash in your tickets. Theoretically, since
+ *MD5* hashes are 128-bit values, you *could* have up to
+ 340,282,366,921,000,000,000,000,000,000,000,000,000 keys/elements (I
+ believe this is 340 unodecillion, but don't quote me).
+
+ STORING
+ When a new key/element is stored, the key (or index number) is first ran
+ through *Digest::MD5* to get a 128-bit signature (example, in hex:
+ b05783b0773d894396d475ced9d2f4f6). Then, the *Master Index* record is
+ checked for the first char of the signature (in this case *b*). If it
+ does not exist, a new *Bucket List* is created for our key (and the next
+ 15 future keys that happen to also have *b* as their first MD5 char).
+ The entire MD5 is written to the *Bucket List* along with the offset of
+ the new *Bucket* record (EOF at this point, unless we are replacing an
+ existing *Bucket*), where the actual data will be stored.
+
+ FETCHING
+ Fetching an existing key/element involves getting a *Digest::MD5* of the
+ key (or index number), then walking along the indexes. If there are
+ enough keys/elements in this DB level, there might be nested indexes,
+ each linked to a particular char of the MD5. Finally, a *Bucket List* is
+ pointed to, which contains up to 16 full MD5 hashes. Each is checked for
+ equality to the key in question. If we found a match, the *Bucket* tag
+ is loaded, where the value and plain key are stored.
+
+ Fetching the plain key occurs when calling the *first_key()* and
+ *next_key()* methods. In this process the indexes are walked
+ systematically, and each key fetched in increasing MD5 order (which is
+ why it appears random). Once the *Bucket* is found, the value is skipped
+ the plain key returned instead. Note: Do not count on keys being fetched
+ as if the MD5 hashes were alphabetically sorted. This only happens on an
+ index-level -- as soon as the *Bucket Lists* are hit, the keys will come
+ out in the order they went in -- so it's pretty much undefined how the
+ keys will come out -- just like Perl's built-in hashes.
+
+AUTHOR
+ Joseph Huckaby, jhuckaby@cpan.org
+
+ Special thanks to Adam Sah and Rich Gaushell! You know why :-)
+
+SEE ALSO
+ perltie(1), Tie::Hash(3), Digest::MD5(3), Fcntl(3), flock(2), lockf(3),
+ nfs(5), Digest::SHA256(3), Crypt::Blowfish(3), Compress::Zlib(3)
+
+LICENSE
+ Copyright (c) 2002-2005 Joseph Huckaby. All Rights Reserved. This is
+ free software, you may use it and distribute it under the same terms as
+ Perl itself.
+
--- /dev/null
+package DBM::Deep;
+
+##
+# DBM::Deep
+#
+# Description:
+# Multi-level database module for storing hash trees, arrays and simple
+# key/value pairs into FTP-able, cross-platform binary database files.
+#
+# Type `perldoc DBM::Deep` for complete documentation.
+#
+# Usage Examples:
+# my %db;
+# tie %db, 'DBM::Deep', 'my_database.db'; # standard tie() method
+#
+# my $db = new DBM::Deep( 'my_database.db' ); # preferred OO method
+#
+# $db->{my_scalar} = 'hello world';
+# $db->{my_hash} = { larry => 'genius', hashes => 'fast' };
+# $db->{my_array} = [ 1, 2, 3, time() ];
+# $db->{my_complex} = [ 'hello', { perl => 'rules' }, 42, 99 ];
+# push @{$db->{my_array}}, 'another value';
+# my @key_list = keys %{$db->{my_hash}};
+# print "This module " . $db->{my_complex}->[1]->{perl} . "!\n";
+#
+# Copyright:
+# (c) 2002-2005 Joseph Huckaby. All Rights Reserved.
+# This program is free software; you can redistribute it and/or
+# modify it under the same terms as Perl itself.
+##
+
+use strict;
+use FileHandle;
+use Fcntl qw/:flock/;
+use Digest::MD5 ();
+use Scalar::Util ();
+use vars qw/$VERSION/;
+
+$VERSION = "0.96";
+
+##
+# Set to 4 and 'N' for 32-bit offset tags (default). Theoretical limit of 4 GB per file.
+# (Perl must be compiled with largefile support for files > 2 GB)
+#
+# Set to 8 and 'Q' for 64-bit offsets. Theoretical limit of 16 XB per file.
+# (Perl must be compiled with largefile and 64-bit long support)
+##
+#my $LONG_SIZE = 4;
+#my $LONG_PACK = 'N';
+
+##
+# Set to 4 and 'N' for 32-bit data length prefixes. Limit of 4 GB for each key/value.
+# Upgrading this is possible (see above) but probably not necessary. If you need
+# more than 4 GB for a single key or value, this module is really not for you :-)
+##
+#my $DATA_LENGTH_SIZE = 4;
+#my $DATA_LENGTH_PACK = 'N';
+my ($LONG_SIZE, $LONG_PACK, $DATA_LENGTH_SIZE, $DATA_LENGTH_PACK);
+
+##
+# Maximum number of buckets per list before another level of indexing is done.
+# Increase this value for slightly greater speed, but larger database files.
+# DO NOT decrease this value below 16, due to risk of recursive reindex overrun.
+##
+my $MAX_BUCKETS = 16;
+
+##
+# Better not adjust anything below here, unless you're me :-)
+##
+
+##
+# Setup digest function for keys
+##
+my ($DIGEST_FUNC, $HASH_SIZE);
+#my $DIGEST_FUNC = \&Digest::MD5::md5;
+
+##
+# Precalculate index and bucket sizes based on values above.
+##
+#my $HASH_SIZE = 16;
+my ($INDEX_SIZE, $BUCKET_SIZE, $BUCKET_LIST_SIZE);
+
+set_digest();
+#set_pack();
+#precalc_sizes();
+
+##
+# Setup file and tag signatures. These should never change.
+##
+sub SIG_FILE () { 'DPDB' }
+sub SIG_HASH () { 'H' }
+sub SIG_ARRAY () { 'A' }
+sub SIG_NULL () { 'N' }
+sub SIG_DATA () { 'D' }
+sub SIG_INDEX () { 'I' }
+sub SIG_BLIST () { 'B' }
+sub SIG_SIZE () { 1 }
+
+##
+# Setup constants for users to pass to new()
+##
+sub TYPE_HASH () { return SIG_HASH; }
+sub TYPE_ARRAY () { return SIG_ARRAY; }
+
+sub new {
+ ##
+ # Class constructor method for Perl OO interface.
+ # Calls tie() and returns blessed reference to tied hash or array,
+ # providing a hybrid OO/tie interface.
+ ##
+ my $class = shift;
+ my $args;
+ if (scalar(@_) > 1) { $args = {@_}; }
+ else { $args = { file => shift }; }
+
+ ##
+ # Check if we want a tied hash or array.
+ ##
+ my $self;
+ if (defined($args->{type}) && $args->{type} eq TYPE_ARRAY) {
+ tie @$self, $class, %$args;
+ }
+ else {
+ tie %$self, $class, %$args;
+ }
+
+ return bless $self, $class;
+}
+
+{
+ my @outer_params = qw( type base_offset );
+ sub init {
+ ##
+ # Setup $self and bless into this class.
+ ##
+ my $class = shift;
+ my $args = shift;
+
+ my $self = {
+ type => TYPE_HASH,
+ base_offset => length(SIG_FILE),
+ root => {
+ file => undef,
+ fh => undef,
+ end => 0,
+ links => 0,
+ autoflush => undef,
+ locking => undef,
+ volatile => undef,
+ debug => undef,
+ mode => 'r+',
+ filter_store_key => undef,
+ filter_store_value => undef,
+ filter_fetch_key => undef,
+ filter_fetch_value => undef,
+ autobless => undef,
+ locked => 0,
+ %$args,
+ },
+ };
+
+ bless $self, $class;
+
+ foreach my $outer_parm ( @outer_params ) {
+ next unless exists $args->{$outer_parm};
+ $self->{$outer_parm} = $args->{$outer_parm}
+ }
+
+ if ( exists $args->{root} ) {
+ $self->{root} = $args->{root};
+ }
+ else {
+ # This is cleanup based on the fact that the $args
+ # coming in is for both the root and non-root items
+ delete $self->root->{$_} for @outer_params;
+ }
+ $self->root->{links}++;
+
+ if (!defined($self->fh)) { $self->open(); }
+
+ return $self;
+ }
+}
+
+sub _get_self { tied( %{$_[0]} ) || $_[0] }
+
+sub TIEHASH {
+ ##
+ # Tied hash constructor method, called by Perl's tie() function.
+ ##
+ my $class = shift;
+ my $args;
+ if (scalar(@_) > 1) { $args = {@_}; }
+ #XXX This use of ref() is bad and is a bug
+ elsif (ref($_[0])) { $args = $_[0]; }
+ else { $args = { file => shift }; }
+
+ return $class->init($args);
+}
+
+sub TIEARRAY {
+##
+# Tied array constructor method, called by Perl's tie() function.
+##
+my $class = shift;
+my $args;
+if (scalar(@_) > 1) { $args = {@_}; }
+ #XXX This use of ref() is bad and is a bug
+ elsif (ref($_[0])) { $args = $_[0]; }
+ else { $args = { file => shift }; }
+
+ return $class->init($args);
+}
+
+sub DESTROY {
+ ##
+ # Class deconstructor. Close file handle if there are no more refs.
+ ##
+ my $self = _get_self($_[0]);
+ return unless $self;
+
+ $self->root->{links}--;
+
+ if (!$self->root->{links}) {
+ $self->close();
+ }
+}
+
+sub open {
+ ##
+ # Open a FileHandle to the database, create if nonexistent.
+ # Make sure file signature matches DeepDB spec.
+ ##
+ my $self = _get_self($_[0]);
+
+ if (defined($self->fh)) { $self->close(); }
+
+ if (!(-e $self->root->{file}) && $self->root->{mode} eq 'r+') {
+ my $temp = FileHandle->new( $self->root->{file}, 'w' );
+ undef $temp;
+ }
+
+ #XXX Convert to set_fh()
+ $self->root->{fh} = FileHandle->new( $self->root->{file}, $self->root->{mode} );
+ if (! defined($self->fh)) {
+ return $self->throw_error("Cannot open file: " . $self->root->{file} . ": $!");
+ }
+
+ binmode $self->fh; # for win32
+ if ($self->root->{autoflush}) {
+ $self->fh->autoflush();
+ }
+
+ my $signature;
+ seek($self->fh, 0, 0);
+ my $bytes_read = $self->fh->read($signature, length(SIG_FILE));
+
+ ##
+ # File is empty -- write signature and master index
+ ##
+ if (!$bytes_read) {
+ seek($self->fh, 0, 0);
+ $self->fh->print(SIG_FILE);
+ $self->root->{end} = length(SIG_FILE);
+ $self->create_tag($self->base_offset, $self->type, chr(0) x $INDEX_SIZE);
+
+ my $plain_key = "[base]";
+ $self->fh->print( pack($DATA_LENGTH_PACK, length($plain_key)) . $plain_key );
+ $self->root->{end} += $DATA_LENGTH_SIZE + length($plain_key);
+ $self->fh->flush();
+
+ return 1;
+ }
+
+ ##
+ # Check signature was valid
+ ##
+ unless ($signature eq SIG_FILE) {
+ $self->close();
+ return $self->throw_error("Signature not found -- file is not a Deep DB");
+ }
+
+ $self->root->{end} = (stat($self->fh))[7];
+
+ ##
+ # Get our type from master index signature
+ ##
+ my $tag = $self->load_tag($self->base_offset);
+#XXX This is a problem - need to verify type, not override it!
+#XXX We probably also want to store the hash algorithm name, not assume anything
+#XXX Convert to set_type() when one is written
+ $self->{type} = $tag->{signature};
+
+ return 1;
+}
+
+sub close {
+ ##
+ # Close database FileHandle
+ ##
+ my $self = _get_self($_[0]);
+ undef $self->root->{fh};
+}
+
+sub create_tag {
+ ##
+ # Given offset, signature and content, create tag and write to disk
+ ##
+ my ($self, $offset, $sig, $content) = @_;
+ my $size = length($content);
+
+ seek($self->fh, $offset, 0);
+ $self->fh->print( $sig . pack($DATA_LENGTH_PACK, $size) . $content );
+
+ if ($offset == $self->root->{end}) {
+ $self->root->{end} += SIG_SIZE + $DATA_LENGTH_SIZE + $size;
+ }
+
+ return {
+ signature => $sig,
+ size => $size,
+ offset => $offset + SIG_SIZE + $DATA_LENGTH_SIZE,
+ content => $content
+ };
+}
+
+sub load_tag {
+ ##
+ # Given offset, load single tag and return signature, size and data
+ ##
+ my $self = shift;
+ my $offset = shift;
+
+ seek($self->fh, $offset, 0);
+ if ($self->fh->eof()) { return; }
+
+ my $sig;
+ $self->fh->read($sig, SIG_SIZE);
+
+ my $size;
+ $self->fh->read($size, $DATA_LENGTH_SIZE);
+ $size = unpack($DATA_LENGTH_PACK, $size);
+
+ my $buffer;
+ $self->fh->read($buffer, $size);
+
+ return {
+ signature => $sig,
+ size => $size,
+ offset => $offset + SIG_SIZE + $DATA_LENGTH_SIZE,
+ content => $buffer
+ };
+}
+
+sub index_lookup {
+ ##
+ # Given index tag, lookup single entry in index and return .
+ ##
+ my $self = shift;
+ my ($tag, $index) = @_;
+
+ my $location = unpack($LONG_PACK, substr($tag->{content}, $index * $LONG_SIZE, $LONG_SIZE) );
+ if (!$location) { return; }
+
+ return $self->load_tag( $location );
+}
+
+sub add_bucket {
+ ##
+ # Adds one key/value pair to bucket list, given offset, MD5 digest of key,
+ # plain (undigested) key and value.
+ ##
+ my $self = shift;
+ my ($tag, $md5, $plain_key, $value) = @_;
+ my $keys = $tag->{content};
+ my $location = 0;
+ my $result = 2;
+
+ my $is_dbm_deep = eval { $value->isa( 'DBM::Deep' ) };
+ my $internal_ref = $is_dbm_deep && ($value->root eq $self->root);
+
+ ##
+ # Iterate through buckets, seeing if this is a new entry or a replace.
+ ##
+ for (my $i=0; $i<$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ my $subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+ if (!$subloc) {
+ ##
+ # Found empty bucket (end of list). Populate and exit loop.
+ ##
+ $result = 2;
+
+ if ($internal_ref) { $location = $value->base_offset; }
+ else { $location = $self->root->{end}; }
+
+ seek($self->fh, $tag->{offset} + ($i * $BUCKET_SIZE), 0);
+ $self->fh->print( $md5 . pack($LONG_PACK, $location) );
+ last;
+ }
+ elsif ($md5 eq $key) {
+ ##
+ # Found existing bucket with same key. Replace with new value.
+ ##
+ $result = 1;
+
+ if ($internal_ref) {
+ $location = $value->base_offset;
+ seek($self->fh, $tag->{offset} + ($i * $BUCKET_SIZE), 0);
+ $self->fh->print( $md5 . pack($LONG_PACK, $location) );
+ }
+ else {
+ seek($self->fh, $subloc + SIG_SIZE, 0);
+ my $size;
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+
+ ##
+ # If value is a hash, array, or raw value with equal or less size, we can
+ # reuse the same content area of the database. Otherwise, we have to create
+ # a new content area at the EOF.
+ ##
+ my $actual_length;
+ my $r = Scalar::Util::reftype( $value ) || '';
+ if ( $r eq 'HASH' || $r eq 'ARRAY' ) { $actual_length = $INDEX_SIZE; }
+ else { $actual_length = length($value); }
+
+ if ($actual_length <= $size) {
+ $location = $subloc;
+ }
+ else {
+ $location = $self->root->{end};
+ seek($self->fh, $tag->{offset} + ($i * $BUCKET_SIZE) + $HASH_SIZE, 0);
+ $self->fh->print( pack($LONG_PACK, $location) );
+ }
+ }
+ last;
+ }
+ } # i loop
+
+ ##
+ # If this is an internal reference, return now.
+ # No need to write value or plain key
+ ##
+#YYY
+ if ($internal_ref) { return $result; }
+
+ ##
+ # If bucket didn't fit into list, split into a new index level
+ ##
+ if (!$location) {
+ seek($self->fh, $tag->{ref_loc}, 0);
+ $self->fh->print( pack($LONG_PACK, $self->root->{end}) );
+
+ my $index_tag = $self->create_tag($self->root->{end}, SIG_INDEX, chr(0) x $INDEX_SIZE);
+ my @offsets = ();
+
+#XXX We've already guaranteed that this cannot be true at YYY
+# if ($internal_ref) {
+# $keys .= $md5 . pack($LONG_PACK, $value->base_offset);
+# $location = $value->base_offset;
+# }
+# else { $keys .= $md5 . pack($LONG_PACK, 0); }
+ $keys .= $md5 . pack($LONG_PACK, 0);
+
+ for (my $i=0; $i<=$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ if ($key) {
+ my $old_subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+ my $num = ord(substr($key, $tag->{ch} + 1, 1));
+
+ if ($offsets[$num]) {
+ my $offset = $offsets[$num] + SIG_SIZE + $DATA_LENGTH_SIZE;
+ seek($self->fh, $offset, 0);
+ my $subkeys;
+ $self->fh->read($subkeys, $BUCKET_LIST_SIZE);
+
+ for (my $k=0; $k<$MAX_BUCKETS; $k++) {
+ my $subloc = unpack($LONG_PACK, substr($subkeys, ($k * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+ if (!$subloc) {
+ seek($self->fh, $offset + ($k * $BUCKET_SIZE), 0);
+ $self->fh->print( $key . pack($LONG_PACK, $old_subloc || $self->root->{end}) );
+ last;
+ }
+ } # k loop
+ }
+ else {
+ $offsets[$num] = $self->root->{end};
+ seek($self->fh, $index_tag->{offset} + ($num * $LONG_SIZE), 0);
+ $self->fh->print( pack($LONG_PACK, $self->root->{end}) );
+
+ my $blist_tag = $self->create_tag($self->root->{end}, SIG_BLIST, chr(0) x $BUCKET_LIST_SIZE);
+
+ seek($self->fh, $blist_tag->{offset}, 0);
+ $self->fh->print( $key . pack($LONG_PACK, $old_subloc || $self->root->{end}) );
+ }
+ } # key is real
+ } # i loop
+
+ $location ||= $self->root->{end};
+ } # re-index bucket list
+
+ ##
+ # Seek to content area and store signature, value and plaintext key
+ ##
+ if ($location) {
+ my $content_length;
+ seek($self->fh, $location, 0);
+
+ ##
+ # Write signature based on content type, set content length and write actual value.
+ ##
+ my $r = Scalar::Util::reftype($value) || '';
+ if ($r eq 'HASH') {
+ $self->fh->print( TYPE_HASH );
+ $self->fh->print( pack($DATA_LENGTH_PACK, $INDEX_SIZE) . chr(0) x $INDEX_SIZE );
+ $content_length = $INDEX_SIZE;
+ }
+ elsif ($r eq 'ARRAY') {
+ $self->fh->print( TYPE_ARRAY );
+ $self->fh->print( pack($DATA_LENGTH_PACK, $INDEX_SIZE) . chr(0) x $INDEX_SIZE );
+ $content_length = $INDEX_SIZE;
+ }
+ elsif (!defined($value)) {
+ $self->fh->print( SIG_NULL );
+ $self->fh->print( pack($DATA_LENGTH_PACK, 0) );
+ $content_length = 0;
+ }
+ else {
+ $self->fh->print( SIG_DATA );
+ $self->fh->print( pack($DATA_LENGTH_PACK, length($value)) . $value );
+ $content_length = length($value);
+ }
+
+ ##
+ # Plain key is stored AFTER value, as keys are typically fetched less often.
+ ##
+ $self->fh->print( pack($DATA_LENGTH_PACK, length($plain_key)) . $plain_key );
+
+ ##
+ # If value is blessed, preserve class name
+ ##
+ my $value_class = Scalar::Util::blessed($value);
+#XXX NO tests for this
+ if ($self->root->{autobless} && defined $value_class) {
+ if ($value_class ne 'DBM::Deep') {
+ ##
+ # Blessed ref -- will restore later
+ ##
+ $self->fh->print( chr(1) );
+ $self->fh->print( pack($DATA_LENGTH_PACK, length($value_class)) . $value_class );
+ $content_length += 1;
+ $content_length += $DATA_LENGTH_SIZE + length($value_class);
+ }
+ else {
+ ##
+ # Simple unblessed ref -- no restore needed
+ ##
+ $self->fh->print( chr(0) );
+ $content_length += 1;
+ }
+ }
+
+ ##
+ # If this is a new content area, advance EOF counter
+ ##
+ if ($location == $self->root->{end}) {
+ $self->root->{end} += SIG_SIZE;
+ $self->root->{end} += $DATA_LENGTH_SIZE + $content_length;
+ $self->root->{end} += $DATA_LENGTH_SIZE + length($plain_key);
+ }
+
+ ##
+ # If content is a hash or array, create new child DeepDB object and
+ # pass each key or element to it.
+ ##
+ if ($r eq 'HASH') {
+ my $branch = DBM::Deep->new(
+ type => TYPE_HASH,
+ base_offset => $location,
+ root => $self->root,
+ );
+ foreach my $key (keys %{$value}) {
+ $branch->{$key} = $value->{$key};
+ }
+ }
+ elsif ($r eq 'ARRAY') {
+ my $branch = DBM::Deep->new(
+ type => TYPE_ARRAY,
+ base_offset => $location,
+ root => $self->root,
+ );
+ my $index = 0;
+ foreach my $element (@{$value}) {
+ $branch->[$index] = $element;
+ $index++;
+ }
+ }
+
+ return $result;
+ }
+
+ return $self->throw_error("Fatal error: indexing failed -- possibly due to corruption in file");
+}
+
+sub get_bucket_value {
+ ##
+ # Fetch single value given tag and MD5 digested key.
+ ##
+ my $self = shift;
+ my ($tag, $md5) = @_;
+ my $keys = $tag->{content};
+
+ ##
+ # Iterate through buckets, looking for a key match
+ ##
+ BUCKET:
+ for (my $i=0; $i<$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ my $subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+
+ if (!$subloc) {
+ ##
+ # Hit end of list, no match
+ ##
+ return;
+ }
+
+ if ( $md5 ne $key ) {
+ next BUCKET;
+ }
+
+ ##
+ # Found match -- seek to offset and read signature
+ ##
+ my $signature;
+ seek($self->fh, $subloc, 0);
+ $self->fh->read($signature, SIG_SIZE);
+
+ ##
+ # If value is a hash or array, return new DeepDB object with correct offset
+ ##
+ if (($signature eq TYPE_HASH) || ($signature eq TYPE_ARRAY)) {
+ my $obj = DBM::Deep->new(
+ type => $signature,
+ base_offset => $subloc,
+ root => $self->root
+ );
+
+#XXX NO tests for this
+ if ($self->root->{autobless}) {
+ ##
+ # Skip over value and plain key to see if object needs
+ # to be re-blessed
+ ##
+ seek($self->fh, $DATA_LENGTH_SIZE + $INDEX_SIZE, 1);
+
+ my $size;
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+ if ($size) { seek($self->fh, $size, 1); }
+
+ my $bless_bit;
+ $self->fh->read($bless_bit, 1);
+ if (ord($bless_bit)) {
+ ##
+ # Yes, object needs to be re-blessed
+ ##
+ my $class_name;
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+ if ($size) { $self->fh->read($class_name, $size); }
+ if ($class_name) { $obj = bless( $obj, $class_name ); }
+ }
+ }
+
+ return $obj;
+ }
+
+ ##
+ # Otherwise return actual value
+ ##
+ elsif ($signature eq SIG_DATA) {
+ my $size;
+ my $value = '';
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+ if ($size) { $self->fh->read($value, $size); }
+ return $value;
+ }
+
+ ##
+ # Key exists, but content is null
+ ##
+ else { return; }
+ } # i loop
+
+ return;
+}
+
+sub delete_bucket {
+ ##
+ # Delete single key/value pair given tag and MD5 digested key.
+ ##
+ my $self = shift;
+ my ($tag, $md5) = @_;
+ my $keys = $tag->{content};
+
+ ##
+ # Iterate through buckets, looking for a key match
+ ##
+ BUCKET:
+ for (my $i=0; $i<$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ my $subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+
+ if (!$subloc) {
+ ##
+ # Hit end of list, no match
+ ##
+ return;
+ }
+
+ if ( $md5 ne $key ) {
+ next BUCKET;
+ }
+
+ ##
+ # Matched key -- delete bucket and return
+ ##
+ seek($self->fh, $tag->{offset} + ($i * $BUCKET_SIZE), 0);
+ $self->fh->print( substr($keys, ($i+1) * $BUCKET_SIZE ) );
+ $self->fh->print( chr(0) x $BUCKET_SIZE );
+
+ return 1;
+ } # i loop
+
+ return;
+}
+
+sub bucket_exists {
+ ##
+ # Check existence of single key given tag and MD5 digested key.
+ ##
+ my $self = shift;
+ my ($tag, $md5) = @_;
+ my $keys = $tag->{content};
+
+ ##
+ # Iterate through buckets, looking for a key match
+ ##
+ BUCKET:
+ for (my $i=0; $i<$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ my $subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+
+ if (!$subloc) {
+ ##
+ # Hit end of list, no match
+ ##
+ return;
+ }
+
+ if ( $md5 ne $key ) {
+ next BUCKET;
+ }
+
+ ##
+ # Matched key -- return true
+ ##
+ return 1;
+ } # i loop
+
+ return;
+}
+
+sub find_bucket_list {
+ ##
+ # Locate offset for bucket list, given digested key
+ ##
+ my $self = shift;
+ my $md5 = shift;
+
+ ##
+ # Locate offset for bucket list using digest index system
+ ##
+ my $ch = 0;
+ my $tag = $self->load_tag($self->base_offset);
+ if (!$tag) { return; }
+
+ while ($tag->{signature} ne SIG_BLIST) {
+ $tag = $self->index_lookup($tag, ord(substr($md5, $ch, 1)));
+ if (!$tag) { return; }
+ $ch++;
+ }
+
+ return $tag;
+}
+
+sub traverse_index {
+ ##
+ # Scan index and recursively step into deeper levels, looking for next key.
+ ##
+ my ($self, $offset, $ch, $force_return_next) = @_;
+ $force_return_next = undef unless $force_return_next;
+
+ my $tag = $self->load_tag( $offset );
+
+ if ($tag->{signature} ne SIG_BLIST) {
+ my $content = $tag->{content};
+ my $start;
+ if ($self->{return_next}) { $start = 0; }
+ else { $start = ord(substr($self->{prev_md5}, $ch, 1)); }
+
+ for (my $index = $start; $index < 256; $index++) {
+ my $subloc = unpack($LONG_PACK, substr($content, $index * $LONG_SIZE, $LONG_SIZE) );
+ if ($subloc) {
+ my $result = $self->traverse_index( $subloc, $ch + 1, $force_return_next );
+ if (defined($result)) { return $result; }
+ }
+ } # index loop
+
+ $self->{return_next} = 1;
+ } # tag is an index
+
+ elsif ($tag->{signature} eq SIG_BLIST) {
+ my $keys = $tag->{content};
+ if ($force_return_next) { $self->{return_next} = 1; }
+
+ ##
+ # Iterate through buckets, looking for a key match
+ ##
+ for (my $i=0; $i<$MAX_BUCKETS; $i++) {
+ my $key = substr($keys, $i * $BUCKET_SIZE, $HASH_SIZE);
+ my $subloc = unpack($LONG_PACK, substr($keys, ($i * $BUCKET_SIZE) + $HASH_SIZE, $LONG_SIZE));
+
+ if (!$subloc) {
+ ##
+ # End of bucket list -- return to outer loop
+ ##
+ $self->{return_next} = 1;
+ last;
+ }
+ elsif ($key eq $self->{prev_md5}) {
+ ##
+ # Located previous key -- return next one found
+ ##
+ $self->{return_next} = 1;
+ next;
+ }
+ elsif ($self->{return_next}) {
+ ##
+ # Seek to bucket location and skip over signature
+ ##
+ seek($self->fh, $subloc + SIG_SIZE, 0);
+
+ ##
+ # Skip over value to get to plain key
+ ##
+ my $size;
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+ if ($size) { seek($self->fh, $size, 1); }
+
+ ##
+ # Read in plain key and return as scalar
+ ##
+ my $plain_key;
+ $self->fh->read($size, $DATA_LENGTH_SIZE); $size = unpack($DATA_LENGTH_PACK, $size);
+ if ($size) { $self->fh->read($plain_key, $size); }
+
+ return $plain_key;
+ }
+ } # bucket loop
+
+ $self->{return_next} = 1;
+ } # tag is a bucket list
+
+ return;
+}
+
+sub get_next_key {
+ ##
+ # Locate next key, given digested previous one
+ ##
+ my $self = _get_self($_[0]);
+
+ $self->{prev_md5} = $_[1] ? $_[1] : undef;
+ $self->{return_next} = 0;
+
+ ##
+ # If the previous key was not specifed, start at the top and
+ # return the first one found.
+ ##
+ if (!$self->{prev_md5}) {
+ $self->{prev_md5} = chr(0) x $HASH_SIZE;
+ $self->{return_next} = 1;
+ }
+
+ return $self->traverse_index( $self->base_offset, 0 );
+}
+
+sub lock {
+ ##
+ # If db locking is set, flock() the db file. If called multiple
+ # times before unlock(), then the same number of unlocks() must
+ # be called before the lock is released.
+ ##
+ my $self = _get_self($_[0]);
+ my ($type) = @_;
+ $type = LOCK_EX unless defined $type;
+
+ if ($self->root->{locking}) {
+ if (!$self->root->{locked}) { flock($self->fh, $type); }
+ $self->root->{locked}++;
+ }
+}
+
+sub unlock {
+ ##
+ # If db locking is set, unlock the db file. See note in lock()
+ # regarding calling lock() multiple times.
+ ##
+ my $self = _get_self($_[0]);
+# my $type = $_[1];
+
+ if ($self->root->{locking} && $self->root->{locked} > 0) {
+ $self->root->{locked}--;
+ if (!$self->root->{locked}) { flock($self->fh, LOCK_UN); }
+ }
+}
+
+#XXX These uses of ref() need verified
+sub copy_node {
+ ##
+ # Copy single level of keys or elements to new DB handle.
+ # Recurse for nested structures
+ ##
+ my $self = _get_self($_[0]);
+ my $db_temp = $_[1];
+
+ if ($self->{type} eq TYPE_HASH) {
+ my $key = $self->first_key();
+ while ($key) {
+ my $value = $self->get($key);
+ if (!ref($value)) { $db_temp->{$key} = $value; }
+ else {
+ my $type = $value->type;
+ if ($type eq TYPE_HASH) { $db_temp->{$key} = {}; }
+ else { $db_temp->{$key} = []; }
+ $value->copy_node( $db_temp->{$key} );
+ }
+ $key = $self->next_key($key);
+ }
+ }
+ else {
+ my $length = $self->length();
+ for (my $index = 0; $index < $length; $index++) {
+ my $value = $self->get($index);
+ if (!ref($value)) { $db_temp->[$index] = $value; }
+ else {
+ my $type = $value->type;
+ if ($type eq TYPE_HASH) { $db_temp->[$index] = {}; }
+ else { $db_temp->[$index] = []; }
+ $value->copy_node( $db_temp->[$index] );
+ }
+ }
+ }
+}
+
+sub export {
+ ##
+ # Recursively export into standard Perl hashes and arrays.
+ ##
+ my $self = _get_self($_[0]);
+
+ my $temp;
+ if ($self->type eq TYPE_HASH) { $temp = {}; }
+ elsif ($self->type eq TYPE_ARRAY) { $temp = []; }
+
+ $self->lock();
+ $self->copy_node( $temp );
+ $self->unlock();
+
+ return $temp;
+}
+
+sub import {
+ ##
+ # Recursively import Perl hash/array structure
+ ##
+ #XXX This use of ref() seems to be ok
+ if (!ref($_[0])) { return; } # Perl calls import() on use -- ignore
+
+ my $self = _get_self($_[0]);
+ my $struct = $_[1];
+
+ #XXX This use of ref() seems to be ok
+ if (!ref($struct)) {
+ ##
+ # struct is not a reference, so just import based on our type
+ ##
+ shift @_;
+
+ if ($self->type eq TYPE_HASH) { $struct = {@_}; }
+ elsif ($self->type eq TYPE_ARRAY) { $struct = [@_]; }
+ }
+
+ my $r = Scalar::Util::reftype($struct) || '';
+ if ($r eq "HASH" && $self->type eq TYPE_HASH) {
+ foreach my $key (keys %$struct) { $self->put($key, $struct->{$key}); }
+ }
+ elsif ($r eq "ARRAY" && $self->type eq TYPE_ARRAY) {
+ $self->push( @$struct );
+ }
+ else {
+ return $self->throw_error("Cannot import: type mismatch");
+ }
+
+ return 1;
+}
+
+sub optimize {
+ ##
+ # Rebuild entire database into new file, then move
+ # it back on top of original.
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->root->{links} > 1) {
+ return $self->throw_error("Cannot optimize: reference count is greater than 1");
+ }
+
+ my $db_temp = DBM::Deep->new(
+ file => $self->root->{file} . '.tmp',
+ type => $self->type
+ );
+ if (!$db_temp) {
+ return $self->throw_error("Cannot optimize: failed to open temp file: $!");
+ }
+
+ $self->lock();
+ $self->copy_node( $db_temp );
+ undef $db_temp;
+
+ ##
+ # Attempt to copy user, group and permissions over to new file
+ ##
+ my @stats = stat($self->fh);
+ my $perms = $stats[2] & 07777;
+ my $uid = $stats[4];
+ my $gid = $stats[5];
+ chown( $uid, $gid, $self->root->{file} . '.tmp' );
+ chmod( $perms, $self->root->{file} . '.tmp' );
+
+ # q.v. perlport for more information on this variable
+ if ( $^O eq 'MSWin32' ) {
+ ##
+ # Potential race condition when optmizing on Win32 with locking.
+ # The Windows filesystem requires that the filehandle be closed
+ # before it is overwritten with rename(). This could be redone
+ # with a soft copy.
+ ##
+ $self->unlock();
+ $self->close();
+ }
+
+ if (!rename $self->root->{file} . '.tmp', $self->root->{file}) {
+ unlink $self->root->{file} . '.tmp';
+ $self->unlock();
+ return $self->throw_error("Optimize failed: Cannot copy temp file over original: $!");
+ }
+
+ $self->unlock();
+ $self->close();
+ $self->open();
+
+ return 1;
+}
+
+sub clone {
+ ##
+ # Make copy of object and return
+ ##
+ my $self = _get_self($_[0]);
+
+ return DBM::Deep->new(
+ type => $self->type,
+ base_offset => $self->base_offset,
+ root => $self->root
+ );
+}
+
+{
+ my %is_legal_filter = map {
+ $_ => ~~1,
+ } qw(
+ store_key store_value
+ fetch_key fetch_value
+ );
+
+ sub set_filter {
+ ##
+ # Setup filter function for storing or fetching the key or value
+ ##
+ my $self = _get_self($_[0]);
+ my $type = lc $_[1];
+ my $func = $_[2] ? $_[2] : undef;
+
+ if ( $is_legal_filter{$type} ) {
+ $self->root->{"filter_$type"} = $func;
+ return 1;
+ }
+
+ return;
+ }
+}
+
+##
+# Accessor methods
+##
+
+sub root {
+ ##
+ # Get access to the root structure
+ ##
+ my $self = _get_self($_[0]);
+ return $self->{root};
+}
+
+sub fh {
+ ##
+ # Get access to the raw FileHandle
+ ##
+ my $self = _get_self($_[0]);
+ return $self->root->{fh};
+}
+
+sub type {
+ ##
+ # Get type of current node (TYPE_HASH or TYPE_ARRAY)
+ ##
+ my $self = _get_self($_[0]);
+ return $self->{type};
+}
+
+sub base_offset {
+ ##
+ # Get base_offset of current node (TYPE_HASH or TYPE_ARRAY)
+ ##
+ my $self = _get_self($_[0]);
+ return $self->{base_offset};
+}
+
+sub error {
+ ##
+ # Get last error string, or undef if no error
+ ##
+ return $_[0]
+ ? ( _get_self($_[0])->{root}->{error} or undef )
+ : $@;
+}
+
+##
+# Utility methods
+##
+
+sub throw_error {
+ ##
+ # Store error string in self
+ ##
+ my $self = _get_self($_[0]);
+ my $error_text = $_[1];
+
+ $self->root->{error} = $error_text;
+
+ unless ($self->root->{debug}) {
+ die "DBM::Deep: $error_text\n";
+ }
+
+ warn "DBM::Deep: $error_text\n";
+ return;
+}
+
+sub clear_error {
+ ##
+ # Clear error state
+ ##
+ my $self = _get_self($_[0]);
+
+ undef $self->root->{error};
+}
+
+sub precalc_sizes {
+ ##
+ # Precalculate index, bucket and bucket list sizes
+ ##
+
+ #XXX I don't like this ...
+ set_pack() unless defined $LONG_SIZE;
+
+ $INDEX_SIZE = 256 * $LONG_SIZE;
+ $BUCKET_SIZE = $HASH_SIZE + $LONG_SIZE;
+ $BUCKET_LIST_SIZE = $MAX_BUCKETS * $BUCKET_SIZE;
+}
+
+sub set_pack {
+ ##
+ # Set pack/unpack modes (see file header for more)
+ ##
+ my ($long_s, $long_p, $data_s, $data_p) = @_;
+
+ $LONG_SIZE = $long_s ? $long_s : 4;
+ $LONG_PACK = $long_p ? $long_p : 'N';
+
+ $DATA_LENGTH_SIZE = $data_s ? $data_s : 4;
+ $DATA_LENGTH_PACK = $data_p ? $data_p : 'N';
+
+ precalc_sizes();
+}
+
+sub set_digest {
+ ##
+ # Set key digest function (default is MD5)
+ ##
+ my ($digest_func, $hash_size) = @_;
+
+ $DIGEST_FUNC = $digest_func ? $digest_func : \&Digest::MD5::md5;
+ $HASH_SIZE = $hash_size ? $hash_size : 16;
+
+ precalc_sizes();
+}
+
+##
+# tie() methods (hashes and arrays)
+##
+
+sub STORE {
+ ##
+ # Store single hash key/value or array element in database.
+ ##
+ my $self = _get_self($_[0]);
+ my $key = ($self->root->{filter_store_key} && $self->type eq TYPE_HASH) ? $self->root->{filter_store_key}->($_[1]) : $_[1];
+ #XXX What is ref() checking here?
+ my $value = ($self->root->{filter_store_value} && !ref($_[2])) ? $self->root->{filter_store_value}->($_[2]) : $_[2];
+
+ my $unpacked_key = $key;
+ if (($self->type eq TYPE_ARRAY) && ($key =~ /^\d+$/)) { $key = pack($LONG_PACK, $key); }
+ my $md5 = $DIGEST_FUNC->($key);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh) && !$self->open()) {
+ return;
+ }
+
+ ##
+ # Request exclusive lock for writing
+ ##
+ $self->lock( LOCK_EX );
+
+ ##
+ # If locking is enabled, set 'end' parameter again, in case another
+ # DB instance appended to our file while we were unlocked.
+ ##
+ if ($self->root->{locking} || $self->root->{volatile}) {
+ $self->root->{end} = (stat($self->fh))[7];
+ }
+
+ ##
+ # Locate offset for bucket list using digest index system
+ ##
+ my $tag = $self->load_tag($self->base_offset);
+ if (!$tag) {
+ $tag = $self->create_tag($self->base_offset, SIG_INDEX, chr(0) x $INDEX_SIZE);
+ }
+
+ my $ch = 0;
+ while ($tag->{signature} ne SIG_BLIST) {
+ my $num = ord(substr($md5, $ch, 1));
+ my $new_tag = $self->index_lookup($tag, $num);
+ if (!$new_tag) {
+ my $ref_loc = $tag->{offset} + ($num * $LONG_SIZE);
+ seek($self->fh, $ref_loc, 0);
+ $self->fh->print( pack($LONG_PACK, $self->root->{end}) );
+
+ $tag = $self->create_tag($self->root->{end}, SIG_BLIST, chr(0) x $BUCKET_LIST_SIZE);
+ $tag->{ref_loc} = $ref_loc;
+ $tag->{ch} = $ch;
+ last;
+ }
+ else {
+ my $ref_loc = $tag->{offset} + ($num * $LONG_SIZE);
+ $tag = $new_tag;
+ $tag->{ref_loc} = $ref_loc;
+ $tag->{ch} = $ch;
+ }
+ $ch++;
+ }
+
+ ##
+ # Add key/value to bucket list
+ ##
+ my $result = $self->add_bucket( $tag, $md5, $key, $value );
+
+ ##
+ # If this object is an array, and bucket was not a replace, and key is numerical,
+ # and index is equal or greater than current length, advance length variable.
+ ##
+ if (($result == 2) && ($self->type eq TYPE_ARRAY) && ($unpacked_key =~ /^\d+$/) && ($unpacked_key >= $self->FETCHSIZE())) {
+ $self->STORESIZE( $unpacked_key + 1 );
+ }
+
+ $self->unlock();
+
+ return $result;
+}
+
+sub FETCH {
+ ##
+ # Fetch single value or element given plain key or array index
+ ##
+ my $self = _get_self($_[0]);
+
+ my $key = $_[1];
+ if ( $self->type eq TYPE_HASH ) {
+ if ( my $filter = $self->root->{filter_store_key} ) {
+ $key = $filter->( $key );
+ }
+ }
+ elsif ( $self->type eq TYPE_ARRAY ) {
+ if ( $key =~ /^\d+$/ ) {
+ $key = pack($LONG_PACK, $key);
+ }
+ }
+
+ my $md5 = $DIGEST_FUNC->($key);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request shared lock for reading
+ ##
+ $self->lock( LOCK_SH );
+
+ my $tag = $self->find_bucket_list( $md5 );
+ if (!$tag) {
+ $self->unlock();
+ return;
+ }
+
+ ##
+ # Get value from bucket list
+ ##
+ my $result = $self->get_bucket_value( $tag, $md5 );
+
+ $self->unlock();
+
+ #XXX What is ref() checking here?
+ return ($result && !ref($result) && $self->root->{filter_fetch_value}) ? $self->root->{filter_fetch_value}->($result) : $result;
+}
+
+sub DELETE {
+ ##
+ # Delete single key/value pair or element given plain key or array index
+ ##
+ my $self = _get_self($_[0]);
+ my $key = ($self->root->{filter_store_key} && $self->type eq TYPE_HASH) ? $self->root->{filter_store_key}->($_[1]) : $_[1];
+
+ my $unpacked_key = $key;
+ if (($self->type eq TYPE_ARRAY) && ($key =~ /^\d+$/)) { $key = pack($LONG_PACK, $key); }
+ my $md5 = $DIGEST_FUNC->($key);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request exclusive lock for writing
+ ##
+ $self->lock( LOCK_EX );
+
+ my $tag = $self->find_bucket_list( $md5 );
+ if (!$tag) {
+ $self->unlock();
+ return;
+ }
+
+ ##
+ # Delete bucket
+ ##
+ my $result = $self->delete_bucket( $tag, $md5 );
+
+ ##
+ # If this object is an array and the key deleted was on the end of the stack,
+ # decrement the length variable.
+ ##
+ if ($result && ($self->type eq TYPE_ARRAY) && ($unpacked_key == $self->FETCHSIZE() - 1)) {
+ $self->STORESIZE( $unpacked_key );
+ }
+
+ $self->unlock();
+
+ return $result;
+}
+
+sub EXISTS {
+ ##
+ # Check if a single key or element exists given plain key or array index
+ ##
+ my $self = _get_self($_[0]);
+ my $key = ($self->root->{filter_store_key} && $self->type eq TYPE_HASH) ? $self->root->{filter_store_key}->($_[1]) : $_[1];
+
+ if (($self->type eq TYPE_ARRAY) && ($key =~ /^\d+$/)) { $key = pack($LONG_PACK, $key); }
+ my $md5 = $DIGEST_FUNC->($key);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request shared lock for reading
+ ##
+ $self->lock( LOCK_SH );
+
+ my $tag = $self->find_bucket_list( $md5 );
+
+ ##
+ # For some reason, the built-in exists() function returns '' for false
+ ##
+ if (!$tag) {
+ $self->unlock();
+ return '';
+ }
+
+ ##
+ # Check if bucket exists and return 1 or ''
+ ##
+ my $result = $self->bucket_exists( $tag, $md5 ) || '';
+
+ $self->unlock();
+
+ return $result;
+}
+
+sub CLEAR {
+ ##
+ # Clear all keys from hash, or all elements from array.
+ ##
+ my $self = _get_self($_[0]);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request exclusive lock for writing
+ ##
+ $self->lock( LOCK_EX );
+
+ seek($self->fh, $self->base_offset, 0);
+ if ($self->fh->eof()) {
+ $self->unlock();
+ return;
+ }
+
+ $self->create_tag($self->base_offset, $self->type, chr(0) x $INDEX_SIZE);
+
+ $self->unlock();
+
+ return 1;
+}
+
+sub FIRSTKEY {
+ ##
+ # Locate and return first key (in no particular order)
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_HASH) {
+ return $self->throw_error("FIRSTKEY method only supported for hashes");
+ }
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request shared lock for reading
+ ##
+ $self->lock( LOCK_SH );
+
+ my $result = $self->get_next_key();
+
+ $self->unlock();
+
+ return ($result && $self->root->{filter_fetch_key}) ? $self->root->{filter_fetch_key}->($result) : $result;
+}
+
+sub NEXTKEY {
+ ##
+ # Return next key (in no particular order), given previous one
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_HASH) {
+ return $self->throw_error("NEXTKEY method only supported for hashes");
+ }
+ my $prev_key = ($self->root->{filter_store_key} && $self->type eq TYPE_HASH) ? $self->root->{filter_store_key}->($_[1]) : $_[1];
+ my $prev_md5 = $DIGEST_FUNC->($prev_key);
+
+ ##
+ # Make sure file is open
+ ##
+ if (!defined($self->fh)) { $self->open(); }
+
+ ##
+ # Request shared lock for reading
+ ##
+ $self->lock( LOCK_SH );
+
+ my $result = $self->get_next_key( $prev_md5 );
+
+ $self->unlock();
+
+ return ($result && $self->root->{filter_fetch_key}) ? $self->root->{filter_fetch_key}->($result) : $result;
+}
+
+##
+# The following methods are for arrays only
+##
+
+sub FETCHSIZE {
+ ##
+ # Return the length of the array
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("FETCHSIZE method only supported for arrays");
+ }
+
+ my $SAVE_FILTER = $self->root->{filter_fetch_value};
+ $self->root->{filter_fetch_value} = undef;
+
+ my $packed_size = $self->FETCH('length');
+
+ $self->root->{filter_fetch_value} = $SAVE_FILTER;
+
+ if ($packed_size) { return int(unpack($LONG_PACK, $packed_size)); }
+ else { return 0; }
+}
+
+sub STORESIZE {
+ ##
+ # Set the length of the array
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("STORESIZE method only supported for arrays");
+ }
+ my $new_length = $_[1];
+
+ my $SAVE_FILTER = $self->root->{filter_store_value};
+ $self->root->{filter_store_value} = undef;
+
+ my $result = $self->STORE('length', pack($LONG_PACK, $new_length));
+
+ $self->root->{filter_store_value} = $SAVE_FILTER;
+
+ return $result;
+}
+
+sub POP {
+ ##
+ # Remove and return the last element on the array
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("POP method only supported for arrays");
+ }
+ my $length = $self->FETCHSIZE();
+
+ if ($length) {
+ my $content = $self->FETCH( $length - 1 );
+ $self->DELETE( $length - 1 );
+ return $content;
+ }
+ else {
+ return;
+ }
+}
+
+sub PUSH {
+ ##
+ # Add new element(s) to the end of the array
+ ##
+ my $self = _get_self(shift);
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("PUSH method only supported for arrays");
+ }
+ my $length = $self->FETCHSIZE();
+
+ while (my $content = shift @_) {
+ $self->STORE( $length, $content );
+ $length++;
+ }
+}
+
+sub SHIFT {
+ ##
+ # Remove and return first element on the array.
+ # Shift over remaining elements to take up space.
+ ##
+ my $self = _get_self($_[0]);
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("SHIFT method only supported for arrays");
+ }
+ my $length = $self->FETCHSIZE();
+
+ if ($length) {
+ my $content = $self->FETCH( 0 );
+
+ ##
+ # Shift elements over and remove last one.
+ ##
+ for (my $i = 0; $i < $length - 1; $i++) {
+ $self->STORE( $i, $self->FETCH($i + 1) );
+ }
+ $self->DELETE( $length - 1 );
+
+ return $content;
+ }
+ else {
+ return;
+ }
+}
+
+sub UNSHIFT {
+ ##
+ # Insert new element(s) at beginning of array.
+ # Shift over other elements to make space.
+ ##
+ my $self = _get_self($_[0]);shift @_;
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("UNSHIFT method only supported for arrays");
+ }
+ my @new_elements = @_;
+ my $length = $self->FETCHSIZE();
+ my $new_size = scalar @new_elements;
+
+ if ($length) {
+ for (my $i = $length - 1; $i >= 0; $i--) {
+ $self->STORE( $i + $new_size, $self->FETCH($i) );
+ }
+ }
+
+ for (my $i = 0; $i < $new_size; $i++) {
+ $self->STORE( $i, $new_elements[$i] );
+ }
+}
+
+sub SPLICE {
+ ##
+ # Splices section of array with optional new section.
+ # Returns deleted section, or last element deleted in scalar context.
+ ##
+ my $self = _get_self($_[0]);shift @_;
+ if ($self->type ne TYPE_ARRAY) {
+ return $self->throw_error("SPLICE method only supported for arrays");
+ }
+ my $length = $self->FETCHSIZE();
+
+ ##
+ # Calculate offset and length of splice
+ ##
+ my $offset = shift || 0;
+ if ($offset < 0) { $offset += $length; }
+
+ my $splice_length;
+ if (scalar @_) { $splice_length = shift; }
+ else { $splice_length = $length - $offset; }
+ if ($splice_length < 0) { $splice_length += ($length - $offset); }
+
+ ##
+ # Setup array with new elements, and copy out old elements for return
+ ##
+ my @new_elements = @_;
+ my $new_size = scalar @new_elements;
+
+ my @old_elements = ();
+ for (my $i = $offset; $i < $offset + $splice_length; $i++) {
+ push @old_elements, $self->FETCH( $i );
+ }
+
+ ##
+ # Adjust array length, and shift elements to accomodate new section.
+ ##
+ if ( $new_size != $splice_length ) {
+ if ($new_size > $splice_length) {
+ for (my $i = $length - 1; $i >= $offset + $splice_length; $i--) {
+ $self->STORE( $i + ($new_size - $splice_length), $self->FETCH($i) );
+ }
+ }
+ else {
+ for (my $i = $offset + $splice_length; $i < $length; $i++) {
+ $self->STORE( $i + ($new_size - $splice_length), $self->FETCH($i) );
+ }
+ for (my $i = 0; $i < $splice_length - $new_size; $i++) {
+ $self->DELETE( $length - 1 );
+ $length--;
+ }
+ }
+ }
+
+ ##
+ # Insert new elements into array
+ ##
+ for (my $i = $offset; $i < $offset + $new_size; $i++) {
+ $self->STORE( $i, shift @new_elements );
+ }
+
+ ##
+ # Return deleted section, or last element in scalar context.
+ ##
+ return wantarray ? @old_elements : $old_elements[-1];
+}
+
+#XXX We don't need to define it.
+#sub EXTEND {
+ ##
+ # Perl will call EXTEND() when the array is likely to grow.
+ # We don't care, but include it for compatibility.
+ ##
+#}
+
+##
+# Public method aliases
+##
+*put = *store = *STORE;
+*get = *fetch = *FETCH;
+*delete = *DELETE;
+*exists = *EXISTS;
+*clear = *CLEAR;
+*first_key = *FIRSTKEY;
+*next_key = *NEXTKEY;
+*length = *FETCHSIZE;
+*pop = *POP;
+*push = *PUSH;
+*shift = *SHIFT;
+*unshift = *UNSHIFT;
+*splice = *SPLICE;
+
+1;
+
+__END__
+
+=head1 NAME
+
+DBM::Deep - A pure perl multi-level hash/array DBM
+
+=head1 SYNOPSIS
+
+ use DBM::Deep;
+ my $db = DBM::Deep->new( "foo.db" );
+
+ $db->{key} = 'value'; # tie() style
+ print $db->{key};
+
+ $db->put('key', 'value'); # OO style
+ print $db->get('key');
+
+ # true multi-level support
+ $db->{my_complex} = [
+ 'hello', { perl => 'rules' },
+ 42, 99 ];
+
+=head1 DESCRIPTION
+
+A unique flat-file database module, written in pure perl. True
+multi-level hash/array support (unlike MLDBM, which is faked), hybrid
+OO / tie() interface, cross-platform FTPable files, and quite fast. Can
+handle millions of keys and unlimited hash levels without significant
+slow-down. Written from the ground-up in pure perl -- this is NOT a
+wrapper around a C-based DBM. Out-of-the-box compatibility with Unix,
+Mac OS X and Windows.
+
+=head1 INSTALLATION
+
+Hopefully you are using CPAN's excellent Perl module, which will download
+and install the module for you. If not, get the tarball, and run these
+commands:
+
+ tar zxf DBM-Deep-*
+ cd DBM-Deep-*
+ perl Makefile.PL
+ make
+ make test
+ make install
+
+=head1 SETUP
+
+Construction can be done OO-style (which is the recommended way), or using
+Perl's tie() function. Both are examined here.
+
+=head2 OO CONSTRUCTION
+
+The recommended way to construct a DBM::Deep object is to use the new()
+method, which gets you a blessed, tied hash or array reference.
+
+ my $db = DBM::Deep->new( "foo.db" );
+
+This opens a new database handle, mapped to the file "foo.db". If this
+file does not exist, it will automatically be created. DB files are
+opened in "r+" (read/write) mode, and the type of object returned is a
+hash, unless otherwise specified (see L<OPTIONS> below).
+
+
+
+You can pass a number of options to the constructor to specify things like
+locking, autoflush, etc. This is done by passing an inline hash:
+
+ my $db = DBM::Deep->new(
+ file => "foo.db",
+ locking => 1,
+ autoflush => 1
+ );
+
+Notice that the filename is now specified I<inside> the hash with
+the "file" parameter, as opposed to being the sole argument to the
+constructor. This is required if any options are specified.
+See L<OPTIONS> below for the complete list.
+
+
+
+You can also start with an array instead of a hash. For this, you must
+specify the C<type> parameter:
+
+ my $db = DBM::Deep->new(
+ file => "foo.db",
+ type => DBM::Deep->TYPE_ARRAY
+ );
+
+B<Note:> Specifing the C<type> parameter only takes effect when beginning
+a new DB file. If you create a DBM::Deep object with an existing file, the
+C<type> will be loaded from the file header, and ignored if it is passed
+to the constructor.
+
+=head2 TIE CONSTRUCTION
+
+Alternatively, you can create a DBM::Deep handle by using Perl's built-in
+tie() function. This is not ideal, because you get only a basic, tied hash
+(or array) which is not blessed, so you can't call any functions on it.
+
+ my %hash;
+ tie %hash, "DBM::Deep", "foo.db";
+
+ my @array;
+ tie @array, "DBM::Deep", "bar.db";
+
+As with the OO constructor, you can replace the DB filename parameter with
+a hash containing one or more options (see L<OPTIONS> just below for the
+complete list).
+
+ tie %hash, "DBM::Deep", {
+ file => "foo.db",
+ locking => 1,
+ autoflush => 1
+ };
+
+=head2 OPTIONS
+
+There are a number of options that can be passed in when constructing your
+DBM::Deep objects. These apply to both the OO- and tie- based approaches.
+
+=over
+
+=item * file
+
+Filename of the DB file to link the handle to. You can pass a full absolute
+filesystem path, partial path, or a plain filename if the file is in the
+current working directory. This is a required parameter.
+
+=item * mode
+
+File open mode (read-only, read-write, etc.) string passed to Perl's FileHandle
+module. This is an optional parameter, and defaults to "r+" (read/write).
+B<Note:> If the default (r+) mode is selected, the file will also be auto-
+created if it doesn't exist.
+
+=item * type
+
+This parameter specifies what type of object to create, a hash or array. Use
+one of these two constants: C<DBM::Deep-E<gt>TYPE_HASH> or C<DBM::Deep-E<gt>TYPE_ARRAY>.
+This only takes effect when beginning a new file. This is an optional
+parameter, and defaults to C<DBM::Deep-E<gt>TYPE_HASH>.
+
+=item * locking
+
+Specifies whether locking is to be enabled. DBM::Deep uses Perl's Fnctl flock()
+function to lock the database in exclusive mode for writes, and shared mode for
+reads. Pass any true value to enable. This affects the base DB handle I<and
+any child hashes or arrays> that use the same DB file. This is an optional
+parameter, and defaults to 0 (disabled). See L<LOCKING> below for more.
+
+=item * autoflush
+
+Specifies whether autoflush is to be enabled on the underlying FileHandle.
+This obviously slows down write operations, but is required if you may have
+multiple processes accessing the same DB file (also consider enable I<locking>
+or at least I<volatile>). Pass any true value to enable. This is an optional
+parameter, and defaults to 0 (disabled).
+
+=item * volatile
+
+If I<volatile> mode is enabled, DBM::Deep will stat() the DB file before each
+STORE() operation. This is required if an outside force may change the size of
+the file between transactions. Locking also implicitly enables volatile. This
+is useful if you want to use a different locking system or write your own. Pass
+any true value to enable. This is an optional parameter, and defaults to 0
+(disabled).
+
+=item * autobless
+
+If I<autobless> mode is enabled, DBM::Deep will preserve blessed hashes, and
+restore them when fetched. This is an B<experimental> feature, and does have
+side-effects. Basically, when hashes are re-blessed into their original
+classes, they are no longer blessed into the DBM::Deep class! So you won't be
+able to call any DBM::Deep methods on them. You have been warned.
+This is an optional parameter, and defaults to 0 (disabled).
+
+=item * filter_*
+
+See L<FILTERS> below.
+
+=item * debug
+
+Setting I<debug> mode will make all errors non-fatal, dump them out to
+STDERR, and continue on. This is for debugging purposes only, and probably
+not what you want. This is an optional parameter, and defaults to 0 (disabled).
+
+=item * fh
+
+Instead of passing a file path, you can instead pass a handle to an pre-opened
+filehandle. Note: Beware of using the magick *DATA handle, as this actually
+contains your entire Perl script, as well as the data following the __DATA__
+marker. This will not work, because DBM::Deep uses absolute seek()s into the
+file. Instead, consider reading *DATA into an IO::Scalar handle, then passing
+in that.
+
+=back
+
+=head1 TIE INTERFACE
+
+With DBM::Deep you can access your databases using Perl's standard hash/array
+syntax. Because all Deep objects are I<tied> to hashes or arrays, you can treat
+them as such. Deep will intercept all reads/writes and direct them to the right
+place -- the DB file. This has nothing to do with the L<TIE CONSTRUCTION>
+section above. This simply tells you how to use DBM::Deep using regular hashes
+and arrays, rather than calling functions like C<get()> and C<put()> (although those
+work too). It is entirely up to you how to want to access your databases.
+
+=head2 HASHES
+
+You can treat any DBM::Deep object like a normal Perl hash reference. Add keys,
+or even nested hashes (or arrays) using standard Perl syntax:
+
+ my $db = DBM::Deep->new( "foo.db" );
+
+ $db->{mykey} = "myvalue";
+ $db->{myhash} = {};
+ $db->{myhash}->{subkey} = "subvalue";
+
+ print $db->{myhash}->{subkey} . "\n";
+
+You can even step through hash keys using the normal Perl C<keys()> function:
+
+ foreach my $key (keys %$db) {
+ print "$key: " . $db->{$key} . "\n";
+ }
+
+Remember that Perl's C<keys()> function extracts I<every> key from the hash and
+pushes them onto an array, all before the loop even begins. If you have an
+extra large hash, this may exhaust Perl's memory. Instead, consider using
+Perl's C<each()> function, which pulls keys/values one at a time, using very
+little memory:
+
+ while (my ($key, $value) = each %$db) {
+ print "$key: $value\n";
+ }
+
+Please note that when using C<each()>, you should always pass a direct
+hash reference, not a lookup. Meaning, you should B<never> do this:
+
+ # NEVER DO THIS
+ while (my ($key, $value) = each %{$db->{foo}}) { # BAD
+
+This causes an infinite loop, because for each iteration, Perl is calling
+FETCH() on the $db handle, resulting in a "new" hash for foo every time, so
+it effectively keeps returning the first key over and over again. Instead,
+assign a temporary variable to C<$db->{foo}>, then pass that to each().
+
+=head2 ARRAYS
+
+As with hashes, you can treat any DBM::Deep object like a normal Perl array
+reference. This includes inserting, removing and manipulating elements,
+and the C<push()>, C<pop()>, C<shift()>, C<unshift()> and C<splice()> functions.
+The object must have first been created using type C<DBM::Deep-E<gt>TYPE_ARRAY>,
+or simply be a nested array reference inside a hash. Example:
+
+ my $db = DBM::Deep->new(
+ file => "foo-array.db",
+ type => DBM::Deep->TYPE_ARRAY
+ );
+
+ $db->[0] = "foo";
+ push @$db, "bar", "baz";
+ unshift @$db, "bah";
+
+ my $last_elem = pop @$db; # baz
+ my $first_elem = shift @$db; # bah
+ my $second_elem = $db->[1]; # bar
+
+ my $num_elements = scalar @$db;
+
+=head1 OO INTERFACE
+
+In addition to the I<tie()> interface, you can also use a standard OO interface
+to manipulate all aspects of DBM::Deep databases. Each type of object (hash or
+array) has its own methods, but both types share the following common methods:
+C<put()>, C<get()>, C<exists()>, C<delete()> and C<clear()>.
+
+=over
+
+=item * put()
+
+Stores a new hash key/value pair, or sets an array element value. Takes two
+arguments, the hash key or array index, and the new value. The value can be
+a scalar, hash ref or array ref. Returns true on success, false on failure.
+
+ $db->put("foo", "bar"); # for hashes
+ $db->put(1, "bar"); # for arrays
+
+=item * get()
+
+Fetches the value of a hash key or array element. Takes one argument: the hash
+key or array index. Returns a scalar, hash ref or array ref, depending on the
+data type stored.
+
+ my $value = $db->get("foo"); # for hashes
+ my $value = $db->get(1); # for arrays
+
+=item * exists()
+
+Checks if a hash key or array index exists. Takes one argument: the hash key
+or array index. Returns true if it exists, false if not.
+
+ if ($db->exists("foo")) { print "yay!\n"; } # for hashes
+ if ($db->exists(1)) { print "yay!\n"; } # for arrays
+
+=item * delete()
+
+Deletes one hash key/value pair or array element. Takes one argument: the hash
+key or array index. Returns true on success, false if not found. For arrays,
+the remaining elements located after the deleted element are NOT moved over.
+The deleted element is essentially just undefined, which is exactly how Perl's
+internal arrays work. Please note that the space occupied by the deleted
+key/value or element is B<not> reused again -- see L<UNUSED SPACE RECOVERY>
+below for details and workarounds.
+
+ $db->delete("foo"); # for hashes
+ $db->delete(1); # for arrays
+
+=item * clear()
+
+Deletes B<all> hash keys or array elements. Takes no arguments. No return
+value. Please note that the space occupied by the deleted keys/values or
+elements is B<not> reused again -- see L<UNUSED SPACE RECOVERY> below for
+details and workarounds.
+
+ $db->clear(); # hashes or arrays
+
+=back
+
+=head2 HASHES
+
+For hashes, DBM::Deep supports all the common methods described above, and the
+following additional methods: C<first_key()> and C<next_key()>.
+
+=over
+
+=item * first_key()
+
+Returns the "first" key in the hash. As with built-in Perl hashes, keys are
+fetched in an undefined order (which appears random). Takes no arguments,
+returns the key as a scalar value.
+
+ my $key = $db->first_key();
+
+=item * next_key()
+
+Returns the "next" key in the hash, given the previous one as the sole argument.
+Returns undef if there are no more keys to be fetched.
+
+ $key = $db->next_key($key);
+
+=back
+
+Here are some examples of using hashes:
+
+ my $db = DBM::Deep->new( "foo.db" );
+
+ $db->put("foo", "bar");
+ print "foo: " . $db->get("foo") . "\n";
+
+ $db->put("baz", {}); # new child hash ref
+ $db->get("baz")->put("buz", "biz");
+ print "buz: " . $db->get("baz")->get("buz") . "\n";
+
+ my $key = $db->first_key();
+ while ($key) {
+ print "$key: " . $db->get($key) . "\n";
+ $key = $db->next_key($key);
+ }
+
+ if ($db->exists("foo")) { $db->delete("foo"); }
+
+=head2 ARRAYS
+
+For arrays, DBM::Deep supports all the common methods described above, and the
+following additional methods: C<length()>, C<push()>, C<pop()>, C<shift()>,
+C<unshift()> and C<splice()>.
+
+=over
+
+=item * length()
+
+Returns the number of elements in the array. Takes no arguments.
+
+ my $len = $db->length();
+
+=item * push()
+
+Adds one or more elements onto the end of the array. Accepts scalars, hash
+refs or array refs. No return value.
+
+ $db->push("foo", "bar", {});
+
+=item * pop()
+
+Fetches the last element in the array, and deletes it. Takes no arguments.
+Returns undef if array is empty. Returns the element value.
+
+ my $elem = $db->pop();
+
+=item * shift()
+
+Fetches the first element in the array, deletes it, then shifts all the
+remaining elements over to take up the space. Returns the element value. This
+method is not recommended with large arrays -- see L<LARGE ARRAYS> below for
+details.
+
+ my $elem = $db->shift();
+
+=item * unshift()
+
+Inserts one or more elements onto the beginning of the array, shifting all
+existing elements over to make room. Accepts scalars, hash refs or array refs.
+No return value. This method is not recommended with large arrays -- see
+<LARGE ARRAYS> below for details.
+
+ $db->unshift("foo", "bar", {});
+
+=item * splice()
+
+Performs exactly like Perl's built-in function of the same name. See L<perldoc
+-f splice> for usage -- it is too complicated to document here. This method is
+not recommended with large arrays -- see L<LARGE ARRAYS> below for details.
+
+=back
+
+Here are some examples of using arrays:
+
+ my $db = DBM::Deep->new(
+ file => "foo.db",
+ type => DBM::Deep->TYPE_ARRAY
+ );
+
+ $db->push("bar", "baz");
+ $db->unshift("foo");
+ $db->put(3, "buz");
+
+ my $len = $db->length();
+ print "length: $len\n"; # 4
+
+ for (my $k=0; $k<$len; $k++) {
+ print "$k: " . $db->get($k) . "\n";
+ }
+
+ $db->splice(1, 2, "biz", "baf");
+
+ while (my $elem = shift @$db) {
+ print "shifted: $elem\n";
+ }
+
+=head1 LOCKING
+
+Enable automatic file locking by passing a true value to the C<locking>
+parameter when constructing your DBM::Deep object (see L<SETUP> above).
+
+ my $db = DBM::Deep->new(
+ file => "foo.db",
+ locking => 1
+ );
+
+This causes Deep to C<flock()> the underlying FileHandle object with exclusive
+mode for writes, and shared mode for reads. This is required if you have
+multiple processes accessing the same database file, to avoid file corruption.
+Please note that C<flock()> does NOT work for files over NFS. See L<DB OVER
+NFS> below for more.
+
+=head2 EXPLICIT LOCKING
+
+You can explicitly lock a database, so it remains locked for multiple
+transactions. This is done by calling the C<lock()> method, and passing an
+optional lock mode argument (defaults to exclusive mode). This is particularly
+useful for things like counters, where the current value needs to be fetched,
+then incremented, then stored again.
+
+ $db->lock();
+ my $counter = $db->get("counter");
+ $counter++;
+ $db->put("counter", $counter);
+ $db->unlock();
+
+ # or...
+
+ $db->lock();
+ $db->{counter}++;
+ $db->unlock();
+
+You can pass C<lock()> an optional argument, which specifies which mode to use
+(exclusive or shared). Use one of these two constants: C<DBM::Deep-E<gt>LOCK_EX>
+or C<DBM::Deep-E<gt>LOCK_SH>. These are passed directly to C<flock()>, and are the
+same as the constants defined in Perl's C<Fcntl> module.
+
+ $db->lock( DBM::Deep->LOCK_SH );
+ # something here
+ $db->unlock();
+
+If you want to implement your own file locking scheme, be sure to create your
+DBM::Deep objects setting the C<volatile> option to true. This hints to Deep
+that the DB file may change between transactions. See L<LOW-LEVEL ACCESS>
+below for more.
+
+=head1 IMPORTING/EXPORTING
+
+You can import existing complex structures by calling the C<import()> method,
+and export an entire database into an in-memory structure using the C<export()>
+method. Both are examined here.
+
+=head2 IMPORTING
+
+Say you have an existing hash with nested hashes/arrays inside it. Instead of
+walking the structure and adding keys/elements to the database as you go,
+simply pass a reference to the C<import()> method. This recursively adds
+everything to an existing DBM::Deep object for you. Here is an example:
+
+ my $struct = {
+ key1 => "value1",
+ key2 => "value2",
+ array1 => [ "elem0", "elem1", "elem2" ],
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2"
+ }
+ };
+
+ my $db = DBM::Deep->new( "foo.db" );
+ $db->import( $struct );
+
+ print $db->{key1} . "\n"; # prints "value1"
+
+This recursively imports the entire C<$struct> object into C<$db>, including
+all nested hashes and arrays. If the DBM::Deep object contains exsiting data,
+keys are merged with the existing ones, replacing if they already exist.
+The C<import()> method can be called on any database level (not just the base
+level), and works with both hash and array DB types.
+
+
+
+B<Note:> Make sure your existing structure has no circular references in it.
+These will cause an infinite loop when importing.
+
+=head2 EXPORTING
+
+Calling the C<export()> method on an existing DBM::Deep object will return
+a reference to a new in-memory copy of the database. The export is done
+recursively, so all nested hashes/arrays are all exported to standard Perl
+objects. Here is an example:
+
+ my $db = DBM::Deep->new( "foo.db" );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ $db->{hash1} = {};
+ $db->{hash1}->{subkey1} = "subvalue1";
+ $db->{hash1}->{subkey2} = "subvalue2";
+
+ my $struct = $db->export();
+
+ print $struct->{key1} . "\n"; # prints "value1"
+
+This makes a complete copy of the database in memory, and returns a reference
+to it. The C<export()> method can be called on any database level (not just
+the base level), and works with both hash and array DB types. Be careful of
+large databases -- you can store a lot more data in a DBM::Deep object than an
+in-memory Perl structure.
+
+
+
+B<Note:> Make sure your database has no circular references in it.
+These will cause an infinite loop when exporting.
+
+=head1 FILTERS
+
+DBM::Deep has a number of hooks where you can specify your own Perl function
+to perform filtering on incoming or outgoing data. This is a perfect
+way to extend the engine, and implement things like real-time compression or
+encryption. Filtering applies to the base DB level, and all child hashes /
+arrays. Filter hooks can be specified when your DBM::Deep object is first
+constructed, or by calling the C<set_filter()> method at any time. There are
+four available filter hooks, described below:
+
+=over
+
+=item * filter_store_key
+
+This filter is called whenever a hash key is stored. It
+is passed the incoming key, and expected to return a transformed key.
+
+=item * filter_store_value
+
+This filter is called whenever a hash key or array element is stored. It
+is passed the incoming value, and expected to return a transformed value.
+
+=item * filter_fetch_key
+
+This filter is called whenever a hash key is fetched (i.e. via
+C<first_key()> or C<next_key()>). It is passed the transformed key,
+and expected to return the plain key.
+
+=item * filter_fetch_value
+
+This filter is called whenever a hash key or array element is fetched.
+It is passed the transformed value, and expected to return the plain value.
+
+=back
+
+Here are the two ways to setup a filter hook:
+
+ my $db = DBM::Deep->new(
+ file => "foo.db",
+ filter_store_value => \&my_filter_store,
+ filter_fetch_value => \&my_filter_fetch
+ );
+
+ # or...
+
+ $db->set_filter( "filter_store_value", \&my_filter_store );
+ $db->set_filter( "filter_fetch_value", \&my_filter_fetch );
+
+Your filter function will be called only when dealing with SCALAR keys or
+values. When nested hashes and arrays are being stored/fetched, filtering
+is bypassed. Filters are called as static functions, passed a single SCALAR
+argument, and expected to return a single SCALAR value. If you want to
+remove a filter, set the function reference to C<undef>:
+
+ $db->set_filter( "filter_store_value", undef );
+
+=head2 REAL-TIME ENCRYPTION EXAMPLE
+
+Here is a working example that uses the I<Crypt::Blowfish> module to
+do real-time encryption / decryption of keys & values with DBM::Deep Filters.
+Please visit L<http://search.cpan.org/search?module=Crypt::Blowfish> for more
+on I<Crypt::Blowfish>. You'll also need the I<Crypt::CBC> module.
+
+ use DBM::Deep;
+ use Crypt::Blowfish;
+ use Crypt::CBC;
+
+ my $cipher = Crypt::CBC->new({
+ 'key' => 'my secret key',
+ 'cipher' => 'Blowfish',
+ 'iv' => '$KJh#(}q',
+ 'regenerate_key' => 0,
+ 'padding' => 'space',
+ 'prepend_iv' => 0
+ });
+
+ my $db = DBM::Deep->new(
+ file => "foo-encrypt.db",
+ filter_store_key => \&my_encrypt,
+ filter_store_value => \&my_encrypt,
+ filter_fetch_key => \&my_decrypt,
+ filter_fetch_value => \&my_decrypt,
+ );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_encrypt {
+ return $cipher->encrypt( $_[0] );
+ }
+ sub my_decrypt {
+ return $cipher->decrypt( $_[0] );
+ }
+
+=head2 REAL-TIME COMPRESSION EXAMPLE
+
+Here is a working example that uses the I<Compress::Zlib> module to do real-time
+compression / decompression of keys & values with DBM::Deep Filters.
+Please visit L<http://search.cpan.org/search?module=Compress::Zlib> for
+more on I<Compress::Zlib>.
+
+ use DBM::Deep;
+ use Compress::Zlib;
+
+ my $db = DBM::Deep->new(
+ file => "foo-compress.db",
+ filter_store_key => \&my_compress,
+ filter_store_value => \&my_compress,
+ filter_fetch_key => \&my_decompress,
+ filter_fetch_value => \&my_decompress,
+ );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_compress {
+ return Compress::Zlib::memGzip( $_[0] ) ;
+ }
+ sub my_decompress {
+ return Compress::Zlib::memGunzip( $_[0] ) ;
+ }
+
+B<Note:> Filtering of keys only applies to hashes. Array "keys" are
+actually numerical index numbers, and are not filtered.
+
+=head1 ERROR HANDLING
+
+Most DBM::Deep methods return a true value for success, and call die() on
+failure. You can wrap calls in an eval block to catch the die. Also, the
+actual error message is stored in an internal scalar, which can be fetched by
+calling the C<error()> method.
+
+ my $db = DBM::Deep->new( "foo.db" ); # create hash
+ eval { $db->push("foo"); }; # ILLEGAL -- push is array-only call
+
+ print $db->error(); # prints error message
+
+You can then call C<clear_error()> to clear the current error state.
+
+ $db->clear_error();
+
+If you set the C<debug> option to true when creating your DBM::Deep object,
+all errors are considered NON-FATAL, and dumped to STDERR. This is only
+for debugging purposes.
+
+=head1 LARGEFILE SUPPORT
+
+If you have a 64-bit system, and your Perl is compiled with both LARGEFILE
+and 64-bit support, you I<may> be able to create databases larger than 2 GB.
+DBM::Deep by default uses 32-bit file offset tags, but these can be changed
+by calling the static C<set_pack()> method before you do anything else.
+
+ DBM::Deep::set_pack(8, 'Q');
+
+This tells DBM::Deep to pack all file offsets with 8-byte (64-bit) quad words
+instead of 32-bit longs. After setting these values your DB files have a
+theoretical maximum size of 16 XB (exabytes).
+
+
+
+B<Note:> Changing these values will B<NOT> work for existing database files.
+Only change this for new files, and make sure it stays set consistently
+throughout the file's life. If you do set these values, you can no longer
+access 32-bit DB files. You can, however, call C<set_pack(4, 'N')> to change
+back to 32-bit mode.
+
+
+
+B<Note:> I have not personally tested files > 2 GB -- all my systems have
+only a 32-bit Perl. However, I have received user reports that this does
+indeed work!
+
+=head1 LOW-LEVEL ACCESS
+
+If you require low-level access to the underlying FileHandle that Deep uses,
+you can call the C<fh()> method, which returns the handle:
+
+ my $fh = $db->fh();
+
+This method can be called on the root level of the datbase, or any child
+hashes or arrays. All levels share a I<root> structure, which contains things
+like the FileHandle, a reference counter, and all your options you specified
+when you created the object. You can get access to this root structure by
+calling the C<root()> method.
+
+ my $root = $db->root();
+
+This is useful for changing options after the object has already been created,
+such as enabling/disabling locking, volatile or debug modes. You can also
+store your own temporary user data in this structure (be wary of name
+collision), which is then accessible from any child hash or array.
+
+=head1 CUSTOM DIGEST ALGORITHM
+
+DBM::Deep by default uses the I<Message Digest 5> (MD5) algorithm for hashing
+keys. However you can override this, and use another algorithm (such as SHA-256)
+or even write your own. But please note that Deep currently expects zero
+collisions, so your algorithm has to be I<perfect>, so to speak.
+Collision detection may be introduced in a later version.
+
+
+
+You can specify a custom digest algorithm by calling the static C<set_digest()>
+function, passing a reference to a subroutine, and the length of the algorithm's
+hashes (in bytes). This is a global static function, which affects ALL Deep
+objects. Here is a working example that uses a 256-bit hash from the
+I<Digest::SHA256> module. Please see
+L<http://search.cpan.org/search?module=Digest::SHA256> for more.
+
+ use DBM::Deep;
+ use Digest::SHA256;
+
+ my $context = Digest::SHA256::new(256);
+
+ DBM::Deep::set_digest( \&my_digest, 32 );
+
+ my $db = DBM::Deep->new( "foo-sha.db" );
+
+ $db->{key1} = "value1";
+ $db->{key2} = "value2";
+ print "key1: " . $db->{key1} . "\n";
+ print "key2: " . $db->{key2} . "\n";
+
+ undef $db;
+ exit;
+
+ sub my_digest {
+ return substr( $context->hash($_[0]), 0, 32 );
+ }
+
+B<Note:> Your returned digest strings must be B<EXACTLY> the number
+of bytes you specify in the C<set_digest()> function (in this case 32).
+
+=head1 CIRCULAR REFERENCES
+
+DBM::Deep has B<experimental> support for circular references. Meaning you
+can have a nested hash key or array element that points to a parent object.
+This relationship is stored in the DB file, and is preserved between sessions.
+Here is an example:
+
+ my $db = DBM::Deep->new( "foo.db" );
+
+ $db->{foo} = "bar";
+ $db->{circle} = $db; # ref to self
+
+ print $db->{foo} . "\n"; # prints "foo"
+ print $db->{circle}->{foo} . "\n"; # prints "foo" again
+
+One catch is, passing the object to a function that recursively walks the
+object tree (such as I<Data::Dumper> or even the built-in C<optimize()> or
+C<export()> methods) will result in an infinite loop. The other catch is,
+if you fetch the I<key> of a circular reference (i.e. using the C<first_key()>
+or C<next_key()> methods), you will get the I<target object's key>, not the
+ref's key. This gets even more interesting with the above example, where
+the I<circle> key points to the base DB object, which technically doesn't
+have a key. So I made DBM::Deep return "[base]" as the key name in that
+special case.
+
+=head1 CAVEATS / ISSUES / BUGS
+
+This section describes all the known issues with DBM::Deep. It you have found
+something that is not listed here, please send e-mail to L<jhuckaby@cpan.org>.
+
+=head2 UNUSED SPACE RECOVERY
+
+One major caveat with Deep is that space occupied by existing keys and
+values is not recovered when they are deleted. Meaning if you keep deleting
+and adding new keys, your file will continuously grow. I am working on this,
+but in the meantime you can call the built-in C<optimize()> method from time to
+time (perhaps in a crontab or something) to recover all your unused space.
+
+ $db->optimize(); # returns true on success
+
+This rebuilds the ENTIRE database into a new file, then moves it on top of
+the original. The new file will have no unused space, thus it will take up as
+little disk space as possible. Please note that this operation can take
+a long time for large files, and you need enough disk space to temporarily hold
+2 copies of your DB file. The temporary file is created in the same directory
+as the original, named with a ".tmp" extension, and is deleted when the
+operation completes. Oh, and if locking is enabled, the DB is automatically
+locked for the entire duration of the copy.
+
+
+
+B<WARNING:> Only call optimize() on the top-level node of the database, and
+make sure there are no child references lying around. Deep keeps a reference
+counter, and if it is greater than 1, optimize() will abort and return undef.
+
+=head2 AUTOVIVIFICATION
+
+Unfortunately, autovivification doesn't work with tied hashes. This appears to
+be a bug in Perl's tie() system, as I<Jakob Schmidt> encountered the very same
+issue with his I<DWH_FIle> module (see L<http://search.cpan.org/search?module=DWH_File>),
+and it is also mentioned in the BUGS section for the I<MLDBM> module <see
+L<http://search.cpan.org/search?module=MLDBM>). Basically, on a new db file,
+this does not work:
+
+ $db->{foo}->{bar} = "hello";
+
+Since "foo" doesn't exist, you cannot add "bar" to it. You end up with "foo"
+being an empty hash. Try this instead, which works fine:
+
+ $db->{foo} = { bar => "hello" };
+
+As of Perl 5.8.7, this bug still exists. I have walked very carefully through
+the execution path, and Perl indeed passes an empty hash to the STORE() method.
+Probably a bug in Perl.
+
+=head2 FILE CORRUPTION
+
+The current level of error handling in Deep is minimal. Files I<are> checked
+for a 32-bit signature on open(), but other corruption in files can cause
+segmentation faults. Deep may try to seek() past the end of a file, or get
+stuck in an infinite loop depending on the level of corruption. File write
+operations are not checked for failure (for speed), so if you happen to run
+out of disk space, Deep will probably fail in a bad way. These things will
+be addressed in a later version of DBM::Deep.
+
+=head2 DB OVER NFS
+
+Beware of using DB files over NFS. Deep uses flock(), which works well on local
+filesystems, but will NOT protect you from file corruption over NFS. I've heard
+about setting up your NFS server with a locking daemon, then using lockf() to
+lock your files, but your milage may vary there as well. From what I
+understand, there is no real way to do it. However, if you need access to the
+underlying FileHandle in Deep for using some other kind of locking scheme like
+lockf(), see the L<LOW-LEVEL ACCESS> section above.
+
+=head2 COPYING OBJECTS
+
+Beware of copying tied objects in Perl. Very strange things can happen.
+Instead, use Deep's C<clone()> method which safely copies the object and
+returns a new, blessed, tied hash or array to the same level in the DB.
+
+ my $copy = $db->clone();
+
+=head2 LARGE ARRAYS
+
+Beware of using C<shift()>, C<unshift()> or C<splice()> with large arrays.
+These functions cause every element in the array to move, which can be murder
+on DBM::Deep, as every element has to be fetched from disk, then stored again in
+a different location. This may be addressed in a later version.
+
+=head1 PERFORMANCE
+
+This section discusses DBM::Deep's speed and memory usage.
+
+=head2 SPEED
+
+Obviously, DBM::Deep isn't going to be as fast as some C-based DBMs, such as
+the almighty I<BerkeleyDB>. But it makes up for it in features like true
+multi-level hash/array support, and cross-platform FTPable files. Even so,
+DBM::Deep is still pretty fast, and the speed stays fairly consistent, even
+with huge databases. Here is some test data:
+
+ Adding 1,000,000 keys to new DB file...
+
+ At 100 keys, avg. speed is 2,703 keys/sec
+ At 200 keys, avg. speed is 2,642 keys/sec
+ At 300 keys, avg. speed is 2,598 keys/sec
+ At 400 keys, avg. speed is 2,578 keys/sec
+ At 500 keys, avg. speed is 2,722 keys/sec
+ At 600 keys, avg. speed is 2,628 keys/sec
+ At 700 keys, avg. speed is 2,700 keys/sec
+ At 800 keys, avg. speed is 2,607 keys/sec
+ At 900 keys, avg. speed is 2,190 keys/sec
+ At 1,000 keys, avg. speed is 2,570 keys/sec
+ At 2,000 keys, avg. speed is 2,417 keys/sec
+ At 3,000 keys, avg. speed is 1,982 keys/sec
+ At 4,000 keys, avg. speed is 1,568 keys/sec
+ At 5,000 keys, avg. speed is 1,533 keys/sec
+ At 6,000 keys, avg. speed is 1,787 keys/sec
+ At 7,000 keys, avg. speed is 1,977 keys/sec
+ At 8,000 keys, avg. speed is 2,028 keys/sec
+ At 9,000 keys, avg. speed is 2,077 keys/sec
+ At 10,000 keys, avg. speed is 2,031 keys/sec
+ At 20,000 keys, avg. speed is 1,970 keys/sec
+ At 30,000 keys, avg. speed is 2,050 keys/sec
+ At 40,000 keys, avg. speed is 2,073 keys/sec
+ At 50,000 keys, avg. speed is 1,973 keys/sec
+ At 60,000 keys, avg. speed is 1,914 keys/sec
+ At 70,000 keys, avg. speed is 2,091 keys/sec
+ At 80,000 keys, avg. speed is 2,103 keys/sec
+ At 90,000 keys, avg. speed is 1,886 keys/sec
+ At 100,000 keys, avg. speed is 1,970 keys/sec
+ At 200,000 keys, avg. speed is 2,053 keys/sec
+ At 300,000 keys, avg. speed is 1,697 keys/sec
+ At 400,000 keys, avg. speed is 1,838 keys/sec
+ At 500,000 keys, avg. speed is 1,941 keys/sec
+ At 600,000 keys, avg. speed is 1,930 keys/sec
+ At 700,000 keys, avg. speed is 1,735 keys/sec
+ At 800,000 keys, avg. speed is 1,795 keys/sec
+ At 900,000 keys, avg. speed is 1,221 keys/sec
+ At 1,000,000 keys, avg. speed is 1,077 keys/sec
+
+This test was performed on a PowerMac G4 1gHz running Mac OS X 10.3.2 & Perl
+5.8.1, with an 80GB Ultra ATA/100 HD spinning at 7200RPM. The hash keys and
+values were between 6 - 12 chars in length. The DB file ended up at 210MB.
+Run time was 12 min 3 sec.
+
+=head2 MEMORY USAGE
+
+One of the great things about DBM::Deep is that it uses very little memory.
+Even with huge databases (1,000,000+ keys) you will not see much increased
+memory on your process. Deep relies solely on the filesystem for storing
+and fetching data. Here is output from I</usr/bin/top> before even opening a
+database handle:
+
+ PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
+ 22831 root 11 0 2716 2716 1296 R 0.0 0.2 0:07 perl
+
+Basically the process is taking 2,716K of memory. And here is the same
+process after storing and fetching 1,000,000 keys:
+
+ PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
+ 22831 root 14 0 2772 2772 1328 R 0.0 0.2 13:32 perl
+
+Notice the memory usage increased by only 56K. Test was performed on a 700mHz
+x86 box running Linux RedHat 7.2 & Perl 5.6.1.
+
+=head1 DB FILE FORMAT
+
+In case you were interested in the underlying DB file format, it is documented
+here in this section. You don't need to know this to use the module, it's just
+included for reference.
+
+=head2 SIGNATURE
+
+DBM::Deep files always start with a 32-bit signature to identify the file type.
+This is at offset 0. The signature is "DPDB" in network byte order. This is
+checked upon each file open().
+
+=head2 TAG
+
+The DBM::Deep file is in a I<tagged format>, meaning each section of the file
+has a standard header containing the type of data, the length of data, and then
+the data itself. The type is a single character (1 byte), the length is a
+32-bit unsigned long in network byte order, and the data is, well, the data.
+Here is how it unfolds:
+
+=head2 MASTER INDEX
+
+Immediately after the 32-bit file signature is the I<Master Index> record.
+This is a standard tag header followed by 1024 bytes (in 32-bit mode) or 2048
+bytes (in 64-bit mode) of data. The type is I<H> for hash or I<A> for array,
+depending on how the DBM::Deep object was constructed.
+
+
+
+The index works by looking at a I<MD5 Hash> of the hash key (or array index
+number). The first 8-bit char of the MD5 signature is the offset into the
+index, multipled by 4 in 32-bit mode, or 8 in 64-bit mode. The value of the
+index element is a file offset of the next tag for the key/element in question,
+which is usually a I<Bucket List> tag (see below).
+
+
+
+The next tag I<could> be another index, depending on how many keys/elements
+exist. See L<RE-INDEXING> below for details.
+
+=head2 BUCKET LIST
+
+A I<Bucket List> is a collection of 16 MD5 hashes for keys/elements, plus
+file offsets to where the actual data is stored. It starts with a standard
+tag header, with type I<B>, and a data size of 320 bytes in 32-bit mode, or
+384 bytes in 64-bit mode. Each MD5 hash is stored in full (16 bytes), plus
+the 32-bit or 64-bit file offset for the I<Bucket> containing the actual data.
+When the list fills up, a I<Re-Index> operation is performed (See
+L<RE-INDEXING> below).
+
+=head2 BUCKET
+
+A I<Bucket> is a tag containing a key/value pair (in hash mode), or a
+index/value pair (in array mode). It starts with a standard tag header with
+type I<D> for scalar data (string, binary, etc.), or it could be a nested
+hash (type I<H>) or array (type I<A>). The value comes just after the tag
+header. The size reported in the tag header is only for the value, but then,
+just after the value is another size (32-bit unsigned long) and then the plain
+key itself. Since the value is likely to be fetched more often than the plain
+key, I figured it would be I<slightly> faster to store the value first.
+
+
+
+If the type is I<H> (hash) or I<A> (array), the value is another I<Master Index>
+record for the nested structure, where the process begins all over again.
+
+=head2 RE-INDEXING
+
+After a I<Bucket List> grows to 16 records, its allocated space in the file is
+exhausted. Then, when another key/element comes in, the list is converted to a
+new index record. However, this index will look at the next char in the MD5
+hash, and arrange new Bucket List pointers accordingly. This process is called
+I<Re-Indexing>. Basically, a new index tag is created at the file EOF, and all
+17 (16 + new one) keys/elements are removed from the old Bucket List and
+inserted into the new index. Several new Bucket Lists are created in the
+process, as a new MD5 char from the key is being examined (it is unlikely that
+the keys will all share the same next char of their MD5s).
+
+
+
+Because of the way the I<MD5> algorithm works, it is impossible to tell exactly
+when the Bucket Lists will turn into indexes, but the first round tends to
+happen right around 4,000 keys. You will see a I<slight> decrease in
+performance here, but it picks back up pretty quick (see L<SPEED> above). Then
+it takes B<a lot> more keys to exhaust the next level of Bucket Lists. It's
+right around 900,000 keys. This process can continue nearly indefinitely --
+right up until the point the I<MD5> signatures start colliding with each other,
+and this is B<EXTREMELY> rare -- like winning the lottery 5 times in a row AND
+getting struck by lightning while you are walking to cash in your tickets.
+Theoretically, since I<MD5> hashes are 128-bit values, you I<could> have up to
+340,282,366,921,000,000,000,000,000,000,000,000,000 keys/elements (I believe
+this is 340 unodecillion, but don't quote me).
+
+=head2 STORING
+
+When a new key/element is stored, the key (or index number) is first ran through
+I<Digest::MD5> to get a 128-bit signature (example, in hex:
+b05783b0773d894396d475ced9d2f4f6). Then, the I<Master Index> record is checked
+for the first char of the signature (in this case I<b>). If it does not exist,
+a new I<Bucket List> is created for our key (and the next 15 future keys that
+happen to also have I<b> as their first MD5 char). The entire MD5 is written
+to the I<Bucket List> along with the offset of the new I<Bucket> record (EOF at
+this point, unless we are replacing an existing I<Bucket>), where the actual
+data will be stored.
+
+=head2 FETCHING
+
+Fetching an existing key/element involves getting a I<Digest::MD5> of the key
+(or index number), then walking along the indexes. If there are enough
+keys/elements in this DB level, there might be nested indexes, each linked to
+a particular char of the MD5. Finally, a I<Bucket List> is pointed to, which
+contains up to 16 full MD5 hashes. Each is checked for equality to the key in
+question. If we found a match, the I<Bucket> tag is loaded, where the value and
+plain key are stored.
+
+
+
+Fetching the plain key occurs when calling the I<first_key()> and I<next_key()>
+methods. In this process the indexes are walked systematically, and each key
+fetched in increasing MD5 order (which is why it appears random). Once the
+I<Bucket> is found, the value is skipped the plain key returned instead.
+B<Note:> Do not count on keys being fetched as if the MD5 hashes were
+alphabetically sorted. This only happens on an index-level -- as soon as the
+I<Bucket Lists> are hit, the keys will come out in the order they went in --
+so it's pretty much undefined how the keys will come out -- just like Perl's
+built-in hashes.
+
+=head1 AUTHOR
+
+Joseph Huckaby, L<jhuckaby@cpan.org>
+
+Special thanks to Adam Sah and Rich Gaushell! You know why :-)
+
+=head1 SEE ALSO
+
+perltie(1), Tie::Hash(3), Digest::MD5(3), Fcntl(3), flock(2), lockf(3), nfs(5),
+Digest::SHA256(3), Crypt::Blowfish(3), Compress::Zlib(3)
+
+=head1 LICENSE
+
+Copyright (c) 2002-2005 Joseph Huckaby. All Rights Reserved.
+This is free software, you may use it and distribute it under the
+same terms as Perl itself.
+
+=cut
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 3;
+
+use_ok( 'DBM::Deep' );
+
+##
+# basic file open
+##
+unlink "t/test.db";
+my $db = eval { DBM::Deep->new( "t/test.db" ) };
+if ( DBM::Deep::error( $db ) || !$db ) {
+ diag "ERROR: " . (DBM::Deep::error($db) || "UNKNOWN\n");
+ Test::More->builder->BAIL_OUT( "Opening a new file fails" );
+}
+
+isa_ok( $db, 'DBM::Deep' );
+ok(1, "We can successfully open a file!" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 42;
+use Test::Exception;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# put/get key
+##
+$db->{key1} = "value1";
+is( $db->get("key1"), "value1", "get() works with hash assignment" );
+is( $db->fetch("key1"), "value1", "... fetch() works with hash assignment" );
+is( $db->{key1}, "value1", "... and hash-access also works" );
+
+$db->put("key2", undef);
+is( $db->get("key2"), undef, "get() works with put()" );
+is( $db->fetch("key2"), undef, "... fetch() works with put()" );
+is( $db->{key2}, undef, "... and hash-access also works" );
+
+$db->store( "key3", "value3" );
+is( $db->get("key3"), "value3", "get() works with store()" );
+is( $db->fetch("key3"), "value3", "... fetch() works with put()" );
+is( $db->{key3}, 'value3', "... and hash-access also works" );
+
+ok( $db->exists("key1"), "exists() function works" );
+ok( exists $db->{key2}, "exists() works against tied hash" );
+
+##
+# count keys
+##
+is( scalar keys %$db, 3, "keys() works against tied hash" );
+
+##
+# step through keys
+##
+my $temphash = {};
+while ( my ($key, $value) = each %$db ) {
+ $temphash->{$key} = $value;
+}
+
+is( $temphash->{key1}, 'value1', "First key copied successfully using tied interface" );
+is( $temphash->{key2}, undef, "Second key copied successfully" );
+is( $temphash->{key3}, 'value3', "Third key copied successfully" );
+
+$temphash = {};
+my $key = $db->first_key();
+while ($key) {
+ $temphash->{$key} = $db->get($key);
+ $key = $db->next_key($key);
+}
+
+is( $temphash->{key1}, 'value1', "First key copied successfully using OO interface" );
+is( $temphash->{key2}, undef, "Second key copied successfully" );
+is( $temphash->{key3}, 'value3', "Third key copied successfully" );
+
+##
+# delete keys
+##
+TODO: {
+ local $TODO = "Delete should return the deleted value";
+ is( delete $db->{key1}, 'value1', "delete through tied inteface works" );
+ is( $db->delete("key2"), undef, "delete through OO inteface works" );
+}
+
+is( scalar keys %$db, 1, "After deleting two keys, 1 remains" );
+
+##
+# delete all keys
+##
+ok( $db->clear(), "clear() returns true" );
+
+is( scalar keys %$db, 0, "After clear(), everything is removed" );
+
+##
+# replace key
+##
+$db->put("key1", "value1");
+is( $db->get("key1"), "value1", "Assignment still works" );
+
+$db->put("key1", "value2");
+is( $db->get("key1"), "value2", "... and replacement works" );
+
+$db->put("key1", "value222222222222222222222222");
+
+is( $db->get("key1"), "value222222222222222222222222", "We set a value before closing the file" );
+
+##
+# Make sure DB still works after closing / opening
+##
+undef $db;
+$db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+is( $db->get("key1"), "value222222222222222222222222", "The value we set is still there after closure" );
+
+##
+# Make sure keys are still fetchable after replacing values
+# with smaller ones (bug found by John Cardenas, DBM::Deep 0.93)
+##
+$db->clear();
+$db->put("key1", "long value here");
+$db->put("key2", "longer value here");
+
+$db->put("key1", "short value");
+$db->put("key2", "shorter v");
+
+my $first_key = $db->first_key();
+my $next_key = $db->next_key($first_key);
+
+ok(
+ (($first_key eq "key1") || ($first_key eq "key2")) &&
+ (($next_key eq "key1") || ($next_key eq "key2")) &&
+ ($first_key ne $next_key)
+ ,"keys() still works if you replace long values with shorter ones"
+);
+
+# These tests verify that the array methods cannot be called on hashtypes.
+# They will be removed once the ARRAY and HASH types are refactored into their own classes.
+
+throws_ok {
+ $db->splice();
+} qr/SPLICE method only supported for arrays/, "Cannot call splice on a hash type";
+
+throws_ok {
+ $db->SPLICE();
+} qr/SPLICE method only supported for arrays/, "Cannot call SPLICE on a hash type";
+
+throws_ok {
+ $db->length();
+} qr/FETCHSIZE method only supported for arrays/, "Cannot call length on a hash type";
+
+throws_ok {
+ $db->FETCHSIZE();
+} qr/FETCHSIZE method only supported for arrays/, "Cannot call FETCHSIZE on a hash type";
+
+throws_ok {
+ $db->STORESIZE();
+} qr/STORESIZE method only supported for arrays/, "Cannot call STORESIZE on a hash type";
+
+throws_ok {
+ $db->POP();
+} qr/POP method only supported for arrays/, "Cannot call POP on a hash type";
+
+throws_ok {
+ $db->pop();
+} qr/POP method only supported for arrays/, "Cannot call pop on a hash type";
+
+throws_ok {
+ $db->PUSH();
+} qr/PUSH method only supported for arrays/, "Cannot call PUSH on a hash type";
+
+throws_ok {
+ $db->push();
+} qr/PUSH method only supported for arrays/, "Cannot call push on a hash type";
+
+throws_ok {
+ $db->SHIFT();
+} qr/SHIFT method only supported for arrays/, "Cannot call SHIFT on a hash type";
+
+throws_ok {
+ $db->shift();
+} qr/SHIFT method only supported for arrays/, "Cannot call shift on a hash type";
+
+throws_ok {
+ $db->UNSHIFT();
+} qr/UNSHIFT method only supported for arrays/, "Cannot call UNSHIFT on a hash type";
+
+throws_ok {
+ $db->unshift();
+} qr/UNSHIFT method only supported for arrays/, "Cannot call unshift on a hash type";
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+
+my $max_keys = 4000;
+plan tests => 2 + $max_keys;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ type => DBM::Deep->TYPE_HASH
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# put/get many keys
+##
+for ( 0 .. $max_keys ) {
+ $db->put( "hello $_" => "there " . $_ * 2 );
+}
+
+for ( 0 .. $max_keys ) {
+ is( $db->get( "hello $_" ), "there " . $_ * 2, "The ${_}th value is correct" );
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 93;
+use Test::Exception;
+
+use_ok( 'DBM::Deep' );
+
+##
+# basic file open
+##
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ type => DBM::Deep->TYPE_ARRAY
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+TODO: {
+ local $TODO = "How is this test ever supposed to pass?";
+ ok( !$db->clear, "If the file has never been written to, clear() returns false" );
+}
+
+##
+# basic put/get/push
+##
+$db->[0] = "elem1";
+$db->push( "elem2" );
+$db->put(2, "elem3");
+$db->store(3, "elem4");
+$db->unshift("elem0");
+
+is( $db->[0], 'elem0', "Array get for shift works" );
+is( $db->[1], 'elem1', "Array get for array set works" );
+is( $db->[2], 'elem2', "Array get for push() works" );
+is( $db->[3], 'elem3', "Array get for put() works" );
+is( $db->[4], 'elem4', "Array get for store() works" );
+
+is( $db->get(0), 'elem0', "get() for shift() works" );
+is( $db->get(1), 'elem1', "get() for array set works" );
+is( $db->get(2), 'elem2', "get() for push() works" );
+is( $db->get(3), 'elem3', "get() for put() works" );
+is( $db->get(4), 'elem4', "get() for store() works" );
+
+is( $db->fetch(0), 'elem0', "fetch() for shift() works" );
+is( $db->fetch(1), 'elem1', "fetch() for array set works" );
+is( $db->fetch(2), 'elem2', "fetch() for push() works" );
+is( $db->fetch(3), 'elem3', "fetch() for put() works" );
+is( $db->fetch(4), 'elem4', "fetch() for store() works" );
+
+is( $db->length, 5, "... and we have five elements" );
+
+is( $db->[-1], $db->[4], "-1st index is 4th value" );
+is( $db->[-2], $db->[3], "-2nd index is 3rd value" );
+is( $db->[-3], $db->[2], "-3rd index is 2nd value" );
+is( $db->[-4], $db->[1], "-4th index is 1st value" );
+is( $db->[-5], $db->[0], "-5th index is 0th value" );
+TODO: {
+ local $TODO = "Going off the end of the array from the back is legal";
+ eval { is( $db->[-6], undef, "-6th index is undef" ); };
+}
+is( $db->length, 5, "... and we have five elements after abortive -6 index lookup" );
+
+my $popped = $db->pop;
+is( $db->length, 4, "... and we have four after popping" );
+is( $db->[0], 'elem0', "0th element still there after popping" );
+is( $db->[1], 'elem1', "1st element still there after popping" );
+is( $db->[2], 'elem2', "2nd element still there after popping" );
+is( $db->[3], 'elem3', "3rd element still there after popping" );
+is( $popped, 'elem4', "Popped value is correct" );
+
+my $shifted = $db->shift;
+is( $db->length, 3, "... and we have three after shifting" );
+is( $db->[0], 'elem1', "0th element still there after shifting" );
+is( $db->[1], 'elem2', "1st element still there after shifting" );
+is( $db->[2], 'elem3', "2nd element still there after shifting" );
+is( $shifted, 'elem0', "Shifted value is correct" );
+
+##
+# delete
+##
+my $deleted = $db->delete(0);
+is( $db->length, 3, "... and we still have three after deleting" );
+is( $db->[0], undef, "0th element now undef" );
+is( $db->[1], 'elem2', "1st element still there after deleting" );
+is( $db->[2], 'elem3', "2nd element still there after deleting" );
+TODO: {
+ local $TODO = "delete on an array element should return the deleted value";
+ is( $deleted, 'elem1', "Deleted value is correct" );
+}
+
+is( $db->delete(99), undef, 'delete on an element not in the array returns undef' );
+is( $db->length, 3, "... and we still have three after a delete on an out-of-range index" );
+
+is( delete $db->[99], undef, 'DELETE on an element not in the array returns undef' );
+is( $db->length, 3, "... and we still have three after a DELETE on an out-of-range index" );
+
+is( $db->delete(-99), undef, 'delete on an element (neg) not in the array returns undef' );
+is( $db->length, 3, "... and we still have three after a DELETE on an out-of-range negative index" );
+
+is( delete $db->[-99], undef, 'DELETE on an element (neg) not in the array returns undef' );
+is( $db->length, 3, "... and we still have three after a DELETE on an out-of-range negative index" );
+
+$deleted = $db->delete(-2);
+is( $db->length, 3, "... and we still have three after deleting" );
+is( $db->[0], undef, "0th element still undef" );
+TODO: {
+ local $TODO = "delete on a negative array element should work";
+ is( $db->[1], undef, "1st element now undef" );
+}
+is( $db->[2], 'elem3', "2nd element still there after deleting" );
+TODO: {
+ local $TODO = "delete on an array element should return the deleted value";
+ is( $deleted, 'elem2', "Deleted value is correct" );
+}
+
+$db->[1] = 'elem2';
+
+##
+# exists
+##
+ok( $db->exists(1), "The 1st value exists" );
+ok( !$db->exists(0), "The 0th value doesn't exists" );
+ok( !$db->exists(22), "The 22nd value doesn't exists" );
+TODO: {
+ local $TODO = "exists on negative values should work";
+ ok( $db->exists(-1), "The -1st value does exists" );
+}
+ok( !$db->exists(-22), "The -22nd value doesn't exists" );
+
+##
+# clear
+##
+ok( $db->clear(), "clear() returns true if the file was ever non-empty" );
+is( $db->length(), 0, "After clear(), no more elements" );
+
+is( $db->pop, undef, "pop on an empty array returns undef" );
+is( $db->length(), 0, "After pop() on empty array, length is still 0" );
+
+is( $db->shift, undef, "shift on an empty array returns undef" );
+is( $db->length(), 0, "After shift() on empty array, length is still 0" );
+
+TODO: {
+ local $TODO = "unshift returns the number of elements in the array";
+ is( $db->unshift( 1, 2, 3 ), 3, "unshift returns the number of elements in the array" );
+ is( $db->unshift( 1, 2, 3 ), 6, "unshift returns the number of elements in the array" );
+ is( $db->push( 1, 2, 3 ), 9, "unshift returns the number of elements in the array" );
+}
+is( $db->length(), 9, "After unshift and push on empty array, length is now 9" );
+
+$db->clear;
+
+##
+# multi-push
+##
+$db->push( 'elem first', "elem middle", "elem last" );
+is( $db->length, 3, "3-element push results in three elements" );
+is($db->[0], "elem first", "First element is 'elem first'");
+is($db->[1], "elem middle", "Second element is 'elem middle'");
+is($db->[2], "elem last", "Third element is 'elem last'");
+
+##
+# splice with length 1
+##
+$db->splice( 1, 1, "middle A", "middle B" );
+is($db->length(), 4);
+is($db->[0], "elem first");
+is($db->[1], "middle A");
+is($db->[2], "middle B");
+is($db->[3], "elem last");
+
+##
+# splice with length of 0
+##
+$db->splice( -1, 0, "middle C" );
+is($db->length(), 5);
+is($db->[0], "elem first");
+is($db->[1], "middle A");
+is($db->[2], "middle B");
+is($db->[3], "middle C");
+is($db->[4], "elem last");
+
+##
+# splice with length of 3
+##
+$db->splice( 1, 3, "middle ABC" );
+is($db->length(), 3);
+is($db->[0], "elem first");
+is($db->[1], "middle ABC");
+is($db->[2], "elem last");
+
+# These tests verify that the hash methods cannot be called on arraytypes.
+# They will be removed once the ARRAY and HASH types are refactored into their own classes.
+
+$db->[0] = [ 1 .. 3 ];
+$db->[1] = { a => 'foo' };
+is( $db->[0]->length, 3, "Reuse of same space with array successful" );
+is( $db->[1]->fetch('a'), 'foo', "Reuse of same space with hash successful" );
+
+throws_ok {
+ $db->FIRSTKEY();
+} qr/FIRSTKEY method only supported for hashes/, "Cannot call FIRSTKEY on an array type";
+
+throws_ok {
+ $db->first_key();
+} qr/FIRSTKEY method only supported for hashes/, "Cannot call first_key on an array type";
+
+throws_ok {
+ $db->NEXTKEY();
+} qr/NEXTKEY method only supported for hashes/, "Cannot call NEXTKEY on an array type";
+
+throws_ok {
+ $db->next_key();
+} qr/NEXTKEY method only supported for hashes/, "Cannot call next_key on an array type";
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+
+my $max_keys = 4000;
+plan tests => 2 + $max_keys;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ type => DBM::Deep->TYPE_ARRAY
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# put/get many keys
+##
+for ( 0 .. $max_keys ) {
+ $db->put( $_ => $_ * 2 );
+}
+
+for ( 0 .. $max_keys ) {
+ is( $db->get( $_ ), $_ * 2, "The ${_}th value is correct" );
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 4;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# cause an error
+##
+eval { $db->push("foo"); }; # ERROR -- array-only method
+
+ok( $db->error() );
+
+$db->clear_error();
+
+ok( !$db->error() );
+undef $db;
+
+open FH, '>t/test.db';
+print FH 'DPDB';
+close FH;
+$db = DBM::Deep->new( "t/test.db" );
+TODO: {
+ local $TODO = "The return value from load_tag() isn't checked in open()";
+ ok( $db->error() );
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test;
+BEGIN { plan tests => 2 }
+
+use DBM::Deep;
+
+##
+# basic file open
+##
+unlink "t/test.db";
+my $db = new DBM::Deep(
+ file => "t/test.db",
+ locking => 1
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# basic put/get
+##
+$db->{key1} = "value1";
+ok( $db->{key1} eq "value1" );
+
+##
+# explicit lock
+##
+$db->lock( DBM::Deep::LOCK_EX );
+$db->{key1} = "value2";
+$db->unlock();
+ok( $db->{key1} eq "value2" );
+
+##
+# close, delete file, exit
+##
+undef $db;
+unlink "t/test.db";
+exit(0);
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+
+my $max_levels = 1000;
+
+plan tests => $max_levels + 5;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db"
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# basic deep hash
+##
+$db->{company} = {};
+$db->{company}->{name} = "My Co.";
+$db->{company}->{employees} = {};
+$db->{company}->{employees}->{"Henry Higgins"} = {};
+$db->{company}->{employees}->{"Henry Higgins"}->{salary} = 90000;
+
+is( $db->{company}->{name}, "My Co.", "Set and retrieved a second-level value" );
+is( $db->{company}->{employees}->{"Henry Higgins"}->{salary}, 90000, "Set and retrieved a fourth-level value" );
+
+##
+# super deep hash
+##
+$db->{base_level} = {};
+my $temp_db = $db->{base_level};
+
+for my $k ( 0 .. $max_levels ) {
+ $temp_db->{"level$k"} = {};
+ $temp_db = $temp_db->{"level$k"};
+}
+$temp_db->{deepkey} = "deepvalue";
+undef $temp_db;
+
+undef $db;
+$db = DBM::Deep->new(
+ file => "t/test.db"
+);
+
+$temp_db = $db->{base_level};
+for my $k ( 0 .. $max_levels ) {
+ $temp_db = $temp_db->{"level$k"};
+ isa_ok( $temp_db, 'DBM::Deep' ) || die "Whoops!";
+}
+is( $temp_db->{deepkey}, "deepvalue", "And we retrieved the value at the bottom of the ocean" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+
+my $max_levels = 1000;
+
+plan tests => $max_levels + 3;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ type => DBM::Deep->TYPE_ARRAY,
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+$db->[0] = [];
+my $temp_db = $db->[0];
+for my $k ( 0 .. $max_levels ) {
+ $temp_db->[$k] = [];
+ $temp_db = $temp_db->[$k];
+}
+$temp_db->[0] = "deepvalue";
+undef $temp_db;
+
+undef $db;
+$db = DBM::Deep->new(
+ file => "t/test.db",
+ type => DBM::Deep->TYPE_ARRAY,
+);
+
+$temp_db = $db->[0];
+for my $k ( 0 .. $max_levels ) {
+ $temp_db = $temp_db->[$k];
+ isa_ok( $temp_db, 'DBM::Deep' ) || die "Whoops!";
+}
+is( $temp_db->[0], "deepvalue", "And we retrieved the value at the bottom of the ocean" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 14;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db"
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# large keys
+##
+my $key1 = "Now is the time for all good men to come to the aid of their country." x 100;
+my $key2 = "The quick brown fox jumped over the lazy, sleeping dog." x 1000;
+my $key3 = "Lorem dolor ipsum latinum suckum causum Ium cannotum rememberum squatum." x 1000;
+
+$db->put($key1, "value1");
+$db->store($key2, "value2");
+$db->{$key3} = "value3";
+
+is( $db->{$key1}, 'value1', "Hash retrieval of put()" );
+is( $db->{$key2}, 'value2', "Hash retrieval of store()" );
+is( $db->{$key3}, 'value3', "Hash retrieval of hashstore" );
+is( $db->get($key1), 'value1', "get() retrieval of put()" );
+is( $db->get($key2), 'value2', "get() retrieval of store()" );
+is( $db->get($key3), 'value3', "get() retrieval of hashstore" );
+is( $db->fetch($key1), 'value1', "fetch() retrieval of put()" );
+is( $db->fetch($key2), 'value2', "fetch() retrieval of store()" );
+is( $db->fetch($key3), 'value3', "fetch() retrieval of hashstore" );
+
+my $test_key = $db->first_key();
+ok(
+ ($test_key eq $key1) ||
+ ($test_key eq $key2) ||
+ ($test_key eq $key3)
+);
+
+$test_key = $db->next_key($test_key);
+ok(
+ ($test_key eq $key1) ||
+ ($test_key eq $key2) ||
+ ($test_key eq $key3)
+);
+
+$test_key = $db->next_key($test_key);
+ok(
+ ($test_key eq $key1) ||
+ ($test_key eq $key2) ||
+ ($test_key eq $key3)
+);
+
+$test_key = $db->next_key($test_key);
+ok( !$test_key );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 5;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ autoflush => 1,
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# create some unused space
+##
+$db->{key1} = "value1";
+$db->{key2} = "value2";
+
+$db->{a} = {};
+$db->{a}{b} = [];
+$db->{a}{c} = 'value2';
+
+my $b = $db->{a}->{b};
+$b->[0] = 1;
+$b->[1] = 2;
+$b->[2] = {};
+$b->[2]->{c} = [];
+
+my $c = $b->[2]->{c};
+$c->[0] = 'd';
+$c->[1] = {};
+$c->[1]->{e} = 'f';
+
+undef $c;
+undef $b;
+
+delete $db->{key2};
+delete $db->{a}{b};
+
+##
+# take byte count readings before, and after optimize
+##
+my $before = (stat($db->fh()))[7];
+my $result = $db->optimize();
+my $after = (stat($db->fh()))[7];
+
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+ok( $result );
+ok( $after < $before ); # make sure file shrunk
+
+is( $db->{key1}, 'value1', "key1's value is still there after optimize" );
+is( $db->{a}{c}, 'value2', "key2's value is still there after optimize" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 14;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+$db->{key1} = "value1";
+
+##
+# clone db handle, make sure both are usable
+##
+my $clone = $db->clone();
+
+is($clone->{key1}, "value1");
+
+$clone->{key2} = "value2";
+$db->{key3} = "value3";
+
+is($db->{key1}, "value1");
+is($db->{key2}, "value2");
+is($db->{key3}, "value3");
+
+is($clone->{key1}, "value1");
+is($clone->{key2}, "value2");
+is($clone->{key3}, "value3");
+
+$db->close;
+undef $db;
+
+is($clone->{key1}, "value1");
+is($clone->{key2}, "value2");
+is($clone->{key3}, "value3");
+
+undef $clone;
+
+$db = DBM::Deep->new(
+ file => "t/test.db",
+);
+
+is($db->{key1}, "value1");
+is($db->{key2}, "value2");
+is($db->{key3}, "value3");
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 2;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+ autoflush => 1
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+$db->{key1} = "value1";
+$db->{key2} = "value2";
+my $before = (stat($db->fh()))[7];
+undef $db;
+
+##
+# set pack to 2-byte (16-bit) words
+##
+DBM::Deep::set_pack(2, 'S');
+
+unlink "t/test.db";
+$db = DBM::Deep->new(
+ file => "t/test.db",
+ autoflush => 1
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+$db->{key1} = "value1";
+$db->{key2} = "value2";
+my $after = (stat($db->fh()))[7];
+undef $db;
+
+ok( $after < $before );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 17;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new(
+ file => "t/test.db",
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+ok( !$db->set_filter( 'floober', sub {} ), "floober isn't a value filter key" );
+
+##
+# First try store filters only (values will be unfiltered)
+##
+ok( $db->set_filter( 'store_key', \&my_filter_store_key ), "set the store_key filter" );
+ok( $db->set_filter( 'store_value', \&my_filter_store_value ), "set the store_value filter" );
+
+$db->{key1} = "value1";
+$db->{key2} = "value2";
+
+is($db->{key1}, "MYFILTERvalue1", "The value for key1 was filtered correctly" );
+is($db->{key2}, "MYFILTERvalue2", "The value for key2 was filtered correctly" );
+
+##
+# Now try fetch filters as well
+##
+ok( $db->set_filter( 'fetch_key', \&my_filter_fetch_key ), "Set the fetch_key filter" );
+ok( $db->set_filter( 'fetch_value', \&my_filter_fetch_value), "Set the fetch_value filter" );
+
+is($db->{key1}, "value1", "Fetchfilters worked right");
+is($db->{key2}, "value2", "Fetchfilters worked right");
+
+##
+# Try fetching keys as well as values
+##
+my $first_key = $db->first_key();
+my $next_key = $db->next_key($first_key);
+
+ok(
+ (($first_key eq "key1") || ($first_key eq "key2")) &&
+ (($next_key eq "key1") || ($next_key eq "key2"))
+);
+
+##
+# Now clear all filters, and make sure all is unfiltered
+##
+ok( $db->set_filter( 'store_key', undef ), "Unset store_key filter" );
+ok( $db->set_filter( 'store_value', undef ), "Unset store_value filter" );
+ok( $db->set_filter( 'fetch_key', undef ), "Unset fetch_key filter" );
+ok( $db->set_filter( 'fetch_value', undef ), "Unset fetch_value filter" );
+
+is($db->{MYFILTERkey1}, "MYFILTERvalue1");
+is($db->{MYFILTERkey2}, "MYFILTERvalue2");
+
+sub my_filter_store_key { return 'MYFILTER' . $_[0]; }
+sub my_filter_store_value { return 'MYFILTER' . $_[0]; }
+
+sub my_filter_fetch_key { $_[0] =~ s/^MYFILTER//; return $_[0]; }
+sub my_filter_fetch_value { $_[0] =~ s/^MYFILTER//; return $_[0]; }
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test;
+BEGIN { plan tests => 13 }
+
+use DBM::Deep;
+
+my $salt = 38473827;
+
+##
+# basic file open
+##
+unlink "t/test.db";
+my $db = new DBM::Deep(
+ file => "t/test.db"
+);
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# Set digest handler
+##
+DBM::Deep::set_digest( \&my_digest, 8 );
+
+##
+# put/get key
+##
+$db->{key1} = "value1";
+ok( $db->{key1} eq "value1" );
+
+$db->put("key2", "value2");
+ok( $db->get("key2") eq "value2" );
+
+##
+# key exists
+##
+ok( $db->exists("key1") );
+ok( exists $db->{key2} );
+
+##
+# count keys
+##
+ok( scalar keys %$db == 2 );
+
+##
+# step through keys
+##
+my $temphash = {};
+while ( my ($key, $value) = each %$db ) {
+ $temphash->{$key} = $value;
+}
+
+ok( ($temphash->{key1} eq "value1") && ($temphash->{key2} eq "value2") );
+
+$temphash = {};
+my $key = $db->first_key();
+while ($key) {
+ $temphash->{$key} = $db->get($key);
+ $key = $db->next_key($key);
+}
+
+ok( ($temphash->{key1} eq "value1") && ($temphash->{key2} eq "value2") );
+
+##
+# delete keys
+##
+ok( delete $db->{key1} );
+ok( $db->delete("key2") );
+
+ok( scalar keys %$db == 0 );
+
+##
+# delete all keys
+##
+$db->put("another", "value");
+$db->clear();
+
+ok( scalar keys %$db == 0 );
+
+##
+# replace key
+##
+$db->put("key1", "value1");
+$db->put("key1", "value2");
+
+ok( $db->get("key1") eq "value2" );
+
+$db->put("key1", "value222222222222222222222222");
+
+ok( $db->get("key1") eq "value222222222222222222222222" );
+
+##
+# close, delete file, exit
+##
+undef $db;
+unlink "t/test.db";
+exit(0);
+
+sub my_digest {
+ ##
+ # Warning: This digest function is for testing ONLY
+ # It is NOT intended for actual use
+ ##
+ my $key = shift;
+ my $num = $salt;
+
+ for (my $k=0; $k<length($key); $k++) {
+ $num += ord( substr($key, $k, 1) );
+ }
+
+ return sprintf("%00000008d", $num);
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 13;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# put/get simple keys
+##
+$db->{key1} = "value1";
+$db->{key2} = "value2";
+
+##
+# Insert circular reference
+##
+$db->{circle} = $db;
+
+##
+# Make sure keys exist in both places
+##
+is( $db->{key1}, 'value1', "The value is there directly" );
+is( $db->{circle}{key1}, 'value1', "The value is there in one loop of the circle" );
+is( $db->{circle}{circle}{key1}, 'value1', "The value is there in two loops of the circle" );
+is( $db->{circle}{circle}{circle}{key1}, 'value1', "The value is there in three loops of the circle" );
+
+##
+# Make sure changes are reflected in both places
+##
+$db->{key1} = "another value";
+
+is( $db->{key1}, 'another value', "The value is there directly" );
+is( $db->{circle}{key1}, 'another value', "The value is there in one loop of the circle" );
+is( $db->{circle}{circle}{key1}, 'another value', "The value is there in two loops of the circle" );
+is( $db->{circle}{circle}{circle}{key1}, 'another value', "The value is there in three loops of the circle" );
+
+$db->{circle}{circle}{circle}{circle}{key1} = "circles";
+
+is( $db->{key1}, 'circles', "The value is there directly" );
+is( $db->{circle}{key1}, 'circles', "The value is there in one loop of the circle" );
+is( $db->{circle}{circle}{key1}, 'circles', "The value is there in two loops of the circle" );
+is( $db->{circle}{circle}{circle}{key1}, 'circles', "The value is there in three loops of the circle" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 2;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# Create structure in memory
+##
+my $struct = {
+ key1 => "value1",
+ key2 => "value2",
+ array1 => [ "elem0", "elem1", "elem2" ],
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2"
+ }
+};
+
+##
+# Import entire thing
+##
+$db->import( $struct );
+undef $struct;
+
+##
+# Make sure everything is there
+##
+ok(
+ ($db->{key1} eq "value1") &&
+ ($db->{key2} eq "value2") &&
+ ($db->{array1} &&
+ ($db->{array1}->[0] eq "elem0") &&
+ ($db->{array1}->[1] eq "elem1") &&
+ ($db->{array1}->[2] eq "elem2")
+ ) &&
+ ($db->{hash1} &&
+ ($db->{hash1}->{subkey1} eq "subvalue1") &&
+ ($db->{hash1}->{subkey2} eq "subvalue2")
+ )
+);
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 2;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# Create structure in DB
+##
+$db->import(
+ key1 => "value1",
+ key2 => "value2",
+ array1 => [ "elem0", "elem1", "elem2" ],
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2"
+ }
+);
+
+##
+# Export entire thing
+##
+my $struct = $db->export();
+
+##
+# close, delete file
+##
+undef $db;
+unlink "t/test.db";
+
+##
+# Make sure everything is here, outside DB
+##
+ok(
+ ($struct->{key1} eq "value1") &&
+ ($struct->{key2} eq "value2") &&
+ ($struct->{array1} &&
+ ($struct->{array1}->[0] eq "elem0") &&
+ ($struct->{array1}->[1] eq "elem1") &&
+ ($struct->{array1}->[2] eq "elem2")
+ ) &&
+ ($struct->{hash1} &&
+ ($struct->{hash1}->{subkey1} eq "subvalue1") &&
+ ($struct->{hash1}->{subkey2} eq "subvalue2")
+ )
+);
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 5;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+unlink "t/test2.db";
+my $db2 = DBM::Deep->new( "t/test2.db" );
+if ($db2->error()) {
+ die "ERROR: " . $db2->error();
+}
+
+##
+# Create structure in $db
+##
+$db->import(
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2"
+ }
+);
+
+is( $db->{hash1}{subkey1}, 'subvalue1', "Value imported correctly" );
+is( $db->{hash1}{subkey2}, 'subvalue2', "Value imported correctly" );
+
+##
+# Cross-ref nested hash accross DB objects
+##
+$db2->{copy} = $db->{hash1};
+
+##
+# close, delete $db
+##
+undef $db;
+unlink "t/test.db";
+
+##
+# Make sure $db2 has copy of $db's hash structure
+##
+is( $db2->{copy}{subkey1}, 'subvalue1', "Value copied correctly" );
+is( $db2->{copy}{subkey2}, 'subvalue2', "Value copied correctly" );
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+BEGIN { plan tests => 10 }
+
+use DBM::Deep;
+
+##
+# testing the various modes of opening a file
+##
+{
+ unlink "t/test.db";
+ my %hash;
+ my $db = tie %hash, 'DBM::Deep', 't/test.db';
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
+
+{
+ unlink "t/test.db";
+ my %hash;
+ my $db = tie %hash, 'DBM::Deep', {
+ file => 't/test.db',
+ };
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
+
+{
+ unlink "t/test.db";
+ my @array;
+ my $db = tie @array, 'DBM::Deep', 't/test.db';
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+
+ TODO: {
+ local $TODO = "TIE_ARRAY doesn't set the type correctly";
+ is( $db->{type}, DBM::Deep->TYPE_ARRAY, "TIE_ARRAY sets the correct type" );
+ }
+}
+
+{
+ unlink "t/test.db";
+ my @array;
+ my $db = tie @array, 'DBM::Deep', {
+ file => 't/test.db',
+ };
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+
+ TODO: {
+ local $TODO = "TIE_ARRAY doesn't set the type correctly";
+ is( $db->{type}, DBM::Deep->TYPE_ARRAY, "TIE_ARRAY sets the correct type" );
+ }
+}
+
+# These are testing the naive use of ref() within TIEHASH and TIEARRAY.
+# They should be doing (Scalar::Util::reftype($_[0]) eq 'HASH') and then
+# erroring out if it's not.
+TODO: {
+ todo_skip "Naive use of ref()", 1;
+ unlink "t/test.db";
+ my %hash;
+ my $db = tie %hash, 'DBM::Deep', [
+ file => 't/test.db',
+ ];
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
+
+TODO: {
+ todo_skip "Naive use of ref()", 1;
+ unlink "t/test.db";
+ my @array;
+ my $db = tie @array, 'DBM::Deep', [
+ file => 't/test.db',
+ ];
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
+
+# These are testing the naive use of the {@_} construct within TIEHASH and
+# TIEARRAY. Instead, they should be checking (@_ % 2 == 0) and erroring out
+# if it's not.
+TODO: {
+ todo_skip( "Naive use of {\@_}", 1 );
+ unlink "t/test.db";
+ my %hash;
+ my $db = tie %hash, 'DBM::Deep',
+ undef, file => 't/test.db'
+ ;
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
+
+TODO: {
+ todo_skip( "Naive use of {\@_}", 1 );
+ unlink "t/test.db";
+ my @array;
+ my $db = tie @array, 'DBM::Deep',
+ undef, file => 't/test.db'
+ ;
+
+ if ($db->error()) {
+ print "ERROR: " . $db->error();
+ ok(0);
+ exit(0);
+ }
+ else { ok(1); }
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+use Test::Exception;
+
+plan tests => 7;
+
+use_ok( 'DBM::Deep' );
+
+# How should one test for creation failure with the tie mechanism?
+
+unlink "t/test.db";
+
+{
+ my %hash;
+ tie %hash, 'DBM::Deep', "t/test.db";
+
+ $hash{key1} = 'value';
+ is( $hash{key1}, 'value', 'Set and retrieved key1' );
+}
+
+{
+ my %hash;
+ tie %hash, 'DBM::Deep', "t/test.db";
+
+ is( $hash{key1}, 'value', 'Set and retrieved key1' );
+
+ is( keys %hash, 1, "There's one key so far" );
+ ok( exists $hash{key1}, "... and it's key1" );
+}
+
+TODO: {
+ local $TODO = "Sig doesn't match, but it's legal??";
+ my @array;
+ throws_ok {
+ tie @array, 'DBM::Deep', {
+ file => 't/test.db',
+ type => DBM::Deep->TYPE_ARRAY,
+ };
+ } qr/DBM::Deep: Cannot open a hash-based file with an array/, "\$SIG_TYPE doesn't match file's type";
+
+ unlink "t/test.db";
+ DBM::Deep->new( file => 't/test.db', type => DBM::Deep->TYPE_ARRAY );
+
+ my %hash;
+ throws_ok {
+ tie %hash, 'DBM::Deep', {
+ file => 't/test.db',
+ type => DBM::Deep->TYPE_HASH,
+ };
+ } qr/DBM::Deep: Cannot open a array-based file with a hash/, "\$SIG_TYPE doesn't match file's type";
+}
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More;
+use Test::Exception;
+
+plan tests => 5;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = new DBM::Deep "t/test.db";
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+$db->{key1} = "value1";
+is( $db->{key1}, "value1", "Value set correctly" );
+
+# Testing to verify that the close() will occur if open is called on an open DB.
+$db->open;
+
+is( $db->{key1}, "value1", "Value still set after re-open" );
+
+throws_ok {
+ my $db = DBM::Deep->new( 't' );
+} qr/^DBM::Deep: Cannot open file: t: /, "Can't open a file we aren't allowed to touch";
+
+throws_ok {
+ my $db = DBM::Deep->new( __FILE__ );
+} qr/^DBM::Deep: Signature not found -- file is not a Deep DB/, "Only DBM::Deep DB files will be opened";
--- /dev/null
+##
+# DBM::Deep Test
+##
+use strict;
+use Test::More tests => 12;
+
+use_ok( 'DBM::Deep' );
+
+unlink "t/test.db";
+my $db = DBM::Deep->new( "t/test.db" );
+if ($db->error()) {
+ die "ERROR: " . $db->error();
+}
+
+##
+# Create structure in $db
+##
+$db->import(
+ hash1 => {
+ subkey1 => "subvalue1",
+ subkey2 => "subvalue2",
+ },
+ hash2 => {
+ subkey3 => 'subvalue3',
+ },
+);
+
+is( $db->{hash1}{subkey1}, 'subvalue1', "Value imported correctly" );
+is( $db->{hash1}{subkey2}, 'subvalue2', "Value imported correctly" );
+
+$db->{copy} = $db->{hash1};
+
+is( $db->{copy}{subkey1}, 'subvalue1', "Value copied correctly" );
+is( $db->{copy}{subkey2}, 'subvalue2', "Value copied correctly" );
+
+$db->{copy}{subkey1} = "another value";
+is( $db->{copy}{subkey1}, 'another value', "New value is set correctly" );
+is( $db->{hash1}{subkey1}, 'another value', "Old value is set to the new one" );
+
+is( scalar(keys %{$db->{hash1}}), 2, "Start with 2 keys in the original" );
+is( scalar(keys %{$db->{copy}}), 2, "Start with 2 keys in the copy" );
+
+delete $db->{copy}{subkey2};
+
+is( scalar(keys %{$db->{copy}}), 1, "Now only have 1 key in the copy" );
+is( scalar(keys %{$db->{hash1}}), 1, "... and only 1 key in the original" );
+
+$db->{copy} = $db->{hash2};
+is( $db->{copy}{subkey3}, 'subvalue3', "After the second copy, we're still good" );