Broke out a large portion of POD into the Cookbook and Internals POD files
[dbsrgits/DBM-Deep.git] / lib / DBM / Deep / Internals.pod
CommitLineData
d8db2929 1=head1 NAME
2
3DBM::Deep::Internals
4
5=head1 DESCRIPTION
6
7This is a document describing the internal workings of L<DBM::Deep/>. It is
8not necessary to read this document if you only intend to be a user. This
9document is intended for people who either want a deeper understanding of
10specifics of how L<DBM::Deep/> works or who wish to help program
11L<DBM::Deep/>.
12
13=head1 CLASS LAYOUT
14
15L<DBM::Deep/> is broken up into five classes in three inheritance hierarchies.
16
17=over 4
18
19=item *
20
21L<DBM::Deep/> is the parent of L<DBM::Deep::Array/> and L<DBM::Deep::Hash/>.
22These classes form the immediate interface to the outside world. They are the
23classes that provide the TIE mechanisms as well as the OO methods.
24
25=item *
26
27L<DBM::Deep::Engine/> is the layer that deals with the mechanics of reading
28and writing to the file. This is where the logic of the file layout is
29handled.
30
31=item *
32
33L<DBM::Deep::File/> is the layer that deals with the physical file. As a
34singleton that every other object has a reference to, it also provides a place
35to handle datastructure-wide items, such as transactions.
36
37=back
38
39=head1 FILE LAYOUT
40
41DBM::Deep uses a tagged file layout. Every section has a tag, a size, and then
42the data.
43
44=head2 File header
45
46=over 4
47
48=item * File Signature
49
50The first four bytes are 'DPDB' in network byte order, signifying that this is
51a DBM::Deep file.
52
53=item * File tag/size
54
55This is the tagging of the file header. The file used by versions prior to
561.00 had a different fifth byte, allowing the difference to the determined.
57
58=item * Version
59
60This is four bytes containing the header version. This lets the header change over time.
61
62=item * Transaction information
63
64The current running transactions are stored here, as is the next transaction
65ID.
66
67=item * Constants
68
69These are the file-wide constants that determine how the file is laid out.
70They can only be set upon file creation.
71
72=back
73
74=head2 Index
75
76The Index parts can be tagged either as Hash, Array, or Index. The latter
77is if there was a reindexing due to a bucketlist growing too large. The others
78are the root index for their respective datatypes. The index consists of a
79tag, a size, and then 256 sections containing file locations. Each section
80corresponds to each value representable in a byte.
81
82The index is used as follows - whenever a hashed key is being looked up, the
83first byte is used to determine which location to go to from the root index.
84Then, if that's also an index, the second byte is used, and so forth until a
85bucketlist is found.
86
87=head2 Bucketlist
88
89This is the part that contains the link to the data section. A bucketlist
90defaults to being 16 buckets long (modifiable by the I<max_buckets>
91parameter used when creating a new file). Each bucket contains an MD5 and a
92location of the appropriate key section.
93
94=head2 Key area
95
96This is the part that handles transactional awareness. There are
97I<max_buckets> sections. Each section contains the location to the data
98section, a transaction ID, and whether that transaction considers this key to
99be deleted or not.
100
101=head2 Data area
102
103This is the part that actual stores the key, value, and class (if
104appropriate). The layout is:
105
106=over 4
107
108=item * tag
109
110=item * length of the value
111
112=item * the actual value
113
114=item * keylength
115
116=item * the actual key
117
118=item * a byte indicating if this value has a classname
119
120=item * the classname (if one is there)
121
122=back
123
124The key is stored after the value because the value is requested more often
125than the key.
126
127=head1 PERFORMANCE
128
129L<DBM::Deep/> is written completely in Perl. It also is a multi-process DBM
130that uses the datafile as a method of synchronizing between multiple
131processes. This is unlike most RDBMSes like MySQL and Oracle. Furthermore,
132unlike all RDBMSes, L<DBM::Deep/> stores both the data and the structure of
133that data as it would appear in a Perl program.
134
135=head2 CPU
136
137DBM::Deep attempts to be CPU-light. As it stores all the data on disk,
138DBM::Deep is I/O-bound, not CPU-bound.
139
140=head2 RAM
141
142DBM::Deep uses extremely little RAM relative to the amount of data you can
143access. You can iterate through a million keys (using C<each()>) without
144increasing your memeory usage at all.
145
146=head2 DISK
147
148DBM::Deep is I/O-bound, pure and simple. The faster your disk, the faster
149DBM::Deep will be. Currently, when performing C<my $x = $db-E<gt>{foo}>, there
150are a minimum of 4 seeks and 1332 + N bytes read (where N is the length of your
151data). (All values assume a medium filesize.) The actions take are:
152
153=over 4
154
155=item 1 Lock the file
156
157=item 1 Perform a stat() to determine if the inode has changed
158
159=item 1 Go to the primary index for the $db (1 seek)
160
161=item 1 Read the tag/size of the primary index (5 bytes)
162
163=item 1 Read the body of the primary index (1024 bytes)
164
165=item 1 Go to the bucketlist for this MD5 (1 seek)
166
167=item 1 Read the tag/size of the bucketlist (5 bytes)
168
169=item 1 Read the body of the bucketlist (144 bytes)
170
171=item 1 Go to the keys location for this MD5 (1 seek)
172
173=item 1 Read the tag/size of the keys section (5 bytes)
174
175=item 1 Read the body of the keys location (144 bytes)
176
177=item 1 Go to the data section that corresponds to this transaction ID. (1 seek)
178
179=item 1 Read the tag/size of the data section (5 bytes)
180
181=item 1 Read the value for this data (N bytes)
182
183=item 1 Unlock the file
184
185=back
186
187Every additional level of indexing (if there are enough keys) requires an
188additional seek and the reading of 1029 additional bytes. If the value is
189blessed, an additional 1 seek and 9 + M bytes are read (where M is the length
190of the classname).
191
192Arrays are (currently) even worse because they're considered "funny hashes"
193with the length stored as just another key. This means that if you do any sort
194of lookup with a negative index, this entire process is performed twice - once
195for the length and once for the value.
196
197=cut