diff options
Diffstat (limited to 'bzrlib/btree_index.py')
-rw-r--r-- | bzrlib/btree_index.py | 1608 |
1 files changed, 1608 insertions, 0 deletions
diff --git a/bzrlib/btree_index.py b/bzrlib/btree_index.py new file mode 100644 index 0000000..6d60490 --- /dev/null +++ b/bzrlib/btree_index.py @@ -0,0 +1,1608 @@ +# Copyright (C) 2008-2011 Canonical Ltd +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 2 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program; if not, write to the Free Software +# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +# + +"""B+Tree indices""" + +from __future__ import absolute_import + +import cStringIO + +from bzrlib.lazy_import import lazy_import +lazy_import(globals(), """ +import bisect +import math +import tempfile +import zlib +""") + +from bzrlib import ( + chunk_writer, + debug, + errors, + fifo_cache, + index, + lru_cache, + osutils, + static_tuple, + trace, + transport, + ) +from bzrlib.index import _OPTION_NODE_REFS, _OPTION_KEY_ELEMENTS, _OPTION_LEN + + +_BTSIGNATURE = "B+Tree Graph Index 2\n" +_OPTION_ROW_LENGTHS = "row_lengths=" +_LEAF_FLAG = "type=leaf\n" +_INTERNAL_FLAG = "type=internal\n" +_INTERNAL_OFFSET = "offset=" + +_RESERVED_HEADER_BYTES = 120 +_PAGE_SIZE = 4096 + +# 4K per page: 4MB - 1000 entries +_NODE_CACHE_SIZE = 1000 + + +class _BuilderRow(object): + """The stored state accumulated while writing out a row in the index. + + :ivar spool: A temporary file used to accumulate nodes for this row + in the tree. + :ivar nodes: The count of nodes emitted so far. + """ + + def __init__(self): + """Create a _BuilderRow.""" + self.nodes = 0 + self.spool = None# tempfile.TemporaryFile(prefix='bzr-index-row-') + self.writer = None + + def finish_node(self, pad=True): + byte_lines, _, padding = self.writer.finish() + if self.nodes == 0: + self.spool = cStringIO.StringIO() + # padded note: + self.spool.write("\x00" * _RESERVED_HEADER_BYTES) + elif self.nodes == 1: + # We got bigger than 1 node, switch to a temp file + spool = tempfile.TemporaryFile(prefix='bzr-index-row-') + spool.write(self.spool.getvalue()) + self.spool = spool + skipped_bytes = 0 + if not pad and padding: + del byte_lines[-1] + skipped_bytes = padding + self.spool.writelines(byte_lines) + remainder = (self.spool.tell() + skipped_bytes) % _PAGE_SIZE + if remainder != 0: + raise AssertionError("incorrect node length: %d, %d" + % (self.spool.tell(), remainder)) + self.nodes += 1 + self.writer = None + + +class _InternalBuilderRow(_BuilderRow): + """The stored state accumulated while writing out internal rows.""" + + def finish_node(self, pad=True): + if not pad: + raise AssertionError("Must pad internal nodes only.") + _BuilderRow.finish_node(self) + + +class _LeafBuilderRow(_BuilderRow): + """The stored state accumulated while writing out a leaf rows.""" + + +class BTreeBuilder(index.GraphIndexBuilder): + """A Builder for B+Tree based Graph indices. + + The resulting graph has the structure: + + _SIGNATURE OPTIONS NODES + _SIGNATURE := 'B+Tree Graph Index 1' NEWLINE + OPTIONS := REF_LISTS KEY_ELEMENTS LENGTH + REF_LISTS := 'node_ref_lists=' DIGITS NEWLINE + KEY_ELEMENTS := 'key_elements=' DIGITS NEWLINE + LENGTH := 'len=' DIGITS NEWLINE + ROW_LENGTHS := 'row_lengths' DIGITS (COMMA DIGITS)* + NODES := NODE_COMPRESSED* + NODE_COMPRESSED:= COMPRESSED_BYTES{4096} + NODE_RAW := INTERNAL | LEAF + INTERNAL := INTERNAL_FLAG POINTERS + LEAF := LEAF_FLAG ROWS + KEY_ELEMENT := Not-whitespace-utf8 + KEY := KEY_ELEMENT (NULL KEY_ELEMENT)* + ROWS := ROW* + ROW := KEY NULL ABSENT? NULL REFERENCES NULL VALUE NEWLINE + ABSENT := 'a' + REFERENCES := REFERENCE_LIST (TAB REFERENCE_LIST){node_ref_lists - 1} + REFERENCE_LIST := (REFERENCE (CR REFERENCE)*)? + REFERENCE := KEY + VALUE := no-newline-no-null-bytes + """ + + def __init__(self, reference_lists=0, key_elements=1, spill_at=100000): + """See GraphIndexBuilder.__init__. + + :param spill_at: Optional parameter controlling the maximum number + of nodes that BTreeBuilder will hold in memory. + """ + index.GraphIndexBuilder.__init__(self, reference_lists=reference_lists, + key_elements=key_elements) + self._spill_at = spill_at + self._backing_indices = [] + # A map of {key: (node_refs, value)} + self._nodes = {} + # Indicate it hasn't been built yet + self._nodes_by_key = None + self._optimize_for_size = False + + def add_node(self, key, value, references=()): + """Add a node to the index. + + If adding the node causes the builder to reach its spill_at threshold, + disk spilling will be triggered. + + :param key: The key. keys are non-empty tuples containing + as many whitespace-free utf8 bytestrings as the key length + defined for this index. + :param references: An iterable of iterables of keys. Each is a + reference to another key. + :param value: The value to associate with the key. It may be any + bytes as long as it does not contain \\0 or \\n. + """ + # Ensure that 'key' is a StaticTuple + key = static_tuple.StaticTuple.from_sequence(key).intern() + # we don't care about absent_references + node_refs, _ = self._check_key_ref_value(key, references, value) + if key in self._nodes: + raise errors.BadIndexDuplicateKey(key, self) + self._nodes[key] = static_tuple.StaticTuple(node_refs, value) + if self._nodes_by_key is not None and self._key_length > 1: + self._update_nodes_by_key(key, value, node_refs) + if len(self._nodes) < self._spill_at: + return + self._spill_mem_keys_to_disk() + + def _spill_mem_keys_to_disk(self): + """Write the in memory keys down to disk to cap memory consumption. + + If we already have some keys written to disk, we will combine them so + as to preserve the sorted order. The algorithm for combining uses + powers of two. So on the first spill, write all mem nodes into a + single index. On the second spill, combine the mem nodes with the nodes + on disk to create a 2x sized disk index and get rid of the first index. + On the third spill, create a single new disk index, which will contain + the mem nodes, and preserve the existing 2x sized index. On the fourth, + combine mem with the first and second indexes, creating a new one of + size 4x. On the fifth create a single new one, etc. + """ + if self._combine_backing_indices: + (new_backing_file, size, + backing_pos) = self._spill_mem_keys_and_combine() + else: + new_backing_file, size = self._spill_mem_keys_without_combining() + # Note: The transport here isn't strictly needed, because we will use + # direct access to the new_backing._file object + new_backing = BTreeGraphIndex(transport.get_transport_from_path('.'), + '<temp>', size) + # GC will clean up the file + new_backing._file = new_backing_file + if self._combine_backing_indices: + if len(self._backing_indices) == backing_pos: + self._backing_indices.append(None) + self._backing_indices[backing_pos] = new_backing + for backing_pos in range(backing_pos): + self._backing_indices[backing_pos] = None + else: + self._backing_indices.append(new_backing) + self._nodes = {} + self._nodes_by_key = None + + def _spill_mem_keys_without_combining(self): + return self._write_nodes(self._iter_mem_nodes(), allow_optimize=False) + + def _spill_mem_keys_and_combine(self): + iterators_to_combine = [self._iter_mem_nodes()] + pos = -1 + for pos, backing in enumerate(self._backing_indices): + if backing is None: + pos -= 1 + break + iterators_to_combine.append(backing.iter_all_entries()) + backing_pos = pos + 1 + new_backing_file, size = \ + self._write_nodes(self._iter_smallest(iterators_to_combine), + allow_optimize=False) + return new_backing_file, size, backing_pos + + def add_nodes(self, nodes): + """Add nodes to the index. + + :param nodes: An iterable of (key, node_refs, value) entries to add. + """ + if self.reference_lists: + for (key, value, node_refs) in nodes: + self.add_node(key, value, node_refs) + else: + for (key, value) in nodes: + self.add_node(key, value) + + def _iter_mem_nodes(self): + """Iterate over the nodes held in memory.""" + nodes = self._nodes + if self.reference_lists: + for key in sorted(nodes): + references, value = nodes[key] + yield self, key, value, references + else: + for key in sorted(nodes): + references, value = nodes[key] + yield self, key, value + + def _iter_smallest(self, iterators_to_combine): + if len(iterators_to_combine) == 1: + for value in iterators_to_combine[0]: + yield value + return + current_values = [] + for iterator in iterators_to_combine: + try: + current_values.append(iterator.next()) + except StopIteration: + current_values.append(None) + last = None + while True: + # Decorate candidates with the value to allow 2.4's min to be used. + candidates = [(item[1][1], item) for item + in enumerate(current_values) if item[1] is not None] + if not len(candidates): + return + selected = min(candidates) + # undecorate back to (pos, node) + selected = selected[1] + if last == selected[1][1]: + raise errors.BadIndexDuplicateKey(last, self) + last = selected[1][1] + # Yield, with self as the index + yield (self,) + selected[1][1:] + pos = selected[0] + try: + current_values[pos] = iterators_to_combine[pos].next() + except StopIteration: + current_values[pos] = None + + def _add_key(self, string_key, line, rows, allow_optimize=True): + """Add a key to the current chunk. + + :param string_key: The key to add. + :param line: The fully serialised key and value. + :param allow_optimize: If set to False, prevent setting the optimize + flag when writing out. This is used by the _spill_mem_keys_to_disk + functionality. + """ + new_leaf = False + if rows[-1].writer is None: + # opening a new leaf chunk; + new_leaf = True + for pos, internal_row in enumerate(rows[:-1]): + # flesh out any internal nodes that are needed to + # preserve the height of the tree + if internal_row.writer is None: + length = _PAGE_SIZE + if internal_row.nodes == 0: + length -= _RESERVED_HEADER_BYTES # padded + if allow_optimize: + optimize_for_size = self._optimize_for_size + else: + optimize_for_size = False + internal_row.writer = chunk_writer.ChunkWriter(length, 0, + optimize_for_size=optimize_for_size) + internal_row.writer.write(_INTERNAL_FLAG) + internal_row.writer.write(_INTERNAL_OFFSET + + str(rows[pos + 1].nodes) + "\n") + # add a new leaf + length = _PAGE_SIZE + if rows[-1].nodes == 0: + length -= _RESERVED_HEADER_BYTES # padded + rows[-1].writer = chunk_writer.ChunkWriter(length, + optimize_for_size=self._optimize_for_size) + rows[-1].writer.write(_LEAF_FLAG) + if rows[-1].writer.write(line): + # if we failed to write, despite having an empty page to write to, + # then line is too big. raising the error avoids infinite recursion + # searching for a suitably large page that will not be found. + if new_leaf: + raise errors.BadIndexKey(string_key) + # this key did not fit in the node: + rows[-1].finish_node() + key_line = string_key + "\n" + new_row = True + for row in reversed(rows[:-1]): + # Mark the start of the next node in the node above. If it + # doesn't fit then propagate upwards until we find one that + # it does fit into. + if row.writer.write(key_line): + row.finish_node() + else: + # We've found a node that can handle the pointer. + new_row = False + break + # If we reached the current root without being able to mark the + # division point, then we need a new root: + if new_row: + # We need a new row + if 'index' in debug.debug_flags: + trace.mutter('Inserting new global row.') + new_row = _InternalBuilderRow() + reserved_bytes = 0 + rows.insert(0, new_row) + # This will be padded, hence the -100 + new_row.writer = chunk_writer.ChunkWriter( + _PAGE_SIZE - _RESERVED_HEADER_BYTES, + reserved_bytes, + optimize_for_size=self._optimize_for_size) + new_row.writer.write(_INTERNAL_FLAG) + new_row.writer.write(_INTERNAL_OFFSET + + str(rows[1].nodes - 1) + "\n") + new_row.writer.write(key_line) + self._add_key(string_key, line, rows, allow_optimize=allow_optimize) + + def _write_nodes(self, node_iterator, allow_optimize=True): + """Write node_iterator out as a B+Tree. + + :param node_iterator: An iterator of sorted nodes. Each node should + match the output given by iter_all_entries. + :param allow_optimize: If set to False, prevent setting the optimize + flag when writing out. This is used by the _spill_mem_keys_to_disk + functionality. + :return: A file handle for a temporary file containing a B+Tree for + the nodes. + """ + # The index rows - rows[0] is the root, rows[1] is the layer under it + # etc. + rows = [] + # forward sorted by key. In future we may consider topological sorting, + # at the cost of table scans for direct lookup, or a second index for + # direct lookup + key_count = 0 + # A stack with the number of nodes of each size. 0 is the root node + # and must always be 1 (if there are any nodes in the tree). + self.row_lengths = [] + # Loop over all nodes adding them to the bottom row + # (rows[-1]). When we finish a chunk in a row, + # propagate the key that didn't fit (comes after the chunk) to the + # row above, transitively. + for node in node_iterator: + if key_count == 0: + # First key triggers the first row + rows.append(_LeafBuilderRow()) + key_count += 1 + string_key, line = _btree_serializer._flatten_node(node, + self.reference_lists) + self._add_key(string_key, line, rows, allow_optimize=allow_optimize) + for row in reversed(rows): + pad = (type(row) != _LeafBuilderRow) + row.finish_node(pad=pad) + lines = [_BTSIGNATURE] + lines.append(_OPTION_NODE_REFS + str(self.reference_lists) + '\n') + lines.append(_OPTION_KEY_ELEMENTS + str(self._key_length) + '\n') + lines.append(_OPTION_LEN + str(key_count) + '\n') + row_lengths = [row.nodes for row in rows] + lines.append(_OPTION_ROW_LENGTHS + ','.join(map(str, row_lengths)) + '\n') + if row_lengths and row_lengths[-1] > 1: + result = tempfile.NamedTemporaryFile(prefix='bzr-index-') + else: + result = cStringIO.StringIO() + result.writelines(lines) + position = sum(map(len, lines)) + root_row = True + if position > _RESERVED_HEADER_BYTES: + raise AssertionError("Could not fit the header in the" + " reserved space: %d > %d" + % (position, _RESERVED_HEADER_BYTES)) + # write the rows out: + for row in rows: + reserved = _RESERVED_HEADER_BYTES # reserved space for first node + row.spool.flush() + row.spool.seek(0) + # copy nodes to the finalised file. + # Special case the first node as it may be prefixed + node = row.spool.read(_PAGE_SIZE) + result.write(node[reserved:]) + if len(node) == _PAGE_SIZE: + result.write("\x00" * (reserved - position)) + position = 0 # Only the root row actually has an offset + copied_len = osutils.pumpfile(row.spool, result) + if copied_len != (row.nodes - 1) * _PAGE_SIZE: + if type(row) != _LeafBuilderRow: + raise AssertionError("Incorrect amount of data copied" + " expected: %d, got: %d" + % ((row.nodes - 1) * _PAGE_SIZE, + copied_len)) + result.flush() + size = result.tell() + result.seek(0) + return result, size + + def finish(self): + """Finalise the index. + + :return: A file handle for a temporary file containing the nodes added + to the index. + """ + return self._write_nodes(self.iter_all_entries())[0] + + def iter_all_entries(self): + """Iterate over all keys within the index + + :return: An iterable of (index, key, value, reference_lists). There is + no defined order for the result iteration - it will be in the most + efficient order for the index (in this case dictionary hash order). + """ + if 'evil' in debug.debug_flags: + trace.mutter_callsite(3, + "iter_all_entries scales with size of history.") + # Doing serial rather than ordered would be faster; but this shouldn't + # be getting called routinely anyway. + iterators = [self._iter_mem_nodes()] + for backing in self._backing_indices: + if backing is not None: + iterators.append(backing.iter_all_entries()) + if len(iterators) == 1: + return iterators[0] + return self._iter_smallest(iterators) + + def iter_entries(self, keys): + """Iterate over keys within the index. + + :param keys: An iterable providing the keys to be retrieved. + :return: An iterable of (index, key, value, reference_lists). There is no + defined order for the result iteration - it will be in the most + efficient order for the index (keys iteration order in this case). + """ + keys = set(keys) + # Note: We don't use keys.intersection() here. If you read the C api, + # set.intersection(other) special cases when other is a set and + # will iterate the smaller of the two and lookup in the other. + # It does *not* do this for any other type (even dict, unlike + # some other set functions.) Since we expect keys is generally << + # self._nodes, it is faster to iterate over it in a list + # comprehension + nodes = self._nodes + local_keys = [key for key in keys if key in nodes] + if self.reference_lists: + for key in local_keys: + node = nodes[key] + yield self, key, node[1], node[0] + else: + for key in local_keys: + node = nodes[key] + yield self, key, node[1] + # Find things that are in backing indices that have not been handled + # yet. + if not self._backing_indices: + return # We won't find anything there either + # Remove all of the keys that we found locally + keys.difference_update(local_keys) + for backing in self._backing_indices: + if backing is None: + continue + if not keys: + return + for node in backing.iter_entries(keys): + keys.remove(node[1]) + yield (self,) + node[1:] + + def iter_entries_prefix(self, keys): + """Iterate over keys within the index using prefix matching. + + Prefix matching is applied within the tuple of a key, not to within + the bytestring of each key element. e.g. if you have the keys ('foo', + 'bar'), ('foobar', 'gam') and do a prefix search for ('foo', None) then + only the former key is returned. + + :param keys: An iterable providing the key prefixes to be retrieved. + Each key prefix takes the form of a tuple the length of a key, but + with the last N elements 'None' rather than a regular bytestring. + The first element cannot be 'None'. + :return: An iterable as per iter_all_entries, but restricted to the + keys with a matching prefix to those supplied. No additional keys + will be returned, and every match that is in the index will be + returned. + """ + # XXX: To much duplication with the GraphIndex class; consider finding + # a good place to pull out the actual common logic. + keys = set(keys) + if not keys: + return + for backing in self._backing_indices: + if backing is None: + continue + for node in backing.iter_entries_prefix(keys): + yield (self,) + node[1:] + if self._key_length == 1: + for key in keys: + # sanity check + if key[0] is None: + raise errors.BadIndexKey(key) + if len(key) != self._key_length: + raise errors.BadIndexKey(key) + try: + node = self._nodes[key] + except KeyError: + continue + if self.reference_lists: + yield self, key, node[1], node[0] + else: + yield self, key, node[1] + return + for key in keys: + # sanity check + if key[0] is None: + raise errors.BadIndexKey(key) + if len(key) != self._key_length: + raise errors.BadIndexKey(key) + # find what it refers to: + key_dict = self._get_nodes_by_key() + elements = list(key) + # find the subdict to return + try: + while len(elements) and elements[0] is not None: + key_dict = key_dict[elements[0]] + elements.pop(0) + except KeyError: + # a non-existant lookup. + continue + if len(elements): + dicts = [key_dict] + while dicts: + key_dict = dicts.pop(-1) + # can't be empty or would not exist + item, value = key_dict.iteritems().next() + if type(value) == dict: + # push keys + dicts.extend(key_dict.itervalues()) + else: + # yield keys + for value in key_dict.itervalues(): + yield (self, ) + tuple(value) + else: + yield (self, ) + key_dict + + def _get_nodes_by_key(self): + if self._nodes_by_key is None: + nodes_by_key = {} + if self.reference_lists: + for key, (references, value) in self._nodes.iteritems(): + key_dict = nodes_by_key + for subkey in key[:-1]: + key_dict = key_dict.setdefault(subkey, {}) + key_dict[key[-1]] = key, value, references + else: + for key, (references, value) in self._nodes.iteritems(): + key_dict = nodes_by_key + for subkey in key[:-1]: + key_dict = key_dict.setdefault(subkey, {}) + key_dict[key[-1]] = key, value + self._nodes_by_key = nodes_by_key + return self._nodes_by_key + + def key_count(self): + """Return an estimate of the number of keys in this index. + + For InMemoryGraphIndex the estimate is exact. + """ + return len(self._nodes) + sum(backing.key_count() for backing in + self._backing_indices if backing is not None) + + def validate(self): + """In memory index's have no known corruption at the moment.""" + + +class _LeafNode(dict): + """A leaf node for a serialised B+Tree index.""" + + __slots__ = ('min_key', 'max_key', '_keys') + + def __init__(self, bytes, key_length, ref_list_length): + """Parse bytes to create a leaf node object.""" + # splitlines mangles the \r delimiters.. don't use it. + key_list = _btree_serializer._parse_leaf_lines(bytes, + key_length, ref_list_length) + if key_list: + self.min_key = key_list[0][0] + self.max_key = key_list[-1][0] + else: + self.min_key = self.max_key = None + super(_LeafNode, self).__init__(key_list) + self._keys = dict(self) + + def all_items(self): + """Return a sorted list of (key, (value, refs)) items""" + items = self.items() + items.sort() + return items + + def all_keys(self): + """Return a sorted list of all keys.""" + keys = self.keys() + keys.sort() + return keys + + +class _InternalNode(object): + """An internal node for a serialised B+Tree index.""" + + __slots__ = ('keys', 'offset') + + def __init__(self, bytes): + """Parse bytes to create an internal node object.""" + # splitlines mangles the \r delimiters.. don't use it. + self.keys = self._parse_lines(bytes.split('\n')) + + def _parse_lines(self, lines): + nodes = [] + self.offset = int(lines[1][7:]) + as_st = static_tuple.StaticTuple.from_sequence + for line in lines[2:]: + if line == '': + break + nodes.append(as_st(map(intern, line.split('\0'))).intern()) + return nodes + + +class BTreeGraphIndex(object): + """Access to nodes via the standard GraphIndex interface for B+Tree's. + + Individual nodes are held in a LRU cache. This holds the root node in + memory except when very large walks are done. + """ + + def __init__(self, transport, name, size, unlimited_cache=False, + offset=0): + """Create a B+Tree index object on the index name. + + :param transport: The transport to read data for the index from. + :param name: The file name of the index on transport. + :param size: Optional size of the index in bytes. This allows + compatibility with the GraphIndex API, as well as ensuring that + the initial read (to read the root node header) can be done + without over-reading even on empty indices, and on small indices + allows single-IO to read the entire index. + :param unlimited_cache: If set to True, then instead of using an + LRUCache with size _NODE_CACHE_SIZE, we will use a dict and always + cache all leaf nodes. + :param offset: The start of the btree index data isn't byte 0 of the + file. Instead it starts at some point later. + """ + self._transport = transport + self._name = name + self._size = size + self._file = None + self._recommended_pages = self._compute_recommended_pages() + self._root_node = None + self._base_offset = offset + self._leaf_factory = _LeafNode + # Default max size is 100,000 leave values + self._leaf_value_cache = None # lru_cache.LRUCache(100*1000) + if unlimited_cache: + self._leaf_node_cache = {} + self._internal_node_cache = {} + else: + self._leaf_node_cache = lru_cache.LRUCache(_NODE_CACHE_SIZE) + # We use a FIFO here just to prevent possible blowout. However, a + # 300k record btree has only 3k leaf nodes, and only 20 internal + # nodes. A value of 100 scales to ~100*100*100 = 1M records. + self._internal_node_cache = fifo_cache.FIFOCache(100) + self._key_count = None + self._row_lengths = None + self._row_offsets = None # Start of each row, [-1] is the end + + def __eq__(self, other): + """Equal when self and other were created with the same parameters.""" + return ( + type(self) == type(other) and + self._transport == other._transport and + self._name == other._name and + self._size == other._size) + + def __ne__(self, other): + return not self.__eq__(other) + + def _get_and_cache_nodes(self, nodes): + """Read nodes and cache them in the lru. + + The nodes list supplied is sorted and then read from disk, each node + being inserted it into the _node_cache. + + Note: Asking for more nodes than the _node_cache can contain will + result in some of the results being immediately discarded, to prevent + this an assertion is raised if more nodes are asked for than are + cachable. + + :return: A dict of {node_pos: node} + """ + found = {} + start_of_leaves = None + for node_pos, node in self._read_nodes(sorted(nodes)): + if node_pos == 0: # Special case + self._root_node = node + else: + if start_of_leaves is None: + start_of_leaves = self._row_offsets[-2] + if node_pos < start_of_leaves: + self._internal_node_cache[node_pos] = node + else: + self._leaf_node_cache[node_pos] = node + found[node_pos] = node + return found + + def _compute_recommended_pages(self): + """Convert transport's recommended_page_size into btree pages. + + recommended_page_size is in bytes, we want to know how many _PAGE_SIZE + pages fit in that length. + """ + recommended_read = self._transport.recommended_page_size() + recommended_pages = int(math.ceil(recommended_read / + float(_PAGE_SIZE))) + return recommended_pages + + def _compute_total_pages_in_index(self): + """How many pages are in the index. + + If we have read the header we will use the value stored there. + Otherwise it will be computed based on the length of the index. + """ + if self._size is None: + raise AssertionError('_compute_total_pages_in_index should not be' + ' called when self._size is None') + if self._root_node is not None: + # This is the number of pages as defined by the header + return self._row_offsets[-1] + # This is the number of pages as defined by the size of the index. They + # should be indentical. + total_pages = int(math.ceil(self._size / float(_PAGE_SIZE))) + return total_pages + + def _expand_offsets(self, offsets): + """Find extra pages to download. + + The idea is that we always want to make big-enough requests (like 64kB + for http), so that we don't waste round trips. So given the entries + that we already have cached and the new pages being downloaded figure + out what other pages we might want to read. + + See also doc/developers/btree_index_prefetch.txt for more details. + + :param offsets: The offsets to be read + :return: A list of offsets to download + """ + if 'index' in debug.debug_flags: + trace.mutter('expanding: %s\toffsets: %s', self._name, offsets) + + if len(offsets) >= self._recommended_pages: + # Don't add more, we are already requesting more than enough + if 'index' in debug.debug_flags: + trace.mutter(' not expanding large request (%s >= %s)', + len(offsets), self._recommended_pages) + return offsets + if self._size is None: + # Don't try anything, because we don't know where the file ends + if 'index' in debug.debug_flags: + trace.mutter(' not expanding without knowing index size') + return offsets + total_pages = self._compute_total_pages_in_index() + cached_offsets = self._get_offsets_to_cached_pages() + # If reading recommended_pages would read the rest of the index, just + # do so. + if total_pages - len(cached_offsets) <= self._recommended_pages: + # Read whatever is left + if cached_offsets: + expanded = [x for x in xrange(total_pages) + if x not in cached_offsets] + else: + expanded = range(total_pages) + if 'index' in debug.debug_flags: + trace.mutter(' reading all unread pages: %s', expanded) + return expanded + + if self._root_node is None: + # ATM on the first read of the root node of a large index, we don't + # bother pre-reading any other pages. This is because the + # likelyhood of actually reading interesting pages is very low. + # See doc/developers/btree_index_prefetch.txt for a discussion, and + # a possible implementation when we are guessing that the second + # layer index is small + final_offsets = offsets + else: + tree_depth = len(self._row_lengths) + if len(cached_offsets) < tree_depth and len(offsets) == 1: + # We haven't read enough to justify expansion + # If we are only going to read the root node, and 1 leaf node, + # then it isn't worth expanding our request. Once we've read at + # least 2 nodes, then we are probably doing a search, and we + # start expanding our requests. + if 'index' in debug.debug_flags: + trace.mutter(' not expanding on first reads') + return offsets + final_offsets = self._expand_to_neighbors(offsets, cached_offsets, + total_pages) + + final_offsets = sorted(final_offsets) + if 'index' in debug.debug_flags: + trace.mutter('expanded: %s', final_offsets) + return final_offsets + + def _expand_to_neighbors(self, offsets, cached_offsets, total_pages): + """Expand requests to neighbors until we have enough pages. + + This is called from _expand_offsets after policy has determined that we + want to expand. + We only want to expand requests within a given layer. We cheat a little + bit and assume all requests will be in the same layer. This is true + given the current design, but if it changes this algorithm may perform + oddly. + + :param offsets: requested offsets + :param cached_offsets: offsets for pages we currently have cached + :return: A set() of offsets after expansion + """ + final_offsets = set(offsets) + first = end = None + new_tips = set(final_offsets) + while len(final_offsets) < self._recommended_pages and new_tips: + next_tips = set() + for pos in new_tips: + if first is None: + first, end = self._find_layer_first_and_end(pos) + previous = pos - 1 + if (previous > 0 + and previous not in cached_offsets + and previous not in final_offsets + and previous >= first): + next_tips.add(previous) + after = pos + 1 + if (after < total_pages + and after not in cached_offsets + and after not in final_offsets + and after < end): + next_tips.add(after) + # This would keep us from going bigger than + # recommended_pages by only expanding the first offsets. + # However, if we are making a 'wide' request, it is + # reasonable to expand all points equally. + # if len(final_offsets) > recommended_pages: + # break + final_offsets.update(next_tips) + new_tips = next_tips + return final_offsets + + def clear_cache(self): + """Clear out any cached/memoized values. + + This can be called at any time, but generally it is used when we have + extracted some information, but don't expect to be requesting any more + from this index. + """ + # Note that we don't touch self._root_node or self._internal_node_cache + # We don't expect either of those to be big, and it can save + # round-trips in the future. We may re-evaluate this if InternalNode + # memory starts to be an issue. + self._leaf_node_cache.clear() + + def external_references(self, ref_list_num): + if self._root_node is None: + self._get_root_node() + if ref_list_num + 1 > self.node_ref_lists: + raise ValueError('No ref list %d, index has %d ref lists' + % (ref_list_num, self.node_ref_lists)) + keys = set() + refs = set() + for node in self.iter_all_entries(): + keys.add(node[1]) + refs.update(node[3][ref_list_num]) + return refs - keys + + def _find_layer_first_and_end(self, offset): + """Find the start/stop nodes for the layer corresponding to offset. + + :return: (first, end) + first is the first node in this layer + end is the first node of the next layer + """ + first = end = 0 + for roffset in self._row_offsets: + first = end + end = roffset + if offset < roffset: + break + return first, end + + def _get_offsets_to_cached_pages(self): + """Determine what nodes we already have cached.""" + cached_offsets = set(self._internal_node_cache.keys()) + cached_offsets.update(self._leaf_node_cache.keys()) + if self._root_node is not None: + cached_offsets.add(0) + return cached_offsets + + def _get_root_node(self): + if self._root_node is None: + # We may not have a root node yet + self._get_internal_nodes([0]) + return self._root_node + + def _get_nodes(self, cache, node_indexes): + found = {} + needed = [] + for idx in node_indexes: + if idx == 0 and self._root_node is not None: + found[0] = self._root_node + continue + try: + found[idx] = cache[idx] + except KeyError: + needed.append(idx) + if not needed: + return found + needed = self._expand_offsets(needed) + found.update(self._get_and_cache_nodes(needed)) + return found + + def _get_internal_nodes(self, node_indexes): + """Get a node, from cache or disk. + + After getting it, the node will be cached. + """ + return self._get_nodes(self._internal_node_cache, node_indexes) + + def _cache_leaf_values(self, nodes): + """Cache directly from key => value, skipping the btree.""" + if self._leaf_value_cache is not None: + for node in nodes.itervalues(): + for key, value in node.all_items(): + if key in self._leaf_value_cache: + # Don't add the rest of the keys, we've seen this node + # before. + break + self._leaf_value_cache[key] = value + + def _get_leaf_nodes(self, node_indexes): + """Get a bunch of nodes, from cache or disk.""" + found = self._get_nodes(self._leaf_node_cache, node_indexes) + self._cache_leaf_values(found) + return found + + def iter_all_entries(self): + """Iterate over all keys within the index. + + :return: An iterable of (index, key, value) or (index, key, value, reference_lists). + The former tuple is used when there are no reference lists in the + index, making the API compatible with simple key:value index types. + There is no defined order for the result iteration - it will be in + the most efficient order for the index. + """ + if 'evil' in debug.debug_flags: + trace.mutter_callsite(3, + "iter_all_entries scales with size of history.") + if not self.key_count(): + return + if self._row_offsets[-1] == 1: + # There is only the root node, and we read that via key_count() + if self.node_ref_lists: + for key, (value, refs) in self._root_node.all_items(): + yield (self, key, value, refs) + else: + for key, (value, refs) in self._root_node.all_items(): + yield (self, key, value) + return + start_of_leaves = self._row_offsets[-2] + end_of_leaves = self._row_offsets[-1] + needed_offsets = range(start_of_leaves, end_of_leaves) + if needed_offsets == [0]: + # Special case when we only have a root node, as we have already + # read everything + nodes = [(0, self._root_node)] + else: + nodes = self._read_nodes(needed_offsets) + # We iterate strictly in-order so that we can use this function + # for spilling index builds to disk. + if self.node_ref_lists: + for _, node in nodes: + for key, (value, refs) in node.all_items(): + yield (self, key, value, refs) + else: + for _, node in nodes: + for key, (value, refs) in node.all_items(): + yield (self, key, value) + + @staticmethod + def _multi_bisect_right(in_keys, fixed_keys): + """Find the positions where each 'in_key' would fit in fixed_keys. + + This is equivalent to doing "bisect_right" on each in_key into + fixed_keys + + :param in_keys: A sorted list of keys to match with fixed_keys + :param fixed_keys: A sorted list of keys to match against + :return: A list of (integer position, [key list]) tuples. + """ + if not in_keys: + return [] + if not fixed_keys: + # no pointers in the fixed_keys list, which means everything must + # fall to the left. + return [(0, in_keys)] + + # TODO: Iterating both lists will generally take M + N steps + # Bisecting each key will generally take M * log2 N steps. + # If we had an efficient way to compare, we could pick the method + # based on which has the fewer number of steps. + # There is also the argument that bisect_right is a compiled + # function, so there is even more to be gained. + # iter_steps = len(in_keys) + len(fixed_keys) + # bisect_steps = len(in_keys) * math.log(len(fixed_keys), 2) + if len(in_keys) == 1: # Bisect will always be faster for M = 1 + return [(bisect.bisect_right(fixed_keys, in_keys[0]), in_keys)] + # elif bisect_steps < iter_steps: + # offsets = {} + # for key in in_keys: + # offsets.setdefault(bisect_right(fixed_keys, key), + # []).append(key) + # return [(o, offsets[o]) for o in sorted(offsets)] + in_keys_iter = iter(in_keys) + fixed_keys_iter = enumerate(fixed_keys) + cur_in_key = in_keys_iter.next() + cur_fixed_offset, cur_fixed_key = fixed_keys_iter.next() + + class InputDone(Exception): pass + class FixedDone(Exception): pass + + output = [] + cur_out = [] + + # TODO: Another possibility is that rather than iterating on each side, + # we could use a combination of bisecting and iterating. For + # example, while cur_in_key < fixed_key, bisect to find its + # point, then iterate all matching keys, then bisect (restricted + # to only the remainder) for the next one, etc. + try: + while True: + if cur_in_key < cur_fixed_key: + cur_keys = [] + cur_out = (cur_fixed_offset, cur_keys) + output.append(cur_out) + while cur_in_key < cur_fixed_key: + cur_keys.append(cur_in_key) + try: + cur_in_key = in_keys_iter.next() + except StopIteration: + raise InputDone + # At this point cur_in_key must be >= cur_fixed_key + # step the cur_fixed_key until we pass the cur key, or walk off + # the end + while cur_in_key >= cur_fixed_key: + try: + cur_fixed_offset, cur_fixed_key = fixed_keys_iter.next() + except StopIteration: + raise FixedDone + except InputDone: + # We consumed all of the input, nothing more to do + pass + except FixedDone: + # There was some input left, but we consumed all of fixed, so we + # have to add one more for the tail + cur_keys = [cur_in_key] + cur_keys.extend(in_keys_iter) + cur_out = (len(fixed_keys), cur_keys) + output.append(cur_out) + return output + + def _walk_through_internal_nodes(self, keys): + """Take the given set of keys, and find the corresponding LeafNodes. + + :param keys: An unsorted iterable of keys to search for + :return: (nodes, index_and_keys) + nodes is a dict mapping {index: LeafNode} + keys_at_index is a list of tuples of [(index, [keys for Leaf])] + """ + # 6 seconds spent in miss_torture using the sorted() line. + # Even with out of order disk IO it seems faster not to sort it when + # large queries are being made. + keys_at_index = [(0, sorted(keys))] + + for row_pos, next_row_start in enumerate(self._row_offsets[1:-1]): + node_indexes = [idx for idx, s_keys in keys_at_index] + nodes = self._get_internal_nodes(node_indexes) + + next_nodes_and_keys = [] + for node_index, sub_keys in keys_at_index: + node = nodes[node_index] + positions = self._multi_bisect_right(sub_keys, node.keys) + node_offset = next_row_start + node.offset + next_nodes_and_keys.extend([(node_offset + pos, s_keys) + for pos, s_keys in positions]) + keys_at_index = next_nodes_and_keys + # We should now be at the _LeafNodes + node_indexes = [idx for idx, s_keys in keys_at_index] + + # TODO: We may *not* want to always read all the nodes in one + # big go. Consider setting a max size on this. + nodes = self._get_leaf_nodes(node_indexes) + return nodes, keys_at_index + + def iter_entries(self, keys): + """Iterate over keys within the index. + + :param keys: An iterable providing the keys to be retrieved. + :return: An iterable as per iter_all_entries, but restricted to the + keys supplied. No additional keys will be returned, and every + key supplied that is in the index will be returned. + """ + # 6 seconds spent in miss_torture using the sorted() line. + # Even with out of order disk IO it seems faster not to sort it when + # large queries are being made. + # However, now that we are doing multi-way bisecting, we need the keys + # in sorted order anyway. We could change the multi-way code to not + # require sorted order. (For example, it bisects for the first node, + # does an in-order search until a key comes before the current point, + # which it then bisects for, etc.) + keys = frozenset(keys) + if not keys: + return + + if not self.key_count(): + return + + needed_keys = [] + if self._leaf_value_cache is None: + needed_keys = keys + else: + for key in keys: + value = self._leaf_value_cache.get(key, None) + if value is not None: + # This key is known not to be here, skip it + value, refs = value + if self.node_ref_lists: + yield (self, key, value, refs) + else: + yield (self, key, value) + else: + needed_keys.append(key) + + last_key = None + needed_keys = keys + if not needed_keys: + return + nodes, nodes_and_keys = self._walk_through_internal_nodes(needed_keys) + for node_index, sub_keys in nodes_and_keys: + if not sub_keys: + continue + node = nodes[node_index] + for next_sub_key in sub_keys: + if next_sub_key in node: + value, refs = node[next_sub_key] + if self.node_ref_lists: + yield (self, next_sub_key, value, refs) + else: + yield (self, next_sub_key, value) + + def _find_ancestors(self, keys, ref_list_num, parent_map, missing_keys): + """Find the parent_map information for the set of keys. + + This populates the parent_map dict and missing_keys set based on the + queried keys. It also can fill out an arbitrary number of parents that + it finds while searching for the supplied keys. + + It is unlikely that you want to call this directly. See + "CombinedGraphIndex.find_ancestry()" for a more appropriate API. + + :param keys: A keys whose ancestry we want to return + Every key will either end up in 'parent_map' or 'missing_keys'. + :param ref_list_num: This index in the ref_lists is the parents we + care about. + :param parent_map: {key: parent_keys} for keys that are present in this + index. This may contain more entries than were in 'keys', that are + reachable ancestors of the keys requested. + :param missing_keys: keys which are known to be missing in this index. + This may include parents that were not directly requested, but we + were able to determine that they are not present in this index. + :return: search_keys parents that were found but not queried to know + if they are missing or present. Callers can re-query this index for + those keys, and they will be placed into parent_map or missing_keys + """ + if not self.key_count(): + # We use key_count() to trigger reading the root node and + # determining info about this BTreeGraphIndex + # If we don't have any keys, then everything is missing + missing_keys.update(keys) + return set() + if ref_list_num >= self.node_ref_lists: + raise ValueError('No ref list %d, index has %d ref lists' + % (ref_list_num, self.node_ref_lists)) + + # The main trick we are trying to accomplish is that when we find a + # key listing its parents, we expect that the parent key is also likely + # to sit on the same page. Allowing us to expand parents quickly + # without suffering the full stack of bisecting, etc. + nodes, nodes_and_keys = self._walk_through_internal_nodes(keys) + + # These are parent keys which could not be immediately resolved on the + # page where the child was present. Note that we may already be + # searching for that key, and it may actually be present [or known + # missing] on one of the other pages we are reading. + # TODO: + # We could try searching for them in the immediate previous or next + # page. If they occur "later" we could put them in a pending lookup + # set, and then for each node we read thereafter we could check to + # see if they are present. + # However, we don't know the impact of keeping this list of things + # that I'm going to search for every node I come across from here on + # out. + # It doesn't handle the case when the parent key is missing on a + # page that we *don't* read. So we already have to handle being + # re-entrant for that. + # Since most keys contain a date string, they are more likely to be + # found earlier in the file than later, but we would know that right + # away (key < min_key), and wouldn't keep searching it on every other + # page that we read. + # Mostly, it is an idea, one which should be benchmarked. + parents_not_on_page = set() + + for node_index, sub_keys in nodes_and_keys: + if not sub_keys: + continue + # sub_keys is all of the keys we are looking for that should exist + # on this page, if they aren't here, then they won't be found + node = nodes[node_index] + parents_to_check = set() + for next_sub_key in sub_keys: + if next_sub_key not in node: + # This one is just not present in the index at all + missing_keys.add(next_sub_key) + else: + value, refs = node[next_sub_key] + parent_keys = refs[ref_list_num] + parent_map[next_sub_key] = parent_keys + parents_to_check.update(parent_keys) + # Don't look for things we've already found + parents_to_check = parents_to_check.difference(parent_map) + # this can be used to test the benefit of having the check loop + # inlined. + # parents_not_on_page.update(parents_to_check) + # continue + while parents_to_check: + next_parents_to_check = set() + for key in parents_to_check: + if key in node: + value, refs = node[key] + parent_keys = refs[ref_list_num] + parent_map[key] = parent_keys + next_parents_to_check.update(parent_keys) + else: + # This parent either is genuinely missing, or should be + # found on another page. Perf test whether it is better + # to check if this node should fit on this page or not. + # in the 'everything-in-one-pack' scenario, this *not* + # doing the check is 237ms vs 243ms. + # So slightly better, but I assume the standard 'lots + # of packs' is going to show a reasonable improvement + # from the check, because it avoids 'going around + # again' for everything that is in another index + # parents_not_on_page.add(key) + # Missing for some reason + if key < node.min_key: + # in the case of bzr.dev, 3.4k/5.3k misses are + # 'earlier' misses (65%) + parents_not_on_page.add(key) + elif key > node.max_key: + # This parent key would be present on a different + # LeafNode + parents_not_on_page.add(key) + else: + # assert key != node.min_key and key != node.max_key + # If it was going to be present, it would be on + # *this* page, so mark it missing. + missing_keys.add(key) + parents_to_check = next_parents_to_check.difference(parent_map) + # Might want to do another .difference() from missing_keys + # parents_not_on_page could have been found on a different page, or be + # known to be missing. So cull out everything that has already been + # found. + search_keys = parents_not_on_page.difference( + parent_map).difference(missing_keys) + return search_keys + + def iter_entries_prefix(self, keys): + """Iterate over keys within the index using prefix matching. + + Prefix matching is applied within the tuple of a key, not to within + the bytestring of each key element. e.g. if you have the keys ('foo', + 'bar'), ('foobar', 'gam') and do a prefix search for ('foo', None) then + only the former key is returned. + + WARNING: Note that this method currently causes a full index parse + unconditionally (which is reasonably appropriate as it is a means for + thunking many small indices into one larger one and still supplies + iter_all_entries at the thunk layer). + + :param keys: An iterable providing the key prefixes to be retrieved. + Each key prefix takes the form of a tuple the length of a key, but + with the last N elements 'None' rather than a regular bytestring. + The first element cannot be 'None'. + :return: An iterable as per iter_all_entries, but restricted to the + keys with a matching prefix to those supplied. No additional keys + will be returned, and every match that is in the index will be + returned. + """ + keys = sorted(set(keys)) + if not keys: + return + # Load if needed to check key lengths + if self._key_count is None: + self._get_root_node() + # TODO: only access nodes that can satisfy the prefixes we are looking + # for. For now, to meet API usage (as this function is not used by + # current bzrlib) just suck the entire index and iterate in memory. + nodes = {} + if self.node_ref_lists: + if self._key_length == 1: + for _1, key, value, refs in self.iter_all_entries(): + nodes[key] = value, refs + else: + nodes_by_key = {} + for _1, key, value, refs in self.iter_all_entries(): + key_value = key, value, refs + # For a key of (foo, bar, baz) create + # _nodes_by_key[foo][bar][baz] = key_value + key_dict = nodes_by_key + for subkey in key[:-1]: + key_dict = key_dict.setdefault(subkey, {}) + key_dict[key[-1]] = key_value + else: + if self._key_length == 1: + for _1, key, value in self.iter_all_entries(): + nodes[key] = value + else: + nodes_by_key = {} + for _1, key, value in self.iter_all_entries(): + key_value = key, value + # For a key of (foo, bar, baz) create + # _nodes_by_key[foo][bar][baz] = key_value + key_dict = nodes_by_key + for subkey in key[:-1]: + key_dict = key_dict.setdefault(subkey, {}) + key_dict[key[-1]] = key_value + if self._key_length == 1: + for key in keys: + # sanity check + if key[0] is None: + raise errors.BadIndexKey(key) + if len(key) != self._key_length: + raise errors.BadIndexKey(key) + try: + if self.node_ref_lists: + value, node_refs = nodes[key] + yield self, key, value, node_refs + else: + yield self, key, nodes[key] + except KeyError: + pass + return + for key in keys: + # sanity check + if key[0] is None: + raise errors.BadIndexKey(key) + if len(key) != self._key_length: + raise errors.BadIndexKey(key) + # find what it refers to: + key_dict = nodes_by_key + elements = list(key) + # find the subdict whose contents should be returned. + try: + while len(elements) and elements[0] is not None: + key_dict = key_dict[elements[0]] + elements.pop(0) + except KeyError: + # a non-existant lookup. + continue + if len(elements): + dicts = [key_dict] + while dicts: + key_dict = dicts.pop(-1) + # can't be empty or would not exist + item, value = key_dict.iteritems().next() + if type(value) == dict: + # push keys + dicts.extend(key_dict.itervalues()) + else: + # yield keys + for value in key_dict.itervalues(): + # each value is the key:value:node refs tuple + # ready to yield. + yield (self, ) + value + else: + # the last thing looked up was a terminal element + yield (self, ) + key_dict + + def key_count(self): + """Return an estimate of the number of keys in this index. + + For BTreeGraphIndex the estimate is exact as it is contained in the + header. + """ + if self._key_count is None: + self._get_root_node() + return self._key_count + + def _compute_row_offsets(self): + """Fill out the _row_offsets attribute based on _row_lengths.""" + offsets = [] + row_offset = 0 + for row in self._row_lengths: + offsets.append(row_offset) + row_offset += row + offsets.append(row_offset) + self._row_offsets = offsets + + def _parse_header_from_bytes(self, bytes): + """Parse the header from a region of bytes. + + :param bytes: The data to parse. + :return: An offset, data tuple such as readv yields, for the unparsed + data. (which may be of length 0). + """ + signature = bytes[0:len(self._signature())] + if not signature == self._signature(): + raise errors.BadIndexFormatSignature(self._name, BTreeGraphIndex) + lines = bytes[len(self._signature()):].splitlines() + options_line = lines[0] + if not options_line.startswith(_OPTION_NODE_REFS): + raise errors.BadIndexOptions(self) + try: + self.node_ref_lists = int(options_line[len(_OPTION_NODE_REFS):]) + except ValueError: + raise errors.BadIndexOptions(self) + options_line = lines[1] + if not options_line.startswith(_OPTION_KEY_ELEMENTS): + raise errors.BadIndexOptions(self) + try: + self._key_length = int(options_line[len(_OPTION_KEY_ELEMENTS):]) + except ValueError: + raise errors.BadIndexOptions(self) + options_line = lines[2] + if not options_line.startswith(_OPTION_LEN): + raise errors.BadIndexOptions(self) + try: + self._key_count = int(options_line[len(_OPTION_LEN):]) + except ValueError: + raise errors.BadIndexOptions(self) + options_line = lines[3] + if not options_line.startswith(_OPTION_ROW_LENGTHS): + raise errors.BadIndexOptions(self) + try: + self._row_lengths = map(int, [length for length in + options_line[len(_OPTION_ROW_LENGTHS):].split(',') + if len(length)]) + except ValueError: + raise errors.BadIndexOptions(self) + self._compute_row_offsets() + + # calculate the bytes we have processed + header_end = (len(signature) + sum(map(len, lines[0:4])) + 4) + return header_end, bytes[header_end:] + + def _read_nodes(self, nodes): + """Read some nodes from disk into the LRU cache. + + This performs a readv to get the node data into memory, and parses each + node, then yields it to the caller. The nodes are requested in the + supplied order. If possible doing sort() on the list before requesting + a read may improve performance. + + :param nodes: The nodes to read. 0 - first node, 1 - second node etc. + :return: None + """ + # may be the byte string of the whole file + bytes = None + # list of (offset, length) regions of the file that should, evenually + # be read in to data_ranges, either from 'bytes' or from the transport + ranges = [] + base_offset = self._base_offset + for index in nodes: + offset = (index * _PAGE_SIZE) + size = _PAGE_SIZE + if index == 0: + # Root node - special case + if self._size: + size = min(_PAGE_SIZE, self._size) + else: + # The only case where we don't know the size, is for very + # small indexes. So we read the whole thing + bytes = self._transport.get_bytes(self._name) + num_bytes = len(bytes) + self._size = num_bytes - base_offset + # the whole thing should be parsed out of 'bytes' + ranges = [(start, min(_PAGE_SIZE, num_bytes - start)) + for start in xrange(base_offset, num_bytes, _PAGE_SIZE)] + break + else: + if offset > self._size: + raise AssertionError('tried to read past the end' + ' of the file %s > %s' + % (offset, self._size)) + size = min(size, self._size - offset) + ranges.append((base_offset + offset, size)) + if not ranges: + return + elif bytes is not None: + # already have the whole file + data_ranges = [(start, bytes[start:start+size]) + for start, size in ranges] + elif self._file is None: + data_ranges = self._transport.readv(self._name, ranges) + else: + data_ranges = [] + for offset, size in ranges: + self._file.seek(offset) + data_ranges.append((offset, self._file.read(size))) + for offset, data in data_ranges: + offset -= base_offset + if offset == 0: + # extract the header + offset, data = self._parse_header_from_bytes(data) + if len(data) == 0: + continue + bytes = zlib.decompress(data) + if bytes.startswith(_LEAF_FLAG): + node = self._leaf_factory(bytes, self._key_length, + self.node_ref_lists) + elif bytes.startswith(_INTERNAL_FLAG): + node = _InternalNode(bytes) + else: + raise AssertionError("Unknown node type for %r" % bytes) + yield offset / _PAGE_SIZE, node + + def _signature(self): + """The file signature for this index type.""" + return _BTSIGNATURE + + def validate(self): + """Validate that everything in the index can be accessed.""" + # just read and parse every node. + self._get_root_node() + if len(self._row_lengths) > 1: + start_node = self._row_offsets[1] + else: + # We shouldn't be reading anything anyway + start_node = 1 + node_end = self._row_offsets[-1] + for node in self._read_nodes(range(start_node, node_end)): + pass + + +_gcchk_factory = _LeafNode + +try: + from bzrlib import _btree_serializer_pyx as _btree_serializer + _gcchk_factory = _btree_serializer._parse_into_chk +except ImportError, e: + osutils.failed_to_load_extension(e) + from bzrlib import _btree_serializer_py as _btree_serializer |