nodemap.py
668 lines
| 21.4 KiB
| text/x-python
|
PythonLexer
r44486 | # nodemap.py - nodemap related code and utilities | |||
# | ||||
# Copyright 2019 Pierre-Yves David <pierre-yves.david@octobus.net> | ||||
# Copyright 2019 George Racinet <georges.racinet@octobus.net> | ||||
# | ||||
# This software may be used and distributed according to the terms of the | ||||
# GNU General Public License version 2 or any later version. | ||||
r44788 | ||||
r44793 | import re | |||
r44788 | import struct | |||
Joerg Sonnenberger
|
r46729 | from ..node import hex | ||
r45292 | ||||
r44788 | from .. import ( | |||
error, | ||||
r49542 | requirements, | |||
r44795 | util, | |||
r44788 | ) | |||
r48093 | from . import docket as docket_mod | |||
r44486 | ||||
class NodeMap(dict): | ||||
def __missing__(self, x): | ||||
raise error.RevlogError(b'unknown node: %s' % x) | ||||
r44788 | ||||
r48852 | def test_race_hook_1(): | |||
"""hook point for test | ||||
This let tests to have things happens between the docket reading and the | ||||
data reading""" | ||||
pass | ||||
r49542 | def post_stream_cleanup(repo): | |||
"""The stream clone might needs to remove some file if persisten nodemap | ||||
was dropped while stream cloning | ||||
""" | ||||
if requirements.REVLOGV1_REQUIREMENT not in repo.requirements: | ||||
return | ||||
if requirements.NODEMAP_REQUIREMENT in repo.requirements: | ||||
return | ||||
unfi = repo.unfiltered() | ||||
delete_nodemap(None, unfi, unfi.changelog) | ||||
delete_nodemap(None, repo, unfi.manifestlog._rootstore._revlog) | ||||
r44790 | def persisted_data(revlog): | |||
"""read the nodemap for a revlog from disk""" | ||||
r47935 | if revlog._nodemap_file is None: | |||
r44790 | return None | |||
r47935 | pdata = revlog.opener.tryread(revlog._nodemap_file) | |||
r44792 | if not pdata: | |||
return None | ||||
offset = 0 | ||||
(version,) = S_VERSION.unpack(pdata[offset : offset + S_VERSION.size]) | ||||
if version != ONDISK_VERSION: | ||||
return None | ||||
offset += S_VERSION.size | ||||
r44807 | headers = S_HEADER.unpack(pdata[offset : offset + S_HEADER.size]) | |||
r45002 | uid_size, tip_rev, data_length, data_unused, tip_node_size = headers | |||
r44792 | offset += S_HEADER.size | |||
r44803 | docket = NodeMapDocket(pdata[offset : offset + uid_size]) | |||
r45002 | offset += uid_size | |||
r44807 | docket.tip_rev = tip_rev | |||
r45002 | docket.tip_node = pdata[offset : offset + tip_node_size] | |||
r44808 | docket.data_length = data_length | |||
docket.data_unused = data_unused | ||||
r44792 | ||||
r44803 | filename = _rawdata_filepath(revlog, docket) | |||
r45296 | use_mmap = revlog.opener.options.get(b"persistent-nodemap.mmap") | |||
r48852 | ||||
test_race_hook_1() | ||||
r44843 | try: | |||
with revlog.opener(filename) as fd: | ||||
if use_mmap: | ||||
r47733 | try: | |||
data = util.buffer(util.mmapread(fd, data_length)) | ||||
except ValueError: | ||||
# raised when the read file is too small | ||||
data = b'' | ||||
r44843 | else: | |||
data = fd.read(data_length) | ||||
Manuel Jacob
|
r50201 | except FileNotFoundError: | ||
return None | ||||
r44811 | if len(data) < data_length: | |||
return None | ||||
return docket, data | ||||
r44790 | ||||
r44789 | def setup_persistent_nodemap(tr, revlog): | |||
"""Install whatever is needed transaction side to persist a nodemap on disk | ||||
(only actually persist the nodemap if this is relevant for this revlog) | ||||
""" | ||||
r44791 | if revlog._inline: | |||
return # inlined revlog are too small for this to be relevant | ||||
r47935 | if revlog._nodemap_file is None: | |||
r44789 | return # we do not use persistent_nodemap on this revlog | |||
r45004 | ||||
# we need to happen after the changelog finalization, in that use "cl-" | ||||
r47935 | callback_id = b"nm-revlog-persistent-nodemap-%s" % revlog._nodemap_file | |||
r44789 | if tr.hasfinalize(callback_id): | |||
return # no need to register again | ||||
r45003 | tr.addpending( | |||
Pulkit Goyal
|
r47195 | callback_id, lambda tr: persist_nodemap(tr, revlog, pending=True) | ||
r45003 | ) | |||
Pulkit Goyal
|
r47195 | tr.addfinalize(callback_id, lambda tr: persist_nodemap(tr, revlog)) | ||
r44986 | ||||
Gregory Szorc
|
r49801 | class _NoTransaction: | ||
Augie Fackler
|
r46554 | """transaction like object to update the nodemap outside a transaction""" | ||
r44986 | ||||
def __init__(self): | ||||
self._postclose = {} | ||||
def addpostclose(self, callback_id, callback_func): | ||||
self._postclose[callback_id] = callback_func | ||||
r44789 | ||||
r45124 | def registertmp(self, *args, **kwargs): | |||
pass | ||||
def addbackup(self, *args, **kwargs): | ||||
pass | ||||
def add(self, *args, **kwargs): | ||||
pass | ||||
r45125 | def addabort(self, *args, **kwargs): | |||
pass | ||||
r45292 | def _report(self, *args): | |||
pass | ||||
r44789 | ||||
r44932 | def update_persistent_nodemap(revlog): | |||
"""update the persistent nodemap right now | ||||
To be used for updating the nodemap on disk outside of a normal transaction | ||||
setup (eg, `debugupdatecache`). | ||||
""" | ||||
r45247 | if revlog._inline: | |||
return # inlined revlog are too small for this to be relevant | ||||
r47935 | if revlog._nodemap_file is None: | |||
r45247 | return # we do not use persistent_nodemap on this revlog | |||
r44986 | notr = _NoTransaction() | |||
Pulkit Goyal
|
r47195 | persist_nodemap(notr, revlog) | ||
r44986 | for k in sorted(notr._postclose): | |||
notr._postclose[k](None) | ||||
r44932 | ||||
Pulkit Goyal
|
r47276 | def delete_nodemap(tr, repo, revlog): | ||
r48244 | """Delete nodemap data on disk for a given revlog""" | |||
r49539 | prefix = revlog.radix | |||
pattern = re.compile(br"(^|/)%s(-[0-9a-f]+\.nd|\.n(\.a)?)$" % prefix) | ||||
dirpath = revlog.opener.dirname(revlog._indexfile) | ||||
for f in revlog.opener.listdir(dirpath): | ||||
if pattern.match(f): | ||||
repo.svfs.tryunlink(f) | ||||
Pulkit Goyal
|
r47276 | |||
Pulkit Goyal
|
r47199 | def persist_nodemap(tr, revlog, pending=False, force=False): | ||
Augie Fackler
|
r46554 | """Write nodemap data on disk for a given revlog""" | ||
r44789 | if getattr(revlog, 'filteredrevs', ()): | |||
raise error.ProgrammingError( | ||||
"cannot persist nodemap of a filtered changelog" | ||||
) | ||||
r47935 | if revlog._nodemap_file is None: | |||
Pulkit Goyal
|
r47199 | if force: | ||
r47935 | revlog._nodemap_file = get_nodemap_file(revlog) | |||
Pulkit Goyal
|
r47199 | else: | ||
msg = "calling persist nodemap on a revlog without the feature enabled" | ||||
raise error.ProgrammingError(msg) | ||||
r44805 | ||||
can_incremental = util.safehasattr(revlog.index, "nodemap_data_incremental") | ||||
ondisk_docket = revlog._nodemap_docket | ||||
r44843 | feed_data = util.safehasattr(revlog.index, "update_nodemap_data") | |||
r45296 | use_mmap = revlog.opener.options.get(b"persistent-nodemap.mmap") | |||
r44805 | ||||
r44809 | data = None | |||
r44805 | # first attemp an incremental update of the data | |||
if can_incremental and ondisk_docket is not None: | ||||
target_docket = revlog._nodemap_docket.copy() | ||||
r44809 | ( | |||
src_docket, | ||||
data_changed_count, | ||||
data, | ||||
) = revlog.index.nodemap_data_incremental() | ||||
r45126 | new_length = target_docket.data_length + len(data) | |||
new_unused = target_docket.data_unused + data_changed_count | ||||
r44809 | if src_docket != target_docket: | |||
data = None | ||||
r45126 | elif new_length <= (new_unused * 10): # under 10% of unused data | |||
data = None | ||||
r44809 | else: | |||
datafile = _rawdata_filepath(revlog, target_docket) | ||||
# EXP-TODO: if this is a cache, this should use a cache vfs, not a | ||||
# store vfs | ||||
r45124 | tr.add(datafile, target_docket.data_length) | |||
r44810 | with revlog.opener(datafile, b'r+') as fd: | |||
fd.seek(target_docket.data_length) | ||||
r44809 | fd.write(data) | |||
r44843 | if feed_data: | |||
if use_mmap: | ||||
fd.seek(0) | ||||
new_data = fd.read(new_length) | ||||
else: | ||||
fd.flush() | ||||
new_data = util.buffer(util.mmapread(fd, new_length)) | ||||
target_docket.data_length = new_length | ||||
r45126 | target_docket.data_unused = new_unused | |||
r44809 | ||||
if data is None: | ||||
r44805 | # otherwise fallback to a full new export | |||
target_docket = NodeMapDocket() | ||||
datafile = _rawdata_filepath(revlog, target_docket) | ||||
if util.safehasattr(revlog.index, "nodemap_data_all"): | ||||
data = revlog.index.nodemap_data_all() | ||||
else: | ||||
data = persistent_data(revlog.index) | ||||
# EXP-TODO: if this is a cache, this should use a cache vfs, not a | ||||
# store vfs | ||||
r45125 | ||||
tryunlink = revlog.opener.tryunlink | ||||
def abortck(tr): | ||||
tryunlink(datafile) | ||||
callback_id = b"delete-%s" % datafile | ||||
# some flavor of the transaction abort does not cleanup new file, it | ||||
# simply empty them. | ||||
tr.addabort(callback_id, abortck) | ||||
r44843 | with revlog.opener(datafile, b'w+') as fd: | |||
r44805 | fd.write(data) | |||
r44843 | if feed_data: | |||
if use_mmap: | ||||
new_data = data | ||||
else: | ||||
fd.flush() | ||||
new_data = util.buffer(util.mmapread(fd, len(data))) | ||||
r44808 | target_docket.data_length = len(data) | |||
r44807 | target_docket.tip_rev = revlog.tiprev() | |||
r45002 | target_docket.tip_node = revlog.node(target_docket.tip_rev) | |||
r44805 | # EXP-TODO: if this is a cache, this should use a cache vfs, not a | |||
# store vfs | ||||
r47935 | file_path = revlog._nodemap_file | |||
r45003 | if pending: | |||
file_path += b'.a' | ||||
r45124 | tr.registertmp(file_path) | |||
else: | ||||
tr.addbackup(file_path) | ||||
r45003 | with revlog.opener(file_path, b'w', atomictemp=True) as fp: | |||
r44805 | fp.write(target_docket.serialize()) | |||
revlog._nodemap_docket = target_docket | ||||
r44843 | if feed_data: | |||
r44812 | revlog.index.update_nodemap_data(target_docket, new_data) | |||
r44805 | # search for old index file in all cases, some older process might have | |||
# left one behind. | ||||
r44803 | olds = _other_rawdata_filepath(revlog, target_docket) | |||
r44793 | if olds: | |||
realvfs = getattr(revlog, '_realopener', revlog.opener) | ||||
def cleanup(tr): | ||||
for oldfile in olds: | ||||
realvfs.tryunlink(oldfile) | ||||
r47935 | callback_id = b"revlog-cleanup-nodemap-%s" % revlog._nodemap_file | |||
r44986 | tr.addpostclose(callback_id, cleanup) | |||
r44792 | ||||
### Nodemap docket file | ||||
# | ||||
# The nodemap data are stored on disk using 2 files: | ||||
# | ||||
# * a raw data files containing a persistent nodemap | ||||
# (see `Nodemap Trie` section) | ||||
# | ||||
# * a small "docket" file containing medatadata | ||||
# | ||||
# While the nodemap data can be multiple tens of megabytes, the "docket" is | ||||
# small, it is easy to update it automatically or to duplicated its content | ||||
# during a transaction. | ||||
# | ||||
# Multiple raw data can exist at the same time (The currently valid one and a | ||||
# new one beind used by an in progress transaction). To accomodate this, the | ||||
# filename hosting the raw data has a variable parts. The exact filename is | ||||
# specified inside the "docket" file. | ||||
# | ||||
# The docket file contains information to find, qualify and validate the raw | ||||
# data. Its content is currently very light, but it will expand as the on disk | ||||
# nodemap gains the necessary features to be used in production. | ||||
r45294 | ONDISK_VERSION = 1 | |||
r44792 | S_VERSION = struct.Struct(">B") | |||
r45002 | S_HEADER = struct.Struct(">BQQQQ") | |||
r44792 | ||||
r48091 | ||||
Gregory Szorc
|
r49801 | class NodeMapDocket: | ||
r44803 | """metadata associated with persistent nodemap data | |||
The persistent data may come from disk or be on their way to disk. | ||||
""" | ||||
def __init__(self, uid=None): | ||||
if uid is None: | ||||
r48093 | uid = docket_mod.make_uid() | |||
r44983 | # a unique identifier for the data file: | |||
# - When new data are appended, it is preserved. | ||||
# - When a new data file is created, a new identifier is generated. | ||||
r44803 | self.uid = uid | |||
r44983 | # the tipmost revision stored in the data file. This revision and all | |||
# revision before it are expected to be encoded in the data file. | ||||
r44807 | self.tip_rev = None | |||
r45002 | # the node of that tipmost revision, if it mismatch the current index | |||
# data the docket is not valid for the current index and should be | ||||
# discarded. | ||||
# | ||||
# note: this method is not perfect as some destructive operation could | ||||
# preserve the same tip_rev + tip_node while altering lower revision. | ||||
# However this multiple other caches have the same vulnerability (eg: | ||||
# brancmap cache). | ||||
self.tip_node = None | ||||
r44983 | # the size (in bytes) of the persisted data to encode the nodemap valid | |||
# for `tip_rev`. | ||||
# - data file shorter than this are corrupted, | ||||
# - any extra data should be ignored. | ||||
r44808 | self.data_length = None | |||
r44983 | # the amount (in bytes) of "dead" data, still in the data file but no | |||
# longer used for the nodemap. | ||||
r44808 | self.data_unused = 0 | |||
r44803 | ||||
def copy(self): | ||||
r44807 | new = NodeMapDocket(uid=self.uid) | |||
new.tip_rev = self.tip_rev | ||||
r45002 | new.tip_node = self.tip_node | |||
r44808 | new.data_length = self.data_length | |||
new.data_unused = self.data_unused | ||||
r44807 | return new | |||
r44803 | ||||
r44809 | def __cmp__(self, other): | |||
if self.uid < other.uid: | ||||
return -1 | ||||
if self.uid > other.uid: | ||||
return 1 | ||||
elif self.data_length < other.data_length: | ||||
return -1 | ||||
elif self.data_length > other.data_length: | ||||
return 1 | ||||
return 0 | ||||
def __eq__(self, other): | ||||
return self.uid == other.uid and self.data_length == other.data_length | ||||
r44803 | def serialize(self): | |||
"""return serialized bytes for a docket using the passed uid""" | ||||
data = [] | ||||
data.append(S_VERSION.pack(ONDISK_VERSION)) | ||||
r44808 | headers = ( | |||
len(self.uid), | ||||
self.tip_rev, | ||||
self.data_length, | ||||
self.data_unused, | ||||
r45002 | len(self.tip_node), | |||
r44808 | ) | |||
r44807 | data.append(S_HEADER.pack(*headers)) | |||
r44803 | data.append(self.uid) | |||
r45002 | data.append(self.tip_node) | |||
r44803 | return b''.join(data) | |||
r44792 | ||||
r44803 | def _rawdata_filepath(revlog, docket): | |||
r44792 | """The (vfs relative) nodemap's rawdata file for a given uid""" | |||
r47923 | prefix = revlog.radix | |||
r44803 | return b"%s-%s.nd" % (prefix, docket.uid) | |||
r44789 | ||||
r44803 | def _other_rawdata_filepath(revlog, docket): | |||
r47923 | prefix = revlog.radix | |||
Augie Fackler
|
r44952 | pattern = re.compile(br"(^|/)%s-[0-9a-f]+\.nd$" % prefix) | ||
r44803 | new_file_path = _rawdata_filepath(revlog, docket) | |||
r44793 | new_file_name = revlog.opener.basename(new_file_path) | |||
dirpath = revlog.opener.dirname(new_file_path) | ||||
others = [] | ||||
for f in revlog.opener.listdir(dirpath): | ||||
if pattern.match(f) and f != new_file_name: | ||||
others.append(f) | ||||
return others | ||||
r44788 | ### Nodemap Trie | |||
# | ||||
# This is a simple reference implementation to compute and persist a nodemap | ||||
# trie. This reference implementation is write only. The python version of this | ||||
# is not expected to be actually used, since it wont provide performance | ||||
# improvement over existing non-persistent C implementation. | ||||
# | ||||
# The nodemap is persisted as Trie using 4bits-address/16-entries block. each | ||||
# revision can be adressed using its node shortest prefix. | ||||
# | ||||
# The trie is stored as a sequence of block. Each block contains 16 entries | ||||
# (signed 64bit integer, big endian). Each entry can be one of the following: | ||||
# | ||||
# * value >= 0 -> index of sub-block | ||||
# * value == -1 -> no value | ||||
Joerg Sonnenberger
|
r46815 | # * value < -1 -> encoded revision: rev = -(value+2) | ||
# | ||||
# See REV_OFFSET and _transform_rev below. | ||||
r44788 | # | |||
# The implementation focus on simplicity, not on performance. A Rust | ||||
# implementation should provide a efficient version of the same binary | ||||
# persistence. This reference python implementation is never meant to be | ||||
# extensively use in production. | ||||
def persistent_data(index): | ||||
Augie Fackler
|
r46554 | """return the persistent binary form for a nodemap for a given index""" | ||
r44788 | trie = _build_trie(index) | |||
return _persist_trie(trie) | ||||
r44805 | def update_persistent_data(index, root, max_idx, last_rev): | |||
Augie Fackler
|
r46554 | """return the incremental update for persistent nodemap from a given index""" | ||
r44808 | changed_block, trie = _update_trie(index, root, last_rev) | |||
return ( | ||||
changed_block * S_BLOCK.size, | ||||
_persist_trie(trie, existing_idx=max_idx), | ||||
) | ||||
r44805 | ||||
r44788 | S_BLOCK = struct.Struct(">" + ("l" * 16)) | |||
NO_ENTRY = -1 | ||||
# rev 0 need to be -2 because 0 is used by block, -1 is a special value. | ||||
REV_OFFSET = 2 | ||||
def _transform_rev(rev): | ||||
"""Return the number used to represent the rev in the tree. | ||||
(or retrieve a rev number from such representation) | ||||
Note that this is an involution, a function equal to its inverse (i.e. | ||||
which gives the identity when applied to itself). | ||||
""" | ||||
return -(rev + REV_OFFSET) | ||||
def _to_int(hex_digit): | ||||
"""turn an hexadecimal digit into a proper integer""" | ||||
return int(hex_digit, 16) | ||||
r44796 | class Block(dict): | |||
"""represent a block of the Trie | ||||
contains up to 16 entry indexed from 0 to 15""" | ||||
r44802 | def __init__(self): | |||
super(Block, self).__init__() | ||||
# If this block exist on disk, here is its ID | ||||
self.ondisk_id = None | ||||
r44797 | def __iter__(self): | |||
return iter(self.get(i) for i in range(16)) | ||||
r44796 | ||||
r44788 | def _build_trie(index): | |||
"""build a nodemap trie | ||||
The nodemap stores revision number for each unique prefix. | ||||
Each block is a dictionary with keys in `[0, 15]`. Values are either | ||||
another block or a revision number. | ||||
""" | ||||
r44796 | root = Block() | |||
r44788 | for rev in range(len(index)): | |||
Joerg Sonnenberger
|
r46729 | current_hex = hex(index[rev][7]) | ||
_insert_into_block(index, 0, root, rev, current_hex) | ||||
r44788 | return root | |||
r44805 | def _update_trie(index, root, last_rev): | |||
"""consume""" | ||||
r44808 | changed = 0 | |||
r44805 | for rev in range(last_rev + 1, len(index)): | |||
Joerg Sonnenberger
|
r46729 | current_hex = hex(index[rev][7]) | ||
changed += _insert_into_block(index, 0, root, rev, current_hex) | ||||
r44808 | return changed, root | |||
r44805 | ||||
r44788 | def _insert_into_block(index, level, block, current_rev, current_hex): | |||
"""insert a new revision in a block | ||||
index: the index we are adding revision for | ||||
level: the depth of the current block in the trie | ||||
block: the block currently being considered | ||||
current_rev: the revision number we are adding | ||||
current_hex: the hexadecimal representation of the of that revision | ||||
""" | ||||
r44808 | changed = 1 | |||
r44805 | if block.ondisk_id is not None: | |||
block.ondisk_id = None | ||||
r44788 | hex_digit = _to_int(current_hex[level : level + 1]) | |||
entry = block.get(hex_digit) | ||||
if entry is None: | ||||
# no entry, simply store the revision number | ||||
block[hex_digit] = current_rev | ||||
elif isinstance(entry, dict): | ||||
# need to recurse to an underlying block | ||||
r44808 | changed += _insert_into_block( | |||
index, level + 1, entry, current_rev, current_hex | ||||
) | ||||
r44788 | else: | |||
# collision with a previously unique prefix, inserting new | ||||
# vertices to fit both entry. | ||||
Joerg Sonnenberger
|
r46729 | other_hex = hex(index[entry][7]) | ||
r44788 | other_rev = entry | |||
r44796 | new = Block() | |||
r44788 | block[hex_digit] = new | |||
_insert_into_block(index, level + 1, new, other_rev, other_hex) | ||||
_insert_into_block(index, level + 1, new, current_rev, current_hex) | ||||
r44808 | return changed | |||
r44788 | ||||
r44805 | def _persist_trie(root, existing_idx=None): | |||
r44788 | """turn a nodemap trie into persistent binary data | |||
See `_build_trie` for nodemap trie structure""" | ||||
block_map = {} | ||||
r44805 | if existing_idx is not None: | |||
base_idx = existing_idx + 1 | ||||
else: | ||||
base_idx = 0 | ||||
r44788 | chunks = [] | |||
for tn in _walk_trie(root): | ||||
r44805 | if tn.ondisk_id is not None: | |||
block_map[id(tn)] = tn.ondisk_id | ||||
else: | ||||
block_map[id(tn)] = len(chunks) + base_idx | ||||
chunks.append(_persist_block(tn, block_map)) | ||||
r44788 | return b''.join(chunks) | |||
def _walk_trie(block): | ||||
"""yield all the block in a trie | ||||
Children blocks are always yield before their parent block. | ||||
""" | ||||
r45186 | for (__, item) in sorted(block.items()): | |||
r44788 | if isinstance(item, dict): | |||
for sub_block in _walk_trie(item): | ||||
yield sub_block | ||||
yield block | ||||
def _persist_block(block_node, block_map): | ||||
"""produce persistent binary data for a single block | ||||
Children block are assumed to be already persisted and present in | ||||
block_map. | ||||
""" | ||||
r44797 | data = tuple(_to_value(v, block_map) for v in block_node) | |||
r44788 | return S_BLOCK.pack(*data) | |||
def _to_value(item, block_map): | ||||
"""persist any value as an integer""" | ||||
if item is None: | ||||
return NO_ENTRY | ||||
elif isinstance(item, dict): | ||||
return block_map[id(item)] | ||||
else: | ||||
return _transform_rev(item) | ||||
r44798 | ||||
def parse_data(data): | ||||
"""parse parse nodemap data into a nodemap Trie""" | ||||
if (len(data) % S_BLOCK.size) != 0: | ||||
Matt Harbison
|
r47511 | msg = b"nodemap data size is not a multiple of block size (%d): %d" | ||
r44798 | raise error.Abort(msg % (S_BLOCK.size, len(data))) | |||
if not data: | ||||
r44805 | return Block(), None | |||
r44798 | block_map = {} | |||
new_blocks = [] | ||||
for i in range(0, len(data), S_BLOCK.size): | ||||
block = Block() | ||||
r44802 | block.ondisk_id = len(block_map) | |||
block_map[block.ondisk_id] = block | ||||
r44798 | block_data = data[i : i + S_BLOCK.size] | |||
values = S_BLOCK.unpack(block_data) | ||||
new_blocks.append((block, values)) | ||||
for b, values in new_blocks: | ||||
for idx, v in enumerate(values): | ||||
if v == NO_ENTRY: | ||||
continue | ||||
elif v >= 0: | ||||
b[idx] = block_map[v] | ||||
else: | ||||
b[idx] = _transform_rev(v) | ||||
r44805 | return block, i // S_BLOCK.size | |||
r44799 | ||||
# debug utility | ||||
def check_data(ui, index, data): | ||||
"""verify that the provided nodemap data are valid for the given idex""" | ||||
ret = 0 | ||||
Arseniy Alekseyev
|
r51403 | ui.status((b"revisions in index: %d\n") % len(index)) | ||
r44805 | root, __ = parse_data(data) | |||
r44799 | all_revs = set(_all_revisions(root)) | |||
Arseniy Alekseyev
|
r51403 | ui.status((b"revisions in nodemap: %d\n") % len(all_revs)) | ||
r44799 | for r in range(len(index)): | |||
if r not in all_revs: | ||||
msg = b" revision missing from nodemap: %d\n" % r | ||||
ui.write_err(msg) | ||||
ret = 1 | ||||
else: | ||||
all_revs.remove(r) | ||||
Joerg Sonnenberger
|
r46729 | nm_rev = _find_node(root, hex(index[r][7])) | ||
r44800 | if nm_rev is None: | |||
msg = b" revision node does not match any entries: %d\n" % r | ||||
ui.write_err(msg) | ||||
ret = 1 | ||||
elif nm_rev != r: | ||||
msg = ( | ||||
b" revision node does not match the expected revision: " | ||||
b"%d != %d\n" % (r, nm_rev) | ||||
) | ||||
ui.write_err(msg) | ||||
ret = 1 | ||||
r44799 | if all_revs: | |||
for r in sorted(all_revs): | ||||
Arseniy Alekseyev
|
r51403 | msg = b" extra revisions in nodemap: %d\n" % r | ||
r44799 | ui.write_err(msg) | |||
ret = 1 | ||||
return ret | ||||
def _all_revisions(root): | ||||
"""return all revisions stored in a Trie""" | ||||
for block in _walk_trie(root): | ||||
for v in block: | ||||
if v is None or isinstance(v, Block): | ||||
continue | ||||
yield v | ||||
r44800 | ||||
def _find_node(block, node): | ||||
"""find the revision associated with a given node""" | ||||
entry = block.get(_to_int(node[0:1])) | ||||
if isinstance(entry, dict): | ||||
return _find_node(entry, node[1:]) | ||||
return entry | ||||
Pulkit Goyal
|
r47196 | |||
r47922 | def get_nodemap_file(revlog): | |||
r48014 | if revlog._trypending: | |||
r47922 | pending_path = revlog.radix + b".n.a" | |||
if revlog.opener.exists(pending_path): | ||||
Pulkit Goyal
|
r47196 | return pending_path | ||
r47922 | return revlog.radix + b".n" | |||