|
|
# manifest.py - manifest revision class for mercurial
|
|
|
#
|
|
|
# Copyright 2005-2007 Olivia Mackall <olivia@selenic.com>
|
|
|
#
|
|
|
# This software may be used and distributed according to the terms of the
|
|
|
# GNU General Public License version 2 or any later version.
|
|
|
|
|
|
from __future__ import annotations
|
|
|
|
|
|
import heapq
|
|
|
import itertools
|
|
|
import struct
|
|
|
import typing
|
|
|
import weakref
|
|
|
|
|
|
from typing import (
|
|
|
ByteString,
|
|
|
Callable,
|
|
|
Collection,
|
|
|
Dict,
|
|
|
Iterable,
|
|
|
Iterator,
|
|
|
List,
|
|
|
Optional,
|
|
|
Set,
|
|
|
Tuple,
|
|
|
Union,
|
|
|
cast,
|
|
|
)
|
|
|
|
|
|
from .i18n import _
|
|
|
from .node import (
|
|
|
bin,
|
|
|
hex,
|
|
|
nullrev,
|
|
|
)
|
|
|
from . import (
|
|
|
encoding,
|
|
|
error,
|
|
|
match as matchmod,
|
|
|
mdiff,
|
|
|
pathutil,
|
|
|
policy,
|
|
|
pycompat,
|
|
|
revlog,
|
|
|
util,
|
|
|
)
|
|
|
from .interfaces import (
|
|
|
repository,
|
|
|
util as interfaceutil,
|
|
|
)
|
|
|
from .revlogutils import (
|
|
|
constants as revlog_constants,
|
|
|
)
|
|
|
|
|
|
parsers = policy.importmod('parsers')
|
|
|
propertycache = util.propertycache
|
|
|
|
|
|
# Allow tests to more easily test the alternate path in manifestdict.fastdelta()
|
|
|
FASTDELTA_TEXTDIFF_THRESHOLD = 1000
|
|
|
|
|
|
|
|
|
def _parse(nodelen, data: bytes):
|
|
|
# This method does a little bit of excessive-looking
|
|
|
# precondition checking. This is so that the behavior of this
|
|
|
# class exactly matches its C counterpart to try and help
|
|
|
# prevent surprise breakage for anyone that develops against
|
|
|
# the pure version.
|
|
|
if data and data[-1:] != b'\n':
|
|
|
raise ValueError(b'Manifest did not end in a newline.')
|
|
|
prev = None
|
|
|
for l in data.splitlines():
|
|
|
if prev is not None and prev > l:
|
|
|
raise ValueError(b'Manifest lines not in sorted order.')
|
|
|
prev = l
|
|
|
f, n = l.split(b'\0')
|
|
|
nl = len(n)
|
|
|
flags = n[-1:]
|
|
|
if flags in _manifestflags:
|
|
|
n = n[:-1]
|
|
|
nl -= 1
|
|
|
else:
|
|
|
flags = b''
|
|
|
if nl != 2 * nodelen:
|
|
|
raise ValueError(b'Invalid manifest line')
|
|
|
|
|
|
yield f, bin(n), flags
|
|
|
|
|
|
|
|
|
def _text(it):
|
|
|
files = []
|
|
|
lines = []
|
|
|
for f, n, fl in it:
|
|
|
files.append(f)
|
|
|
# if this is changed to support newlines in filenames,
|
|
|
# be sure to check the templates/ dir again (especially *-raw.tmpl)
|
|
|
lines.append(b"%s\0%s%s\n" % (f, hex(n), fl))
|
|
|
|
|
|
_checkforbidden(files)
|
|
|
return b''.join(lines)
|
|
|
|
|
|
|
|
|
class lazymanifestiter:
|
|
|
def __init__(self, lm: '_LazyManifest') -> None:
|
|
|
self.pos = 0
|
|
|
self.lm = lm
|
|
|
|
|
|
def __iter__(self) -> 'lazymanifestiter':
|
|
|
return self
|
|
|
|
|
|
def next(self) -> bytes:
|
|
|
try:
|
|
|
data, pos = self.lm._get(self.pos)
|
|
|
except IndexError:
|
|
|
raise StopIteration
|
|
|
if pos == -1:
|
|
|
assert isinstance(data, tuple)
|
|
|
self.pos += 1
|
|
|
return data[0]
|
|
|
assert isinstance(data, bytes)
|
|
|
self.pos += 1
|
|
|
zeropos = data.find(b'\x00', pos)
|
|
|
return data[pos:zeropos]
|
|
|
|
|
|
__next__ = next
|
|
|
|
|
|
|
|
|
class lazymanifestiterentries:
|
|
|
def __init__(self, lm: '_LazyManifest') -> None:
|
|
|
self.lm = lm
|
|
|
self.pos = 0
|
|
|
|
|
|
def __iter__(self) -> 'lazymanifestiterentries':
|
|
|
return self
|
|
|
|
|
|
def next(self) -> Tuple[bytes, bytes, bytes]:
|
|
|
try:
|
|
|
data, pos = self.lm._get(self.pos)
|
|
|
except IndexError:
|
|
|
raise StopIteration
|
|
|
if pos == -1:
|
|
|
assert isinstance(data, tuple)
|
|
|
self.pos += 1
|
|
|
return data
|
|
|
assert isinstance(data, bytes)
|
|
|
zeropos = data.find(b'\x00', pos)
|
|
|
nlpos = data.find(b'\n', pos)
|
|
|
if zeropos == -1 or nlpos == -1 or nlpos < zeropos:
|
|
|
raise error.StorageError(b'Invalid manifest line')
|
|
|
flags = data[nlpos - 1 : nlpos]
|
|
|
if flags in _manifestflags:
|
|
|
hlen = nlpos - zeropos - 2
|
|
|
else:
|
|
|
hlen = nlpos - zeropos - 1
|
|
|
flags = b''
|
|
|
if hlen != 2 * self.lm._nodelen:
|
|
|
raise error.StorageError(b'Invalid manifest line')
|
|
|
hashval = unhexlify(
|
|
|
data, self.lm.extrainfo[self.pos], zeropos + 1, hlen
|
|
|
)
|
|
|
self.pos += 1
|
|
|
return (data[pos:zeropos], hashval, flags)
|
|
|
|
|
|
__next__ = next
|
|
|
|
|
|
|
|
|
def unhexlify(data: bytes, extra: int, pos, length: int):
|
|
|
s = bin(data[pos : pos + length])
|
|
|
if extra:
|
|
|
s += bytes([extra & 0xFF])
|
|
|
return s
|
|
|
|
|
|
|
|
|
def _cmp(a, b):
|
|
|
return (a > b) - (a < b)
|
|
|
|
|
|
|
|
|
_manifestflags = {b'', b'l', b't', b'x'}
|
|
|
|
|
|
|
|
|
class _LazyManifest:
|
|
|
"""A pure python manifest backed by a byte string. It is supplimented with
|
|
|
internal lists as it is modified, until it is compacted back to a pure byte
|
|
|
string.
|
|
|
|
|
|
``data`` is the initial manifest data.
|
|
|
|
|
|
``positions`` is a list of offsets, one per manifest entry. Positive
|
|
|
values are offsets into ``data``, negative values are offsets into the
|
|
|
``extradata`` list. When an entry is removed, its entry is dropped from
|
|
|
``positions``. The values are encoded such that when walking the list and
|
|
|
indexing into ``data`` or ``extradata`` as appropriate, the entries are
|
|
|
sorted by filename.
|
|
|
|
|
|
``extradata`` is a list of (key, hash, flags) for entries that were added or
|
|
|
modified since the manifest was created or compacted.
|
|
|
"""
|
|
|
|
|
|
def __init__(
|
|
|
self,
|
|
|
nodelen: int,
|
|
|
data: bytes,
|
|
|
positions=None,
|
|
|
extrainfo=None,
|
|
|
extradata=None,
|
|
|
hasremovals: bool = False,
|
|
|
):
|
|
|
self._nodelen = nodelen
|
|
|
if positions is None:
|
|
|
self.positions = self.findlines(data)
|
|
|
self.extrainfo = [0] * len(self.positions)
|
|
|
self.data = data
|
|
|
self.extradata = []
|
|
|
self.hasremovals = False
|
|
|
else:
|
|
|
self.positions = positions[:]
|
|
|
self.extrainfo = extrainfo[:]
|
|
|
self.extradata = extradata[:]
|
|
|
self.data = data
|
|
|
self.hasremovals = hasremovals
|
|
|
|
|
|
def findlines(self, data: bytes) -> List[int]:
|
|
|
if not data:
|
|
|
return []
|
|
|
pos = data.find(b"\n")
|
|
|
if pos == -1 or data[-1:] != b'\n':
|
|
|
raise ValueError(b"Manifest did not end in a newline.")
|
|
|
positions = [0]
|
|
|
prev = data[: data.find(b'\x00')]
|
|
|
while pos < len(data) - 1 and pos != -1:
|
|
|
positions.append(pos + 1)
|
|
|
nexts = data[pos + 1 : data.find(b'\x00', pos + 1)]
|
|
|
if nexts < prev:
|
|
|
raise ValueError(b"Manifest lines not in sorted order.")
|
|
|
prev = nexts
|
|
|
pos = data.find(b"\n", pos + 1)
|
|
|
return positions
|
|
|
|
|
|
def _get(
|
|
|
self, index: int
|
|
|
) -> Tuple[Union[bytes, Tuple[bytes, bytes, bytes]], int]:
|
|
|
# get the position encoded in pos:
|
|
|
# positive number is an index in 'data'
|
|
|
# negative number is in extrapieces
|
|
|
pos = self.positions[index]
|
|
|
if pos >= 0:
|
|
|
return self.data, pos
|
|
|
return self.extradata[-pos - 1], -1
|
|
|
|
|
|
def _getkey(self, pos) -> bytes:
|
|
|
if pos >= 0:
|
|
|
return self.data[pos : self.data.find(b'\x00', pos + 1)]
|
|
|
return self.extradata[-pos - 1][0]
|
|
|
|
|
|
def bsearch(self, key: bytes) -> int:
|
|
|
first = 0
|
|
|
last = len(self.positions) - 1
|
|
|
|
|
|
while first <= last:
|
|
|
midpoint = (first + last) // 2
|
|
|
nextpos = self.positions[midpoint]
|
|
|
candidate = self._getkey(nextpos)
|
|
|
r = _cmp(key, candidate)
|
|
|
if r == 0:
|
|
|
return midpoint
|
|
|
else:
|
|
|
if r < 0:
|
|
|
last = midpoint - 1
|
|
|
else:
|
|
|
first = midpoint + 1
|
|
|
return -1
|
|
|
|
|
|
def bsearch2(self, key: bytes) -> Tuple[int, bool]:
|
|
|
# same as the above, but will always return the position
|
|
|
# done for performance reasons
|
|
|
first = 0
|
|
|
last = len(self.positions) - 1
|
|
|
|
|
|
while first <= last:
|
|
|
midpoint = (first + last) // 2
|
|
|
nextpos = self.positions[midpoint]
|
|
|
candidate = self._getkey(nextpos)
|
|
|
r = _cmp(key, candidate)
|
|
|
if r == 0:
|
|
|
return (midpoint, True)
|
|
|
else:
|
|
|
if r < 0:
|
|
|
last = midpoint - 1
|
|
|
else:
|
|
|
first = midpoint + 1
|
|
|
return (first, False)
|
|
|
|
|
|
def __contains__(self, key: bytes) -> bool:
|
|
|
return self.bsearch(key) != -1
|
|
|
|
|
|
def __getitem__(self, key: bytes) -> Tuple[bytes, bytes]:
|
|
|
if not isinstance(key, bytes):
|
|
|
raise TypeError(b"getitem: manifest keys must be a bytes.")
|
|
|
needle = self.bsearch(key)
|
|
|
if needle == -1:
|
|
|
raise KeyError
|
|
|
data, pos = self._get(needle)
|
|
|
if pos == -1:
|
|
|
assert isinstance(data, tuple)
|
|
|
return (data[1], data[2])
|
|
|
|
|
|
assert isinstance(data, bytes)
|
|
|
zeropos = data.find(b'\x00', pos)
|
|
|
nlpos = data.find(b'\n', zeropos)
|
|
|
assert 0 <= needle <= len(self.positions)
|
|
|
assert len(self.extrainfo) == len(self.positions)
|
|
|
if zeropos == -1 or nlpos == -1 or nlpos < zeropos:
|
|
|
raise error.StorageError(b'Invalid manifest line')
|
|
|
hlen = nlpos - zeropos - 1
|
|
|
flags = data[nlpos - 1 : nlpos]
|
|
|
if flags in _manifestflags:
|
|
|
hlen -= 1
|
|
|
else:
|
|
|
flags = b''
|
|
|
if hlen != 2 * self._nodelen:
|
|
|
raise error.StorageError(b'Invalid manifest line')
|
|
|
hashval = unhexlify(data, self.extrainfo[needle], zeropos + 1, hlen)
|
|
|
return (hashval, flags)
|
|
|
|
|
|
def __delitem__(self, key: bytes) -> None:
|
|
|
needle, found = self.bsearch2(key)
|
|
|
if not found:
|
|
|
raise KeyError
|
|
|
cur = self.positions[needle]
|
|
|
self.positions = self.positions[:needle] + self.positions[needle + 1 :]
|
|
|
self.extrainfo = self.extrainfo[:needle] + self.extrainfo[needle + 1 :]
|
|
|
if cur >= 0:
|
|
|
# This does NOT unsort the list as far as the search functions are
|
|
|
# concerned, as they only examine lines mapped by self.positions.
|
|
|
self.data = self.data[:cur] + b'\x00' + self.data[cur + 1 :]
|
|
|
self.hasremovals = True
|
|
|
|
|
|
def __setitem__(self, key: bytes, value: Tuple[bytes, bytes]):
|
|
|
if not isinstance(key, bytes):
|
|
|
raise TypeError(b"setitem: manifest keys must be a byte string.")
|
|
|
if not isinstance(value, tuple) or len(value) != 2:
|
|
|
raise TypeError(
|
|
|
b"Manifest values must be a tuple of (node, flags)."
|
|
|
)
|
|
|
hashval = value[0]
|
|
|
if not isinstance(hashval, bytes) or len(hashval) not in (20, 32):
|
|
|
raise TypeError(b"node must be a 20-byte or 32-byte byte string")
|
|
|
flags = value[1]
|
|
|
if not isinstance(flags, bytes) or len(flags) > 1:
|
|
|
raise TypeError(b"flags must a 0 or 1 byte string, got %r", flags)
|
|
|
needle, found = self.bsearch2(key)
|
|
|
if found:
|
|
|
# put the item
|
|
|
pos = self.positions[needle]
|
|
|
if pos < 0:
|
|
|
self.extradata[-pos - 1] = (key, hashval, value[1])
|
|
|
else:
|
|
|
# just don't bother
|
|
|
self.extradata.append((key, hashval, value[1]))
|
|
|
self.positions[needle] = -len(self.extradata)
|
|
|
else:
|
|
|
# not found, put it in with extra positions
|
|
|
self.extradata.append((key, hashval, value[1]))
|
|
|
self.positions = (
|
|
|
self.positions[:needle]
|
|
|
+ [-len(self.extradata)]
|
|
|
+ self.positions[needle:]
|
|
|
)
|
|
|
self.extrainfo = (
|
|
|
self.extrainfo[:needle] + [0] + self.extrainfo[needle:]
|
|
|
)
|
|
|
|
|
|
def copy(self) -> '_LazyManifest':
|
|
|
# XXX call _compact like in C?
|
|
|
return _lazymanifest(
|
|
|
self._nodelen,
|
|
|
self.data,
|
|
|
self.positions,
|
|
|
self.extrainfo,
|
|
|
self.extradata,
|
|
|
self.hasremovals,
|
|
|
)
|
|
|
|
|
|
def _compact(self) -> None:
|
|
|
# hopefully not called TOO often
|
|
|
if len(self.extradata) == 0 and not self.hasremovals:
|
|
|
return
|
|
|
l = []
|
|
|
i = 0
|
|
|
offset = 0
|
|
|
self.extrainfo = [0] * len(self.positions)
|
|
|
while i < len(self.positions):
|
|
|
if self.positions[i] >= 0:
|
|
|
cur = self.positions[i]
|
|
|
last_cut = cur
|
|
|
|
|
|
# Collect all contiguous entries in the buffer at the current
|
|
|
# offset, breaking out only for added/modified items held in
|
|
|
# extradata, or a deleted line prior to the next position.
|
|
|
while True:
|
|
|
self.positions[i] = offset
|
|
|
i += 1
|
|
|
if i == len(self.positions) or self.positions[i] < 0:
|
|
|
break
|
|
|
|
|
|
# A removed file has no positions[] entry, but does have an
|
|
|
# overwritten first byte. Break out and find the end of the
|
|
|
# current good entry/entries if there is a removed file
|
|
|
# before the next position.
|
|
|
if (
|
|
|
self.hasremovals
|
|
|
and self.data.find(b'\n\x00', cur, self.positions[i])
|
|
|
!= -1
|
|
|
):
|
|
|
break
|
|
|
|
|
|
offset += self.positions[i] - cur
|
|
|
cur = self.positions[i]
|
|
|
end_cut = self.data.find(b'\n', cur)
|
|
|
if end_cut != -1:
|
|
|
end_cut += 1
|
|
|
offset += end_cut - cur
|
|
|
l.append(self.data[last_cut:end_cut])
|
|
|
else:
|
|
|
while i < len(self.positions) and self.positions[i] < 0:
|
|
|
cur = self.positions[i]
|
|
|
t = self.extradata[-cur - 1]
|
|
|
l.append(self._pack(t))
|
|
|
self.positions[i] = offset
|
|
|
# Hashes are either 20 bytes (old sha1s) or 32
|
|
|
# bytes (new non-sha1).
|
|
|
hlen = 20
|
|
|
if len(t[1]) > 25:
|
|
|
hlen = 32
|
|
|
if len(t[1]) > hlen:
|
|
|
self.extrainfo[i] = ord(t[1][hlen + 1])
|
|
|
offset += len(l[-1])
|
|
|
i += 1
|
|
|
self.data = b''.join(l)
|
|
|
self.hasremovals = False
|
|
|
self.extradata = []
|
|
|
|
|
|
def _pack(self, d: Tuple[bytes, bytes, bytes]) -> bytes:
|
|
|
n = d[1]
|
|
|
assert len(n) in (20, 32)
|
|
|
return d[0] + b'\x00' + hex(n) + d[2] + b'\n'
|
|
|
|
|
|
def text(self) -> ByteString:
|
|
|
self._compact()
|
|
|
return self.data
|
|
|
|
|
|
def diff(
|
|
|
self, m2: '_LazyManifest', clean: bool = False
|
|
|
) -> Dict[
|
|
|
bytes,
|
|
|
Optional[
|
|
|
Tuple[Tuple[Optional[bytes], bytes], Tuple[Optional[bytes], bytes]]
|
|
|
],
|
|
|
]:
|
|
|
'''Finds changes between the current manifest and m2.'''
|
|
|
# XXX think whether efficiency matters here
|
|
|
diff = {}
|
|
|
|
|
|
for fn, e1, flags in self.iterentries():
|
|
|
if fn not in m2:
|
|
|
diff[fn] = (e1, flags), (None, b'')
|
|
|
else:
|
|
|
e2 = m2[fn]
|
|
|
if (e1, flags) != e2:
|
|
|
diff[fn] = (e1, flags), e2
|
|
|
elif clean:
|
|
|
diff[fn] = None
|
|
|
|
|
|
for fn, e2, flags in m2.iterentries():
|
|
|
if fn not in self:
|
|
|
diff[fn] = (None, b''), (e2, flags)
|
|
|
|
|
|
return diff
|
|
|
|
|
|
def iterentries(self) -> lazymanifestiterentries:
|
|
|
return lazymanifestiterentries(self)
|
|
|
|
|
|
def iterkeys(self) -> lazymanifestiter:
|
|
|
return lazymanifestiter(self)
|
|
|
|
|
|
def __iter__(self) -> lazymanifestiter:
|
|
|
return lazymanifestiter(self)
|
|
|
|
|
|
def __len__(self) -> int:
|
|
|
return len(self.positions)
|
|
|
|
|
|
def filtercopy(self, filterfn: Callable[[bytes], bool]) -> '_LazyManifest':
|
|
|
# XXX should be optimized
|
|
|
c = _lazymanifest(self._nodelen, b'')
|
|
|
for f, n, fl in self.iterentries():
|
|
|
if filterfn(f):
|
|
|
c[f] = n, fl
|
|
|
return c
|
|
|
|
|
|
|
|
|
try:
|
|
|
_lazymanifest = parsers.lazymanifest
|
|
|
except AttributeError:
|
|
|
_lazymanifest = _LazyManifest
|
|
|
|
|
|
|
|
|
class manifestdict: # (repository.imanifestdict)
|
|
|
def __init__(self, nodelen: int, data: ByteString = b''):
|
|
|
self._nodelen = nodelen
|
|
|
self._lm = _lazymanifest(nodelen, data)
|
|
|
|
|
|
def __getitem__(self, key: bytes) -> bytes:
|
|
|
return self._lm[key][0]
|
|
|
|
|
|
def find(self, key: bytes) -> Tuple[bytes, bytes]:
|
|
|
return self._lm[key]
|
|
|
|
|
|
def __len__(self) -> int:
|
|
|
return len(self._lm)
|
|
|
|
|
|
def __nonzero__(self) -> bool:
|
|
|
# nonzero is covered by the __len__ function, but implementing it here
|
|
|
# makes it easier for extensions to override.
|
|
|
return len(self._lm) != 0
|
|
|
|
|
|
__bool__ = __nonzero__
|
|
|
|
|
|
def set(self, key: bytes, node: bytes, flags: bytes) -> None:
|
|
|
self._lm[key] = node, flags
|
|
|
|
|
|
def __setitem__(self, key: bytes, node: bytes) -> None:
|
|
|
self._lm[key] = node, self.flags(key)
|
|
|
|
|
|
def __contains__(self, key: bytes) -> bool:
|
|
|
if key is None:
|
|
|
return False
|
|
|
return key in self._lm
|
|
|
|
|
|
def __delitem__(self, key: bytes) -> None:
|
|
|
del self._lm[key]
|
|
|
|
|
|
def __iter__(self) -> Iterator[bytes]:
|
|
|
return self._lm.__iter__()
|
|
|
|
|
|
def iterkeys(self) -> Iterator[bytes]:
|
|
|
return self._lm.iterkeys()
|
|
|
|
|
|
def keys(self) -> List[bytes]:
|
|
|
return list(self.iterkeys())
|
|
|
|
|
|
def filesnotin(self, m2, match=None) -> Set[bytes]:
|
|
|
'''Set of files in this manifest that are not in the other'''
|
|
|
if match is not None:
|
|
|
match = matchmod.badmatch(match, lambda path, msg: None)
|
|
|
sm2 = set(m2.walk(match))
|
|
|
return {f for f in self.walk(match) if f not in sm2}
|
|
|
return {f for f in self if f not in m2}
|
|
|
|
|
|
@propertycache
|
|
|
def _dirs(self) -> pathutil.dirs:
|
|
|
return pathutil.dirs(self)
|
|
|
|
|
|
def dirs(self) -> pathutil.dirs:
|
|
|
return self._dirs
|
|
|
|
|
|
def hasdir(self, dir: bytes) -> bool:
|
|
|
return dir in self._dirs
|
|
|
|
|
|
def _filesfastpath(self, match: matchmod.basematcher) -> bool:
|
|
|
"""Checks whether we can correctly and quickly iterate over matcher
|
|
|
files instead of over manifest files."""
|
|
|
files = match.files()
|
|
|
return len(files) < 100 and (
|
|
|
match.isexact()
|
|
|
or (match.prefix() and all(fn in self for fn in files))
|
|
|
)
|
|
|
|
|
|
def walk(self, match: matchmod.basematcher) -> Iterator[bytes]:
|
|
|
"""Generates matching file names.
|
|
|
|
|
|
Equivalent to manifest.matches(match).iterkeys(), but without creating
|
|
|
an entirely new manifest.
|
|
|
|
|
|
It also reports nonexistent files by marking them bad with match.bad().
|
|
|
"""
|
|
|
if match.always():
|
|
|
for f in iter(self):
|
|
|
yield f
|
|
|
return
|
|
|
|
|
|
fset = set(match.files())
|
|
|
|
|
|
# avoid the entire walk if we're only looking for specific files
|
|
|
if self._filesfastpath(match):
|
|
|
for fn in sorted(fset):
|
|
|
if fn in self:
|
|
|
yield fn
|
|
|
return
|
|
|
|
|
|
for fn in self:
|
|
|
if fn in fset:
|
|
|
# specified pattern is the exact name
|
|
|
fset.remove(fn)
|
|
|
if match(fn):
|
|
|
yield fn
|
|
|
|
|
|
# for dirstate.walk, files=[''] means "walk the whole tree".
|
|
|
# follow that here, too
|
|
|
fset.discard(b'')
|
|
|
|
|
|
for fn in sorted(fset):
|
|
|
if not self.hasdir(fn):
|
|
|
match.bad(fn, None)
|
|
|
|
|
|
def _matches(self, match: matchmod.basematcher) -> 'manifestdict':
|
|
|
'''generate a new manifest filtered by the match argument'''
|
|
|
if match.always():
|
|
|
return self.copy()
|
|
|
|
|
|
if self._filesfastpath(match):
|
|
|
m = manifestdict(self._nodelen)
|
|
|
lm = self._lm
|
|
|
for fn in match.files():
|
|
|
if fn in lm:
|
|
|
m._lm[fn] = lm[fn]
|
|
|
return m
|
|
|
|
|
|
m = manifestdict(self._nodelen)
|
|
|
m._lm = self._lm.filtercopy(match)
|
|
|
return m
|
|
|
|
|
|
def diff(
|
|
|
self,
|
|
|
m2: 'manifestdict',
|
|
|
match: Optional[matchmod.basematcher] = None,
|
|
|
clean: bool = False,
|
|
|
) -> Dict[
|
|
|
bytes,
|
|
|
Optional[
|
|
|
Tuple[Tuple[Optional[bytes], bytes], Tuple[Optional[bytes], bytes]]
|
|
|
],
|
|
|
]:
|
|
|
"""Finds changes between the current manifest and m2.
|
|
|
|
|
|
Args:
|
|
|
m2: the manifest to which this manifest should be compared.
|
|
|
clean: if true, include files unchanged between these manifests
|
|
|
with a None value in the returned dictionary.
|
|
|
|
|
|
The result is returned as a dict with filename as key and
|
|
|
values of the form ((n1,fl1),(n2,fl2)), where n1/n2 is the
|
|
|
nodeid in the current/other manifest and fl1/fl2 is the flag
|
|
|
in the current/other manifest. Where the file does not exist,
|
|
|
the nodeid will be None and the flags will be the empty
|
|
|
string.
|
|
|
"""
|
|
|
if match:
|
|
|
m1 = self._matches(match)
|
|
|
m2 = m2._matches(match)
|
|
|
return m1.diff(m2, clean=clean)
|
|
|
return self._lm.diff(m2._lm, clean)
|
|
|
|
|
|
def setflag(self, key: bytes, flag: bytes) -> None:
|
|
|
if flag not in _manifestflags:
|
|
|
raise TypeError(b"Invalid manifest flag set.")
|
|
|
self._lm[key] = self[key], flag
|
|
|
|
|
|
def get(self, key: bytes, default=None) -> Optional[bytes]:
|
|
|
try:
|
|
|
return self._lm[key][0]
|
|
|
except KeyError:
|
|
|
return default
|
|
|
|
|
|
def flags(self, key: bytes) -> bytes:
|
|
|
try:
|
|
|
return self._lm[key][1]
|
|
|
except KeyError:
|
|
|
return b''
|
|
|
|
|
|
def copy(self) -> 'manifestdict':
|
|
|
c = manifestdict(self._nodelen)
|
|
|
c._lm = self._lm.copy()
|
|
|
return c
|
|
|
|
|
|
def items(self) -> Iterator[Tuple[bytes, bytes]]:
|
|
|
return (x[:2] for x in self._lm.iterentries())
|
|
|
|
|
|
def iteritems(self) -> Iterator[Tuple[bytes, bytes]]:
|
|
|
return (x[:2] for x in self._lm.iterentries())
|
|
|
|
|
|
def iterentries(self) -> Iterator[Tuple[bytes, bytes, bytes]]:
|
|
|
return self._lm.iterentries()
|
|
|
|
|
|
def text(self) -> ByteString:
|
|
|
# most likely uses native version
|
|
|
return self._lm.text()
|
|
|
|
|
|
def fastdelta(
|
|
|
self, base: ByteString, changes: Iterable[Tuple[bytes, bool]]
|
|
|
) -> Tuple[ByteString, ByteString]:
|
|
|
"""Given a base manifest text as a bytearray and a list of changes
|
|
|
relative to that text, compute a delta that can be used by revlog.
|
|
|
"""
|
|
|
delta = []
|
|
|
dstart = None
|
|
|
dend = None
|
|
|
dline = [b""]
|
|
|
start = 0
|
|
|
# zero copy representation of base as a buffer
|
|
|
addbuf = util.buffer(base)
|
|
|
|
|
|
changes = list(changes)
|
|
|
if len(changes) < FASTDELTA_TEXTDIFF_THRESHOLD:
|
|
|
# start with a readonly loop that finds the offset of
|
|
|
# each line and creates the deltas
|
|
|
for f, todelete in changes:
|
|
|
# bs will either be the index of the item or the insert point
|
|
|
start, end = _msearch(addbuf, f, start)
|
|
|
if not todelete:
|
|
|
h, fl = self._lm[f]
|
|
|
l = b"%s\0%s%s\n" % (f, hex(h), fl)
|
|
|
else:
|
|
|
if start == end:
|
|
|
# item we want to delete was not found, error out
|
|
|
raise AssertionError(
|
|
|
_(b"failed to remove %s from manifest") % f
|
|
|
)
|
|
|
l = b""
|
|
|
if dstart is not None and dstart <= start and dend >= start:
|
|
|
if dend < end:
|
|
|
dend = end
|
|
|
if l:
|
|
|
dline.append(l)
|
|
|
else:
|
|
|
if dstart is not None:
|
|
|
delta.append((dstart, dend, b"".join(dline)))
|
|
|
dstart = start
|
|
|
dend = end
|
|
|
dline = [l]
|
|
|
|
|
|
if dstart is not None:
|
|
|
delta.append((dstart, dend, b"".join(dline)))
|
|
|
# apply the delta to the base, and get a delta for addrevision
|
|
|
deltatext, arraytext = _addlistdelta(base, delta)
|
|
|
else:
|
|
|
# For large changes, it's much cheaper to just build the text and
|
|
|
# diff it.
|
|
|
arraytext = bytearray(self.text())
|
|
|
deltatext = mdiff.textdiff(
|
|
|
util.buffer(base), util.buffer(arraytext)
|
|
|
)
|
|
|
|
|
|
return arraytext, deltatext
|
|
|
|
|
|
|
|
|
def _msearch(
|
|
|
m: ByteString, s: bytes, lo: int = 0, hi: Optional[int] = None
|
|
|
) -> Tuple[int, int]:
|
|
|
"""return a tuple (start, end) that says where to find s within m.
|
|
|
|
|
|
If the string is found m[start:end] are the line containing
|
|
|
that string. If start == end the string was not found and
|
|
|
they indicate the proper sorted insertion point.
|
|
|
"""
|
|
|
|
|
|
def advance(i: int, c: bytes):
|
|
|
while i < lenm and m[i : i + 1] != c:
|
|
|
i += 1
|
|
|
return i
|
|
|
|
|
|
if not s:
|
|
|
return (lo, lo)
|
|
|
lenm = len(m)
|
|
|
if not hi:
|
|
|
hi = lenm
|
|
|
while lo < hi:
|
|
|
mid = (lo + hi) // 2
|
|
|
start = mid
|
|
|
while start > 0 and m[start - 1 : start] != b'\n':
|
|
|
start -= 1
|
|
|
end = advance(start, b'\0')
|
|
|
if bytes(m[start:end]) < s:
|
|
|
# we know that after the null there are 40 bytes of sha1
|
|
|
# this translates to the bisect lo = mid + 1
|
|
|
lo = advance(end + 40, b'\n') + 1
|
|
|
else:
|
|
|
# this translates to the bisect hi = mid
|
|
|
hi = start
|
|
|
end = advance(lo, b'\0')
|
|
|
found = m[lo:end]
|
|
|
if s == found:
|
|
|
# we know that after the null there are 40 bytes of sha1
|
|
|
end = advance(end + 40, b'\n')
|
|
|
return (lo, end + 1)
|
|
|
else:
|
|
|
return (lo, lo)
|
|
|
|
|
|
|
|
|
def _checkforbidden(l: Iterable[bytes]) -> None:
|
|
|
"""Check filenames for illegal characters."""
|
|
|
for f in l:
|
|
|
if b'\n' in f or b'\r' in f:
|
|
|
raise error.StorageError(
|
|
|
_(b"'\\n' and '\\r' disallowed in filenames: %r")
|
|
|
% pycompat.bytestr(f)
|
|
|
)
|
|
|
|
|
|
|
|
|
# apply the changes collected during the bisect loop to our addlist
|
|
|
# return a delta suitable for addrevision
|
|
|
def _addlistdelta(
|
|
|
addlist: ByteString,
|
|
|
x: Iterable[Tuple[int, int, bytes]],
|
|
|
) -> Tuple[bytes, ByteString]:
|
|
|
# for large addlist arrays, building a new array is cheaper
|
|
|
# than repeatedly modifying the existing one
|
|
|
currentposition = 0
|
|
|
newaddlist = bytearray()
|
|
|
|
|
|
for start, end, content in x:
|
|
|
newaddlist += addlist[currentposition:start]
|
|
|
if content:
|
|
|
newaddlist += bytearray(content)
|
|
|
|
|
|
currentposition = end
|
|
|
|
|
|
newaddlist += addlist[currentposition:]
|
|
|
|
|
|
deltatext = b"".join(
|
|
|
struct.pack(b">lll", start, end, len(content)) + content
|
|
|
for start, end, content in x
|
|
|
)
|
|
|
return deltatext, newaddlist
|
|
|
|
|
|
|
|
|
def _splittopdir(f: bytes) -> Tuple[bytes, bytes]:
|
|
|
if b'/' in f:
|
|
|
dir, subpath = f.split(b'/', 1)
|
|
|
return dir + b'/', subpath
|
|
|
else:
|
|
|
return b'', f
|
|
|
|
|
|
|
|
|
_noop = lambda s: None
|
|
|
|
|
|
|
|
|
class treemanifest: # (repository.imanifestdict)
|
|
|
_dir: bytes
|
|
|
_dirs: Dict[bytes, 'treemanifest']
|
|
|
_dirty: bool
|
|
|
_files: Dict[bytes, bytes]
|
|
|
_flags: Dict[bytes, bytes]
|
|
|
|
|
|
def __init__(self, nodeconstants, dir: bytes = b'', text: bytes = b''):
|
|
|
self._dir = dir
|
|
|
self.nodeconstants = nodeconstants
|
|
|
self._node = self.nodeconstants.nullid
|
|
|
self._nodelen = self.nodeconstants.nodelen
|
|
|
self._loadfunc = _noop
|
|
|
self._copyfunc = _noop
|
|
|
self._dirty = False
|
|
|
self._dirs = {}
|
|
|
self._lazydirs: Dict[
|
|
|
bytes,
|
|
|
Tuple[bytes, Callable[[bytes, bytes], 'treemanifest'], bool],
|
|
|
] = {}
|
|
|
# Using _lazymanifest here is a little slower than plain old dicts
|
|
|
self._files = {}
|
|
|
self._flags = {}
|
|
|
if text:
|
|
|
|
|
|
def readsubtree(subdir, subm):
|
|
|
raise AssertionError(
|
|
|
b'treemanifest constructor only accepts flat manifests'
|
|
|
)
|
|
|
|
|
|
self.parse(text, readsubtree)
|
|
|
self._dirty = True # Mark flat manifest dirty after parsing
|
|
|
|
|
|
def _subpath(self, path: bytes) -> bytes:
|
|
|
return self._dir + path
|
|
|
|
|
|
def _loadalllazy(self) -> None:
|
|
|
selfdirs = self._dirs
|
|
|
subpath = self._subpath
|
|
|
for d, (node, readsubtree, docopy) in self._lazydirs.items():
|
|
|
if docopy:
|
|
|
selfdirs[d] = readsubtree(subpath(d), node).copy()
|
|
|
else:
|
|
|
selfdirs[d] = readsubtree(subpath(d), node)
|
|
|
self._lazydirs.clear()
|
|
|
|
|
|
def _loadlazy(self, d: bytes) -> None:
|
|
|
v = self._lazydirs.get(d)
|
|
|
if v is not None:
|
|
|
node, readsubtree, docopy = v
|
|
|
if docopy:
|
|
|
self._dirs[d] = readsubtree(self._subpath(d), node).copy()
|
|
|
else:
|
|
|
self._dirs[d] = readsubtree(self._subpath(d), node)
|
|
|
del self._lazydirs[d]
|
|
|
|
|
|
def _loadchildrensetlazy(
|
|
|
self, visit: Union[Set[bytes], bytes]
|
|
|
) -> Optional[Set[bytes]]:
|
|
|
if not visit:
|
|
|
return None
|
|
|
if visit == b'all' or visit == b'this':
|
|
|
self._loadalllazy()
|
|
|
return None
|
|
|
|
|
|
visit = cast(Set[bytes], visit)
|
|
|
|
|
|
loadlazy = self._loadlazy
|
|
|
for k in visit:
|
|
|
loadlazy(k + b'/')
|
|
|
return visit
|
|
|
|
|
|
def _loaddifflazy(self, t1: 'treemanifest', t2: 'treemanifest'):
|
|
|
"""load items in t1 and t2 if they're needed for diffing.
|
|
|
|
|
|
The criteria currently is:
|
|
|
- if it's not present in _lazydirs in either t1 or t2, load it in the
|
|
|
other (it may already be loaded or it may not exist, doesn't matter)
|
|
|
- if it's present in _lazydirs in both, compare the nodeid; if it
|
|
|
differs, load it in both
|
|
|
"""
|
|
|
toloadlazy = []
|
|
|
for d, v1 in t1._lazydirs.items():
|
|
|
v2 = t2._lazydirs.get(d)
|
|
|
if v2 is None or v2[0] != v1[0]:
|
|
|
toloadlazy.append(d)
|
|
|
for d, v1 in t2._lazydirs.items():
|
|
|
if d not in t1._lazydirs:
|
|
|
toloadlazy.append(d)
|
|
|
|
|
|
for d in toloadlazy:
|
|
|
t1._loadlazy(d)
|
|
|
t2._loadlazy(d)
|
|
|
|
|
|
def __len__(self) -> int:
|
|
|
self._load()
|
|
|
size = len(self._files)
|
|
|
self._loadalllazy()
|
|
|
for m in self._dirs.values():
|
|
|
size += m.__len__()
|
|
|
return size
|
|
|
|
|
|
def __nonzero__(self) -> bool:
|
|
|
# Faster than "__len__() != 0" since it avoids loading sub-manifests
|
|
|
return not self._isempty()
|
|
|
|
|
|
__bool__ = __nonzero__
|
|
|
|
|
|
def _isempty(self) -> bool:
|
|
|
self._load() # for consistency; already loaded by all callers
|
|
|
# See if we can skip loading everything.
|
|
|
if self._files or (
|
|
|
self._dirs and any(not m._isempty() for m in self._dirs.values())
|
|
|
):
|
|
|
return False
|
|
|
self._loadalllazy()
|
|
|
return not self._dirs or all(m._isempty() for m in self._dirs.values())
|
|
|
|
|
|
@encoding.strmethod
|
|
|
def __repr__(self) -> bytes:
|
|
|
return (
|
|
|
b'<treemanifest dir=%s, node=%s, loaded=%r, dirty=%r at 0x%x>'
|
|
|
% (
|
|
|
self._dir,
|
|
|
hex(self._node),
|
|
|
bool(self._loadfunc is _noop),
|
|
|
self._dirty,
|
|
|
id(self),
|
|
|
)
|
|
|
)
|
|
|
|
|
|
def dir(self) -> bytes:
|
|
|
"""The directory that this tree manifest represents, including a
|
|
|
trailing '/'. Empty string for the repo root directory."""
|
|
|
return self._dir
|
|
|
|
|
|
def node(self) -> bytes:
|
|
|
"""This node of this instance. nullid for unsaved instances. Should
|
|
|
be updated when the instance is read or written from a revlog.
|
|
|
"""
|
|
|
assert not self._dirty
|
|
|
return self._node
|
|
|
|
|
|
def setnode(self, node: bytes) -> None:
|
|
|
self._node = node
|
|
|
self._dirty = False
|
|
|
|
|
|
def iterentries(
|
|
|
self,
|
|
|
) -> Iterator[Tuple[bytes, Union[bytes, 'treemanifest'], bytes]]:
|
|
|
self._load()
|
|
|
self._loadalllazy()
|
|
|
for p, n in sorted(
|
|
|
itertools.chain(self._dirs.items(), self._files.items())
|
|
|
):
|
|
|
if p in self._files:
|
|
|
yield self._subpath(p), n, self._flags.get(p, b'')
|
|
|
else:
|
|
|
for x in n.iterentries():
|
|
|
yield x
|
|
|
|
|
|
def items(self) -> Iterator[Tuple[bytes, Union[bytes, 'treemanifest']]]:
|
|
|
self._load()
|
|
|
self._loadalllazy()
|
|
|
for p, n in sorted(
|
|
|
itertools.chain(self._dirs.items(), self._files.items())
|
|
|
):
|
|
|
if p in self._files:
|
|
|
yield self._subpath(p), n
|
|
|
else:
|
|
|
for f, sn in n.items():
|
|
|
yield f, sn
|
|
|
|
|
|
iteritems = items
|
|
|
|
|
|
def iterkeys(self) -> Iterator[bytes]:
|
|
|
self._load()
|
|
|
self._loadalllazy()
|
|
|
for p in sorted(itertools.chain(self._dirs, self._files)):
|
|
|
if p in self._files:
|
|
|
yield self._subpath(p)
|
|
|
else:
|
|
|
for f in self._dirs[p]:
|
|
|
yield f
|
|
|
|
|
|
def keys(self) -> List[bytes]:
|
|
|
return list(self.iterkeys())
|
|
|
|
|
|
def __iter__(self) -> Iterator[bytes]:
|
|
|
return self.iterkeys()
|
|
|
|
|
|
def __contains__(self, f: bytes) -> bool:
|
|
|
if f is None:
|
|
|
return False
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
if dir not in self._dirs:
|
|
|
return False
|
|
|
|
|
|
return self._dirs[dir].__contains__(subpath)
|
|
|
else:
|
|
|
return f in self._files
|
|
|
|
|
|
def get(self, f: bytes, default: Optional[bytes] = None) -> Optional[bytes]:
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
if dir not in self._dirs:
|
|
|
return default
|
|
|
return self._dirs[dir].get(subpath, default)
|
|
|
else:
|
|
|
return self._files.get(f, default)
|
|
|
|
|
|
def __getitem__(self, f: bytes) -> bytes:
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
return self._dirs[dir].__getitem__(subpath)
|
|
|
else:
|
|
|
return self._files[f]
|
|
|
|
|
|
def flags(self, f: bytes) -> bytes:
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
if dir not in self._dirs:
|
|
|
return b''
|
|
|
return self._dirs[dir].flags(subpath)
|
|
|
else:
|
|
|
if f in self._lazydirs or f in self._dirs:
|
|
|
return b''
|
|
|
return self._flags.get(f, b'')
|
|
|
|
|
|
def find(self, f: bytes) -> Tuple[bytes, bytes]:
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
return self._dirs[dir].find(subpath)
|
|
|
else:
|
|
|
return self._files[f], self._flags.get(f, b'')
|
|
|
|
|
|
def __delitem__(self, f: bytes) -> None:
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
|
|
|
self._dirs[dir].__delitem__(subpath)
|
|
|
# If the directory is now empty, remove it
|
|
|
if self._dirs[dir]._isempty():
|
|
|
del self._dirs[dir]
|
|
|
else:
|
|
|
del self._files[f]
|
|
|
if f in self._flags:
|
|
|
del self._flags[f]
|
|
|
self._dirty = True
|
|
|
|
|
|
def set(self, f: bytes, node: bytes, flags: bytes) -> None:
|
|
|
"""Set both the node and the flags for path f."""
|
|
|
assert node is not None
|
|
|
if flags not in _manifestflags:
|
|
|
raise TypeError(b"Invalid manifest flag set.")
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
if dir not in self._dirs:
|
|
|
self._dirs[dir] = treemanifest(
|
|
|
self.nodeconstants, self._subpath(dir)
|
|
|
)
|
|
|
self._dirs[dir].set(subpath, node, flags)
|
|
|
else:
|
|
|
assert len(node) in (20, 32)
|
|
|
self._files[f] = node
|
|
|
self._flags[f] = flags
|
|
|
self._dirty = True
|
|
|
|
|
|
def __setitem__(self, f: bytes, n: bytes) -> None:
|
|
|
assert n is not None
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
if dir not in self._dirs:
|
|
|
self._dirs[dir] = treemanifest(
|
|
|
self.nodeconstants, self._subpath(dir)
|
|
|
)
|
|
|
self._dirs[dir].__setitem__(subpath, n)
|
|
|
else:
|
|
|
# manifest nodes are either 20 bytes or 32 bytes,
|
|
|
# depending on the hash in use. Assert this as historically
|
|
|
# sometimes extra bytes were added.
|
|
|
assert len(n) in (20, 32)
|
|
|
self._files[f] = n
|
|
|
self._dirty = True
|
|
|
|
|
|
def _load(self) -> None:
|
|
|
if self._loadfunc is not _noop:
|
|
|
lf, self._loadfunc = self._loadfunc, _noop
|
|
|
lf(self)
|
|
|
elif self._copyfunc is not _noop:
|
|
|
cf, self._copyfunc = self._copyfunc, _noop
|
|
|
cf(self)
|
|
|
|
|
|
def setflag(self, f: bytes, flags: bytes) -> None:
|
|
|
"""Set the flags (symlink, executable) for path f."""
|
|
|
if flags not in _manifestflags:
|
|
|
raise TypeError(b"Invalid manifest flag set.")
|
|
|
self._load()
|
|
|
dir, subpath = _splittopdir(f)
|
|
|
if dir:
|
|
|
self._loadlazy(dir)
|
|
|
if dir not in self._dirs:
|
|
|
self._dirs[dir] = treemanifest(
|
|
|
self.nodeconstants, self._subpath(dir)
|
|
|
)
|
|
|
self._dirs[dir].setflag(subpath, flags)
|
|
|
else:
|
|
|
self._flags[f] = flags
|
|
|
self._dirty = True
|
|
|
|
|
|
def copy(self) -> 'treemanifest':
|
|
|
copy = treemanifest(self.nodeconstants, self._dir)
|
|
|
copy._node = self._node
|
|
|
copy._dirty = self._dirty
|
|
|
if self._copyfunc is _noop:
|
|
|
|
|
|
def _copyfunc(s):
|
|
|
self._load()
|
|
|
s._lazydirs = {
|
|
|
d: (n, r, True) for d, (n, r, c) in self._lazydirs.items()
|
|
|
}
|
|
|
sdirs = s._dirs
|
|
|
for d, v in self._dirs.items():
|
|
|
sdirs[d] = v.copy()
|
|
|
s._files = dict.copy(self._files)
|
|
|
s._flags = dict.copy(self._flags)
|
|
|
|
|
|
if self._loadfunc is _noop:
|
|
|
_copyfunc(copy)
|
|
|
else:
|
|
|
copy._copyfunc = _copyfunc
|
|
|
else:
|
|
|
copy._copyfunc = self._copyfunc
|
|
|
return copy
|
|
|
|
|
|
def filesnotin(
|
|
|
self, m2: 'treemanifest', match: Optional[matchmod.basematcher] = None
|
|
|
) -> Set[bytes]:
|
|
|
'''Set of files in this manifest that are not in the other'''
|
|
|
if match and not match.always():
|
|
|
m1 = self._matches(match)
|
|
|
m2 = m2._matches(match)
|
|
|
return m1.filesnotin(m2)
|
|
|
|
|
|
files = set()
|
|
|
|
|
|
def _filesnotin(t1, t2):
|
|
|
if t1._node == t2._node and not t1._dirty and not t2._dirty:
|
|
|
return
|
|
|
t1._load()
|
|
|
t2._load()
|
|
|
self._loaddifflazy(t1, t2)
|
|
|
for d, m1 in t1._dirs.items():
|
|
|
if d in t2._dirs:
|
|
|
m2 = t2._dirs[d]
|
|
|
_filesnotin(m1, m2)
|
|
|
else:
|
|
|
files.update(m1.iterkeys())
|
|
|
|
|
|
for fn in t1._files:
|
|
|
if fn not in t2._files:
|
|
|
files.add(t1._subpath(fn))
|
|
|
|
|
|
_filesnotin(self, m2)
|
|
|
return files
|
|
|
|
|
|
@propertycache
|
|
|
def _alldirs(self) -> pathutil.dirs:
|
|
|
return pathutil.dirs(self)
|
|
|
|
|
|
def dirs(self) -> pathutil.dirs:
|
|
|
return self._alldirs
|
|
|
|
|
|
def hasdir(self, dir: bytes) -> bool:
|
|
|
self._load()
|
|
|
topdir, subdir = _splittopdir(dir)
|
|
|
if topdir:
|
|
|
self._loadlazy(topdir)
|
|
|
if topdir in self._dirs:
|
|
|
return self._dirs[topdir].hasdir(subdir)
|
|
|
return False
|
|
|
dirslash = dir + b'/'
|
|
|
return dirslash in self._dirs or dirslash in self._lazydirs
|
|
|
|
|
|
def walk(self, match: matchmod.basematcher) -> Iterator[bytes]:
|
|
|
"""Generates matching file names.
|
|
|
|
|
|
It also reports nonexistent files by marking them bad with match.bad().
|
|
|
"""
|
|
|
if match.always():
|
|
|
for f in iter(self):
|
|
|
yield f
|
|
|
return
|
|
|
|
|
|
fset = set(match.files())
|
|
|
|
|
|
for fn in self._walk(match):
|
|
|
if fn in fset:
|
|
|
# specified pattern is the exact name
|
|
|
fset.remove(fn)
|
|
|
yield fn
|
|
|
|
|
|
# for dirstate.walk, files=[''] means "walk the whole tree".
|
|
|
# follow that here, too
|
|
|
fset.discard(b'')
|
|
|
|
|
|
for fn in sorted(fset):
|
|
|
if not self.hasdir(fn):
|
|
|
match.bad(fn, None)
|
|
|
|
|
|
def _walk(self, match: matchmod.basematcher) -> Iterator[bytes]:
|
|
|
'''Recursively generates matching file names for walk().'''
|
|
|
visit = match.visitchildrenset(self._dir[:-1])
|
|
|
if not visit:
|
|
|
return
|
|
|
|
|
|
# yield this dir's files and walk its submanifests
|
|
|
self._load()
|
|
|
visit = self._loadchildrensetlazy(visit)
|
|
|
for p in sorted(list(self._dirs) + list(self._files)):
|
|
|
if p in self._files:
|
|
|
fullp = self._subpath(p)
|
|
|
if match(fullp):
|
|
|
yield fullp
|
|
|
else:
|
|
|
if not visit or p[:-1] in visit:
|
|
|
for f in self._dirs[p]._walk(match):
|
|
|
yield f
|
|
|
|
|
|
def _matches(self, match: matchmod.basematcher) -> 'treemanifest':
|
|
|
"""recursively generate a new manifest filtered by the match argument."""
|
|
|
if match.always():
|
|
|
return self.copy()
|
|
|
return self._matches_inner(match)
|
|
|
|
|
|
def _matches_inner(self, match: matchmod.basematcher) -> 'treemanifest':
|
|
|
if match.always():
|
|
|
return self.copy()
|
|
|
|
|
|
visit = match.visitchildrenset(self._dir[:-1])
|
|
|
if visit == b'all':
|
|
|
return self.copy()
|
|
|
ret = treemanifest(self.nodeconstants, self._dir)
|
|
|
if not visit:
|
|
|
return ret
|
|
|
|
|
|
self._load()
|
|
|
for fn in self._files:
|
|
|
# While visitchildrenset *usually* lists only subdirs, this is
|
|
|
# actually up to the matcher and may have some files in the set().
|
|
|
# If visit == 'this', we should obviously look at the files in this
|
|
|
# directory; if visit is a set, and fn is in it, we should inspect
|
|
|
# fn (but no need to inspect things not in the set).
|
|
|
if visit != b'this' and fn not in visit:
|
|
|
continue
|
|
|
fullp = self._subpath(fn)
|
|
|
# visitchildrenset isn't perfect, we still need to call the regular
|
|
|
# matcher code to further filter results.
|
|
|
if not match(fullp):
|
|
|
continue
|
|
|
ret._files[fn] = self._files[fn]
|
|
|
if fn in self._flags:
|
|
|
ret._flags[fn] = self._flags[fn]
|
|
|
|
|
|
visit = self._loadchildrensetlazy(visit)
|
|
|
for dir, subm in self._dirs.items():
|
|
|
if visit and dir[:-1] not in visit:
|
|
|
continue
|
|
|
m = subm._matches_inner(match)
|
|
|
if not m._isempty():
|
|
|
ret._dirs[dir] = m
|
|
|
|
|
|
if not ret._isempty():
|
|
|
ret._dirty = True
|
|
|
return ret
|
|
|
|
|
|
def fastdelta(
|
|
|
self, base: ByteString, changes: Iterable[Tuple[bytes, bool]]
|
|
|
) -> ByteString:
|
|
|
raise FastdeltaUnavailable()
|
|
|
|
|
|
def diff(
|
|
|
self,
|
|
|
m2: 'treemanifest',
|
|
|
match: Optional[matchmod.basematcher] = None,
|
|
|
clean: bool = False,
|
|
|
) -> Dict[
|
|
|
bytes,
|
|
|
Optional[
|
|
|
Tuple[Tuple[Optional[bytes], bytes], Tuple[Optional[bytes], bytes]]
|
|
|
],
|
|
|
]:
|
|
|
"""Finds changes between the current manifest and m2.
|
|
|
|
|
|
Args:
|
|
|
m2: the manifest to which this manifest should be compared.
|
|
|
clean: if true, include files unchanged between these manifests
|
|
|
with a None value in the returned dictionary.
|
|
|
|
|
|
The result is returned as a dict with filename as key and
|
|
|
values of the form ((n1,fl1),(n2,fl2)), where n1/n2 is the
|
|
|
nodeid in the current/other manifest and fl1/fl2 is the flag
|
|
|
in the current/other manifest. Where the file does not exist,
|
|
|
the nodeid will be None and the flags will be the empty
|
|
|
string.
|
|
|
"""
|
|
|
if match and not match.always():
|
|
|
m1 = self._matches(match)
|
|
|
m2 = m2._matches(match)
|
|
|
return m1.diff(m2, clean=clean)
|
|
|
result = {}
|
|
|
emptytree = treemanifest(self.nodeconstants)
|
|
|
|
|
|
def _iterativediff(t1, t2, stack):
|
|
|
"""compares two tree manifests and append new tree-manifests which
|
|
|
needs to be compared to stack"""
|
|
|
if t1._node == t2._node and not t1._dirty and not t2._dirty:
|
|
|
return
|
|
|
t1._load()
|
|
|
t2._load()
|
|
|
self._loaddifflazy(t1, t2)
|
|
|
|
|
|
for d, m1 in t1._dirs.items():
|
|
|
m2 = t2._dirs.get(d, emptytree)
|
|
|
stack.append((m1, m2))
|
|
|
|
|
|
for d, m2 in t2._dirs.items():
|
|
|
if d not in t1._dirs:
|
|
|
stack.append((emptytree, m2))
|
|
|
|
|
|
for fn, n1 in t1._files.items():
|
|
|
fl1 = t1._flags.get(fn, b'')
|
|
|
n2 = t2._files.get(fn, None)
|
|
|
fl2 = t2._flags.get(fn, b'')
|
|
|
if n1 != n2 or fl1 != fl2:
|
|
|
result[t1._subpath(fn)] = ((n1, fl1), (n2, fl2))
|
|
|
elif clean:
|
|
|
result[t1._subpath(fn)] = None
|
|
|
|
|
|
for fn, n2 in t2._files.items():
|
|
|
if fn not in t1._files:
|
|
|
fl2 = t2._flags.get(fn, b'')
|
|
|
result[t2._subpath(fn)] = ((None, b''), (n2, fl2))
|
|
|
|
|
|
stackls = []
|
|
|
_iterativediff(self, m2, stackls)
|
|
|
while stackls:
|
|
|
t1, t2 = stackls.pop()
|
|
|
# stackls is populated in the function call
|
|
|
_iterativediff(t1, t2, stackls)
|
|
|
return result
|
|
|
|
|
|
def unmodifiedsince(self, m2: 'treemanifest') -> bool:
|
|
|
return not self._dirty and not m2._dirty and self._node == m2._node
|
|
|
|
|
|
def parse(
|
|
|
self,
|
|
|
text: bytes,
|
|
|
readsubtree: Callable[[bytes, bytes], 'treemanifest'],
|
|
|
) -> None:
|
|
|
selflazy = self._lazydirs
|
|
|
for f, n, fl in _parse(self._nodelen, text):
|
|
|
if fl == b't':
|
|
|
f = f + b'/'
|
|
|
# False below means "doesn't need to be copied" and can use the
|
|
|
# cached value from readsubtree directly.
|
|
|
selflazy[f] = (n, readsubtree, False)
|
|
|
elif b'/' in f:
|
|
|
# This is a flat manifest, so use __setitem__ and setflag rather
|
|
|
# than assigning directly to _files and _flags, so we can
|
|
|
# assign a path in a subdirectory, and to mark dirty (compared
|
|
|
# to nullid).
|
|
|
self[f] = n
|
|
|
if fl:
|
|
|
self.setflag(f, fl)
|
|
|
else:
|
|
|
# Assigning to _files and _flags avoids marking as dirty,
|
|
|
# and should be a little faster.
|
|
|
self._files[f] = n
|
|
|
if fl:
|
|
|
self._flags[f] = fl
|
|
|
|
|
|
def text(self) -> ByteString:
|
|
|
"""Get the full data of this manifest as a bytestring."""
|
|
|
self._load()
|
|
|
return _text(self.iterentries())
|
|
|
|
|
|
def dirtext(self) -> ByteString:
|
|
|
"""Get the full data of this directory as a bytestring. Make sure that
|
|
|
any submanifests have been written first, so their nodeids are correct.
|
|
|
"""
|
|
|
self._load()
|
|
|
flags = self.flags
|
|
|
lazydirs = [(d[:-1], v[0], b't') for d, v in self._lazydirs.items()]
|
|
|
dirs = [(d[:-1], self._dirs[d]._node, b't') for d in self._dirs]
|
|
|
files = [(f, self._files[f], flags(f)) for f in self._files]
|
|
|
return _text(sorted(dirs + files + lazydirs))
|
|
|
|
|
|
def read(
|
|
|
self,
|
|
|
gettext: Callable[[], ByteString],
|
|
|
readsubtree: Callable[[bytes, bytes], 'treemanifest'],
|
|
|
) -> None:
|
|
|
def _load_for_read(s):
|
|
|
s.parse(gettext(), readsubtree)
|
|
|
s._dirty = False
|
|
|
|
|
|
self._loadfunc = _load_for_read
|
|
|
|
|
|
def writesubtrees(
|
|
|
self,
|
|
|
m1: 'treemanifest',
|
|
|
m2: 'treemanifest',
|
|
|
writesubtree: Callable[
|
|
|
[
|
|
|
Callable[['treemanifest'], None],
|
|
|
bytes,
|
|
|
bytes,
|
|
|
matchmod.basematcher,
|
|
|
],
|
|
|
None,
|
|
|
],
|
|
|
match: matchmod.basematcher,
|
|
|
) -> None:
|
|
|
self._load() # for consistency; should never have any effect here
|
|
|
m1._load()
|
|
|
m2._load()
|
|
|
emptytree = treemanifest(self.nodeconstants)
|
|
|
|
|
|
def getnode(m, d):
|
|
|
ld = m._lazydirs.get(d)
|
|
|
if ld:
|
|
|
return ld[0]
|
|
|
tree = m._dirs.get(d, emptytree)
|
|
|
assert tree is not None # helps pytype
|
|
|
return tree._node
|
|
|
|
|
|
# let's skip investigating things that `match` says we do not need.
|
|
|
visit = match.visitchildrenset(self._dir[:-1])
|
|
|
visit = self._loadchildrensetlazy(visit)
|
|
|
if visit == b'this' or visit == b'all':
|
|
|
visit = None
|
|
|
for d, subm in self._dirs.items():
|
|
|
if visit and d[:-1] not in visit:
|
|
|
continue
|
|
|
subp1 = getnode(m1, d)
|
|
|
subp2 = getnode(m2, d)
|
|
|
if subp1 == self.nodeconstants.nullid:
|
|
|
subp1, subp2 = subp2, subp1
|
|
|
writesubtree(subm, subp1, subp2, match)
|
|
|
|
|
|
def walksubtrees(
|
|
|
self, matcher: Optional[matchmod.basematcher] = None
|
|
|
) -> Iterator['treemanifest']:
|
|
|
"""Returns an iterator of the subtrees of this manifest, including this
|
|
|
manifest itself.
|
|
|
|
|
|
If `matcher` is provided, it only returns subtrees that match.
|
|
|
"""
|
|
|
if matcher and not matcher.visitdir(self._dir[:-1]):
|
|
|
return
|
|
|
if not matcher or matcher(self._dir[:-1]):
|
|
|
yield self
|
|
|
|
|
|
self._load()
|
|
|
# OPT: use visitchildrenset to avoid loading everything.
|
|
|
self._loadalllazy()
|
|
|
for d, subm in self._dirs.items():
|
|
|
for subtree in subm.walksubtrees(matcher=matcher):
|
|
|
yield subtree
|
|
|
|
|
|
|
|
|
class manifestfulltextcache(util.lrucachedict):
|
|
|
"""File-backed LRU cache for the manifest cache
|
|
|
|
|
|
File consists of entries, up to EOF:
|
|
|
|
|
|
- 20 bytes node, 4 bytes length, <length> manifest data
|
|
|
|
|
|
These are written in reverse cache order (oldest to newest).
|
|
|
|
|
|
"""
|
|
|
|
|
|
_file = b'manifestfulltextcache'
|
|
|
|
|
|
def __init__(self, max):
|
|
|
super(manifestfulltextcache, self).__init__(max)
|
|
|
self._dirty = False
|
|
|
self._read = False
|
|
|
self._opener = None
|
|
|
|
|
|
def read(self):
|
|
|
if self._read or self._opener is None:
|
|
|
return
|
|
|
|
|
|
try:
|
|
|
with self._opener(self._file) as fp:
|
|
|
set = super(manifestfulltextcache, self).__setitem__
|
|
|
# ignore trailing data, this is a cache, corruption is skipped
|
|
|
while True:
|
|
|
# TODO do we need to do work here for sha1 portability?
|
|
|
node = fp.read(20)
|
|
|
if len(node) < 20:
|
|
|
break
|
|
|
try:
|
|
|
size = struct.unpack(b'>L', fp.read(4))[0]
|
|
|
except struct.error:
|
|
|
break
|
|
|
value = bytearray(fp.read(size))
|
|
|
if len(value) != size:
|
|
|
break
|
|
|
set(node, value)
|
|
|
except IOError:
|
|
|
# the file is allowed to be missing
|
|
|
pass
|
|
|
|
|
|
self._read = True
|
|
|
self._dirty = False
|
|
|
|
|
|
def write(self):
|
|
|
if not self._dirty or self._opener is None:
|
|
|
return
|
|
|
# rotate backwards to the first used node
|
|
|
try:
|
|
|
with self._opener(
|
|
|
self._file, b'w', atomictemp=True, checkambig=True
|
|
|
) as fp:
|
|
|
node = self._head.prev
|
|
|
while True:
|
|
|
if node.key in self._cache:
|
|
|
fp.write(node.key)
|
|
|
fp.write(struct.pack(b'>L', len(node.value)))
|
|
|
fp.write(node.value)
|
|
|
if node is self._head:
|
|
|
break
|
|
|
node = node.prev
|
|
|
except IOError:
|
|
|
# We could not write the cache (eg: permission error)
|
|
|
# the content can be missing.
|
|
|
#
|
|
|
# We could try harder and see if we could recreate a wcache
|
|
|
# directory were we coudl write too.
|
|
|
#
|
|
|
# XXX the error pass silently, having some way to issue an error
|
|
|
# log `ui.log` would be nice.
|
|
|
pass
|
|
|
|
|
|
def __len__(self):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
return super(manifestfulltextcache, self).__len__()
|
|
|
|
|
|
def __contains__(self, k):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
return super(manifestfulltextcache, self).__contains__(k)
|
|
|
|
|
|
def __iter__(self):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
return super(manifestfulltextcache, self).__iter__()
|
|
|
|
|
|
def __getitem__(self, k):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
# the cache lru order can change on read
|
|
|
setdirty = self._cache.get(k) is not self._head
|
|
|
value = super(manifestfulltextcache, self).__getitem__(k)
|
|
|
if setdirty:
|
|
|
self._dirty = True
|
|
|
return value
|
|
|
|
|
|
def __setitem__(self, k, v):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
super(manifestfulltextcache, self).__setitem__(k, v)
|
|
|
self._dirty = True
|
|
|
|
|
|
def __delitem__(self, k):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
super(manifestfulltextcache, self).__delitem__(k)
|
|
|
self._dirty = True
|
|
|
|
|
|
def get(self, k, default=None):
|
|
|
if not self._read:
|
|
|
self.read()
|
|
|
return super(manifestfulltextcache, self).get(k, default=default)
|
|
|
|
|
|
def clear(self, clear_persisted_data=False):
|
|
|
super(manifestfulltextcache, self).clear()
|
|
|
if clear_persisted_data:
|
|
|
self._dirty = True
|
|
|
self.write()
|
|
|
self._read = False
|
|
|
|
|
|
|
|
|
# and upper bound of what we expect from compression
|
|
|
# (real live value seems to be "3")
|
|
|
MAXCOMPRESSION = 3
|
|
|
|
|
|
|
|
|
class FastdeltaUnavailable(Exception):
|
|
|
"""Exception raised when fastdelta isn't usable on a manifest."""
|
|
|
|
|
|
|
|
|
class manifestrevlog: # (repository.imanifeststorage)
|
|
|
"""A revlog that stores manifest texts. This is responsible for caching the
|
|
|
full-text manifest contents.
|
|
|
"""
|
|
|
|
|
|
def __init__(
|
|
|
self,
|
|
|
nodeconstants,
|
|
|
opener,
|
|
|
tree=b'',
|
|
|
dirlogcache=None,
|
|
|
treemanifest=False,
|
|
|
):
|
|
|
"""Constructs a new manifest revlog
|
|
|
|
|
|
`indexfile` - used by extensions to have two manifests at once, like
|
|
|
when transitioning between flatmanifeset and treemanifests.
|
|
|
|
|
|
`treemanifest` - used to indicate this is a tree manifest revlog. Opener
|
|
|
options can also be used to make this a tree manifest revlog. The opener
|
|
|
option takes precedence, so if it is set to True, we ignore whatever
|
|
|
value is passed in to the constructor.
|
|
|
"""
|
|
|
self.nodeconstants = nodeconstants
|
|
|
# During normal operations, we expect to deal with not more than four
|
|
|
# revs at a time (such as during commit --amend). When rebasing large
|
|
|
# stacks of commits, the number can go up, hence the config knob below.
|
|
|
cachesize = 4
|
|
|
optiontreemanifest = False
|
|
|
persistentnodemap = False
|
|
|
opts = getattr(opener, 'options', None)
|
|
|
if opts is not None:
|
|
|
cachesize = opts.get(b'manifestcachesize', cachesize)
|
|
|
optiontreemanifest = opts.get(b'treemanifest', False)
|
|
|
persistentnodemap = opts.get(b'persistent-nodemap', False)
|
|
|
|
|
|
self._treeondisk = optiontreemanifest or treemanifest
|
|
|
|
|
|
self._fulltextcache = manifestfulltextcache(cachesize)
|
|
|
|
|
|
if tree:
|
|
|
assert self._treeondisk, (tree, b'opts is %r' % opts)
|
|
|
|
|
|
radix = b'00manifest'
|
|
|
if tree:
|
|
|
radix = b"meta/" + tree + radix
|
|
|
|
|
|
self.tree = tree
|
|
|
|
|
|
# The dirlogcache is kept on the root manifest log
|
|
|
if tree:
|
|
|
self._dirlogcache = dirlogcache
|
|
|
else:
|
|
|
self._dirlogcache = {b'': self}
|
|
|
|
|
|
self._revlog = revlog.revlog(
|
|
|
opener,
|
|
|
target=(revlog_constants.KIND_MANIFESTLOG, self.tree),
|
|
|
radix=radix,
|
|
|
# only root indexfile is cached
|
|
|
checkambig=not bool(tree),
|
|
|
mmaplargeindex=True,
|
|
|
upperboundcomp=MAXCOMPRESSION,
|
|
|
persistentnodemap=persistentnodemap,
|
|
|
)
|
|
|
|
|
|
self.index = self._revlog.index
|
|
|
|
|
|
def get_revlog(self):
|
|
|
"""return an actual revlog instance if any
|
|
|
|
|
|
This exist because a lot of code leverage the fact the underlying
|
|
|
storage is a revlog for optimization, so giving simple way to access
|
|
|
the revlog instance helps such code.
|
|
|
"""
|
|
|
return self._revlog
|
|
|
|
|
|
def _setupmanifestcachehooks(self, repo):
|
|
|
"""Persist the manifestfulltextcache on lock release"""
|
|
|
if not hasattr(repo, '_wlockref'):
|
|
|
return
|
|
|
|
|
|
self._fulltextcache._opener = repo.wcachevfs
|
|
|
if repo._currentlock(repo._wlockref) is None:
|
|
|
return
|
|
|
|
|
|
reporef = weakref.ref(repo)
|
|
|
manifestrevlogref = weakref.ref(self)
|
|
|
|
|
|
def persistmanifestcache(success):
|
|
|
# Repo is in an unknown state, do not persist.
|
|
|
if not success:
|
|
|
return
|
|
|
|
|
|
repo = reporef()
|
|
|
self = manifestrevlogref()
|
|
|
if repo is None or self is None:
|
|
|
return
|
|
|
if repo.manifestlog.getstorage(b'') is not self:
|
|
|
# there's a different manifest in play now, abort
|
|
|
return
|
|
|
self._fulltextcache.write()
|
|
|
|
|
|
repo._afterlock(persistmanifestcache)
|
|
|
|
|
|
@property
|
|
|
def fulltextcache(self):
|
|
|
return self._fulltextcache
|
|
|
|
|
|
def clearcaches(self, clear_persisted_data: bool = False) -> None:
|
|
|
self._revlog.clearcaches()
|
|
|
self._fulltextcache.clear(clear_persisted_data=clear_persisted_data)
|
|
|
self._dirlogcache = {self.tree: self}
|
|
|
|
|
|
def dirlog(self, d):
|
|
|
if d:
|
|
|
assert self._treeondisk
|
|
|
if d not in self._dirlogcache:
|
|
|
mfrevlog = manifestrevlog(
|
|
|
self.nodeconstants,
|
|
|
self.opener,
|
|
|
d,
|
|
|
self._dirlogcache,
|
|
|
treemanifest=self._treeondisk,
|
|
|
)
|
|
|
self._dirlogcache[d] = mfrevlog
|
|
|
return self._dirlogcache[d]
|
|
|
|
|
|
def add(
|
|
|
self,
|
|
|
m,
|
|
|
transaction,
|
|
|
link,
|
|
|
p1,
|
|
|
p2,
|
|
|
added: Iterable[bytes],
|
|
|
removed: Iterable[bytes],
|
|
|
readtree=None,
|
|
|
match=None,
|
|
|
):
|
|
|
"""add some manifest entry in to the manifest log
|
|
|
|
|
|
input:
|
|
|
|
|
|
m: the manifest dict we want to store
|
|
|
transaction: the open transaction
|
|
|
p1: manifest-node of p1
|
|
|
p2: manifest-node of p2
|
|
|
added: file added/changed compared to parent
|
|
|
removed: file removed compared to parent
|
|
|
|
|
|
tree manifest input:
|
|
|
|
|
|
readtree: a function to read a subtree
|
|
|
match: a filematcher for the subpart of the tree manifest
|
|
|
"""
|
|
|
try:
|
|
|
if p1 not in self.fulltextcache:
|
|
|
raise FastdeltaUnavailable()
|
|
|
# If our first parent is in the manifest cache, we can
|
|
|
# compute a delta here using properties we know about the
|
|
|
# manifest up-front, which may save time later for the
|
|
|
# revlog layer.
|
|
|
|
|
|
_checkforbidden(added)
|
|
|
# combine the changed lists into one sorted iterator
|
|
|
work = heapq.merge(
|
|
|
[(x, False) for x in sorted(added)],
|
|
|
[(x, True) for x in sorted(removed)],
|
|
|
)
|
|
|
|
|
|
arraytext, deltatext = m.fastdelta(self.fulltextcache[p1], work)
|
|
|
cachedelta = self._revlog.rev(p1), deltatext
|
|
|
text = util.buffer(arraytext)
|
|
|
rev = self._revlog.addrevision(
|
|
|
text, transaction, link, p1, p2, cachedelta
|
|
|
)
|
|
|
n = self._revlog.node(rev)
|
|
|
except FastdeltaUnavailable:
|
|
|
# The first parent manifest isn't already loaded or the
|
|
|
# manifest implementation doesn't support fastdelta, so
|
|
|
# we'll just encode a fulltext of the manifest and pass
|
|
|
# that through to the revlog layer, and let it handle the
|
|
|
# delta process.
|
|
|
if self._treeondisk:
|
|
|
assert readtree, b"readtree must be set for treemanifest writes"
|
|
|
assert match, b"match must be specified for treemanifest writes"
|
|
|
m1 = readtree(self.tree, p1)
|
|
|
m2 = readtree(self.tree, p2)
|
|
|
n = self._addtree(
|
|
|
m, transaction, link, m1, m2, readtree, match=match
|
|
|
)
|
|
|
arraytext = None
|
|
|
else:
|
|
|
text = m.text()
|
|
|
rev = self._revlog.addrevision(text, transaction, link, p1, p2)
|
|
|
n = self._revlog.node(rev)
|
|
|
arraytext = bytearray(text)
|
|
|
|
|
|
if arraytext is not None:
|
|
|
self.fulltextcache[n] = arraytext
|
|
|
|
|
|
return n
|
|
|
|
|
|
def _addtree(self, m, transaction, link, m1, m2, readtree, match):
|
|
|
# If the manifest is unchanged compared to one parent,
|
|
|
# don't write a new revision
|
|
|
if self.tree != b'' and (
|
|
|
m.unmodifiedsince(m1) or m.unmodifiedsince(m2)
|
|
|
):
|
|
|
return m.node()
|
|
|
|
|
|
def writesubtree(subm, subp1, subp2, match):
|
|
|
sublog = self.dirlog(subm.dir())
|
|
|
sublog.add(
|
|
|
subm,
|
|
|
transaction,
|
|
|
link,
|
|
|
subp1,
|
|
|
subp2,
|
|
|
None,
|
|
|
None,
|
|
|
readtree=readtree,
|
|
|
match=match,
|
|
|
)
|
|
|
|
|
|
m.writesubtrees(m1, m2, writesubtree, match)
|
|
|
text = m.dirtext()
|
|
|
n = None
|
|
|
if self.tree != b'':
|
|
|
# Double-check whether contents are unchanged to one parent
|
|
|
if text == m1.dirtext():
|
|
|
n = m1.node()
|
|
|
elif text == m2.dirtext():
|
|
|
n = m2.node()
|
|
|
|
|
|
if not n:
|
|
|
rev = self._revlog.addrevision(
|
|
|
text, transaction, link, m1.node(), m2.node()
|
|
|
)
|
|
|
n = self._revlog.node(rev)
|
|
|
|
|
|
# Save nodeid so parent manifest can calculate its nodeid
|
|
|
m.setnode(n)
|
|
|
return n
|
|
|
|
|
|
def __len__(self):
|
|
|
return len(self._revlog)
|
|
|
|
|
|
def __iter__(self):
|
|
|
return self._revlog.__iter__()
|
|
|
|
|
|
def rev(self, node):
|
|
|
return self._revlog.rev(node)
|
|
|
|
|
|
def node(self, rev):
|
|
|
return self._revlog.node(rev)
|
|
|
|
|
|
def lookup(self, value):
|
|
|
return self._revlog.lookup(value)
|
|
|
|
|
|
def parentrevs(self, rev):
|
|
|
return self._revlog.parentrevs(rev)
|
|
|
|
|
|
def parents(self, node):
|
|
|
return self._revlog.parents(node)
|
|
|
|
|
|
def linkrev(self, rev):
|
|
|
return self._revlog.linkrev(rev)
|
|
|
|
|
|
def checksize(self):
|
|
|
return self._revlog.checksize()
|
|
|
|
|
|
def revision(self, node):
|
|
|
return self._revlog.revision(node)
|
|
|
|
|
|
def rawdata(self, node):
|
|
|
return self._revlog.rawdata(node)
|
|
|
|
|
|
def revdiff(self, rev1, rev2):
|
|
|
return self._revlog.revdiff(rev1, rev2)
|
|
|
|
|
|
def cmp(self, node, text):
|
|
|
return self._revlog.cmp(node, text)
|
|
|
|
|
|
def deltaparent(self, rev):
|
|
|
return self._revlog.deltaparent(rev)
|
|
|
|
|
|
def emitrevisions(
|
|
|
self,
|
|
|
nodes,
|
|
|
nodesorder=None,
|
|
|
revisiondata=False,
|
|
|
assumehaveparentrevisions=False,
|
|
|
deltamode=repository.CG_DELTAMODE_STD,
|
|
|
sidedata_helpers=None,
|
|
|
debug_info=None,
|
|
|
):
|
|
|
return self._revlog.emitrevisions(
|
|
|
nodes,
|
|
|
nodesorder=nodesorder,
|
|
|
revisiondata=revisiondata,
|
|
|
assumehaveparentrevisions=assumehaveparentrevisions,
|
|
|
deltamode=deltamode,
|
|
|
sidedata_helpers=sidedata_helpers,
|
|
|
debug_info=debug_info,
|
|
|
)
|
|
|
|
|
|
def addgroup(
|
|
|
self,
|
|
|
deltas,
|
|
|
linkmapper,
|
|
|
transaction,
|
|
|
alwayscache=False,
|
|
|
addrevisioncb=None,
|
|
|
duplicaterevisioncb=None,
|
|
|
debug_info=None,
|
|
|
delta_base_reuse_policy=None,
|
|
|
):
|
|
|
return self._revlog.addgroup(
|
|
|
deltas,
|
|
|
linkmapper,
|
|
|
transaction,
|
|
|
alwayscache=alwayscache,
|
|
|
addrevisioncb=addrevisioncb,
|
|
|
duplicaterevisioncb=duplicaterevisioncb,
|
|
|
debug_info=debug_info,
|
|
|
delta_base_reuse_policy=delta_base_reuse_policy,
|
|
|
)
|
|
|
|
|
|
def rawsize(self, rev):
|
|
|
return self._revlog.rawsize(rev)
|
|
|
|
|
|
def getstrippoint(self, minlink):
|
|
|
return self._revlog.getstrippoint(minlink)
|
|
|
|
|
|
def strip(self, minlink, transaction):
|
|
|
return self._revlog.strip(minlink, transaction)
|
|
|
|
|
|
def files(self):
|
|
|
return self._revlog.files()
|
|
|
|
|
|
def clone(self, tr, destrevlog, **kwargs):
|
|
|
if not isinstance(destrevlog, manifestrevlog):
|
|
|
raise error.ProgrammingError(b'expected manifestrevlog to clone()')
|
|
|
|
|
|
return self._revlog.clone(tr, destrevlog._revlog, **kwargs)
|
|
|
|
|
|
def storageinfo(
|
|
|
self,
|
|
|
exclusivefiles=False,
|
|
|
sharedfiles=False,
|
|
|
revisionscount=False,
|
|
|
trackedsize=False,
|
|
|
storedsize=False,
|
|
|
):
|
|
|
return self._revlog.storageinfo(
|
|
|
exclusivefiles=exclusivefiles,
|
|
|
sharedfiles=sharedfiles,
|
|
|
revisionscount=revisionscount,
|
|
|
trackedsize=trackedsize,
|
|
|
storedsize=storedsize,
|
|
|
)
|
|
|
|
|
|
@property
|
|
|
def opener(self):
|
|
|
return self._revlog.opener
|
|
|
|
|
|
@opener.setter
|
|
|
def opener(self, value):
|
|
|
self._revlog.opener = value
|
|
|
|
|
|
|
|
|
AnyManifestCtx = Union['manifestctx', 'TreeManifestCtx']
|
|
|
# TODO: drop this in favor of repository.imanifestdict
|
|
|
AnyManifestDict = Union[manifestdict, treemanifest]
|
|
|
|
|
|
|
|
|
class manifestlog: # (repository.imanifestlog)
|
|
|
"""A collection class representing the collection of manifest snapshots
|
|
|
referenced by commits in the repository.
|
|
|
|
|
|
In this situation, 'manifest' refers to the abstract concept of a snapshot
|
|
|
of the list of files in the given commit. Consumers of the output of this
|
|
|
class do not care about the implementation details of the actual manifests
|
|
|
they receive (i.e. tree or flat or lazily loaded, etc)."""
|
|
|
|
|
|
def __init__(self, opener, repo, rootstore, narrowmatch):
|
|
|
self.nodeconstants = repo.nodeconstants
|
|
|
usetreemanifest = False
|
|
|
cachesize = 4
|
|
|
|
|
|
opts = getattr(opener, 'options', None)
|
|
|
if opts is not None:
|
|
|
usetreemanifest = opts.get(b'treemanifest', usetreemanifest)
|
|
|
cachesize = opts.get(b'manifestcachesize', cachesize)
|
|
|
|
|
|
self._treemanifests = usetreemanifest
|
|
|
|
|
|
self._rootstore = rootstore
|
|
|
self._rootstore._setupmanifestcachehooks(repo)
|
|
|
self._narrowmatch = narrowmatch
|
|
|
|
|
|
# A cache of the manifestctx or treemanifestctx for each directory
|
|
|
self._dirmancache = {}
|
|
|
self._dirmancache[b''] = util.lrucachedict(cachesize)
|
|
|
|
|
|
self._cachesize = cachesize
|
|
|
|
|
|
def __getitem__(self, node):
|
|
|
"""Retrieves the manifest instance for the given node. Throws a
|
|
|
LookupError if not found.
|
|
|
"""
|
|
|
return self.get(b'', node)
|
|
|
|
|
|
@property
|
|
|
def narrowed(self):
|
|
|
return not (self._narrowmatch is None or self._narrowmatch.always())
|
|
|
|
|
|
def get(
|
|
|
self, tree: bytes, node: bytes, verify: bool = True
|
|
|
) -> AnyManifestCtx:
|
|
|
"""Retrieves the manifest instance for the given node. Throws a
|
|
|
LookupError if not found.
|
|
|
|
|
|
`verify` - if True an exception will be thrown if the node is not in
|
|
|
the revlog
|
|
|
"""
|
|
|
if node in self._dirmancache.get(tree, ()):
|
|
|
return self._dirmancache[tree][node]
|
|
|
|
|
|
if not self._narrowmatch.always():
|
|
|
if not self._narrowmatch.visitdir(tree[:-1]):
|
|
|
return excludeddirmanifestctx(self.nodeconstants, tree, node)
|
|
|
if tree:
|
|
|
if self._rootstore._treeondisk:
|
|
|
if verify:
|
|
|
# Side-effect is LookupError is raised if node doesn't
|
|
|
# exist.
|
|
|
self.getstorage(tree).rev(node)
|
|
|
|
|
|
m = treemanifestctx(self, tree, node)
|
|
|
else:
|
|
|
raise error.Abort(
|
|
|
_(
|
|
|
b"cannot ask for manifest directory '%s' in a flat "
|
|
|
b"manifest"
|
|
|
)
|
|
|
% tree
|
|
|
)
|
|
|
else:
|
|
|
if verify:
|
|
|
# Side-effect is LookupError is raised if node doesn't exist.
|
|
|
self._rootstore.rev(node)
|
|
|
|
|
|
if self._treemanifests:
|
|
|
m = treemanifestctx(self, b'', node)
|
|
|
else:
|
|
|
m = manifestctx(self, node)
|
|
|
|
|
|
if node != self.nodeconstants.nullid:
|
|
|
mancache = self._dirmancache.get(tree)
|
|
|
if not mancache:
|
|
|
mancache = util.lrucachedict(self._cachesize)
|
|
|
self._dirmancache[tree] = mancache
|
|
|
mancache[node] = m
|
|
|
return m
|
|
|
|
|
|
def getstorage(self, tree):
|
|
|
return self._rootstore.dirlog(tree)
|
|
|
|
|
|
def clearcaches(self, clear_persisted_data: bool = False) -> None:
|
|
|
self._dirmancache.clear()
|
|
|
self._rootstore.clearcaches(clear_persisted_data=clear_persisted_data)
|
|
|
|
|
|
def rev(self, node) -> int:
|
|
|
return self._rootstore.rev(node)
|
|
|
|
|
|
def update_caches(self, transaction) -> None:
|
|
|
return self._rootstore._revlog.update_caches(transaction=transaction)
|
|
|
|
|
|
|
|
|
class memmanifestctx: # (repository.imanifestrevisionwritable)
|
|
|
_manifestdict: manifestdict
|
|
|
|
|
|
def __init__(self, manifestlog):
|
|
|
self._manifestlog = manifestlog
|
|
|
self._manifestdict = manifestdict(manifestlog.nodeconstants.nodelen)
|
|
|
|
|
|
def _storage(self) -> manifestrevlog:
|
|
|
return self._manifestlog.getstorage(b'')
|
|
|
|
|
|
def copy(self) -> 'memmanifestctx':
|
|
|
memmf = memmanifestctx(self._manifestlog)
|
|
|
memmf._manifestdict = self.read().copy()
|
|
|
return memmf
|
|
|
|
|
|
def read(self) -> 'manifestdict':
|
|
|
return self._manifestdict
|
|
|
|
|
|
def write(self, transaction, link, p1, p2, added, removed, match=None):
|
|
|
return self._storage().add(
|
|
|
self._manifestdict,
|
|
|
transaction,
|
|
|
link,
|
|
|
p1,
|
|
|
p2,
|
|
|
added,
|
|
|
removed,
|
|
|
match=match,
|
|
|
)
|
|
|
|
|
|
|
|
|
class manifestctx: # (repository.imanifestrevisionstored)
|
|
|
"""A class representing a single revision of a manifest, including its
|
|
|
contents, its parent revs, and its linkrev.
|
|
|
"""
|
|
|
|
|
|
_data: Optional[manifestdict]
|
|
|
|
|
|
def __init__(self, manifestlog, node):
|
|
|
self._manifestlog = manifestlog
|
|
|
self._data = None
|
|
|
|
|
|
self._node = node
|
|
|
|
|
|
# TODO: We eventually want p1, p2, and linkrev exposed on this class,
|
|
|
# but let's add it later when something needs it and we can load it
|
|
|
# lazily.
|
|
|
# self.p1, self.p2 = store.parents(node)
|
|
|
# rev = store.rev(node)
|
|
|
# self.linkrev = store.linkrev(rev)
|
|
|
|
|
|
def _storage(self) -> 'manifestrevlog':
|
|
|
return self._manifestlog.getstorage(b'')
|
|
|
|
|
|
def node(self) -> bytes:
|
|
|
return self._node
|
|
|
|
|
|
def copy(self) -> memmanifestctx:
|
|
|
memmf = memmanifestctx(self._manifestlog)
|
|
|
memmf._manifestdict = self.read().copy()
|
|
|
return memmf
|
|
|
|
|
|
@propertycache
|
|
|
def parents(self) -> Tuple[bytes, bytes]:
|
|
|
return self._storage().parents(self._node)
|
|
|
|
|
|
def read(self) -> 'manifestdict':
|
|
|
if self._data is None:
|
|
|
nc = self._manifestlog.nodeconstants
|
|
|
if self._node == nc.nullid:
|
|
|
self._data = manifestdict(nc.nodelen)
|
|
|
else:
|
|
|
store = self._storage()
|
|
|
if self._node in store.fulltextcache:
|
|
|
text = pycompat.bytestr(store.fulltextcache[self._node])
|
|
|
else:
|
|
|
text = store.revision(self._node)
|
|
|
arraytext = bytearray(text)
|
|
|
store.fulltextcache[self._node] = arraytext
|
|
|
self._data = manifestdict(nc.nodelen, text)
|
|
|
return self._data
|
|
|
|
|
|
def readfast(self, shallow: bool = False) -> 'manifestdict':
|
|
|
"""Calls either readdelta or read, based on which would be less work.
|
|
|
readdelta is called if the delta is against the p1, and therefore can be
|
|
|
read quickly.
|
|
|
|
|
|
If `shallow` is True, nothing changes since this is a flat manifest.
|
|
|
"""
|
|
|
util.nouideprecwarn(
|
|
|
b'"readfast" is deprecated use "read_any_fast_delta" or "read_delta_parents"',
|
|
|
b"6.9",
|
|
|
stacklevel=2,
|
|
|
)
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
deltaparent = store.deltaparent(r)
|
|
|
if deltaparent != nullrev and deltaparent in store.parentrevs(r):
|
|
|
return self.readdelta()
|
|
|
return self.read()
|
|
|
|
|
|
def readdelta(self, shallow: bool = False) -> 'manifestdict':
|
|
|
"""Returns a manifest containing just the entries that are present
|
|
|
in this manifest, but not in its p1 manifest. This is efficient to read
|
|
|
if the revlog delta is already p1.
|
|
|
|
|
|
Changing the value of `shallow` has no effect on flat manifests.
|
|
|
"""
|
|
|
util.nouideprecwarn(
|
|
|
b'"readfast" is deprecated use "read_any_fast_delta" or "read_delta_new_entries"',
|
|
|
b"6.9",
|
|
|
stacklevel=2,
|
|
|
)
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
|
|
|
def read_any_fast_delta(
|
|
|
self,
|
|
|
valid_bases: Optional[Collection[int]] = None,
|
|
|
*,
|
|
|
shallow: bool = False,
|
|
|
) -> Tuple[Optional[int], manifestdict]:
|
|
|
"""see `imanifestrevisionstored` documentation"""
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
deltaparent = store.deltaparent(r)
|
|
|
if valid_bases is None:
|
|
|
# make sure the next check is True
|
|
|
valid_bases = (deltaparent,)
|
|
|
if deltaparent != nullrev and deltaparent in valid_bases:
|
|
|
d = mdiff.patchtext(store.revdiff(deltaparent, r))
|
|
|
return (
|
|
|
deltaparent,
|
|
|
manifestdict(store.nodeconstants.nodelen, d),
|
|
|
)
|
|
|
return (None, self.read())
|
|
|
|
|
|
def read_delta_parents(
|
|
|
self,
|
|
|
*,
|
|
|
shallow: bool = False,
|
|
|
exact: bool = True,
|
|
|
) -> manifestdict:
|
|
|
"""see `interface.imanifestrevisionbase` documentations"""
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
deltaparent = store.deltaparent(r)
|
|
|
parents = [p for p in store.parentrevs(r) if p is not nullrev]
|
|
|
if not exact and deltaparent in parents:
|
|
|
d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
elif not exact or len(parents) == 0:
|
|
|
return self.read()
|
|
|
elif len(parents) == 1:
|
|
|
p = parents[0]
|
|
|
d = mdiff.patchtext(store.revdiff(p, r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
else:
|
|
|
p1, p2 = parents
|
|
|
d1 = mdiff.patchtext(store.revdiff(p1, r))
|
|
|
d2 = mdiff.patchtext(store.revdiff(p2, r))
|
|
|
d1 = manifestdict(store.nodeconstants.nodelen, d1)
|
|
|
d2 = manifestdict(store.nodeconstants.nodelen, d2)
|
|
|
md = manifestdict(store.nodeconstants.nodelen)
|
|
|
for f, new_node, new_flag in d1.iterentries():
|
|
|
if f not in d2:
|
|
|
continue
|
|
|
if new_node is not None:
|
|
|
md.set(f, new_node, new_flag)
|
|
|
return md
|
|
|
|
|
|
def read_delta_new_entries(self, *, shallow=False) -> manifestdict:
|
|
|
"""see `interface.imanifestrevisionbase` documentations"""
|
|
|
# If we are using narrow, returning a delta against an arbitrary
|
|
|
# changeset might return file outside the narrowspec. This can create
|
|
|
# issue when running validation server side with strict security as
|
|
|
# push from low priviledge usage might be seen as adding new revision
|
|
|
# for files they cannot touch. So we are strict if narrow is involved.
|
|
|
if self._manifestlog.narrowed:
|
|
|
return self.read_delta_parents(shallow=shallow, exact=True)
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
|
|
|
def find(self, key: bytes) -> Tuple[bytes, bytes]:
|
|
|
return self.read().find(key)
|
|
|
|
|
|
|
|
|
class memtreemanifestctx: # (repository.imanifestrevisionwritable)
|
|
|
_treemanifest: treemanifest
|
|
|
|
|
|
def __init__(self, manifestlog, dir=b''):
|
|
|
self._manifestlog = manifestlog
|
|
|
self._dir = dir
|
|
|
self._treemanifest = treemanifest(manifestlog.nodeconstants)
|
|
|
|
|
|
def _storage(self) -> manifestrevlog:
|
|
|
return self._manifestlog.getstorage(b'')
|
|
|
|
|
|
def copy(self) -> 'memtreemanifestctx':
|
|
|
memmf = memtreemanifestctx(self._manifestlog, dir=self._dir)
|
|
|
memmf._treemanifest = self._treemanifest.copy()
|
|
|
return memmf
|
|
|
|
|
|
def read(self) -> 'treemanifest':
|
|
|
return self._treemanifest
|
|
|
|
|
|
def write(self, transaction, link, p1, p2, added, removed, match=None):
|
|
|
def readtree(dir, node):
|
|
|
return self._manifestlog.get(dir, node).read()
|
|
|
|
|
|
return self._storage().add(
|
|
|
self._treemanifest,
|
|
|
transaction,
|
|
|
link,
|
|
|
p1,
|
|
|
p2,
|
|
|
added,
|
|
|
removed,
|
|
|
readtree=readtree,
|
|
|
match=match,
|
|
|
)
|
|
|
|
|
|
|
|
|
class TreeManifestCtx:
|
|
|
_data: Optional[treemanifest]
|
|
|
|
|
|
def __init__(self, manifestlog, dir, node):
|
|
|
self._manifestlog = manifestlog
|
|
|
self._dir = dir
|
|
|
self._data = None
|
|
|
|
|
|
self._node = node
|
|
|
|
|
|
# TODO: Load p1/p2/linkrev lazily. They need to be lazily loaded so that
|
|
|
# we can instantiate treemanifestctx objects for directories we don't
|
|
|
# have on disk.
|
|
|
# self.p1, self.p2 = store.parents(node)
|
|
|
# rev = store.rev(node)
|
|
|
# self.linkrev = store.linkrev(rev)
|
|
|
|
|
|
def _storage(self) -> manifestrevlog:
|
|
|
narrowmatch = self._manifestlog._narrowmatch
|
|
|
if not narrowmatch.always():
|
|
|
if not narrowmatch.visitdir(self._dir[:-1]):
|
|
|
return excludedmanifestrevlog(
|
|
|
self._manifestlog.nodeconstants, self._dir
|
|
|
)
|
|
|
return self._manifestlog.getstorage(self._dir)
|
|
|
|
|
|
def read(self) -> 'treemanifest':
|
|
|
if self._data is None:
|
|
|
store = self._storage()
|
|
|
if self._node == self._manifestlog.nodeconstants.nullid:
|
|
|
self._data = treemanifest(self._manifestlog.nodeconstants)
|
|
|
# TODO accessing non-public API
|
|
|
elif store._treeondisk:
|
|
|
m = treemanifest(self._manifestlog.nodeconstants, dir=self._dir)
|
|
|
|
|
|
def gettext():
|
|
|
return store.revision(self._node)
|
|
|
|
|
|
def readsubtree(dir, subm):
|
|
|
# Set verify to False since we need to be able to create
|
|
|
# subtrees for trees that don't exist on disk.
|
|
|
return self._manifestlog.get(dir, subm, verify=False).read()
|
|
|
|
|
|
m.read(gettext, readsubtree)
|
|
|
m.setnode(self._node)
|
|
|
self._data = m
|
|
|
else:
|
|
|
if self._node in store.fulltextcache:
|
|
|
text = pycompat.bytestr(store.fulltextcache[self._node])
|
|
|
else:
|
|
|
text = store.revision(self._node)
|
|
|
arraytext = bytearray(text)
|
|
|
store.fulltextcache[self._node] = arraytext
|
|
|
self._data = treemanifest(
|
|
|
self._manifestlog.nodeconstants, dir=self._dir, text=text
|
|
|
)
|
|
|
|
|
|
return self._data
|
|
|
|
|
|
def node(self) -> bytes:
|
|
|
return self._node
|
|
|
|
|
|
def copy(self) -> 'memtreemanifestctx':
|
|
|
memmf = memtreemanifestctx(self._manifestlog, dir=self._dir)
|
|
|
memmf._treemanifest = self.read().copy()
|
|
|
return memmf
|
|
|
|
|
|
@propertycache
|
|
|
def parents(self) -> Tuple[bytes, bytes]:
|
|
|
return self._storage().parents(self._node)
|
|
|
|
|
|
def readdelta(self, shallow: bool = False) -> AnyManifestDict:
|
|
|
"""see `imanifestrevisionstored` documentation"""
|
|
|
util.nouideprecwarn(
|
|
|
b'"readdelta" is deprecated use "read_any_fast_delta" or "read_delta_new_entries"',
|
|
|
b"6.9",
|
|
|
stacklevel=2,
|
|
|
)
|
|
|
store = self._storage()
|
|
|
if shallow:
|
|
|
r = store.rev(self._node)
|
|
|
d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
else:
|
|
|
# Need to perform a slow delta
|
|
|
r0 = store.deltaparent(store.rev(self._node))
|
|
|
m0 = self._manifestlog.get(self._dir, store.node(r0)).read()
|
|
|
m1 = self.read()
|
|
|
md = treemanifest(self._manifestlog.nodeconstants, dir=self._dir)
|
|
|
for f, ((n0, fl0), (n1, fl1)) in m0.diff(m1).items():
|
|
|
if n1:
|
|
|
md[f] = n1
|
|
|
if fl1:
|
|
|
md.setflag(f, fl1)
|
|
|
return md
|
|
|
|
|
|
def read_any_fast_delta(
|
|
|
self,
|
|
|
valid_bases: Optional[Collection[int]] = None,
|
|
|
*,
|
|
|
shallow: bool = False,
|
|
|
) -> Tuple[Optional[int], AnyManifestDict]:
|
|
|
"""see `imanifestrevisionstored` documentation"""
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
deltaparent = store.deltaparent(r)
|
|
|
|
|
|
if valid_bases is None:
|
|
|
# make sure the next check is True
|
|
|
valid_bases = (deltaparent,)
|
|
|
can_use_delta = deltaparent != nullrev and deltaparent in valid_bases
|
|
|
|
|
|
if shallow:
|
|
|
if can_use_delta:
|
|
|
return (deltaparent, self._read_storage_delta_shallow())
|
|
|
else:
|
|
|
d = store.revision(self._node)
|
|
|
return (None, manifestdict(store.nodeconstants.nodelen, d))
|
|
|
else:
|
|
|
# note: This use "slow_delta" here is cargo culted from the previous
|
|
|
# implementation. I am not sure it make sense since the goal here is to
|
|
|
# be fast, so why are we computing a delta? On the other hand, tree
|
|
|
# manifest delta as fairly "cheap" and allow for skipping whole part of
|
|
|
# the tree that a full read would access. So it might be a good idea.
|
|
|
#
|
|
|
# If we realize we don't need delta here, we should simply use:
|
|
|
#
|
|
|
# return (None, self.read())
|
|
|
if can_use_delta:
|
|
|
return (None, self._read_storage_slow_delta(base=deltaparent))
|
|
|
else:
|
|
|
parents = [
|
|
|
p
|
|
|
for p in store.parentrevs(r)
|
|
|
if p is not nullrev and p in valid_bases
|
|
|
]
|
|
|
if parents:
|
|
|
best_base = max(parents)
|
|
|
else:
|
|
|
best_base = max(valid_bases)
|
|
|
return (None, self._read_storage_slow_delta(base=best_base))
|
|
|
|
|
|
def _read_storage_delta_shallow(self) -> manifestdict:
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
d = mdiff.patchtext(store.revdiff(store.deltaparent(r), r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
|
|
|
def _read_storage_slow_delta(self, base) -> 'treemanifest':
|
|
|
store = self._storage()
|
|
|
if base is None:
|
|
|
base = store.deltaparent(store.rev(self._node))
|
|
|
m0 = self._manifestlog.get(self._dir, store.node(base)).read()
|
|
|
m1 = self.read()
|
|
|
md = treemanifest(self._manifestlog.nodeconstants, dir=self._dir)
|
|
|
for f, ((n0, fl0), (n1, fl1)) in m0.diff(m1).items():
|
|
|
if n1:
|
|
|
md[f] = n1
|
|
|
if fl1:
|
|
|
md.setflag(f, fl1)
|
|
|
return md
|
|
|
|
|
|
def read_delta_parents(
|
|
|
self,
|
|
|
*,
|
|
|
shallow: bool = False,
|
|
|
exact: bool = True,
|
|
|
) -> AnyManifestDict:
|
|
|
"""see `interface.imanifestrevisionbase` documentations"""
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
parents = [p for p in store.parentrevs(r) if p is not nullrev]
|
|
|
if not exact:
|
|
|
return self.read_any_fast_delta(parents, shallow=shallow)[1]
|
|
|
elif len(parents) == 0:
|
|
|
if shallow:
|
|
|
d = store.revision(self._node)
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
else:
|
|
|
return self.read()
|
|
|
elif len(parents) == 1:
|
|
|
p = parents[0]
|
|
|
if shallow:
|
|
|
d = mdiff.patchtext(store.revdiff(p, r))
|
|
|
return manifestdict(store.nodeconstants.nodelen, d)
|
|
|
else:
|
|
|
return self._read_storage_slow_delta(base=p)
|
|
|
else:
|
|
|
p1, p2 = parents
|
|
|
if shallow:
|
|
|
d1 = mdiff.patchtext(store.revdiff(p1, r))
|
|
|
d2 = mdiff.patchtext(store.revdiff(p2, r))
|
|
|
d1 = manifestdict(store.nodeconstants.nodelen, d1)
|
|
|
d2 = manifestdict(store.nodeconstants.nodelen, d2)
|
|
|
md = manifestdict(store.nodeconstants.nodelen)
|
|
|
for f, new_node, new_flag in d1.iterentries():
|
|
|
if f not in d2:
|
|
|
continue
|
|
|
if new_node is not None:
|
|
|
md.set(f, new_node, new_flag)
|
|
|
return md
|
|
|
else:
|
|
|
m1 = self._manifestlog.get(self._dir, store.node(p1)).read()
|
|
|
m2 = self._manifestlog.get(self._dir, store.node(p2)).read()
|
|
|
mc = self.read()
|
|
|
d1 = m1.diff(mc)
|
|
|
d2 = m2.diff(mc)
|
|
|
md = treemanifest(
|
|
|
self._manifestlog.nodeconstants,
|
|
|
dir=self._dir,
|
|
|
)
|
|
|
for f, new_node, new_flag in d1.iterentries():
|
|
|
if f not in d2:
|
|
|
continue
|
|
|
if new_node is not None:
|
|
|
md.set(f, new_node, new_flag)
|
|
|
return md
|
|
|
|
|
|
def read_delta_new_entries(
|
|
|
self, *, shallow: bool = False
|
|
|
) -> AnyManifestDict:
|
|
|
"""see `interface.imanifestrevisionbase` documentations"""
|
|
|
# If we are using narrow, returning a delta against an arbitrary
|
|
|
# changeset might return file outside the narrowspec. This can create
|
|
|
# issue when running validation server side with strict security as
|
|
|
# push from low priviledge usage might be seen as adding new revision
|
|
|
# for files they cannot touch. So we are strict if narrow is involved.
|
|
|
if self._manifestlog.narrowed:
|
|
|
return self.read_delta_parents(shallow=shallow, exact=True)
|
|
|
# delegate to existing another existing method for simplicity
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
bases = (store.deltaparent(r),)
|
|
|
return self.read_any_fast_delta(bases, shallow=shallow)[1]
|
|
|
|
|
|
def readfast(self, shallow=False) -> AnyManifestDict:
|
|
|
"""Calls either readdelta or read, based on which would be less work.
|
|
|
readdelta is called if the delta is against the p1, and therefore can be
|
|
|
read quickly.
|
|
|
|
|
|
If `shallow` is True, it only returns the entries from this manifest,
|
|
|
and not any submanifests.
|
|
|
"""
|
|
|
util.nouideprecwarn(
|
|
|
b'"readdelta" is deprecated use "read_any_fast_delta" or "read_delta_parents"',
|
|
|
b"6.9",
|
|
|
stacklevel=2,
|
|
|
)
|
|
|
store = self._storage()
|
|
|
r = store.rev(self._node)
|
|
|
deltaparent = store.deltaparent(r)
|
|
|
if deltaparent != nullrev and deltaparent in store.parentrevs(r):
|
|
|
return self.readdelta(shallow=shallow)
|
|
|
|
|
|
if shallow:
|
|
|
return manifestdict(
|
|
|
store.nodeconstants.nodelen, store.revision(self._node)
|
|
|
)
|
|
|
else:
|
|
|
return self.read()
|
|
|
|
|
|
def find(self, key: bytes) -> Tuple[bytes, bytes]:
|
|
|
return self.read().find(key)
|
|
|
|
|
|
|
|
|
treemanifestctx = interfaceutil.implementer(repository.imanifestrevisionstored)(
|
|
|
TreeManifestCtx
|
|
|
)
|
|
|
|
|
|
if typing.TYPE_CHECKING:
|
|
|
treemanifestctx = TreeManifestCtx
|
|
|
|
|
|
|
|
|
class excludeddir(treemanifest):
|
|
|
"""Stand-in for a directory that is excluded from the repository.
|
|
|
|
|
|
With narrowing active on a repository that uses treemanifests,
|
|
|
some of the directory revlogs will be excluded from the resulting
|
|
|
clone. This is a huge storage win for clients, but means we need
|
|
|
some sort of pseudo-manifest to surface to internals so we can
|
|
|
detect a merge conflict outside the narrowspec. That's what this
|
|
|
class is: it stands in for a directory whose node is known, but
|
|
|
whose contents are unknown.
|
|
|
"""
|
|
|
|
|
|
_files: Dict[bytes, bytes]
|
|
|
_flags: Dict[bytes, bytes]
|
|
|
|
|
|
def __init__(self, nodeconstants, dir, node):
|
|
|
super(excludeddir, self).__init__(nodeconstants, dir)
|
|
|
self._node = node
|
|
|
# Add an empty file, which will be included by iterators and such,
|
|
|
# appearing as the directory itself (i.e. something like "dir/")
|
|
|
self._files[b''] = node
|
|
|
self._flags[b''] = b't'
|
|
|
|
|
|
# Manifests outside the narrowspec should never be modified, so avoid
|
|
|
# copying. This makes a noticeable difference when there are very many
|
|
|
# directories outside the narrowspec. Also, it makes sense for the copy to
|
|
|
# be of the same type as the original, which would not happen with the
|
|
|
# super type's copy().
|
|
|
def copy(self):
|
|
|
return self
|
|
|
|
|
|
|
|
|
class excludeddirmanifestctx(treemanifestctx):
|
|
|
"""context wrapper for excludeddir - see that docstring for rationale"""
|
|
|
|
|
|
def __init__(self, nodeconstants, dir, node):
|
|
|
self.nodeconstants = nodeconstants
|
|
|
self._dir = dir
|
|
|
self._node = node
|
|
|
|
|
|
def read(self):
|
|
|
return excludeddir(self.nodeconstants, self._dir, self._node)
|
|
|
|
|
|
def readfast(self, shallow=False):
|
|
|
# special version of readfast since we don't have underlying storage
|
|
|
return self.read()
|
|
|
|
|
|
def write(self, *args):
|
|
|
raise error.ProgrammingError(
|
|
|
b'attempt to write manifest from excluded dir %s' % self._dir
|
|
|
)
|
|
|
|
|
|
|
|
|
class excludedmanifestrevlog(manifestrevlog):
|
|
|
"""Stand-in for excluded treemanifest revlogs.
|
|
|
|
|
|
When narrowing is active on a treemanifest repository, we'll have
|
|
|
references to directories we can't see due to the revlog being
|
|
|
skipped. This class exists to conform to the manifestrevlog
|
|
|
interface for those directories and proactively prevent writes to
|
|
|
outside the narrowspec.
|
|
|
"""
|
|
|
|
|
|
def __init__(self, nodeconstants, dir):
|
|
|
self.nodeconstants = nodeconstants
|
|
|
self._dir = dir
|
|
|
|
|
|
def __len__(self):
|
|
|
raise error.ProgrammingError(
|
|
|
b'attempt to get length of excluded dir %s' % self._dir
|
|
|
)
|
|
|
|
|
|
def rev(self, node):
|
|
|
raise error.ProgrammingError(
|
|
|
b'attempt to get rev from excluded dir %s' % self._dir
|
|
|
)
|
|
|
|
|
|
def linkrev(self, node):
|
|
|
raise error.ProgrammingError(
|
|
|
b'attempt to get linkrev from excluded dir %s' % self._dir
|
|
|
)
|
|
|
|
|
|
def node(self, rev):
|
|
|
raise error.ProgrammingError(
|
|
|
b'attempt to get node from excluded dir %s' % self._dir
|
|
|
)
|
|
|
|
|
|
def add(self, *args, **kwargs):
|
|
|
# We should never write entries in dirlogs outside the narrow clone.
|
|
|
# However, the method still gets called from writesubtree() in
|
|
|
# _addtree(), so we need to handle it. We should possibly make that
|
|
|
# avoid calling add() with a clean manifest (_dirty is always False
|
|
|
# in excludeddir instances).
|
|
|
pass
|
|
|
|