changegroup.py
1062 lines
| 40.3 KiB
| text/x-python
|
PythonLexer
/ mercurial / changegroup.py
Martin Geisler
|
r8226 | # changegroup.py - Mercurial changegroup manipulation functions | ||
# | ||||
# Copyright 2006 Matt Mackall <mpm@selenic.com> | ||||
# | ||||
# This software may be used and distributed according to the terms of the | ||||
Matt Mackall
|
r10263 | # GNU General Public License version 2 or any later version. | ||
Matt Mackall
|
r3877 | |||
Gregory Szorc
|
r25921 | from __future__ import absolute_import | ||
import os | ||||
import struct | ||||
import tempfile | ||||
Pierre-Yves David
|
r20933 | import weakref | ||
Gregory Szorc
|
r25921 | |||
from .i18n import _ | ||||
from .node import ( | ||||
hex, | ||||
nullid, | ||||
nullrev, | ||||
short, | ||||
) | ||||
from . import ( | ||||
branchmap, | ||||
dagutil, | ||||
discovery, | ||||
error, | ||||
mdiff, | ||||
phases, | ||||
util, | ||||
) | ||||
Thomas Arendsen Hein
|
r1981 | |||
Sune Foldager
|
r22390 | _CHANGEGROUPV1_DELTA_HEADER = "20s20s20s20s" | ||
Sune Foldager
|
r23181 | _CHANGEGROUPV2_DELTA_HEADER = "20s20s20s20s20s" | ||
Mike Edgar
|
r27433 | _CHANGEGROUPV3_DELTA_HEADER = ">20s20s20s20s20sH" | ||
Benoit Boissinot
|
r14141 | |||
Mads Kiilerich
|
r13457 | def readexactly(stream, n): | ||
'''read n bytes from stream.read and abort if less was available''' | ||||
s = stream.read(n) | ||||
if len(s) < n: | ||||
Pierre-Yves David
|
r26587 | raise error.Abort(_("stream ended unexpectedly" | ||
Mads Kiilerich
|
r13457 | " (got %d bytes, expected %d)") | ||
% (len(s), n)) | ||||
return s | ||||
def getchunk(stream): | ||||
"""return the next chunk from stream as a string""" | ||||
d = readexactly(stream, 4) | ||||
Thomas Arendsen Hein
|
r1981 | l = struct.unpack(">l", d)[0] | ||
if l <= 4: | ||||
Mads Kiilerich
|
r13458 | if l: | ||
Pierre-Yves David
|
r26587 | raise error.Abort(_("invalid chunk length %d") % l) | ||
Thomas Arendsen Hein
|
r1981 | return "" | ||
Mads Kiilerich
|
r13457 | return readexactly(stream, l - 4) | ||
Thomas Arendsen Hein
|
r1981 | |||
Matt Mackall
|
r5368 | def chunkheader(length): | ||
Greg Ward
|
r9437 | """return a changegroup chunk header (string)""" | ||
Matt Mackall
|
r5368 | return struct.pack(">l", length + 4) | ||
Thomas Arendsen Hein
|
r1981 | |||
def closechunk(): | ||||
Greg Ward
|
r9437 | """return a changegroup chunk header (string) for a zero-length chunk""" | ||
Thomas Arendsen Hein
|
r1981 | return struct.pack(">l", 0) | ||
Eric Sumner
|
r23890 | def combineresults(results): | ||
"""logic to combine 0 or more addchangegroup results into one""" | ||||
changedheads = 0 | ||||
result = 1 | ||||
for ret in results: | ||||
# If any changegroup result is 0, return 0 | ||||
if ret == 0: | ||||
result = 0 | ||||
break | ||||
if ret < -1: | ||||
changedheads += ret + 1 | ||||
elif ret > 1: | ||||
changedheads += ret - 1 | ||||
if changedheads > 0: | ||||
result = 1 + changedheads | ||||
elif changedheads < 0: | ||||
result = -1 + changedheads | ||||
return result | ||||
Pierre-Yves David
|
r26540 | def writechunks(ui, chunks, filename, vfs=None): | ||
"""Write chunks to a file and return its filename. | ||||
Matt Mackall
|
r3659 | |||
Pierre-Yves David
|
r26540 | The stream is assumed to be a bundle file. | ||
Matt Mackall
|
r3659 | Existing files will not be overwritten. | ||
If no filename is specified, a temporary file is created. | ||||
""" | ||||
fh = None | ||||
cleanup = None | ||||
try: | ||||
if filename: | ||||
FUJIWARA Katsunori
|
r20976 | if vfs: | ||
fh = vfs.open(filename, "wb") | ||||
else: | ||||
fh = open(filename, "wb") | ||||
Matt Mackall
|
r3659 | else: | ||
fd, filename = tempfile.mkstemp(prefix="hg-bundle-", suffix=".hg") | ||||
fh = os.fdopen(fd, "wb") | ||||
cleanup = filename | ||||
Pierre-Yves David
|
r26540 | for c in chunks: | ||
fh.write(c) | ||||
Matt Mackall
|
r3659 | cleanup = None | ||
return filename | ||||
finally: | ||||
if fh is not None: | ||||
fh.close() | ||||
if cleanup is not None: | ||||
FUJIWARA Katsunori
|
r20976 | if filename and vfs: | ||
vfs.unlink(cleanup) | ||||
else: | ||||
os.unlink(cleanup) | ||||
Matt Mackall
|
r3660 | |||
Sune Foldager
|
r22390 | class cg1unpacker(object): | ||
Augie Fackler
|
r26708 | """Unpacker for cg1 changegroup streams. | ||
A changegroup unpacker handles the framing of the revision data in | ||||
the wire format. Most consumers will want to use the apply() | ||||
method to add the changes from the changegroup to a repository. | ||||
If you're forwarding a changegroup unmodified to another consumer, | ||||
use getchunks(), which returns an iterator of changegroup | ||||
chunks. This is mostly useful for cases where you need to know the | ||||
data stream has ended by observing the end of the changegroup. | ||||
deltachunk() is useful only if you're applying delta data. Most | ||||
consumers should prefer apply() instead. | ||||
A few other public methods exist. Those are used only for | ||||
bundlerepo and some debug commands - their use is discouraged. | ||||
""" | ||||
Sune Foldager
|
r22390 | deltaheader = _CHANGEGROUPV1_DELTA_HEADER | ||
Benoit Boissinot
|
r14141 | deltaheadersize = struct.calcsize(deltaheader) | ||
Eric Sumner
|
r23896 | version = '01' | ||
Martin von Zweigbergk
|
r27920 | _grouplistcount = 1 # One list of files after the manifests | ||
Matt Mackall
|
r12043 | def __init__(self, fh, alg): | ||
Pierre-Yves David
|
r26267 | if alg == 'UN': | ||
alg = None # get more modern without breaking too much | ||||
Pierre-Yves David
|
r26266 | if not alg in util.decompressors: | ||
Pierre-Yves David
|
r26587 | raise error.Abort(_('unknown stream compression type: %s') | ||
Pierre-Yves David
|
r26266 | % alg) | ||
Pierre-Yves David
|
r26392 | if alg == 'BZ': | ||
alg = '_truncatedBZ' | ||||
Pierre-Yves David
|
r26266 | self._stream = util.decompressors[alg](fh) | ||
Matt Mackall
|
r12044 | self._type = alg | ||
Matt Mackall
|
r12334 | self.callback = None | ||
Augie Fackler
|
r26706 | |||
# These methods (compressed, read, seek, tell) all appear to only | ||||
# be used by bundlerepo, but it's a little hard to tell. | ||||
Matt Mackall
|
r12044 | def compressed(self): | ||
Pierre-Yves David
|
r26267 | return self._type is not None | ||
Matt Mackall
|
r12043 | def read(self, l): | ||
return self._stream.read(l) | ||||
Matt Mackall
|
r12330 | def seek(self, pos): | ||
return self._stream.seek(pos) | ||||
def tell(self): | ||||
Matt Mackall
|
r12332 | return self._stream.tell() | ||
Matt Mackall
|
r12347 | def close(self): | ||
return self._stream.close() | ||||
Matt Mackall
|
r12334 | |||
Augie Fackler
|
r26707 | def _chunklength(self): | ||
Jim Hague
|
r13459 | d = readexactly(self._stream, 4) | ||
Mads Kiilerich
|
r13458 | l = struct.unpack(">l", d)[0] | ||
if l <= 4: | ||||
if l: | ||||
Pierre-Yves David
|
r26587 | raise error.Abort(_("invalid chunk length %d") % l) | ||
Mads Kiilerich
|
r13458 | return 0 | ||
if self.callback: | ||||
Matt Mackall
|
r12334 | self.callback() | ||
Mads Kiilerich
|
r13458 | return l - 4 | ||
Matt Mackall
|
r12334 | |||
Benoit Boissinot
|
r14144 | def changelogheader(self): | ||
"""v10 does not have a changelog header chunk""" | ||||
return {} | ||||
def manifestheader(self): | ||||
"""v10 does not have a manifest header chunk""" | ||||
return {} | ||||
def filelogheader(self): | ||||
"""return the header of the filelogs chunk, v10 only has the filename""" | ||||
Augie Fackler
|
r26707 | l = self._chunklength() | ||
Benoit Boissinot
|
r14144 | if not l: | ||
return {} | ||||
fname = readexactly(self._stream, l) | ||||
Augie Fackler
|
r20675 | return {'filename': fname} | ||
Matt Mackall
|
r12334 | |||
Benoit Boissinot
|
r14141 | def _deltaheader(self, headertuple, prevnode): | ||
node, p1, p2, cs = headertuple | ||||
if prevnode is None: | ||||
deltabase = p1 | ||||
else: | ||||
deltabase = prevnode | ||||
Mike Edgar
|
r27433 | flags = 0 | ||
return node, p1, p2, deltabase, cs, flags | ||||
Benoit Boissinot
|
r14141 | |||
Benoit Boissinot
|
r14144 | def deltachunk(self, prevnode): | ||
Augie Fackler
|
r26707 | l = self._chunklength() | ||
Matt Mackall
|
r12336 | if not l: | ||
return {} | ||||
Benoit Boissinot
|
r14141 | headerdata = readexactly(self._stream, self.deltaheadersize) | ||
header = struct.unpack(self.deltaheader, headerdata) | ||||
delta = readexactly(self._stream, l - self.deltaheadersize) | ||||
Mike Edgar
|
r27433 | node, p1, p2, deltabase, cs, flags = self._deltaheader(header, prevnode) | ||
Augie Fackler
|
r20675 | return {'node': node, 'p1': p1, 'p2': p2, 'cs': cs, | ||
Mike Edgar
|
r27433 | 'deltabase': deltabase, 'delta': delta, 'flags': flags} | ||
Matt Mackall
|
r12336 | |||
Pierre-Yves David
|
r20999 | def getchunks(self): | ||
"""returns all the chunks contains in the bundle | ||||
Used when you need to forward the binary stream to a file or another | ||||
network API. To do so, it parse the changegroup data, otherwise it will | ||||
block in case of sshrepo because it don't know the end of the stream. | ||||
""" | ||||
# an empty chunkgroup is the end of the changegroup | ||||
# a changegroup has at least 2 chunkgroups (changelog and manifest). | ||||
Martin von Zweigbergk
|
r27920 | # after that, changegroup versions 1 and 2 have a series of groups | ||
# with one group per file. changegroup 3 has a series of directory | ||||
# manifests before the files. | ||||
Pierre-Yves David
|
r20999 | count = 0 | ||
Martin von Zweigbergk
|
r27920 | emptycount = 0 | ||
while emptycount < self._grouplistcount: | ||||
Pierre-Yves David
|
r20999 | empty = True | ||
count += 1 | ||||
while True: | ||||
chunk = getchunk(self) | ||||
if not chunk: | ||||
Martin von Zweigbergk
|
r27920 | if empty and count > 2: | ||
emptycount += 1 | ||||
Pierre-Yves David
|
r20999 | break | ||
empty = False | ||||
yield chunkheader(len(chunk)) | ||||
pos = 0 | ||||
while pos < len(chunk): | ||||
next = pos + 2**20 | ||||
yield chunk[pos:next] | ||||
pos = next | ||||
yield closechunk() | ||||
Augie Fackler
|
r26712 | def _unpackmanifests(self, repo, revmap, trp, prog, numchanges): | ||
# We know that we'll never have more manifests than we had | ||||
# changesets. | ||||
self.callback = prog(_('manifests'), numchanges) | ||||
# no need to check for empty manifest group here: | ||||
# if the result of the merge of 1 and 2 is the same in 3 and 4, | ||||
# no new manifest will be created and the manifest group will | ||||
# be empty during the pull | ||||
self.manifestheader() | ||||
repo.manifest.addgroup(self, revmap, trp) | ||||
repo.ui.progress(_('manifests'), None) | ||||
Martin von Zweigbergk
|
r28360 | self.callback = None | ||
Augie Fackler
|
r26712 | |||
Augie Fackler
|
r26695 | def apply(self, repo, srctype, url, emptyok=False, | ||
targetphase=phases.draft, expectedtotal=None): | ||||
"""Add the changegroup returned by source.read() to this repo. | ||||
srctype is a string like 'push', 'pull', or 'unbundle'. url is | ||||
the URL of the repo where this changegroup is coming from. | ||||
Return an integer summarizing the change to this repo: | ||||
- nothing changed or no source: 0 | ||||
- more heads than before: 1+added heads (2..n) | ||||
- fewer heads than before: -1-removed heads (-2..-n) | ||||
- number of heads stays the same: 1 | ||||
""" | ||||
repo = repo.unfiltered() | ||||
def csmap(x): | ||||
repo.ui.debug("add changeset %s\n" % short(x)) | ||||
return len(cl) | ||||
def revmap(x): | ||||
return cl.rev(x) | ||||
changesets = files = revisions = 0 | ||||
Pierre-Yves David
|
r26880 | try: | ||
Bryan O'Sullivan
|
r27867 | with repo.transaction("\n".join([srctype, | ||
util.hidepassword(url)])) as tr: | ||||
# The transaction could have been created before and already | ||||
# carries source information. In this case we use the top | ||||
# level data. We overwrite the argument because we need to use | ||||
# the top level value (if they exist) in this function. | ||||
srctype = tr.hookargs.setdefault('source', srctype) | ||||
url = tr.hookargs.setdefault('url', url) | ||||
repo.hook('prechangegroup', throw=True, **tr.hookargs) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | # write changelog data to temp files so concurrent readers | ||
# will not see an inconsistent view | ||||
cl = repo.changelog | ||||
cl.delayupdate(tr) | ||||
oldheads = cl.heads() | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | trp = weakref.proxy(tr) | ||
# pull off the changeset group | ||||
repo.ui.status(_("adding changesets\n")) | ||||
clstart = len(cl) | ||||
class prog(object): | ||||
def __init__(self, step, total): | ||||
self._step = step | ||||
self._total = total | ||||
self._count = 1 | ||||
def __call__(self): | ||||
repo.ui.progress(self._step, self._count, | ||||
unit=_('chunks'), total=self._total) | ||||
self._count += 1 | ||||
self.callback = prog(_('changesets'), expectedtotal) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | efiles = set() | ||
def onchangelog(cl, node): | ||||
Gregory Szorc
|
r28281 | efiles.update(cl.readfiles(node)) | ||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | self.changelogheader() | ||
srccontent = cl.addgroup(self, csmap, trp, | ||||
addrevisioncb=onchangelog) | ||||
efiles = len(efiles) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | if not (srccontent or emptyok): | ||
raise error.Abort(_("received changelog group is empty")) | ||||
clend = len(cl) | ||||
changesets = clend - clstart | ||||
repo.ui.progress(_('changesets'), None) | ||||
Martin von Zweigbergk
|
r28363 | self.callback = None | ||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | # pull off the manifest group | ||
repo.ui.status(_("adding manifests\n")) | ||||
self._unpackmanifests(repo, revmap, trp, prog, changesets) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | needfiles = {} | ||
if repo.ui.configbool('server', 'validate', default=False): | ||||
# validate incoming csets have their manifests | ||||
for cset in xrange(clstart, clend): | ||||
mfnode = repo.changelog.read( | ||||
repo.changelog.node(cset))[0] | ||||
mfest = repo.manifest.readdelta(mfnode) | ||||
# store file nodes we must see | ||||
for f, n in mfest.iteritems(): | ||||
needfiles.setdefault(f, set()).add(n) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | # process the files | ||
repo.ui.status(_("adding file changes\n")) | ||||
newrevs, newfiles = _addchangegroupfiles( | ||||
Martin von Zweigbergk
|
r28361 | repo, self, revmap, trp, efiles, needfiles) | ||
Bryan O'Sullivan
|
r27867 | revisions += newrevs | ||
files += newfiles | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | dh = 0 | ||
if oldheads: | ||||
heads = cl.heads() | ||||
dh = len(heads) - len(oldheads) | ||||
for h in heads: | ||||
if h not in oldheads and repo[h].closesbranch(): | ||||
dh -= 1 | ||||
htext = "" | ||||
if dh: | ||||
htext = _(" (%+d heads)") % dh | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | repo.ui.status(_("added %d changesets" | ||
" with %d changes to %d files%s\n") | ||||
% (changesets, revisions, files, htext)) | ||||
repo.invalidatevolatilesets() | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | if changesets > 0: | ||
if 'node' not in tr.hookargs: | ||||
tr.hookargs['node'] = hex(cl.node(clstart)) | ||||
tr.hookargs['node_last'] = hex(cl.node(clend - 1)) | ||||
hookargs = dict(tr.hookargs) | ||||
else: | ||||
hookargs = dict(tr.hookargs) | ||||
hookargs['node'] = hex(cl.node(clstart)) | ||||
hookargs['node_last'] = hex(cl.node(clend - 1)) | ||||
repo.hook('pretxnchangegroup', throw=True, **hookargs) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | added = [cl.node(r) for r in xrange(clstart, clend)] | ||
publishing = repo.publishing() | ||||
if srctype in ('push', 'serve'): | ||||
# Old servers can not push the boundary themselves. | ||||
# New servers won't push the boundary if changeset already | ||||
# exists locally as secret | ||||
# | ||||
# We should not use added here but the list of all change in | ||||
# the bundle | ||||
if publishing: | ||||
phases.advanceboundary(repo, tr, phases.public, | ||||
srccontent) | ||||
else: | ||||
# Those changesets have been pushed from the | ||||
# outside, their phases are going to be pushed | ||||
# alongside. Therefor `targetphase` is | ||||
# ignored. | ||||
phases.advanceboundary(repo, tr, phases.draft, | ||||
srccontent) | ||||
phases.retractboundary(repo, tr, phases.draft, added) | ||||
elif srctype != 'strip': | ||||
# publishing only alter behavior during push | ||||
# | ||||
# strip should not touch boundary at all | ||||
phases.retractboundary(repo, tr, targetphase, added) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | if changesets > 0: | ||
if srctype != 'strip': | ||||
# During strip, branchcache is invalid but | ||||
# coming call to `destroyed` will repair it. | ||||
# In other case we can safely update cache on | ||||
# disk. | ||||
branchmap.updatecache(repo.filtered('served')) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | def runhooks(): | ||
# These hooks run when the lock releases, not when the | ||||
# transaction closes. So it's possible for the changelog | ||||
# to have changed since we last saw it. | ||||
if clstart >= len(repo): | ||||
return | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | # forcefully update the on-disk branch cache | ||
repo.ui.debug("updating the branch cache\n") | ||||
repo.hook("changegroup", **hookargs) | ||||
for n in added: | ||||
args = hookargs.copy() | ||||
args['node'] = hex(n) | ||||
del args['node_last'] | ||||
repo.hook("incoming", **args) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | newheads = [h for h in repo.heads() | ||
if h not in oldheads] | ||||
repo.ui.log("incoming", | ||||
"%s incoming changes - new heads: %s\n", | ||||
len(added), | ||||
', '.join([hex(c[:6]) for c in newheads])) | ||||
Augie Fackler
|
r26695 | |||
Bryan O'Sullivan
|
r27867 | tr.addpostclose('changegroup-runhooks-%020i' % clstart, | ||
lambda tr: repo._afterlock(runhooks)) | ||||
Augie Fackler
|
r26695 | finally: | ||
repo.ui.flush() | ||||
# never return 0 here: | ||||
if dh < 0: | ||||
return dh - 1 | ||||
else: | ||||
return dh + 1 | ||||
Sune Foldager
|
r23181 | class cg2unpacker(cg1unpacker): | ||
Augie Fackler
|
r26708 | """Unpacker for cg2 streams. | ||
cg2 streams add support for generaldelta, so the delta header | ||||
format is slightly different. All other features about the data | ||||
remain the same. | ||||
""" | ||||
Sune Foldager
|
r23181 | deltaheader = _CHANGEGROUPV2_DELTA_HEADER | ||
deltaheadersize = struct.calcsize(deltaheader) | ||||
Eric Sumner
|
r23896 | version = '02' | ||
Sune Foldager
|
r23181 | |||
def _deltaheader(self, headertuple, prevnode): | ||||
node, p1, p2, deltabase, cs = headertuple | ||||
Mike Edgar
|
r27433 | flags = 0 | ||
return node, p1, p2, deltabase, cs, flags | ||||
Sune Foldager
|
r23181 | |||
Augie Fackler
|
r27432 | class cg3unpacker(cg2unpacker): | ||
"""Unpacker for cg3 streams. | ||||
Mike Edgar
|
r27433 | cg3 streams add support for exchanging treemanifests and revlog | ||
Martin von Zweigbergk
|
r27753 | flags. It adds the revlog flags to the delta header and an empty chunk | ||
separating manifests and files. | ||||
Augie Fackler
|
r27432 | """ | ||
Mike Edgar
|
r27433 | deltaheader = _CHANGEGROUPV3_DELTA_HEADER | ||
deltaheadersize = struct.calcsize(deltaheader) | ||||
Augie Fackler
|
r27432 | version = '03' | ||
Martin von Zweigbergk
|
r27920 | _grouplistcount = 2 # One list of manifests and one list of files | ||
Augie Fackler
|
r27432 | |||
Mike Edgar
|
r27433 | def _deltaheader(self, headertuple, prevnode): | ||
node, p1, p2, deltabase, cs, flags = headertuple | ||||
return node, p1, p2, deltabase, cs, flags | ||||
Martin von Zweigbergk
|
r27754 | def _unpackmanifests(self, repo, revmap, trp, prog, numchanges): | ||
super(cg3unpacker, self)._unpackmanifests(repo, revmap, trp, prog, | ||||
numchanges) | ||||
while True: | ||||
chunkdata = self.filelogheader() | ||||
if not chunkdata: | ||||
break | ||||
# If we get here, there are directory manifests in the changegroup | ||||
d = chunkdata["filename"] | ||||
repo.ui.debug("adding %s revisions\n" % d) | ||||
dirlog = repo.manifest.dirlog(d) | ||||
if not dirlog.addgroup(self, revmap, trp): | ||||
raise error.Abort(_("received dir revlog group is empty")) | ||||
Matt Mackall
|
r12329 | class headerlessfixup(object): | ||
def __init__(self, fh, h): | ||||
self._h = h | ||||
self._fh = fh | ||||
def read(self, n): | ||||
if self._h: | ||||
d, self._h = self._h[:n], self._h[n:] | ||||
if len(d) < n: | ||||
Mads Kiilerich
|
r13457 | d += readexactly(self._fh, n - len(d)) | ||
Matt Mackall
|
r12329 | return d | ||
Mads Kiilerich
|
r13457 | return readexactly(self._fh, n) | ||
Matt Mackall
|
r12329 | |||
Sune Foldager
|
r22390 | class cg1packer(object): | ||
deltaheader = _CHANGEGROUPV1_DELTA_HEADER | ||||
Eric Sumner
|
r23896 | version = '01' | ||
Sune Foldager
|
r19202 | def __init__(self, repo, bundlecaps=None): | ||
"""Given a source repo, construct a bundler. | ||||
bundlecaps is optional and can be used to specify the set of | ||||
capabilities which can be used to build the bundle. | ||||
""" | ||||
Benoit Boissinot
|
r19201 | # Set of capabilities we can use to build the bundle. | ||
if bundlecaps is None: | ||||
bundlecaps = set() | ||||
self._bundlecaps = bundlecaps | ||||
Matt Mackall
|
r25831 | # experimental config: bundle.reorder | ||
Sune Foldager
|
r19202 | reorder = repo.ui.config('bundle', 'reorder', 'auto') | ||
if reorder == 'auto': | ||||
reorder = None | ||||
else: | ||||
reorder = util.parsebool(reorder) | ||||
self._repo = repo | ||||
self._reorder = reorder | ||||
Benoit Boissinot
|
r19208 | self._progress = repo.ui.progress | ||
Mads Kiilerich
|
r23748 | if self._repo.ui.verbose and not self._repo.ui.debugflag: | ||
self._verbosenote = self._repo.ui.note | ||||
else: | ||||
self._verbosenote = lambda s: None | ||||
Matt Mackall
|
r13831 | def close(self): | ||
return closechunk() | ||||
Sune Foldager
|
r19200 | |||
Matt Mackall
|
r13831 | def fileheader(self, fname): | ||
return chunkheader(len(fname)) + fname | ||||
Sune Foldager
|
r19200 | |||
Augie Fackler
|
r29236 | # Extracted both for clarity and for overriding in extensions. | ||
def _sortgroup(self, revlog, nodelist, lookup): | ||||
"""Sort nodes for change group and turn them into revnums.""" | ||||
# for generaldelta revlogs, we linearize the revs; this will both be | ||||
# much quicker and generate a much smaller bundle | ||||
if (revlog._generaldelta and self._reorder is None) or self._reorder: | ||||
dag = dagutil.revlogdag(revlog) | ||||
return dag.linearize(set(revlog.rev(n) for n in nodelist)) | ||||
else: | ||||
return sorted([revlog.rev(n) for n in nodelist]) | ||||
Martin von Zweigbergk
|
r24912 | def group(self, nodelist, revlog, lookup, units=None): | ||
Sune Foldager
|
r19200 | """Calculate a delta group, yielding a sequence of changegroup chunks | ||
(strings). | ||||
Given a list of changeset revs, return a set of deltas and | ||||
metadata corresponding to nodes. The first delta is | ||||
first parent(nodelist[0]) -> nodelist[0], the receiver is | ||||
guaranteed to have this parent as it has all history before | ||||
these changesets. In the case firstparent is nullrev the | ||||
changegroup starts with a full revision. | ||||
Benoit Boissinot
|
r19208 | |||
If units is not None, progress detail will be generated, units specifies | ||||
the type of revlog that is touched (changelog, manifest, etc.). | ||||
Sune Foldager
|
r19200 | """ | ||
# if we don't have any revisions touched by these changesets, bail | ||||
if len(nodelist) == 0: | ||||
yield self.close() | ||||
return | ||||
Augie Fackler
|
r29236 | revs = self._sortgroup(revlog, nodelist, lookup) | ||
Sune Foldager
|
r19200 | |||
# add the parent of the first rev | ||||
p = revlog.parentrevs(revs[0])[0] | ||||
revs.insert(0, p) | ||||
# build deltas | ||||
Benoit Boissinot
|
r19208 | total = len(revs) - 1 | ||
msgbundling = _('bundling') | ||||
Sune Foldager
|
r19200 | for r in xrange(len(revs) - 1): | ||
Benoit Boissinot
|
r19208 | if units is not None: | ||
self._progress(msgbundling, r + 1, unit=units, total=total) | ||||
Sune Foldager
|
r19200 | prev, curr = revs[r], revs[r + 1] | ||
Benoit Boissinot
|
r19207 | linknode = lookup(revlog.node(curr)) | ||
for c in self.revchunk(revlog, curr, prev, linknode): | ||||
Sune Foldager
|
r19200 | yield c | ||
Martin von Zweigbergk
|
r24901 | if units is not None: | ||
self._progress(msgbundling, None) | ||||
Sune Foldager
|
r19200 | yield self.close() | ||
Durham Goode
|
r19289 | # filter any nodes that claim to be part of the known set | ||
Martin von Zweigbergk
|
r24896 | def prune(self, revlog, missing, commonrevs): | ||
Durham Goode
|
r19289 | rr, rl = revlog.rev, revlog.linkrev | ||
return [n for n in missing if rl(rr(n)) not in commonrevs] | ||||
Martin von Zweigbergk
|
r28228 | def _packmanifests(self, dir, mfnodes, lookuplinknode): | ||
Augie Fackler
|
r26711 | """Pack flat manifests into a changegroup stream.""" | ||
Martin von Zweigbergk
|
r28228 | assert not dir | ||
for chunk in self.group(mfnodes, self._repo.manifest, | ||||
lookuplinknode, units=_('manifests')): | ||||
Augie Fackler
|
r26711 | yield chunk | ||
Martin von Zweigbergk
|
r28228 | |||
def _manifestsdone(self): | ||||
return '' | ||||
Augie Fackler
|
r26711 | |||
Benoit Boissinot
|
r19204 | def generate(self, commonrevs, clnodes, fastpathlinkrev, source): | ||
Sune Foldager
|
r19202 | '''yield a sequence of changegroup chunks (strings)''' | ||
repo = self._repo | ||||
Martin von Zweigbergk
|
r24978 | cl = repo.changelog | ||
Benoit Boissinot
|
r19204 | |||
Durham Goode
|
r23381 | clrevorder = {} | ||
Benoit Boissinot
|
r19204 | mfs = {} # needed manifests | ||
fnodes = {} # needed file nodes | ||||
Martin von Zweigbergk
|
r28241 | changedfiles = set() | ||
Benoit Boissinot
|
r19204 | |||
Benoit Boissinot
|
r19207 | # Callback for the changelog, used to collect changed files and manifest | ||
# nodes. | ||||
# Returns the linkrev node (identity in the changelog case). | ||||
def lookupcl(x): | ||||
c = cl.read(x) | ||||
Durham Goode
|
r23381 | clrevorder[x] = len(clrevorder) | ||
Augie Fackler
|
r27237 | n = c[0] | ||
Benoit Boissinot
|
r19207 | # record the first changeset introducing this manifest version | ||
Augie Fackler
|
r27237 | mfs.setdefault(n, x) | ||
# Record a complete list of potentially-changed files in | ||||
# this manifest. | ||||
Martin von Zweigbergk
|
r28241 | changedfiles.update(c[3]) | ||
Benoit Boissinot
|
r19207 | return x | ||
Benoit Boissinot
|
r19204 | |||
Mads Kiilerich
|
r23748 | self._verbosenote(_('uncompressed size of bundle content:\n')) | ||
size = 0 | ||||
Martin von Zweigbergk
|
r24912 | for chunk in self.group(clnodes, cl, lookupcl, units=_('changesets')): | ||
Mads Kiilerich
|
r23748 | size += len(chunk) | ||
Gregory Szorc
|
r23224 | yield chunk | ||
Mads Kiilerich
|
r23748 | self._verbosenote(_('%8.i (changelog)\n') % size) | ||
Gregory Szorc
|
r23224 | |||
Martin von Zweigbergk
|
r24977 | # We need to make sure that the linkrev in the changegroup refers to | ||
# the first changeset that introduced the manifest or file revision. | ||||
# The fastpath is usually safer than the slowpath, because the filelogs | ||||
# are walked in revlog order. | ||||
# | ||||
# When taking the slowpath with reorder=None and the manifest revlog | ||||
# uses generaldelta, the manifest may be walked in the "wrong" order. | ||||
# Without 'clrevorder', we would get an incorrect linkrev (see fix in | ||||
# cc0ff93d0c0c). | ||||
# | ||||
# When taking the fastpath, we are only vulnerable to reordering | ||||
# of the changelog itself. The changelog never uses generaldelta, so | ||||
# it is only reordered when reorder=True. To handle this case, we | ||||
# simply take the slowpath, which already has the 'clrevorder' logic. | ||||
# This was also fixed in cc0ff93d0c0c. | ||||
Martin von Zweigbergk
|
r24976 | fastpathlinkrev = fastpathlinkrev and not self._reorder | ||
Augie Fackler
|
r27432 | # Treemanifests don't work correctly with fastpathlinkrev | ||
# either, because we don't discover which directory nodes to | ||||
# send along with files. This could probably be fixed. | ||||
fastpathlinkrev = fastpathlinkrev and ( | ||||
'treemanifest' not in repo.requirements) | ||||
Martin von Zweigbergk
|
r28227 | |||
for chunk in self.generatemanifests(commonrevs, clrevorder, | ||||
Martin von Zweigbergk
|
r28241 | fastpathlinkrev, mfs, fnodes): | ||
Martin von Zweigbergk
|
r28227 | yield chunk | ||
mfs.clear() | ||||
clrevs = set(cl.rev(x) for x in clnodes) | ||||
if not fastpathlinkrev: | ||||
def linknodes(unused, fname): | ||||
return fnodes.get(fname, {}) | ||||
else: | ||||
cln = cl.node | ||||
def linknodes(filerevlog, fname): | ||||
llr = filerevlog.linkrev | ||||
fln = filerevlog.node | ||||
revs = ((r, llr(r)) for r in filerevlog) | ||||
return dict((fln(r), cln(lr)) for r, lr in revs if lr in clrevs) | ||||
for chunk in self.generatefiles(changedfiles, linknodes, commonrevs, | ||||
source): | ||||
yield chunk | ||||
yield self.close() | ||||
if clnodes: | ||||
repo.hook('outgoing', node=hex(clnodes[0]), source=source) | ||||
def generatemanifests(self, commonrevs, clrevorder, fastpathlinkrev, mfs, | ||||
Martin von Zweigbergk
|
r28241 | fnodes): | ||
Martin von Zweigbergk
|
r28227 | repo = self._repo | ||
Martin von Zweigbergk
|
r28240 | dirlog = repo.manifest.dirlog | ||
Martin von Zweigbergk
|
r28232 | tmfnodes = {'': mfs} | ||
Martin von Zweigbergk
|
r28227 | |||
Benoit Boissinot
|
r19207 | # Callback for the manifest, used to collect linkrevs for filelog | ||
# revisions. | ||||
# Returns the linkrev node (collected in lookupcl). | ||||
Martin von Zweigbergk
|
r28231 | def makelookupmflinknode(dir): | ||
if fastpathlinkrev: | ||||
assert not dir | ||||
return mfs.__getitem__ | ||||
Augie Fackler
|
r27239 | def lookupmflinknode(x): | ||
"""Callback for looking up the linknode for manifests. | ||||
Augie Fackler
|
r27219 | |||
Augie Fackler
|
r27239 | Returns the linkrev node for the specified manifest. | ||
Augie Fackler
|
r27219 | |||
Augie Fackler
|
r27239 | SIDE EFFECT: | ||
Augie Fackler
|
r27432 | 1) fclnodes gets populated with the list of relevant | ||
file nodes if we're not using fastpathlinkrev | ||||
2) When treemanifests are in use, collects treemanifest nodes | ||||
to send | ||||
Augie Fackler
|
r27219 | |||
Augie Fackler
|
r27432 | Note that this means manifests must be completely sent to | ||
the client before you can trust the list of files and | ||||
treemanifests to send. | ||||
Augie Fackler
|
r27239 | """ | ||
Martin von Zweigbergk
|
r28240 | clnode = tmfnodes[dir][x] | ||
mdata = dirlog(dir).readshallowfast(x) | ||||
Martin von Zweigbergk
|
r28241 | for p, n, fl in mdata.iterentries(): | ||
if fl == 't': # subdirectory manifest | ||||
subdir = dir + p + '/' | ||||
tmfclnodes = tmfnodes.setdefault(subdir, {}) | ||||
tmfclnode = tmfclnodes.setdefault(n, clnode) | ||||
if clrevorder[clnode] < clrevorder[tmfclnode]: | ||||
tmfclnodes[n] = clnode | ||||
else: | ||||
f = dir + p | ||||
Martin von Zweigbergk
|
r28240 | fclnodes = fnodes.setdefault(f, {}) | ||
fclnode = fclnodes.setdefault(n, clnode) | ||||
if clrevorder[clnode] < clrevorder[fclnode]: | ||||
fclnodes[n] = clnode | ||||
Augie Fackler
|
r27239 | return clnode | ||
Martin von Zweigbergk
|
r28231 | return lookupmflinknode | ||
Sune Foldager
|
r19206 | |||
Martin von Zweigbergk
|
r28228 | size = 0 | ||
Martin von Zweigbergk
|
r28232 | while tmfnodes: | ||
dir = min(tmfnodes) | ||||
nodes = tmfnodes[dir] | ||||
Martin von Zweigbergk
|
r28240 | prunednodes = self.prune(dirlog(dir), nodes, commonrevs) | ||
Martin von Zweigbergk
|
r28231 | for x in self._packmanifests(dir, prunednodes, | ||
makelookupmflinknode(dir)): | ||||
Martin von Zweigbergk
|
r28229 | size += len(x) | ||
Martin von Zweigbergk
|
r28228 | yield x | ||
Martin von Zweigbergk
|
r28232 | del tmfnodes[dir] | ||
Martin von Zweigbergk
|
r28229 | self._verbosenote(_('%8.i (manifests)\n') % size) | ||
Martin von Zweigbergk
|
r28228 | yield self._manifestsdone() | ||
Sune Foldager
|
r19206 | |||
Martin von Zweigbergk
|
r24897 | # The 'source' parameter is useful for extensions | ||
Durham Goode
|
r19334 | def generatefiles(self, changedfiles, linknodes, commonrevs, source): | ||
repo = self._repo | ||||
progress = self._progress | ||||
msgbundling = _('bundling') | ||||
total = len(changedfiles) | ||||
# for progress output | ||||
msgfiles = _('files') | ||||
for i, fname in enumerate(sorted(changedfiles)): | ||||
filerevlog = repo.file(fname) | ||||
if not filerevlog: | ||||
Pierre-Yves David
|
r26587 | raise error.Abort(_("empty or missing revlog for %s") % fname) | ||
Durham Goode
|
r19334 | |||
linkrevnodes = linknodes(filerevlog, fname) | ||||
Benoit Boissinot
|
r19207 | # Lookup for filenodes, we collected the linkrev nodes above in the | ||
# fastpath case and with lookupmf in the slowpath case. | ||||
def lookupfilelog(x): | ||||
return linkrevnodes[x] | ||||
Martin von Zweigbergk
|
r24896 | filenodes = self.prune(filerevlog, linkrevnodes, commonrevs) | ||
Sune Foldager
|
r19206 | if filenodes: | ||
Benoit Boissinot
|
r19208 | progress(msgbundling, i + 1, item=fname, unit=msgfiles, | ||
total=total) | ||||
Mads Kiilerich
|
r23748 | h = self.fileheader(fname) | ||
size = len(h) | ||||
yield h | ||||
Martin von Zweigbergk
|
r24912 | for chunk in self.group(filenodes, filerevlog, lookupfilelog): | ||
Mads Kiilerich
|
r23748 | size += len(chunk) | ||
Sune Foldager
|
r19202 | yield chunk | ||
Mads Kiilerich
|
r23748 | self._verbosenote(_('%8.i %s\n') % (size, fname)) | ||
Martin von Zweigbergk
|
r24901 | progress(msgbundling, None) | ||
Sune Foldager
|
r19200 | |||
Sune Foldager
|
r23181 | def deltaparent(self, revlog, rev, p1, p2, prev): | ||
return prev | ||||
Benoit Boissinot
|
r19207 | def revchunk(self, revlog, rev, prev, linknode): | ||
Benoit Boissinot
|
r14143 | node = revlog.node(rev) | ||
p1, p2 = revlog.parentrevs(rev) | ||||
Sune Foldager
|
r23181 | base = self.deltaparent(revlog, rev, p1, p2, prev) | ||
Benoit Boissinot
|
r14143 | |||
prefix = '' | ||||
Mike Edgar
|
r24190 | if revlog.iscensored(base) or revlog.iscensored(rev): | ||
try: | ||||
delta = revlog.revision(node) | ||||
Gregory Szorc
|
r25660 | except error.CensoredNodeError as e: | ||
Mike Edgar
|
r24190 | delta = e.tombstone | ||
if base == nullrev: | ||||
prefix = mdiff.trivialdiffheader(len(delta)) | ||||
else: | ||||
baselen = revlog.rawsize(base) | ||||
prefix = mdiff.replacediffheader(baselen, len(delta)) | ||||
elif base == nullrev: | ||||
Benoit Boissinot
|
r14143 | delta = revlog.revision(node) | ||
prefix = mdiff.trivialdiffheader(len(delta)) | ||||
else: | ||||
delta = revlog.revdiff(base, rev) | ||||
p1n, p2n = revlog.parents(node) | ||||
basenode = revlog.node(base) | ||||
Mike Edgar
|
r27433 | flags = revlog.flags(rev) | ||
meta = self.builddeltaheader(node, p1n, p2n, basenode, linknode, flags) | ||||
Benoit Boissinot
|
r14143 | meta += prefix | ||
l = len(meta) + len(delta) | ||||
Matt Mackall
|
r13831 | yield chunkheader(l) | ||
yield meta | ||||
Benoit Boissinot
|
r14143 | yield delta | ||
Mike Edgar
|
r27433 | def builddeltaheader(self, node, p1n, p2n, basenode, linknode, flags): | ||
Benoit Boissinot
|
r14143 | # do nothing with basenode, it is implicitly the previous one in HG10 | ||
Mike Edgar
|
r27433 | # do nothing with flags, it is implicitly 0 for cg1 and cg2 | ||
Benoit Boissinot
|
r14143 | return struct.pack(self.deltaheader, node, p1n, p2n, linknode) | ||
Pierre-Yves David
|
r20925 | |||
Sune Foldager
|
r23181 | class cg2packer(cg1packer): | ||
Eric Sumner
|
r23896 | version = '02' | ||
Sune Foldager
|
r23181 | deltaheader = _CHANGEGROUPV2_DELTA_HEADER | ||
Martin von Zweigbergk
|
r24911 | def __init__(self, repo, bundlecaps=None): | ||
super(cg2packer, self).__init__(repo, bundlecaps) | ||||
if self._reorder is None: | ||||
# Since generaldelta is directly supported by cg2, reordering | ||||
# generally doesn't help, so we disable it by default (treating | ||||
# bundle.reorder=auto just like bundle.reorder=False). | ||||
self._reorder = False | ||||
Sune Foldager
|
r23181 | |||
def deltaparent(self, revlog, rev, p1, p2, prev): | ||||
dp = revlog.deltaparent(rev) | ||||
# avoid storing full revisions; pick prev in those cases | ||||
# also pick prev when we can't be sure remote has dp | ||||
if dp == nullrev or (dp != p1 and dp != p2 and dp != prev): | ||||
return prev | ||||
return dp | ||||
Mike Edgar
|
r27433 | def builddeltaheader(self, node, p1n, p2n, basenode, linknode, flags): | ||
# Do nothing with flags, it is implicitly 0 in cg1 and cg2 | ||||
Sune Foldager
|
r23181 | return struct.pack(self.deltaheader, node, p1n, p2n, basenode, linknode) | ||
Augie Fackler
|
r27432 | class cg3packer(cg2packer): | ||
version = '03' | ||||
Mike Edgar
|
r27433 | deltaheader = _CHANGEGROUPV3_DELTA_HEADER | ||
Augie Fackler
|
r27432 | |||
Martin von Zweigbergk
|
r28228 | def _packmanifests(self, dir, mfnodes, lookuplinknode): | ||
if dir: | ||||
yield self.fileheader(dir) | ||||
for chunk in self.group(mfnodes, self._repo.manifest.dirlog(dir), | ||||
lookuplinknode, units=_('manifests')): | ||||
yield chunk | ||||
def _manifestsdone(self): | ||||
return self.close() | ||||
Augie Fackler
|
r27432 | |||
Mike Edgar
|
r27433 | def builddeltaheader(self, node, p1n, p2n, basenode, linknode, flags): | ||
return struct.pack( | ||||
self.deltaheader, node, p1n, p2n, basenode, linknode, flags) | ||||
Augie Fackler
|
r27432 | |||
Martin von Zweigbergk
|
r27751 | _packermap = {'01': (cg1packer, cg1unpacker), | ||
Augie Fackler
|
r26709 | # cg2 adds support for exchanging generaldelta | ||
'02': (cg2packer, cg2unpacker), | ||||
Martin von Zweigbergk
|
r27753 | # cg3 adds support for exchanging revlog flags and treemanifests | ||
Augie Fackler
|
r27432 | '03': (cg3packer, cg3unpacker), | ||
Augie Fackler
|
r26709 | } | ||
Pierre-Yves David
|
r23168 | |||
Martin von Zweigbergk
|
r27953 | def allsupportedversions(ui): | ||
Martin von Zweigbergk
|
r27928 | versions = set(_packermap.keys()) | ||
Martin von Zweigbergk
|
r27953 | versions.discard('03') | ||
if (ui.configbool('experimental', 'changegroup3') or | ||||
ui.configbool('experimental', 'treemanifest')): | ||||
versions.add('03') | ||||
return versions | ||||
# Changegroup versions that can be applied to the repo | ||||
def supportedincomingversions(repo): | ||||
versions = allsupportedversions(repo.ui) | ||||
if 'treemanifest' in repo.requirements: | ||||
versions.add('03') | ||||
return versions | ||||
# Changegroup versions that can be created from the repo | ||||
def supportedoutgoingversions(repo): | ||||
versions = allsupportedversions(repo.ui) | ||||
if 'treemanifest' in repo.requirements: | ||||
Martin von Zweigbergk
|
r27928 | # Versions 01 and 02 support only flat manifests and it's just too | ||
# expensive to convert between the flat manifest and tree manifest on | ||||
# the fly. Since tree manifests are hashed differently, all of history | ||||
# would have to be converted. Instead, we simply don't even pretend to | ||||
# support versions 01 and 02. | ||||
versions.discard('01') | ||||
versions.discard('02') | ||||
Martin von Zweigbergk
|
r27953 | versions.add('03') | ||
Martin von Zweigbergk
|
r27752 | return versions | ||
Martin von Zweigbergk
|
r27751 | |||
Martin von Zweigbergk
|
r27929 | def safeversion(repo): | ||
# Finds the smallest version that it's safe to assume clients of the repo | ||||
Martin von Zweigbergk
|
r27931 | # will support. For example, all hg versions that support generaldelta also | ||
# support changegroup 02. | ||||
Martin von Zweigbergk
|
r27953 | versions = supportedoutgoingversions(repo) | ||
Martin von Zweigbergk
|
r27929 | if 'generaldelta' in repo.requirements: | ||
versions.discard('01') | ||||
assert versions | ||||
return min(versions) | ||||
Martin von Zweigbergk
|
r27751 | def getbundler(version, repo, bundlecaps=None): | ||
Martin von Zweigbergk
|
r27953 | assert version in supportedoutgoingversions(repo) | ||
Martin von Zweigbergk
|
r27751 | return _packermap[version][0](repo, bundlecaps) | ||
def getunbundler(version, fh, alg): | ||||
return _packermap[version][1](fh, alg) | ||||
Pierre-Yves David
|
r20926 | def _changegroupinfo(repo, nodes, source): | ||
if repo.ui.verbose or source == 'bundle': | ||||
repo.ui.status(_("%d changesets found\n") % len(nodes)) | ||||
if repo.ui.debugflag: | ||||
repo.ui.debug("list of changesets:\n") | ||||
for node in nodes: | ||||
repo.ui.debug("%s\n" % hex(node)) | ||||
Sune Foldager
|
r23177 | def getsubsetraw(repo, outgoing, bundler, source, fastpath=False): | ||
Pierre-Yves David
|
r20925 | repo = repo.unfiltered() | ||
commonrevs = outgoing.common | ||||
csets = outgoing.missing | ||||
heads = outgoing.missingheads | ||||
# We go through the fast path if we get told to, or if all (unfiltered | ||||
# heads have been requested (since we then know there all linkrevs will | ||||
# be pulled by the client). | ||||
heads.sort() | ||||
fastpathlinkrev = fastpath or ( | ||||
repo.filtername is None and heads == sorted(repo.heads())) | ||||
repo.hook('preoutgoing', throw=True, source=source) | ||||
Pierre-Yves David
|
r20926 | _changegroupinfo(repo, csets, source) | ||
Sune Foldager
|
r23177 | return bundler.generate(commonrevs, csets, fastpathlinkrev, source) | ||
Pierre-Yves David
|
r26595 | def getsubset(repo, outgoing, bundler, source, fastpath=False): | ||
Sune Foldager
|
r23177 | gengroup = getsubsetraw(repo, outgoing, bundler, source, fastpath) | ||
Martin von Zweigbergk
|
r27751 | return getunbundler(bundler.version, util.chunkbuffer(gengroup), None) | ||
Pierre-Yves David
|
r20927 | |||
Eric Sumner
|
r23897 | def changegroupsubset(repo, roots, heads, source, version='01'): | ||
Pierre-Yves David
|
r20927 | """Compute a changegroup consisting of all the nodes that are | ||
descendants of any of the roots and ancestors of any of the heads. | ||||
Return a chunkbuffer object whose read() method will return | ||||
successive changegroup chunks. | ||||
It is fairly complex as determining which filenodes and which | ||||
manifest nodes need to be included for the changeset to be complete | ||||
is non-trivial. | ||||
Another wrinkle is doing the reverse, figuring out which changeset in | ||||
the changegroup a particular filenode or manifestnode belongs to. | ||||
""" | ||||
cl = repo.changelog | ||||
if not roots: | ||||
roots = [nullid] | ||||
discbases = [] | ||||
for n in roots: | ||||
discbases.extend([p for p in cl.parents(n) if p != nullid]) | ||||
Pierre-Yves David
|
r25677 | # TODO: remove call to nodesbetween. | ||
csets, roots, heads = cl.nodesbetween(roots, heads) | ||||
included = set(csets) | ||||
discbases = [n for n in discbases if n not in included] | ||||
Pierre-Yves David
|
r20927 | outgoing = discovery.outgoing(cl, discbases, heads) | ||
Martin von Zweigbergk
|
r27751 | bundler = getbundler(version, repo) | ||
Pierre-Yves David
|
r26595 | return getsubset(repo, outgoing, bundler, source) | ||
Pierre-Yves David
|
r20927 | |||
Sune Foldager
|
r23178 | def getlocalchangegroupraw(repo, source, outgoing, bundlecaps=None, | ||
version='01'): | ||||
Sune Foldager
|
r23177 | """Like getbundle, but taking a discovery.outgoing as an argument. | ||
This is only implemented for local repos and reuses potentially | ||||
precomputed sets in outgoing. Returns a raw changegroup generator.""" | ||||
if not outgoing.missing: | ||||
return None | ||||
Martin von Zweigbergk
|
r27751 | bundler = getbundler(version, repo, bundlecaps) | ||
Sune Foldager
|
r23177 | return getsubsetraw(repo, outgoing, bundler, source) | ||
Pierre-Yves David
|
r26508 | def getlocalchangegroup(repo, source, outgoing, bundlecaps=None, | ||
version='01'): | ||||
Pierre-Yves David
|
r20928 | """Like getbundle, but taking a discovery.outgoing as an argument. | ||
This is only implemented for local repos and reuses potentially | ||||
precomputed sets in outgoing.""" | ||||
if not outgoing.missing: | ||||
return None | ||||
Martin von Zweigbergk
|
r27751 | bundler = getbundler(version, repo, bundlecaps) | ||
Pierre-Yves David
|
r20928 | return getsubset(repo, outgoing, bundler, source) | ||
Gregory Szorc
|
r25400 | def computeoutgoing(repo, heads, common): | ||
Durham Goode
|
r21260 | """Computes which revs are outgoing given a set of common | ||
and a set of heads. | ||||
This is a separate function so extensions can have access to | ||||
the logic. | ||||
Returns a discovery.outgoing object. | ||||
""" | ||||
cl = repo.changelog | ||||
if common: | ||||
hasnode = cl.hasnode | ||||
common = [n for n in common if hasnode(n)] | ||||
else: | ||||
common = [nullid] | ||||
if not heads: | ||||
heads = cl.heads() | ||||
return discovery.outgoing(cl, common, heads) | ||||
Pierre-Yves David
|
r26509 | def getchangegroup(repo, source, heads=None, common=None, bundlecaps=None, | ||
version='01'): | ||||
Pierre-Yves David
|
r20930 | """Like changegroupsubset, but returns the set difference between the | ||
ancestors of heads and the ancestors common. | ||||
If heads is None, use the local heads. If common is None, use [nullid]. | ||||
The nodes in common might not all be known locally due to the way the | ||||
current discovery protocol works. | ||||
""" | ||||
Gregory Szorc
|
r25400 | outgoing = computeoutgoing(repo, heads, common) | ||
Pierre-Yves David
|
r26509 | return getlocalchangegroup(repo, source, outgoing, bundlecaps=bundlecaps, | ||
version=version) | ||||
Pierre-Yves David
|
r20930 | |||
Pierre-Yves David
|
r20931 | def changegroup(repo, basenodes, source): | ||
# to avoid a race we use changegroupsubset() (issue1320) | ||||
return changegroupsubset(repo, basenodes, repo.heads(), source) | ||||
Martin von Zweigbergk
|
r28361 | def _addchangegroupfiles(repo, source, revmap, trp, expectedfiles, needfiles): | ||
Pierre-Yves David
|
r20932 | revisions = 0 | ||
files = 0 | ||||
while True: | ||||
chunkdata = source.filelogheader() | ||||
if not chunkdata: | ||||
break | ||||
Martin von Zweigbergk
|
r28361 | files += 1 | ||
Pierre-Yves David
|
r20932 | f = chunkdata["filename"] | ||
repo.ui.debug("adding %s revisions\n" % f) | ||||
Martin von Zweigbergk
|
r28361 | repo.ui.progress(_('files'), files, unit=_('files'), | ||
total=expectedfiles) | ||||
Martin von Zweigbergk
|
r27754 | fl = repo.file(f) | ||
Pierre-Yves David
|
r20932 | o = len(fl) | ||
Mike Edgar
|
r24120 | try: | ||
if not fl.addgroup(source, revmap, trp): | ||||
Pierre-Yves David
|
r26587 | raise error.Abort(_("received file revlog group is empty")) | ||
Gregory Szorc
|
r25660 | except error.CensoredBaseError as e: | ||
Pierre-Yves David
|
r26587 | raise error.Abort(_("received delta base is censored: %s") % e) | ||
Martin von Zweigbergk
|
r27754 | revisions += len(fl) - o | ||
Pierre-Yves David
|
r20932 | if f in needfiles: | ||
needs = needfiles[f] | ||||
for new in xrange(o, len(fl)): | ||||
n = fl.node(new) | ||||
if n in needs: | ||||
needs.remove(n) | ||||
else: | ||||
Pierre-Yves David
|
r26587 | raise error.Abort( | ||
Pierre-Yves David
|
r20932 | _("received spurious file revlog entry")) | ||
if not needs: | ||||
del needfiles[f] | ||||
repo.ui.progress(_('files'), None) | ||||
for f, needs in needfiles.iteritems(): | ||||
fl = repo.file(f) | ||||
for n in needs: | ||||
try: | ||||
fl.rev(n) | ||||
except error.LookupError: | ||||
Pierre-Yves David
|
r26587 | raise error.Abort( | ||
Pierre-Yves David
|
r20932 | _('missing file data for %s:%s - run hg verify') % | ||
(f, hex(n))) | ||||
return revisions, files | ||||