##// END OF EJS Templates
run-tests: stuff a `python3.exe` into the test bin directory on Windows...
run-tests: stuff a `python3.exe` into the test bin directory on Windows Windows doesn't have `python3.exe` as part of the python.org distribution, and that broke every script with a shebang after c102b704edb5. Windows itself provides a `python3.exe` app execution alias[1], but it is some sort of reparse point that MSYS is incapable of handling[2]. When run by MSYS, it simply prints $ python3 -V - Cannot open That in turn caused every `hghave` check, and test that invokes shebang scripts directly, to fail. Rather than try to patch up every script call to be invoked with `$PYTHON` (and regress when non Windows developers forget), copying the executable into the test binary directory with the new name just works. Since this directory is prepended to the system PATH value, it also overrides the broken execution alias. (The `_tmpbindir` is used instead of `_bindir` because the latter causes python3.exe to be copied into the repo next to hg.exe when `test-run-tests.t` runs. Something runs with this version of the executable and subsequent runs of `run-tests.py` inside `test-run-tests.t` try to copy over it while it is in use, and fail. This avoids the failures and the clutter.) I didn't conditionalize this on py3 because `python3.exe` needs to be present (for the shebangs) even when running py2 tests. It shouldn't matter to these simple scripts, and I think the intention is to make the test runner use py3 always, even if testing a py2 build. For now, still supporting py2 is helping to clean up the mess that is py3 tests. [1] https://stackoverflow.com/a/57168165 [2] https://stackoverflow.com/questions/59148628/solved-unable-to-run-python-3-7-on-windows-10-permission-denied#comment104524397_59148666 Differential Revision: https://phab.mercurial-scm.org/D9543

File last commit:

r46671:74923cb8 default
r46684:cc0b332a default
Show More
engine.py
500 lines | 15.4 KiB | text/x-python | PythonLexer
# upgrade.py - functions for in place upgrade of Mercurial repository
#
# Copyright (c) 2016-present, Gregory Szorc
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
from __future__ import absolute_import
import stat
from ..i18n import _
from ..pycompat import getattr
from .. import (
changelog,
error,
filelog,
manifest,
metadata,
pycompat,
requirements,
revlog,
scmutil,
util,
vfs as vfsmod,
)
def _revlogfrompath(repo, path):
"""Obtain a revlog from a repo path.
An instance of the appropriate class is returned.
"""
if path == b'00changelog.i':
return changelog.changelog(repo.svfs)
elif path.endswith(b'00manifest.i'):
mandir = path[: -len(b'00manifest.i')]
return manifest.manifestrevlog(repo.svfs, tree=mandir)
else:
# reverse of "/".join(("data", path + ".i"))
return filelog.filelog(repo.svfs, path[5:-2])
def _copyrevlog(tr, destrepo, oldrl, unencodedname):
"""copy all relevant files for `oldrl` into `destrepo` store
Files are copied "as is" without any transformation. The copy is performed
without extra checks. Callers are responsible for making sure the copied
content is compatible with format of the destination repository.
"""
oldrl = getattr(oldrl, '_revlog', oldrl)
newrl = _revlogfrompath(destrepo, unencodedname)
newrl = getattr(newrl, '_revlog', newrl)
oldvfs = oldrl.opener
newvfs = newrl.opener
oldindex = oldvfs.join(oldrl.indexfile)
newindex = newvfs.join(newrl.indexfile)
olddata = oldvfs.join(oldrl.datafile)
newdata = newvfs.join(newrl.datafile)
with newvfs(newrl.indexfile, b'w'):
pass # create all the directories
util.copyfile(oldindex, newindex)
copydata = oldrl.opener.exists(oldrl.datafile)
if copydata:
util.copyfile(olddata, newdata)
if not (
unencodedname.endswith(b'00changelog.i')
or unencodedname.endswith(b'00manifest.i')
):
destrepo.svfs.fncache.add(unencodedname)
if copydata:
destrepo.svfs.fncache.add(unencodedname[:-2] + b'.d')
UPGRADE_CHANGELOG = b"changelog"
UPGRADE_MANIFEST = b"manifest"
UPGRADE_FILELOGS = b"all-filelogs"
UPGRADE_ALL_REVLOGS = frozenset(
[UPGRADE_CHANGELOG, UPGRADE_MANIFEST, UPGRADE_FILELOGS]
)
def getsidedatacompanion(srcrepo, dstrepo):
sidedatacompanion = None
removedreqs = srcrepo.requirements - dstrepo.requirements
addedreqs = dstrepo.requirements - srcrepo.requirements
if requirements.SIDEDATA_REQUIREMENT in removedreqs:
def sidedatacompanion(rl, rev):
rl = getattr(rl, '_revlog', rl)
if rl.flags(rev) & revlog.REVIDX_SIDEDATA:
return True, (), {}, 0, 0
return False, (), {}, 0, 0
elif requirements.COPIESSDC_REQUIREMENT in addedreqs:
sidedatacompanion = metadata.getsidedataadder(srcrepo, dstrepo)
elif requirements.COPIESSDC_REQUIREMENT in removedreqs:
sidedatacompanion = metadata.getsidedataremover(srcrepo, dstrepo)
return sidedatacompanion
def matchrevlog(revlogfilter, entry):
"""check if a revlog is selected for cloning.
In other words, are there any updates which need to be done on revlog
or it can be blindly copied.
The store entry is checked against the passed filter"""
if entry.endswith(b'00changelog.i'):
return UPGRADE_CHANGELOG in revlogfilter
elif entry.endswith(b'00manifest.i'):
return UPGRADE_MANIFEST in revlogfilter
return UPGRADE_FILELOGS in revlogfilter
def _clonerevlogs(
ui,
srcrepo,
dstrepo,
tr,
deltareuse,
forcedeltabothparents,
revlogs=UPGRADE_ALL_REVLOGS,
):
"""Copy revlogs between 2 repos."""
revcount = 0
srcsize = 0
srcrawsize = 0
dstsize = 0
fcount = 0
frevcount = 0
fsrcsize = 0
frawsize = 0
fdstsize = 0
mcount = 0
mrevcount = 0
msrcsize = 0
mrawsize = 0
mdstsize = 0
crevcount = 0
csrcsize = 0
crawsize = 0
cdstsize = 0
alldatafiles = list(srcrepo.store.walk())
# Perform a pass to collect metadata. This validates we can open all
# source files and allows a unified progress bar to be displayed.
for unencoded, encoded, size in alldatafiles:
if unencoded.endswith(b'.d'):
continue
rl = _revlogfrompath(srcrepo, unencoded)
info = rl.storageinfo(
exclusivefiles=True,
revisionscount=True,
trackedsize=True,
storedsize=True,
)
revcount += info[b'revisionscount'] or 0
datasize = info[b'storedsize'] or 0
rawsize = info[b'trackedsize'] or 0
srcsize += datasize
srcrawsize += rawsize
# This is for the separate progress bars.
if isinstance(rl, changelog.changelog):
crevcount += len(rl)
csrcsize += datasize
crawsize += rawsize
elif isinstance(rl, manifest.manifestrevlog):
mcount += 1
mrevcount += len(rl)
msrcsize += datasize
mrawsize += rawsize
elif isinstance(rl, filelog.filelog):
fcount += 1
frevcount += len(rl)
fsrcsize += datasize
frawsize += rawsize
else:
error.ProgrammingError(b'unknown revlog type')
if not revcount:
return
ui.status(
_(
b'migrating %d total revisions (%d in filelogs, %d in manifests, '
b'%d in changelog)\n'
)
% (revcount, frevcount, mrevcount, crevcount)
)
ui.status(
_(b'migrating %s in store; %s tracked data\n')
% ((util.bytecount(srcsize), util.bytecount(srcrawsize)))
)
# Used to keep track of progress.
progress = None
def oncopiedrevision(rl, rev, node):
progress.increment()
sidedatacompanion = getsidedatacompanion(srcrepo, dstrepo)
# Do the actual copying.
# FUTURE this operation can be farmed off to worker processes.
seen = set()
for unencoded, encoded, size in alldatafiles:
if unencoded.endswith(b'.d'):
continue
oldrl = _revlogfrompath(srcrepo, unencoded)
if isinstance(oldrl, changelog.changelog) and b'c' not in seen:
ui.status(
_(
b'finished migrating %d manifest revisions across %d '
b'manifests; change in size: %s\n'
)
% (mrevcount, mcount, util.bytecount(mdstsize - msrcsize))
)
ui.status(
_(
b'migrating changelog containing %d revisions '
b'(%s in store; %s tracked data)\n'
)
% (
crevcount,
util.bytecount(csrcsize),
util.bytecount(crawsize),
)
)
seen.add(b'c')
progress = srcrepo.ui.makeprogress(
_(b'changelog revisions'), total=crevcount
)
elif isinstance(oldrl, manifest.manifestrevlog) and b'm' not in seen:
ui.status(
_(
b'finished migrating %d filelog revisions across %d '
b'filelogs; change in size: %s\n'
)
% (frevcount, fcount, util.bytecount(fdstsize - fsrcsize))
)
ui.status(
_(
b'migrating %d manifests containing %d revisions '
b'(%s in store; %s tracked data)\n'
)
% (
mcount,
mrevcount,
util.bytecount(msrcsize),
util.bytecount(mrawsize),
)
)
seen.add(b'm')
if progress:
progress.complete()
progress = srcrepo.ui.makeprogress(
_(b'manifest revisions'), total=mrevcount
)
elif b'f' not in seen:
ui.status(
_(
b'migrating %d filelogs containing %d revisions '
b'(%s in store; %s tracked data)\n'
)
% (
fcount,
frevcount,
util.bytecount(fsrcsize),
util.bytecount(frawsize),
)
)
seen.add(b'f')
if progress:
progress.complete()
progress = srcrepo.ui.makeprogress(
_(b'file revisions'), total=frevcount
)
if matchrevlog(revlogs, unencoded):
ui.note(
_(b'cloning %d revisions from %s\n') % (len(oldrl), unencoded)
)
newrl = _revlogfrompath(dstrepo, unencoded)
oldrl.clone(
tr,
newrl,
addrevisioncb=oncopiedrevision,
deltareuse=deltareuse,
forcedeltabothparents=forcedeltabothparents,
sidedatacompanion=sidedatacompanion,
)
else:
msg = _(b'blindly copying %s containing %i revisions\n')
ui.note(msg % (unencoded, len(oldrl)))
_copyrevlog(tr, dstrepo, oldrl, unencoded)
newrl = _revlogfrompath(dstrepo, unencoded)
info = newrl.storageinfo(storedsize=True)
datasize = info[b'storedsize'] or 0
dstsize += datasize
if isinstance(newrl, changelog.changelog):
cdstsize += datasize
elif isinstance(newrl, manifest.manifestrevlog):
mdstsize += datasize
else:
fdstsize += datasize
progress.complete()
ui.status(
_(
b'finished migrating %d changelog revisions; change in size: '
b'%s\n'
)
% (crevcount, util.bytecount(cdstsize - csrcsize))
)
ui.status(
_(
b'finished migrating %d total revisions; total change in store '
b'size: %s\n'
)
% (revcount, util.bytecount(dstsize - srcsize))
)
def _filterstorefile(srcrepo, dstrepo, requirements, path, mode, st):
"""Determine whether to copy a store file during upgrade.
This function is called when migrating store files from ``srcrepo`` to
``dstrepo`` as part of upgrading a repository.
Args:
srcrepo: repo we are copying from
dstrepo: repo we are copying to
requirements: set of requirements for ``dstrepo``
path: store file being examined
mode: the ``ST_MODE`` file type of ``path``
st: ``stat`` data structure for ``path``
Function should return ``True`` if the file is to be copied.
"""
# Skip revlogs.
if path.endswith((b'.i', b'.d', b'.n', b'.nd')):
return False
# Skip transaction related files.
if path.startswith(b'undo'):
return False
# Only copy regular files.
if mode != stat.S_IFREG:
return False
# Skip other skipped files.
if path in (b'lock', b'fncache'):
return False
return True
def _finishdatamigration(ui, srcrepo, dstrepo, requirements):
"""Hook point for extensions to perform additional actions during upgrade.
This function is called after revlogs and store files have been copied but
before the new store is swapped into the original location.
"""
def upgrade(ui, srcrepo, dstrepo, upgrade_op):
"""Do the low-level work of upgrading a repository.
The upgrade is effectively performed as a copy between a source
repository and a temporary destination repository.
The source repository is unmodified for as long as possible so the
upgrade can abort at any time without causing loss of service for
readers and without corrupting the source repository.
"""
assert srcrepo.currentwlock()
assert dstrepo.currentwlock()
ui.status(
_(
b'(it is safe to interrupt this process any time before '
b'data migration completes)\n'
)
)
if b're-delta-all' in upgrade_op.actions:
deltareuse = revlog.revlog.DELTAREUSENEVER
elif b're-delta-parent' in upgrade_op.actions:
deltareuse = revlog.revlog.DELTAREUSESAMEREVS
elif b're-delta-multibase' in upgrade_op.actions:
deltareuse = revlog.revlog.DELTAREUSESAMEREVS
elif b're-delta-fulladd' in upgrade_op.actions:
deltareuse = revlog.revlog.DELTAREUSEFULLADD
else:
deltareuse = revlog.revlog.DELTAREUSEALWAYS
with dstrepo.transaction(b'upgrade') as tr:
_clonerevlogs(
ui,
srcrepo,
dstrepo,
tr,
deltareuse,
b're-delta-multibase' in upgrade_op.actions,
revlogs=upgrade_op.revlogs_to_process,
)
# Now copy other files in the store directory.
# The sorted() makes execution deterministic.
for p, kind, st in sorted(srcrepo.store.vfs.readdir(b'', stat=True)):
if not _filterstorefile(
srcrepo, dstrepo, upgrade_op.requirements, p, kind, st
):
continue
srcrepo.ui.status(_(b'copying %s\n') % p)
src = srcrepo.store.rawvfs.join(p)
dst = dstrepo.store.rawvfs.join(p)
util.copyfile(src, dst, copystat=True)
_finishdatamigration(ui, srcrepo, dstrepo, requirements)
ui.status(_(b'data fully migrated to temporary repository\n'))
backuppath = pycompat.mkdtemp(prefix=b'upgradebackup.', dir=srcrepo.path)
backupvfs = vfsmod.vfs(backuppath)
# Make a backup of requires file first, as it is the first to be modified.
util.copyfile(srcrepo.vfs.join(b'requires'), backupvfs.join(b'requires'))
# We install an arbitrary requirement that clients must not support
# as a mechanism to lock out new clients during the data swap. This is
# better than allowing a client to continue while the repository is in
# an inconsistent state.
ui.status(
_(
b'marking source repository as being upgraded; clients will be '
b'unable to read from repository\n'
)
)
scmutil.writereporequirements(
srcrepo, srcrepo.requirements | {b'upgradeinprogress'}
)
ui.status(_(b'starting in-place swap of repository data\n'))
ui.status(_(b'replaced files will be backed up at %s\n') % backuppath)
# Now swap in the new store directory. Doing it as a rename should make
# the operation nearly instantaneous and atomic (at least in well-behaved
# environments).
ui.status(_(b'replacing store...\n'))
tstart = util.timer()
util.rename(srcrepo.spath, backupvfs.join(b'store'))
util.rename(dstrepo.spath, srcrepo.spath)
elapsed = util.timer() - tstart
ui.status(
_(
b'store replacement complete; repository was inconsistent for '
b'%0.1fs\n'
)
% elapsed
)
# We first write the requirements file. Any new requirements will lock
# out legacy clients.
ui.status(
_(
b'finalizing requirements file and making repository readable '
b'again\n'
)
)
scmutil.writereporequirements(srcrepo, upgrade_op.requirements)
# The lock file from the old store won't be removed because nothing has a
# reference to its new location. So clean it up manually. Alternatively, we
# could update srcrepo.svfs and other variables to point to the new
# location. This is simpler.
backupvfs.unlink(b'store/lock')
return backuppath