##// END OF EJS Templates
transaction: tests we don't overwrite bookmark activation on abort...
transaction: tests we don't overwrite bookmark activation on abort We actually do not! Great.

File last commit:

r51002:cad3a68c default
r51068:bf27727e stable
Show More
rebase.py
2286 lines | 82.0 KiB | text/x-python | PythonLexer
# rebase.py - rebasing feature for mercurial
#
# Copyright 2008 Stefano Tortarolo <stefano.tortarolo at gmail dot com>
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
'''command to move sets of revisions to a different ancestor
This extension lets you rebase changesets in an existing Mercurial
repository.
For more information:
https://mercurial-scm.org/wiki/RebaseExtension
'''
import os
from mercurial.i18n import _
from mercurial.node import (
nullrev,
short,
wdirrev,
)
from mercurial.pycompat import open
from mercurial import (
bookmarks,
cmdutil,
commands,
copies,
destutil,
dirstateguard,
error,
extensions,
logcmdutil,
merge as mergemod,
mergestate as mergestatemod,
mergeutil,
obsolete,
obsutil,
patch,
phases,
pycompat,
registrar,
repair,
revset,
revsetlang,
rewriteutil,
scmutil,
smartset,
state as statemod,
util,
)
# The following constants are used throughout the rebase module. The ordering of
# their values must be maintained.
# Indicates that a revision needs to be rebased
revtodo = -1
revtodostr = b'-1'
# legacy revstates no longer needed in current code
# -2: nullmerge, -3: revignored, -4: revprecursor, -5: revpruned
legacystates = {b'-2', b'-3', b'-4', b'-5'}
cmdtable = {}
command = registrar.command(cmdtable)
configtable = {}
configitem = registrar.configitem(configtable)
configitem(
b'devel',
b'rebase.force-in-memory-merge',
default=False,
)
# Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
# extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
# be specifying the version(s) of Mercurial they are tested with, or
# leave the attribute unspecified.
testedwith = b'ships-with-hg-core'
def _nothingtorebase():
return 1
def _savegraft(ctx, extra):
s = ctx.extra().get(b'source', None)
if s is not None:
extra[b'source'] = s
s = ctx.extra().get(b'intermediate-source', None)
if s is not None:
extra[b'intermediate-source'] = s
def _savebranch(ctx, extra):
extra[b'branch'] = ctx.branch()
def _destrebase(repo, sourceset, destspace=None):
"""small wrapper around destmerge to pass the right extra args
Please wrap destutil.destmerge instead."""
return destutil.destmerge(
repo,
action=b'rebase',
sourceset=sourceset,
onheadcheck=False,
destspace=destspace,
)
revsetpredicate = registrar.revsetpredicate()
@revsetpredicate(b'_destrebase')
def _revsetdestrebase(repo, subset, x):
# ``_rebasedefaultdest()``
# default destination for rebase.
# # XXX: Currently private because I expect the signature to change.
# # XXX: - bailing out in case of ambiguity vs returning all data.
# i18n: "_rebasedefaultdest" is a keyword
sourceset = None
if x is not None:
sourceset = revset.getset(repo, smartset.fullreposet(repo), x)
return subset & smartset.baseset([_destrebase(repo, sourceset)])
@revsetpredicate(b'_destautoorphanrebase')
def _revsetdestautoorphanrebase(repo, subset, x):
# ``_destautoorphanrebase()``
# automatic rebase destination for a single orphan revision.
unfi = repo.unfiltered()
obsoleted = unfi.revs(b'obsolete()')
src = revset.getset(repo, subset, x).first()
# Empty src or already obsoleted - Do not return a destination
if not src or src in obsoleted:
return smartset.baseset()
dests = destutil.orphanpossibledestination(repo, src)
if len(dests) > 1:
raise error.StateError(
_(b"ambiguous automatic rebase: %r could end up on any of %r")
% (src, dests)
)
# We have zero or one destination, so we can just return here.
return smartset.baseset(dests)
def _ctxdesc(ctx):
"""short description for a context"""
return cmdutil.format_changeset_summary(
ctx.repo().ui, ctx, command=b'rebase'
)
class rebaseruntime:
"""This class is a container for rebase runtime state"""
def __init__(self, repo, ui, inmemory=False, dryrun=False, opts=None):
if opts is None:
opts = {}
# prepared: whether we have rebasestate prepared or not. Currently it
# decides whether "self.repo" is unfiltered or not.
# The rebasestate has explicit hash to hash instructions not depending
# on visibility. If rebasestate exists (in-memory or on-disk), use
# unfiltered repo to avoid visibility issues.
# Before knowing rebasestate (i.e. when starting a new rebase (not
# --continue or --abort)), the original repo should be used so
# visibility-dependent revsets are correct.
self.prepared = False
self.resume = False
self._repo = repo
self.ui = ui
self.opts = opts
self.originalwd = None
self.external = nullrev
# Mapping between the old revision id and either what is the new rebased
# revision or what needs to be done with the old revision. The state
# dict will be what contains most of the rebase progress state.
self.state = {}
self.activebookmark = None
self.destmap = {}
self.skipped = set()
self.collapsef = opts.get('collapse', False)
self.collapsemsg = cmdutil.logmessage(ui, pycompat.byteskwargs(opts))
self.date = opts.get('date', None)
e = opts.get('extrafn') # internal, used by e.g. hgsubversion
self.extrafns = [_savegraft]
if e:
self.extrafns = [e]
self.backupf = ui.configbool(b'rewrite', b'backup-bundle')
self.keepf = opts.get('keep', False)
self.keepbranchesf = opts.get('keepbranches', False)
self.skipemptysuccessorf = rewriteutil.skip_empty_successor(
repo.ui, b'rebase'
)
self.obsolete_with_successor_in_destination = {}
self.obsolete_with_successor_in_rebase_set = set()
self.inmemory = inmemory
self.dryrun = dryrun
self.stateobj = statemod.cmdstate(repo, b'rebasestate')
@property
def repo(self):
if self.prepared:
return self._repo.unfiltered()
else:
return self._repo
def storestatus(self, tr=None):
"""Store the current status to allow recovery"""
if tr:
tr.addfilegenerator(
b'rebasestate',
(b'rebasestate',),
self._writestatus,
location=b'plain',
)
else:
with self.repo.vfs(b"rebasestate", b"w") as f:
self._writestatus(f)
def _writestatus(self, f):
repo = self.repo
assert repo.filtername is None
f.write(repo[self.originalwd].hex() + b'\n')
# was "dest". we now write dest per src root below.
f.write(b'\n')
f.write(repo[self.external].hex() + b'\n')
f.write(b'%d\n' % int(self.collapsef))
f.write(b'%d\n' % int(self.keepf))
f.write(b'%d\n' % int(self.keepbranchesf))
f.write(b'%s\n' % (self.activebookmark or b''))
destmap = self.destmap
for d, v in self.state.items():
oldrev = repo[d].hex()
if v >= 0:
newrev = repo[v].hex()
else:
newrev = b"%d" % v
destnode = repo[destmap[d]].hex()
f.write(b"%s:%s:%s\n" % (oldrev, newrev, destnode))
repo.ui.debug(b'rebase status stored\n')
def restorestatus(self):
"""Restore a previously stored status"""
if not self.stateobj.exists():
cmdutil.wrongtooltocontinue(self.repo, _(b'rebase'))
data = self._read()
self.repo.ui.debug(b'rebase status resumed\n')
self.originalwd = data[b'originalwd']
self.destmap = data[b'destmap']
self.state = data[b'state']
self.skipped = data[b'skipped']
self.collapsef = data[b'collapse']
self.keepf = data[b'keep']
self.keepbranchesf = data[b'keepbranches']
self.external = data[b'external']
self.activebookmark = data[b'activebookmark']
def _read(self):
self.prepared = True
repo = self.repo
assert repo.filtername is None
data = {
b'keepbranches': None,
b'collapse': None,
b'activebookmark': None,
b'external': nullrev,
b'keep': None,
b'originalwd': None,
}
legacydest = None
state = {}
destmap = {}
if True:
f = repo.vfs(b"rebasestate")
for i, l in enumerate(f.read().splitlines()):
if i == 0:
data[b'originalwd'] = repo[l].rev()
elif i == 1:
# this line should be empty in newer version. but legacy
# clients may still use it
if l:
legacydest = repo[l].rev()
elif i == 2:
data[b'external'] = repo[l].rev()
elif i == 3:
data[b'collapse'] = bool(int(l))
elif i == 4:
data[b'keep'] = bool(int(l))
elif i == 5:
data[b'keepbranches'] = bool(int(l))
elif i == 6 and not (len(l) == 81 and b':' in l):
# line 6 is a recent addition, so for backwards
# compatibility check that the line doesn't look like the
# oldrev:newrev lines
data[b'activebookmark'] = l
else:
args = l.split(b':')
oldrev = repo[args[0]].rev()
newrev = args[1]
if newrev in legacystates:
continue
if len(args) > 2:
destrev = repo[args[2]].rev()
else:
destrev = legacydest
destmap[oldrev] = destrev
if newrev == revtodostr:
state[oldrev] = revtodo
# Legacy compat special case
else:
state[oldrev] = repo[newrev].rev()
if data[b'keepbranches'] is None:
raise error.Abort(_(b'.hg/rebasestate is incomplete'))
data[b'destmap'] = destmap
data[b'state'] = state
skipped = set()
# recompute the set of skipped revs
if not data[b'collapse']:
seen = set(destmap.values())
for old, new in sorted(state.items()):
if new != revtodo and new in seen:
skipped.add(old)
seen.add(new)
data[b'skipped'] = skipped
repo.ui.debug(
b'computed skipped revs: %s\n'
% (b' '.join(b'%d' % r for r in sorted(skipped)) or b'')
)
return data
def _handleskippingobsolete(self):
"""Compute structures necessary for skipping obsolete revisions"""
if self.keepf:
return
if not self.ui.configbool(b'experimental', b'rebaseskipobsolete'):
return
obsoleteset = {r for r in self.state if self.repo[r].obsolete()}
(
self.obsolete_with_successor_in_destination,
self.obsolete_with_successor_in_rebase_set,
) = _compute_obsolete_sets(self.repo, obsoleteset, self.destmap)
skippedset = set(self.obsolete_with_successor_in_destination)
skippedset.update(self.obsolete_with_successor_in_rebase_set)
_checkobsrebase(self.repo, self.ui, obsoleteset, skippedset)
if obsolete.isenabled(self.repo, obsolete.allowdivergenceopt):
self.obsolete_with_successor_in_rebase_set = set()
else:
for rev in self.repo.revs(
b'descendants(%ld) and not %ld',
self.obsolete_with_successor_in_rebase_set,
self.obsolete_with_successor_in_rebase_set,
):
self.state.pop(rev, None)
self.destmap.pop(rev, None)
def _prepareabortorcontinue(
self, isabort, backup=True, suppwarns=False, dryrun=False, confirm=False
):
self.resume = True
try:
self.restorestatus()
# Calculate self.obsolete_* sets
self._handleskippingobsolete()
self.collapsemsg = restorecollapsemsg(self.repo, isabort)
except error.RepoLookupError:
if isabort:
clearstatus(self.repo)
clearcollapsemsg(self.repo)
self.repo.ui.warn(
_(
b'rebase aborted (no revision is removed,'
b' only broken state is cleared)\n'
)
)
return 0
else:
msg = _(b'cannot continue inconsistent rebase')
hint = _(b'use "hg rebase --abort" to clear broken state')
raise error.Abort(msg, hint=hint)
if isabort:
backup = backup and self.backupf
return self._abort(
backup=backup,
suppwarns=suppwarns,
dryrun=dryrun,
confirm=confirm,
)
def _preparenewrebase(self, destmap):
if not destmap:
return _nothingtorebase()
result = buildstate(self.repo, destmap, self.collapsef)
if not result:
# Empty state built, nothing to rebase
self.ui.status(_(b'nothing to rebase\n'))
return _nothingtorebase()
(self.originalwd, self.destmap, self.state) = result
if self.collapsef:
dests = set(self.destmap.values())
if len(dests) != 1:
raise error.InputError(
_(b'--collapse does not work with multiple destinations')
)
destrev = next(iter(dests))
destancestors = self.repo.changelog.ancestors(
[destrev], inclusive=True
)
self.external = externalparent(self.repo, self.state, destancestors)
for destrev in sorted(set(destmap.values())):
dest = self.repo[destrev]
if dest.closesbranch() and not self.keepbranchesf:
self.ui.status(_(b'reopening closed branch head %s\n') % dest)
# Calculate self.obsolete_* sets
self._handleskippingobsolete()
if not self.keepf:
rebaseset = set(destmap.keys())
rebaseset -= set(self.obsolete_with_successor_in_destination)
rebaseset -= self.obsolete_with_successor_in_rebase_set
# We have our own divergence-checking in the rebase extension
overrides = {}
if obsolete.isenabled(self.repo, obsolete.createmarkersopt):
overrides = {
(b'experimental', b'evolution.allowdivergence'): b'true'
}
try:
with self.ui.configoverride(overrides):
rewriteutil.precheck(self.repo, rebaseset, action=b'rebase')
except error.Abort as e:
if e.hint is None:
e.hint = _(b'use --keep to keep original changesets')
raise e
self.prepared = True
def _assignworkingcopy(self):
if self.inmemory:
from mercurial.context import overlayworkingctx
self.wctx = overlayworkingctx(self.repo)
self.repo.ui.debug(b"rebasing in memory\n")
else:
self.wctx = self.repo[None]
self.repo.ui.debug(b"rebasing on disk\n")
self.repo.ui.log(
b"rebase",
b"using in-memory rebase: %r\n",
self.inmemory,
rebase_imm_used=self.inmemory,
)
def _performrebase(self, tr):
self._assignworkingcopy()
repo, ui = self.repo, self.ui
if self.keepbranchesf:
# insert _savebranch at the start of extrafns so if
# there's a user-provided extrafn it can clobber branch if
# desired
self.extrafns.insert(0, _savebranch)
if self.collapsef:
branches = set()
for rev in self.state:
branches.add(repo[rev].branch())
if len(branches) > 1:
raise error.InputError(
_(b'cannot collapse multiple named branches')
)
# Keep track of the active bookmarks in order to reset them later
self.activebookmark = self.activebookmark or repo._activebookmark
if self.activebookmark:
bookmarks.deactivate(repo)
# Store the state before we begin so users can run 'hg rebase --abort'
# if we fail before the transaction closes.
self.storestatus()
if tr:
# When using single transaction, store state when transaction
# commits.
self.storestatus(tr)
cands = [k for k, v in self.state.items() if v == revtodo]
p = repo.ui.makeprogress(
_(b"rebasing"), unit=_(b'changesets'), total=len(cands)
)
def progress(ctx):
p.increment(item=(b"%d:%s" % (ctx.rev(), ctx)))
for subset in sortsource(self.destmap):
sortedrevs = self.repo.revs(b'sort(%ld, -topo)', subset)
for rev in sortedrevs:
self._rebasenode(tr, rev, progress)
p.complete()
ui.note(_(b'rebase merging completed\n'))
def _concludenode(self, rev, editor, commitmsg=None):
"""Commit the wd changes with parents p1 and p2.
Reuse commit info from rev but also store useful information in extra.
Return node of committed revision."""
repo = self.repo
ctx = repo[rev]
if commitmsg is None:
commitmsg = ctx.description()
# Skip replacement if collapsing, as that degenerates to p1 for all
# nodes.
if not self.collapsef:
cl = repo.changelog
commitmsg = rewriteutil.update_hash_refs(
repo,
commitmsg,
{
cl.node(oldrev): [cl.node(newrev)]
for oldrev, newrev in self.state.items()
if newrev != revtodo
},
)
date = self.date
if date is None:
date = ctx.date()
extra = {}
if repo.ui.configbool(b'rebase', b'store-source'):
extra = {b'rebase_source': ctx.hex()}
for c in self.extrafns:
c(ctx, extra)
destphase = max(ctx.phase(), phases.draft)
overrides = {
(b'phases', b'new-commit'): destphase,
(b'ui', b'allowemptycommit'): not self.skipemptysuccessorf,
}
with repo.ui.configoverride(overrides, b'rebase'):
if self.inmemory:
newnode = commitmemorynode(
repo,
wctx=self.wctx,
extra=extra,
commitmsg=commitmsg,
editor=editor,
user=ctx.user(),
date=date,
)
else:
newnode = commitnode(
repo,
extra=extra,
commitmsg=commitmsg,
editor=editor,
user=ctx.user(),
date=date,
)
return newnode
def _rebasenode(self, tr, rev, progressfn):
repo, ui, opts = self.repo, self.ui, self.opts
ctx = repo[rev]
desc = _ctxdesc(ctx)
if self.state[rev] == rev:
ui.status(_(b'already rebased %s\n') % desc)
elif rev in self.obsolete_with_successor_in_rebase_set:
msg = (
_(
b'note: not rebasing %s and its descendants as '
b'this would cause divergence\n'
)
% desc
)
repo.ui.status(msg)
self.skipped.add(rev)
elif rev in self.obsolete_with_successor_in_destination:
succ = self.obsolete_with_successor_in_destination[rev]
if succ is None:
msg = _(b'note: not rebasing %s, it has no successor\n') % desc
else:
succdesc = _ctxdesc(repo[succ])
msg = _(
b'note: not rebasing %s, already in destination as %s\n'
) % (desc, succdesc)
repo.ui.status(msg)
# Make clearrebased aware state[rev] is not a true successor
self.skipped.add(rev)
# Record rev as moved to its desired destination in self.state.
# This helps bookmark and working parent movement.
dest = max(
adjustdest(repo, rev, self.destmap, self.state, self.skipped)
)
self.state[rev] = dest
elif self.state[rev] == revtodo:
ui.status(_(b'rebasing %s\n') % desc)
progressfn(ctx)
p1, p2, base = defineparents(
repo,
rev,
self.destmap,
self.state,
self.skipped,
self.obsolete_with_successor_in_destination,
)
if self.resume and self.wctx.p1().rev() == p1:
repo.ui.debug(b'resuming interrupted rebase\n')
self.resume = False
else:
overrides = {(b'ui', b'forcemerge'): opts.get('tool', b'')}
with ui.configoverride(overrides, b'rebase'):
try:
rebasenode(
repo,
rev,
p1,
p2,
base,
self.collapsef,
wctx=self.wctx,
)
except error.InMemoryMergeConflictsError:
if self.dryrun:
raise error.ConflictResolutionRequired(b'rebase')
if self.collapsef:
# TODO: Make the overlayworkingctx reflected
# in the working copy here instead of re-raising
# so the entire rebase operation is retried.
raise
ui.status(
_(
b"hit merge conflicts; rebasing that "
b"commit again in the working copy\n"
)
)
try:
cmdutil.bailifchanged(repo)
except error.Abort:
clearstatus(repo)
clearcollapsemsg(repo)
raise
self.inmemory = False
self._assignworkingcopy()
mergemod.update(repo[p1], wc=self.wctx)
rebasenode(
repo,
rev,
p1,
p2,
base,
self.collapsef,
wctx=self.wctx,
)
if not self.collapsef:
merging = p2 != nullrev
editform = cmdutil.mergeeditform(merging, b'rebase')
editor = cmdutil.getcommiteditor(editform=editform, **opts)
# We need to set parents again here just in case we're continuing
# a rebase started with an old hg version (before 9c9cfecd4600),
# because those old versions would have left us with two dirstate
# parents, and we don't want to create a merge commit here (unless
# we're rebasing a merge commit).
self.wctx.setparents(repo[p1].node(), repo[p2].node())
newnode = self._concludenode(rev, editor)
else:
# Skip commit if we are collapsing
newnode = None
# Update the state
if newnode is not None:
self.state[rev] = repo[newnode].rev()
ui.debug(b'rebased as %s\n' % short(newnode))
if repo[newnode].isempty():
ui.warn(
_(
b'note: created empty successor for %s, its '
b'destination already has all its changes\n'
)
% desc
)
else:
if not self.collapsef:
ui.warn(
_(
b'note: not rebasing %s, its destination already '
b'has all its changes\n'
)
% desc
)
self.skipped.add(rev)
self.state[rev] = p1
ui.debug(b'next revision set to %d\n' % p1)
else:
ui.status(
_(b'already rebased %s as %s\n') % (desc, repo[self.state[rev]])
)
if not tr:
# When not using single transaction, store state after each
# commit is completely done. On InterventionRequired, we thus
# won't store the status. Instead, we'll hit the "len(parents) == 2"
# case and realize that the commit was in progress.
self.storestatus()
def _finishrebase(self):
repo, ui, opts = self.repo, self.ui, self.opts
fm = ui.formatter(b'rebase', pycompat.byteskwargs(opts))
fm.startitem()
if self.collapsef:
p1, p2, _base = defineparents(
repo,
min(self.state),
self.destmap,
self.state,
self.skipped,
self.obsolete_with_successor_in_destination,
)
editopt = opts.get('edit')
editform = b'rebase.collapse'
if self.collapsemsg:
commitmsg = self.collapsemsg
else:
commitmsg = b'Collapsed revision'
for rebased in sorted(self.state):
if rebased not in self.skipped:
commitmsg += b'\n* %s' % repo[rebased].description()
editopt = True
editor = cmdutil.getcommiteditor(edit=editopt, editform=editform)
revtoreuse = max(self.state)
self.wctx.setparents(repo[p1].node(), repo[self.external].node())
newnode = self._concludenode(
revtoreuse, editor, commitmsg=commitmsg
)
if newnode is not None:
newrev = repo[newnode].rev()
for oldrev in self.state:
self.state[oldrev] = newrev
if b'qtip' in repo.tags():
updatemq(repo, self.state, self.skipped, **opts)
# restore original working directory
# (we do this before stripping)
newwd = self.state.get(self.originalwd, self.originalwd)
if newwd < 0:
# original directory is a parent of rebase set root or ignored
newwd = self.originalwd
if newwd not in [c.rev() for c in repo[None].parents()]:
ui.note(_(b"update back to initial working directory parent\n"))
mergemod.update(repo[newwd])
collapsedas = None
if self.collapsef and not self.keepf:
collapsedas = newnode
clearrebased(
ui,
repo,
self.destmap,
self.state,
self.skipped,
collapsedas,
self.keepf,
fm=fm,
backup=self.backupf,
)
clearstatus(repo)
clearcollapsemsg(repo)
ui.note(_(b"rebase completed\n"))
util.unlinkpath(repo.sjoin(b'undo'), ignoremissing=True)
if self.skipped:
skippedlen = len(self.skipped)
ui.note(_(b"%d revisions have been skipped\n") % skippedlen)
fm.end()
if (
self.activebookmark
and self.activebookmark in repo._bookmarks
and repo[b'.'].node() == repo._bookmarks[self.activebookmark]
):
bookmarks.activate(repo, self.activebookmark)
def _abort(self, backup=True, suppwarns=False, dryrun=False, confirm=False):
'''Restore the repository to its original state.'''
repo = self.repo
try:
# If the first commits in the rebased set get skipped during the
# rebase, their values within the state mapping will be the dest
# rev id. The rebased list must must not contain the dest rev
# (issue4896)
rebased = [
s
for r, s in self.state.items()
if s >= 0 and s != r and s != self.destmap[r]
]
immutable = [d for d in rebased if not repo[d].mutable()]
cleanup = True
if immutable:
repo.ui.warn(
_(b"warning: can't clean up public changesets %s\n")
% b', '.join(bytes(repo[r]) for r in immutable),
hint=_(b"see 'hg help phases' for details"),
)
cleanup = False
descendants = set()
if rebased:
descendants = set(repo.changelog.descendants(rebased))
if descendants - set(rebased):
repo.ui.warn(
_(
b"warning: new changesets detected on "
b"destination branch, can't strip\n"
)
)
cleanup = False
if cleanup:
if rebased:
strippoints = [
c.node() for c in repo.set(b'roots(%ld)', rebased)
]
updateifonnodes = set(rebased)
updateifonnodes.update(self.destmap.values())
if not dryrun and not confirm:
updateifonnodes.add(self.originalwd)
shouldupdate = repo[b'.'].rev() in updateifonnodes
# Update away from the rebase if necessary
if shouldupdate:
mergemod.clean_update(repo[self.originalwd])
# Strip from the first rebased revision
if rebased:
repair.strip(repo.ui, repo, strippoints, backup=backup)
if self.activebookmark and self.activebookmark in repo._bookmarks:
bookmarks.activate(repo, self.activebookmark)
finally:
clearstatus(repo)
clearcollapsemsg(repo)
if not suppwarns:
repo.ui.warn(_(b'rebase aborted\n'))
return 0
@command(
b'rebase',
[
(
b's',
b'source',
[],
_(b'rebase the specified changesets and their descendants'),
_(b'REV'),
),
(
b'b',
b'base',
[],
_(b'rebase everything from branching point of specified changeset'),
_(b'REV'),
),
(b'r', b'rev', [], _(b'rebase these revisions'), _(b'REV')),
(
b'd',
b'dest',
b'',
_(b'rebase onto the specified changeset'),
_(b'REV'),
),
(b'', b'collapse', False, _(b'collapse the rebased changesets')),
(
b'm',
b'message',
b'',
_(b'use text as collapse commit message'),
_(b'TEXT'),
),
(b'e', b'edit', False, _(b'invoke editor on commit messages')),
(
b'l',
b'logfile',
b'',
_(b'read collapse commit message from file'),
_(b'FILE'),
),
(b'k', b'keep', False, _(b'keep original changesets')),
(b'', b'keepbranches', False, _(b'keep original branch names')),
(b'D', b'detach', False, _(b'(DEPRECATED)')),
(b'i', b'interactive', False, _(b'(DEPRECATED)')),
(b't', b'tool', b'', _(b'specify merge tool')),
(b'', b'stop', False, _(b'stop interrupted rebase')),
(b'c', b'continue', False, _(b'continue an interrupted rebase')),
(b'a', b'abort', False, _(b'abort an interrupted rebase')),
(
b'',
b'auto-orphans',
b'',
_(
b'automatically rebase orphan revisions '
b'in the specified revset (EXPERIMENTAL)'
),
),
]
+ cmdutil.dryrunopts
+ cmdutil.formatteropts
+ cmdutil.confirmopts,
_(b'[[-s REV]... | [-b REV]... | [-r REV]...] [-d REV] [OPTION]...'),
helpcategory=command.CATEGORY_CHANGE_MANAGEMENT,
)
def rebase(ui, repo, **opts):
"""move changeset (and descendants) to a different branch
Rebase uses repeated merging to graft changesets from one part of
history (the source) onto another (the destination). This can be
useful for linearizing *local* changes relative to a master
development tree.
Published commits cannot be rebased (see :hg:`help phases`).
To copy commits, see :hg:`help graft`.
If you don't specify a destination changeset (``-d/--dest``), rebase
will use the same logic as :hg:`merge` to pick a destination. if
the current branch contains exactly one other head, the other head
is merged with by default. Otherwise, an explicit revision with
which to merge with must be provided. (destination changeset is not
modified by rebasing, but new changesets are added as its
descendants.)
Here are the ways to select changesets:
1. Explicitly select them using ``--rev``.
2. Use ``--source`` to select a root changeset and include all of its
descendants.
3. Use ``--base`` to select a changeset; rebase will find ancestors
and their descendants which are not also ancestors of the destination.
4. If you do not specify any of ``--rev``, ``--source``, or ``--base``,
rebase will use ``--base .`` as above.
If ``--source`` or ``--rev`` is used, special names ``SRC`` and ``ALLSRC``
can be used in ``--dest``. Destination would be calculated per source
revision with ``SRC`` substituted by that single source revision and
``ALLSRC`` substituted by all source revisions.
Rebase will destroy original changesets unless you use ``--keep``.
It will also move your bookmarks (even if you do).
Some changesets may be dropped if they do not contribute changes
(e.g. merges from the destination branch).
Unlike ``merge``, rebase will do nothing if you are at the branch tip of
a named branch with two heads. You will need to explicitly specify source
and/or destination.
If you need to use a tool to automate merge/conflict decisions, you
can specify one with ``--tool``, see :hg:`help merge-tools`.
As a caveat: the tool will not be used to mediate when a file was
deleted, there is no hook presently available for this.
If a rebase is interrupted to manually resolve a conflict, it can be
continued with --continue/-c, aborted with --abort/-a, or stopped with
--stop.
.. container:: verbose
Examples:
- move "local changes" (current commit back to branching point)
to the current branch tip after a pull::
hg rebase
- move a single changeset to the stable branch::
hg rebase -r 5f493448 -d stable
- splice a commit and all its descendants onto another part of history::
hg rebase --source c0c3 --dest 4cf9
- rebase everything on a branch marked by a bookmark onto the
default branch::
hg rebase --base myfeature --dest default
- collapse a sequence of changes into a single commit::
hg rebase --collapse -r 1520:1525 -d .
- move a named branch while preserving its name::
hg rebase -r "branch(featureX)" -d 1.3 --keepbranches
- stabilize orphaned changesets so history looks linear::
hg rebase -r 'orphan()-obsolete()'\
-d 'first(max((successors(max(roots(ALLSRC) & ::SRC)^)-obsolete())::) +\
max(::((roots(ALLSRC) & ::SRC)^)-obsolete()))'
Configuration Options:
You can make rebase require a destination if you set the following config
option::
[commands]
rebase.requiredest = True
By default, rebase will close the transaction after each commit. For
performance purposes, you can configure rebase to use a single transaction
across the entire rebase. WARNING: This setting introduces a significant
risk of losing the work you've done in a rebase if the rebase aborts
unexpectedly::
[rebase]
singletransaction = True
By default, rebase writes to the working copy, but you can configure it to
run in-memory for better performance. When the rebase is not moving the
parent(s) of the working copy (AKA the "currently checked out changesets"),
this may also allow it to run even if the working copy is dirty::
[rebase]
experimental.inmemory = True
Return Values:
Returns 0 on success, 1 if nothing to rebase or there are
unresolved conflicts.
"""
inmemory = ui.configbool(b'rebase', b'experimental.inmemory')
action = cmdutil.check_at_most_one_arg(opts, 'abort', 'stop', 'continue')
if action:
cmdutil.check_incompatible_arguments(
opts, action, ['confirm', 'dry_run']
)
cmdutil.check_incompatible_arguments(
opts, action, ['rev', 'source', 'base', 'dest']
)
cmdutil.check_at_most_one_arg(opts, 'confirm', 'dry_run')
cmdutil.check_at_most_one_arg(opts, 'rev', 'source', 'base')
if action or repo.currenttransaction() is not None:
# in-memory rebase is not compatible with resuming rebases.
# (Or if it is run within a transaction, since the restart logic can
# fail the entire transaction.)
inmemory = False
if opts.get('auto_orphans'):
disallowed_opts = set(opts) - {'auto_orphans'}
cmdutil.check_incompatible_arguments(
opts, 'auto_orphans', disallowed_opts
)
userrevs = list(repo.revs(opts.get('auto_orphans')))
opts['rev'] = [revsetlang.formatspec(b'%ld and orphan()', userrevs)]
opts['dest'] = b'_destautoorphanrebase(SRC)'
if opts.get('dry_run') or opts.get('confirm'):
return _dryrunrebase(ui, repo, action, opts)
elif action == 'stop':
rbsrt = rebaseruntime(repo, ui)
with repo.wlock(), repo.lock():
rbsrt.restorestatus()
if rbsrt.collapsef:
raise error.StateError(_(b"cannot stop in --collapse session"))
allowunstable = obsolete.isenabled(repo, obsolete.allowunstableopt)
if not (rbsrt.keepf or allowunstable):
raise error.StateError(
_(
b"cannot remove original changesets with"
b" unrebased descendants"
),
hint=_(
b'either enable obsmarkers to allow unstable '
b'revisions or use --keep to keep original '
b'changesets'
),
)
# update to the current working revision
# to clear interrupted merge
mergemod.clean_update(repo[rbsrt.originalwd])
rbsrt._finishrebase()
return 0
elif inmemory:
try:
# in-memory merge doesn't support conflicts, so if we hit any, abort
# and re-run as an on-disk merge.
overrides = {(b'rebase', b'singletransaction'): True}
with ui.configoverride(overrides, b'rebase'):
return _dorebase(ui, repo, action, opts, inmemory=inmemory)
except error.InMemoryMergeConflictsError:
if ui.configbool(b'devel', b'rebase.force-in-memory-merge'):
raise
ui.warn(
_(
b'hit merge conflicts; re-running rebase without in-memory'
b' merge\n'
)
)
clearstatus(repo)
clearcollapsemsg(repo)
return _dorebase(ui, repo, action, opts, inmemory=False)
else:
return _dorebase(ui, repo, action, opts)
def _dryrunrebase(ui, repo, action, opts):
rbsrt = rebaseruntime(repo, ui, inmemory=True, dryrun=True, opts=opts)
confirm = opts.get('confirm')
if confirm:
ui.status(_(b'starting in-memory rebase\n'))
else:
ui.status(
_(b'starting dry-run rebase; repository will not be changed\n')
)
with repo.wlock(), repo.lock():
needsabort = True
try:
overrides = {(b'rebase', b'singletransaction'): True}
with ui.configoverride(overrides, b'rebase'):
res = _origrebase(
ui,
repo,
action,
opts,
rbsrt,
)
if res == _nothingtorebase():
needsabort = False
return res
except error.ConflictResolutionRequired:
ui.status(_(b'hit a merge conflict\n'))
return 1
except error.Abort:
needsabort = False
raise
else:
if confirm:
ui.status(_(b'rebase completed successfully\n'))
if not ui.promptchoice(_(b'apply changes (yn)?$$ &Yes $$ &No')):
# finish unfinished rebase
rbsrt._finishrebase()
else:
rbsrt._prepareabortorcontinue(
isabort=True,
backup=False,
suppwarns=True,
confirm=confirm,
)
needsabort = False
else:
ui.status(
_(
b'dry-run rebase completed successfully; run without'
b' -n/--dry-run to perform this rebase\n'
)
)
return 0
finally:
if needsabort:
# no need to store backup in case of dryrun
rbsrt._prepareabortorcontinue(
isabort=True,
backup=False,
suppwarns=True,
dryrun=opts.get('dry_run'),
)
def _dorebase(ui, repo, action, opts, inmemory=False):
rbsrt = rebaseruntime(repo, ui, inmemory, opts=opts)
return _origrebase(ui, repo, action, opts, rbsrt)
def _origrebase(ui, repo, action, opts, rbsrt):
assert action != 'stop'
with repo.wlock(), repo.lock():
if opts.get('interactive'):
try:
if extensions.find(b'histedit'):
enablehistedit = b''
except KeyError:
enablehistedit = b" --config extensions.histedit="
help = b"hg%s help -e histedit" % enablehistedit
msg = (
_(
b"interactive history editing is supported by the "
b"'histedit' extension (see \"%s\")"
)
% help
)
raise error.InputError(msg)
if rbsrt.collapsemsg and not rbsrt.collapsef:
raise error.InputError(
_(b'message can only be specified with collapse')
)
if action:
if rbsrt.collapsef:
raise error.InputError(
_(b'cannot use collapse with continue or abort')
)
if action == 'abort' and opts.get('tool', False):
ui.warn(_(b'tool option will be ignored\n'))
if action == 'continue':
ms = mergestatemod.mergestate.read(repo)
mergeutil.checkunresolved(ms)
retcode = rbsrt._prepareabortorcontinue(isabort=(action == 'abort'))
if retcode is not None:
return retcode
else:
# search default destination in this space
# used in the 'hg pull --rebase' case, see issue 5214.
destspace = opts.get('_destspace')
destmap = _definedestmap(
ui,
repo,
rbsrt.inmemory,
opts.get('dest', None),
opts.get('source', []),
opts.get('base', []),
opts.get('rev', []),
destspace=destspace,
)
retcode = rbsrt._preparenewrebase(destmap)
if retcode is not None:
return retcode
storecollapsemsg(repo, rbsrt.collapsemsg)
tr = None
singletr = ui.configbool(b'rebase', b'singletransaction')
if singletr:
tr = repo.transaction(b'rebase')
# If `rebase.singletransaction` is enabled, wrap the entire operation in
# one transaction here. Otherwise, transactions are obtained when
# committing each node, which is slower but allows partial success.
with util.acceptintervention(tr):
# Same logic for the dirstate guard, except we don't create one when
# rebasing in-memory (it's not needed).
dsguard = None
if singletr and not rbsrt.inmemory:
dsguard = dirstateguard.dirstateguard(repo, b'rebase')
with util.acceptintervention(dsguard):
rbsrt._performrebase(tr)
if not rbsrt.dryrun:
rbsrt._finishrebase()
def _definedestmap(ui, repo, inmemory, destf, srcf, basef, revf, destspace):
"""use revisions argument to define destmap {srcrev: destrev}"""
if revf is None:
revf = []
# destspace is here to work around issues with `hg pull --rebase` see
# issue5214 for details
cmdutil.checkunfinished(repo)
if not inmemory:
cmdutil.bailifchanged(repo)
if ui.configbool(b'commands', b'rebase.requiredest') and not destf:
raise error.InputError(
_(b'you must specify a destination'),
hint=_(b'use: hg rebase -d REV'),
)
dest = None
if revf:
rebaseset = logcmdutil.revrange(repo, revf)
if not rebaseset:
ui.status(_(b'empty "rev" revision set - nothing to rebase\n'))
return None
elif srcf:
src = logcmdutil.revrange(repo, srcf)
if not src:
ui.status(_(b'empty "source" revision set - nothing to rebase\n'))
return None
# `+ (%ld)` to work around `wdir()::` being empty
rebaseset = repo.revs(b'(%ld):: + (%ld)', src, src)
else:
base = logcmdutil.revrange(repo, basef or [b'.'])
if not base:
ui.status(
_(b'empty "base" revision set - ' b"can't compute rebase set\n")
)
return None
if destf:
# --base does not support multiple destinations
dest = logcmdutil.revsingle(repo, destf)
else:
dest = repo[_destrebase(repo, base, destspace=destspace)]
destf = bytes(dest)
roots = [] # selected children of branching points
bpbase = {} # {branchingpoint: [origbase]}
for b in base: # group bases by branching points
bp = repo.revs(b'ancestor(%d, %d)', b, dest.rev()).first()
bpbase[bp] = bpbase.get(bp, []) + [b]
if None in bpbase:
# emulate the old behavior, showing "nothing to rebase" (a better
# behavior may be abort with "cannot find branching point" error)
bpbase.clear()
for bp, bs in bpbase.items(): # calculate roots
roots += list(repo.revs(b'children(%d) & ancestors(%ld)', bp, bs))
rebaseset = repo.revs(b'%ld::', roots)
if not rebaseset:
# transform to list because smartsets are not comparable to
# lists. This should be improved to honor laziness of
# smartset.
if list(base) == [dest.rev()]:
if basef:
ui.status(
_(
b'nothing to rebase - %s is both "base"'
b' and destination\n'
)
% dest
)
else:
ui.status(
_(
b'nothing to rebase - working directory '
b'parent is also destination\n'
)
)
elif not repo.revs(b'%ld - ::%d', base, dest.rev()):
if basef:
ui.status(
_(
b'nothing to rebase - "base" %s is '
b'already an ancestor of destination '
b'%s\n'
)
% (b'+'.join(bytes(repo[r]) for r in base), dest)
)
else:
ui.status(
_(
b'nothing to rebase - working '
b'directory parent is already an '
b'ancestor of destination %s\n'
)
% dest
)
else: # can it happen?
ui.status(
_(b'nothing to rebase from %s to %s\n')
% (b'+'.join(bytes(repo[r]) for r in base), dest)
)
return None
if wdirrev in rebaseset:
raise error.InputError(_(b'cannot rebase the working copy'))
rebasingwcp = repo[b'.'].rev() in rebaseset
ui.log(
b"rebase",
b"rebasing working copy parent: %r\n",
rebasingwcp,
rebase_rebasing_wcp=rebasingwcp,
)
if inmemory and rebasingwcp:
# Check these since we did not before.
cmdutil.checkunfinished(repo)
cmdutil.bailifchanged(repo)
if not destf:
dest = repo[_destrebase(repo, rebaseset, destspace=destspace)]
destf = bytes(dest)
allsrc = revsetlang.formatspec(b'%ld', rebaseset)
alias = {b'ALLSRC': allsrc}
if dest is None:
try:
# fast path: try to resolve dest without SRC alias
dest = scmutil.revsingle(repo, destf, localalias=alias)
except error.RepoLookupError:
# multi-dest path: resolve dest for each SRC separately
destmap = {}
for r in rebaseset:
alias[b'SRC'] = revsetlang.formatspec(b'%d', r)
# use repo.anyrevs instead of scmutil.revsingle because we
# don't want to abort if destset is empty.
destset = repo.anyrevs([destf], user=True, localalias=alias)
size = len(destset)
if size == 1:
destmap[r] = destset.first()
elif size == 0:
ui.note(_(b'skipping %s - empty destination\n') % repo[r])
else:
raise error.InputError(
_(b'rebase destination for %s is not unique') % repo[r]
)
if dest is not None:
# single-dest case: assign dest to each rev in rebaseset
destrev = dest.rev()
destmap = {r: destrev for r in rebaseset} # {srcrev: destrev}
if not destmap:
ui.status(_(b'nothing to rebase - empty destination\n'))
return None
return destmap
def externalparent(repo, state, destancestors):
"""Return the revision that should be used as the second parent
when the revisions in state is collapsed on top of destancestors.
Abort if there is more than one parent.
"""
parents = set()
source = min(state)
for rev in state:
if rev == source:
continue
for p in repo[rev].parents():
if p.rev() not in state and p.rev() not in destancestors:
parents.add(p.rev())
if not parents:
return nullrev
if len(parents) == 1:
return parents.pop()
raise error.StateError(
_(
b'unable to collapse on top of %d, there is more '
b'than one external parent: %s'
)
% (max(destancestors), b', '.join(b"%d" % p for p in sorted(parents)))
)
def commitmemorynode(repo, wctx, editor, extra, user, date, commitmsg):
"""Commit the memory changes with parents p1 and p2.
Return node of committed revision."""
# By convention, ``extra['branch']`` (set by extrafn) clobbers
# ``branch`` (used when passing ``--keepbranches``).
branch = None
if b'branch' in extra:
branch = extra[b'branch']
# FIXME: We call _compact() because it's required to correctly detect
# changed files. This was added to fix a regression shortly before the 5.5
# release. A proper fix will be done in the default branch.
wctx._compact()
memctx = wctx.tomemctx(
commitmsg,
date=date,
extra=extra,
user=user,
branch=branch,
editor=editor,
)
if memctx.isempty() and not repo.ui.configbool(b'ui', b'allowemptycommit'):
return None
commitres = repo.commitctx(memctx)
wctx.clean() # Might be reused
return commitres
def commitnode(repo, editor, extra, user, date, commitmsg):
"""Commit the wd changes with parents p1 and p2.
Return node of committed revision."""
dsguard = util.nullcontextmanager()
if not repo.ui.configbool(b'rebase', b'singletransaction'):
dsguard = dirstateguard.dirstateguard(repo, b'rebase')
with dsguard:
# Commit might fail if unresolved files exist
newnode = repo.commit(
text=commitmsg, user=user, date=date, extra=extra, editor=editor
)
repo.dirstate.setbranch(repo[newnode].branch())
return newnode
def rebasenode(repo, rev, p1, p2, base, collapse, wctx):
"""Rebase a single revision rev on top of p1 using base as merge ancestor"""
# Merge phase
# Update to destination and merge it with local
p1ctx = repo[p1]
if wctx.isinmemory():
wctx.setbase(p1ctx)
else:
if repo[b'.'].rev() != p1:
repo.ui.debug(b" update to %d:%s\n" % (p1, p1ctx))
mergemod.clean_update(p1ctx)
else:
repo.ui.debug(b" already in destination\n")
# This is, alas, necessary to invalidate workingctx's manifest cache,
# as well as other data we litter on it in other places.
wctx = repo[None]
repo.dirstate.write(repo.currenttransaction())
ctx = repo[rev]
repo.ui.debug(b" merge against %d:%s\n" % (rev, ctx))
if base is not None:
repo.ui.debug(b" detach base %d:%s\n" % (base, repo[base]))
# See explanation in merge.graft()
mergeancestor = repo.changelog.isancestor(p1ctx.node(), ctx.node())
stats = mergemod._update(
repo,
rev,
branchmerge=True,
force=True,
ancestor=base,
mergeancestor=mergeancestor,
labels=[b'dest', b'source', b'parent of source'],
wc=wctx,
)
wctx.setparents(p1ctx.node(), repo[p2].node())
if collapse:
copies.graftcopies(wctx, ctx, p1ctx)
else:
# If we're not using --collapse, we need to
# duplicate copies between the revision we're
# rebasing and its first parent.
copies.graftcopies(wctx, ctx, ctx.p1())
if stats.unresolvedcount > 0:
if wctx.isinmemory():
raise error.InMemoryMergeConflictsError()
else:
raise error.ConflictResolutionRequired(b'rebase')
def adjustdest(repo, rev, destmap, state, skipped):
r"""adjust rebase destination given the current rebase state
rev is what is being rebased. Return a list of two revs, which are the
adjusted destinations for rev's p1 and p2, respectively. If a parent is
nullrev, return dest without adjustment for it.
For example, when doing rebasing B+E to F, C to G, rebase will first move B
to B1, and E's destination will be adjusted from F to B1.
B1 <- written during rebasing B
|
F <- original destination of B, E
|
| E <- rev, which is being rebased
| |
| D <- prev, one parent of rev being checked
| |
| x <- skipped, ex. no successor or successor in (::dest)
| |
| C <- rebased as C', different destination
| |
| B <- rebased as B1 C'
|/ |
A G <- destination of C, different
Another example about merge changeset, rebase -r C+G+H -d K, rebase will
first move C to C1, G to G1, and when it's checking H, the adjusted
destinations will be [C1, G1].
H C1 G1
/| | /
F G |/
K | | -> K
| C D |
| |/ |
| B | ...
|/ |/
A A
Besides, adjust dest according to existing rebase information. For example,
B C D B needs to be rebased on top of C, C needs to be rebased on top
\|/ of D. We will rebase C first.
A
C' After rebasing C, when considering B's destination, use C'
| instead of the original C.
B D
\ /
A
"""
# pick already rebased revs with same dest from state as interesting source
dest = destmap[rev]
source = [
s
for s, d in state.items()
if d > 0 and destmap[s] == dest and s not in skipped
]
result = []
for prev in repo.changelog.parentrevs(rev):
adjusted = dest
if prev != nullrev:
candidate = repo.revs(b'max(%ld and (::%d))', source, prev).first()
if candidate is not None:
adjusted = state[candidate]
if adjusted == dest and dest in state:
adjusted = state[dest]
if adjusted == revtodo:
# sortsource should produce an order that makes this impossible
raise error.ProgrammingError(
b'rev %d should be rebased already at this time' % dest
)
result.append(adjusted)
return result
def _checkobsrebase(repo, ui, rebaseobsrevs, rebaseobsskipped):
"""
Abort if rebase will create divergence or rebase is noop because of markers
`rebaseobsrevs`: set of obsolete revision in source
`rebaseobsskipped`: set of revisions from source skipped because they have
successors in destination or no non-obsolete successor.
"""
# Obsolete node with successors not in dest leads to divergence
divergenceok = obsolete.isenabled(repo, obsolete.allowdivergenceopt)
divergencebasecandidates = rebaseobsrevs - rebaseobsskipped
if divergencebasecandidates and not divergenceok:
divhashes = (bytes(repo[r]) for r in divergencebasecandidates)
msg = _(b"this rebase will cause divergences from: %s")
h = _(
b"to force the rebase please set "
b"experimental.evolution.allowdivergence=True"
)
raise error.StateError(msg % (b",".join(divhashes),), hint=h)
def successorrevs(unfi, rev):
"""yield revision numbers for successors of rev"""
assert unfi.filtername is None
get_rev = unfi.changelog.index.get_rev
for s in obsutil.allsuccessors(unfi.obsstore, [unfi[rev].node()]):
r = get_rev(s)
if r is not None:
yield r
def defineparents(repo, rev, destmap, state, skipped, obsskipped):
"""Return new parents and optionally a merge base for rev being rebased
The destination specified by "dest" cannot always be used directly because
previously rebase result could affect destination. For example,
D E rebase -r C+D+E -d B
|/ C will be rebased to C'
B C D's new destination will be C' instead of B
|/ E's new destination will be C' instead of B
A
The new parents of a merge is slightly more complicated. See the comment
block below.
"""
# use unfiltered changelog since successorrevs may return filtered nodes
assert repo.filtername is None
cl = repo.changelog
isancestor = cl.isancestorrev
dest = destmap[rev]
oldps = repo.changelog.parentrevs(rev) # old parents
newps = [nullrev, nullrev] # new parents
dests = adjustdest(repo, rev, destmap, state, skipped)
bases = list(oldps) # merge base candidates, initially just old parents
if all(r == nullrev for r in oldps[1:]):
# For non-merge changeset, just move p to adjusted dest as requested.
newps[0] = dests[0]
else:
# For merge changeset, if we move p to dests[i] unconditionally, both
# parents may change and the end result looks like "the merge loses a
# parent", which is a surprise. This is a limit because "--dest" only
# accepts one dest per src.
#
# Therefore, only move p with reasonable conditions (in this order):
# 1. use dest, if dest is a descendent of (p or one of p's successors)
# 2. use p's rebased result, if p is rebased (state[p] > 0)
#
# Comparing with adjustdest, the logic here does some additional work:
# 1. decide which parents will not be moved towards dest
# 2. if the above decision is "no", should a parent still be moved
# because it was rebased?
#
# For example:
#
# C # "rebase -r C -d D" is an error since none of the parents
# /| # can be moved. "rebase -r B+C -d D" will move C's parent
# A B D # B (using rule "2."), since B will be rebased.
#
# The loop tries to be not rely on the fact that a Mercurial node has
# at most 2 parents.
for i, p in enumerate(oldps):
np = p # new parent
if any(isancestor(x, dests[i]) for x in successorrevs(repo, p)):
np = dests[i]
elif p in state and state[p] > 0:
np = state[p]
# If one parent becomes an ancestor of the other, drop the ancestor
for j, x in enumerate(newps[:i]):
if x == nullrev:
continue
if isancestor(np, x): # CASE-1
np = nullrev
elif isancestor(x, np): # CASE-2
newps[j] = np
np = nullrev
# New parents forming an ancestor relationship does not
# mean the old parents have a similar relationship. Do not
# set bases[x] to nullrev.
bases[j], bases[i] = bases[i], bases[j]
newps[i] = np
# "rebasenode" updates to new p1, and the old p1 will be used as merge
# base. If only p2 changes, merging using unchanged p1 as merge base is
# suboptimal. Therefore swap parents to make the merge sane.
if newps[1] != nullrev and oldps[0] == newps[0]:
assert len(newps) == 2 and len(oldps) == 2
newps.reverse()
bases.reverse()
# No parent change might be an error because we fail to make rev a
# descendent of requested dest. This can happen, for example:
#
# C # rebase -r C -d D
# /| # None of A and B will be changed to D and rebase fails.
# A B D
if set(newps) == set(oldps) and dest not in newps:
raise error.InputError(
_(
b'cannot rebase %d:%s without '
b'moving at least one of its parents'
)
% (rev, repo[rev])
)
# Source should not be ancestor of dest. The check here guarantees it's
# impossible. With multi-dest, the initial check does not cover complex
# cases since we don't have abstractions to dry-run rebase cheaply.
if any(p != nullrev and isancestor(rev, p) for p in newps):
raise error.InputError(_(b'source is ancestor of destination'))
# Check if the merge will contain unwanted changes. That may happen if
# there are multiple special (non-changelog ancestor) merge bases, which
# cannot be handled well by the 3-way merge algorithm. For example:
#
# F
# /|
# D E # "rebase -r D+E+F -d Z", when rebasing F, if "D" was chosen
# | | # as merge base, the difference between D and F will include
# B C # C, so the rebased F will contain C surprisingly. If "E" was
# |/ # chosen, the rebased F will contain B.
# A Z
#
# But our merge base candidates (D and E in above case) could still be
# better than the default (ancestor(F, Z) == null). Therefore still
# pick one (so choose p1 above).
if sum(1 for b in set(bases) if b != nullrev and b not in newps) > 1:
unwanted = [None, None] # unwanted[i]: unwanted revs if choose bases[i]
for i, base in enumerate(bases):
if base == nullrev or base in newps:
continue
# Revisions in the side (not chosen as merge base) branch that
# might contain "surprising" contents
other_bases = set(bases) - {base}
siderevs = list(
repo.revs(b'(%ld %% (%d+%d))', other_bases, base, dest)
)
# If those revisions are covered by rebaseset, the result is good.
# A merge in rebaseset would be considered to cover its ancestors.
if siderevs:
rebaseset = [
r for r, d in state.items() if d > 0 and r not in obsskipped
]
merges = [
r for r in rebaseset if cl.parentrevs(r)[1] != nullrev
]
unwanted[i] = list(
repo.revs(
b'%ld - (::%ld) - %ld', siderevs, merges, rebaseset
)
)
if any(revs is not None for revs in unwanted):
# Choose a merge base that has a minimal number of unwanted revs.
l, i = min(
(len(revs), i)
for i, revs in enumerate(unwanted)
if revs is not None
)
# The merge will include unwanted revisions. Abort now. Revisit this if
# we have a more advanced merge algorithm that handles multiple bases.
if l > 0:
unwanteddesc = _(b' or ').join(
(
b', '.join(b'%d:%s' % (r, repo[r]) for r in revs)
for revs in unwanted
if revs is not None
)
)
raise error.InputError(
_(b'rebasing %d:%s will include unwanted changes from %s')
% (rev, repo[rev], unwanteddesc)
)
# newps[0] should match merge base if possible. Currently, if newps[i]
# is nullrev, the only case is newps[i] and newps[j] (j < i), one is
# the other's ancestor. In that case, it's fine to not swap newps here.
# (see CASE-1 and CASE-2 above)
if i != 0:
if newps[i] != nullrev:
newps[0], newps[i] = newps[i], newps[0]
bases[0], bases[i] = bases[i], bases[0]
# "rebasenode" updates to new p1, use the corresponding merge base.
base = bases[0]
repo.ui.debug(b" future parents are %d and %d\n" % tuple(newps))
return newps[0], newps[1], base
def isagitpatch(repo, patchname):
"""Return true if the given patch is in git format"""
mqpatch = os.path.join(repo.mq.path, patchname)
for line in patch.linereader(open(mqpatch, b'rb')):
if line.startswith(b'diff --git'):
return True
return False
def updatemq(repo, state, skipped, **opts):
"""Update rebased mq patches - finalize and then import them"""
mqrebase = {}
mq = repo.mq
original_series = mq.fullseries[:]
skippedpatches = set()
for p in mq.applied:
rev = repo[p.node].rev()
if rev in state:
repo.ui.debug(
b'revision %d is an mq patch (%s), finalize it.\n'
% (rev, p.name)
)
mqrebase[rev] = (p.name, isagitpatch(repo, p.name))
else:
# Applied but not rebased, not sure this should happen
skippedpatches.add(p.name)
if mqrebase:
mq.finish(repo, mqrebase.keys())
# We must start import from the newest revision
for rev in sorted(mqrebase, reverse=True):
if rev not in skipped:
name, isgit = mqrebase[rev]
repo.ui.note(
_(b'updating mq patch %s to %d:%s\n')
% (name, state[rev], repo[state[rev]])
)
mq.qimport(
repo,
(),
patchname=name,
git=isgit,
rev=[b"%d" % state[rev]],
)
else:
# Rebased and skipped
skippedpatches.add(mqrebase[rev][0])
# Patches were either applied and rebased and imported in
# order, applied and removed or unapplied. Discard the removed
# ones while preserving the original series order and guards.
newseries = [
s
for s in original_series
if mq.guard_re.split(s, 1)[0] not in skippedpatches
]
mq.fullseries[:] = newseries
mq.seriesdirty = True
mq.savedirty()
def storecollapsemsg(repo, collapsemsg):
"""Store the collapse message to allow recovery"""
collapsemsg = collapsemsg or b''
f = repo.vfs(b"last-message.txt", b"w")
f.write(b"%s\n" % collapsemsg)
f.close()
def clearcollapsemsg(repo):
"""Remove collapse message file"""
repo.vfs.unlinkpath(b"last-message.txt", ignoremissing=True)
def restorecollapsemsg(repo, isabort):
"""Restore previously stored collapse message"""
try:
f = repo.vfs(b"last-message.txt")
collapsemsg = f.readline().strip()
f.close()
except FileNotFoundError:
if isabort:
# Oh well, just abort like normal
collapsemsg = b''
else:
raise error.Abort(_(b'missing .hg/last-message.txt for rebase'))
return collapsemsg
def clearstatus(repo):
"""Remove the status files"""
# Make sure the active transaction won't write the state file
tr = repo.currenttransaction()
if tr:
tr.removefilegenerator(b'rebasestate')
repo.vfs.unlinkpath(b"rebasestate", ignoremissing=True)
def sortsource(destmap):
"""yield source revisions in an order that we only rebase things once
If source and destination overlaps, we should filter out revisions
depending on other revisions which hasn't been rebased yet.
Yield a sorted list of revisions each time.
For example, when rebasing A to B, B to C. This function yields [B], then
[A], indicating B needs to be rebased first.
Raise if there is a cycle so the rebase is impossible.
"""
srcset = set(destmap)
while srcset:
srclist = sorted(srcset)
result = []
for r in srclist:
if destmap[r] not in srcset:
result.append(r)
if not result:
raise error.InputError(_(b'source and destination form a cycle'))
srcset -= set(result)
yield result
def buildstate(repo, destmap, collapse):
"""Define which revisions are going to be rebased and where
repo: repo
destmap: {srcrev: destrev}
"""
rebaseset = destmap.keys()
originalwd = repo[b'.'].rev()
# This check isn't strictly necessary, since mq detects commits over an
# applied patch. But it prevents messing up the working directory when
# a partially completed rebase is blocked by mq.
if b'qtip' in repo.tags():
mqapplied = {repo[s.node].rev() for s in repo.mq.applied}
if set(destmap.values()) & mqapplied:
raise error.StateError(_(b'cannot rebase onto an applied mq patch'))
# Get "cycle" error early by exhausting the generator.
sortedsrc = list(sortsource(destmap)) # a list of sorted revs
if not sortedsrc:
raise error.InputError(_(b'no matching revisions'))
# Only check the first batch of revisions to rebase not depending on other
# rebaseset. This means "source is ancestor of destination" for the second
# (and following) batches of revisions are not checked here. We rely on
# "defineparents" to do that check.
roots = list(repo.set(b'roots(%ld)', sortedsrc[0]))
if not roots:
raise error.InputError(_(b'no matching revisions'))
def revof(r):
return r.rev()
roots = sorted(roots, key=revof)
state = dict.fromkeys(rebaseset, revtodo)
emptyrebase = len(sortedsrc) == 1
for root in roots:
dest = repo[destmap[root.rev()]]
commonbase = root.ancestor(dest)
if commonbase == root:
raise error.InputError(_(b'source is ancestor of destination'))
if commonbase == dest:
wctx = repo[None]
if dest == wctx.p1():
# when rebasing to '.', it will use the current wd branch name
samebranch = root.branch() == wctx.branch()
else:
samebranch = root.branch() == dest.branch()
if not collapse and samebranch and dest in root.parents():
# mark the revision as done by setting its new revision
# equal to its old (current) revisions
state[root.rev()] = root.rev()
repo.ui.debug(b'source is a child of destination\n')
continue
emptyrebase = False
repo.ui.debug(b'rebase onto %s starting from %s\n' % (dest, root))
if emptyrebase:
return None
for rev in sorted(state):
parents = [p for p in repo.changelog.parentrevs(rev) if p != nullrev]
# if all parents of this revision are done, then so is this revision
if parents and all((state.get(p) == p for p in parents)):
state[rev] = rev
return originalwd, destmap, state
def clearrebased(
ui,
repo,
destmap,
state,
skipped,
collapsedas=None,
keepf=False,
fm=None,
backup=True,
):
"""dispose of rebased revision at the end of the rebase
If `collapsedas` is not None, the rebase was a collapse whose result if the
`collapsedas` node.
If `keepf` is not True, the rebase has --keep set and no nodes should be
removed (but bookmarks still need to be moved).
If `backup` is False, no backup will be stored when stripping rebased
revisions.
"""
tonode = repo.changelog.node
replacements = {}
moves = {}
stripcleanup = not obsolete.isenabled(repo, obsolete.createmarkersopt)
collapsednodes = []
for rev, newrev in sorted(state.items()):
if newrev >= 0 and newrev != rev:
oldnode = tonode(rev)
newnode = collapsedas or tonode(newrev)
moves[oldnode] = newnode
succs = None
if rev in skipped:
if stripcleanup or not repo[rev].obsolete():
succs = ()
elif collapsedas:
collapsednodes.append(oldnode)
else:
succs = (newnode,)
if succs is not None:
replacements[(oldnode,)] = succs
if collapsednodes:
replacements[tuple(collapsednodes)] = (collapsedas,)
if fm:
hf = fm.hexfunc
fl = fm.formatlist
fd = fm.formatdict
changes = {}
for oldns, newn in replacements.items():
for oldn in oldns:
changes[hf(oldn)] = fl([hf(n) for n in newn], name=b'node')
nodechanges = fd(changes, key=b"oldnode", value=b"newnodes")
fm.data(nodechanges=nodechanges)
if keepf:
replacements = {}
scmutil.cleanupnodes(repo, replacements, b'rebase', moves, backup=backup)
def pullrebase(orig, ui, repo, *args, **opts):
"""Call rebase after pull if the latter has been invoked with --rebase"""
if opts.get('rebase'):
if ui.configbool(b'commands', b'rebase.requiredest'):
msg = _(b'rebase destination required by configuration')
hint = _(b'use hg pull followed by hg rebase -d DEST')
raise error.InputError(msg, hint=hint)
with repo.wlock(), repo.lock():
if opts.get('update'):
del opts['update']
ui.debug(
b'--update and --rebase are not compatible, ignoring '
b'the update flag\n'
)
cmdutil.checkunfinished(repo, skipmerge=True)
cmdutil.bailifchanged(
repo,
hint=_(
b'cannot pull with rebase: '
b'please commit or shelve your changes first'
),
)
revsprepull = len(repo)
origpostincoming = commands.postincoming
def _dummy(*args, **kwargs):
pass
commands.postincoming = _dummy
try:
ret = orig(ui, repo, *args, **opts)
finally:
commands.postincoming = origpostincoming
revspostpull = len(repo)
if revspostpull > revsprepull:
# --rev option from pull conflict with rebase own --rev
# dropping it
if 'rev' in opts:
del opts['rev']
# positional argument from pull conflicts with rebase's own
# --source.
if 'source' in opts:
del opts['source']
# revsprepull is the len of the repo, not revnum of tip.
destspace = list(repo.changelog.revs(start=revsprepull))
opts['_destspace'] = destspace
try:
rebase(ui, repo, **opts)
except error.NoMergeDestAbort:
# we can maybe update instead
rev, _a, _b = destutil.destupdate(repo)
if rev == repo[b'.'].rev():
ui.status(_(b'nothing to rebase\n'))
else:
ui.status(_(b'nothing to rebase - updating instead\n'))
# not passing argument to get the bare update behavior
# with warning and trumpets
commands.update(ui, repo)
else:
if opts.get('tool'):
raise error.InputError(_(b'--tool can only be used with --rebase'))
ret = orig(ui, repo, *args, **opts)
return ret
def _compute_obsolete_sets(repo, rebaseobsrevs, destmap):
"""Figure out what to do about about obsolete revisions
`obsolete_with_successor_in_destination` is a mapping mapping obsolete => successor for all
obsolete nodes to be rebased given in `rebaseobsrevs`.
`obsolete_with_successor_in_rebase_set` is a set with obsolete revisions,
without a successor in destination, that would cause divergence.
"""
obsolete_with_successor_in_destination = {}
obsolete_with_successor_in_rebase_set = set()
cl = repo.changelog
get_rev = cl.index.get_rev
extinctrevs = set(repo.revs(b'extinct()'))
for srcrev in rebaseobsrevs:
srcnode = cl.node(srcrev)
# XXX: more advanced APIs are required to handle split correctly
successors = set(obsutil.allsuccessors(repo.obsstore, [srcnode]))
# obsutil.allsuccessors includes node itself
successors.remove(srcnode)
succrevs = {get_rev(s) for s in successors}
succrevs.discard(None)
if not successors or succrevs.issubset(extinctrevs):
# no successor, or all successors are extinct
obsolete_with_successor_in_destination[srcrev] = None
else:
dstrev = destmap[srcrev]
for succrev in succrevs:
if cl.isancestorrev(succrev, dstrev):
obsolete_with_successor_in_destination[srcrev] = succrev
break
else:
# If 'srcrev' has a successor in rebase set but none in
# destination (which would be catched above), we shall skip it
# and its descendants to avoid divergence.
if srcrev in extinctrevs or any(s in destmap for s in succrevs):
obsolete_with_successor_in_rebase_set.add(srcrev)
return (
obsolete_with_successor_in_destination,
obsolete_with_successor_in_rebase_set,
)
def abortrebase(ui, repo):
with repo.wlock(), repo.lock():
rbsrt = rebaseruntime(repo, ui)
rbsrt._prepareabortorcontinue(isabort=True)
def continuerebase(ui, repo):
with repo.wlock(), repo.lock():
rbsrt = rebaseruntime(repo, ui)
ms = mergestatemod.mergestate.read(repo)
mergeutil.checkunresolved(ms)
retcode = rbsrt._prepareabortorcontinue(isabort=False)
if retcode is not None:
return retcode
rbsrt._performrebase(None)
rbsrt._finishrebase()
def summaryhook(ui, repo):
if not repo.vfs.exists(b'rebasestate'):
return
try:
rbsrt = rebaseruntime(repo, ui, {})
rbsrt.restorestatus()
state = rbsrt.state
except error.RepoLookupError:
# i18n: column positioning for "hg summary"
msg = _(b'rebase: (use "hg rebase --abort" to clear broken state)\n')
ui.write(msg)
return
numrebased = len([i for i in state.values() if i >= 0])
# i18n: column positioning for "hg summary"
ui.write(
_(b'rebase: %s, %s (rebase --continue)\n')
% (
ui.label(_(b'%d rebased'), b'rebase.rebased') % numrebased,
ui.label(_(b'%d remaining'), b'rebase.remaining')
% (len(state) - numrebased),
)
)
def uisetup(ui):
# Replace pull with a decorator to provide --rebase option
entry = extensions.wrapcommand(commands.table, b'pull', pullrebase)
entry[1].append(
(b'', b'rebase', None, _(b"rebase working directory to branch head"))
)
entry[1].append((b't', b'tool', b'', _(b"specify merge tool for rebase")))
cmdutil.summaryhooks.add(b'rebase', summaryhook)
statemod.addunfinished(
b'rebase',
fname=b'rebasestate',
stopflag=True,
continueflag=True,
abortfunc=abortrebase,
continuefunc=continuerebase,
)