##// END OF EJS Templates
streamclone: move applystreamclone() from localrepo.py...
streamclone: move applystreamclone() from localrepo.py Upcoming patches will modernize the streaming clone code. Streaming clone data and code kind of lives in its own world. exchange.py is arguably the most appropriate existing location for it. However, over a dozen patches from now it became apparent that there was a lot of code related to streaming clones and that having it contained within its own module would make it easier to comprehend. So, we establish streamclone.py. It's worth noting that streamclone.py existed a long time ago, last seen in the 1.6 release. It was removed in 04f76a954842. The function was renamed as part of the move because its old name was redundant with the new module name. The only other content change was "self" was renamed to "repo" and minor grammar in the docstring was updated.

File last commit:

r25944:337d010f default
r26441:56527b88 default
Show More
discovery.py
393 lines | 15.5 KiB | text/x-python | PythonLexer
Dirkjan Ochtman
discovery: fix description line
r11313 # discovery.py - protocol changeset discovery functions
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301 #
Dirkjan Ochtman
discovery: fix description line
r11313 # Copyright 2010 Matt Mackall <mpm@selenic.com>
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301 #
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
Gregory Szorc
discovery: use absolute_import
r25944 from __future__ import absolute_import
from .i18n import _
from .node import (
nullid,
short,
)
from . import (
bookmarks,
branchmap,
obsolete,
phases,
setdiscovery,
treediscovery,
util,
)
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Peter Arrenbrecht
discovery: drop findoutgoing and simplify findcommonincoming's api...
r14073 def findcommonincoming(repo, remote, heads=None, force=False):
"""Return a tuple (common, anyincoming, heads) used to identify the common
subset of nodes between repo and remote.
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Peter Arrenbrecht
discovery: drop findoutgoing and simplify findcommonincoming's api...
r14073 "common" is a list of (at least) the heads of the common subset.
"anyincoming" is testable as a boolean indicating if any nodes are missing
locally. If remote does not support getbundle, this actually is a list of
roots of the nodes that would be incoming, to be supplied to
changegroupsubset. No code except for pull should be relying on this fact
any longer.
"heads" is either the supplied heads, or else the remote's heads.
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164
Mads Kiilerich
fix trivial spelling errors
r17424 If you pass heads and they are all known locally, the response lists just
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 these heads in "common" and in "heads".
Peter Arrenbrecht
discovery: resurrect findoutgoing as findcommonoutgoing for extension hooks...
r14213
Please use findcommonoutgoing to compute the set of outgoing nodes to give
extensions a good hook into outgoing.
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301 """
Peter Arrenbrecht
discovery: drop findoutgoing and simplify findcommonincoming's api...
r14073
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 if not remote.capable('getbundle'):
return treediscovery.findcommonincoming(repo, remote, heads, force)
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 if heads:
allknown = True
Pierre-Yves David
discovery: stop using nodemap for membership testing...
r20225 knownnode = repo.changelog.hasnode # no nodemap until it is filtered
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 for h in heads:
Pierre-Yves David
discovery: stop using nodemap for membership testing...
r20225 if not knownnode(h):
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 allknown = False
break
if allknown:
return (heads, False, heads)
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Peter Arrenbrecht
discovery: add new set-based discovery...
r14164 res = setdiscovery.findcommonheads(repo.ui, repo, remote,
abortwhenunrelated=not force)
common, anyinc, srvheads = res
return (list(common), anyinc, heads or list(srvheads))
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Pierre-Yves David
discovery: introduce outgoing object for result of findcommonoutgoing...
r15837 class outgoing(object):
'''Represents the set of nodes present in a local repo but not in a
(possibly) remote one.
Members:
missing is a list of all nodes present in local but not in remote.
common is a list of all nodes shared between the two repos.
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 excluded is the list of missing changeset that shouldn't be sent remotely.
Pierre-Yves David
discovery: introduce outgoing object for result of findcommonoutgoing...
r15837 missingheads is the list of heads of missing.
commonheads is the list of heads of common.
The sets are computed on demand from the heads, unless provided upfront
by discovery.'''
def __init__(self, revlog, commonheads, missingheads):
self.commonheads = commonheads
self.missingheads = missingheads
self._revlog = revlog
self._common = None
self._missing = None
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 self.excluded = []
Pierre-Yves David
discovery: introduce outgoing object for result of findcommonoutgoing...
r15837
def _computecommonmissing(self):
sets = self._revlog.findcommonmissing(self.commonheads,
self.missingheads)
self._common, self._missing = sets
@util.propertycache
def common(self):
if self._common is None:
self._computecommonmissing()
return self._common
@util.propertycache
def missing(self):
if self._missing is None:
self._computecommonmissing()
return self._missing
Brodie Rao
cleanup: eradicate long lines
r16683 def findcommonoutgoing(repo, other, onlyheads=None, force=False,
Matt Mackall
merge with stable
r16746 commoninc=None, portable=False):
Pierre-Yves David
discovery: introduce outgoing object for result of findcommonoutgoing...
r15837 '''Return an outgoing instance to identify the nodes present in repo but
not in other.
Peter Arrenbrecht
discovery: resurrect findoutgoing as findcommonoutgoing for extension hooks...
r14213
Brodie Rao
cleanup: eradicate long lines
r16683 If onlyheads is given, only nodes ancestral to nodes in onlyheads
(inclusive) are included. If you already know the local repo's heads,
passing them in onlyheads is faster than letting them be recomputed here.
Peter Arrenbrecht
discovery: resurrect findoutgoing as findcommonoutgoing for extension hooks...
r14213
Mads Kiilerich
help: fix some instances of 'the the'
r17251 If commoninc is given, it must be the result of a prior call to
Sune Foldager
bundle: make bundles more portable (isue3441)...
r16736 findcommonincoming(repo, other, force) to avoid recomputing it here.
If portable is given, compute more conservative common and missingheads,
to make bundles created from the instance more portable.'''
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 # declare an empty outgoing object to be filled later
og = outgoing(repo.changelog, None, None)
# get common set if not provided
if commoninc is None:
commoninc = findcommonincoming(repo, other, force=force)
og.commonheads, _any, _hds = commoninc
# compute outgoing
Pierre-Yves David
obsolete: do not exchange extinct changesets...
r17206 mayexclude = (repo._phasecache.phaseroots[phases.secret] or repo.obsstore)
if not mayexclude:
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 og.missingheads = onlyheads or repo.heads()
elif onlyheads is None:
# use visible heads as it should be cached
Kevin Bullock
filtering: rename filters to their antonyms...
r18382 og.missingheads = repo.filtered("served").heads()
Patrick Mezard
discovery: add extinct changesets to outgoing.excluded...
r17248 og.excluded = [ctx.node() for ctx in repo.set('secret() or extinct()')]
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 else:
# compute common, missing and exclude secret stuff
sets = repo.changelog.findcommonmissing(og.commonheads, onlyheads)
og._common, allmissing = sets
og._missing = missing = []
Pierre-Yves David
phases: properly register excluded changeset when revision are specified...
r15951 og.excluded = excluded = []
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 for node in allmissing:
Pierre-Yves David
obsolete: do not exchange extinct changesets...
r17206 ctx = repo[node]
Patrick Mezard
discovery: add extinct changesets to outgoing.excluded...
r17248 if ctx.phase() >= phases.secret or ctx.extinct():
excluded.append(node)
else:
missing.append(node)
Pierre-Yves David
obsolete: do not exchange extinct changesets...
r17206 if len(missing) == len(allmissing):
missingheads = onlyheads
else: # update missing heads
Pierre-Yves David
discovery: ensure that missingheads are always heads of everything we tried...
r15955 missingheads = phases.newheads(repo, onlyheads, excluded)
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 og.missingheads = missingheads
Sune Foldager
bundle: make bundles more portable (isue3441)...
r16736 if portable:
# recompute common and missingheads as if -r<rev> had been given for
# each head of missing, and --base <rev> for each head of the proper
# ancestors of missing
og._computecommonmissing()
cl = repo.changelog
missingrevs = set(cl.rev(n) for n in og._missing)
Bryan O'Sullivan
revlog: ancestors(*revs) becomes ancestors(revs) (API)...
r16866 og._common = set(cl.ancestors(missingrevs)) - missingrevs
Sune Foldager
bundle: make bundles more portable (isue3441)...
r16736 commonheads = set(og.commonheads)
og.missingheads = [h for h in og.missingheads if h not in commonheads]
Pierre-Yves David
phases: make outgoing object and discovery aware of exclusion...
r15838 return og
Peter Arrenbrecht
discovery: resurrect findoutgoing as findcommonoutgoing for extension hooks...
r14213
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 def _headssummary(repo, remote, outgoing):
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 """compute a summary of branch and heads status before and after push
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 return {'branch': ([remoteheads], [newheads], [unsyncedheads])} mapping
- branch: the branch name
- remoteheads: the list of remote heads known locally
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 None if the branch is new
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 - newheads: the new remote heads (known locally) with outgoing pushed
- unsyncedheads: the list of remote heads unknown locally.
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 """
cl = repo.changelog
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 headssum = {}
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 # A. Create set of branches involved in the push.
branches = set(repo[n].branch() for n in outgoing.missing)
remotemap = remote.branchmap()
newbranches = branches - set(remotemap)
branches.difference_update(newbranches)
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 # A. register remote heads
remotebranches = set()
for branch, heads in remote.branchmap().iteritems():
remotebranches.add(branch)
known = []
unsynced = []
Pierre-Yves David
discovery: stop using nodemap for membership testing...
r20225 knownnode = cl.hasnode # do not use nodemap until it is filtered
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 for h in heads:
Pierre-Yves David
discovery: stop using nodemap for membership testing...
r20225 if knownnode(h):
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 known.append(h)
else:
unsynced.append(h)
headssum[branch] = (known, list(known), unsynced)
# B. add new branch data
missingctx = list(repo[n] for n in outgoing.missing)
touchedbranches = set()
for ctx in missingctx:
branch = ctx.branch()
touchedbranches.add(branch)
if branch not in headssum:
headssum[branch] = (None, [], [])
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 # C drop data about untouched branches:
for branch in remotebranches - touchedbranches:
del headssum[branch]
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 # D. Update newmap with outgoing changes.
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 # This will possibly add new heads and remove existing ones.
Pierre-Yves David
branchmap: store branchcache in a dedicated object...
r18124 newmap = branchmap.branchcache((branch, heads[1])
for branch, heads in headssum.iteritems()
if heads[0] is not None)
Pierre-Yves David
branchmap: pass revision insteads of changectx to the update function...
r18305 newmap.update(repo, (ctx.rev() for ctx in missingctx))
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 for branch, newheads in newmap.iteritems():
headssum[branch][1][:] = newheads
return headssum
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 def _oldheadssummary(repo, remoteheads, outgoing, inc=False):
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 """Compute branchmapsummary for repo without branchmap support"""
# 1-4b. old servers: Check for new topological heads.
# Construct {old,new}map with branch = None (topological branch).
Pierre-Yves David
branchmap: extract _updatebranchcache from repo
r18120 # (code based on update)
Pierre-Yves David
discovery: stop using nodemap for membership testing...
r20225 knownnode = repo.changelog.hasnode # no nodemap until it is filtered
oldheads = set(h for h in remoteheads if knownnode(h))
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 # all nodes in outgoing.missing are children of either:
# - an element of oldheads
# - another element of outgoing.missing
# - nullrev
# This explains why the new head are very simple to compute.
r = repo.set('heads(%ln + %ln)', oldheads, outgoing.missing)
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 newheads = list(c.node() for c in r)
Pierre-Yves David
discovery: prevent crash on unknown remote heads with old repo (issue4337)...
r22178 # set some unsynced head to issue the "unsynced changes" warning
Jordi Gutiérrez Hermoso
style: kill ersatz if-else ternary operators...
r24306 if inc:
unsynced = set([None])
else:
unsynced = set()
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 return {None: (oldheads, newheads, unsynced)}
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209
Stephen Lee
bookmarks: allow push -B to create a new remote head (issue2372)...
r20184 def checkheads(repo, remote, outgoing, remoteheads, newbranch=False, inc=False,
newbookmarks=[]):
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 """Check that a push won't add any outgoing head
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 raise Abort error and display ui message as needed.
"""
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 # Check for each named branch if we're creating new remote heads.
# To be a remote head after push, node must be either:
# - unknown locally
# - a local outgoing head descended from update
# - a remote head that's known locally and not
# ancestral to an outgoing head
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 if remoteheads == [nullid]:
# remote is empty, nothing to check.
return
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 if remote.capable('branchmap'):
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 headssum = _headssummary(repo, remote, outgoing)
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 else:
Pierre-Yves David
checkheads: simplify the structure build by preprocessing...
r17211 headssum = _oldheadssummary(repo, remoteheads, outgoing, inc)
newbranches = [branch for branch, heads in headssum.iteritems()
if heads[0] is None]
Pierre-Yves David
checkheads: extract branchmap preprocessing...
r17209 # 1. Check for new branches on the remote.
if newbranches and not newbranch: # new branch requires --new-branch
branchnames = ', '.join(sorted(newbranches))
raise util.Abort(_("push creates new remote branches: %s!")
% branchnames,
hint=_("use 'hg push --new-branch' to create"
" new remote branches"))
Dirkjan Ochtman
move discovery methods from localrepo into new discovery module
r11301
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 # 2. Compute newly pushed bookmarks. We don't warn about bookmarked heads.
Pierre-Yves David
checkheads: extract bookmark computation from the branch loop...
r17212 localbookmarks = repo._bookmarks
remotebookmarks = remote.listkeys('bookmarks')
bookmarkedheads = set()
for bm in localbookmarks:
rnode = remotebookmarks.get(bm)
if rnode and rnode in repo:
lctx, rctx = repo[bm], repo[rnode]
Pierre-Yves David
bookmarks: extract valid destination logic in a dedicated function...
r17550 if bookmarks.validdest(repo, rctx, lctx):
Pierre-Yves David
checkheads: extract bookmark computation from the branch loop...
r17212 bookmarkedheads.add(lctx.node())
Stephen Lee
bookmarks: allow push -B to create a new remote head (issue2372)...
r20184 else:
if bm in newbookmarks:
bookmarkedheads.add(repo[bm].node())
Pierre-Yves David
checkheads: extract bookmark computation from the branch loop...
r17212
# 3. Check for new heads.
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 # If there are more heads after the push than before, a suitable
# error message, depending on unsynced status, is displayed.
error = None
Michael O'Connor
discovery: don't compute allfuturecommon when it won't be used...
r24702 # If there is no obsstore, allfuturecommon won't be used, so no
# need to compute it.
if repo.obsstore:
allmissing = set(outgoing.missing)
cctx = repo.set('%ld', outgoing.common)
allfuturecommon = set(c.node() for c in cctx)
allfuturecommon.update(allmissing)
Mads Kiilerich
discovery: process heads in sorted order
r18361 for branch, heads in sorted(headssum.iteritems()):
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 remoteheads, newheads, unsyncedheads = heads
candidate_newhs = set(newheads)
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 # add unsynced data
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 if remoteheads is None:
FUJIWARA Katsunori
discovery: abort also when pushing multiple headed new branch...
r19840 oldhs = set()
else:
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 oldhs = set(remoteheads)
oldhs.update(unsyncedheads)
candidate_newhs.update(unsyncedheads)
dhs = None # delta heads, the new heads on branch
Pierre-Yves David
checkheads: don't warn about unsynced changes that we ill obsolete...
r17549 discardedheads = set()
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 if repo.obsstore:
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 # remove future heads which are actually obsoleted by another
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 # pushed element:
#
Pierre-Yves David
checkheads: don't warn about unsynced changes that we ill obsolete...
r17549 # XXX as above, There are several cases this case does not handle
# XXX properly
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 #
# (1) if <nh> is public, it won't be affected by obsolete marker
# and a new is created
#
# (2) if the new heads have ancestors which are not obsolete and
# not ancestors of any other heads we will have a new head too.
#
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 # These two cases will be easy to handle for known changeset but
# much more tricky for unsynced changes.
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 newhs = set()
for nh in candidate_newhs:
Pierre-Yves David
checkheads: check successors for new heads in both missing and common...
r17548 if nh in repo and repo[nh].phase() <= phases.public:
Pierre-Yves David
checkheads: attend to phases when computing new heads with obsolete...
r17547 newhs.add(nh)
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 else:
Pierre-Yves David
obsolete: have `allsuccessors` takes a list of nodes...
r17827 for suc in obsolete.allsuccessors(repo.obsstore, [nh]):
Pierre-Yves David
checkheads: check successors for new heads in both missing and common...
r17548 if suc != nh and suc in allfuturecommon:
Pierre-Yves David
checkheads: don't warn about unsynced changes that we ill obsolete...
r17549 discardedheads.add(nh)
Pierre-Yves David
checkheads: attend to phases when computing new heads with obsolete...
r17547 break
else:
newhs.add(nh)
Pierre-Yves David
checkheads: take future obsoleted heads into account...
r17214 else:
newhs = candidate_newhs
Mads Kiilerich
discovery: improve "note: unsynced remote changes!" warning...
r20501 unsynced = sorted(h for h in unsyncedheads if h not in discardedheads)
if unsynced:
Pierre-Yves David
discovery: prevent crash on unknown remote heads with old repo (issue4337)...
r22178 if None in unsynced:
# old remote, no heads data
heads = None
elif len(unsynced) <= 4 or repo.ui.verbose:
Mads Kiilerich
discovery: don't report all "unsynced" remote heads (issue4230)...
r21198 heads = ' '.join(short(h) for h in unsynced)
else:
heads = (' '.join(short(h) for h in unsynced[:4]) +
' ' + _("and %s others") % (len(unsynced) - 4))
Pierre-Yves David
discovery: prevent crash on unknown remote heads with old repo (issue4337)...
r22178 if heads is None:
repo.ui.status(_("remote has heads that are "
"not known locally\n"))
elif branch is None:
Mads Kiilerich
discovery: make "note: unsynced remote changes!" less serious than a warning...
r20502 repo.ui.status(_("remote has heads that are "
"not known locally: %s\n") % heads)
Mads Kiilerich
discovery: improve "note: unsynced remote changes!" warning...
r20501 else:
Mads Kiilerich
discovery: make "note: unsynced remote changes!" less serious than a warning...
r20502 repo.ui.status(_("remote has heads on branch '%s' that are "
"not known locally: %s\n") % (branch, heads))
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 if remoteheads is None:
if len(newhs) > 1:
FUJIWARA Katsunori
discovery: abort also when pushing multiple headed new branch...
r19840 dhs = list(newhs)
if error is None:
Mads Kiilerich
discovery: tweak error message for multiple branch heads
r20050 error = (_("push creates new branch '%s' "
"with multiple heads") % (branch))
FUJIWARA Katsunori
discovery: abort also when pushing multiple headed new branch...
r19840 hint = _("merge or"
FUJIWARA Katsunori
discovery: revise hint message introduced by changeset b00ba31313c3...
r19936 " see \"hg help push\" for details about"
FUJIWARA Katsunori
discovery: abort also when pushing multiple headed new branch...
r19840 " pushing new heads")
elif len(newhs) > len(oldhs):
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 # remove bookmarked or existing remote heads from the new heads list
Mads Kiilerich
discovery: process heads in sorted order
r18361 dhs = sorted(newhs - bookmarkedheads - oldhs)
Levi Bard
bookmarks: allow existing remote bookmarks to become heads when pushing
r16835 if dhs:
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 if error is None:
if branch not in ('default', None):
error = _("push creates new remote head %s "
"on branch '%s'!") % (short(dhs[0]), branch)
Stephen Lee
discovery: if a push would create a new head, mention the bookmark name if any
r21580 elif repo[dhs[0]].bookmarks():
error = _("push creates new remote head %s "
"with bookmark '%s'!") % (
short(dhs[0]), repo[dhs[0]].bookmarks()[0])
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 else:
error = _("push creates new remote head %s!"
) % short(dhs[0])
Mads Kiilerich
discovery: cleanup of variable names and comments
r20381 if unsyncedheads:
FUJIWARA Katsunori
push: hide description about "-f" in the hint to prevent from using it easily...
r19934 hint = _("pull and merge or"
" see \"hg help push\" for details about"
" pushing new heads")
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 else:
FUJIWARA Katsunori
push: hide description about "-f" in the hint to prevent from using it easily...
r19934 hint = _("merge or"
" see \"hg help push\" for details about"
" pushing new heads")
Mads Kiilerich
discovery: make note messages for new heads more readable
r20051 if branch is None:
repo.ui.note(_("new remote heads:\n"))
else:
repo.ui.note(_("new remote heads on branch '%s':\n") % branch)
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 for h in dhs:
Mads Kiilerich
discovery: make note messages for new heads more readable
r20051 repo.ui.note((" %s\n") % short(h))
Pierre-Yves David
discovery: diet discovery.prepush from non-discovery code...
r15932 if error:
raise util.Abort(error, hint=hint)