obsutil.py
977 lines
| 34.7 KiB
| text/x-python
|
PythonLexer
/ mercurial / obsutil.py
Boris Feld
|
r32879 | # obsutil.py - utility functions for obsolescence | ||
# | ||||
# Copyright 2017 Boris Feld <boris.feld@octobus.net> | ||||
# | ||||
# This software may be used and distributed according to the terms of the | ||||
# GNU General Public License version 2 or any later version. | ||||
from __future__ import absolute_import | ||||
Boris Feld
|
r34421 | import re | ||
Boris Feld
|
r35589 | from .i18n import _ | ||
r33249 | from . import ( | |||
Boris Feld
|
r35589 | node as nodemod, | ||
r33249 | phases, | |||
Boris Feld
|
r35589 | util, | ||
r33249 | ) | |||
Boris Feld
|
r36625 | from .utils import dateutil | ||
r33249 | ||||
r36971 | ### obsolescence marker flag | |||
## bumpedfix flag | ||||
# | ||||
# When a changeset A' succeed to a changeset A which became public, we call A' | ||||
# "bumped" because it's a successors of a public changesets | ||||
# | ||||
# o A' (bumped) | ||||
# |`: | ||||
# | o A | ||||
# |/ | ||||
# o Z | ||||
# | ||||
# The way to solve this situation is to create a new changeset Ad as children | ||||
# of A. This changeset have the same content than A'. So the diff from A to A' | ||||
# is the same than the diff from A to Ad. Ad is marked as a successors of A' | ||||
# | ||||
# o Ad | ||||
# |`: | ||||
# | x A' | ||||
# |'| | ||||
# o | A | ||||
# |/ | ||||
# o Z | ||||
# | ||||
# But by transitivity Ad is also a successors of A. To avoid having Ad marked | ||||
# as bumped too, we add the `bumpedfix` flag to the marker. <A', (Ad,)>. | ||||
# This flag mean that the successors express the changes between the public and | ||||
# bumped version and fix the situation, breaking the transitivity of | ||||
# "bumped" here. | ||||
bumpedfix = 1 | ||||
usingsha256 = 2 | ||||
r33149 | class marker(object): | |||
"""Wrap obsolete marker raw data""" | ||||
def __init__(self, repo, data): | ||||
# the repo argument will be used to create changectx in later version | ||||
self._repo = repo | ||||
self._data = data | ||||
self._decodedmeta = None | ||||
def __hash__(self): | ||||
return hash(self._data) | ||||
def __eq__(self, other): | ||||
if type(other) != type(self): | ||||
return False | ||||
return self._data == other._data | ||||
Boris Feld
|
r33697 | def prednode(self): | ||
"""Predecessor changeset node identifier""" | ||||
r33149 | return self._data[0] | |||
def succnodes(self): | ||||
"""List of successor changesets node identifiers""" | ||||
return self._data[1] | ||||
def parentnodes(self): | ||||
Boris Feld
|
r33697 | """Parents of the predecessors (None if not recorded)""" | ||
r33149 | return self._data[5] | |||
def metadata(self): | ||||
"""Decoded metadata dictionary""" | ||||
return dict(self._data[3]) | ||||
def date(self): | ||||
"""Creation date as (unixtime, offset)""" | ||||
return self._data[4] | ||||
def flags(self): | ||||
"""The flags field of the marker""" | ||||
return self._data[2] | ||||
r33150 | def getmarkers(repo, nodes=None, exclusive=False): | |||
"""returns markers known in a repository | ||||
If <nodes> is specified, only markers "relevant" to those nodes are are | ||||
returned""" | ||||
if nodes is None: | ||||
rawmarkers = repo.obsstore | ||||
elif exclusive: | ||||
rawmarkers = exclusivemarkers(repo, nodes) | ||||
else: | ||||
rawmarkers = repo.obsstore.relevantmarkers(nodes) | ||||
for markerdata in rawmarkers: | ||||
yield marker(repo, markerdata) | ||||
Boris Feld
|
r32879 | def closestpredecessors(repo, nodeid): | ||
"""yield the list of next predecessors pointing on visible changectx nodes | ||||
This function respect the repoview filtering, filtered revision will be | ||||
considered missing. | ||||
""" | ||||
Boris Feld
|
r33699 | precursors = repo.obsstore.predecessors | ||
Boris Feld
|
r32879 | stack = [nodeid] | ||
seen = set(stack) | ||||
while stack: | ||||
current = stack.pop() | ||||
currentpreccs = precursors.get(current, ()) | ||||
for prec in currentpreccs: | ||||
precnodeid = prec[0] | ||||
# Basic cycle protection | ||||
if precnodeid in seen: | ||||
continue | ||||
seen.add(precnodeid) | ||||
if precnodeid in repo: | ||||
yield precnodeid | ||||
else: | ||||
stack.append(precnodeid) | ||||
r33143 | ||||
Boris Feld
|
r33701 | def allpredecessors(obsstore, nodes, ignoreflags=0): | ||
r33145 | """Yield node for every precursors of <nodes>. | |||
Some precursors may be unknown locally. | ||||
This is a linear yield unsuited to detecting folded changesets. It includes | ||||
initial nodes too.""" | ||||
remaining = set(nodes) | ||||
seen = set(remaining) | ||||
while remaining: | ||||
current = remaining.pop() | ||||
yield current | ||||
Boris Feld
|
r33699 | for mark in obsstore.predecessors.get(current, ()): | ||
r33145 | # ignore marker flagged with specified flag | |||
if mark[2] & ignoreflags: | ||||
continue | ||||
suc = mark[0] | ||||
if suc not in seen: | ||||
seen.add(suc) | ||||
remaining.add(suc) | ||||
r33146 | def allsuccessors(obsstore, nodes, ignoreflags=0): | |||
"""Yield node for every successor of <nodes>. | ||||
Some successors may be unknown locally. | ||||
This is a linear yield unsuited to detecting split changesets. It includes | ||||
initial nodes too.""" | ||||
remaining = set(nodes) | ||||
seen = set(remaining) | ||||
while remaining: | ||||
current = remaining.pop() | ||||
yield current | ||||
for mark in obsstore.successors.get(current, ()): | ||||
# ignore marker flagged with specified flag | ||||
if mark[2] & ignoreflags: | ||||
continue | ||||
for suc in mark[1]: | ||||
if suc not in seen: | ||||
seen.add(suc) | ||||
remaining.add(suc) | ||||
r33144 | def _filterprunes(markers): | |||
"""return a set with no prune markers""" | ||||
return set(m for m in markers if m[1]) | ||||
def exclusivemarkers(repo, nodes): | ||||
"""set of markers relevant to "nodes" but no other locally-known nodes | ||||
This function compute the set of markers "exclusive" to a locally-known | ||||
node. This means we walk the markers starting from <nodes> until we reach a | ||||
locally-known precursors outside of <nodes>. Element of <nodes> with | ||||
locally-known successors outside of <nodes> are ignored (since their | ||||
precursors markers are also relevant to these successors). | ||||
For example: | ||||
# (A0 rewritten as A1) | ||||
# | ||||
# A0 <-1- A1 # Marker "1" is exclusive to A1 | ||||
or | ||||
# (A0 rewritten as AX; AX rewritten as A1; AX is unkown locally) | ||||
# | ||||
# <-1- A0 <-2- AX <-3- A1 # Marker "2,3" are exclusive to A1 | ||||
or | ||||
# (A0 has unknown precursors, A0 rewritten as A1 and A2 (divergence)) | ||||
# | ||||
# <-2- A1 # Marker "2" is exclusive to A0,A1 | ||||
# / | ||||
# <-1- A0 | ||||
# \ | ||||
# <-3- A2 # Marker "3" is exclusive to A0,A2 | ||||
# | ||||
# in addition: | ||||
# | ||||
# Markers "2,3" are exclusive to A1,A2 | ||||
# Markers "1,2,3" are exclusive to A0,A1,A2 | ||||
See test/test-obsolete-bundle-strip.t for more examples. | ||||
An example usage is strip. When stripping a changeset, we also want to | ||||
strip the markers exclusive to this changeset. Otherwise we would have | ||||
"dangling"" obsolescence markers from its precursors: Obsolescence markers | ||||
marking a node as obsolete without any successors available locally. | ||||
As for relevant markers, the prune markers for children will be followed. | ||||
Of course, they will only be followed if the pruned children is | ||||
locally-known. Since the prune markers are relevant to the pruned node. | ||||
However, while prune markers are considered relevant to the parent of the | ||||
pruned changesets, prune markers for locally-known changeset (with no | ||||
successors) are considered exclusive to the pruned nodes. This allows | ||||
to strip the prune markers (with the rest of the exclusive chain) alongside | ||||
the pruned changesets. | ||||
""" | ||||
# running on a filtered repository would be dangerous as markers could be | ||||
# reported as exclusive when they are relevant for other filtered nodes. | ||||
unfi = repo.unfiltered() | ||||
# shortcut to various useful item | ||||
nm = unfi.changelog.nodemap | ||||
Boris Feld
|
r33699 | precursorsmarkers = unfi.obsstore.predecessors | ||
r33144 | successormarkers = unfi.obsstore.successors | |||
childrenmarkers = unfi.obsstore.children | ||||
# exclusive markers (return of the function) | ||||
exclmarkers = set() | ||||
# we need fast membership testing | ||||
nodes = set(nodes) | ||||
# looking for head in the obshistory | ||||
# | ||||
# XXX we are ignoring all issues in regard with cycle for now. | ||||
stack = [n for n in nodes if not _filterprunes(successormarkers.get(n, ()))] | ||||
stack.sort() | ||||
# nodes already stacked | ||||
seennodes = set(stack) | ||||
while stack: | ||||
current = stack.pop() | ||||
# fetch precursors markers | ||||
markers = list(precursorsmarkers.get(current, ())) | ||||
# extend the list with prune markers | ||||
for mark in successormarkers.get(current, ()): | ||||
if not mark[1]: | ||||
markers.append(mark) | ||||
# and markers from children (looking for prune) | ||||
for mark in childrenmarkers.get(current, ()): | ||||
if not mark[1]: | ||||
markers.append(mark) | ||||
# traverse the markers | ||||
for mark in markers: | ||||
if mark in exclmarkers: | ||||
# markers already selected | ||||
continue | ||||
# If the markers is about the current node, select it | ||||
# | ||||
# (this delay the addition of markers from children) | ||||
if mark[1] or mark[0] == current: | ||||
exclmarkers.add(mark) | ||||
# should we keep traversing through the precursors? | ||||
prec = mark[0] | ||||
# nodes in the stack or already processed | ||||
if prec in seennodes: | ||||
continue | ||||
# is this a locally known node ? | ||||
known = prec in nm | ||||
# if locally-known and not in the <nodes> set the traversal | ||||
# stop here. | ||||
if known and prec not in nodes: | ||||
continue | ||||
# do not keep going if there are unselected markers pointing to this | ||||
# nodes. If we end up traversing these unselected markers later the | ||||
# node will be taken care of at that point. | ||||
precmarkers = _filterprunes(successormarkers.get(prec)) | ||||
if precmarkers.issubset(exclmarkers): | ||||
seennodes.add(prec) | ||||
stack.append(prec) | ||||
return exclmarkers | ||||
r33147 | def foreground(repo, nodes): | |||
"""return all nodes in the "foreground" of other node | ||||
The foreground of a revision is anything reachable using parent -> children | ||||
or precursor -> successor relation. It is very similar to "descendant" but | ||||
augmented with obsolescence information. | ||||
Beware that possible obsolescence cycle may result if complex situation. | ||||
""" | ||||
repo = repo.unfiltered() | ||||
foreground = set(repo.set('%ln::', nodes)) | ||||
if repo.obsstore: | ||||
# We only need this complicated logic if there is obsolescence | ||||
# XXX will probably deserve an optimised revset. | ||||
nm = repo.changelog.nodemap | ||||
plen = -1 | ||||
# compute the whole set of successors or descendants | ||||
while len(foreground) != plen: | ||||
plen = len(foreground) | ||||
succs = set(c.node() for c in foreground) | ||||
mutable = [c.node() for c in foreground if c.mutable()] | ||||
succs.update(allsuccessors(repo.obsstore, mutable)) | ||||
known = (n for n in succs if n in nm) | ||||
foreground = set(repo.set('%ln::', known)) | ||||
return set(c.node() for c in foreground) | ||||
Boris Feld
|
r34423 | # effectflag field | ||
# | ||||
# Effect-flag is a 1-byte bit field used to store what changed between a | ||||
# changeset and its successor(s). | ||||
# | ||||
# The effect flag is stored in obs-markers metadata while we iterate on the | ||||
# information design. That's why we have the EFFECTFLAGFIELD. If we come up | ||||
# with an incompatible design for effect flag, we can store a new design under | ||||
# another field name so we don't break readers. We plan to extend the existing | ||||
# obsmarkers bit-field when the effect flag design will be stabilized. | ||||
# | ||||
# The effect-flag is placed behind an experimental flag | ||||
# `effect-flags` set to off by default. | ||||
# | ||||
Boris Feld
|
r34414 | EFFECTFLAGFIELD = "ef1" | ||
Boris Feld
|
r34416 | DESCCHANGED = 1 << 0 # action changed the description | ||
Boris Feld
|
r34421 | METACHANGED = 1 << 1 # action change the meta | ||
Boris Feld
|
r34422 | DIFFCHANGED = 1 << 3 # action change diff introduced by the changeset | ||
Boris Feld
|
r34420 | PARENTCHANGED = 1 << 2 # action change the parent | ||
Boris Feld
|
r34417 | USERCHANGED = 1 << 4 # the user changed | ||
Boris Feld
|
r34418 | DATECHANGED = 1 << 5 # the date changed | ||
Boris Feld
|
r34419 | BRANCHCHANGED = 1 << 6 # the branch changed | ||
Boris Feld
|
r34416 | |||
Boris Feld
|
r34421 | METABLACKLIST = [ | ||
re.compile('^branch$'), | ||||
re.compile('^.*-source$'), | ||||
re.compile('^.*_source$'), | ||||
re.compile('^source$'), | ||||
] | ||||
def metanotblacklisted(metaitem): | ||||
""" Check that the key of a meta item (extrakey, extravalue) does not | ||||
match at least one of the blacklist pattern | ||||
""" | ||||
metakey = metaitem[0] | ||||
return not any(pattern.match(metakey) for pattern in METABLACKLIST) | ||||
Boris Feld
|
r34422 | def _prepare_hunk(hunk): | ||
"""Drop all information but the username and patch""" | ||||
cleanhunk = [] | ||||
for line in hunk.splitlines(): | ||||
if line.startswith(b'# User') or not line.startswith(b'#'): | ||||
if line.startswith(b'@@'): | ||||
line = b'@@\n' | ||||
cleanhunk.append(line) | ||||
return cleanhunk | ||||
def _getdifflines(iterdiff): | ||||
"""return a cleaned up lines""" | ||||
lines = next(iterdiff, None) | ||||
if lines is None: | ||||
return lines | ||||
return _prepare_hunk(lines) | ||||
def _cmpdiff(leftctx, rightctx): | ||||
"""return True if both ctx introduce the "same diff" | ||||
This is a first and basic implementation, with many shortcoming. | ||||
""" | ||||
# Leftctx or right ctx might be filtered, so we need to use the contexts | ||||
# with an unfiltered repository to safely compute the diff | ||||
leftunfi = leftctx._repo.unfiltered()[leftctx.rev()] | ||||
leftdiff = leftunfi.diff(git=1) | ||||
rightunfi = rightctx._repo.unfiltered()[rightctx.rev()] | ||||
rightdiff = rightunfi.diff(git=1) | ||||
left, right = (0, 0) | ||||
while None not in (left, right): | ||||
left = _getdifflines(leftdiff) | ||||
right = _getdifflines(rightdiff) | ||||
if left != right: | ||||
return False | ||||
return True | ||||
Boris Feld
|
r34414 | def geteffectflag(relation): | ||
""" From an obs-marker relation, compute what changed between the | ||||
predecessor and the successor. | ||||
""" | ||||
effects = 0 | ||||
source = relation[0] | ||||
Boris Feld
|
r34416 | for changectx in relation[1]: | ||
# Check if description has changed | ||||
if changectx.description() != source.description(): | ||||
effects |= DESCCHANGED | ||||
Boris Feld
|
r34417 | # Check if user has changed | ||
if changectx.user() != source.user(): | ||||
effects |= USERCHANGED | ||||
Boris Feld
|
r34418 | # Check if date has changed | ||
if changectx.date() != source.date(): | ||||
effects |= DATECHANGED | ||||
Boris Feld
|
r34419 | # Check if branch has changed | ||
if changectx.branch() != source.branch(): | ||||
effects |= BRANCHCHANGED | ||||
Boris Feld
|
r34420 | # Check if at least one of the parent has changed | ||
if changectx.parents() != source.parents(): | ||||
effects |= PARENTCHANGED | ||||
Boris Feld
|
r34421 | # Check if other meta has changed | ||
changeextra = changectx.extra().items() | ||||
Augie Fackler
|
r35893 | ctxmeta = list(filter(metanotblacklisted, changeextra)) | ||
Boris Feld
|
r34421 | |||
sourceextra = source.extra().items() | ||||
Augie Fackler
|
r35893 | srcmeta = list(filter(metanotblacklisted, sourceextra)) | ||
Boris Feld
|
r34421 | |||
if ctxmeta != srcmeta: | ||||
effects |= METACHANGED | ||||
Boris Feld
|
r34422 | # Check if the diff has changed | ||
if not _cmpdiff(source, changectx): | ||||
effects |= DIFFCHANGED | ||||
Boris Feld
|
r34414 | return effects | ||
r33249 | def getobsoleted(repo, tr): | |||
"""return the set of pre-existing revisions obsoleted by a transaction""" | ||||
torev = repo.unfiltered().changelog.nodemap.get | ||||
phase = repo._phasecache.phase | ||||
succsmarkers = repo.obsstore.successors.get | ||||
public = phases.public | ||||
addedmarkers = tr.changes.get('obsmarkers') | ||||
addedrevs = tr.changes.get('revs') | ||||
Joerg Sonnenberger
|
r35309 | seenrevs = set() | ||
r33249 | obsoleted = set() | |||
for mark in addedmarkers: | ||||
node = mark[0] | ||||
rev = torev(node) | ||||
Joerg Sonnenberger
|
r35309 | if rev is None or rev in seenrevs or rev in addedrevs: | ||
r33249 | continue | |||
seenrevs.add(rev) | ||||
if phase(repo, rev) == public: | ||||
continue | ||||
Augie Fackler
|
r33679 | if set(succsmarkers(node) or []).issubset(addedmarkers): | ||
r33249 | obsoleted.add(rev) | |||
return obsoleted | ||||
Boris Feld
|
r33910 | class _succs(list): | ||
"""small class to represent a successors with some metadata about it""" | ||||
Boris Feld
|
r33912 | def __init__(self, *args, **kwargs): | ||
super(_succs, self).__init__(*args, **kwargs) | ||||
self.markers = set() | ||||
Boris Feld
|
r33911 | def copy(self): | ||
Boris Feld
|
r33912 | new = _succs(self) | ||
new.markers = self.markers.copy() | ||||
return new | ||||
Boris Feld
|
r33911 | |||
Boris Feld
|
r33943 | @util.propertycache | ||
def _set(self): | ||||
# immutable | ||||
return set(self) | ||||
def canmerge(self, other): | ||||
return self._set.issubset(other._set) | ||||
Boris Feld
|
r33274 | def successorssets(repo, initialnode, closest=False, cache=None): | ||
r33143 | """Return set of all latest successors of initial nodes | |||
The successors set of a changeset A are the group of revisions that succeed | ||||
A. It succeeds A as a consistent whole, each revision being only a partial | ||||
Boris Feld
|
r33274 | replacement. By default, the successors set contains non-obsolete | ||
changesets only, walking the obsolescence graph until reaching a leaf. If | ||||
'closest' is set to True, closest successors-sets are return (the | ||||
obsolescence walk stops on known changesets). | ||||
r33143 | ||||
This function returns the full list of successor sets which is why it | ||||
returns a list of tuples and not just a single tuple. Each tuple is a valid | ||||
successors set. Note that (A,) may be a valid successors set for changeset A | ||||
(see below). | ||||
In most cases, a changeset A will have a single element (e.g. the changeset | ||||
A is replaced by A') in its successors set. Though, it is also common for a | ||||
changeset A to have no elements in its successor set (e.g. the changeset | ||||
has been pruned). Therefore, the returned list of successors sets will be | ||||
[(A',)] or [], respectively. | ||||
When a changeset A is split into A' and B', however, it will result in a | ||||
successors set containing more than a single element, i.e. [(A',B')]. | ||||
Divergent changesets will result in multiple successors sets, i.e. [(A',), | ||||
(A'')]. | ||||
If a changeset A is not obsolete, then it will conceptually have no | ||||
successors set. To distinguish this from a pruned changeset, the successor | ||||
set will contain itself only, i.e. [(A,)]. | ||||
Boris Feld
|
r33272 | Finally, final successors unknown locally are considered to be pruned | ||
(pruned: obsoleted without any successors). (Final: successors not affected | ||||
by markers). | ||||
r33143 | ||||
Boris Feld
|
r33274 | The 'closest' mode respect the repoview filtering. For example, without | ||
filter it will stop at the first locally known changeset, with 'visible' | ||||
filter it will stop on visible changesets). | ||||
Boris Feld
|
r33272 | The optional `cache` parameter is a dictionary that may contains | ||
precomputed successors sets. It is meant to reuse the computation of a | ||||
previous call to `successorssets` when multiple calls are made at the same | ||||
time. The cache dictionary is updated in place. The caller is responsible | ||||
for its life span. Code that makes multiple calls to `successorssets` | ||||
*should* use this cache mechanism or risk a performance hit. | ||||
Boris Feld
|
r33274 | |||
Since results are different depending of the 'closest' most, the same cache | ||||
cannot be reused for both mode. | ||||
r33143 | """ | |||
succmarkers = repo.obsstore.successors | ||||
# Stack of nodes we search successors sets for | ||||
toproceed = [initialnode] | ||||
# set version of above list for fast loop detection | ||||
# element added to "toproceed" must be added here | ||||
stackedset = set(toproceed) | ||||
if cache is None: | ||||
cache = {} | ||||
# This while loop is the flattened version of a recursive search for | ||||
# successors sets | ||||
# | ||||
# def successorssets(x): | ||||
# successors = directsuccessors(x) | ||||
# ss = [[]] | ||||
# for succ in directsuccessors(x): | ||||
# # product as in itertools cartesian product | ||||
# ss = product(ss, successorssets(succ)) | ||||
# return ss | ||||
# | ||||
# But we can not use plain recursive calls here: | ||||
# - that would blow the python call stack | ||||
# - obsolescence markers may have cycles, we need to handle them. | ||||
# | ||||
# The `toproceed` list act as our call stack. Every node we search | ||||
# successors set for are stacked there. | ||||
# | ||||
# The `stackedset` is set version of this stack used to check if a node is | ||||
# already stacked. This check is used to detect cycles and prevent infinite | ||||
# loop. | ||||
# | ||||
# successors set of all nodes are stored in the `cache` dictionary. | ||||
# | ||||
# After this while loop ends we use the cache to return the successors sets | ||||
# for the node requested by the caller. | ||||
while toproceed: | ||||
# Every iteration tries to compute the successors sets of the topmost | ||||
# node of the stack: CURRENT. | ||||
# | ||||
# There are four possible outcomes: | ||||
# | ||||
# 1) We already know the successors sets of CURRENT: | ||||
# -> mission accomplished, pop it from the stack. | ||||
Boris Feld
|
r33274 | # 2) Stop the walk: | ||
# default case: Node is not obsolete | ||||
# closest case: Node is known at this repo filter level | ||||
# -> the node is its own successors sets. Add it to the cache. | ||||
r33143 | # 3) We do not know successors set of direct successors of CURRENT: | |||
# -> We add those successors to the stack. | ||||
# 4) We know successors sets of all direct successors of CURRENT: | ||||
# -> We can compute CURRENT successors set and add it to the | ||||
# cache. | ||||
# | ||||
current = toproceed[-1] | ||||
Boris Feld
|
r33274 | |||
# case 2 condition is a bit hairy because of closest, | ||||
# we compute it on its own | ||||
case2condition = ((current not in succmarkers) | ||||
or (closest and current != initialnode | ||||
and current in repo)) | ||||
r33143 | if current in cache: | |||
# case (1): We already know the successors sets | ||||
stackedset.remove(toproceed.pop()) | ||||
Boris Feld
|
r33274 | elif case2condition: | ||
# case (2): end of walk. | ||||
r33143 | if current in repo: | |||
Boris Feld
|
r33274 | # We have a valid successors. | ||
Boris Feld
|
r33910 | cache[current] = [_succs((current,))] | ||
r33143 | else: | |||
# Final obsolete version is unknown locally. | ||||
# Do not count that as a valid successors | ||||
cache[current] = [] | ||||
else: | ||||
# cases (3) and (4) | ||||
# | ||||
# We proceed in two phases. Phase 1 aims to distinguish case (3) | ||||
# from case (4): | ||||
# | ||||
# For each direct successors of CURRENT, we check whether its | ||||
# successors sets are known. If they are not, we stack the | ||||
# unknown node and proceed to the next iteration of the while | ||||
# loop. (case 3) | ||||
# | ||||
# During this step, we may detect obsolescence cycles: a node | ||||
# with unknown successors sets but already in the call stack. | ||||
# In such a situation, we arbitrary set the successors sets of | ||||
# the node to nothing (node pruned) to break the cycle. | ||||
# | ||||
# If no break was encountered we proceed to phase 2. | ||||
# | ||||
# Phase 2 computes successors sets of CURRENT (case 4); see details | ||||
# in phase 2 itself. | ||||
# | ||||
# Note the two levels of iteration in each phase. | ||||
# - The first one handles obsolescence markers using CURRENT as | ||||
# precursor (successors markers of CURRENT). | ||||
# | ||||
# Having multiple entry here means divergence. | ||||
# | ||||
# - The second one handles successors defined in each marker. | ||||
# | ||||
# Having none means pruned node, multiple successors means split, | ||||
# single successors are standard replacement. | ||||
# | ||||
for mark in sorted(succmarkers[current]): | ||||
for suc in mark[1]: | ||||
if suc not in cache: | ||||
if suc in stackedset: | ||||
# cycle breaking | ||||
cache[suc] = [] | ||||
else: | ||||
# case (3) If we have not computed successors sets | ||||
# of one of those successors we add it to the | ||||
# `toproceed` stack and stop all work for this | ||||
# iteration. | ||||
toproceed.append(suc) | ||||
stackedset.add(suc) | ||||
break | ||||
else: | ||||
continue | ||||
break | ||||
else: | ||||
# case (4): we know all successors sets of all direct | ||||
# successors | ||||
# | ||||
# Successors set contributed by each marker depends on the | ||||
# successors sets of all its "successors" node. | ||||
# | ||||
# Each different marker is a divergence in the obsolescence | ||||
# history. It contributes successors sets distinct from other | ||||
# markers. | ||||
# | ||||
# Within a marker, a successor may have divergent successors | ||||
# sets. In such a case, the marker will contribute multiple | ||||
# divergent successors sets. If multiple successors have | ||||
# divergent successors sets, a Cartesian product is used. | ||||
# | ||||
# At the end we post-process successors sets to remove | ||||
# duplicated entry and successors set that are strict subset of | ||||
# another one. | ||||
succssets = [] | ||||
for mark in sorted(succmarkers[current]): | ||||
# successors sets contributed by this marker | ||||
Boris Feld
|
r33912 | base = _succs() | ||
base.markers.add(mark) | ||||
markss = [base] | ||||
r33143 | for suc in mark[1]: | |||
# cardinal product with previous successors | ||||
productresult = [] | ||||
for prefix in markss: | ||||
for suffix in cache[suc]: | ||||
Boris Feld
|
r33911 | newss = prefix.copy() | ||
Boris Feld
|
r33912 | newss.markers.update(suffix.markers) | ||
r33143 | for part in suffix: | |||
# do not duplicated entry in successors set | ||||
# first entry wins. | ||||
if part not in newss: | ||||
newss.append(part) | ||||
productresult.append(newss) | ||||
markss = productresult | ||||
succssets.extend(markss) | ||||
# remove duplicated and subset | ||||
seen = [] | ||||
final = [] | ||||
Boris Feld
|
r33944 | candidates = sorted((s for s in succssets if s), | ||
key=len, reverse=True) | ||||
for cand in candidates: | ||||
Boris Feld
|
r33943 | for seensuccs in seen: | ||
if cand.canmerge(seensuccs): | ||||
seensuccs.markers.update(cand.markers) | ||||
r33143 | break | |||
else: | ||||
Boris Feld
|
r33943 | final.append(cand) | ||
seen.append(cand) | ||||
r33143 | final.reverse() # put small successors set first | |||
cache[current] = final | ||||
return cache[initialnode] | ||||
Boris Feld
|
r33913 | |||
def successorsandmarkers(repo, ctx): | ||||
"""compute the raw data needed for computing obsfate | ||||
Returns a list of dict, one dict per successors set | ||||
""" | ||||
if not ctx.obsolete(): | ||||
return None | ||||
ssets = successorssets(repo, ctx.node(), closest=True) | ||||
Boris Feld
|
r33998 | # closestsuccessors returns an empty list for pruned revisions, remap it | ||
# into a list containing an empty list for future processing | ||||
if ssets == []: | ||||
ssets = [[]] | ||||
# Try to recover pruned markers | ||||
succsmap = repo.obsstore.successors | ||||
fullsuccessorsets = [] # successor set + markers | ||||
for sset in ssets: | ||||
if sset: | ||||
fullsuccessorsets.append(sset) | ||||
else: | ||||
# successorsset return an empty set() when ctx or one of its | ||||
# successors is pruned. | ||||
# In this case, walk the obs-markers tree again starting with ctx | ||||
# and find the relevant pruning obs-makers, the ones without | ||||
# successors. | ||||
# Having these markers allow us to compute some information about | ||||
# its fate, like who pruned this changeset and when. | ||||
# XXX we do not catch all prune markers (eg rewritten then pruned) | ||||
# (fix me later) | ||||
foundany = False | ||||
for mark in succsmap.get(ctx.node(), ()): | ||||
if not mark[1]: | ||||
foundany = True | ||||
sset = _succs() | ||||
sset.markers.add(mark) | ||||
fullsuccessorsets.append(sset) | ||||
if not foundany: | ||||
fullsuccessorsets.append(_succs()) | ||||
Boris Feld
|
r33913 | values = [] | ||
Boris Feld
|
r33998 | for sset in fullsuccessorsets: | ||
Boris Feld
|
r33913 | values.append({'successors': sset, 'markers': sset.markers}) | ||
return values | ||||
Boris Feld
|
r33995 | |||
Boris Feld
|
r35589 | def _getobsfate(successorssets): | ||
""" Compute a changeset obsolescence fate based on its successorssets. | ||||
Successors can be the tipmost ones or the immediate ones. This function | ||||
return values are not meant to be shown directly to users, it is meant to | ||||
be used by internal functions only. | ||||
Returns one fate from the following values: | ||||
- pruned | ||||
- diverged | ||||
- superseded | ||||
- superseded_split | ||||
""" | ||||
if len(successorssets) == 0: | ||||
# The commit has been pruned | ||||
return 'pruned' | ||||
elif len(successorssets) > 1: | ||||
return 'diverged' | ||||
else: | ||||
# No divergence, only one set of successors | ||||
successors = successorssets[0] | ||||
if len(successors) == 1: | ||||
return 'superseded' | ||||
else: | ||||
return 'superseded_split' | ||||
Boris Feld
|
r35011 | def obsfateverb(successorset, markers): | ||
""" Return the verb summarizing the successorset and potentially using | ||||
information from the markers | ||||
Boris Feld
|
r33995 | """ | ||
if not successorset: | ||||
verb = 'pruned' | ||||
elif len(successorset) == 1: | ||||
verb = 'rewritten' | ||||
else: | ||||
verb = 'split' | ||||
return verb | ||||
Boris Feld
|
r33996 | |||
Boris Feld
|
r33997 | def markersdates(markers): | ||
"""returns the list of dates for a list of markers | ||||
""" | ||||
return [m[4] for m in markers] | ||||
Boris Feld
|
r33996 | def markersusers(markers): | ||
""" Returns a sorted list of markers users without duplicates | ||||
""" | ||||
markersmeta = [dict(m[3]) for m in markers] | ||||
users = set(meta.get('user') for meta in markersmeta if meta.get('user')) | ||||
return sorted(users) | ||||
Martin von Zweigbergk
|
r34288 | |||
def markersoperations(markers): | ||||
""" Returns a sorted list of markers operations without duplicates | ||||
""" | ||||
markersmeta = [dict(m[3]) for m in markers] | ||||
operations = set(meta.get('operation') for meta in markersmeta | ||||
if meta.get('operation')) | ||||
return sorted(operations) | ||||
Boris Feld
|
r34848 | |||
def obsfateprinter(successors, markers, ui): | ||||
""" Build a obsfate string for a single successorset using all obsfate | ||||
related function defined in obsutil | ||||
""" | ||||
Boris Feld
|
r34851 | quiet = ui.quiet | ||
verbose = ui.verbose | ||||
normal = not verbose and not quiet | ||||
Boris Feld
|
r34848 | line = [] | ||
# Verb | ||||
Boris Feld
|
r35011 | line.append(obsfateverb(successors, markers)) | ||
Boris Feld
|
r34848 | |||
# Operations | ||||
operations = markersoperations(markers) | ||||
if operations: | ||||
line.append(" using %s" % ", ".join(operations)) | ||||
# Successors | ||||
if successors: | ||||
fmtsuccessors = [successors.joinfmt(succ) for succ in successors] | ||||
line.append(" as %s" % ", ".join(fmtsuccessors)) | ||||
# Users | ||||
users = markersusers(markers) | ||||
Boris Feld
|
r34851 | # Filter out current user in not verbose mode to reduce amount of | ||
# information | ||||
if not verbose: | ||||
currentuser = ui.username(acceptempty=True) | ||||
if len(users) == 1 and currentuser in users: | ||||
users = None | ||||
Boris Feld
|
r34848 | |||
Boris Feld
|
r34851 | if (verbose or normal) and users: | ||
Boris Feld
|
r34848 | line.append(" by %s" % ", ".join(users)) | ||
# Date | ||||
dates = markersdates(markers) | ||||
Boris Feld
|
r34874 | if dates and verbose: | ||
Boris Feld
|
r34852 | min_date = min(dates) | ||
max_date = max(dates) | ||||
Boris Feld
|
r34848 | |||
Boris Feld
|
r34852 | if min_date == max_date: | ||
Boris Feld
|
r36625 | fmtmin_date = dateutil.datestr(min_date, '%Y-%m-%d %H:%M %1%2') | ||
Boris Feld
|
r34852 | line.append(" (at %s)" % fmtmin_date) | ||
else: | ||||
Boris Feld
|
r36625 | fmtmin_date = dateutil.datestr(min_date, '%Y-%m-%d %H:%M %1%2') | ||
fmtmax_date = dateutil.datestr(max_date, '%Y-%m-%d %H:%M %1%2') | ||||
Boris Feld
|
r34852 | line.append(" (between %s and %s)" % (fmtmin_date, fmtmax_date)) | ||
Boris Feld
|
r34848 | |||
return "".join(line) | ||||
Boris Feld
|
r35589 | |||
Boris Feld
|
r35627 | |||
filteredmsgtable = { | ||||
"pruned": _("hidden revision '%s' is pruned"), | ||||
"diverged": _("hidden revision '%s' has diverged"), | ||||
"superseded": _("hidden revision '%s' was rewritten as: %s"), | ||||
"superseded_split": _("hidden revision '%s' was split as: %s"), | ||||
"superseded_split_several": _("hidden revision '%s' was split as: %s and " | ||||
"%d more"), | ||||
} | ||||
Boris Feld
|
r35628 | def _getfilteredreason(repo, changeid, ctx): | ||
Boris Feld
|
r35589 | """return a human-friendly string on why a obsolete changeset is hidden | ||
""" | ||||
Boris Feld
|
r35628 | successors = successorssets(repo, ctx.node()) | ||
Boris Feld
|
r35589 | fate = _getobsfate(successors) | ||
# Be more precise in case the revision is superseded | ||||
if fate == 'pruned': | ||||
Boris Feld
|
r35627 | return filteredmsgtable['pruned'] % changeid | ||
Boris Feld
|
r35589 | elif fate == 'diverged': | ||
Boris Feld
|
r35627 | return filteredmsgtable['diverged'] % changeid | ||
Boris Feld
|
r35589 | elif fate == 'superseded': | ||
Boris Feld
|
r35627 | single_successor = nodemod.short(successors[0][0]) | ||
return filteredmsgtable['superseded'] % (changeid, single_successor) | ||||
Boris Feld
|
r35589 | elif fate == 'superseded_split': | ||
succs = [] | ||||
for node_id in successors[0]: | ||||
succs.append(nodemod.short(node_id)) | ||||
if len(succs) <= 2: | ||||
Boris Feld
|
r35627 | fmtsuccs = ', '.join(succs) | ||
return filteredmsgtable['superseded_split'] % (changeid, fmtsuccs) | ||||
Boris Feld
|
r35589 | else: | ||
Boris Feld
|
r35627 | firstsuccessors = ', '.join(succs[:2]) | ||
Boris Feld
|
r35589 | remainingnumber = len(succs) - 2 | ||
Boris Feld
|
r35627 | args = (changeid, firstsuccessors, remainingnumber) | ||
return filteredmsgtable['superseded_split_several'] % args | ||||
r36972 | ||||
def divergentsets(repo, ctx): | ||||
"""Compute sets of commits divergent with a given one""" | ||||
cache = {} | ||||
base = {} | ||||
for n in allpredecessors(repo.obsstore, [ctx.node()]): | ||||
if n == ctx.node(): | ||||
# a node can't be a base for divergence with itself | ||||
continue | ||||
nsuccsets = successorssets(repo, n, cache) | ||||
for nsuccset in nsuccsets: | ||||
if ctx.node() in nsuccset: | ||||
# we are only interested in *other* successor sets | ||||
continue | ||||
if tuple(nsuccset) in base: | ||||
# we already know the latest base for this divergency | ||||
continue | ||||
base[tuple(nsuccset)] = n | ||||
return [{'divergentnodes': divset, 'commonpredecessor': b} | ||||
for divset, b in base.iteritems()] | ||||
def whyunstable(repo, ctx): | ||||
result = [] | ||||
if ctx.orphan(): | ||||
for parent in ctx.parents(): | ||||
kind = None | ||||
if parent.orphan(): | ||||
kind = 'orphan' | ||||
elif parent.obsolete(): | ||||
kind = 'obsolete' | ||||
if kind is not None: | ||||
result.append({'instability': 'orphan', | ||||
'reason': '%s parent' % kind, | ||||
'node': parent.hex()}) | ||||
if ctx.phasedivergent(): | ||||
predecessors = allpredecessors(repo.obsstore, [ctx.node()], | ||||
ignoreflags=bumpedfix) | ||||
immutable = [repo[p] for p in predecessors | ||||
if p in repo and not repo[p].mutable()] | ||||
for predecessor in immutable: | ||||
result.append({'instability': 'phase-divergent', | ||||
'reason': 'immutable predecessor', | ||||
'node': predecessor.hex()}) | ||||
if ctx.contentdivergent(): | ||||
dsets = divergentsets(repo, ctx) | ||||
for dset in dsets: | ||||
divnodes = [repo[n] for n in dset['divergentnodes']] | ||||
result.append({'instability': 'content-divergent', | ||||
'divergentnodes': divnodes, | ||||
'reason': 'predecessor', | ||||
'node': nodemod.hex(dset['commonpredecessor'])}) | ||||
return result | ||||