##// END OF EJS Templates
compression: introduce a `storage.revlog.zstd.level` configuration...
compression: introduce a `storage.revlog.zstd.level` configuration This option control the zstd compression level used when compressing revlog chunk. The usage of zstd for revlog compression has not graduated from experimental yet, but we intend to fix that soon. The option name for the compression level is more straight forward to pick, so this changesets comes first. Having a dedicated option for each compression engine is useful because they don't support the same range of values. I ran the same measurement as for the zlib compression level (in the parent changesets). The variation in repository size is stay mostly in the same (small) range. The "read/write" performance see smallish variation, but are overall much better than zlib. Write performance show the same tend of having better write performance for when reaching high-end compression. Again, we don't intend to change the default zstd compression level (currently: 3) in this series. However this is worth investigating in the future. The Performance comparison of zlib vs zstd is quite impressive. The repository size stay in the same range, but the performance are much better in all situations. Comparison summary ================== We are looking at: - performance range for zlib - performance range for zstd - comparison of default zstd (level-3) to default zlib (level 6) - comparison of the slowest zstd time to the fastest zlib time Read performance: ----------------- | zlib | zstd | cmp | f2s mercurial | 0.170159 - 0.189219 | 0.144127 - 0.149624 | 80% | 88% pypy | 2.679217 - 2.768691 | 1.532317 - 1.705044 | 60% | 63% netbeans | 122.477027 - 141.620281 | 72.996346 - 89.731560 | 58% | 73% mozilla | 147.867662 - 170.572118 | 91.700995 - 105.853099 | 56% | 71% Write performance: ------------------ | zlib | zstd | cmp | f2s mercurial | 53.250304 - 56.2936129 | 40.877025 - 45.677286 | 75% | 86% pypy | 460.721984 - 476.589918 | 270.545409 - 301.002219 | 63% | 65% netbeans | 520.560316 - 715.930400 | 370.356311 - 428.329652 | 55% | 82% mozilla | 739.803002 - 987.056093 | 505.152906 - 591.930683 | 57% | 80% Raw data -------- repo alg lvl .hg/store size 00manifest.d read write mercurial zlib 1 49,402,813 5,963,475 0.170159 53.250304 mercurial zlib 6 47,197,397 5,875,730 0.182820 56.264320 mercurial zlib 9 47,121,596 5,849,781 0.189219 56.293612 mercurial zstd 1 49,737,084 5,966,355 0.144127 40.877025 mercurial zstd 3 48,961,867 5,895,208 0.146376 42.268142 mercurial zstd 5 48,200,592 5,938,676 0.149624 43.162875 mercurial zstd 10 47,833,520 5,913,353 0.145185 44.012489 mercurial zstd 15 47,314,604 5,728,679 0.147686 45.677286 mercurial zstd 20 47,330,502 5,830,539 0.145789 45.025407 mercurial zstd 22 47,330,076 5,830,539 0.143996 44.690460 pypy zlib 1 370,830,572 28,462,425 2.679217 460.721984 pypy zlib 6 340,112,317 27,648,747 2.768691 467.537158 pypy zlib 9 338,360,736 27,639,003 2.763495 476.589918 pypy zstd 1 362,377,479 27,916,214 1.532317 270.545409 pypy zstd 3 354,137,693 27,905,988 1.686718 294.951509 pypy zstd 5 342,640,043 27,655,774 1.705044 301.002219 pypy zstd 10 334,224,327 27,164,493 1.567287 285.186239 pypy zstd 15 329,000,363 26,645,965 1.637729 299.561332 pypy zstd 20 324,534,039 26,199,547 1.526813 302.149827 pypy zstd 22 324,530,595 26,198,932 1.525718 307.821218 netbeans zlib 1 1,281,847,810 165,495,457 122.477027 520.560316 netbeans zlib 6 1,205,284,353 159,161,207 139.876147 715.930400 netbeans zlib 9 1,197,135,671 155,034,586 141.620281 678.297064 netbeans zstd 1 1,259,581,737 160,840,613 72.996346 370.356311 netbeans zstd 3 1,232,978,122 157,691,551 81.622317 396.733087 netbeans zstd 5 1,208,034,075 160,246,880 83.080549 364.342626 netbeans zstd 10 1,188,624,176 156,083,417 79.323935 403.594602 netbeans zstd 15 1,176,973,589 153,859,477 89.731560 428.329652 netbeans zstd 20 1,162,958,258 151,147,535 82.842667 392.335349 netbeans zstd 22 1,162,707,029 151,150,220 82.565695 402.840655 mozilla zlib 1 2,775,497,186 298,527,987 147.867662 751.263721 mozilla zlib 6 2,596,856,420 286,597,671 170.572118 987.056093 mozilla zlib 9 2,587,542,494 287,018,264 163.622338 739.803002 mozilla zstd 1 2,723,159,348 286,617,532 91.700995 570.042751 mozilla zstd 3 2,665,055,001 286,152,013 95.240155 561.412805 mozilla zstd 5 2,607,819,817 288,060,030 101.978048 505.152906 mozilla zstd 10 2,558,761,085 283,967,648 104.113481 497.771202 mozilla zstd 15 2,526,216,060 275,581,300 105.853099 591.930683 mozilla zstd 20 2,485,114,806 266,478,859 95.268795 576.515389 mozilla zstd 22 2,484,869,080 266,456,505 94.429282 572.785537

File last commit:

r41996:c6a5009e default
r42211:bb271ec2 default
Show More
absorb.py
1011 lines | 38.6 KiB | text/x-python | PythonLexer
# absorb.py
#
# Copyright 2016 Facebook, Inc.
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
"""apply working directory changes to changesets (EXPERIMENTAL)
The absorb extension provides a command to use annotate information to
amend modified chunks into the corresponding non-public changesets.
::
[absorb]
# only check 50 recent non-public changesets at most
max-stack-size = 50
# whether to add noise to new commits to avoid obsolescence cycle
add-noise = 1
# make `amend --correlated` a shortcut to the main command
amend-flag = correlated
[color]
absorb.description = yellow
absorb.node = blue bold
absorb.path = bold
"""
# TODO:
# * Rename config items to [commands] namespace
# * Converge getdraftstack() with other code in core
# * move many attributes on fixupstate to be private
from __future__ import absolute_import
import collections
from mercurial.i18n import _
from mercurial import (
cmdutil,
commands,
context,
crecord,
error,
linelog,
mdiff,
node,
obsolete,
patch,
phases,
pycompat,
registrar,
scmutil,
util,
)
from mercurial.utils import (
stringutil,
)
# Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for
# extensions which SHIP WITH MERCURIAL. Non-mainline extensions should
# be specifying the version(s) of Mercurial they are tested with, or
# leave the attribute unspecified.
testedwith = 'ships-with-hg-core'
cmdtable = {}
command = registrar.command(cmdtable)
configtable = {}
configitem = registrar.configitem(configtable)
configitem('absorb', 'add-noise', default=True)
configitem('absorb', 'amend-flag', default=None)
configitem('absorb', 'max-stack-size', default=50)
colortable = {
'absorb.description': 'yellow',
'absorb.node': 'blue bold',
'absorb.path': 'bold',
}
defaultdict = collections.defaultdict
class nullui(object):
"""blank ui object doing nothing"""
debugflag = False
verbose = False
quiet = True
def __getitem__(name):
def nullfunc(*args, **kwds):
return
return nullfunc
class emptyfilecontext(object):
"""minimal filecontext representing an empty file"""
def data(self):
return ''
def node(self):
return node.nullid
def uniq(lst):
"""list -> list. remove duplicated items without changing the order"""
seen = set()
result = []
for x in lst:
if x not in seen:
seen.add(x)
result.append(x)
return result
def getdraftstack(headctx, limit=None):
"""(ctx, int?) -> [ctx]. get a linear stack of non-public changesets.
changesets are sorted in topo order, oldest first.
return at most limit items, if limit is a positive number.
merges are considered as non-draft as well. i.e. every commit
returned has and only has 1 parent.
"""
ctx = headctx
result = []
while ctx.phase() != phases.public:
if limit and len(result) >= limit:
break
parents = ctx.parents()
if len(parents) != 1:
break
result.append(ctx)
ctx = parents[0]
result.reverse()
return result
def getfilestack(stack, path, seenfctxs=None):
"""([ctx], str, set) -> [fctx], {ctx: fctx}
stack is a list of contexts, from old to new. usually they are what
"getdraftstack" returns.
follows renames, but not copies.
seenfctxs is a set of filecontexts that will be considered "immutable".
they are usually what this function returned in earlier calls, useful
to avoid issues that a file was "moved" to multiple places and was then
modified differently, like: "a" was copied to "b", "a" was also copied to
"c" and then "a" was deleted, then both "b" and "c" were "moved" from "a"
and we enforce only one of them to be able to affect "a"'s content.
return an empty list and an empty dict, if the specified path does not
exist in stack[-1] (the top of the stack).
otherwise, return a list of de-duplicated filecontexts, and the map to
convert ctx in the stack to fctx, for possible mutable fctxs. the first item
of the list would be outside the stack and should be considered immutable.
the remaining items are within the stack.
for example, given the following changelog and corresponding filelog
revisions:
changelog: 3----4----5----6----7
filelog: x 0----1----1----2 (x: no such file yet)
- if stack = [5, 6, 7], returns ([0, 1, 2], {5: 1, 6: 1, 7: 2})
- if stack = [3, 4, 5], returns ([e, 0, 1], {4: 0, 5: 1}), where "e" is a
dummy empty filecontext.
- if stack = [2], returns ([], {})
- if stack = [7], returns ([1, 2], {7: 2})
- if stack = [6, 7], returns ([1, 2], {6: 1, 7: 2}), although {6: 1} can be
removed, since 1 is immutable.
"""
if seenfctxs is None:
seenfctxs = set()
assert stack
if path not in stack[-1]:
return [], {}
fctxs = []
fctxmap = {}
pctx = stack[0].p1() # the public (immutable) ctx we stop at
for ctx in reversed(stack):
if path not in ctx: # the file is added in the next commit
pctx = ctx
break
fctx = ctx[path]
fctxs.append(fctx)
if fctx in seenfctxs: # treat fctx as the immutable one
pctx = None # do not add another immutable fctx
break
fctxmap[ctx] = fctx # only for mutable fctxs
copy = fctx.copysource()
if copy:
path = copy # follow rename
if path in ctx: # but do not follow copy
pctx = ctx.p1()
break
if pctx is not None: # need an extra immutable fctx
if path in pctx:
fctxs.append(pctx[path])
else:
fctxs.append(emptyfilecontext())
fctxs.reverse()
# note: we rely on a property of hg: filerev is not reused for linear
# history. i.e. it's impossible to have:
# changelog: 4----5----6 (linear, no merges)
# filelog: 1----2----1
# ^ reuse filerev (impossible)
# because parents are part of the hash. if that's not true, we need to
# remove uniq and find a different way to identify fctxs.
return uniq(fctxs), fctxmap
class overlaystore(patch.filestore):
"""read-only, hybrid store based on a dict and ctx.
memworkingcopy: {path: content}, overrides file contents.
"""
def __init__(self, basectx, memworkingcopy):
self.basectx = basectx
self.memworkingcopy = memworkingcopy
def getfile(self, path):
"""comply with mercurial.patch.filestore.getfile"""
if path not in self.basectx:
return None, None, None
fctx = self.basectx[path]
if path in self.memworkingcopy:
content = self.memworkingcopy[path]
else:
content = fctx.data()
mode = (fctx.islink(), fctx.isexec())
copy = fctx.copysource()
return content, mode, copy
def overlaycontext(memworkingcopy, ctx, parents=None, extra=None):
"""({path: content}, ctx, (p1node, p2node)?, {}?) -> memctx
memworkingcopy overrides file contents.
"""
# parents must contain 2 items: (node1, node2)
if parents is None:
parents = ctx.repo().changelog.parents(ctx.node())
if extra is None:
extra = ctx.extra()
date = ctx.date()
desc = ctx.description()
user = ctx.user()
files = set(ctx.files()).union(memworkingcopy)
store = overlaystore(ctx, memworkingcopy)
return context.memctx(
repo=ctx.repo(), parents=parents, text=desc,
files=files, filectxfn=store, user=user, date=date,
branch=None, extra=extra)
class filefixupstate(object):
"""state needed to apply fixups to a single file
internally, it keeps file contents of several revisions and a linelog.
the linelog uses odd revision numbers for original contents (fctxs passed
to __init__), and even revision numbers for fixups, like:
linelog rev 1: self.fctxs[0] (from an immutable "public" changeset)
linelog rev 2: fixups made to self.fctxs[0]
linelog rev 3: self.fctxs[1] (a child of fctxs[0])
linelog rev 4: fixups made to self.fctxs[1]
...
a typical use is like:
1. call diffwith, to calculate self.fixups
2. (optionally), present self.fixups to the user, or change it
3. call apply, to apply changes
4. read results from "finalcontents", or call getfinalcontent
"""
def __init__(self, fctxs, path, ui=None, opts=None):
"""([fctx], ui or None) -> None
fctxs should be linear, and sorted by topo order - oldest first.
fctxs[0] will be considered as "immutable" and will not be changed.
"""
self.fctxs = fctxs
self.path = path
self.ui = ui or nullui()
self.opts = opts or {}
# following fields are built from fctxs. they exist for perf reason
self.contents = [f.data() for f in fctxs]
self.contentlines = pycompat.maplist(mdiff.splitnewlines, self.contents)
self.linelog = self._buildlinelog()
if self.ui.debugflag:
assert self._checkoutlinelog() == self.contents
# following fields will be filled later
self.chunkstats = [0, 0] # [adopted, total : int]
self.targetlines = [] # [str]
self.fixups = [] # [(linelog rev, a1, a2, b1, b2)]
self.finalcontents = [] # [str]
self.ctxaffected = set()
def diffwith(self, targetfctx, fm=None):
"""calculate fixups needed by examining the differences between
self.fctxs[-1] and targetfctx, chunk by chunk.
targetfctx is the target state we move towards. we may or may not be
able to get there because not all modified chunks can be amended into
a non-public fctx unambiguously.
call this only once, before apply().
update self.fixups, self.chunkstats, and self.targetlines.
"""
a = self.contents[-1]
alines = self.contentlines[-1]
b = targetfctx.data()
blines = mdiff.splitnewlines(b)
self.targetlines = blines
self.linelog.annotate(self.linelog.maxrev)
annotated = self.linelog.annotateresult # [(linelog rev, linenum)]
assert len(annotated) == len(alines)
# add a dummy end line to make insertion at the end easier
if annotated:
dummyendline = (annotated[-1][0], annotated[-1][1] + 1)
annotated.append(dummyendline)
# analyse diff blocks
for chunk in self._alldiffchunks(a, b, alines, blines):
newfixups = self._analysediffchunk(chunk, annotated)
self.chunkstats[0] += bool(newfixups) # 1 or 0
self.chunkstats[1] += 1
self.fixups += newfixups
if fm is not None:
self._showchanges(fm, alines, blines, chunk, newfixups)
def apply(self):
"""apply self.fixups. update self.linelog, self.finalcontents.
call this only once, before getfinalcontent(), after diffwith().
"""
# the following is unnecessary, as it's done by "diffwith":
# self.linelog.annotate(self.linelog.maxrev)
for rev, a1, a2, b1, b2 in reversed(self.fixups):
blines = self.targetlines[b1:b2]
if self.ui.debugflag:
idx = (max(rev - 1, 0)) // 2
self.ui.write(_('%s: chunk %d:%d -> %d lines\n')
% (node.short(self.fctxs[idx].node()),
a1, a2, len(blines)))
self.linelog.replacelines(rev, a1, a2, b1, b2)
if self.opts.get('edit_lines', False):
self.finalcontents = self._checkoutlinelogwithedits()
else:
self.finalcontents = self._checkoutlinelog()
def getfinalcontent(self, fctx):
"""(fctx) -> str. get modified file content for a given filecontext"""
idx = self.fctxs.index(fctx)
return self.finalcontents[idx]
def _analysediffchunk(self, chunk, annotated):
"""analyse a different chunk and return new fixups found
return [] if no lines from the chunk can be safely applied.
the chunk (or lines) cannot be safely applied, if, for example:
- the modified (deleted) lines belong to a public changeset
(self.fctxs[0])
- the chunk is a pure insertion and the adjacent lines (at most 2
lines) belong to different non-public changesets, or do not belong
to any non-public changesets.
- the chunk is modifying lines from different changesets.
in this case, if the number of lines deleted equals to the number
of lines added, assume it's a simple 1:1 map (could be wrong).
otherwise, give up.
- the chunk is modifying lines from a single non-public changeset,
but other revisions touch the area as well. i.e. the lines are
not continuous as seen from the linelog.
"""
a1, a2, b1, b2 = chunk
# find involved indexes from annotate result
involved = annotated[a1:a2]
if not involved and annotated: # a1 == a2 and a is not empty
# pure insertion, check nearby lines. ignore lines belong
# to the public (first) changeset (i.e. annotated[i][0] == 1)
nearbylinenums = {a2, max(0, a1 - 1)}
involved = [annotated[i]
for i in nearbylinenums if annotated[i][0] != 1]
involvedrevs = list(set(r for r, l in involved))
newfixups = []
if len(involvedrevs) == 1 and self._iscontinuous(a1, a2 - 1, True):
# chunk belongs to a single revision
rev = involvedrevs[0]
if rev > 1:
fixuprev = rev + 1
newfixups.append((fixuprev, a1, a2, b1, b2))
elif a2 - a1 == b2 - b1 or b1 == b2:
# 1:1 line mapping, or chunk was deleted
for i in pycompat.xrange(a1, a2):
rev, linenum = annotated[i]
if rev > 1:
if b1 == b2: # deletion, simply remove that single line
nb1 = nb2 = 0
else: # 1:1 line mapping, change the corresponding rev
nb1 = b1 + i - a1
nb2 = nb1 + 1
fixuprev = rev + 1
newfixups.append((fixuprev, i, i + 1, nb1, nb2))
return self._optimizefixups(newfixups)
@staticmethod
def _alldiffchunks(a, b, alines, blines):
"""like mdiff.allblocks, but only care about differences"""
blocks = mdiff.allblocks(a, b, lines1=alines, lines2=blines)
for chunk, btype in blocks:
if btype != '!':
continue
yield chunk
def _buildlinelog(self):
"""calculate the initial linelog based on self.content{,line}s.
this is similar to running a partial "annotate".
"""
llog = linelog.linelog()
a, alines = '', []
for i in pycompat.xrange(len(self.contents)):
b, blines = self.contents[i], self.contentlines[i]
llrev = i * 2 + 1
chunks = self._alldiffchunks(a, b, alines, blines)
for a1, a2, b1, b2 in reversed(list(chunks)):
llog.replacelines(llrev, a1, a2, b1, b2)
a, alines = b, blines
return llog
def _checkoutlinelog(self):
"""() -> [str]. check out file contents from linelog"""
contents = []
for i in pycompat.xrange(len(self.contents)):
rev = (i + 1) * 2
self.linelog.annotate(rev)
content = ''.join(map(self._getline, self.linelog.annotateresult))
contents.append(content)
return contents
def _checkoutlinelogwithedits(self):
"""() -> [str]. prompt all lines for edit"""
alllines = self.linelog.getalllines()
# header
editortext = (_('HG: editing %s\nHG: "y" means the line to the right '
'exists in the changeset to the top\nHG:\n')
% self.fctxs[-1].path())
# [(idx, fctx)]. hide the dummy emptyfilecontext
visiblefctxs = [(i, f)
for i, f in enumerate(self.fctxs)
if not isinstance(f, emptyfilecontext)]
for i, (j, f) in enumerate(visiblefctxs):
editortext += (_('HG: %s/%s %s %s\n') %
('|' * i, '-' * (len(visiblefctxs) - i + 1),
node.short(f.node()),
f.description().split('\n',1)[0]))
editortext += _('HG: %s\n') % ('|' * len(visiblefctxs))
# figure out the lifetime of a line, this is relatively inefficient,
# but probably fine
lineset = defaultdict(lambda: set()) # {(llrev, linenum): {llrev}}
for i, f in visiblefctxs:
self.linelog.annotate((i + 1) * 2)
for l in self.linelog.annotateresult:
lineset[l].add(i)
# append lines
for l in alllines:
editortext += (' %s : %s' %
(''.join([('y' if i in lineset[l] else ' ')
for i, _f in visiblefctxs]),
self._getline(l)))
# run editor
editedtext = self.ui.edit(editortext, '', action='absorb')
if not editedtext:
raise error.Abort(_('empty editor text'))
# parse edited result
contents = ['' for i in self.fctxs]
leftpadpos = 4
colonpos = leftpadpos + len(visiblefctxs) + 1
for l in mdiff.splitnewlines(editedtext):
if l.startswith('HG:'):
continue
if l[colonpos - 1:colonpos + 2] != ' : ':
raise error.Abort(_('malformed line: %s') % l)
linecontent = l[colonpos + 2:]
for i, ch in enumerate(
pycompat.bytestr(l[leftpadpos:colonpos - 1])):
if ch == 'y':
contents[visiblefctxs[i][0]] += linecontent
# chunkstats is hard to calculate if anything changes, therefore
# set them to just a simple value (1, 1).
if editedtext != editortext:
self.chunkstats = [1, 1]
return contents
def _getline(self, lineinfo):
"""((rev, linenum)) -> str. convert rev+line number to line content"""
rev, linenum = lineinfo
if rev & 1: # odd: original line taken from fctxs
return self.contentlines[rev // 2][linenum]
else: # even: fixup line from targetfctx
return self.targetlines[linenum]
def _iscontinuous(self, a1, a2, closedinterval=False):
"""(a1, a2 : int) -> bool
check if these lines are continuous. i.e. no other insertions or
deletions (from other revisions) among these lines.
closedinterval decides whether a2 should be included or not. i.e. is
it [a1, a2), or [a1, a2] ?
"""
if a1 >= a2:
return True
llog = self.linelog
offset1 = llog.getoffset(a1)
offset2 = llog.getoffset(a2) + int(closedinterval)
linesinbetween = llog.getalllines(offset1, offset2)
return len(linesinbetween) == a2 - a1 + int(closedinterval)
def _optimizefixups(self, fixups):
"""[(rev, a1, a2, b1, b2)] -> [(rev, a1, a2, b1, b2)].
merge adjacent fixups to make them less fragmented.
"""
result = []
pcurrentchunk = [[-1, -1, -1, -1, -1]]
def pushchunk():
if pcurrentchunk[0][0] != -1:
result.append(tuple(pcurrentchunk[0]))
for i, chunk in enumerate(fixups):
rev, a1, a2, b1, b2 = chunk
lastrev = pcurrentchunk[0][0]
lasta2 = pcurrentchunk[0][2]
lastb2 = pcurrentchunk[0][4]
if (a1 == lasta2 and b1 == lastb2 and rev == lastrev and
self._iscontinuous(max(a1 - 1, 0), a1)):
# merge into currentchunk
pcurrentchunk[0][2] = a2
pcurrentchunk[0][4] = b2
else:
pushchunk()
pcurrentchunk[0] = list(chunk)
pushchunk()
return result
def _showchanges(self, fm, alines, blines, chunk, fixups):
def trim(line):
if line.endswith('\n'):
line = line[:-1]
return line
# this is not optimized for perf but _showchanges only gets executed
# with an extra command-line flag.
a1, a2, b1, b2 = chunk
aidxs, bidxs = [0] * (a2 - a1), [0] * (b2 - b1)
for idx, fa1, fa2, fb1, fb2 in fixups:
for i in pycompat.xrange(fa1, fa2):
aidxs[i - a1] = (max(idx, 1) - 1) // 2
for i in pycompat.xrange(fb1, fb2):
bidxs[i - b1] = (max(idx, 1) - 1) // 2
fm.startitem()
fm.write('hunk', ' %s\n',
'@@ -%d,%d +%d,%d @@'
% (a1, a2 - a1, b1, b2 - b1), label='diff.hunk')
fm.data(path=self.path, linetype='hunk')
def writeline(idx, diffchar, line, linetype, linelabel):
fm.startitem()
node = ''
if idx:
ctx = self.fctxs[idx]
fm.context(fctx=ctx)
node = ctx.hex()
self.ctxaffected.add(ctx.changectx())
fm.write('node', '%-7.7s ', node, label='absorb.node')
fm.write('diffchar ' + linetype, '%s%s\n', diffchar, line,
label=linelabel)
fm.data(path=self.path, linetype=linetype)
for i in pycompat.xrange(a1, a2):
writeline(aidxs[i - a1], '-', trim(alines[i]), 'deleted',
'diff.deleted')
for i in pycompat.xrange(b1, b2):
writeline(bidxs[i - b1], '+', trim(blines[i]), 'inserted',
'diff.inserted')
class fixupstate(object):
"""state needed to run absorb
internally, it keeps paths and filefixupstates.
a typical use is like filefixupstates:
1. call diffwith, to calculate fixups
2. (optionally), present fixups to the user, or edit fixups
3. call apply, to apply changes to memory
4. call commit, to commit changes to hg database
"""
def __init__(self, stack, ui=None, opts=None):
"""([ctx], ui or None) -> None
stack: should be linear, and sorted by topo order - oldest first.
all commits in stack are considered mutable.
"""
assert stack
self.ui = ui or nullui()
self.opts = opts or {}
self.stack = stack
self.repo = stack[-1].repo().unfiltered()
# following fields will be filled later
self.paths = [] # [str]
self.status = None # ctx.status output
self.fctxmap = {} # {path: {ctx: fctx}}
self.fixupmap = {} # {path: filefixupstate}
self.replacemap = {} # {oldnode: newnode or None}
self.finalnode = None # head after all fixups
self.ctxaffected = set() # ctx that will be absorbed into
def diffwith(self, targetctx, match=None, fm=None):
"""diff and prepare fixups. update self.fixupmap, self.paths"""
# only care about modified files
self.status = self.stack[-1].status(targetctx, match)
self.paths = []
# but if --edit-lines is used, the user may want to edit files
# even if they are not modified
editopt = self.opts.get('edit_lines')
if not self.status.modified and editopt and match:
interestingpaths = match.files()
else:
interestingpaths = self.status.modified
# prepare the filefixupstate
seenfctxs = set()
# sorting is necessary to eliminate ambiguity for the "double move"
# case: "hg cp A B; hg cp A C; hg rm A", then only "B" can affect "A".
for path in sorted(interestingpaths):
self.ui.debug('calculating fixups for %s\n' % path)
targetfctx = targetctx[path]
fctxs, ctx2fctx = getfilestack(self.stack, path, seenfctxs)
# ignore symbolic links or binary, or unchanged files
if any(f.islink() or stringutil.binary(f.data())
for f in [targetfctx] + fctxs
if not isinstance(f, emptyfilecontext)):
continue
if targetfctx.data() == fctxs[-1].data() and not editopt:
continue
seenfctxs.update(fctxs[1:])
self.fctxmap[path] = ctx2fctx
fstate = filefixupstate(fctxs, path, ui=self.ui, opts=self.opts)
if fm is not None:
fm.startitem()
fm.plain('showing changes for ')
fm.write('path', '%s\n', path, label='absorb.path')
fm.data(linetype='path')
fstate.diffwith(targetfctx, fm)
self.fixupmap[path] = fstate
self.paths.append(path)
self.ctxaffected.update(fstate.ctxaffected)
def apply(self):
"""apply fixups to individual filefixupstates"""
for path, state in self.fixupmap.iteritems():
if self.ui.debugflag:
self.ui.write(_('applying fixups to %s\n') % path)
state.apply()
@property
def chunkstats(self):
"""-> {path: chunkstats}. collect chunkstats from filefixupstates"""
return dict((path, state.chunkstats)
for path, state in self.fixupmap.iteritems())
def commit(self):
"""commit changes. update self.finalnode, self.replacemap"""
with self.repo.wlock(), self.repo.lock():
with self.repo.transaction('absorb') as tr:
self._commitstack()
self._movebookmarks(tr)
if self.repo['.'].node() in self.replacemap:
self._moveworkingdirectoryparent()
self._cleanupoldcommits()
return self.finalnode
def printchunkstats(self):
"""print things like '1 of 2 chunk(s) applied'"""
ui = self.ui
chunkstats = self.chunkstats
if ui.verbose:
# chunkstats for each file
for path, stat in chunkstats.iteritems():
if stat[0]:
ui.write(_('%s: %d of %d chunk(s) applied\n')
% (path, stat[0], stat[1]))
elif not ui.quiet:
# a summary for all files
stats = chunkstats.values()
applied, total = (sum(s[i] for s in stats) for i in (0, 1))
ui.write(_('%d of %d chunk(s) applied\n') % (applied, total))
def _commitstack(self):
"""make new commits. update self.finalnode, self.replacemap.
it is splitted from "commit" to avoid too much indentation.
"""
# last node (20-char) committed by us
lastcommitted = None
# p1 which overrides the parent of the next commit, "None" means use
# the original parent unchanged
nextp1 = None
for ctx in self.stack:
memworkingcopy = self._getnewfilecontents(ctx)
if not memworkingcopy and not lastcommitted:
# nothing changed, nothing commited
nextp1 = ctx
continue
if self._willbecomenoop(memworkingcopy, ctx, nextp1):
# changeset is no longer necessary
self.replacemap[ctx.node()] = None
msg = _('became empty and was dropped')
else:
# changeset needs re-commit
nodestr = self._commitsingle(memworkingcopy, ctx, p1=nextp1)
lastcommitted = self.repo[nodestr]
nextp1 = lastcommitted
self.replacemap[ctx.node()] = lastcommitted.node()
if memworkingcopy:
msg = _('%d file(s) changed, became %s') % (
len(memworkingcopy), self._ctx2str(lastcommitted))
else:
msg = _('became %s') % self._ctx2str(lastcommitted)
if self.ui.verbose and msg:
self.ui.write(_('%s: %s\n') % (self._ctx2str(ctx), msg))
self.finalnode = lastcommitted and lastcommitted.node()
def _ctx2str(self, ctx):
if self.ui.debugflag:
return '%d:%s' % (ctx.rev(), ctx.hex())
else:
return '%d:%s' % (ctx.rev(), node.short(ctx.node()))
def _getnewfilecontents(self, ctx):
"""(ctx) -> {path: str}
fetch file contents from filefixupstates.
return the working copy overrides - files different from ctx.
"""
result = {}
for path in self.paths:
ctx2fctx = self.fctxmap[path] # {ctx: fctx}
if ctx not in ctx2fctx:
continue
fctx = ctx2fctx[ctx]
content = fctx.data()
newcontent = self.fixupmap[path].getfinalcontent(fctx)
if content != newcontent:
result[fctx.path()] = newcontent
return result
def _movebookmarks(self, tr):
repo = self.repo
needupdate = [(name, self.replacemap[hsh])
for name, hsh in repo._bookmarks.iteritems()
if hsh in self.replacemap]
changes = []
for name, hsh in needupdate:
if hsh:
changes.append((name, hsh))
if self.ui.verbose:
self.ui.write(_('moving bookmark %s to %s\n')
% (name, node.hex(hsh)))
else:
changes.append((name, None))
if self.ui.verbose:
self.ui.write(_('deleting bookmark %s\n') % name)
repo._bookmarks.applychanges(repo, tr, changes)
def _moveworkingdirectoryparent(self):
if not self.finalnode:
# Find the latest not-{obsoleted,stripped} parent.
revs = self.repo.revs('max(::. - %ln)', self.replacemap.keys())
ctx = self.repo[revs.first()]
self.finalnode = ctx.node()
else:
ctx = self.repo[self.finalnode]
dirstate = self.repo.dirstate
# dirstate.rebuild invalidates fsmonitorstate, causing "hg status" to
# be slow. in absorb's case, no need to invalidate fsmonitorstate.
noop = lambda: 0
restore = noop
if util.safehasattr(dirstate, '_fsmonitorstate'):
bak = dirstate._fsmonitorstate.invalidate
def restore():
dirstate._fsmonitorstate.invalidate = bak
dirstate._fsmonitorstate.invalidate = noop
try:
with dirstate.parentchange():
dirstate.rebuild(ctx.node(), ctx.manifest(), self.paths)
finally:
restore()
@staticmethod
def _willbecomenoop(memworkingcopy, ctx, pctx=None):
"""({path: content}, ctx, ctx) -> bool. test if a commit will be noop
if it will become an empty commit (does not change anything, after the
memworkingcopy overrides), return True. otherwise return False.
"""
if not pctx:
parents = ctx.parents()
if len(parents) != 1:
return False
pctx = parents[0]
# ctx changes more files (not a subset of memworkingcopy)
if not set(ctx.files()).issubset(set(memworkingcopy)):
return False
for path, content in memworkingcopy.iteritems():
if path not in pctx or path not in ctx:
return False
fctx = ctx[path]
pfctx = pctx[path]
if pfctx.flags() != fctx.flags():
return False
if pfctx.data() != content:
return False
return True
def _commitsingle(self, memworkingcopy, ctx, p1=None):
"""(ctx, {path: content}, node) -> node. make a single commit
the commit is a clone from ctx, with a (optionally) different p1, and
different file contents replaced by memworkingcopy.
"""
parents = p1 and (p1, node.nullid)
extra = ctx.extra()
if self._useobsolete and self.ui.configbool('absorb', 'add-noise'):
extra['absorb_source'] = ctx.hex()
mctx = overlaycontext(memworkingcopy, ctx, parents, extra=extra)
return mctx.commit()
@util.propertycache
def _useobsolete(self):
"""() -> bool"""
return obsolete.isenabled(self.repo, obsolete.createmarkersopt)
def _cleanupoldcommits(self):
replacements = {k: ([v] if v is not None else [])
for k, v in self.replacemap.iteritems()}
if replacements:
scmutil.cleanupnodes(self.repo, replacements, operation='absorb',
fixphase=True)
def _parsechunk(hunk):
"""(crecord.uihunk or patch.recordhunk) -> (path, (a1, a2, [bline]))"""
if type(hunk) not in (crecord.uihunk, patch.recordhunk):
return None, None
path = hunk.header.filename()
a1 = hunk.fromline + len(hunk.before) - 1
# remove before and after context
hunk.before = hunk.after = []
buf = util.stringio()
hunk.write(buf)
patchlines = mdiff.splitnewlines(buf.getvalue())
# hunk.prettystr() will update hunk.removed
a2 = a1 + hunk.removed
blines = [l[1:] for l in patchlines[1:] if l[0] != '-']
return path, (a1, a2, blines)
def overlaydiffcontext(ctx, chunks):
"""(ctx, [crecord.uihunk]) -> memctx
return a memctx with some [1] patches (chunks) applied to ctx.
[1]: modifications are handled. renames, mode changes, etc. are ignored.
"""
# sadly the applying-patch logic is hardly reusable, and messy:
# 1. the core logic "_applydiff" is too heavy - it writes .rej files, it
# needs a file stream of a patch and will re-parse it, while we have
# structured hunk objects at hand.
# 2. a lot of different implementations about "chunk" (patch.hunk,
# patch.recordhunk, crecord.uihunk)
# as we only care about applying changes to modified files, no mode
# change, no binary diff, and no renames, it's probably okay to
# re-invent the logic using much simpler code here.
memworkingcopy = {} # {path: content}
patchmap = defaultdict(lambda: []) # {path: [(a1, a2, [bline])]}
for path, info in map(_parsechunk, chunks):
if not path or not info:
continue
patchmap[path].append(info)
for path, patches in patchmap.iteritems():
if path not in ctx or not patches:
continue
patches.sort(reverse=True)
lines = mdiff.splitnewlines(ctx[path].data())
for a1, a2, blines in patches:
lines[a1:a2] = blines
memworkingcopy[path] = ''.join(lines)
return overlaycontext(memworkingcopy, ctx)
def absorb(ui, repo, stack=None, targetctx=None, pats=None, opts=None):
"""pick fixup chunks from targetctx, apply them to stack.
if targetctx is None, the working copy context will be used.
if stack is None, the current draft stack will be used.
return fixupstate.
"""
if stack is None:
limit = ui.configint('absorb', 'max-stack-size')
stack = getdraftstack(repo['.'], limit)
if limit and len(stack) >= limit:
ui.warn(_('absorb: only the recent %d changesets will '
'be analysed\n')
% limit)
if not stack:
raise error.Abort(_('no mutable changeset to change'))
if targetctx is None: # default to working copy
targetctx = repo[None]
if pats is None:
pats = ()
if opts is None:
opts = {}
state = fixupstate(stack, ui=ui, opts=opts)
matcher = scmutil.match(targetctx, pats, opts)
if opts.get('interactive'):
diff = patch.diff(repo, stack[-1].node(), targetctx.node(), matcher)
origchunks = patch.parsepatch(diff)
chunks = cmdutil.recordfilter(ui, origchunks)[0]
targetctx = overlaydiffcontext(stack[-1], chunks)
fm = None
if opts.get('print_changes') or not opts.get('apply_changes'):
fm = ui.formatter('absorb', opts)
state.diffwith(targetctx, matcher, fm)
if fm is not None:
fm.startitem()
fm.write("count", "\n%d changesets affected\n", len(state.ctxaffected))
fm.data(linetype='summary')
for ctx in reversed(stack):
if ctx not in state.ctxaffected:
continue
fm.startitem()
fm.context(ctx=ctx)
fm.data(linetype='changeset')
fm.write('node', '%-7.7s ', ctx.hex(), label='absorb.node')
descfirstline = ctx.description().splitlines()[0]
fm.write('descfirstline', '%s\n', descfirstline,
label='absorb.description')
fm.end()
if not opts.get('dry_run'):
if (not opts.get('apply_changes') and
state.ctxaffected and
ui.promptchoice("apply changes (yn)? $$ &Yes $$ &No", default=1)):
raise error.Abort(_('absorb cancelled\n'))
state.apply()
if state.commit():
state.printchunkstats()
elif not ui.quiet:
ui.write(_('nothing applied\n'))
return state
@command('absorb',
[('a', 'apply-changes', None,
_('apply changes without prompting for confirmation')),
('p', 'print-changes', None,
_('always print which changesets are modified by which changes')),
('i', 'interactive', None,
_('interactively select which chunks to apply (EXPERIMENTAL)')),
('e', 'edit-lines', None,
_('edit what lines belong to which changesets before commit '
'(EXPERIMENTAL)')),
] + commands.dryrunopts + commands.templateopts + commands.walkopts,
_('hg absorb [OPTION] [FILE]...'),
helpcategory=command.CATEGORY_COMMITTING,
helpbasic=True)
def absorbcmd(ui, repo, *pats, **opts):
"""incorporate corrections into the stack of draft changesets
absorb analyzes each change in your working directory and attempts to
amend the changed lines into the changesets in your stack that first
introduced those lines.
If absorb cannot find an unambiguous changeset to amend for a change,
that change will be left in the working directory, untouched. They can be
observed by :hg:`status` or :hg:`diff` afterwards. In other words,
absorb does not write to the working directory.
Changesets outside the revset `::. and not public() and not merge()` will
not be changed.
Changesets that become empty after applying the changes will be deleted.
By default, absorb will show what it plans to do and prompt for
confirmation. If you are confident that the changes will be absorbed
to the correct place, run :hg:`absorb -a` to apply the changes
immediately.
Returns 0 on success, 1 if all chunks were ignored and nothing amended.
"""
opts = pycompat.byteskwargs(opts)
state = absorb(ui, repo, pats=pats, opts=opts)
if sum(s[0] for s in state.chunkstats.values()) == 0:
return 1