context.py
2597 lines
| 91.5 KiB
| text/x-python
|
PythonLexer
/ mercurial / context.py
Matt Mackall
|
r2563 | # context.py - changeset and file context objects for mercurial | ||
# | ||||
Thomas Arendsen Hein
|
r4635 | # Copyright 2006, 2007 Matt Mackall <mpm@selenic.com> | ||
Matt Mackall
|
r2563 | # | ||
Martin Geisler
|
r8225 | # This software may be used and distributed according to the terms of the | ||
Matt Mackall
|
r10263 | # GNU General Public License version 2 or any later version. | ||
Matt Mackall
|
r2563 | |||
Gregory Szorc
|
r27506 | from __future__ import absolute_import | ||
import errno | ||||
Phil Cohen
|
r34686 | import filecmp | ||
Gregory Szorc
|
r27506 | import os | ||
Mads Kiilerich
|
r26604 | import re | ||
Gregory Szorc
|
r27506 | import stat | ||
Mads Kiilerich
|
r26604 | |||
Gregory Szorc
|
r27506 | from .i18n import _ | ||
from .node import ( | ||||
Durham Goode
|
r30361 | addednodeid, | ||
Gregory Szorc
|
r27506 | bin, | ||
hex, | ||||
Durham Goode
|
r30361 | modifiednodeid, | ||
Gregory Szorc
|
r27506 | nullid, | ||
nullrev, | ||||
short, | ||||
wdirid, | ||||
Durham Goode
|
r31258 | wdirnodes, | ||
Yuya Nishihara
|
r32658 | wdirrev, | ||
Gregory Szorc
|
r27506 | ) | ||
Siddharth Agarwal
|
r34433 | from .thirdparty import ( | ||
attr, | ||||
) | ||||
Gregory Szorc
|
r27506 | from . import ( | ||
encoding, | ||||
error, | ||||
fileset, | ||||
match as matchmod, | ||||
mdiff, | ||||
obsolete as obsmod, | ||||
patch, | ||||
Matt Harbison
|
r33501 | pathutil, | ||
Gregory Szorc
|
r27506 | phases, | ||
Augie Fackler
|
r31343 | pycompat, | ||
Gregory Szorc
|
r27506 | repoview, | ||
revlog, | ||||
scmutil, | ||||
Gregory Szorc
|
r33353 | sparse, | ||
Gregory Szorc
|
r27506 | subrepo, | ||
util, | ||||
) | ||||
Matt Mackall
|
r3122 | |||
Matt Mackall
|
r8207 | propertycache = util.propertycache | ||
Dirkjan Ochtman
|
r7368 | |||
Mads Kiilerich
|
r26604 | nonascii = re.compile(r'[^\x21-\x7f]').search | ||
Sean Farley
|
r19537 | class basectx(object): | ||
"""A basectx object represents the common logic for its children: | ||||
changectx: read-only context that is already present in the repo, | ||||
workingctx: a context that represents the working directory and can | ||||
be committed, | ||||
memctx: a context that represents changes in-memory and can also | ||||
be committed.""" | ||||
def __new__(cls, repo, changeid='', *args, **kwargs): | ||||
Sean Farley
|
r19538 | if isinstance(changeid, basectx): | ||
return changeid | ||||
o = super(basectx, cls).__new__(cls) | ||||
o._repo = repo | ||||
o._rev = nullrev | ||||
o._node = nullid | ||||
return o | ||||
Sean Farley
|
r19537 | |||
Augie Fackler
|
r31344 | def __bytes__(self): | ||
Sean Farley
|
r19540 | return short(self.node()) | ||
Yuya Nishihara
|
r33022 | __str__ = encoding.strmethod(__bytes__) | ||
Sean Farley
|
r19545 | def __int__(self): | ||
return self.rev() | ||||
Sean Farley
|
r19546 | def __repr__(self): | ||
Pulkit Goyal
|
r32613 | return r"<%s %s>" % (type(self).__name__, str(self)) | ||
Sean Farley
|
r19546 | |||
Sean Farley
|
r19547 | def __eq__(self, other): | ||
try: | ||||
return type(self) == type(other) and self._rev == other._rev | ||||
except AttributeError: | ||||
return False | ||||
Sean Farley
|
r19548 | def __ne__(self, other): | ||
return not (self == other) | ||||
Sean Farley
|
r19550 | def __contains__(self, key): | ||
return key in self._manifest | ||||
Sean Farley
|
r19551 | def __getitem__(self, key): | ||
return self.filectx(key) | ||||
Sean Farley
|
r19552 | def __iter__(self): | ||
Augie Fackler
|
r24227 | return iter(self._manifest) | ||
Sean Farley
|
r19552 | |||
Durham Goode
|
r31261 | def _buildstatusmanifest(self, status): | ||
"""Builds a manifest that includes the given status results, if this is | ||||
a working copy context. For non-working copy contexts, it just returns | ||||
the normal manifest.""" | ||||
return self.manifest() | ||||
Siddharth Agarwal
|
r21880 | |||
Martin von Zweigbergk
|
r23237 | def _matchstatus(self, other, match): | ||
Martin von Zweigbergk
|
r33937 | """This internal method provides a way for child objects to override the | ||
Sean Farley
|
r21481 | match operator. | ||
""" | ||||
Martin von Zweigbergk
|
r33937 | return match | ||
Sean Farley
|
r21481 | |||
Sean Farley
|
r21471 | def _buildstatus(self, other, s, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown): | ||
Sean Farley
|
r21471 | """build a status with respect to another context""" | ||
Martin von Zweigbergk
|
r23257 | # Load earliest manifest first for caching reasons. More specifically, | ||
# if you have revisions 1000 and 1001, 1001 is probably stored as a | ||||
# delta against 1000. Thus, if you read 1000 first, we'll reconstruct | ||||
# 1000 and cache it so that when you read 1001, we just need to apply a | ||||
# delta to what's in the cache. So that's one full reconstruction + one | ||||
# delta application. | ||||
Durham Goode
|
r31260 | mf2 = None | ||
Martin von Zweigbergk
|
r23238 | if self.rev() is not None and self.rev() < other.rev(): | ||
Durham Goode
|
r31261 | mf2 = self._buildstatusmanifest(s) | ||
mf1 = other._buildstatusmanifest(s) | ||||
Durham Goode
|
r31260 | if mf2 is None: | ||
Durham Goode
|
r31261 | mf2 = self._buildstatusmanifest(s) | ||
Sean Farley
|
r21471 | |||
Augie Fackler
|
r23755 | modified, added = [], [] | ||
removed = [] | ||||
Augie Fackler
|
r23757 | clean = [] | ||
Martin von Zweigbergk
|
r23304 | deleted, unknown, ignored = s.deleted, s.unknown, s.ignored | ||
Martin von Zweigbergk
|
r23085 | deletedset = set(deleted) | ||
Durham Goode
|
r31261 | d = mf1.diff(mf2, match=match, clean=listclean) | ||
Augie Fackler
|
r23757 | for fn, value in d.iteritems(): | ||
Martin von Zweigbergk
|
r23731 | if fn in deletedset: | ||
continue | ||||
Augie Fackler
|
r23757 | if value is None: | ||
clean.append(fn) | ||||
continue | ||||
(node1, flag1), (node2, flag2) = value | ||||
Augie Fackler
|
r23755 | if node1 is None: | ||
added.append(fn) | ||||
elif node2 is None: | ||||
removed.append(fn) | ||||
Martin von Zweigbergk
|
r27749 | elif flag1 != flag2: | ||
modified.append(fn) | ||||
Durham Goode
|
r31258 | elif node2 not in wdirnodes: | ||
Martin von Zweigbergk
|
r27748 | # When comparing files between two commits, we save time by | ||
# not comparing the file contents when the nodeids differ. | ||||
# Note that this means we incorrectly report a reverted change | ||||
# to a file as a modification. | ||||
Martin von Zweigbergk
|
r27747 | modified.append(fn) | ||
Augie Fackler
|
r23755 | elif self[fn].cmp(other[fn]): | ||
modified.append(fn) | ||||
Martin von Zweigbergk
|
r23731 | else: | ||
Augie Fackler
|
r23757 | clean.append(fn) | ||
Augie Fackler
|
r23755 | |||
Pierre-Yves David
|
r21971 | if removed: | ||
# need to filter files if they are already reported as removed | ||||
Durham Goode
|
r31261 | unknown = [fn for fn in unknown if fn not in mf1 and | ||
(not match or match(fn))] | ||||
ignored = [fn for fn in ignored if fn not in mf1 and | ||||
(not match or match(fn))] | ||||
Martin von Zweigbergk
|
r23730 | # if they're deleted, don't report them as removed | ||
removed = [fn for fn in removed if fn not in deletedset] | ||||
Sean Farley
|
r21471 | |||
Martin von Zweigbergk
|
r23302 | return scmutil.status(modified, added, removed, deleted, unknown, | ||
ignored, clean) | ||||
Sean Farley
|
r21471 | |||
Sean Farley
|
r19549 | @propertycache | ||
def substate(self): | ||||
return subrepo.state(self, self._repo.ui) | ||||
Sean Farley
|
r21586 | def subrev(self, subpath): | ||
return self.substate[subpath][1] | ||||
Sean Farley
|
r19541 | def rev(self): | ||
return self._rev | ||||
Sean Farley
|
r19542 | def node(self): | ||
return self._node | ||||
Sean Farley
|
r19543 | def hex(self): | ||
Sean Farley
|
r19544 | return hex(self.node()) | ||
Sean Farley
|
r19553 | def manifest(self): | ||
return self._manifest | ||||
Durham Goode
|
r30344 | def manifestctx(self): | ||
return self._manifestctx | ||||
Matt Harbison
|
r24300 | def repo(self): | ||
return self._repo | ||||
Sean Farley
|
r19554 | def phasestr(self): | ||
return phases.phasenames[self.phase()] | ||||
Sean Farley
|
r19555 | def mutable(self): | ||
return self.phase() > phases.public | ||||
Sean Farley
|
r19541 | |||
Augie Fackler
|
r20400 | def getfileset(self, expr): | ||
return fileset.getfileset(self, expr) | ||||
Sean Farley
|
r19734 | def obsolete(self): | ||
"""True if the changeset is obsolete""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'obsolete') | ||||
def extinct(self): | ||||
"""True if the changeset is extinct""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'extinct') | ||||
def unstable(self): | ||||
Boris Feld
|
r33693 | msg = ("'context.unstable' is deprecated, " | ||
"use 'context.orphan'") | ||||
self._repo.ui.deprecwarn(msg, '4.4') | ||||
return self.orphan() | ||||
def orphan(self): | ||||
Sean Farley
|
r19734 | """True if the changeset is not obsolete but it's ancestor are""" | ||
Boris Feld
|
r33772 | return self.rev() in obsmod.getrevs(self._repo, 'orphan') | ||
Sean Farley
|
r19734 | |||
def bumped(self): | ||||
Boris Feld
|
r33695 | msg = ("'context.bumped' is deprecated, " | ||
"use 'context.phasedivergent'") | ||||
self._repo.ui.deprecwarn(msg, '4.4') | ||||
return self.phasedivergent() | ||||
def phasedivergent(self): | ||||
Sean Farley
|
r19734 | """True if the changeset try to be a successor of a public changeset | ||
Only non-public and non-obsolete changesets may be bumped. | ||||
""" | ||||
Boris Feld
|
r33774 | return self.rev() in obsmod.getrevs(self._repo, 'phasedivergent') | ||
Sean Farley
|
r19734 | |||
def divergent(self): | ||||
Boris Feld
|
r33694 | msg = ("'context.divergent' is deprecated, " | ||
"use 'context.contentdivergent'") | ||||
self._repo.ui.deprecwarn(msg, '4.4') | ||||
return self.contentdivergent() | ||||
def contentdivergent(self): | ||||
Sean Farley
|
r19734 | """Is a successors of a changeset with multiple possible successors set | ||
Only non-public and non-obsolete changesets may be divergent. | ||||
""" | ||||
Boris Feld
|
r33773 | return self.rev() in obsmod.getrevs(self._repo, 'contentdivergent') | ||
Sean Farley
|
r19734 | |||
def troubled(self): | ||||
Boris Feld
|
r33696 | msg = ("'context.troubled' is deprecated, " | ||
"use 'context.isunstable'") | ||||
self._repo.ui.deprecwarn(msg, '4.4') | ||||
Boris Feld
|
r33794 | return self.isunstable() | ||
Boris Feld
|
r33696 | |||
def isunstable(self): | ||||
Sean Farley
|
r19734 | """True if the changeset is either unstable, bumped or divergent""" | ||
Boris Feld
|
r33695 | return self.orphan() or self.phasedivergent() or self.contentdivergent() | ||
Sean Farley
|
r19734 | |||
def troubles(self): | ||||
Boris Feld
|
r33692 | """Keep the old version around in order to avoid breaking extensions | ||
about different return values. | ||||
""" | ||||
msg = ("'context.troubles' is deprecated, " | ||||
"use 'context.instabilities'") | ||||
self._repo.ui.deprecwarn(msg, '4.4') | ||||
Sean Farley
|
r19734 | |||
Boris Feld
|
r33692 | troubles = [] | ||
Boris Feld
|
r33693 | if self.orphan(): | ||
Boris Feld
|
r33692 | troubles.append('orphan') | ||
Boris Feld
|
r33695 | if self.phasedivergent(): | ||
Boris Feld
|
r33692 | troubles.append('bumped') | ||
Boris Feld
|
r33694 | if self.contentdivergent(): | ||
Boris Feld
|
r33692 | troubles.append('divergent') | ||
return troubles | ||||
def instabilities(self): | ||||
"""return the list of instabilities affecting this changeset. | ||||
Instabilities are returned as strings. possible values are: | ||||
Boris Feld
|
r33632 | - orphan, | ||
Boris Feld
|
r33652 | - phase-divergent, | ||
Boris Feld
|
r33651 | - content-divergent. | ||
Sean Farley
|
r19734 | """ | ||
Boris Feld
|
r33692 | instabilities = [] | ||
Boris Feld
|
r33693 | if self.orphan(): | ||
Boris Feld
|
r33692 | instabilities.append('orphan') | ||
Boris Feld
|
r33695 | if self.phasedivergent(): | ||
Boris Feld
|
r33692 | instabilities.append('phase-divergent') | ||
Boris Feld
|
r33694 | if self.contentdivergent(): | ||
Boris Feld
|
r33692 | instabilities.append('content-divergent') | ||
return instabilities | ||||
Sean Farley
|
r19734 | |||
Sean Farley
|
r19556 | def parents(self): | ||
"""return contexts for each parent changeset""" | ||||
return self._parents | ||||
Sean Farley
|
r19557 | def p1(self): | ||
return self._parents[0] | ||||
Sean Farley
|
r19558 | def p2(self): | ||
Gregory Szorc
|
r27064 | parents = self._parents | ||
if len(parents) == 2: | ||||
return parents[1] | ||||
return changectx(self._repo, nullrev) | ||||
Sean Farley
|
r19558 | |||
Sean Farley
|
r19559 | def _fileinfo(self, path): | ||
Pulkit Goyal
|
r32148 | if r'_manifest' in self.__dict__: | ||
Sean Farley
|
r19559 | try: | ||
return self._manifest[path], self._manifest.flags(path) | ||||
except KeyError: | ||||
raise error.ManifestLookupError(self._node, path, | ||||
_('not found in manifest')) | ||||
Pulkit Goyal
|
r32148 | if r'_manifestdelta' in self.__dict__ or path in self.files(): | ||
Sean Farley
|
r19559 | if path in self._manifestdelta: | ||
return (self._manifestdelta[path], | ||||
self._manifestdelta.flags(path)) | ||||
Durham Goode
|
r30340 | mfl = self._repo.manifestlog | ||
try: | ||||
node, flag = mfl[self._changeset.manifest].find(path) | ||||
except KeyError: | ||||
Sean Farley
|
r19559 | raise error.ManifestLookupError(self._node, path, | ||
_('not found in manifest')) | ||||
return node, flag | ||||
Sean Farley
|
r19560 | def filenode(self, path): | ||
return self._fileinfo(path)[0] | ||||
Sean Farley
|
r19561 | def flags(self, path): | ||
try: | ||||
return self._fileinfo(path)[1] | ||||
except error.LookupError: | ||||
return '' | ||||
Matt Harbison
|
r29021 | def sub(self, path, allowcreate=True): | ||
Matt Harbison
|
r25600 | '''return a subrepo for the stored revision of path, never wdir()''' | ||
Matt Harbison
|
r29021 | return subrepo.subrepo(self, path, allowcreate=allowcreate) | ||
Sean Farley
|
r19562 | |||
Matt Harbison
|
r25417 | def nullsub(self, path, pctx): | ||
return subrepo.nullsubrepo(self, path, pctx) | ||||
Matt Harbison
|
r25600 | def workingsub(self, path): | ||
'''return a subrepo for the stored revision, or wdir if this is a wdir | ||||
context. | ||||
''' | ||||
return subrepo.subrepo(self, path, allowwdir=True) | ||||
Gregory Szorc
|
r31388 | def match(self, pats=None, include=None, exclude=None, default='glob', | ||
Matt Harbison
|
r25465 | listsubrepos=False, badfn=None): | ||
Sean Farley
|
r19563 | r = self._repo | ||
Pierre-Yves David
|
r31437 | return matchmod.match(r.root, r.getcwd(), pats, | ||
Sean Farley
|
r19563 | include, exclude, default, | ||
Pierre-Yves David
|
r27234 | auditor=r.nofsauditor, ctx=self, | ||
Matt Harbison
|
r25465 | listsubrepos=listsubrepos, badfn=badfn) | ||
Sean Farley
|
r19563 | |||
Sean Farley
|
r19564 | def diff(self, ctx2=None, match=None, **opts): | ||
"""Returns a diff generator for the given contexts and matcher""" | ||||
if ctx2 is None: | ||||
ctx2 = self.p1() | ||||
Sean Farley
|
r19568 | if ctx2 is not None: | ||
Sean Farley
|
r19564 | ctx2 = self._repo[ctx2] | ||
diffopts = patch.diffopts(self._repo.ui, opts) | ||||
Sean Farley
|
r21834 | return patch.diff(self._repo, ctx2, self, match=match, opts=diffopts) | ||
Sean Farley
|
r19564 | |||
Drew Gottlieb
|
r24323 | def dirs(self): | ||
return self._manifest.dirs() | ||||
Sean Farley
|
r19565 | |||
Drew Gottlieb
|
r24325 | def hasdir(self, dir): | ||
return self._manifest.hasdir(dir) | ||||
Sean Farley
|
r19566 | |||
Sean Farley
|
r21594 | def status(self, other=None, match=None, listignored=False, | ||
listclean=False, listunknown=False, listsubrepos=False): | ||||
"""return status of files between two nodes or node and working | ||||
directory. | ||||
If other is None, compare this node with working directory. | ||||
Pierre-Yves David
|
r21722 | |||
returns (modified, added, removed, deleted, unknown, ignored, clean) | ||||
Sean Farley
|
r21594 | """ | ||
ctx1 = self | ||||
ctx2 = self._repo[other] | ||||
# This next code block is, admittedly, fragile logic that tests for | ||||
# reversing the contexts and wouldn't need to exist if it weren't for | ||||
# the fast (and common) code path of comparing the working directory | ||||
# with its first parent. | ||||
# | ||||
# What we're aiming for here is the ability to call: | ||||
# | ||||
# workingctx.status(parentctx) | ||||
# | ||||
# If we always built the manifest for each context and compared those, | ||||
# then we'd be done. But the special case of the above call means we | ||||
# just copy the manifest of the parent. | ||||
reversed = False | ||||
if (not isinstance(ctx1, changectx) | ||||
and isinstance(ctx2, changectx)): | ||||
reversed = True | ||||
ctx1, ctx2 = ctx2, ctx1 | ||||
Martin von Zweigbergk
|
r33937 | match = match or matchmod.always(self._repo.root, self._repo.getcwd()) | ||
Martin von Zweigbergk
|
r23237 | match = ctx2._matchstatus(ctx1, match) | ||
Martin von Zweigbergk
|
r23304 | r = scmutil.status([], [], [], [], [], [], []) | ||
Sean Farley
|
r21594 | r = ctx2._buildstatus(ctx1, r, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown) | ||
Sean Farley
|
r21594 | |||
if reversed: | ||||
Martin von Zweigbergk
|
r23301 | # Reverse added and removed. Clear deleted, unknown and ignored as | ||
# these make no sense to reverse. | ||||
r = scmutil.status(r.modified, r.removed, r.added, [], [], [], | ||||
r.clean) | ||||
Sean Farley
|
r21594 | |||
if listsubrepos: | ||||
for subpath, sub in scmutil.itersubrepos(ctx1, ctx2): | ||||
try: | ||||
Andrew Zwicky
|
r27183 | rev2 = ctx2.subrev(subpath) | ||
except KeyError: | ||||
# A subrepo that existed in node1 was deleted between | ||||
# node1 and node2 (inclusive). Thus, ctx2's substate | ||||
# won't contain that subpath. The best we can do ignore it. | ||||
rev2 = None | ||||
Martin von Zweigbergk
|
r28017 | submatch = matchmod.subdirmatcher(subpath, match) | ||
Andrew Zwicky
|
r27183 | s = sub.status(rev2, match=submatch, ignored=listignored, | ||
clean=listclean, unknown=listunknown, | ||||
listsubrepos=True) | ||||
for rfiles, sfiles in zip(r, s): | ||||
rfiles.extend("%s/%s" % (subpath, f) for f in sfiles) | ||||
Sean Farley
|
r21594 | |||
for l in r: | ||||
l.sort() | ||||
Sean Farley
|
r21616 | |||
Martin von Zweigbergk
|
r23301 | return r | ||
Sean Farley
|
r21594 | |||
Pierre-Yves David
|
r32006 | def _filterederror(repo, changeid): | ||
"""build an exception to be raised about a filtered changeid | ||||
This is extracted in a function to help extensions (eg: evolve) to | ||||
experiment with various message variants.""" | ||||
if repo.filtername.startswith('visible'): | ||||
msg = _("hidden revision '%s'") % changeid | ||||
hint = _('use --hidden to access hidden revisions') | ||||
return error.FilteredRepoLookupError(msg, hint=hint) | ||||
msg = _("filtered revision '%s' (not in '%s' subset)") | ||||
msg %= (changeid, repo.filtername) | ||||
return error.FilteredRepoLookupError(msg) | ||||
Sean Farley
|
r19537 | class changectx(basectx): | ||
Matt Mackall
|
r2563 | """A changecontext object makes access to data related to a particular | ||
Mads Kiilerich
|
r19951 | changeset convenient. It represents a read-only context already present in | ||
Sean Farley
|
r19537 | the repo.""" | ||
Matt Mackall
|
r6741 | def __init__(self, repo, changeid=''): | ||
Matt Mackall
|
r2563 | """changeid is a revision number, node, or tag""" | ||
Sean Farley
|
r19539 | |||
# since basectx.__new__ already took care of copying the object, we | ||||
# don't need to do anything in __init__, so we just exit here | ||||
if isinstance(changeid, basectx): | ||||
return | ||||
Matt Mackall
|
r6741 | if changeid == '': | ||
changeid = '.' | ||||
Matt Mackall
|
r2563 | self._repo = repo | ||
Matt Mackall
|
r16376 | |||
Pierre-Yves David
|
r23012 | try: | ||
if isinstance(changeid, int): | ||||
Pierre-Yves David
|
r23013 | self._node = repo.changelog.node(changeid) | ||
Pierre-Yves David
|
r23012 | self._rev = changeid | ||
return | ||||
Augie Fackler
|
r31343 | if not pycompat.ispy3 and isinstance(changeid, long): | ||
Pierre-Yves David
|
r23012 | changeid = str(changeid) | ||
if changeid == 'null': | ||||
self._node = nullid | ||||
self._rev = nullrev | ||||
return | ||||
if changeid == 'tip': | ||||
self._node = repo.changelog.tip() | ||||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Martin von Zweigbergk
|
r24050 | if changeid == '.' or changeid == repo.dirstate.p1(): | ||
# this is a hack to delay/avoid loading obsmarkers | ||||
# when we know that '.' won't be hidden | ||||
self._node = repo.dirstate.p1() | ||||
self._rev = repo.unfiltered().changelog.rev(self._node) | ||||
return | ||||
Pierre-Yves David
|
r23012 | if len(changeid) == 20: | ||
try: | ||||
self._node = changeid | ||||
self._rev = repo.changelog.rev(changeid) | ||||
return | ||||
Pierre-Yves David
|
r23017 | except error.FilteredRepoLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except LookupError: | ||
pass | ||||
Pierre-Yves David
|
r18084 | try: | ||
Pierre-Yves David
|
r23012 | r = int(changeid) | ||
Augie Fackler
|
r31351 | if '%d' % r != changeid: | ||
Pierre-Yves David
|
r23012 | raise ValueError | ||
l = len(repo.changelog) | ||||
if r < 0: | ||||
r += l | ||||
Yuya Nishihara
|
r32658 | if r < 0 or r >= l and r != wdirrev: | ||
Pierre-Yves David
|
r23012 | raise ValueError | ||
self._rev = r | ||||
self._node = repo.changelog.node(r) | ||||
Matt Mackall
|
r16376 | return | ||
Pierre-Yves David
|
r23017 | except error.FilteredIndexError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except (ValueError, OverflowError, IndexError): | ||
Matt Mackall
|
r16376 | pass | ||
Pierre-Yves David
|
r23012 | if len(changeid) == 40: | ||
try: | ||||
self._node = bin(changeid) | ||||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Pierre-Yves David
|
r23017 | except error.FilteredLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except (TypeError, LookupError): | ||
pass | ||||
Matt Mackall
|
r16376 | |||
Sean Farley
|
r23560 | # lookup bookmarks through the name interface | ||
try: | ||||
Ryan McElroy
|
r23561 | self._node = repo.names.singlenode(repo, changeid) | ||
Matt Mackall
|
r16376 | self._rev = repo.changelog.rev(self._node) | ||
return | ||||
Sean Farley
|
r23560 | except KeyError: | ||
pass | ||||
Pierre-Yves David
|
r23017 | except error.FilteredRepoLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except error.RepoLookupError: | ||
Matt Mackall
|
r16376 | pass | ||
Pierre-Yves David
|
r23017 | self._node = repo.unfiltered().changelog._partialmatch(changeid) | ||
Pierre-Yves David
|
r23012 | if self._node is not None: | ||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Matt Mackall
|
r16376 | |||
Pierre-Yves David
|
r23012 | # lookup failed | ||
# check if it might have come from damaged dirstate | ||||
# | ||||
# XXX we could avoid the unfiltered if we had a recognizable | ||||
# exception for filtered changeset access | ||||
if changeid in repo.unfiltered().dirstate.parents(): | ||||
msg = _("working directory has unknown parent '%s'!") | ||||
raise error.Abort(msg % short(changeid)) | ||||
try: | ||||
Mads Kiilerich
|
r26604 | if len(changeid) == 20 and nonascii(changeid): | ||
Pierre-Yves David
|
r23012 | changeid = hex(changeid) | ||
except TypeError: | ||||
pass | ||||
Pierre-Yves David
|
r23017 | except (error.FilteredIndexError, error.FilteredLookupError, | ||
error.FilteredRepoLookupError): | ||||
Pierre-Yves David
|
r32006 | raise _filterederror(repo, changeid) | ||
Pierre-Yves David
|
r23013 | except IndexError: | ||
pass | ||||
Matt Mackall
|
r16376 | raise error.RepoLookupError( | ||
_("unknown revision '%s'") % changeid) | ||||
Matt Mackall
|
r2563 | |||
Paul Moore
|
r6469 | def __hash__(self): | ||
try: | ||||
return hash(self._rev) | ||||
except AttributeError: | ||||
return id(self) | ||||
Matt Mackall
|
r3168 | def __nonzero__(self): | ||
Thomas Arendsen Hein
|
r3578 | return self._rev != nullrev | ||
Matt Mackall
|
r3168 | |||
Gregory Szorc
|
r31476 | __bool__ = __nonzero__ | ||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _changeset(self): | ||
Gregory Szorc
|
r28488 | return self._repo.changelog.changelogrevision(self.rev()) | ||
Dirkjan Ochtman
|
r7368 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _manifest(self): | ||
Durham Goode
|
r30344 | return self._manifestctx.read() | ||
Jun Wu
|
r32519 | @property | ||
Durham Goode
|
r30344 | def _manifestctx(self): | ||
return self._repo.manifestlog[self._changeset.manifest] | ||||
Dirkjan Ochtman
|
r7368 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _manifestdelta(self): | ||
Durham Goode
|
r30344 | return self._manifestctx.readdelta() | ||
Dirkjan Ochtman
|
r7368 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _parents(self): | ||
Gregory Szorc
|
r27063 | repo = self._repo | ||
p1, p2 = repo.changelog.parentrevs(self._rev) | ||||
if p2 == nullrev: | ||||
return [changectx(repo, p1)] | ||||
return [changectx(repo, p1), changectx(repo, p2)] | ||||
Matt Mackall
|
r3215 | |||
Matt Mackall
|
r10282 | def changeset(self): | ||
Gregory Szorc
|
r28488 | c = self._changeset | ||
return ( | ||||
c.manifest, | ||||
c.user, | ||||
c.date, | ||||
c.files, | ||||
c.description, | ||||
c.extra, | ||||
) | ||||
Matt Mackall
|
r10282 | def manifestnode(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.manifest | ||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r10282 | def user(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.user | ||
Matt Mackall
|
r10282 | def date(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.date | ||
Matt Mackall
|
r10282 | def files(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.files | ||
Matt Mackall
|
r10282 | def description(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.description | ||
Matt Mackall
|
r10282 | def branch(self): | ||
Gregory Szorc
|
r28488 | return encoding.tolocal(self._changeset.extra.get("branch")) | ||
Brodie Rao
|
r16720 | def closesbranch(self): | ||
Gregory Szorc
|
r28488 | return 'close' in self._changeset.extra | ||
Matt Mackall
|
r10282 | def extra(self): | ||
Gregory Szorc
|
r28488 | return self._changeset.extra | ||
Matt Mackall
|
r10282 | def tags(self): | ||
return self._repo.nodetags(self._node) | ||||
David Soria Parra
|
r13384 | def bookmarks(self): | ||
return self._repo.nodebookmarks(self._node) | ||||
Pierre-Yves David
|
r15421 | def phase(self): | ||
Patrick Mezard
|
r16657 | return self._repo._phasecache.phase(self._repo, self._rev) | ||
Pierre-Yves David
|
r14644 | def hidden(self): | ||
Kevin Bullock
|
r18382 | return self._rev in repoview.filterrevs(self._repo, 'visible') | ||
Matt Mackall
|
r2563 | |||
Phil Cohen
|
r34682 | def isinmemory(self): | ||
return False | ||||
Matt Mackall
|
r2563 | def children(self): | ||
"""return contexts for each child changeset""" | ||||
Benoit Boissinot
|
r2627 | c = self._repo.changelog.children(self._node) | ||
Thomas Arendsen Hein
|
r3673 | return [changectx(self._repo, x) for x in c] | ||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r6876 | def ancestors(self): | ||
Bryan O'Sullivan
|
r16866 | for a in self._repo.changelog.ancestors([self._rev]): | ||
Matt Mackall
|
r6876 | yield changectx(self._repo, a) | ||
def descendants(self): | ||||
Bryan O'Sullivan
|
r16867 | for d in self._repo.changelog.descendants([self._rev]): | ||
Matt Mackall
|
r6876 | yield changectx(self._repo, d) | ||
Benoit Boissinot
|
r3966 | def filectx(self, path, fileid=None, filelog=None): | ||
Matt Mackall
|
r2563 | """get a file context from this changeset""" | ||
Benoit Boissinot
|
r2628 | if fileid is None: | ||
fileid = self.filenode(path) | ||||
Benoit Boissinot
|
r3966 | return filectx(self._repo, path, fileid=fileid, | ||
changectx=self, filelog=filelog) | ||||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r21203 | def ancestor(self, c2, warn=False): | ||
Mads Kiilerich
|
r22389 | """return the "best" ancestor context of self and c2 | ||
If there are multiple candidates, it will show a message and check | ||||
merge.preferancestor configuration before falling back to the | ||||
revlog ancestor.""" | ||||
Matt Mackall
|
r9843 | # deal with workingctxs | ||
n2 = c2._node | ||||
Martin Geisler
|
r13031 | if n2 is None: | ||
Matt Mackall
|
r9843 | n2 = c2._parents[0]._node | ||
Mads Kiilerich
|
r21125 | cahs = self._repo.changelog.commonancestorsheads(self._node, n2) | ||
if not cahs: | ||||
anc = nullid | ||||
elif len(cahs) == 1: | ||||
anc = cahs[0] | ||||
else: | ||||
Matt Mackall
|
r25844 | # experimental config: merge.preferancestor | ||
Boris Feld
|
r34481 | for r in self._repo.ui.configlist('merge', 'preferancestor'): | ||
Mads Kiilerich
|
r22671 | try: | ||
ctx = changectx(self._repo, r) | ||||
except error.RepoLookupError: | ||||
Mads Kiilerich
|
r22180 | continue | ||
Mads Kiilerich
|
r21126 | anc = ctx.node() | ||
if anc in cahs: | ||||
break | ||||
else: | ||||
anc = self._repo.changelog.ancestor(self._node, n2) | ||||
Matt Mackall
|
r21203 | if warn: | ||
self._repo.ui.status( | ||||
(_("note: using %s as ancestor of %s and %s\n") % | ||||
(short(anc), short(self._node), short(n2))) + | ||||
''.join(_(" alternatively, use --config " | ||||
"merge.preferancestor=%s\n") % | ||||
short(n) for n in sorted(cahs) if n != anc)) | ||||
Mads Kiilerich
|
r21125 | return changectx(self._repo, anc) | ||
Matt Mackall
|
r3125 | |||
FUJIWARA Katsunori
|
r17626 | def descendant(self, other): | ||
"""True if other is descendant of this changeset""" | ||||
return self._repo.changelog.descendant(self._rev, other._rev) | ||||
Matt Mackall
|
r6764 | def walk(self, match): | ||
Drew Gottlieb
|
r24646 | '''Generates matching file names.''' | ||
Durham Goode
|
r20292 | |||
Matt Harbison
|
r25435 | # Wrap match.bad method to have message with nodeid | ||
Drew Gottlieb
|
r24646 | def bad(fn, msg): | ||
Matt Harbison
|
r25193 | # The manifest doesn't know about subrepos, so don't complain about | ||
# paths into valid subrepos. | ||||
Matt Mackall
|
r25195 | if any(fn == s or fn.startswith(s + '/') | ||
for s in self.substate): | ||||
Matt Harbison
|
r25193 | return | ||
Matt Harbison
|
r25435 | match.bad(fn, _('no such file in rev %s') % self) | ||
Durham Goode
|
r20292 | |||
Matt Harbison
|
r25435 | m = matchmod.badmatch(match, bad) | ||
return self._manifest.walk(m) | ||||
Matt Mackall
|
r6764 | |||
Siddharth Agarwal
|
r21985 | def matches(self, match): | ||
return self.walk(match) | ||||
Sean Farley
|
r19572 | class basefilectx(object): | ||
"""A filecontext object represents the common logic for its children: | ||||
filectx: read-only access to a filerevision that is already present | ||||
in the repo, | ||||
workingfilectx: a filecontext that represents files from the working | ||||
directory, | ||||
Jun Wu
|
r32239 | memfilectx: a filecontext that represents files in-memory, | ||
overlayfilectx: duplicate another filecontext with some fields overridden. | ||||
""" | ||||
Sean Farley
|
r19573 | @propertycache | ||
def _filelog(self): | ||||
return self._repo.file(self._path) | ||||
Sean Farley
|
r19574 | @propertycache | ||
def _changeid(self): | ||||
Pulkit Goyal
|
r32148 | if r'_changeid' in self.__dict__: | ||
Sean Farley
|
r19574 | return self._changeid | ||
Pulkit Goyal
|
r32148 | elif r'_changectx' in self.__dict__: | ||
Sean Farley
|
r19574 | return self._changectx.rev() | ||
Pulkit Goyal
|
r32148 | elif r'_descendantrev' in self.__dict__: | ||
Matt Mackall
|
r23983 | # this file context was created from a revision with a known | ||
# descendant, we can (lazily) correct for linkrev aliases | ||||
Jun Wu
|
r30275 | return self._adjustlinkrev(self._descendantrev) | ||
Sean Farley
|
r19574 | else: | ||
return self._filelog.linkrev(self._filerev) | ||||
Sean Farley
|
r19575 | @propertycache | ||
def _filenode(self): | ||||
Pulkit Goyal
|
r32148 | if r'_fileid' in self.__dict__: | ||
Sean Farley
|
r19575 | return self._filelog.lookup(self._fileid) | ||
else: | ||||
return self._changectx.filenode(self._path) | ||||
Sean Farley
|
r19576 | @propertycache | ||
def _filerev(self): | ||||
return self._filelog.rev(self._filenode) | ||||
Sean Farley
|
r19577 | @propertycache | ||
def _repopath(self): | ||||
return self._path | ||||
Sean Farley
|
r19578 | def __nonzero__(self): | ||
try: | ||||
self._filenode | ||||
return True | ||||
except error.LookupError: | ||||
# file is missing | ||||
return False | ||||
Gregory Szorc
|
r31476 | __bool__ = __nonzero__ | ||
Yuya Nishihara
|
r33022 | def __bytes__(self): | ||
Mads Kiilerich
|
r30270 | try: | ||
return "%s@%s" % (self.path(), self._changectx) | ||||
except error.LookupError: | ||||
return "%s@???" % self.path() | ||||
Sean Farley
|
r19579 | |||
Yuya Nishihara
|
r33022 | __str__ = encoding.strmethod(__bytes__) | ||
Pulkit Goyal
|
r33019 | |||
Sean Farley
|
r19580 | def __repr__(self): | ||
return "<%s %s>" % (type(self).__name__, str(self)) | ||||
Sean Farley
|
r19581 | def __hash__(self): | ||
try: | ||||
return hash((self._path, self._filenode)) | ||||
except AttributeError: | ||||
return id(self) | ||||
Sean Farley
|
r19582 | def __eq__(self, other): | ||
try: | ||||
return (type(self) == type(other) and self._path == other._path | ||||
and self._filenode == other._filenode) | ||||
except AttributeError: | ||||
return False | ||||
Sean Farley
|
r19583 | def __ne__(self, other): | ||
return not (self == other) | ||||
Sean Farley
|
r19584 | def filerev(self): | ||
return self._filerev | ||||
Sean Farley
|
r19585 | def filenode(self): | ||
return self._filenode | ||||
Jun Wu
|
r32234 | @propertycache | ||
def _flags(self): | ||||
return self._changectx.flags(self._path) | ||||
Sean Farley
|
r19586 | def flags(self): | ||
Jun Wu
|
r32234 | return self._flags | ||
Sean Farley
|
r19587 | def filelog(self): | ||
return self._filelog | ||||
Sean Farley
|
r19588 | def rev(self): | ||
return self._changeid | ||||
Sean Farley
|
r19589 | def linkrev(self): | ||
return self._filelog.linkrev(self._filerev) | ||||
Sean Farley
|
r19590 | def node(self): | ||
return self._changectx.node() | ||||
Sean Farley
|
r19591 | def hex(self): | ||
return self._changectx.hex() | ||||
Sean Farley
|
r19592 | def user(self): | ||
return self._changectx.user() | ||||
Sean Farley
|
r19593 | def date(self): | ||
return self._changectx.date() | ||||
Sean Farley
|
r19594 | def files(self): | ||
return self._changectx.files() | ||||
Sean Farley
|
r19595 | def description(self): | ||
return self._changectx.description() | ||||
Sean Farley
|
r19596 | def branch(self): | ||
return self._changectx.branch() | ||||
Sean Farley
|
r19597 | def extra(self): | ||
return self._changectx.extra() | ||||
Sean Farley
|
r19598 | def phase(self): | ||
return self._changectx.phase() | ||||
Sean Farley
|
r19599 | def phasestr(self): | ||
return self._changectx.phasestr() | ||||
Sean Farley
|
r19600 | def manifest(self): | ||
return self._changectx.manifest() | ||||
Sean Farley
|
r19601 | def changectx(self): | ||
return self._changectx | ||||
Jun Wu
|
r32235 | def renamed(self): | ||
return self._copied | ||||
Matt Harbison
|
r24333 | def repo(self): | ||
return self._repo | ||||
Jun Wu
|
r32236 | def size(self): | ||
return len(self.data()) | ||||
Sean Farley
|
r19584 | |||
Sean Farley
|
r19602 | def path(self): | ||
return self._path | ||||
Sean Farley
|
r19603 | def isbinary(self): | ||
try: | ||||
return util.binary(self.data()) | ||||
except IOError: | ||||
return False | ||||
Sean Farley
|
r22054 | def isexec(self): | ||
return 'x' in self.flags() | ||||
def islink(self): | ||||
return 'l' in self.flags() | ||||
Sean Farley
|
r19603 | |||
Siddharth Agarwal
|
r26978 | def isabsent(self): | ||
"""whether this filectx represents a file not in self._changectx | ||||
This is mainly for merge code to detect change/delete conflicts. This is | ||||
expected to be True for all subclasses of basectx.""" | ||||
return False | ||||
Siddharth Agarwal
|
r26977 | _customcmp = False | ||
Sean Farley
|
r19604 | def cmp(self, fctx): | ||
"""compare with other file context | ||||
returns True if different than fctx. | ||||
""" | ||||
Siddharth Agarwal
|
r26977 | if fctx._customcmp: | ||
return fctx.cmp(self) | ||||
Durham Goode
|
r28116 | if (fctx._filenode is None | ||
Sean Farley
|
r19604 | and (self._repo._encodefilterpats | ||
# if file data starts with '\1\n', empty metadata block is | ||||
# prepended, which adds 4 bytes to filelog.size(). | ||||
or self.size() - 4 == fctx.size()) | ||||
or self.size() == fctx.size()): | ||||
return self._filelog.cmp(self._filenode, fctx.data()) | ||||
return True | ||||
Jun Wu
|
r30275 | def _adjustlinkrev(self, srcrev, inclusive=False): | ||
Mads Kiilerich
|
r24180 | """return the first ancestor of <srcrev> introducing <fnode> | ||
Pierre-Yves David
|
r23979 | |||
If the linkrev of the file revision does not point to an ancestor of | ||||
srcrev, we'll walk down the ancestors until we find one introducing | ||||
this file revision. | ||||
:srcrev: the changeset revision we search ancestors from | ||||
:inclusive: if true, the src revision will also be checked | ||||
""" | ||||
repo = self._repo | ||||
cl = repo.unfiltered().changelog | ||||
Durham Goode
|
r29939 | mfl = repo.manifestlog | ||
Pierre-Yves David
|
r23979 | # fetch the linkrev | ||
Jun Wu
|
r30275 | lkr = self.linkrev() | ||
Pierre-Yves David
|
r23980 | # hack to reuse ancestor computation when searching for renames | ||
memberanc = getattr(self, '_ancestrycontext', None) | ||||
iteranc = None | ||||
Pierre-Yves David
|
r24411 | if srcrev is None: | ||
# wctx case, used by workingfilectx during mergecopy | ||||
revs = [p.rev() for p in self._repo[None].parents()] | ||||
inclusive = True # we skipped the real (revless) source | ||||
else: | ||||
revs = [srcrev] | ||||
Pierre-Yves David
|
r23980 | if memberanc is None: | ||
Pierre-Yves David
|
r24411 | memberanc = iteranc = cl.ancestors(revs, lkr, | ||
inclusive=inclusive) | ||||
Pierre-Yves David
|
r23979 | # check if this linkrev is an ancestor of srcrev | ||
Pierre-Yves David
|
r23980 | if lkr not in memberanc: | ||
if iteranc is None: | ||||
Pierre-Yves David
|
r24410 | iteranc = cl.ancestors(revs, lkr, inclusive=inclusive) | ||
Jun Wu
|
r30275 | fnode = self._filenode | ||
path = self._path | ||||
Pierre-Yves David
|
r23980 | for a in iteranc: | ||
Pierre-Yves David
|
r23979 | ac = cl.read(a) # get changeset data (we avoid object creation) | ||
if path in ac[3]: # checking the 'files' field. | ||||
# The file has been touched, check if the content is | ||||
# similar to the one we search for. | ||||
Durham Goode
|
r29939 | if fnode == mfl[ac[0]].readfast().get(path): | ||
Pierre-Yves David
|
r23979 | return a | ||
# In theory, we should never get out of that loop without a result. | ||||
# But if manifest uses a buggy file revision (not children of the | ||||
# one it replaces) we could. Such a buggy situation will likely | ||||
# result is crash somewhere else at to some point. | ||||
return lkr | ||||
Pierre-Yves David
|
r23703 | def introrev(self): | ||
"""return the rev of the changeset which introduced this file revision | ||||
This method is different from linkrev because it take into account the | ||||
changeset the filectx was created from. It ensures the returned | ||||
revision is one of its ancestors. This prevents bugs from | ||||
'linkrev-shadowing' when a file revision is used by multiple | ||||
changesets. | ||||
""" | ||||
lkr = self.linkrev() | ||||
attrs = vars(self) | ||||
noctx = not ('_changeid' in attrs or '_changectx' in attrs) | ||||
if noctx or self.rev() == lkr: | ||||
return self.linkrev() | ||||
Jun Wu
|
r30275 | return self._adjustlinkrev(self.rev(), inclusive=True) | ||
Pierre-Yves David
|
r23703 | |||
Yuya Nishihara
|
r24816 | def _parentfilectx(self, path, fileid, filelog): | ||
"""create parent filectx keeping ancestry info for _adjustlinkrev()""" | ||||
fctx = filectx(self._repo, path, fileid=fileid, filelog=filelog) | ||||
if '_changeid' in vars(self) or '_changectx' in vars(self): | ||||
# If self is associated with a changeset (probably explicitly | ||||
# fed), ensure the created filectx is associated with a | ||||
# changeset that is an ancestor of self.changectx. | ||||
# This lets us later use _adjustlinkrev to get a correct link. | ||||
fctx._descendantrev = self.rev() | ||||
fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) | ||||
elif '_descendantrev' in vars(self): | ||||
# Otherwise propagate _descendantrev if we have one associated. | ||||
fctx._descendantrev = self._descendantrev | ||||
fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) | ||||
return fctx | ||||
Sean Farley
|
r19605 | def parents(self): | ||
Mads Kiilerich
|
r22201 | _path = self._path | ||
Sean Farley
|
r19605 | fl = self._filelog | ||
Pierre-Yves David
|
r23688 | parents = self._filelog.parents(self._filenode) | ||
pl = [(_path, node, fl) for node in parents if node != nullid] | ||||
Sean Farley
|
r19605 | |||
Pierre-Yves David
|
r23702 | r = fl.renamed(self._filenode) | ||
Sean Farley
|
r19605 | if r: | ||
Pierre-Yves David
|
r23688 | # - In the simple rename case, both parent are nullid, pl is empty. | ||
# - In case of merge, only one of the parent is null id and should | ||||
# be replaced with the rename information. This parent is -always- | ||||
# the first one. | ||||
# | ||||
Mads Kiilerich
|
r24180 | # As null id have always been filtered out in the previous list | ||
Pierre-Yves David
|
r23688 | # comprehension, inserting to 0 will always result in "replacing | ||
# first nullid parent with rename information. | ||||
Pierre-Yves David
|
r23699 | pl.insert(0, (r[0], r[1], self._repo.file(r[0]))) | ||
Sean Farley
|
r19605 | |||
Yuya Nishihara
|
r24816 | return [self._parentfilectx(path, fnode, l) for path, fnode, l in pl] | ||
Sean Farley
|
r19605 | |||
Sean Farley
|
r19606 | def p1(self): | ||
return self.parents()[0] | ||||
Sean Farley
|
r19607 | def p2(self): | ||
p = self.parents() | ||||
if len(p) == 2: | ||||
return p[1] | ||||
return filectx(self._repo, self._path, fileid=-1, filelog=self._filelog) | ||||
Siddharth Agarwal
|
r32486 | def annotate(self, follow=False, linenumber=False, skiprevs=None, | ||
diffopts=None): | ||||
Denis Laxalde
|
r29527 | '''returns a list of tuples of ((ctx, number), line) for each line | ||
Brendan Cully
|
r3172 | in the file, where ctx is the filectx of the node where | ||
Denis Laxalde
|
r29527 | that line was last changed; if linenumber parameter is true, number is | ||
the line number at the first appearance in the managed file, otherwise, | ||||
number has a fixed value of False. | ||||
''' | ||||
Brendan Cully
|
r3172 | |||
Matt Mackall
|
r29223 | def lines(text): | ||
if text.endswith("\n"): | ||||
return text.count("\n") | ||||
Jun Wu
|
r30040 | return text.count("\n") + int(bool(text)) | ||
Matt Mackall
|
r29223 | |||
Denis Laxalde
|
r29527 | if linenumber: | ||
Yuya Nishihara
|
r22192 | def decorate(text, rev): | ||
Siddharth Agarwal
|
r34433 | return ([annotateline(fctx=rev, lineno=i) | ||
for i in xrange(1, lines(text) + 1)], text) | ||||
Yuya Nishihara
|
r22191 | else: | ||
Yuya Nishihara
|
r22192 | def decorate(text, rev): | ||
Siddharth Agarwal
|
r34433 | return ([annotateline(fctx=rev)] * lines(text), text) | ||
FUJIWARA Katsunori
|
r4856 | |||
Matt Mackall
|
r9097 | getlog = util.lrucachefunc(lambda x: self._repo.file(x)) | ||
Brendan Cully
|
r3172 | |||
def parents(f): | ||||
Yuya Nishihara
|
r24862 | # Cut _descendantrev here to mitigate the penalty of lazy linkrev | ||
# adjustment. Otherwise, p._adjustlinkrev() would walk changelog | ||||
# from the topmost introrev (= srcrev) down to p.linkrev() if it | ||||
# isn't an ancestor of the srcrev. | ||||
f._changeid | ||||
Durham Goode
|
r19292 | pl = f.parents() | ||
# Don't return renamed parents if we aren't following. | ||||
if not follow: | ||||
pl = [p for p in pl if p.path() == f.path()] | ||||
Brendan Cully
|
r3172 | |||
Durham Goode
|
r19292 | # renamed filectx won't have a filelog yet, so set it | ||
# from the cache to save time | ||||
for p in pl: | ||||
if not '_filelog' in p.__dict__: | ||||
p._filelog = getlog(p.path()) | ||||
Brendan Cully
|
r3146 | |||
Durham Goode
|
r19292 | return pl | ||
Matt Mackall
|
r3217 | |||
Brendan Cully
|
r3404 | # use linkrev to find the first changeset where self appeared | ||
Pierre-Yves David
|
r23705 | base = self | ||
introrev = self.introrev() | ||||
if self.rev() != introrev: | ||||
Durham Goode
|
r23770 | base = self.filectx(self.filenode(), changeid=introrev) | ||
Yuya Nishihara
|
r24818 | if getattr(base, '_ancestrycontext', None) is None: | ||
cl = self._repo.changelog | ||||
if introrev is None: | ||||
# wctx is not inclusive, but works because _ancestrycontext | ||||
# is used to test filelog revisions | ||||
ac = cl.ancestors([p.rev() for p in base.parents()], | ||||
inclusive=True) | ||||
else: | ||||
ac = cl.ancestors([introrev], inclusive=True) | ||||
Pierre-Yves David
|
r24407 | base._ancestrycontext = ac | ||
Brendan Cully
|
r3404 | |||
Matt Mackall
|
r13552 | # This algorithm would prefer to be recursive, but Python is a | ||
# bit recursion-hostile. Instead we do an iterative | ||||
# depth-first search. | ||||
Jun Wu
|
r29861 | # 1st DFS pre-calculates pcache and needed | ||
Matt Mackall
|
r13552 | visit = [base] | ||
pcache = {} | ||||
Brendan Cully
|
r3404 | needed = {base: 1} | ||
Brendan Cully
|
r3172 | while visit: | ||
Jun Wu
|
r29861 | f = visit.pop() | ||
if f in pcache: | ||||
continue | ||||
pl = parents(f) | ||||
pcache[f] = pl | ||||
for p in pl: | ||||
needed[p] = needed.get(p, 0) + 1 | ||||
if p not in pcache: | ||||
visit.append(p) | ||||
# 2nd DFS does the actual annotate | ||||
visit[:] = [base] | ||||
hist = {} | ||||
while visit: | ||||
Matt Mackall
|
r13552 | f = visit[-1] | ||
Jun Wu
|
r29861 | if f in hist: | ||
visit.pop() | ||||
continue | ||||
Brendan Cully
|
r3172 | |||
Matt Mackall
|
r13552 | ready = True | ||
pl = pcache[f] | ||||
for p in pl: | ||||
if p not in hist: | ||||
ready = False | ||||
visit.append(p) | ||||
if ready: | ||||
visit.pop() | ||||
Jun Wu
|
r29861 | curr = decorate(f.data(), f) | ||
Siddharth Agarwal
|
r32486 | skipchild = False | ||
if skiprevs is not None: | ||||
skipchild = f._changeid in skiprevs | ||||
curr = _annotatepair([hist[p] for p in pl], f, curr, skipchild, | ||||
Siddharth Agarwal
|
r32485 | diffopts) | ||
Matt Mackall
|
r13552 | for p in pl: | ||
if needed[p] == 1: | ||||
del hist[p] | ||||
FUJIWARA Katsunori
|
r19061 | del needed[p] | ||
Matt Mackall
|
r13552 | else: | ||
needed[p] -= 1 | ||||
Matt Mackall
|
r6762 | |||
Matt Mackall
|
r13552 | hist[f] = curr | ||
Jun Wu
|
r29861 | del pcache[f] | ||
Brendan Cully
|
r3172 | |||
Matt Mackall
|
r13552 | return zip(hist[base][0], hist[base][1].splitlines(True)) | ||
Matt Mackall
|
r3124 | |||
Sean Farley
|
r19610 | def ancestors(self, followfirst=False): | ||
visit = {} | ||||
c = self | ||||
Jordi Gutiérrez Hermoso
|
r24306 | if followfirst: | ||
cut = 1 | ||||
else: | ||||
cut = None | ||||
Sean Farley
|
r19610 | while True: | ||
for parent in c.parents()[:cut]: | ||||
Matt Mackall
|
r23981 | visit[(parent.linkrev(), parent.filenode())] = parent | ||
Sean Farley
|
r19610 | if not visit: | ||
break | ||||
c = visit.pop(max(visit)) | ||||
yield c | ||||
Phil Cohen
|
r33902 | def decodeddata(self): | ||
"""Returns `data()` after running repository decoding filters. | ||||
This is often equivalent to how the data would be expressed on disk. | ||||
""" | ||||
return self._repo.wwritedata(self.path(), self.data()) | ||||
Siddharth Agarwal
|
r34433 | @attr.s(slots=True, frozen=True) | ||
class annotateline(object): | ||||
fctx = attr.ib() | ||||
lineno = attr.ib(default=False) | ||||
Siddharth Agarwal
|
r34434 | # Whether this annotation was the result of a skip-annotate. | ||
skip = attr.ib(default=False) | ||||
Siddharth Agarwal
|
r34433 | |||
Siddharth Agarwal
|
r32485 | def _annotatepair(parents, childfctx, child, skipchild, diffopts): | ||
r''' | ||||
Given parent and child fctxes and annotate data for parents, for all lines | ||||
in either parent that match the child, annotate the child with the parent's | ||||
data. | ||||
Additionally, if `skipchild` is True, replace all other lines with parent | ||||
annotate data as well such that child is never blamed for any lines. | ||||
Siddharth Agarwal
|
r34431 | See test-annotate.py for unit tests. | ||
Siddharth Agarwal
|
r32485 | ''' | ||
Siddharth Agarwal
|
r32484 | pblocks = [(parent, mdiff.allblocks(parent[1], child[1], opts=diffopts)) | ||
for parent in parents] | ||||
Siddharth Agarwal
|
r32485 | |||
if skipchild: | ||||
# Need to iterate over the blocks twice -- make it a list | ||||
pblocks = [(p, list(blocks)) for (p, blocks) in pblocks] | ||||
Siddharth Agarwal
|
r32484 | # Mercurial currently prefers p2 over p1 for annotate. | ||
# TODO: change this? | ||||
for parent, blocks in pblocks: | ||||
for (a1, a2, b1, b2), t in blocks: | ||||
# Changed blocks ('!') or blocks made only of blank lines ('~') | ||||
# belong to the child. | ||||
if t == '=': | ||||
child[0][b1:b2] = parent[0][a1:a2] | ||||
Siddharth Agarwal
|
r32485 | |||
if skipchild: | ||||
# Now try and match up anything that couldn't be matched, | ||||
# Reversing pblocks maintains bias towards p2, matching above | ||||
# behavior. | ||||
pblocks.reverse() | ||||
# The heuristics are: | ||||
# * Work on blocks of changed lines (effectively diff hunks with -U0). | ||||
# This could potentially be smarter but works well enough. | ||||
# * For a non-matching section, do a best-effort fit. Match lines in | ||||
# diff hunks 1:1, dropping lines as necessary. | ||||
# * Repeat the last line as a last resort. | ||||
# First, replace as much as possible without repeating the last line. | ||||
remaining = [(parent, []) for parent, _blocks in pblocks] | ||||
for idx, (parent, blocks) in enumerate(pblocks): | ||||
for (a1, a2, b1, b2), _t in blocks: | ||||
if a2 - a1 >= b2 - b1: | ||||
for bk in xrange(b1, b2): | ||||
Siddharth Agarwal
|
r34433 | if child[0][bk].fctx == childfctx: | ||
Siddharth Agarwal
|
r32485 | ak = min(a1 + (bk - b1), a2 - 1) | ||
Siddharth Agarwal
|
r34434 | child[0][bk] = attr.evolve(parent[0][ak], skip=True) | ||
Siddharth Agarwal
|
r32485 | else: | ||
remaining[idx][1].append((a1, a2, b1, b2)) | ||||
# Then, look at anything left, which might involve repeating the last | ||||
# line. | ||||
for parent, blocks in remaining: | ||||
for a1, a2, b1, b2 in blocks: | ||||
for bk in xrange(b1, b2): | ||||
Siddharth Agarwal
|
r34433 | if child[0][bk].fctx == childfctx: | ||
Siddharth Agarwal
|
r32485 | ak = min(a1 + (bk - b1), a2 - 1) | ||
Siddharth Agarwal
|
r34434 | child[0][bk] = attr.evolve(parent[0][ak], skip=True) | ||
Siddharth Agarwal
|
r32483 | return child | ||
Sean Farley
|
r19608 | class filectx(basefilectx): | ||
"""A filecontext object makes access to data related to a particular | ||||
filerevision convenient.""" | ||||
def __init__(self, repo, path, changeid=None, fileid=None, | ||||
filelog=None, changectx=None): | ||||
"""changeid can be a changeset revision, node, or tag. | ||||
fileid can be a file revision or node.""" | ||||
self._repo = repo | ||||
self._path = path | ||||
assert (changeid is not None | ||||
or fileid is not None | ||||
or changectx is not None), \ | ||||
("bad args: changeid=%r, fileid=%r, changectx=%r" | ||||
% (changeid, fileid, changectx)) | ||||
if filelog is not None: | ||||
self._filelog = filelog | ||||
if changeid is not None: | ||||
self._changeid = changeid | ||||
if changectx is not None: | ||||
self._changectx = changectx | ||||
if fileid is not None: | ||||
self._fileid = fileid | ||||
@propertycache | ||||
def _changectx(self): | ||||
try: | ||||
return changectx(self._repo, self._changeid) | ||||
Pierre-Yves David
|
r23687 | except error.FilteredRepoLookupError: | ||
Sean Farley
|
r19608 | # Linkrev may point to any revision in the repository. When the | ||
# repository is filtered this may lead to `filectx` trying to build | ||||
# `changectx` for filtered revision. In such case we fallback to | ||||
# creating `changectx` on the unfiltered version of the reposition. | ||||
# This fallback should not be an issue because `changectx` from | ||||
# `filectx` are not used in complex operations that care about | ||||
# filtering. | ||||
# | ||||
# This fallback is a cheap and dirty fix that prevent several | ||||
# crashes. It does not ensure the behavior is correct. However the | ||||
# behavior was not correct before filtering either and "incorrect | ||||
# behavior" is seen as better as "crash" | ||||
# | ||||
# Linkrevs have several serious troubles with filtering that are | ||||
# complicated to solve. Proper handling of the issue here should be | ||||
# considered when solving linkrev issue are on the table. | ||||
return changectx(self._repo.unfiltered(), self._changeid) | ||||
Durham Goode
|
r23770 | def filectx(self, fileid, changeid=None): | ||
Sean Farley
|
r19608 | '''opens an arbitrary revision of the file without | ||
opening a new filelog''' | ||||
return filectx(self._repo, self._path, fileid=fileid, | ||||
Durham Goode
|
r23770 | filelog=self._filelog, changeid=changeid) | ||
Sean Farley
|
r19608 | |||
Remi Chaintron
|
r30743 | def rawdata(self): | ||
return self._filelog.revision(self._filenode, raw=True) | ||||
Jun Wu
|
r32237 | def rawflags(self): | ||
"""low-level revlog flags""" | ||||
return self._filelog.flags(self._filerev) | ||||
Sean Farley
|
r19608 | def data(self): | ||
Mike Edgar
|
r22932 | try: | ||
return self._filelog.read(self._filenode) | ||||
except error.CensoredNodeError: | ||||
Jun Wu
|
r33499 | if self._repo.ui.config("censor", "policy") == "ignore": | ||
Mike Edgar
|
r22932 | return "" | ||
Pierre-Yves David
|
r26587 | raise error.Abort(_("censored node: %s") % short(self._filenode), | ||
FUJIWARA Katsunori
|
r23110 | hint=_("set censor.policy to ignore errors")) | ||
Mike Edgar
|
r22932 | |||
Sean Farley
|
r19608 | def size(self): | ||
return self._filelog.size(self._filerev) | ||||
Jun Wu
|
r32235 | @propertycache | ||
def _copied(self): | ||||
Sean Farley
|
r19608 | """check if file was actually renamed in this changeset revision | ||
If rename logged in file revision, we report copy for changeset only | ||||
if file revisions linkrev points back to the changeset in question | ||||
or both changeset parents contain different file revisions. | ||||
""" | ||||
renamed = self._filelog.renamed(self._filenode) | ||||
if not renamed: | ||||
return renamed | ||||
if self.rev() == self.linkrev(): | ||||
return renamed | ||||
name = self.path() | ||||
fnode = self._filenode | ||||
for p in self._changectx.parents(): | ||||
try: | ||||
if fnode == p.filenode(name): | ||||
return None | ||||
except error.LookupError: | ||||
pass | ||||
return renamed | ||||
def children(self): | ||||
# hard for renames | ||||
c = self._filelog.children(self._filenode) | ||||
return [filectx(self._repo, self._path, fileid=x, | ||||
filelog=self._filelog) for x in c] | ||||
Sean Farley
|
r19733 | class committablectx(basectx): | ||
"""A committablectx object provides common functionality for a context that | ||||
Sean Farley
|
r19664 | wants the ability to commit, e.g. workingctx or memctx.""" | ||
def __init__(self, repo, text="", user=None, date=None, extra=None, | ||||
changes=None): | ||||
Matt Mackall
|
r3217 | self._repo = repo | ||
self._rev = None | ||||
self._node = None | ||||
Patrick Mezard
|
r6709 | self._text = text | ||
Christian Ebert
|
r6718 | if date: | ||
Patrick Mezard
|
r6709 | self._date = util.parsedate(date) | ||
Matt Mackall
|
r6817 | if user: | ||
self._user = user | ||||
Patrick Mezard
|
r6707 | if changes: | ||
Sean Farley
|
r21592 | self._status = changes | ||
Matt Mackall
|
r3217 | |||
Patrick Mezard
|
r6708 | self._extra = {} | ||
if extra: | ||||
self._extra = extra.copy() | ||||
if 'branch' not in self._extra: | ||||
try: | ||||
Matt Mackall
|
r13047 | branch = encoding.fromlocal(self._repo.dirstate.branch()) | ||
Patrick Mezard
|
r6708 | except UnicodeDecodeError: | ||
Pierre-Yves David
|
r26587 | raise error.Abort(_('branch name not in UTF-8!')) | ||
Patrick Mezard
|
r6708 | self._extra['branch'] = branch | ||
if self._extra['branch'] == '': | ||||
self._extra['branch'] = 'default' | ||||
Pulkit Goyal
|
r32643 | def __bytes__(self): | ||
return bytes(self._parents[0]) + "+" | ||||
Sean Farley
|
r19666 | |||
Yuya Nishihara
|
r33022 | __str__ = encoding.strmethod(__bytes__) | ||
Sean Farley
|
r19667 | def __nonzero__(self): | ||
return True | ||||
Gregory Szorc
|
r31476 | __bool__ = __nonzero__ | ||
Matt Mackall
|
r15337 | def _buildflagfunc(self): | ||
# Create a fallback function for getting file flags when the | ||||
# filesystem doesn't support them | ||||
copiesget = self._repo.dirstate.copies().get | ||||
Gregory Szorc
|
r27064 | parents = self.parents() | ||
if len(parents) < 2: | ||||
Matt Mackall
|
r15337 | # when we have one parent, it's easy: copy from parent | ||
Gregory Szorc
|
r27064 | man = parents[0].manifest() | ||
Matt Mackall
|
r15337 | def func(f): | ||
f = copiesget(f, f) | ||||
return man.flags(f) | ||||
else: | ||||
# merges are tricky: we try to reconstruct the unstored | ||||
# result from the merge (issue1802) | ||||
Gregory Szorc
|
r27064 | p1, p2 = parents | ||
Matt Mackall
|
r15337 | pa = p1.ancestor(p2) | ||
m1, m2, ma = p1.manifest(), p2.manifest(), pa.manifest() | ||||
def func(f): | ||||
f = copiesget(f, f) # may be wrong for merges with copies | ||||
fl1, fl2, fla = m1.flags(f), m2.flags(f), ma.flags(f) | ||||
if fl1 == fl2: | ||||
return fl1 | ||||
if fl1 == fla: | ||||
return fl2 | ||||
if fl2 == fla: | ||||
return fl1 | ||||
return '' # punt for conflicts | ||||
return func | ||||
Sean Farley
|
r19670 | @propertycache | ||
def _flagfunc(self): | ||||
return self._repo.dirstate.flagfunc(self._buildflagfunc) | ||||
Matt Mackall
|
r15337 | @propertycache | ||
Sean Farley
|
r19672 | def _status(self): | ||
Sean Farley
|
r21592 | return self._repo.status() | ||
Sean Farley
|
r19672 | |||
Sean Farley
|
r19674 | @propertycache | ||
def _user(self): | ||||
return self._repo.ui.username() | ||||
Sean Farley
|
r19676 | @propertycache | ||
def _date(self): | ||||
Boris Feld
|
r32409 | ui = self._repo.ui | ||
date = ui.configdate('devel', 'default-date') | ||||
if date is None: | ||||
date = util.makedate() | ||||
return date | ||||
Sean Farley
|
r19676 | |||
Sean Farley
|
r21587 | def subrev(self, subpath): | ||
return None | ||||
Yuya Nishihara
|
r24719 | def manifestnode(self): | ||
return None | ||||
Sean Farley
|
r19675 | def user(self): | ||
return self._user or self._repo.ui.username() | ||||
Sean Farley
|
r19677 | def date(self): | ||
return self._date | ||||
Sean Farley
|
r19678 | def description(self): | ||
return self._text | ||||
Sean Farley
|
r19679 | def files(self): | ||
Martin von Zweigbergk
|
r22916 | return sorted(self._status.modified + self._status.added + | ||
self._status.removed) | ||||
Sean Farley
|
r19675 | |||
Sean Farley
|
r19680 | def modified(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.modified | ||
Sean Farley
|
r19681 | def added(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.added | ||
Sean Farley
|
r19682 | def removed(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.removed | ||
Sean Farley
|
r19683 | def deleted(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.deleted | ||
Sean Farley
|
r19687 | def branch(self): | ||
return encoding.tolocal(self._extra['branch']) | ||||
Sean Farley
|
r19688 | def closesbranch(self): | ||
return 'close' in self._extra | ||||
Sean Farley
|
r19689 | def extra(self): | ||
return self._extra | ||||
Sean Farley
|
r19680 | |||
Phil Cohen
|
r34682 | def isinmemory(self): | ||
return False | ||||
Sean Farley
|
r19690 | def tags(self): | ||
Matt Harbison
|
r25688 | return [] | ||
Sean Farley
|
r19690 | |||
Sean Farley
|
r19691 | def bookmarks(self): | ||
b = [] | ||||
for p in self.parents(): | ||||
b.extend(p.bookmarks()) | ||||
return b | ||||
Sean Farley
|
r19692 | def phase(self): | ||
phase = phases.draft # default phase to draft | ||||
for p in self.parents(): | ||||
phase = max(phase, p.phase()) | ||||
return phase | ||||
Sean Farley
|
r19693 | def hidden(self): | ||
return False | ||||
Sean Farley
|
r19694 | def children(self): | ||
return [] | ||||
Sean Farley
|
r19695 | def flags(self, path): | ||
Pulkit Goyal
|
r32148 | if r'_manifest' in self.__dict__: | ||
Sean Farley
|
r19695 | try: | ||
return self._manifest.flags(path) | ||||
except KeyError: | ||||
return '' | ||||
try: | ||||
return self._flagfunc(path) | ||||
except OSError: | ||||
return '' | ||||
Sean Farley
|
r19696 | def ancestor(self, c2): | ||
Mads Kiilerich
|
r22389 | """return the "best" ancestor context of self and c2""" | ||
Sean Farley
|
r19696 | return self._parents[0].ancestor(c2) # punt on two parents for now | ||
Sean Farley
|
r19697 | def walk(self, match): | ||
Drew Gottlieb
|
r24646 | '''Generates matching file names.''' | ||
Martin von Zweigbergk
|
r34344 | return sorted(self._repo.dirstate.walk(match, | ||
subrepos=sorted(self.substate), | ||||
unknown=True, ignored=False)) | ||||
Sean Farley
|
r19697 | |||
Siddharth Agarwal
|
r21985 | def matches(self, match): | ||
return sorted(self._repo.dirstate.matches(match)) | ||||
Sean Farley
|
r19698 | def ancestors(self): | ||
Durham Goode
|
r23616 | for p in self._parents: | ||
yield p | ||||
Sean Farley
|
r19698 | for a in self._repo.changelog.ancestors( | ||
[p.rev() for p in self._parents]): | ||||
yield changectx(self._repo, a) | ||||
Sean Farley
|
r19699 | def markcommitted(self, node): | ||
"""Perform post-commit cleanup necessary after committing this ctx | ||||
Specifically, this updates backing stores this working context | ||||
wraps to reflect the fact that the changes reflected by this | ||||
workingctx have been committed. For example, it marks | ||||
modified and added files as normal in the dirstate. | ||||
""" | ||||
Augie Fackler
|
r32349 | with self._repo.dirstate.parentchange(): | ||
for f in self.modified() + self.added(): | ||||
self._repo.dirstate.normal(f) | ||||
for f in self.removed(): | ||||
self._repo.dirstate.drop(f) | ||||
self._repo.dirstate.setparents(node) | ||||
Sean Farley
|
r19699 | |||
FUJIWARA Katsunori
|
r25757 | # write changes out explicitly, because nesting wlock at | ||
# runtime may prevent 'wlock.release()' in 'repo.commit()' | ||||
# from immediately doing so for subsequent changing files | ||||
FUJIWARA Katsunori
|
r26748 | self._repo.dirstate.write(self._repo.currenttransaction()) | ||
FUJIWARA Katsunori
|
r25757 | |||
Sean Farley
|
r32610 | def dirty(self, missing=False, merge=True, branch=True): | ||
return False | ||||
Sean Farley
|
r19733 | class workingctx(committablectx): | ||
Sean Farley
|
r19671 | """A workingctx object makes access to data related to | ||
the current working directory convenient. | ||||
date - any valid date string or (unixtime, offset), or None. | ||||
user - username string, or None. | ||||
extra - a dictionary of extra values, or None. | ||||
changes - a list of file lists as returned by localrepo.status() | ||||
or None to use the repository status. | ||||
""" | ||||
def __init__(self, repo, text="", user=None, date=None, extra=None, | ||||
changes=None): | ||||
super(workingctx, self).__init__(repo, text, user, date, extra, changes) | ||||
Matt Mackall
|
r14129 | def __iter__(self): | ||
d = self._repo.dirstate | ||||
for f in d: | ||||
if d[f] != 'r': | ||||
yield f | ||||
Sean Farley
|
r21845 | def __contains__(self, key): | ||
return self._repo.dirstate[key] not in "?r" | ||||
Matt Harbison
|
r25590 | def hex(self): | ||
Yuya Nishihara
|
r25738 | return hex(wdirid) | ||
Matt Harbison
|
r25590 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _parents(self): | ||
p = self._repo.dirstate.parents() | ||||
if p[1] == nullid: | ||||
p = p[:-1] | ||||
Patrick Mezard
|
r17330 | return [changectx(self._repo, x) for x in p] | ||
Matt Mackall
|
r3217 | |||
Benoit Boissinot
|
r3966 | def filectx(self, path, filelog=None): | ||
Matt Mackall
|
r3217 | """get a file context from the working directory""" | ||
Benoit Boissinot
|
r3966 | return workingfilectx(self._repo, path, workingctx=self, | ||
filelog=filelog) | ||||
Matt Mackall
|
r3217 | |||
Patrick Mezard
|
r16491 | def dirty(self, missing=False, merge=True, branch=True): | ||
Matt Mackall
|
r8717 | "check whether a working directory is modified" | ||
Edouard Gomez
|
r11110 | # check subrepos first | ||
Mads Kiilerich
|
r18364 | for s in sorted(self.substate): | ||
Matt Harbison
|
r33364 | if self.sub(s).dirty(missing=missing): | ||
Edouard Gomez
|
r11110 | return True | ||
# check current working dir | ||||
Patrick Mezard
|
r16491 | return ((merge and self.p2()) or | ||
(branch and self.branch() != self.p1().branch()) or | ||||
Matt Mackall
|
r8717 | self.modified() or self.added() or self.removed() or | ||
(missing and self.deleted())) | ||||
Martin Geisler
|
r12270 | def add(self, list, prefix=""): | ||
Bryan O'Sullivan
|
r27809 | with self._repo.wlock(): | ||
ui, ds = self._repo.ui, self._repo.dirstate | ||||
Matt Harbison
|
r33501 | uipath = lambda f: ds.pathto(pathutil.join(prefix, f)) | ||
Dirkjan Ochtman
|
r11303 | rejected = [] | ||
FUJIWARA Katsunori
|
r19900 | lstat = self._repo.wvfs.lstat | ||
Dirkjan Ochtman
|
r11303 | for f in list: | ||
Matt Harbison
|
r33501 | # ds.pathto() returns an absolute file when this is invoked from | ||
# the keyword extension. That gets flagged as non-portable on | ||||
# Windows, since it contains the drive letter and colon. | ||||
scmutil.checkportable(ui, os.path.join(prefix, f)) | ||||
Dirkjan Ochtman
|
r11303 | try: | ||
FUJIWARA Katsunori
|
r19900 | st = lstat(f) | ||
Idan Kamara
|
r14004 | except OSError: | ||
Matt Harbison
|
r33501 | ui.warn(_("%s does not exist!\n") % uipath(f)) | ||
Dirkjan Ochtman
|
r11303 | rejected.append(f) | ||
continue | ||||
if st.st_size > 10000000: | ||||
ui.warn(_("%s: up to %d MB of RAM may be required " | ||||
"to manage this file\n" | ||||
"(use 'hg revert %s' to cancel the " | ||||
"pending addition)\n") | ||||
Matt Harbison
|
r33501 | % (f, 3 * st.st_size // 1000000, uipath(f))) | ||
Dirkjan Ochtman
|
r11303 | if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): | ||
ui.warn(_("%s not added: only files and symlinks " | ||||
Matt Harbison
|
r33501 | "supported currently\n") % uipath(f)) | ||
FUJIWARA Katsunori
|
r19900 | rejected.append(f) | ||
Dirkjan Ochtman
|
r11303 | elif ds[f] in 'amn': | ||
Matt Harbison
|
r33501 | ui.warn(_("%s already tracked!\n") % uipath(f)) | ||
Dirkjan Ochtman
|
r11303 | elif ds[f] == 'r': | ||
ds.normallookup(f) | ||||
else: | ||||
ds.add(f) | ||||
return rejected | ||||
David M. Carr
|
r15912 | def forget(self, files, prefix=""): | ||
Bryan O'Sullivan
|
r27810 | with self._repo.wlock(): | ||
Matt Harbison
|
r33501 | ds = self._repo.dirstate | ||
uipath = lambda f: ds.pathto(pathutil.join(prefix, f)) | ||||
David M. Carr
|
r15912 | rejected = [] | ||
Matt Mackall
|
r14435 | for f in files: | ||
Patrick Mezard
|
r16111 | if f not in self._repo.dirstate: | ||
Matt Harbison
|
r33501 | self._repo.ui.warn(_("%s not tracked!\n") % uipath(f)) | ||
David M. Carr
|
r15912 | rejected.append(f) | ||
Patrick Mezard
|
r16111 | elif self._repo.dirstate[f] != 'a': | ||
self._repo.dirstate.remove(f) | ||||
Dirkjan Ochtman
|
r11303 | else: | ||
Matt Mackall
|
r14434 | self._repo.dirstate.drop(f) | ||
David M. Carr
|
r15912 | return rejected | ||
Dirkjan Ochtman
|
r11303 | |||
def undelete(self, list): | ||||
pctxs = self.parents() | ||||
Bryan O'Sullivan
|
r27811 | with self._repo.wlock(): | ||
Matt Harbison
|
r33501 | ds = self._repo.dirstate | ||
Dirkjan Ochtman
|
r11303 | for f in list: | ||
if self._repo.dirstate[f] != 'r': | ||||
Matt Harbison
|
r33501 | self._repo.ui.warn(_("%s not removed!\n") % ds.pathto(f)) | ||
Dirkjan Ochtman
|
r11303 | else: | ||
Patrick Mezard
|
r12360 | fctx = f in pctxs[0] and pctxs[0][f] or pctxs[1][f] | ||
Dirkjan Ochtman
|
r11303 | t = fctx.data() | ||
self._repo.wwrite(f, t, fctx.flags()) | ||||
self._repo.dirstate.normal(f) | ||||
def copy(self, source, dest): | ||||
FUJIWARA Katsunori
|
r19902 | try: | ||
st = self._repo.wvfs.lstat(dest) | ||||
Gregory Szorc
|
r25660 | except OSError as err: | ||
FUJIWARA Katsunori
|
r19902 | if err.errno != errno.ENOENT: | ||
raise | ||||
Matt Harbison
|
r33501 | self._repo.ui.warn(_("%s does not exist!\n") | ||
% self._repo.dirstate.pathto(dest)) | ||||
FUJIWARA Katsunori
|
r19902 | return | ||
if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): | ||||
Dirkjan Ochtman
|
r11303 | self._repo.ui.warn(_("copy failed: %s is not a file or a " | ||
Matt Harbison
|
r33501 | "symbolic link\n") | ||
% self._repo.dirstate.pathto(dest)) | ||||
Dirkjan Ochtman
|
r11303 | else: | ||
Bryan O'Sullivan
|
r27812 | with self._repo.wlock(): | ||
Pierre-Yves David
|
r23402 | if self._repo.dirstate[dest] in '?': | ||
Dirkjan Ochtman
|
r11303 | self._repo.dirstate.add(dest) | ||
Pierre-Yves David
|
r23402 | elif self._repo.dirstate[dest] in 'r': | ||
self._repo.dirstate.normallookup(dest) | ||||
Dirkjan Ochtman
|
r11303 | self._repo.dirstate.copy(source, dest) | ||
Gregory Szorc
|
r31388 | def match(self, pats=None, include=None, exclude=None, default='glob', | ||
Matt Harbison
|
r25465 | listsubrepos=False, badfn=None): | ||
Matt Harbison
|
r24790 | r = self._repo | ||
# Only a case insensitive filesystem needs magic to translate user input | ||||
# to actual case in the filesystem. | ||||
Martin von Zweigbergk
|
r32400 | icasefs = not util.fscasesensitive(r.root) | ||
return matchmod.match(r.root, r.getcwd(), pats, include, exclude, | ||||
default, auditor=r.auditor, ctx=self, | ||||
listsubrepos=listsubrepos, badfn=badfn, | ||||
icasefs=icasefs) | ||||
Matt Harbison
|
r24790 | |||
Phil Cohen
|
r34106 | def flushall(self): | ||
pass # For overlayworkingfilectx compatibility. | ||||
Sean Farley
|
r21393 | def _filtersuspectsymlink(self, files): | ||
if not files or self._repo.dirstate._checklink: | ||||
return files | ||||
# Symlink placeholders may get non-symlink-like contents | ||||
# via user error or dereferencing by NFS or Samba servers, | ||||
# so we filter out any placeholders that don't look like a | ||||
# symlink | ||||
sane = [] | ||||
for f in files: | ||||
if self.flags(f) == 'l': | ||||
d = self[f].data() | ||||
if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d): | ||||
self._repo.ui.debug('ignoring suspect symlink placeholder' | ||||
' "%s"\n' % f) | ||||
continue | ||||
sane.append(f) | ||||
return sane | ||||
Sean Farley
|
r21395 | def _checklookup(self, files): | ||
# check for any possibly clean files | ||||
if not files: | ||||
Siddharth Agarwal
|
r32651 | return [], [], [] | ||
Sean Farley
|
r21395 | |||
modified = [] | ||||
Siddharth Agarwal
|
r32651 | deleted = [] | ||
Sean Farley
|
r21395 | fixup = [] | ||
pctx = self._parents[0] | ||||
# do a full compare of any files that might have changed | ||||
for f in sorted(files): | ||||
Siddharth Agarwal
|
r32651 | try: | ||
# This will return True for a file that got replaced by a | ||||
# directory in the interim, but fixing that is pretty hard. | ||||
if (f not in pctx or self.flags(f) != pctx.flags(f) | ||||
or pctx[f].cmp(self[f])): | ||||
modified.append(f) | ||||
else: | ||||
fixup.append(f) | ||||
except (IOError, OSError): | ||||
# A file become inaccessible in between? Mark it as deleted, | ||||
# matching dirstate behavior (issue5584). | ||||
# The dirstate has more complex behavior around whether a | ||||
# missing file matches a directory, etc, but we don't need to | ||||
# bother with that: if f has made it to this point, we're sure | ||||
# it's in the dirstate. | ||||
deleted.append(f) | ||||
Sean Farley
|
r21395 | |||
Siddharth Agarwal
|
r32812 | return modified, deleted, fixup | ||
Siddharth Agarwal
|
r32813 | def _poststatusfixup(self, status, fixup): | ||
Siddharth Agarwal
|
r32812 | """update dirstate for files that are actually clean""" | ||
Siddharth Agarwal
|
r32814 | poststatus = self._repo.postdsstatus() | ||
if fixup or poststatus: | ||||
Sean Farley
|
r21395 | try: | ||
FUJIWARA Katsunori
|
r32752 | oldid = self._repo.dirstate.identity() | ||
Sean Farley
|
r21395 | # updating the dirstate is optional | ||
# so we don't wait on the lock | ||||
Siddharth Agarwal
|
r21990 | # wlock can invalidate the dirstate, so cache normal _after_ | ||
# taking the lock | ||||
Bryan O'Sullivan
|
r27813 | with self._repo.wlock(False): | ||
FUJIWARA Katsunori
|
r32752 | if self._repo.dirstate.identity() == oldid: | ||
Siddharth Agarwal
|
r32814 | if fixup: | ||
normal = self._repo.dirstate.normal | ||||
for f in fixup: | ||||
normal(f) | ||||
# write changes out explicitly, because nesting | ||||
# wlock at runtime may prevent 'wlock.release()' | ||||
# after this block from doing so for subsequent | ||||
# changing files | ||||
tr = self._repo.currenttransaction() | ||||
self._repo.dirstate.write(tr) | ||||
if poststatus: | ||||
for ps in poststatus: | ||||
ps(self, status) | ||||
FUJIWARA Katsunori
|
r32752 | else: | ||
# in this case, writing changes out breaks | ||||
# consistency, because .hg/dirstate was | ||||
# already changed simultaneously after last | ||||
# caching (see also issue5584 for detail) | ||||
self._repo.ui.debug('skip updating dirstate: ' | ||||
'identity mismatch\n') | ||||
Sean Farley
|
r21395 | except error.LockError: | ||
pass | ||||
Siddharth Agarwal
|
r32814 | finally: | ||
# Even if the wlock couldn't be grabbed, clear out the list. | ||||
self._repo.clearpostdsstatus() | ||||
Sean Farley
|
r21395 | |||
Martin von Zweigbergk
|
r33938 | def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False): | ||
Sean Farley
|
r21397 | '''Gets the status from the dirstate -- internal use only.''' | ||
subrepos = [] | ||||
if '.hgsub' in self: | ||||
subrepos = sorted(self.substate) | ||||
Martin von Zweigbergk
|
r34345 | cmp, s = self._repo.dirstate.status(match, subrepos, ignored=ignored, | ||
clean=clean, unknown=unknown) | ||||
Sean Farley
|
r21397 | |||
# check for any possibly clean files | ||||
Siddharth Agarwal
|
r32812 | fixup = [] | ||
Sean Farley
|
r21397 | if cmp: | ||
Siddharth Agarwal
|
r32651 | modified2, deleted2, fixup = self._checklookup(cmp) | ||
Martin von Zweigbergk
|
r23303 | s.modified.extend(modified2) | ||
Siddharth Agarwal
|
r32651 | s.deleted.extend(deleted2) | ||
Sean Farley
|
r21397 | |||
Martin von Zweigbergk
|
r34345 | if fixup and clean: | ||
Martin von Zweigbergk
|
r23303 | s.clean.extend(fixup) | ||
Sean Farley
|
r21397 | |||
Siddharth Agarwal
|
r32813 | self._poststatusfixup(s, fixup) | ||
Siddharth Agarwal
|
r32812 | |||
Martin von Zweigbergk
|
r23776 | if match.always(): | ||
# cache for performance | ||||
if s.unknown or s.ignored or s.clean: | ||||
# "_status" is cached with list*=False in the normal route | ||||
self._status = scmutil.status(s.modified, s.added, s.removed, | ||||
s.deleted, [], [], []) | ||||
else: | ||||
self._status = s | ||||
Martin von Zweigbergk
|
r23303 | return s | ||
Sean Farley
|
r21397 | |||
Durham Goode
|
r31259 | @propertycache | ||
def _manifest(self): | ||||
"""generate a manifest corresponding to the values in self._status | ||||
This reuse the file nodeid from parent, but we use special node | ||||
identifiers for added and modified files. This is used by manifests | ||||
merge to see that files are different and by update logic to avoid | ||||
deleting newly added files. | ||||
""" | ||||
return self._buildstatusmanifest(self._status) | ||||
def _buildstatusmanifest(self, status): | ||||
"""Builds a manifest that includes the given status results.""" | ||||
parents = self.parents() | ||||
man = parents[0].manifest().copy() | ||||
ff = self._flagfunc | ||||
for i, l in ((addednodeid, status.added), | ||||
(modifiednodeid, status.modified)): | ||||
for f in l: | ||||
man[f] = i | ||||
try: | ||||
man.setflag(f, ff(f)) | ||||
except OSError: | ||||
pass | ||||
for f in status.deleted + status.removed: | ||||
if f in man: | ||||
del man[f] | ||||
return man | ||||
Sean Farley
|
r21480 | def _buildstatus(self, other, s, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown): | ||
Sean Farley
|
r21480 | """build a status with respect to another context | ||
This includes logic for maintaining the fast path of status when | ||||
comparing the working directory against its parent, which is to skip | ||||
building a new manifest if self (working directory) is not comparing | ||||
against its parent (repo['.']). | ||||
""" | ||||
Martin von Zweigbergk
|
r23239 | s = self._dirstatestatus(match, listignored, listclean, listunknown) | ||
Mads Kiilerich
|
r23543 | # Filter out symlinks that, in the case of FAT32 and NTFS filesystems, | ||
Martin von Zweigbergk
|
r23242 | # might have accidentally ended up with the entire contents of the file | ||
Mads Kiilerich
|
r23543 | # they are supposed to be linking to. | ||
Martin von Zweigbergk
|
r23302 | s.modified[:] = self._filtersuspectsymlink(s.modified) | ||
Sean Farley
|
r21480 | if other != self._repo['.']: | ||
s = super(workingctx, self)._buildstatus(other, s, match, | ||||
listignored, listclean, | ||||
listunknown) | ||||
return s | ||||
Martin von Zweigbergk
|
r23237 | def _matchstatus(self, other, match): | ||
Sean Farley
|
r21482 | """override the match method with a filter for directory patterns | ||
We use inheritance to customize the match.bad method only in cases of | ||||
workingctx since it belongs only to the working directory when | ||||
comparing against the parent changeset. | ||||
If we aren't comparing against the working directory's parent, then we | ||||
just use the default match object sent to us. | ||||
""" | ||||
if other != self._repo['.']: | ||||
def bad(f, msg): | ||||
# 'f' may be a directory pattern from 'match.files()', | ||||
# so 'f not in ctx1' is not enough | ||||
Drew Gottlieb
|
r24326 | if f not in other and not other.hasdir(f): | ||
Sean Farley
|
r21482 | self._repo.ui.warn('%s: %s\n' % | ||
(self._repo.dirstate.pathto(f), msg)) | ||||
match.bad = bad | ||||
return match | ||||
Gregory Szorc
|
r33353 | def markcommitted(self, node): | ||
super(workingctx, self).markcommitted(node) | ||||
sparse.aftercommit(self._repo, node) | ||||
Sean Farley
|
r19733 | class committablefilectx(basefilectx): | ||
"""A committablefilectx provides common functionality for a file context | ||||
that wants the ability to commit, e.g. workingfilectx or memfilectx.""" | ||||
Sean Farley
|
r19701 | def __init__(self, repo, path, filelog=None, ctx=None): | ||
Matt Mackall
|
r3217 | self._repo = repo | ||
self._path = path | ||||
self._changeid = None | ||||
self._filerev = self._filenode = None | ||||
Durham Goode
|
r19149 | if filelog is not None: | ||
Matt Mackall
|
r3217 | self._filelog = filelog | ||
Sean Farley
|
r19702 | if ctx: | ||
self._changectx = ctx | ||||
Sean Farley
|
r19703 | def __nonzero__(self): | ||
return True | ||||
Gregory Szorc
|
r31476 | __bool__ = __nonzero__ | ||
Yuya Nishihara
|
r24420 | def linkrev(self): | ||
# linked to self._changectx no matter if file is modified or not | ||||
return self.rev() | ||||
Matt Mackall
|
r3217 | def parents(self): | ||
'''return parent filectxs, following copies if necessary''' | ||||
Benoit Boissinot
|
r8528 | def filenode(ctx, path): | ||
return ctx._manifest.get(path, nullid) | ||||
path = self._path | ||||
Matt Mackall
|
r3217 | fl = self._filelog | ||
Benoit Boissinot
|
r8528 | pcl = self._changectx._parents | ||
renamed = self.renamed() | ||||
if renamed: | ||||
pl = [renamed + (None,)] | ||||
else: | ||||
pl = [(path, filenode(pcl[0], path), fl)] | ||||
for pc in pcl[1:]: | ||||
pl.append((path, filenode(pc, path), fl)) | ||||
Matt Mackall
|
r3217 | |||
Yuya Nishihara
|
r24817 | return [self._parentfilectx(p, fileid=n, filelog=l) | ||
Matt Mackall
|
r10282 | for p, n, l in pl if n != nullid] | ||
Matt Mackall
|
r3217 | |||
Sean Farley
|
r19705 | def children(self): | ||
return [] | ||||
Sean Farley
|
r19733 | class workingfilectx(committablefilectx): | ||
Sean Farley
|
r19704 | """A workingfilectx object makes access to data related to a particular | ||
file in the working directory convenient.""" | ||||
def __init__(self, repo, path, filelog=None, workingctx=None): | ||||
super(workingfilectx, self).__init__(repo, path, filelog, workingctx) | ||||
@propertycache | ||||
def _changectx(self): | ||||
return workingctx(self._repo) | ||||
def data(self): | ||||
return self._repo.wread(self._path) | ||||
def renamed(self): | ||||
rp = self._repo.dirstate.copied(self._path) | ||||
if not rp: | ||||
return None | ||||
return rp, self._changectx._parents[0]._manifest.get(rp, nullid) | ||||
Matt Mackall
|
r10282 | def size(self): | ||
FUJIWARA Katsunori
|
r19901 | return self._repo.wvfs.lstat(self._path).st_size | ||
Benoit Boissinot
|
r3962 | def date(self): | ||
t, tz = self._changectx.date() | ||||
try: | ||||
Matt Mackall
|
r27016 | return (self._repo.wvfs.lstat(self._path).st_mtime, tz) | ||
Gregory Szorc
|
r25660 | except OSError as err: | ||
Matt Mackall
|
r10282 | if err.errno != errno.ENOENT: | ||
raise | ||||
Benoit Boissinot
|
r3962 | return (t, tz) | ||
Matt Mackall
|
r3310 | |||
Phil Cohen
|
r33283 | def exists(self): | ||
return self._repo.wvfs.exists(self._path) | ||||
def lexists(self): | ||||
return self._repo.wvfs.lexists(self._path) | ||||
Phil Cohen
|
r33086 | def audit(self): | ||
return self._repo.wvfs.audit(self._path) | ||||
Nicolas Dumazet
|
r11702 | def cmp(self, fctx): | ||
"""compare with other file context | ||||
Nicolas Dumazet
|
r11539 | |||
Nicolas Dumazet
|
r11702 | returns True if different than fctx. | ||
Nicolas Dumazet
|
r11539 | """ | ||
Mads Kiilerich
|
r17425 | # fctx should be a filectx (not a workingfilectx) | ||
Nicolas Dumazet
|
r11703 | # invert comparison to reuse the same code path | ||
return fctx.cmp(self) | ||||
Patrick Mezard
|
r6715 | |||
Sean Farley
|
r22073 | def remove(self, ignoremissing=False): | ||
"""wraps unlink for a repo's working directory""" | ||||
Mads Kiilerich
|
r31309 | self._repo.wvfs.unlinkpath(self._path, ignoremissing=ignoremissing) | ||
Sean Farley
|
r22073 | |||
Phil Cohen
|
r33085 | def write(self, data, flags, backgroundclose=False): | ||
Sean Farley
|
r22073 | """wraps repo.wwrite""" | ||
Phil Cohen
|
r33085 | self._repo.wwrite(self._path, data, flags, | ||
backgroundclose=backgroundclose) | ||||
Sean Farley
|
r22073 | |||
Phil Cohen
|
r34788 | def markcopied(self, src): | ||
"""marks this file a copy of `src`""" | ||||
if self._repo.dirstate[self._path] in "nma": | ||||
self._repo.dirstate.copy(src, self._path) | ||||
Phil Cohen
|
r34039 | def clearunknown(self): | ||
"""Removes conflicting items in the working directory so that | ||||
``write()`` can be called successfully. | ||||
""" | ||||
wvfs = self._repo.wvfs | ||||
Mark Thomas
|
r34557 | f = self._path | ||
if wvfs.isdir(f) and not wvfs.islink(f): | ||||
wvfs.rmtree(f, forcibly=True) | ||||
for p in reversed(list(util.finddirs(f))): | ||||
if wvfs.isfileorlink(p): | ||||
wvfs.unlink(p) | ||||
break | ||||
Phil Cohen
|
r34039 | |||
Phil Cohen
|
r33084 | def setflags(self, l, x): | ||
self._repo.wvfs.setflags(self._path, l, x) | ||||
Phil Cohen
|
r34106 | class overlayworkingctx(workingctx): | ||
"""Wraps another mutable context with a write-back cache that can be flushed | ||||
at a later time. | ||||
self._cache[path] maps to a dict with keys: { | ||||
'exists': bool? | ||||
'date': date? | ||||
'data': str? | ||||
'flags': str? | ||||
} | ||||
If `exists` is True, `flags` must be non-None and 'date' is non-None. If it | ||||
is `False`, the file was deleted. | ||||
""" | ||||
def __init__(self, repo, wrappedctx): | ||||
super(overlayworkingctx, self).__init__(repo) | ||||
self._repo = repo | ||||
self._wrappedctx = wrappedctx | ||||
self._clean() | ||||
def data(self, path): | ||||
if self.isdirty(path): | ||||
if self._cache[path]['exists']: | ||||
if self._cache[path]['data']: | ||||
return self._cache[path]['data'] | ||||
else: | ||||
# Must fallback here, too, because we only set flags. | ||||
return self._wrappedctx[path].data() | ||||
else: | ||||
raise error.ProgrammingError("No such file or directory: %s" % | ||||
self._path) | ||||
else: | ||||
return self._wrappedctx[path].data() | ||||
Phil Cohen
|
r34682 | def isinmemory(self): | ||
return True | ||||
Phil Cohen
|
r34106 | def filedate(self, path): | ||
if self.isdirty(path): | ||||
return self._cache[path]['date'] | ||||
else: | ||||
return self._wrappedctx[path].date() | ||||
def flags(self, path): | ||||
if self.isdirty(path): | ||||
if self._cache[path]['exists']: | ||||
return self._cache[path]['flags'] | ||||
else: | ||||
raise error.ProgrammingError("No such file or directory: %s" % | ||||
self._path) | ||||
else: | ||||
return self._wrappedctx[path].flags() | ||||
def write(self, path, data, flags=''): | ||||
if data is None: | ||||
raise error.ProgrammingError("data must be non-None") | ||||
self._markdirty(path, exists=True, data=data, date=util.makedate(), | ||||
flags=flags) | ||||
def setflags(self, path, l, x): | ||||
self._markdirty(path, exists=True, date=util.makedate(), | ||||
flags=(l and 'l' or '') + (x and 'x' or '')) | ||||
def remove(self, path): | ||||
self._markdirty(path, exists=False) | ||||
def exists(self, path): | ||||
"""exists behaves like `lexists`, but needs to follow symlinks and | ||||
return False if they are broken. | ||||
""" | ||||
if self.isdirty(path): | ||||
# If this path exists and is a symlink, "follow" it by calling | ||||
# exists on the destination path. | ||||
if (self._cache[path]['exists'] and | ||||
'l' in self._cache[path]['flags']): | ||||
return self.exists(self._cache[path]['data'].strip()) | ||||
else: | ||||
return self._cache[path]['exists'] | ||||
return self._wrappedctx[path].exists() | ||||
def lexists(self, path): | ||||
"""lexists returns True if the path exists""" | ||||
if self.isdirty(path): | ||||
return self._cache[path]['exists'] | ||||
return self._wrappedctx[path].lexists() | ||||
def size(self, path): | ||||
if self.isdirty(path): | ||||
if self._cache[path]['exists']: | ||||
return len(self._cache[path]['data']) | ||||
else: | ||||
raise error.ProgrammingError("No such file or directory: %s" % | ||||
self._path) | ||||
return self._wrappedctx[path].size() | ||||
def flushall(self): | ||||
for path in self._writeorder: | ||||
entry = self._cache[path] | ||||
if entry['exists']: | ||||
self._wrappedctx[path].clearunknown() | ||||
if entry['data'] is not None: | ||||
if entry['flags'] is None: | ||||
raise error.ProgrammingError('data set but not flags') | ||||
self._wrappedctx[path].write( | ||||
entry['data'], | ||||
entry['flags']) | ||||
else: | ||||
self._wrappedctx[path].setflags( | ||||
'l' in entry['flags'], | ||||
'x' in entry['flags']) | ||||
else: | ||||
self._wrappedctx[path].remove(path) | ||||
self._clean() | ||||
def isdirty(self, path): | ||||
return path in self._cache | ||||
def _clean(self): | ||||
self._cache = {} | ||||
self._writeorder = [] | ||||
def _markdirty(self, path, exists, data=None, date=None, flags=''): | ||||
if path not in self._cache: | ||||
self._writeorder.append(path) | ||||
self._cache[path] = { | ||||
'exists': exists, | ||||
'data': data, | ||||
'date': date, | ||||
'flags': flags, | ||||
} | ||||
def filectx(self, path, filelog=None): | ||||
return overlayworkingfilectx(self._repo, path, parent=self, | ||||
filelog=filelog) | ||||
class overlayworkingfilectx(workingfilectx): | ||||
"""Wrap a ``workingfilectx`` but intercepts all writes into an in-memory | ||||
cache, which can be flushed through later by calling ``flush()``.""" | ||||
def __init__(self, repo, path, filelog=None, parent=None): | ||||
super(overlayworkingfilectx, self).__init__(repo, path, filelog, | ||||
parent) | ||||
self._repo = repo | ||||
self._parent = parent | ||||
self._path = path | ||||
Phil Cohen
|
r34784 | def cmp(self, fctx): | ||
return self.data() != fctx.data() | ||||
Phil Cohen
|
r34106 | def ctx(self): | ||
return self._parent | ||||
def data(self): | ||||
return self._parent.data(self._path) | ||||
def date(self): | ||||
return self._parent.filedate(self._path) | ||||
def exists(self): | ||||
return self.lexists() | ||||
def lexists(self): | ||||
return self._parent.exists(self._path) | ||||
def renamed(self): | ||||
# Copies are currently tracked in the dirstate as before. Straight copy | ||||
# from workingfilectx. | ||||
rp = self._repo.dirstate.copied(self._path) | ||||
if not rp: | ||||
return None | ||||
return rp, self._changectx._parents[0]._manifest.get(rp, nullid) | ||||
def size(self): | ||||
return self._parent.size(self._path) | ||||
def audit(self): | ||||
pass | ||||
def flags(self): | ||||
return self._parent.flags(self._path) | ||||
def setflags(self, islink, isexec): | ||||
return self._parent.setflags(self._path, islink, isexec) | ||||
def write(self, data, flags, backgroundclose=False): | ||||
return self._parent.write(self._path, data, flags) | ||||
def remove(self, ignoremissing=False): | ||||
return self._parent.remove(self._path) | ||||
FUJIWARA Katsunori
|
r23710 | class workingcommitctx(workingctx): | ||
"""A workingcommitctx object makes access to data related to | ||||
the revision being committed convenient. | ||||
This hides changes in the working directory, if they aren't | ||||
committed in this context. | ||||
""" | ||||
def __init__(self, repo, changes, | ||||
text="", user=None, date=None, extra=None): | ||||
super(workingctx, self).__init__(repo, text, user, date, extra, | ||||
changes) | ||||
Martin von Zweigbergk
|
r33938 | def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False): | ||
FUJIWARA Katsunori
|
r23712 | """Return matched files only in ``self._status`` | ||
Uncommitted files appear "clean" via this context, even if | ||||
they aren't actually so in the working directory. | ||||
""" | ||||
if clean: | ||||
clean = [f for f in self._manifest if f not in self._changedset] | ||||
else: | ||||
clean = [] | ||||
return scmutil.status([f for f in self._status.modified if match(f)], | ||||
[f for f in self._status.added if match(f)], | ||||
[f for f in self._status.removed if match(f)], | ||||
[], [], [], clean) | ||||
@propertycache | ||||
def _changedset(self): | ||||
"""Return the set of files changed in this context | ||||
""" | ||||
changed = set(self._status.modified) | ||||
changed.update(self._status.added) | ||||
changed.update(self._status.removed) | ||||
return changed | ||||
Gregory Szorc
|
r27906 | def makecachingfilectxfn(func): | ||
"""Create a filectxfn that caches based on the path. | ||||
We can't use util.cachefunc because it uses all arguments as the cache | ||||
key and this creates a cycle since the arguments include the repo and | ||||
memctx. | ||||
""" | ||||
cache = {} | ||||
def getfilectx(repo, memctx, path): | ||||
if path not in cache: | ||||
cache[path] = func(repo, memctx, path) | ||||
return cache[path] | ||||
return getfilectx | ||||
Sean Farley
|
r32763 | def memfilefromctx(ctx): | ||
"""Given a context return a memfilectx for ctx[path] | ||||
This is a convenience method for building a memctx based on another | ||||
context. | ||||
""" | ||||
def getfilectx(repo, memctx, path): | ||||
fctx = ctx[path] | ||||
# this is weird but apparently we only keep track of one parent | ||||
# (why not only store that instead of a tuple?) | ||||
copied = fctx.renamed() | ||||
if copied: | ||||
copied = copied[0] | ||||
return memfilectx(repo, path, fctx.data(), | ||||
islink=fctx.islink(), isexec=fctx.isexec(), | ||||
copied=copied, memctx=memctx) | ||||
return getfilectx | ||||
Sean Farley
|
r32764 | def memfilefrompatch(patchstore): | ||
"""Given a patch (e.g. patchstore object) return a memfilectx | ||||
This is a convenience method for building a memctx based on a patchstore. | ||||
""" | ||||
def getfilectx(repo, memctx, path): | ||||
data, mode, copied = patchstore.getfile(path) | ||||
if data is None: | ||||
return None | ||||
islink, isexec = mode | ||||
return memfilectx(repo, path, data, islink=islink, | ||||
isexec=isexec, copied=copied, | ||||
memctx=memctx) | ||||
return getfilectx | ||||
Sean Farley
|
r21665 | class memctx(committablectx): | ||
Patrick Mezard
|
r7077 | """Use memctx to perform in-memory commits via localrepo.commitctx(). | ||
Patrick Mezard
|
r6715 | |||
Patrick Mezard
|
r7077 | Revision information is supplied at initialization time while | ||
related files data and is made available through a callback | ||||
mechanism. 'repo' is the current localrepo, 'parents' is a | ||||
sequence of two parent revisions identifiers (pass None for every | ||||
missing parent), 'text' is the commit message and 'files' lists | ||||
names of files touched by the revision (normalized and relative to | ||||
repository root). | ||||
Patrick Mezard
|
r6715 | |||
Patrick Mezard
|
r7077 | filectxfn(repo, memctx, path) is a callable receiving the | ||
repository, the current memctx object and the normalized path of | ||||
requested file, relative to repository root. It is fired by the | ||||
commit function for every file in 'files', but calls order is | ||||
undefined. If the file is available in the revision being | ||||
committed (updated or added), filectxfn returns a memfilectx | ||||
FUJIWARA Katsunori
|
r31612 | object. If the file was removed, filectxfn return None for recent | ||
Mercurial. Moved files are represented by marking the source file | ||||
Patrick Mezard
|
r7077 | removed and the new file added with copy information (see | ||
memfilectx). | ||||
user receives the committer name and defaults to current | ||||
repository username, date is the commit date in any format | ||||
supported by util.parsedate() and defaults to current date, extra | ||||
is a dictionary of metadata or is left empty. | ||||
Patrick Mezard
|
r6715 | """ | ||
Siddharth Agarwal
|
r22313 | |||
# Mercurial <= 3.1 expects the filectxfn to raise IOError for missing files. | ||||
# Extensions that need to retain compatibility across Mercurial 3.1 can use | ||||
# this field to determine what to do in filectxfn. | ||||
_returnnoneformissingfiles = True | ||||
Dirkjan Ochtman
|
r6721 | def __init__(self, repo, parents, text, files, filectxfn, user=None, | ||
Sean Farley
|
r32765 | date=None, extra=None, branch=None, editor=False): | ||
Sean Farley
|
r21666 | super(memctx, self).__init__(repo, text, user, date, extra) | ||
Patrick Mezard
|
r6715 | self._rev = None | ||
self._node = None | ||||
parents = [(p or nullid) for p in parents] | ||||
p1, p2 = parents | ||||
Matt Mackall
|
r6747 | self._parents = [changectx(self._repo, p) for p in (p1, p2)] | ||
Matt Mackall
|
r8209 | files = sorted(set(files)) | ||
FUJIWARA Katsunori
|
r23587 | self._files = files | ||
Sean Farley
|
r32765 | if branch is not None: | ||
self._extra['branch'] = encoding.fromlocal(branch) | ||||
Sean Farley
|
r21938 | self.substate = {} | ||
Patrick Mezard
|
r6715 | |||
Sean Farley
|
r32765 | if isinstance(filectxfn, patch.filestore): | ||
Sean Farley
|
r32781 | filectxfn = memfilefrompatch(filectxfn) | ||
Sean Farley
|
r32765 | elif not callable(filectxfn): | ||
# if store is not callable, wrap it in a function | ||||
Sean Farley
|
r32781 | filectxfn = memfilefromctx(filectxfn) | ||
# memoizing increases performance for e.g. vcs convert scenarios. | ||||
self._filectxfn = makecachingfilectxfn(filectxfn) | ||||
Sean Farley
|
r22072 | |||
FUJIWARA Katsunori
|
r21238 | if editor: | ||
self._text = editor(self._repo, self, []) | ||||
self._repo.savecommitmessage(self._text) | ||||
Patrick Mezard
|
r6715 | def filectx(self, path, filelog=None): | ||
Mads Kiilerich
|
r22296 | """get a file context from the working directory | ||
Returns None if file doesn't exist and should be removed.""" | ||||
Patrick Mezard
|
r6715 | return self._filectxfn(self._repo, self, path) | ||
Alexander Solovyov
|
r11151 | def commit(self): | ||
"""commit context to the repo""" | ||||
return self._repo.commitctx(self) | ||||
Sean Farley
|
r21835 | @propertycache | ||
def _manifest(self): | ||||
"""generate a manifest based on the return values of filectxfn""" | ||||
# keep this simple for now; just worry about p1 | ||||
pctx = self._parents[0] | ||||
man = pctx.manifest().copy() | ||||
FUJIWARA Katsunori
|
r23603 | for f in self._status.modified: | ||
Sean Farley
|
r21835 | p1node = nullid | ||
p2node = nullid | ||||
Sean Farley
|
r22075 | p = pctx[f].parents() # if file isn't in pctx, check p2? | ||
Sean Farley
|
r21835 | if len(p) > 0: | ||
Durham Goode
|
r27983 | p1node = p[0].filenode() | ||
Sean Farley
|
r21835 | if len(p) > 1: | ||
Durham Goode
|
r27983 | p2node = p[1].filenode() | ||
FUJIWARA Katsunori
|
r23603 | man[f] = revlog.hash(self[f].data(), p1node, p2node) | ||
Sean Farley
|
r21835 | |||
FUJIWARA Katsunori
|
r23588 | for f in self._status.added: | ||
man[f] = revlog.hash(self[f].data(), nullid, nullid) | ||||
FUJIWARA Katsunori
|
r23589 | for f in self._status.removed: | ||
if f in man: | ||||
del man[f] | ||||
Sean Farley
|
r21835 | |||
return man | ||||
FUJIWARA Katsunori
|
r23587 | @propertycache | ||
def _status(self): | ||||
"""Calculate exact status from ``files`` specified at construction | ||||
""" | ||||
man1 = self.p1().manifest() | ||||
p2 = self._parents[1] | ||||
# "1 < len(self._parents)" can't be used for checking | ||||
# existence of the 2nd parent, because "memctx._parents" is | ||||
# explicitly initialized by the list, of which length is 2. | ||||
if p2.node() != nullid: | ||||
man2 = p2.manifest() | ||||
managing = lambda f: f in man1 or f in man2 | ||||
else: | ||||
managing = lambda f: f in man1 | ||||
modified, added, removed = [], [], [] | ||||
for f in self._files: | ||||
if not managing(f): | ||||
added.append(f) | ||||
elif self[f]: | ||||
modified.append(f) | ||||
else: | ||||
removed.append(f) | ||||
return scmutil.status(modified, added, removed, [], [], [], []) | ||||
Sean Farley
|
r21835 | |||
Sean Farley
|
r21688 | class memfilectx(committablefilectx): | ||
Patrick Mezard
|
r7077 | """memfilectx represents an in-memory file to commit. | ||
Mads Kiilerich
|
r23139 | See memctx and committablefilectx for more details. | ||
Patrick Mezard
|
r6715 | """ | ||
Sean Farley
|
r21689 | def __init__(self, repo, path, data, islink=False, | ||
isexec=False, copied=None, memctx=None): | ||||
Patrick Mezard
|
r7077 | """ | ||
path is the normalized file path relative to repository root. | ||||
data is the file content as a string. | ||||
islink is True if the file is a symbolic link. | ||||
isexec is True if the file is executable. | ||||
copied is the source file path if current file was copied in the | ||||
revision being committed, or None.""" | ||||
Sean Farley
|
r21689 | super(memfilectx, self).__init__(repo, path, None, memctx) | ||
Patrick Mezard
|
r6715 | self._data = data | ||
self._flags = (islink and 'l' or '') + (isexec and 'x' or '') | ||||
self._copied = None | ||||
if copied: | ||||
self._copied = (copied, nullid) | ||||
Matt Mackall
|
r10282 | def data(self): | ||
return self._data | ||||
Sean Farley
|
r22074 | |||
def remove(self, ignoremissing=False): | ||||
"""wraps unlink for a repo's working directory""" | ||||
# need to figure out what to do here | ||||
del self._changectx[self._path] | ||||
def write(self, data, flags): | ||||
"""wraps repo.wwrite""" | ||||
self._data = data | ||||
Mateusz Kwapich
|
r30567 | |||
Jun Wu
|
r32239 | class overlayfilectx(committablefilectx): | ||
"""Like memfilectx but take an original filectx and optional parameters to | ||||
override parts of it. This is useful when fctx.data() is expensive (i.e. | ||||
flag processor is expensive) and raw data, flags, and filenode could be | ||||
reused (ex. rebase or mode-only amend a REVIDX_EXTSTORED file). | ||||
""" | ||||
def __init__(self, originalfctx, datafunc=None, path=None, flags=None, | ||||
copied=None, ctx=None): | ||||
"""originalfctx: filecontext to duplicate | ||||
datafunc: None or a function to override data (file content). It is a | ||||
function to be lazy. path, flags, copied, ctx: None or overridden value | ||||
copied could be (path, rev), or False. copied could also be just path, | ||||
and will be converted to (path, nullid). This simplifies some callers. | ||||
""" | ||||
if path is None: | ||||
path = originalfctx.path() | ||||
if ctx is None: | ||||
ctx = originalfctx.changectx() | ||||
ctxmatch = lambda: True | ||||
else: | ||||
ctxmatch = lambda: ctx == originalfctx.changectx() | ||||
repo = originalfctx.repo() | ||||
flog = originalfctx.filelog() | ||||
super(overlayfilectx, self).__init__(repo, path, flog, ctx) | ||||
if copied is None: | ||||
copied = originalfctx.renamed() | ||||
copiedmatch = lambda: True | ||||
else: | ||||
if copied and not isinstance(copied, tuple): | ||||
# repo._filecommit will recalculate copyrev so nullid is okay | ||||
copied = (copied, nullid) | ||||
copiedmatch = lambda: copied == originalfctx.renamed() | ||||
# When data, copied (could affect data), ctx (could affect filelog | ||||
# parents) are not overridden, rawdata, rawflags, and filenode may be | ||||
# reused (repo._filecommit should double check filelog parents). | ||||
# | ||||
# path, flags are not hashed in filelog (but in manifestlog) so they do | ||||
# not affect reusable here. | ||||
# | ||||
# If ctx or copied is overridden to a same value with originalfctx, | ||||
# still consider it's reusable. originalfctx.renamed() may be a bit | ||||
# expensive so it's not called unless necessary. Assuming datafunc is | ||||
# always expensive, do not call it for this "reusable" test. | ||||
reusable = datafunc is None and ctxmatch() and copiedmatch() | ||||
if datafunc is None: | ||||
datafunc = originalfctx.data | ||||
if flags is None: | ||||
flags = originalfctx.flags() | ||||
self._datafunc = datafunc | ||||
self._flags = flags | ||||
self._copied = copied | ||||
if reusable: | ||||
# copy extra fields from originalfctx | ||||
attrs = ['rawdata', 'rawflags', '_filenode', '_filerev'] | ||||
Siddharth Agarwal
|
r34432 | for attr_ in attrs: | ||
if util.safehasattr(originalfctx, attr_): | ||||
setattr(self, attr_, getattr(originalfctx, attr_)) | ||||
Jun Wu
|
r32239 | |||
def data(self): | ||||
return self._datafunc() | ||||
Mateusz Kwapich
|
r30567 | class metadataonlyctx(committablectx): | ||
"""Like memctx but it's reusing the manifest of different commit. | ||||
Intended to be used by lightweight operations that are creating | ||||
metadata-only changes. | ||||
Revision information is supplied at initialization time. 'repo' is the | ||||
current localrepo, 'ctx' is original revision which manifest we're reuisng | ||||
'parents' is a sequence of two parent revisions identifiers (pass None for | ||||
every missing parent), 'text' is the commit. | ||||
user receives the committer name and defaults to current repository | ||||
username, date is the commit date in any format supported by | ||||
util.parsedate() and defaults to current date, extra is a dictionary of | ||||
metadata or is left empty. | ||||
""" | ||||
Jun Wu
|
r30609 | def __new__(cls, repo, originalctx, *args, **kwargs): | ||
Mateusz Kwapich
|
r30567 | return super(metadataonlyctx, cls).__new__(cls, repo) | ||
Jun Wu
|
r34000 | def __init__(self, repo, originalctx, parents=None, text=None, user=None, | ||
date=None, extra=None, editor=False): | ||||
if text is None: | ||||
text = originalctx.description() | ||||
Mateusz Kwapich
|
r30567 | super(metadataonlyctx, self).__init__(repo, text, user, date, extra) | ||
self._rev = None | ||||
self._node = None | ||||
self._originalctx = originalctx | ||||
self._manifestnode = originalctx.manifestnode() | ||||
Jun Wu
|
r34000 | if parents is None: | ||
parents = originalctx.parents() | ||||
else: | ||||
parents = [repo[p] for p in parents if p is not None] | ||||
parents = parents[:] | ||||
while len(parents) < 2: | ||||
parents.append(repo[nullid]) | ||||
p1, p2 = self._parents = parents | ||||
Mateusz Kwapich
|
r30567 | |||
# sanity check to ensure that the reused manifest parents are | ||||
# manifests of our commit parents | ||||
mp1, mp2 = self.manifestctx().parents | ||||
Jun Wu
|
r31840 | if p1 != nullid and p1.manifestnode() != mp1: | ||
Mateusz Kwapich
|
r30567 | raise RuntimeError('can\'t reuse the manifest: ' | ||
'its p1 doesn\'t match the new ctx p1') | ||||
Jun Wu
|
r31840 | if p2 != nullid and p2.manifestnode() != mp2: | ||
Mateusz Kwapich
|
r30567 | raise RuntimeError('can\'t reuse the manifest: ' | ||
'its p2 doesn\'t match the new ctx p2') | ||||
self._files = originalctx.files() | ||||
self.substate = {} | ||||
if editor: | ||||
self._text = editor(self._repo, self, []) | ||||
self._repo.savecommitmessage(self._text) | ||||
def manifestnode(self): | ||||
return self._manifestnode | ||||
Jun Wu
|
r32519 | @property | ||
Mateusz Kwapich
|
r30567 | def _manifestctx(self): | ||
return self._repo.manifestlog[self._manifestnode] | ||||
def filectx(self, path, filelog=None): | ||||
return self._originalctx.filectx(path, filelog=filelog) | ||||
def commit(self): | ||||
"""commit context to the repo""" | ||||
return self._repo.commitctx(self) | ||||
@property | ||||
def _manifest(self): | ||||
return self._originalctx.manifest() | ||||
@propertycache | ||||
def _status(self): | ||||
"""Calculate exact status from ``files`` specified in the ``origctx`` | ||||
and parents manifests. | ||||
""" | ||||
man1 = self.p1().manifest() | ||||
p2 = self._parents[1] | ||||
# "1 < len(self._parents)" can't be used for checking | ||||
# existence of the 2nd parent, because "metadataonlyctx._parents" is | ||||
# explicitly initialized by the list, of which length is 2. | ||||
if p2.node() != nullid: | ||||
man2 = p2.manifest() | ||||
managing = lambda f: f in man1 or f in man2 | ||||
else: | ||||
managing = lambda f: f in man1 | ||||
modified, added, removed = [], [], [] | ||||
for f in self._files: | ||||
if not managing(f): | ||||
added.append(f) | ||||
Jun Wu
|
r34001 | elif f in self: | ||
Mateusz Kwapich
|
r30567 | modified.append(f) | ||
else: | ||||
removed.append(f) | ||||
return scmutil.status(modified, added, removed, [], [], [], []) | ||||
Phil Cohen
|
r34053 | |||
class arbitraryfilectx(object): | ||||
"""Allows you to use filectx-like functions on a file in an arbitrary | ||||
location on disk, possibly not in the working directory. | ||||
""" | ||||
Phil Cohen
|
r34686 | def __init__(self, path, repo=None): | ||
# Repo is optional because contrib/simplemerge uses this class. | ||||
self._repo = repo | ||||
Phil Cohen
|
r34053 | self._path = path | ||
Phil Cohen
|
r34686 | def cmp(self, fctx): | ||
if isinstance(fctx, workingfilectx) and self._repo: | ||||
# Add a fast-path for merge if both sides are disk-backed. | ||||
# Note that filecmp uses the opposite return values as cmp. | ||||
return not filecmp.cmp(self.path(), self._repo.wjoin(fctx.path())) | ||||
return self.data() != fctx.data() | ||||
Phil Cohen
|
r34053 | |||
def path(self): | ||||
return self._path | ||||
def flags(self): | ||||
return '' | ||||
def data(self): | ||||
return util.readfile(self._path) | ||||
def decodeddata(self): | ||||
with open(self._path, "rb") as f: | ||||
return f.read() | ||||
def remove(self): | ||||
util.unlink(self._path) | ||||
def write(self, data, flags): | ||||
assert not flags | ||||
with open(self._path, "w") as f: | ||||
f.write(data) | ||||