context.py
1880 lines
| 66.5 KiB
| text/x-python
|
PythonLexer
/ mercurial / context.py
Matt Mackall
|
r2563 | # context.py - changeset and file context objects for mercurial | ||
# | ||||
Thomas Arendsen Hein
|
r4635 | # Copyright 2006, 2007 Matt Mackall <mpm@selenic.com> | ||
Matt Mackall
|
r2563 | # | ||
Martin Geisler
|
r8225 | # This software may be used and distributed according to the terms of the | ||
Matt Mackall
|
r10263 | # GNU General Public License version 2 or any later version. | ||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r16376 | from node import nullid, nullrev, short, hex, bin | ||
Matt Mackall
|
r3891 | from i18n import _ | ||
Mads Kiilerich
|
r20984 | import mdiff, error, util, scmutil, subrepo, patch, encoding, phases | ||
Matt Mackall
|
r14669 | import match as matchmod | ||
Dirkjan Ochtman
|
r11303 | import os, errno, stat | ||
Pierre-Yves David
|
r17469 | import obsolete as obsmod | ||
Pierre-Yves David
|
r18252 | import repoview | ||
Augie Fackler
|
r20400 | import fileset | ||
Sean Farley
|
r21835 | import revlog | ||
Matt Mackall
|
r3122 | |||
Matt Mackall
|
r8207 | propertycache = util.propertycache | ||
Dirkjan Ochtman
|
r7368 | |||
Augie Fackler
|
r23593 | # Phony node value to stand-in for new files in some uses of | ||
# manifests. Manifests support 21-byte hashes for nodes which are | ||||
# dirty in the working copy. | ||||
_newnode = '!' * 21 | ||||
Sean Farley
|
r19537 | class basectx(object): | ||
"""A basectx object represents the common logic for its children: | ||||
changectx: read-only context that is already present in the repo, | ||||
workingctx: a context that represents the working directory and can | ||||
be committed, | ||||
memctx: a context that represents changes in-memory and can also | ||||
be committed.""" | ||||
def __new__(cls, repo, changeid='', *args, **kwargs): | ||||
Sean Farley
|
r19538 | if isinstance(changeid, basectx): | ||
return changeid | ||||
o = super(basectx, cls).__new__(cls) | ||||
o._repo = repo | ||||
o._rev = nullrev | ||||
o._node = nullid | ||||
return o | ||||
Sean Farley
|
r19537 | |||
Sean Farley
|
r19540 | def __str__(self): | ||
return short(self.node()) | ||||
Sean Farley
|
r19545 | def __int__(self): | ||
return self.rev() | ||||
Sean Farley
|
r19546 | def __repr__(self): | ||
return "<%s %s>" % (type(self).__name__, str(self)) | ||||
Sean Farley
|
r19547 | def __eq__(self, other): | ||
try: | ||||
return type(self) == type(other) and self._rev == other._rev | ||||
except AttributeError: | ||||
return False | ||||
Sean Farley
|
r19548 | def __ne__(self, other): | ||
return not (self == other) | ||||
Sean Farley
|
r19550 | def __contains__(self, key): | ||
return key in self._manifest | ||||
Sean Farley
|
r19551 | def __getitem__(self, key): | ||
return self.filectx(key) | ||||
Sean Farley
|
r19552 | def __iter__(self): | ||
Augie Fackler
|
r24227 | return iter(self._manifest) | ||
Sean Farley
|
r19552 | |||
Sean Farley
|
r21466 | def _manifestmatches(self, match, s): | ||
"""generate a new manifest filtered by the match argument | ||||
This method is for internal use only and mainly exists to provide an | ||||
object oriented way for other contexts to customize the manifest | ||||
generation. | ||||
""" | ||||
Martin von Zweigbergk
|
r23305 | return self.manifest().matches(match) | ||
Siddharth Agarwal
|
r21880 | |||
Martin von Zweigbergk
|
r23237 | def _matchstatus(self, other, match): | ||
Sean Farley
|
r21481 | """return match.always if match is none | ||
This internal method provides a way for child objects to override the | ||||
match operator. | ||||
""" | ||||
return match or matchmod.always(self._repo.root, self._repo.getcwd()) | ||||
Sean Farley
|
r21471 | def _buildstatus(self, other, s, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown): | ||
Sean Farley
|
r21471 | """build a status with respect to another context""" | ||
Martin von Zweigbergk
|
r23257 | # Load earliest manifest first for caching reasons. More specifically, | ||
# if you have revisions 1000 and 1001, 1001 is probably stored as a | ||||
# delta against 1000. Thus, if you read 1000 first, we'll reconstruct | ||||
# 1000 and cache it so that when you read 1001, we just need to apply a | ||||
# delta to what's in the cache. So that's one full reconstruction + one | ||||
# delta application. | ||||
Martin von Zweigbergk
|
r23238 | if self.rev() is not None and self.rev() < other.rev(): | ||
self.manifest() | ||||
Sean Farley
|
r21471 | mf1 = other._manifestmatches(match, s) | ||
mf2 = self._manifestmatches(match, s) | ||||
Augie Fackler
|
r23755 | modified, added = [], [] | ||
removed = [] | ||||
Augie Fackler
|
r23757 | clean = [] | ||
Martin von Zweigbergk
|
r23304 | deleted, unknown, ignored = s.deleted, s.unknown, s.ignored | ||
Martin von Zweigbergk
|
r23085 | deletedset = set(deleted) | ||
Augie Fackler
|
r23757 | d = mf1.diff(mf2, clean=listclean) | ||
for fn, value in d.iteritems(): | ||||
Martin von Zweigbergk
|
r23731 | if fn in deletedset: | ||
continue | ||||
Augie Fackler
|
r23757 | if value is None: | ||
clean.append(fn) | ||||
continue | ||||
(node1, flag1), (node2, flag2) = value | ||||
Augie Fackler
|
r23755 | if node1 is None: | ||
added.append(fn) | ||||
elif node2 is None: | ||||
removed.append(fn) | ||||
elif node2 != _newnode: | ||||
# The file was not a new file in mf2, so an entry | ||||
# from diff is really a difference. | ||||
modified.append(fn) | ||||
elif self[fn].cmp(other[fn]): | ||||
# node2 was newnode, but the working file doesn't | ||||
# match the one in mf1. | ||||
modified.append(fn) | ||||
Martin von Zweigbergk
|
r23731 | else: | ||
Augie Fackler
|
r23757 | clean.append(fn) | ||
Augie Fackler
|
r23755 | |||
Pierre-Yves David
|
r21971 | if removed: | ||
# need to filter files if they are already reported as removed | ||||
unknown = [fn for fn in unknown if fn not in mf1] | ||||
ignored = [fn for fn in ignored if fn not in mf1] | ||||
Martin von Zweigbergk
|
r23730 | # if they're deleted, don't report them as removed | ||
removed = [fn for fn in removed if fn not in deletedset] | ||||
Sean Farley
|
r21471 | |||
Martin von Zweigbergk
|
r23302 | return scmutil.status(modified, added, removed, deleted, unknown, | ||
ignored, clean) | ||||
Sean Farley
|
r21471 | |||
Sean Farley
|
r19549 | @propertycache | ||
def substate(self): | ||||
return subrepo.state(self, self._repo.ui) | ||||
Sean Farley
|
r21586 | def subrev(self, subpath): | ||
return self.substate[subpath][1] | ||||
Sean Farley
|
r19541 | def rev(self): | ||
return self._rev | ||||
Sean Farley
|
r19542 | def node(self): | ||
return self._node | ||||
Sean Farley
|
r19543 | def hex(self): | ||
Sean Farley
|
r19544 | return hex(self.node()) | ||
Sean Farley
|
r19553 | def manifest(self): | ||
return self._manifest | ||||
Matt Harbison
|
r24300 | def repo(self): | ||
return self._repo | ||||
Sean Farley
|
r19554 | def phasestr(self): | ||
return phases.phasenames[self.phase()] | ||||
Sean Farley
|
r19555 | def mutable(self): | ||
return self.phase() > phases.public | ||||
Sean Farley
|
r19541 | |||
Augie Fackler
|
r20400 | def getfileset(self, expr): | ||
return fileset.getfileset(self, expr) | ||||
Sean Farley
|
r19734 | def obsolete(self): | ||
"""True if the changeset is obsolete""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'obsolete') | ||||
def extinct(self): | ||||
"""True if the changeset is extinct""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'extinct') | ||||
def unstable(self): | ||||
"""True if the changeset is not obsolete but it's ancestor are""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'unstable') | ||||
def bumped(self): | ||||
"""True if the changeset try to be a successor of a public changeset | ||||
Only non-public and non-obsolete changesets may be bumped. | ||||
""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'bumped') | ||||
def divergent(self): | ||||
"""Is a successors of a changeset with multiple possible successors set | ||||
Only non-public and non-obsolete changesets may be divergent. | ||||
""" | ||||
return self.rev() in obsmod.getrevs(self._repo, 'divergent') | ||||
def troubled(self): | ||||
"""True if the changeset is either unstable, bumped or divergent""" | ||||
return self.unstable() or self.bumped() or self.divergent() | ||||
def troubles(self): | ||||
"""return the list of troubles affecting this changesets. | ||||
Troubles are returned as strings. possible values are: | ||||
- unstable, | ||||
- bumped, | ||||
- divergent. | ||||
""" | ||||
troubles = [] | ||||
if self.unstable(): | ||||
troubles.append('unstable') | ||||
if self.bumped(): | ||||
troubles.append('bumped') | ||||
if self.divergent(): | ||||
troubles.append('divergent') | ||||
return troubles | ||||
Sean Farley
|
r19556 | def parents(self): | ||
"""return contexts for each parent changeset""" | ||||
return self._parents | ||||
Sean Farley
|
r19557 | def p1(self): | ||
return self._parents[0] | ||||
Sean Farley
|
r19558 | def p2(self): | ||
if len(self._parents) == 2: | ||||
return self._parents[1] | ||||
return changectx(self._repo, -1) | ||||
Sean Farley
|
r19559 | def _fileinfo(self, path): | ||
if '_manifest' in self.__dict__: | ||||
try: | ||||
return self._manifest[path], self._manifest.flags(path) | ||||
except KeyError: | ||||
raise error.ManifestLookupError(self._node, path, | ||||
_('not found in manifest')) | ||||
if '_manifestdelta' in self.__dict__ or path in self.files(): | ||||
if path in self._manifestdelta: | ||||
return (self._manifestdelta[path], | ||||
self._manifestdelta.flags(path)) | ||||
node, flag = self._repo.manifest.find(self._changeset[0], path) | ||||
if not node: | ||||
raise error.ManifestLookupError(self._node, path, | ||||
_('not found in manifest')) | ||||
return node, flag | ||||
Sean Farley
|
r19560 | def filenode(self, path): | ||
return self._fileinfo(path)[0] | ||||
Sean Farley
|
r19561 | def flags(self, path): | ||
try: | ||||
return self._fileinfo(path)[1] | ||||
except error.LookupError: | ||||
return '' | ||||
Sean Farley
|
r19562 | def sub(self, path): | ||
return subrepo.subrepo(self, path) | ||||
Sean Farley
|
r19563 | def match(self, pats=[], include=None, exclude=None, default='glob'): | ||
r = self._repo | ||||
return matchmod.match(r.root, r.getcwd(), pats, | ||||
include, exclude, default, | ||||
auditor=r.auditor, ctx=self) | ||||
Sean Farley
|
r19564 | def diff(self, ctx2=None, match=None, **opts): | ||
"""Returns a diff generator for the given contexts and matcher""" | ||||
if ctx2 is None: | ||||
ctx2 = self.p1() | ||||
Sean Farley
|
r19568 | if ctx2 is not None: | ||
Sean Farley
|
r19564 | ctx2 = self._repo[ctx2] | ||
diffopts = patch.diffopts(self._repo.ui, opts) | ||||
Sean Farley
|
r21834 | return patch.diff(self._repo, ctx2, self, match=match, opts=diffopts) | ||
Sean Farley
|
r19564 | |||
Drew Gottlieb
|
r24323 | def dirs(self): | ||
return self._manifest.dirs() | ||||
Sean Farley
|
r19565 | |||
Drew Gottlieb
|
r24325 | def hasdir(self, dir): | ||
return self._manifest.hasdir(dir) | ||||
Sean Farley
|
r19566 | |||
Sean Farley
|
r22055 | def dirty(self, missing=False, merge=True, branch=True): | ||
Sean Farley
|
r19567 | return False | ||
Sean Farley
|
r21594 | def status(self, other=None, match=None, listignored=False, | ||
listclean=False, listunknown=False, listsubrepos=False): | ||||
"""return status of files between two nodes or node and working | ||||
directory. | ||||
If other is None, compare this node with working directory. | ||||
Pierre-Yves David
|
r21722 | |||
returns (modified, added, removed, deleted, unknown, ignored, clean) | ||||
Sean Farley
|
r21594 | """ | ||
ctx1 = self | ||||
ctx2 = self._repo[other] | ||||
# This next code block is, admittedly, fragile logic that tests for | ||||
# reversing the contexts and wouldn't need to exist if it weren't for | ||||
# the fast (and common) code path of comparing the working directory | ||||
# with its first parent. | ||||
# | ||||
# What we're aiming for here is the ability to call: | ||||
# | ||||
# workingctx.status(parentctx) | ||||
# | ||||
# If we always built the manifest for each context and compared those, | ||||
# then we'd be done. But the special case of the above call means we | ||||
# just copy the manifest of the parent. | ||||
reversed = False | ||||
if (not isinstance(ctx1, changectx) | ||||
and isinstance(ctx2, changectx)): | ||||
reversed = True | ||||
ctx1, ctx2 = ctx2, ctx1 | ||||
Martin von Zweigbergk
|
r23237 | match = ctx2._matchstatus(ctx1, match) | ||
Martin von Zweigbergk
|
r23304 | r = scmutil.status([], [], [], [], [], [], []) | ||
Sean Farley
|
r21594 | r = ctx2._buildstatus(ctx1, r, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown) | ||
Sean Farley
|
r21594 | |||
if reversed: | ||||
Martin von Zweigbergk
|
r23301 | # Reverse added and removed. Clear deleted, unknown and ignored as | ||
# these make no sense to reverse. | ||||
r = scmutil.status(r.modified, r.removed, r.added, [], [], [], | ||||
r.clean) | ||||
Sean Farley
|
r21594 | |||
if listsubrepos: | ||||
for subpath, sub in scmutil.itersubrepos(ctx1, ctx2): | ||||
rev2 = ctx2.subrev(subpath) | ||||
try: | ||||
submatch = matchmod.narrowmatcher(subpath, match) | ||||
s = sub.status(rev2, match=submatch, ignored=listignored, | ||||
clean=listclean, unknown=listunknown, | ||||
listsubrepos=True) | ||||
for rfiles, sfiles in zip(r, s): | ||||
rfiles.extend("%s/%s" % (subpath, f) for f in sfiles) | ||||
except error.LookupError: | ||||
self._repo.ui.status(_("skipping missing " | ||||
"subrepository: %s\n") % subpath) | ||||
for l in r: | ||||
l.sort() | ||||
Sean Farley
|
r21616 | |||
Martin von Zweigbergk
|
r23301 | return r | ||
Sean Farley
|
r21594 | |||
Augie Fackler
|
r20035 | def makememctx(repo, parents, text, user, date, branch, files, store, | ||
editor=None): | ||||
def getfilectx(repo, memctx, path): | ||||
Mads Kiilerich
|
r22296 | data, mode, copied = store.getfile(path) | ||
if data is None: | ||||
return None | ||||
islink, isexec = mode | ||||
Sean Farley
|
r21689 | return memfilectx(repo, path, data, islink=islink, isexec=isexec, | ||
copied=copied, memctx=memctx) | ||||
Augie Fackler
|
r20035 | extra = {} | ||
if branch: | ||||
extra['branch'] = encoding.fromlocal(branch) | ||||
ctx = memctx(repo, parents, text, files, getfilectx, user, | ||||
FUJIWARA Katsunori
|
r21238 | date, extra, editor) | ||
Augie Fackler
|
r20035 | return ctx | ||
Sean Farley
|
r19537 | class changectx(basectx): | ||
Matt Mackall
|
r2563 | """A changecontext object makes access to data related to a particular | ||
Mads Kiilerich
|
r19951 | changeset convenient. It represents a read-only context already present in | ||
Sean Farley
|
r19537 | the repo.""" | ||
Matt Mackall
|
r6741 | def __init__(self, repo, changeid=''): | ||
Matt Mackall
|
r2563 | """changeid is a revision number, node, or tag""" | ||
Sean Farley
|
r19539 | |||
# since basectx.__new__ already took care of copying the object, we | ||||
# don't need to do anything in __init__, so we just exit here | ||||
if isinstance(changeid, basectx): | ||||
return | ||||
Matt Mackall
|
r6741 | if changeid == '': | ||
changeid = '.' | ||||
Matt Mackall
|
r2563 | self._repo = repo | ||
Matt Mackall
|
r16376 | |||
Pierre-Yves David
|
r23012 | try: | ||
if isinstance(changeid, int): | ||||
Pierre-Yves David
|
r23013 | self._node = repo.changelog.node(changeid) | ||
Pierre-Yves David
|
r23012 | self._rev = changeid | ||
return | ||||
if isinstance(changeid, long): | ||||
changeid = str(changeid) | ||||
if changeid == 'null': | ||||
self._node = nullid | ||||
self._rev = nullrev | ||||
return | ||||
if changeid == 'tip': | ||||
self._node = repo.changelog.tip() | ||||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Martin von Zweigbergk
|
r24050 | if changeid == '.' or changeid == repo.dirstate.p1(): | ||
# this is a hack to delay/avoid loading obsmarkers | ||||
# when we know that '.' won't be hidden | ||||
self._node = repo.dirstate.p1() | ||||
self._rev = repo.unfiltered().changelog.rev(self._node) | ||||
return | ||||
Pierre-Yves David
|
r23012 | if len(changeid) == 20: | ||
try: | ||||
self._node = changeid | ||||
self._rev = repo.changelog.rev(changeid) | ||||
return | ||||
Pierre-Yves David
|
r23017 | except error.FilteredRepoLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except LookupError: | ||
pass | ||||
Pierre-Yves David
|
r18084 | try: | ||
Pierre-Yves David
|
r23012 | r = int(changeid) | ||
if str(r) != changeid: | ||||
raise ValueError | ||||
l = len(repo.changelog) | ||||
if r < 0: | ||||
r += l | ||||
if r < 0 or r >= l: | ||||
raise ValueError | ||||
self._rev = r | ||||
self._node = repo.changelog.node(r) | ||||
Matt Mackall
|
r16376 | return | ||
Pierre-Yves David
|
r23017 | except error.FilteredIndexError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except (ValueError, OverflowError, IndexError): | ||
Matt Mackall
|
r16376 | pass | ||
Pierre-Yves David
|
r23012 | if len(changeid) == 40: | ||
try: | ||||
self._node = bin(changeid) | ||||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Pierre-Yves David
|
r23017 | except error.FilteredLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except (TypeError, LookupError): | ||
pass | ||||
Matt Mackall
|
r16376 | |||
Sean Farley
|
r23560 | # lookup bookmarks through the name interface | ||
try: | ||||
Ryan McElroy
|
r23561 | self._node = repo.names.singlenode(repo, changeid) | ||
Matt Mackall
|
r16376 | self._rev = repo.changelog.rev(self._node) | ||
return | ||||
Sean Farley
|
r23560 | except KeyError: | ||
pass | ||||
Pierre-Yves David
|
r23017 | except error.FilteredRepoLookupError: | ||
raise | ||||
Pierre-Yves David
|
r23012 | except error.RepoLookupError: | ||
Matt Mackall
|
r16376 | pass | ||
Pierre-Yves David
|
r23017 | self._node = repo.unfiltered().changelog._partialmatch(changeid) | ||
Pierre-Yves David
|
r23012 | if self._node is not None: | ||
self._rev = repo.changelog.rev(self._node) | ||||
return | ||||
Matt Mackall
|
r16376 | |||
Pierre-Yves David
|
r23012 | # lookup failed | ||
# check if it might have come from damaged dirstate | ||||
# | ||||
# XXX we could avoid the unfiltered if we had a recognizable | ||||
# exception for filtered changeset access | ||||
if changeid in repo.unfiltered().dirstate.parents(): | ||||
msg = _("working directory has unknown parent '%s'!") | ||||
raise error.Abort(msg % short(changeid)) | ||||
try: | ||||
if len(changeid) == 20: | ||||
changeid = hex(changeid) | ||||
except TypeError: | ||||
pass | ||||
Pierre-Yves David
|
r23017 | except (error.FilteredIndexError, error.FilteredLookupError, | ||
error.FilteredRepoLookupError): | ||||
Pierre-Yves David
|
r23046 | if repo.filtername == 'visible': | ||
msg = _("hidden revision '%s'") % changeid | ||||
hint = _('use --hidden to access hidden revisions') | ||||
raise error.FilteredRepoLookupError(msg, hint=hint) | ||||
Pierre-Yves David
|
r23045 | msg = _("filtered revision '%s' (not in '%s' subset)") | ||
msg %= (changeid, repo.filtername) | ||||
raise error.FilteredRepoLookupError(msg) | ||||
Pierre-Yves David
|
r23013 | except IndexError: | ||
pass | ||||
Matt Mackall
|
r16376 | raise error.RepoLookupError( | ||
_("unknown revision '%s'") % changeid) | ||||
Matt Mackall
|
r2563 | |||
Paul Moore
|
r6469 | def __hash__(self): | ||
try: | ||||
return hash(self._rev) | ||||
except AttributeError: | ||||
return id(self) | ||||
Matt Mackall
|
r3168 | def __nonzero__(self): | ||
Thomas Arendsen Hein
|
r3578 | return self._rev != nullrev | ||
Matt Mackall
|
r3168 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _changeset(self): | ||
Matt Mackall
|
r16377 | return self._repo.changelog.read(self.rev()) | ||
Dirkjan Ochtman
|
r7368 | |||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _manifest(self): | ||
return self._repo.manifest.read(self._changeset[0]) | ||||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _manifestdelta(self): | ||
return self._repo.manifest.readdelta(self._changeset[0]) | ||||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _parents(self): | ||
p = self._repo.changelog.parentrevs(self._rev) | ||||
if p[1] == nullrev: | ||||
p = p[:-1] | ||||
return [changectx(self._repo, x) for x in p] | ||||
Matt Mackall
|
r3215 | |||
Matt Mackall
|
r10282 | def changeset(self): | ||
return self._changeset | ||||
def manifestnode(self): | ||||
return self._changeset[0] | ||||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r10282 | def user(self): | ||
return self._changeset[1] | ||||
def date(self): | ||||
return self._changeset[2] | ||||
def files(self): | ||||
return self._changeset[3] | ||||
def description(self): | ||||
return self._changeset[4] | ||||
def branch(self): | ||||
Matt Mackall
|
r13047 | return encoding.tolocal(self._changeset[5].get("branch")) | ||
Brodie Rao
|
r16720 | def closesbranch(self): | ||
return 'close' in self._changeset[5] | ||||
Matt Mackall
|
r10282 | def extra(self): | ||
return self._changeset[5] | ||||
def tags(self): | ||||
return self._repo.nodetags(self._node) | ||||
David Soria Parra
|
r13384 | def bookmarks(self): | ||
return self._repo.nodebookmarks(self._node) | ||||
Pierre-Yves David
|
r15421 | def phase(self): | ||
Patrick Mezard
|
r16657 | return self._repo._phasecache.phase(self._repo, self._rev) | ||
Pierre-Yves David
|
r14644 | def hidden(self): | ||
Kevin Bullock
|
r18382 | return self._rev in repoview.filterrevs(self._repo, 'visible') | ||
Matt Mackall
|
r2563 | |||
def children(self): | ||||
"""return contexts for each child changeset""" | ||||
Benoit Boissinot
|
r2627 | c = self._repo.changelog.children(self._node) | ||
Thomas Arendsen Hein
|
r3673 | return [changectx(self._repo, x) for x in c] | ||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r6876 | def ancestors(self): | ||
Bryan O'Sullivan
|
r16866 | for a in self._repo.changelog.ancestors([self._rev]): | ||
Matt Mackall
|
r6876 | yield changectx(self._repo, a) | ||
def descendants(self): | ||||
Bryan O'Sullivan
|
r16867 | for d in self._repo.changelog.descendants([self._rev]): | ||
Matt Mackall
|
r6876 | yield changectx(self._repo, d) | ||
Benoit Boissinot
|
r3966 | def filectx(self, path, fileid=None, filelog=None): | ||
Matt Mackall
|
r2563 | """get a file context from this changeset""" | ||
Benoit Boissinot
|
r2628 | if fileid is None: | ||
fileid = self.filenode(path) | ||||
Benoit Boissinot
|
r3966 | return filectx(self._repo, path, fileid=fileid, | ||
changectx=self, filelog=filelog) | ||||
Matt Mackall
|
r2563 | |||
Matt Mackall
|
r21203 | def ancestor(self, c2, warn=False): | ||
Mads Kiilerich
|
r22389 | """return the "best" ancestor context of self and c2 | ||
If there are multiple candidates, it will show a message and check | ||||
merge.preferancestor configuration before falling back to the | ||||
revlog ancestor.""" | ||||
Matt Mackall
|
r9843 | # deal with workingctxs | ||
n2 = c2._node | ||||
Martin Geisler
|
r13031 | if n2 is None: | ||
Matt Mackall
|
r9843 | n2 = c2._parents[0]._node | ||
Mads Kiilerich
|
r21125 | cahs = self._repo.changelog.commonancestorsheads(self._node, n2) | ||
if not cahs: | ||||
anc = nullid | ||||
elif len(cahs) == 1: | ||||
anc = cahs[0] | ||||
else: | ||||
Mads Kiilerich
|
r21126 | for r in self._repo.ui.configlist('merge', 'preferancestor'): | ||
Mads Kiilerich
|
r22671 | try: | ||
ctx = changectx(self._repo, r) | ||||
except error.RepoLookupError: | ||||
Mads Kiilerich
|
r22180 | continue | ||
Mads Kiilerich
|
r21126 | anc = ctx.node() | ||
if anc in cahs: | ||||
break | ||||
else: | ||||
anc = self._repo.changelog.ancestor(self._node, n2) | ||||
Matt Mackall
|
r21203 | if warn: | ||
self._repo.ui.status( | ||||
(_("note: using %s as ancestor of %s and %s\n") % | ||||
(short(anc), short(self._node), short(n2))) + | ||||
''.join(_(" alternatively, use --config " | ||||
"merge.preferancestor=%s\n") % | ||||
short(n) for n in sorted(cahs) if n != anc)) | ||||
Mads Kiilerich
|
r21125 | return changectx(self._repo, anc) | ||
Matt Mackall
|
r3125 | |||
FUJIWARA Katsunori
|
r17626 | def descendant(self, other): | ||
"""True if other is descendant of this changeset""" | ||||
return self._repo.changelog.descendant(self._rev, other._rev) | ||||
Matt Mackall
|
r6764 | def walk(self, match): | ||
Simon Heimberg
|
r8380 | fset = set(match.files()) | ||
Durham Goode
|
r20292 | # avoid the entire walk if we're only looking for specific files | ||
if fset and not match.anypats(): | ||||
Martin von Zweigbergk
|
r24380 | if util.all(fn in self for fn in fset): | ||
Durham Goode
|
r20292 | for fn in sorted(fset): | ||
if match(fn): | ||||
yield fn | ||||
raise StopIteration | ||||
Matt Mackall
|
r6764 | for fn in self: | ||
FUJIWARA Katsunori
|
r16145 | if fn in fset: | ||
# specified pattern is the exact name | ||||
fset.remove(fn) | ||||
Matt Mackall
|
r6764 | if match(fn): | ||
yield fn | ||||
Martin von Zweigbergk
|
r24381 | # for dirstate.walk, files=['.'] means "walk the whole tree". | ||
# follow that here, too | ||||
fset.discard('.') | ||||
Simon Heimberg
|
r8380 | for fn in sorted(fset): | ||
Drew Gottlieb
|
r24326 | if not self.hasdir(fn): | ||
match.bad(fn, _('no such file in rev %s') % self) | ||||
Matt Mackall
|
r6764 | |||
Siddharth Agarwal
|
r21985 | def matches(self, match): | ||
return self.walk(match) | ||||
Sean Farley
|
r19572 | class basefilectx(object): | ||
"""A filecontext object represents the common logic for its children: | ||||
filectx: read-only access to a filerevision that is already present | ||||
in the repo, | ||||
workingfilectx: a filecontext that represents files from the working | ||||
directory, | ||||
memfilectx: a filecontext that represents files in-memory.""" | ||||
def __new__(cls, repo, path, *args, **kwargs): | ||||
return super(basefilectx, cls).__new__(cls) | ||||
Sean Farley
|
r19573 | @propertycache | ||
def _filelog(self): | ||||
return self._repo.file(self._path) | ||||
Sean Farley
|
r19574 | @propertycache | ||
def _changeid(self): | ||||
if '_changeid' in self.__dict__: | ||||
return self._changeid | ||||
elif '_changectx' in self.__dict__: | ||||
return self._changectx.rev() | ||||
Matt Mackall
|
r23983 | elif '_descendantrev' in self.__dict__: | ||
# this file context was created from a revision with a known | ||||
# descendant, we can (lazily) correct for linkrev aliases | ||||
return self._adjustlinkrev(self._path, self._filelog, | ||||
self._filenode, self._descendantrev) | ||||
Sean Farley
|
r19574 | else: | ||
return self._filelog.linkrev(self._filerev) | ||||
Sean Farley
|
r19575 | @propertycache | ||
def _filenode(self): | ||||
if '_fileid' in self.__dict__: | ||||
return self._filelog.lookup(self._fileid) | ||||
else: | ||||
return self._changectx.filenode(self._path) | ||||
Sean Farley
|
r19576 | @propertycache | ||
def _filerev(self): | ||||
return self._filelog.rev(self._filenode) | ||||
Sean Farley
|
r19577 | @propertycache | ||
def _repopath(self): | ||||
return self._path | ||||
Sean Farley
|
r19578 | def __nonzero__(self): | ||
try: | ||||
self._filenode | ||||
return True | ||||
except error.LookupError: | ||||
# file is missing | ||||
return False | ||||
Sean Farley
|
r19579 | def __str__(self): | ||
Sean Farley
|
r19660 | return "%s@%s" % (self.path(), self._changectx) | ||
Sean Farley
|
r19579 | |||
Sean Farley
|
r19580 | def __repr__(self): | ||
return "<%s %s>" % (type(self).__name__, str(self)) | ||||
Sean Farley
|
r19581 | def __hash__(self): | ||
try: | ||||
return hash((self._path, self._filenode)) | ||||
except AttributeError: | ||||
return id(self) | ||||
Sean Farley
|
r19582 | def __eq__(self, other): | ||
try: | ||||
return (type(self) == type(other) and self._path == other._path | ||||
and self._filenode == other._filenode) | ||||
except AttributeError: | ||||
return False | ||||
Sean Farley
|
r19583 | def __ne__(self, other): | ||
return not (self == other) | ||||
Sean Farley
|
r19584 | def filerev(self): | ||
return self._filerev | ||||
Sean Farley
|
r19585 | def filenode(self): | ||
return self._filenode | ||||
Sean Farley
|
r19586 | def flags(self): | ||
return self._changectx.flags(self._path) | ||||
Sean Farley
|
r19587 | def filelog(self): | ||
return self._filelog | ||||
Sean Farley
|
r19588 | def rev(self): | ||
return self._changeid | ||||
Sean Farley
|
r19589 | def linkrev(self): | ||
return self._filelog.linkrev(self._filerev) | ||||
Sean Farley
|
r19590 | def node(self): | ||
return self._changectx.node() | ||||
Sean Farley
|
r19591 | def hex(self): | ||
return self._changectx.hex() | ||||
Sean Farley
|
r19592 | def user(self): | ||
return self._changectx.user() | ||||
Sean Farley
|
r19593 | def date(self): | ||
return self._changectx.date() | ||||
Sean Farley
|
r19594 | def files(self): | ||
return self._changectx.files() | ||||
Sean Farley
|
r19595 | def description(self): | ||
return self._changectx.description() | ||||
Sean Farley
|
r19596 | def branch(self): | ||
return self._changectx.branch() | ||||
Sean Farley
|
r19597 | def extra(self): | ||
return self._changectx.extra() | ||||
Sean Farley
|
r19598 | def phase(self): | ||
return self._changectx.phase() | ||||
Sean Farley
|
r19599 | def phasestr(self): | ||
return self._changectx.phasestr() | ||||
Sean Farley
|
r19600 | def manifest(self): | ||
return self._changectx.manifest() | ||||
Sean Farley
|
r19601 | def changectx(self): | ||
return self._changectx | ||||
Matt Harbison
|
r24333 | def repo(self): | ||
return self._repo | ||||
Sean Farley
|
r19584 | |||
Sean Farley
|
r19602 | def path(self): | ||
return self._path | ||||
Sean Farley
|
r19603 | def isbinary(self): | ||
try: | ||||
return util.binary(self.data()) | ||||
except IOError: | ||||
return False | ||||
Sean Farley
|
r22054 | def isexec(self): | ||
return 'x' in self.flags() | ||||
def islink(self): | ||||
return 'l' in self.flags() | ||||
Sean Farley
|
r19603 | |||
Sean Farley
|
r19604 | def cmp(self, fctx): | ||
"""compare with other file context | ||||
returns True if different than fctx. | ||||
""" | ||||
if (fctx._filerev is None | ||||
and (self._repo._encodefilterpats | ||||
# if file data starts with '\1\n', empty metadata block is | ||||
# prepended, which adds 4 bytes to filelog.size(). | ||||
or self.size() - 4 == fctx.size()) | ||||
or self.size() == fctx.size()): | ||||
return self._filelog.cmp(self._filenode, fctx.data()) | ||||
return True | ||||
Pierre-Yves David
|
r23979 | def _adjustlinkrev(self, path, filelog, fnode, srcrev, inclusive=False): | ||
Mads Kiilerich
|
r24180 | """return the first ancestor of <srcrev> introducing <fnode> | ||
Pierre-Yves David
|
r23979 | |||
If the linkrev of the file revision does not point to an ancestor of | ||||
srcrev, we'll walk down the ancestors until we find one introducing | ||||
this file revision. | ||||
:repo: a localrepository object (used to access changelog and manifest) | ||||
:path: the file path | ||||
:fnode: the nodeid of the file revision | ||||
:filelog: the filelog of this path | ||||
:srcrev: the changeset revision we search ancestors from | ||||
:inclusive: if true, the src revision will also be checked | ||||
""" | ||||
repo = self._repo | ||||
cl = repo.unfiltered().changelog | ||||
ma = repo.manifest | ||||
# fetch the linkrev | ||||
fr = filelog.rev(fnode) | ||||
lkr = filelog.linkrev(fr) | ||||
Pierre-Yves David
|
r23980 | # hack to reuse ancestor computation when searching for renames | ||
memberanc = getattr(self, '_ancestrycontext', None) | ||||
iteranc = None | ||||
Pierre-Yves David
|
r24411 | if srcrev is None: | ||
# wctx case, used by workingfilectx during mergecopy | ||||
revs = [p.rev() for p in self._repo[None].parents()] | ||||
inclusive = True # we skipped the real (revless) source | ||||
else: | ||||
revs = [srcrev] | ||||
Pierre-Yves David
|
r23980 | if memberanc is None: | ||
Pierre-Yves David
|
r24411 | memberanc = iteranc = cl.ancestors(revs, lkr, | ||
inclusive=inclusive) | ||||
Pierre-Yves David
|
r23979 | # check if this linkrev is an ancestor of srcrev | ||
Pierre-Yves David
|
r23980 | if lkr not in memberanc: | ||
if iteranc is None: | ||||
Pierre-Yves David
|
r24410 | iteranc = cl.ancestors(revs, lkr, inclusive=inclusive) | ||
Pierre-Yves David
|
r23980 | for a in iteranc: | ||
Pierre-Yves David
|
r23979 | ac = cl.read(a) # get changeset data (we avoid object creation) | ||
if path in ac[3]: # checking the 'files' field. | ||||
# The file has been touched, check if the content is | ||||
# similar to the one we search for. | ||||
if fnode == ma.readfast(ac[0]).get(path): | ||||
return a | ||||
# In theory, we should never get out of that loop without a result. | ||||
# But if manifest uses a buggy file revision (not children of the | ||||
# one it replaces) we could. Such a buggy situation will likely | ||||
# result is crash somewhere else at to some point. | ||||
return lkr | ||||
Pierre-Yves David
|
r23703 | def introrev(self): | ||
"""return the rev of the changeset which introduced this file revision | ||||
This method is different from linkrev because it take into account the | ||||
changeset the filectx was created from. It ensures the returned | ||||
revision is one of its ancestors. This prevents bugs from | ||||
'linkrev-shadowing' when a file revision is used by multiple | ||||
changesets. | ||||
""" | ||||
lkr = self.linkrev() | ||||
attrs = vars(self) | ||||
noctx = not ('_changeid' in attrs or '_changectx' in attrs) | ||||
if noctx or self.rev() == lkr: | ||||
return self.linkrev() | ||||
Pierre-Yves David
|
r23979 | return self._adjustlinkrev(self._path, self._filelog, self._filenode, | ||
self.rev(), inclusive=True) | ||||
Pierre-Yves David
|
r23703 | |||
Sean Farley
|
r19605 | def parents(self): | ||
Mads Kiilerich
|
r22201 | _path = self._path | ||
Sean Farley
|
r19605 | fl = self._filelog | ||
Pierre-Yves David
|
r23688 | parents = self._filelog.parents(self._filenode) | ||
pl = [(_path, node, fl) for node in parents if node != nullid] | ||||
Sean Farley
|
r19605 | |||
Pierre-Yves David
|
r23702 | r = fl.renamed(self._filenode) | ||
Sean Farley
|
r19605 | if r: | ||
Pierre-Yves David
|
r23688 | # - In the simple rename case, both parent are nullid, pl is empty. | ||
# - In case of merge, only one of the parent is null id and should | ||||
# be replaced with the rename information. This parent is -always- | ||||
# the first one. | ||||
# | ||||
Mads Kiilerich
|
r24180 | # As null id have always been filtered out in the previous list | ||
Pierre-Yves David
|
r23688 | # comprehension, inserting to 0 will always result in "replacing | ||
# first nullid parent with rename information. | ||||
Pierre-Yves David
|
r23699 | pl.insert(0, (r[0], r[1], self._repo.file(r[0]))) | ||
Sean Farley
|
r19605 | |||
Pierre-Yves David
|
r23702 | ret = [] | ||
for path, fnode, l in pl: | ||||
if '_changeid' in vars(self) or '_changectx' in vars(self): | ||||
# If self is associated with a changeset (probably explicitly | ||||
# fed), ensure the created filectx is associated with a | ||||
# changeset that is an ancestor of self.changectx. | ||||
Matt Mackall
|
r23984 | # This lets us later use _adjustlinkrev to get a correct link. | ||
fctx = filectx(self._repo, path, fileid=fnode, filelog=l) | ||||
fctx._descendantrev = self.rev() | ||||
Pierre-Yves David
|
r23980 | fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) | ||
Matt Mackall
|
r23984 | elif '_descendantrev' in vars(self): | ||
# Otherwise propagate _descendantrev if we have one associated. | ||||
fctx = filectx(self._repo, path, fileid=fnode, filelog=l) | ||||
fctx._descendantrev = self._descendantrev | ||||
fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) | ||||
Pierre-Yves David
|
r23702 | else: | ||
fctx = filectx(self._repo, path, fileid=fnode, filelog=l) | ||||
ret.append(fctx) | ||||
return ret | ||||
Sean Farley
|
r19605 | |||
Sean Farley
|
r19606 | def p1(self): | ||
return self.parents()[0] | ||||
Sean Farley
|
r19607 | def p2(self): | ||
p = self.parents() | ||||
if len(p) == 2: | ||||
return p[1] | ||||
return filectx(self._repo, self._path, fileid=-1, filelog=self._filelog) | ||||
Patrick Mezard
|
r15528 | def annotate(self, follow=False, linenumber=None, diffopts=None): | ||
Brendan Cully
|
r3172 | '''returns a list of tuples of (ctx, line) for each line | ||
in the file, where ctx is the filectx of the node where | ||||
FUJIWARA Katsunori
|
r4856 | that line was last changed. | ||
This returns tuples of ((ctx, linenumber), line) for each line, | ||||
if "linenumber" parameter is NOT "None". | ||||
In such tuples, linenumber means one at the first appearance | ||||
in the managed file. | ||||
To reduce annotation cost, | ||||
this returns fixed value(False is used) as linenumber, | ||||
if "linenumber" parameter is "False".''' | ||||
Brendan Cully
|
r3172 | |||
Yuya Nishihara
|
r22191 | if linenumber is None: | ||
Yuya Nishihara
|
r22192 | def decorate(text, rev): | ||
return ([rev] * len(text.splitlines()), text) | ||||
Yuya Nishihara
|
r22191 | elif linenumber: | ||
Yuya Nishihara
|
r22192 | def decorate(text, rev): | ||
size = len(text.splitlines()) | ||||
return ([(rev, i) for i in xrange(1, size + 1)], text) | ||||
Yuya Nishihara
|
r22191 | else: | ||
Yuya Nishihara
|
r22192 | def decorate(text, rev): | ||
return ([(rev, False)] * len(text.splitlines()), text) | ||||
FUJIWARA Katsunori
|
r4856 | |||
Brendan Cully
|
r3172 | def pair(parent, child): | ||
Patrick Mezard
|
r15528 | blocks = mdiff.allblocks(parent[1], child[1], opts=diffopts, | ||
refine=True) | ||||
for (a1, a2, b1, b2), t in blocks: | ||||
# Changed blocks ('!') or blocks made only of blank lines ('~') | ||||
# belong to the child. | ||||
if t == '=': | ||||
child[0][b1:b2] = parent[0][a1:a2] | ||||
Brendan Cully
|
r3172 | return child | ||
Matt Mackall
|
r9097 | getlog = util.lrucachefunc(lambda x: self._repo.file(x)) | ||
Brendan Cully
|
r3172 | |||
def parents(f): | ||||
Durham Goode
|
r19292 | pl = f.parents() | ||
# Don't return renamed parents if we aren't following. | ||||
if not follow: | ||||
pl = [p for p in pl if p.path() == f.path()] | ||||
Brendan Cully
|
r3172 | |||
Durham Goode
|
r19292 | # renamed filectx won't have a filelog yet, so set it | ||
# from the cache to save time | ||||
for p in pl: | ||||
if not '_filelog' in p.__dict__: | ||||
p._filelog = getlog(p.path()) | ||||
Brendan Cully
|
r3146 | |||
Durham Goode
|
r19292 | return pl | ||
Matt Mackall
|
r3217 | |||
Brendan Cully
|
r3404 | # use linkrev to find the first changeset where self appeared | ||
Pierre-Yves David
|
r23705 | base = self | ||
introrev = self.introrev() | ||||
if self.rev() != introrev: | ||||
Durham Goode
|
r23770 | base = self.filectx(self.filenode(), changeid=introrev) | ||
Pierre-Yves David
|
r24407 | ac = self._repo.changelog.ancestors([introrev], inclusive=True) | ||
base._ancestrycontext = ac | ||||
Brendan Cully
|
r3404 | |||
Matt Mackall
|
r13552 | # This algorithm would prefer to be recursive, but Python is a | ||
# bit recursion-hostile. Instead we do an iterative | ||||
# depth-first search. | ||||
visit = [base] | ||||
hist = {} | ||||
pcache = {} | ||||
Brendan Cully
|
r3404 | needed = {base: 1} | ||
Brendan Cully
|
r3172 | while visit: | ||
Matt Mackall
|
r13552 | f = visit[-1] | ||
FUJIWARA Katsunori
|
r18993 | pcached = f in pcache | ||
if not pcached: | ||||
Matt Mackall
|
r13552 | pcache[f] = parents(f) | ||
Brendan Cully
|
r3172 | |||
Matt Mackall
|
r13552 | ready = True | ||
pl = pcache[f] | ||||
for p in pl: | ||||
if p not in hist: | ||||
ready = False | ||||
visit.append(p) | ||||
FUJIWARA Katsunori
|
r18993 | if not pcached: | ||
Matt Mackall
|
r13552 | needed[p] = needed.get(p, 0) + 1 | ||
if ready: | ||||
visit.pop() | ||||
FUJIWARA Katsunori
|
r18992 | reusable = f in hist | ||
if reusable: | ||||
curr = hist[f] | ||||
else: | ||||
curr = decorate(f.data(), f) | ||||
Matt Mackall
|
r13552 | for p in pl: | ||
FUJIWARA Katsunori
|
r18992 | if not reusable: | ||
curr = pair(hist[p], curr) | ||||
Matt Mackall
|
r13552 | if needed[p] == 1: | ||
del hist[p] | ||||
FUJIWARA Katsunori
|
r19061 | del needed[p] | ||
Matt Mackall
|
r13552 | else: | ||
needed[p] -= 1 | ||||
Matt Mackall
|
r6762 | |||
Matt Mackall
|
r13552 | hist[f] = curr | ||
pcache[f] = [] | ||||
Brendan Cully
|
r3172 | |||
Matt Mackall
|
r13552 | return zip(hist[base][0], hist[base][1].splitlines(True)) | ||
Matt Mackall
|
r3124 | |||
Sean Farley
|
r19610 | def ancestors(self, followfirst=False): | ||
visit = {} | ||||
c = self | ||||
Jordi Gutiérrez Hermoso
|
r24306 | if followfirst: | ||
cut = 1 | ||||
else: | ||||
cut = None | ||||
Sean Farley
|
r19610 | while True: | ||
for parent in c.parents()[:cut]: | ||||
Matt Mackall
|
r23981 | visit[(parent.linkrev(), parent.filenode())] = parent | ||
Sean Farley
|
r19610 | if not visit: | ||
break | ||||
c = visit.pop(max(visit)) | ||||
yield c | ||||
Sean Farley
|
r19608 | class filectx(basefilectx): | ||
"""A filecontext object makes access to data related to a particular | ||||
filerevision convenient.""" | ||||
def __init__(self, repo, path, changeid=None, fileid=None, | ||||
filelog=None, changectx=None): | ||||
"""changeid can be a changeset revision, node, or tag. | ||||
fileid can be a file revision or node.""" | ||||
self._repo = repo | ||||
self._path = path | ||||
assert (changeid is not None | ||||
or fileid is not None | ||||
or changectx is not None), \ | ||||
("bad args: changeid=%r, fileid=%r, changectx=%r" | ||||
% (changeid, fileid, changectx)) | ||||
if filelog is not None: | ||||
self._filelog = filelog | ||||
if changeid is not None: | ||||
self._changeid = changeid | ||||
if changectx is not None: | ||||
self._changectx = changectx | ||||
if fileid is not None: | ||||
self._fileid = fileid | ||||
@propertycache | ||||
def _changectx(self): | ||||
try: | ||||
return changectx(self._repo, self._changeid) | ||||
Pierre-Yves David
|
r23687 | except error.FilteredRepoLookupError: | ||
Sean Farley
|
r19608 | # Linkrev may point to any revision in the repository. When the | ||
# repository is filtered this may lead to `filectx` trying to build | ||||
# `changectx` for filtered revision. In such case we fallback to | ||||
# creating `changectx` on the unfiltered version of the reposition. | ||||
# This fallback should not be an issue because `changectx` from | ||||
# `filectx` are not used in complex operations that care about | ||||
# filtering. | ||||
# | ||||
# This fallback is a cheap and dirty fix that prevent several | ||||
# crashes. It does not ensure the behavior is correct. However the | ||||
# behavior was not correct before filtering either and "incorrect | ||||
# behavior" is seen as better as "crash" | ||||
# | ||||
# Linkrevs have several serious troubles with filtering that are | ||||
# complicated to solve. Proper handling of the issue here should be | ||||
# considered when solving linkrev issue are on the table. | ||||
return changectx(self._repo.unfiltered(), self._changeid) | ||||
Durham Goode
|
r23770 | def filectx(self, fileid, changeid=None): | ||
Sean Farley
|
r19608 | '''opens an arbitrary revision of the file without | ||
opening a new filelog''' | ||||
return filectx(self._repo, self._path, fileid=fileid, | ||||
Durham Goode
|
r23770 | filelog=self._filelog, changeid=changeid) | ||
Sean Farley
|
r19608 | |||
def data(self): | ||||
Mike Edgar
|
r22932 | try: | ||
return self._filelog.read(self._filenode) | ||||
except error.CensoredNodeError: | ||||
if self._repo.ui.config("censor", "policy", "abort") == "ignore": | ||||
return "" | ||||
raise util.Abort(_("censored node: %s") % short(self._filenode), | ||||
FUJIWARA Katsunori
|
r23110 | hint=_("set censor.policy to ignore errors")) | ||
Mike Edgar
|
r22932 | |||
Sean Farley
|
r19608 | def size(self): | ||
return self._filelog.size(self._filerev) | ||||
def renamed(self): | ||||
"""check if file was actually renamed in this changeset revision | ||||
If rename logged in file revision, we report copy for changeset only | ||||
if file revisions linkrev points back to the changeset in question | ||||
or both changeset parents contain different file revisions. | ||||
""" | ||||
renamed = self._filelog.renamed(self._filenode) | ||||
if not renamed: | ||||
return renamed | ||||
if self.rev() == self.linkrev(): | ||||
return renamed | ||||
name = self.path() | ||||
fnode = self._filenode | ||||
for p in self._changectx.parents(): | ||||
try: | ||||
if fnode == p.filenode(name): | ||||
return None | ||||
except error.LookupError: | ||||
pass | ||||
return renamed | ||||
def children(self): | ||||
# hard for renames | ||||
c = self._filelog.children(self._filenode) | ||||
return [filectx(self._repo, self._path, fileid=x, | ||||
filelog=self._filelog) for x in c] | ||||
Sean Farley
|
r19733 | class committablectx(basectx): | ||
"""A committablectx object provides common functionality for a context that | ||||
Sean Farley
|
r19664 | wants the ability to commit, e.g. workingctx or memctx.""" | ||
def __init__(self, repo, text="", user=None, date=None, extra=None, | ||||
changes=None): | ||||
Matt Mackall
|
r3217 | self._repo = repo | ||
self._rev = None | ||||
self._node = None | ||||
Patrick Mezard
|
r6709 | self._text = text | ||
Christian Ebert
|
r6718 | if date: | ||
Patrick Mezard
|
r6709 | self._date = util.parsedate(date) | ||
Matt Mackall
|
r6817 | if user: | ||
self._user = user | ||||
Patrick Mezard
|
r6707 | if changes: | ||
Sean Farley
|
r21592 | self._status = changes | ||
Matt Mackall
|
r3217 | |||
Patrick Mezard
|
r6708 | self._extra = {} | ||
if extra: | ||||
self._extra = extra.copy() | ||||
if 'branch' not in self._extra: | ||||
try: | ||||
Matt Mackall
|
r13047 | branch = encoding.fromlocal(self._repo.dirstate.branch()) | ||
Patrick Mezard
|
r6708 | except UnicodeDecodeError: | ||
raise util.Abort(_('branch name not in UTF-8!')) | ||||
self._extra['branch'] = branch | ||||
if self._extra['branch'] == '': | ||||
self._extra['branch'] = 'default' | ||||
Sean Farley
|
r19666 | def __str__(self): | ||
return str(self._parents[0]) + "+" | ||||
Sean Farley
|
r19667 | def __nonzero__(self): | ||
return True | ||||
Matt Mackall
|
r15337 | def _buildflagfunc(self): | ||
# Create a fallback function for getting file flags when the | ||||
# filesystem doesn't support them | ||||
copiesget = self._repo.dirstate.copies().get | ||||
if len(self._parents) < 2: | ||||
# when we have one parent, it's easy: copy from parent | ||||
man = self._parents[0].manifest() | ||||
def func(f): | ||||
f = copiesget(f, f) | ||||
return man.flags(f) | ||||
else: | ||||
# merges are tricky: we try to reconstruct the unstored | ||||
# result from the merge (issue1802) | ||||
p1, p2 = self._parents | ||||
pa = p1.ancestor(p2) | ||||
m1, m2, ma = p1.manifest(), p2.manifest(), pa.manifest() | ||||
def func(f): | ||||
f = copiesget(f, f) # may be wrong for merges with copies | ||||
fl1, fl2, fla = m1.flags(f), m2.flags(f), ma.flags(f) | ||||
if fl1 == fl2: | ||||
return fl1 | ||||
if fl1 == fla: | ||||
return fl2 | ||||
if fl2 == fla: | ||||
return fl1 | ||||
return '' # punt for conflicts | ||||
return func | ||||
Sean Farley
|
r19670 | @propertycache | ||
def _flagfunc(self): | ||||
return self._repo.dirstate.flagfunc(self._buildflagfunc) | ||||
Matt Mackall
|
r15337 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _manifest(self): | ||
Pierre-Yves David
|
r23410 | """generate a manifest corresponding to the values in self._status | ||
This reuse the file nodeid from parent, but we append an extra letter | ||||
Mads Kiilerich
|
r23543 | when modified. Modified files get an extra 'm' while added files get | ||
an extra 'a'. This is used by manifests merge to see that files | ||||
Pierre-Yves David
|
r23410 | are different and by update logic to avoid deleting newly added files. | ||
""" | ||||
Matt Mackall
|
r3217 | |||
Pierre-Yves David
|
r23401 | man1 = self._parents[0].manifest() | ||
man = man1.copy() | ||||
Benoit Boissinot
|
r10921 | if len(self._parents) > 1: | ||
man2 = self.p2().manifest() | ||||
def getman(f): | ||||
Pierre-Yves David
|
r23401 | if f in man1: | ||
return man1 | ||||
Benoit Boissinot
|
r10921 | return man2 | ||
else: | ||||
Pierre-Yves David
|
r23401 | getman = lambda f: man1 | ||
Matt Mackall
|
r15337 | |||
copied = self._repo.dirstate.copies() | ||||
ff = self._flagfunc | ||||
Martin von Zweigbergk
|
r22916 | for i, l in (("a", self._status.added), ("m", self._status.modified)): | ||
Matt Mackall
|
r3217 | for f in l: | ||
Benoit Boissinot
|
r10921 | orig = copied.get(f, f) | ||
man[f] = getman(orig).get(orig, nullid) + i | ||||
Matt Mackall
|
r3823 | try: | ||
Augie Fackler
|
r22942 | man.setflag(f, ff(f)) | ||
Matt Mackall
|
r3823 | except OSError: | ||
pass | ||||
Matt Mackall
|
r3217 | |||
Martin von Zweigbergk
|
r22916 | for f in self._status.deleted + self._status.removed: | ||
Giorgos Keramidas
|
r3325 | if f in man: | ||
del man[f] | ||||
Matt Mackall
|
r3217 | |||
Dirkjan Ochtman
|
r7368 | return man | ||
Sean Farley
|
r19672 | @propertycache | ||
def _status(self): | ||||
Sean Farley
|
r21592 | return self._repo.status() | ||
Sean Farley
|
r19672 | |||
Sean Farley
|
r19674 | @propertycache | ||
def _user(self): | ||||
return self._repo.ui.username() | ||||
Sean Farley
|
r19676 | @propertycache | ||
def _date(self): | ||||
return util.makedate() | ||||
Sean Farley
|
r21587 | def subrev(self, subpath): | ||
return None | ||||
Sean Farley
|
r19675 | def user(self): | ||
return self._user or self._repo.ui.username() | ||||
Sean Farley
|
r19677 | def date(self): | ||
return self._date | ||||
Sean Farley
|
r19678 | def description(self): | ||
return self._text | ||||
Sean Farley
|
r19679 | def files(self): | ||
Martin von Zweigbergk
|
r22916 | return sorted(self._status.modified + self._status.added + | ||
self._status.removed) | ||||
Sean Farley
|
r19675 | |||
Sean Farley
|
r19680 | def modified(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.modified | ||
Sean Farley
|
r19681 | def added(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.added | ||
Sean Farley
|
r19682 | def removed(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.removed | ||
Sean Farley
|
r19683 | def deleted(self): | ||
Martin von Zweigbergk
|
r22916 | return self._status.deleted | ||
Sean Farley
|
r19687 | def branch(self): | ||
return encoding.tolocal(self._extra['branch']) | ||||
Sean Farley
|
r19688 | def closesbranch(self): | ||
return 'close' in self._extra | ||||
Sean Farley
|
r19689 | def extra(self): | ||
return self._extra | ||||
Sean Farley
|
r19680 | |||
Sean Farley
|
r19690 | def tags(self): | ||
t = [] | ||||
for p in self.parents(): | ||||
t.extend(p.tags()) | ||||
return t | ||||
Sean Farley
|
r19691 | def bookmarks(self): | ||
b = [] | ||||
for p in self.parents(): | ||||
b.extend(p.bookmarks()) | ||||
return b | ||||
Sean Farley
|
r19692 | def phase(self): | ||
phase = phases.draft # default phase to draft | ||||
for p in self.parents(): | ||||
phase = max(phase, p.phase()) | ||||
return phase | ||||
Sean Farley
|
r19693 | def hidden(self): | ||
return False | ||||
Sean Farley
|
r19694 | def children(self): | ||
return [] | ||||
Sean Farley
|
r19695 | def flags(self, path): | ||
if '_manifest' in self.__dict__: | ||||
try: | ||||
return self._manifest.flags(path) | ||||
except KeyError: | ||||
return '' | ||||
try: | ||||
return self._flagfunc(path) | ||||
except OSError: | ||||
return '' | ||||
Sean Farley
|
r19696 | def ancestor(self, c2): | ||
Mads Kiilerich
|
r22389 | """return the "best" ancestor context of self and c2""" | ||
Sean Farley
|
r19696 | return self._parents[0].ancestor(c2) # punt on two parents for now | ||
Sean Farley
|
r19697 | def walk(self, match): | ||
return sorted(self._repo.dirstate.walk(match, sorted(self.substate), | ||||
True, False)) | ||||
Siddharth Agarwal
|
r21985 | def matches(self, match): | ||
return sorted(self._repo.dirstate.matches(match)) | ||||
Sean Farley
|
r19698 | def ancestors(self): | ||
Durham Goode
|
r23616 | for p in self._parents: | ||
yield p | ||||
Sean Farley
|
r19698 | for a in self._repo.changelog.ancestors( | ||
[p.rev() for p in self._parents]): | ||||
yield changectx(self._repo, a) | ||||
Sean Farley
|
r19699 | def markcommitted(self, node): | ||
"""Perform post-commit cleanup necessary after committing this ctx | ||||
Specifically, this updates backing stores this working context | ||||
wraps to reflect the fact that the changes reflected by this | ||||
workingctx have been committed. For example, it marks | ||||
modified and added files as normal in the dirstate. | ||||
""" | ||||
Durham Goode
|
r22405 | self._repo.dirstate.beginparentchange() | ||
Sean Farley
|
r19699 | for f in self.modified() + self.added(): | ||
self._repo.dirstate.normal(f) | ||||
for f in self.removed(): | ||||
self._repo.dirstate.drop(f) | ||||
self._repo.dirstate.setparents(node) | ||||
Durham Goode
|
r22405 | self._repo.dirstate.endparentchange() | ||
Sean Farley
|
r19699 | |||
Sean Farley
|
r19733 | class workingctx(committablectx): | ||
Sean Farley
|
r19671 | """A workingctx object makes access to data related to | ||
the current working directory convenient. | ||||
date - any valid date string or (unixtime, offset), or None. | ||||
user - username string, or None. | ||||
extra - a dictionary of extra values, or None. | ||||
changes - a list of file lists as returned by localrepo.status() | ||||
or None to use the repository status. | ||||
""" | ||||
def __init__(self, repo, text="", user=None, date=None, extra=None, | ||||
changes=None): | ||||
super(workingctx, self).__init__(repo, text, user, date, extra, changes) | ||||
Matt Mackall
|
r14129 | def __iter__(self): | ||
d = self._repo.dirstate | ||||
for f in d: | ||||
if d[f] != 'r': | ||||
yield f | ||||
Sean Farley
|
r21845 | def __contains__(self, key): | ||
return self._repo.dirstate[key] not in "?r" | ||||
Martin Geisler
|
r8157 | @propertycache | ||
Dirkjan Ochtman
|
r7368 | def _parents(self): | ||
p = self._repo.dirstate.parents() | ||||
if p[1] == nullid: | ||||
p = p[:-1] | ||||
Patrick Mezard
|
r17330 | return [changectx(self._repo, x) for x in p] | ||
Matt Mackall
|
r3217 | |||
Benoit Boissinot
|
r3966 | def filectx(self, path, filelog=None): | ||
Matt Mackall
|
r3217 | """get a file context from the working directory""" | ||
Benoit Boissinot
|
r3966 | return workingfilectx(self._repo, path, workingctx=self, | ||
filelog=filelog) | ||||
Matt Mackall
|
r3217 | |||
Patrick Mezard
|
r16491 | def dirty(self, missing=False, merge=True, branch=True): | ||
Matt Mackall
|
r8717 | "check whether a working directory is modified" | ||
Edouard Gomez
|
r11110 | # check subrepos first | ||
Mads Kiilerich
|
r18364 | for s in sorted(self.substate): | ||
Edouard Gomez
|
r11110 | if self.sub(s).dirty(): | ||
return True | ||||
# check current working dir | ||||
Patrick Mezard
|
r16491 | return ((merge and self.p2()) or | ||
(branch and self.branch() != self.p1().branch()) or | ||||
Matt Mackall
|
r8717 | self.modified() or self.added() or self.removed() or | ||
(missing and self.deleted())) | ||||
Martin Geisler
|
r12270 | def add(self, list, prefix=""): | ||
join = lambda f: os.path.join(prefix, f) | ||||
Dirkjan Ochtman
|
r11303 | wlock = self._repo.wlock() | ||
ui, ds = self._repo.ui, self._repo.dirstate | ||||
try: | ||||
rejected = [] | ||||
FUJIWARA Katsunori
|
r19900 | lstat = self._repo.wvfs.lstat | ||
Dirkjan Ochtman
|
r11303 | for f in list: | ||
Adrian Buehlmann
|
r13962 | scmutil.checkportable(ui, join(f)) | ||
Dirkjan Ochtman
|
r11303 | try: | ||
FUJIWARA Katsunori
|
r19900 | st = lstat(f) | ||
Idan Kamara
|
r14004 | except OSError: | ||
Martin Geisler
|
r12270 | ui.warn(_("%s does not exist!\n") % join(f)) | ||
Dirkjan Ochtman
|
r11303 | rejected.append(f) | ||
continue | ||||
if st.st_size > 10000000: | ||||
ui.warn(_("%s: up to %d MB of RAM may be required " | ||||
"to manage this file\n" | ||||
"(use 'hg revert %s' to cancel the " | ||||
"pending addition)\n") | ||||
Martin Geisler
|
r12270 | % (f, 3 * st.st_size // 1000000, join(f))) | ||
Dirkjan Ochtman
|
r11303 | if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): | ||
ui.warn(_("%s not added: only files and symlinks " | ||||
Martin Geisler
|
r12270 | "supported currently\n") % join(f)) | ||
FUJIWARA Katsunori
|
r19900 | rejected.append(f) | ||
Dirkjan Ochtman
|
r11303 | elif ds[f] in 'amn': | ||
Martin Geisler
|
r12270 | ui.warn(_("%s already tracked!\n") % join(f)) | ||
Dirkjan Ochtman
|
r11303 | elif ds[f] == 'r': | ||
ds.normallookup(f) | ||||
else: | ||||
ds.add(f) | ||||
return rejected | ||||
finally: | ||||
wlock.release() | ||||
David M. Carr
|
r15912 | def forget(self, files, prefix=""): | ||
join = lambda f: os.path.join(prefix, f) | ||||
Dirkjan Ochtman
|
r11303 | wlock = self._repo.wlock() | ||
try: | ||||
David M. Carr
|
r15912 | rejected = [] | ||
Matt Mackall
|
r14435 | for f in files: | ||
Patrick Mezard
|
r16111 | if f not in self._repo.dirstate: | ||
David M. Carr
|
r15912 | self._repo.ui.warn(_("%s not tracked!\n") % join(f)) | ||
rejected.append(f) | ||||
Patrick Mezard
|
r16111 | elif self._repo.dirstate[f] != 'a': | ||
self._repo.dirstate.remove(f) | ||||
Dirkjan Ochtman
|
r11303 | else: | ||
Matt Mackall
|
r14434 | self._repo.dirstate.drop(f) | ||
David M. Carr
|
r15912 | return rejected | ||
Dirkjan Ochtman
|
r11303 | finally: | ||
wlock.release() | ||||
def undelete(self, list): | ||||
pctxs = self.parents() | ||||
wlock = self._repo.wlock() | ||||
try: | ||||
for f in list: | ||||
if self._repo.dirstate[f] != 'r': | ||||
self._repo.ui.warn(_("%s not removed!\n") % f) | ||||
else: | ||||
Patrick Mezard
|
r12360 | fctx = f in pctxs[0] and pctxs[0][f] or pctxs[1][f] | ||
Dirkjan Ochtman
|
r11303 | t = fctx.data() | ||
self._repo.wwrite(f, t, fctx.flags()) | ||||
self._repo.dirstate.normal(f) | ||||
finally: | ||||
wlock.release() | ||||
def copy(self, source, dest): | ||||
FUJIWARA Katsunori
|
r19902 | try: | ||
st = self._repo.wvfs.lstat(dest) | ||||
except OSError, err: | ||||
if err.errno != errno.ENOENT: | ||||
raise | ||||
Dirkjan Ochtman
|
r11303 | self._repo.ui.warn(_("%s does not exist!\n") % dest) | ||
FUJIWARA Katsunori
|
r19902 | return | ||
if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): | ||||
Dirkjan Ochtman
|
r11303 | self._repo.ui.warn(_("copy failed: %s is not a file or a " | ||
"symbolic link\n") % dest) | ||||
else: | ||||
wlock = self._repo.wlock() | ||||
try: | ||||
Pierre-Yves David
|
r23402 | if self._repo.dirstate[dest] in '?': | ||
Dirkjan Ochtman
|
r11303 | self._repo.dirstate.add(dest) | ||
Pierre-Yves David
|
r23402 | elif self._repo.dirstate[dest] in 'r': | ||
self._repo.dirstate.normallookup(dest) | ||||
Dirkjan Ochtman
|
r11303 | self._repo.dirstate.copy(source, dest) | ||
finally: | ||||
wlock.release() | ||||
Sean Farley
|
r21393 | def _filtersuspectsymlink(self, files): | ||
if not files or self._repo.dirstate._checklink: | ||||
return files | ||||
# Symlink placeholders may get non-symlink-like contents | ||||
# via user error or dereferencing by NFS or Samba servers, | ||||
# so we filter out any placeholders that don't look like a | ||||
# symlink | ||||
sane = [] | ||||
for f in files: | ||||
if self.flags(f) == 'l': | ||||
d = self[f].data() | ||||
if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d): | ||||
self._repo.ui.debug('ignoring suspect symlink placeholder' | ||||
' "%s"\n' % f) | ||||
continue | ||||
sane.append(f) | ||||
return sane | ||||
Sean Farley
|
r21395 | def _checklookup(self, files): | ||
# check for any possibly clean files | ||||
if not files: | ||||
return [], [] | ||||
modified = [] | ||||
fixup = [] | ||||
pctx = self._parents[0] | ||||
# do a full compare of any files that might have changed | ||||
for f in sorted(files): | ||||
if (f not in pctx or self.flags(f) != pctx.flags(f) | ||||
or pctx[f].cmp(self[f])): | ||||
modified.append(f) | ||||
else: | ||||
fixup.append(f) | ||||
# update dirstate for files that are actually clean | ||||
if fixup: | ||||
try: | ||||
# updating the dirstate is optional | ||||
# so we don't wait on the lock | ||||
Siddharth Agarwal
|
r21990 | # wlock can invalidate the dirstate, so cache normal _after_ | ||
# taking the lock | ||||
wlock = self._repo.wlock(False) | ||||
Sean Farley
|
r21395 | normal = self._repo.dirstate.normal | ||
try: | ||||
for f in fixup: | ||||
normal(f) | ||||
finally: | ||||
wlock.release() | ||||
except error.LockError: | ||||
pass | ||||
return modified, fixup | ||||
Sean Farley
|
r21468 | def _manifestmatches(self, match, s): | ||
"""Slow path for workingctx | ||||
The fast path is when we compare the working directory to its parent | ||||
which means this function is comparing with a non-parent; therefore we | ||||
need to build a manifest and return what matches. | ||||
""" | ||||
mf = self._repo['.']._manifestmatches(match, s) | ||||
Martin von Zweigbergk
|
r23304 | for f in s.modified + s.added: | ||
Augie Fackler
|
r23593 | mf[f] = _newnode | ||
Augie Fackler
|
r22942 | mf.setflag(f, self.flags(f)) | ||
Martin von Zweigbergk
|
r23304 | for f in s.removed: | ||
Sean Farley
|
r21468 | if f in mf: | ||
del mf[f] | ||||
return mf | ||||
Sean Farley
|
r21397 | def _dirstatestatus(self, match=None, ignored=False, clean=False, | ||
unknown=False): | ||||
'''Gets the status from the dirstate -- internal use only.''' | ||||
listignored, listclean, listunknown = ignored, clean, unknown | ||||
match = match or matchmod.always(self._repo.root, self._repo.getcwd()) | ||||
subrepos = [] | ||||
if '.hgsub' in self: | ||||
subrepos = sorted(self.substate) | ||||
Martin von Zweigbergk
|
r22911 | cmp, s = self._repo.dirstate.status(match, subrepos, listignored, | ||
listclean, listunknown) | ||||
Sean Farley
|
r21397 | |||
# check for any possibly clean files | ||||
if cmp: | ||||
modified2, fixup = self._checklookup(cmp) | ||||
Martin von Zweigbergk
|
r23303 | s.modified.extend(modified2) | ||
Sean Farley
|
r21397 | |||
# update dirstate for files that are actually clean | ||||
if fixup and listclean: | ||||
Martin von Zweigbergk
|
r23303 | s.clean.extend(fixup) | ||
Sean Farley
|
r21397 | |||
Martin von Zweigbergk
|
r23776 | if match.always(): | ||
# cache for performance | ||||
if s.unknown or s.ignored or s.clean: | ||||
# "_status" is cached with list*=False in the normal route | ||||
self._status = scmutil.status(s.modified, s.added, s.removed, | ||||
s.deleted, [], [], []) | ||||
else: | ||||
self._status = s | ||||
Martin von Zweigbergk
|
r23303 | return s | ||
Sean Farley
|
r21397 | |||
Sean Farley
|
r21480 | def _buildstatus(self, other, s, match, listignored, listclean, | ||
Sean Farley
|
r21663 | listunknown): | ||
Sean Farley
|
r21480 | """build a status with respect to another context | ||
This includes logic for maintaining the fast path of status when | ||||
comparing the working directory against its parent, which is to skip | ||||
building a new manifest if self (working directory) is not comparing | ||||
against its parent (repo['.']). | ||||
""" | ||||
Martin von Zweigbergk
|
r23239 | s = self._dirstatestatus(match, listignored, listclean, listunknown) | ||
Mads Kiilerich
|
r23543 | # Filter out symlinks that, in the case of FAT32 and NTFS filesystems, | ||
Martin von Zweigbergk
|
r23242 | # might have accidentally ended up with the entire contents of the file | ||
Mads Kiilerich
|
r23543 | # they are supposed to be linking to. | ||
Martin von Zweigbergk
|
r23302 | s.modified[:] = self._filtersuspectsymlink(s.modified) | ||
Sean Farley
|
r21480 | if other != self._repo['.']: | ||
s = super(workingctx, self)._buildstatus(other, s, match, | ||||
listignored, listclean, | ||||
listunknown) | ||||
return s | ||||
Martin von Zweigbergk
|
r23237 | def _matchstatus(self, other, match): | ||
Sean Farley
|
r21482 | """override the match method with a filter for directory patterns | ||
We use inheritance to customize the match.bad method only in cases of | ||||
workingctx since it belongs only to the working directory when | ||||
comparing against the parent changeset. | ||||
If we aren't comparing against the working directory's parent, then we | ||||
just use the default match object sent to us. | ||||
""" | ||||
superself = super(workingctx, self) | ||||
Martin von Zweigbergk
|
r23237 | match = superself._matchstatus(other, match) | ||
Sean Farley
|
r21482 | if other != self._repo['.']: | ||
def bad(f, msg): | ||||
# 'f' may be a directory pattern from 'match.files()', | ||||
# so 'f not in ctx1' is not enough | ||||
Drew Gottlieb
|
r24326 | if f not in other and not other.hasdir(f): | ||
Sean Farley
|
r21482 | self._repo.ui.warn('%s: %s\n' % | ||
(self._repo.dirstate.pathto(f), msg)) | ||||
match.bad = bad | ||||
return match | ||||
Sean Farley
|
r19733 | class committablefilectx(basefilectx): | ||
"""A committablefilectx provides common functionality for a file context | ||||
that wants the ability to commit, e.g. workingfilectx or memfilectx.""" | ||||
Sean Farley
|
r19701 | def __init__(self, repo, path, filelog=None, ctx=None): | ||
Matt Mackall
|
r3217 | self._repo = repo | ||
self._path = path | ||||
self._changeid = None | ||||
self._filerev = self._filenode = None | ||||
Durham Goode
|
r19149 | if filelog is not None: | ||
Matt Mackall
|
r3217 | self._filelog = filelog | ||
Sean Farley
|
r19702 | if ctx: | ||
self._changectx = ctx | ||||
Sean Farley
|
r19703 | def __nonzero__(self): | ||
return True | ||||
Matt Mackall
|
r3217 | def parents(self): | ||
'''return parent filectxs, following copies if necessary''' | ||||
Benoit Boissinot
|
r8528 | def filenode(ctx, path): | ||
return ctx._manifest.get(path, nullid) | ||||
path = self._path | ||||
Matt Mackall
|
r3217 | fl = self._filelog | ||
Benoit Boissinot
|
r8528 | pcl = self._changectx._parents | ||
renamed = self.renamed() | ||||
if renamed: | ||||
pl = [renamed + (None,)] | ||||
else: | ||||
pl = [(path, filenode(pcl[0], path), fl)] | ||||
for pc in pcl[1:]: | ||||
pl.append((path, filenode(pc, path), fl)) | ||||
Matt Mackall
|
r3217 | |||
Thomas Arendsen Hein
|
r3673 | return [filectx(self._repo, p, fileid=n, filelog=l) | ||
Matt Mackall
|
r10282 | for p, n, l in pl if n != nullid] | ||
Matt Mackall
|
r3217 | |||
Sean Farley
|
r19705 | def children(self): | ||
return [] | ||||
Sean Farley
|
r19733 | class workingfilectx(committablefilectx): | ||
Sean Farley
|
r19704 | """A workingfilectx object makes access to data related to a particular | ||
file in the working directory convenient.""" | ||||
def __init__(self, repo, path, filelog=None, workingctx=None): | ||||
super(workingfilectx, self).__init__(repo, path, filelog, workingctx) | ||||
@propertycache | ||||
def _changectx(self): | ||||
return workingctx(self._repo) | ||||
def data(self): | ||||
return self._repo.wread(self._path) | ||||
def renamed(self): | ||||
rp = self._repo.dirstate.copied(self._path) | ||||
if not rp: | ||||
return None | ||||
return rp, self._changectx._parents[0]._manifest.get(rp, nullid) | ||||
Matt Mackall
|
r10282 | def size(self): | ||
FUJIWARA Katsunori
|
r19901 | return self._repo.wvfs.lstat(self._path).st_size | ||
Benoit Boissinot
|
r3962 | def date(self): | ||
t, tz = self._changectx.date() | ||||
try: | ||||
FUJIWARA Katsunori
|
r19901 | return (int(self._repo.wvfs.lstat(self._path).st_mtime), tz) | ||
Benoit Boissinot
|
r3962 | except OSError, err: | ||
Matt Mackall
|
r10282 | if err.errno != errno.ENOENT: | ||
raise | ||||
Benoit Boissinot
|
r3962 | return (t, tz) | ||
Matt Mackall
|
r3310 | |||
Nicolas Dumazet
|
r11702 | def cmp(self, fctx): | ||
"""compare with other file context | ||||
Nicolas Dumazet
|
r11539 | |||
Nicolas Dumazet
|
r11702 | returns True if different than fctx. | ||
Nicolas Dumazet
|
r11539 | """ | ||
Mads Kiilerich
|
r17425 | # fctx should be a filectx (not a workingfilectx) | ||
Nicolas Dumazet
|
r11703 | # invert comparison to reuse the same code path | ||
return fctx.cmp(self) | ||||
Patrick Mezard
|
r6715 | |||
Sean Farley
|
r22073 | def remove(self, ignoremissing=False): | ||
"""wraps unlink for a repo's working directory""" | ||||
util.unlinkpath(self._repo.wjoin(self._path), ignoremissing) | ||||
def write(self, data, flags): | ||||
"""wraps repo.wwrite""" | ||||
self._repo.wwrite(self._path, data, flags) | ||||
FUJIWARA Katsunori
|
r23710 | class workingcommitctx(workingctx): | ||
"""A workingcommitctx object makes access to data related to | ||||
the revision being committed convenient. | ||||
This hides changes in the working directory, if they aren't | ||||
committed in this context. | ||||
""" | ||||
def __init__(self, repo, changes, | ||||
text="", user=None, date=None, extra=None): | ||||
super(workingctx, self).__init__(repo, text, user, date, extra, | ||||
changes) | ||||
FUJIWARA Katsunori
|
r23712 | def _dirstatestatus(self, match=None, ignored=False, clean=False, | ||
unknown=False): | ||||
"""Return matched files only in ``self._status`` | ||||
Uncommitted files appear "clean" via this context, even if | ||||
they aren't actually so in the working directory. | ||||
""" | ||||
match = match or matchmod.always(self._repo.root, self._repo.getcwd()) | ||||
if clean: | ||||
clean = [f for f in self._manifest if f not in self._changedset] | ||||
else: | ||||
clean = [] | ||||
return scmutil.status([f for f in self._status.modified if match(f)], | ||||
[f for f in self._status.added if match(f)], | ||||
[f for f in self._status.removed if match(f)], | ||||
[], [], [], clean) | ||||
@propertycache | ||||
def _changedset(self): | ||||
"""Return the set of files changed in this context | ||||
""" | ||||
changed = set(self._status.modified) | ||||
changed.update(self._status.added) | ||||
changed.update(self._status.removed) | ||||
return changed | ||||
Sean Farley
|
r21665 | class memctx(committablectx): | ||
Patrick Mezard
|
r7077 | """Use memctx to perform in-memory commits via localrepo.commitctx(). | ||
Patrick Mezard
|
r6715 | |||
Patrick Mezard
|
r7077 | Revision information is supplied at initialization time while | ||
related files data and is made available through a callback | ||||
mechanism. 'repo' is the current localrepo, 'parents' is a | ||||
sequence of two parent revisions identifiers (pass None for every | ||||
missing parent), 'text' is the commit message and 'files' lists | ||||
names of files touched by the revision (normalized and relative to | ||||
repository root). | ||||
Patrick Mezard
|
r6715 | |||
Patrick Mezard
|
r7077 | filectxfn(repo, memctx, path) is a callable receiving the | ||
repository, the current memctx object and the normalized path of | ||||
requested file, relative to repository root. It is fired by the | ||||
commit function for every file in 'files', but calls order is | ||||
undefined. If the file is available in the revision being | ||||
committed (updated or added), filectxfn returns a memfilectx | ||||
object. If the file was removed, filectxfn raises an | ||||
IOError. Moved files are represented by marking the source file | ||||
removed and the new file added with copy information (see | ||||
memfilectx). | ||||
user receives the committer name and defaults to current | ||||
repository username, date is the commit date in any format | ||||
supported by util.parsedate() and defaults to current date, extra | ||||
is a dictionary of metadata or is left empty. | ||||
Patrick Mezard
|
r6715 | """ | ||
Siddharth Agarwal
|
r22313 | |||
# Mercurial <= 3.1 expects the filectxfn to raise IOError for missing files. | ||||
# Extensions that need to retain compatibility across Mercurial 3.1 can use | ||||
# this field to determine what to do in filectxfn. | ||||
_returnnoneformissingfiles = True | ||||
Dirkjan Ochtman
|
r6721 | def __init__(self, repo, parents, text, files, filectxfn, user=None, | ||
FUJIWARA Katsunori
|
r21238 | date=None, extra=None, editor=False): | ||
Sean Farley
|
r21666 | super(memctx, self).__init__(repo, text, user, date, extra) | ||
Patrick Mezard
|
r6715 | self._rev = None | ||
self._node = None | ||||
parents = [(p or nullid) for p in parents] | ||||
p1, p2 = parents | ||||
Matt Mackall
|
r6747 | self._parents = [changectx(self._repo, p) for p in (p1, p2)] | ||
Matt Mackall
|
r8209 | files = sorted(set(files)) | ||
FUJIWARA Katsunori
|
r23587 | self._files = files | ||
Sean Farley
|
r21938 | self.substate = {} | ||
Patrick Mezard
|
r6715 | |||
Sean Farley
|
r22072 | # if store is not callable, wrap it in a function | ||
if not callable(filectxfn): | ||||
def getfilectx(repo, memctx, path): | ||||
fctx = filectxfn[path] | ||||
# this is weird but apparently we only keep track of one parent | ||||
# (why not only store that instead of a tuple?) | ||||
copied = fctx.renamed() | ||||
if copied: | ||||
copied = copied[0] | ||||
return memfilectx(repo, path, fctx.data(), | ||||
islink=fctx.islink(), isexec=fctx.isexec(), | ||||
copied=copied, memctx=memctx) | ||||
self._filectxfn = getfilectx | ||||
FUJIWARA Katsunori
|
r23587 | else: | ||
# "util.cachefunc" reduces invocation of possibly expensive | ||||
# "filectxfn" for performance (e.g. converting from another VCS) | ||||
self._filectxfn = util.cachefunc(filectxfn) | ||||
Sean Farley
|
r22072 | |||
Jordi Gutiérrez Hermoso
|
r24306 | if extra: | ||
self._extra = extra.copy() | ||||
else: | ||||
self._extra = {} | ||||
Patrick Mezard
|
r14528 | if self._extra.get('branch', '') == '': | ||
Patrick Mezard
|
r6715 | self._extra['branch'] = 'default' | ||
FUJIWARA Katsunori
|
r21238 | if editor: | ||
self._text = editor(self._repo, self, []) | ||||
self._repo.savecommitmessage(self._text) | ||||
Patrick Mezard
|
r6715 | def filectx(self, path, filelog=None): | ||
Mads Kiilerich
|
r22296 | """get a file context from the working directory | ||
Returns None if file doesn't exist and should be removed.""" | ||||
Patrick Mezard
|
r6715 | return self._filectxfn(self._repo, self, path) | ||
Alexander Solovyov
|
r11151 | def commit(self): | ||
"""commit context to the repo""" | ||||
return self._repo.commitctx(self) | ||||
Sean Farley
|
r21835 | @propertycache | ||
def _manifest(self): | ||||
"""generate a manifest based on the return values of filectxfn""" | ||||
# keep this simple for now; just worry about p1 | ||||
pctx = self._parents[0] | ||||
man = pctx.manifest().copy() | ||||
FUJIWARA Katsunori
|
r23603 | for f in self._status.modified: | ||
Sean Farley
|
r21835 | p1node = nullid | ||
p2node = nullid | ||||
Sean Farley
|
r22075 | p = pctx[f].parents() # if file isn't in pctx, check p2? | ||
Sean Farley
|
r21835 | if len(p) > 0: | ||
p1node = p[0].node() | ||||
if len(p) > 1: | ||||
p2node = p[1].node() | ||||
FUJIWARA Katsunori
|
r23603 | man[f] = revlog.hash(self[f].data(), p1node, p2node) | ||
Sean Farley
|
r21835 | |||
FUJIWARA Katsunori
|
r23588 | for f in self._status.added: | ||
man[f] = revlog.hash(self[f].data(), nullid, nullid) | ||||
FUJIWARA Katsunori
|
r23589 | for f in self._status.removed: | ||
if f in man: | ||||
del man[f] | ||||
Sean Farley
|
r21835 | |||
return man | ||||
FUJIWARA Katsunori
|
r23587 | @propertycache | ||
def _status(self): | ||||
"""Calculate exact status from ``files`` specified at construction | ||||
""" | ||||
man1 = self.p1().manifest() | ||||
p2 = self._parents[1] | ||||
# "1 < len(self._parents)" can't be used for checking | ||||
# existence of the 2nd parent, because "memctx._parents" is | ||||
# explicitly initialized by the list, of which length is 2. | ||||
if p2.node() != nullid: | ||||
man2 = p2.manifest() | ||||
managing = lambda f: f in man1 or f in man2 | ||||
else: | ||||
managing = lambda f: f in man1 | ||||
modified, added, removed = [], [], [] | ||||
for f in self._files: | ||||
if not managing(f): | ||||
added.append(f) | ||||
elif self[f]: | ||||
modified.append(f) | ||||
else: | ||||
removed.append(f) | ||||
return scmutil.status(modified, added, removed, [], [], [], []) | ||||
Sean Farley
|
r21835 | |||
Sean Farley
|
r21688 | class memfilectx(committablefilectx): | ||
Patrick Mezard
|
r7077 | """memfilectx represents an in-memory file to commit. | ||
Mads Kiilerich
|
r23139 | See memctx and committablefilectx for more details. | ||
Patrick Mezard
|
r6715 | """ | ||
Sean Farley
|
r21689 | def __init__(self, repo, path, data, islink=False, | ||
isexec=False, copied=None, memctx=None): | ||||
Patrick Mezard
|
r7077 | """ | ||
path is the normalized file path relative to repository root. | ||||
data is the file content as a string. | ||||
islink is True if the file is a symbolic link. | ||||
isexec is True if the file is executable. | ||||
copied is the source file path if current file was copied in the | ||||
revision being committed, or None.""" | ||||
Sean Farley
|
r21689 | super(memfilectx, self).__init__(repo, path, None, memctx) | ||
Patrick Mezard
|
r6715 | self._data = data | ||
self._flags = (islink and 'l' or '') + (isexec and 'x' or '') | ||||
self._copied = None | ||||
if copied: | ||||
self._copied = (copied, nullid) | ||||
Matt Mackall
|
r10282 | def data(self): | ||
return self._data | ||||
Sean Farley
|
r21710 | def size(self): | ||
return len(self.data()) | ||||
Matt Mackall
|
r10282 | def flags(self): | ||
return self._flags | ||||
def renamed(self): | ||||
return self._copied | ||||
Sean Farley
|
r22074 | |||
def remove(self, ignoremissing=False): | ||||
"""wraps unlink for a repo's working directory""" | ||||
# need to figure out what to do here | ||||
del self._changectx[self._path] | ||||
def write(self, data, flags): | ||||
"""wraps repo.wwrite""" | ||||
self._data = data | ||||