##// END OF EJS Templates
dirstate: change placeholder hash length to 20 bytes...
Durham Goode -
r30360:0298a07f default
parent child Browse files
Show More
@@ -1,1989 +1,1985
1 1 # context.py - changeset and file context objects for mercurial
2 2 #
3 3 # Copyright 2006, 2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import errno
11 11 import os
12 12 import re
13 13 import stat
14 14
15 15 from .i18n import _
16 16 from .node import (
17 17 bin,
18 18 hex,
19 newnodeid,
19 20 nullid,
20 21 nullrev,
21 22 short,
22 23 wdirid,
23 24 )
24 25 from . import (
25 26 encoding,
26 27 error,
27 28 fileset,
28 29 match as matchmod,
29 30 mdiff,
30 31 obsolete as obsmod,
31 32 patch,
32 33 phases,
33 34 repoview,
34 35 revlog,
35 36 scmutil,
36 37 subrepo,
37 38 util,
38 39 )
39 40
40 41 propertycache = util.propertycache
41 42
42 # Phony node value to stand-in for new files in some uses of
43 # manifests. Manifests support 21-byte hashes for nodes which are
44 # dirty in the working copy.
45 _newnode = '!' * 21
46
47 43 nonascii = re.compile(r'[^\x21-\x7f]').search
48 44
49 45 class basectx(object):
50 46 """A basectx object represents the common logic for its children:
51 47 changectx: read-only context that is already present in the repo,
52 48 workingctx: a context that represents the working directory and can
53 49 be committed,
54 50 memctx: a context that represents changes in-memory and can also
55 51 be committed."""
56 52 def __new__(cls, repo, changeid='', *args, **kwargs):
57 53 if isinstance(changeid, basectx):
58 54 return changeid
59 55
60 56 o = super(basectx, cls).__new__(cls)
61 57
62 58 o._repo = repo
63 59 o._rev = nullrev
64 60 o._node = nullid
65 61
66 62 return o
67 63
68 64 def __str__(self):
69 65 return short(self.node())
70 66
71 67 def __int__(self):
72 68 return self.rev()
73 69
74 70 def __repr__(self):
75 71 return "<%s %s>" % (type(self).__name__, str(self))
76 72
77 73 def __eq__(self, other):
78 74 try:
79 75 return type(self) == type(other) and self._rev == other._rev
80 76 except AttributeError:
81 77 return False
82 78
83 79 def __ne__(self, other):
84 80 return not (self == other)
85 81
86 82 def __contains__(self, key):
87 83 return key in self._manifest
88 84
89 85 def __getitem__(self, key):
90 86 return self.filectx(key)
91 87
92 88 def __iter__(self):
93 89 return iter(self._manifest)
94 90
95 91 def _manifestmatches(self, match, s):
96 92 """generate a new manifest filtered by the match argument
97 93
98 94 This method is for internal use only and mainly exists to provide an
99 95 object oriented way for other contexts to customize the manifest
100 96 generation.
101 97 """
102 98 return self.manifest().matches(match)
103 99
104 100 def _matchstatus(self, other, match):
105 101 """return match.always if match is none
106 102
107 103 This internal method provides a way for child objects to override the
108 104 match operator.
109 105 """
110 106 return match or matchmod.always(self._repo.root, self._repo.getcwd())
111 107
112 108 def _buildstatus(self, other, s, match, listignored, listclean,
113 109 listunknown):
114 110 """build a status with respect to another context"""
115 111 # Load earliest manifest first for caching reasons. More specifically,
116 112 # if you have revisions 1000 and 1001, 1001 is probably stored as a
117 113 # delta against 1000. Thus, if you read 1000 first, we'll reconstruct
118 114 # 1000 and cache it so that when you read 1001, we just need to apply a
119 115 # delta to what's in the cache. So that's one full reconstruction + one
120 116 # delta application.
121 117 if self.rev() is not None and self.rev() < other.rev():
122 118 self.manifest()
123 119 mf1 = other._manifestmatches(match, s)
124 120 mf2 = self._manifestmatches(match, s)
125 121
126 122 modified, added = [], []
127 123 removed = []
128 124 clean = []
129 125 deleted, unknown, ignored = s.deleted, s.unknown, s.ignored
130 126 deletedset = set(deleted)
131 127 d = mf1.diff(mf2, clean=listclean)
132 128 for fn, value in d.iteritems():
133 129 if fn in deletedset:
134 130 continue
135 131 if value is None:
136 132 clean.append(fn)
137 133 continue
138 134 (node1, flag1), (node2, flag2) = value
139 135 if node1 is None:
140 136 added.append(fn)
141 137 elif node2 is None:
142 138 removed.append(fn)
143 139 elif flag1 != flag2:
144 140 modified.append(fn)
145 elif node2 != _newnode:
141 elif node2 != newnodeid:
146 142 # When comparing files between two commits, we save time by
147 143 # not comparing the file contents when the nodeids differ.
148 144 # Note that this means we incorrectly report a reverted change
149 145 # to a file as a modification.
150 146 modified.append(fn)
151 147 elif self[fn].cmp(other[fn]):
152 148 modified.append(fn)
153 149 else:
154 150 clean.append(fn)
155 151
156 152 if removed:
157 153 # need to filter files if they are already reported as removed
158 154 unknown = [fn for fn in unknown if fn not in mf1]
159 155 ignored = [fn for fn in ignored if fn not in mf1]
160 156 # if they're deleted, don't report them as removed
161 157 removed = [fn for fn in removed if fn not in deletedset]
162 158
163 159 return scmutil.status(modified, added, removed, deleted, unknown,
164 160 ignored, clean)
165 161
166 162 @propertycache
167 163 def substate(self):
168 164 return subrepo.state(self, self._repo.ui)
169 165
170 166 def subrev(self, subpath):
171 167 return self.substate[subpath][1]
172 168
173 169 def rev(self):
174 170 return self._rev
175 171 def node(self):
176 172 return self._node
177 173 def hex(self):
178 174 return hex(self.node())
179 175 def manifest(self):
180 176 return self._manifest
181 177 def manifestctx(self):
182 178 return self._manifestctx
183 179 def repo(self):
184 180 return self._repo
185 181 def phasestr(self):
186 182 return phases.phasenames[self.phase()]
187 183 def mutable(self):
188 184 return self.phase() > phases.public
189 185
190 186 def getfileset(self, expr):
191 187 return fileset.getfileset(self, expr)
192 188
193 189 def obsolete(self):
194 190 """True if the changeset is obsolete"""
195 191 return self.rev() in obsmod.getrevs(self._repo, 'obsolete')
196 192
197 193 def extinct(self):
198 194 """True if the changeset is extinct"""
199 195 return self.rev() in obsmod.getrevs(self._repo, 'extinct')
200 196
201 197 def unstable(self):
202 198 """True if the changeset is not obsolete but it's ancestor are"""
203 199 return self.rev() in obsmod.getrevs(self._repo, 'unstable')
204 200
205 201 def bumped(self):
206 202 """True if the changeset try to be a successor of a public changeset
207 203
208 204 Only non-public and non-obsolete changesets may be bumped.
209 205 """
210 206 return self.rev() in obsmod.getrevs(self._repo, 'bumped')
211 207
212 208 def divergent(self):
213 209 """Is a successors of a changeset with multiple possible successors set
214 210
215 211 Only non-public and non-obsolete changesets may be divergent.
216 212 """
217 213 return self.rev() in obsmod.getrevs(self._repo, 'divergent')
218 214
219 215 def troubled(self):
220 216 """True if the changeset is either unstable, bumped or divergent"""
221 217 return self.unstable() or self.bumped() or self.divergent()
222 218
223 219 def troubles(self):
224 220 """return the list of troubles affecting this changesets.
225 221
226 222 Troubles are returned as strings. possible values are:
227 223 - unstable,
228 224 - bumped,
229 225 - divergent.
230 226 """
231 227 troubles = []
232 228 if self.unstable():
233 229 troubles.append('unstable')
234 230 if self.bumped():
235 231 troubles.append('bumped')
236 232 if self.divergent():
237 233 troubles.append('divergent')
238 234 return troubles
239 235
240 236 def parents(self):
241 237 """return contexts for each parent changeset"""
242 238 return self._parents
243 239
244 240 def p1(self):
245 241 return self._parents[0]
246 242
247 243 def p2(self):
248 244 parents = self._parents
249 245 if len(parents) == 2:
250 246 return parents[1]
251 247 return changectx(self._repo, nullrev)
252 248
253 249 def _fileinfo(self, path):
254 250 if '_manifest' in self.__dict__:
255 251 try:
256 252 return self._manifest[path], self._manifest.flags(path)
257 253 except KeyError:
258 254 raise error.ManifestLookupError(self._node, path,
259 255 _('not found in manifest'))
260 256 if '_manifestdelta' in self.__dict__ or path in self.files():
261 257 if path in self._manifestdelta:
262 258 return (self._manifestdelta[path],
263 259 self._manifestdelta.flags(path))
264 260 mfl = self._repo.manifestlog
265 261 try:
266 262 node, flag = mfl[self._changeset.manifest].find(path)
267 263 except KeyError:
268 264 raise error.ManifestLookupError(self._node, path,
269 265 _('not found in manifest'))
270 266
271 267 return node, flag
272 268
273 269 def filenode(self, path):
274 270 return self._fileinfo(path)[0]
275 271
276 272 def flags(self, path):
277 273 try:
278 274 return self._fileinfo(path)[1]
279 275 except error.LookupError:
280 276 return ''
281 277
282 278 def sub(self, path, allowcreate=True):
283 279 '''return a subrepo for the stored revision of path, never wdir()'''
284 280 return subrepo.subrepo(self, path, allowcreate=allowcreate)
285 281
286 282 def nullsub(self, path, pctx):
287 283 return subrepo.nullsubrepo(self, path, pctx)
288 284
289 285 def workingsub(self, path):
290 286 '''return a subrepo for the stored revision, or wdir if this is a wdir
291 287 context.
292 288 '''
293 289 return subrepo.subrepo(self, path, allowwdir=True)
294 290
295 291 def match(self, pats=[], include=None, exclude=None, default='glob',
296 292 listsubrepos=False, badfn=None):
297 293 r = self._repo
298 294 return matchmod.match(r.root, r.getcwd(), pats,
299 295 include, exclude, default,
300 296 auditor=r.nofsauditor, ctx=self,
301 297 listsubrepos=listsubrepos, badfn=badfn)
302 298
303 299 def diff(self, ctx2=None, match=None, **opts):
304 300 """Returns a diff generator for the given contexts and matcher"""
305 301 if ctx2 is None:
306 302 ctx2 = self.p1()
307 303 if ctx2 is not None:
308 304 ctx2 = self._repo[ctx2]
309 305 diffopts = patch.diffopts(self._repo.ui, opts)
310 306 return patch.diff(self._repo, ctx2, self, match=match, opts=diffopts)
311 307
312 308 def dirs(self):
313 309 return self._manifest.dirs()
314 310
315 311 def hasdir(self, dir):
316 312 return self._manifest.hasdir(dir)
317 313
318 314 def dirty(self, missing=False, merge=True, branch=True):
319 315 return False
320 316
321 317 def status(self, other=None, match=None, listignored=False,
322 318 listclean=False, listunknown=False, listsubrepos=False):
323 319 """return status of files between two nodes or node and working
324 320 directory.
325 321
326 322 If other is None, compare this node with working directory.
327 323
328 324 returns (modified, added, removed, deleted, unknown, ignored, clean)
329 325 """
330 326
331 327 ctx1 = self
332 328 ctx2 = self._repo[other]
333 329
334 330 # This next code block is, admittedly, fragile logic that tests for
335 331 # reversing the contexts and wouldn't need to exist if it weren't for
336 332 # the fast (and common) code path of comparing the working directory
337 333 # with its first parent.
338 334 #
339 335 # What we're aiming for here is the ability to call:
340 336 #
341 337 # workingctx.status(parentctx)
342 338 #
343 339 # If we always built the manifest for each context and compared those,
344 340 # then we'd be done. But the special case of the above call means we
345 341 # just copy the manifest of the parent.
346 342 reversed = False
347 343 if (not isinstance(ctx1, changectx)
348 344 and isinstance(ctx2, changectx)):
349 345 reversed = True
350 346 ctx1, ctx2 = ctx2, ctx1
351 347
352 348 match = ctx2._matchstatus(ctx1, match)
353 349 r = scmutil.status([], [], [], [], [], [], [])
354 350 r = ctx2._buildstatus(ctx1, r, match, listignored, listclean,
355 351 listunknown)
356 352
357 353 if reversed:
358 354 # Reverse added and removed. Clear deleted, unknown and ignored as
359 355 # these make no sense to reverse.
360 356 r = scmutil.status(r.modified, r.removed, r.added, [], [], [],
361 357 r.clean)
362 358
363 359 if listsubrepos:
364 360 for subpath, sub in scmutil.itersubrepos(ctx1, ctx2):
365 361 try:
366 362 rev2 = ctx2.subrev(subpath)
367 363 except KeyError:
368 364 # A subrepo that existed in node1 was deleted between
369 365 # node1 and node2 (inclusive). Thus, ctx2's substate
370 366 # won't contain that subpath. The best we can do ignore it.
371 367 rev2 = None
372 368 submatch = matchmod.subdirmatcher(subpath, match)
373 369 s = sub.status(rev2, match=submatch, ignored=listignored,
374 370 clean=listclean, unknown=listunknown,
375 371 listsubrepos=True)
376 372 for rfiles, sfiles in zip(r, s):
377 373 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
378 374
379 375 for l in r:
380 376 l.sort()
381 377
382 378 return r
383 379
384 380
385 381 def makememctx(repo, parents, text, user, date, branch, files, store,
386 382 editor=None, extra=None):
387 383 def getfilectx(repo, memctx, path):
388 384 data, mode, copied = store.getfile(path)
389 385 if data is None:
390 386 return None
391 387 islink, isexec = mode
392 388 return memfilectx(repo, path, data, islink=islink, isexec=isexec,
393 389 copied=copied, memctx=memctx)
394 390 if extra is None:
395 391 extra = {}
396 392 if branch:
397 393 extra['branch'] = encoding.fromlocal(branch)
398 394 ctx = memctx(repo, parents, text, files, getfilectx, user,
399 395 date, extra, editor)
400 396 return ctx
401 397
402 398 class changectx(basectx):
403 399 """A changecontext object makes access to data related to a particular
404 400 changeset convenient. It represents a read-only context already present in
405 401 the repo."""
406 402 def __init__(self, repo, changeid=''):
407 403 """changeid is a revision number, node, or tag"""
408 404
409 405 # since basectx.__new__ already took care of copying the object, we
410 406 # don't need to do anything in __init__, so we just exit here
411 407 if isinstance(changeid, basectx):
412 408 return
413 409
414 410 if changeid == '':
415 411 changeid = '.'
416 412 self._repo = repo
417 413
418 414 try:
419 415 if isinstance(changeid, int):
420 416 self._node = repo.changelog.node(changeid)
421 417 self._rev = changeid
422 418 return
423 419 if isinstance(changeid, long):
424 420 changeid = str(changeid)
425 421 if changeid == 'null':
426 422 self._node = nullid
427 423 self._rev = nullrev
428 424 return
429 425 if changeid == 'tip':
430 426 self._node = repo.changelog.tip()
431 427 self._rev = repo.changelog.rev(self._node)
432 428 return
433 429 if changeid == '.' or changeid == repo.dirstate.p1():
434 430 # this is a hack to delay/avoid loading obsmarkers
435 431 # when we know that '.' won't be hidden
436 432 self._node = repo.dirstate.p1()
437 433 self._rev = repo.unfiltered().changelog.rev(self._node)
438 434 return
439 435 if len(changeid) == 20:
440 436 try:
441 437 self._node = changeid
442 438 self._rev = repo.changelog.rev(changeid)
443 439 return
444 440 except error.FilteredRepoLookupError:
445 441 raise
446 442 except LookupError:
447 443 pass
448 444
449 445 try:
450 446 r = int(changeid)
451 447 if str(r) != changeid:
452 448 raise ValueError
453 449 l = len(repo.changelog)
454 450 if r < 0:
455 451 r += l
456 452 if r < 0 or r >= l:
457 453 raise ValueError
458 454 self._rev = r
459 455 self._node = repo.changelog.node(r)
460 456 return
461 457 except error.FilteredIndexError:
462 458 raise
463 459 except (ValueError, OverflowError, IndexError):
464 460 pass
465 461
466 462 if len(changeid) == 40:
467 463 try:
468 464 self._node = bin(changeid)
469 465 self._rev = repo.changelog.rev(self._node)
470 466 return
471 467 except error.FilteredLookupError:
472 468 raise
473 469 except (TypeError, LookupError):
474 470 pass
475 471
476 472 # lookup bookmarks through the name interface
477 473 try:
478 474 self._node = repo.names.singlenode(repo, changeid)
479 475 self._rev = repo.changelog.rev(self._node)
480 476 return
481 477 except KeyError:
482 478 pass
483 479 except error.FilteredRepoLookupError:
484 480 raise
485 481 except error.RepoLookupError:
486 482 pass
487 483
488 484 self._node = repo.unfiltered().changelog._partialmatch(changeid)
489 485 if self._node is not None:
490 486 self._rev = repo.changelog.rev(self._node)
491 487 return
492 488
493 489 # lookup failed
494 490 # check if it might have come from damaged dirstate
495 491 #
496 492 # XXX we could avoid the unfiltered if we had a recognizable
497 493 # exception for filtered changeset access
498 494 if changeid in repo.unfiltered().dirstate.parents():
499 495 msg = _("working directory has unknown parent '%s'!")
500 496 raise error.Abort(msg % short(changeid))
501 497 try:
502 498 if len(changeid) == 20 and nonascii(changeid):
503 499 changeid = hex(changeid)
504 500 except TypeError:
505 501 pass
506 502 except (error.FilteredIndexError, error.FilteredLookupError,
507 503 error.FilteredRepoLookupError):
508 504 if repo.filtername.startswith('visible'):
509 505 msg = _("hidden revision '%s'") % changeid
510 506 hint = _('use --hidden to access hidden revisions')
511 507 raise error.FilteredRepoLookupError(msg, hint=hint)
512 508 msg = _("filtered revision '%s' (not in '%s' subset)")
513 509 msg %= (changeid, repo.filtername)
514 510 raise error.FilteredRepoLookupError(msg)
515 511 except IndexError:
516 512 pass
517 513 raise error.RepoLookupError(
518 514 _("unknown revision '%s'") % changeid)
519 515
520 516 def __hash__(self):
521 517 try:
522 518 return hash(self._rev)
523 519 except AttributeError:
524 520 return id(self)
525 521
526 522 def __nonzero__(self):
527 523 return self._rev != nullrev
528 524
529 525 @propertycache
530 526 def _changeset(self):
531 527 return self._repo.changelog.changelogrevision(self.rev())
532 528
533 529 @propertycache
534 530 def _manifest(self):
535 531 return self._manifestctx.read()
536 532
537 533 @propertycache
538 534 def _manifestctx(self):
539 535 return self._repo.manifestlog[self._changeset.manifest]
540 536
541 537 @propertycache
542 538 def _manifestdelta(self):
543 539 return self._manifestctx.readdelta()
544 540
545 541 @propertycache
546 542 def _parents(self):
547 543 repo = self._repo
548 544 p1, p2 = repo.changelog.parentrevs(self._rev)
549 545 if p2 == nullrev:
550 546 return [changectx(repo, p1)]
551 547 return [changectx(repo, p1), changectx(repo, p2)]
552 548
553 549 def changeset(self):
554 550 c = self._changeset
555 551 return (
556 552 c.manifest,
557 553 c.user,
558 554 c.date,
559 555 c.files,
560 556 c.description,
561 557 c.extra,
562 558 )
563 559 def manifestnode(self):
564 560 return self._changeset.manifest
565 561
566 562 def user(self):
567 563 return self._changeset.user
568 564 def date(self):
569 565 return self._changeset.date
570 566 def files(self):
571 567 return self._changeset.files
572 568 def description(self):
573 569 return self._changeset.description
574 570 def branch(self):
575 571 return encoding.tolocal(self._changeset.extra.get("branch"))
576 572 def closesbranch(self):
577 573 return 'close' in self._changeset.extra
578 574 def extra(self):
579 575 return self._changeset.extra
580 576 def tags(self):
581 577 return self._repo.nodetags(self._node)
582 578 def bookmarks(self):
583 579 return self._repo.nodebookmarks(self._node)
584 580 def phase(self):
585 581 return self._repo._phasecache.phase(self._repo, self._rev)
586 582 def hidden(self):
587 583 return self._rev in repoview.filterrevs(self._repo, 'visible')
588 584
589 585 def children(self):
590 586 """return contexts for each child changeset"""
591 587 c = self._repo.changelog.children(self._node)
592 588 return [changectx(self._repo, x) for x in c]
593 589
594 590 def ancestors(self):
595 591 for a in self._repo.changelog.ancestors([self._rev]):
596 592 yield changectx(self._repo, a)
597 593
598 594 def descendants(self):
599 595 for d in self._repo.changelog.descendants([self._rev]):
600 596 yield changectx(self._repo, d)
601 597
602 598 def filectx(self, path, fileid=None, filelog=None):
603 599 """get a file context from this changeset"""
604 600 if fileid is None:
605 601 fileid = self.filenode(path)
606 602 return filectx(self._repo, path, fileid=fileid,
607 603 changectx=self, filelog=filelog)
608 604
609 605 def ancestor(self, c2, warn=False):
610 606 """return the "best" ancestor context of self and c2
611 607
612 608 If there are multiple candidates, it will show a message and check
613 609 merge.preferancestor configuration before falling back to the
614 610 revlog ancestor."""
615 611 # deal with workingctxs
616 612 n2 = c2._node
617 613 if n2 is None:
618 614 n2 = c2._parents[0]._node
619 615 cahs = self._repo.changelog.commonancestorsheads(self._node, n2)
620 616 if not cahs:
621 617 anc = nullid
622 618 elif len(cahs) == 1:
623 619 anc = cahs[0]
624 620 else:
625 621 # experimental config: merge.preferancestor
626 622 for r in self._repo.ui.configlist('merge', 'preferancestor', ['*']):
627 623 try:
628 624 ctx = changectx(self._repo, r)
629 625 except error.RepoLookupError:
630 626 continue
631 627 anc = ctx.node()
632 628 if anc in cahs:
633 629 break
634 630 else:
635 631 anc = self._repo.changelog.ancestor(self._node, n2)
636 632 if warn:
637 633 self._repo.ui.status(
638 634 (_("note: using %s as ancestor of %s and %s\n") %
639 635 (short(anc), short(self._node), short(n2))) +
640 636 ''.join(_(" alternatively, use --config "
641 637 "merge.preferancestor=%s\n") %
642 638 short(n) for n in sorted(cahs) if n != anc))
643 639 return changectx(self._repo, anc)
644 640
645 641 def descendant(self, other):
646 642 """True if other is descendant of this changeset"""
647 643 return self._repo.changelog.descendant(self._rev, other._rev)
648 644
649 645 def walk(self, match):
650 646 '''Generates matching file names.'''
651 647
652 648 # Wrap match.bad method to have message with nodeid
653 649 def bad(fn, msg):
654 650 # The manifest doesn't know about subrepos, so don't complain about
655 651 # paths into valid subrepos.
656 652 if any(fn == s or fn.startswith(s + '/')
657 653 for s in self.substate):
658 654 return
659 655 match.bad(fn, _('no such file in rev %s') % self)
660 656
661 657 m = matchmod.badmatch(match, bad)
662 658 return self._manifest.walk(m)
663 659
664 660 def matches(self, match):
665 661 return self.walk(match)
666 662
667 663 class basefilectx(object):
668 664 """A filecontext object represents the common logic for its children:
669 665 filectx: read-only access to a filerevision that is already present
670 666 in the repo,
671 667 workingfilectx: a filecontext that represents files from the working
672 668 directory,
673 669 memfilectx: a filecontext that represents files in-memory."""
674 670 def __new__(cls, repo, path, *args, **kwargs):
675 671 return super(basefilectx, cls).__new__(cls)
676 672
677 673 @propertycache
678 674 def _filelog(self):
679 675 return self._repo.file(self._path)
680 676
681 677 @propertycache
682 678 def _changeid(self):
683 679 if '_changeid' in self.__dict__:
684 680 return self._changeid
685 681 elif '_changectx' in self.__dict__:
686 682 return self._changectx.rev()
687 683 elif '_descendantrev' in self.__dict__:
688 684 # this file context was created from a revision with a known
689 685 # descendant, we can (lazily) correct for linkrev aliases
690 686 return self._adjustlinkrev(self._descendantrev)
691 687 else:
692 688 return self._filelog.linkrev(self._filerev)
693 689
694 690 @propertycache
695 691 def _filenode(self):
696 692 if '_fileid' in self.__dict__:
697 693 return self._filelog.lookup(self._fileid)
698 694 else:
699 695 return self._changectx.filenode(self._path)
700 696
701 697 @propertycache
702 698 def _filerev(self):
703 699 return self._filelog.rev(self._filenode)
704 700
705 701 @propertycache
706 702 def _repopath(self):
707 703 return self._path
708 704
709 705 def __nonzero__(self):
710 706 try:
711 707 self._filenode
712 708 return True
713 709 except error.LookupError:
714 710 # file is missing
715 711 return False
716 712
717 713 def __str__(self):
718 714 try:
719 715 return "%s@%s" % (self.path(), self._changectx)
720 716 except error.LookupError:
721 717 return "%s@???" % self.path()
722 718
723 719 def __repr__(self):
724 720 return "<%s %s>" % (type(self).__name__, str(self))
725 721
726 722 def __hash__(self):
727 723 try:
728 724 return hash((self._path, self._filenode))
729 725 except AttributeError:
730 726 return id(self)
731 727
732 728 def __eq__(self, other):
733 729 try:
734 730 return (type(self) == type(other) and self._path == other._path
735 731 and self._filenode == other._filenode)
736 732 except AttributeError:
737 733 return False
738 734
739 735 def __ne__(self, other):
740 736 return not (self == other)
741 737
742 738 def filerev(self):
743 739 return self._filerev
744 740 def filenode(self):
745 741 return self._filenode
746 742 def flags(self):
747 743 return self._changectx.flags(self._path)
748 744 def filelog(self):
749 745 return self._filelog
750 746 def rev(self):
751 747 return self._changeid
752 748 def linkrev(self):
753 749 return self._filelog.linkrev(self._filerev)
754 750 def node(self):
755 751 return self._changectx.node()
756 752 def hex(self):
757 753 return self._changectx.hex()
758 754 def user(self):
759 755 return self._changectx.user()
760 756 def date(self):
761 757 return self._changectx.date()
762 758 def files(self):
763 759 return self._changectx.files()
764 760 def description(self):
765 761 return self._changectx.description()
766 762 def branch(self):
767 763 return self._changectx.branch()
768 764 def extra(self):
769 765 return self._changectx.extra()
770 766 def phase(self):
771 767 return self._changectx.phase()
772 768 def phasestr(self):
773 769 return self._changectx.phasestr()
774 770 def manifest(self):
775 771 return self._changectx.manifest()
776 772 def changectx(self):
777 773 return self._changectx
778 774 def repo(self):
779 775 return self._repo
780 776
781 777 def path(self):
782 778 return self._path
783 779
784 780 def isbinary(self):
785 781 try:
786 782 return util.binary(self.data())
787 783 except IOError:
788 784 return False
789 785 def isexec(self):
790 786 return 'x' in self.flags()
791 787 def islink(self):
792 788 return 'l' in self.flags()
793 789
794 790 def isabsent(self):
795 791 """whether this filectx represents a file not in self._changectx
796 792
797 793 This is mainly for merge code to detect change/delete conflicts. This is
798 794 expected to be True for all subclasses of basectx."""
799 795 return False
800 796
801 797 _customcmp = False
802 798 def cmp(self, fctx):
803 799 """compare with other file context
804 800
805 801 returns True if different than fctx.
806 802 """
807 803 if fctx._customcmp:
808 804 return fctx.cmp(self)
809 805
810 806 if (fctx._filenode is None
811 807 and (self._repo._encodefilterpats
812 808 # if file data starts with '\1\n', empty metadata block is
813 809 # prepended, which adds 4 bytes to filelog.size().
814 810 or self.size() - 4 == fctx.size())
815 811 or self.size() == fctx.size()):
816 812 return self._filelog.cmp(self._filenode, fctx.data())
817 813
818 814 return True
819 815
820 816 def _adjustlinkrev(self, srcrev, inclusive=False):
821 817 """return the first ancestor of <srcrev> introducing <fnode>
822 818
823 819 If the linkrev of the file revision does not point to an ancestor of
824 820 srcrev, we'll walk down the ancestors until we find one introducing
825 821 this file revision.
826 822
827 823 :srcrev: the changeset revision we search ancestors from
828 824 :inclusive: if true, the src revision will also be checked
829 825 """
830 826 repo = self._repo
831 827 cl = repo.unfiltered().changelog
832 828 mfl = repo.manifestlog
833 829 # fetch the linkrev
834 830 lkr = self.linkrev()
835 831 # hack to reuse ancestor computation when searching for renames
836 832 memberanc = getattr(self, '_ancestrycontext', None)
837 833 iteranc = None
838 834 if srcrev is None:
839 835 # wctx case, used by workingfilectx during mergecopy
840 836 revs = [p.rev() for p in self._repo[None].parents()]
841 837 inclusive = True # we skipped the real (revless) source
842 838 else:
843 839 revs = [srcrev]
844 840 if memberanc is None:
845 841 memberanc = iteranc = cl.ancestors(revs, lkr,
846 842 inclusive=inclusive)
847 843 # check if this linkrev is an ancestor of srcrev
848 844 if lkr not in memberanc:
849 845 if iteranc is None:
850 846 iteranc = cl.ancestors(revs, lkr, inclusive=inclusive)
851 847 fnode = self._filenode
852 848 path = self._path
853 849 for a in iteranc:
854 850 ac = cl.read(a) # get changeset data (we avoid object creation)
855 851 if path in ac[3]: # checking the 'files' field.
856 852 # The file has been touched, check if the content is
857 853 # similar to the one we search for.
858 854 if fnode == mfl[ac[0]].readfast().get(path):
859 855 return a
860 856 # In theory, we should never get out of that loop without a result.
861 857 # But if manifest uses a buggy file revision (not children of the
862 858 # one it replaces) we could. Such a buggy situation will likely
863 859 # result is crash somewhere else at to some point.
864 860 return lkr
865 861
866 862 def introrev(self):
867 863 """return the rev of the changeset which introduced this file revision
868 864
869 865 This method is different from linkrev because it take into account the
870 866 changeset the filectx was created from. It ensures the returned
871 867 revision is one of its ancestors. This prevents bugs from
872 868 'linkrev-shadowing' when a file revision is used by multiple
873 869 changesets.
874 870 """
875 871 lkr = self.linkrev()
876 872 attrs = vars(self)
877 873 noctx = not ('_changeid' in attrs or '_changectx' in attrs)
878 874 if noctx or self.rev() == lkr:
879 875 return self.linkrev()
880 876 return self._adjustlinkrev(self.rev(), inclusive=True)
881 877
882 878 def _parentfilectx(self, path, fileid, filelog):
883 879 """create parent filectx keeping ancestry info for _adjustlinkrev()"""
884 880 fctx = filectx(self._repo, path, fileid=fileid, filelog=filelog)
885 881 if '_changeid' in vars(self) or '_changectx' in vars(self):
886 882 # If self is associated with a changeset (probably explicitly
887 883 # fed), ensure the created filectx is associated with a
888 884 # changeset that is an ancestor of self.changectx.
889 885 # This lets us later use _adjustlinkrev to get a correct link.
890 886 fctx._descendantrev = self.rev()
891 887 fctx._ancestrycontext = getattr(self, '_ancestrycontext', None)
892 888 elif '_descendantrev' in vars(self):
893 889 # Otherwise propagate _descendantrev if we have one associated.
894 890 fctx._descendantrev = self._descendantrev
895 891 fctx._ancestrycontext = getattr(self, '_ancestrycontext', None)
896 892 return fctx
897 893
898 894 def parents(self):
899 895 _path = self._path
900 896 fl = self._filelog
901 897 parents = self._filelog.parents(self._filenode)
902 898 pl = [(_path, node, fl) for node in parents if node != nullid]
903 899
904 900 r = fl.renamed(self._filenode)
905 901 if r:
906 902 # - In the simple rename case, both parent are nullid, pl is empty.
907 903 # - In case of merge, only one of the parent is null id and should
908 904 # be replaced with the rename information. This parent is -always-
909 905 # the first one.
910 906 #
911 907 # As null id have always been filtered out in the previous list
912 908 # comprehension, inserting to 0 will always result in "replacing
913 909 # first nullid parent with rename information.
914 910 pl.insert(0, (r[0], r[1], self._repo.file(r[0])))
915 911
916 912 return [self._parentfilectx(path, fnode, l) for path, fnode, l in pl]
917 913
918 914 def p1(self):
919 915 return self.parents()[0]
920 916
921 917 def p2(self):
922 918 p = self.parents()
923 919 if len(p) == 2:
924 920 return p[1]
925 921 return filectx(self._repo, self._path, fileid=-1, filelog=self._filelog)
926 922
927 923 def annotate(self, follow=False, linenumber=False, diffopts=None):
928 924 '''returns a list of tuples of ((ctx, number), line) for each line
929 925 in the file, where ctx is the filectx of the node where
930 926 that line was last changed; if linenumber parameter is true, number is
931 927 the line number at the first appearance in the managed file, otherwise,
932 928 number has a fixed value of False.
933 929 '''
934 930
935 931 def lines(text):
936 932 if text.endswith("\n"):
937 933 return text.count("\n")
938 934 return text.count("\n") + int(bool(text))
939 935
940 936 if linenumber:
941 937 def decorate(text, rev):
942 938 return ([(rev, i) for i in xrange(1, lines(text) + 1)], text)
943 939 else:
944 940 def decorate(text, rev):
945 941 return ([(rev, False)] * lines(text), text)
946 942
947 943 def pair(parent, child):
948 944 blocks = mdiff.allblocks(parent[1], child[1], opts=diffopts)
949 945 for (a1, a2, b1, b2), t in blocks:
950 946 # Changed blocks ('!') or blocks made only of blank lines ('~')
951 947 # belong to the child.
952 948 if t == '=':
953 949 child[0][b1:b2] = parent[0][a1:a2]
954 950 return child
955 951
956 952 getlog = util.lrucachefunc(lambda x: self._repo.file(x))
957 953
958 954 def parents(f):
959 955 # Cut _descendantrev here to mitigate the penalty of lazy linkrev
960 956 # adjustment. Otherwise, p._adjustlinkrev() would walk changelog
961 957 # from the topmost introrev (= srcrev) down to p.linkrev() if it
962 958 # isn't an ancestor of the srcrev.
963 959 f._changeid
964 960 pl = f.parents()
965 961
966 962 # Don't return renamed parents if we aren't following.
967 963 if not follow:
968 964 pl = [p for p in pl if p.path() == f.path()]
969 965
970 966 # renamed filectx won't have a filelog yet, so set it
971 967 # from the cache to save time
972 968 for p in pl:
973 969 if not '_filelog' in p.__dict__:
974 970 p._filelog = getlog(p.path())
975 971
976 972 return pl
977 973
978 974 # use linkrev to find the first changeset where self appeared
979 975 base = self
980 976 introrev = self.introrev()
981 977 if self.rev() != introrev:
982 978 base = self.filectx(self.filenode(), changeid=introrev)
983 979 if getattr(base, '_ancestrycontext', None) is None:
984 980 cl = self._repo.changelog
985 981 if introrev is None:
986 982 # wctx is not inclusive, but works because _ancestrycontext
987 983 # is used to test filelog revisions
988 984 ac = cl.ancestors([p.rev() for p in base.parents()],
989 985 inclusive=True)
990 986 else:
991 987 ac = cl.ancestors([introrev], inclusive=True)
992 988 base._ancestrycontext = ac
993 989
994 990 # This algorithm would prefer to be recursive, but Python is a
995 991 # bit recursion-hostile. Instead we do an iterative
996 992 # depth-first search.
997 993
998 994 # 1st DFS pre-calculates pcache and needed
999 995 visit = [base]
1000 996 pcache = {}
1001 997 needed = {base: 1}
1002 998 while visit:
1003 999 f = visit.pop()
1004 1000 if f in pcache:
1005 1001 continue
1006 1002 pl = parents(f)
1007 1003 pcache[f] = pl
1008 1004 for p in pl:
1009 1005 needed[p] = needed.get(p, 0) + 1
1010 1006 if p not in pcache:
1011 1007 visit.append(p)
1012 1008
1013 1009 # 2nd DFS does the actual annotate
1014 1010 visit[:] = [base]
1015 1011 hist = {}
1016 1012 while visit:
1017 1013 f = visit[-1]
1018 1014 if f in hist:
1019 1015 visit.pop()
1020 1016 continue
1021 1017
1022 1018 ready = True
1023 1019 pl = pcache[f]
1024 1020 for p in pl:
1025 1021 if p not in hist:
1026 1022 ready = False
1027 1023 visit.append(p)
1028 1024 if ready:
1029 1025 visit.pop()
1030 1026 curr = decorate(f.data(), f)
1031 1027 for p in pl:
1032 1028 curr = pair(hist[p], curr)
1033 1029 if needed[p] == 1:
1034 1030 del hist[p]
1035 1031 del needed[p]
1036 1032 else:
1037 1033 needed[p] -= 1
1038 1034
1039 1035 hist[f] = curr
1040 1036 del pcache[f]
1041 1037
1042 1038 return zip(hist[base][0], hist[base][1].splitlines(True))
1043 1039
1044 1040 def ancestors(self, followfirst=False):
1045 1041 visit = {}
1046 1042 c = self
1047 1043 if followfirst:
1048 1044 cut = 1
1049 1045 else:
1050 1046 cut = None
1051 1047
1052 1048 while True:
1053 1049 for parent in c.parents()[:cut]:
1054 1050 visit[(parent.linkrev(), parent.filenode())] = parent
1055 1051 if not visit:
1056 1052 break
1057 1053 c = visit.pop(max(visit))
1058 1054 yield c
1059 1055
1060 1056 class filectx(basefilectx):
1061 1057 """A filecontext object makes access to data related to a particular
1062 1058 filerevision convenient."""
1063 1059 def __init__(self, repo, path, changeid=None, fileid=None,
1064 1060 filelog=None, changectx=None):
1065 1061 """changeid can be a changeset revision, node, or tag.
1066 1062 fileid can be a file revision or node."""
1067 1063 self._repo = repo
1068 1064 self._path = path
1069 1065
1070 1066 assert (changeid is not None
1071 1067 or fileid is not None
1072 1068 or changectx is not None), \
1073 1069 ("bad args: changeid=%r, fileid=%r, changectx=%r"
1074 1070 % (changeid, fileid, changectx))
1075 1071
1076 1072 if filelog is not None:
1077 1073 self._filelog = filelog
1078 1074
1079 1075 if changeid is not None:
1080 1076 self._changeid = changeid
1081 1077 if changectx is not None:
1082 1078 self._changectx = changectx
1083 1079 if fileid is not None:
1084 1080 self._fileid = fileid
1085 1081
1086 1082 @propertycache
1087 1083 def _changectx(self):
1088 1084 try:
1089 1085 return changectx(self._repo, self._changeid)
1090 1086 except error.FilteredRepoLookupError:
1091 1087 # Linkrev may point to any revision in the repository. When the
1092 1088 # repository is filtered this may lead to `filectx` trying to build
1093 1089 # `changectx` for filtered revision. In such case we fallback to
1094 1090 # creating `changectx` on the unfiltered version of the reposition.
1095 1091 # This fallback should not be an issue because `changectx` from
1096 1092 # `filectx` are not used in complex operations that care about
1097 1093 # filtering.
1098 1094 #
1099 1095 # This fallback is a cheap and dirty fix that prevent several
1100 1096 # crashes. It does not ensure the behavior is correct. However the
1101 1097 # behavior was not correct before filtering either and "incorrect
1102 1098 # behavior" is seen as better as "crash"
1103 1099 #
1104 1100 # Linkrevs have several serious troubles with filtering that are
1105 1101 # complicated to solve. Proper handling of the issue here should be
1106 1102 # considered when solving linkrev issue are on the table.
1107 1103 return changectx(self._repo.unfiltered(), self._changeid)
1108 1104
1109 1105 def filectx(self, fileid, changeid=None):
1110 1106 '''opens an arbitrary revision of the file without
1111 1107 opening a new filelog'''
1112 1108 return filectx(self._repo, self._path, fileid=fileid,
1113 1109 filelog=self._filelog, changeid=changeid)
1114 1110
1115 1111 def data(self):
1116 1112 try:
1117 1113 return self._filelog.read(self._filenode)
1118 1114 except error.CensoredNodeError:
1119 1115 if self._repo.ui.config("censor", "policy", "abort") == "ignore":
1120 1116 return ""
1121 1117 raise error.Abort(_("censored node: %s") % short(self._filenode),
1122 1118 hint=_("set censor.policy to ignore errors"))
1123 1119
1124 1120 def size(self):
1125 1121 return self._filelog.size(self._filerev)
1126 1122
1127 1123 def renamed(self):
1128 1124 """check if file was actually renamed in this changeset revision
1129 1125
1130 1126 If rename logged in file revision, we report copy for changeset only
1131 1127 if file revisions linkrev points back to the changeset in question
1132 1128 or both changeset parents contain different file revisions.
1133 1129 """
1134 1130
1135 1131 renamed = self._filelog.renamed(self._filenode)
1136 1132 if not renamed:
1137 1133 return renamed
1138 1134
1139 1135 if self.rev() == self.linkrev():
1140 1136 return renamed
1141 1137
1142 1138 name = self.path()
1143 1139 fnode = self._filenode
1144 1140 for p in self._changectx.parents():
1145 1141 try:
1146 1142 if fnode == p.filenode(name):
1147 1143 return None
1148 1144 except error.LookupError:
1149 1145 pass
1150 1146 return renamed
1151 1147
1152 1148 def children(self):
1153 1149 # hard for renames
1154 1150 c = self._filelog.children(self._filenode)
1155 1151 return [filectx(self._repo, self._path, fileid=x,
1156 1152 filelog=self._filelog) for x in c]
1157 1153
1158 1154 class committablectx(basectx):
1159 1155 """A committablectx object provides common functionality for a context that
1160 1156 wants the ability to commit, e.g. workingctx or memctx."""
1161 1157 def __init__(self, repo, text="", user=None, date=None, extra=None,
1162 1158 changes=None):
1163 1159 self._repo = repo
1164 1160 self._rev = None
1165 1161 self._node = None
1166 1162 self._text = text
1167 1163 if date:
1168 1164 self._date = util.parsedate(date)
1169 1165 if user:
1170 1166 self._user = user
1171 1167 if changes:
1172 1168 self._status = changes
1173 1169
1174 1170 self._extra = {}
1175 1171 if extra:
1176 1172 self._extra = extra.copy()
1177 1173 if 'branch' not in self._extra:
1178 1174 try:
1179 1175 branch = encoding.fromlocal(self._repo.dirstate.branch())
1180 1176 except UnicodeDecodeError:
1181 1177 raise error.Abort(_('branch name not in UTF-8!'))
1182 1178 self._extra['branch'] = branch
1183 1179 if self._extra['branch'] == '':
1184 1180 self._extra['branch'] = 'default'
1185 1181
1186 1182 def __str__(self):
1187 1183 return str(self._parents[0]) + "+"
1188 1184
1189 1185 def __nonzero__(self):
1190 1186 return True
1191 1187
1192 1188 def _buildflagfunc(self):
1193 1189 # Create a fallback function for getting file flags when the
1194 1190 # filesystem doesn't support them
1195 1191
1196 1192 copiesget = self._repo.dirstate.copies().get
1197 1193 parents = self.parents()
1198 1194 if len(parents) < 2:
1199 1195 # when we have one parent, it's easy: copy from parent
1200 1196 man = parents[0].manifest()
1201 1197 def func(f):
1202 1198 f = copiesget(f, f)
1203 1199 return man.flags(f)
1204 1200 else:
1205 1201 # merges are tricky: we try to reconstruct the unstored
1206 1202 # result from the merge (issue1802)
1207 1203 p1, p2 = parents
1208 1204 pa = p1.ancestor(p2)
1209 1205 m1, m2, ma = p1.manifest(), p2.manifest(), pa.manifest()
1210 1206
1211 1207 def func(f):
1212 1208 f = copiesget(f, f) # may be wrong for merges with copies
1213 1209 fl1, fl2, fla = m1.flags(f), m2.flags(f), ma.flags(f)
1214 1210 if fl1 == fl2:
1215 1211 return fl1
1216 1212 if fl1 == fla:
1217 1213 return fl2
1218 1214 if fl2 == fla:
1219 1215 return fl1
1220 1216 return '' # punt for conflicts
1221 1217
1222 1218 return func
1223 1219
1224 1220 @propertycache
1225 1221 def _flagfunc(self):
1226 1222 return self._repo.dirstate.flagfunc(self._buildflagfunc)
1227 1223
1228 1224 @propertycache
1229 1225 def _manifest(self):
1230 1226 """generate a manifest corresponding to the values in self._status
1231 1227
1232 1228 This reuse the file nodeid from parent, but we append an extra letter
1233 1229 when modified. Modified files get an extra 'm' while added files get
1234 1230 an extra 'a'. This is used by manifests merge to see that files
1235 1231 are different and by update logic to avoid deleting newly added files.
1236 1232 """
1237 1233 parents = self.parents()
1238 1234
1239 1235 man1 = parents[0].manifest()
1240 1236 man = man1.copy()
1241 1237 if len(parents) > 1:
1242 1238 man2 = self.p2().manifest()
1243 1239 def getman(f):
1244 1240 if f in man1:
1245 1241 return man1
1246 1242 return man2
1247 1243 else:
1248 1244 getman = lambda f: man1
1249 1245
1250 1246 copied = self._repo.dirstate.copies()
1251 1247 ff = self._flagfunc
1252 1248 for i, l in (("a", self._status.added), ("m", self._status.modified)):
1253 1249 for f in l:
1254 1250 orig = copied.get(f, f)
1255 1251 man[f] = getman(orig).get(orig, nullid) + i
1256 1252 try:
1257 1253 man.setflag(f, ff(f))
1258 1254 except OSError:
1259 1255 pass
1260 1256
1261 1257 for f in self._status.deleted + self._status.removed:
1262 1258 if f in man:
1263 1259 del man[f]
1264 1260
1265 1261 return man
1266 1262
1267 1263 @propertycache
1268 1264 def _status(self):
1269 1265 return self._repo.status()
1270 1266
1271 1267 @propertycache
1272 1268 def _user(self):
1273 1269 return self._repo.ui.username()
1274 1270
1275 1271 @propertycache
1276 1272 def _date(self):
1277 1273 return util.makedate()
1278 1274
1279 1275 def subrev(self, subpath):
1280 1276 return None
1281 1277
1282 1278 def manifestnode(self):
1283 1279 return None
1284 1280 def user(self):
1285 1281 return self._user or self._repo.ui.username()
1286 1282 def date(self):
1287 1283 return self._date
1288 1284 def description(self):
1289 1285 return self._text
1290 1286 def files(self):
1291 1287 return sorted(self._status.modified + self._status.added +
1292 1288 self._status.removed)
1293 1289
1294 1290 def modified(self):
1295 1291 return self._status.modified
1296 1292 def added(self):
1297 1293 return self._status.added
1298 1294 def removed(self):
1299 1295 return self._status.removed
1300 1296 def deleted(self):
1301 1297 return self._status.deleted
1302 1298 def branch(self):
1303 1299 return encoding.tolocal(self._extra['branch'])
1304 1300 def closesbranch(self):
1305 1301 return 'close' in self._extra
1306 1302 def extra(self):
1307 1303 return self._extra
1308 1304
1309 1305 def tags(self):
1310 1306 return []
1311 1307
1312 1308 def bookmarks(self):
1313 1309 b = []
1314 1310 for p in self.parents():
1315 1311 b.extend(p.bookmarks())
1316 1312 return b
1317 1313
1318 1314 def phase(self):
1319 1315 phase = phases.draft # default phase to draft
1320 1316 for p in self.parents():
1321 1317 phase = max(phase, p.phase())
1322 1318 return phase
1323 1319
1324 1320 def hidden(self):
1325 1321 return False
1326 1322
1327 1323 def children(self):
1328 1324 return []
1329 1325
1330 1326 def flags(self, path):
1331 1327 if '_manifest' in self.__dict__:
1332 1328 try:
1333 1329 return self._manifest.flags(path)
1334 1330 except KeyError:
1335 1331 return ''
1336 1332
1337 1333 try:
1338 1334 return self._flagfunc(path)
1339 1335 except OSError:
1340 1336 return ''
1341 1337
1342 1338 def ancestor(self, c2):
1343 1339 """return the "best" ancestor context of self and c2"""
1344 1340 return self._parents[0].ancestor(c2) # punt on two parents for now
1345 1341
1346 1342 def walk(self, match):
1347 1343 '''Generates matching file names.'''
1348 1344 return sorted(self._repo.dirstate.walk(match, sorted(self.substate),
1349 1345 True, False))
1350 1346
1351 1347 def matches(self, match):
1352 1348 return sorted(self._repo.dirstate.matches(match))
1353 1349
1354 1350 def ancestors(self):
1355 1351 for p in self._parents:
1356 1352 yield p
1357 1353 for a in self._repo.changelog.ancestors(
1358 1354 [p.rev() for p in self._parents]):
1359 1355 yield changectx(self._repo, a)
1360 1356
1361 1357 def markcommitted(self, node):
1362 1358 """Perform post-commit cleanup necessary after committing this ctx
1363 1359
1364 1360 Specifically, this updates backing stores this working context
1365 1361 wraps to reflect the fact that the changes reflected by this
1366 1362 workingctx have been committed. For example, it marks
1367 1363 modified and added files as normal in the dirstate.
1368 1364
1369 1365 """
1370 1366
1371 1367 self._repo.dirstate.beginparentchange()
1372 1368 for f in self.modified() + self.added():
1373 1369 self._repo.dirstate.normal(f)
1374 1370 for f in self.removed():
1375 1371 self._repo.dirstate.drop(f)
1376 1372 self._repo.dirstate.setparents(node)
1377 1373 self._repo.dirstate.endparentchange()
1378 1374
1379 1375 # write changes out explicitly, because nesting wlock at
1380 1376 # runtime may prevent 'wlock.release()' in 'repo.commit()'
1381 1377 # from immediately doing so for subsequent changing files
1382 1378 self._repo.dirstate.write(self._repo.currenttransaction())
1383 1379
1384 1380 class workingctx(committablectx):
1385 1381 """A workingctx object makes access to data related to
1386 1382 the current working directory convenient.
1387 1383 date - any valid date string or (unixtime, offset), or None.
1388 1384 user - username string, or None.
1389 1385 extra - a dictionary of extra values, or None.
1390 1386 changes - a list of file lists as returned by localrepo.status()
1391 1387 or None to use the repository status.
1392 1388 """
1393 1389 def __init__(self, repo, text="", user=None, date=None, extra=None,
1394 1390 changes=None):
1395 1391 super(workingctx, self).__init__(repo, text, user, date, extra, changes)
1396 1392
1397 1393 def __iter__(self):
1398 1394 d = self._repo.dirstate
1399 1395 for f in d:
1400 1396 if d[f] != 'r':
1401 1397 yield f
1402 1398
1403 1399 def __contains__(self, key):
1404 1400 return self._repo.dirstate[key] not in "?r"
1405 1401
1406 1402 def hex(self):
1407 1403 return hex(wdirid)
1408 1404
1409 1405 @propertycache
1410 1406 def _parents(self):
1411 1407 p = self._repo.dirstate.parents()
1412 1408 if p[1] == nullid:
1413 1409 p = p[:-1]
1414 1410 return [changectx(self._repo, x) for x in p]
1415 1411
1416 1412 def filectx(self, path, filelog=None):
1417 1413 """get a file context from the working directory"""
1418 1414 return workingfilectx(self._repo, path, workingctx=self,
1419 1415 filelog=filelog)
1420 1416
1421 1417 def dirty(self, missing=False, merge=True, branch=True):
1422 1418 "check whether a working directory is modified"
1423 1419 # check subrepos first
1424 1420 for s in sorted(self.substate):
1425 1421 if self.sub(s).dirty():
1426 1422 return True
1427 1423 # check current working dir
1428 1424 return ((merge and self.p2()) or
1429 1425 (branch and self.branch() != self.p1().branch()) or
1430 1426 self.modified() or self.added() or self.removed() or
1431 1427 (missing and self.deleted()))
1432 1428
1433 1429 def add(self, list, prefix=""):
1434 1430 join = lambda f: os.path.join(prefix, f)
1435 1431 with self._repo.wlock():
1436 1432 ui, ds = self._repo.ui, self._repo.dirstate
1437 1433 rejected = []
1438 1434 lstat = self._repo.wvfs.lstat
1439 1435 for f in list:
1440 1436 scmutil.checkportable(ui, join(f))
1441 1437 try:
1442 1438 st = lstat(f)
1443 1439 except OSError:
1444 1440 ui.warn(_("%s does not exist!\n") % join(f))
1445 1441 rejected.append(f)
1446 1442 continue
1447 1443 if st.st_size > 10000000:
1448 1444 ui.warn(_("%s: up to %d MB of RAM may be required "
1449 1445 "to manage this file\n"
1450 1446 "(use 'hg revert %s' to cancel the "
1451 1447 "pending addition)\n")
1452 1448 % (f, 3 * st.st_size // 1000000, join(f)))
1453 1449 if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)):
1454 1450 ui.warn(_("%s not added: only files and symlinks "
1455 1451 "supported currently\n") % join(f))
1456 1452 rejected.append(f)
1457 1453 elif ds[f] in 'amn':
1458 1454 ui.warn(_("%s already tracked!\n") % join(f))
1459 1455 elif ds[f] == 'r':
1460 1456 ds.normallookup(f)
1461 1457 else:
1462 1458 ds.add(f)
1463 1459 return rejected
1464 1460
1465 1461 def forget(self, files, prefix=""):
1466 1462 join = lambda f: os.path.join(prefix, f)
1467 1463 with self._repo.wlock():
1468 1464 rejected = []
1469 1465 for f in files:
1470 1466 if f not in self._repo.dirstate:
1471 1467 self._repo.ui.warn(_("%s not tracked!\n") % join(f))
1472 1468 rejected.append(f)
1473 1469 elif self._repo.dirstate[f] != 'a':
1474 1470 self._repo.dirstate.remove(f)
1475 1471 else:
1476 1472 self._repo.dirstate.drop(f)
1477 1473 return rejected
1478 1474
1479 1475 def undelete(self, list):
1480 1476 pctxs = self.parents()
1481 1477 with self._repo.wlock():
1482 1478 for f in list:
1483 1479 if self._repo.dirstate[f] != 'r':
1484 1480 self._repo.ui.warn(_("%s not removed!\n") % f)
1485 1481 else:
1486 1482 fctx = f in pctxs[0] and pctxs[0][f] or pctxs[1][f]
1487 1483 t = fctx.data()
1488 1484 self._repo.wwrite(f, t, fctx.flags())
1489 1485 self._repo.dirstate.normal(f)
1490 1486
1491 1487 def copy(self, source, dest):
1492 1488 try:
1493 1489 st = self._repo.wvfs.lstat(dest)
1494 1490 except OSError as err:
1495 1491 if err.errno != errno.ENOENT:
1496 1492 raise
1497 1493 self._repo.ui.warn(_("%s does not exist!\n") % dest)
1498 1494 return
1499 1495 if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)):
1500 1496 self._repo.ui.warn(_("copy failed: %s is not a file or a "
1501 1497 "symbolic link\n") % dest)
1502 1498 else:
1503 1499 with self._repo.wlock():
1504 1500 if self._repo.dirstate[dest] in '?':
1505 1501 self._repo.dirstate.add(dest)
1506 1502 elif self._repo.dirstate[dest] in 'r':
1507 1503 self._repo.dirstate.normallookup(dest)
1508 1504 self._repo.dirstate.copy(source, dest)
1509 1505
1510 1506 def match(self, pats=[], include=None, exclude=None, default='glob',
1511 1507 listsubrepos=False, badfn=None):
1512 1508 r = self._repo
1513 1509
1514 1510 # Only a case insensitive filesystem needs magic to translate user input
1515 1511 # to actual case in the filesystem.
1516 1512 if not util.fscasesensitive(r.root):
1517 1513 return matchmod.icasefsmatcher(r.root, r.getcwd(), pats, include,
1518 1514 exclude, default, r.auditor, self,
1519 1515 listsubrepos=listsubrepos,
1520 1516 badfn=badfn)
1521 1517 return matchmod.match(r.root, r.getcwd(), pats,
1522 1518 include, exclude, default,
1523 1519 auditor=r.auditor, ctx=self,
1524 1520 listsubrepos=listsubrepos, badfn=badfn)
1525 1521
1526 1522 def _filtersuspectsymlink(self, files):
1527 1523 if not files or self._repo.dirstate._checklink:
1528 1524 return files
1529 1525
1530 1526 # Symlink placeholders may get non-symlink-like contents
1531 1527 # via user error or dereferencing by NFS or Samba servers,
1532 1528 # so we filter out any placeholders that don't look like a
1533 1529 # symlink
1534 1530 sane = []
1535 1531 for f in files:
1536 1532 if self.flags(f) == 'l':
1537 1533 d = self[f].data()
1538 1534 if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d):
1539 1535 self._repo.ui.debug('ignoring suspect symlink placeholder'
1540 1536 ' "%s"\n' % f)
1541 1537 continue
1542 1538 sane.append(f)
1543 1539 return sane
1544 1540
1545 1541 def _checklookup(self, files):
1546 1542 # check for any possibly clean files
1547 1543 if not files:
1548 1544 return [], []
1549 1545
1550 1546 modified = []
1551 1547 fixup = []
1552 1548 pctx = self._parents[0]
1553 1549 # do a full compare of any files that might have changed
1554 1550 for f in sorted(files):
1555 1551 if (f not in pctx or self.flags(f) != pctx.flags(f)
1556 1552 or pctx[f].cmp(self[f])):
1557 1553 modified.append(f)
1558 1554 else:
1559 1555 fixup.append(f)
1560 1556
1561 1557 # update dirstate for files that are actually clean
1562 1558 if fixup:
1563 1559 try:
1564 1560 # updating the dirstate is optional
1565 1561 # so we don't wait on the lock
1566 1562 # wlock can invalidate the dirstate, so cache normal _after_
1567 1563 # taking the lock
1568 1564 with self._repo.wlock(False):
1569 1565 normal = self._repo.dirstate.normal
1570 1566 for f in fixup:
1571 1567 normal(f)
1572 1568 # write changes out explicitly, because nesting
1573 1569 # wlock at runtime may prevent 'wlock.release()'
1574 1570 # after this block from doing so for subsequent
1575 1571 # changing files
1576 1572 self._repo.dirstate.write(self._repo.currenttransaction())
1577 1573 except error.LockError:
1578 1574 pass
1579 1575 return modified, fixup
1580 1576
1581 1577 def _manifestmatches(self, match, s):
1582 1578 """Slow path for workingctx
1583 1579
1584 1580 The fast path is when we compare the working directory to its parent
1585 1581 which means this function is comparing with a non-parent; therefore we
1586 1582 need to build a manifest and return what matches.
1587 1583 """
1588 1584 mf = self._repo['.']._manifestmatches(match, s)
1589 1585 for f in s.modified + s.added:
1590 mf[f] = _newnode
1586 mf[f] = newnodeid
1591 1587 mf.setflag(f, self.flags(f))
1592 1588 for f in s.removed:
1593 1589 if f in mf:
1594 1590 del mf[f]
1595 1591 return mf
1596 1592
1597 1593 def _dirstatestatus(self, match=None, ignored=False, clean=False,
1598 1594 unknown=False):
1599 1595 '''Gets the status from the dirstate -- internal use only.'''
1600 1596 listignored, listclean, listunknown = ignored, clean, unknown
1601 1597 match = match or matchmod.always(self._repo.root, self._repo.getcwd())
1602 1598 subrepos = []
1603 1599 if '.hgsub' in self:
1604 1600 subrepos = sorted(self.substate)
1605 1601 cmp, s = self._repo.dirstate.status(match, subrepos, listignored,
1606 1602 listclean, listunknown)
1607 1603
1608 1604 # check for any possibly clean files
1609 1605 if cmp:
1610 1606 modified2, fixup = self._checklookup(cmp)
1611 1607 s.modified.extend(modified2)
1612 1608
1613 1609 # update dirstate for files that are actually clean
1614 1610 if fixup and listclean:
1615 1611 s.clean.extend(fixup)
1616 1612
1617 1613 if match.always():
1618 1614 # cache for performance
1619 1615 if s.unknown or s.ignored or s.clean:
1620 1616 # "_status" is cached with list*=False in the normal route
1621 1617 self._status = scmutil.status(s.modified, s.added, s.removed,
1622 1618 s.deleted, [], [], [])
1623 1619 else:
1624 1620 self._status = s
1625 1621
1626 1622 return s
1627 1623
1628 1624 def _buildstatus(self, other, s, match, listignored, listclean,
1629 1625 listunknown):
1630 1626 """build a status with respect to another context
1631 1627
1632 1628 This includes logic for maintaining the fast path of status when
1633 1629 comparing the working directory against its parent, which is to skip
1634 1630 building a new manifest if self (working directory) is not comparing
1635 1631 against its parent (repo['.']).
1636 1632 """
1637 1633 s = self._dirstatestatus(match, listignored, listclean, listunknown)
1638 1634 # Filter out symlinks that, in the case of FAT32 and NTFS filesystems,
1639 1635 # might have accidentally ended up with the entire contents of the file
1640 1636 # they are supposed to be linking to.
1641 1637 s.modified[:] = self._filtersuspectsymlink(s.modified)
1642 1638 if other != self._repo['.']:
1643 1639 s = super(workingctx, self)._buildstatus(other, s, match,
1644 1640 listignored, listclean,
1645 1641 listunknown)
1646 1642 return s
1647 1643
1648 1644 def _matchstatus(self, other, match):
1649 1645 """override the match method with a filter for directory patterns
1650 1646
1651 1647 We use inheritance to customize the match.bad method only in cases of
1652 1648 workingctx since it belongs only to the working directory when
1653 1649 comparing against the parent changeset.
1654 1650
1655 1651 If we aren't comparing against the working directory's parent, then we
1656 1652 just use the default match object sent to us.
1657 1653 """
1658 1654 superself = super(workingctx, self)
1659 1655 match = superself._matchstatus(other, match)
1660 1656 if other != self._repo['.']:
1661 1657 def bad(f, msg):
1662 1658 # 'f' may be a directory pattern from 'match.files()',
1663 1659 # so 'f not in ctx1' is not enough
1664 1660 if f not in other and not other.hasdir(f):
1665 1661 self._repo.ui.warn('%s: %s\n' %
1666 1662 (self._repo.dirstate.pathto(f), msg))
1667 1663 match.bad = bad
1668 1664 return match
1669 1665
1670 1666 class committablefilectx(basefilectx):
1671 1667 """A committablefilectx provides common functionality for a file context
1672 1668 that wants the ability to commit, e.g. workingfilectx or memfilectx."""
1673 1669 def __init__(self, repo, path, filelog=None, ctx=None):
1674 1670 self._repo = repo
1675 1671 self._path = path
1676 1672 self._changeid = None
1677 1673 self._filerev = self._filenode = None
1678 1674
1679 1675 if filelog is not None:
1680 1676 self._filelog = filelog
1681 1677 if ctx:
1682 1678 self._changectx = ctx
1683 1679
1684 1680 def __nonzero__(self):
1685 1681 return True
1686 1682
1687 1683 def linkrev(self):
1688 1684 # linked to self._changectx no matter if file is modified or not
1689 1685 return self.rev()
1690 1686
1691 1687 def parents(self):
1692 1688 '''return parent filectxs, following copies if necessary'''
1693 1689 def filenode(ctx, path):
1694 1690 return ctx._manifest.get(path, nullid)
1695 1691
1696 1692 path = self._path
1697 1693 fl = self._filelog
1698 1694 pcl = self._changectx._parents
1699 1695 renamed = self.renamed()
1700 1696
1701 1697 if renamed:
1702 1698 pl = [renamed + (None,)]
1703 1699 else:
1704 1700 pl = [(path, filenode(pcl[0], path), fl)]
1705 1701
1706 1702 for pc in pcl[1:]:
1707 1703 pl.append((path, filenode(pc, path), fl))
1708 1704
1709 1705 return [self._parentfilectx(p, fileid=n, filelog=l)
1710 1706 for p, n, l in pl if n != nullid]
1711 1707
1712 1708 def children(self):
1713 1709 return []
1714 1710
1715 1711 class workingfilectx(committablefilectx):
1716 1712 """A workingfilectx object makes access to data related to a particular
1717 1713 file in the working directory convenient."""
1718 1714 def __init__(self, repo, path, filelog=None, workingctx=None):
1719 1715 super(workingfilectx, self).__init__(repo, path, filelog, workingctx)
1720 1716
1721 1717 @propertycache
1722 1718 def _changectx(self):
1723 1719 return workingctx(self._repo)
1724 1720
1725 1721 def data(self):
1726 1722 return self._repo.wread(self._path)
1727 1723 def renamed(self):
1728 1724 rp = self._repo.dirstate.copied(self._path)
1729 1725 if not rp:
1730 1726 return None
1731 1727 return rp, self._changectx._parents[0]._manifest.get(rp, nullid)
1732 1728
1733 1729 def size(self):
1734 1730 return self._repo.wvfs.lstat(self._path).st_size
1735 1731 def date(self):
1736 1732 t, tz = self._changectx.date()
1737 1733 try:
1738 1734 return (self._repo.wvfs.lstat(self._path).st_mtime, tz)
1739 1735 except OSError as err:
1740 1736 if err.errno != errno.ENOENT:
1741 1737 raise
1742 1738 return (t, tz)
1743 1739
1744 1740 def cmp(self, fctx):
1745 1741 """compare with other file context
1746 1742
1747 1743 returns True if different than fctx.
1748 1744 """
1749 1745 # fctx should be a filectx (not a workingfilectx)
1750 1746 # invert comparison to reuse the same code path
1751 1747 return fctx.cmp(self)
1752 1748
1753 1749 def remove(self, ignoremissing=False):
1754 1750 """wraps unlink for a repo's working directory"""
1755 1751 util.unlinkpath(self._repo.wjoin(self._path), ignoremissing)
1756 1752
1757 1753 def write(self, data, flags):
1758 1754 """wraps repo.wwrite"""
1759 1755 self._repo.wwrite(self._path, data, flags)
1760 1756
1761 1757 class workingcommitctx(workingctx):
1762 1758 """A workingcommitctx object makes access to data related to
1763 1759 the revision being committed convenient.
1764 1760
1765 1761 This hides changes in the working directory, if they aren't
1766 1762 committed in this context.
1767 1763 """
1768 1764 def __init__(self, repo, changes,
1769 1765 text="", user=None, date=None, extra=None):
1770 1766 super(workingctx, self).__init__(repo, text, user, date, extra,
1771 1767 changes)
1772 1768
1773 1769 def _dirstatestatus(self, match=None, ignored=False, clean=False,
1774 1770 unknown=False):
1775 1771 """Return matched files only in ``self._status``
1776 1772
1777 1773 Uncommitted files appear "clean" via this context, even if
1778 1774 they aren't actually so in the working directory.
1779 1775 """
1780 1776 match = match or matchmod.always(self._repo.root, self._repo.getcwd())
1781 1777 if clean:
1782 1778 clean = [f for f in self._manifest if f not in self._changedset]
1783 1779 else:
1784 1780 clean = []
1785 1781 return scmutil.status([f for f in self._status.modified if match(f)],
1786 1782 [f for f in self._status.added if match(f)],
1787 1783 [f for f in self._status.removed if match(f)],
1788 1784 [], [], [], clean)
1789 1785
1790 1786 @propertycache
1791 1787 def _changedset(self):
1792 1788 """Return the set of files changed in this context
1793 1789 """
1794 1790 changed = set(self._status.modified)
1795 1791 changed.update(self._status.added)
1796 1792 changed.update(self._status.removed)
1797 1793 return changed
1798 1794
1799 1795 def makecachingfilectxfn(func):
1800 1796 """Create a filectxfn that caches based on the path.
1801 1797
1802 1798 We can't use util.cachefunc because it uses all arguments as the cache
1803 1799 key and this creates a cycle since the arguments include the repo and
1804 1800 memctx.
1805 1801 """
1806 1802 cache = {}
1807 1803
1808 1804 def getfilectx(repo, memctx, path):
1809 1805 if path not in cache:
1810 1806 cache[path] = func(repo, memctx, path)
1811 1807 return cache[path]
1812 1808
1813 1809 return getfilectx
1814 1810
1815 1811 class memctx(committablectx):
1816 1812 """Use memctx to perform in-memory commits via localrepo.commitctx().
1817 1813
1818 1814 Revision information is supplied at initialization time while
1819 1815 related files data and is made available through a callback
1820 1816 mechanism. 'repo' is the current localrepo, 'parents' is a
1821 1817 sequence of two parent revisions identifiers (pass None for every
1822 1818 missing parent), 'text' is the commit message and 'files' lists
1823 1819 names of files touched by the revision (normalized and relative to
1824 1820 repository root).
1825 1821
1826 1822 filectxfn(repo, memctx, path) is a callable receiving the
1827 1823 repository, the current memctx object and the normalized path of
1828 1824 requested file, relative to repository root. It is fired by the
1829 1825 commit function for every file in 'files', but calls order is
1830 1826 undefined. If the file is available in the revision being
1831 1827 committed (updated or added), filectxfn returns a memfilectx
1832 1828 object. If the file was removed, filectxfn raises an
1833 1829 IOError. Moved files are represented by marking the source file
1834 1830 removed and the new file added with copy information (see
1835 1831 memfilectx).
1836 1832
1837 1833 user receives the committer name and defaults to current
1838 1834 repository username, date is the commit date in any format
1839 1835 supported by util.parsedate() and defaults to current date, extra
1840 1836 is a dictionary of metadata or is left empty.
1841 1837 """
1842 1838
1843 1839 # Mercurial <= 3.1 expects the filectxfn to raise IOError for missing files.
1844 1840 # Extensions that need to retain compatibility across Mercurial 3.1 can use
1845 1841 # this field to determine what to do in filectxfn.
1846 1842 _returnnoneformissingfiles = True
1847 1843
1848 1844 def __init__(self, repo, parents, text, files, filectxfn, user=None,
1849 1845 date=None, extra=None, editor=False):
1850 1846 super(memctx, self).__init__(repo, text, user, date, extra)
1851 1847 self._rev = None
1852 1848 self._node = None
1853 1849 parents = [(p or nullid) for p in parents]
1854 1850 p1, p2 = parents
1855 1851 self._parents = [changectx(self._repo, p) for p in (p1, p2)]
1856 1852 files = sorted(set(files))
1857 1853 self._files = files
1858 1854 self.substate = {}
1859 1855
1860 1856 # if store is not callable, wrap it in a function
1861 1857 if not callable(filectxfn):
1862 1858 def getfilectx(repo, memctx, path):
1863 1859 fctx = filectxfn[path]
1864 1860 # this is weird but apparently we only keep track of one parent
1865 1861 # (why not only store that instead of a tuple?)
1866 1862 copied = fctx.renamed()
1867 1863 if copied:
1868 1864 copied = copied[0]
1869 1865 return memfilectx(repo, path, fctx.data(),
1870 1866 islink=fctx.islink(), isexec=fctx.isexec(),
1871 1867 copied=copied, memctx=memctx)
1872 1868 self._filectxfn = getfilectx
1873 1869 else:
1874 1870 # memoizing increases performance for e.g. vcs convert scenarios.
1875 1871 self._filectxfn = makecachingfilectxfn(filectxfn)
1876 1872
1877 1873 if extra:
1878 1874 self._extra = extra.copy()
1879 1875 else:
1880 1876 self._extra = {}
1881 1877
1882 1878 if self._extra.get('branch', '') == '':
1883 1879 self._extra['branch'] = 'default'
1884 1880
1885 1881 if editor:
1886 1882 self._text = editor(self._repo, self, [])
1887 1883 self._repo.savecommitmessage(self._text)
1888 1884
1889 1885 def filectx(self, path, filelog=None):
1890 1886 """get a file context from the working directory
1891 1887
1892 1888 Returns None if file doesn't exist and should be removed."""
1893 1889 return self._filectxfn(self._repo, self, path)
1894 1890
1895 1891 def commit(self):
1896 1892 """commit context to the repo"""
1897 1893 return self._repo.commitctx(self)
1898 1894
1899 1895 @propertycache
1900 1896 def _manifest(self):
1901 1897 """generate a manifest based on the return values of filectxfn"""
1902 1898
1903 1899 # keep this simple for now; just worry about p1
1904 1900 pctx = self._parents[0]
1905 1901 man = pctx.manifest().copy()
1906 1902
1907 1903 for f in self._status.modified:
1908 1904 p1node = nullid
1909 1905 p2node = nullid
1910 1906 p = pctx[f].parents() # if file isn't in pctx, check p2?
1911 1907 if len(p) > 0:
1912 1908 p1node = p[0].filenode()
1913 1909 if len(p) > 1:
1914 1910 p2node = p[1].filenode()
1915 1911 man[f] = revlog.hash(self[f].data(), p1node, p2node)
1916 1912
1917 1913 for f in self._status.added:
1918 1914 man[f] = revlog.hash(self[f].data(), nullid, nullid)
1919 1915
1920 1916 for f in self._status.removed:
1921 1917 if f in man:
1922 1918 del man[f]
1923 1919
1924 1920 return man
1925 1921
1926 1922 @propertycache
1927 1923 def _status(self):
1928 1924 """Calculate exact status from ``files`` specified at construction
1929 1925 """
1930 1926 man1 = self.p1().manifest()
1931 1927 p2 = self._parents[1]
1932 1928 # "1 < len(self._parents)" can't be used for checking
1933 1929 # existence of the 2nd parent, because "memctx._parents" is
1934 1930 # explicitly initialized by the list, of which length is 2.
1935 1931 if p2.node() != nullid:
1936 1932 man2 = p2.manifest()
1937 1933 managing = lambda f: f in man1 or f in man2
1938 1934 else:
1939 1935 managing = lambda f: f in man1
1940 1936
1941 1937 modified, added, removed = [], [], []
1942 1938 for f in self._files:
1943 1939 if not managing(f):
1944 1940 added.append(f)
1945 1941 elif self[f]:
1946 1942 modified.append(f)
1947 1943 else:
1948 1944 removed.append(f)
1949 1945
1950 1946 return scmutil.status(modified, added, removed, [], [], [], [])
1951 1947
1952 1948 class memfilectx(committablefilectx):
1953 1949 """memfilectx represents an in-memory file to commit.
1954 1950
1955 1951 See memctx and committablefilectx for more details.
1956 1952 """
1957 1953 def __init__(self, repo, path, data, islink=False,
1958 1954 isexec=False, copied=None, memctx=None):
1959 1955 """
1960 1956 path is the normalized file path relative to repository root.
1961 1957 data is the file content as a string.
1962 1958 islink is True if the file is a symbolic link.
1963 1959 isexec is True if the file is executable.
1964 1960 copied is the source file path if current file was copied in the
1965 1961 revision being committed, or None."""
1966 1962 super(memfilectx, self).__init__(repo, path, None, memctx)
1967 1963 self._data = data
1968 1964 self._flags = (islink and 'l' or '') + (isexec and 'x' or '')
1969 1965 self._copied = None
1970 1966 if copied:
1971 1967 self._copied = (copied, nullid)
1972 1968
1973 1969 def data(self):
1974 1970 return self._data
1975 1971 def size(self):
1976 1972 return len(self.data())
1977 1973 def flags(self):
1978 1974 return self._flags
1979 1975 def renamed(self):
1980 1976 return self._copied
1981 1977
1982 1978 def remove(self, ignoremissing=False):
1983 1979 """wraps unlink for a repo's working directory"""
1984 1980 # need to figure out what to do here
1985 1981 del self._changectx[self._path]
1986 1982
1987 1983 def write(self, data, flags):
1988 1984 """wraps repo.wwrite"""
1989 1985 self._data = data
@@ -1,714 +1,714
1 1 # copies.py - copy detection for Mercurial
2 2 #
3 3 # Copyright 2008 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import heapq
11 11
12 12 from . import (
13 13 node,
14 14 pathutil,
15 15 scmutil,
16 16 util,
17 17 )
18 18
19 19 def _findlimit(repo, a, b):
20 20 """
21 21 Find the last revision that needs to be checked to ensure that a full
22 22 transitive closure for file copies can be properly calculated.
23 23 Generally, this means finding the earliest revision number that's an
24 24 ancestor of a or b but not both, except when a or b is a direct descendent
25 25 of the other, in which case we can return the minimum revnum of a and b.
26 26 None if no such revision exists.
27 27 """
28 28
29 29 # basic idea:
30 30 # - mark a and b with different sides
31 31 # - if a parent's children are all on the same side, the parent is
32 32 # on that side, otherwise it is on no side
33 33 # - walk the graph in topological order with the help of a heap;
34 34 # - add unseen parents to side map
35 35 # - clear side of any parent that has children on different sides
36 36 # - track number of interesting revs that might still be on a side
37 37 # - track the lowest interesting rev seen
38 38 # - quit when interesting revs is zero
39 39
40 40 cl = repo.changelog
41 41 working = len(cl) # pseudo rev for the working directory
42 42 if a is None:
43 43 a = working
44 44 if b is None:
45 45 b = working
46 46
47 47 side = {a: -1, b: 1}
48 48 visit = [-a, -b]
49 49 heapq.heapify(visit)
50 50 interesting = len(visit)
51 51 hascommonancestor = False
52 52 limit = working
53 53
54 54 while interesting:
55 55 r = -heapq.heappop(visit)
56 56 if r == working:
57 57 parents = [cl.rev(p) for p in repo.dirstate.parents()]
58 58 else:
59 59 parents = cl.parentrevs(r)
60 60 for p in parents:
61 61 if p < 0:
62 62 continue
63 63 if p not in side:
64 64 # first time we see p; add it to visit
65 65 side[p] = side[r]
66 66 if side[p]:
67 67 interesting += 1
68 68 heapq.heappush(visit, -p)
69 69 elif side[p] and side[p] != side[r]:
70 70 # p was interesting but now we know better
71 71 side[p] = 0
72 72 interesting -= 1
73 73 hascommonancestor = True
74 74 if side[r]:
75 75 limit = r # lowest rev visited
76 76 interesting -= 1
77 77
78 78 if not hascommonancestor:
79 79 return None
80 80
81 81 # Consider the following flow (see test-commit-amend.t under issue4405):
82 82 # 1/ File 'a0' committed
83 83 # 2/ File renamed from 'a0' to 'a1' in a new commit (call it 'a1')
84 84 # 3/ Move back to first commit
85 85 # 4/ Create a new commit via revert to contents of 'a1' (call it 'a1-amend')
86 86 # 5/ Rename file from 'a1' to 'a2' and commit --amend 'a1-msg'
87 87 #
88 88 # During the amend in step five, we will be in this state:
89 89 #
90 90 # @ 3 temporary amend commit for a1-amend
91 91 # |
92 92 # o 2 a1-amend
93 93 # |
94 94 # | o 1 a1
95 95 # |/
96 96 # o 0 a0
97 97 #
98 98 # When _findlimit is called, a and b are revs 3 and 0, so limit will be 2,
99 99 # yet the filelog has the copy information in rev 1 and we will not look
100 100 # back far enough unless we also look at the a and b as candidates.
101 101 # This only occurs when a is a descendent of b or visa-versa.
102 102 return min(limit, a, b)
103 103
104 104 def _chain(src, dst, a, b):
105 105 '''chain two sets of copies a->b'''
106 106 t = a.copy()
107 107 for k, v in b.iteritems():
108 108 if v in t:
109 109 # found a chain
110 110 if t[v] != k:
111 111 # file wasn't renamed back to itself
112 112 t[k] = t[v]
113 113 if v not in dst:
114 114 # chain was a rename, not a copy
115 115 del t[v]
116 116 if v in src:
117 117 # file is a copy of an existing file
118 118 t[k] = v
119 119
120 120 # remove criss-crossed copies
121 121 for k, v in t.items():
122 122 if k in src and v in dst:
123 123 del t[k]
124 124
125 125 return t
126 126
127 127 def _tracefile(fctx, am, limit=-1):
128 128 '''return file context that is the ancestor of fctx present in ancestor
129 129 manifest am, stopping after the first ancestor lower than limit'''
130 130
131 131 for f in fctx.ancestors():
132 132 if am.get(f.path(), None) == f.filenode():
133 133 return f
134 134 if limit >= 0 and f.linkrev() < limit and f.rev() < limit:
135 135 return None
136 136
137 137 def _dirstatecopies(d):
138 138 ds = d._repo.dirstate
139 139 c = ds.copies().copy()
140 140 for k in c.keys():
141 141 if ds[k] not in 'anm':
142 142 del c[k]
143 143 return c
144 144
145 145 def _computeforwardmissing(a, b, match=None):
146 146 """Computes which files are in b but not a.
147 147 This is its own function so extensions can easily wrap this call to see what
148 148 files _forwardcopies is about to process.
149 149 """
150 150 ma = a.manifest()
151 151 mb = b.manifest()
152 152 if match:
153 153 ma = ma.matches(match)
154 154 mb = mb.matches(match)
155 155 return mb.filesnotin(ma)
156 156
157 157 def _forwardcopies(a, b, match=None):
158 158 '''find {dst@b: src@a} copy mapping where a is an ancestor of b'''
159 159
160 160 # check for working copy
161 161 w = None
162 162 if b.rev() is None:
163 163 w = b
164 164 b = w.p1()
165 165 if a == b:
166 166 # short-circuit to avoid issues with merge states
167 167 return _dirstatecopies(w)
168 168
169 169 # files might have to be traced back to the fctx parent of the last
170 170 # one-side-only changeset, but not further back than that
171 171 limit = _findlimit(a._repo, a.rev(), b.rev())
172 172 if limit is None:
173 173 limit = -1
174 174 am = a.manifest()
175 175
176 176 # find where new files came from
177 177 # we currently don't try to find where old files went, too expensive
178 178 # this means we can miss a case like 'hg rm b; hg cp a b'
179 179 cm = {}
180 180
181 181 # Computing the forward missing is quite expensive on large manifests, since
182 182 # it compares the entire manifests. We can optimize it in the common use
183 183 # case of computing what copies are in a commit versus its parent (like
184 184 # during a rebase or histedit). Note, we exclude merge commits from this
185 185 # optimization, since the ctx.files() for a merge commit is not correct for
186 186 # this comparison.
187 187 forwardmissingmatch = match
188 188 if not match and b.p1() == a and b.p2().node() == node.nullid:
189 189 forwardmissingmatch = scmutil.matchfiles(a._repo, b.files())
190 190 missing = _computeforwardmissing(a, b, match=forwardmissingmatch)
191 191
192 192 ancestrycontext = a._repo.changelog.ancestors([b.rev()], inclusive=True)
193 193 for f in missing:
194 194 fctx = b[f]
195 195 fctx._ancestrycontext = ancestrycontext
196 196 ofctx = _tracefile(fctx, am, limit)
197 197 if ofctx:
198 198 cm[f] = ofctx.path()
199 199
200 200 # combine copies from dirstate if necessary
201 201 if w is not None:
202 202 cm = _chain(a, w, cm, _dirstatecopies(w))
203 203
204 204 return cm
205 205
206 206 def _backwardrenames(a, b):
207 207 if a._repo.ui.configbool('experimental', 'disablecopytrace'):
208 208 return {}
209 209
210 210 # Even though we're not taking copies into account, 1:n rename situations
211 211 # can still exist (e.g. hg cp a b; hg mv a c). In those cases we
212 212 # arbitrarily pick one of the renames.
213 213 f = _forwardcopies(b, a)
214 214 r = {}
215 215 for k, v in sorted(f.iteritems()):
216 216 # remove copies
217 217 if v in a:
218 218 continue
219 219 r[v] = k
220 220 return r
221 221
222 222 def pathcopies(x, y, match=None):
223 223 '''find {dst@y: src@x} copy mapping for directed compare'''
224 224 if x == y or not x or not y:
225 225 return {}
226 226 a = y.ancestor(x)
227 227 if a == x:
228 228 return _forwardcopies(x, y, match=match)
229 229 if a == y:
230 230 return _backwardrenames(x, y)
231 231 return _chain(x, y, _backwardrenames(x, a),
232 232 _forwardcopies(a, y, match=match))
233 233
234 234 def _computenonoverlap(repo, c1, c2, addedinm1, addedinm2, baselabel=''):
235 235 """Computes, based on addedinm1 and addedinm2, the files exclusive to c1
236 236 and c2. This is its own function so extensions can easily wrap this call
237 237 to see what files mergecopies is about to process.
238 238
239 239 Even though c1 and c2 are not used in this function, they are useful in
240 240 other extensions for being able to read the file nodes of the changed files.
241 241
242 242 "baselabel" can be passed to help distinguish the multiple computations
243 243 done in the graft case.
244 244 """
245 245 u1 = sorted(addedinm1 - addedinm2)
246 246 u2 = sorted(addedinm2 - addedinm1)
247 247
248 248 header = " unmatched files in %s"
249 249 if baselabel:
250 250 header += ' (from %s)' % baselabel
251 251 if u1:
252 252 repo.ui.debug("%s:\n %s\n" % (header % 'local', "\n ".join(u1)))
253 253 if u2:
254 254 repo.ui.debug("%s:\n %s\n" % (header % 'other', "\n ".join(u2)))
255 255 return u1, u2
256 256
257 257 def _makegetfctx(ctx):
258 258 """return a 'getfctx' function suitable for _checkcopies usage
259 259
260 260 We have to re-setup the function building 'filectx' for each
261 261 '_checkcopies' to ensure the linkrev adjustment is properly setup for
262 262 each. Linkrev adjustment is important to avoid bug in rename
263 263 detection. Moreover, having a proper '_ancestrycontext' setup ensures
264 264 the performance impact of this adjustment is kept limited. Without it,
265 265 each file could do a full dag traversal making the time complexity of
266 266 the operation explode (see issue4537).
267 267
268 268 This function exists here mostly to limit the impact on stable. Feel
269 269 free to refactor on default.
270 270 """
271 271 rev = ctx.rev()
272 272 repo = ctx._repo
273 273 ac = getattr(ctx, '_ancestrycontext', None)
274 274 if ac is None:
275 275 revs = [rev]
276 276 if rev is None:
277 277 revs = [p.rev() for p in ctx.parents()]
278 278 ac = repo.changelog.ancestors(revs, inclusive=True)
279 279 ctx._ancestrycontext = ac
280 280 def makectx(f, n):
281 if len(n) != 20: # in a working context?
281 if len(n) != 20 or n in node.wdirnodes: # in a working context?
282 282 if ctx.rev() is None:
283 283 return ctx.filectx(f)
284 284 return repo[None][f]
285 285 fctx = repo.filectx(f, fileid=n)
286 286 # setup only needed for filectx not create from a changectx
287 287 fctx._ancestrycontext = ac
288 288 fctx._descendantrev = rev
289 289 return fctx
290 290 return util.lrucachefunc(makectx)
291 291
292 292 def _combinecopies(copyfrom, copyto, finalcopy, diverge, incompletediverge):
293 293 """combine partial copy paths"""
294 294 remainder = {}
295 295 for f in copyfrom:
296 296 if f in copyto:
297 297 finalcopy[copyto[f]] = copyfrom[f]
298 298 del copyto[f]
299 299 for f in incompletediverge:
300 300 assert f not in diverge
301 301 ic = incompletediverge[f]
302 302 if ic[0] in copyto:
303 303 diverge[f] = [copyto[ic[0]], ic[1]]
304 304 else:
305 305 remainder[f] = ic
306 306 return remainder
307 307
308 308 def mergecopies(repo, c1, c2, base):
309 309 """
310 310 Find moves and copies between context c1 and c2 that are relevant
311 311 for merging. 'base' will be used as the merge base.
312 312
313 313 Returns four dicts: "copy", "movewithdir", "diverge", and
314 314 "renamedelete".
315 315
316 316 "copy" is a mapping from destination name -> source name,
317 317 where source is in c1 and destination is in c2 or vice-versa.
318 318
319 319 "movewithdir" is a mapping from source name -> destination name,
320 320 where the file at source present in one context but not the other
321 321 needs to be moved to destination by the merge process, because the
322 322 other context moved the directory it is in.
323 323
324 324 "diverge" is a mapping of source name -> list of destination names
325 325 for divergent renames.
326 326
327 327 "renamedelete" is a mapping of source name -> list of destination
328 328 names for files deleted in c1 that were renamed in c2 or vice-versa.
329 329 """
330 330 # avoid silly behavior for update from empty dir
331 331 if not c1 or not c2 or c1 == c2:
332 332 return {}, {}, {}, {}
333 333
334 334 # avoid silly behavior for parent -> working dir
335 335 if c2.node() is None and c1.node() == repo.dirstate.p1():
336 336 return repo.dirstate.copies(), {}, {}, {}
337 337
338 338 # Copy trace disabling is explicitly below the node == p1 logic above
339 339 # because the logic above is required for a simple copy to be kept across a
340 340 # rebase.
341 341 if repo.ui.configbool('experimental', 'disablecopytrace'):
342 342 return {}, {}, {}, {}
343 343
344 344 # In certain scenarios (e.g. graft, update or rebase), base can be
345 345 # overridden We still need to know a real common ancestor in this case We
346 346 # can't just compute _c1.ancestor(_c2) and compare it to ca, because there
347 347 # can be multiple common ancestors, e.g. in case of bidmerge. Because our
348 348 # caller may not know if the revision passed in lieu of the CA is a genuine
349 349 # common ancestor or not without explicitly checking it, it's better to
350 350 # determine that here.
351 351 #
352 352 # base.descendant(wc) and base.descendant(base) are False, work around that
353 353 _c1 = c1.p1() if c1.rev() is None else c1
354 354 _c2 = c2.p1() if c2.rev() is None else c2
355 355 # an endpoint is "dirty" if it isn't a descendant of the merge base
356 356 # if we have a dirty endpoint, we need to trigger graft logic, and also
357 357 # keep track of which endpoint is dirty
358 358 dirtyc1 = not (base == _c1 or base.descendant(_c1))
359 359 dirtyc2 = not (base== _c2 or base.descendant(_c2))
360 360 graft = dirtyc1 or dirtyc2
361 361 tca = base
362 362 if graft:
363 363 tca = _c1.ancestor(_c2)
364 364
365 365 limit = _findlimit(repo, c1.rev(), c2.rev())
366 366 if limit is None:
367 367 # no common ancestor, no copies
368 368 return {}, {}, {}, {}
369 369 repo.ui.debug(" searching for copies back to rev %d\n" % limit)
370 370
371 371 m1 = c1.manifest()
372 372 m2 = c2.manifest()
373 373 mb = base.manifest()
374 374
375 375 # gather data from _checkcopies:
376 376 # - diverge = record all diverges in this dict
377 377 # - copy = record all non-divergent copies in this dict
378 378 # - fullcopy = record all copies in this dict
379 379 # - incomplete = record non-divergent partial copies here
380 380 # - incompletediverge = record divergent partial copies here
381 381 diverge = {} # divergence data is shared
382 382 incompletediverge = {}
383 383 data1 = {'copy': {},
384 384 'fullcopy': {},
385 385 'incomplete': {},
386 386 'diverge': diverge,
387 387 'incompletediverge': incompletediverge,
388 388 }
389 389 data2 = {'copy': {},
390 390 'fullcopy': {},
391 391 'incomplete': {},
392 392 'diverge': diverge,
393 393 'incompletediverge': incompletediverge,
394 394 }
395 395
396 396 # find interesting file sets from manifests
397 397 addedinm1 = m1.filesnotin(mb)
398 398 addedinm2 = m2.filesnotin(mb)
399 399 bothnew = sorted(addedinm1 & addedinm2)
400 400 if tca == base:
401 401 # unmatched file from base
402 402 u1r, u2r = _computenonoverlap(repo, c1, c2, addedinm1, addedinm2)
403 403 u1u, u2u = u1r, u2r
404 404 else:
405 405 # unmatched file from base (DAG rotation in the graft case)
406 406 u1r, u2r = _computenonoverlap(repo, c1, c2, addedinm1, addedinm2,
407 407 baselabel='base')
408 408 # unmatched file from topological common ancestors (no DAG rotation)
409 409 # need to recompute this for directory move handling when grafting
410 410 mta = tca.manifest()
411 411 u1u, u2u = _computenonoverlap(repo, c1, c2, m1.filesnotin(mta),
412 412 m2.filesnotin(mta),
413 413 baselabel='topological common ancestor')
414 414
415 415 for f in u1u:
416 416 _checkcopies(c1, f, m1, m2, base, tca, dirtyc1, limit, data1)
417 417
418 418 for f in u2u:
419 419 _checkcopies(c2, f, m2, m1, base, tca, dirtyc2, limit, data2)
420 420
421 421 copy = dict(data1['copy'].items() + data2['copy'].items())
422 422 fullcopy = dict(data1['fullcopy'].items() + data2['fullcopy'].items())
423 423
424 424 if dirtyc1:
425 425 _combinecopies(data2['incomplete'], data1['incomplete'], copy, diverge,
426 426 incompletediverge)
427 427 else:
428 428 _combinecopies(data1['incomplete'], data2['incomplete'], copy, diverge,
429 429 incompletediverge)
430 430
431 431 renamedelete = {}
432 432 renamedeleteset = set()
433 433 divergeset = set()
434 434 for of, fl in diverge.items():
435 435 if len(fl) == 1 or of in c1 or of in c2:
436 436 del diverge[of] # not actually divergent, or not a rename
437 437 if of not in c1 and of not in c2:
438 438 # renamed on one side, deleted on the other side, but filter
439 439 # out files that have been renamed and then deleted
440 440 renamedelete[of] = [f for f in fl if f in c1 or f in c2]
441 441 renamedeleteset.update(fl) # reverse map for below
442 442 else:
443 443 divergeset.update(fl) # reverse map for below
444 444
445 445 if bothnew:
446 446 repo.ui.debug(" unmatched files new in both:\n %s\n"
447 447 % "\n ".join(bothnew))
448 448 bothdiverge = {}
449 449 bothincompletediverge = {}
450 450 remainder = {}
451 451 both1 = {'copy': {},
452 452 'fullcopy': {},
453 453 'incomplete': {},
454 454 'diverge': bothdiverge,
455 455 'incompletediverge': bothincompletediverge
456 456 }
457 457 both2 = {'copy': {},
458 458 'fullcopy': {},
459 459 'incomplete': {},
460 460 'diverge': bothdiverge,
461 461 'incompletediverge': bothincompletediverge
462 462 }
463 463 for f in bothnew:
464 464 _checkcopies(c1, f, m1, m2, base, tca, dirtyc1, limit, both1)
465 465 _checkcopies(c2, f, m2, m1, base, tca, dirtyc2, limit, both2)
466 466 if dirtyc1:
467 467 # incomplete copies may only be found on the "dirty" side for bothnew
468 468 assert not both2['incomplete']
469 469 remainder = _combinecopies({}, both1['incomplete'], copy, bothdiverge,
470 470 bothincompletediverge)
471 471 elif dirtyc2:
472 472 assert not both1['incomplete']
473 473 remainder = _combinecopies({}, both2['incomplete'], copy, bothdiverge,
474 474 bothincompletediverge)
475 475 else:
476 476 # incomplete copies and divergences can't happen outside grafts
477 477 assert not both1['incomplete']
478 478 assert not both2['incomplete']
479 479 assert not bothincompletediverge
480 480 for f in remainder:
481 481 assert f not in bothdiverge
482 482 ic = remainder[f]
483 483 if ic[0] in (m1 if dirtyc1 else m2):
484 484 # backed-out rename on one side, but watch out for deleted files
485 485 bothdiverge[f] = ic
486 486 for of, fl in bothdiverge.items():
487 487 if len(fl) == 2 and fl[0] == fl[1]:
488 488 copy[fl[0]] = of # not actually divergent, just matching renames
489 489
490 490 if fullcopy and repo.ui.debugflag:
491 491 repo.ui.debug(" all copies found (* = to merge, ! = divergent, "
492 492 "% = renamed and deleted):\n")
493 493 for f in sorted(fullcopy):
494 494 note = ""
495 495 if f in copy:
496 496 note += "*"
497 497 if f in divergeset:
498 498 note += "!"
499 499 if f in renamedeleteset:
500 500 note += "%"
501 501 repo.ui.debug(" src: '%s' -> dst: '%s' %s\n" % (fullcopy[f], f,
502 502 note))
503 503 del divergeset
504 504
505 505 if not fullcopy:
506 506 return copy, {}, diverge, renamedelete
507 507
508 508 repo.ui.debug(" checking for directory renames\n")
509 509
510 510 # generate a directory move map
511 511 d1, d2 = c1.dirs(), c2.dirs()
512 512 # Hack for adding '', which is not otherwise added, to d1 and d2
513 513 d1.addpath('/')
514 514 d2.addpath('/')
515 515 invalid = set()
516 516 dirmove = {}
517 517
518 518 # examine each file copy for a potential directory move, which is
519 519 # when all the files in a directory are moved to a new directory
520 520 for dst, src in fullcopy.iteritems():
521 521 dsrc, ddst = pathutil.dirname(src), pathutil.dirname(dst)
522 522 if dsrc in invalid:
523 523 # already seen to be uninteresting
524 524 continue
525 525 elif dsrc in d1 and ddst in d1:
526 526 # directory wasn't entirely moved locally
527 527 invalid.add(dsrc + "/")
528 528 elif dsrc in d2 and ddst in d2:
529 529 # directory wasn't entirely moved remotely
530 530 invalid.add(dsrc + "/")
531 531 elif dsrc + "/" in dirmove and dirmove[dsrc + "/"] != ddst + "/":
532 532 # files from the same directory moved to two different places
533 533 invalid.add(dsrc + "/")
534 534 else:
535 535 # looks good so far
536 536 dirmove[dsrc + "/"] = ddst + "/"
537 537
538 538 for i in invalid:
539 539 if i in dirmove:
540 540 del dirmove[i]
541 541 del d1, d2, invalid
542 542
543 543 if not dirmove:
544 544 return copy, {}, diverge, renamedelete
545 545
546 546 for d in dirmove:
547 547 repo.ui.debug(" discovered dir src: '%s' -> dst: '%s'\n" %
548 548 (d, dirmove[d]))
549 549
550 550 movewithdir = {}
551 551 # check unaccounted nonoverlapping files against directory moves
552 552 for f in u1r + u2r:
553 553 if f not in fullcopy:
554 554 for d in dirmove:
555 555 if f.startswith(d):
556 556 # new file added in a directory that was moved, move it
557 557 df = dirmove[d] + f[len(d):]
558 558 if df not in copy:
559 559 movewithdir[f] = df
560 560 repo.ui.debug((" pending file src: '%s' -> "
561 561 "dst: '%s'\n") % (f, df))
562 562 break
563 563
564 564 return copy, movewithdir, diverge, renamedelete
565 565
566 566 def _related(f1, f2, limit):
567 567 """return True if f1 and f2 filectx have a common ancestor
568 568
569 569 Walk back to common ancestor to see if the two files originate
570 570 from the same file. Since workingfilectx's rev() is None it messes
571 571 up the integer comparison logic, hence the pre-step check for
572 572 None (f1 and f2 can only be workingfilectx's initially).
573 573 """
574 574
575 575 if f1 == f2:
576 576 return f1 # a match
577 577
578 578 g1, g2 = f1.ancestors(), f2.ancestors()
579 579 try:
580 580 f1r, f2r = f1.linkrev(), f2.linkrev()
581 581
582 582 if f1r is None:
583 583 f1 = next(g1)
584 584 if f2r is None:
585 585 f2 = next(g2)
586 586
587 587 while True:
588 588 f1r, f2r = f1.linkrev(), f2.linkrev()
589 589 if f1r > f2r:
590 590 f1 = next(g1)
591 591 elif f2r > f1r:
592 592 f2 = next(g2)
593 593 elif f1 == f2:
594 594 return f1 # a match
595 595 elif f1r == f2r or f1r < limit or f2r < limit:
596 596 return False # copy no longer relevant
597 597 except StopIteration:
598 598 return False
599 599
600 600 def _checkcopies(ctx, f, m1, m2, base, tca, remotebase, limit, data):
601 601 """
602 602 check possible copies of f from m1 to m2
603 603
604 604 ctx = starting context for f in m1
605 605 f = the filename to check (as in m1)
606 606 m1 = the source manifest
607 607 m2 = the destination manifest
608 608 base = the changectx used as a merge base
609 609 tca = topological common ancestor for graft-like scenarios
610 610 remotebase = True if base is outside tca::ctx, False otherwise
611 611 limit = the rev number to not search beyond
612 612 data = dictionary of dictionary to store copy data. (see mergecopies)
613 613
614 614 note: limit is only an optimization, and there is no guarantee that
615 615 irrelevant revisions will not be limited
616 616 there is no easy way to make this algorithm stop in a guaranteed way
617 617 once it "goes behind a certain revision".
618 618 """
619 619
620 620 mb = base.manifest()
621 621 mta = tca.manifest()
622 622 # Might be true if this call is about finding backward renames,
623 623 # This happens in the case of grafts because the DAG is then rotated.
624 624 # If the file exists in both the base and the source, we are not looking
625 625 # for a rename on the source side, but on the part of the DAG that is
626 626 # traversed backwards.
627 627 #
628 628 # In the case there is both backward and forward renames (before and after
629 629 # the base) this is more complicated as we must detect a divergence.
630 630 # We use 'backwards = False' in that case.
631 631 backwards = not remotebase and base != tca and f in mb
632 632 getfctx = _makegetfctx(ctx)
633 633
634 634 if m1[f] == mb.get(f) and not remotebase:
635 635 # Nothing to merge
636 636 return
637 637
638 638 of = None
639 639 seen = set([f])
640 640 for oc in getfctx(f, m1[f]).ancestors():
641 641 ocr = oc.linkrev()
642 642 of = oc.path()
643 643 if of in seen:
644 644 # check limit late - grab last rename before
645 645 if ocr < limit:
646 646 break
647 647 continue
648 648 seen.add(of)
649 649
650 650 # remember for dir rename detection
651 651 if backwards:
652 652 data['fullcopy'][of] = f # grafting backwards through renames
653 653 else:
654 654 data['fullcopy'][f] = of
655 655 if of not in m2:
656 656 continue # no match, keep looking
657 657 if m2[of] == mb.get(of):
658 658 return # no merge needed, quit early
659 659 c2 = getfctx(of, m2[of])
660 660 # c2 might be a plain new file on added on destination side that is
661 661 # unrelated to the droids we are looking for.
662 662 cr = _related(oc, c2, tca.rev())
663 663 if cr and (of == f or of == c2.path()): # non-divergent
664 664 if backwards:
665 665 data['copy'][of] = f
666 666 elif of in mb:
667 667 data['copy'][f] = of
668 668 elif remotebase: # special case: a <- b <- a -> b "ping-pong" rename
669 669 data['copy'][of] = f
670 670 del data['fullcopy'][f]
671 671 data['fullcopy'][of] = f
672 672 else: # divergence w.r.t. graft CA on one side of topological CA
673 673 for sf in seen:
674 674 if sf in mb:
675 675 assert sf not in data['diverge']
676 676 data['diverge'][sf] = [f, of]
677 677 break
678 678 return
679 679
680 680 if of in mta:
681 681 if backwards or remotebase:
682 682 data['incomplete'][of] = f
683 683 else:
684 684 for sf in seen:
685 685 if sf in mb:
686 686 if tca == base:
687 687 data['diverge'].setdefault(sf, []).append(f)
688 688 else:
689 689 data['incompletediverge'][sf] = [of, f]
690 690 return
691 691
692 692 def duplicatecopies(repo, rev, fromrev, skiprev=None):
693 693 '''reproduce copies from fromrev to rev in the dirstate
694 694
695 695 If skiprev is specified, it's a revision that should be used to
696 696 filter copy records. Any copies that occur between fromrev and
697 697 skiprev will not be duplicated, even if they appear in the set of
698 698 copies between fromrev and rev.
699 699 '''
700 700 exclude = {}
701 701 if (skiprev is not None and
702 702 not repo.ui.configbool('experimental', 'disablecopytrace')):
703 703 # disablecopytrace skips this line, but not the entire function because
704 704 # the line below is O(size of the repo) during a rebase, while the rest
705 705 # of the function is much faster (and is required for carrying copy
706 706 # metadata across the rebase anyway).
707 707 exclude = pathcopies(repo[fromrev], repo[skiprev])
708 708 for dst, src in pathcopies(repo[fromrev], repo[rev]).iteritems():
709 709 # copies.pathcopies returns backward renames, so dst might not
710 710 # actually be in the dirstate
711 711 if dst in exclude:
712 712 continue
713 713 if repo.dirstate[dst] in "nma":
714 714 repo.dirstate.copy(src, dst)
@@ -1,26 +1,32
1 1 # node.py - basic nodeid manipulation for mercurial
2 2 #
3 3 # Copyright 2005, 2006 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import binascii
11 11
12 12 # This ugly style has a noticeable effect in manifest parsing
13 13 hex = binascii.hexlify
14 14 bin = binascii.unhexlify
15 15
16 16 nullrev = -1
17 17 nullid = b"\0" * 20
18 18 nullhex = hex(nullid)
19 19
20 # Phony node value to stand-in for new files in some uses of
21 # manifests.
22 newnodeid = '!' * 20
23
24 wdirnodes = set((newnodeid,))
25
20 26 # pseudo identifiers for working directory
21 27 # (they are experimental, so don't add too many dependencies on them)
22 28 wdirrev = 0x7fffffff
23 29 wdirid = b"\xff" * 20
24 30
25 31 def short(node):
26 32 return hex(node[:6])
General Comments 0
You need to be logged in to leave comments. Login now