##// END OF EJS Templates
journal: set Abort hint when failing due to an abandoned transaction
Johan Bjork -
r21274:3b4c7569 default
parent child Browse files
Show More
@@ -1,1912 +1,1913
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7 from node import hex, nullid, short
8 8 from i18n import _
9 9 import urllib
10 10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
11 11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
12 12 import lock as lockmod
13 13 import transaction, store, encoding, exchange, bundle2
14 14 import scmutil, util, extensions, hook, error, revset
15 15 import match as matchmod
16 16 import merge as mergemod
17 17 import tags as tagsmod
18 18 from lock import release
19 19 import weakref, errno, os, time, inspect
20 20 import branchmap, pathutil
21 21 propertycache = util.propertycache
22 22 filecache = scmutil.filecache
23 23
24 24 class repofilecache(filecache):
25 25 """All filecache usage on repo are done for logic that should be unfiltered
26 26 """
27 27
28 28 def __get__(self, repo, type=None):
29 29 return super(repofilecache, self).__get__(repo.unfiltered(), type)
30 30 def __set__(self, repo, value):
31 31 return super(repofilecache, self).__set__(repo.unfiltered(), value)
32 32 def __delete__(self, repo):
33 33 return super(repofilecache, self).__delete__(repo.unfiltered())
34 34
35 35 class storecache(repofilecache):
36 36 """filecache for files in the store"""
37 37 def join(self, obj, fname):
38 38 return obj.sjoin(fname)
39 39
40 40 class unfilteredpropertycache(propertycache):
41 41 """propertycache that apply to unfiltered repo only"""
42 42
43 43 def __get__(self, repo, type=None):
44 44 unfi = repo.unfiltered()
45 45 if unfi is repo:
46 46 return super(unfilteredpropertycache, self).__get__(unfi)
47 47 return getattr(unfi, self.name)
48 48
49 49 class filteredpropertycache(propertycache):
50 50 """propertycache that must take filtering in account"""
51 51
52 52 def cachevalue(self, obj, value):
53 53 object.__setattr__(obj, self.name, value)
54 54
55 55
56 56 def hasunfilteredcache(repo, name):
57 57 """check if a repo has an unfilteredpropertycache value for <name>"""
58 58 return name in vars(repo.unfiltered())
59 59
60 60 def unfilteredmethod(orig):
61 61 """decorate method that always need to be run on unfiltered version"""
62 62 def wrapper(repo, *args, **kwargs):
63 63 return orig(repo.unfiltered(), *args, **kwargs)
64 64 return wrapper
65 65
66 66 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
67 67 'unbundle'))
68 68 legacycaps = moderncaps.union(set(['changegroupsubset']))
69 69
70 70 class localpeer(peer.peerrepository):
71 71 '''peer for a local repo; reflects only the most recent API'''
72 72
73 73 def __init__(self, repo, caps=moderncaps):
74 74 peer.peerrepository.__init__(self)
75 75 self._repo = repo.filtered('served')
76 76 self.ui = repo.ui
77 77 self._caps = repo._restrictcapabilities(caps)
78 78 self.requirements = repo.requirements
79 79 self.supportedformats = repo.supportedformats
80 80
81 81 def close(self):
82 82 self._repo.close()
83 83
84 84 def _capabilities(self):
85 85 return self._caps
86 86
87 87 def local(self):
88 88 return self._repo
89 89
90 90 def canpush(self):
91 91 return True
92 92
93 93 def url(self):
94 94 return self._repo.url()
95 95
96 96 def lookup(self, key):
97 97 return self._repo.lookup(key)
98 98
99 99 def branchmap(self):
100 100 return self._repo.branchmap()
101 101
102 102 def heads(self):
103 103 return self._repo.heads()
104 104
105 105 def known(self, nodes):
106 106 return self._repo.known(nodes)
107 107
108 108 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
109 109 format='HG10', **kwargs):
110 110 cg = exchange.getbundle(self._repo, source, heads=heads,
111 111 common=common, bundlecaps=bundlecaps, **kwargs)
112 112 if bundlecaps is not None and 'HG2X' in bundlecaps:
113 113 # When requesting a bundle2, getbundle returns a stream to make the
114 114 # wire level function happier. We need to build a proper object
115 115 # from it in local peer.
116 116 cg = bundle2.unbundle20(self.ui, cg)
117 117 return cg
118 118
119 119 # TODO We might want to move the next two calls into legacypeer and add
120 120 # unbundle instead.
121 121
122 122 def unbundle(self, cg, heads, url):
123 123 """apply a bundle on a repo
124 124
125 125 This function handles the repo locking itself."""
126 126 try:
127 127 cg = exchange.readbundle(self.ui, cg, None)
128 128 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
129 129 if util.safehasattr(ret, 'getchunks'):
130 130 # This is a bundle20 object, turn it into an unbundler.
131 131 # This little dance should be dropped eventually when the API
132 132 # is finally improved.
133 133 stream = util.chunkbuffer(ret.getchunks())
134 134 ret = bundle2.unbundle20(self.ui, stream)
135 135 return ret
136 136 except error.PushRaced, exc:
137 137 raise error.ResponseError(_('push failed:'), str(exc))
138 138
139 139 def lock(self):
140 140 return self._repo.lock()
141 141
142 142 def addchangegroup(self, cg, source, url):
143 143 return changegroup.addchangegroup(self._repo, cg, source, url)
144 144
145 145 def pushkey(self, namespace, key, old, new):
146 146 return self._repo.pushkey(namespace, key, old, new)
147 147
148 148 def listkeys(self, namespace):
149 149 return self._repo.listkeys(namespace)
150 150
151 151 def debugwireargs(self, one, two, three=None, four=None, five=None):
152 152 '''used to test argument passing over the wire'''
153 153 return "%s %s %s %s %s" % (one, two, three, four, five)
154 154
155 155 class locallegacypeer(localpeer):
156 156 '''peer extension which implements legacy methods too; used for tests with
157 157 restricted capabilities'''
158 158
159 159 def __init__(self, repo):
160 160 localpeer.__init__(self, repo, caps=legacycaps)
161 161
162 162 def branches(self, nodes):
163 163 return self._repo.branches(nodes)
164 164
165 165 def between(self, pairs):
166 166 return self._repo.between(pairs)
167 167
168 168 def changegroup(self, basenodes, source):
169 169 return changegroup.changegroup(self._repo, basenodes, source)
170 170
171 171 def changegroupsubset(self, bases, heads, source):
172 172 return changegroup.changegroupsubset(self._repo, bases, heads, source)
173 173
174 174 class localrepository(object):
175 175
176 176 supportedformats = set(('revlogv1', 'generaldelta'))
177 177 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
178 178 'dotencode'))
179 179 openerreqs = set(('revlogv1', 'generaldelta'))
180 180 requirements = ['revlogv1']
181 181 filtername = None
182 182
183 183 bundle2caps = {'HG2X': ()}
184 184
185 185 # a list of (ui, featureset) functions.
186 186 # only functions defined in module of enabled extensions are invoked
187 187 featuresetupfuncs = set()
188 188
189 189 def _baserequirements(self, create):
190 190 return self.requirements[:]
191 191
192 192 def __init__(self, baseui, path=None, create=False):
193 193 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
194 194 self.wopener = self.wvfs
195 195 self.root = self.wvfs.base
196 196 self.path = self.wvfs.join(".hg")
197 197 self.origroot = path
198 198 self.auditor = pathutil.pathauditor(self.root, self._checknested)
199 199 self.vfs = scmutil.vfs(self.path)
200 200 self.opener = self.vfs
201 201 self.baseui = baseui
202 202 self.ui = baseui.copy()
203 203 self.ui.copy = baseui.copy # prevent copying repo configuration
204 204 # A list of callback to shape the phase if no data were found.
205 205 # Callback are in the form: func(repo, roots) --> processed root.
206 206 # This list it to be filled by extension during repo setup
207 207 self._phasedefaults = []
208 208 try:
209 209 self.ui.readconfig(self.join("hgrc"), self.root)
210 210 extensions.loadall(self.ui)
211 211 except IOError:
212 212 pass
213 213
214 214 if self.featuresetupfuncs:
215 215 self.supported = set(self._basesupported) # use private copy
216 216 extmods = set(m.__name__ for n, m
217 217 in extensions.extensions(self.ui))
218 218 for setupfunc in self.featuresetupfuncs:
219 219 if setupfunc.__module__ in extmods:
220 220 setupfunc(self.ui, self.supported)
221 221 else:
222 222 self.supported = self._basesupported
223 223
224 224 if not self.vfs.isdir():
225 225 if create:
226 226 if not self.wvfs.exists():
227 227 self.wvfs.makedirs()
228 228 self.vfs.makedir(notindexed=True)
229 229 requirements = self._baserequirements(create)
230 230 if self.ui.configbool('format', 'usestore', True):
231 231 self.vfs.mkdir("store")
232 232 requirements.append("store")
233 233 if self.ui.configbool('format', 'usefncache', True):
234 234 requirements.append("fncache")
235 235 if self.ui.configbool('format', 'dotencode', True):
236 236 requirements.append('dotencode')
237 237 # create an invalid changelog
238 238 self.vfs.append(
239 239 "00changelog.i",
240 240 '\0\0\0\2' # represents revlogv2
241 241 ' dummy changelog to prevent using the old repo layout'
242 242 )
243 243 if self.ui.configbool('format', 'generaldelta', False):
244 244 requirements.append("generaldelta")
245 245 requirements = set(requirements)
246 246 else:
247 247 raise error.RepoError(_("repository %s not found") % path)
248 248 elif create:
249 249 raise error.RepoError(_("repository %s already exists") % path)
250 250 else:
251 251 try:
252 252 requirements = scmutil.readrequires(self.vfs, self.supported)
253 253 except IOError, inst:
254 254 if inst.errno != errno.ENOENT:
255 255 raise
256 256 requirements = set()
257 257
258 258 self.sharedpath = self.path
259 259 try:
260 260 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
261 261 realpath=True)
262 262 s = vfs.base
263 263 if not vfs.exists():
264 264 raise error.RepoError(
265 265 _('.hg/sharedpath points to nonexistent directory %s') % s)
266 266 self.sharedpath = s
267 267 except IOError, inst:
268 268 if inst.errno != errno.ENOENT:
269 269 raise
270 270
271 271 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
272 272 self.spath = self.store.path
273 273 self.svfs = self.store.vfs
274 274 self.sopener = self.svfs
275 275 self.sjoin = self.store.join
276 276 self.vfs.createmode = self.store.createmode
277 277 self._applyrequirements(requirements)
278 278 if create:
279 279 self._writerequirements()
280 280
281 281
282 282 self._branchcaches = {}
283 283 self.filterpats = {}
284 284 self._datafilters = {}
285 285 self._transref = self._lockref = self._wlockref = None
286 286
287 287 # A cache for various files under .hg/ that tracks file changes,
288 288 # (used by the filecache decorator)
289 289 #
290 290 # Maps a property name to its util.filecacheentry
291 291 self._filecache = {}
292 292
293 293 # hold sets of revision to be filtered
294 294 # should be cleared when something might have changed the filter value:
295 295 # - new changesets,
296 296 # - phase change,
297 297 # - new obsolescence marker,
298 298 # - working directory parent change,
299 299 # - bookmark changes
300 300 self.filteredrevcache = {}
301 301
302 302 def close(self):
303 303 pass
304 304
305 305 def _restrictcapabilities(self, caps):
306 306 # bundle2 is not ready for prime time, drop it unless explicitly
307 307 # required by the tests (or some brave tester)
308 308 if self.ui.configbool('experimental', 'bundle2-exp', False):
309 309 caps = set(caps)
310 310 capsblob = bundle2.encodecaps(self.bundle2caps)
311 311 caps.add('bundle2-exp=' + urllib.quote(capsblob))
312 312 return caps
313 313
314 314 def _applyrequirements(self, requirements):
315 315 self.requirements = requirements
316 316 self.sopener.options = dict((r, 1) for r in requirements
317 317 if r in self.openerreqs)
318 318 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
319 319 if chunkcachesize is not None:
320 320 self.sopener.options['chunkcachesize'] = chunkcachesize
321 321
322 322 def _writerequirements(self):
323 323 reqfile = self.opener("requires", "w")
324 324 for r in sorted(self.requirements):
325 325 reqfile.write("%s\n" % r)
326 326 reqfile.close()
327 327
328 328 def _checknested(self, path):
329 329 """Determine if path is a legal nested repository."""
330 330 if not path.startswith(self.root):
331 331 return False
332 332 subpath = path[len(self.root) + 1:]
333 333 normsubpath = util.pconvert(subpath)
334 334
335 335 # XXX: Checking against the current working copy is wrong in
336 336 # the sense that it can reject things like
337 337 #
338 338 # $ hg cat -r 10 sub/x.txt
339 339 #
340 340 # if sub/ is no longer a subrepository in the working copy
341 341 # parent revision.
342 342 #
343 343 # However, it can of course also allow things that would have
344 344 # been rejected before, such as the above cat command if sub/
345 345 # is a subrepository now, but was a normal directory before.
346 346 # The old path auditor would have rejected by mistake since it
347 347 # panics when it sees sub/.hg/.
348 348 #
349 349 # All in all, checking against the working copy seems sensible
350 350 # since we want to prevent access to nested repositories on
351 351 # the filesystem *now*.
352 352 ctx = self[None]
353 353 parts = util.splitpath(subpath)
354 354 while parts:
355 355 prefix = '/'.join(parts)
356 356 if prefix in ctx.substate:
357 357 if prefix == normsubpath:
358 358 return True
359 359 else:
360 360 sub = ctx.sub(prefix)
361 361 return sub.checknested(subpath[len(prefix) + 1:])
362 362 else:
363 363 parts.pop()
364 364 return False
365 365
366 366 def peer(self):
367 367 return localpeer(self) # not cached to avoid reference cycle
368 368
369 369 def unfiltered(self):
370 370 """Return unfiltered version of the repository
371 371
372 372 Intended to be overwritten by filtered repo."""
373 373 return self
374 374
375 375 def filtered(self, name):
376 376 """Return a filtered version of a repository"""
377 377 # build a new class with the mixin and the current class
378 378 # (possibly subclass of the repo)
379 379 class proxycls(repoview.repoview, self.unfiltered().__class__):
380 380 pass
381 381 return proxycls(self, name)
382 382
383 383 @repofilecache('bookmarks')
384 384 def _bookmarks(self):
385 385 return bookmarks.bmstore(self)
386 386
387 387 @repofilecache('bookmarks.current')
388 388 def _bookmarkcurrent(self):
389 389 return bookmarks.readcurrent(self)
390 390
391 391 def bookmarkheads(self, bookmark):
392 392 name = bookmark.split('@', 1)[0]
393 393 heads = []
394 394 for mark, n in self._bookmarks.iteritems():
395 395 if mark.split('@', 1)[0] == name:
396 396 heads.append(n)
397 397 return heads
398 398
399 399 @storecache('phaseroots')
400 400 def _phasecache(self):
401 401 return phases.phasecache(self, self._phasedefaults)
402 402
403 403 @storecache('obsstore')
404 404 def obsstore(self):
405 405 store = obsolete.obsstore(self.sopener)
406 406 if store and not obsolete._enabled:
407 407 # message is rare enough to not be translated
408 408 msg = 'obsolete feature not enabled but %i markers found!\n'
409 409 self.ui.warn(msg % len(list(store)))
410 410 return store
411 411
412 412 @storecache('00changelog.i')
413 413 def changelog(self):
414 414 c = changelog.changelog(self.sopener)
415 415 if 'HG_PENDING' in os.environ:
416 416 p = os.environ['HG_PENDING']
417 417 if p.startswith(self.root):
418 418 c.readpending('00changelog.i.a')
419 419 return c
420 420
421 421 @storecache('00manifest.i')
422 422 def manifest(self):
423 423 return manifest.manifest(self.sopener)
424 424
425 425 @repofilecache('dirstate')
426 426 def dirstate(self):
427 427 warned = [0]
428 428 def validate(node):
429 429 try:
430 430 self.changelog.rev(node)
431 431 return node
432 432 except error.LookupError:
433 433 if not warned[0]:
434 434 warned[0] = True
435 435 self.ui.warn(_("warning: ignoring unknown"
436 436 " working parent %s!\n") % short(node))
437 437 return nullid
438 438
439 439 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
440 440
441 441 def __getitem__(self, changeid):
442 442 if changeid is None:
443 443 return context.workingctx(self)
444 444 return context.changectx(self, changeid)
445 445
446 446 def __contains__(self, changeid):
447 447 try:
448 448 return bool(self.lookup(changeid))
449 449 except error.RepoLookupError:
450 450 return False
451 451
452 452 def __nonzero__(self):
453 453 return True
454 454
455 455 def __len__(self):
456 456 return len(self.changelog)
457 457
458 458 def __iter__(self):
459 459 return iter(self.changelog)
460 460
461 461 def revs(self, expr, *args):
462 462 '''Return a list of revisions matching the given revset'''
463 463 expr = revset.formatspec(expr, *args)
464 464 m = revset.match(None, expr)
465 465 return m(self, revset.spanset(self))
466 466
467 467 def set(self, expr, *args):
468 468 '''
469 469 Yield a context for each matching revision, after doing arg
470 470 replacement via revset.formatspec
471 471 '''
472 472 for r in self.revs(expr, *args):
473 473 yield self[r]
474 474
475 475 def url(self):
476 476 return 'file:' + self.root
477 477
478 478 def hook(self, name, throw=False, **args):
479 479 return hook.hook(self.ui, self, name, throw, **args)
480 480
481 481 @unfilteredmethod
482 482 def _tag(self, names, node, message, local, user, date, extra={},
483 483 editor=False):
484 484 if isinstance(names, str):
485 485 names = (names,)
486 486
487 487 branches = self.branchmap()
488 488 for name in names:
489 489 self.hook('pretag', throw=True, node=hex(node), tag=name,
490 490 local=local)
491 491 if name in branches:
492 492 self.ui.warn(_("warning: tag %s conflicts with existing"
493 493 " branch name\n") % name)
494 494
495 495 def writetags(fp, names, munge, prevtags):
496 496 fp.seek(0, 2)
497 497 if prevtags and prevtags[-1] != '\n':
498 498 fp.write('\n')
499 499 for name in names:
500 500 m = munge and munge(name) or name
501 501 if (self._tagscache.tagtypes and
502 502 name in self._tagscache.tagtypes):
503 503 old = self.tags().get(name, nullid)
504 504 fp.write('%s %s\n' % (hex(old), m))
505 505 fp.write('%s %s\n' % (hex(node), m))
506 506 fp.close()
507 507
508 508 prevtags = ''
509 509 if local:
510 510 try:
511 511 fp = self.opener('localtags', 'r+')
512 512 except IOError:
513 513 fp = self.opener('localtags', 'a')
514 514 else:
515 515 prevtags = fp.read()
516 516
517 517 # local tags are stored in the current charset
518 518 writetags(fp, names, None, prevtags)
519 519 for name in names:
520 520 self.hook('tag', node=hex(node), tag=name, local=local)
521 521 return
522 522
523 523 try:
524 524 fp = self.wfile('.hgtags', 'rb+')
525 525 except IOError, e:
526 526 if e.errno != errno.ENOENT:
527 527 raise
528 528 fp = self.wfile('.hgtags', 'ab')
529 529 else:
530 530 prevtags = fp.read()
531 531
532 532 # committed tags are stored in UTF-8
533 533 writetags(fp, names, encoding.fromlocal, prevtags)
534 534
535 535 fp.close()
536 536
537 537 self.invalidatecaches()
538 538
539 539 if '.hgtags' not in self.dirstate:
540 540 self[None].add(['.hgtags'])
541 541
542 542 m = matchmod.exact(self.root, '', ['.hgtags'])
543 543 tagnode = self.commit(message, user, date, extra=extra, match=m,
544 544 editor=editor)
545 545
546 546 for name in names:
547 547 self.hook('tag', node=hex(node), tag=name, local=local)
548 548
549 549 return tagnode
550 550
551 551 def tag(self, names, node, message, local, user, date, editor=False):
552 552 '''tag a revision with one or more symbolic names.
553 553
554 554 names is a list of strings or, when adding a single tag, names may be a
555 555 string.
556 556
557 557 if local is True, the tags are stored in a per-repository file.
558 558 otherwise, they are stored in the .hgtags file, and a new
559 559 changeset is committed with the change.
560 560
561 561 keyword arguments:
562 562
563 563 local: whether to store tags in non-version-controlled file
564 564 (default False)
565 565
566 566 message: commit message to use if committing
567 567
568 568 user: name of user to use if committing
569 569
570 570 date: date tuple to use if committing'''
571 571
572 572 if not local:
573 573 for x in self.status()[:5]:
574 574 if '.hgtags' in x:
575 575 raise util.Abort(_('working copy of .hgtags is changed '
576 576 '(please commit .hgtags manually)'))
577 577
578 578 self.tags() # instantiate the cache
579 579 self._tag(names, node, message, local, user, date, editor=editor)
580 580
581 581 @filteredpropertycache
582 582 def _tagscache(self):
583 583 '''Returns a tagscache object that contains various tags related
584 584 caches.'''
585 585
586 586 # This simplifies its cache management by having one decorated
587 587 # function (this one) and the rest simply fetch things from it.
588 588 class tagscache(object):
589 589 def __init__(self):
590 590 # These two define the set of tags for this repository. tags
591 591 # maps tag name to node; tagtypes maps tag name to 'global' or
592 592 # 'local'. (Global tags are defined by .hgtags across all
593 593 # heads, and local tags are defined in .hg/localtags.)
594 594 # They constitute the in-memory cache of tags.
595 595 self.tags = self.tagtypes = None
596 596
597 597 self.nodetagscache = self.tagslist = None
598 598
599 599 cache = tagscache()
600 600 cache.tags, cache.tagtypes = self._findtags()
601 601
602 602 return cache
603 603
604 604 def tags(self):
605 605 '''return a mapping of tag to node'''
606 606 t = {}
607 607 if self.changelog.filteredrevs:
608 608 tags, tt = self._findtags()
609 609 else:
610 610 tags = self._tagscache.tags
611 611 for k, v in tags.iteritems():
612 612 try:
613 613 # ignore tags to unknown nodes
614 614 self.changelog.rev(v)
615 615 t[k] = v
616 616 except (error.LookupError, ValueError):
617 617 pass
618 618 return t
619 619
620 620 def _findtags(self):
621 621 '''Do the hard work of finding tags. Return a pair of dicts
622 622 (tags, tagtypes) where tags maps tag name to node, and tagtypes
623 623 maps tag name to a string like \'global\' or \'local\'.
624 624 Subclasses or extensions are free to add their own tags, but
625 625 should be aware that the returned dicts will be retained for the
626 626 duration of the localrepo object.'''
627 627
628 628 # XXX what tagtype should subclasses/extensions use? Currently
629 629 # mq and bookmarks add tags, but do not set the tagtype at all.
630 630 # Should each extension invent its own tag type? Should there
631 631 # be one tagtype for all such "virtual" tags? Or is the status
632 632 # quo fine?
633 633
634 634 alltags = {} # map tag name to (node, hist)
635 635 tagtypes = {}
636 636
637 637 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
638 638 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
639 639
640 640 # Build the return dicts. Have to re-encode tag names because
641 641 # the tags module always uses UTF-8 (in order not to lose info
642 642 # writing to the cache), but the rest of Mercurial wants them in
643 643 # local encoding.
644 644 tags = {}
645 645 for (name, (node, hist)) in alltags.iteritems():
646 646 if node != nullid:
647 647 tags[encoding.tolocal(name)] = node
648 648 tags['tip'] = self.changelog.tip()
649 649 tagtypes = dict([(encoding.tolocal(name), value)
650 650 for (name, value) in tagtypes.iteritems()])
651 651 return (tags, tagtypes)
652 652
653 653 def tagtype(self, tagname):
654 654 '''
655 655 return the type of the given tag. result can be:
656 656
657 657 'local' : a local tag
658 658 'global' : a global tag
659 659 None : tag does not exist
660 660 '''
661 661
662 662 return self._tagscache.tagtypes.get(tagname)
663 663
664 664 def tagslist(self):
665 665 '''return a list of tags ordered by revision'''
666 666 if not self._tagscache.tagslist:
667 667 l = []
668 668 for t, n in self.tags().iteritems():
669 669 r = self.changelog.rev(n)
670 670 l.append((r, t, n))
671 671 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
672 672
673 673 return self._tagscache.tagslist
674 674
675 675 def nodetags(self, node):
676 676 '''return the tags associated with a node'''
677 677 if not self._tagscache.nodetagscache:
678 678 nodetagscache = {}
679 679 for t, n in self._tagscache.tags.iteritems():
680 680 nodetagscache.setdefault(n, []).append(t)
681 681 for tags in nodetagscache.itervalues():
682 682 tags.sort()
683 683 self._tagscache.nodetagscache = nodetagscache
684 684 return self._tagscache.nodetagscache.get(node, [])
685 685
686 686 def nodebookmarks(self, node):
687 687 marks = []
688 688 for bookmark, n in self._bookmarks.iteritems():
689 689 if n == node:
690 690 marks.append(bookmark)
691 691 return sorted(marks)
692 692
693 693 def branchmap(self):
694 694 '''returns a dictionary {branch: [branchheads]} with branchheads
695 695 ordered by increasing revision number'''
696 696 branchmap.updatecache(self)
697 697 return self._branchcaches[self.filtername]
698 698
699 699 def branchtip(self, branch):
700 700 '''return the tip node for a given branch'''
701 701 try:
702 702 return self.branchmap().branchtip(branch)
703 703 except KeyError:
704 704 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
705 705
706 706 def lookup(self, key):
707 707 return self[key].node()
708 708
709 709 def lookupbranch(self, key, remote=None):
710 710 repo = remote or self
711 711 if key in repo.branchmap():
712 712 return key
713 713
714 714 repo = (remote and remote.local()) and remote or self
715 715 return repo[key].branch()
716 716
717 717 def known(self, nodes):
718 718 nm = self.changelog.nodemap
719 719 pc = self._phasecache
720 720 result = []
721 721 for n in nodes:
722 722 r = nm.get(n)
723 723 resp = not (r is None or pc.phase(self, r) >= phases.secret)
724 724 result.append(resp)
725 725 return result
726 726
727 727 def local(self):
728 728 return self
729 729
730 730 def cancopy(self):
731 731 # so statichttprepo's override of local() works
732 732 if not self.local():
733 733 return False
734 734 if not self.ui.configbool('phases', 'publish', True):
735 735 return True
736 736 # if publishing we can't copy if there is filtered content
737 737 return not self.filtered('visible').changelog.filteredrevs
738 738
739 739 def join(self, f):
740 740 return os.path.join(self.path, f)
741 741
742 742 def wjoin(self, f):
743 743 return os.path.join(self.root, f)
744 744
745 745 def file(self, f):
746 746 if f[0] == '/':
747 747 f = f[1:]
748 748 return filelog.filelog(self.sopener, f)
749 749
750 750 def changectx(self, changeid):
751 751 return self[changeid]
752 752
753 753 def parents(self, changeid=None):
754 754 '''get list of changectxs for parents of changeid'''
755 755 return self[changeid].parents()
756 756
757 757 def setparents(self, p1, p2=nullid):
758 758 copies = self.dirstate.setparents(p1, p2)
759 759 pctx = self[p1]
760 760 if copies:
761 761 # Adjust copy records, the dirstate cannot do it, it
762 762 # requires access to parents manifests. Preserve them
763 763 # only for entries added to first parent.
764 764 for f in copies:
765 765 if f not in pctx and copies[f] in pctx:
766 766 self.dirstate.copy(copies[f], f)
767 767 if p2 == nullid:
768 768 for f, s in sorted(self.dirstate.copies().items()):
769 769 if f not in pctx and s not in pctx:
770 770 self.dirstate.copy(None, f)
771 771
772 772 def filectx(self, path, changeid=None, fileid=None):
773 773 """changeid can be a changeset revision, node, or tag.
774 774 fileid can be a file revision or node."""
775 775 return context.filectx(self, path, changeid, fileid)
776 776
777 777 def getcwd(self):
778 778 return self.dirstate.getcwd()
779 779
780 780 def pathto(self, f, cwd=None):
781 781 return self.dirstate.pathto(f, cwd)
782 782
783 783 def wfile(self, f, mode='r'):
784 784 return self.wopener(f, mode)
785 785
786 786 def _link(self, f):
787 787 return self.wvfs.islink(f)
788 788
789 789 def _loadfilter(self, filter):
790 790 if filter not in self.filterpats:
791 791 l = []
792 792 for pat, cmd in self.ui.configitems(filter):
793 793 if cmd == '!':
794 794 continue
795 795 mf = matchmod.match(self.root, '', [pat])
796 796 fn = None
797 797 params = cmd
798 798 for name, filterfn in self._datafilters.iteritems():
799 799 if cmd.startswith(name):
800 800 fn = filterfn
801 801 params = cmd[len(name):].lstrip()
802 802 break
803 803 if not fn:
804 804 fn = lambda s, c, **kwargs: util.filter(s, c)
805 805 # Wrap old filters not supporting keyword arguments
806 806 if not inspect.getargspec(fn)[2]:
807 807 oldfn = fn
808 808 fn = lambda s, c, **kwargs: oldfn(s, c)
809 809 l.append((mf, fn, params))
810 810 self.filterpats[filter] = l
811 811 return self.filterpats[filter]
812 812
813 813 def _filter(self, filterpats, filename, data):
814 814 for mf, fn, cmd in filterpats:
815 815 if mf(filename):
816 816 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
817 817 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
818 818 break
819 819
820 820 return data
821 821
822 822 @unfilteredpropertycache
823 823 def _encodefilterpats(self):
824 824 return self._loadfilter('encode')
825 825
826 826 @unfilteredpropertycache
827 827 def _decodefilterpats(self):
828 828 return self._loadfilter('decode')
829 829
830 830 def adddatafilter(self, name, filter):
831 831 self._datafilters[name] = filter
832 832
833 833 def wread(self, filename):
834 834 if self._link(filename):
835 835 data = self.wvfs.readlink(filename)
836 836 else:
837 837 data = self.wopener.read(filename)
838 838 return self._filter(self._encodefilterpats, filename, data)
839 839
840 840 def wwrite(self, filename, data, flags):
841 841 data = self._filter(self._decodefilterpats, filename, data)
842 842 if 'l' in flags:
843 843 self.wopener.symlink(data, filename)
844 844 else:
845 845 self.wopener.write(filename, data)
846 846 if 'x' in flags:
847 847 self.wvfs.setflags(filename, False, True)
848 848
849 849 def wwritedata(self, filename, data):
850 850 return self._filter(self._decodefilterpats, filename, data)
851 851
852 852 def transaction(self, desc, report=None):
853 853 tr = self._transref and self._transref() or None
854 854 if tr and tr.running():
855 855 return tr.nest()
856 856
857 857 # abort here if the journal already exists
858 858 if self.svfs.exists("journal"):
859 859 raise error.RepoError(
860 _("abandoned transaction found - run hg recover"))
860 _("abandoned transaction found"),
861 hint=_("run 'hg recover' to clean up transaction"))
861 862
862 863 def onclose():
863 864 self.store.write(tr)
864 865
865 866 self._writejournal(desc)
866 867 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
867 868 rp = report and report or self.ui.warn
868 869 tr = transaction.transaction(rp, self.sopener,
869 870 "journal",
870 871 aftertrans(renames),
871 872 self.store.createmode,
872 873 onclose)
873 874 self._transref = weakref.ref(tr)
874 875 return tr
875 876
876 877 def _journalfiles(self):
877 878 return ((self.svfs, 'journal'),
878 879 (self.vfs, 'journal.dirstate'),
879 880 (self.vfs, 'journal.branch'),
880 881 (self.vfs, 'journal.desc'),
881 882 (self.vfs, 'journal.bookmarks'),
882 883 (self.svfs, 'journal.phaseroots'))
883 884
884 885 def undofiles(self):
885 886 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
886 887
887 888 def _writejournal(self, desc):
888 889 self.opener.write("journal.dirstate",
889 890 self.opener.tryread("dirstate"))
890 891 self.opener.write("journal.branch",
891 892 encoding.fromlocal(self.dirstate.branch()))
892 893 self.opener.write("journal.desc",
893 894 "%d\n%s\n" % (len(self), desc))
894 895 self.opener.write("journal.bookmarks",
895 896 self.opener.tryread("bookmarks"))
896 897 self.sopener.write("journal.phaseroots",
897 898 self.sopener.tryread("phaseroots"))
898 899
899 900 def recover(self):
900 901 lock = self.lock()
901 902 try:
902 903 if self.svfs.exists("journal"):
903 904 self.ui.status(_("rolling back interrupted transaction\n"))
904 905 transaction.rollback(self.sopener, "journal",
905 906 self.ui.warn)
906 907 self.invalidate()
907 908 return True
908 909 else:
909 910 self.ui.warn(_("no interrupted transaction available\n"))
910 911 return False
911 912 finally:
912 913 lock.release()
913 914
914 915 def rollback(self, dryrun=False, force=False):
915 916 wlock = lock = None
916 917 try:
917 918 wlock = self.wlock()
918 919 lock = self.lock()
919 920 if self.svfs.exists("undo"):
920 921 return self._rollback(dryrun, force)
921 922 else:
922 923 self.ui.warn(_("no rollback information available\n"))
923 924 return 1
924 925 finally:
925 926 release(lock, wlock)
926 927
927 928 @unfilteredmethod # Until we get smarter cache management
928 929 def _rollback(self, dryrun, force):
929 930 ui = self.ui
930 931 try:
931 932 args = self.opener.read('undo.desc').splitlines()
932 933 (oldlen, desc, detail) = (int(args[0]), args[1], None)
933 934 if len(args) >= 3:
934 935 detail = args[2]
935 936 oldtip = oldlen - 1
936 937
937 938 if detail and ui.verbose:
938 939 msg = (_('repository tip rolled back to revision %s'
939 940 ' (undo %s: %s)\n')
940 941 % (oldtip, desc, detail))
941 942 else:
942 943 msg = (_('repository tip rolled back to revision %s'
943 944 ' (undo %s)\n')
944 945 % (oldtip, desc))
945 946 except IOError:
946 947 msg = _('rolling back unknown transaction\n')
947 948 desc = None
948 949
949 950 if not force and self['.'] != self['tip'] and desc == 'commit':
950 951 raise util.Abort(
951 952 _('rollback of last commit while not checked out '
952 953 'may lose data'), hint=_('use -f to force'))
953 954
954 955 ui.status(msg)
955 956 if dryrun:
956 957 return 0
957 958
958 959 parents = self.dirstate.parents()
959 960 self.destroying()
960 961 transaction.rollback(self.sopener, 'undo', ui.warn)
961 962 if self.vfs.exists('undo.bookmarks'):
962 963 self.vfs.rename('undo.bookmarks', 'bookmarks')
963 964 if self.svfs.exists('undo.phaseroots'):
964 965 self.svfs.rename('undo.phaseroots', 'phaseroots')
965 966 self.invalidate()
966 967
967 968 parentgone = (parents[0] not in self.changelog.nodemap or
968 969 parents[1] not in self.changelog.nodemap)
969 970 if parentgone:
970 971 self.vfs.rename('undo.dirstate', 'dirstate')
971 972 try:
972 973 branch = self.opener.read('undo.branch')
973 974 self.dirstate.setbranch(encoding.tolocal(branch))
974 975 except IOError:
975 976 ui.warn(_('named branch could not be reset: '
976 977 'current branch is still \'%s\'\n')
977 978 % self.dirstate.branch())
978 979
979 980 self.dirstate.invalidate()
980 981 parents = tuple([p.rev() for p in self.parents()])
981 982 if len(parents) > 1:
982 983 ui.status(_('working directory now based on '
983 984 'revisions %d and %d\n') % parents)
984 985 else:
985 986 ui.status(_('working directory now based on '
986 987 'revision %d\n') % parents)
987 988 # TODO: if we know which new heads may result from this rollback, pass
988 989 # them to destroy(), which will prevent the branchhead cache from being
989 990 # invalidated.
990 991 self.destroyed()
991 992 return 0
992 993
993 994 def invalidatecaches(self):
994 995
995 996 if '_tagscache' in vars(self):
996 997 # can't use delattr on proxy
997 998 del self.__dict__['_tagscache']
998 999
999 1000 self.unfiltered()._branchcaches.clear()
1000 1001 self.invalidatevolatilesets()
1001 1002
1002 1003 def invalidatevolatilesets(self):
1003 1004 self.filteredrevcache.clear()
1004 1005 obsolete.clearobscaches(self)
1005 1006
1006 1007 def invalidatedirstate(self):
1007 1008 '''Invalidates the dirstate, causing the next call to dirstate
1008 1009 to check if it was modified since the last time it was read,
1009 1010 rereading it if it has.
1010 1011
1011 1012 This is different to dirstate.invalidate() that it doesn't always
1012 1013 rereads the dirstate. Use dirstate.invalidate() if you want to
1013 1014 explicitly read the dirstate again (i.e. restoring it to a previous
1014 1015 known good state).'''
1015 1016 if hasunfilteredcache(self, 'dirstate'):
1016 1017 for k in self.dirstate._filecache:
1017 1018 try:
1018 1019 delattr(self.dirstate, k)
1019 1020 except AttributeError:
1020 1021 pass
1021 1022 delattr(self.unfiltered(), 'dirstate')
1022 1023
1023 1024 def invalidate(self):
1024 1025 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1025 1026 for k in self._filecache:
1026 1027 # dirstate is invalidated separately in invalidatedirstate()
1027 1028 if k == 'dirstate':
1028 1029 continue
1029 1030
1030 1031 try:
1031 1032 delattr(unfiltered, k)
1032 1033 except AttributeError:
1033 1034 pass
1034 1035 self.invalidatecaches()
1035 1036 self.store.invalidatecaches()
1036 1037
1037 1038 def invalidateall(self):
1038 1039 '''Fully invalidates both store and non-store parts, causing the
1039 1040 subsequent operation to reread any outside changes.'''
1040 1041 # extension should hook this to invalidate its caches
1041 1042 self.invalidate()
1042 1043 self.invalidatedirstate()
1043 1044
1044 1045 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1045 1046 try:
1046 1047 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1047 1048 except error.LockHeld, inst:
1048 1049 if not wait:
1049 1050 raise
1050 1051 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1051 1052 (desc, inst.locker))
1052 1053 # default to 600 seconds timeout
1053 1054 l = lockmod.lock(vfs, lockname,
1054 1055 int(self.ui.config("ui", "timeout", "600")),
1055 1056 releasefn, desc=desc)
1056 1057 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1057 1058 if acquirefn:
1058 1059 acquirefn()
1059 1060 return l
1060 1061
1061 1062 def _afterlock(self, callback):
1062 1063 """add a callback to the current repository lock.
1063 1064
1064 1065 The callback will be executed on lock release."""
1065 1066 l = self._lockref and self._lockref()
1066 1067 if l:
1067 1068 l.postrelease.append(callback)
1068 1069 else:
1069 1070 callback()
1070 1071
1071 1072 def lock(self, wait=True):
1072 1073 '''Lock the repository store (.hg/store) and return a weak reference
1073 1074 to the lock. Use this before modifying the store (e.g. committing or
1074 1075 stripping). If you are opening a transaction, get a lock as well.)'''
1075 1076 l = self._lockref and self._lockref()
1076 1077 if l is not None and l.held:
1077 1078 l.lock()
1078 1079 return l
1079 1080
1080 1081 def unlock():
1081 1082 if hasunfilteredcache(self, '_phasecache'):
1082 1083 self._phasecache.write()
1083 1084 for k, ce in self._filecache.items():
1084 1085 if k == 'dirstate' or k not in self.__dict__:
1085 1086 continue
1086 1087 ce.refresh()
1087 1088
1088 1089 l = self._lock(self.svfs, "lock", wait, unlock,
1089 1090 self.invalidate, _('repository %s') % self.origroot)
1090 1091 self._lockref = weakref.ref(l)
1091 1092 return l
1092 1093
1093 1094 def wlock(self, wait=True):
1094 1095 '''Lock the non-store parts of the repository (everything under
1095 1096 .hg except .hg/store) and return a weak reference to the lock.
1096 1097 Use this before modifying files in .hg.'''
1097 1098 l = self._wlockref and self._wlockref()
1098 1099 if l is not None and l.held:
1099 1100 l.lock()
1100 1101 return l
1101 1102
1102 1103 def unlock():
1103 1104 self.dirstate.write()
1104 1105 self._filecache['dirstate'].refresh()
1105 1106
1106 1107 l = self._lock(self.vfs, "wlock", wait, unlock,
1107 1108 self.invalidatedirstate, _('working directory of %s') %
1108 1109 self.origroot)
1109 1110 self._wlockref = weakref.ref(l)
1110 1111 return l
1111 1112
1112 1113 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1113 1114 """
1114 1115 commit an individual file as part of a larger transaction
1115 1116 """
1116 1117
1117 1118 fname = fctx.path()
1118 1119 text = fctx.data()
1119 1120 flog = self.file(fname)
1120 1121 fparent1 = manifest1.get(fname, nullid)
1121 1122 fparent2 = fparent2o = manifest2.get(fname, nullid)
1122 1123
1123 1124 meta = {}
1124 1125 copy = fctx.renamed()
1125 1126 if copy and copy[0] != fname:
1126 1127 # Mark the new revision of this file as a copy of another
1127 1128 # file. This copy data will effectively act as a parent
1128 1129 # of this new revision. If this is a merge, the first
1129 1130 # parent will be the nullid (meaning "look up the copy data")
1130 1131 # and the second one will be the other parent. For example:
1131 1132 #
1132 1133 # 0 --- 1 --- 3 rev1 changes file foo
1133 1134 # \ / rev2 renames foo to bar and changes it
1134 1135 # \- 2 -/ rev3 should have bar with all changes and
1135 1136 # should record that bar descends from
1136 1137 # bar in rev2 and foo in rev1
1137 1138 #
1138 1139 # this allows this merge to succeed:
1139 1140 #
1140 1141 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1141 1142 # \ / merging rev3 and rev4 should use bar@rev2
1142 1143 # \- 2 --- 4 as the merge base
1143 1144 #
1144 1145
1145 1146 cfname = copy[0]
1146 1147 crev = manifest1.get(cfname)
1147 1148 newfparent = fparent2
1148 1149
1149 1150 if manifest2: # branch merge
1150 1151 if fparent2 == nullid or crev is None: # copied on remote side
1151 1152 if cfname in manifest2:
1152 1153 crev = manifest2[cfname]
1153 1154 newfparent = fparent1
1154 1155
1155 1156 # find source in nearest ancestor if we've lost track
1156 1157 if not crev:
1157 1158 self.ui.debug(" %s: searching for copy revision for %s\n" %
1158 1159 (fname, cfname))
1159 1160 for ancestor in self[None].ancestors():
1160 1161 if cfname in ancestor:
1161 1162 crev = ancestor[cfname].filenode()
1162 1163 break
1163 1164
1164 1165 if crev:
1165 1166 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1166 1167 meta["copy"] = cfname
1167 1168 meta["copyrev"] = hex(crev)
1168 1169 fparent1, fparent2 = nullid, newfparent
1169 1170 else:
1170 1171 self.ui.warn(_("warning: can't find ancestor for '%s' "
1171 1172 "copied from '%s'!\n") % (fname, cfname))
1172 1173
1173 1174 elif fparent1 == nullid:
1174 1175 fparent1, fparent2 = fparent2, nullid
1175 1176 elif fparent2 != nullid:
1176 1177 # is one parent an ancestor of the other?
1177 1178 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1178 1179 if fparent1 in fparentancestors:
1179 1180 fparent1, fparent2 = fparent2, nullid
1180 1181 elif fparent2 in fparentancestors:
1181 1182 fparent2 = nullid
1182 1183
1183 1184 # is the file changed?
1184 1185 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1185 1186 changelist.append(fname)
1186 1187 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1187 1188
1188 1189 # are just the flags changed during merge?
1189 1190 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1190 1191 changelist.append(fname)
1191 1192
1192 1193 return fparent1
1193 1194
1194 1195 @unfilteredmethod
1195 1196 def commit(self, text="", user=None, date=None, match=None, force=False,
1196 1197 editor=False, extra={}):
1197 1198 """Add a new revision to current repository.
1198 1199
1199 1200 Revision information is gathered from the working directory,
1200 1201 match can be used to filter the committed files. If editor is
1201 1202 supplied, it is called to get a commit message.
1202 1203 """
1203 1204
1204 1205 def fail(f, msg):
1205 1206 raise util.Abort('%s: %s' % (f, msg))
1206 1207
1207 1208 if not match:
1208 1209 match = matchmod.always(self.root, '')
1209 1210
1210 1211 if not force:
1211 1212 vdirs = []
1212 1213 match.explicitdir = vdirs.append
1213 1214 match.bad = fail
1214 1215
1215 1216 wlock = self.wlock()
1216 1217 try:
1217 1218 wctx = self[None]
1218 1219 merge = len(wctx.parents()) > 1
1219 1220
1220 1221 if (not force and merge and match and
1221 1222 (match.files() or match.anypats())):
1222 1223 raise util.Abort(_('cannot partially commit a merge '
1223 1224 '(do not specify files or patterns)'))
1224 1225
1225 1226 changes = self.status(match=match, clean=force)
1226 1227 if force:
1227 1228 changes[0].extend(changes[6]) # mq may commit unchanged files
1228 1229
1229 1230 # check subrepos
1230 1231 subs = []
1231 1232 commitsubs = set()
1232 1233 newstate = wctx.substate.copy()
1233 1234 # only manage subrepos and .hgsubstate if .hgsub is present
1234 1235 if '.hgsub' in wctx:
1235 1236 # we'll decide whether to track this ourselves, thanks
1236 1237 for c in changes[:3]:
1237 1238 if '.hgsubstate' in c:
1238 1239 c.remove('.hgsubstate')
1239 1240
1240 1241 # compare current state to last committed state
1241 1242 # build new substate based on last committed state
1242 1243 oldstate = wctx.p1().substate
1243 1244 for s in sorted(newstate.keys()):
1244 1245 if not match(s):
1245 1246 # ignore working copy, use old state if present
1246 1247 if s in oldstate:
1247 1248 newstate[s] = oldstate[s]
1248 1249 continue
1249 1250 if not force:
1250 1251 raise util.Abort(
1251 1252 _("commit with new subrepo %s excluded") % s)
1252 1253 if wctx.sub(s).dirty(True):
1253 1254 if not self.ui.configbool('ui', 'commitsubrepos'):
1254 1255 raise util.Abort(
1255 1256 _("uncommitted changes in subrepo %s") % s,
1256 1257 hint=_("use --subrepos for recursive commit"))
1257 1258 subs.append(s)
1258 1259 commitsubs.add(s)
1259 1260 else:
1260 1261 bs = wctx.sub(s).basestate()
1261 1262 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1262 1263 if oldstate.get(s, (None, None, None))[1] != bs:
1263 1264 subs.append(s)
1264 1265
1265 1266 # check for removed subrepos
1266 1267 for p in wctx.parents():
1267 1268 r = [s for s in p.substate if s not in newstate]
1268 1269 subs += [s for s in r if match(s)]
1269 1270 if subs:
1270 1271 if (not match('.hgsub') and
1271 1272 '.hgsub' in (wctx.modified() + wctx.added())):
1272 1273 raise util.Abort(
1273 1274 _("can't commit subrepos without .hgsub"))
1274 1275 changes[0].insert(0, '.hgsubstate')
1275 1276
1276 1277 elif '.hgsub' in changes[2]:
1277 1278 # clean up .hgsubstate when .hgsub is removed
1278 1279 if ('.hgsubstate' in wctx and
1279 1280 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1280 1281 changes[2].insert(0, '.hgsubstate')
1281 1282
1282 1283 # make sure all explicit patterns are matched
1283 1284 if not force and match.files():
1284 1285 matched = set(changes[0] + changes[1] + changes[2])
1285 1286
1286 1287 for f in match.files():
1287 1288 f = self.dirstate.normalize(f)
1288 1289 if f == '.' or f in matched or f in wctx.substate:
1289 1290 continue
1290 1291 if f in changes[3]: # missing
1291 1292 fail(f, _('file not found!'))
1292 1293 if f in vdirs: # visited directory
1293 1294 d = f + '/'
1294 1295 for mf in matched:
1295 1296 if mf.startswith(d):
1296 1297 break
1297 1298 else:
1298 1299 fail(f, _("no match under directory!"))
1299 1300 elif f not in self.dirstate:
1300 1301 fail(f, _("file not tracked!"))
1301 1302
1302 1303 cctx = context.workingctx(self, text, user, date, extra, changes)
1303 1304
1304 1305 if (not force and not extra.get("close") and not merge
1305 1306 and not cctx.files()
1306 1307 and wctx.branch() == wctx.p1().branch()):
1307 1308 return None
1308 1309
1309 1310 if merge and cctx.deleted():
1310 1311 raise util.Abort(_("cannot commit merge with missing files"))
1311 1312
1312 1313 ms = mergemod.mergestate(self)
1313 1314 for f in changes[0]:
1314 1315 if f in ms and ms[f] == 'u':
1315 1316 raise util.Abort(_("unresolved merge conflicts "
1316 1317 "(see hg help resolve)"))
1317 1318
1318 1319 if editor:
1319 1320 cctx._text = editor(self, cctx, subs)
1320 1321 edited = (text != cctx._text)
1321 1322
1322 1323 # Save commit message in case this transaction gets rolled back
1323 1324 # (e.g. by a pretxncommit hook). Leave the content alone on
1324 1325 # the assumption that the user will use the same editor again.
1325 1326 msgfn = self.savecommitmessage(cctx._text)
1326 1327
1327 1328 # commit subs and write new state
1328 1329 if subs:
1329 1330 for s in sorted(commitsubs):
1330 1331 sub = wctx.sub(s)
1331 1332 self.ui.status(_('committing subrepository %s\n') %
1332 1333 subrepo.subrelpath(sub))
1333 1334 sr = sub.commit(cctx._text, user, date)
1334 1335 newstate[s] = (newstate[s][0], sr)
1335 1336 subrepo.writestate(self, newstate)
1336 1337
1337 1338 p1, p2 = self.dirstate.parents()
1338 1339 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1339 1340 try:
1340 1341 self.hook("precommit", throw=True, parent1=hookp1,
1341 1342 parent2=hookp2)
1342 1343 ret = self.commitctx(cctx, True)
1343 1344 except: # re-raises
1344 1345 if edited:
1345 1346 self.ui.write(
1346 1347 _('note: commit message saved in %s\n') % msgfn)
1347 1348 raise
1348 1349
1349 1350 # update bookmarks, dirstate and mergestate
1350 1351 bookmarks.update(self, [p1, p2], ret)
1351 1352 cctx.markcommitted(ret)
1352 1353 ms.reset()
1353 1354 finally:
1354 1355 wlock.release()
1355 1356
1356 1357 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1357 1358 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1358 1359 self._afterlock(commithook)
1359 1360 return ret
1360 1361
1361 1362 @unfilteredmethod
1362 1363 def commitctx(self, ctx, error=False):
1363 1364 """Add a new revision to current repository.
1364 1365 Revision information is passed via the context argument.
1365 1366 """
1366 1367
1367 1368 tr = lock = None
1368 1369 removed = list(ctx.removed())
1369 1370 p1, p2 = ctx.p1(), ctx.p2()
1370 1371 user = ctx.user()
1371 1372
1372 1373 lock = self.lock()
1373 1374 try:
1374 1375 tr = self.transaction("commit")
1375 1376 trp = weakref.proxy(tr)
1376 1377
1377 1378 if ctx.files():
1378 1379 m1 = p1.manifest().copy()
1379 1380 m2 = p2.manifest()
1380 1381
1381 1382 # check in files
1382 1383 new = {}
1383 1384 changed = []
1384 1385 linkrev = len(self)
1385 1386 for f in sorted(ctx.modified() + ctx.added()):
1386 1387 self.ui.note(f + "\n")
1387 1388 try:
1388 1389 fctx = ctx[f]
1389 1390 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1390 1391 changed)
1391 1392 m1.set(f, fctx.flags())
1392 1393 except OSError, inst:
1393 1394 self.ui.warn(_("trouble committing %s!\n") % f)
1394 1395 raise
1395 1396 except IOError, inst:
1396 1397 errcode = getattr(inst, 'errno', errno.ENOENT)
1397 1398 if error or errcode and errcode != errno.ENOENT:
1398 1399 self.ui.warn(_("trouble committing %s!\n") % f)
1399 1400 raise
1400 1401 else:
1401 1402 removed.append(f)
1402 1403
1403 1404 # update manifest
1404 1405 m1.update(new)
1405 1406 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1406 1407 drop = [f for f in removed if f in m1]
1407 1408 for f in drop:
1408 1409 del m1[f]
1409 1410 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1410 1411 p2.manifestnode(), (new, drop))
1411 1412 files = changed + removed
1412 1413 else:
1413 1414 mn = p1.manifestnode()
1414 1415 files = []
1415 1416
1416 1417 # update changelog
1417 1418 self.changelog.delayupdate()
1418 1419 n = self.changelog.add(mn, files, ctx.description(),
1419 1420 trp, p1.node(), p2.node(),
1420 1421 user, ctx.date(), ctx.extra().copy())
1421 1422 p = lambda: self.changelog.writepending() and self.root or ""
1422 1423 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1423 1424 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1424 1425 parent2=xp2, pending=p)
1425 1426 self.changelog.finalize(trp)
1426 1427 # set the new commit is proper phase
1427 1428 targetphase = subrepo.newcommitphase(self.ui, ctx)
1428 1429 if targetphase:
1429 1430 # retract boundary do not alter parent changeset.
1430 1431 # if a parent have higher the resulting phase will
1431 1432 # be compliant anyway
1432 1433 #
1433 1434 # if minimal phase was 0 we don't need to retract anything
1434 1435 phases.retractboundary(self, targetphase, [n])
1435 1436 tr.close()
1436 1437 branchmap.updatecache(self.filtered('served'))
1437 1438 return n
1438 1439 finally:
1439 1440 if tr:
1440 1441 tr.release()
1441 1442 lock.release()
1442 1443
1443 1444 @unfilteredmethod
1444 1445 def destroying(self):
1445 1446 '''Inform the repository that nodes are about to be destroyed.
1446 1447 Intended for use by strip and rollback, so there's a common
1447 1448 place for anything that has to be done before destroying history.
1448 1449
1449 1450 This is mostly useful for saving state that is in memory and waiting
1450 1451 to be flushed when the current lock is released. Because a call to
1451 1452 destroyed is imminent, the repo will be invalidated causing those
1452 1453 changes to stay in memory (waiting for the next unlock), or vanish
1453 1454 completely.
1454 1455 '''
1455 1456 # When using the same lock to commit and strip, the phasecache is left
1456 1457 # dirty after committing. Then when we strip, the repo is invalidated,
1457 1458 # causing those changes to disappear.
1458 1459 if '_phasecache' in vars(self):
1459 1460 self._phasecache.write()
1460 1461
1461 1462 @unfilteredmethod
1462 1463 def destroyed(self):
1463 1464 '''Inform the repository that nodes have been destroyed.
1464 1465 Intended for use by strip and rollback, so there's a common
1465 1466 place for anything that has to be done after destroying history.
1466 1467 '''
1467 1468 # When one tries to:
1468 1469 # 1) destroy nodes thus calling this method (e.g. strip)
1469 1470 # 2) use phasecache somewhere (e.g. commit)
1470 1471 #
1471 1472 # then 2) will fail because the phasecache contains nodes that were
1472 1473 # removed. We can either remove phasecache from the filecache,
1473 1474 # causing it to reload next time it is accessed, or simply filter
1474 1475 # the removed nodes now and write the updated cache.
1475 1476 self._phasecache.filterunknown(self)
1476 1477 self._phasecache.write()
1477 1478
1478 1479 # update the 'served' branch cache to help read only server process
1479 1480 # Thanks to branchcache collaboration this is done from the nearest
1480 1481 # filtered subset and it is expected to be fast.
1481 1482 branchmap.updatecache(self.filtered('served'))
1482 1483
1483 1484 # Ensure the persistent tag cache is updated. Doing it now
1484 1485 # means that the tag cache only has to worry about destroyed
1485 1486 # heads immediately after a strip/rollback. That in turn
1486 1487 # guarantees that "cachetip == currenttip" (comparing both rev
1487 1488 # and node) always means no nodes have been added or destroyed.
1488 1489
1489 1490 # XXX this is suboptimal when qrefresh'ing: we strip the current
1490 1491 # head, refresh the tag cache, then immediately add a new head.
1491 1492 # But I think doing it this way is necessary for the "instant
1492 1493 # tag cache retrieval" case to work.
1493 1494 self.invalidate()
1494 1495
1495 1496 def walk(self, match, node=None):
1496 1497 '''
1497 1498 walk recursively through the directory tree or a given
1498 1499 changeset, finding all files matched by the match
1499 1500 function
1500 1501 '''
1501 1502 return self[node].walk(match)
1502 1503
1503 1504 def status(self, node1='.', node2=None, match=None,
1504 1505 ignored=False, clean=False, unknown=False,
1505 1506 listsubrepos=False):
1506 1507 """return status of files between two nodes or node and working
1507 1508 directory.
1508 1509
1509 1510 If node1 is None, use the first dirstate parent instead.
1510 1511 If node2 is None, compare node1 with working directory.
1511 1512 """
1512 1513
1513 1514 def mfmatches(ctx):
1514 1515 mf = ctx.manifest().copy()
1515 1516 if match.always():
1516 1517 return mf
1517 1518 for fn in mf.keys():
1518 1519 if not match(fn):
1519 1520 del mf[fn]
1520 1521 return mf
1521 1522
1522 1523 ctx1 = self[node1]
1523 1524 ctx2 = self[node2]
1524 1525
1525 1526 working = ctx2.rev() is None
1526 1527 parentworking = working and ctx1 == self['.']
1527 1528 match = match or matchmod.always(self.root, self.getcwd())
1528 1529 listignored, listclean, listunknown = ignored, clean, unknown
1529 1530
1530 1531 # load earliest manifest first for caching reasons
1531 1532 if not working and ctx2.rev() < ctx1.rev():
1532 1533 ctx2.manifest()
1533 1534
1534 1535 if not parentworking:
1535 1536 def bad(f, msg):
1536 1537 # 'f' may be a directory pattern from 'match.files()',
1537 1538 # so 'f not in ctx1' is not enough
1538 1539 if f not in ctx1 and f not in ctx1.dirs():
1539 1540 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1540 1541 match.bad = bad
1541 1542
1542 1543 if working: # we need to scan the working dir
1543 1544 subrepos = []
1544 1545 if '.hgsub' in self.dirstate:
1545 1546 subrepos = sorted(ctx2.substate)
1546 1547 s = self.dirstate.status(match, subrepos, listignored,
1547 1548 listclean, listunknown)
1548 1549 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1549 1550
1550 1551 # check for any possibly clean files
1551 1552 if parentworking and cmp:
1552 1553 fixup = []
1553 1554 # do a full compare of any files that might have changed
1554 1555 for f in sorted(cmp):
1555 1556 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1556 1557 or ctx1[f].cmp(ctx2[f])):
1557 1558 modified.append(f)
1558 1559 else:
1559 1560 fixup.append(f)
1560 1561
1561 1562 # update dirstate for files that are actually clean
1562 1563 if fixup:
1563 1564 if listclean:
1564 1565 clean += fixup
1565 1566
1566 1567 try:
1567 1568 # updating the dirstate is optional
1568 1569 # so we don't wait on the lock
1569 1570 wlock = self.wlock(False)
1570 1571 try:
1571 1572 for f in fixup:
1572 1573 self.dirstate.normal(f)
1573 1574 finally:
1574 1575 wlock.release()
1575 1576 except error.LockError:
1576 1577 pass
1577 1578
1578 1579 if not parentworking:
1579 1580 mf1 = mfmatches(ctx1)
1580 1581 if working:
1581 1582 # we are comparing working dir against non-parent
1582 1583 # generate a pseudo-manifest for the working dir
1583 1584 mf2 = mfmatches(self['.'])
1584 1585 for f in cmp + modified + added:
1585 1586 mf2[f] = None
1586 1587 mf2.set(f, ctx2.flags(f))
1587 1588 for f in removed:
1588 1589 if f in mf2:
1589 1590 del mf2[f]
1590 1591 else:
1591 1592 # we are comparing two revisions
1592 1593 deleted, unknown, ignored = [], [], []
1593 1594 mf2 = mfmatches(ctx2)
1594 1595
1595 1596 modified, added, clean = [], [], []
1596 1597 withflags = mf1.withflags() | mf2.withflags()
1597 1598 for fn, mf2node in mf2.iteritems():
1598 1599 if fn in mf1:
1599 1600 if (fn not in deleted and
1600 1601 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1601 1602 (mf1[fn] != mf2node and
1602 1603 (mf2node or ctx1[fn].cmp(ctx2[fn]))))):
1603 1604 modified.append(fn)
1604 1605 elif listclean:
1605 1606 clean.append(fn)
1606 1607 del mf1[fn]
1607 1608 elif fn not in deleted:
1608 1609 added.append(fn)
1609 1610 removed = mf1.keys()
1610 1611
1611 1612 if working and modified and not self.dirstate._checklink:
1612 1613 # Symlink placeholders may get non-symlink-like contents
1613 1614 # via user error or dereferencing by NFS or Samba servers,
1614 1615 # so we filter out any placeholders that don't look like a
1615 1616 # symlink
1616 1617 sane = []
1617 1618 for f in modified:
1618 1619 if ctx2.flags(f) == 'l':
1619 1620 d = ctx2[f].data()
1620 1621 if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d):
1621 1622 self.ui.debug('ignoring suspect symlink placeholder'
1622 1623 ' "%s"\n' % f)
1623 1624 continue
1624 1625 sane.append(f)
1625 1626 modified = sane
1626 1627
1627 1628 r = modified, added, removed, deleted, unknown, ignored, clean
1628 1629
1629 1630 if listsubrepos:
1630 1631 for subpath, sub in scmutil.itersubrepos(ctx1, ctx2):
1631 1632 if working:
1632 1633 rev2 = None
1633 1634 else:
1634 1635 rev2 = ctx2.substate[subpath][1]
1635 1636 try:
1636 1637 submatch = matchmod.narrowmatcher(subpath, match)
1637 1638 s = sub.status(rev2, match=submatch, ignored=listignored,
1638 1639 clean=listclean, unknown=listunknown,
1639 1640 listsubrepos=True)
1640 1641 for rfiles, sfiles in zip(r, s):
1641 1642 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1642 1643 except error.LookupError:
1643 1644 self.ui.status(_("skipping missing subrepository: %s\n")
1644 1645 % subpath)
1645 1646
1646 1647 for l in r:
1647 1648 l.sort()
1648 1649 return r
1649 1650
1650 1651 def heads(self, start=None):
1651 1652 heads = self.changelog.heads(start)
1652 1653 # sort the output in rev descending order
1653 1654 return sorted(heads, key=self.changelog.rev, reverse=True)
1654 1655
1655 1656 def branchheads(self, branch=None, start=None, closed=False):
1656 1657 '''return a (possibly filtered) list of heads for the given branch
1657 1658
1658 1659 Heads are returned in topological order, from newest to oldest.
1659 1660 If branch is None, use the dirstate branch.
1660 1661 If start is not None, return only heads reachable from start.
1661 1662 If closed is True, return heads that are marked as closed as well.
1662 1663 '''
1663 1664 if branch is None:
1664 1665 branch = self[None].branch()
1665 1666 branches = self.branchmap()
1666 1667 if branch not in branches:
1667 1668 return []
1668 1669 # the cache returns heads ordered lowest to highest
1669 1670 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1670 1671 if start is not None:
1671 1672 # filter out the heads that cannot be reached from startrev
1672 1673 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1673 1674 bheads = [h for h in bheads if h in fbheads]
1674 1675 return bheads
1675 1676
1676 1677 def branches(self, nodes):
1677 1678 if not nodes:
1678 1679 nodes = [self.changelog.tip()]
1679 1680 b = []
1680 1681 for n in nodes:
1681 1682 t = n
1682 1683 while True:
1683 1684 p = self.changelog.parents(n)
1684 1685 if p[1] != nullid or p[0] == nullid:
1685 1686 b.append((t, n, p[0], p[1]))
1686 1687 break
1687 1688 n = p[0]
1688 1689 return b
1689 1690
1690 1691 def between(self, pairs):
1691 1692 r = []
1692 1693
1693 1694 for top, bottom in pairs:
1694 1695 n, l, i = top, [], 0
1695 1696 f = 1
1696 1697
1697 1698 while n != bottom and n != nullid:
1698 1699 p = self.changelog.parents(n)[0]
1699 1700 if i == f:
1700 1701 l.append(n)
1701 1702 f = f * 2
1702 1703 n = p
1703 1704 i += 1
1704 1705
1705 1706 r.append(l)
1706 1707
1707 1708 return r
1708 1709
1709 1710 def pull(self, remote, heads=None, force=False):
1710 1711 return exchange.pull (self, remote, heads, force)
1711 1712
1712 1713 def checkpush(self, pushop):
1713 1714 """Extensions can override this function if additional checks have
1714 1715 to be performed before pushing, or call it if they override push
1715 1716 command.
1716 1717 """
1717 1718 pass
1718 1719
1719 1720 @unfilteredpropertycache
1720 1721 def prepushoutgoinghooks(self):
1721 1722 """Return util.hooks consists of "(repo, remote, outgoing)"
1722 1723 functions, which are called before pushing changesets.
1723 1724 """
1724 1725 return util.hooks()
1725 1726
1726 1727 def push(self, remote, force=False, revs=None, newbranch=False):
1727 1728 return exchange.push(self, remote, force, revs, newbranch)
1728 1729
1729 1730 def stream_in(self, remote, requirements):
1730 1731 lock = self.lock()
1731 1732 try:
1732 1733 # Save remote branchmap. We will use it later
1733 1734 # to speed up branchcache creation
1734 1735 rbranchmap = None
1735 1736 if remote.capable("branchmap"):
1736 1737 rbranchmap = remote.branchmap()
1737 1738
1738 1739 fp = remote.stream_out()
1739 1740 l = fp.readline()
1740 1741 try:
1741 1742 resp = int(l)
1742 1743 except ValueError:
1743 1744 raise error.ResponseError(
1744 1745 _('unexpected response from remote server:'), l)
1745 1746 if resp == 1:
1746 1747 raise util.Abort(_('operation forbidden by server'))
1747 1748 elif resp == 2:
1748 1749 raise util.Abort(_('locking the remote repository failed'))
1749 1750 elif resp != 0:
1750 1751 raise util.Abort(_('the server sent an unknown error code'))
1751 1752 self.ui.status(_('streaming all changes\n'))
1752 1753 l = fp.readline()
1753 1754 try:
1754 1755 total_files, total_bytes = map(int, l.split(' ', 1))
1755 1756 except (ValueError, TypeError):
1756 1757 raise error.ResponseError(
1757 1758 _('unexpected response from remote server:'), l)
1758 1759 self.ui.status(_('%d files to transfer, %s of data\n') %
1759 1760 (total_files, util.bytecount(total_bytes)))
1760 1761 handled_bytes = 0
1761 1762 self.ui.progress(_('clone'), 0, total=total_bytes)
1762 1763 start = time.time()
1763 1764
1764 1765 tr = self.transaction(_('clone'))
1765 1766 try:
1766 1767 for i in xrange(total_files):
1767 1768 # XXX doesn't support '\n' or '\r' in filenames
1768 1769 l = fp.readline()
1769 1770 try:
1770 1771 name, size = l.split('\0', 1)
1771 1772 size = int(size)
1772 1773 except (ValueError, TypeError):
1773 1774 raise error.ResponseError(
1774 1775 _('unexpected response from remote server:'), l)
1775 1776 if self.ui.debugflag:
1776 1777 self.ui.debug('adding %s (%s)\n' %
1777 1778 (name, util.bytecount(size)))
1778 1779 # for backwards compat, name was partially encoded
1779 1780 ofp = self.sopener(store.decodedir(name), 'w')
1780 1781 for chunk in util.filechunkiter(fp, limit=size):
1781 1782 handled_bytes += len(chunk)
1782 1783 self.ui.progress(_('clone'), handled_bytes,
1783 1784 total=total_bytes)
1784 1785 ofp.write(chunk)
1785 1786 ofp.close()
1786 1787 tr.close()
1787 1788 finally:
1788 1789 tr.release()
1789 1790
1790 1791 # Writing straight to files circumvented the inmemory caches
1791 1792 self.invalidate()
1792 1793
1793 1794 elapsed = time.time() - start
1794 1795 if elapsed <= 0:
1795 1796 elapsed = 0.001
1796 1797 self.ui.progress(_('clone'), None)
1797 1798 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1798 1799 (util.bytecount(total_bytes), elapsed,
1799 1800 util.bytecount(total_bytes / elapsed)))
1800 1801
1801 1802 # new requirements = old non-format requirements +
1802 1803 # new format-related
1803 1804 # requirements from the streamed-in repository
1804 1805 requirements.update(set(self.requirements) - self.supportedformats)
1805 1806 self._applyrequirements(requirements)
1806 1807 self._writerequirements()
1807 1808
1808 1809 if rbranchmap:
1809 1810 rbheads = []
1810 1811 for bheads in rbranchmap.itervalues():
1811 1812 rbheads.extend(bheads)
1812 1813
1813 1814 if rbheads:
1814 1815 rtiprev = max((int(self.changelog.rev(node))
1815 1816 for node in rbheads))
1816 1817 cache = branchmap.branchcache(rbranchmap,
1817 1818 self[rtiprev].node(),
1818 1819 rtiprev)
1819 1820 # Try to stick it as low as possible
1820 1821 # filter above served are unlikely to be fetch from a clone
1821 1822 for candidate in ('base', 'immutable', 'served'):
1822 1823 rview = self.filtered(candidate)
1823 1824 if cache.validfor(rview):
1824 1825 self._branchcaches[candidate] = cache
1825 1826 cache.write(rview)
1826 1827 break
1827 1828 self.invalidate()
1828 1829 return len(self.heads()) + 1
1829 1830 finally:
1830 1831 lock.release()
1831 1832
1832 1833 def clone(self, remote, heads=[], stream=False):
1833 1834 '''clone remote repository.
1834 1835
1835 1836 keyword arguments:
1836 1837 heads: list of revs to clone (forces use of pull)
1837 1838 stream: use streaming clone if possible'''
1838 1839
1839 1840 # now, all clients that can request uncompressed clones can
1840 1841 # read repo formats supported by all servers that can serve
1841 1842 # them.
1842 1843
1843 1844 # if revlog format changes, client will have to check version
1844 1845 # and format flags on "stream" capability, and use
1845 1846 # uncompressed only if compatible.
1846 1847
1847 1848 if not stream:
1848 1849 # if the server explicitly prefers to stream (for fast LANs)
1849 1850 stream = remote.capable('stream-preferred')
1850 1851
1851 1852 if stream and not heads:
1852 1853 # 'stream' means remote revlog format is revlogv1 only
1853 1854 if remote.capable('stream'):
1854 1855 return self.stream_in(remote, set(('revlogv1',)))
1855 1856 # otherwise, 'streamreqs' contains the remote revlog format
1856 1857 streamreqs = remote.capable('streamreqs')
1857 1858 if streamreqs:
1858 1859 streamreqs = set(streamreqs.split(','))
1859 1860 # if we support it, stream in and adjust our requirements
1860 1861 if not streamreqs - self.supportedformats:
1861 1862 return self.stream_in(remote, streamreqs)
1862 1863 return self.pull(remote, heads)
1863 1864
1864 1865 def pushkey(self, namespace, key, old, new):
1865 1866 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1866 1867 old=old, new=new)
1867 1868 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1868 1869 ret = pushkey.push(self, namespace, key, old, new)
1869 1870 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1870 1871 ret=ret)
1871 1872 return ret
1872 1873
1873 1874 def listkeys(self, namespace):
1874 1875 self.hook('prelistkeys', throw=True, namespace=namespace)
1875 1876 self.ui.debug('listing keys for "%s"\n' % namespace)
1876 1877 values = pushkey.list(self, namespace)
1877 1878 self.hook('listkeys', namespace=namespace, values=values)
1878 1879 return values
1879 1880
1880 1881 def debugwireargs(self, one, two, three=None, four=None, five=None):
1881 1882 '''used to test argument passing over the wire'''
1882 1883 return "%s %s %s %s %s" % (one, two, three, four, five)
1883 1884
1884 1885 def savecommitmessage(self, text):
1885 1886 fp = self.opener('last-message.txt', 'wb')
1886 1887 try:
1887 1888 fp.write(text)
1888 1889 finally:
1889 1890 fp.close()
1890 1891 return self.pathto(fp.name[len(self.root) + 1:])
1891 1892
1892 1893 # used to avoid circular references so destructors work
1893 1894 def aftertrans(files):
1894 1895 renamefiles = [tuple(t) for t in files]
1895 1896 def a():
1896 1897 for vfs, src, dest in renamefiles:
1897 1898 try:
1898 1899 vfs.rename(src, dest)
1899 1900 except OSError: # journal file does not yet exist
1900 1901 pass
1901 1902 return a
1902 1903
1903 1904 def undoname(fn):
1904 1905 base, name = os.path.split(fn)
1905 1906 assert name.startswith('journal')
1906 1907 return os.path.join(base, name.replace('journal', 'undo', 1))
1907 1908
1908 1909 def instance(ui, path, create):
1909 1910 return localrepository(ui, util.urllocalpath(path), create)
1910 1911
1911 1912 def islocal(path):
1912 1913 return True
@@ -1,36 +1,37
1 1 $ hg init
2 2 $ echo a > a
3 3 $ hg ci -Am0
4 4 adding a
5 5
6 6 $ hg -q clone . foo
7 7
8 8 $ touch .hg/store/journal
9 9
10 10 $ echo foo > a
11 11 $ hg ci -Am0
12 abort: abandoned transaction found - run hg recover!
12 abort: abandoned transaction found!
13 (run 'hg recover' to clean up transaction)
13 14 [255]
14 15
15 16 $ hg recover
16 17 rolling back interrupted transaction
17 18 checking changesets
18 19 checking manifests
19 20 crosschecking files in changesets and manifests
20 21 checking files
21 22 1 files, 1 changesets, 1 total revisions
22 23
23 24 Check that zero-size journals are correctly aborted:
24 25
25 26 #if unix-permissions no-root
26 27 $ hg bundle -qa repo.hg
27 28 $ chmod -w foo/.hg/store/00changelog.i
28 29
29 30 $ hg -R foo unbundle repo.hg
30 31 adding changesets
31 32 abort: Permission denied: $TESTTMP/foo/.hg/store/.00changelog.i-* (glob)
32 33 [255]
33 34
34 35 $ if test -f foo/.hg/store/journal; then echo 'journal exists :-('; fi
35 36 #endif
36 37
General Comments 0
You need to be logged in to leave comments. Login now