##// END OF EJS Templates
validate: check for spurious incoming filelog entries...
Sune Foldager -
r18422:f009804e default
parent child Browse files
Show More
@@ -1,2573 +1,2576 b''
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7 from node import hex, nullid, short
8 8 from i18n import _
9 9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 11 import lock, transaction, store, encoding, base85
12 12 import scmutil, util, extensions, hook, error, revset
13 13 import match as matchmod
14 14 import merge as mergemod
15 15 import tags as tagsmod
16 16 from lock import release
17 17 import weakref, errno, os, time, inspect
18 18 import branchmap
19 19 propertycache = util.propertycache
20 20 filecache = scmutil.filecache
21 21
22 22 class repofilecache(filecache):
23 23 """All filecache usage on repo are done for logic that should be unfiltered
24 24 """
25 25
26 26 def __get__(self, repo, type=None):
27 27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 28 def __set__(self, repo, value):
29 29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 30 def __delete__(self, repo):
31 31 return super(repofilecache, self).__delete__(repo.unfiltered())
32 32
33 33 class storecache(repofilecache):
34 34 """filecache for files in the store"""
35 35 def join(self, obj, fname):
36 36 return obj.sjoin(fname)
37 37
38 38 class unfilteredpropertycache(propertycache):
39 39 """propertycache that apply to unfiltered repo only"""
40 40
41 41 def __get__(self, repo, type=None):
42 42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43 43
44 44 class filteredpropertycache(propertycache):
45 45 """propertycache that must take filtering in account"""
46 46
47 47 def cachevalue(self, obj, value):
48 48 object.__setattr__(obj, self.name, value)
49 49
50 50
51 51 def hasunfilteredcache(repo, name):
52 52 """check if an repo and a unfilteredproperty cached value for <name>"""
53 53 return name in vars(repo.unfiltered())
54 54
55 55 def unfilteredmethod(orig):
56 56 """decorate method that always need to be run on unfiltered version"""
57 57 def wrapper(repo, *args, **kwargs):
58 58 return orig(repo.unfiltered(), *args, **kwargs)
59 59 return wrapper
60 60
61 61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63 63
64 64 class localpeer(peer.peerrepository):
65 65 '''peer for a local repo; reflects only the most recent API'''
66 66
67 67 def __init__(self, repo, caps=MODERNCAPS):
68 68 peer.peerrepository.__init__(self)
69 69 self._repo = repo.filtered('served')
70 70 self.ui = repo.ui
71 71 self._caps = repo._restrictcapabilities(caps)
72 72 self.requirements = repo.requirements
73 73 self.supportedformats = repo.supportedformats
74 74
75 75 def close(self):
76 76 self._repo.close()
77 77
78 78 def _capabilities(self):
79 79 return self._caps
80 80
81 81 def local(self):
82 82 return self._repo
83 83
84 84 def canpush(self):
85 85 return True
86 86
87 87 def url(self):
88 88 return self._repo.url()
89 89
90 90 def lookup(self, key):
91 91 return self._repo.lookup(key)
92 92
93 93 def branchmap(self):
94 94 return self._repo.branchmap()
95 95
96 96 def heads(self):
97 97 return self._repo.heads()
98 98
99 99 def known(self, nodes):
100 100 return self._repo.known(nodes)
101 101
102 102 def getbundle(self, source, heads=None, common=None):
103 103 return self._repo.getbundle(source, heads=heads, common=common)
104 104
105 105 # TODO We might want to move the next two calls into legacypeer and add
106 106 # unbundle instead.
107 107
108 108 def lock(self):
109 109 return self._repo.lock()
110 110
111 111 def addchangegroup(self, cg, source, url):
112 112 return self._repo.addchangegroup(cg, source, url)
113 113
114 114 def pushkey(self, namespace, key, old, new):
115 115 return self._repo.pushkey(namespace, key, old, new)
116 116
117 117 def listkeys(self, namespace):
118 118 return self._repo.listkeys(namespace)
119 119
120 120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 121 '''used to test argument passing over the wire'''
122 122 return "%s %s %s %s %s" % (one, two, three, four, five)
123 123
124 124 class locallegacypeer(localpeer):
125 125 '''peer extension which implements legacy methods too; used for tests with
126 126 restricted capabilities'''
127 127
128 128 def __init__(self, repo):
129 129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130 130
131 131 def branches(self, nodes):
132 132 return self._repo.branches(nodes)
133 133
134 134 def between(self, pairs):
135 135 return self._repo.between(pairs)
136 136
137 137 def changegroup(self, basenodes, source):
138 138 return self._repo.changegroup(basenodes, source)
139 139
140 140 def changegroupsubset(self, bases, heads, source):
141 141 return self._repo.changegroupsubset(bases, heads, source)
142 142
143 143 class localrepository(object):
144 144
145 145 supportedformats = set(('revlogv1', 'generaldelta'))
146 146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 147 'dotencode'))
148 148 openerreqs = set(('revlogv1', 'generaldelta'))
149 149 requirements = ['revlogv1']
150 150 filtername = None
151 151
152 152 def _baserequirements(self, create):
153 153 return self.requirements[:]
154 154
155 155 def __init__(self, baseui, path=None, create=False):
156 156 self.wvfs = scmutil.vfs(path, expand=True)
157 157 self.wopener = self.wvfs
158 158 self.root = self.wvfs.base
159 159 self.path = self.wvfs.join(".hg")
160 160 self.origroot = path
161 161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 162 self.vfs = scmutil.vfs(self.path)
163 163 self.opener = self.vfs
164 164 self.baseui = baseui
165 165 self.ui = baseui.copy()
166 166 # A list of callback to shape the phase if no data were found.
167 167 # Callback are in the form: func(repo, roots) --> processed root.
168 168 # This list it to be filled by extension during repo setup
169 169 self._phasedefaults = []
170 170 try:
171 171 self.ui.readconfig(self.join("hgrc"), self.root)
172 172 extensions.loadall(self.ui)
173 173 except IOError:
174 174 pass
175 175
176 176 if not self.vfs.isdir():
177 177 if create:
178 178 if not self.wvfs.exists():
179 179 self.wvfs.makedirs()
180 180 self.vfs.makedir(notindexed=True)
181 181 requirements = self._baserequirements(create)
182 182 if self.ui.configbool('format', 'usestore', True):
183 183 self.vfs.mkdir("store")
184 184 requirements.append("store")
185 185 if self.ui.configbool('format', 'usefncache', True):
186 186 requirements.append("fncache")
187 187 if self.ui.configbool('format', 'dotencode', True):
188 188 requirements.append('dotencode')
189 189 # create an invalid changelog
190 190 self.vfs.append(
191 191 "00changelog.i",
192 192 '\0\0\0\2' # represents revlogv2
193 193 ' dummy changelog to prevent using the old repo layout'
194 194 )
195 195 if self.ui.configbool('format', 'generaldelta', False):
196 196 requirements.append("generaldelta")
197 197 requirements = set(requirements)
198 198 else:
199 199 raise error.RepoError(_("repository %s not found") % path)
200 200 elif create:
201 201 raise error.RepoError(_("repository %s already exists") % path)
202 202 else:
203 203 try:
204 204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 205 except IOError, inst:
206 206 if inst.errno != errno.ENOENT:
207 207 raise
208 208 requirements = set()
209 209
210 210 self.sharedpath = self.path
211 211 try:
212 212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 213 if not os.path.exists(s):
214 214 raise error.RepoError(
215 215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 216 self.sharedpath = s
217 217 except IOError, inst:
218 218 if inst.errno != errno.ENOENT:
219 219 raise
220 220
221 221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 222 self.spath = self.store.path
223 223 self.svfs = self.store.vfs
224 224 self.sopener = self.svfs
225 225 self.sjoin = self.store.join
226 226 self.vfs.createmode = self.store.createmode
227 227 self._applyrequirements(requirements)
228 228 if create:
229 229 self._writerequirements()
230 230
231 231
232 232 self._branchcaches = {}
233 233 self.filterpats = {}
234 234 self._datafilters = {}
235 235 self._transref = self._lockref = self._wlockref = None
236 236
237 237 # A cache for various files under .hg/ that tracks file changes,
238 238 # (used by the filecache decorator)
239 239 #
240 240 # Maps a property name to its util.filecacheentry
241 241 self._filecache = {}
242 242
243 243 # hold sets of revision to be filtered
244 244 # should be cleared when something might have changed the filter value:
245 245 # - new changesets,
246 246 # - phase change,
247 247 # - new obsolescence marker,
248 248 # - working directory parent change,
249 249 # - bookmark changes
250 250 self.filteredrevcache = {}
251 251
252 252 def close(self):
253 253 pass
254 254
255 255 def _restrictcapabilities(self, caps):
256 256 return caps
257 257
258 258 def _applyrequirements(self, requirements):
259 259 self.requirements = requirements
260 260 self.sopener.options = dict((r, 1) for r in requirements
261 261 if r in self.openerreqs)
262 262
263 263 def _writerequirements(self):
264 264 reqfile = self.opener("requires", "w")
265 265 for r in sorted(self.requirements):
266 266 reqfile.write("%s\n" % r)
267 267 reqfile.close()
268 268
269 269 def _checknested(self, path):
270 270 """Determine if path is a legal nested repository."""
271 271 if not path.startswith(self.root):
272 272 return False
273 273 subpath = path[len(self.root) + 1:]
274 274 normsubpath = util.pconvert(subpath)
275 275
276 276 # XXX: Checking against the current working copy is wrong in
277 277 # the sense that it can reject things like
278 278 #
279 279 # $ hg cat -r 10 sub/x.txt
280 280 #
281 281 # if sub/ is no longer a subrepository in the working copy
282 282 # parent revision.
283 283 #
284 284 # However, it can of course also allow things that would have
285 285 # been rejected before, such as the above cat command if sub/
286 286 # is a subrepository now, but was a normal directory before.
287 287 # The old path auditor would have rejected by mistake since it
288 288 # panics when it sees sub/.hg/.
289 289 #
290 290 # All in all, checking against the working copy seems sensible
291 291 # since we want to prevent access to nested repositories on
292 292 # the filesystem *now*.
293 293 ctx = self[None]
294 294 parts = util.splitpath(subpath)
295 295 while parts:
296 296 prefix = '/'.join(parts)
297 297 if prefix in ctx.substate:
298 298 if prefix == normsubpath:
299 299 return True
300 300 else:
301 301 sub = ctx.sub(prefix)
302 302 return sub.checknested(subpath[len(prefix) + 1:])
303 303 else:
304 304 parts.pop()
305 305 return False
306 306
307 307 def peer(self):
308 308 return localpeer(self) # not cached to avoid reference cycle
309 309
310 310 def unfiltered(self):
311 311 """Return unfiltered version of the repository
312 312
313 313 Intended to be ovewritten by filtered repo."""
314 314 return self
315 315
316 316 def filtered(self, name):
317 317 """Return a filtered version of a repository"""
318 318 # build a new class with the mixin and the current class
319 319 # (possibily subclass of the repo)
320 320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 321 pass
322 322 return proxycls(self, name)
323 323
324 324 @repofilecache('bookmarks')
325 325 def _bookmarks(self):
326 326 return bookmarks.bmstore(self)
327 327
328 328 @repofilecache('bookmarks.current')
329 329 def _bookmarkcurrent(self):
330 330 return bookmarks.readcurrent(self)
331 331
332 332 def bookmarkheads(self, bookmark):
333 333 name = bookmark.split('@', 1)[0]
334 334 heads = []
335 335 for mark, n in self._bookmarks.iteritems():
336 336 if mark.split('@', 1)[0] == name:
337 337 heads.append(n)
338 338 return heads
339 339
340 340 @storecache('phaseroots')
341 341 def _phasecache(self):
342 342 return phases.phasecache(self, self._phasedefaults)
343 343
344 344 @storecache('obsstore')
345 345 def obsstore(self):
346 346 store = obsolete.obsstore(self.sopener)
347 347 if store and not obsolete._enabled:
348 348 # message is rare enough to not be translated
349 349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 350 self.ui.warn(msg % len(list(store)))
351 351 return store
352 352
353 353 @storecache('00changelog.i')
354 354 def changelog(self):
355 355 c = changelog.changelog(self.sopener)
356 356 if 'HG_PENDING' in os.environ:
357 357 p = os.environ['HG_PENDING']
358 358 if p.startswith(self.root):
359 359 c.readpending('00changelog.i.a')
360 360 return c
361 361
362 362 @storecache('00manifest.i')
363 363 def manifest(self):
364 364 return manifest.manifest(self.sopener)
365 365
366 366 @repofilecache('dirstate')
367 367 def dirstate(self):
368 368 warned = [0]
369 369 def validate(node):
370 370 try:
371 371 self.changelog.rev(node)
372 372 return node
373 373 except error.LookupError:
374 374 if not warned[0]:
375 375 warned[0] = True
376 376 self.ui.warn(_("warning: ignoring unknown"
377 377 " working parent %s!\n") % short(node))
378 378 return nullid
379 379
380 380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381 381
382 382 def __getitem__(self, changeid):
383 383 if changeid is None:
384 384 return context.workingctx(self)
385 385 return context.changectx(self, changeid)
386 386
387 387 def __contains__(self, changeid):
388 388 try:
389 389 return bool(self.lookup(changeid))
390 390 except error.RepoLookupError:
391 391 return False
392 392
393 393 def __nonzero__(self):
394 394 return True
395 395
396 396 def __len__(self):
397 397 return len(self.changelog)
398 398
399 399 def __iter__(self):
400 400 return iter(self.changelog)
401 401
402 402 def revs(self, expr, *args):
403 403 '''Return a list of revisions matching the given revset'''
404 404 expr = revset.formatspec(expr, *args)
405 405 m = revset.match(None, expr)
406 406 return [r for r in m(self, list(self))]
407 407
408 408 def set(self, expr, *args):
409 409 '''
410 410 Yield a context for each matching revision, after doing arg
411 411 replacement via revset.formatspec
412 412 '''
413 413 for r in self.revs(expr, *args):
414 414 yield self[r]
415 415
416 416 def url(self):
417 417 return 'file:' + self.root
418 418
419 419 def hook(self, name, throw=False, **args):
420 420 return hook.hook(self.ui, self, name, throw, **args)
421 421
422 422 @unfilteredmethod
423 423 def _tag(self, names, node, message, local, user, date, extra={}):
424 424 if isinstance(names, str):
425 425 names = (names,)
426 426
427 427 branches = self.branchmap()
428 428 for name in names:
429 429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 430 local=local)
431 431 if name in branches:
432 432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 433 " branch name\n") % name)
434 434
435 435 def writetags(fp, names, munge, prevtags):
436 436 fp.seek(0, 2)
437 437 if prevtags and prevtags[-1] != '\n':
438 438 fp.write('\n')
439 439 for name in names:
440 440 m = munge and munge(name) or name
441 441 if (self._tagscache.tagtypes and
442 442 name in self._tagscache.tagtypes):
443 443 old = self.tags().get(name, nullid)
444 444 fp.write('%s %s\n' % (hex(old), m))
445 445 fp.write('%s %s\n' % (hex(node), m))
446 446 fp.close()
447 447
448 448 prevtags = ''
449 449 if local:
450 450 try:
451 451 fp = self.opener('localtags', 'r+')
452 452 except IOError:
453 453 fp = self.opener('localtags', 'a')
454 454 else:
455 455 prevtags = fp.read()
456 456
457 457 # local tags are stored in the current charset
458 458 writetags(fp, names, None, prevtags)
459 459 for name in names:
460 460 self.hook('tag', node=hex(node), tag=name, local=local)
461 461 return
462 462
463 463 try:
464 464 fp = self.wfile('.hgtags', 'rb+')
465 465 except IOError, e:
466 466 if e.errno != errno.ENOENT:
467 467 raise
468 468 fp = self.wfile('.hgtags', 'ab')
469 469 else:
470 470 prevtags = fp.read()
471 471
472 472 # committed tags are stored in UTF-8
473 473 writetags(fp, names, encoding.fromlocal, prevtags)
474 474
475 475 fp.close()
476 476
477 477 self.invalidatecaches()
478 478
479 479 if '.hgtags' not in self.dirstate:
480 480 self[None].add(['.hgtags'])
481 481
482 482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484 484
485 485 for name in names:
486 486 self.hook('tag', node=hex(node), tag=name, local=local)
487 487
488 488 return tagnode
489 489
490 490 def tag(self, names, node, message, local, user, date):
491 491 '''tag a revision with one or more symbolic names.
492 492
493 493 names is a list of strings or, when adding a single tag, names may be a
494 494 string.
495 495
496 496 if local is True, the tags are stored in a per-repository file.
497 497 otherwise, they are stored in the .hgtags file, and a new
498 498 changeset is committed with the change.
499 499
500 500 keyword arguments:
501 501
502 502 local: whether to store tags in non-version-controlled file
503 503 (default False)
504 504
505 505 message: commit message to use if committing
506 506
507 507 user: name of user to use if committing
508 508
509 509 date: date tuple to use if committing'''
510 510
511 511 if not local:
512 512 for x in self.status()[:5]:
513 513 if '.hgtags' in x:
514 514 raise util.Abort(_('working copy of .hgtags is changed '
515 515 '(please commit .hgtags manually)'))
516 516
517 517 self.tags() # instantiate the cache
518 518 self._tag(names, node, message, local, user, date)
519 519
520 520 @filteredpropertycache
521 521 def _tagscache(self):
522 522 '''Returns a tagscache object that contains various tags related
523 523 caches.'''
524 524
525 525 # This simplifies its cache management by having one decorated
526 526 # function (this one) and the rest simply fetch things from it.
527 527 class tagscache(object):
528 528 def __init__(self):
529 529 # These two define the set of tags for this repository. tags
530 530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 531 # 'local'. (Global tags are defined by .hgtags across all
532 532 # heads, and local tags are defined in .hg/localtags.)
533 533 # They constitute the in-memory cache of tags.
534 534 self.tags = self.tagtypes = None
535 535
536 536 self.nodetagscache = self.tagslist = None
537 537
538 538 cache = tagscache()
539 539 cache.tags, cache.tagtypes = self._findtags()
540 540
541 541 return cache
542 542
543 543 def tags(self):
544 544 '''return a mapping of tag to node'''
545 545 t = {}
546 546 if self.changelog.filteredrevs:
547 547 tags, tt = self._findtags()
548 548 else:
549 549 tags = self._tagscache.tags
550 550 for k, v in tags.iteritems():
551 551 try:
552 552 # ignore tags to unknown nodes
553 553 self.changelog.rev(v)
554 554 t[k] = v
555 555 except (error.LookupError, ValueError):
556 556 pass
557 557 return t
558 558
559 559 def _findtags(self):
560 560 '''Do the hard work of finding tags. Return a pair of dicts
561 561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 562 maps tag name to a string like \'global\' or \'local\'.
563 563 Subclasses or extensions are free to add their own tags, but
564 564 should be aware that the returned dicts will be retained for the
565 565 duration of the localrepo object.'''
566 566
567 567 # XXX what tagtype should subclasses/extensions use? Currently
568 568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 569 # Should each extension invent its own tag type? Should there
570 570 # be one tagtype for all such "virtual" tags? Or is the status
571 571 # quo fine?
572 572
573 573 alltags = {} # map tag name to (node, hist)
574 574 tagtypes = {}
575 575
576 576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578 578
579 579 # Build the return dicts. Have to re-encode tag names because
580 580 # the tags module always uses UTF-8 (in order not to lose info
581 581 # writing to the cache), but the rest of Mercurial wants them in
582 582 # local encoding.
583 583 tags = {}
584 584 for (name, (node, hist)) in alltags.iteritems():
585 585 if node != nullid:
586 586 tags[encoding.tolocal(name)] = node
587 587 tags['tip'] = self.changelog.tip()
588 588 tagtypes = dict([(encoding.tolocal(name), value)
589 589 for (name, value) in tagtypes.iteritems()])
590 590 return (tags, tagtypes)
591 591
592 592 def tagtype(self, tagname):
593 593 '''
594 594 return the type of the given tag. result can be:
595 595
596 596 'local' : a local tag
597 597 'global' : a global tag
598 598 None : tag does not exist
599 599 '''
600 600
601 601 return self._tagscache.tagtypes.get(tagname)
602 602
603 603 def tagslist(self):
604 604 '''return a list of tags ordered by revision'''
605 605 if not self._tagscache.tagslist:
606 606 l = []
607 607 for t, n in self.tags().iteritems():
608 608 r = self.changelog.rev(n)
609 609 l.append((r, t, n))
610 610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611 611
612 612 return self._tagscache.tagslist
613 613
614 614 def nodetags(self, node):
615 615 '''return the tags associated with a node'''
616 616 if not self._tagscache.nodetagscache:
617 617 nodetagscache = {}
618 618 for t, n in self._tagscache.tags.iteritems():
619 619 nodetagscache.setdefault(n, []).append(t)
620 620 for tags in nodetagscache.itervalues():
621 621 tags.sort()
622 622 self._tagscache.nodetagscache = nodetagscache
623 623 return self._tagscache.nodetagscache.get(node, [])
624 624
625 625 def nodebookmarks(self, node):
626 626 marks = []
627 627 for bookmark, n in self._bookmarks.iteritems():
628 628 if n == node:
629 629 marks.append(bookmark)
630 630 return sorted(marks)
631 631
632 632 def branchmap(self):
633 633 '''returns a dictionary {branch: [branchheads]}'''
634 634 branchmap.updatecache(self)
635 635 return self._branchcaches[self.filtername]
636 636
637 637
638 638 def _branchtip(self, heads):
639 639 '''return the tipmost branch head in heads'''
640 640 tip = heads[-1]
641 641 for h in reversed(heads):
642 642 if not self[h].closesbranch():
643 643 tip = h
644 644 break
645 645 return tip
646 646
647 647 def branchtip(self, branch):
648 648 '''return the tip node for a given branch'''
649 649 if branch not in self.branchmap():
650 650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 651 return self._branchtip(self.branchmap()[branch])
652 652
653 653 def branchtags(self):
654 654 '''return a dict where branch names map to the tipmost head of
655 655 the branch, open heads come before closed'''
656 656 bt = {}
657 657 for bn, heads in self.branchmap().iteritems():
658 658 bt[bn] = self._branchtip(heads)
659 659 return bt
660 660
661 661 def lookup(self, key):
662 662 return self[key].node()
663 663
664 664 def lookupbranch(self, key, remote=None):
665 665 repo = remote or self
666 666 if key in repo.branchmap():
667 667 return key
668 668
669 669 repo = (remote and remote.local()) and remote or self
670 670 return repo[key].branch()
671 671
672 672 def known(self, nodes):
673 673 nm = self.changelog.nodemap
674 674 pc = self._phasecache
675 675 result = []
676 676 for n in nodes:
677 677 r = nm.get(n)
678 678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 679 result.append(resp)
680 680 return result
681 681
682 682 def local(self):
683 683 return self
684 684
685 685 def cancopy(self):
686 686 return self.local() # so statichttprepo's override of local() works
687 687
688 688 def join(self, f):
689 689 return os.path.join(self.path, f)
690 690
691 691 def wjoin(self, f):
692 692 return os.path.join(self.root, f)
693 693
694 694 def file(self, f):
695 695 if f[0] == '/':
696 696 f = f[1:]
697 697 return filelog.filelog(self.sopener, f)
698 698
699 699 def changectx(self, changeid):
700 700 return self[changeid]
701 701
702 702 def parents(self, changeid=None):
703 703 '''get list of changectxs for parents of changeid'''
704 704 return self[changeid].parents()
705 705
706 706 def setparents(self, p1, p2=nullid):
707 707 copies = self.dirstate.setparents(p1, p2)
708 708 if copies:
709 709 # Adjust copy records, the dirstate cannot do it, it
710 710 # requires access to parents manifests. Preserve them
711 711 # only for entries added to first parent.
712 712 pctx = self[p1]
713 713 for f in copies:
714 714 if f not in pctx and copies[f] in pctx:
715 715 self.dirstate.copy(copies[f], f)
716 716
717 717 def filectx(self, path, changeid=None, fileid=None):
718 718 """changeid can be a changeset revision, node, or tag.
719 719 fileid can be a file revision or node."""
720 720 return context.filectx(self, path, changeid, fileid)
721 721
722 722 def getcwd(self):
723 723 return self.dirstate.getcwd()
724 724
725 725 def pathto(self, f, cwd=None):
726 726 return self.dirstate.pathto(f, cwd)
727 727
728 728 def wfile(self, f, mode='r'):
729 729 return self.wopener(f, mode)
730 730
731 731 def _link(self, f):
732 732 return os.path.islink(self.wjoin(f))
733 733
734 734 def _loadfilter(self, filter):
735 735 if filter not in self.filterpats:
736 736 l = []
737 737 for pat, cmd in self.ui.configitems(filter):
738 738 if cmd == '!':
739 739 continue
740 740 mf = matchmod.match(self.root, '', [pat])
741 741 fn = None
742 742 params = cmd
743 743 for name, filterfn in self._datafilters.iteritems():
744 744 if cmd.startswith(name):
745 745 fn = filterfn
746 746 params = cmd[len(name):].lstrip()
747 747 break
748 748 if not fn:
749 749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 750 # Wrap old filters not supporting keyword arguments
751 751 if not inspect.getargspec(fn)[2]:
752 752 oldfn = fn
753 753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 754 l.append((mf, fn, params))
755 755 self.filterpats[filter] = l
756 756 return self.filterpats[filter]
757 757
758 758 def _filter(self, filterpats, filename, data):
759 759 for mf, fn, cmd in filterpats:
760 760 if mf(filename):
761 761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 763 break
764 764
765 765 return data
766 766
767 767 @unfilteredpropertycache
768 768 def _encodefilterpats(self):
769 769 return self._loadfilter('encode')
770 770
771 771 @unfilteredpropertycache
772 772 def _decodefilterpats(self):
773 773 return self._loadfilter('decode')
774 774
775 775 def adddatafilter(self, name, filter):
776 776 self._datafilters[name] = filter
777 777
778 778 def wread(self, filename):
779 779 if self._link(filename):
780 780 data = os.readlink(self.wjoin(filename))
781 781 else:
782 782 data = self.wopener.read(filename)
783 783 return self._filter(self._encodefilterpats, filename, data)
784 784
785 785 def wwrite(self, filename, data, flags):
786 786 data = self._filter(self._decodefilterpats, filename, data)
787 787 if 'l' in flags:
788 788 self.wopener.symlink(data, filename)
789 789 else:
790 790 self.wopener.write(filename, data)
791 791 if 'x' in flags:
792 792 util.setflags(self.wjoin(filename), False, True)
793 793
794 794 def wwritedata(self, filename, data):
795 795 return self._filter(self._decodefilterpats, filename, data)
796 796
797 797 def transaction(self, desc):
798 798 tr = self._transref and self._transref() or None
799 799 if tr and tr.running():
800 800 return tr.nest()
801 801
802 802 # abort here if the journal already exists
803 803 if os.path.exists(self.sjoin("journal")):
804 804 raise error.RepoError(
805 805 _("abandoned transaction found - run hg recover"))
806 806
807 807 self._writejournal(desc)
808 808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809 809
810 810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 811 self.sjoin("journal"),
812 812 aftertrans(renames),
813 813 self.store.createmode)
814 814 self._transref = weakref.ref(tr)
815 815 return tr
816 816
817 817 def _journalfiles(self):
818 818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 819 self.join('journal.branch'), self.join('journal.desc'),
820 820 self.join('journal.bookmarks'),
821 821 self.sjoin('journal.phaseroots'))
822 822
823 823 def undofiles(self):
824 824 return [undoname(x) for x in self._journalfiles()]
825 825
826 826 def _writejournal(self, desc):
827 827 self.opener.write("journal.dirstate",
828 828 self.opener.tryread("dirstate"))
829 829 self.opener.write("journal.branch",
830 830 encoding.fromlocal(self.dirstate.branch()))
831 831 self.opener.write("journal.desc",
832 832 "%d\n%s\n" % (len(self), desc))
833 833 self.opener.write("journal.bookmarks",
834 834 self.opener.tryread("bookmarks"))
835 835 self.sopener.write("journal.phaseroots",
836 836 self.sopener.tryread("phaseroots"))
837 837
838 838 def recover(self):
839 839 lock = self.lock()
840 840 try:
841 841 if os.path.exists(self.sjoin("journal")):
842 842 self.ui.status(_("rolling back interrupted transaction\n"))
843 843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 844 self.ui.warn)
845 845 self.invalidate()
846 846 return True
847 847 else:
848 848 self.ui.warn(_("no interrupted transaction available\n"))
849 849 return False
850 850 finally:
851 851 lock.release()
852 852
853 853 def rollback(self, dryrun=False, force=False):
854 854 wlock = lock = None
855 855 try:
856 856 wlock = self.wlock()
857 857 lock = self.lock()
858 858 if os.path.exists(self.sjoin("undo")):
859 859 return self._rollback(dryrun, force)
860 860 else:
861 861 self.ui.warn(_("no rollback information available\n"))
862 862 return 1
863 863 finally:
864 864 release(lock, wlock)
865 865
866 866 @unfilteredmethod # Until we get smarter cache management
867 867 def _rollback(self, dryrun, force):
868 868 ui = self.ui
869 869 try:
870 870 args = self.opener.read('undo.desc').splitlines()
871 871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 872 if len(args) >= 3:
873 873 detail = args[2]
874 874 oldtip = oldlen - 1
875 875
876 876 if detail and ui.verbose:
877 877 msg = (_('repository tip rolled back to revision %s'
878 878 ' (undo %s: %s)\n')
879 879 % (oldtip, desc, detail))
880 880 else:
881 881 msg = (_('repository tip rolled back to revision %s'
882 882 ' (undo %s)\n')
883 883 % (oldtip, desc))
884 884 except IOError:
885 885 msg = _('rolling back unknown transaction\n')
886 886 desc = None
887 887
888 888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 889 raise util.Abort(
890 890 _('rollback of last commit while not checked out '
891 891 'may lose data'), hint=_('use -f to force'))
892 892
893 893 ui.status(msg)
894 894 if dryrun:
895 895 return 0
896 896
897 897 parents = self.dirstate.parents()
898 898 self.destroying()
899 899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 900 if os.path.exists(self.join('undo.bookmarks')):
901 901 util.rename(self.join('undo.bookmarks'),
902 902 self.join('bookmarks'))
903 903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 904 util.rename(self.sjoin('undo.phaseroots'),
905 905 self.sjoin('phaseroots'))
906 906 self.invalidate()
907 907
908 908 parentgone = (parents[0] not in self.changelog.nodemap or
909 909 parents[1] not in self.changelog.nodemap)
910 910 if parentgone:
911 911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
912 912 try:
913 913 branch = self.opener.read('undo.branch')
914 914 self.dirstate.setbranch(encoding.tolocal(branch))
915 915 except IOError:
916 916 ui.warn(_('named branch could not be reset: '
917 917 'current branch is still \'%s\'\n')
918 918 % self.dirstate.branch())
919 919
920 920 self.dirstate.invalidate()
921 921 parents = tuple([p.rev() for p in self.parents()])
922 922 if len(parents) > 1:
923 923 ui.status(_('working directory now based on '
924 924 'revisions %d and %d\n') % parents)
925 925 else:
926 926 ui.status(_('working directory now based on '
927 927 'revision %d\n') % parents)
928 928 # TODO: if we know which new heads may result from this rollback, pass
929 929 # them to destroy(), which will prevent the branchhead cache from being
930 930 # invalidated.
931 931 self.destroyed()
932 932 return 0
933 933
934 934 def invalidatecaches(self):
935 935
936 936 if '_tagscache' in vars(self):
937 937 # can't use delattr on proxy
938 938 del self.__dict__['_tagscache']
939 939
940 940 self.unfiltered()._branchcaches.clear()
941 941 self.invalidatevolatilesets()
942 942
943 943 def invalidatevolatilesets(self):
944 944 self.filteredrevcache.clear()
945 945 obsolete.clearobscaches(self)
946 946
947 947 def invalidatedirstate(self):
948 948 '''Invalidates the dirstate, causing the next call to dirstate
949 949 to check if it was modified since the last time it was read,
950 950 rereading it if it has.
951 951
952 952 This is different to dirstate.invalidate() that it doesn't always
953 953 rereads the dirstate. Use dirstate.invalidate() if you want to
954 954 explicitly read the dirstate again (i.e. restoring it to a previous
955 955 known good state).'''
956 956 if hasunfilteredcache(self, 'dirstate'):
957 957 for k in self.dirstate._filecache:
958 958 try:
959 959 delattr(self.dirstate, k)
960 960 except AttributeError:
961 961 pass
962 962 delattr(self.unfiltered(), 'dirstate')
963 963
964 964 def invalidate(self):
965 965 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
966 966 for k in self._filecache:
967 967 # dirstate is invalidated separately in invalidatedirstate()
968 968 if k == 'dirstate':
969 969 continue
970 970
971 971 try:
972 972 delattr(unfiltered, k)
973 973 except AttributeError:
974 974 pass
975 975 self.invalidatecaches()
976 976
977 977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
978 978 try:
979 979 l = lock.lock(lockname, 0, releasefn, desc=desc)
980 980 except error.LockHeld, inst:
981 981 if not wait:
982 982 raise
983 983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
984 984 (desc, inst.locker))
985 985 # default to 600 seconds timeout
986 986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
987 987 releasefn, desc=desc)
988 988 if acquirefn:
989 989 acquirefn()
990 990 return l
991 991
992 992 def _afterlock(self, callback):
993 993 """add a callback to the current repository lock.
994 994
995 995 The callback will be executed on lock release."""
996 996 l = self._lockref and self._lockref()
997 997 if l:
998 998 l.postrelease.append(callback)
999 999 else:
1000 1000 callback()
1001 1001
1002 1002 def lock(self, wait=True):
1003 1003 '''Lock the repository store (.hg/store) and return a weak reference
1004 1004 to the lock. Use this before modifying the store (e.g. committing or
1005 1005 stripping). If you are opening a transaction, get a lock as well.)'''
1006 1006 l = self._lockref and self._lockref()
1007 1007 if l is not None and l.held:
1008 1008 l.lock()
1009 1009 return l
1010 1010
1011 1011 def unlock():
1012 1012 self.store.write()
1013 1013 if hasunfilteredcache(self, '_phasecache'):
1014 1014 self._phasecache.write()
1015 1015 for k, ce in self._filecache.items():
1016 1016 if k == 'dirstate' or k not in self.__dict__:
1017 1017 continue
1018 1018 ce.refresh()
1019 1019
1020 1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1021 1021 self.invalidate, _('repository %s') % self.origroot)
1022 1022 self._lockref = weakref.ref(l)
1023 1023 return l
1024 1024
1025 1025 def wlock(self, wait=True):
1026 1026 '''Lock the non-store parts of the repository (everything under
1027 1027 .hg except .hg/store) and return a weak reference to the lock.
1028 1028 Use this before modifying files in .hg.'''
1029 1029 l = self._wlockref and self._wlockref()
1030 1030 if l is not None and l.held:
1031 1031 l.lock()
1032 1032 return l
1033 1033
1034 1034 def unlock():
1035 1035 self.dirstate.write()
1036 1036 self._filecache['dirstate'].refresh()
1037 1037
1038 1038 l = self._lock(self.join("wlock"), wait, unlock,
1039 1039 self.invalidatedirstate, _('working directory of %s') %
1040 1040 self.origroot)
1041 1041 self._wlockref = weakref.ref(l)
1042 1042 return l
1043 1043
1044 1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1045 1045 """
1046 1046 commit an individual file as part of a larger transaction
1047 1047 """
1048 1048
1049 1049 fname = fctx.path()
1050 1050 text = fctx.data()
1051 1051 flog = self.file(fname)
1052 1052 fparent1 = manifest1.get(fname, nullid)
1053 1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1054 1054
1055 1055 meta = {}
1056 1056 copy = fctx.renamed()
1057 1057 if copy and copy[0] != fname:
1058 1058 # Mark the new revision of this file as a copy of another
1059 1059 # file. This copy data will effectively act as a parent
1060 1060 # of this new revision. If this is a merge, the first
1061 1061 # parent will be the nullid (meaning "look up the copy data")
1062 1062 # and the second one will be the other parent. For example:
1063 1063 #
1064 1064 # 0 --- 1 --- 3 rev1 changes file foo
1065 1065 # \ / rev2 renames foo to bar and changes it
1066 1066 # \- 2 -/ rev3 should have bar with all changes and
1067 1067 # should record that bar descends from
1068 1068 # bar in rev2 and foo in rev1
1069 1069 #
1070 1070 # this allows this merge to succeed:
1071 1071 #
1072 1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1073 1073 # \ / merging rev3 and rev4 should use bar@rev2
1074 1074 # \- 2 --- 4 as the merge base
1075 1075 #
1076 1076
1077 1077 cfname = copy[0]
1078 1078 crev = manifest1.get(cfname)
1079 1079 newfparent = fparent2
1080 1080
1081 1081 if manifest2: # branch merge
1082 1082 if fparent2 == nullid or crev is None: # copied on remote side
1083 1083 if cfname in manifest2:
1084 1084 crev = manifest2[cfname]
1085 1085 newfparent = fparent1
1086 1086
1087 1087 # find source in nearest ancestor if we've lost track
1088 1088 if not crev:
1089 1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1090 1090 (fname, cfname))
1091 1091 for ancestor in self[None].ancestors():
1092 1092 if cfname in ancestor:
1093 1093 crev = ancestor[cfname].filenode()
1094 1094 break
1095 1095
1096 1096 if crev:
1097 1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1098 1098 meta["copy"] = cfname
1099 1099 meta["copyrev"] = hex(crev)
1100 1100 fparent1, fparent2 = nullid, newfparent
1101 1101 else:
1102 1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1103 1103 "copied from '%s'!\n") % (fname, cfname))
1104 1104
1105 1105 elif fparent2 != nullid:
1106 1106 # is one parent an ancestor of the other?
1107 1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1108 1108 if fparentancestor == fparent1:
1109 1109 fparent1, fparent2 = fparent2, nullid
1110 1110 elif fparentancestor == fparent2:
1111 1111 fparent2 = nullid
1112 1112
1113 1113 # is the file changed?
1114 1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1115 1115 changelist.append(fname)
1116 1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1117 1117
1118 1118 # are just the flags changed during merge?
1119 1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1120 1120 changelist.append(fname)
1121 1121
1122 1122 return fparent1
1123 1123
1124 1124 @unfilteredmethod
1125 1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1126 1126 editor=False, extra={}):
1127 1127 """Add a new revision to current repository.
1128 1128
1129 1129 Revision information is gathered from the working directory,
1130 1130 match can be used to filter the committed files. If editor is
1131 1131 supplied, it is called to get a commit message.
1132 1132 """
1133 1133
1134 1134 def fail(f, msg):
1135 1135 raise util.Abort('%s: %s' % (f, msg))
1136 1136
1137 1137 if not match:
1138 1138 match = matchmod.always(self.root, '')
1139 1139
1140 1140 if not force:
1141 1141 vdirs = []
1142 1142 match.dir = vdirs.append
1143 1143 match.bad = fail
1144 1144
1145 1145 wlock = self.wlock()
1146 1146 try:
1147 1147 wctx = self[None]
1148 1148 merge = len(wctx.parents()) > 1
1149 1149
1150 1150 if (not force and merge and match and
1151 1151 (match.files() or match.anypats())):
1152 1152 raise util.Abort(_('cannot partially commit a merge '
1153 1153 '(do not specify files or patterns)'))
1154 1154
1155 1155 changes = self.status(match=match, clean=force)
1156 1156 if force:
1157 1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1158 1158
1159 1159 # check subrepos
1160 1160 subs = []
1161 1161 commitsubs = set()
1162 1162 newstate = wctx.substate.copy()
1163 1163 # only manage subrepos and .hgsubstate if .hgsub is present
1164 1164 if '.hgsub' in wctx:
1165 1165 # we'll decide whether to track this ourselves, thanks
1166 1166 if '.hgsubstate' in changes[0]:
1167 1167 changes[0].remove('.hgsubstate')
1168 1168 if '.hgsubstate' in changes[2]:
1169 1169 changes[2].remove('.hgsubstate')
1170 1170
1171 1171 # compare current state to last committed state
1172 1172 # build new substate based on last committed state
1173 1173 oldstate = wctx.p1().substate
1174 1174 for s in sorted(newstate.keys()):
1175 1175 if not match(s):
1176 1176 # ignore working copy, use old state if present
1177 1177 if s in oldstate:
1178 1178 newstate[s] = oldstate[s]
1179 1179 continue
1180 1180 if not force:
1181 1181 raise util.Abort(
1182 1182 _("commit with new subrepo %s excluded") % s)
1183 1183 if wctx.sub(s).dirty(True):
1184 1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1185 1185 raise util.Abort(
1186 1186 _("uncommitted changes in subrepo %s") % s,
1187 1187 hint=_("use --subrepos for recursive commit"))
1188 1188 subs.append(s)
1189 1189 commitsubs.add(s)
1190 1190 else:
1191 1191 bs = wctx.sub(s).basestate()
1192 1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1193 1193 if oldstate.get(s, (None, None, None))[1] != bs:
1194 1194 subs.append(s)
1195 1195
1196 1196 # check for removed subrepos
1197 1197 for p in wctx.parents():
1198 1198 r = [s for s in p.substate if s not in newstate]
1199 1199 subs += [s for s in r if match(s)]
1200 1200 if subs:
1201 1201 if (not match('.hgsub') and
1202 1202 '.hgsub' in (wctx.modified() + wctx.added())):
1203 1203 raise util.Abort(
1204 1204 _("can't commit subrepos without .hgsub"))
1205 1205 changes[0].insert(0, '.hgsubstate')
1206 1206
1207 1207 elif '.hgsub' in changes[2]:
1208 1208 # clean up .hgsubstate when .hgsub is removed
1209 1209 if ('.hgsubstate' in wctx and
1210 1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1211 1211 changes[2].insert(0, '.hgsubstate')
1212 1212
1213 1213 # make sure all explicit patterns are matched
1214 1214 if not force and match.files():
1215 1215 matched = set(changes[0] + changes[1] + changes[2])
1216 1216
1217 1217 for f in match.files():
1218 1218 f = self.dirstate.normalize(f)
1219 1219 if f == '.' or f in matched or f in wctx.substate:
1220 1220 continue
1221 1221 if f in changes[3]: # missing
1222 1222 fail(f, _('file not found!'))
1223 1223 if f in vdirs: # visited directory
1224 1224 d = f + '/'
1225 1225 for mf in matched:
1226 1226 if mf.startswith(d):
1227 1227 break
1228 1228 else:
1229 1229 fail(f, _("no match under directory!"))
1230 1230 elif f not in self.dirstate:
1231 1231 fail(f, _("file not tracked!"))
1232 1232
1233 1233 if (not force and not extra.get("close") and not merge
1234 1234 and not (changes[0] or changes[1] or changes[2])
1235 1235 and wctx.branch() == wctx.p1().branch()):
1236 1236 return None
1237 1237
1238 1238 if merge and changes[3]:
1239 1239 raise util.Abort(_("cannot commit merge with missing files"))
1240 1240
1241 1241 ms = mergemod.mergestate(self)
1242 1242 for f in changes[0]:
1243 1243 if f in ms and ms[f] == 'u':
1244 1244 raise util.Abort(_("unresolved merge conflicts "
1245 1245 "(see hg help resolve)"))
1246 1246
1247 1247 cctx = context.workingctx(self, text, user, date, extra, changes)
1248 1248 if editor:
1249 1249 cctx._text = editor(self, cctx, subs)
1250 1250 edited = (text != cctx._text)
1251 1251
1252 1252 # commit subs and write new state
1253 1253 if subs:
1254 1254 for s in sorted(commitsubs):
1255 1255 sub = wctx.sub(s)
1256 1256 self.ui.status(_('committing subrepository %s\n') %
1257 1257 subrepo.subrelpath(sub))
1258 1258 sr = sub.commit(cctx._text, user, date)
1259 1259 newstate[s] = (newstate[s][0], sr)
1260 1260 subrepo.writestate(self, newstate)
1261 1261
1262 1262 # Save commit message in case this transaction gets rolled back
1263 1263 # (e.g. by a pretxncommit hook). Leave the content alone on
1264 1264 # the assumption that the user will use the same editor again.
1265 1265 msgfn = self.savecommitmessage(cctx._text)
1266 1266
1267 1267 p1, p2 = self.dirstate.parents()
1268 1268 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1269 1269 try:
1270 1270 self.hook("precommit", throw=True, parent1=hookp1,
1271 1271 parent2=hookp2)
1272 1272 ret = self.commitctx(cctx, True)
1273 1273 except: # re-raises
1274 1274 if edited:
1275 1275 self.ui.write(
1276 1276 _('note: commit message saved in %s\n') % msgfn)
1277 1277 raise
1278 1278
1279 1279 # update bookmarks, dirstate and mergestate
1280 1280 bookmarks.update(self, [p1, p2], ret)
1281 1281 for f in changes[0] + changes[1]:
1282 1282 self.dirstate.normal(f)
1283 1283 for f in changes[2]:
1284 1284 self.dirstate.drop(f)
1285 1285 self.dirstate.setparents(ret)
1286 1286 ms.reset()
1287 1287 finally:
1288 1288 wlock.release()
1289 1289
1290 1290 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1291 1291 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1292 1292 self._afterlock(commithook)
1293 1293 return ret
1294 1294
1295 1295 @unfilteredmethod
1296 1296 def commitctx(self, ctx, error=False):
1297 1297 """Add a new revision to current repository.
1298 1298 Revision information is passed via the context argument.
1299 1299 """
1300 1300
1301 1301 tr = lock = None
1302 1302 removed = list(ctx.removed())
1303 1303 p1, p2 = ctx.p1(), ctx.p2()
1304 1304 user = ctx.user()
1305 1305
1306 1306 lock = self.lock()
1307 1307 try:
1308 1308 tr = self.transaction("commit")
1309 1309 trp = weakref.proxy(tr)
1310 1310
1311 1311 if ctx.files():
1312 1312 m1 = p1.manifest().copy()
1313 1313 m2 = p2.manifest()
1314 1314
1315 1315 # check in files
1316 1316 new = {}
1317 1317 changed = []
1318 1318 linkrev = len(self)
1319 1319 for f in sorted(ctx.modified() + ctx.added()):
1320 1320 self.ui.note(f + "\n")
1321 1321 try:
1322 1322 fctx = ctx[f]
1323 1323 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1324 1324 changed)
1325 1325 m1.set(f, fctx.flags())
1326 1326 except OSError, inst:
1327 1327 self.ui.warn(_("trouble committing %s!\n") % f)
1328 1328 raise
1329 1329 except IOError, inst:
1330 1330 errcode = getattr(inst, 'errno', errno.ENOENT)
1331 1331 if error or errcode and errcode != errno.ENOENT:
1332 1332 self.ui.warn(_("trouble committing %s!\n") % f)
1333 1333 raise
1334 1334 else:
1335 1335 removed.append(f)
1336 1336
1337 1337 # update manifest
1338 1338 m1.update(new)
1339 1339 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1340 1340 drop = [f for f in removed if f in m1]
1341 1341 for f in drop:
1342 1342 del m1[f]
1343 1343 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1344 1344 p2.manifestnode(), (new, drop))
1345 1345 files = changed + removed
1346 1346 else:
1347 1347 mn = p1.manifestnode()
1348 1348 files = []
1349 1349
1350 1350 # update changelog
1351 1351 self.changelog.delayupdate()
1352 1352 n = self.changelog.add(mn, files, ctx.description(),
1353 1353 trp, p1.node(), p2.node(),
1354 1354 user, ctx.date(), ctx.extra().copy())
1355 1355 p = lambda: self.changelog.writepending() and self.root or ""
1356 1356 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1357 1357 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1358 1358 parent2=xp2, pending=p)
1359 1359 self.changelog.finalize(trp)
1360 1360 # set the new commit is proper phase
1361 1361 targetphase = phases.newcommitphase(self.ui)
1362 1362 if targetphase:
1363 1363 # retract boundary do not alter parent changeset.
1364 1364 # if a parent have higher the resulting phase will
1365 1365 # be compliant anyway
1366 1366 #
1367 1367 # if minimal phase was 0 we don't need to retract anything
1368 1368 phases.retractboundary(self, targetphase, [n])
1369 1369 tr.close()
1370 1370 branchmap.updatecache(self.filtered('served'))
1371 1371 return n
1372 1372 finally:
1373 1373 if tr:
1374 1374 tr.release()
1375 1375 lock.release()
1376 1376
1377 1377 @unfilteredmethod
1378 1378 def destroying(self):
1379 1379 '''Inform the repository that nodes are about to be destroyed.
1380 1380 Intended for use by strip and rollback, so there's a common
1381 1381 place for anything that has to be done before destroying history.
1382 1382
1383 1383 This is mostly useful for saving state that is in memory and waiting
1384 1384 to be flushed when the current lock is released. Because a call to
1385 1385 destroyed is imminent, the repo will be invalidated causing those
1386 1386 changes to stay in memory (waiting for the next unlock), or vanish
1387 1387 completely.
1388 1388 '''
1389 1389 # When using the same lock to commit and strip, the phasecache is left
1390 1390 # dirty after committing. Then when we strip, the repo is invalidated,
1391 1391 # causing those changes to disappear.
1392 1392 if '_phasecache' in vars(self):
1393 1393 self._phasecache.write()
1394 1394
1395 1395 @unfilteredmethod
1396 1396 def destroyed(self):
1397 1397 '''Inform the repository that nodes have been destroyed.
1398 1398 Intended for use by strip and rollback, so there's a common
1399 1399 place for anything that has to be done after destroying history.
1400 1400
1401 1401 If you know the branchheadcache was uptodate before nodes were removed
1402 1402 and you also know the set of candidate new heads that may have resulted
1403 1403 from the destruction, you can set newheadnodes. This will enable the
1404 1404 code to update the branchheads cache, rather than having future code
1405 1405 decide it's invalid and regenerating it from scratch.
1406 1406 '''
1407 1407 # When one tries to:
1408 1408 # 1) destroy nodes thus calling this method (e.g. strip)
1409 1409 # 2) use phasecache somewhere (e.g. commit)
1410 1410 #
1411 1411 # then 2) will fail because the phasecache contains nodes that were
1412 1412 # removed. We can either remove phasecache from the filecache,
1413 1413 # causing it to reload next time it is accessed, or simply filter
1414 1414 # the removed nodes now and write the updated cache.
1415 1415 if '_phasecache' in self._filecache:
1416 1416 self._phasecache.filterunknown(self)
1417 1417 self._phasecache.write()
1418 1418
1419 1419 # update the 'served' branch cache to help read only server process
1420 1420 # Thanks to branchcach collaboration this is done from the nearest
1421 1421 # filtered subset and it is expected to be fast.
1422 1422 branchmap.updatecache(self.filtered('served'))
1423 1423
1424 1424 # Ensure the persistent tag cache is updated. Doing it now
1425 1425 # means that the tag cache only has to worry about destroyed
1426 1426 # heads immediately after a strip/rollback. That in turn
1427 1427 # guarantees that "cachetip == currenttip" (comparing both rev
1428 1428 # and node) always means no nodes have been added or destroyed.
1429 1429
1430 1430 # XXX this is suboptimal when qrefresh'ing: we strip the current
1431 1431 # head, refresh the tag cache, then immediately add a new head.
1432 1432 # But I think doing it this way is necessary for the "instant
1433 1433 # tag cache retrieval" case to work.
1434 1434 self.invalidate()
1435 1435
1436 1436 def walk(self, match, node=None):
1437 1437 '''
1438 1438 walk recursively through the directory tree or a given
1439 1439 changeset, finding all files matched by the match
1440 1440 function
1441 1441 '''
1442 1442 return self[node].walk(match)
1443 1443
1444 1444 def status(self, node1='.', node2=None, match=None,
1445 1445 ignored=False, clean=False, unknown=False,
1446 1446 listsubrepos=False):
1447 1447 """return status of files between two nodes or node and working
1448 1448 directory.
1449 1449
1450 1450 If node1 is None, use the first dirstate parent instead.
1451 1451 If node2 is None, compare node1 with working directory.
1452 1452 """
1453 1453
1454 1454 def mfmatches(ctx):
1455 1455 mf = ctx.manifest().copy()
1456 1456 if match.always():
1457 1457 return mf
1458 1458 for fn in mf.keys():
1459 1459 if not match(fn):
1460 1460 del mf[fn]
1461 1461 return mf
1462 1462
1463 1463 if isinstance(node1, context.changectx):
1464 1464 ctx1 = node1
1465 1465 else:
1466 1466 ctx1 = self[node1]
1467 1467 if isinstance(node2, context.changectx):
1468 1468 ctx2 = node2
1469 1469 else:
1470 1470 ctx2 = self[node2]
1471 1471
1472 1472 working = ctx2.rev() is None
1473 1473 parentworking = working and ctx1 == self['.']
1474 1474 match = match or matchmod.always(self.root, self.getcwd())
1475 1475 listignored, listclean, listunknown = ignored, clean, unknown
1476 1476
1477 1477 # load earliest manifest first for caching reasons
1478 1478 if not working and ctx2.rev() < ctx1.rev():
1479 1479 ctx2.manifest()
1480 1480
1481 1481 if not parentworking:
1482 1482 def bad(f, msg):
1483 1483 # 'f' may be a directory pattern from 'match.files()',
1484 1484 # so 'f not in ctx1' is not enough
1485 1485 if f not in ctx1 and f not in ctx1.dirs():
1486 1486 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1487 1487 match.bad = bad
1488 1488
1489 1489 if working: # we need to scan the working dir
1490 1490 subrepos = []
1491 1491 if '.hgsub' in self.dirstate:
1492 1492 subrepos = sorted(ctx2.substate)
1493 1493 s = self.dirstate.status(match, subrepos, listignored,
1494 1494 listclean, listunknown)
1495 1495 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1496 1496
1497 1497 # check for any possibly clean files
1498 1498 if parentworking and cmp:
1499 1499 fixup = []
1500 1500 # do a full compare of any files that might have changed
1501 1501 for f in sorted(cmp):
1502 1502 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1503 1503 or ctx1[f].cmp(ctx2[f])):
1504 1504 modified.append(f)
1505 1505 else:
1506 1506 fixup.append(f)
1507 1507
1508 1508 # update dirstate for files that are actually clean
1509 1509 if fixup:
1510 1510 if listclean:
1511 1511 clean += fixup
1512 1512
1513 1513 try:
1514 1514 # updating the dirstate is optional
1515 1515 # so we don't wait on the lock
1516 1516 wlock = self.wlock(False)
1517 1517 try:
1518 1518 for f in fixup:
1519 1519 self.dirstate.normal(f)
1520 1520 finally:
1521 1521 wlock.release()
1522 1522 except error.LockError:
1523 1523 pass
1524 1524
1525 1525 if not parentworking:
1526 1526 mf1 = mfmatches(ctx1)
1527 1527 if working:
1528 1528 # we are comparing working dir against non-parent
1529 1529 # generate a pseudo-manifest for the working dir
1530 1530 mf2 = mfmatches(self['.'])
1531 1531 for f in cmp + modified + added:
1532 1532 mf2[f] = None
1533 1533 mf2.set(f, ctx2.flags(f))
1534 1534 for f in removed:
1535 1535 if f in mf2:
1536 1536 del mf2[f]
1537 1537 else:
1538 1538 # we are comparing two revisions
1539 1539 deleted, unknown, ignored = [], [], []
1540 1540 mf2 = mfmatches(ctx2)
1541 1541
1542 1542 modified, added, clean = [], [], []
1543 1543 withflags = mf1.withflags() | mf2.withflags()
1544 1544 for fn in mf2:
1545 1545 if fn in mf1:
1546 1546 if (fn not in deleted and
1547 1547 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1548 1548 (mf1[fn] != mf2[fn] and
1549 1549 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1550 1550 modified.append(fn)
1551 1551 elif listclean:
1552 1552 clean.append(fn)
1553 1553 del mf1[fn]
1554 1554 elif fn not in deleted:
1555 1555 added.append(fn)
1556 1556 removed = mf1.keys()
1557 1557
1558 1558 if working and modified and not self.dirstate._checklink:
1559 1559 # Symlink placeholders may get non-symlink-like contents
1560 1560 # via user error or dereferencing by NFS or Samba servers,
1561 1561 # so we filter out any placeholders that don't look like a
1562 1562 # symlink
1563 1563 sane = []
1564 1564 for f in modified:
1565 1565 if ctx2.flags(f) == 'l':
1566 1566 d = ctx2[f].data()
1567 1567 if len(d) >= 1024 or '\n' in d or util.binary(d):
1568 1568 self.ui.debug('ignoring suspect symlink placeholder'
1569 1569 ' "%s"\n' % f)
1570 1570 continue
1571 1571 sane.append(f)
1572 1572 modified = sane
1573 1573
1574 1574 r = modified, added, removed, deleted, unknown, ignored, clean
1575 1575
1576 1576 if listsubrepos:
1577 1577 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1578 1578 if working:
1579 1579 rev2 = None
1580 1580 else:
1581 1581 rev2 = ctx2.substate[subpath][1]
1582 1582 try:
1583 1583 submatch = matchmod.narrowmatcher(subpath, match)
1584 1584 s = sub.status(rev2, match=submatch, ignored=listignored,
1585 1585 clean=listclean, unknown=listunknown,
1586 1586 listsubrepos=True)
1587 1587 for rfiles, sfiles in zip(r, s):
1588 1588 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1589 1589 except error.LookupError:
1590 1590 self.ui.status(_("skipping missing subrepository: %s\n")
1591 1591 % subpath)
1592 1592
1593 1593 for l in r:
1594 1594 l.sort()
1595 1595 return r
1596 1596
1597 1597 def heads(self, start=None):
1598 1598 heads = self.changelog.heads(start)
1599 1599 # sort the output in rev descending order
1600 1600 return sorted(heads, key=self.changelog.rev, reverse=True)
1601 1601
1602 1602 def branchheads(self, branch=None, start=None, closed=False):
1603 1603 '''return a (possibly filtered) list of heads for the given branch
1604 1604
1605 1605 Heads are returned in topological order, from newest to oldest.
1606 1606 If branch is None, use the dirstate branch.
1607 1607 If start is not None, return only heads reachable from start.
1608 1608 If closed is True, return heads that are marked as closed as well.
1609 1609 '''
1610 1610 if branch is None:
1611 1611 branch = self[None].branch()
1612 1612 branches = self.branchmap()
1613 1613 if branch not in branches:
1614 1614 return []
1615 1615 # the cache returns heads ordered lowest to highest
1616 1616 bheads = list(reversed(branches[branch]))
1617 1617 if start is not None:
1618 1618 # filter out the heads that cannot be reached from startrev
1619 1619 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1620 1620 bheads = [h for h in bheads if h in fbheads]
1621 1621 if not closed:
1622 1622 bheads = [h for h in bheads if not self[h].closesbranch()]
1623 1623 return bheads
1624 1624
1625 1625 def branches(self, nodes):
1626 1626 if not nodes:
1627 1627 nodes = [self.changelog.tip()]
1628 1628 b = []
1629 1629 for n in nodes:
1630 1630 t = n
1631 1631 while True:
1632 1632 p = self.changelog.parents(n)
1633 1633 if p[1] != nullid or p[0] == nullid:
1634 1634 b.append((t, n, p[0], p[1]))
1635 1635 break
1636 1636 n = p[0]
1637 1637 return b
1638 1638
1639 1639 def between(self, pairs):
1640 1640 r = []
1641 1641
1642 1642 for top, bottom in pairs:
1643 1643 n, l, i = top, [], 0
1644 1644 f = 1
1645 1645
1646 1646 while n != bottom and n != nullid:
1647 1647 p = self.changelog.parents(n)[0]
1648 1648 if i == f:
1649 1649 l.append(n)
1650 1650 f = f * 2
1651 1651 n = p
1652 1652 i += 1
1653 1653
1654 1654 r.append(l)
1655 1655
1656 1656 return r
1657 1657
1658 1658 def pull(self, remote, heads=None, force=False):
1659 1659 # don't open transaction for nothing or you break future useful
1660 1660 # rollback call
1661 1661 tr = None
1662 1662 trname = 'pull\n' + util.hidepassword(remote.url())
1663 1663 lock = self.lock()
1664 1664 try:
1665 1665 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1666 1666 force=force)
1667 1667 common, fetch, rheads = tmp
1668 1668 if not fetch:
1669 1669 self.ui.status(_("no changes found\n"))
1670 1670 added = []
1671 1671 result = 0
1672 1672 else:
1673 1673 tr = self.transaction(trname)
1674 1674 if heads is None and list(common) == [nullid]:
1675 1675 self.ui.status(_("requesting all changes\n"))
1676 1676 elif heads is None and remote.capable('changegroupsubset'):
1677 1677 # issue1320, avoid a race if remote changed after discovery
1678 1678 heads = rheads
1679 1679
1680 1680 if remote.capable('getbundle'):
1681 1681 cg = remote.getbundle('pull', common=common,
1682 1682 heads=heads or rheads)
1683 1683 elif heads is None:
1684 1684 cg = remote.changegroup(fetch, 'pull')
1685 1685 elif not remote.capable('changegroupsubset'):
1686 1686 raise util.Abort(_("partial pull cannot be done because "
1687 1687 "other repository doesn't support "
1688 1688 "changegroupsubset."))
1689 1689 else:
1690 1690 cg = remote.changegroupsubset(fetch, heads, 'pull')
1691 1691 clstart = len(self.changelog)
1692 1692 result = self.addchangegroup(cg, 'pull', remote.url())
1693 1693 clend = len(self.changelog)
1694 1694 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1695 1695
1696 1696 # compute target subset
1697 1697 if heads is None:
1698 1698 # We pulled every thing possible
1699 1699 # sync on everything common
1700 1700 subset = common + added
1701 1701 else:
1702 1702 # We pulled a specific subset
1703 1703 # sync on this subset
1704 1704 subset = heads
1705 1705
1706 1706 # Get remote phases data from remote
1707 1707 remotephases = remote.listkeys('phases')
1708 1708 publishing = bool(remotephases.get('publishing', False))
1709 1709 if remotephases and not publishing:
1710 1710 # remote is new and unpublishing
1711 1711 pheads, _dr = phases.analyzeremotephases(self, subset,
1712 1712 remotephases)
1713 1713 phases.advanceboundary(self, phases.public, pheads)
1714 1714 phases.advanceboundary(self, phases.draft, subset)
1715 1715 else:
1716 1716 # Remote is old or publishing all common changesets
1717 1717 # should be seen as public
1718 1718 phases.advanceboundary(self, phases.public, subset)
1719 1719
1720 1720 if obsolete._enabled:
1721 1721 self.ui.debug('fetching remote obsolete markers\n')
1722 1722 remoteobs = remote.listkeys('obsolete')
1723 1723 if 'dump0' in remoteobs:
1724 1724 if tr is None:
1725 1725 tr = self.transaction(trname)
1726 1726 for key in sorted(remoteobs, reverse=True):
1727 1727 if key.startswith('dump'):
1728 1728 data = base85.b85decode(remoteobs[key])
1729 1729 self.obsstore.mergemarkers(tr, data)
1730 1730 self.invalidatevolatilesets()
1731 1731 if tr is not None:
1732 1732 tr.close()
1733 1733 finally:
1734 1734 if tr is not None:
1735 1735 tr.release()
1736 1736 lock.release()
1737 1737
1738 1738 return result
1739 1739
1740 1740 def checkpush(self, force, revs):
1741 1741 """Extensions can override this function if additional checks have
1742 1742 to be performed before pushing, or call it if they override push
1743 1743 command.
1744 1744 """
1745 1745 pass
1746 1746
1747 1747 def push(self, remote, force=False, revs=None, newbranch=False):
1748 1748 '''Push outgoing changesets (limited by revs) from the current
1749 1749 repository to remote. Return an integer:
1750 1750 - None means nothing to push
1751 1751 - 0 means HTTP error
1752 1752 - 1 means we pushed and remote head count is unchanged *or*
1753 1753 we have outgoing changesets but refused to push
1754 1754 - other values as described by addchangegroup()
1755 1755 '''
1756 1756 # there are two ways to push to remote repo:
1757 1757 #
1758 1758 # addchangegroup assumes local user can lock remote
1759 1759 # repo (local filesystem, old ssh servers).
1760 1760 #
1761 1761 # unbundle assumes local user cannot lock remote repo (new ssh
1762 1762 # servers, http servers).
1763 1763
1764 1764 if not remote.canpush():
1765 1765 raise util.Abort(_("destination does not support push"))
1766 1766 unfi = self.unfiltered()
1767 1767 # get local lock as we might write phase data
1768 1768 locallock = self.lock()
1769 1769 try:
1770 1770 self.checkpush(force, revs)
1771 1771 lock = None
1772 1772 unbundle = remote.capable('unbundle')
1773 1773 if not unbundle:
1774 1774 lock = remote.lock()
1775 1775 try:
1776 1776 # discovery
1777 1777 fci = discovery.findcommonincoming
1778 1778 commoninc = fci(unfi, remote, force=force)
1779 1779 common, inc, remoteheads = commoninc
1780 1780 fco = discovery.findcommonoutgoing
1781 1781 outgoing = fco(unfi, remote, onlyheads=revs,
1782 1782 commoninc=commoninc, force=force)
1783 1783
1784 1784
1785 1785 if not outgoing.missing:
1786 1786 # nothing to push
1787 1787 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1788 1788 ret = None
1789 1789 else:
1790 1790 # something to push
1791 1791 if not force:
1792 1792 # if self.obsstore == False --> no obsolete
1793 1793 # then, save the iteration
1794 1794 if unfi.obsstore:
1795 1795 # this message are here for 80 char limit reason
1796 1796 mso = _("push includes obsolete changeset: %s!")
1797 1797 mst = "push includes %s changeset: %s!"
1798 1798 # plain versions for i18n tool to detect them
1799 1799 _("push includes unstable changeset: %s!")
1800 1800 _("push includes bumped changeset: %s!")
1801 1801 _("push includes divergent changeset: %s!")
1802 1802 # If we are to push if there is at least one
1803 1803 # obsolete or unstable changeset in missing, at
1804 1804 # least one of the missinghead will be obsolete or
1805 1805 # unstable. So checking heads only is ok
1806 1806 for node in outgoing.missingheads:
1807 1807 ctx = unfi[node]
1808 1808 if ctx.obsolete():
1809 1809 raise util.Abort(mso % ctx)
1810 1810 elif ctx.troubled():
1811 1811 raise util.Abort(_(mst)
1812 1812 % (ctx.troubles()[0],
1813 1813 ctx))
1814 1814 discovery.checkheads(unfi, remote, outgoing,
1815 1815 remoteheads, newbranch,
1816 1816 bool(inc))
1817 1817
1818 1818 # create a changegroup from local
1819 1819 if revs is None and not outgoing.excluded:
1820 1820 # push everything,
1821 1821 # use the fast path, no race possible on push
1822 1822 cg = self._changegroup(outgoing.missing, 'push')
1823 1823 else:
1824 1824 cg = self.getlocalbundle('push', outgoing)
1825 1825
1826 1826 # apply changegroup to remote
1827 1827 if unbundle:
1828 1828 # local repo finds heads on server, finds out what
1829 1829 # revs it must push. once revs transferred, if server
1830 1830 # finds it has different heads (someone else won
1831 1831 # commit/push race), server aborts.
1832 1832 if force:
1833 1833 remoteheads = ['force']
1834 1834 # ssh: return remote's addchangegroup()
1835 1835 # http: return remote's addchangegroup() or 0 for error
1836 1836 ret = remote.unbundle(cg, remoteheads, 'push')
1837 1837 else:
1838 1838 # we return an integer indicating remote head count
1839 1839 # change
1840 1840 ret = remote.addchangegroup(cg, 'push', self.url())
1841 1841
1842 1842 if ret:
1843 1843 # push succeed, synchronize target of the push
1844 1844 cheads = outgoing.missingheads
1845 1845 elif revs is None:
1846 1846 # All out push fails. synchronize all common
1847 1847 cheads = outgoing.commonheads
1848 1848 else:
1849 1849 # I want cheads = heads(::missingheads and ::commonheads)
1850 1850 # (missingheads is revs with secret changeset filtered out)
1851 1851 #
1852 1852 # This can be expressed as:
1853 1853 # cheads = ( (missingheads and ::commonheads)
1854 1854 # + (commonheads and ::missingheads))"
1855 1855 # )
1856 1856 #
1857 1857 # while trying to push we already computed the following:
1858 1858 # common = (::commonheads)
1859 1859 # missing = ((commonheads::missingheads) - commonheads)
1860 1860 #
1861 1861 # We can pick:
1862 1862 # * missingheads part of common (::commonheads)
1863 1863 common = set(outgoing.common)
1864 1864 cheads = [node for node in revs if node in common]
1865 1865 # and
1866 1866 # * commonheads parents on missing
1867 1867 revset = unfi.set('%ln and parents(roots(%ln))',
1868 1868 outgoing.commonheads,
1869 1869 outgoing.missing)
1870 1870 cheads.extend(c.node() for c in revset)
1871 1871 # even when we don't push, exchanging phase data is useful
1872 1872 remotephases = remote.listkeys('phases')
1873 1873 if not remotephases: # old server or public only repo
1874 1874 phases.advanceboundary(self, phases.public, cheads)
1875 1875 # don't push any phase data as there is nothing to push
1876 1876 else:
1877 1877 ana = phases.analyzeremotephases(self, cheads, remotephases)
1878 1878 pheads, droots = ana
1879 1879 ### Apply remote phase on local
1880 1880 if remotephases.get('publishing', False):
1881 1881 phases.advanceboundary(self, phases.public, cheads)
1882 1882 else: # publish = False
1883 1883 phases.advanceboundary(self, phases.public, pheads)
1884 1884 phases.advanceboundary(self, phases.draft, cheads)
1885 1885 ### Apply local phase on remote
1886 1886
1887 1887 # Get the list of all revs draft on remote by public here.
1888 1888 # XXX Beware that revset break if droots is not strictly
1889 1889 # XXX root we may want to ensure it is but it is costly
1890 1890 outdated = unfi.set('heads((%ln::%ln) and public())',
1891 1891 droots, cheads)
1892 1892 for newremotehead in outdated:
1893 1893 r = remote.pushkey('phases',
1894 1894 newremotehead.hex(),
1895 1895 str(phases.draft),
1896 1896 str(phases.public))
1897 1897 if not r:
1898 1898 self.ui.warn(_('updating %s to public failed!\n')
1899 1899 % newremotehead)
1900 1900 self.ui.debug('try to push obsolete markers to remote\n')
1901 1901 if (obsolete._enabled and self.obsstore and
1902 1902 'obsolete' in remote.listkeys('namespaces')):
1903 1903 rslts = []
1904 1904 remotedata = self.listkeys('obsolete')
1905 1905 for key in sorted(remotedata, reverse=True):
1906 1906 # reverse sort to ensure we end with dump0
1907 1907 data = remotedata[key]
1908 1908 rslts.append(remote.pushkey('obsolete', key, '', data))
1909 1909 if [r for r in rslts if not r]:
1910 1910 msg = _('failed to push some obsolete markers!\n')
1911 1911 self.ui.warn(msg)
1912 1912 finally:
1913 1913 if lock is not None:
1914 1914 lock.release()
1915 1915 finally:
1916 1916 locallock.release()
1917 1917
1918 1918 self.ui.debug("checking for updated bookmarks\n")
1919 1919 rb = remote.listkeys('bookmarks')
1920 1920 for k in rb.keys():
1921 1921 if k in unfi._bookmarks:
1922 1922 nr, nl = rb[k], hex(self._bookmarks[k])
1923 1923 if nr in unfi:
1924 1924 cr = unfi[nr]
1925 1925 cl = unfi[nl]
1926 1926 if bookmarks.validdest(unfi, cr, cl):
1927 1927 r = remote.pushkey('bookmarks', k, nr, nl)
1928 1928 if r:
1929 1929 self.ui.status(_("updating bookmark %s\n") % k)
1930 1930 else:
1931 1931 self.ui.warn(_('updating bookmark %s'
1932 1932 ' failed!\n') % k)
1933 1933
1934 1934 return ret
1935 1935
1936 1936 def changegroupinfo(self, nodes, source):
1937 1937 if self.ui.verbose or source == 'bundle':
1938 1938 self.ui.status(_("%d changesets found\n") % len(nodes))
1939 1939 if self.ui.debugflag:
1940 1940 self.ui.debug("list of changesets:\n")
1941 1941 for node in nodes:
1942 1942 self.ui.debug("%s\n" % hex(node))
1943 1943
1944 1944 def changegroupsubset(self, bases, heads, source):
1945 1945 """Compute a changegroup consisting of all the nodes that are
1946 1946 descendants of any of the bases and ancestors of any of the heads.
1947 1947 Return a chunkbuffer object whose read() method will return
1948 1948 successive changegroup chunks.
1949 1949
1950 1950 It is fairly complex as determining which filenodes and which
1951 1951 manifest nodes need to be included for the changeset to be complete
1952 1952 is non-trivial.
1953 1953
1954 1954 Another wrinkle is doing the reverse, figuring out which changeset in
1955 1955 the changegroup a particular filenode or manifestnode belongs to.
1956 1956 """
1957 1957 cl = self.changelog
1958 1958 if not bases:
1959 1959 bases = [nullid]
1960 1960 csets, bases, heads = cl.nodesbetween(bases, heads)
1961 1961 # We assume that all ancestors of bases are known
1962 1962 common = cl.ancestors([cl.rev(n) for n in bases])
1963 1963 return self._changegroupsubset(common, csets, heads, source)
1964 1964
1965 1965 def getlocalbundle(self, source, outgoing):
1966 1966 """Like getbundle, but taking a discovery.outgoing as an argument.
1967 1967
1968 1968 This is only implemented for local repos and reuses potentially
1969 1969 precomputed sets in outgoing."""
1970 1970 if not outgoing.missing:
1971 1971 return None
1972 1972 return self._changegroupsubset(outgoing.common,
1973 1973 outgoing.missing,
1974 1974 outgoing.missingheads,
1975 1975 source)
1976 1976
1977 1977 def getbundle(self, source, heads=None, common=None):
1978 1978 """Like changegroupsubset, but returns the set difference between the
1979 1979 ancestors of heads and the ancestors common.
1980 1980
1981 1981 If heads is None, use the local heads. If common is None, use [nullid].
1982 1982
1983 1983 The nodes in common might not all be known locally due to the way the
1984 1984 current discovery protocol works.
1985 1985 """
1986 1986 cl = self.changelog
1987 1987 if common:
1988 1988 hasnode = cl.hasnode
1989 1989 common = [n for n in common if hasnode(n)]
1990 1990 else:
1991 1991 common = [nullid]
1992 1992 if not heads:
1993 1993 heads = cl.heads()
1994 1994 return self.getlocalbundle(source,
1995 1995 discovery.outgoing(cl, common, heads))
1996 1996
1997 1997 @unfilteredmethod
1998 1998 def _changegroupsubset(self, commonrevs, csets, heads, source):
1999 1999
2000 2000 cl = self.changelog
2001 2001 mf = self.manifest
2002 2002 mfs = {} # needed manifests
2003 2003 fnodes = {} # needed file nodes
2004 2004 changedfiles = set()
2005 2005 fstate = ['', {}]
2006 2006 count = [0, 0]
2007 2007
2008 2008 # can we go through the fast path ?
2009 2009 heads.sort()
2010 2010 if heads == sorted(self.heads()):
2011 2011 return self._changegroup(csets, source)
2012 2012
2013 2013 # slow path
2014 2014 self.hook('preoutgoing', throw=True, source=source)
2015 2015 self.changegroupinfo(csets, source)
2016 2016
2017 2017 # filter any nodes that claim to be part of the known set
2018 2018 def prune(revlog, missing):
2019 2019 rr, rl = revlog.rev, revlog.linkrev
2020 2020 return [n for n in missing
2021 2021 if rl(rr(n)) not in commonrevs]
2022 2022
2023 2023 progress = self.ui.progress
2024 2024 _bundling = _('bundling')
2025 2025 _changesets = _('changesets')
2026 2026 _manifests = _('manifests')
2027 2027 _files = _('files')
2028 2028
2029 2029 def lookup(revlog, x):
2030 2030 if revlog == cl:
2031 2031 c = cl.read(x)
2032 2032 changedfiles.update(c[3])
2033 2033 mfs.setdefault(c[0], x)
2034 2034 count[0] += 1
2035 2035 progress(_bundling, count[0],
2036 2036 unit=_changesets, total=count[1])
2037 2037 return x
2038 2038 elif revlog == mf:
2039 2039 clnode = mfs[x]
2040 2040 mdata = mf.readfast(x)
2041 2041 for f, n in mdata.iteritems():
2042 2042 if f in changedfiles:
2043 2043 fnodes[f].setdefault(n, clnode)
2044 2044 count[0] += 1
2045 2045 progress(_bundling, count[0],
2046 2046 unit=_manifests, total=count[1])
2047 2047 return clnode
2048 2048 else:
2049 2049 progress(_bundling, count[0], item=fstate[0],
2050 2050 unit=_files, total=count[1])
2051 2051 return fstate[1][x]
2052 2052
2053 2053 bundler = changegroup.bundle10(lookup)
2054 2054 reorder = self.ui.config('bundle', 'reorder', 'auto')
2055 2055 if reorder == 'auto':
2056 2056 reorder = None
2057 2057 else:
2058 2058 reorder = util.parsebool(reorder)
2059 2059
2060 2060 def gengroup():
2061 2061 # Create a changenode group generator that will call our functions
2062 2062 # back to lookup the owning changenode and collect information.
2063 2063 count[:] = [0, len(csets)]
2064 2064 for chunk in cl.group(csets, bundler, reorder=reorder):
2065 2065 yield chunk
2066 2066 progress(_bundling, None)
2067 2067
2068 2068 # Create a generator for the manifestnodes that calls our lookup
2069 2069 # and data collection functions back.
2070 2070 for f in changedfiles:
2071 2071 fnodes[f] = {}
2072 2072 count[:] = [0, len(mfs)]
2073 2073 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2074 2074 yield chunk
2075 2075 progress(_bundling, None)
2076 2076
2077 2077 mfs.clear()
2078 2078
2079 2079 # Go through all our files in order sorted by name.
2080 2080 count[:] = [0, len(changedfiles)]
2081 2081 for fname in sorted(changedfiles):
2082 2082 filerevlog = self.file(fname)
2083 2083 if not len(filerevlog):
2084 2084 raise util.Abort(_("empty or missing revlog for %s")
2085 2085 % fname)
2086 2086 fstate[0] = fname
2087 2087 fstate[1] = fnodes.pop(fname, {})
2088 2088
2089 2089 nodelist = prune(filerevlog, fstate[1])
2090 2090 if nodelist:
2091 2091 count[0] += 1
2092 2092 yield bundler.fileheader(fname)
2093 2093 for chunk in filerevlog.group(nodelist, bundler, reorder):
2094 2094 yield chunk
2095 2095
2096 2096 # Signal that no more groups are left.
2097 2097 yield bundler.close()
2098 2098 progress(_bundling, None)
2099 2099
2100 2100 if csets:
2101 2101 self.hook('outgoing', node=hex(csets[0]), source=source)
2102 2102
2103 2103 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2104 2104
2105 2105 def changegroup(self, basenodes, source):
2106 2106 # to avoid a race we use changegroupsubset() (issue1320)
2107 2107 return self.changegroupsubset(basenodes, self.heads(), source)
2108 2108
2109 2109 @unfilteredmethod
2110 2110 def _changegroup(self, nodes, source):
2111 2111 """Compute the changegroup of all nodes that we have that a recipient
2112 2112 doesn't. Return a chunkbuffer object whose read() method will return
2113 2113 successive changegroup chunks.
2114 2114
2115 2115 This is much easier than the previous function as we can assume that
2116 2116 the recipient has any changenode we aren't sending them.
2117 2117
2118 2118 nodes is the set of nodes to send"""
2119 2119
2120 2120 cl = self.changelog
2121 2121 mf = self.manifest
2122 2122 mfs = {}
2123 2123 changedfiles = set()
2124 2124 fstate = ['']
2125 2125 count = [0, 0]
2126 2126
2127 2127 self.hook('preoutgoing', throw=True, source=source)
2128 2128 self.changegroupinfo(nodes, source)
2129 2129
2130 2130 revset = set([cl.rev(n) for n in nodes])
2131 2131
2132 2132 def gennodelst(log):
2133 2133 ln, llr = log.node, log.linkrev
2134 2134 return [ln(r) for r in log if llr(r) in revset]
2135 2135
2136 2136 progress = self.ui.progress
2137 2137 _bundling = _('bundling')
2138 2138 _changesets = _('changesets')
2139 2139 _manifests = _('manifests')
2140 2140 _files = _('files')
2141 2141
2142 2142 def lookup(revlog, x):
2143 2143 if revlog == cl:
2144 2144 c = cl.read(x)
2145 2145 changedfiles.update(c[3])
2146 2146 mfs.setdefault(c[0], x)
2147 2147 count[0] += 1
2148 2148 progress(_bundling, count[0],
2149 2149 unit=_changesets, total=count[1])
2150 2150 return x
2151 2151 elif revlog == mf:
2152 2152 count[0] += 1
2153 2153 progress(_bundling, count[0],
2154 2154 unit=_manifests, total=count[1])
2155 2155 return cl.node(revlog.linkrev(revlog.rev(x)))
2156 2156 else:
2157 2157 progress(_bundling, count[0], item=fstate[0],
2158 2158 total=count[1], unit=_files)
2159 2159 return cl.node(revlog.linkrev(revlog.rev(x)))
2160 2160
2161 2161 bundler = changegroup.bundle10(lookup)
2162 2162 reorder = self.ui.config('bundle', 'reorder', 'auto')
2163 2163 if reorder == 'auto':
2164 2164 reorder = None
2165 2165 else:
2166 2166 reorder = util.parsebool(reorder)
2167 2167
2168 2168 def gengroup():
2169 2169 '''yield a sequence of changegroup chunks (strings)'''
2170 2170 # construct a list of all changed files
2171 2171
2172 2172 count[:] = [0, len(nodes)]
2173 2173 for chunk in cl.group(nodes, bundler, reorder=reorder):
2174 2174 yield chunk
2175 2175 progress(_bundling, None)
2176 2176
2177 2177 count[:] = [0, len(mfs)]
2178 2178 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2179 2179 yield chunk
2180 2180 progress(_bundling, None)
2181 2181
2182 2182 count[:] = [0, len(changedfiles)]
2183 2183 for fname in sorted(changedfiles):
2184 2184 filerevlog = self.file(fname)
2185 2185 if not len(filerevlog):
2186 2186 raise util.Abort(_("empty or missing revlog for %s")
2187 2187 % fname)
2188 2188 fstate[0] = fname
2189 2189 nodelist = gennodelst(filerevlog)
2190 2190 if nodelist:
2191 2191 count[0] += 1
2192 2192 yield bundler.fileheader(fname)
2193 2193 for chunk in filerevlog.group(nodelist, bundler, reorder):
2194 2194 yield chunk
2195 2195 yield bundler.close()
2196 2196 progress(_bundling, None)
2197 2197
2198 2198 if nodes:
2199 2199 self.hook('outgoing', node=hex(nodes[0]), source=source)
2200 2200
2201 2201 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2202 2202
2203 2203 @unfilteredmethod
2204 2204 def addchangegroup(self, source, srctype, url, emptyok=False):
2205 2205 """Add the changegroup returned by source.read() to this repo.
2206 2206 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2207 2207 the URL of the repo where this changegroup is coming from.
2208 2208
2209 2209 Return an integer summarizing the change to this repo:
2210 2210 - nothing changed or no source: 0
2211 2211 - more heads than before: 1+added heads (2..n)
2212 2212 - fewer heads than before: -1-removed heads (-2..-n)
2213 2213 - number of heads stays the same: 1
2214 2214 """
2215 2215 def csmap(x):
2216 2216 self.ui.debug("add changeset %s\n" % short(x))
2217 2217 return len(cl)
2218 2218
2219 2219 def revmap(x):
2220 2220 return cl.rev(x)
2221 2221
2222 2222 if not source:
2223 2223 return 0
2224 2224
2225 2225 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2226 2226
2227 2227 changesets = files = revisions = 0
2228 2228 efiles = set()
2229 2229
2230 2230 # write changelog data to temp files so concurrent readers will not see
2231 2231 # inconsistent view
2232 2232 cl = self.changelog
2233 2233 cl.delayupdate()
2234 2234 oldheads = cl.heads()
2235 2235
2236 2236 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2237 2237 try:
2238 2238 trp = weakref.proxy(tr)
2239 2239 # pull off the changeset group
2240 2240 self.ui.status(_("adding changesets\n"))
2241 2241 clstart = len(cl)
2242 2242 class prog(object):
2243 2243 step = _('changesets')
2244 2244 count = 1
2245 2245 ui = self.ui
2246 2246 total = None
2247 2247 def __call__(self):
2248 2248 self.ui.progress(self.step, self.count, unit=_('chunks'),
2249 2249 total=self.total)
2250 2250 self.count += 1
2251 2251 pr = prog()
2252 2252 source.callback = pr
2253 2253
2254 2254 source.changelogheader()
2255 2255 srccontent = cl.addgroup(source, csmap, trp)
2256 2256 if not (srccontent or emptyok):
2257 2257 raise util.Abort(_("received changelog group is empty"))
2258 2258 clend = len(cl)
2259 2259 changesets = clend - clstart
2260 2260 for c in xrange(clstart, clend):
2261 2261 efiles.update(self[c].files())
2262 2262 efiles = len(efiles)
2263 2263 self.ui.progress(_('changesets'), None)
2264 2264
2265 2265 # pull off the manifest group
2266 2266 self.ui.status(_("adding manifests\n"))
2267 2267 pr.step = _('manifests')
2268 2268 pr.count = 1
2269 2269 pr.total = changesets # manifests <= changesets
2270 2270 # no need to check for empty manifest group here:
2271 2271 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2272 2272 # no new manifest will be created and the manifest group will
2273 2273 # be empty during the pull
2274 2274 source.manifestheader()
2275 2275 self.manifest.addgroup(source, revmap, trp)
2276 2276 self.ui.progress(_('manifests'), None)
2277 2277
2278 2278 needfiles = {}
2279 2279 if self.ui.configbool('server', 'validate', default=False):
2280 2280 # validate incoming csets have their manifests
2281 2281 for cset in xrange(clstart, clend):
2282 2282 mfest = self.changelog.read(self.changelog.node(cset))[0]
2283 2283 mfest = self.manifest.readdelta(mfest)
2284 2284 # store file nodes we must see
2285 2285 for f, n in mfest.iteritems():
2286 2286 needfiles.setdefault(f, set()).add(n)
2287 2287
2288 2288 # process the files
2289 2289 self.ui.status(_("adding file changes\n"))
2290 2290 pr.step = _('files')
2291 2291 pr.count = 1
2292 2292 pr.total = efiles
2293 2293 source.callback = None
2294 2294
2295 2295 while True:
2296 2296 chunkdata = source.filelogheader()
2297 2297 if not chunkdata:
2298 2298 break
2299 2299 f = chunkdata["filename"]
2300 2300 self.ui.debug("adding %s revisions\n" % f)
2301 2301 pr()
2302 2302 fl = self.file(f)
2303 2303 o = len(fl)
2304 2304 if not fl.addgroup(source, revmap, trp):
2305 2305 raise util.Abort(_("received file revlog group is empty"))
2306 2306 revisions += len(fl) - o
2307 2307 files += 1
2308 2308 if f in needfiles:
2309 2309 needs = needfiles[f]
2310 2310 for new in xrange(o, len(fl)):
2311 2311 n = fl.node(new)
2312 2312 if n in needs:
2313 2313 needs.remove(n)
2314 else:
2315 raise util.Abort(
2316 _("received spurious file revlog entry"))
2314 2317 if not needs:
2315 2318 del needfiles[f]
2316 2319 self.ui.progress(_('files'), None)
2317 2320
2318 2321 for f, needs in needfiles.iteritems():
2319 2322 fl = self.file(f)
2320 2323 for n in needs:
2321 2324 try:
2322 2325 fl.rev(n)
2323 2326 except error.LookupError:
2324 2327 raise util.Abort(
2325 2328 _('missing file data for %s:%s - run hg verify') %
2326 2329 (f, hex(n)))
2327 2330
2328 2331 dh = 0
2329 2332 if oldheads:
2330 2333 heads = cl.heads()
2331 2334 dh = len(heads) - len(oldheads)
2332 2335 for h in heads:
2333 2336 if h not in oldheads and self[h].closesbranch():
2334 2337 dh -= 1
2335 2338 htext = ""
2336 2339 if dh:
2337 2340 htext = _(" (%+d heads)") % dh
2338 2341
2339 2342 self.ui.status(_("added %d changesets"
2340 2343 " with %d changes to %d files%s\n")
2341 2344 % (changesets, revisions, files, htext))
2342 2345 self.invalidatevolatilesets()
2343 2346
2344 2347 if changesets > 0:
2345 2348 p = lambda: cl.writepending() and self.root or ""
2346 2349 self.hook('pretxnchangegroup', throw=True,
2347 2350 node=hex(cl.node(clstart)), source=srctype,
2348 2351 url=url, pending=p)
2349 2352
2350 2353 added = [cl.node(r) for r in xrange(clstart, clend)]
2351 2354 publishing = self.ui.configbool('phases', 'publish', True)
2352 2355 if srctype == 'push':
2353 2356 # Old server can not push the boundary themself.
2354 2357 # New server won't push the boundary if changeset already
2355 2358 # existed locally as secrete
2356 2359 #
2357 2360 # We should not use added here but the list of all change in
2358 2361 # the bundle
2359 2362 if publishing:
2360 2363 phases.advanceboundary(self, phases.public, srccontent)
2361 2364 else:
2362 2365 phases.advanceboundary(self, phases.draft, srccontent)
2363 2366 phases.retractboundary(self, phases.draft, added)
2364 2367 elif srctype != 'strip':
2365 2368 # publishing only alter behavior during push
2366 2369 #
2367 2370 # strip should not touch boundary at all
2368 2371 phases.retractboundary(self, phases.draft, added)
2369 2372
2370 2373 # make changelog see real files again
2371 2374 cl.finalize(trp)
2372 2375
2373 2376 tr.close()
2374 2377
2375 2378 if changesets > 0:
2376 2379 if srctype != 'strip':
2377 2380 # During strip, branchcache is invalid but coming call to
2378 2381 # `destroyed` will repair it.
2379 2382 # In other case we can safely update cache on disk.
2380 2383 branchmap.updatecache(self.filtered('served'))
2381 2384 def runhooks():
2382 2385 # forcefully update the on-disk branch cache
2383 2386 self.ui.debug("updating the branch cache\n")
2384 2387 self.hook("changegroup", node=hex(cl.node(clstart)),
2385 2388 source=srctype, url=url)
2386 2389
2387 2390 for n in added:
2388 2391 self.hook("incoming", node=hex(n), source=srctype,
2389 2392 url=url)
2390 2393 self._afterlock(runhooks)
2391 2394
2392 2395 finally:
2393 2396 tr.release()
2394 2397 # never return 0 here:
2395 2398 if dh < 0:
2396 2399 return dh - 1
2397 2400 else:
2398 2401 return dh + 1
2399 2402
2400 2403 def stream_in(self, remote, requirements):
2401 2404 lock = self.lock()
2402 2405 try:
2403 2406 # Save remote branchmap. We will use it later
2404 2407 # to speed up branchcache creation
2405 2408 rbranchmap = None
2406 2409 if remote.capable("branchmap"):
2407 2410 rbranchmap = remote.branchmap()
2408 2411
2409 2412 fp = remote.stream_out()
2410 2413 l = fp.readline()
2411 2414 try:
2412 2415 resp = int(l)
2413 2416 except ValueError:
2414 2417 raise error.ResponseError(
2415 2418 _('unexpected response from remote server:'), l)
2416 2419 if resp == 1:
2417 2420 raise util.Abort(_('operation forbidden by server'))
2418 2421 elif resp == 2:
2419 2422 raise util.Abort(_('locking the remote repository failed'))
2420 2423 elif resp != 0:
2421 2424 raise util.Abort(_('the server sent an unknown error code'))
2422 2425 self.ui.status(_('streaming all changes\n'))
2423 2426 l = fp.readline()
2424 2427 try:
2425 2428 total_files, total_bytes = map(int, l.split(' ', 1))
2426 2429 except (ValueError, TypeError):
2427 2430 raise error.ResponseError(
2428 2431 _('unexpected response from remote server:'), l)
2429 2432 self.ui.status(_('%d files to transfer, %s of data\n') %
2430 2433 (total_files, util.bytecount(total_bytes)))
2431 2434 handled_bytes = 0
2432 2435 self.ui.progress(_('clone'), 0, total=total_bytes)
2433 2436 start = time.time()
2434 2437 for i in xrange(total_files):
2435 2438 # XXX doesn't support '\n' or '\r' in filenames
2436 2439 l = fp.readline()
2437 2440 try:
2438 2441 name, size = l.split('\0', 1)
2439 2442 size = int(size)
2440 2443 except (ValueError, TypeError):
2441 2444 raise error.ResponseError(
2442 2445 _('unexpected response from remote server:'), l)
2443 2446 if self.ui.debugflag:
2444 2447 self.ui.debug('adding %s (%s)\n' %
2445 2448 (name, util.bytecount(size)))
2446 2449 # for backwards compat, name was partially encoded
2447 2450 ofp = self.sopener(store.decodedir(name), 'w')
2448 2451 for chunk in util.filechunkiter(fp, limit=size):
2449 2452 handled_bytes += len(chunk)
2450 2453 self.ui.progress(_('clone'), handled_bytes,
2451 2454 total=total_bytes)
2452 2455 ofp.write(chunk)
2453 2456 ofp.close()
2454 2457 elapsed = time.time() - start
2455 2458 if elapsed <= 0:
2456 2459 elapsed = 0.001
2457 2460 self.ui.progress(_('clone'), None)
2458 2461 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2459 2462 (util.bytecount(total_bytes), elapsed,
2460 2463 util.bytecount(total_bytes / elapsed)))
2461 2464
2462 2465 # new requirements = old non-format requirements +
2463 2466 # new format-related
2464 2467 # requirements from the streamed-in repository
2465 2468 requirements.update(set(self.requirements) - self.supportedformats)
2466 2469 self._applyrequirements(requirements)
2467 2470 self._writerequirements()
2468 2471
2469 2472 if rbranchmap:
2470 2473 rbheads = []
2471 2474 for bheads in rbranchmap.itervalues():
2472 2475 rbheads.extend(bheads)
2473 2476
2474 2477 if rbheads:
2475 2478 rtiprev = max((int(self.changelog.rev(node))
2476 2479 for node in rbheads))
2477 2480 cache = branchmap.branchcache(rbranchmap,
2478 2481 self[rtiprev].node(),
2479 2482 rtiprev)
2480 2483 # Try to stick it as low as possible
2481 2484 # filter above served are unlikely to be fetch from a clone
2482 2485 for candidate in ('base', 'immutable', 'served'):
2483 2486 rview = self.filtered(candidate)
2484 2487 if cache.validfor(rview):
2485 2488 self._branchcaches[candidate] = cache
2486 2489 cache.write(rview)
2487 2490 break
2488 2491 self.invalidate()
2489 2492 return len(self.heads()) + 1
2490 2493 finally:
2491 2494 lock.release()
2492 2495
2493 2496 def clone(self, remote, heads=[], stream=False):
2494 2497 '''clone remote repository.
2495 2498
2496 2499 keyword arguments:
2497 2500 heads: list of revs to clone (forces use of pull)
2498 2501 stream: use streaming clone if possible'''
2499 2502
2500 2503 # now, all clients that can request uncompressed clones can
2501 2504 # read repo formats supported by all servers that can serve
2502 2505 # them.
2503 2506
2504 2507 # if revlog format changes, client will have to check version
2505 2508 # and format flags on "stream" capability, and use
2506 2509 # uncompressed only if compatible.
2507 2510
2508 2511 if not stream:
2509 2512 # if the server explicitly prefers to stream (for fast LANs)
2510 2513 stream = remote.capable('stream-preferred')
2511 2514
2512 2515 if stream and not heads:
2513 2516 # 'stream' means remote revlog format is revlogv1 only
2514 2517 if remote.capable('stream'):
2515 2518 return self.stream_in(remote, set(('revlogv1',)))
2516 2519 # otherwise, 'streamreqs' contains the remote revlog format
2517 2520 streamreqs = remote.capable('streamreqs')
2518 2521 if streamreqs:
2519 2522 streamreqs = set(streamreqs.split(','))
2520 2523 # if we support it, stream in and adjust our requirements
2521 2524 if not streamreqs - self.supportedformats:
2522 2525 return self.stream_in(remote, streamreqs)
2523 2526 return self.pull(remote, heads)
2524 2527
2525 2528 def pushkey(self, namespace, key, old, new):
2526 2529 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2527 2530 old=old, new=new)
2528 2531 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2529 2532 ret = pushkey.push(self, namespace, key, old, new)
2530 2533 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2531 2534 ret=ret)
2532 2535 return ret
2533 2536
2534 2537 def listkeys(self, namespace):
2535 2538 self.hook('prelistkeys', throw=True, namespace=namespace)
2536 2539 self.ui.debug('listing keys for "%s"\n' % namespace)
2537 2540 values = pushkey.list(self, namespace)
2538 2541 self.hook('listkeys', namespace=namespace, values=values)
2539 2542 return values
2540 2543
2541 2544 def debugwireargs(self, one, two, three=None, four=None, five=None):
2542 2545 '''used to test argument passing over the wire'''
2543 2546 return "%s %s %s %s %s" % (one, two, three, four, five)
2544 2547
2545 2548 def savecommitmessage(self, text):
2546 2549 fp = self.opener('last-message.txt', 'wb')
2547 2550 try:
2548 2551 fp.write(text)
2549 2552 finally:
2550 2553 fp.close()
2551 2554 return self.pathto(fp.name[len(self.root) + 1:])
2552 2555
2553 2556 # used to avoid circular references so destructors work
2554 2557 def aftertrans(files):
2555 2558 renamefiles = [tuple(t) for t in files]
2556 2559 def a():
2557 2560 for src, dest in renamefiles:
2558 2561 try:
2559 2562 util.rename(src, dest)
2560 2563 except OSError: # journal file does not yet exist
2561 2564 pass
2562 2565 return a
2563 2566
2564 2567 def undoname(fn):
2565 2568 base, name = os.path.split(fn)
2566 2569 assert name.startswith('journal')
2567 2570 return os.path.join(base, name.replace('journal', 'undo', 1))
2568 2571
2569 2572 def instance(ui, path, create):
2570 2573 return localrepository(ui, util.urllocalpath(path), create)
2571 2574
2572 2575 def islocal(path):
2573 2576 return True
@@ -1,53 +1,92 b''
1 1 $ hg init test
2 2 $ cd test
3 3
4 4 $ cat > .hg/hgrc <<EOF
5 5 > [server]
6 6 > validate=1
7 7 > EOF
8 8
9 9 $ echo alpha > alpha
10 10 $ echo beta > beta
11 11 $ hg addr
12 12 adding alpha
13 13 adding beta
14 14 $ hg ci -m 1
15 15
16 16 $ cd ..
17 17 $ hg clone test test-clone
18 18 updating to branch default
19 19 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
20 20
21 Test spurious filelog entries:
22
21 23 $ cd test-clone
24 $ echo blah >> beta
25 $ cp .hg/store/data/beta.i tmp1
26 $ hg ci -m 2
27 $ cp .hg/store/data/beta.i tmp2
28 $ hg -q rollback
29 $ mv tmp2 .hg/store/data/beta.i
30 $ echo blah >> beta
31 $ hg ci -m '2 (corrupt)'
32
33 Expected to fail:
34
35 $ hg verify
36 checking changesets
37 checking manifests
38 crosschecking files in changesets and manifests
39 checking files
40 beta@1: dddc47b3ba30 not in manifests
41 2 files, 2 changesets, 4 total revisions
42 1 integrity errors encountered!
43 (first damaged changeset appears to be 1)
44 [1]
45
46 $ hg push
47 pushing to $TESTTMP/test
48 searching for changes
49 adding changesets
50 adding manifests
51 adding file changes
52 transaction abort!
53 rollback completed
54 abort: received spurious file revlog entry
55 [255]
56
57 $ hg -q rollback
58 $ mv tmp1 .hg/store/data/beta.i
59 $ echo beta > beta
60
61 Test missing filelog entries:
62
22 63 $ cp .hg/store/data/beta.i tmp
23 64 $ echo blah >> beta
24 65 $ hg ci -m '2 (corrupt)'
25 66 $ mv tmp .hg/store/data/beta.i
26 67
27 68 Expected to fail:
28 69
29 70 $ hg verify
30 71 checking changesets
31 72 checking manifests
32 73 crosschecking files in changesets and manifests
33 74 checking files
34 75 beta@1: dddc47b3ba30 in manifests not found
35 76 2 files, 2 changesets, 2 total revisions
36 77 1 integrity errors encountered!
37 78 (first damaged changeset appears to be 1)
38 79 [1]
39 80
40 Expected to fail:
41
42 81 $ hg push
43 82 pushing to $TESTTMP/test (glob)
44 83 searching for changes
45 84 adding changesets
46 85 adding manifests
47 86 adding file changes
48 87 transaction abort!
49 88 rollback completed
50 89 abort: missing file data for beta:dddc47b3ba30e54484720ce0f4f768a0f4b6efb9 - run hg verify
51 90 [255]
52 91
53 92 $ cd ..
General Comments 0
You need to be logged in to leave comments. Login now