##// END OF EJS Templates
localrepo: only use 'bookmarksinstore' requirement if we have 'store'...
Pulkit Goyal -
r45857:dc457177 default
parent child Browse files
Show More
@@ -1,3473 +1,3496 b''
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import errno
11 11 import os
12 12 import random
13 13 import sys
14 14 import time
15 15 import weakref
16 16
17 17 from .i18n import _
18 18 from .node import (
19 19 bin,
20 20 hex,
21 21 nullid,
22 22 nullrev,
23 23 short,
24 24 )
25 25 from .pycompat import (
26 26 delattr,
27 27 getattr,
28 28 )
29 29 from . import (
30 30 bookmarks,
31 31 branchmap,
32 32 bundle2,
33 33 changegroup,
34 34 color,
35 35 commit,
36 36 context,
37 37 dirstate,
38 38 dirstateguard,
39 39 discovery,
40 40 encoding,
41 41 error,
42 42 exchange,
43 43 extensions,
44 44 filelog,
45 45 hook,
46 46 lock as lockmod,
47 47 match as matchmod,
48 48 mergestate as mergestatemod,
49 49 mergeutil,
50 50 namespaces,
51 51 narrowspec,
52 52 obsolete,
53 53 pathutil,
54 54 phases,
55 55 pushkey,
56 56 pycompat,
57 57 rcutil,
58 58 repoview,
59 59 revset,
60 60 revsetlang,
61 61 scmutil,
62 62 sparse,
63 63 store as storemod,
64 64 subrepoutil,
65 65 tags as tagsmod,
66 66 transaction,
67 67 txnutil,
68 68 util,
69 69 vfs as vfsmod,
70 70 )
71 71
72 72 from .interfaces import (
73 73 repository,
74 74 util as interfaceutil,
75 75 )
76 76
77 77 from .utils import (
78 78 hashutil,
79 79 procutil,
80 80 stringutil,
81 81 )
82 82
83 83 from .revlogutils import constants as revlogconst
84 84
85 85 release = lockmod.release
86 86 urlerr = util.urlerr
87 87 urlreq = util.urlreq
88 88
89 89 # set of (path, vfs-location) tuples. vfs-location is:
90 90 # - 'plain for vfs relative paths
91 91 # - '' for svfs relative paths
92 92 _cachedfiles = set()
93 93
94 94
95 95 class _basefilecache(scmutil.filecache):
96 96 """All filecache usage on repo are done for logic that should be unfiltered
97 97 """
98 98
99 99 def __get__(self, repo, type=None):
100 100 if repo is None:
101 101 return self
102 102 # proxy to unfiltered __dict__ since filtered repo has no entry
103 103 unfi = repo.unfiltered()
104 104 try:
105 105 return unfi.__dict__[self.sname]
106 106 except KeyError:
107 107 pass
108 108 return super(_basefilecache, self).__get__(unfi, type)
109 109
110 110 def set(self, repo, value):
111 111 return super(_basefilecache, self).set(repo.unfiltered(), value)
112 112
113 113
114 114 class repofilecache(_basefilecache):
115 115 """filecache for files in .hg but outside of .hg/store"""
116 116
117 117 def __init__(self, *paths):
118 118 super(repofilecache, self).__init__(*paths)
119 119 for path in paths:
120 120 _cachedfiles.add((path, b'plain'))
121 121
122 122 def join(self, obj, fname):
123 123 return obj.vfs.join(fname)
124 124
125 125
126 126 class storecache(_basefilecache):
127 127 """filecache for files in the store"""
128 128
129 129 def __init__(self, *paths):
130 130 super(storecache, self).__init__(*paths)
131 131 for path in paths:
132 132 _cachedfiles.add((path, b''))
133 133
134 134 def join(self, obj, fname):
135 135 return obj.sjoin(fname)
136 136
137 137
138 138 class mixedrepostorecache(_basefilecache):
139 139 """filecache for a mix files in .hg/store and outside"""
140 140
141 141 def __init__(self, *pathsandlocations):
142 142 # scmutil.filecache only uses the path for passing back into our
143 143 # join(), so we can safely pass a list of paths and locations
144 144 super(mixedrepostorecache, self).__init__(*pathsandlocations)
145 145 _cachedfiles.update(pathsandlocations)
146 146
147 147 def join(self, obj, fnameandlocation):
148 148 fname, location = fnameandlocation
149 149 if location == b'plain':
150 150 return obj.vfs.join(fname)
151 151 else:
152 152 if location != b'':
153 153 raise error.ProgrammingError(
154 154 b'unexpected location: %s' % location
155 155 )
156 156 return obj.sjoin(fname)
157 157
158 158
159 159 def isfilecached(repo, name):
160 160 """check if a repo has already cached "name" filecache-ed property
161 161
162 162 This returns (cachedobj-or-None, iscached) tuple.
163 163 """
164 164 cacheentry = repo.unfiltered()._filecache.get(name, None)
165 165 if not cacheentry:
166 166 return None, False
167 167 return cacheentry.obj, True
168 168
169 169
170 170 class unfilteredpropertycache(util.propertycache):
171 171 """propertycache that apply to unfiltered repo only"""
172 172
173 173 def __get__(self, repo, type=None):
174 174 unfi = repo.unfiltered()
175 175 if unfi is repo:
176 176 return super(unfilteredpropertycache, self).__get__(unfi)
177 177 return getattr(unfi, self.name)
178 178
179 179
180 180 class filteredpropertycache(util.propertycache):
181 181 """propertycache that must take filtering in account"""
182 182
183 183 def cachevalue(self, obj, value):
184 184 object.__setattr__(obj, self.name, value)
185 185
186 186
187 187 def hasunfilteredcache(repo, name):
188 188 """check if a repo has an unfilteredpropertycache value for <name>"""
189 189 return name in vars(repo.unfiltered())
190 190
191 191
192 192 def unfilteredmethod(orig):
193 193 """decorate method that always need to be run on unfiltered version"""
194 194
195 195 def wrapper(repo, *args, **kwargs):
196 196 return orig(repo.unfiltered(), *args, **kwargs)
197 197
198 198 return wrapper
199 199
200 200
201 201 moderncaps = {
202 202 b'lookup',
203 203 b'branchmap',
204 204 b'pushkey',
205 205 b'known',
206 206 b'getbundle',
207 207 b'unbundle',
208 208 }
209 209 legacycaps = moderncaps.union({b'changegroupsubset'})
210 210
211 211
212 212 @interfaceutil.implementer(repository.ipeercommandexecutor)
213 213 class localcommandexecutor(object):
214 214 def __init__(self, peer):
215 215 self._peer = peer
216 216 self._sent = False
217 217 self._closed = False
218 218
219 219 def __enter__(self):
220 220 return self
221 221
222 222 def __exit__(self, exctype, excvalue, exctb):
223 223 self.close()
224 224
225 225 def callcommand(self, command, args):
226 226 if self._sent:
227 227 raise error.ProgrammingError(
228 228 b'callcommand() cannot be used after sendcommands()'
229 229 )
230 230
231 231 if self._closed:
232 232 raise error.ProgrammingError(
233 233 b'callcommand() cannot be used after close()'
234 234 )
235 235
236 236 # We don't need to support anything fancy. Just call the named
237 237 # method on the peer and return a resolved future.
238 238 fn = getattr(self._peer, pycompat.sysstr(command))
239 239
240 240 f = pycompat.futures.Future()
241 241
242 242 try:
243 243 result = fn(**pycompat.strkwargs(args))
244 244 except Exception:
245 245 pycompat.future_set_exception_info(f, sys.exc_info()[1:])
246 246 else:
247 247 f.set_result(result)
248 248
249 249 return f
250 250
251 251 def sendcommands(self):
252 252 self._sent = True
253 253
254 254 def close(self):
255 255 self._closed = True
256 256
257 257
258 258 @interfaceutil.implementer(repository.ipeercommands)
259 259 class localpeer(repository.peer):
260 260 '''peer for a local repo; reflects only the most recent API'''
261 261
262 262 def __init__(self, repo, caps=None):
263 263 super(localpeer, self).__init__()
264 264
265 265 if caps is None:
266 266 caps = moderncaps.copy()
267 267 self._repo = repo.filtered(b'served')
268 268 self.ui = repo.ui
269 269 self._caps = repo._restrictcapabilities(caps)
270 270
271 271 # Begin of _basepeer interface.
272 272
273 273 def url(self):
274 274 return self._repo.url()
275 275
276 276 def local(self):
277 277 return self._repo
278 278
279 279 def peer(self):
280 280 return self
281 281
282 282 def canpush(self):
283 283 return True
284 284
285 285 def close(self):
286 286 self._repo.close()
287 287
288 288 # End of _basepeer interface.
289 289
290 290 # Begin of _basewirecommands interface.
291 291
292 292 def branchmap(self):
293 293 return self._repo.branchmap()
294 294
295 295 def capabilities(self):
296 296 return self._caps
297 297
298 298 def clonebundles(self):
299 299 return self._repo.tryread(b'clonebundles.manifest')
300 300
301 301 def debugwireargs(self, one, two, three=None, four=None, five=None):
302 302 """Used to test argument passing over the wire"""
303 303 return b"%s %s %s %s %s" % (
304 304 one,
305 305 two,
306 306 pycompat.bytestr(three),
307 307 pycompat.bytestr(four),
308 308 pycompat.bytestr(five),
309 309 )
310 310
311 311 def getbundle(
312 312 self, source, heads=None, common=None, bundlecaps=None, **kwargs
313 313 ):
314 314 chunks = exchange.getbundlechunks(
315 315 self._repo,
316 316 source,
317 317 heads=heads,
318 318 common=common,
319 319 bundlecaps=bundlecaps,
320 320 **kwargs
321 321 )[1]
322 322 cb = util.chunkbuffer(chunks)
323 323
324 324 if exchange.bundle2requested(bundlecaps):
325 325 # When requesting a bundle2, getbundle returns a stream to make the
326 326 # wire level function happier. We need to build a proper object
327 327 # from it in local peer.
328 328 return bundle2.getunbundler(self.ui, cb)
329 329 else:
330 330 return changegroup.getunbundler(b'01', cb, None)
331 331
332 332 def heads(self):
333 333 return self._repo.heads()
334 334
335 335 def known(self, nodes):
336 336 return self._repo.known(nodes)
337 337
338 338 def listkeys(self, namespace):
339 339 return self._repo.listkeys(namespace)
340 340
341 341 def lookup(self, key):
342 342 return self._repo.lookup(key)
343 343
344 344 def pushkey(self, namespace, key, old, new):
345 345 return self._repo.pushkey(namespace, key, old, new)
346 346
347 347 def stream_out(self):
348 348 raise error.Abort(_(b'cannot perform stream clone against local peer'))
349 349
350 350 def unbundle(self, bundle, heads, url):
351 351 """apply a bundle on a repo
352 352
353 353 This function handles the repo locking itself."""
354 354 try:
355 355 try:
356 356 bundle = exchange.readbundle(self.ui, bundle, None)
357 357 ret = exchange.unbundle(self._repo, bundle, heads, b'push', url)
358 358 if util.safehasattr(ret, b'getchunks'):
359 359 # This is a bundle20 object, turn it into an unbundler.
360 360 # This little dance should be dropped eventually when the
361 361 # API is finally improved.
362 362 stream = util.chunkbuffer(ret.getchunks())
363 363 ret = bundle2.getunbundler(self.ui, stream)
364 364 return ret
365 365 except Exception as exc:
366 366 # If the exception contains output salvaged from a bundle2
367 367 # reply, we need to make sure it is printed before continuing
368 368 # to fail. So we build a bundle2 with such output and consume
369 369 # it directly.
370 370 #
371 371 # This is not very elegant but allows a "simple" solution for
372 372 # issue4594
373 373 output = getattr(exc, '_bundle2salvagedoutput', ())
374 374 if output:
375 375 bundler = bundle2.bundle20(self._repo.ui)
376 376 for out in output:
377 377 bundler.addpart(out)
378 378 stream = util.chunkbuffer(bundler.getchunks())
379 379 b = bundle2.getunbundler(self.ui, stream)
380 380 bundle2.processbundle(self._repo, b)
381 381 raise
382 382 except error.PushRaced as exc:
383 383 raise error.ResponseError(
384 384 _(b'push failed:'), stringutil.forcebytestr(exc)
385 385 )
386 386
387 387 # End of _basewirecommands interface.
388 388
389 389 # Begin of peer interface.
390 390
391 391 def commandexecutor(self):
392 392 return localcommandexecutor(self)
393 393
394 394 # End of peer interface.
395 395
396 396
397 397 @interfaceutil.implementer(repository.ipeerlegacycommands)
398 398 class locallegacypeer(localpeer):
399 399 '''peer extension which implements legacy methods too; used for tests with
400 400 restricted capabilities'''
401 401
402 402 def __init__(self, repo):
403 403 super(locallegacypeer, self).__init__(repo, caps=legacycaps)
404 404
405 405 # Begin of baselegacywirecommands interface.
406 406
407 407 def between(self, pairs):
408 408 return self._repo.between(pairs)
409 409
410 410 def branches(self, nodes):
411 411 return self._repo.branches(nodes)
412 412
413 413 def changegroup(self, nodes, source):
414 414 outgoing = discovery.outgoing(
415 415 self._repo, missingroots=nodes, ancestorsof=self._repo.heads()
416 416 )
417 417 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
418 418
419 419 def changegroupsubset(self, bases, heads, source):
420 420 outgoing = discovery.outgoing(
421 421 self._repo, missingroots=bases, ancestorsof=heads
422 422 )
423 423 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
424 424
425 425 # End of baselegacywirecommands interface.
426 426
427 427
428 428 # Increment the sub-version when the revlog v2 format changes to lock out old
429 429 # clients.
430 430 REVLOGV2_REQUIREMENT = b'exp-revlogv2.1'
431 431
432 432 # A repository with the sparserevlog feature will have delta chains that
433 433 # can spread over a larger span. Sparse reading cuts these large spans into
434 434 # pieces, so that each piece isn't too big.
435 435 # Without the sparserevlog capability, reading from the repository could use
436 436 # huge amounts of memory, because the whole span would be read at once,
437 437 # including all the intermediate revisions that aren't pertinent for the chain.
438 438 # This is why once a repository has enabled sparse-read, it becomes required.
439 439 SPARSEREVLOG_REQUIREMENT = b'sparserevlog'
440 440
441 441 # A repository with the sidedataflag requirement will allow to store extra
442 442 # information for revision without altering their original hashes.
443 443 SIDEDATA_REQUIREMENT = b'exp-sidedata-flag'
444 444
445 445 # A repository with the the copies-sidedata-changeset requirement will store
446 446 # copies related information in changeset's sidedata.
447 447 COPIESSDC_REQUIREMENT = b'exp-copies-sidedata-changeset'
448 448
449 449 # The repository use persistent nodemap for the changelog and the manifest.
450 450 NODEMAP_REQUIREMENT = b'persistent-nodemap'
451 451
452 452 # Functions receiving (ui, features) that extensions can register to impact
453 453 # the ability to load repositories with custom requirements. Only
454 454 # functions defined in loaded extensions are called.
455 455 #
456 456 # The function receives a set of requirement strings that the repository
457 457 # is capable of opening. Functions will typically add elements to the
458 458 # set to reflect that the extension knows how to handle that requirements.
459 459 featuresetupfuncs = set()
460 460
461 461
462 462 def makelocalrepository(baseui, path, intents=None):
463 463 """Create a local repository object.
464 464
465 465 Given arguments needed to construct a local repository, this function
466 466 performs various early repository loading functionality (such as
467 467 reading the ``.hg/requires`` and ``.hg/hgrc`` files), validates that
468 468 the repository can be opened, derives a type suitable for representing
469 469 that repository, and returns an instance of it.
470 470
471 471 The returned object conforms to the ``repository.completelocalrepository``
472 472 interface.
473 473
474 474 The repository type is derived by calling a series of factory functions
475 475 for each aspect/interface of the final repository. These are defined by
476 476 ``REPO_INTERFACES``.
477 477
478 478 Each factory function is called to produce a type implementing a specific
479 479 interface. The cumulative list of returned types will be combined into a
480 480 new type and that type will be instantiated to represent the local
481 481 repository.
482 482
483 483 The factory functions each receive various state that may be consulted
484 484 as part of deriving a type.
485 485
486 486 Extensions should wrap these factory functions to customize repository type
487 487 creation. Note that an extension's wrapped function may be called even if
488 488 that extension is not loaded for the repo being constructed. Extensions
489 489 should check if their ``__name__`` appears in the
490 490 ``extensionmodulenames`` set passed to the factory function and no-op if
491 491 not.
492 492 """
493 493 ui = baseui.copy()
494 494 # Prevent copying repo configuration.
495 495 ui.copy = baseui.copy
496 496
497 497 # Working directory VFS rooted at repository root.
498 498 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
499 499
500 500 # Main VFS for .hg/ directory.
501 501 hgpath = wdirvfs.join(b'.hg')
502 502 hgvfs = vfsmod.vfs(hgpath, cacheaudited=True)
503 503
504 504 # The .hg/ path should exist and should be a directory. All other
505 505 # cases are errors.
506 506 if not hgvfs.isdir():
507 507 try:
508 508 hgvfs.stat()
509 509 except OSError as e:
510 510 if e.errno != errno.ENOENT:
511 511 raise
512 512 except ValueError as e:
513 513 # Can be raised on Python 3.8 when path is invalid.
514 514 raise error.Abort(
515 515 _(b'invalid path %s: %s') % (path, pycompat.bytestr(e))
516 516 )
517 517
518 518 raise error.RepoError(_(b'repository %s not found') % path)
519 519
520 520 # .hg/requires file contains a newline-delimited list of
521 521 # features/capabilities the opener (us) must have in order to use
522 522 # the repository. This file was introduced in Mercurial 0.9.2,
523 523 # which means very old repositories may not have one. We assume
524 524 # a missing file translates to no requirements.
525 525 try:
526 526 requirements = set(hgvfs.read(b'requires').splitlines())
527 527 except IOError as e:
528 528 if e.errno != errno.ENOENT:
529 529 raise
530 530 requirements = set()
531 531
532 532 # The .hg/hgrc file may load extensions or contain config options
533 533 # that influence repository construction. Attempt to load it and
534 534 # process any new extensions that it may have pulled in.
535 535 if loadhgrc(ui, wdirvfs, hgvfs, requirements):
536 536 afterhgrcload(ui, wdirvfs, hgvfs, requirements)
537 537 extensions.loadall(ui)
538 538 extensions.populateui(ui)
539 539
540 540 # Set of module names of extensions loaded for this repository.
541 541 extensionmodulenames = {m.__name__ for n, m in extensions.extensions(ui)}
542 542
543 543 supportedrequirements = gathersupportedrequirements(ui)
544 544
545 545 # We first validate the requirements are known.
546 546 ensurerequirementsrecognized(requirements, supportedrequirements)
547 547
548 548 # Then we validate that the known set is reasonable to use together.
549 549 ensurerequirementscompatible(ui, requirements)
550 550
551 551 # TODO there are unhandled edge cases related to opening repositories with
552 552 # shared storage. If storage is shared, we should also test for requirements
553 553 # compatibility in the pointed-to repo. This entails loading the .hg/hgrc in
554 554 # that repo, as that repo may load extensions needed to open it. This is a
555 555 # bit complicated because we don't want the other hgrc to overwrite settings
556 556 # in this hgrc.
557 557 #
558 558 # This bug is somewhat mitigated by the fact that we copy the .hg/requires
559 559 # file when sharing repos. But if a requirement is added after the share is
560 560 # performed, thereby introducing a new requirement for the opener, we may
561 561 # will not see that and could encounter a run-time error interacting with
562 562 # that shared store since it has an unknown-to-us requirement.
563 563
564 564 # At this point, we know we should be capable of opening the repository.
565 565 # Now get on with doing that.
566 566
567 567 features = set()
568 568
569 569 # The "store" part of the repository holds versioned data. How it is
570 570 # accessed is determined by various requirements. The ``shared`` or
571 571 # ``relshared`` requirements indicate the store lives in the path contained
572 572 # in the ``.hg/sharedpath`` file. This is an absolute path for
573 573 # ``shared`` and relative to ``.hg/`` for ``relshared``.
574 574 if b'shared' in requirements or b'relshared' in requirements:
575 575 sharedpath = hgvfs.read(b'sharedpath').rstrip(b'\n')
576 576 if b'relshared' in requirements:
577 577 sharedpath = hgvfs.join(sharedpath)
578 578
579 579 sharedvfs = vfsmod.vfs(sharedpath, realpath=True)
580 580
581 581 if not sharedvfs.exists():
582 582 raise error.RepoError(
583 583 _(b'.hg/sharedpath points to nonexistent directory %s')
584 584 % sharedvfs.base
585 585 )
586 586
587 587 features.add(repository.REPO_FEATURE_SHARED_STORAGE)
588 588
589 589 storebasepath = sharedvfs.base
590 590 cachepath = sharedvfs.join(b'cache')
591 591 else:
592 592 storebasepath = hgvfs.base
593 593 cachepath = hgvfs.join(b'cache')
594 594 wcachepath = hgvfs.join(b'wcache')
595 595
596 596 # The store has changed over time and the exact layout is dictated by
597 597 # requirements. The store interface abstracts differences across all
598 598 # of them.
599 599 store = makestore(
600 600 requirements,
601 601 storebasepath,
602 602 lambda base: vfsmod.vfs(base, cacheaudited=True),
603 603 )
604 604 hgvfs.createmode = store.createmode
605 605
606 606 storevfs = store.vfs
607 607 storevfs.options = resolvestorevfsoptions(ui, requirements, features)
608 608
609 609 # The cache vfs is used to manage cache files.
610 610 cachevfs = vfsmod.vfs(cachepath, cacheaudited=True)
611 611 cachevfs.createmode = store.createmode
612 612 # The cache vfs is used to manage cache files related to the working copy
613 613 wcachevfs = vfsmod.vfs(wcachepath, cacheaudited=True)
614 614 wcachevfs.createmode = store.createmode
615 615
616 616 # Now resolve the type for the repository object. We do this by repeatedly
617 617 # calling a factory function to produces types for specific aspects of the
618 618 # repo's operation. The aggregate returned types are used as base classes
619 619 # for a dynamically-derived type, which will represent our new repository.
620 620
621 621 bases = []
622 622 extrastate = {}
623 623
624 624 for iface, fn in REPO_INTERFACES:
625 625 # We pass all potentially useful state to give extensions tons of
626 626 # flexibility.
627 627 typ = fn()(
628 628 ui=ui,
629 629 intents=intents,
630 630 requirements=requirements,
631 631 features=features,
632 632 wdirvfs=wdirvfs,
633 633 hgvfs=hgvfs,
634 634 store=store,
635 635 storevfs=storevfs,
636 636 storeoptions=storevfs.options,
637 637 cachevfs=cachevfs,
638 638 wcachevfs=wcachevfs,
639 639 extensionmodulenames=extensionmodulenames,
640 640 extrastate=extrastate,
641 641 baseclasses=bases,
642 642 )
643 643
644 644 if not isinstance(typ, type):
645 645 raise error.ProgrammingError(
646 646 b'unable to construct type for %s' % iface
647 647 )
648 648
649 649 bases.append(typ)
650 650
651 651 # type() allows you to use characters in type names that wouldn't be
652 652 # recognized as Python symbols in source code. We abuse that to add
653 653 # rich information about our constructed repo.
654 654 name = pycompat.sysstr(
655 655 b'derivedrepo:%s<%s>' % (wdirvfs.base, b','.join(sorted(requirements)))
656 656 )
657 657
658 658 cls = type(name, tuple(bases), {})
659 659
660 660 return cls(
661 661 baseui=baseui,
662 662 ui=ui,
663 663 origroot=path,
664 664 wdirvfs=wdirvfs,
665 665 hgvfs=hgvfs,
666 666 requirements=requirements,
667 667 supportedrequirements=supportedrequirements,
668 668 sharedpath=storebasepath,
669 669 store=store,
670 670 cachevfs=cachevfs,
671 671 wcachevfs=wcachevfs,
672 672 features=features,
673 673 intents=intents,
674 674 )
675 675
676 676
677 677 def loadhgrc(ui, wdirvfs, hgvfs, requirements):
678 678 """Load hgrc files/content into a ui instance.
679 679
680 680 This is called during repository opening to load any additional
681 681 config files or settings relevant to the current repository.
682 682
683 683 Returns a bool indicating whether any additional configs were loaded.
684 684
685 685 Extensions should monkeypatch this function to modify how per-repo
686 686 configs are loaded. For example, an extension may wish to pull in
687 687 configs from alternate files or sources.
688 688 """
689 689 if not rcutil.use_repo_hgrc():
690 690 return False
691 691 try:
692 692 ui.readconfig(hgvfs.join(b'hgrc'), root=wdirvfs.base)
693 693 return True
694 694 except IOError:
695 695 return False
696 696
697 697
698 698 def afterhgrcload(ui, wdirvfs, hgvfs, requirements):
699 699 """Perform additional actions after .hg/hgrc is loaded.
700 700
701 701 This function is called during repository loading immediately after
702 702 the .hg/hgrc file is loaded and before per-repo extensions are loaded.
703 703
704 704 The function can be used to validate configs, automatically add
705 705 options (including extensions) based on requirements, etc.
706 706 """
707 707
708 708 # Map of requirements to list of extensions to load automatically when
709 709 # requirement is present.
710 710 autoextensions = {
711 711 b'git': [b'git'],
712 712 b'largefiles': [b'largefiles'],
713 713 b'lfs': [b'lfs'],
714 714 }
715 715
716 716 for requirement, names in sorted(autoextensions.items()):
717 717 if requirement not in requirements:
718 718 continue
719 719
720 720 for name in names:
721 721 if not ui.hasconfig(b'extensions', name):
722 722 ui.setconfig(b'extensions', name, b'', source=b'autoload')
723 723
724 724
725 725 def gathersupportedrequirements(ui):
726 726 """Determine the complete set of recognized requirements."""
727 727 # Start with all requirements supported by this file.
728 728 supported = set(localrepository._basesupported)
729 729
730 730 # Execute ``featuresetupfuncs`` entries if they belong to an extension
731 731 # relevant to this ui instance.
732 732 modules = {m.__name__ for n, m in extensions.extensions(ui)}
733 733
734 734 for fn in featuresetupfuncs:
735 735 if fn.__module__ in modules:
736 736 fn(ui, supported)
737 737
738 738 # Add derived requirements from registered compression engines.
739 739 for name in util.compengines:
740 740 engine = util.compengines[name]
741 741 if engine.available() and engine.revlogheader():
742 742 supported.add(b'exp-compression-%s' % name)
743 743 if engine.name() == b'zstd':
744 744 supported.add(b'revlog-compression-zstd')
745 745
746 746 return supported
747 747
748 748
749 749 def ensurerequirementsrecognized(requirements, supported):
750 750 """Validate that a set of local requirements is recognized.
751 751
752 752 Receives a set of requirements. Raises an ``error.RepoError`` if there
753 753 exists any requirement in that set that currently loaded code doesn't
754 754 recognize.
755 755
756 756 Returns a set of supported requirements.
757 757 """
758 758 missing = set()
759 759
760 760 for requirement in requirements:
761 761 if requirement in supported:
762 762 continue
763 763
764 764 if not requirement or not requirement[0:1].isalnum():
765 765 raise error.RequirementError(_(b'.hg/requires file is corrupt'))
766 766
767 767 missing.add(requirement)
768 768
769 769 if missing:
770 770 raise error.RequirementError(
771 771 _(b'repository requires features unknown to this Mercurial: %s')
772 772 % b' '.join(sorted(missing)),
773 773 hint=_(
774 774 b'see https://mercurial-scm.org/wiki/MissingRequirement '
775 775 b'for more information'
776 776 ),
777 777 )
778 778
779 779
780 780 def ensurerequirementscompatible(ui, requirements):
781 781 """Validates that a set of recognized requirements is mutually compatible.
782 782
783 783 Some requirements may not be compatible with others or require
784 784 config options that aren't enabled. This function is called during
785 785 repository opening to ensure that the set of requirements needed
786 786 to open a repository is sane and compatible with config options.
787 787
788 788 Extensions can monkeypatch this function to perform additional
789 789 checking.
790 790
791 791 ``error.RepoError`` should be raised on failure.
792 792 """
793 793 if b'exp-sparse' in requirements and not sparse.enabled:
794 794 raise error.RepoError(
795 795 _(
796 796 b'repository is using sparse feature but '
797 797 b'sparse is not enabled; enable the '
798 798 b'"sparse" extensions to access'
799 799 )
800 800 )
801 801
802 802
803 803 def makestore(requirements, path, vfstype):
804 804 """Construct a storage object for a repository."""
805 805 if b'store' in requirements:
806 806 if b'fncache' in requirements:
807 807 return storemod.fncachestore(
808 808 path, vfstype, b'dotencode' in requirements
809 809 )
810 810
811 811 return storemod.encodedstore(path, vfstype)
812 812
813 813 return storemod.basicstore(path, vfstype)
814 814
815 815
816 816 def resolvestorevfsoptions(ui, requirements, features):
817 817 """Resolve the options to pass to the store vfs opener.
818 818
819 819 The returned dict is used to influence behavior of the storage layer.
820 820 """
821 821 options = {}
822 822
823 823 if b'treemanifest' in requirements:
824 824 options[b'treemanifest'] = True
825 825
826 826 # experimental config: format.manifestcachesize
827 827 manifestcachesize = ui.configint(b'format', b'manifestcachesize')
828 828 if manifestcachesize is not None:
829 829 options[b'manifestcachesize'] = manifestcachesize
830 830
831 831 # In the absence of another requirement superseding a revlog-related
832 832 # requirement, we have to assume the repo is using revlog version 0.
833 833 # This revlog format is super old and we don't bother trying to parse
834 834 # opener options for it because those options wouldn't do anything
835 835 # meaningful on such old repos.
836 836 if b'revlogv1' in requirements or REVLOGV2_REQUIREMENT in requirements:
837 837 options.update(resolverevlogstorevfsoptions(ui, requirements, features))
838 838 else: # explicitly mark repo as using revlogv0
839 839 options[b'revlogv0'] = True
840 840
841 841 if COPIESSDC_REQUIREMENT in requirements:
842 842 options[b'copies-storage'] = b'changeset-sidedata'
843 843 else:
844 844 writecopiesto = ui.config(b'experimental', b'copies.write-to')
845 845 copiesextramode = (b'changeset-only', b'compatibility')
846 846 if writecopiesto in copiesextramode:
847 847 options[b'copies-storage'] = b'extra'
848 848
849 849 return options
850 850
851 851
852 852 def resolverevlogstorevfsoptions(ui, requirements, features):
853 853 """Resolve opener options specific to revlogs."""
854 854
855 855 options = {}
856 856 options[b'flagprocessors'] = {}
857 857
858 858 if b'revlogv1' in requirements:
859 859 options[b'revlogv1'] = True
860 860 if REVLOGV2_REQUIREMENT in requirements:
861 861 options[b'revlogv2'] = True
862 862
863 863 if b'generaldelta' in requirements:
864 864 options[b'generaldelta'] = True
865 865
866 866 # experimental config: format.chunkcachesize
867 867 chunkcachesize = ui.configint(b'format', b'chunkcachesize')
868 868 if chunkcachesize is not None:
869 869 options[b'chunkcachesize'] = chunkcachesize
870 870
871 871 deltabothparents = ui.configbool(
872 872 b'storage', b'revlog.optimize-delta-parent-choice'
873 873 )
874 874 options[b'deltabothparents'] = deltabothparents
875 875
876 876 lazydelta = ui.configbool(b'storage', b'revlog.reuse-external-delta')
877 877 lazydeltabase = False
878 878 if lazydelta:
879 879 lazydeltabase = ui.configbool(
880 880 b'storage', b'revlog.reuse-external-delta-parent'
881 881 )
882 882 if lazydeltabase is None:
883 883 lazydeltabase = not scmutil.gddeltaconfig(ui)
884 884 options[b'lazydelta'] = lazydelta
885 885 options[b'lazydeltabase'] = lazydeltabase
886 886
887 887 chainspan = ui.configbytes(b'experimental', b'maxdeltachainspan')
888 888 if 0 <= chainspan:
889 889 options[b'maxdeltachainspan'] = chainspan
890 890
891 891 mmapindexthreshold = ui.configbytes(b'experimental', b'mmapindexthreshold')
892 892 if mmapindexthreshold is not None:
893 893 options[b'mmapindexthreshold'] = mmapindexthreshold
894 894
895 895 withsparseread = ui.configbool(b'experimental', b'sparse-read')
896 896 srdensitythres = float(
897 897 ui.config(b'experimental', b'sparse-read.density-threshold')
898 898 )
899 899 srmingapsize = ui.configbytes(b'experimental', b'sparse-read.min-gap-size')
900 900 options[b'with-sparse-read'] = withsparseread
901 901 options[b'sparse-read-density-threshold'] = srdensitythres
902 902 options[b'sparse-read-min-gap-size'] = srmingapsize
903 903
904 904 sparserevlog = SPARSEREVLOG_REQUIREMENT in requirements
905 905 options[b'sparse-revlog'] = sparserevlog
906 906 if sparserevlog:
907 907 options[b'generaldelta'] = True
908 908
909 909 sidedata = SIDEDATA_REQUIREMENT in requirements
910 910 options[b'side-data'] = sidedata
911 911
912 912 maxchainlen = None
913 913 if sparserevlog:
914 914 maxchainlen = revlogconst.SPARSE_REVLOG_MAX_CHAIN_LENGTH
915 915 # experimental config: format.maxchainlen
916 916 maxchainlen = ui.configint(b'format', b'maxchainlen', maxchainlen)
917 917 if maxchainlen is not None:
918 918 options[b'maxchainlen'] = maxchainlen
919 919
920 920 for r in requirements:
921 921 # we allow multiple compression engine requirement to co-exist because
922 922 # strickly speaking, revlog seems to support mixed compression style.
923 923 #
924 924 # The compression used for new entries will be "the last one"
925 925 prefix = r.startswith
926 926 if prefix(b'revlog-compression-') or prefix(b'exp-compression-'):
927 927 options[b'compengine'] = r.split(b'-', 2)[2]
928 928
929 929 options[b'zlib.level'] = ui.configint(b'storage', b'revlog.zlib.level')
930 930 if options[b'zlib.level'] is not None:
931 931 if not (0 <= options[b'zlib.level'] <= 9):
932 932 msg = _(b'invalid value for `storage.revlog.zlib.level` config: %d')
933 933 raise error.Abort(msg % options[b'zlib.level'])
934 934 options[b'zstd.level'] = ui.configint(b'storage', b'revlog.zstd.level')
935 935 if options[b'zstd.level'] is not None:
936 936 if not (0 <= options[b'zstd.level'] <= 22):
937 937 msg = _(b'invalid value for `storage.revlog.zstd.level` config: %d')
938 938 raise error.Abort(msg % options[b'zstd.level'])
939 939
940 940 if repository.NARROW_REQUIREMENT in requirements:
941 941 options[b'enableellipsis'] = True
942 942
943 943 if ui.configbool(b'experimental', b'rust.index'):
944 944 options[b'rust.index'] = True
945 945 if NODEMAP_REQUIREMENT in requirements:
946 946 options[b'persistent-nodemap'] = True
947 947 if ui.configbool(b'storage', b'revlog.nodemap.mmap'):
948 948 options[b'persistent-nodemap.mmap'] = True
949 949 epnm = ui.config(b'storage', b'revlog.nodemap.mode')
950 950 options[b'persistent-nodemap.mode'] = epnm
951 951 if ui.configbool(b'devel', b'persistent-nodemap'):
952 952 options[b'devel-force-nodemap'] = True
953 953
954 954 return options
955 955
956 956
957 957 def makemain(**kwargs):
958 958 """Produce a type conforming to ``ilocalrepositorymain``."""
959 959 return localrepository
960 960
961 961
962 962 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
963 963 class revlogfilestorage(object):
964 964 """File storage when using revlogs."""
965 965
966 966 def file(self, path):
967 967 if path[0] == b'/':
968 968 path = path[1:]
969 969
970 970 return filelog.filelog(self.svfs, path)
971 971
972 972
973 973 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
974 974 class revlognarrowfilestorage(object):
975 975 """File storage when using revlogs and narrow files."""
976 976
977 977 def file(self, path):
978 978 if path[0] == b'/':
979 979 path = path[1:]
980 980
981 981 return filelog.narrowfilelog(self.svfs, path, self._storenarrowmatch)
982 982
983 983
984 984 def makefilestorage(requirements, features, **kwargs):
985 985 """Produce a type conforming to ``ilocalrepositoryfilestorage``."""
986 986 features.add(repository.REPO_FEATURE_REVLOG_FILE_STORAGE)
987 987 features.add(repository.REPO_FEATURE_STREAM_CLONE)
988 988
989 989 if repository.NARROW_REQUIREMENT in requirements:
990 990 return revlognarrowfilestorage
991 991 else:
992 992 return revlogfilestorage
993 993
994 994
995 995 # List of repository interfaces and factory functions for them. Each
996 996 # will be called in order during ``makelocalrepository()`` to iteratively
997 997 # derive the final type for a local repository instance. We capture the
998 998 # function as a lambda so we don't hold a reference and the module-level
999 999 # functions can be wrapped.
1000 1000 REPO_INTERFACES = [
1001 1001 (repository.ilocalrepositorymain, lambda: makemain),
1002 1002 (repository.ilocalrepositoryfilestorage, lambda: makefilestorage),
1003 1003 ]
1004 1004
1005 1005
1006 1006 @interfaceutil.implementer(repository.ilocalrepositorymain)
1007 1007 class localrepository(object):
1008 1008 """Main class for representing local repositories.
1009 1009
1010 1010 All local repositories are instances of this class.
1011 1011
1012 1012 Constructed on its own, instances of this class are not usable as
1013 1013 repository objects. To obtain a usable repository object, call
1014 1014 ``hg.repository()``, ``localrepo.instance()``, or
1015 1015 ``localrepo.makelocalrepository()``. The latter is the lowest-level.
1016 1016 ``instance()`` adds support for creating new repositories.
1017 1017 ``hg.repository()`` adds more extension integration, including calling
1018 1018 ``reposetup()``. Generally speaking, ``hg.repository()`` should be
1019 1019 used.
1020 1020 """
1021 1021
1022 1022 # obsolete experimental requirements:
1023 1023 # - manifestv2: An experimental new manifest format that allowed
1024 1024 # for stem compression of long paths. Experiment ended up not
1025 1025 # being successful (repository sizes went up due to worse delta
1026 1026 # chains), and the code was deleted in 4.6.
1027 1027 supportedformats = {
1028 1028 b'revlogv1',
1029 1029 b'generaldelta',
1030 1030 b'treemanifest',
1031 1031 COPIESSDC_REQUIREMENT,
1032 1032 REVLOGV2_REQUIREMENT,
1033 1033 SIDEDATA_REQUIREMENT,
1034 1034 SPARSEREVLOG_REQUIREMENT,
1035 1035 NODEMAP_REQUIREMENT,
1036 1036 bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT,
1037 1037 }
1038 1038 _basesupported = supportedformats | {
1039 1039 b'store',
1040 1040 b'fncache',
1041 1041 b'shared',
1042 1042 b'relshared',
1043 1043 b'dotencode',
1044 1044 b'exp-sparse',
1045 1045 b'internal-phase',
1046 1046 }
1047 1047
1048 1048 # list of prefix for file which can be written without 'wlock'
1049 1049 # Extensions should extend this list when needed
1050 1050 _wlockfreeprefix = {
1051 1051 # We migh consider requiring 'wlock' for the next
1052 1052 # two, but pretty much all the existing code assume
1053 1053 # wlock is not needed so we keep them excluded for
1054 1054 # now.
1055 1055 b'hgrc',
1056 1056 b'requires',
1057 1057 # XXX cache is a complicatged business someone
1058 1058 # should investigate this in depth at some point
1059 1059 b'cache/',
1060 1060 # XXX shouldn't be dirstate covered by the wlock?
1061 1061 b'dirstate',
1062 1062 # XXX bisect was still a bit too messy at the time
1063 1063 # this changeset was introduced. Someone should fix
1064 1064 # the remainig bit and drop this line
1065 1065 b'bisect.state',
1066 1066 }
1067 1067
1068 1068 def __init__(
1069 1069 self,
1070 1070 baseui,
1071 1071 ui,
1072 1072 origroot,
1073 1073 wdirvfs,
1074 1074 hgvfs,
1075 1075 requirements,
1076 1076 supportedrequirements,
1077 1077 sharedpath,
1078 1078 store,
1079 1079 cachevfs,
1080 1080 wcachevfs,
1081 1081 features,
1082 1082 intents=None,
1083 1083 ):
1084 1084 """Create a new local repository instance.
1085 1085
1086 1086 Most callers should use ``hg.repository()``, ``localrepo.instance()``,
1087 1087 or ``localrepo.makelocalrepository()`` for obtaining a new repository
1088 1088 object.
1089 1089
1090 1090 Arguments:
1091 1091
1092 1092 baseui
1093 1093 ``ui.ui`` instance that ``ui`` argument was based off of.
1094 1094
1095 1095 ui
1096 1096 ``ui.ui`` instance for use by the repository.
1097 1097
1098 1098 origroot
1099 1099 ``bytes`` path to working directory root of this repository.
1100 1100
1101 1101 wdirvfs
1102 1102 ``vfs.vfs`` rooted at the working directory.
1103 1103
1104 1104 hgvfs
1105 1105 ``vfs.vfs`` rooted at .hg/
1106 1106
1107 1107 requirements
1108 1108 ``set`` of bytestrings representing repository opening requirements.
1109 1109
1110 1110 supportedrequirements
1111 1111 ``set`` of bytestrings representing repository requirements that we
1112 1112 know how to open. May be a supetset of ``requirements``.
1113 1113
1114 1114 sharedpath
1115 1115 ``bytes`` Defining path to storage base directory. Points to a
1116 1116 ``.hg/`` directory somewhere.
1117 1117
1118 1118 store
1119 1119 ``store.basicstore`` (or derived) instance providing access to
1120 1120 versioned storage.
1121 1121
1122 1122 cachevfs
1123 1123 ``vfs.vfs`` used for cache files.
1124 1124
1125 1125 wcachevfs
1126 1126 ``vfs.vfs`` used for cache files related to the working copy.
1127 1127
1128 1128 features
1129 1129 ``set`` of bytestrings defining features/capabilities of this
1130 1130 instance.
1131 1131
1132 1132 intents
1133 1133 ``set`` of system strings indicating what this repo will be used
1134 1134 for.
1135 1135 """
1136 1136 self.baseui = baseui
1137 1137 self.ui = ui
1138 1138 self.origroot = origroot
1139 1139 # vfs rooted at working directory.
1140 1140 self.wvfs = wdirvfs
1141 1141 self.root = wdirvfs.base
1142 1142 # vfs rooted at .hg/. Used to access most non-store paths.
1143 1143 self.vfs = hgvfs
1144 1144 self.path = hgvfs.base
1145 1145 self.requirements = requirements
1146 1146 self.supported = supportedrequirements
1147 1147 self.sharedpath = sharedpath
1148 1148 self.store = store
1149 1149 self.cachevfs = cachevfs
1150 1150 self.wcachevfs = wcachevfs
1151 1151 self.features = features
1152 1152
1153 1153 self.filtername = None
1154 1154
1155 1155 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1156 1156 b'devel', b'check-locks'
1157 1157 ):
1158 1158 self.vfs.audit = self._getvfsward(self.vfs.audit)
1159 1159 # A list of callback to shape the phase if no data were found.
1160 1160 # Callback are in the form: func(repo, roots) --> processed root.
1161 1161 # This list it to be filled by extension during repo setup
1162 1162 self._phasedefaults = []
1163 1163
1164 1164 color.setup(self.ui)
1165 1165
1166 1166 self.spath = self.store.path
1167 1167 self.svfs = self.store.vfs
1168 1168 self.sjoin = self.store.join
1169 1169 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1170 1170 b'devel', b'check-locks'
1171 1171 ):
1172 1172 if util.safehasattr(self.svfs, b'vfs'): # this is filtervfs
1173 1173 self.svfs.vfs.audit = self._getsvfsward(self.svfs.vfs.audit)
1174 1174 else: # standard vfs
1175 1175 self.svfs.audit = self._getsvfsward(self.svfs.audit)
1176 1176
1177 1177 self._dirstatevalidatewarned = False
1178 1178
1179 1179 self._branchcaches = branchmap.BranchMapCache()
1180 1180 self._revbranchcache = None
1181 1181 self._filterpats = {}
1182 1182 self._datafilters = {}
1183 1183 self._transref = self._lockref = self._wlockref = None
1184 1184
1185 1185 # A cache for various files under .hg/ that tracks file changes,
1186 1186 # (used by the filecache decorator)
1187 1187 #
1188 1188 # Maps a property name to its util.filecacheentry
1189 1189 self._filecache = {}
1190 1190
1191 1191 # hold sets of revision to be filtered
1192 1192 # should be cleared when something might have changed the filter value:
1193 1193 # - new changesets,
1194 1194 # - phase change,
1195 1195 # - new obsolescence marker,
1196 1196 # - working directory parent change,
1197 1197 # - bookmark changes
1198 1198 self.filteredrevcache = {}
1199 1199
1200 1200 # post-dirstate-status hooks
1201 1201 self._postdsstatus = []
1202 1202
1203 1203 # generic mapping between names and nodes
1204 1204 self.names = namespaces.namespaces()
1205 1205
1206 1206 # Key to signature value.
1207 1207 self._sparsesignaturecache = {}
1208 1208 # Signature to cached matcher instance.
1209 1209 self._sparsematchercache = {}
1210 1210
1211 1211 self._extrafilterid = repoview.extrafilter(ui)
1212 1212
1213 1213 self.filecopiesmode = None
1214 1214 if COPIESSDC_REQUIREMENT in self.requirements:
1215 1215 self.filecopiesmode = b'changeset-sidedata'
1216 1216
1217 1217 def _getvfsward(self, origfunc):
1218 1218 """build a ward for self.vfs"""
1219 1219 rref = weakref.ref(self)
1220 1220
1221 1221 def checkvfs(path, mode=None):
1222 1222 ret = origfunc(path, mode=mode)
1223 1223 repo = rref()
1224 1224 if (
1225 1225 repo is None
1226 1226 or not util.safehasattr(repo, b'_wlockref')
1227 1227 or not util.safehasattr(repo, b'_lockref')
1228 1228 ):
1229 1229 return
1230 1230 if mode in (None, b'r', b'rb'):
1231 1231 return
1232 1232 if path.startswith(repo.path):
1233 1233 # truncate name relative to the repository (.hg)
1234 1234 path = path[len(repo.path) + 1 :]
1235 1235 if path.startswith(b'cache/'):
1236 1236 msg = b'accessing cache with vfs instead of cachevfs: "%s"'
1237 1237 repo.ui.develwarn(msg % path, stacklevel=3, config=b"cache-vfs")
1238 1238 # path prefixes covered by 'lock'
1239 1239 vfs_path_prefixes = (b'journal.', b'undo.', b'strip-backup/')
1240 1240 if any(path.startswith(prefix) for prefix in vfs_path_prefixes):
1241 1241 if repo._currentlock(repo._lockref) is None:
1242 1242 repo.ui.develwarn(
1243 1243 b'write with no lock: "%s"' % path,
1244 1244 stacklevel=3,
1245 1245 config=b'check-locks',
1246 1246 )
1247 1247 elif repo._currentlock(repo._wlockref) is None:
1248 1248 # rest of vfs files are covered by 'wlock'
1249 1249 #
1250 1250 # exclude special files
1251 1251 for prefix in self._wlockfreeprefix:
1252 1252 if path.startswith(prefix):
1253 1253 return
1254 1254 repo.ui.develwarn(
1255 1255 b'write with no wlock: "%s"' % path,
1256 1256 stacklevel=3,
1257 1257 config=b'check-locks',
1258 1258 )
1259 1259 return ret
1260 1260
1261 1261 return checkvfs
1262 1262
1263 1263 def _getsvfsward(self, origfunc):
1264 1264 """build a ward for self.svfs"""
1265 1265 rref = weakref.ref(self)
1266 1266
1267 1267 def checksvfs(path, mode=None):
1268 1268 ret = origfunc(path, mode=mode)
1269 1269 repo = rref()
1270 1270 if repo is None or not util.safehasattr(repo, b'_lockref'):
1271 1271 return
1272 1272 if mode in (None, b'r', b'rb'):
1273 1273 return
1274 1274 if path.startswith(repo.sharedpath):
1275 1275 # truncate name relative to the repository (.hg)
1276 1276 path = path[len(repo.sharedpath) + 1 :]
1277 1277 if repo._currentlock(repo._lockref) is None:
1278 1278 repo.ui.develwarn(
1279 1279 b'write with no lock: "%s"' % path, stacklevel=4
1280 1280 )
1281 1281 return ret
1282 1282
1283 1283 return checksvfs
1284 1284
1285 1285 def close(self):
1286 1286 self._writecaches()
1287 1287
1288 1288 def _writecaches(self):
1289 1289 if self._revbranchcache:
1290 1290 self._revbranchcache.write()
1291 1291
1292 1292 def _restrictcapabilities(self, caps):
1293 1293 if self.ui.configbool(b'experimental', b'bundle2-advertise'):
1294 1294 caps = set(caps)
1295 1295 capsblob = bundle2.encodecaps(
1296 1296 bundle2.getrepocaps(self, role=b'client')
1297 1297 )
1298 1298 caps.add(b'bundle2=' + urlreq.quote(capsblob))
1299 1299 return caps
1300 1300
1301 1301 # Don't cache auditor/nofsauditor, or you'll end up with reference cycle:
1302 1302 # self -> auditor -> self._checknested -> self
1303 1303
1304 1304 @property
1305 1305 def auditor(self):
1306 1306 # This is only used by context.workingctx.match in order to
1307 1307 # detect files in subrepos.
1308 1308 return pathutil.pathauditor(self.root, callback=self._checknested)
1309 1309
1310 1310 @property
1311 1311 def nofsauditor(self):
1312 1312 # This is only used by context.basectx.match in order to detect
1313 1313 # files in subrepos.
1314 1314 return pathutil.pathauditor(
1315 1315 self.root, callback=self._checknested, realfs=False, cached=True
1316 1316 )
1317 1317
1318 1318 def _checknested(self, path):
1319 1319 """Determine if path is a legal nested repository."""
1320 1320 if not path.startswith(self.root):
1321 1321 return False
1322 1322 subpath = path[len(self.root) + 1 :]
1323 1323 normsubpath = util.pconvert(subpath)
1324 1324
1325 1325 # XXX: Checking against the current working copy is wrong in
1326 1326 # the sense that it can reject things like
1327 1327 #
1328 1328 # $ hg cat -r 10 sub/x.txt
1329 1329 #
1330 1330 # if sub/ is no longer a subrepository in the working copy
1331 1331 # parent revision.
1332 1332 #
1333 1333 # However, it can of course also allow things that would have
1334 1334 # been rejected before, such as the above cat command if sub/
1335 1335 # is a subrepository now, but was a normal directory before.
1336 1336 # The old path auditor would have rejected by mistake since it
1337 1337 # panics when it sees sub/.hg/.
1338 1338 #
1339 1339 # All in all, checking against the working copy seems sensible
1340 1340 # since we want to prevent access to nested repositories on
1341 1341 # the filesystem *now*.
1342 1342 ctx = self[None]
1343 1343 parts = util.splitpath(subpath)
1344 1344 while parts:
1345 1345 prefix = b'/'.join(parts)
1346 1346 if prefix in ctx.substate:
1347 1347 if prefix == normsubpath:
1348 1348 return True
1349 1349 else:
1350 1350 sub = ctx.sub(prefix)
1351 1351 return sub.checknested(subpath[len(prefix) + 1 :])
1352 1352 else:
1353 1353 parts.pop()
1354 1354 return False
1355 1355
1356 1356 def peer(self):
1357 1357 return localpeer(self) # not cached to avoid reference cycle
1358 1358
1359 1359 def unfiltered(self):
1360 1360 """Return unfiltered version of the repository
1361 1361
1362 1362 Intended to be overwritten by filtered repo."""
1363 1363 return self
1364 1364
1365 1365 def filtered(self, name, visibilityexceptions=None):
1366 1366 """Return a filtered version of a repository
1367 1367
1368 1368 The `name` parameter is the identifier of the requested view. This
1369 1369 will return a repoview object set "exactly" to the specified view.
1370 1370
1371 1371 This function does not apply recursive filtering to a repository. For
1372 1372 example calling `repo.filtered("served")` will return a repoview using
1373 1373 the "served" view, regardless of the initial view used by `repo`.
1374 1374
1375 1375 In other word, there is always only one level of `repoview` "filtering".
1376 1376 """
1377 1377 if self._extrafilterid is not None and b'%' not in name:
1378 1378 name = name + b'%' + self._extrafilterid
1379 1379
1380 1380 cls = repoview.newtype(self.unfiltered().__class__)
1381 1381 return cls(self, name, visibilityexceptions)
1382 1382
1383 1383 @mixedrepostorecache(
1384 1384 (b'bookmarks', b'plain'),
1385 1385 (b'bookmarks.current', b'plain'),
1386 1386 (b'bookmarks', b''),
1387 1387 (b'00changelog.i', b''),
1388 1388 )
1389 1389 def _bookmarks(self):
1390 1390 # Since the multiple files involved in the transaction cannot be
1391 1391 # written atomically (with current repository format), there is a race
1392 1392 # condition here.
1393 1393 #
1394 1394 # 1) changelog content A is read
1395 1395 # 2) outside transaction update changelog to content B
1396 1396 # 3) outside transaction update bookmark file referring to content B
1397 1397 # 4) bookmarks file content is read and filtered against changelog-A
1398 1398 #
1399 1399 # When this happens, bookmarks against nodes missing from A are dropped.
1400 1400 #
1401 1401 # Having this happening during read is not great, but it become worse
1402 1402 # when this happen during write because the bookmarks to the "unknown"
1403 1403 # nodes will be dropped for good. However, writes happen within locks.
1404 1404 # This locking makes it possible to have a race free consistent read.
1405 1405 # For this purpose data read from disc before locking are
1406 1406 # "invalidated" right after the locks are taken. This invalidations are
1407 1407 # "light", the `filecache` mechanism keep the data in memory and will
1408 1408 # reuse them if the underlying files did not changed. Not parsing the
1409 1409 # same data multiple times helps performances.
1410 1410 #
1411 1411 # Unfortunately in the case describe above, the files tracked by the
1412 1412 # bookmarks file cache might not have changed, but the in-memory
1413 1413 # content is still "wrong" because we used an older changelog content
1414 1414 # to process the on-disk data. So after locking, the changelog would be
1415 1415 # refreshed but `_bookmarks` would be preserved.
1416 1416 # Adding `00changelog.i` to the list of tracked file is not
1417 1417 # enough, because at the time we build the content for `_bookmarks` in
1418 1418 # (4), the changelog file has already diverged from the content used
1419 1419 # for loading `changelog` in (1)
1420 1420 #
1421 1421 # To prevent the issue, we force the changelog to be explicitly
1422 1422 # reloaded while computing `_bookmarks`. The data race can still happen
1423 1423 # without the lock (with a narrower window), but it would no longer go
1424 1424 # undetected during the lock time refresh.
1425 1425 #
1426 1426 # The new schedule is as follow
1427 1427 #
1428 1428 # 1) filecache logic detect that `_bookmarks` needs to be computed
1429 1429 # 2) cachestat for `bookmarks` and `changelog` are captured (for book)
1430 1430 # 3) We force `changelog` filecache to be tested
1431 1431 # 4) cachestat for `changelog` are captured (for changelog)
1432 1432 # 5) `_bookmarks` is computed and cached
1433 1433 #
1434 1434 # The step in (3) ensure we have a changelog at least as recent as the
1435 1435 # cache stat computed in (1). As a result at locking time:
1436 1436 # * if the changelog did not changed since (1) -> we can reuse the data
1437 1437 # * otherwise -> the bookmarks get refreshed.
1438 1438 self._refreshchangelog()
1439 1439 return bookmarks.bmstore(self)
1440 1440
1441 1441 def _refreshchangelog(self):
1442 1442 """make sure the in memory changelog match the on-disk one"""
1443 1443 if 'changelog' in vars(self) and self.currenttransaction() is None:
1444 1444 del self.changelog
1445 1445
1446 1446 @property
1447 1447 def _activebookmark(self):
1448 1448 return self._bookmarks.active
1449 1449
1450 1450 # _phasesets depend on changelog. what we need is to call
1451 1451 # _phasecache.invalidate() if '00changelog.i' was changed, but it
1452 1452 # can't be easily expressed in filecache mechanism.
1453 1453 @storecache(b'phaseroots', b'00changelog.i')
1454 1454 def _phasecache(self):
1455 1455 return phases.phasecache(self, self._phasedefaults)
1456 1456
1457 1457 @storecache(b'obsstore')
1458 1458 def obsstore(self):
1459 1459 return obsolete.makestore(self.ui, self)
1460 1460
1461 1461 @storecache(b'00changelog.i')
1462 1462 def changelog(self):
1463 1463 # load dirstate before changelog to avoid race see issue6303
1464 1464 self.dirstate.prefetch_parents()
1465 1465 return self.store.changelog(txnutil.mayhavepending(self.root))
1466 1466
1467 1467 @storecache(b'00manifest.i')
1468 1468 def manifestlog(self):
1469 1469 return self.store.manifestlog(self, self._storenarrowmatch)
1470 1470
1471 1471 @repofilecache(b'dirstate')
1472 1472 def dirstate(self):
1473 1473 return self._makedirstate()
1474 1474
1475 1475 def _makedirstate(self):
1476 1476 """Extension point for wrapping the dirstate per-repo."""
1477 1477 sparsematchfn = lambda: sparse.matcher(self)
1478 1478
1479 1479 return dirstate.dirstate(
1480 1480 self.vfs, self.ui, self.root, self._dirstatevalidate, sparsematchfn
1481 1481 )
1482 1482
1483 1483 def _dirstatevalidate(self, node):
1484 1484 try:
1485 1485 self.changelog.rev(node)
1486 1486 return node
1487 1487 except error.LookupError:
1488 1488 if not self._dirstatevalidatewarned:
1489 1489 self._dirstatevalidatewarned = True
1490 1490 self.ui.warn(
1491 1491 _(b"warning: ignoring unknown working parent %s!\n")
1492 1492 % short(node)
1493 1493 )
1494 1494 return nullid
1495 1495
1496 1496 @storecache(narrowspec.FILENAME)
1497 1497 def narrowpats(self):
1498 1498 """matcher patterns for this repository's narrowspec
1499 1499
1500 1500 A tuple of (includes, excludes).
1501 1501 """
1502 1502 return narrowspec.load(self)
1503 1503
1504 1504 @storecache(narrowspec.FILENAME)
1505 1505 def _storenarrowmatch(self):
1506 1506 if repository.NARROW_REQUIREMENT not in self.requirements:
1507 1507 return matchmod.always()
1508 1508 include, exclude = self.narrowpats
1509 1509 return narrowspec.match(self.root, include=include, exclude=exclude)
1510 1510
1511 1511 @storecache(narrowspec.FILENAME)
1512 1512 def _narrowmatch(self):
1513 1513 if repository.NARROW_REQUIREMENT not in self.requirements:
1514 1514 return matchmod.always()
1515 1515 narrowspec.checkworkingcopynarrowspec(self)
1516 1516 include, exclude = self.narrowpats
1517 1517 return narrowspec.match(self.root, include=include, exclude=exclude)
1518 1518
1519 1519 def narrowmatch(self, match=None, includeexact=False):
1520 1520 """matcher corresponding the the repo's narrowspec
1521 1521
1522 1522 If `match` is given, then that will be intersected with the narrow
1523 1523 matcher.
1524 1524
1525 1525 If `includeexact` is True, then any exact matches from `match` will
1526 1526 be included even if they're outside the narrowspec.
1527 1527 """
1528 1528 if match:
1529 1529 if includeexact and not self._narrowmatch.always():
1530 1530 # do not exclude explicitly-specified paths so that they can
1531 1531 # be warned later on
1532 1532 em = matchmod.exact(match.files())
1533 1533 nm = matchmod.unionmatcher([self._narrowmatch, em])
1534 1534 return matchmod.intersectmatchers(match, nm)
1535 1535 return matchmod.intersectmatchers(match, self._narrowmatch)
1536 1536 return self._narrowmatch
1537 1537
1538 1538 def setnarrowpats(self, newincludes, newexcludes):
1539 1539 narrowspec.save(self, newincludes, newexcludes)
1540 1540 self.invalidate(clearfilecache=True)
1541 1541
1542 1542 @unfilteredpropertycache
1543 1543 def _quick_access_changeid_null(self):
1544 1544 return {
1545 1545 b'null': (nullrev, nullid),
1546 1546 nullrev: (nullrev, nullid),
1547 1547 nullid: (nullrev, nullid),
1548 1548 }
1549 1549
1550 1550 @unfilteredpropertycache
1551 1551 def _quick_access_changeid_wc(self):
1552 1552 # also fast path access to the working copy parents
1553 1553 # however, only do it for filter that ensure wc is visible.
1554 1554 quick = {}
1555 1555 cl = self.unfiltered().changelog
1556 1556 for node in self.dirstate.parents():
1557 1557 if node == nullid:
1558 1558 continue
1559 1559 rev = cl.index.get_rev(node)
1560 1560 if rev is None:
1561 1561 # unknown working copy parent case:
1562 1562 #
1563 1563 # skip the fast path and let higher code deal with it
1564 1564 continue
1565 1565 pair = (rev, node)
1566 1566 quick[rev] = pair
1567 1567 quick[node] = pair
1568 1568 # also add the parents of the parents
1569 1569 for r in cl.parentrevs(rev):
1570 1570 if r == nullrev:
1571 1571 continue
1572 1572 n = cl.node(r)
1573 1573 pair = (r, n)
1574 1574 quick[r] = pair
1575 1575 quick[n] = pair
1576 1576 p1node = self.dirstate.p1()
1577 1577 if p1node != nullid:
1578 1578 quick[b'.'] = quick[p1node]
1579 1579 return quick
1580 1580
1581 1581 @unfilteredmethod
1582 1582 def _quick_access_changeid_invalidate(self):
1583 1583 if '_quick_access_changeid_wc' in vars(self):
1584 1584 del self.__dict__['_quick_access_changeid_wc']
1585 1585
1586 1586 @property
1587 1587 def _quick_access_changeid(self):
1588 1588 """an helper dictionnary for __getitem__ calls
1589 1589
1590 1590 This contains a list of symbol we can recognise right away without
1591 1591 further processing.
1592 1592 """
1593 1593 mapping = self._quick_access_changeid_null
1594 1594 if self.filtername in repoview.filter_has_wc:
1595 1595 mapping = mapping.copy()
1596 1596 mapping.update(self._quick_access_changeid_wc)
1597 1597 return mapping
1598 1598
1599 1599 def __getitem__(self, changeid):
1600 1600 # dealing with special cases
1601 1601 if changeid is None:
1602 1602 return context.workingctx(self)
1603 1603 if isinstance(changeid, context.basectx):
1604 1604 return changeid
1605 1605
1606 1606 # dealing with multiple revisions
1607 1607 if isinstance(changeid, slice):
1608 1608 # wdirrev isn't contiguous so the slice shouldn't include it
1609 1609 return [
1610 1610 self[i]
1611 1611 for i in pycompat.xrange(*changeid.indices(len(self)))
1612 1612 if i not in self.changelog.filteredrevs
1613 1613 ]
1614 1614
1615 1615 # dealing with some special values
1616 1616 quick_access = self._quick_access_changeid.get(changeid)
1617 1617 if quick_access is not None:
1618 1618 rev, node = quick_access
1619 1619 return context.changectx(self, rev, node, maybe_filtered=False)
1620 1620 if changeid == b'tip':
1621 1621 node = self.changelog.tip()
1622 1622 rev = self.changelog.rev(node)
1623 1623 return context.changectx(self, rev, node)
1624 1624
1625 1625 # dealing with arbitrary values
1626 1626 try:
1627 1627 if isinstance(changeid, int):
1628 1628 node = self.changelog.node(changeid)
1629 1629 rev = changeid
1630 1630 elif changeid == b'.':
1631 1631 # this is a hack to delay/avoid loading obsmarkers
1632 1632 # when we know that '.' won't be hidden
1633 1633 node = self.dirstate.p1()
1634 1634 rev = self.unfiltered().changelog.rev(node)
1635 1635 elif len(changeid) == 20:
1636 1636 try:
1637 1637 node = changeid
1638 1638 rev = self.changelog.rev(changeid)
1639 1639 except error.FilteredLookupError:
1640 1640 changeid = hex(changeid) # for the error message
1641 1641 raise
1642 1642 except LookupError:
1643 1643 # check if it might have come from damaged dirstate
1644 1644 #
1645 1645 # XXX we could avoid the unfiltered if we had a recognizable
1646 1646 # exception for filtered changeset access
1647 1647 if (
1648 1648 self.local()
1649 1649 and changeid in self.unfiltered().dirstate.parents()
1650 1650 ):
1651 1651 msg = _(b"working directory has unknown parent '%s'!")
1652 1652 raise error.Abort(msg % short(changeid))
1653 1653 changeid = hex(changeid) # for the error message
1654 1654 raise
1655 1655
1656 1656 elif len(changeid) == 40:
1657 1657 node = bin(changeid)
1658 1658 rev = self.changelog.rev(node)
1659 1659 else:
1660 1660 raise error.ProgrammingError(
1661 1661 b"unsupported changeid '%s' of type %s"
1662 1662 % (changeid, pycompat.bytestr(type(changeid)))
1663 1663 )
1664 1664
1665 1665 return context.changectx(self, rev, node)
1666 1666
1667 1667 except (error.FilteredIndexError, error.FilteredLookupError):
1668 1668 raise error.FilteredRepoLookupError(
1669 1669 _(b"filtered revision '%s'") % pycompat.bytestr(changeid)
1670 1670 )
1671 1671 except (IndexError, LookupError):
1672 1672 raise error.RepoLookupError(
1673 1673 _(b"unknown revision '%s'") % pycompat.bytestr(changeid)
1674 1674 )
1675 1675 except error.WdirUnsupported:
1676 1676 return context.workingctx(self)
1677 1677
1678 1678 def __contains__(self, changeid):
1679 1679 """True if the given changeid exists
1680 1680
1681 1681 error.AmbiguousPrefixLookupError is raised if an ambiguous node
1682 1682 specified.
1683 1683 """
1684 1684 try:
1685 1685 self[changeid]
1686 1686 return True
1687 1687 except error.RepoLookupError:
1688 1688 return False
1689 1689
1690 1690 def __nonzero__(self):
1691 1691 return True
1692 1692
1693 1693 __bool__ = __nonzero__
1694 1694
1695 1695 def __len__(self):
1696 1696 # no need to pay the cost of repoview.changelog
1697 1697 unfi = self.unfiltered()
1698 1698 return len(unfi.changelog)
1699 1699
1700 1700 def __iter__(self):
1701 1701 return iter(self.changelog)
1702 1702
1703 1703 def revs(self, expr, *args):
1704 1704 '''Find revisions matching a revset.
1705 1705
1706 1706 The revset is specified as a string ``expr`` that may contain
1707 1707 %-formatting to escape certain types. See ``revsetlang.formatspec``.
1708 1708
1709 1709 Revset aliases from the configuration are not expanded. To expand
1710 1710 user aliases, consider calling ``scmutil.revrange()`` or
1711 1711 ``repo.anyrevs([expr], user=True)``.
1712 1712
1713 1713 Returns a smartset.abstractsmartset, which is a list-like interface
1714 1714 that contains integer revisions.
1715 1715 '''
1716 1716 tree = revsetlang.spectree(expr, *args)
1717 1717 return revset.makematcher(tree)(self)
1718 1718
1719 1719 def set(self, expr, *args):
1720 1720 '''Find revisions matching a revset and emit changectx instances.
1721 1721
1722 1722 This is a convenience wrapper around ``revs()`` that iterates the
1723 1723 result and is a generator of changectx instances.
1724 1724
1725 1725 Revset aliases from the configuration are not expanded. To expand
1726 1726 user aliases, consider calling ``scmutil.revrange()``.
1727 1727 '''
1728 1728 for r in self.revs(expr, *args):
1729 1729 yield self[r]
1730 1730
1731 1731 def anyrevs(self, specs, user=False, localalias=None):
1732 1732 '''Find revisions matching one of the given revsets.
1733 1733
1734 1734 Revset aliases from the configuration are not expanded by default. To
1735 1735 expand user aliases, specify ``user=True``. To provide some local
1736 1736 definitions overriding user aliases, set ``localalias`` to
1737 1737 ``{name: definitionstring}``.
1738 1738 '''
1739 1739 if specs == [b'null']:
1740 1740 return revset.baseset([nullrev])
1741 1741 if specs == [b'.']:
1742 1742 quick_data = self._quick_access_changeid.get(b'.')
1743 1743 if quick_data is not None:
1744 1744 return revset.baseset([quick_data[0]])
1745 1745 if user:
1746 1746 m = revset.matchany(
1747 1747 self.ui,
1748 1748 specs,
1749 1749 lookup=revset.lookupfn(self),
1750 1750 localalias=localalias,
1751 1751 )
1752 1752 else:
1753 1753 m = revset.matchany(None, specs, localalias=localalias)
1754 1754 return m(self)
1755 1755
1756 1756 def url(self):
1757 1757 return b'file:' + self.root
1758 1758
1759 1759 def hook(self, name, throw=False, **args):
1760 1760 """Call a hook, passing this repo instance.
1761 1761
1762 1762 This a convenience method to aid invoking hooks. Extensions likely
1763 1763 won't call this unless they have registered a custom hook or are
1764 1764 replacing code that is expected to call a hook.
1765 1765 """
1766 1766 return hook.hook(self.ui, self, name, throw, **args)
1767 1767
1768 1768 @filteredpropertycache
1769 1769 def _tagscache(self):
1770 1770 '''Returns a tagscache object that contains various tags related
1771 1771 caches.'''
1772 1772
1773 1773 # This simplifies its cache management by having one decorated
1774 1774 # function (this one) and the rest simply fetch things from it.
1775 1775 class tagscache(object):
1776 1776 def __init__(self):
1777 1777 # These two define the set of tags for this repository. tags
1778 1778 # maps tag name to node; tagtypes maps tag name to 'global' or
1779 1779 # 'local'. (Global tags are defined by .hgtags across all
1780 1780 # heads, and local tags are defined in .hg/localtags.)
1781 1781 # They constitute the in-memory cache of tags.
1782 1782 self.tags = self.tagtypes = None
1783 1783
1784 1784 self.nodetagscache = self.tagslist = None
1785 1785
1786 1786 cache = tagscache()
1787 1787 cache.tags, cache.tagtypes = self._findtags()
1788 1788
1789 1789 return cache
1790 1790
1791 1791 def tags(self):
1792 1792 '''return a mapping of tag to node'''
1793 1793 t = {}
1794 1794 if self.changelog.filteredrevs:
1795 1795 tags, tt = self._findtags()
1796 1796 else:
1797 1797 tags = self._tagscache.tags
1798 1798 rev = self.changelog.rev
1799 1799 for k, v in pycompat.iteritems(tags):
1800 1800 try:
1801 1801 # ignore tags to unknown nodes
1802 1802 rev(v)
1803 1803 t[k] = v
1804 1804 except (error.LookupError, ValueError):
1805 1805 pass
1806 1806 return t
1807 1807
1808 1808 def _findtags(self):
1809 1809 '''Do the hard work of finding tags. Return a pair of dicts
1810 1810 (tags, tagtypes) where tags maps tag name to node, and tagtypes
1811 1811 maps tag name to a string like \'global\' or \'local\'.
1812 1812 Subclasses or extensions are free to add their own tags, but
1813 1813 should be aware that the returned dicts will be retained for the
1814 1814 duration of the localrepo object.'''
1815 1815
1816 1816 # XXX what tagtype should subclasses/extensions use? Currently
1817 1817 # mq and bookmarks add tags, but do not set the tagtype at all.
1818 1818 # Should each extension invent its own tag type? Should there
1819 1819 # be one tagtype for all such "virtual" tags? Or is the status
1820 1820 # quo fine?
1821 1821
1822 1822 # map tag name to (node, hist)
1823 1823 alltags = tagsmod.findglobaltags(self.ui, self)
1824 1824 # map tag name to tag type
1825 1825 tagtypes = {tag: b'global' for tag in alltags}
1826 1826
1827 1827 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
1828 1828
1829 1829 # Build the return dicts. Have to re-encode tag names because
1830 1830 # the tags module always uses UTF-8 (in order not to lose info
1831 1831 # writing to the cache), but the rest of Mercurial wants them in
1832 1832 # local encoding.
1833 1833 tags = {}
1834 1834 for (name, (node, hist)) in pycompat.iteritems(alltags):
1835 1835 if node != nullid:
1836 1836 tags[encoding.tolocal(name)] = node
1837 1837 tags[b'tip'] = self.changelog.tip()
1838 1838 tagtypes = {
1839 1839 encoding.tolocal(name): value
1840 1840 for (name, value) in pycompat.iteritems(tagtypes)
1841 1841 }
1842 1842 return (tags, tagtypes)
1843 1843
1844 1844 def tagtype(self, tagname):
1845 1845 '''
1846 1846 return the type of the given tag. result can be:
1847 1847
1848 1848 'local' : a local tag
1849 1849 'global' : a global tag
1850 1850 None : tag does not exist
1851 1851 '''
1852 1852
1853 1853 return self._tagscache.tagtypes.get(tagname)
1854 1854
1855 1855 def tagslist(self):
1856 1856 '''return a list of tags ordered by revision'''
1857 1857 if not self._tagscache.tagslist:
1858 1858 l = []
1859 1859 for t, n in pycompat.iteritems(self.tags()):
1860 1860 l.append((self.changelog.rev(n), t, n))
1861 1861 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
1862 1862
1863 1863 return self._tagscache.tagslist
1864 1864
1865 1865 def nodetags(self, node):
1866 1866 '''return the tags associated with a node'''
1867 1867 if not self._tagscache.nodetagscache:
1868 1868 nodetagscache = {}
1869 1869 for t, n in pycompat.iteritems(self._tagscache.tags):
1870 1870 nodetagscache.setdefault(n, []).append(t)
1871 1871 for tags in pycompat.itervalues(nodetagscache):
1872 1872 tags.sort()
1873 1873 self._tagscache.nodetagscache = nodetagscache
1874 1874 return self._tagscache.nodetagscache.get(node, [])
1875 1875
1876 1876 def nodebookmarks(self, node):
1877 1877 """return the list of bookmarks pointing to the specified node"""
1878 1878 return self._bookmarks.names(node)
1879 1879
1880 1880 def branchmap(self):
1881 1881 '''returns a dictionary {branch: [branchheads]} with branchheads
1882 1882 ordered by increasing revision number'''
1883 1883 return self._branchcaches[self]
1884 1884
1885 1885 @unfilteredmethod
1886 1886 def revbranchcache(self):
1887 1887 if not self._revbranchcache:
1888 1888 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
1889 1889 return self._revbranchcache
1890 1890
1891 1891 def branchtip(self, branch, ignoremissing=False):
1892 1892 '''return the tip node for a given branch
1893 1893
1894 1894 If ignoremissing is True, then this method will not raise an error.
1895 1895 This is helpful for callers that only expect None for a missing branch
1896 1896 (e.g. namespace).
1897 1897
1898 1898 '''
1899 1899 try:
1900 1900 return self.branchmap().branchtip(branch)
1901 1901 except KeyError:
1902 1902 if not ignoremissing:
1903 1903 raise error.RepoLookupError(_(b"unknown branch '%s'") % branch)
1904 1904 else:
1905 1905 pass
1906 1906
1907 1907 def lookup(self, key):
1908 1908 node = scmutil.revsymbol(self, key).node()
1909 1909 if node is None:
1910 1910 raise error.RepoLookupError(_(b"unknown revision '%s'") % key)
1911 1911 return node
1912 1912
1913 1913 def lookupbranch(self, key):
1914 1914 if self.branchmap().hasbranch(key):
1915 1915 return key
1916 1916
1917 1917 return scmutil.revsymbol(self, key).branch()
1918 1918
1919 1919 def known(self, nodes):
1920 1920 cl = self.changelog
1921 1921 get_rev = cl.index.get_rev
1922 1922 filtered = cl.filteredrevs
1923 1923 result = []
1924 1924 for n in nodes:
1925 1925 r = get_rev(n)
1926 1926 resp = not (r is None or r in filtered)
1927 1927 result.append(resp)
1928 1928 return result
1929 1929
1930 1930 def local(self):
1931 1931 return self
1932 1932
1933 1933 def publishing(self):
1934 1934 # it's safe (and desirable) to trust the publish flag unconditionally
1935 1935 # so that we don't finalize changes shared between users via ssh or nfs
1936 1936 return self.ui.configbool(b'phases', b'publish', untrusted=True)
1937 1937
1938 1938 def cancopy(self):
1939 1939 # so statichttprepo's override of local() works
1940 1940 if not self.local():
1941 1941 return False
1942 1942 if not self.publishing():
1943 1943 return True
1944 1944 # if publishing we can't copy if there is filtered content
1945 1945 return not self.filtered(b'visible').changelog.filteredrevs
1946 1946
1947 1947 def shared(self):
1948 1948 '''the type of shared repository (None if not shared)'''
1949 1949 if self.sharedpath != self.path:
1950 1950 return b'store'
1951 1951 return None
1952 1952
1953 1953 def wjoin(self, f, *insidef):
1954 1954 return self.vfs.reljoin(self.root, f, *insidef)
1955 1955
1956 1956 def setparents(self, p1, p2=nullid):
1957 1957 self[None].setparents(p1, p2)
1958 1958 self._quick_access_changeid_invalidate()
1959 1959
1960 1960 def filectx(self, path, changeid=None, fileid=None, changectx=None):
1961 1961 """changeid must be a changeset revision, if specified.
1962 1962 fileid can be a file revision or node."""
1963 1963 return context.filectx(
1964 1964 self, path, changeid, fileid, changectx=changectx
1965 1965 )
1966 1966
1967 1967 def getcwd(self):
1968 1968 return self.dirstate.getcwd()
1969 1969
1970 1970 def pathto(self, f, cwd=None):
1971 1971 return self.dirstate.pathto(f, cwd)
1972 1972
1973 1973 def _loadfilter(self, filter):
1974 1974 if filter not in self._filterpats:
1975 1975 l = []
1976 1976 for pat, cmd in self.ui.configitems(filter):
1977 1977 if cmd == b'!':
1978 1978 continue
1979 1979 mf = matchmod.match(self.root, b'', [pat])
1980 1980 fn = None
1981 1981 params = cmd
1982 1982 for name, filterfn in pycompat.iteritems(self._datafilters):
1983 1983 if cmd.startswith(name):
1984 1984 fn = filterfn
1985 1985 params = cmd[len(name) :].lstrip()
1986 1986 break
1987 1987 if not fn:
1988 1988 fn = lambda s, c, **kwargs: procutil.filter(s, c)
1989 1989 fn.__name__ = 'commandfilter'
1990 1990 # Wrap old filters not supporting keyword arguments
1991 1991 if not pycompat.getargspec(fn)[2]:
1992 1992 oldfn = fn
1993 1993 fn = lambda s, c, oldfn=oldfn, **kwargs: oldfn(s, c)
1994 1994 fn.__name__ = 'compat-' + oldfn.__name__
1995 1995 l.append((mf, fn, params))
1996 1996 self._filterpats[filter] = l
1997 1997 return self._filterpats[filter]
1998 1998
1999 1999 def _filter(self, filterpats, filename, data):
2000 2000 for mf, fn, cmd in filterpats:
2001 2001 if mf(filename):
2002 2002 self.ui.debug(
2003 2003 b"filtering %s through %s\n"
2004 2004 % (filename, cmd or pycompat.sysbytes(fn.__name__))
2005 2005 )
2006 2006 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
2007 2007 break
2008 2008
2009 2009 return data
2010 2010
2011 2011 @unfilteredpropertycache
2012 2012 def _encodefilterpats(self):
2013 2013 return self._loadfilter(b'encode')
2014 2014
2015 2015 @unfilteredpropertycache
2016 2016 def _decodefilterpats(self):
2017 2017 return self._loadfilter(b'decode')
2018 2018
2019 2019 def adddatafilter(self, name, filter):
2020 2020 self._datafilters[name] = filter
2021 2021
2022 2022 def wread(self, filename):
2023 2023 if self.wvfs.islink(filename):
2024 2024 data = self.wvfs.readlink(filename)
2025 2025 else:
2026 2026 data = self.wvfs.read(filename)
2027 2027 return self._filter(self._encodefilterpats, filename, data)
2028 2028
2029 2029 def wwrite(self, filename, data, flags, backgroundclose=False, **kwargs):
2030 2030 """write ``data`` into ``filename`` in the working directory
2031 2031
2032 2032 This returns length of written (maybe decoded) data.
2033 2033 """
2034 2034 data = self._filter(self._decodefilterpats, filename, data)
2035 2035 if b'l' in flags:
2036 2036 self.wvfs.symlink(data, filename)
2037 2037 else:
2038 2038 self.wvfs.write(
2039 2039 filename, data, backgroundclose=backgroundclose, **kwargs
2040 2040 )
2041 2041 if b'x' in flags:
2042 2042 self.wvfs.setflags(filename, False, True)
2043 2043 else:
2044 2044 self.wvfs.setflags(filename, False, False)
2045 2045 return len(data)
2046 2046
2047 2047 def wwritedata(self, filename, data):
2048 2048 return self._filter(self._decodefilterpats, filename, data)
2049 2049
2050 2050 def currenttransaction(self):
2051 2051 """return the current transaction or None if non exists"""
2052 2052 if self._transref:
2053 2053 tr = self._transref()
2054 2054 else:
2055 2055 tr = None
2056 2056
2057 2057 if tr and tr.running():
2058 2058 return tr
2059 2059 return None
2060 2060
2061 2061 def transaction(self, desc, report=None):
2062 2062 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
2063 2063 b'devel', b'check-locks'
2064 2064 ):
2065 2065 if self._currentlock(self._lockref) is None:
2066 2066 raise error.ProgrammingError(b'transaction requires locking')
2067 2067 tr = self.currenttransaction()
2068 2068 if tr is not None:
2069 2069 return tr.nest(name=desc)
2070 2070
2071 2071 # abort here if the journal already exists
2072 2072 if self.svfs.exists(b"journal"):
2073 2073 raise error.RepoError(
2074 2074 _(b"abandoned transaction found"),
2075 2075 hint=_(b"run 'hg recover' to clean up transaction"),
2076 2076 )
2077 2077
2078 2078 idbase = b"%.40f#%f" % (random.random(), time.time())
2079 2079 ha = hex(hashutil.sha1(idbase).digest())
2080 2080 txnid = b'TXN:' + ha
2081 2081 self.hook(b'pretxnopen', throw=True, txnname=desc, txnid=txnid)
2082 2082
2083 2083 self._writejournal(desc)
2084 2084 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
2085 2085 if report:
2086 2086 rp = report
2087 2087 else:
2088 2088 rp = self.ui.warn
2089 2089 vfsmap = {b'plain': self.vfs, b'store': self.svfs} # root of .hg/
2090 2090 # we must avoid cyclic reference between repo and transaction.
2091 2091 reporef = weakref.ref(self)
2092 2092 # Code to track tag movement
2093 2093 #
2094 2094 # Since tags are all handled as file content, it is actually quite hard
2095 2095 # to track these movement from a code perspective. So we fallback to a
2096 2096 # tracking at the repository level. One could envision to track changes
2097 2097 # to the '.hgtags' file through changegroup apply but that fails to
2098 2098 # cope with case where transaction expose new heads without changegroup
2099 2099 # being involved (eg: phase movement).
2100 2100 #
2101 2101 # For now, We gate the feature behind a flag since this likely comes
2102 2102 # with performance impacts. The current code run more often than needed
2103 2103 # and do not use caches as much as it could. The current focus is on
2104 2104 # the behavior of the feature so we disable it by default. The flag
2105 2105 # will be removed when we are happy with the performance impact.
2106 2106 #
2107 2107 # Once this feature is no longer experimental move the following
2108 2108 # documentation to the appropriate help section:
2109 2109 #
2110 2110 # The ``HG_TAG_MOVED`` variable will be set if the transaction touched
2111 2111 # tags (new or changed or deleted tags). In addition the details of
2112 2112 # these changes are made available in a file at:
2113 2113 # ``REPOROOT/.hg/changes/tags.changes``.
2114 2114 # Make sure you check for HG_TAG_MOVED before reading that file as it
2115 2115 # might exist from a previous transaction even if no tag were touched
2116 2116 # in this one. Changes are recorded in a line base format::
2117 2117 #
2118 2118 # <action> <hex-node> <tag-name>\n
2119 2119 #
2120 2120 # Actions are defined as follow:
2121 2121 # "-R": tag is removed,
2122 2122 # "+A": tag is added,
2123 2123 # "-M": tag is moved (old value),
2124 2124 # "+M": tag is moved (new value),
2125 2125 tracktags = lambda x: None
2126 2126 # experimental config: experimental.hook-track-tags
2127 2127 shouldtracktags = self.ui.configbool(
2128 2128 b'experimental', b'hook-track-tags'
2129 2129 )
2130 2130 if desc != b'strip' and shouldtracktags:
2131 2131 oldheads = self.changelog.headrevs()
2132 2132
2133 2133 def tracktags(tr2):
2134 2134 repo = reporef()
2135 2135 oldfnodes = tagsmod.fnoderevs(repo.ui, repo, oldheads)
2136 2136 newheads = repo.changelog.headrevs()
2137 2137 newfnodes = tagsmod.fnoderevs(repo.ui, repo, newheads)
2138 2138 # notes: we compare lists here.
2139 2139 # As we do it only once buiding set would not be cheaper
2140 2140 changes = tagsmod.difftags(repo.ui, repo, oldfnodes, newfnodes)
2141 2141 if changes:
2142 2142 tr2.hookargs[b'tag_moved'] = b'1'
2143 2143 with repo.vfs(
2144 2144 b'changes/tags.changes', b'w', atomictemp=True
2145 2145 ) as changesfile:
2146 2146 # note: we do not register the file to the transaction
2147 2147 # because we needs it to still exist on the transaction
2148 2148 # is close (for txnclose hooks)
2149 2149 tagsmod.writediff(changesfile, changes)
2150 2150
2151 2151 def validate(tr2):
2152 2152 """will run pre-closing hooks"""
2153 2153 # XXX the transaction API is a bit lacking here so we take a hacky
2154 2154 # path for now
2155 2155 #
2156 2156 # We cannot add this as a "pending" hooks since the 'tr.hookargs'
2157 2157 # dict is copied before these run. In addition we needs the data
2158 2158 # available to in memory hooks too.
2159 2159 #
2160 2160 # Moreover, we also need to make sure this runs before txnclose
2161 2161 # hooks and there is no "pending" mechanism that would execute
2162 2162 # logic only if hooks are about to run.
2163 2163 #
2164 2164 # Fixing this limitation of the transaction is also needed to track
2165 2165 # other families of changes (bookmarks, phases, obsolescence).
2166 2166 #
2167 2167 # This will have to be fixed before we remove the experimental
2168 2168 # gating.
2169 2169 tracktags(tr2)
2170 2170 repo = reporef()
2171 2171
2172 2172 singleheadopt = (b'experimental', b'single-head-per-branch')
2173 2173 singlehead = repo.ui.configbool(*singleheadopt)
2174 2174 if singlehead:
2175 2175 singleheadsub = repo.ui.configsuboptions(*singleheadopt)[1]
2176 2176 accountclosed = singleheadsub.get(
2177 2177 b"account-closed-heads", False
2178 2178 )
2179 2179 scmutil.enforcesinglehead(repo, tr2, desc, accountclosed)
2180 2180 if hook.hashook(repo.ui, b'pretxnclose-bookmark'):
2181 2181 for name, (old, new) in sorted(
2182 2182 tr.changes[b'bookmarks'].items()
2183 2183 ):
2184 2184 args = tr.hookargs.copy()
2185 2185 args.update(bookmarks.preparehookargs(name, old, new))
2186 2186 repo.hook(
2187 2187 b'pretxnclose-bookmark',
2188 2188 throw=True,
2189 2189 **pycompat.strkwargs(args)
2190 2190 )
2191 2191 if hook.hashook(repo.ui, b'pretxnclose-phase'):
2192 2192 cl = repo.unfiltered().changelog
2193 2193 for revs, (old, new) in tr.changes[b'phases']:
2194 2194 for rev in revs:
2195 2195 args = tr.hookargs.copy()
2196 2196 node = hex(cl.node(rev))
2197 2197 args.update(phases.preparehookargs(node, old, new))
2198 2198 repo.hook(
2199 2199 b'pretxnclose-phase',
2200 2200 throw=True,
2201 2201 **pycompat.strkwargs(args)
2202 2202 )
2203 2203
2204 2204 repo.hook(
2205 2205 b'pretxnclose', throw=True, **pycompat.strkwargs(tr.hookargs)
2206 2206 )
2207 2207
2208 2208 def releasefn(tr, success):
2209 2209 repo = reporef()
2210 2210 if repo is None:
2211 2211 # If the repo has been GC'd (and this release function is being
2212 2212 # called from transaction.__del__), there's not much we can do,
2213 2213 # so just leave the unfinished transaction there and let the
2214 2214 # user run `hg recover`.
2215 2215 return
2216 2216 if success:
2217 2217 # this should be explicitly invoked here, because
2218 2218 # in-memory changes aren't written out at closing
2219 2219 # transaction, if tr.addfilegenerator (via
2220 2220 # dirstate.write or so) isn't invoked while
2221 2221 # transaction running
2222 2222 repo.dirstate.write(None)
2223 2223 else:
2224 2224 # discard all changes (including ones already written
2225 2225 # out) in this transaction
2226 2226 narrowspec.restorebackup(self, b'journal.narrowspec')
2227 2227 narrowspec.restorewcbackup(self, b'journal.narrowspec.dirstate')
2228 2228 repo.dirstate.restorebackup(None, b'journal.dirstate')
2229 2229
2230 2230 repo.invalidate(clearfilecache=True)
2231 2231
2232 2232 tr = transaction.transaction(
2233 2233 rp,
2234 2234 self.svfs,
2235 2235 vfsmap,
2236 2236 b"journal",
2237 2237 b"undo",
2238 2238 aftertrans(renames),
2239 2239 self.store.createmode,
2240 2240 validator=validate,
2241 2241 releasefn=releasefn,
2242 2242 checkambigfiles=_cachedfiles,
2243 2243 name=desc,
2244 2244 )
2245 2245 tr.changes[b'origrepolen'] = len(self)
2246 2246 tr.changes[b'obsmarkers'] = set()
2247 2247 tr.changes[b'phases'] = []
2248 2248 tr.changes[b'bookmarks'] = {}
2249 2249
2250 2250 tr.hookargs[b'txnid'] = txnid
2251 2251 tr.hookargs[b'txnname'] = desc
2252 2252 tr.hookargs[b'changes'] = tr.changes
2253 2253 # note: writing the fncache only during finalize mean that the file is
2254 2254 # outdated when running hooks. As fncache is used for streaming clone,
2255 2255 # this is not expected to break anything that happen during the hooks.
2256 2256 tr.addfinalize(b'flush-fncache', self.store.write)
2257 2257
2258 2258 def txnclosehook(tr2):
2259 2259 """To be run if transaction is successful, will schedule a hook run
2260 2260 """
2261 2261 # Don't reference tr2 in hook() so we don't hold a reference.
2262 2262 # This reduces memory consumption when there are multiple
2263 2263 # transactions per lock. This can likely go away if issue5045
2264 2264 # fixes the function accumulation.
2265 2265 hookargs = tr2.hookargs
2266 2266
2267 2267 def hookfunc(unused_success):
2268 2268 repo = reporef()
2269 2269 if hook.hashook(repo.ui, b'txnclose-bookmark'):
2270 2270 bmchanges = sorted(tr.changes[b'bookmarks'].items())
2271 2271 for name, (old, new) in bmchanges:
2272 2272 args = tr.hookargs.copy()
2273 2273 args.update(bookmarks.preparehookargs(name, old, new))
2274 2274 repo.hook(
2275 2275 b'txnclose-bookmark',
2276 2276 throw=False,
2277 2277 **pycompat.strkwargs(args)
2278 2278 )
2279 2279
2280 2280 if hook.hashook(repo.ui, b'txnclose-phase'):
2281 2281 cl = repo.unfiltered().changelog
2282 2282 phasemv = sorted(
2283 2283 tr.changes[b'phases'], key=lambda r: r[0][0]
2284 2284 )
2285 2285 for revs, (old, new) in phasemv:
2286 2286 for rev in revs:
2287 2287 args = tr.hookargs.copy()
2288 2288 node = hex(cl.node(rev))
2289 2289 args.update(phases.preparehookargs(node, old, new))
2290 2290 repo.hook(
2291 2291 b'txnclose-phase',
2292 2292 throw=False,
2293 2293 **pycompat.strkwargs(args)
2294 2294 )
2295 2295
2296 2296 repo.hook(
2297 2297 b'txnclose', throw=False, **pycompat.strkwargs(hookargs)
2298 2298 )
2299 2299
2300 2300 reporef()._afterlock(hookfunc)
2301 2301
2302 2302 tr.addfinalize(b'txnclose-hook', txnclosehook)
2303 2303 # Include a leading "-" to make it happen before the transaction summary
2304 2304 # reports registered via scmutil.registersummarycallback() whose names
2305 2305 # are 00-txnreport etc. That way, the caches will be warm when the
2306 2306 # callbacks run.
2307 2307 tr.addpostclose(b'-warm-cache', self._buildcacheupdater(tr))
2308 2308
2309 2309 def txnaborthook(tr2):
2310 2310 """To be run if transaction is aborted
2311 2311 """
2312 2312 reporef().hook(
2313 2313 b'txnabort', throw=False, **pycompat.strkwargs(tr2.hookargs)
2314 2314 )
2315 2315
2316 2316 tr.addabort(b'txnabort-hook', txnaborthook)
2317 2317 # avoid eager cache invalidation. in-memory data should be identical
2318 2318 # to stored data if transaction has no error.
2319 2319 tr.addpostclose(b'refresh-filecachestats', self._refreshfilecachestats)
2320 2320 self._transref = weakref.ref(tr)
2321 2321 scmutil.registersummarycallback(self, tr, desc)
2322 2322 return tr
2323 2323
2324 2324 def _journalfiles(self):
2325 2325 return (
2326 2326 (self.svfs, b'journal'),
2327 2327 (self.svfs, b'journal.narrowspec'),
2328 2328 (self.vfs, b'journal.narrowspec.dirstate'),
2329 2329 (self.vfs, b'journal.dirstate'),
2330 2330 (self.vfs, b'journal.branch'),
2331 2331 (self.vfs, b'journal.desc'),
2332 2332 (bookmarks.bookmarksvfs(self), b'journal.bookmarks'),
2333 2333 (self.svfs, b'journal.phaseroots'),
2334 2334 )
2335 2335
2336 2336 def undofiles(self):
2337 2337 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
2338 2338
2339 2339 @unfilteredmethod
2340 2340 def _writejournal(self, desc):
2341 2341 self.dirstate.savebackup(None, b'journal.dirstate')
2342 2342 narrowspec.savewcbackup(self, b'journal.narrowspec.dirstate')
2343 2343 narrowspec.savebackup(self, b'journal.narrowspec')
2344 2344 self.vfs.write(
2345 2345 b"journal.branch", encoding.fromlocal(self.dirstate.branch())
2346 2346 )
2347 2347 self.vfs.write(b"journal.desc", b"%d\n%s\n" % (len(self), desc))
2348 2348 bookmarksvfs = bookmarks.bookmarksvfs(self)
2349 2349 bookmarksvfs.write(
2350 2350 b"journal.bookmarks", bookmarksvfs.tryread(b"bookmarks")
2351 2351 )
2352 2352 self.svfs.write(b"journal.phaseroots", self.svfs.tryread(b"phaseroots"))
2353 2353
2354 2354 def recover(self):
2355 2355 with self.lock():
2356 2356 if self.svfs.exists(b"journal"):
2357 2357 self.ui.status(_(b"rolling back interrupted transaction\n"))
2358 2358 vfsmap = {
2359 2359 b'': self.svfs,
2360 2360 b'plain': self.vfs,
2361 2361 }
2362 2362 transaction.rollback(
2363 2363 self.svfs,
2364 2364 vfsmap,
2365 2365 b"journal",
2366 2366 self.ui.warn,
2367 2367 checkambigfiles=_cachedfiles,
2368 2368 )
2369 2369 self.invalidate()
2370 2370 return True
2371 2371 else:
2372 2372 self.ui.warn(_(b"no interrupted transaction available\n"))
2373 2373 return False
2374 2374
2375 2375 def rollback(self, dryrun=False, force=False):
2376 2376 wlock = lock = dsguard = None
2377 2377 try:
2378 2378 wlock = self.wlock()
2379 2379 lock = self.lock()
2380 2380 if self.svfs.exists(b"undo"):
2381 2381 dsguard = dirstateguard.dirstateguard(self, b'rollback')
2382 2382
2383 2383 return self._rollback(dryrun, force, dsguard)
2384 2384 else:
2385 2385 self.ui.warn(_(b"no rollback information available\n"))
2386 2386 return 1
2387 2387 finally:
2388 2388 release(dsguard, lock, wlock)
2389 2389
2390 2390 @unfilteredmethod # Until we get smarter cache management
2391 2391 def _rollback(self, dryrun, force, dsguard):
2392 2392 ui = self.ui
2393 2393 try:
2394 2394 args = self.vfs.read(b'undo.desc').splitlines()
2395 2395 (oldlen, desc, detail) = (int(args[0]), args[1], None)
2396 2396 if len(args) >= 3:
2397 2397 detail = args[2]
2398 2398 oldtip = oldlen - 1
2399 2399
2400 2400 if detail and ui.verbose:
2401 2401 msg = _(
2402 2402 b'repository tip rolled back to revision %d'
2403 2403 b' (undo %s: %s)\n'
2404 2404 ) % (oldtip, desc, detail)
2405 2405 else:
2406 2406 msg = _(
2407 2407 b'repository tip rolled back to revision %d (undo %s)\n'
2408 2408 ) % (oldtip, desc)
2409 2409 except IOError:
2410 2410 msg = _(b'rolling back unknown transaction\n')
2411 2411 desc = None
2412 2412
2413 2413 if not force and self[b'.'] != self[b'tip'] and desc == b'commit':
2414 2414 raise error.Abort(
2415 2415 _(
2416 2416 b'rollback of last commit while not checked out '
2417 2417 b'may lose data'
2418 2418 ),
2419 2419 hint=_(b'use -f to force'),
2420 2420 )
2421 2421
2422 2422 ui.status(msg)
2423 2423 if dryrun:
2424 2424 return 0
2425 2425
2426 2426 parents = self.dirstate.parents()
2427 2427 self.destroying()
2428 2428 vfsmap = {b'plain': self.vfs, b'': self.svfs}
2429 2429 transaction.rollback(
2430 2430 self.svfs, vfsmap, b'undo', ui.warn, checkambigfiles=_cachedfiles
2431 2431 )
2432 2432 bookmarksvfs = bookmarks.bookmarksvfs(self)
2433 2433 if bookmarksvfs.exists(b'undo.bookmarks'):
2434 2434 bookmarksvfs.rename(
2435 2435 b'undo.bookmarks', b'bookmarks', checkambig=True
2436 2436 )
2437 2437 if self.svfs.exists(b'undo.phaseroots'):
2438 2438 self.svfs.rename(b'undo.phaseroots', b'phaseroots', checkambig=True)
2439 2439 self.invalidate()
2440 2440
2441 2441 has_node = self.changelog.index.has_node
2442 2442 parentgone = any(not has_node(p) for p in parents)
2443 2443 if parentgone:
2444 2444 # prevent dirstateguard from overwriting already restored one
2445 2445 dsguard.close()
2446 2446
2447 2447 narrowspec.restorebackup(self, b'undo.narrowspec')
2448 2448 narrowspec.restorewcbackup(self, b'undo.narrowspec.dirstate')
2449 2449 self.dirstate.restorebackup(None, b'undo.dirstate')
2450 2450 try:
2451 2451 branch = self.vfs.read(b'undo.branch')
2452 2452 self.dirstate.setbranch(encoding.tolocal(branch))
2453 2453 except IOError:
2454 2454 ui.warn(
2455 2455 _(
2456 2456 b'named branch could not be reset: '
2457 2457 b'current branch is still \'%s\'\n'
2458 2458 )
2459 2459 % self.dirstate.branch()
2460 2460 )
2461 2461
2462 2462 parents = tuple([p.rev() for p in self[None].parents()])
2463 2463 if len(parents) > 1:
2464 2464 ui.status(
2465 2465 _(
2466 2466 b'working directory now based on '
2467 2467 b'revisions %d and %d\n'
2468 2468 )
2469 2469 % parents
2470 2470 )
2471 2471 else:
2472 2472 ui.status(
2473 2473 _(b'working directory now based on revision %d\n') % parents
2474 2474 )
2475 2475 mergestatemod.mergestate.clean(self, self[b'.'].node())
2476 2476
2477 2477 # TODO: if we know which new heads may result from this rollback, pass
2478 2478 # them to destroy(), which will prevent the branchhead cache from being
2479 2479 # invalidated.
2480 2480 self.destroyed()
2481 2481 return 0
2482 2482
2483 2483 def _buildcacheupdater(self, newtransaction):
2484 2484 """called during transaction to build the callback updating cache
2485 2485
2486 2486 Lives on the repository to help extension who might want to augment
2487 2487 this logic. For this purpose, the created transaction is passed to the
2488 2488 method.
2489 2489 """
2490 2490 # we must avoid cyclic reference between repo and transaction.
2491 2491 reporef = weakref.ref(self)
2492 2492
2493 2493 def updater(tr):
2494 2494 repo = reporef()
2495 2495 repo.updatecaches(tr)
2496 2496
2497 2497 return updater
2498 2498
2499 2499 @unfilteredmethod
2500 2500 def updatecaches(self, tr=None, full=False):
2501 2501 """warm appropriate caches
2502 2502
2503 2503 If this function is called after a transaction closed. The transaction
2504 2504 will be available in the 'tr' argument. This can be used to selectively
2505 2505 update caches relevant to the changes in that transaction.
2506 2506
2507 2507 If 'full' is set, make sure all caches the function knows about have
2508 2508 up-to-date data. Even the ones usually loaded more lazily.
2509 2509 """
2510 2510 if tr is not None and tr.hookargs.get(b'source') == b'strip':
2511 2511 # During strip, many caches are invalid but
2512 2512 # later call to `destroyed` will refresh them.
2513 2513 return
2514 2514
2515 2515 if tr is None or tr.changes[b'origrepolen'] < len(self):
2516 2516 # accessing the 'ser ved' branchmap should refresh all the others,
2517 2517 self.ui.debug(b'updating the branch cache\n')
2518 2518 self.filtered(b'served').branchmap()
2519 2519 self.filtered(b'served.hidden').branchmap()
2520 2520
2521 2521 if full:
2522 2522 unfi = self.unfiltered()
2523 2523
2524 2524 self.changelog.update_caches(transaction=tr)
2525 2525 self.manifestlog.update_caches(transaction=tr)
2526 2526
2527 2527 rbc = unfi.revbranchcache()
2528 2528 for r in unfi.changelog:
2529 2529 rbc.branchinfo(r)
2530 2530 rbc.write()
2531 2531
2532 2532 # ensure the working copy parents are in the manifestfulltextcache
2533 2533 for ctx in self[b'.'].parents():
2534 2534 ctx.manifest() # accessing the manifest is enough
2535 2535
2536 2536 # accessing fnode cache warms the cache
2537 2537 tagsmod.fnoderevs(self.ui, unfi, unfi.changelog.revs())
2538 2538 # accessing tags warm the cache
2539 2539 self.tags()
2540 2540 self.filtered(b'served').tags()
2541 2541
2542 2542 # The `full` arg is documented as updating even the lazily-loaded
2543 2543 # caches immediately, so we're forcing a write to cause these caches
2544 2544 # to be warmed up even if they haven't explicitly been requested
2545 2545 # yet (if they've never been used by hg, they won't ever have been
2546 2546 # written, even if they're a subset of another kind of cache that
2547 2547 # *has* been used).
2548 2548 for filt in repoview.filtertable.keys():
2549 2549 filtered = self.filtered(filt)
2550 2550 filtered.branchmap().write(filtered)
2551 2551
2552 2552 def invalidatecaches(self):
2553 2553
2554 2554 if '_tagscache' in vars(self):
2555 2555 # can't use delattr on proxy
2556 2556 del self.__dict__['_tagscache']
2557 2557
2558 2558 self._branchcaches.clear()
2559 2559 self.invalidatevolatilesets()
2560 2560 self._sparsesignaturecache.clear()
2561 2561
2562 2562 def invalidatevolatilesets(self):
2563 2563 self.filteredrevcache.clear()
2564 2564 obsolete.clearobscaches(self)
2565 2565 self._quick_access_changeid_invalidate()
2566 2566
2567 2567 def invalidatedirstate(self):
2568 2568 '''Invalidates the dirstate, causing the next call to dirstate
2569 2569 to check if it was modified since the last time it was read,
2570 2570 rereading it if it has.
2571 2571
2572 2572 This is different to dirstate.invalidate() that it doesn't always
2573 2573 rereads the dirstate. Use dirstate.invalidate() if you want to
2574 2574 explicitly read the dirstate again (i.e. restoring it to a previous
2575 2575 known good state).'''
2576 2576 if hasunfilteredcache(self, 'dirstate'):
2577 2577 for k in self.dirstate._filecache:
2578 2578 try:
2579 2579 delattr(self.dirstate, k)
2580 2580 except AttributeError:
2581 2581 pass
2582 2582 delattr(self.unfiltered(), 'dirstate')
2583 2583
2584 2584 def invalidate(self, clearfilecache=False):
2585 2585 '''Invalidates both store and non-store parts other than dirstate
2586 2586
2587 2587 If a transaction is running, invalidation of store is omitted,
2588 2588 because discarding in-memory changes might cause inconsistency
2589 2589 (e.g. incomplete fncache causes unintentional failure, but
2590 2590 redundant one doesn't).
2591 2591 '''
2592 2592 unfiltered = self.unfiltered() # all file caches are stored unfiltered
2593 2593 for k in list(self._filecache.keys()):
2594 2594 # dirstate is invalidated separately in invalidatedirstate()
2595 2595 if k == b'dirstate':
2596 2596 continue
2597 2597 if (
2598 2598 k == b'changelog'
2599 2599 and self.currenttransaction()
2600 2600 and self.changelog._delayed
2601 2601 ):
2602 2602 # The changelog object may store unwritten revisions. We don't
2603 2603 # want to lose them.
2604 2604 # TODO: Solve the problem instead of working around it.
2605 2605 continue
2606 2606
2607 2607 if clearfilecache:
2608 2608 del self._filecache[k]
2609 2609 try:
2610 2610 delattr(unfiltered, k)
2611 2611 except AttributeError:
2612 2612 pass
2613 2613 self.invalidatecaches()
2614 2614 if not self.currenttransaction():
2615 2615 # TODO: Changing contents of store outside transaction
2616 2616 # causes inconsistency. We should make in-memory store
2617 2617 # changes detectable, and abort if changed.
2618 2618 self.store.invalidatecaches()
2619 2619
2620 2620 def invalidateall(self):
2621 2621 '''Fully invalidates both store and non-store parts, causing the
2622 2622 subsequent operation to reread any outside changes.'''
2623 2623 # extension should hook this to invalidate its caches
2624 2624 self.invalidate()
2625 2625 self.invalidatedirstate()
2626 2626
2627 2627 @unfilteredmethod
2628 2628 def _refreshfilecachestats(self, tr):
2629 2629 """Reload stats of cached files so that they are flagged as valid"""
2630 2630 for k, ce in self._filecache.items():
2631 2631 k = pycompat.sysstr(k)
2632 2632 if k == 'dirstate' or k not in self.__dict__:
2633 2633 continue
2634 2634 ce.refresh()
2635 2635
2636 2636 def _lock(
2637 2637 self,
2638 2638 vfs,
2639 2639 lockname,
2640 2640 wait,
2641 2641 releasefn,
2642 2642 acquirefn,
2643 2643 desc,
2644 2644 inheritchecker=None,
2645 2645 parentenvvar=None,
2646 2646 ):
2647 2647 parentlock = None
2648 2648 # the contents of parentenvvar are used by the underlying lock to
2649 2649 # determine whether it can be inherited
2650 2650 if parentenvvar is not None:
2651 2651 parentlock = encoding.environ.get(parentenvvar)
2652 2652
2653 2653 timeout = 0
2654 2654 warntimeout = 0
2655 2655 if wait:
2656 2656 timeout = self.ui.configint(b"ui", b"timeout")
2657 2657 warntimeout = self.ui.configint(b"ui", b"timeout.warn")
2658 2658 # internal config: ui.signal-safe-lock
2659 2659 signalsafe = self.ui.configbool(b'ui', b'signal-safe-lock')
2660 2660
2661 2661 l = lockmod.trylock(
2662 2662 self.ui,
2663 2663 vfs,
2664 2664 lockname,
2665 2665 timeout,
2666 2666 warntimeout,
2667 2667 releasefn=releasefn,
2668 2668 acquirefn=acquirefn,
2669 2669 desc=desc,
2670 2670 inheritchecker=inheritchecker,
2671 2671 parentlock=parentlock,
2672 2672 signalsafe=signalsafe,
2673 2673 )
2674 2674 return l
2675 2675
2676 2676 def _afterlock(self, callback):
2677 2677 """add a callback to be run when the repository is fully unlocked
2678 2678
2679 2679 The callback will be executed when the outermost lock is released
2680 2680 (with wlock being higher level than 'lock')."""
2681 2681 for ref in (self._wlockref, self._lockref):
2682 2682 l = ref and ref()
2683 2683 if l and l.held:
2684 2684 l.postrelease.append(callback)
2685 2685 break
2686 2686 else: # no lock have been found.
2687 2687 callback(True)
2688 2688
2689 2689 def lock(self, wait=True):
2690 2690 '''Lock the repository store (.hg/store) and return a weak reference
2691 2691 to the lock. Use this before modifying the store (e.g. committing or
2692 2692 stripping). If you are opening a transaction, get a lock as well.)
2693 2693
2694 2694 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2695 2695 'wlock' first to avoid a dead-lock hazard.'''
2696 2696 l = self._currentlock(self._lockref)
2697 2697 if l is not None:
2698 2698 l.lock()
2699 2699 return l
2700 2700
2701 2701 l = self._lock(
2702 2702 vfs=self.svfs,
2703 2703 lockname=b"lock",
2704 2704 wait=wait,
2705 2705 releasefn=None,
2706 2706 acquirefn=self.invalidate,
2707 2707 desc=_(b'repository %s') % self.origroot,
2708 2708 )
2709 2709 self._lockref = weakref.ref(l)
2710 2710 return l
2711 2711
2712 2712 def _wlockchecktransaction(self):
2713 2713 if self.currenttransaction() is not None:
2714 2714 raise error.LockInheritanceContractViolation(
2715 2715 b'wlock cannot be inherited in the middle of a transaction'
2716 2716 )
2717 2717
2718 2718 def wlock(self, wait=True):
2719 2719 '''Lock the non-store parts of the repository (everything under
2720 2720 .hg except .hg/store) and return a weak reference to the lock.
2721 2721
2722 2722 Use this before modifying files in .hg.
2723 2723
2724 2724 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2725 2725 'wlock' first to avoid a dead-lock hazard.'''
2726 2726 l = self._wlockref and self._wlockref()
2727 2727 if l is not None and l.held:
2728 2728 l.lock()
2729 2729 return l
2730 2730
2731 2731 # We do not need to check for non-waiting lock acquisition. Such
2732 2732 # acquisition would not cause dead-lock as they would just fail.
2733 2733 if wait and (
2734 2734 self.ui.configbool(b'devel', b'all-warnings')
2735 2735 or self.ui.configbool(b'devel', b'check-locks')
2736 2736 ):
2737 2737 if self._currentlock(self._lockref) is not None:
2738 2738 self.ui.develwarn(b'"wlock" acquired after "lock"')
2739 2739
2740 2740 def unlock():
2741 2741 if self.dirstate.pendingparentchange():
2742 2742 self.dirstate.invalidate()
2743 2743 else:
2744 2744 self.dirstate.write(None)
2745 2745
2746 2746 self._filecache[b'dirstate'].refresh()
2747 2747
2748 2748 l = self._lock(
2749 2749 self.vfs,
2750 2750 b"wlock",
2751 2751 wait,
2752 2752 unlock,
2753 2753 self.invalidatedirstate,
2754 2754 _(b'working directory of %s') % self.origroot,
2755 2755 inheritchecker=self._wlockchecktransaction,
2756 2756 parentenvvar=b'HG_WLOCK_LOCKER',
2757 2757 )
2758 2758 self._wlockref = weakref.ref(l)
2759 2759 return l
2760 2760
2761 2761 def _currentlock(self, lockref):
2762 2762 """Returns the lock if it's held, or None if it's not."""
2763 2763 if lockref is None:
2764 2764 return None
2765 2765 l = lockref()
2766 2766 if l is None or not l.held:
2767 2767 return None
2768 2768 return l
2769 2769
2770 2770 def currentwlock(self):
2771 2771 """Returns the wlock if it's held, or None if it's not."""
2772 2772 return self._currentlock(self._wlockref)
2773 2773
2774 2774 def checkcommitpatterns(self, wctx, match, status, fail):
2775 2775 """check for commit arguments that aren't committable"""
2776 2776 if match.isexact() or match.prefix():
2777 2777 matched = set(status.modified + status.added + status.removed)
2778 2778
2779 2779 for f in match.files():
2780 2780 f = self.dirstate.normalize(f)
2781 2781 if f == b'.' or f in matched or f in wctx.substate:
2782 2782 continue
2783 2783 if f in status.deleted:
2784 2784 fail(f, _(b'file not found!'))
2785 2785 # Is it a directory that exists or used to exist?
2786 2786 if self.wvfs.isdir(f) or wctx.p1().hasdir(f):
2787 2787 d = f + b'/'
2788 2788 for mf in matched:
2789 2789 if mf.startswith(d):
2790 2790 break
2791 2791 else:
2792 2792 fail(f, _(b"no match under directory!"))
2793 2793 elif f not in self.dirstate:
2794 2794 fail(f, _(b"file not tracked!"))
2795 2795
2796 2796 @unfilteredmethod
2797 2797 def commit(
2798 2798 self,
2799 2799 text=b"",
2800 2800 user=None,
2801 2801 date=None,
2802 2802 match=None,
2803 2803 force=False,
2804 2804 editor=None,
2805 2805 extra=None,
2806 2806 ):
2807 2807 """Add a new revision to current repository.
2808 2808
2809 2809 Revision information is gathered from the working directory,
2810 2810 match can be used to filter the committed files. If editor is
2811 2811 supplied, it is called to get a commit message.
2812 2812 """
2813 2813 if extra is None:
2814 2814 extra = {}
2815 2815
2816 2816 def fail(f, msg):
2817 2817 raise error.Abort(b'%s: %s' % (f, msg))
2818 2818
2819 2819 if not match:
2820 2820 match = matchmod.always()
2821 2821
2822 2822 if not force:
2823 2823 match.bad = fail
2824 2824
2825 2825 # lock() for recent changelog (see issue4368)
2826 2826 with self.wlock(), self.lock():
2827 2827 wctx = self[None]
2828 2828 merge = len(wctx.parents()) > 1
2829 2829
2830 2830 if not force and merge and not match.always():
2831 2831 raise error.Abort(
2832 2832 _(
2833 2833 b'cannot partially commit a merge '
2834 2834 b'(do not specify files or patterns)'
2835 2835 )
2836 2836 )
2837 2837
2838 2838 status = self.status(match=match, clean=force)
2839 2839 if force:
2840 2840 status.modified.extend(
2841 2841 status.clean
2842 2842 ) # mq may commit clean files
2843 2843
2844 2844 # check subrepos
2845 2845 subs, commitsubs, newstate = subrepoutil.precommit(
2846 2846 self.ui, wctx, status, match, force=force
2847 2847 )
2848 2848
2849 2849 # make sure all explicit patterns are matched
2850 2850 if not force:
2851 2851 self.checkcommitpatterns(wctx, match, status, fail)
2852 2852
2853 2853 cctx = context.workingcommitctx(
2854 2854 self, status, text, user, date, extra
2855 2855 )
2856 2856
2857 2857 ms = mergestatemod.mergestate.read(self)
2858 2858 mergeutil.checkunresolved(ms)
2859 2859
2860 2860 # internal config: ui.allowemptycommit
2861 2861 if cctx.isempty() and not self.ui.configbool(
2862 2862 b'ui', b'allowemptycommit'
2863 2863 ):
2864 2864 self.ui.debug(b'nothing to commit, clearing merge state\n')
2865 2865 ms.reset()
2866 2866 return None
2867 2867
2868 2868 if merge and cctx.deleted():
2869 2869 raise error.Abort(_(b"cannot commit merge with missing files"))
2870 2870
2871 2871 if editor:
2872 2872 cctx._text = editor(self, cctx, subs)
2873 2873 edited = text != cctx._text
2874 2874
2875 2875 # Save commit message in case this transaction gets rolled back
2876 2876 # (e.g. by a pretxncommit hook). Leave the content alone on
2877 2877 # the assumption that the user will use the same editor again.
2878 2878 msgfn = self.savecommitmessage(cctx._text)
2879 2879
2880 2880 # commit subs and write new state
2881 2881 if subs:
2882 2882 uipathfn = scmutil.getuipathfn(self)
2883 2883 for s in sorted(commitsubs):
2884 2884 sub = wctx.sub(s)
2885 2885 self.ui.status(
2886 2886 _(b'committing subrepository %s\n')
2887 2887 % uipathfn(subrepoutil.subrelpath(sub))
2888 2888 )
2889 2889 sr = sub.commit(cctx._text, user, date)
2890 2890 newstate[s] = (newstate[s][0], sr)
2891 2891 subrepoutil.writestate(self, newstate)
2892 2892
2893 2893 p1, p2 = self.dirstate.parents()
2894 2894 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or b'')
2895 2895 try:
2896 2896 self.hook(
2897 2897 b"precommit", throw=True, parent1=hookp1, parent2=hookp2
2898 2898 )
2899 2899 with self.transaction(b'commit'):
2900 2900 ret = self.commitctx(cctx, True)
2901 2901 # update bookmarks, dirstate and mergestate
2902 2902 bookmarks.update(self, [p1, p2], ret)
2903 2903 cctx.markcommitted(ret)
2904 2904 ms.reset()
2905 2905 except: # re-raises
2906 2906 if edited:
2907 2907 self.ui.write(
2908 2908 _(b'note: commit message saved in %s\n') % msgfn
2909 2909 )
2910 2910 self.ui.write(
2911 2911 _(
2912 2912 b"note: use 'hg commit --logfile "
2913 2913 b".hg/last-message.txt --edit' to reuse it\n"
2914 2914 )
2915 2915 )
2916 2916 raise
2917 2917
2918 2918 def commithook(unused_success):
2919 2919 # hack for command that use a temporary commit (eg: histedit)
2920 2920 # temporary commit got stripped before hook release
2921 2921 if self.changelog.hasnode(ret):
2922 2922 self.hook(
2923 2923 b"commit", node=hex(ret), parent1=hookp1, parent2=hookp2
2924 2924 )
2925 2925
2926 2926 self._afterlock(commithook)
2927 2927 return ret
2928 2928
2929 2929 @unfilteredmethod
2930 2930 def commitctx(self, ctx, error=False, origctx=None):
2931 2931 return commit.commitctx(self, ctx, error=error, origctx=origctx)
2932 2932
2933 2933 @unfilteredmethod
2934 2934 def destroying(self):
2935 2935 '''Inform the repository that nodes are about to be destroyed.
2936 2936 Intended for use by strip and rollback, so there's a common
2937 2937 place for anything that has to be done before destroying history.
2938 2938
2939 2939 This is mostly useful for saving state that is in memory and waiting
2940 2940 to be flushed when the current lock is released. Because a call to
2941 2941 destroyed is imminent, the repo will be invalidated causing those
2942 2942 changes to stay in memory (waiting for the next unlock), or vanish
2943 2943 completely.
2944 2944 '''
2945 2945 # When using the same lock to commit and strip, the phasecache is left
2946 2946 # dirty after committing. Then when we strip, the repo is invalidated,
2947 2947 # causing those changes to disappear.
2948 2948 if '_phasecache' in vars(self):
2949 2949 self._phasecache.write()
2950 2950
2951 2951 @unfilteredmethod
2952 2952 def destroyed(self):
2953 2953 '''Inform the repository that nodes have been destroyed.
2954 2954 Intended for use by strip and rollback, so there's a common
2955 2955 place for anything that has to be done after destroying history.
2956 2956 '''
2957 2957 # When one tries to:
2958 2958 # 1) destroy nodes thus calling this method (e.g. strip)
2959 2959 # 2) use phasecache somewhere (e.g. commit)
2960 2960 #
2961 2961 # then 2) will fail because the phasecache contains nodes that were
2962 2962 # removed. We can either remove phasecache from the filecache,
2963 2963 # causing it to reload next time it is accessed, or simply filter
2964 2964 # the removed nodes now and write the updated cache.
2965 2965 self._phasecache.filterunknown(self)
2966 2966 self._phasecache.write()
2967 2967
2968 2968 # refresh all repository caches
2969 2969 self.updatecaches()
2970 2970
2971 2971 # Ensure the persistent tag cache is updated. Doing it now
2972 2972 # means that the tag cache only has to worry about destroyed
2973 2973 # heads immediately after a strip/rollback. That in turn
2974 2974 # guarantees that "cachetip == currenttip" (comparing both rev
2975 2975 # and node) always means no nodes have been added or destroyed.
2976 2976
2977 2977 # XXX this is suboptimal when qrefresh'ing: we strip the current
2978 2978 # head, refresh the tag cache, then immediately add a new head.
2979 2979 # But I think doing it this way is necessary for the "instant
2980 2980 # tag cache retrieval" case to work.
2981 2981 self.invalidate()
2982 2982
2983 2983 def status(
2984 2984 self,
2985 2985 node1=b'.',
2986 2986 node2=None,
2987 2987 match=None,
2988 2988 ignored=False,
2989 2989 clean=False,
2990 2990 unknown=False,
2991 2991 listsubrepos=False,
2992 2992 ):
2993 2993 '''a convenience method that calls node1.status(node2)'''
2994 2994 return self[node1].status(
2995 2995 node2, match, ignored, clean, unknown, listsubrepos
2996 2996 )
2997 2997
2998 2998 def addpostdsstatus(self, ps):
2999 2999 """Add a callback to run within the wlock, at the point at which status
3000 3000 fixups happen.
3001 3001
3002 3002 On status completion, callback(wctx, status) will be called with the
3003 3003 wlock held, unless the dirstate has changed from underneath or the wlock
3004 3004 couldn't be grabbed.
3005 3005
3006 3006 Callbacks should not capture and use a cached copy of the dirstate --
3007 3007 it might change in the meanwhile. Instead, they should access the
3008 3008 dirstate via wctx.repo().dirstate.
3009 3009
3010 3010 This list is emptied out after each status run -- extensions should
3011 3011 make sure it adds to this list each time dirstate.status is called.
3012 3012 Extensions should also make sure they don't call this for statuses
3013 3013 that don't involve the dirstate.
3014 3014 """
3015 3015
3016 3016 # The list is located here for uniqueness reasons -- it is actually
3017 3017 # managed by the workingctx, but that isn't unique per-repo.
3018 3018 self._postdsstatus.append(ps)
3019 3019
3020 3020 def postdsstatus(self):
3021 3021 """Used by workingctx to get the list of post-dirstate-status hooks."""
3022 3022 return self._postdsstatus
3023 3023
3024 3024 def clearpostdsstatus(self):
3025 3025 """Used by workingctx to clear post-dirstate-status hooks."""
3026 3026 del self._postdsstatus[:]
3027 3027
3028 3028 def heads(self, start=None):
3029 3029 if start is None:
3030 3030 cl = self.changelog
3031 3031 headrevs = reversed(cl.headrevs())
3032 3032 return [cl.node(rev) for rev in headrevs]
3033 3033
3034 3034 heads = self.changelog.heads(start)
3035 3035 # sort the output in rev descending order
3036 3036 return sorted(heads, key=self.changelog.rev, reverse=True)
3037 3037
3038 3038 def branchheads(self, branch=None, start=None, closed=False):
3039 3039 '''return a (possibly filtered) list of heads for the given branch
3040 3040
3041 3041 Heads are returned in topological order, from newest to oldest.
3042 3042 If branch is None, use the dirstate branch.
3043 3043 If start is not None, return only heads reachable from start.
3044 3044 If closed is True, return heads that are marked as closed as well.
3045 3045 '''
3046 3046 if branch is None:
3047 3047 branch = self[None].branch()
3048 3048 branches = self.branchmap()
3049 3049 if not branches.hasbranch(branch):
3050 3050 return []
3051 3051 # the cache returns heads ordered lowest to highest
3052 3052 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
3053 3053 if start is not None:
3054 3054 # filter out the heads that cannot be reached from startrev
3055 3055 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
3056 3056 bheads = [h for h in bheads if h in fbheads]
3057 3057 return bheads
3058 3058
3059 3059 def branches(self, nodes):
3060 3060 if not nodes:
3061 3061 nodes = [self.changelog.tip()]
3062 3062 b = []
3063 3063 for n in nodes:
3064 3064 t = n
3065 3065 while True:
3066 3066 p = self.changelog.parents(n)
3067 3067 if p[1] != nullid or p[0] == nullid:
3068 3068 b.append((t, n, p[0], p[1]))
3069 3069 break
3070 3070 n = p[0]
3071 3071 return b
3072 3072
3073 3073 def between(self, pairs):
3074 3074 r = []
3075 3075
3076 3076 for top, bottom in pairs:
3077 3077 n, l, i = top, [], 0
3078 3078 f = 1
3079 3079
3080 3080 while n != bottom and n != nullid:
3081 3081 p = self.changelog.parents(n)[0]
3082 3082 if i == f:
3083 3083 l.append(n)
3084 3084 f = f * 2
3085 3085 n = p
3086 3086 i += 1
3087 3087
3088 3088 r.append(l)
3089 3089
3090 3090 return r
3091 3091
3092 3092 def checkpush(self, pushop):
3093 3093 """Extensions can override this function if additional checks have
3094 3094 to be performed before pushing, or call it if they override push
3095 3095 command.
3096 3096 """
3097 3097
3098 3098 @unfilteredpropertycache
3099 3099 def prepushoutgoinghooks(self):
3100 3100 """Return util.hooks consists of a pushop with repo, remote, outgoing
3101 3101 methods, which are called before pushing changesets.
3102 3102 """
3103 3103 return util.hooks()
3104 3104
3105 3105 def pushkey(self, namespace, key, old, new):
3106 3106 try:
3107 3107 tr = self.currenttransaction()
3108 3108 hookargs = {}
3109 3109 if tr is not None:
3110 3110 hookargs.update(tr.hookargs)
3111 3111 hookargs = pycompat.strkwargs(hookargs)
3112 3112 hookargs['namespace'] = namespace
3113 3113 hookargs['key'] = key
3114 3114 hookargs['old'] = old
3115 3115 hookargs['new'] = new
3116 3116 self.hook(b'prepushkey', throw=True, **hookargs)
3117 3117 except error.HookAbort as exc:
3118 3118 self.ui.write_err(_(b"pushkey-abort: %s\n") % exc)
3119 3119 if exc.hint:
3120 3120 self.ui.write_err(_(b"(%s)\n") % exc.hint)
3121 3121 return False
3122 3122 self.ui.debug(b'pushing key for "%s:%s"\n' % (namespace, key))
3123 3123 ret = pushkey.push(self, namespace, key, old, new)
3124 3124
3125 3125 def runhook(unused_success):
3126 3126 self.hook(
3127 3127 b'pushkey',
3128 3128 namespace=namespace,
3129 3129 key=key,
3130 3130 old=old,
3131 3131 new=new,
3132 3132 ret=ret,
3133 3133 )
3134 3134
3135 3135 self._afterlock(runhook)
3136 3136 return ret
3137 3137
3138 3138 def listkeys(self, namespace):
3139 3139 self.hook(b'prelistkeys', throw=True, namespace=namespace)
3140 3140 self.ui.debug(b'listing keys for "%s"\n' % namespace)
3141 3141 values = pushkey.list(self, namespace)
3142 3142 self.hook(b'listkeys', namespace=namespace, values=values)
3143 3143 return values
3144 3144
3145 3145 def debugwireargs(self, one, two, three=None, four=None, five=None):
3146 3146 '''used to test argument passing over the wire'''
3147 3147 return b"%s %s %s %s %s" % (
3148 3148 one,
3149 3149 two,
3150 3150 pycompat.bytestr(three),
3151 3151 pycompat.bytestr(four),
3152 3152 pycompat.bytestr(five),
3153 3153 )
3154 3154
3155 3155 def savecommitmessage(self, text):
3156 3156 fp = self.vfs(b'last-message.txt', b'wb')
3157 3157 try:
3158 3158 fp.write(text)
3159 3159 finally:
3160 3160 fp.close()
3161 3161 return self.pathto(fp.name[len(self.root) + 1 :])
3162 3162
3163 3163
3164 3164 # used to avoid circular references so destructors work
3165 3165 def aftertrans(files):
3166 3166 renamefiles = [tuple(t) for t in files]
3167 3167
3168 3168 def a():
3169 3169 for vfs, src, dest in renamefiles:
3170 3170 # if src and dest refer to a same file, vfs.rename is a no-op,
3171 3171 # leaving both src and dest on disk. delete dest to make sure
3172 3172 # the rename couldn't be such a no-op.
3173 3173 vfs.tryunlink(dest)
3174 3174 try:
3175 3175 vfs.rename(src, dest)
3176 3176 except OSError: # journal file does not yet exist
3177 3177 pass
3178 3178
3179 3179 return a
3180 3180
3181 3181
3182 3182 def undoname(fn):
3183 3183 base, name = os.path.split(fn)
3184 3184 assert name.startswith(b'journal')
3185 3185 return os.path.join(base, name.replace(b'journal', b'undo', 1))
3186 3186
3187 3187
3188 3188 def instance(ui, path, create, intents=None, createopts=None):
3189 3189 localpath = util.urllocalpath(path)
3190 3190 if create:
3191 3191 createrepository(ui, localpath, createopts=createopts)
3192 3192
3193 3193 return makelocalrepository(ui, localpath, intents=intents)
3194 3194
3195 3195
3196 3196 def islocal(path):
3197 3197 return True
3198 3198
3199 3199
3200 3200 def defaultcreateopts(ui, createopts=None):
3201 3201 """Populate the default creation options for a repository.
3202 3202
3203 3203 A dictionary of explicitly requested creation options can be passed
3204 3204 in. Missing keys will be populated.
3205 3205 """
3206 3206 createopts = dict(createopts or {})
3207 3207
3208 3208 if b'backend' not in createopts:
3209 3209 # experimental config: storage.new-repo-backend
3210 3210 createopts[b'backend'] = ui.config(b'storage', b'new-repo-backend')
3211 3211
3212 3212 return createopts
3213 3213
3214 3214
3215 3215 def newreporequirements(ui, createopts):
3216 3216 """Determine the set of requirements for a new local repository.
3217 3217
3218 3218 Extensions can wrap this function to specify custom requirements for
3219 3219 new repositories.
3220 3220 """
3221 3221 # If the repo is being created from a shared repository, we copy
3222 3222 # its requirements.
3223 3223 if b'sharedrepo' in createopts:
3224 3224 requirements = set(createopts[b'sharedrepo'].requirements)
3225 3225 if createopts.get(b'sharedrelative'):
3226 3226 requirements.add(b'relshared')
3227 3227 else:
3228 3228 requirements.add(b'shared')
3229 3229
3230 3230 return requirements
3231 3231
3232 3232 if b'backend' not in createopts:
3233 3233 raise error.ProgrammingError(
3234 3234 b'backend key not present in createopts; '
3235 3235 b'was defaultcreateopts() called?'
3236 3236 )
3237 3237
3238 3238 if createopts[b'backend'] != b'revlogv1':
3239 3239 raise error.Abort(
3240 3240 _(
3241 3241 b'unable to determine repository requirements for '
3242 3242 b'storage backend: %s'
3243 3243 )
3244 3244 % createopts[b'backend']
3245 3245 )
3246 3246
3247 3247 requirements = {b'revlogv1'}
3248 3248 if ui.configbool(b'format', b'usestore'):
3249 3249 requirements.add(b'store')
3250 3250 if ui.configbool(b'format', b'usefncache'):
3251 3251 requirements.add(b'fncache')
3252 3252 if ui.configbool(b'format', b'dotencode'):
3253 3253 requirements.add(b'dotencode')
3254 3254
3255 3255 compengines = ui.configlist(b'format', b'revlog-compression')
3256 3256 for compengine in compengines:
3257 3257 if compengine in util.compengines:
3258 3258 break
3259 3259 else:
3260 3260 raise error.Abort(
3261 3261 _(
3262 3262 b'compression engines %s defined by '
3263 3263 b'format.revlog-compression not available'
3264 3264 )
3265 3265 % b', '.join(b'"%s"' % e for e in compengines),
3266 3266 hint=_(
3267 3267 b'run "hg debuginstall" to list available '
3268 3268 b'compression engines'
3269 3269 ),
3270 3270 )
3271 3271
3272 3272 # zlib is the historical default and doesn't need an explicit requirement.
3273 3273 if compengine == b'zstd':
3274 3274 requirements.add(b'revlog-compression-zstd')
3275 3275 elif compengine != b'zlib':
3276 3276 requirements.add(b'exp-compression-%s' % compengine)
3277 3277
3278 3278 if scmutil.gdinitconfig(ui):
3279 3279 requirements.add(b'generaldelta')
3280 3280 if ui.configbool(b'format', b'sparse-revlog'):
3281 3281 requirements.add(SPARSEREVLOG_REQUIREMENT)
3282 3282
3283 3283 # experimental config: format.exp-use-side-data
3284 3284 if ui.configbool(b'format', b'exp-use-side-data'):
3285 3285 requirements.add(SIDEDATA_REQUIREMENT)
3286 3286 # experimental config: format.exp-use-copies-side-data-changeset
3287 3287 if ui.configbool(b'format', b'exp-use-copies-side-data-changeset'):
3288 3288 requirements.add(SIDEDATA_REQUIREMENT)
3289 3289 requirements.add(COPIESSDC_REQUIREMENT)
3290 3290 if ui.configbool(b'experimental', b'treemanifest'):
3291 3291 requirements.add(b'treemanifest')
3292 3292
3293 3293 revlogv2 = ui.config(b'experimental', b'revlogv2')
3294 3294 if revlogv2 == b'enable-unstable-format-and-corrupt-my-data':
3295 3295 requirements.remove(b'revlogv1')
3296 3296 # generaldelta is implied by revlogv2.
3297 3297 requirements.discard(b'generaldelta')
3298 3298 requirements.add(REVLOGV2_REQUIREMENT)
3299 3299 # experimental config: format.internal-phase
3300 3300 if ui.configbool(b'format', b'internal-phase'):
3301 3301 requirements.add(b'internal-phase')
3302 3302
3303 3303 if createopts.get(b'narrowfiles'):
3304 3304 requirements.add(repository.NARROW_REQUIREMENT)
3305 3305
3306 3306 if createopts.get(b'lfs'):
3307 3307 requirements.add(b'lfs')
3308 3308
3309 3309 if ui.configbool(b'format', b'bookmarks-in-store'):
3310 3310 requirements.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3311 3311
3312 3312 if ui.configbool(b'format', b'use-persistent-nodemap'):
3313 3313 requirements.add(NODEMAP_REQUIREMENT)
3314 3314
3315 3315 return requirements
3316 3316
3317 3317
3318 def checkrequirementscompat(ui, requirements):
3319 """ Checks compatibility of repository requirements enabled and disabled.
3320
3321 Returns a set of requirements which needs to be dropped because dependend
3322 requirements are not enabled. Also warns users about it """
3323
3324 dropped = set()
3325
3326 if b'store' not in requirements:
3327 if bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT in requirements:
3328 ui.warn(
3329 _(
3330 b'ignoring enabled \'format.bookmarks-in-store\' config '
3331 b'beacuse it is incompatible with disabled '
3332 b'\'format.usestore\' config\n'
3333 )
3334 )
3335 dropped.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3336
3337 return dropped
3338
3339
3318 3340 def filterknowncreateopts(ui, createopts):
3319 3341 """Filters a dict of repo creation options against options that are known.
3320 3342
3321 3343 Receives a dict of repo creation options and returns a dict of those
3322 3344 options that we don't know how to handle.
3323 3345
3324 3346 This function is called as part of repository creation. If the
3325 3347 returned dict contains any items, repository creation will not
3326 3348 be allowed, as it means there was a request to create a repository
3327 3349 with options not recognized by loaded code.
3328 3350
3329 3351 Extensions can wrap this function to filter out creation options
3330 3352 they know how to handle.
3331 3353 """
3332 3354 known = {
3333 3355 b'backend',
3334 3356 b'lfs',
3335 3357 b'narrowfiles',
3336 3358 b'sharedrepo',
3337 3359 b'sharedrelative',
3338 3360 b'shareditems',
3339 3361 b'shallowfilestore',
3340 3362 }
3341 3363
3342 3364 return {k: v for k, v in createopts.items() if k not in known}
3343 3365
3344 3366
3345 3367 def createrepository(ui, path, createopts=None):
3346 3368 """Create a new repository in a vfs.
3347 3369
3348 3370 ``path`` path to the new repo's working directory.
3349 3371 ``createopts`` options for the new repository.
3350 3372
3351 3373 The following keys for ``createopts`` are recognized:
3352 3374
3353 3375 backend
3354 3376 The storage backend to use.
3355 3377 lfs
3356 3378 Repository will be created with ``lfs`` requirement. The lfs extension
3357 3379 will automatically be loaded when the repository is accessed.
3358 3380 narrowfiles
3359 3381 Set up repository to support narrow file storage.
3360 3382 sharedrepo
3361 3383 Repository object from which storage should be shared.
3362 3384 sharedrelative
3363 3385 Boolean indicating if the path to the shared repo should be
3364 3386 stored as relative. By default, the pointer to the "parent" repo
3365 3387 is stored as an absolute path.
3366 3388 shareditems
3367 3389 Set of items to share to the new repository (in addition to storage).
3368 3390 shallowfilestore
3369 3391 Indicates that storage for files should be shallow (not all ancestor
3370 3392 revisions are known).
3371 3393 """
3372 3394 createopts = defaultcreateopts(ui, createopts=createopts)
3373 3395
3374 3396 unknownopts = filterknowncreateopts(ui, createopts)
3375 3397
3376 3398 if not isinstance(unknownopts, dict):
3377 3399 raise error.ProgrammingError(
3378 3400 b'filterknowncreateopts() did not return a dict'
3379 3401 )
3380 3402
3381 3403 if unknownopts:
3382 3404 raise error.Abort(
3383 3405 _(
3384 3406 b'unable to create repository because of unknown '
3385 3407 b'creation option: %s'
3386 3408 )
3387 3409 % b', '.join(sorted(unknownopts)),
3388 3410 hint=_(b'is a required extension not loaded?'),
3389 3411 )
3390 3412
3391 3413 requirements = newreporequirements(ui, createopts=createopts)
3414 requirements -= checkrequirementscompat(ui, requirements)
3392 3415
3393 3416 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
3394 3417
3395 3418 hgvfs = vfsmod.vfs(wdirvfs.join(b'.hg'))
3396 3419 if hgvfs.exists():
3397 3420 raise error.RepoError(_(b'repository %s already exists') % path)
3398 3421
3399 3422 if b'sharedrepo' in createopts:
3400 3423 sharedpath = createopts[b'sharedrepo'].sharedpath
3401 3424
3402 3425 if createopts.get(b'sharedrelative'):
3403 3426 try:
3404 3427 sharedpath = os.path.relpath(sharedpath, hgvfs.base)
3405 3428 except (IOError, ValueError) as e:
3406 3429 # ValueError is raised on Windows if the drive letters differ
3407 3430 # on each path.
3408 3431 raise error.Abort(
3409 3432 _(b'cannot calculate relative path'),
3410 3433 hint=stringutil.forcebytestr(e),
3411 3434 )
3412 3435
3413 3436 if not wdirvfs.exists():
3414 3437 wdirvfs.makedirs()
3415 3438
3416 3439 hgvfs.makedir(notindexed=True)
3417 3440 if b'sharedrepo' not in createopts:
3418 3441 hgvfs.mkdir(b'cache')
3419 3442 hgvfs.mkdir(b'wcache')
3420 3443
3421 3444 if b'store' in requirements and b'sharedrepo' not in createopts:
3422 3445 hgvfs.mkdir(b'store')
3423 3446
3424 3447 # We create an invalid changelog outside the store so very old
3425 3448 # Mercurial versions (which didn't know about the requirements
3426 3449 # file) encounter an error on reading the changelog. This
3427 3450 # effectively locks out old clients and prevents them from
3428 3451 # mucking with a repo in an unknown format.
3429 3452 #
3430 3453 # The revlog header has version 2, which won't be recognized by
3431 3454 # such old clients.
3432 3455 hgvfs.append(
3433 3456 b'00changelog.i',
3434 3457 b'\0\0\0\2 dummy changelog to prevent using the old repo '
3435 3458 b'layout',
3436 3459 )
3437 3460
3438 3461 scmutil.writerequires(hgvfs, requirements)
3439 3462
3440 3463 # Write out file telling readers where to find the shared store.
3441 3464 if b'sharedrepo' in createopts:
3442 3465 hgvfs.write(b'sharedpath', sharedpath)
3443 3466
3444 3467 if createopts.get(b'shareditems'):
3445 3468 shared = b'\n'.join(sorted(createopts[b'shareditems'])) + b'\n'
3446 3469 hgvfs.write(b'shared', shared)
3447 3470
3448 3471
3449 3472 def poisonrepository(repo):
3450 3473 """Poison a repository instance so it can no longer be used."""
3451 3474 # Perform any cleanup on the instance.
3452 3475 repo.close()
3453 3476
3454 3477 # Our strategy is to replace the type of the object with one that
3455 3478 # has all attribute lookups result in error.
3456 3479 #
3457 3480 # But we have to allow the close() method because some constructors
3458 3481 # of repos call close() on repo references.
3459 3482 class poisonedrepository(object):
3460 3483 def __getattribute__(self, item):
3461 3484 if item == 'close':
3462 3485 return object.__getattribute__(self, item)
3463 3486
3464 3487 raise error.ProgrammingError(
3465 3488 b'repo instances should not be used after unshare'
3466 3489 )
3467 3490
3468 3491 def close(self):
3469 3492 pass
3470 3493
3471 3494 # We may have a repoview, which intercepts __setattr__. So be sure
3472 3495 # we operate at the lowest level possible.
3473 3496 object.__setattr__(repo, '__class__', poisonedrepository)
@@ -1,281 +1,286 b''
1 1 #testcases vfs svfs
2 2
3 3 $ echo "[extensions]" >> $HGRCPATH
4 4 $ echo "share = " >> $HGRCPATH
5 5
6 6 #if svfs
7 7 $ echo "[format]" >> $HGRCPATH
8 8 $ echo "bookmarks-in-store = yes " >> $HGRCPATH
9 9 #endif
10 10
11 11 prepare repo1
12 12
13 13 $ hg init repo1
14 14 $ cd repo1
15 15 $ echo a > a
16 16 $ hg commit -A -m'init'
17 17 adding a
18 18 $ echo a >> a
19 19 $ hg commit -m'change in shared clone'
20 20 $ echo b > b
21 21 $ hg commit -A -m'another file'
22 22 adding b
23 23
24 24 share it
25 25
26 26 $ cd ..
27 27 $ hg share repo1 repo2
28 28 updating working directory
29 29 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
30 30
31 31 test sharing bookmarks
32 32
33 33 $ hg share -B repo1 repo3
34 34 updating working directory
35 35 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
36 36 $ cd repo1
37 37 $ hg bookmark bm1
38 38 $ hg bookmarks
39 39 * bm1 2:c2e0ac586386
40 40 $ cd ../repo2
41 41 $ hg book bm2
42 42 $ hg bookmarks
43 43 bm1 2:c2e0ac586386 (svfs !)
44 44 * bm2 2:c2e0ac586386
45 45 $ cd ../repo3
46 46 $ hg bookmarks
47 47 bm1 2:c2e0ac586386
48 48 bm2 2:c2e0ac586386 (svfs !)
49 49 $ hg book bm3
50 50 $ hg bookmarks
51 51 bm1 2:c2e0ac586386
52 52 bm2 2:c2e0ac586386 (svfs !)
53 53 * bm3 2:c2e0ac586386
54 54 $ cd ../repo1
55 55 $ hg bookmarks
56 56 * bm1 2:c2e0ac586386
57 57 bm2 2:c2e0ac586386 (svfs !)
58 58 bm3 2:c2e0ac586386
59 59
60 60 check whether HG_PENDING makes pending changes only in relatd
61 61 repositories visible to an external hook.
62 62
63 63 In "hg share" case, another transaction can't run in other
64 64 repositories sharing same source repository, because starting
65 65 transaction requires locking store of source repository.
66 66
67 67 Therefore, this test scenario ignores checking visibility of
68 68 .hg/bookmarks.pending in repo2, which shares repo1 without bookmarks.
69 69
70 70 $ cat > $TESTTMP/checkbookmarks.sh <<EOF
71 71 > echo "@repo1"
72 72 > hg -R "$TESTTMP/repo1" bookmarks
73 73 > echo "@repo2"
74 74 > hg -R "$TESTTMP/repo2" bookmarks
75 75 > echo "@repo3"
76 76 > hg -R "$TESTTMP/repo3" bookmarks
77 77 > exit 1 # to avoid adding new bookmark for subsequent tests
78 78 > EOF
79 79
80 80 $ cd ../repo1
81 81 $ hg --config hooks.pretxnclose="sh $TESTTMP/checkbookmarks.sh" -q book bmX
82 82 @repo1
83 83 bm1 2:c2e0ac586386
84 84 bm2 2:c2e0ac586386 (svfs !)
85 85 bm3 2:c2e0ac586386
86 86 * bmX 2:c2e0ac586386
87 87 @repo2
88 88 bm1 2:c2e0ac586386 (svfs !)
89 89 * bm2 2:c2e0ac586386
90 90 bm3 2:c2e0ac586386 (svfs !)
91 91 @repo3
92 92 bm1 2:c2e0ac586386
93 93 bm2 2:c2e0ac586386 (svfs !)
94 94 * bm3 2:c2e0ac586386
95 95 bmX 2:c2e0ac586386 (vfs !)
96 96 transaction abort!
97 97 rollback completed
98 98 abort: pretxnclose hook exited with status 1
99 99 [255]
100 100 $ hg book bm1
101 101
102 102 FYI, in contrast to above test, bmX is invisible in repo1 (= shared
103 103 src), because (1) HG_PENDING refers only repo3 and (2)
104 104 "bookmarks.pending" is written only into repo3.
105 105
106 106 $ cd ../repo3
107 107 $ hg --config hooks.pretxnclose="sh $TESTTMP/checkbookmarks.sh" -q book bmX
108 108 @repo1
109 109 * bm1 2:c2e0ac586386
110 110 bm2 2:c2e0ac586386 (svfs !)
111 111 bm3 2:c2e0ac586386
112 112 @repo2
113 113 bm1 2:c2e0ac586386 (svfs !)
114 114 * bm2 2:c2e0ac586386
115 115 bm3 2:c2e0ac586386 (svfs !)
116 116 @repo3
117 117 bm1 2:c2e0ac586386
118 118 bm2 2:c2e0ac586386 (svfs !)
119 119 bm3 2:c2e0ac586386
120 120 * bmX 2:c2e0ac586386
121 121 transaction abort!
122 122 rollback completed
123 123 abort: pretxnclose hook exited with status 1
124 124 [255]
125 125 $ hg book bm3
126 126
127 127 clean up bm2 since it's uninteresting (not shared in the vfs case and
128 128 same as bm3 in the svfs case)
129 129 $ cd ../repo2
130 130 $ hg book -d bm2
131 131
132 132 $ cd ../repo1
133 133
134 134 test that commits work
135 135
136 136 $ echo 'shared bookmarks' > a
137 137 $ hg commit -m 'testing shared bookmarks'
138 138 $ hg bookmarks
139 139 * bm1 3:b87954705719
140 140 bm3 2:c2e0ac586386
141 141 $ cd ../repo3
142 142 $ hg bookmarks
143 143 bm1 3:b87954705719
144 144 * bm3 2:c2e0ac586386
145 145 $ echo 'more shared bookmarks' > a
146 146 $ hg commit -m 'testing shared bookmarks'
147 147 created new head
148 148 $ hg bookmarks
149 149 bm1 3:b87954705719
150 150 * bm3 4:62f4ded848e4
151 151 $ cd ../repo1
152 152 $ hg bookmarks
153 153 * bm1 3:b87954705719
154 154 bm3 4:62f4ded848e4
155 155 $ cd ..
156 156
157 157 test pushing bookmarks works
158 158
159 159 $ hg clone repo3 repo4
160 160 updating to branch default
161 161 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
162 162 $ cd repo4
163 163 $ hg boo bm4
164 164 $ echo foo > b
165 165 $ hg commit -m 'foo in b'
166 166 $ hg boo
167 167 bm1 3:b87954705719
168 168 bm3 4:62f4ded848e4
169 169 * bm4 5:92793bfc8cad
170 170 $ hg push -B bm4
171 171 pushing to $TESTTMP/repo3
172 172 searching for changes
173 173 adding changesets
174 174 adding manifests
175 175 adding file changes
176 176 added 1 changesets with 1 changes to 1 files
177 177 exporting bookmark bm4
178 178 $ cd ../repo1
179 179 $ hg bookmarks
180 180 * bm1 3:b87954705719
181 181 bm3 4:62f4ded848e4
182 182 bm4 5:92793bfc8cad
183 183 $ cd ../repo3
184 184 $ hg bookmarks
185 185 bm1 3:b87954705719
186 186 * bm3 4:62f4ded848e4
187 187 bm4 5:92793bfc8cad
188 188 $ cd ..
189 189
190 190 test behavior when sharing a shared repo
191 191
192 192 $ hg share -B repo3 missingdir/repo5
193 193 updating working directory
194 194 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
195 195 $ cd missingdir/repo5
196 196 $ hg book
197 197 bm1 3:b87954705719
198 198 bm3 4:62f4ded848e4
199 199 bm4 5:92793bfc8cad
200 200 $ cd ../..
201 201
202 202 test what happens when an active bookmark is deleted
203 203
204 204 $ cd repo1
205 205 $ hg boo -d bm3
206 206 $ hg boo
207 207 * bm1 3:b87954705719
208 208 bm4 5:92793bfc8cad
209 209 $ cd ../repo3
210 210 $ hg boo
211 211 bm1 3:b87954705719
212 212 bm4 5:92793bfc8cad
213 213 $ cd ..
214 214
215 215 verify that bookmarks are not written on failed transaction
216 216
217 217 $ cat > failpullbookmarks.py << EOF
218 218 > """A small extension that makes bookmark pulls fail, for testing"""
219 219 > from __future__ import absolute_import
220 220 > from mercurial import (
221 221 > error,
222 222 > exchange,
223 223 > extensions,
224 224 > )
225 225 > def _pullbookmarks(orig, pullop):
226 226 > orig(pullop)
227 227 > raise error.HookAbort('forced failure by extension')
228 228 > def extsetup(ui):
229 229 > extensions.wrapfunction(exchange, '_pullbookmarks', _pullbookmarks)
230 230 > EOF
231 231 $ cd repo4
232 232 $ hg boo
233 233 bm1 3:b87954705719
234 234 bm3 4:62f4ded848e4
235 235 * bm4 5:92793bfc8cad
236 236 $ cd ../repo3
237 237 $ hg boo
238 238 bm1 3:b87954705719
239 239 bm4 5:92793bfc8cad
240 240 $ hg --config "extensions.failpullbookmarks=$TESTTMP/failpullbookmarks.py" pull $TESTTMP/repo4
241 241 pulling from $TESTTMP/repo4
242 242 searching for changes
243 243 no changes found
244 244 adding remote bookmark bm3
245 245 abort: forced failure by extension
246 246 [255]
247 247 $ hg boo
248 248 bm1 3:b87954705719
249 249 bm4 5:92793bfc8cad
250 250 $ hg pull $TESTTMP/repo4
251 251 pulling from $TESTTMP/repo4
252 252 searching for changes
253 253 no changes found
254 254 adding remote bookmark bm3
255 255 1 local changesets published
256 256 $ hg boo
257 257 bm1 3:b87954705719
258 258 * bm3 4:62f4ded848e4
259 259 bm4 5:92793bfc8cad
260 260 $ cd ..
261 261
262 262 verify bookmark behavior after unshare
263 263
264 264 $ cd repo3
265 265 $ hg unshare
266 266 $ hg boo
267 267 bm1 3:b87954705719
268 268 * bm3 4:62f4ded848e4
269 269 bm4 5:92793bfc8cad
270 270 $ hg boo -d bm4
271 271 $ hg boo bm5
272 272 $ hg boo
273 273 bm1 3:b87954705719
274 274 bm3 4:62f4ded848e4
275 275 * bm5 4:62f4ded848e4
276 276 $ cd ../repo1
277 277 $ hg boo
278 278 * bm1 3:b87954705719
279 279 bm3 4:62f4ded848e4
280 280 bm4 5:92793bfc8cad
281 281 $ cd ..
282
283 Test that if store is disabled, we drop the bookmarksinstore requirement
284
285 $ hg init brokenrepo --config format.bookmarks-in-store=True --config format.usestore=false
286 ignoring enabled 'format.bookmarks-in-store' config beacuse it is incompatible with disabled 'format.usestore' config
General Comments 0
You need to be logged in to leave comments. Login now