##// END OF EJS Templates
commitctx: document _filecommit input and output...
marmoute -
r45624:41b32cbc default
parent child Browse files
Show More
@@ -1,3795 +1,3811 b''
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import errno
11 11 import os
12 12 import random
13 13 import sys
14 14 import time
15 15 import weakref
16 16
17 17 from .i18n import _
18 18 from .node import (
19 19 bin,
20 20 hex,
21 21 nullid,
22 22 nullrev,
23 23 short,
24 24 )
25 25 from .pycompat import (
26 26 delattr,
27 27 getattr,
28 28 )
29 29 from . import (
30 30 bookmarks,
31 31 branchmap,
32 32 bundle2,
33 33 changegroup,
34 34 color,
35 35 context,
36 36 dirstate,
37 37 dirstateguard,
38 38 discovery,
39 39 encoding,
40 40 error,
41 41 exchange,
42 42 extensions,
43 43 filelog,
44 44 hook,
45 45 lock as lockmod,
46 46 match as matchmod,
47 47 mergestate as mergestatemod,
48 48 mergeutil,
49 49 metadata,
50 50 namespaces,
51 51 narrowspec,
52 52 obsolete,
53 53 pathutil,
54 54 phases,
55 55 pushkey,
56 56 pycompat,
57 57 rcutil,
58 58 repoview,
59 59 revset,
60 60 revsetlang,
61 61 scmutil,
62 62 sparse,
63 63 store as storemod,
64 64 subrepoutil,
65 65 tags as tagsmod,
66 66 transaction,
67 67 txnutil,
68 68 util,
69 69 vfs as vfsmod,
70 70 )
71 71
72 72 from .interfaces import (
73 73 repository,
74 74 util as interfaceutil,
75 75 )
76 76
77 77 from .utils import (
78 78 hashutil,
79 79 procutil,
80 80 stringutil,
81 81 )
82 82
83 83 from .revlogutils import constants as revlogconst
84 84
85 85 release = lockmod.release
86 86 urlerr = util.urlerr
87 87 urlreq = util.urlreq
88 88
89 89 # set of (path, vfs-location) tuples. vfs-location is:
90 90 # - 'plain for vfs relative paths
91 91 # - '' for svfs relative paths
92 92 _cachedfiles = set()
93 93
94 94
95 95 class _basefilecache(scmutil.filecache):
96 96 """All filecache usage on repo are done for logic that should be unfiltered
97 97 """
98 98
99 99 def __get__(self, repo, type=None):
100 100 if repo is None:
101 101 return self
102 102 # proxy to unfiltered __dict__ since filtered repo has no entry
103 103 unfi = repo.unfiltered()
104 104 try:
105 105 return unfi.__dict__[self.sname]
106 106 except KeyError:
107 107 pass
108 108 return super(_basefilecache, self).__get__(unfi, type)
109 109
110 110 def set(self, repo, value):
111 111 return super(_basefilecache, self).set(repo.unfiltered(), value)
112 112
113 113
114 114 class repofilecache(_basefilecache):
115 115 """filecache for files in .hg but outside of .hg/store"""
116 116
117 117 def __init__(self, *paths):
118 118 super(repofilecache, self).__init__(*paths)
119 119 for path in paths:
120 120 _cachedfiles.add((path, b'plain'))
121 121
122 122 def join(self, obj, fname):
123 123 return obj.vfs.join(fname)
124 124
125 125
126 126 class storecache(_basefilecache):
127 127 """filecache for files in the store"""
128 128
129 129 def __init__(self, *paths):
130 130 super(storecache, self).__init__(*paths)
131 131 for path in paths:
132 132 _cachedfiles.add((path, b''))
133 133
134 134 def join(self, obj, fname):
135 135 return obj.sjoin(fname)
136 136
137 137
138 138 class mixedrepostorecache(_basefilecache):
139 139 """filecache for a mix files in .hg/store and outside"""
140 140
141 141 def __init__(self, *pathsandlocations):
142 142 # scmutil.filecache only uses the path for passing back into our
143 143 # join(), so we can safely pass a list of paths and locations
144 144 super(mixedrepostorecache, self).__init__(*pathsandlocations)
145 145 _cachedfiles.update(pathsandlocations)
146 146
147 147 def join(self, obj, fnameandlocation):
148 148 fname, location = fnameandlocation
149 149 if location == b'plain':
150 150 return obj.vfs.join(fname)
151 151 else:
152 152 if location != b'':
153 153 raise error.ProgrammingError(
154 154 b'unexpected location: %s' % location
155 155 )
156 156 return obj.sjoin(fname)
157 157
158 158
159 159 def isfilecached(repo, name):
160 160 """check if a repo has already cached "name" filecache-ed property
161 161
162 162 This returns (cachedobj-or-None, iscached) tuple.
163 163 """
164 164 cacheentry = repo.unfiltered()._filecache.get(name, None)
165 165 if not cacheentry:
166 166 return None, False
167 167 return cacheentry.obj, True
168 168
169 169
170 170 class unfilteredpropertycache(util.propertycache):
171 171 """propertycache that apply to unfiltered repo only"""
172 172
173 173 def __get__(self, repo, type=None):
174 174 unfi = repo.unfiltered()
175 175 if unfi is repo:
176 176 return super(unfilteredpropertycache, self).__get__(unfi)
177 177 return getattr(unfi, self.name)
178 178
179 179
180 180 class filteredpropertycache(util.propertycache):
181 181 """propertycache that must take filtering in account"""
182 182
183 183 def cachevalue(self, obj, value):
184 184 object.__setattr__(obj, self.name, value)
185 185
186 186
187 187 def hasunfilteredcache(repo, name):
188 188 """check if a repo has an unfilteredpropertycache value for <name>"""
189 189 return name in vars(repo.unfiltered())
190 190
191 191
192 192 def unfilteredmethod(orig):
193 193 """decorate method that always need to be run on unfiltered version"""
194 194
195 195 def wrapper(repo, *args, **kwargs):
196 196 return orig(repo.unfiltered(), *args, **kwargs)
197 197
198 198 return wrapper
199 199
200 200
201 201 moderncaps = {
202 202 b'lookup',
203 203 b'branchmap',
204 204 b'pushkey',
205 205 b'known',
206 206 b'getbundle',
207 207 b'unbundle',
208 208 }
209 209 legacycaps = moderncaps.union({b'changegroupsubset'})
210 210
211 211
212 212 @interfaceutil.implementer(repository.ipeercommandexecutor)
213 213 class localcommandexecutor(object):
214 214 def __init__(self, peer):
215 215 self._peer = peer
216 216 self._sent = False
217 217 self._closed = False
218 218
219 219 def __enter__(self):
220 220 return self
221 221
222 222 def __exit__(self, exctype, excvalue, exctb):
223 223 self.close()
224 224
225 225 def callcommand(self, command, args):
226 226 if self._sent:
227 227 raise error.ProgrammingError(
228 228 b'callcommand() cannot be used after sendcommands()'
229 229 )
230 230
231 231 if self._closed:
232 232 raise error.ProgrammingError(
233 233 b'callcommand() cannot be used after close()'
234 234 )
235 235
236 236 # We don't need to support anything fancy. Just call the named
237 237 # method on the peer and return a resolved future.
238 238 fn = getattr(self._peer, pycompat.sysstr(command))
239 239
240 240 f = pycompat.futures.Future()
241 241
242 242 try:
243 243 result = fn(**pycompat.strkwargs(args))
244 244 except Exception:
245 245 pycompat.future_set_exception_info(f, sys.exc_info()[1:])
246 246 else:
247 247 f.set_result(result)
248 248
249 249 return f
250 250
251 251 def sendcommands(self):
252 252 self._sent = True
253 253
254 254 def close(self):
255 255 self._closed = True
256 256
257 257
258 258 @interfaceutil.implementer(repository.ipeercommands)
259 259 class localpeer(repository.peer):
260 260 '''peer for a local repo; reflects only the most recent API'''
261 261
262 262 def __init__(self, repo, caps=None):
263 263 super(localpeer, self).__init__()
264 264
265 265 if caps is None:
266 266 caps = moderncaps.copy()
267 267 self._repo = repo.filtered(b'served')
268 268 self.ui = repo.ui
269 269 self._caps = repo._restrictcapabilities(caps)
270 270
271 271 # Begin of _basepeer interface.
272 272
273 273 def url(self):
274 274 return self._repo.url()
275 275
276 276 def local(self):
277 277 return self._repo
278 278
279 279 def peer(self):
280 280 return self
281 281
282 282 def canpush(self):
283 283 return True
284 284
285 285 def close(self):
286 286 self._repo.close()
287 287
288 288 # End of _basepeer interface.
289 289
290 290 # Begin of _basewirecommands interface.
291 291
292 292 def branchmap(self):
293 293 return self._repo.branchmap()
294 294
295 295 def capabilities(self):
296 296 return self._caps
297 297
298 298 def clonebundles(self):
299 299 return self._repo.tryread(b'clonebundles.manifest')
300 300
301 301 def debugwireargs(self, one, two, three=None, four=None, five=None):
302 302 """Used to test argument passing over the wire"""
303 303 return b"%s %s %s %s %s" % (
304 304 one,
305 305 two,
306 306 pycompat.bytestr(three),
307 307 pycompat.bytestr(four),
308 308 pycompat.bytestr(five),
309 309 )
310 310
311 311 def getbundle(
312 312 self, source, heads=None, common=None, bundlecaps=None, **kwargs
313 313 ):
314 314 chunks = exchange.getbundlechunks(
315 315 self._repo,
316 316 source,
317 317 heads=heads,
318 318 common=common,
319 319 bundlecaps=bundlecaps,
320 320 **kwargs
321 321 )[1]
322 322 cb = util.chunkbuffer(chunks)
323 323
324 324 if exchange.bundle2requested(bundlecaps):
325 325 # When requesting a bundle2, getbundle returns a stream to make the
326 326 # wire level function happier. We need to build a proper object
327 327 # from it in local peer.
328 328 return bundle2.getunbundler(self.ui, cb)
329 329 else:
330 330 return changegroup.getunbundler(b'01', cb, None)
331 331
332 332 def heads(self):
333 333 return self._repo.heads()
334 334
335 335 def known(self, nodes):
336 336 return self._repo.known(nodes)
337 337
338 338 def listkeys(self, namespace):
339 339 return self._repo.listkeys(namespace)
340 340
341 341 def lookup(self, key):
342 342 return self._repo.lookup(key)
343 343
344 344 def pushkey(self, namespace, key, old, new):
345 345 return self._repo.pushkey(namespace, key, old, new)
346 346
347 347 def stream_out(self):
348 348 raise error.Abort(_(b'cannot perform stream clone against local peer'))
349 349
350 350 def unbundle(self, bundle, heads, url):
351 351 """apply a bundle on a repo
352 352
353 353 This function handles the repo locking itself."""
354 354 try:
355 355 try:
356 356 bundle = exchange.readbundle(self.ui, bundle, None)
357 357 ret = exchange.unbundle(self._repo, bundle, heads, b'push', url)
358 358 if util.safehasattr(ret, b'getchunks'):
359 359 # This is a bundle20 object, turn it into an unbundler.
360 360 # This little dance should be dropped eventually when the
361 361 # API is finally improved.
362 362 stream = util.chunkbuffer(ret.getchunks())
363 363 ret = bundle2.getunbundler(self.ui, stream)
364 364 return ret
365 365 except Exception as exc:
366 366 # If the exception contains output salvaged from a bundle2
367 367 # reply, we need to make sure it is printed before continuing
368 368 # to fail. So we build a bundle2 with such output and consume
369 369 # it directly.
370 370 #
371 371 # This is not very elegant but allows a "simple" solution for
372 372 # issue4594
373 373 output = getattr(exc, '_bundle2salvagedoutput', ())
374 374 if output:
375 375 bundler = bundle2.bundle20(self._repo.ui)
376 376 for out in output:
377 377 bundler.addpart(out)
378 378 stream = util.chunkbuffer(bundler.getchunks())
379 379 b = bundle2.getunbundler(self.ui, stream)
380 380 bundle2.processbundle(self._repo, b)
381 381 raise
382 382 except error.PushRaced as exc:
383 383 raise error.ResponseError(
384 384 _(b'push failed:'), stringutil.forcebytestr(exc)
385 385 )
386 386
387 387 # End of _basewirecommands interface.
388 388
389 389 # Begin of peer interface.
390 390
391 391 def commandexecutor(self):
392 392 return localcommandexecutor(self)
393 393
394 394 # End of peer interface.
395 395
396 396
397 397 @interfaceutil.implementer(repository.ipeerlegacycommands)
398 398 class locallegacypeer(localpeer):
399 399 '''peer extension which implements legacy methods too; used for tests with
400 400 restricted capabilities'''
401 401
402 402 def __init__(self, repo):
403 403 super(locallegacypeer, self).__init__(repo, caps=legacycaps)
404 404
405 405 # Begin of baselegacywirecommands interface.
406 406
407 407 def between(self, pairs):
408 408 return self._repo.between(pairs)
409 409
410 410 def branches(self, nodes):
411 411 return self._repo.branches(nodes)
412 412
413 413 def changegroup(self, nodes, source):
414 414 outgoing = discovery.outgoing(
415 415 self._repo, missingroots=nodes, missingheads=self._repo.heads()
416 416 )
417 417 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
418 418
419 419 def changegroupsubset(self, bases, heads, source):
420 420 outgoing = discovery.outgoing(
421 421 self._repo, missingroots=bases, missingheads=heads
422 422 )
423 423 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
424 424
425 425 # End of baselegacywirecommands interface.
426 426
427 427
428 428 # Increment the sub-version when the revlog v2 format changes to lock out old
429 429 # clients.
430 430 REVLOGV2_REQUIREMENT = b'exp-revlogv2.1'
431 431
432 432 # A repository with the sparserevlog feature will have delta chains that
433 433 # can spread over a larger span. Sparse reading cuts these large spans into
434 434 # pieces, so that each piece isn't too big.
435 435 # Without the sparserevlog capability, reading from the repository could use
436 436 # huge amounts of memory, because the whole span would be read at once,
437 437 # including all the intermediate revisions that aren't pertinent for the chain.
438 438 # This is why once a repository has enabled sparse-read, it becomes required.
439 439 SPARSEREVLOG_REQUIREMENT = b'sparserevlog'
440 440
441 441 # A repository with the sidedataflag requirement will allow to store extra
442 442 # information for revision without altering their original hashes.
443 443 SIDEDATA_REQUIREMENT = b'exp-sidedata-flag'
444 444
445 445 # A repository with the the copies-sidedata-changeset requirement will store
446 446 # copies related information in changeset's sidedata.
447 447 COPIESSDC_REQUIREMENT = b'exp-copies-sidedata-changeset'
448 448
449 449 # The repository use persistent nodemap for the changelog and the manifest.
450 450 NODEMAP_REQUIREMENT = b'persistent-nodemap'
451 451
452 452 # Functions receiving (ui, features) that extensions can register to impact
453 453 # the ability to load repositories with custom requirements. Only
454 454 # functions defined in loaded extensions are called.
455 455 #
456 456 # The function receives a set of requirement strings that the repository
457 457 # is capable of opening. Functions will typically add elements to the
458 458 # set to reflect that the extension knows how to handle that requirements.
459 459 featuresetupfuncs = set()
460 460
461 461
462 462 def makelocalrepository(baseui, path, intents=None):
463 463 """Create a local repository object.
464 464
465 465 Given arguments needed to construct a local repository, this function
466 466 performs various early repository loading functionality (such as
467 467 reading the ``.hg/requires`` and ``.hg/hgrc`` files), validates that
468 468 the repository can be opened, derives a type suitable for representing
469 469 that repository, and returns an instance of it.
470 470
471 471 The returned object conforms to the ``repository.completelocalrepository``
472 472 interface.
473 473
474 474 The repository type is derived by calling a series of factory functions
475 475 for each aspect/interface of the final repository. These are defined by
476 476 ``REPO_INTERFACES``.
477 477
478 478 Each factory function is called to produce a type implementing a specific
479 479 interface. The cumulative list of returned types will be combined into a
480 480 new type and that type will be instantiated to represent the local
481 481 repository.
482 482
483 483 The factory functions each receive various state that may be consulted
484 484 as part of deriving a type.
485 485
486 486 Extensions should wrap these factory functions to customize repository type
487 487 creation. Note that an extension's wrapped function may be called even if
488 488 that extension is not loaded for the repo being constructed. Extensions
489 489 should check if their ``__name__`` appears in the
490 490 ``extensionmodulenames`` set passed to the factory function and no-op if
491 491 not.
492 492 """
493 493 ui = baseui.copy()
494 494 # Prevent copying repo configuration.
495 495 ui.copy = baseui.copy
496 496
497 497 # Working directory VFS rooted at repository root.
498 498 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
499 499
500 500 # Main VFS for .hg/ directory.
501 501 hgpath = wdirvfs.join(b'.hg')
502 502 hgvfs = vfsmod.vfs(hgpath, cacheaudited=True)
503 503
504 504 # The .hg/ path should exist and should be a directory. All other
505 505 # cases are errors.
506 506 if not hgvfs.isdir():
507 507 try:
508 508 hgvfs.stat()
509 509 except OSError as e:
510 510 if e.errno != errno.ENOENT:
511 511 raise
512 512 except ValueError as e:
513 513 # Can be raised on Python 3.8 when path is invalid.
514 514 raise error.Abort(
515 515 _(b'invalid path %s: %s') % (path, pycompat.bytestr(e))
516 516 )
517 517
518 518 raise error.RepoError(_(b'repository %s not found') % path)
519 519
520 520 # .hg/requires file contains a newline-delimited list of
521 521 # features/capabilities the opener (us) must have in order to use
522 522 # the repository. This file was introduced in Mercurial 0.9.2,
523 523 # which means very old repositories may not have one. We assume
524 524 # a missing file translates to no requirements.
525 525 try:
526 526 requirements = set(hgvfs.read(b'requires').splitlines())
527 527 except IOError as e:
528 528 if e.errno != errno.ENOENT:
529 529 raise
530 530 requirements = set()
531 531
532 532 # The .hg/hgrc file may load extensions or contain config options
533 533 # that influence repository construction. Attempt to load it and
534 534 # process any new extensions that it may have pulled in.
535 535 if loadhgrc(ui, wdirvfs, hgvfs, requirements):
536 536 afterhgrcload(ui, wdirvfs, hgvfs, requirements)
537 537 extensions.loadall(ui)
538 538 extensions.populateui(ui)
539 539
540 540 # Set of module names of extensions loaded for this repository.
541 541 extensionmodulenames = {m.__name__ for n, m in extensions.extensions(ui)}
542 542
543 543 supportedrequirements = gathersupportedrequirements(ui)
544 544
545 545 # We first validate the requirements are known.
546 546 ensurerequirementsrecognized(requirements, supportedrequirements)
547 547
548 548 # Then we validate that the known set is reasonable to use together.
549 549 ensurerequirementscompatible(ui, requirements)
550 550
551 551 # TODO there are unhandled edge cases related to opening repositories with
552 552 # shared storage. If storage is shared, we should also test for requirements
553 553 # compatibility in the pointed-to repo. This entails loading the .hg/hgrc in
554 554 # that repo, as that repo may load extensions needed to open it. This is a
555 555 # bit complicated because we don't want the other hgrc to overwrite settings
556 556 # in this hgrc.
557 557 #
558 558 # This bug is somewhat mitigated by the fact that we copy the .hg/requires
559 559 # file when sharing repos. But if a requirement is added after the share is
560 560 # performed, thereby introducing a new requirement for the opener, we may
561 561 # will not see that and could encounter a run-time error interacting with
562 562 # that shared store since it has an unknown-to-us requirement.
563 563
564 564 # At this point, we know we should be capable of opening the repository.
565 565 # Now get on with doing that.
566 566
567 567 features = set()
568 568
569 569 # The "store" part of the repository holds versioned data. How it is
570 570 # accessed is determined by various requirements. The ``shared`` or
571 571 # ``relshared`` requirements indicate the store lives in the path contained
572 572 # in the ``.hg/sharedpath`` file. This is an absolute path for
573 573 # ``shared`` and relative to ``.hg/`` for ``relshared``.
574 574 if b'shared' in requirements or b'relshared' in requirements:
575 575 sharedpath = hgvfs.read(b'sharedpath').rstrip(b'\n')
576 576 if b'relshared' in requirements:
577 577 sharedpath = hgvfs.join(sharedpath)
578 578
579 579 sharedvfs = vfsmod.vfs(sharedpath, realpath=True)
580 580
581 581 if not sharedvfs.exists():
582 582 raise error.RepoError(
583 583 _(b'.hg/sharedpath points to nonexistent directory %s')
584 584 % sharedvfs.base
585 585 )
586 586
587 587 features.add(repository.REPO_FEATURE_SHARED_STORAGE)
588 588
589 589 storebasepath = sharedvfs.base
590 590 cachepath = sharedvfs.join(b'cache')
591 591 else:
592 592 storebasepath = hgvfs.base
593 593 cachepath = hgvfs.join(b'cache')
594 594 wcachepath = hgvfs.join(b'wcache')
595 595
596 596 # The store has changed over time and the exact layout is dictated by
597 597 # requirements. The store interface abstracts differences across all
598 598 # of them.
599 599 store = makestore(
600 600 requirements,
601 601 storebasepath,
602 602 lambda base: vfsmod.vfs(base, cacheaudited=True),
603 603 )
604 604 hgvfs.createmode = store.createmode
605 605
606 606 storevfs = store.vfs
607 607 storevfs.options = resolvestorevfsoptions(ui, requirements, features)
608 608
609 609 # The cache vfs is used to manage cache files.
610 610 cachevfs = vfsmod.vfs(cachepath, cacheaudited=True)
611 611 cachevfs.createmode = store.createmode
612 612 # The cache vfs is used to manage cache files related to the working copy
613 613 wcachevfs = vfsmod.vfs(wcachepath, cacheaudited=True)
614 614 wcachevfs.createmode = store.createmode
615 615
616 616 # Now resolve the type for the repository object. We do this by repeatedly
617 617 # calling a factory function to produces types for specific aspects of the
618 618 # repo's operation. The aggregate returned types are used as base classes
619 619 # for a dynamically-derived type, which will represent our new repository.
620 620
621 621 bases = []
622 622 extrastate = {}
623 623
624 624 for iface, fn in REPO_INTERFACES:
625 625 # We pass all potentially useful state to give extensions tons of
626 626 # flexibility.
627 627 typ = fn()(
628 628 ui=ui,
629 629 intents=intents,
630 630 requirements=requirements,
631 631 features=features,
632 632 wdirvfs=wdirvfs,
633 633 hgvfs=hgvfs,
634 634 store=store,
635 635 storevfs=storevfs,
636 636 storeoptions=storevfs.options,
637 637 cachevfs=cachevfs,
638 638 wcachevfs=wcachevfs,
639 639 extensionmodulenames=extensionmodulenames,
640 640 extrastate=extrastate,
641 641 baseclasses=bases,
642 642 )
643 643
644 644 if not isinstance(typ, type):
645 645 raise error.ProgrammingError(
646 646 b'unable to construct type for %s' % iface
647 647 )
648 648
649 649 bases.append(typ)
650 650
651 651 # type() allows you to use characters in type names that wouldn't be
652 652 # recognized as Python symbols in source code. We abuse that to add
653 653 # rich information about our constructed repo.
654 654 name = pycompat.sysstr(
655 655 b'derivedrepo:%s<%s>' % (wdirvfs.base, b','.join(sorted(requirements)))
656 656 )
657 657
658 658 cls = type(name, tuple(bases), {})
659 659
660 660 return cls(
661 661 baseui=baseui,
662 662 ui=ui,
663 663 origroot=path,
664 664 wdirvfs=wdirvfs,
665 665 hgvfs=hgvfs,
666 666 requirements=requirements,
667 667 supportedrequirements=supportedrequirements,
668 668 sharedpath=storebasepath,
669 669 store=store,
670 670 cachevfs=cachevfs,
671 671 wcachevfs=wcachevfs,
672 672 features=features,
673 673 intents=intents,
674 674 )
675 675
676 676
677 677 def loadhgrc(ui, wdirvfs, hgvfs, requirements):
678 678 """Load hgrc files/content into a ui instance.
679 679
680 680 This is called during repository opening to load any additional
681 681 config files or settings relevant to the current repository.
682 682
683 683 Returns a bool indicating whether any additional configs were loaded.
684 684
685 685 Extensions should monkeypatch this function to modify how per-repo
686 686 configs are loaded. For example, an extension may wish to pull in
687 687 configs from alternate files or sources.
688 688 """
689 689 if not rcutil.use_repo_hgrc():
690 690 return False
691 691 try:
692 692 ui.readconfig(hgvfs.join(b'hgrc'), root=wdirvfs.base)
693 693 return True
694 694 except IOError:
695 695 return False
696 696
697 697
698 698 def afterhgrcload(ui, wdirvfs, hgvfs, requirements):
699 699 """Perform additional actions after .hg/hgrc is loaded.
700 700
701 701 This function is called during repository loading immediately after
702 702 the .hg/hgrc file is loaded and before per-repo extensions are loaded.
703 703
704 704 The function can be used to validate configs, automatically add
705 705 options (including extensions) based on requirements, etc.
706 706 """
707 707
708 708 # Map of requirements to list of extensions to load automatically when
709 709 # requirement is present.
710 710 autoextensions = {
711 711 b'git': [b'git'],
712 712 b'largefiles': [b'largefiles'],
713 713 b'lfs': [b'lfs'],
714 714 }
715 715
716 716 for requirement, names in sorted(autoextensions.items()):
717 717 if requirement not in requirements:
718 718 continue
719 719
720 720 for name in names:
721 721 if not ui.hasconfig(b'extensions', name):
722 722 ui.setconfig(b'extensions', name, b'', source=b'autoload')
723 723
724 724
725 725 def gathersupportedrequirements(ui):
726 726 """Determine the complete set of recognized requirements."""
727 727 # Start with all requirements supported by this file.
728 728 supported = set(localrepository._basesupported)
729 729
730 730 # Execute ``featuresetupfuncs`` entries if they belong to an extension
731 731 # relevant to this ui instance.
732 732 modules = {m.__name__ for n, m in extensions.extensions(ui)}
733 733
734 734 for fn in featuresetupfuncs:
735 735 if fn.__module__ in modules:
736 736 fn(ui, supported)
737 737
738 738 # Add derived requirements from registered compression engines.
739 739 for name in util.compengines:
740 740 engine = util.compengines[name]
741 741 if engine.available() and engine.revlogheader():
742 742 supported.add(b'exp-compression-%s' % name)
743 743 if engine.name() == b'zstd':
744 744 supported.add(b'revlog-compression-zstd')
745 745
746 746 return supported
747 747
748 748
749 749 def ensurerequirementsrecognized(requirements, supported):
750 750 """Validate that a set of local requirements is recognized.
751 751
752 752 Receives a set of requirements. Raises an ``error.RepoError`` if there
753 753 exists any requirement in that set that currently loaded code doesn't
754 754 recognize.
755 755
756 756 Returns a set of supported requirements.
757 757 """
758 758 missing = set()
759 759
760 760 for requirement in requirements:
761 761 if requirement in supported:
762 762 continue
763 763
764 764 if not requirement or not requirement[0:1].isalnum():
765 765 raise error.RequirementError(_(b'.hg/requires file is corrupt'))
766 766
767 767 missing.add(requirement)
768 768
769 769 if missing:
770 770 raise error.RequirementError(
771 771 _(b'repository requires features unknown to this Mercurial: %s')
772 772 % b' '.join(sorted(missing)),
773 773 hint=_(
774 774 b'see https://mercurial-scm.org/wiki/MissingRequirement '
775 775 b'for more information'
776 776 ),
777 777 )
778 778
779 779
780 780 def ensurerequirementscompatible(ui, requirements):
781 781 """Validates that a set of recognized requirements is mutually compatible.
782 782
783 783 Some requirements may not be compatible with others or require
784 784 config options that aren't enabled. This function is called during
785 785 repository opening to ensure that the set of requirements needed
786 786 to open a repository is sane and compatible with config options.
787 787
788 788 Extensions can monkeypatch this function to perform additional
789 789 checking.
790 790
791 791 ``error.RepoError`` should be raised on failure.
792 792 """
793 793 if b'exp-sparse' in requirements and not sparse.enabled:
794 794 raise error.RepoError(
795 795 _(
796 796 b'repository is using sparse feature but '
797 797 b'sparse is not enabled; enable the '
798 798 b'"sparse" extensions to access'
799 799 )
800 800 )
801 801
802 802
803 803 def makestore(requirements, path, vfstype):
804 804 """Construct a storage object for a repository."""
805 805 if b'store' in requirements:
806 806 if b'fncache' in requirements:
807 807 return storemod.fncachestore(
808 808 path, vfstype, b'dotencode' in requirements
809 809 )
810 810
811 811 return storemod.encodedstore(path, vfstype)
812 812
813 813 return storemod.basicstore(path, vfstype)
814 814
815 815
816 816 def resolvestorevfsoptions(ui, requirements, features):
817 817 """Resolve the options to pass to the store vfs opener.
818 818
819 819 The returned dict is used to influence behavior of the storage layer.
820 820 """
821 821 options = {}
822 822
823 823 if b'treemanifest' in requirements:
824 824 options[b'treemanifest'] = True
825 825
826 826 # experimental config: format.manifestcachesize
827 827 manifestcachesize = ui.configint(b'format', b'manifestcachesize')
828 828 if manifestcachesize is not None:
829 829 options[b'manifestcachesize'] = manifestcachesize
830 830
831 831 # In the absence of another requirement superseding a revlog-related
832 832 # requirement, we have to assume the repo is using revlog version 0.
833 833 # This revlog format is super old and we don't bother trying to parse
834 834 # opener options for it because those options wouldn't do anything
835 835 # meaningful on such old repos.
836 836 if b'revlogv1' in requirements or REVLOGV2_REQUIREMENT in requirements:
837 837 options.update(resolverevlogstorevfsoptions(ui, requirements, features))
838 838 else: # explicitly mark repo as using revlogv0
839 839 options[b'revlogv0'] = True
840 840
841 841 if COPIESSDC_REQUIREMENT in requirements:
842 842 options[b'copies-storage'] = b'changeset-sidedata'
843 843 else:
844 844 writecopiesto = ui.config(b'experimental', b'copies.write-to')
845 845 copiesextramode = (b'changeset-only', b'compatibility')
846 846 if writecopiesto in copiesextramode:
847 847 options[b'copies-storage'] = b'extra'
848 848
849 849 return options
850 850
851 851
852 852 def resolverevlogstorevfsoptions(ui, requirements, features):
853 853 """Resolve opener options specific to revlogs."""
854 854
855 855 options = {}
856 856 options[b'flagprocessors'] = {}
857 857
858 858 if b'revlogv1' in requirements:
859 859 options[b'revlogv1'] = True
860 860 if REVLOGV2_REQUIREMENT in requirements:
861 861 options[b'revlogv2'] = True
862 862
863 863 if b'generaldelta' in requirements:
864 864 options[b'generaldelta'] = True
865 865
866 866 # experimental config: format.chunkcachesize
867 867 chunkcachesize = ui.configint(b'format', b'chunkcachesize')
868 868 if chunkcachesize is not None:
869 869 options[b'chunkcachesize'] = chunkcachesize
870 870
871 871 deltabothparents = ui.configbool(
872 872 b'storage', b'revlog.optimize-delta-parent-choice'
873 873 )
874 874 options[b'deltabothparents'] = deltabothparents
875 875
876 876 lazydelta = ui.configbool(b'storage', b'revlog.reuse-external-delta')
877 877 lazydeltabase = False
878 878 if lazydelta:
879 879 lazydeltabase = ui.configbool(
880 880 b'storage', b'revlog.reuse-external-delta-parent'
881 881 )
882 882 if lazydeltabase is None:
883 883 lazydeltabase = not scmutil.gddeltaconfig(ui)
884 884 options[b'lazydelta'] = lazydelta
885 885 options[b'lazydeltabase'] = lazydeltabase
886 886
887 887 chainspan = ui.configbytes(b'experimental', b'maxdeltachainspan')
888 888 if 0 <= chainspan:
889 889 options[b'maxdeltachainspan'] = chainspan
890 890
891 891 mmapindexthreshold = ui.configbytes(b'experimental', b'mmapindexthreshold')
892 892 if mmapindexthreshold is not None:
893 893 options[b'mmapindexthreshold'] = mmapindexthreshold
894 894
895 895 withsparseread = ui.configbool(b'experimental', b'sparse-read')
896 896 srdensitythres = float(
897 897 ui.config(b'experimental', b'sparse-read.density-threshold')
898 898 )
899 899 srmingapsize = ui.configbytes(b'experimental', b'sparse-read.min-gap-size')
900 900 options[b'with-sparse-read'] = withsparseread
901 901 options[b'sparse-read-density-threshold'] = srdensitythres
902 902 options[b'sparse-read-min-gap-size'] = srmingapsize
903 903
904 904 sparserevlog = SPARSEREVLOG_REQUIREMENT in requirements
905 905 options[b'sparse-revlog'] = sparserevlog
906 906 if sparserevlog:
907 907 options[b'generaldelta'] = True
908 908
909 909 sidedata = SIDEDATA_REQUIREMENT in requirements
910 910 options[b'side-data'] = sidedata
911 911
912 912 maxchainlen = None
913 913 if sparserevlog:
914 914 maxchainlen = revlogconst.SPARSE_REVLOG_MAX_CHAIN_LENGTH
915 915 # experimental config: format.maxchainlen
916 916 maxchainlen = ui.configint(b'format', b'maxchainlen', maxchainlen)
917 917 if maxchainlen is not None:
918 918 options[b'maxchainlen'] = maxchainlen
919 919
920 920 for r in requirements:
921 921 # we allow multiple compression engine requirement to co-exist because
922 922 # strickly speaking, revlog seems to support mixed compression style.
923 923 #
924 924 # The compression used for new entries will be "the last one"
925 925 prefix = r.startswith
926 926 if prefix(b'revlog-compression-') or prefix(b'exp-compression-'):
927 927 options[b'compengine'] = r.split(b'-', 2)[2]
928 928
929 929 options[b'zlib.level'] = ui.configint(b'storage', b'revlog.zlib.level')
930 930 if options[b'zlib.level'] is not None:
931 931 if not (0 <= options[b'zlib.level'] <= 9):
932 932 msg = _(b'invalid value for `storage.revlog.zlib.level` config: %d')
933 933 raise error.Abort(msg % options[b'zlib.level'])
934 934 options[b'zstd.level'] = ui.configint(b'storage', b'revlog.zstd.level')
935 935 if options[b'zstd.level'] is not None:
936 936 if not (0 <= options[b'zstd.level'] <= 22):
937 937 msg = _(b'invalid value for `storage.revlog.zstd.level` config: %d')
938 938 raise error.Abort(msg % options[b'zstd.level'])
939 939
940 940 if repository.NARROW_REQUIREMENT in requirements:
941 941 options[b'enableellipsis'] = True
942 942
943 943 if ui.configbool(b'experimental', b'rust.index'):
944 944 options[b'rust.index'] = True
945 945 if NODEMAP_REQUIREMENT in requirements:
946 946 options[b'persistent-nodemap'] = True
947 947 if ui.configbool(b'storage', b'revlog.nodemap.mmap'):
948 948 options[b'persistent-nodemap.mmap'] = True
949 949 epnm = ui.config(b'storage', b'revlog.nodemap.mode')
950 950 options[b'persistent-nodemap.mode'] = epnm
951 951 if ui.configbool(b'devel', b'persistent-nodemap'):
952 952 options[b'devel-force-nodemap'] = True
953 953
954 954 return options
955 955
956 956
957 957 def makemain(**kwargs):
958 958 """Produce a type conforming to ``ilocalrepositorymain``."""
959 959 return localrepository
960 960
961 961
962 962 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
963 963 class revlogfilestorage(object):
964 964 """File storage when using revlogs."""
965 965
966 966 def file(self, path):
967 967 if path[0] == b'/':
968 968 path = path[1:]
969 969
970 970 return filelog.filelog(self.svfs, path)
971 971
972 972
973 973 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
974 974 class revlognarrowfilestorage(object):
975 975 """File storage when using revlogs and narrow files."""
976 976
977 977 def file(self, path):
978 978 if path[0] == b'/':
979 979 path = path[1:]
980 980
981 981 return filelog.narrowfilelog(self.svfs, path, self._storenarrowmatch)
982 982
983 983
984 984 def makefilestorage(requirements, features, **kwargs):
985 985 """Produce a type conforming to ``ilocalrepositoryfilestorage``."""
986 986 features.add(repository.REPO_FEATURE_REVLOG_FILE_STORAGE)
987 987 features.add(repository.REPO_FEATURE_STREAM_CLONE)
988 988
989 989 if repository.NARROW_REQUIREMENT in requirements:
990 990 return revlognarrowfilestorage
991 991 else:
992 992 return revlogfilestorage
993 993
994 994
995 995 # List of repository interfaces and factory functions for them. Each
996 996 # will be called in order during ``makelocalrepository()`` to iteratively
997 997 # derive the final type for a local repository instance. We capture the
998 998 # function as a lambda so we don't hold a reference and the module-level
999 999 # functions can be wrapped.
1000 1000 REPO_INTERFACES = [
1001 1001 (repository.ilocalrepositorymain, lambda: makemain),
1002 1002 (repository.ilocalrepositoryfilestorage, lambda: makefilestorage),
1003 1003 ]
1004 1004
1005 1005
1006 1006 @interfaceutil.implementer(repository.ilocalrepositorymain)
1007 1007 class localrepository(object):
1008 1008 """Main class for representing local repositories.
1009 1009
1010 1010 All local repositories are instances of this class.
1011 1011
1012 1012 Constructed on its own, instances of this class are not usable as
1013 1013 repository objects. To obtain a usable repository object, call
1014 1014 ``hg.repository()``, ``localrepo.instance()``, or
1015 1015 ``localrepo.makelocalrepository()``. The latter is the lowest-level.
1016 1016 ``instance()`` adds support for creating new repositories.
1017 1017 ``hg.repository()`` adds more extension integration, including calling
1018 1018 ``reposetup()``. Generally speaking, ``hg.repository()`` should be
1019 1019 used.
1020 1020 """
1021 1021
1022 1022 # obsolete experimental requirements:
1023 1023 # - manifestv2: An experimental new manifest format that allowed
1024 1024 # for stem compression of long paths. Experiment ended up not
1025 1025 # being successful (repository sizes went up due to worse delta
1026 1026 # chains), and the code was deleted in 4.6.
1027 1027 supportedformats = {
1028 1028 b'revlogv1',
1029 1029 b'generaldelta',
1030 1030 b'treemanifest',
1031 1031 COPIESSDC_REQUIREMENT,
1032 1032 REVLOGV2_REQUIREMENT,
1033 1033 SIDEDATA_REQUIREMENT,
1034 1034 SPARSEREVLOG_REQUIREMENT,
1035 1035 NODEMAP_REQUIREMENT,
1036 1036 bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT,
1037 1037 }
1038 1038 _basesupported = supportedformats | {
1039 1039 b'store',
1040 1040 b'fncache',
1041 1041 b'shared',
1042 1042 b'relshared',
1043 1043 b'dotencode',
1044 1044 b'exp-sparse',
1045 1045 b'internal-phase',
1046 1046 }
1047 1047
1048 1048 # list of prefix for file which can be written without 'wlock'
1049 1049 # Extensions should extend this list when needed
1050 1050 _wlockfreeprefix = {
1051 1051 # We migh consider requiring 'wlock' for the next
1052 1052 # two, but pretty much all the existing code assume
1053 1053 # wlock is not needed so we keep them excluded for
1054 1054 # now.
1055 1055 b'hgrc',
1056 1056 b'requires',
1057 1057 # XXX cache is a complicatged business someone
1058 1058 # should investigate this in depth at some point
1059 1059 b'cache/',
1060 1060 # XXX shouldn't be dirstate covered by the wlock?
1061 1061 b'dirstate',
1062 1062 # XXX bisect was still a bit too messy at the time
1063 1063 # this changeset was introduced. Someone should fix
1064 1064 # the remainig bit and drop this line
1065 1065 b'bisect.state',
1066 1066 }
1067 1067
1068 1068 def __init__(
1069 1069 self,
1070 1070 baseui,
1071 1071 ui,
1072 1072 origroot,
1073 1073 wdirvfs,
1074 1074 hgvfs,
1075 1075 requirements,
1076 1076 supportedrequirements,
1077 1077 sharedpath,
1078 1078 store,
1079 1079 cachevfs,
1080 1080 wcachevfs,
1081 1081 features,
1082 1082 intents=None,
1083 1083 ):
1084 1084 """Create a new local repository instance.
1085 1085
1086 1086 Most callers should use ``hg.repository()``, ``localrepo.instance()``,
1087 1087 or ``localrepo.makelocalrepository()`` for obtaining a new repository
1088 1088 object.
1089 1089
1090 1090 Arguments:
1091 1091
1092 1092 baseui
1093 1093 ``ui.ui`` instance that ``ui`` argument was based off of.
1094 1094
1095 1095 ui
1096 1096 ``ui.ui`` instance for use by the repository.
1097 1097
1098 1098 origroot
1099 1099 ``bytes`` path to working directory root of this repository.
1100 1100
1101 1101 wdirvfs
1102 1102 ``vfs.vfs`` rooted at the working directory.
1103 1103
1104 1104 hgvfs
1105 1105 ``vfs.vfs`` rooted at .hg/
1106 1106
1107 1107 requirements
1108 1108 ``set`` of bytestrings representing repository opening requirements.
1109 1109
1110 1110 supportedrequirements
1111 1111 ``set`` of bytestrings representing repository requirements that we
1112 1112 know how to open. May be a supetset of ``requirements``.
1113 1113
1114 1114 sharedpath
1115 1115 ``bytes`` Defining path to storage base directory. Points to a
1116 1116 ``.hg/`` directory somewhere.
1117 1117
1118 1118 store
1119 1119 ``store.basicstore`` (or derived) instance providing access to
1120 1120 versioned storage.
1121 1121
1122 1122 cachevfs
1123 1123 ``vfs.vfs`` used for cache files.
1124 1124
1125 1125 wcachevfs
1126 1126 ``vfs.vfs`` used for cache files related to the working copy.
1127 1127
1128 1128 features
1129 1129 ``set`` of bytestrings defining features/capabilities of this
1130 1130 instance.
1131 1131
1132 1132 intents
1133 1133 ``set`` of system strings indicating what this repo will be used
1134 1134 for.
1135 1135 """
1136 1136 self.baseui = baseui
1137 1137 self.ui = ui
1138 1138 self.origroot = origroot
1139 1139 # vfs rooted at working directory.
1140 1140 self.wvfs = wdirvfs
1141 1141 self.root = wdirvfs.base
1142 1142 # vfs rooted at .hg/. Used to access most non-store paths.
1143 1143 self.vfs = hgvfs
1144 1144 self.path = hgvfs.base
1145 1145 self.requirements = requirements
1146 1146 self.supported = supportedrequirements
1147 1147 self.sharedpath = sharedpath
1148 1148 self.store = store
1149 1149 self.cachevfs = cachevfs
1150 1150 self.wcachevfs = wcachevfs
1151 1151 self.features = features
1152 1152
1153 1153 self.filtername = None
1154 1154
1155 1155 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1156 1156 b'devel', b'check-locks'
1157 1157 ):
1158 1158 self.vfs.audit = self._getvfsward(self.vfs.audit)
1159 1159 # A list of callback to shape the phase if no data were found.
1160 1160 # Callback are in the form: func(repo, roots) --> processed root.
1161 1161 # This list it to be filled by extension during repo setup
1162 1162 self._phasedefaults = []
1163 1163
1164 1164 color.setup(self.ui)
1165 1165
1166 1166 self.spath = self.store.path
1167 1167 self.svfs = self.store.vfs
1168 1168 self.sjoin = self.store.join
1169 1169 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1170 1170 b'devel', b'check-locks'
1171 1171 ):
1172 1172 if util.safehasattr(self.svfs, b'vfs'): # this is filtervfs
1173 1173 self.svfs.vfs.audit = self._getsvfsward(self.svfs.vfs.audit)
1174 1174 else: # standard vfs
1175 1175 self.svfs.audit = self._getsvfsward(self.svfs.audit)
1176 1176
1177 1177 self._dirstatevalidatewarned = False
1178 1178
1179 1179 self._branchcaches = branchmap.BranchMapCache()
1180 1180 self._revbranchcache = None
1181 1181 self._filterpats = {}
1182 1182 self._datafilters = {}
1183 1183 self._transref = self._lockref = self._wlockref = None
1184 1184
1185 1185 # A cache for various files under .hg/ that tracks file changes,
1186 1186 # (used by the filecache decorator)
1187 1187 #
1188 1188 # Maps a property name to its util.filecacheentry
1189 1189 self._filecache = {}
1190 1190
1191 1191 # hold sets of revision to be filtered
1192 1192 # should be cleared when something might have changed the filter value:
1193 1193 # - new changesets,
1194 1194 # - phase change,
1195 1195 # - new obsolescence marker,
1196 1196 # - working directory parent change,
1197 1197 # - bookmark changes
1198 1198 self.filteredrevcache = {}
1199 1199
1200 1200 # post-dirstate-status hooks
1201 1201 self._postdsstatus = []
1202 1202
1203 1203 # generic mapping between names and nodes
1204 1204 self.names = namespaces.namespaces()
1205 1205
1206 1206 # Key to signature value.
1207 1207 self._sparsesignaturecache = {}
1208 1208 # Signature to cached matcher instance.
1209 1209 self._sparsematchercache = {}
1210 1210
1211 1211 self._extrafilterid = repoview.extrafilter(ui)
1212 1212
1213 1213 self.filecopiesmode = None
1214 1214 if COPIESSDC_REQUIREMENT in self.requirements:
1215 1215 self.filecopiesmode = b'changeset-sidedata'
1216 1216
1217 1217 def _getvfsward(self, origfunc):
1218 1218 """build a ward for self.vfs"""
1219 1219 rref = weakref.ref(self)
1220 1220
1221 1221 def checkvfs(path, mode=None):
1222 1222 ret = origfunc(path, mode=mode)
1223 1223 repo = rref()
1224 1224 if (
1225 1225 repo is None
1226 1226 or not util.safehasattr(repo, b'_wlockref')
1227 1227 or not util.safehasattr(repo, b'_lockref')
1228 1228 ):
1229 1229 return
1230 1230 if mode in (None, b'r', b'rb'):
1231 1231 return
1232 1232 if path.startswith(repo.path):
1233 1233 # truncate name relative to the repository (.hg)
1234 1234 path = path[len(repo.path) + 1 :]
1235 1235 if path.startswith(b'cache/'):
1236 1236 msg = b'accessing cache with vfs instead of cachevfs: "%s"'
1237 1237 repo.ui.develwarn(msg % path, stacklevel=3, config=b"cache-vfs")
1238 1238 # path prefixes covered by 'lock'
1239 1239 vfs_path_prefixes = (b'journal.', b'undo.', b'strip-backup/')
1240 1240 if any(path.startswith(prefix) for prefix in vfs_path_prefixes):
1241 1241 if repo._currentlock(repo._lockref) is None:
1242 1242 repo.ui.develwarn(
1243 1243 b'write with no lock: "%s"' % path,
1244 1244 stacklevel=3,
1245 1245 config=b'check-locks',
1246 1246 )
1247 1247 elif repo._currentlock(repo._wlockref) is None:
1248 1248 # rest of vfs files are covered by 'wlock'
1249 1249 #
1250 1250 # exclude special files
1251 1251 for prefix in self._wlockfreeprefix:
1252 1252 if path.startswith(prefix):
1253 1253 return
1254 1254 repo.ui.develwarn(
1255 1255 b'write with no wlock: "%s"' % path,
1256 1256 stacklevel=3,
1257 1257 config=b'check-locks',
1258 1258 )
1259 1259 return ret
1260 1260
1261 1261 return checkvfs
1262 1262
1263 1263 def _getsvfsward(self, origfunc):
1264 1264 """build a ward for self.svfs"""
1265 1265 rref = weakref.ref(self)
1266 1266
1267 1267 def checksvfs(path, mode=None):
1268 1268 ret = origfunc(path, mode=mode)
1269 1269 repo = rref()
1270 1270 if repo is None or not util.safehasattr(repo, b'_lockref'):
1271 1271 return
1272 1272 if mode in (None, b'r', b'rb'):
1273 1273 return
1274 1274 if path.startswith(repo.sharedpath):
1275 1275 # truncate name relative to the repository (.hg)
1276 1276 path = path[len(repo.sharedpath) + 1 :]
1277 1277 if repo._currentlock(repo._lockref) is None:
1278 1278 repo.ui.develwarn(
1279 1279 b'write with no lock: "%s"' % path, stacklevel=4
1280 1280 )
1281 1281 return ret
1282 1282
1283 1283 return checksvfs
1284 1284
1285 1285 def close(self):
1286 1286 self._writecaches()
1287 1287
1288 1288 def _writecaches(self):
1289 1289 if self._revbranchcache:
1290 1290 self._revbranchcache.write()
1291 1291
1292 1292 def _restrictcapabilities(self, caps):
1293 1293 if self.ui.configbool(b'experimental', b'bundle2-advertise'):
1294 1294 caps = set(caps)
1295 1295 capsblob = bundle2.encodecaps(
1296 1296 bundle2.getrepocaps(self, role=b'client')
1297 1297 )
1298 1298 caps.add(b'bundle2=' + urlreq.quote(capsblob))
1299 1299 return caps
1300 1300
1301 1301 def _writerequirements(self):
1302 1302 scmutil.writerequires(self.vfs, self.requirements)
1303 1303
1304 1304 # Don't cache auditor/nofsauditor, or you'll end up with reference cycle:
1305 1305 # self -> auditor -> self._checknested -> self
1306 1306
1307 1307 @property
1308 1308 def auditor(self):
1309 1309 # This is only used by context.workingctx.match in order to
1310 1310 # detect files in subrepos.
1311 1311 return pathutil.pathauditor(self.root, callback=self._checknested)
1312 1312
1313 1313 @property
1314 1314 def nofsauditor(self):
1315 1315 # This is only used by context.basectx.match in order to detect
1316 1316 # files in subrepos.
1317 1317 return pathutil.pathauditor(
1318 1318 self.root, callback=self._checknested, realfs=False, cached=True
1319 1319 )
1320 1320
1321 1321 def _checknested(self, path):
1322 1322 """Determine if path is a legal nested repository."""
1323 1323 if not path.startswith(self.root):
1324 1324 return False
1325 1325 subpath = path[len(self.root) + 1 :]
1326 1326 normsubpath = util.pconvert(subpath)
1327 1327
1328 1328 # XXX: Checking against the current working copy is wrong in
1329 1329 # the sense that it can reject things like
1330 1330 #
1331 1331 # $ hg cat -r 10 sub/x.txt
1332 1332 #
1333 1333 # if sub/ is no longer a subrepository in the working copy
1334 1334 # parent revision.
1335 1335 #
1336 1336 # However, it can of course also allow things that would have
1337 1337 # been rejected before, such as the above cat command if sub/
1338 1338 # is a subrepository now, but was a normal directory before.
1339 1339 # The old path auditor would have rejected by mistake since it
1340 1340 # panics when it sees sub/.hg/.
1341 1341 #
1342 1342 # All in all, checking against the working copy seems sensible
1343 1343 # since we want to prevent access to nested repositories on
1344 1344 # the filesystem *now*.
1345 1345 ctx = self[None]
1346 1346 parts = util.splitpath(subpath)
1347 1347 while parts:
1348 1348 prefix = b'/'.join(parts)
1349 1349 if prefix in ctx.substate:
1350 1350 if prefix == normsubpath:
1351 1351 return True
1352 1352 else:
1353 1353 sub = ctx.sub(prefix)
1354 1354 return sub.checknested(subpath[len(prefix) + 1 :])
1355 1355 else:
1356 1356 parts.pop()
1357 1357 return False
1358 1358
1359 1359 def peer(self):
1360 1360 return localpeer(self) # not cached to avoid reference cycle
1361 1361
1362 1362 def unfiltered(self):
1363 1363 """Return unfiltered version of the repository
1364 1364
1365 1365 Intended to be overwritten by filtered repo."""
1366 1366 return self
1367 1367
1368 1368 def filtered(self, name, visibilityexceptions=None):
1369 1369 """Return a filtered version of a repository
1370 1370
1371 1371 The `name` parameter is the identifier of the requested view. This
1372 1372 will return a repoview object set "exactly" to the specified view.
1373 1373
1374 1374 This function does not apply recursive filtering to a repository. For
1375 1375 example calling `repo.filtered("served")` will return a repoview using
1376 1376 the "served" view, regardless of the initial view used by `repo`.
1377 1377
1378 1378 In other word, there is always only one level of `repoview` "filtering".
1379 1379 """
1380 1380 if self._extrafilterid is not None and b'%' not in name:
1381 1381 name = name + b'%' + self._extrafilterid
1382 1382
1383 1383 cls = repoview.newtype(self.unfiltered().__class__)
1384 1384 return cls(self, name, visibilityexceptions)
1385 1385
1386 1386 @mixedrepostorecache(
1387 1387 (b'bookmarks', b'plain'),
1388 1388 (b'bookmarks.current', b'plain'),
1389 1389 (b'bookmarks', b''),
1390 1390 (b'00changelog.i', b''),
1391 1391 )
1392 1392 def _bookmarks(self):
1393 1393 # Since the multiple files involved in the transaction cannot be
1394 1394 # written atomically (with current repository format), there is a race
1395 1395 # condition here.
1396 1396 #
1397 1397 # 1) changelog content A is read
1398 1398 # 2) outside transaction update changelog to content B
1399 1399 # 3) outside transaction update bookmark file referring to content B
1400 1400 # 4) bookmarks file content is read and filtered against changelog-A
1401 1401 #
1402 1402 # When this happens, bookmarks against nodes missing from A are dropped.
1403 1403 #
1404 1404 # Having this happening during read is not great, but it become worse
1405 1405 # when this happen during write because the bookmarks to the "unknown"
1406 1406 # nodes will be dropped for good. However, writes happen within locks.
1407 1407 # This locking makes it possible to have a race free consistent read.
1408 1408 # For this purpose data read from disc before locking are
1409 1409 # "invalidated" right after the locks are taken. This invalidations are
1410 1410 # "light", the `filecache` mechanism keep the data in memory and will
1411 1411 # reuse them if the underlying files did not changed. Not parsing the
1412 1412 # same data multiple times helps performances.
1413 1413 #
1414 1414 # Unfortunately in the case describe above, the files tracked by the
1415 1415 # bookmarks file cache might not have changed, but the in-memory
1416 1416 # content is still "wrong" because we used an older changelog content
1417 1417 # to process the on-disk data. So after locking, the changelog would be
1418 1418 # refreshed but `_bookmarks` would be preserved.
1419 1419 # Adding `00changelog.i` to the list of tracked file is not
1420 1420 # enough, because at the time we build the content for `_bookmarks` in
1421 1421 # (4), the changelog file has already diverged from the content used
1422 1422 # for loading `changelog` in (1)
1423 1423 #
1424 1424 # To prevent the issue, we force the changelog to be explicitly
1425 1425 # reloaded while computing `_bookmarks`. The data race can still happen
1426 1426 # without the lock (with a narrower window), but it would no longer go
1427 1427 # undetected during the lock time refresh.
1428 1428 #
1429 1429 # The new schedule is as follow
1430 1430 #
1431 1431 # 1) filecache logic detect that `_bookmarks` needs to be computed
1432 1432 # 2) cachestat for `bookmarks` and `changelog` are captured (for book)
1433 1433 # 3) We force `changelog` filecache to be tested
1434 1434 # 4) cachestat for `changelog` are captured (for changelog)
1435 1435 # 5) `_bookmarks` is computed and cached
1436 1436 #
1437 1437 # The step in (3) ensure we have a changelog at least as recent as the
1438 1438 # cache stat computed in (1). As a result at locking time:
1439 1439 # * if the changelog did not changed since (1) -> we can reuse the data
1440 1440 # * otherwise -> the bookmarks get refreshed.
1441 1441 self._refreshchangelog()
1442 1442 return bookmarks.bmstore(self)
1443 1443
1444 1444 def _refreshchangelog(self):
1445 1445 """make sure the in memory changelog match the on-disk one"""
1446 1446 if 'changelog' in vars(self) and self.currenttransaction() is None:
1447 1447 del self.changelog
1448 1448
1449 1449 @property
1450 1450 def _activebookmark(self):
1451 1451 return self._bookmarks.active
1452 1452
1453 1453 # _phasesets depend on changelog. what we need is to call
1454 1454 # _phasecache.invalidate() if '00changelog.i' was changed, but it
1455 1455 # can't be easily expressed in filecache mechanism.
1456 1456 @storecache(b'phaseroots', b'00changelog.i')
1457 1457 def _phasecache(self):
1458 1458 return phases.phasecache(self, self._phasedefaults)
1459 1459
1460 1460 @storecache(b'obsstore')
1461 1461 def obsstore(self):
1462 1462 return obsolete.makestore(self.ui, self)
1463 1463
1464 1464 @storecache(b'00changelog.i')
1465 1465 def changelog(self):
1466 1466 # load dirstate before changelog to avoid race see issue6303
1467 1467 self.dirstate.prefetch_parents()
1468 1468 return self.store.changelog(txnutil.mayhavepending(self.root))
1469 1469
1470 1470 @storecache(b'00manifest.i')
1471 1471 def manifestlog(self):
1472 1472 return self.store.manifestlog(self, self._storenarrowmatch)
1473 1473
1474 1474 @repofilecache(b'dirstate')
1475 1475 def dirstate(self):
1476 1476 return self._makedirstate()
1477 1477
1478 1478 def _makedirstate(self):
1479 1479 """Extension point for wrapping the dirstate per-repo."""
1480 1480 sparsematchfn = lambda: sparse.matcher(self)
1481 1481
1482 1482 return dirstate.dirstate(
1483 1483 self.vfs, self.ui, self.root, self._dirstatevalidate, sparsematchfn
1484 1484 )
1485 1485
1486 1486 def _dirstatevalidate(self, node):
1487 1487 try:
1488 1488 self.changelog.rev(node)
1489 1489 return node
1490 1490 except error.LookupError:
1491 1491 if not self._dirstatevalidatewarned:
1492 1492 self._dirstatevalidatewarned = True
1493 1493 self.ui.warn(
1494 1494 _(b"warning: ignoring unknown working parent %s!\n")
1495 1495 % short(node)
1496 1496 )
1497 1497 return nullid
1498 1498
1499 1499 @storecache(narrowspec.FILENAME)
1500 1500 def narrowpats(self):
1501 1501 """matcher patterns for this repository's narrowspec
1502 1502
1503 1503 A tuple of (includes, excludes).
1504 1504 """
1505 1505 return narrowspec.load(self)
1506 1506
1507 1507 @storecache(narrowspec.FILENAME)
1508 1508 def _storenarrowmatch(self):
1509 1509 if repository.NARROW_REQUIREMENT not in self.requirements:
1510 1510 return matchmod.always()
1511 1511 include, exclude = self.narrowpats
1512 1512 return narrowspec.match(self.root, include=include, exclude=exclude)
1513 1513
1514 1514 @storecache(narrowspec.FILENAME)
1515 1515 def _narrowmatch(self):
1516 1516 if repository.NARROW_REQUIREMENT not in self.requirements:
1517 1517 return matchmod.always()
1518 1518 narrowspec.checkworkingcopynarrowspec(self)
1519 1519 include, exclude = self.narrowpats
1520 1520 return narrowspec.match(self.root, include=include, exclude=exclude)
1521 1521
1522 1522 def narrowmatch(self, match=None, includeexact=False):
1523 1523 """matcher corresponding the the repo's narrowspec
1524 1524
1525 1525 If `match` is given, then that will be intersected with the narrow
1526 1526 matcher.
1527 1527
1528 1528 If `includeexact` is True, then any exact matches from `match` will
1529 1529 be included even if they're outside the narrowspec.
1530 1530 """
1531 1531 if match:
1532 1532 if includeexact and not self._narrowmatch.always():
1533 1533 # do not exclude explicitly-specified paths so that they can
1534 1534 # be warned later on
1535 1535 em = matchmod.exact(match.files())
1536 1536 nm = matchmod.unionmatcher([self._narrowmatch, em])
1537 1537 return matchmod.intersectmatchers(match, nm)
1538 1538 return matchmod.intersectmatchers(match, self._narrowmatch)
1539 1539 return self._narrowmatch
1540 1540
1541 1541 def setnarrowpats(self, newincludes, newexcludes):
1542 1542 narrowspec.save(self, newincludes, newexcludes)
1543 1543 self.invalidate(clearfilecache=True)
1544 1544
1545 1545 @unfilteredpropertycache
1546 1546 def _quick_access_changeid_null(self):
1547 1547 return {
1548 1548 b'null': (nullrev, nullid),
1549 1549 nullrev: (nullrev, nullid),
1550 1550 nullid: (nullrev, nullid),
1551 1551 }
1552 1552
1553 1553 @unfilteredpropertycache
1554 1554 def _quick_access_changeid_wc(self):
1555 1555 # also fast path access to the working copy parents
1556 1556 # however, only do it for filter that ensure wc is visible.
1557 1557 quick = {}
1558 1558 cl = self.unfiltered().changelog
1559 1559 for node in self.dirstate.parents():
1560 1560 if node == nullid:
1561 1561 continue
1562 1562 rev = cl.index.get_rev(node)
1563 1563 if rev is None:
1564 1564 # unknown working copy parent case:
1565 1565 #
1566 1566 # skip the fast path and let higher code deal with it
1567 1567 continue
1568 1568 pair = (rev, node)
1569 1569 quick[rev] = pair
1570 1570 quick[node] = pair
1571 1571 # also add the parents of the parents
1572 1572 for r in cl.parentrevs(rev):
1573 1573 if r == nullrev:
1574 1574 continue
1575 1575 n = cl.node(r)
1576 1576 pair = (r, n)
1577 1577 quick[r] = pair
1578 1578 quick[n] = pair
1579 1579 p1node = self.dirstate.p1()
1580 1580 if p1node != nullid:
1581 1581 quick[b'.'] = quick[p1node]
1582 1582 return quick
1583 1583
1584 1584 @unfilteredmethod
1585 1585 def _quick_access_changeid_invalidate(self):
1586 1586 if '_quick_access_changeid_wc' in vars(self):
1587 1587 del self.__dict__['_quick_access_changeid_wc']
1588 1588
1589 1589 @property
1590 1590 def _quick_access_changeid(self):
1591 1591 """an helper dictionnary for __getitem__ calls
1592 1592
1593 1593 This contains a list of symbol we can recognise right away without
1594 1594 further processing.
1595 1595 """
1596 1596 mapping = self._quick_access_changeid_null
1597 1597 if self.filtername in repoview.filter_has_wc:
1598 1598 mapping = mapping.copy()
1599 1599 mapping.update(self._quick_access_changeid_wc)
1600 1600 return mapping
1601 1601
1602 1602 def __getitem__(self, changeid):
1603 1603 # dealing with special cases
1604 1604 if changeid is None:
1605 1605 return context.workingctx(self)
1606 1606 if isinstance(changeid, context.basectx):
1607 1607 return changeid
1608 1608
1609 1609 # dealing with multiple revisions
1610 1610 if isinstance(changeid, slice):
1611 1611 # wdirrev isn't contiguous so the slice shouldn't include it
1612 1612 return [
1613 1613 self[i]
1614 1614 for i in pycompat.xrange(*changeid.indices(len(self)))
1615 1615 if i not in self.changelog.filteredrevs
1616 1616 ]
1617 1617
1618 1618 # dealing with some special values
1619 1619 quick_access = self._quick_access_changeid.get(changeid)
1620 1620 if quick_access is not None:
1621 1621 rev, node = quick_access
1622 1622 return context.changectx(self, rev, node, maybe_filtered=False)
1623 1623 if changeid == b'tip':
1624 1624 node = self.changelog.tip()
1625 1625 rev = self.changelog.rev(node)
1626 1626 return context.changectx(self, rev, node)
1627 1627
1628 1628 # dealing with arbitrary values
1629 1629 try:
1630 1630 if isinstance(changeid, int):
1631 1631 node = self.changelog.node(changeid)
1632 1632 rev = changeid
1633 1633 elif changeid == b'.':
1634 1634 # this is a hack to delay/avoid loading obsmarkers
1635 1635 # when we know that '.' won't be hidden
1636 1636 node = self.dirstate.p1()
1637 1637 rev = self.unfiltered().changelog.rev(node)
1638 1638 elif len(changeid) == 20:
1639 1639 try:
1640 1640 node = changeid
1641 1641 rev = self.changelog.rev(changeid)
1642 1642 except error.FilteredLookupError:
1643 1643 changeid = hex(changeid) # for the error message
1644 1644 raise
1645 1645 except LookupError:
1646 1646 # check if it might have come from damaged dirstate
1647 1647 #
1648 1648 # XXX we could avoid the unfiltered if we had a recognizable
1649 1649 # exception for filtered changeset access
1650 1650 if (
1651 1651 self.local()
1652 1652 and changeid in self.unfiltered().dirstate.parents()
1653 1653 ):
1654 1654 msg = _(b"working directory has unknown parent '%s'!")
1655 1655 raise error.Abort(msg % short(changeid))
1656 1656 changeid = hex(changeid) # for the error message
1657 1657 raise
1658 1658
1659 1659 elif len(changeid) == 40:
1660 1660 node = bin(changeid)
1661 1661 rev = self.changelog.rev(node)
1662 1662 else:
1663 1663 raise error.ProgrammingError(
1664 1664 b"unsupported changeid '%s' of type %s"
1665 1665 % (changeid, pycompat.bytestr(type(changeid)))
1666 1666 )
1667 1667
1668 1668 return context.changectx(self, rev, node)
1669 1669
1670 1670 except (error.FilteredIndexError, error.FilteredLookupError):
1671 1671 raise error.FilteredRepoLookupError(
1672 1672 _(b"filtered revision '%s'") % pycompat.bytestr(changeid)
1673 1673 )
1674 1674 except (IndexError, LookupError):
1675 1675 raise error.RepoLookupError(
1676 1676 _(b"unknown revision '%s'") % pycompat.bytestr(changeid)
1677 1677 )
1678 1678 except error.WdirUnsupported:
1679 1679 return context.workingctx(self)
1680 1680
1681 1681 def __contains__(self, changeid):
1682 1682 """True if the given changeid exists
1683 1683
1684 1684 error.AmbiguousPrefixLookupError is raised if an ambiguous node
1685 1685 specified.
1686 1686 """
1687 1687 try:
1688 1688 self[changeid]
1689 1689 return True
1690 1690 except error.RepoLookupError:
1691 1691 return False
1692 1692
1693 1693 def __nonzero__(self):
1694 1694 return True
1695 1695
1696 1696 __bool__ = __nonzero__
1697 1697
1698 1698 def __len__(self):
1699 1699 # no need to pay the cost of repoview.changelog
1700 1700 unfi = self.unfiltered()
1701 1701 return len(unfi.changelog)
1702 1702
1703 1703 def __iter__(self):
1704 1704 return iter(self.changelog)
1705 1705
1706 1706 def revs(self, expr, *args):
1707 1707 '''Find revisions matching a revset.
1708 1708
1709 1709 The revset is specified as a string ``expr`` that may contain
1710 1710 %-formatting to escape certain types. See ``revsetlang.formatspec``.
1711 1711
1712 1712 Revset aliases from the configuration are not expanded. To expand
1713 1713 user aliases, consider calling ``scmutil.revrange()`` or
1714 1714 ``repo.anyrevs([expr], user=True)``.
1715 1715
1716 1716 Returns a smartset.abstractsmartset, which is a list-like interface
1717 1717 that contains integer revisions.
1718 1718 '''
1719 1719 tree = revsetlang.spectree(expr, *args)
1720 1720 return revset.makematcher(tree)(self)
1721 1721
1722 1722 def set(self, expr, *args):
1723 1723 '''Find revisions matching a revset and emit changectx instances.
1724 1724
1725 1725 This is a convenience wrapper around ``revs()`` that iterates the
1726 1726 result and is a generator of changectx instances.
1727 1727
1728 1728 Revset aliases from the configuration are not expanded. To expand
1729 1729 user aliases, consider calling ``scmutil.revrange()``.
1730 1730 '''
1731 1731 for r in self.revs(expr, *args):
1732 1732 yield self[r]
1733 1733
1734 1734 def anyrevs(self, specs, user=False, localalias=None):
1735 1735 '''Find revisions matching one of the given revsets.
1736 1736
1737 1737 Revset aliases from the configuration are not expanded by default. To
1738 1738 expand user aliases, specify ``user=True``. To provide some local
1739 1739 definitions overriding user aliases, set ``localalias`` to
1740 1740 ``{name: definitionstring}``.
1741 1741 '''
1742 1742 if specs == [b'null']:
1743 1743 return revset.baseset([nullrev])
1744 1744 if specs == [b'.']:
1745 1745 quick_data = self._quick_access_changeid.get(b'.')
1746 1746 if quick_data is not None:
1747 1747 return revset.baseset([quick_data[0]])
1748 1748 if user:
1749 1749 m = revset.matchany(
1750 1750 self.ui,
1751 1751 specs,
1752 1752 lookup=revset.lookupfn(self),
1753 1753 localalias=localalias,
1754 1754 )
1755 1755 else:
1756 1756 m = revset.matchany(None, specs, localalias=localalias)
1757 1757 return m(self)
1758 1758
1759 1759 def url(self):
1760 1760 return b'file:' + self.root
1761 1761
1762 1762 def hook(self, name, throw=False, **args):
1763 1763 """Call a hook, passing this repo instance.
1764 1764
1765 1765 This a convenience method to aid invoking hooks. Extensions likely
1766 1766 won't call this unless they have registered a custom hook or are
1767 1767 replacing code that is expected to call a hook.
1768 1768 """
1769 1769 return hook.hook(self.ui, self, name, throw, **args)
1770 1770
1771 1771 @filteredpropertycache
1772 1772 def _tagscache(self):
1773 1773 '''Returns a tagscache object that contains various tags related
1774 1774 caches.'''
1775 1775
1776 1776 # This simplifies its cache management by having one decorated
1777 1777 # function (this one) and the rest simply fetch things from it.
1778 1778 class tagscache(object):
1779 1779 def __init__(self):
1780 1780 # These two define the set of tags for this repository. tags
1781 1781 # maps tag name to node; tagtypes maps tag name to 'global' or
1782 1782 # 'local'. (Global tags are defined by .hgtags across all
1783 1783 # heads, and local tags are defined in .hg/localtags.)
1784 1784 # They constitute the in-memory cache of tags.
1785 1785 self.tags = self.tagtypes = None
1786 1786
1787 1787 self.nodetagscache = self.tagslist = None
1788 1788
1789 1789 cache = tagscache()
1790 1790 cache.tags, cache.tagtypes = self._findtags()
1791 1791
1792 1792 return cache
1793 1793
1794 1794 def tags(self):
1795 1795 '''return a mapping of tag to node'''
1796 1796 t = {}
1797 1797 if self.changelog.filteredrevs:
1798 1798 tags, tt = self._findtags()
1799 1799 else:
1800 1800 tags = self._tagscache.tags
1801 1801 rev = self.changelog.rev
1802 1802 for k, v in pycompat.iteritems(tags):
1803 1803 try:
1804 1804 # ignore tags to unknown nodes
1805 1805 rev(v)
1806 1806 t[k] = v
1807 1807 except (error.LookupError, ValueError):
1808 1808 pass
1809 1809 return t
1810 1810
1811 1811 def _findtags(self):
1812 1812 '''Do the hard work of finding tags. Return a pair of dicts
1813 1813 (tags, tagtypes) where tags maps tag name to node, and tagtypes
1814 1814 maps tag name to a string like \'global\' or \'local\'.
1815 1815 Subclasses or extensions are free to add their own tags, but
1816 1816 should be aware that the returned dicts will be retained for the
1817 1817 duration of the localrepo object.'''
1818 1818
1819 1819 # XXX what tagtype should subclasses/extensions use? Currently
1820 1820 # mq and bookmarks add tags, but do not set the tagtype at all.
1821 1821 # Should each extension invent its own tag type? Should there
1822 1822 # be one tagtype for all such "virtual" tags? Or is the status
1823 1823 # quo fine?
1824 1824
1825 1825 # map tag name to (node, hist)
1826 1826 alltags = tagsmod.findglobaltags(self.ui, self)
1827 1827 # map tag name to tag type
1828 1828 tagtypes = {tag: b'global' for tag in alltags}
1829 1829
1830 1830 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
1831 1831
1832 1832 # Build the return dicts. Have to re-encode tag names because
1833 1833 # the tags module always uses UTF-8 (in order not to lose info
1834 1834 # writing to the cache), but the rest of Mercurial wants them in
1835 1835 # local encoding.
1836 1836 tags = {}
1837 1837 for (name, (node, hist)) in pycompat.iteritems(alltags):
1838 1838 if node != nullid:
1839 1839 tags[encoding.tolocal(name)] = node
1840 1840 tags[b'tip'] = self.changelog.tip()
1841 1841 tagtypes = {
1842 1842 encoding.tolocal(name): value
1843 1843 for (name, value) in pycompat.iteritems(tagtypes)
1844 1844 }
1845 1845 return (tags, tagtypes)
1846 1846
1847 1847 def tagtype(self, tagname):
1848 1848 '''
1849 1849 return the type of the given tag. result can be:
1850 1850
1851 1851 'local' : a local tag
1852 1852 'global' : a global tag
1853 1853 None : tag does not exist
1854 1854 '''
1855 1855
1856 1856 return self._tagscache.tagtypes.get(tagname)
1857 1857
1858 1858 def tagslist(self):
1859 1859 '''return a list of tags ordered by revision'''
1860 1860 if not self._tagscache.tagslist:
1861 1861 l = []
1862 1862 for t, n in pycompat.iteritems(self.tags()):
1863 1863 l.append((self.changelog.rev(n), t, n))
1864 1864 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
1865 1865
1866 1866 return self._tagscache.tagslist
1867 1867
1868 1868 def nodetags(self, node):
1869 1869 '''return the tags associated with a node'''
1870 1870 if not self._tagscache.nodetagscache:
1871 1871 nodetagscache = {}
1872 1872 for t, n in pycompat.iteritems(self._tagscache.tags):
1873 1873 nodetagscache.setdefault(n, []).append(t)
1874 1874 for tags in pycompat.itervalues(nodetagscache):
1875 1875 tags.sort()
1876 1876 self._tagscache.nodetagscache = nodetagscache
1877 1877 return self._tagscache.nodetagscache.get(node, [])
1878 1878
1879 1879 def nodebookmarks(self, node):
1880 1880 """return the list of bookmarks pointing to the specified node"""
1881 1881 return self._bookmarks.names(node)
1882 1882
1883 1883 def branchmap(self):
1884 1884 '''returns a dictionary {branch: [branchheads]} with branchheads
1885 1885 ordered by increasing revision number'''
1886 1886 return self._branchcaches[self]
1887 1887
1888 1888 @unfilteredmethod
1889 1889 def revbranchcache(self):
1890 1890 if not self._revbranchcache:
1891 1891 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
1892 1892 return self._revbranchcache
1893 1893
1894 1894 def branchtip(self, branch, ignoremissing=False):
1895 1895 '''return the tip node for a given branch
1896 1896
1897 1897 If ignoremissing is True, then this method will not raise an error.
1898 1898 This is helpful for callers that only expect None for a missing branch
1899 1899 (e.g. namespace).
1900 1900
1901 1901 '''
1902 1902 try:
1903 1903 return self.branchmap().branchtip(branch)
1904 1904 except KeyError:
1905 1905 if not ignoremissing:
1906 1906 raise error.RepoLookupError(_(b"unknown branch '%s'") % branch)
1907 1907 else:
1908 1908 pass
1909 1909
1910 1910 def lookup(self, key):
1911 1911 node = scmutil.revsymbol(self, key).node()
1912 1912 if node is None:
1913 1913 raise error.RepoLookupError(_(b"unknown revision '%s'") % key)
1914 1914 return node
1915 1915
1916 1916 def lookupbranch(self, key):
1917 1917 if self.branchmap().hasbranch(key):
1918 1918 return key
1919 1919
1920 1920 return scmutil.revsymbol(self, key).branch()
1921 1921
1922 1922 def known(self, nodes):
1923 1923 cl = self.changelog
1924 1924 get_rev = cl.index.get_rev
1925 1925 filtered = cl.filteredrevs
1926 1926 result = []
1927 1927 for n in nodes:
1928 1928 r = get_rev(n)
1929 1929 resp = not (r is None or r in filtered)
1930 1930 result.append(resp)
1931 1931 return result
1932 1932
1933 1933 def local(self):
1934 1934 return self
1935 1935
1936 1936 def publishing(self):
1937 1937 # it's safe (and desirable) to trust the publish flag unconditionally
1938 1938 # so that we don't finalize changes shared between users via ssh or nfs
1939 1939 return self.ui.configbool(b'phases', b'publish', untrusted=True)
1940 1940
1941 1941 def cancopy(self):
1942 1942 # so statichttprepo's override of local() works
1943 1943 if not self.local():
1944 1944 return False
1945 1945 if not self.publishing():
1946 1946 return True
1947 1947 # if publishing we can't copy if there is filtered content
1948 1948 return not self.filtered(b'visible').changelog.filteredrevs
1949 1949
1950 1950 def shared(self):
1951 1951 '''the type of shared repository (None if not shared)'''
1952 1952 if self.sharedpath != self.path:
1953 1953 return b'store'
1954 1954 return None
1955 1955
1956 1956 def wjoin(self, f, *insidef):
1957 1957 return self.vfs.reljoin(self.root, f, *insidef)
1958 1958
1959 1959 def setparents(self, p1, p2=nullid):
1960 1960 self[None].setparents(p1, p2)
1961 1961 self._quick_access_changeid_invalidate()
1962 1962
1963 1963 def filectx(self, path, changeid=None, fileid=None, changectx=None):
1964 1964 """changeid must be a changeset revision, if specified.
1965 1965 fileid can be a file revision or node."""
1966 1966 return context.filectx(
1967 1967 self, path, changeid, fileid, changectx=changectx
1968 1968 )
1969 1969
1970 1970 def getcwd(self):
1971 1971 return self.dirstate.getcwd()
1972 1972
1973 1973 def pathto(self, f, cwd=None):
1974 1974 return self.dirstate.pathto(f, cwd)
1975 1975
1976 1976 def _loadfilter(self, filter):
1977 1977 if filter not in self._filterpats:
1978 1978 l = []
1979 1979 for pat, cmd in self.ui.configitems(filter):
1980 1980 if cmd == b'!':
1981 1981 continue
1982 1982 mf = matchmod.match(self.root, b'', [pat])
1983 1983 fn = None
1984 1984 params = cmd
1985 1985 for name, filterfn in pycompat.iteritems(self._datafilters):
1986 1986 if cmd.startswith(name):
1987 1987 fn = filterfn
1988 1988 params = cmd[len(name) :].lstrip()
1989 1989 break
1990 1990 if not fn:
1991 1991 fn = lambda s, c, **kwargs: procutil.filter(s, c)
1992 1992 fn.__name__ = 'commandfilter'
1993 1993 # Wrap old filters not supporting keyword arguments
1994 1994 if not pycompat.getargspec(fn)[2]:
1995 1995 oldfn = fn
1996 1996 fn = lambda s, c, oldfn=oldfn, **kwargs: oldfn(s, c)
1997 1997 fn.__name__ = 'compat-' + oldfn.__name__
1998 1998 l.append((mf, fn, params))
1999 1999 self._filterpats[filter] = l
2000 2000 return self._filterpats[filter]
2001 2001
2002 2002 def _filter(self, filterpats, filename, data):
2003 2003 for mf, fn, cmd in filterpats:
2004 2004 if mf(filename):
2005 2005 self.ui.debug(
2006 2006 b"filtering %s through %s\n"
2007 2007 % (filename, cmd or pycompat.sysbytes(fn.__name__))
2008 2008 )
2009 2009 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
2010 2010 break
2011 2011
2012 2012 return data
2013 2013
2014 2014 @unfilteredpropertycache
2015 2015 def _encodefilterpats(self):
2016 2016 return self._loadfilter(b'encode')
2017 2017
2018 2018 @unfilteredpropertycache
2019 2019 def _decodefilterpats(self):
2020 2020 return self._loadfilter(b'decode')
2021 2021
2022 2022 def adddatafilter(self, name, filter):
2023 2023 self._datafilters[name] = filter
2024 2024
2025 2025 def wread(self, filename):
2026 2026 if self.wvfs.islink(filename):
2027 2027 data = self.wvfs.readlink(filename)
2028 2028 else:
2029 2029 data = self.wvfs.read(filename)
2030 2030 return self._filter(self._encodefilterpats, filename, data)
2031 2031
2032 2032 def wwrite(self, filename, data, flags, backgroundclose=False, **kwargs):
2033 2033 """write ``data`` into ``filename`` in the working directory
2034 2034
2035 2035 This returns length of written (maybe decoded) data.
2036 2036 """
2037 2037 data = self._filter(self._decodefilterpats, filename, data)
2038 2038 if b'l' in flags:
2039 2039 self.wvfs.symlink(data, filename)
2040 2040 else:
2041 2041 self.wvfs.write(
2042 2042 filename, data, backgroundclose=backgroundclose, **kwargs
2043 2043 )
2044 2044 if b'x' in flags:
2045 2045 self.wvfs.setflags(filename, False, True)
2046 2046 else:
2047 2047 self.wvfs.setflags(filename, False, False)
2048 2048 return len(data)
2049 2049
2050 2050 def wwritedata(self, filename, data):
2051 2051 return self._filter(self._decodefilterpats, filename, data)
2052 2052
2053 2053 def currenttransaction(self):
2054 2054 """return the current transaction or None if non exists"""
2055 2055 if self._transref:
2056 2056 tr = self._transref()
2057 2057 else:
2058 2058 tr = None
2059 2059
2060 2060 if tr and tr.running():
2061 2061 return tr
2062 2062 return None
2063 2063
2064 2064 def transaction(self, desc, report=None):
2065 2065 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
2066 2066 b'devel', b'check-locks'
2067 2067 ):
2068 2068 if self._currentlock(self._lockref) is None:
2069 2069 raise error.ProgrammingError(b'transaction requires locking')
2070 2070 tr = self.currenttransaction()
2071 2071 if tr is not None:
2072 2072 return tr.nest(name=desc)
2073 2073
2074 2074 # abort here if the journal already exists
2075 2075 if self.svfs.exists(b"journal"):
2076 2076 raise error.RepoError(
2077 2077 _(b"abandoned transaction found"),
2078 2078 hint=_(b"run 'hg recover' to clean up transaction"),
2079 2079 )
2080 2080
2081 2081 idbase = b"%.40f#%f" % (random.random(), time.time())
2082 2082 ha = hex(hashutil.sha1(idbase).digest())
2083 2083 txnid = b'TXN:' + ha
2084 2084 self.hook(b'pretxnopen', throw=True, txnname=desc, txnid=txnid)
2085 2085
2086 2086 self._writejournal(desc)
2087 2087 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
2088 2088 if report:
2089 2089 rp = report
2090 2090 else:
2091 2091 rp = self.ui.warn
2092 2092 vfsmap = {b'plain': self.vfs, b'store': self.svfs} # root of .hg/
2093 2093 # we must avoid cyclic reference between repo and transaction.
2094 2094 reporef = weakref.ref(self)
2095 2095 # Code to track tag movement
2096 2096 #
2097 2097 # Since tags are all handled as file content, it is actually quite hard
2098 2098 # to track these movement from a code perspective. So we fallback to a
2099 2099 # tracking at the repository level. One could envision to track changes
2100 2100 # to the '.hgtags' file through changegroup apply but that fails to
2101 2101 # cope with case where transaction expose new heads without changegroup
2102 2102 # being involved (eg: phase movement).
2103 2103 #
2104 2104 # For now, We gate the feature behind a flag since this likely comes
2105 2105 # with performance impacts. The current code run more often than needed
2106 2106 # and do not use caches as much as it could. The current focus is on
2107 2107 # the behavior of the feature so we disable it by default. The flag
2108 2108 # will be removed when we are happy with the performance impact.
2109 2109 #
2110 2110 # Once this feature is no longer experimental move the following
2111 2111 # documentation to the appropriate help section:
2112 2112 #
2113 2113 # The ``HG_TAG_MOVED`` variable will be set if the transaction touched
2114 2114 # tags (new or changed or deleted tags). In addition the details of
2115 2115 # these changes are made available in a file at:
2116 2116 # ``REPOROOT/.hg/changes/tags.changes``.
2117 2117 # Make sure you check for HG_TAG_MOVED before reading that file as it
2118 2118 # might exist from a previous transaction even if no tag were touched
2119 2119 # in this one. Changes are recorded in a line base format::
2120 2120 #
2121 2121 # <action> <hex-node> <tag-name>\n
2122 2122 #
2123 2123 # Actions are defined as follow:
2124 2124 # "-R": tag is removed,
2125 2125 # "+A": tag is added,
2126 2126 # "-M": tag is moved (old value),
2127 2127 # "+M": tag is moved (new value),
2128 2128 tracktags = lambda x: None
2129 2129 # experimental config: experimental.hook-track-tags
2130 2130 shouldtracktags = self.ui.configbool(
2131 2131 b'experimental', b'hook-track-tags'
2132 2132 )
2133 2133 if desc != b'strip' and shouldtracktags:
2134 2134 oldheads = self.changelog.headrevs()
2135 2135
2136 2136 def tracktags(tr2):
2137 2137 repo = reporef()
2138 2138 oldfnodes = tagsmod.fnoderevs(repo.ui, repo, oldheads)
2139 2139 newheads = repo.changelog.headrevs()
2140 2140 newfnodes = tagsmod.fnoderevs(repo.ui, repo, newheads)
2141 2141 # notes: we compare lists here.
2142 2142 # As we do it only once buiding set would not be cheaper
2143 2143 changes = tagsmod.difftags(repo.ui, repo, oldfnodes, newfnodes)
2144 2144 if changes:
2145 2145 tr2.hookargs[b'tag_moved'] = b'1'
2146 2146 with repo.vfs(
2147 2147 b'changes/tags.changes', b'w', atomictemp=True
2148 2148 ) as changesfile:
2149 2149 # note: we do not register the file to the transaction
2150 2150 # because we needs it to still exist on the transaction
2151 2151 # is close (for txnclose hooks)
2152 2152 tagsmod.writediff(changesfile, changes)
2153 2153
2154 2154 def validate(tr2):
2155 2155 """will run pre-closing hooks"""
2156 2156 # XXX the transaction API is a bit lacking here so we take a hacky
2157 2157 # path for now
2158 2158 #
2159 2159 # We cannot add this as a "pending" hooks since the 'tr.hookargs'
2160 2160 # dict is copied before these run. In addition we needs the data
2161 2161 # available to in memory hooks too.
2162 2162 #
2163 2163 # Moreover, we also need to make sure this runs before txnclose
2164 2164 # hooks and there is no "pending" mechanism that would execute
2165 2165 # logic only if hooks are about to run.
2166 2166 #
2167 2167 # Fixing this limitation of the transaction is also needed to track
2168 2168 # other families of changes (bookmarks, phases, obsolescence).
2169 2169 #
2170 2170 # This will have to be fixed before we remove the experimental
2171 2171 # gating.
2172 2172 tracktags(tr2)
2173 2173 repo = reporef()
2174 2174
2175 2175 singleheadopt = (b'experimental', b'single-head-per-branch')
2176 2176 singlehead = repo.ui.configbool(*singleheadopt)
2177 2177 if singlehead:
2178 2178 singleheadsub = repo.ui.configsuboptions(*singleheadopt)[1]
2179 2179 accountclosed = singleheadsub.get(
2180 2180 b"account-closed-heads", False
2181 2181 )
2182 2182 scmutil.enforcesinglehead(repo, tr2, desc, accountclosed)
2183 2183 if hook.hashook(repo.ui, b'pretxnclose-bookmark'):
2184 2184 for name, (old, new) in sorted(
2185 2185 tr.changes[b'bookmarks'].items()
2186 2186 ):
2187 2187 args = tr.hookargs.copy()
2188 2188 args.update(bookmarks.preparehookargs(name, old, new))
2189 2189 repo.hook(
2190 2190 b'pretxnclose-bookmark',
2191 2191 throw=True,
2192 2192 **pycompat.strkwargs(args)
2193 2193 )
2194 2194 if hook.hashook(repo.ui, b'pretxnclose-phase'):
2195 2195 cl = repo.unfiltered().changelog
2196 2196 for revs, (old, new) in tr.changes[b'phases']:
2197 2197 for rev in revs:
2198 2198 args = tr.hookargs.copy()
2199 2199 node = hex(cl.node(rev))
2200 2200 args.update(phases.preparehookargs(node, old, new))
2201 2201 repo.hook(
2202 2202 b'pretxnclose-phase',
2203 2203 throw=True,
2204 2204 **pycompat.strkwargs(args)
2205 2205 )
2206 2206
2207 2207 repo.hook(
2208 2208 b'pretxnclose', throw=True, **pycompat.strkwargs(tr.hookargs)
2209 2209 )
2210 2210
2211 2211 def releasefn(tr, success):
2212 2212 repo = reporef()
2213 2213 if repo is None:
2214 2214 # If the repo has been GC'd (and this release function is being
2215 2215 # called from transaction.__del__), there's not much we can do,
2216 2216 # so just leave the unfinished transaction there and let the
2217 2217 # user run `hg recover`.
2218 2218 return
2219 2219 if success:
2220 2220 # this should be explicitly invoked here, because
2221 2221 # in-memory changes aren't written out at closing
2222 2222 # transaction, if tr.addfilegenerator (via
2223 2223 # dirstate.write or so) isn't invoked while
2224 2224 # transaction running
2225 2225 repo.dirstate.write(None)
2226 2226 else:
2227 2227 # discard all changes (including ones already written
2228 2228 # out) in this transaction
2229 2229 narrowspec.restorebackup(self, b'journal.narrowspec')
2230 2230 narrowspec.restorewcbackup(self, b'journal.narrowspec.dirstate')
2231 2231 repo.dirstate.restorebackup(None, b'journal.dirstate')
2232 2232
2233 2233 repo.invalidate(clearfilecache=True)
2234 2234
2235 2235 tr = transaction.transaction(
2236 2236 rp,
2237 2237 self.svfs,
2238 2238 vfsmap,
2239 2239 b"journal",
2240 2240 b"undo",
2241 2241 aftertrans(renames),
2242 2242 self.store.createmode,
2243 2243 validator=validate,
2244 2244 releasefn=releasefn,
2245 2245 checkambigfiles=_cachedfiles,
2246 2246 name=desc,
2247 2247 )
2248 2248 tr.changes[b'origrepolen'] = len(self)
2249 2249 tr.changes[b'obsmarkers'] = set()
2250 2250 tr.changes[b'phases'] = []
2251 2251 tr.changes[b'bookmarks'] = {}
2252 2252
2253 2253 tr.hookargs[b'txnid'] = txnid
2254 2254 tr.hookargs[b'txnname'] = desc
2255 2255 tr.hookargs[b'changes'] = tr.changes
2256 2256 # note: writing the fncache only during finalize mean that the file is
2257 2257 # outdated when running hooks. As fncache is used for streaming clone,
2258 2258 # this is not expected to break anything that happen during the hooks.
2259 2259 tr.addfinalize(b'flush-fncache', self.store.write)
2260 2260
2261 2261 def txnclosehook(tr2):
2262 2262 """To be run if transaction is successful, will schedule a hook run
2263 2263 """
2264 2264 # Don't reference tr2 in hook() so we don't hold a reference.
2265 2265 # This reduces memory consumption when there are multiple
2266 2266 # transactions per lock. This can likely go away if issue5045
2267 2267 # fixes the function accumulation.
2268 2268 hookargs = tr2.hookargs
2269 2269
2270 2270 def hookfunc(unused_success):
2271 2271 repo = reporef()
2272 2272 if hook.hashook(repo.ui, b'txnclose-bookmark'):
2273 2273 bmchanges = sorted(tr.changes[b'bookmarks'].items())
2274 2274 for name, (old, new) in bmchanges:
2275 2275 args = tr.hookargs.copy()
2276 2276 args.update(bookmarks.preparehookargs(name, old, new))
2277 2277 repo.hook(
2278 2278 b'txnclose-bookmark',
2279 2279 throw=False,
2280 2280 **pycompat.strkwargs(args)
2281 2281 )
2282 2282
2283 2283 if hook.hashook(repo.ui, b'txnclose-phase'):
2284 2284 cl = repo.unfiltered().changelog
2285 2285 phasemv = sorted(
2286 2286 tr.changes[b'phases'], key=lambda r: r[0][0]
2287 2287 )
2288 2288 for revs, (old, new) in phasemv:
2289 2289 for rev in revs:
2290 2290 args = tr.hookargs.copy()
2291 2291 node = hex(cl.node(rev))
2292 2292 args.update(phases.preparehookargs(node, old, new))
2293 2293 repo.hook(
2294 2294 b'txnclose-phase',
2295 2295 throw=False,
2296 2296 **pycompat.strkwargs(args)
2297 2297 )
2298 2298
2299 2299 repo.hook(
2300 2300 b'txnclose', throw=False, **pycompat.strkwargs(hookargs)
2301 2301 )
2302 2302
2303 2303 reporef()._afterlock(hookfunc)
2304 2304
2305 2305 tr.addfinalize(b'txnclose-hook', txnclosehook)
2306 2306 # Include a leading "-" to make it happen before the transaction summary
2307 2307 # reports registered via scmutil.registersummarycallback() whose names
2308 2308 # are 00-txnreport etc. That way, the caches will be warm when the
2309 2309 # callbacks run.
2310 2310 tr.addpostclose(b'-warm-cache', self._buildcacheupdater(tr))
2311 2311
2312 2312 def txnaborthook(tr2):
2313 2313 """To be run if transaction is aborted
2314 2314 """
2315 2315 reporef().hook(
2316 2316 b'txnabort', throw=False, **pycompat.strkwargs(tr2.hookargs)
2317 2317 )
2318 2318
2319 2319 tr.addabort(b'txnabort-hook', txnaborthook)
2320 2320 # avoid eager cache invalidation. in-memory data should be identical
2321 2321 # to stored data if transaction has no error.
2322 2322 tr.addpostclose(b'refresh-filecachestats', self._refreshfilecachestats)
2323 2323 self._transref = weakref.ref(tr)
2324 2324 scmutil.registersummarycallback(self, tr, desc)
2325 2325 return tr
2326 2326
2327 2327 def _journalfiles(self):
2328 2328 return (
2329 2329 (self.svfs, b'journal'),
2330 2330 (self.svfs, b'journal.narrowspec'),
2331 2331 (self.vfs, b'journal.narrowspec.dirstate'),
2332 2332 (self.vfs, b'journal.dirstate'),
2333 2333 (self.vfs, b'journal.branch'),
2334 2334 (self.vfs, b'journal.desc'),
2335 2335 (bookmarks.bookmarksvfs(self), b'journal.bookmarks'),
2336 2336 (self.svfs, b'journal.phaseroots'),
2337 2337 )
2338 2338
2339 2339 def undofiles(self):
2340 2340 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
2341 2341
2342 2342 @unfilteredmethod
2343 2343 def _writejournal(self, desc):
2344 2344 self.dirstate.savebackup(None, b'journal.dirstate')
2345 2345 narrowspec.savewcbackup(self, b'journal.narrowspec.dirstate')
2346 2346 narrowspec.savebackup(self, b'journal.narrowspec')
2347 2347 self.vfs.write(
2348 2348 b"journal.branch", encoding.fromlocal(self.dirstate.branch())
2349 2349 )
2350 2350 self.vfs.write(b"journal.desc", b"%d\n%s\n" % (len(self), desc))
2351 2351 bookmarksvfs = bookmarks.bookmarksvfs(self)
2352 2352 bookmarksvfs.write(
2353 2353 b"journal.bookmarks", bookmarksvfs.tryread(b"bookmarks")
2354 2354 )
2355 2355 self.svfs.write(b"journal.phaseroots", self.svfs.tryread(b"phaseroots"))
2356 2356
2357 2357 def recover(self):
2358 2358 with self.lock():
2359 2359 if self.svfs.exists(b"journal"):
2360 2360 self.ui.status(_(b"rolling back interrupted transaction\n"))
2361 2361 vfsmap = {
2362 2362 b'': self.svfs,
2363 2363 b'plain': self.vfs,
2364 2364 }
2365 2365 transaction.rollback(
2366 2366 self.svfs,
2367 2367 vfsmap,
2368 2368 b"journal",
2369 2369 self.ui.warn,
2370 2370 checkambigfiles=_cachedfiles,
2371 2371 )
2372 2372 self.invalidate()
2373 2373 return True
2374 2374 else:
2375 2375 self.ui.warn(_(b"no interrupted transaction available\n"))
2376 2376 return False
2377 2377
2378 2378 def rollback(self, dryrun=False, force=False):
2379 2379 wlock = lock = dsguard = None
2380 2380 try:
2381 2381 wlock = self.wlock()
2382 2382 lock = self.lock()
2383 2383 if self.svfs.exists(b"undo"):
2384 2384 dsguard = dirstateguard.dirstateguard(self, b'rollback')
2385 2385
2386 2386 return self._rollback(dryrun, force, dsguard)
2387 2387 else:
2388 2388 self.ui.warn(_(b"no rollback information available\n"))
2389 2389 return 1
2390 2390 finally:
2391 2391 release(dsguard, lock, wlock)
2392 2392
2393 2393 @unfilteredmethod # Until we get smarter cache management
2394 2394 def _rollback(self, dryrun, force, dsguard):
2395 2395 ui = self.ui
2396 2396 try:
2397 2397 args = self.vfs.read(b'undo.desc').splitlines()
2398 2398 (oldlen, desc, detail) = (int(args[0]), args[1], None)
2399 2399 if len(args) >= 3:
2400 2400 detail = args[2]
2401 2401 oldtip = oldlen - 1
2402 2402
2403 2403 if detail and ui.verbose:
2404 2404 msg = _(
2405 2405 b'repository tip rolled back to revision %d'
2406 2406 b' (undo %s: %s)\n'
2407 2407 ) % (oldtip, desc, detail)
2408 2408 else:
2409 2409 msg = _(
2410 2410 b'repository tip rolled back to revision %d (undo %s)\n'
2411 2411 ) % (oldtip, desc)
2412 2412 except IOError:
2413 2413 msg = _(b'rolling back unknown transaction\n')
2414 2414 desc = None
2415 2415
2416 2416 if not force and self[b'.'] != self[b'tip'] and desc == b'commit':
2417 2417 raise error.Abort(
2418 2418 _(
2419 2419 b'rollback of last commit while not checked out '
2420 2420 b'may lose data'
2421 2421 ),
2422 2422 hint=_(b'use -f to force'),
2423 2423 )
2424 2424
2425 2425 ui.status(msg)
2426 2426 if dryrun:
2427 2427 return 0
2428 2428
2429 2429 parents = self.dirstate.parents()
2430 2430 self.destroying()
2431 2431 vfsmap = {b'plain': self.vfs, b'': self.svfs}
2432 2432 transaction.rollback(
2433 2433 self.svfs, vfsmap, b'undo', ui.warn, checkambigfiles=_cachedfiles
2434 2434 )
2435 2435 bookmarksvfs = bookmarks.bookmarksvfs(self)
2436 2436 if bookmarksvfs.exists(b'undo.bookmarks'):
2437 2437 bookmarksvfs.rename(
2438 2438 b'undo.bookmarks', b'bookmarks', checkambig=True
2439 2439 )
2440 2440 if self.svfs.exists(b'undo.phaseroots'):
2441 2441 self.svfs.rename(b'undo.phaseroots', b'phaseroots', checkambig=True)
2442 2442 self.invalidate()
2443 2443
2444 2444 has_node = self.changelog.index.has_node
2445 2445 parentgone = any(not has_node(p) for p in parents)
2446 2446 if parentgone:
2447 2447 # prevent dirstateguard from overwriting already restored one
2448 2448 dsguard.close()
2449 2449
2450 2450 narrowspec.restorebackup(self, b'undo.narrowspec')
2451 2451 narrowspec.restorewcbackup(self, b'undo.narrowspec.dirstate')
2452 2452 self.dirstate.restorebackup(None, b'undo.dirstate')
2453 2453 try:
2454 2454 branch = self.vfs.read(b'undo.branch')
2455 2455 self.dirstate.setbranch(encoding.tolocal(branch))
2456 2456 except IOError:
2457 2457 ui.warn(
2458 2458 _(
2459 2459 b'named branch could not be reset: '
2460 2460 b'current branch is still \'%s\'\n'
2461 2461 )
2462 2462 % self.dirstate.branch()
2463 2463 )
2464 2464
2465 2465 parents = tuple([p.rev() for p in self[None].parents()])
2466 2466 if len(parents) > 1:
2467 2467 ui.status(
2468 2468 _(
2469 2469 b'working directory now based on '
2470 2470 b'revisions %d and %d\n'
2471 2471 )
2472 2472 % parents
2473 2473 )
2474 2474 else:
2475 2475 ui.status(
2476 2476 _(b'working directory now based on revision %d\n') % parents
2477 2477 )
2478 2478 mergestatemod.mergestate.clean(self, self[b'.'].node())
2479 2479
2480 2480 # TODO: if we know which new heads may result from this rollback, pass
2481 2481 # them to destroy(), which will prevent the branchhead cache from being
2482 2482 # invalidated.
2483 2483 self.destroyed()
2484 2484 return 0
2485 2485
2486 2486 def _buildcacheupdater(self, newtransaction):
2487 2487 """called during transaction to build the callback updating cache
2488 2488
2489 2489 Lives on the repository to help extension who might want to augment
2490 2490 this logic. For this purpose, the created transaction is passed to the
2491 2491 method.
2492 2492 """
2493 2493 # we must avoid cyclic reference between repo and transaction.
2494 2494 reporef = weakref.ref(self)
2495 2495
2496 2496 def updater(tr):
2497 2497 repo = reporef()
2498 2498 repo.updatecaches(tr)
2499 2499
2500 2500 return updater
2501 2501
2502 2502 @unfilteredmethod
2503 2503 def updatecaches(self, tr=None, full=False):
2504 2504 """warm appropriate caches
2505 2505
2506 2506 If this function is called after a transaction closed. The transaction
2507 2507 will be available in the 'tr' argument. This can be used to selectively
2508 2508 update caches relevant to the changes in that transaction.
2509 2509
2510 2510 If 'full' is set, make sure all caches the function knows about have
2511 2511 up-to-date data. Even the ones usually loaded more lazily.
2512 2512 """
2513 2513 if tr is not None and tr.hookargs.get(b'source') == b'strip':
2514 2514 # During strip, many caches are invalid but
2515 2515 # later call to `destroyed` will refresh them.
2516 2516 return
2517 2517
2518 2518 if tr is None or tr.changes[b'origrepolen'] < len(self):
2519 2519 # accessing the 'ser ved' branchmap should refresh all the others,
2520 2520 self.ui.debug(b'updating the branch cache\n')
2521 2521 self.filtered(b'served').branchmap()
2522 2522 self.filtered(b'served.hidden').branchmap()
2523 2523
2524 2524 if full:
2525 2525 unfi = self.unfiltered()
2526 2526
2527 2527 self.changelog.update_caches(transaction=tr)
2528 2528 self.manifestlog.update_caches(transaction=tr)
2529 2529
2530 2530 rbc = unfi.revbranchcache()
2531 2531 for r in unfi.changelog:
2532 2532 rbc.branchinfo(r)
2533 2533 rbc.write()
2534 2534
2535 2535 # ensure the working copy parents are in the manifestfulltextcache
2536 2536 for ctx in self[b'.'].parents():
2537 2537 ctx.manifest() # accessing the manifest is enough
2538 2538
2539 2539 # accessing fnode cache warms the cache
2540 2540 tagsmod.fnoderevs(self.ui, unfi, unfi.changelog.revs())
2541 2541 # accessing tags warm the cache
2542 2542 self.tags()
2543 2543 self.filtered(b'served').tags()
2544 2544
2545 2545 # The `full` arg is documented as updating even the lazily-loaded
2546 2546 # caches immediately, so we're forcing a write to cause these caches
2547 2547 # to be warmed up even if they haven't explicitly been requested
2548 2548 # yet (if they've never been used by hg, they won't ever have been
2549 2549 # written, even if they're a subset of another kind of cache that
2550 2550 # *has* been used).
2551 2551 for filt in repoview.filtertable.keys():
2552 2552 filtered = self.filtered(filt)
2553 2553 filtered.branchmap().write(filtered)
2554 2554
2555 2555 def invalidatecaches(self):
2556 2556
2557 2557 if '_tagscache' in vars(self):
2558 2558 # can't use delattr on proxy
2559 2559 del self.__dict__['_tagscache']
2560 2560
2561 2561 self._branchcaches.clear()
2562 2562 self.invalidatevolatilesets()
2563 2563 self._sparsesignaturecache.clear()
2564 2564
2565 2565 def invalidatevolatilesets(self):
2566 2566 self.filteredrevcache.clear()
2567 2567 obsolete.clearobscaches(self)
2568 2568 self._quick_access_changeid_invalidate()
2569 2569
2570 2570 def invalidatedirstate(self):
2571 2571 '''Invalidates the dirstate, causing the next call to dirstate
2572 2572 to check if it was modified since the last time it was read,
2573 2573 rereading it if it has.
2574 2574
2575 2575 This is different to dirstate.invalidate() that it doesn't always
2576 2576 rereads the dirstate. Use dirstate.invalidate() if you want to
2577 2577 explicitly read the dirstate again (i.e. restoring it to a previous
2578 2578 known good state).'''
2579 2579 if hasunfilteredcache(self, 'dirstate'):
2580 2580 for k in self.dirstate._filecache:
2581 2581 try:
2582 2582 delattr(self.dirstate, k)
2583 2583 except AttributeError:
2584 2584 pass
2585 2585 delattr(self.unfiltered(), 'dirstate')
2586 2586
2587 2587 def invalidate(self, clearfilecache=False):
2588 2588 '''Invalidates both store and non-store parts other than dirstate
2589 2589
2590 2590 If a transaction is running, invalidation of store is omitted,
2591 2591 because discarding in-memory changes might cause inconsistency
2592 2592 (e.g. incomplete fncache causes unintentional failure, but
2593 2593 redundant one doesn't).
2594 2594 '''
2595 2595 unfiltered = self.unfiltered() # all file caches are stored unfiltered
2596 2596 for k in list(self._filecache.keys()):
2597 2597 # dirstate is invalidated separately in invalidatedirstate()
2598 2598 if k == b'dirstate':
2599 2599 continue
2600 2600 if (
2601 2601 k == b'changelog'
2602 2602 and self.currenttransaction()
2603 2603 and self.changelog._delayed
2604 2604 ):
2605 2605 # The changelog object may store unwritten revisions. We don't
2606 2606 # want to lose them.
2607 2607 # TODO: Solve the problem instead of working around it.
2608 2608 continue
2609 2609
2610 2610 if clearfilecache:
2611 2611 del self._filecache[k]
2612 2612 try:
2613 2613 delattr(unfiltered, k)
2614 2614 except AttributeError:
2615 2615 pass
2616 2616 self.invalidatecaches()
2617 2617 if not self.currenttransaction():
2618 2618 # TODO: Changing contents of store outside transaction
2619 2619 # causes inconsistency. We should make in-memory store
2620 2620 # changes detectable, and abort if changed.
2621 2621 self.store.invalidatecaches()
2622 2622
2623 2623 def invalidateall(self):
2624 2624 '''Fully invalidates both store and non-store parts, causing the
2625 2625 subsequent operation to reread any outside changes.'''
2626 2626 # extension should hook this to invalidate its caches
2627 2627 self.invalidate()
2628 2628 self.invalidatedirstate()
2629 2629
2630 2630 @unfilteredmethod
2631 2631 def _refreshfilecachestats(self, tr):
2632 2632 """Reload stats of cached files so that they are flagged as valid"""
2633 2633 for k, ce in self._filecache.items():
2634 2634 k = pycompat.sysstr(k)
2635 2635 if k == 'dirstate' or k not in self.__dict__:
2636 2636 continue
2637 2637 ce.refresh()
2638 2638
2639 2639 def _lock(
2640 2640 self,
2641 2641 vfs,
2642 2642 lockname,
2643 2643 wait,
2644 2644 releasefn,
2645 2645 acquirefn,
2646 2646 desc,
2647 2647 inheritchecker=None,
2648 2648 parentenvvar=None,
2649 2649 ):
2650 2650 parentlock = None
2651 2651 # the contents of parentenvvar are used by the underlying lock to
2652 2652 # determine whether it can be inherited
2653 2653 if parentenvvar is not None:
2654 2654 parentlock = encoding.environ.get(parentenvvar)
2655 2655
2656 2656 timeout = 0
2657 2657 warntimeout = 0
2658 2658 if wait:
2659 2659 timeout = self.ui.configint(b"ui", b"timeout")
2660 2660 warntimeout = self.ui.configint(b"ui", b"timeout.warn")
2661 2661 # internal config: ui.signal-safe-lock
2662 2662 signalsafe = self.ui.configbool(b'ui', b'signal-safe-lock')
2663 2663
2664 2664 l = lockmod.trylock(
2665 2665 self.ui,
2666 2666 vfs,
2667 2667 lockname,
2668 2668 timeout,
2669 2669 warntimeout,
2670 2670 releasefn=releasefn,
2671 2671 acquirefn=acquirefn,
2672 2672 desc=desc,
2673 2673 inheritchecker=inheritchecker,
2674 2674 parentlock=parentlock,
2675 2675 signalsafe=signalsafe,
2676 2676 )
2677 2677 return l
2678 2678
2679 2679 def _afterlock(self, callback):
2680 2680 """add a callback to be run when the repository is fully unlocked
2681 2681
2682 2682 The callback will be executed when the outermost lock is released
2683 2683 (with wlock being higher level than 'lock')."""
2684 2684 for ref in (self._wlockref, self._lockref):
2685 2685 l = ref and ref()
2686 2686 if l and l.held:
2687 2687 l.postrelease.append(callback)
2688 2688 break
2689 2689 else: # no lock have been found.
2690 2690 callback(True)
2691 2691
2692 2692 def lock(self, wait=True):
2693 2693 '''Lock the repository store (.hg/store) and return a weak reference
2694 2694 to the lock. Use this before modifying the store (e.g. committing or
2695 2695 stripping). If you are opening a transaction, get a lock as well.)
2696 2696
2697 2697 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2698 2698 'wlock' first to avoid a dead-lock hazard.'''
2699 2699 l = self._currentlock(self._lockref)
2700 2700 if l is not None:
2701 2701 l.lock()
2702 2702 return l
2703 2703
2704 2704 l = self._lock(
2705 2705 vfs=self.svfs,
2706 2706 lockname=b"lock",
2707 2707 wait=wait,
2708 2708 releasefn=None,
2709 2709 acquirefn=self.invalidate,
2710 2710 desc=_(b'repository %s') % self.origroot,
2711 2711 )
2712 2712 self._lockref = weakref.ref(l)
2713 2713 return l
2714 2714
2715 2715 def _wlockchecktransaction(self):
2716 2716 if self.currenttransaction() is not None:
2717 2717 raise error.LockInheritanceContractViolation(
2718 2718 b'wlock cannot be inherited in the middle of a transaction'
2719 2719 )
2720 2720
2721 2721 def wlock(self, wait=True):
2722 2722 '''Lock the non-store parts of the repository (everything under
2723 2723 .hg except .hg/store) and return a weak reference to the lock.
2724 2724
2725 2725 Use this before modifying files in .hg.
2726 2726
2727 2727 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2728 2728 'wlock' first to avoid a dead-lock hazard.'''
2729 2729 l = self._wlockref and self._wlockref()
2730 2730 if l is not None and l.held:
2731 2731 l.lock()
2732 2732 return l
2733 2733
2734 2734 # We do not need to check for non-waiting lock acquisition. Such
2735 2735 # acquisition would not cause dead-lock as they would just fail.
2736 2736 if wait and (
2737 2737 self.ui.configbool(b'devel', b'all-warnings')
2738 2738 or self.ui.configbool(b'devel', b'check-locks')
2739 2739 ):
2740 2740 if self._currentlock(self._lockref) is not None:
2741 2741 self.ui.develwarn(b'"wlock" acquired after "lock"')
2742 2742
2743 2743 def unlock():
2744 2744 if self.dirstate.pendingparentchange():
2745 2745 self.dirstate.invalidate()
2746 2746 else:
2747 2747 self.dirstate.write(None)
2748 2748
2749 2749 self._filecache[b'dirstate'].refresh()
2750 2750
2751 2751 l = self._lock(
2752 2752 self.vfs,
2753 2753 b"wlock",
2754 2754 wait,
2755 2755 unlock,
2756 2756 self.invalidatedirstate,
2757 2757 _(b'working directory of %s') % self.origroot,
2758 2758 inheritchecker=self._wlockchecktransaction,
2759 2759 parentenvvar=b'HG_WLOCK_LOCKER',
2760 2760 )
2761 2761 self._wlockref = weakref.ref(l)
2762 2762 return l
2763 2763
2764 2764 def _currentlock(self, lockref):
2765 2765 """Returns the lock if it's held, or None if it's not."""
2766 2766 if lockref is None:
2767 2767 return None
2768 2768 l = lockref()
2769 2769 if l is None or not l.held:
2770 2770 return None
2771 2771 return l
2772 2772
2773 2773 def currentwlock(self):
2774 2774 """Returns the wlock if it's held, or None if it's not."""
2775 2775 return self._currentlock(self._wlockref)
2776 2776
2777 2777 def _filecommit(
2778 2778 self,
2779 2779 fctx,
2780 2780 manifest1,
2781 2781 manifest2,
2782 2782 linkrev,
2783 2783 tr,
2784 2784 changelist,
2785 2785 includecopymeta,
2786 2786 ):
2787 2787 """
2788 2788 commit an individual file as part of a larger transaction
2789
2790 input:
2791
2792 fctx: a file context with the content we are trying to commit
2793 manifest1: manifest of changeset first parent
2794 manifest2: manifest of changeset second parent
2795 linkrev: revision number of the changeset being created
2796 tr: current transation
2797 changelist: list of file being changed (modified inplace)
2798 individual: boolean, set to False to skip storing the copy data
2799 (only used by the Google specific feature of using
2800 changeset extra as copy source of truth).
2801
2802 output:
2803
2804 The resulting filenode
2789 2805 """
2790 2806
2791 2807 fname = fctx.path()
2792 2808 fparent1 = manifest1.get(fname, nullid)
2793 2809 fparent2 = manifest2.get(fname, nullid)
2794 2810 if isinstance(fctx, context.filectx):
2795 2811 node = fctx.filenode()
2796 2812 if node in [fparent1, fparent2]:
2797 2813 self.ui.debug(b'reusing %s filelog entry\n' % fname)
2798 2814 if (
2799 2815 fparent1 != nullid
2800 2816 and manifest1.flags(fname) != fctx.flags()
2801 2817 ) or (
2802 2818 fparent2 != nullid
2803 2819 and manifest2.flags(fname) != fctx.flags()
2804 2820 ):
2805 2821 changelist.append(fname)
2806 2822 return node
2807 2823
2808 2824 flog = self.file(fname)
2809 2825 meta = {}
2810 2826 cfname = fctx.copysource()
2811 2827 if cfname and cfname != fname:
2812 2828 # Mark the new revision of this file as a copy of another
2813 2829 # file. This copy data will effectively act as a parent
2814 2830 # of this new revision. If this is a merge, the first
2815 2831 # parent will be the nullid (meaning "look up the copy data")
2816 2832 # and the second one will be the other parent. For example:
2817 2833 #
2818 2834 # 0 --- 1 --- 3 rev1 changes file foo
2819 2835 # \ / rev2 renames foo to bar and changes it
2820 2836 # \- 2 -/ rev3 should have bar with all changes and
2821 2837 # should record that bar descends from
2822 2838 # bar in rev2 and foo in rev1
2823 2839 #
2824 2840 # this allows this merge to succeed:
2825 2841 #
2826 2842 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
2827 2843 # \ / merging rev3 and rev4 should use bar@rev2
2828 2844 # \- 2 --- 4 as the merge base
2829 2845 #
2830 2846
2831 2847 cnode = manifest1.get(cfname)
2832 2848 newfparent = fparent2
2833 2849
2834 2850 if manifest2: # branch merge
2835 2851 if fparent2 == nullid or cnode is None: # copied on remote side
2836 2852 if cfname in manifest2:
2837 2853 cnode = manifest2[cfname]
2838 2854 newfparent = fparent1
2839 2855
2840 2856 # Here, we used to search backwards through history to try to find
2841 2857 # where the file copy came from if the source of a copy was not in
2842 2858 # the parent directory. However, this doesn't actually make sense to
2843 2859 # do (what does a copy from something not in your working copy even
2844 2860 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
2845 2861 # the user that copy information was dropped, so if they didn't
2846 2862 # expect this outcome it can be fixed, but this is the correct
2847 2863 # behavior in this circumstance.
2848 2864
2849 2865 if cnode:
2850 2866 self.ui.debug(
2851 2867 b" %s: copy %s:%s\n" % (fname, cfname, hex(cnode))
2852 2868 )
2853 2869 if includecopymeta:
2854 2870 meta[b"copy"] = cfname
2855 2871 meta[b"copyrev"] = hex(cnode)
2856 2872 fparent1, fparent2 = nullid, newfparent
2857 2873 else:
2858 2874 self.ui.warn(
2859 2875 _(
2860 2876 b"warning: can't find ancestor for '%s' "
2861 2877 b"copied from '%s'!\n"
2862 2878 )
2863 2879 % (fname, cfname)
2864 2880 )
2865 2881
2866 2882 elif fparent1 == nullid:
2867 2883 fparent1, fparent2 = fparent2, nullid
2868 2884 elif fparent2 != nullid:
2869 2885 # is one parent an ancestor of the other?
2870 2886 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
2871 2887 if fparent1 in fparentancestors:
2872 2888 fparent1, fparent2 = fparent2, nullid
2873 2889 elif fparent2 in fparentancestors:
2874 2890 fparent2 = nullid
2875 2891 elif not fparentancestors:
2876 2892 # TODO: this whole if-else might be simplified much more
2877 2893 ms = mergestatemod.mergestate.read(self)
2878 2894 if (
2879 2895 fname in ms
2880 2896 and ms[fname] == mergestatemod.MERGE_RECORD_MERGED_OTHER
2881 2897 ):
2882 2898 fparent1, fparent2 = fparent2, nullid
2883 2899
2884 2900 # is the file changed?
2885 2901 text = fctx.data()
2886 2902 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
2887 2903 changelist.append(fname)
2888 2904 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
2889 2905 # are just the flags changed during merge?
2890 2906 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
2891 2907 changelist.append(fname)
2892 2908
2893 2909 return fparent1
2894 2910
2895 2911 def checkcommitpatterns(self, wctx, match, status, fail):
2896 2912 """check for commit arguments that aren't committable"""
2897 2913 if match.isexact() or match.prefix():
2898 2914 matched = set(status.modified + status.added + status.removed)
2899 2915
2900 2916 for f in match.files():
2901 2917 f = self.dirstate.normalize(f)
2902 2918 if f == b'.' or f in matched or f in wctx.substate:
2903 2919 continue
2904 2920 if f in status.deleted:
2905 2921 fail(f, _(b'file not found!'))
2906 2922 # Is it a directory that exists or used to exist?
2907 2923 if self.wvfs.isdir(f) or wctx.p1().hasdir(f):
2908 2924 d = f + b'/'
2909 2925 for mf in matched:
2910 2926 if mf.startswith(d):
2911 2927 break
2912 2928 else:
2913 2929 fail(f, _(b"no match under directory!"))
2914 2930 elif f not in self.dirstate:
2915 2931 fail(f, _(b"file not tracked!"))
2916 2932
2917 2933 @unfilteredmethod
2918 2934 def commit(
2919 2935 self,
2920 2936 text=b"",
2921 2937 user=None,
2922 2938 date=None,
2923 2939 match=None,
2924 2940 force=False,
2925 2941 editor=None,
2926 2942 extra=None,
2927 2943 ):
2928 2944 """Add a new revision to current repository.
2929 2945
2930 2946 Revision information is gathered from the working directory,
2931 2947 match can be used to filter the committed files. If editor is
2932 2948 supplied, it is called to get a commit message.
2933 2949 """
2934 2950 if extra is None:
2935 2951 extra = {}
2936 2952
2937 2953 def fail(f, msg):
2938 2954 raise error.Abort(b'%s: %s' % (f, msg))
2939 2955
2940 2956 if not match:
2941 2957 match = matchmod.always()
2942 2958
2943 2959 if not force:
2944 2960 match.bad = fail
2945 2961
2946 2962 # lock() for recent changelog (see issue4368)
2947 2963 with self.wlock(), self.lock():
2948 2964 wctx = self[None]
2949 2965 merge = len(wctx.parents()) > 1
2950 2966
2951 2967 if not force and merge and not match.always():
2952 2968 raise error.Abort(
2953 2969 _(
2954 2970 b'cannot partially commit a merge '
2955 2971 b'(do not specify files or patterns)'
2956 2972 )
2957 2973 )
2958 2974
2959 2975 status = self.status(match=match, clean=force)
2960 2976 if force:
2961 2977 status.modified.extend(
2962 2978 status.clean
2963 2979 ) # mq may commit clean files
2964 2980
2965 2981 # check subrepos
2966 2982 subs, commitsubs, newstate = subrepoutil.precommit(
2967 2983 self.ui, wctx, status, match, force=force
2968 2984 )
2969 2985
2970 2986 # make sure all explicit patterns are matched
2971 2987 if not force:
2972 2988 self.checkcommitpatterns(wctx, match, status, fail)
2973 2989
2974 2990 cctx = context.workingcommitctx(
2975 2991 self, status, text, user, date, extra
2976 2992 )
2977 2993
2978 2994 ms = mergestatemod.mergestate.read(self)
2979 2995 mergeutil.checkunresolved(ms)
2980 2996
2981 2997 # internal config: ui.allowemptycommit
2982 2998 allowemptycommit = (
2983 2999 wctx.branch() != wctx.p1().branch()
2984 3000 or extra.get(b'close')
2985 3001 or merge
2986 3002 or cctx.files()
2987 3003 or self.ui.configbool(b'ui', b'allowemptycommit')
2988 3004 )
2989 3005 if not allowemptycommit:
2990 3006 self.ui.debug(b'nothing to commit, clearing merge state\n')
2991 3007 ms.reset()
2992 3008 return None
2993 3009
2994 3010 if merge and cctx.deleted():
2995 3011 raise error.Abort(_(b"cannot commit merge with missing files"))
2996 3012
2997 3013 if editor:
2998 3014 cctx._text = editor(self, cctx, subs)
2999 3015 edited = text != cctx._text
3000 3016
3001 3017 # Save commit message in case this transaction gets rolled back
3002 3018 # (e.g. by a pretxncommit hook). Leave the content alone on
3003 3019 # the assumption that the user will use the same editor again.
3004 3020 msgfn = self.savecommitmessage(cctx._text)
3005 3021
3006 3022 # commit subs and write new state
3007 3023 if subs:
3008 3024 uipathfn = scmutil.getuipathfn(self)
3009 3025 for s in sorted(commitsubs):
3010 3026 sub = wctx.sub(s)
3011 3027 self.ui.status(
3012 3028 _(b'committing subrepository %s\n')
3013 3029 % uipathfn(subrepoutil.subrelpath(sub))
3014 3030 )
3015 3031 sr = sub.commit(cctx._text, user, date)
3016 3032 newstate[s] = (newstate[s][0], sr)
3017 3033 subrepoutil.writestate(self, newstate)
3018 3034
3019 3035 p1, p2 = self.dirstate.parents()
3020 3036 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or b'')
3021 3037 try:
3022 3038 self.hook(
3023 3039 b"precommit", throw=True, parent1=hookp1, parent2=hookp2
3024 3040 )
3025 3041 with self.transaction(b'commit'):
3026 3042 ret = self.commitctx(cctx, True)
3027 3043 # update bookmarks, dirstate and mergestate
3028 3044 bookmarks.update(self, [p1, p2], ret)
3029 3045 cctx.markcommitted(ret)
3030 3046 ms.reset()
3031 3047 except: # re-raises
3032 3048 if edited:
3033 3049 self.ui.write(
3034 3050 _(b'note: commit message saved in %s\n') % msgfn
3035 3051 )
3036 3052 self.ui.write(
3037 3053 _(
3038 3054 b"note: use 'hg commit --logfile "
3039 3055 b".hg/last-message.txt --edit' to reuse it\n"
3040 3056 )
3041 3057 )
3042 3058 raise
3043 3059
3044 3060 def commithook(unused_success):
3045 3061 # hack for command that use a temporary commit (eg: histedit)
3046 3062 # temporary commit got stripped before hook release
3047 3063 if self.changelog.hasnode(ret):
3048 3064 self.hook(
3049 3065 b"commit", node=hex(ret), parent1=hookp1, parent2=hookp2
3050 3066 )
3051 3067
3052 3068 self._afterlock(commithook)
3053 3069 return ret
3054 3070
3055 3071 @unfilteredmethod
3056 3072 def commitctx(self, ctx, error=False, origctx=None):
3057 3073 """Add a new revision to current repository.
3058 3074 Revision information is passed via the context argument.
3059 3075
3060 3076 ctx.files() should list all files involved in this commit, i.e.
3061 3077 modified/added/removed files. On merge, it may be wider than the
3062 3078 ctx.files() to be committed, since any file nodes derived directly
3063 3079 from p1 or p2 are excluded from the committed ctx.files().
3064 3080
3065 3081 origctx is for convert to work around the problem that bug
3066 3082 fixes to the files list in changesets change hashes. For
3067 3083 convert to be the identity, it can pass an origctx and this
3068 3084 function will use the same files list when it makes sense to
3069 3085 do so.
3070 3086 """
3071 3087
3072 3088 p1, p2 = ctx.p1(), ctx.p2()
3073 3089 user = ctx.user()
3074 3090
3075 3091 if self.filecopiesmode == b'changeset-sidedata':
3076 3092 writechangesetcopy = True
3077 3093 writefilecopymeta = True
3078 3094 writecopiesto = None
3079 3095 else:
3080 3096 writecopiesto = self.ui.config(b'experimental', b'copies.write-to')
3081 3097 writefilecopymeta = writecopiesto != b'changeset-only'
3082 3098 writechangesetcopy = writecopiesto in (
3083 3099 b'changeset-only',
3084 3100 b'compatibility',
3085 3101 )
3086 3102 p1copies, p2copies = None, None
3087 3103 if writechangesetcopy:
3088 3104 p1copies = ctx.p1copies()
3089 3105 p2copies = ctx.p2copies()
3090 3106 filesadded, filesremoved = None, None
3091 3107 with self.lock(), self.transaction(b"commit") as tr:
3092 3108 trp = weakref.proxy(tr)
3093 3109
3094 3110 if ctx.manifestnode():
3095 3111 # reuse an existing manifest revision
3096 3112 self.ui.debug(b'reusing known manifest\n')
3097 3113 mn = ctx.manifestnode()
3098 3114 files = ctx.files()
3099 3115 if writechangesetcopy:
3100 3116 filesadded = ctx.filesadded()
3101 3117 filesremoved = ctx.filesremoved()
3102 3118 elif ctx.files():
3103 3119 m1ctx = p1.manifestctx()
3104 3120 m2ctx = p2.manifestctx()
3105 3121 mctx = m1ctx.copy()
3106 3122
3107 3123 m = mctx.read()
3108 3124 m1 = m1ctx.read()
3109 3125 m2 = m2ctx.read()
3110 3126
3111 3127 # check in files
3112 3128 added = []
3113 3129 changed = []
3114 3130 removed = list(ctx.removed())
3115 3131 linkrev = len(self)
3116 3132 self.ui.note(_(b"committing files:\n"))
3117 3133 uipathfn = scmutil.getuipathfn(self)
3118 3134 for f in sorted(ctx.modified() + ctx.added()):
3119 3135 self.ui.note(uipathfn(f) + b"\n")
3120 3136 try:
3121 3137 fctx = ctx[f]
3122 3138 if fctx is None:
3123 3139 removed.append(f)
3124 3140 else:
3125 3141 added.append(f)
3126 3142 m[f] = self._filecommit(
3127 3143 fctx,
3128 3144 m1,
3129 3145 m2,
3130 3146 linkrev,
3131 3147 trp,
3132 3148 changed,
3133 3149 writefilecopymeta,
3134 3150 )
3135 3151 m.setflag(f, fctx.flags())
3136 3152 except OSError:
3137 3153 self.ui.warn(
3138 3154 _(b"trouble committing %s!\n") % uipathfn(f)
3139 3155 )
3140 3156 raise
3141 3157 except IOError as inst:
3142 3158 errcode = getattr(inst, 'errno', errno.ENOENT)
3143 3159 if error or errcode and errcode != errno.ENOENT:
3144 3160 self.ui.warn(
3145 3161 _(b"trouble committing %s!\n") % uipathfn(f)
3146 3162 )
3147 3163 raise
3148 3164
3149 3165 # update manifest
3150 3166 removed = [f for f in removed if f in m1 or f in m2]
3151 3167 drop = sorted([f for f in removed if f in m])
3152 3168 for f in drop:
3153 3169 del m[f]
3154 3170 if p2.rev() != nullrev:
3155 3171 rf = metadata.get_removal_filter(ctx, (p1, p2, m1, m2))
3156 3172 removed = [f for f in removed if not rf(f)]
3157 3173
3158 3174 files = changed + removed
3159 3175 md = None
3160 3176 if not files:
3161 3177 # if no "files" actually changed in terms of the changelog,
3162 3178 # try hard to detect unmodified manifest entry so that the
3163 3179 # exact same commit can be reproduced later on convert.
3164 3180 md = m1.diff(m, scmutil.matchfiles(self, ctx.files()))
3165 3181 if not files and md:
3166 3182 self.ui.debug(
3167 3183 b'not reusing manifest (no file change in '
3168 3184 b'changelog, but manifest differs)\n'
3169 3185 )
3170 3186 if files or md:
3171 3187 self.ui.note(_(b"committing manifest\n"))
3172 3188 # we're using narrowmatch here since it's already applied at
3173 3189 # other stages (such as dirstate.walk), so we're already
3174 3190 # ignoring things outside of narrowspec in most cases. The
3175 3191 # one case where we might have files outside the narrowspec
3176 3192 # at this point is merges, and we already error out in the
3177 3193 # case where the merge has files outside of the narrowspec,
3178 3194 # so this is safe.
3179 3195 mn = mctx.write(
3180 3196 trp,
3181 3197 linkrev,
3182 3198 p1.manifestnode(),
3183 3199 p2.manifestnode(),
3184 3200 added,
3185 3201 drop,
3186 3202 match=self.narrowmatch(),
3187 3203 )
3188 3204
3189 3205 if writechangesetcopy:
3190 3206 filesadded = [
3191 3207 f for f in changed if not (f in m1 or f in m2)
3192 3208 ]
3193 3209 filesremoved = removed
3194 3210 else:
3195 3211 self.ui.debug(
3196 3212 b'reusing manifest from p1 (listed files '
3197 3213 b'actually unchanged)\n'
3198 3214 )
3199 3215 mn = p1.manifestnode()
3200 3216 else:
3201 3217 self.ui.debug(b'reusing manifest from p1 (no file change)\n')
3202 3218 mn = p1.manifestnode()
3203 3219 files = []
3204 3220
3205 3221 if writecopiesto == b'changeset-only':
3206 3222 # If writing only to changeset extras, use None to indicate that
3207 3223 # no entry should be written. If writing to both, write an empty
3208 3224 # entry to prevent the reader from falling back to reading
3209 3225 # filelogs.
3210 3226 p1copies = p1copies or None
3211 3227 p2copies = p2copies or None
3212 3228 filesadded = filesadded or None
3213 3229 filesremoved = filesremoved or None
3214 3230
3215 3231 if origctx and origctx.manifestnode() == mn:
3216 3232 files = origctx.files()
3217 3233
3218 3234 # update changelog
3219 3235 self.ui.note(_(b"committing changelog\n"))
3220 3236 self.changelog.delayupdate(tr)
3221 3237 n = self.changelog.add(
3222 3238 mn,
3223 3239 files,
3224 3240 ctx.description(),
3225 3241 trp,
3226 3242 p1.node(),
3227 3243 p2.node(),
3228 3244 user,
3229 3245 ctx.date(),
3230 3246 ctx.extra().copy(),
3231 3247 p1copies,
3232 3248 p2copies,
3233 3249 filesadded,
3234 3250 filesremoved,
3235 3251 )
3236 3252 xp1, xp2 = p1.hex(), p2 and p2.hex() or b''
3237 3253 self.hook(
3238 3254 b'pretxncommit',
3239 3255 throw=True,
3240 3256 node=hex(n),
3241 3257 parent1=xp1,
3242 3258 parent2=xp2,
3243 3259 )
3244 3260 # set the new commit is proper phase
3245 3261 targetphase = subrepoutil.newcommitphase(self.ui, ctx)
3246 3262 if targetphase:
3247 3263 # retract boundary do not alter parent changeset.
3248 3264 # if a parent have higher the resulting phase will
3249 3265 # be compliant anyway
3250 3266 #
3251 3267 # if minimal phase was 0 we don't need to retract anything
3252 3268 phases.registernew(self, tr, targetphase, [n])
3253 3269 return n
3254 3270
3255 3271 @unfilteredmethod
3256 3272 def destroying(self):
3257 3273 '''Inform the repository that nodes are about to be destroyed.
3258 3274 Intended for use by strip and rollback, so there's a common
3259 3275 place for anything that has to be done before destroying history.
3260 3276
3261 3277 This is mostly useful for saving state that is in memory and waiting
3262 3278 to be flushed when the current lock is released. Because a call to
3263 3279 destroyed is imminent, the repo will be invalidated causing those
3264 3280 changes to stay in memory (waiting for the next unlock), or vanish
3265 3281 completely.
3266 3282 '''
3267 3283 # When using the same lock to commit and strip, the phasecache is left
3268 3284 # dirty after committing. Then when we strip, the repo is invalidated,
3269 3285 # causing those changes to disappear.
3270 3286 if '_phasecache' in vars(self):
3271 3287 self._phasecache.write()
3272 3288
3273 3289 @unfilteredmethod
3274 3290 def destroyed(self):
3275 3291 '''Inform the repository that nodes have been destroyed.
3276 3292 Intended for use by strip and rollback, so there's a common
3277 3293 place for anything that has to be done after destroying history.
3278 3294 '''
3279 3295 # When one tries to:
3280 3296 # 1) destroy nodes thus calling this method (e.g. strip)
3281 3297 # 2) use phasecache somewhere (e.g. commit)
3282 3298 #
3283 3299 # then 2) will fail because the phasecache contains nodes that were
3284 3300 # removed. We can either remove phasecache from the filecache,
3285 3301 # causing it to reload next time it is accessed, or simply filter
3286 3302 # the removed nodes now and write the updated cache.
3287 3303 self._phasecache.filterunknown(self)
3288 3304 self._phasecache.write()
3289 3305
3290 3306 # refresh all repository caches
3291 3307 self.updatecaches()
3292 3308
3293 3309 # Ensure the persistent tag cache is updated. Doing it now
3294 3310 # means that the tag cache only has to worry about destroyed
3295 3311 # heads immediately after a strip/rollback. That in turn
3296 3312 # guarantees that "cachetip == currenttip" (comparing both rev
3297 3313 # and node) always means no nodes have been added or destroyed.
3298 3314
3299 3315 # XXX this is suboptimal when qrefresh'ing: we strip the current
3300 3316 # head, refresh the tag cache, then immediately add a new head.
3301 3317 # But I think doing it this way is necessary for the "instant
3302 3318 # tag cache retrieval" case to work.
3303 3319 self.invalidate()
3304 3320
3305 3321 def status(
3306 3322 self,
3307 3323 node1=b'.',
3308 3324 node2=None,
3309 3325 match=None,
3310 3326 ignored=False,
3311 3327 clean=False,
3312 3328 unknown=False,
3313 3329 listsubrepos=False,
3314 3330 ):
3315 3331 '''a convenience method that calls node1.status(node2)'''
3316 3332 return self[node1].status(
3317 3333 node2, match, ignored, clean, unknown, listsubrepos
3318 3334 )
3319 3335
3320 3336 def addpostdsstatus(self, ps):
3321 3337 """Add a callback to run within the wlock, at the point at which status
3322 3338 fixups happen.
3323 3339
3324 3340 On status completion, callback(wctx, status) will be called with the
3325 3341 wlock held, unless the dirstate has changed from underneath or the wlock
3326 3342 couldn't be grabbed.
3327 3343
3328 3344 Callbacks should not capture and use a cached copy of the dirstate --
3329 3345 it might change in the meanwhile. Instead, they should access the
3330 3346 dirstate via wctx.repo().dirstate.
3331 3347
3332 3348 This list is emptied out after each status run -- extensions should
3333 3349 make sure it adds to this list each time dirstate.status is called.
3334 3350 Extensions should also make sure they don't call this for statuses
3335 3351 that don't involve the dirstate.
3336 3352 """
3337 3353
3338 3354 # The list is located here for uniqueness reasons -- it is actually
3339 3355 # managed by the workingctx, but that isn't unique per-repo.
3340 3356 self._postdsstatus.append(ps)
3341 3357
3342 3358 def postdsstatus(self):
3343 3359 """Used by workingctx to get the list of post-dirstate-status hooks."""
3344 3360 return self._postdsstatus
3345 3361
3346 3362 def clearpostdsstatus(self):
3347 3363 """Used by workingctx to clear post-dirstate-status hooks."""
3348 3364 del self._postdsstatus[:]
3349 3365
3350 3366 def heads(self, start=None):
3351 3367 if start is None:
3352 3368 cl = self.changelog
3353 3369 headrevs = reversed(cl.headrevs())
3354 3370 return [cl.node(rev) for rev in headrevs]
3355 3371
3356 3372 heads = self.changelog.heads(start)
3357 3373 # sort the output in rev descending order
3358 3374 return sorted(heads, key=self.changelog.rev, reverse=True)
3359 3375
3360 3376 def branchheads(self, branch=None, start=None, closed=False):
3361 3377 '''return a (possibly filtered) list of heads for the given branch
3362 3378
3363 3379 Heads are returned in topological order, from newest to oldest.
3364 3380 If branch is None, use the dirstate branch.
3365 3381 If start is not None, return only heads reachable from start.
3366 3382 If closed is True, return heads that are marked as closed as well.
3367 3383 '''
3368 3384 if branch is None:
3369 3385 branch = self[None].branch()
3370 3386 branches = self.branchmap()
3371 3387 if not branches.hasbranch(branch):
3372 3388 return []
3373 3389 # the cache returns heads ordered lowest to highest
3374 3390 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
3375 3391 if start is not None:
3376 3392 # filter out the heads that cannot be reached from startrev
3377 3393 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
3378 3394 bheads = [h for h in bheads if h in fbheads]
3379 3395 return bheads
3380 3396
3381 3397 def branches(self, nodes):
3382 3398 if not nodes:
3383 3399 nodes = [self.changelog.tip()]
3384 3400 b = []
3385 3401 for n in nodes:
3386 3402 t = n
3387 3403 while True:
3388 3404 p = self.changelog.parents(n)
3389 3405 if p[1] != nullid or p[0] == nullid:
3390 3406 b.append((t, n, p[0], p[1]))
3391 3407 break
3392 3408 n = p[0]
3393 3409 return b
3394 3410
3395 3411 def between(self, pairs):
3396 3412 r = []
3397 3413
3398 3414 for top, bottom in pairs:
3399 3415 n, l, i = top, [], 0
3400 3416 f = 1
3401 3417
3402 3418 while n != bottom and n != nullid:
3403 3419 p = self.changelog.parents(n)[0]
3404 3420 if i == f:
3405 3421 l.append(n)
3406 3422 f = f * 2
3407 3423 n = p
3408 3424 i += 1
3409 3425
3410 3426 r.append(l)
3411 3427
3412 3428 return r
3413 3429
3414 3430 def checkpush(self, pushop):
3415 3431 """Extensions can override this function if additional checks have
3416 3432 to be performed before pushing, or call it if they override push
3417 3433 command.
3418 3434 """
3419 3435
3420 3436 @unfilteredpropertycache
3421 3437 def prepushoutgoinghooks(self):
3422 3438 """Return util.hooks consists of a pushop with repo, remote, outgoing
3423 3439 methods, which are called before pushing changesets.
3424 3440 """
3425 3441 return util.hooks()
3426 3442
3427 3443 def pushkey(self, namespace, key, old, new):
3428 3444 try:
3429 3445 tr = self.currenttransaction()
3430 3446 hookargs = {}
3431 3447 if tr is not None:
3432 3448 hookargs.update(tr.hookargs)
3433 3449 hookargs = pycompat.strkwargs(hookargs)
3434 3450 hookargs['namespace'] = namespace
3435 3451 hookargs['key'] = key
3436 3452 hookargs['old'] = old
3437 3453 hookargs['new'] = new
3438 3454 self.hook(b'prepushkey', throw=True, **hookargs)
3439 3455 except error.HookAbort as exc:
3440 3456 self.ui.write_err(_(b"pushkey-abort: %s\n") % exc)
3441 3457 if exc.hint:
3442 3458 self.ui.write_err(_(b"(%s)\n") % exc.hint)
3443 3459 return False
3444 3460 self.ui.debug(b'pushing key for "%s:%s"\n' % (namespace, key))
3445 3461 ret = pushkey.push(self, namespace, key, old, new)
3446 3462
3447 3463 def runhook(unused_success):
3448 3464 self.hook(
3449 3465 b'pushkey',
3450 3466 namespace=namespace,
3451 3467 key=key,
3452 3468 old=old,
3453 3469 new=new,
3454 3470 ret=ret,
3455 3471 )
3456 3472
3457 3473 self._afterlock(runhook)
3458 3474 return ret
3459 3475
3460 3476 def listkeys(self, namespace):
3461 3477 self.hook(b'prelistkeys', throw=True, namespace=namespace)
3462 3478 self.ui.debug(b'listing keys for "%s"\n' % namespace)
3463 3479 values = pushkey.list(self, namespace)
3464 3480 self.hook(b'listkeys', namespace=namespace, values=values)
3465 3481 return values
3466 3482
3467 3483 def debugwireargs(self, one, two, three=None, four=None, five=None):
3468 3484 '''used to test argument passing over the wire'''
3469 3485 return b"%s %s %s %s %s" % (
3470 3486 one,
3471 3487 two,
3472 3488 pycompat.bytestr(three),
3473 3489 pycompat.bytestr(four),
3474 3490 pycompat.bytestr(five),
3475 3491 )
3476 3492
3477 3493 def savecommitmessage(self, text):
3478 3494 fp = self.vfs(b'last-message.txt', b'wb')
3479 3495 try:
3480 3496 fp.write(text)
3481 3497 finally:
3482 3498 fp.close()
3483 3499 return self.pathto(fp.name[len(self.root) + 1 :])
3484 3500
3485 3501
3486 3502 # used to avoid circular references so destructors work
3487 3503 def aftertrans(files):
3488 3504 renamefiles = [tuple(t) for t in files]
3489 3505
3490 3506 def a():
3491 3507 for vfs, src, dest in renamefiles:
3492 3508 # if src and dest refer to a same file, vfs.rename is a no-op,
3493 3509 # leaving both src and dest on disk. delete dest to make sure
3494 3510 # the rename couldn't be such a no-op.
3495 3511 vfs.tryunlink(dest)
3496 3512 try:
3497 3513 vfs.rename(src, dest)
3498 3514 except OSError: # journal file does not yet exist
3499 3515 pass
3500 3516
3501 3517 return a
3502 3518
3503 3519
3504 3520 def undoname(fn):
3505 3521 base, name = os.path.split(fn)
3506 3522 assert name.startswith(b'journal')
3507 3523 return os.path.join(base, name.replace(b'journal', b'undo', 1))
3508 3524
3509 3525
3510 3526 def instance(ui, path, create, intents=None, createopts=None):
3511 3527 localpath = util.urllocalpath(path)
3512 3528 if create:
3513 3529 createrepository(ui, localpath, createopts=createopts)
3514 3530
3515 3531 return makelocalrepository(ui, localpath, intents=intents)
3516 3532
3517 3533
3518 3534 def islocal(path):
3519 3535 return True
3520 3536
3521 3537
3522 3538 def defaultcreateopts(ui, createopts=None):
3523 3539 """Populate the default creation options for a repository.
3524 3540
3525 3541 A dictionary of explicitly requested creation options can be passed
3526 3542 in. Missing keys will be populated.
3527 3543 """
3528 3544 createopts = dict(createopts or {})
3529 3545
3530 3546 if b'backend' not in createopts:
3531 3547 # experimental config: storage.new-repo-backend
3532 3548 createopts[b'backend'] = ui.config(b'storage', b'new-repo-backend')
3533 3549
3534 3550 return createopts
3535 3551
3536 3552
3537 3553 def newreporequirements(ui, createopts):
3538 3554 """Determine the set of requirements for a new local repository.
3539 3555
3540 3556 Extensions can wrap this function to specify custom requirements for
3541 3557 new repositories.
3542 3558 """
3543 3559 # If the repo is being created from a shared repository, we copy
3544 3560 # its requirements.
3545 3561 if b'sharedrepo' in createopts:
3546 3562 requirements = set(createopts[b'sharedrepo'].requirements)
3547 3563 if createopts.get(b'sharedrelative'):
3548 3564 requirements.add(b'relshared')
3549 3565 else:
3550 3566 requirements.add(b'shared')
3551 3567
3552 3568 return requirements
3553 3569
3554 3570 if b'backend' not in createopts:
3555 3571 raise error.ProgrammingError(
3556 3572 b'backend key not present in createopts; '
3557 3573 b'was defaultcreateopts() called?'
3558 3574 )
3559 3575
3560 3576 if createopts[b'backend'] != b'revlogv1':
3561 3577 raise error.Abort(
3562 3578 _(
3563 3579 b'unable to determine repository requirements for '
3564 3580 b'storage backend: %s'
3565 3581 )
3566 3582 % createopts[b'backend']
3567 3583 )
3568 3584
3569 3585 requirements = {b'revlogv1'}
3570 3586 if ui.configbool(b'format', b'usestore'):
3571 3587 requirements.add(b'store')
3572 3588 if ui.configbool(b'format', b'usefncache'):
3573 3589 requirements.add(b'fncache')
3574 3590 if ui.configbool(b'format', b'dotencode'):
3575 3591 requirements.add(b'dotencode')
3576 3592
3577 3593 compengines = ui.configlist(b'format', b'revlog-compression')
3578 3594 for compengine in compengines:
3579 3595 if compengine in util.compengines:
3580 3596 break
3581 3597 else:
3582 3598 raise error.Abort(
3583 3599 _(
3584 3600 b'compression engines %s defined by '
3585 3601 b'format.revlog-compression not available'
3586 3602 )
3587 3603 % b', '.join(b'"%s"' % e for e in compengines),
3588 3604 hint=_(
3589 3605 b'run "hg debuginstall" to list available '
3590 3606 b'compression engines'
3591 3607 ),
3592 3608 )
3593 3609
3594 3610 # zlib is the historical default and doesn't need an explicit requirement.
3595 3611 if compengine == b'zstd':
3596 3612 requirements.add(b'revlog-compression-zstd')
3597 3613 elif compengine != b'zlib':
3598 3614 requirements.add(b'exp-compression-%s' % compengine)
3599 3615
3600 3616 if scmutil.gdinitconfig(ui):
3601 3617 requirements.add(b'generaldelta')
3602 3618 if ui.configbool(b'format', b'sparse-revlog'):
3603 3619 requirements.add(SPARSEREVLOG_REQUIREMENT)
3604 3620
3605 3621 # experimental config: format.exp-use-side-data
3606 3622 if ui.configbool(b'format', b'exp-use-side-data'):
3607 3623 requirements.add(SIDEDATA_REQUIREMENT)
3608 3624 # experimental config: format.exp-use-copies-side-data-changeset
3609 3625 if ui.configbool(b'format', b'exp-use-copies-side-data-changeset'):
3610 3626 requirements.add(SIDEDATA_REQUIREMENT)
3611 3627 requirements.add(COPIESSDC_REQUIREMENT)
3612 3628 if ui.configbool(b'experimental', b'treemanifest'):
3613 3629 requirements.add(b'treemanifest')
3614 3630
3615 3631 revlogv2 = ui.config(b'experimental', b'revlogv2')
3616 3632 if revlogv2 == b'enable-unstable-format-and-corrupt-my-data':
3617 3633 requirements.remove(b'revlogv1')
3618 3634 # generaldelta is implied by revlogv2.
3619 3635 requirements.discard(b'generaldelta')
3620 3636 requirements.add(REVLOGV2_REQUIREMENT)
3621 3637 # experimental config: format.internal-phase
3622 3638 if ui.configbool(b'format', b'internal-phase'):
3623 3639 requirements.add(b'internal-phase')
3624 3640
3625 3641 if createopts.get(b'narrowfiles'):
3626 3642 requirements.add(repository.NARROW_REQUIREMENT)
3627 3643
3628 3644 if createopts.get(b'lfs'):
3629 3645 requirements.add(b'lfs')
3630 3646
3631 3647 if ui.configbool(b'format', b'bookmarks-in-store'):
3632 3648 requirements.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3633 3649
3634 3650 if ui.configbool(b'format', b'use-persistent-nodemap'):
3635 3651 requirements.add(NODEMAP_REQUIREMENT)
3636 3652
3637 3653 return requirements
3638 3654
3639 3655
3640 3656 def filterknowncreateopts(ui, createopts):
3641 3657 """Filters a dict of repo creation options against options that are known.
3642 3658
3643 3659 Receives a dict of repo creation options and returns a dict of those
3644 3660 options that we don't know how to handle.
3645 3661
3646 3662 This function is called as part of repository creation. If the
3647 3663 returned dict contains any items, repository creation will not
3648 3664 be allowed, as it means there was a request to create a repository
3649 3665 with options not recognized by loaded code.
3650 3666
3651 3667 Extensions can wrap this function to filter out creation options
3652 3668 they know how to handle.
3653 3669 """
3654 3670 known = {
3655 3671 b'backend',
3656 3672 b'lfs',
3657 3673 b'narrowfiles',
3658 3674 b'sharedrepo',
3659 3675 b'sharedrelative',
3660 3676 b'shareditems',
3661 3677 b'shallowfilestore',
3662 3678 }
3663 3679
3664 3680 return {k: v for k, v in createopts.items() if k not in known}
3665 3681
3666 3682
3667 3683 def createrepository(ui, path, createopts=None):
3668 3684 """Create a new repository in a vfs.
3669 3685
3670 3686 ``path`` path to the new repo's working directory.
3671 3687 ``createopts`` options for the new repository.
3672 3688
3673 3689 The following keys for ``createopts`` are recognized:
3674 3690
3675 3691 backend
3676 3692 The storage backend to use.
3677 3693 lfs
3678 3694 Repository will be created with ``lfs`` requirement. The lfs extension
3679 3695 will automatically be loaded when the repository is accessed.
3680 3696 narrowfiles
3681 3697 Set up repository to support narrow file storage.
3682 3698 sharedrepo
3683 3699 Repository object from which storage should be shared.
3684 3700 sharedrelative
3685 3701 Boolean indicating if the path to the shared repo should be
3686 3702 stored as relative. By default, the pointer to the "parent" repo
3687 3703 is stored as an absolute path.
3688 3704 shareditems
3689 3705 Set of items to share to the new repository (in addition to storage).
3690 3706 shallowfilestore
3691 3707 Indicates that storage for files should be shallow (not all ancestor
3692 3708 revisions are known).
3693 3709 """
3694 3710 createopts = defaultcreateopts(ui, createopts=createopts)
3695 3711
3696 3712 unknownopts = filterknowncreateopts(ui, createopts)
3697 3713
3698 3714 if not isinstance(unknownopts, dict):
3699 3715 raise error.ProgrammingError(
3700 3716 b'filterknowncreateopts() did not return a dict'
3701 3717 )
3702 3718
3703 3719 if unknownopts:
3704 3720 raise error.Abort(
3705 3721 _(
3706 3722 b'unable to create repository because of unknown '
3707 3723 b'creation option: %s'
3708 3724 )
3709 3725 % b', '.join(sorted(unknownopts)),
3710 3726 hint=_(b'is a required extension not loaded?'),
3711 3727 )
3712 3728
3713 3729 requirements = newreporequirements(ui, createopts=createopts)
3714 3730
3715 3731 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
3716 3732
3717 3733 hgvfs = vfsmod.vfs(wdirvfs.join(b'.hg'))
3718 3734 if hgvfs.exists():
3719 3735 raise error.RepoError(_(b'repository %s already exists') % path)
3720 3736
3721 3737 if b'sharedrepo' in createopts:
3722 3738 sharedpath = createopts[b'sharedrepo'].sharedpath
3723 3739
3724 3740 if createopts.get(b'sharedrelative'):
3725 3741 try:
3726 3742 sharedpath = os.path.relpath(sharedpath, hgvfs.base)
3727 3743 except (IOError, ValueError) as e:
3728 3744 # ValueError is raised on Windows if the drive letters differ
3729 3745 # on each path.
3730 3746 raise error.Abort(
3731 3747 _(b'cannot calculate relative path'),
3732 3748 hint=stringutil.forcebytestr(e),
3733 3749 )
3734 3750
3735 3751 if not wdirvfs.exists():
3736 3752 wdirvfs.makedirs()
3737 3753
3738 3754 hgvfs.makedir(notindexed=True)
3739 3755 if b'sharedrepo' not in createopts:
3740 3756 hgvfs.mkdir(b'cache')
3741 3757 hgvfs.mkdir(b'wcache')
3742 3758
3743 3759 if b'store' in requirements and b'sharedrepo' not in createopts:
3744 3760 hgvfs.mkdir(b'store')
3745 3761
3746 3762 # We create an invalid changelog outside the store so very old
3747 3763 # Mercurial versions (which didn't know about the requirements
3748 3764 # file) encounter an error on reading the changelog. This
3749 3765 # effectively locks out old clients and prevents them from
3750 3766 # mucking with a repo in an unknown format.
3751 3767 #
3752 3768 # The revlog header has version 2, which won't be recognized by
3753 3769 # such old clients.
3754 3770 hgvfs.append(
3755 3771 b'00changelog.i',
3756 3772 b'\0\0\0\2 dummy changelog to prevent using the old repo '
3757 3773 b'layout',
3758 3774 )
3759 3775
3760 3776 scmutil.writerequires(hgvfs, requirements)
3761 3777
3762 3778 # Write out file telling readers where to find the shared store.
3763 3779 if b'sharedrepo' in createopts:
3764 3780 hgvfs.write(b'sharedpath', sharedpath)
3765 3781
3766 3782 if createopts.get(b'shareditems'):
3767 3783 shared = b'\n'.join(sorted(createopts[b'shareditems'])) + b'\n'
3768 3784 hgvfs.write(b'shared', shared)
3769 3785
3770 3786
3771 3787 def poisonrepository(repo):
3772 3788 """Poison a repository instance so it can no longer be used."""
3773 3789 # Perform any cleanup on the instance.
3774 3790 repo.close()
3775 3791
3776 3792 # Our strategy is to replace the type of the object with one that
3777 3793 # has all attribute lookups result in error.
3778 3794 #
3779 3795 # But we have to allow the close() method because some constructors
3780 3796 # of repos call close() on repo references.
3781 3797 class poisonedrepository(object):
3782 3798 def __getattribute__(self, item):
3783 3799 if item == 'close':
3784 3800 return object.__getattribute__(self, item)
3785 3801
3786 3802 raise error.ProgrammingError(
3787 3803 b'repo instances should not be used after unshare'
3788 3804 )
3789 3805
3790 3806 def close(self):
3791 3807 pass
3792 3808
3793 3809 # We may have a repoview, which intercepts __setattr__. So be sure
3794 3810 # we operate at the lowest level possible.
3795 3811 object.__setattr__(repo, '__class__', poisonedrepository)
General Comments 0
You need to be logged in to leave comments. Login now