##// END OF EJS Templates
commitctx: extract the function in a dedicated module...
marmoute -
r45759:ae5c1a3b default
parent child Browse files
Show More
@@ -0,0 +1,215 b''
1 # commit.py - fonction to perform commit
2 #
3 # This software may be used and distributed according to the terms of the
4 # GNU General Public License version 2 or any later version.
5
6 from __future__ import absolute_import
7
8 import errno
9 import weakref
10
11 from .i18n import _
12 from .node import (
13 hex,
14 nullrev,
15 )
16
17 from . import (
18 metadata,
19 phases,
20 scmutil,
21 subrepoutil,
22 )
23
24
25 def commitctx(repo, ctx, error=False, origctx=None):
26 """Add a new revision to the target repository.
27 Revision information is passed via the context argument.
28
29 ctx.files() should list all files involved in this commit, i.e.
30 modified/added/removed files. On merge, it may be wider than the
31 ctx.files() to be committed, since any file nodes derived directly
32 from p1 or p2 are excluded from the committed ctx.files().
33
34 origctx is for convert to work around the problem that bug
35 fixes to the files list in changesets change hashes. For
36 convert to be the identity, it can pass an origctx and this
37 function will use the same files list when it makes sense to
38 do so.
39 """
40 repo = repo.unfiltered()
41
42 p1, p2 = ctx.p1(), ctx.p2()
43 user = ctx.user()
44
45 if repo.filecopiesmode == b'changeset-sidedata':
46 writechangesetcopy = True
47 writefilecopymeta = True
48 writecopiesto = None
49 else:
50 writecopiesto = repo.ui.config(b'experimental', b'copies.write-to')
51 writefilecopymeta = writecopiesto != b'changeset-only'
52 writechangesetcopy = writecopiesto in (
53 b'changeset-only',
54 b'compatibility',
55 )
56 p1copies, p2copies = None, None
57 if writechangesetcopy:
58 p1copies = ctx.p1copies()
59 p2copies = ctx.p2copies()
60 filesadded, filesremoved = None, None
61 with repo.lock(), repo.transaction(b"commit") as tr:
62 trp = weakref.proxy(tr)
63
64 if ctx.manifestnode():
65 # reuse an existing manifest revision
66 repo.ui.debug(b'reusing known manifest\n')
67 mn = ctx.manifestnode()
68 files = ctx.files()
69 if writechangesetcopy:
70 filesadded = ctx.filesadded()
71 filesremoved = ctx.filesremoved()
72 elif not ctx.files():
73 repo.ui.debug(b'reusing manifest from p1 (no file change)\n')
74 mn = p1.manifestnode()
75 files = []
76 else:
77 m1ctx = p1.manifestctx()
78 m2ctx = p2.manifestctx()
79 mctx = m1ctx.copy()
80
81 m = mctx.read()
82 m1 = m1ctx.read()
83 m2 = m2ctx.read()
84
85 # check in files
86 added = []
87 filesadded = []
88 removed = list(ctx.removed())
89 touched = []
90 linkrev = len(repo)
91 repo.ui.note(_(b"committing files:\n"))
92 uipathfn = scmutil.getuipathfn(repo)
93 for f in sorted(ctx.modified() + ctx.added()):
94 repo.ui.note(uipathfn(f) + b"\n")
95 try:
96 fctx = ctx[f]
97 if fctx is None:
98 removed.append(f)
99 else:
100 added.append(f)
101 m[f], is_touched = repo._filecommit(
102 fctx, m1, m2, linkrev, trp, writefilecopymeta,
103 )
104 if is_touched:
105 touched.append(f)
106 if writechangesetcopy and is_touched == 'added':
107 filesadded.append(f)
108 m.setflag(f, fctx.flags())
109 except OSError:
110 repo.ui.warn(_(b"trouble committing %s!\n") % uipathfn(f))
111 raise
112 except IOError as inst:
113 errcode = getattr(inst, 'errno', errno.ENOENT)
114 if error or errcode and errcode != errno.ENOENT:
115 repo.ui.warn(
116 _(b"trouble committing %s!\n") % uipathfn(f)
117 )
118 raise
119
120 # update manifest
121 removed = [f for f in removed if f in m1 or f in m2]
122 drop = sorted([f for f in removed if f in m])
123 for f in drop:
124 del m[f]
125 if p2.rev() != nullrev:
126 rf = metadata.get_removal_filter(ctx, (p1, p2, m1, m2))
127 removed = [f for f in removed if not rf(f)]
128
129 touched.extend(removed)
130
131 if writechangesetcopy:
132 filesremoved = removed
133
134 files = touched
135 md = None
136 if not files:
137 # if no "files" actually changed in terms of the changelog,
138 # try hard to detect unmodified manifest entry so that the
139 # exact same commit can be reproduced later on convert.
140 md = m1.diff(m, scmutil.matchfiles(repo, ctx.files()))
141 if not files and md:
142 repo.ui.debug(
143 b'not reusing manifest (no file change in '
144 b'changelog, but manifest differs)\n'
145 )
146 if files or md:
147 repo.ui.note(_(b"committing manifest\n"))
148 # we're using narrowmatch here since it's already applied at
149 # other stages (such as dirstate.walk), so we're already
150 # ignoring things outside of narrowspec in most cases. The
151 # one case where we might have files outside the narrowspec
152 # at this point is merges, and we already error out in the
153 # case where the merge has files outside of the narrowspec,
154 # so this is safe.
155 mn = mctx.write(
156 trp,
157 linkrev,
158 p1.manifestnode(),
159 p2.manifestnode(),
160 added,
161 drop,
162 match=repo.narrowmatch(),
163 )
164 else:
165 repo.ui.debug(
166 b'reusing manifest from p1 (listed files '
167 b'actually unchanged)\n'
168 )
169 mn = p1.manifestnode()
170
171 if writecopiesto == b'changeset-only':
172 # If writing only to changeset extras, use None to indicate that
173 # no entry should be written. If writing to both, write an empty
174 # entry to prevent the reader from falling back to reading
175 # filelogs.
176 p1copies = p1copies or None
177 p2copies = p2copies or None
178 filesadded = filesadded or None
179 filesremoved = filesremoved or None
180
181 if origctx and origctx.manifestnode() == mn:
182 files = origctx.files()
183
184 # update changelog
185 repo.ui.note(_(b"committing changelog\n"))
186 repo.changelog.delayupdate(tr)
187 n = repo.changelog.add(
188 mn,
189 files,
190 ctx.description(),
191 trp,
192 p1.node(),
193 p2.node(),
194 user,
195 ctx.date(),
196 ctx.extra().copy(),
197 p1copies,
198 p2copies,
199 filesadded,
200 filesremoved,
201 )
202 xp1, xp2 = p1.hex(), p2 and p2.hex() or b''
203 repo.hook(
204 b'pretxncommit', throw=True, node=hex(n), parent1=xp1, parent2=xp2,
205 )
206 # set the new commit is proper phase
207 targetphase = subrepoutil.newcommitphase(repo.ui, ctx)
208 if targetphase:
209 # retract boundary do not alter parent changeset.
210 # if a parent have higher the resulting phase will
211 # be compliant anyway
212 #
213 # if minimal phase was 0 we don't need to retract anything
214 phases.registernew(repo, tr, targetphase, [n])
215 return n
@@ -1,3806 +1,3612 b''
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import errno
11 11 import os
12 12 import random
13 13 import sys
14 14 import time
15 15 import weakref
16 16
17 17 from .i18n import _
18 18 from .node import (
19 19 bin,
20 20 hex,
21 21 nullid,
22 22 nullrev,
23 23 short,
24 24 )
25 25 from .pycompat import (
26 26 delattr,
27 27 getattr,
28 28 )
29 29 from . import (
30 30 bookmarks,
31 31 branchmap,
32 32 bundle2,
33 33 changegroup,
34 34 color,
35 commit,
35 36 context,
36 37 dirstate,
37 38 dirstateguard,
38 39 discovery,
39 40 encoding,
40 41 error,
41 42 exchange,
42 43 extensions,
43 44 filelog,
44 45 hook,
45 46 lock as lockmod,
46 47 match as matchmod,
47 48 mergestate as mergestatemod,
48 49 mergeutil,
49 metadata,
50 50 namespaces,
51 51 narrowspec,
52 52 obsolete,
53 53 pathutil,
54 54 phases,
55 55 pushkey,
56 56 pycompat,
57 57 rcutil,
58 58 repoview,
59 59 revset,
60 60 revsetlang,
61 61 scmutil,
62 62 sparse,
63 63 store as storemod,
64 64 subrepoutil,
65 65 tags as tagsmod,
66 66 transaction,
67 67 txnutil,
68 68 util,
69 69 vfs as vfsmod,
70 70 )
71 71
72 72 from .interfaces import (
73 73 repository,
74 74 util as interfaceutil,
75 75 )
76 76
77 77 from .utils import (
78 78 hashutil,
79 79 procutil,
80 80 stringutil,
81 81 )
82 82
83 83 from .revlogutils import constants as revlogconst
84 84
85 85 release = lockmod.release
86 86 urlerr = util.urlerr
87 87 urlreq = util.urlreq
88 88
89 89 # set of (path, vfs-location) tuples. vfs-location is:
90 90 # - 'plain for vfs relative paths
91 91 # - '' for svfs relative paths
92 92 _cachedfiles = set()
93 93
94 94
95 95 class _basefilecache(scmutil.filecache):
96 96 """All filecache usage on repo are done for logic that should be unfiltered
97 97 """
98 98
99 99 def __get__(self, repo, type=None):
100 100 if repo is None:
101 101 return self
102 102 # proxy to unfiltered __dict__ since filtered repo has no entry
103 103 unfi = repo.unfiltered()
104 104 try:
105 105 return unfi.__dict__[self.sname]
106 106 except KeyError:
107 107 pass
108 108 return super(_basefilecache, self).__get__(unfi, type)
109 109
110 110 def set(self, repo, value):
111 111 return super(_basefilecache, self).set(repo.unfiltered(), value)
112 112
113 113
114 114 class repofilecache(_basefilecache):
115 115 """filecache for files in .hg but outside of .hg/store"""
116 116
117 117 def __init__(self, *paths):
118 118 super(repofilecache, self).__init__(*paths)
119 119 for path in paths:
120 120 _cachedfiles.add((path, b'plain'))
121 121
122 122 def join(self, obj, fname):
123 123 return obj.vfs.join(fname)
124 124
125 125
126 126 class storecache(_basefilecache):
127 127 """filecache for files in the store"""
128 128
129 129 def __init__(self, *paths):
130 130 super(storecache, self).__init__(*paths)
131 131 for path in paths:
132 132 _cachedfiles.add((path, b''))
133 133
134 134 def join(self, obj, fname):
135 135 return obj.sjoin(fname)
136 136
137 137
138 138 class mixedrepostorecache(_basefilecache):
139 139 """filecache for a mix files in .hg/store and outside"""
140 140
141 141 def __init__(self, *pathsandlocations):
142 142 # scmutil.filecache only uses the path for passing back into our
143 143 # join(), so we can safely pass a list of paths and locations
144 144 super(mixedrepostorecache, self).__init__(*pathsandlocations)
145 145 _cachedfiles.update(pathsandlocations)
146 146
147 147 def join(self, obj, fnameandlocation):
148 148 fname, location = fnameandlocation
149 149 if location == b'plain':
150 150 return obj.vfs.join(fname)
151 151 else:
152 152 if location != b'':
153 153 raise error.ProgrammingError(
154 154 b'unexpected location: %s' % location
155 155 )
156 156 return obj.sjoin(fname)
157 157
158 158
159 159 def isfilecached(repo, name):
160 160 """check if a repo has already cached "name" filecache-ed property
161 161
162 162 This returns (cachedobj-or-None, iscached) tuple.
163 163 """
164 164 cacheentry = repo.unfiltered()._filecache.get(name, None)
165 165 if not cacheentry:
166 166 return None, False
167 167 return cacheentry.obj, True
168 168
169 169
170 170 class unfilteredpropertycache(util.propertycache):
171 171 """propertycache that apply to unfiltered repo only"""
172 172
173 173 def __get__(self, repo, type=None):
174 174 unfi = repo.unfiltered()
175 175 if unfi is repo:
176 176 return super(unfilteredpropertycache, self).__get__(unfi)
177 177 return getattr(unfi, self.name)
178 178
179 179
180 180 class filteredpropertycache(util.propertycache):
181 181 """propertycache that must take filtering in account"""
182 182
183 183 def cachevalue(self, obj, value):
184 184 object.__setattr__(obj, self.name, value)
185 185
186 186
187 187 def hasunfilteredcache(repo, name):
188 188 """check if a repo has an unfilteredpropertycache value for <name>"""
189 189 return name in vars(repo.unfiltered())
190 190
191 191
192 192 def unfilteredmethod(orig):
193 193 """decorate method that always need to be run on unfiltered version"""
194 194
195 195 def wrapper(repo, *args, **kwargs):
196 196 return orig(repo.unfiltered(), *args, **kwargs)
197 197
198 198 return wrapper
199 199
200 200
201 201 moderncaps = {
202 202 b'lookup',
203 203 b'branchmap',
204 204 b'pushkey',
205 205 b'known',
206 206 b'getbundle',
207 207 b'unbundle',
208 208 }
209 209 legacycaps = moderncaps.union({b'changegroupsubset'})
210 210
211 211
212 212 @interfaceutil.implementer(repository.ipeercommandexecutor)
213 213 class localcommandexecutor(object):
214 214 def __init__(self, peer):
215 215 self._peer = peer
216 216 self._sent = False
217 217 self._closed = False
218 218
219 219 def __enter__(self):
220 220 return self
221 221
222 222 def __exit__(self, exctype, excvalue, exctb):
223 223 self.close()
224 224
225 225 def callcommand(self, command, args):
226 226 if self._sent:
227 227 raise error.ProgrammingError(
228 228 b'callcommand() cannot be used after sendcommands()'
229 229 )
230 230
231 231 if self._closed:
232 232 raise error.ProgrammingError(
233 233 b'callcommand() cannot be used after close()'
234 234 )
235 235
236 236 # We don't need to support anything fancy. Just call the named
237 237 # method on the peer and return a resolved future.
238 238 fn = getattr(self._peer, pycompat.sysstr(command))
239 239
240 240 f = pycompat.futures.Future()
241 241
242 242 try:
243 243 result = fn(**pycompat.strkwargs(args))
244 244 except Exception:
245 245 pycompat.future_set_exception_info(f, sys.exc_info()[1:])
246 246 else:
247 247 f.set_result(result)
248 248
249 249 return f
250 250
251 251 def sendcommands(self):
252 252 self._sent = True
253 253
254 254 def close(self):
255 255 self._closed = True
256 256
257 257
258 258 @interfaceutil.implementer(repository.ipeercommands)
259 259 class localpeer(repository.peer):
260 260 '''peer for a local repo; reflects only the most recent API'''
261 261
262 262 def __init__(self, repo, caps=None):
263 263 super(localpeer, self).__init__()
264 264
265 265 if caps is None:
266 266 caps = moderncaps.copy()
267 267 self._repo = repo.filtered(b'served')
268 268 self.ui = repo.ui
269 269 self._caps = repo._restrictcapabilities(caps)
270 270
271 271 # Begin of _basepeer interface.
272 272
273 273 def url(self):
274 274 return self._repo.url()
275 275
276 276 def local(self):
277 277 return self._repo
278 278
279 279 def peer(self):
280 280 return self
281 281
282 282 def canpush(self):
283 283 return True
284 284
285 285 def close(self):
286 286 self._repo.close()
287 287
288 288 # End of _basepeer interface.
289 289
290 290 # Begin of _basewirecommands interface.
291 291
292 292 def branchmap(self):
293 293 return self._repo.branchmap()
294 294
295 295 def capabilities(self):
296 296 return self._caps
297 297
298 298 def clonebundles(self):
299 299 return self._repo.tryread(b'clonebundles.manifest')
300 300
301 301 def debugwireargs(self, one, two, three=None, four=None, five=None):
302 302 """Used to test argument passing over the wire"""
303 303 return b"%s %s %s %s %s" % (
304 304 one,
305 305 two,
306 306 pycompat.bytestr(three),
307 307 pycompat.bytestr(four),
308 308 pycompat.bytestr(five),
309 309 )
310 310
311 311 def getbundle(
312 312 self, source, heads=None, common=None, bundlecaps=None, **kwargs
313 313 ):
314 314 chunks = exchange.getbundlechunks(
315 315 self._repo,
316 316 source,
317 317 heads=heads,
318 318 common=common,
319 319 bundlecaps=bundlecaps,
320 320 **kwargs
321 321 )[1]
322 322 cb = util.chunkbuffer(chunks)
323 323
324 324 if exchange.bundle2requested(bundlecaps):
325 325 # When requesting a bundle2, getbundle returns a stream to make the
326 326 # wire level function happier. We need to build a proper object
327 327 # from it in local peer.
328 328 return bundle2.getunbundler(self.ui, cb)
329 329 else:
330 330 return changegroup.getunbundler(b'01', cb, None)
331 331
332 332 def heads(self):
333 333 return self._repo.heads()
334 334
335 335 def known(self, nodes):
336 336 return self._repo.known(nodes)
337 337
338 338 def listkeys(self, namespace):
339 339 return self._repo.listkeys(namespace)
340 340
341 341 def lookup(self, key):
342 342 return self._repo.lookup(key)
343 343
344 344 def pushkey(self, namespace, key, old, new):
345 345 return self._repo.pushkey(namespace, key, old, new)
346 346
347 347 def stream_out(self):
348 348 raise error.Abort(_(b'cannot perform stream clone against local peer'))
349 349
350 350 def unbundle(self, bundle, heads, url):
351 351 """apply a bundle on a repo
352 352
353 353 This function handles the repo locking itself."""
354 354 try:
355 355 try:
356 356 bundle = exchange.readbundle(self.ui, bundle, None)
357 357 ret = exchange.unbundle(self._repo, bundle, heads, b'push', url)
358 358 if util.safehasattr(ret, b'getchunks'):
359 359 # This is a bundle20 object, turn it into an unbundler.
360 360 # This little dance should be dropped eventually when the
361 361 # API is finally improved.
362 362 stream = util.chunkbuffer(ret.getchunks())
363 363 ret = bundle2.getunbundler(self.ui, stream)
364 364 return ret
365 365 except Exception as exc:
366 366 # If the exception contains output salvaged from a bundle2
367 367 # reply, we need to make sure it is printed before continuing
368 368 # to fail. So we build a bundle2 with such output and consume
369 369 # it directly.
370 370 #
371 371 # This is not very elegant but allows a "simple" solution for
372 372 # issue4594
373 373 output = getattr(exc, '_bundle2salvagedoutput', ())
374 374 if output:
375 375 bundler = bundle2.bundle20(self._repo.ui)
376 376 for out in output:
377 377 bundler.addpart(out)
378 378 stream = util.chunkbuffer(bundler.getchunks())
379 379 b = bundle2.getunbundler(self.ui, stream)
380 380 bundle2.processbundle(self._repo, b)
381 381 raise
382 382 except error.PushRaced as exc:
383 383 raise error.ResponseError(
384 384 _(b'push failed:'), stringutil.forcebytestr(exc)
385 385 )
386 386
387 387 # End of _basewirecommands interface.
388 388
389 389 # Begin of peer interface.
390 390
391 391 def commandexecutor(self):
392 392 return localcommandexecutor(self)
393 393
394 394 # End of peer interface.
395 395
396 396
397 397 @interfaceutil.implementer(repository.ipeerlegacycommands)
398 398 class locallegacypeer(localpeer):
399 399 '''peer extension which implements legacy methods too; used for tests with
400 400 restricted capabilities'''
401 401
402 402 def __init__(self, repo):
403 403 super(locallegacypeer, self).__init__(repo, caps=legacycaps)
404 404
405 405 # Begin of baselegacywirecommands interface.
406 406
407 407 def between(self, pairs):
408 408 return self._repo.between(pairs)
409 409
410 410 def branches(self, nodes):
411 411 return self._repo.branches(nodes)
412 412
413 413 def changegroup(self, nodes, source):
414 414 outgoing = discovery.outgoing(
415 415 self._repo, missingroots=nodes, ancestorsof=self._repo.heads()
416 416 )
417 417 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
418 418
419 419 def changegroupsubset(self, bases, heads, source):
420 420 outgoing = discovery.outgoing(
421 421 self._repo, missingroots=bases, ancestorsof=heads
422 422 )
423 423 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
424 424
425 425 # End of baselegacywirecommands interface.
426 426
427 427
428 428 # Increment the sub-version when the revlog v2 format changes to lock out old
429 429 # clients.
430 430 REVLOGV2_REQUIREMENT = b'exp-revlogv2.1'
431 431
432 432 # A repository with the sparserevlog feature will have delta chains that
433 433 # can spread over a larger span. Sparse reading cuts these large spans into
434 434 # pieces, so that each piece isn't too big.
435 435 # Without the sparserevlog capability, reading from the repository could use
436 436 # huge amounts of memory, because the whole span would be read at once,
437 437 # including all the intermediate revisions that aren't pertinent for the chain.
438 438 # This is why once a repository has enabled sparse-read, it becomes required.
439 439 SPARSEREVLOG_REQUIREMENT = b'sparserevlog'
440 440
441 441 # A repository with the sidedataflag requirement will allow to store extra
442 442 # information for revision without altering their original hashes.
443 443 SIDEDATA_REQUIREMENT = b'exp-sidedata-flag'
444 444
445 445 # A repository with the the copies-sidedata-changeset requirement will store
446 446 # copies related information in changeset's sidedata.
447 447 COPIESSDC_REQUIREMENT = b'exp-copies-sidedata-changeset'
448 448
449 449 # The repository use persistent nodemap for the changelog and the manifest.
450 450 NODEMAP_REQUIREMENT = b'persistent-nodemap'
451 451
452 452 # Functions receiving (ui, features) that extensions can register to impact
453 453 # the ability to load repositories with custom requirements. Only
454 454 # functions defined in loaded extensions are called.
455 455 #
456 456 # The function receives a set of requirement strings that the repository
457 457 # is capable of opening. Functions will typically add elements to the
458 458 # set to reflect that the extension knows how to handle that requirements.
459 459 featuresetupfuncs = set()
460 460
461 461
462 462 def makelocalrepository(baseui, path, intents=None):
463 463 """Create a local repository object.
464 464
465 465 Given arguments needed to construct a local repository, this function
466 466 performs various early repository loading functionality (such as
467 467 reading the ``.hg/requires`` and ``.hg/hgrc`` files), validates that
468 468 the repository can be opened, derives a type suitable for representing
469 469 that repository, and returns an instance of it.
470 470
471 471 The returned object conforms to the ``repository.completelocalrepository``
472 472 interface.
473 473
474 474 The repository type is derived by calling a series of factory functions
475 475 for each aspect/interface of the final repository. These are defined by
476 476 ``REPO_INTERFACES``.
477 477
478 478 Each factory function is called to produce a type implementing a specific
479 479 interface. The cumulative list of returned types will be combined into a
480 480 new type and that type will be instantiated to represent the local
481 481 repository.
482 482
483 483 The factory functions each receive various state that may be consulted
484 484 as part of deriving a type.
485 485
486 486 Extensions should wrap these factory functions to customize repository type
487 487 creation. Note that an extension's wrapped function may be called even if
488 488 that extension is not loaded for the repo being constructed. Extensions
489 489 should check if their ``__name__`` appears in the
490 490 ``extensionmodulenames`` set passed to the factory function and no-op if
491 491 not.
492 492 """
493 493 ui = baseui.copy()
494 494 # Prevent copying repo configuration.
495 495 ui.copy = baseui.copy
496 496
497 497 # Working directory VFS rooted at repository root.
498 498 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
499 499
500 500 # Main VFS for .hg/ directory.
501 501 hgpath = wdirvfs.join(b'.hg')
502 502 hgvfs = vfsmod.vfs(hgpath, cacheaudited=True)
503 503
504 504 # The .hg/ path should exist and should be a directory. All other
505 505 # cases are errors.
506 506 if not hgvfs.isdir():
507 507 try:
508 508 hgvfs.stat()
509 509 except OSError as e:
510 510 if e.errno != errno.ENOENT:
511 511 raise
512 512 except ValueError as e:
513 513 # Can be raised on Python 3.8 when path is invalid.
514 514 raise error.Abort(
515 515 _(b'invalid path %s: %s') % (path, pycompat.bytestr(e))
516 516 )
517 517
518 518 raise error.RepoError(_(b'repository %s not found') % path)
519 519
520 520 # .hg/requires file contains a newline-delimited list of
521 521 # features/capabilities the opener (us) must have in order to use
522 522 # the repository. This file was introduced in Mercurial 0.9.2,
523 523 # which means very old repositories may not have one. We assume
524 524 # a missing file translates to no requirements.
525 525 try:
526 526 requirements = set(hgvfs.read(b'requires').splitlines())
527 527 except IOError as e:
528 528 if e.errno != errno.ENOENT:
529 529 raise
530 530 requirements = set()
531 531
532 532 # The .hg/hgrc file may load extensions or contain config options
533 533 # that influence repository construction. Attempt to load it and
534 534 # process any new extensions that it may have pulled in.
535 535 if loadhgrc(ui, wdirvfs, hgvfs, requirements):
536 536 afterhgrcload(ui, wdirvfs, hgvfs, requirements)
537 537 extensions.loadall(ui)
538 538 extensions.populateui(ui)
539 539
540 540 # Set of module names of extensions loaded for this repository.
541 541 extensionmodulenames = {m.__name__ for n, m in extensions.extensions(ui)}
542 542
543 543 supportedrequirements = gathersupportedrequirements(ui)
544 544
545 545 # We first validate the requirements are known.
546 546 ensurerequirementsrecognized(requirements, supportedrequirements)
547 547
548 548 # Then we validate that the known set is reasonable to use together.
549 549 ensurerequirementscompatible(ui, requirements)
550 550
551 551 # TODO there are unhandled edge cases related to opening repositories with
552 552 # shared storage. If storage is shared, we should also test for requirements
553 553 # compatibility in the pointed-to repo. This entails loading the .hg/hgrc in
554 554 # that repo, as that repo may load extensions needed to open it. This is a
555 555 # bit complicated because we don't want the other hgrc to overwrite settings
556 556 # in this hgrc.
557 557 #
558 558 # This bug is somewhat mitigated by the fact that we copy the .hg/requires
559 559 # file when sharing repos. But if a requirement is added after the share is
560 560 # performed, thereby introducing a new requirement for the opener, we may
561 561 # will not see that and could encounter a run-time error interacting with
562 562 # that shared store since it has an unknown-to-us requirement.
563 563
564 564 # At this point, we know we should be capable of opening the repository.
565 565 # Now get on with doing that.
566 566
567 567 features = set()
568 568
569 569 # The "store" part of the repository holds versioned data. How it is
570 570 # accessed is determined by various requirements. The ``shared`` or
571 571 # ``relshared`` requirements indicate the store lives in the path contained
572 572 # in the ``.hg/sharedpath`` file. This is an absolute path for
573 573 # ``shared`` and relative to ``.hg/`` for ``relshared``.
574 574 if b'shared' in requirements or b'relshared' in requirements:
575 575 sharedpath = hgvfs.read(b'sharedpath').rstrip(b'\n')
576 576 if b'relshared' in requirements:
577 577 sharedpath = hgvfs.join(sharedpath)
578 578
579 579 sharedvfs = vfsmod.vfs(sharedpath, realpath=True)
580 580
581 581 if not sharedvfs.exists():
582 582 raise error.RepoError(
583 583 _(b'.hg/sharedpath points to nonexistent directory %s')
584 584 % sharedvfs.base
585 585 )
586 586
587 587 features.add(repository.REPO_FEATURE_SHARED_STORAGE)
588 588
589 589 storebasepath = sharedvfs.base
590 590 cachepath = sharedvfs.join(b'cache')
591 591 else:
592 592 storebasepath = hgvfs.base
593 593 cachepath = hgvfs.join(b'cache')
594 594 wcachepath = hgvfs.join(b'wcache')
595 595
596 596 # The store has changed over time and the exact layout is dictated by
597 597 # requirements. The store interface abstracts differences across all
598 598 # of them.
599 599 store = makestore(
600 600 requirements,
601 601 storebasepath,
602 602 lambda base: vfsmod.vfs(base, cacheaudited=True),
603 603 )
604 604 hgvfs.createmode = store.createmode
605 605
606 606 storevfs = store.vfs
607 607 storevfs.options = resolvestorevfsoptions(ui, requirements, features)
608 608
609 609 # The cache vfs is used to manage cache files.
610 610 cachevfs = vfsmod.vfs(cachepath, cacheaudited=True)
611 611 cachevfs.createmode = store.createmode
612 612 # The cache vfs is used to manage cache files related to the working copy
613 613 wcachevfs = vfsmod.vfs(wcachepath, cacheaudited=True)
614 614 wcachevfs.createmode = store.createmode
615 615
616 616 # Now resolve the type for the repository object. We do this by repeatedly
617 617 # calling a factory function to produces types for specific aspects of the
618 618 # repo's operation. The aggregate returned types are used as base classes
619 619 # for a dynamically-derived type, which will represent our new repository.
620 620
621 621 bases = []
622 622 extrastate = {}
623 623
624 624 for iface, fn in REPO_INTERFACES:
625 625 # We pass all potentially useful state to give extensions tons of
626 626 # flexibility.
627 627 typ = fn()(
628 628 ui=ui,
629 629 intents=intents,
630 630 requirements=requirements,
631 631 features=features,
632 632 wdirvfs=wdirvfs,
633 633 hgvfs=hgvfs,
634 634 store=store,
635 635 storevfs=storevfs,
636 636 storeoptions=storevfs.options,
637 637 cachevfs=cachevfs,
638 638 wcachevfs=wcachevfs,
639 639 extensionmodulenames=extensionmodulenames,
640 640 extrastate=extrastate,
641 641 baseclasses=bases,
642 642 )
643 643
644 644 if not isinstance(typ, type):
645 645 raise error.ProgrammingError(
646 646 b'unable to construct type for %s' % iface
647 647 )
648 648
649 649 bases.append(typ)
650 650
651 651 # type() allows you to use characters in type names that wouldn't be
652 652 # recognized as Python symbols in source code. We abuse that to add
653 653 # rich information about our constructed repo.
654 654 name = pycompat.sysstr(
655 655 b'derivedrepo:%s<%s>' % (wdirvfs.base, b','.join(sorted(requirements)))
656 656 )
657 657
658 658 cls = type(name, tuple(bases), {})
659 659
660 660 return cls(
661 661 baseui=baseui,
662 662 ui=ui,
663 663 origroot=path,
664 664 wdirvfs=wdirvfs,
665 665 hgvfs=hgvfs,
666 666 requirements=requirements,
667 667 supportedrequirements=supportedrequirements,
668 668 sharedpath=storebasepath,
669 669 store=store,
670 670 cachevfs=cachevfs,
671 671 wcachevfs=wcachevfs,
672 672 features=features,
673 673 intents=intents,
674 674 )
675 675
676 676
677 677 def loadhgrc(ui, wdirvfs, hgvfs, requirements):
678 678 """Load hgrc files/content into a ui instance.
679 679
680 680 This is called during repository opening to load any additional
681 681 config files or settings relevant to the current repository.
682 682
683 683 Returns a bool indicating whether any additional configs were loaded.
684 684
685 685 Extensions should monkeypatch this function to modify how per-repo
686 686 configs are loaded. For example, an extension may wish to pull in
687 687 configs from alternate files or sources.
688 688 """
689 689 if not rcutil.use_repo_hgrc():
690 690 return False
691 691 try:
692 692 ui.readconfig(hgvfs.join(b'hgrc'), root=wdirvfs.base)
693 693 return True
694 694 except IOError:
695 695 return False
696 696
697 697
698 698 def afterhgrcload(ui, wdirvfs, hgvfs, requirements):
699 699 """Perform additional actions after .hg/hgrc is loaded.
700 700
701 701 This function is called during repository loading immediately after
702 702 the .hg/hgrc file is loaded and before per-repo extensions are loaded.
703 703
704 704 The function can be used to validate configs, automatically add
705 705 options (including extensions) based on requirements, etc.
706 706 """
707 707
708 708 # Map of requirements to list of extensions to load automatically when
709 709 # requirement is present.
710 710 autoextensions = {
711 711 b'git': [b'git'],
712 712 b'largefiles': [b'largefiles'],
713 713 b'lfs': [b'lfs'],
714 714 }
715 715
716 716 for requirement, names in sorted(autoextensions.items()):
717 717 if requirement not in requirements:
718 718 continue
719 719
720 720 for name in names:
721 721 if not ui.hasconfig(b'extensions', name):
722 722 ui.setconfig(b'extensions', name, b'', source=b'autoload')
723 723
724 724
725 725 def gathersupportedrequirements(ui):
726 726 """Determine the complete set of recognized requirements."""
727 727 # Start with all requirements supported by this file.
728 728 supported = set(localrepository._basesupported)
729 729
730 730 # Execute ``featuresetupfuncs`` entries if they belong to an extension
731 731 # relevant to this ui instance.
732 732 modules = {m.__name__ for n, m in extensions.extensions(ui)}
733 733
734 734 for fn in featuresetupfuncs:
735 735 if fn.__module__ in modules:
736 736 fn(ui, supported)
737 737
738 738 # Add derived requirements from registered compression engines.
739 739 for name in util.compengines:
740 740 engine = util.compengines[name]
741 741 if engine.available() and engine.revlogheader():
742 742 supported.add(b'exp-compression-%s' % name)
743 743 if engine.name() == b'zstd':
744 744 supported.add(b'revlog-compression-zstd')
745 745
746 746 return supported
747 747
748 748
749 749 def ensurerequirementsrecognized(requirements, supported):
750 750 """Validate that a set of local requirements is recognized.
751 751
752 752 Receives a set of requirements. Raises an ``error.RepoError`` if there
753 753 exists any requirement in that set that currently loaded code doesn't
754 754 recognize.
755 755
756 756 Returns a set of supported requirements.
757 757 """
758 758 missing = set()
759 759
760 760 for requirement in requirements:
761 761 if requirement in supported:
762 762 continue
763 763
764 764 if not requirement or not requirement[0:1].isalnum():
765 765 raise error.RequirementError(_(b'.hg/requires file is corrupt'))
766 766
767 767 missing.add(requirement)
768 768
769 769 if missing:
770 770 raise error.RequirementError(
771 771 _(b'repository requires features unknown to this Mercurial: %s')
772 772 % b' '.join(sorted(missing)),
773 773 hint=_(
774 774 b'see https://mercurial-scm.org/wiki/MissingRequirement '
775 775 b'for more information'
776 776 ),
777 777 )
778 778
779 779
780 780 def ensurerequirementscompatible(ui, requirements):
781 781 """Validates that a set of recognized requirements is mutually compatible.
782 782
783 783 Some requirements may not be compatible with others or require
784 784 config options that aren't enabled. This function is called during
785 785 repository opening to ensure that the set of requirements needed
786 786 to open a repository is sane and compatible with config options.
787 787
788 788 Extensions can monkeypatch this function to perform additional
789 789 checking.
790 790
791 791 ``error.RepoError`` should be raised on failure.
792 792 """
793 793 if b'exp-sparse' in requirements and not sparse.enabled:
794 794 raise error.RepoError(
795 795 _(
796 796 b'repository is using sparse feature but '
797 797 b'sparse is not enabled; enable the '
798 798 b'"sparse" extensions to access'
799 799 )
800 800 )
801 801
802 802
803 803 def makestore(requirements, path, vfstype):
804 804 """Construct a storage object for a repository."""
805 805 if b'store' in requirements:
806 806 if b'fncache' in requirements:
807 807 return storemod.fncachestore(
808 808 path, vfstype, b'dotencode' in requirements
809 809 )
810 810
811 811 return storemod.encodedstore(path, vfstype)
812 812
813 813 return storemod.basicstore(path, vfstype)
814 814
815 815
816 816 def resolvestorevfsoptions(ui, requirements, features):
817 817 """Resolve the options to pass to the store vfs opener.
818 818
819 819 The returned dict is used to influence behavior of the storage layer.
820 820 """
821 821 options = {}
822 822
823 823 if b'treemanifest' in requirements:
824 824 options[b'treemanifest'] = True
825 825
826 826 # experimental config: format.manifestcachesize
827 827 manifestcachesize = ui.configint(b'format', b'manifestcachesize')
828 828 if manifestcachesize is not None:
829 829 options[b'manifestcachesize'] = manifestcachesize
830 830
831 831 # In the absence of another requirement superseding a revlog-related
832 832 # requirement, we have to assume the repo is using revlog version 0.
833 833 # This revlog format is super old and we don't bother trying to parse
834 834 # opener options for it because those options wouldn't do anything
835 835 # meaningful on such old repos.
836 836 if b'revlogv1' in requirements or REVLOGV2_REQUIREMENT in requirements:
837 837 options.update(resolverevlogstorevfsoptions(ui, requirements, features))
838 838 else: # explicitly mark repo as using revlogv0
839 839 options[b'revlogv0'] = True
840 840
841 841 if COPIESSDC_REQUIREMENT in requirements:
842 842 options[b'copies-storage'] = b'changeset-sidedata'
843 843 else:
844 844 writecopiesto = ui.config(b'experimental', b'copies.write-to')
845 845 copiesextramode = (b'changeset-only', b'compatibility')
846 846 if writecopiesto in copiesextramode:
847 847 options[b'copies-storage'] = b'extra'
848 848
849 849 return options
850 850
851 851
852 852 def resolverevlogstorevfsoptions(ui, requirements, features):
853 853 """Resolve opener options specific to revlogs."""
854 854
855 855 options = {}
856 856 options[b'flagprocessors'] = {}
857 857
858 858 if b'revlogv1' in requirements:
859 859 options[b'revlogv1'] = True
860 860 if REVLOGV2_REQUIREMENT in requirements:
861 861 options[b'revlogv2'] = True
862 862
863 863 if b'generaldelta' in requirements:
864 864 options[b'generaldelta'] = True
865 865
866 866 # experimental config: format.chunkcachesize
867 867 chunkcachesize = ui.configint(b'format', b'chunkcachesize')
868 868 if chunkcachesize is not None:
869 869 options[b'chunkcachesize'] = chunkcachesize
870 870
871 871 deltabothparents = ui.configbool(
872 872 b'storage', b'revlog.optimize-delta-parent-choice'
873 873 )
874 874 options[b'deltabothparents'] = deltabothparents
875 875
876 876 lazydelta = ui.configbool(b'storage', b'revlog.reuse-external-delta')
877 877 lazydeltabase = False
878 878 if lazydelta:
879 879 lazydeltabase = ui.configbool(
880 880 b'storage', b'revlog.reuse-external-delta-parent'
881 881 )
882 882 if lazydeltabase is None:
883 883 lazydeltabase = not scmutil.gddeltaconfig(ui)
884 884 options[b'lazydelta'] = lazydelta
885 885 options[b'lazydeltabase'] = lazydeltabase
886 886
887 887 chainspan = ui.configbytes(b'experimental', b'maxdeltachainspan')
888 888 if 0 <= chainspan:
889 889 options[b'maxdeltachainspan'] = chainspan
890 890
891 891 mmapindexthreshold = ui.configbytes(b'experimental', b'mmapindexthreshold')
892 892 if mmapindexthreshold is not None:
893 893 options[b'mmapindexthreshold'] = mmapindexthreshold
894 894
895 895 withsparseread = ui.configbool(b'experimental', b'sparse-read')
896 896 srdensitythres = float(
897 897 ui.config(b'experimental', b'sparse-read.density-threshold')
898 898 )
899 899 srmingapsize = ui.configbytes(b'experimental', b'sparse-read.min-gap-size')
900 900 options[b'with-sparse-read'] = withsparseread
901 901 options[b'sparse-read-density-threshold'] = srdensitythres
902 902 options[b'sparse-read-min-gap-size'] = srmingapsize
903 903
904 904 sparserevlog = SPARSEREVLOG_REQUIREMENT in requirements
905 905 options[b'sparse-revlog'] = sparserevlog
906 906 if sparserevlog:
907 907 options[b'generaldelta'] = True
908 908
909 909 sidedata = SIDEDATA_REQUIREMENT in requirements
910 910 options[b'side-data'] = sidedata
911 911
912 912 maxchainlen = None
913 913 if sparserevlog:
914 914 maxchainlen = revlogconst.SPARSE_REVLOG_MAX_CHAIN_LENGTH
915 915 # experimental config: format.maxchainlen
916 916 maxchainlen = ui.configint(b'format', b'maxchainlen', maxchainlen)
917 917 if maxchainlen is not None:
918 918 options[b'maxchainlen'] = maxchainlen
919 919
920 920 for r in requirements:
921 921 # we allow multiple compression engine requirement to co-exist because
922 922 # strickly speaking, revlog seems to support mixed compression style.
923 923 #
924 924 # The compression used for new entries will be "the last one"
925 925 prefix = r.startswith
926 926 if prefix(b'revlog-compression-') or prefix(b'exp-compression-'):
927 927 options[b'compengine'] = r.split(b'-', 2)[2]
928 928
929 929 options[b'zlib.level'] = ui.configint(b'storage', b'revlog.zlib.level')
930 930 if options[b'zlib.level'] is not None:
931 931 if not (0 <= options[b'zlib.level'] <= 9):
932 932 msg = _(b'invalid value for `storage.revlog.zlib.level` config: %d')
933 933 raise error.Abort(msg % options[b'zlib.level'])
934 934 options[b'zstd.level'] = ui.configint(b'storage', b'revlog.zstd.level')
935 935 if options[b'zstd.level'] is not None:
936 936 if not (0 <= options[b'zstd.level'] <= 22):
937 937 msg = _(b'invalid value for `storage.revlog.zstd.level` config: %d')
938 938 raise error.Abort(msg % options[b'zstd.level'])
939 939
940 940 if repository.NARROW_REQUIREMENT in requirements:
941 941 options[b'enableellipsis'] = True
942 942
943 943 if ui.configbool(b'experimental', b'rust.index'):
944 944 options[b'rust.index'] = True
945 945 if NODEMAP_REQUIREMENT in requirements:
946 946 options[b'persistent-nodemap'] = True
947 947 if ui.configbool(b'storage', b'revlog.nodemap.mmap'):
948 948 options[b'persistent-nodemap.mmap'] = True
949 949 epnm = ui.config(b'storage', b'revlog.nodemap.mode')
950 950 options[b'persistent-nodemap.mode'] = epnm
951 951 if ui.configbool(b'devel', b'persistent-nodemap'):
952 952 options[b'devel-force-nodemap'] = True
953 953
954 954 return options
955 955
956 956
957 957 def makemain(**kwargs):
958 958 """Produce a type conforming to ``ilocalrepositorymain``."""
959 959 return localrepository
960 960
961 961
962 962 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
963 963 class revlogfilestorage(object):
964 964 """File storage when using revlogs."""
965 965
966 966 def file(self, path):
967 967 if path[0] == b'/':
968 968 path = path[1:]
969 969
970 970 return filelog.filelog(self.svfs, path)
971 971
972 972
973 973 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
974 974 class revlognarrowfilestorage(object):
975 975 """File storage when using revlogs and narrow files."""
976 976
977 977 def file(self, path):
978 978 if path[0] == b'/':
979 979 path = path[1:]
980 980
981 981 return filelog.narrowfilelog(self.svfs, path, self._storenarrowmatch)
982 982
983 983
984 984 def makefilestorage(requirements, features, **kwargs):
985 985 """Produce a type conforming to ``ilocalrepositoryfilestorage``."""
986 986 features.add(repository.REPO_FEATURE_REVLOG_FILE_STORAGE)
987 987 features.add(repository.REPO_FEATURE_STREAM_CLONE)
988 988
989 989 if repository.NARROW_REQUIREMENT in requirements:
990 990 return revlognarrowfilestorage
991 991 else:
992 992 return revlogfilestorage
993 993
994 994
995 995 # List of repository interfaces and factory functions for them. Each
996 996 # will be called in order during ``makelocalrepository()`` to iteratively
997 997 # derive the final type for a local repository instance. We capture the
998 998 # function as a lambda so we don't hold a reference and the module-level
999 999 # functions can be wrapped.
1000 1000 REPO_INTERFACES = [
1001 1001 (repository.ilocalrepositorymain, lambda: makemain),
1002 1002 (repository.ilocalrepositoryfilestorage, lambda: makefilestorage),
1003 1003 ]
1004 1004
1005 1005
1006 1006 @interfaceutil.implementer(repository.ilocalrepositorymain)
1007 1007 class localrepository(object):
1008 1008 """Main class for representing local repositories.
1009 1009
1010 1010 All local repositories are instances of this class.
1011 1011
1012 1012 Constructed on its own, instances of this class are not usable as
1013 1013 repository objects. To obtain a usable repository object, call
1014 1014 ``hg.repository()``, ``localrepo.instance()``, or
1015 1015 ``localrepo.makelocalrepository()``. The latter is the lowest-level.
1016 1016 ``instance()`` adds support for creating new repositories.
1017 1017 ``hg.repository()`` adds more extension integration, including calling
1018 1018 ``reposetup()``. Generally speaking, ``hg.repository()`` should be
1019 1019 used.
1020 1020 """
1021 1021
1022 1022 # obsolete experimental requirements:
1023 1023 # - manifestv2: An experimental new manifest format that allowed
1024 1024 # for stem compression of long paths. Experiment ended up not
1025 1025 # being successful (repository sizes went up due to worse delta
1026 1026 # chains), and the code was deleted in 4.6.
1027 1027 supportedformats = {
1028 1028 b'revlogv1',
1029 1029 b'generaldelta',
1030 1030 b'treemanifest',
1031 1031 COPIESSDC_REQUIREMENT,
1032 1032 REVLOGV2_REQUIREMENT,
1033 1033 SIDEDATA_REQUIREMENT,
1034 1034 SPARSEREVLOG_REQUIREMENT,
1035 1035 NODEMAP_REQUIREMENT,
1036 1036 bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT,
1037 1037 }
1038 1038 _basesupported = supportedformats | {
1039 1039 b'store',
1040 1040 b'fncache',
1041 1041 b'shared',
1042 1042 b'relshared',
1043 1043 b'dotencode',
1044 1044 b'exp-sparse',
1045 1045 b'internal-phase',
1046 1046 }
1047 1047
1048 1048 # list of prefix for file which can be written without 'wlock'
1049 1049 # Extensions should extend this list when needed
1050 1050 _wlockfreeprefix = {
1051 1051 # We migh consider requiring 'wlock' for the next
1052 1052 # two, but pretty much all the existing code assume
1053 1053 # wlock is not needed so we keep them excluded for
1054 1054 # now.
1055 1055 b'hgrc',
1056 1056 b'requires',
1057 1057 # XXX cache is a complicatged business someone
1058 1058 # should investigate this in depth at some point
1059 1059 b'cache/',
1060 1060 # XXX shouldn't be dirstate covered by the wlock?
1061 1061 b'dirstate',
1062 1062 # XXX bisect was still a bit too messy at the time
1063 1063 # this changeset was introduced. Someone should fix
1064 1064 # the remainig bit and drop this line
1065 1065 b'bisect.state',
1066 1066 }
1067 1067
1068 1068 def __init__(
1069 1069 self,
1070 1070 baseui,
1071 1071 ui,
1072 1072 origroot,
1073 1073 wdirvfs,
1074 1074 hgvfs,
1075 1075 requirements,
1076 1076 supportedrequirements,
1077 1077 sharedpath,
1078 1078 store,
1079 1079 cachevfs,
1080 1080 wcachevfs,
1081 1081 features,
1082 1082 intents=None,
1083 1083 ):
1084 1084 """Create a new local repository instance.
1085 1085
1086 1086 Most callers should use ``hg.repository()``, ``localrepo.instance()``,
1087 1087 or ``localrepo.makelocalrepository()`` for obtaining a new repository
1088 1088 object.
1089 1089
1090 1090 Arguments:
1091 1091
1092 1092 baseui
1093 1093 ``ui.ui`` instance that ``ui`` argument was based off of.
1094 1094
1095 1095 ui
1096 1096 ``ui.ui`` instance for use by the repository.
1097 1097
1098 1098 origroot
1099 1099 ``bytes`` path to working directory root of this repository.
1100 1100
1101 1101 wdirvfs
1102 1102 ``vfs.vfs`` rooted at the working directory.
1103 1103
1104 1104 hgvfs
1105 1105 ``vfs.vfs`` rooted at .hg/
1106 1106
1107 1107 requirements
1108 1108 ``set`` of bytestrings representing repository opening requirements.
1109 1109
1110 1110 supportedrequirements
1111 1111 ``set`` of bytestrings representing repository requirements that we
1112 1112 know how to open. May be a supetset of ``requirements``.
1113 1113
1114 1114 sharedpath
1115 1115 ``bytes`` Defining path to storage base directory. Points to a
1116 1116 ``.hg/`` directory somewhere.
1117 1117
1118 1118 store
1119 1119 ``store.basicstore`` (or derived) instance providing access to
1120 1120 versioned storage.
1121 1121
1122 1122 cachevfs
1123 1123 ``vfs.vfs`` used for cache files.
1124 1124
1125 1125 wcachevfs
1126 1126 ``vfs.vfs`` used for cache files related to the working copy.
1127 1127
1128 1128 features
1129 1129 ``set`` of bytestrings defining features/capabilities of this
1130 1130 instance.
1131 1131
1132 1132 intents
1133 1133 ``set`` of system strings indicating what this repo will be used
1134 1134 for.
1135 1135 """
1136 1136 self.baseui = baseui
1137 1137 self.ui = ui
1138 1138 self.origroot = origroot
1139 1139 # vfs rooted at working directory.
1140 1140 self.wvfs = wdirvfs
1141 1141 self.root = wdirvfs.base
1142 1142 # vfs rooted at .hg/. Used to access most non-store paths.
1143 1143 self.vfs = hgvfs
1144 1144 self.path = hgvfs.base
1145 1145 self.requirements = requirements
1146 1146 self.supported = supportedrequirements
1147 1147 self.sharedpath = sharedpath
1148 1148 self.store = store
1149 1149 self.cachevfs = cachevfs
1150 1150 self.wcachevfs = wcachevfs
1151 1151 self.features = features
1152 1152
1153 1153 self.filtername = None
1154 1154
1155 1155 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1156 1156 b'devel', b'check-locks'
1157 1157 ):
1158 1158 self.vfs.audit = self._getvfsward(self.vfs.audit)
1159 1159 # A list of callback to shape the phase if no data were found.
1160 1160 # Callback are in the form: func(repo, roots) --> processed root.
1161 1161 # This list it to be filled by extension during repo setup
1162 1162 self._phasedefaults = []
1163 1163
1164 1164 color.setup(self.ui)
1165 1165
1166 1166 self.spath = self.store.path
1167 1167 self.svfs = self.store.vfs
1168 1168 self.sjoin = self.store.join
1169 1169 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1170 1170 b'devel', b'check-locks'
1171 1171 ):
1172 1172 if util.safehasattr(self.svfs, b'vfs'): # this is filtervfs
1173 1173 self.svfs.vfs.audit = self._getsvfsward(self.svfs.vfs.audit)
1174 1174 else: # standard vfs
1175 1175 self.svfs.audit = self._getsvfsward(self.svfs.audit)
1176 1176
1177 1177 self._dirstatevalidatewarned = False
1178 1178
1179 1179 self._branchcaches = branchmap.BranchMapCache()
1180 1180 self._revbranchcache = None
1181 1181 self._filterpats = {}
1182 1182 self._datafilters = {}
1183 1183 self._transref = self._lockref = self._wlockref = None
1184 1184
1185 1185 # A cache for various files under .hg/ that tracks file changes,
1186 1186 # (used by the filecache decorator)
1187 1187 #
1188 1188 # Maps a property name to its util.filecacheentry
1189 1189 self._filecache = {}
1190 1190
1191 1191 # hold sets of revision to be filtered
1192 1192 # should be cleared when something might have changed the filter value:
1193 1193 # - new changesets,
1194 1194 # - phase change,
1195 1195 # - new obsolescence marker,
1196 1196 # - working directory parent change,
1197 1197 # - bookmark changes
1198 1198 self.filteredrevcache = {}
1199 1199
1200 1200 # post-dirstate-status hooks
1201 1201 self._postdsstatus = []
1202 1202
1203 1203 # generic mapping between names and nodes
1204 1204 self.names = namespaces.namespaces()
1205 1205
1206 1206 # Key to signature value.
1207 1207 self._sparsesignaturecache = {}
1208 1208 # Signature to cached matcher instance.
1209 1209 self._sparsematchercache = {}
1210 1210
1211 1211 self._extrafilterid = repoview.extrafilter(ui)
1212 1212
1213 1213 self.filecopiesmode = None
1214 1214 if COPIESSDC_REQUIREMENT in self.requirements:
1215 1215 self.filecopiesmode = b'changeset-sidedata'
1216 1216
1217 1217 def _getvfsward(self, origfunc):
1218 1218 """build a ward for self.vfs"""
1219 1219 rref = weakref.ref(self)
1220 1220
1221 1221 def checkvfs(path, mode=None):
1222 1222 ret = origfunc(path, mode=mode)
1223 1223 repo = rref()
1224 1224 if (
1225 1225 repo is None
1226 1226 or not util.safehasattr(repo, b'_wlockref')
1227 1227 or not util.safehasattr(repo, b'_lockref')
1228 1228 ):
1229 1229 return
1230 1230 if mode in (None, b'r', b'rb'):
1231 1231 return
1232 1232 if path.startswith(repo.path):
1233 1233 # truncate name relative to the repository (.hg)
1234 1234 path = path[len(repo.path) + 1 :]
1235 1235 if path.startswith(b'cache/'):
1236 1236 msg = b'accessing cache with vfs instead of cachevfs: "%s"'
1237 1237 repo.ui.develwarn(msg % path, stacklevel=3, config=b"cache-vfs")
1238 1238 # path prefixes covered by 'lock'
1239 1239 vfs_path_prefixes = (b'journal.', b'undo.', b'strip-backup/')
1240 1240 if any(path.startswith(prefix) for prefix in vfs_path_prefixes):
1241 1241 if repo._currentlock(repo._lockref) is None:
1242 1242 repo.ui.develwarn(
1243 1243 b'write with no lock: "%s"' % path,
1244 1244 stacklevel=3,
1245 1245 config=b'check-locks',
1246 1246 )
1247 1247 elif repo._currentlock(repo._wlockref) is None:
1248 1248 # rest of vfs files are covered by 'wlock'
1249 1249 #
1250 1250 # exclude special files
1251 1251 for prefix in self._wlockfreeprefix:
1252 1252 if path.startswith(prefix):
1253 1253 return
1254 1254 repo.ui.develwarn(
1255 1255 b'write with no wlock: "%s"' % path,
1256 1256 stacklevel=3,
1257 1257 config=b'check-locks',
1258 1258 )
1259 1259 return ret
1260 1260
1261 1261 return checkvfs
1262 1262
1263 1263 def _getsvfsward(self, origfunc):
1264 1264 """build a ward for self.svfs"""
1265 1265 rref = weakref.ref(self)
1266 1266
1267 1267 def checksvfs(path, mode=None):
1268 1268 ret = origfunc(path, mode=mode)
1269 1269 repo = rref()
1270 1270 if repo is None or not util.safehasattr(repo, b'_lockref'):
1271 1271 return
1272 1272 if mode in (None, b'r', b'rb'):
1273 1273 return
1274 1274 if path.startswith(repo.sharedpath):
1275 1275 # truncate name relative to the repository (.hg)
1276 1276 path = path[len(repo.sharedpath) + 1 :]
1277 1277 if repo._currentlock(repo._lockref) is None:
1278 1278 repo.ui.develwarn(
1279 1279 b'write with no lock: "%s"' % path, stacklevel=4
1280 1280 )
1281 1281 return ret
1282 1282
1283 1283 return checksvfs
1284 1284
1285 1285 def close(self):
1286 1286 self._writecaches()
1287 1287
1288 1288 def _writecaches(self):
1289 1289 if self._revbranchcache:
1290 1290 self._revbranchcache.write()
1291 1291
1292 1292 def _restrictcapabilities(self, caps):
1293 1293 if self.ui.configbool(b'experimental', b'bundle2-advertise'):
1294 1294 caps = set(caps)
1295 1295 capsblob = bundle2.encodecaps(
1296 1296 bundle2.getrepocaps(self, role=b'client')
1297 1297 )
1298 1298 caps.add(b'bundle2=' + urlreq.quote(capsblob))
1299 1299 return caps
1300 1300
1301 1301 # Don't cache auditor/nofsauditor, or you'll end up with reference cycle:
1302 1302 # self -> auditor -> self._checknested -> self
1303 1303
1304 1304 @property
1305 1305 def auditor(self):
1306 1306 # This is only used by context.workingctx.match in order to
1307 1307 # detect files in subrepos.
1308 1308 return pathutil.pathauditor(self.root, callback=self._checknested)
1309 1309
1310 1310 @property
1311 1311 def nofsauditor(self):
1312 1312 # This is only used by context.basectx.match in order to detect
1313 1313 # files in subrepos.
1314 1314 return pathutil.pathauditor(
1315 1315 self.root, callback=self._checknested, realfs=False, cached=True
1316 1316 )
1317 1317
1318 1318 def _checknested(self, path):
1319 1319 """Determine if path is a legal nested repository."""
1320 1320 if not path.startswith(self.root):
1321 1321 return False
1322 1322 subpath = path[len(self.root) + 1 :]
1323 1323 normsubpath = util.pconvert(subpath)
1324 1324
1325 1325 # XXX: Checking against the current working copy is wrong in
1326 1326 # the sense that it can reject things like
1327 1327 #
1328 1328 # $ hg cat -r 10 sub/x.txt
1329 1329 #
1330 1330 # if sub/ is no longer a subrepository in the working copy
1331 1331 # parent revision.
1332 1332 #
1333 1333 # However, it can of course also allow things that would have
1334 1334 # been rejected before, such as the above cat command if sub/
1335 1335 # is a subrepository now, but was a normal directory before.
1336 1336 # The old path auditor would have rejected by mistake since it
1337 1337 # panics when it sees sub/.hg/.
1338 1338 #
1339 1339 # All in all, checking against the working copy seems sensible
1340 1340 # since we want to prevent access to nested repositories on
1341 1341 # the filesystem *now*.
1342 1342 ctx = self[None]
1343 1343 parts = util.splitpath(subpath)
1344 1344 while parts:
1345 1345 prefix = b'/'.join(parts)
1346 1346 if prefix in ctx.substate:
1347 1347 if prefix == normsubpath:
1348 1348 return True
1349 1349 else:
1350 1350 sub = ctx.sub(prefix)
1351 1351 return sub.checknested(subpath[len(prefix) + 1 :])
1352 1352 else:
1353 1353 parts.pop()
1354 1354 return False
1355 1355
1356 1356 def peer(self):
1357 1357 return localpeer(self) # not cached to avoid reference cycle
1358 1358
1359 1359 def unfiltered(self):
1360 1360 """Return unfiltered version of the repository
1361 1361
1362 1362 Intended to be overwritten by filtered repo."""
1363 1363 return self
1364 1364
1365 1365 def filtered(self, name, visibilityexceptions=None):
1366 1366 """Return a filtered version of a repository
1367 1367
1368 1368 The `name` parameter is the identifier of the requested view. This
1369 1369 will return a repoview object set "exactly" to the specified view.
1370 1370
1371 1371 This function does not apply recursive filtering to a repository. For
1372 1372 example calling `repo.filtered("served")` will return a repoview using
1373 1373 the "served" view, regardless of the initial view used by `repo`.
1374 1374
1375 1375 In other word, there is always only one level of `repoview` "filtering".
1376 1376 """
1377 1377 if self._extrafilterid is not None and b'%' not in name:
1378 1378 name = name + b'%' + self._extrafilterid
1379 1379
1380 1380 cls = repoview.newtype(self.unfiltered().__class__)
1381 1381 return cls(self, name, visibilityexceptions)
1382 1382
1383 1383 @mixedrepostorecache(
1384 1384 (b'bookmarks', b'plain'),
1385 1385 (b'bookmarks.current', b'plain'),
1386 1386 (b'bookmarks', b''),
1387 1387 (b'00changelog.i', b''),
1388 1388 )
1389 1389 def _bookmarks(self):
1390 1390 # Since the multiple files involved in the transaction cannot be
1391 1391 # written atomically (with current repository format), there is a race
1392 1392 # condition here.
1393 1393 #
1394 1394 # 1) changelog content A is read
1395 1395 # 2) outside transaction update changelog to content B
1396 1396 # 3) outside transaction update bookmark file referring to content B
1397 1397 # 4) bookmarks file content is read and filtered against changelog-A
1398 1398 #
1399 1399 # When this happens, bookmarks against nodes missing from A are dropped.
1400 1400 #
1401 1401 # Having this happening during read is not great, but it become worse
1402 1402 # when this happen during write because the bookmarks to the "unknown"
1403 1403 # nodes will be dropped for good. However, writes happen within locks.
1404 1404 # This locking makes it possible to have a race free consistent read.
1405 1405 # For this purpose data read from disc before locking are
1406 1406 # "invalidated" right after the locks are taken. This invalidations are
1407 1407 # "light", the `filecache` mechanism keep the data in memory and will
1408 1408 # reuse them if the underlying files did not changed. Not parsing the
1409 1409 # same data multiple times helps performances.
1410 1410 #
1411 1411 # Unfortunately in the case describe above, the files tracked by the
1412 1412 # bookmarks file cache might not have changed, but the in-memory
1413 1413 # content is still "wrong" because we used an older changelog content
1414 1414 # to process the on-disk data. So after locking, the changelog would be
1415 1415 # refreshed but `_bookmarks` would be preserved.
1416 1416 # Adding `00changelog.i` to the list of tracked file is not
1417 1417 # enough, because at the time we build the content for `_bookmarks` in
1418 1418 # (4), the changelog file has already diverged from the content used
1419 1419 # for loading `changelog` in (1)
1420 1420 #
1421 1421 # To prevent the issue, we force the changelog to be explicitly
1422 1422 # reloaded while computing `_bookmarks`. The data race can still happen
1423 1423 # without the lock (with a narrower window), but it would no longer go
1424 1424 # undetected during the lock time refresh.
1425 1425 #
1426 1426 # The new schedule is as follow
1427 1427 #
1428 1428 # 1) filecache logic detect that `_bookmarks` needs to be computed
1429 1429 # 2) cachestat for `bookmarks` and `changelog` are captured (for book)
1430 1430 # 3) We force `changelog` filecache to be tested
1431 1431 # 4) cachestat for `changelog` are captured (for changelog)
1432 1432 # 5) `_bookmarks` is computed and cached
1433 1433 #
1434 1434 # The step in (3) ensure we have a changelog at least as recent as the
1435 1435 # cache stat computed in (1). As a result at locking time:
1436 1436 # * if the changelog did not changed since (1) -> we can reuse the data
1437 1437 # * otherwise -> the bookmarks get refreshed.
1438 1438 self._refreshchangelog()
1439 1439 return bookmarks.bmstore(self)
1440 1440
1441 1441 def _refreshchangelog(self):
1442 1442 """make sure the in memory changelog match the on-disk one"""
1443 1443 if 'changelog' in vars(self) and self.currenttransaction() is None:
1444 1444 del self.changelog
1445 1445
1446 1446 @property
1447 1447 def _activebookmark(self):
1448 1448 return self._bookmarks.active
1449 1449
1450 1450 # _phasesets depend on changelog. what we need is to call
1451 1451 # _phasecache.invalidate() if '00changelog.i' was changed, but it
1452 1452 # can't be easily expressed in filecache mechanism.
1453 1453 @storecache(b'phaseroots', b'00changelog.i')
1454 1454 def _phasecache(self):
1455 1455 return phases.phasecache(self, self._phasedefaults)
1456 1456
1457 1457 @storecache(b'obsstore')
1458 1458 def obsstore(self):
1459 1459 return obsolete.makestore(self.ui, self)
1460 1460
1461 1461 @storecache(b'00changelog.i')
1462 1462 def changelog(self):
1463 1463 # load dirstate before changelog to avoid race see issue6303
1464 1464 self.dirstate.prefetch_parents()
1465 1465 return self.store.changelog(txnutil.mayhavepending(self.root))
1466 1466
1467 1467 @storecache(b'00manifest.i')
1468 1468 def manifestlog(self):
1469 1469 return self.store.manifestlog(self, self._storenarrowmatch)
1470 1470
1471 1471 @repofilecache(b'dirstate')
1472 1472 def dirstate(self):
1473 1473 return self._makedirstate()
1474 1474
1475 1475 def _makedirstate(self):
1476 1476 """Extension point for wrapping the dirstate per-repo."""
1477 1477 sparsematchfn = lambda: sparse.matcher(self)
1478 1478
1479 1479 return dirstate.dirstate(
1480 1480 self.vfs, self.ui, self.root, self._dirstatevalidate, sparsematchfn
1481 1481 )
1482 1482
1483 1483 def _dirstatevalidate(self, node):
1484 1484 try:
1485 1485 self.changelog.rev(node)
1486 1486 return node
1487 1487 except error.LookupError:
1488 1488 if not self._dirstatevalidatewarned:
1489 1489 self._dirstatevalidatewarned = True
1490 1490 self.ui.warn(
1491 1491 _(b"warning: ignoring unknown working parent %s!\n")
1492 1492 % short(node)
1493 1493 )
1494 1494 return nullid
1495 1495
1496 1496 @storecache(narrowspec.FILENAME)
1497 1497 def narrowpats(self):
1498 1498 """matcher patterns for this repository's narrowspec
1499 1499
1500 1500 A tuple of (includes, excludes).
1501 1501 """
1502 1502 return narrowspec.load(self)
1503 1503
1504 1504 @storecache(narrowspec.FILENAME)
1505 1505 def _storenarrowmatch(self):
1506 1506 if repository.NARROW_REQUIREMENT not in self.requirements:
1507 1507 return matchmod.always()
1508 1508 include, exclude = self.narrowpats
1509 1509 return narrowspec.match(self.root, include=include, exclude=exclude)
1510 1510
1511 1511 @storecache(narrowspec.FILENAME)
1512 1512 def _narrowmatch(self):
1513 1513 if repository.NARROW_REQUIREMENT not in self.requirements:
1514 1514 return matchmod.always()
1515 1515 narrowspec.checkworkingcopynarrowspec(self)
1516 1516 include, exclude = self.narrowpats
1517 1517 return narrowspec.match(self.root, include=include, exclude=exclude)
1518 1518
1519 1519 def narrowmatch(self, match=None, includeexact=False):
1520 1520 """matcher corresponding the the repo's narrowspec
1521 1521
1522 1522 If `match` is given, then that will be intersected with the narrow
1523 1523 matcher.
1524 1524
1525 1525 If `includeexact` is True, then any exact matches from `match` will
1526 1526 be included even if they're outside the narrowspec.
1527 1527 """
1528 1528 if match:
1529 1529 if includeexact and not self._narrowmatch.always():
1530 1530 # do not exclude explicitly-specified paths so that they can
1531 1531 # be warned later on
1532 1532 em = matchmod.exact(match.files())
1533 1533 nm = matchmod.unionmatcher([self._narrowmatch, em])
1534 1534 return matchmod.intersectmatchers(match, nm)
1535 1535 return matchmod.intersectmatchers(match, self._narrowmatch)
1536 1536 return self._narrowmatch
1537 1537
1538 1538 def setnarrowpats(self, newincludes, newexcludes):
1539 1539 narrowspec.save(self, newincludes, newexcludes)
1540 1540 self.invalidate(clearfilecache=True)
1541 1541
1542 1542 @unfilteredpropertycache
1543 1543 def _quick_access_changeid_null(self):
1544 1544 return {
1545 1545 b'null': (nullrev, nullid),
1546 1546 nullrev: (nullrev, nullid),
1547 1547 nullid: (nullrev, nullid),
1548 1548 }
1549 1549
1550 1550 @unfilteredpropertycache
1551 1551 def _quick_access_changeid_wc(self):
1552 1552 # also fast path access to the working copy parents
1553 1553 # however, only do it for filter that ensure wc is visible.
1554 1554 quick = {}
1555 1555 cl = self.unfiltered().changelog
1556 1556 for node in self.dirstate.parents():
1557 1557 if node == nullid:
1558 1558 continue
1559 1559 rev = cl.index.get_rev(node)
1560 1560 if rev is None:
1561 1561 # unknown working copy parent case:
1562 1562 #
1563 1563 # skip the fast path and let higher code deal with it
1564 1564 continue
1565 1565 pair = (rev, node)
1566 1566 quick[rev] = pair
1567 1567 quick[node] = pair
1568 1568 # also add the parents of the parents
1569 1569 for r in cl.parentrevs(rev):
1570 1570 if r == nullrev:
1571 1571 continue
1572 1572 n = cl.node(r)
1573 1573 pair = (r, n)
1574 1574 quick[r] = pair
1575 1575 quick[n] = pair
1576 1576 p1node = self.dirstate.p1()
1577 1577 if p1node != nullid:
1578 1578 quick[b'.'] = quick[p1node]
1579 1579 return quick
1580 1580
1581 1581 @unfilteredmethod
1582 1582 def _quick_access_changeid_invalidate(self):
1583 1583 if '_quick_access_changeid_wc' in vars(self):
1584 1584 del self.__dict__['_quick_access_changeid_wc']
1585 1585
1586 1586 @property
1587 1587 def _quick_access_changeid(self):
1588 1588 """an helper dictionnary for __getitem__ calls
1589 1589
1590 1590 This contains a list of symbol we can recognise right away without
1591 1591 further processing.
1592 1592 """
1593 1593 mapping = self._quick_access_changeid_null
1594 1594 if self.filtername in repoview.filter_has_wc:
1595 1595 mapping = mapping.copy()
1596 1596 mapping.update(self._quick_access_changeid_wc)
1597 1597 return mapping
1598 1598
1599 1599 def __getitem__(self, changeid):
1600 1600 # dealing with special cases
1601 1601 if changeid is None:
1602 1602 return context.workingctx(self)
1603 1603 if isinstance(changeid, context.basectx):
1604 1604 return changeid
1605 1605
1606 1606 # dealing with multiple revisions
1607 1607 if isinstance(changeid, slice):
1608 1608 # wdirrev isn't contiguous so the slice shouldn't include it
1609 1609 return [
1610 1610 self[i]
1611 1611 for i in pycompat.xrange(*changeid.indices(len(self)))
1612 1612 if i not in self.changelog.filteredrevs
1613 1613 ]
1614 1614
1615 1615 # dealing with some special values
1616 1616 quick_access = self._quick_access_changeid.get(changeid)
1617 1617 if quick_access is not None:
1618 1618 rev, node = quick_access
1619 1619 return context.changectx(self, rev, node, maybe_filtered=False)
1620 1620 if changeid == b'tip':
1621 1621 node = self.changelog.tip()
1622 1622 rev = self.changelog.rev(node)
1623 1623 return context.changectx(self, rev, node)
1624 1624
1625 1625 # dealing with arbitrary values
1626 1626 try:
1627 1627 if isinstance(changeid, int):
1628 1628 node = self.changelog.node(changeid)
1629 1629 rev = changeid
1630 1630 elif changeid == b'.':
1631 1631 # this is a hack to delay/avoid loading obsmarkers
1632 1632 # when we know that '.' won't be hidden
1633 1633 node = self.dirstate.p1()
1634 1634 rev = self.unfiltered().changelog.rev(node)
1635 1635 elif len(changeid) == 20:
1636 1636 try:
1637 1637 node = changeid
1638 1638 rev = self.changelog.rev(changeid)
1639 1639 except error.FilteredLookupError:
1640 1640 changeid = hex(changeid) # for the error message
1641 1641 raise
1642 1642 except LookupError:
1643 1643 # check if it might have come from damaged dirstate
1644 1644 #
1645 1645 # XXX we could avoid the unfiltered if we had a recognizable
1646 1646 # exception for filtered changeset access
1647 1647 if (
1648 1648 self.local()
1649 1649 and changeid in self.unfiltered().dirstate.parents()
1650 1650 ):
1651 1651 msg = _(b"working directory has unknown parent '%s'!")
1652 1652 raise error.Abort(msg % short(changeid))
1653 1653 changeid = hex(changeid) # for the error message
1654 1654 raise
1655 1655
1656 1656 elif len(changeid) == 40:
1657 1657 node = bin(changeid)
1658 1658 rev = self.changelog.rev(node)
1659 1659 else:
1660 1660 raise error.ProgrammingError(
1661 1661 b"unsupported changeid '%s' of type %s"
1662 1662 % (changeid, pycompat.bytestr(type(changeid)))
1663 1663 )
1664 1664
1665 1665 return context.changectx(self, rev, node)
1666 1666
1667 1667 except (error.FilteredIndexError, error.FilteredLookupError):
1668 1668 raise error.FilteredRepoLookupError(
1669 1669 _(b"filtered revision '%s'") % pycompat.bytestr(changeid)
1670 1670 )
1671 1671 except (IndexError, LookupError):
1672 1672 raise error.RepoLookupError(
1673 1673 _(b"unknown revision '%s'") % pycompat.bytestr(changeid)
1674 1674 )
1675 1675 except error.WdirUnsupported:
1676 1676 return context.workingctx(self)
1677 1677
1678 1678 def __contains__(self, changeid):
1679 1679 """True if the given changeid exists
1680 1680
1681 1681 error.AmbiguousPrefixLookupError is raised if an ambiguous node
1682 1682 specified.
1683 1683 """
1684 1684 try:
1685 1685 self[changeid]
1686 1686 return True
1687 1687 except error.RepoLookupError:
1688 1688 return False
1689 1689
1690 1690 def __nonzero__(self):
1691 1691 return True
1692 1692
1693 1693 __bool__ = __nonzero__
1694 1694
1695 1695 def __len__(self):
1696 1696 # no need to pay the cost of repoview.changelog
1697 1697 unfi = self.unfiltered()
1698 1698 return len(unfi.changelog)
1699 1699
1700 1700 def __iter__(self):
1701 1701 return iter(self.changelog)
1702 1702
1703 1703 def revs(self, expr, *args):
1704 1704 '''Find revisions matching a revset.
1705 1705
1706 1706 The revset is specified as a string ``expr`` that may contain
1707 1707 %-formatting to escape certain types. See ``revsetlang.formatspec``.
1708 1708
1709 1709 Revset aliases from the configuration are not expanded. To expand
1710 1710 user aliases, consider calling ``scmutil.revrange()`` or
1711 1711 ``repo.anyrevs([expr], user=True)``.
1712 1712
1713 1713 Returns a smartset.abstractsmartset, which is a list-like interface
1714 1714 that contains integer revisions.
1715 1715 '''
1716 1716 tree = revsetlang.spectree(expr, *args)
1717 1717 return revset.makematcher(tree)(self)
1718 1718
1719 1719 def set(self, expr, *args):
1720 1720 '''Find revisions matching a revset and emit changectx instances.
1721 1721
1722 1722 This is a convenience wrapper around ``revs()`` that iterates the
1723 1723 result and is a generator of changectx instances.
1724 1724
1725 1725 Revset aliases from the configuration are not expanded. To expand
1726 1726 user aliases, consider calling ``scmutil.revrange()``.
1727 1727 '''
1728 1728 for r in self.revs(expr, *args):
1729 1729 yield self[r]
1730 1730
1731 1731 def anyrevs(self, specs, user=False, localalias=None):
1732 1732 '''Find revisions matching one of the given revsets.
1733 1733
1734 1734 Revset aliases from the configuration are not expanded by default. To
1735 1735 expand user aliases, specify ``user=True``. To provide some local
1736 1736 definitions overriding user aliases, set ``localalias`` to
1737 1737 ``{name: definitionstring}``.
1738 1738 '''
1739 1739 if specs == [b'null']:
1740 1740 return revset.baseset([nullrev])
1741 1741 if specs == [b'.']:
1742 1742 quick_data = self._quick_access_changeid.get(b'.')
1743 1743 if quick_data is not None:
1744 1744 return revset.baseset([quick_data[0]])
1745 1745 if user:
1746 1746 m = revset.matchany(
1747 1747 self.ui,
1748 1748 specs,
1749 1749 lookup=revset.lookupfn(self),
1750 1750 localalias=localalias,
1751 1751 )
1752 1752 else:
1753 1753 m = revset.matchany(None, specs, localalias=localalias)
1754 1754 return m(self)
1755 1755
1756 1756 def url(self):
1757 1757 return b'file:' + self.root
1758 1758
1759 1759 def hook(self, name, throw=False, **args):
1760 1760 """Call a hook, passing this repo instance.
1761 1761
1762 1762 This a convenience method to aid invoking hooks. Extensions likely
1763 1763 won't call this unless they have registered a custom hook or are
1764 1764 replacing code that is expected to call a hook.
1765 1765 """
1766 1766 return hook.hook(self.ui, self, name, throw, **args)
1767 1767
1768 1768 @filteredpropertycache
1769 1769 def _tagscache(self):
1770 1770 '''Returns a tagscache object that contains various tags related
1771 1771 caches.'''
1772 1772
1773 1773 # This simplifies its cache management by having one decorated
1774 1774 # function (this one) and the rest simply fetch things from it.
1775 1775 class tagscache(object):
1776 1776 def __init__(self):
1777 1777 # These two define the set of tags for this repository. tags
1778 1778 # maps tag name to node; tagtypes maps tag name to 'global' or
1779 1779 # 'local'. (Global tags are defined by .hgtags across all
1780 1780 # heads, and local tags are defined in .hg/localtags.)
1781 1781 # They constitute the in-memory cache of tags.
1782 1782 self.tags = self.tagtypes = None
1783 1783
1784 1784 self.nodetagscache = self.tagslist = None
1785 1785
1786 1786 cache = tagscache()
1787 1787 cache.tags, cache.tagtypes = self._findtags()
1788 1788
1789 1789 return cache
1790 1790
1791 1791 def tags(self):
1792 1792 '''return a mapping of tag to node'''
1793 1793 t = {}
1794 1794 if self.changelog.filteredrevs:
1795 1795 tags, tt = self._findtags()
1796 1796 else:
1797 1797 tags = self._tagscache.tags
1798 1798 rev = self.changelog.rev
1799 1799 for k, v in pycompat.iteritems(tags):
1800 1800 try:
1801 1801 # ignore tags to unknown nodes
1802 1802 rev(v)
1803 1803 t[k] = v
1804 1804 except (error.LookupError, ValueError):
1805 1805 pass
1806 1806 return t
1807 1807
1808 1808 def _findtags(self):
1809 1809 '''Do the hard work of finding tags. Return a pair of dicts
1810 1810 (tags, tagtypes) where tags maps tag name to node, and tagtypes
1811 1811 maps tag name to a string like \'global\' or \'local\'.
1812 1812 Subclasses or extensions are free to add their own tags, but
1813 1813 should be aware that the returned dicts will be retained for the
1814 1814 duration of the localrepo object.'''
1815 1815
1816 1816 # XXX what tagtype should subclasses/extensions use? Currently
1817 1817 # mq and bookmarks add tags, but do not set the tagtype at all.
1818 1818 # Should each extension invent its own tag type? Should there
1819 1819 # be one tagtype for all such "virtual" tags? Or is the status
1820 1820 # quo fine?
1821 1821
1822 1822 # map tag name to (node, hist)
1823 1823 alltags = tagsmod.findglobaltags(self.ui, self)
1824 1824 # map tag name to tag type
1825 1825 tagtypes = {tag: b'global' for tag in alltags}
1826 1826
1827 1827 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
1828 1828
1829 1829 # Build the return dicts. Have to re-encode tag names because
1830 1830 # the tags module always uses UTF-8 (in order not to lose info
1831 1831 # writing to the cache), but the rest of Mercurial wants them in
1832 1832 # local encoding.
1833 1833 tags = {}
1834 1834 for (name, (node, hist)) in pycompat.iteritems(alltags):
1835 1835 if node != nullid:
1836 1836 tags[encoding.tolocal(name)] = node
1837 1837 tags[b'tip'] = self.changelog.tip()
1838 1838 tagtypes = {
1839 1839 encoding.tolocal(name): value
1840 1840 for (name, value) in pycompat.iteritems(tagtypes)
1841 1841 }
1842 1842 return (tags, tagtypes)
1843 1843
1844 1844 def tagtype(self, tagname):
1845 1845 '''
1846 1846 return the type of the given tag. result can be:
1847 1847
1848 1848 'local' : a local tag
1849 1849 'global' : a global tag
1850 1850 None : tag does not exist
1851 1851 '''
1852 1852
1853 1853 return self._tagscache.tagtypes.get(tagname)
1854 1854
1855 1855 def tagslist(self):
1856 1856 '''return a list of tags ordered by revision'''
1857 1857 if not self._tagscache.tagslist:
1858 1858 l = []
1859 1859 for t, n in pycompat.iteritems(self.tags()):
1860 1860 l.append((self.changelog.rev(n), t, n))
1861 1861 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
1862 1862
1863 1863 return self._tagscache.tagslist
1864 1864
1865 1865 def nodetags(self, node):
1866 1866 '''return the tags associated with a node'''
1867 1867 if not self._tagscache.nodetagscache:
1868 1868 nodetagscache = {}
1869 1869 for t, n in pycompat.iteritems(self._tagscache.tags):
1870 1870 nodetagscache.setdefault(n, []).append(t)
1871 1871 for tags in pycompat.itervalues(nodetagscache):
1872 1872 tags.sort()
1873 1873 self._tagscache.nodetagscache = nodetagscache
1874 1874 return self._tagscache.nodetagscache.get(node, [])
1875 1875
1876 1876 def nodebookmarks(self, node):
1877 1877 """return the list of bookmarks pointing to the specified node"""
1878 1878 return self._bookmarks.names(node)
1879 1879
1880 1880 def branchmap(self):
1881 1881 '''returns a dictionary {branch: [branchheads]} with branchheads
1882 1882 ordered by increasing revision number'''
1883 1883 return self._branchcaches[self]
1884 1884
1885 1885 @unfilteredmethod
1886 1886 def revbranchcache(self):
1887 1887 if not self._revbranchcache:
1888 1888 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
1889 1889 return self._revbranchcache
1890 1890
1891 1891 def branchtip(self, branch, ignoremissing=False):
1892 1892 '''return the tip node for a given branch
1893 1893
1894 1894 If ignoremissing is True, then this method will not raise an error.
1895 1895 This is helpful for callers that only expect None for a missing branch
1896 1896 (e.g. namespace).
1897 1897
1898 1898 '''
1899 1899 try:
1900 1900 return self.branchmap().branchtip(branch)
1901 1901 except KeyError:
1902 1902 if not ignoremissing:
1903 1903 raise error.RepoLookupError(_(b"unknown branch '%s'") % branch)
1904 1904 else:
1905 1905 pass
1906 1906
1907 1907 def lookup(self, key):
1908 1908 node = scmutil.revsymbol(self, key).node()
1909 1909 if node is None:
1910 1910 raise error.RepoLookupError(_(b"unknown revision '%s'") % key)
1911 1911 return node
1912 1912
1913 1913 def lookupbranch(self, key):
1914 1914 if self.branchmap().hasbranch(key):
1915 1915 return key
1916 1916
1917 1917 return scmutil.revsymbol(self, key).branch()
1918 1918
1919 1919 def known(self, nodes):
1920 1920 cl = self.changelog
1921 1921 get_rev = cl.index.get_rev
1922 1922 filtered = cl.filteredrevs
1923 1923 result = []
1924 1924 for n in nodes:
1925 1925 r = get_rev(n)
1926 1926 resp = not (r is None or r in filtered)
1927 1927 result.append(resp)
1928 1928 return result
1929 1929
1930 1930 def local(self):
1931 1931 return self
1932 1932
1933 1933 def publishing(self):
1934 1934 # it's safe (and desirable) to trust the publish flag unconditionally
1935 1935 # so that we don't finalize changes shared between users via ssh or nfs
1936 1936 return self.ui.configbool(b'phases', b'publish', untrusted=True)
1937 1937
1938 1938 def cancopy(self):
1939 1939 # so statichttprepo's override of local() works
1940 1940 if not self.local():
1941 1941 return False
1942 1942 if not self.publishing():
1943 1943 return True
1944 1944 # if publishing we can't copy if there is filtered content
1945 1945 return not self.filtered(b'visible').changelog.filteredrevs
1946 1946
1947 1947 def shared(self):
1948 1948 '''the type of shared repository (None if not shared)'''
1949 1949 if self.sharedpath != self.path:
1950 1950 return b'store'
1951 1951 return None
1952 1952
1953 1953 def wjoin(self, f, *insidef):
1954 1954 return self.vfs.reljoin(self.root, f, *insidef)
1955 1955
1956 1956 def setparents(self, p1, p2=nullid):
1957 1957 self[None].setparents(p1, p2)
1958 1958 self._quick_access_changeid_invalidate()
1959 1959
1960 1960 def filectx(self, path, changeid=None, fileid=None, changectx=None):
1961 1961 """changeid must be a changeset revision, if specified.
1962 1962 fileid can be a file revision or node."""
1963 1963 return context.filectx(
1964 1964 self, path, changeid, fileid, changectx=changectx
1965 1965 )
1966 1966
1967 1967 def getcwd(self):
1968 1968 return self.dirstate.getcwd()
1969 1969
1970 1970 def pathto(self, f, cwd=None):
1971 1971 return self.dirstate.pathto(f, cwd)
1972 1972
1973 1973 def _loadfilter(self, filter):
1974 1974 if filter not in self._filterpats:
1975 1975 l = []
1976 1976 for pat, cmd in self.ui.configitems(filter):
1977 1977 if cmd == b'!':
1978 1978 continue
1979 1979 mf = matchmod.match(self.root, b'', [pat])
1980 1980 fn = None
1981 1981 params = cmd
1982 1982 for name, filterfn in pycompat.iteritems(self._datafilters):
1983 1983 if cmd.startswith(name):
1984 1984 fn = filterfn
1985 1985 params = cmd[len(name) :].lstrip()
1986 1986 break
1987 1987 if not fn:
1988 1988 fn = lambda s, c, **kwargs: procutil.filter(s, c)
1989 1989 fn.__name__ = 'commandfilter'
1990 1990 # Wrap old filters not supporting keyword arguments
1991 1991 if not pycompat.getargspec(fn)[2]:
1992 1992 oldfn = fn
1993 1993 fn = lambda s, c, oldfn=oldfn, **kwargs: oldfn(s, c)
1994 1994 fn.__name__ = 'compat-' + oldfn.__name__
1995 1995 l.append((mf, fn, params))
1996 1996 self._filterpats[filter] = l
1997 1997 return self._filterpats[filter]
1998 1998
1999 1999 def _filter(self, filterpats, filename, data):
2000 2000 for mf, fn, cmd in filterpats:
2001 2001 if mf(filename):
2002 2002 self.ui.debug(
2003 2003 b"filtering %s through %s\n"
2004 2004 % (filename, cmd or pycompat.sysbytes(fn.__name__))
2005 2005 )
2006 2006 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
2007 2007 break
2008 2008
2009 2009 return data
2010 2010
2011 2011 @unfilteredpropertycache
2012 2012 def _encodefilterpats(self):
2013 2013 return self._loadfilter(b'encode')
2014 2014
2015 2015 @unfilteredpropertycache
2016 2016 def _decodefilterpats(self):
2017 2017 return self._loadfilter(b'decode')
2018 2018
2019 2019 def adddatafilter(self, name, filter):
2020 2020 self._datafilters[name] = filter
2021 2021
2022 2022 def wread(self, filename):
2023 2023 if self.wvfs.islink(filename):
2024 2024 data = self.wvfs.readlink(filename)
2025 2025 else:
2026 2026 data = self.wvfs.read(filename)
2027 2027 return self._filter(self._encodefilterpats, filename, data)
2028 2028
2029 2029 def wwrite(self, filename, data, flags, backgroundclose=False, **kwargs):
2030 2030 """write ``data`` into ``filename`` in the working directory
2031 2031
2032 2032 This returns length of written (maybe decoded) data.
2033 2033 """
2034 2034 data = self._filter(self._decodefilterpats, filename, data)
2035 2035 if b'l' in flags:
2036 2036 self.wvfs.symlink(data, filename)
2037 2037 else:
2038 2038 self.wvfs.write(
2039 2039 filename, data, backgroundclose=backgroundclose, **kwargs
2040 2040 )
2041 2041 if b'x' in flags:
2042 2042 self.wvfs.setflags(filename, False, True)
2043 2043 else:
2044 2044 self.wvfs.setflags(filename, False, False)
2045 2045 return len(data)
2046 2046
2047 2047 def wwritedata(self, filename, data):
2048 2048 return self._filter(self._decodefilterpats, filename, data)
2049 2049
2050 2050 def currenttransaction(self):
2051 2051 """return the current transaction or None if non exists"""
2052 2052 if self._transref:
2053 2053 tr = self._transref()
2054 2054 else:
2055 2055 tr = None
2056 2056
2057 2057 if tr and tr.running():
2058 2058 return tr
2059 2059 return None
2060 2060
2061 2061 def transaction(self, desc, report=None):
2062 2062 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
2063 2063 b'devel', b'check-locks'
2064 2064 ):
2065 2065 if self._currentlock(self._lockref) is None:
2066 2066 raise error.ProgrammingError(b'transaction requires locking')
2067 2067 tr = self.currenttransaction()
2068 2068 if tr is not None:
2069 2069 return tr.nest(name=desc)
2070 2070
2071 2071 # abort here if the journal already exists
2072 2072 if self.svfs.exists(b"journal"):
2073 2073 raise error.RepoError(
2074 2074 _(b"abandoned transaction found"),
2075 2075 hint=_(b"run 'hg recover' to clean up transaction"),
2076 2076 )
2077 2077
2078 2078 idbase = b"%.40f#%f" % (random.random(), time.time())
2079 2079 ha = hex(hashutil.sha1(idbase).digest())
2080 2080 txnid = b'TXN:' + ha
2081 2081 self.hook(b'pretxnopen', throw=True, txnname=desc, txnid=txnid)
2082 2082
2083 2083 self._writejournal(desc)
2084 2084 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
2085 2085 if report:
2086 2086 rp = report
2087 2087 else:
2088 2088 rp = self.ui.warn
2089 2089 vfsmap = {b'plain': self.vfs, b'store': self.svfs} # root of .hg/
2090 2090 # we must avoid cyclic reference between repo and transaction.
2091 2091 reporef = weakref.ref(self)
2092 2092 # Code to track tag movement
2093 2093 #
2094 2094 # Since tags are all handled as file content, it is actually quite hard
2095 2095 # to track these movement from a code perspective. So we fallback to a
2096 2096 # tracking at the repository level. One could envision to track changes
2097 2097 # to the '.hgtags' file through changegroup apply but that fails to
2098 2098 # cope with case where transaction expose new heads without changegroup
2099 2099 # being involved (eg: phase movement).
2100 2100 #
2101 2101 # For now, We gate the feature behind a flag since this likely comes
2102 2102 # with performance impacts. The current code run more often than needed
2103 2103 # and do not use caches as much as it could. The current focus is on
2104 2104 # the behavior of the feature so we disable it by default. The flag
2105 2105 # will be removed when we are happy with the performance impact.
2106 2106 #
2107 2107 # Once this feature is no longer experimental move the following
2108 2108 # documentation to the appropriate help section:
2109 2109 #
2110 2110 # The ``HG_TAG_MOVED`` variable will be set if the transaction touched
2111 2111 # tags (new or changed or deleted tags). In addition the details of
2112 2112 # these changes are made available in a file at:
2113 2113 # ``REPOROOT/.hg/changes/tags.changes``.
2114 2114 # Make sure you check for HG_TAG_MOVED before reading that file as it
2115 2115 # might exist from a previous transaction even if no tag were touched
2116 2116 # in this one. Changes are recorded in a line base format::
2117 2117 #
2118 2118 # <action> <hex-node> <tag-name>\n
2119 2119 #
2120 2120 # Actions are defined as follow:
2121 2121 # "-R": tag is removed,
2122 2122 # "+A": tag is added,
2123 2123 # "-M": tag is moved (old value),
2124 2124 # "+M": tag is moved (new value),
2125 2125 tracktags = lambda x: None
2126 2126 # experimental config: experimental.hook-track-tags
2127 2127 shouldtracktags = self.ui.configbool(
2128 2128 b'experimental', b'hook-track-tags'
2129 2129 )
2130 2130 if desc != b'strip' and shouldtracktags:
2131 2131 oldheads = self.changelog.headrevs()
2132 2132
2133 2133 def tracktags(tr2):
2134 2134 repo = reporef()
2135 2135 oldfnodes = tagsmod.fnoderevs(repo.ui, repo, oldheads)
2136 2136 newheads = repo.changelog.headrevs()
2137 2137 newfnodes = tagsmod.fnoderevs(repo.ui, repo, newheads)
2138 2138 # notes: we compare lists here.
2139 2139 # As we do it only once buiding set would not be cheaper
2140 2140 changes = tagsmod.difftags(repo.ui, repo, oldfnodes, newfnodes)
2141 2141 if changes:
2142 2142 tr2.hookargs[b'tag_moved'] = b'1'
2143 2143 with repo.vfs(
2144 2144 b'changes/tags.changes', b'w', atomictemp=True
2145 2145 ) as changesfile:
2146 2146 # note: we do not register the file to the transaction
2147 2147 # because we needs it to still exist on the transaction
2148 2148 # is close (for txnclose hooks)
2149 2149 tagsmod.writediff(changesfile, changes)
2150 2150
2151 2151 def validate(tr2):
2152 2152 """will run pre-closing hooks"""
2153 2153 # XXX the transaction API is a bit lacking here so we take a hacky
2154 2154 # path for now
2155 2155 #
2156 2156 # We cannot add this as a "pending" hooks since the 'tr.hookargs'
2157 2157 # dict is copied before these run. In addition we needs the data
2158 2158 # available to in memory hooks too.
2159 2159 #
2160 2160 # Moreover, we also need to make sure this runs before txnclose
2161 2161 # hooks and there is no "pending" mechanism that would execute
2162 2162 # logic only if hooks are about to run.
2163 2163 #
2164 2164 # Fixing this limitation of the transaction is also needed to track
2165 2165 # other families of changes (bookmarks, phases, obsolescence).
2166 2166 #
2167 2167 # This will have to be fixed before we remove the experimental
2168 2168 # gating.
2169 2169 tracktags(tr2)
2170 2170 repo = reporef()
2171 2171
2172 2172 singleheadopt = (b'experimental', b'single-head-per-branch')
2173 2173 singlehead = repo.ui.configbool(*singleheadopt)
2174 2174 if singlehead:
2175 2175 singleheadsub = repo.ui.configsuboptions(*singleheadopt)[1]
2176 2176 accountclosed = singleheadsub.get(
2177 2177 b"account-closed-heads", False
2178 2178 )
2179 2179 scmutil.enforcesinglehead(repo, tr2, desc, accountclosed)
2180 2180 if hook.hashook(repo.ui, b'pretxnclose-bookmark'):
2181 2181 for name, (old, new) in sorted(
2182 2182 tr.changes[b'bookmarks'].items()
2183 2183 ):
2184 2184 args = tr.hookargs.copy()
2185 2185 args.update(bookmarks.preparehookargs(name, old, new))
2186 2186 repo.hook(
2187 2187 b'pretxnclose-bookmark',
2188 2188 throw=True,
2189 2189 **pycompat.strkwargs(args)
2190 2190 )
2191 2191 if hook.hashook(repo.ui, b'pretxnclose-phase'):
2192 2192 cl = repo.unfiltered().changelog
2193 2193 for revs, (old, new) in tr.changes[b'phases']:
2194 2194 for rev in revs:
2195 2195 args = tr.hookargs.copy()
2196 2196 node = hex(cl.node(rev))
2197 2197 args.update(phases.preparehookargs(node, old, new))
2198 2198 repo.hook(
2199 2199 b'pretxnclose-phase',
2200 2200 throw=True,
2201 2201 **pycompat.strkwargs(args)
2202 2202 )
2203 2203
2204 2204 repo.hook(
2205 2205 b'pretxnclose', throw=True, **pycompat.strkwargs(tr.hookargs)
2206 2206 )
2207 2207
2208 2208 def releasefn(tr, success):
2209 2209 repo = reporef()
2210 2210 if repo is None:
2211 2211 # If the repo has been GC'd (and this release function is being
2212 2212 # called from transaction.__del__), there's not much we can do,
2213 2213 # so just leave the unfinished transaction there and let the
2214 2214 # user run `hg recover`.
2215 2215 return
2216 2216 if success:
2217 2217 # this should be explicitly invoked here, because
2218 2218 # in-memory changes aren't written out at closing
2219 2219 # transaction, if tr.addfilegenerator (via
2220 2220 # dirstate.write or so) isn't invoked while
2221 2221 # transaction running
2222 2222 repo.dirstate.write(None)
2223 2223 else:
2224 2224 # discard all changes (including ones already written
2225 2225 # out) in this transaction
2226 2226 narrowspec.restorebackup(self, b'journal.narrowspec')
2227 2227 narrowspec.restorewcbackup(self, b'journal.narrowspec.dirstate')
2228 2228 repo.dirstate.restorebackup(None, b'journal.dirstate')
2229 2229
2230 2230 repo.invalidate(clearfilecache=True)
2231 2231
2232 2232 tr = transaction.transaction(
2233 2233 rp,
2234 2234 self.svfs,
2235 2235 vfsmap,
2236 2236 b"journal",
2237 2237 b"undo",
2238 2238 aftertrans(renames),
2239 2239 self.store.createmode,
2240 2240 validator=validate,
2241 2241 releasefn=releasefn,
2242 2242 checkambigfiles=_cachedfiles,
2243 2243 name=desc,
2244 2244 )
2245 2245 tr.changes[b'origrepolen'] = len(self)
2246 2246 tr.changes[b'obsmarkers'] = set()
2247 2247 tr.changes[b'phases'] = []
2248 2248 tr.changes[b'bookmarks'] = {}
2249 2249
2250 2250 tr.hookargs[b'txnid'] = txnid
2251 2251 tr.hookargs[b'txnname'] = desc
2252 2252 tr.hookargs[b'changes'] = tr.changes
2253 2253 # note: writing the fncache only during finalize mean that the file is
2254 2254 # outdated when running hooks. As fncache is used for streaming clone,
2255 2255 # this is not expected to break anything that happen during the hooks.
2256 2256 tr.addfinalize(b'flush-fncache', self.store.write)
2257 2257
2258 2258 def txnclosehook(tr2):
2259 2259 """To be run if transaction is successful, will schedule a hook run
2260 2260 """
2261 2261 # Don't reference tr2 in hook() so we don't hold a reference.
2262 2262 # This reduces memory consumption when there are multiple
2263 2263 # transactions per lock. This can likely go away if issue5045
2264 2264 # fixes the function accumulation.
2265 2265 hookargs = tr2.hookargs
2266 2266
2267 2267 def hookfunc(unused_success):
2268 2268 repo = reporef()
2269 2269 if hook.hashook(repo.ui, b'txnclose-bookmark'):
2270 2270 bmchanges = sorted(tr.changes[b'bookmarks'].items())
2271 2271 for name, (old, new) in bmchanges:
2272 2272 args = tr.hookargs.copy()
2273 2273 args.update(bookmarks.preparehookargs(name, old, new))
2274 2274 repo.hook(
2275 2275 b'txnclose-bookmark',
2276 2276 throw=False,
2277 2277 **pycompat.strkwargs(args)
2278 2278 )
2279 2279
2280 2280 if hook.hashook(repo.ui, b'txnclose-phase'):
2281 2281 cl = repo.unfiltered().changelog
2282 2282 phasemv = sorted(
2283 2283 tr.changes[b'phases'], key=lambda r: r[0][0]
2284 2284 )
2285 2285 for revs, (old, new) in phasemv:
2286 2286 for rev in revs:
2287 2287 args = tr.hookargs.copy()
2288 2288 node = hex(cl.node(rev))
2289 2289 args.update(phases.preparehookargs(node, old, new))
2290 2290 repo.hook(
2291 2291 b'txnclose-phase',
2292 2292 throw=False,
2293 2293 **pycompat.strkwargs(args)
2294 2294 )
2295 2295
2296 2296 repo.hook(
2297 2297 b'txnclose', throw=False, **pycompat.strkwargs(hookargs)
2298 2298 )
2299 2299
2300 2300 reporef()._afterlock(hookfunc)
2301 2301
2302 2302 tr.addfinalize(b'txnclose-hook', txnclosehook)
2303 2303 # Include a leading "-" to make it happen before the transaction summary
2304 2304 # reports registered via scmutil.registersummarycallback() whose names
2305 2305 # are 00-txnreport etc. That way, the caches will be warm when the
2306 2306 # callbacks run.
2307 2307 tr.addpostclose(b'-warm-cache', self._buildcacheupdater(tr))
2308 2308
2309 2309 def txnaborthook(tr2):
2310 2310 """To be run if transaction is aborted
2311 2311 """
2312 2312 reporef().hook(
2313 2313 b'txnabort', throw=False, **pycompat.strkwargs(tr2.hookargs)
2314 2314 )
2315 2315
2316 2316 tr.addabort(b'txnabort-hook', txnaborthook)
2317 2317 # avoid eager cache invalidation. in-memory data should be identical
2318 2318 # to stored data if transaction has no error.
2319 2319 tr.addpostclose(b'refresh-filecachestats', self._refreshfilecachestats)
2320 2320 self._transref = weakref.ref(tr)
2321 2321 scmutil.registersummarycallback(self, tr, desc)
2322 2322 return tr
2323 2323
2324 2324 def _journalfiles(self):
2325 2325 return (
2326 2326 (self.svfs, b'journal'),
2327 2327 (self.svfs, b'journal.narrowspec'),
2328 2328 (self.vfs, b'journal.narrowspec.dirstate'),
2329 2329 (self.vfs, b'journal.dirstate'),
2330 2330 (self.vfs, b'journal.branch'),
2331 2331 (self.vfs, b'journal.desc'),
2332 2332 (bookmarks.bookmarksvfs(self), b'journal.bookmarks'),
2333 2333 (self.svfs, b'journal.phaseroots'),
2334 2334 )
2335 2335
2336 2336 def undofiles(self):
2337 2337 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
2338 2338
2339 2339 @unfilteredmethod
2340 2340 def _writejournal(self, desc):
2341 2341 self.dirstate.savebackup(None, b'journal.dirstate')
2342 2342 narrowspec.savewcbackup(self, b'journal.narrowspec.dirstate')
2343 2343 narrowspec.savebackup(self, b'journal.narrowspec')
2344 2344 self.vfs.write(
2345 2345 b"journal.branch", encoding.fromlocal(self.dirstate.branch())
2346 2346 )
2347 2347 self.vfs.write(b"journal.desc", b"%d\n%s\n" % (len(self), desc))
2348 2348 bookmarksvfs = bookmarks.bookmarksvfs(self)
2349 2349 bookmarksvfs.write(
2350 2350 b"journal.bookmarks", bookmarksvfs.tryread(b"bookmarks")
2351 2351 )
2352 2352 self.svfs.write(b"journal.phaseroots", self.svfs.tryread(b"phaseroots"))
2353 2353
2354 2354 def recover(self):
2355 2355 with self.lock():
2356 2356 if self.svfs.exists(b"journal"):
2357 2357 self.ui.status(_(b"rolling back interrupted transaction\n"))
2358 2358 vfsmap = {
2359 2359 b'': self.svfs,
2360 2360 b'plain': self.vfs,
2361 2361 }
2362 2362 transaction.rollback(
2363 2363 self.svfs,
2364 2364 vfsmap,
2365 2365 b"journal",
2366 2366 self.ui.warn,
2367 2367 checkambigfiles=_cachedfiles,
2368 2368 )
2369 2369 self.invalidate()
2370 2370 return True
2371 2371 else:
2372 2372 self.ui.warn(_(b"no interrupted transaction available\n"))
2373 2373 return False
2374 2374
2375 2375 def rollback(self, dryrun=False, force=False):
2376 2376 wlock = lock = dsguard = None
2377 2377 try:
2378 2378 wlock = self.wlock()
2379 2379 lock = self.lock()
2380 2380 if self.svfs.exists(b"undo"):
2381 2381 dsguard = dirstateguard.dirstateguard(self, b'rollback')
2382 2382
2383 2383 return self._rollback(dryrun, force, dsguard)
2384 2384 else:
2385 2385 self.ui.warn(_(b"no rollback information available\n"))
2386 2386 return 1
2387 2387 finally:
2388 2388 release(dsguard, lock, wlock)
2389 2389
2390 2390 @unfilteredmethod # Until we get smarter cache management
2391 2391 def _rollback(self, dryrun, force, dsguard):
2392 2392 ui = self.ui
2393 2393 try:
2394 2394 args = self.vfs.read(b'undo.desc').splitlines()
2395 2395 (oldlen, desc, detail) = (int(args[0]), args[1], None)
2396 2396 if len(args) >= 3:
2397 2397 detail = args[2]
2398 2398 oldtip = oldlen - 1
2399 2399
2400 2400 if detail and ui.verbose:
2401 2401 msg = _(
2402 2402 b'repository tip rolled back to revision %d'
2403 2403 b' (undo %s: %s)\n'
2404 2404 ) % (oldtip, desc, detail)
2405 2405 else:
2406 2406 msg = _(
2407 2407 b'repository tip rolled back to revision %d (undo %s)\n'
2408 2408 ) % (oldtip, desc)
2409 2409 except IOError:
2410 2410 msg = _(b'rolling back unknown transaction\n')
2411 2411 desc = None
2412 2412
2413 2413 if not force and self[b'.'] != self[b'tip'] and desc == b'commit':
2414 2414 raise error.Abort(
2415 2415 _(
2416 2416 b'rollback of last commit while not checked out '
2417 2417 b'may lose data'
2418 2418 ),
2419 2419 hint=_(b'use -f to force'),
2420 2420 )
2421 2421
2422 2422 ui.status(msg)
2423 2423 if dryrun:
2424 2424 return 0
2425 2425
2426 2426 parents = self.dirstate.parents()
2427 2427 self.destroying()
2428 2428 vfsmap = {b'plain': self.vfs, b'': self.svfs}
2429 2429 transaction.rollback(
2430 2430 self.svfs, vfsmap, b'undo', ui.warn, checkambigfiles=_cachedfiles
2431 2431 )
2432 2432 bookmarksvfs = bookmarks.bookmarksvfs(self)
2433 2433 if bookmarksvfs.exists(b'undo.bookmarks'):
2434 2434 bookmarksvfs.rename(
2435 2435 b'undo.bookmarks', b'bookmarks', checkambig=True
2436 2436 )
2437 2437 if self.svfs.exists(b'undo.phaseroots'):
2438 2438 self.svfs.rename(b'undo.phaseroots', b'phaseroots', checkambig=True)
2439 2439 self.invalidate()
2440 2440
2441 2441 has_node = self.changelog.index.has_node
2442 2442 parentgone = any(not has_node(p) for p in parents)
2443 2443 if parentgone:
2444 2444 # prevent dirstateguard from overwriting already restored one
2445 2445 dsguard.close()
2446 2446
2447 2447 narrowspec.restorebackup(self, b'undo.narrowspec')
2448 2448 narrowspec.restorewcbackup(self, b'undo.narrowspec.dirstate')
2449 2449 self.dirstate.restorebackup(None, b'undo.dirstate')
2450 2450 try:
2451 2451 branch = self.vfs.read(b'undo.branch')
2452 2452 self.dirstate.setbranch(encoding.tolocal(branch))
2453 2453 except IOError:
2454 2454 ui.warn(
2455 2455 _(
2456 2456 b'named branch could not be reset: '
2457 2457 b'current branch is still \'%s\'\n'
2458 2458 )
2459 2459 % self.dirstate.branch()
2460 2460 )
2461 2461
2462 2462 parents = tuple([p.rev() for p in self[None].parents()])
2463 2463 if len(parents) > 1:
2464 2464 ui.status(
2465 2465 _(
2466 2466 b'working directory now based on '
2467 2467 b'revisions %d and %d\n'
2468 2468 )
2469 2469 % parents
2470 2470 )
2471 2471 else:
2472 2472 ui.status(
2473 2473 _(b'working directory now based on revision %d\n') % parents
2474 2474 )
2475 2475 mergestatemod.mergestate.clean(self, self[b'.'].node())
2476 2476
2477 2477 # TODO: if we know which new heads may result from this rollback, pass
2478 2478 # them to destroy(), which will prevent the branchhead cache from being
2479 2479 # invalidated.
2480 2480 self.destroyed()
2481 2481 return 0
2482 2482
2483 2483 def _buildcacheupdater(self, newtransaction):
2484 2484 """called during transaction to build the callback updating cache
2485 2485
2486 2486 Lives on the repository to help extension who might want to augment
2487 2487 this logic. For this purpose, the created transaction is passed to the
2488 2488 method.
2489 2489 """
2490 2490 # we must avoid cyclic reference between repo and transaction.
2491 2491 reporef = weakref.ref(self)
2492 2492
2493 2493 def updater(tr):
2494 2494 repo = reporef()
2495 2495 repo.updatecaches(tr)
2496 2496
2497 2497 return updater
2498 2498
2499 2499 @unfilteredmethod
2500 2500 def updatecaches(self, tr=None, full=False):
2501 2501 """warm appropriate caches
2502 2502
2503 2503 If this function is called after a transaction closed. The transaction
2504 2504 will be available in the 'tr' argument. This can be used to selectively
2505 2505 update caches relevant to the changes in that transaction.
2506 2506
2507 2507 If 'full' is set, make sure all caches the function knows about have
2508 2508 up-to-date data. Even the ones usually loaded more lazily.
2509 2509 """
2510 2510 if tr is not None and tr.hookargs.get(b'source') == b'strip':
2511 2511 # During strip, many caches are invalid but
2512 2512 # later call to `destroyed` will refresh them.
2513 2513 return
2514 2514
2515 2515 if tr is None or tr.changes[b'origrepolen'] < len(self):
2516 2516 # accessing the 'ser ved' branchmap should refresh all the others,
2517 2517 self.ui.debug(b'updating the branch cache\n')
2518 2518 self.filtered(b'served').branchmap()
2519 2519 self.filtered(b'served.hidden').branchmap()
2520 2520
2521 2521 if full:
2522 2522 unfi = self.unfiltered()
2523 2523
2524 2524 self.changelog.update_caches(transaction=tr)
2525 2525 self.manifestlog.update_caches(transaction=tr)
2526 2526
2527 2527 rbc = unfi.revbranchcache()
2528 2528 for r in unfi.changelog:
2529 2529 rbc.branchinfo(r)
2530 2530 rbc.write()
2531 2531
2532 2532 # ensure the working copy parents are in the manifestfulltextcache
2533 2533 for ctx in self[b'.'].parents():
2534 2534 ctx.manifest() # accessing the manifest is enough
2535 2535
2536 2536 # accessing fnode cache warms the cache
2537 2537 tagsmod.fnoderevs(self.ui, unfi, unfi.changelog.revs())
2538 2538 # accessing tags warm the cache
2539 2539 self.tags()
2540 2540 self.filtered(b'served').tags()
2541 2541
2542 2542 # The `full` arg is documented as updating even the lazily-loaded
2543 2543 # caches immediately, so we're forcing a write to cause these caches
2544 2544 # to be warmed up even if they haven't explicitly been requested
2545 2545 # yet (if they've never been used by hg, they won't ever have been
2546 2546 # written, even if they're a subset of another kind of cache that
2547 2547 # *has* been used).
2548 2548 for filt in repoview.filtertable.keys():
2549 2549 filtered = self.filtered(filt)
2550 2550 filtered.branchmap().write(filtered)
2551 2551
2552 2552 def invalidatecaches(self):
2553 2553
2554 2554 if '_tagscache' in vars(self):
2555 2555 # can't use delattr on proxy
2556 2556 del self.__dict__['_tagscache']
2557 2557
2558 2558 self._branchcaches.clear()
2559 2559 self.invalidatevolatilesets()
2560 2560 self._sparsesignaturecache.clear()
2561 2561
2562 2562 def invalidatevolatilesets(self):
2563 2563 self.filteredrevcache.clear()
2564 2564 obsolete.clearobscaches(self)
2565 2565 self._quick_access_changeid_invalidate()
2566 2566
2567 2567 def invalidatedirstate(self):
2568 2568 '''Invalidates the dirstate, causing the next call to dirstate
2569 2569 to check if it was modified since the last time it was read,
2570 2570 rereading it if it has.
2571 2571
2572 2572 This is different to dirstate.invalidate() that it doesn't always
2573 2573 rereads the dirstate. Use dirstate.invalidate() if you want to
2574 2574 explicitly read the dirstate again (i.e. restoring it to a previous
2575 2575 known good state).'''
2576 2576 if hasunfilteredcache(self, 'dirstate'):
2577 2577 for k in self.dirstate._filecache:
2578 2578 try:
2579 2579 delattr(self.dirstate, k)
2580 2580 except AttributeError:
2581 2581 pass
2582 2582 delattr(self.unfiltered(), 'dirstate')
2583 2583
2584 2584 def invalidate(self, clearfilecache=False):
2585 2585 '''Invalidates both store and non-store parts other than dirstate
2586 2586
2587 2587 If a transaction is running, invalidation of store is omitted,
2588 2588 because discarding in-memory changes might cause inconsistency
2589 2589 (e.g. incomplete fncache causes unintentional failure, but
2590 2590 redundant one doesn't).
2591 2591 '''
2592 2592 unfiltered = self.unfiltered() # all file caches are stored unfiltered
2593 2593 for k in list(self._filecache.keys()):
2594 2594 # dirstate is invalidated separately in invalidatedirstate()
2595 2595 if k == b'dirstate':
2596 2596 continue
2597 2597 if (
2598 2598 k == b'changelog'
2599 2599 and self.currenttransaction()
2600 2600 and self.changelog._delayed
2601 2601 ):
2602 2602 # The changelog object may store unwritten revisions. We don't
2603 2603 # want to lose them.
2604 2604 # TODO: Solve the problem instead of working around it.
2605 2605 continue
2606 2606
2607 2607 if clearfilecache:
2608 2608 del self._filecache[k]
2609 2609 try:
2610 2610 delattr(unfiltered, k)
2611 2611 except AttributeError:
2612 2612 pass
2613 2613 self.invalidatecaches()
2614 2614 if not self.currenttransaction():
2615 2615 # TODO: Changing contents of store outside transaction
2616 2616 # causes inconsistency. We should make in-memory store
2617 2617 # changes detectable, and abort if changed.
2618 2618 self.store.invalidatecaches()
2619 2619
2620 2620 def invalidateall(self):
2621 2621 '''Fully invalidates both store and non-store parts, causing the
2622 2622 subsequent operation to reread any outside changes.'''
2623 2623 # extension should hook this to invalidate its caches
2624 2624 self.invalidate()
2625 2625 self.invalidatedirstate()
2626 2626
2627 2627 @unfilteredmethod
2628 2628 def _refreshfilecachestats(self, tr):
2629 2629 """Reload stats of cached files so that they are flagged as valid"""
2630 2630 for k, ce in self._filecache.items():
2631 2631 k = pycompat.sysstr(k)
2632 2632 if k == 'dirstate' or k not in self.__dict__:
2633 2633 continue
2634 2634 ce.refresh()
2635 2635
2636 2636 def _lock(
2637 2637 self,
2638 2638 vfs,
2639 2639 lockname,
2640 2640 wait,
2641 2641 releasefn,
2642 2642 acquirefn,
2643 2643 desc,
2644 2644 inheritchecker=None,
2645 2645 parentenvvar=None,
2646 2646 ):
2647 2647 parentlock = None
2648 2648 # the contents of parentenvvar are used by the underlying lock to
2649 2649 # determine whether it can be inherited
2650 2650 if parentenvvar is not None:
2651 2651 parentlock = encoding.environ.get(parentenvvar)
2652 2652
2653 2653 timeout = 0
2654 2654 warntimeout = 0
2655 2655 if wait:
2656 2656 timeout = self.ui.configint(b"ui", b"timeout")
2657 2657 warntimeout = self.ui.configint(b"ui", b"timeout.warn")
2658 2658 # internal config: ui.signal-safe-lock
2659 2659 signalsafe = self.ui.configbool(b'ui', b'signal-safe-lock')
2660 2660
2661 2661 l = lockmod.trylock(
2662 2662 self.ui,
2663 2663 vfs,
2664 2664 lockname,
2665 2665 timeout,
2666 2666 warntimeout,
2667 2667 releasefn=releasefn,
2668 2668 acquirefn=acquirefn,
2669 2669 desc=desc,
2670 2670 inheritchecker=inheritchecker,
2671 2671 parentlock=parentlock,
2672 2672 signalsafe=signalsafe,
2673 2673 )
2674 2674 return l
2675 2675
2676 2676 def _afterlock(self, callback):
2677 2677 """add a callback to be run when the repository is fully unlocked
2678 2678
2679 2679 The callback will be executed when the outermost lock is released
2680 2680 (with wlock being higher level than 'lock')."""
2681 2681 for ref in (self._wlockref, self._lockref):
2682 2682 l = ref and ref()
2683 2683 if l and l.held:
2684 2684 l.postrelease.append(callback)
2685 2685 break
2686 2686 else: # no lock have been found.
2687 2687 callback(True)
2688 2688
2689 2689 def lock(self, wait=True):
2690 2690 '''Lock the repository store (.hg/store) and return a weak reference
2691 2691 to the lock. Use this before modifying the store (e.g. committing or
2692 2692 stripping). If you are opening a transaction, get a lock as well.)
2693 2693
2694 2694 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2695 2695 'wlock' first to avoid a dead-lock hazard.'''
2696 2696 l = self._currentlock(self._lockref)
2697 2697 if l is not None:
2698 2698 l.lock()
2699 2699 return l
2700 2700
2701 2701 l = self._lock(
2702 2702 vfs=self.svfs,
2703 2703 lockname=b"lock",
2704 2704 wait=wait,
2705 2705 releasefn=None,
2706 2706 acquirefn=self.invalidate,
2707 2707 desc=_(b'repository %s') % self.origroot,
2708 2708 )
2709 2709 self._lockref = weakref.ref(l)
2710 2710 return l
2711 2711
2712 2712 def _wlockchecktransaction(self):
2713 2713 if self.currenttransaction() is not None:
2714 2714 raise error.LockInheritanceContractViolation(
2715 2715 b'wlock cannot be inherited in the middle of a transaction'
2716 2716 )
2717 2717
2718 2718 def wlock(self, wait=True):
2719 2719 '''Lock the non-store parts of the repository (everything under
2720 2720 .hg except .hg/store) and return a weak reference to the lock.
2721 2721
2722 2722 Use this before modifying files in .hg.
2723 2723
2724 2724 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2725 2725 'wlock' first to avoid a dead-lock hazard.'''
2726 2726 l = self._wlockref and self._wlockref()
2727 2727 if l is not None and l.held:
2728 2728 l.lock()
2729 2729 return l
2730 2730
2731 2731 # We do not need to check for non-waiting lock acquisition. Such
2732 2732 # acquisition would not cause dead-lock as they would just fail.
2733 2733 if wait and (
2734 2734 self.ui.configbool(b'devel', b'all-warnings')
2735 2735 or self.ui.configbool(b'devel', b'check-locks')
2736 2736 ):
2737 2737 if self._currentlock(self._lockref) is not None:
2738 2738 self.ui.develwarn(b'"wlock" acquired after "lock"')
2739 2739
2740 2740 def unlock():
2741 2741 if self.dirstate.pendingparentchange():
2742 2742 self.dirstate.invalidate()
2743 2743 else:
2744 2744 self.dirstate.write(None)
2745 2745
2746 2746 self._filecache[b'dirstate'].refresh()
2747 2747
2748 2748 l = self._lock(
2749 2749 self.vfs,
2750 2750 b"wlock",
2751 2751 wait,
2752 2752 unlock,
2753 2753 self.invalidatedirstate,
2754 2754 _(b'working directory of %s') % self.origroot,
2755 2755 inheritchecker=self._wlockchecktransaction,
2756 2756 parentenvvar=b'HG_WLOCK_LOCKER',
2757 2757 )
2758 2758 self._wlockref = weakref.ref(l)
2759 2759 return l
2760 2760
2761 2761 def _currentlock(self, lockref):
2762 2762 """Returns the lock if it's held, or None if it's not."""
2763 2763 if lockref is None:
2764 2764 return None
2765 2765 l = lockref()
2766 2766 if l is None or not l.held:
2767 2767 return None
2768 2768 return l
2769 2769
2770 2770 def currentwlock(self):
2771 2771 """Returns the wlock if it's held, or None if it's not."""
2772 2772 return self._currentlock(self._wlockref)
2773 2773
2774 2774 def _filecommit(
2775 2775 self, fctx, manifest1, manifest2, linkrev, tr, includecopymeta,
2776 2776 ):
2777 2777 """
2778 2778 commit an individual file as part of a larger transaction
2779 2779
2780 2780 input:
2781 2781
2782 2782 fctx: a file context with the content we are trying to commit
2783 2783 manifest1: manifest of changeset first parent
2784 2784 manifest2: manifest of changeset second parent
2785 2785 linkrev: revision number of the changeset being created
2786 2786 tr: current transation
2787 2787 individual: boolean, set to False to skip storing the copy data
2788 2788 (only used by the Google specific feature of using
2789 2789 changeset extra as copy source of truth).
2790 2790
2791 2791 output: (filenode, touched)
2792 2792
2793 2793 filenode: the filenode that should be used by this changeset
2794 2794 touched: one of: None, 'added' or 'modified'
2795 2795 """
2796 2796
2797 2797 fname = fctx.path()
2798 2798 fparent1 = manifest1.get(fname, nullid)
2799 2799 fparent2 = manifest2.get(fname, nullid)
2800 2800 touched = None
2801 2801 if fparent1 == fparent2 == nullid:
2802 2802 touched = 'added'
2803 2803
2804 2804 if isinstance(fctx, context.filectx):
2805 2805 # This block fast path most comparisons which are usually done. It
2806 2806 # assumes that bare filectx is used and no merge happened, hence no
2807 2807 # need to create a new file revision in this case.
2808 2808 node = fctx.filenode()
2809 2809 if node in [fparent1, fparent2]:
2810 2810 self.ui.debug(b'reusing %s filelog entry\n' % fname)
2811 2811 if (
2812 2812 fparent1 != nullid
2813 2813 and manifest1.flags(fname) != fctx.flags()
2814 2814 ) or (
2815 2815 fparent2 != nullid
2816 2816 and manifest2.flags(fname) != fctx.flags()
2817 2817 ):
2818 2818 touched = 'modified'
2819 2819 return node, touched
2820 2820
2821 2821 flog = self.file(fname)
2822 2822 meta = {}
2823 2823 cfname = fctx.copysource()
2824 2824 fnode = None
2825 2825
2826 2826 if cfname and cfname != fname:
2827 2827 # Mark the new revision of this file as a copy of another
2828 2828 # file. This copy data will effectively act as a parent
2829 2829 # of this new revision. If this is a merge, the first
2830 2830 # parent will be the nullid (meaning "look up the copy data")
2831 2831 # and the second one will be the other parent. For example:
2832 2832 #
2833 2833 # 0 --- 1 --- 3 rev1 changes file foo
2834 2834 # \ / rev2 renames foo to bar and changes it
2835 2835 # \- 2 -/ rev3 should have bar with all changes and
2836 2836 # should record that bar descends from
2837 2837 # bar in rev2 and foo in rev1
2838 2838 #
2839 2839 # this allows this merge to succeed:
2840 2840 #
2841 2841 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
2842 2842 # \ / merging rev3 and rev4 should use bar@rev2
2843 2843 # \- 2 --- 4 as the merge base
2844 2844 #
2845 2845
2846 2846 cnode = manifest1.get(cfname)
2847 2847 newfparent = fparent2
2848 2848
2849 2849 if manifest2: # branch merge
2850 2850 if fparent2 == nullid or cnode is None: # copied on remote side
2851 2851 if cfname in manifest2:
2852 2852 cnode = manifest2[cfname]
2853 2853 newfparent = fparent1
2854 2854
2855 2855 # Here, we used to search backwards through history to try to find
2856 2856 # where the file copy came from if the source of a copy was not in
2857 2857 # the parent directory. However, this doesn't actually make sense to
2858 2858 # do (what does a copy from something not in your working copy even
2859 2859 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
2860 2860 # the user that copy information was dropped, so if they didn't
2861 2861 # expect this outcome it can be fixed, but this is the correct
2862 2862 # behavior in this circumstance.
2863 2863
2864 2864 if cnode:
2865 2865 self.ui.debug(
2866 2866 b" %s: copy %s:%s\n" % (fname, cfname, hex(cnode))
2867 2867 )
2868 2868 if includecopymeta:
2869 2869 meta[b"copy"] = cfname
2870 2870 meta[b"copyrev"] = hex(cnode)
2871 2871 fparent1, fparent2 = nullid, newfparent
2872 2872 else:
2873 2873 self.ui.warn(
2874 2874 _(
2875 2875 b"warning: can't find ancestor for '%s' "
2876 2876 b"copied from '%s'!\n"
2877 2877 )
2878 2878 % (fname, cfname)
2879 2879 )
2880 2880
2881 2881 elif fparent1 == nullid:
2882 2882 fparent1, fparent2 = fparent2, nullid
2883 2883 elif fparent2 != nullid:
2884 2884 # is one parent an ancestor of the other?
2885 2885 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
2886 2886 if fparent1 in fparentancestors:
2887 2887 fparent1, fparent2 = fparent2, nullid
2888 2888 elif fparent2 in fparentancestors:
2889 2889 fparent2 = nullid
2890 2890 elif not fparentancestors:
2891 2891 # TODO: this whole if-else might be simplified much more
2892 2892 ms = mergestatemod.mergestate.read(self)
2893 2893 if (
2894 2894 fname in ms
2895 2895 and ms[fname] == mergestatemod.MERGE_RECORD_MERGED_OTHER
2896 2896 ):
2897 2897 fparent1, fparent2 = fparent2, nullid
2898 2898
2899 2899 # is the file changed?
2900 2900 text = fctx.data()
2901 2901 if fparent2 != nullid or meta or flog.cmp(fparent1, text):
2902 2902 if touched is None: # do not overwrite added
2903 2903 touched = 'modified'
2904 2904 fnode = flog.add(text, meta, tr, linkrev, fparent1, fparent2)
2905 2905 # are just the flags changed during merge?
2906 2906 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
2907 2907 touched = 'modified'
2908 2908 fnode = fparent1
2909 2909 else:
2910 2910 fnode = fparent1
2911 2911 return fnode, touched
2912 2912
2913 2913 def checkcommitpatterns(self, wctx, match, status, fail):
2914 2914 """check for commit arguments that aren't committable"""
2915 2915 if match.isexact() or match.prefix():
2916 2916 matched = set(status.modified + status.added + status.removed)
2917 2917
2918 2918 for f in match.files():
2919 2919 f = self.dirstate.normalize(f)
2920 2920 if f == b'.' or f in matched or f in wctx.substate:
2921 2921 continue
2922 2922 if f in status.deleted:
2923 2923 fail(f, _(b'file not found!'))
2924 2924 # Is it a directory that exists or used to exist?
2925 2925 if self.wvfs.isdir(f) or wctx.p1().hasdir(f):
2926 2926 d = f + b'/'
2927 2927 for mf in matched:
2928 2928 if mf.startswith(d):
2929 2929 break
2930 2930 else:
2931 2931 fail(f, _(b"no match under directory!"))
2932 2932 elif f not in self.dirstate:
2933 2933 fail(f, _(b"file not tracked!"))
2934 2934
2935 2935 @unfilteredmethod
2936 2936 def commit(
2937 2937 self,
2938 2938 text=b"",
2939 2939 user=None,
2940 2940 date=None,
2941 2941 match=None,
2942 2942 force=False,
2943 2943 editor=None,
2944 2944 extra=None,
2945 2945 ):
2946 2946 """Add a new revision to current repository.
2947 2947
2948 2948 Revision information is gathered from the working directory,
2949 2949 match can be used to filter the committed files. If editor is
2950 2950 supplied, it is called to get a commit message.
2951 2951 """
2952 2952 if extra is None:
2953 2953 extra = {}
2954 2954
2955 2955 def fail(f, msg):
2956 2956 raise error.Abort(b'%s: %s' % (f, msg))
2957 2957
2958 2958 if not match:
2959 2959 match = matchmod.always()
2960 2960
2961 2961 if not force:
2962 2962 match.bad = fail
2963 2963
2964 2964 # lock() for recent changelog (see issue4368)
2965 2965 with self.wlock(), self.lock():
2966 2966 wctx = self[None]
2967 2967 merge = len(wctx.parents()) > 1
2968 2968
2969 2969 if not force and merge and not match.always():
2970 2970 raise error.Abort(
2971 2971 _(
2972 2972 b'cannot partially commit a merge '
2973 2973 b'(do not specify files or patterns)'
2974 2974 )
2975 2975 )
2976 2976
2977 2977 status = self.status(match=match, clean=force)
2978 2978 if force:
2979 2979 status.modified.extend(
2980 2980 status.clean
2981 2981 ) # mq may commit clean files
2982 2982
2983 2983 # check subrepos
2984 2984 subs, commitsubs, newstate = subrepoutil.precommit(
2985 2985 self.ui, wctx, status, match, force=force
2986 2986 )
2987 2987
2988 2988 # make sure all explicit patterns are matched
2989 2989 if not force:
2990 2990 self.checkcommitpatterns(wctx, match, status, fail)
2991 2991
2992 2992 cctx = context.workingcommitctx(
2993 2993 self, status, text, user, date, extra
2994 2994 )
2995 2995
2996 2996 ms = mergestatemod.mergestate.read(self)
2997 2997 mergeutil.checkunresolved(ms)
2998 2998
2999 2999 # internal config: ui.allowemptycommit
3000 3000 if cctx.isempty() and not self.ui.configbool(
3001 3001 b'ui', b'allowemptycommit'
3002 3002 ):
3003 3003 self.ui.debug(b'nothing to commit, clearing merge state\n')
3004 3004 ms.reset()
3005 3005 return None
3006 3006
3007 3007 if merge and cctx.deleted():
3008 3008 raise error.Abort(_(b"cannot commit merge with missing files"))
3009 3009
3010 3010 if editor:
3011 3011 cctx._text = editor(self, cctx, subs)
3012 3012 edited = text != cctx._text
3013 3013
3014 3014 # Save commit message in case this transaction gets rolled back
3015 3015 # (e.g. by a pretxncommit hook). Leave the content alone on
3016 3016 # the assumption that the user will use the same editor again.
3017 3017 msgfn = self.savecommitmessage(cctx._text)
3018 3018
3019 3019 # commit subs and write new state
3020 3020 if subs:
3021 3021 uipathfn = scmutil.getuipathfn(self)
3022 3022 for s in sorted(commitsubs):
3023 3023 sub = wctx.sub(s)
3024 3024 self.ui.status(
3025 3025 _(b'committing subrepository %s\n')
3026 3026 % uipathfn(subrepoutil.subrelpath(sub))
3027 3027 )
3028 3028 sr = sub.commit(cctx._text, user, date)
3029 3029 newstate[s] = (newstate[s][0], sr)
3030 3030 subrepoutil.writestate(self, newstate)
3031 3031
3032 3032 p1, p2 = self.dirstate.parents()
3033 3033 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or b'')
3034 3034 try:
3035 3035 self.hook(
3036 3036 b"precommit", throw=True, parent1=hookp1, parent2=hookp2
3037 3037 )
3038 3038 with self.transaction(b'commit'):
3039 3039 ret = self.commitctx(cctx, True)
3040 3040 # update bookmarks, dirstate and mergestate
3041 3041 bookmarks.update(self, [p1, p2], ret)
3042 3042 cctx.markcommitted(ret)
3043 3043 ms.reset()
3044 3044 except: # re-raises
3045 3045 if edited:
3046 3046 self.ui.write(
3047 3047 _(b'note: commit message saved in %s\n') % msgfn
3048 3048 )
3049 3049 self.ui.write(
3050 3050 _(
3051 3051 b"note: use 'hg commit --logfile "
3052 3052 b".hg/last-message.txt --edit' to reuse it\n"
3053 3053 )
3054 3054 )
3055 3055 raise
3056 3056
3057 3057 def commithook(unused_success):
3058 3058 # hack for command that use a temporary commit (eg: histedit)
3059 3059 # temporary commit got stripped before hook release
3060 3060 if self.changelog.hasnode(ret):
3061 3061 self.hook(
3062 3062 b"commit", node=hex(ret), parent1=hookp1, parent2=hookp2
3063 3063 )
3064 3064
3065 3065 self._afterlock(commithook)
3066 3066 return ret
3067 3067
3068 3068 @unfilteredmethod
3069 3069 def commitctx(self, ctx, error=False, origctx=None):
3070 """Add a new revision to current repository.
3071 Revision information is passed via the context argument.
3072
3073 ctx.files() should list all files involved in this commit, i.e.
3074 modified/added/removed files. On merge, it may be wider than the
3075 ctx.files() to be committed, since any file nodes derived directly
3076 from p1 or p2 are excluded from the committed ctx.files().
3077
3078 origctx is for convert to work around the problem that bug
3079 fixes to the files list in changesets change hashes. For
3080 convert to be the identity, it can pass an origctx and this
3081 function will use the same files list when it makes sense to
3082 do so.
3083 """
3084
3085 p1, p2 = ctx.p1(), ctx.p2()
3086 user = ctx.user()
3087
3088 if self.filecopiesmode == b'changeset-sidedata':
3089 writechangesetcopy = True
3090 writefilecopymeta = True
3091 writecopiesto = None
3092 else:
3093 writecopiesto = self.ui.config(b'experimental', b'copies.write-to')
3094 writefilecopymeta = writecopiesto != b'changeset-only'
3095 writechangesetcopy = writecopiesto in (
3096 b'changeset-only',
3097 b'compatibility',
3098 )
3099 p1copies, p2copies = None, None
3100 if writechangesetcopy:
3101 p1copies = ctx.p1copies()
3102 p2copies = ctx.p2copies()
3103 filesadded, filesremoved = None, None
3104 with self.lock(), self.transaction(b"commit") as tr:
3105 trp = weakref.proxy(tr)
3106
3107 if ctx.manifestnode():
3108 # reuse an existing manifest revision
3109 self.ui.debug(b'reusing known manifest\n')
3110 mn = ctx.manifestnode()
3111 files = ctx.files()
3112 if writechangesetcopy:
3113 filesadded = ctx.filesadded()
3114 filesremoved = ctx.filesremoved()
3115 elif not ctx.files():
3116 self.ui.debug(b'reusing manifest from p1 (no file change)\n')
3117 mn = p1.manifestnode()
3118 files = []
3119 else:
3120 m1ctx = p1.manifestctx()
3121 m2ctx = p2.manifestctx()
3122 mctx = m1ctx.copy()
3123
3124 m = mctx.read()
3125 m1 = m1ctx.read()
3126 m2 = m2ctx.read()
3127
3128 # check in files
3129 added = []
3130 filesadded = []
3131 removed = list(ctx.removed())
3132 touched = []
3133 linkrev = len(self)
3134 self.ui.note(_(b"committing files:\n"))
3135 uipathfn = scmutil.getuipathfn(self)
3136 for f in sorted(ctx.modified() + ctx.added()):
3137 self.ui.note(uipathfn(f) + b"\n")
3138 try:
3139 fctx = ctx[f]
3140 if fctx is None:
3141 removed.append(f)
3142 else:
3143 added.append(f)
3144 m[f], is_touched = self._filecommit(
3145 fctx, m1, m2, linkrev, trp, writefilecopymeta,
3146 )
3147 if is_touched:
3148 touched.append(f)
3149 if writechangesetcopy and is_touched == 'added':
3150 filesadded.append(f)
3151 m.setflag(f, fctx.flags())
3152 except OSError:
3153 self.ui.warn(
3154 _(b"trouble committing %s!\n") % uipathfn(f)
3155 )
3156 raise
3157 except IOError as inst:
3158 errcode = getattr(inst, 'errno', errno.ENOENT)
3159 if error or errcode and errcode != errno.ENOENT:
3160 self.ui.warn(
3161 _(b"trouble committing %s!\n") % uipathfn(f)
3162 )
3163 raise
3164
3165 # update manifest
3166 removed = [f for f in removed if f in m1 or f in m2]
3167 drop = sorted([f for f in removed if f in m])
3168 for f in drop:
3169 del m[f]
3170 if p2.rev() != nullrev:
3171 rf = metadata.get_removal_filter(ctx, (p1, p2, m1, m2))
3172 removed = [f for f in removed if not rf(f)]
3173
3174 touched.extend(removed)
3175
3176 if writechangesetcopy:
3177 filesremoved = removed
3178
3179 files = touched
3180 md = None
3181 if not files:
3182 # if no "files" actually changed in terms of the changelog,
3183 # try hard to detect unmodified manifest entry so that the
3184 # exact same commit can be reproduced later on convert.
3185 md = m1.diff(m, scmutil.matchfiles(self, ctx.files()))
3186 if not files and md:
3187 self.ui.debug(
3188 b'not reusing manifest (no file change in '
3189 b'changelog, but manifest differs)\n'
3190 )
3191 if files or md:
3192 self.ui.note(_(b"committing manifest\n"))
3193 # we're using narrowmatch here since it's already applied at
3194 # other stages (such as dirstate.walk), so we're already
3195 # ignoring things outside of narrowspec in most cases. The
3196 # one case where we might have files outside the narrowspec
3197 # at this point is merges, and we already error out in the
3198 # case where the merge has files outside of the narrowspec,
3199 # so this is safe.
3200 mn = mctx.write(
3201 trp,
3202 linkrev,
3203 p1.manifestnode(),
3204 p2.manifestnode(),
3205 added,
3206 drop,
3207 match=self.narrowmatch(),
3208 )
3209 else:
3210 self.ui.debug(
3211 b'reusing manifest from p1 (listed files '
3212 b'actually unchanged)\n'
3213 )
3214 mn = p1.manifestnode()
3215
3216 if writecopiesto == b'changeset-only':
3217 # If writing only to changeset extras, use None to indicate that
3218 # no entry should be written. If writing to both, write an empty
3219 # entry to prevent the reader from falling back to reading
3220 # filelogs.
3221 p1copies = p1copies or None
3222 p2copies = p2copies or None
3223 filesadded = filesadded or None
3224 filesremoved = filesremoved or None
3225
3226 if origctx and origctx.manifestnode() == mn:
3227 files = origctx.files()
3228
3229 # update changelog
3230 self.ui.note(_(b"committing changelog\n"))
3231 self.changelog.delayupdate(tr)
3232 n = self.changelog.add(
3233 mn,
3234 files,
3235 ctx.description(),
3236 trp,
3237 p1.node(),
3238 p2.node(),
3239 user,
3240 ctx.date(),
3241 ctx.extra().copy(),
3242 p1copies,
3243 p2copies,
3244 filesadded,
3245 filesremoved,
3246 )
3247 xp1, xp2 = p1.hex(), p2 and p2.hex() or b''
3248 self.hook(
3249 b'pretxncommit',
3250 throw=True,
3251 node=hex(n),
3252 parent1=xp1,
3253 parent2=xp2,
3254 )
3255 # set the new commit is proper phase
3256 targetphase = subrepoutil.newcommitphase(self.ui, ctx)
3257 if targetphase:
3258 # retract boundary do not alter parent changeset.
3259 # if a parent have higher the resulting phase will
3260 # be compliant anyway
3261 #
3262 # if minimal phase was 0 we don't need to retract anything
3263 phases.registernew(self, tr, targetphase, [n])
3264 return n
3070 return commit.commitctx(self, ctx, error=error, origctx=origctx)
3265 3071
3266 3072 @unfilteredmethod
3267 3073 def destroying(self):
3268 3074 '''Inform the repository that nodes are about to be destroyed.
3269 3075 Intended for use by strip and rollback, so there's a common
3270 3076 place for anything that has to be done before destroying history.
3271 3077
3272 3078 This is mostly useful for saving state that is in memory and waiting
3273 3079 to be flushed when the current lock is released. Because a call to
3274 3080 destroyed is imminent, the repo will be invalidated causing those
3275 3081 changes to stay in memory (waiting for the next unlock), or vanish
3276 3082 completely.
3277 3083 '''
3278 3084 # When using the same lock to commit and strip, the phasecache is left
3279 3085 # dirty after committing. Then when we strip, the repo is invalidated,
3280 3086 # causing those changes to disappear.
3281 3087 if '_phasecache' in vars(self):
3282 3088 self._phasecache.write()
3283 3089
3284 3090 @unfilteredmethod
3285 3091 def destroyed(self):
3286 3092 '''Inform the repository that nodes have been destroyed.
3287 3093 Intended for use by strip and rollback, so there's a common
3288 3094 place for anything that has to be done after destroying history.
3289 3095 '''
3290 3096 # When one tries to:
3291 3097 # 1) destroy nodes thus calling this method (e.g. strip)
3292 3098 # 2) use phasecache somewhere (e.g. commit)
3293 3099 #
3294 3100 # then 2) will fail because the phasecache contains nodes that were
3295 3101 # removed. We can either remove phasecache from the filecache,
3296 3102 # causing it to reload next time it is accessed, or simply filter
3297 3103 # the removed nodes now and write the updated cache.
3298 3104 self._phasecache.filterunknown(self)
3299 3105 self._phasecache.write()
3300 3106
3301 3107 # refresh all repository caches
3302 3108 self.updatecaches()
3303 3109
3304 3110 # Ensure the persistent tag cache is updated. Doing it now
3305 3111 # means that the tag cache only has to worry about destroyed
3306 3112 # heads immediately after a strip/rollback. That in turn
3307 3113 # guarantees that "cachetip == currenttip" (comparing both rev
3308 3114 # and node) always means no nodes have been added or destroyed.
3309 3115
3310 3116 # XXX this is suboptimal when qrefresh'ing: we strip the current
3311 3117 # head, refresh the tag cache, then immediately add a new head.
3312 3118 # But I think doing it this way is necessary for the "instant
3313 3119 # tag cache retrieval" case to work.
3314 3120 self.invalidate()
3315 3121
3316 3122 def status(
3317 3123 self,
3318 3124 node1=b'.',
3319 3125 node2=None,
3320 3126 match=None,
3321 3127 ignored=False,
3322 3128 clean=False,
3323 3129 unknown=False,
3324 3130 listsubrepos=False,
3325 3131 ):
3326 3132 '''a convenience method that calls node1.status(node2)'''
3327 3133 return self[node1].status(
3328 3134 node2, match, ignored, clean, unknown, listsubrepos
3329 3135 )
3330 3136
3331 3137 def addpostdsstatus(self, ps):
3332 3138 """Add a callback to run within the wlock, at the point at which status
3333 3139 fixups happen.
3334 3140
3335 3141 On status completion, callback(wctx, status) will be called with the
3336 3142 wlock held, unless the dirstate has changed from underneath or the wlock
3337 3143 couldn't be grabbed.
3338 3144
3339 3145 Callbacks should not capture and use a cached copy of the dirstate --
3340 3146 it might change in the meanwhile. Instead, they should access the
3341 3147 dirstate via wctx.repo().dirstate.
3342 3148
3343 3149 This list is emptied out after each status run -- extensions should
3344 3150 make sure it adds to this list each time dirstate.status is called.
3345 3151 Extensions should also make sure they don't call this for statuses
3346 3152 that don't involve the dirstate.
3347 3153 """
3348 3154
3349 3155 # The list is located here for uniqueness reasons -- it is actually
3350 3156 # managed by the workingctx, but that isn't unique per-repo.
3351 3157 self._postdsstatus.append(ps)
3352 3158
3353 3159 def postdsstatus(self):
3354 3160 """Used by workingctx to get the list of post-dirstate-status hooks."""
3355 3161 return self._postdsstatus
3356 3162
3357 3163 def clearpostdsstatus(self):
3358 3164 """Used by workingctx to clear post-dirstate-status hooks."""
3359 3165 del self._postdsstatus[:]
3360 3166
3361 3167 def heads(self, start=None):
3362 3168 if start is None:
3363 3169 cl = self.changelog
3364 3170 headrevs = reversed(cl.headrevs())
3365 3171 return [cl.node(rev) for rev in headrevs]
3366 3172
3367 3173 heads = self.changelog.heads(start)
3368 3174 # sort the output in rev descending order
3369 3175 return sorted(heads, key=self.changelog.rev, reverse=True)
3370 3176
3371 3177 def branchheads(self, branch=None, start=None, closed=False):
3372 3178 '''return a (possibly filtered) list of heads for the given branch
3373 3179
3374 3180 Heads are returned in topological order, from newest to oldest.
3375 3181 If branch is None, use the dirstate branch.
3376 3182 If start is not None, return only heads reachable from start.
3377 3183 If closed is True, return heads that are marked as closed as well.
3378 3184 '''
3379 3185 if branch is None:
3380 3186 branch = self[None].branch()
3381 3187 branches = self.branchmap()
3382 3188 if not branches.hasbranch(branch):
3383 3189 return []
3384 3190 # the cache returns heads ordered lowest to highest
3385 3191 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
3386 3192 if start is not None:
3387 3193 # filter out the heads that cannot be reached from startrev
3388 3194 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
3389 3195 bheads = [h for h in bheads if h in fbheads]
3390 3196 return bheads
3391 3197
3392 3198 def branches(self, nodes):
3393 3199 if not nodes:
3394 3200 nodes = [self.changelog.tip()]
3395 3201 b = []
3396 3202 for n in nodes:
3397 3203 t = n
3398 3204 while True:
3399 3205 p = self.changelog.parents(n)
3400 3206 if p[1] != nullid or p[0] == nullid:
3401 3207 b.append((t, n, p[0], p[1]))
3402 3208 break
3403 3209 n = p[0]
3404 3210 return b
3405 3211
3406 3212 def between(self, pairs):
3407 3213 r = []
3408 3214
3409 3215 for top, bottom in pairs:
3410 3216 n, l, i = top, [], 0
3411 3217 f = 1
3412 3218
3413 3219 while n != bottom and n != nullid:
3414 3220 p = self.changelog.parents(n)[0]
3415 3221 if i == f:
3416 3222 l.append(n)
3417 3223 f = f * 2
3418 3224 n = p
3419 3225 i += 1
3420 3226
3421 3227 r.append(l)
3422 3228
3423 3229 return r
3424 3230
3425 3231 def checkpush(self, pushop):
3426 3232 """Extensions can override this function if additional checks have
3427 3233 to be performed before pushing, or call it if they override push
3428 3234 command.
3429 3235 """
3430 3236
3431 3237 @unfilteredpropertycache
3432 3238 def prepushoutgoinghooks(self):
3433 3239 """Return util.hooks consists of a pushop with repo, remote, outgoing
3434 3240 methods, which are called before pushing changesets.
3435 3241 """
3436 3242 return util.hooks()
3437 3243
3438 3244 def pushkey(self, namespace, key, old, new):
3439 3245 try:
3440 3246 tr = self.currenttransaction()
3441 3247 hookargs = {}
3442 3248 if tr is not None:
3443 3249 hookargs.update(tr.hookargs)
3444 3250 hookargs = pycompat.strkwargs(hookargs)
3445 3251 hookargs['namespace'] = namespace
3446 3252 hookargs['key'] = key
3447 3253 hookargs['old'] = old
3448 3254 hookargs['new'] = new
3449 3255 self.hook(b'prepushkey', throw=True, **hookargs)
3450 3256 except error.HookAbort as exc:
3451 3257 self.ui.write_err(_(b"pushkey-abort: %s\n") % exc)
3452 3258 if exc.hint:
3453 3259 self.ui.write_err(_(b"(%s)\n") % exc.hint)
3454 3260 return False
3455 3261 self.ui.debug(b'pushing key for "%s:%s"\n' % (namespace, key))
3456 3262 ret = pushkey.push(self, namespace, key, old, new)
3457 3263
3458 3264 def runhook(unused_success):
3459 3265 self.hook(
3460 3266 b'pushkey',
3461 3267 namespace=namespace,
3462 3268 key=key,
3463 3269 old=old,
3464 3270 new=new,
3465 3271 ret=ret,
3466 3272 )
3467 3273
3468 3274 self._afterlock(runhook)
3469 3275 return ret
3470 3276
3471 3277 def listkeys(self, namespace):
3472 3278 self.hook(b'prelistkeys', throw=True, namespace=namespace)
3473 3279 self.ui.debug(b'listing keys for "%s"\n' % namespace)
3474 3280 values = pushkey.list(self, namespace)
3475 3281 self.hook(b'listkeys', namespace=namespace, values=values)
3476 3282 return values
3477 3283
3478 3284 def debugwireargs(self, one, two, three=None, four=None, five=None):
3479 3285 '''used to test argument passing over the wire'''
3480 3286 return b"%s %s %s %s %s" % (
3481 3287 one,
3482 3288 two,
3483 3289 pycompat.bytestr(three),
3484 3290 pycompat.bytestr(four),
3485 3291 pycompat.bytestr(five),
3486 3292 )
3487 3293
3488 3294 def savecommitmessage(self, text):
3489 3295 fp = self.vfs(b'last-message.txt', b'wb')
3490 3296 try:
3491 3297 fp.write(text)
3492 3298 finally:
3493 3299 fp.close()
3494 3300 return self.pathto(fp.name[len(self.root) + 1 :])
3495 3301
3496 3302
3497 3303 # used to avoid circular references so destructors work
3498 3304 def aftertrans(files):
3499 3305 renamefiles = [tuple(t) for t in files]
3500 3306
3501 3307 def a():
3502 3308 for vfs, src, dest in renamefiles:
3503 3309 # if src and dest refer to a same file, vfs.rename is a no-op,
3504 3310 # leaving both src and dest on disk. delete dest to make sure
3505 3311 # the rename couldn't be such a no-op.
3506 3312 vfs.tryunlink(dest)
3507 3313 try:
3508 3314 vfs.rename(src, dest)
3509 3315 except OSError: # journal file does not yet exist
3510 3316 pass
3511 3317
3512 3318 return a
3513 3319
3514 3320
3515 3321 def undoname(fn):
3516 3322 base, name = os.path.split(fn)
3517 3323 assert name.startswith(b'journal')
3518 3324 return os.path.join(base, name.replace(b'journal', b'undo', 1))
3519 3325
3520 3326
3521 3327 def instance(ui, path, create, intents=None, createopts=None):
3522 3328 localpath = util.urllocalpath(path)
3523 3329 if create:
3524 3330 createrepository(ui, localpath, createopts=createopts)
3525 3331
3526 3332 return makelocalrepository(ui, localpath, intents=intents)
3527 3333
3528 3334
3529 3335 def islocal(path):
3530 3336 return True
3531 3337
3532 3338
3533 3339 def defaultcreateopts(ui, createopts=None):
3534 3340 """Populate the default creation options for a repository.
3535 3341
3536 3342 A dictionary of explicitly requested creation options can be passed
3537 3343 in. Missing keys will be populated.
3538 3344 """
3539 3345 createopts = dict(createopts or {})
3540 3346
3541 3347 if b'backend' not in createopts:
3542 3348 # experimental config: storage.new-repo-backend
3543 3349 createopts[b'backend'] = ui.config(b'storage', b'new-repo-backend')
3544 3350
3545 3351 return createopts
3546 3352
3547 3353
3548 3354 def newreporequirements(ui, createopts):
3549 3355 """Determine the set of requirements for a new local repository.
3550 3356
3551 3357 Extensions can wrap this function to specify custom requirements for
3552 3358 new repositories.
3553 3359 """
3554 3360 # If the repo is being created from a shared repository, we copy
3555 3361 # its requirements.
3556 3362 if b'sharedrepo' in createopts:
3557 3363 requirements = set(createopts[b'sharedrepo'].requirements)
3558 3364 if createopts.get(b'sharedrelative'):
3559 3365 requirements.add(b'relshared')
3560 3366 else:
3561 3367 requirements.add(b'shared')
3562 3368
3563 3369 return requirements
3564 3370
3565 3371 if b'backend' not in createopts:
3566 3372 raise error.ProgrammingError(
3567 3373 b'backend key not present in createopts; '
3568 3374 b'was defaultcreateopts() called?'
3569 3375 )
3570 3376
3571 3377 if createopts[b'backend'] != b'revlogv1':
3572 3378 raise error.Abort(
3573 3379 _(
3574 3380 b'unable to determine repository requirements for '
3575 3381 b'storage backend: %s'
3576 3382 )
3577 3383 % createopts[b'backend']
3578 3384 )
3579 3385
3580 3386 requirements = {b'revlogv1'}
3581 3387 if ui.configbool(b'format', b'usestore'):
3582 3388 requirements.add(b'store')
3583 3389 if ui.configbool(b'format', b'usefncache'):
3584 3390 requirements.add(b'fncache')
3585 3391 if ui.configbool(b'format', b'dotencode'):
3586 3392 requirements.add(b'dotencode')
3587 3393
3588 3394 compengines = ui.configlist(b'format', b'revlog-compression')
3589 3395 for compengine in compengines:
3590 3396 if compengine in util.compengines:
3591 3397 break
3592 3398 else:
3593 3399 raise error.Abort(
3594 3400 _(
3595 3401 b'compression engines %s defined by '
3596 3402 b'format.revlog-compression not available'
3597 3403 )
3598 3404 % b', '.join(b'"%s"' % e for e in compengines),
3599 3405 hint=_(
3600 3406 b'run "hg debuginstall" to list available '
3601 3407 b'compression engines'
3602 3408 ),
3603 3409 )
3604 3410
3605 3411 # zlib is the historical default and doesn't need an explicit requirement.
3606 3412 if compengine == b'zstd':
3607 3413 requirements.add(b'revlog-compression-zstd')
3608 3414 elif compengine != b'zlib':
3609 3415 requirements.add(b'exp-compression-%s' % compengine)
3610 3416
3611 3417 if scmutil.gdinitconfig(ui):
3612 3418 requirements.add(b'generaldelta')
3613 3419 if ui.configbool(b'format', b'sparse-revlog'):
3614 3420 requirements.add(SPARSEREVLOG_REQUIREMENT)
3615 3421
3616 3422 # experimental config: format.exp-use-side-data
3617 3423 if ui.configbool(b'format', b'exp-use-side-data'):
3618 3424 requirements.add(SIDEDATA_REQUIREMENT)
3619 3425 # experimental config: format.exp-use-copies-side-data-changeset
3620 3426 if ui.configbool(b'format', b'exp-use-copies-side-data-changeset'):
3621 3427 requirements.add(SIDEDATA_REQUIREMENT)
3622 3428 requirements.add(COPIESSDC_REQUIREMENT)
3623 3429 if ui.configbool(b'experimental', b'treemanifest'):
3624 3430 requirements.add(b'treemanifest')
3625 3431
3626 3432 revlogv2 = ui.config(b'experimental', b'revlogv2')
3627 3433 if revlogv2 == b'enable-unstable-format-and-corrupt-my-data':
3628 3434 requirements.remove(b'revlogv1')
3629 3435 # generaldelta is implied by revlogv2.
3630 3436 requirements.discard(b'generaldelta')
3631 3437 requirements.add(REVLOGV2_REQUIREMENT)
3632 3438 # experimental config: format.internal-phase
3633 3439 if ui.configbool(b'format', b'internal-phase'):
3634 3440 requirements.add(b'internal-phase')
3635 3441
3636 3442 if createopts.get(b'narrowfiles'):
3637 3443 requirements.add(repository.NARROW_REQUIREMENT)
3638 3444
3639 3445 if createopts.get(b'lfs'):
3640 3446 requirements.add(b'lfs')
3641 3447
3642 3448 if ui.configbool(b'format', b'bookmarks-in-store'):
3643 3449 requirements.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3644 3450
3645 3451 if ui.configbool(b'format', b'use-persistent-nodemap'):
3646 3452 requirements.add(NODEMAP_REQUIREMENT)
3647 3453
3648 3454 return requirements
3649 3455
3650 3456
3651 3457 def filterknowncreateopts(ui, createopts):
3652 3458 """Filters a dict of repo creation options against options that are known.
3653 3459
3654 3460 Receives a dict of repo creation options and returns a dict of those
3655 3461 options that we don't know how to handle.
3656 3462
3657 3463 This function is called as part of repository creation. If the
3658 3464 returned dict contains any items, repository creation will not
3659 3465 be allowed, as it means there was a request to create a repository
3660 3466 with options not recognized by loaded code.
3661 3467
3662 3468 Extensions can wrap this function to filter out creation options
3663 3469 they know how to handle.
3664 3470 """
3665 3471 known = {
3666 3472 b'backend',
3667 3473 b'lfs',
3668 3474 b'narrowfiles',
3669 3475 b'sharedrepo',
3670 3476 b'sharedrelative',
3671 3477 b'shareditems',
3672 3478 b'shallowfilestore',
3673 3479 }
3674 3480
3675 3481 return {k: v for k, v in createopts.items() if k not in known}
3676 3482
3677 3483
3678 3484 def createrepository(ui, path, createopts=None):
3679 3485 """Create a new repository in a vfs.
3680 3486
3681 3487 ``path`` path to the new repo's working directory.
3682 3488 ``createopts`` options for the new repository.
3683 3489
3684 3490 The following keys for ``createopts`` are recognized:
3685 3491
3686 3492 backend
3687 3493 The storage backend to use.
3688 3494 lfs
3689 3495 Repository will be created with ``lfs`` requirement. The lfs extension
3690 3496 will automatically be loaded when the repository is accessed.
3691 3497 narrowfiles
3692 3498 Set up repository to support narrow file storage.
3693 3499 sharedrepo
3694 3500 Repository object from which storage should be shared.
3695 3501 sharedrelative
3696 3502 Boolean indicating if the path to the shared repo should be
3697 3503 stored as relative. By default, the pointer to the "parent" repo
3698 3504 is stored as an absolute path.
3699 3505 shareditems
3700 3506 Set of items to share to the new repository (in addition to storage).
3701 3507 shallowfilestore
3702 3508 Indicates that storage for files should be shallow (not all ancestor
3703 3509 revisions are known).
3704 3510 """
3705 3511 createopts = defaultcreateopts(ui, createopts=createopts)
3706 3512
3707 3513 unknownopts = filterknowncreateopts(ui, createopts)
3708 3514
3709 3515 if not isinstance(unknownopts, dict):
3710 3516 raise error.ProgrammingError(
3711 3517 b'filterknowncreateopts() did not return a dict'
3712 3518 )
3713 3519
3714 3520 if unknownopts:
3715 3521 raise error.Abort(
3716 3522 _(
3717 3523 b'unable to create repository because of unknown '
3718 3524 b'creation option: %s'
3719 3525 )
3720 3526 % b', '.join(sorted(unknownopts)),
3721 3527 hint=_(b'is a required extension not loaded?'),
3722 3528 )
3723 3529
3724 3530 requirements = newreporequirements(ui, createopts=createopts)
3725 3531
3726 3532 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
3727 3533
3728 3534 hgvfs = vfsmod.vfs(wdirvfs.join(b'.hg'))
3729 3535 if hgvfs.exists():
3730 3536 raise error.RepoError(_(b'repository %s already exists') % path)
3731 3537
3732 3538 if b'sharedrepo' in createopts:
3733 3539 sharedpath = createopts[b'sharedrepo'].sharedpath
3734 3540
3735 3541 if createopts.get(b'sharedrelative'):
3736 3542 try:
3737 3543 sharedpath = os.path.relpath(sharedpath, hgvfs.base)
3738 3544 except (IOError, ValueError) as e:
3739 3545 # ValueError is raised on Windows if the drive letters differ
3740 3546 # on each path.
3741 3547 raise error.Abort(
3742 3548 _(b'cannot calculate relative path'),
3743 3549 hint=stringutil.forcebytestr(e),
3744 3550 )
3745 3551
3746 3552 if not wdirvfs.exists():
3747 3553 wdirvfs.makedirs()
3748 3554
3749 3555 hgvfs.makedir(notindexed=True)
3750 3556 if b'sharedrepo' not in createopts:
3751 3557 hgvfs.mkdir(b'cache')
3752 3558 hgvfs.mkdir(b'wcache')
3753 3559
3754 3560 if b'store' in requirements and b'sharedrepo' not in createopts:
3755 3561 hgvfs.mkdir(b'store')
3756 3562
3757 3563 # We create an invalid changelog outside the store so very old
3758 3564 # Mercurial versions (which didn't know about the requirements
3759 3565 # file) encounter an error on reading the changelog. This
3760 3566 # effectively locks out old clients and prevents them from
3761 3567 # mucking with a repo in an unknown format.
3762 3568 #
3763 3569 # The revlog header has version 2, which won't be recognized by
3764 3570 # such old clients.
3765 3571 hgvfs.append(
3766 3572 b'00changelog.i',
3767 3573 b'\0\0\0\2 dummy changelog to prevent using the old repo '
3768 3574 b'layout',
3769 3575 )
3770 3576
3771 3577 scmutil.writerequires(hgvfs, requirements)
3772 3578
3773 3579 # Write out file telling readers where to find the shared store.
3774 3580 if b'sharedrepo' in createopts:
3775 3581 hgvfs.write(b'sharedpath', sharedpath)
3776 3582
3777 3583 if createopts.get(b'shareditems'):
3778 3584 shared = b'\n'.join(sorted(createopts[b'shareditems'])) + b'\n'
3779 3585 hgvfs.write(b'shared', shared)
3780 3586
3781 3587
3782 3588 def poisonrepository(repo):
3783 3589 """Poison a repository instance so it can no longer be used."""
3784 3590 # Perform any cleanup on the instance.
3785 3591 repo.close()
3786 3592
3787 3593 # Our strategy is to replace the type of the object with one that
3788 3594 # has all attribute lookups result in error.
3789 3595 #
3790 3596 # But we have to allow the close() method because some constructors
3791 3597 # of repos call close() on repo references.
3792 3598 class poisonedrepository(object):
3793 3599 def __getattribute__(self, item):
3794 3600 if item == 'close':
3795 3601 return object.__getattribute__(self, item)
3796 3602
3797 3603 raise error.ProgrammingError(
3798 3604 b'repo instances should not be used after unshare'
3799 3605 )
3800 3606
3801 3607 def close(self):
3802 3608 pass
3803 3609
3804 3610 # We may have a repoview, which intercepts __setattr__. So be sure
3805 3611 # we operate at the lowest level possible.
3806 3612 object.__setattr__(repo, '__class__', poisonedrepository)
General Comments 0
You need to be logged in to leave comments. Login now