##// END OF EJS Templates
localrepo: refactor logic to calculate sharedvfs in separate fn...
Pulkit Goyal -
r45912:665e9115 default
parent child Browse files
Show More
@@ -1,3504 +1,3521 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7
7
8 from __future__ import absolute_import
8 from __future__ import absolute_import
9
9
10 import errno
10 import errno
11 import os
11 import os
12 import random
12 import random
13 import sys
13 import sys
14 import time
14 import time
15 import weakref
15 import weakref
16
16
17 from .i18n import _
17 from .i18n import _
18 from .node import (
18 from .node import (
19 bin,
19 bin,
20 hex,
20 hex,
21 nullid,
21 nullid,
22 nullrev,
22 nullrev,
23 short,
23 short,
24 )
24 )
25 from .pycompat import (
25 from .pycompat import (
26 delattr,
26 delattr,
27 getattr,
27 getattr,
28 )
28 )
29 from . import (
29 from . import (
30 bookmarks,
30 bookmarks,
31 branchmap,
31 branchmap,
32 bundle2,
32 bundle2,
33 changegroup,
33 changegroup,
34 color,
34 color,
35 commit,
35 commit,
36 context,
36 context,
37 dirstate,
37 dirstate,
38 dirstateguard,
38 dirstateguard,
39 discovery,
39 discovery,
40 encoding,
40 encoding,
41 error,
41 error,
42 exchange,
42 exchange,
43 extensions,
43 extensions,
44 filelog,
44 filelog,
45 hook,
45 hook,
46 lock as lockmod,
46 lock as lockmod,
47 match as matchmod,
47 match as matchmod,
48 mergestate as mergestatemod,
48 mergestate as mergestatemod,
49 mergeutil,
49 mergeutil,
50 namespaces,
50 namespaces,
51 narrowspec,
51 narrowspec,
52 obsolete,
52 obsolete,
53 pathutil,
53 pathutil,
54 phases,
54 phases,
55 pushkey,
55 pushkey,
56 pycompat,
56 pycompat,
57 rcutil,
57 rcutil,
58 repoview,
58 repoview,
59 revset,
59 revset,
60 revsetlang,
60 revsetlang,
61 scmutil,
61 scmutil,
62 sparse,
62 sparse,
63 store as storemod,
63 store as storemod,
64 subrepoutil,
64 subrepoutil,
65 tags as tagsmod,
65 tags as tagsmod,
66 transaction,
66 transaction,
67 txnutil,
67 txnutil,
68 util,
68 util,
69 vfs as vfsmod,
69 vfs as vfsmod,
70 )
70 )
71
71
72 from .interfaces import (
72 from .interfaces import (
73 repository,
73 repository,
74 util as interfaceutil,
74 util as interfaceutil,
75 )
75 )
76
76
77 from .utils import (
77 from .utils import (
78 hashutil,
78 hashutil,
79 procutil,
79 procutil,
80 stringutil,
80 stringutil,
81 )
81 )
82
82
83 from .revlogutils import constants as revlogconst
83 from .revlogutils import constants as revlogconst
84
84
85 release = lockmod.release
85 release = lockmod.release
86 urlerr = util.urlerr
86 urlerr = util.urlerr
87 urlreq = util.urlreq
87 urlreq = util.urlreq
88
88
89 # set of (path, vfs-location) tuples. vfs-location is:
89 # set of (path, vfs-location) tuples. vfs-location is:
90 # - 'plain for vfs relative paths
90 # - 'plain for vfs relative paths
91 # - '' for svfs relative paths
91 # - '' for svfs relative paths
92 _cachedfiles = set()
92 _cachedfiles = set()
93
93
94
94
95 class _basefilecache(scmutil.filecache):
95 class _basefilecache(scmutil.filecache):
96 """All filecache usage on repo are done for logic that should be unfiltered
96 """All filecache usage on repo are done for logic that should be unfiltered
97 """
97 """
98
98
99 def __get__(self, repo, type=None):
99 def __get__(self, repo, type=None):
100 if repo is None:
100 if repo is None:
101 return self
101 return self
102 # proxy to unfiltered __dict__ since filtered repo has no entry
102 # proxy to unfiltered __dict__ since filtered repo has no entry
103 unfi = repo.unfiltered()
103 unfi = repo.unfiltered()
104 try:
104 try:
105 return unfi.__dict__[self.sname]
105 return unfi.__dict__[self.sname]
106 except KeyError:
106 except KeyError:
107 pass
107 pass
108 return super(_basefilecache, self).__get__(unfi, type)
108 return super(_basefilecache, self).__get__(unfi, type)
109
109
110 def set(self, repo, value):
110 def set(self, repo, value):
111 return super(_basefilecache, self).set(repo.unfiltered(), value)
111 return super(_basefilecache, self).set(repo.unfiltered(), value)
112
112
113
113
114 class repofilecache(_basefilecache):
114 class repofilecache(_basefilecache):
115 """filecache for files in .hg but outside of .hg/store"""
115 """filecache for files in .hg but outside of .hg/store"""
116
116
117 def __init__(self, *paths):
117 def __init__(self, *paths):
118 super(repofilecache, self).__init__(*paths)
118 super(repofilecache, self).__init__(*paths)
119 for path in paths:
119 for path in paths:
120 _cachedfiles.add((path, b'plain'))
120 _cachedfiles.add((path, b'plain'))
121
121
122 def join(self, obj, fname):
122 def join(self, obj, fname):
123 return obj.vfs.join(fname)
123 return obj.vfs.join(fname)
124
124
125
125
126 class storecache(_basefilecache):
126 class storecache(_basefilecache):
127 """filecache for files in the store"""
127 """filecache for files in the store"""
128
128
129 def __init__(self, *paths):
129 def __init__(self, *paths):
130 super(storecache, self).__init__(*paths)
130 super(storecache, self).__init__(*paths)
131 for path in paths:
131 for path in paths:
132 _cachedfiles.add((path, b''))
132 _cachedfiles.add((path, b''))
133
133
134 def join(self, obj, fname):
134 def join(self, obj, fname):
135 return obj.sjoin(fname)
135 return obj.sjoin(fname)
136
136
137
137
138 class mixedrepostorecache(_basefilecache):
138 class mixedrepostorecache(_basefilecache):
139 """filecache for a mix files in .hg/store and outside"""
139 """filecache for a mix files in .hg/store and outside"""
140
140
141 def __init__(self, *pathsandlocations):
141 def __init__(self, *pathsandlocations):
142 # scmutil.filecache only uses the path for passing back into our
142 # scmutil.filecache only uses the path for passing back into our
143 # join(), so we can safely pass a list of paths and locations
143 # join(), so we can safely pass a list of paths and locations
144 super(mixedrepostorecache, self).__init__(*pathsandlocations)
144 super(mixedrepostorecache, self).__init__(*pathsandlocations)
145 _cachedfiles.update(pathsandlocations)
145 _cachedfiles.update(pathsandlocations)
146
146
147 def join(self, obj, fnameandlocation):
147 def join(self, obj, fnameandlocation):
148 fname, location = fnameandlocation
148 fname, location = fnameandlocation
149 if location == b'plain':
149 if location == b'plain':
150 return obj.vfs.join(fname)
150 return obj.vfs.join(fname)
151 else:
151 else:
152 if location != b'':
152 if location != b'':
153 raise error.ProgrammingError(
153 raise error.ProgrammingError(
154 b'unexpected location: %s' % location
154 b'unexpected location: %s' % location
155 )
155 )
156 return obj.sjoin(fname)
156 return obj.sjoin(fname)
157
157
158
158
159 def isfilecached(repo, name):
159 def isfilecached(repo, name):
160 """check if a repo has already cached "name" filecache-ed property
160 """check if a repo has already cached "name" filecache-ed property
161
161
162 This returns (cachedobj-or-None, iscached) tuple.
162 This returns (cachedobj-or-None, iscached) tuple.
163 """
163 """
164 cacheentry = repo.unfiltered()._filecache.get(name, None)
164 cacheentry = repo.unfiltered()._filecache.get(name, None)
165 if not cacheentry:
165 if not cacheentry:
166 return None, False
166 return None, False
167 return cacheentry.obj, True
167 return cacheentry.obj, True
168
168
169
169
170 class unfilteredpropertycache(util.propertycache):
170 class unfilteredpropertycache(util.propertycache):
171 """propertycache that apply to unfiltered repo only"""
171 """propertycache that apply to unfiltered repo only"""
172
172
173 def __get__(self, repo, type=None):
173 def __get__(self, repo, type=None):
174 unfi = repo.unfiltered()
174 unfi = repo.unfiltered()
175 if unfi is repo:
175 if unfi is repo:
176 return super(unfilteredpropertycache, self).__get__(unfi)
176 return super(unfilteredpropertycache, self).__get__(unfi)
177 return getattr(unfi, self.name)
177 return getattr(unfi, self.name)
178
178
179
179
180 class filteredpropertycache(util.propertycache):
180 class filteredpropertycache(util.propertycache):
181 """propertycache that must take filtering in account"""
181 """propertycache that must take filtering in account"""
182
182
183 def cachevalue(self, obj, value):
183 def cachevalue(self, obj, value):
184 object.__setattr__(obj, self.name, value)
184 object.__setattr__(obj, self.name, value)
185
185
186
186
187 def hasunfilteredcache(repo, name):
187 def hasunfilteredcache(repo, name):
188 """check if a repo has an unfilteredpropertycache value for <name>"""
188 """check if a repo has an unfilteredpropertycache value for <name>"""
189 return name in vars(repo.unfiltered())
189 return name in vars(repo.unfiltered())
190
190
191
191
192 def unfilteredmethod(orig):
192 def unfilteredmethod(orig):
193 """decorate method that always need to be run on unfiltered version"""
193 """decorate method that always need to be run on unfiltered version"""
194
194
195 def wrapper(repo, *args, **kwargs):
195 def wrapper(repo, *args, **kwargs):
196 return orig(repo.unfiltered(), *args, **kwargs)
196 return orig(repo.unfiltered(), *args, **kwargs)
197
197
198 return wrapper
198 return wrapper
199
199
200
200
201 moderncaps = {
201 moderncaps = {
202 b'lookup',
202 b'lookup',
203 b'branchmap',
203 b'branchmap',
204 b'pushkey',
204 b'pushkey',
205 b'known',
205 b'known',
206 b'getbundle',
206 b'getbundle',
207 b'unbundle',
207 b'unbundle',
208 }
208 }
209 legacycaps = moderncaps.union({b'changegroupsubset'})
209 legacycaps = moderncaps.union({b'changegroupsubset'})
210
210
211
211
212 @interfaceutil.implementer(repository.ipeercommandexecutor)
212 @interfaceutil.implementer(repository.ipeercommandexecutor)
213 class localcommandexecutor(object):
213 class localcommandexecutor(object):
214 def __init__(self, peer):
214 def __init__(self, peer):
215 self._peer = peer
215 self._peer = peer
216 self._sent = False
216 self._sent = False
217 self._closed = False
217 self._closed = False
218
218
219 def __enter__(self):
219 def __enter__(self):
220 return self
220 return self
221
221
222 def __exit__(self, exctype, excvalue, exctb):
222 def __exit__(self, exctype, excvalue, exctb):
223 self.close()
223 self.close()
224
224
225 def callcommand(self, command, args):
225 def callcommand(self, command, args):
226 if self._sent:
226 if self._sent:
227 raise error.ProgrammingError(
227 raise error.ProgrammingError(
228 b'callcommand() cannot be used after sendcommands()'
228 b'callcommand() cannot be used after sendcommands()'
229 )
229 )
230
230
231 if self._closed:
231 if self._closed:
232 raise error.ProgrammingError(
232 raise error.ProgrammingError(
233 b'callcommand() cannot be used after close()'
233 b'callcommand() cannot be used after close()'
234 )
234 )
235
235
236 # We don't need to support anything fancy. Just call the named
236 # We don't need to support anything fancy. Just call the named
237 # method on the peer and return a resolved future.
237 # method on the peer and return a resolved future.
238 fn = getattr(self._peer, pycompat.sysstr(command))
238 fn = getattr(self._peer, pycompat.sysstr(command))
239
239
240 f = pycompat.futures.Future()
240 f = pycompat.futures.Future()
241
241
242 try:
242 try:
243 result = fn(**pycompat.strkwargs(args))
243 result = fn(**pycompat.strkwargs(args))
244 except Exception:
244 except Exception:
245 pycompat.future_set_exception_info(f, sys.exc_info()[1:])
245 pycompat.future_set_exception_info(f, sys.exc_info()[1:])
246 else:
246 else:
247 f.set_result(result)
247 f.set_result(result)
248
248
249 return f
249 return f
250
250
251 def sendcommands(self):
251 def sendcommands(self):
252 self._sent = True
252 self._sent = True
253
253
254 def close(self):
254 def close(self):
255 self._closed = True
255 self._closed = True
256
256
257
257
258 @interfaceutil.implementer(repository.ipeercommands)
258 @interfaceutil.implementer(repository.ipeercommands)
259 class localpeer(repository.peer):
259 class localpeer(repository.peer):
260 '''peer for a local repo; reflects only the most recent API'''
260 '''peer for a local repo; reflects only the most recent API'''
261
261
262 def __init__(self, repo, caps=None):
262 def __init__(self, repo, caps=None):
263 super(localpeer, self).__init__()
263 super(localpeer, self).__init__()
264
264
265 if caps is None:
265 if caps is None:
266 caps = moderncaps.copy()
266 caps = moderncaps.copy()
267 self._repo = repo.filtered(b'served')
267 self._repo = repo.filtered(b'served')
268 self.ui = repo.ui
268 self.ui = repo.ui
269 self._caps = repo._restrictcapabilities(caps)
269 self._caps = repo._restrictcapabilities(caps)
270
270
271 # Begin of _basepeer interface.
271 # Begin of _basepeer interface.
272
272
273 def url(self):
273 def url(self):
274 return self._repo.url()
274 return self._repo.url()
275
275
276 def local(self):
276 def local(self):
277 return self._repo
277 return self._repo
278
278
279 def peer(self):
279 def peer(self):
280 return self
280 return self
281
281
282 def canpush(self):
282 def canpush(self):
283 return True
283 return True
284
284
285 def close(self):
285 def close(self):
286 self._repo.close()
286 self._repo.close()
287
287
288 # End of _basepeer interface.
288 # End of _basepeer interface.
289
289
290 # Begin of _basewirecommands interface.
290 # Begin of _basewirecommands interface.
291
291
292 def branchmap(self):
292 def branchmap(self):
293 return self._repo.branchmap()
293 return self._repo.branchmap()
294
294
295 def capabilities(self):
295 def capabilities(self):
296 return self._caps
296 return self._caps
297
297
298 def clonebundles(self):
298 def clonebundles(self):
299 return self._repo.tryread(b'clonebundles.manifest')
299 return self._repo.tryread(b'clonebundles.manifest')
300
300
301 def debugwireargs(self, one, two, three=None, four=None, five=None):
301 def debugwireargs(self, one, two, three=None, four=None, five=None):
302 """Used to test argument passing over the wire"""
302 """Used to test argument passing over the wire"""
303 return b"%s %s %s %s %s" % (
303 return b"%s %s %s %s %s" % (
304 one,
304 one,
305 two,
305 two,
306 pycompat.bytestr(three),
306 pycompat.bytestr(three),
307 pycompat.bytestr(four),
307 pycompat.bytestr(four),
308 pycompat.bytestr(five),
308 pycompat.bytestr(five),
309 )
309 )
310
310
311 def getbundle(
311 def getbundle(
312 self, source, heads=None, common=None, bundlecaps=None, **kwargs
312 self, source, heads=None, common=None, bundlecaps=None, **kwargs
313 ):
313 ):
314 chunks = exchange.getbundlechunks(
314 chunks = exchange.getbundlechunks(
315 self._repo,
315 self._repo,
316 source,
316 source,
317 heads=heads,
317 heads=heads,
318 common=common,
318 common=common,
319 bundlecaps=bundlecaps,
319 bundlecaps=bundlecaps,
320 **kwargs
320 **kwargs
321 )[1]
321 )[1]
322 cb = util.chunkbuffer(chunks)
322 cb = util.chunkbuffer(chunks)
323
323
324 if exchange.bundle2requested(bundlecaps):
324 if exchange.bundle2requested(bundlecaps):
325 # When requesting a bundle2, getbundle returns a stream to make the
325 # When requesting a bundle2, getbundle returns a stream to make the
326 # wire level function happier. We need to build a proper object
326 # wire level function happier. We need to build a proper object
327 # from it in local peer.
327 # from it in local peer.
328 return bundle2.getunbundler(self.ui, cb)
328 return bundle2.getunbundler(self.ui, cb)
329 else:
329 else:
330 return changegroup.getunbundler(b'01', cb, None)
330 return changegroup.getunbundler(b'01', cb, None)
331
331
332 def heads(self):
332 def heads(self):
333 return self._repo.heads()
333 return self._repo.heads()
334
334
335 def known(self, nodes):
335 def known(self, nodes):
336 return self._repo.known(nodes)
336 return self._repo.known(nodes)
337
337
338 def listkeys(self, namespace):
338 def listkeys(self, namespace):
339 return self._repo.listkeys(namespace)
339 return self._repo.listkeys(namespace)
340
340
341 def lookup(self, key):
341 def lookup(self, key):
342 return self._repo.lookup(key)
342 return self._repo.lookup(key)
343
343
344 def pushkey(self, namespace, key, old, new):
344 def pushkey(self, namespace, key, old, new):
345 return self._repo.pushkey(namespace, key, old, new)
345 return self._repo.pushkey(namespace, key, old, new)
346
346
347 def stream_out(self):
347 def stream_out(self):
348 raise error.Abort(_(b'cannot perform stream clone against local peer'))
348 raise error.Abort(_(b'cannot perform stream clone against local peer'))
349
349
350 def unbundle(self, bundle, heads, url):
350 def unbundle(self, bundle, heads, url):
351 """apply a bundle on a repo
351 """apply a bundle on a repo
352
352
353 This function handles the repo locking itself."""
353 This function handles the repo locking itself."""
354 try:
354 try:
355 try:
355 try:
356 bundle = exchange.readbundle(self.ui, bundle, None)
356 bundle = exchange.readbundle(self.ui, bundle, None)
357 ret = exchange.unbundle(self._repo, bundle, heads, b'push', url)
357 ret = exchange.unbundle(self._repo, bundle, heads, b'push', url)
358 if util.safehasattr(ret, b'getchunks'):
358 if util.safehasattr(ret, b'getchunks'):
359 # This is a bundle20 object, turn it into an unbundler.
359 # This is a bundle20 object, turn it into an unbundler.
360 # This little dance should be dropped eventually when the
360 # This little dance should be dropped eventually when the
361 # API is finally improved.
361 # API is finally improved.
362 stream = util.chunkbuffer(ret.getchunks())
362 stream = util.chunkbuffer(ret.getchunks())
363 ret = bundle2.getunbundler(self.ui, stream)
363 ret = bundle2.getunbundler(self.ui, stream)
364 return ret
364 return ret
365 except Exception as exc:
365 except Exception as exc:
366 # If the exception contains output salvaged from a bundle2
366 # If the exception contains output salvaged from a bundle2
367 # reply, we need to make sure it is printed before continuing
367 # reply, we need to make sure it is printed before continuing
368 # to fail. So we build a bundle2 with such output and consume
368 # to fail. So we build a bundle2 with such output and consume
369 # it directly.
369 # it directly.
370 #
370 #
371 # This is not very elegant but allows a "simple" solution for
371 # This is not very elegant but allows a "simple" solution for
372 # issue4594
372 # issue4594
373 output = getattr(exc, '_bundle2salvagedoutput', ())
373 output = getattr(exc, '_bundle2salvagedoutput', ())
374 if output:
374 if output:
375 bundler = bundle2.bundle20(self._repo.ui)
375 bundler = bundle2.bundle20(self._repo.ui)
376 for out in output:
376 for out in output:
377 bundler.addpart(out)
377 bundler.addpart(out)
378 stream = util.chunkbuffer(bundler.getchunks())
378 stream = util.chunkbuffer(bundler.getchunks())
379 b = bundle2.getunbundler(self.ui, stream)
379 b = bundle2.getunbundler(self.ui, stream)
380 bundle2.processbundle(self._repo, b)
380 bundle2.processbundle(self._repo, b)
381 raise
381 raise
382 except error.PushRaced as exc:
382 except error.PushRaced as exc:
383 raise error.ResponseError(
383 raise error.ResponseError(
384 _(b'push failed:'), stringutil.forcebytestr(exc)
384 _(b'push failed:'), stringutil.forcebytestr(exc)
385 )
385 )
386
386
387 # End of _basewirecommands interface.
387 # End of _basewirecommands interface.
388
388
389 # Begin of peer interface.
389 # Begin of peer interface.
390
390
391 def commandexecutor(self):
391 def commandexecutor(self):
392 return localcommandexecutor(self)
392 return localcommandexecutor(self)
393
393
394 # End of peer interface.
394 # End of peer interface.
395
395
396
396
397 @interfaceutil.implementer(repository.ipeerlegacycommands)
397 @interfaceutil.implementer(repository.ipeerlegacycommands)
398 class locallegacypeer(localpeer):
398 class locallegacypeer(localpeer):
399 '''peer extension which implements legacy methods too; used for tests with
399 '''peer extension which implements legacy methods too; used for tests with
400 restricted capabilities'''
400 restricted capabilities'''
401
401
402 def __init__(self, repo):
402 def __init__(self, repo):
403 super(locallegacypeer, self).__init__(repo, caps=legacycaps)
403 super(locallegacypeer, self).__init__(repo, caps=legacycaps)
404
404
405 # Begin of baselegacywirecommands interface.
405 # Begin of baselegacywirecommands interface.
406
406
407 def between(self, pairs):
407 def between(self, pairs):
408 return self._repo.between(pairs)
408 return self._repo.between(pairs)
409
409
410 def branches(self, nodes):
410 def branches(self, nodes):
411 return self._repo.branches(nodes)
411 return self._repo.branches(nodes)
412
412
413 def changegroup(self, nodes, source):
413 def changegroup(self, nodes, source):
414 outgoing = discovery.outgoing(
414 outgoing = discovery.outgoing(
415 self._repo, missingroots=nodes, ancestorsof=self._repo.heads()
415 self._repo, missingroots=nodes, ancestorsof=self._repo.heads()
416 )
416 )
417 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
417 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
418
418
419 def changegroupsubset(self, bases, heads, source):
419 def changegroupsubset(self, bases, heads, source):
420 outgoing = discovery.outgoing(
420 outgoing = discovery.outgoing(
421 self._repo, missingroots=bases, ancestorsof=heads
421 self._repo, missingroots=bases, ancestorsof=heads
422 )
422 )
423 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
423 return changegroup.makechangegroup(self._repo, outgoing, b'01', source)
424
424
425 # End of baselegacywirecommands interface.
425 # End of baselegacywirecommands interface.
426
426
427
427
428 # Increment the sub-version when the revlog v2 format changes to lock out old
428 # Increment the sub-version when the revlog v2 format changes to lock out old
429 # clients.
429 # clients.
430 REVLOGV2_REQUIREMENT = b'exp-revlogv2.1'
430 REVLOGV2_REQUIREMENT = b'exp-revlogv2.1'
431
431
432 # A repository with the sparserevlog feature will have delta chains that
432 # A repository with the sparserevlog feature will have delta chains that
433 # can spread over a larger span. Sparse reading cuts these large spans into
433 # can spread over a larger span. Sparse reading cuts these large spans into
434 # pieces, so that each piece isn't too big.
434 # pieces, so that each piece isn't too big.
435 # Without the sparserevlog capability, reading from the repository could use
435 # Without the sparserevlog capability, reading from the repository could use
436 # huge amounts of memory, because the whole span would be read at once,
436 # huge amounts of memory, because the whole span would be read at once,
437 # including all the intermediate revisions that aren't pertinent for the chain.
437 # including all the intermediate revisions that aren't pertinent for the chain.
438 # This is why once a repository has enabled sparse-read, it becomes required.
438 # This is why once a repository has enabled sparse-read, it becomes required.
439 SPARSEREVLOG_REQUIREMENT = b'sparserevlog'
439 SPARSEREVLOG_REQUIREMENT = b'sparserevlog'
440
440
441 # A repository with the sidedataflag requirement will allow to store extra
441 # A repository with the sidedataflag requirement will allow to store extra
442 # information for revision without altering their original hashes.
442 # information for revision without altering their original hashes.
443 SIDEDATA_REQUIREMENT = b'exp-sidedata-flag'
443 SIDEDATA_REQUIREMENT = b'exp-sidedata-flag'
444
444
445 # A repository with the the copies-sidedata-changeset requirement will store
445 # A repository with the the copies-sidedata-changeset requirement will store
446 # copies related information in changeset's sidedata.
446 # copies related information in changeset's sidedata.
447 COPIESSDC_REQUIREMENT = b'exp-copies-sidedata-changeset'
447 COPIESSDC_REQUIREMENT = b'exp-copies-sidedata-changeset'
448
448
449 # The repository use persistent nodemap for the changelog and the manifest.
449 # The repository use persistent nodemap for the changelog and the manifest.
450 NODEMAP_REQUIREMENT = b'persistent-nodemap'
450 NODEMAP_REQUIREMENT = b'persistent-nodemap'
451
451
452 # Functions receiving (ui, features) that extensions can register to impact
452 # Functions receiving (ui, features) that extensions can register to impact
453 # the ability to load repositories with custom requirements. Only
453 # the ability to load repositories with custom requirements. Only
454 # functions defined in loaded extensions are called.
454 # functions defined in loaded extensions are called.
455 #
455 #
456 # The function receives a set of requirement strings that the repository
456 # The function receives a set of requirement strings that the repository
457 # is capable of opening. Functions will typically add elements to the
457 # is capable of opening. Functions will typically add elements to the
458 # set to reflect that the extension knows how to handle that requirements.
458 # set to reflect that the extension knows how to handle that requirements.
459 featuresetupfuncs = set()
459 featuresetupfuncs = set()
460
460
461
461
462 def _getsharedvfs(hgvfs, requirements):
463 """ returns the vfs object pointing to root of shared source
464 repo for a shared repository
465
466 hgvfs is vfs pointing at .hg/ of current repo (shared one)
467 requirements is a set of requirements of current repo (shared one)
468 """
469 # The ``shared`` or ``relshared`` requirements indicate the
470 # store lives in the path contained in the ``.hg/sharedpath`` file.
471 # This is an absolute path for ``shared`` and relative to
472 # ``.hg/`` for ``relshared``.
473 sharedpath = hgvfs.read(b'sharedpath').rstrip(b'\n')
474 if b'relshared' in requirements:
475 sharedpath = hgvfs.join(sharedpath)
476
477 sharedvfs = vfsmod.vfs(sharedpath, realpath=True)
478
479 if not sharedvfs.exists():
480 raise error.RepoError(
481 _(b'.hg/sharedpath points to nonexistent directory %s')
482 % sharedvfs.base
483 )
484 return sharedvfs
485
486
462 def makelocalrepository(baseui, path, intents=None):
487 def makelocalrepository(baseui, path, intents=None):
463 """Create a local repository object.
488 """Create a local repository object.
464
489
465 Given arguments needed to construct a local repository, this function
490 Given arguments needed to construct a local repository, this function
466 performs various early repository loading functionality (such as
491 performs various early repository loading functionality (such as
467 reading the ``.hg/requires`` and ``.hg/hgrc`` files), validates that
492 reading the ``.hg/requires`` and ``.hg/hgrc`` files), validates that
468 the repository can be opened, derives a type suitable for representing
493 the repository can be opened, derives a type suitable for representing
469 that repository, and returns an instance of it.
494 that repository, and returns an instance of it.
470
495
471 The returned object conforms to the ``repository.completelocalrepository``
496 The returned object conforms to the ``repository.completelocalrepository``
472 interface.
497 interface.
473
498
474 The repository type is derived by calling a series of factory functions
499 The repository type is derived by calling a series of factory functions
475 for each aspect/interface of the final repository. These are defined by
500 for each aspect/interface of the final repository. These are defined by
476 ``REPO_INTERFACES``.
501 ``REPO_INTERFACES``.
477
502
478 Each factory function is called to produce a type implementing a specific
503 Each factory function is called to produce a type implementing a specific
479 interface. The cumulative list of returned types will be combined into a
504 interface. The cumulative list of returned types will be combined into a
480 new type and that type will be instantiated to represent the local
505 new type and that type will be instantiated to represent the local
481 repository.
506 repository.
482
507
483 The factory functions each receive various state that may be consulted
508 The factory functions each receive various state that may be consulted
484 as part of deriving a type.
509 as part of deriving a type.
485
510
486 Extensions should wrap these factory functions to customize repository type
511 Extensions should wrap these factory functions to customize repository type
487 creation. Note that an extension's wrapped function may be called even if
512 creation. Note that an extension's wrapped function may be called even if
488 that extension is not loaded for the repo being constructed. Extensions
513 that extension is not loaded for the repo being constructed. Extensions
489 should check if their ``__name__`` appears in the
514 should check if their ``__name__`` appears in the
490 ``extensionmodulenames`` set passed to the factory function and no-op if
515 ``extensionmodulenames`` set passed to the factory function and no-op if
491 not.
516 not.
492 """
517 """
493 ui = baseui.copy()
518 ui = baseui.copy()
494 # Prevent copying repo configuration.
519 # Prevent copying repo configuration.
495 ui.copy = baseui.copy
520 ui.copy = baseui.copy
496
521
497 # Working directory VFS rooted at repository root.
522 # Working directory VFS rooted at repository root.
498 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
523 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
499
524
500 # Main VFS for .hg/ directory.
525 # Main VFS for .hg/ directory.
501 hgpath = wdirvfs.join(b'.hg')
526 hgpath = wdirvfs.join(b'.hg')
502 hgvfs = vfsmod.vfs(hgpath, cacheaudited=True)
527 hgvfs = vfsmod.vfs(hgpath, cacheaudited=True)
528 # Whether this repository is shared one or not
529 shared = False
530 # If this repository is shared, vfs pointing to shared repo
531 sharedvfs = None
503
532
504 # The .hg/ path should exist and should be a directory. All other
533 # The .hg/ path should exist and should be a directory. All other
505 # cases are errors.
534 # cases are errors.
506 if not hgvfs.isdir():
535 if not hgvfs.isdir():
507 try:
536 try:
508 hgvfs.stat()
537 hgvfs.stat()
509 except OSError as e:
538 except OSError as e:
510 if e.errno != errno.ENOENT:
539 if e.errno != errno.ENOENT:
511 raise
540 raise
512 except ValueError as e:
541 except ValueError as e:
513 # Can be raised on Python 3.8 when path is invalid.
542 # Can be raised on Python 3.8 when path is invalid.
514 raise error.Abort(
543 raise error.Abort(
515 _(b'invalid path %s: %s') % (path, pycompat.bytestr(e))
544 _(b'invalid path %s: %s') % (path, pycompat.bytestr(e))
516 )
545 )
517
546
518 raise error.RepoError(_(b'repository %s not found') % path)
547 raise error.RepoError(_(b'repository %s not found') % path)
519
548
520 # .hg/requires file contains a newline-delimited list of
549 # .hg/requires file contains a newline-delimited list of
521 # features/capabilities the opener (us) must have in order to use
550 # features/capabilities the opener (us) must have in order to use
522 # the repository. This file was introduced in Mercurial 0.9.2,
551 # the repository. This file was introduced in Mercurial 0.9.2,
523 # which means very old repositories may not have one. We assume
552 # which means very old repositories may not have one. We assume
524 # a missing file translates to no requirements.
553 # a missing file translates to no requirements.
525 try:
554 try:
526 requirements = set(hgvfs.read(b'requires').splitlines())
555 requirements = set(hgvfs.read(b'requires').splitlines())
527 except IOError as e:
556 except IOError as e:
528 if e.errno != errno.ENOENT:
557 if e.errno != errno.ENOENT:
529 raise
558 raise
530 requirements = set()
559 requirements = set()
531
560
532 # The .hg/hgrc file may load extensions or contain config options
561 # The .hg/hgrc file may load extensions or contain config options
533 # that influence repository construction. Attempt to load it and
562 # that influence repository construction. Attempt to load it and
534 # process any new extensions that it may have pulled in.
563 # process any new extensions that it may have pulled in.
535 if loadhgrc(ui, wdirvfs, hgvfs, requirements):
564 if loadhgrc(ui, wdirvfs, hgvfs, requirements):
536 afterhgrcload(ui, wdirvfs, hgvfs, requirements)
565 afterhgrcload(ui, wdirvfs, hgvfs, requirements)
537 extensions.loadall(ui)
566 extensions.loadall(ui)
538 extensions.populateui(ui)
567 extensions.populateui(ui)
539
568
540 # Set of module names of extensions loaded for this repository.
569 # Set of module names of extensions loaded for this repository.
541 extensionmodulenames = {m.__name__ for n, m in extensions.extensions(ui)}
570 extensionmodulenames = {m.__name__ for n, m in extensions.extensions(ui)}
542
571
543 supportedrequirements = gathersupportedrequirements(ui)
572 supportedrequirements = gathersupportedrequirements(ui)
544
573
545 # We first validate the requirements are known.
574 # We first validate the requirements are known.
546 ensurerequirementsrecognized(requirements, supportedrequirements)
575 ensurerequirementsrecognized(requirements, supportedrequirements)
547
576
548 # Then we validate that the known set is reasonable to use together.
577 # Then we validate that the known set is reasonable to use together.
549 ensurerequirementscompatible(ui, requirements)
578 ensurerequirementscompatible(ui, requirements)
550
579
551 # TODO there are unhandled edge cases related to opening repositories with
580 # TODO there are unhandled edge cases related to opening repositories with
552 # shared storage. If storage is shared, we should also test for requirements
581 # shared storage. If storage is shared, we should also test for requirements
553 # compatibility in the pointed-to repo. This entails loading the .hg/hgrc in
582 # compatibility in the pointed-to repo. This entails loading the .hg/hgrc in
554 # that repo, as that repo may load extensions needed to open it. This is a
583 # that repo, as that repo may load extensions needed to open it. This is a
555 # bit complicated because we don't want the other hgrc to overwrite settings
584 # bit complicated because we don't want the other hgrc to overwrite settings
556 # in this hgrc.
585 # in this hgrc.
557 #
586 #
558 # This bug is somewhat mitigated by the fact that we copy the .hg/requires
587 # This bug is somewhat mitigated by the fact that we copy the .hg/requires
559 # file when sharing repos. But if a requirement is added after the share is
588 # file when sharing repos. But if a requirement is added after the share is
560 # performed, thereby introducing a new requirement for the opener, we may
589 # performed, thereby introducing a new requirement for the opener, we may
561 # will not see that and could encounter a run-time error interacting with
590 # will not see that and could encounter a run-time error interacting with
562 # that shared store since it has an unknown-to-us requirement.
591 # that shared store since it has an unknown-to-us requirement.
563
592
564 # At this point, we know we should be capable of opening the repository.
593 # At this point, we know we should be capable of opening the repository.
565 # Now get on with doing that.
594 # Now get on with doing that.
566
595
567 features = set()
596 features = set()
568
597
569 # The "store" part of the repository holds versioned data. How it is
598 # The "store" part of the repository holds versioned data. How it is
570 # accessed is determined by various requirements. The ``shared`` or
599 # accessed is determined by various requirements. If `shared` or
571 # ``relshared`` requirements indicate the store lives in the path contained
600 # `relshared` requirements are present, this indicates current repository
572 # in the ``.hg/sharedpath`` file. This is an absolute path for
601 # is a share and store exists in path mentioned in `.hg/sharedpath`
573 # ``shared`` and relative to ``.hg/`` for ``relshared``.
602 shared = b'shared' in requirements or b'relshared' in requirements
574 if b'shared' in requirements or b'relshared' in requirements:
603 if shared:
575 sharedpath = hgvfs.read(b'sharedpath').rstrip(b'\n')
604 sharedvfs = _getsharedvfs(hgvfs, requirements)
576 if b'relshared' in requirements:
577 sharedpath = hgvfs.join(sharedpath)
578
579 sharedvfs = vfsmod.vfs(sharedpath, realpath=True)
580
581 if not sharedvfs.exists():
582 raise error.RepoError(
583 _(b'.hg/sharedpath points to nonexistent directory %s')
584 % sharedvfs.base
585 )
586
587 features.add(repository.REPO_FEATURE_SHARED_STORAGE)
588
589 storebasepath = sharedvfs.base
605 storebasepath = sharedvfs.base
590 cachepath = sharedvfs.join(b'cache')
606 cachepath = sharedvfs.join(b'cache')
607 features.add(repository.REPO_FEATURE_SHARED_STORAGE)
591 else:
608 else:
592 storebasepath = hgvfs.base
609 storebasepath = hgvfs.base
593 cachepath = hgvfs.join(b'cache')
610 cachepath = hgvfs.join(b'cache')
594 wcachepath = hgvfs.join(b'wcache')
611 wcachepath = hgvfs.join(b'wcache')
595
612
596 # The store has changed over time and the exact layout is dictated by
613 # The store has changed over time and the exact layout is dictated by
597 # requirements. The store interface abstracts differences across all
614 # requirements. The store interface abstracts differences across all
598 # of them.
615 # of them.
599 store = makestore(
616 store = makestore(
600 requirements,
617 requirements,
601 storebasepath,
618 storebasepath,
602 lambda base: vfsmod.vfs(base, cacheaudited=True),
619 lambda base: vfsmod.vfs(base, cacheaudited=True),
603 )
620 )
604 hgvfs.createmode = store.createmode
621 hgvfs.createmode = store.createmode
605
622
606 storevfs = store.vfs
623 storevfs = store.vfs
607 storevfs.options = resolvestorevfsoptions(ui, requirements, features)
624 storevfs.options = resolvestorevfsoptions(ui, requirements, features)
608
625
609 # The cache vfs is used to manage cache files.
626 # The cache vfs is used to manage cache files.
610 cachevfs = vfsmod.vfs(cachepath, cacheaudited=True)
627 cachevfs = vfsmod.vfs(cachepath, cacheaudited=True)
611 cachevfs.createmode = store.createmode
628 cachevfs.createmode = store.createmode
612 # The cache vfs is used to manage cache files related to the working copy
629 # The cache vfs is used to manage cache files related to the working copy
613 wcachevfs = vfsmod.vfs(wcachepath, cacheaudited=True)
630 wcachevfs = vfsmod.vfs(wcachepath, cacheaudited=True)
614 wcachevfs.createmode = store.createmode
631 wcachevfs.createmode = store.createmode
615
632
616 # Now resolve the type for the repository object. We do this by repeatedly
633 # Now resolve the type for the repository object. We do this by repeatedly
617 # calling a factory function to produces types for specific aspects of the
634 # calling a factory function to produces types for specific aspects of the
618 # repo's operation. The aggregate returned types are used as base classes
635 # repo's operation. The aggregate returned types are used as base classes
619 # for a dynamically-derived type, which will represent our new repository.
636 # for a dynamically-derived type, which will represent our new repository.
620
637
621 bases = []
638 bases = []
622 extrastate = {}
639 extrastate = {}
623
640
624 for iface, fn in REPO_INTERFACES:
641 for iface, fn in REPO_INTERFACES:
625 # We pass all potentially useful state to give extensions tons of
642 # We pass all potentially useful state to give extensions tons of
626 # flexibility.
643 # flexibility.
627 typ = fn()(
644 typ = fn()(
628 ui=ui,
645 ui=ui,
629 intents=intents,
646 intents=intents,
630 requirements=requirements,
647 requirements=requirements,
631 features=features,
648 features=features,
632 wdirvfs=wdirvfs,
649 wdirvfs=wdirvfs,
633 hgvfs=hgvfs,
650 hgvfs=hgvfs,
634 store=store,
651 store=store,
635 storevfs=storevfs,
652 storevfs=storevfs,
636 storeoptions=storevfs.options,
653 storeoptions=storevfs.options,
637 cachevfs=cachevfs,
654 cachevfs=cachevfs,
638 wcachevfs=wcachevfs,
655 wcachevfs=wcachevfs,
639 extensionmodulenames=extensionmodulenames,
656 extensionmodulenames=extensionmodulenames,
640 extrastate=extrastate,
657 extrastate=extrastate,
641 baseclasses=bases,
658 baseclasses=bases,
642 )
659 )
643
660
644 if not isinstance(typ, type):
661 if not isinstance(typ, type):
645 raise error.ProgrammingError(
662 raise error.ProgrammingError(
646 b'unable to construct type for %s' % iface
663 b'unable to construct type for %s' % iface
647 )
664 )
648
665
649 bases.append(typ)
666 bases.append(typ)
650
667
651 # type() allows you to use characters in type names that wouldn't be
668 # type() allows you to use characters in type names that wouldn't be
652 # recognized as Python symbols in source code. We abuse that to add
669 # recognized as Python symbols in source code. We abuse that to add
653 # rich information about our constructed repo.
670 # rich information about our constructed repo.
654 name = pycompat.sysstr(
671 name = pycompat.sysstr(
655 b'derivedrepo:%s<%s>' % (wdirvfs.base, b','.join(sorted(requirements)))
672 b'derivedrepo:%s<%s>' % (wdirvfs.base, b','.join(sorted(requirements)))
656 )
673 )
657
674
658 cls = type(name, tuple(bases), {})
675 cls = type(name, tuple(bases), {})
659
676
660 return cls(
677 return cls(
661 baseui=baseui,
678 baseui=baseui,
662 ui=ui,
679 ui=ui,
663 origroot=path,
680 origroot=path,
664 wdirvfs=wdirvfs,
681 wdirvfs=wdirvfs,
665 hgvfs=hgvfs,
682 hgvfs=hgvfs,
666 requirements=requirements,
683 requirements=requirements,
667 supportedrequirements=supportedrequirements,
684 supportedrequirements=supportedrequirements,
668 sharedpath=storebasepath,
685 sharedpath=storebasepath,
669 store=store,
686 store=store,
670 cachevfs=cachevfs,
687 cachevfs=cachevfs,
671 wcachevfs=wcachevfs,
688 wcachevfs=wcachevfs,
672 features=features,
689 features=features,
673 intents=intents,
690 intents=intents,
674 )
691 )
675
692
676
693
677 def loadhgrc(ui, wdirvfs, hgvfs, requirements):
694 def loadhgrc(ui, wdirvfs, hgvfs, requirements):
678 """Load hgrc files/content into a ui instance.
695 """Load hgrc files/content into a ui instance.
679
696
680 This is called during repository opening to load any additional
697 This is called during repository opening to load any additional
681 config files or settings relevant to the current repository.
698 config files or settings relevant to the current repository.
682
699
683 Returns a bool indicating whether any additional configs were loaded.
700 Returns a bool indicating whether any additional configs were loaded.
684
701
685 Extensions should monkeypatch this function to modify how per-repo
702 Extensions should monkeypatch this function to modify how per-repo
686 configs are loaded. For example, an extension may wish to pull in
703 configs are loaded. For example, an extension may wish to pull in
687 configs from alternate files or sources.
704 configs from alternate files or sources.
688 """
705 """
689 if not rcutil.use_repo_hgrc():
706 if not rcutil.use_repo_hgrc():
690 return False
707 return False
691 try:
708 try:
692 ui.readconfig(hgvfs.join(b'hgrc'), root=wdirvfs.base)
709 ui.readconfig(hgvfs.join(b'hgrc'), root=wdirvfs.base)
693 return True
710 return True
694 except IOError:
711 except IOError:
695 return False
712 return False
696
713
697
714
698 def afterhgrcload(ui, wdirvfs, hgvfs, requirements):
715 def afterhgrcload(ui, wdirvfs, hgvfs, requirements):
699 """Perform additional actions after .hg/hgrc is loaded.
716 """Perform additional actions after .hg/hgrc is loaded.
700
717
701 This function is called during repository loading immediately after
718 This function is called during repository loading immediately after
702 the .hg/hgrc file is loaded and before per-repo extensions are loaded.
719 the .hg/hgrc file is loaded and before per-repo extensions are loaded.
703
720
704 The function can be used to validate configs, automatically add
721 The function can be used to validate configs, automatically add
705 options (including extensions) based on requirements, etc.
722 options (including extensions) based on requirements, etc.
706 """
723 """
707
724
708 # Map of requirements to list of extensions to load automatically when
725 # Map of requirements to list of extensions to load automatically when
709 # requirement is present.
726 # requirement is present.
710 autoextensions = {
727 autoextensions = {
711 b'git': [b'git'],
728 b'git': [b'git'],
712 b'largefiles': [b'largefiles'],
729 b'largefiles': [b'largefiles'],
713 b'lfs': [b'lfs'],
730 b'lfs': [b'lfs'],
714 }
731 }
715
732
716 for requirement, names in sorted(autoextensions.items()):
733 for requirement, names in sorted(autoextensions.items()):
717 if requirement not in requirements:
734 if requirement not in requirements:
718 continue
735 continue
719
736
720 for name in names:
737 for name in names:
721 if not ui.hasconfig(b'extensions', name):
738 if not ui.hasconfig(b'extensions', name):
722 ui.setconfig(b'extensions', name, b'', source=b'autoload')
739 ui.setconfig(b'extensions', name, b'', source=b'autoload')
723
740
724
741
725 def gathersupportedrequirements(ui):
742 def gathersupportedrequirements(ui):
726 """Determine the complete set of recognized requirements."""
743 """Determine the complete set of recognized requirements."""
727 # Start with all requirements supported by this file.
744 # Start with all requirements supported by this file.
728 supported = set(localrepository._basesupported)
745 supported = set(localrepository._basesupported)
729
746
730 # Execute ``featuresetupfuncs`` entries if they belong to an extension
747 # Execute ``featuresetupfuncs`` entries if they belong to an extension
731 # relevant to this ui instance.
748 # relevant to this ui instance.
732 modules = {m.__name__ for n, m in extensions.extensions(ui)}
749 modules = {m.__name__ for n, m in extensions.extensions(ui)}
733
750
734 for fn in featuresetupfuncs:
751 for fn in featuresetupfuncs:
735 if fn.__module__ in modules:
752 if fn.__module__ in modules:
736 fn(ui, supported)
753 fn(ui, supported)
737
754
738 # Add derived requirements from registered compression engines.
755 # Add derived requirements from registered compression engines.
739 for name in util.compengines:
756 for name in util.compengines:
740 engine = util.compengines[name]
757 engine = util.compengines[name]
741 if engine.available() and engine.revlogheader():
758 if engine.available() and engine.revlogheader():
742 supported.add(b'exp-compression-%s' % name)
759 supported.add(b'exp-compression-%s' % name)
743 if engine.name() == b'zstd':
760 if engine.name() == b'zstd':
744 supported.add(b'revlog-compression-zstd')
761 supported.add(b'revlog-compression-zstd')
745
762
746 return supported
763 return supported
747
764
748
765
749 def ensurerequirementsrecognized(requirements, supported):
766 def ensurerequirementsrecognized(requirements, supported):
750 """Validate that a set of local requirements is recognized.
767 """Validate that a set of local requirements is recognized.
751
768
752 Receives a set of requirements. Raises an ``error.RepoError`` if there
769 Receives a set of requirements. Raises an ``error.RepoError`` if there
753 exists any requirement in that set that currently loaded code doesn't
770 exists any requirement in that set that currently loaded code doesn't
754 recognize.
771 recognize.
755
772
756 Returns a set of supported requirements.
773 Returns a set of supported requirements.
757 """
774 """
758 missing = set()
775 missing = set()
759
776
760 for requirement in requirements:
777 for requirement in requirements:
761 if requirement in supported:
778 if requirement in supported:
762 continue
779 continue
763
780
764 if not requirement or not requirement[0:1].isalnum():
781 if not requirement or not requirement[0:1].isalnum():
765 raise error.RequirementError(_(b'.hg/requires file is corrupt'))
782 raise error.RequirementError(_(b'.hg/requires file is corrupt'))
766
783
767 missing.add(requirement)
784 missing.add(requirement)
768
785
769 if missing:
786 if missing:
770 raise error.RequirementError(
787 raise error.RequirementError(
771 _(b'repository requires features unknown to this Mercurial: %s')
788 _(b'repository requires features unknown to this Mercurial: %s')
772 % b' '.join(sorted(missing)),
789 % b' '.join(sorted(missing)),
773 hint=_(
790 hint=_(
774 b'see https://mercurial-scm.org/wiki/MissingRequirement '
791 b'see https://mercurial-scm.org/wiki/MissingRequirement '
775 b'for more information'
792 b'for more information'
776 ),
793 ),
777 )
794 )
778
795
779
796
780 def ensurerequirementscompatible(ui, requirements):
797 def ensurerequirementscompatible(ui, requirements):
781 """Validates that a set of recognized requirements is mutually compatible.
798 """Validates that a set of recognized requirements is mutually compatible.
782
799
783 Some requirements may not be compatible with others or require
800 Some requirements may not be compatible with others or require
784 config options that aren't enabled. This function is called during
801 config options that aren't enabled. This function is called during
785 repository opening to ensure that the set of requirements needed
802 repository opening to ensure that the set of requirements needed
786 to open a repository is sane and compatible with config options.
803 to open a repository is sane and compatible with config options.
787
804
788 Extensions can monkeypatch this function to perform additional
805 Extensions can monkeypatch this function to perform additional
789 checking.
806 checking.
790
807
791 ``error.RepoError`` should be raised on failure.
808 ``error.RepoError`` should be raised on failure.
792 """
809 """
793 if b'exp-sparse' in requirements and not sparse.enabled:
810 if b'exp-sparse' in requirements and not sparse.enabled:
794 raise error.RepoError(
811 raise error.RepoError(
795 _(
812 _(
796 b'repository is using sparse feature but '
813 b'repository is using sparse feature but '
797 b'sparse is not enabled; enable the '
814 b'sparse is not enabled; enable the '
798 b'"sparse" extensions to access'
815 b'"sparse" extensions to access'
799 )
816 )
800 )
817 )
801
818
802
819
803 def makestore(requirements, path, vfstype):
820 def makestore(requirements, path, vfstype):
804 """Construct a storage object for a repository."""
821 """Construct a storage object for a repository."""
805 if b'store' in requirements:
822 if b'store' in requirements:
806 if b'fncache' in requirements:
823 if b'fncache' in requirements:
807 return storemod.fncachestore(
824 return storemod.fncachestore(
808 path, vfstype, b'dotencode' in requirements
825 path, vfstype, b'dotencode' in requirements
809 )
826 )
810
827
811 return storemod.encodedstore(path, vfstype)
828 return storemod.encodedstore(path, vfstype)
812
829
813 return storemod.basicstore(path, vfstype)
830 return storemod.basicstore(path, vfstype)
814
831
815
832
816 def resolvestorevfsoptions(ui, requirements, features):
833 def resolvestorevfsoptions(ui, requirements, features):
817 """Resolve the options to pass to the store vfs opener.
834 """Resolve the options to pass to the store vfs opener.
818
835
819 The returned dict is used to influence behavior of the storage layer.
836 The returned dict is used to influence behavior of the storage layer.
820 """
837 """
821 options = {}
838 options = {}
822
839
823 if b'treemanifest' in requirements:
840 if b'treemanifest' in requirements:
824 options[b'treemanifest'] = True
841 options[b'treemanifest'] = True
825
842
826 # experimental config: format.manifestcachesize
843 # experimental config: format.manifestcachesize
827 manifestcachesize = ui.configint(b'format', b'manifestcachesize')
844 manifestcachesize = ui.configint(b'format', b'manifestcachesize')
828 if manifestcachesize is not None:
845 if manifestcachesize is not None:
829 options[b'manifestcachesize'] = manifestcachesize
846 options[b'manifestcachesize'] = manifestcachesize
830
847
831 # In the absence of another requirement superseding a revlog-related
848 # In the absence of another requirement superseding a revlog-related
832 # requirement, we have to assume the repo is using revlog version 0.
849 # requirement, we have to assume the repo is using revlog version 0.
833 # This revlog format is super old and we don't bother trying to parse
850 # This revlog format is super old and we don't bother trying to parse
834 # opener options for it because those options wouldn't do anything
851 # opener options for it because those options wouldn't do anything
835 # meaningful on such old repos.
852 # meaningful on such old repos.
836 if b'revlogv1' in requirements or REVLOGV2_REQUIREMENT in requirements:
853 if b'revlogv1' in requirements or REVLOGV2_REQUIREMENT in requirements:
837 options.update(resolverevlogstorevfsoptions(ui, requirements, features))
854 options.update(resolverevlogstorevfsoptions(ui, requirements, features))
838 else: # explicitly mark repo as using revlogv0
855 else: # explicitly mark repo as using revlogv0
839 options[b'revlogv0'] = True
856 options[b'revlogv0'] = True
840
857
841 if COPIESSDC_REQUIREMENT in requirements:
858 if COPIESSDC_REQUIREMENT in requirements:
842 options[b'copies-storage'] = b'changeset-sidedata'
859 options[b'copies-storage'] = b'changeset-sidedata'
843 else:
860 else:
844 writecopiesto = ui.config(b'experimental', b'copies.write-to')
861 writecopiesto = ui.config(b'experimental', b'copies.write-to')
845 copiesextramode = (b'changeset-only', b'compatibility')
862 copiesextramode = (b'changeset-only', b'compatibility')
846 if writecopiesto in copiesextramode:
863 if writecopiesto in copiesextramode:
847 options[b'copies-storage'] = b'extra'
864 options[b'copies-storage'] = b'extra'
848
865
849 return options
866 return options
850
867
851
868
852 def resolverevlogstorevfsoptions(ui, requirements, features):
869 def resolverevlogstorevfsoptions(ui, requirements, features):
853 """Resolve opener options specific to revlogs."""
870 """Resolve opener options specific to revlogs."""
854
871
855 options = {}
872 options = {}
856 options[b'flagprocessors'] = {}
873 options[b'flagprocessors'] = {}
857
874
858 if b'revlogv1' in requirements:
875 if b'revlogv1' in requirements:
859 options[b'revlogv1'] = True
876 options[b'revlogv1'] = True
860 if REVLOGV2_REQUIREMENT in requirements:
877 if REVLOGV2_REQUIREMENT in requirements:
861 options[b'revlogv2'] = True
878 options[b'revlogv2'] = True
862
879
863 if b'generaldelta' in requirements:
880 if b'generaldelta' in requirements:
864 options[b'generaldelta'] = True
881 options[b'generaldelta'] = True
865
882
866 # experimental config: format.chunkcachesize
883 # experimental config: format.chunkcachesize
867 chunkcachesize = ui.configint(b'format', b'chunkcachesize')
884 chunkcachesize = ui.configint(b'format', b'chunkcachesize')
868 if chunkcachesize is not None:
885 if chunkcachesize is not None:
869 options[b'chunkcachesize'] = chunkcachesize
886 options[b'chunkcachesize'] = chunkcachesize
870
887
871 deltabothparents = ui.configbool(
888 deltabothparents = ui.configbool(
872 b'storage', b'revlog.optimize-delta-parent-choice'
889 b'storage', b'revlog.optimize-delta-parent-choice'
873 )
890 )
874 options[b'deltabothparents'] = deltabothparents
891 options[b'deltabothparents'] = deltabothparents
875
892
876 lazydelta = ui.configbool(b'storage', b'revlog.reuse-external-delta')
893 lazydelta = ui.configbool(b'storage', b'revlog.reuse-external-delta')
877 lazydeltabase = False
894 lazydeltabase = False
878 if lazydelta:
895 if lazydelta:
879 lazydeltabase = ui.configbool(
896 lazydeltabase = ui.configbool(
880 b'storage', b'revlog.reuse-external-delta-parent'
897 b'storage', b'revlog.reuse-external-delta-parent'
881 )
898 )
882 if lazydeltabase is None:
899 if lazydeltabase is None:
883 lazydeltabase = not scmutil.gddeltaconfig(ui)
900 lazydeltabase = not scmutil.gddeltaconfig(ui)
884 options[b'lazydelta'] = lazydelta
901 options[b'lazydelta'] = lazydelta
885 options[b'lazydeltabase'] = lazydeltabase
902 options[b'lazydeltabase'] = lazydeltabase
886
903
887 chainspan = ui.configbytes(b'experimental', b'maxdeltachainspan')
904 chainspan = ui.configbytes(b'experimental', b'maxdeltachainspan')
888 if 0 <= chainspan:
905 if 0 <= chainspan:
889 options[b'maxdeltachainspan'] = chainspan
906 options[b'maxdeltachainspan'] = chainspan
890
907
891 mmapindexthreshold = ui.configbytes(b'experimental', b'mmapindexthreshold')
908 mmapindexthreshold = ui.configbytes(b'experimental', b'mmapindexthreshold')
892 if mmapindexthreshold is not None:
909 if mmapindexthreshold is not None:
893 options[b'mmapindexthreshold'] = mmapindexthreshold
910 options[b'mmapindexthreshold'] = mmapindexthreshold
894
911
895 withsparseread = ui.configbool(b'experimental', b'sparse-read')
912 withsparseread = ui.configbool(b'experimental', b'sparse-read')
896 srdensitythres = float(
913 srdensitythres = float(
897 ui.config(b'experimental', b'sparse-read.density-threshold')
914 ui.config(b'experimental', b'sparse-read.density-threshold')
898 )
915 )
899 srmingapsize = ui.configbytes(b'experimental', b'sparse-read.min-gap-size')
916 srmingapsize = ui.configbytes(b'experimental', b'sparse-read.min-gap-size')
900 options[b'with-sparse-read'] = withsparseread
917 options[b'with-sparse-read'] = withsparseread
901 options[b'sparse-read-density-threshold'] = srdensitythres
918 options[b'sparse-read-density-threshold'] = srdensitythres
902 options[b'sparse-read-min-gap-size'] = srmingapsize
919 options[b'sparse-read-min-gap-size'] = srmingapsize
903
920
904 sparserevlog = SPARSEREVLOG_REQUIREMENT in requirements
921 sparserevlog = SPARSEREVLOG_REQUIREMENT in requirements
905 options[b'sparse-revlog'] = sparserevlog
922 options[b'sparse-revlog'] = sparserevlog
906 if sparserevlog:
923 if sparserevlog:
907 options[b'generaldelta'] = True
924 options[b'generaldelta'] = True
908
925
909 sidedata = SIDEDATA_REQUIREMENT in requirements
926 sidedata = SIDEDATA_REQUIREMENT in requirements
910 options[b'side-data'] = sidedata
927 options[b'side-data'] = sidedata
911
928
912 maxchainlen = None
929 maxchainlen = None
913 if sparserevlog:
930 if sparserevlog:
914 maxchainlen = revlogconst.SPARSE_REVLOG_MAX_CHAIN_LENGTH
931 maxchainlen = revlogconst.SPARSE_REVLOG_MAX_CHAIN_LENGTH
915 # experimental config: format.maxchainlen
932 # experimental config: format.maxchainlen
916 maxchainlen = ui.configint(b'format', b'maxchainlen', maxchainlen)
933 maxchainlen = ui.configint(b'format', b'maxchainlen', maxchainlen)
917 if maxchainlen is not None:
934 if maxchainlen is not None:
918 options[b'maxchainlen'] = maxchainlen
935 options[b'maxchainlen'] = maxchainlen
919
936
920 for r in requirements:
937 for r in requirements:
921 # we allow multiple compression engine requirement to co-exist because
938 # we allow multiple compression engine requirement to co-exist because
922 # strickly speaking, revlog seems to support mixed compression style.
939 # strickly speaking, revlog seems to support mixed compression style.
923 #
940 #
924 # The compression used for new entries will be "the last one"
941 # The compression used for new entries will be "the last one"
925 prefix = r.startswith
942 prefix = r.startswith
926 if prefix(b'revlog-compression-') or prefix(b'exp-compression-'):
943 if prefix(b'revlog-compression-') or prefix(b'exp-compression-'):
927 options[b'compengine'] = r.split(b'-', 2)[2]
944 options[b'compengine'] = r.split(b'-', 2)[2]
928
945
929 options[b'zlib.level'] = ui.configint(b'storage', b'revlog.zlib.level')
946 options[b'zlib.level'] = ui.configint(b'storage', b'revlog.zlib.level')
930 if options[b'zlib.level'] is not None:
947 if options[b'zlib.level'] is not None:
931 if not (0 <= options[b'zlib.level'] <= 9):
948 if not (0 <= options[b'zlib.level'] <= 9):
932 msg = _(b'invalid value for `storage.revlog.zlib.level` config: %d')
949 msg = _(b'invalid value for `storage.revlog.zlib.level` config: %d')
933 raise error.Abort(msg % options[b'zlib.level'])
950 raise error.Abort(msg % options[b'zlib.level'])
934 options[b'zstd.level'] = ui.configint(b'storage', b'revlog.zstd.level')
951 options[b'zstd.level'] = ui.configint(b'storage', b'revlog.zstd.level')
935 if options[b'zstd.level'] is not None:
952 if options[b'zstd.level'] is not None:
936 if not (0 <= options[b'zstd.level'] <= 22):
953 if not (0 <= options[b'zstd.level'] <= 22):
937 msg = _(b'invalid value for `storage.revlog.zstd.level` config: %d')
954 msg = _(b'invalid value for `storage.revlog.zstd.level` config: %d')
938 raise error.Abort(msg % options[b'zstd.level'])
955 raise error.Abort(msg % options[b'zstd.level'])
939
956
940 if repository.NARROW_REQUIREMENT in requirements:
957 if repository.NARROW_REQUIREMENT in requirements:
941 options[b'enableellipsis'] = True
958 options[b'enableellipsis'] = True
942
959
943 if ui.configbool(b'experimental', b'rust.index'):
960 if ui.configbool(b'experimental', b'rust.index'):
944 options[b'rust.index'] = True
961 options[b'rust.index'] = True
945 if NODEMAP_REQUIREMENT in requirements:
962 if NODEMAP_REQUIREMENT in requirements:
946 options[b'persistent-nodemap'] = True
963 options[b'persistent-nodemap'] = True
947 if ui.configbool(b'storage', b'revlog.nodemap.mmap'):
964 if ui.configbool(b'storage', b'revlog.nodemap.mmap'):
948 options[b'persistent-nodemap.mmap'] = True
965 options[b'persistent-nodemap.mmap'] = True
949 epnm = ui.config(b'storage', b'revlog.nodemap.mode')
966 epnm = ui.config(b'storage', b'revlog.nodemap.mode')
950 options[b'persistent-nodemap.mode'] = epnm
967 options[b'persistent-nodemap.mode'] = epnm
951 if ui.configbool(b'devel', b'persistent-nodemap'):
968 if ui.configbool(b'devel', b'persistent-nodemap'):
952 options[b'devel-force-nodemap'] = True
969 options[b'devel-force-nodemap'] = True
953
970
954 return options
971 return options
955
972
956
973
957 def makemain(**kwargs):
974 def makemain(**kwargs):
958 """Produce a type conforming to ``ilocalrepositorymain``."""
975 """Produce a type conforming to ``ilocalrepositorymain``."""
959 return localrepository
976 return localrepository
960
977
961
978
962 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
979 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
963 class revlogfilestorage(object):
980 class revlogfilestorage(object):
964 """File storage when using revlogs."""
981 """File storage when using revlogs."""
965
982
966 def file(self, path):
983 def file(self, path):
967 if path[0] == b'/':
984 if path[0] == b'/':
968 path = path[1:]
985 path = path[1:]
969
986
970 return filelog.filelog(self.svfs, path)
987 return filelog.filelog(self.svfs, path)
971
988
972
989
973 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
990 @interfaceutil.implementer(repository.ilocalrepositoryfilestorage)
974 class revlognarrowfilestorage(object):
991 class revlognarrowfilestorage(object):
975 """File storage when using revlogs and narrow files."""
992 """File storage when using revlogs and narrow files."""
976
993
977 def file(self, path):
994 def file(self, path):
978 if path[0] == b'/':
995 if path[0] == b'/':
979 path = path[1:]
996 path = path[1:]
980
997
981 return filelog.narrowfilelog(self.svfs, path, self._storenarrowmatch)
998 return filelog.narrowfilelog(self.svfs, path, self._storenarrowmatch)
982
999
983
1000
984 def makefilestorage(requirements, features, **kwargs):
1001 def makefilestorage(requirements, features, **kwargs):
985 """Produce a type conforming to ``ilocalrepositoryfilestorage``."""
1002 """Produce a type conforming to ``ilocalrepositoryfilestorage``."""
986 features.add(repository.REPO_FEATURE_REVLOG_FILE_STORAGE)
1003 features.add(repository.REPO_FEATURE_REVLOG_FILE_STORAGE)
987 features.add(repository.REPO_FEATURE_STREAM_CLONE)
1004 features.add(repository.REPO_FEATURE_STREAM_CLONE)
988
1005
989 if repository.NARROW_REQUIREMENT in requirements:
1006 if repository.NARROW_REQUIREMENT in requirements:
990 return revlognarrowfilestorage
1007 return revlognarrowfilestorage
991 else:
1008 else:
992 return revlogfilestorage
1009 return revlogfilestorage
993
1010
994
1011
995 # List of repository interfaces and factory functions for them. Each
1012 # List of repository interfaces and factory functions for them. Each
996 # will be called in order during ``makelocalrepository()`` to iteratively
1013 # will be called in order during ``makelocalrepository()`` to iteratively
997 # derive the final type for a local repository instance. We capture the
1014 # derive the final type for a local repository instance. We capture the
998 # function as a lambda so we don't hold a reference and the module-level
1015 # function as a lambda so we don't hold a reference and the module-level
999 # functions can be wrapped.
1016 # functions can be wrapped.
1000 REPO_INTERFACES = [
1017 REPO_INTERFACES = [
1001 (repository.ilocalrepositorymain, lambda: makemain),
1018 (repository.ilocalrepositorymain, lambda: makemain),
1002 (repository.ilocalrepositoryfilestorage, lambda: makefilestorage),
1019 (repository.ilocalrepositoryfilestorage, lambda: makefilestorage),
1003 ]
1020 ]
1004
1021
1005
1022
1006 @interfaceutil.implementer(repository.ilocalrepositorymain)
1023 @interfaceutil.implementer(repository.ilocalrepositorymain)
1007 class localrepository(object):
1024 class localrepository(object):
1008 """Main class for representing local repositories.
1025 """Main class for representing local repositories.
1009
1026
1010 All local repositories are instances of this class.
1027 All local repositories are instances of this class.
1011
1028
1012 Constructed on its own, instances of this class are not usable as
1029 Constructed on its own, instances of this class are not usable as
1013 repository objects. To obtain a usable repository object, call
1030 repository objects. To obtain a usable repository object, call
1014 ``hg.repository()``, ``localrepo.instance()``, or
1031 ``hg.repository()``, ``localrepo.instance()``, or
1015 ``localrepo.makelocalrepository()``. The latter is the lowest-level.
1032 ``localrepo.makelocalrepository()``. The latter is the lowest-level.
1016 ``instance()`` adds support for creating new repositories.
1033 ``instance()`` adds support for creating new repositories.
1017 ``hg.repository()`` adds more extension integration, including calling
1034 ``hg.repository()`` adds more extension integration, including calling
1018 ``reposetup()``. Generally speaking, ``hg.repository()`` should be
1035 ``reposetup()``. Generally speaking, ``hg.repository()`` should be
1019 used.
1036 used.
1020 """
1037 """
1021
1038
1022 # obsolete experimental requirements:
1039 # obsolete experimental requirements:
1023 # - manifestv2: An experimental new manifest format that allowed
1040 # - manifestv2: An experimental new manifest format that allowed
1024 # for stem compression of long paths. Experiment ended up not
1041 # for stem compression of long paths. Experiment ended up not
1025 # being successful (repository sizes went up due to worse delta
1042 # being successful (repository sizes went up due to worse delta
1026 # chains), and the code was deleted in 4.6.
1043 # chains), and the code was deleted in 4.6.
1027 supportedformats = {
1044 supportedformats = {
1028 b'revlogv1',
1045 b'revlogv1',
1029 b'generaldelta',
1046 b'generaldelta',
1030 b'treemanifest',
1047 b'treemanifest',
1031 COPIESSDC_REQUIREMENT,
1048 COPIESSDC_REQUIREMENT,
1032 REVLOGV2_REQUIREMENT,
1049 REVLOGV2_REQUIREMENT,
1033 SIDEDATA_REQUIREMENT,
1050 SIDEDATA_REQUIREMENT,
1034 SPARSEREVLOG_REQUIREMENT,
1051 SPARSEREVLOG_REQUIREMENT,
1035 NODEMAP_REQUIREMENT,
1052 NODEMAP_REQUIREMENT,
1036 bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT,
1053 bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT,
1037 }
1054 }
1038 _basesupported = supportedformats | {
1055 _basesupported = supportedformats | {
1039 b'store',
1056 b'store',
1040 b'fncache',
1057 b'fncache',
1041 b'shared',
1058 b'shared',
1042 b'relshared',
1059 b'relshared',
1043 b'dotencode',
1060 b'dotencode',
1044 b'exp-sparse',
1061 b'exp-sparse',
1045 b'internal-phase',
1062 b'internal-phase',
1046 }
1063 }
1047
1064
1048 # list of prefix for file which can be written without 'wlock'
1065 # list of prefix for file which can be written without 'wlock'
1049 # Extensions should extend this list when needed
1066 # Extensions should extend this list when needed
1050 _wlockfreeprefix = {
1067 _wlockfreeprefix = {
1051 # We migh consider requiring 'wlock' for the next
1068 # We migh consider requiring 'wlock' for the next
1052 # two, but pretty much all the existing code assume
1069 # two, but pretty much all the existing code assume
1053 # wlock is not needed so we keep them excluded for
1070 # wlock is not needed so we keep them excluded for
1054 # now.
1071 # now.
1055 b'hgrc',
1072 b'hgrc',
1056 b'requires',
1073 b'requires',
1057 # XXX cache is a complicatged business someone
1074 # XXX cache is a complicatged business someone
1058 # should investigate this in depth at some point
1075 # should investigate this in depth at some point
1059 b'cache/',
1076 b'cache/',
1060 # XXX shouldn't be dirstate covered by the wlock?
1077 # XXX shouldn't be dirstate covered by the wlock?
1061 b'dirstate',
1078 b'dirstate',
1062 # XXX bisect was still a bit too messy at the time
1079 # XXX bisect was still a bit too messy at the time
1063 # this changeset was introduced. Someone should fix
1080 # this changeset was introduced. Someone should fix
1064 # the remainig bit and drop this line
1081 # the remainig bit and drop this line
1065 b'bisect.state',
1082 b'bisect.state',
1066 }
1083 }
1067
1084
1068 def __init__(
1085 def __init__(
1069 self,
1086 self,
1070 baseui,
1087 baseui,
1071 ui,
1088 ui,
1072 origroot,
1089 origroot,
1073 wdirvfs,
1090 wdirvfs,
1074 hgvfs,
1091 hgvfs,
1075 requirements,
1092 requirements,
1076 supportedrequirements,
1093 supportedrequirements,
1077 sharedpath,
1094 sharedpath,
1078 store,
1095 store,
1079 cachevfs,
1096 cachevfs,
1080 wcachevfs,
1097 wcachevfs,
1081 features,
1098 features,
1082 intents=None,
1099 intents=None,
1083 ):
1100 ):
1084 """Create a new local repository instance.
1101 """Create a new local repository instance.
1085
1102
1086 Most callers should use ``hg.repository()``, ``localrepo.instance()``,
1103 Most callers should use ``hg.repository()``, ``localrepo.instance()``,
1087 or ``localrepo.makelocalrepository()`` for obtaining a new repository
1104 or ``localrepo.makelocalrepository()`` for obtaining a new repository
1088 object.
1105 object.
1089
1106
1090 Arguments:
1107 Arguments:
1091
1108
1092 baseui
1109 baseui
1093 ``ui.ui`` instance that ``ui`` argument was based off of.
1110 ``ui.ui`` instance that ``ui`` argument was based off of.
1094
1111
1095 ui
1112 ui
1096 ``ui.ui`` instance for use by the repository.
1113 ``ui.ui`` instance for use by the repository.
1097
1114
1098 origroot
1115 origroot
1099 ``bytes`` path to working directory root of this repository.
1116 ``bytes`` path to working directory root of this repository.
1100
1117
1101 wdirvfs
1118 wdirvfs
1102 ``vfs.vfs`` rooted at the working directory.
1119 ``vfs.vfs`` rooted at the working directory.
1103
1120
1104 hgvfs
1121 hgvfs
1105 ``vfs.vfs`` rooted at .hg/
1122 ``vfs.vfs`` rooted at .hg/
1106
1123
1107 requirements
1124 requirements
1108 ``set`` of bytestrings representing repository opening requirements.
1125 ``set`` of bytestrings representing repository opening requirements.
1109
1126
1110 supportedrequirements
1127 supportedrequirements
1111 ``set`` of bytestrings representing repository requirements that we
1128 ``set`` of bytestrings representing repository requirements that we
1112 know how to open. May be a supetset of ``requirements``.
1129 know how to open. May be a supetset of ``requirements``.
1113
1130
1114 sharedpath
1131 sharedpath
1115 ``bytes`` Defining path to storage base directory. Points to a
1132 ``bytes`` Defining path to storage base directory. Points to a
1116 ``.hg/`` directory somewhere.
1133 ``.hg/`` directory somewhere.
1117
1134
1118 store
1135 store
1119 ``store.basicstore`` (or derived) instance providing access to
1136 ``store.basicstore`` (or derived) instance providing access to
1120 versioned storage.
1137 versioned storage.
1121
1138
1122 cachevfs
1139 cachevfs
1123 ``vfs.vfs`` used for cache files.
1140 ``vfs.vfs`` used for cache files.
1124
1141
1125 wcachevfs
1142 wcachevfs
1126 ``vfs.vfs`` used for cache files related to the working copy.
1143 ``vfs.vfs`` used for cache files related to the working copy.
1127
1144
1128 features
1145 features
1129 ``set`` of bytestrings defining features/capabilities of this
1146 ``set`` of bytestrings defining features/capabilities of this
1130 instance.
1147 instance.
1131
1148
1132 intents
1149 intents
1133 ``set`` of system strings indicating what this repo will be used
1150 ``set`` of system strings indicating what this repo will be used
1134 for.
1151 for.
1135 """
1152 """
1136 self.baseui = baseui
1153 self.baseui = baseui
1137 self.ui = ui
1154 self.ui = ui
1138 self.origroot = origroot
1155 self.origroot = origroot
1139 # vfs rooted at working directory.
1156 # vfs rooted at working directory.
1140 self.wvfs = wdirvfs
1157 self.wvfs = wdirvfs
1141 self.root = wdirvfs.base
1158 self.root = wdirvfs.base
1142 # vfs rooted at .hg/. Used to access most non-store paths.
1159 # vfs rooted at .hg/. Used to access most non-store paths.
1143 self.vfs = hgvfs
1160 self.vfs = hgvfs
1144 self.path = hgvfs.base
1161 self.path = hgvfs.base
1145 self.requirements = requirements
1162 self.requirements = requirements
1146 self.supported = supportedrequirements
1163 self.supported = supportedrequirements
1147 self.sharedpath = sharedpath
1164 self.sharedpath = sharedpath
1148 self.store = store
1165 self.store = store
1149 self.cachevfs = cachevfs
1166 self.cachevfs = cachevfs
1150 self.wcachevfs = wcachevfs
1167 self.wcachevfs = wcachevfs
1151 self.features = features
1168 self.features = features
1152
1169
1153 self.filtername = None
1170 self.filtername = None
1154
1171
1155 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1172 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1156 b'devel', b'check-locks'
1173 b'devel', b'check-locks'
1157 ):
1174 ):
1158 self.vfs.audit = self._getvfsward(self.vfs.audit)
1175 self.vfs.audit = self._getvfsward(self.vfs.audit)
1159 # A list of callback to shape the phase if no data were found.
1176 # A list of callback to shape the phase if no data were found.
1160 # Callback are in the form: func(repo, roots) --> processed root.
1177 # Callback are in the form: func(repo, roots) --> processed root.
1161 # This list it to be filled by extension during repo setup
1178 # This list it to be filled by extension during repo setup
1162 self._phasedefaults = []
1179 self._phasedefaults = []
1163
1180
1164 color.setup(self.ui)
1181 color.setup(self.ui)
1165
1182
1166 self.spath = self.store.path
1183 self.spath = self.store.path
1167 self.svfs = self.store.vfs
1184 self.svfs = self.store.vfs
1168 self.sjoin = self.store.join
1185 self.sjoin = self.store.join
1169 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1186 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
1170 b'devel', b'check-locks'
1187 b'devel', b'check-locks'
1171 ):
1188 ):
1172 if util.safehasattr(self.svfs, b'vfs'): # this is filtervfs
1189 if util.safehasattr(self.svfs, b'vfs'): # this is filtervfs
1173 self.svfs.vfs.audit = self._getsvfsward(self.svfs.vfs.audit)
1190 self.svfs.vfs.audit = self._getsvfsward(self.svfs.vfs.audit)
1174 else: # standard vfs
1191 else: # standard vfs
1175 self.svfs.audit = self._getsvfsward(self.svfs.audit)
1192 self.svfs.audit = self._getsvfsward(self.svfs.audit)
1176
1193
1177 self._dirstatevalidatewarned = False
1194 self._dirstatevalidatewarned = False
1178
1195
1179 self._branchcaches = branchmap.BranchMapCache()
1196 self._branchcaches = branchmap.BranchMapCache()
1180 self._revbranchcache = None
1197 self._revbranchcache = None
1181 self._filterpats = {}
1198 self._filterpats = {}
1182 self._datafilters = {}
1199 self._datafilters = {}
1183 self._transref = self._lockref = self._wlockref = None
1200 self._transref = self._lockref = self._wlockref = None
1184
1201
1185 # A cache for various files under .hg/ that tracks file changes,
1202 # A cache for various files under .hg/ that tracks file changes,
1186 # (used by the filecache decorator)
1203 # (used by the filecache decorator)
1187 #
1204 #
1188 # Maps a property name to its util.filecacheentry
1205 # Maps a property name to its util.filecacheentry
1189 self._filecache = {}
1206 self._filecache = {}
1190
1207
1191 # hold sets of revision to be filtered
1208 # hold sets of revision to be filtered
1192 # should be cleared when something might have changed the filter value:
1209 # should be cleared when something might have changed the filter value:
1193 # - new changesets,
1210 # - new changesets,
1194 # - phase change,
1211 # - phase change,
1195 # - new obsolescence marker,
1212 # - new obsolescence marker,
1196 # - working directory parent change,
1213 # - working directory parent change,
1197 # - bookmark changes
1214 # - bookmark changes
1198 self.filteredrevcache = {}
1215 self.filteredrevcache = {}
1199
1216
1200 # post-dirstate-status hooks
1217 # post-dirstate-status hooks
1201 self._postdsstatus = []
1218 self._postdsstatus = []
1202
1219
1203 # generic mapping between names and nodes
1220 # generic mapping between names and nodes
1204 self.names = namespaces.namespaces()
1221 self.names = namespaces.namespaces()
1205
1222
1206 # Key to signature value.
1223 # Key to signature value.
1207 self._sparsesignaturecache = {}
1224 self._sparsesignaturecache = {}
1208 # Signature to cached matcher instance.
1225 # Signature to cached matcher instance.
1209 self._sparsematchercache = {}
1226 self._sparsematchercache = {}
1210
1227
1211 self._extrafilterid = repoview.extrafilter(ui)
1228 self._extrafilterid = repoview.extrafilter(ui)
1212
1229
1213 self.filecopiesmode = None
1230 self.filecopiesmode = None
1214 if COPIESSDC_REQUIREMENT in self.requirements:
1231 if COPIESSDC_REQUIREMENT in self.requirements:
1215 self.filecopiesmode = b'changeset-sidedata'
1232 self.filecopiesmode = b'changeset-sidedata'
1216
1233
1217 def _getvfsward(self, origfunc):
1234 def _getvfsward(self, origfunc):
1218 """build a ward for self.vfs"""
1235 """build a ward for self.vfs"""
1219 rref = weakref.ref(self)
1236 rref = weakref.ref(self)
1220
1237
1221 def checkvfs(path, mode=None):
1238 def checkvfs(path, mode=None):
1222 ret = origfunc(path, mode=mode)
1239 ret = origfunc(path, mode=mode)
1223 repo = rref()
1240 repo = rref()
1224 if (
1241 if (
1225 repo is None
1242 repo is None
1226 or not util.safehasattr(repo, b'_wlockref')
1243 or not util.safehasattr(repo, b'_wlockref')
1227 or not util.safehasattr(repo, b'_lockref')
1244 or not util.safehasattr(repo, b'_lockref')
1228 ):
1245 ):
1229 return
1246 return
1230 if mode in (None, b'r', b'rb'):
1247 if mode in (None, b'r', b'rb'):
1231 return
1248 return
1232 if path.startswith(repo.path):
1249 if path.startswith(repo.path):
1233 # truncate name relative to the repository (.hg)
1250 # truncate name relative to the repository (.hg)
1234 path = path[len(repo.path) + 1 :]
1251 path = path[len(repo.path) + 1 :]
1235 if path.startswith(b'cache/'):
1252 if path.startswith(b'cache/'):
1236 msg = b'accessing cache with vfs instead of cachevfs: "%s"'
1253 msg = b'accessing cache with vfs instead of cachevfs: "%s"'
1237 repo.ui.develwarn(msg % path, stacklevel=3, config=b"cache-vfs")
1254 repo.ui.develwarn(msg % path, stacklevel=3, config=b"cache-vfs")
1238 # path prefixes covered by 'lock'
1255 # path prefixes covered by 'lock'
1239 vfs_path_prefixes = (b'journal.', b'undo.', b'strip-backup/')
1256 vfs_path_prefixes = (b'journal.', b'undo.', b'strip-backup/')
1240 if any(path.startswith(prefix) for prefix in vfs_path_prefixes):
1257 if any(path.startswith(prefix) for prefix in vfs_path_prefixes):
1241 if repo._currentlock(repo._lockref) is None:
1258 if repo._currentlock(repo._lockref) is None:
1242 repo.ui.develwarn(
1259 repo.ui.develwarn(
1243 b'write with no lock: "%s"' % path,
1260 b'write with no lock: "%s"' % path,
1244 stacklevel=3,
1261 stacklevel=3,
1245 config=b'check-locks',
1262 config=b'check-locks',
1246 )
1263 )
1247 elif repo._currentlock(repo._wlockref) is None:
1264 elif repo._currentlock(repo._wlockref) is None:
1248 # rest of vfs files are covered by 'wlock'
1265 # rest of vfs files are covered by 'wlock'
1249 #
1266 #
1250 # exclude special files
1267 # exclude special files
1251 for prefix in self._wlockfreeprefix:
1268 for prefix in self._wlockfreeprefix:
1252 if path.startswith(prefix):
1269 if path.startswith(prefix):
1253 return
1270 return
1254 repo.ui.develwarn(
1271 repo.ui.develwarn(
1255 b'write with no wlock: "%s"' % path,
1272 b'write with no wlock: "%s"' % path,
1256 stacklevel=3,
1273 stacklevel=3,
1257 config=b'check-locks',
1274 config=b'check-locks',
1258 )
1275 )
1259 return ret
1276 return ret
1260
1277
1261 return checkvfs
1278 return checkvfs
1262
1279
1263 def _getsvfsward(self, origfunc):
1280 def _getsvfsward(self, origfunc):
1264 """build a ward for self.svfs"""
1281 """build a ward for self.svfs"""
1265 rref = weakref.ref(self)
1282 rref = weakref.ref(self)
1266
1283
1267 def checksvfs(path, mode=None):
1284 def checksvfs(path, mode=None):
1268 ret = origfunc(path, mode=mode)
1285 ret = origfunc(path, mode=mode)
1269 repo = rref()
1286 repo = rref()
1270 if repo is None or not util.safehasattr(repo, b'_lockref'):
1287 if repo is None or not util.safehasattr(repo, b'_lockref'):
1271 return
1288 return
1272 if mode in (None, b'r', b'rb'):
1289 if mode in (None, b'r', b'rb'):
1273 return
1290 return
1274 if path.startswith(repo.sharedpath):
1291 if path.startswith(repo.sharedpath):
1275 # truncate name relative to the repository (.hg)
1292 # truncate name relative to the repository (.hg)
1276 path = path[len(repo.sharedpath) + 1 :]
1293 path = path[len(repo.sharedpath) + 1 :]
1277 if repo._currentlock(repo._lockref) is None:
1294 if repo._currentlock(repo._lockref) is None:
1278 repo.ui.develwarn(
1295 repo.ui.develwarn(
1279 b'write with no lock: "%s"' % path, stacklevel=4
1296 b'write with no lock: "%s"' % path, stacklevel=4
1280 )
1297 )
1281 return ret
1298 return ret
1282
1299
1283 return checksvfs
1300 return checksvfs
1284
1301
1285 def close(self):
1302 def close(self):
1286 self._writecaches()
1303 self._writecaches()
1287
1304
1288 def _writecaches(self):
1305 def _writecaches(self):
1289 if self._revbranchcache:
1306 if self._revbranchcache:
1290 self._revbranchcache.write()
1307 self._revbranchcache.write()
1291
1308
1292 def _restrictcapabilities(self, caps):
1309 def _restrictcapabilities(self, caps):
1293 if self.ui.configbool(b'experimental', b'bundle2-advertise'):
1310 if self.ui.configbool(b'experimental', b'bundle2-advertise'):
1294 caps = set(caps)
1311 caps = set(caps)
1295 capsblob = bundle2.encodecaps(
1312 capsblob = bundle2.encodecaps(
1296 bundle2.getrepocaps(self, role=b'client')
1313 bundle2.getrepocaps(self, role=b'client')
1297 )
1314 )
1298 caps.add(b'bundle2=' + urlreq.quote(capsblob))
1315 caps.add(b'bundle2=' + urlreq.quote(capsblob))
1299 return caps
1316 return caps
1300
1317
1301 # Don't cache auditor/nofsauditor, or you'll end up with reference cycle:
1318 # Don't cache auditor/nofsauditor, or you'll end up with reference cycle:
1302 # self -> auditor -> self._checknested -> self
1319 # self -> auditor -> self._checknested -> self
1303
1320
1304 @property
1321 @property
1305 def auditor(self):
1322 def auditor(self):
1306 # This is only used by context.workingctx.match in order to
1323 # This is only used by context.workingctx.match in order to
1307 # detect files in subrepos.
1324 # detect files in subrepos.
1308 return pathutil.pathauditor(self.root, callback=self._checknested)
1325 return pathutil.pathauditor(self.root, callback=self._checknested)
1309
1326
1310 @property
1327 @property
1311 def nofsauditor(self):
1328 def nofsauditor(self):
1312 # This is only used by context.basectx.match in order to detect
1329 # This is only used by context.basectx.match in order to detect
1313 # files in subrepos.
1330 # files in subrepos.
1314 return pathutil.pathauditor(
1331 return pathutil.pathauditor(
1315 self.root, callback=self._checknested, realfs=False, cached=True
1332 self.root, callback=self._checknested, realfs=False, cached=True
1316 )
1333 )
1317
1334
1318 def _checknested(self, path):
1335 def _checknested(self, path):
1319 """Determine if path is a legal nested repository."""
1336 """Determine if path is a legal nested repository."""
1320 if not path.startswith(self.root):
1337 if not path.startswith(self.root):
1321 return False
1338 return False
1322 subpath = path[len(self.root) + 1 :]
1339 subpath = path[len(self.root) + 1 :]
1323 normsubpath = util.pconvert(subpath)
1340 normsubpath = util.pconvert(subpath)
1324
1341
1325 # XXX: Checking against the current working copy is wrong in
1342 # XXX: Checking against the current working copy is wrong in
1326 # the sense that it can reject things like
1343 # the sense that it can reject things like
1327 #
1344 #
1328 # $ hg cat -r 10 sub/x.txt
1345 # $ hg cat -r 10 sub/x.txt
1329 #
1346 #
1330 # if sub/ is no longer a subrepository in the working copy
1347 # if sub/ is no longer a subrepository in the working copy
1331 # parent revision.
1348 # parent revision.
1332 #
1349 #
1333 # However, it can of course also allow things that would have
1350 # However, it can of course also allow things that would have
1334 # been rejected before, such as the above cat command if sub/
1351 # been rejected before, such as the above cat command if sub/
1335 # is a subrepository now, but was a normal directory before.
1352 # is a subrepository now, but was a normal directory before.
1336 # The old path auditor would have rejected by mistake since it
1353 # The old path auditor would have rejected by mistake since it
1337 # panics when it sees sub/.hg/.
1354 # panics when it sees sub/.hg/.
1338 #
1355 #
1339 # All in all, checking against the working copy seems sensible
1356 # All in all, checking against the working copy seems sensible
1340 # since we want to prevent access to nested repositories on
1357 # since we want to prevent access to nested repositories on
1341 # the filesystem *now*.
1358 # the filesystem *now*.
1342 ctx = self[None]
1359 ctx = self[None]
1343 parts = util.splitpath(subpath)
1360 parts = util.splitpath(subpath)
1344 while parts:
1361 while parts:
1345 prefix = b'/'.join(parts)
1362 prefix = b'/'.join(parts)
1346 if prefix in ctx.substate:
1363 if prefix in ctx.substate:
1347 if prefix == normsubpath:
1364 if prefix == normsubpath:
1348 return True
1365 return True
1349 else:
1366 else:
1350 sub = ctx.sub(prefix)
1367 sub = ctx.sub(prefix)
1351 return sub.checknested(subpath[len(prefix) + 1 :])
1368 return sub.checknested(subpath[len(prefix) + 1 :])
1352 else:
1369 else:
1353 parts.pop()
1370 parts.pop()
1354 return False
1371 return False
1355
1372
1356 def peer(self):
1373 def peer(self):
1357 return localpeer(self) # not cached to avoid reference cycle
1374 return localpeer(self) # not cached to avoid reference cycle
1358
1375
1359 def unfiltered(self):
1376 def unfiltered(self):
1360 """Return unfiltered version of the repository
1377 """Return unfiltered version of the repository
1361
1378
1362 Intended to be overwritten by filtered repo."""
1379 Intended to be overwritten by filtered repo."""
1363 return self
1380 return self
1364
1381
1365 def filtered(self, name, visibilityexceptions=None):
1382 def filtered(self, name, visibilityexceptions=None):
1366 """Return a filtered version of a repository
1383 """Return a filtered version of a repository
1367
1384
1368 The `name` parameter is the identifier of the requested view. This
1385 The `name` parameter is the identifier of the requested view. This
1369 will return a repoview object set "exactly" to the specified view.
1386 will return a repoview object set "exactly" to the specified view.
1370
1387
1371 This function does not apply recursive filtering to a repository. For
1388 This function does not apply recursive filtering to a repository. For
1372 example calling `repo.filtered("served")` will return a repoview using
1389 example calling `repo.filtered("served")` will return a repoview using
1373 the "served" view, regardless of the initial view used by `repo`.
1390 the "served" view, regardless of the initial view used by `repo`.
1374
1391
1375 In other word, there is always only one level of `repoview` "filtering".
1392 In other word, there is always only one level of `repoview` "filtering".
1376 """
1393 """
1377 if self._extrafilterid is not None and b'%' not in name:
1394 if self._extrafilterid is not None and b'%' not in name:
1378 name = name + b'%' + self._extrafilterid
1395 name = name + b'%' + self._extrafilterid
1379
1396
1380 cls = repoview.newtype(self.unfiltered().__class__)
1397 cls = repoview.newtype(self.unfiltered().__class__)
1381 return cls(self, name, visibilityexceptions)
1398 return cls(self, name, visibilityexceptions)
1382
1399
1383 @mixedrepostorecache(
1400 @mixedrepostorecache(
1384 (b'bookmarks', b'plain'),
1401 (b'bookmarks', b'plain'),
1385 (b'bookmarks.current', b'plain'),
1402 (b'bookmarks.current', b'plain'),
1386 (b'bookmarks', b''),
1403 (b'bookmarks', b''),
1387 (b'00changelog.i', b''),
1404 (b'00changelog.i', b''),
1388 )
1405 )
1389 def _bookmarks(self):
1406 def _bookmarks(self):
1390 # Since the multiple files involved in the transaction cannot be
1407 # Since the multiple files involved in the transaction cannot be
1391 # written atomically (with current repository format), there is a race
1408 # written atomically (with current repository format), there is a race
1392 # condition here.
1409 # condition here.
1393 #
1410 #
1394 # 1) changelog content A is read
1411 # 1) changelog content A is read
1395 # 2) outside transaction update changelog to content B
1412 # 2) outside transaction update changelog to content B
1396 # 3) outside transaction update bookmark file referring to content B
1413 # 3) outside transaction update bookmark file referring to content B
1397 # 4) bookmarks file content is read and filtered against changelog-A
1414 # 4) bookmarks file content is read and filtered against changelog-A
1398 #
1415 #
1399 # When this happens, bookmarks against nodes missing from A are dropped.
1416 # When this happens, bookmarks against nodes missing from A are dropped.
1400 #
1417 #
1401 # Having this happening during read is not great, but it become worse
1418 # Having this happening during read is not great, but it become worse
1402 # when this happen during write because the bookmarks to the "unknown"
1419 # when this happen during write because the bookmarks to the "unknown"
1403 # nodes will be dropped for good. However, writes happen within locks.
1420 # nodes will be dropped for good. However, writes happen within locks.
1404 # This locking makes it possible to have a race free consistent read.
1421 # This locking makes it possible to have a race free consistent read.
1405 # For this purpose data read from disc before locking are
1422 # For this purpose data read from disc before locking are
1406 # "invalidated" right after the locks are taken. This invalidations are
1423 # "invalidated" right after the locks are taken. This invalidations are
1407 # "light", the `filecache` mechanism keep the data in memory and will
1424 # "light", the `filecache` mechanism keep the data in memory and will
1408 # reuse them if the underlying files did not changed. Not parsing the
1425 # reuse them if the underlying files did not changed. Not parsing the
1409 # same data multiple times helps performances.
1426 # same data multiple times helps performances.
1410 #
1427 #
1411 # Unfortunately in the case describe above, the files tracked by the
1428 # Unfortunately in the case describe above, the files tracked by the
1412 # bookmarks file cache might not have changed, but the in-memory
1429 # bookmarks file cache might not have changed, but the in-memory
1413 # content is still "wrong" because we used an older changelog content
1430 # content is still "wrong" because we used an older changelog content
1414 # to process the on-disk data. So after locking, the changelog would be
1431 # to process the on-disk data. So after locking, the changelog would be
1415 # refreshed but `_bookmarks` would be preserved.
1432 # refreshed but `_bookmarks` would be preserved.
1416 # Adding `00changelog.i` to the list of tracked file is not
1433 # Adding `00changelog.i` to the list of tracked file is not
1417 # enough, because at the time we build the content for `_bookmarks` in
1434 # enough, because at the time we build the content for `_bookmarks` in
1418 # (4), the changelog file has already diverged from the content used
1435 # (4), the changelog file has already diverged from the content used
1419 # for loading `changelog` in (1)
1436 # for loading `changelog` in (1)
1420 #
1437 #
1421 # To prevent the issue, we force the changelog to be explicitly
1438 # To prevent the issue, we force the changelog to be explicitly
1422 # reloaded while computing `_bookmarks`. The data race can still happen
1439 # reloaded while computing `_bookmarks`. The data race can still happen
1423 # without the lock (with a narrower window), but it would no longer go
1440 # without the lock (with a narrower window), but it would no longer go
1424 # undetected during the lock time refresh.
1441 # undetected during the lock time refresh.
1425 #
1442 #
1426 # The new schedule is as follow
1443 # The new schedule is as follow
1427 #
1444 #
1428 # 1) filecache logic detect that `_bookmarks` needs to be computed
1445 # 1) filecache logic detect that `_bookmarks` needs to be computed
1429 # 2) cachestat for `bookmarks` and `changelog` are captured (for book)
1446 # 2) cachestat for `bookmarks` and `changelog` are captured (for book)
1430 # 3) We force `changelog` filecache to be tested
1447 # 3) We force `changelog` filecache to be tested
1431 # 4) cachestat for `changelog` are captured (for changelog)
1448 # 4) cachestat for `changelog` are captured (for changelog)
1432 # 5) `_bookmarks` is computed and cached
1449 # 5) `_bookmarks` is computed and cached
1433 #
1450 #
1434 # The step in (3) ensure we have a changelog at least as recent as the
1451 # The step in (3) ensure we have a changelog at least as recent as the
1435 # cache stat computed in (1). As a result at locking time:
1452 # cache stat computed in (1). As a result at locking time:
1436 # * if the changelog did not changed since (1) -> we can reuse the data
1453 # * if the changelog did not changed since (1) -> we can reuse the data
1437 # * otherwise -> the bookmarks get refreshed.
1454 # * otherwise -> the bookmarks get refreshed.
1438 self._refreshchangelog()
1455 self._refreshchangelog()
1439 return bookmarks.bmstore(self)
1456 return bookmarks.bmstore(self)
1440
1457
1441 def _refreshchangelog(self):
1458 def _refreshchangelog(self):
1442 """make sure the in memory changelog match the on-disk one"""
1459 """make sure the in memory changelog match the on-disk one"""
1443 if 'changelog' in vars(self) and self.currenttransaction() is None:
1460 if 'changelog' in vars(self) and self.currenttransaction() is None:
1444 del self.changelog
1461 del self.changelog
1445
1462
1446 @property
1463 @property
1447 def _activebookmark(self):
1464 def _activebookmark(self):
1448 return self._bookmarks.active
1465 return self._bookmarks.active
1449
1466
1450 # _phasesets depend on changelog. what we need is to call
1467 # _phasesets depend on changelog. what we need is to call
1451 # _phasecache.invalidate() if '00changelog.i' was changed, but it
1468 # _phasecache.invalidate() if '00changelog.i' was changed, but it
1452 # can't be easily expressed in filecache mechanism.
1469 # can't be easily expressed in filecache mechanism.
1453 @storecache(b'phaseroots', b'00changelog.i')
1470 @storecache(b'phaseroots', b'00changelog.i')
1454 def _phasecache(self):
1471 def _phasecache(self):
1455 return phases.phasecache(self, self._phasedefaults)
1472 return phases.phasecache(self, self._phasedefaults)
1456
1473
1457 @storecache(b'obsstore')
1474 @storecache(b'obsstore')
1458 def obsstore(self):
1475 def obsstore(self):
1459 return obsolete.makestore(self.ui, self)
1476 return obsolete.makestore(self.ui, self)
1460
1477
1461 @storecache(b'00changelog.i')
1478 @storecache(b'00changelog.i')
1462 def changelog(self):
1479 def changelog(self):
1463 # load dirstate before changelog to avoid race see issue6303
1480 # load dirstate before changelog to avoid race see issue6303
1464 self.dirstate.prefetch_parents()
1481 self.dirstate.prefetch_parents()
1465 return self.store.changelog(txnutil.mayhavepending(self.root))
1482 return self.store.changelog(txnutil.mayhavepending(self.root))
1466
1483
1467 @storecache(b'00manifest.i')
1484 @storecache(b'00manifest.i')
1468 def manifestlog(self):
1485 def manifestlog(self):
1469 return self.store.manifestlog(self, self._storenarrowmatch)
1486 return self.store.manifestlog(self, self._storenarrowmatch)
1470
1487
1471 @repofilecache(b'dirstate')
1488 @repofilecache(b'dirstate')
1472 def dirstate(self):
1489 def dirstate(self):
1473 return self._makedirstate()
1490 return self._makedirstate()
1474
1491
1475 def _makedirstate(self):
1492 def _makedirstate(self):
1476 """Extension point for wrapping the dirstate per-repo."""
1493 """Extension point for wrapping the dirstate per-repo."""
1477 sparsematchfn = lambda: sparse.matcher(self)
1494 sparsematchfn = lambda: sparse.matcher(self)
1478
1495
1479 return dirstate.dirstate(
1496 return dirstate.dirstate(
1480 self.vfs, self.ui, self.root, self._dirstatevalidate, sparsematchfn
1497 self.vfs, self.ui, self.root, self._dirstatevalidate, sparsematchfn
1481 )
1498 )
1482
1499
1483 def _dirstatevalidate(self, node):
1500 def _dirstatevalidate(self, node):
1484 try:
1501 try:
1485 self.changelog.rev(node)
1502 self.changelog.rev(node)
1486 return node
1503 return node
1487 except error.LookupError:
1504 except error.LookupError:
1488 if not self._dirstatevalidatewarned:
1505 if not self._dirstatevalidatewarned:
1489 self._dirstatevalidatewarned = True
1506 self._dirstatevalidatewarned = True
1490 self.ui.warn(
1507 self.ui.warn(
1491 _(b"warning: ignoring unknown working parent %s!\n")
1508 _(b"warning: ignoring unknown working parent %s!\n")
1492 % short(node)
1509 % short(node)
1493 )
1510 )
1494 return nullid
1511 return nullid
1495
1512
1496 @storecache(narrowspec.FILENAME)
1513 @storecache(narrowspec.FILENAME)
1497 def narrowpats(self):
1514 def narrowpats(self):
1498 """matcher patterns for this repository's narrowspec
1515 """matcher patterns for this repository's narrowspec
1499
1516
1500 A tuple of (includes, excludes).
1517 A tuple of (includes, excludes).
1501 """
1518 """
1502 return narrowspec.load(self)
1519 return narrowspec.load(self)
1503
1520
1504 @storecache(narrowspec.FILENAME)
1521 @storecache(narrowspec.FILENAME)
1505 def _storenarrowmatch(self):
1522 def _storenarrowmatch(self):
1506 if repository.NARROW_REQUIREMENT not in self.requirements:
1523 if repository.NARROW_REQUIREMENT not in self.requirements:
1507 return matchmod.always()
1524 return matchmod.always()
1508 include, exclude = self.narrowpats
1525 include, exclude = self.narrowpats
1509 return narrowspec.match(self.root, include=include, exclude=exclude)
1526 return narrowspec.match(self.root, include=include, exclude=exclude)
1510
1527
1511 @storecache(narrowspec.FILENAME)
1528 @storecache(narrowspec.FILENAME)
1512 def _narrowmatch(self):
1529 def _narrowmatch(self):
1513 if repository.NARROW_REQUIREMENT not in self.requirements:
1530 if repository.NARROW_REQUIREMENT not in self.requirements:
1514 return matchmod.always()
1531 return matchmod.always()
1515 narrowspec.checkworkingcopynarrowspec(self)
1532 narrowspec.checkworkingcopynarrowspec(self)
1516 include, exclude = self.narrowpats
1533 include, exclude = self.narrowpats
1517 return narrowspec.match(self.root, include=include, exclude=exclude)
1534 return narrowspec.match(self.root, include=include, exclude=exclude)
1518
1535
1519 def narrowmatch(self, match=None, includeexact=False):
1536 def narrowmatch(self, match=None, includeexact=False):
1520 """matcher corresponding the the repo's narrowspec
1537 """matcher corresponding the the repo's narrowspec
1521
1538
1522 If `match` is given, then that will be intersected with the narrow
1539 If `match` is given, then that will be intersected with the narrow
1523 matcher.
1540 matcher.
1524
1541
1525 If `includeexact` is True, then any exact matches from `match` will
1542 If `includeexact` is True, then any exact matches from `match` will
1526 be included even if they're outside the narrowspec.
1543 be included even if they're outside the narrowspec.
1527 """
1544 """
1528 if match:
1545 if match:
1529 if includeexact and not self._narrowmatch.always():
1546 if includeexact and not self._narrowmatch.always():
1530 # do not exclude explicitly-specified paths so that they can
1547 # do not exclude explicitly-specified paths so that they can
1531 # be warned later on
1548 # be warned later on
1532 em = matchmod.exact(match.files())
1549 em = matchmod.exact(match.files())
1533 nm = matchmod.unionmatcher([self._narrowmatch, em])
1550 nm = matchmod.unionmatcher([self._narrowmatch, em])
1534 return matchmod.intersectmatchers(match, nm)
1551 return matchmod.intersectmatchers(match, nm)
1535 return matchmod.intersectmatchers(match, self._narrowmatch)
1552 return matchmod.intersectmatchers(match, self._narrowmatch)
1536 return self._narrowmatch
1553 return self._narrowmatch
1537
1554
1538 def setnarrowpats(self, newincludes, newexcludes):
1555 def setnarrowpats(self, newincludes, newexcludes):
1539 narrowspec.save(self, newincludes, newexcludes)
1556 narrowspec.save(self, newincludes, newexcludes)
1540 self.invalidate(clearfilecache=True)
1557 self.invalidate(clearfilecache=True)
1541
1558
1542 @unfilteredpropertycache
1559 @unfilteredpropertycache
1543 def _quick_access_changeid_null(self):
1560 def _quick_access_changeid_null(self):
1544 return {
1561 return {
1545 b'null': (nullrev, nullid),
1562 b'null': (nullrev, nullid),
1546 nullrev: (nullrev, nullid),
1563 nullrev: (nullrev, nullid),
1547 nullid: (nullrev, nullid),
1564 nullid: (nullrev, nullid),
1548 }
1565 }
1549
1566
1550 @unfilteredpropertycache
1567 @unfilteredpropertycache
1551 def _quick_access_changeid_wc(self):
1568 def _quick_access_changeid_wc(self):
1552 # also fast path access to the working copy parents
1569 # also fast path access to the working copy parents
1553 # however, only do it for filter that ensure wc is visible.
1570 # however, only do it for filter that ensure wc is visible.
1554 quick = {}
1571 quick = {}
1555 cl = self.unfiltered().changelog
1572 cl = self.unfiltered().changelog
1556 for node in self.dirstate.parents():
1573 for node in self.dirstate.parents():
1557 if node == nullid:
1574 if node == nullid:
1558 continue
1575 continue
1559 rev = cl.index.get_rev(node)
1576 rev = cl.index.get_rev(node)
1560 if rev is None:
1577 if rev is None:
1561 # unknown working copy parent case:
1578 # unknown working copy parent case:
1562 #
1579 #
1563 # skip the fast path and let higher code deal with it
1580 # skip the fast path and let higher code deal with it
1564 continue
1581 continue
1565 pair = (rev, node)
1582 pair = (rev, node)
1566 quick[rev] = pair
1583 quick[rev] = pair
1567 quick[node] = pair
1584 quick[node] = pair
1568 # also add the parents of the parents
1585 # also add the parents of the parents
1569 for r in cl.parentrevs(rev):
1586 for r in cl.parentrevs(rev):
1570 if r == nullrev:
1587 if r == nullrev:
1571 continue
1588 continue
1572 n = cl.node(r)
1589 n = cl.node(r)
1573 pair = (r, n)
1590 pair = (r, n)
1574 quick[r] = pair
1591 quick[r] = pair
1575 quick[n] = pair
1592 quick[n] = pair
1576 p1node = self.dirstate.p1()
1593 p1node = self.dirstate.p1()
1577 if p1node != nullid:
1594 if p1node != nullid:
1578 quick[b'.'] = quick[p1node]
1595 quick[b'.'] = quick[p1node]
1579 return quick
1596 return quick
1580
1597
1581 @unfilteredmethod
1598 @unfilteredmethod
1582 def _quick_access_changeid_invalidate(self):
1599 def _quick_access_changeid_invalidate(self):
1583 if '_quick_access_changeid_wc' in vars(self):
1600 if '_quick_access_changeid_wc' in vars(self):
1584 del self.__dict__['_quick_access_changeid_wc']
1601 del self.__dict__['_quick_access_changeid_wc']
1585
1602
1586 @property
1603 @property
1587 def _quick_access_changeid(self):
1604 def _quick_access_changeid(self):
1588 """an helper dictionnary for __getitem__ calls
1605 """an helper dictionnary for __getitem__ calls
1589
1606
1590 This contains a list of symbol we can recognise right away without
1607 This contains a list of symbol we can recognise right away without
1591 further processing.
1608 further processing.
1592 """
1609 """
1593 mapping = self._quick_access_changeid_null
1610 mapping = self._quick_access_changeid_null
1594 if self.filtername in repoview.filter_has_wc:
1611 if self.filtername in repoview.filter_has_wc:
1595 mapping = mapping.copy()
1612 mapping = mapping.copy()
1596 mapping.update(self._quick_access_changeid_wc)
1613 mapping.update(self._quick_access_changeid_wc)
1597 return mapping
1614 return mapping
1598
1615
1599 def __getitem__(self, changeid):
1616 def __getitem__(self, changeid):
1600 # dealing with special cases
1617 # dealing with special cases
1601 if changeid is None:
1618 if changeid is None:
1602 return context.workingctx(self)
1619 return context.workingctx(self)
1603 if isinstance(changeid, context.basectx):
1620 if isinstance(changeid, context.basectx):
1604 return changeid
1621 return changeid
1605
1622
1606 # dealing with multiple revisions
1623 # dealing with multiple revisions
1607 if isinstance(changeid, slice):
1624 if isinstance(changeid, slice):
1608 # wdirrev isn't contiguous so the slice shouldn't include it
1625 # wdirrev isn't contiguous so the slice shouldn't include it
1609 return [
1626 return [
1610 self[i]
1627 self[i]
1611 for i in pycompat.xrange(*changeid.indices(len(self)))
1628 for i in pycompat.xrange(*changeid.indices(len(self)))
1612 if i not in self.changelog.filteredrevs
1629 if i not in self.changelog.filteredrevs
1613 ]
1630 ]
1614
1631
1615 # dealing with some special values
1632 # dealing with some special values
1616 quick_access = self._quick_access_changeid.get(changeid)
1633 quick_access = self._quick_access_changeid.get(changeid)
1617 if quick_access is not None:
1634 if quick_access is not None:
1618 rev, node = quick_access
1635 rev, node = quick_access
1619 return context.changectx(self, rev, node, maybe_filtered=False)
1636 return context.changectx(self, rev, node, maybe_filtered=False)
1620 if changeid == b'tip':
1637 if changeid == b'tip':
1621 node = self.changelog.tip()
1638 node = self.changelog.tip()
1622 rev = self.changelog.rev(node)
1639 rev = self.changelog.rev(node)
1623 return context.changectx(self, rev, node)
1640 return context.changectx(self, rev, node)
1624
1641
1625 # dealing with arbitrary values
1642 # dealing with arbitrary values
1626 try:
1643 try:
1627 if isinstance(changeid, int):
1644 if isinstance(changeid, int):
1628 node = self.changelog.node(changeid)
1645 node = self.changelog.node(changeid)
1629 rev = changeid
1646 rev = changeid
1630 elif changeid == b'.':
1647 elif changeid == b'.':
1631 # this is a hack to delay/avoid loading obsmarkers
1648 # this is a hack to delay/avoid loading obsmarkers
1632 # when we know that '.' won't be hidden
1649 # when we know that '.' won't be hidden
1633 node = self.dirstate.p1()
1650 node = self.dirstate.p1()
1634 rev = self.unfiltered().changelog.rev(node)
1651 rev = self.unfiltered().changelog.rev(node)
1635 elif len(changeid) == 20:
1652 elif len(changeid) == 20:
1636 try:
1653 try:
1637 node = changeid
1654 node = changeid
1638 rev = self.changelog.rev(changeid)
1655 rev = self.changelog.rev(changeid)
1639 except error.FilteredLookupError:
1656 except error.FilteredLookupError:
1640 changeid = hex(changeid) # for the error message
1657 changeid = hex(changeid) # for the error message
1641 raise
1658 raise
1642 except LookupError:
1659 except LookupError:
1643 # check if it might have come from damaged dirstate
1660 # check if it might have come from damaged dirstate
1644 #
1661 #
1645 # XXX we could avoid the unfiltered if we had a recognizable
1662 # XXX we could avoid the unfiltered if we had a recognizable
1646 # exception for filtered changeset access
1663 # exception for filtered changeset access
1647 if (
1664 if (
1648 self.local()
1665 self.local()
1649 and changeid in self.unfiltered().dirstate.parents()
1666 and changeid in self.unfiltered().dirstate.parents()
1650 ):
1667 ):
1651 msg = _(b"working directory has unknown parent '%s'!")
1668 msg = _(b"working directory has unknown parent '%s'!")
1652 raise error.Abort(msg % short(changeid))
1669 raise error.Abort(msg % short(changeid))
1653 changeid = hex(changeid) # for the error message
1670 changeid = hex(changeid) # for the error message
1654 raise
1671 raise
1655
1672
1656 elif len(changeid) == 40:
1673 elif len(changeid) == 40:
1657 node = bin(changeid)
1674 node = bin(changeid)
1658 rev = self.changelog.rev(node)
1675 rev = self.changelog.rev(node)
1659 else:
1676 else:
1660 raise error.ProgrammingError(
1677 raise error.ProgrammingError(
1661 b"unsupported changeid '%s' of type %s"
1678 b"unsupported changeid '%s' of type %s"
1662 % (changeid, pycompat.bytestr(type(changeid)))
1679 % (changeid, pycompat.bytestr(type(changeid)))
1663 )
1680 )
1664
1681
1665 return context.changectx(self, rev, node)
1682 return context.changectx(self, rev, node)
1666
1683
1667 except (error.FilteredIndexError, error.FilteredLookupError):
1684 except (error.FilteredIndexError, error.FilteredLookupError):
1668 raise error.FilteredRepoLookupError(
1685 raise error.FilteredRepoLookupError(
1669 _(b"filtered revision '%s'") % pycompat.bytestr(changeid)
1686 _(b"filtered revision '%s'") % pycompat.bytestr(changeid)
1670 )
1687 )
1671 except (IndexError, LookupError):
1688 except (IndexError, LookupError):
1672 raise error.RepoLookupError(
1689 raise error.RepoLookupError(
1673 _(b"unknown revision '%s'") % pycompat.bytestr(changeid)
1690 _(b"unknown revision '%s'") % pycompat.bytestr(changeid)
1674 )
1691 )
1675 except error.WdirUnsupported:
1692 except error.WdirUnsupported:
1676 return context.workingctx(self)
1693 return context.workingctx(self)
1677
1694
1678 def __contains__(self, changeid):
1695 def __contains__(self, changeid):
1679 """True if the given changeid exists
1696 """True if the given changeid exists
1680
1697
1681 error.AmbiguousPrefixLookupError is raised if an ambiguous node
1698 error.AmbiguousPrefixLookupError is raised if an ambiguous node
1682 specified.
1699 specified.
1683 """
1700 """
1684 try:
1701 try:
1685 self[changeid]
1702 self[changeid]
1686 return True
1703 return True
1687 except error.RepoLookupError:
1704 except error.RepoLookupError:
1688 return False
1705 return False
1689
1706
1690 def __nonzero__(self):
1707 def __nonzero__(self):
1691 return True
1708 return True
1692
1709
1693 __bool__ = __nonzero__
1710 __bool__ = __nonzero__
1694
1711
1695 def __len__(self):
1712 def __len__(self):
1696 # no need to pay the cost of repoview.changelog
1713 # no need to pay the cost of repoview.changelog
1697 unfi = self.unfiltered()
1714 unfi = self.unfiltered()
1698 return len(unfi.changelog)
1715 return len(unfi.changelog)
1699
1716
1700 def __iter__(self):
1717 def __iter__(self):
1701 return iter(self.changelog)
1718 return iter(self.changelog)
1702
1719
1703 def revs(self, expr, *args):
1720 def revs(self, expr, *args):
1704 '''Find revisions matching a revset.
1721 '''Find revisions matching a revset.
1705
1722
1706 The revset is specified as a string ``expr`` that may contain
1723 The revset is specified as a string ``expr`` that may contain
1707 %-formatting to escape certain types. See ``revsetlang.formatspec``.
1724 %-formatting to escape certain types. See ``revsetlang.formatspec``.
1708
1725
1709 Revset aliases from the configuration are not expanded. To expand
1726 Revset aliases from the configuration are not expanded. To expand
1710 user aliases, consider calling ``scmutil.revrange()`` or
1727 user aliases, consider calling ``scmutil.revrange()`` or
1711 ``repo.anyrevs([expr], user=True)``.
1728 ``repo.anyrevs([expr], user=True)``.
1712
1729
1713 Returns a smartset.abstractsmartset, which is a list-like interface
1730 Returns a smartset.abstractsmartset, which is a list-like interface
1714 that contains integer revisions.
1731 that contains integer revisions.
1715 '''
1732 '''
1716 tree = revsetlang.spectree(expr, *args)
1733 tree = revsetlang.spectree(expr, *args)
1717 return revset.makematcher(tree)(self)
1734 return revset.makematcher(tree)(self)
1718
1735
1719 def set(self, expr, *args):
1736 def set(self, expr, *args):
1720 '''Find revisions matching a revset and emit changectx instances.
1737 '''Find revisions matching a revset and emit changectx instances.
1721
1738
1722 This is a convenience wrapper around ``revs()`` that iterates the
1739 This is a convenience wrapper around ``revs()`` that iterates the
1723 result and is a generator of changectx instances.
1740 result and is a generator of changectx instances.
1724
1741
1725 Revset aliases from the configuration are not expanded. To expand
1742 Revset aliases from the configuration are not expanded. To expand
1726 user aliases, consider calling ``scmutil.revrange()``.
1743 user aliases, consider calling ``scmutil.revrange()``.
1727 '''
1744 '''
1728 for r in self.revs(expr, *args):
1745 for r in self.revs(expr, *args):
1729 yield self[r]
1746 yield self[r]
1730
1747
1731 def anyrevs(self, specs, user=False, localalias=None):
1748 def anyrevs(self, specs, user=False, localalias=None):
1732 '''Find revisions matching one of the given revsets.
1749 '''Find revisions matching one of the given revsets.
1733
1750
1734 Revset aliases from the configuration are not expanded by default. To
1751 Revset aliases from the configuration are not expanded by default. To
1735 expand user aliases, specify ``user=True``. To provide some local
1752 expand user aliases, specify ``user=True``. To provide some local
1736 definitions overriding user aliases, set ``localalias`` to
1753 definitions overriding user aliases, set ``localalias`` to
1737 ``{name: definitionstring}``.
1754 ``{name: definitionstring}``.
1738 '''
1755 '''
1739 if specs == [b'null']:
1756 if specs == [b'null']:
1740 return revset.baseset([nullrev])
1757 return revset.baseset([nullrev])
1741 if specs == [b'.']:
1758 if specs == [b'.']:
1742 quick_data = self._quick_access_changeid.get(b'.')
1759 quick_data = self._quick_access_changeid.get(b'.')
1743 if quick_data is not None:
1760 if quick_data is not None:
1744 return revset.baseset([quick_data[0]])
1761 return revset.baseset([quick_data[0]])
1745 if user:
1762 if user:
1746 m = revset.matchany(
1763 m = revset.matchany(
1747 self.ui,
1764 self.ui,
1748 specs,
1765 specs,
1749 lookup=revset.lookupfn(self),
1766 lookup=revset.lookupfn(self),
1750 localalias=localalias,
1767 localalias=localalias,
1751 )
1768 )
1752 else:
1769 else:
1753 m = revset.matchany(None, specs, localalias=localalias)
1770 m = revset.matchany(None, specs, localalias=localalias)
1754 return m(self)
1771 return m(self)
1755
1772
1756 def url(self):
1773 def url(self):
1757 return b'file:' + self.root
1774 return b'file:' + self.root
1758
1775
1759 def hook(self, name, throw=False, **args):
1776 def hook(self, name, throw=False, **args):
1760 """Call a hook, passing this repo instance.
1777 """Call a hook, passing this repo instance.
1761
1778
1762 This a convenience method to aid invoking hooks. Extensions likely
1779 This a convenience method to aid invoking hooks. Extensions likely
1763 won't call this unless they have registered a custom hook or are
1780 won't call this unless they have registered a custom hook or are
1764 replacing code that is expected to call a hook.
1781 replacing code that is expected to call a hook.
1765 """
1782 """
1766 return hook.hook(self.ui, self, name, throw, **args)
1783 return hook.hook(self.ui, self, name, throw, **args)
1767
1784
1768 @filteredpropertycache
1785 @filteredpropertycache
1769 def _tagscache(self):
1786 def _tagscache(self):
1770 '''Returns a tagscache object that contains various tags related
1787 '''Returns a tagscache object that contains various tags related
1771 caches.'''
1788 caches.'''
1772
1789
1773 # This simplifies its cache management by having one decorated
1790 # This simplifies its cache management by having one decorated
1774 # function (this one) and the rest simply fetch things from it.
1791 # function (this one) and the rest simply fetch things from it.
1775 class tagscache(object):
1792 class tagscache(object):
1776 def __init__(self):
1793 def __init__(self):
1777 # These two define the set of tags for this repository. tags
1794 # These two define the set of tags for this repository. tags
1778 # maps tag name to node; tagtypes maps tag name to 'global' or
1795 # maps tag name to node; tagtypes maps tag name to 'global' or
1779 # 'local'. (Global tags are defined by .hgtags across all
1796 # 'local'. (Global tags are defined by .hgtags across all
1780 # heads, and local tags are defined in .hg/localtags.)
1797 # heads, and local tags are defined in .hg/localtags.)
1781 # They constitute the in-memory cache of tags.
1798 # They constitute the in-memory cache of tags.
1782 self.tags = self.tagtypes = None
1799 self.tags = self.tagtypes = None
1783
1800
1784 self.nodetagscache = self.tagslist = None
1801 self.nodetagscache = self.tagslist = None
1785
1802
1786 cache = tagscache()
1803 cache = tagscache()
1787 cache.tags, cache.tagtypes = self._findtags()
1804 cache.tags, cache.tagtypes = self._findtags()
1788
1805
1789 return cache
1806 return cache
1790
1807
1791 def tags(self):
1808 def tags(self):
1792 '''return a mapping of tag to node'''
1809 '''return a mapping of tag to node'''
1793 t = {}
1810 t = {}
1794 if self.changelog.filteredrevs:
1811 if self.changelog.filteredrevs:
1795 tags, tt = self._findtags()
1812 tags, tt = self._findtags()
1796 else:
1813 else:
1797 tags = self._tagscache.tags
1814 tags = self._tagscache.tags
1798 rev = self.changelog.rev
1815 rev = self.changelog.rev
1799 for k, v in pycompat.iteritems(tags):
1816 for k, v in pycompat.iteritems(tags):
1800 try:
1817 try:
1801 # ignore tags to unknown nodes
1818 # ignore tags to unknown nodes
1802 rev(v)
1819 rev(v)
1803 t[k] = v
1820 t[k] = v
1804 except (error.LookupError, ValueError):
1821 except (error.LookupError, ValueError):
1805 pass
1822 pass
1806 return t
1823 return t
1807
1824
1808 def _findtags(self):
1825 def _findtags(self):
1809 '''Do the hard work of finding tags. Return a pair of dicts
1826 '''Do the hard work of finding tags. Return a pair of dicts
1810 (tags, tagtypes) where tags maps tag name to node, and tagtypes
1827 (tags, tagtypes) where tags maps tag name to node, and tagtypes
1811 maps tag name to a string like \'global\' or \'local\'.
1828 maps tag name to a string like \'global\' or \'local\'.
1812 Subclasses or extensions are free to add their own tags, but
1829 Subclasses or extensions are free to add their own tags, but
1813 should be aware that the returned dicts will be retained for the
1830 should be aware that the returned dicts will be retained for the
1814 duration of the localrepo object.'''
1831 duration of the localrepo object.'''
1815
1832
1816 # XXX what tagtype should subclasses/extensions use? Currently
1833 # XXX what tagtype should subclasses/extensions use? Currently
1817 # mq and bookmarks add tags, but do not set the tagtype at all.
1834 # mq and bookmarks add tags, but do not set the tagtype at all.
1818 # Should each extension invent its own tag type? Should there
1835 # Should each extension invent its own tag type? Should there
1819 # be one tagtype for all such "virtual" tags? Or is the status
1836 # be one tagtype for all such "virtual" tags? Or is the status
1820 # quo fine?
1837 # quo fine?
1821
1838
1822 # map tag name to (node, hist)
1839 # map tag name to (node, hist)
1823 alltags = tagsmod.findglobaltags(self.ui, self)
1840 alltags = tagsmod.findglobaltags(self.ui, self)
1824 # map tag name to tag type
1841 # map tag name to tag type
1825 tagtypes = {tag: b'global' for tag in alltags}
1842 tagtypes = {tag: b'global' for tag in alltags}
1826
1843
1827 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
1844 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
1828
1845
1829 # Build the return dicts. Have to re-encode tag names because
1846 # Build the return dicts. Have to re-encode tag names because
1830 # the tags module always uses UTF-8 (in order not to lose info
1847 # the tags module always uses UTF-8 (in order not to lose info
1831 # writing to the cache), but the rest of Mercurial wants them in
1848 # writing to the cache), but the rest of Mercurial wants them in
1832 # local encoding.
1849 # local encoding.
1833 tags = {}
1850 tags = {}
1834 for (name, (node, hist)) in pycompat.iteritems(alltags):
1851 for (name, (node, hist)) in pycompat.iteritems(alltags):
1835 if node != nullid:
1852 if node != nullid:
1836 tags[encoding.tolocal(name)] = node
1853 tags[encoding.tolocal(name)] = node
1837 tags[b'tip'] = self.changelog.tip()
1854 tags[b'tip'] = self.changelog.tip()
1838 tagtypes = {
1855 tagtypes = {
1839 encoding.tolocal(name): value
1856 encoding.tolocal(name): value
1840 for (name, value) in pycompat.iteritems(tagtypes)
1857 for (name, value) in pycompat.iteritems(tagtypes)
1841 }
1858 }
1842 return (tags, tagtypes)
1859 return (tags, tagtypes)
1843
1860
1844 def tagtype(self, tagname):
1861 def tagtype(self, tagname):
1845 '''
1862 '''
1846 return the type of the given tag. result can be:
1863 return the type of the given tag. result can be:
1847
1864
1848 'local' : a local tag
1865 'local' : a local tag
1849 'global' : a global tag
1866 'global' : a global tag
1850 None : tag does not exist
1867 None : tag does not exist
1851 '''
1868 '''
1852
1869
1853 return self._tagscache.tagtypes.get(tagname)
1870 return self._tagscache.tagtypes.get(tagname)
1854
1871
1855 def tagslist(self):
1872 def tagslist(self):
1856 '''return a list of tags ordered by revision'''
1873 '''return a list of tags ordered by revision'''
1857 if not self._tagscache.tagslist:
1874 if not self._tagscache.tagslist:
1858 l = []
1875 l = []
1859 for t, n in pycompat.iteritems(self.tags()):
1876 for t, n in pycompat.iteritems(self.tags()):
1860 l.append((self.changelog.rev(n), t, n))
1877 l.append((self.changelog.rev(n), t, n))
1861 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
1878 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
1862
1879
1863 return self._tagscache.tagslist
1880 return self._tagscache.tagslist
1864
1881
1865 def nodetags(self, node):
1882 def nodetags(self, node):
1866 '''return the tags associated with a node'''
1883 '''return the tags associated with a node'''
1867 if not self._tagscache.nodetagscache:
1884 if not self._tagscache.nodetagscache:
1868 nodetagscache = {}
1885 nodetagscache = {}
1869 for t, n in pycompat.iteritems(self._tagscache.tags):
1886 for t, n in pycompat.iteritems(self._tagscache.tags):
1870 nodetagscache.setdefault(n, []).append(t)
1887 nodetagscache.setdefault(n, []).append(t)
1871 for tags in pycompat.itervalues(nodetagscache):
1888 for tags in pycompat.itervalues(nodetagscache):
1872 tags.sort()
1889 tags.sort()
1873 self._tagscache.nodetagscache = nodetagscache
1890 self._tagscache.nodetagscache = nodetagscache
1874 return self._tagscache.nodetagscache.get(node, [])
1891 return self._tagscache.nodetagscache.get(node, [])
1875
1892
1876 def nodebookmarks(self, node):
1893 def nodebookmarks(self, node):
1877 """return the list of bookmarks pointing to the specified node"""
1894 """return the list of bookmarks pointing to the specified node"""
1878 return self._bookmarks.names(node)
1895 return self._bookmarks.names(node)
1879
1896
1880 def branchmap(self):
1897 def branchmap(self):
1881 '''returns a dictionary {branch: [branchheads]} with branchheads
1898 '''returns a dictionary {branch: [branchheads]} with branchheads
1882 ordered by increasing revision number'''
1899 ordered by increasing revision number'''
1883 return self._branchcaches[self]
1900 return self._branchcaches[self]
1884
1901
1885 @unfilteredmethod
1902 @unfilteredmethod
1886 def revbranchcache(self):
1903 def revbranchcache(self):
1887 if not self._revbranchcache:
1904 if not self._revbranchcache:
1888 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
1905 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
1889 return self._revbranchcache
1906 return self._revbranchcache
1890
1907
1891 def branchtip(self, branch, ignoremissing=False):
1908 def branchtip(self, branch, ignoremissing=False):
1892 '''return the tip node for a given branch
1909 '''return the tip node for a given branch
1893
1910
1894 If ignoremissing is True, then this method will not raise an error.
1911 If ignoremissing is True, then this method will not raise an error.
1895 This is helpful for callers that only expect None for a missing branch
1912 This is helpful for callers that only expect None for a missing branch
1896 (e.g. namespace).
1913 (e.g. namespace).
1897
1914
1898 '''
1915 '''
1899 try:
1916 try:
1900 return self.branchmap().branchtip(branch)
1917 return self.branchmap().branchtip(branch)
1901 except KeyError:
1918 except KeyError:
1902 if not ignoremissing:
1919 if not ignoremissing:
1903 raise error.RepoLookupError(_(b"unknown branch '%s'") % branch)
1920 raise error.RepoLookupError(_(b"unknown branch '%s'") % branch)
1904 else:
1921 else:
1905 pass
1922 pass
1906
1923
1907 def lookup(self, key):
1924 def lookup(self, key):
1908 node = scmutil.revsymbol(self, key).node()
1925 node = scmutil.revsymbol(self, key).node()
1909 if node is None:
1926 if node is None:
1910 raise error.RepoLookupError(_(b"unknown revision '%s'") % key)
1927 raise error.RepoLookupError(_(b"unknown revision '%s'") % key)
1911 return node
1928 return node
1912
1929
1913 def lookupbranch(self, key):
1930 def lookupbranch(self, key):
1914 if self.branchmap().hasbranch(key):
1931 if self.branchmap().hasbranch(key):
1915 return key
1932 return key
1916
1933
1917 return scmutil.revsymbol(self, key).branch()
1934 return scmutil.revsymbol(self, key).branch()
1918
1935
1919 def known(self, nodes):
1936 def known(self, nodes):
1920 cl = self.changelog
1937 cl = self.changelog
1921 get_rev = cl.index.get_rev
1938 get_rev = cl.index.get_rev
1922 filtered = cl.filteredrevs
1939 filtered = cl.filteredrevs
1923 result = []
1940 result = []
1924 for n in nodes:
1941 for n in nodes:
1925 r = get_rev(n)
1942 r = get_rev(n)
1926 resp = not (r is None or r in filtered)
1943 resp = not (r is None or r in filtered)
1927 result.append(resp)
1944 result.append(resp)
1928 return result
1945 return result
1929
1946
1930 def local(self):
1947 def local(self):
1931 return self
1948 return self
1932
1949
1933 def publishing(self):
1950 def publishing(self):
1934 # it's safe (and desirable) to trust the publish flag unconditionally
1951 # it's safe (and desirable) to trust the publish flag unconditionally
1935 # so that we don't finalize changes shared between users via ssh or nfs
1952 # so that we don't finalize changes shared between users via ssh or nfs
1936 return self.ui.configbool(b'phases', b'publish', untrusted=True)
1953 return self.ui.configbool(b'phases', b'publish', untrusted=True)
1937
1954
1938 def cancopy(self):
1955 def cancopy(self):
1939 # so statichttprepo's override of local() works
1956 # so statichttprepo's override of local() works
1940 if not self.local():
1957 if not self.local():
1941 return False
1958 return False
1942 if not self.publishing():
1959 if not self.publishing():
1943 return True
1960 return True
1944 # if publishing we can't copy if there is filtered content
1961 # if publishing we can't copy if there is filtered content
1945 return not self.filtered(b'visible').changelog.filteredrevs
1962 return not self.filtered(b'visible').changelog.filteredrevs
1946
1963
1947 def shared(self):
1964 def shared(self):
1948 '''the type of shared repository (None if not shared)'''
1965 '''the type of shared repository (None if not shared)'''
1949 if self.sharedpath != self.path:
1966 if self.sharedpath != self.path:
1950 return b'store'
1967 return b'store'
1951 return None
1968 return None
1952
1969
1953 def wjoin(self, f, *insidef):
1970 def wjoin(self, f, *insidef):
1954 return self.vfs.reljoin(self.root, f, *insidef)
1971 return self.vfs.reljoin(self.root, f, *insidef)
1955
1972
1956 def setparents(self, p1, p2=nullid):
1973 def setparents(self, p1, p2=nullid):
1957 self[None].setparents(p1, p2)
1974 self[None].setparents(p1, p2)
1958 self._quick_access_changeid_invalidate()
1975 self._quick_access_changeid_invalidate()
1959
1976
1960 def filectx(self, path, changeid=None, fileid=None, changectx=None):
1977 def filectx(self, path, changeid=None, fileid=None, changectx=None):
1961 """changeid must be a changeset revision, if specified.
1978 """changeid must be a changeset revision, if specified.
1962 fileid can be a file revision or node."""
1979 fileid can be a file revision or node."""
1963 return context.filectx(
1980 return context.filectx(
1964 self, path, changeid, fileid, changectx=changectx
1981 self, path, changeid, fileid, changectx=changectx
1965 )
1982 )
1966
1983
1967 def getcwd(self):
1984 def getcwd(self):
1968 return self.dirstate.getcwd()
1985 return self.dirstate.getcwd()
1969
1986
1970 def pathto(self, f, cwd=None):
1987 def pathto(self, f, cwd=None):
1971 return self.dirstate.pathto(f, cwd)
1988 return self.dirstate.pathto(f, cwd)
1972
1989
1973 def _loadfilter(self, filter):
1990 def _loadfilter(self, filter):
1974 if filter not in self._filterpats:
1991 if filter not in self._filterpats:
1975 l = []
1992 l = []
1976 for pat, cmd in self.ui.configitems(filter):
1993 for pat, cmd in self.ui.configitems(filter):
1977 if cmd == b'!':
1994 if cmd == b'!':
1978 continue
1995 continue
1979 mf = matchmod.match(self.root, b'', [pat])
1996 mf = matchmod.match(self.root, b'', [pat])
1980 fn = None
1997 fn = None
1981 params = cmd
1998 params = cmd
1982 for name, filterfn in pycompat.iteritems(self._datafilters):
1999 for name, filterfn in pycompat.iteritems(self._datafilters):
1983 if cmd.startswith(name):
2000 if cmd.startswith(name):
1984 fn = filterfn
2001 fn = filterfn
1985 params = cmd[len(name) :].lstrip()
2002 params = cmd[len(name) :].lstrip()
1986 break
2003 break
1987 if not fn:
2004 if not fn:
1988 fn = lambda s, c, **kwargs: procutil.filter(s, c)
2005 fn = lambda s, c, **kwargs: procutil.filter(s, c)
1989 fn.__name__ = 'commandfilter'
2006 fn.__name__ = 'commandfilter'
1990 # Wrap old filters not supporting keyword arguments
2007 # Wrap old filters not supporting keyword arguments
1991 if not pycompat.getargspec(fn)[2]:
2008 if not pycompat.getargspec(fn)[2]:
1992 oldfn = fn
2009 oldfn = fn
1993 fn = lambda s, c, oldfn=oldfn, **kwargs: oldfn(s, c)
2010 fn = lambda s, c, oldfn=oldfn, **kwargs: oldfn(s, c)
1994 fn.__name__ = 'compat-' + oldfn.__name__
2011 fn.__name__ = 'compat-' + oldfn.__name__
1995 l.append((mf, fn, params))
2012 l.append((mf, fn, params))
1996 self._filterpats[filter] = l
2013 self._filterpats[filter] = l
1997 return self._filterpats[filter]
2014 return self._filterpats[filter]
1998
2015
1999 def _filter(self, filterpats, filename, data):
2016 def _filter(self, filterpats, filename, data):
2000 for mf, fn, cmd in filterpats:
2017 for mf, fn, cmd in filterpats:
2001 if mf(filename):
2018 if mf(filename):
2002 self.ui.debug(
2019 self.ui.debug(
2003 b"filtering %s through %s\n"
2020 b"filtering %s through %s\n"
2004 % (filename, cmd or pycompat.sysbytes(fn.__name__))
2021 % (filename, cmd or pycompat.sysbytes(fn.__name__))
2005 )
2022 )
2006 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
2023 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
2007 break
2024 break
2008
2025
2009 return data
2026 return data
2010
2027
2011 @unfilteredpropertycache
2028 @unfilteredpropertycache
2012 def _encodefilterpats(self):
2029 def _encodefilterpats(self):
2013 return self._loadfilter(b'encode')
2030 return self._loadfilter(b'encode')
2014
2031
2015 @unfilteredpropertycache
2032 @unfilteredpropertycache
2016 def _decodefilterpats(self):
2033 def _decodefilterpats(self):
2017 return self._loadfilter(b'decode')
2034 return self._loadfilter(b'decode')
2018
2035
2019 def adddatafilter(self, name, filter):
2036 def adddatafilter(self, name, filter):
2020 self._datafilters[name] = filter
2037 self._datafilters[name] = filter
2021
2038
2022 def wread(self, filename):
2039 def wread(self, filename):
2023 if self.wvfs.islink(filename):
2040 if self.wvfs.islink(filename):
2024 data = self.wvfs.readlink(filename)
2041 data = self.wvfs.readlink(filename)
2025 else:
2042 else:
2026 data = self.wvfs.read(filename)
2043 data = self.wvfs.read(filename)
2027 return self._filter(self._encodefilterpats, filename, data)
2044 return self._filter(self._encodefilterpats, filename, data)
2028
2045
2029 def wwrite(self, filename, data, flags, backgroundclose=False, **kwargs):
2046 def wwrite(self, filename, data, flags, backgroundclose=False, **kwargs):
2030 """write ``data`` into ``filename`` in the working directory
2047 """write ``data`` into ``filename`` in the working directory
2031
2048
2032 This returns length of written (maybe decoded) data.
2049 This returns length of written (maybe decoded) data.
2033 """
2050 """
2034 data = self._filter(self._decodefilterpats, filename, data)
2051 data = self._filter(self._decodefilterpats, filename, data)
2035 if b'l' in flags:
2052 if b'l' in flags:
2036 self.wvfs.symlink(data, filename)
2053 self.wvfs.symlink(data, filename)
2037 else:
2054 else:
2038 self.wvfs.write(
2055 self.wvfs.write(
2039 filename, data, backgroundclose=backgroundclose, **kwargs
2056 filename, data, backgroundclose=backgroundclose, **kwargs
2040 )
2057 )
2041 if b'x' in flags:
2058 if b'x' in flags:
2042 self.wvfs.setflags(filename, False, True)
2059 self.wvfs.setflags(filename, False, True)
2043 else:
2060 else:
2044 self.wvfs.setflags(filename, False, False)
2061 self.wvfs.setflags(filename, False, False)
2045 return len(data)
2062 return len(data)
2046
2063
2047 def wwritedata(self, filename, data):
2064 def wwritedata(self, filename, data):
2048 return self._filter(self._decodefilterpats, filename, data)
2065 return self._filter(self._decodefilterpats, filename, data)
2049
2066
2050 def currenttransaction(self):
2067 def currenttransaction(self):
2051 """return the current transaction or None if non exists"""
2068 """return the current transaction or None if non exists"""
2052 if self._transref:
2069 if self._transref:
2053 tr = self._transref()
2070 tr = self._transref()
2054 else:
2071 else:
2055 tr = None
2072 tr = None
2056
2073
2057 if tr and tr.running():
2074 if tr and tr.running():
2058 return tr
2075 return tr
2059 return None
2076 return None
2060
2077
2061 def transaction(self, desc, report=None):
2078 def transaction(self, desc, report=None):
2062 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
2079 if self.ui.configbool(b'devel', b'all-warnings') or self.ui.configbool(
2063 b'devel', b'check-locks'
2080 b'devel', b'check-locks'
2064 ):
2081 ):
2065 if self._currentlock(self._lockref) is None:
2082 if self._currentlock(self._lockref) is None:
2066 raise error.ProgrammingError(b'transaction requires locking')
2083 raise error.ProgrammingError(b'transaction requires locking')
2067 tr = self.currenttransaction()
2084 tr = self.currenttransaction()
2068 if tr is not None:
2085 if tr is not None:
2069 return tr.nest(name=desc)
2086 return tr.nest(name=desc)
2070
2087
2071 # abort here if the journal already exists
2088 # abort here if the journal already exists
2072 if self.svfs.exists(b"journal"):
2089 if self.svfs.exists(b"journal"):
2073 raise error.RepoError(
2090 raise error.RepoError(
2074 _(b"abandoned transaction found"),
2091 _(b"abandoned transaction found"),
2075 hint=_(b"run 'hg recover' to clean up transaction"),
2092 hint=_(b"run 'hg recover' to clean up transaction"),
2076 )
2093 )
2077
2094
2078 idbase = b"%.40f#%f" % (random.random(), time.time())
2095 idbase = b"%.40f#%f" % (random.random(), time.time())
2079 ha = hex(hashutil.sha1(idbase).digest())
2096 ha = hex(hashutil.sha1(idbase).digest())
2080 txnid = b'TXN:' + ha
2097 txnid = b'TXN:' + ha
2081 self.hook(b'pretxnopen', throw=True, txnname=desc, txnid=txnid)
2098 self.hook(b'pretxnopen', throw=True, txnname=desc, txnid=txnid)
2082
2099
2083 self._writejournal(desc)
2100 self._writejournal(desc)
2084 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
2101 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
2085 if report:
2102 if report:
2086 rp = report
2103 rp = report
2087 else:
2104 else:
2088 rp = self.ui.warn
2105 rp = self.ui.warn
2089 vfsmap = {b'plain': self.vfs, b'store': self.svfs} # root of .hg/
2106 vfsmap = {b'plain': self.vfs, b'store': self.svfs} # root of .hg/
2090 # we must avoid cyclic reference between repo and transaction.
2107 # we must avoid cyclic reference between repo and transaction.
2091 reporef = weakref.ref(self)
2108 reporef = weakref.ref(self)
2092 # Code to track tag movement
2109 # Code to track tag movement
2093 #
2110 #
2094 # Since tags are all handled as file content, it is actually quite hard
2111 # Since tags are all handled as file content, it is actually quite hard
2095 # to track these movement from a code perspective. So we fallback to a
2112 # to track these movement from a code perspective. So we fallback to a
2096 # tracking at the repository level. One could envision to track changes
2113 # tracking at the repository level. One could envision to track changes
2097 # to the '.hgtags' file through changegroup apply but that fails to
2114 # to the '.hgtags' file through changegroup apply but that fails to
2098 # cope with case where transaction expose new heads without changegroup
2115 # cope with case where transaction expose new heads without changegroup
2099 # being involved (eg: phase movement).
2116 # being involved (eg: phase movement).
2100 #
2117 #
2101 # For now, We gate the feature behind a flag since this likely comes
2118 # For now, We gate the feature behind a flag since this likely comes
2102 # with performance impacts. The current code run more often than needed
2119 # with performance impacts. The current code run more often than needed
2103 # and do not use caches as much as it could. The current focus is on
2120 # and do not use caches as much as it could. The current focus is on
2104 # the behavior of the feature so we disable it by default. The flag
2121 # the behavior of the feature so we disable it by default. The flag
2105 # will be removed when we are happy with the performance impact.
2122 # will be removed when we are happy with the performance impact.
2106 #
2123 #
2107 # Once this feature is no longer experimental move the following
2124 # Once this feature is no longer experimental move the following
2108 # documentation to the appropriate help section:
2125 # documentation to the appropriate help section:
2109 #
2126 #
2110 # The ``HG_TAG_MOVED`` variable will be set if the transaction touched
2127 # The ``HG_TAG_MOVED`` variable will be set if the transaction touched
2111 # tags (new or changed or deleted tags). In addition the details of
2128 # tags (new or changed or deleted tags). In addition the details of
2112 # these changes are made available in a file at:
2129 # these changes are made available in a file at:
2113 # ``REPOROOT/.hg/changes/tags.changes``.
2130 # ``REPOROOT/.hg/changes/tags.changes``.
2114 # Make sure you check for HG_TAG_MOVED before reading that file as it
2131 # Make sure you check for HG_TAG_MOVED before reading that file as it
2115 # might exist from a previous transaction even if no tag were touched
2132 # might exist from a previous transaction even if no tag were touched
2116 # in this one. Changes are recorded in a line base format::
2133 # in this one. Changes are recorded in a line base format::
2117 #
2134 #
2118 # <action> <hex-node> <tag-name>\n
2135 # <action> <hex-node> <tag-name>\n
2119 #
2136 #
2120 # Actions are defined as follow:
2137 # Actions are defined as follow:
2121 # "-R": tag is removed,
2138 # "-R": tag is removed,
2122 # "+A": tag is added,
2139 # "+A": tag is added,
2123 # "-M": tag is moved (old value),
2140 # "-M": tag is moved (old value),
2124 # "+M": tag is moved (new value),
2141 # "+M": tag is moved (new value),
2125 tracktags = lambda x: None
2142 tracktags = lambda x: None
2126 # experimental config: experimental.hook-track-tags
2143 # experimental config: experimental.hook-track-tags
2127 shouldtracktags = self.ui.configbool(
2144 shouldtracktags = self.ui.configbool(
2128 b'experimental', b'hook-track-tags'
2145 b'experimental', b'hook-track-tags'
2129 )
2146 )
2130 if desc != b'strip' and shouldtracktags:
2147 if desc != b'strip' and shouldtracktags:
2131 oldheads = self.changelog.headrevs()
2148 oldheads = self.changelog.headrevs()
2132
2149
2133 def tracktags(tr2):
2150 def tracktags(tr2):
2134 repo = reporef()
2151 repo = reporef()
2135 oldfnodes = tagsmod.fnoderevs(repo.ui, repo, oldheads)
2152 oldfnodes = tagsmod.fnoderevs(repo.ui, repo, oldheads)
2136 newheads = repo.changelog.headrevs()
2153 newheads = repo.changelog.headrevs()
2137 newfnodes = tagsmod.fnoderevs(repo.ui, repo, newheads)
2154 newfnodes = tagsmod.fnoderevs(repo.ui, repo, newheads)
2138 # notes: we compare lists here.
2155 # notes: we compare lists here.
2139 # As we do it only once buiding set would not be cheaper
2156 # As we do it only once buiding set would not be cheaper
2140 changes = tagsmod.difftags(repo.ui, repo, oldfnodes, newfnodes)
2157 changes = tagsmod.difftags(repo.ui, repo, oldfnodes, newfnodes)
2141 if changes:
2158 if changes:
2142 tr2.hookargs[b'tag_moved'] = b'1'
2159 tr2.hookargs[b'tag_moved'] = b'1'
2143 with repo.vfs(
2160 with repo.vfs(
2144 b'changes/tags.changes', b'w', atomictemp=True
2161 b'changes/tags.changes', b'w', atomictemp=True
2145 ) as changesfile:
2162 ) as changesfile:
2146 # note: we do not register the file to the transaction
2163 # note: we do not register the file to the transaction
2147 # because we needs it to still exist on the transaction
2164 # because we needs it to still exist on the transaction
2148 # is close (for txnclose hooks)
2165 # is close (for txnclose hooks)
2149 tagsmod.writediff(changesfile, changes)
2166 tagsmod.writediff(changesfile, changes)
2150
2167
2151 def validate(tr2):
2168 def validate(tr2):
2152 """will run pre-closing hooks"""
2169 """will run pre-closing hooks"""
2153 # XXX the transaction API is a bit lacking here so we take a hacky
2170 # XXX the transaction API is a bit lacking here so we take a hacky
2154 # path for now
2171 # path for now
2155 #
2172 #
2156 # We cannot add this as a "pending" hooks since the 'tr.hookargs'
2173 # We cannot add this as a "pending" hooks since the 'tr.hookargs'
2157 # dict is copied before these run. In addition we needs the data
2174 # dict is copied before these run. In addition we needs the data
2158 # available to in memory hooks too.
2175 # available to in memory hooks too.
2159 #
2176 #
2160 # Moreover, we also need to make sure this runs before txnclose
2177 # Moreover, we also need to make sure this runs before txnclose
2161 # hooks and there is no "pending" mechanism that would execute
2178 # hooks and there is no "pending" mechanism that would execute
2162 # logic only if hooks are about to run.
2179 # logic only if hooks are about to run.
2163 #
2180 #
2164 # Fixing this limitation of the transaction is also needed to track
2181 # Fixing this limitation of the transaction is also needed to track
2165 # other families of changes (bookmarks, phases, obsolescence).
2182 # other families of changes (bookmarks, phases, obsolescence).
2166 #
2183 #
2167 # This will have to be fixed before we remove the experimental
2184 # This will have to be fixed before we remove the experimental
2168 # gating.
2185 # gating.
2169 tracktags(tr2)
2186 tracktags(tr2)
2170 repo = reporef()
2187 repo = reporef()
2171
2188
2172 singleheadopt = (b'experimental', b'single-head-per-branch')
2189 singleheadopt = (b'experimental', b'single-head-per-branch')
2173 singlehead = repo.ui.configbool(*singleheadopt)
2190 singlehead = repo.ui.configbool(*singleheadopt)
2174 if singlehead:
2191 if singlehead:
2175 singleheadsub = repo.ui.configsuboptions(*singleheadopt)[1]
2192 singleheadsub = repo.ui.configsuboptions(*singleheadopt)[1]
2176 accountclosed = singleheadsub.get(
2193 accountclosed = singleheadsub.get(
2177 b"account-closed-heads", False
2194 b"account-closed-heads", False
2178 )
2195 )
2179 scmutil.enforcesinglehead(repo, tr2, desc, accountclosed)
2196 scmutil.enforcesinglehead(repo, tr2, desc, accountclosed)
2180 if hook.hashook(repo.ui, b'pretxnclose-bookmark'):
2197 if hook.hashook(repo.ui, b'pretxnclose-bookmark'):
2181 for name, (old, new) in sorted(
2198 for name, (old, new) in sorted(
2182 tr.changes[b'bookmarks'].items()
2199 tr.changes[b'bookmarks'].items()
2183 ):
2200 ):
2184 args = tr.hookargs.copy()
2201 args = tr.hookargs.copy()
2185 args.update(bookmarks.preparehookargs(name, old, new))
2202 args.update(bookmarks.preparehookargs(name, old, new))
2186 repo.hook(
2203 repo.hook(
2187 b'pretxnclose-bookmark',
2204 b'pretxnclose-bookmark',
2188 throw=True,
2205 throw=True,
2189 **pycompat.strkwargs(args)
2206 **pycompat.strkwargs(args)
2190 )
2207 )
2191 if hook.hashook(repo.ui, b'pretxnclose-phase'):
2208 if hook.hashook(repo.ui, b'pretxnclose-phase'):
2192 cl = repo.unfiltered().changelog
2209 cl = repo.unfiltered().changelog
2193 for revs, (old, new) in tr.changes[b'phases']:
2210 for revs, (old, new) in tr.changes[b'phases']:
2194 for rev in revs:
2211 for rev in revs:
2195 args = tr.hookargs.copy()
2212 args = tr.hookargs.copy()
2196 node = hex(cl.node(rev))
2213 node = hex(cl.node(rev))
2197 args.update(phases.preparehookargs(node, old, new))
2214 args.update(phases.preparehookargs(node, old, new))
2198 repo.hook(
2215 repo.hook(
2199 b'pretxnclose-phase',
2216 b'pretxnclose-phase',
2200 throw=True,
2217 throw=True,
2201 **pycompat.strkwargs(args)
2218 **pycompat.strkwargs(args)
2202 )
2219 )
2203
2220
2204 repo.hook(
2221 repo.hook(
2205 b'pretxnclose', throw=True, **pycompat.strkwargs(tr.hookargs)
2222 b'pretxnclose', throw=True, **pycompat.strkwargs(tr.hookargs)
2206 )
2223 )
2207
2224
2208 def releasefn(tr, success):
2225 def releasefn(tr, success):
2209 repo = reporef()
2226 repo = reporef()
2210 if repo is None:
2227 if repo is None:
2211 # If the repo has been GC'd (and this release function is being
2228 # If the repo has been GC'd (and this release function is being
2212 # called from transaction.__del__), there's not much we can do,
2229 # called from transaction.__del__), there's not much we can do,
2213 # so just leave the unfinished transaction there and let the
2230 # so just leave the unfinished transaction there and let the
2214 # user run `hg recover`.
2231 # user run `hg recover`.
2215 return
2232 return
2216 if success:
2233 if success:
2217 # this should be explicitly invoked here, because
2234 # this should be explicitly invoked here, because
2218 # in-memory changes aren't written out at closing
2235 # in-memory changes aren't written out at closing
2219 # transaction, if tr.addfilegenerator (via
2236 # transaction, if tr.addfilegenerator (via
2220 # dirstate.write or so) isn't invoked while
2237 # dirstate.write or so) isn't invoked while
2221 # transaction running
2238 # transaction running
2222 repo.dirstate.write(None)
2239 repo.dirstate.write(None)
2223 else:
2240 else:
2224 # discard all changes (including ones already written
2241 # discard all changes (including ones already written
2225 # out) in this transaction
2242 # out) in this transaction
2226 narrowspec.restorebackup(self, b'journal.narrowspec')
2243 narrowspec.restorebackup(self, b'journal.narrowspec')
2227 narrowspec.restorewcbackup(self, b'journal.narrowspec.dirstate')
2244 narrowspec.restorewcbackup(self, b'journal.narrowspec.dirstate')
2228 repo.dirstate.restorebackup(None, b'journal.dirstate')
2245 repo.dirstate.restorebackup(None, b'journal.dirstate')
2229
2246
2230 repo.invalidate(clearfilecache=True)
2247 repo.invalidate(clearfilecache=True)
2231
2248
2232 tr = transaction.transaction(
2249 tr = transaction.transaction(
2233 rp,
2250 rp,
2234 self.svfs,
2251 self.svfs,
2235 vfsmap,
2252 vfsmap,
2236 b"journal",
2253 b"journal",
2237 b"undo",
2254 b"undo",
2238 aftertrans(renames),
2255 aftertrans(renames),
2239 self.store.createmode,
2256 self.store.createmode,
2240 validator=validate,
2257 validator=validate,
2241 releasefn=releasefn,
2258 releasefn=releasefn,
2242 checkambigfiles=_cachedfiles,
2259 checkambigfiles=_cachedfiles,
2243 name=desc,
2260 name=desc,
2244 )
2261 )
2245 tr.changes[b'origrepolen'] = len(self)
2262 tr.changes[b'origrepolen'] = len(self)
2246 tr.changes[b'obsmarkers'] = set()
2263 tr.changes[b'obsmarkers'] = set()
2247 tr.changes[b'phases'] = []
2264 tr.changes[b'phases'] = []
2248 tr.changes[b'bookmarks'] = {}
2265 tr.changes[b'bookmarks'] = {}
2249
2266
2250 tr.hookargs[b'txnid'] = txnid
2267 tr.hookargs[b'txnid'] = txnid
2251 tr.hookargs[b'txnname'] = desc
2268 tr.hookargs[b'txnname'] = desc
2252 tr.hookargs[b'changes'] = tr.changes
2269 tr.hookargs[b'changes'] = tr.changes
2253 # note: writing the fncache only during finalize mean that the file is
2270 # note: writing the fncache only during finalize mean that the file is
2254 # outdated when running hooks. As fncache is used for streaming clone,
2271 # outdated when running hooks. As fncache is used for streaming clone,
2255 # this is not expected to break anything that happen during the hooks.
2272 # this is not expected to break anything that happen during the hooks.
2256 tr.addfinalize(b'flush-fncache', self.store.write)
2273 tr.addfinalize(b'flush-fncache', self.store.write)
2257
2274
2258 def txnclosehook(tr2):
2275 def txnclosehook(tr2):
2259 """To be run if transaction is successful, will schedule a hook run
2276 """To be run if transaction is successful, will schedule a hook run
2260 """
2277 """
2261 # Don't reference tr2 in hook() so we don't hold a reference.
2278 # Don't reference tr2 in hook() so we don't hold a reference.
2262 # This reduces memory consumption when there are multiple
2279 # This reduces memory consumption when there are multiple
2263 # transactions per lock. This can likely go away if issue5045
2280 # transactions per lock. This can likely go away if issue5045
2264 # fixes the function accumulation.
2281 # fixes the function accumulation.
2265 hookargs = tr2.hookargs
2282 hookargs = tr2.hookargs
2266
2283
2267 def hookfunc(unused_success):
2284 def hookfunc(unused_success):
2268 repo = reporef()
2285 repo = reporef()
2269 if hook.hashook(repo.ui, b'txnclose-bookmark'):
2286 if hook.hashook(repo.ui, b'txnclose-bookmark'):
2270 bmchanges = sorted(tr.changes[b'bookmarks'].items())
2287 bmchanges = sorted(tr.changes[b'bookmarks'].items())
2271 for name, (old, new) in bmchanges:
2288 for name, (old, new) in bmchanges:
2272 args = tr.hookargs.copy()
2289 args = tr.hookargs.copy()
2273 args.update(bookmarks.preparehookargs(name, old, new))
2290 args.update(bookmarks.preparehookargs(name, old, new))
2274 repo.hook(
2291 repo.hook(
2275 b'txnclose-bookmark',
2292 b'txnclose-bookmark',
2276 throw=False,
2293 throw=False,
2277 **pycompat.strkwargs(args)
2294 **pycompat.strkwargs(args)
2278 )
2295 )
2279
2296
2280 if hook.hashook(repo.ui, b'txnclose-phase'):
2297 if hook.hashook(repo.ui, b'txnclose-phase'):
2281 cl = repo.unfiltered().changelog
2298 cl = repo.unfiltered().changelog
2282 phasemv = sorted(
2299 phasemv = sorted(
2283 tr.changes[b'phases'], key=lambda r: r[0][0]
2300 tr.changes[b'phases'], key=lambda r: r[0][0]
2284 )
2301 )
2285 for revs, (old, new) in phasemv:
2302 for revs, (old, new) in phasemv:
2286 for rev in revs:
2303 for rev in revs:
2287 args = tr.hookargs.copy()
2304 args = tr.hookargs.copy()
2288 node = hex(cl.node(rev))
2305 node = hex(cl.node(rev))
2289 args.update(phases.preparehookargs(node, old, new))
2306 args.update(phases.preparehookargs(node, old, new))
2290 repo.hook(
2307 repo.hook(
2291 b'txnclose-phase',
2308 b'txnclose-phase',
2292 throw=False,
2309 throw=False,
2293 **pycompat.strkwargs(args)
2310 **pycompat.strkwargs(args)
2294 )
2311 )
2295
2312
2296 repo.hook(
2313 repo.hook(
2297 b'txnclose', throw=False, **pycompat.strkwargs(hookargs)
2314 b'txnclose', throw=False, **pycompat.strkwargs(hookargs)
2298 )
2315 )
2299
2316
2300 reporef()._afterlock(hookfunc)
2317 reporef()._afterlock(hookfunc)
2301
2318
2302 tr.addfinalize(b'txnclose-hook', txnclosehook)
2319 tr.addfinalize(b'txnclose-hook', txnclosehook)
2303 # Include a leading "-" to make it happen before the transaction summary
2320 # Include a leading "-" to make it happen before the transaction summary
2304 # reports registered via scmutil.registersummarycallback() whose names
2321 # reports registered via scmutil.registersummarycallback() whose names
2305 # are 00-txnreport etc. That way, the caches will be warm when the
2322 # are 00-txnreport etc. That way, the caches will be warm when the
2306 # callbacks run.
2323 # callbacks run.
2307 tr.addpostclose(b'-warm-cache', self._buildcacheupdater(tr))
2324 tr.addpostclose(b'-warm-cache', self._buildcacheupdater(tr))
2308
2325
2309 def txnaborthook(tr2):
2326 def txnaborthook(tr2):
2310 """To be run if transaction is aborted
2327 """To be run if transaction is aborted
2311 """
2328 """
2312 reporef().hook(
2329 reporef().hook(
2313 b'txnabort', throw=False, **pycompat.strkwargs(tr2.hookargs)
2330 b'txnabort', throw=False, **pycompat.strkwargs(tr2.hookargs)
2314 )
2331 )
2315
2332
2316 tr.addabort(b'txnabort-hook', txnaborthook)
2333 tr.addabort(b'txnabort-hook', txnaborthook)
2317 # avoid eager cache invalidation. in-memory data should be identical
2334 # avoid eager cache invalidation. in-memory data should be identical
2318 # to stored data if transaction has no error.
2335 # to stored data if transaction has no error.
2319 tr.addpostclose(b'refresh-filecachestats', self._refreshfilecachestats)
2336 tr.addpostclose(b'refresh-filecachestats', self._refreshfilecachestats)
2320 self._transref = weakref.ref(tr)
2337 self._transref = weakref.ref(tr)
2321 scmutil.registersummarycallback(self, tr, desc)
2338 scmutil.registersummarycallback(self, tr, desc)
2322 return tr
2339 return tr
2323
2340
2324 def _journalfiles(self):
2341 def _journalfiles(self):
2325 return (
2342 return (
2326 (self.svfs, b'journal'),
2343 (self.svfs, b'journal'),
2327 (self.svfs, b'journal.narrowspec'),
2344 (self.svfs, b'journal.narrowspec'),
2328 (self.vfs, b'journal.narrowspec.dirstate'),
2345 (self.vfs, b'journal.narrowspec.dirstate'),
2329 (self.vfs, b'journal.dirstate'),
2346 (self.vfs, b'journal.dirstate'),
2330 (self.vfs, b'journal.branch'),
2347 (self.vfs, b'journal.branch'),
2331 (self.vfs, b'journal.desc'),
2348 (self.vfs, b'journal.desc'),
2332 (bookmarks.bookmarksvfs(self), b'journal.bookmarks'),
2349 (bookmarks.bookmarksvfs(self), b'journal.bookmarks'),
2333 (self.svfs, b'journal.phaseroots'),
2350 (self.svfs, b'journal.phaseroots'),
2334 )
2351 )
2335
2352
2336 def undofiles(self):
2353 def undofiles(self):
2337 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
2354 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
2338
2355
2339 @unfilteredmethod
2356 @unfilteredmethod
2340 def _writejournal(self, desc):
2357 def _writejournal(self, desc):
2341 self.dirstate.savebackup(None, b'journal.dirstate')
2358 self.dirstate.savebackup(None, b'journal.dirstate')
2342 narrowspec.savewcbackup(self, b'journal.narrowspec.dirstate')
2359 narrowspec.savewcbackup(self, b'journal.narrowspec.dirstate')
2343 narrowspec.savebackup(self, b'journal.narrowspec')
2360 narrowspec.savebackup(self, b'journal.narrowspec')
2344 self.vfs.write(
2361 self.vfs.write(
2345 b"journal.branch", encoding.fromlocal(self.dirstate.branch())
2362 b"journal.branch", encoding.fromlocal(self.dirstate.branch())
2346 )
2363 )
2347 self.vfs.write(b"journal.desc", b"%d\n%s\n" % (len(self), desc))
2364 self.vfs.write(b"journal.desc", b"%d\n%s\n" % (len(self), desc))
2348 bookmarksvfs = bookmarks.bookmarksvfs(self)
2365 bookmarksvfs = bookmarks.bookmarksvfs(self)
2349 bookmarksvfs.write(
2366 bookmarksvfs.write(
2350 b"journal.bookmarks", bookmarksvfs.tryread(b"bookmarks")
2367 b"journal.bookmarks", bookmarksvfs.tryread(b"bookmarks")
2351 )
2368 )
2352 self.svfs.write(b"journal.phaseroots", self.svfs.tryread(b"phaseroots"))
2369 self.svfs.write(b"journal.phaseroots", self.svfs.tryread(b"phaseroots"))
2353
2370
2354 def recover(self):
2371 def recover(self):
2355 with self.lock():
2372 with self.lock():
2356 if self.svfs.exists(b"journal"):
2373 if self.svfs.exists(b"journal"):
2357 self.ui.status(_(b"rolling back interrupted transaction\n"))
2374 self.ui.status(_(b"rolling back interrupted transaction\n"))
2358 vfsmap = {
2375 vfsmap = {
2359 b'': self.svfs,
2376 b'': self.svfs,
2360 b'plain': self.vfs,
2377 b'plain': self.vfs,
2361 }
2378 }
2362 transaction.rollback(
2379 transaction.rollback(
2363 self.svfs,
2380 self.svfs,
2364 vfsmap,
2381 vfsmap,
2365 b"journal",
2382 b"journal",
2366 self.ui.warn,
2383 self.ui.warn,
2367 checkambigfiles=_cachedfiles,
2384 checkambigfiles=_cachedfiles,
2368 )
2385 )
2369 self.invalidate()
2386 self.invalidate()
2370 return True
2387 return True
2371 else:
2388 else:
2372 self.ui.warn(_(b"no interrupted transaction available\n"))
2389 self.ui.warn(_(b"no interrupted transaction available\n"))
2373 return False
2390 return False
2374
2391
2375 def rollback(self, dryrun=False, force=False):
2392 def rollback(self, dryrun=False, force=False):
2376 wlock = lock = dsguard = None
2393 wlock = lock = dsguard = None
2377 try:
2394 try:
2378 wlock = self.wlock()
2395 wlock = self.wlock()
2379 lock = self.lock()
2396 lock = self.lock()
2380 if self.svfs.exists(b"undo"):
2397 if self.svfs.exists(b"undo"):
2381 dsguard = dirstateguard.dirstateguard(self, b'rollback')
2398 dsguard = dirstateguard.dirstateguard(self, b'rollback')
2382
2399
2383 return self._rollback(dryrun, force, dsguard)
2400 return self._rollback(dryrun, force, dsguard)
2384 else:
2401 else:
2385 self.ui.warn(_(b"no rollback information available\n"))
2402 self.ui.warn(_(b"no rollback information available\n"))
2386 return 1
2403 return 1
2387 finally:
2404 finally:
2388 release(dsguard, lock, wlock)
2405 release(dsguard, lock, wlock)
2389
2406
2390 @unfilteredmethod # Until we get smarter cache management
2407 @unfilteredmethod # Until we get smarter cache management
2391 def _rollback(self, dryrun, force, dsguard):
2408 def _rollback(self, dryrun, force, dsguard):
2392 ui = self.ui
2409 ui = self.ui
2393 try:
2410 try:
2394 args = self.vfs.read(b'undo.desc').splitlines()
2411 args = self.vfs.read(b'undo.desc').splitlines()
2395 (oldlen, desc, detail) = (int(args[0]), args[1], None)
2412 (oldlen, desc, detail) = (int(args[0]), args[1], None)
2396 if len(args) >= 3:
2413 if len(args) >= 3:
2397 detail = args[2]
2414 detail = args[2]
2398 oldtip = oldlen - 1
2415 oldtip = oldlen - 1
2399
2416
2400 if detail and ui.verbose:
2417 if detail and ui.verbose:
2401 msg = _(
2418 msg = _(
2402 b'repository tip rolled back to revision %d'
2419 b'repository tip rolled back to revision %d'
2403 b' (undo %s: %s)\n'
2420 b' (undo %s: %s)\n'
2404 ) % (oldtip, desc, detail)
2421 ) % (oldtip, desc, detail)
2405 else:
2422 else:
2406 msg = _(
2423 msg = _(
2407 b'repository tip rolled back to revision %d (undo %s)\n'
2424 b'repository tip rolled back to revision %d (undo %s)\n'
2408 ) % (oldtip, desc)
2425 ) % (oldtip, desc)
2409 except IOError:
2426 except IOError:
2410 msg = _(b'rolling back unknown transaction\n')
2427 msg = _(b'rolling back unknown transaction\n')
2411 desc = None
2428 desc = None
2412
2429
2413 if not force and self[b'.'] != self[b'tip'] and desc == b'commit':
2430 if not force and self[b'.'] != self[b'tip'] and desc == b'commit':
2414 raise error.Abort(
2431 raise error.Abort(
2415 _(
2432 _(
2416 b'rollback of last commit while not checked out '
2433 b'rollback of last commit while not checked out '
2417 b'may lose data'
2434 b'may lose data'
2418 ),
2435 ),
2419 hint=_(b'use -f to force'),
2436 hint=_(b'use -f to force'),
2420 )
2437 )
2421
2438
2422 ui.status(msg)
2439 ui.status(msg)
2423 if dryrun:
2440 if dryrun:
2424 return 0
2441 return 0
2425
2442
2426 parents = self.dirstate.parents()
2443 parents = self.dirstate.parents()
2427 self.destroying()
2444 self.destroying()
2428 vfsmap = {b'plain': self.vfs, b'': self.svfs}
2445 vfsmap = {b'plain': self.vfs, b'': self.svfs}
2429 transaction.rollback(
2446 transaction.rollback(
2430 self.svfs, vfsmap, b'undo', ui.warn, checkambigfiles=_cachedfiles
2447 self.svfs, vfsmap, b'undo', ui.warn, checkambigfiles=_cachedfiles
2431 )
2448 )
2432 bookmarksvfs = bookmarks.bookmarksvfs(self)
2449 bookmarksvfs = bookmarks.bookmarksvfs(self)
2433 if bookmarksvfs.exists(b'undo.bookmarks'):
2450 if bookmarksvfs.exists(b'undo.bookmarks'):
2434 bookmarksvfs.rename(
2451 bookmarksvfs.rename(
2435 b'undo.bookmarks', b'bookmarks', checkambig=True
2452 b'undo.bookmarks', b'bookmarks', checkambig=True
2436 )
2453 )
2437 if self.svfs.exists(b'undo.phaseroots'):
2454 if self.svfs.exists(b'undo.phaseroots'):
2438 self.svfs.rename(b'undo.phaseroots', b'phaseroots', checkambig=True)
2455 self.svfs.rename(b'undo.phaseroots', b'phaseroots', checkambig=True)
2439 self.invalidate()
2456 self.invalidate()
2440
2457
2441 has_node = self.changelog.index.has_node
2458 has_node = self.changelog.index.has_node
2442 parentgone = any(not has_node(p) for p in parents)
2459 parentgone = any(not has_node(p) for p in parents)
2443 if parentgone:
2460 if parentgone:
2444 # prevent dirstateguard from overwriting already restored one
2461 # prevent dirstateguard from overwriting already restored one
2445 dsguard.close()
2462 dsguard.close()
2446
2463
2447 narrowspec.restorebackup(self, b'undo.narrowspec')
2464 narrowspec.restorebackup(self, b'undo.narrowspec')
2448 narrowspec.restorewcbackup(self, b'undo.narrowspec.dirstate')
2465 narrowspec.restorewcbackup(self, b'undo.narrowspec.dirstate')
2449 self.dirstate.restorebackup(None, b'undo.dirstate')
2466 self.dirstate.restorebackup(None, b'undo.dirstate')
2450 try:
2467 try:
2451 branch = self.vfs.read(b'undo.branch')
2468 branch = self.vfs.read(b'undo.branch')
2452 self.dirstate.setbranch(encoding.tolocal(branch))
2469 self.dirstate.setbranch(encoding.tolocal(branch))
2453 except IOError:
2470 except IOError:
2454 ui.warn(
2471 ui.warn(
2455 _(
2472 _(
2456 b'named branch could not be reset: '
2473 b'named branch could not be reset: '
2457 b'current branch is still \'%s\'\n'
2474 b'current branch is still \'%s\'\n'
2458 )
2475 )
2459 % self.dirstate.branch()
2476 % self.dirstate.branch()
2460 )
2477 )
2461
2478
2462 parents = tuple([p.rev() for p in self[None].parents()])
2479 parents = tuple([p.rev() for p in self[None].parents()])
2463 if len(parents) > 1:
2480 if len(parents) > 1:
2464 ui.status(
2481 ui.status(
2465 _(
2482 _(
2466 b'working directory now based on '
2483 b'working directory now based on '
2467 b'revisions %d and %d\n'
2484 b'revisions %d and %d\n'
2468 )
2485 )
2469 % parents
2486 % parents
2470 )
2487 )
2471 else:
2488 else:
2472 ui.status(
2489 ui.status(
2473 _(b'working directory now based on revision %d\n') % parents
2490 _(b'working directory now based on revision %d\n') % parents
2474 )
2491 )
2475 mergestatemod.mergestate.clean(self, self[b'.'].node())
2492 mergestatemod.mergestate.clean(self, self[b'.'].node())
2476
2493
2477 # TODO: if we know which new heads may result from this rollback, pass
2494 # TODO: if we know which new heads may result from this rollback, pass
2478 # them to destroy(), which will prevent the branchhead cache from being
2495 # them to destroy(), which will prevent the branchhead cache from being
2479 # invalidated.
2496 # invalidated.
2480 self.destroyed()
2497 self.destroyed()
2481 return 0
2498 return 0
2482
2499
2483 def _buildcacheupdater(self, newtransaction):
2500 def _buildcacheupdater(self, newtransaction):
2484 """called during transaction to build the callback updating cache
2501 """called during transaction to build the callback updating cache
2485
2502
2486 Lives on the repository to help extension who might want to augment
2503 Lives on the repository to help extension who might want to augment
2487 this logic. For this purpose, the created transaction is passed to the
2504 this logic. For this purpose, the created transaction is passed to the
2488 method.
2505 method.
2489 """
2506 """
2490 # we must avoid cyclic reference between repo and transaction.
2507 # we must avoid cyclic reference between repo and transaction.
2491 reporef = weakref.ref(self)
2508 reporef = weakref.ref(self)
2492
2509
2493 def updater(tr):
2510 def updater(tr):
2494 repo = reporef()
2511 repo = reporef()
2495 repo.updatecaches(tr)
2512 repo.updatecaches(tr)
2496
2513
2497 return updater
2514 return updater
2498
2515
2499 @unfilteredmethod
2516 @unfilteredmethod
2500 def updatecaches(self, tr=None, full=False):
2517 def updatecaches(self, tr=None, full=False):
2501 """warm appropriate caches
2518 """warm appropriate caches
2502
2519
2503 If this function is called after a transaction closed. The transaction
2520 If this function is called after a transaction closed. The transaction
2504 will be available in the 'tr' argument. This can be used to selectively
2521 will be available in the 'tr' argument. This can be used to selectively
2505 update caches relevant to the changes in that transaction.
2522 update caches relevant to the changes in that transaction.
2506
2523
2507 If 'full' is set, make sure all caches the function knows about have
2524 If 'full' is set, make sure all caches the function knows about have
2508 up-to-date data. Even the ones usually loaded more lazily.
2525 up-to-date data. Even the ones usually loaded more lazily.
2509 """
2526 """
2510 if tr is not None and tr.hookargs.get(b'source') == b'strip':
2527 if tr is not None and tr.hookargs.get(b'source') == b'strip':
2511 # During strip, many caches are invalid but
2528 # During strip, many caches are invalid but
2512 # later call to `destroyed` will refresh them.
2529 # later call to `destroyed` will refresh them.
2513 return
2530 return
2514
2531
2515 if tr is None or tr.changes[b'origrepolen'] < len(self):
2532 if tr is None or tr.changes[b'origrepolen'] < len(self):
2516 # accessing the 'ser ved' branchmap should refresh all the others,
2533 # accessing the 'ser ved' branchmap should refresh all the others,
2517 self.ui.debug(b'updating the branch cache\n')
2534 self.ui.debug(b'updating the branch cache\n')
2518 self.filtered(b'served').branchmap()
2535 self.filtered(b'served').branchmap()
2519 self.filtered(b'served.hidden').branchmap()
2536 self.filtered(b'served.hidden').branchmap()
2520
2537
2521 if full:
2538 if full:
2522 unfi = self.unfiltered()
2539 unfi = self.unfiltered()
2523
2540
2524 self.changelog.update_caches(transaction=tr)
2541 self.changelog.update_caches(transaction=tr)
2525 self.manifestlog.update_caches(transaction=tr)
2542 self.manifestlog.update_caches(transaction=tr)
2526
2543
2527 rbc = unfi.revbranchcache()
2544 rbc = unfi.revbranchcache()
2528 for r in unfi.changelog:
2545 for r in unfi.changelog:
2529 rbc.branchinfo(r)
2546 rbc.branchinfo(r)
2530 rbc.write()
2547 rbc.write()
2531
2548
2532 # ensure the working copy parents are in the manifestfulltextcache
2549 # ensure the working copy parents are in the manifestfulltextcache
2533 for ctx in self[b'.'].parents():
2550 for ctx in self[b'.'].parents():
2534 ctx.manifest() # accessing the manifest is enough
2551 ctx.manifest() # accessing the manifest is enough
2535
2552
2536 # accessing fnode cache warms the cache
2553 # accessing fnode cache warms the cache
2537 tagsmod.fnoderevs(self.ui, unfi, unfi.changelog.revs())
2554 tagsmod.fnoderevs(self.ui, unfi, unfi.changelog.revs())
2538 # accessing tags warm the cache
2555 # accessing tags warm the cache
2539 self.tags()
2556 self.tags()
2540 self.filtered(b'served').tags()
2557 self.filtered(b'served').tags()
2541
2558
2542 # The `full` arg is documented as updating even the lazily-loaded
2559 # The `full` arg is documented as updating even the lazily-loaded
2543 # caches immediately, so we're forcing a write to cause these caches
2560 # caches immediately, so we're forcing a write to cause these caches
2544 # to be warmed up even if they haven't explicitly been requested
2561 # to be warmed up even if they haven't explicitly been requested
2545 # yet (if they've never been used by hg, they won't ever have been
2562 # yet (if they've never been used by hg, they won't ever have been
2546 # written, even if they're a subset of another kind of cache that
2563 # written, even if they're a subset of another kind of cache that
2547 # *has* been used).
2564 # *has* been used).
2548 for filt in repoview.filtertable.keys():
2565 for filt in repoview.filtertable.keys():
2549 filtered = self.filtered(filt)
2566 filtered = self.filtered(filt)
2550 filtered.branchmap().write(filtered)
2567 filtered.branchmap().write(filtered)
2551
2568
2552 def invalidatecaches(self):
2569 def invalidatecaches(self):
2553
2570
2554 if '_tagscache' in vars(self):
2571 if '_tagscache' in vars(self):
2555 # can't use delattr on proxy
2572 # can't use delattr on proxy
2556 del self.__dict__['_tagscache']
2573 del self.__dict__['_tagscache']
2557
2574
2558 self._branchcaches.clear()
2575 self._branchcaches.clear()
2559 self.invalidatevolatilesets()
2576 self.invalidatevolatilesets()
2560 self._sparsesignaturecache.clear()
2577 self._sparsesignaturecache.clear()
2561
2578
2562 def invalidatevolatilesets(self):
2579 def invalidatevolatilesets(self):
2563 self.filteredrevcache.clear()
2580 self.filteredrevcache.clear()
2564 obsolete.clearobscaches(self)
2581 obsolete.clearobscaches(self)
2565 self._quick_access_changeid_invalidate()
2582 self._quick_access_changeid_invalidate()
2566
2583
2567 def invalidatedirstate(self):
2584 def invalidatedirstate(self):
2568 '''Invalidates the dirstate, causing the next call to dirstate
2585 '''Invalidates the dirstate, causing the next call to dirstate
2569 to check if it was modified since the last time it was read,
2586 to check if it was modified since the last time it was read,
2570 rereading it if it has.
2587 rereading it if it has.
2571
2588
2572 This is different to dirstate.invalidate() that it doesn't always
2589 This is different to dirstate.invalidate() that it doesn't always
2573 rereads the dirstate. Use dirstate.invalidate() if you want to
2590 rereads the dirstate. Use dirstate.invalidate() if you want to
2574 explicitly read the dirstate again (i.e. restoring it to a previous
2591 explicitly read the dirstate again (i.e. restoring it to a previous
2575 known good state).'''
2592 known good state).'''
2576 if hasunfilteredcache(self, 'dirstate'):
2593 if hasunfilteredcache(self, 'dirstate'):
2577 for k in self.dirstate._filecache:
2594 for k in self.dirstate._filecache:
2578 try:
2595 try:
2579 delattr(self.dirstate, k)
2596 delattr(self.dirstate, k)
2580 except AttributeError:
2597 except AttributeError:
2581 pass
2598 pass
2582 delattr(self.unfiltered(), 'dirstate')
2599 delattr(self.unfiltered(), 'dirstate')
2583
2600
2584 def invalidate(self, clearfilecache=False):
2601 def invalidate(self, clearfilecache=False):
2585 '''Invalidates both store and non-store parts other than dirstate
2602 '''Invalidates both store and non-store parts other than dirstate
2586
2603
2587 If a transaction is running, invalidation of store is omitted,
2604 If a transaction is running, invalidation of store is omitted,
2588 because discarding in-memory changes might cause inconsistency
2605 because discarding in-memory changes might cause inconsistency
2589 (e.g. incomplete fncache causes unintentional failure, but
2606 (e.g. incomplete fncache causes unintentional failure, but
2590 redundant one doesn't).
2607 redundant one doesn't).
2591 '''
2608 '''
2592 unfiltered = self.unfiltered() # all file caches are stored unfiltered
2609 unfiltered = self.unfiltered() # all file caches are stored unfiltered
2593 for k in list(self._filecache.keys()):
2610 for k in list(self._filecache.keys()):
2594 # dirstate is invalidated separately in invalidatedirstate()
2611 # dirstate is invalidated separately in invalidatedirstate()
2595 if k == b'dirstate':
2612 if k == b'dirstate':
2596 continue
2613 continue
2597 if (
2614 if (
2598 k == b'changelog'
2615 k == b'changelog'
2599 and self.currenttransaction()
2616 and self.currenttransaction()
2600 and self.changelog._delayed
2617 and self.changelog._delayed
2601 ):
2618 ):
2602 # The changelog object may store unwritten revisions. We don't
2619 # The changelog object may store unwritten revisions. We don't
2603 # want to lose them.
2620 # want to lose them.
2604 # TODO: Solve the problem instead of working around it.
2621 # TODO: Solve the problem instead of working around it.
2605 continue
2622 continue
2606
2623
2607 if clearfilecache:
2624 if clearfilecache:
2608 del self._filecache[k]
2625 del self._filecache[k]
2609 try:
2626 try:
2610 delattr(unfiltered, k)
2627 delattr(unfiltered, k)
2611 except AttributeError:
2628 except AttributeError:
2612 pass
2629 pass
2613 self.invalidatecaches()
2630 self.invalidatecaches()
2614 if not self.currenttransaction():
2631 if not self.currenttransaction():
2615 # TODO: Changing contents of store outside transaction
2632 # TODO: Changing contents of store outside transaction
2616 # causes inconsistency. We should make in-memory store
2633 # causes inconsistency. We should make in-memory store
2617 # changes detectable, and abort if changed.
2634 # changes detectable, and abort if changed.
2618 self.store.invalidatecaches()
2635 self.store.invalidatecaches()
2619
2636
2620 def invalidateall(self):
2637 def invalidateall(self):
2621 '''Fully invalidates both store and non-store parts, causing the
2638 '''Fully invalidates both store and non-store parts, causing the
2622 subsequent operation to reread any outside changes.'''
2639 subsequent operation to reread any outside changes.'''
2623 # extension should hook this to invalidate its caches
2640 # extension should hook this to invalidate its caches
2624 self.invalidate()
2641 self.invalidate()
2625 self.invalidatedirstate()
2642 self.invalidatedirstate()
2626
2643
2627 @unfilteredmethod
2644 @unfilteredmethod
2628 def _refreshfilecachestats(self, tr):
2645 def _refreshfilecachestats(self, tr):
2629 """Reload stats of cached files so that they are flagged as valid"""
2646 """Reload stats of cached files so that they are flagged as valid"""
2630 for k, ce in self._filecache.items():
2647 for k, ce in self._filecache.items():
2631 k = pycompat.sysstr(k)
2648 k = pycompat.sysstr(k)
2632 if k == 'dirstate' or k not in self.__dict__:
2649 if k == 'dirstate' or k not in self.__dict__:
2633 continue
2650 continue
2634 ce.refresh()
2651 ce.refresh()
2635
2652
2636 def _lock(
2653 def _lock(
2637 self,
2654 self,
2638 vfs,
2655 vfs,
2639 lockname,
2656 lockname,
2640 wait,
2657 wait,
2641 releasefn,
2658 releasefn,
2642 acquirefn,
2659 acquirefn,
2643 desc,
2660 desc,
2644 inheritchecker=None,
2661 inheritchecker=None,
2645 parentenvvar=None,
2662 parentenvvar=None,
2646 ):
2663 ):
2647 parentlock = None
2664 parentlock = None
2648 # the contents of parentenvvar are used by the underlying lock to
2665 # the contents of parentenvvar are used by the underlying lock to
2649 # determine whether it can be inherited
2666 # determine whether it can be inherited
2650 if parentenvvar is not None:
2667 if parentenvvar is not None:
2651 parentlock = encoding.environ.get(parentenvvar)
2668 parentlock = encoding.environ.get(parentenvvar)
2652
2669
2653 timeout = 0
2670 timeout = 0
2654 warntimeout = 0
2671 warntimeout = 0
2655 if wait:
2672 if wait:
2656 timeout = self.ui.configint(b"ui", b"timeout")
2673 timeout = self.ui.configint(b"ui", b"timeout")
2657 warntimeout = self.ui.configint(b"ui", b"timeout.warn")
2674 warntimeout = self.ui.configint(b"ui", b"timeout.warn")
2658 # internal config: ui.signal-safe-lock
2675 # internal config: ui.signal-safe-lock
2659 signalsafe = self.ui.configbool(b'ui', b'signal-safe-lock')
2676 signalsafe = self.ui.configbool(b'ui', b'signal-safe-lock')
2660
2677
2661 l = lockmod.trylock(
2678 l = lockmod.trylock(
2662 self.ui,
2679 self.ui,
2663 vfs,
2680 vfs,
2664 lockname,
2681 lockname,
2665 timeout,
2682 timeout,
2666 warntimeout,
2683 warntimeout,
2667 releasefn=releasefn,
2684 releasefn=releasefn,
2668 acquirefn=acquirefn,
2685 acquirefn=acquirefn,
2669 desc=desc,
2686 desc=desc,
2670 inheritchecker=inheritchecker,
2687 inheritchecker=inheritchecker,
2671 parentlock=parentlock,
2688 parentlock=parentlock,
2672 signalsafe=signalsafe,
2689 signalsafe=signalsafe,
2673 )
2690 )
2674 return l
2691 return l
2675
2692
2676 def _afterlock(self, callback):
2693 def _afterlock(self, callback):
2677 """add a callback to be run when the repository is fully unlocked
2694 """add a callback to be run when the repository is fully unlocked
2678
2695
2679 The callback will be executed when the outermost lock is released
2696 The callback will be executed when the outermost lock is released
2680 (with wlock being higher level than 'lock')."""
2697 (with wlock being higher level than 'lock')."""
2681 for ref in (self._wlockref, self._lockref):
2698 for ref in (self._wlockref, self._lockref):
2682 l = ref and ref()
2699 l = ref and ref()
2683 if l and l.held:
2700 if l and l.held:
2684 l.postrelease.append(callback)
2701 l.postrelease.append(callback)
2685 break
2702 break
2686 else: # no lock have been found.
2703 else: # no lock have been found.
2687 callback(True)
2704 callback(True)
2688
2705
2689 def lock(self, wait=True):
2706 def lock(self, wait=True):
2690 '''Lock the repository store (.hg/store) and return a weak reference
2707 '''Lock the repository store (.hg/store) and return a weak reference
2691 to the lock. Use this before modifying the store (e.g. committing or
2708 to the lock. Use this before modifying the store (e.g. committing or
2692 stripping). If you are opening a transaction, get a lock as well.)
2709 stripping). If you are opening a transaction, get a lock as well.)
2693
2710
2694 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2711 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2695 'wlock' first to avoid a dead-lock hazard.'''
2712 'wlock' first to avoid a dead-lock hazard.'''
2696 l = self._currentlock(self._lockref)
2713 l = self._currentlock(self._lockref)
2697 if l is not None:
2714 if l is not None:
2698 l.lock()
2715 l.lock()
2699 return l
2716 return l
2700
2717
2701 l = self._lock(
2718 l = self._lock(
2702 vfs=self.svfs,
2719 vfs=self.svfs,
2703 lockname=b"lock",
2720 lockname=b"lock",
2704 wait=wait,
2721 wait=wait,
2705 releasefn=None,
2722 releasefn=None,
2706 acquirefn=self.invalidate,
2723 acquirefn=self.invalidate,
2707 desc=_(b'repository %s') % self.origroot,
2724 desc=_(b'repository %s') % self.origroot,
2708 )
2725 )
2709 self._lockref = weakref.ref(l)
2726 self._lockref = weakref.ref(l)
2710 return l
2727 return l
2711
2728
2712 def _wlockchecktransaction(self):
2729 def _wlockchecktransaction(self):
2713 if self.currenttransaction() is not None:
2730 if self.currenttransaction() is not None:
2714 raise error.LockInheritanceContractViolation(
2731 raise error.LockInheritanceContractViolation(
2715 b'wlock cannot be inherited in the middle of a transaction'
2732 b'wlock cannot be inherited in the middle of a transaction'
2716 )
2733 )
2717
2734
2718 def wlock(self, wait=True):
2735 def wlock(self, wait=True):
2719 '''Lock the non-store parts of the repository (everything under
2736 '''Lock the non-store parts of the repository (everything under
2720 .hg except .hg/store) and return a weak reference to the lock.
2737 .hg except .hg/store) and return a weak reference to the lock.
2721
2738
2722 Use this before modifying files in .hg.
2739 Use this before modifying files in .hg.
2723
2740
2724 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2741 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
2725 'wlock' first to avoid a dead-lock hazard.'''
2742 'wlock' first to avoid a dead-lock hazard.'''
2726 l = self._wlockref and self._wlockref()
2743 l = self._wlockref and self._wlockref()
2727 if l is not None and l.held:
2744 if l is not None and l.held:
2728 l.lock()
2745 l.lock()
2729 return l
2746 return l
2730
2747
2731 # We do not need to check for non-waiting lock acquisition. Such
2748 # We do not need to check for non-waiting lock acquisition. Such
2732 # acquisition would not cause dead-lock as they would just fail.
2749 # acquisition would not cause dead-lock as they would just fail.
2733 if wait and (
2750 if wait and (
2734 self.ui.configbool(b'devel', b'all-warnings')
2751 self.ui.configbool(b'devel', b'all-warnings')
2735 or self.ui.configbool(b'devel', b'check-locks')
2752 or self.ui.configbool(b'devel', b'check-locks')
2736 ):
2753 ):
2737 if self._currentlock(self._lockref) is not None:
2754 if self._currentlock(self._lockref) is not None:
2738 self.ui.develwarn(b'"wlock" acquired after "lock"')
2755 self.ui.develwarn(b'"wlock" acquired after "lock"')
2739
2756
2740 def unlock():
2757 def unlock():
2741 if self.dirstate.pendingparentchange():
2758 if self.dirstate.pendingparentchange():
2742 self.dirstate.invalidate()
2759 self.dirstate.invalidate()
2743 else:
2760 else:
2744 self.dirstate.write(None)
2761 self.dirstate.write(None)
2745
2762
2746 self._filecache[b'dirstate'].refresh()
2763 self._filecache[b'dirstate'].refresh()
2747
2764
2748 l = self._lock(
2765 l = self._lock(
2749 self.vfs,
2766 self.vfs,
2750 b"wlock",
2767 b"wlock",
2751 wait,
2768 wait,
2752 unlock,
2769 unlock,
2753 self.invalidatedirstate,
2770 self.invalidatedirstate,
2754 _(b'working directory of %s') % self.origroot,
2771 _(b'working directory of %s') % self.origroot,
2755 inheritchecker=self._wlockchecktransaction,
2772 inheritchecker=self._wlockchecktransaction,
2756 parentenvvar=b'HG_WLOCK_LOCKER',
2773 parentenvvar=b'HG_WLOCK_LOCKER',
2757 )
2774 )
2758 self._wlockref = weakref.ref(l)
2775 self._wlockref = weakref.ref(l)
2759 return l
2776 return l
2760
2777
2761 def _currentlock(self, lockref):
2778 def _currentlock(self, lockref):
2762 """Returns the lock if it's held, or None if it's not."""
2779 """Returns the lock if it's held, or None if it's not."""
2763 if lockref is None:
2780 if lockref is None:
2764 return None
2781 return None
2765 l = lockref()
2782 l = lockref()
2766 if l is None or not l.held:
2783 if l is None or not l.held:
2767 return None
2784 return None
2768 return l
2785 return l
2769
2786
2770 def currentwlock(self):
2787 def currentwlock(self):
2771 """Returns the wlock if it's held, or None if it's not."""
2788 """Returns the wlock if it's held, or None if it's not."""
2772 return self._currentlock(self._wlockref)
2789 return self._currentlock(self._wlockref)
2773
2790
2774 def checkcommitpatterns(self, wctx, match, status, fail):
2791 def checkcommitpatterns(self, wctx, match, status, fail):
2775 """check for commit arguments that aren't committable"""
2792 """check for commit arguments that aren't committable"""
2776 if match.isexact() or match.prefix():
2793 if match.isexact() or match.prefix():
2777 matched = set(status.modified + status.added + status.removed)
2794 matched = set(status.modified + status.added + status.removed)
2778
2795
2779 for f in match.files():
2796 for f in match.files():
2780 f = self.dirstate.normalize(f)
2797 f = self.dirstate.normalize(f)
2781 if f == b'.' or f in matched or f in wctx.substate:
2798 if f == b'.' or f in matched or f in wctx.substate:
2782 continue
2799 continue
2783 if f in status.deleted:
2800 if f in status.deleted:
2784 fail(f, _(b'file not found!'))
2801 fail(f, _(b'file not found!'))
2785 # Is it a directory that exists or used to exist?
2802 # Is it a directory that exists or used to exist?
2786 if self.wvfs.isdir(f) or wctx.p1().hasdir(f):
2803 if self.wvfs.isdir(f) or wctx.p1().hasdir(f):
2787 d = f + b'/'
2804 d = f + b'/'
2788 for mf in matched:
2805 for mf in matched:
2789 if mf.startswith(d):
2806 if mf.startswith(d):
2790 break
2807 break
2791 else:
2808 else:
2792 fail(f, _(b"no match under directory!"))
2809 fail(f, _(b"no match under directory!"))
2793 elif f not in self.dirstate:
2810 elif f not in self.dirstate:
2794 fail(f, _(b"file not tracked!"))
2811 fail(f, _(b"file not tracked!"))
2795
2812
2796 @unfilteredmethod
2813 @unfilteredmethod
2797 def commit(
2814 def commit(
2798 self,
2815 self,
2799 text=b"",
2816 text=b"",
2800 user=None,
2817 user=None,
2801 date=None,
2818 date=None,
2802 match=None,
2819 match=None,
2803 force=False,
2820 force=False,
2804 editor=None,
2821 editor=None,
2805 extra=None,
2822 extra=None,
2806 ):
2823 ):
2807 """Add a new revision to current repository.
2824 """Add a new revision to current repository.
2808
2825
2809 Revision information is gathered from the working directory,
2826 Revision information is gathered from the working directory,
2810 match can be used to filter the committed files. If editor is
2827 match can be used to filter the committed files. If editor is
2811 supplied, it is called to get a commit message.
2828 supplied, it is called to get a commit message.
2812 """
2829 """
2813 if extra is None:
2830 if extra is None:
2814 extra = {}
2831 extra = {}
2815
2832
2816 def fail(f, msg):
2833 def fail(f, msg):
2817 raise error.Abort(b'%s: %s' % (f, msg))
2834 raise error.Abort(b'%s: %s' % (f, msg))
2818
2835
2819 if not match:
2836 if not match:
2820 match = matchmod.always()
2837 match = matchmod.always()
2821
2838
2822 if not force:
2839 if not force:
2823 match.bad = fail
2840 match.bad = fail
2824
2841
2825 # lock() for recent changelog (see issue4368)
2842 # lock() for recent changelog (see issue4368)
2826 with self.wlock(), self.lock():
2843 with self.wlock(), self.lock():
2827 wctx = self[None]
2844 wctx = self[None]
2828 merge = len(wctx.parents()) > 1
2845 merge = len(wctx.parents()) > 1
2829
2846
2830 if not force and merge and not match.always():
2847 if not force and merge and not match.always():
2831 raise error.Abort(
2848 raise error.Abort(
2832 _(
2849 _(
2833 b'cannot partially commit a merge '
2850 b'cannot partially commit a merge '
2834 b'(do not specify files or patterns)'
2851 b'(do not specify files or patterns)'
2835 )
2852 )
2836 )
2853 )
2837
2854
2838 status = self.status(match=match, clean=force)
2855 status = self.status(match=match, clean=force)
2839 if force:
2856 if force:
2840 status.modified.extend(
2857 status.modified.extend(
2841 status.clean
2858 status.clean
2842 ) # mq may commit clean files
2859 ) # mq may commit clean files
2843
2860
2844 # check subrepos
2861 # check subrepos
2845 subs, commitsubs, newstate = subrepoutil.precommit(
2862 subs, commitsubs, newstate = subrepoutil.precommit(
2846 self.ui, wctx, status, match, force=force
2863 self.ui, wctx, status, match, force=force
2847 )
2864 )
2848
2865
2849 # make sure all explicit patterns are matched
2866 # make sure all explicit patterns are matched
2850 if not force:
2867 if not force:
2851 self.checkcommitpatterns(wctx, match, status, fail)
2868 self.checkcommitpatterns(wctx, match, status, fail)
2852
2869
2853 cctx = context.workingcommitctx(
2870 cctx = context.workingcommitctx(
2854 self, status, text, user, date, extra
2871 self, status, text, user, date, extra
2855 )
2872 )
2856
2873
2857 ms = mergestatemod.mergestate.read(self)
2874 ms = mergestatemod.mergestate.read(self)
2858 mergeutil.checkunresolved(ms)
2875 mergeutil.checkunresolved(ms)
2859
2876
2860 # internal config: ui.allowemptycommit
2877 # internal config: ui.allowemptycommit
2861 if cctx.isempty() and not self.ui.configbool(
2878 if cctx.isempty() and not self.ui.configbool(
2862 b'ui', b'allowemptycommit'
2879 b'ui', b'allowemptycommit'
2863 ):
2880 ):
2864 self.ui.debug(b'nothing to commit, clearing merge state\n')
2881 self.ui.debug(b'nothing to commit, clearing merge state\n')
2865 ms.reset()
2882 ms.reset()
2866 return None
2883 return None
2867
2884
2868 if merge and cctx.deleted():
2885 if merge and cctx.deleted():
2869 raise error.Abort(_(b"cannot commit merge with missing files"))
2886 raise error.Abort(_(b"cannot commit merge with missing files"))
2870
2887
2871 if editor:
2888 if editor:
2872 cctx._text = editor(self, cctx, subs)
2889 cctx._text = editor(self, cctx, subs)
2873 edited = text != cctx._text
2890 edited = text != cctx._text
2874
2891
2875 # Save commit message in case this transaction gets rolled back
2892 # Save commit message in case this transaction gets rolled back
2876 # (e.g. by a pretxncommit hook). Leave the content alone on
2893 # (e.g. by a pretxncommit hook). Leave the content alone on
2877 # the assumption that the user will use the same editor again.
2894 # the assumption that the user will use the same editor again.
2878 msgfn = self.savecommitmessage(cctx._text)
2895 msgfn = self.savecommitmessage(cctx._text)
2879
2896
2880 # commit subs and write new state
2897 # commit subs and write new state
2881 if subs:
2898 if subs:
2882 uipathfn = scmutil.getuipathfn(self)
2899 uipathfn = scmutil.getuipathfn(self)
2883 for s in sorted(commitsubs):
2900 for s in sorted(commitsubs):
2884 sub = wctx.sub(s)
2901 sub = wctx.sub(s)
2885 self.ui.status(
2902 self.ui.status(
2886 _(b'committing subrepository %s\n')
2903 _(b'committing subrepository %s\n')
2887 % uipathfn(subrepoutil.subrelpath(sub))
2904 % uipathfn(subrepoutil.subrelpath(sub))
2888 )
2905 )
2889 sr = sub.commit(cctx._text, user, date)
2906 sr = sub.commit(cctx._text, user, date)
2890 newstate[s] = (newstate[s][0], sr)
2907 newstate[s] = (newstate[s][0], sr)
2891 subrepoutil.writestate(self, newstate)
2908 subrepoutil.writestate(self, newstate)
2892
2909
2893 p1, p2 = self.dirstate.parents()
2910 p1, p2 = self.dirstate.parents()
2894 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or b'')
2911 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or b'')
2895 try:
2912 try:
2896 self.hook(
2913 self.hook(
2897 b"precommit", throw=True, parent1=hookp1, parent2=hookp2
2914 b"precommit", throw=True, parent1=hookp1, parent2=hookp2
2898 )
2915 )
2899 with self.transaction(b'commit'):
2916 with self.transaction(b'commit'):
2900 ret = self.commitctx(cctx, True)
2917 ret = self.commitctx(cctx, True)
2901 # update bookmarks, dirstate and mergestate
2918 # update bookmarks, dirstate and mergestate
2902 bookmarks.update(self, [p1, p2], ret)
2919 bookmarks.update(self, [p1, p2], ret)
2903 cctx.markcommitted(ret)
2920 cctx.markcommitted(ret)
2904 ms.reset()
2921 ms.reset()
2905 except: # re-raises
2922 except: # re-raises
2906 if edited:
2923 if edited:
2907 self.ui.write(
2924 self.ui.write(
2908 _(b'note: commit message saved in %s\n') % msgfn
2925 _(b'note: commit message saved in %s\n') % msgfn
2909 )
2926 )
2910 self.ui.write(
2927 self.ui.write(
2911 _(
2928 _(
2912 b"note: use 'hg commit --logfile "
2929 b"note: use 'hg commit --logfile "
2913 b".hg/last-message.txt --edit' to reuse it\n"
2930 b".hg/last-message.txt --edit' to reuse it\n"
2914 )
2931 )
2915 )
2932 )
2916 raise
2933 raise
2917
2934
2918 def commithook(unused_success):
2935 def commithook(unused_success):
2919 # hack for command that use a temporary commit (eg: histedit)
2936 # hack for command that use a temporary commit (eg: histedit)
2920 # temporary commit got stripped before hook release
2937 # temporary commit got stripped before hook release
2921 if self.changelog.hasnode(ret):
2938 if self.changelog.hasnode(ret):
2922 self.hook(
2939 self.hook(
2923 b"commit", node=hex(ret), parent1=hookp1, parent2=hookp2
2940 b"commit", node=hex(ret), parent1=hookp1, parent2=hookp2
2924 )
2941 )
2925
2942
2926 self._afterlock(commithook)
2943 self._afterlock(commithook)
2927 return ret
2944 return ret
2928
2945
2929 @unfilteredmethod
2946 @unfilteredmethod
2930 def commitctx(self, ctx, error=False, origctx=None):
2947 def commitctx(self, ctx, error=False, origctx=None):
2931 return commit.commitctx(self, ctx, error=error, origctx=origctx)
2948 return commit.commitctx(self, ctx, error=error, origctx=origctx)
2932
2949
2933 @unfilteredmethod
2950 @unfilteredmethod
2934 def destroying(self):
2951 def destroying(self):
2935 '''Inform the repository that nodes are about to be destroyed.
2952 '''Inform the repository that nodes are about to be destroyed.
2936 Intended for use by strip and rollback, so there's a common
2953 Intended for use by strip and rollback, so there's a common
2937 place for anything that has to be done before destroying history.
2954 place for anything that has to be done before destroying history.
2938
2955
2939 This is mostly useful for saving state that is in memory and waiting
2956 This is mostly useful for saving state that is in memory and waiting
2940 to be flushed when the current lock is released. Because a call to
2957 to be flushed when the current lock is released. Because a call to
2941 destroyed is imminent, the repo will be invalidated causing those
2958 destroyed is imminent, the repo will be invalidated causing those
2942 changes to stay in memory (waiting for the next unlock), or vanish
2959 changes to stay in memory (waiting for the next unlock), or vanish
2943 completely.
2960 completely.
2944 '''
2961 '''
2945 # When using the same lock to commit and strip, the phasecache is left
2962 # When using the same lock to commit and strip, the phasecache is left
2946 # dirty after committing. Then when we strip, the repo is invalidated,
2963 # dirty after committing. Then when we strip, the repo is invalidated,
2947 # causing those changes to disappear.
2964 # causing those changes to disappear.
2948 if '_phasecache' in vars(self):
2965 if '_phasecache' in vars(self):
2949 self._phasecache.write()
2966 self._phasecache.write()
2950
2967
2951 @unfilteredmethod
2968 @unfilteredmethod
2952 def destroyed(self):
2969 def destroyed(self):
2953 '''Inform the repository that nodes have been destroyed.
2970 '''Inform the repository that nodes have been destroyed.
2954 Intended for use by strip and rollback, so there's a common
2971 Intended for use by strip and rollback, so there's a common
2955 place for anything that has to be done after destroying history.
2972 place for anything that has to be done after destroying history.
2956 '''
2973 '''
2957 # When one tries to:
2974 # When one tries to:
2958 # 1) destroy nodes thus calling this method (e.g. strip)
2975 # 1) destroy nodes thus calling this method (e.g. strip)
2959 # 2) use phasecache somewhere (e.g. commit)
2976 # 2) use phasecache somewhere (e.g. commit)
2960 #
2977 #
2961 # then 2) will fail because the phasecache contains nodes that were
2978 # then 2) will fail because the phasecache contains nodes that were
2962 # removed. We can either remove phasecache from the filecache,
2979 # removed. We can either remove phasecache from the filecache,
2963 # causing it to reload next time it is accessed, or simply filter
2980 # causing it to reload next time it is accessed, or simply filter
2964 # the removed nodes now and write the updated cache.
2981 # the removed nodes now and write the updated cache.
2965 self._phasecache.filterunknown(self)
2982 self._phasecache.filterunknown(self)
2966 self._phasecache.write()
2983 self._phasecache.write()
2967
2984
2968 # refresh all repository caches
2985 # refresh all repository caches
2969 self.updatecaches()
2986 self.updatecaches()
2970
2987
2971 # Ensure the persistent tag cache is updated. Doing it now
2988 # Ensure the persistent tag cache is updated. Doing it now
2972 # means that the tag cache only has to worry about destroyed
2989 # means that the tag cache only has to worry about destroyed
2973 # heads immediately after a strip/rollback. That in turn
2990 # heads immediately after a strip/rollback. That in turn
2974 # guarantees that "cachetip == currenttip" (comparing both rev
2991 # guarantees that "cachetip == currenttip" (comparing both rev
2975 # and node) always means no nodes have been added or destroyed.
2992 # and node) always means no nodes have been added or destroyed.
2976
2993
2977 # XXX this is suboptimal when qrefresh'ing: we strip the current
2994 # XXX this is suboptimal when qrefresh'ing: we strip the current
2978 # head, refresh the tag cache, then immediately add a new head.
2995 # head, refresh the tag cache, then immediately add a new head.
2979 # But I think doing it this way is necessary for the "instant
2996 # But I think doing it this way is necessary for the "instant
2980 # tag cache retrieval" case to work.
2997 # tag cache retrieval" case to work.
2981 self.invalidate()
2998 self.invalidate()
2982
2999
2983 def status(
3000 def status(
2984 self,
3001 self,
2985 node1=b'.',
3002 node1=b'.',
2986 node2=None,
3003 node2=None,
2987 match=None,
3004 match=None,
2988 ignored=False,
3005 ignored=False,
2989 clean=False,
3006 clean=False,
2990 unknown=False,
3007 unknown=False,
2991 listsubrepos=False,
3008 listsubrepos=False,
2992 ):
3009 ):
2993 '''a convenience method that calls node1.status(node2)'''
3010 '''a convenience method that calls node1.status(node2)'''
2994 return self[node1].status(
3011 return self[node1].status(
2995 node2, match, ignored, clean, unknown, listsubrepos
3012 node2, match, ignored, clean, unknown, listsubrepos
2996 )
3013 )
2997
3014
2998 def addpostdsstatus(self, ps):
3015 def addpostdsstatus(self, ps):
2999 """Add a callback to run within the wlock, at the point at which status
3016 """Add a callback to run within the wlock, at the point at which status
3000 fixups happen.
3017 fixups happen.
3001
3018
3002 On status completion, callback(wctx, status) will be called with the
3019 On status completion, callback(wctx, status) will be called with the
3003 wlock held, unless the dirstate has changed from underneath or the wlock
3020 wlock held, unless the dirstate has changed from underneath or the wlock
3004 couldn't be grabbed.
3021 couldn't be grabbed.
3005
3022
3006 Callbacks should not capture and use a cached copy of the dirstate --
3023 Callbacks should not capture and use a cached copy of the dirstate --
3007 it might change in the meanwhile. Instead, they should access the
3024 it might change in the meanwhile. Instead, they should access the
3008 dirstate via wctx.repo().dirstate.
3025 dirstate via wctx.repo().dirstate.
3009
3026
3010 This list is emptied out after each status run -- extensions should
3027 This list is emptied out after each status run -- extensions should
3011 make sure it adds to this list each time dirstate.status is called.
3028 make sure it adds to this list each time dirstate.status is called.
3012 Extensions should also make sure they don't call this for statuses
3029 Extensions should also make sure they don't call this for statuses
3013 that don't involve the dirstate.
3030 that don't involve the dirstate.
3014 """
3031 """
3015
3032
3016 # The list is located here for uniqueness reasons -- it is actually
3033 # The list is located here for uniqueness reasons -- it is actually
3017 # managed by the workingctx, but that isn't unique per-repo.
3034 # managed by the workingctx, but that isn't unique per-repo.
3018 self._postdsstatus.append(ps)
3035 self._postdsstatus.append(ps)
3019
3036
3020 def postdsstatus(self):
3037 def postdsstatus(self):
3021 """Used by workingctx to get the list of post-dirstate-status hooks."""
3038 """Used by workingctx to get the list of post-dirstate-status hooks."""
3022 return self._postdsstatus
3039 return self._postdsstatus
3023
3040
3024 def clearpostdsstatus(self):
3041 def clearpostdsstatus(self):
3025 """Used by workingctx to clear post-dirstate-status hooks."""
3042 """Used by workingctx to clear post-dirstate-status hooks."""
3026 del self._postdsstatus[:]
3043 del self._postdsstatus[:]
3027
3044
3028 def heads(self, start=None):
3045 def heads(self, start=None):
3029 if start is None:
3046 if start is None:
3030 cl = self.changelog
3047 cl = self.changelog
3031 headrevs = reversed(cl.headrevs())
3048 headrevs = reversed(cl.headrevs())
3032 return [cl.node(rev) for rev in headrevs]
3049 return [cl.node(rev) for rev in headrevs]
3033
3050
3034 heads = self.changelog.heads(start)
3051 heads = self.changelog.heads(start)
3035 # sort the output in rev descending order
3052 # sort the output in rev descending order
3036 return sorted(heads, key=self.changelog.rev, reverse=True)
3053 return sorted(heads, key=self.changelog.rev, reverse=True)
3037
3054
3038 def branchheads(self, branch=None, start=None, closed=False):
3055 def branchheads(self, branch=None, start=None, closed=False):
3039 '''return a (possibly filtered) list of heads for the given branch
3056 '''return a (possibly filtered) list of heads for the given branch
3040
3057
3041 Heads are returned in topological order, from newest to oldest.
3058 Heads are returned in topological order, from newest to oldest.
3042 If branch is None, use the dirstate branch.
3059 If branch is None, use the dirstate branch.
3043 If start is not None, return only heads reachable from start.
3060 If start is not None, return only heads reachable from start.
3044 If closed is True, return heads that are marked as closed as well.
3061 If closed is True, return heads that are marked as closed as well.
3045 '''
3062 '''
3046 if branch is None:
3063 if branch is None:
3047 branch = self[None].branch()
3064 branch = self[None].branch()
3048 branches = self.branchmap()
3065 branches = self.branchmap()
3049 if not branches.hasbranch(branch):
3066 if not branches.hasbranch(branch):
3050 return []
3067 return []
3051 # the cache returns heads ordered lowest to highest
3068 # the cache returns heads ordered lowest to highest
3052 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
3069 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
3053 if start is not None:
3070 if start is not None:
3054 # filter out the heads that cannot be reached from startrev
3071 # filter out the heads that cannot be reached from startrev
3055 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
3072 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
3056 bheads = [h for h in bheads if h in fbheads]
3073 bheads = [h for h in bheads if h in fbheads]
3057 return bheads
3074 return bheads
3058
3075
3059 def branches(self, nodes):
3076 def branches(self, nodes):
3060 if not nodes:
3077 if not nodes:
3061 nodes = [self.changelog.tip()]
3078 nodes = [self.changelog.tip()]
3062 b = []
3079 b = []
3063 for n in nodes:
3080 for n in nodes:
3064 t = n
3081 t = n
3065 while True:
3082 while True:
3066 p = self.changelog.parents(n)
3083 p = self.changelog.parents(n)
3067 if p[1] != nullid or p[0] == nullid:
3084 if p[1] != nullid or p[0] == nullid:
3068 b.append((t, n, p[0], p[1]))
3085 b.append((t, n, p[0], p[1]))
3069 break
3086 break
3070 n = p[0]
3087 n = p[0]
3071 return b
3088 return b
3072
3089
3073 def between(self, pairs):
3090 def between(self, pairs):
3074 r = []
3091 r = []
3075
3092
3076 for top, bottom in pairs:
3093 for top, bottom in pairs:
3077 n, l, i = top, [], 0
3094 n, l, i = top, [], 0
3078 f = 1
3095 f = 1
3079
3096
3080 while n != bottom and n != nullid:
3097 while n != bottom and n != nullid:
3081 p = self.changelog.parents(n)[0]
3098 p = self.changelog.parents(n)[0]
3082 if i == f:
3099 if i == f:
3083 l.append(n)
3100 l.append(n)
3084 f = f * 2
3101 f = f * 2
3085 n = p
3102 n = p
3086 i += 1
3103 i += 1
3087
3104
3088 r.append(l)
3105 r.append(l)
3089
3106
3090 return r
3107 return r
3091
3108
3092 def checkpush(self, pushop):
3109 def checkpush(self, pushop):
3093 """Extensions can override this function if additional checks have
3110 """Extensions can override this function if additional checks have
3094 to be performed before pushing, or call it if they override push
3111 to be performed before pushing, or call it if they override push
3095 command.
3112 command.
3096 """
3113 """
3097
3114
3098 @unfilteredpropertycache
3115 @unfilteredpropertycache
3099 def prepushoutgoinghooks(self):
3116 def prepushoutgoinghooks(self):
3100 """Return util.hooks consists of a pushop with repo, remote, outgoing
3117 """Return util.hooks consists of a pushop with repo, remote, outgoing
3101 methods, which are called before pushing changesets.
3118 methods, which are called before pushing changesets.
3102 """
3119 """
3103 return util.hooks()
3120 return util.hooks()
3104
3121
3105 def pushkey(self, namespace, key, old, new):
3122 def pushkey(self, namespace, key, old, new):
3106 try:
3123 try:
3107 tr = self.currenttransaction()
3124 tr = self.currenttransaction()
3108 hookargs = {}
3125 hookargs = {}
3109 if tr is not None:
3126 if tr is not None:
3110 hookargs.update(tr.hookargs)
3127 hookargs.update(tr.hookargs)
3111 hookargs = pycompat.strkwargs(hookargs)
3128 hookargs = pycompat.strkwargs(hookargs)
3112 hookargs['namespace'] = namespace
3129 hookargs['namespace'] = namespace
3113 hookargs['key'] = key
3130 hookargs['key'] = key
3114 hookargs['old'] = old
3131 hookargs['old'] = old
3115 hookargs['new'] = new
3132 hookargs['new'] = new
3116 self.hook(b'prepushkey', throw=True, **hookargs)
3133 self.hook(b'prepushkey', throw=True, **hookargs)
3117 except error.HookAbort as exc:
3134 except error.HookAbort as exc:
3118 self.ui.write_err(_(b"pushkey-abort: %s\n") % exc)
3135 self.ui.write_err(_(b"pushkey-abort: %s\n") % exc)
3119 if exc.hint:
3136 if exc.hint:
3120 self.ui.write_err(_(b"(%s)\n") % exc.hint)
3137 self.ui.write_err(_(b"(%s)\n") % exc.hint)
3121 return False
3138 return False
3122 self.ui.debug(b'pushing key for "%s:%s"\n' % (namespace, key))
3139 self.ui.debug(b'pushing key for "%s:%s"\n' % (namespace, key))
3123 ret = pushkey.push(self, namespace, key, old, new)
3140 ret = pushkey.push(self, namespace, key, old, new)
3124
3141
3125 def runhook(unused_success):
3142 def runhook(unused_success):
3126 self.hook(
3143 self.hook(
3127 b'pushkey',
3144 b'pushkey',
3128 namespace=namespace,
3145 namespace=namespace,
3129 key=key,
3146 key=key,
3130 old=old,
3147 old=old,
3131 new=new,
3148 new=new,
3132 ret=ret,
3149 ret=ret,
3133 )
3150 )
3134
3151
3135 self._afterlock(runhook)
3152 self._afterlock(runhook)
3136 return ret
3153 return ret
3137
3154
3138 def listkeys(self, namespace):
3155 def listkeys(self, namespace):
3139 self.hook(b'prelistkeys', throw=True, namespace=namespace)
3156 self.hook(b'prelistkeys', throw=True, namespace=namespace)
3140 self.ui.debug(b'listing keys for "%s"\n' % namespace)
3157 self.ui.debug(b'listing keys for "%s"\n' % namespace)
3141 values = pushkey.list(self, namespace)
3158 values = pushkey.list(self, namespace)
3142 self.hook(b'listkeys', namespace=namespace, values=values)
3159 self.hook(b'listkeys', namespace=namespace, values=values)
3143 return values
3160 return values
3144
3161
3145 def debugwireargs(self, one, two, three=None, four=None, five=None):
3162 def debugwireargs(self, one, two, three=None, four=None, five=None):
3146 '''used to test argument passing over the wire'''
3163 '''used to test argument passing over the wire'''
3147 return b"%s %s %s %s %s" % (
3164 return b"%s %s %s %s %s" % (
3148 one,
3165 one,
3149 two,
3166 two,
3150 pycompat.bytestr(three),
3167 pycompat.bytestr(three),
3151 pycompat.bytestr(four),
3168 pycompat.bytestr(four),
3152 pycompat.bytestr(five),
3169 pycompat.bytestr(five),
3153 )
3170 )
3154
3171
3155 def savecommitmessage(self, text):
3172 def savecommitmessage(self, text):
3156 fp = self.vfs(b'last-message.txt', b'wb')
3173 fp = self.vfs(b'last-message.txt', b'wb')
3157 try:
3174 try:
3158 fp.write(text)
3175 fp.write(text)
3159 finally:
3176 finally:
3160 fp.close()
3177 fp.close()
3161 return self.pathto(fp.name[len(self.root) + 1 :])
3178 return self.pathto(fp.name[len(self.root) + 1 :])
3162
3179
3163
3180
3164 # used to avoid circular references so destructors work
3181 # used to avoid circular references so destructors work
3165 def aftertrans(files):
3182 def aftertrans(files):
3166 renamefiles = [tuple(t) for t in files]
3183 renamefiles = [tuple(t) for t in files]
3167
3184
3168 def a():
3185 def a():
3169 for vfs, src, dest in renamefiles:
3186 for vfs, src, dest in renamefiles:
3170 # if src and dest refer to a same file, vfs.rename is a no-op,
3187 # if src and dest refer to a same file, vfs.rename is a no-op,
3171 # leaving both src and dest on disk. delete dest to make sure
3188 # leaving both src and dest on disk. delete dest to make sure
3172 # the rename couldn't be such a no-op.
3189 # the rename couldn't be such a no-op.
3173 vfs.tryunlink(dest)
3190 vfs.tryunlink(dest)
3174 try:
3191 try:
3175 vfs.rename(src, dest)
3192 vfs.rename(src, dest)
3176 except OSError: # journal file does not yet exist
3193 except OSError: # journal file does not yet exist
3177 pass
3194 pass
3178
3195
3179 return a
3196 return a
3180
3197
3181
3198
3182 def undoname(fn):
3199 def undoname(fn):
3183 base, name = os.path.split(fn)
3200 base, name = os.path.split(fn)
3184 assert name.startswith(b'journal')
3201 assert name.startswith(b'journal')
3185 return os.path.join(base, name.replace(b'journal', b'undo', 1))
3202 return os.path.join(base, name.replace(b'journal', b'undo', 1))
3186
3203
3187
3204
3188 def instance(ui, path, create, intents=None, createopts=None):
3205 def instance(ui, path, create, intents=None, createopts=None):
3189 localpath = util.urllocalpath(path)
3206 localpath = util.urllocalpath(path)
3190 if create:
3207 if create:
3191 createrepository(ui, localpath, createopts=createopts)
3208 createrepository(ui, localpath, createopts=createopts)
3192
3209
3193 return makelocalrepository(ui, localpath, intents=intents)
3210 return makelocalrepository(ui, localpath, intents=intents)
3194
3211
3195
3212
3196 def islocal(path):
3213 def islocal(path):
3197 return True
3214 return True
3198
3215
3199
3216
3200 def defaultcreateopts(ui, createopts=None):
3217 def defaultcreateopts(ui, createopts=None):
3201 """Populate the default creation options for a repository.
3218 """Populate the default creation options for a repository.
3202
3219
3203 A dictionary of explicitly requested creation options can be passed
3220 A dictionary of explicitly requested creation options can be passed
3204 in. Missing keys will be populated.
3221 in. Missing keys will be populated.
3205 """
3222 """
3206 createopts = dict(createopts or {})
3223 createopts = dict(createopts or {})
3207
3224
3208 if b'backend' not in createopts:
3225 if b'backend' not in createopts:
3209 # experimental config: storage.new-repo-backend
3226 # experimental config: storage.new-repo-backend
3210 createopts[b'backend'] = ui.config(b'storage', b'new-repo-backend')
3227 createopts[b'backend'] = ui.config(b'storage', b'new-repo-backend')
3211
3228
3212 return createopts
3229 return createopts
3213
3230
3214
3231
3215 def newreporequirements(ui, createopts):
3232 def newreporequirements(ui, createopts):
3216 """Determine the set of requirements for a new local repository.
3233 """Determine the set of requirements for a new local repository.
3217
3234
3218 Extensions can wrap this function to specify custom requirements for
3235 Extensions can wrap this function to specify custom requirements for
3219 new repositories.
3236 new repositories.
3220 """
3237 """
3221 # If the repo is being created from a shared repository, we copy
3238 # If the repo is being created from a shared repository, we copy
3222 # its requirements.
3239 # its requirements.
3223 if b'sharedrepo' in createopts:
3240 if b'sharedrepo' in createopts:
3224 requirements = set(createopts[b'sharedrepo'].requirements)
3241 requirements = set(createopts[b'sharedrepo'].requirements)
3225 if createopts.get(b'sharedrelative'):
3242 if createopts.get(b'sharedrelative'):
3226 requirements.add(b'relshared')
3243 requirements.add(b'relshared')
3227 else:
3244 else:
3228 requirements.add(b'shared')
3245 requirements.add(b'shared')
3229
3246
3230 return requirements
3247 return requirements
3231
3248
3232 if b'backend' not in createopts:
3249 if b'backend' not in createopts:
3233 raise error.ProgrammingError(
3250 raise error.ProgrammingError(
3234 b'backend key not present in createopts; '
3251 b'backend key not present in createopts; '
3235 b'was defaultcreateopts() called?'
3252 b'was defaultcreateopts() called?'
3236 )
3253 )
3237
3254
3238 if createopts[b'backend'] != b'revlogv1':
3255 if createopts[b'backend'] != b'revlogv1':
3239 raise error.Abort(
3256 raise error.Abort(
3240 _(
3257 _(
3241 b'unable to determine repository requirements for '
3258 b'unable to determine repository requirements for '
3242 b'storage backend: %s'
3259 b'storage backend: %s'
3243 )
3260 )
3244 % createopts[b'backend']
3261 % createopts[b'backend']
3245 )
3262 )
3246
3263
3247 requirements = {b'revlogv1'}
3264 requirements = {b'revlogv1'}
3248 if ui.configbool(b'format', b'usestore'):
3265 if ui.configbool(b'format', b'usestore'):
3249 requirements.add(b'store')
3266 requirements.add(b'store')
3250 if ui.configbool(b'format', b'usefncache'):
3267 if ui.configbool(b'format', b'usefncache'):
3251 requirements.add(b'fncache')
3268 requirements.add(b'fncache')
3252 if ui.configbool(b'format', b'dotencode'):
3269 if ui.configbool(b'format', b'dotencode'):
3253 requirements.add(b'dotencode')
3270 requirements.add(b'dotencode')
3254
3271
3255 compengines = ui.configlist(b'format', b'revlog-compression')
3272 compengines = ui.configlist(b'format', b'revlog-compression')
3256 for compengine in compengines:
3273 for compengine in compengines:
3257 if compengine in util.compengines:
3274 if compengine in util.compengines:
3258 break
3275 break
3259 else:
3276 else:
3260 raise error.Abort(
3277 raise error.Abort(
3261 _(
3278 _(
3262 b'compression engines %s defined by '
3279 b'compression engines %s defined by '
3263 b'format.revlog-compression not available'
3280 b'format.revlog-compression not available'
3264 )
3281 )
3265 % b', '.join(b'"%s"' % e for e in compengines),
3282 % b', '.join(b'"%s"' % e for e in compengines),
3266 hint=_(
3283 hint=_(
3267 b'run "hg debuginstall" to list available '
3284 b'run "hg debuginstall" to list available '
3268 b'compression engines'
3285 b'compression engines'
3269 ),
3286 ),
3270 )
3287 )
3271
3288
3272 # zlib is the historical default and doesn't need an explicit requirement.
3289 # zlib is the historical default and doesn't need an explicit requirement.
3273 if compengine == b'zstd':
3290 if compengine == b'zstd':
3274 requirements.add(b'revlog-compression-zstd')
3291 requirements.add(b'revlog-compression-zstd')
3275 elif compengine != b'zlib':
3292 elif compengine != b'zlib':
3276 requirements.add(b'exp-compression-%s' % compengine)
3293 requirements.add(b'exp-compression-%s' % compengine)
3277
3294
3278 if scmutil.gdinitconfig(ui):
3295 if scmutil.gdinitconfig(ui):
3279 requirements.add(b'generaldelta')
3296 requirements.add(b'generaldelta')
3280 if ui.configbool(b'format', b'sparse-revlog'):
3297 if ui.configbool(b'format', b'sparse-revlog'):
3281 requirements.add(SPARSEREVLOG_REQUIREMENT)
3298 requirements.add(SPARSEREVLOG_REQUIREMENT)
3282
3299
3283 # experimental config: format.exp-use-side-data
3300 # experimental config: format.exp-use-side-data
3284 if ui.configbool(b'format', b'exp-use-side-data'):
3301 if ui.configbool(b'format', b'exp-use-side-data'):
3285 requirements.add(SIDEDATA_REQUIREMENT)
3302 requirements.add(SIDEDATA_REQUIREMENT)
3286 # experimental config: format.exp-use-copies-side-data-changeset
3303 # experimental config: format.exp-use-copies-side-data-changeset
3287 if ui.configbool(b'format', b'exp-use-copies-side-data-changeset'):
3304 if ui.configbool(b'format', b'exp-use-copies-side-data-changeset'):
3288 requirements.add(SIDEDATA_REQUIREMENT)
3305 requirements.add(SIDEDATA_REQUIREMENT)
3289 requirements.add(COPIESSDC_REQUIREMENT)
3306 requirements.add(COPIESSDC_REQUIREMENT)
3290 if ui.configbool(b'experimental', b'treemanifest'):
3307 if ui.configbool(b'experimental', b'treemanifest'):
3291 requirements.add(b'treemanifest')
3308 requirements.add(b'treemanifest')
3292
3309
3293 revlogv2 = ui.config(b'experimental', b'revlogv2')
3310 revlogv2 = ui.config(b'experimental', b'revlogv2')
3294 if revlogv2 == b'enable-unstable-format-and-corrupt-my-data':
3311 if revlogv2 == b'enable-unstable-format-and-corrupt-my-data':
3295 requirements.remove(b'revlogv1')
3312 requirements.remove(b'revlogv1')
3296 # generaldelta is implied by revlogv2.
3313 # generaldelta is implied by revlogv2.
3297 requirements.discard(b'generaldelta')
3314 requirements.discard(b'generaldelta')
3298 requirements.add(REVLOGV2_REQUIREMENT)
3315 requirements.add(REVLOGV2_REQUIREMENT)
3299 # experimental config: format.internal-phase
3316 # experimental config: format.internal-phase
3300 if ui.configbool(b'format', b'internal-phase'):
3317 if ui.configbool(b'format', b'internal-phase'):
3301 requirements.add(b'internal-phase')
3318 requirements.add(b'internal-phase')
3302
3319
3303 if createopts.get(b'narrowfiles'):
3320 if createopts.get(b'narrowfiles'):
3304 requirements.add(repository.NARROW_REQUIREMENT)
3321 requirements.add(repository.NARROW_REQUIREMENT)
3305
3322
3306 if createopts.get(b'lfs'):
3323 if createopts.get(b'lfs'):
3307 requirements.add(b'lfs')
3324 requirements.add(b'lfs')
3308
3325
3309 if ui.configbool(b'format', b'bookmarks-in-store'):
3326 if ui.configbool(b'format', b'bookmarks-in-store'):
3310 requirements.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3327 requirements.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3311
3328
3312 if ui.configbool(b'format', b'use-persistent-nodemap'):
3329 if ui.configbool(b'format', b'use-persistent-nodemap'):
3313 requirements.add(NODEMAP_REQUIREMENT)
3330 requirements.add(NODEMAP_REQUIREMENT)
3314
3331
3315 return requirements
3332 return requirements
3316
3333
3317
3334
3318 def checkrequirementscompat(ui, requirements):
3335 def checkrequirementscompat(ui, requirements):
3319 """ Checks compatibility of repository requirements enabled and disabled.
3336 """ Checks compatibility of repository requirements enabled and disabled.
3320
3337
3321 Returns a set of requirements which needs to be dropped because dependend
3338 Returns a set of requirements which needs to be dropped because dependend
3322 requirements are not enabled. Also warns users about it """
3339 requirements are not enabled. Also warns users about it """
3323
3340
3324 dropped = set()
3341 dropped = set()
3325
3342
3326 if b'store' not in requirements:
3343 if b'store' not in requirements:
3327 if bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT in requirements:
3344 if bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT in requirements:
3328 ui.warn(
3345 ui.warn(
3329 _(
3346 _(
3330 b'ignoring enabled \'format.bookmarks-in-store\' config '
3347 b'ignoring enabled \'format.bookmarks-in-store\' config '
3331 b'beacuse it is incompatible with disabled '
3348 b'beacuse it is incompatible with disabled '
3332 b'\'format.usestore\' config\n'
3349 b'\'format.usestore\' config\n'
3333 )
3350 )
3334 )
3351 )
3335 dropped.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3352 dropped.add(bookmarks.BOOKMARKS_IN_STORE_REQUIREMENT)
3336
3353
3337 if b'shared' in requirements or b'relshared' in requirements:
3354 if b'shared' in requirements or b'relshared' in requirements:
3338 raise error.Abort(
3355 raise error.Abort(
3339 _(
3356 _(
3340 b"cannot create shared repository as source was created"
3357 b"cannot create shared repository as source was created"
3341 b" with 'format.usestore' config disabled"
3358 b" with 'format.usestore' config disabled"
3342 )
3359 )
3343 )
3360 )
3344
3361
3345 return dropped
3362 return dropped
3346
3363
3347
3364
3348 def filterknowncreateopts(ui, createopts):
3365 def filterknowncreateopts(ui, createopts):
3349 """Filters a dict of repo creation options against options that are known.
3366 """Filters a dict of repo creation options against options that are known.
3350
3367
3351 Receives a dict of repo creation options and returns a dict of those
3368 Receives a dict of repo creation options and returns a dict of those
3352 options that we don't know how to handle.
3369 options that we don't know how to handle.
3353
3370
3354 This function is called as part of repository creation. If the
3371 This function is called as part of repository creation. If the
3355 returned dict contains any items, repository creation will not
3372 returned dict contains any items, repository creation will not
3356 be allowed, as it means there was a request to create a repository
3373 be allowed, as it means there was a request to create a repository
3357 with options not recognized by loaded code.
3374 with options not recognized by loaded code.
3358
3375
3359 Extensions can wrap this function to filter out creation options
3376 Extensions can wrap this function to filter out creation options
3360 they know how to handle.
3377 they know how to handle.
3361 """
3378 """
3362 known = {
3379 known = {
3363 b'backend',
3380 b'backend',
3364 b'lfs',
3381 b'lfs',
3365 b'narrowfiles',
3382 b'narrowfiles',
3366 b'sharedrepo',
3383 b'sharedrepo',
3367 b'sharedrelative',
3384 b'sharedrelative',
3368 b'shareditems',
3385 b'shareditems',
3369 b'shallowfilestore',
3386 b'shallowfilestore',
3370 }
3387 }
3371
3388
3372 return {k: v for k, v in createopts.items() if k not in known}
3389 return {k: v for k, v in createopts.items() if k not in known}
3373
3390
3374
3391
3375 def createrepository(ui, path, createopts=None):
3392 def createrepository(ui, path, createopts=None):
3376 """Create a new repository in a vfs.
3393 """Create a new repository in a vfs.
3377
3394
3378 ``path`` path to the new repo's working directory.
3395 ``path`` path to the new repo's working directory.
3379 ``createopts`` options for the new repository.
3396 ``createopts`` options for the new repository.
3380
3397
3381 The following keys for ``createopts`` are recognized:
3398 The following keys for ``createopts`` are recognized:
3382
3399
3383 backend
3400 backend
3384 The storage backend to use.
3401 The storage backend to use.
3385 lfs
3402 lfs
3386 Repository will be created with ``lfs`` requirement. The lfs extension
3403 Repository will be created with ``lfs`` requirement. The lfs extension
3387 will automatically be loaded when the repository is accessed.
3404 will automatically be loaded when the repository is accessed.
3388 narrowfiles
3405 narrowfiles
3389 Set up repository to support narrow file storage.
3406 Set up repository to support narrow file storage.
3390 sharedrepo
3407 sharedrepo
3391 Repository object from which storage should be shared.
3408 Repository object from which storage should be shared.
3392 sharedrelative
3409 sharedrelative
3393 Boolean indicating if the path to the shared repo should be
3410 Boolean indicating if the path to the shared repo should be
3394 stored as relative. By default, the pointer to the "parent" repo
3411 stored as relative. By default, the pointer to the "parent" repo
3395 is stored as an absolute path.
3412 is stored as an absolute path.
3396 shareditems
3413 shareditems
3397 Set of items to share to the new repository (in addition to storage).
3414 Set of items to share to the new repository (in addition to storage).
3398 shallowfilestore
3415 shallowfilestore
3399 Indicates that storage for files should be shallow (not all ancestor
3416 Indicates that storage for files should be shallow (not all ancestor
3400 revisions are known).
3417 revisions are known).
3401 """
3418 """
3402 createopts = defaultcreateopts(ui, createopts=createopts)
3419 createopts = defaultcreateopts(ui, createopts=createopts)
3403
3420
3404 unknownopts = filterknowncreateopts(ui, createopts)
3421 unknownopts = filterknowncreateopts(ui, createopts)
3405
3422
3406 if not isinstance(unknownopts, dict):
3423 if not isinstance(unknownopts, dict):
3407 raise error.ProgrammingError(
3424 raise error.ProgrammingError(
3408 b'filterknowncreateopts() did not return a dict'
3425 b'filterknowncreateopts() did not return a dict'
3409 )
3426 )
3410
3427
3411 if unknownopts:
3428 if unknownopts:
3412 raise error.Abort(
3429 raise error.Abort(
3413 _(
3430 _(
3414 b'unable to create repository because of unknown '
3431 b'unable to create repository because of unknown '
3415 b'creation option: %s'
3432 b'creation option: %s'
3416 )
3433 )
3417 % b', '.join(sorted(unknownopts)),
3434 % b', '.join(sorted(unknownopts)),
3418 hint=_(b'is a required extension not loaded?'),
3435 hint=_(b'is a required extension not loaded?'),
3419 )
3436 )
3420
3437
3421 requirements = newreporequirements(ui, createopts=createopts)
3438 requirements = newreporequirements(ui, createopts=createopts)
3422 requirements -= checkrequirementscompat(ui, requirements)
3439 requirements -= checkrequirementscompat(ui, requirements)
3423
3440
3424 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
3441 wdirvfs = vfsmod.vfs(path, expandpath=True, realpath=True)
3425
3442
3426 hgvfs = vfsmod.vfs(wdirvfs.join(b'.hg'))
3443 hgvfs = vfsmod.vfs(wdirvfs.join(b'.hg'))
3427 if hgvfs.exists():
3444 if hgvfs.exists():
3428 raise error.RepoError(_(b'repository %s already exists') % path)
3445 raise error.RepoError(_(b'repository %s already exists') % path)
3429
3446
3430 if b'sharedrepo' in createopts:
3447 if b'sharedrepo' in createopts:
3431 sharedpath = createopts[b'sharedrepo'].sharedpath
3448 sharedpath = createopts[b'sharedrepo'].sharedpath
3432
3449
3433 if createopts.get(b'sharedrelative'):
3450 if createopts.get(b'sharedrelative'):
3434 try:
3451 try:
3435 sharedpath = os.path.relpath(sharedpath, hgvfs.base)
3452 sharedpath = os.path.relpath(sharedpath, hgvfs.base)
3436 except (IOError, ValueError) as e:
3453 except (IOError, ValueError) as e:
3437 # ValueError is raised on Windows if the drive letters differ
3454 # ValueError is raised on Windows if the drive letters differ
3438 # on each path.
3455 # on each path.
3439 raise error.Abort(
3456 raise error.Abort(
3440 _(b'cannot calculate relative path'),
3457 _(b'cannot calculate relative path'),
3441 hint=stringutil.forcebytestr(e),
3458 hint=stringutil.forcebytestr(e),
3442 )
3459 )
3443
3460
3444 if not wdirvfs.exists():
3461 if not wdirvfs.exists():
3445 wdirvfs.makedirs()
3462 wdirvfs.makedirs()
3446
3463
3447 hgvfs.makedir(notindexed=True)
3464 hgvfs.makedir(notindexed=True)
3448 if b'sharedrepo' not in createopts:
3465 if b'sharedrepo' not in createopts:
3449 hgvfs.mkdir(b'cache')
3466 hgvfs.mkdir(b'cache')
3450 hgvfs.mkdir(b'wcache')
3467 hgvfs.mkdir(b'wcache')
3451
3468
3452 if b'store' in requirements and b'sharedrepo' not in createopts:
3469 if b'store' in requirements and b'sharedrepo' not in createopts:
3453 hgvfs.mkdir(b'store')
3470 hgvfs.mkdir(b'store')
3454
3471
3455 # We create an invalid changelog outside the store so very old
3472 # We create an invalid changelog outside the store so very old
3456 # Mercurial versions (which didn't know about the requirements
3473 # Mercurial versions (which didn't know about the requirements
3457 # file) encounter an error on reading the changelog. This
3474 # file) encounter an error on reading the changelog. This
3458 # effectively locks out old clients and prevents them from
3475 # effectively locks out old clients and prevents them from
3459 # mucking with a repo in an unknown format.
3476 # mucking with a repo in an unknown format.
3460 #
3477 #
3461 # The revlog header has version 2, which won't be recognized by
3478 # The revlog header has version 2, which won't be recognized by
3462 # such old clients.
3479 # such old clients.
3463 hgvfs.append(
3480 hgvfs.append(
3464 b'00changelog.i',
3481 b'00changelog.i',
3465 b'\0\0\0\2 dummy changelog to prevent using the old repo '
3482 b'\0\0\0\2 dummy changelog to prevent using the old repo '
3466 b'layout',
3483 b'layout',
3467 )
3484 )
3468
3485
3469 scmutil.writerequires(hgvfs, requirements)
3486 scmutil.writerequires(hgvfs, requirements)
3470
3487
3471 # Write out file telling readers where to find the shared store.
3488 # Write out file telling readers where to find the shared store.
3472 if b'sharedrepo' in createopts:
3489 if b'sharedrepo' in createopts:
3473 hgvfs.write(b'sharedpath', sharedpath)
3490 hgvfs.write(b'sharedpath', sharedpath)
3474
3491
3475 if createopts.get(b'shareditems'):
3492 if createopts.get(b'shareditems'):
3476 shared = b'\n'.join(sorted(createopts[b'shareditems'])) + b'\n'
3493 shared = b'\n'.join(sorted(createopts[b'shareditems'])) + b'\n'
3477 hgvfs.write(b'shared', shared)
3494 hgvfs.write(b'shared', shared)
3478
3495
3479
3496
3480 def poisonrepository(repo):
3497 def poisonrepository(repo):
3481 """Poison a repository instance so it can no longer be used."""
3498 """Poison a repository instance so it can no longer be used."""
3482 # Perform any cleanup on the instance.
3499 # Perform any cleanup on the instance.
3483 repo.close()
3500 repo.close()
3484
3501
3485 # Our strategy is to replace the type of the object with one that
3502 # Our strategy is to replace the type of the object with one that
3486 # has all attribute lookups result in error.
3503 # has all attribute lookups result in error.
3487 #
3504 #
3488 # But we have to allow the close() method because some constructors
3505 # But we have to allow the close() method because some constructors
3489 # of repos call close() on repo references.
3506 # of repos call close() on repo references.
3490 class poisonedrepository(object):
3507 class poisonedrepository(object):
3491 def __getattribute__(self, item):
3508 def __getattribute__(self, item):
3492 if item == 'close':
3509 if item == 'close':
3493 return object.__getattribute__(self, item)
3510 return object.__getattribute__(self, item)
3494
3511
3495 raise error.ProgrammingError(
3512 raise error.ProgrammingError(
3496 b'repo instances should not be used after unshare'
3513 b'repo instances should not be used after unshare'
3497 )
3514 )
3498
3515
3499 def close(self):
3516 def close(self):
3500 pass
3517 pass
3501
3518
3502 # We may have a repoview, which intercepts __setattr__. So be sure
3519 # We may have a repoview, which intercepts __setattr__. So be sure
3503 # we operate at the lowest level possible.
3520 # we operate at the lowest level possible.
3504 object.__setattr__(repo, '__class__', poisonedrepository)
3521 object.__setattr__(repo, '__class__', poisonedrepository)
General Comments 0
You need to be logged in to leave comments. Login now