##// END OF EJS Templates
rollback: clear resolve state (issue4593)
Matt Mackall -
r24784:59406b8b default
parent child Browse files
Show More
@@ -1,1928 +1,1931 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import urllib
9 import urllib
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
12 import lock as lockmod
12 import lock as lockmod
13 import transaction, store, encoding, exchange, bundle2
13 import transaction, store, encoding, exchange, bundle2
14 import scmutil, util, extensions, hook, error, revset
14 import scmutil, util, extensions, hook, error, revset
15 import match as matchmod
15 import match as matchmod
16 import merge as mergemod
16 import merge as mergemod
17 import tags as tagsmod
17 import tags as tagsmod
18 from lock import release
18 from lock import release
19 import weakref, errno, os, time, inspect
19 import weakref, errno, os, time, inspect
20 import branchmap, pathutil
20 import branchmap, pathutil
21 import namespaces
21 import namespaces
22 propertycache = util.propertycache
22 propertycache = util.propertycache
23 filecache = scmutil.filecache
23 filecache = scmutil.filecache
24
24
25 class repofilecache(filecache):
25 class repofilecache(filecache):
26 """All filecache usage on repo are done for logic that should be unfiltered
26 """All filecache usage on repo are done for logic that should be unfiltered
27 """
27 """
28
28
29 def __get__(self, repo, type=None):
29 def __get__(self, repo, type=None):
30 return super(repofilecache, self).__get__(repo.unfiltered(), type)
30 return super(repofilecache, self).__get__(repo.unfiltered(), type)
31 def __set__(self, repo, value):
31 def __set__(self, repo, value):
32 return super(repofilecache, self).__set__(repo.unfiltered(), value)
32 return super(repofilecache, self).__set__(repo.unfiltered(), value)
33 def __delete__(self, repo):
33 def __delete__(self, repo):
34 return super(repofilecache, self).__delete__(repo.unfiltered())
34 return super(repofilecache, self).__delete__(repo.unfiltered())
35
35
36 class storecache(repofilecache):
36 class storecache(repofilecache):
37 """filecache for files in the store"""
37 """filecache for files in the store"""
38 def join(self, obj, fname):
38 def join(self, obj, fname):
39 return obj.sjoin(fname)
39 return obj.sjoin(fname)
40
40
41 class unfilteredpropertycache(propertycache):
41 class unfilteredpropertycache(propertycache):
42 """propertycache that apply to unfiltered repo only"""
42 """propertycache that apply to unfiltered repo only"""
43
43
44 def __get__(self, repo, type=None):
44 def __get__(self, repo, type=None):
45 unfi = repo.unfiltered()
45 unfi = repo.unfiltered()
46 if unfi is repo:
46 if unfi is repo:
47 return super(unfilteredpropertycache, self).__get__(unfi)
47 return super(unfilteredpropertycache, self).__get__(unfi)
48 return getattr(unfi, self.name)
48 return getattr(unfi, self.name)
49
49
50 class filteredpropertycache(propertycache):
50 class filteredpropertycache(propertycache):
51 """propertycache that must take filtering in account"""
51 """propertycache that must take filtering in account"""
52
52
53 def cachevalue(self, obj, value):
53 def cachevalue(self, obj, value):
54 object.__setattr__(obj, self.name, value)
54 object.__setattr__(obj, self.name, value)
55
55
56
56
57 def hasunfilteredcache(repo, name):
57 def hasunfilteredcache(repo, name):
58 """check if a repo has an unfilteredpropertycache value for <name>"""
58 """check if a repo has an unfilteredpropertycache value for <name>"""
59 return name in vars(repo.unfiltered())
59 return name in vars(repo.unfiltered())
60
60
61 def unfilteredmethod(orig):
61 def unfilteredmethod(orig):
62 """decorate method that always need to be run on unfiltered version"""
62 """decorate method that always need to be run on unfiltered version"""
63 def wrapper(repo, *args, **kwargs):
63 def wrapper(repo, *args, **kwargs):
64 return orig(repo.unfiltered(), *args, **kwargs)
64 return orig(repo.unfiltered(), *args, **kwargs)
65 return wrapper
65 return wrapper
66
66
67 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
67 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
68 'unbundle'))
68 'unbundle'))
69 legacycaps = moderncaps.union(set(['changegroupsubset']))
69 legacycaps = moderncaps.union(set(['changegroupsubset']))
70
70
71 class localpeer(peer.peerrepository):
71 class localpeer(peer.peerrepository):
72 '''peer for a local repo; reflects only the most recent API'''
72 '''peer for a local repo; reflects only the most recent API'''
73
73
74 def __init__(self, repo, caps=moderncaps):
74 def __init__(self, repo, caps=moderncaps):
75 peer.peerrepository.__init__(self)
75 peer.peerrepository.__init__(self)
76 self._repo = repo.filtered('served')
76 self._repo = repo.filtered('served')
77 self.ui = repo.ui
77 self.ui = repo.ui
78 self._caps = repo._restrictcapabilities(caps)
78 self._caps = repo._restrictcapabilities(caps)
79 self.requirements = repo.requirements
79 self.requirements = repo.requirements
80 self.supportedformats = repo.supportedformats
80 self.supportedformats = repo.supportedformats
81
81
82 def close(self):
82 def close(self):
83 self._repo.close()
83 self._repo.close()
84
84
85 def _capabilities(self):
85 def _capabilities(self):
86 return self._caps
86 return self._caps
87
87
88 def local(self):
88 def local(self):
89 return self._repo
89 return self._repo
90
90
91 def canpush(self):
91 def canpush(self):
92 return True
92 return True
93
93
94 def url(self):
94 def url(self):
95 return self._repo.url()
95 return self._repo.url()
96
96
97 def lookup(self, key):
97 def lookup(self, key):
98 return self._repo.lookup(key)
98 return self._repo.lookup(key)
99
99
100 def branchmap(self):
100 def branchmap(self):
101 return self._repo.branchmap()
101 return self._repo.branchmap()
102
102
103 def heads(self):
103 def heads(self):
104 return self._repo.heads()
104 return self._repo.heads()
105
105
106 def known(self, nodes):
106 def known(self, nodes):
107 return self._repo.known(nodes)
107 return self._repo.known(nodes)
108
108
109 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
109 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
110 **kwargs):
110 **kwargs):
111 cg = exchange.getbundle(self._repo, source, heads=heads,
111 cg = exchange.getbundle(self._repo, source, heads=heads,
112 common=common, bundlecaps=bundlecaps, **kwargs)
112 common=common, bundlecaps=bundlecaps, **kwargs)
113 if bundlecaps is not None and 'HG20' in bundlecaps:
113 if bundlecaps is not None and 'HG20' in bundlecaps:
114 # When requesting a bundle2, getbundle returns a stream to make the
114 # When requesting a bundle2, getbundle returns a stream to make the
115 # wire level function happier. We need to build a proper object
115 # wire level function happier. We need to build a proper object
116 # from it in local peer.
116 # from it in local peer.
117 cg = bundle2.getunbundler(self.ui, cg)
117 cg = bundle2.getunbundler(self.ui, cg)
118 return cg
118 return cg
119
119
120 # TODO We might want to move the next two calls into legacypeer and add
120 # TODO We might want to move the next two calls into legacypeer and add
121 # unbundle instead.
121 # unbundle instead.
122
122
123 def unbundle(self, cg, heads, url):
123 def unbundle(self, cg, heads, url):
124 """apply a bundle on a repo
124 """apply a bundle on a repo
125
125
126 This function handles the repo locking itself."""
126 This function handles the repo locking itself."""
127 try:
127 try:
128 cg = exchange.readbundle(self.ui, cg, None)
128 cg = exchange.readbundle(self.ui, cg, None)
129 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
129 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
130 if util.safehasattr(ret, 'getchunks'):
130 if util.safehasattr(ret, 'getchunks'):
131 # This is a bundle20 object, turn it into an unbundler.
131 # This is a bundle20 object, turn it into an unbundler.
132 # This little dance should be dropped eventually when the API
132 # This little dance should be dropped eventually when the API
133 # is finally improved.
133 # is finally improved.
134 stream = util.chunkbuffer(ret.getchunks())
134 stream = util.chunkbuffer(ret.getchunks())
135 ret = bundle2.getunbundler(self.ui, stream)
135 ret = bundle2.getunbundler(self.ui, stream)
136 return ret
136 return ret
137 except error.PushRaced, exc:
137 except error.PushRaced, exc:
138 raise error.ResponseError(_('push failed:'), str(exc))
138 raise error.ResponseError(_('push failed:'), str(exc))
139
139
140 def lock(self):
140 def lock(self):
141 return self._repo.lock()
141 return self._repo.lock()
142
142
143 def addchangegroup(self, cg, source, url):
143 def addchangegroup(self, cg, source, url):
144 return changegroup.addchangegroup(self._repo, cg, source, url)
144 return changegroup.addchangegroup(self._repo, cg, source, url)
145
145
146 def pushkey(self, namespace, key, old, new):
146 def pushkey(self, namespace, key, old, new):
147 return self._repo.pushkey(namespace, key, old, new)
147 return self._repo.pushkey(namespace, key, old, new)
148
148
149 def listkeys(self, namespace):
149 def listkeys(self, namespace):
150 return self._repo.listkeys(namespace)
150 return self._repo.listkeys(namespace)
151
151
152 def debugwireargs(self, one, two, three=None, four=None, five=None):
152 def debugwireargs(self, one, two, three=None, four=None, five=None):
153 '''used to test argument passing over the wire'''
153 '''used to test argument passing over the wire'''
154 return "%s %s %s %s %s" % (one, two, three, four, five)
154 return "%s %s %s %s %s" % (one, two, three, four, five)
155
155
156 class locallegacypeer(localpeer):
156 class locallegacypeer(localpeer):
157 '''peer extension which implements legacy methods too; used for tests with
157 '''peer extension which implements legacy methods too; used for tests with
158 restricted capabilities'''
158 restricted capabilities'''
159
159
160 def __init__(self, repo):
160 def __init__(self, repo):
161 localpeer.__init__(self, repo, caps=legacycaps)
161 localpeer.__init__(self, repo, caps=legacycaps)
162
162
163 def branches(self, nodes):
163 def branches(self, nodes):
164 return self._repo.branches(nodes)
164 return self._repo.branches(nodes)
165
165
166 def between(self, pairs):
166 def between(self, pairs):
167 return self._repo.between(pairs)
167 return self._repo.between(pairs)
168
168
169 def changegroup(self, basenodes, source):
169 def changegroup(self, basenodes, source):
170 return changegroup.changegroup(self._repo, basenodes, source)
170 return changegroup.changegroup(self._repo, basenodes, source)
171
171
172 def changegroupsubset(self, bases, heads, source):
172 def changegroupsubset(self, bases, heads, source):
173 return changegroup.changegroupsubset(self._repo, bases, heads, source)
173 return changegroup.changegroupsubset(self._repo, bases, heads, source)
174
174
175 class localrepository(object):
175 class localrepository(object):
176
176
177 supportedformats = set(('revlogv1', 'generaldelta', 'manifestv2'))
177 supportedformats = set(('revlogv1', 'generaldelta', 'manifestv2'))
178 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
178 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
179 'dotencode'))
179 'dotencode'))
180 openerreqs = set(('revlogv1', 'generaldelta', 'manifestv2'))
180 openerreqs = set(('revlogv1', 'generaldelta', 'manifestv2'))
181 requirements = ['revlogv1']
181 requirements = ['revlogv1']
182 filtername = None
182 filtername = None
183
183
184 # a list of (ui, featureset) functions.
184 # a list of (ui, featureset) functions.
185 # only functions defined in module of enabled extensions are invoked
185 # only functions defined in module of enabled extensions are invoked
186 featuresetupfuncs = set()
186 featuresetupfuncs = set()
187
187
188 def _baserequirements(self, create):
188 def _baserequirements(self, create):
189 return self.requirements[:]
189 return self.requirements[:]
190
190
191 def __init__(self, baseui, path=None, create=False):
191 def __init__(self, baseui, path=None, create=False):
192 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
192 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
193 self.wopener = self.wvfs
193 self.wopener = self.wvfs
194 self.root = self.wvfs.base
194 self.root = self.wvfs.base
195 self.path = self.wvfs.join(".hg")
195 self.path = self.wvfs.join(".hg")
196 self.origroot = path
196 self.origroot = path
197 self.auditor = pathutil.pathauditor(self.root, self._checknested)
197 self.auditor = pathutil.pathauditor(self.root, self._checknested)
198 self.vfs = scmutil.vfs(self.path)
198 self.vfs = scmutil.vfs(self.path)
199 self.opener = self.vfs
199 self.opener = self.vfs
200 self.baseui = baseui
200 self.baseui = baseui
201 self.ui = baseui.copy()
201 self.ui = baseui.copy()
202 self.ui.copy = baseui.copy # prevent copying repo configuration
202 self.ui.copy = baseui.copy # prevent copying repo configuration
203 # A list of callback to shape the phase if no data were found.
203 # A list of callback to shape the phase if no data were found.
204 # Callback are in the form: func(repo, roots) --> processed root.
204 # Callback are in the form: func(repo, roots) --> processed root.
205 # This list it to be filled by extension during repo setup
205 # This list it to be filled by extension during repo setup
206 self._phasedefaults = []
206 self._phasedefaults = []
207 try:
207 try:
208 self.ui.readconfig(self.join("hgrc"), self.root)
208 self.ui.readconfig(self.join("hgrc"), self.root)
209 extensions.loadall(self.ui)
209 extensions.loadall(self.ui)
210 except IOError:
210 except IOError:
211 pass
211 pass
212
212
213 if self.featuresetupfuncs:
213 if self.featuresetupfuncs:
214 self.supported = set(self._basesupported) # use private copy
214 self.supported = set(self._basesupported) # use private copy
215 extmods = set(m.__name__ for n, m
215 extmods = set(m.__name__ for n, m
216 in extensions.extensions(self.ui))
216 in extensions.extensions(self.ui))
217 for setupfunc in self.featuresetupfuncs:
217 for setupfunc in self.featuresetupfuncs:
218 if setupfunc.__module__ in extmods:
218 if setupfunc.__module__ in extmods:
219 setupfunc(self.ui, self.supported)
219 setupfunc(self.ui, self.supported)
220 else:
220 else:
221 self.supported = self._basesupported
221 self.supported = self._basesupported
222
222
223 if not self.vfs.isdir():
223 if not self.vfs.isdir():
224 if create:
224 if create:
225 if not self.wvfs.exists():
225 if not self.wvfs.exists():
226 self.wvfs.makedirs()
226 self.wvfs.makedirs()
227 self.vfs.makedir(notindexed=True)
227 self.vfs.makedir(notindexed=True)
228 requirements = self._baserequirements(create)
228 requirements = self._baserequirements(create)
229 if self.ui.configbool('format', 'usestore', True):
229 if self.ui.configbool('format', 'usestore', True):
230 self.vfs.mkdir("store")
230 self.vfs.mkdir("store")
231 requirements.append("store")
231 requirements.append("store")
232 if self.ui.configbool('format', 'usefncache', True):
232 if self.ui.configbool('format', 'usefncache', True):
233 requirements.append("fncache")
233 requirements.append("fncache")
234 if self.ui.configbool('format', 'dotencode', True):
234 if self.ui.configbool('format', 'dotencode', True):
235 requirements.append('dotencode')
235 requirements.append('dotencode')
236 # create an invalid changelog
236 # create an invalid changelog
237 self.vfs.append(
237 self.vfs.append(
238 "00changelog.i",
238 "00changelog.i",
239 '\0\0\0\2' # represents revlogv2
239 '\0\0\0\2' # represents revlogv2
240 ' dummy changelog to prevent using the old repo layout'
240 ' dummy changelog to prevent using the old repo layout'
241 )
241 )
242 if self.ui.configbool('format', 'generaldelta', False):
242 if self.ui.configbool('format', 'generaldelta', False):
243 requirements.append("generaldelta")
243 requirements.append("generaldelta")
244 if self.ui.configbool('experimental', 'manifestv2', False):
244 if self.ui.configbool('experimental', 'manifestv2', False):
245 requirements.append("manifestv2")
245 requirements.append("manifestv2")
246 requirements = set(requirements)
246 requirements = set(requirements)
247 else:
247 else:
248 raise error.RepoError(_("repository %s not found") % path)
248 raise error.RepoError(_("repository %s not found") % path)
249 elif create:
249 elif create:
250 raise error.RepoError(_("repository %s already exists") % path)
250 raise error.RepoError(_("repository %s already exists") % path)
251 else:
251 else:
252 try:
252 try:
253 requirements = scmutil.readrequires(self.vfs, self.supported)
253 requirements = scmutil.readrequires(self.vfs, self.supported)
254 except IOError, inst:
254 except IOError, inst:
255 if inst.errno != errno.ENOENT:
255 if inst.errno != errno.ENOENT:
256 raise
256 raise
257 requirements = set()
257 requirements = set()
258
258
259 self.sharedpath = self.path
259 self.sharedpath = self.path
260 try:
260 try:
261 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
261 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
262 realpath=True)
262 realpath=True)
263 s = vfs.base
263 s = vfs.base
264 if not vfs.exists():
264 if not vfs.exists():
265 raise error.RepoError(
265 raise error.RepoError(
266 _('.hg/sharedpath points to nonexistent directory %s') % s)
266 _('.hg/sharedpath points to nonexistent directory %s') % s)
267 self.sharedpath = s
267 self.sharedpath = s
268 except IOError, inst:
268 except IOError, inst:
269 if inst.errno != errno.ENOENT:
269 if inst.errno != errno.ENOENT:
270 raise
270 raise
271
271
272 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
272 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
273 self.spath = self.store.path
273 self.spath = self.store.path
274 self.svfs = self.store.vfs
274 self.svfs = self.store.vfs
275 self.sopener = self.svfs
275 self.sopener = self.svfs
276 self.sjoin = self.store.join
276 self.sjoin = self.store.join
277 self.vfs.createmode = self.store.createmode
277 self.vfs.createmode = self.store.createmode
278 self._applyrequirements(requirements)
278 self._applyrequirements(requirements)
279 if create:
279 if create:
280 self._writerequirements()
280 self._writerequirements()
281
281
282
282
283 self._branchcaches = {}
283 self._branchcaches = {}
284 self._revbranchcache = None
284 self._revbranchcache = None
285 self.filterpats = {}
285 self.filterpats = {}
286 self._datafilters = {}
286 self._datafilters = {}
287 self._transref = self._lockref = self._wlockref = None
287 self._transref = self._lockref = self._wlockref = None
288
288
289 # A cache for various files under .hg/ that tracks file changes,
289 # A cache for various files under .hg/ that tracks file changes,
290 # (used by the filecache decorator)
290 # (used by the filecache decorator)
291 #
291 #
292 # Maps a property name to its util.filecacheentry
292 # Maps a property name to its util.filecacheentry
293 self._filecache = {}
293 self._filecache = {}
294
294
295 # hold sets of revision to be filtered
295 # hold sets of revision to be filtered
296 # should be cleared when something might have changed the filter value:
296 # should be cleared when something might have changed the filter value:
297 # - new changesets,
297 # - new changesets,
298 # - phase change,
298 # - phase change,
299 # - new obsolescence marker,
299 # - new obsolescence marker,
300 # - working directory parent change,
300 # - working directory parent change,
301 # - bookmark changes
301 # - bookmark changes
302 self.filteredrevcache = {}
302 self.filteredrevcache = {}
303
303
304 # generic mapping between names and nodes
304 # generic mapping between names and nodes
305 self.names = namespaces.namespaces()
305 self.names = namespaces.namespaces()
306
306
307 def close(self):
307 def close(self):
308 self._writecaches()
308 self._writecaches()
309
309
310 def _writecaches(self):
310 def _writecaches(self):
311 if self._revbranchcache:
311 if self._revbranchcache:
312 self._revbranchcache.write()
312 self._revbranchcache.write()
313
313
314 def _restrictcapabilities(self, caps):
314 def _restrictcapabilities(self, caps):
315 if self.ui.configbool('experimental', 'bundle2-advertise', True):
315 if self.ui.configbool('experimental', 'bundle2-advertise', True):
316 caps = set(caps)
316 caps = set(caps)
317 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
317 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
318 caps.add('bundle2=' + urllib.quote(capsblob))
318 caps.add('bundle2=' + urllib.quote(capsblob))
319 return caps
319 return caps
320
320
321 def _applyrequirements(self, requirements):
321 def _applyrequirements(self, requirements):
322 self.requirements = requirements
322 self.requirements = requirements
323 self.svfs.options = dict((r, 1) for r in requirements
323 self.svfs.options = dict((r, 1) for r in requirements
324 if r in self.openerreqs)
324 if r in self.openerreqs)
325 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
325 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
326 if chunkcachesize is not None:
326 if chunkcachesize is not None:
327 self.svfs.options['chunkcachesize'] = chunkcachesize
327 self.svfs.options['chunkcachesize'] = chunkcachesize
328 maxchainlen = self.ui.configint('format', 'maxchainlen')
328 maxchainlen = self.ui.configint('format', 'maxchainlen')
329 if maxchainlen is not None:
329 if maxchainlen is not None:
330 self.svfs.options['maxchainlen'] = maxchainlen
330 self.svfs.options['maxchainlen'] = maxchainlen
331 manifestcachesize = self.ui.configint('format', 'manifestcachesize')
331 manifestcachesize = self.ui.configint('format', 'manifestcachesize')
332 if manifestcachesize is not None:
332 if manifestcachesize is not None:
333 self.svfs.options['manifestcachesize'] = manifestcachesize
333 self.svfs.options['manifestcachesize'] = manifestcachesize
334 usetreemanifest = self.ui.configbool('experimental', 'treemanifest')
334 usetreemanifest = self.ui.configbool('experimental', 'treemanifest')
335 if usetreemanifest is not None:
335 if usetreemanifest is not None:
336 self.svfs.options['usetreemanifest'] = usetreemanifest
336 self.svfs.options['usetreemanifest'] = usetreemanifest
337
337
338 def _writerequirements(self):
338 def _writerequirements(self):
339 reqfile = self.vfs("requires", "w")
339 reqfile = self.vfs("requires", "w")
340 for r in sorted(self.requirements):
340 for r in sorted(self.requirements):
341 reqfile.write("%s\n" % r)
341 reqfile.write("%s\n" % r)
342 reqfile.close()
342 reqfile.close()
343
343
344 def _checknested(self, path):
344 def _checknested(self, path):
345 """Determine if path is a legal nested repository."""
345 """Determine if path is a legal nested repository."""
346 if not path.startswith(self.root):
346 if not path.startswith(self.root):
347 return False
347 return False
348 subpath = path[len(self.root) + 1:]
348 subpath = path[len(self.root) + 1:]
349 normsubpath = util.pconvert(subpath)
349 normsubpath = util.pconvert(subpath)
350
350
351 # XXX: Checking against the current working copy is wrong in
351 # XXX: Checking against the current working copy is wrong in
352 # the sense that it can reject things like
352 # the sense that it can reject things like
353 #
353 #
354 # $ hg cat -r 10 sub/x.txt
354 # $ hg cat -r 10 sub/x.txt
355 #
355 #
356 # if sub/ is no longer a subrepository in the working copy
356 # if sub/ is no longer a subrepository in the working copy
357 # parent revision.
357 # parent revision.
358 #
358 #
359 # However, it can of course also allow things that would have
359 # However, it can of course also allow things that would have
360 # been rejected before, such as the above cat command if sub/
360 # been rejected before, such as the above cat command if sub/
361 # is a subrepository now, but was a normal directory before.
361 # is a subrepository now, but was a normal directory before.
362 # The old path auditor would have rejected by mistake since it
362 # The old path auditor would have rejected by mistake since it
363 # panics when it sees sub/.hg/.
363 # panics when it sees sub/.hg/.
364 #
364 #
365 # All in all, checking against the working copy seems sensible
365 # All in all, checking against the working copy seems sensible
366 # since we want to prevent access to nested repositories on
366 # since we want to prevent access to nested repositories on
367 # the filesystem *now*.
367 # the filesystem *now*.
368 ctx = self[None]
368 ctx = self[None]
369 parts = util.splitpath(subpath)
369 parts = util.splitpath(subpath)
370 while parts:
370 while parts:
371 prefix = '/'.join(parts)
371 prefix = '/'.join(parts)
372 if prefix in ctx.substate:
372 if prefix in ctx.substate:
373 if prefix == normsubpath:
373 if prefix == normsubpath:
374 return True
374 return True
375 else:
375 else:
376 sub = ctx.sub(prefix)
376 sub = ctx.sub(prefix)
377 return sub.checknested(subpath[len(prefix) + 1:])
377 return sub.checknested(subpath[len(prefix) + 1:])
378 else:
378 else:
379 parts.pop()
379 parts.pop()
380 return False
380 return False
381
381
382 def peer(self):
382 def peer(self):
383 return localpeer(self) # not cached to avoid reference cycle
383 return localpeer(self) # not cached to avoid reference cycle
384
384
385 def unfiltered(self):
385 def unfiltered(self):
386 """Return unfiltered version of the repository
386 """Return unfiltered version of the repository
387
387
388 Intended to be overwritten by filtered repo."""
388 Intended to be overwritten by filtered repo."""
389 return self
389 return self
390
390
391 def filtered(self, name):
391 def filtered(self, name):
392 """Return a filtered version of a repository"""
392 """Return a filtered version of a repository"""
393 # build a new class with the mixin and the current class
393 # build a new class with the mixin and the current class
394 # (possibly subclass of the repo)
394 # (possibly subclass of the repo)
395 class proxycls(repoview.repoview, self.unfiltered().__class__):
395 class proxycls(repoview.repoview, self.unfiltered().__class__):
396 pass
396 pass
397 return proxycls(self, name)
397 return proxycls(self, name)
398
398
399 @repofilecache('bookmarks')
399 @repofilecache('bookmarks')
400 def _bookmarks(self):
400 def _bookmarks(self):
401 return bookmarks.bmstore(self)
401 return bookmarks.bmstore(self)
402
402
403 @repofilecache('bookmarks.current')
403 @repofilecache('bookmarks.current')
404 def _bookmarkcurrent(self):
404 def _bookmarkcurrent(self):
405 return bookmarks.readcurrent(self)
405 return bookmarks.readcurrent(self)
406
406
407 def bookmarkheads(self, bookmark):
407 def bookmarkheads(self, bookmark):
408 name = bookmark.split('@', 1)[0]
408 name = bookmark.split('@', 1)[0]
409 heads = []
409 heads = []
410 for mark, n in self._bookmarks.iteritems():
410 for mark, n in self._bookmarks.iteritems():
411 if mark.split('@', 1)[0] == name:
411 if mark.split('@', 1)[0] == name:
412 heads.append(n)
412 heads.append(n)
413 return heads
413 return heads
414
414
415 @storecache('phaseroots')
415 @storecache('phaseroots')
416 def _phasecache(self):
416 def _phasecache(self):
417 return phases.phasecache(self, self._phasedefaults)
417 return phases.phasecache(self, self._phasedefaults)
418
418
419 @storecache('obsstore')
419 @storecache('obsstore')
420 def obsstore(self):
420 def obsstore(self):
421 # read default format for new obsstore.
421 # read default format for new obsstore.
422 defaultformat = self.ui.configint('format', 'obsstore-version', None)
422 defaultformat = self.ui.configint('format', 'obsstore-version', None)
423 # rely on obsstore class default when possible.
423 # rely on obsstore class default when possible.
424 kwargs = {}
424 kwargs = {}
425 if defaultformat is not None:
425 if defaultformat is not None:
426 kwargs['defaultformat'] = defaultformat
426 kwargs['defaultformat'] = defaultformat
427 readonly = not obsolete.isenabled(self, obsolete.createmarkersopt)
427 readonly = not obsolete.isenabled(self, obsolete.createmarkersopt)
428 store = obsolete.obsstore(self.svfs, readonly=readonly,
428 store = obsolete.obsstore(self.svfs, readonly=readonly,
429 **kwargs)
429 **kwargs)
430 if store and readonly:
430 if store and readonly:
431 self.ui.warn(
431 self.ui.warn(
432 _('obsolete feature not enabled but %i markers found!\n')
432 _('obsolete feature not enabled but %i markers found!\n')
433 % len(list(store)))
433 % len(list(store)))
434 return store
434 return store
435
435
436 @storecache('00changelog.i')
436 @storecache('00changelog.i')
437 def changelog(self):
437 def changelog(self):
438 c = changelog.changelog(self.svfs)
438 c = changelog.changelog(self.svfs)
439 if 'HG_PENDING' in os.environ:
439 if 'HG_PENDING' in os.environ:
440 p = os.environ['HG_PENDING']
440 p = os.environ['HG_PENDING']
441 if p.startswith(self.root):
441 if p.startswith(self.root):
442 c.readpending('00changelog.i.a')
442 c.readpending('00changelog.i.a')
443 return c
443 return c
444
444
445 @storecache('00manifest.i')
445 @storecache('00manifest.i')
446 def manifest(self):
446 def manifest(self):
447 return manifest.manifest(self.svfs)
447 return manifest.manifest(self.svfs)
448
448
449 @repofilecache('dirstate')
449 @repofilecache('dirstate')
450 def dirstate(self):
450 def dirstate(self):
451 warned = [0]
451 warned = [0]
452 def validate(node):
452 def validate(node):
453 try:
453 try:
454 self.changelog.rev(node)
454 self.changelog.rev(node)
455 return node
455 return node
456 except error.LookupError:
456 except error.LookupError:
457 if not warned[0]:
457 if not warned[0]:
458 warned[0] = True
458 warned[0] = True
459 self.ui.warn(_("warning: ignoring unknown"
459 self.ui.warn(_("warning: ignoring unknown"
460 " working parent %s!\n") % short(node))
460 " working parent %s!\n") % short(node))
461 return nullid
461 return nullid
462
462
463 return dirstate.dirstate(self.vfs, self.ui, self.root, validate)
463 return dirstate.dirstate(self.vfs, self.ui, self.root, validate)
464
464
465 def __getitem__(self, changeid):
465 def __getitem__(self, changeid):
466 if changeid is None:
466 if changeid is None:
467 return context.workingctx(self)
467 return context.workingctx(self)
468 if isinstance(changeid, slice):
468 if isinstance(changeid, slice):
469 return [context.changectx(self, i)
469 return [context.changectx(self, i)
470 for i in xrange(*changeid.indices(len(self)))
470 for i in xrange(*changeid.indices(len(self)))
471 if i not in self.changelog.filteredrevs]
471 if i not in self.changelog.filteredrevs]
472 return context.changectx(self, changeid)
472 return context.changectx(self, changeid)
473
473
474 def __contains__(self, changeid):
474 def __contains__(self, changeid):
475 try:
475 try:
476 self[changeid]
476 self[changeid]
477 return True
477 return True
478 except error.RepoLookupError:
478 except error.RepoLookupError:
479 return False
479 return False
480
480
481 def __nonzero__(self):
481 def __nonzero__(self):
482 return True
482 return True
483
483
484 def __len__(self):
484 def __len__(self):
485 return len(self.changelog)
485 return len(self.changelog)
486
486
487 def __iter__(self):
487 def __iter__(self):
488 return iter(self.changelog)
488 return iter(self.changelog)
489
489
490 def revs(self, expr, *args):
490 def revs(self, expr, *args):
491 '''Return a list of revisions matching the given revset'''
491 '''Return a list of revisions matching the given revset'''
492 expr = revset.formatspec(expr, *args)
492 expr = revset.formatspec(expr, *args)
493 m = revset.match(None, expr)
493 m = revset.match(None, expr)
494 return m(self)
494 return m(self)
495
495
496 def set(self, expr, *args):
496 def set(self, expr, *args):
497 '''
497 '''
498 Yield a context for each matching revision, after doing arg
498 Yield a context for each matching revision, after doing arg
499 replacement via revset.formatspec
499 replacement via revset.formatspec
500 '''
500 '''
501 for r in self.revs(expr, *args):
501 for r in self.revs(expr, *args):
502 yield self[r]
502 yield self[r]
503
503
504 def url(self):
504 def url(self):
505 return 'file:' + self.root
505 return 'file:' + self.root
506
506
507 def hook(self, name, throw=False, **args):
507 def hook(self, name, throw=False, **args):
508 """Call a hook, passing this repo instance.
508 """Call a hook, passing this repo instance.
509
509
510 This a convenience method to aid invoking hooks. Extensions likely
510 This a convenience method to aid invoking hooks. Extensions likely
511 won't call this unless they have registered a custom hook or are
511 won't call this unless they have registered a custom hook or are
512 replacing code that is expected to call a hook.
512 replacing code that is expected to call a hook.
513 """
513 """
514 return hook.hook(self.ui, self, name, throw, **args)
514 return hook.hook(self.ui, self, name, throw, **args)
515
515
516 @unfilteredmethod
516 @unfilteredmethod
517 def _tag(self, names, node, message, local, user, date, extra={},
517 def _tag(self, names, node, message, local, user, date, extra={},
518 editor=False):
518 editor=False):
519 if isinstance(names, str):
519 if isinstance(names, str):
520 names = (names,)
520 names = (names,)
521
521
522 branches = self.branchmap()
522 branches = self.branchmap()
523 for name in names:
523 for name in names:
524 self.hook('pretag', throw=True, node=hex(node), tag=name,
524 self.hook('pretag', throw=True, node=hex(node), tag=name,
525 local=local)
525 local=local)
526 if name in branches:
526 if name in branches:
527 self.ui.warn(_("warning: tag %s conflicts with existing"
527 self.ui.warn(_("warning: tag %s conflicts with existing"
528 " branch name\n") % name)
528 " branch name\n") % name)
529
529
530 def writetags(fp, names, munge, prevtags):
530 def writetags(fp, names, munge, prevtags):
531 fp.seek(0, 2)
531 fp.seek(0, 2)
532 if prevtags and prevtags[-1] != '\n':
532 if prevtags and prevtags[-1] != '\n':
533 fp.write('\n')
533 fp.write('\n')
534 for name in names:
534 for name in names:
535 if munge:
535 if munge:
536 m = munge(name)
536 m = munge(name)
537 else:
537 else:
538 m = name
538 m = name
539
539
540 if (self._tagscache.tagtypes and
540 if (self._tagscache.tagtypes and
541 name in self._tagscache.tagtypes):
541 name in self._tagscache.tagtypes):
542 old = self.tags().get(name, nullid)
542 old = self.tags().get(name, nullid)
543 fp.write('%s %s\n' % (hex(old), m))
543 fp.write('%s %s\n' % (hex(old), m))
544 fp.write('%s %s\n' % (hex(node), m))
544 fp.write('%s %s\n' % (hex(node), m))
545 fp.close()
545 fp.close()
546
546
547 prevtags = ''
547 prevtags = ''
548 if local:
548 if local:
549 try:
549 try:
550 fp = self.vfs('localtags', 'r+')
550 fp = self.vfs('localtags', 'r+')
551 except IOError:
551 except IOError:
552 fp = self.vfs('localtags', 'a')
552 fp = self.vfs('localtags', 'a')
553 else:
553 else:
554 prevtags = fp.read()
554 prevtags = fp.read()
555
555
556 # local tags are stored in the current charset
556 # local tags are stored in the current charset
557 writetags(fp, names, None, prevtags)
557 writetags(fp, names, None, prevtags)
558 for name in names:
558 for name in names:
559 self.hook('tag', node=hex(node), tag=name, local=local)
559 self.hook('tag', node=hex(node), tag=name, local=local)
560 return
560 return
561
561
562 try:
562 try:
563 fp = self.wfile('.hgtags', 'rb+')
563 fp = self.wfile('.hgtags', 'rb+')
564 except IOError, e:
564 except IOError, e:
565 if e.errno != errno.ENOENT:
565 if e.errno != errno.ENOENT:
566 raise
566 raise
567 fp = self.wfile('.hgtags', 'ab')
567 fp = self.wfile('.hgtags', 'ab')
568 else:
568 else:
569 prevtags = fp.read()
569 prevtags = fp.read()
570
570
571 # committed tags are stored in UTF-8
571 # committed tags are stored in UTF-8
572 writetags(fp, names, encoding.fromlocal, prevtags)
572 writetags(fp, names, encoding.fromlocal, prevtags)
573
573
574 fp.close()
574 fp.close()
575
575
576 self.invalidatecaches()
576 self.invalidatecaches()
577
577
578 if '.hgtags' not in self.dirstate:
578 if '.hgtags' not in self.dirstate:
579 self[None].add(['.hgtags'])
579 self[None].add(['.hgtags'])
580
580
581 m = matchmod.exact(self.root, '', ['.hgtags'])
581 m = matchmod.exact(self.root, '', ['.hgtags'])
582 tagnode = self.commit(message, user, date, extra=extra, match=m,
582 tagnode = self.commit(message, user, date, extra=extra, match=m,
583 editor=editor)
583 editor=editor)
584
584
585 for name in names:
585 for name in names:
586 self.hook('tag', node=hex(node), tag=name, local=local)
586 self.hook('tag', node=hex(node), tag=name, local=local)
587
587
588 return tagnode
588 return tagnode
589
589
590 def tag(self, names, node, message, local, user, date, editor=False):
590 def tag(self, names, node, message, local, user, date, editor=False):
591 '''tag a revision with one or more symbolic names.
591 '''tag a revision with one or more symbolic names.
592
592
593 names is a list of strings or, when adding a single tag, names may be a
593 names is a list of strings or, when adding a single tag, names may be a
594 string.
594 string.
595
595
596 if local is True, the tags are stored in a per-repository file.
596 if local is True, the tags are stored in a per-repository file.
597 otherwise, they are stored in the .hgtags file, and a new
597 otherwise, they are stored in the .hgtags file, and a new
598 changeset is committed with the change.
598 changeset is committed with the change.
599
599
600 keyword arguments:
600 keyword arguments:
601
601
602 local: whether to store tags in non-version-controlled file
602 local: whether to store tags in non-version-controlled file
603 (default False)
603 (default False)
604
604
605 message: commit message to use if committing
605 message: commit message to use if committing
606
606
607 user: name of user to use if committing
607 user: name of user to use if committing
608
608
609 date: date tuple to use if committing'''
609 date: date tuple to use if committing'''
610
610
611 if not local:
611 if not local:
612 m = matchmod.exact(self.root, '', ['.hgtags'])
612 m = matchmod.exact(self.root, '', ['.hgtags'])
613 if util.any(self.status(match=m, unknown=True, ignored=True)):
613 if util.any(self.status(match=m, unknown=True, ignored=True)):
614 raise util.Abort(_('working copy of .hgtags is changed'),
614 raise util.Abort(_('working copy of .hgtags is changed'),
615 hint=_('please commit .hgtags manually'))
615 hint=_('please commit .hgtags manually'))
616
616
617 self.tags() # instantiate the cache
617 self.tags() # instantiate the cache
618 self._tag(names, node, message, local, user, date, editor=editor)
618 self._tag(names, node, message, local, user, date, editor=editor)
619
619
620 @filteredpropertycache
620 @filteredpropertycache
621 def _tagscache(self):
621 def _tagscache(self):
622 '''Returns a tagscache object that contains various tags related
622 '''Returns a tagscache object that contains various tags related
623 caches.'''
623 caches.'''
624
624
625 # This simplifies its cache management by having one decorated
625 # This simplifies its cache management by having one decorated
626 # function (this one) and the rest simply fetch things from it.
626 # function (this one) and the rest simply fetch things from it.
627 class tagscache(object):
627 class tagscache(object):
628 def __init__(self):
628 def __init__(self):
629 # These two define the set of tags for this repository. tags
629 # These two define the set of tags for this repository. tags
630 # maps tag name to node; tagtypes maps tag name to 'global' or
630 # maps tag name to node; tagtypes maps tag name to 'global' or
631 # 'local'. (Global tags are defined by .hgtags across all
631 # 'local'. (Global tags are defined by .hgtags across all
632 # heads, and local tags are defined in .hg/localtags.)
632 # heads, and local tags are defined in .hg/localtags.)
633 # They constitute the in-memory cache of tags.
633 # They constitute the in-memory cache of tags.
634 self.tags = self.tagtypes = None
634 self.tags = self.tagtypes = None
635
635
636 self.nodetagscache = self.tagslist = None
636 self.nodetagscache = self.tagslist = None
637
637
638 cache = tagscache()
638 cache = tagscache()
639 cache.tags, cache.tagtypes = self._findtags()
639 cache.tags, cache.tagtypes = self._findtags()
640
640
641 return cache
641 return cache
642
642
643 def tags(self):
643 def tags(self):
644 '''return a mapping of tag to node'''
644 '''return a mapping of tag to node'''
645 t = {}
645 t = {}
646 if self.changelog.filteredrevs:
646 if self.changelog.filteredrevs:
647 tags, tt = self._findtags()
647 tags, tt = self._findtags()
648 else:
648 else:
649 tags = self._tagscache.tags
649 tags = self._tagscache.tags
650 for k, v in tags.iteritems():
650 for k, v in tags.iteritems():
651 try:
651 try:
652 # ignore tags to unknown nodes
652 # ignore tags to unknown nodes
653 self.changelog.rev(v)
653 self.changelog.rev(v)
654 t[k] = v
654 t[k] = v
655 except (error.LookupError, ValueError):
655 except (error.LookupError, ValueError):
656 pass
656 pass
657 return t
657 return t
658
658
659 def _findtags(self):
659 def _findtags(self):
660 '''Do the hard work of finding tags. Return a pair of dicts
660 '''Do the hard work of finding tags. Return a pair of dicts
661 (tags, tagtypes) where tags maps tag name to node, and tagtypes
661 (tags, tagtypes) where tags maps tag name to node, and tagtypes
662 maps tag name to a string like \'global\' or \'local\'.
662 maps tag name to a string like \'global\' or \'local\'.
663 Subclasses or extensions are free to add their own tags, but
663 Subclasses or extensions are free to add their own tags, but
664 should be aware that the returned dicts will be retained for the
664 should be aware that the returned dicts will be retained for the
665 duration of the localrepo object.'''
665 duration of the localrepo object.'''
666
666
667 # XXX what tagtype should subclasses/extensions use? Currently
667 # XXX what tagtype should subclasses/extensions use? Currently
668 # mq and bookmarks add tags, but do not set the tagtype at all.
668 # mq and bookmarks add tags, but do not set the tagtype at all.
669 # Should each extension invent its own tag type? Should there
669 # Should each extension invent its own tag type? Should there
670 # be one tagtype for all such "virtual" tags? Or is the status
670 # be one tagtype for all such "virtual" tags? Or is the status
671 # quo fine?
671 # quo fine?
672
672
673 alltags = {} # map tag name to (node, hist)
673 alltags = {} # map tag name to (node, hist)
674 tagtypes = {}
674 tagtypes = {}
675
675
676 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
676 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
677 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
677 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
678
678
679 # Build the return dicts. Have to re-encode tag names because
679 # Build the return dicts. Have to re-encode tag names because
680 # the tags module always uses UTF-8 (in order not to lose info
680 # the tags module always uses UTF-8 (in order not to lose info
681 # writing to the cache), but the rest of Mercurial wants them in
681 # writing to the cache), but the rest of Mercurial wants them in
682 # local encoding.
682 # local encoding.
683 tags = {}
683 tags = {}
684 for (name, (node, hist)) in alltags.iteritems():
684 for (name, (node, hist)) in alltags.iteritems():
685 if node != nullid:
685 if node != nullid:
686 tags[encoding.tolocal(name)] = node
686 tags[encoding.tolocal(name)] = node
687 tags['tip'] = self.changelog.tip()
687 tags['tip'] = self.changelog.tip()
688 tagtypes = dict([(encoding.tolocal(name), value)
688 tagtypes = dict([(encoding.tolocal(name), value)
689 for (name, value) in tagtypes.iteritems()])
689 for (name, value) in tagtypes.iteritems()])
690 return (tags, tagtypes)
690 return (tags, tagtypes)
691
691
692 def tagtype(self, tagname):
692 def tagtype(self, tagname):
693 '''
693 '''
694 return the type of the given tag. result can be:
694 return the type of the given tag. result can be:
695
695
696 'local' : a local tag
696 'local' : a local tag
697 'global' : a global tag
697 'global' : a global tag
698 None : tag does not exist
698 None : tag does not exist
699 '''
699 '''
700
700
701 return self._tagscache.tagtypes.get(tagname)
701 return self._tagscache.tagtypes.get(tagname)
702
702
703 def tagslist(self):
703 def tagslist(self):
704 '''return a list of tags ordered by revision'''
704 '''return a list of tags ordered by revision'''
705 if not self._tagscache.tagslist:
705 if not self._tagscache.tagslist:
706 l = []
706 l = []
707 for t, n in self.tags().iteritems():
707 for t, n in self.tags().iteritems():
708 l.append((self.changelog.rev(n), t, n))
708 l.append((self.changelog.rev(n), t, n))
709 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
709 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
710
710
711 return self._tagscache.tagslist
711 return self._tagscache.tagslist
712
712
713 def nodetags(self, node):
713 def nodetags(self, node):
714 '''return the tags associated with a node'''
714 '''return the tags associated with a node'''
715 if not self._tagscache.nodetagscache:
715 if not self._tagscache.nodetagscache:
716 nodetagscache = {}
716 nodetagscache = {}
717 for t, n in self._tagscache.tags.iteritems():
717 for t, n in self._tagscache.tags.iteritems():
718 nodetagscache.setdefault(n, []).append(t)
718 nodetagscache.setdefault(n, []).append(t)
719 for tags in nodetagscache.itervalues():
719 for tags in nodetagscache.itervalues():
720 tags.sort()
720 tags.sort()
721 self._tagscache.nodetagscache = nodetagscache
721 self._tagscache.nodetagscache = nodetagscache
722 return self._tagscache.nodetagscache.get(node, [])
722 return self._tagscache.nodetagscache.get(node, [])
723
723
724 def nodebookmarks(self, node):
724 def nodebookmarks(self, node):
725 marks = []
725 marks = []
726 for bookmark, n in self._bookmarks.iteritems():
726 for bookmark, n in self._bookmarks.iteritems():
727 if n == node:
727 if n == node:
728 marks.append(bookmark)
728 marks.append(bookmark)
729 return sorted(marks)
729 return sorted(marks)
730
730
731 def branchmap(self):
731 def branchmap(self):
732 '''returns a dictionary {branch: [branchheads]} with branchheads
732 '''returns a dictionary {branch: [branchheads]} with branchheads
733 ordered by increasing revision number'''
733 ordered by increasing revision number'''
734 branchmap.updatecache(self)
734 branchmap.updatecache(self)
735 return self._branchcaches[self.filtername]
735 return self._branchcaches[self.filtername]
736
736
737 @unfilteredmethod
737 @unfilteredmethod
738 def revbranchcache(self):
738 def revbranchcache(self):
739 if not self._revbranchcache:
739 if not self._revbranchcache:
740 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
740 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
741 return self._revbranchcache
741 return self._revbranchcache
742
742
743 def branchtip(self, branch, ignoremissing=False):
743 def branchtip(self, branch, ignoremissing=False):
744 '''return the tip node for a given branch
744 '''return the tip node for a given branch
745
745
746 If ignoremissing is True, then this method will not raise an error.
746 If ignoremissing is True, then this method will not raise an error.
747 This is helpful for callers that only expect None for a missing branch
747 This is helpful for callers that only expect None for a missing branch
748 (e.g. namespace).
748 (e.g. namespace).
749
749
750 '''
750 '''
751 try:
751 try:
752 return self.branchmap().branchtip(branch)
752 return self.branchmap().branchtip(branch)
753 except KeyError:
753 except KeyError:
754 if not ignoremissing:
754 if not ignoremissing:
755 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
755 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
756 else:
756 else:
757 pass
757 pass
758
758
759 def lookup(self, key):
759 def lookup(self, key):
760 return self[key].node()
760 return self[key].node()
761
761
762 def lookupbranch(self, key, remote=None):
762 def lookupbranch(self, key, remote=None):
763 repo = remote or self
763 repo = remote or self
764 if key in repo.branchmap():
764 if key in repo.branchmap():
765 return key
765 return key
766
766
767 repo = (remote and remote.local()) and remote or self
767 repo = (remote and remote.local()) and remote or self
768 return repo[key].branch()
768 return repo[key].branch()
769
769
770 def known(self, nodes):
770 def known(self, nodes):
771 nm = self.changelog.nodemap
771 nm = self.changelog.nodemap
772 pc = self._phasecache
772 pc = self._phasecache
773 result = []
773 result = []
774 for n in nodes:
774 for n in nodes:
775 r = nm.get(n)
775 r = nm.get(n)
776 resp = not (r is None or pc.phase(self, r) >= phases.secret)
776 resp = not (r is None or pc.phase(self, r) >= phases.secret)
777 result.append(resp)
777 result.append(resp)
778 return result
778 return result
779
779
780 def local(self):
780 def local(self):
781 return self
781 return self
782
782
783 def cancopy(self):
783 def cancopy(self):
784 # so statichttprepo's override of local() works
784 # so statichttprepo's override of local() works
785 if not self.local():
785 if not self.local():
786 return False
786 return False
787 if not self.ui.configbool('phases', 'publish', True):
787 if not self.ui.configbool('phases', 'publish', True):
788 return True
788 return True
789 # if publishing we can't copy if there is filtered content
789 # if publishing we can't copy if there is filtered content
790 return not self.filtered('visible').changelog.filteredrevs
790 return not self.filtered('visible').changelog.filteredrevs
791
791
792 def shared(self):
792 def shared(self):
793 '''the type of shared repository (None if not shared)'''
793 '''the type of shared repository (None if not shared)'''
794 if self.sharedpath != self.path:
794 if self.sharedpath != self.path:
795 return 'store'
795 return 'store'
796 return None
796 return None
797
797
798 def join(self, f, *insidef):
798 def join(self, f, *insidef):
799 return self.vfs.join(os.path.join(f, *insidef))
799 return self.vfs.join(os.path.join(f, *insidef))
800
800
801 def wjoin(self, f, *insidef):
801 def wjoin(self, f, *insidef):
802 return self.vfs.reljoin(self.root, f, *insidef)
802 return self.vfs.reljoin(self.root, f, *insidef)
803
803
804 def file(self, f):
804 def file(self, f):
805 if f[0] == '/':
805 if f[0] == '/':
806 f = f[1:]
806 f = f[1:]
807 return filelog.filelog(self.svfs, f)
807 return filelog.filelog(self.svfs, f)
808
808
809 def changectx(self, changeid):
809 def changectx(self, changeid):
810 return self[changeid]
810 return self[changeid]
811
811
812 def parents(self, changeid=None):
812 def parents(self, changeid=None):
813 '''get list of changectxs for parents of changeid'''
813 '''get list of changectxs for parents of changeid'''
814 return self[changeid].parents()
814 return self[changeid].parents()
815
815
816 def setparents(self, p1, p2=nullid):
816 def setparents(self, p1, p2=nullid):
817 self.dirstate.beginparentchange()
817 self.dirstate.beginparentchange()
818 copies = self.dirstate.setparents(p1, p2)
818 copies = self.dirstate.setparents(p1, p2)
819 pctx = self[p1]
819 pctx = self[p1]
820 if copies:
820 if copies:
821 # Adjust copy records, the dirstate cannot do it, it
821 # Adjust copy records, the dirstate cannot do it, it
822 # requires access to parents manifests. Preserve them
822 # requires access to parents manifests. Preserve them
823 # only for entries added to first parent.
823 # only for entries added to first parent.
824 for f in copies:
824 for f in copies:
825 if f not in pctx and copies[f] in pctx:
825 if f not in pctx and copies[f] in pctx:
826 self.dirstate.copy(copies[f], f)
826 self.dirstate.copy(copies[f], f)
827 if p2 == nullid:
827 if p2 == nullid:
828 for f, s in sorted(self.dirstate.copies().items()):
828 for f, s in sorted(self.dirstate.copies().items()):
829 if f not in pctx and s not in pctx:
829 if f not in pctx and s not in pctx:
830 self.dirstate.copy(None, f)
830 self.dirstate.copy(None, f)
831 self.dirstate.endparentchange()
831 self.dirstate.endparentchange()
832
832
833 def filectx(self, path, changeid=None, fileid=None):
833 def filectx(self, path, changeid=None, fileid=None):
834 """changeid can be a changeset revision, node, or tag.
834 """changeid can be a changeset revision, node, or tag.
835 fileid can be a file revision or node."""
835 fileid can be a file revision or node."""
836 return context.filectx(self, path, changeid, fileid)
836 return context.filectx(self, path, changeid, fileid)
837
837
838 def getcwd(self):
838 def getcwd(self):
839 return self.dirstate.getcwd()
839 return self.dirstate.getcwd()
840
840
841 def pathto(self, f, cwd=None):
841 def pathto(self, f, cwd=None):
842 return self.dirstate.pathto(f, cwd)
842 return self.dirstate.pathto(f, cwd)
843
843
844 def wfile(self, f, mode='r'):
844 def wfile(self, f, mode='r'):
845 return self.wvfs(f, mode)
845 return self.wvfs(f, mode)
846
846
847 def _link(self, f):
847 def _link(self, f):
848 return self.wvfs.islink(f)
848 return self.wvfs.islink(f)
849
849
850 def _loadfilter(self, filter):
850 def _loadfilter(self, filter):
851 if filter not in self.filterpats:
851 if filter not in self.filterpats:
852 l = []
852 l = []
853 for pat, cmd in self.ui.configitems(filter):
853 for pat, cmd in self.ui.configitems(filter):
854 if cmd == '!':
854 if cmd == '!':
855 continue
855 continue
856 mf = matchmod.match(self.root, '', [pat])
856 mf = matchmod.match(self.root, '', [pat])
857 fn = None
857 fn = None
858 params = cmd
858 params = cmd
859 for name, filterfn in self._datafilters.iteritems():
859 for name, filterfn in self._datafilters.iteritems():
860 if cmd.startswith(name):
860 if cmd.startswith(name):
861 fn = filterfn
861 fn = filterfn
862 params = cmd[len(name):].lstrip()
862 params = cmd[len(name):].lstrip()
863 break
863 break
864 if not fn:
864 if not fn:
865 fn = lambda s, c, **kwargs: util.filter(s, c)
865 fn = lambda s, c, **kwargs: util.filter(s, c)
866 # Wrap old filters not supporting keyword arguments
866 # Wrap old filters not supporting keyword arguments
867 if not inspect.getargspec(fn)[2]:
867 if not inspect.getargspec(fn)[2]:
868 oldfn = fn
868 oldfn = fn
869 fn = lambda s, c, **kwargs: oldfn(s, c)
869 fn = lambda s, c, **kwargs: oldfn(s, c)
870 l.append((mf, fn, params))
870 l.append((mf, fn, params))
871 self.filterpats[filter] = l
871 self.filterpats[filter] = l
872 return self.filterpats[filter]
872 return self.filterpats[filter]
873
873
874 def _filter(self, filterpats, filename, data):
874 def _filter(self, filterpats, filename, data):
875 for mf, fn, cmd in filterpats:
875 for mf, fn, cmd in filterpats:
876 if mf(filename):
876 if mf(filename):
877 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
877 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
878 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
878 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
879 break
879 break
880
880
881 return data
881 return data
882
882
883 @unfilteredpropertycache
883 @unfilteredpropertycache
884 def _encodefilterpats(self):
884 def _encodefilterpats(self):
885 return self._loadfilter('encode')
885 return self._loadfilter('encode')
886
886
887 @unfilteredpropertycache
887 @unfilteredpropertycache
888 def _decodefilterpats(self):
888 def _decodefilterpats(self):
889 return self._loadfilter('decode')
889 return self._loadfilter('decode')
890
890
891 def adddatafilter(self, name, filter):
891 def adddatafilter(self, name, filter):
892 self._datafilters[name] = filter
892 self._datafilters[name] = filter
893
893
894 def wread(self, filename):
894 def wread(self, filename):
895 if self._link(filename):
895 if self._link(filename):
896 data = self.wvfs.readlink(filename)
896 data = self.wvfs.readlink(filename)
897 else:
897 else:
898 data = self.wvfs.read(filename)
898 data = self.wvfs.read(filename)
899 return self._filter(self._encodefilterpats, filename, data)
899 return self._filter(self._encodefilterpats, filename, data)
900
900
901 def wwrite(self, filename, data, flags):
901 def wwrite(self, filename, data, flags):
902 data = self._filter(self._decodefilterpats, filename, data)
902 data = self._filter(self._decodefilterpats, filename, data)
903 if 'l' in flags:
903 if 'l' in flags:
904 self.wvfs.symlink(data, filename)
904 self.wvfs.symlink(data, filename)
905 else:
905 else:
906 self.wvfs.write(filename, data)
906 self.wvfs.write(filename, data)
907 if 'x' in flags:
907 if 'x' in flags:
908 self.wvfs.setflags(filename, False, True)
908 self.wvfs.setflags(filename, False, True)
909
909
910 def wwritedata(self, filename, data):
910 def wwritedata(self, filename, data):
911 return self._filter(self._decodefilterpats, filename, data)
911 return self._filter(self._decodefilterpats, filename, data)
912
912
913 def currenttransaction(self):
913 def currenttransaction(self):
914 """return the current transaction or None if non exists"""
914 """return the current transaction or None if non exists"""
915 if self._transref:
915 if self._transref:
916 tr = self._transref()
916 tr = self._transref()
917 else:
917 else:
918 tr = None
918 tr = None
919
919
920 if tr and tr.running():
920 if tr and tr.running():
921 return tr
921 return tr
922 return None
922 return None
923
923
924 def transaction(self, desc, report=None):
924 def transaction(self, desc, report=None):
925 if (self.ui.configbool('devel', 'all')
925 if (self.ui.configbool('devel', 'all')
926 or self.ui.configbool('devel', 'check-locks')):
926 or self.ui.configbool('devel', 'check-locks')):
927 l = self._lockref and self._lockref()
927 l = self._lockref and self._lockref()
928 if l is None or not l.held:
928 if l is None or not l.held:
929 scmutil.develwarn(self.ui, 'transaction with no lock')
929 scmutil.develwarn(self.ui, 'transaction with no lock')
930 tr = self.currenttransaction()
930 tr = self.currenttransaction()
931 if tr is not None:
931 if tr is not None:
932 return tr.nest()
932 return tr.nest()
933
933
934 # abort here if the journal already exists
934 # abort here if the journal already exists
935 if self.svfs.exists("journal"):
935 if self.svfs.exists("journal"):
936 raise error.RepoError(
936 raise error.RepoError(
937 _("abandoned transaction found"),
937 _("abandoned transaction found"),
938 hint=_("run 'hg recover' to clean up transaction"))
938 hint=_("run 'hg recover' to clean up transaction"))
939
939
940 self.hook('pretxnopen', throw=True, txnname=desc)
940 self.hook('pretxnopen', throw=True, txnname=desc)
941
941
942 self._writejournal(desc)
942 self._writejournal(desc)
943 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
943 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
944 if report:
944 if report:
945 rp = report
945 rp = report
946 else:
946 else:
947 rp = self.ui.warn
947 rp = self.ui.warn
948 vfsmap = {'plain': self.vfs} # root of .hg/
948 vfsmap = {'plain': self.vfs} # root of .hg/
949 # we must avoid cyclic reference between repo and transaction.
949 # we must avoid cyclic reference between repo and transaction.
950 reporef = weakref.ref(self)
950 reporef = weakref.ref(self)
951 def validate(tr):
951 def validate(tr):
952 """will run pre-closing hooks"""
952 """will run pre-closing hooks"""
953 pending = lambda: tr.writepending() and self.root or ""
953 pending = lambda: tr.writepending() and self.root or ""
954 reporef().hook('pretxnclose', throw=True, pending=pending,
954 reporef().hook('pretxnclose', throw=True, pending=pending,
955 xnname=desc, **tr.hookargs)
955 xnname=desc, **tr.hookargs)
956
956
957 tr = transaction.transaction(rp, self.sopener, vfsmap,
957 tr = transaction.transaction(rp, self.sopener, vfsmap,
958 "journal",
958 "journal",
959 "undo",
959 "undo",
960 aftertrans(renames),
960 aftertrans(renames),
961 self.store.createmode,
961 self.store.createmode,
962 validator=validate)
962 validator=validate)
963
963
964 trid = 'TXN:' + util.sha1("%s#%f" % (id(tr), time.time())).hexdigest()
964 trid = 'TXN:' + util.sha1("%s#%f" % (id(tr), time.time())).hexdigest()
965 tr.hookargs['TXNID'] = trid
965 tr.hookargs['TXNID'] = trid
966 # note: writing the fncache only during finalize mean that the file is
966 # note: writing the fncache only during finalize mean that the file is
967 # outdated when running hooks. As fncache is used for streaming clone,
967 # outdated when running hooks. As fncache is used for streaming clone,
968 # this is not expected to break anything that happen during the hooks.
968 # this is not expected to break anything that happen during the hooks.
969 tr.addfinalize('flush-fncache', self.store.write)
969 tr.addfinalize('flush-fncache', self.store.write)
970 def txnclosehook(tr2):
970 def txnclosehook(tr2):
971 """To be run if transaction is successful, will schedule a hook run
971 """To be run if transaction is successful, will schedule a hook run
972 """
972 """
973 def hook():
973 def hook():
974 reporef().hook('txnclose', throw=False, txnname=desc,
974 reporef().hook('txnclose', throw=False, txnname=desc,
975 **tr2.hookargs)
975 **tr2.hookargs)
976 reporef()._afterlock(hook)
976 reporef()._afterlock(hook)
977 tr.addfinalize('txnclose-hook', txnclosehook)
977 tr.addfinalize('txnclose-hook', txnclosehook)
978 self._transref = weakref.ref(tr)
978 self._transref = weakref.ref(tr)
979 return tr
979 return tr
980
980
981 def _journalfiles(self):
981 def _journalfiles(self):
982 return ((self.svfs, 'journal'),
982 return ((self.svfs, 'journal'),
983 (self.vfs, 'journal.dirstate'),
983 (self.vfs, 'journal.dirstate'),
984 (self.vfs, 'journal.branch'),
984 (self.vfs, 'journal.branch'),
985 (self.vfs, 'journal.desc'),
985 (self.vfs, 'journal.desc'),
986 (self.vfs, 'journal.bookmarks'),
986 (self.vfs, 'journal.bookmarks'),
987 (self.svfs, 'journal.phaseroots'))
987 (self.svfs, 'journal.phaseroots'))
988
988
989 def undofiles(self):
989 def undofiles(self):
990 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
990 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
991
991
992 def _writejournal(self, desc):
992 def _writejournal(self, desc):
993 self.vfs.write("journal.dirstate",
993 self.vfs.write("journal.dirstate",
994 self.vfs.tryread("dirstate"))
994 self.vfs.tryread("dirstate"))
995 self.vfs.write("journal.branch",
995 self.vfs.write("journal.branch",
996 encoding.fromlocal(self.dirstate.branch()))
996 encoding.fromlocal(self.dirstate.branch()))
997 self.vfs.write("journal.desc",
997 self.vfs.write("journal.desc",
998 "%d\n%s\n" % (len(self), desc))
998 "%d\n%s\n" % (len(self), desc))
999 self.vfs.write("journal.bookmarks",
999 self.vfs.write("journal.bookmarks",
1000 self.vfs.tryread("bookmarks"))
1000 self.vfs.tryread("bookmarks"))
1001 self.svfs.write("journal.phaseroots",
1001 self.svfs.write("journal.phaseroots",
1002 self.svfs.tryread("phaseroots"))
1002 self.svfs.tryread("phaseroots"))
1003
1003
1004 def recover(self):
1004 def recover(self):
1005 lock = self.lock()
1005 lock = self.lock()
1006 try:
1006 try:
1007 if self.svfs.exists("journal"):
1007 if self.svfs.exists("journal"):
1008 self.ui.status(_("rolling back interrupted transaction\n"))
1008 self.ui.status(_("rolling back interrupted transaction\n"))
1009 vfsmap = {'': self.svfs,
1009 vfsmap = {'': self.svfs,
1010 'plain': self.vfs,}
1010 'plain': self.vfs,}
1011 transaction.rollback(self.svfs, vfsmap, "journal",
1011 transaction.rollback(self.svfs, vfsmap, "journal",
1012 self.ui.warn)
1012 self.ui.warn)
1013 self.invalidate()
1013 self.invalidate()
1014 return True
1014 return True
1015 else:
1015 else:
1016 self.ui.warn(_("no interrupted transaction available\n"))
1016 self.ui.warn(_("no interrupted transaction available\n"))
1017 return False
1017 return False
1018 finally:
1018 finally:
1019 lock.release()
1019 lock.release()
1020
1020
1021 def rollback(self, dryrun=False, force=False):
1021 def rollback(self, dryrun=False, force=False):
1022 wlock = lock = None
1022 wlock = lock = None
1023 try:
1023 try:
1024 wlock = self.wlock()
1024 wlock = self.wlock()
1025 lock = self.lock()
1025 lock = self.lock()
1026 if self.svfs.exists("undo"):
1026 if self.svfs.exists("undo"):
1027 return self._rollback(dryrun, force)
1027 return self._rollback(dryrun, force)
1028 else:
1028 else:
1029 self.ui.warn(_("no rollback information available\n"))
1029 self.ui.warn(_("no rollback information available\n"))
1030 return 1
1030 return 1
1031 finally:
1031 finally:
1032 release(lock, wlock)
1032 release(lock, wlock)
1033
1033
1034 @unfilteredmethod # Until we get smarter cache management
1034 @unfilteredmethod # Until we get smarter cache management
1035 def _rollback(self, dryrun, force):
1035 def _rollback(self, dryrun, force):
1036 ui = self.ui
1036 ui = self.ui
1037 try:
1037 try:
1038 args = self.vfs.read('undo.desc').splitlines()
1038 args = self.vfs.read('undo.desc').splitlines()
1039 (oldlen, desc, detail) = (int(args[0]), args[1], None)
1039 (oldlen, desc, detail) = (int(args[0]), args[1], None)
1040 if len(args) >= 3:
1040 if len(args) >= 3:
1041 detail = args[2]
1041 detail = args[2]
1042 oldtip = oldlen - 1
1042 oldtip = oldlen - 1
1043
1043
1044 if detail and ui.verbose:
1044 if detail and ui.verbose:
1045 msg = (_('repository tip rolled back to revision %s'
1045 msg = (_('repository tip rolled back to revision %s'
1046 ' (undo %s: %s)\n')
1046 ' (undo %s: %s)\n')
1047 % (oldtip, desc, detail))
1047 % (oldtip, desc, detail))
1048 else:
1048 else:
1049 msg = (_('repository tip rolled back to revision %s'
1049 msg = (_('repository tip rolled back to revision %s'
1050 ' (undo %s)\n')
1050 ' (undo %s)\n')
1051 % (oldtip, desc))
1051 % (oldtip, desc))
1052 except IOError:
1052 except IOError:
1053 msg = _('rolling back unknown transaction\n')
1053 msg = _('rolling back unknown transaction\n')
1054 desc = None
1054 desc = None
1055
1055
1056 if not force and self['.'] != self['tip'] and desc == 'commit':
1056 if not force and self['.'] != self['tip'] and desc == 'commit':
1057 raise util.Abort(
1057 raise util.Abort(
1058 _('rollback of last commit while not checked out '
1058 _('rollback of last commit while not checked out '
1059 'may lose data'), hint=_('use -f to force'))
1059 'may lose data'), hint=_('use -f to force'))
1060
1060
1061 ui.status(msg)
1061 ui.status(msg)
1062 if dryrun:
1062 if dryrun:
1063 return 0
1063 return 0
1064
1064
1065 parents = self.dirstate.parents()
1065 parents = self.dirstate.parents()
1066 self.destroying()
1066 self.destroying()
1067 vfsmap = {'plain': self.vfs, '': self.svfs}
1067 vfsmap = {'plain': self.vfs, '': self.svfs}
1068 transaction.rollback(self.svfs, vfsmap, 'undo', ui.warn)
1068 transaction.rollback(self.svfs, vfsmap, 'undo', ui.warn)
1069 if self.vfs.exists('undo.bookmarks'):
1069 if self.vfs.exists('undo.bookmarks'):
1070 self.vfs.rename('undo.bookmarks', 'bookmarks')
1070 self.vfs.rename('undo.bookmarks', 'bookmarks')
1071 if self.svfs.exists('undo.phaseroots'):
1071 if self.svfs.exists('undo.phaseroots'):
1072 self.svfs.rename('undo.phaseroots', 'phaseroots')
1072 self.svfs.rename('undo.phaseroots', 'phaseroots')
1073 self.invalidate()
1073 self.invalidate()
1074
1074
1075 parentgone = (parents[0] not in self.changelog.nodemap or
1075 parentgone = (parents[0] not in self.changelog.nodemap or
1076 parents[1] not in self.changelog.nodemap)
1076 parents[1] not in self.changelog.nodemap)
1077 if parentgone:
1077 if parentgone:
1078 self.vfs.rename('undo.dirstate', 'dirstate')
1078 self.vfs.rename('undo.dirstate', 'dirstate')
1079 try:
1079 try:
1080 branch = self.vfs.read('undo.branch')
1080 branch = self.vfs.read('undo.branch')
1081 self.dirstate.setbranch(encoding.tolocal(branch))
1081 self.dirstate.setbranch(encoding.tolocal(branch))
1082 except IOError:
1082 except IOError:
1083 ui.warn(_('named branch could not be reset: '
1083 ui.warn(_('named branch could not be reset: '
1084 'current branch is still \'%s\'\n')
1084 'current branch is still \'%s\'\n')
1085 % self.dirstate.branch())
1085 % self.dirstate.branch())
1086
1086
1087 self.dirstate.invalidate()
1087 self.dirstate.invalidate()
1088 parents = tuple([p.rev() for p in self.parents()])
1088 parents = tuple([p.rev() for p in self.parents()])
1089 if len(parents) > 1:
1089 if len(parents) > 1:
1090 ui.status(_('working directory now based on '
1090 ui.status(_('working directory now based on '
1091 'revisions %d and %d\n') % parents)
1091 'revisions %d and %d\n') % parents)
1092 else:
1092 else:
1093 ui.status(_('working directory now based on '
1093 ui.status(_('working directory now based on '
1094 'revision %d\n') % parents)
1094 'revision %d\n') % parents)
1095 ms = mergemod.mergestate(self)
1096 ms.reset(self['.'].node())
1097
1095 # TODO: if we know which new heads may result from this rollback, pass
1098 # TODO: if we know which new heads may result from this rollback, pass
1096 # them to destroy(), which will prevent the branchhead cache from being
1099 # them to destroy(), which will prevent the branchhead cache from being
1097 # invalidated.
1100 # invalidated.
1098 self.destroyed()
1101 self.destroyed()
1099 return 0
1102 return 0
1100
1103
1101 def invalidatecaches(self):
1104 def invalidatecaches(self):
1102
1105
1103 if '_tagscache' in vars(self):
1106 if '_tagscache' in vars(self):
1104 # can't use delattr on proxy
1107 # can't use delattr on proxy
1105 del self.__dict__['_tagscache']
1108 del self.__dict__['_tagscache']
1106
1109
1107 self.unfiltered()._branchcaches.clear()
1110 self.unfiltered()._branchcaches.clear()
1108 self.invalidatevolatilesets()
1111 self.invalidatevolatilesets()
1109
1112
1110 def invalidatevolatilesets(self):
1113 def invalidatevolatilesets(self):
1111 self.filteredrevcache.clear()
1114 self.filteredrevcache.clear()
1112 obsolete.clearobscaches(self)
1115 obsolete.clearobscaches(self)
1113
1116
1114 def invalidatedirstate(self):
1117 def invalidatedirstate(self):
1115 '''Invalidates the dirstate, causing the next call to dirstate
1118 '''Invalidates the dirstate, causing the next call to dirstate
1116 to check if it was modified since the last time it was read,
1119 to check if it was modified since the last time it was read,
1117 rereading it if it has.
1120 rereading it if it has.
1118
1121
1119 This is different to dirstate.invalidate() that it doesn't always
1122 This is different to dirstate.invalidate() that it doesn't always
1120 rereads the dirstate. Use dirstate.invalidate() if you want to
1123 rereads the dirstate. Use dirstate.invalidate() if you want to
1121 explicitly read the dirstate again (i.e. restoring it to a previous
1124 explicitly read the dirstate again (i.e. restoring it to a previous
1122 known good state).'''
1125 known good state).'''
1123 if hasunfilteredcache(self, 'dirstate'):
1126 if hasunfilteredcache(self, 'dirstate'):
1124 for k in self.dirstate._filecache:
1127 for k in self.dirstate._filecache:
1125 try:
1128 try:
1126 delattr(self.dirstate, k)
1129 delattr(self.dirstate, k)
1127 except AttributeError:
1130 except AttributeError:
1128 pass
1131 pass
1129 delattr(self.unfiltered(), 'dirstate')
1132 delattr(self.unfiltered(), 'dirstate')
1130
1133
1131 def invalidate(self):
1134 def invalidate(self):
1132 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1135 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1133 for k in self._filecache:
1136 for k in self._filecache:
1134 # dirstate is invalidated separately in invalidatedirstate()
1137 # dirstate is invalidated separately in invalidatedirstate()
1135 if k == 'dirstate':
1138 if k == 'dirstate':
1136 continue
1139 continue
1137
1140
1138 try:
1141 try:
1139 delattr(unfiltered, k)
1142 delattr(unfiltered, k)
1140 except AttributeError:
1143 except AttributeError:
1141 pass
1144 pass
1142 self.invalidatecaches()
1145 self.invalidatecaches()
1143 self.store.invalidatecaches()
1146 self.store.invalidatecaches()
1144
1147
1145 def invalidateall(self):
1148 def invalidateall(self):
1146 '''Fully invalidates both store and non-store parts, causing the
1149 '''Fully invalidates both store and non-store parts, causing the
1147 subsequent operation to reread any outside changes.'''
1150 subsequent operation to reread any outside changes.'''
1148 # extension should hook this to invalidate its caches
1151 # extension should hook this to invalidate its caches
1149 self.invalidate()
1152 self.invalidate()
1150 self.invalidatedirstate()
1153 self.invalidatedirstate()
1151
1154
1152 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1155 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1153 try:
1156 try:
1154 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1157 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1155 except error.LockHeld, inst:
1158 except error.LockHeld, inst:
1156 if not wait:
1159 if not wait:
1157 raise
1160 raise
1158 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1161 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1159 (desc, inst.locker))
1162 (desc, inst.locker))
1160 # default to 600 seconds timeout
1163 # default to 600 seconds timeout
1161 l = lockmod.lock(vfs, lockname,
1164 l = lockmod.lock(vfs, lockname,
1162 int(self.ui.config("ui", "timeout", "600")),
1165 int(self.ui.config("ui", "timeout", "600")),
1163 releasefn, desc=desc)
1166 releasefn, desc=desc)
1164 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1167 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1165 if acquirefn:
1168 if acquirefn:
1166 acquirefn()
1169 acquirefn()
1167 return l
1170 return l
1168
1171
1169 def _afterlock(self, callback):
1172 def _afterlock(self, callback):
1170 """add a callback to the current repository lock.
1173 """add a callback to the current repository lock.
1171
1174
1172 The callback will be executed on lock release."""
1175 The callback will be executed on lock release."""
1173 l = self._lockref and self._lockref()
1176 l = self._lockref and self._lockref()
1174 if l:
1177 if l:
1175 l.postrelease.append(callback)
1178 l.postrelease.append(callback)
1176 else:
1179 else:
1177 callback()
1180 callback()
1178
1181
1179 def lock(self, wait=True):
1182 def lock(self, wait=True):
1180 '''Lock the repository store (.hg/store) and return a weak reference
1183 '''Lock the repository store (.hg/store) and return a weak reference
1181 to the lock. Use this before modifying the store (e.g. committing or
1184 to the lock. Use this before modifying the store (e.g. committing or
1182 stripping). If you are opening a transaction, get a lock as well.)
1185 stripping). If you are opening a transaction, get a lock as well.)
1183
1186
1184 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
1187 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
1185 'wlock' first to avoid a dead-lock hazard.'''
1188 'wlock' first to avoid a dead-lock hazard.'''
1186 l = self._lockref and self._lockref()
1189 l = self._lockref and self._lockref()
1187 if l is not None and l.held:
1190 if l is not None and l.held:
1188 l.lock()
1191 l.lock()
1189 return l
1192 return l
1190
1193
1191 def unlock():
1194 def unlock():
1192 for k, ce in self._filecache.items():
1195 for k, ce in self._filecache.items():
1193 if k == 'dirstate' or k not in self.__dict__:
1196 if k == 'dirstate' or k not in self.__dict__:
1194 continue
1197 continue
1195 ce.refresh()
1198 ce.refresh()
1196
1199
1197 l = self._lock(self.svfs, "lock", wait, unlock,
1200 l = self._lock(self.svfs, "lock", wait, unlock,
1198 self.invalidate, _('repository %s') % self.origroot)
1201 self.invalidate, _('repository %s') % self.origroot)
1199 self._lockref = weakref.ref(l)
1202 self._lockref = weakref.ref(l)
1200 return l
1203 return l
1201
1204
1202 def wlock(self, wait=True):
1205 def wlock(self, wait=True):
1203 '''Lock the non-store parts of the repository (everything under
1206 '''Lock the non-store parts of the repository (everything under
1204 .hg except .hg/store) and return a weak reference to the lock.
1207 .hg except .hg/store) and return a weak reference to the lock.
1205
1208
1206 Use this before modifying files in .hg.
1209 Use this before modifying files in .hg.
1207
1210
1208 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
1211 If both 'lock' and 'wlock' must be acquired, ensure you always acquires
1209 'wlock' first to avoid a dead-lock hazard.'''
1212 'wlock' first to avoid a dead-lock hazard.'''
1210 l = self._wlockref and self._wlockref()
1213 l = self._wlockref and self._wlockref()
1211 if l is not None and l.held:
1214 if l is not None and l.held:
1212 l.lock()
1215 l.lock()
1213 return l
1216 return l
1214
1217
1215 # We do not need to check for non-waiting lock aquisition. Such
1218 # We do not need to check for non-waiting lock aquisition. Such
1216 # acquisition would not cause dead-lock as they would just fail.
1219 # acquisition would not cause dead-lock as they would just fail.
1217 if wait and (self.ui.configbool('devel', 'all')
1220 if wait and (self.ui.configbool('devel', 'all')
1218 or self.ui.configbool('devel', 'check-locks')):
1221 or self.ui.configbool('devel', 'check-locks')):
1219 l = self._lockref and self._lockref()
1222 l = self._lockref and self._lockref()
1220 if l is not None and l.held:
1223 if l is not None and l.held:
1221 scmutil.develwarn(self.ui, '"wlock" acquired after "lock"')
1224 scmutil.develwarn(self.ui, '"wlock" acquired after "lock"')
1222
1225
1223 def unlock():
1226 def unlock():
1224 if self.dirstate.pendingparentchange():
1227 if self.dirstate.pendingparentchange():
1225 self.dirstate.invalidate()
1228 self.dirstate.invalidate()
1226 else:
1229 else:
1227 self.dirstate.write()
1230 self.dirstate.write()
1228
1231
1229 self._filecache['dirstate'].refresh()
1232 self._filecache['dirstate'].refresh()
1230
1233
1231 l = self._lock(self.vfs, "wlock", wait, unlock,
1234 l = self._lock(self.vfs, "wlock", wait, unlock,
1232 self.invalidatedirstate, _('working directory of %s') %
1235 self.invalidatedirstate, _('working directory of %s') %
1233 self.origroot)
1236 self.origroot)
1234 self._wlockref = weakref.ref(l)
1237 self._wlockref = weakref.ref(l)
1235 return l
1238 return l
1236
1239
1237 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1240 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1238 """
1241 """
1239 commit an individual file as part of a larger transaction
1242 commit an individual file as part of a larger transaction
1240 """
1243 """
1241
1244
1242 fname = fctx.path()
1245 fname = fctx.path()
1243 fparent1 = manifest1.get(fname, nullid)
1246 fparent1 = manifest1.get(fname, nullid)
1244 fparent2 = manifest2.get(fname, nullid)
1247 fparent2 = manifest2.get(fname, nullid)
1245 if isinstance(fctx, context.filectx):
1248 if isinstance(fctx, context.filectx):
1246 node = fctx.filenode()
1249 node = fctx.filenode()
1247 if node in [fparent1, fparent2]:
1250 if node in [fparent1, fparent2]:
1248 self.ui.debug('reusing %s filelog entry\n' % fname)
1251 self.ui.debug('reusing %s filelog entry\n' % fname)
1249 return node
1252 return node
1250
1253
1251 flog = self.file(fname)
1254 flog = self.file(fname)
1252 meta = {}
1255 meta = {}
1253 copy = fctx.renamed()
1256 copy = fctx.renamed()
1254 if copy and copy[0] != fname:
1257 if copy and copy[0] != fname:
1255 # Mark the new revision of this file as a copy of another
1258 # Mark the new revision of this file as a copy of another
1256 # file. This copy data will effectively act as a parent
1259 # file. This copy data will effectively act as a parent
1257 # of this new revision. If this is a merge, the first
1260 # of this new revision. If this is a merge, the first
1258 # parent will be the nullid (meaning "look up the copy data")
1261 # parent will be the nullid (meaning "look up the copy data")
1259 # and the second one will be the other parent. For example:
1262 # and the second one will be the other parent. For example:
1260 #
1263 #
1261 # 0 --- 1 --- 3 rev1 changes file foo
1264 # 0 --- 1 --- 3 rev1 changes file foo
1262 # \ / rev2 renames foo to bar and changes it
1265 # \ / rev2 renames foo to bar and changes it
1263 # \- 2 -/ rev3 should have bar with all changes and
1266 # \- 2 -/ rev3 should have bar with all changes and
1264 # should record that bar descends from
1267 # should record that bar descends from
1265 # bar in rev2 and foo in rev1
1268 # bar in rev2 and foo in rev1
1266 #
1269 #
1267 # this allows this merge to succeed:
1270 # this allows this merge to succeed:
1268 #
1271 #
1269 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1272 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1270 # \ / merging rev3 and rev4 should use bar@rev2
1273 # \ / merging rev3 and rev4 should use bar@rev2
1271 # \- 2 --- 4 as the merge base
1274 # \- 2 --- 4 as the merge base
1272 #
1275 #
1273
1276
1274 cfname = copy[0]
1277 cfname = copy[0]
1275 crev = manifest1.get(cfname)
1278 crev = manifest1.get(cfname)
1276 newfparent = fparent2
1279 newfparent = fparent2
1277
1280
1278 if manifest2: # branch merge
1281 if manifest2: # branch merge
1279 if fparent2 == nullid or crev is None: # copied on remote side
1282 if fparent2 == nullid or crev is None: # copied on remote side
1280 if cfname in manifest2:
1283 if cfname in manifest2:
1281 crev = manifest2[cfname]
1284 crev = manifest2[cfname]
1282 newfparent = fparent1
1285 newfparent = fparent1
1283
1286
1284 # Here, we used to search backwards through history to try to find
1287 # Here, we used to search backwards through history to try to find
1285 # where the file copy came from if the source of a copy was not in
1288 # where the file copy came from if the source of a copy was not in
1286 # the parent directory. However, this doesn't actually make sense to
1289 # the parent directory. However, this doesn't actually make sense to
1287 # do (what does a copy from something not in your working copy even
1290 # do (what does a copy from something not in your working copy even
1288 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
1291 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
1289 # the user that copy information was dropped, so if they didn't
1292 # the user that copy information was dropped, so if they didn't
1290 # expect this outcome it can be fixed, but this is the correct
1293 # expect this outcome it can be fixed, but this is the correct
1291 # behavior in this circumstance.
1294 # behavior in this circumstance.
1292
1295
1293 if crev:
1296 if crev:
1294 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1297 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1295 meta["copy"] = cfname
1298 meta["copy"] = cfname
1296 meta["copyrev"] = hex(crev)
1299 meta["copyrev"] = hex(crev)
1297 fparent1, fparent2 = nullid, newfparent
1300 fparent1, fparent2 = nullid, newfparent
1298 else:
1301 else:
1299 self.ui.warn(_("warning: can't find ancestor for '%s' "
1302 self.ui.warn(_("warning: can't find ancestor for '%s' "
1300 "copied from '%s'!\n") % (fname, cfname))
1303 "copied from '%s'!\n") % (fname, cfname))
1301
1304
1302 elif fparent1 == nullid:
1305 elif fparent1 == nullid:
1303 fparent1, fparent2 = fparent2, nullid
1306 fparent1, fparent2 = fparent2, nullid
1304 elif fparent2 != nullid:
1307 elif fparent2 != nullid:
1305 # is one parent an ancestor of the other?
1308 # is one parent an ancestor of the other?
1306 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1309 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1307 if fparent1 in fparentancestors:
1310 if fparent1 in fparentancestors:
1308 fparent1, fparent2 = fparent2, nullid
1311 fparent1, fparent2 = fparent2, nullid
1309 elif fparent2 in fparentancestors:
1312 elif fparent2 in fparentancestors:
1310 fparent2 = nullid
1313 fparent2 = nullid
1311
1314
1312 # is the file changed?
1315 # is the file changed?
1313 text = fctx.data()
1316 text = fctx.data()
1314 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1317 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1315 changelist.append(fname)
1318 changelist.append(fname)
1316 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1319 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1317 # are just the flags changed during merge?
1320 # are just the flags changed during merge?
1318 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1321 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1319 changelist.append(fname)
1322 changelist.append(fname)
1320
1323
1321 return fparent1
1324 return fparent1
1322
1325
1323 @unfilteredmethod
1326 @unfilteredmethod
1324 def commit(self, text="", user=None, date=None, match=None, force=False,
1327 def commit(self, text="", user=None, date=None, match=None, force=False,
1325 editor=False, extra={}):
1328 editor=False, extra={}):
1326 """Add a new revision to current repository.
1329 """Add a new revision to current repository.
1327
1330
1328 Revision information is gathered from the working directory,
1331 Revision information is gathered from the working directory,
1329 match can be used to filter the committed files. If editor is
1332 match can be used to filter the committed files. If editor is
1330 supplied, it is called to get a commit message.
1333 supplied, it is called to get a commit message.
1331 """
1334 """
1332
1335
1333 def fail(f, msg):
1336 def fail(f, msg):
1334 raise util.Abort('%s: %s' % (f, msg))
1337 raise util.Abort('%s: %s' % (f, msg))
1335
1338
1336 if not match:
1339 if not match:
1337 match = matchmod.always(self.root, '')
1340 match = matchmod.always(self.root, '')
1338
1341
1339 if not force:
1342 if not force:
1340 vdirs = []
1343 vdirs = []
1341 match.explicitdir = vdirs.append
1344 match.explicitdir = vdirs.append
1342 match.bad = fail
1345 match.bad = fail
1343
1346
1344 wlock = self.wlock()
1347 wlock = self.wlock()
1345 try:
1348 try:
1346 wctx = self[None]
1349 wctx = self[None]
1347 merge = len(wctx.parents()) > 1
1350 merge = len(wctx.parents()) > 1
1348
1351
1349 if not force and merge and not match.always():
1352 if not force and merge and not match.always():
1350 raise util.Abort(_('cannot partially commit a merge '
1353 raise util.Abort(_('cannot partially commit a merge '
1351 '(do not specify files or patterns)'))
1354 '(do not specify files or patterns)'))
1352
1355
1353 status = self.status(match=match, clean=force)
1356 status = self.status(match=match, clean=force)
1354 if force:
1357 if force:
1355 status.modified.extend(status.clean) # mq may commit clean files
1358 status.modified.extend(status.clean) # mq may commit clean files
1356
1359
1357 # check subrepos
1360 # check subrepos
1358 subs = []
1361 subs = []
1359 commitsubs = set()
1362 commitsubs = set()
1360 newstate = wctx.substate.copy()
1363 newstate = wctx.substate.copy()
1361 # only manage subrepos and .hgsubstate if .hgsub is present
1364 # only manage subrepos and .hgsubstate if .hgsub is present
1362 if '.hgsub' in wctx:
1365 if '.hgsub' in wctx:
1363 # we'll decide whether to track this ourselves, thanks
1366 # we'll decide whether to track this ourselves, thanks
1364 for c in status.modified, status.added, status.removed:
1367 for c in status.modified, status.added, status.removed:
1365 if '.hgsubstate' in c:
1368 if '.hgsubstate' in c:
1366 c.remove('.hgsubstate')
1369 c.remove('.hgsubstate')
1367
1370
1368 # compare current state to last committed state
1371 # compare current state to last committed state
1369 # build new substate based on last committed state
1372 # build new substate based on last committed state
1370 oldstate = wctx.p1().substate
1373 oldstate = wctx.p1().substate
1371 for s in sorted(newstate.keys()):
1374 for s in sorted(newstate.keys()):
1372 if not match(s):
1375 if not match(s):
1373 # ignore working copy, use old state if present
1376 # ignore working copy, use old state if present
1374 if s in oldstate:
1377 if s in oldstate:
1375 newstate[s] = oldstate[s]
1378 newstate[s] = oldstate[s]
1376 continue
1379 continue
1377 if not force:
1380 if not force:
1378 raise util.Abort(
1381 raise util.Abort(
1379 _("commit with new subrepo %s excluded") % s)
1382 _("commit with new subrepo %s excluded") % s)
1380 dirtyreason = wctx.sub(s).dirtyreason(True)
1383 dirtyreason = wctx.sub(s).dirtyreason(True)
1381 if dirtyreason:
1384 if dirtyreason:
1382 if not self.ui.configbool('ui', 'commitsubrepos'):
1385 if not self.ui.configbool('ui', 'commitsubrepos'):
1383 raise util.Abort(dirtyreason,
1386 raise util.Abort(dirtyreason,
1384 hint=_("use --subrepos for recursive commit"))
1387 hint=_("use --subrepos for recursive commit"))
1385 subs.append(s)
1388 subs.append(s)
1386 commitsubs.add(s)
1389 commitsubs.add(s)
1387 else:
1390 else:
1388 bs = wctx.sub(s).basestate()
1391 bs = wctx.sub(s).basestate()
1389 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1392 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1390 if oldstate.get(s, (None, None, None))[1] != bs:
1393 if oldstate.get(s, (None, None, None))[1] != bs:
1391 subs.append(s)
1394 subs.append(s)
1392
1395
1393 # check for removed subrepos
1396 # check for removed subrepos
1394 for p in wctx.parents():
1397 for p in wctx.parents():
1395 r = [s for s in p.substate if s not in newstate]
1398 r = [s for s in p.substate if s not in newstate]
1396 subs += [s for s in r if match(s)]
1399 subs += [s for s in r if match(s)]
1397 if subs:
1400 if subs:
1398 if (not match('.hgsub') and
1401 if (not match('.hgsub') and
1399 '.hgsub' in (wctx.modified() + wctx.added())):
1402 '.hgsub' in (wctx.modified() + wctx.added())):
1400 raise util.Abort(
1403 raise util.Abort(
1401 _("can't commit subrepos without .hgsub"))
1404 _("can't commit subrepos without .hgsub"))
1402 status.modified.insert(0, '.hgsubstate')
1405 status.modified.insert(0, '.hgsubstate')
1403
1406
1404 elif '.hgsub' in status.removed:
1407 elif '.hgsub' in status.removed:
1405 # clean up .hgsubstate when .hgsub is removed
1408 # clean up .hgsubstate when .hgsub is removed
1406 if ('.hgsubstate' in wctx and
1409 if ('.hgsubstate' in wctx and
1407 '.hgsubstate' not in (status.modified + status.added +
1410 '.hgsubstate' not in (status.modified + status.added +
1408 status.removed)):
1411 status.removed)):
1409 status.removed.insert(0, '.hgsubstate')
1412 status.removed.insert(0, '.hgsubstate')
1410
1413
1411 # make sure all explicit patterns are matched
1414 # make sure all explicit patterns are matched
1412 if not force and match.files():
1415 if not force and match.files():
1413 matched = set(status.modified + status.added + status.removed)
1416 matched = set(status.modified + status.added + status.removed)
1414
1417
1415 for f in match.files():
1418 for f in match.files():
1416 f = self.dirstate.normalize(f)
1419 f = self.dirstate.normalize(f)
1417 if f == '.' or f in matched or f in wctx.substate:
1420 if f == '.' or f in matched or f in wctx.substate:
1418 continue
1421 continue
1419 if f in status.deleted:
1422 if f in status.deleted:
1420 fail(f, _('file not found!'))
1423 fail(f, _('file not found!'))
1421 if f in vdirs: # visited directory
1424 if f in vdirs: # visited directory
1422 d = f + '/'
1425 d = f + '/'
1423 for mf in matched:
1426 for mf in matched:
1424 if mf.startswith(d):
1427 if mf.startswith(d):
1425 break
1428 break
1426 else:
1429 else:
1427 fail(f, _("no match under directory!"))
1430 fail(f, _("no match under directory!"))
1428 elif f not in self.dirstate:
1431 elif f not in self.dirstate:
1429 fail(f, _("file not tracked!"))
1432 fail(f, _("file not tracked!"))
1430
1433
1431 cctx = context.workingcommitctx(self, status,
1434 cctx = context.workingcommitctx(self, status,
1432 text, user, date, extra)
1435 text, user, date, extra)
1433
1436
1434 if (not force and not extra.get("close") and not merge
1437 if (not force and not extra.get("close") and not merge
1435 and not cctx.files()
1438 and not cctx.files()
1436 and wctx.branch() == wctx.p1().branch()):
1439 and wctx.branch() == wctx.p1().branch()):
1437 return None
1440 return None
1438
1441
1439 if merge and cctx.deleted():
1442 if merge and cctx.deleted():
1440 raise util.Abort(_("cannot commit merge with missing files"))
1443 raise util.Abort(_("cannot commit merge with missing files"))
1441
1444
1442 ms = mergemod.mergestate(self)
1445 ms = mergemod.mergestate(self)
1443 for f in status.modified:
1446 for f in status.modified:
1444 if f in ms and ms[f] == 'u':
1447 if f in ms and ms[f] == 'u':
1445 raise util.Abort(_('unresolved merge conflicts '
1448 raise util.Abort(_('unresolved merge conflicts '
1446 '(see "hg help resolve")'))
1449 '(see "hg help resolve")'))
1447
1450
1448 if editor:
1451 if editor:
1449 cctx._text = editor(self, cctx, subs)
1452 cctx._text = editor(self, cctx, subs)
1450 edited = (text != cctx._text)
1453 edited = (text != cctx._text)
1451
1454
1452 # Save commit message in case this transaction gets rolled back
1455 # Save commit message in case this transaction gets rolled back
1453 # (e.g. by a pretxncommit hook). Leave the content alone on
1456 # (e.g. by a pretxncommit hook). Leave the content alone on
1454 # the assumption that the user will use the same editor again.
1457 # the assumption that the user will use the same editor again.
1455 msgfn = self.savecommitmessage(cctx._text)
1458 msgfn = self.savecommitmessage(cctx._text)
1456
1459
1457 # commit subs and write new state
1460 # commit subs and write new state
1458 if subs:
1461 if subs:
1459 for s in sorted(commitsubs):
1462 for s in sorted(commitsubs):
1460 sub = wctx.sub(s)
1463 sub = wctx.sub(s)
1461 self.ui.status(_('committing subrepository %s\n') %
1464 self.ui.status(_('committing subrepository %s\n') %
1462 subrepo.subrelpath(sub))
1465 subrepo.subrelpath(sub))
1463 sr = sub.commit(cctx._text, user, date)
1466 sr = sub.commit(cctx._text, user, date)
1464 newstate[s] = (newstate[s][0], sr)
1467 newstate[s] = (newstate[s][0], sr)
1465 subrepo.writestate(self, newstate)
1468 subrepo.writestate(self, newstate)
1466
1469
1467 p1, p2 = self.dirstate.parents()
1470 p1, p2 = self.dirstate.parents()
1468 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1471 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1469 try:
1472 try:
1470 self.hook("precommit", throw=True, parent1=hookp1,
1473 self.hook("precommit", throw=True, parent1=hookp1,
1471 parent2=hookp2)
1474 parent2=hookp2)
1472 ret = self.commitctx(cctx, True)
1475 ret = self.commitctx(cctx, True)
1473 except: # re-raises
1476 except: # re-raises
1474 if edited:
1477 if edited:
1475 self.ui.write(
1478 self.ui.write(
1476 _('note: commit message saved in %s\n') % msgfn)
1479 _('note: commit message saved in %s\n') % msgfn)
1477 raise
1480 raise
1478
1481
1479 # update bookmarks, dirstate and mergestate
1482 # update bookmarks, dirstate and mergestate
1480 bookmarks.update(self, [p1, p2], ret)
1483 bookmarks.update(self, [p1, p2], ret)
1481 cctx.markcommitted(ret)
1484 cctx.markcommitted(ret)
1482 ms.reset()
1485 ms.reset()
1483 finally:
1486 finally:
1484 wlock.release()
1487 wlock.release()
1485
1488
1486 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1489 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1487 # hack for command that use a temporary commit (eg: histedit)
1490 # hack for command that use a temporary commit (eg: histedit)
1488 # temporary commit got stripped before hook release
1491 # temporary commit got stripped before hook release
1489 if node in self:
1492 if node in self:
1490 self.hook("commit", node=node, parent1=parent1,
1493 self.hook("commit", node=node, parent1=parent1,
1491 parent2=parent2)
1494 parent2=parent2)
1492 self._afterlock(commithook)
1495 self._afterlock(commithook)
1493 return ret
1496 return ret
1494
1497
1495 @unfilteredmethod
1498 @unfilteredmethod
1496 def commitctx(self, ctx, error=False):
1499 def commitctx(self, ctx, error=False):
1497 """Add a new revision to current repository.
1500 """Add a new revision to current repository.
1498 Revision information is passed via the context argument.
1501 Revision information is passed via the context argument.
1499 """
1502 """
1500
1503
1501 tr = None
1504 tr = None
1502 p1, p2 = ctx.p1(), ctx.p2()
1505 p1, p2 = ctx.p1(), ctx.p2()
1503 user = ctx.user()
1506 user = ctx.user()
1504
1507
1505 lock = self.lock()
1508 lock = self.lock()
1506 try:
1509 try:
1507 tr = self.transaction("commit")
1510 tr = self.transaction("commit")
1508 trp = weakref.proxy(tr)
1511 trp = weakref.proxy(tr)
1509
1512
1510 if ctx.files():
1513 if ctx.files():
1511 m1 = p1.manifest()
1514 m1 = p1.manifest()
1512 m2 = p2.manifest()
1515 m2 = p2.manifest()
1513 m = m1.copy()
1516 m = m1.copy()
1514
1517
1515 # check in files
1518 # check in files
1516 added = []
1519 added = []
1517 changed = []
1520 changed = []
1518 removed = list(ctx.removed())
1521 removed = list(ctx.removed())
1519 linkrev = len(self)
1522 linkrev = len(self)
1520 self.ui.note(_("committing files:\n"))
1523 self.ui.note(_("committing files:\n"))
1521 for f in sorted(ctx.modified() + ctx.added()):
1524 for f in sorted(ctx.modified() + ctx.added()):
1522 self.ui.note(f + "\n")
1525 self.ui.note(f + "\n")
1523 try:
1526 try:
1524 fctx = ctx[f]
1527 fctx = ctx[f]
1525 if fctx is None:
1528 if fctx is None:
1526 removed.append(f)
1529 removed.append(f)
1527 else:
1530 else:
1528 added.append(f)
1531 added.append(f)
1529 m[f] = self._filecommit(fctx, m1, m2, linkrev,
1532 m[f] = self._filecommit(fctx, m1, m2, linkrev,
1530 trp, changed)
1533 trp, changed)
1531 m.setflag(f, fctx.flags())
1534 m.setflag(f, fctx.flags())
1532 except OSError, inst:
1535 except OSError, inst:
1533 self.ui.warn(_("trouble committing %s!\n") % f)
1536 self.ui.warn(_("trouble committing %s!\n") % f)
1534 raise
1537 raise
1535 except IOError, inst:
1538 except IOError, inst:
1536 errcode = getattr(inst, 'errno', errno.ENOENT)
1539 errcode = getattr(inst, 'errno', errno.ENOENT)
1537 if error or errcode and errcode != errno.ENOENT:
1540 if error or errcode and errcode != errno.ENOENT:
1538 self.ui.warn(_("trouble committing %s!\n") % f)
1541 self.ui.warn(_("trouble committing %s!\n") % f)
1539 raise
1542 raise
1540
1543
1541 # update manifest
1544 # update manifest
1542 self.ui.note(_("committing manifest\n"))
1545 self.ui.note(_("committing manifest\n"))
1543 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1546 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1544 drop = [f for f in removed if f in m]
1547 drop = [f for f in removed if f in m]
1545 for f in drop:
1548 for f in drop:
1546 del m[f]
1549 del m[f]
1547 mn = self.manifest.add(m, trp, linkrev,
1550 mn = self.manifest.add(m, trp, linkrev,
1548 p1.manifestnode(), p2.manifestnode(),
1551 p1.manifestnode(), p2.manifestnode(),
1549 added, drop)
1552 added, drop)
1550 files = changed + removed
1553 files = changed + removed
1551 else:
1554 else:
1552 mn = p1.manifestnode()
1555 mn = p1.manifestnode()
1553 files = []
1556 files = []
1554
1557
1555 # update changelog
1558 # update changelog
1556 self.ui.note(_("committing changelog\n"))
1559 self.ui.note(_("committing changelog\n"))
1557 self.changelog.delayupdate(tr)
1560 self.changelog.delayupdate(tr)
1558 n = self.changelog.add(mn, files, ctx.description(),
1561 n = self.changelog.add(mn, files, ctx.description(),
1559 trp, p1.node(), p2.node(),
1562 trp, p1.node(), p2.node(),
1560 user, ctx.date(), ctx.extra().copy())
1563 user, ctx.date(), ctx.extra().copy())
1561 p = lambda: tr.writepending() and self.root or ""
1564 p = lambda: tr.writepending() and self.root or ""
1562 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1565 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1563 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1566 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1564 parent2=xp2, pending=p)
1567 parent2=xp2, pending=p)
1565 # set the new commit is proper phase
1568 # set the new commit is proper phase
1566 targetphase = subrepo.newcommitphase(self.ui, ctx)
1569 targetphase = subrepo.newcommitphase(self.ui, ctx)
1567 if targetphase:
1570 if targetphase:
1568 # retract boundary do not alter parent changeset.
1571 # retract boundary do not alter parent changeset.
1569 # if a parent have higher the resulting phase will
1572 # if a parent have higher the resulting phase will
1570 # be compliant anyway
1573 # be compliant anyway
1571 #
1574 #
1572 # if minimal phase was 0 we don't need to retract anything
1575 # if minimal phase was 0 we don't need to retract anything
1573 phases.retractboundary(self, tr, targetphase, [n])
1576 phases.retractboundary(self, tr, targetphase, [n])
1574 tr.close()
1577 tr.close()
1575 branchmap.updatecache(self.filtered('served'))
1578 branchmap.updatecache(self.filtered('served'))
1576 return n
1579 return n
1577 finally:
1580 finally:
1578 if tr:
1581 if tr:
1579 tr.release()
1582 tr.release()
1580 lock.release()
1583 lock.release()
1581
1584
1582 @unfilteredmethod
1585 @unfilteredmethod
1583 def destroying(self):
1586 def destroying(self):
1584 '''Inform the repository that nodes are about to be destroyed.
1587 '''Inform the repository that nodes are about to be destroyed.
1585 Intended for use by strip and rollback, so there's a common
1588 Intended for use by strip and rollback, so there's a common
1586 place for anything that has to be done before destroying history.
1589 place for anything that has to be done before destroying history.
1587
1590
1588 This is mostly useful for saving state that is in memory and waiting
1591 This is mostly useful for saving state that is in memory and waiting
1589 to be flushed when the current lock is released. Because a call to
1592 to be flushed when the current lock is released. Because a call to
1590 destroyed is imminent, the repo will be invalidated causing those
1593 destroyed is imminent, the repo will be invalidated causing those
1591 changes to stay in memory (waiting for the next unlock), or vanish
1594 changes to stay in memory (waiting for the next unlock), or vanish
1592 completely.
1595 completely.
1593 '''
1596 '''
1594 # When using the same lock to commit and strip, the phasecache is left
1597 # When using the same lock to commit and strip, the phasecache is left
1595 # dirty after committing. Then when we strip, the repo is invalidated,
1598 # dirty after committing. Then when we strip, the repo is invalidated,
1596 # causing those changes to disappear.
1599 # causing those changes to disappear.
1597 if '_phasecache' in vars(self):
1600 if '_phasecache' in vars(self):
1598 self._phasecache.write()
1601 self._phasecache.write()
1599
1602
1600 @unfilteredmethod
1603 @unfilteredmethod
1601 def destroyed(self):
1604 def destroyed(self):
1602 '''Inform the repository that nodes have been destroyed.
1605 '''Inform the repository that nodes have been destroyed.
1603 Intended for use by strip and rollback, so there's a common
1606 Intended for use by strip and rollback, so there's a common
1604 place for anything that has to be done after destroying history.
1607 place for anything that has to be done after destroying history.
1605 '''
1608 '''
1606 # When one tries to:
1609 # When one tries to:
1607 # 1) destroy nodes thus calling this method (e.g. strip)
1610 # 1) destroy nodes thus calling this method (e.g. strip)
1608 # 2) use phasecache somewhere (e.g. commit)
1611 # 2) use phasecache somewhere (e.g. commit)
1609 #
1612 #
1610 # then 2) will fail because the phasecache contains nodes that were
1613 # then 2) will fail because the phasecache contains nodes that were
1611 # removed. We can either remove phasecache from the filecache,
1614 # removed. We can either remove phasecache from the filecache,
1612 # causing it to reload next time it is accessed, or simply filter
1615 # causing it to reload next time it is accessed, or simply filter
1613 # the removed nodes now and write the updated cache.
1616 # the removed nodes now and write the updated cache.
1614 self._phasecache.filterunknown(self)
1617 self._phasecache.filterunknown(self)
1615 self._phasecache.write()
1618 self._phasecache.write()
1616
1619
1617 # update the 'served' branch cache to help read only server process
1620 # update the 'served' branch cache to help read only server process
1618 # Thanks to branchcache collaboration this is done from the nearest
1621 # Thanks to branchcache collaboration this is done from the nearest
1619 # filtered subset and it is expected to be fast.
1622 # filtered subset and it is expected to be fast.
1620 branchmap.updatecache(self.filtered('served'))
1623 branchmap.updatecache(self.filtered('served'))
1621
1624
1622 # Ensure the persistent tag cache is updated. Doing it now
1625 # Ensure the persistent tag cache is updated. Doing it now
1623 # means that the tag cache only has to worry about destroyed
1626 # means that the tag cache only has to worry about destroyed
1624 # heads immediately after a strip/rollback. That in turn
1627 # heads immediately after a strip/rollback. That in turn
1625 # guarantees that "cachetip == currenttip" (comparing both rev
1628 # guarantees that "cachetip == currenttip" (comparing both rev
1626 # and node) always means no nodes have been added or destroyed.
1629 # and node) always means no nodes have been added or destroyed.
1627
1630
1628 # XXX this is suboptimal when qrefresh'ing: we strip the current
1631 # XXX this is suboptimal when qrefresh'ing: we strip the current
1629 # head, refresh the tag cache, then immediately add a new head.
1632 # head, refresh the tag cache, then immediately add a new head.
1630 # But I think doing it this way is necessary for the "instant
1633 # But I think doing it this way is necessary for the "instant
1631 # tag cache retrieval" case to work.
1634 # tag cache retrieval" case to work.
1632 self.invalidate()
1635 self.invalidate()
1633
1636
1634 def walk(self, match, node=None):
1637 def walk(self, match, node=None):
1635 '''
1638 '''
1636 walk recursively through the directory tree or a given
1639 walk recursively through the directory tree or a given
1637 changeset, finding all files matched by the match
1640 changeset, finding all files matched by the match
1638 function
1641 function
1639 '''
1642 '''
1640 return self[node].walk(match)
1643 return self[node].walk(match)
1641
1644
1642 def status(self, node1='.', node2=None, match=None,
1645 def status(self, node1='.', node2=None, match=None,
1643 ignored=False, clean=False, unknown=False,
1646 ignored=False, clean=False, unknown=False,
1644 listsubrepos=False):
1647 listsubrepos=False):
1645 '''a convenience method that calls node1.status(node2)'''
1648 '''a convenience method that calls node1.status(node2)'''
1646 return self[node1].status(node2, match, ignored, clean, unknown,
1649 return self[node1].status(node2, match, ignored, clean, unknown,
1647 listsubrepos)
1650 listsubrepos)
1648
1651
1649 def heads(self, start=None):
1652 def heads(self, start=None):
1650 heads = self.changelog.heads(start)
1653 heads = self.changelog.heads(start)
1651 # sort the output in rev descending order
1654 # sort the output in rev descending order
1652 return sorted(heads, key=self.changelog.rev, reverse=True)
1655 return sorted(heads, key=self.changelog.rev, reverse=True)
1653
1656
1654 def branchheads(self, branch=None, start=None, closed=False):
1657 def branchheads(self, branch=None, start=None, closed=False):
1655 '''return a (possibly filtered) list of heads for the given branch
1658 '''return a (possibly filtered) list of heads for the given branch
1656
1659
1657 Heads are returned in topological order, from newest to oldest.
1660 Heads are returned in topological order, from newest to oldest.
1658 If branch is None, use the dirstate branch.
1661 If branch is None, use the dirstate branch.
1659 If start is not None, return only heads reachable from start.
1662 If start is not None, return only heads reachable from start.
1660 If closed is True, return heads that are marked as closed as well.
1663 If closed is True, return heads that are marked as closed as well.
1661 '''
1664 '''
1662 if branch is None:
1665 if branch is None:
1663 branch = self[None].branch()
1666 branch = self[None].branch()
1664 branches = self.branchmap()
1667 branches = self.branchmap()
1665 if branch not in branches:
1668 if branch not in branches:
1666 return []
1669 return []
1667 # the cache returns heads ordered lowest to highest
1670 # the cache returns heads ordered lowest to highest
1668 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1671 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1669 if start is not None:
1672 if start is not None:
1670 # filter out the heads that cannot be reached from startrev
1673 # filter out the heads that cannot be reached from startrev
1671 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1674 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1672 bheads = [h for h in bheads if h in fbheads]
1675 bheads = [h for h in bheads if h in fbheads]
1673 return bheads
1676 return bheads
1674
1677
1675 def branches(self, nodes):
1678 def branches(self, nodes):
1676 if not nodes:
1679 if not nodes:
1677 nodes = [self.changelog.tip()]
1680 nodes = [self.changelog.tip()]
1678 b = []
1681 b = []
1679 for n in nodes:
1682 for n in nodes:
1680 t = n
1683 t = n
1681 while True:
1684 while True:
1682 p = self.changelog.parents(n)
1685 p = self.changelog.parents(n)
1683 if p[1] != nullid or p[0] == nullid:
1686 if p[1] != nullid or p[0] == nullid:
1684 b.append((t, n, p[0], p[1]))
1687 b.append((t, n, p[0], p[1]))
1685 break
1688 break
1686 n = p[0]
1689 n = p[0]
1687 return b
1690 return b
1688
1691
1689 def between(self, pairs):
1692 def between(self, pairs):
1690 r = []
1693 r = []
1691
1694
1692 for top, bottom in pairs:
1695 for top, bottom in pairs:
1693 n, l, i = top, [], 0
1696 n, l, i = top, [], 0
1694 f = 1
1697 f = 1
1695
1698
1696 while n != bottom and n != nullid:
1699 while n != bottom and n != nullid:
1697 p = self.changelog.parents(n)[0]
1700 p = self.changelog.parents(n)[0]
1698 if i == f:
1701 if i == f:
1699 l.append(n)
1702 l.append(n)
1700 f = f * 2
1703 f = f * 2
1701 n = p
1704 n = p
1702 i += 1
1705 i += 1
1703
1706
1704 r.append(l)
1707 r.append(l)
1705
1708
1706 return r
1709 return r
1707
1710
1708 def checkpush(self, pushop):
1711 def checkpush(self, pushop):
1709 """Extensions can override this function if additional checks have
1712 """Extensions can override this function if additional checks have
1710 to be performed before pushing, or call it if they override push
1713 to be performed before pushing, or call it if they override push
1711 command.
1714 command.
1712 """
1715 """
1713 pass
1716 pass
1714
1717
1715 @unfilteredpropertycache
1718 @unfilteredpropertycache
1716 def prepushoutgoinghooks(self):
1719 def prepushoutgoinghooks(self):
1717 """Return util.hooks consists of "(repo, remote, outgoing)"
1720 """Return util.hooks consists of "(repo, remote, outgoing)"
1718 functions, which are called before pushing changesets.
1721 functions, which are called before pushing changesets.
1719 """
1722 """
1720 return util.hooks()
1723 return util.hooks()
1721
1724
1722 def stream_in(self, remote, requirements):
1725 def stream_in(self, remote, requirements):
1723 lock = self.lock()
1726 lock = self.lock()
1724 try:
1727 try:
1725 # Save remote branchmap. We will use it later
1728 # Save remote branchmap. We will use it later
1726 # to speed up branchcache creation
1729 # to speed up branchcache creation
1727 rbranchmap = None
1730 rbranchmap = None
1728 if remote.capable("branchmap"):
1731 if remote.capable("branchmap"):
1729 rbranchmap = remote.branchmap()
1732 rbranchmap = remote.branchmap()
1730
1733
1731 fp = remote.stream_out()
1734 fp = remote.stream_out()
1732 l = fp.readline()
1735 l = fp.readline()
1733 try:
1736 try:
1734 resp = int(l)
1737 resp = int(l)
1735 except ValueError:
1738 except ValueError:
1736 raise error.ResponseError(
1739 raise error.ResponseError(
1737 _('unexpected response from remote server:'), l)
1740 _('unexpected response from remote server:'), l)
1738 if resp == 1:
1741 if resp == 1:
1739 raise util.Abort(_('operation forbidden by server'))
1742 raise util.Abort(_('operation forbidden by server'))
1740 elif resp == 2:
1743 elif resp == 2:
1741 raise util.Abort(_('locking the remote repository failed'))
1744 raise util.Abort(_('locking the remote repository failed'))
1742 elif resp != 0:
1745 elif resp != 0:
1743 raise util.Abort(_('the server sent an unknown error code'))
1746 raise util.Abort(_('the server sent an unknown error code'))
1744 self.ui.status(_('streaming all changes\n'))
1747 self.ui.status(_('streaming all changes\n'))
1745 l = fp.readline()
1748 l = fp.readline()
1746 try:
1749 try:
1747 total_files, total_bytes = map(int, l.split(' ', 1))
1750 total_files, total_bytes = map(int, l.split(' ', 1))
1748 except (ValueError, TypeError):
1751 except (ValueError, TypeError):
1749 raise error.ResponseError(
1752 raise error.ResponseError(
1750 _('unexpected response from remote server:'), l)
1753 _('unexpected response from remote server:'), l)
1751 self.ui.status(_('%d files to transfer, %s of data\n') %
1754 self.ui.status(_('%d files to transfer, %s of data\n') %
1752 (total_files, util.bytecount(total_bytes)))
1755 (total_files, util.bytecount(total_bytes)))
1753 handled_bytes = 0
1756 handled_bytes = 0
1754 self.ui.progress(_('clone'), 0, total=total_bytes)
1757 self.ui.progress(_('clone'), 0, total=total_bytes)
1755 start = time.time()
1758 start = time.time()
1756
1759
1757 tr = self.transaction(_('clone'))
1760 tr = self.transaction(_('clone'))
1758 try:
1761 try:
1759 for i in xrange(total_files):
1762 for i in xrange(total_files):
1760 # XXX doesn't support '\n' or '\r' in filenames
1763 # XXX doesn't support '\n' or '\r' in filenames
1761 l = fp.readline()
1764 l = fp.readline()
1762 try:
1765 try:
1763 name, size = l.split('\0', 1)
1766 name, size = l.split('\0', 1)
1764 size = int(size)
1767 size = int(size)
1765 except (ValueError, TypeError):
1768 except (ValueError, TypeError):
1766 raise error.ResponseError(
1769 raise error.ResponseError(
1767 _('unexpected response from remote server:'), l)
1770 _('unexpected response from remote server:'), l)
1768 if self.ui.debugflag:
1771 if self.ui.debugflag:
1769 self.ui.debug('adding %s (%s)\n' %
1772 self.ui.debug('adding %s (%s)\n' %
1770 (name, util.bytecount(size)))
1773 (name, util.bytecount(size)))
1771 # for backwards compat, name was partially encoded
1774 # for backwards compat, name was partially encoded
1772 ofp = self.svfs(store.decodedir(name), 'w')
1775 ofp = self.svfs(store.decodedir(name), 'w')
1773 for chunk in util.filechunkiter(fp, limit=size):
1776 for chunk in util.filechunkiter(fp, limit=size):
1774 handled_bytes += len(chunk)
1777 handled_bytes += len(chunk)
1775 self.ui.progress(_('clone'), handled_bytes,
1778 self.ui.progress(_('clone'), handled_bytes,
1776 total=total_bytes)
1779 total=total_bytes)
1777 ofp.write(chunk)
1780 ofp.write(chunk)
1778 ofp.close()
1781 ofp.close()
1779 tr.close()
1782 tr.close()
1780 finally:
1783 finally:
1781 tr.release()
1784 tr.release()
1782
1785
1783 # Writing straight to files circumvented the inmemory caches
1786 # Writing straight to files circumvented the inmemory caches
1784 self.invalidate()
1787 self.invalidate()
1785
1788
1786 elapsed = time.time() - start
1789 elapsed = time.time() - start
1787 if elapsed <= 0:
1790 if elapsed <= 0:
1788 elapsed = 0.001
1791 elapsed = 0.001
1789 self.ui.progress(_('clone'), None)
1792 self.ui.progress(_('clone'), None)
1790 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1793 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1791 (util.bytecount(total_bytes), elapsed,
1794 (util.bytecount(total_bytes), elapsed,
1792 util.bytecount(total_bytes / elapsed)))
1795 util.bytecount(total_bytes / elapsed)))
1793
1796
1794 # new requirements = old non-format requirements +
1797 # new requirements = old non-format requirements +
1795 # new format-related
1798 # new format-related
1796 # requirements from the streamed-in repository
1799 # requirements from the streamed-in repository
1797 requirements.update(set(self.requirements) - self.supportedformats)
1800 requirements.update(set(self.requirements) - self.supportedformats)
1798 self._applyrequirements(requirements)
1801 self._applyrequirements(requirements)
1799 self._writerequirements()
1802 self._writerequirements()
1800
1803
1801 if rbranchmap:
1804 if rbranchmap:
1802 rbheads = []
1805 rbheads = []
1803 closed = []
1806 closed = []
1804 for bheads in rbranchmap.itervalues():
1807 for bheads in rbranchmap.itervalues():
1805 rbheads.extend(bheads)
1808 rbheads.extend(bheads)
1806 for h in bheads:
1809 for h in bheads:
1807 r = self.changelog.rev(h)
1810 r = self.changelog.rev(h)
1808 b, c = self.changelog.branchinfo(r)
1811 b, c = self.changelog.branchinfo(r)
1809 if c:
1812 if c:
1810 closed.append(h)
1813 closed.append(h)
1811
1814
1812 if rbheads:
1815 if rbheads:
1813 rtiprev = max((int(self.changelog.rev(node))
1816 rtiprev = max((int(self.changelog.rev(node))
1814 for node in rbheads))
1817 for node in rbheads))
1815 cache = branchmap.branchcache(rbranchmap,
1818 cache = branchmap.branchcache(rbranchmap,
1816 self[rtiprev].node(),
1819 self[rtiprev].node(),
1817 rtiprev,
1820 rtiprev,
1818 closednodes=closed)
1821 closednodes=closed)
1819 # Try to stick it as low as possible
1822 # Try to stick it as low as possible
1820 # filter above served are unlikely to be fetch from a clone
1823 # filter above served are unlikely to be fetch from a clone
1821 for candidate in ('base', 'immutable', 'served'):
1824 for candidate in ('base', 'immutable', 'served'):
1822 rview = self.filtered(candidate)
1825 rview = self.filtered(candidate)
1823 if cache.validfor(rview):
1826 if cache.validfor(rview):
1824 self._branchcaches[candidate] = cache
1827 self._branchcaches[candidate] = cache
1825 cache.write(rview)
1828 cache.write(rview)
1826 break
1829 break
1827 self.invalidate()
1830 self.invalidate()
1828 return len(self.heads()) + 1
1831 return len(self.heads()) + 1
1829 finally:
1832 finally:
1830 lock.release()
1833 lock.release()
1831
1834
1832 def clone(self, remote, heads=[], stream=None):
1835 def clone(self, remote, heads=[], stream=None):
1833 '''clone remote repository.
1836 '''clone remote repository.
1834
1837
1835 keyword arguments:
1838 keyword arguments:
1836 heads: list of revs to clone (forces use of pull)
1839 heads: list of revs to clone (forces use of pull)
1837 stream: use streaming clone if possible'''
1840 stream: use streaming clone if possible'''
1838
1841
1839 # now, all clients that can request uncompressed clones can
1842 # now, all clients that can request uncompressed clones can
1840 # read repo formats supported by all servers that can serve
1843 # read repo formats supported by all servers that can serve
1841 # them.
1844 # them.
1842
1845
1843 # if revlog format changes, client will have to check version
1846 # if revlog format changes, client will have to check version
1844 # and format flags on "stream" capability, and use
1847 # and format flags on "stream" capability, and use
1845 # uncompressed only if compatible.
1848 # uncompressed only if compatible.
1846
1849
1847 if stream is None:
1850 if stream is None:
1848 # if the server explicitly prefers to stream (for fast LANs)
1851 # if the server explicitly prefers to stream (for fast LANs)
1849 stream = remote.capable('stream-preferred')
1852 stream = remote.capable('stream-preferred')
1850
1853
1851 if stream and not heads:
1854 if stream and not heads:
1852 # 'stream' means remote revlog format is revlogv1 only
1855 # 'stream' means remote revlog format is revlogv1 only
1853 if remote.capable('stream'):
1856 if remote.capable('stream'):
1854 self.stream_in(remote, set(('revlogv1',)))
1857 self.stream_in(remote, set(('revlogv1',)))
1855 else:
1858 else:
1856 # otherwise, 'streamreqs' contains the remote revlog format
1859 # otherwise, 'streamreqs' contains the remote revlog format
1857 streamreqs = remote.capable('streamreqs')
1860 streamreqs = remote.capable('streamreqs')
1858 if streamreqs:
1861 if streamreqs:
1859 streamreqs = set(streamreqs.split(','))
1862 streamreqs = set(streamreqs.split(','))
1860 # if we support it, stream in and adjust our requirements
1863 # if we support it, stream in and adjust our requirements
1861 if not streamreqs - self.supportedformats:
1864 if not streamreqs - self.supportedformats:
1862 self.stream_in(remote, streamreqs)
1865 self.stream_in(remote, streamreqs)
1863
1866
1864 quiet = self.ui.backupconfig('ui', 'quietbookmarkmove')
1867 quiet = self.ui.backupconfig('ui', 'quietbookmarkmove')
1865 try:
1868 try:
1866 self.ui.setconfig('ui', 'quietbookmarkmove', True, 'clone')
1869 self.ui.setconfig('ui', 'quietbookmarkmove', True, 'clone')
1867 ret = exchange.pull(self, remote, heads).cgresult
1870 ret = exchange.pull(self, remote, heads).cgresult
1868 finally:
1871 finally:
1869 self.ui.restoreconfig(quiet)
1872 self.ui.restoreconfig(quiet)
1870 return ret
1873 return ret
1871
1874
1872 def pushkey(self, namespace, key, old, new):
1875 def pushkey(self, namespace, key, old, new):
1873 try:
1876 try:
1874 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1877 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1875 old=old, new=new)
1878 old=old, new=new)
1876 except error.HookAbort, exc:
1879 except error.HookAbort, exc:
1877 self.ui.write_err(_("pushkey-abort: %s\n") % exc)
1880 self.ui.write_err(_("pushkey-abort: %s\n") % exc)
1878 if exc.hint:
1881 if exc.hint:
1879 self.ui.write_err(_("(%s)\n") % exc.hint)
1882 self.ui.write_err(_("(%s)\n") % exc.hint)
1880 return False
1883 return False
1881 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1884 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1882 ret = pushkey.push(self, namespace, key, old, new)
1885 ret = pushkey.push(self, namespace, key, old, new)
1883 def runhook():
1886 def runhook():
1884 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1887 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1885 ret=ret)
1888 ret=ret)
1886 self._afterlock(runhook)
1889 self._afterlock(runhook)
1887 return ret
1890 return ret
1888
1891
1889 def listkeys(self, namespace):
1892 def listkeys(self, namespace):
1890 self.hook('prelistkeys', throw=True, namespace=namespace)
1893 self.hook('prelistkeys', throw=True, namespace=namespace)
1891 self.ui.debug('listing keys for "%s"\n' % namespace)
1894 self.ui.debug('listing keys for "%s"\n' % namespace)
1892 values = pushkey.list(self, namespace)
1895 values = pushkey.list(self, namespace)
1893 self.hook('listkeys', namespace=namespace, values=values)
1896 self.hook('listkeys', namespace=namespace, values=values)
1894 return values
1897 return values
1895
1898
1896 def debugwireargs(self, one, two, three=None, four=None, five=None):
1899 def debugwireargs(self, one, two, three=None, four=None, five=None):
1897 '''used to test argument passing over the wire'''
1900 '''used to test argument passing over the wire'''
1898 return "%s %s %s %s %s" % (one, two, three, four, five)
1901 return "%s %s %s %s %s" % (one, two, three, four, five)
1899
1902
1900 def savecommitmessage(self, text):
1903 def savecommitmessage(self, text):
1901 fp = self.vfs('last-message.txt', 'wb')
1904 fp = self.vfs('last-message.txt', 'wb')
1902 try:
1905 try:
1903 fp.write(text)
1906 fp.write(text)
1904 finally:
1907 finally:
1905 fp.close()
1908 fp.close()
1906 return self.pathto(fp.name[len(self.root) + 1:])
1909 return self.pathto(fp.name[len(self.root) + 1:])
1907
1910
1908 # used to avoid circular references so destructors work
1911 # used to avoid circular references so destructors work
1909 def aftertrans(files):
1912 def aftertrans(files):
1910 renamefiles = [tuple(t) for t in files]
1913 renamefiles = [tuple(t) for t in files]
1911 def a():
1914 def a():
1912 for vfs, src, dest in renamefiles:
1915 for vfs, src, dest in renamefiles:
1913 try:
1916 try:
1914 vfs.rename(src, dest)
1917 vfs.rename(src, dest)
1915 except OSError: # journal file does not yet exist
1918 except OSError: # journal file does not yet exist
1916 pass
1919 pass
1917 return a
1920 return a
1918
1921
1919 def undoname(fn):
1922 def undoname(fn):
1920 base, name = os.path.split(fn)
1923 base, name = os.path.split(fn)
1921 assert name.startswith('journal')
1924 assert name.startswith('journal')
1922 return os.path.join(base, name.replace('journal', 'undo', 1))
1925 return os.path.join(base, name.replace('journal', 'undo', 1))
1923
1926
1924 def instance(ui, path, create):
1927 def instance(ui, path, create):
1925 return localrepository(ui, util.urllocalpath(path), create)
1928 return localrepository(ui, util.urllocalpath(path), create)
1926
1929
1927 def islocal(path):
1930 def islocal(path):
1928 return True
1931 return True
General Comments 0
You need to be logged in to leave comments. Login now