##// END OF EJS Templates
obsolete: mark warning for translation...
Matt Mackall -
r24414:f247fbfe default
parent child Browse files
Show More
@@ -1,1923 +1,1923 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import urllib
9 import urllib
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
12 import lock as lockmod
12 import lock as lockmod
13 import transaction, store, encoding, exchange, bundle2
13 import transaction, store, encoding, exchange, bundle2
14 import scmutil, util, extensions, hook, error, revset
14 import scmutil, util, extensions, hook, error, revset
15 import match as matchmod
15 import match as matchmod
16 import merge as mergemod
16 import merge as mergemod
17 import tags as tagsmod
17 import tags as tagsmod
18 from lock import release
18 from lock import release
19 import weakref, errno, os, time, inspect
19 import weakref, errno, os, time, inspect
20 import branchmap, pathutil
20 import branchmap, pathutil
21 import namespaces
21 import namespaces
22 propertycache = util.propertycache
22 propertycache = util.propertycache
23 filecache = scmutil.filecache
23 filecache = scmutil.filecache
24
24
25 class repofilecache(filecache):
25 class repofilecache(filecache):
26 """All filecache usage on repo are done for logic that should be unfiltered
26 """All filecache usage on repo are done for logic that should be unfiltered
27 """
27 """
28
28
29 def __get__(self, repo, type=None):
29 def __get__(self, repo, type=None):
30 return super(repofilecache, self).__get__(repo.unfiltered(), type)
30 return super(repofilecache, self).__get__(repo.unfiltered(), type)
31 def __set__(self, repo, value):
31 def __set__(self, repo, value):
32 return super(repofilecache, self).__set__(repo.unfiltered(), value)
32 return super(repofilecache, self).__set__(repo.unfiltered(), value)
33 def __delete__(self, repo):
33 def __delete__(self, repo):
34 return super(repofilecache, self).__delete__(repo.unfiltered())
34 return super(repofilecache, self).__delete__(repo.unfiltered())
35
35
36 class storecache(repofilecache):
36 class storecache(repofilecache):
37 """filecache for files in the store"""
37 """filecache for files in the store"""
38 def join(self, obj, fname):
38 def join(self, obj, fname):
39 return obj.sjoin(fname)
39 return obj.sjoin(fname)
40
40
41 class unfilteredpropertycache(propertycache):
41 class unfilteredpropertycache(propertycache):
42 """propertycache that apply to unfiltered repo only"""
42 """propertycache that apply to unfiltered repo only"""
43
43
44 def __get__(self, repo, type=None):
44 def __get__(self, repo, type=None):
45 unfi = repo.unfiltered()
45 unfi = repo.unfiltered()
46 if unfi is repo:
46 if unfi is repo:
47 return super(unfilteredpropertycache, self).__get__(unfi)
47 return super(unfilteredpropertycache, self).__get__(unfi)
48 return getattr(unfi, self.name)
48 return getattr(unfi, self.name)
49
49
50 class filteredpropertycache(propertycache):
50 class filteredpropertycache(propertycache):
51 """propertycache that must take filtering in account"""
51 """propertycache that must take filtering in account"""
52
52
53 def cachevalue(self, obj, value):
53 def cachevalue(self, obj, value):
54 object.__setattr__(obj, self.name, value)
54 object.__setattr__(obj, self.name, value)
55
55
56
56
57 def hasunfilteredcache(repo, name):
57 def hasunfilteredcache(repo, name):
58 """check if a repo has an unfilteredpropertycache value for <name>"""
58 """check if a repo has an unfilteredpropertycache value for <name>"""
59 return name in vars(repo.unfiltered())
59 return name in vars(repo.unfiltered())
60
60
61 def unfilteredmethod(orig):
61 def unfilteredmethod(orig):
62 """decorate method that always need to be run on unfiltered version"""
62 """decorate method that always need to be run on unfiltered version"""
63 def wrapper(repo, *args, **kwargs):
63 def wrapper(repo, *args, **kwargs):
64 return orig(repo.unfiltered(), *args, **kwargs)
64 return orig(repo.unfiltered(), *args, **kwargs)
65 return wrapper
65 return wrapper
66
66
67 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
67 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
68 'unbundle'))
68 'unbundle'))
69 legacycaps = moderncaps.union(set(['changegroupsubset']))
69 legacycaps = moderncaps.union(set(['changegroupsubset']))
70
70
71 class localpeer(peer.peerrepository):
71 class localpeer(peer.peerrepository):
72 '''peer for a local repo; reflects only the most recent API'''
72 '''peer for a local repo; reflects only the most recent API'''
73
73
74 def __init__(self, repo, caps=moderncaps):
74 def __init__(self, repo, caps=moderncaps):
75 peer.peerrepository.__init__(self)
75 peer.peerrepository.__init__(self)
76 self._repo = repo.filtered('served')
76 self._repo = repo.filtered('served')
77 self.ui = repo.ui
77 self.ui = repo.ui
78 self._caps = repo._restrictcapabilities(caps)
78 self._caps = repo._restrictcapabilities(caps)
79 self.requirements = repo.requirements
79 self.requirements = repo.requirements
80 self.supportedformats = repo.supportedformats
80 self.supportedformats = repo.supportedformats
81
81
82 def close(self):
82 def close(self):
83 self._repo.close()
83 self._repo.close()
84
84
85 def _capabilities(self):
85 def _capabilities(self):
86 return self._caps
86 return self._caps
87
87
88 def local(self):
88 def local(self):
89 return self._repo
89 return self._repo
90
90
91 def canpush(self):
91 def canpush(self):
92 return True
92 return True
93
93
94 def url(self):
94 def url(self):
95 return self._repo.url()
95 return self._repo.url()
96
96
97 def lookup(self, key):
97 def lookup(self, key):
98 return self._repo.lookup(key)
98 return self._repo.lookup(key)
99
99
100 def branchmap(self):
100 def branchmap(self):
101 return self._repo.branchmap()
101 return self._repo.branchmap()
102
102
103 def heads(self):
103 def heads(self):
104 return self._repo.heads()
104 return self._repo.heads()
105
105
106 def known(self, nodes):
106 def known(self, nodes):
107 return self._repo.known(nodes)
107 return self._repo.known(nodes)
108
108
109 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
109 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
110 format='HG10', **kwargs):
110 format='HG10', **kwargs):
111 cg = exchange.getbundle(self._repo, source, heads=heads,
111 cg = exchange.getbundle(self._repo, source, heads=heads,
112 common=common, bundlecaps=bundlecaps, **kwargs)
112 common=common, bundlecaps=bundlecaps, **kwargs)
113 if bundlecaps is not None and 'HG2Y' in bundlecaps:
113 if bundlecaps is not None and 'HG2Y' in bundlecaps:
114 # When requesting a bundle2, getbundle returns a stream to make the
114 # When requesting a bundle2, getbundle returns a stream to make the
115 # wire level function happier. We need to build a proper object
115 # wire level function happier. We need to build a proper object
116 # from it in local peer.
116 # from it in local peer.
117 cg = bundle2.unbundle20(self.ui, cg)
117 cg = bundle2.unbundle20(self.ui, cg)
118 return cg
118 return cg
119
119
120 # TODO We might want to move the next two calls into legacypeer and add
120 # TODO We might want to move the next two calls into legacypeer and add
121 # unbundle instead.
121 # unbundle instead.
122
122
123 def unbundle(self, cg, heads, url):
123 def unbundle(self, cg, heads, url):
124 """apply a bundle on a repo
124 """apply a bundle on a repo
125
125
126 This function handles the repo locking itself."""
126 This function handles the repo locking itself."""
127 try:
127 try:
128 cg = exchange.readbundle(self.ui, cg, None)
128 cg = exchange.readbundle(self.ui, cg, None)
129 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
129 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
130 if util.safehasattr(ret, 'getchunks'):
130 if util.safehasattr(ret, 'getchunks'):
131 # This is a bundle20 object, turn it into an unbundler.
131 # This is a bundle20 object, turn it into an unbundler.
132 # This little dance should be dropped eventually when the API
132 # This little dance should be dropped eventually when the API
133 # is finally improved.
133 # is finally improved.
134 stream = util.chunkbuffer(ret.getchunks())
134 stream = util.chunkbuffer(ret.getchunks())
135 ret = bundle2.unbundle20(self.ui, stream)
135 ret = bundle2.unbundle20(self.ui, stream)
136 return ret
136 return ret
137 except error.PushRaced, exc:
137 except error.PushRaced, exc:
138 raise error.ResponseError(_('push failed:'), str(exc))
138 raise error.ResponseError(_('push failed:'), str(exc))
139
139
140 def lock(self):
140 def lock(self):
141 return self._repo.lock()
141 return self._repo.lock()
142
142
143 def addchangegroup(self, cg, source, url):
143 def addchangegroup(self, cg, source, url):
144 return changegroup.addchangegroup(self._repo, cg, source, url)
144 return changegroup.addchangegroup(self._repo, cg, source, url)
145
145
146 def pushkey(self, namespace, key, old, new):
146 def pushkey(self, namespace, key, old, new):
147 return self._repo.pushkey(namespace, key, old, new)
147 return self._repo.pushkey(namespace, key, old, new)
148
148
149 def listkeys(self, namespace):
149 def listkeys(self, namespace):
150 return self._repo.listkeys(namespace)
150 return self._repo.listkeys(namespace)
151
151
152 def debugwireargs(self, one, two, three=None, four=None, five=None):
152 def debugwireargs(self, one, two, three=None, four=None, five=None):
153 '''used to test argument passing over the wire'''
153 '''used to test argument passing over the wire'''
154 return "%s %s %s %s %s" % (one, two, three, four, five)
154 return "%s %s %s %s %s" % (one, two, three, four, five)
155
155
156 class locallegacypeer(localpeer):
156 class locallegacypeer(localpeer):
157 '''peer extension which implements legacy methods too; used for tests with
157 '''peer extension which implements legacy methods too; used for tests with
158 restricted capabilities'''
158 restricted capabilities'''
159
159
160 def __init__(self, repo):
160 def __init__(self, repo):
161 localpeer.__init__(self, repo, caps=legacycaps)
161 localpeer.__init__(self, repo, caps=legacycaps)
162
162
163 def branches(self, nodes):
163 def branches(self, nodes):
164 return self._repo.branches(nodes)
164 return self._repo.branches(nodes)
165
165
166 def between(self, pairs):
166 def between(self, pairs):
167 return self._repo.between(pairs)
167 return self._repo.between(pairs)
168
168
169 def changegroup(self, basenodes, source):
169 def changegroup(self, basenodes, source):
170 return changegroup.changegroup(self._repo, basenodes, source)
170 return changegroup.changegroup(self._repo, basenodes, source)
171
171
172 def changegroupsubset(self, bases, heads, source):
172 def changegroupsubset(self, bases, heads, source):
173 return changegroup.changegroupsubset(self._repo, bases, heads, source)
173 return changegroup.changegroupsubset(self._repo, bases, heads, source)
174
174
175 class localrepository(object):
175 class localrepository(object):
176
176
177 supportedformats = set(('revlogv1', 'generaldelta'))
177 supportedformats = set(('revlogv1', 'generaldelta'))
178 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
178 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
179 'dotencode'))
179 'dotencode'))
180 openerreqs = set(('revlogv1', 'generaldelta'))
180 openerreqs = set(('revlogv1', 'generaldelta'))
181 requirements = ['revlogv1']
181 requirements = ['revlogv1']
182 filtername = None
182 filtername = None
183
183
184 # a list of (ui, featureset) functions.
184 # a list of (ui, featureset) functions.
185 # only functions defined in module of enabled extensions are invoked
185 # only functions defined in module of enabled extensions are invoked
186 featuresetupfuncs = set()
186 featuresetupfuncs = set()
187
187
188 def _baserequirements(self, create):
188 def _baserequirements(self, create):
189 return self.requirements[:]
189 return self.requirements[:]
190
190
191 def __init__(self, baseui, path=None, create=False):
191 def __init__(self, baseui, path=None, create=False):
192 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
192 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
193 self.wopener = self.wvfs
193 self.wopener = self.wvfs
194 self.root = self.wvfs.base
194 self.root = self.wvfs.base
195 self.path = self.wvfs.join(".hg")
195 self.path = self.wvfs.join(".hg")
196 self.origroot = path
196 self.origroot = path
197 self.auditor = pathutil.pathauditor(self.root, self._checknested)
197 self.auditor = pathutil.pathauditor(self.root, self._checknested)
198 self.vfs = scmutil.vfs(self.path)
198 self.vfs = scmutil.vfs(self.path)
199 self.opener = self.vfs
199 self.opener = self.vfs
200 self.baseui = baseui
200 self.baseui = baseui
201 self.ui = baseui.copy()
201 self.ui = baseui.copy()
202 self.ui.copy = baseui.copy # prevent copying repo configuration
202 self.ui.copy = baseui.copy # prevent copying repo configuration
203 # A list of callback to shape the phase if no data were found.
203 # A list of callback to shape the phase if no data were found.
204 # Callback are in the form: func(repo, roots) --> processed root.
204 # Callback are in the form: func(repo, roots) --> processed root.
205 # This list it to be filled by extension during repo setup
205 # This list it to be filled by extension during repo setup
206 self._phasedefaults = []
206 self._phasedefaults = []
207 try:
207 try:
208 self.ui.readconfig(self.join("hgrc"), self.root)
208 self.ui.readconfig(self.join("hgrc"), self.root)
209 extensions.loadall(self.ui)
209 extensions.loadall(self.ui)
210 except IOError:
210 except IOError:
211 pass
211 pass
212
212
213 if self.featuresetupfuncs:
213 if self.featuresetupfuncs:
214 self.supported = set(self._basesupported) # use private copy
214 self.supported = set(self._basesupported) # use private copy
215 extmods = set(m.__name__ for n, m
215 extmods = set(m.__name__ for n, m
216 in extensions.extensions(self.ui))
216 in extensions.extensions(self.ui))
217 for setupfunc in self.featuresetupfuncs:
217 for setupfunc in self.featuresetupfuncs:
218 if setupfunc.__module__ in extmods:
218 if setupfunc.__module__ in extmods:
219 setupfunc(self.ui, self.supported)
219 setupfunc(self.ui, self.supported)
220 else:
220 else:
221 self.supported = self._basesupported
221 self.supported = self._basesupported
222
222
223 if not self.vfs.isdir():
223 if not self.vfs.isdir():
224 if create:
224 if create:
225 if not self.wvfs.exists():
225 if not self.wvfs.exists():
226 self.wvfs.makedirs()
226 self.wvfs.makedirs()
227 self.vfs.makedir(notindexed=True)
227 self.vfs.makedir(notindexed=True)
228 requirements = self._baserequirements(create)
228 requirements = self._baserequirements(create)
229 if self.ui.configbool('format', 'usestore', True):
229 if self.ui.configbool('format', 'usestore', True):
230 self.vfs.mkdir("store")
230 self.vfs.mkdir("store")
231 requirements.append("store")
231 requirements.append("store")
232 if self.ui.configbool('format', 'usefncache', True):
232 if self.ui.configbool('format', 'usefncache', True):
233 requirements.append("fncache")
233 requirements.append("fncache")
234 if self.ui.configbool('format', 'dotencode', True):
234 if self.ui.configbool('format', 'dotencode', True):
235 requirements.append('dotencode')
235 requirements.append('dotencode')
236 # create an invalid changelog
236 # create an invalid changelog
237 self.vfs.append(
237 self.vfs.append(
238 "00changelog.i",
238 "00changelog.i",
239 '\0\0\0\2' # represents revlogv2
239 '\0\0\0\2' # represents revlogv2
240 ' dummy changelog to prevent using the old repo layout'
240 ' dummy changelog to prevent using the old repo layout'
241 )
241 )
242 if self.ui.configbool('format', 'generaldelta', False):
242 if self.ui.configbool('format', 'generaldelta', False):
243 requirements.append("generaldelta")
243 requirements.append("generaldelta")
244 requirements = set(requirements)
244 requirements = set(requirements)
245 else:
245 else:
246 raise error.RepoError(_("repository %s not found") % path)
246 raise error.RepoError(_("repository %s not found") % path)
247 elif create:
247 elif create:
248 raise error.RepoError(_("repository %s already exists") % path)
248 raise error.RepoError(_("repository %s already exists") % path)
249 else:
249 else:
250 try:
250 try:
251 requirements = scmutil.readrequires(self.vfs, self.supported)
251 requirements = scmutil.readrequires(self.vfs, self.supported)
252 except IOError, inst:
252 except IOError, inst:
253 if inst.errno != errno.ENOENT:
253 if inst.errno != errno.ENOENT:
254 raise
254 raise
255 requirements = set()
255 requirements = set()
256
256
257 self.sharedpath = self.path
257 self.sharedpath = self.path
258 try:
258 try:
259 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
259 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
260 realpath=True)
260 realpath=True)
261 s = vfs.base
261 s = vfs.base
262 if not vfs.exists():
262 if not vfs.exists():
263 raise error.RepoError(
263 raise error.RepoError(
264 _('.hg/sharedpath points to nonexistent directory %s') % s)
264 _('.hg/sharedpath points to nonexistent directory %s') % s)
265 self.sharedpath = s
265 self.sharedpath = s
266 except IOError, inst:
266 except IOError, inst:
267 if inst.errno != errno.ENOENT:
267 if inst.errno != errno.ENOENT:
268 raise
268 raise
269
269
270 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
270 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
271 self.spath = self.store.path
271 self.spath = self.store.path
272 self.svfs = self.store.vfs
272 self.svfs = self.store.vfs
273 self.sopener = self.svfs
273 self.sopener = self.svfs
274 self.sjoin = self.store.join
274 self.sjoin = self.store.join
275 self.vfs.createmode = self.store.createmode
275 self.vfs.createmode = self.store.createmode
276 self._applyrequirements(requirements)
276 self._applyrequirements(requirements)
277 if create:
277 if create:
278 self._writerequirements()
278 self._writerequirements()
279
279
280
280
281 self._branchcaches = {}
281 self._branchcaches = {}
282 self._revbranchcache = None
282 self._revbranchcache = None
283 self.filterpats = {}
283 self.filterpats = {}
284 self._datafilters = {}
284 self._datafilters = {}
285 self._transref = self._lockref = self._wlockref = None
285 self._transref = self._lockref = self._wlockref = None
286
286
287 # A cache for various files under .hg/ that tracks file changes,
287 # A cache for various files under .hg/ that tracks file changes,
288 # (used by the filecache decorator)
288 # (used by the filecache decorator)
289 #
289 #
290 # Maps a property name to its util.filecacheentry
290 # Maps a property name to its util.filecacheentry
291 self._filecache = {}
291 self._filecache = {}
292
292
293 # hold sets of revision to be filtered
293 # hold sets of revision to be filtered
294 # should be cleared when something might have changed the filter value:
294 # should be cleared when something might have changed the filter value:
295 # - new changesets,
295 # - new changesets,
296 # - phase change,
296 # - phase change,
297 # - new obsolescence marker,
297 # - new obsolescence marker,
298 # - working directory parent change,
298 # - working directory parent change,
299 # - bookmark changes
299 # - bookmark changes
300 self.filteredrevcache = {}
300 self.filteredrevcache = {}
301
301
302 # generic mapping between names and nodes
302 # generic mapping between names and nodes
303 self.names = namespaces.namespaces()
303 self.names = namespaces.namespaces()
304
304
305 def close(self):
305 def close(self):
306 self._writecaches()
306 self._writecaches()
307
307
308 def _writecaches(self):
308 def _writecaches(self):
309 if self._revbranchcache:
309 if self._revbranchcache:
310 self._revbranchcache.write()
310 self._revbranchcache.write()
311
311
312 def _restrictcapabilities(self, caps):
312 def _restrictcapabilities(self, caps):
313 # bundle2 is not ready for prime time, drop it unless explicitly
313 # bundle2 is not ready for prime time, drop it unless explicitly
314 # required by the tests (or some brave tester)
314 # required by the tests (or some brave tester)
315 if self.ui.configbool('experimental', 'bundle2-exp', False):
315 if self.ui.configbool('experimental', 'bundle2-exp', False):
316 caps = set(caps)
316 caps = set(caps)
317 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
317 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
318 caps.add('bundle2-exp=' + urllib.quote(capsblob))
318 caps.add('bundle2-exp=' + urllib.quote(capsblob))
319 return caps
319 return caps
320
320
321 def _applyrequirements(self, requirements):
321 def _applyrequirements(self, requirements):
322 self.requirements = requirements
322 self.requirements = requirements
323 self.svfs.options = dict((r, 1) for r in requirements
323 self.svfs.options = dict((r, 1) for r in requirements
324 if r in self.openerreqs)
324 if r in self.openerreqs)
325 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
325 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
326 if chunkcachesize is not None:
326 if chunkcachesize is not None:
327 self.svfs.options['chunkcachesize'] = chunkcachesize
327 self.svfs.options['chunkcachesize'] = chunkcachesize
328 maxchainlen = self.ui.configint('format', 'maxchainlen')
328 maxchainlen = self.ui.configint('format', 'maxchainlen')
329 if maxchainlen is not None:
329 if maxchainlen is not None:
330 self.svfs.options['maxchainlen'] = maxchainlen
330 self.svfs.options['maxchainlen'] = maxchainlen
331 manifestcachesize = self.ui.configint('format', 'manifestcachesize')
331 manifestcachesize = self.ui.configint('format', 'manifestcachesize')
332 if manifestcachesize is not None:
332 if manifestcachesize is not None:
333 self.svfs.options['manifestcachesize'] = manifestcachesize
333 self.svfs.options['manifestcachesize'] = manifestcachesize
334 usetreemanifest = self.ui.configbool('experimental', 'treemanifest')
334 usetreemanifest = self.ui.configbool('experimental', 'treemanifest')
335 if usetreemanifest is not None:
335 if usetreemanifest is not None:
336 self.svfs.options['usetreemanifest'] = usetreemanifest
336 self.svfs.options['usetreemanifest'] = usetreemanifest
337
337
338 def _writerequirements(self):
338 def _writerequirements(self):
339 reqfile = self.vfs("requires", "w")
339 reqfile = self.vfs("requires", "w")
340 for r in sorted(self.requirements):
340 for r in sorted(self.requirements):
341 reqfile.write("%s\n" % r)
341 reqfile.write("%s\n" % r)
342 reqfile.close()
342 reqfile.close()
343
343
344 def _checknested(self, path):
344 def _checknested(self, path):
345 """Determine if path is a legal nested repository."""
345 """Determine if path is a legal nested repository."""
346 if not path.startswith(self.root):
346 if not path.startswith(self.root):
347 return False
347 return False
348 subpath = path[len(self.root) + 1:]
348 subpath = path[len(self.root) + 1:]
349 normsubpath = util.pconvert(subpath)
349 normsubpath = util.pconvert(subpath)
350
350
351 # XXX: Checking against the current working copy is wrong in
351 # XXX: Checking against the current working copy is wrong in
352 # the sense that it can reject things like
352 # the sense that it can reject things like
353 #
353 #
354 # $ hg cat -r 10 sub/x.txt
354 # $ hg cat -r 10 sub/x.txt
355 #
355 #
356 # if sub/ is no longer a subrepository in the working copy
356 # if sub/ is no longer a subrepository in the working copy
357 # parent revision.
357 # parent revision.
358 #
358 #
359 # However, it can of course also allow things that would have
359 # However, it can of course also allow things that would have
360 # been rejected before, such as the above cat command if sub/
360 # been rejected before, such as the above cat command if sub/
361 # is a subrepository now, but was a normal directory before.
361 # is a subrepository now, but was a normal directory before.
362 # The old path auditor would have rejected by mistake since it
362 # The old path auditor would have rejected by mistake since it
363 # panics when it sees sub/.hg/.
363 # panics when it sees sub/.hg/.
364 #
364 #
365 # All in all, checking against the working copy seems sensible
365 # All in all, checking against the working copy seems sensible
366 # since we want to prevent access to nested repositories on
366 # since we want to prevent access to nested repositories on
367 # the filesystem *now*.
367 # the filesystem *now*.
368 ctx = self[None]
368 ctx = self[None]
369 parts = util.splitpath(subpath)
369 parts = util.splitpath(subpath)
370 while parts:
370 while parts:
371 prefix = '/'.join(parts)
371 prefix = '/'.join(parts)
372 if prefix in ctx.substate:
372 if prefix in ctx.substate:
373 if prefix == normsubpath:
373 if prefix == normsubpath:
374 return True
374 return True
375 else:
375 else:
376 sub = ctx.sub(prefix)
376 sub = ctx.sub(prefix)
377 return sub.checknested(subpath[len(prefix) + 1:])
377 return sub.checknested(subpath[len(prefix) + 1:])
378 else:
378 else:
379 parts.pop()
379 parts.pop()
380 return False
380 return False
381
381
382 def peer(self):
382 def peer(self):
383 return localpeer(self) # not cached to avoid reference cycle
383 return localpeer(self) # not cached to avoid reference cycle
384
384
385 def unfiltered(self):
385 def unfiltered(self):
386 """Return unfiltered version of the repository
386 """Return unfiltered version of the repository
387
387
388 Intended to be overwritten by filtered repo."""
388 Intended to be overwritten by filtered repo."""
389 return self
389 return self
390
390
391 def filtered(self, name):
391 def filtered(self, name):
392 """Return a filtered version of a repository"""
392 """Return a filtered version of a repository"""
393 # build a new class with the mixin and the current class
393 # build a new class with the mixin and the current class
394 # (possibly subclass of the repo)
394 # (possibly subclass of the repo)
395 class proxycls(repoview.repoview, self.unfiltered().__class__):
395 class proxycls(repoview.repoview, self.unfiltered().__class__):
396 pass
396 pass
397 return proxycls(self, name)
397 return proxycls(self, name)
398
398
399 @repofilecache('bookmarks')
399 @repofilecache('bookmarks')
400 def _bookmarks(self):
400 def _bookmarks(self):
401 return bookmarks.bmstore(self)
401 return bookmarks.bmstore(self)
402
402
403 @repofilecache('bookmarks.current')
403 @repofilecache('bookmarks.current')
404 def _bookmarkcurrent(self):
404 def _bookmarkcurrent(self):
405 return bookmarks.readcurrent(self)
405 return bookmarks.readcurrent(self)
406
406
407 def bookmarkheads(self, bookmark):
407 def bookmarkheads(self, bookmark):
408 name = bookmark.split('@', 1)[0]
408 name = bookmark.split('@', 1)[0]
409 heads = []
409 heads = []
410 for mark, n in self._bookmarks.iteritems():
410 for mark, n in self._bookmarks.iteritems():
411 if mark.split('@', 1)[0] == name:
411 if mark.split('@', 1)[0] == name:
412 heads.append(n)
412 heads.append(n)
413 return heads
413 return heads
414
414
415 @storecache('phaseroots')
415 @storecache('phaseroots')
416 def _phasecache(self):
416 def _phasecache(self):
417 return phases.phasecache(self, self._phasedefaults)
417 return phases.phasecache(self, self._phasedefaults)
418
418
419 @storecache('obsstore')
419 @storecache('obsstore')
420 def obsstore(self):
420 def obsstore(self):
421 # read default format for new obsstore.
421 # read default format for new obsstore.
422 defaultformat = self.ui.configint('format', 'obsstore-version', None)
422 defaultformat = self.ui.configint('format', 'obsstore-version', None)
423 # rely on obsstore class default when possible.
423 # rely on obsstore class default when possible.
424 kwargs = {}
424 kwargs = {}
425 if defaultformat is not None:
425 if defaultformat is not None:
426 kwargs['defaultformat'] = defaultformat
426 kwargs['defaultformat'] = defaultformat
427 readonly = not obsolete.isenabled(self, obsolete.createmarkersopt)
427 readonly = not obsolete.isenabled(self, obsolete.createmarkersopt)
428 store = obsolete.obsstore(self.svfs, readonly=readonly,
428 store = obsolete.obsstore(self.svfs, readonly=readonly,
429 **kwargs)
429 **kwargs)
430 if store and readonly:
430 if store and readonly:
431 # message is rare enough to not be translated
431 self.ui.warn(
432 msg = 'obsolete feature not enabled but %i markers found!\n'
432 _('obsolete feature not enabled but %i markers found!\n')
433 self.ui.warn(msg % len(list(store)))
433 % len(list(store)))
434 return store
434 return store
435
435
436 @storecache('00changelog.i')
436 @storecache('00changelog.i')
437 def changelog(self):
437 def changelog(self):
438 c = changelog.changelog(self.svfs)
438 c = changelog.changelog(self.svfs)
439 if 'HG_PENDING' in os.environ:
439 if 'HG_PENDING' in os.environ:
440 p = os.environ['HG_PENDING']
440 p = os.environ['HG_PENDING']
441 if p.startswith(self.root):
441 if p.startswith(self.root):
442 c.readpending('00changelog.i.a')
442 c.readpending('00changelog.i.a')
443 return c
443 return c
444
444
445 @storecache('00manifest.i')
445 @storecache('00manifest.i')
446 def manifest(self):
446 def manifest(self):
447 return manifest.manifest(self.svfs)
447 return manifest.manifest(self.svfs)
448
448
449 @repofilecache('dirstate')
449 @repofilecache('dirstate')
450 def dirstate(self):
450 def dirstate(self):
451 warned = [0]
451 warned = [0]
452 def validate(node):
452 def validate(node):
453 try:
453 try:
454 self.changelog.rev(node)
454 self.changelog.rev(node)
455 return node
455 return node
456 except error.LookupError:
456 except error.LookupError:
457 if not warned[0]:
457 if not warned[0]:
458 warned[0] = True
458 warned[0] = True
459 self.ui.warn(_("warning: ignoring unknown"
459 self.ui.warn(_("warning: ignoring unknown"
460 " working parent %s!\n") % short(node))
460 " working parent %s!\n") % short(node))
461 return nullid
461 return nullid
462
462
463 return dirstate.dirstate(self.vfs, self.ui, self.root, validate)
463 return dirstate.dirstate(self.vfs, self.ui, self.root, validate)
464
464
465 def __getitem__(self, changeid):
465 def __getitem__(self, changeid):
466 if changeid is None:
466 if changeid is None:
467 return context.workingctx(self)
467 return context.workingctx(self)
468 if isinstance(changeid, slice):
468 if isinstance(changeid, slice):
469 return [context.changectx(self, i)
469 return [context.changectx(self, i)
470 for i in xrange(*changeid.indices(len(self)))
470 for i in xrange(*changeid.indices(len(self)))
471 if i not in self.changelog.filteredrevs]
471 if i not in self.changelog.filteredrevs]
472 return context.changectx(self, changeid)
472 return context.changectx(self, changeid)
473
473
474 def __contains__(self, changeid):
474 def __contains__(self, changeid):
475 try:
475 try:
476 self[changeid]
476 self[changeid]
477 return True
477 return True
478 except error.RepoLookupError:
478 except error.RepoLookupError:
479 return False
479 return False
480
480
481 def __nonzero__(self):
481 def __nonzero__(self):
482 return True
482 return True
483
483
484 def __len__(self):
484 def __len__(self):
485 return len(self.changelog)
485 return len(self.changelog)
486
486
487 def __iter__(self):
487 def __iter__(self):
488 return iter(self.changelog)
488 return iter(self.changelog)
489
489
490 def revs(self, expr, *args):
490 def revs(self, expr, *args):
491 '''Return a list of revisions matching the given revset'''
491 '''Return a list of revisions matching the given revset'''
492 expr = revset.formatspec(expr, *args)
492 expr = revset.formatspec(expr, *args)
493 m = revset.match(None, expr)
493 m = revset.match(None, expr)
494 return m(self)
494 return m(self)
495
495
496 def set(self, expr, *args):
496 def set(self, expr, *args):
497 '''
497 '''
498 Yield a context for each matching revision, after doing arg
498 Yield a context for each matching revision, after doing arg
499 replacement via revset.formatspec
499 replacement via revset.formatspec
500 '''
500 '''
501 for r in self.revs(expr, *args):
501 for r in self.revs(expr, *args):
502 yield self[r]
502 yield self[r]
503
503
504 def url(self):
504 def url(self):
505 return 'file:' + self.root
505 return 'file:' + self.root
506
506
507 def hook(self, name, throw=False, **args):
507 def hook(self, name, throw=False, **args):
508 """Call a hook, passing this repo instance.
508 """Call a hook, passing this repo instance.
509
509
510 This a convenience method to aid invoking hooks. Extensions likely
510 This a convenience method to aid invoking hooks. Extensions likely
511 won't call this unless they have registered a custom hook or are
511 won't call this unless they have registered a custom hook or are
512 replacing code that is expected to call a hook.
512 replacing code that is expected to call a hook.
513 """
513 """
514 return hook.hook(self.ui, self, name, throw, **args)
514 return hook.hook(self.ui, self, name, throw, **args)
515
515
516 @unfilteredmethod
516 @unfilteredmethod
517 def _tag(self, names, node, message, local, user, date, extra={},
517 def _tag(self, names, node, message, local, user, date, extra={},
518 editor=False):
518 editor=False):
519 if isinstance(names, str):
519 if isinstance(names, str):
520 names = (names,)
520 names = (names,)
521
521
522 branches = self.branchmap()
522 branches = self.branchmap()
523 for name in names:
523 for name in names:
524 self.hook('pretag', throw=True, node=hex(node), tag=name,
524 self.hook('pretag', throw=True, node=hex(node), tag=name,
525 local=local)
525 local=local)
526 if name in branches:
526 if name in branches:
527 self.ui.warn(_("warning: tag %s conflicts with existing"
527 self.ui.warn(_("warning: tag %s conflicts with existing"
528 " branch name\n") % name)
528 " branch name\n") % name)
529
529
530 def writetags(fp, names, munge, prevtags):
530 def writetags(fp, names, munge, prevtags):
531 fp.seek(0, 2)
531 fp.seek(0, 2)
532 if prevtags and prevtags[-1] != '\n':
532 if prevtags and prevtags[-1] != '\n':
533 fp.write('\n')
533 fp.write('\n')
534 for name in names:
534 for name in names:
535 if munge:
535 if munge:
536 m = munge(name)
536 m = munge(name)
537 else:
537 else:
538 m = name
538 m = name
539
539
540 if (self._tagscache.tagtypes and
540 if (self._tagscache.tagtypes and
541 name in self._tagscache.tagtypes):
541 name in self._tagscache.tagtypes):
542 old = self.tags().get(name, nullid)
542 old = self.tags().get(name, nullid)
543 fp.write('%s %s\n' % (hex(old), m))
543 fp.write('%s %s\n' % (hex(old), m))
544 fp.write('%s %s\n' % (hex(node), m))
544 fp.write('%s %s\n' % (hex(node), m))
545 fp.close()
545 fp.close()
546
546
547 prevtags = ''
547 prevtags = ''
548 if local:
548 if local:
549 try:
549 try:
550 fp = self.vfs('localtags', 'r+')
550 fp = self.vfs('localtags', 'r+')
551 except IOError:
551 except IOError:
552 fp = self.vfs('localtags', 'a')
552 fp = self.vfs('localtags', 'a')
553 else:
553 else:
554 prevtags = fp.read()
554 prevtags = fp.read()
555
555
556 # local tags are stored in the current charset
556 # local tags are stored in the current charset
557 writetags(fp, names, None, prevtags)
557 writetags(fp, names, None, prevtags)
558 for name in names:
558 for name in names:
559 self.hook('tag', node=hex(node), tag=name, local=local)
559 self.hook('tag', node=hex(node), tag=name, local=local)
560 return
560 return
561
561
562 try:
562 try:
563 fp = self.wfile('.hgtags', 'rb+')
563 fp = self.wfile('.hgtags', 'rb+')
564 except IOError, e:
564 except IOError, e:
565 if e.errno != errno.ENOENT:
565 if e.errno != errno.ENOENT:
566 raise
566 raise
567 fp = self.wfile('.hgtags', 'ab')
567 fp = self.wfile('.hgtags', 'ab')
568 else:
568 else:
569 prevtags = fp.read()
569 prevtags = fp.read()
570
570
571 # committed tags are stored in UTF-8
571 # committed tags are stored in UTF-8
572 writetags(fp, names, encoding.fromlocal, prevtags)
572 writetags(fp, names, encoding.fromlocal, prevtags)
573
573
574 fp.close()
574 fp.close()
575
575
576 self.invalidatecaches()
576 self.invalidatecaches()
577
577
578 if '.hgtags' not in self.dirstate:
578 if '.hgtags' not in self.dirstate:
579 self[None].add(['.hgtags'])
579 self[None].add(['.hgtags'])
580
580
581 m = matchmod.exact(self.root, '', ['.hgtags'])
581 m = matchmod.exact(self.root, '', ['.hgtags'])
582 tagnode = self.commit(message, user, date, extra=extra, match=m,
582 tagnode = self.commit(message, user, date, extra=extra, match=m,
583 editor=editor)
583 editor=editor)
584
584
585 for name in names:
585 for name in names:
586 self.hook('tag', node=hex(node), tag=name, local=local)
586 self.hook('tag', node=hex(node), tag=name, local=local)
587
587
588 return tagnode
588 return tagnode
589
589
590 def tag(self, names, node, message, local, user, date, editor=False):
590 def tag(self, names, node, message, local, user, date, editor=False):
591 '''tag a revision with one or more symbolic names.
591 '''tag a revision with one or more symbolic names.
592
592
593 names is a list of strings or, when adding a single tag, names may be a
593 names is a list of strings or, when adding a single tag, names may be a
594 string.
594 string.
595
595
596 if local is True, the tags are stored in a per-repository file.
596 if local is True, the tags are stored in a per-repository file.
597 otherwise, they are stored in the .hgtags file, and a new
597 otherwise, they are stored in the .hgtags file, and a new
598 changeset is committed with the change.
598 changeset is committed with the change.
599
599
600 keyword arguments:
600 keyword arguments:
601
601
602 local: whether to store tags in non-version-controlled file
602 local: whether to store tags in non-version-controlled file
603 (default False)
603 (default False)
604
604
605 message: commit message to use if committing
605 message: commit message to use if committing
606
606
607 user: name of user to use if committing
607 user: name of user to use if committing
608
608
609 date: date tuple to use if committing'''
609 date: date tuple to use if committing'''
610
610
611 if not local:
611 if not local:
612 m = matchmod.exact(self.root, '', ['.hgtags'])
612 m = matchmod.exact(self.root, '', ['.hgtags'])
613 if util.any(self.status(match=m, unknown=True, ignored=True)):
613 if util.any(self.status(match=m, unknown=True, ignored=True)):
614 raise util.Abort(_('working copy of .hgtags is changed'),
614 raise util.Abort(_('working copy of .hgtags is changed'),
615 hint=_('please commit .hgtags manually'))
615 hint=_('please commit .hgtags manually'))
616
616
617 self.tags() # instantiate the cache
617 self.tags() # instantiate the cache
618 self._tag(names, node, message, local, user, date, editor=editor)
618 self._tag(names, node, message, local, user, date, editor=editor)
619
619
620 @filteredpropertycache
620 @filteredpropertycache
621 def _tagscache(self):
621 def _tagscache(self):
622 '''Returns a tagscache object that contains various tags related
622 '''Returns a tagscache object that contains various tags related
623 caches.'''
623 caches.'''
624
624
625 # This simplifies its cache management by having one decorated
625 # This simplifies its cache management by having one decorated
626 # function (this one) and the rest simply fetch things from it.
626 # function (this one) and the rest simply fetch things from it.
627 class tagscache(object):
627 class tagscache(object):
628 def __init__(self):
628 def __init__(self):
629 # These two define the set of tags for this repository. tags
629 # These two define the set of tags for this repository. tags
630 # maps tag name to node; tagtypes maps tag name to 'global' or
630 # maps tag name to node; tagtypes maps tag name to 'global' or
631 # 'local'. (Global tags are defined by .hgtags across all
631 # 'local'. (Global tags are defined by .hgtags across all
632 # heads, and local tags are defined in .hg/localtags.)
632 # heads, and local tags are defined in .hg/localtags.)
633 # They constitute the in-memory cache of tags.
633 # They constitute the in-memory cache of tags.
634 self.tags = self.tagtypes = None
634 self.tags = self.tagtypes = None
635
635
636 self.nodetagscache = self.tagslist = None
636 self.nodetagscache = self.tagslist = None
637
637
638 cache = tagscache()
638 cache = tagscache()
639 cache.tags, cache.tagtypes = self._findtags()
639 cache.tags, cache.tagtypes = self._findtags()
640
640
641 return cache
641 return cache
642
642
643 def tags(self):
643 def tags(self):
644 '''return a mapping of tag to node'''
644 '''return a mapping of tag to node'''
645 t = {}
645 t = {}
646 if self.changelog.filteredrevs:
646 if self.changelog.filteredrevs:
647 tags, tt = self._findtags()
647 tags, tt = self._findtags()
648 else:
648 else:
649 tags = self._tagscache.tags
649 tags = self._tagscache.tags
650 for k, v in tags.iteritems():
650 for k, v in tags.iteritems():
651 try:
651 try:
652 # ignore tags to unknown nodes
652 # ignore tags to unknown nodes
653 self.changelog.rev(v)
653 self.changelog.rev(v)
654 t[k] = v
654 t[k] = v
655 except (error.LookupError, ValueError):
655 except (error.LookupError, ValueError):
656 pass
656 pass
657 return t
657 return t
658
658
659 def _findtags(self):
659 def _findtags(self):
660 '''Do the hard work of finding tags. Return a pair of dicts
660 '''Do the hard work of finding tags. Return a pair of dicts
661 (tags, tagtypes) where tags maps tag name to node, and tagtypes
661 (tags, tagtypes) where tags maps tag name to node, and tagtypes
662 maps tag name to a string like \'global\' or \'local\'.
662 maps tag name to a string like \'global\' or \'local\'.
663 Subclasses or extensions are free to add their own tags, but
663 Subclasses or extensions are free to add their own tags, but
664 should be aware that the returned dicts will be retained for the
664 should be aware that the returned dicts will be retained for the
665 duration of the localrepo object.'''
665 duration of the localrepo object.'''
666
666
667 # XXX what tagtype should subclasses/extensions use? Currently
667 # XXX what tagtype should subclasses/extensions use? Currently
668 # mq and bookmarks add tags, but do not set the tagtype at all.
668 # mq and bookmarks add tags, but do not set the tagtype at all.
669 # Should each extension invent its own tag type? Should there
669 # Should each extension invent its own tag type? Should there
670 # be one tagtype for all such "virtual" tags? Or is the status
670 # be one tagtype for all such "virtual" tags? Or is the status
671 # quo fine?
671 # quo fine?
672
672
673 alltags = {} # map tag name to (node, hist)
673 alltags = {} # map tag name to (node, hist)
674 tagtypes = {}
674 tagtypes = {}
675
675
676 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
676 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
677 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
677 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
678
678
679 # Build the return dicts. Have to re-encode tag names because
679 # Build the return dicts. Have to re-encode tag names because
680 # the tags module always uses UTF-8 (in order not to lose info
680 # the tags module always uses UTF-8 (in order not to lose info
681 # writing to the cache), but the rest of Mercurial wants them in
681 # writing to the cache), but the rest of Mercurial wants them in
682 # local encoding.
682 # local encoding.
683 tags = {}
683 tags = {}
684 for (name, (node, hist)) in alltags.iteritems():
684 for (name, (node, hist)) in alltags.iteritems():
685 if node != nullid:
685 if node != nullid:
686 tags[encoding.tolocal(name)] = node
686 tags[encoding.tolocal(name)] = node
687 tags['tip'] = self.changelog.tip()
687 tags['tip'] = self.changelog.tip()
688 tagtypes = dict([(encoding.tolocal(name), value)
688 tagtypes = dict([(encoding.tolocal(name), value)
689 for (name, value) in tagtypes.iteritems()])
689 for (name, value) in tagtypes.iteritems()])
690 return (tags, tagtypes)
690 return (tags, tagtypes)
691
691
692 def tagtype(self, tagname):
692 def tagtype(self, tagname):
693 '''
693 '''
694 return the type of the given tag. result can be:
694 return the type of the given tag. result can be:
695
695
696 'local' : a local tag
696 'local' : a local tag
697 'global' : a global tag
697 'global' : a global tag
698 None : tag does not exist
698 None : tag does not exist
699 '''
699 '''
700
700
701 return self._tagscache.tagtypes.get(tagname)
701 return self._tagscache.tagtypes.get(tagname)
702
702
703 def tagslist(self):
703 def tagslist(self):
704 '''return a list of tags ordered by revision'''
704 '''return a list of tags ordered by revision'''
705 if not self._tagscache.tagslist:
705 if not self._tagscache.tagslist:
706 l = []
706 l = []
707 for t, n in self.tags().iteritems():
707 for t, n in self.tags().iteritems():
708 l.append((self.changelog.rev(n), t, n))
708 l.append((self.changelog.rev(n), t, n))
709 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
709 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
710
710
711 return self._tagscache.tagslist
711 return self._tagscache.tagslist
712
712
713 def nodetags(self, node):
713 def nodetags(self, node):
714 '''return the tags associated with a node'''
714 '''return the tags associated with a node'''
715 if not self._tagscache.nodetagscache:
715 if not self._tagscache.nodetagscache:
716 nodetagscache = {}
716 nodetagscache = {}
717 for t, n in self._tagscache.tags.iteritems():
717 for t, n in self._tagscache.tags.iteritems():
718 nodetagscache.setdefault(n, []).append(t)
718 nodetagscache.setdefault(n, []).append(t)
719 for tags in nodetagscache.itervalues():
719 for tags in nodetagscache.itervalues():
720 tags.sort()
720 tags.sort()
721 self._tagscache.nodetagscache = nodetagscache
721 self._tagscache.nodetagscache = nodetagscache
722 return self._tagscache.nodetagscache.get(node, [])
722 return self._tagscache.nodetagscache.get(node, [])
723
723
724 def nodebookmarks(self, node):
724 def nodebookmarks(self, node):
725 marks = []
725 marks = []
726 for bookmark, n in self._bookmarks.iteritems():
726 for bookmark, n in self._bookmarks.iteritems():
727 if n == node:
727 if n == node:
728 marks.append(bookmark)
728 marks.append(bookmark)
729 return sorted(marks)
729 return sorted(marks)
730
730
731 def branchmap(self):
731 def branchmap(self):
732 '''returns a dictionary {branch: [branchheads]} with branchheads
732 '''returns a dictionary {branch: [branchheads]} with branchheads
733 ordered by increasing revision number'''
733 ordered by increasing revision number'''
734 branchmap.updatecache(self)
734 branchmap.updatecache(self)
735 return self._branchcaches[self.filtername]
735 return self._branchcaches[self.filtername]
736
736
737 @unfilteredmethod
737 @unfilteredmethod
738 def revbranchcache(self):
738 def revbranchcache(self):
739 if not self._revbranchcache:
739 if not self._revbranchcache:
740 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
740 self._revbranchcache = branchmap.revbranchcache(self.unfiltered())
741 return self._revbranchcache
741 return self._revbranchcache
742
742
743 def branchtip(self, branch, ignoremissing=False):
743 def branchtip(self, branch, ignoremissing=False):
744 '''return the tip node for a given branch
744 '''return the tip node for a given branch
745
745
746 If ignoremissing is True, then this method will not raise an error.
746 If ignoremissing is True, then this method will not raise an error.
747 This is helpful for callers that only expect None for a missing branch
747 This is helpful for callers that only expect None for a missing branch
748 (e.g. namespace).
748 (e.g. namespace).
749
749
750 '''
750 '''
751 try:
751 try:
752 return self.branchmap().branchtip(branch)
752 return self.branchmap().branchtip(branch)
753 except KeyError:
753 except KeyError:
754 if not ignoremissing:
754 if not ignoremissing:
755 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
755 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
756 else:
756 else:
757 pass
757 pass
758
758
759 def lookup(self, key):
759 def lookup(self, key):
760 return self[key].node()
760 return self[key].node()
761
761
762 def lookupbranch(self, key, remote=None):
762 def lookupbranch(self, key, remote=None):
763 repo = remote or self
763 repo = remote or self
764 if key in repo.branchmap():
764 if key in repo.branchmap():
765 return key
765 return key
766
766
767 repo = (remote and remote.local()) and remote or self
767 repo = (remote and remote.local()) and remote or self
768 return repo[key].branch()
768 return repo[key].branch()
769
769
770 def known(self, nodes):
770 def known(self, nodes):
771 nm = self.changelog.nodemap
771 nm = self.changelog.nodemap
772 pc = self._phasecache
772 pc = self._phasecache
773 result = []
773 result = []
774 for n in nodes:
774 for n in nodes:
775 r = nm.get(n)
775 r = nm.get(n)
776 resp = not (r is None or pc.phase(self, r) >= phases.secret)
776 resp = not (r is None or pc.phase(self, r) >= phases.secret)
777 result.append(resp)
777 result.append(resp)
778 return result
778 return result
779
779
780 def local(self):
780 def local(self):
781 return self
781 return self
782
782
783 def cancopy(self):
783 def cancopy(self):
784 # so statichttprepo's override of local() works
784 # so statichttprepo's override of local() works
785 if not self.local():
785 if not self.local():
786 return False
786 return False
787 if not self.ui.configbool('phases', 'publish', True):
787 if not self.ui.configbool('phases', 'publish', True):
788 return True
788 return True
789 # if publishing we can't copy if there is filtered content
789 # if publishing we can't copy if there is filtered content
790 return not self.filtered('visible').changelog.filteredrevs
790 return not self.filtered('visible').changelog.filteredrevs
791
791
792 def shared(self):
792 def shared(self):
793 '''the type of shared repository (None if not shared)'''
793 '''the type of shared repository (None if not shared)'''
794 if self.sharedpath != self.path:
794 if self.sharedpath != self.path:
795 return 'store'
795 return 'store'
796 return None
796 return None
797
797
798 def join(self, f, *insidef):
798 def join(self, f, *insidef):
799 return self.vfs.join(os.path.join(f, *insidef))
799 return self.vfs.join(os.path.join(f, *insidef))
800
800
801 def wjoin(self, f, *insidef):
801 def wjoin(self, f, *insidef):
802 return self.vfs.reljoin(self.root, f, *insidef)
802 return self.vfs.reljoin(self.root, f, *insidef)
803
803
804 def file(self, f):
804 def file(self, f):
805 if f[0] == '/':
805 if f[0] == '/':
806 f = f[1:]
806 f = f[1:]
807 return filelog.filelog(self.svfs, f)
807 return filelog.filelog(self.svfs, f)
808
808
809 def changectx(self, changeid):
809 def changectx(self, changeid):
810 return self[changeid]
810 return self[changeid]
811
811
812 def parents(self, changeid=None):
812 def parents(self, changeid=None):
813 '''get list of changectxs for parents of changeid'''
813 '''get list of changectxs for parents of changeid'''
814 return self[changeid].parents()
814 return self[changeid].parents()
815
815
816 def setparents(self, p1, p2=nullid):
816 def setparents(self, p1, p2=nullid):
817 self.dirstate.beginparentchange()
817 self.dirstate.beginparentchange()
818 copies = self.dirstate.setparents(p1, p2)
818 copies = self.dirstate.setparents(p1, p2)
819 pctx = self[p1]
819 pctx = self[p1]
820 if copies:
820 if copies:
821 # Adjust copy records, the dirstate cannot do it, it
821 # Adjust copy records, the dirstate cannot do it, it
822 # requires access to parents manifests. Preserve them
822 # requires access to parents manifests. Preserve them
823 # only for entries added to first parent.
823 # only for entries added to first parent.
824 for f in copies:
824 for f in copies:
825 if f not in pctx and copies[f] in pctx:
825 if f not in pctx and copies[f] in pctx:
826 self.dirstate.copy(copies[f], f)
826 self.dirstate.copy(copies[f], f)
827 if p2 == nullid:
827 if p2 == nullid:
828 for f, s in sorted(self.dirstate.copies().items()):
828 for f, s in sorted(self.dirstate.copies().items()):
829 if f not in pctx and s not in pctx:
829 if f not in pctx and s not in pctx:
830 self.dirstate.copy(None, f)
830 self.dirstate.copy(None, f)
831 self.dirstate.endparentchange()
831 self.dirstate.endparentchange()
832
832
833 def filectx(self, path, changeid=None, fileid=None):
833 def filectx(self, path, changeid=None, fileid=None):
834 """changeid can be a changeset revision, node, or tag.
834 """changeid can be a changeset revision, node, or tag.
835 fileid can be a file revision or node."""
835 fileid can be a file revision or node."""
836 return context.filectx(self, path, changeid, fileid)
836 return context.filectx(self, path, changeid, fileid)
837
837
838 def getcwd(self):
838 def getcwd(self):
839 return self.dirstate.getcwd()
839 return self.dirstate.getcwd()
840
840
841 def pathto(self, f, cwd=None):
841 def pathto(self, f, cwd=None):
842 return self.dirstate.pathto(f, cwd)
842 return self.dirstate.pathto(f, cwd)
843
843
844 def wfile(self, f, mode='r'):
844 def wfile(self, f, mode='r'):
845 return self.wvfs(f, mode)
845 return self.wvfs(f, mode)
846
846
847 def _link(self, f):
847 def _link(self, f):
848 return self.wvfs.islink(f)
848 return self.wvfs.islink(f)
849
849
850 def _loadfilter(self, filter):
850 def _loadfilter(self, filter):
851 if filter not in self.filterpats:
851 if filter not in self.filterpats:
852 l = []
852 l = []
853 for pat, cmd in self.ui.configitems(filter):
853 for pat, cmd in self.ui.configitems(filter):
854 if cmd == '!':
854 if cmd == '!':
855 continue
855 continue
856 mf = matchmod.match(self.root, '', [pat])
856 mf = matchmod.match(self.root, '', [pat])
857 fn = None
857 fn = None
858 params = cmd
858 params = cmd
859 for name, filterfn in self._datafilters.iteritems():
859 for name, filterfn in self._datafilters.iteritems():
860 if cmd.startswith(name):
860 if cmd.startswith(name):
861 fn = filterfn
861 fn = filterfn
862 params = cmd[len(name):].lstrip()
862 params = cmd[len(name):].lstrip()
863 break
863 break
864 if not fn:
864 if not fn:
865 fn = lambda s, c, **kwargs: util.filter(s, c)
865 fn = lambda s, c, **kwargs: util.filter(s, c)
866 # Wrap old filters not supporting keyword arguments
866 # Wrap old filters not supporting keyword arguments
867 if not inspect.getargspec(fn)[2]:
867 if not inspect.getargspec(fn)[2]:
868 oldfn = fn
868 oldfn = fn
869 fn = lambda s, c, **kwargs: oldfn(s, c)
869 fn = lambda s, c, **kwargs: oldfn(s, c)
870 l.append((mf, fn, params))
870 l.append((mf, fn, params))
871 self.filterpats[filter] = l
871 self.filterpats[filter] = l
872 return self.filterpats[filter]
872 return self.filterpats[filter]
873
873
874 def _filter(self, filterpats, filename, data):
874 def _filter(self, filterpats, filename, data):
875 for mf, fn, cmd in filterpats:
875 for mf, fn, cmd in filterpats:
876 if mf(filename):
876 if mf(filename):
877 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
877 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
878 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
878 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
879 break
879 break
880
880
881 return data
881 return data
882
882
883 @unfilteredpropertycache
883 @unfilteredpropertycache
884 def _encodefilterpats(self):
884 def _encodefilterpats(self):
885 return self._loadfilter('encode')
885 return self._loadfilter('encode')
886
886
887 @unfilteredpropertycache
887 @unfilteredpropertycache
888 def _decodefilterpats(self):
888 def _decodefilterpats(self):
889 return self._loadfilter('decode')
889 return self._loadfilter('decode')
890
890
891 def adddatafilter(self, name, filter):
891 def adddatafilter(self, name, filter):
892 self._datafilters[name] = filter
892 self._datafilters[name] = filter
893
893
894 def wread(self, filename):
894 def wread(self, filename):
895 if self._link(filename):
895 if self._link(filename):
896 data = self.wvfs.readlink(filename)
896 data = self.wvfs.readlink(filename)
897 else:
897 else:
898 data = self.wvfs.read(filename)
898 data = self.wvfs.read(filename)
899 return self._filter(self._encodefilterpats, filename, data)
899 return self._filter(self._encodefilterpats, filename, data)
900
900
901 def wwrite(self, filename, data, flags):
901 def wwrite(self, filename, data, flags):
902 data = self._filter(self._decodefilterpats, filename, data)
902 data = self._filter(self._decodefilterpats, filename, data)
903 if 'l' in flags:
903 if 'l' in flags:
904 self.wvfs.symlink(data, filename)
904 self.wvfs.symlink(data, filename)
905 else:
905 else:
906 self.wvfs.write(filename, data)
906 self.wvfs.write(filename, data)
907 if 'x' in flags:
907 if 'x' in flags:
908 self.wvfs.setflags(filename, False, True)
908 self.wvfs.setflags(filename, False, True)
909
909
910 def wwritedata(self, filename, data):
910 def wwritedata(self, filename, data):
911 return self._filter(self._decodefilterpats, filename, data)
911 return self._filter(self._decodefilterpats, filename, data)
912
912
913 def currenttransaction(self):
913 def currenttransaction(self):
914 """return the current transaction or None if non exists"""
914 """return the current transaction or None if non exists"""
915 if self._transref:
915 if self._transref:
916 tr = self._transref()
916 tr = self._transref()
917 else:
917 else:
918 tr = None
918 tr = None
919
919
920 if tr and tr.running():
920 if tr and tr.running():
921 return tr
921 return tr
922 return None
922 return None
923
923
924 def transaction(self, desc, report=None):
924 def transaction(self, desc, report=None):
925 if (self.ui.configbool('devel', 'all')
925 if (self.ui.configbool('devel', 'all')
926 or self.ui.configbool('devel', 'check-locks')):
926 or self.ui.configbool('devel', 'check-locks')):
927 l = self._lockref and self._lockref()
927 l = self._lockref and self._lockref()
928 if l is None or not l.held:
928 if l is None or not l.held:
929 msg = 'transaction with no lock\n'
929 msg = 'transaction with no lock\n'
930 if self.ui.tracebackflag:
930 if self.ui.tracebackflag:
931 util.debugstacktrace(msg, 1)
931 util.debugstacktrace(msg, 1)
932 else:
932 else:
933 self.ui.write_err(msg)
933 self.ui.write_err(msg)
934 tr = self.currenttransaction()
934 tr = self.currenttransaction()
935 if tr is not None:
935 if tr is not None:
936 return tr.nest()
936 return tr.nest()
937
937
938 # abort here if the journal already exists
938 # abort here if the journal already exists
939 if self.svfs.exists("journal"):
939 if self.svfs.exists("journal"):
940 raise error.RepoError(
940 raise error.RepoError(
941 _("abandoned transaction found"),
941 _("abandoned transaction found"),
942 hint=_("run 'hg recover' to clean up transaction"))
942 hint=_("run 'hg recover' to clean up transaction"))
943
943
944 self.hook('pretxnopen', throw=True, txnname=desc)
944 self.hook('pretxnopen', throw=True, txnname=desc)
945
945
946 self._writejournal(desc)
946 self._writejournal(desc)
947 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
947 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
948 if report:
948 if report:
949 rp = report
949 rp = report
950 else:
950 else:
951 rp = self.ui.warn
951 rp = self.ui.warn
952 vfsmap = {'plain': self.vfs} # root of .hg/
952 vfsmap = {'plain': self.vfs} # root of .hg/
953 # we must avoid cyclic reference between repo and transaction.
953 # we must avoid cyclic reference between repo and transaction.
954 reporef = weakref.ref(self)
954 reporef = weakref.ref(self)
955 def validate(tr):
955 def validate(tr):
956 """will run pre-closing hooks"""
956 """will run pre-closing hooks"""
957 pending = lambda: tr.writepending() and self.root or ""
957 pending = lambda: tr.writepending() and self.root or ""
958 reporef().hook('pretxnclose', throw=True, pending=pending,
958 reporef().hook('pretxnclose', throw=True, pending=pending,
959 xnname=desc)
959 xnname=desc)
960
960
961 tr = transaction.transaction(rp, self.sopener, vfsmap,
961 tr = transaction.transaction(rp, self.sopener, vfsmap,
962 "journal",
962 "journal",
963 "undo",
963 "undo",
964 aftertrans(renames),
964 aftertrans(renames),
965 self.store.createmode,
965 self.store.createmode,
966 validator=validate)
966 validator=validate)
967 # note: writing the fncache only during finalize mean that the file is
967 # note: writing the fncache only during finalize mean that the file is
968 # outdated when running hooks. As fncache is used for streaming clone,
968 # outdated when running hooks. As fncache is used for streaming clone,
969 # this is not expected to break anything that happen during the hooks.
969 # this is not expected to break anything that happen during the hooks.
970 tr.addfinalize('flush-fncache', self.store.write)
970 tr.addfinalize('flush-fncache', self.store.write)
971 def txnclosehook(tr2):
971 def txnclosehook(tr2):
972 """To be run if transaction is successful, will schedule a hook run
972 """To be run if transaction is successful, will schedule a hook run
973 """
973 """
974 def hook():
974 def hook():
975 reporef().hook('txnclose', throw=False, txnname=desc,
975 reporef().hook('txnclose', throw=False, txnname=desc,
976 **tr2.hookargs)
976 **tr2.hookargs)
977 reporef()._afterlock(hook)
977 reporef()._afterlock(hook)
978 tr.addfinalize('txnclose-hook', txnclosehook)
978 tr.addfinalize('txnclose-hook', txnclosehook)
979 self._transref = weakref.ref(tr)
979 self._transref = weakref.ref(tr)
980 return tr
980 return tr
981
981
982 def _journalfiles(self):
982 def _journalfiles(self):
983 return ((self.svfs, 'journal'),
983 return ((self.svfs, 'journal'),
984 (self.vfs, 'journal.dirstate'),
984 (self.vfs, 'journal.dirstate'),
985 (self.vfs, 'journal.branch'),
985 (self.vfs, 'journal.branch'),
986 (self.vfs, 'journal.desc'),
986 (self.vfs, 'journal.desc'),
987 (self.vfs, 'journal.bookmarks'),
987 (self.vfs, 'journal.bookmarks'),
988 (self.svfs, 'journal.phaseroots'))
988 (self.svfs, 'journal.phaseroots'))
989
989
990 def undofiles(self):
990 def undofiles(self):
991 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
991 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
992
992
993 def _writejournal(self, desc):
993 def _writejournal(self, desc):
994 self.vfs.write("journal.dirstate",
994 self.vfs.write("journal.dirstate",
995 self.vfs.tryread("dirstate"))
995 self.vfs.tryread("dirstate"))
996 self.vfs.write("journal.branch",
996 self.vfs.write("journal.branch",
997 encoding.fromlocal(self.dirstate.branch()))
997 encoding.fromlocal(self.dirstate.branch()))
998 self.vfs.write("journal.desc",
998 self.vfs.write("journal.desc",
999 "%d\n%s\n" % (len(self), desc))
999 "%d\n%s\n" % (len(self), desc))
1000 self.vfs.write("journal.bookmarks",
1000 self.vfs.write("journal.bookmarks",
1001 self.vfs.tryread("bookmarks"))
1001 self.vfs.tryread("bookmarks"))
1002 self.svfs.write("journal.phaseroots",
1002 self.svfs.write("journal.phaseroots",
1003 self.svfs.tryread("phaseroots"))
1003 self.svfs.tryread("phaseroots"))
1004
1004
1005 def recover(self):
1005 def recover(self):
1006 lock = self.lock()
1006 lock = self.lock()
1007 try:
1007 try:
1008 if self.svfs.exists("journal"):
1008 if self.svfs.exists("journal"):
1009 self.ui.status(_("rolling back interrupted transaction\n"))
1009 self.ui.status(_("rolling back interrupted transaction\n"))
1010 vfsmap = {'': self.svfs,
1010 vfsmap = {'': self.svfs,
1011 'plain': self.vfs,}
1011 'plain': self.vfs,}
1012 transaction.rollback(self.svfs, vfsmap, "journal",
1012 transaction.rollback(self.svfs, vfsmap, "journal",
1013 self.ui.warn)
1013 self.ui.warn)
1014 self.invalidate()
1014 self.invalidate()
1015 return True
1015 return True
1016 else:
1016 else:
1017 self.ui.warn(_("no interrupted transaction available\n"))
1017 self.ui.warn(_("no interrupted transaction available\n"))
1018 return False
1018 return False
1019 finally:
1019 finally:
1020 lock.release()
1020 lock.release()
1021
1021
1022 def rollback(self, dryrun=False, force=False):
1022 def rollback(self, dryrun=False, force=False):
1023 wlock = lock = None
1023 wlock = lock = None
1024 try:
1024 try:
1025 wlock = self.wlock()
1025 wlock = self.wlock()
1026 lock = self.lock()
1026 lock = self.lock()
1027 if self.svfs.exists("undo"):
1027 if self.svfs.exists("undo"):
1028 return self._rollback(dryrun, force)
1028 return self._rollback(dryrun, force)
1029 else:
1029 else:
1030 self.ui.warn(_("no rollback information available\n"))
1030 self.ui.warn(_("no rollback information available\n"))
1031 return 1
1031 return 1
1032 finally:
1032 finally:
1033 release(lock, wlock)
1033 release(lock, wlock)
1034
1034
1035 @unfilteredmethod # Until we get smarter cache management
1035 @unfilteredmethod # Until we get smarter cache management
1036 def _rollback(self, dryrun, force):
1036 def _rollback(self, dryrun, force):
1037 ui = self.ui
1037 ui = self.ui
1038 try:
1038 try:
1039 args = self.vfs.read('undo.desc').splitlines()
1039 args = self.vfs.read('undo.desc').splitlines()
1040 (oldlen, desc, detail) = (int(args[0]), args[1], None)
1040 (oldlen, desc, detail) = (int(args[0]), args[1], None)
1041 if len(args) >= 3:
1041 if len(args) >= 3:
1042 detail = args[2]
1042 detail = args[2]
1043 oldtip = oldlen - 1
1043 oldtip = oldlen - 1
1044
1044
1045 if detail and ui.verbose:
1045 if detail and ui.verbose:
1046 msg = (_('repository tip rolled back to revision %s'
1046 msg = (_('repository tip rolled back to revision %s'
1047 ' (undo %s: %s)\n')
1047 ' (undo %s: %s)\n')
1048 % (oldtip, desc, detail))
1048 % (oldtip, desc, detail))
1049 else:
1049 else:
1050 msg = (_('repository tip rolled back to revision %s'
1050 msg = (_('repository tip rolled back to revision %s'
1051 ' (undo %s)\n')
1051 ' (undo %s)\n')
1052 % (oldtip, desc))
1052 % (oldtip, desc))
1053 except IOError:
1053 except IOError:
1054 msg = _('rolling back unknown transaction\n')
1054 msg = _('rolling back unknown transaction\n')
1055 desc = None
1055 desc = None
1056
1056
1057 if not force and self['.'] != self['tip'] and desc == 'commit':
1057 if not force and self['.'] != self['tip'] and desc == 'commit':
1058 raise util.Abort(
1058 raise util.Abort(
1059 _('rollback of last commit while not checked out '
1059 _('rollback of last commit while not checked out '
1060 'may lose data'), hint=_('use -f to force'))
1060 'may lose data'), hint=_('use -f to force'))
1061
1061
1062 ui.status(msg)
1062 ui.status(msg)
1063 if dryrun:
1063 if dryrun:
1064 return 0
1064 return 0
1065
1065
1066 parents = self.dirstate.parents()
1066 parents = self.dirstate.parents()
1067 self.destroying()
1067 self.destroying()
1068 vfsmap = {'plain': self.vfs, '': self.svfs}
1068 vfsmap = {'plain': self.vfs, '': self.svfs}
1069 transaction.rollback(self.svfs, vfsmap, 'undo', ui.warn)
1069 transaction.rollback(self.svfs, vfsmap, 'undo', ui.warn)
1070 if self.vfs.exists('undo.bookmarks'):
1070 if self.vfs.exists('undo.bookmarks'):
1071 self.vfs.rename('undo.bookmarks', 'bookmarks')
1071 self.vfs.rename('undo.bookmarks', 'bookmarks')
1072 if self.svfs.exists('undo.phaseroots'):
1072 if self.svfs.exists('undo.phaseroots'):
1073 self.svfs.rename('undo.phaseroots', 'phaseroots')
1073 self.svfs.rename('undo.phaseroots', 'phaseroots')
1074 self.invalidate()
1074 self.invalidate()
1075
1075
1076 parentgone = (parents[0] not in self.changelog.nodemap or
1076 parentgone = (parents[0] not in self.changelog.nodemap or
1077 parents[1] not in self.changelog.nodemap)
1077 parents[1] not in self.changelog.nodemap)
1078 if parentgone:
1078 if parentgone:
1079 self.vfs.rename('undo.dirstate', 'dirstate')
1079 self.vfs.rename('undo.dirstate', 'dirstate')
1080 try:
1080 try:
1081 branch = self.vfs.read('undo.branch')
1081 branch = self.vfs.read('undo.branch')
1082 self.dirstate.setbranch(encoding.tolocal(branch))
1082 self.dirstate.setbranch(encoding.tolocal(branch))
1083 except IOError:
1083 except IOError:
1084 ui.warn(_('named branch could not be reset: '
1084 ui.warn(_('named branch could not be reset: '
1085 'current branch is still \'%s\'\n')
1085 'current branch is still \'%s\'\n')
1086 % self.dirstate.branch())
1086 % self.dirstate.branch())
1087
1087
1088 self.dirstate.invalidate()
1088 self.dirstate.invalidate()
1089 parents = tuple([p.rev() for p in self.parents()])
1089 parents = tuple([p.rev() for p in self.parents()])
1090 if len(parents) > 1:
1090 if len(parents) > 1:
1091 ui.status(_('working directory now based on '
1091 ui.status(_('working directory now based on '
1092 'revisions %d and %d\n') % parents)
1092 'revisions %d and %d\n') % parents)
1093 else:
1093 else:
1094 ui.status(_('working directory now based on '
1094 ui.status(_('working directory now based on '
1095 'revision %d\n') % parents)
1095 'revision %d\n') % parents)
1096 # TODO: if we know which new heads may result from this rollback, pass
1096 # TODO: if we know which new heads may result from this rollback, pass
1097 # them to destroy(), which will prevent the branchhead cache from being
1097 # them to destroy(), which will prevent the branchhead cache from being
1098 # invalidated.
1098 # invalidated.
1099 self.destroyed()
1099 self.destroyed()
1100 return 0
1100 return 0
1101
1101
1102 def invalidatecaches(self):
1102 def invalidatecaches(self):
1103
1103
1104 if '_tagscache' in vars(self):
1104 if '_tagscache' in vars(self):
1105 # can't use delattr on proxy
1105 # can't use delattr on proxy
1106 del self.__dict__['_tagscache']
1106 del self.__dict__['_tagscache']
1107
1107
1108 self.unfiltered()._branchcaches.clear()
1108 self.unfiltered()._branchcaches.clear()
1109 self.invalidatevolatilesets()
1109 self.invalidatevolatilesets()
1110
1110
1111 def invalidatevolatilesets(self):
1111 def invalidatevolatilesets(self):
1112 self.filteredrevcache.clear()
1112 self.filteredrevcache.clear()
1113 obsolete.clearobscaches(self)
1113 obsolete.clearobscaches(self)
1114
1114
1115 def invalidatedirstate(self):
1115 def invalidatedirstate(self):
1116 '''Invalidates the dirstate, causing the next call to dirstate
1116 '''Invalidates the dirstate, causing the next call to dirstate
1117 to check if it was modified since the last time it was read,
1117 to check if it was modified since the last time it was read,
1118 rereading it if it has.
1118 rereading it if it has.
1119
1119
1120 This is different to dirstate.invalidate() that it doesn't always
1120 This is different to dirstate.invalidate() that it doesn't always
1121 rereads the dirstate. Use dirstate.invalidate() if you want to
1121 rereads the dirstate. Use dirstate.invalidate() if you want to
1122 explicitly read the dirstate again (i.e. restoring it to a previous
1122 explicitly read the dirstate again (i.e. restoring it to a previous
1123 known good state).'''
1123 known good state).'''
1124 if hasunfilteredcache(self, 'dirstate'):
1124 if hasunfilteredcache(self, 'dirstate'):
1125 for k in self.dirstate._filecache:
1125 for k in self.dirstate._filecache:
1126 try:
1126 try:
1127 delattr(self.dirstate, k)
1127 delattr(self.dirstate, k)
1128 except AttributeError:
1128 except AttributeError:
1129 pass
1129 pass
1130 delattr(self.unfiltered(), 'dirstate')
1130 delattr(self.unfiltered(), 'dirstate')
1131
1131
1132 def invalidate(self):
1132 def invalidate(self):
1133 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1133 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1134 for k in self._filecache:
1134 for k in self._filecache:
1135 # dirstate is invalidated separately in invalidatedirstate()
1135 # dirstate is invalidated separately in invalidatedirstate()
1136 if k == 'dirstate':
1136 if k == 'dirstate':
1137 continue
1137 continue
1138
1138
1139 try:
1139 try:
1140 delattr(unfiltered, k)
1140 delattr(unfiltered, k)
1141 except AttributeError:
1141 except AttributeError:
1142 pass
1142 pass
1143 self.invalidatecaches()
1143 self.invalidatecaches()
1144 self.store.invalidatecaches()
1144 self.store.invalidatecaches()
1145
1145
1146 def invalidateall(self):
1146 def invalidateall(self):
1147 '''Fully invalidates both store and non-store parts, causing the
1147 '''Fully invalidates both store and non-store parts, causing the
1148 subsequent operation to reread any outside changes.'''
1148 subsequent operation to reread any outside changes.'''
1149 # extension should hook this to invalidate its caches
1149 # extension should hook this to invalidate its caches
1150 self.invalidate()
1150 self.invalidate()
1151 self.invalidatedirstate()
1151 self.invalidatedirstate()
1152
1152
1153 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1153 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1154 try:
1154 try:
1155 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1155 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1156 except error.LockHeld, inst:
1156 except error.LockHeld, inst:
1157 if not wait:
1157 if not wait:
1158 raise
1158 raise
1159 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1159 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1160 (desc, inst.locker))
1160 (desc, inst.locker))
1161 # default to 600 seconds timeout
1161 # default to 600 seconds timeout
1162 l = lockmod.lock(vfs, lockname,
1162 l = lockmod.lock(vfs, lockname,
1163 int(self.ui.config("ui", "timeout", "600")),
1163 int(self.ui.config("ui", "timeout", "600")),
1164 releasefn, desc=desc)
1164 releasefn, desc=desc)
1165 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1165 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1166 if acquirefn:
1166 if acquirefn:
1167 acquirefn()
1167 acquirefn()
1168 return l
1168 return l
1169
1169
1170 def _afterlock(self, callback):
1170 def _afterlock(self, callback):
1171 """add a callback to the current repository lock.
1171 """add a callback to the current repository lock.
1172
1172
1173 The callback will be executed on lock release."""
1173 The callback will be executed on lock release."""
1174 l = self._lockref and self._lockref()
1174 l = self._lockref and self._lockref()
1175 if l:
1175 if l:
1176 l.postrelease.append(callback)
1176 l.postrelease.append(callback)
1177 else:
1177 else:
1178 callback()
1178 callback()
1179
1179
1180 def lock(self, wait=True):
1180 def lock(self, wait=True):
1181 '''Lock the repository store (.hg/store) and return a weak reference
1181 '''Lock the repository store (.hg/store) and return a weak reference
1182 to the lock. Use this before modifying the store (e.g. committing or
1182 to the lock. Use this before modifying the store (e.g. committing or
1183 stripping). If you are opening a transaction, get a lock as well.)'''
1183 stripping). If you are opening a transaction, get a lock as well.)'''
1184 l = self._lockref and self._lockref()
1184 l = self._lockref and self._lockref()
1185 if l is not None and l.held:
1185 if l is not None and l.held:
1186 l.lock()
1186 l.lock()
1187 return l
1187 return l
1188
1188
1189 def unlock():
1189 def unlock():
1190 for k, ce in self._filecache.items():
1190 for k, ce in self._filecache.items():
1191 if k == 'dirstate' or k not in self.__dict__:
1191 if k == 'dirstate' or k not in self.__dict__:
1192 continue
1192 continue
1193 ce.refresh()
1193 ce.refresh()
1194
1194
1195 l = self._lock(self.svfs, "lock", wait, unlock,
1195 l = self._lock(self.svfs, "lock", wait, unlock,
1196 self.invalidate, _('repository %s') % self.origroot)
1196 self.invalidate, _('repository %s') % self.origroot)
1197 self._lockref = weakref.ref(l)
1197 self._lockref = weakref.ref(l)
1198 return l
1198 return l
1199
1199
1200 def wlock(self, wait=True):
1200 def wlock(self, wait=True):
1201 '''Lock the non-store parts of the repository (everything under
1201 '''Lock the non-store parts of the repository (everything under
1202 .hg except .hg/store) and return a weak reference to the lock.
1202 .hg except .hg/store) and return a weak reference to the lock.
1203 Use this before modifying files in .hg.'''
1203 Use this before modifying files in .hg.'''
1204 if (self.ui.configbool('devel', 'all')
1204 if (self.ui.configbool('devel', 'all')
1205 or self.ui.configbool('devel', 'check-locks')):
1205 or self.ui.configbool('devel', 'check-locks')):
1206 l = self._lockref and self._lockref()
1206 l = self._lockref and self._lockref()
1207 if l is not None and l.held:
1207 if l is not None and l.held:
1208 msg = '"lock" taken before "wlock"\n'
1208 msg = '"lock" taken before "wlock"\n'
1209 if self.ui.tracebackflag:
1209 if self.ui.tracebackflag:
1210 util.debugstacktrace(msg, 1)
1210 util.debugstacktrace(msg, 1)
1211 else:
1211 else:
1212 self.ui.write_err(msg)
1212 self.ui.write_err(msg)
1213 l = self._wlockref and self._wlockref()
1213 l = self._wlockref and self._wlockref()
1214 if l is not None and l.held:
1214 if l is not None and l.held:
1215 l.lock()
1215 l.lock()
1216 return l
1216 return l
1217
1217
1218 def unlock():
1218 def unlock():
1219 if self.dirstate.pendingparentchange():
1219 if self.dirstate.pendingparentchange():
1220 self.dirstate.invalidate()
1220 self.dirstate.invalidate()
1221 else:
1221 else:
1222 self.dirstate.write()
1222 self.dirstate.write()
1223
1223
1224 self._filecache['dirstate'].refresh()
1224 self._filecache['dirstate'].refresh()
1225
1225
1226 l = self._lock(self.vfs, "wlock", wait, unlock,
1226 l = self._lock(self.vfs, "wlock", wait, unlock,
1227 self.invalidatedirstate, _('working directory of %s') %
1227 self.invalidatedirstate, _('working directory of %s') %
1228 self.origroot)
1228 self.origroot)
1229 self._wlockref = weakref.ref(l)
1229 self._wlockref = weakref.ref(l)
1230 return l
1230 return l
1231
1231
1232 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1232 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1233 """
1233 """
1234 commit an individual file as part of a larger transaction
1234 commit an individual file as part of a larger transaction
1235 """
1235 """
1236
1236
1237 fname = fctx.path()
1237 fname = fctx.path()
1238 fparent1 = manifest1.get(fname, nullid)
1238 fparent1 = manifest1.get(fname, nullid)
1239 fparent2 = manifest2.get(fname, nullid)
1239 fparent2 = manifest2.get(fname, nullid)
1240 if isinstance(fctx, context.filectx):
1240 if isinstance(fctx, context.filectx):
1241 node = fctx.filenode()
1241 node = fctx.filenode()
1242 if node in [fparent1, fparent2]:
1242 if node in [fparent1, fparent2]:
1243 self.ui.debug('reusing %s filelog entry\n' % fname)
1243 self.ui.debug('reusing %s filelog entry\n' % fname)
1244 return node
1244 return node
1245
1245
1246 flog = self.file(fname)
1246 flog = self.file(fname)
1247 meta = {}
1247 meta = {}
1248 copy = fctx.renamed()
1248 copy = fctx.renamed()
1249 if copy and copy[0] != fname:
1249 if copy and copy[0] != fname:
1250 # Mark the new revision of this file as a copy of another
1250 # Mark the new revision of this file as a copy of another
1251 # file. This copy data will effectively act as a parent
1251 # file. This copy data will effectively act as a parent
1252 # of this new revision. If this is a merge, the first
1252 # of this new revision. If this is a merge, the first
1253 # parent will be the nullid (meaning "look up the copy data")
1253 # parent will be the nullid (meaning "look up the copy data")
1254 # and the second one will be the other parent. For example:
1254 # and the second one will be the other parent. For example:
1255 #
1255 #
1256 # 0 --- 1 --- 3 rev1 changes file foo
1256 # 0 --- 1 --- 3 rev1 changes file foo
1257 # \ / rev2 renames foo to bar and changes it
1257 # \ / rev2 renames foo to bar and changes it
1258 # \- 2 -/ rev3 should have bar with all changes and
1258 # \- 2 -/ rev3 should have bar with all changes and
1259 # should record that bar descends from
1259 # should record that bar descends from
1260 # bar in rev2 and foo in rev1
1260 # bar in rev2 and foo in rev1
1261 #
1261 #
1262 # this allows this merge to succeed:
1262 # this allows this merge to succeed:
1263 #
1263 #
1264 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1264 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1265 # \ / merging rev3 and rev4 should use bar@rev2
1265 # \ / merging rev3 and rev4 should use bar@rev2
1266 # \- 2 --- 4 as the merge base
1266 # \- 2 --- 4 as the merge base
1267 #
1267 #
1268
1268
1269 cfname = copy[0]
1269 cfname = copy[0]
1270 crev = manifest1.get(cfname)
1270 crev = manifest1.get(cfname)
1271 newfparent = fparent2
1271 newfparent = fparent2
1272
1272
1273 if manifest2: # branch merge
1273 if manifest2: # branch merge
1274 if fparent2 == nullid or crev is None: # copied on remote side
1274 if fparent2 == nullid or crev is None: # copied on remote side
1275 if cfname in manifest2:
1275 if cfname in manifest2:
1276 crev = manifest2[cfname]
1276 crev = manifest2[cfname]
1277 newfparent = fparent1
1277 newfparent = fparent1
1278
1278
1279 # Here, we used to search backwards through history to try to find
1279 # Here, we used to search backwards through history to try to find
1280 # where the file copy came from if the source of a copy was not in
1280 # where the file copy came from if the source of a copy was not in
1281 # the parent directory. However, this doesn't actually make sense to
1281 # the parent directory. However, this doesn't actually make sense to
1282 # do (what does a copy from something not in your working copy even
1282 # do (what does a copy from something not in your working copy even
1283 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
1283 # mean?) and it causes bugs (eg, issue4476). Instead, we will warn
1284 # the user that copy information was dropped, so if they didn't
1284 # the user that copy information was dropped, so if they didn't
1285 # expect this outcome it can be fixed, but this is the correct
1285 # expect this outcome it can be fixed, but this is the correct
1286 # behavior in this circumstance.
1286 # behavior in this circumstance.
1287
1287
1288 if crev:
1288 if crev:
1289 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1289 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1290 meta["copy"] = cfname
1290 meta["copy"] = cfname
1291 meta["copyrev"] = hex(crev)
1291 meta["copyrev"] = hex(crev)
1292 fparent1, fparent2 = nullid, newfparent
1292 fparent1, fparent2 = nullid, newfparent
1293 else:
1293 else:
1294 self.ui.warn(_("warning: can't find ancestor for '%s' "
1294 self.ui.warn(_("warning: can't find ancestor for '%s' "
1295 "copied from '%s'!\n") % (fname, cfname))
1295 "copied from '%s'!\n") % (fname, cfname))
1296
1296
1297 elif fparent1 == nullid:
1297 elif fparent1 == nullid:
1298 fparent1, fparent2 = fparent2, nullid
1298 fparent1, fparent2 = fparent2, nullid
1299 elif fparent2 != nullid:
1299 elif fparent2 != nullid:
1300 # is one parent an ancestor of the other?
1300 # is one parent an ancestor of the other?
1301 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1301 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1302 if fparent1 in fparentancestors:
1302 if fparent1 in fparentancestors:
1303 fparent1, fparent2 = fparent2, nullid
1303 fparent1, fparent2 = fparent2, nullid
1304 elif fparent2 in fparentancestors:
1304 elif fparent2 in fparentancestors:
1305 fparent2 = nullid
1305 fparent2 = nullid
1306
1306
1307 # is the file changed?
1307 # is the file changed?
1308 text = fctx.data()
1308 text = fctx.data()
1309 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1309 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1310 changelist.append(fname)
1310 changelist.append(fname)
1311 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1311 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1312 # are just the flags changed during merge?
1312 # are just the flags changed during merge?
1313 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1313 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1314 changelist.append(fname)
1314 changelist.append(fname)
1315
1315
1316 return fparent1
1316 return fparent1
1317
1317
1318 @unfilteredmethod
1318 @unfilteredmethod
1319 def commit(self, text="", user=None, date=None, match=None, force=False,
1319 def commit(self, text="", user=None, date=None, match=None, force=False,
1320 editor=False, extra={}):
1320 editor=False, extra={}):
1321 """Add a new revision to current repository.
1321 """Add a new revision to current repository.
1322
1322
1323 Revision information is gathered from the working directory,
1323 Revision information is gathered from the working directory,
1324 match can be used to filter the committed files. If editor is
1324 match can be used to filter the committed files. If editor is
1325 supplied, it is called to get a commit message.
1325 supplied, it is called to get a commit message.
1326 """
1326 """
1327
1327
1328 def fail(f, msg):
1328 def fail(f, msg):
1329 raise util.Abort('%s: %s' % (f, msg))
1329 raise util.Abort('%s: %s' % (f, msg))
1330
1330
1331 if not match:
1331 if not match:
1332 match = matchmod.always(self.root, '')
1332 match = matchmod.always(self.root, '')
1333
1333
1334 if not force:
1334 if not force:
1335 vdirs = []
1335 vdirs = []
1336 match.explicitdir = vdirs.append
1336 match.explicitdir = vdirs.append
1337 match.bad = fail
1337 match.bad = fail
1338
1338
1339 wlock = self.wlock()
1339 wlock = self.wlock()
1340 try:
1340 try:
1341 wctx = self[None]
1341 wctx = self[None]
1342 merge = len(wctx.parents()) > 1
1342 merge = len(wctx.parents()) > 1
1343
1343
1344 if not force and merge and not match.always():
1344 if not force and merge and not match.always():
1345 raise util.Abort(_('cannot partially commit a merge '
1345 raise util.Abort(_('cannot partially commit a merge '
1346 '(do not specify files or patterns)'))
1346 '(do not specify files or patterns)'))
1347
1347
1348 status = self.status(match=match, clean=force)
1348 status = self.status(match=match, clean=force)
1349 if force:
1349 if force:
1350 status.modified.extend(status.clean) # mq may commit clean files
1350 status.modified.extend(status.clean) # mq may commit clean files
1351
1351
1352 # check subrepos
1352 # check subrepos
1353 subs = []
1353 subs = []
1354 commitsubs = set()
1354 commitsubs = set()
1355 newstate = wctx.substate.copy()
1355 newstate = wctx.substate.copy()
1356 # only manage subrepos and .hgsubstate if .hgsub is present
1356 # only manage subrepos and .hgsubstate if .hgsub is present
1357 if '.hgsub' in wctx:
1357 if '.hgsub' in wctx:
1358 # we'll decide whether to track this ourselves, thanks
1358 # we'll decide whether to track this ourselves, thanks
1359 for c in status.modified, status.added, status.removed:
1359 for c in status.modified, status.added, status.removed:
1360 if '.hgsubstate' in c:
1360 if '.hgsubstate' in c:
1361 c.remove('.hgsubstate')
1361 c.remove('.hgsubstate')
1362
1362
1363 # compare current state to last committed state
1363 # compare current state to last committed state
1364 # build new substate based on last committed state
1364 # build new substate based on last committed state
1365 oldstate = wctx.p1().substate
1365 oldstate = wctx.p1().substate
1366 for s in sorted(newstate.keys()):
1366 for s in sorted(newstate.keys()):
1367 if not match(s):
1367 if not match(s):
1368 # ignore working copy, use old state if present
1368 # ignore working copy, use old state if present
1369 if s in oldstate:
1369 if s in oldstate:
1370 newstate[s] = oldstate[s]
1370 newstate[s] = oldstate[s]
1371 continue
1371 continue
1372 if not force:
1372 if not force:
1373 raise util.Abort(
1373 raise util.Abort(
1374 _("commit with new subrepo %s excluded") % s)
1374 _("commit with new subrepo %s excluded") % s)
1375 if wctx.sub(s).dirty(True):
1375 if wctx.sub(s).dirty(True):
1376 if not self.ui.configbool('ui', 'commitsubrepos'):
1376 if not self.ui.configbool('ui', 'commitsubrepos'):
1377 raise util.Abort(
1377 raise util.Abort(
1378 _("uncommitted changes in subrepo %s") % s,
1378 _("uncommitted changes in subrepo %s") % s,
1379 hint=_("use --subrepos for recursive commit"))
1379 hint=_("use --subrepos for recursive commit"))
1380 subs.append(s)
1380 subs.append(s)
1381 commitsubs.add(s)
1381 commitsubs.add(s)
1382 else:
1382 else:
1383 bs = wctx.sub(s).basestate()
1383 bs = wctx.sub(s).basestate()
1384 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1384 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1385 if oldstate.get(s, (None, None, None))[1] != bs:
1385 if oldstate.get(s, (None, None, None))[1] != bs:
1386 subs.append(s)
1386 subs.append(s)
1387
1387
1388 # check for removed subrepos
1388 # check for removed subrepos
1389 for p in wctx.parents():
1389 for p in wctx.parents():
1390 r = [s for s in p.substate if s not in newstate]
1390 r = [s for s in p.substate if s not in newstate]
1391 subs += [s for s in r if match(s)]
1391 subs += [s for s in r if match(s)]
1392 if subs:
1392 if subs:
1393 if (not match('.hgsub') and
1393 if (not match('.hgsub') and
1394 '.hgsub' in (wctx.modified() + wctx.added())):
1394 '.hgsub' in (wctx.modified() + wctx.added())):
1395 raise util.Abort(
1395 raise util.Abort(
1396 _("can't commit subrepos without .hgsub"))
1396 _("can't commit subrepos without .hgsub"))
1397 status.modified.insert(0, '.hgsubstate')
1397 status.modified.insert(0, '.hgsubstate')
1398
1398
1399 elif '.hgsub' in status.removed:
1399 elif '.hgsub' in status.removed:
1400 # clean up .hgsubstate when .hgsub is removed
1400 # clean up .hgsubstate when .hgsub is removed
1401 if ('.hgsubstate' in wctx and
1401 if ('.hgsubstate' in wctx and
1402 '.hgsubstate' not in (status.modified + status.added +
1402 '.hgsubstate' not in (status.modified + status.added +
1403 status.removed)):
1403 status.removed)):
1404 status.removed.insert(0, '.hgsubstate')
1404 status.removed.insert(0, '.hgsubstate')
1405
1405
1406 # make sure all explicit patterns are matched
1406 # make sure all explicit patterns are matched
1407 if not force and match.files():
1407 if not force and match.files():
1408 matched = set(status.modified + status.added + status.removed)
1408 matched = set(status.modified + status.added + status.removed)
1409
1409
1410 for f in match.files():
1410 for f in match.files():
1411 f = self.dirstate.normalize(f)
1411 f = self.dirstate.normalize(f)
1412 if f == '.' or f in matched or f in wctx.substate:
1412 if f == '.' or f in matched or f in wctx.substate:
1413 continue
1413 continue
1414 if f in status.deleted:
1414 if f in status.deleted:
1415 fail(f, _('file not found!'))
1415 fail(f, _('file not found!'))
1416 if f in vdirs: # visited directory
1416 if f in vdirs: # visited directory
1417 d = f + '/'
1417 d = f + '/'
1418 for mf in matched:
1418 for mf in matched:
1419 if mf.startswith(d):
1419 if mf.startswith(d):
1420 break
1420 break
1421 else:
1421 else:
1422 fail(f, _("no match under directory!"))
1422 fail(f, _("no match under directory!"))
1423 elif f not in self.dirstate:
1423 elif f not in self.dirstate:
1424 fail(f, _("file not tracked!"))
1424 fail(f, _("file not tracked!"))
1425
1425
1426 cctx = context.workingcommitctx(self, status,
1426 cctx = context.workingcommitctx(self, status,
1427 text, user, date, extra)
1427 text, user, date, extra)
1428
1428
1429 if (not force and not extra.get("close") and not merge
1429 if (not force and not extra.get("close") and not merge
1430 and not cctx.files()
1430 and not cctx.files()
1431 and wctx.branch() == wctx.p1().branch()):
1431 and wctx.branch() == wctx.p1().branch()):
1432 return None
1432 return None
1433
1433
1434 if merge and cctx.deleted():
1434 if merge and cctx.deleted():
1435 raise util.Abort(_("cannot commit merge with missing files"))
1435 raise util.Abort(_("cannot commit merge with missing files"))
1436
1436
1437 ms = mergemod.mergestate(self)
1437 ms = mergemod.mergestate(self)
1438 for f in status.modified:
1438 for f in status.modified:
1439 if f in ms and ms[f] == 'u':
1439 if f in ms and ms[f] == 'u':
1440 raise util.Abort(_('unresolved merge conflicts '
1440 raise util.Abort(_('unresolved merge conflicts '
1441 '(see "hg help resolve")'))
1441 '(see "hg help resolve")'))
1442
1442
1443 if editor:
1443 if editor:
1444 cctx._text = editor(self, cctx, subs)
1444 cctx._text = editor(self, cctx, subs)
1445 edited = (text != cctx._text)
1445 edited = (text != cctx._text)
1446
1446
1447 # Save commit message in case this transaction gets rolled back
1447 # Save commit message in case this transaction gets rolled back
1448 # (e.g. by a pretxncommit hook). Leave the content alone on
1448 # (e.g. by a pretxncommit hook). Leave the content alone on
1449 # the assumption that the user will use the same editor again.
1449 # the assumption that the user will use the same editor again.
1450 msgfn = self.savecommitmessage(cctx._text)
1450 msgfn = self.savecommitmessage(cctx._text)
1451
1451
1452 # commit subs and write new state
1452 # commit subs and write new state
1453 if subs:
1453 if subs:
1454 for s in sorted(commitsubs):
1454 for s in sorted(commitsubs):
1455 sub = wctx.sub(s)
1455 sub = wctx.sub(s)
1456 self.ui.status(_('committing subrepository %s\n') %
1456 self.ui.status(_('committing subrepository %s\n') %
1457 subrepo.subrelpath(sub))
1457 subrepo.subrelpath(sub))
1458 sr = sub.commit(cctx._text, user, date)
1458 sr = sub.commit(cctx._text, user, date)
1459 newstate[s] = (newstate[s][0], sr)
1459 newstate[s] = (newstate[s][0], sr)
1460 subrepo.writestate(self, newstate)
1460 subrepo.writestate(self, newstate)
1461
1461
1462 p1, p2 = self.dirstate.parents()
1462 p1, p2 = self.dirstate.parents()
1463 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1463 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1464 try:
1464 try:
1465 self.hook("precommit", throw=True, parent1=hookp1,
1465 self.hook("precommit", throw=True, parent1=hookp1,
1466 parent2=hookp2)
1466 parent2=hookp2)
1467 ret = self.commitctx(cctx, True)
1467 ret = self.commitctx(cctx, True)
1468 except: # re-raises
1468 except: # re-raises
1469 if edited:
1469 if edited:
1470 self.ui.write(
1470 self.ui.write(
1471 _('note: commit message saved in %s\n') % msgfn)
1471 _('note: commit message saved in %s\n') % msgfn)
1472 raise
1472 raise
1473
1473
1474 # update bookmarks, dirstate and mergestate
1474 # update bookmarks, dirstate and mergestate
1475 bookmarks.update(self, [p1, p2], ret)
1475 bookmarks.update(self, [p1, p2], ret)
1476 cctx.markcommitted(ret)
1476 cctx.markcommitted(ret)
1477 ms.reset()
1477 ms.reset()
1478 finally:
1478 finally:
1479 wlock.release()
1479 wlock.release()
1480
1480
1481 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1481 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1482 # hack for command that use a temporary commit (eg: histedit)
1482 # hack for command that use a temporary commit (eg: histedit)
1483 # temporary commit got stripped before hook release
1483 # temporary commit got stripped before hook release
1484 if node in self:
1484 if node in self:
1485 self.hook("commit", node=node, parent1=parent1,
1485 self.hook("commit", node=node, parent1=parent1,
1486 parent2=parent2)
1486 parent2=parent2)
1487 self._afterlock(commithook)
1487 self._afterlock(commithook)
1488 return ret
1488 return ret
1489
1489
1490 @unfilteredmethod
1490 @unfilteredmethod
1491 def commitctx(self, ctx, error=False):
1491 def commitctx(self, ctx, error=False):
1492 """Add a new revision to current repository.
1492 """Add a new revision to current repository.
1493 Revision information is passed via the context argument.
1493 Revision information is passed via the context argument.
1494 """
1494 """
1495
1495
1496 tr = None
1496 tr = None
1497 p1, p2 = ctx.p1(), ctx.p2()
1497 p1, p2 = ctx.p1(), ctx.p2()
1498 user = ctx.user()
1498 user = ctx.user()
1499
1499
1500 lock = self.lock()
1500 lock = self.lock()
1501 try:
1501 try:
1502 tr = self.transaction("commit")
1502 tr = self.transaction("commit")
1503 trp = weakref.proxy(tr)
1503 trp = weakref.proxy(tr)
1504
1504
1505 if ctx.files():
1505 if ctx.files():
1506 m1 = p1.manifest()
1506 m1 = p1.manifest()
1507 m2 = p2.manifest()
1507 m2 = p2.manifest()
1508 m = m1.copy()
1508 m = m1.copy()
1509
1509
1510 # check in files
1510 # check in files
1511 added = []
1511 added = []
1512 changed = []
1512 changed = []
1513 removed = list(ctx.removed())
1513 removed = list(ctx.removed())
1514 linkrev = len(self)
1514 linkrev = len(self)
1515 self.ui.note(_("committing files:\n"))
1515 self.ui.note(_("committing files:\n"))
1516 for f in sorted(ctx.modified() + ctx.added()):
1516 for f in sorted(ctx.modified() + ctx.added()):
1517 self.ui.note(f + "\n")
1517 self.ui.note(f + "\n")
1518 try:
1518 try:
1519 fctx = ctx[f]
1519 fctx = ctx[f]
1520 if fctx is None:
1520 if fctx is None:
1521 removed.append(f)
1521 removed.append(f)
1522 else:
1522 else:
1523 added.append(f)
1523 added.append(f)
1524 m[f] = self._filecommit(fctx, m1, m2, linkrev,
1524 m[f] = self._filecommit(fctx, m1, m2, linkrev,
1525 trp, changed)
1525 trp, changed)
1526 m.setflag(f, fctx.flags())
1526 m.setflag(f, fctx.flags())
1527 except OSError, inst:
1527 except OSError, inst:
1528 self.ui.warn(_("trouble committing %s!\n") % f)
1528 self.ui.warn(_("trouble committing %s!\n") % f)
1529 raise
1529 raise
1530 except IOError, inst:
1530 except IOError, inst:
1531 errcode = getattr(inst, 'errno', errno.ENOENT)
1531 errcode = getattr(inst, 'errno', errno.ENOENT)
1532 if error or errcode and errcode != errno.ENOENT:
1532 if error or errcode and errcode != errno.ENOENT:
1533 self.ui.warn(_("trouble committing %s!\n") % f)
1533 self.ui.warn(_("trouble committing %s!\n") % f)
1534 raise
1534 raise
1535
1535
1536 # update manifest
1536 # update manifest
1537 self.ui.note(_("committing manifest\n"))
1537 self.ui.note(_("committing manifest\n"))
1538 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1538 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1539 drop = [f for f in removed if f in m]
1539 drop = [f for f in removed if f in m]
1540 for f in drop:
1540 for f in drop:
1541 del m[f]
1541 del m[f]
1542 mn = self.manifest.add(m, trp, linkrev,
1542 mn = self.manifest.add(m, trp, linkrev,
1543 p1.manifestnode(), p2.manifestnode(),
1543 p1.manifestnode(), p2.manifestnode(),
1544 added, drop)
1544 added, drop)
1545 files = changed + removed
1545 files = changed + removed
1546 else:
1546 else:
1547 mn = p1.manifestnode()
1547 mn = p1.manifestnode()
1548 files = []
1548 files = []
1549
1549
1550 # update changelog
1550 # update changelog
1551 self.ui.note(_("committing changelog\n"))
1551 self.ui.note(_("committing changelog\n"))
1552 self.changelog.delayupdate(tr)
1552 self.changelog.delayupdate(tr)
1553 n = self.changelog.add(mn, files, ctx.description(),
1553 n = self.changelog.add(mn, files, ctx.description(),
1554 trp, p1.node(), p2.node(),
1554 trp, p1.node(), p2.node(),
1555 user, ctx.date(), ctx.extra().copy())
1555 user, ctx.date(), ctx.extra().copy())
1556 p = lambda: tr.writepending() and self.root or ""
1556 p = lambda: tr.writepending() and self.root or ""
1557 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1557 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1558 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1558 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1559 parent2=xp2, pending=p)
1559 parent2=xp2, pending=p)
1560 # set the new commit is proper phase
1560 # set the new commit is proper phase
1561 targetphase = subrepo.newcommitphase(self.ui, ctx)
1561 targetphase = subrepo.newcommitphase(self.ui, ctx)
1562 if targetphase:
1562 if targetphase:
1563 # retract boundary do not alter parent changeset.
1563 # retract boundary do not alter parent changeset.
1564 # if a parent have higher the resulting phase will
1564 # if a parent have higher the resulting phase will
1565 # be compliant anyway
1565 # be compliant anyway
1566 #
1566 #
1567 # if minimal phase was 0 we don't need to retract anything
1567 # if minimal phase was 0 we don't need to retract anything
1568 phases.retractboundary(self, tr, targetphase, [n])
1568 phases.retractboundary(self, tr, targetphase, [n])
1569 tr.close()
1569 tr.close()
1570 branchmap.updatecache(self.filtered('served'))
1570 branchmap.updatecache(self.filtered('served'))
1571 return n
1571 return n
1572 finally:
1572 finally:
1573 if tr:
1573 if tr:
1574 tr.release()
1574 tr.release()
1575 lock.release()
1575 lock.release()
1576
1576
1577 @unfilteredmethod
1577 @unfilteredmethod
1578 def destroying(self):
1578 def destroying(self):
1579 '''Inform the repository that nodes are about to be destroyed.
1579 '''Inform the repository that nodes are about to be destroyed.
1580 Intended for use by strip and rollback, so there's a common
1580 Intended for use by strip and rollback, so there's a common
1581 place for anything that has to be done before destroying history.
1581 place for anything that has to be done before destroying history.
1582
1582
1583 This is mostly useful for saving state that is in memory and waiting
1583 This is mostly useful for saving state that is in memory and waiting
1584 to be flushed when the current lock is released. Because a call to
1584 to be flushed when the current lock is released. Because a call to
1585 destroyed is imminent, the repo will be invalidated causing those
1585 destroyed is imminent, the repo will be invalidated causing those
1586 changes to stay in memory (waiting for the next unlock), or vanish
1586 changes to stay in memory (waiting for the next unlock), or vanish
1587 completely.
1587 completely.
1588 '''
1588 '''
1589 # When using the same lock to commit and strip, the phasecache is left
1589 # When using the same lock to commit and strip, the phasecache is left
1590 # dirty after committing. Then when we strip, the repo is invalidated,
1590 # dirty after committing. Then when we strip, the repo is invalidated,
1591 # causing those changes to disappear.
1591 # causing those changes to disappear.
1592 if '_phasecache' in vars(self):
1592 if '_phasecache' in vars(self):
1593 self._phasecache.write()
1593 self._phasecache.write()
1594
1594
1595 @unfilteredmethod
1595 @unfilteredmethod
1596 def destroyed(self):
1596 def destroyed(self):
1597 '''Inform the repository that nodes have been destroyed.
1597 '''Inform the repository that nodes have been destroyed.
1598 Intended for use by strip and rollback, so there's a common
1598 Intended for use by strip and rollback, so there's a common
1599 place for anything that has to be done after destroying history.
1599 place for anything that has to be done after destroying history.
1600 '''
1600 '''
1601 # When one tries to:
1601 # When one tries to:
1602 # 1) destroy nodes thus calling this method (e.g. strip)
1602 # 1) destroy nodes thus calling this method (e.g. strip)
1603 # 2) use phasecache somewhere (e.g. commit)
1603 # 2) use phasecache somewhere (e.g. commit)
1604 #
1604 #
1605 # then 2) will fail because the phasecache contains nodes that were
1605 # then 2) will fail because the phasecache contains nodes that were
1606 # removed. We can either remove phasecache from the filecache,
1606 # removed. We can either remove phasecache from the filecache,
1607 # causing it to reload next time it is accessed, or simply filter
1607 # causing it to reload next time it is accessed, or simply filter
1608 # the removed nodes now and write the updated cache.
1608 # the removed nodes now and write the updated cache.
1609 self._phasecache.filterunknown(self)
1609 self._phasecache.filterunknown(self)
1610 self._phasecache.write()
1610 self._phasecache.write()
1611
1611
1612 # update the 'served' branch cache to help read only server process
1612 # update the 'served' branch cache to help read only server process
1613 # Thanks to branchcache collaboration this is done from the nearest
1613 # Thanks to branchcache collaboration this is done from the nearest
1614 # filtered subset and it is expected to be fast.
1614 # filtered subset and it is expected to be fast.
1615 branchmap.updatecache(self.filtered('served'))
1615 branchmap.updatecache(self.filtered('served'))
1616
1616
1617 # Ensure the persistent tag cache is updated. Doing it now
1617 # Ensure the persistent tag cache is updated. Doing it now
1618 # means that the tag cache only has to worry about destroyed
1618 # means that the tag cache only has to worry about destroyed
1619 # heads immediately after a strip/rollback. That in turn
1619 # heads immediately after a strip/rollback. That in turn
1620 # guarantees that "cachetip == currenttip" (comparing both rev
1620 # guarantees that "cachetip == currenttip" (comparing both rev
1621 # and node) always means no nodes have been added or destroyed.
1621 # and node) always means no nodes have been added or destroyed.
1622
1622
1623 # XXX this is suboptimal when qrefresh'ing: we strip the current
1623 # XXX this is suboptimal when qrefresh'ing: we strip the current
1624 # head, refresh the tag cache, then immediately add a new head.
1624 # head, refresh the tag cache, then immediately add a new head.
1625 # But I think doing it this way is necessary for the "instant
1625 # But I think doing it this way is necessary for the "instant
1626 # tag cache retrieval" case to work.
1626 # tag cache retrieval" case to work.
1627 self.invalidate()
1627 self.invalidate()
1628
1628
1629 def walk(self, match, node=None):
1629 def walk(self, match, node=None):
1630 '''
1630 '''
1631 walk recursively through the directory tree or a given
1631 walk recursively through the directory tree or a given
1632 changeset, finding all files matched by the match
1632 changeset, finding all files matched by the match
1633 function
1633 function
1634 '''
1634 '''
1635 return self[node].walk(match)
1635 return self[node].walk(match)
1636
1636
1637 def status(self, node1='.', node2=None, match=None,
1637 def status(self, node1='.', node2=None, match=None,
1638 ignored=False, clean=False, unknown=False,
1638 ignored=False, clean=False, unknown=False,
1639 listsubrepos=False):
1639 listsubrepos=False):
1640 '''a convenience method that calls node1.status(node2)'''
1640 '''a convenience method that calls node1.status(node2)'''
1641 return self[node1].status(node2, match, ignored, clean, unknown,
1641 return self[node1].status(node2, match, ignored, clean, unknown,
1642 listsubrepos)
1642 listsubrepos)
1643
1643
1644 def heads(self, start=None):
1644 def heads(self, start=None):
1645 heads = self.changelog.heads(start)
1645 heads = self.changelog.heads(start)
1646 # sort the output in rev descending order
1646 # sort the output in rev descending order
1647 return sorted(heads, key=self.changelog.rev, reverse=True)
1647 return sorted(heads, key=self.changelog.rev, reverse=True)
1648
1648
1649 def branchheads(self, branch=None, start=None, closed=False):
1649 def branchheads(self, branch=None, start=None, closed=False):
1650 '''return a (possibly filtered) list of heads for the given branch
1650 '''return a (possibly filtered) list of heads for the given branch
1651
1651
1652 Heads are returned in topological order, from newest to oldest.
1652 Heads are returned in topological order, from newest to oldest.
1653 If branch is None, use the dirstate branch.
1653 If branch is None, use the dirstate branch.
1654 If start is not None, return only heads reachable from start.
1654 If start is not None, return only heads reachable from start.
1655 If closed is True, return heads that are marked as closed as well.
1655 If closed is True, return heads that are marked as closed as well.
1656 '''
1656 '''
1657 if branch is None:
1657 if branch is None:
1658 branch = self[None].branch()
1658 branch = self[None].branch()
1659 branches = self.branchmap()
1659 branches = self.branchmap()
1660 if branch not in branches:
1660 if branch not in branches:
1661 return []
1661 return []
1662 # the cache returns heads ordered lowest to highest
1662 # the cache returns heads ordered lowest to highest
1663 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1663 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1664 if start is not None:
1664 if start is not None:
1665 # filter out the heads that cannot be reached from startrev
1665 # filter out the heads that cannot be reached from startrev
1666 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1666 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1667 bheads = [h for h in bheads if h in fbheads]
1667 bheads = [h for h in bheads if h in fbheads]
1668 return bheads
1668 return bheads
1669
1669
1670 def branches(self, nodes):
1670 def branches(self, nodes):
1671 if not nodes:
1671 if not nodes:
1672 nodes = [self.changelog.tip()]
1672 nodes = [self.changelog.tip()]
1673 b = []
1673 b = []
1674 for n in nodes:
1674 for n in nodes:
1675 t = n
1675 t = n
1676 while True:
1676 while True:
1677 p = self.changelog.parents(n)
1677 p = self.changelog.parents(n)
1678 if p[1] != nullid or p[0] == nullid:
1678 if p[1] != nullid or p[0] == nullid:
1679 b.append((t, n, p[0], p[1]))
1679 b.append((t, n, p[0], p[1]))
1680 break
1680 break
1681 n = p[0]
1681 n = p[0]
1682 return b
1682 return b
1683
1683
1684 def between(self, pairs):
1684 def between(self, pairs):
1685 r = []
1685 r = []
1686
1686
1687 for top, bottom in pairs:
1687 for top, bottom in pairs:
1688 n, l, i = top, [], 0
1688 n, l, i = top, [], 0
1689 f = 1
1689 f = 1
1690
1690
1691 while n != bottom and n != nullid:
1691 while n != bottom and n != nullid:
1692 p = self.changelog.parents(n)[0]
1692 p = self.changelog.parents(n)[0]
1693 if i == f:
1693 if i == f:
1694 l.append(n)
1694 l.append(n)
1695 f = f * 2
1695 f = f * 2
1696 n = p
1696 n = p
1697 i += 1
1697 i += 1
1698
1698
1699 r.append(l)
1699 r.append(l)
1700
1700
1701 return r
1701 return r
1702
1702
1703 def checkpush(self, pushop):
1703 def checkpush(self, pushop):
1704 """Extensions can override this function if additional checks have
1704 """Extensions can override this function if additional checks have
1705 to be performed before pushing, or call it if they override push
1705 to be performed before pushing, or call it if they override push
1706 command.
1706 command.
1707 """
1707 """
1708 pass
1708 pass
1709
1709
1710 @unfilteredpropertycache
1710 @unfilteredpropertycache
1711 def prepushoutgoinghooks(self):
1711 def prepushoutgoinghooks(self):
1712 """Return util.hooks consists of "(repo, remote, outgoing)"
1712 """Return util.hooks consists of "(repo, remote, outgoing)"
1713 functions, which are called before pushing changesets.
1713 functions, which are called before pushing changesets.
1714 """
1714 """
1715 return util.hooks()
1715 return util.hooks()
1716
1716
1717 def stream_in(self, remote, requirements):
1717 def stream_in(self, remote, requirements):
1718 lock = self.lock()
1718 lock = self.lock()
1719 try:
1719 try:
1720 # Save remote branchmap. We will use it later
1720 # Save remote branchmap. We will use it later
1721 # to speed up branchcache creation
1721 # to speed up branchcache creation
1722 rbranchmap = None
1722 rbranchmap = None
1723 if remote.capable("branchmap"):
1723 if remote.capable("branchmap"):
1724 rbranchmap = remote.branchmap()
1724 rbranchmap = remote.branchmap()
1725
1725
1726 fp = remote.stream_out()
1726 fp = remote.stream_out()
1727 l = fp.readline()
1727 l = fp.readline()
1728 try:
1728 try:
1729 resp = int(l)
1729 resp = int(l)
1730 except ValueError:
1730 except ValueError:
1731 raise error.ResponseError(
1731 raise error.ResponseError(
1732 _('unexpected response from remote server:'), l)
1732 _('unexpected response from remote server:'), l)
1733 if resp == 1:
1733 if resp == 1:
1734 raise util.Abort(_('operation forbidden by server'))
1734 raise util.Abort(_('operation forbidden by server'))
1735 elif resp == 2:
1735 elif resp == 2:
1736 raise util.Abort(_('locking the remote repository failed'))
1736 raise util.Abort(_('locking the remote repository failed'))
1737 elif resp != 0:
1737 elif resp != 0:
1738 raise util.Abort(_('the server sent an unknown error code'))
1738 raise util.Abort(_('the server sent an unknown error code'))
1739 self.ui.status(_('streaming all changes\n'))
1739 self.ui.status(_('streaming all changes\n'))
1740 l = fp.readline()
1740 l = fp.readline()
1741 try:
1741 try:
1742 total_files, total_bytes = map(int, l.split(' ', 1))
1742 total_files, total_bytes = map(int, l.split(' ', 1))
1743 except (ValueError, TypeError):
1743 except (ValueError, TypeError):
1744 raise error.ResponseError(
1744 raise error.ResponseError(
1745 _('unexpected response from remote server:'), l)
1745 _('unexpected response from remote server:'), l)
1746 self.ui.status(_('%d files to transfer, %s of data\n') %
1746 self.ui.status(_('%d files to transfer, %s of data\n') %
1747 (total_files, util.bytecount(total_bytes)))
1747 (total_files, util.bytecount(total_bytes)))
1748 handled_bytes = 0
1748 handled_bytes = 0
1749 self.ui.progress(_('clone'), 0, total=total_bytes)
1749 self.ui.progress(_('clone'), 0, total=total_bytes)
1750 start = time.time()
1750 start = time.time()
1751
1751
1752 tr = self.transaction(_('clone'))
1752 tr = self.transaction(_('clone'))
1753 try:
1753 try:
1754 for i in xrange(total_files):
1754 for i in xrange(total_files):
1755 # XXX doesn't support '\n' or '\r' in filenames
1755 # XXX doesn't support '\n' or '\r' in filenames
1756 l = fp.readline()
1756 l = fp.readline()
1757 try:
1757 try:
1758 name, size = l.split('\0', 1)
1758 name, size = l.split('\0', 1)
1759 size = int(size)
1759 size = int(size)
1760 except (ValueError, TypeError):
1760 except (ValueError, TypeError):
1761 raise error.ResponseError(
1761 raise error.ResponseError(
1762 _('unexpected response from remote server:'), l)
1762 _('unexpected response from remote server:'), l)
1763 if self.ui.debugflag:
1763 if self.ui.debugflag:
1764 self.ui.debug('adding %s (%s)\n' %
1764 self.ui.debug('adding %s (%s)\n' %
1765 (name, util.bytecount(size)))
1765 (name, util.bytecount(size)))
1766 # for backwards compat, name was partially encoded
1766 # for backwards compat, name was partially encoded
1767 ofp = self.svfs(store.decodedir(name), 'w')
1767 ofp = self.svfs(store.decodedir(name), 'w')
1768 for chunk in util.filechunkiter(fp, limit=size):
1768 for chunk in util.filechunkiter(fp, limit=size):
1769 handled_bytes += len(chunk)
1769 handled_bytes += len(chunk)
1770 self.ui.progress(_('clone'), handled_bytes,
1770 self.ui.progress(_('clone'), handled_bytes,
1771 total=total_bytes)
1771 total=total_bytes)
1772 ofp.write(chunk)
1772 ofp.write(chunk)
1773 ofp.close()
1773 ofp.close()
1774 tr.close()
1774 tr.close()
1775 finally:
1775 finally:
1776 tr.release()
1776 tr.release()
1777
1777
1778 # Writing straight to files circumvented the inmemory caches
1778 # Writing straight to files circumvented the inmemory caches
1779 self.invalidate()
1779 self.invalidate()
1780
1780
1781 elapsed = time.time() - start
1781 elapsed = time.time() - start
1782 if elapsed <= 0:
1782 if elapsed <= 0:
1783 elapsed = 0.001
1783 elapsed = 0.001
1784 self.ui.progress(_('clone'), None)
1784 self.ui.progress(_('clone'), None)
1785 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1785 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1786 (util.bytecount(total_bytes), elapsed,
1786 (util.bytecount(total_bytes), elapsed,
1787 util.bytecount(total_bytes / elapsed)))
1787 util.bytecount(total_bytes / elapsed)))
1788
1788
1789 # new requirements = old non-format requirements +
1789 # new requirements = old non-format requirements +
1790 # new format-related
1790 # new format-related
1791 # requirements from the streamed-in repository
1791 # requirements from the streamed-in repository
1792 requirements.update(set(self.requirements) - self.supportedformats)
1792 requirements.update(set(self.requirements) - self.supportedformats)
1793 self._applyrequirements(requirements)
1793 self._applyrequirements(requirements)
1794 self._writerequirements()
1794 self._writerequirements()
1795
1795
1796 if rbranchmap:
1796 if rbranchmap:
1797 rbheads = []
1797 rbheads = []
1798 closed = []
1798 closed = []
1799 for bheads in rbranchmap.itervalues():
1799 for bheads in rbranchmap.itervalues():
1800 rbheads.extend(bheads)
1800 rbheads.extend(bheads)
1801 for h in bheads:
1801 for h in bheads:
1802 r = self.changelog.rev(h)
1802 r = self.changelog.rev(h)
1803 b, c = self.changelog.branchinfo(r)
1803 b, c = self.changelog.branchinfo(r)
1804 if c:
1804 if c:
1805 closed.append(h)
1805 closed.append(h)
1806
1806
1807 if rbheads:
1807 if rbheads:
1808 rtiprev = max((int(self.changelog.rev(node))
1808 rtiprev = max((int(self.changelog.rev(node))
1809 for node in rbheads))
1809 for node in rbheads))
1810 cache = branchmap.branchcache(rbranchmap,
1810 cache = branchmap.branchcache(rbranchmap,
1811 self[rtiprev].node(),
1811 self[rtiprev].node(),
1812 rtiprev,
1812 rtiprev,
1813 closednodes=closed)
1813 closednodes=closed)
1814 # Try to stick it as low as possible
1814 # Try to stick it as low as possible
1815 # filter above served are unlikely to be fetch from a clone
1815 # filter above served are unlikely to be fetch from a clone
1816 for candidate in ('base', 'immutable', 'served'):
1816 for candidate in ('base', 'immutable', 'served'):
1817 rview = self.filtered(candidate)
1817 rview = self.filtered(candidate)
1818 if cache.validfor(rview):
1818 if cache.validfor(rview):
1819 self._branchcaches[candidate] = cache
1819 self._branchcaches[candidate] = cache
1820 cache.write(rview)
1820 cache.write(rview)
1821 break
1821 break
1822 self.invalidate()
1822 self.invalidate()
1823 return len(self.heads()) + 1
1823 return len(self.heads()) + 1
1824 finally:
1824 finally:
1825 lock.release()
1825 lock.release()
1826
1826
1827 def clone(self, remote, heads=[], stream=None):
1827 def clone(self, remote, heads=[], stream=None):
1828 '''clone remote repository.
1828 '''clone remote repository.
1829
1829
1830 keyword arguments:
1830 keyword arguments:
1831 heads: list of revs to clone (forces use of pull)
1831 heads: list of revs to clone (forces use of pull)
1832 stream: use streaming clone if possible'''
1832 stream: use streaming clone if possible'''
1833
1833
1834 # now, all clients that can request uncompressed clones can
1834 # now, all clients that can request uncompressed clones can
1835 # read repo formats supported by all servers that can serve
1835 # read repo formats supported by all servers that can serve
1836 # them.
1836 # them.
1837
1837
1838 # if revlog format changes, client will have to check version
1838 # if revlog format changes, client will have to check version
1839 # and format flags on "stream" capability, and use
1839 # and format flags on "stream" capability, and use
1840 # uncompressed only if compatible.
1840 # uncompressed only if compatible.
1841
1841
1842 if stream is None:
1842 if stream is None:
1843 # if the server explicitly prefers to stream (for fast LANs)
1843 # if the server explicitly prefers to stream (for fast LANs)
1844 stream = remote.capable('stream-preferred')
1844 stream = remote.capable('stream-preferred')
1845
1845
1846 if stream and not heads:
1846 if stream and not heads:
1847 # 'stream' means remote revlog format is revlogv1 only
1847 # 'stream' means remote revlog format is revlogv1 only
1848 if remote.capable('stream'):
1848 if remote.capable('stream'):
1849 self.stream_in(remote, set(('revlogv1',)))
1849 self.stream_in(remote, set(('revlogv1',)))
1850 else:
1850 else:
1851 # otherwise, 'streamreqs' contains the remote revlog format
1851 # otherwise, 'streamreqs' contains the remote revlog format
1852 streamreqs = remote.capable('streamreqs')
1852 streamreqs = remote.capable('streamreqs')
1853 if streamreqs:
1853 if streamreqs:
1854 streamreqs = set(streamreqs.split(','))
1854 streamreqs = set(streamreqs.split(','))
1855 # if we support it, stream in and adjust our requirements
1855 # if we support it, stream in and adjust our requirements
1856 if not streamreqs - self.supportedformats:
1856 if not streamreqs - self.supportedformats:
1857 self.stream_in(remote, streamreqs)
1857 self.stream_in(remote, streamreqs)
1858
1858
1859 quiet = self.ui.backupconfig('ui', 'quietbookmarkmove')
1859 quiet = self.ui.backupconfig('ui', 'quietbookmarkmove')
1860 try:
1860 try:
1861 self.ui.setconfig('ui', 'quietbookmarkmove', True, 'clone')
1861 self.ui.setconfig('ui', 'quietbookmarkmove', True, 'clone')
1862 ret = exchange.pull(self, remote, heads).cgresult
1862 ret = exchange.pull(self, remote, heads).cgresult
1863 finally:
1863 finally:
1864 self.ui.restoreconfig(quiet)
1864 self.ui.restoreconfig(quiet)
1865 return ret
1865 return ret
1866
1866
1867 def pushkey(self, namespace, key, old, new):
1867 def pushkey(self, namespace, key, old, new):
1868 try:
1868 try:
1869 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1869 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1870 old=old, new=new)
1870 old=old, new=new)
1871 except error.HookAbort, exc:
1871 except error.HookAbort, exc:
1872 self.ui.write_err(_("pushkey-abort: %s\n") % exc)
1872 self.ui.write_err(_("pushkey-abort: %s\n") % exc)
1873 if exc.hint:
1873 if exc.hint:
1874 self.ui.write_err(_("(%s)\n") % exc.hint)
1874 self.ui.write_err(_("(%s)\n") % exc.hint)
1875 return False
1875 return False
1876 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1876 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1877 ret = pushkey.push(self, namespace, key, old, new)
1877 ret = pushkey.push(self, namespace, key, old, new)
1878 def runhook():
1878 def runhook():
1879 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1879 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1880 ret=ret)
1880 ret=ret)
1881 self._afterlock(runhook)
1881 self._afterlock(runhook)
1882 return ret
1882 return ret
1883
1883
1884 def listkeys(self, namespace):
1884 def listkeys(self, namespace):
1885 self.hook('prelistkeys', throw=True, namespace=namespace)
1885 self.hook('prelistkeys', throw=True, namespace=namespace)
1886 self.ui.debug('listing keys for "%s"\n' % namespace)
1886 self.ui.debug('listing keys for "%s"\n' % namespace)
1887 values = pushkey.list(self, namespace)
1887 values = pushkey.list(self, namespace)
1888 self.hook('listkeys', namespace=namespace, values=values)
1888 self.hook('listkeys', namespace=namespace, values=values)
1889 return values
1889 return values
1890
1890
1891 def debugwireargs(self, one, two, three=None, four=None, five=None):
1891 def debugwireargs(self, one, two, three=None, four=None, five=None):
1892 '''used to test argument passing over the wire'''
1892 '''used to test argument passing over the wire'''
1893 return "%s %s %s %s %s" % (one, two, three, four, five)
1893 return "%s %s %s %s %s" % (one, two, three, four, five)
1894
1894
1895 def savecommitmessage(self, text):
1895 def savecommitmessage(self, text):
1896 fp = self.vfs('last-message.txt', 'wb')
1896 fp = self.vfs('last-message.txt', 'wb')
1897 try:
1897 try:
1898 fp.write(text)
1898 fp.write(text)
1899 finally:
1899 finally:
1900 fp.close()
1900 fp.close()
1901 return self.pathto(fp.name[len(self.root) + 1:])
1901 return self.pathto(fp.name[len(self.root) + 1:])
1902
1902
1903 # used to avoid circular references so destructors work
1903 # used to avoid circular references so destructors work
1904 def aftertrans(files):
1904 def aftertrans(files):
1905 renamefiles = [tuple(t) for t in files]
1905 renamefiles = [tuple(t) for t in files]
1906 def a():
1906 def a():
1907 for vfs, src, dest in renamefiles:
1907 for vfs, src, dest in renamefiles:
1908 try:
1908 try:
1909 vfs.rename(src, dest)
1909 vfs.rename(src, dest)
1910 except OSError: # journal file does not yet exist
1910 except OSError: # journal file does not yet exist
1911 pass
1911 pass
1912 return a
1912 return a
1913
1913
1914 def undoname(fn):
1914 def undoname(fn):
1915 base, name = os.path.split(fn)
1915 base, name = os.path.split(fn)
1916 assert name.startswith('journal')
1916 assert name.startswith('journal')
1917 return os.path.join(base, name.replace('journal', 'undo', 1))
1917 return os.path.join(base, name.replace('journal', 'undo', 1))
1918
1918
1919 def instance(ui, path, create):
1919 def instance(ui, path, create):
1920 return localrepository(ui, util.urllocalpath(path), create)
1920 return localrepository(ui, util.urllocalpath(path), create)
1921
1921
1922 def islocal(path):
1922 def islocal(path):
1923 return True
1923 return True
General Comments 0
You need to be logged in to leave comments. Login now