##// END OF EJS Templates
locarepo: remove the `push` method (API)...
Pierre-Yves David -
r22620:4d52e6eb default
parent child Browse files
Show More
@@ -1,1780 +1,1777 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import urllib
9 import urllib
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
12 import lock as lockmod
12 import lock as lockmod
13 import transaction, store, encoding, exchange, bundle2
13 import transaction, store, encoding, exchange, bundle2
14 import scmutil, util, extensions, hook, error, revset
14 import scmutil, util, extensions, hook, error, revset
15 import match as matchmod
15 import match as matchmod
16 import merge as mergemod
16 import merge as mergemod
17 import tags as tagsmod
17 import tags as tagsmod
18 from lock import release
18 from lock import release
19 import weakref, errno, os, time, inspect
19 import weakref, errno, os, time, inspect
20 import branchmap, pathutil
20 import branchmap, pathutil
21 propertycache = util.propertycache
21 propertycache = util.propertycache
22 filecache = scmutil.filecache
22 filecache = scmutil.filecache
23
23
24 class repofilecache(filecache):
24 class repofilecache(filecache):
25 """All filecache usage on repo are done for logic that should be unfiltered
25 """All filecache usage on repo are done for logic that should be unfiltered
26 """
26 """
27
27
28 def __get__(self, repo, type=None):
28 def __get__(self, repo, type=None):
29 return super(repofilecache, self).__get__(repo.unfiltered(), type)
29 return super(repofilecache, self).__get__(repo.unfiltered(), type)
30 def __set__(self, repo, value):
30 def __set__(self, repo, value):
31 return super(repofilecache, self).__set__(repo.unfiltered(), value)
31 return super(repofilecache, self).__set__(repo.unfiltered(), value)
32 def __delete__(self, repo):
32 def __delete__(self, repo):
33 return super(repofilecache, self).__delete__(repo.unfiltered())
33 return super(repofilecache, self).__delete__(repo.unfiltered())
34
34
35 class storecache(repofilecache):
35 class storecache(repofilecache):
36 """filecache for files in the store"""
36 """filecache for files in the store"""
37 def join(self, obj, fname):
37 def join(self, obj, fname):
38 return obj.sjoin(fname)
38 return obj.sjoin(fname)
39
39
40 class unfilteredpropertycache(propertycache):
40 class unfilteredpropertycache(propertycache):
41 """propertycache that apply to unfiltered repo only"""
41 """propertycache that apply to unfiltered repo only"""
42
42
43 def __get__(self, repo, type=None):
43 def __get__(self, repo, type=None):
44 unfi = repo.unfiltered()
44 unfi = repo.unfiltered()
45 if unfi is repo:
45 if unfi is repo:
46 return super(unfilteredpropertycache, self).__get__(unfi)
46 return super(unfilteredpropertycache, self).__get__(unfi)
47 return getattr(unfi, self.name)
47 return getattr(unfi, self.name)
48
48
49 class filteredpropertycache(propertycache):
49 class filteredpropertycache(propertycache):
50 """propertycache that must take filtering in account"""
50 """propertycache that must take filtering in account"""
51
51
52 def cachevalue(self, obj, value):
52 def cachevalue(self, obj, value):
53 object.__setattr__(obj, self.name, value)
53 object.__setattr__(obj, self.name, value)
54
54
55
55
56 def hasunfilteredcache(repo, name):
56 def hasunfilteredcache(repo, name):
57 """check if a repo has an unfilteredpropertycache value for <name>"""
57 """check if a repo has an unfilteredpropertycache value for <name>"""
58 return name in vars(repo.unfiltered())
58 return name in vars(repo.unfiltered())
59
59
60 def unfilteredmethod(orig):
60 def unfilteredmethod(orig):
61 """decorate method that always need to be run on unfiltered version"""
61 """decorate method that always need to be run on unfiltered version"""
62 def wrapper(repo, *args, **kwargs):
62 def wrapper(repo, *args, **kwargs):
63 return orig(repo.unfiltered(), *args, **kwargs)
63 return orig(repo.unfiltered(), *args, **kwargs)
64 return wrapper
64 return wrapper
65
65
66 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
66 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
67 'unbundle'))
67 'unbundle'))
68 legacycaps = moderncaps.union(set(['changegroupsubset']))
68 legacycaps = moderncaps.union(set(['changegroupsubset']))
69
69
70 class localpeer(peer.peerrepository):
70 class localpeer(peer.peerrepository):
71 '''peer for a local repo; reflects only the most recent API'''
71 '''peer for a local repo; reflects only the most recent API'''
72
72
73 def __init__(self, repo, caps=moderncaps):
73 def __init__(self, repo, caps=moderncaps):
74 peer.peerrepository.__init__(self)
74 peer.peerrepository.__init__(self)
75 self._repo = repo.filtered('served')
75 self._repo = repo.filtered('served')
76 self.ui = repo.ui
76 self.ui = repo.ui
77 self._caps = repo._restrictcapabilities(caps)
77 self._caps = repo._restrictcapabilities(caps)
78 self.requirements = repo.requirements
78 self.requirements = repo.requirements
79 self.supportedformats = repo.supportedformats
79 self.supportedformats = repo.supportedformats
80
80
81 def close(self):
81 def close(self):
82 self._repo.close()
82 self._repo.close()
83
83
84 def _capabilities(self):
84 def _capabilities(self):
85 return self._caps
85 return self._caps
86
86
87 def local(self):
87 def local(self):
88 return self._repo
88 return self._repo
89
89
90 def canpush(self):
90 def canpush(self):
91 return True
91 return True
92
92
93 def url(self):
93 def url(self):
94 return self._repo.url()
94 return self._repo.url()
95
95
96 def lookup(self, key):
96 def lookup(self, key):
97 return self._repo.lookup(key)
97 return self._repo.lookup(key)
98
98
99 def branchmap(self):
99 def branchmap(self):
100 return self._repo.branchmap()
100 return self._repo.branchmap()
101
101
102 def heads(self):
102 def heads(self):
103 return self._repo.heads()
103 return self._repo.heads()
104
104
105 def known(self, nodes):
105 def known(self, nodes):
106 return self._repo.known(nodes)
106 return self._repo.known(nodes)
107
107
108 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
108 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
109 format='HG10', **kwargs):
109 format='HG10', **kwargs):
110 cg = exchange.getbundle(self._repo, source, heads=heads,
110 cg = exchange.getbundle(self._repo, source, heads=heads,
111 common=common, bundlecaps=bundlecaps, **kwargs)
111 common=common, bundlecaps=bundlecaps, **kwargs)
112 if bundlecaps is not None and 'HG2X' in bundlecaps:
112 if bundlecaps is not None and 'HG2X' in bundlecaps:
113 # When requesting a bundle2, getbundle returns a stream to make the
113 # When requesting a bundle2, getbundle returns a stream to make the
114 # wire level function happier. We need to build a proper object
114 # wire level function happier. We need to build a proper object
115 # from it in local peer.
115 # from it in local peer.
116 cg = bundle2.unbundle20(self.ui, cg)
116 cg = bundle2.unbundle20(self.ui, cg)
117 return cg
117 return cg
118
118
119 # TODO We might want to move the next two calls into legacypeer and add
119 # TODO We might want to move the next two calls into legacypeer and add
120 # unbundle instead.
120 # unbundle instead.
121
121
122 def unbundle(self, cg, heads, url):
122 def unbundle(self, cg, heads, url):
123 """apply a bundle on a repo
123 """apply a bundle on a repo
124
124
125 This function handles the repo locking itself."""
125 This function handles the repo locking itself."""
126 try:
126 try:
127 cg = exchange.readbundle(self.ui, cg, None)
127 cg = exchange.readbundle(self.ui, cg, None)
128 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
128 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
129 if util.safehasattr(ret, 'getchunks'):
129 if util.safehasattr(ret, 'getchunks'):
130 # This is a bundle20 object, turn it into an unbundler.
130 # This is a bundle20 object, turn it into an unbundler.
131 # This little dance should be dropped eventually when the API
131 # This little dance should be dropped eventually when the API
132 # is finally improved.
132 # is finally improved.
133 stream = util.chunkbuffer(ret.getchunks())
133 stream = util.chunkbuffer(ret.getchunks())
134 ret = bundle2.unbundle20(self.ui, stream)
134 ret = bundle2.unbundle20(self.ui, stream)
135 return ret
135 return ret
136 except error.PushRaced, exc:
136 except error.PushRaced, exc:
137 raise error.ResponseError(_('push failed:'), str(exc))
137 raise error.ResponseError(_('push failed:'), str(exc))
138
138
139 def lock(self):
139 def lock(self):
140 return self._repo.lock()
140 return self._repo.lock()
141
141
142 def addchangegroup(self, cg, source, url):
142 def addchangegroup(self, cg, source, url):
143 return changegroup.addchangegroup(self._repo, cg, source, url)
143 return changegroup.addchangegroup(self._repo, cg, source, url)
144
144
145 def pushkey(self, namespace, key, old, new):
145 def pushkey(self, namespace, key, old, new):
146 return self._repo.pushkey(namespace, key, old, new)
146 return self._repo.pushkey(namespace, key, old, new)
147
147
148 def listkeys(self, namespace):
148 def listkeys(self, namespace):
149 return self._repo.listkeys(namespace)
149 return self._repo.listkeys(namespace)
150
150
151 def debugwireargs(self, one, two, three=None, four=None, five=None):
151 def debugwireargs(self, one, two, three=None, four=None, five=None):
152 '''used to test argument passing over the wire'''
152 '''used to test argument passing over the wire'''
153 return "%s %s %s %s %s" % (one, two, three, four, five)
153 return "%s %s %s %s %s" % (one, two, three, four, five)
154
154
155 class locallegacypeer(localpeer):
155 class locallegacypeer(localpeer):
156 '''peer extension which implements legacy methods too; used for tests with
156 '''peer extension which implements legacy methods too; used for tests with
157 restricted capabilities'''
157 restricted capabilities'''
158
158
159 def __init__(self, repo):
159 def __init__(self, repo):
160 localpeer.__init__(self, repo, caps=legacycaps)
160 localpeer.__init__(self, repo, caps=legacycaps)
161
161
162 def branches(self, nodes):
162 def branches(self, nodes):
163 return self._repo.branches(nodes)
163 return self._repo.branches(nodes)
164
164
165 def between(self, pairs):
165 def between(self, pairs):
166 return self._repo.between(pairs)
166 return self._repo.between(pairs)
167
167
168 def changegroup(self, basenodes, source):
168 def changegroup(self, basenodes, source):
169 return changegroup.changegroup(self._repo, basenodes, source)
169 return changegroup.changegroup(self._repo, basenodes, source)
170
170
171 def changegroupsubset(self, bases, heads, source):
171 def changegroupsubset(self, bases, heads, source):
172 return changegroup.changegroupsubset(self._repo, bases, heads, source)
172 return changegroup.changegroupsubset(self._repo, bases, heads, source)
173
173
174 class localrepository(object):
174 class localrepository(object):
175
175
176 supportedformats = set(('revlogv1', 'generaldelta'))
176 supportedformats = set(('revlogv1', 'generaldelta'))
177 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
177 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
178 'dotencode'))
178 'dotencode'))
179 openerreqs = set(('revlogv1', 'generaldelta'))
179 openerreqs = set(('revlogv1', 'generaldelta'))
180 requirements = ['revlogv1']
180 requirements = ['revlogv1']
181 filtername = None
181 filtername = None
182
182
183 # a list of (ui, featureset) functions.
183 # a list of (ui, featureset) functions.
184 # only functions defined in module of enabled extensions are invoked
184 # only functions defined in module of enabled extensions are invoked
185 featuresetupfuncs = set()
185 featuresetupfuncs = set()
186
186
187 def _baserequirements(self, create):
187 def _baserequirements(self, create):
188 return self.requirements[:]
188 return self.requirements[:]
189
189
190 def __init__(self, baseui, path=None, create=False):
190 def __init__(self, baseui, path=None, create=False):
191 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
191 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
192 self.wopener = self.wvfs
192 self.wopener = self.wvfs
193 self.root = self.wvfs.base
193 self.root = self.wvfs.base
194 self.path = self.wvfs.join(".hg")
194 self.path = self.wvfs.join(".hg")
195 self.origroot = path
195 self.origroot = path
196 self.auditor = pathutil.pathauditor(self.root, self._checknested)
196 self.auditor = pathutil.pathauditor(self.root, self._checknested)
197 self.vfs = scmutil.vfs(self.path)
197 self.vfs = scmutil.vfs(self.path)
198 self.opener = self.vfs
198 self.opener = self.vfs
199 self.baseui = baseui
199 self.baseui = baseui
200 self.ui = baseui.copy()
200 self.ui = baseui.copy()
201 self.ui.copy = baseui.copy # prevent copying repo configuration
201 self.ui.copy = baseui.copy # prevent copying repo configuration
202 # A list of callback to shape the phase if no data were found.
202 # A list of callback to shape the phase if no data were found.
203 # Callback are in the form: func(repo, roots) --> processed root.
203 # Callback are in the form: func(repo, roots) --> processed root.
204 # This list it to be filled by extension during repo setup
204 # This list it to be filled by extension during repo setup
205 self._phasedefaults = []
205 self._phasedefaults = []
206 try:
206 try:
207 self.ui.readconfig(self.join("hgrc"), self.root)
207 self.ui.readconfig(self.join("hgrc"), self.root)
208 extensions.loadall(self.ui)
208 extensions.loadall(self.ui)
209 except IOError:
209 except IOError:
210 pass
210 pass
211
211
212 if self.featuresetupfuncs:
212 if self.featuresetupfuncs:
213 self.supported = set(self._basesupported) # use private copy
213 self.supported = set(self._basesupported) # use private copy
214 extmods = set(m.__name__ for n, m
214 extmods = set(m.__name__ for n, m
215 in extensions.extensions(self.ui))
215 in extensions.extensions(self.ui))
216 for setupfunc in self.featuresetupfuncs:
216 for setupfunc in self.featuresetupfuncs:
217 if setupfunc.__module__ in extmods:
217 if setupfunc.__module__ in extmods:
218 setupfunc(self.ui, self.supported)
218 setupfunc(self.ui, self.supported)
219 else:
219 else:
220 self.supported = self._basesupported
220 self.supported = self._basesupported
221
221
222 if not self.vfs.isdir():
222 if not self.vfs.isdir():
223 if create:
223 if create:
224 if not self.wvfs.exists():
224 if not self.wvfs.exists():
225 self.wvfs.makedirs()
225 self.wvfs.makedirs()
226 self.vfs.makedir(notindexed=True)
226 self.vfs.makedir(notindexed=True)
227 requirements = self._baserequirements(create)
227 requirements = self._baserequirements(create)
228 if self.ui.configbool('format', 'usestore', True):
228 if self.ui.configbool('format', 'usestore', True):
229 self.vfs.mkdir("store")
229 self.vfs.mkdir("store")
230 requirements.append("store")
230 requirements.append("store")
231 if self.ui.configbool('format', 'usefncache', True):
231 if self.ui.configbool('format', 'usefncache', True):
232 requirements.append("fncache")
232 requirements.append("fncache")
233 if self.ui.configbool('format', 'dotencode', True):
233 if self.ui.configbool('format', 'dotencode', True):
234 requirements.append('dotencode')
234 requirements.append('dotencode')
235 # create an invalid changelog
235 # create an invalid changelog
236 self.vfs.append(
236 self.vfs.append(
237 "00changelog.i",
237 "00changelog.i",
238 '\0\0\0\2' # represents revlogv2
238 '\0\0\0\2' # represents revlogv2
239 ' dummy changelog to prevent using the old repo layout'
239 ' dummy changelog to prevent using the old repo layout'
240 )
240 )
241 if self.ui.configbool('format', 'generaldelta', False):
241 if self.ui.configbool('format', 'generaldelta', False):
242 requirements.append("generaldelta")
242 requirements.append("generaldelta")
243 requirements = set(requirements)
243 requirements = set(requirements)
244 else:
244 else:
245 raise error.RepoError(_("repository %s not found") % path)
245 raise error.RepoError(_("repository %s not found") % path)
246 elif create:
246 elif create:
247 raise error.RepoError(_("repository %s already exists") % path)
247 raise error.RepoError(_("repository %s already exists") % path)
248 else:
248 else:
249 try:
249 try:
250 requirements = scmutil.readrequires(self.vfs, self.supported)
250 requirements = scmutil.readrequires(self.vfs, self.supported)
251 except IOError, inst:
251 except IOError, inst:
252 if inst.errno != errno.ENOENT:
252 if inst.errno != errno.ENOENT:
253 raise
253 raise
254 requirements = set()
254 requirements = set()
255
255
256 self.sharedpath = self.path
256 self.sharedpath = self.path
257 try:
257 try:
258 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
258 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
259 realpath=True)
259 realpath=True)
260 s = vfs.base
260 s = vfs.base
261 if not vfs.exists():
261 if not vfs.exists():
262 raise error.RepoError(
262 raise error.RepoError(
263 _('.hg/sharedpath points to nonexistent directory %s') % s)
263 _('.hg/sharedpath points to nonexistent directory %s') % s)
264 self.sharedpath = s
264 self.sharedpath = s
265 except IOError, inst:
265 except IOError, inst:
266 if inst.errno != errno.ENOENT:
266 if inst.errno != errno.ENOENT:
267 raise
267 raise
268
268
269 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
269 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
270 self.spath = self.store.path
270 self.spath = self.store.path
271 self.svfs = self.store.vfs
271 self.svfs = self.store.vfs
272 self.sopener = self.svfs
272 self.sopener = self.svfs
273 self.sjoin = self.store.join
273 self.sjoin = self.store.join
274 self.vfs.createmode = self.store.createmode
274 self.vfs.createmode = self.store.createmode
275 self._applyrequirements(requirements)
275 self._applyrequirements(requirements)
276 if create:
276 if create:
277 self._writerequirements()
277 self._writerequirements()
278
278
279
279
280 self._branchcaches = {}
280 self._branchcaches = {}
281 self.filterpats = {}
281 self.filterpats = {}
282 self._datafilters = {}
282 self._datafilters = {}
283 self._transref = self._lockref = self._wlockref = None
283 self._transref = self._lockref = self._wlockref = None
284
284
285 # A cache for various files under .hg/ that tracks file changes,
285 # A cache for various files under .hg/ that tracks file changes,
286 # (used by the filecache decorator)
286 # (used by the filecache decorator)
287 #
287 #
288 # Maps a property name to its util.filecacheentry
288 # Maps a property name to its util.filecacheentry
289 self._filecache = {}
289 self._filecache = {}
290
290
291 # hold sets of revision to be filtered
291 # hold sets of revision to be filtered
292 # should be cleared when something might have changed the filter value:
292 # should be cleared when something might have changed the filter value:
293 # - new changesets,
293 # - new changesets,
294 # - phase change,
294 # - phase change,
295 # - new obsolescence marker,
295 # - new obsolescence marker,
296 # - working directory parent change,
296 # - working directory parent change,
297 # - bookmark changes
297 # - bookmark changes
298 self.filteredrevcache = {}
298 self.filteredrevcache = {}
299
299
300 def close(self):
300 def close(self):
301 pass
301 pass
302
302
303 def _restrictcapabilities(self, caps):
303 def _restrictcapabilities(self, caps):
304 # bundle2 is not ready for prime time, drop it unless explicitly
304 # bundle2 is not ready for prime time, drop it unless explicitly
305 # required by the tests (or some brave tester)
305 # required by the tests (or some brave tester)
306 if self.ui.configbool('experimental', 'bundle2-exp', False):
306 if self.ui.configbool('experimental', 'bundle2-exp', False):
307 caps = set(caps)
307 caps = set(caps)
308 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
308 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
309 caps.add('bundle2-exp=' + urllib.quote(capsblob))
309 caps.add('bundle2-exp=' + urllib.quote(capsblob))
310 return caps
310 return caps
311
311
312 def _applyrequirements(self, requirements):
312 def _applyrequirements(self, requirements):
313 self.requirements = requirements
313 self.requirements = requirements
314 self.sopener.options = dict((r, 1) for r in requirements
314 self.sopener.options = dict((r, 1) for r in requirements
315 if r in self.openerreqs)
315 if r in self.openerreqs)
316 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
316 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
317 if chunkcachesize is not None:
317 if chunkcachesize is not None:
318 self.sopener.options['chunkcachesize'] = chunkcachesize
318 self.sopener.options['chunkcachesize'] = chunkcachesize
319
319
320 def _writerequirements(self):
320 def _writerequirements(self):
321 reqfile = self.opener("requires", "w")
321 reqfile = self.opener("requires", "w")
322 for r in sorted(self.requirements):
322 for r in sorted(self.requirements):
323 reqfile.write("%s\n" % r)
323 reqfile.write("%s\n" % r)
324 reqfile.close()
324 reqfile.close()
325
325
326 def _checknested(self, path):
326 def _checknested(self, path):
327 """Determine if path is a legal nested repository."""
327 """Determine if path is a legal nested repository."""
328 if not path.startswith(self.root):
328 if not path.startswith(self.root):
329 return False
329 return False
330 subpath = path[len(self.root) + 1:]
330 subpath = path[len(self.root) + 1:]
331 normsubpath = util.pconvert(subpath)
331 normsubpath = util.pconvert(subpath)
332
332
333 # XXX: Checking against the current working copy is wrong in
333 # XXX: Checking against the current working copy is wrong in
334 # the sense that it can reject things like
334 # the sense that it can reject things like
335 #
335 #
336 # $ hg cat -r 10 sub/x.txt
336 # $ hg cat -r 10 sub/x.txt
337 #
337 #
338 # if sub/ is no longer a subrepository in the working copy
338 # if sub/ is no longer a subrepository in the working copy
339 # parent revision.
339 # parent revision.
340 #
340 #
341 # However, it can of course also allow things that would have
341 # However, it can of course also allow things that would have
342 # been rejected before, such as the above cat command if sub/
342 # been rejected before, such as the above cat command if sub/
343 # is a subrepository now, but was a normal directory before.
343 # is a subrepository now, but was a normal directory before.
344 # The old path auditor would have rejected by mistake since it
344 # The old path auditor would have rejected by mistake since it
345 # panics when it sees sub/.hg/.
345 # panics when it sees sub/.hg/.
346 #
346 #
347 # All in all, checking against the working copy seems sensible
347 # All in all, checking against the working copy seems sensible
348 # since we want to prevent access to nested repositories on
348 # since we want to prevent access to nested repositories on
349 # the filesystem *now*.
349 # the filesystem *now*.
350 ctx = self[None]
350 ctx = self[None]
351 parts = util.splitpath(subpath)
351 parts = util.splitpath(subpath)
352 while parts:
352 while parts:
353 prefix = '/'.join(parts)
353 prefix = '/'.join(parts)
354 if prefix in ctx.substate:
354 if prefix in ctx.substate:
355 if prefix == normsubpath:
355 if prefix == normsubpath:
356 return True
356 return True
357 else:
357 else:
358 sub = ctx.sub(prefix)
358 sub = ctx.sub(prefix)
359 return sub.checknested(subpath[len(prefix) + 1:])
359 return sub.checknested(subpath[len(prefix) + 1:])
360 else:
360 else:
361 parts.pop()
361 parts.pop()
362 return False
362 return False
363
363
364 def peer(self):
364 def peer(self):
365 return localpeer(self) # not cached to avoid reference cycle
365 return localpeer(self) # not cached to avoid reference cycle
366
366
367 def unfiltered(self):
367 def unfiltered(self):
368 """Return unfiltered version of the repository
368 """Return unfiltered version of the repository
369
369
370 Intended to be overwritten by filtered repo."""
370 Intended to be overwritten by filtered repo."""
371 return self
371 return self
372
372
373 def filtered(self, name):
373 def filtered(self, name):
374 """Return a filtered version of a repository"""
374 """Return a filtered version of a repository"""
375 # build a new class with the mixin and the current class
375 # build a new class with the mixin and the current class
376 # (possibly subclass of the repo)
376 # (possibly subclass of the repo)
377 class proxycls(repoview.repoview, self.unfiltered().__class__):
377 class proxycls(repoview.repoview, self.unfiltered().__class__):
378 pass
378 pass
379 return proxycls(self, name)
379 return proxycls(self, name)
380
380
381 @repofilecache('bookmarks')
381 @repofilecache('bookmarks')
382 def _bookmarks(self):
382 def _bookmarks(self):
383 return bookmarks.bmstore(self)
383 return bookmarks.bmstore(self)
384
384
385 @repofilecache('bookmarks.current')
385 @repofilecache('bookmarks.current')
386 def _bookmarkcurrent(self):
386 def _bookmarkcurrent(self):
387 return bookmarks.readcurrent(self)
387 return bookmarks.readcurrent(self)
388
388
389 def bookmarkheads(self, bookmark):
389 def bookmarkheads(self, bookmark):
390 name = bookmark.split('@', 1)[0]
390 name = bookmark.split('@', 1)[0]
391 heads = []
391 heads = []
392 for mark, n in self._bookmarks.iteritems():
392 for mark, n in self._bookmarks.iteritems():
393 if mark.split('@', 1)[0] == name:
393 if mark.split('@', 1)[0] == name:
394 heads.append(n)
394 heads.append(n)
395 return heads
395 return heads
396
396
397 @storecache('phaseroots')
397 @storecache('phaseroots')
398 def _phasecache(self):
398 def _phasecache(self):
399 return phases.phasecache(self, self._phasedefaults)
399 return phases.phasecache(self, self._phasedefaults)
400
400
401 @storecache('obsstore')
401 @storecache('obsstore')
402 def obsstore(self):
402 def obsstore(self):
403 store = obsolete.obsstore(self.sopener)
403 store = obsolete.obsstore(self.sopener)
404 if store and not obsolete._enabled:
404 if store and not obsolete._enabled:
405 # message is rare enough to not be translated
405 # message is rare enough to not be translated
406 msg = 'obsolete feature not enabled but %i markers found!\n'
406 msg = 'obsolete feature not enabled but %i markers found!\n'
407 self.ui.warn(msg % len(list(store)))
407 self.ui.warn(msg % len(list(store)))
408 return store
408 return store
409
409
410 @storecache('00changelog.i')
410 @storecache('00changelog.i')
411 def changelog(self):
411 def changelog(self):
412 c = changelog.changelog(self.sopener)
412 c = changelog.changelog(self.sopener)
413 if 'HG_PENDING' in os.environ:
413 if 'HG_PENDING' in os.environ:
414 p = os.environ['HG_PENDING']
414 p = os.environ['HG_PENDING']
415 if p.startswith(self.root):
415 if p.startswith(self.root):
416 c.readpending('00changelog.i.a')
416 c.readpending('00changelog.i.a')
417 return c
417 return c
418
418
419 @storecache('00manifest.i')
419 @storecache('00manifest.i')
420 def manifest(self):
420 def manifest(self):
421 return manifest.manifest(self.sopener)
421 return manifest.manifest(self.sopener)
422
422
423 @repofilecache('dirstate')
423 @repofilecache('dirstate')
424 def dirstate(self):
424 def dirstate(self):
425 warned = [0]
425 warned = [0]
426 def validate(node):
426 def validate(node):
427 try:
427 try:
428 self.changelog.rev(node)
428 self.changelog.rev(node)
429 return node
429 return node
430 except error.LookupError:
430 except error.LookupError:
431 if not warned[0]:
431 if not warned[0]:
432 warned[0] = True
432 warned[0] = True
433 self.ui.warn(_("warning: ignoring unknown"
433 self.ui.warn(_("warning: ignoring unknown"
434 " working parent %s!\n") % short(node))
434 " working parent %s!\n") % short(node))
435 return nullid
435 return nullid
436
436
437 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
437 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
438
438
439 def __getitem__(self, changeid):
439 def __getitem__(self, changeid):
440 if changeid is None:
440 if changeid is None:
441 return context.workingctx(self)
441 return context.workingctx(self)
442 return context.changectx(self, changeid)
442 return context.changectx(self, changeid)
443
443
444 def __contains__(self, changeid):
444 def __contains__(self, changeid):
445 try:
445 try:
446 return bool(self.lookup(changeid))
446 return bool(self.lookup(changeid))
447 except error.RepoLookupError:
447 except error.RepoLookupError:
448 return False
448 return False
449
449
450 def __nonzero__(self):
450 def __nonzero__(self):
451 return True
451 return True
452
452
453 def __len__(self):
453 def __len__(self):
454 return len(self.changelog)
454 return len(self.changelog)
455
455
456 def __iter__(self):
456 def __iter__(self):
457 return iter(self.changelog)
457 return iter(self.changelog)
458
458
459 def revs(self, expr, *args):
459 def revs(self, expr, *args):
460 '''Return a list of revisions matching the given revset'''
460 '''Return a list of revisions matching the given revset'''
461 expr = revset.formatspec(expr, *args)
461 expr = revset.formatspec(expr, *args)
462 m = revset.match(None, expr)
462 m = revset.match(None, expr)
463 return m(self, revset.spanset(self))
463 return m(self, revset.spanset(self))
464
464
465 def set(self, expr, *args):
465 def set(self, expr, *args):
466 '''
466 '''
467 Yield a context for each matching revision, after doing arg
467 Yield a context for each matching revision, after doing arg
468 replacement via revset.formatspec
468 replacement via revset.formatspec
469 '''
469 '''
470 for r in self.revs(expr, *args):
470 for r in self.revs(expr, *args):
471 yield self[r]
471 yield self[r]
472
472
473 def url(self):
473 def url(self):
474 return 'file:' + self.root
474 return 'file:' + self.root
475
475
476 def hook(self, name, throw=False, **args):
476 def hook(self, name, throw=False, **args):
477 """Call a hook, passing this repo instance.
477 """Call a hook, passing this repo instance.
478
478
479 This a convenience method to aid invoking hooks. Extensions likely
479 This a convenience method to aid invoking hooks. Extensions likely
480 won't call this unless they have registered a custom hook or are
480 won't call this unless they have registered a custom hook or are
481 replacing code that is expected to call a hook.
481 replacing code that is expected to call a hook.
482 """
482 """
483 return hook.hook(self.ui, self, name, throw, **args)
483 return hook.hook(self.ui, self, name, throw, **args)
484
484
485 @unfilteredmethod
485 @unfilteredmethod
486 def _tag(self, names, node, message, local, user, date, extra={},
486 def _tag(self, names, node, message, local, user, date, extra={},
487 editor=False):
487 editor=False):
488 if isinstance(names, str):
488 if isinstance(names, str):
489 names = (names,)
489 names = (names,)
490
490
491 branches = self.branchmap()
491 branches = self.branchmap()
492 for name in names:
492 for name in names:
493 self.hook('pretag', throw=True, node=hex(node), tag=name,
493 self.hook('pretag', throw=True, node=hex(node), tag=name,
494 local=local)
494 local=local)
495 if name in branches:
495 if name in branches:
496 self.ui.warn(_("warning: tag %s conflicts with existing"
496 self.ui.warn(_("warning: tag %s conflicts with existing"
497 " branch name\n") % name)
497 " branch name\n") % name)
498
498
499 def writetags(fp, names, munge, prevtags):
499 def writetags(fp, names, munge, prevtags):
500 fp.seek(0, 2)
500 fp.seek(0, 2)
501 if prevtags and prevtags[-1] != '\n':
501 if prevtags and prevtags[-1] != '\n':
502 fp.write('\n')
502 fp.write('\n')
503 for name in names:
503 for name in names:
504 m = munge and munge(name) or name
504 m = munge and munge(name) or name
505 if (self._tagscache.tagtypes and
505 if (self._tagscache.tagtypes and
506 name in self._tagscache.tagtypes):
506 name in self._tagscache.tagtypes):
507 old = self.tags().get(name, nullid)
507 old = self.tags().get(name, nullid)
508 fp.write('%s %s\n' % (hex(old), m))
508 fp.write('%s %s\n' % (hex(old), m))
509 fp.write('%s %s\n' % (hex(node), m))
509 fp.write('%s %s\n' % (hex(node), m))
510 fp.close()
510 fp.close()
511
511
512 prevtags = ''
512 prevtags = ''
513 if local:
513 if local:
514 try:
514 try:
515 fp = self.opener('localtags', 'r+')
515 fp = self.opener('localtags', 'r+')
516 except IOError:
516 except IOError:
517 fp = self.opener('localtags', 'a')
517 fp = self.opener('localtags', 'a')
518 else:
518 else:
519 prevtags = fp.read()
519 prevtags = fp.read()
520
520
521 # local tags are stored in the current charset
521 # local tags are stored in the current charset
522 writetags(fp, names, None, prevtags)
522 writetags(fp, names, None, prevtags)
523 for name in names:
523 for name in names:
524 self.hook('tag', node=hex(node), tag=name, local=local)
524 self.hook('tag', node=hex(node), tag=name, local=local)
525 return
525 return
526
526
527 try:
527 try:
528 fp = self.wfile('.hgtags', 'rb+')
528 fp = self.wfile('.hgtags', 'rb+')
529 except IOError, e:
529 except IOError, e:
530 if e.errno != errno.ENOENT:
530 if e.errno != errno.ENOENT:
531 raise
531 raise
532 fp = self.wfile('.hgtags', 'ab')
532 fp = self.wfile('.hgtags', 'ab')
533 else:
533 else:
534 prevtags = fp.read()
534 prevtags = fp.read()
535
535
536 # committed tags are stored in UTF-8
536 # committed tags are stored in UTF-8
537 writetags(fp, names, encoding.fromlocal, prevtags)
537 writetags(fp, names, encoding.fromlocal, prevtags)
538
538
539 fp.close()
539 fp.close()
540
540
541 self.invalidatecaches()
541 self.invalidatecaches()
542
542
543 if '.hgtags' not in self.dirstate:
543 if '.hgtags' not in self.dirstate:
544 self[None].add(['.hgtags'])
544 self[None].add(['.hgtags'])
545
545
546 m = matchmod.exact(self.root, '', ['.hgtags'])
546 m = matchmod.exact(self.root, '', ['.hgtags'])
547 tagnode = self.commit(message, user, date, extra=extra, match=m,
547 tagnode = self.commit(message, user, date, extra=extra, match=m,
548 editor=editor)
548 editor=editor)
549
549
550 for name in names:
550 for name in names:
551 self.hook('tag', node=hex(node), tag=name, local=local)
551 self.hook('tag', node=hex(node), tag=name, local=local)
552
552
553 return tagnode
553 return tagnode
554
554
555 def tag(self, names, node, message, local, user, date, editor=False):
555 def tag(self, names, node, message, local, user, date, editor=False):
556 '''tag a revision with one or more symbolic names.
556 '''tag a revision with one or more symbolic names.
557
557
558 names is a list of strings or, when adding a single tag, names may be a
558 names is a list of strings or, when adding a single tag, names may be a
559 string.
559 string.
560
560
561 if local is True, the tags are stored in a per-repository file.
561 if local is True, the tags are stored in a per-repository file.
562 otherwise, they are stored in the .hgtags file, and a new
562 otherwise, they are stored in the .hgtags file, and a new
563 changeset is committed with the change.
563 changeset is committed with the change.
564
564
565 keyword arguments:
565 keyword arguments:
566
566
567 local: whether to store tags in non-version-controlled file
567 local: whether to store tags in non-version-controlled file
568 (default False)
568 (default False)
569
569
570 message: commit message to use if committing
570 message: commit message to use if committing
571
571
572 user: name of user to use if committing
572 user: name of user to use if committing
573
573
574 date: date tuple to use if committing'''
574 date: date tuple to use if committing'''
575
575
576 if not local:
576 if not local:
577 for x in self.status()[:5]:
577 for x in self.status()[:5]:
578 if '.hgtags' in x:
578 if '.hgtags' in x:
579 raise util.Abort(_('working copy of .hgtags is changed '
579 raise util.Abort(_('working copy of .hgtags is changed '
580 '(please commit .hgtags manually)'))
580 '(please commit .hgtags manually)'))
581
581
582 self.tags() # instantiate the cache
582 self.tags() # instantiate the cache
583 self._tag(names, node, message, local, user, date, editor=editor)
583 self._tag(names, node, message, local, user, date, editor=editor)
584
584
585 @filteredpropertycache
585 @filteredpropertycache
586 def _tagscache(self):
586 def _tagscache(self):
587 '''Returns a tagscache object that contains various tags related
587 '''Returns a tagscache object that contains various tags related
588 caches.'''
588 caches.'''
589
589
590 # This simplifies its cache management by having one decorated
590 # This simplifies its cache management by having one decorated
591 # function (this one) and the rest simply fetch things from it.
591 # function (this one) and the rest simply fetch things from it.
592 class tagscache(object):
592 class tagscache(object):
593 def __init__(self):
593 def __init__(self):
594 # These two define the set of tags for this repository. tags
594 # These two define the set of tags for this repository. tags
595 # maps tag name to node; tagtypes maps tag name to 'global' or
595 # maps tag name to node; tagtypes maps tag name to 'global' or
596 # 'local'. (Global tags are defined by .hgtags across all
596 # 'local'. (Global tags are defined by .hgtags across all
597 # heads, and local tags are defined in .hg/localtags.)
597 # heads, and local tags are defined in .hg/localtags.)
598 # They constitute the in-memory cache of tags.
598 # They constitute the in-memory cache of tags.
599 self.tags = self.tagtypes = None
599 self.tags = self.tagtypes = None
600
600
601 self.nodetagscache = self.tagslist = None
601 self.nodetagscache = self.tagslist = None
602
602
603 cache = tagscache()
603 cache = tagscache()
604 cache.tags, cache.tagtypes = self._findtags()
604 cache.tags, cache.tagtypes = self._findtags()
605
605
606 return cache
606 return cache
607
607
608 def tags(self):
608 def tags(self):
609 '''return a mapping of tag to node'''
609 '''return a mapping of tag to node'''
610 t = {}
610 t = {}
611 if self.changelog.filteredrevs:
611 if self.changelog.filteredrevs:
612 tags, tt = self._findtags()
612 tags, tt = self._findtags()
613 else:
613 else:
614 tags = self._tagscache.tags
614 tags = self._tagscache.tags
615 for k, v in tags.iteritems():
615 for k, v in tags.iteritems():
616 try:
616 try:
617 # ignore tags to unknown nodes
617 # ignore tags to unknown nodes
618 self.changelog.rev(v)
618 self.changelog.rev(v)
619 t[k] = v
619 t[k] = v
620 except (error.LookupError, ValueError):
620 except (error.LookupError, ValueError):
621 pass
621 pass
622 return t
622 return t
623
623
624 def _findtags(self):
624 def _findtags(self):
625 '''Do the hard work of finding tags. Return a pair of dicts
625 '''Do the hard work of finding tags. Return a pair of dicts
626 (tags, tagtypes) where tags maps tag name to node, and tagtypes
626 (tags, tagtypes) where tags maps tag name to node, and tagtypes
627 maps tag name to a string like \'global\' or \'local\'.
627 maps tag name to a string like \'global\' or \'local\'.
628 Subclasses or extensions are free to add their own tags, but
628 Subclasses or extensions are free to add their own tags, but
629 should be aware that the returned dicts will be retained for the
629 should be aware that the returned dicts will be retained for the
630 duration of the localrepo object.'''
630 duration of the localrepo object.'''
631
631
632 # XXX what tagtype should subclasses/extensions use? Currently
632 # XXX what tagtype should subclasses/extensions use? Currently
633 # mq and bookmarks add tags, but do not set the tagtype at all.
633 # mq and bookmarks add tags, but do not set the tagtype at all.
634 # Should each extension invent its own tag type? Should there
634 # Should each extension invent its own tag type? Should there
635 # be one tagtype for all such "virtual" tags? Or is the status
635 # be one tagtype for all such "virtual" tags? Or is the status
636 # quo fine?
636 # quo fine?
637
637
638 alltags = {} # map tag name to (node, hist)
638 alltags = {} # map tag name to (node, hist)
639 tagtypes = {}
639 tagtypes = {}
640
640
641 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
641 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
642 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
642 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
643
643
644 # Build the return dicts. Have to re-encode tag names because
644 # Build the return dicts. Have to re-encode tag names because
645 # the tags module always uses UTF-8 (in order not to lose info
645 # the tags module always uses UTF-8 (in order not to lose info
646 # writing to the cache), but the rest of Mercurial wants them in
646 # writing to the cache), but the rest of Mercurial wants them in
647 # local encoding.
647 # local encoding.
648 tags = {}
648 tags = {}
649 for (name, (node, hist)) in alltags.iteritems():
649 for (name, (node, hist)) in alltags.iteritems():
650 if node != nullid:
650 if node != nullid:
651 tags[encoding.tolocal(name)] = node
651 tags[encoding.tolocal(name)] = node
652 tags['tip'] = self.changelog.tip()
652 tags['tip'] = self.changelog.tip()
653 tagtypes = dict([(encoding.tolocal(name), value)
653 tagtypes = dict([(encoding.tolocal(name), value)
654 for (name, value) in tagtypes.iteritems()])
654 for (name, value) in tagtypes.iteritems()])
655 return (tags, tagtypes)
655 return (tags, tagtypes)
656
656
657 def tagtype(self, tagname):
657 def tagtype(self, tagname):
658 '''
658 '''
659 return the type of the given tag. result can be:
659 return the type of the given tag. result can be:
660
660
661 'local' : a local tag
661 'local' : a local tag
662 'global' : a global tag
662 'global' : a global tag
663 None : tag does not exist
663 None : tag does not exist
664 '''
664 '''
665
665
666 return self._tagscache.tagtypes.get(tagname)
666 return self._tagscache.tagtypes.get(tagname)
667
667
668 def tagslist(self):
668 def tagslist(self):
669 '''return a list of tags ordered by revision'''
669 '''return a list of tags ordered by revision'''
670 if not self._tagscache.tagslist:
670 if not self._tagscache.tagslist:
671 l = []
671 l = []
672 for t, n in self.tags().iteritems():
672 for t, n in self.tags().iteritems():
673 l.append((self.changelog.rev(n), t, n))
673 l.append((self.changelog.rev(n), t, n))
674 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
674 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
675
675
676 return self._tagscache.tagslist
676 return self._tagscache.tagslist
677
677
678 def nodetags(self, node):
678 def nodetags(self, node):
679 '''return the tags associated with a node'''
679 '''return the tags associated with a node'''
680 if not self._tagscache.nodetagscache:
680 if not self._tagscache.nodetagscache:
681 nodetagscache = {}
681 nodetagscache = {}
682 for t, n in self._tagscache.tags.iteritems():
682 for t, n in self._tagscache.tags.iteritems():
683 nodetagscache.setdefault(n, []).append(t)
683 nodetagscache.setdefault(n, []).append(t)
684 for tags in nodetagscache.itervalues():
684 for tags in nodetagscache.itervalues():
685 tags.sort()
685 tags.sort()
686 self._tagscache.nodetagscache = nodetagscache
686 self._tagscache.nodetagscache = nodetagscache
687 return self._tagscache.nodetagscache.get(node, [])
687 return self._tagscache.nodetagscache.get(node, [])
688
688
689 def nodebookmarks(self, node):
689 def nodebookmarks(self, node):
690 marks = []
690 marks = []
691 for bookmark, n in self._bookmarks.iteritems():
691 for bookmark, n in self._bookmarks.iteritems():
692 if n == node:
692 if n == node:
693 marks.append(bookmark)
693 marks.append(bookmark)
694 return sorted(marks)
694 return sorted(marks)
695
695
696 def branchmap(self):
696 def branchmap(self):
697 '''returns a dictionary {branch: [branchheads]} with branchheads
697 '''returns a dictionary {branch: [branchheads]} with branchheads
698 ordered by increasing revision number'''
698 ordered by increasing revision number'''
699 branchmap.updatecache(self)
699 branchmap.updatecache(self)
700 return self._branchcaches[self.filtername]
700 return self._branchcaches[self.filtername]
701
701
702 def branchtip(self, branch):
702 def branchtip(self, branch):
703 '''return the tip node for a given branch'''
703 '''return the tip node for a given branch'''
704 try:
704 try:
705 return self.branchmap().branchtip(branch)
705 return self.branchmap().branchtip(branch)
706 except KeyError:
706 except KeyError:
707 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
707 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
708
708
709 def lookup(self, key):
709 def lookup(self, key):
710 return self[key].node()
710 return self[key].node()
711
711
712 def lookupbranch(self, key, remote=None):
712 def lookupbranch(self, key, remote=None):
713 repo = remote or self
713 repo = remote or self
714 if key in repo.branchmap():
714 if key in repo.branchmap():
715 return key
715 return key
716
716
717 repo = (remote and remote.local()) and remote or self
717 repo = (remote and remote.local()) and remote or self
718 return repo[key].branch()
718 return repo[key].branch()
719
719
720 def known(self, nodes):
720 def known(self, nodes):
721 nm = self.changelog.nodemap
721 nm = self.changelog.nodemap
722 pc = self._phasecache
722 pc = self._phasecache
723 result = []
723 result = []
724 for n in nodes:
724 for n in nodes:
725 r = nm.get(n)
725 r = nm.get(n)
726 resp = not (r is None or pc.phase(self, r) >= phases.secret)
726 resp = not (r is None or pc.phase(self, r) >= phases.secret)
727 result.append(resp)
727 result.append(resp)
728 return result
728 return result
729
729
730 def local(self):
730 def local(self):
731 return self
731 return self
732
732
733 def cancopy(self):
733 def cancopy(self):
734 # so statichttprepo's override of local() works
734 # so statichttprepo's override of local() works
735 if not self.local():
735 if not self.local():
736 return False
736 return False
737 if not self.ui.configbool('phases', 'publish', True):
737 if not self.ui.configbool('phases', 'publish', True):
738 return True
738 return True
739 # if publishing we can't copy if there is filtered content
739 # if publishing we can't copy if there is filtered content
740 return not self.filtered('visible').changelog.filteredrevs
740 return not self.filtered('visible').changelog.filteredrevs
741
741
742 def join(self, f, *insidef):
742 def join(self, f, *insidef):
743 return os.path.join(self.path, f, *insidef)
743 return os.path.join(self.path, f, *insidef)
744
744
745 def wjoin(self, f, *insidef):
745 def wjoin(self, f, *insidef):
746 return os.path.join(self.root, f, *insidef)
746 return os.path.join(self.root, f, *insidef)
747
747
748 def file(self, f):
748 def file(self, f):
749 if f[0] == '/':
749 if f[0] == '/':
750 f = f[1:]
750 f = f[1:]
751 return filelog.filelog(self.sopener, f)
751 return filelog.filelog(self.sopener, f)
752
752
753 def changectx(self, changeid):
753 def changectx(self, changeid):
754 return self[changeid]
754 return self[changeid]
755
755
756 def parents(self, changeid=None):
756 def parents(self, changeid=None):
757 '''get list of changectxs for parents of changeid'''
757 '''get list of changectxs for parents of changeid'''
758 return self[changeid].parents()
758 return self[changeid].parents()
759
759
760 def setparents(self, p1, p2=nullid):
760 def setparents(self, p1, p2=nullid):
761 self.dirstate.beginparentchange()
761 self.dirstate.beginparentchange()
762 copies = self.dirstate.setparents(p1, p2)
762 copies = self.dirstate.setparents(p1, p2)
763 pctx = self[p1]
763 pctx = self[p1]
764 if copies:
764 if copies:
765 # Adjust copy records, the dirstate cannot do it, it
765 # Adjust copy records, the dirstate cannot do it, it
766 # requires access to parents manifests. Preserve them
766 # requires access to parents manifests. Preserve them
767 # only for entries added to first parent.
767 # only for entries added to first parent.
768 for f in copies:
768 for f in copies:
769 if f not in pctx and copies[f] in pctx:
769 if f not in pctx and copies[f] in pctx:
770 self.dirstate.copy(copies[f], f)
770 self.dirstate.copy(copies[f], f)
771 if p2 == nullid:
771 if p2 == nullid:
772 for f, s in sorted(self.dirstate.copies().items()):
772 for f, s in sorted(self.dirstate.copies().items()):
773 if f not in pctx and s not in pctx:
773 if f not in pctx and s not in pctx:
774 self.dirstate.copy(None, f)
774 self.dirstate.copy(None, f)
775 self.dirstate.endparentchange()
775 self.dirstate.endparentchange()
776
776
777 def filectx(self, path, changeid=None, fileid=None):
777 def filectx(self, path, changeid=None, fileid=None):
778 """changeid can be a changeset revision, node, or tag.
778 """changeid can be a changeset revision, node, or tag.
779 fileid can be a file revision or node."""
779 fileid can be a file revision or node."""
780 return context.filectx(self, path, changeid, fileid)
780 return context.filectx(self, path, changeid, fileid)
781
781
782 def getcwd(self):
782 def getcwd(self):
783 return self.dirstate.getcwd()
783 return self.dirstate.getcwd()
784
784
785 def pathto(self, f, cwd=None):
785 def pathto(self, f, cwd=None):
786 return self.dirstate.pathto(f, cwd)
786 return self.dirstate.pathto(f, cwd)
787
787
788 def wfile(self, f, mode='r'):
788 def wfile(self, f, mode='r'):
789 return self.wopener(f, mode)
789 return self.wopener(f, mode)
790
790
791 def _link(self, f):
791 def _link(self, f):
792 return self.wvfs.islink(f)
792 return self.wvfs.islink(f)
793
793
794 def _loadfilter(self, filter):
794 def _loadfilter(self, filter):
795 if filter not in self.filterpats:
795 if filter not in self.filterpats:
796 l = []
796 l = []
797 for pat, cmd in self.ui.configitems(filter):
797 for pat, cmd in self.ui.configitems(filter):
798 if cmd == '!':
798 if cmd == '!':
799 continue
799 continue
800 mf = matchmod.match(self.root, '', [pat])
800 mf = matchmod.match(self.root, '', [pat])
801 fn = None
801 fn = None
802 params = cmd
802 params = cmd
803 for name, filterfn in self._datafilters.iteritems():
803 for name, filterfn in self._datafilters.iteritems():
804 if cmd.startswith(name):
804 if cmd.startswith(name):
805 fn = filterfn
805 fn = filterfn
806 params = cmd[len(name):].lstrip()
806 params = cmd[len(name):].lstrip()
807 break
807 break
808 if not fn:
808 if not fn:
809 fn = lambda s, c, **kwargs: util.filter(s, c)
809 fn = lambda s, c, **kwargs: util.filter(s, c)
810 # Wrap old filters not supporting keyword arguments
810 # Wrap old filters not supporting keyword arguments
811 if not inspect.getargspec(fn)[2]:
811 if not inspect.getargspec(fn)[2]:
812 oldfn = fn
812 oldfn = fn
813 fn = lambda s, c, **kwargs: oldfn(s, c)
813 fn = lambda s, c, **kwargs: oldfn(s, c)
814 l.append((mf, fn, params))
814 l.append((mf, fn, params))
815 self.filterpats[filter] = l
815 self.filterpats[filter] = l
816 return self.filterpats[filter]
816 return self.filterpats[filter]
817
817
818 def _filter(self, filterpats, filename, data):
818 def _filter(self, filterpats, filename, data):
819 for mf, fn, cmd in filterpats:
819 for mf, fn, cmd in filterpats:
820 if mf(filename):
820 if mf(filename):
821 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
821 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
822 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
822 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
823 break
823 break
824
824
825 return data
825 return data
826
826
827 @unfilteredpropertycache
827 @unfilteredpropertycache
828 def _encodefilterpats(self):
828 def _encodefilterpats(self):
829 return self._loadfilter('encode')
829 return self._loadfilter('encode')
830
830
831 @unfilteredpropertycache
831 @unfilteredpropertycache
832 def _decodefilterpats(self):
832 def _decodefilterpats(self):
833 return self._loadfilter('decode')
833 return self._loadfilter('decode')
834
834
835 def adddatafilter(self, name, filter):
835 def adddatafilter(self, name, filter):
836 self._datafilters[name] = filter
836 self._datafilters[name] = filter
837
837
838 def wread(self, filename):
838 def wread(self, filename):
839 if self._link(filename):
839 if self._link(filename):
840 data = self.wvfs.readlink(filename)
840 data = self.wvfs.readlink(filename)
841 else:
841 else:
842 data = self.wopener.read(filename)
842 data = self.wopener.read(filename)
843 return self._filter(self._encodefilterpats, filename, data)
843 return self._filter(self._encodefilterpats, filename, data)
844
844
845 def wwrite(self, filename, data, flags):
845 def wwrite(self, filename, data, flags):
846 data = self._filter(self._decodefilterpats, filename, data)
846 data = self._filter(self._decodefilterpats, filename, data)
847 if 'l' in flags:
847 if 'l' in flags:
848 self.wopener.symlink(data, filename)
848 self.wopener.symlink(data, filename)
849 else:
849 else:
850 self.wopener.write(filename, data)
850 self.wopener.write(filename, data)
851 if 'x' in flags:
851 if 'x' in flags:
852 self.wvfs.setflags(filename, False, True)
852 self.wvfs.setflags(filename, False, True)
853
853
854 def wwritedata(self, filename, data):
854 def wwritedata(self, filename, data):
855 return self._filter(self._decodefilterpats, filename, data)
855 return self._filter(self._decodefilterpats, filename, data)
856
856
857 def transaction(self, desc, report=None):
857 def transaction(self, desc, report=None):
858 tr = self._transref and self._transref() or None
858 tr = self._transref and self._transref() or None
859 if tr and tr.running():
859 if tr and tr.running():
860 return tr.nest()
860 return tr.nest()
861
861
862 # abort here if the journal already exists
862 # abort here if the journal already exists
863 if self.svfs.exists("journal"):
863 if self.svfs.exists("journal"):
864 raise error.RepoError(
864 raise error.RepoError(
865 _("abandoned transaction found"),
865 _("abandoned transaction found"),
866 hint=_("run 'hg recover' to clean up transaction"))
866 hint=_("run 'hg recover' to clean up transaction"))
867
867
868 def onclose():
868 def onclose():
869 self.store.write(self._transref())
869 self.store.write(self._transref())
870
870
871 self._writejournal(desc)
871 self._writejournal(desc)
872 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
872 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
873 rp = report and report or self.ui.warn
873 rp = report and report or self.ui.warn
874 tr = transaction.transaction(rp, self.sopener,
874 tr = transaction.transaction(rp, self.sopener,
875 "journal",
875 "journal",
876 aftertrans(renames),
876 aftertrans(renames),
877 self.store.createmode,
877 self.store.createmode,
878 onclose)
878 onclose)
879 self._transref = weakref.ref(tr)
879 self._transref = weakref.ref(tr)
880 return tr
880 return tr
881
881
882 def _journalfiles(self):
882 def _journalfiles(self):
883 return ((self.svfs, 'journal'),
883 return ((self.svfs, 'journal'),
884 (self.vfs, 'journal.dirstate'),
884 (self.vfs, 'journal.dirstate'),
885 (self.vfs, 'journal.branch'),
885 (self.vfs, 'journal.branch'),
886 (self.vfs, 'journal.desc'),
886 (self.vfs, 'journal.desc'),
887 (self.vfs, 'journal.bookmarks'),
887 (self.vfs, 'journal.bookmarks'),
888 (self.svfs, 'journal.phaseroots'))
888 (self.svfs, 'journal.phaseroots'))
889
889
890 def undofiles(self):
890 def undofiles(self):
891 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
891 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
892
892
893 def _writejournal(self, desc):
893 def _writejournal(self, desc):
894 self.opener.write("journal.dirstate",
894 self.opener.write("journal.dirstate",
895 self.opener.tryread("dirstate"))
895 self.opener.tryread("dirstate"))
896 self.opener.write("journal.branch",
896 self.opener.write("journal.branch",
897 encoding.fromlocal(self.dirstate.branch()))
897 encoding.fromlocal(self.dirstate.branch()))
898 self.opener.write("journal.desc",
898 self.opener.write("journal.desc",
899 "%d\n%s\n" % (len(self), desc))
899 "%d\n%s\n" % (len(self), desc))
900 self.opener.write("journal.bookmarks",
900 self.opener.write("journal.bookmarks",
901 self.opener.tryread("bookmarks"))
901 self.opener.tryread("bookmarks"))
902 self.sopener.write("journal.phaseroots",
902 self.sopener.write("journal.phaseroots",
903 self.sopener.tryread("phaseroots"))
903 self.sopener.tryread("phaseroots"))
904
904
905 def recover(self):
905 def recover(self):
906 lock = self.lock()
906 lock = self.lock()
907 try:
907 try:
908 if self.svfs.exists("journal"):
908 if self.svfs.exists("journal"):
909 self.ui.status(_("rolling back interrupted transaction\n"))
909 self.ui.status(_("rolling back interrupted transaction\n"))
910 transaction.rollback(self.sopener, "journal",
910 transaction.rollback(self.sopener, "journal",
911 self.ui.warn)
911 self.ui.warn)
912 self.invalidate()
912 self.invalidate()
913 return True
913 return True
914 else:
914 else:
915 self.ui.warn(_("no interrupted transaction available\n"))
915 self.ui.warn(_("no interrupted transaction available\n"))
916 return False
916 return False
917 finally:
917 finally:
918 lock.release()
918 lock.release()
919
919
920 def rollback(self, dryrun=False, force=False):
920 def rollback(self, dryrun=False, force=False):
921 wlock = lock = None
921 wlock = lock = None
922 try:
922 try:
923 wlock = self.wlock()
923 wlock = self.wlock()
924 lock = self.lock()
924 lock = self.lock()
925 if self.svfs.exists("undo"):
925 if self.svfs.exists("undo"):
926 return self._rollback(dryrun, force)
926 return self._rollback(dryrun, force)
927 else:
927 else:
928 self.ui.warn(_("no rollback information available\n"))
928 self.ui.warn(_("no rollback information available\n"))
929 return 1
929 return 1
930 finally:
930 finally:
931 release(lock, wlock)
931 release(lock, wlock)
932
932
933 @unfilteredmethod # Until we get smarter cache management
933 @unfilteredmethod # Until we get smarter cache management
934 def _rollback(self, dryrun, force):
934 def _rollback(self, dryrun, force):
935 ui = self.ui
935 ui = self.ui
936 try:
936 try:
937 args = self.opener.read('undo.desc').splitlines()
937 args = self.opener.read('undo.desc').splitlines()
938 (oldlen, desc, detail) = (int(args[0]), args[1], None)
938 (oldlen, desc, detail) = (int(args[0]), args[1], None)
939 if len(args) >= 3:
939 if len(args) >= 3:
940 detail = args[2]
940 detail = args[2]
941 oldtip = oldlen - 1
941 oldtip = oldlen - 1
942
942
943 if detail and ui.verbose:
943 if detail and ui.verbose:
944 msg = (_('repository tip rolled back to revision %s'
944 msg = (_('repository tip rolled back to revision %s'
945 ' (undo %s: %s)\n')
945 ' (undo %s: %s)\n')
946 % (oldtip, desc, detail))
946 % (oldtip, desc, detail))
947 else:
947 else:
948 msg = (_('repository tip rolled back to revision %s'
948 msg = (_('repository tip rolled back to revision %s'
949 ' (undo %s)\n')
949 ' (undo %s)\n')
950 % (oldtip, desc))
950 % (oldtip, desc))
951 except IOError:
951 except IOError:
952 msg = _('rolling back unknown transaction\n')
952 msg = _('rolling back unknown transaction\n')
953 desc = None
953 desc = None
954
954
955 if not force and self['.'] != self['tip'] and desc == 'commit':
955 if not force and self['.'] != self['tip'] and desc == 'commit':
956 raise util.Abort(
956 raise util.Abort(
957 _('rollback of last commit while not checked out '
957 _('rollback of last commit while not checked out '
958 'may lose data'), hint=_('use -f to force'))
958 'may lose data'), hint=_('use -f to force'))
959
959
960 ui.status(msg)
960 ui.status(msg)
961 if dryrun:
961 if dryrun:
962 return 0
962 return 0
963
963
964 parents = self.dirstate.parents()
964 parents = self.dirstate.parents()
965 self.destroying()
965 self.destroying()
966 transaction.rollback(self.sopener, 'undo', ui.warn)
966 transaction.rollback(self.sopener, 'undo', ui.warn)
967 if self.vfs.exists('undo.bookmarks'):
967 if self.vfs.exists('undo.bookmarks'):
968 self.vfs.rename('undo.bookmarks', 'bookmarks')
968 self.vfs.rename('undo.bookmarks', 'bookmarks')
969 if self.svfs.exists('undo.phaseroots'):
969 if self.svfs.exists('undo.phaseroots'):
970 self.svfs.rename('undo.phaseroots', 'phaseroots')
970 self.svfs.rename('undo.phaseroots', 'phaseroots')
971 self.invalidate()
971 self.invalidate()
972
972
973 parentgone = (parents[0] not in self.changelog.nodemap or
973 parentgone = (parents[0] not in self.changelog.nodemap or
974 parents[1] not in self.changelog.nodemap)
974 parents[1] not in self.changelog.nodemap)
975 if parentgone:
975 if parentgone:
976 self.vfs.rename('undo.dirstate', 'dirstate')
976 self.vfs.rename('undo.dirstate', 'dirstate')
977 try:
977 try:
978 branch = self.opener.read('undo.branch')
978 branch = self.opener.read('undo.branch')
979 self.dirstate.setbranch(encoding.tolocal(branch))
979 self.dirstate.setbranch(encoding.tolocal(branch))
980 except IOError:
980 except IOError:
981 ui.warn(_('named branch could not be reset: '
981 ui.warn(_('named branch could not be reset: '
982 'current branch is still \'%s\'\n')
982 'current branch is still \'%s\'\n')
983 % self.dirstate.branch())
983 % self.dirstate.branch())
984
984
985 self.dirstate.invalidate()
985 self.dirstate.invalidate()
986 parents = tuple([p.rev() for p in self.parents()])
986 parents = tuple([p.rev() for p in self.parents()])
987 if len(parents) > 1:
987 if len(parents) > 1:
988 ui.status(_('working directory now based on '
988 ui.status(_('working directory now based on '
989 'revisions %d and %d\n') % parents)
989 'revisions %d and %d\n') % parents)
990 else:
990 else:
991 ui.status(_('working directory now based on '
991 ui.status(_('working directory now based on '
992 'revision %d\n') % parents)
992 'revision %d\n') % parents)
993 # TODO: if we know which new heads may result from this rollback, pass
993 # TODO: if we know which new heads may result from this rollback, pass
994 # them to destroy(), which will prevent the branchhead cache from being
994 # them to destroy(), which will prevent the branchhead cache from being
995 # invalidated.
995 # invalidated.
996 self.destroyed()
996 self.destroyed()
997 return 0
997 return 0
998
998
999 def invalidatecaches(self):
999 def invalidatecaches(self):
1000
1000
1001 if '_tagscache' in vars(self):
1001 if '_tagscache' in vars(self):
1002 # can't use delattr on proxy
1002 # can't use delattr on proxy
1003 del self.__dict__['_tagscache']
1003 del self.__dict__['_tagscache']
1004
1004
1005 self.unfiltered()._branchcaches.clear()
1005 self.unfiltered()._branchcaches.clear()
1006 self.invalidatevolatilesets()
1006 self.invalidatevolatilesets()
1007
1007
1008 def invalidatevolatilesets(self):
1008 def invalidatevolatilesets(self):
1009 self.filteredrevcache.clear()
1009 self.filteredrevcache.clear()
1010 obsolete.clearobscaches(self)
1010 obsolete.clearobscaches(self)
1011
1011
1012 def invalidatedirstate(self):
1012 def invalidatedirstate(self):
1013 '''Invalidates the dirstate, causing the next call to dirstate
1013 '''Invalidates the dirstate, causing the next call to dirstate
1014 to check if it was modified since the last time it was read,
1014 to check if it was modified since the last time it was read,
1015 rereading it if it has.
1015 rereading it if it has.
1016
1016
1017 This is different to dirstate.invalidate() that it doesn't always
1017 This is different to dirstate.invalidate() that it doesn't always
1018 rereads the dirstate. Use dirstate.invalidate() if you want to
1018 rereads the dirstate. Use dirstate.invalidate() if you want to
1019 explicitly read the dirstate again (i.e. restoring it to a previous
1019 explicitly read the dirstate again (i.e. restoring it to a previous
1020 known good state).'''
1020 known good state).'''
1021 if hasunfilteredcache(self, 'dirstate'):
1021 if hasunfilteredcache(self, 'dirstate'):
1022 for k in self.dirstate._filecache:
1022 for k in self.dirstate._filecache:
1023 try:
1023 try:
1024 delattr(self.dirstate, k)
1024 delattr(self.dirstate, k)
1025 except AttributeError:
1025 except AttributeError:
1026 pass
1026 pass
1027 delattr(self.unfiltered(), 'dirstate')
1027 delattr(self.unfiltered(), 'dirstate')
1028
1028
1029 def invalidate(self):
1029 def invalidate(self):
1030 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1030 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1031 for k in self._filecache:
1031 for k in self._filecache:
1032 # dirstate is invalidated separately in invalidatedirstate()
1032 # dirstate is invalidated separately in invalidatedirstate()
1033 if k == 'dirstate':
1033 if k == 'dirstate':
1034 continue
1034 continue
1035
1035
1036 try:
1036 try:
1037 delattr(unfiltered, k)
1037 delattr(unfiltered, k)
1038 except AttributeError:
1038 except AttributeError:
1039 pass
1039 pass
1040 self.invalidatecaches()
1040 self.invalidatecaches()
1041 self.store.invalidatecaches()
1041 self.store.invalidatecaches()
1042
1042
1043 def invalidateall(self):
1043 def invalidateall(self):
1044 '''Fully invalidates both store and non-store parts, causing the
1044 '''Fully invalidates both store and non-store parts, causing the
1045 subsequent operation to reread any outside changes.'''
1045 subsequent operation to reread any outside changes.'''
1046 # extension should hook this to invalidate its caches
1046 # extension should hook this to invalidate its caches
1047 self.invalidate()
1047 self.invalidate()
1048 self.invalidatedirstate()
1048 self.invalidatedirstate()
1049
1049
1050 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1050 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1051 try:
1051 try:
1052 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1052 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1053 except error.LockHeld, inst:
1053 except error.LockHeld, inst:
1054 if not wait:
1054 if not wait:
1055 raise
1055 raise
1056 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1056 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1057 (desc, inst.locker))
1057 (desc, inst.locker))
1058 # default to 600 seconds timeout
1058 # default to 600 seconds timeout
1059 l = lockmod.lock(vfs, lockname,
1059 l = lockmod.lock(vfs, lockname,
1060 int(self.ui.config("ui", "timeout", "600")),
1060 int(self.ui.config("ui", "timeout", "600")),
1061 releasefn, desc=desc)
1061 releasefn, desc=desc)
1062 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1062 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1063 if acquirefn:
1063 if acquirefn:
1064 acquirefn()
1064 acquirefn()
1065 return l
1065 return l
1066
1066
1067 def _afterlock(self, callback):
1067 def _afterlock(self, callback):
1068 """add a callback to the current repository lock.
1068 """add a callback to the current repository lock.
1069
1069
1070 The callback will be executed on lock release."""
1070 The callback will be executed on lock release."""
1071 l = self._lockref and self._lockref()
1071 l = self._lockref and self._lockref()
1072 if l:
1072 if l:
1073 l.postrelease.append(callback)
1073 l.postrelease.append(callback)
1074 else:
1074 else:
1075 callback()
1075 callback()
1076
1076
1077 def lock(self, wait=True):
1077 def lock(self, wait=True):
1078 '''Lock the repository store (.hg/store) and return a weak reference
1078 '''Lock the repository store (.hg/store) and return a weak reference
1079 to the lock. Use this before modifying the store (e.g. committing or
1079 to the lock. Use this before modifying the store (e.g. committing or
1080 stripping). If you are opening a transaction, get a lock as well.)'''
1080 stripping). If you are opening a transaction, get a lock as well.)'''
1081 l = self._lockref and self._lockref()
1081 l = self._lockref and self._lockref()
1082 if l is not None and l.held:
1082 if l is not None and l.held:
1083 l.lock()
1083 l.lock()
1084 return l
1084 return l
1085
1085
1086 def unlock():
1086 def unlock():
1087 for k, ce in self._filecache.items():
1087 for k, ce in self._filecache.items():
1088 if k == 'dirstate' or k not in self.__dict__:
1088 if k == 'dirstate' or k not in self.__dict__:
1089 continue
1089 continue
1090 ce.refresh()
1090 ce.refresh()
1091
1091
1092 l = self._lock(self.svfs, "lock", wait, unlock,
1092 l = self._lock(self.svfs, "lock", wait, unlock,
1093 self.invalidate, _('repository %s') % self.origroot)
1093 self.invalidate, _('repository %s') % self.origroot)
1094 self._lockref = weakref.ref(l)
1094 self._lockref = weakref.ref(l)
1095 return l
1095 return l
1096
1096
1097 def wlock(self, wait=True):
1097 def wlock(self, wait=True):
1098 '''Lock the non-store parts of the repository (everything under
1098 '''Lock the non-store parts of the repository (everything under
1099 .hg except .hg/store) and return a weak reference to the lock.
1099 .hg except .hg/store) and return a weak reference to the lock.
1100 Use this before modifying files in .hg.'''
1100 Use this before modifying files in .hg.'''
1101 l = self._wlockref and self._wlockref()
1101 l = self._wlockref and self._wlockref()
1102 if l is not None and l.held:
1102 if l is not None and l.held:
1103 l.lock()
1103 l.lock()
1104 return l
1104 return l
1105
1105
1106 def unlock():
1106 def unlock():
1107 if self.dirstate.pendingparentchange():
1107 if self.dirstate.pendingparentchange():
1108 self.dirstate.invalidate()
1108 self.dirstate.invalidate()
1109 else:
1109 else:
1110 self.dirstate.write()
1110 self.dirstate.write()
1111
1111
1112 self._filecache['dirstate'].refresh()
1112 self._filecache['dirstate'].refresh()
1113
1113
1114 l = self._lock(self.vfs, "wlock", wait, unlock,
1114 l = self._lock(self.vfs, "wlock", wait, unlock,
1115 self.invalidatedirstate, _('working directory of %s') %
1115 self.invalidatedirstate, _('working directory of %s') %
1116 self.origroot)
1116 self.origroot)
1117 self._wlockref = weakref.ref(l)
1117 self._wlockref = weakref.ref(l)
1118 return l
1118 return l
1119
1119
1120 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1120 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1121 """
1121 """
1122 commit an individual file as part of a larger transaction
1122 commit an individual file as part of a larger transaction
1123 """
1123 """
1124
1124
1125 fname = fctx.path()
1125 fname = fctx.path()
1126 text = fctx.data()
1126 text = fctx.data()
1127 flog = self.file(fname)
1127 flog = self.file(fname)
1128 fparent1 = manifest1.get(fname, nullid)
1128 fparent1 = manifest1.get(fname, nullid)
1129 fparent2 = manifest2.get(fname, nullid)
1129 fparent2 = manifest2.get(fname, nullid)
1130
1130
1131 meta = {}
1131 meta = {}
1132 copy = fctx.renamed()
1132 copy = fctx.renamed()
1133 if copy and copy[0] != fname:
1133 if copy and copy[0] != fname:
1134 # Mark the new revision of this file as a copy of another
1134 # Mark the new revision of this file as a copy of another
1135 # file. This copy data will effectively act as a parent
1135 # file. This copy data will effectively act as a parent
1136 # of this new revision. If this is a merge, the first
1136 # of this new revision. If this is a merge, the first
1137 # parent will be the nullid (meaning "look up the copy data")
1137 # parent will be the nullid (meaning "look up the copy data")
1138 # and the second one will be the other parent. For example:
1138 # and the second one will be the other parent. For example:
1139 #
1139 #
1140 # 0 --- 1 --- 3 rev1 changes file foo
1140 # 0 --- 1 --- 3 rev1 changes file foo
1141 # \ / rev2 renames foo to bar and changes it
1141 # \ / rev2 renames foo to bar and changes it
1142 # \- 2 -/ rev3 should have bar with all changes and
1142 # \- 2 -/ rev3 should have bar with all changes and
1143 # should record that bar descends from
1143 # should record that bar descends from
1144 # bar in rev2 and foo in rev1
1144 # bar in rev2 and foo in rev1
1145 #
1145 #
1146 # this allows this merge to succeed:
1146 # this allows this merge to succeed:
1147 #
1147 #
1148 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1148 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1149 # \ / merging rev3 and rev4 should use bar@rev2
1149 # \ / merging rev3 and rev4 should use bar@rev2
1150 # \- 2 --- 4 as the merge base
1150 # \- 2 --- 4 as the merge base
1151 #
1151 #
1152
1152
1153 cfname = copy[0]
1153 cfname = copy[0]
1154 crev = manifest1.get(cfname)
1154 crev = manifest1.get(cfname)
1155 newfparent = fparent2
1155 newfparent = fparent2
1156
1156
1157 if manifest2: # branch merge
1157 if manifest2: # branch merge
1158 if fparent2 == nullid or crev is None: # copied on remote side
1158 if fparent2 == nullid or crev is None: # copied on remote side
1159 if cfname in manifest2:
1159 if cfname in manifest2:
1160 crev = manifest2[cfname]
1160 crev = manifest2[cfname]
1161 newfparent = fparent1
1161 newfparent = fparent1
1162
1162
1163 # find source in nearest ancestor if we've lost track
1163 # find source in nearest ancestor if we've lost track
1164 if not crev:
1164 if not crev:
1165 self.ui.debug(" %s: searching for copy revision for %s\n" %
1165 self.ui.debug(" %s: searching for copy revision for %s\n" %
1166 (fname, cfname))
1166 (fname, cfname))
1167 for ancestor in self[None].ancestors():
1167 for ancestor in self[None].ancestors():
1168 if cfname in ancestor:
1168 if cfname in ancestor:
1169 crev = ancestor[cfname].filenode()
1169 crev = ancestor[cfname].filenode()
1170 break
1170 break
1171
1171
1172 if crev:
1172 if crev:
1173 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1173 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1174 meta["copy"] = cfname
1174 meta["copy"] = cfname
1175 meta["copyrev"] = hex(crev)
1175 meta["copyrev"] = hex(crev)
1176 fparent1, fparent2 = nullid, newfparent
1176 fparent1, fparent2 = nullid, newfparent
1177 else:
1177 else:
1178 self.ui.warn(_("warning: can't find ancestor for '%s' "
1178 self.ui.warn(_("warning: can't find ancestor for '%s' "
1179 "copied from '%s'!\n") % (fname, cfname))
1179 "copied from '%s'!\n") % (fname, cfname))
1180
1180
1181 elif fparent1 == nullid:
1181 elif fparent1 == nullid:
1182 fparent1, fparent2 = fparent2, nullid
1182 fparent1, fparent2 = fparent2, nullid
1183 elif fparent2 != nullid:
1183 elif fparent2 != nullid:
1184 # is one parent an ancestor of the other?
1184 # is one parent an ancestor of the other?
1185 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1185 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1186 if fparent1 in fparentancestors:
1186 if fparent1 in fparentancestors:
1187 fparent1, fparent2 = fparent2, nullid
1187 fparent1, fparent2 = fparent2, nullid
1188 elif fparent2 in fparentancestors:
1188 elif fparent2 in fparentancestors:
1189 fparent2 = nullid
1189 fparent2 = nullid
1190
1190
1191 # is the file changed?
1191 # is the file changed?
1192 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1192 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1193 changelist.append(fname)
1193 changelist.append(fname)
1194 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1194 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1195 # are just the flags changed during merge?
1195 # are just the flags changed during merge?
1196 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1196 elif fname in manifest1 and manifest1.flags(fname) != fctx.flags():
1197 changelist.append(fname)
1197 changelist.append(fname)
1198
1198
1199 return fparent1
1199 return fparent1
1200
1200
1201 @unfilteredmethod
1201 @unfilteredmethod
1202 def commit(self, text="", user=None, date=None, match=None, force=False,
1202 def commit(self, text="", user=None, date=None, match=None, force=False,
1203 editor=False, extra={}):
1203 editor=False, extra={}):
1204 """Add a new revision to current repository.
1204 """Add a new revision to current repository.
1205
1205
1206 Revision information is gathered from the working directory,
1206 Revision information is gathered from the working directory,
1207 match can be used to filter the committed files. If editor is
1207 match can be used to filter the committed files. If editor is
1208 supplied, it is called to get a commit message.
1208 supplied, it is called to get a commit message.
1209 """
1209 """
1210
1210
1211 def fail(f, msg):
1211 def fail(f, msg):
1212 raise util.Abort('%s: %s' % (f, msg))
1212 raise util.Abort('%s: %s' % (f, msg))
1213
1213
1214 if not match:
1214 if not match:
1215 match = matchmod.always(self.root, '')
1215 match = matchmod.always(self.root, '')
1216
1216
1217 if not force:
1217 if not force:
1218 vdirs = []
1218 vdirs = []
1219 match.explicitdir = vdirs.append
1219 match.explicitdir = vdirs.append
1220 match.bad = fail
1220 match.bad = fail
1221
1221
1222 wlock = self.wlock()
1222 wlock = self.wlock()
1223 try:
1223 try:
1224 wctx = self[None]
1224 wctx = self[None]
1225 merge = len(wctx.parents()) > 1
1225 merge = len(wctx.parents()) > 1
1226
1226
1227 if (not force and merge and match and
1227 if (not force and merge and match and
1228 (match.files() or match.anypats())):
1228 (match.files() or match.anypats())):
1229 raise util.Abort(_('cannot partially commit a merge '
1229 raise util.Abort(_('cannot partially commit a merge '
1230 '(do not specify files or patterns)'))
1230 '(do not specify files or patterns)'))
1231
1231
1232 changes = self.status(match=match, clean=force)
1232 changes = self.status(match=match, clean=force)
1233 if force:
1233 if force:
1234 changes[0].extend(changes[6]) # mq may commit unchanged files
1234 changes[0].extend(changes[6]) # mq may commit unchanged files
1235
1235
1236 # check subrepos
1236 # check subrepos
1237 subs = []
1237 subs = []
1238 commitsubs = set()
1238 commitsubs = set()
1239 newstate = wctx.substate.copy()
1239 newstate = wctx.substate.copy()
1240 # only manage subrepos and .hgsubstate if .hgsub is present
1240 # only manage subrepos and .hgsubstate if .hgsub is present
1241 if '.hgsub' in wctx:
1241 if '.hgsub' in wctx:
1242 # we'll decide whether to track this ourselves, thanks
1242 # we'll decide whether to track this ourselves, thanks
1243 for c in changes[:3]:
1243 for c in changes[:3]:
1244 if '.hgsubstate' in c:
1244 if '.hgsubstate' in c:
1245 c.remove('.hgsubstate')
1245 c.remove('.hgsubstate')
1246
1246
1247 # compare current state to last committed state
1247 # compare current state to last committed state
1248 # build new substate based on last committed state
1248 # build new substate based on last committed state
1249 oldstate = wctx.p1().substate
1249 oldstate = wctx.p1().substate
1250 for s in sorted(newstate.keys()):
1250 for s in sorted(newstate.keys()):
1251 if not match(s):
1251 if not match(s):
1252 # ignore working copy, use old state if present
1252 # ignore working copy, use old state if present
1253 if s in oldstate:
1253 if s in oldstate:
1254 newstate[s] = oldstate[s]
1254 newstate[s] = oldstate[s]
1255 continue
1255 continue
1256 if not force:
1256 if not force:
1257 raise util.Abort(
1257 raise util.Abort(
1258 _("commit with new subrepo %s excluded") % s)
1258 _("commit with new subrepo %s excluded") % s)
1259 if wctx.sub(s).dirty(True):
1259 if wctx.sub(s).dirty(True):
1260 if not self.ui.configbool('ui', 'commitsubrepos'):
1260 if not self.ui.configbool('ui', 'commitsubrepos'):
1261 raise util.Abort(
1261 raise util.Abort(
1262 _("uncommitted changes in subrepo %s") % s,
1262 _("uncommitted changes in subrepo %s") % s,
1263 hint=_("use --subrepos for recursive commit"))
1263 hint=_("use --subrepos for recursive commit"))
1264 subs.append(s)
1264 subs.append(s)
1265 commitsubs.add(s)
1265 commitsubs.add(s)
1266 else:
1266 else:
1267 bs = wctx.sub(s).basestate()
1267 bs = wctx.sub(s).basestate()
1268 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1268 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1269 if oldstate.get(s, (None, None, None))[1] != bs:
1269 if oldstate.get(s, (None, None, None))[1] != bs:
1270 subs.append(s)
1270 subs.append(s)
1271
1271
1272 # check for removed subrepos
1272 # check for removed subrepos
1273 for p in wctx.parents():
1273 for p in wctx.parents():
1274 r = [s for s in p.substate if s not in newstate]
1274 r = [s for s in p.substate if s not in newstate]
1275 subs += [s for s in r if match(s)]
1275 subs += [s for s in r if match(s)]
1276 if subs:
1276 if subs:
1277 if (not match('.hgsub') and
1277 if (not match('.hgsub') and
1278 '.hgsub' in (wctx.modified() + wctx.added())):
1278 '.hgsub' in (wctx.modified() + wctx.added())):
1279 raise util.Abort(
1279 raise util.Abort(
1280 _("can't commit subrepos without .hgsub"))
1280 _("can't commit subrepos without .hgsub"))
1281 changes[0].insert(0, '.hgsubstate')
1281 changes[0].insert(0, '.hgsubstate')
1282
1282
1283 elif '.hgsub' in changes[2]:
1283 elif '.hgsub' in changes[2]:
1284 # clean up .hgsubstate when .hgsub is removed
1284 # clean up .hgsubstate when .hgsub is removed
1285 if ('.hgsubstate' in wctx and
1285 if ('.hgsubstate' in wctx and
1286 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1286 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1287 changes[2].insert(0, '.hgsubstate')
1287 changes[2].insert(0, '.hgsubstate')
1288
1288
1289 # make sure all explicit patterns are matched
1289 # make sure all explicit patterns are matched
1290 if not force and match.files():
1290 if not force and match.files():
1291 matched = set(changes[0] + changes[1] + changes[2])
1291 matched = set(changes[0] + changes[1] + changes[2])
1292
1292
1293 for f in match.files():
1293 for f in match.files():
1294 f = self.dirstate.normalize(f)
1294 f = self.dirstate.normalize(f)
1295 if f == '.' or f in matched or f in wctx.substate:
1295 if f == '.' or f in matched or f in wctx.substate:
1296 continue
1296 continue
1297 if f in changes[3]: # missing
1297 if f in changes[3]: # missing
1298 fail(f, _('file not found!'))
1298 fail(f, _('file not found!'))
1299 if f in vdirs: # visited directory
1299 if f in vdirs: # visited directory
1300 d = f + '/'
1300 d = f + '/'
1301 for mf in matched:
1301 for mf in matched:
1302 if mf.startswith(d):
1302 if mf.startswith(d):
1303 break
1303 break
1304 else:
1304 else:
1305 fail(f, _("no match under directory!"))
1305 fail(f, _("no match under directory!"))
1306 elif f not in self.dirstate:
1306 elif f not in self.dirstate:
1307 fail(f, _("file not tracked!"))
1307 fail(f, _("file not tracked!"))
1308
1308
1309 cctx = context.workingctx(self, text, user, date, extra, changes)
1309 cctx = context.workingctx(self, text, user, date, extra, changes)
1310
1310
1311 if (not force and not extra.get("close") and not merge
1311 if (not force and not extra.get("close") and not merge
1312 and not cctx.files()
1312 and not cctx.files()
1313 and wctx.branch() == wctx.p1().branch()):
1313 and wctx.branch() == wctx.p1().branch()):
1314 return None
1314 return None
1315
1315
1316 if merge and cctx.deleted():
1316 if merge and cctx.deleted():
1317 raise util.Abort(_("cannot commit merge with missing files"))
1317 raise util.Abort(_("cannot commit merge with missing files"))
1318
1318
1319 ms = mergemod.mergestate(self)
1319 ms = mergemod.mergestate(self)
1320 for f in changes[0]:
1320 for f in changes[0]:
1321 if f in ms and ms[f] == 'u':
1321 if f in ms and ms[f] == 'u':
1322 raise util.Abort(_("unresolved merge conflicts "
1322 raise util.Abort(_("unresolved merge conflicts "
1323 "(see hg help resolve)"))
1323 "(see hg help resolve)"))
1324
1324
1325 if editor:
1325 if editor:
1326 cctx._text = editor(self, cctx, subs)
1326 cctx._text = editor(self, cctx, subs)
1327 edited = (text != cctx._text)
1327 edited = (text != cctx._text)
1328
1328
1329 # Save commit message in case this transaction gets rolled back
1329 # Save commit message in case this transaction gets rolled back
1330 # (e.g. by a pretxncommit hook). Leave the content alone on
1330 # (e.g. by a pretxncommit hook). Leave the content alone on
1331 # the assumption that the user will use the same editor again.
1331 # the assumption that the user will use the same editor again.
1332 msgfn = self.savecommitmessage(cctx._text)
1332 msgfn = self.savecommitmessage(cctx._text)
1333
1333
1334 # commit subs and write new state
1334 # commit subs and write new state
1335 if subs:
1335 if subs:
1336 for s in sorted(commitsubs):
1336 for s in sorted(commitsubs):
1337 sub = wctx.sub(s)
1337 sub = wctx.sub(s)
1338 self.ui.status(_('committing subrepository %s\n') %
1338 self.ui.status(_('committing subrepository %s\n') %
1339 subrepo.subrelpath(sub))
1339 subrepo.subrelpath(sub))
1340 sr = sub.commit(cctx._text, user, date)
1340 sr = sub.commit(cctx._text, user, date)
1341 newstate[s] = (newstate[s][0], sr)
1341 newstate[s] = (newstate[s][0], sr)
1342 subrepo.writestate(self, newstate)
1342 subrepo.writestate(self, newstate)
1343
1343
1344 p1, p2 = self.dirstate.parents()
1344 p1, p2 = self.dirstate.parents()
1345 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1345 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1346 try:
1346 try:
1347 self.hook("precommit", throw=True, parent1=hookp1,
1347 self.hook("precommit", throw=True, parent1=hookp1,
1348 parent2=hookp2)
1348 parent2=hookp2)
1349 ret = self.commitctx(cctx, True)
1349 ret = self.commitctx(cctx, True)
1350 except: # re-raises
1350 except: # re-raises
1351 if edited:
1351 if edited:
1352 self.ui.write(
1352 self.ui.write(
1353 _('note: commit message saved in %s\n') % msgfn)
1353 _('note: commit message saved in %s\n') % msgfn)
1354 raise
1354 raise
1355
1355
1356 # update bookmarks, dirstate and mergestate
1356 # update bookmarks, dirstate and mergestate
1357 bookmarks.update(self, [p1, p2], ret)
1357 bookmarks.update(self, [p1, p2], ret)
1358 cctx.markcommitted(ret)
1358 cctx.markcommitted(ret)
1359 ms.reset()
1359 ms.reset()
1360 finally:
1360 finally:
1361 wlock.release()
1361 wlock.release()
1362
1362
1363 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1363 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1364 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1364 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1365 self._afterlock(commithook)
1365 self._afterlock(commithook)
1366 return ret
1366 return ret
1367
1367
1368 @unfilteredmethod
1368 @unfilteredmethod
1369 def commitctx(self, ctx, error=False):
1369 def commitctx(self, ctx, error=False):
1370 """Add a new revision to current repository.
1370 """Add a new revision to current repository.
1371 Revision information is passed via the context argument.
1371 Revision information is passed via the context argument.
1372 """
1372 """
1373
1373
1374 tr = lock = None
1374 tr = lock = None
1375 removed = list(ctx.removed())
1375 removed = list(ctx.removed())
1376 p1, p2 = ctx.p1(), ctx.p2()
1376 p1, p2 = ctx.p1(), ctx.p2()
1377 user = ctx.user()
1377 user = ctx.user()
1378
1378
1379 lock = self.lock()
1379 lock = self.lock()
1380 try:
1380 try:
1381 tr = self.transaction("commit")
1381 tr = self.transaction("commit")
1382 trp = weakref.proxy(tr)
1382 trp = weakref.proxy(tr)
1383
1383
1384 if ctx.files():
1384 if ctx.files():
1385 m1 = p1.manifest().copy()
1385 m1 = p1.manifest().copy()
1386 m2 = p2.manifest()
1386 m2 = p2.manifest()
1387
1387
1388 # check in files
1388 # check in files
1389 new = {}
1389 new = {}
1390 changed = []
1390 changed = []
1391 linkrev = len(self)
1391 linkrev = len(self)
1392 for f in sorted(ctx.modified() + ctx.added()):
1392 for f in sorted(ctx.modified() + ctx.added()):
1393 self.ui.note(f + "\n")
1393 self.ui.note(f + "\n")
1394 try:
1394 try:
1395 fctx = ctx[f]
1395 fctx = ctx[f]
1396 if fctx is None:
1396 if fctx is None:
1397 removed.append(f)
1397 removed.append(f)
1398 else:
1398 else:
1399 new[f] = self._filecommit(fctx, m1, m2, linkrev,
1399 new[f] = self._filecommit(fctx, m1, m2, linkrev,
1400 trp, changed)
1400 trp, changed)
1401 m1.set(f, fctx.flags())
1401 m1.set(f, fctx.flags())
1402 except OSError, inst:
1402 except OSError, inst:
1403 self.ui.warn(_("trouble committing %s!\n") % f)
1403 self.ui.warn(_("trouble committing %s!\n") % f)
1404 raise
1404 raise
1405 except IOError, inst:
1405 except IOError, inst:
1406 errcode = getattr(inst, 'errno', errno.ENOENT)
1406 errcode = getattr(inst, 'errno', errno.ENOENT)
1407 if error or errcode and errcode != errno.ENOENT:
1407 if error or errcode and errcode != errno.ENOENT:
1408 self.ui.warn(_("trouble committing %s!\n") % f)
1408 self.ui.warn(_("trouble committing %s!\n") % f)
1409 raise
1409 raise
1410
1410
1411 # update manifest
1411 # update manifest
1412 m1.update(new)
1412 m1.update(new)
1413 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1413 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1414 drop = [f for f in removed if f in m1]
1414 drop = [f for f in removed if f in m1]
1415 for f in drop:
1415 for f in drop:
1416 del m1[f]
1416 del m1[f]
1417 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1417 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1418 p2.manifestnode(), (new, drop))
1418 p2.manifestnode(), (new, drop))
1419 files = changed + removed
1419 files = changed + removed
1420 else:
1420 else:
1421 mn = p1.manifestnode()
1421 mn = p1.manifestnode()
1422 files = []
1422 files = []
1423
1423
1424 # update changelog
1424 # update changelog
1425 self.changelog.delayupdate()
1425 self.changelog.delayupdate()
1426 n = self.changelog.add(mn, files, ctx.description(),
1426 n = self.changelog.add(mn, files, ctx.description(),
1427 trp, p1.node(), p2.node(),
1427 trp, p1.node(), p2.node(),
1428 user, ctx.date(), ctx.extra().copy())
1428 user, ctx.date(), ctx.extra().copy())
1429 p = lambda: self.changelog.writepending() and self.root or ""
1429 p = lambda: self.changelog.writepending() and self.root or ""
1430 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1430 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1431 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1431 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1432 parent2=xp2, pending=p)
1432 parent2=xp2, pending=p)
1433 self.changelog.finalize(trp)
1433 self.changelog.finalize(trp)
1434 # set the new commit is proper phase
1434 # set the new commit is proper phase
1435 targetphase = subrepo.newcommitphase(self.ui, ctx)
1435 targetphase = subrepo.newcommitphase(self.ui, ctx)
1436 if targetphase:
1436 if targetphase:
1437 # retract boundary do not alter parent changeset.
1437 # retract boundary do not alter parent changeset.
1438 # if a parent have higher the resulting phase will
1438 # if a parent have higher the resulting phase will
1439 # be compliant anyway
1439 # be compliant anyway
1440 #
1440 #
1441 # if minimal phase was 0 we don't need to retract anything
1441 # if minimal phase was 0 we don't need to retract anything
1442 phases.retractboundary(self, tr, targetphase, [n])
1442 phases.retractboundary(self, tr, targetphase, [n])
1443 tr.close()
1443 tr.close()
1444 branchmap.updatecache(self.filtered('served'))
1444 branchmap.updatecache(self.filtered('served'))
1445 return n
1445 return n
1446 finally:
1446 finally:
1447 if tr:
1447 if tr:
1448 tr.release()
1448 tr.release()
1449 lock.release()
1449 lock.release()
1450
1450
1451 @unfilteredmethod
1451 @unfilteredmethod
1452 def destroying(self):
1452 def destroying(self):
1453 '''Inform the repository that nodes are about to be destroyed.
1453 '''Inform the repository that nodes are about to be destroyed.
1454 Intended for use by strip and rollback, so there's a common
1454 Intended for use by strip and rollback, so there's a common
1455 place for anything that has to be done before destroying history.
1455 place for anything that has to be done before destroying history.
1456
1456
1457 This is mostly useful for saving state that is in memory and waiting
1457 This is mostly useful for saving state that is in memory and waiting
1458 to be flushed when the current lock is released. Because a call to
1458 to be flushed when the current lock is released. Because a call to
1459 destroyed is imminent, the repo will be invalidated causing those
1459 destroyed is imminent, the repo will be invalidated causing those
1460 changes to stay in memory (waiting for the next unlock), or vanish
1460 changes to stay in memory (waiting for the next unlock), or vanish
1461 completely.
1461 completely.
1462 '''
1462 '''
1463 # When using the same lock to commit and strip, the phasecache is left
1463 # When using the same lock to commit and strip, the phasecache is left
1464 # dirty after committing. Then when we strip, the repo is invalidated,
1464 # dirty after committing. Then when we strip, the repo is invalidated,
1465 # causing those changes to disappear.
1465 # causing those changes to disappear.
1466 if '_phasecache' in vars(self):
1466 if '_phasecache' in vars(self):
1467 self._phasecache.write()
1467 self._phasecache.write()
1468
1468
1469 @unfilteredmethod
1469 @unfilteredmethod
1470 def destroyed(self):
1470 def destroyed(self):
1471 '''Inform the repository that nodes have been destroyed.
1471 '''Inform the repository that nodes have been destroyed.
1472 Intended for use by strip and rollback, so there's a common
1472 Intended for use by strip and rollback, so there's a common
1473 place for anything that has to be done after destroying history.
1473 place for anything that has to be done after destroying history.
1474 '''
1474 '''
1475 # When one tries to:
1475 # When one tries to:
1476 # 1) destroy nodes thus calling this method (e.g. strip)
1476 # 1) destroy nodes thus calling this method (e.g. strip)
1477 # 2) use phasecache somewhere (e.g. commit)
1477 # 2) use phasecache somewhere (e.g. commit)
1478 #
1478 #
1479 # then 2) will fail because the phasecache contains nodes that were
1479 # then 2) will fail because the phasecache contains nodes that were
1480 # removed. We can either remove phasecache from the filecache,
1480 # removed. We can either remove phasecache from the filecache,
1481 # causing it to reload next time it is accessed, or simply filter
1481 # causing it to reload next time it is accessed, or simply filter
1482 # the removed nodes now and write the updated cache.
1482 # the removed nodes now and write the updated cache.
1483 self._phasecache.filterunknown(self)
1483 self._phasecache.filterunknown(self)
1484 self._phasecache.write()
1484 self._phasecache.write()
1485
1485
1486 # update the 'served' branch cache to help read only server process
1486 # update the 'served' branch cache to help read only server process
1487 # Thanks to branchcache collaboration this is done from the nearest
1487 # Thanks to branchcache collaboration this is done from the nearest
1488 # filtered subset and it is expected to be fast.
1488 # filtered subset and it is expected to be fast.
1489 branchmap.updatecache(self.filtered('served'))
1489 branchmap.updatecache(self.filtered('served'))
1490
1490
1491 # Ensure the persistent tag cache is updated. Doing it now
1491 # Ensure the persistent tag cache is updated. Doing it now
1492 # means that the tag cache only has to worry about destroyed
1492 # means that the tag cache only has to worry about destroyed
1493 # heads immediately after a strip/rollback. That in turn
1493 # heads immediately after a strip/rollback. That in turn
1494 # guarantees that "cachetip == currenttip" (comparing both rev
1494 # guarantees that "cachetip == currenttip" (comparing both rev
1495 # and node) always means no nodes have been added or destroyed.
1495 # and node) always means no nodes have been added or destroyed.
1496
1496
1497 # XXX this is suboptimal when qrefresh'ing: we strip the current
1497 # XXX this is suboptimal when qrefresh'ing: we strip the current
1498 # head, refresh the tag cache, then immediately add a new head.
1498 # head, refresh the tag cache, then immediately add a new head.
1499 # But I think doing it this way is necessary for the "instant
1499 # But I think doing it this way is necessary for the "instant
1500 # tag cache retrieval" case to work.
1500 # tag cache retrieval" case to work.
1501 self.invalidate()
1501 self.invalidate()
1502
1502
1503 def walk(self, match, node=None):
1503 def walk(self, match, node=None):
1504 '''
1504 '''
1505 walk recursively through the directory tree or a given
1505 walk recursively through the directory tree or a given
1506 changeset, finding all files matched by the match
1506 changeset, finding all files matched by the match
1507 function
1507 function
1508 '''
1508 '''
1509 return self[node].walk(match)
1509 return self[node].walk(match)
1510
1510
1511 def status(self, node1='.', node2=None, match=None,
1511 def status(self, node1='.', node2=None, match=None,
1512 ignored=False, clean=False, unknown=False,
1512 ignored=False, clean=False, unknown=False,
1513 listsubrepos=False):
1513 listsubrepos=False):
1514 '''a convenience method that calls node1.status(node2)'''
1514 '''a convenience method that calls node1.status(node2)'''
1515 return self[node1].status(node2, match, ignored, clean, unknown,
1515 return self[node1].status(node2, match, ignored, clean, unknown,
1516 listsubrepos)
1516 listsubrepos)
1517
1517
1518 def heads(self, start=None):
1518 def heads(self, start=None):
1519 heads = self.changelog.heads(start)
1519 heads = self.changelog.heads(start)
1520 # sort the output in rev descending order
1520 # sort the output in rev descending order
1521 return sorted(heads, key=self.changelog.rev, reverse=True)
1521 return sorted(heads, key=self.changelog.rev, reverse=True)
1522
1522
1523 def branchheads(self, branch=None, start=None, closed=False):
1523 def branchheads(self, branch=None, start=None, closed=False):
1524 '''return a (possibly filtered) list of heads for the given branch
1524 '''return a (possibly filtered) list of heads for the given branch
1525
1525
1526 Heads are returned in topological order, from newest to oldest.
1526 Heads are returned in topological order, from newest to oldest.
1527 If branch is None, use the dirstate branch.
1527 If branch is None, use the dirstate branch.
1528 If start is not None, return only heads reachable from start.
1528 If start is not None, return only heads reachable from start.
1529 If closed is True, return heads that are marked as closed as well.
1529 If closed is True, return heads that are marked as closed as well.
1530 '''
1530 '''
1531 if branch is None:
1531 if branch is None:
1532 branch = self[None].branch()
1532 branch = self[None].branch()
1533 branches = self.branchmap()
1533 branches = self.branchmap()
1534 if branch not in branches:
1534 if branch not in branches:
1535 return []
1535 return []
1536 # the cache returns heads ordered lowest to highest
1536 # the cache returns heads ordered lowest to highest
1537 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1537 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1538 if start is not None:
1538 if start is not None:
1539 # filter out the heads that cannot be reached from startrev
1539 # filter out the heads that cannot be reached from startrev
1540 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1540 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1541 bheads = [h for h in bheads if h in fbheads]
1541 bheads = [h for h in bheads if h in fbheads]
1542 return bheads
1542 return bheads
1543
1543
1544 def branches(self, nodes):
1544 def branches(self, nodes):
1545 if not nodes:
1545 if not nodes:
1546 nodes = [self.changelog.tip()]
1546 nodes = [self.changelog.tip()]
1547 b = []
1547 b = []
1548 for n in nodes:
1548 for n in nodes:
1549 t = n
1549 t = n
1550 while True:
1550 while True:
1551 p = self.changelog.parents(n)
1551 p = self.changelog.parents(n)
1552 if p[1] != nullid or p[0] == nullid:
1552 if p[1] != nullid or p[0] == nullid:
1553 b.append((t, n, p[0], p[1]))
1553 b.append((t, n, p[0], p[1]))
1554 break
1554 break
1555 n = p[0]
1555 n = p[0]
1556 return b
1556 return b
1557
1557
1558 def between(self, pairs):
1558 def between(self, pairs):
1559 r = []
1559 r = []
1560
1560
1561 for top, bottom in pairs:
1561 for top, bottom in pairs:
1562 n, l, i = top, [], 0
1562 n, l, i = top, [], 0
1563 f = 1
1563 f = 1
1564
1564
1565 while n != bottom and n != nullid:
1565 while n != bottom and n != nullid:
1566 p = self.changelog.parents(n)[0]
1566 p = self.changelog.parents(n)[0]
1567 if i == f:
1567 if i == f:
1568 l.append(n)
1568 l.append(n)
1569 f = f * 2
1569 f = f * 2
1570 n = p
1570 n = p
1571 i += 1
1571 i += 1
1572
1572
1573 r.append(l)
1573 r.append(l)
1574
1574
1575 return r
1575 return r
1576
1576
1577 def pull(self, remote, heads=None, force=False):
1577 def pull(self, remote, heads=None, force=False):
1578 return exchange.pull (self, remote, heads, force)
1578 return exchange.pull (self, remote, heads, force)
1579
1579
1580 def checkpush(self, pushop):
1580 def checkpush(self, pushop):
1581 """Extensions can override this function if additional checks have
1581 """Extensions can override this function if additional checks have
1582 to be performed before pushing, or call it if they override push
1582 to be performed before pushing, or call it if they override push
1583 command.
1583 command.
1584 """
1584 """
1585 pass
1585 pass
1586
1586
1587 @unfilteredpropertycache
1587 @unfilteredpropertycache
1588 def prepushoutgoinghooks(self):
1588 def prepushoutgoinghooks(self):
1589 """Return util.hooks consists of "(repo, remote, outgoing)"
1589 """Return util.hooks consists of "(repo, remote, outgoing)"
1590 functions, which are called before pushing changesets.
1590 functions, which are called before pushing changesets.
1591 """
1591 """
1592 return util.hooks()
1592 return util.hooks()
1593
1593
1594 def push(self, remote, force=False, revs=None, newbranch=False):
1595 return exchange.push(self, remote, force, revs, newbranch).cgresult
1596
1597 def stream_in(self, remote, requirements):
1594 def stream_in(self, remote, requirements):
1598 lock = self.lock()
1595 lock = self.lock()
1599 try:
1596 try:
1600 # Save remote branchmap. We will use it later
1597 # Save remote branchmap. We will use it later
1601 # to speed up branchcache creation
1598 # to speed up branchcache creation
1602 rbranchmap = None
1599 rbranchmap = None
1603 if remote.capable("branchmap"):
1600 if remote.capable("branchmap"):
1604 rbranchmap = remote.branchmap()
1601 rbranchmap = remote.branchmap()
1605
1602
1606 fp = remote.stream_out()
1603 fp = remote.stream_out()
1607 l = fp.readline()
1604 l = fp.readline()
1608 try:
1605 try:
1609 resp = int(l)
1606 resp = int(l)
1610 except ValueError:
1607 except ValueError:
1611 raise error.ResponseError(
1608 raise error.ResponseError(
1612 _('unexpected response from remote server:'), l)
1609 _('unexpected response from remote server:'), l)
1613 if resp == 1:
1610 if resp == 1:
1614 raise util.Abort(_('operation forbidden by server'))
1611 raise util.Abort(_('operation forbidden by server'))
1615 elif resp == 2:
1612 elif resp == 2:
1616 raise util.Abort(_('locking the remote repository failed'))
1613 raise util.Abort(_('locking the remote repository failed'))
1617 elif resp != 0:
1614 elif resp != 0:
1618 raise util.Abort(_('the server sent an unknown error code'))
1615 raise util.Abort(_('the server sent an unknown error code'))
1619 self.ui.status(_('streaming all changes\n'))
1616 self.ui.status(_('streaming all changes\n'))
1620 l = fp.readline()
1617 l = fp.readline()
1621 try:
1618 try:
1622 total_files, total_bytes = map(int, l.split(' ', 1))
1619 total_files, total_bytes = map(int, l.split(' ', 1))
1623 except (ValueError, TypeError):
1620 except (ValueError, TypeError):
1624 raise error.ResponseError(
1621 raise error.ResponseError(
1625 _('unexpected response from remote server:'), l)
1622 _('unexpected response from remote server:'), l)
1626 self.ui.status(_('%d files to transfer, %s of data\n') %
1623 self.ui.status(_('%d files to transfer, %s of data\n') %
1627 (total_files, util.bytecount(total_bytes)))
1624 (total_files, util.bytecount(total_bytes)))
1628 handled_bytes = 0
1625 handled_bytes = 0
1629 self.ui.progress(_('clone'), 0, total=total_bytes)
1626 self.ui.progress(_('clone'), 0, total=total_bytes)
1630 start = time.time()
1627 start = time.time()
1631
1628
1632 tr = self.transaction(_('clone'))
1629 tr = self.transaction(_('clone'))
1633 try:
1630 try:
1634 for i in xrange(total_files):
1631 for i in xrange(total_files):
1635 # XXX doesn't support '\n' or '\r' in filenames
1632 # XXX doesn't support '\n' or '\r' in filenames
1636 l = fp.readline()
1633 l = fp.readline()
1637 try:
1634 try:
1638 name, size = l.split('\0', 1)
1635 name, size = l.split('\0', 1)
1639 size = int(size)
1636 size = int(size)
1640 except (ValueError, TypeError):
1637 except (ValueError, TypeError):
1641 raise error.ResponseError(
1638 raise error.ResponseError(
1642 _('unexpected response from remote server:'), l)
1639 _('unexpected response from remote server:'), l)
1643 if self.ui.debugflag:
1640 if self.ui.debugflag:
1644 self.ui.debug('adding %s (%s)\n' %
1641 self.ui.debug('adding %s (%s)\n' %
1645 (name, util.bytecount(size)))
1642 (name, util.bytecount(size)))
1646 # for backwards compat, name was partially encoded
1643 # for backwards compat, name was partially encoded
1647 ofp = self.sopener(store.decodedir(name), 'w')
1644 ofp = self.sopener(store.decodedir(name), 'w')
1648 for chunk in util.filechunkiter(fp, limit=size):
1645 for chunk in util.filechunkiter(fp, limit=size):
1649 handled_bytes += len(chunk)
1646 handled_bytes += len(chunk)
1650 self.ui.progress(_('clone'), handled_bytes,
1647 self.ui.progress(_('clone'), handled_bytes,
1651 total=total_bytes)
1648 total=total_bytes)
1652 ofp.write(chunk)
1649 ofp.write(chunk)
1653 ofp.close()
1650 ofp.close()
1654 tr.close()
1651 tr.close()
1655 finally:
1652 finally:
1656 tr.release()
1653 tr.release()
1657
1654
1658 # Writing straight to files circumvented the inmemory caches
1655 # Writing straight to files circumvented the inmemory caches
1659 self.invalidate()
1656 self.invalidate()
1660
1657
1661 elapsed = time.time() - start
1658 elapsed = time.time() - start
1662 if elapsed <= 0:
1659 if elapsed <= 0:
1663 elapsed = 0.001
1660 elapsed = 0.001
1664 self.ui.progress(_('clone'), None)
1661 self.ui.progress(_('clone'), None)
1665 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1662 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1666 (util.bytecount(total_bytes), elapsed,
1663 (util.bytecount(total_bytes), elapsed,
1667 util.bytecount(total_bytes / elapsed)))
1664 util.bytecount(total_bytes / elapsed)))
1668
1665
1669 # new requirements = old non-format requirements +
1666 # new requirements = old non-format requirements +
1670 # new format-related
1667 # new format-related
1671 # requirements from the streamed-in repository
1668 # requirements from the streamed-in repository
1672 requirements.update(set(self.requirements) - self.supportedformats)
1669 requirements.update(set(self.requirements) - self.supportedformats)
1673 self._applyrequirements(requirements)
1670 self._applyrequirements(requirements)
1674 self._writerequirements()
1671 self._writerequirements()
1675
1672
1676 if rbranchmap:
1673 if rbranchmap:
1677 rbheads = []
1674 rbheads = []
1678 for bheads in rbranchmap.itervalues():
1675 for bheads in rbranchmap.itervalues():
1679 rbheads.extend(bheads)
1676 rbheads.extend(bheads)
1680
1677
1681 if rbheads:
1678 if rbheads:
1682 rtiprev = max((int(self.changelog.rev(node))
1679 rtiprev = max((int(self.changelog.rev(node))
1683 for node in rbheads))
1680 for node in rbheads))
1684 cache = branchmap.branchcache(rbranchmap,
1681 cache = branchmap.branchcache(rbranchmap,
1685 self[rtiprev].node(),
1682 self[rtiprev].node(),
1686 rtiprev)
1683 rtiprev)
1687 # Try to stick it as low as possible
1684 # Try to stick it as low as possible
1688 # filter above served are unlikely to be fetch from a clone
1685 # filter above served are unlikely to be fetch from a clone
1689 for candidate in ('base', 'immutable', 'served'):
1686 for candidate in ('base', 'immutable', 'served'):
1690 rview = self.filtered(candidate)
1687 rview = self.filtered(candidate)
1691 if cache.validfor(rview):
1688 if cache.validfor(rview):
1692 self._branchcaches[candidate] = cache
1689 self._branchcaches[candidate] = cache
1693 cache.write(rview)
1690 cache.write(rview)
1694 break
1691 break
1695 self.invalidate()
1692 self.invalidate()
1696 return len(self.heads()) + 1
1693 return len(self.heads()) + 1
1697 finally:
1694 finally:
1698 lock.release()
1695 lock.release()
1699
1696
1700 def clone(self, remote, heads=[], stream=False):
1697 def clone(self, remote, heads=[], stream=False):
1701 '''clone remote repository.
1698 '''clone remote repository.
1702
1699
1703 keyword arguments:
1700 keyword arguments:
1704 heads: list of revs to clone (forces use of pull)
1701 heads: list of revs to clone (forces use of pull)
1705 stream: use streaming clone if possible'''
1702 stream: use streaming clone if possible'''
1706
1703
1707 # now, all clients that can request uncompressed clones can
1704 # now, all clients that can request uncompressed clones can
1708 # read repo formats supported by all servers that can serve
1705 # read repo formats supported by all servers that can serve
1709 # them.
1706 # them.
1710
1707
1711 # if revlog format changes, client will have to check version
1708 # if revlog format changes, client will have to check version
1712 # and format flags on "stream" capability, and use
1709 # and format flags on "stream" capability, and use
1713 # uncompressed only if compatible.
1710 # uncompressed only if compatible.
1714
1711
1715 if not stream:
1712 if not stream:
1716 # if the server explicitly prefers to stream (for fast LANs)
1713 # if the server explicitly prefers to stream (for fast LANs)
1717 stream = remote.capable('stream-preferred')
1714 stream = remote.capable('stream-preferred')
1718
1715
1719 if stream and not heads:
1716 if stream and not heads:
1720 # 'stream' means remote revlog format is revlogv1 only
1717 # 'stream' means remote revlog format is revlogv1 only
1721 if remote.capable('stream'):
1718 if remote.capable('stream'):
1722 return self.stream_in(remote, set(('revlogv1',)))
1719 return self.stream_in(remote, set(('revlogv1',)))
1723 # otherwise, 'streamreqs' contains the remote revlog format
1720 # otherwise, 'streamreqs' contains the remote revlog format
1724 streamreqs = remote.capable('streamreqs')
1721 streamreqs = remote.capable('streamreqs')
1725 if streamreqs:
1722 if streamreqs:
1726 streamreqs = set(streamreqs.split(','))
1723 streamreqs = set(streamreqs.split(','))
1727 # if we support it, stream in and adjust our requirements
1724 # if we support it, stream in and adjust our requirements
1728 if not streamreqs - self.supportedformats:
1725 if not streamreqs - self.supportedformats:
1729 return self.stream_in(remote, streamreqs)
1726 return self.stream_in(remote, streamreqs)
1730 return self.pull(remote, heads)
1727 return self.pull(remote, heads)
1731
1728
1732 def pushkey(self, namespace, key, old, new):
1729 def pushkey(self, namespace, key, old, new):
1733 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1730 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1734 old=old, new=new)
1731 old=old, new=new)
1735 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1732 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1736 ret = pushkey.push(self, namespace, key, old, new)
1733 ret = pushkey.push(self, namespace, key, old, new)
1737 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1734 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1738 ret=ret)
1735 ret=ret)
1739 return ret
1736 return ret
1740
1737
1741 def listkeys(self, namespace):
1738 def listkeys(self, namespace):
1742 self.hook('prelistkeys', throw=True, namespace=namespace)
1739 self.hook('prelistkeys', throw=True, namespace=namespace)
1743 self.ui.debug('listing keys for "%s"\n' % namespace)
1740 self.ui.debug('listing keys for "%s"\n' % namespace)
1744 values = pushkey.list(self, namespace)
1741 values = pushkey.list(self, namespace)
1745 self.hook('listkeys', namespace=namespace, values=values)
1742 self.hook('listkeys', namespace=namespace, values=values)
1746 return values
1743 return values
1747
1744
1748 def debugwireargs(self, one, two, three=None, four=None, five=None):
1745 def debugwireargs(self, one, two, three=None, four=None, five=None):
1749 '''used to test argument passing over the wire'''
1746 '''used to test argument passing over the wire'''
1750 return "%s %s %s %s %s" % (one, two, three, four, five)
1747 return "%s %s %s %s %s" % (one, two, three, four, five)
1751
1748
1752 def savecommitmessage(self, text):
1749 def savecommitmessage(self, text):
1753 fp = self.opener('last-message.txt', 'wb')
1750 fp = self.opener('last-message.txt', 'wb')
1754 try:
1751 try:
1755 fp.write(text)
1752 fp.write(text)
1756 finally:
1753 finally:
1757 fp.close()
1754 fp.close()
1758 return self.pathto(fp.name[len(self.root) + 1:])
1755 return self.pathto(fp.name[len(self.root) + 1:])
1759
1756
1760 # used to avoid circular references so destructors work
1757 # used to avoid circular references so destructors work
1761 def aftertrans(files):
1758 def aftertrans(files):
1762 renamefiles = [tuple(t) for t in files]
1759 renamefiles = [tuple(t) for t in files]
1763 def a():
1760 def a():
1764 for vfs, src, dest in renamefiles:
1761 for vfs, src, dest in renamefiles:
1765 try:
1762 try:
1766 vfs.rename(src, dest)
1763 vfs.rename(src, dest)
1767 except OSError: # journal file does not yet exist
1764 except OSError: # journal file does not yet exist
1768 pass
1765 pass
1769 return a
1766 return a
1770
1767
1771 def undoname(fn):
1768 def undoname(fn):
1772 base, name = os.path.split(fn)
1769 base, name = os.path.split(fn)
1773 assert name.startswith('journal')
1770 assert name.startswith('journal')
1774 return os.path.join(base, name.replace('journal', 'undo', 1))
1771 return os.path.join(base, name.replace('journal', 'undo', 1))
1775
1772
1776 def instance(ui, path, create):
1773 def instance(ui, path, create):
1777 return localrepository(ui, util.urllocalpath(path), create)
1774 return localrepository(ui, util.urllocalpath(path), create)
1778
1775
1779 def islocal(path):
1776 def islocal(path):
1780 return True
1777 return True
General Comments 0
You need to be logged in to leave comments. Login now