##// END OF EJS Templates
localrepo: use "vfs.readlink()" instead of "os.readlink()"
FUJIWARA Katsunori -
r18950:647e3b0c default
parent child Browse files
Show More
@@ -1,2594 +1,2594 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import lock, transaction, store, encoding, base85
11 import lock, transaction, store, encoding, base85
12 import scmutil, util, extensions, hook, error, revset
12 import scmutil, util, extensions, hook, error, revset
13 import match as matchmod
13 import match as matchmod
14 import merge as mergemod
14 import merge as mergemod
15 import tags as tagsmod
15 import tags as tagsmod
16 from lock import release
16 from lock import release
17 import weakref, errno, os, time, inspect
17 import weakref, errno, os, time, inspect
18 import branchmap
18 import branchmap
19 propertycache = util.propertycache
19 propertycache = util.propertycache
20 filecache = scmutil.filecache
20 filecache = scmutil.filecache
21
21
22 class repofilecache(filecache):
22 class repofilecache(filecache):
23 """All filecache usage on repo are done for logic that should be unfiltered
23 """All filecache usage on repo are done for logic that should be unfiltered
24 """
24 """
25
25
26 def __get__(self, repo, type=None):
26 def __get__(self, repo, type=None):
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 def __set__(self, repo, value):
28 def __set__(self, repo, value):
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 def __delete__(self, repo):
30 def __delete__(self, repo):
31 return super(repofilecache, self).__delete__(repo.unfiltered())
31 return super(repofilecache, self).__delete__(repo.unfiltered())
32
32
33 class storecache(repofilecache):
33 class storecache(repofilecache):
34 """filecache for files in the store"""
34 """filecache for files in the store"""
35 def join(self, obj, fname):
35 def join(self, obj, fname):
36 return obj.sjoin(fname)
36 return obj.sjoin(fname)
37
37
38 class unfilteredpropertycache(propertycache):
38 class unfilteredpropertycache(propertycache):
39 """propertycache that apply to unfiltered repo only"""
39 """propertycache that apply to unfiltered repo only"""
40
40
41 def __get__(self, repo, type=None):
41 def __get__(self, repo, type=None):
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43
43
44 class filteredpropertycache(propertycache):
44 class filteredpropertycache(propertycache):
45 """propertycache that must take filtering in account"""
45 """propertycache that must take filtering in account"""
46
46
47 def cachevalue(self, obj, value):
47 def cachevalue(self, obj, value):
48 object.__setattr__(obj, self.name, value)
48 object.__setattr__(obj, self.name, value)
49
49
50
50
51 def hasunfilteredcache(repo, name):
51 def hasunfilteredcache(repo, name):
52 """check if a repo has an unfilteredpropertycache value for <name>"""
52 """check if a repo has an unfilteredpropertycache value for <name>"""
53 return name in vars(repo.unfiltered())
53 return name in vars(repo.unfiltered())
54
54
55 def unfilteredmethod(orig):
55 def unfilteredmethod(orig):
56 """decorate method that always need to be run on unfiltered version"""
56 """decorate method that always need to be run on unfiltered version"""
57 def wrapper(repo, *args, **kwargs):
57 def wrapper(repo, *args, **kwargs):
58 return orig(repo.unfiltered(), *args, **kwargs)
58 return orig(repo.unfiltered(), *args, **kwargs)
59 return wrapper
59 return wrapper
60
60
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63
63
64 class localpeer(peer.peerrepository):
64 class localpeer(peer.peerrepository):
65 '''peer for a local repo; reflects only the most recent API'''
65 '''peer for a local repo; reflects only the most recent API'''
66
66
67 def __init__(self, repo, caps=MODERNCAPS):
67 def __init__(self, repo, caps=MODERNCAPS):
68 peer.peerrepository.__init__(self)
68 peer.peerrepository.__init__(self)
69 self._repo = repo.filtered('served')
69 self._repo = repo.filtered('served')
70 self.ui = repo.ui
70 self.ui = repo.ui
71 self._caps = repo._restrictcapabilities(caps)
71 self._caps = repo._restrictcapabilities(caps)
72 self.requirements = repo.requirements
72 self.requirements = repo.requirements
73 self.supportedformats = repo.supportedformats
73 self.supportedformats = repo.supportedformats
74
74
75 def close(self):
75 def close(self):
76 self._repo.close()
76 self._repo.close()
77
77
78 def _capabilities(self):
78 def _capabilities(self):
79 return self._caps
79 return self._caps
80
80
81 def local(self):
81 def local(self):
82 return self._repo
82 return self._repo
83
83
84 def canpush(self):
84 def canpush(self):
85 return True
85 return True
86
86
87 def url(self):
87 def url(self):
88 return self._repo.url()
88 return self._repo.url()
89
89
90 def lookup(self, key):
90 def lookup(self, key):
91 return self._repo.lookup(key)
91 return self._repo.lookup(key)
92
92
93 def branchmap(self):
93 def branchmap(self):
94 return self._repo.branchmap()
94 return self._repo.branchmap()
95
95
96 def heads(self):
96 def heads(self):
97 return self._repo.heads()
97 return self._repo.heads()
98
98
99 def known(self, nodes):
99 def known(self, nodes):
100 return self._repo.known(nodes)
100 return self._repo.known(nodes)
101
101
102 def getbundle(self, source, heads=None, common=None):
102 def getbundle(self, source, heads=None, common=None):
103 return self._repo.getbundle(source, heads=heads, common=common)
103 return self._repo.getbundle(source, heads=heads, common=common)
104
104
105 # TODO We might want to move the next two calls into legacypeer and add
105 # TODO We might want to move the next two calls into legacypeer and add
106 # unbundle instead.
106 # unbundle instead.
107
107
108 def lock(self):
108 def lock(self):
109 return self._repo.lock()
109 return self._repo.lock()
110
110
111 def addchangegroup(self, cg, source, url):
111 def addchangegroup(self, cg, source, url):
112 return self._repo.addchangegroup(cg, source, url)
112 return self._repo.addchangegroup(cg, source, url)
113
113
114 def pushkey(self, namespace, key, old, new):
114 def pushkey(self, namespace, key, old, new):
115 return self._repo.pushkey(namespace, key, old, new)
115 return self._repo.pushkey(namespace, key, old, new)
116
116
117 def listkeys(self, namespace):
117 def listkeys(self, namespace):
118 return self._repo.listkeys(namespace)
118 return self._repo.listkeys(namespace)
119
119
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 '''used to test argument passing over the wire'''
121 '''used to test argument passing over the wire'''
122 return "%s %s %s %s %s" % (one, two, three, four, five)
122 return "%s %s %s %s %s" % (one, two, three, four, five)
123
123
124 class locallegacypeer(localpeer):
124 class locallegacypeer(localpeer):
125 '''peer extension which implements legacy methods too; used for tests with
125 '''peer extension which implements legacy methods too; used for tests with
126 restricted capabilities'''
126 restricted capabilities'''
127
127
128 def __init__(self, repo):
128 def __init__(self, repo):
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130
130
131 def branches(self, nodes):
131 def branches(self, nodes):
132 return self._repo.branches(nodes)
132 return self._repo.branches(nodes)
133
133
134 def between(self, pairs):
134 def between(self, pairs):
135 return self._repo.between(pairs)
135 return self._repo.between(pairs)
136
136
137 def changegroup(self, basenodes, source):
137 def changegroup(self, basenodes, source):
138 return self._repo.changegroup(basenodes, source)
138 return self._repo.changegroup(basenodes, source)
139
139
140 def changegroupsubset(self, bases, heads, source):
140 def changegroupsubset(self, bases, heads, source):
141 return self._repo.changegroupsubset(bases, heads, source)
141 return self._repo.changegroupsubset(bases, heads, source)
142
142
143 class localrepository(object):
143 class localrepository(object):
144
144
145 supportedformats = set(('revlogv1', 'generaldelta'))
145 supportedformats = set(('revlogv1', 'generaldelta'))
146 supported = supportedformats | set(('store', 'fncache', 'shared',
146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 'dotencode'))
147 'dotencode'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
149 requirements = ['revlogv1']
149 requirements = ['revlogv1']
150 filtername = None
150 filtername = None
151
151
152 def _baserequirements(self, create):
152 def _baserequirements(self, create):
153 return self.requirements[:]
153 return self.requirements[:]
154
154
155 def __init__(self, baseui, path=None, create=False):
155 def __init__(self, baseui, path=None, create=False):
156 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
156 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
157 self.wopener = self.wvfs
157 self.wopener = self.wvfs
158 self.root = self.wvfs.base
158 self.root = self.wvfs.base
159 self.path = self.wvfs.join(".hg")
159 self.path = self.wvfs.join(".hg")
160 self.origroot = path
160 self.origroot = path
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 self.vfs = scmutil.vfs(self.path)
162 self.vfs = scmutil.vfs(self.path)
163 self.opener = self.vfs
163 self.opener = self.vfs
164 self.baseui = baseui
164 self.baseui = baseui
165 self.ui = baseui.copy()
165 self.ui = baseui.copy()
166 # A list of callback to shape the phase if no data were found.
166 # A list of callback to shape the phase if no data were found.
167 # Callback are in the form: func(repo, roots) --> processed root.
167 # Callback are in the form: func(repo, roots) --> processed root.
168 # This list it to be filled by extension during repo setup
168 # This list it to be filled by extension during repo setup
169 self._phasedefaults = []
169 self._phasedefaults = []
170 try:
170 try:
171 self.ui.readconfig(self.join("hgrc"), self.root)
171 self.ui.readconfig(self.join("hgrc"), self.root)
172 extensions.loadall(self.ui)
172 extensions.loadall(self.ui)
173 except IOError:
173 except IOError:
174 pass
174 pass
175
175
176 if not self.vfs.isdir():
176 if not self.vfs.isdir():
177 if create:
177 if create:
178 if not self.wvfs.exists():
178 if not self.wvfs.exists():
179 self.wvfs.makedirs()
179 self.wvfs.makedirs()
180 self.vfs.makedir(notindexed=True)
180 self.vfs.makedir(notindexed=True)
181 requirements = self._baserequirements(create)
181 requirements = self._baserequirements(create)
182 if self.ui.configbool('format', 'usestore', True):
182 if self.ui.configbool('format', 'usestore', True):
183 self.vfs.mkdir("store")
183 self.vfs.mkdir("store")
184 requirements.append("store")
184 requirements.append("store")
185 if self.ui.configbool('format', 'usefncache', True):
185 if self.ui.configbool('format', 'usefncache', True):
186 requirements.append("fncache")
186 requirements.append("fncache")
187 if self.ui.configbool('format', 'dotencode', True):
187 if self.ui.configbool('format', 'dotencode', True):
188 requirements.append('dotencode')
188 requirements.append('dotencode')
189 # create an invalid changelog
189 # create an invalid changelog
190 self.vfs.append(
190 self.vfs.append(
191 "00changelog.i",
191 "00changelog.i",
192 '\0\0\0\2' # represents revlogv2
192 '\0\0\0\2' # represents revlogv2
193 ' dummy changelog to prevent using the old repo layout'
193 ' dummy changelog to prevent using the old repo layout'
194 )
194 )
195 if self.ui.configbool('format', 'generaldelta', False):
195 if self.ui.configbool('format', 'generaldelta', False):
196 requirements.append("generaldelta")
196 requirements.append("generaldelta")
197 requirements = set(requirements)
197 requirements = set(requirements)
198 else:
198 else:
199 raise error.RepoError(_("repository %s not found") % path)
199 raise error.RepoError(_("repository %s not found") % path)
200 elif create:
200 elif create:
201 raise error.RepoError(_("repository %s already exists") % path)
201 raise error.RepoError(_("repository %s already exists") % path)
202 else:
202 else:
203 try:
203 try:
204 requirements = scmutil.readrequires(self.vfs, self.supported)
204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 except IOError, inst:
205 except IOError, inst:
206 if inst.errno != errno.ENOENT:
206 if inst.errno != errno.ENOENT:
207 raise
207 raise
208 requirements = set()
208 requirements = set()
209
209
210 self.sharedpath = self.path
210 self.sharedpath = self.path
211 try:
211 try:
212 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
212 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
213 realpath=True)
213 realpath=True)
214 s = vfs.base
214 s = vfs.base
215 if not vfs.exists():
215 if not vfs.exists():
216 raise error.RepoError(
216 raise error.RepoError(
217 _('.hg/sharedpath points to nonexistent directory %s') % s)
217 _('.hg/sharedpath points to nonexistent directory %s') % s)
218 self.sharedpath = s
218 self.sharedpath = s
219 except IOError, inst:
219 except IOError, inst:
220 if inst.errno != errno.ENOENT:
220 if inst.errno != errno.ENOENT:
221 raise
221 raise
222
222
223 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
223 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
224 self.spath = self.store.path
224 self.spath = self.store.path
225 self.svfs = self.store.vfs
225 self.svfs = self.store.vfs
226 self.sopener = self.svfs
226 self.sopener = self.svfs
227 self.sjoin = self.store.join
227 self.sjoin = self.store.join
228 self.vfs.createmode = self.store.createmode
228 self.vfs.createmode = self.store.createmode
229 self._applyrequirements(requirements)
229 self._applyrequirements(requirements)
230 if create:
230 if create:
231 self._writerequirements()
231 self._writerequirements()
232
232
233
233
234 self._branchcaches = {}
234 self._branchcaches = {}
235 self.filterpats = {}
235 self.filterpats = {}
236 self._datafilters = {}
236 self._datafilters = {}
237 self._transref = self._lockref = self._wlockref = None
237 self._transref = self._lockref = self._wlockref = None
238
238
239 # A cache for various files under .hg/ that tracks file changes,
239 # A cache for various files under .hg/ that tracks file changes,
240 # (used by the filecache decorator)
240 # (used by the filecache decorator)
241 #
241 #
242 # Maps a property name to its util.filecacheentry
242 # Maps a property name to its util.filecacheentry
243 self._filecache = {}
243 self._filecache = {}
244
244
245 # hold sets of revision to be filtered
245 # hold sets of revision to be filtered
246 # should be cleared when something might have changed the filter value:
246 # should be cleared when something might have changed the filter value:
247 # - new changesets,
247 # - new changesets,
248 # - phase change,
248 # - phase change,
249 # - new obsolescence marker,
249 # - new obsolescence marker,
250 # - working directory parent change,
250 # - working directory parent change,
251 # - bookmark changes
251 # - bookmark changes
252 self.filteredrevcache = {}
252 self.filteredrevcache = {}
253
253
254 def close(self):
254 def close(self):
255 pass
255 pass
256
256
257 def _restrictcapabilities(self, caps):
257 def _restrictcapabilities(self, caps):
258 return caps
258 return caps
259
259
260 def _applyrequirements(self, requirements):
260 def _applyrequirements(self, requirements):
261 self.requirements = requirements
261 self.requirements = requirements
262 self.sopener.options = dict((r, 1) for r in requirements
262 self.sopener.options = dict((r, 1) for r in requirements
263 if r in self.openerreqs)
263 if r in self.openerreqs)
264
264
265 def _writerequirements(self):
265 def _writerequirements(self):
266 reqfile = self.opener("requires", "w")
266 reqfile = self.opener("requires", "w")
267 for r in sorted(self.requirements):
267 for r in sorted(self.requirements):
268 reqfile.write("%s\n" % r)
268 reqfile.write("%s\n" % r)
269 reqfile.close()
269 reqfile.close()
270
270
271 def _checknested(self, path):
271 def _checknested(self, path):
272 """Determine if path is a legal nested repository."""
272 """Determine if path is a legal nested repository."""
273 if not path.startswith(self.root):
273 if not path.startswith(self.root):
274 return False
274 return False
275 subpath = path[len(self.root) + 1:]
275 subpath = path[len(self.root) + 1:]
276 normsubpath = util.pconvert(subpath)
276 normsubpath = util.pconvert(subpath)
277
277
278 # XXX: Checking against the current working copy is wrong in
278 # XXX: Checking against the current working copy is wrong in
279 # the sense that it can reject things like
279 # the sense that it can reject things like
280 #
280 #
281 # $ hg cat -r 10 sub/x.txt
281 # $ hg cat -r 10 sub/x.txt
282 #
282 #
283 # if sub/ is no longer a subrepository in the working copy
283 # if sub/ is no longer a subrepository in the working copy
284 # parent revision.
284 # parent revision.
285 #
285 #
286 # However, it can of course also allow things that would have
286 # However, it can of course also allow things that would have
287 # been rejected before, such as the above cat command if sub/
287 # been rejected before, such as the above cat command if sub/
288 # is a subrepository now, but was a normal directory before.
288 # is a subrepository now, but was a normal directory before.
289 # The old path auditor would have rejected by mistake since it
289 # The old path auditor would have rejected by mistake since it
290 # panics when it sees sub/.hg/.
290 # panics when it sees sub/.hg/.
291 #
291 #
292 # All in all, checking against the working copy seems sensible
292 # All in all, checking against the working copy seems sensible
293 # since we want to prevent access to nested repositories on
293 # since we want to prevent access to nested repositories on
294 # the filesystem *now*.
294 # the filesystem *now*.
295 ctx = self[None]
295 ctx = self[None]
296 parts = util.splitpath(subpath)
296 parts = util.splitpath(subpath)
297 while parts:
297 while parts:
298 prefix = '/'.join(parts)
298 prefix = '/'.join(parts)
299 if prefix in ctx.substate:
299 if prefix in ctx.substate:
300 if prefix == normsubpath:
300 if prefix == normsubpath:
301 return True
301 return True
302 else:
302 else:
303 sub = ctx.sub(prefix)
303 sub = ctx.sub(prefix)
304 return sub.checknested(subpath[len(prefix) + 1:])
304 return sub.checknested(subpath[len(prefix) + 1:])
305 else:
305 else:
306 parts.pop()
306 parts.pop()
307 return False
307 return False
308
308
309 def peer(self):
309 def peer(self):
310 return localpeer(self) # not cached to avoid reference cycle
310 return localpeer(self) # not cached to avoid reference cycle
311
311
312 def unfiltered(self):
312 def unfiltered(self):
313 """Return unfiltered version of the repository
313 """Return unfiltered version of the repository
314
314
315 Intended to be overwritten by filtered repo."""
315 Intended to be overwritten by filtered repo."""
316 return self
316 return self
317
317
318 def filtered(self, name):
318 def filtered(self, name):
319 """Return a filtered version of a repository"""
319 """Return a filtered version of a repository"""
320 # build a new class with the mixin and the current class
320 # build a new class with the mixin and the current class
321 # (possibly subclass of the repo)
321 # (possibly subclass of the repo)
322 class proxycls(repoview.repoview, self.unfiltered().__class__):
322 class proxycls(repoview.repoview, self.unfiltered().__class__):
323 pass
323 pass
324 return proxycls(self, name)
324 return proxycls(self, name)
325
325
326 @repofilecache('bookmarks')
326 @repofilecache('bookmarks')
327 def _bookmarks(self):
327 def _bookmarks(self):
328 return bookmarks.bmstore(self)
328 return bookmarks.bmstore(self)
329
329
330 @repofilecache('bookmarks.current')
330 @repofilecache('bookmarks.current')
331 def _bookmarkcurrent(self):
331 def _bookmarkcurrent(self):
332 return bookmarks.readcurrent(self)
332 return bookmarks.readcurrent(self)
333
333
334 def bookmarkheads(self, bookmark):
334 def bookmarkheads(self, bookmark):
335 name = bookmark.split('@', 1)[0]
335 name = bookmark.split('@', 1)[0]
336 heads = []
336 heads = []
337 for mark, n in self._bookmarks.iteritems():
337 for mark, n in self._bookmarks.iteritems():
338 if mark.split('@', 1)[0] == name:
338 if mark.split('@', 1)[0] == name:
339 heads.append(n)
339 heads.append(n)
340 return heads
340 return heads
341
341
342 @storecache('phaseroots')
342 @storecache('phaseroots')
343 def _phasecache(self):
343 def _phasecache(self):
344 return phases.phasecache(self, self._phasedefaults)
344 return phases.phasecache(self, self._phasedefaults)
345
345
346 @storecache('obsstore')
346 @storecache('obsstore')
347 def obsstore(self):
347 def obsstore(self):
348 store = obsolete.obsstore(self.sopener)
348 store = obsolete.obsstore(self.sopener)
349 if store and not obsolete._enabled:
349 if store and not obsolete._enabled:
350 # message is rare enough to not be translated
350 # message is rare enough to not be translated
351 msg = 'obsolete feature not enabled but %i markers found!\n'
351 msg = 'obsolete feature not enabled but %i markers found!\n'
352 self.ui.warn(msg % len(list(store)))
352 self.ui.warn(msg % len(list(store)))
353 return store
353 return store
354
354
355 @storecache('00changelog.i')
355 @storecache('00changelog.i')
356 def changelog(self):
356 def changelog(self):
357 c = changelog.changelog(self.sopener)
357 c = changelog.changelog(self.sopener)
358 if 'HG_PENDING' in os.environ:
358 if 'HG_PENDING' in os.environ:
359 p = os.environ['HG_PENDING']
359 p = os.environ['HG_PENDING']
360 if p.startswith(self.root):
360 if p.startswith(self.root):
361 c.readpending('00changelog.i.a')
361 c.readpending('00changelog.i.a')
362 return c
362 return c
363
363
364 @storecache('00manifest.i')
364 @storecache('00manifest.i')
365 def manifest(self):
365 def manifest(self):
366 return manifest.manifest(self.sopener)
366 return manifest.manifest(self.sopener)
367
367
368 @repofilecache('dirstate')
368 @repofilecache('dirstate')
369 def dirstate(self):
369 def dirstate(self):
370 warned = [0]
370 warned = [0]
371 def validate(node):
371 def validate(node):
372 try:
372 try:
373 self.changelog.rev(node)
373 self.changelog.rev(node)
374 return node
374 return node
375 except error.LookupError:
375 except error.LookupError:
376 if not warned[0]:
376 if not warned[0]:
377 warned[0] = True
377 warned[0] = True
378 self.ui.warn(_("warning: ignoring unknown"
378 self.ui.warn(_("warning: ignoring unknown"
379 " working parent %s!\n") % short(node))
379 " working parent %s!\n") % short(node))
380 return nullid
380 return nullid
381
381
382 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
382 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
383
383
384 def __getitem__(self, changeid):
384 def __getitem__(self, changeid):
385 if changeid is None:
385 if changeid is None:
386 return context.workingctx(self)
386 return context.workingctx(self)
387 return context.changectx(self, changeid)
387 return context.changectx(self, changeid)
388
388
389 def __contains__(self, changeid):
389 def __contains__(self, changeid):
390 try:
390 try:
391 return bool(self.lookup(changeid))
391 return bool(self.lookup(changeid))
392 except error.RepoLookupError:
392 except error.RepoLookupError:
393 return False
393 return False
394
394
395 def __nonzero__(self):
395 def __nonzero__(self):
396 return True
396 return True
397
397
398 def __len__(self):
398 def __len__(self):
399 return len(self.changelog)
399 return len(self.changelog)
400
400
401 def __iter__(self):
401 def __iter__(self):
402 return iter(self.changelog)
402 return iter(self.changelog)
403
403
404 def revs(self, expr, *args):
404 def revs(self, expr, *args):
405 '''Return a list of revisions matching the given revset'''
405 '''Return a list of revisions matching the given revset'''
406 expr = revset.formatspec(expr, *args)
406 expr = revset.formatspec(expr, *args)
407 m = revset.match(None, expr)
407 m = revset.match(None, expr)
408 return [r for r in m(self, list(self))]
408 return [r for r in m(self, list(self))]
409
409
410 def set(self, expr, *args):
410 def set(self, expr, *args):
411 '''
411 '''
412 Yield a context for each matching revision, after doing arg
412 Yield a context for each matching revision, after doing arg
413 replacement via revset.formatspec
413 replacement via revset.formatspec
414 '''
414 '''
415 for r in self.revs(expr, *args):
415 for r in self.revs(expr, *args):
416 yield self[r]
416 yield self[r]
417
417
418 def url(self):
418 def url(self):
419 return 'file:' + self.root
419 return 'file:' + self.root
420
420
421 def hook(self, name, throw=False, **args):
421 def hook(self, name, throw=False, **args):
422 return hook.hook(self.ui, self, name, throw, **args)
422 return hook.hook(self.ui, self, name, throw, **args)
423
423
424 @unfilteredmethod
424 @unfilteredmethod
425 def _tag(self, names, node, message, local, user, date, extra={}):
425 def _tag(self, names, node, message, local, user, date, extra={}):
426 if isinstance(names, str):
426 if isinstance(names, str):
427 names = (names,)
427 names = (names,)
428
428
429 branches = self.branchmap()
429 branches = self.branchmap()
430 for name in names:
430 for name in names:
431 self.hook('pretag', throw=True, node=hex(node), tag=name,
431 self.hook('pretag', throw=True, node=hex(node), tag=name,
432 local=local)
432 local=local)
433 if name in branches:
433 if name in branches:
434 self.ui.warn(_("warning: tag %s conflicts with existing"
434 self.ui.warn(_("warning: tag %s conflicts with existing"
435 " branch name\n") % name)
435 " branch name\n") % name)
436
436
437 def writetags(fp, names, munge, prevtags):
437 def writetags(fp, names, munge, prevtags):
438 fp.seek(0, 2)
438 fp.seek(0, 2)
439 if prevtags and prevtags[-1] != '\n':
439 if prevtags and prevtags[-1] != '\n':
440 fp.write('\n')
440 fp.write('\n')
441 for name in names:
441 for name in names:
442 m = munge and munge(name) or name
442 m = munge and munge(name) or name
443 if (self._tagscache.tagtypes and
443 if (self._tagscache.tagtypes and
444 name in self._tagscache.tagtypes):
444 name in self._tagscache.tagtypes):
445 old = self.tags().get(name, nullid)
445 old = self.tags().get(name, nullid)
446 fp.write('%s %s\n' % (hex(old), m))
446 fp.write('%s %s\n' % (hex(old), m))
447 fp.write('%s %s\n' % (hex(node), m))
447 fp.write('%s %s\n' % (hex(node), m))
448 fp.close()
448 fp.close()
449
449
450 prevtags = ''
450 prevtags = ''
451 if local:
451 if local:
452 try:
452 try:
453 fp = self.opener('localtags', 'r+')
453 fp = self.opener('localtags', 'r+')
454 except IOError:
454 except IOError:
455 fp = self.opener('localtags', 'a')
455 fp = self.opener('localtags', 'a')
456 else:
456 else:
457 prevtags = fp.read()
457 prevtags = fp.read()
458
458
459 # local tags are stored in the current charset
459 # local tags are stored in the current charset
460 writetags(fp, names, None, prevtags)
460 writetags(fp, names, None, prevtags)
461 for name in names:
461 for name in names:
462 self.hook('tag', node=hex(node), tag=name, local=local)
462 self.hook('tag', node=hex(node), tag=name, local=local)
463 return
463 return
464
464
465 try:
465 try:
466 fp = self.wfile('.hgtags', 'rb+')
466 fp = self.wfile('.hgtags', 'rb+')
467 except IOError, e:
467 except IOError, e:
468 if e.errno != errno.ENOENT:
468 if e.errno != errno.ENOENT:
469 raise
469 raise
470 fp = self.wfile('.hgtags', 'ab')
470 fp = self.wfile('.hgtags', 'ab')
471 else:
471 else:
472 prevtags = fp.read()
472 prevtags = fp.read()
473
473
474 # committed tags are stored in UTF-8
474 # committed tags are stored in UTF-8
475 writetags(fp, names, encoding.fromlocal, prevtags)
475 writetags(fp, names, encoding.fromlocal, prevtags)
476
476
477 fp.close()
477 fp.close()
478
478
479 self.invalidatecaches()
479 self.invalidatecaches()
480
480
481 if '.hgtags' not in self.dirstate:
481 if '.hgtags' not in self.dirstate:
482 self[None].add(['.hgtags'])
482 self[None].add(['.hgtags'])
483
483
484 m = matchmod.exact(self.root, '', ['.hgtags'])
484 m = matchmod.exact(self.root, '', ['.hgtags'])
485 tagnode = self.commit(message, user, date, extra=extra, match=m)
485 tagnode = self.commit(message, user, date, extra=extra, match=m)
486
486
487 for name in names:
487 for name in names:
488 self.hook('tag', node=hex(node), tag=name, local=local)
488 self.hook('tag', node=hex(node), tag=name, local=local)
489
489
490 return tagnode
490 return tagnode
491
491
492 def tag(self, names, node, message, local, user, date):
492 def tag(self, names, node, message, local, user, date):
493 '''tag a revision with one or more symbolic names.
493 '''tag a revision with one or more symbolic names.
494
494
495 names is a list of strings or, when adding a single tag, names may be a
495 names is a list of strings or, when adding a single tag, names may be a
496 string.
496 string.
497
497
498 if local is True, the tags are stored in a per-repository file.
498 if local is True, the tags are stored in a per-repository file.
499 otherwise, they are stored in the .hgtags file, and a new
499 otherwise, they are stored in the .hgtags file, and a new
500 changeset is committed with the change.
500 changeset is committed with the change.
501
501
502 keyword arguments:
502 keyword arguments:
503
503
504 local: whether to store tags in non-version-controlled file
504 local: whether to store tags in non-version-controlled file
505 (default False)
505 (default False)
506
506
507 message: commit message to use if committing
507 message: commit message to use if committing
508
508
509 user: name of user to use if committing
509 user: name of user to use if committing
510
510
511 date: date tuple to use if committing'''
511 date: date tuple to use if committing'''
512
512
513 if not local:
513 if not local:
514 for x in self.status()[:5]:
514 for x in self.status()[:5]:
515 if '.hgtags' in x:
515 if '.hgtags' in x:
516 raise util.Abort(_('working copy of .hgtags is changed '
516 raise util.Abort(_('working copy of .hgtags is changed '
517 '(please commit .hgtags manually)'))
517 '(please commit .hgtags manually)'))
518
518
519 self.tags() # instantiate the cache
519 self.tags() # instantiate the cache
520 self._tag(names, node, message, local, user, date)
520 self._tag(names, node, message, local, user, date)
521
521
522 @filteredpropertycache
522 @filteredpropertycache
523 def _tagscache(self):
523 def _tagscache(self):
524 '''Returns a tagscache object that contains various tags related
524 '''Returns a tagscache object that contains various tags related
525 caches.'''
525 caches.'''
526
526
527 # This simplifies its cache management by having one decorated
527 # This simplifies its cache management by having one decorated
528 # function (this one) and the rest simply fetch things from it.
528 # function (this one) and the rest simply fetch things from it.
529 class tagscache(object):
529 class tagscache(object):
530 def __init__(self):
530 def __init__(self):
531 # These two define the set of tags for this repository. tags
531 # These two define the set of tags for this repository. tags
532 # maps tag name to node; tagtypes maps tag name to 'global' or
532 # maps tag name to node; tagtypes maps tag name to 'global' or
533 # 'local'. (Global tags are defined by .hgtags across all
533 # 'local'. (Global tags are defined by .hgtags across all
534 # heads, and local tags are defined in .hg/localtags.)
534 # heads, and local tags are defined in .hg/localtags.)
535 # They constitute the in-memory cache of tags.
535 # They constitute the in-memory cache of tags.
536 self.tags = self.tagtypes = None
536 self.tags = self.tagtypes = None
537
537
538 self.nodetagscache = self.tagslist = None
538 self.nodetagscache = self.tagslist = None
539
539
540 cache = tagscache()
540 cache = tagscache()
541 cache.tags, cache.tagtypes = self._findtags()
541 cache.tags, cache.tagtypes = self._findtags()
542
542
543 return cache
543 return cache
544
544
545 def tags(self):
545 def tags(self):
546 '''return a mapping of tag to node'''
546 '''return a mapping of tag to node'''
547 t = {}
547 t = {}
548 if self.changelog.filteredrevs:
548 if self.changelog.filteredrevs:
549 tags, tt = self._findtags()
549 tags, tt = self._findtags()
550 else:
550 else:
551 tags = self._tagscache.tags
551 tags = self._tagscache.tags
552 for k, v in tags.iteritems():
552 for k, v in tags.iteritems():
553 try:
553 try:
554 # ignore tags to unknown nodes
554 # ignore tags to unknown nodes
555 self.changelog.rev(v)
555 self.changelog.rev(v)
556 t[k] = v
556 t[k] = v
557 except (error.LookupError, ValueError):
557 except (error.LookupError, ValueError):
558 pass
558 pass
559 return t
559 return t
560
560
561 def _findtags(self):
561 def _findtags(self):
562 '''Do the hard work of finding tags. Return a pair of dicts
562 '''Do the hard work of finding tags. Return a pair of dicts
563 (tags, tagtypes) where tags maps tag name to node, and tagtypes
563 (tags, tagtypes) where tags maps tag name to node, and tagtypes
564 maps tag name to a string like \'global\' or \'local\'.
564 maps tag name to a string like \'global\' or \'local\'.
565 Subclasses or extensions are free to add their own tags, but
565 Subclasses or extensions are free to add their own tags, but
566 should be aware that the returned dicts will be retained for the
566 should be aware that the returned dicts will be retained for the
567 duration of the localrepo object.'''
567 duration of the localrepo object.'''
568
568
569 # XXX what tagtype should subclasses/extensions use? Currently
569 # XXX what tagtype should subclasses/extensions use? Currently
570 # mq and bookmarks add tags, but do not set the tagtype at all.
570 # mq and bookmarks add tags, but do not set the tagtype at all.
571 # Should each extension invent its own tag type? Should there
571 # Should each extension invent its own tag type? Should there
572 # be one tagtype for all such "virtual" tags? Or is the status
572 # be one tagtype for all such "virtual" tags? Or is the status
573 # quo fine?
573 # quo fine?
574
574
575 alltags = {} # map tag name to (node, hist)
575 alltags = {} # map tag name to (node, hist)
576 tagtypes = {}
576 tagtypes = {}
577
577
578 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
578 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
579 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
579 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
580
580
581 # Build the return dicts. Have to re-encode tag names because
581 # Build the return dicts. Have to re-encode tag names because
582 # the tags module always uses UTF-8 (in order not to lose info
582 # the tags module always uses UTF-8 (in order not to lose info
583 # writing to the cache), but the rest of Mercurial wants them in
583 # writing to the cache), but the rest of Mercurial wants them in
584 # local encoding.
584 # local encoding.
585 tags = {}
585 tags = {}
586 for (name, (node, hist)) in alltags.iteritems():
586 for (name, (node, hist)) in alltags.iteritems():
587 if node != nullid:
587 if node != nullid:
588 tags[encoding.tolocal(name)] = node
588 tags[encoding.tolocal(name)] = node
589 tags['tip'] = self.changelog.tip()
589 tags['tip'] = self.changelog.tip()
590 tagtypes = dict([(encoding.tolocal(name), value)
590 tagtypes = dict([(encoding.tolocal(name), value)
591 for (name, value) in tagtypes.iteritems()])
591 for (name, value) in tagtypes.iteritems()])
592 return (tags, tagtypes)
592 return (tags, tagtypes)
593
593
594 def tagtype(self, tagname):
594 def tagtype(self, tagname):
595 '''
595 '''
596 return the type of the given tag. result can be:
596 return the type of the given tag. result can be:
597
597
598 'local' : a local tag
598 'local' : a local tag
599 'global' : a global tag
599 'global' : a global tag
600 None : tag does not exist
600 None : tag does not exist
601 '''
601 '''
602
602
603 return self._tagscache.tagtypes.get(tagname)
603 return self._tagscache.tagtypes.get(tagname)
604
604
605 def tagslist(self):
605 def tagslist(self):
606 '''return a list of tags ordered by revision'''
606 '''return a list of tags ordered by revision'''
607 if not self._tagscache.tagslist:
607 if not self._tagscache.tagslist:
608 l = []
608 l = []
609 for t, n in self.tags().iteritems():
609 for t, n in self.tags().iteritems():
610 r = self.changelog.rev(n)
610 r = self.changelog.rev(n)
611 l.append((r, t, n))
611 l.append((r, t, n))
612 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
612 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
613
613
614 return self._tagscache.tagslist
614 return self._tagscache.tagslist
615
615
616 def nodetags(self, node):
616 def nodetags(self, node):
617 '''return the tags associated with a node'''
617 '''return the tags associated with a node'''
618 if not self._tagscache.nodetagscache:
618 if not self._tagscache.nodetagscache:
619 nodetagscache = {}
619 nodetagscache = {}
620 for t, n in self._tagscache.tags.iteritems():
620 for t, n in self._tagscache.tags.iteritems():
621 nodetagscache.setdefault(n, []).append(t)
621 nodetagscache.setdefault(n, []).append(t)
622 for tags in nodetagscache.itervalues():
622 for tags in nodetagscache.itervalues():
623 tags.sort()
623 tags.sort()
624 self._tagscache.nodetagscache = nodetagscache
624 self._tagscache.nodetagscache = nodetagscache
625 return self._tagscache.nodetagscache.get(node, [])
625 return self._tagscache.nodetagscache.get(node, [])
626
626
627 def nodebookmarks(self, node):
627 def nodebookmarks(self, node):
628 marks = []
628 marks = []
629 for bookmark, n in self._bookmarks.iteritems():
629 for bookmark, n in self._bookmarks.iteritems():
630 if n == node:
630 if n == node:
631 marks.append(bookmark)
631 marks.append(bookmark)
632 return sorted(marks)
632 return sorted(marks)
633
633
634 def branchmap(self):
634 def branchmap(self):
635 '''returns a dictionary {branch: [branchheads]}'''
635 '''returns a dictionary {branch: [branchheads]}'''
636 branchmap.updatecache(self)
636 branchmap.updatecache(self)
637 return self._branchcaches[self.filtername]
637 return self._branchcaches[self.filtername]
638
638
639
639
640 def _branchtip(self, heads):
640 def _branchtip(self, heads):
641 '''return the tipmost branch head in heads'''
641 '''return the tipmost branch head in heads'''
642 tip = heads[-1]
642 tip = heads[-1]
643 for h in reversed(heads):
643 for h in reversed(heads):
644 if not self[h].closesbranch():
644 if not self[h].closesbranch():
645 tip = h
645 tip = h
646 break
646 break
647 return tip
647 return tip
648
648
649 def branchtip(self, branch):
649 def branchtip(self, branch):
650 '''return the tip node for a given branch'''
650 '''return the tip node for a given branch'''
651 if branch not in self.branchmap():
651 if branch not in self.branchmap():
652 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
652 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
653 return self._branchtip(self.branchmap()[branch])
653 return self._branchtip(self.branchmap()[branch])
654
654
655 def branchtags(self):
655 def branchtags(self):
656 '''return a dict where branch names map to the tipmost head of
656 '''return a dict where branch names map to the tipmost head of
657 the branch, open heads come before closed'''
657 the branch, open heads come before closed'''
658 bt = {}
658 bt = {}
659 for bn, heads in self.branchmap().iteritems():
659 for bn, heads in self.branchmap().iteritems():
660 bt[bn] = self._branchtip(heads)
660 bt[bn] = self._branchtip(heads)
661 return bt
661 return bt
662
662
663 def lookup(self, key):
663 def lookup(self, key):
664 return self[key].node()
664 return self[key].node()
665
665
666 def lookupbranch(self, key, remote=None):
666 def lookupbranch(self, key, remote=None):
667 repo = remote or self
667 repo = remote or self
668 if key in repo.branchmap():
668 if key in repo.branchmap():
669 return key
669 return key
670
670
671 repo = (remote and remote.local()) and remote or self
671 repo = (remote and remote.local()) and remote or self
672 return repo[key].branch()
672 return repo[key].branch()
673
673
674 def known(self, nodes):
674 def known(self, nodes):
675 nm = self.changelog.nodemap
675 nm = self.changelog.nodemap
676 pc = self._phasecache
676 pc = self._phasecache
677 result = []
677 result = []
678 for n in nodes:
678 for n in nodes:
679 r = nm.get(n)
679 r = nm.get(n)
680 resp = not (r is None or pc.phase(self, r) >= phases.secret)
680 resp = not (r is None or pc.phase(self, r) >= phases.secret)
681 result.append(resp)
681 result.append(resp)
682 return result
682 return result
683
683
684 def local(self):
684 def local(self):
685 return self
685 return self
686
686
687 def cancopy(self):
687 def cancopy(self):
688 return self.local() # so statichttprepo's override of local() works
688 return self.local() # so statichttprepo's override of local() works
689
689
690 def join(self, f):
690 def join(self, f):
691 return os.path.join(self.path, f)
691 return os.path.join(self.path, f)
692
692
693 def wjoin(self, f):
693 def wjoin(self, f):
694 return os.path.join(self.root, f)
694 return os.path.join(self.root, f)
695
695
696 def file(self, f):
696 def file(self, f):
697 if f[0] == '/':
697 if f[0] == '/':
698 f = f[1:]
698 f = f[1:]
699 return filelog.filelog(self.sopener, f)
699 return filelog.filelog(self.sopener, f)
700
700
701 def changectx(self, changeid):
701 def changectx(self, changeid):
702 return self[changeid]
702 return self[changeid]
703
703
704 def parents(self, changeid=None):
704 def parents(self, changeid=None):
705 '''get list of changectxs for parents of changeid'''
705 '''get list of changectxs for parents of changeid'''
706 return self[changeid].parents()
706 return self[changeid].parents()
707
707
708 def setparents(self, p1, p2=nullid):
708 def setparents(self, p1, p2=nullid):
709 copies = self.dirstate.setparents(p1, p2)
709 copies = self.dirstate.setparents(p1, p2)
710 pctx = self[p1]
710 pctx = self[p1]
711 if copies:
711 if copies:
712 # Adjust copy records, the dirstate cannot do it, it
712 # Adjust copy records, the dirstate cannot do it, it
713 # requires access to parents manifests. Preserve them
713 # requires access to parents manifests. Preserve them
714 # only for entries added to first parent.
714 # only for entries added to first parent.
715 for f in copies:
715 for f in copies:
716 if f not in pctx and copies[f] in pctx:
716 if f not in pctx and copies[f] in pctx:
717 self.dirstate.copy(copies[f], f)
717 self.dirstate.copy(copies[f], f)
718 if p2 == nullid:
718 if p2 == nullid:
719 for f, s in sorted(self.dirstate.copies().items()):
719 for f, s in sorted(self.dirstate.copies().items()):
720 if f not in pctx and s not in pctx:
720 if f not in pctx and s not in pctx:
721 self.dirstate.copy(None, f)
721 self.dirstate.copy(None, f)
722
722
723 def filectx(self, path, changeid=None, fileid=None):
723 def filectx(self, path, changeid=None, fileid=None):
724 """changeid can be a changeset revision, node, or tag.
724 """changeid can be a changeset revision, node, or tag.
725 fileid can be a file revision or node."""
725 fileid can be a file revision or node."""
726 return context.filectx(self, path, changeid, fileid)
726 return context.filectx(self, path, changeid, fileid)
727
727
728 def getcwd(self):
728 def getcwd(self):
729 return self.dirstate.getcwd()
729 return self.dirstate.getcwd()
730
730
731 def pathto(self, f, cwd=None):
731 def pathto(self, f, cwd=None):
732 return self.dirstate.pathto(f, cwd)
732 return self.dirstate.pathto(f, cwd)
733
733
734 def wfile(self, f, mode='r'):
734 def wfile(self, f, mode='r'):
735 return self.wopener(f, mode)
735 return self.wopener(f, mode)
736
736
737 def _link(self, f):
737 def _link(self, f):
738 return self.wvfs.islink(f)
738 return self.wvfs.islink(f)
739
739
740 def _loadfilter(self, filter):
740 def _loadfilter(self, filter):
741 if filter not in self.filterpats:
741 if filter not in self.filterpats:
742 l = []
742 l = []
743 for pat, cmd in self.ui.configitems(filter):
743 for pat, cmd in self.ui.configitems(filter):
744 if cmd == '!':
744 if cmd == '!':
745 continue
745 continue
746 mf = matchmod.match(self.root, '', [pat])
746 mf = matchmod.match(self.root, '', [pat])
747 fn = None
747 fn = None
748 params = cmd
748 params = cmd
749 for name, filterfn in self._datafilters.iteritems():
749 for name, filterfn in self._datafilters.iteritems():
750 if cmd.startswith(name):
750 if cmd.startswith(name):
751 fn = filterfn
751 fn = filterfn
752 params = cmd[len(name):].lstrip()
752 params = cmd[len(name):].lstrip()
753 break
753 break
754 if not fn:
754 if not fn:
755 fn = lambda s, c, **kwargs: util.filter(s, c)
755 fn = lambda s, c, **kwargs: util.filter(s, c)
756 # Wrap old filters not supporting keyword arguments
756 # Wrap old filters not supporting keyword arguments
757 if not inspect.getargspec(fn)[2]:
757 if not inspect.getargspec(fn)[2]:
758 oldfn = fn
758 oldfn = fn
759 fn = lambda s, c, **kwargs: oldfn(s, c)
759 fn = lambda s, c, **kwargs: oldfn(s, c)
760 l.append((mf, fn, params))
760 l.append((mf, fn, params))
761 self.filterpats[filter] = l
761 self.filterpats[filter] = l
762 return self.filterpats[filter]
762 return self.filterpats[filter]
763
763
764 def _filter(self, filterpats, filename, data):
764 def _filter(self, filterpats, filename, data):
765 for mf, fn, cmd in filterpats:
765 for mf, fn, cmd in filterpats:
766 if mf(filename):
766 if mf(filename):
767 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
767 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
768 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
768 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
769 break
769 break
770
770
771 return data
771 return data
772
772
773 @unfilteredpropertycache
773 @unfilteredpropertycache
774 def _encodefilterpats(self):
774 def _encodefilterpats(self):
775 return self._loadfilter('encode')
775 return self._loadfilter('encode')
776
776
777 @unfilteredpropertycache
777 @unfilteredpropertycache
778 def _decodefilterpats(self):
778 def _decodefilterpats(self):
779 return self._loadfilter('decode')
779 return self._loadfilter('decode')
780
780
781 def adddatafilter(self, name, filter):
781 def adddatafilter(self, name, filter):
782 self._datafilters[name] = filter
782 self._datafilters[name] = filter
783
783
784 def wread(self, filename):
784 def wread(self, filename):
785 if self._link(filename):
785 if self._link(filename):
786 data = os.readlink(self.wjoin(filename))
786 data = self.wvfs.readlink(filename)
787 else:
787 else:
788 data = self.wopener.read(filename)
788 data = self.wopener.read(filename)
789 return self._filter(self._encodefilterpats, filename, data)
789 return self._filter(self._encodefilterpats, filename, data)
790
790
791 def wwrite(self, filename, data, flags):
791 def wwrite(self, filename, data, flags):
792 data = self._filter(self._decodefilterpats, filename, data)
792 data = self._filter(self._decodefilterpats, filename, data)
793 if 'l' in flags:
793 if 'l' in flags:
794 self.wopener.symlink(data, filename)
794 self.wopener.symlink(data, filename)
795 else:
795 else:
796 self.wopener.write(filename, data)
796 self.wopener.write(filename, data)
797 if 'x' in flags:
797 if 'x' in flags:
798 util.setflags(self.wjoin(filename), False, True)
798 util.setflags(self.wjoin(filename), False, True)
799
799
800 def wwritedata(self, filename, data):
800 def wwritedata(self, filename, data):
801 return self._filter(self._decodefilterpats, filename, data)
801 return self._filter(self._decodefilterpats, filename, data)
802
802
803 def transaction(self, desc):
803 def transaction(self, desc):
804 tr = self._transref and self._transref() or None
804 tr = self._transref and self._transref() or None
805 if tr and tr.running():
805 if tr and tr.running():
806 return tr.nest()
806 return tr.nest()
807
807
808 # abort here if the journal already exists
808 # abort here if the journal already exists
809 if self.svfs.exists("journal"):
809 if self.svfs.exists("journal"):
810 raise error.RepoError(
810 raise error.RepoError(
811 _("abandoned transaction found - run hg recover"))
811 _("abandoned transaction found - run hg recover"))
812
812
813 self._writejournal(desc)
813 self._writejournal(desc)
814 renames = [(x, undoname(x)) for x in self._journalfiles()]
814 renames = [(x, undoname(x)) for x in self._journalfiles()]
815
815
816 tr = transaction.transaction(self.ui.warn, self.sopener,
816 tr = transaction.transaction(self.ui.warn, self.sopener,
817 self.sjoin("journal"),
817 self.sjoin("journal"),
818 aftertrans(renames),
818 aftertrans(renames),
819 self.store.createmode)
819 self.store.createmode)
820 self._transref = weakref.ref(tr)
820 self._transref = weakref.ref(tr)
821 return tr
821 return tr
822
822
823 def _journalfiles(self):
823 def _journalfiles(self):
824 return (self.sjoin('journal'), self.join('journal.dirstate'),
824 return (self.sjoin('journal'), self.join('journal.dirstate'),
825 self.join('journal.branch'), self.join('journal.desc'),
825 self.join('journal.branch'), self.join('journal.desc'),
826 self.join('journal.bookmarks'),
826 self.join('journal.bookmarks'),
827 self.sjoin('journal.phaseroots'))
827 self.sjoin('journal.phaseroots'))
828
828
829 def undofiles(self):
829 def undofiles(self):
830 return [undoname(x) for x in self._journalfiles()]
830 return [undoname(x) for x in self._journalfiles()]
831
831
832 def _writejournal(self, desc):
832 def _writejournal(self, desc):
833 self.opener.write("journal.dirstate",
833 self.opener.write("journal.dirstate",
834 self.opener.tryread("dirstate"))
834 self.opener.tryread("dirstate"))
835 self.opener.write("journal.branch",
835 self.opener.write("journal.branch",
836 encoding.fromlocal(self.dirstate.branch()))
836 encoding.fromlocal(self.dirstate.branch()))
837 self.opener.write("journal.desc",
837 self.opener.write("journal.desc",
838 "%d\n%s\n" % (len(self), desc))
838 "%d\n%s\n" % (len(self), desc))
839 self.opener.write("journal.bookmarks",
839 self.opener.write("journal.bookmarks",
840 self.opener.tryread("bookmarks"))
840 self.opener.tryread("bookmarks"))
841 self.sopener.write("journal.phaseroots",
841 self.sopener.write("journal.phaseroots",
842 self.sopener.tryread("phaseroots"))
842 self.sopener.tryread("phaseroots"))
843
843
844 def recover(self):
844 def recover(self):
845 lock = self.lock()
845 lock = self.lock()
846 try:
846 try:
847 if self.svfs.exists("journal"):
847 if self.svfs.exists("journal"):
848 self.ui.status(_("rolling back interrupted transaction\n"))
848 self.ui.status(_("rolling back interrupted transaction\n"))
849 transaction.rollback(self.sopener, self.sjoin("journal"),
849 transaction.rollback(self.sopener, self.sjoin("journal"),
850 self.ui.warn)
850 self.ui.warn)
851 self.invalidate()
851 self.invalidate()
852 return True
852 return True
853 else:
853 else:
854 self.ui.warn(_("no interrupted transaction available\n"))
854 self.ui.warn(_("no interrupted transaction available\n"))
855 return False
855 return False
856 finally:
856 finally:
857 lock.release()
857 lock.release()
858
858
859 def rollback(self, dryrun=False, force=False):
859 def rollback(self, dryrun=False, force=False):
860 wlock = lock = None
860 wlock = lock = None
861 try:
861 try:
862 wlock = self.wlock()
862 wlock = self.wlock()
863 lock = self.lock()
863 lock = self.lock()
864 if self.svfs.exists("undo"):
864 if self.svfs.exists("undo"):
865 return self._rollback(dryrun, force)
865 return self._rollback(dryrun, force)
866 else:
866 else:
867 self.ui.warn(_("no rollback information available\n"))
867 self.ui.warn(_("no rollback information available\n"))
868 return 1
868 return 1
869 finally:
869 finally:
870 release(lock, wlock)
870 release(lock, wlock)
871
871
872 @unfilteredmethod # Until we get smarter cache management
872 @unfilteredmethod # Until we get smarter cache management
873 def _rollback(self, dryrun, force):
873 def _rollback(self, dryrun, force):
874 ui = self.ui
874 ui = self.ui
875 try:
875 try:
876 args = self.opener.read('undo.desc').splitlines()
876 args = self.opener.read('undo.desc').splitlines()
877 (oldlen, desc, detail) = (int(args[0]), args[1], None)
877 (oldlen, desc, detail) = (int(args[0]), args[1], None)
878 if len(args) >= 3:
878 if len(args) >= 3:
879 detail = args[2]
879 detail = args[2]
880 oldtip = oldlen - 1
880 oldtip = oldlen - 1
881
881
882 if detail and ui.verbose:
882 if detail and ui.verbose:
883 msg = (_('repository tip rolled back to revision %s'
883 msg = (_('repository tip rolled back to revision %s'
884 ' (undo %s: %s)\n')
884 ' (undo %s: %s)\n')
885 % (oldtip, desc, detail))
885 % (oldtip, desc, detail))
886 else:
886 else:
887 msg = (_('repository tip rolled back to revision %s'
887 msg = (_('repository tip rolled back to revision %s'
888 ' (undo %s)\n')
888 ' (undo %s)\n')
889 % (oldtip, desc))
889 % (oldtip, desc))
890 except IOError:
890 except IOError:
891 msg = _('rolling back unknown transaction\n')
891 msg = _('rolling back unknown transaction\n')
892 desc = None
892 desc = None
893
893
894 if not force and self['.'] != self['tip'] and desc == 'commit':
894 if not force and self['.'] != self['tip'] and desc == 'commit':
895 raise util.Abort(
895 raise util.Abort(
896 _('rollback of last commit while not checked out '
896 _('rollback of last commit while not checked out '
897 'may lose data'), hint=_('use -f to force'))
897 'may lose data'), hint=_('use -f to force'))
898
898
899 ui.status(msg)
899 ui.status(msg)
900 if dryrun:
900 if dryrun:
901 return 0
901 return 0
902
902
903 parents = self.dirstate.parents()
903 parents = self.dirstate.parents()
904 self.destroying()
904 self.destroying()
905 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
905 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
906 if self.vfs.exists('undo.bookmarks'):
906 if self.vfs.exists('undo.bookmarks'):
907 self.vfs.rename('undo.bookmarks', 'bookmarks')
907 self.vfs.rename('undo.bookmarks', 'bookmarks')
908 if self.svfs.exists('undo.phaseroots'):
908 if self.svfs.exists('undo.phaseroots'):
909 self.svfs.rename('undo.phaseroots', 'phaseroots')
909 self.svfs.rename('undo.phaseroots', 'phaseroots')
910 self.invalidate()
910 self.invalidate()
911
911
912 parentgone = (parents[0] not in self.changelog.nodemap or
912 parentgone = (parents[0] not in self.changelog.nodemap or
913 parents[1] not in self.changelog.nodemap)
913 parents[1] not in self.changelog.nodemap)
914 if parentgone:
914 if parentgone:
915 self.vfs.rename('undo.dirstate', 'dirstate')
915 self.vfs.rename('undo.dirstate', 'dirstate')
916 try:
916 try:
917 branch = self.opener.read('undo.branch')
917 branch = self.opener.read('undo.branch')
918 self.dirstate.setbranch(encoding.tolocal(branch))
918 self.dirstate.setbranch(encoding.tolocal(branch))
919 except IOError:
919 except IOError:
920 ui.warn(_('named branch could not be reset: '
920 ui.warn(_('named branch could not be reset: '
921 'current branch is still \'%s\'\n')
921 'current branch is still \'%s\'\n')
922 % self.dirstate.branch())
922 % self.dirstate.branch())
923
923
924 self.dirstate.invalidate()
924 self.dirstate.invalidate()
925 parents = tuple([p.rev() for p in self.parents()])
925 parents = tuple([p.rev() for p in self.parents()])
926 if len(parents) > 1:
926 if len(parents) > 1:
927 ui.status(_('working directory now based on '
927 ui.status(_('working directory now based on '
928 'revisions %d and %d\n') % parents)
928 'revisions %d and %d\n') % parents)
929 else:
929 else:
930 ui.status(_('working directory now based on '
930 ui.status(_('working directory now based on '
931 'revision %d\n') % parents)
931 'revision %d\n') % parents)
932 # TODO: if we know which new heads may result from this rollback, pass
932 # TODO: if we know which new heads may result from this rollback, pass
933 # them to destroy(), which will prevent the branchhead cache from being
933 # them to destroy(), which will prevent the branchhead cache from being
934 # invalidated.
934 # invalidated.
935 self.destroyed()
935 self.destroyed()
936 return 0
936 return 0
937
937
938 def invalidatecaches(self):
938 def invalidatecaches(self):
939
939
940 if '_tagscache' in vars(self):
940 if '_tagscache' in vars(self):
941 # can't use delattr on proxy
941 # can't use delattr on proxy
942 del self.__dict__['_tagscache']
942 del self.__dict__['_tagscache']
943
943
944 self.unfiltered()._branchcaches.clear()
944 self.unfiltered()._branchcaches.clear()
945 self.invalidatevolatilesets()
945 self.invalidatevolatilesets()
946
946
947 def invalidatevolatilesets(self):
947 def invalidatevolatilesets(self):
948 self.filteredrevcache.clear()
948 self.filteredrevcache.clear()
949 obsolete.clearobscaches(self)
949 obsolete.clearobscaches(self)
950
950
951 def invalidatedirstate(self):
951 def invalidatedirstate(self):
952 '''Invalidates the dirstate, causing the next call to dirstate
952 '''Invalidates the dirstate, causing the next call to dirstate
953 to check if it was modified since the last time it was read,
953 to check if it was modified since the last time it was read,
954 rereading it if it has.
954 rereading it if it has.
955
955
956 This is different to dirstate.invalidate() that it doesn't always
956 This is different to dirstate.invalidate() that it doesn't always
957 rereads the dirstate. Use dirstate.invalidate() if you want to
957 rereads the dirstate. Use dirstate.invalidate() if you want to
958 explicitly read the dirstate again (i.e. restoring it to a previous
958 explicitly read the dirstate again (i.e. restoring it to a previous
959 known good state).'''
959 known good state).'''
960 if hasunfilteredcache(self, 'dirstate'):
960 if hasunfilteredcache(self, 'dirstate'):
961 for k in self.dirstate._filecache:
961 for k in self.dirstate._filecache:
962 try:
962 try:
963 delattr(self.dirstate, k)
963 delattr(self.dirstate, k)
964 except AttributeError:
964 except AttributeError:
965 pass
965 pass
966 delattr(self.unfiltered(), 'dirstate')
966 delattr(self.unfiltered(), 'dirstate')
967
967
968 def invalidate(self):
968 def invalidate(self):
969 unfiltered = self.unfiltered() # all file caches are stored unfiltered
969 unfiltered = self.unfiltered() # all file caches are stored unfiltered
970 for k in self._filecache:
970 for k in self._filecache:
971 # dirstate is invalidated separately in invalidatedirstate()
971 # dirstate is invalidated separately in invalidatedirstate()
972 if k == 'dirstate':
972 if k == 'dirstate':
973 continue
973 continue
974
974
975 try:
975 try:
976 delattr(unfiltered, k)
976 delattr(unfiltered, k)
977 except AttributeError:
977 except AttributeError:
978 pass
978 pass
979 self.invalidatecaches()
979 self.invalidatecaches()
980
980
981 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
981 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
982 try:
982 try:
983 l = lock.lock(lockname, 0, releasefn, desc=desc)
983 l = lock.lock(lockname, 0, releasefn, desc=desc)
984 except error.LockHeld, inst:
984 except error.LockHeld, inst:
985 if not wait:
985 if not wait:
986 raise
986 raise
987 self.ui.warn(_("waiting for lock on %s held by %r\n") %
987 self.ui.warn(_("waiting for lock on %s held by %r\n") %
988 (desc, inst.locker))
988 (desc, inst.locker))
989 # default to 600 seconds timeout
989 # default to 600 seconds timeout
990 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
990 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
991 releasefn, desc=desc)
991 releasefn, desc=desc)
992 if acquirefn:
992 if acquirefn:
993 acquirefn()
993 acquirefn()
994 return l
994 return l
995
995
996 def _afterlock(self, callback):
996 def _afterlock(self, callback):
997 """add a callback to the current repository lock.
997 """add a callback to the current repository lock.
998
998
999 The callback will be executed on lock release."""
999 The callback will be executed on lock release."""
1000 l = self._lockref and self._lockref()
1000 l = self._lockref and self._lockref()
1001 if l:
1001 if l:
1002 l.postrelease.append(callback)
1002 l.postrelease.append(callback)
1003 else:
1003 else:
1004 callback()
1004 callback()
1005
1005
1006 def lock(self, wait=True):
1006 def lock(self, wait=True):
1007 '''Lock the repository store (.hg/store) and return a weak reference
1007 '''Lock the repository store (.hg/store) and return a weak reference
1008 to the lock. Use this before modifying the store (e.g. committing or
1008 to the lock. Use this before modifying the store (e.g. committing or
1009 stripping). If you are opening a transaction, get a lock as well.)'''
1009 stripping). If you are opening a transaction, get a lock as well.)'''
1010 l = self._lockref and self._lockref()
1010 l = self._lockref and self._lockref()
1011 if l is not None and l.held:
1011 if l is not None and l.held:
1012 l.lock()
1012 l.lock()
1013 return l
1013 return l
1014
1014
1015 def unlock():
1015 def unlock():
1016 self.store.write()
1016 self.store.write()
1017 if hasunfilteredcache(self, '_phasecache'):
1017 if hasunfilteredcache(self, '_phasecache'):
1018 self._phasecache.write()
1018 self._phasecache.write()
1019 for k, ce in self._filecache.items():
1019 for k, ce in self._filecache.items():
1020 if k == 'dirstate' or k not in self.__dict__:
1020 if k == 'dirstate' or k not in self.__dict__:
1021 continue
1021 continue
1022 ce.refresh()
1022 ce.refresh()
1023
1023
1024 l = self._lock(self.sjoin("lock"), wait, unlock,
1024 l = self._lock(self.sjoin("lock"), wait, unlock,
1025 self.invalidate, _('repository %s') % self.origroot)
1025 self.invalidate, _('repository %s') % self.origroot)
1026 self._lockref = weakref.ref(l)
1026 self._lockref = weakref.ref(l)
1027 return l
1027 return l
1028
1028
1029 def wlock(self, wait=True):
1029 def wlock(self, wait=True):
1030 '''Lock the non-store parts of the repository (everything under
1030 '''Lock the non-store parts of the repository (everything under
1031 .hg except .hg/store) and return a weak reference to the lock.
1031 .hg except .hg/store) and return a weak reference to the lock.
1032 Use this before modifying files in .hg.'''
1032 Use this before modifying files in .hg.'''
1033 l = self._wlockref and self._wlockref()
1033 l = self._wlockref and self._wlockref()
1034 if l is not None and l.held:
1034 if l is not None and l.held:
1035 l.lock()
1035 l.lock()
1036 return l
1036 return l
1037
1037
1038 def unlock():
1038 def unlock():
1039 self.dirstate.write()
1039 self.dirstate.write()
1040 self._filecache['dirstate'].refresh()
1040 self._filecache['dirstate'].refresh()
1041
1041
1042 l = self._lock(self.join("wlock"), wait, unlock,
1042 l = self._lock(self.join("wlock"), wait, unlock,
1043 self.invalidatedirstate, _('working directory of %s') %
1043 self.invalidatedirstate, _('working directory of %s') %
1044 self.origroot)
1044 self.origroot)
1045 self._wlockref = weakref.ref(l)
1045 self._wlockref = weakref.ref(l)
1046 return l
1046 return l
1047
1047
1048 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1048 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1049 """
1049 """
1050 commit an individual file as part of a larger transaction
1050 commit an individual file as part of a larger transaction
1051 """
1051 """
1052
1052
1053 fname = fctx.path()
1053 fname = fctx.path()
1054 text = fctx.data()
1054 text = fctx.data()
1055 flog = self.file(fname)
1055 flog = self.file(fname)
1056 fparent1 = manifest1.get(fname, nullid)
1056 fparent1 = manifest1.get(fname, nullid)
1057 fparent2 = fparent2o = manifest2.get(fname, nullid)
1057 fparent2 = fparent2o = manifest2.get(fname, nullid)
1058
1058
1059 meta = {}
1059 meta = {}
1060 copy = fctx.renamed()
1060 copy = fctx.renamed()
1061 if copy and copy[0] != fname:
1061 if copy and copy[0] != fname:
1062 # Mark the new revision of this file as a copy of another
1062 # Mark the new revision of this file as a copy of another
1063 # file. This copy data will effectively act as a parent
1063 # file. This copy data will effectively act as a parent
1064 # of this new revision. If this is a merge, the first
1064 # of this new revision. If this is a merge, the first
1065 # parent will be the nullid (meaning "look up the copy data")
1065 # parent will be the nullid (meaning "look up the copy data")
1066 # and the second one will be the other parent. For example:
1066 # and the second one will be the other parent. For example:
1067 #
1067 #
1068 # 0 --- 1 --- 3 rev1 changes file foo
1068 # 0 --- 1 --- 3 rev1 changes file foo
1069 # \ / rev2 renames foo to bar and changes it
1069 # \ / rev2 renames foo to bar and changes it
1070 # \- 2 -/ rev3 should have bar with all changes and
1070 # \- 2 -/ rev3 should have bar with all changes and
1071 # should record that bar descends from
1071 # should record that bar descends from
1072 # bar in rev2 and foo in rev1
1072 # bar in rev2 and foo in rev1
1073 #
1073 #
1074 # this allows this merge to succeed:
1074 # this allows this merge to succeed:
1075 #
1075 #
1076 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1076 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1077 # \ / merging rev3 and rev4 should use bar@rev2
1077 # \ / merging rev3 and rev4 should use bar@rev2
1078 # \- 2 --- 4 as the merge base
1078 # \- 2 --- 4 as the merge base
1079 #
1079 #
1080
1080
1081 cfname = copy[0]
1081 cfname = copy[0]
1082 crev = manifest1.get(cfname)
1082 crev = manifest1.get(cfname)
1083 newfparent = fparent2
1083 newfparent = fparent2
1084
1084
1085 if manifest2: # branch merge
1085 if manifest2: # branch merge
1086 if fparent2 == nullid or crev is None: # copied on remote side
1086 if fparent2 == nullid or crev is None: # copied on remote side
1087 if cfname in manifest2:
1087 if cfname in manifest2:
1088 crev = manifest2[cfname]
1088 crev = manifest2[cfname]
1089 newfparent = fparent1
1089 newfparent = fparent1
1090
1090
1091 # find source in nearest ancestor if we've lost track
1091 # find source in nearest ancestor if we've lost track
1092 if not crev:
1092 if not crev:
1093 self.ui.debug(" %s: searching for copy revision for %s\n" %
1093 self.ui.debug(" %s: searching for copy revision for %s\n" %
1094 (fname, cfname))
1094 (fname, cfname))
1095 for ancestor in self[None].ancestors():
1095 for ancestor in self[None].ancestors():
1096 if cfname in ancestor:
1096 if cfname in ancestor:
1097 crev = ancestor[cfname].filenode()
1097 crev = ancestor[cfname].filenode()
1098 break
1098 break
1099
1099
1100 if crev:
1100 if crev:
1101 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1101 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1102 meta["copy"] = cfname
1102 meta["copy"] = cfname
1103 meta["copyrev"] = hex(crev)
1103 meta["copyrev"] = hex(crev)
1104 fparent1, fparent2 = nullid, newfparent
1104 fparent1, fparent2 = nullid, newfparent
1105 else:
1105 else:
1106 self.ui.warn(_("warning: can't find ancestor for '%s' "
1106 self.ui.warn(_("warning: can't find ancestor for '%s' "
1107 "copied from '%s'!\n") % (fname, cfname))
1107 "copied from '%s'!\n") % (fname, cfname))
1108
1108
1109 elif fparent2 != nullid:
1109 elif fparent2 != nullid:
1110 # is one parent an ancestor of the other?
1110 # is one parent an ancestor of the other?
1111 fparentancestor = flog.ancestor(fparent1, fparent2)
1111 fparentancestor = flog.ancestor(fparent1, fparent2)
1112 if fparentancestor == fparent1:
1112 if fparentancestor == fparent1:
1113 fparent1, fparent2 = fparent2, nullid
1113 fparent1, fparent2 = fparent2, nullid
1114 elif fparentancestor == fparent2:
1114 elif fparentancestor == fparent2:
1115 fparent2 = nullid
1115 fparent2 = nullid
1116
1116
1117 # is the file changed?
1117 # is the file changed?
1118 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1118 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1119 changelist.append(fname)
1119 changelist.append(fname)
1120 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1120 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1121
1121
1122 # are just the flags changed during merge?
1122 # are just the flags changed during merge?
1123 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1123 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1124 changelist.append(fname)
1124 changelist.append(fname)
1125
1125
1126 return fparent1
1126 return fparent1
1127
1127
1128 @unfilteredmethod
1128 @unfilteredmethod
1129 def commit(self, text="", user=None, date=None, match=None, force=False,
1129 def commit(self, text="", user=None, date=None, match=None, force=False,
1130 editor=False, extra={}):
1130 editor=False, extra={}):
1131 """Add a new revision to current repository.
1131 """Add a new revision to current repository.
1132
1132
1133 Revision information is gathered from the working directory,
1133 Revision information is gathered from the working directory,
1134 match can be used to filter the committed files. If editor is
1134 match can be used to filter the committed files. If editor is
1135 supplied, it is called to get a commit message.
1135 supplied, it is called to get a commit message.
1136 """
1136 """
1137
1137
1138 def fail(f, msg):
1138 def fail(f, msg):
1139 raise util.Abort('%s: %s' % (f, msg))
1139 raise util.Abort('%s: %s' % (f, msg))
1140
1140
1141 if not match:
1141 if not match:
1142 match = matchmod.always(self.root, '')
1142 match = matchmod.always(self.root, '')
1143
1143
1144 if not force:
1144 if not force:
1145 vdirs = []
1145 vdirs = []
1146 match.dir = vdirs.append
1146 match.dir = vdirs.append
1147 match.bad = fail
1147 match.bad = fail
1148
1148
1149 wlock = self.wlock()
1149 wlock = self.wlock()
1150 try:
1150 try:
1151 wctx = self[None]
1151 wctx = self[None]
1152 merge = len(wctx.parents()) > 1
1152 merge = len(wctx.parents()) > 1
1153
1153
1154 if (not force and merge and match and
1154 if (not force and merge and match and
1155 (match.files() or match.anypats())):
1155 (match.files() or match.anypats())):
1156 raise util.Abort(_('cannot partially commit a merge '
1156 raise util.Abort(_('cannot partially commit a merge '
1157 '(do not specify files or patterns)'))
1157 '(do not specify files or patterns)'))
1158
1158
1159 changes = self.status(match=match, clean=force)
1159 changes = self.status(match=match, clean=force)
1160 if force:
1160 if force:
1161 changes[0].extend(changes[6]) # mq may commit unchanged files
1161 changes[0].extend(changes[6]) # mq may commit unchanged files
1162
1162
1163 # check subrepos
1163 # check subrepos
1164 subs = []
1164 subs = []
1165 commitsubs = set()
1165 commitsubs = set()
1166 newstate = wctx.substate.copy()
1166 newstate = wctx.substate.copy()
1167 # only manage subrepos and .hgsubstate if .hgsub is present
1167 # only manage subrepos and .hgsubstate if .hgsub is present
1168 if '.hgsub' in wctx:
1168 if '.hgsub' in wctx:
1169 # we'll decide whether to track this ourselves, thanks
1169 # we'll decide whether to track this ourselves, thanks
1170 if '.hgsubstate' in changes[0]:
1170 if '.hgsubstate' in changes[0]:
1171 changes[0].remove('.hgsubstate')
1171 changes[0].remove('.hgsubstate')
1172 if '.hgsubstate' in changes[2]:
1172 if '.hgsubstate' in changes[2]:
1173 changes[2].remove('.hgsubstate')
1173 changes[2].remove('.hgsubstate')
1174
1174
1175 # compare current state to last committed state
1175 # compare current state to last committed state
1176 # build new substate based on last committed state
1176 # build new substate based on last committed state
1177 oldstate = wctx.p1().substate
1177 oldstate = wctx.p1().substate
1178 for s in sorted(newstate.keys()):
1178 for s in sorted(newstate.keys()):
1179 if not match(s):
1179 if not match(s):
1180 # ignore working copy, use old state if present
1180 # ignore working copy, use old state if present
1181 if s in oldstate:
1181 if s in oldstate:
1182 newstate[s] = oldstate[s]
1182 newstate[s] = oldstate[s]
1183 continue
1183 continue
1184 if not force:
1184 if not force:
1185 raise util.Abort(
1185 raise util.Abort(
1186 _("commit with new subrepo %s excluded") % s)
1186 _("commit with new subrepo %s excluded") % s)
1187 if wctx.sub(s).dirty(True):
1187 if wctx.sub(s).dirty(True):
1188 if not self.ui.configbool('ui', 'commitsubrepos'):
1188 if not self.ui.configbool('ui', 'commitsubrepos'):
1189 raise util.Abort(
1189 raise util.Abort(
1190 _("uncommitted changes in subrepo %s") % s,
1190 _("uncommitted changes in subrepo %s") % s,
1191 hint=_("use --subrepos for recursive commit"))
1191 hint=_("use --subrepos for recursive commit"))
1192 subs.append(s)
1192 subs.append(s)
1193 commitsubs.add(s)
1193 commitsubs.add(s)
1194 else:
1194 else:
1195 bs = wctx.sub(s).basestate()
1195 bs = wctx.sub(s).basestate()
1196 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1196 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1197 if oldstate.get(s, (None, None, None))[1] != bs:
1197 if oldstate.get(s, (None, None, None))[1] != bs:
1198 subs.append(s)
1198 subs.append(s)
1199
1199
1200 # check for removed subrepos
1200 # check for removed subrepos
1201 for p in wctx.parents():
1201 for p in wctx.parents():
1202 r = [s for s in p.substate if s not in newstate]
1202 r = [s for s in p.substate if s not in newstate]
1203 subs += [s for s in r if match(s)]
1203 subs += [s for s in r if match(s)]
1204 if subs:
1204 if subs:
1205 if (not match('.hgsub') and
1205 if (not match('.hgsub') and
1206 '.hgsub' in (wctx.modified() + wctx.added())):
1206 '.hgsub' in (wctx.modified() + wctx.added())):
1207 raise util.Abort(
1207 raise util.Abort(
1208 _("can't commit subrepos without .hgsub"))
1208 _("can't commit subrepos without .hgsub"))
1209 changes[0].insert(0, '.hgsubstate')
1209 changes[0].insert(0, '.hgsubstate')
1210
1210
1211 elif '.hgsub' in changes[2]:
1211 elif '.hgsub' in changes[2]:
1212 # clean up .hgsubstate when .hgsub is removed
1212 # clean up .hgsubstate when .hgsub is removed
1213 if ('.hgsubstate' in wctx and
1213 if ('.hgsubstate' in wctx and
1214 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1214 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1215 changes[2].insert(0, '.hgsubstate')
1215 changes[2].insert(0, '.hgsubstate')
1216
1216
1217 # make sure all explicit patterns are matched
1217 # make sure all explicit patterns are matched
1218 if not force and match.files():
1218 if not force and match.files():
1219 matched = set(changes[0] + changes[1] + changes[2])
1219 matched = set(changes[0] + changes[1] + changes[2])
1220
1220
1221 for f in match.files():
1221 for f in match.files():
1222 f = self.dirstate.normalize(f)
1222 f = self.dirstate.normalize(f)
1223 if f == '.' or f in matched or f in wctx.substate:
1223 if f == '.' or f in matched or f in wctx.substate:
1224 continue
1224 continue
1225 if f in changes[3]: # missing
1225 if f in changes[3]: # missing
1226 fail(f, _('file not found!'))
1226 fail(f, _('file not found!'))
1227 if f in vdirs: # visited directory
1227 if f in vdirs: # visited directory
1228 d = f + '/'
1228 d = f + '/'
1229 for mf in matched:
1229 for mf in matched:
1230 if mf.startswith(d):
1230 if mf.startswith(d):
1231 break
1231 break
1232 else:
1232 else:
1233 fail(f, _("no match under directory!"))
1233 fail(f, _("no match under directory!"))
1234 elif f not in self.dirstate:
1234 elif f not in self.dirstate:
1235 fail(f, _("file not tracked!"))
1235 fail(f, _("file not tracked!"))
1236
1236
1237 cctx = context.workingctx(self, text, user, date, extra, changes)
1237 cctx = context.workingctx(self, text, user, date, extra, changes)
1238
1238
1239 if (not force and not extra.get("close") and not merge
1239 if (not force and not extra.get("close") and not merge
1240 and not cctx.files()
1240 and not cctx.files()
1241 and wctx.branch() == wctx.p1().branch()):
1241 and wctx.branch() == wctx.p1().branch()):
1242 return None
1242 return None
1243
1243
1244 if merge and cctx.deleted():
1244 if merge and cctx.deleted():
1245 raise util.Abort(_("cannot commit merge with missing files"))
1245 raise util.Abort(_("cannot commit merge with missing files"))
1246
1246
1247 ms = mergemod.mergestate(self)
1247 ms = mergemod.mergestate(self)
1248 for f in changes[0]:
1248 for f in changes[0]:
1249 if f in ms and ms[f] == 'u':
1249 if f in ms and ms[f] == 'u':
1250 raise util.Abort(_("unresolved merge conflicts "
1250 raise util.Abort(_("unresolved merge conflicts "
1251 "(see hg help resolve)"))
1251 "(see hg help resolve)"))
1252
1252
1253 if editor:
1253 if editor:
1254 cctx._text = editor(self, cctx, subs)
1254 cctx._text = editor(self, cctx, subs)
1255 edited = (text != cctx._text)
1255 edited = (text != cctx._text)
1256
1256
1257 # commit subs and write new state
1257 # commit subs and write new state
1258 if subs:
1258 if subs:
1259 for s in sorted(commitsubs):
1259 for s in sorted(commitsubs):
1260 sub = wctx.sub(s)
1260 sub = wctx.sub(s)
1261 self.ui.status(_('committing subrepository %s\n') %
1261 self.ui.status(_('committing subrepository %s\n') %
1262 subrepo.subrelpath(sub))
1262 subrepo.subrelpath(sub))
1263 sr = sub.commit(cctx._text, user, date)
1263 sr = sub.commit(cctx._text, user, date)
1264 newstate[s] = (newstate[s][0], sr)
1264 newstate[s] = (newstate[s][0], sr)
1265 subrepo.writestate(self, newstate)
1265 subrepo.writestate(self, newstate)
1266
1266
1267 # Save commit message in case this transaction gets rolled back
1267 # Save commit message in case this transaction gets rolled back
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1269 # the assumption that the user will use the same editor again.
1269 # the assumption that the user will use the same editor again.
1270 msgfn = self.savecommitmessage(cctx._text)
1270 msgfn = self.savecommitmessage(cctx._text)
1271
1271
1272 p1, p2 = self.dirstate.parents()
1272 p1, p2 = self.dirstate.parents()
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1274 try:
1274 try:
1275 self.hook("precommit", throw=True, parent1=hookp1,
1275 self.hook("precommit", throw=True, parent1=hookp1,
1276 parent2=hookp2)
1276 parent2=hookp2)
1277 ret = self.commitctx(cctx, True)
1277 ret = self.commitctx(cctx, True)
1278 except: # re-raises
1278 except: # re-raises
1279 if edited:
1279 if edited:
1280 self.ui.write(
1280 self.ui.write(
1281 _('note: commit message saved in %s\n') % msgfn)
1281 _('note: commit message saved in %s\n') % msgfn)
1282 raise
1282 raise
1283
1283
1284 # update bookmarks, dirstate and mergestate
1284 # update bookmarks, dirstate and mergestate
1285 bookmarks.update(self, [p1, p2], ret)
1285 bookmarks.update(self, [p1, p2], ret)
1286 cctx.markcommitted(ret)
1286 cctx.markcommitted(ret)
1287 ms.reset()
1287 ms.reset()
1288 finally:
1288 finally:
1289 wlock.release()
1289 wlock.release()
1290
1290
1291 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1291 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1292 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1292 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1293 self._afterlock(commithook)
1293 self._afterlock(commithook)
1294 return ret
1294 return ret
1295
1295
1296 @unfilteredmethod
1296 @unfilteredmethod
1297 def commitctx(self, ctx, error=False):
1297 def commitctx(self, ctx, error=False):
1298 """Add a new revision to current repository.
1298 """Add a new revision to current repository.
1299 Revision information is passed via the context argument.
1299 Revision information is passed via the context argument.
1300 """
1300 """
1301
1301
1302 tr = lock = None
1302 tr = lock = None
1303 removed = list(ctx.removed())
1303 removed = list(ctx.removed())
1304 p1, p2 = ctx.p1(), ctx.p2()
1304 p1, p2 = ctx.p1(), ctx.p2()
1305 user = ctx.user()
1305 user = ctx.user()
1306
1306
1307 lock = self.lock()
1307 lock = self.lock()
1308 try:
1308 try:
1309 tr = self.transaction("commit")
1309 tr = self.transaction("commit")
1310 trp = weakref.proxy(tr)
1310 trp = weakref.proxy(tr)
1311
1311
1312 if ctx.files():
1312 if ctx.files():
1313 m1 = p1.manifest().copy()
1313 m1 = p1.manifest().copy()
1314 m2 = p2.manifest()
1314 m2 = p2.manifest()
1315
1315
1316 # check in files
1316 # check in files
1317 new = {}
1317 new = {}
1318 changed = []
1318 changed = []
1319 linkrev = len(self)
1319 linkrev = len(self)
1320 for f in sorted(ctx.modified() + ctx.added()):
1320 for f in sorted(ctx.modified() + ctx.added()):
1321 self.ui.note(f + "\n")
1321 self.ui.note(f + "\n")
1322 try:
1322 try:
1323 fctx = ctx[f]
1323 fctx = ctx[f]
1324 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1324 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1325 changed)
1325 changed)
1326 m1.set(f, fctx.flags())
1326 m1.set(f, fctx.flags())
1327 except OSError, inst:
1327 except OSError, inst:
1328 self.ui.warn(_("trouble committing %s!\n") % f)
1328 self.ui.warn(_("trouble committing %s!\n") % f)
1329 raise
1329 raise
1330 except IOError, inst:
1330 except IOError, inst:
1331 errcode = getattr(inst, 'errno', errno.ENOENT)
1331 errcode = getattr(inst, 'errno', errno.ENOENT)
1332 if error or errcode and errcode != errno.ENOENT:
1332 if error or errcode and errcode != errno.ENOENT:
1333 self.ui.warn(_("trouble committing %s!\n") % f)
1333 self.ui.warn(_("trouble committing %s!\n") % f)
1334 raise
1334 raise
1335 else:
1335 else:
1336 removed.append(f)
1336 removed.append(f)
1337
1337
1338 # update manifest
1338 # update manifest
1339 m1.update(new)
1339 m1.update(new)
1340 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1340 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1341 drop = [f for f in removed if f in m1]
1341 drop = [f for f in removed if f in m1]
1342 for f in drop:
1342 for f in drop:
1343 del m1[f]
1343 del m1[f]
1344 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1344 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1345 p2.manifestnode(), (new, drop))
1345 p2.manifestnode(), (new, drop))
1346 files = changed + removed
1346 files = changed + removed
1347 else:
1347 else:
1348 mn = p1.manifestnode()
1348 mn = p1.manifestnode()
1349 files = []
1349 files = []
1350
1350
1351 # update changelog
1351 # update changelog
1352 self.changelog.delayupdate()
1352 self.changelog.delayupdate()
1353 n = self.changelog.add(mn, files, ctx.description(),
1353 n = self.changelog.add(mn, files, ctx.description(),
1354 trp, p1.node(), p2.node(),
1354 trp, p1.node(), p2.node(),
1355 user, ctx.date(), ctx.extra().copy())
1355 user, ctx.date(), ctx.extra().copy())
1356 p = lambda: self.changelog.writepending() and self.root or ""
1356 p = lambda: self.changelog.writepending() and self.root or ""
1357 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1357 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1358 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1358 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1359 parent2=xp2, pending=p)
1359 parent2=xp2, pending=p)
1360 self.changelog.finalize(trp)
1360 self.changelog.finalize(trp)
1361 # set the new commit is proper phase
1361 # set the new commit is proper phase
1362 targetphase = phases.newcommitphase(self.ui)
1362 targetphase = phases.newcommitphase(self.ui)
1363 if targetphase:
1363 if targetphase:
1364 # retract boundary do not alter parent changeset.
1364 # retract boundary do not alter parent changeset.
1365 # if a parent have higher the resulting phase will
1365 # if a parent have higher the resulting phase will
1366 # be compliant anyway
1366 # be compliant anyway
1367 #
1367 #
1368 # if minimal phase was 0 we don't need to retract anything
1368 # if minimal phase was 0 we don't need to retract anything
1369 phases.retractboundary(self, targetphase, [n])
1369 phases.retractboundary(self, targetphase, [n])
1370 tr.close()
1370 tr.close()
1371 branchmap.updatecache(self.filtered('served'))
1371 branchmap.updatecache(self.filtered('served'))
1372 return n
1372 return n
1373 finally:
1373 finally:
1374 if tr:
1374 if tr:
1375 tr.release()
1375 tr.release()
1376 lock.release()
1376 lock.release()
1377
1377
1378 @unfilteredmethod
1378 @unfilteredmethod
1379 def destroying(self):
1379 def destroying(self):
1380 '''Inform the repository that nodes are about to be destroyed.
1380 '''Inform the repository that nodes are about to be destroyed.
1381 Intended for use by strip and rollback, so there's a common
1381 Intended for use by strip and rollback, so there's a common
1382 place for anything that has to be done before destroying history.
1382 place for anything that has to be done before destroying history.
1383
1383
1384 This is mostly useful for saving state that is in memory and waiting
1384 This is mostly useful for saving state that is in memory and waiting
1385 to be flushed when the current lock is released. Because a call to
1385 to be flushed when the current lock is released. Because a call to
1386 destroyed is imminent, the repo will be invalidated causing those
1386 destroyed is imminent, the repo will be invalidated causing those
1387 changes to stay in memory (waiting for the next unlock), or vanish
1387 changes to stay in memory (waiting for the next unlock), or vanish
1388 completely.
1388 completely.
1389 '''
1389 '''
1390 # When using the same lock to commit and strip, the phasecache is left
1390 # When using the same lock to commit and strip, the phasecache is left
1391 # dirty after committing. Then when we strip, the repo is invalidated,
1391 # dirty after committing. Then when we strip, the repo is invalidated,
1392 # causing those changes to disappear.
1392 # causing those changes to disappear.
1393 if '_phasecache' in vars(self):
1393 if '_phasecache' in vars(self):
1394 self._phasecache.write()
1394 self._phasecache.write()
1395
1395
1396 @unfilteredmethod
1396 @unfilteredmethod
1397 def destroyed(self):
1397 def destroyed(self):
1398 '''Inform the repository that nodes have been destroyed.
1398 '''Inform the repository that nodes have been destroyed.
1399 Intended for use by strip and rollback, so there's a common
1399 Intended for use by strip and rollback, so there's a common
1400 place for anything that has to be done after destroying history.
1400 place for anything that has to be done after destroying history.
1401 '''
1401 '''
1402 # When one tries to:
1402 # When one tries to:
1403 # 1) destroy nodes thus calling this method (e.g. strip)
1403 # 1) destroy nodes thus calling this method (e.g. strip)
1404 # 2) use phasecache somewhere (e.g. commit)
1404 # 2) use phasecache somewhere (e.g. commit)
1405 #
1405 #
1406 # then 2) will fail because the phasecache contains nodes that were
1406 # then 2) will fail because the phasecache contains nodes that were
1407 # removed. We can either remove phasecache from the filecache,
1407 # removed. We can either remove phasecache from the filecache,
1408 # causing it to reload next time it is accessed, or simply filter
1408 # causing it to reload next time it is accessed, or simply filter
1409 # the removed nodes now and write the updated cache.
1409 # the removed nodes now and write the updated cache.
1410 self._phasecache.filterunknown(self)
1410 self._phasecache.filterunknown(self)
1411 self._phasecache.write()
1411 self._phasecache.write()
1412
1412
1413 # update the 'served' branch cache to help read only server process
1413 # update the 'served' branch cache to help read only server process
1414 # Thanks to branchcache collaboration this is done from the nearest
1414 # Thanks to branchcache collaboration this is done from the nearest
1415 # filtered subset and it is expected to be fast.
1415 # filtered subset and it is expected to be fast.
1416 branchmap.updatecache(self.filtered('served'))
1416 branchmap.updatecache(self.filtered('served'))
1417
1417
1418 # Ensure the persistent tag cache is updated. Doing it now
1418 # Ensure the persistent tag cache is updated. Doing it now
1419 # means that the tag cache only has to worry about destroyed
1419 # means that the tag cache only has to worry about destroyed
1420 # heads immediately after a strip/rollback. That in turn
1420 # heads immediately after a strip/rollback. That in turn
1421 # guarantees that "cachetip == currenttip" (comparing both rev
1421 # guarantees that "cachetip == currenttip" (comparing both rev
1422 # and node) always means no nodes have been added or destroyed.
1422 # and node) always means no nodes have been added or destroyed.
1423
1423
1424 # XXX this is suboptimal when qrefresh'ing: we strip the current
1424 # XXX this is suboptimal when qrefresh'ing: we strip the current
1425 # head, refresh the tag cache, then immediately add a new head.
1425 # head, refresh the tag cache, then immediately add a new head.
1426 # But I think doing it this way is necessary for the "instant
1426 # But I think doing it this way is necessary for the "instant
1427 # tag cache retrieval" case to work.
1427 # tag cache retrieval" case to work.
1428 self.invalidate()
1428 self.invalidate()
1429
1429
1430 def walk(self, match, node=None):
1430 def walk(self, match, node=None):
1431 '''
1431 '''
1432 walk recursively through the directory tree or a given
1432 walk recursively through the directory tree or a given
1433 changeset, finding all files matched by the match
1433 changeset, finding all files matched by the match
1434 function
1434 function
1435 '''
1435 '''
1436 return self[node].walk(match)
1436 return self[node].walk(match)
1437
1437
1438 def status(self, node1='.', node2=None, match=None,
1438 def status(self, node1='.', node2=None, match=None,
1439 ignored=False, clean=False, unknown=False,
1439 ignored=False, clean=False, unknown=False,
1440 listsubrepos=False):
1440 listsubrepos=False):
1441 """return status of files between two nodes or node and working
1441 """return status of files between two nodes or node and working
1442 directory.
1442 directory.
1443
1443
1444 If node1 is None, use the first dirstate parent instead.
1444 If node1 is None, use the first dirstate parent instead.
1445 If node2 is None, compare node1 with working directory.
1445 If node2 is None, compare node1 with working directory.
1446 """
1446 """
1447
1447
1448 def mfmatches(ctx):
1448 def mfmatches(ctx):
1449 mf = ctx.manifest().copy()
1449 mf = ctx.manifest().copy()
1450 if match.always():
1450 if match.always():
1451 return mf
1451 return mf
1452 for fn in mf.keys():
1452 for fn in mf.keys():
1453 if not match(fn):
1453 if not match(fn):
1454 del mf[fn]
1454 del mf[fn]
1455 return mf
1455 return mf
1456
1456
1457 if isinstance(node1, context.changectx):
1457 if isinstance(node1, context.changectx):
1458 ctx1 = node1
1458 ctx1 = node1
1459 else:
1459 else:
1460 ctx1 = self[node1]
1460 ctx1 = self[node1]
1461 if isinstance(node2, context.changectx):
1461 if isinstance(node2, context.changectx):
1462 ctx2 = node2
1462 ctx2 = node2
1463 else:
1463 else:
1464 ctx2 = self[node2]
1464 ctx2 = self[node2]
1465
1465
1466 working = ctx2.rev() is None
1466 working = ctx2.rev() is None
1467 parentworking = working and ctx1 == self['.']
1467 parentworking = working and ctx1 == self['.']
1468 match = match or matchmod.always(self.root, self.getcwd())
1468 match = match or matchmod.always(self.root, self.getcwd())
1469 listignored, listclean, listunknown = ignored, clean, unknown
1469 listignored, listclean, listunknown = ignored, clean, unknown
1470
1470
1471 # load earliest manifest first for caching reasons
1471 # load earliest manifest first for caching reasons
1472 if not working and ctx2.rev() < ctx1.rev():
1472 if not working and ctx2.rev() < ctx1.rev():
1473 ctx2.manifest()
1473 ctx2.manifest()
1474
1474
1475 if not parentworking:
1475 if not parentworking:
1476 def bad(f, msg):
1476 def bad(f, msg):
1477 # 'f' may be a directory pattern from 'match.files()',
1477 # 'f' may be a directory pattern from 'match.files()',
1478 # so 'f not in ctx1' is not enough
1478 # so 'f not in ctx1' is not enough
1479 if f not in ctx1 and f not in ctx1.dirs():
1479 if f not in ctx1 and f not in ctx1.dirs():
1480 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1480 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1481 match.bad = bad
1481 match.bad = bad
1482
1482
1483 if working: # we need to scan the working dir
1483 if working: # we need to scan the working dir
1484 subrepos = []
1484 subrepos = []
1485 if '.hgsub' in self.dirstate:
1485 if '.hgsub' in self.dirstate:
1486 subrepos = sorted(ctx2.substate)
1486 subrepos = sorted(ctx2.substate)
1487 s = self.dirstate.status(match, subrepos, listignored,
1487 s = self.dirstate.status(match, subrepos, listignored,
1488 listclean, listunknown)
1488 listclean, listunknown)
1489 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1489 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1490
1490
1491 # check for any possibly clean files
1491 # check for any possibly clean files
1492 if parentworking and cmp:
1492 if parentworking and cmp:
1493 fixup = []
1493 fixup = []
1494 # do a full compare of any files that might have changed
1494 # do a full compare of any files that might have changed
1495 for f in sorted(cmp):
1495 for f in sorted(cmp):
1496 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1496 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1497 or ctx1[f].cmp(ctx2[f])):
1497 or ctx1[f].cmp(ctx2[f])):
1498 modified.append(f)
1498 modified.append(f)
1499 else:
1499 else:
1500 fixup.append(f)
1500 fixup.append(f)
1501
1501
1502 # update dirstate for files that are actually clean
1502 # update dirstate for files that are actually clean
1503 if fixup:
1503 if fixup:
1504 if listclean:
1504 if listclean:
1505 clean += fixup
1505 clean += fixup
1506
1506
1507 try:
1507 try:
1508 # updating the dirstate is optional
1508 # updating the dirstate is optional
1509 # so we don't wait on the lock
1509 # so we don't wait on the lock
1510 wlock = self.wlock(False)
1510 wlock = self.wlock(False)
1511 try:
1511 try:
1512 for f in fixup:
1512 for f in fixup:
1513 self.dirstate.normal(f)
1513 self.dirstate.normal(f)
1514 finally:
1514 finally:
1515 wlock.release()
1515 wlock.release()
1516 except error.LockError:
1516 except error.LockError:
1517 pass
1517 pass
1518
1518
1519 if not parentworking:
1519 if not parentworking:
1520 mf1 = mfmatches(ctx1)
1520 mf1 = mfmatches(ctx1)
1521 if working:
1521 if working:
1522 # we are comparing working dir against non-parent
1522 # we are comparing working dir against non-parent
1523 # generate a pseudo-manifest for the working dir
1523 # generate a pseudo-manifest for the working dir
1524 mf2 = mfmatches(self['.'])
1524 mf2 = mfmatches(self['.'])
1525 for f in cmp + modified + added:
1525 for f in cmp + modified + added:
1526 mf2[f] = None
1526 mf2[f] = None
1527 mf2.set(f, ctx2.flags(f))
1527 mf2.set(f, ctx2.flags(f))
1528 for f in removed:
1528 for f in removed:
1529 if f in mf2:
1529 if f in mf2:
1530 del mf2[f]
1530 del mf2[f]
1531 else:
1531 else:
1532 # we are comparing two revisions
1532 # we are comparing two revisions
1533 deleted, unknown, ignored = [], [], []
1533 deleted, unknown, ignored = [], [], []
1534 mf2 = mfmatches(ctx2)
1534 mf2 = mfmatches(ctx2)
1535
1535
1536 modified, added, clean = [], [], []
1536 modified, added, clean = [], [], []
1537 withflags = mf1.withflags() | mf2.withflags()
1537 withflags = mf1.withflags() | mf2.withflags()
1538 for fn, mf2node in mf2.iteritems():
1538 for fn, mf2node in mf2.iteritems():
1539 if fn in mf1:
1539 if fn in mf1:
1540 if (fn not in deleted and
1540 if (fn not in deleted and
1541 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1541 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1542 (mf1[fn] != mf2node and
1542 (mf1[fn] != mf2node and
1543 (mf2node or ctx1[fn].cmp(ctx2[fn]))))):
1543 (mf2node or ctx1[fn].cmp(ctx2[fn]))))):
1544 modified.append(fn)
1544 modified.append(fn)
1545 elif listclean:
1545 elif listclean:
1546 clean.append(fn)
1546 clean.append(fn)
1547 del mf1[fn]
1547 del mf1[fn]
1548 elif fn not in deleted:
1548 elif fn not in deleted:
1549 added.append(fn)
1549 added.append(fn)
1550 removed = mf1.keys()
1550 removed = mf1.keys()
1551
1551
1552 if working and modified and not self.dirstate._checklink:
1552 if working and modified and not self.dirstate._checklink:
1553 # Symlink placeholders may get non-symlink-like contents
1553 # Symlink placeholders may get non-symlink-like contents
1554 # via user error or dereferencing by NFS or Samba servers,
1554 # via user error or dereferencing by NFS or Samba servers,
1555 # so we filter out any placeholders that don't look like a
1555 # so we filter out any placeholders that don't look like a
1556 # symlink
1556 # symlink
1557 sane = []
1557 sane = []
1558 for f in modified:
1558 for f in modified:
1559 if ctx2.flags(f) == 'l':
1559 if ctx2.flags(f) == 'l':
1560 d = ctx2[f].data()
1560 d = ctx2[f].data()
1561 if len(d) >= 1024 or '\n' in d or util.binary(d):
1561 if len(d) >= 1024 or '\n' in d or util.binary(d):
1562 self.ui.debug('ignoring suspect symlink placeholder'
1562 self.ui.debug('ignoring suspect symlink placeholder'
1563 ' "%s"\n' % f)
1563 ' "%s"\n' % f)
1564 continue
1564 continue
1565 sane.append(f)
1565 sane.append(f)
1566 modified = sane
1566 modified = sane
1567
1567
1568 r = modified, added, removed, deleted, unknown, ignored, clean
1568 r = modified, added, removed, deleted, unknown, ignored, clean
1569
1569
1570 if listsubrepos:
1570 if listsubrepos:
1571 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1571 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1572 if working:
1572 if working:
1573 rev2 = None
1573 rev2 = None
1574 else:
1574 else:
1575 rev2 = ctx2.substate[subpath][1]
1575 rev2 = ctx2.substate[subpath][1]
1576 try:
1576 try:
1577 submatch = matchmod.narrowmatcher(subpath, match)
1577 submatch = matchmod.narrowmatcher(subpath, match)
1578 s = sub.status(rev2, match=submatch, ignored=listignored,
1578 s = sub.status(rev2, match=submatch, ignored=listignored,
1579 clean=listclean, unknown=listunknown,
1579 clean=listclean, unknown=listunknown,
1580 listsubrepos=True)
1580 listsubrepos=True)
1581 for rfiles, sfiles in zip(r, s):
1581 for rfiles, sfiles in zip(r, s):
1582 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1582 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1583 except error.LookupError:
1583 except error.LookupError:
1584 self.ui.status(_("skipping missing subrepository: %s\n")
1584 self.ui.status(_("skipping missing subrepository: %s\n")
1585 % subpath)
1585 % subpath)
1586
1586
1587 for l in r:
1587 for l in r:
1588 l.sort()
1588 l.sort()
1589 return r
1589 return r
1590
1590
1591 def heads(self, start=None):
1591 def heads(self, start=None):
1592 heads = self.changelog.heads(start)
1592 heads = self.changelog.heads(start)
1593 # sort the output in rev descending order
1593 # sort the output in rev descending order
1594 return sorted(heads, key=self.changelog.rev, reverse=True)
1594 return sorted(heads, key=self.changelog.rev, reverse=True)
1595
1595
1596 def branchheads(self, branch=None, start=None, closed=False):
1596 def branchheads(self, branch=None, start=None, closed=False):
1597 '''return a (possibly filtered) list of heads for the given branch
1597 '''return a (possibly filtered) list of heads for the given branch
1598
1598
1599 Heads are returned in topological order, from newest to oldest.
1599 Heads are returned in topological order, from newest to oldest.
1600 If branch is None, use the dirstate branch.
1600 If branch is None, use the dirstate branch.
1601 If start is not None, return only heads reachable from start.
1601 If start is not None, return only heads reachable from start.
1602 If closed is True, return heads that are marked as closed as well.
1602 If closed is True, return heads that are marked as closed as well.
1603 '''
1603 '''
1604 if branch is None:
1604 if branch is None:
1605 branch = self[None].branch()
1605 branch = self[None].branch()
1606 branches = self.branchmap()
1606 branches = self.branchmap()
1607 if branch not in branches:
1607 if branch not in branches:
1608 return []
1608 return []
1609 # the cache returns heads ordered lowest to highest
1609 # the cache returns heads ordered lowest to highest
1610 bheads = list(reversed(branches[branch]))
1610 bheads = list(reversed(branches[branch]))
1611 if start is not None:
1611 if start is not None:
1612 # filter out the heads that cannot be reached from startrev
1612 # filter out the heads that cannot be reached from startrev
1613 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1613 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1614 bheads = [h for h in bheads if h in fbheads]
1614 bheads = [h for h in bheads if h in fbheads]
1615 if not closed:
1615 if not closed:
1616 bheads = [h for h in bheads if not self[h].closesbranch()]
1616 bheads = [h for h in bheads if not self[h].closesbranch()]
1617 return bheads
1617 return bheads
1618
1618
1619 def branches(self, nodes):
1619 def branches(self, nodes):
1620 if not nodes:
1620 if not nodes:
1621 nodes = [self.changelog.tip()]
1621 nodes = [self.changelog.tip()]
1622 b = []
1622 b = []
1623 for n in nodes:
1623 for n in nodes:
1624 t = n
1624 t = n
1625 while True:
1625 while True:
1626 p = self.changelog.parents(n)
1626 p = self.changelog.parents(n)
1627 if p[1] != nullid or p[0] == nullid:
1627 if p[1] != nullid or p[0] == nullid:
1628 b.append((t, n, p[0], p[1]))
1628 b.append((t, n, p[0], p[1]))
1629 break
1629 break
1630 n = p[0]
1630 n = p[0]
1631 return b
1631 return b
1632
1632
1633 def between(self, pairs):
1633 def between(self, pairs):
1634 r = []
1634 r = []
1635
1635
1636 for top, bottom in pairs:
1636 for top, bottom in pairs:
1637 n, l, i = top, [], 0
1637 n, l, i = top, [], 0
1638 f = 1
1638 f = 1
1639
1639
1640 while n != bottom and n != nullid:
1640 while n != bottom and n != nullid:
1641 p = self.changelog.parents(n)[0]
1641 p = self.changelog.parents(n)[0]
1642 if i == f:
1642 if i == f:
1643 l.append(n)
1643 l.append(n)
1644 f = f * 2
1644 f = f * 2
1645 n = p
1645 n = p
1646 i += 1
1646 i += 1
1647
1647
1648 r.append(l)
1648 r.append(l)
1649
1649
1650 return r
1650 return r
1651
1651
1652 def pull(self, remote, heads=None, force=False):
1652 def pull(self, remote, heads=None, force=False):
1653 # don't open transaction for nothing or you break future useful
1653 # don't open transaction for nothing or you break future useful
1654 # rollback call
1654 # rollback call
1655 tr = None
1655 tr = None
1656 trname = 'pull\n' + util.hidepassword(remote.url())
1656 trname = 'pull\n' + util.hidepassword(remote.url())
1657 lock = self.lock()
1657 lock = self.lock()
1658 try:
1658 try:
1659 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1659 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1660 force=force)
1660 force=force)
1661 common, fetch, rheads = tmp
1661 common, fetch, rheads = tmp
1662 if not fetch:
1662 if not fetch:
1663 self.ui.status(_("no changes found\n"))
1663 self.ui.status(_("no changes found\n"))
1664 added = []
1664 added = []
1665 result = 0
1665 result = 0
1666 else:
1666 else:
1667 tr = self.transaction(trname)
1667 tr = self.transaction(trname)
1668 if heads is None and list(common) == [nullid]:
1668 if heads is None and list(common) == [nullid]:
1669 self.ui.status(_("requesting all changes\n"))
1669 self.ui.status(_("requesting all changes\n"))
1670 elif heads is None and remote.capable('changegroupsubset'):
1670 elif heads is None and remote.capable('changegroupsubset'):
1671 # issue1320, avoid a race if remote changed after discovery
1671 # issue1320, avoid a race if remote changed after discovery
1672 heads = rheads
1672 heads = rheads
1673
1673
1674 if remote.capable('getbundle'):
1674 if remote.capable('getbundle'):
1675 cg = remote.getbundle('pull', common=common,
1675 cg = remote.getbundle('pull', common=common,
1676 heads=heads or rheads)
1676 heads=heads or rheads)
1677 elif heads is None:
1677 elif heads is None:
1678 cg = remote.changegroup(fetch, 'pull')
1678 cg = remote.changegroup(fetch, 'pull')
1679 elif not remote.capable('changegroupsubset'):
1679 elif not remote.capable('changegroupsubset'):
1680 raise util.Abort(_("partial pull cannot be done because "
1680 raise util.Abort(_("partial pull cannot be done because "
1681 "other repository doesn't support "
1681 "other repository doesn't support "
1682 "changegroupsubset."))
1682 "changegroupsubset."))
1683 else:
1683 else:
1684 cg = remote.changegroupsubset(fetch, heads, 'pull')
1684 cg = remote.changegroupsubset(fetch, heads, 'pull')
1685 # we use unfiltered changelog here because hidden revision must
1685 # we use unfiltered changelog here because hidden revision must
1686 # be taken in account for phase synchronization. They may
1686 # be taken in account for phase synchronization. They may
1687 # becomes public and becomes visible again.
1687 # becomes public and becomes visible again.
1688 cl = self.unfiltered().changelog
1688 cl = self.unfiltered().changelog
1689 clstart = len(cl)
1689 clstart = len(cl)
1690 result = self.addchangegroup(cg, 'pull', remote.url())
1690 result = self.addchangegroup(cg, 'pull', remote.url())
1691 clend = len(cl)
1691 clend = len(cl)
1692 added = [cl.node(r) for r in xrange(clstart, clend)]
1692 added = [cl.node(r) for r in xrange(clstart, clend)]
1693
1693
1694 # compute target subset
1694 # compute target subset
1695 if heads is None:
1695 if heads is None:
1696 # We pulled every thing possible
1696 # We pulled every thing possible
1697 # sync on everything common
1697 # sync on everything common
1698 subset = common + added
1698 subset = common + added
1699 else:
1699 else:
1700 # We pulled a specific subset
1700 # We pulled a specific subset
1701 # sync on this subset
1701 # sync on this subset
1702 subset = heads
1702 subset = heads
1703
1703
1704 # Get remote phases data from remote
1704 # Get remote phases data from remote
1705 remotephases = remote.listkeys('phases')
1705 remotephases = remote.listkeys('phases')
1706 publishing = bool(remotephases.get('publishing', False))
1706 publishing = bool(remotephases.get('publishing', False))
1707 if remotephases and not publishing:
1707 if remotephases and not publishing:
1708 # remote is new and unpublishing
1708 # remote is new and unpublishing
1709 pheads, _dr = phases.analyzeremotephases(self, subset,
1709 pheads, _dr = phases.analyzeremotephases(self, subset,
1710 remotephases)
1710 remotephases)
1711 phases.advanceboundary(self, phases.public, pheads)
1711 phases.advanceboundary(self, phases.public, pheads)
1712 phases.advanceboundary(self, phases.draft, subset)
1712 phases.advanceboundary(self, phases.draft, subset)
1713 else:
1713 else:
1714 # Remote is old or publishing all common changesets
1714 # Remote is old or publishing all common changesets
1715 # should be seen as public
1715 # should be seen as public
1716 phases.advanceboundary(self, phases.public, subset)
1716 phases.advanceboundary(self, phases.public, subset)
1717
1717
1718 if obsolete._enabled:
1718 if obsolete._enabled:
1719 self.ui.debug('fetching remote obsolete markers\n')
1719 self.ui.debug('fetching remote obsolete markers\n')
1720 remoteobs = remote.listkeys('obsolete')
1720 remoteobs = remote.listkeys('obsolete')
1721 if 'dump0' in remoteobs:
1721 if 'dump0' in remoteobs:
1722 if tr is None:
1722 if tr is None:
1723 tr = self.transaction(trname)
1723 tr = self.transaction(trname)
1724 for key in sorted(remoteobs, reverse=True):
1724 for key in sorted(remoteobs, reverse=True):
1725 if key.startswith('dump'):
1725 if key.startswith('dump'):
1726 data = base85.b85decode(remoteobs[key])
1726 data = base85.b85decode(remoteobs[key])
1727 self.obsstore.mergemarkers(tr, data)
1727 self.obsstore.mergemarkers(tr, data)
1728 self.invalidatevolatilesets()
1728 self.invalidatevolatilesets()
1729 if tr is not None:
1729 if tr is not None:
1730 tr.close()
1730 tr.close()
1731 finally:
1731 finally:
1732 if tr is not None:
1732 if tr is not None:
1733 tr.release()
1733 tr.release()
1734 lock.release()
1734 lock.release()
1735
1735
1736 return result
1736 return result
1737
1737
1738 def checkpush(self, force, revs):
1738 def checkpush(self, force, revs):
1739 """Extensions can override this function if additional checks have
1739 """Extensions can override this function if additional checks have
1740 to be performed before pushing, or call it if they override push
1740 to be performed before pushing, or call it if they override push
1741 command.
1741 command.
1742 """
1742 """
1743 pass
1743 pass
1744
1744
1745 def push(self, remote, force=False, revs=None, newbranch=False):
1745 def push(self, remote, force=False, revs=None, newbranch=False):
1746 '''Push outgoing changesets (limited by revs) from the current
1746 '''Push outgoing changesets (limited by revs) from the current
1747 repository to remote. Return an integer:
1747 repository to remote. Return an integer:
1748 - None means nothing to push
1748 - None means nothing to push
1749 - 0 means HTTP error
1749 - 0 means HTTP error
1750 - 1 means we pushed and remote head count is unchanged *or*
1750 - 1 means we pushed and remote head count is unchanged *or*
1751 we have outgoing changesets but refused to push
1751 we have outgoing changesets but refused to push
1752 - other values as described by addchangegroup()
1752 - other values as described by addchangegroup()
1753 '''
1753 '''
1754 # there are two ways to push to remote repo:
1754 # there are two ways to push to remote repo:
1755 #
1755 #
1756 # addchangegroup assumes local user can lock remote
1756 # addchangegroup assumes local user can lock remote
1757 # repo (local filesystem, old ssh servers).
1757 # repo (local filesystem, old ssh servers).
1758 #
1758 #
1759 # unbundle assumes local user cannot lock remote repo (new ssh
1759 # unbundle assumes local user cannot lock remote repo (new ssh
1760 # servers, http servers).
1760 # servers, http servers).
1761
1761
1762 if not remote.canpush():
1762 if not remote.canpush():
1763 raise util.Abort(_("destination does not support push"))
1763 raise util.Abort(_("destination does not support push"))
1764 unfi = self.unfiltered()
1764 unfi = self.unfiltered()
1765 # get local lock as we might write phase data
1765 # get local lock as we might write phase data
1766 locallock = self.lock()
1766 locallock = self.lock()
1767 try:
1767 try:
1768 self.checkpush(force, revs)
1768 self.checkpush(force, revs)
1769 lock = None
1769 lock = None
1770 unbundle = remote.capable('unbundle')
1770 unbundle = remote.capable('unbundle')
1771 if not unbundle:
1771 if not unbundle:
1772 lock = remote.lock()
1772 lock = remote.lock()
1773 try:
1773 try:
1774 # discovery
1774 # discovery
1775 fci = discovery.findcommonincoming
1775 fci = discovery.findcommonincoming
1776 commoninc = fci(unfi, remote, force=force)
1776 commoninc = fci(unfi, remote, force=force)
1777 common, inc, remoteheads = commoninc
1777 common, inc, remoteheads = commoninc
1778 fco = discovery.findcommonoutgoing
1778 fco = discovery.findcommonoutgoing
1779 outgoing = fco(unfi, remote, onlyheads=revs,
1779 outgoing = fco(unfi, remote, onlyheads=revs,
1780 commoninc=commoninc, force=force)
1780 commoninc=commoninc, force=force)
1781
1781
1782
1782
1783 if not outgoing.missing:
1783 if not outgoing.missing:
1784 # nothing to push
1784 # nothing to push
1785 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1785 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1786 ret = None
1786 ret = None
1787 else:
1787 else:
1788 # something to push
1788 # something to push
1789 if not force:
1789 if not force:
1790 # if self.obsstore == False --> no obsolete
1790 # if self.obsstore == False --> no obsolete
1791 # then, save the iteration
1791 # then, save the iteration
1792 if unfi.obsstore:
1792 if unfi.obsstore:
1793 # this message are here for 80 char limit reason
1793 # this message are here for 80 char limit reason
1794 mso = _("push includes obsolete changeset: %s!")
1794 mso = _("push includes obsolete changeset: %s!")
1795 mst = "push includes %s changeset: %s!"
1795 mst = "push includes %s changeset: %s!"
1796 # plain versions for i18n tool to detect them
1796 # plain versions for i18n tool to detect them
1797 _("push includes unstable changeset: %s!")
1797 _("push includes unstable changeset: %s!")
1798 _("push includes bumped changeset: %s!")
1798 _("push includes bumped changeset: %s!")
1799 _("push includes divergent changeset: %s!")
1799 _("push includes divergent changeset: %s!")
1800 # If we are to push if there is at least one
1800 # If we are to push if there is at least one
1801 # obsolete or unstable changeset in missing, at
1801 # obsolete or unstable changeset in missing, at
1802 # least one of the missinghead will be obsolete or
1802 # least one of the missinghead will be obsolete or
1803 # unstable. So checking heads only is ok
1803 # unstable. So checking heads only is ok
1804 for node in outgoing.missingheads:
1804 for node in outgoing.missingheads:
1805 ctx = unfi[node]
1805 ctx = unfi[node]
1806 if ctx.obsolete():
1806 if ctx.obsolete():
1807 raise util.Abort(mso % ctx)
1807 raise util.Abort(mso % ctx)
1808 elif ctx.troubled():
1808 elif ctx.troubled():
1809 raise util.Abort(_(mst)
1809 raise util.Abort(_(mst)
1810 % (ctx.troubles()[0],
1810 % (ctx.troubles()[0],
1811 ctx))
1811 ctx))
1812 discovery.checkheads(unfi, remote, outgoing,
1812 discovery.checkheads(unfi, remote, outgoing,
1813 remoteheads, newbranch,
1813 remoteheads, newbranch,
1814 bool(inc))
1814 bool(inc))
1815
1815
1816 # create a changegroup from local
1816 # create a changegroup from local
1817 if revs is None and not outgoing.excluded:
1817 if revs is None and not outgoing.excluded:
1818 # push everything,
1818 # push everything,
1819 # use the fast path, no race possible on push
1819 # use the fast path, no race possible on push
1820 cg = self._changegroup(outgoing.missing, 'push')
1820 cg = self._changegroup(outgoing.missing, 'push')
1821 else:
1821 else:
1822 cg = self.getlocalbundle('push', outgoing)
1822 cg = self.getlocalbundle('push', outgoing)
1823
1823
1824 # apply changegroup to remote
1824 # apply changegroup to remote
1825 if unbundle:
1825 if unbundle:
1826 # local repo finds heads on server, finds out what
1826 # local repo finds heads on server, finds out what
1827 # revs it must push. once revs transferred, if server
1827 # revs it must push. once revs transferred, if server
1828 # finds it has different heads (someone else won
1828 # finds it has different heads (someone else won
1829 # commit/push race), server aborts.
1829 # commit/push race), server aborts.
1830 if force:
1830 if force:
1831 remoteheads = ['force']
1831 remoteheads = ['force']
1832 # ssh: return remote's addchangegroup()
1832 # ssh: return remote's addchangegroup()
1833 # http: return remote's addchangegroup() or 0 for error
1833 # http: return remote's addchangegroup() or 0 for error
1834 ret = remote.unbundle(cg, remoteheads, 'push')
1834 ret = remote.unbundle(cg, remoteheads, 'push')
1835 else:
1835 else:
1836 # we return an integer indicating remote head count
1836 # we return an integer indicating remote head count
1837 # change
1837 # change
1838 ret = remote.addchangegroup(cg, 'push', self.url())
1838 ret = remote.addchangegroup(cg, 'push', self.url())
1839
1839
1840 if ret:
1840 if ret:
1841 # push succeed, synchronize target of the push
1841 # push succeed, synchronize target of the push
1842 cheads = outgoing.missingheads
1842 cheads = outgoing.missingheads
1843 elif revs is None:
1843 elif revs is None:
1844 # All out push fails. synchronize all common
1844 # All out push fails. synchronize all common
1845 cheads = outgoing.commonheads
1845 cheads = outgoing.commonheads
1846 else:
1846 else:
1847 # I want cheads = heads(::missingheads and ::commonheads)
1847 # I want cheads = heads(::missingheads and ::commonheads)
1848 # (missingheads is revs with secret changeset filtered out)
1848 # (missingheads is revs with secret changeset filtered out)
1849 #
1849 #
1850 # This can be expressed as:
1850 # This can be expressed as:
1851 # cheads = ( (missingheads and ::commonheads)
1851 # cheads = ( (missingheads and ::commonheads)
1852 # + (commonheads and ::missingheads))"
1852 # + (commonheads and ::missingheads))"
1853 # )
1853 # )
1854 #
1854 #
1855 # while trying to push we already computed the following:
1855 # while trying to push we already computed the following:
1856 # common = (::commonheads)
1856 # common = (::commonheads)
1857 # missing = ((commonheads::missingheads) - commonheads)
1857 # missing = ((commonheads::missingheads) - commonheads)
1858 #
1858 #
1859 # We can pick:
1859 # We can pick:
1860 # * missingheads part of common (::commonheads)
1860 # * missingheads part of common (::commonheads)
1861 common = set(outgoing.common)
1861 common = set(outgoing.common)
1862 cheads = [node for node in revs if node in common]
1862 cheads = [node for node in revs if node in common]
1863 # and
1863 # and
1864 # * commonheads parents on missing
1864 # * commonheads parents on missing
1865 revset = unfi.set('%ln and parents(roots(%ln))',
1865 revset = unfi.set('%ln and parents(roots(%ln))',
1866 outgoing.commonheads,
1866 outgoing.commonheads,
1867 outgoing.missing)
1867 outgoing.missing)
1868 cheads.extend(c.node() for c in revset)
1868 cheads.extend(c.node() for c in revset)
1869 # even when we don't push, exchanging phase data is useful
1869 # even when we don't push, exchanging phase data is useful
1870 remotephases = remote.listkeys('phases')
1870 remotephases = remote.listkeys('phases')
1871 if (self.ui.configbool('ui', '_usedassubrepo', False)
1871 if (self.ui.configbool('ui', '_usedassubrepo', False)
1872 and remotephases # server supports phases
1872 and remotephases # server supports phases
1873 and ret is None # nothing was pushed
1873 and ret is None # nothing was pushed
1874 and remotephases.get('publishing', False)):
1874 and remotephases.get('publishing', False)):
1875 # When:
1875 # When:
1876 # - this is a subrepo push
1876 # - this is a subrepo push
1877 # - and remote support phase
1877 # - and remote support phase
1878 # - and no changeset was pushed
1878 # - and no changeset was pushed
1879 # - and remote is publishing
1879 # - and remote is publishing
1880 # We may be in issue 3871 case!
1880 # We may be in issue 3871 case!
1881 # We drop the possible phase synchronisation done by
1881 # We drop the possible phase synchronisation done by
1882 # courtesy to publish changesets possibly locally draft
1882 # courtesy to publish changesets possibly locally draft
1883 # on the remote.
1883 # on the remote.
1884 remotephases = {'publishing': 'True'}
1884 remotephases = {'publishing': 'True'}
1885 if not remotephases: # old server or public only repo
1885 if not remotephases: # old server or public only repo
1886 phases.advanceboundary(self, phases.public, cheads)
1886 phases.advanceboundary(self, phases.public, cheads)
1887 # don't push any phase data as there is nothing to push
1887 # don't push any phase data as there is nothing to push
1888 else:
1888 else:
1889 ana = phases.analyzeremotephases(self, cheads, remotephases)
1889 ana = phases.analyzeremotephases(self, cheads, remotephases)
1890 pheads, droots = ana
1890 pheads, droots = ana
1891 ### Apply remote phase on local
1891 ### Apply remote phase on local
1892 if remotephases.get('publishing', False):
1892 if remotephases.get('publishing', False):
1893 phases.advanceboundary(self, phases.public, cheads)
1893 phases.advanceboundary(self, phases.public, cheads)
1894 else: # publish = False
1894 else: # publish = False
1895 phases.advanceboundary(self, phases.public, pheads)
1895 phases.advanceboundary(self, phases.public, pheads)
1896 phases.advanceboundary(self, phases.draft, cheads)
1896 phases.advanceboundary(self, phases.draft, cheads)
1897 ### Apply local phase on remote
1897 ### Apply local phase on remote
1898
1898
1899 # Get the list of all revs draft on remote by public here.
1899 # Get the list of all revs draft on remote by public here.
1900 # XXX Beware that revset break if droots is not strictly
1900 # XXX Beware that revset break if droots is not strictly
1901 # XXX root we may want to ensure it is but it is costly
1901 # XXX root we may want to ensure it is but it is costly
1902 outdated = unfi.set('heads((%ln::%ln) and public())',
1902 outdated = unfi.set('heads((%ln::%ln) and public())',
1903 droots, cheads)
1903 droots, cheads)
1904 for newremotehead in outdated:
1904 for newremotehead in outdated:
1905 r = remote.pushkey('phases',
1905 r = remote.pushkey('phases',
1906 newremotehead.hex(),
1906 newremotehead.hex(),
1907 str(phases.draft),
1907 str(phases.draft),
1908 str(phases.public))
1908 str(phases.public))
1909 if not r:
1909 if not r:
1910 self.ui.warn(_('updating %s to public failed!\n')
1910 self.ui.warn(_('updating %s to public failed!\n')
1911 % newremotehead)
1911 % newremotehead)
1912 self.ui.debug('try to push obsolete markers to remote\n')
1912 self.ui.debug('try to push obsolete markers to remote\n')
1913 if (obsolete._enabled and self.obsstore and
1913 if (obsolete._enabled and self.obsstore and
1914 'obsolete' in remote.listkeys('namespaces')):
1914 'obsolete' in remote.listkeys('namespaces')):
1915 rslts = []
1915 rslts = []
1916 remotedata = self.listkeys('obsolete')
1916 remotedata = self.listkeys('obsolete')
1917 for key in sorted(remotedata, reverse=True):
1917 for key in sorted(remotedata, reverse=True):
1918 # reverse sort to ensure we end with dump0
1918 # reverse sort to ensure we end with dump0
1919 data = remotedata[key]
1919 data = remotedata[key]
1920 rslts.append(remote.pushkey('obsolete', key, '', data))
1920 rslts.append(remote.pushkey('obsolete', key, '', data))
1921 if [r for r in rslts if not r]:
1921 if [r for r in rslts if not r]:
1922 msg = _('failed to push some obsolete markers!\n')
1922 msg = _('failed to push some obsolete markers!\n')
1923 self.ui.warn(msg)
1923 self.ui.warn(msg)
1924 finally:
1924 finally:
1925 if lock is not None:
1925 if lock is not None:
1926 lock.release()
1926 lock.release()
1927 finally:
1927 finally:
1928 locallock.release()
1928 locallock.release()
1929
1929
1930 self.ui.debug("checking for updated bookmarks\n")
1930 self.ui.debug("checking for updated bookmarks\n")
1931 rb = remote.listkeys('bookmarks')
1931 rb = remote.listkeys('bookmarks')
1932 for k in rb.keys():
1932 for k in rb.keys():
1933 if k in unfi._bookmarks:
1933 if k in unfi._bookmarks:
1934 nr, nl = rb[k], hex(self._bookmarks[k])
1934 nr, nl = rb[k], hex(self._bookmarks[k])
1935 if nr in unfi:
1935 if nr in unfi:
1936 cr = unfi[nr]
1936 cr = unfi[nr]
1937 cl = unfi[nl]
1937 cl = unfi[nl]
1938 if bookmarks.validdest(unfi, cr, cl):
1938 if bookmarks.validdest(unfi, cr, cl):
1939 r = remote.pushkey('bookmarks', k, nr, nl)
1939 r = remote.pushkey('bookmarks', k, nr, nl)
1940 if r:
1940 if r:
1941 self.ui.status(_("updating bookmark %s\n") % k)
1941 self.ui.status(_("updating bookmark %s\n") % k)
1942 else:
1942 else:
1943 self.ui.warn(_('updating bookmark %s'
1943 self.ui.warn(_('updating bookmark %s'
1944 ' failed!\n') % k)
1944 ' failed!\n') % k)
1945
1945
1946 return ret
1946 return ret
1947
1947
1948 def changegroupinfo(self, nodes, source):
1948 def changegroupinfo(self, nodes, source):
1949 if self.ui.verbose or source == 'bundle':
1949 if self.ui.verbose or source == 'bundle':
1950 self.ui.status(_("%d changesets found\n") % len(nodes))
1950 self.ui.status(_("%d changesets found\n") % len(nodes))
1951 if self.ui.debugflag:
1951 if self.ui.debugflag:
1952 self.ui.debug("list of changesets:\n")
1952 self.ui.debug("list of changesets:\n")
1953 for node in nodes:
1953 for node in nodes:
1954 self.ui.debug("%s\n" % hex(node))
1954 self.ui.debug("%s\n" % hex(node))
1955
1955
1956 def changegroupsubset(self, bases, heads, source):
1956 def changegroupsubset(self, bases, heads, source):
1957 """Compute a changegroup consisting of all the nodes that are
1957 """Compute a changegroup consisting of all the nodes that are
1958 descendants of any of the bases and ancestors of any of the heads.
1958 descendants of any of the bases and ancestors of any of the heads.
1959 Return a chunkbuffer object whose read() method will return
1959 Return a chunkbuffer object whose read() method will return
1960 successive changegroup chunks.
1960 successive changegroup chunks.
1961
1961
1962 It is fairly complex as determining which filenodes and which
1962 It is fairly complex as determining which filenodes and which
1963 manifest nodes need to be included for the changeset to be complete
1963 manifest nodes need to be included for the changeset to be complete
1964 is non-trivial.
1964 is non-trivial.
1965
1965
1966 Another wrinkle is doing the reverse, figuring out which changeset in
1966 Another wrinkle is doing the reverse, figuring out which changeset in
1967 the changegroup a particular filenode or manifestnode belongs to.
1967 the changegroup a particular filenode or manifestnode belongs to.
1968 """
1968 """
1969 cl = self.changelog
1969 cl = self.changelog
1970 if not bases:
1970 if not bases:
1971 bases = [nullid]
1971 bases = [nullid]
1972 csets, bases, heads = cl.nodesbetween(bases, heads)
1972 csets, bases, heads = cl.nodesbetween(bases, heads)
1973 # We assume that all ancestors of bases are known
1973 # We assume that all ancestors of bases are known
1974 common = cl.ancestors([cl.rev(n) for n in bases])
1974 common = cl.ancestors([cl.rev(n) for n in bases])
1975 return self._changegroupsubset(common, csets, heads, source)
1975 return self._changegroupsubset(common, csets, heads, source)
1976
1976
1977 def getlocalbundle(self, source, outgoing):
1977 def getlocalbundle(self, source, outgoing):
1978 """Like getbundle, but taking a discovery.outgoing as an argument.
1978 """Like getbundle, but taking a discovery.outgoing as an argument.
1979
1979
1980 This is only implemented for local repos and reuses potentially
1980 This is only implemented for local repos and reuses potentially
1981 precomputed sets in outgoing."""
1981 precomputed sets in outgoing."""
1982 if not outgoing.missing:
1982 if not outgoing.missing:
1983 return None
1983 return None
1984 return self._changegroupsubset(outgoing.common,
1984 return self._changegroupsubset(outgoing.common,
1985 outgoing.missing,
1985 outgoing.missing,
1986 outgoing.missingheads,
1986 outgoing.missingheads,
1987 source)
1987 source)
1988
1988
1989 def getbundle(self, source, heads=None, common=None):
1989 def getbundle(self, source, heads=None, common=None):
1990 """Like changegroupsubset, but returns the set difference between the
1990 """Like changegroupsubset, but returns the set difference between the
1991 ancestors of heads and the ancestors common.
1991 ancestors of heads and the ancestors common.
1992
1992
1993 If heads is None, use the local heads. If common is None, use [nullid].
1993 If heads is None, use the local heads. If common is None, use [nullid].
1994
1994
1995 The nodes in common might not all be known locally due to the way the
1995 The nodes in common might not all be known locally due to the way the
1996 current discovery protocol works.
1996 current discovery protocol works.
1997 """
1997 """
1998 cl = self.changelog
1998 cl = self.changelog
1999 if common:
1999 if common:
2000 hasnode = cl.hasnode
2000 hasnode = cl.hasnode
2001 common = [n for n in common if hasnode(n)]
2001 common = [n for n in common if hasnode(n)]
2002 else:
2002 else:
2003 common = [nullid]
2003 common = [nullid]
2004 if not heads:
2004 if not heads:
2005 heads = cl.heads()
2005 heads = cl.heads()
2006 return self.getlocalbundle(source,
2006 return self.getlocalbundle(source,
2007 discovery.outgoing(cl, common, heads))
2007 discovery.outgoing(cl, common, heads))
2008
2008
2009 @unfilteredmethod
2009 @unfilteredmethod
2010 def _changegroupsubset(self, commonrevs, csets, heads, source):
2010 def _changegroupsubset(self, commonrevs, csets, heads, source):
2011
2011
2012 cl = self.changelog
2012 cl = self.changelog
2013 mf = self.manifest
2013 mf = self.manifest
2014 mfs = {} # needed manifests
2014 mfs = {} # needed manifests
2015 fnodes = {} # needed file nodes
2015 fnodes = {} # needed file nodes
2016 changedfiles = set()
2016 changedfiles = set()
2017 fstate = ['', {}]
2017 fstate = ['', {}]
2018 count = [0, 0]
2018 count = [0, 0]
2019
2019
2020 # can we go through the fast path ?
2020 # can we go through the fast path ?
2021 heads.sort()
2021 heads.sort()
2022 if heads == sorted(self.heads()):
2022 if heads == sorted(self.heads()):
2023 return self._changegroup(csets, source)
2023 return self._changegroup(csets, source)
2024
2024
2025 # slow path
2025 # slow path
2026 self.hook('preoutgoing', throw=True, source=source)
2026 self.hook('preoutgoing', throw=True, source=source)
2027 self.changegroupinfo(csets, source)
2027 self.changegroupinfo(csets, source)
2028
2028
2029 # filter any nodes that claim to be part of the known set
2029 # filter any nodes that claim to be part of the known set
2030 def prune(revlog, missing):
2030 def prune(revlog, missing):
2031 rr, rl = revlog.rev, revlog.linkrev
2031 rr, rl = revlog.rev, revlog.linkrev
2032 return [n for n in missing
2032 return [n for n in missing
2033 if rl(rr(n)) not in commonrevs]
2033 if rl(rr(n)) not in commonrevs]
2034
2034
2035 progress = self.ui.progress
2035 progress = self.ui.progress
2036 _bundling = _('bundling')
2036 _bundling = _('bundling')
2037 _changesets = _('changesets')
2037 _changesets = _('changesets')
2038 _manifests = _('manifests')
2038 _manifests = _('manifests')
2039 _files = _('files')
2039 _files = _('files')
2040
2040
2041 def lookup(revlog, x):
2041 def lookup(revlog, x):
2042 if revlog == cl:
2042 if revlog == cl:
2043 c = cl.read(x)
2043 c = cl.read(x)
2044 changedfiles.update(c[3])
2044 changedfiles.update(c[3])
2045 mfs.setdefault(c[0], x)
2045 mfs.setdefault(c[0], x)
2046 count[0] += 1
2046 count[0] += 1
2047 progress(_bundling, count[0],
2047 progress(_bundling, count[0],
2048 unit=_changesets, total=count[1])
2048 unit=_changesets, total=count[1])
2049 return x
2049 return x
2050 elif revlog == mf:
2050 elif revlog == mf:
2051 clnode = mfs[x]
2051 clnode = mfs[x]
2052 mdata = mf.readfast(x)
2052 mdata = mf.readfast(x)
2053 for f, n in mdata.iteritems():
2053 for f, n in mdata.iteritems():
2054 if f in changedfiles:
2054 if f in changedfiles:
2055 fnodes[f].setdefault(n, clnode)
2055 fnodes[f].setdefault(n, clnode)
2056 count[0] += 1
2056 count[0] += 1
2057 progress(_bundling, count[0],
2057 progress(_bundling, count[0],
2058 unit=_manifests, total=count[1])
2058 unit=_manifests, total=count[1])
2059 return clnode
2059 return clnode
2060 else:
2060 else:
2061 progress(_bundling, count[0], item=fstate[0],
2061 progress(_bundling, count[0], item=fstate[0],
2062 unit=_files, total=count[1])
2062 unit=_files, total=count[1])
2063 return fstate[1][x]
2063 return fstate[1][x]
2064
2064
2065 bundler = changegroup.bundle10(lookup)
2065 bundler = changegroup.bundle10(lookup)
2066 reorder = self.ui.config('bundle', 'reorder', 'auto')
2066 reorder = self.ui.config('bundle', 'reorder', 'auto')
2067 if reorder == 'auto':
2067 if reorder == 'auto':
2068 reorder = None
2068 reorder = None
2069 else:
2069 else:
2070 reorder = util.parsebool(reorder)
2070 reorder = util.parsebool(reorder)
2071
2071
2072 def gengroup():
2072 def gengroup():
2073 # Create a changenode group generator that will call our functions
2073 # Create a changenode group generator that will call our functions
2074 # back to lookup the owning changenode and collect information.
2074 # back to lookup the owning changenode and collect information.
2075 count[:] = [0, len(csets)]
2075 count[:] = [0, len(csets)]
2076 for chunk in cl.group(csets, bundler, reorder=reorder):
2076 for chunk in cl.group(csets, bundler, reorder=reorder):
2077 yield chunk
2077 yield chunk
2078 progress(_bundling, None)
2078 progress(_bundling, None)
2079
2079
2080 # Create a generator for the manifestnodes that calls our lookup
2080 # Create a generator for the manifestnodes that calls our lookup
2081 # and data collection functions back.
2081 # and data collection functions back.
2082 for f in changedfiles:
2082 for f in changedfiles:
2083 fnodes[f] = {}
2083 fnodes[f] = {}
2084 count[:] = [0, len(mfs)]
2084 count[:] = [0, len(mfs)]
2085 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2085 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2086 yield chunk
2086 yield chunk
2087 progress(_bundling, None)
2087 progress(_bundling, None)
2088
2088
2089 mfs.clear()
2089 mfs.clear()
2090
2090
2091 # Go through all our files in order sorted by name.
2091 # Go through all our files in order sorted by name.
2092 count[:] = [0, len(changedfiles)]
2092 count[:] = [0, len(changedfiles)]
2093 for fname in sorted(changedfiles):
2093 for fname in sorted(changedfiles):
2094 filerevlog = self.file(fname)
2094 filerevlog = self.file(fname)
2095 if not len(filerevlog):
2095 if not len(filerevlog):
2096 raise util.Abort(_("empty or missing revlog for %s")
2096 raise util.Abort(_("empty or missing revlog for %s")
2097 % fname)
2097 % fname)
2098 fstate[0] = fname
2098 fstate[0] = fname
2099 fstate[1] = fnodes.pop(fname, {})
2099 fstate[1] = fnodes.pop(fname, {})
2100
2100
2101 nodelist = prune(filerevlog, fstate[1])
2101 nodelist = prune(filerevlog, fstate[1])
2102 if nodelist:
2102 if nodelist:
2103 count[0] += 1
2103 count[0] += 1
2104 yield bundler.fileheader(fname)
2104 yield bundler.fileheader(fname)
2105 for chunk in filerevlog.group(nodelist, bundler, reorder):
2105 for chunk in filerevlog.group(nodelist, bundler, reorder):
2106 yield chunk
2106 yield chunk
2107
2107
2108 # Signal that no more groups are left.
2108 # Signal that no more groups are left.
2109 yield bundler.close()
2109 yield bundler.close()
2110 progress(_bundling, None)
2110 progress(_bundling, None)
2111
2111
2112 if csets:
2112 if csets:
2113 self.hook('outgoing', node=hex(csets[0]), source=source)
2113 self.hook('outgoing', node=hex(csets[0]), source=source)
2114
2114
2115 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2115 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2116
2116
2117 def changegroup(self, basenodes, source):
2117 def changegroup(self, basenodes, source):
2118 # to avoid a race we use changegroupsubset() (issue1320)
2118 # to avoid a race we use changegroupsubset() (issue1320)
2119 return self.changegroupsubset(basenodes, self.heads(), source)
2119 return self.changegroupsubset(basenodes, self.heads(), source)
2120
2120
2121 @unfilteredmethod
2121 @unfilteredmethod
2122 def _changegroup(self, nodes, source):
2122 def _changegroup(self, nodes, source):
2123 """Compute the changegroup of all nodes that we have that a recipient
2123 """Compute the changegroup of all nodes that we have that a recipient
2124 doesn't. Return a chunkbuffer object whose read() method will return
2124 doesn't. Return a chunkbuffer object whose read() method will return
2125 successive changegroup chunks.
2125 successive changegroup chunks.
2126
2126
2127 This is much easier than the previous function as we can assume that
2127 This is much easier than the previous function as we can assume that
2128 the recipient has any changenode we aren't sending them.
2128 the recipient has any changenode we aren't sending them.
2129
2129
2130 nodes is the set of nodes to send"""
2130 nodes is the set of nodes to send"""
2131
2131
2132 cl = self.changelog
2132 cl = self.changelog
2133 mf = self.manifest
2133 mf = self.manifest
2134 mfs = {}
2134 mfs = {}
2135 changedfiles = set()
2135 changedfiles = set()
2136 fstate = ['']
2136 fstate = ['']
2137 count = [0, 0]
2137 count = [0, 0]
2138
2138
2139 self.hook('preoutgoing', throw=True, source=source)
2139 self.hook('preoutgoing', throw=True, source=source)
2140 self.changegroupinfo(nodes, source)
2140 self.changegroupinfo(nodes, source)
2141
2141
2142 revset = set([cl.rev(n) for n in nodes])
2142 revset = set([cl.rev(n) for n in nodes])
2143
2143
2144 def gennodelst(log):
2144 def gennodelst(log):
2145 ln, llr = log.node, log.linkrev
2145 ln, llr = log.node, log.linkrev
2146 return [ln(r) for r in log if llr(r) in revset]
2146 return [ln(r) for r in log if llr(r) in revset]
2147
2147
2148 progress = self.ui.progress
2148 progress = self.ui.progress
2149 _bundling = _('bundling')
2149 _bundling = _('bundling')
2150 _changesets = _('changesets')
2150 _changesets = _('changesets')
2151 _manifests = _('manifests')
2151 _manifests = _('manifests')
2152 _files = _('files')
2152 _files = _('files')
2153
2153
2154 def lookup(revlog, x):
2154 def lookup(revlog, x):
2155 if revlog == cl:
2155 if revlog == cl:
2156 c = cl.read(x)
2156 c = cl.read(x)
2157 changedfiles.update(c[3])
2157 changedfiles.update(c[3])
2158 mfs.setdefault(c[0], x)
2158 mfs.setdefault(c[0], x)
2159 count[0] += 1
2159 count[0] += 1
2160 progress(_bundling, count[0],
2160 progress(_bundling, count[0],
2161 unit=_changesets, total=count[1])
2161 unit=_changesets, total=count[1])
2162 return x
2162 return x
2163 elif revlog == mf:
2163 elif revlog == mf:
2164 count[0] += 1
2164 count[0] += 1
2165 progress(_bundling, count[0],
2165 progress(_bundling, count[0],
2166 unit=_manifests, total=count[1])
2166 unit=_manifests, total=count[1])
2167 return cl.node(revlog.linkrev(revlog.rev(x)))
2167 return cl.node(revlog.linkrev(revlog.rev(x)))
2168 else:
2168 else:
2169 progress(_bundling, count[0], item=fstate[0],
2169 progress(_bundling, count[0], item=fstate[0],
2170 total=count[1], unit=_files)
2170 total=count[1], unit=_files)
2171 return cl.node(revlog.linkrev(revlog.rev(x)))
2171 return cl.node(revlog.linkrev(revlog.rev(x)))
2172
2172
2173 bundler = changegroup.bundle10(lookup)
2173 bundler = changegroup.bundle10(lookup)
2174 reorder = self.ui.config('bundle', 'reorder', 'auto')
2174 reorder = self.ui.config('bundle', 'reorder', 'auto')
2175 if reorder == 'auto':
2175 if reorder == 'auto':
2176 reorder = None
2176 reorder = None
2177 else:
2177 else:
2178 reorder = util.parsebool(reorder)
2178 reorder = util.parsebool(reorder)
2179
2179
2180 def gengroup():
2180 def gengroup():
2181 '''yield a sequence of changegroup chunks (strings)'''
2181 '''yield a sequence of changegroup chunks (strings)'''
2182 # construct a list of all changed files
2182 # construct a list of all changed files
2183
2183
2184 count[:] = [0, len(nodes)]
2184 count[:] = [0, len(nodes)]
2185 for chunk in cl.group(nodes, bundler, reorder=reorder):
2185 for chunk in cl.group(nodes, bundler, reorder=reorder):
2186 yield chunk
2186 yield chunk
2187 progress(_bundling, None)
2187 progress(_bundling, None)
2188
2188
2189 count[:] = [0, len(mfs)]
2189 count[:] = [0, len(mfs)]
2190 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2190 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2191 yield chunk
2191 yield chunk
2192 progress(_bundling, None)
2192 progress(_bundling, None)
2193
2193
2194 count[:] = [0, len(changedfiles)]
2194 count[:] = [0, len(changedfiles)]
2195 for fname in sorted(changedfiles):
2195 for fname in sorted(changedfiles):
2196 filerevlog = self.file(fname)
2196 filerevlog = self.file(fname)
2197 if not len(filerevlog):
2197 if not len(filerevlog):
2198 raise util.Abort(_("empty or missing revlog for %s")
2198 raise util.Abort(_("empty or missing revlog for %s")
2199 % fname)
2199 % fname)
2200 fstate[0] = fname
2200 fstate[0] = fname
2201 nodelist = gennodelst(filerevlog)
2201 nodelist = gennodelst(filerevlog)
2202 if nodelist:
2202 if nodelist:
2203 count[0] += 1
2203 count[0] += 1
2204 yield bundler.fileheader(fname)
2204 yield bundler.fileheader(fname)
2205 for chunk in filerevlog.group(nodelist, bundler, reorder):
2205 for chunk in filerevlog.group(nodelist, bundler, reorder):
2206 yield chunk
2206 yield chunk
2207 yield bundler.close()
2207 yield bundler.close()
2208 progress(_bundling, None)
2208 progress(_bundling, None)
2209
2209
2210 if nodes:
2210 if nodes:
2211 self.hook('outgoing', node=hex(nodes[0]), source=source)
2211 self.hook('outgoing', node=hex(nodes[0]), source=source)
2212
2212
2213 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2213 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2214
2214
2215 @unfilteredmethod
2215 @unfilteredmethod
2216 def addchangegroup(self, source, srctype, url, emptyok=False):
2216 def addchangegroup(self, source, srctype, url, emptyok=False):
2217 """Add the changegroup returned by source.read() to this repo.
2217 """Add the changegroup returned by source.read() to this repo.
2218 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2218 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2219 the URL of the repo where this changegroup is coming from.
2219 the URL of the repo where this changegroup is coming from.
2220
2220
2221 Return an integer summarizing the change to this repo:
2221 Return an integer summarizing the change to this repo:
2222 - nothing changed or no source: 0
2222 - nothing changed or no source: 0
2223 - more heads than before: 1+added heads (2..n)
2223 - more heads than before: 1+added heads (2..n)
2224 - fewer heads than before: -1-removed heads (-2..-n)
2224 - fewer heads than before: -1-removed heads (-2..-n)
2225 - number of heads stays the same: 1
2225 - number of heads stays the same: 1
2226 """
2226 """
2227 def csmap(x):
2227 def csmap(x):
2228 self.ui.debug("add changeset %s\n" % short(x))
2228 self.ui.debug("add changeset %s\n" % short(x))
2229 return len(cl)
2229 return len(cl)
2230
2230
2231 def revmap(x):
2231 def revmap(x):
2232 return cl.rev(x)
2232 return cl.rev(x)
2233
2233
2234 if not source:
2234 if not source:
2235 return 0
2235 return 0
2236
2236
2237 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2237 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2238
2238
2239 changesets = files = revisions = 0
2239 changesets = files = revisions = 0
2240 efiles = set()
2240 efiles = set()
2241
2241
2242 # write changelog data to temp files so concurrent readers will not see
2242 # write changelog data to temp files so concurrent readers will not see
2243 # inconsistent view
2243 # inconsistent view
2244 cl = self.changelog
2244 cl = self.changelog
2245 cl.delayupdate()
2245 cl.delayupdate()
2246 oldheads = cl.heads()
2246 oldheads = cl.heads()
2247
2247
2248 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2248 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2249 try:
2249 try:
2250 trp = weakref.proxy(tr)
2250 trp = weakref.proxy(tr)
2251 # pull off the changeset group
2251 # pull off the changeset group
2252 self.ui.status(_("adding changesets\n"))
2252 self.ui.status(_("adding changesets\n"))
2253 clstart = len(cl)
2253 clstart = len(cl)
2254 class prog(object):
2254 class prog(object):
2255 step = _('changesets')
2255 step = _('changesets')
2256 count = 1
2256 count = 1
2257 ui = self.ui
2257 ui = self.ui
2258 total = None
2258 total = None
2259 def __call__(self):
2259 def __call__(self):
2260 self.ui.progress(self.step, self.count, unit=_('chunks'),
2260 self.ui.progress(self.step, self.count, unit=_('chunks'),
2261 total=self.total)
2261 total=self.total)
2262 self.count += 1
2262 self.count += 1
2263 pr = prog()
2263 pr = prog()
2264 source.callback = pr
2264 source.callback = pr
2265
2265
2266 source.changelogheader()
2266 source.changelogheader()
2267 srccontent = cl.addgroup(source, csmap, trp)
2267 srccontent = cl.addgroup(source, csmap, trp)
2268 if not (srccontent or emptyok):
2268 if not (srccontent or emptyok):
2269 raise util.Abort(_("received changelog group is empty"))
2269 raise util.Abort(_("received changelog group is empty"))
2270 clend = len(cl)
2270 clend = len(cl)
2271 changesets = clend - clstart
2271 changesets = clend - clstart
2272 for c in xrange(clstart, clend):
2272 for c in xrange(clstart, clend):
2273 efiles.update(self[c].files())
2273 efiles.update(self[c].files())
2274 efiles = len(efiles)
2274 efiles = len(efiles)
2275 self.ui.progress(_('changesets'), None)
2275 self.ui.progress(_('changesets'), None)
2276
2276
2277 # pull off the manifest group
2277 # pull off the manifest group
2278 self.ui.status(_("adding manifests\n"))
2278 self.ui.status(_("adding manifests\n"))
2279 pr.step = _('manifests')
2279 pr.step = _('manifests')
2280 pr.count = 1
2280 pr.count = 1
2281 pr.total = changesets # manifests <= changesets
2281 pr.total = changesets # manifests <= changesets
2282 # no need to check for empty manifest group here:
2282 # no need to check for empty manifest group here:
2283 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2283 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2284 # no new manifest will be created and the manifest group will
2284 # no new manifest will be created and the manifest group will
2285 # be empty during the pull
2285 # be empty during the pull
2286 source.manifestheader()
2286 source.manifestheader()
2287 self.manifest.addgroup(source, revmap, trp)
2287 self.manifest.addgroup(source, revmap, trp)
2288 self.ui.progress(_('manifests'), None)
2288 self.ui.progress(_('manifests'), None)
2289
2289
2290 needfiles = {}
2290 needfiles = {}
2291 if self.ui.configbool('server', 'validate', default=False):
2291 if self.ui.configbool('server', 'validate', default=False):
2292 # validate incoming csets have their manifests
2292 # validate incoming csets have their manifests
2293 for cset in xrange(clstart, clend):
2293 for cset in xrange(clstart, clend):
2294 mfest = self.changelog.read(self.changelog.node(cset))[0]
2294 mfest = self.changelog.read(self.changelog.node(cset))[0]
2295 mfest = self.manifest.readdelta(mfest)
2295 mfest = self.manifest.readdelta(mfest)
2296 # store file nodes we must see
2296 # store file nodes we must see
2297 for f, n in mfest.iteritems():
2297 for f, n in mfest.iteritems():
2298 needfiles.setdefault(f, set()).add(n)
2298 needfiles.setdefault(f, set()).add(n)
2299
2299
2300 # process the files
2300 # process the files
2301 self.ui.status(_("adding file changes\n"))
2301 self.ui.status(_("adding file changes\n"))
2302 pr.step = _('files')
2302 pr.step = _('files')
2303 pr.count = 1
2303 pr.count = 1
2304 pr.total = efiles
2304 pr.total = efiles
2305 source.callback = None
2305 source.callback = None
2306
2306
2307 while True:
2307 while True:
2308 chunkdata = source.filelogheader()
2308 chunkdata = source.filelogheader()
2309 if not chunkdata:
2309 if not chunkdata:
2310 break
2310 break
2311 f = chunkdata["filename"]
2311 f = chunkdata["filename"]
2312 self.ui.debug("adding %s revisions\n" % f)
2312 self.ui.debug("adding %s revisions\n" % f)
2313 pr()
2313 pr()
2314 fl = self.file(f)
2314 fl = self.file(f)
2315 o = len(fl)
2315 o = len(fl)
2316 if not fl.addgroup(source, revmap, trp):
2316 if not fl.addgroup(source, revmap, trp):
2317 raise util.Abort(_("received file revlog group is empty"))
2317 raise util.Abort(_("received file revlog group is empty"))
2318 revisions += len(fl) - o
2318 revisions += len(fl) - o
2319 files += 1
2319 files += 1
2320 if f in needfiles:
2320 if f in needfiles:
2321 needs = needfiles[f]
2321 needs = needfiles[f]
2322 for new in xrange(o, len(fl)):
2322 for new in xrange(o, len(fl)):
2323 n = fl.node(new)
2323 n = fl.node(new)
2324 if n in needs:
2324 if n in needs:
2325 needs.remove(n)
2325 needs.remove(n)
2326 else:
2326 else:
2327 raise util.Abort(
2327 raise util.Abort(
2328 _("received spurious file revlog entry"))
2328 _("received spurious file revlog entry"))
2329 if not needs:
2329 if not needs:
2330 del needfiles[f]
2330 del needfiles[f]
2331 self.ui.progress(_('files'), None)
2331 self.ui.progress(_('files'), None)
2332
2332
2333 for f, needs in needfiles.iteritems():
2333 for f, needs in needfiles.iteritems():
2334 fl = self.file(f)
2334 fl = self.file(f)
2335 for n in needs:
2335 for n in needs:
2336 try:
2336 try:
2337 fl.rev(n)
2337 fl.rev(n)
2338 except error.LookupError:
2338 except error.LookupError:
2339 raise util.Abort(
2339 raise util.Abort(
2340 _('missing file data for %s:%s - run hg verify') %
2340 _('missing file data for %s:%s - run hg verify') %
2341 (f, hex(n)))
2341 (f, hex(n)))
2342
2342
2343 dh = 0
2343 dh = 0
2344 if oldheads:
2344 if oldheads:
2345 heads = cl.heads()
2345 heads = cl.heads()
2346 dh = len(heads) - len(oldheads)
2346 dh = len(heads) - len(oldheads)
2347 for h in heads:
2347 for h in heads:
2348 if h not in oldheads and self[h].closesbranch():
2348 if h not in oldheads and self[h].closesbranch():
2349 dh -= 1
2349 dh -= 1
2350 htext = ""
2350 htext = ""
2351 if dh:
2351 if dh:
2352 htext = _(" (%+d heads)") % dh
2352 htext = _(" (%+d heads)") % dh
2353
2353
2354 self.ui.status(_("added %d changesets"
2354 self.ui.status(_("added %d changesets"
2355 " with %d changes to %d files%s\n")
2355 " with %d changes to %d files%s\n")
2356 % (changesets, revisions, files, htext))
2356 % (changesets, revisions, files, htext))
2357 self.invalidatevolatilesets()
2357 self.invalidatevolatilesets()
2358
2358
2359 if changesets > 0:
2359 if changesets > 0:
2360 p = lambda: cl.writepending() and self.root or ""
2360 p = lambda: cl.writepending() and self.root or ""
2361 self.hook('pretxnchangegroup', throw=True,
2361 self.hook('pretxnchangegroup', throw=True,
2362 node=hex(cl.node(clstart)), source=srctype,
2362 node=hex(cl.node(clstart)), source=srctype,
2363 url=url, pending=p)
2363 url=url, pending=p)
2364
2364
2365 added = [cl.node(r) for r in xrange(clstart, clend)]
2365 added = [cl.node(r) for r in xrange(clstart, clend)]
2366 publishing = self.ui.configbool('phases', 'publish', True)
2366 publishing = self.ui.configbool('phases', 'publish', True)
2367 if srctype == 'push':
2367 if srctype == 'push':
2368 # Old server can not push the boundary themself.
2368 # Old server can not push the boundary themself.
2369 # New server won't push the boundary if changeset already
2369 # New server won't push the boundary if changeset already
2370 # existed locally as secrete
2370 # existed locally as secrete
2371 #
2371 #
2372 # We should not use added here but the list of all change in
2372 # We should not use added here but the list of all change in
2373 # the bundle
2373 # the bundle
2374 if publishing:
2374 if publishing:
2375 phases.advanceboundary(self, phases.public, srccontent)
2375 phases.advanceboundary(self, phases.public, srccontent)
2376 else:
2376 else:
2377 phases.advanceboundary(self, phases.draft, srccontent)
2377 phases.advanceboundary(self, phases.draft, srccontent)
2378 phases.retractboundary(self, phases.draft, added)
2378 phases.retractboundary(self, phases.draft, added)
2379 elif srctype != 'strip':
2379 elif srctype != 'strip':
2380 # publishing only alter behavior during push
2380 # publishing only alter behavior during push
2381 #
2381 #
2382 # strip should not touch boundary at all
2382 # strip should not touch boundary at all
2383 phases.retractboundary(self, phases.draft, added)
2383 phases.retractboundary(self, phases.draft, added)
2384
2384
2385 # make changelog see real files again
2385 # make changelog see real files again
2386 cl.finalize(trp)
2386 cl.finalize(trp)
2387
2387
2388 tr.close()
2388 tr.close()
2389
2389
2390 if changesets > 0:
2390 if changesets > 0:
2391 if srctype != 'strip':
2391 if srctype != 'strip':
2392 # During strip, branchcache is invalid but coming call to
2392 # During strip, branchcache is invalid but coming call to
2393 # `destroyed` will repair it.
2393 # `destroyed` will repair it.
2394 # In other case we can safely update cache on disk.
2394 # In other case we can safely update cache on disk.
2395 branchmap.updatecache(self.filtered('served'))
2395 branchmap.updatecache(self.filtered('served'))
2396 def runhooks():
2396 def runhooks():
2397 # forcefully update the on-disk branch cache
2397 # forcefully update the on-disk branch cache
2398 self.ui.debug("updating the branch cache\n")
2398 self.ui.debug("updating the branch cache\n")
2399 self.hook("changegroup", node=hex(cl.node(clstart)),
2399 self.hook("changegroup", node=hex(cl.node(clstart)),
2400 source=srctype, url=url)
2400 source=srctype, url=url)
2401
2401
2402 for n in added:
2402 for n in added:
2403 self.hook("incoming", node=hex(n), source=srctype,
2403 self.hook("incoming", node=hex(n), source=srctype,
2404 url=url)
2404 url=url)
2405
2405
2406 newheads = [h for h in self.heads() if h not in oldheads]
2406 newheads = [h for h in self.heads() if h not in oldheads]
2407 self.ui.log("incoming",
2407 self.ui.log("incoming",
2408 "%s incoming changes - new heads: %s\n",
2408 "%s incoming changes - new heads: %s\n",
2409 len(added),
2409 len(added),
2410 ', '.join([hex(c[:6]) for c in newheads]))
2410 ', '.join([hex(c[:6]) for c in newheads]))
2411 self._afterlock(runhooks)
2411 self._afterlock(runhooks)
2412
2412
2413 finally:
2413 finally:
2414 tr.release()
2414 tr.release()
2415 # never return 0 here:
2415 # never return 0 here:
2416 if dh < 0:
2416 if dh < 0:
2417 return dh - 1
2417 return dh - 1
2418 else:
2418 else:
2419 return dh + 1
2419 return dh + 1
2420
2420
2421 def stream_in(self, remote, requirements):
2421 def stream_in(self, remote, requirements):
2422 lock = self.lock()
2422 lock = self.lock()
2423 try:
2423 try:
2424 # Save remote branchmap. We will use it later
2424 # Save remote branchmap. We will use it later
2425 # to speed up branchcache creation
2425 # to speed up branchcache creation
2426 rbranchmap = None
2426 rbranchmap = None
2427 if remote.capable("branchmap"):
2427 if remote.capable("branchmap"):
2428 rbranchmap = remote.branchmap()
2428 rbranchmap = remote.branchmap()
2429
2429
2430 fp = remote.stream_out()
2430 fp = remote.stream_out()
2431 l = fp.readline()
2431 l = fp.readline()
2432 try:
2432 try:
2433 resp = int(l)
2433 resp = int(l)
2434 except ValueError:
2434 except ValueError:
2435 raise error.ResponseError(
2435 raise error.ResponseError(
2436 _('unexpected response from remote server:'), l)
2436 _('unexpected response from remote server:'), l)
2437 if resp == 1:
2437 if resp == 1:
2438 raise util.Abort(_('operation forbidden by server'))
2438 raise util.Abort(_('operation forbidden by server'))
2439 elif resp == 2:
2439 elif resp == 2:
2440 raise util.Abort(_('locking the remote repository failed'))
2440 raise util.Abort(_('locking the remote repository failed'))
2441 elif resp != 0:
2441 elif resp != 0:
2442 raise util.Abort(_('the server sent an unknown error code'))
2442 raise util.Abort(_('the server sent an unknown error code'))
2443 self.ui.status(_('streaming all changes\n'))
2443 self.ui.status(_('streaming all changes\n'))
2444 l = fp.readline()
2444 l = fp.readline()
2445 try:
2445 try:
2446 total_files, total_bytes = map(int, l.split(' ', 1))
2446 total_files, total_bytes = map(int, l.split(' ', 1))
2447 except (ValueError, TypeError):
2447 except (ValueError, TypeError):
2448 raise error.ResponseError(
2448 raise error.ResponseError(
2449 _('unexpected response from remote server:'), l)
2449 _('unexpected response from remote server:'), l)
2450 self.ui.status(_('%d files to transfer, %s of data\n') %
2450 self.ui.status(_('%d files to transfer, %s of data\n') %
2451 (total_files, util.bytecount(total_bytes)))
2451 (total_files, util.bytecount(total_bytes)))
2452 handled_bytes = 0
2452 handled_bytes = 0
2453 self.ui.progress(_('clone'), 0, total=total_bytes)
2453 self.ui.progress(_('clone'), 0, total=total_bytes)
2454 start = time.time()
2454 start = time.time()
2455 for i in xrange(total_files):
2455 for i in xrange(total_files):
2456 # XXX doesn't support '\n' or '\r' in filenames
2456 # XXX doesn't support '\n' or '\r' in filenames
2457 l = fp.readline()
2457 l = fp.readline()
2458 try:
2458 try:
2459 name, size = l.split('\0', 1)
2459 name, size = l.split('\0', 1)
2460 size = int(size)
2460 size = int(size)
2461 except (ValueError, TypeError):
2461 except (ValueError, TypeError):
2462 raise error.ResponseError(
2462 raise error.ResponseError(
2463 _('unexpected response from remote server:'), l)
2463 _('unexpected response from remote server:'), l)
2464 if self.ui.debugflag:
2464 if self.ui.debugflag:
2465 self.ui.debug('adding %s (%s)\n' %
2465 self.ui.debug('adding %s (%s)\n' %
2466 (name, util.bytecount(size)))
2466 (name, util.bytecount(size)))
2467 # for backwards compat, name was partially encoded
2467 # for backwards compat, name was partially encoded
2468 ofp = self.sopener(store.decodedir(name), 'w')
2468 ofp = self.sopener(store.decodedir(name), 'w')
2469 for chunk in util.filechunkiter(fp, limit=size):
2469 for chunk in util.filechunkiter(fp, limit=size):
2470 handled_bytes += len(chunk)
2470 handled_bytes += len(chunk)
2471 self.ui.progress(_('clone'), handled_bytes,
2471 self.ui.progress(_('clone'), handled_bytes,
2472 total=total_bytes)
2472 total=total_bytes)
2473 ofp.write(chunk)
2473 ofp.write(chunk)
2474 ofp.close()
2474 ofp.close()
2475 elapsed = time.time() - start
2475 elapsed = time.time() - start
2476 if elapsed <= 0:
2476 if elapsed <= 0:
2477 elapsed = 0.001
2477 elapsed = 0.001
2478 self.ui.progress(_('clone'), None)
2478 self.ui.progress(_('clone'), None)
2479 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2479 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2480 (util.bytecount(total_bytes), elapsed,
2480 (util.bytecount(total_bytes), elapsed,
2481 util.bytecount(total_bytes / elapsed)))
2481 util.bytecount(total_bytes / elapsed)))
2482
2482
2483 # new requirements = old non-format requirements +
2483 # new requirements = old non-format requirements +
2484 # new format-related
2484 # new format-related
2485 # requirements from the streamed-in repository
2485 # requirements from the streamed-in repository
2486 requirements.update(set(self.requirements) - self.supportedformats)
2486 requirements.update(set(self.requirements) - self.supportedformats)
2487 self._applyrequirements(requirements)
2487 self._applyrequirements(requirements)
2488 self._writerequirements()
2488 self._writerequirements()
2489
2489
2490 if rbranchmap:
2490 if rbranchmap:
2491 rbheads = []
2491 rbheads = []
2492 for bheads in rbranchmap.itervalues():
2492 for bheads in rbranchmap.itervalues():
2493 rbheads.extend(bheads)
2493 rbheads.extend(bheads)
2494
2494
2495 if rbheads:
2495 if rbheads:
2496 rtiprev = max((int(self.changelog.rev(node))
2496 rtiprev = max((int(self.changelog.rev(node))
2497 for node in rbheads))
2497 for node in rbheads))
2498 cache = branchmap.branchcache(rbranchmap,
2498 cache = branchmap.branchcache(rbranchmap,
2499 self[rtiprev].node(),
2499 self[rtiprev].node(),
2500 rtiprev)
2500 rtiprev)
2501 # Try to stick it as low as possible
2501 # Try to stick it as low as possible
2502 # filter above served are unlikely to be fetch from a clone
2502 # filter above served are unlikely to be fetch from a clone
2503 for candidate in ('base', 'immutable', 'served'):
2503 for candidate in ('base', 'immutable', 'served'):
2504 rview = self.filtered(candidate)
2504 rview = self.filtered(candidate)
2505 if cache.validfor(rview):
2505 if cache.validfor(rview):
2506 self._branchcaches[candidate] = cache
2506 self._branchcaches[candidate] = cache
2507 cache.write(rview)
2507 cache.write(rview)
2508 break
2508 break
2509 self.invalidate()
2509 self.invalidate()
2510 return len(self.heads()) + 1
2510 return len(self.heads()) + 1
2511 finally:
2511 finally:
2512 lock.release()
2512 lock.release()
2513
2513
2514 def clone(self, remote, heads=[], stream=False):
2514 def clone(self, remote, heads=[], stream=False):
2515 '''clone remote repository.
2515 '''clone remote repository.
2516
2516
2517 keyword arguments:
2517 keyword arguments:
2518 heads: list of revs to clone (forces use of pull)
2518 heads: list of revs to clone (forces use of pull)
2519 stream: use streaming clone if possible'''
2519 stream: use streaming clone if possible'''
2520
2520
2521 # now, all clients that can request uncompressed clones can
2521 # now, all clients that can request uncompressed clones can
2522 # read repo formats supported by all servers that can serve
2522 # read repo formats supported by all servers that can serve
2523 # them.
2523 # them.
2524
2524
2525 # if revlog format changes, client will have to check version
2525 # if revlog format changes, client will have to check version
2526 # and format flags on "stream" capability, and use
2526 # and format flags on "stream" capability, and use
2527 # uncompressed only if compatible.
2527 # uncompressed only if compatible.
2528
2528
2529 if not stream:
2529 if not stream:
2530 # if the server explicitly prefers to stream (for fast LANs)
2530 # if the server explicitly prefers to stream (for fast LANs)
2531 stream = remote.capable('stream-preferred')
2531 stream = remote.capable('stream-preferred')
2532
2532
2533 if stream and not heads:
2533 if stream and not heads:
2534 # 'stream' means remote revlog format is revlogv1 only
2534 # 'stream' means remote revlog format is revlogv1 only
2535 if remote.capable('stream'):
2535 if remote.capable('stream'):
2536 return self.stream_in(remote, set(('revlogv1',)))
2536 return self.stream_in(remote, set(('revlogv1',)))
2537 # otherwise, 'streamreqs' contains the remote revlog format
2537 # otherwise, 'streamreqs' contains the remote revlog format
2538 streamreqs = remote.capable('streamreqs')
2538 streamreqs = remote.capable('streamreqs')
2539 if streamreqs:
2539 if streamreqs:
2540 streamreqs = set(streamreqs.split(','))
2540 streamreqs = set(streamreqs.split(','))
2541 # if we support it, stream in and adjust our requirements
2541 # if we support it, stream in and adjust our requirements
2542 if not streamreqs - self.supportedformats:
2542 if not streamreqs - self.supportedformats:
2543 return self.stream_in(remote, streamreqs)
2543 return self.stream_in(remote, streamreqs)
2544 return self.pull(remote, heads)
2544 return self.pull(remote, heads)
2545
2545
2546 def pushkey(self, namespace, key, old, new):
2546 def pushkey(self, namespace, key, old, new):
2547 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2547 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2548 old=old, new=new)
2548 old=old, new=new)
2549 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2549 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2550 ret = pushkey.push(self, namespace, key, old, new)
2550 ret = pushkey.push(self, namespace, key, old, new)
2551 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2551 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2552 ret=ret)
2552 ret=ret)
2553 return ret
2553 return ret
2554
2554
2555 def listkeys(self, namespace):
2555 def listkeys(self, namespace):
2556 self.hook('prelistkeys', throw=True, namespace=namespace)
2556 self.hook('prelistkeys', throw=True, namespace=namespace)
2557 self.ui.debug('listing keys for "%s"\n' % namespace)
2557 self.ui.debug('listing keys for "%s"\n' % namespace)
2558 values = pushkey.list(self, namespace)
2558 values = pushkey.list(self, namespace)
2559 self.hook('listkeys', namespace=namespace, values=values)
2559 self.hook('listkeys', namespace=namespace, values=values)
2560 return values
2560 return values
2561
2561
2562 def debugwireargs(self, one, two, three=None, four=None, five=None):
2562 def debugwireargs(self, one, two, three=None, four=None, five=None):
2563 '''used to test argument passing over the wire'''
2563 '''used to test argument passing over the wire'''
2564 return "%s %s %s %s %s" % (one, two, three, four, five)
2564 return "%s %s %s %s %s" % (one, two, three, four, five)
2565
2565
2566 def savecommitmessage(self, text):
2566 def savecommitmessage(self, text):
2567 fp = self.opener('last-message.txt', 'wb')
2567 fp = self.opener('last-message.txt', 'wb')
2568 try:
2568 try:
2569 fp.write(text)
2569 fp.write(text)
2570 finally:
2570 finally:
2571 fp.close()
2571 fp.close()
2572 return self.pathto(fp.name[len(self.root) + 1:])
2572 return self.pathto(fp.name[len(self.root) + 1:])
2573
2573
2574 # used to avoid circular references so destructors work
2574 # used to avoid circular references so destructors work
2575 def aftertrans(files):
2575 def aftertrans(files):
2576 renamefiles = [tuple(t) for t in files]
2576 renamefiles = [tuple(t) for t in files]
2577 def a():
2577 def a():
2578 for src, dest in renamefiles:
2578 for src, dest in renamefiles:
2579 try:
2579 try:
2580 util.rename(src, dest)
2580 util.rename(src, dest)
2581 except OSError: # journal file does not yet exist
2581 except OSError: # journal file does not yet exist
2582 pass
2582 pass
2583 return a
2583 return a
2584
2584
2585 def undoname(fn):
2585 def undoname(fn):
2586 base, name = os.path.split(fn)
2586 base, name = os.path.split(fn)
2587 assert name.startswith('journal')
2587 assert name.startswith('journal')
2588 return os.path.join(base, name.replace('journal', 'undo', 1))
2588 return os.path.join(base, name.replace('journal', 'undo', 1))
2589
2589
2590 def instance(ui, path, create):
2590 def instance(ui, path, create):
2591 return localrepository(ui, util.urllocalpath(path), create)
2591 return localrepository(ui, util.urllocalpath(path), create)
2592
2592
2593 def islocal(path):
2593 def islocal(path):
2594 return True
2594 return True
@@ -1,945 +1,948 b''
1 # scmutil.py - Mercurial core utility functions
1 # scmutil.py - Mercurial core utility functions
2 #
2 #
3 # Copyright Matt Mackall <mpm@selenic.com>
3 # Copyright Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7
7
8 from i18n import _
8 from i18n import _
9 from mercurial.node import nullrev
9 from mercurial.node import nullrev
10 import util, error, osutil, revset, similar, encoding, phases, parsers
10 import util, error, osutil, revset, similar, encoding, phases, parsers
11 import match as matchmod
11 import match as matchmod
12 import os, errno, re, stat, glob
12 import os, errno, re, stat, glob
13
13
14 if os.name == 'nt':
14 if os.name == 'nt':
15 import scmwindows as scmplatform
15 import scmwindows as scmplatform
16 else:
16 else:
17 import scmposix as scmplatform
17 import scmposix as scmplatform
18
18
19 systemrcpath = scmplatform.systemrcpath
19 systemrcpath = scmplatform.systemrcpath
20 userrcpath = scmplatform.userrcpath
20 userrcpath = scmplatform.userrcpath
21
21
22 def nochangesfound(ui, repo, excluded=None):
22 def nochangesfound(ui, repo, excluded=None):
23 '''Report no changes for push/pull, excluded is None or a list of
23 '''Report no changes for push/pull, excluded is None or a list of
24 nodes excluded from the push/pull.
24 nodes excluded from the push/pull.
25 '''
25 '''
26 secretlist = []
26 secretlist = []
27 if excluded:
27 if excluded:
28 for n in excluded:
28 for n in excluded:
29 if n not in repo:
29 if n not in repo:
30 # discovery should not have included the filtered revision,
30 # discovery should not have included the filtered revision,
31 # we have to explicitly exclude it until discovery is cleanup.
31 # we have to explicitly exclude it until discovery is cleanup.
32 continue
32 continue
33 ctx = repo[n]
33 ctx = repo[n]
34 if ctx.phase() >= phases.secret and not ctx.extinct():
34 if ctx.phase() >= phases.secret and not ctx.extinct():
35 secretlist.append(n)
35 secretlist.append(n)
36
36
37 if secretlist:
37 if secretlist:
38 ui.status(_("no changes found (ignored %d secret changesets)\n")
38 ui.status(_("no changes found (ignored %d secret changesets)\n")
39 % len(secretlist))
39 % len(secretlist))
40 else:
40 else:
41 ui.status(_("no changes found\n"))
41 ui.status(_("no changes found\n"))
42
42
43 def checknewlabel(repo, lbl, kind):
43 def checknewlabel(repo, lbl, kind):
44 if lbl in ['tip', '.', 'null']:
44 if lbl in ['tip', '.', 'null']:
45 raise util.Abort(_("the name '%s' is reserved") % lbl)
45 raise util.Abort(_("the name '%s' is reserved") % lbl)
46 for c in (':', '\0', '\n', '\r'):
46 for c in (':', '\0', '\n', '\r'):
47 if c in lbl:
47 if c in lbl:
48 raise util.Abort(_("%r cannot be used in a name") % c)
48 raise util.Abort(_("%r cannot be used in a name") % c)
49 try:
49 try:
50 int(lbl)
50 int(lbl)
51 raise util.Abort(_("a %s cannot have an integer as its name") % kind)
51 raise util.Abort(_("a %s cannot have an integer as its name") % kind)
52 except ValueError:
52 except ValueError:
53 pass
53 pass
54
54
55 def checkfilename(f):
55 def checkfilename(f):
56 '''Check that the filename f is an acceptable filename for a tracked file'''
56 '''Check that the filename f is an acceptable filename for a tracked file'''
57 if '\r' in f or '\n' in f:
57 if '\r' in f or '\n' in f:
58 raise util.Abort(_("'\\n' and '\\r' disallowed in filenames: %r") % f)
58 raise util.Abort(_("'\\n' and '\\r' disallowed in filenames: %r") % f)
59
59
60 def checkportable(ui, f):
60 def checkportable(ui, f):
61 '''Check if filename f is portable and warn or abort depending on config'''
61 '''Check if filename f is portable and warn or abort depending on config'''
62 checkfilename(f)
62 checkfilename(f)
63 abort, warn = checkportabilityalert(ui)
63 abort, warn = checkportabilityalert(ui)
64 if abort or warn:
64 if abort or warn:
65 msg = util.checkwinfilename(f)
65 msg = util.checkwinfilename(f)
66 if msg:
66 if msg:
67 msg = "%s: %r" % (msg, f)
67 msg = "%s: %r" % (msg, f)
68 if abort:
68 if abort:
69 raise util.Abort(msg)
69 raise util.Abort(msg)
70 ui.warn(_("warning: %s\n") % msg)
70 ui.warn(_("warning: %s\n") % msg)
71
71
72 def checkportabilityalert(ui):
72 def checkportabilityalert(ui):
73 '''check if the user's config requests nothing, a warning, or abort for
73 '''check if the user's config requests nothing, a warning, or abort for
74 non-portable filenames'''
74 non-portable filenames'''
75 val = ui.config('ui', 'portablefilenames', 'warn')
75 val = ui.config('ui', 'portablefilenames', 'warn')
76 lval = val.lower()
76 lval = val.lower()
77 bval = util.parsebool(val)
77 bval = util.parsebool(val)
78 abort = os.name == 'nt' or lval == 'abort'
78 abort = os.name == 'nt' or lval == 'abort'
79 warn = bval or lval == 'warn'
79 warn = bval or lval == 'warn'
80 if bval is None and not (warn or abort or lval == 'ignore'):
80 if bval is None and not (warn or abort or lval == 'ignore'):
81 raise error.ConfigError(
81 raise error.ConfigError(
82 _("ui.portablefilenames value is invalid ('%s')") % val)
82 _("ui.portablefilenames value is invalid ('%s')") % val)
83 return abort, warn
83 return abort, warn
84
84
85 class casecollisionauditor(object):
85 class casecollisionauditor(object):
86 def __init__(self, ui, abort, dirstate):
86 def __init__(self, ui, abort, dirstate):
87 self._ui = ui
87 self._ui = ui
88 self._abort = abort
88 self._abort = abort
89 allfiles = '\0'.join(dirstate._map)
89 allfiles = '\0'.join(dirstate._map)
90 self._loweredfiles = set(encoding.lower(allfiles).split('\0'))
90 self._loweredfiles = set(encoding.lower(allfiles).split('\0'))
91 self._dirstate = dirstate
91 self._dirstate = dirstate
92 # The purpose of _newfiles is so that we don't complain about
92 # The purpose of _newfiles is so that we don't complain about
93 # case collisions if someone were to call this object with the
93 # case collisions if someone were to call this object with the
94 # same filename twice.
94 # same filename twice.
95 self._newfiles = set()
95 self._newfiles = set()
96
96
97 def __call__(self, f):
97 def __call__(self, f):
98 fl = encoding.lower(f)
98 fl = encoding.lower(f)
99 if (fl in self._loweredfiles and f not in self._dirstate and
99 if (fl in self._loweredfiles and f not in self._dirstate and
100 f not in self._newfiles):
100 f not in self._newfiles):
101 msg = _('possible case-folding collision for %s') % f
101 msg = _('possible case-folding collision for %s') % f
102 if self._abort:
102 if self._abort:
103 raise util.Abort(msg)
103 raise util.Abort(msg)
104 self._ui.warn(_("warning: %s\n") % msg)
104 self._ui.warn(_("warning: %s\n") % msg)
105 self._loweredfiles.add(fl)
105 self._loweredfiles.add(fl)
106 self._newfiles.add(f)
106 self._newfiles.add(f)
107
107
108 class pathauditor(object):
108 class pathauditor(object):
109 '''ensure that a filesystem path contains no banned components.
109 '''ensure that a filesystem path contains no banned components.
110 the following properties of a path are checked:
110 the following properties of a path are checked:
111
111
112 - ends with a directory separator
112 - ends with a directory separator
113 - under top-level .hg
113 - under top-level .hg
114 - starts at the root of a windows drive
114 - starts at the root of a windows drive
115 - contains ".."
115 - contains ".."
116 - traverses a symlink (e.g. a/symlink_here/b)
116 - traverses a symlink (e.g. a/symlink_here/b)
117 - inside a nested repository (a callback can be used to approve
117 - inside a nested repository (a callback can be used to approve
118 some nested repositories, e.g., subrepositories)
118 some nested repositories, e.g., subrepositories)
119 '''
119 '''
120
120
121 def __init__(self, root, callback=None):
121 def __init__(self, root, callback=None):
122 self.audited = set()
122 self.audited = set()
123 self.auditeddir = set()
123 self.auditeddir = set()
124 self.root = root
124 self.root = root
125 self.callback = callback
125 self.callback = callback
126 if os.path.lexists(root) and not util.checkcase(root):
126 if os.path.lexists(root) and not util.checkcase(root):
127 self.normcase = util.normcase
127 self.normcase = util.normcase
128 else:
128 else:
129 self.normcase = lambda x: x
129 self.normcase = lambda x: x
130
130
131 def __call__(self, path):
131 def __call__(self, path):
132 '''Check the relative path.
132 '''Check the relative path.
133 path may contain a pattern (e.g. foodir/**.txt)'''
133 path may contain a pattern (e.g. foodir/**.txt)'''
134
134
135 path = util.localpath(path)
135 path = util.localpath(path)
136 normpath = self.normcase(path)
136 normpath = self.normcase(path)
137 if normpath in self.audited:
137 if normpath in self.audited:
138 return
138 return
139 # AIX ignores "/" at end of path, others raise EISDIR.
139 # AIX ignores "/" at end of path, others raise EISDIR.
140 if util.endswithsep(path):
140 if util.endswithsep(path):
141 raise util.Abort(_("path ends in directory separator: %s") % path)
141 raise util.Abort(_("path ends in directory separator: %s") % path)
142 parts = util.splitpath(path)
142 parts = util.splitpath(path)
143 if (os.path.splitdrive(path)[0]
143 if (os.path.splitdrive(path)[0]
144 or parts[0].lower() in ('.hg', '.hg.', '')
144 or parts[0].lower() in ('.hg', '.hg.', '')
145 or os.pardir in parts):
145 or os.pardir in parts):
146 raise util.Abort(_("path contains illegal component: %s") % path)
146 raise util.Abort(_("path contains illegal component: %s") % path)
147 if '.hg' in path.lower():
147 if '.hg' in path.lower():
148 lparts = [p.lower() for p in parts]
148 lparts = [p.lower() for p in parts]
149 for p in '.hg', '.hg.':
149 for p in '.hg', '.hg.':
150 if p in lparts[1:]:
150 if p in lparts[1:]:
151 pos = lparts.index(p)
151 pos = lparts.index(p)
152 base = os.path.join(*parts[:pos])
152 base = os.path.join(*parts[:pos])
153 raise util.Abort(_("path '%s' is inside nested repo %r")
153 raise util.Abort(_("path '%s' is inside nested repo %r")
154 % (path, base))
154 % (path, base))
155
155
156 normparts = util.splitpath(normpath)
156 normparts = util.splitpath(normpath)
157 assert len(parts) == len(normparts)
157 assert len(parts) == len(normparts)
158
158
159 parts.pop()
159 parts.pop()
160 normparts.pop()
160 normparts.pop()
161 prefixes = []
161 prefixes = []
162 while parts:
162 while parts:
163 prefix = os.sep.join(parts)
163 prefix = os.sep.join(parts)
164 normprefix = os.sep.join(normparts)
164 normprefix = os.sep.join(normparts)
165 if normprefix in self.auditeddir:
165 if normprefix in self.auditeddir:
166 break
166 break
167 curpath = os.path.join(self.root, prefix)
167 curpath = os.path.join(self.root, prefix)
168 try:
168 try:
169 st = os.lstat(curpath)
169 st = os.lstat(curpath)
170 except OSError, err:
170 except OSError, err:
171 # EINVAL can be raised as invalid path syntax under win32.
171 # EINVAL can be raised as invalid path syntax under win32.
172 # They must be ignored for patterns can be checked too.
172 # They must be ignored for patterns can be checked too.
173 if err.errno not in (errno.ENOENT, errno.ENOTDIR, errno.EINVAL):
173 if err.errno not in (errno.ENOENT, errno.ENOTDIR, errno.EINVAL):
174 raise
174 raise
175 else:
175 else:
176 if stat.S_ISLNK(st.st_mode):
176 if stat.S_ISLNK(st.st_mode):
177 raise util.Abort(
177 raise util.Abort(
178 _('path %r traverses symbolic link %r')
178 _('path %r traverses symbolic link %r')
179 % (path, prefix))
179 % (path, prefix))
180 elif (stat.S_ISDIR(st.st_mode) and
180 elif (stat.S_ISDIR(st.st_mode) and
181 os.path.isdir(os.path.join(curpath, '.hg'))):
181 os.path.isdir(os.path.join(curpath, '.hg'))):
182 if not self.callback or not self.callback(curpath):
182 if not self.callback or not self.callback(curpath):
183 raise util.Abort(_("path '%s' is inside nested "
183 raise util.Abort(_("path '%s' is inside nested "
184 "repo %r")
184 "repo %r")
185 % (path, prefix))
185 % (path, prefix))
186 prefixes.append(normprefix)
186 prefixes.append(normprefix)
187 parts.pop()
187 parts.pop()
188 normparts.pop()
188 normparts.pop()
189
189
190 self.audited.add(normpath)
190 self.audited.add(normpath)
191 # only add prefixes to the cache after checking everything: we don't
191 # only add prefixes to the cache after checking everything: we don't
192 # want to add "foo/bar/baz" before checking if there's a "foo/.hg"
192 # want to add "foo/bar/baz" before checking if there's a "foo/.hg"
193 self.auditeddir.update(prefixes)
193 self.auditeddir.update(prefixes)
194
194
195 def check(self, path):
195 def check(self, path):
196 try:
196 try:
197 self(path)
197 self(path)
198 return True
198 return True
199 except (OSError, util.Abort):
199 except (OSError, util.Abort):
200 return False
200 return False
201
201
202 class abstractvfs(object):
202 class abstractvfs(object):
203 """Abstract base class; cannot be instantiated"""
203 """Abstract base class; cannot be instantiated"""
204
204
205 def __init__(self, *args, **kwargs):
205 def __init__(self, *args, **kwargs):
206 '''Prevent instantiation; don't call this from subclasses.'''
206 '''Prevent instantiation; don't call this from subclasses.'''
207 raise NotImplementedError('attempted instantiating ' + str(type(self)))
207 raise NotImplementedError('attempted instantiating ' + str(type(self)))
208
208
209 def tryread(self, path):
209 def tryread(self, path):
210 '''gracefully return an empty string for missing files'''
210 '''gracefully return an empty string for missing files'''
211 try:
211 try:
212 return self.read(path)
212 return self.read(path)
213 except IOError, inst:
213 except IOError, inst:
214 if inst.errno != errno.ENOENT:
214 if inst.errno != errno.ENOENT:
215 raise
215 raise
216 return ""
216 return ""
217
217
218 def read(self, path):
218 def read(self, path):
219 fp = self(path, 'rb')
219 fp = self(path, 'rb')
220 try:
220 try:
221 return fp.read()
221 return fp.read()
222 finally:
222 finally:
223 fp.close()
223 fp.close()
224
224
225 def write(self, path, data):
225 def write(self, path, data):
226 fp = self(path, 'wb')
226 fp = self(path, 'wb')
227 try:
227 try:
228 return fp.write(data)
228 return fp.write(data)
229 finally:
229 finally:
230 fp.close()
230 fp.close()
231
231
232 def append(self, path, data):
232 def append(self, path, data):
233 fp = self(path, 'ab')
233 fp = self(path, 'ab')
234 try:
234 try:
235 return fp.write(data)
235 return fp.write(data)
236 finally:
236 finally:
237 fp.close()
237 fp.close()
238
238
239 def exists(self, path=None):
239 def exists(self, path=None):
240 return os.path.exists(self.join(path))
240 return os.path.exists(self.join(path))
241
241
242 def isdir(self, path=None):
242 def isdir(self, path=None):
243 return os.path.isdir(self.join(path))
243 return os.path.isdir(self.join(path))
244
244
245 def islink(self, path=None):
245 def islink(self, path=None):
246 return os.path.islink(self.join(path))
246 return os.path.islink(self.join(path))
247
247
248 def makedir(self, path=None, notindexed=True):
248 def makedir(self, path=None, notindexed=True):
249 return util.makedir(self.join(path), notindexed)
249 return util.makedir(self.join(path), notindexed)
250
250
251 def makedirs(self, path=None, mode=None):
251 def makedirs(self, path=None, mode=None):
252 return util.makedirs(self.join(path), mode)
252 return util.makedirs(self.join(path), mode)
253
253
254 def mkdir(self, path=None):
254 def mkdir(self, path=None):
255 return os.mkdir(self.join(path))
255 return os.mkdir(self.join(path))
256
256
257 def readdir(self, path=None, stat=None, skip=None):
257 def readdir(self, path=None, stat=None, skip=None):
258 return osutil.listdir(self.join(path), stat, skip)
258 return osutil.listdir(self.join(path), stat, skip)
259
259
260 def rename(self, src, dst):
260 def rename(self, src, dst):
261 return util.rename(self.join(src), self.join(dst))
261 return util.rename(self.join(src), self.join(dst))
262
262
263 def readlink(self, path):
264 return os.readlink(self.join(path))
265
263 def stat(self, path=None):
266 def stat(self, path=None):
264 return os.stat(self.join(path))
267 return os.stat(self.join(path))
265
268
266 class vfs(abstractvfs):
269 class vfs(abstractvfs):
267 '''Operate files relative to a base directory
270 '''Operate files relative to a base directory
268
271
269 This class is used to hide the details of COW semantics and
272 This class is used to hide the details of COW semantics and
270 remote file access from higher level code.
273 remote file access from higher level code.
271 '''
274 '''
272 def __init__(self, base, audit=True, expandpath=False, realpath=False):
275 def __init__(self, base, audit=True, expandpath=False, realpath=False):
273 if expandpath:
276 if expandpath:
274 base = util.expandpath(base)
277 base = util.expandpath(base)
275 if realpath:
278 if realpath:
276 base = os.path.realpath(base)
279 base = os.path.realpath(base)
277 self.base = base
280 self.base = base
278 self._setmustaudit(audit)
281 self._setmustaudit(audit)
279 self.createmode = None
282 self.createmode = None
280 self._trustnlink = None
283 self._trustnlink = None
281
284
282 def _getmustaudit(self):
285 def _getmustaudit(self):
283 return self._audit
286 return self._audit
284
287
285 def _setmustaudit(self, onoff):
288 def _setmustaudit(self, onoff):
286 self._audit = onoff
289 self._audit = onoff
287 if onoff:
290 if onoff:
288 self.audit = pathauditor(self.base)
291 self.audit = pathauditor(self.base)
289 else:
292 else:
290 self.audit = util.always
293 self.audit = util.always
291
294
292 mustaudit = property(_getmustaudit, _setmustaudit)
295 mustaudit = property(_getmustaudit, _setmustaudit)
293
296
294 @util.propertycache
297 @util.propertycache
295 def _cansymlink(self):
298 def _cansymlink(self):
296 return util.checklink(self.base)
299 return util.checklink(self.base)
297
300
298 @util.propertycache
301 @util.propertycache
299 def _chmod(self):
302 def _chmod(self):
300 return util.checkexec(self.base)
303 return util.checkexec(self.base)
301
304
302 def _fixfilemode(self, name):
305 def _fixfilemode(self, name):
303 if self.createmode is None or not self._chmod:
306 if self.createmode is None or not self._chmod:
304 return
307 return
305 os.chmod(name, self.createmode & 0666)
308 os.chmod(name, self.createmode & 0666)
306
309
307 def __call__(self, path, mode="r", text=False, atomictemp=False):
310 def __call__(self, path, mode="r", text=False, atomictemp=False):
308 if self._audit:
311 if self._audit:
309 r = util.checkosfilename(path)
312 r = util.checkosfilename(path)
310 if r:
313 if r:
311 raise util.Abort("%s: %r" % (r, path))
314 raise util.Abort("%s: %r" % (r, path))
312 self.audit(path)
315 self.audit(path)
313 f = self.join(path)
316 f = self.join(path)
314
317
315 if not text and "b" not in mode:
318 if not text and "b" not in mode:
316 mode += "b" # for that other OS
319 mode += "b" # for that other OS
317
320
318 nlink = -1
321 nlink = -1
319 if mode not in ('r', 'rb'):
322 if mode not in ('r', 'rb'):
320 dirname, basename = util.split(f)
323 dirname, basename = util.split(f)
321 # If basename is empty, then the path is malformed because it points
324 # If basename is empty, then the path is malformed because it points
322 # to a directory. Let the posixfile() call below raise IOError.
325 # to a directory. Let the posixfile() call below raise IOError.
323 if basename:
326 if basename:
324 if atomictemp:
327 if atomictemp:
325 util.ensuredirs(dirname, self.createmode)
328 util.ensuredirs(dirname, self.createmode)
326 return util.atomictempfile(f, mode, self.createmode)
329 return util.atomictempfile(f, mode, self.createmode)
327 try:
330 try:
328 if 'w' in mode:
331 if 'w' in mode:
329 util.unlink(f)
332 util.unlink(f)
330 nlink = 0
333 nlink = 0
331 else:
334 else:
332 # nlinks() may behave differently for files on Windows
335 # nlinks() may behave differently for files on Windows
333 # shares if the file is open.
336 # shares if the file is open.
334 fd = util.posixfile(f)
337 fd = util.posixfile(f)
335 nlink = util.nlinks(f)
338 nlink = util.nlinks(f)
336 if nlink < 1:
339 if nlink < 1:
337 nlink = 2 # force mktempcopy (issue1922)
340 nlink = 2 # force mktempcopy (issue1922)
338 fd.close()
341 fd.close()
339 except (OSError, IOError), e:
342 except (OSError, IOError), e:
340 if e.errno != errno.ENOENT:
343 if e.errno != errno.ENOENT:
341 raise
344 raise
342 nlink = 0
345 nlink = 0
343 util.ensuredirs(dirname, self.createmode)
346 util.ensuredirs(dirname, self.createmode)
344 if nlink > 0:
347 if nlink > 0:
345 if self._trustnlink is None:
348 if self._trustnlink is None:
346 self._trustnlink = nlink > 1 or util.checknlink(f)
349 self._trustnlink = nlink > 1 or util.checknlink(f)
347 if nlink > 1 or not self._trustnlink:
350 if nlink > 1 or not self._trustnlink:
348 util.rename(util.mktempcopy(f), f)
351 util.rename(util.mktempcopy(f), f)
349 fp = util.posixfile(f, mode)
352 fp = util.posixfile(f, mode)
350 if nlink == 0:
353 if nlink == 0:
351 self._fixfilemode(f)
354 self._fixfilemode(f)
352 return fp
355 return fp
353
356
354 def symlink(self, src, dst):
357 def symlink(self, src, dst):
355 self.audit(dst)
358 self.audit(dst)
356 linkname = self.join(dst)
359 linkname = self.join(dst)
357 try:
360 try:
358 os.unlink(linkname)
361 os.unlink(linkname)
359 except OSError:
362 except OSError:
360 pass
363 pass
361
364
362 util.ensuredirs(os.path.dirname(linkname), self.createmode)
365 util.ensuredirs(os.path.dirname(linkname), self.createmode)
363
366
364 if self._cansymlink:
367 if self._cansymlink:
365 try:
368 try:
366 os.symlink(src, linkname)
369 os.symlink(src, linkname)
367 except OSError, err:
370 except OSError, err:
368 raise OSError(err.errno, _('could not symlink to %r: %s') %
371 raise OSError(err.errno, _('could not symlink to %r: %s') %
369 (src, err.strerror), linkname)
372 (src, err.strerror), linkname)
370 else:
373 else:
371 self.write(dst, src)
374 self.write(dst, src)
372
375
373 def join(self, path):
376 def join(self, path):
374 if path:
377 if path:
375 return os.path.join(self.base, path)
378 return os.path.join(self.base, path)
376 else:
379 else:
377 return self.base
380 return self.base
378
381
379 opener = vfs
382 opener = vfs
380
383
381 class auditvfs(object):
384 class auditvfs(object):
382 def __init__(self, vfs):
385 def __init__(self, vfs):
383 self.vfs = vfs
386 self.vfs = vfs
384
387
385 def _getmustaudit(self):
388 def _getmustaudit(self):
386 return self.vfs.mustaudit
389 return self.vfs.mustaudit
387
390
388 def _setmustaudit(self, onoff):
391 def _setmustaudit(self, onoff):
389 self.vfs.mustaudit = onoff
392 self.vfs.mustaudit = onoff
390
393
391 mustaudit = property(_getmustaudit, _setmustaudit)
394 mustaudit = property(_getmustaudit, _setmustaudit)
392
395
393 class filtervfs(abstractvfs, auditvfs):
396 class filtervfs(abstractvfs, auditvfs):
394 '''Wrapper vfs for filtering filenames with a function.'''
397 '''Wrapper vfs for filtering filenames with a function.'''
395
398
396 def __init__(self, vfs, filter):
399 def __init__(self, vfs, filter):
397 auditvfs.__init__(self, vfs)
400 auditvfs.__init__(self, vfs)
398 self._filter = filter
401 self._filter = filter
399
402
400 def __call__(self, path, *args, **kwargs):
403 def __call__(self, path, *args, **kwargs):
401 return self.vfs(self._filter(path), *args, **kwargs)
404 return self.vfs(self._filter(path), *args, **kwargs)
402
405
403 def join(self, path):
406 def join(self, path):
404 if path:
407 if path:
405 return self.vfs.join(self._filter(path))
408 return self.vfs.join(self._filter(path))
406 else:
409 else:
407 return self.vfs.join(path)
410 return self.vfs.join(path)
408
411
409 filteropener = filtervfs
412 filteropener = filtervfs
410
413
411 class readonlyvfs(abstractvfs, auditvfs):
414 class readonlyvfs(abstractvfs, auditvfs):
412 '''Wrapper vfs preventing any writing.'''
415 '''Wrapper vfs preventing any writing.'''
413
416
414 def __init__(self, vfs):
417 def __init__(self, vfs):
415 auditvfs.__init__(self, vfs)
418 auditvfs.__init__(self, vfs)
416
419
417 def __call__(self, path, mode='r', *args, **kw):
420 def __call__(self, path, mode='r', *args, **kw):
418 if mode not in ('r', 'rb'):
421 if mode not in ('r', 'rb'):
419 raise util.Abort('this vfs is read only')
422 raise util.Abort('this vfs is read only')
420 return self.vfs(path, mode, *args, **kw)
423 return self.vfs(path, mode, *args, **kw)
421
424
422
425
423 def canonpath(root, cwd, myname, auditor=None):
426 def canonpath(root, cwd, myname, auditor=None):
424 '''return the canonical path of myname, given cwd and root'''
427 '''return the canonical path of myname, given cwd and root'''
425 if util.endswithsep(root):
428 if util.endswithsep(root):
426 rootsep = root
429 rootsep = root
427 else:
430 else:
428 rootsep = root + os.sep
431 rootsep = root + os.sep
429 name = myname
432 name = myname
430 if not os.path.isabs(name):
433 if not os.path.isabs(name):
431 name = os.path.join(root, cwd, name)
434 name = os.path.join(root, cwd, name)
432 name = os.path.normpath(name)
435 name = os.path.normpath(name)
433 if auditor is None:
436 if auditor is None:
434 auditor = pathauditor(root)
437 auditor = pathauditor(root)
435 if name != rootsep and name.startswith(rootsep):
438 if name != rootsep and name.startswith(rootsep):
436 name = name[len(rootsep):]
439 name = name[len(rootsep):]
437 auditor(name)
440 auditor(name)
438 return util.pconvert(name)
441 return util.pconvert(name)
439 elif name == root:
442 elif name == root:
440 return ''
443 return ''
441 else:
444 else:
442 # Determine whether `name' is in the hierarchy at or beneath `root',
445 # Determine whether `name' is in the hierarchy at or beneath `root',
443 # by iterating name=dirname(name) until that causes no change (can't
446 # by iterating name=dirname(name) until that causes no change (can't
444 # check name == '/', because that doesn't work on windows). The list
447 # check name == '/', because that doesn't work on windows). The list
445 # `rel' holds the reversed list of components making up the relative
448 # `rel' holds the reversed list of components making up the relative
446 # file name we want.
449 # file name we want.
447 rel = []
450 rel = []
448 while True:
451 while True:
449 try:
452 try:
450 s = util.samefile(name, root)
453 s = util.samefile(name, root)
451 except OSError:
454 except OSError:
452 s = False
455 s = False
453 if s:
456 if s:
454 if not rel:
457 if not rel:
455 # name was actually the same as root (maybe a symlink)
458 # name was actually the same as root (maybe a symlink)
456 return ''
459 return ''
457 rel.reverse()
460 rel.reverse()
458 name = os.path.join(*rel)
461 name = os.path.join(*rel)
459 auditor(name)
462 auditor(name)
460 return util.pconvert(name)
463 return util.pconvert(name)
461 dirname, basename = util.split(name)
464 dirname, basename = util.split(name)
462 rel.append(basename)
465 rel.append(basename)
463 if dirname == name:
466 if dirname == name:
464 break
467 break
465 name = dirname
468 name = dirname
466
469
467 raise util.Abort(_("%s not under root '%s'") % (myname, root))
470 raise util.Abort(_("%s not under root '%s'") % (myname, root))
468
471
469 def walkrepos(path, followsym=False, seen_dirs=None, recurse=False):
472 def walkrepos(path, followsym=False, seen_dirs=None, recurse=False):
470 '''yield every hg repository under path, always recursively.
473 '''yield every hg repository under path, always recursively.
471 The recurse flag will only control recursion into repo working dirs'''
474 The recurse flag will only control recursion into repo working dirs'''
472 def errhandler(err):
475 def errhandler(err):
473 if err.filename == path:
476 if err.filename == path:
474 raise err
477 raise err
475 samestat = getattr(os.path, 'samestat', None)
478 samestat = getattr(os.path, 'samestat', None)
476 if followsym and samestat is not None:
479 if followsym and samestat is not None:
477 def adddir(dirlst, dirname):
480 def adddir(dirlst, dirname):
478 match = False
481 match = False
479 dirstat = os.stat(dirname)
482 dirstat = os.stat(dirname)
480 for lstdirstat in dirlst:
483 for lstdirstat in dirlst:
481 if samestat(dirstat, lstdirstat):
484 if samestat(dirstat, lstdirstat):
482 match = True
485 match = True
483 break
486 break
484 if not match:
487 if not match:
485 dirlst.append(dirstat)
488 dirlst.append(dirstat)
486 return not match
489 return not match
487 else:
490 else:
488 followsym = False
491 followsym = False
489
492
490 if (seen_dirs is None) and followsym:
493 if (seen_dirs is None) and followsym:
491 seen_dirs = []
494 seen_dirs = []
492 adddir(seen_dirs, path)
495 adddir(seen_dirs, path)
493 for root, dirs, files in os.walk(path, topdown=True, onerror=errhandler):
496 for root, dirs, files in os.walk(path, topdown=True, onerror=errhandler):
494 dirs.sort()
497 dirs.sort()
495 if '.hg' in dirs:
498 if '.hg' in dirs:
496 yield root # found a repository
499 yield root # found a repository
497 qroot = os.path.join(root, '.hg', 'patches')
500 qroot = os.path.join(root, '.hg', 'patches')
498 if os.path.isdir(os.path.join(qroot, '.hg')):
501 if os.path.isdir(os.path.join(qroot, '.hg')):
499 yield qroot # we have a patch queue repo here
502 yield qroot # we have a patch queue repo here
500 if recurse:
503 if recurse:
501 # avoid recursing inside the .hg directory
504 # avoid recursing inside the .hg directory
502 dirs.remove('.hg')
505 dirs.remove('.hg')
503 else:
506 else:
504 dirs[:] = [] # don't descend further
507 dirs[:] = [] # don't descend further
505 elif followsym:
508 elif followsym:
506 newdirs = []
509 newdirs = []
507 for d in dirs:
510 for d in dirs:
508 fname = os.path.join(root, d)
511 fname = os.path.join(root, d)
509 if adddir(seen_dirs, fname):
512 if adddir(seen_dirs, fname):
510 if os.path.islink(fname):
513 if os.path.islink(fname):
511 for hgname in walkrepos(fname, True, seen_dirs):
514 for hgname in walkrepos(fname, True, seen_dirs):
512 yield hgname
515 yield hgname
513 else:
516 else:
514 newdirs.append(d)
517 newdirs.append(d)
515 dirs[:] = newdirs
518 dirs[:] = newdirs
516
519
517 def osrcpath():
520 def osrcpath():
518 '''return default os-specific hgrc search path'''
521 '''return default os-specific hgrc search path'''
519 path = systemrcpath()
522 path = systemrcpath()
520 path.extend(userrcpath())
523 path.extend(userrcpath())
521 path = [os.path.normpath(f) for f in path]
524 path = [os.path.normpath(f) for f in path]
522 return path
525 return path
523
526
524 _rcpath = None
527 _rcpath = None
525
528
526 def rcpath():
529 def rcpath():
527 '''return hgrc search path. if env var HGRCPATH is set, use it.
530 '''return hgrc search path. if env var HGRCPATH is set, use it.
528 for each item in path, if directory, use files ending in .rc,
531 for each item in path, if directory, use files ending in .rc,
529 else use item.
532 else use item.
530 make HGRCPATH empty to only look in .hg/hgrc of current repo.
533 make HGRCPATH empty to only look in .hg/hgrc of current repo.
531 if no HGRCPATH, use default os-specific path.'''
534 if no HGRCPATH, use default os-specific path.'''
532 global _rcpath
535 global _rcpath
533 if _rcpath is None:
536 if _rcpath is None:
534 if 'HGRCPATH' in os.environ:
537 if 'HGRCPATH' in os.environ:
535 _rcpath = []
538 _rcpath = []
536 for p in os.environ['HGRCPATH'].split(os.pathsep):
539 for p in os.environ['HGRCPATH'].split(os.pathsep):
537 if not p:
540 if not p:
538 continue
541 continue
539 p = util.expandpath(p)
542 p = util.expandpath(p)
540 if os.path.isdir(p):
543 if os.path.isdir(p):
541 for f, kind in osutil.listdir(p):
544 for f, kind in osutil.listdir(p):
542 if f.endswith('.rc'):
545 if f.endswith('.rc'):
543 _rcpath.append(os.path.join(p, f))
546 _rcpath.append(os.path.join(p, f))
544 else:
547 else:
545 _rcpath.append(p)
548 _rcpath.append(p)
546 else:
549 else:
547 _rcpath = osrcpath()
550 _rcpath = osrcpath()
548 return _rcpath
551 return _rcpath
549
552
550 def revsingle(repo, revspec, default='.'):
553 def revsingle(repo, revspec, default='.'):
551 if not revspec:
554 if not revspec:
552 return repo[default]
555 return repo[default]
553
556
554 l = revrange(repo, [revspec])
557 l = revrange(repo, [revspec])
555 if len(l) < 1:
558 if len(l) < 1:
556 raise util.Abort(_('empty revision set'))
559 raise util.Abort(_('empty revision set'))
557 return repo[l[-1]]
560 return repo[l[-1]]
558
561
559 def revpair(repo, revs):
562 def revpair(repo, revs):
560 if not revs:
563 if not revs:
561 return repo.dirstate.p1(), None
564 return repo.dirstate.p1(), None
562
565
563 l = revrange(repo, revs)
566 l = revrange(repo, revs)
564
567
565 if len(l) == 0:
568 if len(l) == 0:
566 if revs:
569 if revs:
567 raise util.Abort(_('empty revision range'))
570 raise util.Abort(_('empty revision range'))
568 return repo.dirstate.p1(), None
571 return repo.dirstate.p1(), None
569
572
570 if len(l) == 1 and len(revs) == 1 and _revrangesep not in revs[0]:
573 if len(l) == 1 and len(revs) == 1 and _revrangesep not in revs[0]:
571 return repo.lookup(l[0]), None
574 return repo.lookup(l[0]), None
572
575
573 return repo.lookup(l[0]), repo.lookup(l[-1])
576 return repo.lookup(l[0]), repo.lookup(l[-1])
574
577
575 _revrangesep = ':'
578 _revrangesep = ':'
576
579
577 def revrange(repo, revs):
580 def revrange(repo, revs):
578 """Yield revision as strings from a list of revision specifications."""
581 """Yield revision as strings from a list of revision specifications."""
579
582
580 def revfix(repo, val, defval):
583 def revfix(repo, val, defval):
581 if not val and val != 0 and defval is not None:
584 if not val and val != 0 and defval is not None:
582 return defval
585 return defval
583 return repo[val].rev()
586 return repo[val].rev()
584
587
585 seen, l = set(), []
588 seen, l = set(), []
586 for spec in revs:
589 for spec in revs:
587 if l and not seen:
590 if l and not seen:
588 seen = set(l)
591 seen = set(l)
589 # attempt to parse old-style ranges first to deal with
592 # attempt to parse old-style ranges first to deal with
590 # things like old-tag which contain query metacharacters
593 # things like old-tag which contain query metacharacters
591 try:
594 try:
592 if isinstance(spec, int):
595 if isinstance(spec, int):
593 seen.add(spec)
596 seen.add(spec)
594 l.append(spec)
597 l.append(spec)
595 continue
598 continue
596
599
597 if _revrangesep in spec:
600 if _revrangesep in spec:
598 start, end = spec.split(_revrangesep, 1)
601 start, end = spec.split(_revrangesep, 1)
599 start = revfix(repo, start, 0)
602 start = revfix(repo, start, 0)
600 end = revfix(repo, end, len(repo) - 1)
603 end = revfix(repo, end, len(repo) - 1)
601 if end == nullrev and start <= 0:
604 if end == nullrev and start <= 0:
602 start = nullrev
605 start = nullrev
603 rangeiter = repo.changelog.revs(start, end)
606 rangeiter = repo.changelog.revs(start, end)
604 if not seen and not l:
607 if not seen and not l:
605 # by far the most common case: revs = ["-1:0"]
608 # by far the most common case: revs = ["-1:0"]
606 l = list(rangeiter)
609 l = list(rangeiter)
607 # defer syncing seen until next iteration
610 # defer syncing seen until next iteration
608 continue
611 continue
609 newrevs = set(rangeiter)
612 newrevs = set(rangeiter)
610 if seen:
613 if seen:
611 newrevs.difference_update(seen)
614 newrevs.difference_update(seen)
612 seen.update(newrevs)
615 seen.update(newrevs)
613 else:
616 else:
614 seen = newrevs
617 seen = newrevs
615 l.extend(sorted(newrevs, reverse=start > end))
618 l.extend(sorted(newrevs, reverse=start > end))
616 continue
619 continue
617 elif spec and spec in repo: # single unquoted rev
620 elif spec and spec in repo: # single unquoted rev
618 rev = revfix(repo, spec, None)
621 rev = revfix(repo, spec, None)
619 if rev in seen:
622 if rev in seen:
620 continue
623 continue
621 seen.add(rev)
624 seen.add(rev)
622 l.append(rev)
625 l.append(rev)
623 continue
626 continue
624 except error.RepoLookupError:
627 except error.RepoLookupError:
625 pass
628 pass
626
629
627 # fall through to new-style queries if old-style fails
630 # fall through to new-style queries if old-style fails
628 m = revset.match(repo.ui, spec)
631 m = revset.match(repo.ui, spec)
629 dl = [r for r in m(repo, list(repo)) if r not in seen]
632 dl = [r for r in m(repo, list(repo)) if r not in seen]
630 l.extend(dl)
633 l.extend(dl)
631 seen.update(dl)
634 seen.update(dl)
632
635
633 return l
636 return l
634
637
635 def expandpats(pats):
638 def expandpats(pats):
636 if not util.expandglobs:
639 if not util.expandglobs:
637 return list(pats)
640 return list(pats)
638 ret = []
641 ret = []
639 for p in pats:
642 for p in pats:
640 kind, name = matchmod._patsplit(p, None)
643 kind, name = matchmod._patsplit(p, None)
641 if kind is None:
644 if kind is None:
642 try:
645 try:
643 globbed = glob.glob(name)
646 globbed = glob.glob(name)
644 except re.error:
647 except re.error:
645 globbed = [name]
648 globbed = [name]
646 if globbed:
649 if globbed:
647 ret.extend(globbed)
650 ret.extend(globbed)
648 continue
651 continue
649 ret.append(p)
652 ret.append(p)
650 return ret
653 return ret
651
654
652 def matchandpats(ctx, pats=[], opts={}, globbed=False, default='relpath'):
655 def matchandpats(ctx, pats=[], opts={}, globbed=False, default='relpath'):
653 if pats == ("",):
656 if pats == ("",):
654 pats = []
657 pats = []
655 if not globbed and default == 'relpath':
658 if not globbed and default == 'relpath':
656 pats = expandpats(pats or [])
659 pats = expandpats(pats or [])
657
660
658 m = ctx.match(pats, opts.get('include'), opts.get('exclude'),
661 m = ctx.match(pats, opts.get('include'), opts.get('exclude'),
659 default)
662 default)
660 def badfn(f, msg):
663 def badfn(f, msg):
661 ctx._repo.ui.warn("%s: %s\n" % (m.rel(f), msg))
664 ctx._repo.ui.warn("%s: %s\n" % (m.rel(f), msg))
662 m.bad = badfn
665 m.bad = badfn
663 return m, pats
666 return m, pats
664
667
665 def match(ctx, pats=[], opts={}, globbed=False, default='relpath'):
668 def match(ctx, pats=[], opts={}, globbed=False, default='relpath'):
666 return matchandpats(ctx, pats, opts, globbed, default)[0]
669 return matchandpats(ctx, pats, opts, globbed, default)[0]
667
670
668 def matchall(repo):
671 def matchall(repo):
669 return matchmod.always(repo.root, repo.getcwd())
672 return matchmod.always(repo.root, repo.getcwd())
670
673
671 def matchfiles(repo, files):
674 def matchfiles(repo, files):
672 return matchmod.exact(repo.root, repo.getcwd(), files)
675 return matchmod.exact(repo.root, repo.getcwd(), files)
673
676
674 def addremove(repo, pats=[], opts={}, dry_run=None, similarity=None):
677 def addremove(repo, pats=[], opts={}, dry_run=None, similarity=None):
675 if dry_run is None:
678 if dry_run is None:
676 dry_run = opts.get('dry_run')
679 dry_run = opts.get('dry_run')
677 if similarity is None:
680 if similarity is None:
678 similarity = float(opts.get('similarity') or 0)
681 similarity = float(opts.get('similarity') or 0)
679 # we'd use status here, except handling of symlinks and ignore is tricky
682 # we'd use status here, except handling of symlinks and ignore is tricky
680 added, unknown, deleted, removed = [], [], [], []
683 added, unknown, deleted, removed = [], [], [], []
681 audit_path = pathauditor(repo.root)
684 audit_path = pathauditor(repo.root)
682 m = match(repo[None], pats, opts)
685 m = match(repo[None], pats, opts)
683 rejected = []
686 rejected = []
684 m.bad = lambda x, y: rejected.append(x)
687 m.bad = lambda x, y: rejected.append(x)
685
688
686 ctx = repo[None]
689 ctx = repo[None]
687 dirstate = repo.dirstate
690 dirstate = repo.dirstate
688 walkresults = dirstate.walk(m, sorted(ctx.substate), True, False)
691 walkresults = dirstate.walk(m, sorted(ctx.substate), True, False)
689 for abs, st in walkresults.iteritems():
692 for abs, st in walkresults.iteritems():
690 dstate = dirstate[abs]
693 dstate = dirstate[abs]
691 if dstate == '?' and audit_path.check(abs):
694 if dstate == '?' and audit_path.check(abs):
692 unknown.append(abs)
695 unknown.append(abs)
693 elif dstate != 'r' and not st:
696 elif dstate != 'r' and not st:
694 deleted.append(abs)
697 deleted.append(abs)
695 # for finding renames
698 # for finding renames
696 elif dstate == 'r':
699 elif dstate == 'r':
697 removed.append(abs)
700 removed.append(abs)
698 elif dstate == 'a':
701 elif dstate == 'a':
699 added.append(abs)
702 added.append(abs)
700
703
701 unknownset = set(unknown)
704 unknownset = set(unknown)
702 toprint = unknownset.copy()
705 toprint = unknownset.copy()
703 toprint.update(deleted)
706 toprint.update(deleted)
704 for abs in sorted(toprint):
707 for abs in sorted(toprint):
705 if repo.ui.verbose or not m.exact(abs):
708 if repo.ui.verbose or not m.exact(abs):
706 rel = m.rel(abs)
709 rel = m.rel(abs)
707 if abs in unknownset:
710 if abs in unknownset:
708 status = _('adding %s\n') % ((pats and rel) or abs)
711 status = _('adding %s\n') % ((pats and rel) or abs)
709 else:
712 else:
710 status = _('removing %s\n') % ((pats and rel) or abs)
713 status = _('removing %s\n') % ((pats and rel) or abs)
711 repo.ui.status(status)
714 repo.ui.status(status)
712
715
713 copies = {}
716 copies = {}
714 if similarity > 0:
717 if similarity > 0:
715 for old, new, score in similar.findrenames(repo,
718 for old, new, score in similar.findrenames(repo,
716 added + unknown, removed + deleted, similarity):
719 added + unknown, removed + deleted, similarity):
717 if repo.ui.verbose or not m.exact(old) or not m.exact(new):
720 if repo.ui.verbose or not m.exact(old) or not m.exact(new):
718 repo.ui.status(_('recording removal of %s as rename to %s '
721 repo.ui.status(_('recording removal of %s as rename to %s '
719 '(%d%% similar)\n') %
722 '(%d%% similar)\n') %
720 (m.rel(old), m.rel(new), score * 100))
723 (m.rel(old), m.rel(new), score * 100))
721 copies[new] = old
724 copies[new] = old
722
725
723 if not dry_run:
726 if not dry_run:
724 wctx = repo[None]
727 wctx = repo[None]
725 wlock = repo.wlock()
728 wlock = repo.wlock()
726 try:
729 try:
727 wctx.forget(deleted)
730 wctx.forget(deleted)
728 wctx.add(unknown)
731 wctx.add(unknown)
729 for new, old in copies.iteritems():
732 for new, old in copies.iteritems():
730 wctx.copy(old, new)
733 wctx.copy(old, new)
731 finally:
734 finally:
732 wlock.release()
735 wlock.release()
733
736
734 for f in rejected:
737 for f in rejected:
735 if f in m.files():
738 if f in m.files():
736 return 1
739 return 1
737 return 0
740 return 0
738
741
739 def dirstatecopy(ui, repo, wctx, src, dst, dryrun=False, cwd=None):
742 def dirstatecopy(ui, repo, wctx, src, dst, dryrun=False, cwd=None):
740 """Update the dirstate to reflect the intent of copying src to dst. For
743 """Update the dirstate to reflect the intent of copying src to dst. For
741 different reasons it might not end with dst being marked as copied from src.
744 different reasons it might not end with dst being marked as copied from src.
742 """
745 """
743 origsrc = repo.dirstate.copied(src) or src
746 origsrc = repo.dirstate.copied(src) or src
744 if dst == origsrc: # copying back a copy?
747 if dst == origsrc: # copying back a copy?
745 if repo.dirstate[dst] not in 'mn' and not dryrun:
748 if repo.dirstate[dst] not in 'mn' and not dryrun:
746 repo.dirstate.normallookup(dst)
749 repo.dirstate.normallookup(dst)
747 else:
750 else:
748 if repo.dirstate[origsrc] == 'a' and origsrc == src:
751 if repo.dirstate[origsrc] == 'a' and origsrc == src:
749 if not ui.quiet:
752 if not ui.quiet:
750 ui.warn(_("%s has not been committed yet, so no copy "
753 ui.warn(_("%s has not been committed yet, so no copy "
751 "data will be stored for %s.\n")
754 "data will be stored for %s.\n")
752 % (repo.pathto(origsrc, cwd), repo.pathto(dst, cwd)))
755 % (repo.pathto(origsrc, cwd), repo.pathto(dst, cwd)))
753 if repo.dirstate[dst] in '?r' and not dryrun:
756 if repo.dirstate[dst] in '?r' and not dryrun:
754 wctx.add([dst])
757 wctx.add([dst])
755 elif not dryrun:
758 elif not dryrun:
756 wctx.copy(origsrc, dst)
759 wctx.copy(origsrc, dst)
757
760
758 def readrequires(opener, supported):
761 def readrequires(opener, supported):
759 '''Reads and parses .hg/requires and checks if all entries found
762 '''Reads and parses .hg/requires and checks if all entries found
760 are in the list of supported features.'''
763 are in the list of supported features.'''
761 requirements = set(opener.read("requires").splitlines())
764 requirements = set(opener.read("requires").splitlines())
762 missings = []
765 missings = []
763 for r in requirements:
766 for r in requirements:
764 if r not in supported:
767 if r not in supported:
765 if not r or not r[0].isalnum():
768 if not r or not r[0].isalnum():
766 raise error.RequirementError(_(".hg/requires file is corrupt"))
769 raise error.RequirementError(_(".hg/requires file is corrupt"))
767 missings.append(r)
770 missings.append(r)
768 missings.sort()
771 missings.sort()
769 if missings:
772 if missings:
770 raise error.RequirementError(
773 raise error.RequirementError(
771 _("unknown repository format: requires features '%s' (upgrade "
774 _("unknown repository format: requires features '%s' (upgrade "
772 "Mercurial)") % "', '".join(missings))
775 "Mercurial)") % "', '".join(missings))
773 return requirements
776 return requirements
774
777
775 class filecacheentry(object):
778 class filecacheentry(object):
776 def __init__(self, path, stat=True):
779 def __init__(self, path, stat=True):
777 self.path = path
780 self.path = path
778 self.cachestat = None
781 self.cachestat = None
779 self._cacheable = None
782 self._cacheable = None
780
783
781 if stat:
784 if stat:
782 self.cachestat = filecacheentry.stat(self.path)
785 self.cachestat = filecacheentry.stat(self.path)
783
786
784 if self.cachestat:
787 if self.cachestat:
785 self._cacheable = self.cachestat.cacheable()
788 self._cacheable = self.cachestat.cacheable()
786 else:
789 else:
787 # None means we don't know yet
790 # None means we don't know yet
788 self._cacheable = None
791 self._cacheable = None
789
792
790 def refresh(self):
793 def refresh(self):
791 if self.cacheable():
794 if self.cacheable():
792 self.cachestat = filecacheentry.stat(self.path)
795 self.cachestat = filecacheentry.stat(self.path)
793
796
794 def cacheable(self):
797 def cacheable(self):
795 if self._cacheable is not None:
798 if self._cacheable is not None:
796 return self._cacheable
799 return self._cacheable
797
800
798 # we don't know yet, assume it is for now
801 # we don't know yet, assume it is for now
799 return True
802 return True
800
803
801 def changed(self):
804 def changed(self):
802 # no point in going further if we can't cache it
805 # no point in going further if we can't cache it
803 if not self.cacheable():
806 if not self.cacheable():
804 return True
807 return True
805
808
806 newstat = filecacheentry.stat(self.path)
809 newstat = filecacheentry.stat(self.path)
807
810
808 # we may not know if it's cacheable yet, check again now
811 # we may not know if it's cacheable yet, check again now
809 if newstat and self._cacheable is None:
812 if newstat and self._cacheable is None:
810 self._cacheable = newstat.cacheable()
813 self._cacheable = newstat.cacheable()
811
814
812 # check again
815 # check again
813 if not self._cacheable:
816 if not self._cacheable:
814 return True
817 return True
815
818
816 if self.cachestat != newstat:
819 if self.cachestat != newstat:
817 self.cachestat = newstat
820 self.cachestat = newstat
818 return True
821 return True
819 else:
822 else:
820 return False
823 return False
821
824
822 @staticmethod
825 @staticmethod
823 def stat(path):
826 def stat(path):
824 try:
827 try:
825 return util.cachestat(path)
828 return util.cachestat(path)
826 except OSError, e:
829 except OSError, e:
827 if e.errno != errno.ENOENT:
830 if e.errno != errno.ENOENT:
828 raise
831 raise
829
832
830 class filecache(object):
833 class filecache(object):
831 '''A property like decorator that tracks a file under .hg/ for updates.
834 '''A property like decorator that tracks a file under .hg/ for updates.
832
835
833 Records stat info when called in _filecache.
836 Records stat info when called in _filecache.
834
837
835 On subsequent calls, compares old stat info with new info, and recreates
838 On subsequent calls, compares old stat info with new info, and recreates
836 the object when needed, updating the new stat info in _filecache.
839 the object when needed, updating the new stat info in _filecache.
837
840
838 Mercurial either atomic renames or appends for files under .hg,
841 Mercurial either atomic renames or appends for files under .hg,
839 so to ensure the cache is reliable we need the filesystem to be able
842 so to ensure the cache is reliable we need the filesystem to be able
840 to tell us if a file has been replaced. If it can't, we fallback to
843 to tell us if a file has been replaced. If it can't, we fallback to
841 recreating the object on every call (essentially the same behaviour as
844 recreating the object on every call (essentially the same behaviour as
842 propertycache).'''
845 propertycache).'''
843 def __init__(self, path):
846 def __init__(self, path):
844 self.path = path
847 self.path = path
845
848
846 def join(self, obj, fname):
849 def join(self, obj, fname):
847 """Used to compute the runtime path of the cached file.
850 """Used to compute the runtime path of the cached file.
848
851
849 Users should subclass filecache and provide their own version of this
852 Users should subclass filecache and provide their own version of this
850 function to call the appropriate join function on 'obj' (an instance
853 function to call the appropriate join function on 'obj' (an instance
851 of the class that its member function was decorated).
854 of the class that its member function was decorated).
852 """
855 """
853 return obj.join(fname)
856 return obj.join(fname)
854
857
855 def __call__(self, func):
858 def __call__(self, func):
856 self.func = func
859 self.func = func
857 self.name = func.__name__
860 self.name = func.__name__
858 return self
861 return self
859
862
860 def __get__(self, obj, type=None):
863 def __get__(self, obj, type=None):
861 # do we need to check if the file changed?
864 # do we need to check if the file changed?
862 if self.name in obj.__dict__:
865 if self.name in obj.__dict__:
863 assert self.name in obj._filecache, self.name
866 assert self.name in obj._filecache, self.name
864 return obj.__dict__[self.name]
867 return obj.__dict__[self.name]
865
868
866 entry = obj._filecache.get(self.name)
869 entry = obj._filecache.get(self.name)
867
870
868 if entry:
871 if entry:
869 if entry.changed():
872 if entry.changed():
870 entry.obj = self.func(obj)
873 entry.obj = self.func(obj)
871 else:
874 else:
872 path = self.join(obj, self.path)
875 path = self.join(obj, self.path)
873
876
874 # We stat -before- creating the object so our cache doesn't lie if
877 # We stat -before- creating the object so our cache doesn't lie if
875 # a writer modified between the time we read and stat
878 # a writer modified between the time we read and stat
876 entry = filecacheentry(path)
879 entry = filecacheentry(path)
877 entry.obj = self.func(obj)
880 entry.obj = self.func(obj)
878
881
879 obj._filecache[self.name] = entry
882 obj._filecache[self.name] = entry
880
883
881 obj.__dict__[self.name] = entry.obj
884 obj.__dict__[self.name] = entry.obj
882 return entry.obj
885 return entry.obj
883
886
884 def __set__(self, obj, value):
887 def __set__(self, obj, value):
885 if self.name not in obj._filecache:
888 if self.name not in obj._filecache:
886 # we add an entry for the missing value because X in __dict__
889 # we add an entry for the missing value because X in __dict__
887 # implies X in _filecache
890 # implies X in _filecache
888 ce = filecacheentry(self.join(obj, self.path), False)
891 ce = filecacheentry(self.join(obj, self.path), False)
889 obj._filecache[self.name] = ce
892 obj._filecache[self.name] = ce
890 else:
893 else:
891 ce = obj._filecache[self.name]
894 ce = obj._filecache[self.name]
892
895
893 ce.obj = value # update cached copy
896 ce.obj = value # update cached copy
894 obj.__dict__[self.name] = value # update copy returned by obj.x
897 obj.__dict__[self.name] = value # update copy returned by obj.x
895
898
896 def __delete__(self, obj):
899 def __delete__(self, obj):
897 try:
900 try:
898 del obj.__dict__[self.name]
901 del obj.__dict__[self.name]
899 except KeyError:
902 except KeyError:
900 raise AttributeError(self.name)
903 raise AttributeError(self.name)
901
904
902 class dirs(object):
905 class dirs(object):
903 '''a multiset of directory names from a dirstate or manifest'''
906 '''a multiset of directory names from a dirstate or manifest'''
904
907
905 def __init__(self, map, skip=None):
908 def __init__(self, map, skip=None):
906 self._dirs = {}
909 self._dirs = {}
907 addpath = self.addpath
910 addpath = self.addpath
908 if util.safehasattr(map, 'iteritems') and skip is not None:
911 if util.safehasattr(map, 'iteritems') and skip is not None:
909 for f, s in map.iteritems():
912 for f, s in map.iteritems():
910 if s[0] != skip:
913 if s[0] != skip:
911 addpath(f)
914 addpath(f)
912 else:
915 else:
913 for f in map:
916 for f in map:
914 addpath(f)
917 addpath(f)
915
918
916 def addpath(self, path):
919 def addpath(self, path):
917 dirs = self._dirs
920 dirs = self._dirs
918 for base in finddirs(path):
921 for base in finddirs(path):
919 if base in dirs:
922 if base in dirs:
920 dirs[base] += 1
923 dirs[base] += 1
921 return
924 return
922 dirs[base] = 1
925 dirs[base] = 1
923
926
924 def delpath(self, path):
927 def delpath(self, path):
925 dirs = self._dirs
928 dirs = self._dirs
926 for base in finddirs(path):
929 for base in finddirs(path):
927 if dirs[base] > 1:
930 if dirs[base] > 1:
928 dirs[base] -= 1
931 dirs[base] -= 1
929 return
932 return
930 del dirs[base]
933 del dirs[base]
931
934
932 def __iter__(self):
935 def __iter__(self):
933 return self._dirs.iterkeys()
936 return self._dirs.iterkeys()
934
937
935 def __contains__(self, d):
938 def __contains__(self, d):
936 return d in self._dirs
939 return d in self._dirs
937
940
938 if util.safehasattr(parsers, 'dirs'):
941 if util.safehasattr(parsers, 'dirs'):
939 dirs = parsers.dirs
942 dirs = parsers.dirs
940
943
941 def finddirs(path):
944 def finddirs(path):
942 pos = path.rfind('/')
945 pos = path.rfind('/')
943 while pos != -1:
946 while pos != -1:
944 yield path[:pos]
947 yield path[:pos]
945 pos = path.rfind('/', 0, pos)
948 pos = path.rfind('/', 0, pos)
General Comments 0
You need to be logged in to leave comments. Login now