##// END OF EJS Templates
destroyed: keep the filecache in sync with __dict__ (issue3335) (issue3693) (issue3743)...
Idan Kamara -
r18313:3e4a944c default
parent child Browse files
Show More
@@ -1,2586 +1,2583 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import lock, transaction, store, encoding, base85
11 import lock, transaction, store, encoding, base85
12 import scmutil, util, extensions, hook, error, revset
12 import scmutil, util, extensions, hook, error, revset
13 import match as matchmod
13 import match as matchmod
14 import merge as mergemod
14 import merge as mergemod
15 import tags as tagsmod
15 import tags as tagsmod
16 from lock import release
16 from lock import release
17 import weakref, errno, os, time, inspect
17 import weakref, errno, os, time, inspect
18 import branchmap
18 import branchmap
19 propertycache = util.propertycache
19 propertycache = util.propertycache
20 filecache = scmutil.filecache
20 filecache = scmutil.filecache
21
21
22 class repofilecache(filecache):
22 class repofilecache(filecache):
23 """All filecache usage on repo are done for logic that should be unfiltered
23 """All filecache usage on repo are done for logic that should be unfiltered
24 """
24 """
25
25
26 def __get__(self, repo, type=None):
26 def __get__(self, repo, type=None):
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 def __set__(self, repo, value):
28 def __set__(self, repo, value):
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 def __delete__(self, repo):
30 def __delete__(self, repo):
31 return super(repofilecache, self).__delete__(repo.unfiltered())
31 return super(repofilecache, self).__delete__(repo.unfiltered())
32
32
33 class storecache(repofilecache):
33 class storecache(repofilecache):
34 """filecache for files in the store"""
34 """filecache for files in the store"""
35 def join(self, obj, fname):
35 def join(self, obj, fname):
36 return obj.sjoin(fname)
36 return obj.sjoin(fname)
37
37
38 class unfilteredpropertycache(propertycache):
38 class unfilteredpropertycache(propertycache):
39 """propertycache that apply to unfiltered repo only"""
39 """propertycache that apply to unfiltered repo only"""
40
40
41 def __get__(self, repo, type=None):
41 def __get__(self, repo, type=None):
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43
43
44 class filteredpropertycache(propertycache):
44 class filteredpropertycache(propertycache):
45 """propertycache that must take filtering in account"""
45 """propertycache that must take filtering in account"""
46
46
47 def cachevalue(self, obj, value):
47 def cachevalue(self, obj, value):
48 object.__setattr__(obj, self.name, value)
48 object.__setattr__(obj, self.name, value)
49
49
50
50
51 def hasunfilteredcache(repo, name):
51 def hasunfilteredcache(repo, name):
52 """check if an repo and a unfilteredproperty cached value for <name>"""
52 """check if an repo and a unfilteredproperty cached value for <name>"""
53 return name in vars(repo.unfiltered())
53 return name in vars(repo.unfiltered())
54
54
55 def unfilteredmethod(orig):
55 def unfilteredmethod(orig):
56 """decorate method that always need to be run on unfiltered version"""
56 """decorate method that always need to be run on unfiltered version"""
57 def wrapper(repo, *args, **kwargs):
57 def wrapper(repo, *args, **kwargs):
58 return orig(repo.unfiltered(), *args, **kwargs)
58 return orig(repo.unfiltered(), *args, **kwargs)
59 return wrapper
59 return wrapper
60
60
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63
63
64 class localpeer(peer.peerrepository):
64 class localpeer(peer.peerrepository):
65 '''peer for a local repo; reflects only the most recent API'''
65 '''peer for a local repo; reflects only the most recent API'''
66
66
67 def __init__(self, repo, caps=MODERNCAPS):
67 def __init__(self, repo, caps=MODERNCAPS):
68 peer.peerrepository.__init__(self)
68 peer.peerrepository.__init__(self)
69 self._repo = repo.filtered('unserved')
69 self._repo = repo.filtered('unserved')
70 self.ui = repo.ui
70 self.ui = repo.ui
71 self._caps = repo._restrictcapabilities(caps)
71 self._caps = repo._restrictcapabilities(caps)
72 self.requirements = repo.requirements
72 self.requirements = repo.requirements
73 self.supportedformats = repo.supportedformats
73 self.supportedformats = repo.supportedformats
74
74
75 def close(self):
75 def close(self):
76 self._repo.close()
76 self._repo.close()
77
77
78 def _capabilities(self):
78 def _capabilities(self):
79 return self._caps
79 return self._caps
80
80
81 def local(self):
81 def local(self):
82 return self._repo
82 return self._repo
83
83
84 def canpush(self):
84 def canpush(self):
85 return True
85 return True
86
86
87 def url(self):
87 def url(self):
88 return self._repo.url()
88 return self._repo.url()
89
89
90 def lookup(self, key):
90 def lookup(self, key):
91 return self._repo.lookup(key)
91 return self._repo.lookup(key)
92
92
93 def branchmap(self):
93 def branchmap(self):
94 return self._repo.branchmap()
94 return self._repo.branchmap()
95
95
96 def heads(self):
96 def heads(self):
97 return self._repo.heads()
97 return self._repo.heads()
98
98
99 def known(self, nodes):
99 def known(self, nodes):
100 return self._repo.known(nodes)
100 return self._repo.known(nodes)
101
101
102 def getbundle(self, source, heads=None, common=None):
102 def getbundle(self, source, heads=None, common=None):
103 return self._repo.getbundle(source, heads=heads, common=common)
103 return self._repo.getbundle(source, heads=heads, common=common)
104
104
105 # TODO We might want to move the next two calls into legacypeer and add
105 # TODO We might want to move the next two calls into legacypeer and add
106 # unbundle instead.
106 # unbundle instead.
107
107
108 def lock(self):
108 def lock(self):
109 return self._repo.lock()
109 return self._repo.lock()
110
110
111 def addchangegroup(self, cg, source, url):
111 def addchangegroup(self, cg, source, url):
112 return self._repo.addchangegroup(cg, source, url)
112 return self._repo.addchangegroup(cg, source, url)
113
113
114 def pushkey(self, namespace, key, old, new):
114 def pushkey(self, namespace, key, old, new):
115 return self._repo.pushkey(namespace, key, old, new)
115 return self._repo.pushkey(namespace, key, old, new)
116
116
117 def listkeys(self, namespace):
117 def listkeys(self, namespace):
118 return self._repo.listkeys(namespace)
118 return self._repo.listkeys(namespace)
119
119
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 '''used to test argument passing over the wire'''
121 '''used to test argument passing over the wire'''
122 return "%s %s %s %s %s" % (one, two, three, four, five)
122 return "%s %s %s %s %s" % (one, two, three, four, five)
123
123
124 class locallegacypeer(localpeer):
124 class locallegacypeer(localpeer):
125 '''peer extension which implements legacy methods too; used for tests with
125 '''peer extension which implements legacy methods too; used for tests with
126 restricted capabilities'''
126 restricted capabilities'''
127
127
128 def __init__(self, repo):
128 def __init__(self, repo):
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130
130
131 def branches(self, nodes):
131 def branches(self, nodes):
132 return self._repo.branches(nodes)
132 return self._repo.branches(nodes)
133
133
134 def between(self, pairs):
134 def between(self, pairs):
135 return self._repo.between(pairs)
135 return self._repo.between(pairs)
136
136
137 def changegroup(self, basenodes, source):
137 def changegroup(self, basenodes, source):
138 return self._repo.changegroup(basenodes, source)
138 return self._repo.changegroup(basenodes, source)
139
139
140 def changegroupsubset(self, bases, heads, source):
140 def changegroupsubset(self, bases, heads, source):
141 return self._repo.changegroupsubset(bases, heads, source)
141 return self._repo.changegroupsubset(bases, heads, source)
142
142
143 class localrepository(object):
143 class localrepository(object):
144
144
145 supportedformats = set(('revlogv1', 'generaldelta'))
145 supportedformats = set(('revlogv1', 'generaldelta'))
146 supported = supportedformats | set(('store', 'fncache', 'shared',
146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 'dotencode'))
147 'dotencode'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
149 requirements = ['revlogv1']
149 requirements = ['revlogv1']
150 filtername = None
150 filtername = None
151
151
152 def _baserequirements(self, create):
152 def _baserequirements(self, create):
153 return self.requirements[:]
153 return self.requirements[:]
154
154
155 def __init__(self, baseui, path=None, create=False):
155 def __init__(self, baseui, path=None, create=False):
156 self.wvfs = scmutil.vfs(path, expand=True)
156 self.wvfs = scmutil.vfs(path, expand=True)
157 self.wopener = self.wvfs
157 self.wopener = self.wvfs
158 self.root = self.wvfs.base
158 self.root = self.wvfs.base
159 self.path = self.wvfs.join(".hg")
159 self.path = self.wvfs.join(".hg")
160 self.origroot = path
160 self.origroot = path
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 self.vfs = scmutil.vfs(self.path)
162 self.vfs = scmutil.vfs(self.path)
163 self.opener = self.vfs
163 self.opener = self.vfs
164 self.baseui = baseui
164 self.baseui = baseui
165 self.ui = baseui.copy()
165 self.ui = baseui.copy()
166 # A list of callback to shape the phase if no data were found.
166 # A list of callback to shape the phase if no data were found.
167 # Callback are in the form: func(repo, roots) --> processed root.
167 # Callback are in the form: func(repo, roots) --> processed root.
168 # This list it to be filled by extension during repo setup
168 # This list it to be filled by extension during repo setup
169 self._phasedefaults = []
169 self._phasedefaults = []
170 try:
170 try:
171 self.ui.readconfig(self.join("hgrc"), self.root)
171 self.ui.readconfig(self.join("hgrc"), self.root)
172 extensions.loadall(self.ui)
172 extensions.loadall(self.ui)
173 except IOError:
173 except IOError:
174 pass
174 pass
175
175
176 if not self.vfs.isdir():
176 if not self.vfs.isdir():
177 if create:
177 if create:
178 if not self.wvfs.exists():
178 if not self.wvfs.exists():
179 self.wvfs.makedirs()
179 self.wvfs.makedirs()
180 self.vfs.makedir(notindexed=True)
180 self.vfs.makedir(notindexed=True)
181 requirements = self._baserequirements(create)
181 requirements = self._baserequirements(create)
182 if self.ui.configbool('format', 'usestore', True):
182 if self.ui.configbool('format', 'usestore', True):
183 self.vfs.mkdir("store")
183 self.vfs.mkdir("store")
184 requirements.append("store")
184 requirements.append("store")
185 if self.ui.configbool('format', 'usefncache', True):
185 if self.ui.configbool('format', 'usefncache', True):
186 requirements.append("fncache")
186 requirements.append("fncache")
187 if self.ui.configbool('format', 'dotencode', True):
187 if self.ui.configbool('format', 'dotencode', True):
188 requirements.append('dotencode')
188 requirements.append('dotencode')
189 # create an invalid changelog
189 # create an invalid changelog
190 self.vfs.append(
190 self.vfs.append(
191 "00changelog.i",
191 "00changelog.i",
192 '\0\0\0\2' # represents revlogv2
192 '\0\0\0\2' # represents revlogv2
193 ' dummy changelog to prevent using the old repo layout'
193 ' dummy changelog to prevent using the old repo layout'
194 )
194 )
195 if self.ui.configbool('format', 'generaldelta', False):
195 if self.ui.configbool('format', 'generaldelta', False):
196 requirements.append("generaldelta")
196 requirements.append("generaldelta")
197 requirements = set(requirements)
197 requirements = set(requirements)
198 else:
198 else:
199 raise error.RepoError(_("repository %s not found") % path)
199 raise error.RepoError(_("repository %s not found") % path)
200 elif create:
200 elif create:
201 raise error.RepoError(_("repository %s already exists") % path)
201 raise error.RepoError(_("repository %s already exists") % path)
202 else:
202 else:
203 try:
203 try:
204 requirements = scmutil.readrequires(self.vfs, self.supported)
204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 except IOError, inst:
205 except IOError, inst:
206 if inst.errno != errno.ENOENT:
206 if inst.errno != errno.ENOENT:
207 raise
207 raise
208 requirements = set()
208 requirements = set()
209
209
210 self.sharedpath = self.path
210 self.sharedpath = self.path
211 try:
211 try:
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 if not os.path.exists(s):
213 if not os.path.exists(s):
214 raise error.RepoError(
214 raise error.RepoError(
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 self.sharedpath = s
216 self.sharedpath = s
217 except IOError, inst:
217 except IOError, inst:
218 if inst.errno != errno.ENOENT:
218 if inst.errno != errno.ENOENT:
219 raise
219 raise
220
220
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 self.spath = self.store.path
222 self.spath = self.store.path
223 self.svfs = self.store.vfs
223 self.svfs = self.store.vfs
224 self.sopener = self.svfs
224 self.sopener = self.svfs
225 self.sjoin = self.store.join
225 self.sjoin = self.store.join
226 self.vfs.createmode = self.store.createmode
226 self.vfs.createmode = self.store.createmode
227 self._applyrequirements(requirements)
227 self._applyrequirements(requirements)
228 if create:
228 if create:
229 self._writerequirements()
229 self._writerequirements()
230
230
231
231
232 self._branchcaches = {}
232 self._branchcaches = {}
233 self.filterpats = {}
233 self.filterpats = {}
234 self._datafilters = {}
234 self._datafilters = {}
235 self._transref = self._lockref = self._wlockref = None
235 self._transref = self._lockref = self._wlockref = None
236
236
237 # A cache for various files under .hg/ that tracks file changes,
237 # A cache for various files under .hg/ that tracks file changes,
238 # (used by the filecache decorator)
238 # (used by the filecache decorator)
239 #
239 #
240 # Maps a property name to its util.filecacheentry
240 # Maps a property name to its util.filecacheentry
241 self._filecache = {}
241 self._filecache = {}
242
242
243 # hold sets of revision to be filtered
243 # hold sets of revision to be filtered
244 # should be cleared when something might have changed the filter value:
244 # should be cleared when something might have changed the filter value:
245 # - new changesets,
245 # - new changesets,
246 # - phase change,
246 # - phase change,
247 # - new obsolescence marker,
247 # - new obsolescence marker,
248 # - working directory parent change,
248 # - working directory parent change,
249 # - bookmark changes
249 # - bookmark changes
250 self.filteredrevcache = {}
250 self.filteredrevcache = {}
251
251
252 def close(self):
252 def close(self):
253 pass
253 pass
254
254
255 def _restrictcapabilities(self, caps):
255 def _restrictcapabilities(self, caps):
256 return caps
256 return caps
257
257
258 def _applyrequirements(self, requirements):
258 def _applyrequirements(self, requirements):
259 self.requirements = requirements
259 self.requirements = requirements
260 self.sopener.options = dict((r, 1) for r in requirements
260 self.sopener.options = dict((r, 1) for r in requirements
261 if r in self.openerreqs)
261 if r in self.openerreqs)
262
262
263 def _writerequirements(self):
263 def _writerequirements(self):
264 reqfile = self.opener("requires", "w")
264 reqfile = self.opener("requires", "w")
265 for r in self.requirements:
265 for r in self.requirements:
266 reqfile.write("%s\n" % r)
266 reqfile.write("%s\n" % r)
267 reqfile.close()
267 reqfile.close()
268
268
269 def _checknested(self, path):
269 def _checknested(self, path):
270 """Determine if path is a legal nested repository."""
270 """Determine if path is a legal nested repository."""
271 if not path.startswith(self.root):
271 if not path.startswith(self.root):
272 return False
272 return False
273 subpath = path[len(self.root) + 1:]
273 subpath = path[len(self.root) + 1:]
274 normsubpath = util.pconvert(subpath)
274 normsubpath = util.pconvert(subpath)
275
275
276 # XXX: Checking against the current working copy is wrong in
276 # XXX: Checking against the current working copy is wrong in
277 # the sense that it can reject things like
277 # the sense that it can reject things like
278 #
278 #
279 # $ hg cat -r 10 sub/x.txt
279 # $ hg cat -r 10 sub/x.txt
280 #
280 #
281 # if sub/ is no longer a subrepository in the working copy
281 # if sub/ is no longer a subrepository in the working copy
282 # parent revision.
282 # parent revision.
283 #
283 #
284 # However, it can of course also allow things that would have
284 # However, it can of course also allow things that would have
285 # been rejected before, such as the above cat command if sub/
285 # been rejected before, such as the above cat command if sub/
286 # is a subrepository now, but was a normal directory before.
286 # is a subrepository now, but was a normal directory before.
287 # The old path auditor would have rejected by mistake since it
287 # The old path auditor would have rejected by mistake since it
288 # panics when it sees sub/.hg/.
288 # panics when it sees sub/.hg/.
289 #
289 #
290 # All in all, checking against the working copy seems sensible
290 # All in all, checking against the working copy seems sensible
291 # since we want to prevent access to nested repositories on
291 # since we want to prevent access to nested repositories on
292 # the filesystem *now*.
292 # the filesystem *now*.
293 ctx = self[None]
293 ctx = self[None]
294 parts = util.splitpath(subpath)
294 parts = util.splitpath(subpath)
295 while parts:
295 while parts:
296 prefix = '/'.join(parts)
296 prefix = '/'.join(parts)
297 if prefix in ctx.substate:
297 if prefix in ctx.substate:
298 if prefix == normsubpath:
298 if prefix == normsubpath:
299 return True
299 return True
300 else:
300 else:
301 sub = ctx.sub(prefix)
301 sub = ctx.sub(prefix)
302 return sub.checknested(subpath[len(prefix) + 1:])
302 return sub.checknested(subpath[len(prefix) + 1:])
303 else:
303 else:
304 parts.pop()
304 parts.pop()
305 return False
305 return False
306
306
307 def peer(self):
307 def peer(self):
308 return localpeer(self) # not cached to avoid reference cycle
308 return localpeer(self) # not cached to avoid reference cycle
309
309
310 def unfiltered(self):
310 def unfiltered(self):
311 """Return unfiltered version of the repository
311 """Return unfiltered version of the repository
312
312
313 Intended to be ovewritten by filtered repo."""
313 Intended to be ovewritten by filtered repo."""
314 return self
314 return self
315
315
316 def filtered(self, name):
316 def filtered(self, name):
317 """Return a filtered version of a repository"""
317 """Return a filtered version of a repository"""
318 # build a new class with the mixin and the current class
318 # build a new class with the mixin and the current class
319 # (possibily subclass of the repo)
319 # (possibily subclass of the repo)
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 pass
321 pass
322 return proxycls(self, name)
322 return proxycls(self, name)
323
323
324 @repofilecache('bookmarks')
324 @repofilecache('bookmarks')
325 def _bookmarks(self):
325 def _bookmarks(self):
326 return bookmarks.bmstore(self)
326 return bookmarks.bmstore(self)
327
327
328 @repofilecache('bookmarks.current')
328 @repofilecache('bookmarks.current')
329 def _bookmarkcurrent(self):
329 def _bookmarkcurrent(self):
330 return bookmarks.readcurrent(self)
330 return bookmarks.readcurrent(self)
331
331
332 def bookmarkheads(self, bookmark):
332 def bookmarkheads(self, bookmark):
333 name = bookmark.split('@', 1)[0]
333 name = bookmark.split('@', 1)[0]
334 heads = []
334 heads = []
335 for mark, n in self._bookmarks.iteritems():
335 for mark, n in self._bookmarks.iteritems():
336 if mark.split('@', 1)[0] == name:
336 if mark.split('@', 1)[0] == name:
337 heads.append(n)
337 heads.append(n)
338 return heads
338 return heads
339
339
340 @storecache('phaseroots')
340 @storecache('phaseroots')
341 def _phasecache(self):
341 def _phasecache(self):
342 return phases.phasecache(self, self._phasedefaults)
342 return phases.phasecache(self, self._phasedefaults)
343
343
344 @storecache('obsstore')
344 @storecache('obsstore')
345 def obsstore(self):
345 def obsstore(self):
346 store = obsolete.obsstore(self.sopener)
346 store = obsolete.obsstore(self.sopener)
347 if store and not obsolete._enabled:
347 if store and not obsolete._enabled:
348 # message is rare enough to not be translated
348 # message is rare enough to not be translated
349 msg = 'obsolete feature not enabled but %i markers found!\n'
349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 self.ui.warn(msg % len(list(store)))
350 self.ui.warn(msg % len(list(store)))
351 return store
351 return store
352
352
353 @storecache('00changelog.i')
353 @storecache('00changelog.i')
354 def changelog(self):
354 def changelog(self):
355 c = changelog.changelog(self.sopener)
355 c = changelog.changelog(self.sopener)
356 if 'HG_PENDING' in os.environ:
356 if 'HG_PENDING' in os.environ:
357 p = os.environ['HG_PENDING']
357 p = os.environ['HG_PENDING']
358 if p.startswith(self.root):
358 if p.startswith(self.root):
359 c.readpending('00changelog.i.a')
359 c.readpending('00changelog.i.a')
360 return c
360 return c
361
361
362 @storecache('00manifest.i')
362 @storecache('00manifest.i')
363 def manifest(self):
363 def manifest(self):
364 return manifest.manifest(self.sopener)
364 return manifest.manifest(self.sopener)
365
365
366 @repofilecache('dirstate')
366 @repofilecache('dirstate')
367 def dirstate(self):
367 def dirstate(self):
368 warned = [0]
368 warned = [0]
369 def validate(node):
369 def validate(node):
370 try:
370 try:
371 self.changelog.rev(node)
371 self.changelog.rev(node)
372 return node
372 return node
373 except error.LookupError:
373 except error.LookupError:
374 if not warned[0]:
374 if not warned[0]:
375 warned[0] = True
375 warned[0] = True
376 self.ui.warn(_("warning: ignoring unknown"
376 self.ui.warn(_("warning: ignoring unknown"
377 " working parent %s!\n") % short(node))
377 " working parent %s!\n") % short(node))
378 return nullid
378 return nullid
379
379
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381
381
382 def __getitem__(self, changeid):
382 def __getitem__(self, changeid):
383 if changeid is None:
383 if changeid is None:
384 return context.workingctx(self)
384 return context.workingctx(self)
385 return context.changectx(self, changeid)
385 return context.changectx(self, changeid)
386
386
387 def __contains__(self, changeid):
387 def __contains__(self, changeid):
388 try:
388 try:
389 return bool(self.lookup(changeid))
389 return bool(self.lookup(changeid))
390 except error.RepoLookupError:
390 except error.RepoLookupError:
391 return False
391 return False
392
392
393 def __nonzero__(self):
393 def __nonzero__(self):
394 return True
394 return True
395
395
396 def __len__(self):
396 def __len__(self):
397 return len(self.changelog)
397 return len(self.changelog)
398
398
399 def __iter__(self):
399 def __iter__(self):
400 return iter(self.changelog)
400 return iter(self.changelog)
401
401
402 def revs(self, expr, *args):
402 def revs(self, expr, *args):
403 '''Return a list of revisions matching the given revset'''
403 '''Return a list of revisions matching the given revset'''
404 expr = revset.formatspec(expr, *args)
404 expr = revset.formatspec(expr, *args)
405 m = revset.match(None, expr)
405 m = revset.match(None, expr)
406 return [r for r in m(self, list(self))]
406 return [r for r in m(self, list(self))]
407
407
408 def set(self, expr, *args):
408 def set(self, expr, *args):
409 '''
409 '''
410 Yield a context for each matching revision, after doing arg
410 Yield a context for each matching revision, after doing arg
411 replacement via revset.formatspec
411 replacement via revset.formatspec
412 '''
412 '''
413 for r in self.revs(expr, *args):
413 for r in self.revs(expr, *args):
414 yield self[r]
414 yield self[r]
415
415
416 def url(self):
416 def url(self):
417 return 'file:' + self.root
417 return 'file:' + self.root
418
418
419 def hook(self, name, throw=False, **args):
419 def hook(self, name, throw=False, **args):
420 return hook.hook(self.ui, self, name, throw, **args)
420 return hook.hook(self.ui, self, name, throw, **args)
421
421
422 @unfilteredmethod
422 @unfilteredmethod
423 def _tag(self, names, node, message, local, user, date, extra={}):
423 def _tag(self, names, node, message, local, user, date, extra={}):
424 if isinstance(names, str):
424 if isinstance(names, str):
425 names = (names,)
425 names = (names,)
426
426
427 branches = self.branchmap()
427 branches = self.branchmap()
428 for name in names:
428 for name in names:
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 local=local)
430 local=local)
431 if name in branches:
431 if name in branches:
432 self.ui.warn(_("warning: tag %s conflicts with existing"
432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 " branch name\n") % name)
433 " branch name\n") % name)
434
434
435 def writetags(fp, names, munge, prevtags):
435 def writetags(fp, names, munge, prevtags):
436 fp.seek(0, 2)
436 fp.seek(0, 2)
437 if prevtags and prevtags[-1] != '\n':
437 if prevtags and prevtags[-1] != '\n':
438 fp.write('\n')
438 fp.write('\n')
439 for name in names:
439 for name in names:
440 m = munge and munge(name) or name
440 m = munge and munge(name) or name
441 if (self._tagscache.tagtypes and
441 if (self._tagscache.tagtypes and
442 name in self._tagscache.tagtypes):
442 name in self._tagscache.tagtypes):
443 old = self.tags().get(name, nullid)
443 old = self.tags().get(name, nullid)
444 fp.write('%s %s\n' % (hex(old), m))
444 fp.write('%s %s\n' % (hex(old), m))
445 fp.write('%s %s\n' % (hex(node), m))
445 fp.write('%s %s\n' % (hex(node), m))
446 fp.close()
446 fp.close()
447
447
448 prevtags = ''
448 prevtags = ''
449 if local:
449 if local:
450 try:
450 try:
451 fp = self.opener('localtags', 'r+')
451 fp = self.opener('localtags', 'r+')
452 except IOError:
452 except IOError:
453 fp = self.opener('localtags', 'a')
453 fp = self.opener('localtags', 'a')
454 else:
454 else:
455 prevtags = fp.read()
455 prevtags = fp.read()
456
456
457 # local tags are stored in the current charset
457 # local tags are stored in the current charset
458 writetags(fp, names, None, prevtags)
458 writetags(fp, names, None, prevtags)
459 for name in names:
459 for name in names:
460 self.hook('tag', node=hex(node), tag=name, local=local)
460 self.hook('tag', node=hex(node), tag=name, local=local)
461 return
461 return
462
462
463 try:
463 try:
464 fp = self.wfile('.hgtags', 'rb+')
464 fp = self.wfile('.hgtags', 'rb+')
465 except IOError, e:
465 except IOError, e:
466 if e.errno != errno.ENOENT:
466 if e.errno != errno.ENOENT:
467 raise
467 raise
468 fp = self.wfile('.hgtags', 'ab')
468 fp = self.wfile('.hgtags', 'ab')
469 else:
469 else:
470 prevtags = fp.read()
470 prevtags = fp.read()
471
471
472 # committed tags are stored in UTF-8
472 # committed tags are stored in UTF-8
473 writetags(fp, names, encoding.fromlocal, prevtags)
473 writetags(fp, names, encoding.fromlocal, prevtags)
474
474
475 fp.close()
475 fp.close()
476
476
477 self.invalidatecaches()
477 self.invalidatecaches()
478
478
479 if '.hgtags' not in self.dirstate:
479 if '.hgtags' not in self.dirstate:
480 self[None].add(['.hgtags'])
480 self[None].add(['.hgtags'])
481
481
482 m = matchmod.exact(self.root, '', ['.hgtags'])
482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484
484
485 for name in names:
485 for name in names:
486 self.hook('tag', node=hex(node), tag=name, local=local)
486 self.hook('tag', node=hex(node), tag=name, local=local)
487
487
488 return tagnode
488 return tagnode
489
489
490 def tag(self, names, node, message, local, user, date):
490 def tag(self, names, node, message, local, user, date):
491 '''tag a revision with one or more symbolic names.
491 '''tag a revision with one or more symbolic names.
492
492
493 names is a list of strings or, when adding a single tag, names may be a
493 names is a list of strings or, when adding a single tag, names may be a
494 string.
494 string.
495
495
496 if local is True, the tags are stored in a per-repository file.
496 if local is True, the tags are stored in a per-repository file.
497 otherwise, they are stored in the .hgtags file, and a new
497 otherwise, they are stored in the .hgtags file, and a new
498 changeset is committed with the change.
498 changeset is committed with the change.
499
499
500 keyword arguments:
500 keyword arguments:
501
501
502 local: whether to store tags in non-version-controlled file
502 local: whether to store tags in non-version-controlled file
503 (default False)
503 (default False)
504
504
505 message: commit message to use if committing
505 message: commit message to use if committing
506
506
507 user: name of user to use if committing
507 user: name of user to use if committing
508
508
509 date: date tuple to use if committing'''
509 date: date tuple to use if committing'''
510
510
511 if not local:
511 if not local:
512 for x in self.status()[:5]:
512 for x in self.status()[:5]:
513 if '.hgtags' in x:
513 if '.hgtags' in x:
514 raise util.Abort(_('working copy of .hgtags is changed '
514 raise util.Abort(_('working copy of .hgtags is changed '
515 '(please commit .hgtags manually)'))
515 '(please commit .hgtags manually)'))
516
516
517 self.tags() # instantiate the cache
517 self.tags() # instantiate the cache
518 self._tag(names, node, message, local, user, date)
518 self._tag(names, node, message, local, user, date)
519
519
520 @filteredpropertycache
520 @filteredpropertycache
521 def _tagscache(self):
521 def _tagscache(self):
522 '''Returns a tagscache object that contains various tags related
522 '''Returns a tagscache object that contains various tags related
523 caches.'''
523 caches.'''
524
524
525 # This simplifies its cache management by having one decorated
525 # This simplifies its cache management by having one decorated
526 # function (this one) and the rest simply fetch things from it.
526 # function (this one) and the rest simply fetch things from it.
527 class tagscache(object):
527 class tagscache(object):
528 def __init__(self):
528 def __init__(self):
529 # These two define the set of tags for this repository. tags
529 # These two define the set of tags for this repository. tags
530 # maps tag name to node; tagtypes maps tag name to 'global' or
530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 # 'local'. (Global tags are defined by .hgtags across all
531 # 'local'. (Global tags are defined by .hgtags across all
532 # heads, and local tags are defined in .hg/localtags.)
532 # heads, and local tags are defined in .hg/localtags.)
533 # They constitute the in-memory cache of tags.
533 # They constitute the in-memory cache of tags.
534 self.tags = self.tagtypes = None
534 self.tags = self.tagtypes = None
535
535
536 self.nodetagscache = self.tagslist = None
536 self.nodetagscache = self.tagslist = None
537
537
538 cache = tagscache()
538 cache = tagscache()
539 cache.tags, cache.tagtypes = self._findtags()
539 cache.tags, cache.tagtypes = self._findtags()
540
540
541 return cache
541 return cache
542
542
543 def tags(self):
543 def tags(self):
544 '''return a mapping of tag to node'''
544 '''return a mapping of tag to node'''
545 t = {}
545 t = {}
546 if self.changelog.filteredrevs:
546 if self.changelog.filteredrevs:
547 tags, tt = self._findtags()
547 tags, tt = self._findtags()
548 else:
548 else:
549 tags = self._tagscache.tags
549 tags = self._tagscache.tags
550 for k, v in tags.iteritems():
550 for k, v in tags.iteritems():
551 try:
551 try:
552 # ignore tags to unknown nodes
552 # ignore tags to unknown nodes
553 self.changelog.rev(v)
553 self.changelog.rev(v)
554 t[k] = v
554 t[k] = v
555 except (error.LookupError, ValueError):
555 except (error.LookupError, ValueError):
556 pass
556 pass
557 return t
557 return t
558
558
559 def _findtags(self):
559 def _findtags(self):
560 '''Do the hard work of finding tags. Return a pair of dicts
560 '''Do the hard work of finding tags. Return a pair of dicts
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 maps tag name to a string like \'global\' or \'local\'.
562 maps tag name to a string like \'global\' or \'local\'.
563 Subclasses or extensions are free to add their own tags, but
563 Subclasses or extensions are free to add their own tags, but
564 should be aware that the returned dicts will be retained for the
564 should be aware that the returned dicts will be retained for the
565 duration of the localrepo object.'''
565 duration of the localrepo object.'''
566
566
567 # XXX what tagtype should subclasses/extensions use? Currently
567 # XXX what tagtype should subclasses/extensions use? Currently
568 # mq and bookmarks add tags, but do not set the tagtype at all.
568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 # Should each extension invent its own tag type? Should there
569 # Should each extension invent its own tag type? Should there
570 # be one tagtype for all such "virtual" tags? Or is the status
570 # be one tagtype for all such "virtual" tags? Or is the status
571 # quo fine?
571 # quo fine?
572
572
573 alltags = {} # map tag name to (node, hist)
573 alltags = {} # map tag name to (node, hist)
574 tagtypes = {}
574 tagtypes = {}
575
575
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578
578
579 # Build the return dicts. Have to re-encode tag names because
579 # Build the return dicts. Have to re-encode tag names because
580 # the tags module always uses UTF-8 (in order not to lose info
580 # the tags module always uses UTF-8 (in order not to lose info
581 # writing to the cache), but the rest of Mercurial wants them in
581 # writing to the cache), but the rest of Mercurial wants them in
582 # local encoding.
582 # local encoding.
583 tags = {}
583 tags = {}
584 for (name, (node, hist)) in alltags.iteritems():
584 for (name, (node, hist)) in alltags.iteritems():
585 if node != nullid:
585 if node != nullid:
586 tags[encoding.tolocal(name)] = node
586 tags[encoding.tolocal(name)] = node
587 tags['tip'] = self.changelog.tip()
587 tags['tip'] = self.changelog.tip()
588 tagtypes = dict([(encoding.tolocal(name), value)
588 tagtypes = dict([(encoding.tolocal(name), value)
589 for (name, value) in tagtypes.iteritems()])
589 for (name, value) in tagtypes.iteritems()])
590 return (tags, tagtypes)
590 return (tags, tagtypes)
591
591
592 def tagtype(self, tagname):
592 def tagtype(self, tagname):
593 '''
593 '''
594 return the type of the given tag. result can be:
594 return the type of the given tag. result can be:
595
595
596 'local' : a local tag
596 'local' : a local tag
597 'global' : a global tag
597 'global' : a global tag
598 None : tag does not exist
598 None : tag does not exist
599 '''
599 '''
600
600
601 return self._tagscache.tagtypes.get(tagname)
601 return self._tagscache.tagtypes.get(tagname)
602
602
603 def tagslist(self):
603 def tagslist(self):
604 '''return a list of tags ordered by revision'''
604 '''return a list of tags ordered by revision'''
605 if not self._tagscache.tagslist:
605 if not self._tagscache.tagslist:
606 l = []
606 l = []
607 for t, n in self.tags().iteritems():
607 for t, n in self.tags().iteritems():
608 r = self.changelog.rev(n)
608 r = self.changelog.rev(n)
609 l.append((r, t, n))
609 l.append((r, t, n))
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611
611
612 return self._tagscache.tagslist
612 return self._tagscache.tagslist
613
613
614 def nodetags(self, node):
614 def nodetags(self, node):
615 '''return the tags associated with a node'''
615 '''return the tags associated with a node'''
616 if not self._tagscache.nodetagscache:
616 if not self._tagscache.nodetagscache:
617 nodetagscache = {}
617 nodetagscache = {}
618 for t, n in self._tagscache.tags.iteritems():
618 for t, n in self._tagscache.tags.iteritems():
619 nodetagscache.setdefault(n, []).append(t)
619 nodetagscache.setdefault(n, []).append(t)
620 for tags in nodetagscache.itervalues():
620 for tags in nodetagscache.itervalues():
621 tags.sort()
621 tags.sort()
622 self._tagscache.nodetagscache = nodetagscache
622 self._tagscache.nodetagscache = nodetagscache
623 return self._tagscache.nodetagscache.get(node, [])
623 return self._tagscache.nodetagscache.get(node, [])
624
624
625 def nodebookmarks(self, node):
625 def nodebookmarks(self, node):
626 marks = []
626 marks = []
627 for bookmark, n in self._bookmarks.iteritems():
627 for bookmark, n in self._bookmarks.iteritems():
628 if n == node:
628 if n == node:
629 marks.append(bookmark)
629 marks.append(bookmark)
630 return sorted(marks)
630 return sorted(marks)
631
631
632 def branchmap(self):
632 def branchmap(self):
633 '''returns a dictionary {branch: [branchheads]}'''
633 '''returns a dictionary {branch: [branchheads]}'''
634 branchmap.updatecache(self)
634 branchmap.updatecache(self)
635 return self._branchcaches[self.filtername]
635 return self._branchcaches[self.filtername]
636
636
637
637
638 def _branchtip(self, heads):
638 def _branchtip(self, heads):
639 '''return the tipmost branch head in heads'''
639 '''return the tipmost branch head in heads'''
640 tip = heads[-1]
640 tip = heads[-1]
641 for h in reversed(heads):
641 for h in reversed(heads):
642 if not self[h].closesbranch():
642 if not self[h].closesbranch():
643 tip = h
643 tip = h
644 break
644 break
645 return tip
645 return tip
646
646
647 def branchtip(self, branch):
647 def branchtip(self, branch):
648 '''return the tip node for a given branch'''
648 '''return the tip node for a given branch'''
649 if branch not in self.branchmap():
649 if branch not in self.branchmap():
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 return self._branchtip(self.branchmap()[branch])
651 return self._branchtip(self.branchmap()[branch])
652
652
653 def branchtags(self):
653 def branchtags(self):
654 '''return a dict where branch names map to the tipmost head of
654 '''return a dict where branch names map to the tipmost head of
655 the branch, open heads come before closed'''
655 the branch, open heads come before closed'''
656 bt = {}
656 bt = {}
657 for bn, heads in self.branchmap().iteritems():
657 for bn, heads in self.branchmap().iteritems():
658 bt[bn] = self._branchtip(heads)
658 bt[bn] = self._branchtip(heads)
659 return bt
659 return bt
660
660
661 def lookup(self, key):
661 def lookup(self, key):
662 return self[key].node()
662 return self[key].node()
663
663
664 def lookupbranch(self, key, remote=None):
664 def lookupbranch(self, key, remote=None):
665 repo = remote or self
665 repo = remote or self
666 if key in repo.branchmap():
666 if key in repo.branchmap():
667 return key
667 return key
668
668
669 repo = (remote and remote.local()) and remote or self
669 repo = (remote and remote.local()) and remote or self
670 return repo[key].branch()
670 return repo[key].branch()
671
671
672 def known(self, nodes):
672 def known(self, nodes):
673 nm = self.changelog.nodemap
673 nm = self.changelog.nodemap
674 pc = self._phasecache
674 pc = self._phasecache
675 result = []
675 result = []
676 for n in nodes:
676 for n in nodes:
677 r = nm.get(n)
677 r = nm.get(n)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 result.append(resp)
679 result.append(resp)
680 return result
680 return result
681
681
682 def local(self):
682 def local(self):
683 return self
683 return self
684
684
685 def cancopy(self):
685 def cancopy(self):
686 return self.local() # so statichttprepo's override of local() works
686 return self.local() # so statichttprepo's override of local() works
687
687
688 def join(self, f):
688 def join(self, f):
689 return os.path.join(self.path, f)
689 return os.path.join(self.path, f)
690
690
691 def wjoin(self, f):
691 def wjoin(self, f):
692 return os.path.join(self.root, f)
692 return os.path.join(self.root, f)
693
693
694 def file(self, f):
694 def file(self, f):
695 if f[0] == '/':
695 if f[0] == '/':
696 f = f[1:]
696 f = f[1:]
697 return filelog.filelog(self.sopener, f)
697 return filelog.filelog(self.sopener, f)
698
698
699 def changectx(self, changeid):
699 def changectx(self, changeid):
700 return self[changeid]
700 return self[changeid]
701
701
702 def parents(self, changeid=None):
702 def parents(self, changeid=None):
703 '''get list of changectxs for parents of changeid'''
703 '''get list of changectxs for parents of changeid'''
704 return self[changeid].parents()
704 return self[changeid].parents()
705
705
706 def setparents(self, p1, p2=nullid):
706 def setparents(self, p1, p2=nullid):
707 copies = self.dirstate.setparents(p1, p2)
707 copies = self.dirstate.setparents(p1, p2)
708 if copies:
708 if copies:
709 # Adjust copy records, the dirstate cannot do it, it
709 # Adjust copy records, the dirstate cannot do it, it
710 # requires access to parents manifests. Preserve them
710 # requires access to parents manifests. Preserve them
711 # only for entries added to first parent.
711 # only for entries added to first parent.
712 pctx = self[p1]
712 pctx = self[p1]
713 for f in copies:
713 for f in copies:
714 if f not in pctx and copies[f] in pctx:
714 if f not in pctx and copies[f] in pctx:
715 self.dirstate.copy(copies[f], f)
715 self.dirstate.copy(copies[f], f)
716
716
717 def filectx(self, path, changeid=None, fileid=None):
717 def filectx(self, path, changeid=None, fileid=None):
718 """changeid can be a changeset revision, node, or tag.
718 """changeid can be a changeset revision, node, or tag.
719 fileid can be a file revision or node."""
719 fileid can be a file revision or node."""
720 return context.filectx(self, path, changeid, fileid)
720 return context.filectx(self, path, changeid, fileid)
721
721
722 def getcwd(self):
722 def getcwd(self):
723 return self.dirstate.getcwd()
723 return self.dirstate.getcwd()
724
724
725 def pathto(self, f, cwd=None):
725 def pathto(self, f, cwd=None):
726 return self.dirstate.pathto(f, cwd)
726 return self.dirstate.pathto(f, cwd)
727
727
728 def wfile(self, f, mode='r'):
728 def wfile(self, f, mode='r'):
729 return self.wopener(f, mode)
729 return self.wopener(f, mode)
730
730
731 def _link(self, f):
731 def _link(self, f):
732 return os.path.islink(self.wjoin(f))
732 return os.path.islink(self.wjoin(f))
733
733
734 def _loadfilter(self, filter):
734 def _loadfilter(self, filter):
735 if filter not in self.filterpats:
735 if filter not in self.filterpats:
736 l = []
736 l = []
737 for pat, cmd in self.ui.configitems(filter):
737 for pat, cmd in self.ui.configitems(filter):
738 if cmd == '!':
738 if cmd == '!':
739 continue
739 continue
740 mf = matchmod.match(self.root, '', [pat])
740 mf = matchmod.match(self.root, '', [pat])
741 fn = None
741 fn = None
742 params = cmd
742 params = cmd
743 for name, filterfn in self._datafilters.iteritems():
743 for name, filterfn in self._datafilters.iteritems():
744 if cmd.startswith(name):
744 if cmd.startswith(name):
745 fn = filterfn
745 fn = filterfn
746 params = cmd[len(name):].lstrip()
746 params = cmd[len(name):].lstrip()
747 break
747 break
748 if not fn:
748 if not fn:
749 fn = lambda s, c, **kwargs: util.filter(s, c)
749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 # Wrap old filters not supporting keyword arguments
750 # Wrap old filters not supporting keyword arguments
751 if not inspect.getargspec(fn)[2]:
751 if not inspect.getargspec(fn)[2]:
752 oldfn = fn
752 oldfn = fn
753 fn = lambda s, c, **kwargs: oldfn(s, c)
753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 l.append((mf, fn, params))
754 l.append((mf, fn, params))
755 self.filterpats[filter] = l
755 self.filterpats[filter] = l
756 return self.filterpats[filter]
756 return self.filterpats[filter]
757
757
758 def _filter(self, filterpats, filename, data):
758 def _filter(self, filterpats, filename, data):
759 for mf, fn, cmd in filterpats:
759 for mf, fn, cmd in filterpats:
760 if mf(filename):
760 if mf(filename):
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 break
763 break
764
764
765 return data
765 return data
766
766
767 @unfilteredpropertycache
767 @unfilteredpropertycache
768 def _encodefilterpats(self):
768 def _encodefilterpats(self):
769 return self._loadfilter('encode')
769 return self._loadfilter('encode')
770
770
771 @unfilteredpropertycache
771 @unfilteredpropertycache
772 def _decodefilterpats(self):
772 def _decodefilterpats(self):
773 return self._loadfilter('decode')
773 return self._loadfilter('decode')
774
774
775 def adddatafilter(self, name, filter):
775 def adddatafilter(self, name, filter):
776 self._datafilters[name] = filter
776 self._datafilters[name] = filter
777
777
778 def wread(self, filename):
778 def wread(self, filename):
779 if self._link(filename):
779 if self._link(filename):
780 data = os.readlink(self.wjoin(filename))
780 data = os.readlink(self.wjoin(filename))
781 else:
781 else:
782 data = self.wopener.read(filename)
782 data = self.wopener.read(filename)
783 return self._filter(self._encodefilterpats, filename, data)
783 return self._filter(self._encodefilterpats, filename, data)
784
784
785 def wwrite(self, filename, data, flags):
785 def wwrite(self, filename, data, flags):
786 data = self._filter(self._decodefilterpats, filename, data)
786 data = self._filter(self._decodefilterpats, filename, data)
787 if 'l' in flags:
787 if 'l' in flags:
788 self.wopener.symlink(data, filename)
788 self.wopener.symlink(data, filename)
789 else:
789 else:
790 self.wopener.write(filename, data)
790 self.wopener.write(filename, data)
791 if 'x' in flags:
791 if 'x' in flags:
792 util.setflags(self.wjoin(filename), False, True)
792 util.setflags(self.wjoin(filename), False, True)
793
793
794 def wwritedata(self, filename, data):
794 def wwritedata(self, filename, data):
795 return self._filter(self._decodefilterpats, filename, data)
795 return self._filter(self._decodefilterpats, filename, data)
796
796
797 def transaction(self, desc):
797 def transaction(self, desc):
798 tr = self._transref and self._transref() or None
798 tr = self._transref and self._transref() or None
799 if tr and tr.running():
799 if tr and tr.running():
800 return tr.nest()
800 return tr.nest()
801
801
802 # abort here if the journal already exists
802 # abort here if the journal already exists
803 if os.path.exists(self.sjoin("journal")):
803 if os.path.exists(self.sjoin("journal")):
804 raise error.RepoError(
804 raise error.RepoError(
805 _("abandoned transaction found - run hg recover"))
805 _("abandoned transaction found - run hg recover"))
806
806
807 self._writejournal(desc)
807 self._writejournal(desc)
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809
809
810 tr = transaction.transaction(self.ui.warn, self.sopener,
810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 self.sjoin("journal"),
811 self.sjoin("journal"),
812 aftertrans(renames),
812 aftertrans(renames),
813 self.store.createmode)
813 self.store.createmode)
814 self._transref = weakref.ref(tr)
814 self._transref = weakref.ref(tr)
815 return tr
815 return tr
816
816
817 def _journalfiles(self):
817 def _journalfiles(self):
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 self.join('journal.branch'), self.join('journal.desc'),
819 self.join('journal.branch'), self.join('journal.desc'),
820 self.join('journal.bookmarks'),
820 self.join('journal.bookmarks'),
821 self.sjoin('journal.phaseroots'))
821 self.sjoin('journal.phaseroots'))
822
822
823 def undofiles(self):
823 def undofiles(self):
824 return [undoname(x) for x in self._journalfiles()]
824 return [undoname(x) for x in self._journalfiles()]
825
825
826 def _writejournal(self, desc):
826 def _writejournal(self, desc):
827 self.opener.write("journal.dirstate",
827 self.opener.write("journal.dirstate",
828 self.opener.tryread("dirstate"))
828 self.opener.tryread("dirstate"))
829 self.opener.write("journal.branch",
829 self.opener.write("journal.branch",
830 encoding.fromlocal(self.dirstate.branch()))
830 encoding.fromlocal(self.dirstate.branch()))
831 self.opener.write("journal.desc",
831 self.opener.write("journal.desc",
832 "%d\n%s\n" % (len(self), desc))
832 "%d\n%s\n" % (len(self), desc))
833 self.opener.write("journal.bookmarks",
833 self.opener.write("journal.bookmarks",
834 self.opener.tryread("bookmarks"))
834 self.opener.tryread("bookmarks"))
835 self.sopener.write("journal.phaseroots",
835 self.sopener.write("journal.phaseroots",
836 self.sopener.tryread("phaseroots"))
836 self.sopener.tryread("phaseroots"))
837
837
838 def recover(self):
838 def recover(self):
839 lock = self.lock()
839 lock = self.lock()
840 try:
840 try:
841 if os.path.exists(self.sjoin("journal")):
841 if os.path.exists(self.sjoin("journal")):
842 self.ui.status(_("rolling back interrupted transaction\n"))
842 self.ui.status(_("rolling back interrupted transaction\n"))
843 transaction.rollback(self.sopener, self.sjoin("journal"),
843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 self.ui.warn)
844 self.ui.warn)
845 self.invalidate()
845 self.invalidate()
846 return True
846 return True
847 else:
847 else:
848 self.ui.warn(_("no interrupted transaction available\n"))
848 self.ui.warn(_("no interrupted transaction available\n"))
849 return False
849 return False
850 finally:
850 finally:
851 lock.release()
851 lock.release()
852
852
853 def rollback(self, dryrun=False, force=False):
853 def rollback(self, dryrun=False, force=False):
854 wlock = lock = None
854 wlock = lock = None
855 try:
855 try:
856 wlock = self.wlock()
856 wlock = self.wlock()
857 lock = self.lock()
857 lock = self.lock()
858 if os.path.exists(self.sjoin("undo")):
858 if os.path.exists(self.sjoin("undo")):
859 return self._rollback(dryrun, force)
859 return self._rollback(dryrun, force)
860 else:
860 else:
861 self.ui.warn(_("no rollback information available\n"))
861 self.ui.warn(_("no rollback information available\n"))
862 return 1
862 return 1
863 finally:
863 finally:
864 release(lock, wlock)
864 release(lock, wlock)
865
865
866 @unfilteredmethod # Until we get smarter cache management
866 @unfilteredmethod # Until we get smarter cache management
867 def _rollback(self, dryrun, force):
867 def _rollback(self, dryrun, force):
868 ui = self.ui
868 ui = self.ui
869 try:
869 try:
870 args = self.opener.read('undo.desc').splitlines()
870 args = self.opener.read('undo.desc').splitlines()
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 if len(args) >= 3:
872 if len(args) >= 3:
873 detail = args[2]
873 detail = args[2]
874 oldtip = oldlen - 1
874 oldtip = oldlen - 1
875
875
876 if detail and ui.verbose:
876 if detail and ui.verbose:
877 msg = (_('repository tip rolled back to revision %s'
877 msg = (_('repository tip rolled back to revision %s'
878 ' (undo %s: %s)\n')
878 ' (undo %s: %s)\n')
879 % (oldtip, desc, detail))
879 % (oldtip, desc, detail))
880 else:
880 else:
881 msg = (_('repository tip rolled back to revision %s'
881 msg = (_('repository tip rolled back to revision %s'
882 ' (undo %s)\n')
882 ' (undo %s)\n')
883 % (oldtip, desc))
883 % (oldtip, desc))
884 except IOError:
884 except IOError:
885 msg = _('rolling back unknown transaction\n')
885 msg = _('rolling back unknown transaction\n')
886 desc = None
886 desc = None
887
887
888 if not force and self['.'] != self['tip'] and desc == 'commit':
888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 raise util.Abort(
889 raise util.Abort(
890 _('rollback of last commit while not checked out '
890 _('rollback of last commit while not checked out '
891 'may lose data'), hint=_('use -f to force'))
891 'may lose data'), hint=_('use -f to force'))
892
892
893 ui.status(msg)
893 ui.status(msg)
894 if dryrun:
894 if dryrun:
895 return 0
895 return 0
896
896
897 parents = self.dirstate.parents()
897 parents = self.dirstate.parents()
898 self.destroying()
898 self.destroying()
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 if os.path.exists(self.join('undo.bookmarks')):
900 if os.path.exists(self.join('undo.bookmarks')):
901 util.rename(self.join('undo.bookmarks'),
901 util.rename(self.join('undo.bookmarks'),
902 self.join('bookmarks'))
902 self.join('bookmarks'))
903 if os.path.exists(self.sjoin('undo.phaseroots')):
903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 util.rename(self.sjoin('undo.phaseroots'),
904 util.rename(self.sjoin('undo.phaseroots'),
905 self.sjoin('phaseroots'))
905 self.sjoin('phaseroots'))
906 self.invalidate()
906 self.invalidate()
907
907
908 # Discard all cache entries to force reloading everything.
908 # Discard all cache entries to force reloading everything.
909 self._filecache.clear()
909 self._filecache.clear()
910
910
911 parentgone = (parents[0] not in self.changelog.nodemap or
911 parentgone = (parents[0] not in self.changelog.nodemap or
912 parents[1] not in self.changelog.nodemap)
912 parents[1] not in self.changelog.nodemap)
913 if parentgone:
913 if parentgone:
914 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
914 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
915 try:
915 try:
916 branch = self.opener.read('undo.branch')
916 branch = self.opener.read('undo.branch')
917 self.dirstate.setbranch(encoding.tolocal(branch))
917 self.dirstate.setbranch(encoding.tolocal(branch))
918 except IOError:
918 except IOError:
919 ui.warn(_('named branch could not be reset: '
919 ui.warn(_('named branch could not be reset: '
920 'current branch is still \'%s\'\n')
920 'current branch is still \'%s\'\n')
921 % self.dirstate.branch())
921 % self.dirstate.branch())
922
922
923 self.dirstate.invalidate()
923 self.dirstate.invalidate()
924 parents = tuple([p.rev() for p in self.parents()])
924 parents = tuple([p.rev() for p in self.parents()])
925 if len(parents) > 1:
925 if len(parents) > 1:
926 ui.status(_('working directory now based on '
926 ui.status(_('working directory now based on '
927 'revisions %d and %d\n') % parents)
927 'revisions %d and %d\n') % parents)
928 else:
928 else:
929 ui.status(_('working directory now based on '
929 ui.status(_('working directory now based on '
930 'revision %d\n') % parents)
930 'revision %d\n') % parents)
931 # TODO: if we know which new heads may result from this rollback, pass
931 # TODO: if we know which new heads may result from this rollback, pass
932 # them to destroy(), which will prevent the branchhead cache from being
932 # them to destroy(), which will prevent the branchhead cache from being
933 # invalidated.
933 # invalidated.
934 self.destroyed()
934 self.destroyed()
935 return 0
935 return 0
936
936
937 def invalidatecaches(self):
937 def invalidatecaches(self):
938
938
939 if '_tagscache' in vars(self):
939 if '_tagscache' in vars(self):
940 # can't use delattr on proxy
940 # can't use delattr on proxy
941 del self.__dict__['_tagscache']
941 del self.__dict__['_tagscache']
942
942
943 self.unfiltered()._branchcaches.clear()
943 self.unfiltered()._branchcaches.clear()
944 self.invalidatevolatilesets()
944 self.invalidatevolatilesets()
945
945
946 def invalidatevolatilesets(self):
946 def invalidatevolatilesets(self):
947 self.filteredrevcache.clear()
947 self.filteredrevcache.clear()
948 obsolete.clearobscaches(self)
948 obsolete.clearobscaches(self)
949
949
950 def invalidatedirstate(self):
950 def invalidatedirstate(self):
951 '''Invalidates the dirstate, causing the next call to dirstate
951 '''Invalidates the dirstate, causing the next call to dirstate
952 to check if it was modified since the last time it was read,
952 to check if it was modified since the last time it was read,
953 rereading it if it has.
953 rereading it if it has.
954
954
955 This is different to dirstate.invalidate() that it doesn't always
955 This is different to dirstate.invalidate() that it doesn't always
956 rereads the dirstate. Use dirstate.invalidate() if you want to
956 rereads the dirstate. Use dirstate.invalidate() if you want to
957 explicitly read the dirstate again (i.e. restoring it to a previous
957 explicitly read the dirstate again (i.e. restoring it to a previous
958 known good state).'''
958 known good state).'''
959 if hasunfilteredcache(self, 'dirstate'):
959 if hasunfilteredcache(self, 'dirstate'):
960 for k in self.dirstate._filecache:
960 for k in self.dirstate._filecache:
961 try:
961 try:
962 delattr(self.dirstate, k)
962 delattr(self.dirstate, k)
963 except AttributeError:
963 except AttributeError:
964 pass
964 pass
965 delattr(self.unfiltered(), 'dirstate')
965 delattr(self.unfiltered(), 'dirstate')
966
966
967 def invalidate(self):
967 def invalidate(self):
968 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
968 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
969 for k in self._filecache:
969 for k in self._filecache:
970 # dirstate is invalidated separately in invalidatedirstate()
970 # dirstate is invalidated separately in invalidatedirstate()
971 if k == 'dirstate':
971 if k == 'dirstate':
972 continue
972 continue
973
973
974 try:
974 try:
975 delattr(unfiltered, k)
975 delattr(unfiltered, k)
976 except AttributeError:
976 except AttributeError:
977 pass
977 pass
978 self.invalidatecaches()
978 self.invalidatecaches()
979
979
980 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
980 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
981 try:
981 try:
982 l = lock.lock(lockname, 0, releasefn, desc=desc)
982 l = lock.lock(lockname, 0, releasefn, desc=desc)
983 except error.LockHeld, inst:
983 except error.LockHeld, inst:
984 if not wait:
984 if not wait:
985 raise
985 raise
986 self.ui.warn(_("waiting for lock on %s held by %r\n") %
986 self.ui.warn(_("waiting for lock on %s held by %r\n") %
987 (desc, inst.locker))
987 (desc, inst.locker))
988 # default to 600 seconds timeout
988 # default to 600 seconds timeout
989 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
989 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
990 releasefn, desc=desc)
990 releasefn, desc=desc)
991 if acquirefn:
991 if acquirefn:
992 acquirefn()
992 acquirefn()
993 return l
993 return l
994
994
995 def _afterlock(self, callback):
995 def _afterlock(self, callback):
996 """add a callback to the current repository lock.
996 """add a callback to the current repository lock.
997
997
998 The callback will be executed on lock release."""
998 The callback will be executed on lock release."""
999 l = self._lockref and self._lockref()
999 l = self._lockref and self._lockref()
1000 if l:
1000 if l:
1001 l.postrelease.append(callback)
1001 l.postrelease.append(callback)
1002 else:
1002 else:
1003 callback()
1003 callback()
1004
1004
1005 def lock(self, wait=True):
1005 def lock(self, wait=True):
1006 '''Lock the repository store (.hg/store) and return a weak reference
1006 '''Lock the repository store (.hg/store) and return a weak reference
1007 to the lock. Use this before modifying the store (e.g. committing or
1007 to the lock. Use this before modifying the store (e.g. committing or
1008 stripping). If you are opening a transaction, get a lock as well.)'''
1008 stripping). If you are opening a transaction, get a lock as well.)'''
1009 l = self._lockref and self._lockref()
1009 l = self._lockref and self._lockref()
1010 if l is not None and l.held:
1010 if l is not None and l.held:
1011 l.lock()
1011 l.lock()
1012 return l
1012 return l
1013
1013
1014 def unlock():
1014 def unlock():
1015 self.store.write()
1015 self.store.write()
1016 if hasunfilteredcache(self, '_phasecache'):
1016 if hasunfilteredcache(self, '_phasecache'):
1017 self._phasecache.write()
1017 self._phasecache.write()
1018 for k, ce in self._filecache.items():
1018 for k, ce in self._filecache.items():
1019 if k == 'dirstate' or k not in self.__dict__:
1019 if k == 'dirstate' or k not in self.__dict__:
1020 continue
1020 continue
1021 ce.refresh()
1021 ce.refresh()
1022
1022
1023 l = self._lock(self.sjoin("lock"), wait, unlock,
1023 l = self._lock(self.sjoin("lock"), wait, unlock,
1024 self.invalidate, _('repository %s') % self.origroot)
1024 self.invalidate, _('repository %s') % self.origroot)
1025 self._lockref = weakref.ref(l)
1025 self._lockref = weakref.ref(l)
1026 return l
1026 return l
1027
1027
1028 def wlock(self, wait=True):
1028 def wlock(self, wait=True):
1029 '''Lock the non-store parts of the repository (everything under
1029 '''Lock the non-store parts of the repository (everything under
1030 .hg except .hg/store) and return a weak reference to the lock.
1030 .hg except .hg/store) and return a weak reference to the lock.
1031 Use this before modifying files in .hg.'''
1031 Use this before modifying files in .hg.'''
1032 l = self._wlockref and self._wlockref()
1032 l = self._wlockref and self._wlockref()
1033 if l is not None and l.held:
1033 if l is not None and l.held:
1034 l.lock()
1034 l.lock()
1035 return l
1035 return l
1036
1036
1037 def unlock():
1037 def unlock():
1038 self.dirstate.write()
1038 self.dirstate.write()
1039 ce = self._filecache.get('dirstate')
1039 ce = self._filecache.get('dirstate')
1040 if ce:
1040 if ce:
1041 ce.refresh()
1041 ce.refresh()
1042
1042
1043 l = self._lock(self.join("wlock"), wait, unlock,
1043 l = self._lock(self.join("wlock"), wait, unlock,
1044 self.invalidatedirstate, _('working directory of %s') %
1044 self.invalidatedirstate, _('working directory of %s') %
1045 self.origroot)
1045 self.origroot)
1046 self._wlockref = weakref.ref(l)
1046 self._wlockref = weakref.ref(l)
1047 return l
1047 return l
1048
1048
1049 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1049 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1050 """
1050 """
1051 commit an individual file as part of a larger transaction
1051 commit an individual file as part of a larger transaction
1052 """
1052 """
1053
1053
1054 fname = fctx.path()
1054 fname = fctx.path()
1055 text = fctx.data()
1055 text = fctx.data()
1056 flog = self.file(fname)
1056 flog = self.file(fname)
1057 fparent1 = manifest1.get(fname, nullid)
1057 fparent1 = manifest1.get(fname, nullid)
1058 fparent2 = fparent2o = manifest2.get(fname, nullid)
1058 fparent2 = fparent2o = manifest2.get(fname, nullid)
1059
1059
1060 meta = {}
1060 meta = {}
1061 copy = fctx.renamed()
1061 copy = fctx.renamed()
1062 if copy and copy[0] != fname:
1062 if copy and copy[0] != fname:
1063 # Mark the new revision of this file as a copy of another
1063 # Mark the new revision of this file as a copy of another
1064 # file. This copy data will effectively act as a parent
1064 # file. This copy data will effectively act as a parent
1065 # of this new revision. If this is a merge, the first
1065 # of this new revision. If this is a merge, the first
1066 # parent will be the nullid (meaning "look up the copy data")
1066 # parent will be the nullid (meaning "look up the copy data")
1067 # and the second one will be the other parent. For example:
1067 # and the second one will be the other parent. For example:
1068 #
1068 #
1069 # 0 --- 1 --- 3 rev1 changes file foo
1069 # 0 --- 1 --- 3 rev1 changes file foo
1070 # \ / rev2 renames foo to bar and changes it
1070 # \ / rev2 renames foo to bar and changes it
1071 # \- 2 -/ rev3 should have bar with all changes and
1071 # \- 2 -/ rev3 should have bar with all changes and
1072 # should record that bar descends from
1072 # should record that bar descends from
1073 # bar in rev2 and foo in rev1
1073 # bar in rev2 and foo in rev1
1074 #
1074 #
1075 # this allows this merge to succeed:
1075 # this allows this merge to succeed:
1076 #
1076 #
1077 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1077 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1078 # \ / merging rev3 and rev4 should use bar@rev2
1078 # \ / merging rev3 and rev4 should use bar@rev2
1079 # \- 2 --- 4 as the merge base
1079 # \- 2 --- 4 as the merge base
1080 #
1080 #
1081
1081
1082 cfname = copy[0]
1082 cfname = copy[0]
1083 crev = manifest1.get(cfname)
1083 crev = manifest1.get(cfname)
1084 newfparent = fparent2
1084 newfparent = fparent2
1085
1085
1086 if manifest2: # branch merge
1086 if manifest2: # branch merge
1087 if fparent2 == nullid or crev is None: # copied on remote side
1087 if fparent2 == nullid or crev is None: # copied on remote side
1088 if cfname in manifest2:
1088 if cfname in manifest2:
1089 crev = manifest2[cfname]
1089 crev = manifest2[cfname]
1090 newfparent = fparent1
1090 newfparent = fparent1
1091
1091
1092 # find source in nearest ancestor if we've lost track
1092 # find source in nearest ancestor if we've lost track
1093 if not crev:
1093 if not crev:
1094 self.ui.debug(" %s: searching for copy revision for %s\n" %
1094 self.ui.debug(" %s: searching for copy revision for %s\n" %
1095 (fname, cfname))
1095 (fname, cfname))
1096 for ancestor in self[None].ancestors():
1096 for ancestor in self[None].ancestors():
1097 if cfname in ancestor:
1097 if cfname in ancestor:
1098 crev = ancestor[cfname].filenode()
1098 crev = ancestor[cfname].filenode()
1099 break
1099 break
1100
1100
1101 if crev:
1101 if crev:
1102 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1102 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1103 meta["copy"] = cfname
1103 meta["copy"] = cfname
1104 meta["copyrev"] = hex(crev)
1104 meta["copyrev"] = hex(crev)
1105 fparent1, fparent2 = nullid, newfparent
1105 fparent1, fparent2 = nullid, newfparent
1106 else:
1106 else:
1107 self.ui.warn(_("warning: can't find ancestor for '%s' "
1107 self.ui.warn(_("warning: can't find ancestor for '%s' "
1108 "copied from '%s'!\n") % (fname, cfname))
1108 "copied from '%s'!\n") % (fname, cfname))
1109
1109
1110 elif fparent2 != nullid:
1110 elif fparent2 != nullid:
1111 # is one parent an ancestor of the other?
1111 # is one parent an ancestor of the other?
1112 fparentancestor = flog.ancestor(fparent1, fparent2)
1112 fparentancestor = flog.ancestor(fparent1, fparent2)
1113 if fparentancestor == fparent1:
1113 if fparentancestor == fparent1:
1114 fparent1, fparent2 = fparent2, nullid
1114 fparent1, fparent2 = fparent2, nullid
1115 elif fparentancestor == fparent2:
1115 elif fparentancestor == fparent2:
1116 fparent2 = nullid
1116 fparent2 = nullid
1117
1117
1118 # is the file changed?
1118 # is the file changed?
1119 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1119 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1120 changelist.append(fname)
1120 changelist.append(fname)
1121 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1121 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1122
1122
1123 # are just the flags changed during merge?
1123 # are just the flags changed during merge?
1124 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1124 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1125 changelist.append(fname)
1125 changelist.append(fname)
1126
1126
1127 return fparent1
1127 return fparent1
1128
1128
1129 @unfilteredmethod
1129 @unfilteredmethod
1130 def commit(self, text="", user=None, date=None, match=None, force=False,
1130 def commit(self, text="", user=None, date=None, match=None, force=False,
1131 editor=False, extra={}):
1131 editor=False, extra={}):
1132 """Add a new revision to current repository.
1132 """Add a new revision to current repository.
1133
1133
1134 Revision information is gathered from the working directory,
1134 Revision information is gathered from the working directory,
1135 match can be used to filter the committed files. If editor is
1135 match can be used to filter the committed files. If editor is
1136 supplied, it is called to get a commit message.
1136 supplied, it is called to get a commit message.
1137 """
1137 """
1138
1138
1139 def fail(f, msg):
1139 def fail(f, msg):
1140 raise util.Abort('%s: %s' % (f, msg))
1140 raise util.Abort('%s: %s' % (f, msg))
1141
1141
1142 if not match:
1142 if not match:
1143 match = matchmod.always(self.root, '')
1143 match = matchmod.always(self.root, '')
1144
1144
1145 if not force:
1145 if not force:
1146 vdirs = []
1146 vdirs = []
1147 match.dir = vdirs.append
1147 match.dir = vdirs.append
1148 match.bad = fail
1148 match.bad = fail
1149
1149
1150 wlock = self.wlock()
1150 wlock = self.wlock()
1151 try:
1151 try:
1152 wctx = self[None]
1152 wctx = self[None]
1153 merge = len(wctx.parents()) > 1
1153 merge = len(wctx.parents()) > 1
1154
1154
1155 if (not force and merge and match and
1155 if (not force and merge and match and
1156 (match.files() or match.anypats())):
1156 (match.files() or match.anypats())):
1157 raise util.Abort(_('cannot partially commit a merge '
1157 raise util.Abort(_('cannot partially commit a merge '
1158 '(do not specify files or patterns)'))
1158 '(do not specify files or patterns)'))
1159
1159
1160 changes = self.status(match=match, clean=force)
1160 changes = self.status(match=match, clean=force)
1161 if force:
1161 if force:
1162 changes[0].extend(changes[6]) # mq may commit unchanged files
1162 changes[0].extend(changes[6]) # mq may commit unchanged files
1163
1163
1164 # check subrepos
1164 # check subrepos
1165 subs = []
1165 subs = []
1166 commitsubs = set()
1166 commitsubs = set()
1167 newstate = wctx.substate.copy()
1167 newstate = wctx.substate.copy()
1168 # only manage subrepos and .hgsubstate if .hgsub is present
1168 # only manage subrepos and .hgsubstate if .hgsub is present
1169 if '.hgsub' in wctx:
1169 if '.hgsub' in wctx:
1170 # we'll decide whether to track this ourselves, thanks
1170 # we'll decide whether to track this ourselves, thanks
1171 if '.hgsubstate' in changes[0]:
1171 if '.hgsubstate' in changes[0]:
1172 changes[0].remove('.hgsubstate')
1172 changes[0].remove('.hgsubstate')
1173 if '.hgsubstate' in changes[2]:
1173 if '.hgsubstate' in changes[2]:
1174 changes[2].remove('.hgsubstate')
1174 changes[2].remove('.hgsubstate')
1175
1175
1176 # compare current state to last committed state
1176 # compare current state to last committed state
1177 # build new substate based on last committed state
1177 # build new substate based on last committed state
1178 oldstate = wctx.p1().substate
1178 oldstate = wctx.p1().substate
1179 for s in sorted(newstate.keys()):
1179 for s in sorted(newstate.keys()):
1180 if not match(s):
1180 if not match(s):
1181 # ignore working copy, use old state if present
1181 # ignore working copy, use old state if present
1182 if s in oldstate:
1182 if s in oldstate:
1183 newstate[s] = oldstate[s]
1183 newstate[s] = oldstate[s]
1184 continue
1184 continue
1185 if not force:
1185 if not force:
1186 raise util.Abort(
1186 raise util.Abort(
1187 _("commit with new subrepo %s excluded") % s)
1187 _("commit with new subrepo %s excluded") % s)
1188 if wctx.sub(s).dirty(True):
1188 if wctx.sub(s).dirty(True):
1189 if not self.ui.configbool('ui', 'commitsubrepos'):
1189 if not self.ui.configbool('ui', 'commitsubrepos'):
1190 raise util.Abort(
1190 raise util.Abort(
1191 _("uncommitted changes in subrepo %s") % s,
1191 _("uncommitted changes in subrepo %s") % s,
1192 hint=_("use --subrepos for recursive commit"))
1192 hint=_("use --subrepos for recursive commit"))
1193 subs.append(s)
1193 subs.append(s)
1194 commitsubs.add(s)
1194 commitsubs.add(s)
1195 else:
1195 else:
1196 bs = wctx.sub(s).basestate()
1196 bs = wctx.sub(s).basestate()
1197 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1197 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1198 if oldstate.get(s, (None, None, None))[1] != bs:
1198 if oldstate.get(s, (None, None, None))[1] != bs:
1199 subs.append(s)
1199 subs.append(s)
1200
1200
1201 # check for removed subrepos
1201 # check for removed subrepos
1202 for p in wctx.parents():
1202 for p in wctx.parents():
1203 r = [s for s in p.substate if s not in newstate]
1203 r = [s for s in p.substate if s not in newstate]
1204 subs += [s for s in r if match(s)]
1204 subs += [s for s in r if match(s)]
1205 if subs:
1205 if subs:
1206 if (not match('.hgsub') and
1206 if (not match('.hgsub') and
1207 '.hgsub' in (wctx.modified() + wctx.added())):
1207 '.hgsub' in (wctx.modified() + wctx.added())):
1208 raise util.Abort(
1208 raise util.Abort(
1209 _("can't commit subrepos without .hgsub"))
1209 _("can't commit subrepos without .hgsub"))
1210 changes[0].insert(0, '.hgsubstate')
1210 changes[0].insert(0, '.hgsubstate')
1211
1211
1212 elif '.hgsub' in changes[2]:
1212 elif '.hgsub' in changes[2]:
1213 # clean up .hgsubstate when .hgsub is removed
1213 # clean up .hgsubstate when .hgsub is removed
1214 if ('.hgsubstate' in wctx and
1214 if ('.hgsubstate' in wctx and
1215 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1215 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1216 changes[2].insert(0, '.hgsubstate')
1216 changes[2].insert(0, '.hgsubstate')
1217
1217
1218 # make sure all explicit patterns are matched
1218 # make sure all explicit patterns are matched
1219 if not force and match.files():
1219 if not force and match.files():
1220 matched = set(changes[0] + changes[1] + changes[2])
1220 matched = set(changes[0] + changes[1] + changes[2])
1221
1221
1222 for f in match.files():
1222 for f in match.files():
1223 f = self.dirstate.normalize(f)
1223 f = self.dirstate.normalize(f)
1224 if f == '.' or f in matched or f in wctx.substate:
1224 if f == '.' or f in matched or f in wctx.substate:
1225 continue
1225 continue
1226 if f in changes[3]: # missing
1226 if f in changes[3]: # missing
1227 fail(f, _('file not found!'))
1227 fail(f, _('file not found!'))
1228 if f in vdirs: # visited directory
1228 if f in vdirs: # visited directory
1229 d = f + '/'
1229 d = f + '/'
1230 for mf in matched:
1230 for mf in matched:
1231 if mf.startswith(d):
1231 if mf.startswith(d):
1232 break
1232 break
1233 else:
1233 else:
1234 fail(f, _("no match under directory!"))
1234 fail(f, _("no match under directory!"))
1235 elif f not in self.dirstate:
1235 elif f not in self.dirstate:
1236 fail(f, _("file not tracked!"))
1236 fail(f, _("file not tracked!"))
1237
1237
1238 if (not force and not extra.get("close") and not merge
1238 if (not force and not extra.get("close") and not merge
1239 and not (changes[0] or changes[1] or changes[2])
1239 and not (changes[0] or changes[1] or changes[2])
1240 and wctx.branch() == wctx.p1().branch()):
1240 and wctx.branch() == wctx.p1().branch()):
1241 return None
1241 return None
1242
1242
1243 if merge and changes[3]:
1243 if merge and changes[3]:
1244 raise util.Abort(_("cannot commit merge with missing files"))
1244 raise util.Abort(_("cannot commit merge with missing files"))
1245
1245
1246 ms = mergemod.mergestate(self)
1246 ms = mergemod.mergestate(self)
1247 for f in changes[0]:
1247 for f in changes[0]:
1248 if f in ms and ms[f] == 'u':
1248 if f in ms and ms[f] == 'u':
1249 raise util.Abort(_("unresolved merge conflicts "
1249 raise util.Abort(_("unresolved merge conflicts "
1250 "(see hg help resolve)"))
1250 "(see hg help resolve)"))
1251
1251
1252 cctx = context.workingctx(self, text, user, date, extra, changes)
1252 cctx = context.workingctx(self, text, user, date, extra, changes)
1253 if editor:
1253 if editor:
1254 cctx._text = editor(self, cctx, subs)
1254 cctx._text = editor(self, cctx, subs)
1255 edited = (text != cctx._text)
1255 edited = (text != cctx._text)
1256
1256
1257 # commit subs and write new state
1257 # commit subs and write new state
1258 if subs:
1258 if subs:
1259 for s in sorted(commitsubs):
1259 for s in sorted(commitsubs):
1260 sub = wctx.sub(s)
1260 sub = wctx.sub(s)
1261 self.ui.status(_('committing subrepository %s\n') %
1261 self.ui.status(_('committing subrepository %s\n') %
1262 subrepo.subrelpath(sub))
1262 subrepo.subrelpath(sub))
1263 sr = sub.commit(cctx._text, user, date)
1263 sr = sub.commit(cctx._text, user, date)
1264 newstate[s] = (newstate[s][0], sr)
1264 newstate[s] = (newstate[s][0], sr)
1265 subrepo.writestate(self, newstate)
1265 subrepo.writestate(self, newstate)
1266
1266
1267 # Save commit message in case this transaction gets rolled back
1267 # Save commit message in case this transaction gets rolled back
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1269 # the assumption that the user will use the same editor again.
1269 # the assumption that the user will use the same editor again.
1270 msgfn = self.savecommitmessage(cctx._text)
1270 msgfn = self.savecommitmessage(cctx._text)
1271
1271
1272 p1, p2 = self.dirstate.parents()
1272 p1, p2 = self.dirstate.parents()
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1274 try:
1274 try:
1275 self.hook("precommit", throw=True, parent1=hookp1,
1275 self.hook("precommit", throw=True, parent1=hookp1,
1276 parent2=hookp2)
1276 parent2=hookp2)
1277 ret = self.commitctx(cctx, True)
1277 ret = self.commitctx(cctx, True)
1278 except: # re-raises
1278 except: # re-raises
1279 if edited:
1279 if edited:
1280 self.ui.write(
1280 self.ui.write(
1281 _('note: commit message saved in %s\n') % msgfn)
1281 _('note: commit message saved in %s\n') % msgfn)
1282 raise
1282 raise
1283
1283
1284 # update bookmarks, dirstate and mergestate
1284 # update bookmarks, dirstate and mergestate
1285 bookmarks.update(self, [p1, p2], ret)
1285 bookmarks.update(self, [p1, p2], ret)
1286 for f in changes[0] + changes[1]:
1286 for f in changes[0] + changes[1]:
1287 self.dirstate.normal(f)
1287 self.dirstate.normal(f)
1288 for f in changes[2]:
1288 for f in changes[2]:
1289 self.dirstate.drop(f)
1289 self.dirstate.drop(f)
1290 self.dirstate.setparents(ret)
1290 self.dirstate.setparents(ret)
1291 ms.reset()
1291 ms.reset()
1292 finally:
1292 finally:
1293 wlock.release()
1293 wlock.release()
1294
1294
1295 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1295 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1296 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1296 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1297 self._afterlock(commithook)
1297 self._afterlock(commithook)
1298 return ret
1298 return ret
1299
1299
1300 @unfilteredmethod
1300 @unfilteredmethod
1301 def commitctx(self, ctx, error=False):
1301 def commitctx(self, ctx, error=False):
1302 """Add a new revision to current repository.
1302 """Add a new revision to current repository.
1303 Revision information is passed via the context argument.
1303 Revision information is passed via the context argument.
1304 """
1304 """
1305
1305
1306 tr = lock = None
1306 tr = lock = None
1307 removed = list(ctx.removed())
1307 removed = list(ctx.removed())
1308 p1, p2 = ctx.p1(), ctx.p2()
1308 p1, p2 = ctx.p1(), ctx.p2()
1309 user = ctx.user()
1309 user = ctx.user()
1310
1310
1311 lock = self.lock()
1311 lock = self.lock()
1312 try:
1312 try:
1313 tr = self.transaction("commit")
1313 tr = self.transaction("commit")
1314 trp = weakref.proxy(tr)
1314 trp = weakref.proxy(tr)
1315
1315
1316 if ctx.files():
1316 if ctx.files():
1317 m1 = p1.manifest().copy()
1317 m1 = p1.manifest().copy()
1318 m2 = p2.manifest()
1318 m2 = p2.manifest()
1319
1319
1320 # check in files
1320 # check in files
1321 new = {}
1321 new = {}
1322 changed = []
1322 changed = []
1323 linkrev = len(self)
1323 linkrev = len(self)
1324 for f in sorted(ctx.modified() + ctx.added()):
1324 for f in sorted(ctx.modified() + ctx.added()):
1325 self.ui.note(f + "\n")
1325 self.ui.note(f + "\n")
1326 try:
1326 try:
1327 fctx = ctx[f]
1327 fctx = ctx[f]
1328 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1328 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1329 changed)
1329 changed)
1330 m1.set(f, fctx.flags())
1330 m1.set(f, fctx.flags())
1331 except OSError, inst:
1331 except OSError, inst:
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1333 raise
1333 raise
1334 except IOError, inst:
1334 except IOError, inst:
1335 errcode = getattr(inst, 'errno', errno.ENOENT)
1335 errcode = getattr(inst, 'errno', errno.ENOENT)
1336 if error or errcode and errcode != errno.ENOENT:
1336 if error or errcode and errcode != errno.ENOENT:
1337 self.ui.warn(_("trouble committing %s!\n") % f)
1337 self.ui.warn(_("trouble committing %s!\n") % f)
1338 raise
1338 raise
1339 else:
1339 else:
1340 removed.append(f)
1340 removed.append(f)
1341
1341
1342 # update manifest
1342 # update manifest
1343 m1.update(new)
1343 m1.update(new)
1344 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1344 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1345 drop = [f for f in removed if f in m1]
1345 drop = [f for f in removed if f in m1]
1346 for f in drop:
1346 for f in drop:
1347 del m1[f]
1347 del m1[f]
1348 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1348 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1349 p2.manifestnode(), (new, drop))
1349 p2.manifestnode(), (new, drop))
1350 files = changed + removed
1350 files = changed + removed
1351 else:
1351 else:
1352 mn = p1.manifestnode()
1352 mn = p1.manifestnode()
1353 files = []
1353 files = []
1354
1354
1355 # update changelog
1355 # update changelog
1356 self.changelog.delayupdate()
1356 self.changelog.delayupdate()
1357 n = self.changelog.add(mn, files, ctx.description(),
1357 n = self.changelog.add(mn, files, ctx.description(),
1358 trp, p1.node(), p2.node(),
1358 trp, p1.node(), p2.node(),
1359 user, ctx.date(), ctx.extra().copy())
1359 user, ctx.date(), ctx.extra().copy())
1360 p = lambda: self.changelog.writepending() and self.root or ""
1360 p = lambda: self.changelog.writepending() and self.root or ""
1361 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1361 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1362 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1362 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1363 parent2=xp2, pending=p)
1363 parent2=xp2, pending=p)
1364 self.changelog.finalize(trp)
1364 self.changelog.finalize(trp)
1365 # set the new commit is proper phase
1365 # set the new commit is proper phase
1366 targetphase = phases.newcommitphase(self.ui)
1366 targetphase = phases.newcommitphase(self.ui)
1367 if targetphase:
1367 if targetphase:
1368 # retract boundary do not alter parent changeset.
1368 # retract boundary do not alter parent changeset.
1369 # if a parent have higher the resulting phase will
1369 # if a parent have higher the resulting phase will
1370 # be compliant anyway
1370 # be compliant anyway
1371 #
1371 #
1372 # if minimal phase was 0 we don't need to retract anything
1372 # if minimal phase was 0 we don't need to retract anything
1373 phases.retractboundary(self, targetphase, [n])
1373 phases.retractboundary(self, targetphase, [n])
1374 tr.close()
1374 tr.close()
1375 branchmap.updatecache(self)
1375 branchmap.updatecache(self)
1376 return n
1376 return n
1377 finally:
1377 finally:
1378 if tr:
1378 if tr:
1379 tr.release()
1379 tr.release()
1380 lock.release()
1380 lock.release()
1381
1381
1382 @unfilteredmethod
1382 @unfilteredmethod
1383 def destroying(self):
1383 def destroying(self):
1384 '''Inform the repository that nodes are about to be destroyed.
1384 '''Inform the repository that nodes are about to be destroyed.
1385 Intended for use by strip and rollback, so there's a common
1385 Intended for use by strip and rollback, so there's a common
1386 place for anything that has to be done before destroying history.
1386 place for anything that has to be done before destroying history.
1387
1387
1388 This is mostly useful for saving state that is in memory and waiting
1388 This is mostly useful for saving state that is in memory and waiting
1389 to be flushed when the current lock is released. Because a call to
1389 to be flushed when the current lock is released. Because a call to
1390 destroyed is imminent, the repo will be invalidated causing those
1390 destroyed is imminent, the repo will be invalidated causing those
1391 changes to stay in memory (waiting for the next unlock), or vanish
1391 changes to stay in memory (waiting for the next unlock), or vanish
1392 completely.
1392 completely.
1393 '''
1393 '''
1394 # It simplifies the logic around updating the branchheads cache if we
1394 # It simplifies the logic around updating the branchheads cache if we
1395 # only have to consider the effect of the stripped revisions and not
1395 # only have to consider the effect of the stripped revisions and not
1396 # revisions missing because the cache is out-of-date.
1396 # revisions missing because the cache is out-of-date.
1397 branchmap.updatecache(self)
1397 branchmap.updatecache(self)
1398
1398
1399 # When using the same lock to commit and strip, the phasecache is left
1399 # When using the same lock to commit and strip, the phasecache is left
1400 # dirty after committing. Then when we strip, the repo is invalidated,
1400 # dirty after committing. Then when we strip, the repo is invalidated,
1401 # causing those changes to disappear.
1401 # causing those changes to disappear.
1402 if '_phasecache' in vars(self):
1402 if '_phasecache' in vars(self):
1403 self._phasecache.write()
1403 self._phasecache.write()
1404
1404
1405 @unfilteredmethod
1405 @unfilteredmethod
1406 def destroyed(self, newheadnodes=None):
1406 def destroyed(self, newheadnodes=None):
1407 '''Inform the repository that nodes have been destroyed.
1407 '''Inform the repository that nodes have been destroyed.
1408 Intended for use by strip and rollback, so there's a common
1408 Intended for use by strip and rollback, so there's a common
1409 place for anything that has to be done after destroying history.
1409 place for anything that has to be done after destroying history.
1410
1410
1411 If you know the branchheadcache was uptodate before nodes were removed
1411 If you know the branchheadcache was uptodate before nodes were removed
1412 and you also know the set of candidate new heads that may have resulted
1412 and you also know the set of candidate new heads that may have resulted
1413 from the destruction, you can set newheadnodes. This will enable the
1413 from the destruction, you can set newheadnodes. This will enable the
1414 code to update the branchheads cache, rather than having future code
1414 code to update the branchheads cache, rather than having future code
1415 decide it's invalid and regenerating it from scratch.
1415 decide it's invalid and regenerating it from scratch.
1416 '''
1416 '''
1417 # When one tries to:
1417 # When one tries to:
1418 # 1) destroy nodes thus calling this method (e.g. strip)
1418 # 1) destroy nodes thus calling this method (e.g. strip)
1419 # 2) use phasecache somewhere (e.g. commit)
1419 # 2) use phasecache somewhere (e.g. commit)
1420 #
1420 #
1421 # then 2) will fail because the phasecache contains nodes that were
1421 # then 2) will fail because the phasecache contains nodes that were
1422 # removed. We can either remove phasecache from the filecache,
1422 # removed. We can either remove phasecache from the filecache,
1423 # causing it to reload next time it is accessed, or simply filter
1423 # causing it to reload next time it is accessed, or simply filter
1424 # the removed nodes now and write the updated cache.
1424 # the removed nodes now and write the updated cache.
1425 if '_phasecache' in self._filecache:
1425 if '_phasecache' in self._filecache:
1426 self._phasecache.filterunknown(self)
1426 self._phasecache.filterunknown(self)
1427 self._phasecache.write()
1427 self._phasecache.write()
1428
1428
1429 # If we have info, newheadnodes, on how to update the branch cache, do
1429 # If we have info, newheadnodes, on how to update the branch cache, do
1430 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1430 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1431 # will be caught the next time it is read.
1431 # will be caught the next time it is read.
1432 if newheadnodes:
1432 if newheadnodes:
1433 cl = self.changelog
1433 cl = self.changelog
1434 revgen = (cl.rev(node) for node in newheadnodes
1434 revgen = (cl.rev(node) for node in newheadnodes
1435 if cl.hasnode(node))
1435 if cl.hasnode(node))
1436 cache = self._branchcaches[None]
1436 cache = self._branchcaches[None]
1437 cache.update(self, revgen)
1437 cache.update(self, revgen)
1438 cache.write(self)
1438 cache.write(self)
1439
1439
1440 # Ensure the persistent tag cache is updated. Doing it now
1440 # Ensure the persistent tag cache is updated. Doing it now
1441 # means that the tag cache only has to worry about destroyed
1441 # means that the tag cache only has to worry about destroyed
1442 # heads immediately after a strip/rollback. That in turn
1442 # heads immediately after a strip/rollback. That in turn
1443 # guarantees that "cachetip == currenttip" (comparing both rev
1443 # guarantees that "cachetip == currenttip" (comparing both rev
1444 # and node) always means no nodes have been added or destroyed.
1444 # and node) always means no nodes have been added or destroyed.
1445
1445
1446 # XXX this is suboptimal when qrefresh'ing: we strip the current
1446 # XXX this is suboptimal when qrefresh'ing: we strip the current
1447 # head, refresh the tag cache, then immediately add a new head.
1447 # head, refresh the tag cache, then immediately add a new head.
1448 # But I think doing it this way is necessary for the "instant
1448 # But I think doing it this way is necessary for the "instant
1449 # tag cache retrieval" case to work.
1449 # tag cache retrieval" case to work.
1450 self.invalidatecaches()
1450 self.invalidate()
1451
1452 # Discard all cache entries to force reloading everything.
1453 self._filecache.clear()
1454
1451
1455 def walk(self, match, node=None):
1452 def walk(self, match, node=None):
1456 '''
1453 '''
1457 walk recursively through the directory tree or a given
1454 walk recursively through the directory tree or a given
1458 changeset, finding all files matched by the match
1455 changeset, finding all files matched by the match
1459 function
1456 function
1460 '''
1457 '''
1461 return self[node].walk(match)
1458 return self[node].walk(match)
1462
1459
1463 def status(self, node1='.', node2=None, match=None,
1460 def status(self, node1='.', node2=None, match=None,
1464 ignored=False, clean=False, unknown=False,
1461 ignored=False, clean=False, unknown=False,
1465 listsubrepos=False):
1462 listsubrepos=False):
1466 """return status of files between two nodes or node and working
1463 """return status of files between two nodes or node and working
1467 directory.
1464 directory.
1468
1465
1469 If node1 is None, use the first dirstate parent instead.
1466 If node1 is None, use the first dirstate parent instead.
1470 If node2 is None, compare node1 with working directory.
1467 If node2 is None, compare node1 with working directory.
1471 """
1468 """
1472
1469
1473 def mfmatches(ctx):
1470 def mfmatches(ctx):
1474 mf = ctx.manifest().copy()
1471 mf = ctx.manifest().copy()
1475 if match.always():
1472 if match.always():
1476 return mf
1473 return mf
1477 for fn in mf.keys():
1474 for fn in mf.keys():
1478 if not match(fn):
1475 if not match(fn):
1479 del mf[fn]
1476 del mf[fn]
1480 return mf
1477 return mf
1481
1478
1482 if isinstance(node1, context.changectx):
1479 if isinstance(node1, context.changectx):
1483 ctx1 = node1
1480 ctx1 = node1
1484 else:
1481 else:
1485 ctx1 = self[node1]
1482 ctx1 = self[node1]
1486 if isinstance(node2, context.changectx):
1483 if isinstance(node2, context.changectx):
1487 ctx2 = node2
1484 ctx2 = node2
1488 else:
1485 else:
1489 ctx2 = self[node2]
1486 ctx2 = self[node2]
1490
1487
1491 working = ctx2.rev() is None
1488 working = ctx2.rev() is None
1492 parentworking = working and ctx1 == self['.']
1489 parentworking = working and ctx1 == self['.']
1493 match = match or matchmod.always(self.root, self.getcwd())
1490 match = match or matchmod.always(self.root, self.getcwd())
1494 listignored, listclean, listunknown = ignored, clean, unknown
1491 listignored, listclean, listunknown = ignored, clean, unknown
1495
1492
1496 # load earliest manifest first for caching reasons
1493 # load earliest manifest first for caching reasons
1497 if not working and ctx2.rev() < ctx1.rev():
1494 if not working and ctx2.rev() < ctx1.rev():
1498 ctx2.manifest()
1495 ctx2.manifest()
1499
1496
1500 if not parentworking:
1497 if not parentworking:
1501 def bad(f, msg):
1498 def bad(f, msg):
1502 # 'f' may be a directory pattern from 'match.files()',
1499 # 'f' may be a directory pattern from 'match.files()',
1503 # so 'f not in ctx1' is not enough
1500 # so 'f not in ctx1' is not enough
1504 if f not in ctx1 and f not in ctx1.dirs():
1501 if f not in ctx1 and f not in ctx1.dirs():
1505 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1502 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1506 match.bad = bad
1503 match.bad = bad
1507
1504
1508 if working: # we need to scan the working dir
1505 if working: # we need to scan the working dir
1509 subrepos = []
1506 subrepos = []
1510 if '.hgsub' in self.dirstate:
1507 if '.hgsub' in self.dirstate:
1511 subrepos = ctx2.substate.keys()
1508 subrepos = ctx2.substate.keys()
1512 s = self.dirstate.status(match, subrepos, listignored,
1509 s = self.dirstate.status(match, subrepos, listignored,
1513 listclean, listunknown)
1510 listclean, listunknown)
1514 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1511 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1515
1512
1516 # check for any possibly clean files
1513 # check for any possibly clean files
1517 if parentworking and cmp:
1514 if parentworking and cmp:
1518 fixup = []
1515 fixup = []
1519 # do a full compare of any files that might have changed
1516 # do a full compare of any files that might have changed
1520 for f in sorted(cmp):
1517 for f in sorted(cmp):
1521 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1518 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1522 or ctx1[f].cmp(ctx2[f])):
1519 or ctx1[f].cmp(ctx2[f])):
1523 modified.append(f)
1520 modified.append(f)
1524 else:
1521 else:
1525 fixup.append(f)
1522 fixup.append(f)
1526
1523
1527 # update dirstate for files that are actually clean
1524 # update dirstate for files that are actually clean
1528 if fixup:
1525 if fixup:
1529 if listclean:
1526 if listclean:
1530 clean += fixup
1527 clean += fixup
1531
1528
1532 try:
1529 try:
1533 # updating the dirstate is optional
1530 # updating the dirstate is optional
1534 # so we don't wait on the lock
1531 # so we don't wait on the lock
1535 wlock = self.wlock(False)
1532 wlock = self.wlock(False)
1536 try:
1533 try:
1537 for f in fixup:
1534 for f in fixup:
1538 self.dirstate.normal(f)
1535 self.dirstate.normal(f)
1539 finally:
1536 finally:
1540 wlock.release()
1537 wlock.release()
1541 except error.LockError:
1538 except error.LockError:
1542 pass
1539 pass
1543
1540
1544 if not parentworking:
1541 if not parentworking:
1545 mf1 = mfmatches(ctx1)
1542 mf1 = mfmatches(ctx1)
1546 if working:
1543 if working:
1547 # we are comparing working dir against non-parent
1544 # we are comparing working dir against non-parent
1548 # generate a pseudo-manifest for the working dir
1545 # generate a pseudo-manifest for the working dir
1549 mf2 = mfmatches(self['.'])
1546 mf2 = mfmatches(self['.'])
1550 for f in cmp + modified + added:
1547 for f in cmp + modified + added:
1551 mf2[f] = None
1548 mf2[f] = None
1552 mf2.set(f, ctx2.flags(f))
1549 mf2.set(f, ctx2.flags(f))
1553 for f in removed:
1550 for f in removed:
1554 if f in mf2:
1551 if f in mf2:
1555 del mf2[f]
1552 del mf2[f]
1556 else:
1553 else:
1557 # we are comparing two revisions
1554 # we are comparing two revisions
1558 deleted, unknown, ignored = [], [], []
1555 deleted, unknown, ignored = [], [], []
1559 mf2 = mfmatches(ctx2)
1556 mf2 = mfmatches(ctx2)
1560
1557
1561 modified, added, clean = [], [], []
1558 modified, added, clean = [], [], []
1562 withflags = mf1.withflags() | mf2.withflags()
1559 withflags = mf1.withflags() | mf2.withflags()
1563 for fn in mf2:
1560 for fn in mf2:
1564 if fn in mf1:
1561 if fn in mf1:
1565 if (fn not in deleted and
1562 if (fn not in deleted and
1566 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1563 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1567 (mf1[fn] != mf2[fn] and
1564 (mf1[fn] != mf2[fn] and
1568 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1565 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1569 modified.append(fn)
1566 modified.append(fn)
1570 elif listclean:
1567 elif listclean:
1571 clean.append(fn)
1568 clean.append(fn)
1572 del mf1[fn]
1569 del mf1[fn]
1573 elif fn not in deleted:
1570 elif fn not in deleted:
1574 added.append(fn)
1571 added.append(fn)
1575 removed = mf1.keys()
1572 removed = mf1.keys()
1576
1573
1577 if working and modified and not self.dirstate._checklink:
1574 if working and modified and not self.dirstate._checklink:
1578 # Symlink placeholders may get non-symlink-like contents
1575 # Symlink placeholders may get non-symlink-like contents
1579 # via user error or dereferencing by NFS or Samba servers,
1576 # via user error or dereferencing by NFS or Samba servers,
1580 # so we filter out any placeholders that don't look like a
1577 # so we filter out any placeholders that don't look like a
1581 # symlink
1578 # symlink
1582 sane = []
1579 sane = []
1583 for f in modified:
1580 for f in modified:
1584 if ctx2.flags(f) == 'l':
1581 if ctx2.flags(f) == 'l':
1585 d = ctx2[f].data()
1582 d = ctx2[f].data()
1586 if len(d) >= 1024 or '\n' in d or util.binary(d):
1583 if len(d) >= 1024 or '\n' in d or util.binary(d):
1587 self.ui.debug('ignoring suspect symlink placeholder'
1584 self.ui.debug('ignoring suspect symlink placeholder'
1588 ' "%s"\n' % f)
1585 ' "%s"\n' % f)
1589 continue
1586 continue
1590 sane.append(f)
1587 sane.append(f)
1591 modified = sane
1588 modified = sane
1592
1589
1593 r = modified, added, removed, deleted, unknown, ignored, clean
1590 r = modified, added, removed, deleted, unknown, ignored, clean
1594
1591
1595 if listsubrepos:
1592 if listsubrepos:
1596 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1593 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1597 if working:
1594 if working:
1598 rev2 = None
1595 rev2 = None
1599 else:
1596 else:
1600 rev2 = ctx2.substate[subpath][1]
1597 rev2 = ctx2.substate[subpath][1]
1601 try:
1598 try:
1602 submatch = matchmod.narrowmatcher(subpath, match)
1599 submatch = matchmod.narrowmatcher(subpath, match)
1603 s = sub.status(rev2, match=submatch, ignored=listignored,
1600 s = sub.status(rev2, match=submatch, ignored=listignored,
1604 clean=listclean, unknown=listunknown,
1601 clean=listclean, unknown=listunknown,
1605 listsubrepos=True)
1602 listsubrepos=True)
1606 for rfiles, sfiles in zip(r, s):
1603 for rfiles, sfiles in zip(r, s):
1607 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1604 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1608 except error.LookupError:
1605 except error.LookupError:
1609 self.ui.status(_("skipping missing subrepository: %s\n")
1606 self.ui.status(_("skipping missing subrepository: %s\n")
1610 % subpath)
1607 % subpath)
1611
1608
1612 for l in r:
1609 for l in r:
1613 l.sort()
1610 l.sort()
1614 return r
1611 return r
1615
1612
1616 def heads(self, start=None):
1613 def heads(self, start=None):
1617 heads = self.changelog.heads(start)
1614 heads = self.changelog.heads(start)
1618 # sort the output in rev descending order
1615 # sort the output in rev descending order
1619 return sorted(heads, key=self.changelog.rev, reverse=True)
1616 return sorted(heads, key=self.changelog.rev, reverse=True)
1620
1617
1621 def branchheads(self, branch=None, start=None, closed=False):
1618 def branchheads(self, branch=None, start=None, closed=False):
1622 '''return a (possibly filtered) list of heads for the given branch
1619 '''return a (possibly filtered) list of heads for the given branch
1623
1620
1624 Heads are returned in topological order, from newest to oldest.
1621 Heads are returned in topological order, from newest to oldest.
1625 If branch is None, use the dirstate branch.
1622 If branch is None, use the dirstate branch.
1626 If start is not None, return only heads reachable from start.
1623 If start is not None, return only heads reachable from start.
1627 If closed is True, return heads that are marked as closed as well.
1624 If closed is True, return heads that are marked as closed as well.
1628 '''
1625 '''
1629 if branch is None:
1626 if branch is None:
1630 branch = self[None].branch()
1627 branch = self[None].branch()
1631 branches = self.branchmap()
1628 branches = self.branchmap()
1632 if branch not in branches:
1629 if branch not in branches:
1633 return []
1630 return []
1634 # the cache returns heads ordered lowest to highest
1631 # the cache returns heads ordered lowest to highest
1635 bheads = list(reversed(branches[branch]))
1632 bheads = list(reversed(branches[branch]))
1636 if start is not None:
1633 if start is not None:
1637 # filter out the heads that cannot be reached from startrev
1634 # filter out the heads that cannot be reached from startrev
1638 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1635 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1639 bheads = [h for h in bheads if h in fbheads]
1636 bheads = [h for h in bheads if h in fbheads]
1640 if not closed:
1637 if not closed:
1641 bheads = [h for h in bheads if not self[h].closesbranch()]
1638 bheads = [h for h in bheads if not self[h].closesbranch()]
1642 return bheads
1639 return bheads
1643
1640
1644 def branches(self, nodes):
1641 def branches(self, nodes):
1645 if not nodes:
1642 if not nodes:
1646 nodes = [self.changelog.tip()]
1643 nodes = [self.changelog.tip()]
1647 b = []
1644 b = []
1648 for n in nodes:
1645 for n in nodes:
1649 t = n
1646 t = n
1650 while True:
1647 while True:
1651 p = self.changelog.parents(n)
1648 p = self.changelog.parents(n)
1652 if p[1] != nullid or p[0] == nullid:
1649 if p[1] != nullid or p[0] == nullid:
1653 b.append((t, n, p[0], p[1]))
1650 b.append((t, n, p[0], p[1]))
1654 break
1651 break
1655 n = p[0]
1652 n = p[0]
1656 return b
1653 return b
1657
1654
1658 def between(self, pairs):
1655 def between(self, pairs):
1659 r = []
1656 r = []
1660
1657
1661 for top, bottom in pairs:
1658 for top, bottom in pairs:
1662 n, l, i = top, [], 0
1659 n, l, i = top, [], 0
1663 f = 1
1660 f = 1
1664
1661
1665 while n != bottom and n != nullid:
1662 while n != bottom and n != nullid:
1666 p = self.changelog.parents(n)[0]
1663 p = self.changelog.parents(n)[0]
1667 if i == f:
1664 if i == f:
1668 l.append(n)
1665 l.append(n)
1669 f = f * 2
1666 f = f * 2
1670 n = p
1667 n = p
1671 i += 1
1668 i += 1
1672
1669
1673 r.append(l)
1670 r.append(l)
1674
1671
1675 return r
1672 return r
1676
1673
1677 def pull(self, remote, heads=None, force=False):
1674 def pull(self, remote, heads=None, force=False):
1678 # don't open transaction for nothing or you break future useful
1675 # don't open transaction for nothing or you break future useful
1679 # rollback call
1676 # rollback call
1680 tr = None
1677 tr = None
1681 trname = 'pull\n' + util.hidepassword(remote.url())
1678 trname = 'pull\n' + util.hidepassword(remote.url())
1682 lock = self.lock()
1679 lock = self.lock()
1683 try:
1680 try:
1684 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1681 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1685 force=force)
1682 force=force)
1686 common, fetch, rheads = tmp
1683 common, fetch, rheads = tmp
1687 if not fetch:
1684 if not fetch:
1688 self.ui.status(_("no changes found\n"))
1685 self.ui.status(_("no changes found\n"))
1689 added = []
1686 added = []
1690 result = 0
1687 result = 0
1691 else:
1688 else:
1692 tr = self.transaction(trname)
1689 tr = self.transaction(trname)
1693 if heads is None and list(common) == [nullid]:
1690 if heads is None and list(common) == [nullid]:
1694 self.ui.status(_("requesting all changes\n"))
1691 self.ui.status(_("requesting all changes\n"))
1695 elif heads is None and remote.capable('changegroupsubset'):
1692 elif heads is None and remote.capable('changegroupsubset'):
1696 # issue1320, avoid a race if remote changed after discovery
1693 # issue1320, avoid a race if remote changed after discovery
1697 heads = rheads
1694 heads = rheads
1698
1695
1699 if remote.capable('getbundle'):
1696 if remote.capable('getbundle'):
1700 cg = remote.getbundle('pull', common=common,
1697 cg = remote.getbundle('pull', common=common,
1701 heads=heads or rheads)
1698 heads=heads or rheads)
1702 elif heads is None:
1699 elif heads is None:
1703 cg = remote.changegroup(fetch, 'pull')
1700 cg = remote.changegroup(fetch, 'pull')
1704 elif not remote.capable('changegroupsubset'):
1701 elif not remote.capable('changegroupsubset'):
1705 raise util.Abort(_("partial pull cannot be done because "
1702 raise util.Abort(_("partial pull cannot be done because "
1706 "other repository doesn't support "
1703 "other repository doesn't support "
1707 "changegroupsubset."))
1704 "changegroupsubset."))
1708 else:
1705 else:
1709 cg = remote.changegroupsubset(fetch, heads, 'pull')
1706 cg = remote.changegroupsubset(fetch, heads, 'pull')
1710 clstart = len(self.changelog)
1707 clstart = len(self.changelog)
1711 result = self.addchangegroup(cg, 'pull', remote.url())
1708 result = self.addchangegroup(cg, 'pull', remote.url())
1712 clend = len(self.changelog)
1709 clend = len(self.changelog)
1713 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1710 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1714
1711
1715 # compute target subset
1712 # compute target subset
1716 if heads is None:
1713 if heads is None:
1717 # We pulled every thing possible
1714 # We pulled every thing possible
1718 # sync on everything common
1715 # sync on everything common
1719 subset = common + added
1716 subset = common + added
1720 else:
1717 else:
1721 # We pulled a specific subset
1718 # We pulled a specific subset
1722 # sync on this subset
1719 # sync on this subset
1723 subset = heads
1720 subset = heads
1724
1721
1725 # Get remote phases data from remote
1722 # Get remote phases data from remote
1726 remotephases = remote.listkeys('phases')
1723 remotephases = remote.listkeys('phases')
1727 publishing = bool(remotephases.get('publishing', False))
1724 publishing = bool(remotephases.get('publishing', False))
1728 if remotephases and not publishing:
1725 if remotephases and not publishing:
1729 # remote is new and unpublishing
1726 # remote is new and unpublishing
1730 pheads, _dr = phases.analyzeremotephases(self, subset,
1727 pheads, _dr = phases.analyzeremotephases(self, subset,
1731 remotephases)
1728 remotephases)
1732 phases.advanceboundary(self, phases.public, pheads)
1729 phases.advanceboundary(self, phases.public, pheads)
1733 phases.advanceboundary(self, phases.draft, subset)
1730 phases.advanceboundary(self, phases.draft, subset)
1734 else:
1731 else:
1735 # Remote is old or publishing all common changesets
1732 # Remote is old or publishing all common changesets
1736 # should be seen as public
1733 # should be seen as public
1737 phases.advanceboundary(self, phases.public, subset)
1734 phases.advanceboundary(self, phases.public, subset)
1738
1735
1739 if obsolete._enabled:
1736 if obsolete._enabled:
1740 self.ui.debug('fetching remote obsolete markers\n')
1737 self.ui.debug('fetching remote obsolete markers\n')
1741 remoteobs = remote.listkeys('obsolete')
1738 remoteobs = remote.listkeys('obsolete')
1742 if 'dump0' in remoteobs:
1739 if 'dump0' in remoteobs:
1743 if tr is None:
1740 if tr is None:
1744 tr = self.transaction(trname)
1741 tr = self.transaction(trname)
1745 for key in sorted(remoteobs, reverse=True):
1742 for key in sorted(remoteobs, reverse=True):
1746 if key.startswith('dump'):
1743 if key.startswith('dump'):
1747 data = base85.b85decode(remoteobs[key])
1744 data = base85.b85decode(remoteobs[key])
1748 self.obsstore.mergemarkers(tr, data)
1745 self.obsstore.mergemarkers(tr, data)
1749 self.invalidatevolatilesets()
1746 self.invalidatevolatilesets()
1750 if tr is not None:
1747 if tr is not None:
1751 tr.close()
1748 tr.close()
1752 finally:
1749 finally:
1753 if tr is not None:
1750 if tr is not None:
1754 tr.release()
1751 tr.release()
1755 lock.release()
1752 lock.release()
1756
1753
1757 return result
1754 return result
1758
1755
1759 def checkpush(self, force, revs):
1756 def checkpush(self, force, revs):
1760 """Extensions can override this function if additional checks have
1757 """Extensions can override this function if additional checks have
1761 to be performed before pushing, or call it if they override push
1758 to be performed before pushing, or call it if they override push
1762 command.
1759 command.
1763 """
1760 """
1764 pass
1761 pass
1765
1762
1766 def push(self, remote, force=False, revs=None, newbranch=False):
1763 def push(self, remote, force=False, revs=None, newbranch=False):
1767 '''Push outgoing changesets (limited by revs) from the current
1764 '''Push outgoing changesets (limited by revs) from the current
1768 repository to remote. Return an integer:
1765 repository to remote. Return an integer:
1769 - None means nothing to push
1766 - None means nothing to push
1770 - 0 means HTTP error
1767 - 0 means HTTP error
1771 - 1 means we pushed and remote head count is unchanged *or*
1768 - 1 means we pushed and remote head count is unchanged *or*
1772 we have outgoing changesets but refused to push
1769 we have outgoing changesets but refused to push
1773 - other values as described by addchangegroup()
1770 - other values as described by addchangegroup()
1774 '''
1771 '''
1775 # there are two ways to push to remote repo:
1772 # there are two ways to push to remote repo:
1776 #
1773 #
1777 # addchangegroup assumes local user can lock remote
1774 # addchangegroup assumes local user can lock remote
1778 # repo (local filesystem, old ssh servers).
1775 # repo (local filesystem, old ssh servers).
1779 #
1776 #
1780 # unbundle assumes local user cannot lock remote repo (new ssh
1777 # unbundle assumes local user cannot lock remote repo (new ssh
1781 # servers, http servers).
1778 # servers, http servers).
1782
1779
1783 if not remote.canpush():
1780 if not remote.canpush():
1784 raise util.Abort(_("destination does not support push"))
1781 raise util.Abort(_("destination does not support push"))
1785 unfi = self.unfiltered()
1782 unfi = self.unfiltered()
1786 # get local lock as we might write phase data
1783 # get local lock as we might write phase data
1787 locallock = self.lock()
1784 locallock = self.lock()
1788 try:
1785 try:
1789 self.checkpush(force, revs)
1786 self.checkpush(force, revs)
1790 lock = None
1787 lock = None
1791 unbundle = remote.capable('unbundle')
1788 unbundle = remote.capable('unbundle')
1792 if not unbundle:
1789 if not unbundle:
1793 lock = remote.lock()
1790 lock = remote.lock()
1794 try:
1791 try:
1795 # discovery
1792 # discovery
1796 fci = discovery.findcommonincoming
1793 fci = discovery.findcommonincoming
1797 commoninc = fci(unfi, remote, force=force)
1794 commoninc = fci(unfi, remote, force=force)
1798 common, inc, remoteheads = commoninc
1795 common, inc, remoteheads = commoninc
1799 fco = discovery.findcommonoutgoing
1796 fco = discovery.findcommonoutgoing
1800 outgoing = fco(unfi, remote, onlyheads=revs,
1797 outgoing = fco(unfi, remote, onlyheads=revs,
1801 commoninc=commoninc, force=force)
1798 commoninc=commoninc, force=force)
1802
1799
1803
1800
1804 if not outgoing.missing:
1801 if not outgoing.missing:
1805 # nothing to push
1802 # nothing to push
1806 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1803 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1807 ret = None
1804 ret = None
1808 else:
1805 else:
1809 # something to push
1806 # something to push
1810 if not force:
1807 if not force:
1811 # if self.obsstore == False --> no obsolete
1808 # if self.obsstore == False --> no obsolete
1812 # then, save the iteration
1809 # then, save the iteration
1813 if unfi.obsstore:
1810 if unfi.obsstore:
1814 # this message are here for 80 char limit reason
1811 # this message are here for 80 char limit reason
1815 mso = _("push includes obsolete changeset: %s!")
1812 mso = _("push includes obsolete changeset: %s!")
1816 mst = "push includes %s changeset: %s!"
1813 mst = "push includes %s changeset: %s!"
1817 # plain versions for i18n tool to detect them
1814 # plain versions for i18n tool to detect them
1818 _("push includes unstable changeset: %s!")
1815 _("push includes unstable changeset: %s!")
1819 _("push includes bumped changeset: %s!")
1816 _("push includes bumped changeset: %s!")
1820 _("push includes divergent changeset: %s!")
1817 _("push includes divergent changeset: %s!")
1821 # If we are to push if there is at least one
1818 # If we are to push if there is at least one
1822 # obsolete or unstable changeset in missing, at
1819 # obsolete or unstable changeset in missing, at
1823 # least one of the missinghead will be obsolete or
1820 # least one of the missinghead will be obsolete or
1824 # unstable. So checking heads only is ok
1821 # unstable. So checking heads only is ok
1825 for node in outgoing.missingheads:
1822 for node in outgoing.missingheads:
1826 ctx = unfi[node]
1823 ctx = unfi[node]
1827 if ctx.obsolete():
1824 if ctx.obsolete():
1828 raise util.Abort(mso % ctx)
1825 raise util.Abort(mso % ctx)
1829 elif ctx.troubled():
1826 elif ctx.troubled():
1830 raise util.Abort(_(mst)
1827 raise util.Abort(_(mst)
1831 % (ctx.troubles()[0],
1828 % (ctx.troubles()[0],
1832 ctx))
1829 ctx))
1833 discovery.checkheads(unfi, remote, outgoing,
1830 discovery.checkheads(unfi, remote, outgoing,
1834 remoteheads, newbranch,
1831 remoteheads, newbranch,
1835 bool(inc))
1832 bool(inc))
1836
1833
1837 # create a changegroup from local
1834 # create a changegroup from local
1838 if revs is None and not outgoing.excluded:
1835 if revs is None and not outgoing.excluded:
1839 # push everything,
1836 # push everything,
1840 # use the fast path, no race possible on push
1837 # use the fast path, no race possible on push
1841 cg = self._changegroup(outgoing.missing, 'push')
1838 cg = self._changegroup(outgoing.missing, 'push')
1842 else:
1839 else:
1843 cg = self.getlocalbundle('push', outgoing)
1840 cg = self.getlocalbundle('push', outgoing)
1844
1841
1845 # apply changegroup to remote
1842 # apply changegroup to remote
1846 if unbundle:
1843 if unbundle:
1847 # local repo finds heads on server, finds out what
1844 # local repo finds heads on server, finds out what
1848 # revs it must push. once revs transferred, if server
1845 # revs it must push. once revs transferred, if server
1849 # finds it has different heads (someone else won
1846 # finds it has different heads (someone else won
1850 # commit/push race), server aborts.
1847 # commit/push race), server aborts.
1851 if force:
1848 if force:
1852 remoteheads = ['force']
1849 remoteheads = ['force']
1853 # ssh: return remote's addchangegroup()
1850 # ssh: return remote's addchangegroup()
1854 # http: return remote's addchangegroup() or 0 for error
1851 # http: return remote's addchangegroup() or 0 for error
1855 ret = remote.unbundle(cg, remoteheads, 'push')
1852 ret = remote.unbundle(cg, remoteheads, 'push')
1856 else:
1853 else:
1857 # we return an integer indicating remote head count
1854 # we return an integer indicating remote head count
1858 # change
1855 # change
1859 ret = remote.addchangegroup(cg, 'push', self.url())
1856 ret = remote.addchangegroup(cg, 'push', self.url())
1860
1857
1861 if ret:
1858 if ret:
1862 # push succeed, synchronize target of the push
1859 # push succeed, synchronize target of the push
1863 cheads = outgoing.missingheads
1860 cheads = outgoing.missingheads
1864 elif revs is None:
1861 elif revs is None:
1865 # All out push fails. synchronize all common
1862 # All out push fails. synchronize all common
1866 cheads = outgoing.commonheads
1863 cheads = outgoing.commonheads
1867 else:
1864 else:
1868 # I want cheads = heads(::missingheads and ::commonheads)
1865 # I want cheads = heads(::missingheads and ::commonheads)
1869 # (missingheads is revs with secret changeset filtered out)
1866 # (missingheads is revs with secret changeset filtered out)
1870 #
1867 #
1871 # This can be expressed as:
1868 # This can be expressed as:
1872 # cheads = ( (missingheads and ::commonheads)
1869 # cheads = ( (missingheads and ::commonheads)
1873 # + (commonheads and ::missingheads))"
1870 # + (commonheads and ::missingheads))"
1874 # )
1871 # )
1875 #
1872 #
1876 # while trying to push we already computed the following:
1873 # while trying to push we already computed the following:
1877 # common = (::commonheads)
1874 # common = (::commonheads)
1878 # missing = ((commonheads::missingheads) - commonheads)
1875 # missing = ((commonheads::missingheads) - commonheads)
1879 #
1876 #
1880 # We can pick:
1877 # We can pick:
1881 # * missingheads part of common (::commonheads)
1878 # * missingheads part of common (::commonheads)
1882 common = set(outgoing.common)
1879 common = set(outgoing.common)
1883 cheads = [node for node in revs if node in common]
1880 cheads = [node for node in revs if node in common]
1884 # and
1881 # and
1885 # * commonheads parents on missing
1882 # * commonheads parents on missing
1886 revset = unfi.set('%ln and parents(roots(%ln))',
1883 revset = unfi.set('%ln and parents(roots(%ln))',
1887 outgoing.commonheads,
1884 outgoing.commonheads,
1888 outgoing.missing)
1885 outgoing.missing)
1889 cheads.extend(c.node() for c in revset)
1886 cheads.extend(c.node() for c in revset)
1890 # even when we don't push, exchanging phase data is useful
1887 # even when we don't push, exchanging phase data is useful
1891 remotephases = remote.listkeys('phases')
1888 remotephases = remote.listkeys('phases')
1892 if not remotephases: # old server or public only repo
1889 if not remotephases: # old server or public only repo
1893 phases.advanceboundary(self, phases.public, cheads)
1890 phases.advanceboundary(self, phases.public, cheads)
1894 # don't push any phase data as there is nothing to push
1891 # don't push any phase data as there is nothing to push
1895 else:
1892 else:
1896 ana = phases.analyzeremotephases(self, cheads, remotephases)
1893 ana = phases.analyzeremotephases(self, cheads, remotephases)
1897 pheads, droots = ana
1894 pheads, droots = ana
1898 ### Apply remote phase on local
1895 ### Apply remote phase on local
1899 if remotephases.get('publishing', False):
1896 if remotephases.get('publishing', False):
1900 phases.advanceboundary(self, phases.public, cheads)
1897 phases.advanceboundary(self, phases.public, cheads)
1901 else: # publish = False
1898 else: # publish = False
1902 phases.advanceboundary(self, phases.public, pheads)
1899 phases.advanceboundary(self, phases.public, pheads)
1903 phases.advanceboundary(self, phases.draft, cheads)
1900 phases.advanceboundary(self, phases.draft, cheads)
1904 ### Apply local phase on remote
1901 ### Apply local phase on remote
1905
1902
1906 # Get the list of all revs draft on remote by public here.
1903 # Get the list of all revs draft on remote by public here.
1907 # XXX Beware that revset break if droots is not strictly
1904 # XXX Beware that revset break if droots is not strictly
1908 # XXX root we may want to ensure it is but it is costly
1905 # XXX root we may want to ensure it is but it is costly
1909 outdated = unfi.set('heads((%ln::%ln) and public())',
1906 outdated = unfi.set('heads((%ln::%ln) and public())',
1910 droots, cheads)
1907 droots, cheads)
1911 for newremotehead in outdated:
1908 for newremotehead in outdated:
1912 r = remote.pushkey('phases',
1909 r = remote.pushkey('phases',
1913 newremotehead.hex(),
1910 newremotehead.hex(),
1914 str(phases.draft),
1911 str(phases.draft),
1915 str(phases.public))
1912 str(phases.public))
1916 if not r:
1913 if not r:
1917 self.ui.warn(_('updating %s to public failed!\n')
1914 self.ui.warn(_('updating %s to public failed!\n')
1918 % newremotehead)
1915 % newremotehead)
1919 self.ui.debug('try to push obsolete markers to remote\n')
1916 self.ui.debug('try to push obsolete markers to remote\n')
1920 if (obsolete._enabled and self.obsstore and
1917 if (obsolete._enabled and self.obsstore and
1921 'obsolete' in remote.listkeys('namespaces')):
1918 'obsolete' in remote.listkeys('namespaces')):
1922 rslts = []
1919 rslts = []
1923 remotedata = self.listkeys('obsolete')
1920 remotedata = self.listkeys('obsolete')
1924 for key in sorted(remotedata, reverse=True):
1921 for key in sorted(remotedata, reverse=True):
1925 # reverse sort to ensure we end with dump0
1922 # reverse sort to ensure we end with dump0
1926 data = remotedata[key]
1923 data = remotedata[key]
1927 rslts.append(remote.pushkey('obsolete', key, '', data))
1924 rslts.append(remote.pushkey('obsolete', key, '', data))
1928 if [r for r in rslts if not r]:
1925 if [r for r in rslts if not r]:
1929 msg = _('failed to push some obsolete markers!\n')
1926 msg = _('failed to push some obsolete markers!\n')
1930 self.ui.warn(msg)
1927 self.ui.warn(msg)
1931 finally:
1928 finally:
1932 if lock is not None:
1929 if lock is not None:
1933 lock.release()
1930 lock.release()
1934 finally:
1931 finally:
1935 locallock.release()
1932 locallock.release()
1936
1933
1937 self.ui.debug("checking for updated bookmarks\n")
1934 self.ui.debug("checking for updated bookmarks\n")
1938 rb = remote.listkeys('bookmarks')
1935 rb = remote.listkeys('bookmarks')
1939 for k in rb.keys():
1936 for k in rb.keys():
1940 if k in unfi._bookmarks:
1937 if k in unfi._bookmarks:
1941 nr, nl = rb[k], hex(self._bookmarks[k])
1938 nr, nl = rb[k], hex(self._bookmarks[k])
1942 if nr in unfi:
1939 if nr in unfi:
1943 cr = unfi[nr]
1940 cr = unfi[nr]
1944 cl = unfi[nl]
1941 cl = unfi[nl]
1945 if bookmarks.validdest(unfi, cr, cl):
1942 if bookmarks.validdest(unfi, cr, cl):
1946 r = remote.pushkey('bookmarks', k, nr, nl)
1943 r = remote.pushkey('bookmarks', k, nr, nl)
1947 if r:
1944 if r:
1948 self.ui.status(_("updating bookmark %s\n") % k)
1945 self.ui.status(_("updating bookmark %s\n") % k)
1949 else:
1946 else:
1950 self.ui.warn(_('updating bookmark %s'
1947 self.ui.warn(_('updating bookmark %s'
1951 ' failed!\n') % k)
1948 ' failed!\n') % k)
1952
1949
1953 return ret
1950 return ret
1954
1951
1955 def changegroupinfo(self, nodes, source):
1952 def changegroupinfo(self, nodes, source):
1956 if self.ui.verbose or source == 'bundle':
1953 if self.ui.verbose or source == 'bundle':
1957 self.ui.status(_("%d changesets found\n") % len(nodes))
1954 self.ui.status(_("%d changesets found\n") % len(nodes))
1958 if self.ui.debugflag:
1955 if self.ui.debugflag:
1959 self.ui.debug("list of changesets:\n")
1956 self.ui.debug("list of changesets:\n")
1960 for node in nodes:
1957 for node in nodes:
1961 self.ui.debug("%s\n" % hex(node))
1958 self.ui.debug("%s\n" % hex(node))
1962
1959
1963 def changegroupsubset(self, bases, heads, source):
1960 def changegroupsubset(self, bases, heads, source):
1964 """Compute a changegroup consisting of all the nodes that are
1961 """Compute a changegroup consisting of all the nodes that are
1965 descendants of any of the bases and ancestors of any of the heads.
1962 descendants of any of the bases and ancestors of any of the heads.
1966 Return a chunkbuffer object whose read() method will return
1963 Return a chunkbuffer object whose read() method will return
1967 successive changegroup chunks.
1964 successive changegroup chunks.
1968
1965
1969 It is fairly complex as determining which filenodes and which
1966 It is fairly complex as determining which filenodes and which
1970 manifest nodes need to be included for the changeset to be complete
1967 manifest nodes need to be included for the changeset to be complete
1971 is non-trivial.
1968 is non-trivial.
1972
1969
1973 Another wrinkle is doing the reverse, figuring out which changeset in
1970 Another wrinkle is doing the reverse, figuring out which changeset in
1974 the changegroup a particular filenode or manifestnode belongs to.
1971 the changegroup a particular filenode or manifestnode belongs to.
1975 """
1972 """
1976 cl = self.changelog
1973 cl = self.changelog
1977 if not bases:
1974 if not bases:
1978 bases = [nullid]
1975 bases = [nullid]
1979 csets, bases, heads = cl.nodesbetween(bases, heads)
1976 csets, bases, heads = cl.nodesbetween(bases, heads)
1980 # We assume that all ancestors of bases are known
1977 # We assume that all ancestors of bases are known
1981 common = cl.ancestors([cl.rev(n) for n in bases])
1978 common = cl.ancestors([cl.rev(n) for n in bases])
1982 return self._changegroupsubset(common, csets, heads, source)
1979 return self._changegroupsubset(common, csets, heads, source)
1983
1980
1984 def getlocalbundle(self, source, outgoing):
1981 def getlocalbundle(self, source, outgoing):
1985 """Like getbundle, but taking a discovery.outgoing as an argument.
1982 """Like getbundle, but taking a discovery.outgoing as an argument.
1986
1983
1987 This is only implemented for local repos and reuses potentially
1984 This is only implemented for local repos and reuses potentially
1988 precomputed sets in outgoing."""
1985 precomputed sets in outgoing."""
1989 if not outgoing.missing:
1986 if not outgoing.missing:
1990 return None
1987 return None
1991 return self._changegroupsubset(outgoing.common,
1988 return self._changegroupsubset(outgoing.common,
1992 outgoing.missing,
1989 outgoing.missing,
1993 outgoing.missingheads,
1990 outgoing.missingheads,
1994 source)
1991 source)
1995
1992
1996 def getbundle(self, source, heads=None, common=None):
1993 def getbundle(self, source, heads=None, common=None):
1997 """Like changegroupsubset, but returns the set difference between the
1994 """Like changegroupsubset, but returns the set difference between the
1998 ancestors of heads and the ancestors common.
1995 ancestors of heads and the ancestors common.
1999
1996
2000 If heads is None, use the local heads. If common is None, use [nullid].
1997 If heads is None, use the local heads. If common is None, use [nullid].
2001
1998
2002 The nodes in common might not all be known locally due to the way the
1999 The nodes in common might not all be known locally due to the way the
2003 current discovery protocol works.
2000 current discovery protocol works.
2004 """
2001 """
2005 cl = self.changelog
2002 cl = self.changelog
2006 if common:
2003 if common:
2007 hasnode = cl.hasnode
2004 hasnode = cl.hasnode
2008 common = [n for n in common if hasnode(n)]
2005 common = [n for n in common if hasnode(n)]
2009 else:
2006 else:
2010 common = [nullid]
2007 common = [nullid]
2011 if not heads:
2008 if not heads:
2012 heads = cl.heads()
2009 heads = cl.heads()
2013 return self.getlocalbundle(source,
2010 return self.getlocalbundle(source,
2014 discovery.outgoing(cl, common, heads))
2011 discovery.outgoing(cl, common, heads))
2015
2012
2016 @unfilteredmethod
2013 @unfilteredmethod
2017 def _changegroupsubset(self, commonrevs, csets, heads, source):
2014 def _changegroupsubset(self, commonrevs, csets, heads, source):
2018
2015
2019 cl = self.changelog
2016 cl = self.changelog
2020 mf = self.manifest
2017 mf = self.manifest
2021 mfs = {} # needed manifests
2018 mfs = {} # needed manifests
2022 fnodes = {} # needed file nodes
2019 fnodes = {} # needed file nodes
2023 changedfiles = set()
2020 changedfiles = set()
2024 fstate = ['', {}]
2021 fstate = ['', {}]
2025 count = [0, 0]
2022 count = [0, 0]
2026
2023
2027 # can we go through the fast path ?
2024 # can we go through the fast path ?
2028 heads.sort()
2025 heads.sort()
2029 if heads == sorted(self.heads()):
2026 if heads == sorted(self.heads()):
2030 return self._changegroup(csets, source)
2027 return self._changegroup(csets, source)
2031
2028
2032 # slow path
2029 # slow path
2033 self.hook('preoutgoing', throw=True, source=source)
2030 self.hook('preoutgoing', throw=True, source=source)
2034 self.changegroupinfo(csets, source)
2031 self.changegroupinfo(csets, source)
2035
2032
2036 # filter any nodes that claim to be part of the known set
2033 # filter any nodes that claim to be part of the known set
2037 def prune(revlog, missing):
2034 def prune(revlog, missing):
2038 rr, rl = revlog.rev, revlog.linkrev
2035 rr, rl = revlog.rev, revlog.linkrev
2039 return [n for n in missing
2036 return [n for n in missing
2040 if rl(rr(n)) not in commonrevs]
2037 if rl(rr(n)) not in commonrevs]
2041
2038
2042 progress = self.ui.progress
2039 progress = self.ui.progress
2043 _bundling = _('bundling')
2040 _bundling = _('bundling')
2044 _changesets = _('changesets')
2041 _changesets = _('changesets')
2045 _manifests = _('manifests')
2042 _manifests = _('manifests')
2046 _files = _('files')
2043 _files = _('files')
2047
2044
2048 def lookup(revlog, x):
2045 def lookup(revlog, x):
2049 if revlog == cl:
2046 if revlog == cl:
2050 c = cl.read(x)
2047 c = cl.read(x)
2051 changedfiles.update(c[3])
2048 changedfiles.update(c[3])
2052 mfs.setdefault(c[0], x)
2049 mfs.setdefault(c[0], x)
2053 count[0] += 1
2050 count[0] += 1
2054 progress(_bundling, count[0],
2051 progress(_bundling, count[0],
2055 unit=_changesets, total=count[1])
2052 unit=_changesets, total=count[1])
2056 return x
2053 return x
2057 elif revlog == mf:
2054 elif revlog == mf:
2058 clnode = mfs[x]
2055 clnode = mfs[x]
2059 mdata = mf.readfast(x)
2056 mdata = mf.readfast(x)
2060 for f, n in mdata.iteritems():
2057 for f, n in mdata.iteritems():
2061 if f in changedfiles:
2058 if f in changedfiles:
2062 fnodes[f].setdefault(n, clnode)
2059 fnodes[f].setdefault(n, clnode)
2063 count[0] += 1
2060 count[0] += 1
2064 progress(_bundling, count[0],
2061 progress(_bundling, count[0],
2065 unit=_manifests, total=count[1])
2062 unit=_manifests, total=count[1])
2066 return clnode
2063 return clnode
2067 else:
2064 else:
2068 progress(_bundling, count[0], item=fstate[0],
2065 progress(_bundling, count[0], item=fstate[0],
2069 unit=_files, total=count[1])
2066 unit=_files, total=count[1])
2070 return fstate[1][x]
2067 return fstate[1][x]
2071
2068
2072 bundler = changegroup.bundle10(lookup)
2069 bundler = changegroup.bundle10(lookup)
2073 reorder = self.ui.config('bundle', 'reorder', 'auto')
2070 reorder = self.ui.config('bundle', 'reorder', 'auto')
2074 if reorder == 'auto':
2071 if reorder == 'auto':
2075 reorder = None
2072 reorder = None
2076 else:
2073 else:
2077 reorder = util.parsebool(reorder)
2074 reorder = util.parsebool(reorder)
2078
2075
2079 def gengroup():
2076 def gengroup():
2080 # Create a changenode group generator that will call our functions
2077 # Create a changenode group generator that will call our functions
2081 # back to lookup the owning changenode and collect information.
2078 # back to lookup the owning changenode and collect information.
2082 count[:] = [0, len(csets)]
2079 count[:] = [0, len(csets)]
2083 for chunk in cl.group(csets, bundler, reorder=reorder):
2080 for chunk in cl.group(csets, bundler, reorder=reorder):
2084 yield chunk
2081 yield chunk
2085 progress(_bundling, None)
2082 progress(_bundling, None)
2086
2083
2087 # Create a generator for the manifestnodes that calls our lookup
2084 # Create a generator for the manifestnodes that calls our lookup
2088 # and data collection functions back.
2085 # and data collection functions back.
2089 for f in changedfiles:
2086 for f in changedfiles:
2090 fnodes[f] = {}
2087 fnodes[f] = {}
2091 count[:] = [0, len(mfs)]
2088 count[:] = [0, len(mfs)]
2092 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2089 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2093 yield chunk
2090 yield chunk
2094 progress(_bundling, None)
2091 progress(_bundling, None)
2095
2092
2096 mfs.clear()
2093 mfs.clear()
2097
2094
2098 # Go through all our files in order sorted by name.
2095 # Go through all our files in order sorted by name.
2099 count[:] = [0, len(changedfiles)]
2096 count[:] = [0, len(changedfiles)]
2100 for fname in sorted(changedfiles):
2097 for fname in sorted(changedfiles):
2101 filerevlog = self.file(fname)
2098 filerevlog = self.file(fname)
2102 if not len(filerevlog):
2099 if not len(filerevlog):
2103 raise util.Abort(_("empty or missing revlog for %s")
2100 raise util.Abort(_("empty or missing revlog for %s")
2104 % fname)
2101 % fname)
2105 fstate[0] = fname
2102 fstate[0] = fname
2106 fstate[1] = fnodes.pop(fname, {})
2103 fstate[1] = fnodes.pop(fname, {})
2107
2104
2108 nodelist = prune(filerevlog, fstate[1])
2105 nodelist = prune(filerevlog, fstate[1])
2109 if nodelist:
2106 if nodelist:
2110 count[0] += 1
2107 count[0] += 1
2111 yield bundler.fileheader(fname)
2108 yield bundler.fileheader(fname)
2112 for chunk in filerevlog.group(nodelist, bundler, reorder):
2109 for chunk in filerevlog.group(nodelist, bundler, reorder):
2113 yield chunk
2110 yield chunk
2114
2111
2115 # Signal that no more groups are left.
2112 # Signal that no more groups are left.
2116 yield bundler.close()
2113 yield bundler.close()
2117 progress(_bundling, None)
2114 progress(_bundling, None)
2118
2115
2119 if csets:
2116 if csets:
2120 self.hook('outgoing', node=hex(csets[0]), source=source)
2117 self.hook('outgoing', node=hex(csets[0]), source=source)
2121
2118
2122 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2119 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2123
2120
2124 def changegroup(self, basenodes, source):
2121 def changegroup(self, basenodes, source):
2125 # to avoid a race we use changegroupsubset() (issue1320)
2122 # to avoid a race we use changegroupsubset() (issue1320)
2126 return self.changegroupsubset(basenodes, self.heads(), source)
2123 return self.changegroupsubset(basenodes, self.heads(), source)
2127
2124
2128 @unfilteredmethod
2125 @unfilteredmethod
2129 def _changegroup(self, nodes, source):
2126 def _changegroup(self, nodes, source):
2130 """Compute the changegroup of all nodes that we have that a recipient
2127 """Compute the changegroup of all nodes that we have that a recipient
2131 doesn't. Return a chunkbuffer object whose read() method will return
2128 doesn't. Return a chunkbuffer object whose read() method will return
2132 successive changegroup chunks.
2129 successive changegroup chunks.
2133
2130
2134 This is much easier than the previous function as we can assume that
2131 This is much easier than the previous function as we can assume that
2135 the recipient has any changenode we aren't sending them.
2132 the recipient has any changenode we aren't sending them.
2136
2133
2137 nodes is the set of nodes to send"""
2134 nodes is the set of nodes to send"""
2138
2135
2139 cl = self.changelog
2136 cl = self.changelog
2140 mf = self.manifest
2137 mf = self.manifest
2141 mfs = {}
2138 mfs = {}
2142 changedfiles = set()
2139 changedfiles = set()
2143 fstate = ['']
2140 fstate = ['']
2144 count = [0, 0]
2141 count = [0, 0]
2145
2142
2146 self.hook('preoutgoing', throw=True, source=source)
2143 self.hook('preoutgoing', throw=True, source=source)
2147 self.changegroupinfo(nodes, source)
2144 self.changegroupinfo(nodes, source)
2148
2145
2149 revset = set([cl.rev(n) for n in nodes])
2146 revset = set([cl.rev(n) for n in nodes])
2150
2147
2151 def gennodelst(log):
2148 def gennodelst(log):
2152 ln, llr = log.node, log.linkrev
2149 ln, llr = log.node, log.linkrev
2153 return [ln(r) for r in log if llr(r) in revset]
2150 return [ln(r) for r in log if llr(r) in revset]
2154
2151
2155 progress = self.ui.progress
2152 progress = self.ui.progress
2156 _bundling = _('bundling')
2153 _bundling = _('bundling')
2157 _changesets = _('changesets')
2154 _changesets = _('changesets')
2158 _manifests = _('manifests')
2155 _manifests = _('manifests')
2159 _files = _('files')
2156 _files = _('files')
2160
2157
2161 def lookup(revlog, x):
2158 def lookup(revlog, x):
2162 if revlog == cl:
2159 if revlog == cl:
2163 c = cl.read(x)
2160 c = cl.read(x)
2164 changedfiles.update(c[3])
2161 changedfiles.update(c[3])
2165 mfs.setdefault(c[0], x)
2162 mfs.setdefault(c[0], x)
2166 count[0] += 1
2163 count[0] += 1
2167 progress(_bundling, count[0],
2164 progress(_bundling, count[0],
2168 unit=_changesets, total=count[1])
2165 unit=_changesets, total=count[1])
2169 return x
2166 return x
2170 elif revlog == mf:
2167 elif revlog == mf:
2171 count[0] += 1
2168 count[0] += 1
2172 progress(_bundling, count[0],
2169 progress(_bundling, count[0],
2173 unit=_manifests, total=count[1])
2170 unit=_manifests, total=count[1])
2174 return cl.node(revlog.linkrev(revlog.rev(x)))
2171 return cl.node(revlog.linkrev(revlog.rev(x)))
2175 else:
2172 else:
2176 progress(_bundling, count[0], item=fstate[0],
2173 progress(_bundling, count[0], item=fstate[0],
2177 total=count[1], unit=_files)
2174 total=count[1], unit=_files)
2178 return cl.node(revlog.linkrev(revlog.rev(x)))
2175 return cl.node(revlog.linkrev(revlog.rev(x)))
2179
2176
2180 bundler = changegroup.bundle10(lookup)
2177 bundler = changegroup.bundle10(lookup)
2181 reorder = self.ui.config('bundle', 'reorder', 'auto')
2178 reorder = self.ui.config('bundle', 'reorder', 'auto')
2182 if reorder == 'auto':
2179 if reorder == 'auto':
2183 reorder = None
2180 reorder = None
2184 else:
2181 else:
2185 reorder = util.parsebool(reorder)
2182 reorder = util.parsebool(reorder)
2186
2183
2187 def gengroup():
2184 def gengroup():
2188 '''yield a sequence of changegroup chunks (strings)'''
2185 '''yield a sequence of changegroup chunks (strings)'''
2189 # construct a list of all changed files
2186 # construct a list of all changed files
2190
2187
2191 count[:] = [0, len(nodes)]
2188 count[:] = [0, len(nodes)]
2192 for chunk in cl.group(nodes, bundler, reorder=reorder):
2189 for chunk in cl.group(nodes, bundler, reorder=reorder):
2193 yield chunk
2190 yield chunk
2194 progress(_bundling, None)
2191 progress(_bundling, None)
2195
2192
2196 count[:] = [0, len(mfs)]
2193 count[:] = [0, len(mfs)]
2197 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2194 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2198 yield chunk
2195 yield chunk
2199 progress(_bundling, None)
2196 progress(_bundling, None)
2200
2197
2201 count[:] = [0, len(changedfiles)]
2198 count[:] = [0, len(changedfiles)]
2202 for fname in sorted(changedfiles):
2199 for fname in sorted(changedfiles):
2203 filerevlog = self.file(fname)
2200 filerevlog = self.file(fname)
2204 if not len(filerevlog):
2201 if not len(filerevlog):
2205 raise util.Abort(_("empty or missing revlog for %s")
2202 raise util.Abort(_("empty or missing revlog for %s")
2206 % fname)
2203 % fname)
2207 fstate[0] = fname
2204 fstate[0] = fname
2208 nodelist = gennodelst(filerevlog)
2205 nodelist = gennodelst(filerevlog)
2209 if nodelist:
2206 if nodelist:
2210 count[0] += 1
2207 count[0] += 1
2211 yield bundler.fileheader(fname)
2208 yield bundler.fileheader(fname)
2212 for chunk in filerevlog.group(nodelist, bundler, reorder):
2209 for chunk in filerevlog.group(nodelist, bundler, reorder):
2213 yield chunk
2210 yield chunk
2214 yield bundler.close()
2211 yield bundler.close()
2215 progress(_bundling, None)
2212 progress(_bundling, None)
2216
2213
2217 if nodes:
2214 if nodes:
2218 self.hook('outgoing', node=hex(nodes[0]), source=source)
2215 self.hook('outgoing', node=hex(nodes[0]), source=source)
2219
2216
2220 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2217 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2221
2218
2222 @unfilteredmethod
2219 @unfilteredmethod
2223 def addchangegroup(self, source, srctype, url, emptyok=False):
2220 def addchangegroup(self, source, srctype, url, emptyok=False):
2224 """Add the changegroup returned by source.read() to this repo.
2221 """Add the changegroup returned by source.read() to this repo.
2225 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2222 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2226 the URL of the repo where this changegroup is coming from.
2223 the URL of the repo where this changegroup is coming from.
2227
2224
2228 Return an integer summarizing the change to this repo:
2225 Return an integer summarizing the change to this repo:
2229 - nothing changed or no source: 0
2226 - nothing changed or no source: 0
2230 - more heads than before: 1+added heads (2..n)
2227 - more heads than before: 1+added heads (2..n)
2231 - fewer heads than before: -1-removed heads (-2..-n)
2228 - fewer heads than before: -1-removed heads (-2..-n)
2232 - number of heads stays the same: 1
2229 - number of heads stays the same: 1
2233 """
2230 """
2234 def csmap(x):
2231 def csmap(x):
2235 self.ui.debug("add changeset %s\n" % short(x))
2232 self.ui.debug("add changeset %s\n" % short(x))
2236 return len(cl)
2233 return len(cl)
2237
2234
2238 def revmap(x):
2235 def revmap(x):
2239 return cl.rev(x)
2236 return cl.rev(x)
2240
2237
2241 if not source:
2238 if not source:
2242 return 0
2239 return 0
2243
2240
2244 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2241 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2245
2242
2246 changesets = files = revisions = 0
2243 changesets = files = revisions = 0
2247 efiles = set()
2244 efiles = set()
2248
2245
2249 # write changelog data to temp files so concurrent readers will not see
2246 # write changelog data to temp files so concurrent readers will not see
2250 # inconsistent view
2247 # inconsistent view
2251 cl = self.changelog
2248 cl = self.changelog
2252 cl.delayupdate()
2249 cl.delayupdate()
2253 oldheads = cl.heads()
2250 oldheads = cl.heads()
2254
2251
2255 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2252 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2256 try:
2253 try:
2257 trp = weakref.proxy(tr)
2254 trp = weakref.proxy(tr)
2258 # pull off the changeset group
2255 # pull off the changeset group
2259 self.ui.status(_("adding changesets\n"))
2256 self.ui.status(_("adding changesets\n"))
2260 clstart = len(cl)
2257 clstart = len(cl)
2261 class prog(object):
2258 class prog(object):
2262 step = _('changesets')
2259 step = _('changesets')
2263 count = 1
2260 count = 1
2264 ui = self.ui
2261 ui = self.ui
2265 total = None
2262 total = None
2266 def __call__(self):
2263 def __call__(self):
2267 self.ui.progress(self.step, self.count, unit=_('chunks'),
2264 self.ui.progress(self.step, self.count, unit=_('chunks'),
2268 total=self.total)
2265 total=self.total)
2269 self.count += 1
2266 self.count += 1
2270 pr = prog()
2267 pr = prog()
2271 source.callback = pr
2268 source.callback = pr
2272
2269
2273 source.changelogheader()
2270 source.changelogheader()
2274 srccontent = cl.addgroup(source, csmap, trp)
2271 srccontent = cl.addgroup(source, csmap, trp)
2275 if not (srccontent or emptyok):
2272 if not (srccontent or emptyok):
2276 raise util.Abort(_("received changelog group is empty"))
2273 raise util.Abort(_("received changelog group is empty"))
2277 clend = len(cl)
2274 clend = len(cl)
2278 changesets = clend - clstart
2275 changesets = clend - clstart
2279 for c in xrange(clstart, clend):
2276 for c in xrange(clstart, clend):
2280 efiles.update(self[c].files())
2277 efiles.update(self[c].files())
2281 efiles = len(efiles)
2278 efiles = len(efiles)
2282 self.ui.progress(_('changesets'), None)
2279 self.ui.progress(_('changesets'), None)
2283
2280
2284 # pull off the manifest group
2281 # pull off the manifest group
2285 self.ui.status(_("adding manifests\n"))
2282 self.ui.status(_("adding manifests\n"))
2286 pr.step = _('manifests')
2283 pr.step = _('manifests')
2287 pr.count = 1
2284 pr.count = 1
2288 pr.total = changesets # manifests <= changesets
2285 pr.total = changesets # manifests <= changesets
2289 # no need to check for empty manifest group here:
2286 # no need to check for empty manifest group here:
2290 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2287 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2291 # no new manifest will be created and the manifest group will
2288 # no new manifest will be created and the manifest group will
2292 # be empty during the pull
2289 # be empty during the pull
2293 source.manifestheader()
2290 source.manifestheader()
2294 self.manifest.addgroup(source, revmap, trp)
2291 self.manifest.addgroup(source, revmap, trp)
2295 self.ui.progress(_('manifests'), None)
2292 self.ui.progress(_('manifests'), None)
2296
2293
2297 needfiles = {}
2294 needfiles = {}
2298 if self.ui.configbool('server', 'validate', default=False):
2295 if self.ui.configbool('server', 'validate', default=False):
2299 # validate incoming csets have their manifests
2296 # validate incoming csets have their manifests
2300 for cset in xrange(clstart, clend):
2297 for cset in xrange(clstart, clend):
2301 mfest = self.changelog.read(self.changelog.node(cset))[0]
2298 mfest = self.changelog.read(self.changelog.node(cset))[0]
2302 mfest = self.manifest.readdelta(mfest)
2299 mfest = self.manifest.readdelta(mfest)
2303 # store file nodes we must see
2300 # store file nodes we must see
2304 for f, n in mfest.iteritems():
2301 for f, n in mfest.iteritems():
2305 needfiles.setdefault(f, set()).add(n)
2302 needfiles.setdefault(f, set()).add(n)
2306
2303
2307 # process the files
2304 # process the files
2308 self.ui.status(_("adding file changes\n"))
2305 self.ui.status(_("adding file changes\n"))
2309 pr.step = _('files')
2306 pr.step = _('files')
2310 pr.count = 1
2307 pr.count = 1
2311 pr.total = efiles
2308 pr.total = efiles
2312 source.callback = None
2309 source.callback = None
2313
2310
2314 while True:
2311 while True:
2315 chunkdata = source.filelogheader()
2312 chunkdata = source.filelogheader()
2316 if not chunkdata:
2313 if not chunkdata:
2317 break
2314 break
2318 f = chunkdata["filename"]
2315 f = chunkdata["filename"]
2319 self.ui.debug("adding %s revisions\n" % f)
2316 self.ui.debug("adding %s revisions\n" % f)
2320 pr()
2317 pr()
2321 fl = self.file(f)
2318 fl = self.file(f)
2322 o = len(fl)
2319 o = len(fl)
2323 if not fl.addgroup(source, revmap, trp):
2320 if not fl.addgroup(source, revmap, trp):
2324 raise util.Abort(_("received file revlog group is empty"))
2321 raise util.Abort(_("received file revlog group is empty"))
2325 revisions += len(fl) - o
2322 revisions += len(fl) - o
2326 files += 1
2323 files += 1
2327 if f in needfiles:
2324 if f in needfiles:
2328 needs = needfiles[f]
2325 needs = needfiles[f]
2329 for new in xrange(o, len(fl)):
2326 for new in xrange(o, len(fl)):
2330 n = fl.node(new)
2327 n = fl.node(new)
2331 if n in needs:
2328 if n in needs:
2332 needs.remove(n)
2329 needs.remove(n)
2333 if not needs:
2330 if not needs:
2334 del needfiles[f]
2331 del needfiles[f]
2335 self.ui.progress(_('files'), None)
2332 self.ui.progress(_('files'), None)
2336
2333
2337 for f, needs in needfiles.iteritems():
2334 for f, needs in needfiles.iteritems():
2338 fl = self.file(f)
2335 fl = self.file(f)
2339 for n in needs:
2336 for n in needs:
2340 try:
2337 try:
2341 fl.rev(n)
2338 fl.rev(n)
2342 except error.LookupError:
2339 except error.LookupError:
2343 raise util.Abort(
2340 raise util.Abort(
2344 _('missing file data for %s:%s - run hg verify') %
2341 _('missing file data for %s:%s - run hg verify') %
2345 (f, hex(n)))
2342 (f, hex(n)))
2346
2343
2347 dh = 0
2344 dh = 0
2348 if oldheads:
2345 if oldheads:
2349 heads = cl.heads()
2346 heads = cl.heads()
2350 dh = len(heads) - len(oldheads)
2347 dh = len(heads) - len(oldheads)
2351 for h in heads:
2348 for h in heads:
2352 if h not in oldheads and self[h].closesbranch():
2349 if h not in oldheads and self[h].closesbranch():
2353 dh -= 1
2350 dh -= 1
2354 htext = ""
2351 htext = ""
2355 if dh:
2352 if dh:
2356 htext = _(" (%+d heads)") % dh
2353 htext = _(" (%+d heads)") % dh
2357
2354
2358 self.ui.status(_("added %d changesets"
2355 self.ui.status(_("added %d changesets"
2359 " with %d changes to %d files%s\n")
2356 " with %d changes to %d files%s\n")
2360 % (changesets, revisions, files, htext))
2357 % (changesets, revisions, files, htext))
2361 self.invalidatevolatilesets()
2358 self.invalidatevolatilesets()
2362
2359
2363 if changesets > 0:
2360 if changesets > 0:
2364 p = lambda: cl.writepending() and self.root or ""
2361 p = lambda: cl.writepending() and self.root or ""
2365 self.hook('pretxnchangegroup', throw=True,
2362 self.hook('pretxnchangegroup', throw=True,
2366 node=hex(cl.node(clstart)), source=srctype,
2363 node=hex(cl.node(clstart)), source=srctype,
2367 url=url, pending=p)
2364 url=url, pending=p)
2368
2365
2369 added = [cl.node(r) for r in xrange(clstart, clend)]
2366 added = [cl.node(r) for r in xrange(clstart, clend)]
2370 publishing = self.ui.configbool('phases', 'publish', True)
2367 publishing = self.ui.configbool('phases', 'publish', True)
2371 if srctype == 'push':
2368 if srctype == 'push':
2372 # Old server can not push the boundary themself.
2369 # Old server can not push the boundary themself.
2373 # New server won't push the boundary if changeset already
2370 # New server won't push the boundary if changeset already
2374 # existed locally as secrete
2371 # existed locally as secrete
2375 #
2372 #
2376 # We should not use added here but the list of all change in
2373 # We should not use added here but the list of all change in
2377 # the bundle
2374 # the bundle
2378 if publishing:
2375 if publishing:
2379 phases.advanceboundary(self, phases.public, srccontent)
2376 phases.advanceboundary(self, phases.public, srccontent)
2380 else:
2377 else:
2381 phases.advanceboundary(self, phases.draft, srccontent)
2378 phases.advanceboundary(self, phases.draft, srccontent)
2382 phases.retractboundary(self, phases.draft, added)
2379 phases.retractboundary(self, phases.draft, added)
2383 elif srctype != 'strip':
2380 elif srctype != 'strip':
2384 # publishing only alter behavior during push
2381 # publishing only alter behavior during push
2385 #
2382 #
2386 # strip should not touch boundary at all
2383 # strip should not touch boundary at all
2387 phases.retractboundary(self, phases.draft, added)
2384 phases.retractboundary(self, phases.draft, added)
2388
2385
2389 # make changelog see real files again
2386 # make changelog see real files again
2390 cl.finalize(trp)
2387 cl.finalize(trp)
2391
2388
2392 tr.close()
2389 tr.close()
2393
2390
2394 if changesets > 0:
2391 if changesets > 0:
2395 if srctype != 'strip':
2392 if srctype != 'strip':
2396 # During strip, branchcache is invalid but coming call to
2393 # During strip, branchcache is invalid but coming call to
2397 # `destroyed` will repair it.
2394 # `destroyed` will repair it.
2398 # In other case we can safely update cache on disk.
2395 # In other case we can safely update cache on disk.
2399 branchmap.updatecache(self)
2396 branchmap.updatecache(self)
2400 def runhooks():
2397 def runhooks():
2401 # forcefully update the on-disk branch cache
2398 # forcefully update the on-disk branch cache
2402 self.ui.debug("updating the branch cache\n")
2399 self.ui.debug("updating the branch cache\n")
2403 self.hook("changegroup", node=hex(cl.node(clstart)),
2400 self.hook("changegroup", node=hex(cl.node(clstart)),
2404 source=srctype, url=url)
2401 source=srctype, url=url)
2405
2402
2406 for n in added:
2403 for n in added:
2407 self.hook("incoming", node=hex(n), source=srctype,
2404 self.hook("incoming", node=hex(n), source=srctype,
2408 url=url)
2405 url=url)
2409 self._afterlock(runhooks)
2406 self._afterlock(runhooks)
2410
2407
2411 finally:
2408 finally:
2412 tr.release()
2409 tr.release()
2413 # never return 0 here:
2410 # never return 0 here:
2414 if dh < 0:
2411 if dh < 0:
2415 return dh - 1
2412 return dh - 1
2416 else:
2413 else:
2417 return dh + 1
2414 return dh + 1
2418
2415
2419 def stream_in(self, remote, requirements):
2416 def stream_in(self, remote, requirements):
2420 lock = self.lock()
2417 lock = self.lock()
2421 try:
2418 try:
2422 # Save remote branchmap. We will use it later
2419 # Save remote branchmap. We will use it later
2423 # to speed up branchcache creation
2420 # to speed up branchcache creation
2424 rbranchmap = None
2421 rbranchmap = None
2425 if remote.capable("branchmap"):
2422 if remote.capable("branchmap"):
2426 rbranchmap = remote.branchmap()
2423 rbranchmap = remote.branchmap()
2427
2424
2428 fp = remote.stream_out()
2425 fp = remote.stream_out()
2429 l = fp.readline()
2426 l = fp.readline()
2430 try:
2427 try:
2431 resp = int(l)
2428 resp = int(l)
2432 except ValueError:
2429 except ValueError:
2433 raise error.ResponseError(
2430 raise error.ResponseError(
2434 _('unexpected response from remote server:'), l)
2431 _('unexpected response from remote server:'), l)
2435 if resp == 1:
2432 if resp == 1:
2436 raise util.Abort(_('operation forbidden by server'))
2433 raise util.Abort(_('operation forbidden by server'))
2437 elif resp == 2:
2434 elif resp == 2:
2438 raise util.Abort(_('locking the remote repository failed'))
2435 raise util.Abort(_('locking the remote repository failed'))
2439 elif resp != 0:
2436 elif resp != 0:
2440 raise util.Abort(_('the server sent an unknown error code'))
2437 raise util.Abort(_('the server sent an unknown error code'))
2441 self.ui.status(_('streaming all changes\n'))
2438 self.ui.status(_('streaming all changes\n'))
2442 l = fp.readline()
2439 l = fp.readline()
2443 try:
2440 try:
2444 total_files, total_bytes = map(int, l.split(' ', 1))
2441 total_files, total_bytes = map(int, l.split(' ', 1))
2445 except (ValueError, TypeError):
2442 except (ValueError, TypeError):
2446 raise error.ResponseError(
2443 raise error.ResponseError(
2447 _('unexpected response from remote server:'), l)
2444 _('unexpected response from remote server:'), l)
2448 self.ui.status(_('%d files to transfer, %s of data\n') %
2445 self.ui.status(_('%d files to transfer, %s of data\n') %
2449 (total_files, util.bytecount(total_bytes)))
2446 (total_files, util.bytecount(total_bytes)))
2450 handled_bytes = 0
2447 handled_bytes = 0
2451 self.ui.progress(_('clone'), 0, total=total_bytes)
2448 self.ui.progress(_('clone'), 0, total=total_bytes)
2452 start = time.time()
2449 start = time.time()
2453 for i in xrange(total_files):
2450 for i in xrange(total_files):
2454 # XXX doesn't support '\n' or '\r' in filenames
2451 # XXX doesn't support '\n' or '\r' in filenames
2455 l = fp.readline()
2452 l = fp.readline()
2456 try:
2453 try:
2457 name, size = l.split('\0', 1)
2454 name, size = l.split('\0', 1)
2458 size = int(size)
2455 size = int(size)
2459 except (ValueError, TypeError):
2456 except (ValueError, TypeError):
2460 raise error.ResponseError(
2457 raise error.ResponseError(
2461 _('unexpected response from remote server:'), l)
2458 _('unexpected response from remote server:'), l)
2462 if self.ui.debugflag:
2459 if self.ui.debugflag:
2463 self.ui.debug('adding %s (%s)\n' %
2460 self.ui.debug('adding %s (%s)\n' %
2464 (name, util.bytecount(size)))
2461 (name, util.bytecount(size)))
2465 # for backwards compat, name was partially encoded
2462 # for backwards compat, name was partially encoded
2466 ofp = self.sopener(store.decodedir(name), 'w')
2463 ofp = self.sopener(store.decodedir(name), 'w')
2467 for chunk in util.filechunkiter(fp, limit=size):
2464 for chunk in util.filechunkiter(fp, limit=size):
2468 handled_bytes += len(chunk)
2465 handled_bytes += len(chunk)
2469 self.ui.progress(_('clone'), handled_bytes,
2466 self.ui.progress(_('clone'), handled_bytes,
2470 total=total_bytes)
2467 total=total_bytes)
2471 ofp.write(chunk)
2468 ofp.write(chunk)
2472 ofp.close()
2469 ofp.close()
2473 elapsed = time.time() - start
2470 elapsed = time.time() - start
2474 if elapsed <= 0:
2471 if elapsed <= 0:
2475 elapsed = 0.001
2472 elapsed = 0.001
2476 self.ui.progress(_('clone'), None)
2473 self.ui.progress(_('clone'), None)
2477 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2474 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2478 (util.bytecount(total_bytes), elapsed,
2475 (util.bytecount(total_bytes), elapsed,
2479 util.bytecount(total_bytes / elapsed)))
2476 util.bytecount(total_bytes / elapsed)))
2480
2477
2481 # new requirements = old non-format requirements +
2478 # new requirements = old non-format requirements +
2482 # new format-related
2479 # new format-related
2483 # requirements from the streamed-in repository
2480 # requirements from the streamed-in repository
2484 requirements.update(set(self.requirements) - self.supportedformats)
2481 requirements.update(set(self.requirements) - self.supportedformats)
2485 self._applyrequirements(requirements)
2482 self._applyrequirements(requirements)
2486 self._writerequirements()
2483 self._writerequirements()
2487
2484
2488 if rbranchmap:
2485 if rbranchmap:
2489 rbheads = []
2486 rbheads = []
2490 for bheads in rbranchmap.itervalues():
2487 for bheads in rbranchmap.itervalues():
2491 rbheads.extend(bheads)
2488 rbheads.extend(bheads)
2492
2489
2493 if rbheads:
2490 if rbheads:
2494 rtiprev = max((int(self.changelog.rev(node))
2491 rtiprev = max((int(self.changelog.rev(node))
2495 for node in rbheads))
2492 for node in rbheads))
2496 cache = branchmap.branchcache(rbranchmap,
2493 cache = branchmap.branchcache(rbranchmap,
2497 self[rtiprev].node(),
2494 self[rtiprev].node(),
2498 rtiprev)
2495 rtiprev)
2499 self._branchcaches[None] = cache
2496 self._branchcaches[None] = cache
2500 cache.write(self.unfiltered())
2497 cache.write(self.unfiltered())
2501 self.invalidate()
2498 self.invalidate()
2502 return len(self.heads()) + 1
2499 return len(self.heads()) + 1
2503 finally:
2500 finally:
2504 lock.release()
2501 lock.release()
2505
2502
2506 def clone(self, remote, heads=[], stream=False):
2503 def clone(self, remote, heads=[], stream=False):
2507 '''clone remote repository.
2504 '''clone remote repository.
2508
2505
2509 keyword arguments:
2506 keyword arguments:
2510 heads: list of revs to clone (forces use of pull)
2507 heads: list of revs to clone (forces use of pull)
2511 stream: use streaming clone if possible'''
2508 stream: use streaming clone if possible'''
2512
2509
2513 # now, all clients that can request uncompressed clones can
2510 # now, all clients that can request uncompressed clones can
2514 # read repo formats supported by all servers that can serve
2511 # read repo formats supported by all servers that can serve
2515 # them.
2512 # them.
2516
2513
2517 # if revlog format changes, client will have to check version
2514 # if revlog format changes, client will have to check version
2518 # and format flags on "stream" capability, and use
2515 # and format flags on "stream" capability, and use
2519 # uncompressed only if compatible.
2516 # uncompressed only if compatible.
2520
2517
2521 if not stream:
2518 if not stream:
2522 # if the server explicitly prefers to stream (for fast LANs)
2519 # if the server explicitly prefers to stream (for fast LANs)
2523 stream = remote.capable('stream-preferred')
2520 stream = remote.capable('stream-preferred')
2524
2521
2525 if stream and not heads:
2522 if stream and not heads:
2526 # 'stream' means remote revlog format is revlogv1 only
2523 # 'stream' means remote revlog format is revlogv1 only
2527 if remote.capable('stream'):
2524 if remote.capable('stream'):
2528 return self.stream_in(remote, set(('revlogv1',)))
2525 return self.stream_in(remote, set(('revlogv1',)))
2529 # otherwise, 'streamreqs' contains the remote revlog format
2526 # otherwise, 'streamreqs' contains the remote revlog format
2530 streamreqs = remote.capable('streamreqs')
2527 streamreqs = remote.capable('streamreqs')
2531 if streamreqs:
2528 if streamreqs:
2532 streamreqs = set(streamreqs.split(','))
2529 streamreqs = set(streamreqs.split(','))
2533 # if we support it, stream in and adjust our requirements
2530 # if we support it, stream in and adjust our requirements
2534 if not streamreqs - self.supportedformats:
2531 if not streamreqs - self.supportedformats:
2535 return self.stream_in(remote, streamreqs)
2532 return self.stream_in(remote, streamreqs)
2536 return self.pull(remote, heads)
2533 return self.pull(remote, heads)
2537
2534
2538 def pushkey(self, namespace, key, old, new):
2535 def pushkey(self, namespace, key, old, new):
2539 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2536 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2540 old=old, new=new)
2537 old=old, new=new)
2541 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2538 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2542 ret = pushkey.push(self, namespace, key, old, new)
2539 ret = pushkey.push(self, namespace, key, old, new)
2543 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2540 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2544 ret=ret)
2541 ret=ret)
2545 return ret
2542 return ret
2546
2543
2547 def listkeys(self, namespace):
2544 def listkeys(self, namespace):
2548 self.hook('prelistkeys', throw=True, namespace=namespace)
2545 self.hook('prelistkeys', throw=True, namespace=namespace)
2549 self.ui.debug('listing keys for "%s"\n' % namespace)
2546 self.ui.debug('listing keys for "%s"\n' % namespace)
2550 values = pushkey.list(self, namespace)
2547 values = pushkey.list(self, namespace)
2551 self.hook('listkeys', namespace=namespace, values=values)
2548 self.hook('listkeys', namespace=namespace, values=values)
2552 return values
2549 return values
2553
2550
2554 def debugwireargs(self, one, two, three=None, four=None, five=None):
2551 def debugwireargs(self, one, two, three=None, four=None, five=None):
2555 '''used to test argument passing over the wire'''
2552 '''used to test argument passing over the wire'''
2556 return "%s %s %s %s %s" % (one, two, three, four, five)
2553 return "%s %s %s %s %s" % (one, two, three, four, five)
2557
2554
2558 def savecommitmessage(self, text):
2555 def savecommitmessage(self, text):
2559 fp = self.opener('last-message.txt', 'wb')
2556 fp = self.opener('last-message.txt', 'wb')
2560 try:
2557 try:
2561 fp.write(text)
2558 fp.write(text)
2562 finally:
2559 finally:
2563 fp.close()
2560 fp.close()
2564 return self.pathto(fp.name[len(self.root) + 1:])
2561 return self.pathto(fp.name[len(self.root) + 1:])
2565
2562
2566 # used to avoid circular references so destructors work
2563 # used to avoid circular references so destructors work
2567 def aftertrans(files):
2564 def aftertrans(files):
2568 renamefiles = [tuple(t) for t in files]
2565 renamefiles = [tuple(t) for t in files]
2569 def a():
2566 def a():
2570 for src, dest in renamefiles:
2567 for src, dest in renamefiles:
2571 try:
2568 try:
2572 util.rename(src, dest)
2569 util.rename(src, dest)
2573 except OSError: # journal file does not yet exist
2570 except OSError: # journal file does not yet exist
2574 pass
2571 pass
2575 return a
2572 return a
2576
2573
2577 def undoname(fn):
2574 def undoname(fn):
2578 base, name = os.path.split(fn)
2575 base, name = os.path.split(fn)
2579 assert name.startswith('journal')
2576 assert name.startswith('journal')
2580 return os.path.join(base, name.replace('journal', 'undo', 1))
2577 return os.path.join(base, name.replace('journal', 'undo', 1))
2581
2578
2582 def instance(ui, path, create):
2579 def instance(ui, path, create):
2583 return localrepository(ui, util.urllocalpath(path), create)
2580 return localrepository(ui, util.urllocalpath(path), create)
2584
2581
2585 def islocal(path):
2582 def islocal(path):
2586 return True
2583 return True
@@ -1,95 +1,111 b''
1 import sys, os, subprocess
1 import sys, os, subprocess
2
2
3 if subprocess.call(['python', '%s/hghave' % os.environ['TESTDIR'],
3 if subprocess.call(['python', '%s/hghave' % os.environ['TESTDIR'],
4 'cacheable']):
4 'cacheable']):
5 sys.exit(80)
5 sys.exit(80)
6
6
7 from mercurial import util, scmutil, extensions
7 from mercurial import util, scmutil, extensions, hg, ui
8
8
9 filecache = scmutil.filecache
9 filecache = scmutil.filecache
10
10
11 class fakerepo(object):
11 class fakerepo(object):
12 def __init__(self):
12 def __init__(self):
13 self._filecache = {}
13 self._filecache = {}
14
14
15 def join(self, p):
15 def join(self, p):
16 return p
16 return p
17
17
18 def sjoin(self, p):
18 def sjoin(self, p):
19 return p
19 return p
20
20
21 @filecache('x')
21 @filecache('x')
22 def cached(self):
22 def cached(self):
23 print 'creating'
23 print 'creating'
24
24
25 def invalidate(self):
25 def invalidate(self):
26 for k in self._filecache:
26 for k in self._filecache:
27 try:
27 try:
28 delattr(self, k)
28 delattr(self, k)
29 except AttributeError:
29 except AttributeError:
30 pass
30 pass
31
31
32 def basic(repo):
32 def basic(repo):
33 # file doesn't exist, calls function
33 # file doesn't exist, calls function
34 repo.cached
34 repo.cached
35
35
36 repo.invalidate()
36 repo.invalidate()
37 # file still doesn't exist, uses cache
37 # file still doesn't exist, uses cache
38 repo.cached
38 repo.cached
39
39
40 # create empty file
40 # create empty file
41 f = open('x', 'w')
41 f = open('x', 'w')
42 f.close()
42 f.close()
43 repo.invalidate()
43 repo.invalidate()
44 # should recreate the object
44 # should recreate the object
45 repo.cached
45 repo.cached
46
46
47 f = open('x', 'w')
47 f = open('x', 'w')
48 f.write('a')
48 f.write('a')
49 f.close()
49 f.close()
50 repo.invalidate()
50 repo.invalidate()
51 # should recreate the object
51 # should recreate the object
52 repo.cached
52 repo.cached
53
53
54 repo.invalidate()
54 repo.invalidate()
55 # stats file again, nothing changed, reuses object
55 # stats file again, nothing changed, reuses object
56 repo.cached
56 repo.cached
57
57
58 # atomic replace file, size doesn't change
58 # atomic replace file, size doesn't change
59 # hopefully st_mtime doesn't change as well so this doesn't use the cache
59 # hopefully st_mtime doesn't change as well so this doesn't use the cache
60 # because of inode change
60 # because of inode change
61 f = scmutil.opener('.')('x', 'w', atomictemp=True)
61 f = scmutil.opener('.')('x', 'w', atomictemp=True)
62 f.write('b')
62 f.write('b')
63 f.close()
63 f.close()
64
64
65 repo.invalidate()
65 repo.invalidate()
66 repo.cached
66 repo.cached
67
67
68 def fakeuncacheable():
68 def fakeuncacheable():
69 def wrapcacheable(orig, *args, **kwargs):
69 def wrapcacheable(orig, *args, **kwargs):
70 return False
70 return False
71
71
72 def wrapinit(orig, *args, **kwargs):
72 def wrapinit(orig, *args, **kwargs):
73 pass
73 pass
74
74
75 originit = extensions.wrapfunction(util.cachestat, '__init__', wrapinit)
75 originit = extensions.wrapfunction(util.cachestat, '__init__', wrapinit)
76 origcacheable = extensions.wrapfunction(util.cachestat, 'cacheable',
76 origcacheable = extensions.wrapfunction(util.cachestat, 'cacheable',
77 wrapcacheable)
77 wrapcacheable)
78
78
79 try:
79 try:
80 os.remove('x')
80 os.remove('x')
81 except OSError:
81 except OSError:
82 pass
82 pass
83
83
84 basic(fakerepo())
84 basic(fakerepo())
85
85
86 util.cachestat.cacheable = origcacheable
86 util.cachestat.cacheable = origcacheable
87 util.cachestat.__init__ = originit
87 util.cachestat.__init__ = originit
88
88
89 def test_filecache_synced():
90 # test old behaviour that caused filecached properties to go out of sync
91 os.system('hg init && echo a >> a && hg ci -qAm.')
92 repo = hg.repository(ui.ui())
93 # first rollback clears the filecache, but changelog to stays in __dict__
94 repo.rollback()
95 repo.commit('.')
96 # second rollback comes along and touches the changelog externally
97 # (file is moved)
98 repo.rollback()
99 # but since changelog isn't under the filecache control anymore, we don't
100 # see that it changed, and return the old changelog without reconstructing
101 # it
102 repo.commit('.')
103
89 print 'basic:'
104 print 'basic:'
90 print
105 print
91 basic(fakerepo())
106 basic(fakerepo())
92 print
107 print
93 print 'fakeuncacheable:'
108 print 'fakeuncacheable:'
94 print
109 print
95 fakeuncacheable()
110 fakeuncacheable()
111 test_filecache_synced()
@@ -1,15 +1,19 b''
1 basic:
1 basic:
2
2
3 creating
3 creating
4 creating
4 creating
5 creating
5 creating
6 creating
6 creating
7
7
8 fakeuncacheable:
8 fakeuncacheable:
9
9
10 creating
10 creating
11 creating
11 creating
12 creating
12 creating
13 creating
13 creating
14 creating
14 creating
15 creating
15 creating
16 repository tip rolled back to revision -1 (undo commit)
17 working directory now based on revision -1
18 repository tip rolled back to revision -1 (undo commit)
19 working directory now based on revision -1
General Comments 0
You need to be logged in to leave comments. Login now