##// END OF EJS Templates
localrepo: update the branchmap when destroying nodes...
Idan Kamara -
r18311:08407464 default
parent child Browse files
Show More
@@ -1,2577 +1,2580 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import lock, transaction, store, encoding, base85
11 import lock, transaction, store, encoding, base85
12 import scmutil, util, extensions, hook, error, revset
12 import scmutil, util, extensions, hook, error, revset
13 import match as matchmod
13 import match as matchmod
14 import merge as mergemod
14 import merge as mergemod
15 import tags as tagsmod
15 import tags as tagsmod
16 from lock import release
16 from lock import release
17 import weakref, errno, os, time, inspect
17 import weakref, errno, os, time, inspect
18 import branchmap
18 import branchmap
19 propertycache = util.propertycache
19 propertycache = util.propertycache
20 filecache = scmutil.filecache
20 filecache = scmutil.filecache
21
21
22 class repofilecache(filecache):
22 class repofilecache(filecache):
23 """All filecache usage on repo are done for logic that should be unfiltered
23 """All filecache usage on repo are done for logic that should be unfiltered
24 """
24 """
25
25
26 def __get__(self, repo, type=None):
26 def __get__(self, repo, type=None):
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 def __set__(self, repo, value):
28 def __set__(self, repo, value):
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 def __delete__(self, repo):
30 def __delete__(self, repo):
31 return super(repofilecache, self).__delete__(repo.unfiltered())
31 return super(repofilecache, self).__delete__(repo.unfiltered())
32
32
33 class storecache(repofilecache):
33 class storecache(repofilecache):
34 """filecache for files in the store"""
34 """filecache for files in the store"""
35 def join(self, obj, fname):
35 def join(self, obj, fname):
36 return obj.sjoin(fname)
36 return obj.sjoin(fname)
37
37
38 class unfilteredpropertycache(propertycache):
38 class unfilteredpropertycache(propertycache):
39 """propertycache that apply to unfiltered repo only"""
39 """propertycache that apply to unfiltered repo only"""
40
40
41 def __get__(self, repo, type=None):
41 def __get__(self, repo, type=None):
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43
43
44 class filteredpropertycache(propertycache):
44 class filteredpropertycache(propertycache):
45 """propertycache that must take filtering in account"""
45 """propertycache that must take filtering in account"""
46
46
47 def cachevalue(self, obj, value):
47 def cachevalue(self, obj, value):
48 object.__setattr__(obj, self.name, value)
48 object.__setattr__(obj, self.name, value)
49
49
50
50
51 def hasunfilteredcache(repo, name):
51 def hasunfilteredcache(repo, name):
52 """check if an repo and a unfilteredproperty cached value for <name>"""
52 """check if an repo and a unfilteredproperty cached value for <name>"""
53 return name in vars(repo.unfiltered())
53 return name in vars(repo.unfiltered())
54
54
55 def unfilteredmethod(orig):
55 def unfilteredmethod(orig):
56 """decorate method that always need to be run on unfiltered version"""
56 """decorate method that always need to be run on unfiltered version"""
57 def wrapper(repo, *args, **kwargs):
57 def wrapper(repo, *args, **kwargs):
58 return orig(repo.unfiltered(), *args, **kwargs)
58 return orig(repo.unfiltered(), *args, **kwargs)
59 return wrapper
59 return wrapper
60
60
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63
63
64 class localpeer(peer.peerrepository):
64 class localpeer(peer.peerrepository):
65 '''peer for a local repo; reflects only the most recent API'''
65 '''peer for a local repo; reflects only the most recent API'''
66
66
67 def __init__(self, repo, caps=MODERNCAPS):
67 def __init__(self, repo, caps=MODERNCAPS):
68 peer.peerrepository.__init__(self)
68 peer.peerrepository.__init__(self)
69 self._repo = repo.filtered('unserved')
69 self._repo = repo.filtered('unserved')
70 self.ui = repo.ui
70 self.ui = repo.ui
71 self._caps = repo._restrictcapabilities(caps)
71 self._caps = repo._restrictcapabilities(caps)
72 self.requirements = repo.requirements
72 self.requirements = repo.requirements
73 self.supportedformats = repo.supportedformats
73 self.supportedformats = repo.supportedformats
74
74
75 def close(self):
75 def close(self):
76 self._repo.close()
76 self._repo.close()
77
77
78 def _capabilities(self):
78 def _capabilities(self):
79 return self._caps
79 return self._caps
80
80
81 def local(self):
81 def local(self):
82 return self._repo
82 return self._repo
83
83
84 def canpush(self):
84 def canpush(self):
85 return True
85 return True
86
86
87 def url(self):
87 def url(self):
88 return self._repo.url()
88 return self._repo.url()
89
89
90 def lookup(self, key):
90 def lookup(self, key):
91 return self._repo.lookup(key)
91 return self._repo.lookup(key)
92
92
93 def branchmap(self):
93 def branchmap(self):
94 return self._repo.branchmap()
94 return self._repo.branchmap()
95
95
96 def heads(self):
96 def heads(self):
97 return self._repo.heads()
97 return self._repo.heads()
98
98
99 def known(self, nodes):
99 def known(self, nodes):
100 return self._repo.known(nodes)
100 return self._repo.known(nodes)
101
101
102 def getbundle(self, source, heads=None, common=None):
102 def getbundle(self, source, heads=None, common=None):
103 return self._repo.getbundle(source, heads=heads, common=common)
103 return self._repo.getbundle(source, heads=heads, common=common)
104
104
105 # TODO We might want to move the next two calls into legacypeer and add
105 # TODO We might want to move the next two calls into legacypeer and add
106 # unbundle instead.
106 # unbundle instead.
107
107
108 def lock(self):
108 def lock(self):
109 return self._repo.lock()
109 return self._repo.lock()
110
110
111 def addchangegroup(self, cg, source, url):
111 def addchangegroup(self, cg, source, url):
112 return self._repo.addchangegroup(cg, source, url)
112 return self._repo.addchangegroup(cg, source, url)
113
113
114 def pushkey(self, namespace, key, old, new):
114 def pushkey(self, namespace, key, old, new):
115 return self._repo.pushkey(namespace, key, old, new)
115 return self._repo.pushkey(namespace, key, old, new)
116
116
117 def listkeys(self, namespace):
117 def listkeys(self, namespace):
118 return self._repo.listkeys(namespace)
118 return self._repo.listkeys(namespace)
119
119
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 '''used to test argument passing over the wire'''
121 '''used to test argument passing over the wire'''
122 return "%s %s %s %s %s" % (one, two, three, four, five)
122 return "%s %s %s %s %s" % (one, two, three, four, five)
123
123
124 class locallegacypeer(localpeer):
124 class locallegacypeer(localpeer):
125 '''peer extension which implements legacy methods too; used for tests with
125 '''peer extension which implements legacy methods too; used for tests with
126 restricted capabilities'''
126 restricted capabilities'''
127
127
128 def __init__(self, repo):
128 def __init__(self, repo):
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130
130
131 def branches(self, nodes):
131 def branches(self, nodes):
132 return self._repo.branches(nodes)
132 return self._repo.branches(nodes)
133
133
134 def between(self, pairs):
134 def between(self, pairs):
135 return self._repo.between(pairs)
135 return self._repo.between(pairs)
136
136
137 def changegroup(self, basenodes, source):
137 def changegroup(self, basenodes, source):
138 return self._repo.changegroup(basenodes, source)
138 return self._repo.changegroup(basenodes, source)
139
139
140 def changegroupsubset(self, bases, heads, source):
140 def changegroupsubset(self, bases, heads, source):
141 return self._repo.changegroupsubset(bases, heads, source)
141 return self._repo.changegroupsubset(bases, heads, source)
142
142
143 class localrepository(object):
143 class localrepository(object):
144
144
145 supportedformats = set(('revlogv1', 'generaldelta'))
145 supportedformats = set(('revlogv1', 'generaldelta'))
146 supported = supportedformats | set(('store', 'fncache', 'shared',
146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 'dotencode'))
147 'dotencode'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
149 requirements = ['revlogv1']
149 requirements = ['revlogv1']
150 filtername = None
150 filtername = None
151
151
152 def _baserequirements(self, create):
152 def _baserequirements(self, create):
153 return self.requirements[:]
153 return self.requirements[:]
154
154
155 def __init__(self, baseui, path=None, create=False):
155 def __init__(self, baseui, path=None, create=False):
156 self.wvfs = scmutil.vfs(path, expand=True)
156 self.wvfs = scmutil.vfs(path, expand=True)
157 self.wopener = self.wvfs
157 self.wopener = self.wvfs
158 self.root = self.wvfs.base
158 self.root = self.wvfs.base
159 self.path = self.wvfs.join(".hg")
159 self.path = self.wvfs.join(".hg")
160 self.origroot = path
160 self.origroot = path
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 self.vfs = scmutil.vfs(self.path)
162 self.vfs = scmutil.vfs(self.path)
163 self.opener = self.vfs
163 self.opener = self.vfs
164 self.baseui = baseui
164 self.baseui = baseui
165 self.ui = baseui.copy()
165 self.ui = baseui.copy()
166 # A list of callback to shape the phase if no data were found.
166 # A list of callback to shape the phase if no data were found.
167 # Callback are in the form: func(repo, roots) --> processed root.
167 # Callback are in the form: func(repo, roots) --> processed root.
168 # This list it to be filled by extension during repo setup
168 # This list it to be filled by extension during repo setup
169 self._phasedefaults = []
169 self._phasedefaults = []
170 try:
170 try:
171 self.ui.readconfig(self.join("hgrc"), self.root)
171 self.ui.readconfig(self.join("hgrc"), self.root)
172 extensions.loadall(self.ui)
172 extensions.loadall(self.ui)
173 except IOError:
173 except IOError:
174 pass
174 pass
175
175
176 if not self.vfs.isdir():
176 if not self.vfs.isdir():
177 if create:
177 if create:
178 if not self.wvfs.exists():
178 if not self.wvfs.exists():
179 self.wvfs.makedirs()
179 self.wvfs.makedirs()
180 self.vfs.makedir(notindexed=True)
180 self.vfs.makedir(notindexed=True)
181 requirements = self._baserequirements(create)
181 requirements = self._baserequirements(create)
182 if self.ui.configbool('format', 'usestore', True):
182 if self.ui.configbool('format', 'usestore', True):
183 self.vfs.mkdir("store")
183 self.vfs.mkdir("store")
184 requirements.append("store")
184 requirements.append("store")
185 if self.ui.configbool('format', 'usefncache', True):
185 if self.ui.configbool('format', 'usefncache', True):
186 requirements.append("fncache")
186 requirements.append("fncache")
187 if self.ui.configbool('format', 'dotencode', True):
187 if self.ui.configbool('format', 'dotencode', True):
188 requirements.append('dotencode')
188 requirements.append('dotencode')
189 # create an invalid changelog
189 # create an invalid changelog
190 self.vfs.append(
190 self.vfs.append(
191 "00changelog.i",
191 "00changelog.i",
192 '\0\0\0\2' # represents revlogv2
192 '\0\0\0\2' # represents revlogv2
193 ' dummy changelog to prevent using the old repo layout'
193 ' dummy changelog to prevent using the old repo layout'
194 )
194 )
195 if self.ui.configbool('format', 'generaldelta', False):
195 if self.ui.configbool('format', 'generaldelta', False):
196 requirements.append("generaldelta")
196 requirements.append("generaldelta")
197 requirements = set(requirements)
197 requirements = set(requirements)
198 else:
198 else:
199 raise error.RepoError(_("repository %s not found") % path)
199 raise error.RepoError(_("repository %s not found") % path)
200 elif create:
200 elif create:
201 raise error.RepoError(_("repository %s already exists") % path)
201 raise error.RepoError(_("repository %s already exists") % path)
202 else:
202 else:
203 try:
203 try:
204 requirements = scmutil.readrequires(self.vfs, self.supported)
204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 except IOError, inst:
205 except IOError, inst:
206 if inst.errno != errno.ENOENT:
206 if inst.errno != errno.ENOENT:
207 raise
207 raise
208 requirements = set()
208 requirements = set()
209
209
210 self.sharedpath = self.path
210 self.sharedpath = self.path
211 try:
211 try:
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 if not os.path.exists(s):
213 if not os.path.exists(s):
214 raise error.RepoError(
214 raise error.RepoError(
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 self.sharedpath = s
216 self.sharedpath = s
217 except IOError, inst:
217 except IOError, inst:
218 if inst.errno != errno.ENOENT:
218 if inst.errno != errno.ENOENT:
219 raise
219 raise
220
220
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 self.spath = self.store.path
222 self.spath = self.store.path
223 self.svfs = self.store.vfs
223 self.svfs = self.store.vfs
224 self.sopener = self.svfs
224 self.sopener = self.svfs
225 self.sjoin = self.store.join
225 self.sjoin = self.store.join
226 self.vfs.createmode = self.store.createmode
226 self.vfs.createmode = self.store.createmode
227 self._applyrequirements(requirements)
227 self._applyrequirements(requirements)
228 if create:
228 if create:
229 self._writerequirements()
229 self._writerequirements()
230
230
231
231
232 self._branchcaches = {}
232 self._branchcaches = {}
233 self.filterpats = {}
233 self.filterpats = {}
234 self._datafilters = {}
234 self._datafilters = {}
235 self._transref = self._lockref = self._wlockref = None
235 self._transref = self._lockref = self._wlockref = None
236
236
237 # A cache for various files under .hg/ that tracks file changes,
237 # A cache for various files under .hg/ that tracks file changes,
238 # (used by the filecache decorator)
238 # (used by the filecache decorator)
239 #
239 #
240 # Maps a property name to its util.filecacheentry
240 # Maps a property name to its util.filecacheentry
241 self._filecache = {}
241 self._filecache = {}
242
242
243 # hold sets of revision to be filtered
243 # hold sets of revision to be filtered
244 # should be cleared when something might have changed the filter value:
244 # should be cleared when something might have changed the filter value:
245 # - new changesets,
245 # - new changesets,
246 # - phase change,
246 # - phase change,
247 # - new obsolescence marker,
247 # - new obsolescence marker,
248 # - working directory parent change,
248 # - working directory parent change,
249 # - bookmark changes
249 # - bookmark changes
250 self.filteredrevcache = {}
250 self.filteredrevcache = {}
251
251
252 def close(self):
252 def close(self):
253 pass
253 pass
254
254
255 def _restrictcapabilities(self, caps):
255 def _restrictcapabilities(self, caps):
256 return caps
256 return caps
257
257
258 def _applyrequirements(self, requirements):
258 def _applyrequirements(self, requirements):
259 self.requirements = requirements
259 self.requirements = requirements
260 self.sopener.options = dict((r, 1) for r in requirements
260 self.sopener.options = dict((r, 1) for r in requirements
261 if r in self.openerreqs)
261 if r in self.openerreqs)
262
262
263 def _writerequirements(self):
263 def _writerequirements(self):
264 reqfile = self.opener("requires", "w")
264 reqfile = self.opener("requires", "w")
265 for r in self.requirements:
265 for r in self.requirements:
266 reqfile.write("%s\n" % r)
266 reqfile.write("%s\n" % r)
267 reqfile.close()
267 reqfile.close()
268
268
269 def _checknested(self, path):
269 def _checknested(self, path):
270 """Determine if path is a legal nested repository."""
270 """Determine if path is a legal nested repository."""
271 if not path.startswith(self.root):
271 if not path.startswith(self.root):
272 return False
272 return False
273 subpath = path[len(self.root) + 1:]
273 subpath = path[len(self.root) + 1:]
274 normsubpath = util.pconvert(subpath)
274 normsubpath = util.pconvert(subpath)
275
275
276 # XXX: Checking against the current working copy is wrong in
276 # XXX: Checking against the current working copy is wrong in
277 # the sense that it can reject things like
277 # the sense that it can reject things like
278 #
278 #
279 # $ hg cat -r 10 sub/x.txt
279 # $ hg cat -r 10 sub/x.txt
280 #
280 #
281 # if sub/ is no longer a subrepository in the working copy
281 # if sub/ is no longer a subrepository in the working copy
282 # parent revision.
282 # parent revision.
283 #
283 #
284 # However, it can of course also allow things that would have
284 # However, it can of course also allow things that would have
285 # been rejected before, such as the above cat command if sub/
285 # been rejected before, such as the above cat command if sub/
286 # is a subrepository now, but was a normal directory before.
286 # is a subrepository now, but was a normal directory before.
287 # The old path auditor would have rejected by mistake since it
287 # The old path auditor would have rejected by mistake since it
288 # panics when it sees sub/.hg/.
288 # panics when it sees sub/.hg/.
289 #
289 #
290 # All in all, checking against the working copy seems sensible
290 # All in all, checking against the working copy seems sensible
291 # since we want to prevent access to nested repositories on
291 # since we want to prevent access to nested repositories on
292 # the filesystem *now*.
292 # the filesystem *now*.
293 ctx = self[None]
293 ctx = self[None]
294 parts = util.splitpath(subpath)
294 parts = util.splitpath(subpath)
295 while parts:
295 while parts:
296 prefix = '/'.join(parts)
296 prefix = '/'.join(parts)
297 if prefix in ctx.substate:
297 if prefix in ctx.substate:
298 if prefix == normsubpath:
298 if prefix == normsubpath:
299 return True
299 return True
300 else:
300 else:
301 sub = ctx.sub(prefix)
301 sub = ctx.sub(prefix)
302 return sub.checknested(subpath[len(prefix) + 1:])
302 return sub.checknested(subpath[len(prefix) + 1:])
303 else:
303 else:
304 parts.pop()
304 parts.pop()
305 return False
305 return False
306
306
307 def peer(self):
307 def peer(self):
308 return localpeer(self) # not cached to avoid reference cycle
308 return localpeer(self) # not cached to avoid reference cycle
309
309
310 def unfiltered(self):
310 def unfiltered(self):
311 """Return unfiltered version of the repository
311 """Return unfiltered version of the repository
312
312
313 Intended to be ovewritten by filtered repo."""
313 Intended to be ovewritten by filtered repo."""
314 return self
314 return self
315
315
316 def filtered(self, name):
316 def filtered(self, name):
317 """Return a filtered version of a repository"""
317 """Return a filtered version of a repository"""
318 # build a new class with the mixin and the current class
318 # build a new class with the mixin and the current class
319 # (possibily subclass of the repo)
319 # (possibily subclass of the repo)
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 pass
321 pass
322 return proxycls(self, name)
322 return proxycls(self, name)
323
323
324 @repofilecache('bookmarks')
324 @repofilecache('bookmarks')
325 def _bookmarks(self):
325 def _bookmarks(self):
326 return bookmarks.bmstore(self)
326 return bookmarks.bmstore(self)
327
327
328 @repofilecache('bookmarks.current')
328 @repofilecache('bookmarks.current')
329 def _bookmarkcurrent(self):
329 def _bookmarkcurrent(self):
330 return bookmarks.readcurrent(self)
330 return bookmarks.readcurrent(self)
331
331
332 def bookmarkheads(self, bookmark):
332 def bookmarkheads(self, bookmark):
333 name = bookmark.split('@', 1)[0]
333 name = bookmark.split('@', 1)[0]
334 heads = []
334 heads = []
335 for mark, n in self._bookmarks.iteritems():
335 for mark, n in self._bookmarks.iteritems():
336 if mark.split('@', 1)[0] == name:
336 if mark.split('@', 1)[0] == name:
337 heads.append(n)
337 heads.append(n)
338 return heads
338 return heads
339
339
340 @storecache('phaseroots')
340 @storecache('phaseroots')
341 def _phasecache(self):
341 def _phasecache(self):
342 return phases.phasecache(self, self._phasedefaults)
342 return phases.phasecache(self, self._phasedefaults)
343
343
344 @storecache('obsstore')
344 @storecache('obsstore')
345 def obsstore(self):
345 def obsstore(self):
346 store = obsolete.obsstore(self.sopener)
346 store = obsolete.obsstore(self.sopener)
347 if store and not obsolete._enabled:
347 if store and not obsolete._enabled:
348 # message is rare enough to not be translated
348 # message is rare enough to not be translated
349 msg = 'obsolete feature not enabled but %i markers found!\n'
349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 self.ui.warn(msg % len(list(store)))
350 self.ui.warn(msg % len(list(store)))
351 return store
351 return store
352
352
353 @storecache('00changelog.i')
353 @storecache('00changelog.i')
354 def changelog(self):
354 def changelog(self):
355 c = changelog.changelog(self.sopener)
355 c = changelog.changelog(self.sopener)
356 if 'HG_PENDING' in os.environ:
356 if 'HG_PENDING' in os.environ:
357 p = os.environ['HG_PENDING']
357 p = os.environ['HG_PENDING']
358 if p.startswith(self.root):
358 if p.startswith(self.root):
359 c.readpending('00changelog.i.a')
359 c.readpending('00changelog.i.a')
360 return c
360 return c
361
361
362 @storecache('00manifest.i')
362 @storecache('00manifest.i')
363 def manifest(self):
363 def manifest(self):
364 return manifest.manifest(self.sopener)
364 return manifest.manifest(self.sopener)
365
365
366 @repofilecache('dirstate')
366 @repofilecache('dirstate')
367 def dirstate(self):
367 def dirstate(self):
368 warned = [0]
368 warned = [0]
369 def validate(node):
369 def validate(node):
370 try:
370 try:
371 self.changelog.rev(node)
371 self.changelog.rev(node)
372 return node
372 return node
373 except error.LookupError:
373 except error.LookupError:
374 if not warned[0]:
374 if not warned[0]:
375 warned[0] = True
375 warned[0] = True
376 self.ui.warn(_("warning: ignoring unknown"
376 self.ui.warn(_("warning: ignoring unknown"
377 " working parent %s!\n") % short(node))
377 " working parent %s!\n") % short(node))
378 return nullid
378 return nullid
379
379
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381
381
382 def __getitem__(self, changeid):
382 def __getitem__(self, changeid):
383 if changeid is None:
383 if changeid is None:
384 return context.workingctx(self)
384 return context.workingctx(self)
385 return context.changectx(self, changeid)
385 return context.changectx(self, changeid)
386
386
387 def __contains__(self, changeid):
387 def __contains__(self, changeid):
388 try:
388 try:
389 return bool(self.lookup(changeid))
389 return bool(self.lookup(changeid))
390 except error.RepoLookupError:
390 except error.RepoLookupError:
391 return False
391 return False
392
392
393 def __nonzero__(self):
393 def __nonzero__(self):
394 return True
394 return True
395
395
396 def __len__(self):
396 def __len__(self):
397 return len(self.changelog)
397 return len(self.changelog)
398
398
399 def __iter__(self):
399 def __iter__(self):
400 return iter(self.changelog)
400 return iter(self.changelog)
401
401
402 def revs(self, expr, *args):
402 def revs(self, expr, *args):
403 '''Return a list of revisions matching the given revset'''
403 '''Return a list of revisions matching the given revset'''
404 expr = revset.formatspec(expr, *args)
404 expr = revset.formatspec(expr, *args)
405 m = revset.match(None, expr)
405 m = revset.match(None, expr)
406 return [r for r in m(self, list(self))]
406 return [r for r in m(self, list(self))]
407
407
408 def set(self, expr, *args):
408 def set(self, expr, *args):
409 '''
409 '''
410 Yield a context for each matching revision, after doing arg
410 Yield a context for each matching revision, after doing arg
411 replacement via revset.formatspec
411 replacement via revset.formatspec
412 '''
412 '''
413 for r in self.revs(expr, *args):
413 for r in self.revs(expr, *args):
414 yield self[r]
414 yield self[r]
415
415
416 def url(self):
416 def url(self):
417 return 'file:' + self.root
417 return 'file:' + self.root
418
418
419 def hook(self, name, throw=False, **args):
419 def hook(self, name, throw=False, **args):
420 return hook.hook(self.ui, self, name, throw, **args)
420 return hook.hook(self.ui, self, name, throw, **args)
421
421
422 @unfilteredmethod
422 @unfilteredmethod
423 def _tag(self, names, node, message, local, user, date, extra={}):
423 def _tag(self, names, node, message, local, user, date, extra={}):
424 if isinstance(names, str):
424 if isinstance(names, str):
425 names = (names,)
425 names = (names,)
426
426
427 branches = self.branchmap()
427 branches = self.branchmap()
428 for name in names:
428 for name in names:
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 local=local)
430 local=local)
431 if name in branches:
431 if name in branches:
432 self.ui.warn(_("warning: tag %s conflicts with existing"
432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 " branch name\n") % name)
433 " branch name\n") % name)
434
434
435 def writetags(fp, names, munge, prevtags):
435 def writetags(fp, names, munge, prevtags):
436 fp.seek(0, 2)
436 fp.seek(0, 2)
437 if prevtags and prevtags[-1] != '\n':
437 if prevtags and prevtags[-1] != '\n':
438 fp.write('\n')
438 fp.write('\n')
439 for name in names:
439 for name in names:
440 m = munge and munge(name) or name
440 m = munge and munge(name) or name
441 if (self._tagscache.tagtypes and
441 if (self._tagscache.tagtypes and
442 name in self._tagscache.tagtypes):
442 name in self._tagscache.tagtypes):
443 old = self.tags().get(name, nullid)
443 old = self.tags().get(name, nullid)
444 fp.write('%s %s\n' % (hex(old), m))
444 fp.write('%s %s\n' % (hex(old), m))
445 fp.write('%s %s\n' % (hex(node), m))
445 fp.write('%s %s\n' % (hex(node), m))
446 fp.close()
446 fp.close()
447
447
448 prevtags = ''
448 prevtags = ''
449 if local:
449 if local:
450 try:
450 try:
451 fp = self.opener('localtags', 'r+')
451 fp = self.opener('localtags', 'r+')
452 except IOError:
452 except IOError:
453 fp = self.opener('localtags', 'a')
453 fp = self.opener('localtags', 'a')
454 else:
454 else:
455 prevtags = fp.read()
455 prevtags = fp.read()
456
456
457 # local tags are stored in the current charset
457 # local tags are stored in the current charset
458 writetags(fp, names, None, prevtags)
458 writetags(fp, names, None, prevtags)
459 for name in names:
459 for name in names:
460 self.hook('tag', node=hex(node), tag=name, local=local)
460 self.hook('tag', node=hex(node), tag=name, local=local)
461 return
461 return
462
462
463 try:
463 try:
464 fp = self.wfile('.hgtags', 'rb+')
464 fp = self.wfile('.hgtags', 'rb+')
465 except IOError, e:
465 except IOError, e:
466 if e.errno != errno.ENOENT:
466 if e.errno != errno.ENOENT:
467 raise
467 raise
468 fp = self.wfile('.hgtags', 'ab')
468 fp = self.wfile('.hgtags', 'ab')
469 else:
469 else:
470 prevtags = fp.read()
470 prevtags = fp.read()
471
471
472 # committed tags are stored in UTF-8
472 # committed tags are stored in UTF-8
473 writetags(fp, names, encoding.fromlocal, prevtags)
473 writetags(fp, names, encoding.fromlocal, prevtags)
474
474
475 fp.close()
475 fp.close()
476
476
477 self.invalidatecaches()
477 self.invalidatecaches()
478
478
479 if '.hgtags' not in self.dirstate:
479 if '.hgtags' not in self.dirstate:
480 self[None].add(['.hgtags'])
480 self[None].add(['.hgtags'])
481
481
482 m = matchmod.exact(self.root, '', ['.hgtags'])
482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484
484
485 for name in names:
485 for name in names:
486 self.hook('tag', node=hex(node), tag=name, local=local)
486 self.hook('tag', node=hex(node), tag=name, local=local)
487
487
488 return tagnode
488 return tagnode
489
489
490 def tag(self, names, node, message, local, user, date):
490 def tag(self, names, node, message, local, user, date):
491 '''tag a revision with one or more symbolic names.
491 '''tag a revision with one or more symbolic names.
492
492
493 names is a list of strings or, when adding a single tag, names may be a
493 names is a list of strings or, when adding a single tag, names may be a
494 string.
494 string.
495
495
496 if local is True, the tags are stored in a per-repository file.
496 if local is True, the tags are stored in a per-repository file.
497 otherwise, they are stored in the .hgtags file, and a new
497 otherwise, they are stored in the .hgtags file, and a new
498 changeset is committed with the change.
498 changeset is committed with the change.
499
499
500 keyword arguments:
500 keyword arguments:
501
501
502 local: whether to store tags in non-version-controlled file
502 local: whether to store tags in non-version-controlled file
503 (default False)
503 (default False)
504
504
505 message: commit message to use if committing
505 message: commit message to use if committing
506
506
507 user: name of user to use if committing
507 user: name of user to use if committing
508
508
509 date: date tuple to use if committing'''
509 date: date tuple to use if committing'''
510
510
511 if not local:
511 if not local:
512 for x in self.status()[:5]:
512 for x in self.status()[:5]:
513 if '.hgtags' in x:
513 if '.hgtags' in x:
514 raise util.Abort(_('working copy of .hgtags is changed '
514 raise util.Abort(_('working copy of .hgtags is changed '
515 '(please commit .hgtags manually)'))
515 '(please commit .hgtags manually)'))
516
516
517 self.tags() # instantiate the cache
517 self.tags() # instantiate the cache
518 self._tag(names, node, message, local, user, date)
518 self._tag(names, node, message, local, user, date)
519
519
520 @filteredpropertycache
520 @filteredpropertycache
521 def _tagscache(self):
521 def _tagscache(self):
522 '''Returns a tagscache object that contains various tags related
522 '''Returns a tagscache object that contains various tags related
523 caches.'''
523 caches.'''
524
524
525 # This simplifies its cache management by having one decorated
525 # This simplifies its cache management by having one decorated
526 # function (this one) and the rest simply fetch things from it.
526 # function (this one) and the rest simply fetch things from it.
527 class tagscache(object):
527 class tagscache(object):
528 def __init__(self):
528 def __init__(self):
529 # These two define the set of tags for this repository. tags
529 # These two define the set of tags for this repository. tags
530 # maps tag name to node; tagtypes maps tag name to 'global' or
530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 # 'local'. (Global tags are defined by .hgtags across all
531 # 'local'. (Global tags are defined by .hgtags across all
532 # heads, and local tags are defined in .hg/localtags.)
532 # heads, and local tags are defined in .hg/localtags.)
533 # They constitute the in-memory cache of tags.
533 # They constitute the in-memory cache of tags.
534 self.tags = self.tagtypes = None
534 self.tags = self.tagtypes = None
535
535
536 self.nodetagscache = self.tagslist = None
536 self.nodetagscache = self.tagslist = None
537
537
538 cache = tagscache()
538 cache = tagscache()
539 cache.tags, cache.tagtypes = self._findtags()
539 cache.tags, cache.tagtypes = self._findtags()
540
540
541 return cache
541 return cache
542
542
543 def tags(self):
543 def tags(self):
544 '''return a mapping of tag to node'''
544 '''return a mapping of tag to node'''
545 t = {}
545 t = {}
546 if self.changelog.filteredrevs:
546 if self.changelog.filteredrevs:
547 tags, tt = self._findtags()
547 tags, tt = self._findtags()
548 else:
548 else:
549 tags = self._tagscache.tags
549 tags = self._tagscache.tags
550 for k, v in tags.iteritems():
550 for k, v in tags.iteritems():
551 try:
551 try:
552 # ignore tags to unknown nodes
552 # ignore tags to unknown nodes
553 self.changelog.rev(v)
553 self.changelog.rev(v)
554 t[k] = v
554 t[k] = v
555 except (error.LookupError, ValueError):
555 except (error.LookupError, ValueError):
556 pass
556 pass
557 return t
557 return t
558
558
559 def _findtags(self):
559 def _findtags(self):
560 '''Do the hard work of finding tags. Return a pair of dicts
560 '''Do the hard work of finding tags. Return a pair of dicts
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 maps tag name to a string like \'global\' or \'local\'.
562 maps tag name to a string like \'global\' or \'local\'.
563 Subclasses or extensions are free to add their own tags, but
563 Subclasses or extensions are free to add their own tags, but
564 should be aware that the returned dicts will be retained for the
564 should be aware that the returned dicts will be retained for the
565 duration of the localrepo object.'''
565 duration of the localrepo object.'''
566
566
567 # XXX what tagtype should subclasses/extensions use? Currently
567 # XXX what tagtype should subclasses/extensions use? Currently
568 # mq and bookmarks add tags, but do not set the tagtype at all.
568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 # Should each extension invent its own tag type? Should there
569 # Should each extension invent its own tag type? Should there
570 # be one tagtype for all such "virtual" tags? Or is the status
570 # be one tagtype for all such "virtual" tags? Or is the status
571 # quo fine?
571 # quo fine?
572
572
573 alltags = {} # map tag name to (node, hist)
573 alltags = {} # map tag name to (node, hist)
574 tagtypes = {}
574 tagtypes = {}
575
575
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578
578
579 # Build the return dicts. Have to re-encode tag names because
579 # Build the return dicts. Have to re-encode tag names because
580 # the tags module always uses UTF-8 (in order not to lose info
580 # the tags module always uses UTF-8 (in order not to lose info
581 # writing to the cache), but the rest of Mercurial wants them in
581 # writing to the cache), but the rest of Mercurial wants them in
582 # local encoding.
582 # local encoding.
583 tags = {}
583 tags = {}
584 for (name, (node, hist)) in alltags.iteritems():
584 for (name, (node, hist)) in alltags.iteritems():
585 if node != nullid:
585 if node != nullid:
586 tags[encoding.tolocal(name)] = node
586 tags[encoding.tolocal(name)] = node
587 tags['tip'] = self.changelog.tip()
587 tags['tip'] = self.changelog.tip()
588 tagtypes = dict([(encoding.tolocal(name), value)
588 tagtypes = dict([(encoding.tolocal(name), value)
589 for (name, value) in tagtypes.iteritems()])
589 for (name, value) in tagtypes.iteritems()])
590 return (tags, tagtypes)
590 return (tags, tagtypes)
591
591
592 def tagtype(self, tagname):
592 def tagtype(self, tagname):
593 '''
593 '''
594 return the type of the given tag. result can be:
594 return the type of the given tag. result can be:
595
595
596 'local' : a local tag
596 'local' : a local tag
597 'global' : a global tag
597 'global' : a global tag
598 None : tag does not exist
598 None : tag does not exist
599 '''
599 '''
600
600
601 return self._tagscache.tagtypes.get(tagname)
601 return self._tagscache.tagtypes.get(tagname)
602
602
603 def tagslist(self):
603 def tagslist(self):
604 '''return a list of tags ordered by revision'''
604 '''return a list of tags ordered by revision'''
605 if not self._tagscache.tagslist:
605 if not self._tagscache.tagslist:
606 l = []
606 l = []
607 for t, n in self.tags().iteritems():
607 for t, n in self.tags().iteritems():
608 r = self.changelog.rev(n)
608 r = self.changelog.rev(n)
609 l.append((r, t, n))
609 l.append((r, t, n))
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611
611
612 return self._tagscache.tagslist
612 return self._tagscache.tagslist
613
613
614 def nodetags(self, node):
614 def nodetags(self, node):
615 '''return the tags associated with a node'''
615 '''return the tags associated with a node'''
616 if not self._tagscache.nodetagscache:
616 if not self._tagscache.nodetagscache:
617 nodetagscache = {}
617 nodetagscache = {}
618 for t, n in self._tagscache.tags.iteritems():
618 for t, n in self._tagscache.tags.iteritems():
619 nodetagscache.setdefault(n, []).append(t)
619 nodetagscache.setdefault(n, []).append(t)
620 for tags in nodetagscache.itervalues():
620 for tags in nodetagscache.itervalues():
621 tags.sort()
621 tags.sort()
622 self._tagscache.nodetagscache = nodetagscache
622 self._tagscache.nodetagscache = nodetagscache
623 return self._tagscache.nodetagscache.get(node, [])
623 return self._tagscache.nodetagscache.get(node, [])
624
624
625 def nodebookmarks(self, node):
625 def nodebookmarks(self, node):
626 marks = []
626 marks = []
627 for bookmark, n in self._bookmarks.iteritems():
627 for bookmark, n in self._bookmarks.iteritems():
628 if n == node:
628 if n == node:
629 marks.append(bookmark)
629 marks.append(bookmark)
630 return sorted(marks)
630 return sorted(marks)
631
631
632 def branchmap(self):
632 def branchmap(self):
633 '''returns a dictionary {branch: [branchheads]}'''
633 '''returns a dictionary {branch: [branchheads]}'''
634 branchmap.updatecache(self)
634 branchmap.updatecache(self)
635 return self._branchcaches[self.filtername]
635 return self._branchcaches[self.filtername]
636
636
637
637
638 def _branchtip(self, heads):
638 def _branchtip(self, heads):
639 '''return the tipmost branch head in heads'''
639 '''return the tipmost branch head in heads'''
640 tip = heads[-1]
640 tip = heads[-1]
641 for h in reversed(heads):
641 for h in reversed(heads):
642 if not self[h].closesbranch():
642 if not self[h].closesbranch():
643 tip = h
643 tip = h
644 break
644 break
645 return tip
645 return tip
646
646
647 def branchtip(self, branch):
647 def branchtip(self, branch):
648 '''return the tip node for a given branch'''
648 '''return the tip node for a given branch'''
649 if branch not in self.branchmap():
649 if branch not in self.branchmap():
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 return self._branchtip(self.branchmap()[branch])
651 return self._branchtip(self.branchmap()[branch])
652
652
653 def branchtags(self):
653 def branchtags(self):
654 '''return a dict where branch names map to the tipmost head of
654 '''return a dict where branch names map to the tipmost head of
655 the branch, open heads come before closed'''
655 the branch, open heads come before closed'''
656 bt = {}
656 bt = {}
657 for bn, heads in self.branchmap().iteritems():
657 for bn, heads in self.branchmap().iteritems():
658 bt[bn] = self._branchtip(heads)
658 bt[bn] = self._branchtip(heads)
659 return bt
659 return bt
660
660
661 def lookup(self, key):
661 def lookup(self, key):
662 return self[key].node()
662 return self[key].node()
663
663
664 def lookupbranch(self, key, remote=None):
664 def lookupbranch(self, key, remote=None):
665 repo = remote or self
665 repo = remote or self
666 if key in repo.branchmap():
666 if key in repo.branchmap():
667 return key
667 return key
668
668
669 repo = (remote and remote.local()) and remote or self
669 repo = (remote and remote.local()) and remote or self
670 return repo[key].branch()
670 return repo[key].branch()
671
671
672 def known(self, nodes):
672 def known(self, nodes):
673 nm = self.changelog.nodemap
673 nm = self.changelog.nodemap
674 pc = self._phasecache
674 pc = self._phasecache
675 result = []
675 result = []
676 for n in nodes:
676 for n in nodes:
677 r = nm.get(n)
677 r = nm.get(n)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 result.append(resp)
679 result.append(resp)
680 return result
680 return result
681
681
682 def local(self):
682 def local(self):
683 return self
683 return self
684
684
685 def cancopy(self):
685 def cancopy(self):
686 return self.local() # so statichttprepo's override of local() works
686 return self.local() # so statichttprepo's override of local() works
687
687
688 def join(self, f):
688 def join(self, f):
689 return os.path.join(self.path, f)
689 return os.path.join(self.path, f)
690
690
691 def wjoin(self, f):
691 def wjoin(self, f):
692 return os.path.join(self.root, f)
692 return os.path.join(self.root, f)
693
693
694 def file(self, f):
694 def file(self, f):
695 if f[0] == '/':
695 if f[0] == '/':
696 f = f[1:]
696 f = f[1:]
697 return filelog.filelog(self.sopener, f)
697 return filelog.filelog(self.sopener, f)
698
698
699 def changectx(self, changeid):
699 def changectx(self, changeid):
700 return self[changeid]
700 return self[changeid]
701
701
702 def parents(self, changeid=None):
702 def parents(self, changeid=None):
703 '''get list of changectxs for parents of changeid'''
703 '''get list of changectxs for parents of changeid'''
704 return self[changeid].parents()
704 return self[changeid].parents()
705
705
706 def setparents(self, p1, p2=nullid):
706 def setparents(self, p1, p2=nullid):
707 copies = self.dirstate.setparents(p1, p2)
707 copies = self.dirstate.setparents(p1, p2)
708 if copies:
708 if copies:
709 # Adjust copy records, the dirstate cannot do it, it
709 # Adjust copy records, the dirstate cannot do it, it
710 # requires access to parents manifests. Preserve them
710 # requires access to parents manifests. Preserve them
711 # only for entries added to first parent.
711 # only for entries added to first parent.
712 pctx = self[p1]
712 pctx = self[p1]
713 for f in copies:
713 for f in copies:
714 if f not in pctx and copies[f] in pctx:
714 if f not in pctx and copies[f] in pctx:
715 self.dirstate.copy(copies[f], f)
715 self.dirstate.copy(copies[f], f)
716
716
717 def filectx(self, path, changeid=None, fileid=None):
717 def filectx(self, path, changeid=None, fileid=None):
718 """changeid can be a changeset revision, node, or tag.
718 """changeid can be a changeset revision, node, or tag.
719 fileid can be a file revision or node."""
719 fileid can be a file revision or node."""
720 return context.filectx(self, path, changeid, fileid)
720 return context.filectx(self, path, changeid, fileid)
721
721
722 def getcwd(self):
722 def getcwd(self):
723 return self.dirstate.getcwd()
723 return self.dirstate.getcwd()
724
724
725 def pathto(self, f, cwd=None):
725 def pathto(self, f, cwd=None):
726 return self.dirstate.pathto(f, cwd)
726 return self.dirstate.pathto(f, cwd)
727
727
728 def wfile(self, f, mode='r'):
728 def wfile(self, f, mode='r'):
729 return self.wopener(f, mode)
729 return self.wopener(f, mode)
730
730
731 def _link(self, f):
731 def _link(self, f):
732 return os.path.islink(self.wjoin(f))
732 return os.path.islink(self.wjoin(f))
733
733
734 def _loadfilter(self, filter):
734 def _loadfilter(self, filter):
735 if filter not in self.filterpats:
735 if filter not in self.filterpats:
736 l = []
736 l = []
737 for pat, cmd in self.ui.configitems(filter):
737 for pat, cmd in self.ui.configitems(filter):
738 if cmd == '!':
738 if cmd == '!':
739 continue
739 continue
740 mf = matchmod.match(self.root, '', [pat])
740 mf = matchmod.match(self.root, '', [pat])
741 fn = None
741 fn = None
742 params = cmd
742 params = cmd
743 for name, filterfn in self._datafilters.iteritems():
743 for name, filterfn in self._datafilters.iteritems():
744 if cmd.startswith(name):
744 if cmd.startswith(name):
745 fn = filterfn
745 fn = filterfn
746 params = cmd[len(name):].lstrip()
746 params = cmd[len(name):].lstrip()
747 break
747 break
748 if not fn:
748 if not fn:
749 fn = lambda s, c, **kwargs: util.filter(s, c)
749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 # Wrap old filters not supporting keyword arguments
750 # Wrap old filters not supporting keyword arguments
751 if not inspect.getargspec(fn)[2]:
751 if not inspect.getargspec(fn)[2]:
752 oldfn = fn
752 oldfn = fn
753 fn = lambda s, c, **kwargs: oldfn(s, c)
753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 l.append((mf, fn, params))
754 l.append((mf, fn, params))
755 self.filterpats[filter] = l
755 self.filterpats[filter] = l
756 return self.filterpats[filter]
756 return self.filterpats[filter]
757
757
758 def _filter(self, filterpats, filename, data):
758 def _filter(self, filterpats, filename, data):
759 for mf, fn, cmd in filterpats:
759 for mf, fn, cmd in filterpats:
760 if mf(filename):
760 if mf(filename):
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 break
763 break
764
764
765 return data
765 return data
766
766
767 @unfilteredpropertycache
767 @unfilteredpropertycache
768 def _encodefilterpats(self):
768 def _encodefilterpats(self):
769 return self._loadfilter('encode')
769 return self._loadfilter('encode')
770
770
771 @unfilteredpropertycache
771 @unfilteredpropertycache
772 def _decodefilterpats(self):
772 def _decodefilterpats(self):
773 return self._loadfilter('decode')
773 return self._loadfilter('decode')
774
774
775 def adddatafilter(self, name, filter):
775 def adddatafilter(self, name, filter):
776 self._datafilters[name] = filter
776 self._datafilters[name] = filter
777
777
778 def wread(self, filename):
778 def wread(self, filename):
779 if self._link(filename):
779 if self._link(filename):
780 data = os.readlink(self.wjoin(filename))
780 data = os.readlink(self.wjoin(filename))
781 else:
781 else:
782 data = self.wopener.read(filename)
782 data = self.wopener.read(filename)
783 return self._filter(self._encodefilterpats, filename, data)
783 return self._filter(self._encodefilterpats, filename, data)
784
784
785 def wwrite(self, filename, data, flags):
785 def wwrite(self, filename, data, flags):
786 data = self._filter(self._decodefilterpats, filename, data)
786 data = self._filter(self._decodefilterpats, filename, data)
787 if 'l' in flags:
787 if 'l' in flags:
788 self.wopener.symlink(data, filename)
788 self.wopener.symlink(data, filename)
789 else:
789 else:
790 self.wopener.write(filename, data)
790 self.wopener.write(filename, data)
791 if 'x' in flags:
791 if 'x' in flags:
792 util.setflags(self.wjoin(filename), False, True)
792 util.setflags(self.wjoin(filename), False, True)
793
793
794 def wwritedata(self, filename, data):
794 def wwritedata(self, filename, data):
795 return self._filter(self._decodefilterpats, filename, data)
795 return self._filter(self._decodefilterpats, filename, data)
796
796
797 def transaction(self, desc):
797 def transaction(self, desc):
798 tr = self._transref and self._transref() or None
798 tr = self._transref and self._transref() or None
799 if tr and tr.running():
799 if tr and tr.running():
800 return tr.nest()
800 return tr.nest()
801
801
802 # abort here if the journal already exists
802 # abort here if the journal already exists
803 if os.path.exists(self.sjoin("journal")):
803 if os.path.exists(self.sjoin("journal")):
804 raise error.RepoError(
804 raise error.RepoError(
805 _("abandoned transaction found - run hg recover"))
805 _("abandoned transaction found - run hg recover"))
806
806
807 self._writejournal(desc)
807 self._writejournal(desc)
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809
809
810 tr = transaction.transaction(self.ui.warn, self.sopener,
810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 self.sjoin("journal"),
811 self.sjoin("journal"),
812 aftertrans(renames),
812 aftertrans(renames),
813 self.store.createmode)
813 self.store.createmode)
814 self._transref = weakref.ref(tr)
814 self._transref = weakref.ref(tr)
815 return tr
815 return tr
816
816
817 def _journalfiles(self):
817 def _journalfiles(self):
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 self.join('journal.branch'), self.join('journal.desc'),
819 self.join('journal.branch'), self.join('journal.desc'),
820 self.join('journal.bookmarks'),
820 self.join('journal.bookmarks'),
821 self.sjoin('journal.phaseroots'))
821 self.sjoin('journal.phaseroots'))
822
822
823 def undofiles(self):
823 def undofiles(self):
824 return [undoname(x) for x in self._journalfiles()]
824 return [undoname(x) for x in self._journalfiles()]
825
825
826 def _writejournal(self, desc):
826 def _writejournal(self, desc):
827 self.opener.write("journal.dirstate",
827 self.opener.write("journal.dirstate",
828 self.opener.tryread("dirstate"))
828 self.opener.tryread("dirstate"))
829 self.opener.write("journal.branch",
829 self.opener.write("journal.branch",
830 encoding.fromlocal(self.dirstate.branch()))
830 encoding.fromlocal(self.dirstate.branch()))
831 self.opener.write("journal.desc",
831 self.opener.write("journal.desc",
832 "%d\n%s\n" % (len(self), desc))
832 "%d\n%s\n" % (len(self), desc))
833 self.opener.write("journal.bookmarks",
833 self.opener.write("journal.bookmarks",
834 self.opener.tryread("bookmarks"))
834 self.opener.tryread("bookmarks"))
835 self.sopener.write("journal.phaseroots",
835 self.sopener.write("journal.phaseroots",
836 self.sopener.tryread("phaseroots"))
836 self.sopener.tryread("phaseroots"))
837
837
838 def recover(self):
838 def recover(self):
839 lock = self.lock()
839 lock = self.lock()
840 try:
840 try:
841 if os.path.exists(self.sjoin("journal")):
841 if os.path.exists(self.sjoin("journal")):
842 self.ui.status(_("rolling back interrupted transaction\n"))
842 self.ui.status(_("rolling back interrupted transaction\n"))
843 transaction.rollback(self.sopener, self.sjoin("journal"),
843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 self.ui.warn)
844 self.ui.warn)
845 self.invalidate()
845 self.invalidate()
846 return True
846 return True
847 else:
847 else:
848 self.ui.warn(_("no interrupted transaction available\n"))
848 self.ui.warn(_("no interrupted transaction available\n"))
849 return False
849 return False
850 finally:
850 finally:
851 lock.release()
851 lock.release()
852
852
853 def rollback(self, dryrun=False, force=False):
853 def rollback(self, dryrun=False, force=False):
854 wlock = lock = None
854 wlock = lock = None
855 try:
855 try:
856 wlock = self.wlock()
856 wlock = self.wlock()
857 lock = self.lock()
857 lock = self.lock()
858 if os.path.exists(self.sjoin("undo")):
858 if os.path.exists(self.sjoin("undo")):
859 return self._rollback(dryrun, force)
859 return self._rollback(dryrun, force)
860 else:
860 else:
861 self.ui.warn(_("no rollback information available\n"))
861 self.ui.warn(_("no rollback information available\n"))
862 return 1
862 return 1
863 finally:
863 finally:
864 release(lock, wlock)
864 release(lock, wlock)
865
865
866 @unfilteredmethod # Until we get smarter cache management
866 @unfilteredmethod # Until we get smarter cache management
867 def _rollback(self, dryrun, force):
867 def _rollback(self, dryrun, force):
868 ui = self.ui
868 ui = self.ui
869 try:
869 try:
870 args = self.opener.read('undo.desc').splitlines()
870 args = self.opener.read('undo.desc').splitlines()
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 if len(args) >= 3:
872 if len(args) >= 3:
873 detail = args[2]
873 detail = args[2]
874 oldtip = oldlen - 1
874 oldtip = oldlen - 1
875
875
876 if detail and ui.verbose:
876 if detail and ui.verbose:
877 msg = (_('repository tip rolled back to revision %s'
877 msg = (_('repository tip rolled back to revision %s'
878 ' (undo %s: %s)\n')
878 ' (undo %s: %s)\n')
879 % (oldtip, desc, detail))
879 % (oldtip, desc, detail))
880 else:
880 else:
881 msg = (_('repository tip rolled back to revision %s'
881 msg = (_('repository tip rolled back to revision %s'
882 ' (undo %s)\n')
882 ' (undo %s)\n')
883 % (oldtip, desc))
883 % (oldtip, desc))
884 except IOError:
884 except IOError:
885 msg = _('rolling back unknown transaction\n')
885 msg = _('rolling back unknown transaction\n')
886 desc = None
886 desc = None
887
887
888 if not force and self['.'] != self['tip'] and desc == 'commit':
888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 raise util.Abort(
889 raise util.Abort(
890 _('rollback of last commit while not checked out '
890 _('rollback of last commit while not checked out '
891 'may lose data'), hint=_('use -f to force'))
891 'may lose data'), hint=_('use -f to force'))
892
892
893 ui.status(msg)
893 ui.status(msg)
894 if dryrun:
894 if dryrun:
895 return 0
895 return 0
896
896
897 parents = self.dirstate.parents()
897 parents = self.dirstate.parents()
898 self.destroying()
898 self.destroying()
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 if os.path.exists(self.join('undo.bookmarks')):
900 if os.path.exists(self.join('undo.bookmarks')):
901 util.rename(self.join('undo.bookmarks'),
901 util.rename(self.join('undo.bookmarks'),
902 self.join('bookmarks'))
902 self.join('bookmarks'))
903 if os.path.exists(self.sjoin('undo.phaseroots')):
903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 util.rename(self.sjoin('undo.phaseroots'),
904 util.rename(self.sjoin('undo.phaseroots'),
905 self.sjoin('phaseroots'))
905 self.sjoin('phaseroots'))
906 self.invalidate()
906 self.invalidate()
907
907
908 # Discard all cache entries to force reloading everything.
908 # Discard all cache entries to force reloading everything.
909 self._filecache.clear()
909 self._filecache.clear()
910
910
911 parentgone = (parents[0] not in self.changelog.nodemap or
911 parentgone = (parents[0] not in self.changelog.nodemap or
912 parents[1] not in self.changelog.nodemap)
912 parents[1] not in self.changelog.nodemap)
913 if parentgone:
913 if parentgone:
914 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
914 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
915 try:
915 try:
916 branch = self.opener.read('undo.branch')
916 branch = self.opener.read('undo.branch')
917 self.dirstate.setbranch(encoding.tolocal(branch))
917 self.dirstate.setbranch(encoding.tolocal(branch))
918 except IOError:
918 except IOError:
919 ui.warn(_('named branch could not be reset: '
919 ui.warn(_('named branch could not be reset: '
920 'current branch is still \'%s\'\n')
920 'current branch is still \'%s\'\n')
921 % self.dirstate.branch())
921 % self.dirstate.branch())
922
922
923 self.dirstate.invalidate()
923 self.dirstate.invalidate()
924 parents = tuple([p.rev() for p in self.parents()])
924 parents = tuple([p.rev() for p in self.parents()])
925 if len(parents) > 1:
925 if len(parents) > 1:
926 ui.status(_('working directory now based on '
926 ui.status(_('working directory now based on '
927 'revisions %d and %d\n') % parents)
927 'revisions %d and %d\n') % parents)
928 else:
928 else:
929 ui.status(_('working directory now based on '
929 ui.status(_('working directory now based on '
930 'revision %d\n') % parents)
930 'revision %d\n') % parents)
931 # TODO: if we know which new heads may result from this rollback, pass
931 # TODO: if we know which new heads may result from this rollback, pass
932 # them to destroy(), which will prevent the branchhead cache from being
932 # them to destroy(), which will prevent the branchhead cache from being
933 # invalidated.
933 # invalidated.
934 self.destroyed()
934 self.destroyed()
935 return 0
935 return 0
936
936
937 def invalidatecaches(self):
937 def invalidatecaches(self):
938
938
939 if '_tagscache' in vars(self):
939 if '_tagscache' in vars(self):
940 # can't use delattr on proxy
940 # can't use delattr on proxy
941 del self.__dict__['_tagscache']
941 del self.__dict__['_tagscache']
942
942
943 self.unfiltered()._branchcaches.clear()
943 self.unfiltered()._branchcaches.clear()
944 self.invalidatevolatilesets()
944 self.invalidatevolatilesets()
945
945
946 def invalidatevolatilesets(self):
946 def invalidatevolatilesets(self):
947 self.filteredrevcache.clear()
947 self.filteredrevcache.clear()
948 obsolete.clearobscaches(self)
948 obsolete.clearobscaches(self)
949
949
950 def invalidatedirstate(self):
950 def invalidatedirstate(self):
951 '''Invalidates the dirstate, causing the next call to dirstate
951 '''Invalidates the dirstate, causing the next call to dirstate
952 to check if it was modified since the last time it was read,
952 to check if it was modified since the last time it was read,
953 rereading it if it has.
953 rereading it if it has.
954
954
955 This is different to dirstate.invalidate() that it doesn't always
955 This is different to dirstate.invalidate() that it doesn't always
956 rereads the dirstate. Use dirstate.invalidate() if you want to
956 rereads the dirstate. Use dirstate.invalidate() if you want to
957 explicitly read the dirstate again (i.e. restoring it to a previous
957 explicitly read the dirstate again (i.e. restoring it to a previous
958 known good state).'''
958 known good state).'''
959 if hasunfilteredcache(self, 'dirstate'):
959 if hasunfilteredcache(self, 'dirstate'):
960 for k in self.dirstate._filecache:
960 for k in self.dirstate._filecache:
961 try:
961 try:
962 delattr(self.dirstate, k)
962 delattr(self.dirstate, k)
963 except AttributeError:
963 except AttributeError:
964 pass
964 pass
965 delattr(self.unfiltered(), 'dirstate')
965 delattr(self.unfiltered(), 'dirstate')
966
966
967 def invalidate(self):
967 def invalidate(self):
968 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
968 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
969 for k in self._filecache:
969 for k in self._filecache:
970 # dirstate is invalidated separately in invalidatedirstate()
970 # dirstate is invalidated separately in invalidatedirstate()
971 if k == 'dirstate':
971 if k == 'dirstate':
972 continue
972 continue
973
973
974 try:
974 try:
975 delattr(unfiltered, k)
975 delattr(unfiltered, k)
976 except AttributeError:
976 except AttributeError:
977 pass
977 pass
978 self.invalidatecaches()
978 self.invalidatecaches()
979
979
980 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
980 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
981 try:
981 try:
982 l = lock.lock(lockname, 0, releasefn, desc=desc)
982 l = lock.lock(lockname, 0, releasefn, desc=desc)
983 except error.LockHeld, inst:
983 except error.LockHeld, inst:
984 if not wait:
984 if not wait:
985 raise
985 raise
986 self.ui.warn(_("waiting for lock on %s held by %r\n") %
986 self.ui.warn(_("waiting for lock on %s held by %r\n") %
987 (desc, inst.locker))
987 (desc, inst.locker))
988 # default to 600 seconds timeout
988 # default to 600 seconds timeout
989 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
989 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
990 releasefn, desc=desc)
990 releasefn, desc=desc)
991 if acquirefn:
991 if acquirefn:
992 acquirefn()
992 acquirefn()
993 return l
993 return l
994
994
995 def _afterlock(self, callback):
995 def _afterlock(self, callback):
996 """add a callback to the current repository lock.
996 """add a callback to the current repository lock.
997
997
998 The callback will be executed on lock release."""
998 The callback will be executed on lock release."""
999 l = self._lockref and self._lockref()
999 l = self._lockref and self._lockref()
1000 if l:
1000 if l:
1001 l.postrelease.append(callback)
1001 l.postrelease.append(callback)
1002 else:
1002 else:
1003 callback()
1003 callback()
1004
1004
1005 def lock(self, wait=True):
1005 def lock(self, wait=True):
1006 '''Lock the repository store (.hg/store) and return a weak reference
1006 '''Lock the repository store (.hg/store) and return a weak reference
1007 to the lock. Use this before modifying the store (e.g. committing or
1007 to the lock. Use this before modifying the store (e.g. committing or
1008 stripping). If you are opening a transaction, get a lock as well.)'''
1008 stripping). If you are opening a transaction, get a lock as well.)'''
1009 l = self._lockref and self._lockref()
1009 l = self._lockref and self._lockref()
1010 if l is not None and l.held:
1010 if l is not None and l.held:
1011 l.lock()
1011 l.lock()
1012 return l
1012 return l
1013
1013
1014 def unlock():
1014 def unlock():
1015 self.store.write()
1015 self.store.write()
1016 if hasunfilteredcache(self, '_phasecache'):
1016 if hasunfilteredcache(self, '_phasecache'):
1017 self._phasecache.write()
1017 self._phasecache.write()
1018 for k, ce in self._filecache.items():
1018 for k, ce in self._filecache.items():
1019 if k == 'dirstate' or k not in self.__dict__:
1019 if k == 'dirstate' or k not in self.__dict__:
1020 continue
1020 continue
1021 ce.refresh()
1021 ce.refresh()
1022
1022
1023 l = self._lock(self.sjoin("lock"), wait, unlock,
1023 l = self._lock(self.sjoin("lock"), wait, unlock,
1024 self.invalidate, _('repository %s') % self.origroot)
1024 self.invalidate, _('repository %s') % self.origroot)
1025 self._lockref = weakref.ref(l)
1025 self._lockref = weakref.ref(l)
1026 return l
1026 return l
1027
1027
1028 def wlock(self, wait=True):
1028 def wlock(self, wait=True):
1029 '''Lock the non-store parts of the repository (everything under
1029 '''Lock the non-store parts of the repository (everything under
1030 .hg except .hg/store) and return a weak reference to the lock.
1030 .hg except .hg/store) and return a weak reference to the lock.
1031 Use this before modifying files in .hg.'''
1031 Use this before modifying files in .hg.'''
1032 l = self._wlockref and self._wlockref()
1032 l = self._wlockref and self._wlockref()
1033 if l is not None and l.held:
1033 if l is not None and l.held:
1034 l.lock()
1034 l.lock()
1035 return l
1035 return l
1036
1036
1037 def unlock():
1037 def unlock():
1038 self.dirstate.write()
1038 self.dirstate.write()
1039 ce = self._filecache.get('dirstate')
1039 ce = self._filecache.get('dirstate')
1040 if ce:
1040 if ce:
1041 ce.refresh()
1041 ce.refresh()
1042
1042
1043 l = self._lock(self.join("wlock"), wait, unlock,
1043 l = self._lock(self.join("wlock"), wait, unlock,
1044 self.invalidatedirstate, _('working directory of %s') %
1044 self.invalidatedirstate, _('working directory of %s') %
1045 self.origroot)
1045 self.origroot)
1046 self._wlockref = weakref.ref(l)
1046 self._wlockref = weakref.ref(l)
1047 return l
1047 return l
1048
1048
1049 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1049 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1050 """
1050 """
1051 commit an individual file as part of a larger transaction
1051 commit an individual file as part of a larger transaction
1052 """
1052 """
1053
1053
1054 fname = fctx.path()
1054 fname = fctx.path()
1055 text = fctx.data()
1055 text = fctx.data()
1056 flog = self.file(fname)
1056 flog = self.file(fname)
1057 fparent1 = manifest1.get(fname, nullid)
1057 fparent1 = manifest1.get(fname, nullid)
1058 fparent2 = fparent2o = manifest2.get(fname, nullid)
1058 fparent2 = fparent2o = manifest2.get(fname, nullid)
1059
1059
1060 meta = {}
1060 meta = {}
1061 copy = fctx.renamed()
1061 copy = fctx.renamed()
1062 if copy and copy[0] != fname:
1062 if copy and copy[0] != fname:
1063 # Mark the new revision of this file as a copy of another
1063 # Mark the new revision of this file as a copy of another
1064 # file. This copy data will effectively act as a parent
1064 # file. This copy data will effectively act as a parent
1065 # of this new revision. If this is a merge, the first
1065 # of this new revision. If this is a merge, the first
1066 # parent will be the nullid (meaning "look up the copy data")
1066 # parent will be the nullid (meaning "look up the copy data")
1067 # and the second one will be the other parent. For example:
1067 # and the second one will be the other parent. For example:
1068 #
1068 #
1069 # 0 --- 1 --- 3 rev1 changes file foo
1069 # 0 --- 1 --- 3 rev1 changes file foo
1070 # \ / rev2 renames foo to bar and changes it
1070 # \ / rev2 renames foo to bar and changes it
1071 # \- 2 -/ rev3 should have bar with all changes and
1071 # \- 2 -/ rev3 should have bar with all changes and
1072 # should record that bar descends from
1072 # should record that bar descends from
1073 # bar in rev2 and foo in rev1
1073 # bar in rev2 and foo in rev1
1074 #
1074 #
1075 # this allows this merge to succeed:
1075 # this allows this merge to succeed:
1076 #
1076 #
1077 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1077 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1078 # \ / merging rev3 and rev4 should use bar@rev2
1078 # \ / merging rev3 and rev4 should use bar@rev2
1079 # \- 2 --- 4 as the merge base
1079 # \- 2 --- 4 as the merge base
1080 #
1080 #
1081
1081
1082 cfname = copy[0]
1082 cfname = copy[0]
1083 crev = manifest1.get(cfname)
1083 crev = manifest1.get(cfname)
1084 newfparent = fparent2
1084 newfparent = fparent2
1085
1085
1086 if manifest2: # branch merge
1086 if manifest2: # branch merge
1087 if fparent2 == nullid or crev is None: # copied on remote side
1087 if fparent2 == nullid or crev is None: # copied on remote side
1088 if cfname in manifest2:
1088 if cfname in manifest2:
1089 crev = manifest2[cfname]
1089 crev = manifest2[cfname]
1090 newfparent = fparent1
1090 newfparent = fparent1
1091
1091
1092 # find source in nearest ancestor if we've lost track
1092 # find source in nearest ancestor if we've lost track
1093 if not crev:
1093 if not crev:
1094 self.ui.debug(" %s: searching for copy revision for %s\n" %
1094 self.ui.debug(" %s: searching for copy revision for %s\n" %
1095 (fname, cfname))
1095 (fname, cfname))
1096 for ancestor in self[None].ancestors():
1096 for ancestor in self[None].ancestors():
1097 if cfname in ancestor:
1097 if cfname in ancestor:
1098 crev = ancestor[cfname].filenode()
1098 crev = ancestor[cfname].filenode()
1099 break
1099 break
1100
1100
1101 if crev:
1101 if crev:
1102 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1102 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1103 meta["copy"] = cfname
1103 meta["copy"] = cfname
1104 meta["copyrev"] = hex(crev)
1104 meta["copyrev"] = hex(crev)
1105 fparent1, fparent2 = nullid, newfparent
1105 fparent1, fparent2 = nullid, newfparent
1106 else:
1106 else:
1107 self.ui.warn(_("warning: can't find ancestor for '%s' "
1107 self.ui.warn(_("warning: can't find ancestor for '%s' "
1108 "copied from '%s'!\n") % (fname, cfname))
1108 "copied from '%s'!\n") % (fname, cfname))
1109
1109
1110 elif fparent2 != nullid:
1110 elif fparent2 != nullid:
1111 # is one parent an ancestor of the other?
1111 # is one parent an ancestor of the other?
1112 fparentancestor = flog.ancestor(fparent1, fparent2)
1112 fparentancestor = flog.ancestor(fparent1, fparent2)
1113 if fparentancestor == fparent1:
1113 if fparentancestor == fparent1:
1114 fparent1, fparent2 = fparent2, nullid
1114 fparent1, fparent2 = fparent2, nullid
1115 elif fparentancestor == fparent2:
1115 elif fparentancestor == fparent2:
1116 fparent2 = nullid
1116 fparent2 = nullid
1117
1117
1118 # is the file changed?
1118 # is the file changed?
1119 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1119 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1120 changelist.append(fname)
1120 changelist.append(fname)
1121 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1121 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1122
1122
1123 # are just the flags changed during merge?
1123 # are just the flags changed during merge?
1124 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1124 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1125 changelist.append(fname)
1125 changelist.append(fname)
1126
1126
1127 return fparent1
1127 return fparent1
1128
1128
1129 @unfilteredmethod
1129 @unfilteredmethod
1130 def commit(self, text="", user=None, date=None, match=None, force=False,
1130 def commit(self, text="", user=None, date=None, match=None, force=False,
1131 editor=False, extra={}):
1131 editor=False, extra={}):
1132 """Add a new revision to current repository.
1132 """Add a new revision to current repository.
1133
1133
1134 Revision information is gathered from the working directory,
1134 Revision information is gathered from the working directory,
1135 match can be used to filter the committed files. If editor is
1135 match can be used to filter the committed files. If editor is
1136 supplied, it is called to get a commit message.
1136 supplied, it is called to get a commit message.
1137 """
1137 """
1138
1138
1139 def fail(f, msg):
1139 def fail(f, msg):
1140 raise util.Abort('%s: %s' % (f, msg))
1140 raise util.Abort('%s: %s' % (f, msg))
1141
1141
1142 if not match:
1142 if not match:
1143 match = matchmod.always(self.root, '')
1143 match = matchmod.always(self.root, '')
1144
1144
1145 if not force:
1145 if not force:
1146 vdirs = []
1146 vdirs = []
1147 match.dir = vdirs.append
1147 match.dir = vdirs.append
1148 match.bad = fail
1148 match.bad = fail
1149
1149
1150 wlock = self.wlock()
1150 wlock = self.wlock()
1151 try:
1151 try:
1152 wctx = self[None]
1152 wctx = self[None]
1153 merge = len(wctx.parents()) > 1
1153 merge = len(wctx.parents()) > 1
1154
1154
1155 if (not force and merge and match and
1155 if (not force and merge and match and
1156 (match.files() or match.anypats())):
1156 (match.files() or match.anypats())):
1157 raise util.Abort(_('cannot partially commit a merge '
1157 raise util.Abort(_('cannot partially commit a merge '
1158 '(do not specify files or patterns)'))
1158 '(do not specify files or patterns)'))
1159
1159
1160 changes = self.status(match=match, clean=force)
1160 changes = self.status(match=match, clean=force)
1161 if force:
1161 if force:
1162 changes[0].extend(changes[6]) # mq may commit unchanged files
1162 changes[0].extend(changes[6]) # mq may commit unchanged files
1163
1163
1164 # check subrepos
1164 # check subrepos
1165 subs = []
1165 subs = []
1166 commitsubs = set()
1166 commitsubs = set()
1167 newstate = wctx.substate.copy()
1167 newstate = wctx.substate.copy()
1168 # only manage subrepos and .hgsubstate if .hgsub is present
1168 # only manage subrepos and .hgsubstate if .hgsub is present
1169 if '.hgsub' in wctx:
1169 if '.hgsub' in wctx:
1170 # we'll decide whether to track this ourselves, thanks
1170 # we'll decide whether to track this ourselves, thanks
1171 if '.hgsubstate' in changes[0]:
1171 if '.hgsubstate' in changes[0]:
1172 changes[0].remove('.hgsubstate')
1172 changes[0].remove('.hgsubstate')
1173 if '.hgsubstate' in changes[2]:
1173 if '.hgsubstate' in changes[2]:
1174 changes[2].remove('.hgsubstate')
1174 changes[2].remove('.hgsubstate')
1175
1175
1176 # compare current state to last committed state
1176 # compare current state to last committed state
1177 # build new substate based on last committed state
1177 # build new substate based on last committed state
1178 oldstate = wctx.p1().substate
1178 oldstate = wctx.p1().substate
1179 for s in sorted(newstate.keys()):
1179 for s in sorted(newstate.keys()):
1180 if not match(s):
1180 if not match(s):
1181 # ignore working copy, use old state if present
1181 # ignore working copy, use old state if present
1182 if s in oldstate:
1182 if s in oldstate:
1183 newstate[s] = oldstate[s]
1183 newstate[s] = oldstate[s]
1184 continue
1184 continue
1185 if not force:
1185 if not force:
1186 raise util.Abort(
1186 raise util.Abort(
1187 _("commit with new subrepo %s excluded") % s)
1187 _("commit with new subrepo %s excluded") % s)
1188 if wctx.sub(s).dirty(True):
1188 if wctx.sub(s).dirty(True):
1189 if not self.ui.configbool('ui', 'commitsubrepos'):
1189 if not self.ui.configbool('ui', 'commitsubrepos'):
1190 raise util.Abort(
1190 raise util.Abort(
1191 _("uncommitted changes in subrepo %s") % s,
1191 _("uncommitted changes in subrepo %s") % s,
1192 hint=_("use --subrepos for recursive commit"))
1192 hint=_("use --subrepos for recursive commit"))
1193 subs.append(s)
1193 subs.append(s)
1194 commitsubs.add(s)
1194 commitsubs.add(s)
1195 else:
1195 else:
1196 bs = wctx.sub(s).basestate()
1196 bs = wctx.sub(s).basestate()
1197 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1197 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1198 if oldstate.get(s, (None, None, None))[1] != bs:
1198 if oldstate.get(s, (None, None, None))[1] != bs:
1199 subs.append(s)
1199 subs.append(s)
1200
1200
1201 # check for removed subrepos
1201 # check for removed subrepos
1202 for p in wctx.parents():
1202 for p in wctx.parents():
1203 r = [s for s in p.substate if s not in newstate]
1203 r = [s for s in p.substate if s not in newstate]
1204 subs += [s for s in r if match(s)]
1204 subs += [s for s in r if match(s)]
1205 if subs:
1205 if subs:
1206 if (not match('.hgsub') and
1206 if (not match('.hgsub') and
1207 '.hgsub' in (wctx.modified() + wctx.added())):
1207 '.hgsub' in (wctx.modified() + wctx.added())):
1208 raise util.Abort(
1208 raise util.Abort(
1209 _("can't commit subrepos without .hgsub"))
1209 _("can't commit subrepos without .hgsub"))
1210 changes[0].insert(0, '.hgsubstate')
1210 changes[0].insert(0, '.hgsubstate')
1211
1211
1212 elif '.hgsub' in changes[2]:
1212 elif '.hgsub' in changes[2]:
1213 # clean up .hgsubstate when .hgsub is removed
1213 # clean up .hgsubstate when .hgsub is removed
1214 if ('.hgsubstate' in wctx and
1214 if ('.hgsubstate' in wctx and
1215 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1215 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1216 changes[2].insert(0, '.hgsubstate')
1216 changes[2].insert(0, '.hgsubstate')
1217
1217
1218 # make sure all explicit patterns are matched
1218 # make sure all explicit patterns are matched
1219 if not force and match.files():
1219 if not force and match.files():
1220 matched = set(changes[0] + changes[1] + changes[2])
1220 matched = set(changes[0] + changes[1] + changes[2])
1221
1221
1222 for f in match.files():
1222 for f in match.files():
1223 f = self.dirstate.normalize(f)
1223 f = self.dirstate.normalize(f)
1224 if f == '.' or f in matched or f in wctx.substate:
1224 if f == '.' or f in matched or f in wctx.substate:
1225 continue
1225 continue
1226 if f in changes[3]: # missing
1226 if f in changes[3]: # missing
1227 fail(f, _('file not found!'))
1227 fail(f, _('file not found!'))
1228 if f in vdirs: # visited directory
1228 if f in vdirs: # visited directory
1229 d = f + '/'
1229 d = f + '/'
1230 for mf in matched:
1230 for mf in matched:
1231 if mf.startswith(d):
1231 if mf.startswith(d):
1232 break
1232 break
1233 else:
1233 else:
1234 fail(f, _("no match under directory!"))
1234 fail(f, _("no match under directory!"))
1235 elif f not in self.dirstate:
1235 elif f not in self.dirstate:
1236 fail(f, _("file not tracked!"))
1236 fail(f, _("file not tracked!"))
1237
1237
1238 if (not force and not extra.get("close") and not merge
1238 if (not force and not extra.get("close") and not merge
1239 and not (changes[0] or changes[1] or changes[2])
1239 and not (changes[0] or changes[1] or changes[2])
1240 and wctx.branch() == wctx.p1().branch()):
1240 and wctx.branch() == wctx.p1().branch()):
1241 return None
1241 return None
1242
1242
1243 if merge and changes[3]:
1243 if merge and changes[3]:
1244 raise util.Abort(_("cannot commit merge with missing files"))
1244 raise util.Abort(_("cannot commit merge with missing files"))
1245
1245
1246 ms = mergemod.mergestate(self)
1246 ms = mergemod.mergestate(self)
1247 for f in changes[0]:
1247 for f in changes[0]:
1248 if f in ms and ms[f] == 'u':
1248 if f in ms and ms[f] == 'u':
1249 raise util.Abort(_("unresolved merge conflicts "
1249 raise util.Abort(_("unresolved merge conflicts "
1250 "(see hg help resolve)"))
1250 "(see hg help resolve)"))
1251
1251
1252 cctx = context.workingctx(self, text, user, date, extra, changes)
1252 cctx = context.workingctx(self, text, user, date, extra, changes)
1253 if editor:
1253 if editor:
1254 cctx._text = editor(self, cctx, subs)
1254 cctx._text = editor(self, cctx, subs)
1255 edited = (text != cctx._text)
1255 edited = (text != cctx._text)
1256
1256
1257 # commit subs and write new state
1257 # commit subs and write new state
1258 if subs:
1258 if subs:
1259 for s in sorted(commitsubs):
1259 for s in sorted(commitsubs):
1260 sub = wctx.sub(s)
1260 sub = wctx.sub(s)
1261 self.ui.status(_('committing subrepository %s\n') %
1261 self.ui.status(_('committing subrepository %s\n') %
1262 subrepo.subrelpath(sub))
1262 subrepo.subrelpath(sub))
1263 sr = sub.commit(cctx._text, user, date)
1263 sr = sub.commit(cctx._text, user, date)
1264 newstate[s] = (newstate[s][0], sr)
1264 newstate[s] = (newstate[s][0], sr)
1265 subrepo.writestate(self, newstate)
1265 subrepo.writestate(self, newstate)
1266
1266
1267 # Save commit message in case this transaction gets rolled back
1267 # Save commit message in case this transaction gets rolled back
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1268 # (e.g. by a pretxncommit hook). Leave the content alone on
1269 # the assumption that the user will use the same editor again.
1269 # the assumption that the user will use the same editor again.
1270 msgfn = self.savecommitmessage(cctx._text)
1270 msgfn = self.savecommitmessage(cctx._text)
1271
1271
1272 p1, p2 = self.dirstate.parents()
1272 p1, p2 = self.dirstate.parents()
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1273 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1274 try:
1274 try:
1275 self.hook("precommit", throw=True, parent1=hookp1,
1275 self.hook("precommit", throw=True, parent1=hookp1,
1276 parent2=hookp2)
1276 parent2=hookp2)
1277 ret = self.commitctx(cctx, True)
1277 ret = self.commitctx(cctx, True)
1278 except: # re-raises
1278 except: # re-raises
1279 if edited:
1279 if edited:
1280 self.ui.write(
1280 self.ui.write(
1281 _('note: commit message saved in %s\n') % msgfn)
1281 _('note: commit message saved in %s\n') % msgfn)
1282 raise
1282 raise
1283
1283
1284 # update bookmarks, dirstate and mergestate
1284 # update bookmarks, dirstate and mergestate
1285 bookmarks.update(self, [p1, p2], ret)
1285 bookmarks.update(self, [p1, p2], ret)
1286 for f in changes[0] + changes[1]:
1286 for f in changes[0] + changes[1]:
1287 self.dirstate.normal(f)
1287 self.dirstate.normal(f)
1288 for f in changes[2]:
1288 for f in changes[2]:
1289 self.dirstate.drop(f)
1289 self.dirstate.drop(f)
1290 self.dirstate.setparents(ret)
1290 self.dirstate.setparents(ret)
1291 ms.reset()
1291 ms.reset()
1292 finally:
1292 finally:
1293 wlock.release()
1293 wlock.release()
1294
1294
1295 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1295 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1296 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1296 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1297 self._afterlock(commithook)
1297 self._afterlock(commithook)
1298 return ret
1298 return ret
1299
1299
1300 @unfilteredmethod
1300 @unfilteredmethod
1301 def commitctx(self, ctx, error=False):
1301 def commitctx(self, ctx, error=False):
1302 """Add a new revision to current repository.
1302 """Add a new revision to current repository.
1303 Revision information is passed via the context argument.
1303 Revision information is passed via the context argument.
1304 """
1304 """
1305
1305
1306 tr = lock = None
1306 tr = lock = None
1307 removed = list(ctx.removed())
1307 removed = list(ctx.removed())
1308 p1, p2 = ctx.p1(), ctx.p2()
1308 p1, p2 = ctx.p1(), ctx.p2()
1309 user = ctx.user()
1309 user = ctx.user()
1310
1310
1311 lock = self.lock()
1311 lock = self.lock()
1312 try:
1312 try:
1313 tr = self.transaction("commit")
1313 tr = self.transaction("commit")
1314 trp = weakref.proxy(tr)
1314 trp = weakref.proxy(tr)
1315
1315
1316 if ctx.files():
1316 if ctx.files():
1317 m1 = p1.manifest().copy()
1317 m1 = p1.manifest().copy()
1318 m2 = p2.manifest()
1318 m2 = p2.manifest()
1319
1319
1320 # check in files
1320 # check in files
1321 new = {}
1321 new = {}
1322 changed = []
1322 changed = []
1323 linkrev = len(self)
1323 linkrev = len(self)
1324 for f in sorted(ctx.modified() + ctx.added()):
1324 for f in sorted(ctx.modified() + ctx.added()):
1325 self.ui.note(f + "\n")
1325 self.ui.note(f + "\n")
1326 try:
1326 try:
1327 fctx = ctx[f]
1327 fctx = ctx[f]
1328 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1328 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1329 changed)
1329 changed)
1330 m1.set(f, fctx.flags())
1330 m1.set(f, fctx.flags())
1331 except OSError, inst:
1331 except OSError, inst:
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1333 raise
1333 raise
1334 except IOError, inst:
1334 except IOError, inst:
1335 errcode = getattr(inst, 'errno', errno.ENOENT)
1335 errcode = getattr(inst, 'errno', errno.ENOENT)
1336 if error or errcode and errcode != errno.ENOENT:
1336 if error or errcode and errcode != errno.ENOENT:
1337 self.ui.warn(_("trouble committing %s!\n") % f)
1337 self.ui.warn(_("trouble committing %s!\n") % f)
1338 raise
1338 raise
1339 else:
1339 else:
1340 removed.append(f)
1340 removed.append(f)
1341
1341
1342 # update manifest
1342 # update manifest
1343 m1.update(new)
1343 m1.update(new)
1344 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1344 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1345 drop = [f for f in removed if f in m1]
1345 drop = [f for f in removed if f in m1]
1346 for f in drop:
1346 for f in drop:
1347 del m1[f]
1347 del m1[f]
1348 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1348 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1349 p2.manifestnode(), (new, drop))
1349 p2.manifestnode(), (new, drop))
1350 files = changed + removed
1350 files = changed + removed
1351 else:
1351 else:
1352 mn = p1.manifestnode()
1352 mn = p1.manifestnode()
1353 files = []
1353 files = []
1354
1354
1355 # update changelog
1355 # update changelog
1356 self.changelog.delayupdate()
1356 self.changelog.delayupdate()
1357 n = self.changelog.add(mn, files, ctx.description(),
1357 n = self.changelog.add(mn, files, ctx.description(),
1358 trp, p1.node(), p2.node(),
1358 trp, p1.node(), p2.node(),
1359 user, ctx.date(), ctx.extra().copy())
1359 user, ctx.date(), ctx.extra().copy())
1360 p = lambda: self.changelog.writepending() and self.root or ""
1360 p = lambda: self.changelog.writepending() and self.root or ""
1361 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1361 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1362 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1362 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1363 parent2=xp2, pending=p)
1363 parent2=xp2, pending=p)
1364 self.changelog.finalize(trp)
1364 self.changelog.finalize(trp)
1365 # set the new commit is proper phase
1365 # set the new commit is proper phase
1366 targetphase = phases.newcommitphase(self.ui)
1366 targetphase = phases.newcommitphase(self.ui)
1367 if targetphase:
1367 if targetphase:
1368 # retract boundary do not alter parent changeset.
1368 # retract boundary do not alter parent changeset.
1369 # if a parent have higher the resulting phase will
1369 # if a parent have higher the resulting phase will
1370 # be compliant anyway
1370 # be compliant anyway
1371 #
1371 #
1372 # if minimal phase was 0 we don't need to retract anything
1372 # if minimal phase was 0 we don't need to retract anything
1373 phases.retractboundary(self, targetphase, [n])
1373 phases.retractboundary(self, targetphase, [n])
1374 tr.close()
1374 tr.close()
1375 branchmap.updatecache(self)
1375 branchmap.updatecache(self)
1376 return n
1376 return n
1377 finally:
1377 finally:
1378 if tr:
1378 if tr:
1379 tr.release()
1379 tr.release()
1380 lock.release()
1380 lock.release()
1381
1381
1382 @unfilteredmethod
1382 @unfilteredmethod
1383 def destroying(self):
1383 def destroying(self):
1384 '''Inform the repository that nodes are about to be destroyed.
1384 '''Inform the repository that nodes are about to be destroyed.
1385 Intended for use by strip and rollback, so there's a common
1385 Intended for use by strip and rollback, so there's a common
1386 place for anything that has to be done before destroying history.
1386 place for anything that has to be done before destroying history.
1387
1387
1388 This is mostly useful for saving state that is in memory and waiting
1388 This is mostly useful for saving state that is in memory and waiting
1389 to be flushed when the current lock is released. Because a call to
1389 to be flushed when the current lock is released. Because a call to
1390 destroyed is imminent, the repo will be invalidated causing those
1390 destroyed is imminent, the repo will be invalidated causing those
1391 changes to stay in memory (waiting for the next unlock), or vanish
1391 changes to stay in memory (waiting for the next unlock), or vanish
1392 completely.
1392 completely.
1393 '''
1393 '''
1394 pass
1394 # It simplifies the logic around updating the branchheads cache if we
1395 # only have to consider the effect of the stripped revisions and not
1396 # revisions missing because the cache is out-of-date.
1397 branchmap.updatecache(self)
1395
1398
1396 @unfilteredmethod
1399 @unfilteredmethod
1397 def destroyed(self, newheadnodes=None):
1400 def destroyed(self, newheadnodes=None):
1398 '''Inform the repository that nodes have been destroyed.
1401 '''Inform the repository that nodes have been destroyed.
1399 Intended for use by strip and rollback, so there's a common
1402 Intended for use by strip and rollback, so there's a common
1400 place for anything that has to be done after destroying history.
1403 place for anything that has to be done after destroying history.
1401
1404
1402 If you know the branchheadcache was uptodate before nodes were removed
1405 If you know the branchheadcache was uptodate before nodes were removed
1403 and you also know the set of candidate new heads that may have resulted
1406 and you also know the set of candidate new heads that may have resulted
1404 from the destruction, you can set newheadnodes. This will enable the
1407 from the destruction, you can set newheadnodes. This will enable the
1405 code to update the branchheads cache, rather than having future code
1408 code to update the branchheads cache, rather than having future code
1406 decide it's invalid and regenerating it from scratch.
1409 decide it's invalid and regenerating it from scratch.
1407 '''
1410 '''
1408 # When one tries to:
1411 # When one tries to:
1409 # 1) destroy nodes thus calling this method (e.g. strip)
1412 # 1) destroy nodes thus calling this method (e.g. strip)
1410 # 2) use phasecache somewhere (e.g. commit)
1413 # 2) use phasecache somewhere (e.g. commit)
1411 #
1414 #
1412 # then 2) will fail because the phasecache contains nodes that were
1415 # then 2) will fail because the phasecache contains nodes that were
1413 # removed. We can either remove phasecache from the filecache,
1416 # removed. We can either remove phasecache from the filecache,
1414 # causing it to reload next time it is accessed, or simply filter
1417 # causing it to reload next time it is accessed, or simply filter
1415 # the removed nodes now and write the updated cache.
1418 # the removed nodes now and write the updated cache.
1416 if '_phasecache' in self._filecache:
1419 if '_phasecache' in self._filecache:
1417 self._phasecache.filterunknown(self)
1420 self._phasecache.filterunknown(self)
1418 self._phasecache.write()
1421 self._phasecache.write()
1419
1422
1420 # If we have info, newheadnodes, on how to update the branch cache, do
1423 # If we have info, newheadnodes, on how to update the branch cache, do
1421 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1424 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1422 # will be caught the next time it is read.
1425 # will be caught the next time it is read.
1423 if newheadnodes:
1426 if newheadnodes:
1424 cl = self.changelog
1427 cl = self.changelog
1425 revgen = (cl.rev(node) for node in newheadnodes
1428 revgen = (cl.rev(node) for node in newheadnodes
1426 if cl.hasnode(node))
1429 if cl.hasnode(node))
1427 cache = self._branchcaches[None]
1430 cache = self._branchcaches[None]
1428 cache.update(self, revgen)
1431 cache.update(self, revgen)
1429 cache.write(self)
1432 cache.write(self)
1430
1433
1431 # Ensure the persistent tag cache is updated. Doing it now
1434 # Ensure the persistent tag cache is updated. Doing it now
1432 # means that the tag cache only has to worry about destroyed
1435 # means that the tag cache only has to worry about destroyed
1433 # heads immediately after a strip/rollback. That in turn
1436 # heads immediately after a strip/rollback. That in turn
1434 # guarantees that "cachetip == currenttip" (comparing both rev
1437 # guarantees that "cachetip == currenttip" (comparing both rev
1435 # and node) always means no nodes have been added or destroyed.
1438 # and node) always means no nodes have been added or destroyed.
1436
1439
1437 # XXX this is suboptimal when qrefresh'ing: we strip the current
1440 # XXX this is suboptimal when qrefresh'ing: we strip the current
1438 # head, refresh the tag cache, then immediately add a new head.
1441 # head, refresh the tag cache, then immediately add a new head.
1439 # But I think doing it this way is necessary for the "instant
1442 # But I think doing it this way is necessary for the "instant
1440 # tag cache retrieval" case to work.
1443 # tag cache retrieval" case to work.
1441 self.invalidatecaches()
1444 self.invalidatecaches()
1442
1445
1443 # Discard all cache entries to force reloading everything.
1446 # Discard all cache entries to force reloading everything.
1444 self._filecache.clear()
1447 self._filecache.clear()
1445
1448
1446 def walk(self, match, node=None):
1449 def walk(self, match, node=None):
1447 '''
1450 '''
1448 walk recursively through the directory tree or a given
1451 walk recursively through the directory tree or a given
1449 changeset, finding all files matched by the match
1452 changeset, finding all files matched by the match
1450 function
1453 function
1451 '''
1454 '''
1452 return self[node].walk(match)
1455 return self[node].walk(match)
1453
1456
1454 def status(self, node1='.', node2=None, match=None,
1457 def status(self, node1='.', node2=None, match=None,
1455 ignored=False, clean=False, unknown=False,
1458 ignored=False, clean=False, unknown=False,
1456 listsubrepos=False):
1459 listsubrepos=False):
1457 """return status of files between two nodes or node and working
1460 """return status of files between two nodes or node and working
1458 directory.
1461 directory.
1459
1462
1460 If node1 is None, use the first dirstate parent instead.
1463 If node1 is None, use the first dirstate parent instead.
1461 If node2 is None, compare node1 with working directory.
1464 If node2 is None, compare node1 with working directory.
1462 """
1465 """
1463
1466
1464 def mfmatches(ctx):
1467 def mfmatches(ctx):
1465 mf = ctx.manifest().copy()
1468 mf = ctx.manifest().copy()
1466 if match.always():
1469 if match.always():
1467 return mf
1470 return mf
1468 for fn in mf.keys():
1471 for fn in mf.keys():
1469 if not match(fn):
1472 if not match(fn):
1470 del mf[fn]
1473 del mf[fn]
1471 return mf
1474 return mf
1472
1475
1473 if isinstance(node1, context.changectx):
1476 if isinstance(node1, context.changectx):
1474 ctx1 = node1
1477 ctx1 = node1
1475 else:
1478 else:
1476 ctx1 = self[node1]
1479 ctx1 = self[node1]
1477 if isinstance(node2, context.changectx):
1480 if isinstance(node2, context.changectx):
1478 ctx2 = node2
1481 ctx2 = node2
1479 else:
1482 else:
1480 ctx2 = self[node2]
1483 ctx2 = self[node2]
1481
1484
1482 working = ctx2.rev() is None
1485 working = ctx2.rev() is None
1483 parentworking = working and ctx1 == self['.']
1486 parentworking = working and ctx1 == self['.']
1484 match = match or matchmod.always(self.root, self.getcwd())
1487 match = match or matchmod.always(self.root, self.getcwd())
1485 listignored, listclean, listunknown = ignored, clean, unknown
1488 listignored, listclean, listunknown = ignored, clean, unknown
1486
1489
1487 # load earliest manifest first for caching reasons
1490 # load earliest manifest first for caching reasons
1488 if not working and ctx2.rev() < ctx1.rev():
1491 if not working and ctx2.rev() < ctx1.rev():
1489 ctx2.manifest()
1492 ctx2.manifest()
1490
1493
1491 if not parentworking:
1494 if not parentworking:
1492 def bad(f, msg):
1495 def bad(f, msg):
1493 # 'f' may be a directory pattern from 'match.files()',
1496 # 'f' may be a directory pattern from 'match.files()',
1494 # so 'f not in ctx1' is not enough
1497 # so 'f not in ctx1' is not enough
1495 if f not in ctx1 and f not in ctx1.dirs():
1498 if f not in ctx1 and f not in ctx1.dirs():
1496 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1499 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1497 match.bad = bad
1500 match.bad = bad
1498
1501
1499 if working: # we need to scan the working dir
1502 if working: # we need to scan the working dir
1500 subrepos = []
1503 subrepos = []
1501 if '.hgsub' in self.dirstate:
1504 if '.hgsub' in self.dirstate:
1502 subrepos = ctx2.substate.keys()
1505 subrepos = ctx2.substate.keys()
1503 s = self.dirstate.status(match, subrepos, listignored,
1506 s = self.dirstate.status(match, subrepos, listignored,
1504 listclean, listunknown)
1507 listclean, listunknown)
1505 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1508 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1506
1509
1507 # check for any possibly clean files
1510 # check for any possibly clean files
1508 if parentworking and cmp:
1511 if parentworking and cmp:
1509 fixup = []
1512 fixup = []
1510 # do a full compare of any files that might have changed
1513 # do a full compare of any files that might have changed
1511 for f in sorted(cmp):
1514 for f in sorted(cmp):
1512 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1515 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1513 or ctx1[f].cmp(ctx2[f])):
1516 or ctx1[f].cmp(ctx2[f])):
1514 modified.append(f)
1517 modified.append(f)
1515 else:
1518 else:
1516 fixup.append(f)
1519 fixup.append(f)
1517
1520
1518 # update dirstate for files that are actually clean
1521 # update dirstate for files that are actually clean
1519 if fixup:
1522 if fixup:
1520 if listclean:
1523 if listclean:
1521 clean += fixup
1524 clean += fixup
1522
1525
1523 try:
1526 try:
1524 # updating the dirstate is optional
1527 # updating the dirstate is optional
1525 # so we don't wait on the lock
1528 # so we don't wait on the lock
1526 wlock = self.wlock(False)
1529 wlock = self.wlock(False)
1527 try:
1530 try:
1528 for f in fixup:
1531 for f in fixup:
1529 self.dirstate.normal(f)
1532 self.dirstate.normal(f)
1530 finally:
1533 finally:
1531 wlock.release()
1534 wlock.release()
1532 except error.LockError:
1535 except error.LockError:
1533 pass
1536 pass
1534
1537
1535 if not parentworking:
1538 if not parentworking:
1536 mf1 = mfmatches(ctx1)
1539 mf1 = mfmatches(ctx1)
1537 if working:
1540 if working:
1538 # we are comparing working dir against non-parent
1541 # we are comparing working dir against non-parent
1539 # generate a pseudo-manifest for the working dir
1542 # generate a pseudo-manifest for the working dir
1540 mf2 = mfmatches(self['.'])
1543 mf2 = mfmatches(self['.'])
1541 for f in cmp + modified + added:
1544 for f in cmp + modified + added:
1542 mf2[f] = None
1545 mf2[f] = None
1543 mf2.set(f, ctx2.flags(f))
1546 mf2.set(f, ctx2.flags(f))
1544 for f in removed:
1547 for f in removed:
1545 if f in mf2:
1548 if f in mf2:
1546 del mf2[f]
1549 del mf2[f]
1547 else:
1550 else:
1548 # we are comparing two revisions
1551 # we are comparing two revisions
1549 deleted, unknown, ignored = [], [], []
1552 deleted, unknown, ignored = [], [], []
1550 mf2 = mfmatches(ctx2)
1553 mf2 = mfmatches(ctx2)
1551
1554
1552 modified, added, clean = [], [], []
1555 modified, added, clean = [], [], []
1553 withflags = mf1.withflags() | mf2.withflags()
1556 withflags = mf1.withflags() | mf2.withflags()
1554 for fn in mf2:
1557 for fn in mf2:
1555 if fn in mf1:
1558 if fn in mf1:
1556 if (fn not in deleted and
1559 if (fn not in deleted and
1557 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1560 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1558 (mf1[fn] != mf2[fn] and
1561 (mf1[fn] != mf2[fn] and
1559 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1562 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1560 modified.append(fn)
1563 modified.append(fn)
1561 elif listclean:
1564 elif listclean:
1562 clean.append(fn)
1565 clean.append(fn)
1563 del mf1[fn]
1566 del mf1[fn]
1564 elif fn not in deleted:
1567 elif fn not in deleted:
1565 added.append(fn)
1568 added.append(fn)
1566 removed = mf1.keys()
1569 removed = mf1.keys()
1567
1570
1568 if working and modified and not self.dirstate._checklink:
1571 if working and modified and not self.dirstate._checklink:
1569 # Symlink placeholders may get non-symlink-like contents
1572 # Symlink placeholders may get non-symlink-like contents
1570 # via user error or dereferencing by NFS or Samba servers,
1573 # via user error or dereferencing by NFS or Samba servers,
1571 # so we filter out any placeholders that don't look like a
1574 # so we filter out any placeholders that don't look like a
1572 # symlink
1575 # symlink
1573 sane = []
1576 sane = []
1574 for f in modified:
1577 for f in modified:
1575 if ctx2.flags(f) == 'l':
1578 if ctx2.flags(f) == 'l':
1576 d = ctx2[f].data()
1579 d = ctx2[f].data()
1577 if len(d) >= 1024 or '\n' in d or util.binary(d):
1580 if len(d) >= 1024 or '\n' in d or util.binary(d):
1578 self.ui.debug('ignoring suspect symlink placeholder'
1581 self.ui.debug('ignoring suspect symlink placeholder'
1579 ' "%s"\n' % f)
1582 ' "%s"\n' % f)
1580 continue
1583 continue
1581 sane.append(f)
1584 sane.append(f)
1582 modified = sane
1585 modified = sane
1583
1586
1584 r = modified, added, removed, deleted, unknown, ignored, clean
1587 r = modified, added, removed, deleted, unknown, ignored, clean
1585
1588
1586 if listsubrepos:
1589 if listsubrepos:
1587 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1590 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1588 if working:
1591 if working:
1589 rev2 = None
1592 rev2 = None
1590 else:
1593 else:
1591 rev2 = ctx2.substate[subpath][1]
1594 rev2 = ctx2.substate[subpath][1]
1592 try:
1595 try:
1593 submatch = matchmod.narrowmatcher(subpath, match)
1596 submatch = matchmod.narrowmatcher(subpath, match)
1594 s = sub.status(rev2, match=submatch, ignored=listignored,
1597 s = sub.status(rev2, match=submatch, ignored=listignored,
1595 clean=listclean, unknown=listunknown,
1598 clean=listclean, unknown=listunknown,
1596 listsubrepos=True)
1599 listsubrepos=True)
1597 for rfiles, sfiles in zip(r, s):
1600 for rfiles, sfiles in zip(r, s):
1598 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1601 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1599 except error.LookupError:
1602 except error.LookupError:
1600 self.ui.status(_("skipping missing subrepository: %s\n")
1603 self.ui.status(_("skipping missing subrepository: %s\n")
1601 % subpath)
1604 % subpath)
1602
1605
1603 for l in r:
1606 for l in r:
1604 l.sort()
1607 l.sort()
1605 return r
1608 return r
1606
1609
1607 def heads(self, start=None):
1610 def heads(self, start=None):
1608 heads = self.changelog.heads(start)
1611 heads = self.changelog.heads(start)
1609 # sort the output in rev descending order
1612 # sort the output in rev descending order
1610 return sorted(heads, key=self.changelog.rev, reverse=True)
1613 return sorted(heads, key=self.changelog.rev, reverse=True)
1611
1614
1612 def branchheads(self, branch=None, start=None, closed=False):
1615 def branchheads(self, branch=None, start=None, closed=False):
1613 '''return a (possibly filtered) list of heads for the given branch
1616 '''return a (possibly filtered) list of heads for the given branch
1614
1617
1615 Heads are returned in topological order, from newest to oldest.
1618 Heads are returned in topological order, from newest to oldest.
1616 If branch is None, use the dirstate branch.
1619 If branch is None, use the dirstate branch.
1617 If start is not None, return only heads reachable from start.
1620 If start is not None, return only heads reachable from start.
1618 If closed is True, return heads that are marked as closed as well.
1621 If closed is True, return heads that are marked as closed as well.
1619 '''
1622 '''
1620 if branch is None:
1623 if branch is None:
1621 branch = self[None].branch()
1624 branch = self[None].branch()
1622 branches = self.branchmap()
1625 branches = self.branchmap()
1623 if branch not in branches:
1626 if branch not in branches:
1624 return []
1627 return []
1625 # the cache returns heads ordered lowest to highest
1628 # the cache returns heads ordered lowest to highest
1626 bheads = list(reversed(branches[branch]))
1629 bheads = list(reversed(branches[branch]))
1627 if start is not None:
1630 if start is not None:
1628 # filter out the heads that cannot be reached from startrev
1631 # filter out the heads that cannot be reached from startrev
1629 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1632 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1630 bheads = [h for h in bheads if h in fbheads]
1633 bheads = [h for h in bheads if h in fbheads]
1631 if not closed:
1634 if not closed:
1632 bheads = [h for h in bheads if not self[h].closesbranch()]
1635 bheads = [h for h in bheads if not self[h].closesbranch()]
1633 return bheads
1636 return bheads
1634
1637
1635 def branches(self, nodes):
1638 def branches(self, nodes):
1636 if not nodes:
1639 if not nodes:
1637 nodes = [self.changelog.tip()]
1640 nodes = [self.changelog.tip()]
1638 b = []
1641 b = []
1639 for n in nodes:
1642 for n in nodes:
1640 t = n
1643 t = n
1641 while True:
1644 while True:
1642 p = self.changelog.parents(n)
1645 p = self.changelog.parents(n)
1643 if p[1] != nullid or p[0] == nullid:
1646 if p[1] != nullid or p[0] == nullid:
1644 b.append((t, n, p[0], p[1]))
1647 b.append((t, n, p[0], p[1]))
1645 break
1648 break
1646 n = p[0]
1649 n = p[0]
1647 return b
1650 return b
1648
1651
1649 def between(self, pairs):
1652 def between(self, pairs):
1650 r = []
1653 r = []
1651
1654
1652 for top, bottom in pairs:
1655 for top, bottom in pairs:
1653 n, l, i = top, [], 0
1656 n, l, i = top, [], 0
1654 f = 1
1657 f = 1
1655
1658
1656 while n != bottom and n != nullid:
1659 while n != bottom and n != nullid:
1657 p = self.changelog.parents(n)[0]
1660 p = self.changelog.parents(n)[0]
1658 if i == f:
1661 if i == f:
1659 l.append(n)
1662 l.append(n)
1660 f = f * 2
1663 f = f * 2
1661 n = p
1664 n = p
1662 i += 1
1665 i += 1
1663
1666
1664 r.append(l)
1667 r.append(l)
1665
1668
1666 return r
1669 return r
1667
1670
1668 def pull(self, remote, heads=None, force=False):
1671 def pull(self, remote, heads=None, force=False):
1669 # don't open transaction for nothing or you break future useful
1672 # don't open transaction for nothing or you break future useful
1670 # rollback call
1673 # rollback call
1671 tr = None
1674 tr = None
1672 trname = 'pull\n' + util.hidepassword(remote.url())
1675 trname = 'pull\n' + util.hidepassword(remote.url())
1673 lock = self.lock()
1676 lock = self.lock()
1674 try:
1677 try:
1675 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1678 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1676 force=force)
1679 force=force)
1677 common, fetch, rheads = tmp
1680 common, fetch, rheads = tmp
1678 if not fetch:
1681 if not fetch:
1679 self.ui.status(_("no changes found\n"))
1682 self.ui.status(_("no changes found\n"))
1680 added = []
1683 added = []
1681 result = 0
1684 result = 0
1682 else:
1685 else:
1683 tr = self.transaction(trname)
1686 tr = self.transaction(trname)
1684 if heads is None and list(common) == [nullid]:
1687 if heads is None and list(common) == [nullid]:
1685 self.ui.status(_("requesting all changes\n"))
1688 self.ui.status(_("requesting all changes\n"))
1686 elif heads is None and remote.capable('changegroupsubset'):
1689 elif heads is None and remote.capable('changegroupsubset'):
1687 # issue1320, avoid a race if remote changed after discovery
1690 # issue1320, avoid a race if remote changed after discovery
1688 heads = rheads
1691 heads = rheads
1689
1692
1690 if remote.capable('getbundle'):
1693 if remote.capable('getbundle'):
1691 cg = remote.getbundle('pull', common=common,
1694 cg = remote.getbundle('pull', common=common,
1692 heads=heads or rheads)
1695 heads=heads or rheads)
1693 elif heads is None:
1696 elif heads is None:
1694 cg = remote.changegroup(fetch, 'pull')
1697 cg = remote.changegroup(fetch, 'pull')
1695 elif not remote.capable('changegroupsubset'):
1698 elif not remote.capable('changegroupsubset'):
1696 raise util.Abort(_("partial pull cannot be done because "
1699 raise util.Abort(_("partial pull cannot be done because "
1697 "other repository doesn't support "
1700 "other repository doesn't support "
1698 "changegroupsubset."))
1701 "changegroupsubset."))
1699 else:
1702 else:
1700 cg = remote.changegroupsubset(fetch, heads, 'pull')
1703 cg = remote.changegroupsubset(fetch, heads, 'pull')
1701 clstart = len(self.changelog)
1704 clstart = len(self.changelog)
1702 result = self.addchangegroup(cg, 'pull', remote.url())
1705 result = self.addchangegroup(cg, 'pull', remote.url())
1703 clend = len(self.changelog)
1706 clend = len(self.changelog)
1704 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1707 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1705
1708
1706 # compute target subset
1709 # compute target subset
1707 if heads is None:
1710 if heads is None:
1708 # We pulled every thing possible
1711 # We pulled every thing possible
1709 # sync on everything common
1712 # sync on everything common
1710 subset = common + added
1713 subset = common + added
1711 else:
1714 else:
1712 # We pulled a specific subset
1715 # We pulled a specific subset
1713 # sync on this subset
1716 # sync on this subset
1714 subset = heads
1717 subset = heads
1715
1718
1716 # Get remote phases data from remote
1719 # Get remote phases data from remote
1717 remotephases = remote.listkeys('phases')
1720 remotephases = remote.listkeys('phases')
1718 publishing = bool(remotephases.get('publishing', False))
1721 publishing = bool(remotephases.get('publishing', False))
1719 if remotephases and not publishing:
1722 if remotephases and not publishing:
1720 # remote is new and unpublishing
1723 # remote is new and unpublishing
1721 pheads, _dr = phases.analyzeremotephases(self, subset,
1724 pheads, _dr = phases.analyzeremotephases(self, subset,
1722 remotephases)
1725 remotephases)
1723 phases.advanceboundary(self, phases.public, pheads)
1726 phases.advanceboundary(self, phases.public, pheads)
1724 phases.advanceboundary(self, phases.draft, subset)
1727 phases.advanceboundary(self, phases.draft, subset)
1725 else:
1728 else:
1726 # Remote is old or publishing all common changesets
1729 # Remote is old or publishing all common changesets
1727 # should be seen as public
1730 # should be seen as public
1728 phases.advanceboundary(self, phases.public, subset)
1731 phases.advanceboundary(self, phases.public, subset)
1729
1732
1730 if obsolete._enabled:
1733 if obsolete._enabled:
1731 self.ui.debug('fetching remote obsolete markers\n')
1734 self.ui.debug('fetching remote obsolete markers\n')
1732 remoteobs = remote.listkeys('obsolete')
1735 remoteobs = remote.listkeys('obsolete')
1733 if 'dump0' in remoteobs:
1736 if 'dump0' in remoteobs:
1734 if tr is None:
1737 if tr is None:
1735 tr = self.transaction(trname)
1738 tr = self.transaction(trname)
1736 for key in sorted(remoteobs, reverse=True):
1739 for key in sorted(remoteobs, reverse=True):
1737 if key.startswith('dump'):
1740 if key.startswith('dump'):
1738 data = base85.b85decode(remoteobs[key])
1741 data = base85.b85decode(remoteobs[key])
1739 self.obsstore.mergemarkers(tr, data)
1742 self.obsstore.mergemarkers(tr, data)
1740 self.invalidatevolatilesets()
1743 self.invalidatevolatilesets()
1741 if tr is not None:
1744 if tr is not None:
1742 tr.close()
1745 tr.close()
1743 finally:
1746 finally:
1744 if tr is not None:
1747 if tr is not None:
1745 tr.release()
1748 tr.release()
1746 lock.release()
1749 lock.release()
1747
1750
1748 return result
1751 return result
1749
1752
1750 def checkpush(self, force, revs):
1753 def checkpush(self, force, revs):
1751 """Extensions can override this function if additional checks have
1754 """Extensions can override this function if additional checks have
1752 to be performed before pushing, or call it if they override push
1755 to be performed before pushing, or call it if they override push
1753 command.
1756 command.
1754 """
1757 """
1755 pass
1758 pass
1756
1759
1757 def push(self, remote, force=False, revs=None, newbranch=False):
1760 def push(self, remote, force=False, revs=None, newbranch=False):
1758 '''Push outgoing changesets (limited by revs) from the current
1761 '''Push outgoing changesets (limited by revs) from the current
1759 repository to remote. Return an integer:
1762 repository to remote. Return an integer:
1760 - None means nothing to push
1763 - None means nothing to push
1761 - 0 means HTTP error
1764 - 0 means HTTP error
1762 - 1 means we pushed and remote head count is unchanged *or*
1765 - 1 means we pushed and remote head count is unchanged *or*
1763 we have outgoing changesets but refused to push
1766 we have outgoing changesets but refused to push
1764 - other values as described by addchangegroup()
1767 - other values as described by addchangegroup()
1765 '''
1768 '''
1766 # there are two ways to push to remote repo:
1769 # there are two ways to push to remote repo:
1767 #
1770 #
1768 # addchangegroup assumes local user can lock remote
1771 # addchangegroup assumes local user can lock remote
1769 # repo (local filesystem, old ssh servers).
1772 # repo (local filesystem, old ssh servers).
1770 #
1773 #
1771 # unbundle assumes local user cannot lock remote repo (new ssh
1774 # unbundle assumes local user cannot lock remote repo (new ssh
1772 # servers, http servers).
1775 # servers, http servers).
1773
1776
1774 if not remote.canpush():
1777 if not remote.canpush():
1775 raise util.Abort(_("destination does not support push"))
1778 raise util.Abort(_("destination does not support push"))
1776 unfi = self.unfiltered()
1779 unfi = self.unfiltered()
1777 # get local lock as we might write phase data
1780 # get local lock as we might write phase data
1778 locallock = self.lock()
1781 locallock = self.lock()
1779 try:
1782 try:
1780 self.checkpush(force, revs)
1783 self.checkpush(force, revs)
1781 lock = None
1784 lock = None
1782 unbundle = remote.capable('unbundle')
1785 unbundle = remote.capable('unbundle')
1783 if not unbundle:
1786 if not unbundle:
1784 lock = remote.lock()
1787 lock = remote.lock()
1785 try:
1788 try:
1786 # discovery
1789 # discovery
1787 fci = discovery.findcommonincoming
1790 fci = discovery.findcommonincoming
1788 commoninc = fci(unfi, remote, force=force)
1791 commoninc = fci(unfi, remote, force=force)
1789 common, inc, remoteheads = commoninc
1792 common, inc, remoteheads = commoninc
1790 fco = discovery.findcommonoutgoing
1793 fco = discovery.findcommonoutgoing
1791 outgoing = fco(unfi, remote, onlyheads=revs,
1794 outgoing = fco(unfi, remote, onlyheads=revs,
1792 commoninc=commoninc, force=force)
1795 commoninc=commoninc, force=force)
1793
1796
1794
1797
1795 if not outgoing.missing:
1798 if not outgoing.missing:
1796 # nothing to push
1799 # nothing to push
1797 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1800 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1798 ret = None
1801 ret = None
1799 else:
1802 else:
1800 # something to push
1803 # something to push
1801 if not force:
1804 if not force:
1802 # if self.obsstore == False --> no obsolete
1805 # if self.obsstore == False --> no obsolete
1803 # then, save the iteration
1806 # then, save the iteration
1804 if unfi.obsstore:
1807 if unfi.obsstore:
1805 # this message are here for 80 char limit reason
1808 # this message are here for 80 char limit reason
1806 mso = _("push includes obsolete changeset: %s!")
1809 mso = _("push includes obsolete changeset: %s!")
1807 mst = "push includes %s changeset: %s!"
1810 mst = "push includes %s changeset: %s!"
1808 # plain versions for i18n tool to detect them
1811 # plain versions for i18n tool to detect them
1809 _("push includes unstable changeset: %s!")
1812 _("push includes unstable changeset: %s!")
1810 _("push includes bumped changeset: %s!")
1813 _("push includes bumped changeset: %s!")
1811 _("push includes divergent changeset: %s!")
1814 _("push includes divergent changeset: %s!")
1812 # If we are to push if there is at least one
1815 # If we are to push if there is at least one
1813 # obsolete or unstable changeset in missing, at
1816 # obsolete or unstable changeset in missing, at
1814 # least one of the missinghead will be obsolete or
1817 # least one of the missinghead will be obsolete or
1815 # unstable. So checking heads only is ok
1818 # unstable. So checking heads only is ok
1816 for node in outgoing.missingheads:
1819 for node in outgoing.missingheads:
1817 ctx = unfi[node]
1820 ctx = unfi[node]
1818 if ctx.obsolete():
1821 if ctx.obsolete():
1819 raise util.Abort(mso % ctx)
1822 raise util.Abort(mso % ctx)
1820 elif ctx.troubled():
1823 elif ctx.troubled():
1821 raise util.Abort(_(mst)
1824 raise util.Abort(_(mst)
1822 % (ctx.troubles()[0],
1825 % (ctx.troubles()[0],
1823 ctx))
1826 ctx))
1824 discovery.checkheads(unfi, remote, outgoing,
1827 discovery.checkheads(unfi, remote, outgoing,
1825 remoteheads, newbranch,
1828 remoteheads, newbranch,
1826 bool(inc))
1829 bool(inc))
1827
1830
1828 # create a changegroup from local
1831 # create a changegroup from local
1829 if revs is None and not outgoing.excluded:
1832 if revs is None and not outgoing.excluded:
1830 # push everything,
1833 # push everything,
1831 # use the fast path, no race possible on push
1834 # use the fast path, no race possible on push
1832 cg = self._changegroup(outgoing.missing, 'push')
1835 cg = self._changegroup(outgoing.missing, 'push')
1833 else:
1836 else:
1834 cg = self.getlocalbundle('push', outgoing)
1837 cg = self.getlocalbundle('push', outgoing)
1835
1838
1836 # apply changegroup to remote
1839 # apply changegroup to remote
1837 if unbundle:
1840 if unbundle:
1838 # local repo finds heads on server, finds out what
1841 # local repo finds heads on server, finds out what
1839 # revs it must push. once revs transferred, if server
1842 # revs it must push. once revs transferred, if server
1840 # finds it has different heads (someone else won
1843 # finds it has different heads (someone else won
1841 # commit/push race), server aborts.
1844 # commit/push race), server aborts.
1842 if force:
1845 if force:
1843 remoteheads = ['force']
1846 remoteheads = ['force']
1844 # ssh: return remote's addchangegroup()
1847 # ssh: return remote's addchangegroup()
1845 # http: return remote's addchangegroup() or 0 for error
1848 # http: return remote's addchangegroup() or 0 for error
1846 ret = remote.unbundle(cg, remoteheads, 'push')
1849 ret = remote.unbundle(cg, remoteheads, 'push')
1847 else:
1850 else:
1848 # we return an integer indicating remote head count
1851 # we return an integer indicating remote head count
1849 # change
1852 # change
1850 ret = remote.addchangegroup(cg, 'push', self.url())
1853 ret = remote.addchangegroup(cg, 'push', self.url())
1851
1854
1852 if ret:
1855 if ret:
1853 # push succeed, synchronize target of the push
1856 # push succeed, synchronize target of the push
1854 cheads = outgoing.missingheads
1857 cheads = outgoing.missingheads
1855 elif revs is None:
1858 elif revs is None:
1856 # All out push fails. synchronize all common
1859 # All out push fails. synchronize all common
1857 cheads = outgoing.commonheads
1860 cheads = outgoing.commonheads
1858 else:
1861 else:
1859 # I want cheads = heads(::missingheads and ::commonheads)
1862 # I want cheads = heads(::missingheads and ::commonheads)
1860 # (missingheads is revs with secret changeset filtered out)
1863 # (missingheads is revs with secret changeset filtered out)
1861 #
1864 #
1862 # This can be expressed as:
1865 # This can be expressed as:
1863 # cheads = ( (missingheads and ::commonheads)
1866 # cheads = ( (missingheads and ::commonheads)
1864 # + (commonheads and ::missingheads))"
1867 # + (commonheads and ::missingheads))"
1865 # )
1868 # )
1866 #
1869 #
1867 # while trying to push we already computed the following:
1870 # while trying to push we already computed the following:
1868 # common = (::commonheads)
1871 # common = (::commonheads)
1869 # missing = ((commonheads::missingheads) - commonheads)
1872 # missing = ((commonheads::missingheads) - commonheads)
1870 #
1873 #
1871 # We can pick:
1874 # We can pick:
1872 # * missingheads part of common (::commonheads)
1875 # * missingheads part of common (::commonheads)
1873 common = set(outgoing.common)
1876 common = set(outgoing.common)
1874 cheads = [node for node in revs if node in common]
1877 cheads = [node for node in revs if node in common]
1875 # and
1878 # and
1876 # * commonheads parents on missing
1879 # * commonheads parents on missing
1877 revset = unfi.set('%ln and parents(roots(%ln))',
1880 revset = unfi.set('%ln and parents(roots(%ln))',
1878 outgoing.commonheads,
1881 outgoing.commonheads,
1879 outgoing.missing)
1882 outgoing.missing)
1880 cheads.extend(c.node() for c in revset)
1883 cheads.extend(c.node() for c in revset)
1881 # even when we don't push, exchanging phase data is useful
1884 # even when we don't push, exchanging phase data is useful
1882 remotephases = remote.listkeys('phases')
1885 remotephases = remote.listkeys('phases')
1883 if not remotephases: # old server or public only repo
1886 if not remotephases: # old server or public only repo
1884 phases.advanceboundary(self, phases.public, cheads)
1887 phases.advanceboundary(self, phases.public, cheads)
1885 # don't push any phase data as there is nothing to push
1888 # don't push any phase data as there is nothing to push
1886 else:
1889 else:
1887 ana = phases.analyzeremotephases(self, cheads, remotephases)
1890 ana = phases.analyzeremotephases(self, cheads, remotephases)
1888 pheads, droots = ana
1891 pheads, droots = ana
1889 ### Apply remote phase on local
1892 ### Apply remote phase on local
1890 if remotephases.get('publishing', False):
1893 if remotephases.get('publishing', False):
1891 phases.advanceboundary(self, phases.public, cheads)
1894 phases.advanceboundary(self, phases.public, cheads)
1892 else: # publish = False
1895 else: # publish = False
1893 phases.advanceboundary(self, phases.public, pheads)
1896 phases.advanceboundary(self, phases.public, pheads)
1894 phases.advanceboundary(self, phases.draft, cheads)
1897 phases.advanceboundary(self, phases.draft, cheads)
1895 ### Apply local phase on remote
1898 ### Apply local phase on remote
1896
1899
1897 # Get the list of all revs draft on remote by public here.
1900 # Get the list of all revs draft on remote by public here.
1898 # XXX Beware that revset break if droots is not strictly
1901 # XXX Beware that revset break if droots is not strictly
1899 # XXX root we may want to ensure it is but it is costly
1902 # XXX root we may want to ensure it is but it is costly
1900 outdated = unfi.set('heads((%ln::%ln) and public())',
1903 outdated = unfi.set('heads((%ln::%ln) and public())',
1901 droots, cheads)
1904 droots, cheads)
1902 for newremotehead in outdated:
1905 for newremotehead in outdated:
1903 r = remote.pushkey('phases',
1906 r = remote.pushkey('phases',
1904 newremotehead.hex(),
1907 newremotehead.hex(),
1905 str(phases.draft),
1908 str(phases.draft),
1906 str(phases.public))
1909 str(phases.public))
1907 if not r:
1910 if not r:
1908 self.ui.warn(_('updating %s to public failed!\n')
1911 self.ui.warn(_('updating %s to public failed!\n')
1909 % newremotehead)
1912 % newremotehead)
1910 self.ui.debug('try to push obsolete markers to remote\n')
1913 self.ui.debug('try to push obsolete markers to remote\n')
1911 if (obsolete._enabled and self.obsstore and
1914 if (obsolete._enabled and self.obsstore and
1912 'obsolete' in remote.listkeys('namespaces')):
1915 'obsolete' in remote.listkeys('namespaces')):
1913 rslts = []
1916 rslts = []
1914 remotedata = self.listkeys('obsolete')
1917 remotedata = self.listkeys('obsolete')
1915 for key in sorted(remotedata, reverse=True):
1918 for key in sorted(remotedata, reverse=True):
1916 # reverse sort to ensure we end with dump0
1919 # reverse sort to ensure we end with dump0
1917 data = remotedata[key]
1920 data = remotedata[key]
1918 rslts.append(remote.pushkey('obsolete', key, '', data))
1921 rslts.append(remote.pushkey('obsolete', key, '', data))
1919 if [r for r in rslts if not r]:
1922 if [r for r in rslts if not r]:
1920 msg = _('failed to push some obsolete markers!\n')
1923 msg = _('failed to push some obsolete markers!\n')
1921 self.ui.warn(msg)
1924 self.ui.warn(msg)
1922 finally:
1925 finally:
1923 if lock is not None:
1926 if lock is not None:
1924 lock.release()
1927 lock.release()
1925 finally:
1928 finally:
1926 locallock.release()
1929 locallock.release()
1927
1930
1928 self.ui.debug("checking for updated bookmarks\n")
1931 self.ui.debug("checking for updated bookmarks\n")
1929 rb = remote.listkeys('bookmarks')
1932 rb = remote.listkeys('bookmarks')
1930 for k in rb.keys():
1933 for k in rb.keys():
1931 if k in unfi._bookmarks:
1934 if k in unfi._bookmarks:
1932 nr, nl = rb[k], hex(self._bookmarks[k])
1935 nr, nl = rb[k], hex(self._bookmarks[k])
1933 if nr in unfi:
1936 if nr in unfi:
1934 cr = unfi[nr]
1937 cr = unfi[nr]
1935 cl = unfi[nl]
1938 cl = unfi[nl]
1936 if bookmarks.validdest(unfi, cr, cl):
1939 if bookmarks.validdest(unfi, cr, cl):
1937 r = remote.pushkey('bookmarks', k, nr, nl)
1940 r = remote.pushkey('bookmarks', k, nr, nl)
1938 if r:
1941 if r:
1939 self.ui.status(_("updating bookmark %s\n") % k)
1942 self.ui.status(_("updating bookmark %s\n") % k)
1940 else:
1943 else:
1941 self.ui.warn(_('updating bookmark %s'
1944 self.ui.warn(_('updating bookmark %s'
1942 ' failed!\n') % k)
1945 ' failed!\n') % k)
1943
1946
1944 return ret
1947 return ret
1945
1948
1946 def changegroupinfo(self, nodes, source):
1949 def changegroupinfo(self, nodes, source):
1947 if self.ui.verbose or source == 'bundle':
1950 if self.ui.verbose or source == 'bundle':
1948 self.ui.status(_("%d changesets found\n") % len(nodes))
1951 self.ui.status(_("%d changesets found\n") % len(nodes))
1949 if self.ui.debugflag:
1952 if self.ui.debugflag:
1950 self.ui.debug("list of changesets:\n")
1953 self.ui.debug("list of changesets:\n")
1951 for node in nodes:
1954 for node in nodes:
1952 self.ui.debug("%s\n" % hex(node))
1955 self.ui.debug("%s\n" % hex(node))
1953
1956
1954 def changegroupsubset(self, bases, heads, source):
1957 def changegroupsubset(self, bases, heads, source):
1955 """Compute a changegroup consisting of all the nodes that are
1958 """Compute a changegroup consisting of all the nodes that are
1956 descendants of any of the bases and ancestors of any of the heads.
1959 descendants of any of the bases and ancestors of any of the heads.
1957 Return a chunkbuffer object whose read() method will return
1960 Return a chunkbuffer object whose read() method will return
1958 successive changegroup chunks.
1961 successive changegroup chunks.
1959
1962
1960 It is fairly complex as determining which filenodes and which
1963 It is fairly complex as determining which filenodes and which
1961 manifest nodes need to be included for the changeset to be complete
1964 manifest nodes need to be included for the changeset to be complete
1962 is non-trivial.
1965 is non-trivial.
1963
1966
1964 Another wrinkle is doing the reverse, figuring out which changeset in
1967 Another wrinkle is doing the reverse, figuring out which changeset in
1965 the changegroup a particular filenode or manifestnode belongs to.
1968 the changegroup a particular filenode or manifestnode belongs to.
1966 """
1969 """
1967 cl = self.changelog
1970 cl = self.changelog
1968 if not bases:
1971 if not bases:
1969 bases = [nullid]
1972 bases = [nullid]
1970 csets, bases, heads = cl.nodesbetween(bases, heads)
1973 csets, bases, heads = cl.nodesbetween(bases, heads)
1971 # We assume that all ancestors of bases are known
1974 # We assume that all ancestors of bases are known
1972 common = cl.ancestors([cl.rev(n) for n in bases])
1975 common = cl.ancestors([cl.rev(n) for n in bases])
1973 return self._changegroupsubset(common, csets, heads, source)
1976 return self._changegroupsubset(common, csets, heads, source)
1974
1977
1975 def getlocalbundle(self, source, outgoing):
1978 def getlocalbundle(self, source, outgoing):
1976 """Like getbundle, but taking a discovery.outgoing as an argument.
1979 """Like getbundle, but taking a discovery.outgoing as an argument.
1977
1980
1978 This is only implemented for local repos and reuses potentially
1981 This is only implemented for local repos and reuses potentially
1979 precomputed sets in outgoing."""
1982 precomputed sets in outgoing."""
1980 if not outgoing.missing:
1983 if not outgoing.missing:
1981 return None
1984 return None
1982 return self._changegroupsubset(outgoing.common,
1985 return self._changegroupsubset(outgoing.common,
1983 outgoing.missing,
1986 outgoing.missing,
1984 outgoing.missingheads,
1987 outgoing.missingheads,
1985 source)
1988 source)
1986
1989
1987 def getbundle(self, source, heads=None, common=None):
1990 def getbundle(self, source, heads=None, common=None):
1988 """Like changegroupsubset, but returns the set difference between the
1991 """Like changegroupsubset, but returns the set difference between the
1989 ancestors of heads and the ancestors common.
1992 ancestors of heads and the ancestors common.
1990
1993
1991 If heads is None, use the local heads. If common is None, use [nullid].
1994 If heads is None, use the local heads. If common is None, use [nullid].
1992
1995
1993 The nodes in common might not all be known locally due to the way the
1996 The nodes in common might not all be known locally due to the way the
1994 current discovery protocol works.
1997 current discovery protocol works.
1995 """
1998 """
1996 cl = self.changelog
1999 cl = self.changelog
1997 if common:
2000 if common:
1998 hasnode = cl.hasnode
2001 hasnode = cl.hasnode
1999 common = [n for n in common if hasnode(n)]
2002 common = [n for n in common if hasnode(n)]
2000 else:
2003 else:
2001 common = [nullid]
2004 common = [nullid]
2002 if not heads:
2005 if not heads:
2003 heads = cl.heads()
2006 heads = cl.heads()
2004 return self.getlocalbundle(source,
2007 return self.getlocalbundle(source,
2005 discovery.outgoing(cl, common, heads))
2008 discovery.outgoing(cl, common, heads))
2006
2009
2007 @unfilteredmethod
2010 @unfilteredmethod
2008 def _changegroupsubset(self, commonrevs, csets, heads, source):
2011 def _changegroupsubset(self, commonrevs, csets, heads, source):
2009
2012
2010 cl = self.changelog
2013 cl = self.changelog
2011 mf = self.manifest
2014 mf = self.manifest
2012 mfs = {} # needed manifests
2015 mfs = {} # needed manifests
2013 fnodes = {} # needed file nodes
2016 fnodes = {} # needed file nodes
2014 changedfiles = set()
2017 changedfiles = set()
2015 fstate = ['', {}]
2018 fstate = ['', {}]
2016 count = [0, 0]
2019 count = [0, 0]
2017
2020
2018 # can we go through the fast path ?
2021 # can we go through the fast path ?
2019 heads.sort()
2022 heads.sort()
2020 if heads == sorted(self.heads()):
2023 if heads == sorted(self.heads()):
2021 return self._changegroup(csets, source)
2024 return self._changegroup(csets, source)
2022
2025
2023 # slow path
2026 # slow path
2024 self.hook('preoutgoing', throw=True, source=source)
2027 self.hook('preoutgoing', throw=True, source=source)
2025 self.changegroupinfo(csets, source)
2028 self.changegroupinfo(csets, source)
2026
2029
2027 # filter any nodes that claim to be part of the known set
2030 # filter any nodes that claim to be part of the known set
2028 def prune(revlog, missing):
2031 def prune(revlog, missing):
2029 rr, rl = revlog.rev, revlog.linkrev
2032 rr, rl = revlog.rev, revlog.linkrev
2030 return [n for n in missing
2033 return [n for n in missing
2031 if rl(rr(n)) not in commonrevs]
2034 if rl(rr(n)) not in commonrevs]
2032
2035
2033 progress = self.ui.progress
2036 progress = self.ui.progress
2034 _bundling = _('bundling')
2037 _bundling = _('bundling')
2035 _changesets = _('changesets')
2038 _changesets = _('changesets')
2036 _manifests = _('manifests')
2039 _manifests = _('manifests')
2037 _files = _('files')
2040 _files = _('files')
2038
2041
2039 def lookup(revlog, x):
2042 def lookup(revlog, x):
2040 if revlog == cl:
2043 if revlog == cl:
2041 c = cl.read(x)
2044 c = cl.read(x)
2042 changedfiles.update(c[3])
2045 changedfiles.update(c[3])
2043 mfs.setdefault(c[0], x)
2046 mfs.setdefault(c[0], x)
2044 count[0] += 1
2047 count[0] += 1
2045 progress(_bundling, count[0],
2048 progress(_bundling, count[0],
2046 unit=_changesets, total=count[1])
2049 unit=_changesets, total=count[1])
2047 return x
2050 return x
2048 elif revlog == mf:
2051 elif revlog == mf:
2049 clnode = mfs[x]
2052 clnode = mfs[x]
2050 mdata = mf.readfast(x)
2053 mdata = mf.readfast(x)
2051 for f, n in mdata.iteritems():
2054 for f, n in mdata.iteritems():
2052 if f in changedfiles:
2055 if f in changedfiles:
2053 fnodes[f].setdefault(n, clnode)
2056 fnodes[f].setdefault(n, clnode)
2054 count[0] += 1
2057 count[0] += 1
2055 progress(_bundling, count[0],
2058 progress(_bundling, count[0],
2056 unit=_manifests, total=count[1])
2059 unit=_manifests, total=count[1])
2057 return clnode
2060 return clnode
2058 else:
2061 else:
2059 progress(_bundling, count[0], item=fstate[0],
2062 progress(_bundling, count[0], item=fstate[0],
2060 unit=_files, total=count[1])
2063 unit=_files, total=count[1])
2061 return fstate[1][x]
2064 return fstate[1][x]
2062
2065
2063 bundler = changegroup.bundle10(lookup)
2066 bundler = changegroup.bundle10(lookup)
2064 reorder = self.ui.config('bundle', 'reorder', 'auto')
2067 reorder = self.ui.config('bundle', 'reorder', 'auto')
2065 if reorder == 'auto':
2068 if reorder == 'auto':
2066 reorder = None
2069 reorder = None
2067 else:
2070 else:
2068 reorder = util.parsebool(reorder)
2071 reorder = util.parsebool(reorder)
2069
2072
2070 def gengroup():
2073 def gengroup():
2071 # Create a changenode group generator that will call our functions
2074 # Create a changenode group generator that will call our functions
2072 # back to lookup the owning changenode and collect information.
2075 # back to lookup the owning changenode and collect information.
2073 count[:] = [0, len(csets)]
2076 count[:] = [0, len(csets)]
2074 for chunk in cl.group(csets, bundler, reorder=reorder):
2077 for chunk in cl.group(csets, bundler, reorder=reorder):
2075 yield chunk
2078 yield chunk
2076 progress(_bundling, None)
2079 progress(_bundling, None)
2077
2080
2078 # Create a generator for the manifestnodes that calls our lookup
2081 # Create a generator for the manifestnodes that calls our lookup
2079 # and data collection functions back.
2082 # and data collection functions back.
2080 for f in changedfiles:
2083 for f in changedfiles:
2081 fnodes[f] = {}
2084 fnodes[f] = {}
2082 count[:] = [0, len(mfs)]
2085 count[:] = [0, len(mfs)]
2083 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2086 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2084 yield chunk
2087 yield chunk
2085 progress(_bundling, None)
2088 progress(_bundling, None)
2086
2089
2087 mfs.clear()
2090 mfs.clear()
2088
2091
2089 # Go through all our files in order sorted by name.
2092 # Go through all our files in order sorted by name.
2090 count[:] = [0, len(changedfiles)]
2093 count[:] = [0, len(changedfiles)]
2091 for fname in sorted(changedfiles):
2094 for fname in sorted(changedfiles):
2092 filerevlog = self.file(fname)
2095 filerevlog = self.file(fname)
2093 if not len(filerevlog):
2096 if not len(filerevlog):
2094 raise util.Abort(_("empty or missing revlog for %s")
2097 raise util.Abort(_("empty or missing revlog for %s")
2095 % fname)
2098 % fname)
2096 fstate[0] = fname
2099 fstate[0] = fname
2097 fstate[1] = fnodes.pop(fname, {})
2100 fstate[1] = fnodes.pop(fname, {})
2098
2101
2099 nodelist = prune(filerevlog, fstate[1])
2102 nodelist = prune(filerevlog, fstate[1])
2100 if nodelist:
2103 if nodelist:
2101 count[0] += 1
2104 count[0] += 1
2102 yield bundler.fileheader(fname)
2105 yield bundler.fileheader(fname)
2103 for chunk in filerevlog.group(nodelist, bundler, reorder):
2106 for chunk in filerevlog.group(nodelist, bundler, reorder):
2104 yield chunk
2107 yield chunk
2105
2108
2106 # Signal that no more groups are left.
2109 # Signal that no more groups are left.
2107 yield bundler.close()
2110 yield bundler.close()
2108 progress(_bundling, None)
2111 progress(_bundling, None)
2109
2112
2110 if csets:
2113 if csets:
2111 self.hook('outgoing', node=hex(csets[0]), source=source)
2114 self.hook('outgoing', node=hex(csets[0]), source=source)
2112
2115
2113 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2116 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2114
2117
2115 def changegroup(self, basenodes, source):
2118 def changegroup(self, basenodes, source):
2116 # to avoid a race we use changegroupsubset() (issue1320)
2119 # to avoid a race we use changegroupsubset() (issue1320)
2117 return self.changegroupsubset(basenodes, self.heads(), source)
2120 return self.changegroupsubset(basenodes, self.heads(), source)
2118
2121
2119 @unfilteredmethod
2122 @unfilteredmethod
2120 def _changegroup(self, nodes, source):
2123 def _changegroup(self, nodes, source):
2121 """Compute the changegroup of all nodes that we have that a recipient
2124 """Compute the changegroup of all nodes that we have that a recipient
2122 doesn't. Return a chunkbuffer object whose read() method will return
2125 doesn't. Return a chunkbuffer object whose read() method will return
2123 successive changegroup chunks.
2126 successive changegroup chunks.
2124
2127
2125 This is much easier than the previous function as we can assume that
2128 This is much easier than the previous function as we can assume that
2126 the recipient has any changenode we aren't sending them.
2129 the recipient has any changenode we aren't sending them.
2127
2130
2128 nodes is the set of nodes to send"""
2131 nodes is the set of nodes to send"""
2129
2132
2130 cl = self.changelog
2133 cl = self.changelog
2131 mf = self.manifest
2134 mf = self.manifest
2132 mfs = {}
2135 mfs = {}
2133 changedfiles = set()
2136 changedfiles = set()
2134 fstate = ['']
2137 fstate = ['']
2135 count = [0, 0]
2138 count = [0, 0]
2136
2139
2137 self.hook('preoutgoing', throw=True, source=source)
2140 self.hook('preoutgoing', throw=True, source=source)
2138 self.changegroupinfo(nodes, source)
2141 self.changegroupinfo(nodes, source)
2139
2142
2140 revset = set([cl.rev(n) for n in nodes])
2143 revset = set([cl.rev(n) for n in nodes])
2141
2144
2142 def gennodelst(log):
2145 def gennodelst(log):
2143 ln, llr = log.node, log.linkrev
2146 ln, llr = log.node, log.linkrev
2144 return [ln(r) for r in log if llr(r) in revset]
2147 return [ln(r) for r in log if llr(r) in revset]
2145
2148
2146 progress = self.ui.progress
2149 progress = self.ui.progress
2147 _bundling = _('bundling')
2150 _bundling = _('bundling')
2148 _changesets = _('changesets')
2151 _changesets = _('changesets')
2149 _manifests = _('manifests')
2152 _manifests = _('manifests')
2150 _files = _('files')
2153 _files = _('files')
2151
2154
2152 def lookup(revlog, x):
2155 def lookup(revlog, x):
2153 if revlog == cl:
2156 if revlog == cl:
2154 c = cl.read(x)
2157 c = cl.read(x)
2155 changedfiles.update(c[3])
2158 changedfiles.update(c[3])
2156 mfs.setdefault(c[0], x)
2159 mfs.setdefault(c[0], x)
2157 count[0] += 1
2160 count[0] += 1
2158 progress(_bundling, count[0],
2161 progress(_bundling, count[0],
2159 unit=_changesets, total=count[1])
2162 unit=_changesets, total=count[1])
2160 return x
2163 return x
2161 elif revlog == mf:
2164 elif revlog == mf:
2162 count[0] += 1
2165 count[0] += 1
2163 progress(_bundling, count[0],
2166 progress(_bundling, count[0],
2164 unit=_manifests, total=count[1])
2167 unit=_manifests, total=count[1])
2165 return cl.node(revlog.linkrev(revlog.rev(x)))
2168 return cl.node(revlog.linkrev(revlog.rev(x)))
2166 else:
2169 else:
2167 progress(_bundling, count[0], item=fstate[0],
2170 progress(_bundling, count[0], item=fstate[0],
2168 total=count[1], unit=_files)
2171 total=count[1], unit=_files)
2169 return cl.node(revlog.linkrev(revlog.rev(x)))
2172 return cl.node(revlog.linkrev(revlog.rev(x)))
2170
2173
2171 bundler = changegroup.bundle10(lookup)
2174 bundler = changegroup.bundle10(lookup)
2172 reorder = self.ui.config('bundle', 'reorder', 'auto')
2175 reorder = self.ui.config('bundle', 'reorder', 'auto')
2173 if reorder == 'auto':
2176 if reorder == 'auto':
2174 reorder = None
2177 reorder = None
2175 else:
2178 else:
2176 reorder = util.parsebool(reorder)
2179 reorder = util.parsebool(reorder)
2177
2180
2178 def gengroup():
2181 def gengroup():
2179 '''yield a sequence of changegroup chunks (strings)'''
2182 '''yield a sequence of changegroup chunks (strings)'''
2180 # construct a list of all changed files
2183 # construct a list of all changed files
2181
2184
2182 count[:] = [0, len(nodes)]
2185 count[:] = [0, len(nodes)]
2183 for chunk in cl.group(nodes, bundler, reorder=reorder):
2186 for chunk in cl.group(nodes, bundler, reorder=reorder):
2184 yield chunk
2187 yield chunk
2185 progress(_bundling, None)
2188 progress(_bundling, None)
2186
2189
2187 count[:] = [0, len(mfs)]
2190 count[:] = [0, len(mfs)]
2188 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2191 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2189 yield chunk
2192 yield chunk
2190 progress(_bundling, None)
2193 progress(_bundling, None)
2191
2194
2192 count[:] = [0, len(changedfiles)]
2195 count[:] = [0, len(changedfiles)]
2193 for fname in sorted(changedfiles):
2196 for fname in sorted(changedfiles):
2194 filerevlog = self.file(fname)
2197 filerevlog = self.file(fname)
2195 if not len(filerevlog):
2198 if not len(filerevlog):
2196 raise util.Abort(_("empty or missing revlog for %s")
2199 raise util.Abort(_("empty or missing revlog for %s")
2197 % fname)
2200 % fname)
2198 fstate[0] = fname
2201 fstate[0] = fname
2199 nodelist = gennodelst(filerevlog)
2202 nodelist = gennodelst(filerevlog)
2200 if nodelist:
2203 if nodelist:
2201 count[0] += 1
2204 count[0] += 1
2202 yield bundler.fileheader(fname)
2205 yield bundler.fileheader(fname)
2203 for chunk in filerevlog.group(nodelist, bundler, reorder):
2206 for chunk in filerevlog.group(nodelist, bundler, reorder):
2204 yield chunk
2207 yield chunk
2205 yield bundler.close()
2208 yield bundler.close()
2206 progress(_bundling, None)
2209 progress(_bundling, None)
2207
2210
2208 if nodes:
2211 if nodes:
2209 self.hook('outgoing', node=hex(nodes[0]), source=source)
2212 self.hook('outgoing', node=hex(nodes[0]), source=source)
2210
2213
2211 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2214 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2212
2215
2213 @unfilteredmethod
2216 @unfilteredmethod
2214 def addchangegroup(self, source, srctype, url, emptyok=False):
2217 def addchangegroup(self, source, srctype, url, emptyok=False):
2215 """Add the changegroup returned by source.read() to this repo.
2218 """Add the changegroup returned by source.read() to this repo.
2216 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2219 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2217 the URL of the repo where this changegroup is coming from.
2220 the URL of the repo where this changegroup is coming from.
2218
2221
2219 Return an integer summarizing the change to this repo:
2222 Return an integer summarizing the change to this repo:
2220 - nothing changed or no source: 0
2223 - nothing changed or no source: 0
2221 - more heads than before: 1+added heads (2..n)
2224 - more heads than before: 1+added heads (2..n)
2222 - fewer heads than before: -1-removed heads (-2..-n)
2225 - fewer heads than before: -1-removed heads (-2..-n)
2223 - number of heads stays the same: 1
2226 - number of heads stays the same: 1
2224 """
2227 """
2225 def csmap(x):
2228 def csmap(x):
2226 self.ui.debug("add changeset %s\n" % short(x))
2229 self.ui.debug("add changeset %s\n" % short(x))
2227 return len(cl)
2230 return len(cl)
2228
2231
2229 def revmap(x):
2232 def revmap(x):
2230 return cl.rev(x)
2233 return cl.rev(x)
2231
2234
2232 if not source:
2235 if not source:
2233 return 0
2236 return 0
2234
2237
2235 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2238 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2236
2239
2237 changesets = files = revisions = 0
2240 changesets = files = revisions = 0
2238 efiles = set()
2241 efiles = set()
2239
2242
2240 # write changelog data to temp files so concurrent readers will not see
2243 # write changelog data to temp files so concurrent readers will not see
2241 # inconsistent view
2244 # inconsistent view
2242 cl = self.changelog
2245 cl = self.changelog
2243 cl.delayupdate()
2246 cl.delayupdate()
2244 oldheads = cl.heads()
2247 oldheads = cl.heads()
2245
2248
2246 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2249 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2247 try:
2250 try:
2248 trp = weakref.proxy(tr)
2251 trp = weakref.proxy(tr)
2249 # pull off the changeset group
2252 # pull off the changeset group
2250 self.ui.status(_("adding changesets\n"))
2253 self.ui.status(_("adding changesets\n"))
2251 clstart = len(cl)
2254 clstart = len(cl)
2252 class prog(object):
2255 class prog(object):
2253 step = _('changesets')
2256 step = _('changesets')
2254 count = 1
2257 count = 1
2255 ui = self.ui
2258 ui = self.ui
2256 total = None
2259 total = None
2257 def __call__(self):
2260 def __call__(self):
2258 self.ui.progress(self.step, self.count, unit=_('chunks'),
2261 self.ui.progress(self.step, self.count, unit=_('chunks'),
2259 total=self.total)
2262 total=self.total)
2260 self.count += 1
2263 self.count += 1
2261 pr = prog()
2264 pr = prog()
2262 source.callback = pr
2265 source.callback = pr
2263
2266
2264 source.changelogheader()
2267 source.changelogheader()
2265 srccontent = cl.addgroup(source, csmap, trp)
2268 srccontent = cl.addgroup(source, csmap, trp)
2266 if not (srccontent or emptyok):
2269 if not (srccontent or emptyok):
2267 raise util.Abort(_("received changelog group is empty"))
2270 raise util.Abort(_("received changelog group is empty"))
2268 clend = len(cl)
2271 clend = len(cl)
2269 changesets = clend - clstart
2272 changesets = clend - clstart
2270 for c in xrange(clstart, clend):
2273 for c in xrange(clstart, clend):
2271 efiles.update(self[c].files())
2274 efiles.update(self[c].files())
2272 efiles = len(efiles)
2275 efiles = len(efiles)
2273 self.ui.progress(_('changesets'), None)
2276 self.ui.progress(_('changesets'), None)
2274
2277
2275 # pull off the manifest group
2278 # pull off the manifest group
2276 self.ui.status(_("adding manifests\n"))
2279 self.ui.status(_("adding manifests\n"))
2277 pr.step = _('manifests')
2280 pr.step = _('manifests')
2278 pr.count = 1
2281 pr.count = 1
2279 pr.total = changesets # manifests <= changesets
2282 pr.total = changesets # manifests <= changesets
2280 # no need to check for empty manifest group here:
2283 # no need to check for empty manifest group here:
2281 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2284 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2282 # no new manifest will be created and the manifest group will
2285 # no new manifest will be created and the manifest group will
2283 # be empty during the pull
2286 # be empty during the pull
2284 source.manifestheader()
2287 source.manifestheader()
2285 self.manifest.addgroup(source, revmap, trp)
2288 self.manifest.addgroup(source, revmap, trp)
2286 self.ui.progress(_('manifests'), None)
2289 self.ui.progress(_('manifests'), None)
2287
2290
2288 needfiles = {}
2291 needfiles = {}
2289 if self.ui.configbool('server', 'validate', default=False):
2292 if self.ui.configbool('server', 'validate', default=False):
2290 # validate incoming csets have their manifests
2293 # validate incoming csets have their manifests
2291 for cset in xrange(clstart, clend):
2294 for cset in xrange(clstart, clend):
2292 mfest = self.changelog.read(self.changelog.node(cset))[0]
2295 mfest = self.changelog.read(self.changelog.node(cset))[0]
2293 mfest = self.manifest.readdelta(mfest)
2296 mfest = self.manifest.readdelta(mfest)
2294 # store file nodes we must see
2297 # store file nodes we must see
2295 for f, n in mfest.iteritems():
2298 for f, n in mfest.iteritems():
2296 needfiles.setdefault(f, set()).add(n)
2299 needfiles.setdefault(f, set()).add(n)
2297
2300
2298 # process the files
2301 # process the files
2299 self.ui.status(_("adding file changes\n"))
2302 self.ui.status(_("adding file changes\n"))
2300 pr.step = _('files')
2303 pr.step = _('files')
2301 pr.count = 1
2304 pr.count = 1
2302 pr.total = efiles
2305 pr.total = efiles
2303 source.callback = None
2306 source.callback = None
2304
2307
2305 while True:
2308 while True:
2306 chunkdata = source.filelogheader()
2309 chunkdata = source.filelogheader()
2307 if not chunkdata:
2310 if not chunkdata:
2308 break
2311 break
2309 f = chunkdata["filename"]
2312 f = chunkdata["filename"]
2310 self.ui.debug("adding %s revisions\n" % f)
2313 self.ui.debug("adding %s revisions\n" % f)
2311 pr()
2314 pr()
2312 fl = self.file(f)
2315 fl = self.file(f)
2313 o = len(fl)
2316 o = len(fl)
2314 if not fl.addgroup(source, revmap, trp):
2317 if not fl.addgroup(source, revmap, trp):
2315 raise util.Abort(_("received file revlog group is empty"))
2318 raise util.Abort(_("received file revlog group is empty"))
2316 revisions += len(fl) - o
2319 revisions += len(fl) - o
2317 files += 1
2320 files += 1
2318 if f in needfiles:
2321 if f in needfiles:
2319 needs = needfiles[f]
2322 needs = needfiles[f]
2320 for new in xrange(o, len(fl)):
2323 for new in xrange(o, len(fl)):
2321 n = fl.node(new)
2324 n = fl.node(new)
2322 if n in needs:
2325 if n in needs:
2323 needs.remove(n)
2326 needs.remove(n)
2324 if not needs:
2327 if not needs:
2325 del needfiles[f]
2328 del needfiles[f]
2326 self.ui.progress(_('files'), None)
2329 self.ui.progress(_('files'), None)
2327
2330
2328 for f, needs in needfiles.iteritems():
2331 for f, needs in needfiles.iteritems():
2329 fl = self.file(f)
2332 fl = self.file(f)
2330 for n in needs:
2333 for n in needs:
2331 try:
2334 try:
2332 fl.rev(n)
2335 fl.rev(n)
2333 except error.LookupError:
2336 except error.LookupError:
2334 raise util.Abort(
2337 raise util.Abort(
2335 _('missing file data for %s:%s - run hg verify') %
2338 _('missing file data for %s:%s - run hg verify') %
2336 (f, hex(n)))
2339 (f, hex(n)))
2337
2340
2338 dh = 0
2341 dh = 0
2339 if oldheads:
2342 if oldheads:
2340 heads = cl.heads()
2343 heads = cl.heads()
2341 dh = len(heads) - len(oldheads)
2344 dh = len(heads) - len(oldheads)
2342 for h in heads:
2345 for h in heads:
2343 if h not in oldheads and self[h].closesbranch():
2346 if h not in oldheads and self[h].closesbranch():
2344 dh -= 1
2347 dh -= 1
2345 htext = ""
2348 htext = ""
2346 if dh:
2349 if dh:
2347 htext = _(" (%+d heads)") % dh
2350 htext = _(" (%+d heads)") % dh
2348
2351
2349 self.ui.status(_("added %d changesets"
2352 self.ui.status(_("added %d changesets"
2350 " with %d changes to %d files%s\n")
2353 " with %d changes to %d files%s\n")
2351 % (changesets, revisions, files, htext))
2354 % (changesets, revisions, files, htext))
2352 self.invalidatevolatilesets()
2355 self.invalidatevolatilesets()
2353
2356
2354 if changesets > 0:
2357 if changesets > 0:
2355 p = lambda: cl.writepending() and self.root or ""
2358 p = lambda: cl.writepending() and self.root or ""
2356 self.hook('pretxnchangegroup', throw=True,
2359 self.hook('pretxnchangegroup', throw=True,
2357 node=hex(cl.node(clstart)), source=srctype,
2360 node=hex(cl.node(clstart)), source=srctype,
2358 url=url, pending=p)
2361 url=url, pending=p)
2359
2362
2360 added = [cl.node(r) for r in xrange(clstart, clend)]
2363 added = [cl.node(r) for r in xrange(clstart, clend)]
2361 publishing = self.ui.configbool('phases', 'publish', True)
2364 publishing = self.ui.configbool('phases', 'publish', True)
2362 if srctype == 'push':
2365 if srctype == 'push':
2363 # Old server can not push the boundary themself.
2366 # Old server can not push the boundary themself.
2364 # New server won't push the boundary if changeset already
2367 # New server won't push the boundary if changeset already
2365 # existed locally as secrete
2368 # existed locally as secrete
2366 #
2369 #
2367 # We should not use added here but the list of all change in
2370 # We should not use added here but the list of all change in
2368 # the bundle
2371 # the bundle
2369 if publishing:
2372 if publishing:
2370 phases.advanceboundary(self, phases.public, srccontent)
2373 phases.advanceboundary(self, phases.public, srccontent)
2371 else:
2374 else:
2372 phases.advanceboundary(self, phases.draft, srccontent)
2375 phases.advanceboundary(self, phases.draft, srccontent)
2373 phases.retractboundary(self, phases.draft, added)
2376 phases.retractboundary(self, phases.draft, added)
2374 elif srctype != 'strip':
2377 elif srctype != 'strip':
2375 # publishing only alter behavior during push
2378 # publishing only alter behavior during push
2376 #
2379 #
2377 # strip should not touch boundary at all
2380 # strip should not touch boundary at all
2378 phases.retractboundary(self, phases.draft, added)
2381 phases.retractboundary(self, phases.draft, added)
2379
2382
2380 # make changelog see real files again
2383 # make changelog see real files again
2381 cl.finalize(trp)
2384 cl.finalize(trp)
2382
2385
2383 tr.close()
2386 tr.close()
2384
2387
2385 if changesets > 0:
2388 if changesets > 0:
2386 if srctype != 'strip':
2389 if srctype != 'strip':
2387 # During strip, branchcache is invalid but coming call to
2390 # During strip, branchcache is invalid but coming call to
2388 # `destroyed` will repair it.
2391 # `destroyed` will repair it.
2389 # In other case we can safely update cache on disk.
2392 # In other case we can safely update cache on disk.
2390 branchmap.updatecache(self)
2393 branchmap.updatecache(self)
2391 def runhooks():
2394 def runhooks():
2392 # forcefully update the on-disk branch cache
2395 # forcefully update the on-disk branch cache
2393 self.ui.debug("updating the branch cache\n")
2396 self.ui.debug("updating the branch cache\n")
2394 self.hook("changegroup", node=hex(cl.node(clstart)),
2397 self.hook("changegroup", node=hex(cl.node(clstart)),
2395 source=srctype, url=url)
2398 source=srctype, url=url)
2396
2399
2397 for n in added:
2400 for n in added:
2398 self.hook("incoming", node=hex(n), source=srctype,
2401 self.hook("incoming", node=hex(n), source=srctype,
2399 url=url)
2402 url=url)
2400 self._afterlock(runhooks)
2403 self._afterlock(runhooks)
2401
2404
2402 finally:
2405 finally:
2403 tr.release()
2406 tr.release()
2404 # never return 0 here:
2407 # never return 0 here:
2405 if dh < 0:
2408 if dh < 0:
2406 return dh - 1
2409 return dh - 1
2407 else:
2410 else:
2408 return dh + 1
2411 return dh + 1
2409
2412
2410 def stream_in(self, remote, requirements):
2413 def stream_in(self, remote, requirements):
2411 lock = self.lock()
2414 lock = self.lock()
2412 try:
2415 try:
2413 # Save remote branchmap. We will use it later
2416 # Save remote branchmap. We will use it later
2414 # to speed up branchcache creation
2417 # to speed up branchcache creation
2415 rbranchmap = None
2418 rbranchmap = None
2416 if remote.capable("branchmap"):
2419 if remote.capable("branchmap"):
2417 rbranchmap = remote.branchmap()
2420 rbranchmap = remote.branchmap()
2418
2421
2419 fp = remote.stream_out()
2422 fp = remote.stream_out()
2420 l = fp.readline()
2423 l = fp.readline()
2421 try:
2424 try:
2422 resp = int(l)
2425 resp = int(l)
2423 except ValueError:
2426 except ValueError:
2424 raise error.ResponseError(
2427 raise error.ResponseError(
2425 _('unexpected response from remote server:'), l)
2428 _('unexpected response from remote server:'), l)
2426 if resp == 1:
2429 if resp == 1:
2427 raise util.Abort(_('operation forbidden by server'))
2430 raise util.Abort(_('operation forbidden by server'))
2428 elif resp == 2:
2431 elif resp == 2:
2429 raise util.Abort(_('locking the remote repository failed'))
2432 raise util.Abort(_('locking the remote repository failed'))
2430 elif resp != 0:
2433 elif resp != 0:
2431 raise util.Abort(_('the server sent an unknown error code'))
2434 raise util.Abort(_('the server sent an unknown error code'))
2432 self.ui.status(_('streaming all changes\n'))
2435 self.ui.status(_('streaming all changes\n'))
2433 l = fp.readline()
2436 l = fp.readline()
2434 try:
2437 try:
2435 total_files, total_bytes = map(int, l.split(' ', 1))
2438 total_files, total_bytes = map(int, l.split(' ', 1))
2436 except (ValueError, TypeError):
2439 except (ValueError, TypeError):
2437 raise error.ResponseError(
2440 raise error.ResponseError(
2438 _('unexpected response from remote server:'), l)
2441 _('unexpected response from remote server:'), l)
2439 self.ui.status(_('%d files to transfer, %s of data\n') %
2442 self.ui.status(_('%d files to transfer, %s of data\n') %
2440 (total_files, util.bytecount(total_bytes)))
2443 (total_files, util.bytecount(total_bytes)))
2441 handled_bytes = 0
2444 handled_bytes = 0
2442 self.ui.progress(_('clone'), 0, total=total_bytes)
2445 self.ui.progress(_('clone'), 0, total=total_bytes)
2443 start = time.time()
2446 start = time.time()
2444 for i in xrange(total_files):
2447 for i in xrange(total_files):
2445 # XXX doesn't support '\n' or '\r' in filenames
2448 # XXX doesn't support '\n' or '\r' in filenames
2446 l = fp.readline()
2449 l = fp.readline()
2447 try:
2450 try:
2448 name, size = l.split('\0', 1)
2451 name, size = l.split('\0', 1)
2449 size = int(size)
2452 size = int(size)
2450 except (ValueError, TypeError):
2453 except (ValueError, TypeError):
2451 raise error.ResponseError(
2454 raise error.ResponseError(
2452 _('unexpected response from remote server:'), l)
2455 _('unexpected response from remote server:'), l)
2453 if self.ui.debugflag:
2456 if self.ui.debugflag:
2454 self.ui.debug('adding %s (%s)\n' %
2457 self.ui.debug('adding %s (%s)\n' %
2455 (name, util.bytecount(size)))
2458 (name, util.bytecount(size)))
2456 # for backwards compat, name was partially encoded
2459 # for backwards compat, name was partially encoded
2457 ofp = self.sopener(store.decodedir(name), 'w')
2460 ofp = self.sopener(store.decodedir(name), 'w')
2458 for chunk in util.filechunkiter(fp, limit=size):
2461 for chunk in util.filechunkiter(fp, limit=size):
2459 handled_bytes += len(chunk)
2462 handled_bytes += len(chunk)
2460 self.ui.progress(_('clone'), handled_bytes,
2463 self.ui.progress(_('clone'), handled_bytes,
2461 total=total_bytes)
2464 total=total_bytes)
2462 ofp.write(chunk)
2465 ofp.write(chunk)
2463 ofp.close()
2466 ofp.close()
2464 elapsed = time.time() - start
2467 elapsed = time.time() - start
2465 if elapsed <= 0:
2468 if elapsed <= 0:
2466 elapsed = 0.001
2469 elapsed = 0.001
2467 self.ui.progress(_('clone'), None)
2470 self.ui.progress(_('clone'), None)
2468 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2471 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2469 (util.bytecount(total_bytes), elapsed,
2472 (util.bytecount(total_bytes), elapsed,
2470 util.bytecount(total_bytes / elapsed)))
2473 util.bytecount(total_bytes / elapsed)))
2471
2474
2472 # new requirements = old non-format requirements +
2475 # new requirements = old non-format requirements +
2473 # new format-related
2476 # new format-related
2474 # requirements from the streamed-in repository
2477 # requirements from the streamed-in repository
2475 requirements.update(set(self.requirements) - self.supportedformats)
2478 requirements.update(set(self.requirements) - self.supportedformats)
2476 self._applyrequirements(requirements)
2479 self._applyrequirements(requirements)
2477 self._writerequirements()
2480 self._writerequirements()
2478
2481
2479 if rbranchmap:
2482 if rbranchmap:
2480 rbheads = []
2483 rbheads = []
2481 for bheads in rbranchmap.itervalues():
2484 for bheads in rbranchmap.itervalues():
2482 rbheads.extend(bheads)
2485 rbheads.extend(bheads)
2483
2486
2484 if rbheads:
2487 if rbheads:
2485 rtiprev = max((int(self.changelog.rev(node))
2488 rtiprev = max((int(self.changelog.rev(node))
2486 for node in rbheads))
2489 for node in rbheads))
2487 cache = branchmap.branchcache(rbranchmap,
2490 cache = branchmap.branchcache(rbranchmap,
2488 self[rtiprev].node(),
2491 self[rtiprev].node(),
2489 rtiprev)
2492 rtiprev)
2490 self._branchcaches[None] = cache
2493 self._branchcaches[None] = cache
2491 cache.write(self.unfiltered())
2494 cache.write(self.unfiltered())
2492 self.invalidate()
2495 self.invalidate()
2493 return len(self.heads()) + 1
2496 return len(self.heads()) + 1
2494 finally:
2497 finally:
2495 lock.release()
2498 lock.release()
2496
2499
2497 def clone(self, remote, heads=[], stream=False):
2500 def clone(self, remote, heads=[], stream=False):
2498 '''clone remote repository.
2501 '''clone remote repository.
2499
2502
2500 keyword arguments:
2503 keyword arguments:
2501 heads: list of revs to clone (forces use of pull)
2504 heads: list of revs to clone (forces use of pull)
2502 stream: use streaming clone if possible'''
2505 stream: use streaming clone if possible'''
2503
2506
2504 # now, all clients that can request uncompressed clones can
2507 # now, all clients that can request uncompressed clones can
2505 # read repo formats supported by all servers that can serve
2508 # read repo formats supported by all servers that can serve
2506 # them.
2509 # them.
2507
2510
2508 # if revlog format changes, client will have to check version
2511 # if revlog format changes, client will have to check version
2509 # and format flags on "stream" capability, and use
2512 # and format flags on "stream" capability, and use
2510 # uncompressed only if compatible.
2513 # uncompressed only if compatible.
2511
2514
2512 if not stream:
2515 if not stream:
2513 # if the server explicitly prefers to stream (for fast LANs)
2516 # if the server explicitly prefers to stream (for fast LANs)
2514 stream = remote.capable('stream-preferred')
2517 stream = remote.capable('stream-preferred')
2515
2518
2516 if stream and not heads:
2519 if stream and not heads:
2517 # 'stream' means remote revlog format is revlogv1 only
2520 # 'stream' means remote revlog format is revlogv1 only
2518 if remote.capable('stream'):
2521 if remote.capable('stream'):
2519 return self.stream_in(remote, set(('revlogv1',)))
2522 return self.stream_in(remote, set(('revlogv1',)))
2520 # otherwise, 'streamreqs' contains the remote revlog format
2523 # otherwise, 'streamreqs' contains the remote revlog format
2521 streamreqs = remote.capable('streamreqs')
2524 streamreqs = remote.capable('streamreqs')
2522 if streamreqs:
2525 if streamreqs:
2523 streamreqs = set(streamreqs.split(','))
2526 streamreqs = set(streamreqs.split(','))
2524 # if we support it, stream in and adjust our requirements
2527 # if we support it, stream in and adjust our requirements
2525 if not streamreqs - self.supportedformats:
2528 if not streamreqs - self.supportedformats:
2526 return self.stream_in(remote, streamreqs)
2529 return self.stream_in(remote, streamreqs)
2527 return self.pull(remote, heads)
2530 return self.pull(remote, heads)
2528
2531
2529 def pushkey(self, namespace, key, old, new):
2532 def pushkey(self, namespace, key, old, new):
2530 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2533 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2531 old=old, new=new)
2534 old=old, new=new)
2532 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2535 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2533 ret = pushkey.push(self, namespace, key, old, new)
2536 ret = pushkey.push(self, namespace, key, old, new)
2534 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2537 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2535 ret=ret)
2538 ret=ret)
2536 return ret
2539 return ret
2537
2540
2538 def listkeys(self, namespace):
2541 def listkeys(self, namespace):
2539 self.hook('prelistkeys', throw=True, namespace=namespace)
2542 self.hook('prelistkeys', throw=True, namespace=namespace)
2540 self.ui.debug('listing keys for "%s"\n' % namespace)
2543 self.ui.debug('listing keys for "%s"\n' % namespace)
2541 values = pushkey.list(self, namespace)
2544 values = pushkey.list(self, namespace)
2542 self.hook('listkeys', namespace=namespace, values=values)
2545 self.hook('listkeys', namespace=namespace, values=values)
2543 return values
2546 return values
2544
2547
2545 def debugwireargs(self, one, two, three=None, four=None, five=None):
2548 def debugwireargs(self, one, two, three=None, four=None, five=None):
2546 '''used to test argument passing over the wire'''
2549 '''used to test argument passing over the wire'''
2547 return "%s %s %s %s %s" % (one, two, three, four, five)
2550 return "%s %s %s %s %s" % (one, two, three, four, five)
2548
2551
2549 def savecommitmessage(self, text):
2552 def savecommitmessage(self, text):
2550 fp = self.opener('last-message.txt', 'wb')
2553 fp = self.opener('last-message.txt', 'wb')
2551 try:
2554 try:
2552 fp.write(text)
2555 fp.write(text)
2553 finally:
2556 finally:
2554 fp.close()
2557 fp.close()
2555 return self.pathto(fp.name[len(self.root) + 1:])
2558 return self.pathto(fp.name[len(self.root) + 1:])
2556
2559
2557 # used to avoid circular references so destructors work
2560 # used to avoid circular references so destructors work
2558 def aftertrans(files):
2561 def aftertrans(files):
2559 renamefiles = [tuple(t) for t in files]
2562 renamefiles = [tuple(t) for t in files]
2560 def a():
2563 def a():
2561 for src, dest in renamefiles:
2564 for src, dest in renamefiles:
2562 try:
2565 try:
2563 util.rename(src, dest)
2566 util.rename(src, dest)
2564 except OSError: # journal file does not yet exist
2567 except OSError: # journal file does not yet exist
2565 pass
2568 pass
2566 return a
2569 return a
2567
2570
2568 def undoname(fn):
2571 def undoname(fn):
2569 base, name = os.path.split(fn)
2572 base, name = os.path.split(fn)
2570 assert name.startswith('journal')
2573 assert name.startswith('journal')
2571 return os.path.join(base, name.replace('journal', 'undo', 1))
2574 return os.path.join(base, name.replace('journal', 'undo', 1))
2572
2575
2573 def instance(ui, path, create):
2576 def instance(ui, path, create):
2574 return localrepository(ui, util.urllocalpath(path), create)
2577 return localrepository(ui, util.urllocalpath(path), create)
2575
2578
2576 def islocal(path):
2579 def islocal(path):
2577 return True
2580 return True
@@ -1,204 +1,200 b''
1 # repair.py - functions for repository repair for mercurial
1 # repair.py - functions for repository repair for mercurial
2 #
2 #
3 # Copyright 2005, 2006 Chris Mason <mason@suse.com>
3 # Copyright 2005, 2006 Chris Mason <mason@suse.com>
4 # Copyright 2007 Matt Mackall
4 # Copyright 2007 Matt Mackall
5 #
5 #
6 # This software may be used and distributed according to the terms of the
6 # This software may be used and distributed according to the terms of the
7 # GNU General Public License version 2 or any later version.
7 # GNU General Public License version 2 or any later version.
8
8
9 from mercurial import changegroup, branchmap
9 from mercurial import changegroup
10 from mercurial.node import short
10 from mercurial.node import short
11 from mercurial.i18n import _
11 from mercurial.i18n import _
12 import os
12 import os
13 import errno
13 import errno
14
14
15 def _bundle(repo, bases, heads, node, suffix, compress=True):
15 def _bundle(repo, bases, heads, node, suffix, compress=True):
16 """create a bundle with the specified revisions as a backup"""
16 """create a bundle with the specified revisions as a backup"""
17 cg = repo.changegroupsubset(bases, heads, 'strip')
17 cg = repo.changegroupsubset(bases, heads, 'strip')
18 backupdir = repo.join("strip-backup")
18 backupdir = repo.join("strip-backup")
19 if not os.path.isdir(backupdir):
19 if not os.path.isdir(backupdir):
20 os.mkdir(backupdir)
20 os.mkdir(backupdir)
21 name = os.path.join(backupdir, "%s-%s.hg" % (short(node), suffix))
21 name = os.path.join(backupdir, "%s-%s.hg" % (short(node), suffix))
22 if compress:
22 if compress:
23 bundletype = "HG10BZ"
23 bundletype = "HG10BZ"
24 else:
24 else:
25 bundletype = "HG10UN"
25 bundletype = "HG10UN"
26 return changegroup.writebundle(cg, name, bundletype)
26 return changegroup.writebundle(cg, name, bundletype)
27
27
28 def _collectfiles(repo, striprev):
28 def _collectfiles(repo, striprev):
29 """find out the filelogs affected by the strip"""
29 """find out the filelogs affected by the strip"""
30 files = set()
30 files = set()
31
31
32 for x in xrange(striprev, len(repo)):
32 for x in xrange(striprev, len(repo)):
33 files.update(repo[x].files())
33 files.update(repo[x].files())
34
34
35 return sorted(files)
35 return sorted(files)
36
36
37 def _collectbrokencsets(repo, files, striprev):
37 def _collectbrokencsets(repo, files, striprev):
38 """return the changesets which will be broken by the truncation"""
38 """return the changesets which will be broken by the truncation"""
39 s = set()
39 s = set()
40 def collectone(revlog):
40 def collectone(revlog):
41 linkgen = (revlog.linkrev(i) for i in revlog)
41 linkgen = (revlog.linkrev(i) for i in revlog)
42 # find the truncation point of the revlog
42 # find the truncation point of the revlog
43 for lrev in linkgen:
43 for lrev in linkgen:
44 if lrev >= striprev:
44 if lrev >= striprev:
45 break
45 break
46 # see if any revision after this point has a linkrev
46 # see if any revision after this point has a linkrev
47 # less than striprev (those will be broken by strip)
47 # less than striprev (those will be broken by strip)
48 for lrev in linkgen:
48 for lrev in linkgen:
49 if lrev < striprev:
49 if lrev < striprev:
50 s.add(lrev)
50 s.add(lrev)
51
51
52 collectone(repo.manifest)
52 collectone(repo.manifest)
53 for fname in files:
53 for fname in files:
54 collectone(repo.file(fname))
54 collectone(repo.file(fname))
55
55
56 return s
56 return s
57
57
58 def strip(ui, repo, nodelist, backup="all", topic='backup'):
58 def strip(ui, repo, nodelist, backup="all", topic='backup'):
59 repo = repo.unfiltered()
59 repo = repo.unfiltered()
60 # It simplifies the logic around updating the branchheads cache if we only
61 # have to consider the effect of the stripped revisions and not revisions
62 # missing because the cache is out-of-date.
63 branchmap.updatecache(repo)
64 repo.destroying()
60 repo.destroying()
65
61
66 cl = repo.changelog
62 cl = repo.changelog
67 # TODO handle undo of merge sets
63 # TODO handle undo of merge sets
68 if isinstance(nodelist, str):
64 if isinstance(nodelist, str):
69 nodelist = [nodelist]
65 nodelist = [nodelist]
70 striplist = [cl.rev(node) for node in nodelist]
66 striplist = [cl.rev(node) for node in nodelist]
71 striprev = min(striplist)
67 striprev = min(striplist)
72
68
73 # Generate set of branches who will have nodes stripped.
69 # Generate set of branches who will have nodes stripped.
74 striprevs = repo.revs("%ld::", striplist)
70 striprevs = repo.revs("%ld::", striplist)
75 stripbranches = set([repo[rev].branch() for rev in striprevs])
71 stripbranches = set([repo[rev].branch() for rev in striprevs])
76
72
77 # Set of potential new heads resulting from the strip. The parents of any
73 # Set of potential new heads resulting from the strip. The parents of any
78 # node removed could be a new head because the node to be removed could have
74 # node removed could be a new head because the node to be removed could have
79 # been the only child of the parent.
75 # been the only child of the parent.
80 newheadrevs = repo.revs("parents(%ld::) - %ld::", striprevs, striprevs)
76 newheadrevs = repo.revs("parents(%ld::) - %ld::", striprevs, striprevs)
81 newheadnodes = set([cl.node(rev) for rev in newheadrevs])
77 newheadnodes = set([cl.node(rev) for rev in newheadrevs])
82 newheadbranches = set([repo[rev].branch() for rev in newheadrevs])
78 newheadbranches = set([repo[rev].branch() for rev in newheadrevs])
83
79
84 keeppartialbundle = backup == 'strip'
80 keeppartialbundle = backup == 'strip'
85
81
86 # Some revisions with rev > striprev may not be descendants of striprev.
82 # Some revisions with rev > striprev may not be descendants of striprev.
87 # We have to find these revisions and put them in a bundle, so that
83 # We have to find these revisions and put them in a bundle, so that
88 # we can restore them after the truncations.
84 # we can restore them after the truncations.
89 # To create the bundle we use repo.changegroupsubset which requires
85 # To create the bundle we use repo.changegroupsubset which requires
90 # the list of heads and bases of the set of interesting revisions.
86 # the list of heads and bases of the set of interesting revisions.
91 # (head = revision in the set that has no descendant in the set;
87 # (head = revision in the set that has no descendant in the set;
92 # base = revision in the set that has no ancestor in the set)
88 # base = revision in the set that has no ancestor in the set)
93 tostrip = set(striplist)
89 tostrip = set(striplist)
94 for rev in striplist:
90 for rev in striplist:
95 for desc in cl.descendants([rev]):
91 for desc in cl.descendants([rev]):
96 tostrip.add(desc)
92 tostrip.add(desc)
97
93
98 files = _collectfiles(repo, striprev)
94 files = _collectfiles(repo, striprev)
99 saverevs = _collectbrokencsets(repo, files, striprev)
95 saverevs = _collectbrokencsets(repo, files, striprev)
100
96
101 # compute heads
97 # compute heads
102 saveheads = set(saverevs)
98 saveheads = set(saverevs)
103 for r in xrange(striprev + 1, len(cl)):
99 for r in xrange(striprev + 1, len(cl)):
104 if r not in tostrip:
100 if r not in tostrip:
105 saverevs.add(r)
101 saverevs.add(r)
106 saveheads.difference_update(cl.parentrevs(r))
102 saveheads.difference_update(cl.parentrevs(r))
107 saveheads.add(r)
103 saveheads.add(r)
108 saveheads = [cl.node(r) for r in saveheads]
104 saveheads = [cl.node(r) for r in saveheads]
109
105
110 # compute base nodes
106 # compute base nodes
111 if saverevs:
107 if saverevs:
112 descendants = set(cl.descendants(saverevs))
108 descendants = set(cl.descendants(saverevs))
113 saverevs.difference_update(descendants)
109 saverevs.difference_update(descendants)
114 savebases = [cl.node(r) for r in saverevs]
110 savebases = [cl.node(r) for r in saverevs]
115 stripbases = [cl.node(r) for r in tostrip]
111 stripbases = [cl.node(r) for r in tostrip]
116
112
117 # For a set s, max(parents(s) - s) is the same as max(heads(::s - s)), but
113 # For a set s, max(parents(s) - s) is the same as max(heads(::s - s)), but
118 # is much faster
114 # is much faster
119 newbmtarget = repo.revs('max(parents(%ld) - (%ld))', tostrip, tostrip)
115 newbmtarget = repo.revs('max(parents(%ld) - (%ld))', tostrip, tostrip)
120 if newbmtarget:
116 if newbmtarget:
121 newbmtarget = repo[newbmtarget[0]].node()
117 newbmtarget = repo[newbmtarget[0]].node()
122 else:
118 else:
123 newbmtarget = '.'
119 newbmtarget = '.'
124
120
125 bm = repo._bookmarks
121 bm = repo._bookmarks
126 updatebm = []
122 updatebm = []
127 for m in bm:
123 for m in bm:
128 rev = repo[bm[m]].rev()
124 rev = repo[bm[m]].rev()
129 if rev in tostrip:
125 if rev in tostrip:
130 updatebm.append(m)
126 updatebm.append(m)
131
127
132 # create a changegroup for all the branches we need to keep
128 # create a changegroup for all the branches we need to keep
133 backupfile = None
129 backupfile = None
134 if backup == "all":
130 if backup == "all":
135 backupfile = _bundle(repo, stripbases, cl.heads(), node, topic)
131 backupfile = _bundle(repo, stripbases, cl.heads(), node, topic)
136 repo.ui.status(_("saved backup bundle to %s\n") % backupfile)
132 repo.ui.status(_("saved backup bundle to %s\n") % backupfile)
137 if saveheads or savebases:
133 if saveheads or savebases:
138 # do not compress partial bundle if we remove it from disk later
134 # do not compress partial bundle if we remove it from disk later
139 chgrpfile = _bundle(repo, savebases, saveheads, node, 'temp',
135 chgrpfile = _bundle(repo, savebases, saveheads, node, 'temp',
140 compress=keeppartialbundle)
136 compress=keeppartialbundle)
141
137
142 mfst = repo.manifest
138 mfst = repo.manifest
143
139
144 tr = repo.transaction("strip")
140 tr = repo.transaction("strip")
145 offset = len(tr.entries)
141 offset = len(tr.entries)
146
142
147 try:
143 try:
148 tr.startgroup()
144 tr.startgroup()
149 cl.strip(striprev, tr)
145 cl.strip(striprev, tr)
150 mfst.strip(striprev, tr)
146 mfst.strip(striprev, tr)
151 for fn in files:
147 for fn in files:
152 repo.file(fn).strip(striprev, tr)
148 repo.file(fn).strip(striprev, tr)
153 tr.endgroup()
149 tr.endgroup()
154
150
155 try:
151 try:
156 for i in xrange(offset, len(tr.entries)):
152 for i in xrange(offset, len(tr.entries)):
157 file, troffset, ignore = tr.entries[i]
153 file, troffset, ignore = tr.entries[i]
158 repo.sopener(file, 'a').truncate(troffset)
154 repo.sopener(file, 'a').truncate(troffset)
159 tr.close()
155 tr.close()
160 except: # re-raises
156 except: # re-raises
161 tr.abort()
157 tr.abort()
162 raise
158 raise
163
159
164 if saveheads or savebases:
160 if saveheads or savebases:
165 ui.note(_("adding branch\n"))
161 ui.note(_("adding branch\n"))
166 f = open(chgrpfile, "rb")
162 f = open(chgrpfile, "rb")
167 gen = changegroup.readbundle(f, chgrpfile)
163 gen = changegroup.readbundle(f, chgrpfile)
168 if not repo.ui.verbose:
164 if not repo.ui.verbose:
169 # silence internal shuffling chatter
165 # silence internal shuffling chatter
170 repo.ui.pushbuffer()
166 repo.ui.pushbuffer()
171 repo.addchangegroup(gen, 'strip', 'bundle:' + chgrpfile, True)
167 repo.addchangegroup(gen, 'strip', 'bundle:' + chgrpfile, True)
172 if not repo.ui.verbose:
168 if not repo.ui.verbose:
173 repo.ui.popbuffer()
169 repo.ui.popbuffer()
174 f.close()
170 f.close()
175 if not keeppartialbundle:
171 if not keeppartialbundle:
176 os.unlink(chgrpfile)
172 os.unlink(chgrpfile)
177
173
178 # remove undo files
174 # remove undo files
179 for undofile in repo.undofiles():
175 for undofile in repo.undofiles():
180 try:
176 try:
181 os.unlink(undofile)
177 os.unlink(undofile)
182 except OSError, e:
178 except OSError, e:
183 if e.errno != errno.ENOENT:
179 if e.errno != errno.ENOENT:
184 ui.warn(_('error removing %s: %s\n') % (undofile, str(e)))
180 ui.warn(_('error removing %s: %s\n') % (undofile, str(e)))
185
181
186 for m in updatebm:
182 for m in updatebm:
187 bm[m] = repo[newbmtarget].node()
183 bm[m] = repo[newbmtarget].node()
188 bm.write()
184 bm.write()
189 except: # re-raises
185 except: # re-raises
190 if backupfile:
186 if backupfile:
191 ui.warn(_("strip failed, full bundle stored in '%s'\n")
187 ui.warn(_("strip failed, full bundle stored in '%s'\n")
192 % backupfile)
188 % backupfile)
193 elif saveheads:
189 elif saveheads:
194 ui.warn(_("strip failed, partial bundle stored in '%s'\n")
190 ui.warn(_("strip failed, partial bundle stored in '%s'\n")
195 % chgrpfile)
191 % chgrpfile)
196 raise
192 raise
197
193
198 if len(stripbranches) == 1 and len(newheadbranches) == 1 \
194 if len(stripbranches) == 1 and len(newheadbranches) == 1 \
199 and stripbranches == newheadbranches:
195 and stripbranches == newheadbranches:
200 repo.destroyed(newheadnodes)
196 repo.destroyed(newheadnodes)
201 else:
197 else:
202 # Multiple branches involved in strip. Will allow branchcache to become
198 # Multiple branches involved in strip. Will allow branchcache to become
203 # invalid and later on rebuilt from scratch
199 # invalid and later on rebuilt from scratch
204 repo.destroyed()
200 repo.destroyed()
General Comments 0
You need to be logged in to leave comments. Login now