##// END OF EJS Templates
validate: check for spurious incoming filelog entries...
Sune Foldager -
r18422:f009804e default
parent child Browse files
Show More
@@ -1,2573 +1,2576 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import lock, transaction, store, encoding, base85
11 import lock, transaction, store, encoding, base85
12 import scmutil, util, extensions, hook, error, revset
12 import scmutil, util, extensions, hook, error, revset
13 import match as matchmod
13 import match as matchmod
14 import merge as mergemod
14 import merge as mergemod
15 import tags as tagsmod
15 import tags as tagsmod
16 from lock import release
16 from lock import release
17 import weakref, errno, os, time, inspect
17 import weakref, errno, os, time, inspect
18 import branchmap
18 import branchmap
19 propertycache = util.propertycache
19 propertycache = util.propertycache
20 filecache = scmutil.filecache
20 filecache = scmutil.filecache
21
21
22 class repofilecache(filecache):
22 class repofilecache(filecache):
23 """All filecache usage on repo are done for logic that should be unfiltered
23 """All filecache usage on repo are done for logic that should be unfiltered
24 """
24 """
25
25
26 def __get__(self, repo, type=None):
26 def __get__(self, repo, type=None):
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 def __set__(self, repo, value):
28 def __set__(self, repo, value):
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 def __delete__(self, repo):
30 def __delete__(self, repo):
31 return super(repofilecache, self).__delete__(repo.unfiltered())
31 return super(repofilecache, self).__delete__(repo.unfiltered())
32
32
33 class storecache(repofilecache):
33 class storecache(repofilecache):
34 """filecache for files in the store"""
34 """filecache for files in the store"""
35 def join(self, obj, fname):
35 def join(self, obj, fname):
36 return obj.sjoin(fname)
36 return obj.sjoin(fname)
37
37
38 class unfilteredpropertycache(propertycache):
38 class unfilteredpropertycache(propertycache):
39 """propertycache that apply to unfiltered repo only"""
39 """propertycache that apply to unfiltered repo only"""
40
40
41 def __get__(self, repo, type=None):
41 def __get__(self, repo, type=None):
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43
43
44 class filteredpropertycache(propertycache):
44 class filteredpropertycache(propertycache):
45 """propertycache that must take filtering in account"""
45 """propertycache that must take filtering in account"""
46
46
47 def cachevalue(self, obj, value):
47 def cachevalue(self, obj, value):
48 object.__setattr__(obj, self.name, value)
48 object.__setattr__(obj, self.name, value)
49
49
50
50
51 def hasunfilteredcache(repo, name):
51 def hasunfilteredcache(repo, name):
52 """check if an repo and a unfilteredproperty cached value for <name>"""
52 """check if an repo and a unfilteredproperty cached value for <name>"""
53 return name in vars(repo.unfiltered())
53 return name in vars(repo.unfiltered())
54
54
55 def unfilteredmethod(orig):
55 def unfilteredmethod(orig):
56 """decorate method that always need to be run on unfiltered version"""
56 """decorate method that always need to be run on unfiltered version"""
57 def wrapper(repo, *args, **kwargs):
57 def wrapper(repo, *args, **kwargs):
58 return orig(repo.unfiltered(), *args, **kwargs)
58 return orig(repo.unfiltered(), *args, **kwargs)
59 return wrapper
59 return wrapper
60
60
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63
63
64 class localpeer(peer.peerrepository):
64 class localpeer(peer.peerrepository):
65 '''peer for a local repo; reflects only the most recent API'''
65 '''peer for a local repo; reflects only the most recent API'''
66
66
67 def __init__(self, repo, caps=MODERNCAPS):
67 def __init__(self, repo, caps=MODERNCAPS):
68 peer.peerrepository.__init__(self)
68 peer.peerrepository.__init__(self)
69 self._repo = repo.filtered('served')
69 self._repo = repo.filtered('served')
70 self.ui = repo.ui
70 self.ui = repo.ui
71 self._caps = repo._restrictcapabilities(caps)
71 self._caps = repo._restrictcapabilities(caps)
72 self.requirements = repo.requirements
72 self.requirements = repo.requirements
73 self.supportedformats = repo.supportedformats
73 self.supportedformats = repo.supportedformats
74
74
75 def close(self):
75 def close(self):
76 self._repo.close()
76 self._repo.close()
77
77
78 def _capabilities(self):
78 def _capabilities(self):
79 return self._caps
79 return self._caps
80
80
81 def local(self):
81 def local(self):
82 return self._repo
82 return self._repo
83
83
84 def canpush(self):
84 def canpush(self):
85 return True
85 return True
86
86
87 def url(self):
87 def url(self):
88 return self._repo.url()
88 return self._repo.url()
89
89
90 def lookup(self, key):
90 def lookup(self, key):
91 return self._repo.lookup(key)
91 return self._repo.lookup(key)
92
92
93 def branchmap(self):
93 def branchmap(self):
94 return self._repo.branchmap()
94 return self._repo.branchmap()
95
95
96 def heads(self):
96 def heads(self):
97 return self._repo.heads()
97 return self._repo.heads()
98
98
99 def known(self, nodes):
99 def known(self, nodes):
100 return self._repo.known(nodes)
100 return self._repo.known(nodes)
101
101
102 def getbundle(self, source, heads=None, common=None):
102 def getbundle(self, source, heads=None, common=None):
103 return self._repo.getbundle(source, heads=heads, common=common)
103 return self._repo.getbundle(source, heads=heads, common=common)
104
104
105 # TODO We might want to move the next two calls into legacypeer and add
105 # TODO We might want to move the next two calls into legacypeer and add
106 # unbundle instead.
106 # unbundle instead.
107
107
108 def lock(self):
108 def lock(self):
109 return self._repo.lock()
109 return self._repo.lock()
110
110
111 def addchangegroup(self, cg, source, url):
111 def addchangegroup(self, cg, source, url):
112 return self._repo.addchangegroup(cg, source, url)
112 return self._repo.addchangegroup(cg, source, url)
113
113
114 def pushkey(self, namespace, key, old, new):
114 def pushkey(self, namespace, key, old, new):
115 return self._repo.pushkey(namespace, key, old, new)
115 return self._repo.pushkey(namespace, key, old, new)
116
116
117 def listkeys(self, namespace):
117 def listkeys(self, namespace):
118 return self._repo.listkeys(namespace)
118 return self._repo.listkeys(namespace)
119
119
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 '''used to test argument passing over the wire'''
121 '''used to test argument passing over the wire'''
122 return "%s %s %s %s %s" % (one, two, three, four, five)
122 return "%s %s %s %s %s" % (one, two, three, four, five)
123
123
124 class locallegacypeer(localpeer):
124 class locallegacypeer(localpeer):
125 '''peer extension which implements legacy methods too; used for tests with
125 '''peer extension which implements legacy methods too; used for tests with
126 restricted capabilities'''
126 restricted capabilities'''
127
127
128 def __init__(self, repo):
128 def __init__(self, repo):
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130
130
131 def branches(self, nodes):
131 def branches(self, nodes):
132 return self._repo.branches(nodes)
132 return self._repo.branches(nodes)
133
133
134 def between(self, pairs):
134 def between(self, pairs):
135 return self._repo.between(pairs)
135 return self._repo.between(pairs)
136
136
137 def changegroup(self, basenodes, source):
137 def changegroup(self, basenodes, source):
138 return self._repo.changegroup(basenodes, source)
138 return self._repo.changegroup(basenodes, source)
139
139
140 def changegroupsubset(self, bases, heads, source):
140 def changegroupsubset(self, bases, heads, source):
141 return self._repo.changegroupsubset(bases, heads, source)
141 return self._repo.changegroupsubset(bases, heads, source)
142
142
143 class localrepository(object):
143 class localrepository(object):
144
144
145 supportedformats = set(('revlogv1', 'generaldelta'))
145 supportedformats = set(('revlogv1', 'generaldelta'))
146 supported = supportedformats | set(('store', 'fncache', 'shared',
146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 'dotencode'))
147 'dotencode'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
149 requirements = ['revlogv1']
149 requirements = ['revlogv1']
150 filtername = None
150 filtername = None
151
151
152 def _baserequirements(self, create):
152 def _baserequirements(self, create):
153 return self.requirements[:]
153 return self.requirements[:]
154
154
155 def __init__(self, baseui, path=None, create=False):
155 def __init__(self, baseui, path=None, create=False):
156 self.wvfs = scmutil.vfs(path, expand=True)
156 self.wvfs = scmutil.vfs(path, expand=True)
157 self.wopener = self.wvfs
157 self.wopener = self.wvfs
158 self.root = self.wvfs.base
158 self.root = self.wvfs.base
159 self.path = self.wvfs.join(".hg")
159 self.path = self.wvfs.join(".hg")
160 self.origroot = path
160 self.origroot = path
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 self.vfs = scmutil.vfs(self.path)
162 self.vfs = scmutil.vfs(self.path)
163 self.opener = self.vfs
163 self.opener = self.vfs
164 self.baseui = baseui
164 self.baseui = baseui
165 self.ui = baseui.copy()
165 self.ui = baseui.copy()
166 # A list of callback to shape the phase if no data were found.
166 # A list of callback to shape the phase if no data were found.
167 # Callback are in the form: func(repo, roots) --> processed root.
167 # Callback are in the form: func(repo, roots) --> processed root.
168 # This list it to be filled by extension during repo setup
168 # This list it to be filled by extension during repo setup
169 self._phasedefaults = []
169 self._phasedefaults = []
170 try:
170 try:
171 self.ui.readconfig(self.join("hgrc"), self.root)
171 self.ui.readconfig(self.join("hgrc"), self.root)
172 extensions.loadall(self.ui)
172 extensions.loadall(self.ui)
173 except IOError:
173 except IOError:
174 pass
174 pass
175
175
176 if not self.vfs.isdir():
176 if not self.vfs.isdir():
177 if create:
177 if create:
178 if not self.wvfs.exists():
178 if not self.wvfs.exists():
179 self.wvfs.makedirs()
179 self.wvfs.makedirs()
180 self.vfs.makedir(notindexed=True)
180 self.vfs.makedir(notindexed=True)
181 requirements = self._baserequirements(create)
181 requirements = self._baserequirements(create)
182 if self.ui.configbool('format', 'usestore', True):
182 if self.ui.configbool('format', 'usestore', True):
183 self.vfs.mkdir("store")
183 self.vfs.mkdir("store")
184 requirements.append("store")
184 requirements.append("store")
185 if self.ui.configbool('format', 'usefncache', True):
185 if self.ui.configbool('format', 'usefncache', True):
186 requirements.append("fncache")
186 requirements.append("fncache")
187 if self.ui.configbool('format', 'dotencode', True):
187 if self.ui.configbool('format', 'dotencode', True):
188 requirements.append('dotencode')
188 requirements.append('dotencode')
189 # create an invalid changelog
189 # create an invalid changelog
190 self.vfs.append(
190 self.vfs.append(
191 "00changelog.i",
191 "00changelog.i",
192 '\0\0\0\2' # represents revlogv2
192 '\0\0\0\2' # represents revlogv2
193 ' dummy changelog to prevent using the old repo layout'
193 ' dummy changelog to prevent using the old repo layout'
194 )
194 )
195 if self.ui.configbool('format', 'generaldelta', False):
195 if self.ui.configbool('format', 'generaldelta', False):
196 requirements.append("generaldelta")
196 requirements.append("generaldelta")
197 requirements = set(requirements)
197 requirements = set(requirements)
198 else:
198 else:
199 raise error.RepoError(_("repository %s not found") % path)
199 raise error.RepoError(_("repository %s not found") % path)
200 elif create:
200 elif create:
201 raise error.RepoError(_("repository %s already exists") % path)
201 raise error.RepoError(_("repository %s already exists") % path)
202 else:
202 else:
203 try:
203 try:
204 requirements = scmutil.readrequires(self.vfs, self.supported)
204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 except IOError, inst:
205 except IOError, inst:
206 if inst.errno != errno.ENOENT:
206 if inst.errno != errno.ENOENT:
207 raise
207 raise
208 requirements = set()
208 requirements = set()
209
209
210 self.sharedpath = self.path
210 self.sharedpath = self.path
211 try:
211 try:
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 if not os.path.exists(s):
213 if not os.path.exists(s):
214 raise error.RepoError(
214 raise error.RepoError(
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 self.sharedpath = s
216 self.sharedpath = s
217 except IOError, inst:
217 except IOError, inst:
218 if inst.errno != errno.ENOENT:
218 if inst.errno != errno.ENOENT:
219 raise
219 raise
220
220
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 self.spath = self.store.path
222 self.spath = self.store.path
223 self.svfs = self.store.vfs
223 self.svfs = self.store.vfs
224 self.sopener = self.svfs
224 self.sopener = self.svfs
225 self.sjoin = self.store.join
225 self.sjoin = self.store.join
226 self.vfs.createmode = self.store.createmode
226 self.vfs.createmode = self.store.createmode
227 self._applyrequirements(requirements)
227 self._applyrequirements(requirements)
228 if create:
228 if create:
229 self._writerequirements()
229 self._writerequirements()
230
230
231
231
232 self._branchcaches = {}
232 self._branchcaches = {}
233 self.filterpats = {}
233 self.filterpats = {}
234 self._datafilters = {}
234 self._datafilters = {}
235 self._transref = self._lockref = self._wlockref = None
235 self._transref = self._lockref = self._wlockref = None
236
236
237 # A cache for various files under .hg/ that tracks file changes,
237 # A cache for various files under .hg/ that tracks file changes,
238 # (used by the filecache decorator)
238 # (used by the filecache decorator)
239 #
239 #
240 # Maps a property name to its util.filecacheentry
240 # Maps a property name to its util.filecacheentry
241 self._filecache = {}
241 self._filecache = {}
242
242
243 # hold sets of revision to be filtered
243 # hold sets of revision to be filtered
244 # should be cleared when something might have changed the filter value:
244 # should be cleared when something might have changed the filter value:
245 # - new changesets,
245 # - new changesets,
246 # - phase change,
246 # - phase change,
247 # - new obsolescence marker,
247 # - new obsolescence marker,
248 # - working directory parent change,
248 # - working directory parent change,
249 # - bookmark changes
249 # - bookmark changes
250 self.filteredrevcache = {}
250 self.filteredrevcache = {}
251
251
252 def close(self):
252 def close(self):
253 pass
253 pass
254
254
255 def _restrictcapabilities(self, caps):
255 def _restrictcapabilities(self, caps):
256 return caps
256 return caps
257
257
258 def _applyrequirements(self, requirements):
258 def _applyrequirements(self, requirements):
259 self.requirements = requirements
259 self.requirements = requirements
260 self.sopener.options = dict((r, 1) for r in requirements
260 self.sopener.options = dict((r, 1) for r in requirements
261 if r in self.openerreqs)
261 if r in self.openerreqs)
262
262
263 def _writerequirements(self):
263 def _writerequirements(self):
264 reqfile = self.opener("requires", "w")
264 reqfile = self.opener("requires", "w")
265 for r in sorted(self.requirements):
265 for r in sorted(self.requirements):
266 reqfile.write("%s\n" % r)
266 reqfile.write("%s\n" % r)
267 reqfile.close()
267 reqfile.close()
268
268
269 def _checknested(self, path):
269 def _checknested(self, path):
270 """Determine if path is a legal nested repository."""
270 """Determine if path is a legal nested repository."""
271 if not path.startswith(self.root):
271 if not path.startswith(self.root):
272 return False
272 return False
273 subpath = path[len(self.root) + 1:]
273 subpath = path[len(self.root) + 1:]
274 normsubpath = util.pconvert(subpath)
274 normsubpath = util.pconvert(subpath)
275
275
276 # XXX: Checking against the current working copy is wrong in
276 # XXX: Checking against the current working copy is wrong in
277 # the sense that it can reject things like
277 # the sense that it can reject things like
278 #
278 #
279 # $ hg cat -r 10 sub/x.txt
279 # $ hg cat -r 10 sub/x.txt
280 #
280 #
281 # if sub/ is no longer a subrepository in the working copy
281 # if sub/ is no longer a subrepository in the working copy
282 # parent revision.
282 # parent revision.
283 #
283 #
284 # However, it can of course also allow things that would have
284 # However, it can of course also allow things that would have
285 # been rejected before, such as the above cat command if sub/
285 # been rejected before, such as the above cat command if sub/
286 # is a subrepository now, but was a normal directory before.
286 # is a subrepository now, but was a normal directory before.
287 # The old path auditor would have rejected by mistake since it
287 # The old path auditor would have rejected by mistake since it
288 # panics when it sees sub/.hg/.
288 # panics when it sees sub/.hg/.
289 #
289 #
290 # All in all, checking against the working copy seems sensible
290 # All in all, checking against the working copy seems sensible
291 # since we want to prevent access to nested repositories on
291 # since we want to prevent access to nested repositories on
292 # the filesystem *now*.
292 # the filesystem *now*.
293 ctx = self[None]
293 ctx = self[None]
294 parts = util.splitpath(subpath)
294 parts = util.splitpath(subpath)
295 while parts:
295 while parts:
296 prefix = '/'.join(parts)
296 prefix = '/'.join(parts)
297 if prefix in ctx.substate:
297 if prefix in ctx.substate:
298 if prefix == normsubpath:
298 if prefix == normsubpath:
299 return True
299 return True
300 else:
300 else:
301 sub = ctx.sub(prefix)
301 sub = ctx.sub(prefix)
302 return sub.checknested(subpath[len(prefix) + 1:])
302 return sub.checknested(subpath[len(prefix) + 1:])
303 else:
303 else:
304 parts.pop()
304 parts.pop()
305 return False
305 return False
306
306
307 def peer(self):
307 def peer(self):
308 return localpeer(self) # not cached to avoid reference cycle
308 return localpeer(self) # not cached to avoid reference cycle
309
309
310 def unfiltered(self):
310 def unfiltered(self):
311 """Return unfiltered version of the repository
311 """Return unfiltered version of the repository
312
312
313 Intended to be ovewritten by filtered repo."""
313 Intended to be ovewritten by filtered repo."""
314 return self
314 return self
315
315
316 def filtered(self, name):
316 def filtered(self, name):
317 """Return a filtered version of a repository"""
317 """Return a filtered version of a repository"""
318 # build a new class with the mixin and the current class
318 # build a new class with the mixin and the current class
319 # (possibily subclass of the repo)
319 # (possibily subclass of the repo)
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 pass
321 pass
322 return proxycls(self, name)
322 return proxycls(self, name)
323
323
324 @repofilecache('bookmarks')
324 @repofilecache('bookmarks')
325 def _bookmarks(self):
325 def _bookmarks(self):
326 return bookmarks.bmstore(self)
326 return bookmarks.bmstore(self)
327
327
328 @repofilecache('bookmarks.current')
328 @repofilecache('bookmarks.current')
329 def _bookmarkcurrent(self):
329 def _bookmarkcurrent(self):
330 return bookmarks.readcurrent(self)
330 return bookmarks.readcurrent(self)
331
331
332 def bookmarkheads(self, bookmark):
332 def bookmarkheads(self, bookmark):
333 name = bookmark.split('@', 1)[0]
333 name = bookmark.split('@', 1)[0]
334 heads = []
334 heads = []
335 for mark, n in self._bookmarks.iteritems():
335 for mark, n in self._bookmarks.iteritems():
336 if mark.split('@', 1)[0] == name:
336 if mark.split('@', 1)[0] == name:
337 heads.append(n)
337 heads.append(n)
338 return heads
338 return heads
339
339
340 @storecache('phaseroots')
340 @storecache('phaseroots')
341 def _phasecache(self):
341 def _phasecache(self):
342 return phases.phasecache(self, self._phasedefaults)
342 return phases.phasecache(self, self._phasedefaults)
343
343
344 @storecache('obsstore')
344 @storecache('obsstore')
345 def obsstore(self):
345 def obsstore(self):
346 store = obsolete.obsstore(self.sopener)
346 store = obsolete.obsstore(self.sopener)
347 if store and not obsolete._enabled:
347 if store and not obsolete._enabled:
348 # message is rare enough to not be translated
348 # message is rare enough to not be translated
349 msg = 'obsolete feature not enabled but %i markers found!\n'
349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 self.ui.warn(msg % len(list(store)))
350 self.ui.warn(msg % len(list(store)))
351 return store
351 return store
352
352
353 @storecache('00changelog.i')
353 @storecache('00changelog.i')
354 def changelog(self):
354 def changelog(self):
355 c = changelog.changelog(self.sopener)
355 c = changelog.changelog(self.sopener)
356 if 'HG_PENDING' in os.environ:
356 if 'HG_PENDING' in os.environ:
357 p = os.environ['HG_PENDING']
357 p = os.environ['HG_PENDING']
358 if p.startswith(self.root):
358 if p.startswith(self.root):
359 c.readpending('00changelog.i.a')
359 c.readpending('00changelog.i.a')
360 return c
360 return c
361
361
362 @storecache('00manifest.i')
362 @storecache('00manifest.i')
363 def manifest(self):
363 def manifest(self):
364 return manifest.manifest(self.sopener)
364 return manifest.manifest(self.sopener)
365
365
366 @repofilecache('dirstate')
366 @repofilecache('dirstate')
367 def dirstate(self):
367 def dirstate(self):
368 warned = [0]
368 warned = [0]
369 def validate(node):
369 def validate(node):
370 try:
370 try:
371 self.changelog.rev(node)
371 self.changelog.rev(node)
372 return node
372 return node
373 except error.LookupError:
373 except error.LookupError:
374 if not warned[0]:
374 if not warned[0]:
375 warned[0] = True
375 warned[0] = True
376 self.ui.warn(_("warning: ignoring unknown"
376 self.ui.warn(_("warning: ignoring unknown"
377 " working parent %s!\n") % short(node))
377 " working parent %s!\n") % short(node))
378 return nullid
378 return nullid
379
379
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381
381
382 def __getitem__(self, changeid):
382 def __getitem__(self, changeid):
383 if changeid is None:
383 if changeid is None:
384 return context.workingctx(self)
384 return context.workingctx(self)
385 return context.changectx(self, changeid)
385 return context.changectx(self, changeid)
386
386
387 def __contains__(self, changeid):
387 def __contains__(self, changeid):
388 try:
388 try:
389 return bool(self.lookup(changeid))
389 return bool(self.lookup(changeid))
390 except error.RepoLookupError:
390 except error.RepoLookupError:
391 return False
391 return False
392
392
393 def __nonzero__(self):
393 def __nonzero__(self):
394 return True
394 return True
395
395
396 def __len__(self):
396 def __len__(self):
397 return len(self.changelog)
397 return len(self.changelog)
398
398
399 def __iter__(self):
399 def __iter__(self):
400 return iter(self.changelog)
400 return iter(self.changelog)
401
401
402 def revs(self, expr, *args):
402 def revs(self, expr, *args):
403 '''Return a list of revisions matching the given revset'''
403 '''Return a list of revisions matching the given revset'''
404 expr = revset.formatspec(expr, *args)
404 expr = revset.formatspec(expr, *args)
405 m = revset.match(None, expr)
405 m = revset.match(None, expr)
406 return [r for r in m(self, list(self))]
406 return [r for r in m(self, list(self))]
407
407
408 def set(self, expr, *args):
408 def set(self, expr, *args):
409 '''
409 '''
410 Yield a context for each matching revision, after doing arg
410 Yield a context for each matching revision, after doing arg
411 replacement via revset.formatspec
411 replacement via revset.formatspec
412 '''
412 '''
413 for r in self.revs(expr, *args):
413 for r in self.revs(expr, *args):
414 yield self[r]
414 yield self[r]
415
415
416 def url(self):
416 def url(self):
417 return 'file:' + self.root
417 return 'file:' + self.root
418
418
419 def hook(self, name, throw=False, **args):
419 def hook(self, name, throw=False, **args):
420 return hook.hook(self.ui, self, name, throw, **args)
420 return hook.hook(self.ui, self, name, throw, **args)
421
421
422 @unfilteredmethod
422 @unfilteredmethod
423 def _tag(self, names, node, message, local, user, date, extra={}):
423 def _tag(self, names, node, message, local, user, date, extra={}):
424 if isinstance(names, str):
424 if isinstance(names, str):
425 names = (names,)
425 names = (names,)
426
426
427 branches = self.branchmap()
427 branches = self.branchmap()
428 for name in names:
428 for name in names:
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 local=local)
430 local=local)
431 if name in branches:
431 if name in branches:
432 self.ui.warn(_("warning: tag %s conflicts with existing"
432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 " branch name\n") % name)
433 " branch name\n") % name)
434
434
435 def writetags(fp, names, munge, prevtags):
435 def writetags(fp, names, munge, prevtags):
436 fp.seek(0, 2)
436 fp.seek(0, 2)
437 if prevtags and prevtags[-1] != '\n':
437 if prevtags and prevtags[-1] != '\n':
438 fp.write('\n')
438 fp.write('\n')
439 for name in names:
439 for name in names:
440 m = munge and munge(name) or name
440 m = munge and munge(name) or name
441 if (self._tagscache.tagtypes and
441 if (self._tagscache.tagtypes and
442 name in self._tagscache.tagtypes):
442 name in self._tagscache.tagtypes):
443 old = self.tags().get(name, nullid)
443 old = self.tags().get(name, nullid)
444 fp.write('%s %s\n' % (hex(old), m))
444 fp.write('%s %s\n' % (hex(old), m))
445 fp.write('%s %s\n' % (hex(node), m))
445 fp.write('%s %s\n' % (hex(node), m))
446 fp.close()
446 fp.close()
447
447
448 prevtags = ''
448 prevtags = ''
449 if local:
449 if local:
450 try:
450 try:
451 fp = self.opener('localtags', 'r+')
451 fp = self.opener('localtags', 'r+')
452 except IOError:
452 except IOError:
453 fp = self.opener('localtags', 'a')
453 fp = self.opener('localtags', 'a')
454 else:
454 else:
455 prevtags = fp.read()
455 prevtags = fp.read()
456
456
457 # local tags are stored in the current charset
457 # local tags are stored in the current charset
458 writetags(fp, names, None, prevtags)
458 writetags(fp, names, None, prevtags)
459 for name in names:
459 for name in names:
460 self.hook('tag', node=hex(node), tag=name, local=local)
460 self.hook('tag', node=hex(node), tag=name, local=local)
461 return
461 return
462
462
463 try:
463 try:
464 fp = self.wfile('.hgtags', 'rb+')
464 fp = self.wfile('.hgtags', 'rb+')
465 except IOError, e:
465 except IOError, e:
466 if e.errno != errno.ENOENT:
466 if e.errno != errno.ENOENT:
467 raise
467 raise
468 fp = self.wfile('.hgtags', 'ab')
468 fp = self.wfile('.hgtags', 'ab')
469 else:
469 else:
470 prevtags = fp.read()
470 prevtags = fp.read()
471
471
472 # committed tags are stored in UTF-8
472 # committed tags are stored in UTF-8
473 writetags(fp, names, encoding.fromlocal, prevtags)
473 writetags(fp, names, encoding.fromlocal, prevtags)
474
474
475 fp.close()
475 fp.close()
476
476
477 self.invalidatecaches()
477 self.invalidatecaches()
478
478
479 if '.hgtags' not in self.dirstate:
479 if '.hgtags' not in self.dirstate:
480 self[None].add(['.hgtags'])
480 self[None].add(['.hgtags'])
481
481
482 m = matchmod.exact(self.root, '', ['.hgtags'])
482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484
484
485 for name in names:
485 for name in names:
486 self.hook('tag', node=hex(node), tag=name, local=local)
486 self.hook('tag', node=hex(node), tag=name, local=local)
487
487
488 return tagnode
488 return tagnode
489
489
490 def tag(self, names, node, message, local, user, date):
490 def tag(self, names, node, message, local, user, date):
491 '''tag a revision with one or more symbolic names.
491 '''tag a revision with one or more symbolic names.
492
492
493 names is a list of strings or, when adding a single tag, names may be a
493 names is a list of strings or, when adding a single tag, names may be a
494 string.
494 string.
495
495
496 if local is True, the tags are stored in a per-repository file.
496 if local is True, the tags are stored in a per-repository file.
497 otherwise, they are stored in the .hgtags file, and a new
497 otherwise, they are stored in the .hgtags file, and a new
498 changeset is committed with the change.
498 changeset is committed with the change.
499
499
500 keyword arguments:
500 keyword arguments:
501
501
502 local: whether to store tags in non-version-controlled file
502 local: whether to store tags in non-version-controlled file
503 (default False)
503 (default False)
504
504
505 message: commit message to use if committing
505 message: commit message to use if committing
506
506
507 user: name of user to use if committing
507 user: name of user to use if committing
508
508
509 date: date tuple to use if committing'''
509 date: date tuple to use if committing'''
510
510
511 if not local:
511 if not local:
512 for x in self.status()[:5]:
512 for x in self.status()[:5]:
513 if '.hgtags' in x:
513 if '.hgtags' in x:
514 raise util.Abort(_('working copy of .hgtags is changed '
514 raise util.Abort(_('working copy of .hgtags is changed '
515 '(please commit .hgtags manually)'))
515 '(please commit .hgtags manually)'))
516
516
517 self.tags() # instantiate the cache
517 self.tags() # instantiate the cache
518 self._tag(names, node, message, local, user, date)
518 self._tag(names, node, message, local, user, date)
519
519
520 @filteredpropertycache
520 @filteredpropertycache
521 def _tagscache(self):
521 def _tagscache(self):
522 '''Returns a tagscache object that contains various tags related
522 '''Returns a tagscache object that contains various tags related
523 caches.'''
523 caches.'''
524
524
525 # This simplifies its cache management by having one decorated
525 # This simplifies its cache management by having one decorated
526 # function (this one) and the rest simply fetch things from it.
526 # function (this one) and the rest simply fetch things from it.
527 class tagscache(object):
527 class tagscache(object):
528 def __init__(self):
528 def __init__(self):
529 # These two define the set of tags for this repository. tags
529 # These two define the set of tags for this repository. tags
530 # maps tag name to node; tagtypes maps tag name to 'global' or
530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 # 'local'. (Global tags are defined by .hgtags across all
531 # 'local'. (Global tags are defined by .hgtags across all
532 # heads, and local tags are defined in .hg/localtags.)
532 # heads, and local tags are defined in .hg/localtags.)
533 # They constitute the in-memory cache of tags.
533 # They constitute the in-memory cache of tags.
534 self.tags = self.tagtypes = None
534 self.tags = self.tagtypes = None
535
535
536 self.nodetagscache = self.tagslist = None
536 self.nodetagscache = self.tagslist = None
537
537
538 cache = tagscache()
538 cache = tagscache()
539 cache.tags, cache.tagtypes = self._findtags()
539 cache.tags, cache.tagtypes = self._findtags()
540
540
541 return cache
541 return cache
542
542
543 def tags(self):
543 def tags(self):
544 '''return a mapping of tag to node'''
544 '''return a mapping of tag to node'''
545 t = {}
545 t = {}
546 if self.changelog.filteredrevs:
546 if self.changelog.filteredrevs:
547 tags, tt = self._findtags()
547 tags, tt = self._findtags()
548 else:
548 else:
549 tags = self._tagscache.tags
549 tags = self._tagscache.tags
550 for k, v in tags.iteritems():
550 for k, v in tags.iteritems():
551 try:
551 try:
552 # ignore tags to unknown nodes
552 # ignore tags to unknown nodes
553 self.changelog.rev(v)
553 self.changelog.rev(v)
554 t[k] = v
554 t[k] = v
555 except (error.LookupError, ValueError):
555 except (error.LookupError, ValueError):
556 pass
556 pass
557 return t
557 return t
558
558
559 def _findtags(self):
559 def _findtags(self):
560 '''Do the hard work of finding tags. Return a pair of dicts
560 '''Do the hard work of finding tags. Return a pair of dicts
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 maps tag name to a string like \'global\' or \'local\'.
562 maps tag name to a string like \'global\' or \'local\'.
563 Subclasses or extensions are free to add their own tags, but
563 Subclasses or extensions are free to add their own tags, but
564 should be aware that the returned dicts will be retained for the
564 should be aware that the returned dicts will be retained for the
565 duration of the localrepo object.'''
565 duration of the localrepo object.'''
566
566
567 # XXX what tagtype should subclasses/extensions use? Currently
567 # XXX what tagtype should subclasses/extensions use? Currently
568 # mq and bookmarks add tags, but do not set the tagtype at all.
568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 # Should each extension invent its own tag type? Should there
569 # Should each extension invent its own tag type? Should there
570 # be one tagtype for all such "virtual" tags? Or is the status
570 # be one tagtype for all such "virtual" tags? Or is the status
571 # quo fine?
571 # quo fine?
572
572
573 alltags = {} # map tag name to (node, hist)
573 alltags = {} # map tag name to (node, hist)
574 tagtypes = {}
574 tagtypes = {}
575
575
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578
578
579 # Build the return dicts. Have to re-encode tag names because
579 # Build the return dicts. Have to re-encode tag names because
580 # the tags module always uses UTF-8 (in order not to lose info
580 # the tags module always uses UTF-8 (in order not to lose info
581 # writing to the cache), but the rest of Mercurial wants them in
581 # writing to the cache), but the rest of Mercurial wants them in
582 # local encoding.
582 # local encoding.
583 tags = {}
583 tags = {}
584 for (name, (node, hist)) in alltags.iteritems():
584 for (name, (node, hist)) in alltags.iteritems():
585 if node != nullid:
585 if node != nullid:
586 tags[encoding.tolocal(name)] = node
586 tags[encoding.tolocal(name)] = node
587 tags['tip'] = self.changelog.tip()
587 tags['tip'] = self.changelog.tip()
588 tagtypes = dict([(encoding.tolocal(name), value)
588 tagtypes = dict([(encoding.tolocal(name), value)
589 for (name, value) in tagtypes.iteritems()])
589 for (name, value) in tagtypes.iteritems()])
590 return (tags, tagtypes)
590 return (tags, tagtypes)
591
591
592 def tagtype(self, tagname):
592 def tagtype(self, tagname):
593 '''
593 '''
594 return the type of the given tag. result can be:
594 return the type of the given tag. result can be:
595
595
596 'local' : a local tag
596 'local' : a local tag
597 'global' : a global tag
597 'global' : a global tag
598 None : tag does not exist
598 None : tag does not exist
599 '''
599 '''
600
600
601 return self._tagscache.tagtypes.get(tagname)
601 return self._tagscache.tagtypes.get(tagname)
602
602
603 def tagslist(self):
603 def tagslist(self):
604 '''return a list of tags ordered by revision'''
604 '''return a list of tags ordered by revision'''
605 if not self._tagscache.tagslist:
605 if not self._tagscache.tagslist:
606 l = []
606 l = []
607 for t, n in self.tags().iteritems():
607 for t, n in self.tags().iteritems():
608 r = self.changelog.rev(n)
608 r = self.changelog.rev(n)
609 l.append((r, t, n))
609 l.append((r, t, n))
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611
611
612 return self._tagscache.tagslist
612 return self._tagscache.tagslist
613
613
614 def nodetags(self, node):
614 def nodetags(self, node):
615 '''return the tags associated with a node'''
615 '''return the tags associated with a node'''
616 if not self._tagscache.nodetagscache:
616 if not self._tagscache.nodetagscache:
617 nodetagscache = {}
617 nodetagscache = {}
618 for t, n in self._tagscache.tags.iteritems():
618 for t, n in self._tagscache.tags.iteritems():
619 nodetagscache.setdefault(n, []).append(t)
619 nodetagscache.setdefault(n, []).append(t)
620 for tags in nodetagscache.itervalues():
620 for tags in nodetagscache.itervalues():
621 tags.sort()
621 tags.sort()
622 self._tagscache.nodetagscache = nodetagscache
622 self._tagscache.nodetagscache = nodetagscache
623 return self._tagscache.nodetagscache.get(node, [])
623 return self._tagscache.nodetagscache.get(node, [])
624
624
625 def nodebookmarks(self, node):
625 def nodebookmarks(self, node):
626 marks = []
626 marks = []
627 for bookmark, n in self._bookmarks.iteritems():
627 for bookmark, n in self._bookmarks.iteritems():
628 if n == node:
628 if n == node:
629 marks.append(bookmark)
629 marks.append(bookmark)
630 return sorted(marks)
630 return sorted(marks)
631
631
632 def branchmap(self):
632 def branchmap(self):
633 '''returns a dictionary {branch: [branchheads]}'''
633 '''returns a dictionary {branch: [branchheads]}'''
634 branchmap.updatecache(self)
634 branchmap.updatecache(self)
635 return self._branchcaches[self.filtername]
635 return self._branchcaches[self.filtername]
636
636
637
637
638 def _branchtip(self, heads):
638 def _branchtip(self, heads):
639 '''return the tipmost branch head in heads'''
639 '''return the tipmost branch head in heads'''
640 tip = heads[-1]
640 tip = heads[-1]
641 for h in reversed(heads):
641 for h in reversed(heads):
642 if not self[h].closesbranch():
642 if not self[h].closesbranch():
643 tip = h
643 tip = h
644 break
644 break
645 return tip
645 return tip
646
646
647 def branchtip(self, branch):
647 def branchtip(self, branch):
648 '''return the tip node for a given branch'''
648 '''return the tip node for a given branch'''
649 if branch not in self.branchmap():
649 if branch not in self.branchmap():
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 return self._branchtip(self.branchmap()[branch])
651 return self._branchtip(self.branchmap()[branch])
652
652
653 def branchtags(self):
653 def branchtags(self):
654 '''return a dict where branch names map to the tipmost head of
654 '''return a dict where branch names map to the tipmost head of
655 the branch, open heads come before closed'''
655 the branch, open heads come before closed'''
656 bt = {}
656 bt = {}
657 for bn, heads in self.branchmap().iteritems():
657 for bn, heads in self.branchmap().iteritems():
658 bt[bn] = self._branchtip(heads)
658 bt[bn] = self._branchtip(heads)
659 return bt
659 return bt
660
660
661 def lookup(self, key):
661 def lookup(self, key):
662 return self[key].node()
662 return self[key].node()
663
663
664 def lookupbranch(self, key, remote=None):
664 def lookupbranch(self, key, remote=None):
665 repo = remote or self
665 repo = remote or self
666 if key in repo.branchmap():
666 if key in repo.branchmap():
667 return key
667 return key
668
668
669 repo = (remote and remote.local()) and remote or self
669 repo = (remote and remote.local()) and remote or self
670 return repo[key].branch()
670 return repo[key].branch()
671
671
672 def known(self, nodes):
672 def known(self, nodes):
673 nm = self.changelog.nodemap
673 nm = self.changelog.nodemap
674 pc = self._phasecache
674 pc = self._phasecache
675 result = []
675 result = []
676 for n in nodes:
676 for n in nodes:
677 r = nm.get(n)
677 r = nm.get(n)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 result.append(resp)
679 result.append(resp)
680 return result
680 return result
681
681
682 def local(self):
682 def local(self):
683 return self
683 return self
684
684
685 def cancopy(self):
685 def cancopy(self):
686 return self.local() # so statichttprepo's override of local() works
686 return self.local() # so statichttprepo's override of local() works
687
687
688 def join(self, f):
688 def join(self, f):
689 return os.path.join(self.path, f)
689 return os.path.join(self.path, f)
690
690
691 def wjoin(self, f):
691 def wjoin(self, f):
692 return os.path.join(self.root, f)
692 return os.path.join(self.root, f)
693
693
694 def file(self, f):
694 def file(self, f):
695 if f[0] == '/':
695 if f[0] == '/':
696 f = f[1:]
696 f = f[1:]
697 return filelog.filelog(self.sopener, f)
697 return filelog.filelog(self.sopener, f)
698
698
699 def changectx(self, changeid):
699 def changectx(self, changeid):
700 return self[changeid]
700 return self[changeid]
701
701
702 def parents(self, changeid=None):
702 def parents(self, changeid=None):
703 '''get list of changectxs for parents of changeid'''
703 '''get list of changectxs for parents of changeid'''
704 return self[changeid].parents()
704 return self[changeid].parents()
705
705
706 def setparents(self, p1, p2=nullid):
706 def setparents(self, p1, p2=nullid):
707 copies = self.dirstate.setparents(p1, p2)
707 copies = self.dirstate.setparents(p1, p2)
708 if copies:
708 if copies:
709 # Adjust copy records, the dirstate cannot do it, it
709 # Adjust copy records, the dirstate cannot do it, it
710 # requires access to parents manifests. Preserve them
710 # requires access to parents manifests. Preserve them
711 # only for entries added to first parent.
711 # only for entries added to first parent.
712 pctx = self[p1]
712 pctx = self[p1]
713 for f in copies:
713 for f in copies:
714 if f not in pctx and copies[f] in pctx:
714 if f not in pctx and copies[f] in pctx:
715 self.dirstate.copy(copies[f], f)
715 self.dirstate.copy(copies[f], f)
716
716
717 def filectx(self, path, changeid=None, fileid=None):
717 def filectx(self, path, changeid=None, fileid=None):
718 """changeid can be a changeset revision, node, or tag.
718 """changeid can be a changeset revision, node, or tag.
719 fileid can be a file revision or node."""
719 fileid can be a file revision or node."""
720 return context.filectx(self, path, changeid, fileid)
720 return context.filectx(self, path, changeid, fileid)
721
721
722 def getcwd(self):
722 def getcwd(self):
723 return self.dirstate.getcwd()
723 return self.dirstate.getcwd()
724
724
725 def pathto(self, f, cwd=None):
725 def pathto(self, f, cwd=None):
726 return self.dirstate.pathto(f, cwd)
726 return self.dirstate.pathto(f, cwd)
727
727
728 def wfile(self, f, mode='r'):
728 def wfile(self, f, mode='r'):
729 return self.wopener(f, mode)
729 return self.wopener(f, mode)
730
730
731 def _link(self, f):
731 def _link(self, f):
732 return os.path.islink(self.wjoin(f))
732 return os.path.islink(self.wjoin(f))
733
733
734 def _loadfilter(self, filter):
734 def _loadfilter(self, filter):
735 if filter not in self.filterpats:
735 if filter not in self.filterpats:
736 l = []
736 l = []
737 for pat, cmd in self.ui.configitems(filter):
737 for pat, cmd in self.ui.configitems(filter):
738 if cmd == '!':
738 if cmd == '!':
739 continue
739 continue
740 mf = matchmod.match(self.root, '', [pat])
740 mf = matchmod.match(self.root, '', [pat])
741 fn = None
741 fn = None
742 params = cmd
742 params = cmd
743 for name, filterfn in self._datafilters.iteritems():
743 for name, filterfn in self._datafilters.iteritems():
744 if cmd.startswith(name):
744 if cmd.startswith(name):
745 fn = filterfn
745 fn = filterfn
746 params = cmd[len(name):].lstrip()
746 params = cmd[len(name):].lstrip()
747 break
747 break
748 if not fn:
748 if not fn:
749 fn = lambda s, c, **kwargs: util.filter(s, c)
749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 # Wrap old filters not supporting keyword arguments
750 # Wrap old filters not supporting keyword arguments
751 if not inspect.getargspec(fn)[2]:
751 if not inspect.getargspec(fn)[2]:
752 oldfn = fn
752 oldfn = fn
753 fn = lambda s, c, **kwargs: oldfn(s, c)
753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 l.append((mf, fn, params))
754 l.append((mf, fn, params))
755 self.filterpats[filter] = l
755 self.filterpats[filter] = l
756 return self.filterpats[filter]
756 return self.filterpats[filter]
757
757
758 def _filter(self, filterpats, filename, data):
758 def _filter(self, filterpats, filename, data):
759 for mf, fn, cmd in filterpats:
759 for mf, fn, cmd in filterpats:
760 if mf(filename):
760 if mf(filename):
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 break
763 break
764
764
765 return data
765 return data
766
766
767 @unfilteredpropertycache
767 @unfilteredpropertycache
768 def _encodefilterpats(self):
768 def _encodefilterpats(self):
769 return self._loadfilter('encode')
769 return self._loadfilter('encode')
770
770
771 @unfilteredpropertycache
771 @unfilteredpropertycache
772 def _decodefilterpats(self):
772 def _decodefilterpats(self):
773 return self._loadfilter('decode')
773 return self._loadfilter('decode')
774
774
775 def adddatafilter(self, name, filter):
775 def adddatafilter(self, name, filter):
776 self._datafilters[name] = filter
776 self._datafilters[name] = filter
777
777
778 def wread(self, filename):
778 def wread(self, filename):
779 if self._link(filename):
779 if self._link(filename):
780 data = os.readlink(self.wjoin(filename))
780 data = os.readlink(self.wjoin(filename))
781 else:
781 else:
782 data = self.wopener.read(filename)
782 data = self.wopener.read(filename)
783 return self._filter(self._encodefilterpats, filename, data)
783 return self._filter(self._encodefilterpats, filename, data)
784
784
785 def wwrite(self, filename, data, flags):
785 def wwrite(self, filename, data, flags):
786 data = self._filter(self._decodefilterpats, filename, data)
786 data = self._filter(self._decodefilterpats, filename, data)
787 if 'l' in flags:
787 if 'l' in flags:
788 self.wopener.symlink(data, filename)
788 self.wopener.symlink(data, filename)
789 else:
789 else:
790 self.wopener.write(filename, data)
790 self.wopener.write(filename, data)
791 if 'x' in flags:
791 if 'x' in flags:
792 util.setflags(self.wjoin(filename), False, True)
792 util.setflags(self.wjoin(filename), False, True)
793
793
794 def wwritedata(self, filename, data):
794 def wwritedata(self, filename, data):
795 return self._filter(self._decodefilterpats, filename, data)
795 return self._filter(self._decodefilterpats, filename, data)
796
796
797 def transaction(self, desc):
797 def transaction(self, desc):
798 tr = self._transref and self._transref() or None
798 tr = self._transref and self._transref() or None
799 if tr and tr.running():
799 if tr and tr.running():
800 return tr.nest()
800 return tr.nest()
801
801
802 # abort here if the journal already exists
802 # abort here if the journal already exists
803 if os.path.exists(self.sjoin("journal")):
803 if os.path.exists(self.sjoin("journal")):
804 raise error.RepoError(
804 raise error.RepoError(
805 _("abandoned transaction found - run hg recover"))
805 _("abandoned transaction found - run hg recover"))
806
806
807 self._writejournal(desc)
807 self._writejournal(desc)
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809
809
810 tr = transaction.transaction(self.ui.warn, self.sopener,
810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 self.sjoin("journal"),
811 self.sjoin("journal"),
812 aftertrans(renames),
812 aftertrans(renames),
813 self.store.createmode)
813 self.store.createmode)
814 self._transref = weakref.ref(tr)
814 self._transref = weakref.ref(tr)
815 return tr
815 return tr
816
816
817 def _journalfiles(self):
817 def _journalfiles(self):
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 self.join('journal.branch'), self.join('journal.desc'),
819 self.join('journal.branch'), self.join('journal.desc'),
820 self.join('journal.bookmarks'),
820 self.join('journal.bookmarks'),
821 self.sjoin('journal.phaseroots'))
821 self.sjoin('journal.phaseroots'))
822
822
823 def undofiles(self):
823 def undofiles(self):
824 return [undoname(x) for x in self._journalfiles()]
824 return [undoname(x) for x in self._journalfiles()]
825
825
826 def _writejournal(self, desc):
826 def _writejournal(self, desc):
827 self.opener.write("journal.dirstate",
827 self.opener.write("journal.dirstate",
828 self.opener.tryread("dirstate"))
828 self.opener.tryread("dirstate"))
829 self.opener.write("journal.branch",
829 self.opener.write("journal.branch",
830 encoding.fromlocal(self.dirstate.branch()))
830 encoding.fromlocal(self.dirstate.branch()))
831 self.opener.write("journal.desc",
831 self.opener.write("journal.desc",
832 "%d\n%s\n" % (len(self), desc))
832 "%d\n%s\n" % (len(self), desc))
833 self.opener.write("journal.bookmarks",
833 self.opener.write("journal.bookmarks",
834 self.opener.tryread("bookmarks"))
834 self.opener.tryread("bookmarks"))
835 self.sopener.write("journal.phaseroots",
835 self.sopener.write("journal.phaseroots",
836 self.sopener.tryread("phaseroots"))
836 self.sopener.tryread("phaseroots"))
837
837
838 def recover(self):
838 def recover(self):
839 lock = self.lock()
839 lock = self.lock()
840 try:
840 try:
841 if os.path.exists(self.sjoin("journal")):
841 if os.path.exists(self.sjoin("journal")):
842 self.ui.status(_("rolling back interrupted transaction\n"))
842 self.ui.status(_("rolling back interrupted transaction\n"))
843 transaction.rollback(self.sopener, self.sjoin("journal"),
843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 self.ui.warn)
844 self.ui.warn)
845 self.invalidate()
845 self.invalidate()
846 return True
846 return True
847 else:
847 else:
848 self.ui.warn(_("no interrupted transaction available\n"))
848 self.ui.warn(_("no interrupted transaction available\n"))
849 return False
849 return False
850 finally:
850 finally:
851 lock.release()
851 lock.release()
852
852
853 def rollback(self, dryrun=False, force=False):
853 def rollback(self, dryrun=False, force=False):
854 wlock = lock = None
854 wlock = lock = None
855 try:
855 try:
856 wlock = self.wlock()
856 wlock = self.wlock()
857 lock = self.lock()
857 lock = self.lock()
858 if os.path.exists(self.sjoin("undo")):
858 if os.path.exists(self.sjoin("undo")):
859 return self._rollback(dryrun, force)
859 return self._rollback(dryrun, force)
860 else:
860 else:
861 self.ui.warn(_("no rollback information available\n"))
861 self.ui.warn(_("no rollback information available\n"))
862 return 1
862 return 1
863 finally:
863 finally:
864 release(lock, wlock)
864 release(lock, wlock)
865
865
866 @unfilteredmethod # Until we get smarter cache management
866 @unfilteredmethod # Until we get smarter cache management
867 def _rollback(self, dryrun, force):
867 def _rollback(self, dryrun, force):
868 ui = self.ui
868 ui = self.ui
869 try:
869 try:
870 args = self.opener.read('undo.desc').splitlines()
870 args = self.opener.read('undo.desc').splitlines()
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 if len(args) >= 3:
872 if len(args) >= 3:
873 detail = args[2]
873 detail = args[2]
874 oldtip = oldlen - 1
874 oldtip = oldlen - 1
875
875
876 if detail and ui.verbose:
876 if detail and ui.verbose:
877 msg = (_('repository tip rolled back to revision %s'
877 msg = (_('repository tip rolled back to revision %s'
878 ' (undo %s: %s)\n')
878 ' (undo %s: %s)\n')
879 % (oldtip, desc, detail))
879 % (oldtip, desc, detail))
880 else:
880 else:
881 msg = (_('repository tip rolled back to revision %s'
881 msg = (_('repository tip rolled back to revision %s'
882 ' (undo %s)\n')
882 ' (undo %s)\n')
883 % (oldtip, desc))
883 % (oldtip, desc))
884 except IOError:
884 except IOError:
885 msg = _('rolling back unknown transaction\n')
885 msg = _('rolling back unknown transaction\n')
886 desc = None
886 desc = None
887
887
888 if not force and self['.'] != self['tip'] and desc == 'commit':
888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 raise util.Abort(
889 raise util.Abort(
890 _('rollback of last commit while not checked out '
890 _('rollback of last commit while not checked out '
891 'may lose data'), hint=_('use -f to force'))
891 'may lose data'), hint=_('use -f to force'))
892
892
893 ui.status(msg)
893 ui.status(msg)
894 if dryrun:
894 if dryrun:
895 return 0
895 return 0
896
896
897 parents = self.dirstate.parents()
897 parents = self.dirstate.parents()
898 self.destroying()
898 self.destroying()
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 if os.path.exists(self.join('undo.bookmarks')):
900 if os.path.exists(self.join('undo.bookmarks')):
901 util.rename(self.join('undo.bookmarks'),
901 util.rename(self.join('undo.bookmarks'),
902 self.join('bookmarks'))
902 self.join('bookmarks'))
903 if os.path.exists(self.sjoin('undo.phaseroots')):
903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 util.rename(self.sjoin('undo.phaseroots'),
904 util.rename(self.sjoin('undo.phaseroots'),
905 self.sjoin('phaseroots'))
905 self.sjoin('phaseroots'))
906 self.invalidate()
906 self.invalidate()
907
907
908 parentgone = (parents[0] not in self.changelog.nodemap or
908 parentgone = (parents[0] not in self.changelog.nodemap or
909 parents[1] not in self.changelog.nodemap)
909 parents[1] not in self.changelog.nodemap)
910 if parentgone:
910 if parentgone:
911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
912 try:
912 try:
913 branch = self.opener.read('undo.branch')
913 branch = self.opener.read('undo.branch')
914 self.dirstate.setbranch(encoding.tolocal(branch))
914 self.dirstate.setbranch(encoding.tolocal(branch))
915 except IOError:
915 except IOError:
916 ui.warn(_('named branch could not be reset: '
916 ui.warn(_('named branch could not be reset: '
917 'current branch is still \'%s\'\n')
917 'current branch is still \'%s\'\n')
918 % self.dirstate.branch())
918 % self.dirstate.branch())
919
919
920 self.dirstate.invalidate()
920 self.dirstate.invalidate()
921 parents = tuple([p.rev() for p in self.parents()])
921 parents = tuple([p.rev() for p in self.parents()])
922 if len(parents) > 1:
922 if len(parents) > 1:
923 ui.status(_('working directory now based on '
923 ui.status(_('working directory now based on '
924 'revisions %d and %d\n') % parents)
924 'revisions %d and %d\n') % parents)
925 else:
925 else:
926 ui.status(_('working directory now based on '
926 ui.status(_('working directory now based on '
927 'revision %d\n') % parents)
927 'revision %d\n') % parents)
928 # TODO: if we know which new heads may result from this rollback, pass
928 # TODO: if we know which new heads may result from this rollback, pass
929 # them to destroy(), which will prevent the branchhead cache from being
929 # them to destroy(), which will prevent the branchhead cache from being
930 # invalidated.
930 # invalidated.
931 self.destroyed()
931 self.destroyed()
932 return 0
932 return 0
933
933
934 def invalidatecaches(self):
934 def invalidatecaches(self):
935
935
936 if '_tagscache' in vars(self):
936 if '_tagscache' in vars(self):
937 # can't use delattr on proxy
937 # can't use delattr on proxy
938 del self.__dict__['_tagscache']
938 del self.__dict__['_tagscache']
939
939
940 self.unfiltered()._branchcaches.clear()
940 self.unfiltered()._branchcaches.clear()
941 self.invalidatevolatilesets()
941 self.invalidatevolatilesets()
942
942
943 def invalidatevolatilesets(self):
943 def invalidatevolatilesets(self):
944 self.filteredrevcache.clear()
944 self.filteredrevcache.clear()
945 obsolete.clearobscaches(self)
945 obsolete.clearobscaches(self)
946
946
947 def invalidatedirstate(self):
947 def invalidatedirstate(self):
948 '''Invalidates the dirstate, causing the next call to dirstate
948 '''Invalidates the dirstate, causing the next call to dirstate
949 to check if it was modified since the last time it was read,
949 to check if it was modified since the last time it was read,
950 rereading it if it has.
950 rereading it if it has.
951
951
952 This is different to dirstate.invalidate() that it doesn't always
952 This is different to dirstate.invalidate() that it doesn't always
953 rereads the dirstate. Use dirstate.invalidate() if you want to
953 rereads the dirstate. Use dirstate.invalidate() if you want to
954 explicitly read the dirstate again (i.e. restoring it to a previous
954 explicitly read the dirstate again (i.e. restoring it to a previous
955 known good state).'''
955 known good state).'''
956 if hasunfilteredcache(self, 'dirstate'):
956 if hasunfilteredcache(self, 'dirstate'):
957 for k in self.dirstate._filecache:
957 for k in self.dirstate._filecache:
958 try:
958 try:
959 delattr(self.dirstate, k)
959 delattr(self.dirstate, k)
960 except AttributeError:
960 except AttributeError:
961 pass
961 pass
962 delattr(self.unfiltered(), 'dirstate')
962 delattr(self.unfiltered(), 'dirstate')
963
963
964 def invalidate(self):
964 def invalidate(self):
965 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
965 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
966 for k in self._filecache:
966 for k in self._filecache:
967 # dirstate is invalidated separately in invalidatedirstate()
967 # dirstate is invalidated separately in invalidatedirstate()
968 if k == 'dirstate':
968 if k == 'dirstate':
969 continue
969 continue
970
970
971 try:
971 try:
972 delattr(unfiltered, k)
972 delattr(unfiltered, k)
973 except AttributeError:
973 except AttributeError:
974 pass
974 pass
975 self.invalidatecaches()
975 self.invalidatecaches()
976
976
977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
978 try:
978 try:
979 l = lock.lock(lockname, 0, releasefn, desc=desc)
979 l = lock.lock(lockname, 0, releasefn, desc=desc)
980 except error.LockHeld, inst:
980 except error.LockHeld, inst:
981 if not wait:
981 if not wait:
982 raise
982 raise
983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
984 (desc, inst.locker))
984 (desc, inst.locker))
985 # default to 600 seconds timeout
985 # default to 600 seconds timeout
986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
987 releasefn, desc=desc)
987 releasefn, desc=desc)
988 if acquirefn:
988 if acquirefn:
989 acquirefn()
989 acquirefn()
990 return l
990 return l
991
991
992 def _afterlock(self, callback):
992 def _afterlock(self, callback):
993 """add a callback to the current repository lock.
993 """add a callback to the current repository lock.
994
994
995 The callback will be executed on lock release."""
995 The callback will be executed on lock release."""
996 l = self._lockref and self._lockref()
996 l = self._lockref and self._lockref()
997 if l:
997 if l:
998 l.postrelease.append(callback)
998 l.postrelease.append(callback)
999 else:
999 else:
1000 callback()
1000 callback()
1001
1001
1002 def lock(self, wait=True):
1002 def lock(self, wait=True):
1003 '''Lock the repository store (.hg/store) and return a weak reference
1003 '''Lock the repository store (.hg/store) and return a weak reference
1004 to the lock. Use this before modifying the store (e.g. committing or
1004 to the lock. Use this before modifying the store (e.g. committing or
1005 stripping). If you are opening a transaction, get a lock as well.)'''
1005 stripping). If you are opening a transaction, get a lock as well.)'''
1006 l = self._lockref and self._lockref()
1006 l = self._lockref and self._lockref()
1007 if l is not None and l.held:
1007 if l is not None and l.held:
1008 l.lock()
1008 l.lock()
1009 return l
1009 return l
1010
1010
1011 def unlock():
1011 def unlock():
1012 self.store.write()
1012 self.store.write()
1013 if hasunfilteredcache(self, '_phasecache'):
1013 if hasunfilteredcache(self, '_phasecache'):
1014 self._phasecache.write()
1014 self._phasecache.write()
1015 for k, ce in self._filecache.items():
1015 for k, ce in self._filecache.items():
1016 if k == 'dirstate' or k not in self.__dict__:
1016 if k == 'dirstate' or k not in self.__dict__:
1017 continue
1017 continue
1018 ce.refresh()
1018 ce.refresh()
1019
1019
1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1021 self.invalidate, _('repository %s') % self.origroot)
1021 self.invalidate, _('repository %s') % self.origroot)
1022 self._lockref = weakref.ref(l)
1022 self._lockref = weakref.ref(l)
1023 return l
1023 return l
1024
1024
1025 def wlock(self, wait=True):
1025 def wlock(self, wait=True):
1026 '''Lock the non-store parts of the repository (everything under
1026 '''Lock the non-store parts of the repository (everything under
1027 .hg except .hg/store) and return a weak reference to the lock.
1027 .hg except .hg/store) and return a weak reference to the lock.
1028 Use this before modifying files in .hg.'''
1028 Use this before modifying files in .hg.'''
1029 l = self._wlockref and self._wlockref()
1029 l = self._wlockref and self._wlockref()
1030 if l is not None and l.held:
1030 if l is not None and l.held:
1031 l.lock()
1031 l.lock()
1032 return l
1032 return l
1033
1033
1034 def unlock():
1034 def unlock():
1035 self.dirstate.write()
1035 self.dirstate.write()
1036 self._filecache['dirstate'].refresh()
1036 self._filecache['dirstate'].refresh()
1037
1037
1038 l = self._lock(self.join("wlock"), wait, unlock,
1038 l = self._lock(self.join("wlock"), wait, unlock,
1039 self.invalidatedirstate, _('working directory of %s') %
1039 self.invalidatedirstate, _('working directory of %s') %
1040 self.origroot)
1040 self.origroot)
1041 self._wlockref = weakref.ref(l)
1041 self._wlockref = weakref.ref(l)
1042 return l
1042 return l
1043
1043
1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1045 """
1045 """
1046 commit an individual file as part of a larger transaction
1046 commit an individual file as part of a larger transaction
1047 """
1047 """
1048
1048
1049 fname = fctx.path()
1049 fname = fctx.path()
1050 text = fctx.data()
1050 text = fctx.data()
1051 flog = self.file(fname)
1051 flog = self.file(fname)
1052 fparent1 = manifest1.get(fname, nullid)
1052 fparent1 = manifest1.get(fname, nullid)
1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1054
1054
1055 meta = {}
1055 meta = {}
1056 copy = fctx.renamed()
1056 copy = fctx.renamed()
1057 if copy and copy[0] != fname:
1057 if copy and copy[0] != fname:
1058 # Mark the new revision of this file as a copy of another
1058 # Mark the new revision of this file as a copy of another
1059 # file. This copy data will effectively act as a parent
1059 # file. This copy data will effectively act as a parent
1060 # of this new revision. If this is a merge, the first
1060 # of this new revision. If this is a merge, the first
1061 # parent will be the nullid (meaning "look up the copy data")
1061 # parent will be the nullid (meaning "look up the copy data")
1062 # and the second one will be the other parent. For example:
1062 # and the second one will be the other parent. For example:
1063 #
1063 #
1064 # 0 --- 1 --- 3 rev1 changes file foo
1064 # 0 --- 1 --- 3 rev1 changes file foo
1065 # \ / rev2 renames foo to bar and changes it
1065 # \ / rev2 renames foo to bar and changes it
1066 # \- 2 -/ rev3 should have bar with all changes and
1066 # \- 2 -/ rev3 should have bar with all changes and
1067 # should record that bar descends from
1067 # should record that bar descends from
1068 # bar in rev2 and foo in rev1
1068 # bar in rev2 and foo in rev1
1069 #
1069 #
1070 # this allows this merge to succeed:
1070 # this allows this merge to succeed:
1071 #
1071 #
1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1073 # \ / merging rev3 and rev4 should use bar@rev2
1073 # \ / merging rev3 and rev4 should use bar@rev2
1074 # \- 2 --- 4 as the merge base
1074 # \- 2 --- 4 as the merge base
1075 #
1075 #
1076
1076
1077 cfname = copy[0]
1077 cfname = copy[0]
1078 crev = manifest1.get(cfname)
1078 crev = manifest1.get(cfname)
1079 newfparent = fparent2
1079 newfparent = fparent2
1080
1080
1081 if manifest2: # branch merge
1081 if manifest2: # branch merge
1082 if fparent2 == nullid or crev is None: # copied on remote side
1082 if fparent2 == nullid or crev is None: # copied on remote side
1083 if cfname in manifest2:
1083 if cfname in manifest2:
1084 crev = manifest2[cfname]
1084 crev = manifest2[cfname]
1085 newfparent = fparent1
1085 newfparent = fparent1
1086
1086
1087 # find source in nearest ancestor if we've lost track
1087 # find source in nearest ancestor if we've lost track
1088 if not crev:
1088 if not crev:
1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1090 (fname, cfname))
1090 (fname, cfname))
1091 for ancestor in self[None].ancestors():
1091 for ancestor in self[None].ancestors():
1092 if cfname in ancestor:
1092 if cfname in ancestor:
1093 crev = ancestor[cfname].filenode()
1093 crev = ancestor[cfname].filenode()
1094 break
1094 break
1095
1095
1096 if crev:
1096 if crev:
1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1098 meta["copy"] = cfname
1098 meta["copy"] = cfname
1099 meta["copyrev"] = hex(crev)
1099 meta["copyrev"] = hex(crev)
1100 fparent1, fparent2 = nullid, newfparent
1100 fparent1, fparent2 = nullid, newfparent
1101 else:
1101 else:
1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1103 "copied from '%s'!\n") % (fname, cfname))
1103 "copied from '%s'!\n") % (fname, cfname))
1104
1104
1105 elif fparent2 != nullid:
1105 elif fparent2 != nullid:
1106 # is one parent an ancestor of the other?
1106 # is one parent an ancestor of the other?
1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1108 if fparentancestor == fparent1:
1108 if fparentancestor == fparent1:
1109 fparent1, fparent2 = fparent2, nullid
1109 fparent1, fparent2 = fparent2, nullid
1110 elif fparentancestor == fparent2:
1110 elif fparentancestor == fparent2:
1111 fparent2 = nullid
1111 fparent2 = nullid
1112
1112
1113 # is the file changed?
1113 # is the file changed?
1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1115 changelist.append(fname)
1115 changelist.append(fname)
1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1117
1117
1118 # are just the flags changed during merge?
1118 # are just the flags changed during merge?
1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1120 changelist.append(fname)
1120 changelist.append(fname)
1121
1121
1122 return fparent1
1122 return fparent1
1123
1123
1124 @unfilteredmethod
1124 @unfilteredmethod
1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1126 editor=False, extra={}):
1126 editor=False, extra={}):
1127 """Add a new revision to current repository.
1127 """Add a new revision to current repository.
1128
1128
1129 Revision information is gathered from the working directory,
1129 Revision information is gathered from the working directory,
1130 match can be used to filter the committed files. If editor is
1130 match can be used to filter the committed files. If editor is
1131 supplied, it is called to get a commit message.
1131 supplied, it is called to get a commit message.
1132 """
1132 """
1133
1133
1134 def fail(f, msg):
1134 def fail(f, msg):
1135 raise util.Abort('%s: %s' % (f, msg))
1135 raise util.Abort('%s: %s' % (f, msg))
1136
1136
1137 if not match:
1137 if not match:
1138 match = matchmod.always(self.root, '')
1138 match = matchmod.always(self.root, '')
1139
1139
1140 if not force:
1140 if not force:
1141 vdirs = []
1141 vdirs = []
1142 match.dir = vdirs.append
1142 match.dir = vdirs.append
1143 match.bad = fail
1143 match.bad = fail
1144
1144
1145 wlock = self.wlock()
1145 wlock = self.wlock()
1146 try:
1146 try:
1147 wctx = self[None]
1147 wctx = self[None]
1148 merge = len(wctx.parents()) > 1
1148 merge = len(wctx.parents()) > 1
1149
1149
1150 if (not force and merge and match and
1150 if (not force and merge and match and
1151 (match.files() or match.anypats())):
1151 (match.files() or match.anypats())):
1152 raise util.Abort(_('cannot partially commit a merge '
1152 raise util.Abort(_('cannot partially commit a merge '
1153 '(do not specify files or patterns)'))
1153 '(do not specify files or patterns)'))
1154
1154
1155 changes = self.status(match=match, clean=force)
1155 changes = self.status(match=match, clean=force)
1156 if force:
1156 if force:
1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1158
1158
1159 # check subrepos
1159 # check subrepos
1160 subs = []
1160 subs = []
1161 commitsubs = set()
1161 commitsubs = set()
1162 newstate = wctx.substate.copy()
1162 newstate = wctx.substate.copy()
1163 # only manage subrepos and .hgsubstate if .hgsub is present
1163 # only manage subrepos and .hgsubstate if .hgsub is present
1164 if '.hgsub' in wctx:
1164 if '.hgsub' in wctx:
1165 # we'll decide whether to track this ourselves, thanks
1165 # we'll decide whether to track this ourselves, thanks
1166 if '.hgsubstate' in changes[0]:
1166 if '.hgsubstate' in changes[0]:
1167 changes[0].remove('.hgsubstate')
1167 changes[0].remove('.hgsubstate')
1168 if '.hgsubstate' in changes[2]:
1168 if '.hgsubstate' in changes[2]:
1169 changes[2].remove('.hgsubstate')
1169 changes[2].remove('.hgsubstate')
1170
1170
1171 # compare current state to last committed state
1171 # compare current state to last committed state
1172 # build new substate based on last committed state
1172 # build new substate based on last committed state
1173 oldstate = wctx.p1().substate
1173 oldstate = wctx.p1().substate
1174 for s in sorted(newstate.keys()):
1174 for s in sorted(newstate.keys()):
1175 if not match(s):
1175 if not match(s):
1176 # ignore working copy, use old state if present
1176 # ignore working copy, use old state if present
1177 if s in oldstate:
1177 if s in oldstate:
1178 newstate[s] = oldstate[s]
1178 newstate[s] = oldstate[s]
1179 continue
1179 continue
1180 if not force:
1180 if not force:
1181 raise util.Abort(
1181 raise util.Abort(
1182 _("commit with new subrepo %s excluded") % s)
1182 _("commit with new subrepo %s excluded") % s)
1183 if wctx.sub(s).dirty(True):
1183 if wctx.sub(s).dirty(True):
1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1185 raise util.Abort(
1185 raise util.Abort(
1186 _("uncommitted changes in subrepo %s") % s,
1186 _("uncommitted changes in subrepo %s") % s,
1187 hint=_("use --subrepos for recursive commit"))
1187 hint=_("use --subrepos for recursive commit"))
1188 subs.append(s)
1188 subs.append(s)
1189 commitsubs.add(s)
1189 commitsubs.add(s)
1190 else:
1190 else:
1191 bs = wctx.sub(s).basestate()
1191 bs = wctx.sub(s).basestate()
1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1193 if oldstate.get(s, (None, None, None))[1] != bs:
1193 if oldstate.get(s, (None, None, None))[1] != bs:
1194 subs.append(s)
1194 subs.append(s)
1195
1195
1196 # check for removed subrepos
1196 # check for removed subrepos
1197 for p in wctx.parents():
1197 for p in wctx.parents():
1198 r = [s for s in p.substate if s not in newstate]
1198 r = [s for s in p.substate if s not in newstate]
1199 subs += [s for s in r if match(s)]
1199 subs += [s for s in r if match(s)]
1200 if subs:
1200 if subs:
1201 if (not match('.hgsub') and
1201 if (not match('.hgsub') and
1202 '.hgsub' in (wctx.modified() + wctx.added())):
1202 '.hgsub' in (wctx.modified() + wctx.added())):
1203 raise util.Abort(
1203 raise util.Abort(
1204 _("can't commit subrepos without .hgsub"))
1204 _("can't commit subrepos without .hgsub"))
1205 changes[0].insert(0, '.hgsubstate')
1205 changes[0].insert(0, '.hgsubstate')
1206
1206
1207 elif '.hgsub' in changes[2]:
1207 elif '.hgsub' in changes[2]:
1208 # clean up .hgsubstate when .hgsub is removed
1208 # clean up .hgsubstate when .hgsub is removed
1209 if ('.hgsubstate' in wctx and
1209 if ('.hgsubstate' in wctx and
1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1211 changes[2].insert(0, '.hgsubstate')
1211 changes[2].insert(0, '.hgsubstate')
1212
1212
1213 # make sure all explicit patterns are matched
1213 # make sure all explicit patterns are matched
1214 if not force and match.files():
1214 if not force and match.files():
1215 matched = set(changes[0] + changes[1] + changes[2])
1215 matched = set(changes[0] + changes[1] + changes[2])
1216
1216
1217 for f in match.files():
1217 for f in match.files():
1218 f = self.dirstate.normalize(f)
1218 f = self.dirstate.normalize(f)
1219 if f == '.' or f in matched or f in wctx.substate:
1219 if f == '.' or f in matched or f in wctx.substate:
1220 continue
1220 continue
1221 if f in changes[3]: # missing
1221 if f in changes[3]: # missing
1222 fail(f, _('file not found!'))
1222 fail(f, _('file not found!'))
1223 if f in vdirs: # visited directory
1223 if f in vdirs: # visited directory
1224 d = f + '/'
1224 d = f + '/'
1225 for mf in matched:
1225 for mf in matched:
1226 if mf.startswith(d):
1226 if mf.startswith(d):
1227 break
1227 break
1228 else:
1228 else:
1229 fail(f, _("no match under directory!"))
1229 fail(f, _("no match under directory!"))
1230 elif f not in self.dirstate:
1230 elif f not in self.dirstate:
1231 fail(f, _("file not tracked!"))
1231 fail(f, _("file not tracked!"))
1232
1232
1233 if (not force and not extra.get("close") and not merge
1233 if (not force and not extra.get("close") and not merge
1234 and not (changes[0] or changes[1] or changes[2])
1234 and not (changes[0] or changes[1] or changes[2])
1235 and wctx.branch() == wctx.p1().branch()):
1235 and wctx.branch() == wctx.p1().branch()):
1236 return None
1236 return None
1237
1237
1238 if merge and changes[3]:
1238 if merge and changes[3]:
1239 raise util.Abort(_("cannot commit merge with missing files"))
1239 raise util.Abort(_("cannot commit merge with missing files"))
1240
1240
1241 ms = mergemod.mergestate(self)
1241 ms = mergemod.mergestate(self)
1242 for f in changes[0]:
1242 for f in changes[0]:
1243 if f in ms and ms[f] == 'u':
1243 if f in ms and ms[f] == 'u':
1244 raise util.Abort(_("unresolved merge conflicts "
1244 raise util.Abort(_("unresolved merge conflicts "
1245 "(see hg help resolve)"))
1245 "(see hg help resolve)"))
1246
1246
1247 cctx = context.workingctx(self, text, user, date, extra, changes)
1247 cctx = context.workingctx(self, text, user, date, extra, changes)
1248 if editor:
1248 if editor:
1249 cctx._text = editor(self, cctx, subs)
1249 cctx._text = editor(self, cctx, subs)
1250 edited = (text != cctx._text)
1250 edited = (text != cctx._text)
1251
1251
1252 # commit subs and write new state
1252 # commit subs and write new state
1253 if subs:
1253 if subs:
1254 for s in sorted(commitsubs):
1254 for s in sorted(commitsubs):
1255 sub = wctx.sub(s)
1255 sub = wctx.sub(s)
1256 self.ui.status(_('committing subrepository %s\n') %
1256 self.ui.status(_('committing subrepository %s\n') %
1257 subrepo.subrelpath(sub))
1257 subrepo.subrelpath(sub))
1258 sr = sub.commit(cctx._text, user, date)
1258 sr = sub.commit(cctx._text, user, date)
1259 newstate[s] = (newstate[s][0], sr)
1259 newstate[s] = (newstate[s][0], sr)
1260 subrepo.writestate(self, newstate)
1260 subrepo.writestate(self, newstate)
1261
1261
1262 # Save commit message in case this transaction gets rolled back
1262 # Save commit message in case this transaction gets rolled back
1263 # (e.g. by a pretxncommit hook). Leave the content alone on
1263 # (e.g. by a pretxncommit hook). Leave the content alone on
1264 # the assumption that the user will use the same editor again.
1264 # the assumption that the user will use the same editor again.
1265 msgfn = self.savecommitmessage(cctx._text)
1265 msgfn = self.savecommitmessage(cctx._text)
1266
1266
1267 p1, p2 = self.dirstate.parents()
1267 p1, p2 = self.dirstate.parents()
1268 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1268 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1269 try:
1269 try:
1270 self.hook("precommit", throw=True, parent1=hookp1,
1270 self.hook("precommit", throw=True, parent1=hookp1,
1271 parent2=hookp2)
1271 parent2=hookp2)
1272 ret = self.commitctx(cctx, True)
1272 ret = self.commitctx(cctx, True)
1273 except: # re-raises
1273 except: # re-raises
1274 if edited:
1274 if edited:
1275 self.ui.write(
1275 self.ui.write(
1276 _('note: commit message saved in %s\n') % msgfn)
1276 _('note: commit message saved in %s\n') % msgfn)
1277 raise
1277 raise
1278
1278
1279 # update bookmarks, dirstate and mergestate
1279 # update bookmarks, dirstate and mergestate
1280 bookmarks.update(self, [p1, p2], ret)
1280 bookmarks.update(self, [p1, p2], ret)
1281 for f in changes[0] + changes[1]:
1281 for f in changes[0] + changes[1]:
1282 self.dirstate.normal(f)
1282 self.dirstate.normal(f)
1283 for f in changes[2]:
1283 for f in changes[2]:
1284 self.dirstate.drop(f)
1284 self.dirstate.drop(f)
1285 self.dirstate.setparents(ret)
1285 self.dirstate.setparents(ret)
1286 ms.reset()
1286 ms.reset()
1287 finally:
1287 finally:
1288 wlock.release()
1288 wlock.release()
1289
1289
1290 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1290 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1291 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1291 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1292 self._afterlock(commithook)
1292 self._afterlock(commithook)
1293 return ret
1293 return ret
1294
1294
1295 @unfilteredmethod
1295 @unfilteredmethod
1296 def commitctx(self, ctx, error=False):
1296 def commitctx(self, ctx, error=False):
1297 """Add a new revision to current repository.
1297 """Add a new revision to current repository.
1298 Revision information is passed via the context argument.
1298 Revision information is passed via the context argument.
1299 """
1299 """
1300
1300
1301 tr = lock = None
1301 tr = lock = None
1302 removed = list(ctx.removed())
1302 removed = list(ctx.removed())
1303 p1, p2 = ctx.p1(), ctx.p2()
1303 p1, p2 = ctx.p1(), ctx.p2()
1304 user = ctx.user()
1304 user = ctx.user()
1305
1305
1306 lock = self.lock()
1306 lock = self.lock()
1307 try:
1307 try:
1308 tr = self.transaction("commit")
1308 tr = self.transaction("commit")
1309 trp = weakref.proxy(tr)
1309 trp = weakref.proxy(tr)
1310
1310
1311 if ctx.files():
1311 if ctx.files():
1312 m1 = p1.manifest().copy()
1312 m1 = p1.manifest().copy()
1313 m2 = p2.manifest()
1313 m2 = p2.manifest()
1314
1314
1315 # check in files
1315 # check in files
1316 new = {}
1316 new = {}
1317 changed = []
1317 changed = []
1318 linkrev = len(self)
1318 linkrev = len(self)
1319 for f in sorted(ctx.modified() + ctx.added()):
1319 for f in sorted(ctx.modified() + ctx.added()):
1320 self.ui.note(f + "\n")
1320 self.ui.note(f + "\n")
1321 try:
1321 try:
1322 fctx = ctx[f]
1322 fctx = ctx[f]
1323 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1323 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1324 changed)
1324 changed)
1325 m1.set(f, fctx.flags())
1325 m1.set(f, fctx.flags())
1326 except OSError, inst:
1326 except OSError, inst:
1327 self.ui.warn(_("trouble committing %s!\n") % f)
1327 self.ui.warn(_("trouble committing %s!\n") % f)
1328 raise
1328 raise
1329 except IOError, inst:
1329 except IOError, inst:
1330 errcode = getattr(inst, 'errno', errno.ENOENT)
1330 errcode = getattr(inst, 'errno', errno.ENOENT)
1331 if error or errcode and errcode != errno.ENOENT:
1331 if error or errcode and errcode != errno.ENOENT:
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1333 raise
1333 raise
1334 else:
1334 else:
1335 removed.append(f)
1335 removed.append(f)
1336
1336
1337 # update manifest
1337 # update manifest
1338 m1.update(new)
1338 m1.update(new)
1339 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1339 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1340 drop = [f for f in removed if f in m1]
1340 drop = [f for f in removed if f in m1]
1341 for f in drop:
1341 for f in drop:
1342 del m1[f]
1342 del m1[f]
1343 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1343 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1344 p2.manifestnode(), (new, drop))
1344 p2.manifestnode(), (new, drop))
1345 files = changed + removed
1345 files = changed + removed
1346 else:
1346 else:
1347 mn = p1.manifestnode()
1347 mn = p1.manifestnode()
1348 files = []
1348 files = []
1349
1349
1350 # update changelog
1350 # update changelog
1351 self.changelog.delayupdate()
1351 self.changelog.delayupdate()
1352 n = self.changelog.add(mn, files, ctx.description(),
1352 n = self.changelog.add(mn, files, ctx.description(),
1353 trp, p1.node(), p2.node(),
1353 trp, p1.node(), p2.node(),
1354 user, ctx.date(), ctx.extra().copy())
1354 user, ctx.date(), ctx.extra().copy())
1355 p = lambda: self.changelog.writepending() and self.root or ""
1355 p = lambda: self.changelog.writepending() and self.root or ""
1356 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1356 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1357 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1357 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1358 parent2=xp2, pending=p)
1358 parent2=xp2, pending=p)
1359 self.changelog.finalize(trp)
1359 self.changelog.finalize(trp)
1360 # set the new commit is proper phase
1360 # set the new commit is proper phase
1361 targetphase = phases.newcommitphase(self.ui)
1361 targetphase = phases.newcommitphase(self.ui)
1362 if targetphase:
1362 if targetphase:
1363 # retract boundary do not alter parent changeset.
1363 # retract boundary do not alter parent changeset.
1364 # if a parent have higher the resulting phase will
1364 # if a parent have higher the resulting phase will
1365 # be compliant anyway
1365 # be compliant anyway
1366 #
1366 #
1367 # if minimal phase was 0 we don't need to retract anything
1367 # if minimal phase was 0 we don't need to retract anything
1368 phases.retractboundary(self, targetphase, [n])
1368 phases.retractboundary(self, targetphase, [n])
1369 tr.close()
1369 tr.close()
1370 branchmap.updatecache(self.filtered('served'))
1370 branchmap.updatecache(self.filtered('served'))
1371 return n
1371 return n
1372 finally:
1372 finally:
1373 if tr:
1373 if tr:
1374 tr.release()
1374 tr.release()
1375 lock.release()
1375 lock.release()
1376
1376
1377 @unfilteredmethod
1377 @unfilteredmethod
1378 def destroying(self):
1378 def destroying(self):
1379 '''Inform the repository that nodes are about to be destroyed.
1379 '''Inform the repository that nodes are about to be destroyed.
1380 Intended for use by strip and rollback, so there's a common
1380 Intended for use by strip and rollback, so there's a common
1381 place for anything that has to be done before destroying history.
1381 place for anything that has to be done before destroying history.
1382
1382
1383 This is mostly useful for saving state that is in memory and waiting
1383 This is mostly useful for saving state that is in memory and waiting
1384 to be flushed when the current lock is released. Because a call to
1384 to be flushed when the current lock is released. Because a call to
1385 destroyed is imminent, the repo will be invalidated causing those
1385 destroyed is imminent, the repo will be invalidated causing those
1386 changes to stay in memory (waiting for the next unlock), or vanish
1386 changes to stay in memory (waiting for the next unlock), or vanish
1387 completely.
1387 completely.
1388 '''
1388 '''
1389 # When using the same lock to commit and strip, the phasecache is left
1389 # When using the same lock to commit and strip, the phasecache is left
1390 # dirty after committing. Then when we strip, the repo is invalidated,
1390 # dirty after committing. Then when we strip, the repo is invalidated,
1391 # causing those changes to disappear.
1391 # causing those changes to disappear.
1392 if '_phasecache' in vars(self):
1392 if '_phasecache' in vars(self):
1393 self._phasecache.write()
1393 self._phasecache.write()
1394
1394
1395 @unfilteredmethod
1395 @unfilteredmethod
1396 def destroyed(self):
1396 def destroyed(self):
1397 '''Inform the repository that nodes have been destroyed.
1397 '''Inform the repository that nodes have been destroyed.
1398 Intended for use by strip and rollback, so there's a common
1398 Intended for use by strip and rollback, so there's a common
1399 place for anything that has to be done after destroying history.
1399 place for anything that has to be done after destroying history.
1400
1400
1401 If you know the branchheadcache was uptodate before nodes were removed
1401 If you know the branchheadcache was uptodate before nodes were removed
1402 and you also know the set of candidate new heads that may have resulted
1402 and you also know the set of candidate new heads that may have resulted
1403 from the destruction, you can set newheadnodes. This will enable the
1403 from the destruction, you can set newheadnodes. This will enable the
1404 code to update the branchheads cache, rather than having future code
1404 code to update the branchheads cache, rather than having future code
1405 decide it's invalid and regenerating it from scratch.
1405 decide it's invalid and regenerating it from scratch.
1406 '''
1406 '''
1407 # When one tries to:
1407 # When one tries to:
1408 # 1) destroy nodes thus calling this method (e.g. strip)
1408 # 1) destroy nodes thus calling this method (e.g. strip)
1409 # 2) use phasecache somewhere (e.g. commit)
1409 # 2) use phasecache somewhere (e.g. commit)
1410 #
1410 #
1411 # then 2) will fail because the phasecache contains nodes that were
1411 # then 2) will fail because the phasecache contains nodes that were
1412 # removed. We can either remove phasecache from the filecache,
1412 # removed. We can either remove phasecache from the filecache,
1413 # causing it to reload next time it is accessed, or simply filter
1413 # causing it to reload next time it is accessed, or simply filter
1414 # the removed nodes now and write the updated cache.
1414 # the removed nodes now and write the updated cache.
1415 if '_phasecache' in self._filecache:
1415 if '_phasecache' in self._filecache:
1416 self._phasecache.filterunknown(self)
1416 self._phasecache.filterunknown(self)
1417 self._phasecache.write()
1417 self._phasecache.write()
1418
1418
1419 # update the 'served' branch cache to help read only server process
1419 # update the 'served' branch cache to help read only server process
1420 # Thanks to branchcach collaboration this is done from the nearest
1420 # Thanks to branchcach collaboration this is done from the nearest
1421 # filtered subset and it is expected to be fast.
1421 # filtered subset and it is expected to be fast.
1422 branchmap.updatecache(self.filtered('served'))
1422 branchmap.updatecache(self.filtered('served'))
1423
1423
1424 # Ensure the persistent tag cache is updated. Doing it now
1424 # Ensure the persistent tag cache is updated. Doing it now
1425 # means that the tag cache only has to worry about destroyed
1425 # means that the tag cache only has to worry about destroyed
1426 # heads immediately after a strip/rollback. That in turn
1426 # heads immediately after a strip/rollback. That in turn
1427 # guarantees that "cachetip == currenttip" (comparing both rev
1427 # guarantees that "cachetip == currenttip" (comparing both rev
1428 # and node) always means no nodes have been added or destroyed.
1428 # and node) always means no nodes have been added or destroyed.
1429
1429
1430 # XXX this is suboptimal when qrefresh'ing: we strip the current
1430 # XXX this is suboptimal when qrefresh'ing: we strip the current
1431 # head, refresh the tag cache, then immediately add a new head.
1431 # head, refresh the tag cache, then immediately add a new head.
1432 # But I think doing it this way is necessary for the "instant
1432 # But I think doing it this way is necessary for the "instant
1433 # tag cache retrieval" case to work.
1433 # tag cache retrieval" case to work.
1434 self.invalidate()
1434 self.invalidate()
1435
1435
1436 def walk(self, match, node=None):
1436 def walk(self, match, node=None):
1437 '''
1437 '''
1438 walk recursively through the directory tree or a given
1438 walk recursively through the directory tree or a given
1439 changeset, finding all files matched by the match
1439 changeset, finding all files matched by the match
1440 function
1440 function
1441 '''
1441 '''
1442 return self[node].walk(match)
1442 return self[node].walk(match)
1443
1443
1444 def status(self, node1='.', node2=None, match=None,
1444 def status(self, node1='.', node2=None, match=None,
1445 ignored=False, clean=False, unknown=False,
1445 ignored=False, clean=False, unknown=False,
1446 listsubrepos=False):
1446 listsubrepos=False):
1447 """return status of files between two nodes or node and working
1447 """return status of files between two nodes or node and working
1448 directory.
1448 directory.
1449
1449
1450 If node1 is None, use the first dirstate parent instead.
1450 If node1 is None, use the first dirstate parent instead.
1451 If node2 is None, compare node1 with working directory.
1451 If node2 is None, compare node1 with working directory.
1452 """
1452 """
1453
1453
1454 def mfmatches(ctx):
1454 def mfmatches(ctx):
1455 mf = ctx.manifest().copy()
1455 mf = ctx.manifest().copy()
1456 if match.always():
1456 if match.always():
1457 return mf
1457 return mf
1458 for fn in mf.keys():
1458 for fn in mf.keys():
1459 if not match(fn):
1459 if not match(fn):
1460 del mf[fn]
1460 del mf[fn]
1461 return mf
1461 return mf
1462
1462
1463 if isinstance(node1, context.changectx):
1463 if isinstance(node1, context.changectx):
1464 ctx1 = node1
1464 ctx1 = node1
1465 else:
1465 else:
1466 ctx1 = self[node1]
1466 ctx1 = self[node1]
1467 if isinstance(node2, context.changectx):
1467 if isinstance(node2, context.changectx):
1468 ctx2 = node2
1468 ctx2 = node2
1469 else:
1469 else:
1470 ctx2 = self[node2]
1470 ctx2 = self[node2]
1471
1471
1472 working = ctx2.rev() is None
1472 working = ctx2.rev() is None
1473 parentworking = working and ctx1 == self['.']
1473 parentworking = working and ctx1 == self['.']
1474 match = match or matchmod.always(self.root, self.getcwd())
1474 match = match or matchmod.always(self.root, self.getcwd())
1475 listignored, listclean, listunknown = ignored, clean, unknown
1475 listignored, listclean, listunknown = ignored, clean, unknown
1476
1476
1477 # load earliest manifest first for caching reasons
1477 # load earliest manifest first for caching reasons
1478 if not working and ctx2.rev() < ctx1.rev():
1478 if not working and ctx2.rev() < ctx1.rev():
1479 ctx2.manifest()
1479 ctx2.manifest()
1480
1480
1481 if not parentworking:
1481 if not parentworking:
1482 def bad(f, msg):
1482 def bad(f, msg):
1483 # 'f' may be a directory pattern from 'match.files()',
1483 # 'f' may be a directory pattern from 'match.files()',
1484 # so 'f not in ctx1' is not enough
1484 # so 'f not in ctx1' is not enough
1485 if f not in ctx1 and f not in ctx1.dirs():
1485 if f not in ctx1 and f not in ctx1.dirs():
1486 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1486 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1487 match.bad = bad
1487 match.bad = bad
1488
1488
1489 if working: # we need to scan the working dir
1489 if working: # we need to scan the working dir
1490 subrepos = []
1490 subrepos = []
1491 if '.hgsub' in self.dirstate:
1491 if '.hgsub' in self.dirstate:
1492 subrepos = sorted(ctx2.substate)
1492 subrepos = sorted(ctx2.substate)
1493 s = self.dirstate.status(match, subrepos, listignored,
1493 s = self.dirstate.status(match, subrepos, listignored,
1494 listclean, listunknown)
1494 listclean, listunknown)
1495 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1495 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1496
1496
1497 # check for any possibly clean files
1497 # check for any possibly clean files
1498 if parentworking and cmp:
1498 if parentworking and cmp:
1499 fixup = []
1499 fixup = []
1500 # do a full compare of any files that might have changed
1500 # do a full compare of any files that might have changed
1501 for f in sorted(cmp):
1501 for f in sorted(cmp):
1502 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1502 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1503 or ctx1[f].cmp(ctx2[f])):
1503 or ctx1[f].cmp(ctx2[f])):
1504 modified.append(f)
1504 modified.append(f)
1505 else:
1505 else:
1506 fixup.append(f)
1506 fixup.append(f)
1507
1507
1508 # update dirstate for files that are actually clean
1508 # update dirstate for files that are actually clean
1509 if fixup:
1509 if fixup:
1510 if listclean:
1510 if listclean:
1511 clean += fixup
1511 clean += fixup
1512
1512
1513 try:
1513 try:
1514 # updating the dirstate is optional
1514 # updating the dirstate is optional
1515 # so we don't wait on the lock
1515 # so we don't wait on the lock
1516 wlock = self.wlock(False)
1516 wlock = self.wlock(False)
1517 try:
1517 try:
1518 for f in fixup:
1518 for f in fixup:
1519 self.dirstate.normal(f)
1519 self.dirstate.normal(f)
1520 finally:
1520 finally:
1521 wlock.release()
1521 wlock.release()
1522 except error.LockError:
1522 except error.LockError:
1523 pass
1523 pass
1524
1524
1525 if not parentworking:
1525 if not parentworking:
1526 mf1 = mfmatches(ctx1)
1526 mf1 = mfmatches(ctx1)
1527 if working:
1527 if working:
1528 # we are comparing working dir against non-parent
1528 # we are comparing working dir against non-parent
1529 # generate a pseudo-manifest for the working dir
1529 # generate a pseudo-manifest for the working dir
1530 mf2 = mfmatches(self['.'])
1530 mf2 = mfmatches(self['.'])
1531 for f in cmp + modified + added:
1531 for f in cmp + modified + added:
1532 mf2[f] = None
1532 mf2[f] = None
1533 mf2.set(f, ctx2.flags(f))
1533 mf2.set(f, ctx2.flags(f))
1534 for f in removed:
1534 for f in removed:
1535 if f in mf2:
1535 if f in mf2:
1536 del mf2[f]
1536 del mf2[f]
1537 else:
1537 else:
1538 # we are comparing two revisions
1538 # we are comparing two revisions
1539 deleted, unknown, ignored = [], [], []
1539 deleted, unknown, ignored = [], [], []
1540 mf2 = mfmatches(ctx2)
1540 mf2 = mfmatches(ctx2)
1541
1541
1542 modified, added, clean = [], [], []
1542 modified, added, clean = [], [], []
1543 withflags = mf1.withflags() | mf2.withflags()
1543 withflags = mf1.withflags() | mf2.withflags()
1544 for fn in mf2:
1544 for fn in mf2:
1545 if fn in mf1:
1545 if fn in mf1:
1546 if (fn not in deleted and
1546 if (fn not in deleted and
1547 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1547 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1548 (mf1[fn] != mf2[fn] and
1548 (mf1[fn] != mf2[fn] and
1549 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1549 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1550 modified.append(fn)
1550 modified.append(fn)
1551 elif listclean:
1551 elif listclean:
1552 clean.append(fn)
1552 clean.append(fn)
1553 del mf1[fn]
1553 del mf1[fn]
1554 elif fn not in deleted:
1554 elif fn not in deleted:
1555 added.append(fn)
1555 added.append(fn)
1556 removed = mf1.keys()
1556 removed = mf1.keys()
1557
1557
1558 if working and modified and not self.dirstate._checklink:
1558 if working and modified and not self.dirstate._checklink:
1559 # Symlink placeholders may get non-symlink-like contents
1559 # Symlink placeholders may get non-symlink-like contents
1560 # via user error or dereferencing by NFS or Samba servers,
1560 # via user error or dereferencing by NFS or Samba servers,
1561 # so we filter out any placeholders that don't look like a
1561 # so we filter out any placeholders that don't look like a
1562 # symlink
1562 # symlink
1563 sane = []
1563 sane = []
1564 for f in modified:
1564 for f in modified:
1565 if ctx2.flags(f) == 'l':
1565 if ctx2.flags(f) == 'l':
1566 d = ctx2[f].data()
1566 d = ctx2[f].data()
1567 if len(d) >= 1024 or '\n' in d or util.binary(d):
1567 if len(d) >= 1024 or '\n' in d or util.binary(d):
1568 self.ui.debug('ignoring suspect symlink placeholder'
1568 self.ui.debug('ignoring suspect symlink placeholder'
1569 ' "%s"\n' % f)
1569 ' "%s"\n' % f)
1570 continue
1570 continue
1571 sane.append(f)
1571 sane.append(f)
1572 modified = sane
1572 modified = sane
1573
1573
1574 r = modified, added, removed, deleted, unknown, ignored, clean
1574 r = modified, added, removed, deleted, unknown, ignored, clean
1575
1575
1576 if listsubrepos:
1576 if listsubrepos:
1577 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1577 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1578 if working:
1578 if working:
1579 rev2 = None
1579 rev2 = None
1580 else:
1580 else:
1581 rev2 = ctx2.substate[subpath][1]
1581 rev2 = ctx2.substate[subpath][1]
1582 try:
1582 try:
1583 submatch = matchmod.narrowmatcher(subpath, match)
1583 submatch = matchmod.narrowmatcher(subpath, match)
1584 s = sub.status(rev2, match=submatch, ignored=listignored,
1584 s = sub.status(rev2, match=submatch, ignored=listignored,
1585 clean=listclean, unknown=listunknown,
1585 clean=listclean, unknown=listunknown,
1586 listsubrepos=True)
1586 listsubrepos=True)
1587 for rfiles, sfiles in zip(r, s):
1587 for rfiles, sfiles in zip(r, s):
1588 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1588 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1589 except error.LookupError:
1589 except error.LookupError:
1590 self.ui.status(_("skipping missing subrepository: %s\n")
1590 self.ui.status(_("skipping missing subrepository: %s\n")
1591 % subpath)
1591 % subpath)
1592
1592
1593 for l in r:
1593 for l in r:
1594 l.sort()
1594 l.sort()
1595 return r
1595 return r
1596
1596
1597 def heads(self, start=None):
1597 def heads(self, start=None):
1598 heads = self.changelog.heads(start)
1598 heads = self.changelog.heads(start)
1599 # sort the output in rev descending order
1599 # sort the output in rev descending order
1600 return sorted(heads, key=self.changelog.rev, reverse=True)
1600 return sorted(heads, key=self.changelog.rev, reverse=True)
1601
1601
1602 def branchheads(self, branch=None, start=None, closed=False):
1602 def branchheads(self, branch=None, start=None, closed=False):
1603 '''return a (possibly filtered) list of heads for the given branch
1603 '''return a (possibly filtered) list of heads for the given branch
1604
1604
1605 Heads are returned in topological order, from newest to oldest.
1605 Heads are returned in topological order, from newest to oldest.
1606 If branch is None, use the dirstate branch.
1606 If branch is None, use the dirstate branch.
1607 If start is not None, return only heads reachable from start.
1607 If start is not None, return only heads reachable from start.
1608 If closed is True, return heads that are marked as closed as well.
1608 If closed is True, return heads that are marked as closed as well.
1609 '''
1609 '''
1610 if branch is None:
1610 if branch is None:
1611 branch = self[None].branch()
1611 branch = self[None].branch()
1612 branches = self.branchmap()
1612 branches = self.branchmap()
1613 if branch not in branches:
1613 if branch not in branches:
1614 return []
1614 return []
1615 # the cache returns heads ordered lowest to highest
1615 # the cache returns heads ordered lowest to highest
1616 bheads = list(reversed(branches[branch]))
1616 bheads = list(reversed(branches[branch]))
1617 if start is not None:
1617 if start is not None:
1618 # filter out the heads that cannot be reached from startrev
1618 # filter out the heads that cannot be reached from startrev
1619 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1619 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1620 bheads = [h for h in bheads if h in fbheads]
1620 bheads = [h for h in bheads if h in fbheads]
1621 if not closed:
1621 if not closed:
1622 bheads = [h for h in bheads if not self[h].closesbranch()]
1622 bheads = [h for h in bheads if not self[h].closesbranch()]
1623 return bheads
1623 return bheads
1624
1624
1625 def branches(self, nodes):
1625 def branches(self, nodes):
1626 if not nodes:
1626 if not nodes:
1627 nodes = [self.changelog.tip()]
1627 nodes = [self.changelog.tip()]
1628 b = []
1628 b = []
1629 for n in nodes:
1629 for n in nodes:
1630 t = n
1630 t = n
1631 while True:
1631 while True:
1632 p = self.changelog.parents(n)
1632 p = self.changelog.parents(n)
1633 if p[1] != nullid or p[0] == nullid:
1633 if p[1] != nullid or p[0] == nullid:
1634 b.append((t, n, p[0], p[1]))
1634 b.append((t, n, p[0], p[1]))
1635 break
1635 break
1636 n = p[0]
1636 n = p[0]
1637 return b
1637 return b
1638
1638
1639 def between(self, pairs):
1639 def between(self, pairs):
1640 r = []
1640 r = []
1641
1641
1642 for top, bottom in pairs:
1642 for top, bottom in pairs:
1643 n, l, i = top, [], 0
1643 n, l, i = top, [], 0
1644 f = 1
1644 f = 1
1645
1645
1646 while n != bottom and n != nullid:
1646 while n != bottom and n != nullid:
1647 p = self.changelog.parents(n)[0]
1647 p = self.changelog.parents(n)[0]
1648 if i == f:
1648 if i == f:
1649 l.append(n)
1649 l.append(n)
1650 f = f * 2
1650 f = f * 2
1651 n = p
1651 n = p
1652 i += 1
1652 i += 1
1653
1653
1654 r.append(l)
1654 r.append(l)
1655
1655
1656 return r
1656 return r
1657
1657
1658 def pull(self, remote, heads=None, force=False):
1658 def pull(self, remote, heads=None, force=False):
1659 # don't open transaction for nothing or you break future useful
1659 # don't open transaction for nothing or you break future useful
1660 # rollback call
1660 # rollback call
1661 tr = None
1661 tr = None
1662 trname = 'pull\n' + util.hidepassword(remote.url())
1662 trname = 'pull\n' + util.hidepassword(remote.url())
1663 lock = self.lock()
1663 lock = self.lock()
1664 try:
1664 try:
1665 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1665 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1666 force=force)
1666 force=force)
1667 common, fetch, rheads = tmp
1667 common, fetch, rheads = tmp
1668 if not fetch:
1668 if not fetch:
1669 self.ui.status(_("no changes found\n"))
1669 self.ui.status(_("no changes found\n"))
1670 added = []
1670 added = []
1671 result = 0
1671 result = 0
1672 else:
1672 else:
1673 tr = self.transaction(trname)
1673 tr = self.transaction(trname)
1674 if heads is None and list(common) == [nullid]:
1674 if heads is None and list(common) == [nullid]:
1675 self.ui.status(_("requesting all changes\n"))
1675 self.ui.status(_("requesting all changes\n"))
1676 elif heads is None and remote.capable('changegroupsubset'):
1676 elif heads is None and remote.capable('changegroupsubset'):
1677 # issue1320, avoid a race if remote changed after discovery
1677 # issue1320, avoid a race if remote changed after discovery
1678 heads = rheads
1678 heads = rheads
1679
1679
1680 if remote.capable('getbundle'):
1680 if remote.capable('getbundle'):
1681 cg = remote.getbundle('pull', common=common,
1681 cg = remote.getbundle('pull', common=common,
1682 heads=heads or rheads)
1682 heads=heads or rheads)
1683 elif heads is None:
1683 elif heads is None:
1684 cg = remote.changegroup(fetch, 'pull')
1684 cg = remote.changegroup(fetch, 'pull')
1685 elif not remote.capable('changegroupsubset'):
1685 elif not remote.capable('changegroupsubset'):
1686 raise util.Abort(_("partial pull cannot be done because "
1686 raise util.Abort(_("partial pull cannot be done because "
1687 "other repository doesn't support "
1687 "other repository doesn't support "
1688 "changegroupsubset."))
1688 "changegroupsubset."))
1689 else:
1689 else:
1690 cg = remote.changegroupsubset(fetch, heads, 'pull')
1690 cg = remote.changegroupsubset(fetch, heads, 'pull')
1691 clstart = len(self.changelog)
1691 clstart = len(self.changelog)
1692 result = self.addchangegroup(cg, 'pull', remote.url())
1692 result = self.addchangegroup(cg, 'pull', remote.url())
1693 clend = len(self.changelog)
1693 clend = len(self.changelog)
1694 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1694 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1695
1695
1696 # compute target subset
1696 # compute target subset
1697 if heads is None:
1697 if heads is None:
1698 # We pulled every thing possible
1698 # We pulled every thing possible
1699 # sync on everything common
1699 # sync on everything common
1700 subset = common + added
1700 subset = common + added
1701 else:
1701 else:
1702 # We pulled a specific subset
1702 # We pulled a specific subset
1703 # sync on this subset
1703 # sync on this subset
1704 subset = heads
1704 subset = heads
1705
1705
1706 # Get remote phases data from remote
1706 # Get remote phases data from remote
1707 remotephases = remote.listkeys('phases')
1707 remotephases = remote.listkeys('phases')
1708 publishing = bool(remotephases.get('publishing', False))
1708 publishing = bool(remotephases.get('publishing', False))
1709 if remotephases and not publishing:
1709 if remotephases and not publishing:
1710 # remote is new and unpublishing
1710 # remote is new and unpublishing
1711 pheads, _dr = phases.analyzeremotephases(self, subset,
1711 pheads, _dr = phases.analyzeremotephases(self, subset,
1712 remotephases)
1712 remotephases)
1713 phases.advanceboundary(self, phases.public, pheads)
1713 phases.advanceboundary(self, phases.public, pheads)
1714 phases.advanceboundary(self, phases.draft, subset)
1714 phases.advanceboundary(self, phases.draft, subset)
1715 else:
1715 else:
1716 # Remote is old or publishing all common changesets
1716 # Remote is old or publishing all common changesets
1717 # should be seen as public
1717 # should be seen as public
1718 phases.advanceboundary(self, phases.public, subset)
1718 phases.advanceboundary(self, phases.public, subset)
1719
1719
1720 if obsolete._enabled:
1720 if obsolete._enabled:
1721 self.ui.debug('fetching remote obsolete markers\n')
1721 self.ui.debug('fetching remote obsolete markers\n')
1722 remoteobs = remote.listkeys('obsolete')
1722 remoteobs = remote.listkeys('obsolete')
1723 if 'dump0' in remoteobs:
1723 if 'dump0' in remoteobs:
1724 if tr is None:
1724 if tr is None:
1725 tr = self.transaction(trname)
1725 tr = self.transaction(trname)
1726 for key in sorted(remoteobs, reverse=True):
1726 for key in sorted(remoteobs, reverse=True):
1727 if key.startswith('dump'):
1727 if key.startswith('dump'):
1728 data = base85.b85decode(remoteobs[key])
1728 data = base85.b85decode(remoteobs[key])
1729 self.obsstore.mergemarkers(tr, data)
1729 self.obsstore.mergemarkers(tr, data)
1730 self.invalidatevolatilesets()
1730 self.invalidatevolatilesets()
1731 if tr is not None:
1731 if tr is not None:
1732 tr.close()
1732 tr.close()
1733 finally:
1733 finally:
1734 if tr is not None:
1734 if tr is not None:
1735 tr.release()
1735 tr.release()
1736 lock.release()
1736 lock.release()
1737
1737
1738 return result
1738 return result
1739
1739
1740 def checkpush(self, force, revs):
1740 def checkpush(self, force, revs):
1741 """Extensions can override this function if additional checks have
1741 """Extensions can override this function if additional checks have
1742 to be performed before pushing, or call it if they override push
1742 to be performed before pushing, or call it if they override push
1743 command.
1743 command.
1744 """
1744 """
1745 pass
1745 pass
1746
1746
1747 def push(self, remote, force=False, revs=None, newbranch=False):
1747 def push(self, remote, force=False, revs=None, newbranch=False):
1748 '''Push outgoing changesets (limited by revs) from the current
1748 '''Push outgoing changesets (limited by revs) from the current
1749 repository to remote. Return an integer:
1749 repository to remote. Return an integer:
1750 - None means nothing to push
1750 - None means nothing to push
1751 - 0 means HTTP error
1751 - 0 means HTTP error
1752 - 1 means we pushed and remote head count is unchanged *or*
1752 - 1 means we pushed and remote head count is unchanged *or*
1753 we have outgoing changesets but refused to push
1753 we have outgoing changesets but refused to push
1754 - other values as described by addchangegroup()
1754 - other values as described by addchangegroup()
1755 '''
1755 '''
1756 # there are two ways to push to remote repo:
1756 # there are two ways to push to remote repo:
1757 #
1757 #
1758 # addchangegroup assumes local user can lock remote
1758 # addchangegroup assumes local user can lock remote
1759 # repo (local filesystem, old ssh servers).
1759 # repo (local filesystem, old ssh servers).
1760 #
1760 #
1761 # unbundle assumes local user cannot lock remote repo (new ssh
1761 # unbundle assumes local user cannot lock remote repo (new ssh
1762 # servers, http servers).
1762 # servers, http servers).
1763
1763
1764 if not remote.canpush():
1764 if not remote.canpush():
1765 raise util.Abort(_("destination does not support push"))
1765 raise util.Abort(_("destination does not support push"))
1766 unfi = self.unfiltered()
1766 unfi = self.unfiltered()
1767 # get local lock as we might write phase data
1767 # get local lock as we might write phase data
1768 locallock = self.lock()
1768 locallock = self.lock()
1769 try:
1769 try:
1770 self.checkpush(force, revs)
1770 self.checkpush(force, revs)
1771 lock = None
1771 lock = None
1772 unbundle = remote.capable('unbundle')
1772 unbundle = remote.capable('unbundle')
1773 if not unbundle:
1773 if not unbundle:
1774 lock = remote.lock()
1774 lock = remote.lock()
1775 try:
1775 try:
1776 # discovery
1776 # discovery
1777 fci = discovery.findcommonincoming
1777 fci = discovery.findcommonincoming
1778 commoninc = fci(unfi, remote, force=force)
1778 commoninc = fci(unfi, remote, force=force)
1779 common, inc, remoteheads = commoninc
1779 common, inc, remoteheads = commoninc
1780 fco = discovery.findcommonoutgoing
1780 fco = discovery.findcommonoutgoing
1781 outgoing = fco(unfi, remote, onlyheads=revs,
1781 outgoing = fco(unfi, remote, onlyheads=revs,
1782 commoninc=commoninc, force=force)
1782 commoninc=commoninc, force=force)
1783
1783
1784
1784
1785 if not outgoing.missing:
1785 if not outgoing.missing:
1786 # nothing to push
1786 # nothing to push
1787 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1787 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1788 ret = None
1788 ret = None
1789 else:
1789 else:
1790 # something to push
1790 # something to push
1791 if not force:
1791 if not force:
1792 # if self.obsstore == False --> no obsolete
1792 # if self.obsstore == False --> no obsolete
1793 # then, save the iteration
1793 # then, save the iteration
1794 if unfi.obsstore:
1794 if unfi.obsstore:
1795 # this message are here for 80 char limit reason
1795 # this message are here for 80 char limit reason
1796 mso = _("push includes obsolete changeset: %s!")
1796 mso = _("push includes obsolete changeset: %s!")
1797 mst = "push includes %s changeset: %s!"
1797 mst = "push includes %s changeset: %s!"
1798 # plain versions for i18n tool to detect them
1798 # plain versions for i18n tool to detect them
1799 _("push includes unstable changeset: %s!")
1799 _("push includes unstable changeset: %s!")
1800 _("push includes bumped changeset: %s!")
1800 _("push includes bumped changeset: %s!")
1801 _("push includes divergent changeset: %s!")
1801 _("push includes divergent changeset: %s!")
1802 # If we are to push if there is at least one
1802 # If we are to push if there is at least one
1803 # obsolete or unstable changeset in missing, at
1803 # obsolete or unstable changeset in missing, at
1804 # least one of the missinghead will be obsolete or
1804 # least one of the missinghead will be obsolete or
1805 # unstable. So checking heads only is ok
1805 # unstable. So checking heads only is ok
1806 for node in outgoing.missingheads:
1806 for node in outgoing.missingheads:
1807 ctx = unfi[node]
1807 ctx = unfi[node]
1808 if ctx.obsolete():
1808 if ctx.obsolete():
1809 raise util.Abort(mso % ctx)
1809 raise util.Abort(mso % ctx)
1810 elif ctx.troubled():
1810 elif ctx.troubled():
1811 raise util.Abort(_(mst)
1811 raise util.Abort(_(mst)
1812 % (ctx.troubles()[0],
1812 % (ctx.troubles()[0],
1813 ctx))
1813 ctx))
1814 discovery.checkheads(unfi, remote, outgoing,
1814 discovery.checkheads(unfi, remote, outgoing,
1815 remoteheads, newbranch,
1815 remoteheads, newbranch,
1816 bool(inc))
1816 bool(inc))
1817
1817
1818 # create a changegroup from local
1818 # create a changegroup from local
1819 if revs is None and not outgoing.excluded:
1819 if revs is None and not outgoing.excluded:
1820 # push everything,
1820 # push everything,
1821 # use the fast path, no race possible on push
1821 # use the fast path, no race possible on push
1822 cg = self._changegroup(outgoing.missing, 'push')
1822 cg = self._changegroup(outgoing.missing, 'push')
1823 else:
1823 else:
1824 cg = self.getlocalbundle('push', outgoing)
1824 cg = self.getlocalbundle('push', outgoing)
1825
1825
1826 # apply changegroup to remote
1826 # apply changegroup to remote
1827 if unbundle:
1827 if unbundle:
1828 # local repo finds heads on server, finds out what
1828 # local repo finds heads on server, finds out what
1829 # revs it must push. once revs transferred, if server
1829 # revs it must push. once revs transferred, if server
1830 # finds it has different heads (someone else won
1830 # finds it has different heads (someone else won
1831 # commit/push race), server aborts.
1831 # commit/push race), server aborts.
1832 if force:
1832 if force:
1833 remoteheads = ['force']
1833 remoteheads = ['force']
1834 # ssh: return remote's addchangegroup()
1834 # ssh: return remote's addchangegroup()
1835 # http: return remote's addchangegroup() or 0 for error
1835 # http: return remote's addchangegroup() or 0 for error
1836 ret = remote.unbundle(cg, remoteheads, 'push')
1836 ret = remote.unbundle(cg, remoteheads, 'push')
1837 else:
1837 else:
1838 # we return an integer indicating remote head count
1838 # we return an integer indicating remote head count
1839 # change
1839 # change
1840 ret = remote.addchangegroup(cg, 'push', self.url())
1840 ret = remote.addchangegroup(cg, 'push', self.url())
1841
1841
1842 if ret:
1842 if ret:
1843 # push succeed, synchronize target of the push
1843 # push succeed, synchronize target of the push
1844 cheads = outgoing.missingheads
1844 cheads = outgoing.missingheads
1845 elif revs is None:
1845 elif revs is None:
1846 # All out push fails. synchronize all common
1846 # All out push fails. synchronize all common
1847 cheads = outgoing.commonheads
1847 cheads = outgoing.commonheads
1848 else:
1848 else:
1849 # I want cheads = heads(::missingheads and ::commonheads)
1849 # I want cheads = heads(::missingheads and ::commonheads)
1850 # (missingheads is revs with secret changeset filtered out)
1850 # (missingheads is revs with secret changeset filtered out)
1851 #
1851 #
1852 # This can be expressed as:
1852 # This can be expressed as:
1853 # cheads = ( (missingheads and ::commonheads)
1853 # cheads = ( (missingheads and ::commonheads)
1854 # + (commonheads and ::missingheads))"
1854 # + (commonheads and ::missingheads))"
1855 # )
1855 # )
1856 #
1856 #
1857 # while trying to push we already computed the following:
1857 # while trying to push we already computed the following:
1858 # common = (::commonheads)
1858 # common = (::commonheads)
1859 # missing = ((commonheads::missingheads) - commonheads)
1859 # missing = ((commonheads::missingheads) - commonheads)
1860 #
1860 #
1861 # We can pick:
1861 # We can pick:
1862 # * missingheads part of common (::commonheads)
1862 # * missingheads part of common (::commonheads)
1863 common = set(outgoing.common)
1863 common = set(outgoing.common)
1864 cheads = [node for node in revs if node in common]
1864 cheads = [node for node in revs if node in common]
1865 # and
1865 # and
1866 # * commonheads parents on missing
1866 # * commonheads parents on missing
1867 revset = unfi.set('%ln and parents(roots(%ln))',
1867 revset = unfi.set('%ln and parents(roots(%ln))',
1868 outgoing.commonheads,
1868 outgoing.commonheads,
1869 outgoing.missing)
1869 outgoing.missing)
1870 cheads.extend(c.node() for c in revset)
1870 cheads.extend(c.node() for c in revset)
1871 # even when we don't push, exchanging phase data is useful
1871 # even when we don't push, exchanging phase data is useful
1872 remotephases = remote.listkeys('phases')
1872 remotephases = remote.listkeys('phases')
1873 if not remotephases: # old server or public only repo
1873 if not remotephases: # old server or public only repo
1874 phases.advanceboundary(self, phases.public, cheads)
1874 phases.advanceboundary(self, phases.public, cheads)
1875 # don't push any phase data as there is nothing to push
1875 # don't push any phase data as there is nothing to push
1876 else:
1876 else:
1877 ana = phases.analyzeremotephases(self, cheads, remotephases)
1877 ana = phases.analyzeremotephases(self, cheads, remotephases)
1878 pheads, droots = ana
1878 pheads, droots = ana
1879 ### Apply remote phase on local
1879 ### Apply remote phase on local
1880 if remotephases.get('publishing', False):
1880 if remotephases.get('publishing', False):
1881 phases.advanceboundary(self, phases.public, cheads)
1881 phases.advanceboundary(self, phases.public, cheads)
1882 else: # publish = False
1882 else: # publish = False
1883 phases.advanceboundary(self, phases.public, pheads)
1883 phases.advanceboundary(self, phases.public, pheads)
1884 phases.advanceboundary(self, phases.draft, cheads)
1884 phases.advanceboundary(self, phases.draft, cheads)
1885 ### Apply local phase on remote
1885 ### Apply local phase on remote
1886
1886
1887 # Get the list of all revs draft on remote by public here.
1887 # Get the list of all revs draft on remote by public here.
1888 # XXX Beware that revset break if droots is not strictly
1888 # XXX Beware that revset break if droots is not strictly
1889 # XXX root we may want to ensure it is but it is costly
1889 # XXX root we may want to ensure it is but it is costly
1890 outdated = unfi.set('heads((%ln::%ln) and public())',
1890 outdated = unfi.set('heads((%ln::%ln) and public())',
1891 droots, cheads)
1891 droots, cheads)
1892 for newremotehead in outdated:
1892 for newremotehead in outdated:
1893 r = remote.pushkey('phases',
1893 r = remote.pushkey('phases',
1894 newremotehead.hex(),
1894 newremotehead.hex(),
1895 str(phases.draft),
1895 str(phases.draft),
1896 str(phases.public))
1896 str(phases.public))
1897 if not r:
1897 if not r:
1898 self.ui.warn(_('updating %s to public failed!\n')
1898 self.ui.warn(_('updating %s to public failed!\n')
1899 % newremotehead)
1899 % newremotehead)
1900 self.ui.debug('try to push obsolete markers to remote\n')
1900 self.ui.debug('try to push obsolete markers to remote\n')
1901 if (obsolete._enabled and self.obsstore and
1901 if (obsolete._enabled and self.obsstore and
1902 'obsolete' in remote.listkeys('namespaces')):
1902 'obsolete' in remote.listkeys('namespaces')):
1903 rslts = []
1903 rslts = []
1904 remotedata = self.listkeys('obsolete')
1904 remotedata = self.listkeys('obsolete')
1905 for key in sorted(remotedata, reverse=True):
1905 for key in sorted(remotedata, reverse=True):
1906 # reverse sort to ensure we end with dump0
1906 # reverse sort to ensure we end with dump0
1907 data = remotedata[key]
1907 data = remotedata[key]
1908 rslts.append(remote.pushkey('obsolete', key, '', data))
1908 rslts.append(remote.pushkey('obsolete', key, '', data))
1909 if [r for r in rslts if not r]:
1909 if [r for r in rslts if not r]:
1910 msg = _('failed to push some obsolete markers!\n')
1910 msg = _('failed to push some obsolete markers!\n')
1911 self.ui.warn(msg)
1911 self.ui.warn(msg)
1912 finally:
1912 finally:
1913 if lock is not None:
1913 if lock is not None:
1914 lock.release()
1914 lock.release()
1915 finally:
1915 finally:
1916 locallock.release()
1916 locallock.release()
1917
1917
1918 self.ui.debug("checking for updated bookmarks\n")
1918 self.ui.debug("checking for updated bookmarks\n")
1919 rb = remote.listkeys('bookmarks')
1919 rb = remote.listkeys('bookmarks')
1920 for k in rb.keys():
1920 for k in rb.keys():
1921 if k in unfi._bookmarks:
1921 if k in unfi._bookmarks:
1922 nr, nl = rb[k], hex(self._bookmarks[k])
1922 nr, nl = rb[k], hex(self._bookmarks[k])
1923 if nr in unfi:
1923 if nr in unfi:
1924 cr = unfi[nr]
1924 cr = unfi[nr]
1925 cl = unfi[nl]
1925 cl = unfi[nl]
1926 if bookmarks.validdest(unfi, cr, cl):
1926 if bookmarks.validdest(unfi, cr, cl):
1927 r = remote.pushkey('bookmarks', k, nr, nl)
1927 r = remote.pushkey('bookmarks', k, nr, nl)
1928 if r:
1928 if r:
1929 self.ui.status(_("updating bookmark %s\n") % k)
1929 self.ui.status(_("updating bookmark %s\n") % k)
1930 else:
1930 else:
1931 self.ui.warn(_('updating bookmark %s'
1931 self.ui.warn(_('updating bookmark %s'
1932 ' failed!\n') % k)
1932 ' failed!\n') % k)
1933
1933
1934 return ret
1934 return ret
1935
1935
1936 def changegroupinfo(self, nodes, source):
1936 def changegroupinfo(self, nodes, source):
1937 if self.ui.verbose or source == 'bundle':
1937 if self.ui.verbose or source == 'bundle':
1938 self.ui.status(_("%d changesets found\n") % len(nodes))
1938 self.ui.status(_("%d changesets found\n") % len(nodes))
1939 if self.ui.debugflag:
1939 if self.ui.debugflag:
1940 self.ui.debug("list of changesets:\n")
1940 self.ui.debug("list of changesets:\n")
1941 for node in nodes:
1941 for node in nodes:
1942 self.ui.debug("%s\n" % hex(node))
1942 self.ui.debug("%s\n" % hex(node))
1943
1943
1944 def changegroupsubset(self, bases, heads, source):
1944 def changegroupsubset(self, bases, heads, source):
1945 """Compute a changegroup consisting of all the nodes that are
1945 """Compute a changegroup consisting of all the nodes that are
1946 descendants of any of the bases and ancestors of any of the heads.
1946 descendants of any of the bases and ancestors of any of the heads.
1947 Return a chunkbuffer object whose read() method will return
1947 Return a chunkbuffer object whose read() method will return
1948 successive changegroup chunks.
1948 successive changegroup chunks.
1949
1949
1950 It is fairly complex as determining which filenodes and which
1950 It is fairly complex as determining which filenodes and which
1951 manifest nodes need to be included for the changeset to be complete
1951 manifest nodes need to be included for the changeset to be complete
1952 is non-trivial.
1952 is non-trivial.
1953
1953
1954 Another wrinkle is doing the reverse, figuring out which changeset in
1954 Another wrinkle is doing the reverse, figuring out which changeset in
1955 the changegroup a particular filenode or manifestnode belongs to.
1955 the changegroup a particular filenode or manifestnode belongs to.
1956 """
1956 """
1957 cl = self.changelog
1957 cl = self.changelog
1958 if not bases:
1958 if not bases:
1959 bases = [nullid]
1959 bases = [nullid]
1960 csets, bases, heads = cl.nodesbetween(bases, heads)
1960 csets, bases, heads = cl.nodesbetween(bases, heads)
1961 # We assume that all ancestors of bases are known
1961 # We assume that all ancestors of bases are known
1962 common = cl.ancestors([cl.rev(n) for n in bases])
1962 common = cl.ancestors([cl.rev(n) for n in bases])
1963 return self._changegroupsubset(common, csets, heads, source)
1963 return self._changegroupsubset(common, csets, heads, source)
1964
1964
1965 def getlocalbundle(self, source, outgoing):
1965 def getlocalbundle(self, source, outgoing):
1966 """Like getbundle, but taking a discovery.outgoing as an argument.
1966 """Like getbundle, but taking a discovery.outgoing as an argument.
1967
1967
1968 This is only implemented for local repos and reuses potentially
1968 This is only implemented for local repos and reuses potentially
1969 precomputed sets in outgoing."""
1969 precomputed sets in outgoing."""
1970 if not outgoing.missing:
1970 if not outgoing.missing:
1971 return None
1971 return None
1972 return self._changegroupsubset(outgoing.common,
1972 return self._changegroupsubset(outgoing.common,
1973 outgoing.missing,
1973 outgoing.missing,
1974 outgoing.missingheads,
1974 outgoing.missingheads,
1975 source)
1975 source)
1976
1976
1977 def getbundle(self, source, heads=None, common=None):
1977 def getbundle(self, source, heads=None, common=None):
1978 """Like changegroupsubset, but returns the set difference between the
1978 """Like changegroupsubset, but returns the set difference between the
1979 ancestors of heads and the ancestors common.
1979 ancestors of heads and the ancestors common.
1980
1980
1981 If heads is None, use the local heads. If common is None, use [nullid].
1981 If heads is None, use the local heads. If common is None, use [nullid].
1982
1982
1983 The nodes in common might not all be known locally due to the way the
1983 The nodes in common might not all be known locally due to the way the
1984 current discovery protocol works.
1984 current discovery protocol works.
1985 """
1985 """
1986 cl = self.changelog
1986 cl = self.changelog
1987 if common:
1987 if common:
1988 hasnode = cl.hasnode
1988 hasnode = cl.hasnode
1989 common = [n for n in common if hasnode(n)]
1989 common = [n for n in common if hasnode(n)]
1990 else:
1990 else:
1991 common = [nullid]
1991 common = [nullid]
1992 if not heads:
1992 if not heads:
1993 heads = cl.heads()
1993 heads = cl.heads()
1994 return self.getlocalbundle(source,
1994 return self.getlocalbundle(source,
1995 discovery.outgoing(cl, common, heads))
1995 discovery.outgoing(cl, common, heads))
1996
1996
1997 @unfilteredmethod
1997 @unfilteredmethod
1998 def _changegroupsubset(self, commonrevs, csets, heads, source):
1998 def _changegroupsubset(self, commonrevs, csets, heads, source):
1999
1999
2000 cl = self.changelog
2000 cl = self.changelog
2001 mf = self.manifest
2001 mf = self.manifest
2002 mfs = {} # needed manifests
2002 mfs = {} # needed manifests
2003 fnodes = {} # needed file nodes
2003 fnodes = {} # needed file nodes
2004 changedfiles = set()
2004 changedfiles = set()
2005 fstate = ['', {}]
2005 fstate = ['', {}]
2006 count = [0, 0]
2006 count = [0, 0]
2007
2007
2008 # can we go through the fast path ?
2008 # can we go through the fast path ?
2009 heads.sort()
2009 heads.sort()
2010 if heads == sorted(self.heads()):
2010 if heads == sorted(self.heads()):
2011 return self._changegroup(csets, source)
2011 return self._changegroup(csets, source)
2012
2012
2013 # slow path
2013 # slow path
2014 self.hook('preoutgoing', throw=True, source=source)
2014 self.hook('preoutgoing', throw=True, source=source)
2015 self.changegroupinfo(csets, source)
2015 self.changegroupinfo(csets, source)
2016
2016
2017 # filter any nodes that claim to be part of the known set
2017 # filter any nodes that claim to be part of the known set
2018 def prune(revlog, missing):
2018 def prune(revlog, missing):
2019 rr, rl = revlog.rev, revlog.linkrev
2019 rr, rl = revlog.rev, revlog.linkrev
2020 return [n for n in missing
2020 return [n for n in missing
2021 if rl(rr(n)) not in commonrevs]
2021 if rl(rr(n)) not in commonrevs]
2022
2022
2023 progress = self.ui.progress
2023 progress = self.ui.progress
2024 _bundling = _('bundling')
2024 _bundling = _('bundling')
2025 _changesets = _('changesets')
2025 _changesets = _('changesets')
2026 _manifests = _('manifests')
2026 _manifests = _('manifests')
2027 _files = _('files')
2027 _files = _('files')
2028
2028
2029 def lookup(revlog, x):
2029 def lookup(revlog, x):
2030 if revlog == cl:
2030 if revlog == cl:
2031 c = cl.read(x)
2031 c = cl.read(x)
2032 changedfiles.update(c[3])
2032 changedfiles.update(c[3])
2033 mfs.setdefault(c[0], x)
2033 mfs.setdefault(c[0], x)
2034 count[0] += 1
2034 count[0] += 1
2035 progress(_bundling, count[0],
2035 progress(_bundling, count[0],
2036 unit=_changesets, total=count[1])
2036 unit=_changesets, total=count[1])
2037 return x
2037 return x
2038 elif revlog == mf:
2038 elif revlog == mf:
2039 clnode = mfs[x]
2039 clnode = mfs[x]
2040 mdata = mf.readfast(x)
2040 mdata = mf.readfast(x)
2041 for f, n in mdata.iteritems():
2041 for f, n in mdata.iteritems():
2042 if f in changedfiles:
2042 if f in changedfiles:
2043 fnodes[f].setdefault(n, clnode)
2043 fnodes[f].setdefault(n, clnode)
2044 count[0] += 1
2044 count[0] += 1
2045 progress(_bundling, count[0],
2045 progress(_bundling, count[0],
2046 unit=_manifests, total=count[1])
2046 unit=_manifests, total=count[1])
2047 return clnode
2047 return clnode
2048 else:
2048 else:
2049 progress(_bundling, count[0], item=fstate[0],
2049 progress(_bundling, count[0], item=fstate[0],
2050 unit=_files, total=count[1])
2050 unit=_files, total=count[1])
2051 return fstate[1][x]
2051 return fstate[1][x]
2052
2052
2053 bundler = changegroup.bundle10(lookup)
2053 bundler = changegroup.bundle10(lookup)
2054 reorder = self.ui.config('bundle', 'reorder', 'auto')
2054 reorder = self.ui.config('bundle', 'reorder', 'auto')
2055 if reorder == 'auto':
2055 if reorder == 'auto':
2056 reorder = None
2056 reorder = None
2057 else:
2057 else:
2058 reorder = util.parsebool(reorder)
2058 reorder = util.parsebool(reorder)
2059
2059
2060 def gengroup():
2060 def gengroup():
2061 # Create a changenode group generator that will call our functions
2061 # Create a changenode group generator that will call our functions
2062 # back to lookup the owning changenode and collect information.
2062 # back to lookup the owning changenode and collect information.
2063 count[:] = [0, len(csets)]
2063 count[:] = [0, len(csets)]
2064 for chunk in cl.group(csets, bundler, reorder=reorder):
2064 for chunk in cl.group(csets, bundler, reorder=reorder):
2065 yield chunk
2065 yield chunk
2066 progress(_bundling, None)
2066 progress(_bundling, None)
2067
2067
2068 # Create a generator for the manifestnodes that calls our lookup
2068 # Create a generator for the manifestnodes that calls our lookup
2069 # and data collection functions back.
2069 # and data collection functions back.
2070 for f in changedfiles:
2070 for f in changedfiles:
2071 fnodes[f] = {}
2071 fnodes[f] = {}
2072 count[:] = [0, len(mfs)]
2072 count[:] = [0, len(mfs)]
2073 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2073 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2074 yield chunk
2074 yield chunk
2075 progress(_bundling, None)
2075 progress(_bundling, None)
2076
2076
2077 mfs.clear()
2077 mfs.clear()
2078
2078
2079 # Go through all our files in order sorted by name.
2079 # Go through all our files in order sorted by name.
2080 count[:] = [0, len(changedfiles)]
2080 count[:] = [0, len(changedfiles)]
2081 for fname in sorted(changedfiles):
2081 for fname in sorted(changedfiles):
2082 filerevlog = self.file(fname)
2082 filerevlog = self.file(fname)
2083 if not len(filerevlog):
2083 if not len(filerevlog):
2084 raise util.Abort(_("empty or missing revlog for %s")
2084 raise util.Abort(_("empty or missing revlog for %s")
2085 % fname)
2085 % fname)
2086 fstate[0] = fname
2086 fstate[0] = fname
2087 fstate[1] = fnodes.pop(fname, {})
2087 fstate[1] = fnodes.pop(fname, {})
2088
2088
2089 nodelist = prune(filerevlog, fstate[1])
2089 nodelist = prune(filerevlog, fstate[1])
2090 if nodelist:
2090 if nodelist:
2091 count[0] += 1
2091 count[0] += 1
2092 yield bundler.fileheader(fname)
2092 yield bundler.fileheader(fname)
2093 for chunk in filerevlog.group(nodelist, bundler, reorder):
2093 for chunk in filerevlog.group(nodelist, bundler, reorder):
2094 yield chunk
2094 yield chunk
2095
2095
2096 # Signal that no more groups are left.
2096 # Signal that no more groups are left.
2097 yield bundler.close()
2097 yield bundler.close()
2098 progress(_bundling, None)
2098 progress(_bundling, None)
2099
2099
2100 if csets:
2100 if csets:
2101 self.hook('outgoing', node=hex(csets[0]), source=source)
2101 self.hook('outgoing', node=hex(csets[0]), source=source)
2102
2102
2103 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2103 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2104
2104
2105 def changegroup(self, basenodes, source):
2105 def changegroup(self, basenodes, source):
2106 # to avoid a race we use changegroupsubset() (issue1320)
2106 # to avoid a race we use changegroupsubset() (issue1320)
2107 return self.changegroupsubset(basenodes, self.heads(), source)
2107 return self.changegroupsubset(basenodes, self.heads(), source)
2108
2108
2109 @unfilteredmethod
2109 @unfilteredmethod
2110 def _changegroup(self, nodes, source):
2110 def _changegroup(self, nodes, source):
2111 """Compute the changegroup of all nodes that we have that a recipient
2111 """Compute the changegroup of all nodes that we have that a recipient
2112 doesn't. Return a chunkbuffer object whose read() method will return
2112 doesn't. Return a chunkbuffer object whose read() method will return
2113 successive changegroup chunks.
2113 successive changegroup chunks.
2114
2114
2115 This is much easier than the previous function as we can assume that
2115 This is much easier than the previous function as we can assume that
2116 the recipient has any changenode we aren't sending them.
2116 the recipient has any changenode we aren't sending them.
2117
2117
2118 nodes is the set of nodes to send"""
2118 nodes is the set of nodes to send"""
2119
2119
2120 cl = self.changelog
2120 cl = self.changelog
2121 mf = self.manifest
2121 mf = self.manifest
2122 mfs = {}
2122 mfs = {}
2123 changedfiles = set()
2123 changedfiles = set()
2124 fstate = ['']
2124 fstate = ['']
2125 count = [0, 0]
2125 count = [0, 0]
2126
2126
2127 self.hook('preoutgoing', throw=True, source=source)
2127 self.hook('preoutgoing', throw=True, source=source)
2128 self.changegroupinfo(nodes, source)
2128 self.changegroupinfo(nodes, source)
2129
2129
2130 revset = set([cl.rev(n) for n in nodes])
2130 revset = set([cl.rev(n) for n in nodes])
2131
2131
2132 def gennodelst(log):
2132 def gennodelst(log):
2133 ln, llr = log.node, log.linkrev
2133 ln, llr = log.node, log.linkrev
2134 return [ln(r) for r in log if llr(r) in revset]
2134 return [ln(r) for r in log if llr(r) in revset]
2135
2135
2136 progress = self.ui.progress
2136 progress = self.ui.progress
2137 _bundling = _('bundling')
2137 _bundling = _('bundling')
2138 _changesets = _('changesets')
2138 _changesets = _('changesets')
2139 _manifests = _('manifests')
2139 _manifests = _('manifests')
2140 _files = _('files')
2140 _files = _('files')
2141
2141
2142 def lookup(revlog, x):
2142 def lookup(revlog, x):
2143 if revlog == cl:
2143 if revlog == cl:
2144 c = cl.read(x)
2144 c = cl.read(x)
2145 changedfiles.update(c[3])
2145 changedfiles.update(c[3])
2146 mfs.setdefault(c[0], x)
2146 mfs.setdefault(c[0], x)
2147 count[0] += 1
2147 count[0] += 1
2148 progress(_bundling, count[0],
2148 progress(_bundling, count[0],
2149 unit=_changesets, total=count[1])
2149 unit=_changesets, total=count[1])
2150 return x
2150 return x
2151 elif revlog == mf:
2151 elif revlog == mf:
2152 count[0] += 1
2152 count[0] += 1
2153 progress(_bundling, count[0],
2153 progress(_bundling, count[0],
2154 unit=_manifests, total=count[1])
2154 unit=_manifests, total=count[1])
2155 return cl.node(revlog.linkrev(revlog.rev(x)))
2155 return cl.node(revlog.linkrev(revlog.rev(x)))
2156 else:
2156 else:
2157 progress(_bundling, count[0], item=fstate[0],
2157 progress(_bundling, count[0], item=fstate[0],
2158 total=count[1], unit=_files)
2158 total=count[1], unit=_files)
2159 return cl.node(revlog.linkrev(revlog.rev(x)))
2159 return cl.node(revlog.linkrev(revlog.rev(x)))
2160
2160
2161 bundler = changegroup.bundle10(lookup)
2161 bundler = changegroup.bundle10(lookup)
2162 reorder = self.ui.config('bundle', 'reorder', 'auto')
2162 reorder = self.ui.config('bundle', 'reorder', 'auto')
2163 if reorder == 'auto':
2163 if reorder == 'auto':
2164 reorder = None
2164 reorder = None
2165 else:
2165 else:
2166 reorder = util.parsebool(reorder)
2166 reorder = util.parsebool(reorder)
2167
2167
2168 def gengroup():
2168 def gengroup():
2169 '''yield a sequence of changegroup chunks (strings)'''
2169 '''yield a sequence of changegroup chunks (strings)'''
2170 # construct a list of all changed files
2170 # construct a list of all changed files
2171
2171
2172 count[:] = [0, len(nodes)]
2172 count[:] = [0, len(nodes)]
2173 for chunk in cl.group(nodes, bundler, reorder=reorder):
2173 for chunk in cl.group(nodes, bundler, reorder=reorder):
2174 yield chunk
2174 yield chunk
2175 progress(_bundling, None)
2175 progress(_bundling, None)
2176
2176
2177 count[:] = [0, len(mfs)]
2177 count[:] = [0, len(mfs)]
2178 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2178 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2179 yield chunk
2179 yield chunk
2180 progress(_bundling, None)
2180 progress(_bundling, None)
2181
2181
2182 count[:] = [0, len(changedfiles)]
2182 count[:] = [0, len(changedfiles)]
2183 for fname in sorted(changedfiles):
2183 for fname in sorted(changedfiles):
2184 filerevlog = self.file(fname)
2184 filerevlog = self.file(fname)
2185 if not len(filerevlog):
2185 if not len(filerevlog):
2186 raise util.Abort(_("empty or missing revlog for %s")
2186 raise util.Abort(_("empty or missing revlog for %s")
2187 % fname)
2187 % fname)
2188 fstate[0] = fname
2188 fstate[0] = fname
2189 nodelist = gennodelst(filerevlog)
2189 nodelist = gennodelst(filerevlog)
2190 if nodelist:
2190 if nodelist:
2191 count[0] += 1
2191 count[0] += 1
2192 yield bundler.fileheader(fname)
2192 yield bundler.fileheader(fname)
2193 for chunk in filerevlog.group(nodelist, bundler, reorder):
2193 for chunk in filerevlog.group(nodelist, bundler, reorder):
2194 yield chunk
2194 yield chunk
2195 yield bundler.close()
2195 yield bundler.close()
2196 progress(_bundling, None)
2196 progress(_bundling, None)
2197
2197
2198 if nodes:
2198 if nodes:
2199 self.hook('outgoing', node=hex(nodes[0]), source=source)
2199 self.hook('outgoing', node=hex(nodes[0]), source=source)
2200
2200
2201 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2201 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2202
2202
2203 @unfilteredmethod
2203 @unfilteredmethod
2204 def addchangegroup(self, source, srctype, url, emptyok=False):
2204 def addchangegroup(self, source, srctype, url, emptyok=False):
2205 """Add the changegroup returned by source.read() to this repo.
2205 """Add the changegroup returned by source.read() to this repo.
2206 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2206 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2207 the URL of the repo where this changegroup is coming from.
2207 the URL of the repo where this changegroup is coming from.
2208
2208
2209 Return an integer summarizing the change to this repo:
2209 Return an integer summarizing the change to this repo:
2210 - nothing changed or no source: 0
2210 - nothing changed or no source: 0
2211 - more heads than before: 1+added heads (2..n)
2211 - more heads than before: 1+added heads (2..n)
2212 - fewer heads than before: -1-removed heads (-2..-n)
2212 - fewer heads than before: -1-removed heads (-2..-n)
2213 - number of heads stays the same: 1
2213 - number of heads stays the same: 1
2214 """
2214 """
2215 def csmap(x):
2215 def csmap(x):
2216 self.ui.debug("add changeset %s\n" % short(x))
2216 self.ui.debug("add changeset %s\n" % short(x))
2217 return len(cl)
2217 return len(cl)
2218
2218
2219 def revmap(x):
2219 def revmap(x):
2220 return cl.rev(x)
2220 return cl.rev(x)
2221
2221
2222 if not source:
2222 if not source:
2223 return 0
2223 return 0
2224
2224
2225 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2225 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2226
2226
2227 changesets = files = revisions = 0
2227 changesets = files = revisions = 0
2228 efiles = set()
2228 efiles = set()
2229
2229
2230 # write changelog data to temp files so concurrent readers will not see
2230 # write changelog data to temp files so concurrent readers will not see
2231 # inconsistent view
2231 # inconsistent view
2232 cl = self.changelog
2232 cl = self.changelog
2233 cl.delayupdate()
2233 cl.delayupdate()
2234 oldheads = cl.heads()
2234 oldheads = cl.heads()
2235
2235
2236 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2236 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2237 try:
2237 try:
2238 trp = weakref.proxy(tr)
2238 trp = weakref.proxy(tr)
2239 # pull off the changeset group
2239 # pull off the changeset group
2240 self.ui.status(_("adding changesets\n"))
2240 self.ui.status(_("adding changesets\n"))
2241 clstart = len(cl)
2241 clstart = len(cl)
2242 class prog(object):
2242 class prog(object):
2243 step = _('changesets')
2243 step = _('changesets')
2244 count = 1
2244 count = 1
2245 ui = self.ui
2245 ui = self.ui
2246 total = None
2246 total = None
2247 def __call__(self):
2247 def __call__(self):
2248 self.ui.progress(self.step, self.count, unit=_('chunks'),
2248 self.ui.progress(self.step, self.count, unit=_('chunks'),
2249 total=self.total)
2249 total=self.total)
2250 self.count += 1
2250 self.count += 1
2251 pr = prog()
2251 pr = prog()
2252 source.callback = pr
2252 source.callback = pr
2253
2253
2254 source.changelogheader()
2254 source.changelogheader()
2255 srccontent = cl.addgroup(source, csmap, trp)
2255 srccontent = cl.addgroup(source, csmap, trp)
2256 if not (srccontent or emptyok):
2256 if not (srccontent or emptyok):
2257 raise util.Abort(_("received changelog group is empty"))
2257 raise util.Abort(_("received changelog group is empty"))
2258 clend = len(cl)
2258 clend = len(cl)
2259 changesets = clend - clstart
2259 changesets = clend - clstart
2260 for c in xrange(clstart, clend):
2260 for c in xrange(clstart, clend):
2261 efiles.update(self[c].files())
2261 efiles.update(self[c].files())
2262 efiles = len(efiles)
2262 efiles = len(efiles)
2263 self.ui.progress(_('changesets'), None)
2263 self.ui.progress(_('changesets'), None)
2264
2264
2265 # pull off the manifest group
2265 # pull off the manifest group
2266 self.ui.status(_("adding manifests\n"))
2266 self.ui.status(_("adding manifests\n"))
2267 pr.step = _('manifests')
2267 pr.step = _('manifests')
2268 pr.count = 1
2268 pr.count = 1
2269 pr.total = changesets # manifests <= changesets
2269 pr.total = changesets # manifests <= changesets
2270 # no need to check for empty manifest group here:
2270 # no need to check for empty manifest group here:
2271 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2271 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2272 # no new manifest will be created and the manifest group will
2272 # no new manifest will be created and the manifest group will
2273 # be empty during the pull
2273 # be empty during the pull
2274 source.manifestheader()
2274 source.manifestheader()
2275 self.manifest.addgroup(source, revmap, trp)
2275 self.manifest.addgroup(source, revmap, trp)
2276 self.ui.progress(_('manifests'), None)
2276 self.ui.progress(_('manifests'), None)
2277
2277
2278 needfiles = {}
2278 needfiles = {}
2279 if self.ui.configbool('server', 'validate', default=False):
2279 if self.ui.configbool('server', 'validate', default=False):
2280 # validate incoming csets have their manifests
2280 # validate incoming csets have their manifests
2281 for cset in xrange(clstart, clend):
2281 for cset in xrange(clstart, clend):
2282 mfest = self.changelog.read(self.changelog.node(cset))[0]
2282 mfest = self.changelog.read(self.changelog.node(cset))[0]
2283 mfest = self.manifest.readdelta(mfest)
2283 mfest = self.manifest.readdelta(mfest)
2284 # store file nodes we must see
2284 # store file nodes we must see
2285 for f, n in mfest.iteritems():
2285 for f, n in mfest.iteritems():
2286 needfiles.setdefault(f, set()).add(n)
2286 needfiles.setdefault(f, set()).add(n)
2287
2287
2288 # process the files
2288 # process the files
2289 self.ui.status(_("adding file changes\n"))
2289 self.ui.status(_("adding file changes\n"))
2290 pr.step = _('files')
2290 pr.step = _('files')
2291 pr.count = 1
2291 pr.count = 1
2292 pr.total = efiles
2292 pr.total = efiles
2293 source.callback = None
2293 source.callback = None
2294
2294
2295 while True:
2295 while True:
2296 chunkdata = source.filelogheader()
2296 chunkdata = source.filelogheader()
2297 if not chunkdata:
2297 if not chunkdata:
2298 break
2298 break
2299 f = chunkdata["filename"]
2299 f = chunkdata["filename"]
2300 self.ui.debug("adding %s revisions\n" % f)
2300 self.ui.debug("adding %s revisions\n" % f)
2301 pr()
2301 pr()
2302 fl = self.file(f)
2302 fl = self.file(f)
2303 o = len(fl)
2303 o = len(fl)
2304 if not fl.addgroup(source, revmap, trp):
2304 if not fl.addgroup(source, revmap, trp):
2305 raise util.Abort(_("received file revlog group is empty"))
2305 raise util.Abort(_("received file revlog group is empty"))
2306 revisions += len(fl) - o
2306 revisions += len(fl) - o
2307 files += 1
2307 files += 1
2308 if f in needfiles:
2308 if f in needfiles:
2309 needs = needfiles[f]
2309 needs = needfiles[f]
2310 for new in xrange(o, len(fl)):
2310 for new in xrange(o, len(fl)):
2311 n = fl.node(new)
2311 n = fl.node(new)
2312 if n in needs:
2312 if n in needs:
2313 needs.remove(n)
2313 needs.remove(n)
2314 else:
2315 raise util.Abort(
2316 _("received spurious file revlog entry"))
2314 if not needs:
2317 if not needs:
2315 del needfiles[f]
2318 del needfiles[f]
2316 self.ui.progress(_('files'), None)
2319 self.ui.progress(_('files'), None)
2317
2320
2318 for f, needs in needfiles.iteritems():
2321 for f, needs in needfiles.iteritems():
2319 fl = self.file(f)
2322 fl = self.file(f)
2320 for n in needs:
2323 for n in needs:
2321 try:
2324 try:
2322 fl.rev(n)
2325 fl.rev(n)
2323 except error.LookupError:
2326 except error.LookupError:
2324 raise util.Abort(
2327 raise util.Abort(
2325 _('missing file data for %s:%s - run hg verify') %
2328 _('missing file data for %s:%s - run hg verify') %
2326 (f, hex(n)))
2329 (f, hex(n)))
2327
2330
2328 dh = 0
2331 dh = 0
2329 if oldheads:
2332 if oldheads:
2330 heads = cl.heads()
2333 heads = cl.heads()
2331 dh = len(heads) - len(oldheads)
2334 dh = len(heads) - len(oldheads)
2332 for h in heads:
2335 for h in heads:
2333 if h not in oldheads and self[h].closesbranch():
2336 if h not in oldheads and self[h].closesbranch():
2334 dh -= 1
2337 dh -= 1
2335 htext = ""
2338 htext = ""
2336 if dh:
2339 if dh:
2337 htext = _(" (%+d heads)") % dh
2340 htext = _(" (%+d heads)") % dh
2338
2341
2339 self.ui.status(_("added %d changesets"
2342 self.ui.status(_("added %d changesets"
2340 " with %d changes to %d files%s\n")
2343 " with %d changes to %d files%s\n")
2341 % (changesets, revisions, files, htext))
2344 % (changesets, revisions, files, htext))
2342 self.invalidatevolatilesets()
2345 self.invalidatevolatilesets()
2343
2346
2344 if changesets > 0:
2347 if changesets > 0:
2345 p = lambda: cl.writepending() and self.root or ""
2348 p = lambda: cl.writepending() and self.root or ""
2346 self.hook('pretxnchangegroup', throw=True,
2349 self.hook('pretxnchangegroup', throw=True,
2347 node=hex(cl.node(clstart)), source=srctype,
2350 node=hex(cl.node(clstart)), source=srctype,
2348 url=url, pending=p)
2351 url=url, pending=p)
2349
2352
2350 added = [cl.node(r) for r in xrange(clstart, clend)]
2353 added = [cl.node(r) for r in xrange(clstart, clend)]
2351 publishing = self.ui.configbool('phases', 'publish', True)
2354 publishing = self.ui.configbool('phases', 'publish', True)
2352 if srctype == 'push':
2355 if srctype == 'push':
2353 # Old server can not push the boundary themself.
2356 # Old server can not push the boundary themself.
2354 # New server won't push the boundary if changeset already
2357 # New server won't push the boundary if changeset already
2355 # existed locally as secrete
2358 # existed locally as secrete
2356 #
2359 #
2357 # We should not use added here but the list of all change in
2360 # We should not use added here but the list of all change in
2358 # the bundle
2361 # the bundle
2359 if publishing:
2362 if publishing:
2360 phases.advanceboundary(self, phases.public, srccontent)
2363 phases.advanceboundary(self, phases.public, srccontent)
2361 else:
2364 else:
2362 phases.advanceboundary(self, phases.draft, srccontent)
2365 phases.advanceboundary(self, phases.draft, srccontent)
2363 phases.retractboundary(self, phases.draft, added)
2366 phases.retractboundary(self, phases.draft, added)
2364 elif srctype != 'strip':
2367 elif srctype != 'strip':
2365 # publishing only alter behavior during push
2368 # publishing only alter behavior during push
2366 #
2369 #
2367 # strip should not touch boundary at all
2370 # strip should not touch boundary at all
2368 phases.retractboundary(self, phases.draft, added)
2371 phases.retractboundary(self, phases.draft, added)
2369
2372
2370 # make changelog see real files again
2373 # make changelog see real files again
2371 cl.finalize(trp)
2374 cl.finalize(trp)
2372
2375
2373 tr.close()
2376 tr.close()
2374
2377
2375 if changesets > 0:
2378 if changesets > 0:
2376 if srctype != 'strip':
2379 if srctype != 'strip':
2377 # During strip, branchcache is invalid but coming call to
2380 # During strip, branchcache is invalid but coming call to
2378 # `destroyed` will repair it.
2381 # `destroyed` will repair it.
2379 # In other case we can safely update cache on disk.
2382 # In other case we can safely update cache on disk.
2380 branchmap.updatecache(self.filtered('served'))
2383 branchmap.updatecache(self.filtered('served'))
2381 def runhooks():
2384 def runhooks():
2382 # forcefully update the on-disk branch cache
2385 # forcefully update the on-disk branch cache
2383 self.ui.debug("updating the branch cache\n")
2386 self.ui.debug("updating the branch cache\n")
2384 self.hook("changegroup", node=hex(cl.node(clstart)),
2387 self.hook("changegroup", node=hex(cl.node(clstart)),
2385 source=srctype, url=url)
2388 source=srctype, url=url)
2386
2389
2387 for n in added:
2390 for n in added:
2388 self.hook("incoming", node=hex(n), source=srctype,
2391 self.hook("incoming", node=hex(n), source=srctype,
2389 url=url)
2392 url=url)
2390 self._afterlock(runhooks)
2393 self._afterlock(runhooks)
2391
2394
2392 finally:
2395 finally:
2393 tr.release()
2396 tr.release()
2394 # never return 0 here:
2397 # never return 0 here:
2395 if dh < 0:
2398 if dh < 0:
2396 return dh - 1
2399 return dh - 1
2397 else:
2400 else:
2398 return dh + 1
2401 return dh + 1
2399
2402
2400 def stream_in(self, remote, requirements):
2403 def stream_in(self, remote, requirements):
2401 lock = self.lock()
2404 lock = self.lock()
2402 try:
2405 try:
2403 # Save remote branchmap. We will use it later
2406 # Save remote branchmap. We will use it later
2404 # to speed up branchcache creation
2407 # to speed up branchcache creation
2405 rbranchmap = None
2408 rbranchmap = None
2406 if remote.capable("branchmap"):
2409 if remote.capable("branchmap"):
2407 rbranchmap = remote.branchmap()
2410 rbranchmap = remote.branchmap()
2408
2411
2409 fp = remote.stream_out()
2412 fp = remote.stream_out()
2410 l = fp.readline()
2413 l = fp.readline()
2411 try:
2414 try:
2412 resp = int(l)
2415 resp = int(l)
2413 except ValueError:
2416 except ValueError:
2414 raise error.ResponseError(
2417 raise error.ResponseError(
2415 _('unexpected response from remote server:'), l)
2418 _('unexpected response from remote server:'), l)
2416 if resp == 1:
2419 if resp == 1:
2417 raise util.Abort(_('operation forbidden by server'))
2420 raise util.Abort(_('operation forbidden by server'))
2418 elif resp == 2:
2421 elif resp == 2:
2419 raise util.Abort(_('locking the remote repository failed'))
2422 raise util.Abort(_('locking the remote repository failed'))
2420 elif resp != 0:
2423 elif resp != 0:
2421 raise util.Abort(_('the server sent an unknown error code'))
2424 raise util.Abort(_('the server sent an unknown error code'))
2422 self.ui.status(_('streaming all changes\n'))
2425 self.ui.status(_('streaming all changes\n'))
2423 l = fp.readline()
2426 l = fp.readline()
2424 try:
2427 try:
2425 total_files, total_bytes = map(int, l.split(' ', 1))
2428 total_files, total_bytes = map(int, l.split(' ', 1))
2426 except (ValueError, TypeError):
2429 except (ValueError, TypeError):
2427 raise error.ResponseError(
2430 raise error.ResponseError(
2428 _('unexpected response from remote server:'), l)
2431 _('unexpected response from remote server:'), l)
2429 self.ui.status(_('%d files to transfer, %s of data\n') %
2432 self.ui.status(_('%d files to transfer, %s of data\n') %
2430 (total_files, util.bytecount(total_bytes)))
2433 (total_files, util.bytecount(total_bytes)))
2431 handled_bytes = 0
2434 handled_bytes = 0
2432 self.ui.progress(_('clone'), 0, total=total_bytes)
2435 self.ui.progress(_('clone'), 0, total=total_bytes)
2433 start = time.time()
2436 start = time.time()
2434 for i in xrange(total_files):
2437 for i in xrange(total_files):
2435 # XXX doesn't support '\n' or '\r' in filenames
2438 # XXX doesn't support '\n' or '\r' in filenames
2436 l = fp.readline()
2439 l = fp.readline()
2437 try:
2440 try:
2438 name, size = l.split('\0', 1)
2441 name, size = l.split('\0', 1)
2439 size = int(size)
2442 size = int(size)
2440 except (ValueError, TypeError):
2443 except (ValueError, TypeError):
2441 raise error.ResponseError(
2444 raise error.ResponseError(
2442 _('unexpected response from remote server:'), l)
2445 _('unexpected response from remote server:'), l)
2443 if self.ui.debugflag:
2446 if self.ui.debugflag:
2444 self.ui.debug('adding %s (%s)\n' %
2447 self.ui.debug('adding %s (%s)\n' %
2445 (name, util.bytecount(size)))
2448 (name, util.bytecount(size)))
2446 # for backwards compat, name was partially encoded
2449 # for backwards compat, name was partially encoded
2447 ofp = self.sopener(store.decodedir(name), 'w')
2450 ofp = self.sopener(store.decodedir(name), 'w')
2448 for chunk in util.filechunkiter(fp, limit=size):
2451 for chunk in util.filechunkiter(fp, limit=size):
2449 handled_bytes += len(chunk)
2452 handled_bytes += len(chunk)
2450 self.ui.progress(_('clone'), handled_bytes,
2453 self.ui.progress(_('clone'), handled_bytes,
2451 total=total_bytes)
2454 total=total_bytes)
2452 ofp.write(chunk)
2455 ofp.write(chunk)
2453 ofp.close()
2456 ofp.close()
2454 elapsed = time.time() - start
2457 elapsed = time.time() - start
2455 if elapsed <= 0:
2458 if elapsed <= 0:
2456 elapsed = 0.001
2459 elapsed = 0.001
2457 self.ui.progress(_('clone'), None)
2460 self.ui.progress(_('clone'), None)
2458 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2461 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2459 (util.bytecount(total_bytes), elapsed,
2462 (util.bytecount(total_bytes), elapsed,
2460 util.bytecount(total_bytes / elapsed)))
2463 util.bytecount(total_bytes / elapsed)))
2461
2464
2462 # new requirements = old non-format requirements +
2465 # new requirements = old non-format requirements +
2463 # new format-related
2466 # new format-related
2464 # requirements from the streamed-in repository
2467 # requirements from the streamed-in repository
2465 requirements.update(set(self.requirements) - self.supportedformats)
2468 requirements.update(set(self.requirements) - self.supportedformats)
2466 self._applyrequirements(requirements)
2469 self._applyrequirements(requirements)
2467 self._writerequirements()
2470 self._writerequirements()
2468
2471
2469 if rbranchmap:
2472 if rbranchmap:
2470 rbheads = []
2473 rbheads = []
2471 for bheads in rbranchmap.itervalues():
2474 for bheads in rbranchmap.itervalues():
2472 rbheads.extend(bheads)
2475 rbheads.extend(bheads)
2473
2476
2474 if rbheads:
2477 if rbheads:
2475 rtiprev = max((int(self.changelog.rev(node))
2478 rtiprev = max((int(self.changelog.rev(node))
2476 for node in rbheads))
2479 for node in rbheads))
2477 cache = branchmap.branchcache(rbranchmap,
2480 cache = branchmap.branchcache(rbranchmap,
2478 self[rtiprev].node(),
2481 self[rtiprev].node(),
2479 rtiprev)
2482 rtiprev)
2480 # Try to stick it as low as possible
2483 # Try to stick it as low as possible
2481 # filter above served are unlikely to be fetch from a clone
2484 # filter above served are unlikely to be fetch from a clone
2482 for candidate in ('base', 'immutable', 'served'):
2485 for candidate in ('base', 'immutable', 'served'):
2483 rview = self.filtered(candidate)
2486 rview = self.filtered(candidate)
2484 if cache.validfor(rview):
2487 if cache.validfor(rview):
2485 self._branchcaches[candidate] = cache
2488 self._branchcaches[candidate] = cache
2486 cache.write(rview)
2489 cache.write(rview)
2487 break
2490 break
2488 self.invalidate()
2491 self.invalidate()
2489 return len(self.heads()) + 1
2492 return len(self.heads()) + 1
2490 finally:
2493 finally:
2491 lock.release()
2494 lock.release()
2492
2495
2493 def clone(self, remote, heads=[], stream=False):
2496 def clone(self, remote, heads=[], stream=False):
2494 '''clone remote repository.
2497 '''clone remote repository.
2495
2498
2496 keyword arguments:
2499 keyword arguments:
2497 heads: list of revs to clone (forces use of pull)
2500 heads: list of revs to clone (forces use of pull)
2498 stream: use streaming clone if possible'''
2501 stream: use streaming clone if possible'''
2499
2502
2500 # now, all clients that can request uncompressed clones can
2503 # now, all clients that can request uncompressed clones can
2501 # read repo formats supported by all servers that can serve
2504 # read repo formats supported by all servers that can serve
2502 # them.
2505 # them.
2503
2506
2504 # if revlog format changes, client will have to check version
2507 # if revlog format changes, client will have to check version
2505 # and format flags on "stream" capability, and use
2508 # and format flags on "stream" capability, and use
2506 # uncompressed only if compatible.
2509 # uncompressed only if compatible.
2507
2510
2508 if not stream:
2511 if not stream:
2509 # if the server explicitly prefers to stream (for fast LANs)
2512 # if the server explicitly prefers to stream (for fast LANs)
2510 stream = remote.capable('stream-preferred')
2513 stream = remote.capable('stream-preferred')
2511
2514
2512 if stream and not heads:
2515 if stream and not heads:
2513 # 'stream' means remote revlog format is revlogv1 only
2516 # 'stream' means remote revlog format is revlogv1 only
2514 if remote.capable('stream'):
2517 if remote.capable('stream'):
2515 return self.stream_in(remote, set(('revlogv1',)))
2518 return self.stream_in(remote, set(('revlogv1',)))
2516 # otherwise, 'streamreqs' contains the remote revlog format
2519 # otherwise, 'streamreqs' contains the remote revlog format
2517 streamreqs = remote.capable('streamreqs')
2520 streamreqs = remote.capable('streamreqs')
2518 if streamreqs:
2521 if streamreqs:
2519 streamreqs = set(streamreqs.split(','))
2522 streamreqs = set(streamreqs.split(','))
2520 # if we support it, stream in and adjust our requirements
2523 # if we support it, stream in and adjust our requirements
2521 if not streamreqs - self.supportedformats:
2524 if not streamreqs - self.supportedformats:
2522 return self.stream_in(remote, streamreqs)
2525 return self.stream_in(remote, streamreqs)
2523 return self.pull(remote, heads)
2526 return self.pull(remote, heads)
2524
2527
2525 def pushkey(self, namespace, key, old, new):
2528 def pushkey(self, namespace, key, old, new):
2526 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2529 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2527 old=old, new=new)
2530 old=old, new=new)
2528 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2531 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2529 ret = pushkey.push(self, namespace, key, old, new)
2532 ret = pushkey.push(self, namespace, key, old, new)
2530 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2533 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2531 ret=ret)
2534 ret=ret)
2532 return ret
2535 return ret
2533
2536
2534 def listkeys(self, namespace):
2537 def listkeys(self, namespace):
2535 self.hook('prelistkeys', throw=True, namespace=namespace)
2538 self.hook('prelistkeys', throw=True, namespace=namespace)
2536 self.ui.debug('listing keys for "%s"\n' % namespace)
2539 self.ui.debug('listing keys for "%s"\n' % namespace)
2537 values = pushkey.list(self, namespace)
2540 values = pushkey.list(self, namespace)
2538 self.hook('listkeys', namespace=namespace, values=values)
2541 self.hook('listkeys', namespace=namespace, values=values)
2539 return values
2542 return values
2540
2543
2541 def debugwireargs(self, one, two, three=None, four=None, five=None):
2544 def debugwireargs(self, one, two, three=None, four=None, five=None):
2542 '''used to test argument passing over the wire'''
2545 '''used to test argument passing over the wire'''
2543 return "%s %s %s %s %s" % (one, two, three, four, five)
2546 return "%s %s %s %s %s" % (one, two, three, four, five)
2544
2547
2545 def savecommitmessage(self, text):
2548 def savecommitmessage(self, text):
2546 fp = self.opener('last-message.txt', 'wb')
2549 fp = self.opener('last-message.txt', 'wb')
2547 try:
2550 try:
2548 fp.write(text)
2551 fp.write(text)
2549 finally:
2552 finally:
2550 fp.close()
2553 fp.close()
2551 return self.pathto(fp.name[len(self.root) + 1:])
2554 return self.pathto(fp.name[len(self.root) + 1:])
2552
2555
2553 # used to avoid circular references so destructors work
2556 # used to avoid circular references so destructors work
2554 def aftertrans(files):
2557 def aftertrans(files):
2555 renamefiles = [tuple(t) for t in files]
2558 renamefiles = [tuple(t) for t in files]
2556 def a():
2559 def a():
2557 for src, dest in renamefiles:
2560 for src, dest in renamefiles:
2558 try:
2561 try:
2559 util.rename(src, dest)
2562 util.rename(src, dest)
2560 except OSError: # journal file does not yet exist
2563 except OSError: # journal file does not yet exist
2561 pass
2564 pass
2562 return a
2565 return a
2563
2566
2564 def undoname(fn):
2567 def undoname(fn):
2565 base, name = os.path.split(fn)
2568 base, name = os.path.split(fn)
2566 assert name.startswith('journal')
2569 assert name.startswith('journal')
2567 return os.path.join(base, name.replace('journal', 'undo', 1))
2570 return os.path.join(base, name.replace('journal', 'undo', 1))
2568
2571
2569 def instance(ui, path, create):
2572 def instance(ui, path, create):
2570 return localrepository(ui, util.urllocalpath(path), create)
2573 return localrepository(ui, util.urllocalpath(path), create)
2571
2574
2572 def islocal(path):
2575 def islocal(path):
2573 return True
2576 return True
@@ -1,53 +1,92 b''
1 $ hg init test
1 $ hg init test
2 $ cd test
2 $ cd test
3
3
4 $ cat > .hg/hgrc <<EOF
4 $ cat > .hg/hgrc <<EOF
5 > [server]
5 > [server]
6 > validate=1
6 > validate=1
7 > EOF
7 > EOF
8
8
9 $ echo alpha > alpha
9 $ echo alpha > alpha
10 $ echo beta > beta
10 $ echo beta > beta
11 $ hg addr
11 $ hg addr
12 adding alpha
12 adding alpha
13 adding beta
13 adding beta
14 $ hg ci -m 1
14 $ hg ci -m 1
15
15
16 $ cd ..
16 $ cd ..
17 $ hg clone test test-clone
17 $ hg clone test test-clone
18 updating to branch default
18 updating to branch default
19 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
19 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
20
20
21 Test spurious filelog entries:
22
21 $ cd test-clone
23 $ cd test-clone
24 $ echo blah >> beta
25 $ cp .hg/store/data/beta.i tmp1
26 $ hg ci -m 2
27 $ cp .hg/store/data/beta.i tmp2
28 $ hg -q rollback
29 $ mv tmp2 .hg/store/data/beta.i
30 $ echo blah >> beta
31 $ hg ci -m '2 (corrupt)'
32
33 Expected to fail:
34
35 $ hg verify
36 checking changesets
37 checking manifests
38 crosschecking files in changesets and manifests
39 checking files
40 beta@1: dddc47b3ba30 not in manifests
41 2 files, 2 changesets, 4 total revisions
42 1 integrity errors encountered!
43 (first damaged changeset appears to be 1)
44 [1]
45
46 $ hg push
47 pushing to $TESTTMP/test
48 searching for changes
49 adding changesets
50 adding manifests
51 adding file changes
52 transaction abort!
53 rollback completed
54 abort: received spurious file revlog entry
55 [255]
56
57 $ hg -q rollback
58 $ mv tmp1 .hg/store/data/beta.i
59 $ echo beta > beta
60
61 Test missing filelog entries:
62
22 $ cp .hg/store/data/beta.i tmp
63 $ cp .hg/store/data/beta.i tmp
23 $ echo blah >> beta
64 $ echo blah >> beta
24 $ hg ci -m '2 (corrupt)'
65 $ hg ci -m '2 (corrupt)'
25 $ mv tmp .hg/store/data/beta.i
66 $ mv tmp .hg/store/data/beta.i
26
67
27 Expected to fail:
68 Expected to fail:
28
69
29 $ hg verify
70 $ hg verify
30 checking changesets
71 checking changesets
31 checking manifests
72 checking manifests
32 crosschecking files in changesets and manifests
73 crosschecking files in changesets and manifests
33 checking files
74 checking files
34 beta@1: dddc47b3ba30 in manifests not found
75 beta@1: dddc47b3ba30 in manifests not found
35 2 files, 2 changesets, 2 total revisions
76 2 files, 2 changesets, 2 total revisions
36 1 integrity errors encountered!
77 1 integrity errors encountered!
37 (first damaged changeset appears to be 1)
78 (first damaged changeset appears to be 1)
38 [1]
79 [1]
39
80
40 Expected to fail:
41
42 $ hg push
81 $ hg push
43 pushing to $TESTTMP/test (glob)
82 pushing to $TESTTMP/test (glob)
44 searching for changes
83 searching for changes
45 adding changesets
84 adding changesets
46 adding manifests
85 adding manifests
47 adding file changes
86 adding file changes
48 transaction abort!
87 transaction abort!
49 rollback completed
88 rollback completed
50 abort: missing file data for beta:dddc47b3ba30e54484720ce0f4f768a0f4b6efb9 - run hg verify
89 abort: missing file data for beta:dddc47b3ba30e54484720ce0f4f768a0f4b6efb9 - run hg verify
51 [255]
90 [255]
52
91
53 $ cd ..
92 $ cd ..
General Comments 0
You need to be logged in to leave comments. Login now