##// END OF EJS Templates
branchmap: update cache of 'unserved' filter on new changesets...
Pierre-Yves David -
r18394:50104481 default
parent child Browse files
Show More
@@ -1,2578 +1,2578 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
9 import peer, changegroup, subrepo, discovery, pushkey, obsolete, repoview
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
10 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import lock, transaction, store, encoding, base85
11 import lock, transaction, store, encoding, base85
12 import scmutil, util, extensions, hook, error, revset
12 import scmutil, util, extensions, hook, error, revset
13 import match as matchmod
13 import match as matchmod
14 import merge as mergemod
14 import merge as mergemod
15 import tags as tagsmod
15 import tags as tagsmod
16 from lock import release
16 from lock import release
17 import weakref, errno, os, time, inspect
17 import weakref, errno, os, time, inspect
18 import branchmap
18 import branchmap
19 propertycache = util.propertycache
19 propertycache = util.propertycache
20 filecache = scmutil.filecache
20 filecache = scmutil.filecache
21
21
22 class repofilecache(filecache):
22 class repofilecache(filecache):
23 """All filecache usage on repo are done for logic that should be unfiltered
23 """All filecache usage on repo are done for logic that should be unfiltered
24 """
24 """
25
25
26 def __get__(self, repo, type=None):
26 def __get__(self, repo, type=None):
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
27 return super(repofilecache, self).__get__(repo.unfiltered(), type)
28 def __set__(self, repo, value):
28 def __set__(self, repo, value):
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
29 return super(repofilecache, self).__set__(repo.unfiltered(), value)
30 def __delete__(self, repo):
30 def __delete__(self, repo):
31 return super(repofilecache, self).__delete__(repo.unfiltered())
31 return super(repofilecache, self).__delete__(repo.unfiltered())
32
32
33 class storecache(repofilecache):
33 class storecache(repofilecache):
34 """filecache for files in the store"""
34 """filecache for files in the store"""
35 def join(self, obj, fname):
35 def join(self, obj, fname):
36 return obj.sjoin(fname)
36 return obj.sjoin(fname)
37
37
38 class unfilteredpropertycache(propertycache):
38 class unfilteredpropertycache(propertycache):
39 """propertycache that apply to unfiltered repo only"""
39 """propertycache that apply to unfiltered repo only"""
40
40
41 def __get__(self, repo, type=None):
41 def __get__(self, repo, type=None):
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
42 return super(unfilteredpropertycache, self).__get__(repo.unfiltered())
43
43
44 class filteredpropertycache(propertycache):
44 class filteredpropertycache(propertycache):
45 """propertycache that must take filtering in account"""
45 """propertycache that must take filtering in account"""
46
46
47 def cachevalue(self, obj, value):
47 def cachevalue(self, obj, value):
48 object.__setattr__(obj, self.name, value)
48 object.__setattr__(obj, self.name, value)
49
49
50
50
51 def hasunfilteredcache(repo, name):
51 def hasunfilteredcache(repo, name):
52 """check if an repo and a unfilteredproperty cached value for <name>"""
52 """check if an repo and a unfilteredproperty cached value for <name>"""
53 return name in vars(repo.unfiltered())
53 return name in vars(repo.unfiltered())
54
54
55 def unfilteredmethod(orig):
55 def unfilteredmethod(orig):
56 """decorate method that always need to be run on unfiltered version"""
56 """decorate method that always need to be run on unfiltered version"""
57 def wrapper(repo, *args, **kwargs):
57 def wrapper(repo, *args, **kwargs):
58 return orig(repo.unfiltered(), *args, **kwargs)
58 return orig(repo.unfiltered(), *args, **kwargs)
59 return wrapper
59 return wrapper
60
60
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
61 MODERNCAPS = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle'))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
62 LEGACYCAPS = MODERNCAPS.union(set(['changegroupsubset']))
63
63
64 class localpeer(peer.peerrepository):
64 class localpeer(peer.peerrepository):
65 '''peer for a local repo; reflects only the most recent API'''
65 '''peer for a local repo; reflects only the most recent API'''
66
66
67 def __init__(self, repo, caps=MODERNCAPS):
67 def __init__(self, repo, caps=MODERNCAPS):
68 peer.peerrepository.__init__(self)
68 peer.peerrepository.__init__(self)
69 self._repo = repo.filtered('served')
69 self._repo = repo.filtered('served')
70 self.ui = repo.ui
70 self.ui = repo.ui
71 self._caps = repo._restrictcapabilities(caps)
71 self._caps = repo._restrictcapabilities(caps)
72 self.requirements = repo.requirements
72 self.requirements = repo.requirements
73 self.supportedformats = repo.supportedformats
73 self.supportedformats = repo.supportedformats
74
74
75 def close(self):
75 def close(self):
76 self._repo.close()
76 self._repo.close()
77
77
78 def _capabilities(self):
78 def _capabilities(self):
79 return self._caps
79 return self._caps
80
80
81 def local(self):
81 def local(self):
82 return self._repo
82 return self._repo
83
83
84 def canpush(self):
84 def canpush(self):
85 return True
85 return True
86
86
87 def url(self):
87 def url(self):
88 return self._repo.url()
88 return self._repo.url()
89
89
90 def lookup(self, key):
90 def lookup(self, key):
91 return self._repo.lookup(key)
91 return self._repo.lookup(key)
92
92
93 def branchmap(self):
93 def branchmap(self):
94 return self._repo.branchmap()
94 return self._repo.branchmap()
95
95
96 def heads(self):
96 def heads(self):
97 return self._repo.heads()
97 return self._repo.heads()
98
98
99 def known(self, nodes):
99 def known(self, nodes):
100 return self._repo.known(nodes)
100 return self._repo.known(nodes)
101
101
102 def getbundle(self, source, heads=None, common=None):
102 def getbundle(self, source, heads=None, common=None):
103 return self._repo.getbundle(source, heads=heads, common=common)
103 return self._repo.getbundle(source, heads=heads, common=common)
104
104
105 # TODO We might want to move the next two calls into legacypeer and add
105 # TODO We might want to move the next two calls into legacypeer and add
106 # unbundle instead.
106 # unbundle instead.
107
107
108 def lock(self):
108 def lock(self):
109 return self._repo.lock()
109 return self._repo.lock()
110
110
111 def addchangegroup(self, cg, source, url):
111 def addchangegroup(self, cg, source, url):
112 return self._repo.addchangegroup(cg, source, url)
112 return self._repo.addchangegroup(cg, source, url)
113
113
114 def pushkey(self, namespace, key, old, new):
114 def pushkey(self, namespace, key, old, new):
115 return self._repo.pushkey(namespace, key, old, new)
115 return self._repo.pushkey(namespace, key, old, new)
116
116
117 def listkeys(self, namespace):
117 def listkeys(self, namespace):
118 return self._repo.listkeys(namespace)
118 return self._repo.listkeys(namespace)
119
119
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
120 def debugwireargs(self, one, two, three=None, four=None, five=None):
121 '''used to test argument passing over the wire'''
121 '''used to test argument passing over the wire'''
122 return "%s %s %s %s %s" % (one, two, three, four, five)
122 return "%s %s %s %s %s" % (one, two, three, four, five)
123
123
124 class locallegacypeer(localpeer):
124 class locallegacypeer(localpeer):
125 '''peer extension which implements legacy methods too; used for tests with
125 '''peer extension which implements legacy methods too; used for tests with
126 restricted capabilities'''
126 restricted capabilities'''
127
127
128 def __init__(self, repo):
128 def __init__(self, repo):
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
129 localpeer.__init__(self, repo, caps=LEGACYCAPS)
130
130
131 def branches(self, nodes):
131 def branches(self, nodes):
132 return self._repo.branches(nodes)
132 return self._repo.branches(nodes)
133
133
134 def between(self, pairs):
134 def between(self, pairs):
135 return self._repo.between(pairs)
135 return self._repo.between(pairs)
136
136
137 def changegroup(self, basenodes, source):
137 def changegroup(self, basenodes, source):
138 return self._repo.changegroup(basenodes, source)
138 return self._repo.changegroup(basenodes, source)
139
139
140 def changegroupsubset(self, bases, heads, source):
140 def changegroupsubset(self, bases, heads, source):
141 return self._repo.changegroupsubset(bases, heads, source)
141 return self._repo.changegroupsubset(bases, heads, source)
142
142
143 class localrepository(object):
143 class localrepository(object):
144
144
145 supportedformats = set(('revlogv1', 'generaldelta'))
145 supportedformats = set(('revlogv1', 'generaldelta'))
146 supported = supportedformats | set(('store', 'fncache', 'shared',
146 supported = supportedformats | set(('store', 'fncache', 'shared',
147 'dotencode'))
147 'dotencode'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
148 openerreqs = set(('revlogv1', 'generaldelta'))
149 requirements = ['revlogv1']
149 requirements = ['revlogv1']
150 filtername = None
150 filtername = None
151
151
152 def _baserequirements(self, create):
152 def _baserequirements(self, create):
153 return self.requirements[:]
153 return self.requirements[:]
154
154
155 def __init__(self, baseui, path=None, create=False):
155 def __init__(self, baseui, path=None, create=False):
156 self.wvfs = scmutil.vfs(path, expand=True)
156 self.wvfs = scmutil.vfs(path, expand=True)
157 self.wopener = self.wvfs
157 self.wopener = self.wvfs
158 self.root = self.wvfs.base
158 self.root = self.wvfs.base
159 self.path = self.wvfs.join(".hg")
159 self.path = self.wvfs.join(".hg")
160 self.origroot = path
160 self.origroot = path
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
161 self.auditor = scmutil.pathauditor(self.root, self._checknested)
162 self.vfs = scmutil.vfs(self.path)
162 self.vfs = scmutil.vfs(self.path)
163 self.opener = self.vfs
163 self.opener = self.vfs
164 self.baseui = baseui
164 self.baseui = baseui
165 self.ui = baseui.copy()
165 self.ui = baseui.copy()
166 # A list of callback to shape the phase if no data were found.
166 # A list of callback to shape the phase if no data were found.
167 # Callback are in the form: func(repo, roots) --> processed root.
167 # Callback are in the form: func(repo, roots) --> processed root.
168 # This list it to be filled by extension during repo setup
168 # This list it to be filled by extension during repo setup
169 self._phasedefaults = []
169 self._phasedefaults = []
170 try:
170 try:
171 self.ui.readconfig(self.join("hgrc"), self.root)
171 self.ui.readconfig(self.join("hgrc"), self.root)
172 extensions.loadall(self.ui)
172 extensions.loadall(self.ui)
173 except IOError:
173 except IOError:
174 pass
174 pass
175
175
176 if not self.vfs.isdir():
176 if not self.vfs.isdir():
177 if create:
177 if create:
178 if not self.wvfs.exists():
178 if not self.wvfs.exists():
179 self.wvfs.makedirs()
179 self.wvfs.makedirs()
180 self.vfs.makedir(notindexed=True)
180 self.vfs.makedir(notindexed=True)
181 requirements = self._baserequirements(create)
181 requirements = self._baserequirements(create)
182 if self.ui.configbool('format', 'usestore', True):
182 if self.ui.configbool('format', 'usestore', True):
183 self.vfs.mkdir("store")
183 self.vfs.mkdir("store")
184 requirements.append("store")
184 requirements.append("store")
185 if self.ui.configbool('format', 'usefncache', True):
185 if self.ui.configbool('format', 'usefncache', True):
186 requirements.append("fncache")
186 requirements.append("fncache")
187 if self.ui.configbool('format', 'dotencode', True):
187 if self.ui.configbool('format', 'dotencode', True):
188 requirements.append('dotencode')
188 requirements.append('dotencode')
189 # create an invalid changelog
189 # create an invalid changelog
190 self.vfs.append(
190 self.vfs.append(
191 "00changelog.i",
191 "00changelog.i",
192 '\0\0\0\2' # represents revlogv2
192 '\0\0\0\2' # represents revlogv2
193 ' dummy changelog to prevent using the old repo layout'
193 ' dummy changelog to prevent using the old repo layout'
194 )
194 )
195 if self.ui.configbool('format', 'generaldelta', False):
195 if self.ui.configbool('format', 'generaldelta', False):
196 requirements.append("generaldelta")
196 requirements.append("generaldelta")
197 requirements = set(requirements)
197 requirements = set(requirements)
198 else:
198 else:
199 raise error.RepoError(_("repository %s not found") % path)
199 raise error.RepoError(_("repository %s not found") % path)
200 elif create:
200 elif create:
201 raise error.RepoError(_("repository %s already exists") % path)
201 raise error.RepoError(_("repository %s already exists") % path)
202 else:
202 else:
203 try:
203 try:
204 requirements = scmutil.readrequires(self.vfs, self.supported)
204 requirements = scmutil.readrequires(self.vfs, self.supported)
205 except IOError, inst:
205 except IOError, inst:
206 if inst.errno != errno.ENOENT:
206 if inst.errno != errno.ENOENT:
207 raise
207 raise
208 requirements = set()
208 requirements = set()
209
209
210 self.sharedpath = self.path
210 self.sharedpath = self.path
211 try:
211 try:
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
212 s = os.path.realpath(self.opener.read("sharedpath").rstrip('\n'))
213 if not os.path.exists(s):
213 if not os.path.exists(s):
214 raise error.RepoError(
214 raise error.RepoError(
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
215 _('.hg/sharedpath points to nonexistent directory %s') % s)
216 self.sharedpath = s
216 self.sharedpath = s
217 except IOError, inst:
217 except IOError, inst:
218 if inst.errno != errno.ENOENT:
218 if inst.errno != errno.ENOENT:
219 raise
219 raise
220
220
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
221 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
222 self.spath = self.store.path
222 self.spath = self.store.path
223 self.svfs = self.store.vfs
223 self.svfs = self.store.vfs
224 self.sopener = self.svfs
224 self.sopener = self.svfs
225 self.sjoin = self.store.join
225 self.sjoin = self.store.join
226 self.vfs.createmode = self.store.createmode
226 self.vfs.createmode = self.store.createmode
227 self._applyrequirements(requirements)
227 self._applyrequirements(requirements)
228 if create:
228 if create:
229 self._writerequirements()
229 self._writerequirements()
230
230
231
231
232 self._branchcaches = {}
232 self._branchcaches = {}
233 self.filterpats = {}
233 self.filterpats = {}
234 self._datafilters = {}
234 self._datafilters = {}
235 self._transref = self._lockref = self._wlockref = None
235 self._transref = self._lockref = self._wlockref = None
236
236
237 # A cache for various files under .hg/ that tracks file changes,
237 # A cache for various files under .hg/ that tracks file changes,
238 # (used by the filecache decorator)
238 # (used by the filecache decorator)
239 #
239 #
240 # Maps a property name to its util.filecacheentry
240 # Maps a property name to its util.filecacheentry
241 self._filecache = {}
241 self._filecache = {}
242
242
243 # hold sets of revision to be filtered
243 # hold sets of revision to be filtered
244 # should be cleared when something might have changed the filter value:
244 # should be cleared when something might have changed the filter value:
245 # - new changesets,
245 # - new changesets,
246 # - phase change,
246 # - phase change,
247 # - new obsolescence marker,
247 # - new obsolescence marker,
248 # - working directory parent change,
248 # - working directory parent change,
249 # - bookmark changes
249 # - bookmark changes
250 self.filteredrevcache = {}
250 self.filteredrevcache = {}
251
251
252 def close(self):
252 def close(self):
253 pass
253 pass
254
254
255 def _restrictcapabilities(self, caps):
255 def _restrictcapabilities(self, caps):
256 return caps
256 return caps
257
257
258 def _applyrequirements(self, requirements):
258 def _applyrequirements(self, requirements):
259 self.requirements = requirements
259 self.requirements = requirements
260 self.sopener.options = dict((r, 1) for r in requirements
260 self.sopener.options = dict((r, 1) for r in requirements
261 if r in self.openerreqs)
261 if r in self.openerreqs)
262
262
263 def _writerequirements(self):
263 def _writerequirements(self):
264 reqfile = self.opener("requires", "w")
264 reqfile = self.opener("requires", "w")
265 for r in sorted(self.requirements):
265 for r in sorted(self.requirements):
266 reqfile.write("%s\n" % r)
266 reqfile.write("%s\n" % r)
267 reqfile.close()
267 reqfile.close()
268
268
269 def _checknested(self, path):
269 def _checknested(self, path):
270 """Determine if path is a legal nested repository."""
270 """Determine if path is a legal nested repository."""
271 if not path.startswith(self.root):
271 if not path.startswith(self.root):
272 return False
272 return False
273 subpath = path[len(self.root) + 1:]
273 subpath = path[len(self.root) + 1:]
274 normsubpath = util.pconvert(subpath)
274 normsubpath = util.pconvert(subpath)
275
275
276 # XXX: Checking against the current working copy is wrong in
276 # XXX: Checking against the current working copy is wrong in
277 # the sense that it can reject things like
277 # the sense that it can reject things like
278 #
278 #
279 # $ hg cat -r 10 sub/x.txt
279 # $ hg cat -r 10 sub/x.txt
280 #
280 #
281 # if sub/ is no longer a subrepository in the working copy
281 # if sub/ is no longer a subrepository in the working copy
282 # parent revision.
282 # parent revision.
283 #
283 #
284 # However, it can of course also allow things that would have
284 # However, it can of course also allow things that would have
285 # been rejected before, such as the above cat command if sub/
285 # been rejected before, such as the above cat command if sub/
286 # is a subrepository now, but was a normal directory before.
286 # is a subrepository now, but was a normal directory before.
287 # The old path auditor would have rejected by mistake since it
287 # The old path auditor would have rejected by mistake since it
288 # panics when it sees sub/.hg/.
288 # panics when it sees sub/.hg/.
289 #
289 #
290 # All in all, checking against the working copy seems sensible
290 # All in all, checking against the working copy seems sensible
291 # since we want to prevent access to nested repositories on
291 # since we want to prevent access to nested repositories on
292 # the filesystem *now*.
292 # the filesystem *now*.
293 ctx = self[None]
293 ctx = self[None]
294 parts = util.splitpath(subpath)
294 parts = util.splitpath(subpath)
295 while parts:
295 while parts:
296 prefix = '/'.join(parts)
296 prefix = '/'.join(parts)
297 if prefix in ctx.substate:
297 if prefix in ctx.substate:
298 if prefix == normsubpath:
298 if prefix == normsubpath:
299 return True
299 return True
300 else:
300 else:
301 sub = ctx.sub(prefix)
301 sub = ctx.sub(prefix)
302 return sub.checknested(subpath[len(prefix) + 1:])
302 return sub.checknested(subpath[len(prefix) + 1:])
303 else:
303 else:
304 parts.pop()
304 parts.pop()
305 return False
305 return False
306
306
307 def peer(self):
307 def peer(self):
308 return localpeer(self) # not cached to avoid reference cycle
308 return localpeer(self) # not cached to avoid reference cycle
309
309
310 def unfiltered(self):
310 def unfiltered(self):
311 """Return unfiltered version of the repository
311 """Return unfiltered version of the repository
312
312
313 Intended to be ovewritten by filtered repo."""
313 Intended to be ovewritten by filtered repo."""
314 return self
314 return self
315
315
316 def filtered(self, name):
316 def filtered(self, name):
317 """Return a filtered version of a repository"""
317 """Return a filtered version of a repository"""
318 # build a new class with the mixin and the current class
318 # build a new class with the mixin and the current class
319 # (possibily subclass of the repo)
319 # (possibily subclass of the repo)
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
320 class proxycls(repoview.repoview, self.unfiltered().__class__):
321 pass
321 pass
322 return proxycls(self, name)
322 return proxycls(self, name)
323
323
324 @repofilecache('bookmarks')
324 @repofilecache('bookmarks')
325 def _bookmarks(self):
325 def _bookmarks(self):
326 return bookmarks.bmstore(self)
326 return bookmarks.bmstore(self)
327
327
328 @repofilecache('bookmarks.current')
328 @repofilecache('bookmarks.current')
329 def _bookmarkcurrent(self):
329 def _bookmarkcurrent(self):
330 return bookmarks.readcurrent(self)
330 return bookmarks.readcurrent(self)
331
331
332 def bookmarkheads(self, bookmark):
332 def bookmarkheads(self, bookmark):
333 name = bookmark.split('@', 1)[0]
333 name = bookmark.split('@', 1)[0]
334 heads = []
334 heads = []
335 for mark, n in self._bookmarks.iteritems():
335 for mark, n in self._bookmarks.iteritems():
336 if mark.split('@', 1)[0] == name:
336 if mark.split('@', 1)[0] == name:
337 heads.append(n)
337 heads.append(n)
338 return heads
338 return heads
339
339
340 @storecache('phaseroots')
340 @storecache('phaseroots')
341 def _phasecache(self):
341 def _phasecache(self):
342 return phases.phasecache(self, self._phasedefaults)
342 return phases.phasecache(self, self._phasedefaults)
343
343
344 @storecache('obsstore')
344 @storecache('obsstore')
345 def obsstore(self):
345 def obsstore(self):
346 store = obsolete.obsstore(self.sopener)
346 store = obsolete.obsstore(self.sopener)
347 if store and not obsolete._enabled:
347 if store and not obsolete._enabled:
348 # message is rare enough to not be translated
348 # message is rare enough to not be translated
349 msg = 'obsolete feature not enabled but %i markers found!\n'
349 msg = 'obsolete feature not enabled but %i markers found!\n'
350 self.ui.warn(msg % len(list(store)))
350 self.ui.warn(msg % len(list(store)))
351 return store
351 return store
352
352
353 @storecache('00changelog.i')
353 @storecache('00changelog.i')
354 def changelog(self):
354 def changelog(self):
355 c = changelog.changelog(self.sopener)
355 c = changelog.changelog(self.sopener)
356 if 'HG_PENDING' in os.environ:
356 if 'HG_PENDING' in os.environ:
357 p = os.environ['HG_PENDING']
357 p = os.environ['HG_PENDING']
358 if p.startswith(self.root):
358 if p.startswith(self.root):
359 c.readpending('00changelog.i.a')
359 c.readpending('00changelog.i.a')
360 return c
360 return c
361
361
362 @storecache('00manifest.i')
362 @storecache('00manifest.i')
363 def manifest(self):
363 def manifest(self):
364 return manifest.manifest(self.sopener)
364 return manifest.manifest(self.sopener)
365
365
366 @repofilecache('dirstate')
366 @repofilecache('dirstate')
367 def dirstate(self):
367 def dirstate(self):
368 warned = [0]
368 warned = [0]
369 def validate(node):
369 def validate(node):
370 try:
370 try:
371 self.changelog.rev(node)
371 self.changelog.rev(node)
372 return node
372 return node
373 except error.LookupError:
373 except error.LookupError:
374 if not warned[0]:
374 if not warned[0]:
375 warned[0] = True
375 warned[0] = True
376 self.ui.warn(_("warning: ignoring unknown"
376 self.ui.warn(_("warning: ignoring unknown"
377 " working parent %s!\n") % short(node))
377 " working parent %s!\n") % short(node))
378 return nullid
378 return nullid
379
379
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
380 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
381
381
382 def __getitem__(self, changeid):
382 def __getitem__(self, changeid):
383 if changeid is None:
383 if changeid is None:
384 return context.workingctx(self)
384 return context.workingctx(self)
385 return context.changectx(self, changeid)
385 return context.changectx(self, changeid)
386
386
387 def __contains__(self, changeid):
387 def __contains__(self, changeid):
388 try:
388 try:
389 return bool(self.lookup(changeid))
389 return bool(self.lookup(changeid))
390 except error.RepoLookupError:
390 except error.RepoLookupError:
391 return False
391 return False
392
392
393 def __nonzero__(self):
393 def __nonzero__(self):
394 return True
394 return True
395
395
396 def __len__(self):
396 def __len__(self):
397 return len(self.changelog)
397 return len(self.changelog)
398
398
399 def __iter__(self):
399 def __iter__(self):
400 return iter(self.changelog)
400 return iter(self.changelog)
401
401
402 def revs(self, expr, *args):
402 def revs(self, expr, *args):
403 '''Return a list of revisions matching the given revset'''
403 '''Return a list of revisions matching the given revset'''
404 expr = revset.formatspec(expr, *args)
404 expr = revset.formatspec(expr, *args)
405 m = revset.match(None, expr)
405 m = revset.match(None, expr)
406 return [r for r in m(self, list(self))]
406 return [r for r in m(self, list(self))]
407
407
408 def set(self, expr, *args):
408 def set(self, expr, *args):
409 '''
409 '''
410 Yield a context for each matching revision, after doing arg
410 Yield a context for each matching revision, after doing arg
411 replacement via revset.formatspec
411 replacement via revset.formatspec
412 '''
412 '''
413 for r in self.revs(expr, *args):
413 for r in self.revs(expr, *args):
414 yield self[r]
414 yield self[r]
415
415
416 def url(self):
416 def url(self):
417 return 'file:' + self.root
417 return 'file:' + self.root
418
418
419 def hook(self, name, throw=False, **args):
419 def hook(self, name, throw=False, **args):
420 return hook.hook(self.ui, self, name, throw, **args)
420 return hook.hook(self.ui, self, name, throw, **args)
421
421
422 @unfilteredmethod
422 @unfilteredmethod
423 def _tag(self, names, node, message, local, user, date, extra={}):
423 def _tag(self, names, node, message, local, user, date, extra={}):
424 if isinstance(names, str):
424 if isinstance(names, str):
425 names = (names,)
425 names = (names,)
426
426
427 branches = self.branchmap()
427 branches = self.branchmap()
428 for name in names:
428 for name in names:
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
429 self.hook('pretag', throw=True, node=hex(node), tag=name,
430 local=local)
430 local=local)
431 if name in branches:
431 if name in branches:
432 self.ui.warn(_("warning: tag %s conflicts with existing"
432 self.ui.warn(_("warning: tag %s conflicts with existing"
433 " branch name\n") % name)
433 " branch name\n") % name)
434
434
435 def writetags(fp, names, munge, prevtags):
435 def writetags(fp, names, munge, prevtags):
436 fp.seek(0, 2)
436 fp.seek(0, 2)
437 if prevtags and prevtags[-1] != '\n':
437 if prevtags and prevtags[-1] != '\n':
438 fp.write('\n')
438 fp.write('\n')
439 for name in names:
439 for name in names:
440 m = munge and munge(name) or name
440 m = munge and munge(name) or name
441 if (self._tagscache.tagtypes and
441 if (self._tagscache.tagtypes and
442 name in self._tagscache.tagtypes):
442 name in self._tagscache.tagtypes):
443 old = self.tags().get(name, nullid)
443 old = self.tags().get(name, nullid)
444 fp.write('%s %s\n' % (hex(old), m))
444 fp.write('%s %s\n' % (hex(old), m))
445 fp.write('%s %s\n' % (hex(node), m))
445 fp.write('%s %s\n' % (hex(node), m))
446 fp.close()
446 fp.close()
447
447
448 prevtags = ''
448 prevtags = ''
449 if local:
449 if local:
450 try:
450 try:
451 fp = self.opener('localtags', 'r+')
451 fp = self.opener('localtags', 'r+')
452 except IOError:
452 except IOError:
453 fp = self.opener('localtags', 'a')
453 fp = self.opener('localtags', 'a')
454 else:
454 else:
455 prevtags = fp.read()
455 prevtags = fp.read()
456
456
457 # local tags are stored in the current charset
457 # local tags are stored in the current charset
458 writetags(fp, names, None, prevtags)
458 writetags(fp, names, None, prevtags)
459 for name in names:
459 for name in names:
460 self.hook('tag', node=hex(node), tag=name, local=local)
460 self.hook('tag', node=hex(node), tag=name, local=local)
461 return
461 return
462
462
463 try:
463 try:
464 fp = self.wfile('.hgtags', 'rb+')
464 fp = self.wfile('.hgtags', 'rb+')
465 except IOError, e:
465 except IOError, e:
466 if e.errno != errno.ENOENT:
466 if e.errno != errno.ENOENT:
467 raise
467 raise
468 fp = self.wfile('.hgtags', 'ab')
468 fp = self.wfile('.hgtags', 'ab')
469 else:
469 else:
470 prevtags = fp.read()
470 prevtags = fp.read()
471
471
472 # committed tags are stored in UTF-8
472 # committed tags are stored in UTF-8
473 writetags(fp, names, encoding.fromlocal, prevtags)
473 writetags(fp, names, encoding.fromlocal, prevtags)
474
474
475 fp.close()
475 fp.close()
476
476
477 self.invalidatecaches()
477 self.invalidatecaches()
478
478
479 if '.hgtags' not in self.dirstate:
479 if '.hgtags' not in self.dirstate:
480 self[None].add(['.hgtags'])
480 self[None].add(['.hgtags'])
481
481
482 m = matchmod.exact(self.root, '', ['.hgtags'])
482 m = matchmod.exact(self.root, '', ['.hgtags'])
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
483 tagnode = self.commit(message, user, date, extra=extra, match=m)
484
484
485 for name in names:
485 for name in names:
486 self.hook('tag', node=hex(node), tag=name, local=local)
486 self.hook('tag', node=hex(node), tag=name, local=local)
487
487
488 return tagnode
488 return tagnode
489
489
490 def tag(self, names, node, message, local, user, date):
490 def tag(self, names, node, message, local, user, date):
491 '''tag a revision with one or more symbolic names.
491 '''tag a revision with one or more symbolic names.
492
492
493 names is a list of strings or, when adding a single tag, names may be a
493 names is a list of strings or, when adding a single tag, names may be a
494 string.
494 string.
495
495
496 if local is True, the tags are stored in a per-repository file.
496 if local is True, the tags are stored in a per-repository file.
497 otherwise, they are stored in the .hgtags file, and a new
497 otherwise, they are stored in the .hgtags file, and a new
498 changeset is committed with the change.
498 changeset is committed with the change.
499
499
500 keyword arguments:
500 keyword arguments:
501
501
502 local: whether to store tags in non-version-controlled file
502 local: whether to store tags in non-version-controlled file
503 (default False)
503 (default False)
504
504
505 message: commit message to use if committing
505 message: commit message to use if committing
506
506
507 user: name of user to use if committing
507 user: name of user to use if committing
508
508
509 date: date tuple to use if committing'''
509 date: date tuple to use if committing'''
510
510
511 if not local:
511 if not local:
512 for x in self.status()[:5]:
512 for x in self.status()[:5]:
513 if '.hgtags' in x:
513 if '.hgtags' in x:
514 raise util.Abort(_('working copy of .hgtags is changed '
514 raise util.Abort(_('working copy of .hgtags is changed '
515 '(please commit .hgtags manually)'))
515 '(please commit .hgtags manually)'))
516
516
517 self.tags() # instantiate the cache
517 self.tags() # instantiate the cache
518 self._tag(names, node, message, local, user, date)
518 self._tag(names, node, message, local, user, date)
519
519
520 @filteredpropertycache
520 @filteredpropertycache
521 def _tagscache(self):
521 def _tagscache(self):
522 '''Returns a tagscache object that contains various tags related
522 '''Returns a tagscache object that contains various tags related
523 caches.'''
523 caches.'''
524
524
525 # This simplifies its cache management by having one decorated
525 # This simplifies its cache management by having one decorated
526 # function (this one) and the rest simply fetch things from it.
526 # function (this one) and the rest simply fetch things from it.
527 class tagscache(object):
527 class tagscache(object):
528 def __init__(self):
528 def __init__(self):
529 # These two define the set of tags for this repository. tags
529 # These two define the set of tags for this repository. tags
530 # maps tag name to node; tagtypes maps tag name to 'global' or
530 # maps tag name to node; tagtypes maps tag name to 'global' or
531 # 'local'. (Global tags are defined by .hgtags across all
531 # 'local'. (Global tags are defined by .hgtags across all
532 # heads, and local tags are defined in .hg/localtags.)
532 # heads, and local tags are defined in .hg/localtags.)
533 # They constitute the in-memory cache of tags.
533 # They constitute the in-memory cache of tags.
534 self.tags = self.tagtypes = None
534 self.tags = self.tagtypes = None
535
535
536 self.nodetagscache = self.tagslist = None
536 self.nodetagscache = self.tagslist = None
537
537
538 cache = tagscache()
538 cache = tagscache()
539 cache.tags, cache.tagtypes = self._findtags()
539 cache.tags, cache.tagtypes = self._findtags()
540
540
541 return cache
541 return cache
542
542
543 def tags(self):
543 def tags(self):
544 '''return a mapping of tag to node'''
544 '''return a mapping of tag to node'''
545 t = {}
545 t = {}
546 if self.changelog.filteredrevs:
546 if self.changelog.filteredrevs:
547 tags, tt = self._findtags()
547 tags, tt = self._findtags()
548 else:
548 else:
549 tags = self._tagscache.tags
549 tags = self._tagscache.tags
550 for k, v in tags.iteritems():
550 for k, v in tags.iteritems():
551 try:
551 try:
552 # ignore tags to unknown nodes
552 # ignore tags to unknown nodes
553 self.changelog.rev(v)
553 self.changelog.rev(v)
554 t[k] = v
554 t[k] = v
555 except (error.LookupError, ValueError):
555 except (error.LookupError, ValueError):
556 pass
556 pass
557 return t
557 return t
558
558
559 def _findtags(self):
559 def _findtags(self):
560 '''Do the hard work of finding tags. Return a pair of dicts
560 '''Do the hard work of finding tags. Return a pair of dicts
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
561 (tags, tagtypes) where tags maps tag name to node, and tagtypes
562 maps tag name to a string like \'global\' or \'local\'.
562 maps tag name to a string like \'global\' or \'local\'.
563 Subclasses or extensions are free to add their own tags, but
563 Subclasses or extensions are free to add their own tags, but
564 should be aware that the returned dicts will be retained for the
564 should be aware that the returned dicts will be retained for the
565 duration of the localrepo object.'''
565 duration of the localrepo object.'''
566
566
567 # XXX what tagtype should subclasses/extensions use? Currently
567 # XXX what tagtype should subclasses/extensions use? Currently
568 # mq and bookmarks add tags, but do not set the tagtype at all.
568 # mq and bookmarks add tags, but do not set the tagtype at all.
569 # Should each extension invent its own tag type? Should there
569 # Should each extension invent its own tag type? Should there
570 # be one tagtype for all such "virtual" tags? Or is the status
570 # be one tagtype for all such "virtual" tags? Or is the status
571 # quo fine?
571 # quo fine?
572
572
573 alltags = {} # map tag name to (node, hist)
573 alltags = {} # map tag name to (node, hist)
574 tagtypes = {}
574 tagtypes = {}
575
575
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
576 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
577 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
578
578
579 # Build the return dicts. Have to re-encode tag names because
579 # Build the return dicts. Have to re-encode tag names because
580 # the tags module always uses UTF-8 (in order not to lose info
580 # the tags module always uses UTF-8 (in order not to lose info
581 # writing to the cache), but the rest of Mercurial wants them in
581 # writing to the cache), but the rest of Mercurial wants them in
582 # local encoding.
582 # local encoding.
583 tags = {}
583 tags = {}
584 for (name, (node, hist)) in alltags.iteritems():
584 for (name, (node, hist)) in alltags.iteritems():
585 if node != nullid:
585 if node != nullid:
586 tags[encoding.tolocal(name)] = node
586 tags[encoding.tolocal(name)] = node
587 tags['tip'] = self.changelog.tip()
587 tags['tip'] = self.changelog.tip()
588 tagtypes = dict([(encoding.tolocal(name), value)
588 tagtypes = dict([(encoding.tolocal(name), value)
589 for (name, value) in tagtypes.iteritems()])
589 for (name, value) in tagtypes.iteritems()])
590 return (tags, tagtypes)
590 return (tags, tagtypes)
591
591
592 def tagtype(self, tagname):
592 def tagtype(self, tagname):
593 '''
593 '''
594 return the type of the given tag. result can be:
594 return the type of the given tag. result can be:
595
595
596 'local' : a local tag
596 'local' : a local tag
597 'global' : a global tag
597 'global' : a global tag
598 None : tag does not exist
598 None : tag does not exist
599 '''
599 '''
600
600
601 return self._tagscache.tagtypes.get(tagname)
601 return self._tagscache.tagtypes.get(tagname)
602
602
603 def tagslist(self):
603 def tagslist(self):
604 '''return a list of tags ordered by revision'''
604 '''return a list of tags ordered by revision'''
605 if not self._tagscache.tagslist:
605 if not self._tagscache.tagslist:
606 l = []
606 l = []
607 for t, n in self.tags().iteritems():
607 for t, n in self.tags().iteritems():
608 r = self.changelog.rev(n)
608 r = self.changelog.rev(n)
609 l.append((r, t, n))
609 l.append((r, t, n))
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
610 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
611
611
612 return self._tagscache.tagslist
612 return self._tagscache.tagslist
613
613
614 def nodetags(self, node):
614 def nodetags(self, node):
615 '''return the tags associated with a node'''
615 '''return the tags associated with a node'''
616 if not self._tagscache.nodetagscache:
616 if not self._tagscache.nodetagscache:
617 nodetagscache = {}
617 nodetagscache = {}
618 for t, n in self._tagscache.tags.iteritems():
618 for t, n in self._tagscache.tags.iteritems():
619 nodetagscache.setdefault(n, []).append(t)
619 nodetagscache.setdefault(n, []).append(t)
620 for tags in nodetagscache.itervalues():
620 for tags in nodetagscache.itervalues():
621 tags.sort()
621 tags.sort()
622 self._tagscache.nodetagscache = nodetagscache
622 self._tagscache.nodetagscache = nodetagscache
623 return self._tagscache.nodetagscache.get(node, [])
623 return self._tagscache.nodetagscache.get(node, [])
624
624
625 def nodebookmarks(self, node):
625 def nodebookmarks(self, node):
626 marks = []
626 marks = []
627 for bookmark, n in self._bookmarks.iteritems():
627 for bookmark, n in self._bookmarks.iteritems():
628 if n == node:
628 if n == node:
629 marks.append(bookmark)
629 marks.append(bookmark)
630 return sorted(marks)
630 return sorted(marks)
631
631
632 def branchmap(self):
632 def branchmap(self):
633 '''returns a dictionary {branch: [branchheads]}'''
633 '''returns a dictionary {branch: [branchheads]}'''
634 branchmap.updatecache(self)
634 branchmap.updatecache(self)
635 return self._branchcaches[self.filtername]
635 return self._branchcaches[self.filtername]
636
636
637
637
638 def _branchtip(self, heads):
638 def _branchtip(self, heads):
639 '''return the tipmost branch head in heads'''
639 '''return the tipmost branch head in heads'''
640 tip = heads[-1]
640 tip = heads[-1]
641 for h in reversed(heads):
641 for h in reversed(heads):
642 if not self[h].closesbranch():
642 if not self[h].closesbranch():
643 tip = h
643 tip = h
644 break
644 break
645 return tip
645 return tip
646
646
647 def branchtip(self, branch):
647 def branchtip(self, branch):
648 '''return the tip node for a given branch'''
648 '''return the tip node for a given branch'''
649 if branch not in self.branchmap():
649 if branch not in self.branchmap():
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
650 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
651 return self._branchtip(self.branchmap()[branch])
651 return self._branchtip(self.branchmap()[branch])
652
652
653 def branchtags(self):
653 def branchtags(self):
654 '''return a dict where branch names map to the tipmost head of
654 '''return a dict where branch names map to the tipmost head of
655 the branch, open heads come before closed'''
655 the branch, open heads come before closed'''
656 bt = {}
656 bt = {}
657 for bn, heads in self.branchmap().iteritems():
657 for bn, heads in self.branchmap().iteritems():
658 bt[bn] = self._branchtip(heads)
658 bt[bn] = self._branchtip(heads)
659 return bt
659 return bt
660
660
661 def lookup(self, key):
661 def lookup(self, key):
662 return self[key].node()
662 return self[key].node()
663
663
664 def lookupbranch(self, key, remote=None):
664 def lookupbranch(self, key, remote=None):
665 repo = remote or self
665 repo = remote or self
666 if key in repo.branchmap():
666 if key in repo.branchmap():
667 return key
667 return key
668
668
669 repo = (remote and remote.local()) and remote or self
669 repo = (remote and remote.local()) and remote or self
670 return repo[key].branch()
670 return repo[key].branch()
671
671
672 def known(self, nodes):
672 def known(self, nodes):
673 nm = self.changelog.nodemap
673 nm = self.changelog.nodemap
674 pc = self._phasecache
674 pc = self._phasecache
675 result = []
675 result = []
676 for n in nodes:
676 for n in nodes:
677 r = nm.get(n)
677 r = nm.get(n)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
678 resp = not (r is None or pc.phase(self, r) >= phases.secret)
679 result.append(resp)
679 result.append(resp)
680 return result
680 return result
681
681
682 def local(self):
682 def local(self):
683 return self
683 return self
684
684
685 def cancopy(self):
685 def cancopy(self):
686 return self.local() # so statichttprepo's override of local() works
686 return self.local() # so statichttprepo's override of local() works
687
687
688 def join(self, f):
688 def join(self, f):
689 return os.path.join(self.path, f)
689 return os.path.join(self.path, f)
690
690
691 def wjoin(self, f):
691 def wjoin(self, f):
692 return os.path.join(self.root, f)
692 return os.path.join(self.root, f)
693
693
694 def file(self, f):
694 def file(self, f):
695 if f[0] == '/':
695 if f[0] == '/':
696 f = f[1:]
696 f = f[1:]
697 return filelog.filelog(self.sopener, f)
697 return filelog.filelog(self.sopener, f)
698
698
699 def changectx(self, changeid):
699 def changectx(self, changeid):
700 return self[changeid]
700 return self[changeid]
701
701
702 def parents(self, changeid=None):
702 def parents(self, changeid=None):
703 '''get list of changectxs for parents of changeid'''
703 '''get list of changectxs for parents of changeid'''
704 return self[changeid].parents()
704 return self[changeid].parents()
705
705
706 def setparents(self, p1, p2=nullid):
706 def setparents(self, p1, p2=nullid):
707 copies = self.dirstate.setparents(p1, p2)
707 copies = self.dirstate.setparents(p1, p2)
708 if copies:
708 if copies:
709 # Adjust copy records, the dirstate cannot do it, it
709 # Adjust copy records, the dirstate cannot do it, it
710 # requires access to parents manifests. Preserve them
710 # requires access to parents manifests. Preserve them
711 # only for entries added to first parent.
711 # only for entries added to first parent.
712 pctx = self[p1]
712 pctx = self[p1]
713 for f in copies:
713 for f in copies:
714 if f not in pctx and copies[f] in pctx:
714 if f not in pctx and copies[f] in pctx:
715 self.dirstate.copy(copies[f], f)
715 self.dirstate.copy(copies[f], f)
716
716
717 def filectx(self, path, changeid=None, fileid=None):
717 def filectx(self, path, changeid=None, fileid=None):
718 """changeid can be a changeset revision, node, or tag.
718 """changeid can be a changeset revision, node, or tag.
719 fileid can be a file revision or node."""
719 fileid can be a file revision or node."""
720 return context.filectx(self, path, changeid, fileid)
720 return context.filectx(self, path, changeid, fileid)
721
721
722 def getcwd(self):
722 def getcwd(self):
723 return self.dirstate.getcwd()
723 return self.dirstate.getcwd()
724
724
725 def pathto(self, f, cwd=None):
725 def pathto(self, f, cwd=None):
726 return self.dirstate.pathto(f, cwd)
726 return self.dirstate.pathto(f, cwd)
727
727
728 def wfile(self, f, mode='r'):
728 def wfile(self, f, mode='r'):
729 return self.wopener(f, mode)
729 return self.wopener(f, mode)
730
730
731 def _link(self, f):
731 def _link(self, f):
732 return os.path.islink(self.wjoin(f))
732 return os.path.islink(self.wjoin(f))
733
733
734 def _loadfilter(self, filter):
734 def _loadfilter(self, filter):
735 if filter not in self.filterpats:
735 if filter not in self.filterpats:
736 l = []
736 l = []
737 for pat, cmd in self.ui.configitems(filter):
737 for pat, cmd in self.ui.configitems(filter):
738 if cmd == '!':
738 if cmd == '!':
739 continue
739 continue
740 mf = matchmod.match(self.root, '', [pat])
740 mf = matchmod.match(self.root, '', [pat])
741 fn = None
741 fn = None
742 params = cmd
742 params = cmd
743 for name, filterfn in self._datafilters.iteritems():
743 for name, filterfn in self._datafilters.iteritems():
744 if cmd.startswith(name):
744 if cmd.startswith(name):
745 fn = filterfn
745 fn = filterfn
746 params = cmd[len(name):].lstrip()
746 params = cmd[len(name):].lstrip()
747 break
747 break
748 if not fn:
748 if not fn:
749 fn = lambda s, c, **kwargs: util.filter(s, c)
749 fn = lambda s, c, **kwargs: util.filter(s, c)
750 # Wrap old filters not supporting keyword arguments
750 # Wrap old filters not supporting keyword arguments
751 if not inspect.getargspec(fn)[2]:
751 if not inspect.getargspec(fn)[2]:
752 oldfn = fn
752 oldfn = fn
753 fn = lambda s, c, **kwargs: oldfn(s, c)
753 fn = lambda s, c, **kwargs: oldfn(s, c)
754 l.append((mf, fn, params))
754 l.append((mf, fn, params))
755 self.filterpats[filter] = l
755 self.filterpats[filter] = l
756 return self.filterpats[filter]
756 return self.filterpats[filter]
757
757
758 def _filter(self, filterpats, filename, data):
758 def _filter(self, filterpats, filename, data):
759 for mf, fn, cmd in filterpats:
759 for mf, fn, cmd in filterpats:
760 if mf(filename):
760 if mf(filename):
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
761 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
762 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
763 break
763 break
764
764
765 return data
765 return data
766
766
767 @unfilteredpropertycache
767 @unfilteredpropertycache
768 def _encodefilterpats(self):
768 def _encodefilterpats(self):
769 return self._loadfilter('encode')
769 return self._loadfilter('encode')
770
770
771 @unfilteredpropertycache
771 @unfilteredpropertycache
772 def _decodefilterpats(self):
772 def _decodefilterpats(self):
773 return self._loadfilter('decode')
773 return self._loadfilter('decode')
774
774
775 def adddatafilter(self, name, filter):
775 def adddatafilter(self, name, filter):
776 self._datafilters[name] = filter
776 self._datafilters[name] = filter
777
777
778 def wread(self, filename):
778 def wread(self, filename):
779 if self._link(filename):
779 if self._link(filename):
780 data = os.readlink(self.wjoin(filename))
780 data = os.readlink(self.wjoin(filename))
781 else:
781 else:
782 data = self.wopener.read(filename)
782 data = self.wopener.read(filename)
783 return self._filter(self._encodefilterpats, filename, data)
783 return self._filter(self._encodefilterpats, filename, data)
784
784
785 def wwrite(self, filename, data, flags):
785 def wwrite(self, filename, data, flags):
786 data = self._filter(self._decodefilterpats, filename, data)
786 data = self._filter(self._decodefilterpats, filename, data)
787 if 'l' in flags:
787 if 'l' in flags:
788 self.wopener.symlink(data, filename)
788 self.wopener.symlink(data, filename)
789 else:
789 else:
790 self.wopener.write(filename, data)
790 self.wopener.write(filename, data)
791 if 'x' in flags:
791 if 'x' in flags:
792 util.setflags(self.wjoin(filename), False, True)
792 util.setflags(self.wjoin(filename), False, True)
793
793
794 def wwritedata(self, filename, data):
794 def wwritedata(self, filename, data):
795 return self._filter(self._decodefilterpats, filename, data)
795 return self._filter(self._decodefilterpats, filename, data)
796
796
797 def transaction(self, desc):
797 def transaction(self, desc):
798 tr = self._transref and self._transref() or None
798 tr = self._transref and self._transref() or None
799 if tr and tr.running():
799 if tr and tr.running():
800 return tr.nest()
800 return tr.nest()
801
801
802 # abort here if the journal already exists
802 # abort here if the journal already exists
803 if os.path.exists(self.sjoin("journal")):
803 if os.path.exists(self.sjoin("journal")):
804 raise error.RepoError(
804 raise error.RepoError(
805 _("abandoned transaction found - run hg recover"))
805 _("abandoned transaction found - run hg recover"))
806
806
807 self._writejournal(desc)
807 self._writejournal(desc)
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
808 renames = [(x, undoname(x)) for x in self._journalfiles()]
809
809
810 tr = transaction.transaction(self.ui.warn, self.sopener,
810 tr = transaction.transaction(self.ui.warn, self.sopener,
811 self.sjoin("journal"),
811 self.sjoin("journal"),
812 aftertrans(renames),
812 aftertrans(renames),
813 self.store.createmode)
813 self.store.createmode)
814 self._transref = weakref.ref(tr)
814 self._transref = weakref.ref(tr)
815 return tr
815 return tr
816
816
817 def _journalfiles(self):
817 def _journalfiles(self):
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
818 return (self.sjoin('journal'), self.join('journal.dirstate'),
819 self.join('journal.branch'), self.join('journal.desc'),
819 self.join('journal.branch'), self.join('journal.desc'),
820 self.join('journal.bookmarks'),
820 self.join('journal.bookmarks'),
821 self.sjoin('journal.phaseroots'))
821 self.sjoin('journal.phaseroots'))
822
822
823 def undofiles(self):
823 def undofiles(self):
824 return [undoname(x) for x in self._journalfiles()]
824 return [undoname(x) for x in self._journalfiles()]
825
825
826 def _writejournal(self, desc):
826 def _writejournal(self, desc):
827 self.opener.write("journal.dirstate",
827 self.opener.write("journal.dirstate",
828 self.opener.tryread("dirstate"))
828 self.opener.tryread("dirstate"))
829 self.opener.write("journal.branch",
829 self.opener.write("journal.branch",
830 encoding.fromlocal(self.dirstate.branch()))
830 encoding.fromlocal(self.dirstate.branch()))
831 self.opener.write("journal.desc",
831 self.opener.write("journal.desc",
832 "%d\n%s\n" % (len(self), desc))
832 "%d\n%s\n" % (len(self), desc))
833 self.opener.write("journal.bookmarks",
833 self.opener.write("journal.bookmarks",
834 self.opener.tryread("bookmarks"))
834 self.opener.tryread("bookmarks"))
835 self.sopener.write("journal.phaseroots",
835 self.sopener.write("journal.phaseroots",
836 self.sopener.tryread("phaseroots"))
836 self.sopener.tryread("phaseroots"))
837
837
838 def recover(self):
838 def recover(self):
839 lock = self.lock()
839 lock = self.lock()
840 try:
840 try:
841 if os.path.exists(self.sjoin("journal")):
841 if os.path.exists(self.sjoin("journal")):
842 self.ui.status(_("rolling back interrupted transaction\n"))
842 self.ui.status(_("rolling back interrupted transaction\n"))
843 transaction.rollback(self.sopener, self.sjoin("journal"),
843 transaction.rollback(self.sopener, self.sjoin("journal"),
844 self.ui.warn)
844 self.ui.warn)
845 self.invalidate()
845 self.invalidate()
846 return True
846 return True
847 else:
847 else:
848 self.ui.warn(_("no interrupted transaction available\n"))
848 self.ui.warn(_("no interrupted transaction available\n"))
849 return False
849 return False
850 finally:
850 finally:
851 lock.release()
851 lock.release()
852
852
853 def rollback(self, dryrun=False, force=False):
853 def rollback(self, dryrun=False, force=False):
854 wlock = lock = None
854 wlock = lock = None
855 try:
855 try:
856 wlock = self.wlock()
856 wlock = self.wlock()
857 lock = self.lock()
857 lock = self.lock()
858 if os.path.exists(self.sjoin("undo")):
858 if os.path.exists(self.sjoin("undo")):
859 return self._rollback(dryrun, force)
859 return self._rollback(dryrun, force)
860 else:
860 else:
861 self.ui.warn(_("no rollback information available\n"))
861 self.ui.warn(_("no rollback information available\n"))
862 return 1
862 return 1
863 finally:
863 finally:
864 release(lock, wlock)
864 release(lock, wlock)
865
865
866 @unfilteredmethod # Until we get smarter cache management
866 @unfilteredmethod # Until we get smarter cache management
867 def _rollback(self, dryrun, force):
867 def _rollback(self, dryrun, force):
868 ui = self.ui
868 ui = self.ui
869 try:
869 try:
870 args = self.opener.read('undo.desc').splitlines()
870 args = self.opener.read('undo.desc').splitlines()
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
871 (oldlen, desc, detail) = (int(args[0]), args[1], None)
872 if len(args) >= 3:
872 if len(args) >= 3:
873 detail = args[2]
873 detail = args[2]
874 oldtip = oldlen - 1
874 oldtip = oldlen - 1
875
875
876 if detail and ui.verbose:
876 if detail and ui.verbose:
877 msg = (_('repository tip rolled back to revision %s'
877 msg = (_('repository tip rolled back to revision %s'
878 ' (undo %s: %s)\n')
878 ' (undo %s: %s)\n')
879 % (oldtip, desc, detail))
879 % (oldtip, desc, detail))
880 else:
880 else:
881 msg = (_('repository tip rolled back to revision %s'
881 msg = (_('repository tip rolled back to revision %s'
882 ' (undo %s)\n')
882 ' (undo %s)\n')
883 % (oldtip, desc))
883 % (oldtip, desc))
884 except IOError:
884 except IOError:
885 msg = _('rolling back unknown transaction\n')
885 msg = _('rolling back unknown transaction\n')
886 desc = None
886 desc = None
887
887
888 if not force and self['.'] != self['tip'] and desc == 'commit':
888 if not force and self['.'] != self['tip'] and desc == 'commit':
889 raise util.Abort(
889 raise util.Abort(
890 _('rollback of last commit while not checked out '
890 _('rollback of last commit while not checked out '
891 'may lose data'), hint=_('use -f to force'))
891 'may lose data'), hint=_('use -f to force'))
892
892
893 ui.status(msg)
893 ui.status(msg)
894 if dryrun:
894 if dryrun:
895 return 0
895 return 0
896
896
897 parents = self.dirstate.parents()
897 parents = self.dirstate.parents()
898 self.destroying()
898 self.destroying()
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
899 transaction.rollback(self.sopener, self.sjoin('undo'), ui.warn)
900 if os.path.exists(self.join('undo.bookmarks')):
900 if os.path.exists(self.join('undo.bookmarks')):
901 util.rename(self.join('undo.bookmarks'),
901 util.rename(self.join('undo.bookmarks'),
902 self.join('bookmarks'))
902 self.join('bookmarks'))
903 if os.path.exists(self.sjoin('undo.phaseroots')):
903 if os.path.exists(self.sjoin('undo.phaseroots')):
904 util.rename(self.sjoin('undo.phaseroots'),
904 util.rename(self.sjoin('undo.phaseroots'),
905 self.sjoin('phaseroots'))
905 self.sjoin('phaseroots'))
906 self.invalidate()
906 self.invalidate()
907
907
908 parentgone = (parents[0] not in self.changelog.nodemap or
908 parentgone = (parents[0] not in self.changelog.nodemap or
909 parents[1] not in self.changelog.nodemap)
909 parents[1] not in self.changelog.nodemap)
910 if parentgone:
910 if parentgone:
911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
911 util.rename(self.join('undo.dirstate'), self.join('dirstate'))
912 try:
912 try:
913 branch = self.opener.read('undo.branch')
913 branch = self.opener.read('undo.branch')
914 self.dirstate.setbranch(encoding.tolocal(branch))
914 self.dirstate.setbranch(encoding.tolocal(branch))
915 except IOError:
915 except IOError:
916 ui.warn(_('named branch could not be reset: '
916 ui.warn(_('named branch could not be reset: '
917 'current branch is still \'%s\'\n')
917 'current branch is still \'%s\'\n')
918 % self.dirstate.branch())
918 % self.dirstate.branch())
919
919
920 self.dirstate.invalidate()
920 self.dirstate.invalidate()
921 parents = tuple([p.rev() for p in self.parents()])
921 parents = tuple([p.rev() for p in self.parents()])
922 if len(parents) > 1:
922 if len(parents) > 1:
923 ui.status(_('working directory now based on '
923 ui.status(_('working directory now based on '
924 'revisions %d and %d\n') % parents)
924 'revisions %d and %d\n') % parents)
925 else:
925 else:
926 ui.status(_('working directory now based on '
926 ui.status(_('working directory now based on '
927 'revision %d\n') % parents)
927 'revision %d\n') % parents)
928 # TODO: if we know which new heads may result from this rollback, pass
928 # TODO: if we know which new heads may result from this rollback, pass
929 # them to destroy(), which will prevent the branchhead cache from being
929 # them to destroy(), which will prevent the branchhead cache from being
930 # invalidated.
930 # invalidated.
931 self.destroyed()
931 self.destroyed()
932 return 0
932 return 0
933
933
934 def invalidatecaches(self):
934 def invalidatecaches(self):
935
935
936 if '_tagscache' in vars(self):
936 if '_tagscache' in vars(self):
937 # can't use delattr on proxy
937 # can't use delattr on proxy
938 del self.__dict__['_tagscache']
938 del self.__dict__['_tagscache']
939
939
940 self.unfiltered()._branchcaches.clear()
940 self.unfiltered()._branchcaches.clear()
941 self.invalidatevolatilesets()
941 self.invalidatevolatilesets()
942
942
943 def invalidatevolatilesets(self):
943 def invalidatevolatilesets(self):
944 self.filteredrevcache.clear()
944 self.filteredrevcache.clear()
945 obsolete.clearobscaches(self)
945 obsolete.clearobscaches(self)
946
946
947 def invalidatedirstate(self):
947 def invalidatedirstate(self):
948 '''Invalidates the dirstate, causing the next call to dirstate
948 '''Invalidates the dirstate, causing the next call to dirstate
949 to check if it was modified since the last time it was read,
949 to check if it was modified since the last time it was read,
950 rereading it if it has.
950 rereading it if it has.
951
951
952 This is different to dirstate.invalidate() that it doesn't always
952 This is different to dirstate.invalidate() that it doesn't always
953 rereads the dirstate. Use dirstate.invalidate() if you want to
953 rereads the dirstate. Use dirstate.invalidate() if you want to
954 explicitly read the dirstate again (i.e. restoring it to a previous
954 explicitly read the dirstate again (i.e. restoring it to a previous
955 known good state).'''
955 known good state).'''
956 if hasunfilteredcache(self, 'dirstate'):
956 if hasunfilteredcache(self, 'dirstate'):
957 for k in self.dirstate._filecache:
957 for k in self.dirstate._filecache:
958 try:
958 try:
959 delattr(self.dirstate, k)
959 delattr(self.dirstate, k)
960 except AttributeError:
960 except AttributeError:
961 pass
961 pass
962 delattr(self.unfiltered(), 'dirstate')
962 delattr(self.unfiltered(), 'dirstate')
963
963
964 def invalidate(self):
964 def invalidate(self):
965 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
965 unfiltered = self.unfiltered() # all filecaches are stored on unfiltered
966 for k in self._filecache:
966 for k in self._filecache:
967 # dirstate is invalidated separately in invalidatedirstate()
967 # dirstate is invalidated separately in invalidatedirstate()
968 if k == 'dirstate':
968 if k == 'dirstate':
969 continue
969 continue
970
970
971 try:
971 try:
972 delattr(unfiltered, k)
972 delattr(unfiltered, k)
973 except AttributeError:
973 except AttributeError:
974 pass
974 pass
975 self.invalidatecaches()
975 self.invalidatecaches()
976
976
977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
977 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
978 try:
978 try:
979 l = lock.lock(lockname, 0, releasefn, desc=desc)
979 l = lock.lock(lockname, 0, releasefn, desc=desc)
980 except error.LockHeld, inst:
980 except error.LockHeld, inst:
981 if not wait:
981 if not wait:
982 raise
982 raise
983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
983 self.ui.warn(_("waiting for lock on %s held by %r\n") %
984 (desc, inst.locker))
984 (desc, inst.locker))
985 # default to 600 seconds timeout
985 # default to 600 seconds timeout
986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
986 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
987 releasefn, desc=desc)
987 releasefn, desc=desc)
988 if acquirefn:
988 if acquirefn:
989 acquirefn()
989 acquirefn()
990 return l
990 return l
991
991
992 def _afterlock(self, callback):
992 def _afterlock(self, callback):
993 """add a callback to the current repository lock.
993 """add a callback to the current repository lock.
994
994
995 The callback will be executed on lock release."""
995 The callback will be executed on lock release."""
996 l = self._lockref and self._lockref()
996 l = self._lockref and self._lockref()
997 if l:
997 if l:
998 l.postrelease.append(callback)
998 l.postrelease.append(callback)
999 else:
999 else:
1000 callback()
1000 callback()
1001
1001
1002 def lock(self, wait=True):
1002 def lock(self, wait=True):
1003 '''Lock the repository store (.hg/store) and return a weak reference
1003 '''Lock the repository store (.hg/store) and return a weak reference
1004 to the lock. Use this before modifying the store (e.g. committing or
1004 to the lock. Use this before modifying the store (e.g. committing or
1005 stripping). If you are opening a transaction, get a lock as well.)'''
1005 stripping). If you are opening a transaction, get a lock as well.)'''
1006 l = self._lockref and self._lockref()
1006 l = self._lockref and self._lockref()
1007 if l is not None and l.held:
1007 if l is not None and l.held:
1008 l.lock()
1008 l.lock()
1009 return l
1009 return l
1010
1010
1011 def unlock():
1011 def unlock():
1012 self.store.write()
1012 self.store.write()
1013 if hasunfilteredcache(self, '_phasecache'):
1013 if hasunfilteredcache(self, '_phasecache'):
1014 self._phasecache.write()
1014 self._phasecache.write()
1015 for k, ce in self._filecache.items():
1015 for k, ce in self._filecache.items():
1016 if k == 'dirstate' or k not in self.__dict__:
1016 if k == 'dirstate' or k not in self.__dict__:
1017 continue
1017 continue
1018 ce.refresh()
1018 ce.refresh()
1019
1019
1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1020 l = self._lock(self.sjoin("lock"), wait, unlock,
1021 self.invalidate, _('repository %s') % self.origroot)
1021 self.invalidate, _('repository %s') % self.origroot)
1022 self._lockref = weakref.ref(l)
1022 self._lockref = weakref.ref(l)
1023 return l
1023 return l
1024
1024
1025 def wlock(self, wait=True):
1025 def wlock(self, wait=True):
1026 '''Lock the non-store parts of the repository (everything under
1026 '''Lock the non-store parts of the repository (everything under
1027 .hg except .hg/store) and return a weak reference to the lock.
1027 .hg except .hg/store) and return a weak reference to the lock.
1028 Use this before modifying files in .hg.'''
1028 Use this before modifying files in .hg.'''
1029 l = self._wlockref and self._wlockref()
1029 l = self._wlockref and self._wlockref()
1030 if l is not None and l.held:
1030 if l is not None and l.held:
1031 l.lock()
1031 l.lock()
1032 return l
1032 return l
1033
1033
1034 def unlock():
1034 def unlock():
1035 self.dirstate.write()
1035 self.dirstate.write()
1036 self._filecache['dirstate'].refresh()
1036 self._filecache['dirstate'].refresh()
1037
1037
1038 l = self._lock(self.join("wlock"), wait, unlock,
1038 l = self._lock(self.join("wlock"), wait, unlock,
1039 self.invalidatedirstate, _('working directory of %s') %
1039 self.invalidatedirstate, _('working directory of %s') %
1040 self.origroot)
1040 self.origroot)
1041 self._wlockref = weakref.ref(l)
1041 self._wlockref = weakref.ref(l)
1042 return l
1042 return l
1043
1043
1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1044 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1045 """
1045 """
1046 commit an individual file as part of a larger transaction
1046 commit an individual file as part of a larger transaction
1047 """
1047 """
1048
1048
1049 fname = fctx.path()
1049 fname = fctx.path()
1050 text = fctx.data()
1050 text = fctx.data()
1051 flog = self.file(fname)
1051 flog = self.file(fname)
1052 fparent1 = manifest1.get(fname, nullid)
1052 fparent1 = manifest1.get(fname, nullid)
1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1053 fparent2 = fparent2o = manifest2.get(fname, nullid)
1054
1054
1055 meta = {}
1055 meta = {}
1056 copy = fctx.renamed()
1056 copy = fctx.renamed()
1057 if copy and copy[0] != fname:
1057 if copy and copy[0] != fname:
1058 # Mark the new revision of this file as a copy of another
1058 # Mark the new revision of this file as a copy of another
1059 # file. This copy data will effectively act as a parent
1059 # file. This copy data will effectively act as a parent
1060 # of this new revision. If this is a merge, the first
1060 # of this new revision. If this is a merge, the first
1061 # parent will be the nullid (meaning "look up the copy data")
1061 # parent will be the nullid (meaning "look up the copy data")
1062 # and the second one will be the other parent. For example:
1062 # and the second one will be the other parent. For example:
1063 #
1063 #
1064 # 0 --- 1 --- 3 rev1 changes file foo
1064 # 0 --- 1 --- 3 rev1 changes file foo
1065 # \ / rev2 renames foo to bar and changes it
1065 # \ / rev2 renames foo to bar and changes it
1066 # \- 2 -/ rev3 should have bar with all changes and
1066 # \- 2 -/ rev3 should have bar with all changes and
1067 # should record that bar descends from
1067 # should record that bar descends from
1068 # bar in rev2 and foo in rev1
1068 # bar in rev2 and foo in rev1
1069 #
1069 #
1070 # this allows this merge to succeed:
1070 # this allows this merge to succeed:
1071 #
1071 #
1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1072 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1073 # \ / merging rev3 and rev4 should use bar@rev2
1073 # \ / merging rev3 and rev4 should use bar@rev2
1074 # \- 2 --- 4 as the merge base
1074 # \- 2 --- 4 as the merge base
1075 #
1075 #
1076
1076
1077 cfname = copy[0]
1077 cfname = copy[0]
1078 crev = manifest1.get(cfname)
1078 crev = manifest1.get(cfname)
1079 newfparent = fparent2
1079 newfparent = fparent2
1080
1080
1081 if manifest2: # branch merge
1081 if manifest2: # branch merge
1082 if fparent2 == nullid or crev is None: # copied on remote side
1082 if fparent2 == nullid or crev is None: # copied on remote side
1083 if cfname in manifest2:
1083 if cfname in manifest2:
1084 crev = manifest2[cfname]
1084 crev = manifest2[cfname]
1085 newfparent = fparent1
1085 newfparent = fparent1
1086
1086
1087 # find source in nearest ancestor if we've lost track
1087 # find source in nearest ancestor if we've lost track
1088 if not crev:
1088 if not crev:
1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1089 self.ui.debug(" %s: searching for copy revision for %s\n" %
1090 (fname, cfname))
1090 (fname, cfname))
1091 for ancestor in self[None].ancestors():
1091 for ancestor in self[None].ancestors():
1092 if cfname in ancestor:
1092 if cfname in ancestor:
1093 crev = ancestor[cfname].filenode()
1093 crev = ancestor[cfname].filenode()
1094 break
1094 break
1095
1095
1096 if crev:
1096 if crev:
1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1097 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1098 meta["copy"] = cfname
1098 meta["copy"] = cfname
1099 meta["copyrev"] = hex(crev)
1099 meta["copyrev"] = hex(crev)
1100 fparent1, fparent2 = nullid, newfparent
1100 fparent1, fparent2 = nullid, newfparent
1101 else:
1101 else:
1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1102 self.ui.warn(_("warning: can't find ancestor for '%s' "
1103 "copied from '%s'!\n") % (fname, cfname))
1103 "copied from '%s'!\n") % (fname, cfname))
1104
1104
1105 elif fparent2 != nullid:
1105 elif fparent2 != nullid:
1106 # is one parent an ancestor of the other?
1106 # is one parent an ancestor of the other?
1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1107 fparentancestor = flog.ancestor(fparent1, fparent2)
1108 if fparentancestor == fparent1:
1108 if fparentancestor == fparent1:
1109 fparent1, fparent2 = fparent2, nullid
1109 fparent1, fparent2 = fparent2, nullid
1110 elif fparentancestor == fparent2:
1110 elif fparentancestor == fparent2:
1111 fparent2 = nullid
1111 fparent2 = nullid
1112
1112
1113 # is the file changed?
1113 # is the file changed?
1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1114 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1115 changelist.append(fname)
1115 changelist.append(fname)
1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1116 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1117
1117
1118 # are just the flags changed during merge?
1118 # are just the flags changed during merge?
1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1119 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1120 changelist.append(fname)
1120 changelist.append(fname)
1121
1121
1122 return fparent1
1122 return fparent1
1123
1123
1124 @unfilteredmethod
1124 @unfilteredmethod
1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1125 def commit(self, text="", user=None, date=None, match=None, force=False,
1126 editor=False, extra={}):
1126 editor=False, extra={}):
1127 """Add a new revision to current repository.
1127 """Add a new revision to current repository.
1128
1128
1129 Revision information is gathered from the working directory,
1129 Revision information is gathered from the working directory,
1130 match can be used to filter the committed files. If editor is
1130 match can be used to filter the committed files. If editor is
1131 supplied, it is called to get a commit message.
1131 supplied, it is called to get a commit message.
1132 """
1132 """
1133
1133
1134 def fail(f, msg):
1134 def fail(f, msg):
1135 raise util.Abort('%s: %s' % (f, msg))
1135 raise util.Abort('%s: %s' % (f, msg))
1136
1136
1137 if not match:
1137 if not match:
1138 match = matchmod.always(self.root, '')
1138 match = matchmod.always(self.root, '')
1139
1139
1140 if not force:
1140 if not force:
1141 vdirs = []
1141 vdirs = []
1142 match.dir = vdirs.append
1142 match.dir = vdirs.append
1143 match.bad = fail
1143 match.bad = fail
1144
1144
1145 wlock = self.wlock()
1145 wlock = self.wlock()
1146 try:
1146 try:
1147 wctx = self[None]
1147 wctx = self[None]
1148 merge = len(wctx.parents()) > 1
1148 merge = len(wctx.parents()) > 1
1149
1149
1150 if (not force and merge and match and
1150 if (not force and merge and match and
1151 (match.files() or match.anypats())):
1151 (match.files() or match.anypats())):
1152 raise util.Abort(_('cannot partially commit a merge '
1152 raise util.Abort(_('cannot partially commit a merge '
1153 '(do not specify files or patterns)'))
1153 '(do not specify files or patterns)'))
1154
1154
1155 changes = self.status(match=match, clean=force)
1155 changes = self.status(match=match, clean=force)
1156 if force:
1156 if force:
1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1157 changes[0].extend(changes[6]) # mq may commit unchanged files
1158
1158
1159 # check subrepos
1159 # check subrepos
1160 subs = []
1160 subs = []
1161 commitsubs = set()
1161 commitsubs = set()
1162 newstate = wctx.substate.copy()
1162 newstate = wctx.substate.copy()
1163 # only manage subrepos and .hgsubstate if .hgsub is present
1163 # only manage subrepos and .hgsubstate if .hgsub is present
1164 if '.hgsub' in wctx:
1164 if '.hgsub' in wctx:
1165 # we'll decide whether to track this ourselves, thanks
1165 # we'll decide whether to track this ourselves, thanks
1166 if '.hgsubstate' in changes[0]:
1166 if '.hgsubstate' in changes[0]:
1167 changes[0].remove('.hgsubstate')
1167 changes[0].remove('.hgsubstate')
1168 if '.hgsubstate' in changes[2]:
1168 if '.hgsubstate' in changes[2]:
1169 changes[2].remove('.hgsubstate')
1169 changes[2].remove('.hgsubstate')
1170
1170
1171 # compare current state to last committed state
1171 # compare current state to last committed state
1172 # build new substate based on last committed state
1172 # build new substate based on last committed state
1173 oldstate = wctx.p1().substate
1173 oldstate = wctx.p1().substate
1174 for s in sorted(newstate.keys()):
1174 for s in sorted(newstate.keys()):
1175 if not match(s):
1175 if not match(s):
1176 # ignore working copy, use old state if present
1176 # ignore working copy, use old state if present
1177 if s in oldstate:
1177 if s in oldstate:
1178 newstate[s] = oldstate[s]
1178 newstate[s] = oldstate[s]
1179 continue
1179 continue
1180 if not force:
1180 if not force:
1181 raise util.Abort(
1181 raise util.Abort(
1182 _("commit with new subrepo %s excluded") % s)
1182 _("commit with new subrepo %s excluded") % s)
1183 if wctx.sub(s).dirty(True):
1183 if wctx.sub(s).dirty(True):
1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1184 if not self.ui.configbool('ui', 'commitsubrepos'):
1185 raise util.Abort(
1185 raise util.Abort(
1186 _("uncommitted changes in subrepo %s") % s,
1186 _("uncommitted changes in subrepo %s") % s,
1187 hint=_("use --subrepos for recursive commit"))
1187 hint=_("use --subrepos for recursive commit"))
1188 subs.append(s)
1188 subs.append(s)
1189 commitsubs.add(s)
1189 commitsubs.add(s)
1190 else:
1190 else:
1191 bs = wctx.sub(s).basestate()
1191 bs = wctx.sub(s).basestate()
1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1192 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1193 if oldstate.get(s, (None, None, None))[1] != bs:
1193 if oldstate.get(s, (None, None, None))[1] != bs:
1194 subs.append(s)
1194 subs.append(s)
1195
1195
1196 # check for removed subrepos
1196 # check for removed subrepos
1197 for p in wctx.parents():
1197 for p in wctx.parents():
1198 r = [s for s in p.substate if s not in newstate]
1198 r = [s for s in p.substate if s not in newstate]
1199 subs += [s for s in r if match(s)]
1199 subs += [s for s in r if match(s)]
1200 if subs:
1200 if subs:
1201 if (not match('.hgsub') and
1201 if (not match('.hgsub') and
1202 '.hgsub' in (wctx.modified() + wctx.added())):
1202 '.hgsub' in (wctx.modified() + wctx.added())):
1203 raise util.Abort(
1203 raise util.Abort(
1204 _("can't commit subrepos without .hgsub"))
1204 _("can't commit subrepos without .hgsub"))
1205 changes[0].insert(0, '.hgsubstate')
1205 changes[0].insert(0, '.hgsubstate')
1206
1206
1207 elif '.hgsub' in changes[2]:
1207 elif '.hgsub' in changes[2]:
1208 # clean up .hgsubstate when .hgsub is removed
1208 # clean up .hgsubstate when .hgsub is removed
1209 if ('.hgsubstate' in wctx and
1209 if ('.hgsubstate' in wctx and
1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1210 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1211 changes[2].insert(0, '.hgsubstate')
1211 changes[2].insert(0, '.hgsubstate')
1212
1212
1213 # make sure all explicit patterns are matched
1213 # make sure all explicit patterns are matched
1214 if not force and match.files():
1214 if not force and match.files():
1215 matched = set(changes[0] + changes[1] + changes[2])
1215 matched = set(changes[0] + changes[1] + changes[2])
1216
1216
1217 for f in match.files():
1217 for f in match.files():
1218 f = self.dirstate.normalize(f)
1218 f = self.dirstate.normalize(f)
1219 if f == '.' or f in matched or f in wctx.substate:
1219 if f == '.' or f in matched or f in wctx.substate:
1220 continue
1220 continue
1221 if f in changes[3]: # missing
1221 if f in changes[3]: # missing
1222 fail(f, _('file not found!'))
1222 fail(f, _('file not found!'))
1223 if f in vdirs: # visited directory
1223 if f in vdirs: # visited directory
1224 d = f + '/'
1224 d = f + '/'
1225 for mf in matched:
1225 for mf in matched:
1226 if mf.startswith(d):
1226 if mf.startswith(d):
1227 break
1227 break
1228 else:
1228 else:
1229 fail(f, _("no match under directory!"))
1229 fail(f, _("no match under directory!"))
1230 elif f not in self.dirstate:
1230 elif f not in self.dirstate:
1231 fail(f, _("file not tracked!"))
1231 fail(f, _("file not tracked!"))
1232
1232
1233 if (not force and not extra.get("close") and not merge
1233 if (not force and not extra.get("close") and not merge
1234 and not (changes[0] or changes[1] or changes[2])
1234 and not (changes[0] or changes[1] or changes[2])
1235 and wctx.branch() == wctx.p1().branch()):
1235 and wctx.branch() == wctx.p1().branch()):
1236 return None
1236 return None
1237
1237
1238 if merge and changes[3]:
1238 if merge and changes[3]:
1239 raise util.Abort(_("cannot commit merge with missing files"))
1239 raise util.Abort(_("cannot commit merge with missing files"))
1240
1240
1241 ms = mergemod.mergestate(self)
1241 ms = mergemod.mergestate(self)
1242 for f in changes[0]:
1242 for f in changes[0]:
1243 if f in ms and ms[f] == 'u':
1243 if f in ms and ms[f] == 'u':
1244 raise util.Abort(_("unresolved merge conflicts "
1244 raise util.Abort(_("unresolved merge conflicts "
1245 "(see hg help resolve)"))
1245 "(see hg help resolve)"))
1246
1246
1247 cctx = context.workingctx(self, text, user, date, extra, changes)
1247 cctx = context.workingctx(self, text, user, date, extra, changes)
1248 if editor:
1248 if editor:
1249 cctx._text = editor(self, cctx, subs)
1249 cctx._text = editor(self, cctx, subs)
1250 edited = (text != cctx._text)
1250 edited = (text != cctx._text)
1251
1251
1252 # commit subs and write new state
1252 # commit subs and write new state
1253 if subs:
1253 if subs:
1254 for s in sorted(commitsubs):
1254 for s in sorted(commitsubs):
1255 sub = wctx.sub(s)
1255 sub = wctx.sub(s)
1256 self.ui.status(_('committing subrepository %s\n') %
1256 self.ui.status(_('committing subrepository %s\n') %
1257 subrepo.subrelpath(sub))
1257 subrepo.subrelpath(sub))
1258 sr = sub.commit(cctx._text, user, date)
1258 sr = sub.commit(cctx._text, user, date)
1259 newstate[s] = (newstate[s][0], sr)
1259 newstate[s] = (newstate[s][0], sr)
1260 subrepo.writestate(self, newstate)
1260 subrepo.writestate(self, newstate)
1261
1261
1262 # Save commit message in case this transaction gets rolled back
1262 # Save commit message in case this transaction gets rolled back
1263 # (e.g. by a pretxncommit hook). Leave the content alone on
1263 # (e.g. by a pretxncommit hook). Leave the content alone on
1264 # the assumption that the user will use the same editor again.
1264 # the assumption that the user will use the same editor again.
1265 msgfn = self.savecommitmessage(cctx._text)
1265 msgfn = self.savecommitmessage(cctx._text)
1266
1266
1267 p1, p2 = self.dirstate.parents()
1267 p1, p2 = self.dirstate.parents()
1268 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1268 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1269 try:
1269 try:
1270 self.hook("precommit", throw=True, parent1=hookp1,
1270 self.hook("precommit", throw=True, parent1=hookp1,
1271 parent2=hookp2)
1271 parent2=hookp2)
1272 ret = self.commitctx(cctx, True)
1272 ret = self.commitctx(cctx, True)
1273 except: # re-raises
1273 except: # re-raises
1274 if edited:
1274 if edited:
1275 self.ui.write(
1275 self.ui.write(
1276 _('note: commit message saved in %s\n') % msgfn)
1276 _('note: commit message saved in %s\n') % msgfn)
1277 raise
1277 raise
1278
1278
1279 # update bookmarks, dirstate and mergestate
1279 # update bookmarks, dirstate and mergestate
1280 bookmarks.update(self, [p1, p2], ret)
1280 bookmarks.update(self, [p1, p2], ret)
1281 for f in changes[0] + changes[1]:
1281 for f in changes[0] + changes[1]:
1282 self.dirstate.normal(f)
1282 self.dirstate.normal(f)
1283 for f in changes[2]:
1283 for f in changes[2]:
1284 self.dirstate.drop(f)
1284 self.dirstate.drop(f)
1285 self.dirstate.setparents(ret)
1285 self.dirstate.setparents(ret)
1286 ms.reset()
1286 ms.reset()
1287 finally:
1287 finally:
1288 wlock.release()
1288 wlock.release()
1289
1289
1290 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1290 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1291 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1291 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1292 self._afterlock(commithook)
1292 self._afterlock(commithook)
1293 return ret
1293 return ret
1294
1294
1295 @unfilteredmethod
1295 @unfilteredmethod
1296 def commitctx(self, ctx, error=False):
1296 def commitctx(self, ctx, error=False):
1297 """Add a new revision to current repository.
1297 """Add a new revision to current repository.
1298 Revision information is passed via the context argument.
1298 Revision information is passed via the context argument.
1299 """
1299 """
1300
1300
1301 tr = lock = None
1301 tr = lock = None
1302 removed = list(ctx.removed())
1302 removed = list(ctx.removed())
1303 p1, p2 = ctx.p1(), ctx.p2()
1303 p1, p2 = ctx.p1(), ctx.p2()
1304 user = ctx.user()
1304 user = ctx.user()
1305
1305
1306 lock = self.lock()
1306 lock = self.lock()
1307 try:
1307 try:
1308 tr = self.transaction("commit")
1308 tr = self.transaction("commit")
1309 trp = weakref.proxy(tr)
1309 trp = weakref.proxy(tr)
1310
1310
1311 if ctx.files():
1311 if ctx.files():
1312 m1 = p1.manifest().copy()
1312 m1 = p1.manifest().copy()
1313 m2 = p2.manifest()
1313 m2 = p2.manifest()
1314
1314
1315 # check in files
1315 # check in files
1316 new = {}
1316 new = {}
1317 changed = []
1317 changed = []
1318 linkrev = len(self)
1318 linkrev = len(self)
1319 for f in sorted(ctx.modified() + ctx.added()):
1319 for f in sorted(ctx.modified() + ctx.added()):
1320 self.ui.note(f + "\n")
1320 self.ui.note(f + "\n")
1321 try:
1321 try:
1322 fctx = ctx[f]
1322 fctx = ctx[f]
1323 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1323 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1324 changed)
1324 changed)
1325 m1.set(f, fctx.flags())
1325 m1.set(f, fctx.flags())
1326 except OSError, inst:
1326 except OSError, inst:
1327 self.ui.warn(_("trouble committing %s!\n") % f)
1327 self.ui.warn(_("trouble committing %s!\n") % f)
1328 raise
1328 raise
1329 except IOError, inst:
1329 except IOError, inst:
1330 errcode = getattr(inst, 'errno', errno.ENOENT)
1330 errcode = getattr(inst, 'errno', errno.ENOENT)
1331 if error or errcode and errcode != errno.ENOENT:
1331 if error or errcode and errcode != errno.ENOENT:
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1332 self.ui.warn(_("trouble committing %s!\n") % f)
1333 raise
1333 raise
1334 else:
1334 else:
1335 removed.append(f)
1335 removed.append(f)
1336
1336
1337 # update manifest
1337 # update manifest
1338 m1.update(new)
1338 m1.update(new)
1339 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1339 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1340 drop = [f for f in removed if f in m1]
1340 drop = [f for f in removed if f in m1]
1341 for f in drop:
1341 for f in drop:
1342 del m1[f]
1342 del m1[f]
1343 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1343 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1344 p2.manifestnode(), (new, drop))
1344 p2.manifestnode(), (new, drop))
1345 files = changed + removed
1345 files = changed + removed
1346 else:
1346 else:
1347 mn = p1.manifestnode()
1347 mn = p1.manifestnode()
1348 files = []
1348 files = []
1349
1349
1350 # update changelog
1350 # update changelog
1351 self.changelog.delayupdate()
1351 self.changelog.delayupdate()
1352 n = self.changelog.add(mn, files, ctx.description(),
1352 n = self.changelog.add(mn, files, ctx.description(),
1353 trp, p1.node(), p2.node(),
1353 trp, p1.node(), p2.node(),
1354 user, ctx.date(), ctx.extra().copy())
1354 user, ctx.date(), ctx.extra().copy())
1355 p = lambda: self.changelog.writepending() and self.root or ""
1355 p = lambda: self.changelog.writepending() and self.root or ""
1356 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1356 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1357 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1357 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1358 parent2=xp2, pending=p)
1358 parent2=xp2, pending=p)
1359 self.changelog.finalize(trp)
1359 self.changelog.finalize(trp)
1360 # set the new commit is proper phase
1360 # set the new commit is proper phase
1361 targetphase = phases.newcommitphase(self.ui)
1361 targetphase = phases.newcommitphase(self.ui)
1362 if targetphase:
1362 if targetphase:
1363 # retract boundary do not alter parent changeset.
1363 # retract boundary do not alter parent changeset.
1364 # if a parent have higher the resulting phase will
1364 # if a parent have higher the resulting phase will
1365 # be compliant anyway
1365 # be compliant anyway
1366 #
1366 #
1367 # if minimal phase was 0 we don't need to retract anything
1367 # if minimal phase was 0 we don't need to retract anything
1368 phases.retractboundary(self, targetphase, [n])
1368 phases.retractboundary(self, targetphase, [n])
1369 tr.close()
1369 tr.close()
1370 branchmap.updatecache(self)
1370 branchmap.updatecache(self.filtered('served'))
1371 return n
1371 return n
1372 finally:
1372 finally:
1373 if tr:
1373 if tr:
1374 tr.release()
1374 tr.release()
1375 lock.release()
1375 lock.release()
1376
1376
1377 @unfilteredmethod
1377 @unfilteredmethod
1378 def destroying(self):
1378 def destroying(self):
1379 '''Inform the repository that nodes are about to be destroyed.
1379 '''Inform the repository that nodes are about to be destroyed.
1380 Intended for use by strip and rollback, so there's a common
1380 Intended for use by strip and rollback, so there's a common
1381 place for anything that has to be done before destroying history.
1381 place for anything that has to be done before destroying history.
1382
1382
1383 This is mostly useful for saving state that is in memory and waiting
1383 This is mostly useful for saving state that is in memory and waiting
1384 to be flushed when the current lock is released. Because a call to
1384 to be flushed when the current lock is released. Because a call to
1385 destroyed is imminent, the repo will be invalidated causing those
1385 destroyed is imminent, the repo will be invalidated causing those
1386 changes to stay in memory (waiting for the next unlock), or vanish
1386 changes to stay in memory (waiting for the next unlock), or vanish
1387 completely.
1387 completely.
1388 '''
1388 '''
1389 # It simplifies the logic around updating the branchheads cache if we
1389 # It simplifies the logic around updating the branchheads cache if we
1390 # only have to consider the effect of the stripped revisions and not
1390 # only have to consider the effect of the stripped revisions and not
1391 # revisions missing because the cache is out-of-date.
1391 # revisions missing because the cache is out-of-date.
1392 branchmap.updatecache(self)
1392 branchmap.updatecache(self)
1393
1393
1394 # When using the same lock to commit and strip, the phasecache is left
1394 # When using the same lock to commit and strip, the phasecache is left
1395 # dirty after committing. Then when we strip, the repo is invalidated,
1395 # dirty after committing. Then when we strip, the repo is invalidated,
1396 # causing those changes to disappear.
1396 # causing those changes to disappear.
1397 if '_phasecache' in vars(self):
1397 if '_phasecache' in vars(self):
1398 self._phasecache.write()
1398 self._phasecache.write()
1399
1399
1400 @unfilteredmethod
1400 @unfilteredmethod
1401 def destroyed(self, newheadnodes=None):
1401 def destroyed(self, newheadnodes=None):
1402 '''Inform the repository that nodes have been destroyed.
1402 '''Inform the repository that nodes have been destroyed.
1403 Intended for use by strip and rollback, so there's a common
1403 Intended for use by strip and rollback, so there's a common
1404 place for anything that has to be done after destroying history.
1404 place for anything that has to be done after destroying history.
1405
1405
1406 If you know the branchheadcache was uptodate before nodes were removed
1406 If you know the branchheadcache was uptodate before nodes were removed
1407 and you also know the set of candidate new heads that may have resulted
1407 and you also know the set of candidate new heads that may have resulted
1408 from the destruction, you can set newheadnodes. This will enable the
1408 from the destruction, you can set newheadnodes. This will enable the
1409 code to update the branchheads cache, rather than having future code
1409 code to update the branchheads cache, rather than having future code
1410 decide it's invalid and regenerating it from scratch.
1410 decide it's invalid and regenerating it from scratch.
1411 '''
1411 '''
1412 # When one tries to:
1412 # When one tries to:
1413 # 1) destroy nodes thus calling this method (e.g. strip)
1413 # 1) destroy nodes thus calling this method (e.g. strip)
1414 # 2) use phasecache somewhere (e.g. commit)
1414 # 2) use phasecache somewhere (e.g. commit)
1415 #
1415 #
1416 # then 2) will fail because the phasecache contains nodes that were
1416 # then 2) will fail because the phasecache contains nodes that were
1417 # removed. We can either remove phasecache from the filecache,
1417 # removed. We can either remove phasecache from the filecache,
1418 # causing it to reload next time it is accessed, or simply filter
1418 # causing it to reload next time it is accessed, or simply filter
1419 # the removed nodes now and write the updated cache.
1419 # the removed nodes now and write the updated cache.
1420 if '_phasecache' in self._filecache:
1420 if '_phasecache' in self._filecache:
1421 self._phasecache.filterunknown(self)
1421 self._phasecache.filterunknown(self)
1422 self._phasecache.write()
1422 self._phasecache.write()
1423
1423
1424 # If we have info, newheadnodes, on how to update the branch cache, do
1424 # If we have info, newheadnodes, on how to update the branch cache, do
1425 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1425 # it, Otherwise, since nodes were destroyed, the cache is stale and this
1426 # will be caught the next time it is read.
1426 # will be caught the next time it is read.
1427 if newheadnodes:
1427 if newheadnodes:
1428 cl = self.changelog
1428 cl = self.changelog
1429 revgen = (cl.rev(node) for node in newheadnodes
1429 revgen = (cl.rev(node) for node in newheadnodes
1430 if cl.hasnode(node))
1430 if cl.hasnode(node))
1431 cache = self._branchcaches[None]
1431 cache = self._branchcaches[None]
1432 cache.update(self, revgen)
1432 cache.update(self, revgen)
1433 cache.write(self)
1433 cache.write(self)
1434
1434
1435 # Ensure the persistent tag cache is updated. Doing it now
1435 # Ensure the persistent tag cache is updated. Doing it now
1436 # means that the tag cache only has to worry about destroyed
1436 # means that the tag cache only has to worry about destroyed
1437 # heads immediately after a strip/rollback. That in turn
1437 # heads immediately after a strip/rollback. That in turn
1438 # guarantees that "cachetip == currenttip" (comparing both rev
1438 # guarantees that "cachetip == currenttip" (comparing both rev
1439 # and node) always means no nodes have been added or destroyed.
1439 # and node) always means no nodes have been added or destroyed.
1440
1440
1441 # XXX this is suboptimal when qrefresh'ing: we strip the current
1441 # XXX this is suboptimal when qrefresh'ing: we strip the current
1442 # head, refresh the tag cache, then immediately add a new head.
1442 # head, refresh the tag cache, then immediately add a new head.
1443 # But I think doing it this way is necessary for the "instant
1443 # But I think doing it this way is necessary for the "instant
1444 # tag cache retrieval" case to work.
1444 # tag cache retrieval" case to work.
1445 self.invalidate()
1445 self.invalidate()
1446
1446
1447 def walk(self, match, node=None):
1447 def walk(self, match, node=None):
1448 '''
1448 '''
1449 walk recursively through the directory tree or a given
1449 walk recursively through the directory tree or a given
1450 changeset, finding all files matched by the match
1450 changeset, finding all files matched by the match
1451 function
1451 function
1452 '''
1452 '''
1453 return self[node].walk(match)
1453 return self[node].walk(match)
1454
1454
1455 def status(self, node1='.', node2=None, match=None,
1455 def status(self, node1='.', node2=None, match=None,
1456 ignored=False, clean=False, unknown=False,
1456 ignored=False, clean=False, unknown=False,
1457 listsubrepos=False):
1457 listsubrepos=False):
1458 """return status of files between two nodes or node and working
1458 """return status of files between two nodes or node and working
1459 directory.
1459 directory.
1460
1460
1461 If node1 is None, use the first dirstate parent instead.
1461 If node1 is None, use the first dirstate parent instead.
1462 If node2 is None, compare node1 with working directory.
1462 If node2 is None, compare node1 with working directory.
1463 """
1463 """
1464
1464
1465 def mfmatches(ctx):
1465 def mfmatches(ctx):
1466 mf = ctx.manifest().copy()
1466 mf = ctx.manifest().copy()
1467 if match.always():
1467 if match.always():
1468 return mf
1468 return mf
1469 for fn in mf.keys():
1469 for fn in mf.keys():
1470 if not match(fn):
1470 if not match(fn):
1471 del mf[fn]
1471 del mf[fn]
1472 return mf
1472 return mf
1473
1473
1474 if isinstance(node1, context.changectx):
1474 if isinstance(node1, context.changectx):
1475 ctx1 = node1
1475 ctx1 = node1
1476 else:
1476 else:
1477 ctx1 = self[node1]
1477 ctx1 = self[node1]
1478 if isinstance(node2, context.changectx):
1478 if isinstance(node2, context.changectx):
1479 ctx2 = node2
1479 ctx2 = node2
1480 else:
1480 else:
1481 ctx2 = self[node2]
1481 ctx2 = self[node2]
1482
1482
1483 working = ctx2.rev() is None
1483 working = ctx2.rev() is None
1484 parentworking = working and ctx1 == self['.']
1484 parentworking = working and ctx1 == self['.']
1485 match = match or matchmod.always(self.root, self.getcwd())
1485 match = match or matchmod.always(self.root, self.getcwd())
1486 listignored, listclean, listunknown = ignored, clean, unknown
1486 listignored, listclean, listunknown = ignored, clean, unknown
1487
1487
1488 # load earliest manifest first for caching reasons
1488 # load earliest manifest first for caching reasons
1489 if not working and ctx2.rev() < ctx1.rev():
1489 if not working and ctx2.rev() < ctx1.rev():
1490 ctx2.manifest()
1490 ctx2.manifest()
1491
1491
1492 if not parentworking:
1492 if not parentworking:
1493 def bad(f, msg):
1493 def bad(f, msg):
1494 # 'f' may be a directory pattern from 'match.files()',
1494 # 'f' may be a directory pattern from 'match.files()',
1495 # so 'f not in ctx1' is not enough
1495 # so 'f not in ctx1' is not enough
1496 if f not in ctx1 and f not in ctx1.dirs():
1496 if f not in ctx1 and f not in ctx1.dirs():
1497 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1497 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1498 match.bad = bad
1498 match.bad = bad
1499
1499
1500 if working: # we need to scan the working dir
1500 if working: # we need to scan the working dir
1501 subrepos = []
1501 subrepos = []
1502 if '.hgsub' in self.dirstate:
1502 if '.hgsub' in self.dirstate:
1503 subrepos = sorted(ctx2.substate)
1503 subrepos = sorted(ctx2.substate)
1504 s = self.dirstate.status(match, subrepos, listignored,
1504 s = self.dirstate.status(match, subrepos, listignored,
1505 listclean, listunknown)
1505 listclean, listunknown)
1506 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1506 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1507
1507
1508 # check for any possibly clean files
1508 # check for any possibly clean files
1509 if parentworking and cmp:
1509 if parentworking and cmp:
1510 fixup = []
1510 fixup = []
1511 # do a full compare of any files that might have changed
1511 # do a full compare of any files that might have changed
1512 for f in sorted(cmp):
1512 for f in sorted(cmp):
1513 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1513 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1514 or ctx1[f].cmp(ctx2[f])):
1514 or ctx1[f].cmp(ctx2[f])):
1515 modified.append(f)
1515 modified.append(f)
1516 else:
1516 else:
1517 fixup.append(f)
1517 fixup.append(f)
1518
1518
1519 # update dirstate for files that are actually clean
1519 # update dirstate for files that are actually clean
1520 if fixup:
1520 if fixup:
1521 if listclean:
1521 if listclean:
1522 clean += fixup
1522 clean += fixup
1523
1523
1524 try:
1524 try:
1525 # updating the dirstate is optional
1525 # updating the dirstate is optional
1526 # so we don't wait on the lock
1526 # so we don't wait on the lock
1527 wlock = self.wlock(False)
1527 wlock = self.wlock(False)
1528 try:
1528 try:
1529 for f in fixup:
1529 for f in fixup:
1530 self.dirstate.normal(f)
1530 self.dirstate.normal(f)
1531 finally:
1531 finally:
1532 wlock.release()
1532 wlock.release()
1533 except error.LockError:
1533 except error.LockError:
1534 pass
1534 pass
1535
1535
1536 if not parentworking:
1536 if not parentworking:
1537 mf1 = mfmatches(ctx1)
1537 mf1 = mfmatches(ctx1)
1538 if working:
1538 if working:
1539 # we are comparing working dir against non-parent
1539 # we are comparing working dir against non-parent
1540 # generate a pseudo-manifest for the working dir
1540 # generate a pseudo-manifest for the working dir
1541 mf2 = mfmatches(self['.'])
1541 mf2 = mfmatches(self['.'])
1542 for f in cmp + modified + added:
1542 for f in cmp + modified + added:
1543 mf2[f] = None
1543 mf2[f] = None
1544 mf2.set(f, ctx2.flags(f))
1544 mf2.set(f, ctx2.flags(f))
1545 for f in removed:
1545 for f in removed:
1546 if f in mf2:
1546 if f in mf2:
1547 del mf2[f]
1547 del mf2[f]
1548 else:
1548 else:
1549 # we are comparing two revisions
1549 # we are comparing two revisions
1550 deleted, unknown, ignored = [], [], []
1550 deleted, unknown, ignored = [], [], []
1551 mf2 = mfmatches(ctx2)
1551 mf2 = mfmatches(ctx2)
1552
1552
1553 modified, added, clean = [], [], []
1553 modified, added, clean = [], [], []
1554 withflags = mf1.withflags() | mf2.withflags()
1554 withflags = mf1.withflags() | mf2.withflags()
1555 for fn in mf2:
1555 for fn in mf2:
1556 if fn in mf1:
1556 if fn in mf1:
1557 if (fn not in deleted and
1557 if (fn not in deleted and
1558 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1558 ((fn in withflags and mf1.flags(fn) != mf2.flags(fn)) or
1559 (mf1[fn] != mf2[fn] and
1559 (mf1[fn] != mf2[fn] and
1560 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1560 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1561 modified.append(fn)
1561 modified.append(fn)
1562 elif listclean:
1562 elif listclean:
1563 clean.append(fn)
1563 clean.append(fn)
1564 del mf1[fn]
1564 del mf1[fn]
1565 elif fn not in deleted:
1565 elif fn not in deleted:
1566 added.append(fn)
1566 added.append(fn)
1567 removed = mf1.keys()
1567 removed = mf1.keys()
1568
1568
1569 if working and modified and not self.dirstate._checklink:
1569 if working and modified and not self.dirstate._checklink:
1570 # Symlink placeholders may get non-symlink-like contents
1570 # Symlink placeholders may get non-symlink-like contents
1571 # via user error or dereferencing by NFS or Samba servers,
1571 # via user error or dereferencing by NFS or Samba servers,
1572 # so we filter out any placeholders that don't look like a
1572 # so we filter out any placeholders that don't look like a
1573 # symlink
1573 # symlink
1574 sane = []
1574 sane = []
1575 for f in modified:
1575 for f in modified:
1576 if ctx2.flags(f) == 'l':
1576 if ctx2.flags(f) == 'l':
1577 d = ctx2[f].data()
1577 d = ctx2[f].data()
1578 if len(d) >= 1024 or '\n' in d or util.binary(d):
1578 if len(d) >= 1024 or '\n' in d or util.binary(d):
1579 self.ui.debug('ignoring suspect symlink placeholder'
1579 self.ui.debug('ignoring suspect symlink placeholder'
1580 ' "%s"\n' % f)
1580 ' "%s"\n' % f)
1581 continue
1581 continue
1582 sane.append(f)
1582 sane.append(f)
1583 modified = sane
1583 modified = sane
1584
1584
1585 r = modified, added, removed, deleted, unknown, ignored, clean
1585 r = modified, added, removed, deleted, unknown, ignored, clean
1586
1586
1587 if listsubrepos:
1587 if listsubrepos:
1588 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1588 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1589 if working:
1589 if working:
1590 rev2 = None
1590 rev2 = None
1591 else:
1591 else:
1592 rev2 = ctx2.substate[subpath][1]
1592 rev2 = ctx2.substate[subpath][1]
1593 try:
1593 try:
1594 submatch = matchmod.narrowmatcher(subpath, match)
1594 submatch = matchmod.narrowmatcher(subpath, match)
1595 s = sub.status(rev2, match=submatch, ignored=listignored,
1595 s = sub.status(rev2, match=submatch, ignored=listignored,
1596 clean=listclean, unknown=listunknown,
1596 clean=listclean, unknown=listunknown,
1597 listsubrepos=True)
1597 listsubrepos=True)
1598 for rfiles, sfiles in zip(r, s):
1598 for rfiles, sfiles in zip(r, s):
1599 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1599 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1600 except error.LookupError:
1600 except error.LookupError:
1601 self.ui.status(_("skipping missing subrepository: %s\n")
1601 self.ui.status(_("skipping missing subrepository: %s\n")
1602 % subpath)
1602 % subpath)
1603
1603
1604 for l in r:
1604 for l in r:
1605 l.sort()
1605 l.sort()
1606 return r
1606 return r
1607
1607
1608 def heads(self, start=None):
1608 def heads(self, start=None):
1609 heads = self.changelog.heads(start)
1609 heads = self.changelog.heads(start)
1610 # sort the output in rev descending order
1610 # sort the output in rev descending order
1611 return sorted(heads, key=self.changelog.rev, reverse=True)
1611 return sorted(heads, key=self.changelog.rev, reverse=True)
1612
1612
1613 def branchheads(self, branch=None, start=None, closed=False):
1613 def branchheads(self, branch=None, start=None, closed=False):
1614 '''return a (possibly filtered) list of heads for the given branch
1614 '''return a (possibly filtered) list of heads for the given branch
1615
1615
1616 Heads are returned in topological order, from newest to oldest.
1616 Heads are returned in topological order, from newest to oldest.
1617 If branch is None, use the dirstate branch.
1617 If branch is None, use the dirstate branch.
1618 If start is not None, return only heads reachable from start.
1618 If start is not None, return only heads reachable from start.
1619 If closed is True, return heads that are marked as closed as well.
1619 If closed is True, return heads that are marked as closed as well.
1620 '''
1620 '''
1621 if branch is None:
1621 if branch is None:
1622 branch = self[None].branch()
1622 branch = self[None].branch()
1623 branches = self.branchmap()
1623 branches = self.branchmap()
1624 if branch not in branches:
1624 if branch not in branches:
1625 return []
1625 return []
1626 # the cache returns heads ordered lowest to highest
1626 # the cache returns heads ordered lowest to highest
1627 bheads = list(reversed(branches[branch]))
1627 bheads = list(reversed(branches[branch]))
1628 if start is not None:
1628 if start is not None:
1629 # filter out the heads that cannot be reached from startrev
1629 # filter out the heads that cannot be reached from startrev
1630 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1630 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1631 bheads = [h for h in bheads if h in fbheads]
1631 bheads = [h for h in bheads if h in fbheads]
1632 if not closed:
1632 if not closed:
1633 bheads = [h for h in bheads if not self[h].closesbranch()]
1633 bheads = [h for h in bheads if not self[h].closesbranch()]
1634 return bheads
1634 return bheads
1635
1635
1636 def branches(self, nodes):
1636 def branches(self, nodes):
1637 if not nodes:
1637 if not nodes:
1638 nodes = [self.changelog.tip()]
1638 nodes = [self.changelog.tip()]
1639 b = []
1639 b = []
1640 for n in nodes:
1640 for n in nodes:
1641 t = n
1641 t = n
1642 while True:
1642 while True:
1643 p = self.changelog.parents(n)
1643 p = self.changelog.parents(n)
1644 if p[1] != nullid or p[0] == nullid:
1644 if p[1] != nullid or p[0] == nullid:
1645 b.append((t, n, p[0], p[1]))
1645 b.append((t, n, p[0], p[1]))
1646 break
1646 break
1647 n = p[0]
1647 n = p[0]
1648 return b
1648 return b
1649
1649
1650 def between(self, pairs):
1650 def between(self, pairs):
1651 r = []
1651 r = []
1652
1652
1653 for top, bottom in pairs:
1653 for top, bottom in pairs:
1654 n, l, i = top, [], 0
1654 n, l, i = top, [], 0
1655 f = 1
1655 f = 1
1656
1656
1657 while n != bottom and n != nullid:
1657 while n != bottom and n != nullid:
1658 p = self.changelog.parents(n)[0]
1658 p = self.changelog.parents(n)[0]
1659 if i == f:
1659 if i == f:
1660 l.append(n)
1660 l.append(n)
1661 f = f * 2
1661 f = f * 2
1662 n = p
1662 n = p
1663 i += 1
1663 i += 1
1664
1664
1665 r.append(l)
1665 r.append(l)
1666
1666
1667 return r
1667 return r
1668
1668
1669 def pull(self, remote, heads=None, force=False):
1669 def pull(self, remote, heads=None, force=False):
1670 # don't open transaction for nothing or you break future useful
1670 # don't open transaction for nothing or you break future useful
1671 # rollback call
1671 # rollback call
1672 tr = None
1672 tr = None
1673 trname = 'pull\n' + util.hidepassword(remote.url())
1673 trname = 'pull\n' + util.hidepassword(remote.url())
1674 lock = self.lock()
1674 lock = self.lock()
1675 try:
1675 try:
1676 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1676 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1677 force=force)
1677 force=force)
1678 common, fetch, rheads = tmp
1678 common, fetch, rheads = tmp
1679 if not fetch:
1679 if not fetch:
1680 self.ui.status(_("no changes found\n"))
1680 self.ui.status(_("no changes found\n"))
1681 added = []
1681 added = []
1682 result = 0
1682 result = 0
1683 else:
1683 else:
1684 tr = self.transaction(trname)
1684 tr = self.transaction(trname)
1685 if heads is None and list(common) == [nullid]:
1685 if heads is None and list(common) == [nullid]:
1686 self.ui.status(_("requesting all changes\n"))
1686 self.ui.status(_("requesting all changes\n"))
1687 elif heads is None and remote.capable('changegroupsubset'):
1687 elif heads is None and remote.capable('changegroupsubset'):
1688 # issue1320, avoid a race if remote changed after discovery
1688 # issue1320, avoid a race if remote changed after discovery
1689 heads = rheads
1689 heads = rheads
1690
1690
1691 if remote.capable('getbundle'):
1691 if remote.capable('getbundle'):
1692 cg = remote.getbundle('pull', common=common,
1692 cg = remote.getbundle('pull', common=common,
1693 heads=heads or rheads)
1693 heads=heads or rheads)
1694 elif heads is None:
1694 elif heads is None:
1695 cg = remote.changegroup(fetch, 'pull')
1695 cg = remote.changegroup(fetch, 'pull')
1696 elif not remote.capable('changegroupsubset'):
1696 elif not remote.capable('changegroupsubset'):
1697 raise util.Abort(_("partial pull cannot be done because "
1697 raise util.Abort(_("partial pull cannot be done because "
1698 "other repository doesn't support "
1698 "other repository doesn't support "
1699 "changegroupsubset."))
1699 "changegroupsubset."))
1700 else:
1700 else:
1701 cg = remote.changegroupsubset(fetch, heads, 'pull')
1701 cg = remote.changegroupsubset(fetch, heads, 'pull')
1702 clstart = len(self.changelog)
1702 clstart = len(self.changelog)
1703 result = self.addchangegroup(cg, 'pull', remote.url())
1703 result = self.addchangegroup(cg, 'pull', remote.url())
1704 clend = len(self.changelog)
1704 clend = len(self.changelog)
1705 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1705 added = [self.changelog.node(r) for r in xrange(clstart, clend)]
1706
1706
1707 # compute target subset
1707 # compute target subset
1708 if heads is None:
1708 if heads is None:
1709 # We pulled every thing possible
1709 # We pulled every thing possible
1710 # sync on everything common
1710 # sync on everything common
1711 subset = common + added
1711 subset = common + added
1712 else:
1712 else:
1713 # We pulled a specific subset
1713 # We pulled a specific subset
1714 # sync on this subset
1714 # sync on this subset
1715 subset = heads
1715 subset = heads
1716
1716
1717 # Get remote phases data from remote
1717 # Get remote phases data from remote
1718 remotephases = remote.listkeys('phases')
1718 remotephases = remote.listkeys('phases')
1719 publishing = bool(remotephases.get('publishing', False))
1719 publishing = bool(remotephases.get('publishing', False))
1720 if remotephases and not publishing:
1720 if remotephases and not publishing:
1721 # remote is new and unpublishing
1721 # remote is new and unpublishing
1722 pheads, _dr = phases.analyzeremotephases(self, subset,
1722 pheads, _dr = phases.analyzeremotephases(self, subset,
1723 remotephases)
1723 remotephases)
1724 phases.advanceboundary(self, phases.public, pheads)
1724 phases.advanceboundary(self, phases.public, pheads)
1725 phases.advanceboundary(self, phases.draft, subset)
1725 phases.advanceboundary(self, phases.draft, subset)
1726 else:
1726 else:
1727 # Remote is old or publishing all common changesets
1727 # Remote is old or publishing all common changesets
1728 # should be seen as public
1728 # should be seen as public
1729 phases.advanceboundary(self, phases.public, subset)
1729 phases.advanceboundary(self, phases.public, subset)
1730
1730
1731 if obsolete._enabled:
1731 if obsolete._enabled:
1732 self.ui.debug('fetching remote obsolete markers\n')
1732 self.ui.debug('fetching remote obsolete markers\n')
1733 remoteobs = remote.listkeys('obsolete')
1733 remoteobs = remote.listkeys('obsolete')
1734 if 'dump0' in remoteobs:
1734 if 'dump0' in remoteobs:
1735 if tr is None:
1735 if tr is None:
1736 tr = self.transaction(trname)
1736 tr = self.transaction(trname)
1737 for key in sorted(remoteobs, reverse=True):
1737 for key in sorted(remoteobs, reverse=True):
1738 if key.startswith('dump'):
1738 if key.startswith('dump'):
1739 data = base85.b85decode(remoteobs[key])
1739 data = base85.b85decode(remoteobs[key])
1740 self.obsstore.mergemarkers(tr, data)
1740 self.obsstore.mergemarkers(tr, data)
1741 self.invalidatevolatilesets()
1741 self.invalidatevolatilesets()
1742 if tr is not None:
1742 if tr is not None:
1743 tr.close()
1743 tr.close()
1744 finally:
1744 finally:
1745 if tr is not None:
1745 if tr is not None:
1746 tr.release()
1746 tr.release()
1747 lock.release()
1747 lock.release()
1748
1748
1749 return result
1749 return result
1750
1750
1751 def checkpush(self, force, revs):
1751 def checkpush(self, force, revs):
1752 """Extensions can override this function if additional checks have
1752 """Extensions can override this function if additional checks have
1753 to be performed before pushing, or call it if they override push
1753 to be performed before pushing, or call it if they override push
1754 command.
1754 command.
1755 """
1755 """
1756 pass
1756 pass
1757
1757
1758 def push(self, remote, force=False, revs=None, newbranch=False):
1758 def push(self, remote, force=False, revs=None, newbranch=False):
1759 '''Push outgoing changesets (limited by revs) from the current
1759 '''Push outgoing changesets (limited by revs) from the current
1760 repository to remote. Return an integer:
1760 repository to remote. Return an integer:
1761 - None means nothing to push
1761 - None means nothing to push
1762 - 0 means HTTP error
1762 - 0 means HTTP error
1763 - 1 means we pushed and remote head count is unchanged *or*
1763 - 1 means we pushed and remote head count is unchanged *or*
1764 we have outgoing changesets but refused to push
1764 we have outgoing changesets but refused to push
1765 - other values as described by addchangegroup()
1765 - other values as described by addchangegroup()
1766 '''
1766 '''
1767 # there are two ways to push to remote repo:
1767 # there are two ways to push to remote repo:
1768 #
1768 #
1769 # addchangegroup assumes local user can lock remote
1769 # addchangegroup assumes local user can lock remote
1770 # repo (local filesystem, old ssh servers).
1770 # repo (local filesystem, old ssh servers).
1771 #
1771 #
1772 # unbundle assumes local user cannot lock remote repo (new ssh
1772 # unbundle assumes local user cannot lock remote repo (new ssh
1773 # servers, http servers).
1773 # servers, http servers).
1774
1774
1775 if not remote.canpush():
1775 if not remote.canpush():
1776 raise util.Abort(_("destination does not support push"))
1776 raise util.Abort(_("destination does not support push"))
1777 unfi = self.unfiltered()
1777 unfi = self.unfiltered()
1778 # get local lock as we might write phase data
1778 # get local lock as we might write phase data
1779 locallock = self.lock()
1779 locallock = self.lock()
1780 try:
1780 try:
1781 self.checkpush(force, revs)
1781 self.checkpush(force, revs)
1782 lock = None
1782 lock = None
1783 unbundle = remote.capable('unbundle')
1783 unbundle = remote.capable('unbundle')
1784 if not unbundle:
1784 if not unbundle:
1785 lock = remote.lock()
1785 lock = remote.lock()
1786 try:
1786 try:
1787 # discovery
1787 # discovery
1788 fci = discovery.findcommonincoming
1788 fci = discovery.findcommonincoming
1789 commoninc = fci(unfi, remote, force=force)
1789 commoninc = fci(unfi, remote, force=force)
1790 common, inc, remoteheads = commoninc
1790 common, inc, remoteheads = commoninc
1791 fco = discovery.findcommonoutgoing
1791 fco = discovery.findcommonoutgoing
1792 outgoing = fco(unfi, remote, onlyheads=revs,
1792 outgoing = fco(unfi, remote, onlyheads=revs,
1793 commoninc=commoninc, force=force)
1793 commoninc=commoninc, force=force)
1794
1794
1795
1795
1796 if not outgoing.missing:
1796 if not outgoing.missing:
1797 # nothing to push
1797 # nothing to push
1798 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1798 scmutil.nochangesfound(unfi.ui, unfi, outgoing.excluded)
1799 ret = None
1799 ret = None
1800 else:
1800 else:
1801 # something to push
1801 # something to push
1802 if not force:
1802 if not force:
1803 # if self.obsstore == False --> no obsolete
1803 # if self.obsstore == False --> no obsolete
1804 # then, save the iteration
1804 # then, save the iteration
1805 if unfi.obsstore:
1805 if unfi.obsstore:
1806 # this message are here for 80 char limit reason
1806 # this message are here for 80 char limit reason
1807 mso = _("push includes obsolete changeset: %s!")
1807 mso = _("push includes obsolete changeset: %s!")
1808 mst = "push includes %s changeset: %s!"
1808 mst = "push includes %s changeset: %s!"
1809 # plain versions for i18n tool to detect them
1809 # plain versions for i18n tool to detect them
1810 _("push includes unstable changeset: %s!")
1810 _("push includes unstable changeset: %s!")
1811 _("push includes bumped changeset: %s!")
1811 _("push includes bumped changeset: %s!")
1812 _("push includes divergent changeset: %s!")
1812 _("push includes divergent changeset: %s!")
1813 # If we are to push if there is at least one
1813 # If we are to push if there is at least one
1814 # obsolete or unstable changeset in missing, at
1814 # obsolete or unstable changeset in missing, at
1815 # least one of the missinghead will be obsolete or
1815 # least one of the missinghead will be obsolete or
1816 # unstable. So checking heads only is ok
1816 # unstable. So checking heads only is ok
1817 for node in outgoing.missingheads:
1817 for node in outgoing.missingheads:
1818 ctx = unfi[node]
1818 ctx = unfi[node]
1819 if ctx.obsolete():
1819 if ctx.obsolete():
1820 raise util.Abort(mso % ctx)
1820 raise util.Abort(mso % ctx)
1821 elif ctx.troubled():
1821 elif ctx.troubled():
1822 raise util.Abort(_(mst)
1822 raise util.Abort(_(mst)
1823 % (ctx.troubles()[0],
1823 % (ctx.troubles()[0],
1824 ctx))
1824 ctx))
1825 discovery.checkheads(unfi, remote, outgoing,
1825 discovery.checkheads(unfi, remote, outgoing,
1826 remoteheads, newbranch,
1826 remoteheads, newbranch,
1827 bool(inc))
1827 bool(inc))
1828
1828
1829 # create a changegroup from local
1829 # create a changegroup from local
1830 if revs is None and not outgoing.excluded:
1830 if revs is None and not outgoing.excluded:
1831 # push everything,
1831 # push everything,
1832 # use the fast path, no race possible on push
1832 # use the fast path, no race possible on push
1833 cg = self._changegroup(outgoing.missing, 'push')
1833 cg = self._changegroup(outgoing.missing, 'push')
1834 else:
1834 else:
1835 cg = self.getlocalbundle('push', outgoing)
1835 cg = self.getlocalbundle('push', outgoing)
1836
1836
1837 # apply changegroup to remote
1837 # apply changegroup to remote
1838 if unbundle:
1838 if unbundle:
1839 # local repo finds heads on server, finds out what
1839 # local repo finds heads on server, finds out what
1840 # revs it must push. once revs transferred, if server
1840 # revs it must push. once revs transferred, if server
1841 # finds it has different heads (someone else won
1841 # finds it has different heads (someone else won
1842 # commit/push race), server aborts.
1842 # commit/push race), server aborts.
1843 if force:
1843 if force:
1844 remoteheads = ['force']
1844 remoteheads = ['force']
1845 # ssh: return remote's addchangegroup()
1845 # ssh: return remote's addchangegroup()
1846 # http: return remote's addchangegroup() or 0 for error
1846 # http: return remote's addchangegroup() or 0 for error
1847 ret = remote.unbundle(cg, remoteheads, 'push')
1847 ret = remote.unbundle(cg, remoteheads, 'push')
1848 else:
1848 else:
1849 # we return an integer indicating remote head count
1849 # we return an integer indicating remote head count
1850 # change
1850 # change
1851 ret = remote.addchangegroup(cg, 'push', self.url())
1851 ret = remote.addchangegroup(cg, 'push', self.url())
1852
1852
1853 if ret:
1853 if ret:
1854 # push succeed, synchronize target of the push
1854 # push succeed, synchronize target of the push
1855 cheads = outgoing.missingheads
1855 cheads = outgoing.missingheads
1856 elif revs is None:
1856 elif revs is None:
1857 # All out push fails. synchronize all common
1857 # All out push fails. synchronize all common
1858 cheads = outgoing.commonheads
1858 cheads = outgoing.commonheads
1859 else:
1859 else:
1860 # I want cheads = heads(::missingheads and ::commonheads)
1860 # I want cheads = heads(::missingheads and ::commonheads)
1861 # (missingheads is revs with secret changeset filtered out)
1861 # (missingheads is revs with secret changeset filtered out)
1862 #
1862 #
1863 # This can be expressed as:
1863 # This can be expressed as:
1864 # cheads = ( (missingheads and ::commonheads)
1864 # cheads = ( (missingheads and ::commonheads)
1865 # + (commonheads and ::missingheads))"
1865 # + (commonheads and ::missingheads))"
1866 # )
1866 # )
1867 #
1867 #
1868 # while trying to push we already computed the following:
1868 # while trying to push we already computed the following:
1869 # common = (::commonheads)
1869 # common = (::commonheads)
1870 # missing = ((commonheads::missingheads) - commonheads)
1870 # missing = ((commonheads::missingheads) - commonheads)
1871 #
1871 #
1872 # We can pick:
1872 # We can pick:
1873 # * missingheads part of common (::commonheads)
1873 # * missingheads part of common (::commonheads)
1874 common = set(outgoing.common)
1874 common = set(outgoing.common)
1875 cheads = [node for node in revs if node in common]
1875 cheads = [node for node in revs if node in common]
1876 # and
1876 # and
1877 # * commonheads parents on missing
1877 # * commonheads parents on missing
1878 revset = unfi.set('%ln and parents(roots(%ln))',
1878 revset = unfi.set('%ln and parents(roots(%ln))',
1879 outgoing.commonheads,
1879 outgoing.commonheads,
1880 outgoing.missing)
1880 outgoing.missing)
1881 cheads.extend(c.node() for c in revset)
1881 cheads.extend(c.node() for c in revset)
1882 # even when we don't push, exchanging phase data is useful
1882 # even when we don't push, exchanging phase data is useful
1883 remotephases = remote.listkeys('phases')
1883 remotephases = remote.listkeys('phases')
1884 if not remotephases: # old server or public only repo
1884 if not remotephases: # old server or public only repo
1885 phases.advanceboundary(self, phases.public, cheads)
1885 phases.advanceboundary(self, phases.public, cheads)
1886 # don't push any phase data as there is nothing to push
1886 # don't push any phase data as there is nothing to push
1887 else:
1887 else:
1888 ana = phases.analyzeremotephases(self, cheads, remotephases)
1888 ana = phases.analyzeremotephases(self, cheads, remotephases)
1889 pheads, droots = ana
1889 pheads, droots = ana
1890 ### Apply remote phase on local
1890 ### Apply remote phase on local
1891 if remotephases.get('publishing', False):
1891 if remotephases.get('publishing', False):
1892 phases.advanceboundary(self, phases.public, cheads)
1892 phases.advanceboundary(self, phases.public, cheads)
1893 else: # publish = False
1893 else: # publish = False
1894 phases.advanceboundary(self, phases.public, pheads)
1894 phases.advanceboundary(self, phases.public, pheads)
1895 phases.advanceboundary(self, phases.draft, cheads)
1895 phases.advanceboundary(self, phases.draft, cheads)
1896 ### Apply local phase on remote
1896 ### Apply local phase on remote
1897
1897
1898 # Get the list of all revs draft on remote by public here.
1898 # Get the list of all revs draft on remote by public here.
1899 # XXX Beware that revset break if droots is not strictly
1899 # XXX Beware that revset break if droots is not strictly
1900 # XXX root we may want to ensure it is but it is costly
1900 # XXX root we may want to ensure it is but it is costly
1901 outdated = unfi.set('heads((%ln::%ln) and public())',
1901 outdated = unfi.set('heads((%ln::%ln) and public())',
1902 droots, cheads)
1902 droots, cheads)
1903 for newremotehead in outdated:
1903 for newremotehead in outdated:
1904 r = remote.pushkey('phases',
1904 r = remote.pushkey('phases',
1905 newremotehead.hex(),
1905 newremotehead.hex(),
1906 str(phases.draft),
1906 str(phases.draft),
1907 str(phases.public))
1907 str(phases.public))
1908 if not r:
1908 if not r:
1909 self.ui.warn(_('updating %s to public failed!\n')
1909 self.ui.warn(_('updating %s to public failed!\n')
1910 % newremotehead)
1910 % newremotehead)
1911 self.ui.debug('try to push obsolete markers to remote\n')
1911 self.ui.debug('try to push obsolete markers to remote\n')
1912 if (obsolete._enabled and self.obsstore and
1912 if (obsolete._enabled and self.obsstore and
1913 'obsolete' in remote.listkeys('namespaces')):
1913 'obsolete' in remote.listkeys('namespaces')):
1914 rslts = []
1914 rslts = []
1915 remotedata = self.listkeys('obsolete')
1915 remotedata = self.listkeys('obsolete')
1916 for key in sorted(remotedata, reverse=True):
1916 for key in sorted(remotedata, reverse=True):
1917 # reverse sort to ensure we end with dump0
1917 # reverse sort to ensure we end with dump0
1918 data = remotedata[key]
1918 data = remotedata[key]
1919 rslts.append(remote.pushkey('obsolete', key, '', data))
1919 rslts.append(remote.pushkey('obsolete', key, '', data))
1920 if [r for r in rslts if not r]:
1920 if [r for r in rslts if not r]:
1921 msg = _('failed to push some obsolete markers!\n')
1921 msg = _('failed to push some obsolete markers!\n')
1922 self.ui.warn(msg)
1922 self.ui.warn(msg)
1923 finally:
1923 finally:
1924 if lock is not None:
1924 if lock is not None:
1925 lock.release()
1925 lock.release()
1926 finally:
1926 finally:
1927 locallock.release()
1927 locallock.release()
1928
1928
1929 self.ui.debug("checking for updated bookmarks\n")
1929 self.ui.debug("checking for updated bookmarks\n")
1930 rb = remote.listkeys('bookmarks')
1930 rb = remote.listkeys('bookmarks')
1931 for k in rb.keys():
1931 for k in rb.keys():
1932 if k in unfi._bookmarks:
1932 if k in unfi._bookmarks:
1933 nr, nl = rb[k], hex(self._bookmarks[k])
1933 nr, nl = rb[k], hex(self._bookmarks[k])
1934 if nr in unfi:
1934 if nr in unfi:
1935 cr = unfi[nr]
1935 cr = unfi[nr]
1936 cl = unfi[nl]
1936 cl = unfi[nl]
1937 if bookmarks.validdest(unfi, cr, cl):
1937 if bookmarks.validdest(unfi, cr, cl):
1938 r = remote.pushkey('bookmarks', k, nr, nl)
1938 r = remote.pushkey('bookmarks', k, nr, nl)
1939 if r:
1939 if r:
1940 self.ui.status(_("updating bookmark %s\n") % k)
1940 self.ui.status(_("updating bookmark %s\n") % k)
1941 else:
1941 else:
1942 self.ui.warn(_('updating bookmark %s'
1942 self.ui.warn(_('updating bookmark %s'
1943 ' failed!\n') % k)
1943 ' failed!\n') % k)
1944
1944
1945 return ret
1945 return ret
1946
1946
1947 def changegroupinfo(self, nodes, source):
1947 def changegroupinfo(self, nodes, source):
1948 if self.ui.verbose or source == 'bundle':
1948 if self.ui.verbose or source == 'bundle':
1949 self.ui.status(_("%d changesets found\n") % len(nodes))
1949 self.ui.status(_("%d changesets found\n") % len(nodes))
1950 if self.ui.debugflag:
1950 if self.ui.debugflag:
1951 self.ui.debug("list of changesets:\n")
1951 self.ui.debug("list of changesets:\n")
1952 for node in nodes:
1952 for node in nodes:
1953 self.ui.debug("%s\n" % hex(node))
1953 self.ui.debug("%s\n" % hex(node))
1954
1954
1955 def changegroupsubset(self, bases, heads, source):
1955 def changegroupsubset(self, bases, heads, source):
1956 """Compute a changegroup consisting of all the nodes that are
1956 """Compute a changegroup consisting of all the nodes that are
1957 descendants of any of the bases and ancestors of any of the heads.
1957 descendants of any of the bases and ancestors of any of the heads.
1958 Return a chunkbuffer object whose read() method will return
1958 Return a chunkbuffer object whose read() method will return
1959 successive changegroup chunks.
1959 successive changegroup chunks.
1960
1960
1961 It is fairly complex as determining which filenodes and which
1961 It is fairly complex as determining which filenodes and which
1962 manifest nodes need to be included for the changeset to be complete
1962 manifest nodes need to be included for the changeset to be complete
1963 is non-trivial.
1963 is non-trivial.
1964
1964
1965 Another wrinkle is doing the reverse, figuring out which changeset in
1965 Another wrinkle is doing the reverse, figuring out which changeset in
1966 the changegroup a particular filenode or manifestnode belongs to.
1966 the changegroup a particular filenode or manifestnode belongs to.
1967 """
1967 """
1968 cl = self.changelog
1968 cl = self.changelog
1969 if not bases:
1969 if not bases:
1970 bases = [nullid]
1970 bases = [nullid]
1971 csets, bases, heads = cl.nodesbetween(bases, heads)
1971 csets, bases, heads = cl.nodesbetween(bases, heads)
1972 # We assume that all ancestors of bases are known
1972 # We assume that all ancestors of bases are known
1973 common = cl.ancestors([cl.rev(n) for n in bases])
1973 common = cl.ancestors([cl.rev(n) for n in bases])
1974 return self._changegroupsubset(common, csets, heads, source)
1974 return self._changegroupsubset(common, csets, heads, source)
1975
1975
1976 def getlocalbundle(self, source, outgoing):
1976 def getlocalbundle(self, source, outgoing):
1977 """Like getbundle, but taking a discovery.outgoing as an argument.
1977 """Like getbundle, but taking a discovery.outgoing as an argument.
1978
1978
1979 This is only implemented for local repos and reuses potentially
1979 This is only implemented for local repos and reuses potentially
1980 precomputed sets in outgoing."""
1980 precomputed sets in outgoing."""
1981 if not outgoing.missing:
1981 if not outgoing.missing:
1982 return None
1982 return None
1983 return self._changegroupsubset(outgoing.common,
1983 return self._changegroupsubset(outgoing.common,
1984 outgoing.missing,
1984 outgoing.missing,
1985 outgoing.missingheads,
1985 outgoing.missingheads,
1986 source)
1986 source)
1987
1987
1988 def getbundle(self, source, heads=None, common=None):
1988 def getbundle(self, source, heads=None, common=None):
1989 """Like changegroupsubset, but returns the set difference between the
1989 """Like changegroupsubset, but returns the set difference between the
1990 ancestors of heads and the ancestors common.
1990 ancestors of heads and the ancestors common.
1991
1991
1992 If heads is None, use the local heads. If common is None, use [nullid].
1992 If heads is None, use the local heads. If common is None, use [nullid].
1993
1993
1994 The nodes in common might not all be known locally due to the way the
1994 The nodes in common might not all be known locally due to the way the
1995 current discovery protocol works.
1995 current discovery protocol works.
1996 """
1996 """
1997 cl = self.changelog
1997 cl = self.changelog
1998 if common:
1998 if common:
1999 hasnode = cl.hasnode
1999 hasnode = cl.hasnode
2000 common = [n for n in common if hasnode(n)]
2000 common = [n for n in common if hasnode(n)]
2001 else:
2001 else:
2002 common = [nullid]
2002 common = [nullid]
2003 if not heads:
2003 if not heads:
2004 heads = cl.heads()
2004 heads = cl.heads()
2005 return self.getlocalbundle(source,
2005 return self.getlocalbundle(source,
2006 discovery.outgoing(cl, common, heads))
2006 discovery.outgoing(cl, common, heads))
2007
2007
2008 @unfilteredmethod
2008 @unfilteredmethod
2009 def _changegroupsubset(self, commonrevs, csets, heads, source):
2009 def _changegroupsubset(self, commonrevs, csets, heads, source):
2010
2010
2011 cl = self.changelog
2011 cl = self.changelog
2012 mf = self.manifest
2012 mf = self.manifest
2013 mfs = {} # needed manifests
2013 mfs = {} # needed manifests
2014 fnodes = {} # needed file nodes
2014 fnodes = {} # needed file nodes
2015 changedfiles = set()
2015 changedfiles = set()
2016 fstate = ['', {}]
2016 fstate = ['', {}]
2017 count = [0, 0]
2017 count = [0, 0]
2018
2018
2019 # can we go through the fast path ?
2019 # can we go through the fast path ?
2020 heads.sort()
2020 heads.sort()
2021 if heads == sorted(self.heads()):
2021 if heads == sorted(self.heads()):
2022 return self._changegroup(csets, source)
2022 return self._changegroup(csets, source)
2023
2023
2024 # slow path
2024 # slow path
2025 self.hook('preoutgoing', throw=True, source=source)
2025 self.hook('preoutgoing', throw=True, source=source)
2026 self.changegroupinfo(csets, source)
2026 self.changegroupinfo(csets, source)
2027
2027
2028 # filter any nodes that claim to be part of the known set
2028 # filter any nodes that claim to be part of the known set
2029 def prune(revlog, missing):
2029 def prune(revlog, missing):
2030 rr, rl = revlog.rev, revlog.linkrev
2030 rr, rl = revlog.rev, revlog.linkrev
2031 return [n for n in missing
2031 return [n for n in missing
2032 if rl(rr(n)) not in commonrevs]
2032 if rl(rr(n)) not in commonrevs]
2033
2033
2034 progress = self.ui.progress
2034 progress = self.ui.progress
2035 _bundling = _('bundling')
2035 _bundling = _('bundling')
2036 _changesets = _('changesets')
2036 _changesets = _('changesets')
2037 _manifests = _('manifests')
2037 _manifests = _('manifests')
2038 _files = _('files')
2038 _files = _('files')
2039
2039
2040 def lookup(revlog, x):
2040 def lookup(revlog, x):
2041 if revlog == cl:
2041 if revlog == cl:
2042 c = cl.read(x)
2042 c = cl.read(x)
2043 changedfiles.update(c[3])
2043 changedfiles.update(c[3])
2044 mfs.setdefault(c[0], x)
2044 mfs.setdefault(c[0], x)
2045 count[0] += 1
2045 count[0] += 1
2046 progress(_bundling, count[0],
2046 progress(_bundling, count[0],
2047 unit=_changesets, total=count[1])
2047 unit=_changesets, total=count[1])
2048 return x
2048 return x
2049 elif revlog == mf:
2049 elif revlog == mf:
2050 clnode = mfs[x]
2050 clnode = mfs[x]
2051 mdata = mf.readfast(x)
2051 mdata = mf.readfast(x)
2052 for f, n in mdata.iteritems():
2052 for f, n in mdata.iteritems():
2053 if f in changedfiles:
2053 if f in changedfiles:
2054 fnodes[f].setdefault(n, clnode)
2054 fnodes[f].setdefault(n, clnode)
2055 count[0] += 1
2055 count[0] += 1
2056 progress(_bundling, count[0],
2056 progress(_bundling, count[0],
2057 unit=_manifests, total=count[1])
2057 unit=_manifests, total=count[1])
2058 return clnode
2058 return clnode
2059 else:
2059 else:
2060 progress(_bundling, count[0], item=fstate[0],
2060 progress(_bundling, count[0], item=fstate[0],
2061 unit=_files, total=count[1])
2061 unit=_files, total=count[1])
2062 return fstate[1][x]
2062 return fstate[1][x]
2063
2063
2064 bundler = changegroup.bundle10(lookup)
2064 bundler = changegroup.bundle10(lookup)
2065 reorder = self.ui.config('bundle', 'reorder', 'auto')
2065 reorder = self.ui.config('bundle', 'reorder', 'auto')
2066 if reorder == 'auto':
2066 if reorder == 'auto':
2067 reorder = None
2067 reorder = None
2068 else:
2068 else:
2069 reorder = util.parsebool(reorder)
2069 reorder = util.parsebool(reorder)
2070
2070
2071 def gengroup():
2071 def gengroup():
2072 # Create a changenode group generator that will call our functions
2072 # Create a changenode group generator that will call our functions
2073 # back to lookup the owning changenode and collect information.
2073 # back to lookup the owning changenode and collect information.
2074 count[:] = [0, len(csets)]
2074 count[:] = [0, len(csets)]
2075 for chunk in cl.group(csets, bundler, reorder=reorder):
2075 for chunk in cl.group(csets, bundler, reorder=reorder):
2076 yield chunk
2076 yield chunk
2077 progress(_bundling, None)
2077 progress(_bundling, None)
2078
2078
2079 # Create a generator for the manifestnodes that calls our lookup
2079 # Create a generator for the manifestnodes that calls our lookup
2080 # and data collection functions back.
2080 # and data collection functions back.
2081 for f in changedfiles:
2081 for f in changedfiles:
2082 fnodes[f] = {}
2082 fnodes[f] = {}
2083 count[:] = [0, len(mfs)]
2083 count[:] = [0, len(mfs)]
2084 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2084 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
2085 yield chunk
2085 yield chunk
2086 progress(_bundling, None)
2086 progress(_bundling, None)
2087
2087
2088 mfs.clear()
2088 mfs.clear()
2089
2089
2090 # Go through all our files in order sorted by name.
2090 # Go through all our files in order sorted by name.
2091 count[:] = [0, len(changedfiles)]
2091 count[:] = [0, len(changedfiles)]
2092 for fname in sorted(changedfiles):
2092 for fname in sorted(changedfiles):
2093 filerevlog = self.file(fname)
2093 filerevlog = self.file(fname)
2094 if not len(filerevlog):
2094 if not len(filerevlog):
2095 raise util.Abort(_("empty or missing revlog for %s")
2095 raise util.Abort(_("empty or missing revlog for %s")
2096 % fname)
2096 % fname)
2097 fstate[0] = fname
2097 fstate[0] = fname
2098 fstate[1] = fnodes.pop(fname, {})
2098 fstate[1] = fnodes.pop(fname, {})
2099
2099
2100 nodelist = prune(filerevlog, fstate[1])
2100 nodelist = prune(filerevlog, fstate[1])
2101 if nodelist:
2101 if nodelist:
2102 count[0] += 1
2102 count[0] += 1
2103 yield bundler.fileheader(fname)
2103 yield bundler.fileheader(fname)
2104 for chunk in filerevlog.group(nodelist, bundler, reorder):
2104 for chunk in filerevlog.group(nodelist, bundler, reorder):
2105 yield chunk
2105 yield chunk
2106
2106
2107 # Signal that no more groups are left.
2107 # Signal that no more groups are left.
2108 yield bundler.close()
2108 yield bundler.close()
2109 progress(_bundling, None)
2109 progress(_bundling, None)
2110
2110
2111 if csets:
2111 if csets:
2112 self.hook('outgoing', node=hex(csets[0]), source=source)
2112 self.hook('outgoing', node=hex(csets[0]), source=source)
2113
2113
2114 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2114 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2115
2115
2116 def changegroup(self, basenodes, source):
2116 def changegroup(self, basenodes, source):
2117 # to avoid a race we use changegroupsubset() (issue1320)
2117 # to avoid a race we use changegroupsubset() (issue1320)
2118 return self.changegroupsubset(basenodes, self.heads(), source)
2118 return self.changegroupsubset(basenodes, self.heads(), source)
2119
2119
2120 @unfilteredmethod
2120 @unfilteredmethod
2121 def _changegroup(self, nodes, source):
2121 def _changegroup(self, nodes, source):
2122 """Compute the changegroup of all nodes that we have that a recipient
2122 """Compute the changegroup of all nodes that we have that a recipient
2123 doesn't. Return a chunkbuffer object whose read() method will return
2123 doesn't. Return a chunkbuffer object whose read() method will return
2124 successive changegroup chunks.
2124 successive changegroup chunks.
2125
2125
2126 This is much easier than the previous function as we can assume that
2126 This is much easier than the previous function as we can assume that
2127 the recipient has any changenode we aren't sending them.
2127 the recipient has any changenode we aren't sending them.
2128
2128
2129 nodes is the set of nodes to send"""
2129 nodes is the set of nodes to send"""
2130
2130
2131 cl = self.changelog
2131 cl = self.changelog
2132 mf = self.manifest
2132 mf = self.manifest
2133 mfs = {}
2133 mfs = {}
2134 changedfiles = set()
2134 changedfiles = set()
2135 fstate = ['']
2135 fstate = ['']
2136 count = [0, 0]
2136 count = [0, 0]
2137
2137
2138 self.hook('preoutgoing', throw=True, source=source)
2138 self.hook('preoutgoing', throw=True, source=source)
2139 self.changegroupinfo(nodes, source)
2139 self.changegroupinfo(nodes, source)
2140
2140
2141 revset = set([cl.rev(n) for n in nodes])
2141 revset = set([cl.rev(n) for n in nodes])
2142
2142
2143 def gennodelst(log):
2143 def gennodelst(log):
2144 ln, llr = log.node, log.linkrev
2144 ln, llr = log.node, log.linkrev
2145 return [ln(r) for r in log if llr(r) in revset]
2145 return [ln(r) for r in log if llr(r) in revset]
2146
2146
2147 progress = self.ui.progress
2147 progress = self.ui.progress
2148 _bundling = _('bundling')
2148 _bundling = _('bundling')
2149 _changesets = _('changesets')
2149 _changesets = _('changesets')
2150 _manifests = _('manifests')
2150 _manifests = _('manifests')
2151 _files = _('files')
2151 _files = _('files')
2152
2152
2153 def lookup(revlog, x):
2153 def lookup(revlog, x):
2154 if revlog == cl:
2154 if revlog == cl:
2155 c = cl.read(x)
2155 c = cl.read(x)
2156 changedfiles.update(c[3])
2156 changedfiles.update(c[3])
2157 mfs.setdefault(c[0], x)
2157 mfs.setdefault(c[0], x)
2158 count[0] += 1
2158 count[0] += 1
2159 progress(_bundling, count[0],
2159 progress(_bundling, count[0],
2160 unit=_changesets, total=count[1])
2160 unit=_changesets, total=count[1])
2161 return x
2161 return x
2162 elif revlog == mf:
2162 elif revlog == mf:
2163 count[0] += 1
2163 count[0] += 1
2164 progress(_bundling, count[0],
2164 progress(_bundling, count[0],
2165 unit=_manifests, total=count[1])
2165 unit=_manifests, total=count[1])
2166 return cl.node(revlog.linkrev(revlog.rev(x)))
2166 return cl.node(revlog.linkrev(revlog.rev(x)))
2167 else:
2167 else:
2168 progress(_bundling, count[0], item=fstate[0],
2168 progress(_bundling, count[0], item=fstate[0],
2169 total=count[1], unit=_files)
2169 total=count[1], unit=_files)
2170 return cl.node(revlog.linkrev(revlog.rev(x)))
2170 return cl.node(revlog.linkrev(revlog.rev(x)))
2171
2171
2172 bundler = changegroup.bundle10(lookup)
2172 bundler = changegroup.bundle10(lookup)
2173 reorder = self.ui.config('bundle', 'reorder', 'auto')
2173 reorder = self.ui.config('bundle', 'reorder', 'auto')
2174 if reorder == 'auto':
2174 if reorder == 'auto':
2175 reorder = None
2175 reorder = None
2176 else:
2176 else:
2177 reorder = util.parsebool(reorder)
2177 reorder = util.parsebool(reorder)
2178
2178
2179 def gengroup():
2179 def gengroup():
2180 '''yield a sequence of changegroup chunks (strings)'''
2180 '''yield a sequence of changegroup chunks (strings)'''
2181 # construct a list of all changed files
2181 # construct a list of all changed files
2182
2182
2183 count[:] = [0, len(nodes)]
2183 count[:] = [0, len(nodes)]
2184 for chunk in cl.group(nodes, bundler, reorder=reorder):
2184 for chunk in cl.group(nodes, bundler, reorder=reorder):
2185 yield chunk
2185 yield chunk
2186 progress(_bundling, None)
2186 progress(_bundling, None)
2187
2187
2188 count[:] = [0, len(mfs)]
2188 count[:] = [0, len(mfs)]
2189 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2189 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
2190 yield chunk
2190 yield chunk
2191 progress(_bundling, None)
2191 progress(_bundling, None)
2192
2192
2193 count[:] = [0, len(changedfiles)]
2193 count[:] = [0, len(changedfiles)]
2194 for fname in sorted(changedfiles):
2194 for fname in sorted(changedfiles):
2195 filerevlog = self.file(fname)
2195 filerevlog = self.file(fname)
2196 if not len(filerevlog):
2196 if not len(filerevlog):
2197 raise util.Abort(_("empty or missing revlog for %s")
2197 raise util.Abort(_("empty or missing revlog for %s")
2198 % fname)
2198 % fname)
2199 fstate[0] = fname
2199 fstate[0] = fname
2200 nodelist = gennodelst(filerevlog)
2200 nodelist = gennodelst(filerevlog)
2201 if nodelist:
2201 if nodelist:
2202 count[0] += 1
2202 count[0] += 1
2203 yield bundler.fileheader(fname)
2203 yield bundler.fileheader(fname)
2204 for chunk in filerevlog.group(nodelist, bundler, reorder):
2204 for chunk in filerevlog.group(nodelist, bundler, reorder):
2205 yield chunk
2205 yield chunk
2206 yield bundler.close()
2206 yield bundler.close()
2207 progress(_bundling, None)
2207 progress(_bundling, None)
2208
2208
2209 if nodes:
2209 if nodes:
2210 self.hook('outgoing', node=hex(nodes[0]), source=source)
2210 self.hook('outgoing', node=hex(nodes[0]), source=source)
2211
2211
2212 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2212 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
2213
2213
2214 @unfilteredmethod
2214 @unfilteredmethod
2215 def addchangegroup(self, source, srctype, url, emptyok=False):
2215 def addchangegroup(self, source, srctype, url, emptyok=False):
2216 """Add the changegroup returned by source.read() to this repo.
2216 """Add the changegroup returned by source.read() to this repo.
2217 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2217 srctype is a string like 'push', 'pull', or 'unbundle'. url is
2218 the URL of the repo where this changegroup is coming from.
2218 the URL of the repo where this changegroup is coming from.
2219
2219
2220 Return an integer summarizing the change to this repo:
2220 Return an integer summarizing the change to this repo:
2221 - nothing changed or no source: 0
2221 - nothing changed or no source: 0
2222 - more heads than before: 1+added heads (2..n)
2222 - more heads than before: 1+added heads (2..n)
2223 - fewer heads than before: -1-removed heads (-2..-n)
2223 - fewer heads than before: -1-removed heads (-2..-n)
2224 - number of heads stays the same: 1
2224 - number of heads stays the same: 1
2225 """
2225 """
2226 def csmap(x):
2226 def csmap(x):
2227 self.ui.debug("add changeset %s\n" % short(x))
2227 self.ui.debug("add changeset %s\n" % short(x))
2228 return len(cl)
2228 return len(cl)
2229
2229
2230 def revmap(x):
2230 def revmap(x):
2231 return cl.rev(x)
2231 return cl.rev(x)
2232
2232
2233 if not source:
2233 if not source:
2234 return 0
2234 return 0
2235
2235
2236 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2236 self.hook('prechangegroup', throw=True, source=srctype, url=url)
2237
2237
2238 changesets = files = revisions = 0
2238 changesets = files = revisions = 0
2239 efiles = set()
2239 efiles = set()
2240
2240
2241 # write changelog data to temp files so concurrent readers will not see
2241 # write changelog data to temp files so concurrent readers will not see
2242 # inconsistent view
2242 # inconsistent view
2243 cl = self.changelog
2243 cl = self.changelog
2244 cl.delayupdate()
2244 cl.delayupdate()
2245 oldheads = cl.heads()
2245 oldheads = cl.heads()
2246
2246
2247 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2247 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
2248 try:
2248 try:
2249 trp = weakref.proxy(tr)
2249 trp = weakref.proxy(tr)
2250 # pull off the changeset group
2250 # pull off the changeset group
2251 self.ui.status(_("adding changesets\n"))
2251 self.ui.status(_("adding changesets\n"))
2252 clstart = len(cl)
2252 clstart = len(cl)
2253 class prog(object):
2253 class prog(object):
2254 step = _('changesets')
2254 step = _('changesets')
2255 count = 1
2255 count = 1
2256 ui = self.ui
2256 ui = self.ui
2257 total = None
2257 total = None
2258 def __call__(self):
2258 def __call__(self):
2259 self.ui.progress(self.step, self.count, unit=_('chunks'),
2259 self.ui.progress(self.step, self.count, unit=_('chunks'),
2260 total=self.total)
2260 total=self.total)
2261 self.count += 1
2261 self.count += 1
2262 pr = prog()
2262 pr = prog()
2263 source.callback = pr
2263 source.callback = pr
2264
2264
2265 source.changelogheader()
2265 source.changelogheader()
2266 srccontent = cl.addgroup(source, csmap, trp)
2266 srccontent = cl.addgroup(source, csmap, trp)
2267 if not (srccontent or emptyok):
2267 if not (srccontent or emptyok):
2268 raise util.Abort(_("received changelog group is empty"))
2268 raise util.Abort(_("received changelog group is empty"))
2269 clend = len(cl)
2269 clend = len(cl)
2270 changesets = clend - clstart
2270 changesets = clend - clstart
2271 for c in xrange(clstart, clend):
2271 for c in xrange(clstart, clend):
2272 efiles.update(self[c].files())
2272 efiles.update(self[c].files())
2273 efiles = len(efiles)
2273 efiles = len(efiles)
2274 self.ui.progress(_('changesets'), None)
2274 self.ui.progress(_('changesets'), None)
2275
2275
2276 # pull off the manifest group
2276 # pull off the manifest group
2277 self.ui.status(_("adding manifests\n"))
2277 self.ui.status(_("adding manifests\n"))
2278 pr.step = _('manifests')
2278 pr.step = _('manifests')
2279 pr.count = 1
2279 pr.count = 1
2280 pr.total = changesets # manifests <= changesets
2280 pr.total = changesets # manifests <= changesets
2281 # no need to check for empty manifest group here:
2281 # no need to check for empty manifest group here:
2282 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2282 # if the result of the merge of 1 and 2 is the same in 3 and 4,
2283 # no new manifest will be created and the manifest group will
2283 # no new manifest will be created and the manifest group will
2284 # be empty during the pull
2284 # be empty during the pull
2285 source.manifestheader()
2285 source.manifestheader()
2286 self.manifest.addgroup(source, revmap, trp)
2286 self.manifest.addgroup(source, revmap, trp)
2287 self.ui.progress(_('manifests'), None)
2287 self.ui.progress(_('manifests'), None)
2288
2288
2289 needfiles = {}
2289 needfiles = {}
2290 if self.ui.configbool('server', 'validate', default=False):
2290 if self.ui.configbool('server', 'validate', default=False):
2291 # validate incoming csets have their manifests
2291 # validate incoming csets have their manifests
2292 for cset in xrange(clstart, clend):
2292 for cset in xrange(clstart, clend):
2293 mfest = self.changelog.read(self.changelog.node(cset))[0]
2293 mfest = self.changelog.read(self.changelog.node(cset))[0]
2294 mfest = self.manifest.readdelta(mfest)
2294 mfest = self.manifest.readdelta(mfest)
2295 # store file nodes we must see
2295 # store file nodes we must see
2296 for f, n in mfest.iteritems():
2296 for f, n in mfest.iteritems():
2297 needfiles.setdefault(f, set()).add(n)
2297 needfiles.setdefault(f, set()).add(n)
2298
2298
2299 # process the files
2299 # process the files
2300 self.ui.status(_("adding file changes\n"))
2300 self.ui.status(_("adding file changes\n"))
2301 pr.step = _('files')
2301 pr.step = _('files')
2302 pr.count = 1
2302 pr.count = 1
2303 pr.total = efiles
2303 pr.total = efiles
2304 source.callback = None
2304 source.callback = None
2305
2305
2306 while True:
2306 while True:
2307 chunkdata = source.filelogheader()
2307 chunkdata = source.filelogheader()
2308 if not chunkdata:
2308 if not chunkdata:
2309 break
2309 break
2310 f = chunkdata["filename"]
2310 f = chunkdata["filename"]
2311 self.ui.debug("adding %s revisions\n" % f)
2311 self.ui.debug("adding %s revisions\n" % f)
2312 pr()
2312 pr()
2313 fl = self.file(f)
2313 fl = self.file(f)
2314 o = len(fl)
2314 o = len(fl)
2315 if not fl.addgroup(source, revmap, trp):
2315 if not fl.addgroup(source, revmap, trp):
2316 raise util.Abort(_("received file revlog group is empty"))
2316 raise util.Abort(_("received file revlog group is empty"))
2317 revisions += len(fl) - o
2317 revisions += len(fl) - o
2318 files += 1
2318 files += 1
2319 if f in needfiles:
2319 if f in needfiles:
2320 needs = needfiles[f]
2320 needs = needfiles[f]
2321 for new in xrange(o, len(fl)):
2321 for new in xrange(o, len(fl)):
2322 n = fl.node(new)
2322 n = fl.node(new)
2323 if n in needs:
2323 if n in needs:
2324 needs.remove(n)
2324 needs.remove(n)
2325 if not needs:
2325 if not needs:
2326 del needfiles[f]
2326 del needfiles[f]
2327 self.ui.progress(_('files'), None)
2327 self.ui.progress(_('files'), None)
2328
2328
2329 for f, needs in needfiles.iteritems():
2329 for f, needs in needfiles.iteritems():
2330 fl = self.file(f)
2330 fl = self.file(f)
2331 for n in needs:
2331 for n in needs:
2332 try:
2332 try:
2333 fl.rev(n)
2333 fl.rev(n)
2334 except error.LookupError:
2334 except error.LookupError:
2335 raise util.Abort(
2335 raise util.Abort(
2336 _('missing file data for %s:%s - run hg verify') %
2336 _('missing file data for %s:%s - run hg verify') %
2337 (f, hex(n)))
2337 (f, hex(n)))
2338
2338
2339 dh = 0
2339 dh = 0
2340 if oldheads:
2340 if oldheads:
2341 heads = cl.heads()
2341 heads = cl.heads()
2342 dh = len(heads) - len(oldheads)
2342 dh = len(heads) - len(oldheads)
2343 for h in heads:
2343 for h in heads:
2344 if h not in oldheads and self[h].closesbranch():
2344 if h not in oldheads and self[h].closesbranch():
2345 dh -= 1
2345 dh -= 1
2346 htext = ""
2346 htext = ""
2347 if dh:
2347 if dh:
2348 htext = _(" (%+d heads)") % dh
2348 htext = _(" (%+d heads)") % dh
2349
2349
2350 self.ui.status(_("added %d changesets"
2350 self.ui.status(_("added %d changesets"
2351 " with %d changes to %d files%s\n")
2351 " with %d changes to %d files%s\n")
2352 % (changesets, revisions, files, htext))
2352 % (changesets, revisions, files, htext))
2353 self.invalidatevolatilesets()
2353 self.invalidatevolatilesets()
2354
2354
2355 if changesets > 0:
2355 if changesets > 0:
2356 p = lambda: cl.writepending() and self.root or ""
2356 p = lambda: cl.writepending() and self.root or ""
2357 self.hook('pretxnchangegroup', throw=True,
2357 self.hook('pretxnchangegroup', throw=True,
2358 node=hex(cl.node(clstart)), source=srctype,
2358 node=hex(cl.node(clstart)), source=srctype,
2359 url=url, pending=p)
2359 url=url, pending=p)
2360
2360
2361 added = [cl.node(r) for r in xrange(clstart, clend)]
2361 added = [cl.node(r) for r in xrange(clstart, clend)]
2362 publishing = self.ui.configbool('phases', 'publish', True)
2362 publishing = self.ui.configbool('phases', 'publish', True)
2363 if srctype == 'push':
2363 if srctype == 'push':
2364 # Old server can not push the boundary themself.
2364 # Old server can not push the boundary themself.
2365 # New server won't push the boundary if changeset already
2365 # New server won't push the boundary if changeset already
2366 # existed locally as secrete
2366 # existed locally as secrete
2367 #
2367 #
2368 # We should not use added here but the list of all change in
2368 # We should not use added here but the list of all change in
2369 # the bundle
2369 # the bundle
2370 if publishing:
2370 if publishing:
2371 phases.advanceboundary(self, phases.public, srccontent)
2371 phases.advanceboundary(self, phases.public, srccontent)
2372 else:
2372 else:
2373 phases.advanceboundary(self, phases.draft, srccontent)
2373 phases.advanceboundary(self, phases.draft, srccontent)
2374 phases.retractboundary(self, phases.draft, added)
2374 phases.retractboundary(self, phases.draft, added)
2375 elif srctype != 'strip':
2375 elif srctype != 'strip':
2376 # publishing only alter behavior during push
2376 # publishing only alter behavior during push
2377 #
2377 #
2378 # strip should not touch boundary at all
2378 # strip should not touch boundary at all
2379 phases.retractboundary(self, phases.draft, added)
2379 phases.retractboundary(self, phases.draft, added)
2380
2380
2381 # make changelog see real files again
2381 # make changelog see real files again
2382 cl.finalize(trp)
2382 cl.finalize(trp)
2383
2383
2384 tr.close()
2384 tr.close()
2385
2385
2386 if changesets > 0:
2386 if changesets > 0:
2387 if srctype != 'strip':
2387 if srctype != 'strip':
2388 # During strip, branchcache is invalid but coming call to
2388 # During strip, branchcache is invalid but coming call to
2389 # `destroyed` will repair it.
2389 # `destroyed` will repair it.
2390 # In other case we can safely update cache on disk.
2390 # In other case we can safely update cache on disk.
2391 branchmap.updatecache(self)
2391 branchmap.updatecache(self.filtered('served'))
2392 def runhooks():
2392 def runhooks():
2393 # forcefully update the on-disk branch cache
2393 # forcefully update the on-disk branch cache
2394 self.ui.debug("updating the branch cache\n")
2394 self.ui.debug("updating the branch cache\n")
2395 self.hook("changegroup", node=hex(cl.node(clstart)),
2395 self.hook("changegroup", node=hex(cl.node(clstart)),
2396 source=srctype, url=url)
2396 source=srctype, url=url)
2397
2397
2398 for n in added:
2398 for n in added:
2399 self.hook("incoming", node=hex(n), source=srctype,
2399 self.hook("incoming", node=hex(n), source=srctype,
2400 url=url)
2400 url=url)
2401 self._afterlock(runhooks)
2401 self._afterlock(runhooks)
2402
2402
2403 finally:
2403 finally:
2404 tr.release()
2404 tr.release()
2405 # never return 0 here:
2405 # never return 0 here:
2406 if dh < 0:
2406 if dh < 0:
2407 return dh - 1
2407 return dh - 1
2408 else:
2408 else:
2409 return dh + 1
2409 return dh + 1
2410
2410
2411 def stream_in(self, remote, requirements):
2411 def stream_in(self, remote, requirements):
2412 lock = self.lock()
2412 lock = self.lock()
2413 try:
2413 try:
2414 # Save remote branchmap. We will use it later
2414 # Save remote branchmap. We will use it later
2415 # to speed up branchcache creation
2415 # to speed up branchcache creation
2416 rbranchmap = None
2416 rbranchmap = None
2417 if remote.capable("branchmap"):
2417 if remote.capable("branchmap"):
2418 rbranchmap = remote.branchmap()
2418 rbranchmap = remote.branchmap()
2419
2419
2420 fp = remote.stream_out()
2420 fp = remote.stream_out()
2421 l = fp.readline()
2421 l = fp.readline()
2422 try:
2422 try:
2423 resp = int(l)
2423 resp = int(l)
2424 except ValueError:
2424 except ValueError:
2425 raise error.ResponseError(
2425 raise error.ResponseError(
2426 _('unexpected response from remote server:'), l)
2426 _('unexpected response from remote server:'), l)
2427 if resp == 1:
2427 if resp == 1:
2428 raise util.Abort(_('operation forbidden by server'))
2428 raise util.Abort(_('operation forbidden by server'))
2429 elif resp == 2:
2429 elif resp == 2:
2430 raise util.Abort(_('locking the remote repository failed'))
2430 raise util.Abort(_('locking the remote repository failed'))
2431 elif resp != 0:
2431 elif resp != 0:
2432 raise util.Abort(_('the server sent an unknown error code'))
2432 raise util.Abort(_('the server sent an unknown error code'))
2433 self.ui.status(_('streaming all changes\n'))
2433 self.ui.status(_('streaming all changes\n'))
2434 l = fp.readline()
2434 l = fp.readline()
2435 try:
2435 try:
2436 total_files, total_bytes = map(int, l.split(' ', 1))
2436 total_files, total_bytes = map(int, l.split(' ', 1))
2437 except (ValueError, TypeError):
2437 except (ValueError, TypeError):
2438 raise error.ResponseError(
2438 raise error.ResponseError(
2439 _('unexpected response from remote server:'), l)
2439 _('unexpected response from remote server:'), l)
2440 self.ui.status(_('%d files to transfer, %s of data\n') %
2440 self.ui.status(_('%d files to transfer, %s of data\n') %
2441 (total_files, util.bytecount(total_bytes)))
2441 (total_files, util.bytecount(total_bytes)))
2442 handled_bytes = 0
2442 handled_bytes = 0
2443 self.ui.progress(_('clone'), 0, total=total_bytes)
2443 self.ui.progress(_('clone'), 0, total=total_bytes)
2444 start = time.time()
2444 start = time.time()
2445 for i in xrange(total_files):
2445 for i in xrange(total_files):
2446 # XXX doesn't support '\n' or '\r' in filenames
2446 # XXX doesn't support '\n' or '\r' in filenames
2447 l = fp.readline()
2447 l = fp.readline()
2448 try:
2448 try:
2449 name, size = l.split('\0', 1)
2449 name, size = l.split('\0', 1)
2450 size = int(size)
2450 size = int(size)
2451 except (ValueError, TypeError):
2451 except (ValueError, TypeError):
2452 raise error.ResponseError(
2452 raise error.ResponseError(
2453 _('unexpected response from remote server:'), l)
2453 _('unexpected response from remote server:'), l)
2454 if self.ui.debugflag:
2454 if self.ui.debugflag:
2455 self.ui.debug('adding %s (%s)\n' %
2455 self.ui.debug('adding %s (%s)\n' %
2456 (name, util.bytecount(size)))
2456 (name, util.bytecount(size)))
2457 # for backwards compat, name was partially encoded
2457 # for backwards compat, name was partially encoded
2458 ofp = self.sopener(store.decodedir(name), 'w')
2458 ofp = self.sopener(store.decodedir(name), 'w')
2459 for chunk in util.filechunkiter(fp, limit=size):
2459 for chunk in util.filechunkiter(fp, limit=size):
2460 handled_bytes += len(chunk)
2460 handled_bytes += len(chunk)
2461 self.ui.progress(_('clone'), handled_bytes,
2461 self.ui.progress(_('clone'), handled_bytes,
2462 total=total_bytes)
2462 total=total_bytes)
2463 ofp.write(chunk)
2463 ofp.write(chunk)
2464 ofp.close()
2464 ofp.close()
2465 elapsed = time.time() - start
2465 elapsed = time.time() - start
2466 if elapsed <= 0:
2466 if elapsed <= 0:
2467 elapsed = 0.001
2467 elapsed = 0.001
2468 self.ui.progress(_('clone'), None)
2468 self.ui.progress(_('clone'), None)
2469 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2469 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
2470 (util.bytecount(total_bytes), elapsed,
2470 (util.bytecount(total_bytes), elapsed,
2471 util.bytecount(total_bytes / elapsed)))
2471 util.bytecount(total_bytes / elapsed)))
2472
2472
2473 # new requirements = old non-format requirements +
2473 # new requirements = old non-format requirements +
2474 # new format-related
2474 # new format-related
2475 # requirements from the streamed-in repository
2475 # requirements from the streamed-in repository
2476 requirements.update(set(self.requirements) - self.supportedformats)
2476 requirements.update(set(self.requirements) - self.supportedformats)
2477 self._applyrequirements(requirements)
2477 self._applyrequirements(requirements)
2478 self._writerequirements()
2478 self._writerequirements()
2479
2479
2480 if rbranchmap:
2480 if rbranchmap:
2481 rbheads = []
2481 rbheads = []
2482 for bheads in rbranchmap.itervalues():
2482 for bheads in rbranchmap.itervalues():
2483 rbheads.extend(bheads)
2483 rbheads.extend(bheads)
2484
2484
2485 if rbheads:
2485 if rbheads:
2486 rtiprev = max((int(self.changelog.rev(node))
2486 rtiprev = max((int(self.changelog.rev(node))
2487 for node in rbheads))
2487 for node in rbheads))
2488 cache = branchmap.branchcache(rbranchmap,
2488 cache = branchmap.branchcache(rbranchmap,
2489 self[rtiprev].node(),
2489 self[rtiprev].node(),
2490 rtiprev)
2490 rtiprev)
2491 self._branchcaches[None] = cache
2491 self._branchcaches[None] = cache
2492 cache.write(self.unfiltered())
2492 cache.write(self.unfiltered())
2493 self.invalidate()
2493 self.invalidate()
2494 return len(self.heads()) + 1
2494 return len(self.heads()) + 1
2495 finally:
2495 finally:
2496 lock.release()
2496 lock.release()
2497
2497
2498 def clone(self, remote, heads=[], stream=False):
2498 def clone(self, remote, heads=[], stream=False):
2499 '''clone remote repository.
2499 '''clone remote repository.
2500
2500
2501 keyword arguments:
2501 keyword arguments:
2502 heads: list of revs to clone (forces use of pull)
2502 heads: list of revs to clone (forces use of pull)
2503 stream: use streaming clone if possible'''
2503 stream: use streaming clone if possible'''
2504
2504
2505 # now, all clients that can request uncompressed clones can
2505 # now, all clients that can request uncompressed clones can
2506 # read repo formats supported by all servers that can serve
2506 # read repo formats supported by all servers that can serve
2507 # them.
2507 # them.
2508
2508
2509 # if revlog format changes, client will have to check version
2509 # if revlog format changes, client will have to check version
2510 # and format flags on "stream" capability, and use
2510 # and format flags on "stream" capability, and use
2511 # uncompressed only if compatible.
2511 # uncompressed only if compatible.
2512
2512
2513 if not stream:
2513 if not stream:
2514 # if the server explicitly prefers to stream (for fast LANs)
2514 # if the server explicitly prefers to stream (for fast LANs)
2515 stream = remote.capable('stream-preferred')
2515 stream = remote.capable('stream-preferred')
2516
2516
2517 if stream and not heads:
2517 if stream and not heads:
2518 # 'stream' means remote revlog format is revlogv1 only
2518 # 'stream' means remote revlog format is revlogv1 only
2519 if remote.capable('stream'):
2519 if remote.capable('stream'):
2520 return self.stream_in(remote, set(('revlogv1',)))
2520 return self.stream_in(remote, set(('revlogv1',)))
2521 # otherwise, 'streamreqs' contains the remote revlog format
2521 # otherwise, 'streamreqs' contains the remote revlog format
2522 streamreqs = remote.capable('streamreqs')
2522 streamreqs = remote.capable('streamreqs')
2523 if streamreqs:
2523 if streamreqs:
2524 streamreqs = set(streamreqs.split(','))
2524 streamreqs = set(streamreqs.split(','))
2525 # if we support it, stream in and adjust our requirements
2525 # if we support it, stream in and adjust our requirements
2526 if not streamreqs - self.supportedformats:
2526 if not streamreqs - self.supportedformats:
2527 return self.stream_in(remote, streamreqs)
2527 return self.stream_in(remote, streamreqs)
2528 return self.pull(remote, heads)
2528 return self.pull(remote, heads)
2529
2529
2530 def pushkey(self, namespace, key, old, new):
2530 def pushkey(self, namespace, key, old, new):
2531 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2531 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
2532 old=old, new=new)
2532 old=old, new=new)
2533 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2533 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
2534 ret = pushkey.push(self, namespace, key, old, new)
2534 ret = pushkey.push(self, namespace, key, old, new)
2535 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2535 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
2536 ret=ret)
2536 ret=ret)
2537 return ret
2537 return ret
2538
2538
2539 def listkeys(self, namespace):
2539 def listkeys(self, namespace):
2540 self.hook('prelistkeys', throw=True, namespace=namespace)
2540 self.hook('prelistkeys', throw=True, namespace=namespace)
2541 self.ui.debug('listing keys for "%s"\n' % namespace)
2541 self.ui.debug('listing keys for "%s"\n' % namespace)
2542 values = pushkey.list(self, namespace)
2542 values = pushkey.list(self, namespace)
2543 self.hook('listkeys', namespace=namespace, values=values)
2543 self.hook('listkeys', namespace=namespace, values=values)
2544 return values
2544 return values
2545
2545
2546 def debugwireargs(self, one, two, three=None, four=None, five=None):
2546 def debugwireargs(self, one, two, three=None, four=None, five=None):
2547 '''used to test argument passing over the wire'''
2547 '''used to test argument passing over the wire'''
2548 return "%s %s %s %s %s" % (one, two, three, four, five)
2548 return "%s %s %s %s %s" % (one, two, three, four, five)
2549
2549
2550 def savecommitmessage(self, text):
2550 def savecommitmessage(self, text):
2551 fp = self.opener('last-message.txt', 'wb')
2551 fp = self.opener('last-message.txt', 'wb')
2552 try:
2552 try:
2553 fp.write(text)
2553 fp.write(text)
2554 finally:
2554 finally:
2555 fp.close()
2555 fp.close()
2556 return self.pathto(fp.name[len(self.root) + 1:])
2556 return self.pathto(fp.name[len(self.root) + 1:])
2557
2557
2558 # used to avoid circular references so destructors work
2558 # used to avoid circular references so destructors work
2559 def aftertrans(files):
2559 def aftertrans(files):
2560 renamefiles = [tuple(t) for t in files]
2560 renamefiles = [tuple(t) for t in files]
2561 def a():
2561 def a():
2562 for src, dest in renamefiles:
2562 for src, dest in renamefiles:
2563 try:
2563 try:
2564 util.rename(src, dest)
2564 util.rename(src, dest)
2565 except OSError: # journal file does not yet exist
2565 except OSError: # journal file does not yet exist
2566 pass
2566 pass
2567 return a
2567 return a
2568
2568
2569 def undoname(fn):
2569 def undoname(fn):
2570 base, name = os.path.split(fn)
2570 base, name = os.path.split(fn)
2571 assert name.startswith('journal')
2571 assert name.startswith('journal')
2572 return os.path.join(base, name.replace('journal', 'undo', 1))
2572 return os.path.join(base, name.replace('journal', 'undo', 1))
2573
2573
2574 def instance(ui, path, create):
2574 def instance(ui, path, create):
2575 return localrepository(ui, util.urllocalpath(path), create)
2575 return localrepository(ui, util.urllocalpath(path), create)
2576
2576
2577 def islocal(path):
2577 def islocal(path):
2578 return True
2578 return True
@@ -1,1155 +1,1153 b''
1 $ cat <<EOF >> $HGRCPATH
1 $ cat <<EOF >> $HGRCPATH
2 > [extensions]
2 > [extensions]
3 > keyword =
3 > keyword =
4 > mq =
4 > mq =
5 > notify =
5 > notify =
6 > record =
6 > record =
7 > transplant =
7 > transplant =
8 > [ui]
8 > [ui]
9 > interactive = true
9 > interactive = true
10 > EOF
10 > EOF
11
11
12 hide outer repo
12 hide outer repo
13 $ hg init
13 $ hg init
14
14
15 Run kwdemo before [keyword] files are set up
15 Run kwdemo before [keyword] files are set up
16 as it would succeed without uisetup otherwise
16 as it would succeed without uisetup otherwise
17
17
18 $ hg --quiet kwdemo
18 $ hg --quiet kwdemo
19 [extensions]
19 [extensions]
20 keyword =
20 keyword =
21 [keyword]
21 [keyword]
22 demo.txt =
22 demo.txt =
23 [keywordset]
23 [keywordset]
24 svn = False
24 svn = False
25 [keywordmaps]
25 [keywordmaps]
26 Author = {author|user}
26 Author = {author|user}
27 Date = {date|utcdate}
27 Date = {date|utcdate}
28 Header = {root}/{file},v {node|short} {date|utcdate} {author|user}
28 Header = {root}/{file},v {node|short} {date|utcdate} {author|user}
29 Id = {file|basename},v {node|short} {date|utcdate} {author|user}
29 Id = {file|basename},v {node|short} {date|utcdate} {author|user}
30 RCSFile = {file|basename},v
30 RCSFile = {file|basename},v
31 RCSfile = {file|basename},v
31 RCSfile = {file|basename},v
32 Revision = {node|short}
32 Revision = {node|short}
33 Source = {root}/{file},v
33 Source = {root}/{file},v
34 $Author: test $
34 $Author: test $
35 $Date: ????/??/?? ??:??:?? $ (glob)
35 $Date: ????/??/?? ??:??:?? $ (glob)
36 $Header: */demo.txt,v ???????????? ????/??/?? ??:??:?? test $ (glob)
36 $Header: */demo.txt,v ???????????? ????/??/?? ??:??:?? test $ (glob)
37 $Id: demo.txt,v ???????????? ????/??/?? ??:??:?? test $ (glob)
37 $Id: demo.txt,v ???????????? ????/??/?? ??:??:?? test $ (glob)
38 $RCSFile: demo.txt,v $
38 $RCSFile: demo.txt,v $
39 $RCSfile: demo.txt,v $
39 $RCSfile: demo.txt,v $
40 $Revision: ???????????? $ (glob)
40 $Revision: ???????????? $ (glob)
41 $Source: */demo.txt,v $ (glob)
41 $Source: */demo.txt,v $ (glob)
42
42
43 $ hg --quiet kwdemo "Branch = {branches}"
43 $ hg --quiet kwdemo "Branch = {branches}"
44 [extensions]
44 [extensions]
45 keyword =
45 keyword =
46 [keyword]
46 [keyword]
47 demo.txt =
47 demo.txt =
48 [keywordset]
48 [keywordset]
49 svn = False
49 svn = False
50 [keywordmaps]
50 [keywordmaps]
51 Branch = {branches}
51 Branch = {branches}
52 $Branch: demobranch $
52 $Branch: demobranch $
53
53
54 $ cat <<EOF >> $HGRCPATH
54 $ cat <<EOF >> $HGRCPATH
55 > [keyword]
55 > [keyword]
56 > ** =
56 > ** =
57 > b = ignore
57 > b = ignore
58 > i = ignore
58 > i = ignore
59 > [hooks]
59 > [hooks]
60 > EOF
60 > EOF
61 $ cp $HGRCPATH $HGRCPATH.nohooks
61 $ cp $HGRCPATH $HGRCPATH.nohooks
62 > cat <<EOF >> $HGRCPATH
62 > cat <<EOF >> $HGRCPATH
63 > commit=
63 > commit=
64 > commit.test=cp a hooktest
64 > commit.test=cp a hooktest
65 > EOF
65 > EOF
66
66
67 $ hg init Test-bndl
67 $ hg init Test-bndl
68 $ cd Test-bndl
68 $ cd Test-bndl
69
69
70 kwshrink should exit silently in empty/invalid repo
70 kwshrink should exit silently in empty/invalid repo
71
71
72 $ hg kwshrink
72 $ hg kwshrink
73
73
74 Symlinks cannot be created on Windows.
74 Symlinks cannot be created on Windows.
75 A bundle to test this was made with:
75 A bundle to test this was made with:
76 hg init t
76 hg init t
77 cd t
77 cd t
78 echo a > a
78 echo a > a
79 ln -s a sym
79 ln -s a sym
80 hg add sym
80 hg add sym
81 hg ci -m addsym -u mercurial
81 hg ci -m addsym -u mercurial
82 hg bundle --base null ../test-keyword.hg
82 hg bundle --base null ../test-keyword.hg
83
83
84 $ hg pull -u "$TESTDIR"/bundles/test-keyword.hg
84 $ hg pull -u "$TESTDIR"/bundles/test-keyword.hg
85 pulling from *test-keyword.hg (glob)
85 pulling from *test-keyword.hg (glob)
86 requesting all changes
86 requesting all changes
87 adding changesets
87 adding changesets
88 adding manifests
88 adding manifests
89 adding file changes
89 adding file changes
90 added 1 changesets with 1 changes to 1 files
90 added 1 changesets with 1 changes to 1 files
91 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
91 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
92
92
93 $ echo 'expand $Id$' > a
93 $ echo 'expand $Id$' > a
94 $ echo 'do not process $Id:' >> a
94 $ echo 'do not process $Id:' >> a
95 $ echo 'xxx $' >> a
95 $ echo 'xxx $' >> a
96 $ echo 'ignore $Id$' > b
96 $ echo 'ignore $Id$' > b
97
97
98 Output files as they were created
98 Output files as they were created
99
99
100 $ cat a b
100 $ cat a b
101 expand $Id$
101 expand $Id$
102 do not process $Id:
102 do not process $Id:
103 xxx $
103 xxx $
104 ignore $Id$
104 ignore $Id$
105
105
106 no kwfiles
106 no kwfiles
107
107
108 $ hg kwfiles
108 $ hg kwfiles
109
109
110 untracked candidates
110 untracked candidates
111
111
112 $ hg -v kwfiles --unknown
112 $ hg -v kwfiles --unknown
113 k a
113 k a
114
114
115 Add files and check status
115 Add files and check status
116
116
117 $ hg addremove
117 $ hg addremove
118 adding a
118 adding a
119 adding b
119 adding b
120 $ hg status
120 $ hg status
121 A a
121 A a
122 A b
122 A b
123
123
124
124
125 Default keyword expansion including commit hook
125 Default keyword expansion including commit hook
126 Interrupted commit should not change state or run commit hook
126 Interrupted commit should not change state or run commit hook
127
127
128 $ hg --debug commit
128 $ hg --debug commit
129 abort: empty commit message
129 abort: empty commit message
130 [255]
130 [255]
131 $ hg status
131 $ hg status
132 A a
132 A a
133 A b
133 A b
134
134
135 Commit with several checks
135 Commit with several checks
136
136
137 $ hg --debug commit -mabsym -u 'User Name <user@example.com>'
137 $ hg --debug commit -mabsym -u 'User Name <user@example.com>'
138 a
138 a
139 b
139 b
140 overwriting a expanding keywords
140 overwriting a expanding keywords
141 running hook commit.test: cp a hooktest
141 running hook commit.test: cp a hooktest
142 committed changeset 1:ef63ca68695bc9495032c6fda1350c71e6d256e9
142 committed changeset 1:ef63ca68695bc9495032c6fda1350c71e6d256e9
143 $ hg status
143 $ hg status
144 ? hooktest
144 ? hooktest
145 $ hg debugrebuildstate
145 $ hg debugrebuildstate
146 $ hg --quiet identify
146 $ hg --quiet identify
147 ef63ca68695b
147 ef63ca68695b
148
148
149 cat files in working directory with keywords expanded
149 cat files in working directory with keywords expanded
150
150
151 $ cat a b
151 $ cat a b
152 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
152 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
153 do not process $Id:
153 do not process $Id:
154 xxx $
154 xxx $
155 ignore $Id$
155 ignore $Id$
156
156
157 hg cat files and symlink, no expansion
157 hg cat files and symlink, no expansion
158
158
159 $ hg cat sym a b && echo
159 $ hg cat sym a b && echo
160 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
160 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
161 do not process $Id:
161 do not process $Id:
162 xxx $
162 xxx $
163 ignore $Id$
163 ignore $Id$
164 a
164 a
165
165
166 $ diff a hooktest
166 $ diff a hooktest
167
167
168 $ cp $HGRCPATH.nohooks $HGRCPATH
168 $ cp $HGRCPATH.nohooks $HGRCPATH
169 $ rm hooktest
169 $ rm hooktest
170
170
171 hg status of kw-ignored binary file starting with '\1\n'
171 hg status of kw-ignored binary file starting with '\1\n'
172
172
173 >>> open("i", "wb").write("\1\nfoo")
173 >>> open("i", "wb").write("\1\nfoo")
174 $ hg -q commit -Am metasep i
174 $ hg -q commit -Am metasep i
175 $ hg status
175 $ hg status
176 >>> open("i", "wb").write("\1\nbar")
176 >>> open("i", "wb").write("\1\nbar")
177 $ hg status
177 $ hg status
178 M i
178 M i
179 $ hg -q commit -m "modify metasep" i
179 $ hg -q commit -m "modify metasep" i
180 $ hg status --rev 2:3
180 $ hg status --rev 2:3
181 M i
181 M i
182 $ touch empty
182 $ touch empty
183 $ hg -q commit -A -m "another file"
183 $ hg -q commit -A -m "another file"
184 $ hg status -A --rev 3:4 i
184 $ hg status -A --rev 3:4 i
185 C i
185 C i
186
186
187 $ hg -q strip -n 2
187 $ hg -q strip -n 2
188
188
189 Test hook execution
189 Test hook execution
190
190
191 bundle
191 bundle
192
192
193 $ hg bundle --base null ../kw.hg
193 $ hg bundle --base null ../kw.hg
194 2 changesets found
194 2 changesets found
195 $ cd ..
195 $ cd ..
196 $ hg init Test
196 $ hg init Test
197 $ cd Test
197 $ cd Test
198
198
199 Notify on pull to check whether keywords stay as is in email
199 Notify on pull to check whether keywords stay as is in email
200 ie. if patch.diff wrapper acts as it should
200 ie. if patch.diff wrapper acts as it should
201
201
202 $ cat <<EOF >> $HGRCPATH
202 $ cat <<EOF >> $HGRCPATH
203 > [hooks]
203 > [hooks]
204 > incoming.notify = python:hgext.notify.hook
204 > incoming.notify = python:hgext.notify.hook
205 > [notify]
205 > [notify]
206 > sources = pull
206 > sources = pull
207 > diffstat = False
207 > diffstat = False
208 > maxsubject = 15
208 > maxsubject = 15
209 > [reposubs]
209 > [reposubs]
210 > * = Test
210 > * = Test
211 > EOF
211 > EOF
212
212
213 Pull from bundle and trigger notify
213 Pull from bundle and trigger notify
214
214
215 $ hg pull -u ../kw.hg
215 $ hg pull -u ../kw.hg
216 pulling from ../kw.hg
216 pulling from ../kw.hg
217 requesting all changes
217 requesting all changes
218 adding changesets
218 adding changesets
219 adding manifests
219 adding manifests
220 adding file changes
220 adding file changes
221 added 2 changesets with 3 changes to 3 files
221 added 2 changesets with 3 changes to 3 files
222 Content-Type: text/plain; charset="us-ascii"
222 Content-Type: text/plain; charset="us-ascii"
223 MIME-Version: 1.0
223 MIME-Version: 1.0
224 Content-Transfer-Encoding: 7bit
224 Content-Transfer-Encoding: 7bit
225 Date: * (glob)
225 Date: * (glob)
226 Subject: changeset in...
226 Subject: changeset in...
227 From: mercurial
227 From: mercurial
228 X-Hg-Notification: changeset a2392c293916
228 X-Hg-Notification: changeset a2392c293916
229 Message-Id: <hg.a2392c293916*> (glob)
229 Message-Id: <hg.a2392c293916*> (glob)
230 To: Test
230 To: Test
231
231
232 changeset a2392c293916 in $TESTTMP/Test (glob)
232 changeset a2392c293916 in $TESTTMP/Test (glob)
233 details: $TESTTMP/Test?cmd=changeset;node=a2392c293916
233 details: $TESTTMP/Test?cmd=changeset;node=a2392c293916
234 description:
234 description:
235 addsym
235 addsym
236
236
237 diffs (6 lines):
237 diffs (6 lines):
238
238
239 diff -r 000000000000 -r a2392c293916 sym
239 diff -r 000000000000 -r a2392c293916 sym
240 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
240 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
241 +++ b/sym Sat Feb 09 20:25:47 2008 +0100
241 +++ b/sym Sat Feb 09 20:25:47 2008 +0100
242 @@ -0,0 +1,1 @@
242 @@ -0,0 +1,1 @@
243 +a
243 +a
244 \ No newline at end of file
244 \ No newline at end of file
245 Content-Type: text/plain; charset="us-ascii"
245 Content-Type: text/plain; charset="us-ascii"
246 MIME-Version: 1.0
246 MIME-Version: 1.0
247 Content-Transfer-Encoding: 7bit
247 Content-Transfer-Encoding: 7bit
248 Date:* (glob)
248 Date:* (glob)
249 Subject: changeset in...
249 Subject: changeset in...
250 From: User Name <user@example.com>
250 From: User Name <user@example.com>
251 X-Hg-Notification: changeset ef63ca68695b
251 X-Hg-Notification: changeset ef63ca68695b
252 Message-Id: <hg.ef63ca68695b*> (glob)
252 Message-Id: <hg.ef63ca68695b*> (glob)
253 To: Test
253 To: Test
254
254
255 changeset ef63ca68695b in $TESTTMP/Test (glob)
255 changeset ef63ca68695b in $TESTTMP/Test (glob)
256 details: $TESTTMP/Test?cmd=changeset;node=ef63ca68695b
256 details: $TESTTMP/Test?cmd=changeset;node=ef63ca68695b
257 description:
257 description:
258 absym
258 absym
259
259
260 diffs (12 lines):
260 diffs (12 lines):
261
261
262 diff -r a2392c293916 -r ef63ca68695b a
262 diff -r a2392c293916 -r ef63ca68695b a
263 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
263 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
264 +++ b/a Thu Jan 01 00:00:00 1970 +0000
264 +++ b/a Thu Jan 01 00:00:00 1970 +0000
265 @@ -0,0 +1,3 @@
265 @@ -0,0 +1,3 @@
266 +expand $Id$
266 +expand $Id$
267 +do not process $Id:
267 +do not process $Id:
268 +xxx $
268 +xxx $
269 diff -r a2392c293916 -r ef63ca68695b b
269 diff -r a2392c293916 -r ef63ca68695b b
270 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
270 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
271 +++ b/b Thu Jan 01 00:00:00 1970 +0000
271 +++ b/b Thu Jan 01 00:00:00 1970 +0000
272 @@ -0,0 +1,1 @@
272 @@ -0,0 +1,1 @@
273 +ignore $Id$
273 +ignore $Id$
274 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
274 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
275
275
276 $ cp $HGRCPATH.nohooks $HGRCPATH
276 $ cp $HGRCPATH.nohooks $HGRCPATH
277
277
278 Touch files and check with status
278 Touch files and check with status
279
279
280 $ touch a b
280 $ touch a b
281 $ hg status
281 $ hg status
282
282
283 Update and expand
283 Update and expand
284
284
285 $ rm sym a b
285 $ rm sym a b
286 $ hg update -C
286 $ hg update -C
287 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
287 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
288 $ cat a b
288 $ cat a b
289 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
289 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
290 do not process $Id:
290 do not process $Id:
291 xxx $
291 xxx $
292 ignore $Id$
292 ignore $Id$
293
293
294 Check whether expansion is filewise and file mode is preserved
294 Check whether expansion is filewise and file mode is preserved
295
295
296 $ echo '$Id$' > c
296 $ echo '$Id$' > c
297 $ echo 'tests for different changenodes' >> c
297 $ echo 'tests for different changenodes' >> c
298 #if unix-permissions
298 #if unix-permissions
299 $ chmod 600 c
299 $ chmod 600 c
300 $ ls -l c | cut -b 1-10
300 $ ls -l c | cut -b 1-10
301 -rw-------
301 -rw-------
302 #endif
302 #endif
303
303
304 commit file c
304 commit file c
305
305
306 $ hg commit -A -mcndiff -d '1 0' -u 'User Name <user@example.com>'
306 $ hg commit -A -mcndiff -d '1 0' -u 'User Name <user@example.com>'
307 adding c
307 adding c
308 #if unix-permissions
308 #if unix-permissions
309 $ ls -l c | cut -b 1-10
309 $ ls -l c | cut -b 1-10
310 -rw-------
310 -rw-------
311 #endif
311 #endif
312
312
313 force expansion
313 force expansion
314
314
315 $ hg -v kwexpand
315 $ hg -v kwexpand
316 overwriting a expanding keywords
316 overwriting a expanding keywords
317 overwriting c expanding keywords
317 overwriting c expanding keywords
318
318
319 compare changenodes in a and c
319 compare changenodes in a and c
320
320
321 $ cat a c
321 $ cat a c
322 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
322 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
323 do not process $Id:
323 do not process $Id:
324 xxx $
324 xxx $
325 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
325 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
326 tests for different changenodes
326 tests for different changenodes
327
327
328 record
328 record
329
329
330 $ echo '$Id$' > r
330 $ echo '$Id$' > r
331 $ hg add r
331 $ hg add r
332
332
333 record chunk
333 record chunk
334
334
335 >>> lines = open('a', 'rb').readlines()
335 >>> lines = open('a', 'rb').readlines()
336 >>> lines.insert(1, 'foo\n')
336 >>> lines.insert(1, 'foo\n')
337 >>> lines.append('bar\n')
337 >>> lines.append('bar\n')
338 >>> open('a', 'wb').writelines(lines)
338 >>> open('a', 'wb').writelines(lines)
339 $ hg record -d '10 1' -m rectest a<<EOF
339 $ hg record -d '10 1' -m rectest a<<EOF
340 > y
340 > y
341 > y
341 > y
342 > n
342 > n
343 > EOF
343 > EOF
344 diff --git a/a b/a
344 diff --git a/a b/a
345 2 hunks, 2 lines changed
345 2 hunks, 2 lines changed
346 examine changes to 'a'? [Ynesfdaq?]
346 examine changes to 'a'? [Ynesfdaq?]
347 @@ -1,3 +1,4 @@
347 @@ -1,3 +1,4 @@
348 expand $Id$
348 expand $Id$
349 +foo
349 +foo
350 do not process $Id:
350 do not process $Id:
351 xxx $
351 xxx $
352 record change 1/2 to 'a'? [Ynesfdaq?]
352 record change 1/2 to 'a'? [Ynesfdaq?]
353 @@ -2,2 +3,3 @@
353 @@ -2,2 +3,3 @@
354 do not process $Id:
354 do not process $Id:
355 xxx $
355 xxx $
356 +bar
356 +bar
357 record change 2/2 to 'a'? [Ynesfdaq?]
357 record change 2/2 to 'a'? [Ynesfdaq?]
358
358
359 $ hg identify
359 $ hg identify
360 5f5eb23505c3+ tip
360 5f5eb23505c3+ tip
361 $ hg status
361 $ hg status
362 M a
362 M a
363 A r
363 A r
364
364
365 Cat modified file a
365 Cat modified file a
366
366
367 $ cat a
367 $ cat a
368 expand $Id: a,v 5f5eb23505c3 1970/01/01 00:00:10 test $
368 expand $Id: a,v 5f5eb23505c3 1970/01/01 00:00:10 test $
369 foo
369 foo
370 do not process $Id:
370 do not process $Id:
371 xxx $
371 xxx $
372 bar
372 bar
373
373
374 Diff remaining chunk
374 Diff remaining chunk
375
375
376 $ hg diff a
376 $ hg diff a
377 diff -r 5f5eb23505c3 a
377 diff -r 5f5eb23505c3 a
378 --- a/a Thu Jan 01 00:00:09 1970 -0000
378 --- a/a Thu Jan 01 00:00:09 1970 -0000
379 +++ b/a * (glob)
379 +++ b/a * (glob)
380 @@ -2,3 +2,4 @@
380 @@ -2,3 +2,4 @@
381 foo
381 foo
382 do not process $Id:
382 do not process $Id:
383 xxx $
383 xxx $
384 +bar
384 +bar
385
385
386 $ hg rollback
386 $ hg rollback
387 repository tip rolled back to revision 2 (undo commit)
387 repository tip rolled back to revision 2 (undo commit)
388 working directory now based on revision 2
388 working directory now based on revision 2
389
389
390 Record all chunks in file a
390 Record all chunks in file a
391
391
392 $ echo foo > msg
392 $ echo foo > msg
393
393
394 - do not use "hg record -m" here!
394 - do not use "hg record -m" here!
395
395
396 $ hg record -l msg -d '11 1' a<<EOF
396 $ hg record -l msg -d '11 1' a<<EOF
397 > y
397 > y
398 > y
398 > y
399 > y
399 > y
400 > EOF
400 > EOF
401 diff --git a/a b/a
401 diff --git a/a b/a
402 2 hunks, 2 lines changed
402 2 hunks, 2 lines changed
403 examine changes to 'a'? [Ynesfdaq?]
403 examine changes to 'a'? [Ynesfdaq?]
404 @@ -1,3 +1,4 @@
404 @@ -1,3 +1,4 @@
405 expand $Id$
405 expand $Id$
406 +foo
406 +foo
407 do not process $Id:
407 do not process $Id:
408 xxx $
408 xxx $
409 record change 1/2 to 'a'? [Ynesfdaq?]
409 record change 1/2 to 'a'? [Ynesfdaq?]
410 @@ -2,2 +3,3 @@
410 @@ -2,2 +3,3 @@
411 do not process $Id:
411 do not process $Id:
412 xxx $
412 xxx $
413 +bar
413 +bar
414 record change 2/2 to 'a'? [Ynesfdaq?]
414 record change 2/2 to 'a'? [Ynesfdaq?]
415
415
416 File a should be clean
416 File a should be clean
417
417
418 $ hg status -A a
418 $ hg status -A a
419 C a
419 C a
420
420
421 rollback and revert expansion
421 rollback and revert expansion
422
422
423 $ cat a
423 $ cat a
424 expand $Id: a,v 78e0a02d76aa 1970/01/01 00:00:11 test $
424 expand $Id: a,v 78e0a02d76aa 1970/01/01 00:00:11 test $
425 foo
425 foo
426 do not process $Id:
426 do not process $Id:
427 xxx $
427 xxx $
428 bar
428 bar
429 $ hg --verbose rollback
429 $ hg --verbose rollback
430 repository tip rolled back to revision 2 (undo commit)
430 repository tip rolled back to revision 2 (undo commit)
431 working directory now based on revision 2
431 working directory now based on revision 2
432 overwriting a expanding keywords
432 overwriting a expanding keywords
433 $ hg status a
433 $ hg status a
434 M a
434 M a
435 $ cat a
435 $ cat a
436 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
436 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
437 foo
437 foo
438 do not process $Id:
438 do not process $Id:
439 xxx $
439 xxx $
440 bar
440 bar
441 $ echo '$Id$' > y
441 $ echo '$Id$' > y
442 $ echo '$Id$' > z
442 $ echo '$Id$' > z
443 $ hg add y
443 $ hg add y
444 $ hg commit -Am "rollback only" z
444 $ hg commit -Am "rollback only" z
445 $ cat z
445 $ cat z
446 $Id: z,v 45a5d3adce53 1970/01/01 00:00:00 test $
446 $Id: z,v 45a5d3adce53 1970/01/01 00:00:00 test $
447 $ hg --verbose rollback
447 $ hg --verbose rollback
448 repository tip rolled back to revision 2 (undo commit)
448 repository tip rolled back to revision 2 (undo commit)
449 working directory now based on revision 2
449 working directory now based on revision 2
450 overwriting z shrinking keywords
450 overwriting z shrinking keywords
451
451
452 Only z should be overwritten
452 Only z should be overwritten
453
453
454 $ hg status a y z
454 $ hg status a y z
455 M a
455 M a
456 A y
456 A y
457 A z
457 A z
458 $ cat z
458 $ cat z
459 $Id$
459 $Id$
460 $ hg forget y z
460 $ hg forget y z
461 $ rm y z
461 $ rm y z
462
462
463 record added file alone
463 record added file alone
464
464
465 $ hg -v record -l msg -d '12 2' r<<EOF
465 $ hg -v record -l msg -d '12 2' r<<EOF
466 > y
466 > y
467 > EOF
467 > EOF
468 diff --git a/r b/r
468 diff --git a/r b/r
469 new file mode 100644
469 new file mode 100644
470 examine changes to 'r'? [Ynesfdaq?]
470 examine changes to 'r'? [Ynesfdaq?]
471 r
471 r
472 committed changeset 3:82a2f715724d
472 committed changeset 3:82a2f715724d
473 overwriting r expanding keywords
473 overwriting r expanding keywords
474 - status call required for dirstate.normallookup() check
474 - status call required for dirstate.normallookup() check
475 $ hg status r
475 $ hg status r
476 $ hg --verbose rollback
476 $ hg --verbose rollback
477 repository tip rolled back to revision 2 (undo commit)
477 repository tip rolled back to revision 2 (undo commit)
478 working directory now based on revision 2
478 working directory now based on revision 2
479 overwriting r shrinking keywords
479 overwriting r shrinking keywords
480 $ hg forget r
480 $ hg forget r
481 $ rm msg r
481 $ rm msg r
482 $ hg update -C
482 $ hg update -C
483 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
483 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
484
484
485 record added keyword ignored file
485 record added keyword ignored file
486
486
487 $ echo '$Id$' > i
487 $ echo '$Id$' > i
488 $ hg add i
488 $ hg add i
489 $ hg --verbose record -d '13 1' -m recignored<<EOF
489 $ hg --verbose record -d '13 1' -m recignored<<EOF
490 > y
490 > y
491 > EOF
491 > EOF
492 diff --git a/i b/i
492 diff --git a/i b/i
493 new file mode 100644
493 new file mode 100644
494 examine changes to 'i'? [Ynesfdaq?]
494 examine changes to 'i'? [Ynesfdaq?]
495 i
495 i
496 committed changeset 3:9f40ceb5a072
496 committed changeset 3:9f40ceb5a072
497 $ cat i
497 $ cat i
498 $Id$
498 $Id$
499 $ hg -q rollback
499 $ hg -q rollback
500 $ hg forget i
500 $ hg forget i
501 $ rm i
501 $ rm i
502
502
503 amend
503 amend
504
504
505 $ echo amend >> a
505 $ echo amend >> a
506 $ echo amend >> b
506 $ echo amend >> b
507 $ hg -q commit -d '14 1' -m 'prepare amend'
507 $ hg -q commit -d '14 1' -m 'prepare amend'
508
508
509 $ hg --debug commit --amend -d '15 1' -m 'amend without changes' | grep keywords
509 $ hg --debug commit --amend -d '15 1' -m 'amend without changes' | grep keywords
510 overwriting a expanding keywords
510 overwriting a expanding keywords
511 $ hg -q id
511 $ hg -q id
512 67d8c481a6be
512 67d8c481a6be
513 $ head -1 a
513 $ head -1 a
514 expand $Id: a,v 67d8c481a6be 1970/01/01 00:00:15 test $
514 expand $Id: a,v 67d8c481a6be 1970/01/01 00:00:15 test $
515
515
516 $ hg -q strip -n tip
516 $ hg -q strip -n tip
517
517
518 Test patch queue repo
518 Test patch queue repo
519
519
520 $ hg init --mq
520 $ hg init --mq
521 $ hg qimport -r tip -n mqtest.diff
521 $ hg qimport -r tip -n mqtest.diff
522 $ hg commit --mq -m mqtest
522 $ hg commit --mq -m mqtest
523
523
524 Keywords should not be expanded in patch
524 Keywords should not be expanded in patch
525
525
526 $ cat .hg/patches/mqtest.diff
526 $ cat .hg/patches/mqtest.diff
527 # HG changeset patch
527 # HG changeset patch
528 # User User Name <user@example.com>
528 # User User Name <user@example.com>
529 # Date 1 0
529 # Date 1 0
530 # Node ID 40a904bbbe4cd4ab0a1f28411e35db26341a40ad
530 # Node ID 40a904bbbe4cd4ab0a1f28411e35db26341a40ad
531 # Parent ef63ca68695bc9495032c6fda1350c71e6d256e9
531 # Parent ef63ca68695bc9495032c6fda1350c71e6d256e9
532 cndiff
532 cndiff
533
533
534 diff -r ef63ca68695b -r 40a904bbbe4c c
534 diff -r ef63ca68695b -r 40a904bbbe4c c
535 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
535 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
536 +++ b/c Thu Jan 01 00:00:01 1970 +0000
536 +++ b/c Thu Jan 01 00:00:01 1970 +0000
537 @@ -0,0 +1,2 @@
537 @@ -0,0 +1,2 @@
538 +$Id$
538 +$Id$
539 +tests for different changenodes
539 +tests for different changenodes
540
540
541 $ hg qpop
541 $ hg qpop
542 popping mqtest.diff
542 popping mqtest.diff
543 patch queue now empty
543 patch queue now empty
544
544
545 qgoto, implying qpush, should expand
545 qgoto, implying qpush, should expand
546
546
547 $ hg qgoto mqtest.diff
547 $ hg qgoto mqtest.diff
548 applying mqtest.diff
548 applying mqtest.diff
549 now at: mqtest.diff
549 now at: mqtest.diff
550 $ cat c
550 $ cat c
551 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
551 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
552 tests for different changenodes
552 tests for different changenodes
553 $ hg cat c
553 $ hg cat c
554 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
554 $Id: c,v 40a904bbbe4c 1970/01/01 00:00:01 user $
555 tests for different changenodes
555 tests for different changenodes
556
556
557 Keywords should not be expanded in filelog
557 Keywords should not be expanded in filelog
558
558
559 $ hg --config 'extensions.keyword=!' cat c
559 $ hg --config 'extensions.keyword=!' cat c
560 $Id$
560 $Id$
561 tests for different changenodes
561 tests for different changenodes
562
562
563 qpop and move on
563 qpop and move on
564
564
565 $ hg qpop
565 $ hg qpop
566 popping mqtest.diff
566 popping mqtest.diff
567 patch queue now empty
567 patch queue now empty
568
568
569 Copy and show added kwfiles
569 Copy and show added kwfiles
570
570
571 $ hg cp a c
571 $ hg cp a c
572 $ hg kwfiles
572 $ hg kwfiles
573 a
573 a
574 c
574 c
575
575
576 Commit and show expansion in original and copy
576 Commit and show expansion in original and copy
577
577
578 $ hg --debug commit -ma2c -d '1 0' -u 'User Name <user@example.com>'
578 $ hg --debug commit -ma2c -d '1 0' -u 'User Name <user@example.com>'
579 invalid branchheads cache (served): tip differs
579 invalid branchheads cache (served): tip differs
580 c
580 c
581 c: copy a:0045e12f6c5791aac80ca6cbfd97709a88307292
581 c: copy a:0045e12f6c5791aac80ca6cbfd97709a88307292
582 invalid branchheads cache (served): tip differs
582 overwriting c expanding keywords
583 overwriting c expanding keywords
583 committed changeset 2:25736cf2f5cbe41f6be4e6784ef6ecf9f3bbcc7d
584 committed changeset 2:25736cf2f5cbe41f6be4e6784ef6ecf9f3bbcc7d
584 $ cat a c
585 $ cat a c
585 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
586 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
586 do not process $Id:
587 do not process $Id:
587 xxx $
588 xxx $
588 expand $Id: c,v 25736cf2f5cb 1970/01/01 00:00:01 user $
589 expand $Id: c,v 25736cf2f5cb 1970/01/01 00:00:01 user $
589 do not process $Id:
590 do not process $Id:
590 xxx $
591 xxx $
591
592
592 Touch copied c and check its status
593 Touch copied c and check its status
593
594
594 $ touch c
595 $ touch c
595 $ hg status
596 $ hg status
596
597
597 Copy kwfile to keyword ignored file unexpanding keywords
598 Copy kwfile to keyword ignored file unexpanding keywords
598
599
599 $ hg --verbose copy a i
600 $ hg --verbose copy a i
600 copying a to i
601 copying a to i
601 overwriting i shrinking keywords
602 overwriting i shrinking keywords
602 $ head -n 1 i
603 $ head -n 1 i
603 expand $Id$
604 expand $Id$
604 $ hg forget i
605 $ hg forget i
605 $ rm i
606 $ rm i
606
607
607 Copy ignored file to ignored file: no overwriting
608 Copy ignored file to ignored file: no overwriting
608
609
609 $ hg --verbose copy b i
610 $ hg --verbose copy b i
610 copying b to i
611 copying b to i
611 $ hg forget i
612 $ hg forget i
612 $ rm i
613 $ rm i
613
614
614 cp symlink file; hg cp -A symlink file (part1)
615 cp symlink file; hg cp -A symlink file (part1)
615 - copied symlink points to kwfile: overwrite
616 - copied symlink points to kwfile: overwrite
616
617
617 #if symlink
618 #if symlink
618 $ cp sym i
619 $ cp sym i
619 $ ls -l i
620 $ ls -l i
620 -rw-r--r--* (glob)
621 -rw-r--r--* (glob)
621 $ head -1 i
622 $ head -1 i
622 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
623 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
623 $ hg copy --after --verbose sym i
624 $ hg copy --after --verbose sym i
624 copying sym to i
625 copying sym to i
625 overwriting i shrinking keywords
626 overwriting i shrinking keywords
626 $ head -1 i
627 $ head -1 i
627 expand $Id$
628 expand $Id$
628 $ hg forget i
629 $ hg forget i
629 $ rm i
630 $ rm i
630 #endif
631 #endif
631
632
632 Test different options of hg kwfiles
633 Test different options of hg kwfiles
633
634
634 $ hg kwfiles
635 $ hg kwfiles
635 a
636 a
636 c
637 c
637 $ hg -v kwfiles --ignore
638 $ hg -v kwfiles --ignore
638 I b
639 I b
639 I sym
640 I sym
640 $ hg kwfiles --all
641 $ hg kwfiles --all
641 K a
642 K a
642 K c
643 K c
643 I b
644 I b
644 I sym
645 I sym
645
646
646 Diff specific revision
647 Diff specific revision
647
648
648 $ hg diff --rev 1
649 $ hg diff --rev 1
649 diff -r ef63ca68695b c
650 diff -r ef63ca68695b c
650 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
651 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
651 +++ b/c * (glob)
652 +++ b/c * (glob)
652 @@ -0,0 +1,3 @@
653 @@ -0,0 +1,3 @@
653 +expand $Id$
654 +expand $Id$
654 +do not process $Id:
655 +do not process $Id:
655 +xxx $
656 +xxx $
656
657
657 Status after rollback:
658 Status after rollback:
658
659
659 $ hg rollback
660 $ hg rollback
660 repository tip rolled back to revision 1 (undo commit)
661 repository tip rolled back to revision 1 (undo commit)
661 working directory now based on revision 1
662 working directory now based on revision 1
662 $ hg status
663 $ hg status
663 A c
664 A c
664 $ hg update --clean
665 $ hg update --clean
665 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
666 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
666
667
667 #if symlink
668 #if symlink
668
669
669 cp symlink file; hg cp -A symlink file (part2)
670 cp symlink file; hg cp -A symlink file (part2)
670 - copied symlink points to kw ignored file: do not overwrite
671 - copied symlink points to kw ignored file: do not overwrite
671
672
672 $ cat a > i
673 $ cat a > i
673 $ ln -s i symignored
674 $ ln -s i symignored
674 $ hg commit -Am 'fake expansion in ignored and symlink' i symignored
675 $ hg commit -Am 'fake expansion in ignored and symlink' i symignored
675 $ cp symignored x
676 $ cp symignored x
676 $ hg copy --after --verbose symignored x
677 $ hg copy --after --verbose symignored x
677 copying symignored to x
678 copying symignored to x
678 $ head -n 1 x
679 $ head -n 1 x
679 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
680 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
680 $ hg forget x
681 $ hg forget x
681 $ rm x
682 $ rm x
682
683
683 $ hg rollback
684 $ hg rollback
684 repository tip rolled back to revision 1 (undo commit)
685 repository tip rolled back to revision 1 (undo commit)
685 working directory now based on revision 1
686 working directory now based on revision 1
686 $ hg update --clean
687 $ hg update --clean
687 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
688 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
688 $ rm i symignored
689 $ rm i symignored
689
690
690 #endif
691 #endif
691
692
692 Custom keywordmaps as argument to kwdemo
693 Custom keywordmaps as argument to kwdemo
693
694
694 $ hg --quiet kwdemo "Xinfo = {author}: {desc}"
695 $ hg --quiet kwdemo "Xinfo = {author}: {desc}"
695 [extensions]
696 [extensions]
696 keyword =
697 keyword =
697 [keyword]
698 [keyword]
698 ** =
699 ** =
699 b = ignore
700 b = ignore
700 demo.txt =
701 demo.txt =
701 i = ignore
702 i = ignore
702 [keywordset]
703 [keywordset]
703 svn = False
704 svn = False
704 [keywordmaps]
705 [keywordmaps]
705 Xinfo = {author}: {desc}
706 Xinfo = {author}: {desc}
706 $Xinfo: test: hg keyword configuration and expansion example $
707 $Xinfo: test: hg keyword configuration and expansion example $
707
708
708 Configure custom keywordmaps
709 Configure custom keywordmaps
709
710
710 $ cat <<EOF >>$HGRCPATH
711 $ cat <<EOF >>$HGRCPATH
711 > [keywordmaps]
712 > [keywordmaps]
712 > Id = {file} {node|short} {date|rfc822date} {author|user}
713 > Id = {file} {node|short} {date|rfc822date} {author|user}
713 > Xinfo = {author}: {desc}
714 > Xinfo = {author}: {desc}
714 > EOF
715 > EOF
715
716
716 Cat and hg cat files before custom expansion
717 Cat and hg cat files before custom expansion
717
718
718 $ cat a b
719 $ cat a b
719 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
720 expand $Id: a,v ef63ca68695b 1970/01/01 00:00:00 user $
720 do not process $Id:
721 do not process $Id:
721 xxx $
722 xxx $
722 ignore $Id$
723 ignore $Id$
723 $ hg cat sym a b && echo
724 $ hg cat sym a b && echo
724 expand $Id: a ef63ca68695b Thu, 01 Jan 1970 00:00:00 +0000 user $
725 expand $Id: a ef63ca68695b Thu, 01 Jan 1970 00:00:00 +0000 user $
725 do not process $Id:
726 do not process $Id:
726 xxx $
727 xxx $
727 ignore $Id$
728 ignore $Id$
728 a
729 a
729
730
730 Write custom keyword and prepare multi-line commit message
731 Write custom keyword and prepare multi-line commit message
731
732
732 $ echo '$Xinfo$' >> a
733 $ echo '$Xinfo$' >> a
733 $ cat <<EOF >> log
734 $ cat <<EOF >> log
734 > firstline
735 > firstline
735 > secondline
736 > secondline
736 > EOF
737 > EOF
737
738
738 Interrupted commit should not change state
739 Interrupted commit should not change state
739
740
740 $ hg commit
741 $ hg commit
741 abort: empty commit message
742 abort: empty commit message
742 [255]
743 [255]
743 $ hg status
744 $ hg status
744 M a
745 M a
745 ? c
746 ? c
746 ? log
747 ? log
747
748
748 Commit with multi-line message and custom expansion
749 Commit with multi-line message and custom expansion
749
750
750 |Note:
751 |Note:
751 |
752 |
752 | After the last rollback, the "unserved" branchheads cache became invalid, but
753 | After the last rollback, the "unserved" branchheads cache became invalid, but
753 | all changesets in the repo were public. For filtering this means:
754 | all changesets in the repo were public. For filtering this means:
754 | "mutable" == "unserved" == ΓΈ.
755 | "mutable" == "unserved" == ΓΈ.
755 |
756 |
756 | As the "unserved" cache is invalid, we fall back to the "mutable" cache. But
757 | As the "unserved" cache is invalid, we fall back to the "mutable" cache. But
757 | no update is needed between "mutable" and "unserved" and the "unserved" cache
758 | no update is needed between "mutable" and "unserved" and the "unserved" cache
758 | is not updated on disk. The on-disk version therefore stays invalid for some
759 | is not updated on disk. The on-disk version therefore stays invalid for some
759 | time. This explains why the "unserved" branchheads cache is detected as
760 | time. This explains why the "unserved" branchheads cache is detected as
760 | invalid here.
761 | invalid here.
761
762
762 $ hg --debug commit -l log -d '2 0' -u 'User Name <user@example.com>'
763 $ hg --debug commit -l log -d '2 0' -u 'User Name <user@example.com>'
763 invalid branchheads cache (served): tip differs
764 invalid branchheads cache (served): tip differs
764 a
765 a
765 invalid branchheads cache: tip differs
766 invalid branchheads cache (served): tip differs
766 invalid branchheads cache (served): tip differs
767 overwriting a expanding keywords
767 overwriting a expanding keywords
768 committed changeset 2:bb948857c743469b22bbf51f7ec8112279ca5d83
768 committed changeset 2:bb948857c743469b22bbf51f7ec8112279ca5d83
769 $ rm log
769 $ rm log
770
770
771 Stat, verify and show custom expansion (firstline)
771 Stat, verify and show custom expansion (firstline)
772
772
773 $ hg status
773 $ hg status
774 ? c
774 ? c
775 $ hg verify
775 $ hg verify
776 checking changesets
776 checking changesets
777 checking manifests
777 checking manifests
778 crosschecking files in changesets and manifests
778 crosschecking files in changesets and manifests
779 checking files
779 checking files
780 3 files, 3 changesets, 4 total revisions
780 3 files, 3 changesets, 4 total revisions
781 $ cat a b
781 $ cat a b
782 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
782 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
783 do not process $Id:
783 do not process $Id:
784 xxx $
784 xxx $
785 $Xinfo: User Name <user@example.com>: firstline $
785 $Xinfo: User Name <user@example.com>: firstline $
786 ignore $Id$
786 ignore $Id$
787 $ hg cat sym a b && echo
787 $ hg cat sym a b && echo
788 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
788 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
789 do not process $Id:
789 do not process $Id:
790 xxx $
790 xxx $
791 $Xinfo: User Name <user@example.com>: firstline $
791 $Xinfo: User Name <user@example.com>: firstline $
792 ignore $Id$
792 ignore $Id$
793 a
793 a
794
794
795 annotate
795 annotate
796
796
797 $ hg annotate a
797 $ hg annotate a
798 1: expand $Id$
798 1: expand $Id$
799 1: do not process $Id:
799 1: do not process $Id:
800 1: xxx $
800 1: xxx $
801 2: $Xinfo$
801 2: $Xinfo$
802
802
803 remove with status checks
803 remove with status checks
804
804
805 $ hg debugrebuildstate
805 $ hg debugrebuildstate
806 $ hg remove a
806 $ hg remove a
807 $ hg --debug commit -m rma
807 $ hg --debug commit -m rma
808 invalid branchheads cache: tip differs
809 committed changeset 3:d14c712653769de926994cf7fbb06c8fbd68f012
808 committed changeset 3:d14c712653769de926994cf7fbb06c8fbd68f012
810 $ hg status
809 $ hg status
811 ? c
810 ? c
812
811
813 Rollback, revert, and check expansion
812 Rollback, revert, and check expansion
814
813
815 $ hg rollback
814 $ hg rollback
816 repository tip rolled back to revision 2 (undo commit)
815 repository tip rolled back to revision 2 (undo commit)
817 working directory now based on revision 2
816 working directory now based on revision 2
818 $ hg status
817 $ hg status
819 R a
818 R a
820 ? c
819 ? c
821 $ hg revert --no-backup --rev tip a
820 $ hg revert --no-backup --rev tip a
822 $ cat a
821 $ cat a
823 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
822 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
824 do not process $Id:
823 do not process $Id:
825 xxx $
824 xxx $
826 $Xinfo: User Name <user@example.com>: firstline $
825 $Xinfo: User Name <user@example.com>: firstline $
827
826
828 Clone to test global and local configurations
827 Clone to test global and local configurations
829
828
830 $ cd ..
829 $ cd ..
831
830
832 Expansion in destination with global configuration
831 Expansion in destination with global configuration
833
832
834 $ hg --quiet clone Test globalconf
833 $ hg --quiet clone Test globalconf
835 $ cat globalconf/a
834 $ cat globalconf/a
836 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
835 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
837 do not process $Id:
836 do not process $Id:
838 xxx $
837 xxx $
839 $Xinfo: User Name <user@example.com>: firstline $
838 $Xinfo: User Name <user@example.com>: firstline $
840
839
841 No expansion in destination with local configuration in origin only
840 No expansion in destination with local configuration in origin only
842
841
843 $ hg --quiet --config 'keyword.**=ignore' clone Test localconf
842 $ hg --quiet --config 'keyword.**=ignore' clone Test localconf
844 $ cat localconf/a
843 $ cat localconf/a
845 expand $Id$
844 expand $Id$
846 do not process $Id:
845 do not process $Id:
847 xxx $
846 xxx $
848 $Xinfo$
847 $Xinfo$
849
848
850 Clone to test incoming
849 Clone to test incoming
851
850
852 $ hg clone -r1 Test Test-a
851 $ hg clone -r1 Test Test-a
853 adding changesets
852 adding changesets
854 adding manifests
853 adding manifests
855 adding file changes
854 adding file changes
856 added 2 changesets with 3 changes to 3 files
855 added 2 changesets with 3 changes to 3 files
857 updating to branch default
856 updating to branch default
858 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
857 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
859 $ cd Test-a
858 $ cd Test-a
860 $ cat <<EOF >> .hg/hgrc
859 $ cat <<EOF >> .hg/hgrc
861 > [paths]
860 > [paths]
862 > default = ../Test
861 > default = ../Test
863 > EOF
862 > EOF
864 $ hg incoming
863 $ hg incoming
865 comparing with $TESTTMP/Test (glob)
864 comparing with $TESTTMP/Test (glob)
866 searching for changes
865 searching for changes
867 changeset: 2:bb948857c743
866 changeset: 2:bb948857c743
868 tag: tip
867 tag: tip
869 user: User Name <user@example.com>
868 user: User Name <user@example.com>
870 date: Thu Jan 01 00:00:02 1970 +0000
869 date: Thu Jan 01 00:00:02 1970 +0000
871 summary: firstline
870 summary: firstline
872
871
873 Imported patch should not be rejected
872 Imported patch should not be rejected
874
873
875 >>> import re
874 >>> import re
876 >>> text = re.sub(r'(Id.*)', r'\1 rejecttest', open('a').read())
875 >>> text = re.sub(r'(Id.*)', r'\1 rejecttest', open('a').read())
877 >>> open('a', 'wb').write(text)
876 >>> open('a', 'wb').write(text)
878 $ hg --debug commit -m'rejects?' -d '3 0' -u 'User Name <user@example.com>'
877 $ hg --debug commit -m'rejects?' -d '3 0' -u 'User Name <user@example.com>'
879 a
878 a
880 overwriting a expanding keywords
879 overwriting a expanding keywords
881 committed changeset 2:85e279d709ffc28c9fdd1b868570985fc3d87082
880 committed changeset 2:85e279d709ffc28c9fdd1b868570985fc3d87082
882 $ hg export -o ../rejecttest.diff tip
881 $ hg export -o ../rejecttest.diff tip
883 $ cd ../Test
882 $ cd ../Test
884 $ hg import ../rejecttest.diff
883 $ hg import ../rejecttest.diff
885 applying ../rejecttest.diff
884 applying ../rejecttest.diff
886 $ cat a b
885 $ cat a b
887 expand $Id: a 4e0994474d25 Thu, 01 Jan 1970 00:00:03 +0000 user $ rejecttest
886 expand $Id: a 4e0994474d25 Thu, 01 Jan 1970 00:00:03 +0000 user $ rejecttest
888 do not process $Id: rejecttest
887 do not process $Id: rejecttest
889 xxx $
888 xxx $
890 $Xinfo: User Name <user@example.com>: rejects? $
889 $Xinfo: User Name <user@example.com>: rejects? $
891 ignore $Id$
890 ignore $Id$
892
891
893 $ hg rollback
892 $ hg rollback
894 repository tip rolled back to revision 2 (undo import)
893 repository tip rolled back to revision 2 (undo import)
895 working directory now based on revision 2
894 working directory now based on revision 2
896 $ hg update --clean
895 $ hg update --clean
897 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
896 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
898
897
899 kwexpand/kwshrink on selected files
898 kwexpand/kwshrink on selected files
900
899
901 $ mkdir x
900 $ mkdir x
902 $ hg copy a x/a
901 $ hg copy a x/a
903 $ hg --verbose kwshrink a
902 $ hg --verbose kwshrink a
904 overwriting a shrinking keywords
903 overwriting a shrinking keywords
905 - sleep required for dirstate.normal() check
904 - sleep required for dirstate.normal() check
906 $ sleep 1
905 $ sleep 1
907 $ hg status a
906 $ hg status a
908 $ hg --verbose kwexpand a
907 $ hg --verbose kwexpand a
909 overwriting a expanding keywords
908 overwriting a expanding keywords
910 $ hg status a
909 $ hg status a
911
910
912 kwexpand x/a should abort
911 kwexpand x/a should abort
913
912
914 $ hg --verbose kwexpand x/a
913 $ hg --verbose kwexpand x/a
915 abort: outstanding uncommitted changes
914 abort: outstanding uncommitted changes
916 [255]
915 [255]
917 $ cd x
916 $ cd x
918 $ hg --debug commit -m xa -d '3 0' -u 'User Name <user@example.com>'
917 $ hg --debug commit -m xa -d '3 0' -u 'User Name <user@example.com>'
919 x/a
918 x/a
920 x/a: copy a:779c764182ce5d43e2b1eb66ce06d7b47bfe342e
919 x/a: copy a:779c764182ce5d43e2b1eb66ce06d7b47bfe342e
921 invalid branchheads cache: tip differs
922 overwriting x/a expanding keywords
920 overwriting x/a expanding keywords
923 committed changeset 3:b4560182a3f9a358179fd2d835c15e9da379c1e4
921 committed changeset 3:b4560182a3f9a358179fd2d835c15e9da379c1e4
924 $ cat a
922 $ cat a
925 expand $Id: x/a b4560182a3f9 Thu, 01 Jan 1970 00:00:03 +0000 user $
923 expand $Id: x/a b4560182a3f9 Thu, 01 Jan 1970 00:00:03 +0000 user $
926 do not process $Id:
924 do not process $Id:
927 xxx $
925 xxx $
928 $Xinfo: User Name <user@example.com>: xa $
926 $Xinfo: User Name <user@example.com>: xa $
929
927
930 kwshrink a inside directory x
928 kwshrink a inside directory x
931
929
932 $ hg --verbose kwshrink a
930 $ hg --verbose kwshrink a
933 overwriting x/a shrinking keywords
931 overwriting x/a shrinking keywords
934 $ cat a
932 $ cat a
935 expand $Id$
933 expand $Id$
936 do not process $Id:
934 do not process $Id:
937 xxx $
935 xxx $
938 $Xinfo$
936 $Xinfo$
939 $ cd ..
937 $ cd ..
940
938
941 kwexpand nonexistent
939 kwexpand nonexistent
942
940
943 $ hg kwexpand nonexistent
941 $ hg kwexpand nonexistent
944 nonexistent:* (glob)
942 nonexistent:* (glob)
945
943
946
944
947 #if serve
945 #if serve
948 hg serve
946 hg serve
949 - expand with hgweb file
947 - expand with hgweb file
950 - no expansion with hgweb annotate/changeset/filediff
948 - no expansion with hgweb annotate/changeset/filediff
951 - check errors
949 - check errors
952
950
953 $ hg serve -p $HGPORT -d --pid-file=hg.pid -A access.log -E errors.log
951 $ hg serve -p $HGPORT -d --pid-file=hg.pid -A access.log -E errors.log
954 $ cat hg.pid >> $DAEMON_PIDS
952 $ cat hg.pid >> $DAEMON_PIDS
955 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'file/tip/a/?style=raw'
953 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'file/tip/a/?style=raw'
956 200 Script output follows
954 200 Script output follows
957
955
958 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
956 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
959 do not process $Id:
957 do not process $Id:
960 xxx $
958 xxx $
961 $Xinfo: User Name <user@example.com>: firstline $
959 $Xinfo: User Name <user@example.com>: firstline $
962 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'annotate/tip/a/?style=raw'
960 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'annotate/tip/a/?style=raw'
963 200 Script output follows
961 200 Script output follows
964
962
965
963
966 user@1: expand $Id$
964 user@1: expand $Id$
967 user@1: do not process $Id:
965 user@1: do not process $Id:
968 user@1: xxx $
966 user@1: xxx $
969 user@2: $Xinfo$
967 user@2: $Xinfo$
970
968
971
969
972
970
973
971
974 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'rev/tip/?style=raw'
972 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'rev/tip/?style=raw'
975 200 Script output follows
973 200 Script output follows
976
974
977
975
978 # HG changeset patch
976 # HG changeset patch
979 # User User Name <user@example.com>
977 # User User Name <user@example.com>
980 # Date 3 0
978 # Date 3 0
981 # Node ID b4560182a3f9a358179fd2d835c15e9da379c1e4
979 # Node ID b4560182a3f9a358179fd2d835c15e9da379c1e4
982 # Parent bb948857c743469b22bbf51f7ec8112279ca5d83
980 # Parent bb948857c743469b22bbf51f7ec8112279ca5d83
983 xa
981 xa
984
982
985 diff -r bb948857c743 -r b4560182a3f9 x/a
983 diff -r bb948857c743 -r b4560182a3f9 x/a
986 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
984 --- /dev/null Thu Jan 01 00:00:00 1970 +0000
987 +++ b/x/a Thu Jan 01 00:00:03 1970 +0000
985 +++ b/x/a Thu Jan 01 00:00:03 1970 +0000
988 @@ -0,0 +1,4 @@
986 @@ -0,0 +1,4 @@
989 +expand $Id$
987 +expand $Id$
990 +do not process $Id:
988 +do not process $Id:
991 +xxx $
989 +xxx $
992 +$Xinfo$
990 +$Xinfo$
993
991
994 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'diff/bb948857c743/a?style=raw'
992 $ "$TESTDIR/get-with-headers.py" localhost:$HGPORT 'diff/bb948857c743/a?style=raw'
995 200 Script output follows
993 200 Script output follows
996
994
997
995
998 diff -r ef63ca68695b -r bb948857c743 a
996 diff -r ef63ca68695b -r bb948857c743 a
999 --- a/a Thu Jan 01 00:00:00 1970 +0000
997 --- a/a Thu Jan 01 00:00:00 1970 +0000
1000 +++ b/a Thu Jan 01 00:00:02 1970 +0000
998 +++ b/a Thu Jan 01 00:00:02 1970 +0000
1001 @@ -1,3 +1,4 @@
999 @@ -1,3 +1,4 @@
1002 expand $Id$
1000 expand $Id$
1003 do not process $Id:
1001 do not process $Id:
1004 xxx $
1002 xxx $
1005 +$Xinfo$
1003 +$Xinfo$
1006
1004
1007
1005
1008
1006
1009
1007
1010 $ cat errors.log
1008 $ cat errors.log
1011 #endif
1009 #endif
1012
1010
1013 Prepare merge and resolve tests
1011 Prepare merge and resolve tests
1014
1012
1015 $ echo '$Id$' > m
1013 $ echo '$Id$' > m
1016 $ hg add m
1014 $ hg add m
1017 $ hg commit -m 4kw
1015 $ hg commit -m 4kw
1018 $ echo foo >> m
1016 $ echo foo >> m
1019 $ hg commit -m 5foo
1017 $ hg commit -m 5foo
1020
1018
1021 simplemerge
1019 simplemerge
1022
1020
1023 $ hg update 4
1021 $ hg update 4
1024 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1022 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1025 $ echo foo >> m
1023 $ echo foo >> m
1026 $ hg commit -m 6foo
1024 $ hg commit -m 6foo
1027 created new head
1025 created new head
1028 $ hg merge
1026 $ hg merge
1029 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
1027 0 files updated, 0 files merged, 0 files removed, 0 files unresolved
1030 (branch merge, don't forget to commit)
1028 (branch merge, don't forget to commit)
1031 $ hg commit -m simplemerge
1029 $ hg commit -m simplemerge
1032 $ cat m
1030 $ cat m
1033 $Id: m 27d48ee14f67 Thu, 01 Jan 1970 00:00:00 +0000 test $
1031 $Id: m 27d48ee14f67 Thu, 01 Jan 1970 00:00:00 +0000 test $
1034 foo
1032 foo
1035
1033
1036 conflict: keyword should stay outside conflict zone
1034 conflict: keyword should stay outside conflict zone
1037
1035
1038 $ hg update 4
1036 $ hg update 4
1039 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1037 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1040 $ echo bar >> m
1038 $ echo bar >> m
1041 $ hg commit -m 8bar
1039 $ hg commit -m 8bar
1042 created new head
1040 created new head
1043 $ hg merge
1041 $ hg merge
1044 merging m
1042 merging m
1045 warning: conflicts during merge.
1043 warning: conflicts during merge.
1046 merging m incomplete! (edit conflicts, then use 'hg resolve --mark')
1044 merging m incomplete! (edit conflicts, then use 'hg resolve --mark')
1047 0 files updated, 0 files merged, 0 files removed, 1 files unresolved
1045 0 files updated, 0 files merged, 0 files removed, 1 files unresolved
1048 use 'hg resolve' to retry unresolved file merges or 'hg update -C .' to abandon
1046 use 'hg resolve' to retry unresolved file merges or 'hg update -C .' to abandon
1049 [1]
1047 [1]
1050 $ cat m
1048 $ cat m
1051 $Id$
1049 $Id$
1052 <<<<<<< local
1050 <<<<<<< local
1053 bar
1051 bar
1054 =======
1052 =======
1055 foo
1053 foo
1056 >>>>>>> other
1054 >>>>>>> other
1057
1055
1058 resolve to local
1056 resolve to local
1059
1057
1060 $ HGMERGE=internal:local hg resolve -a
1058 $ HGMERGE=internal:local hg resolve -a
1061 $ hg commit -m localresolve
1059 $ hg commit -m localresolve
1062 $ cat m
1060 $ cat m
1063 $Id: m 800511b3a22d Thu, 01 Jan 1970 00:00:00 +0000 test $
1061 $Id: m 800511b3a22d Thu, 01 Jan 1970 00:00:00 +0000 test $
1064 bar
1062 bar
1065
1063
1066 Test restricted mode with transplant -b
1064 Test restricted mode with transplant -b
1067
1065
1068 $ hg update 6
1066 $ hg update 6
1069 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1067 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
1070 $ hg branch foo
1068 $ hg branch foo
1071 marked working directory as branch foo
1069 marked working directory as branch foo
1072 (branches are permanent and global, did you want a bookmark?)
1070 (branches are permanent and global, did you want a bookmark?)
1073 $ mv a a.bak
1071 $ mv a a.bak
1074 $ echo foobranch > a
1072 $ echo foobranch > a
1075 $ cat a.bak >> a
1073 $ cat a.bak >> a
1076 $ rm a.bak
1074 $ rm a.bak
1077 $ hg commit -m 9foobranch
1075 $ hg commit -m 9foobranch
1078 $ hg update default
1076 $ hg update default
1079 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
1077 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
1080 $ hg -y transplant -b foo tip
1078 $ hg -y transplant -b foo tip
1081 applying 4aa30d025d50
1079 applying 4aa30d025d50
1082 4aa30d025d50 transplanted to e00abbf63521
1080 4aa30d025d50 transplanted to e00abbf63521
1083
1081
1084 Expansion in changeset but not in file
1082 Expansion in changeset but not in file
1085
1083
1086 $ hg tip -p
1084 $ hg tip -p
1087 changeset: 11:e00abbf63521
1085 changeset: 11:e00abbf63521
1088 tag: tip
1086 tag: tip
1089 parent: 9:800511b3a22d
1087 parent: 9:800511b3a22d
1090 user: test
1088 user: test
1091 date: Thu Jan 01 00:00:00 1970 +0000
1089 date: Thu Jan 01 00:00:00 1970 +0000
1092 summary: 9foobranch
1090 summary: 9foobranch
1093
1091
1094 diff -r 800511b3a22d -r e00abbf63521 a
1092 diff -r 800511b3a22d -r e00abbf63521 a
1095 --- a/a Thu Jan 01 00:00:00 1970 +0000
1093 --- a/a Thu Jan 01 00:00:00 1970 +0000
1096 +++ b/a Thu Jan 01 00:00:00 1970 +0000
1094 +++ b/a Thu Jan 01 00:00:00 1970 +0000
1097 @@ -1,3 +1,4 @@
1095 @@ -1,3 +1,4 @@
1098 +foobranch
1096 +foobranch
1099 expand $Id$
1097 expand $Id$
1100 do not process $Id:
1098 do not process $Id:
1101 xxx $
1099 xxx $
1102
1100
1103 $ head -n 2 a
1101 $ head -n 2 a
1104 foobranch
1102 foobranch
1105 expand $Id: a e00abbf63521 Thu, 01 Jan 1970 00:00:00 +0000 test $
1103 expand $Id: a e00abbf63521 Thu, 01 Jan 1970 00:00:00 +0000 test $
1106
1104
1107 Turn off expansion
1105 Turn off expansion
1108
1106
1109 $ hg -q rollback
1107 $ hg -q rollback
1110 $ hg -q update -C
1108 $ hg -q update -C
1111
1109
1112 kwshrink with unknown file u
1110 kwshrink with unknown file u
1113
1111
1114 $ cp a u
1112 $ cp a u
1115 $ hg --verbose kwshrink
1113 $ hg --verbose kwshrink
1116 overwriting a shrinking keywords
1114 overwriting a shrinking keywords
1117 overwriting m shrinking keywords
1115 overwriting m shrinking keywords
1118 overwriting x/a shrinking keywords
1116 overwriting x/a shrinking keywords
1119
1117
1120 Keywords shrunk in working directory, but not yet disabled
1118 Keywords shrunk in working directory, but not yet disabled
1121 - cat shows unexpanded keywords
1119 - cat shows unexpanded keywords
1122 - hg cat shows expanded keywords
1120 - hg cat shows expanded keywords
1123
1121
1124 $ cat a b
1122 $ cat a b
1125 expand $Id$
1123 expand $Id$
1126 do not process $Id:
1124 do not process $Id:
1127 xxx $
1125 xxx $
1128 $Xinfo$
1126 $Xinfo$
1129 ignore $Id$
1127 ignore $Id$
1130 $ hg cat sym a b && echo
1128 $ hg cat sym a b && echo
1131 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
1129 expand $Id: a bb948857c743 Thu, 01 Jan 1970 00:00:02 +0000 user $
1132 do not process $Id:
1130 do not process $Id:
1133 xxx $
1131 xxx $
1134 $Xinfo: User Name <user@example.com>: firstline $
1132 $Xinfo: User Name <user@example.com>: firstline $
1135 ignore $Id$
1133 ignore $Id$
1136 a
1134 a
1137
1135
1138 Now disable keyword expansion
1136 Now disable keyword expansion
1139
1137
1140 $ rm "$HGRCPATH"
1138 $ rm "$HGRCPATH"
1141 $ cat a b
1139 $ cat a b
1142 expand $Id$
1140 expand $Id$
1143 do not process $Id:
1141 do not process $Id:
1144 xxx $
1142 xxx $
1145 $Xinfo$
1143 $Xinfo$
1146 ignore $Id$
1144 ignore $Id$
1147 $ hg cat sym a b && echo
1145 $ hg cat sym a b && echo
1148 expand $Id$
1146 expand $Id$
1149 do not process $Id:
1147 do not process $Id:
1150 xxx $
1148 xxx $
1151 $Xinfo$
1149 $Xinfo$
1152 ignore $Id$
1150 ignore $Id$
1153 a
1151 a
1154
1152
1155 $ cd ..
1153 $ cd ..
@@ -1,108 +1,109 b''
1 $ "$TESTDIR/hghave" symlink || exit 80
1 $ "$TESTDIR/hghave" symlink || exit 80
2
2
3 $ echo "[extensions]" >> $HGRCPATH
3 $ echo "[extensions]" >> $HGRCPATH
4 $ echo "mq=" >> $HGRCPATH
4 $ echo "mq=" >> $HGRCPATH
5
5
6 $ hg init
6 $ hg init
7 $ hg qinit
7 $ hg qinit
8 $ hg qnew base.patch
8 $ hg qnew base.patch
9 $ echo aaa > a
9 $ echo aaa > a
10 $ echo bbb > b
10 $ echo bbb > b
11 $ echo ccc > c
11 $ echo ccc > c
12 $ hg add a b c
12 $ hg add a b c
13 $ hg qrefresh
13 $ hg qrefresh
14 $ "$TESTDIR/readlink.py" a
14 $ "$TESTDIR/readlink.py" a
15 a -> a not a symlink
15 a -> a not a symlink
16
16
17
17
18 test replacing a file with a symlink
18 test replacing a file with a symlink
19
19
20 $ hg qnew symlink.patch
20 $ hg qnew symlink.patch
21 $ rm a
21 $ rm a
22 $ ln -s b a
22 $ ln -s b a
23 $ hg qrefresh --git
23 $ hg qrefresh --git
24 $ "$TESTDIR/readlink.py" a
24 $ "$TESTDIR/readlink.py" a
25 a -> b
25 a -> b
26
26
27 $ hg qpop
27 $ hg qpop
28 popping symlink.patch
28 popping symlink.patch
29 now at: base.patch
29 now at: base.patch
30 $ hg qpush
30 $ hg qpush
31 applying symlink.patch
31 applying symlink.patch
32 now at: symlink.patch
32 now at: symlink.patch
33 $ "$TESTDIR/readlink.py" a
33 $ "$TESTDIR/readlink.py" a
34 a -> b
34 a -> b
35
35
36
36
37 test updating a symlink
37 test updating a symlink
38
38
39 $ rm a
39 $ rm a
40 $ ln -s c a
40 $ ln -s c a
41 $ hg qnew --git -f updatelink
41 $ hg qnew --git -f updatelink
42 $ "$TESTDIR/readlink.py" a
42 $ "$TESTDIR/readlink.py" a
43 a -> c
43 a -> c
44 $ hg qpop
44 $ hg qpop
45 popping updatelink
45 popping updatelink
46 now at: symlink.patch
46 now at: symlink.patch
47 $ hg qpush --debug
47 $ hg qpush --debug
48 invalid branchheads cache (served): tip differs
48 applying updatelink
49 applying updatelink
49 patching file a
50 patching file a
50 a
51 a
51 now at: updatelink
52 now at: updatelink
52 $ "$TESTDIR/readlink.py" a
53 $ "$TESTDIR/readlink.py" a
53 a -> c
54 a -> c
54 $ hg st
55 $ hg st
55
56
56
57
57 test replacing a symlink with a file
58 test replacing a symlink with a file
58
59
59 $ ln -s c s
60 $ ln -s c s
60 $ hg add s
61 $ hg add s
61 $ hg qnew --git -f addlink
62 $ hg qnew --git -f addlink
62 $ rm s
63 $ rm s
63 $ echo sss > s
64 $ echo sss > s
64 $ hg qnew --git -f replacelinkwithfile
65 $ hg qnew --git -f replacelinkwithfile
65 $ hg qpop
66 $ hg qpop
66 popping replacelinkwithfile
67 popping replacelinkwithfile
67 now at: addlink
68 now at: addlink
68 $ hg qpush
69 $ hg qpush
69 applying replacelinkwithfile
70 applying replacelinkwithfile
70 now at: replacelinkwithfile
71 now at: replacelinkwithfile
71 $ cat s
72 $ cat s
72 sss
73 sss
73 $ hg st
74 $ hg st
74
75
75
76
76 test symlink removal
77 test symlink removal
77
78
78 $ hg qnew removesl.patch
79 $ hg qnew removesl.patch
79 $ hg rm a
80 $ hg rm a
80 $ hg qrefresh --git
81 $ hg qrefresh --git
81 $ hg qpop
82 $ hg qpop
82 popping removesl.patch
83 popping removesl.patch
83 now at: replacelinkwithfile
84 now at: replacelinkwithfile
84 $ hg qpush
85 $ hg qpush
85 applying removesl.patch
86 applying removesl.patch
86 now at: removesl.patch
87 now at: removesl.patch
87 $ hg st -c
88 $ hg st -c
88 C b
89 C b
89 C c
90 C c
90 C s
91 C s
91
92
92 replace broken symlink with another broken symlink
93 replace broken symlink with another broken symlink
93
94
94 $ ln -s linka linka
95 $ ln -s linka linka
95 $ hg add linka
96 $ hg add linka
96 $ hg qnew link
97 $ hg qnew link
97 $ hg mv linka linkb
98 $ hg mv linka linkb
98 $ rm linkb
99 $ rm linkb
99 $ ln -s linkb linkb
100 $ ln -s linkb linkb
100 $ hg qnew movelink
101 $ hg qnew movelink
101 $ hg qpop
102 $ hg qpop
102 popping movelink
103 popping movelink
103 now at: link
104 now at: link
104 $ hg qpush
105 $ hg qpush
105 applying movelink
106 applying movelink
106 now at: movelink
107 now at: movelink
107 $ "$TESTDIR/readlink.py" linkb
108 $ "$TESTDIR/readlink.py" linkb
108 linkb -> linkb
109 linkb -> linkb
@@ -1,503 +1,513 b''
1 $ hglog() { hg log --template "{rev} {phaseidx} {desc}\n" $*; }
1 $ hglog() { hg log --template "{rev} {phaseidx} {desc}\n" $*; }
2 $ mkcommit() {
2 $ mkcommit() {
3 > echo "$1" > "$1"
3 > echo "$1" > "$1"
4 > hg add "$1"
4 > hg add "$1"
5 > message="$1"
5 > message="$1"
6 > shift
6 > shift
7 > hg ci -m "$message" $*
7 > hg ci -m "$message" $*
8 > }
8 > }
9
9
10 $ hg init initialrepo
10 $ hg init initialrepo
11 $ cd initialrepo
11 $ cd initialrepo
12
12
13 Cannot change null revision phase
13 Cannot change null revision phase
14
14
15 $ hg phase --force --secret null
15 $ hg phase --force --secret null
16 abort: cannot change null revision phase
16 abort: cannot change null revision phase
17 [255]
17 [255]
18 $ hg phase null
18 $ hg phase null
19 -1: public
19 -1: public
20
20
21 $ mkcommit A
21 $ mkcommit A
22
22
23 New commit are draft by default
23 New commit are draft by default
24
24
25 $ hglog
25 $ hglog
26 0 1 A
26 0 1 A
27
27
28 Following commit are draft too
28 Following commit are draft too
29
29
30 $ mkcommit B
30 $ mkcommit B
31
31
32 $ hglog
32 $ hglog
33 1 1 B
33 1 1 B
34 0 1 A
34 0 1 A
35
35
36 Draft commit are properly created over public one:
36 Draft commit are properly created over public one:
37
37
38 $ hg phase --public .
38 $ hg phase --public .
39 $ hglog
39 $ hglog
40 1 0 B
40 1 0 B
41 0 0 A
41 0 0 A
42
42
43 $ mkcommit C
43 $ mkcommit C
44 $ mkcommit D
44 $ mkcommit D
45
45
46 $ hglog
46 $ hglog
47 3 1 D
47 3 1 D
48 2 1 C
48 2 1 C
49 1 0 B
49 1 0 B
50 0 0 A
50 0 0 A
51
51
52 Test creating changeset as secret
52 Test creating changeset as secret
53
53
54 $ mkcommit E --config phases.new-commit='secret'
54 $ mkcommit E --config phases.new-commit='secret'
55 $ hglog
55 $ hglog
56 4 2 E
56 4 2 E
57 3 1 D
57 3 1 D
58 2 1 C
58 2 1 C
59 1 0 B
59 1 0 B
60 0 0 A
60 0 0 A
61
61
62 Test the secret property is inherited
62 Test the secret property is inherited
63
63
64 $ mkcommit H
64 $ mkcommit H
65 $ hglog
65 $ hglog
66 5 2 H
66 5 2 H
67 4 2 E
67 4 2 E
68 3 1 D
68 3 1 D
69 2 1 C
69 2 1 C
70 1 0 B
70 1 0 B
71 0 0 A
71 0 0 A
72
72
73 Even on merge
73 Even on merge
74
74
75 $ hg up -q 1
75 $ hg up -q 1
76 $ mkcommit "B'"
76 $ mkcommit "B'"
77 created new head
77 created new head
78 $ hglog
78 $ hglog
79 6 1 B'
79 6 1 B'
80 5 2 H
80 5 2 H
81 4 2 E
81 4 2 E
82 3 1 D
82 3 1 D
83 2 1 C
83 2 1 C
84 1 0 B
84 1 0 B
85 0 0 A
85 0 0 A
86 $ hg merge 4 # E
86 $ hg merge 4 # E
87 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
87 3 files updated, 0 files merged, 0 files removed, 0 files unresolved
88 (branch merge, don't forget to commit)
88 (branch merge, don't forget to commit)
89 $ hg ci -m "merge B' and E"
89 $ hg ci -m "merge B' and E"
90 $ hglog
90 $ hglog
91 7 2 merge B' and E
91 7 2 merge B' and E
92 6 1 B'
92 6 1 B'
93 5 2 H
93 5 2 H
94 4 2 E
94 4 2 E
95 3 1 D
95 3 1 D
96 2 1 C
96 2 1 C
97 1 0 B
97 1 0 B
98 0 0 A
98 0 0 A
99
99
100 Test secret changeset are not pushed
100 Test secret changeset are not pushed
101
101
102 $ hg init ../push-dest
102 $ hg init ../push-dest
103 $ cat > ../push-dest/.hg/hgrc << EOF
103 $ cat > ../push-dest/.hg/hgrc << EOF
104 > [phases]
104 > [phases]
105 > publish=False
105 > publish=False
106 > EOF
106 > EOF
107 $ hg outgoing ../push-dest --template='{rev} {phase} {desc|firstline}\n'
107 $ hg outgoing ../push-dest --template='{rev} {phase} {desc|firstline}\n'
108 comparing with ../push-dest
108 comparing with ../push-dest
109 searching for changes
109 searching for changes
110 0 public A
110 0 public A
111 1 public B
111 1 public B
112 2 draft C
112 2 draft C
113 3 draft D
113 3 draft D
114 6 draft B'
114 6 draft B'
115 $ hg outgoing -r 'branch(default)' ../push-dest --template='{rev} {phase} {desc|firstline}\n'
115 $ hg outgoing -r 'branch(default)' ../push-dest --template='{rev} {phase} {desc|firstline}\n'
116 comparing with ../push-dest
116 comparing with ../push-dest
117 searching for changes
117 searching for changes
118 0 public A
118 0 public A
119 1 public B
119 1 public B
120 2 draft C
120 2 draft C
121 3 draft D
121 3 draft D
122 6 draft B'
122 6 draft B'
123
123
124 $ hg push ../push-dest -f # force because we push multiple heads
124 $ hg push ../push-dest -f # force because we push multiple heads
125 pushing to ../push-dest
125 pushing to ../push-dest
126 searching for changes
126 searching for changes
127 adding changesets
127 adding changesets
128 adding manifests
128 adding manifests
129 adding file changes
129 adding file changes
130 added 5 changesets with 5 changes to 5 files (+1 heads)
130 added 5 changesets with 5 changes to 5 files (+1 heads)
131 $ hglog
131 $ hglog
132 7 2 merge B' and E
132 7 2 merge B' and E
133 6 1 B'
133 6 1 B'
134 5 2 H
134 5 2 H
135 4 2 E
135 4 2 E
136 3 1 D
136 3 1 D
137 2 1 C
137 2 1 C
138 1 0 B
138 1 0 B
139 0 0 A
139 0 0 A
140 $ cd ../push-dest
140 $ cd ../push-dest
141 $ hglog
141 $ hglog
142 4 1 B'
142 4 1 B'
143 3 1 D
143 3 1 D
144 2 1 C
144 2 1 C
145 1 0 B
145 1 0 B
146 0 0 A
146 0 0 A
147
147
148 (Issue3303)
148 (Issue3303)
149 Check that remote secret changeset are ignore when checking creation of remote heads
149 Check that remote secret changeset are ignore when checking creation of remote heads
150
150
151 We add a secret head into the push destination. This secreat head shadow a
151 We add a secret head into the push destination. This secreat head shadow a
152 visible shared between the initial repo and the push destination.
152 visible shared between the initial repo and the push destination.
153
153
154 $ hg up -q 4 # B'
154 $ hg up -q 4 # B'
155 $ mkcommit Z --config phases.new-commit=secret
155 $ mkcommit Z --config phases.new-commit=secret
156 $ hg phase .
156 $ hg phase .
157 5: secret
157 5: secret
158
158
159 # We now try to push a new public changeset that descend from the common public
159 # We now try to push a new public changeset that descend from the common public
160 # head shadowed by the remote secret head.
160 # head shadowed by the remote secret head.
161
161
162 $ cd ../initialrepo
162 $ cd ../initialrepo
163 $ hg up -q 6 #B'
163 $ hg up -q 6 #B'
164 $ mkcommit I
164 $ mkcommit I
165 created new head
165 created new head
166 $ hg push ../push-dest
166 $ hg push ../push-dest
167 pushing to ../push-dest
167 pushing to ../push-dest
168 searching for changes
168 searching for changes
169 adding changesets
169 adding changesets
170 adding manifests
170 adding manifests
171 adding file changes
171 adding file changes
172 added 1 changesets with 1 changes to 1 files (+1 heads)
172 added 1 changesets with 1 changes to 1 files (+1 heads)
173
173
174 :note: The "(+1 heads)" is wrong as we do not had any visible head
174 :note: The "(+1 heads)" is wrong as we do not had any visible head
175
175
176 check that branch cache with "unserved" filter are properly computed and stored
176 check that branch cache with "unserved" filter are properly computed and stored
177
177
178 $ ls ../push-dest/.hg/cache/branchheads*
178 $ ls ../push-dest/.hg/cache/branchheads*
179 ../push-dest/.hg/cache/branchheads-served
179 ../push-dest/.hg/cache/branchheads-served
180 $ cat ../push-dest/.hg/cache/branchheads-served
181 6d6770faffce199f1fddd1cf87f6f026138cf061 6 465891ffab3c47a3c23792f7dc84156e19a90722
182 b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e default
183 6d6770faffce199f1fddd1cf87f6f026138cf061 default
184 $ hg heads -R ../push-dest --template '{rev}:{node} {phase}\n' #update visible cache too
185 6:6d6770faffce199f1fddd1cf87f6f026138cf061 draft
186 5:2713879da13d6eea1ff22b442a5a87cb31a7ce6a secret
187 3:b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e draft
188 $ ls ../push-dest/.hg/cache/branchheads*
189 ../push-dest/.hg/cache/branchheads-served
180 ../push-dest/.hg/cache/branchheads-visible
190 ../push-dest/.hg/cache/branchheads-visible
191 $ cat ../push-dest/.hg/cache/branchheads-served
192 6d6770faffce199f1fddd1cf87f6f026138cf061 6 465891ffab3c47a3c23792f7dc84156e19a90722
193 b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e default
194 6d6770faffce199f1fddd1cf87f6f026138cf061 default
181 $ cat ../push-dest/.hg/cache/branchheads-visible
195 $ cat ../push-dest/.hg/cache/branchheads-visible
182 6d6770faffce199f1fddd1cf87f6f026138cf061 6
196 6d6770faffce199f1fddd1cf87f6f026138cf061 6
183 b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e default
197 b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e default
184 2713879da13d6eea1ff22b442a5a87cb31a7ce6a default
198 2713879da13d6eea1ff22b442a5a87cb31a7ce6a default
185 6d6770faffce199f1fddd1cf87f6f026138cf061 default
199 6d6770faffce199f1fddd1cf87f6f026138cf061 default
186 $ cat ../push-dest/.hg/cache/branchheads-served
187 cf9fe039dfd67e829edf6522a45de057b5c86519 4
188 b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e default
189 cf9fe039dfd67e829edf6522a45de057b5c86519 default
190
200
191
201
192 Restore condition prior extra insertion.
202 Restore condition prior extra insertion.
193 $ hg -q --config extensions.mq= strip .
203 $ hg -q --config extensions.mq= strip .
194 $ hg up -q 7
204 $ hg up -q 7
195 $ cd ..
205 $ cd ..
196
206
197 Test secret changeset are not pull
207 Test secret changeset are not pull
198
208
199 $ hg init pull-dest
209 $ hg init pull-dest
200 $ cd pull-dest
210 $ cd pull-dest
201 $ hg pull ../initialrepo
211 $ hg pull ../initialrepo
202 pulling from ../initialrepo
212 pulling from ../initialrepo
203 requesting all changes
213 requesting all changes
204 adding changesets
214 adding changesets
205 adding manifests
215 adding manifests
206 adding file changes
216 adding file changes
207 added 5 changesets with 5 changes to 5 files (+1 heads)
217 added 5 changesets with 5 changes to 5 files (+1 heads)
208 (run 'hg heads' to see heads, 'hg merge' to merge)
218 (run 'hg heads' to see heads, 'hg merge' to merge)
209 $ hglog
219 $ hglog
210 4 0 B'
220 4 0 B'
211 3 0 D
221 3 0 D
212 2 0 C
222 2 0 C
213 1 0 B
223 1 0 B
214 0 0 A
224 0 0 A
215 $ cd ..
225 $ cd ..
216
226
217 But secret can still be bundled explicitly
227 But secret can still be bundled explicitly
218
228
219 $ cd initialrepo
229 $ cd initialrepo
220 $ hg bundle --base '4^' -r 'children(4)' ../secret-bundle.hg
230 $ hg bundle --base '4^' -r 'children(4)' ../secret-bundle.hg
221 4 changesets found
231 4 changesets found
222 $ cd ..
232 $ cd ..
223
233
224 Test secret changeset are not cloned
234 Test secret changeset are not cloned
225 (during local clone)
235 (during local clone)
226
236
227 $ hg clone -qU initialrepo clone-dest
237 $ hg clone -qU initialrepo clone-dest
228 $ hglog -R clone-dest
238 $ hglog -R clone-dest
229 4 0 B'
239 4 0 B'
230 3 0 D
240 3 0 D
231 2 0 C
241 2 0 C
232 1 0 B
242 1 0 B
233 0 0 A
243 0 0 A
234
244
235 Test revset
245 Test revset
236
246
237 $ cd initialrepo
247 $ cd initialrepo
238 $ hglog -r 'public()'
248 $ hglog -r 'public()'
239 0 0 A
249 0 0 A
240 1 0 B
250 1 0 B
241 $ hglog -r 'draft()'
251 $ hglog -r 'draft()'
242 2 1 C
252 2 1 C
243 3 1 D
253 3 1 D
244 6 1 B'
254 6 1 B'
245 $ hglog -r 'secret()'
255 $ hglog -r 'secret()'
246 4 2 E
256 4 2 E
247 5 2 H
257 5 2 H
248 7 2 merge B' and E
258 7 2 merge B' and E
249
259
250 test that phase are displayed in log at debug level
260 test that phase are displayed in log at debug level
251
261
252 $ hg log --debug
262 $ hg log --debug
253 changeset: 7:17a481b3bccb796c0521ae97903d81c52bfee4af
263 changeset: 7:17a481b3bccb796c0521ae97903d81c52bfee4af
254 tag: tip
264 tag: tip
255 phase: secret
265 phase: secret
256 parent: 6:cf9fe039dfd67e829edf6522a45de057b5c86519
266 parent: 6:cf9fe039dfd67e829edf6522a45de057b5c86519
257 parent: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
267 parent: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
258 manifest: 7:5e724ffacba267b2ab726c91fc8b650710deaaa8
268 manifest: 7:5e724ffacba267b2ab726c91fc8b650710deaaa8
259 user: test
269 user: test
260 date: Thu Jan 01 00:00:00 1970 +0000
270 date: Thu Jan 01 00:00:00 1970 +0000
261 files+: C D E
271 files+: C D E
262 extra: branch=default
272 extra: branch=default
263 description:
273 description:
264 merge B' and E
274 merge B' and E
265
275
266
276
267 changeset: 6:cf9fe039dfd67e829edf6522a45de057b5c86519
277 changeset: 6:cf9fe039dfd67e829edf6522a45de057b5c86519
268 phase: draft
278 phase: draft
269 parent: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
279 parent: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
270 parent: -1:0000000000000000000000000000000000000000
280 parent: -1:0000000000000000000000000000000000000000
271 manifest: 6:ab8bfef2392903058bf4ebb9e7746e8d7026b27a
281 manifest: 6:ab8bfef2392903058bf4ebb9e7746e8d7026b27a
272 user: test
282 user: test
273 date: Thu Jan 01 00:00:00 1970 +0000
283 date: Thu Jan 01 00:00:00 1970 +0000
274 files+: B'
284 files+: B'
275 extra: branch=default
285 extra: branch=default
276 description:
286 description:
277 B'
287 B'
278
288
279
289
280 changeset: 5:a030c6be5127abc010fcbff1851536552e6951a8
290 changeset: 5:a030c6be5127abc010fcbff1851536552e6951a8
281 phase: secret
291 phase: secret
282 parent: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
292 parent: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
283 parent: -1:0000000000000000000000000000000000000000
293 parent: -1:0000000000000000000000000000000000000000
284 manifest: 5:5c710aa854874fe3d5fa7192e77bdb314cc08b5a
294 manifest: 5:5c710aa854874fe3d5fa7192e77bdb314cc08b5a
285 user: test
295 user: test
286 date: Thu Jan 01 00:00:00 1970 +0000
296 date: Thu Jan 01 00:00:00 1970 +0000
287 files+: H
297 files+: H
288 extra: branch=default
298 extra: branch=default
289 description:
299 description:
290 H
300 H
291
301
292
302
293 changeset: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
303 changeset: 4:a603bfb5a83e312131cebcd05353c217d4d21dde
294 phase: secret
304 phase: secret
295 parent: 3:b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e
305 parent: 3:b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e
296 parent: -1:0000000000000000000000000000000000000000
306 parent: -1:0000000000000000000000000000000000000000
297 manifest: 4:7173fd1c27119750b959e3a0f47ed78abe75d6dc
307 manifest: 4:7173fd1c27119750b959e3a0f47ed78abe75d6dc
298 user: test
308 user: test
299 date: Thu Jan 01 00:00:00 1970 +0000
309 date: Thu Jan 01 00:00:00 1970 +0000
300 files+: E
310 files+: E
301 extra: branch=default
311 extra: branch=default
302 description:
312 description:
303 E
313 E
304
314
305
315
306 changeset: 3:b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e
316 changeset: 3:b3325c91a4d916bcc4cdc83ea3fe4ece46a42f6e
307 phase: draft
317 phase: draft
308 parent: 2:f838bfaca5c7226600ebcfd84f3c3c13a28d3757
318 parent: 2:f838bfaca5c7226600ebcfd84f3c3c13a28d3757
309 parent: -1:0000000000000000000000000000000000000000
319 parent: -1:0000000000000000000000000000000000000000
310 manifest: 3:6e1f4c47ecb533ffd0c8e52cdc88afb6cd39e20c
320 manifest: 3:6e1f4c47ecb533ffd0c8e52cdc88afb6cd39e20c
311 user: test
321 user: test
312 date: Thu Jan 01 00:00:00 1970 +0000
322 date: Thu Jan 01 00:00:00 1970 +0000
313 files+: D
323 files+: D
314 extra: branch=default
324 extra: branch=default
315 description:
325 description:
316 D
326 D
317
327
318
328
319 changeset: 2:f838bfaca5c7226600ebcfd84f3c3c13a28d3757
329 changeset: 2:f838bfaca5c7226600ebcfd84f3c3c13a28d3757
320 phase: draft
330 phase: draft
321 parent: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
331 parent: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
322 parent: -1:0000000000000000000000000000000000000000
332 parent: -1:0000000000000000000000000000000000000000
323 manifest: 2:66a5a01817fdf5239c273802b5b7618d051c89e4
333 manifest: 2:66a5a01817fdf5239c273802b5b7618d051c89e4
324 user: test
334 user: test
325 date: Thu Jan 01 00:00:00 1970 +0000
335 date: Thu Jan 01 00:00:00 1970 +0000
326 files+: C
336 files+: C
327 extra: branch=default
337 extra: branch=default
328 description:
338 description:
329 C
339 C
330
340
331
341
332 changeset: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
342 changeset: 1:27547f69f25460a52fff66ad004e58da7ad3fb56
333 parent: 0:4a2df7238c3b48766b5e22fafbb8a2f506ec8256
343 parent: 0:4a2df7238c3b48766b5e22fafbb8a2f506ec8256
334 parent: -1:0000000000000000000000000000000000000000
344 parent: -1:0000000000000000000000000000000000000000
335 manifest: 1:cb5cbbc1bfbf24cc34b9e8c16914e9caa2d2a7fd
345 manifest: 1:cb5cbbc1bfbf24cc34b9e8c16914e9caa2d2a7fd
336 user: test
346 user: test
337 date: Thu Jan 01 00:00:00 1970 +0000
347 date: Thu Jan 01 00:00:00 1970 +0000
338 files+: B
348 files+: B
339 extra: branch=default
349 extra: branch=default
340 description:
350 description:
341 B
351 B
342
352
343
353
344 changeset: 0:4a2df7238c3b48766b5e22fafbb8a2f506ec8256
354 changeset: 0:4a2df7238c3b48766b5e22fafbb8a2f506ec8256
345 parent: -1:0000000000000000000000000000000000000000
355 parent: -1:0000000000000000000000000000000000000000
346 parent: -1:0000000000000000000000000000000000000000
356 parent: -1:0000000000000000000000000000000000000000
347 manifest: 0:007d8c9d88841325f5c6b06371b35b4e8a2b1a83
357 manifest: 0:007d8c9d88841325f5c6b06371b35b4e8a2b1a83
348 user: test
358 user: test
349 date: Thu Jan 01 00:00:00 1970 +0000
359 date: Thu Jan 01 00:00:00 1970 +0000
350 files+: A
360 files+: A
351 extra: branch=default
361 extra: branch=default
352 description:
362 description:
353 A
363 A
354
364
355
365
356
366
357
367
358 (Issue3707)
368 (Issue3707)
359 test invalid phase name
369 test invalid phase name
360
370
361 $ mkcommit I --config phases.new-commit='babar'
371 $ mkcommit I --config phases.new-commit='babar'
362 transaction abort!
372 transaction abort!
363 rollback completed
373 rollback completed
364 abort: phases.new-commit: not a valid phase name ('babar')
374 abort: phases.new-commit: not a valid phase name ('babar')
365 [255]
375 [255]
366 Test phase command
376 Test phase command
367 ===================
377 ===================
368
378
369 initial picture
379 initial picture
370
380
371 $ cat >> $HGRCPATH << EOF
381 $ cat >> $HGRCPATH << EOF
372 > [extensions]
382 > [extensions]
373 > hgext.graphlog=
383 > hgext.graphlog=
374 > EOF
384 > EOF
375 $ hg log -G --template "{rev} {phase} {desc}\n"
385 $ hg log -G --template "{rev} {phase} {desc}\n"
376 @ 7 secret merge B' and E
386 @ 7 secret merge B' and E
377 |\
387 |\
378 | o 6 draft B'
388 | o 6 draft B'
379 | |
389 | |
380 +---o 5 secret H
390 +---o 5 secret H
381 | |
391 | |
382 o | 4 secret E
392 o | 4 secret E
383 | |
393 | |
384 o | 3 draft D
394 o | 3 draft D
385 | |
395 | |
386 o | 2 draft C
396 o | 2 draft C
387 |/
397 |/
388 o 1 public B
398 o 1 public B
389 |
399 |
390 o 0 public A
400 o 0 public A
391
401
392
402
393 display changesets phase
403 display changesets phase
394
404
395 (mixing -r and plain rev specification)
405 (mixing -r and plain rev specification)
396
406
397 $ hg phase 1::4 -r 7
407 $ hg phase 1::4 -r 7
398 1: public
408 1: public
399 2: draft
409 2: draft
400 3: draft
410 3: draft
401 4: secret
411 4: secret
402 7: secret
412 7: secret
403
413
404
414
405 move changeset forward
415 move changeset forward
406
416
407 (with -r option)
417 (with -r option)
408
418
409 $ hg phase --public -r 2
419 $ hg phase --public -r 2
410 $ hg log -G --template "{rev} {phase} {desc}\n"
420 $ hg log -G --template "{rev} {phase} {desc}\n"
411 @ 7 secret merge B' and E
421 @ 7 secret merge B' and E
412 |\
422 |\
413 | o 6 draft B'
423 | o 6 draft B'
414 | |
424 | |
415 +---o 5 secret H
425 +---o 5 secret H
416 | |
426 | |
417 o | 4 secret E
427 o | 4 secret E
418 | |
428 | |
419 o | 3 draft D
429 o | 3 draft D
420 | |
430 | |
421 o | 2 public C
431 o | 2 public C
422 |/
432 |/
423 o 1 public B
433 o 1 public B
424 |
434 |
425 o 0 public A
435 o 0 public A
426
436
427
437
428 move changeset backward
438 move changeset backward
429
439
430 (without -r option)
440 (without -r option)
431
441
432 $ hg phase --draft --force 2
442 $ hg phase --draft --force 2
433 $ hg log -G --template "{rev} {phase} {desc}\n"
443 $ hg log -G --template "{rev} {phase} {desc}\n"
434 @ 7 secret merge B' and E
444 @ 7 secret merge B' and E
435 |\
445 |\
436 | o 6 draft B'
446 | o 6 draft B'
437 | |
447 | |
438 +---o 5 secret H
448 +---o 5 secret H
439 | |
449 | |
440 o | 4 secret E
450 o | 4 secret E
441 | |
451 | |
442 o | 3 draft D
452 o | 3 draft D
443 | |
453 | |
444 o | 2 draft C
454 o | 2 draft C
445 |/
455 |/
446 o 1 public B
456 o 1 public B
447 |
457 |
448 o 0 public A
458 o 0 public A
449
459
450
460
451 move changeset forward and backward
461 move changeset forward and backward
452
462
453 $ hg phase --draft --force 1::4
463 $ hg phase --draft --force 1::4
454 $ hg log -G --template "{rev} {phase} {desc}\n"
464 $ hg log -G --template "{rev} {phase} {desc}\n"
455 @ 7 secret merge B' and E
465 @ 7 secret merge B' and E
456 |\
466 |\
457 | o 6 draft B'
467 | o 6 draft B'
458 | |
468 | |
459 +---o 5 secret H
469 +---o 5 secret H
460 | |
470 | |
461 o | 4 draft E
471 o | 4 draft E
462 | |
472 | |
463 o | 3 draft D
473 o | 3 draft D
464 | |
474 | |
465 o | 2 draft C
475 o | 2 draft C
466 |/
476 |/
467 o 1 draft B
477 o 1 draft B
468 |
478 |
469 o 0 public A
479 o 0 public A
470
480
471 test partial failure
481 test partial failure
472
482
473 $ hg phase --public 7
483 $ hg phase --public 7
474 $ hg phase --draft '5 or 7'
484 $ hg phase --draft '5 or 7'
475 cannot move 1 changesets to a more permissive phase, use --force
485 cannot move 1 changesets to a more permissive phase, use --force
476 phase changed for 1 changesets
486 phase changed for 1 changesets
477 [1]
487 [1]
478 $ hg log -G --template "{rev} {phase} {desc}\n"
488 $ hg log -G --template "{rev} {phase} {desc}\n"
479 @ 7 public merge B' and E
489 @ 7 public merge B' and E
480 |\
490 |\
481 | o 6 public B'
491 | o 6 public B'
482 | |
492 | |
483 +---o 5 draft H
493 +---o 5 draft H
484 | |
494 | |
485 o | 4 public E
495 o | 4 public E
486 | |
496 | |
487 o | 3 public D
497 o | 3 public D
488 | |
498 | |
489 o | 2 public C
499 o | 2 public C
490 |/
500 |/
491 o 1 public B
501 o 1 public B
492 |
502 |
493 o 0 public A
503 o 0 public A
494
504
495
505
496 test complete failure
506 test complete failure
497
507
498 $ hg phase --draft 7
508 $ hg phase --draft 7
499 cannot move 1 changesets to a more permissive phase, use --force
509 cannot move 1 changesets to a more permissive phase, use --force
500 no phases changed
510 no phases changed
501 [1]
511 [1]
502
512
503 $ cd ..
513 $ cd ..
General Comments 0
You need to be logged in to leave comments. Login now