##// END OF EJS Templates
dirstate: add begin/endparentchange to dirstate...
Durham Goode -
r22404:12bc7f06 default
parent child Browse files
Show More
@@ -1,893 +1,918 b''
1 # dirstate.py - working directory tracking for mercurial
1 # dirstate.py - working directory tracking for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7
7
8 from node import nullid
8 from node import nullid
9 from i18n import _
9 from i18n import _
10 import scmutil, util, ignore, osutil, parsers, encoding, pathutil
10 import scmutil, util, ignore, osutil, parsers, encoding, pathutil
11 import os, stat, errno, gc
11 import os, stat, errno, gc
12
12
13 propertycache = util.propertycache
13 propertycache = util.propertycache
14 filecache = scmutil.filecache
14 filecache = scmutil.filecache
15 _rangemask = 0x7fffffff
15 _rangemask = 0x7fffffff
16
16
17 dirstatetuple = parsers.dirstatetuple
17 dirstatetuple = parsers.dirstatetuple
18
18
19 class repocache(filecache):
19 class repocache(filecache):
20 """filecache for files in .hg/"""
20 """filecache for files in .hg/"""
21 def join(self, obj, fname):
21 def join(self, obj, fname):
22 return obj._opener.join(fname)
22 return obj._opener.join(fname)
23
23
24 class rootcache(filecache):
24 class rootcache(filecache):
25 """filecache for files in the repository root"""
25 """filecache for files in the repository root"""
26 def join(self, obj, fname):
26 def join(self, obj, fname):
27 return obj._join(fname)
27 return obj._join(fname)
28
28
29 class dirstate(object):
29 class dirstate(object):
30
30
31 def __init__(self, opener, ui, root, validate):
31 def __init__(self, opener, ui, root, validate):
32 '''Create a new dirstate object.
32 '''Create a new dirstate object.
33
33
34 opener is an open()-like callable that can be used to open the
34 opener is an open()-like callable that can be used to open the
35 dirstate file; root is the root of the directory tracked by
35 dirstate file; root is the root of the directory tracked by
36 the dirstate.
36 the dirstate.
37 '''
37 '''
38 self._opener = opener
38 self._opener = opener
39 self._validate = validate
39 self._validate = validate
40 self._root = root
40 self._root = root
41 self._rootdir = os.path.join(root, '')
41 self._rootdir = os.path.join(root, '')
42 self._dirty = False
42 self._dirty = False
43 self._dirtypl = False
43 self._dirtypl = False
44 self._lastnormaltime = 0
44 self._lastnormaltime = 0
45 self._ui = ui
45 self._ui = ui
46 self._filecache = {}
46 self._filecache = {}
47 self._parentwriters = 0
48
49 def beginparentchange(self):
50 '''Marks the beginning of a set of changes that involve changing
51 the dirstate parents. If there is an exception during this time,
52 the dirstate will not be written when the wlock is released. This
53 prevents writing an incoherent dirstate where the parent doesn't
54 match the contents.
55 '''
56 self._parentwriters += 1
57
58 def endparentchange(self):
59 '''Marks the end of a set of changes that involve changing the
60 dirstate parents. Once all parent changes have been marked done,
61 the wlock will be free to write the dirstate on release.
62 '''
63 if self._parentwriters > 0:
64 self._parentwriters -= 1
65
66 def pendingparentchange(self):
67 '''Returns true if the dirstate is in the middle of a set of changes
68 that modify the dirstate parent.
69 '''
70 return self._parentwriters > 0
47
71
48 @propertycache
72 @propertycache
49 def _map(self):
73 def _map(self):
50 '''Return the dirstate contents as a map from filename to
74 '''Return the dirstate contents as a map from filename to
51 (state, mode, size, time).'''
75 (state, mode, size, time).'''
52 self._read()
76 self._read()
53 return self._map
77 return self._map
54
78
55 @propertycache
79 @propertycache
56 def _copymap(self):
80 def _copymap(self):
57 self._read()
81 self._read()
58 return self._copymap
82 return self._copymap
59
83
60 @propertycache
84 @propertycache
61 def _foldmap(self):
85 def _foldmap(self):
62 f = {}
86 f = {}
63 for name, s in self._map.iteritems():
87 for name, s in self._map.iteritems():
64 if s[0] != 'r':
88 if s[0] != 'r':
65 f[util.normcase(name)] = name
89 f[util.normcase(name)] = name
66 for name in self._dirs:
90 for name in self._dirs:
67 f[util.normcase(name)] = name
91 f[util.normcase(name)] = name
68 f['.'] = '.' # prevents useless util.fspath() invocation
92 f['.'] = '.' # prevents useless util.fspath() invocation
69 return f
93 return f
70
94
71 @repocache('branch')
95 @repocache('branch')
72 def _branch(self):
96 def _branch(self):
73 try:
97 try:
74 return self._opener.read("branch").strip() or "default"
98 return self._opener.read("branch").strip() or "default"
75 except IOError, inst:
99 except IOError, inst:
76 if inst.errno != errno.ENOENT:
100 if inst.errno != errno.ENOENT:
77 raise
101 raise
78 return "default"
102 return "default"
79
103
80 @propertycache
104 @propertycache
81 def _pl(self):
105 def _pl(self):
82 try:
106 try:
83 fp = self._opener("dirstate")
107 fp = self._opener("dirstate")
84 st = fp.read(40)
108 st = fp.read(40)
85 fp.close()
109 fp.close()
86 l = len(st)
110 l = len(st)
87 if l == 40:
111 if l == 40:
88 return st[:20], st[20:40]
112 return st[:20], st[20:40]
89 elif l > 0 and l < 40:
113 elif l > 0 and l < 40:
90 raise util.Abort(_('working directory state appears damaged!'))
114 raise util.Abort(_('working directory state appears damaged!'))
91 except IOError, err:
115 except IOError, err:
92 if err.errno != errno.ENOENT:
116 if err.errno != errno.ENOENT:
93 raise
117 raise
94 return [nullid, nullid]
118 return [nullid, nullid]
95
119
96 @propertycache
120 @propertycache
97 def _dirs(self):
121 def _dirs(self):
98 return scmutil.dirs(self._map, 'r')
122 return scmutil.dirs(self._map, 'r')
99
123
100 def dirs(self):
124 def dirs(self):
101 return self._dirs
125 return self._dirs
102
126
103 @rootcache('.hgignore')
127 @rootcache('.hgignore')
104 def _ignore(self):
128 def _ignore(self):
105 files = [self._join('.hgignore')]
129 files = [self._join('.hgignore')]
106 for name, path in self._ui.configitems("ui"):
130 for name, path in self._ui.configitems("ui"):
107 if name == 'ignore' or name.startswith('ignore.'):
131 if name == 'ignore' or name.startswith('ignore.'):
108 files.append(util.expandpath(path))
132 files.append(util.expandpath(path))
109 return ignore.ignore(self._root, files, self._ui.warn)
133 return ignore.ignore(self._root, files, self._ui.warn)
110
134
111 @propertycache
135 @propertycache
112 def _slash(self):
136 def _slash(self):
113 return self._ui.configbool('ui', 'slash') and os.sep != '/'
137 return self._ui.configbool('ui', 'slash') and os.sep != '/'
114
138
115 @propertycache
139 @propertycache
116 def _checklink(self):
140 def _checklink(self):
117 return util.checklink(self._root)
141 return util.checklink(self._root)
118
142
119 @propertycache
143 @propertycache
120 def _checkexec(self):
144 def _checkexec(self):
121 return util.checkexec(self._root)
145 return util.checkexec(self._root)
122
146
123 @propertycache
147 @propertycache
124 def _checkcase(self):
148 def _checkcase(self):
125 return not util.checkcase(self._join('.hg'))
149 return not util.checkcase(self._join('.hg'))
126
150
127 def _join(self, f):
151 def _join(self, f):
128 # much faster than os.path.join()
152 # much faster than os.path.join()
129 # it's safe because f is always a relative path
153 # it's safe because f is always a relative path
130 return self._rootdir + f
154 return self._rootdir + f
131
155
132 def flagfunc(self, buildfallback):
156 def flagfunc(self, buildfallback):
133 if self._checklink and self._checkexec:
157 if self._checklink and self._checkexec:
134 def f(x):
158 def f(x):
135 try:
159 try:
136 st = os.lstat(self._join(x))
160 st = os.lstat(self._join(x))
137 if util.statislink(st):
161 if util.statislink(st):
138 return 'l'
162 return 'l'
139 if util.statisexec(st):
163 if util.statisexec(st):
140 return 'x'
164 return 'x'
141 except OSError:
165 except OSError:
142 pass
166 pass
143 return ''
167 return ''
144 return f
168 return f
145
169
146 fallback = buildfallback()
170 fallback = buildfallback()
147 if self._checklink:
171 if self._checklink:
148 def f(x):
172 def f(x):
149 if os.path.islink(self._join(x)):
173 if os.path.islink(self._join(x)):
150 return 'l'
174 return 'l'
151 if 'x' in fallback(x):
175 if 'x' in fallback(x):
152 return 'x'
176 return 'x'
153 return ''
177 return ''
154 return f
178 return f
155 if self._checkexec:
179 if self._checkexec:
156 def f(x):
180 def f(x):
157 if 'l' in fallback(x):
181 if 'l' in fallback(x):
158 return 'l'
182 return 'l'
159 if util.isexec(self._join(x)):
183 if util.isexec(self._join(x)):
160 return 'x'
184 return 'x'
161 return ''
185 return ''
162 return f
186 return f
163 else:
187 else:
164 return fallback
188 return fallback
165
189
166 @propertycache
190 @propertycache
167 def _cwd(self):
191 def _cwd(self):
168 return os.getcwd()
192 return os.getcwd()
169
193
170 def getcwd(self):
194 def getcwd(self):
171 cwd = self._cwd
195 cwd = self._cwd
172 if cwd == self._root:
196 if cwd == self._root:
173 return ''
197 return ''
174 # self._root ends with a path separator if self._root is '/' or 'C:\'
198 # self._root ends with a path separator if self._root is '/' or 'C:\'
175 rootsep = self._root
199 rootsep = self._root
176 if not util.endswithsep(rootsep):
200 if not util.endswithsep(rootsep):
177 rootsep += os.sep
201 rootsep += os.sep
178 if cwd.startswith(rootsep):
202 if cwd.startswith(rootsep):
179 return cwd[len(rootsep):]
203 return cwd[len(rootsep):]
180 else:
204 else:
181 # we're outside the repo. return an absolute path.
205 # we're outside the repo. return an absolute path.
182 return cwd
206 return cwd
183
207
184 def pathto(self, f, cwd=None):
208 def pathto(self, f, cwd=None):
185 if cwd is None:
209 if cwd is None:
186 cwd = self.getcwd()
210 cwd = self.getcwd()
187 path = util.pathto(self._root, cwd, f)
211 path = util.pathto(self._root, cwd, f)
188 if self._slash:
212 if self._slash:
189 return util.pconvert(path)
213 return util.pconvert(path)
190 return path
214 return path
191
215
192 def __getitem__(self, key):
216 def __getitem__(self, key):
193 '''Return the current state of key (a filename) in the dirstate.
217 '''Return the current state of key (a filename) in the dirstate.
194
218
195 States are:
219 States are:
196 n normal
220 n normal
197 m needs merging
221 m needs merging
198 r marked for removal
222 r marked for removal
199 a marked for addition
223 a marked for addition
200 ? not tracked
224 ? not tracked
201 '''
225 '''
202 return self._map.get(key, ("?",))[0]
226 return self._map.get(key, ("?",))[0]
203
227
204 def __contains__(self, key):
228 def __contains__(self, key):
205 return key in self._map
229 return key in self._map
206
230
207 def __iter__(self):
231 def __iter__(self):
208 for x in sorted(self._map):
232 for x in sorted(self._map):
209 yield x
233 yield x
210
234
211 def iteritems(self):
235 def iteritems(self):
212 return self._map.iteritems()
236 return self._map.iteritems()
213
237
214 def parents(self):
238 def parents(self):
215 return [self._validate(p) for p in self._pl]
239 return [self._validate(p) for p in self._pl]
216
240
217 def p1(self):
241 def p1(self):
218 return self._validate(self._pl[0])
242 return self._validate(self._pl[0])
219
243
220 def p2(self):
244 def p2(self):
221 return self._validate(self._pl[1])
245 return self._validate(self._pl[1])
222
246
223 def branch(self):
247 def branch(self):
224 return encoding.tolocal(self._branch)
248 return encoding.tolocal(self._branch)
225
249
226 def setparents(self, p1, p2=nullid):
250 def setparents(self, p1, p2=nullid):
227 """Set dirstate parents to p1 and p2.
251 """Set dirstate parents to p1 and p2.
228
252
229 When moving from two parents to one, 'm' merged entries a
253 When moving from two parents to one, 'm' merged entries a
230 adjusted to normal and previous copy records discarded and
254 adjusted to normal and previous copy records discarded and
231 returned by the call.
255 returned by the call.
232
256
233 See localrepo.setparents()
257 See localrepo.setparents()
234 """
258 """
235 self._dirty = self._dirtypl = True
259 self._dirty = self._dirtypl = True
236 oldp2 = self._pl[1]
260 oldp2 = self._pl[1]
237 self._pl = p1, p2
261 self._pl = p1, p2
238 copies = {}
262 copies = {}
239 if oldp2 != nullid and p2 == nullid:
263 if oldp2 != nullid and p2 == nullid:
240 # Discard 'm' markers when moving away from a merge state
264 # Discard 'm' markers when moving away from a merge state
241 for f, s in self._map.iteritems():
265 for f, s in self._map.iteritems():
242 if s[0] == 'm':
266 if s[0] == 'm':
243 if f in self._copymap:
267 if f in self._copymap:
244 copies[f] = self._copymap[f]
268 copies[f] = self._copymap[f]
245 self.normallookup(f)
269 self.normallookup(f)
246 return copies
270 return copies
247
271
248 def setbranch(self, branch):
272 def setbranch(self, branch):
249 self._branch = encoding.fromlocal(branch)
273 self._branch = encoding.fromlocal(branch)
250 f = self._opener('branch', 'w', atomictemp=True)
274 f = self._opener('branch', 'w', atomictemp=True)
251 try:
275 try:
252 f.write(self._branch + '\n')
276 f.write(self._branch + '\n')
253 f.close()
277 f.close()
254
278
255 # make sure filecache has the correct stat info for _branch after
279 # make sure filecache has the correct stat info for _branch after
256 # replacing the underlying file
280 # replacing the underlying file
257 ce = self._filecache['_branch']
281 ce = self._filecache['_branch']
258 if ce:
282 if ce:
259 ce.refresh()
283 ce.refresh()
260 except: # re-raises
284 except: # re-raises
261 f.discard()
285 f.discard()
262 raise
286 raise
263
287
264 def _read(self):
288 def _read(self):
265 self._map = {}
289 self._map = {}
266 self._copymap = {}
290 self._copymap = {}
267 try:
291 try:
268 st = self._opener.read("dirstate")
292 st = self._opener.read("dirstate")
269 except IOError, err:
293 except IOError, err:
270 if err.errno != errno.ENOENT:
294 if err.errno != errno.ENOENT:
271 raise
295 raise
272 return
296 return
273 if not st:
297 if not st:
274 return
298 return
275
299
276 # Python's garbage collector triggers a GC each time a certain number
300 # Python's garbage collector triggers a GC each time a certain number
277 # of container objects (the number being defined by
301 # of container objects (the number being defined by
278 # gc.get_threshold()) are allocated. parse_dirstate creates a tuple
302 # gc.get_threshold()) are allocated. parse_dirstate creates a tuple
279 # for each file in the dirstate. The C version then immediately marks
303 # for each file in the dirstate. The C version then immediately marks
280 # them as not to be tracked by the collector. However, this has no
304 # them as not to be tracked by the collector. However, this has no
281 # effect on when GCs are triggered, only on what objects the GC looks
305 # effect on when GCs are triggered, only on what objects the GC looks
282 # into. This means that O(number of files) GCs are unavoidable.
306 # into. This means that O(number of files) GCs are unavoidable.
283 # Depending on when in the process's lifetime the dirstate is parsed,
307 # Depending on when in the process's lifetime the dirstate is parsed,
284 # this can get very expensive. As a workaround, disable GC while
308 # this can get very expensive. As a workaround, disable GC while
285 # parsing the dirstate.
309 # parsing the dirstate.
286 gcenabled = gc.isenabled()
310 gcenabled = gc.isenabled()
287 gc.disable()
311 gc.disable()
288 try:
312 try:
289 p = parsers.parse_dirstate(self._map, self._copymap, st)
313 p = parsers.parse_dirstate(self._map, self._copymap, st)
290 finally:
314 finally:
291 if gcenabled:
315 if gcenabled:
292 gc.enable()
316 gc.enable()
293 if not self._dirtypl:
317 if not self._dirtypl:
294 self._pl = p
318 self._pl = p
295
319
296 def invalidate(self):
320 def invalidate(self):
297 for a in ("_map", "_copymap", "_foldmap", "_branch", "_pl", "_dirs",
321 for a in ("_map", "_copymap", "_foldmap", "_branch", "_pl", "_dirs",
298 "_ignore"):
322 "_ignore"):
299 if a in self.__dict__:
323 if a in self.__dict__:
300 delattr(self, a)
324 delattr(self, a)
301 self._lastnormaltime = 0
325 self._lastnormaltime = 0
302 self._dirty = False
326 self._dirty = False
327 self._parentwriters = 0
303
328
304 def copy(self, source, dest):
329 def copy(self, source, dest):
305 """Mark dest as a copy of source. Unmark dest if source is None."""
330 """Mark dest as a copy of source. Unmark dest if source is None."""
306 if source == dest:
331 if source == dest:
307 return
332 return
308 self._dirty = True
333 self._dirty = True
309 if source is not None:
334 if source is not None:
310 self._copymap[dest] = source
335 self._copymap[dest] = source
311 elif dest in self._copymap:
336 elif dest in self._copymap:
312 del self._copymap[dest]
337 del self._copymap[dest]
313
338
314 def copied(self, file):
339 def copied(self, file):
315 return self._copymap.get(file, None)
340 return self._copymap.get(file, None)
316
341
317 def copies(self):
342 def copies(self):
318 return self._copymap
343 return self._copymap
319
344
320 def _droppath(self, f):
345 def _droppath(self, f):
321 if self[f] not in "?r" and "_dirs" in self.__dict__:
346 if self[f] not in "?r" and "_dirs" in self.__dict__:
322 self._dirs.delpath(f)
347 self._dirs.delpath(f)
323
348
324 def _addpath(self, f, state, mode, size, mtime):
349 def _addpath(self, f, state, mode, size, mtime):
325 oldstate = self[f]
350 oldstate = self[f]
326 if state == 'a' or oldstate == 'r':
351 if state == 'a' or oldstate == 'r':
327 scmutil.checkfilename(f)
352 scmutil.checkfilename(f)
328 if f in self._dirs:
353 if f in self._dirs:
329 raise util.Abort(_('directory %r already in dirstate') % f)
354 raise util.Abort(_('directory %r already in dirstate') % f)
330 # shadows
355 # shadows
331 for d in scmutil.finddirs(f):
356 for d in scmutil.finddirs(f):
332 if d in self._dirs:
357 if d in self._dirs:
333 break
358 break
334 if d in self._map and self[d] != 'r':
359 if d in self._map and self[d] != 'r':
335 raise util.Abort(
360 raise util.Abort(
336 _('file %r in dirstate clashes with %r') % (d, f))
361 _('file %r in dirstate clashes with %r') % (d, f))
337 if oldstate in "?r" and "_dirs" in self.__dict__:
362 if oldstate in "?r" and "_dirs" in self.__dict__:
338 self._dirs.addpath(f)
363 self._dirs.addpath(f)
339 self._dirty = True
364 self._dirty = True
340 self._map[f] = dirstatetuple(state, mode, size, mtime)
365 self._map[f] = dirstatetuple(state, mode, size, mtime)
341
366
342 def normal(self, f):
367 def normal(self, f):
343 '''Mark a file normal and clean.'''
368 '''Mark a file normal and clean.'''
344 s = os.lstat(self._join(f))
369 s = os.lstat(self._join(f))
345 mtime = int(s.st_mtime)
370 mtime = int(s.st_mtime)
346 self._addpath(f, 'n', s.st_mode,
371 self._addpath(f, 'n', s.st_mode,
347 s.st_size & _rangemask, mtime & _rangemask)
372 s.st_size & _rangemask, mtime & _rangemask)
348 if f in self._copymap:
373 if f in self._copymap:
349 del self._copymap[f]
374 del self._copymap[f]
350 if mtime > self._lastnormaltime:
375 if mtime > self._lastnormaltime:
351 # Remember the most recent modification timeslot for status(),
376 # Remember the most recent modification timeslot for status(),
352 # to make sure we won't miss future size-preserving file content
377 # to make sure we won't miss future size-preserving file content
353 # modifications that happen within the same timeslot.
378 # modifications that happen within the same timeslot.
354 self._lastnormaltime = mtime
379 self._lastnormaltime = mtime
355
380
356 def normallookup(self, f):
381 def normallookup(self, f):
357 '''Mark a file normal, but possibly dirty.'''
382 '''Mark a file normal, but possibly dirty.'''
358 if self._pl[1] != nullid and f in self._map:
383 if self._pl[1] != nullid and f in self._map:
359 # if there is a merge going on and the file was either
384 # if there is a merge going on and the file was either
360 # in state 'm' (-1) or coming from other parent (-2) before
385 # in state 'm' (-1) or coming from other parent (-2) before
361 # being removed, restore that state.
386 # being removed, restore that state.
362 entry = self._map[f]
387 entry = self._map[f]
363 if entry[0] == 'r' and entry[2] in (-1, -2):
388 if entry[0] == 'r' and entry[2] in (-1, -2):
364 source = self._copymap.get(f)
389 source = self._copymap.get(f)
365 if entry[2] == -1:
390 if entry[2] == -1:
366 self.merge(f)
391 self.merge(f)
367 elif entry[2] == -2:
392 elif entry[2] == -2:
368 self.otherparent(f)
393 self.otherparent(f)
369 if source:
394 if source:
370 self.copy(source, f)
395 self.copy(source, f)
371 return
396 return
372 if entry[0] == 'm' or entry[0] == 'n' and entry[2] == -2:
397 if entry[0] == 'm' or entry[0] == 'n' and entry[2] == -2:
373 return
398 return
374 self._addpath(f, 'n', 0, -1, -1)
399 self._addpath(f, 'n', 0, -1, -1)
375 if f in self._copymap:
400 if f in self._copymap:
376 del self._copymap[f]
401 del self._copymap[f]
377
402
378 def otherparent(self, f):
403 def otherparent(self, f):
379 '''Mark as coming from the other parent, always dirty.'''
404 '''Mark as coming from the other parent, always dirty.'''
380 if self._pl[1] == nullid:
405 if self._pl[1] == nullid:
381 raise util.Abort(_("setting %r to other parent "
406 raise util.Abort(_("setting %r to other parent "
382 "only allowed in merges") % f)
407 "only allowed in merges") % f)
383 self._addpath(f, 'n', 0, -2, -1)
408 self._addpath(f, 'n', 0, -2, -1)
384 if f in self._copymap:
409 if f in self._copymap:
385 del self._copymap[f]
410 del self._copymap[f]
386
411
387 def add(self, f):
412 def add(self, f):
388 '''Mark a file added.'''
413 '''Mark a file added.'''
389 self._addpath(f, 'a', 0, -1, -1)
414 self._addpath(f, 'a', 0, -1, -1)
390 if f in self._copymap:
415 if f in self._copymap:
391 del self._copymap[f]
416 del self._copymap[f]
392
417
393 def remove(self, f):
418 def remove(self, f):
394 '''Mark a file removed.'''
419 '''Mark a file removed.'''
395 self._dirty = True
420 self._dirty = True
396 self._droppath(f)
421 self._droppath(f)
397 size = 0
422 size = 0
398 if self._pl[1] != nullid and f in self._map:
423 if self._pl[1] != nullid and f in self._map:
399 # backup the previous state
424 # backup the previous state
400 entry = self._map[f]
425 entry = self._map[f]
401 if entry[0] == 'm': # merge
426 if entry[0] == 'm': # merge
402 size = -1
427 size = -1
403 elif entry[0] == 'n' and entry[2] == -2: # other parent
428 elif entry[0] == 'n' and entry[2] == -2: # other parent
404 size = -2
429 size = -2
405 self._map[f] = dirstatetuple('r', 0, size, 0)
430 self._map[f] = dirstatetuple('r', 0, size, 0)
406 if size == 0 and f in self._copymap:
431 if size == 0 and f in self._copymap:
407 del self._copymap[f]
432 del self._copymap[f]
408
433
409 def merge(self, f):
434 def merge(self, f):
410 '''Mark a file merged.'''
435 '''Mark a file merged.'''
411 if self._pl[1] == nullid:
436 if self._pl[1] == nullid:
412 return self.normallookup(f)
437 return self.normallookup(f)
413 s = os.lstat(self._join(f))
438 s = os.lstat(self._join(f))
414 self._addpath(f, 'm', s.st_mode,
439 self._addpath(f, 'm', s.st_mode,
415 s.st_size & _rangemask, int(s.st_mtime) & _rangemask)
440 s.st_size & _rangemask, int(s.st_mtime) & _rangemask)
416 if f in self._copymap:
441 if f in self._copymap:
417 del self._copymap[f]
442 del self._copymap[f]
418
443
419 def drop(self, f):
444 def drop(self, f):
420 '''Drop a file from the dirstate'''
445 '''Drop a file from the dirstate'''
421 if f in self._map:
446 if f in self._map:
422 self._dirty = True
447 self._dirty = True
423 self._droppath(f)
448 self._droppath(f)
424 del self._map[f]
449 del self._map[f]
425
450
426 def _normalize(self, path, isknown, ignoremissing=False, exists=None):
451 def _normalize(self, path, isknown, ignoremissing=False, exists=None):
427 normed = util.normcase(path)
452 normed = util.normcase(path)
428 folded = self._foldmap.get(normed, None)
453 folded = self._foldmap.get(normed, None)
429 if folded is None:
454 if folded is None:
430 if isknown:
455 if isknown:
431 folded = path
456 folded = path
432 else:
457 else:
433 if exists is None:
458 if exists is None:
434 exists = os.path.lexists(os.path.join(self._root, path))
459 exists = os.path.lexists(os.path.join(self._root, path))
435 if not exists:
460 if not exists:
436 # Maybe a path component exists
461 # Maybe a path component exists
437 if not ignoremissing and '/' in path:
462 if not ignoremissing and '/' in path:
438 d, f = path.rsplit('/', 1)
463 d, f = path.rsplit('/', 1)
439 d = self._normalize(d, isknown, ignoremissing, None)
464 d = self._normalize(d, isknown, ignoremissing, None)
440 folded = d + "/" + f
465 folded = d + "/" + f
441 else:
466 else:
442 # No path components, preserve original case
467 # No path components, preserve original case
443 folded = path
468 folded = path
444 else:
469 else:
445 # recursively normalize leading directory components
470 # recursively normalize leading directory components
446 # against dirstate
471 # against dirstate
447 if '/' in normed:
472 if '/' in normed:
448 d, f = normed.rsplit('/', 1)
473 d, f = normed.rsplit('/', 1)
449 d = self._normalize(d, isknown, ignoremissing, True)
474 d = self._normalize(d, isknown, ignoremissing, True)
450 r = self._root + "/" + d
475 r = self._root + "/" + d
451 folded = d + "/" + util.fspath(f, r)
476 folded = d + "/" + util.fspath(f, r)
452 else:
477 else:
453 folded = util.fspath(normed, self._root)
478 folded = util.fspath(normed, self._root)
454 self._foldmap[normed] = folded
479 self._foldmap[normed] = folded
455
480
456 return folded
481 return folded
457
482
458 def normalize(self, path, isknown=False, ignoremissing=False):
483 def normalize(self, path, isknown=False, ignoremissing=False):
459 '''
484 '''
460 normalize the case of a pathname when on a casefolding filesystem
485 normalize the case of a pathname when on a casefolding filesystem
461
486
462 isknown specifies whether the filename came from walking the
487 isknown specifies whether the filename came from walking the
463 disk, to avoid extra filesystem access.
488 disk, to avoid extra filesystem access.
464
489
465 If ignoremissing is True, missing path are returned
490 If ignoremissing is True, missing path are returned
466 unchanged. Otherwise, we try harder to normalize possibly
491 unchanged. Otherwise, we try harder to normalize possibly
467 existing path components.
492 existing path components.
468
493
469 The normalized case is determined based on the following precedence:
494 The normalized case is determined based on the following precedence:
470
495
471 - version of name already stored in the dirstate
496 - version of name already stored in the dirstate
472 - version of name stored on disk
497 - version of name stored on disk
473 - version provided via command arguments
498 - version provided via command arguments
474 '''
499 '''
475
500
476 if self._checkcase:
501 if self._checkcase:
477 return self._normalize(path, isknown, ignoremissing)
502 return self._normalize(path, isknown, ignoremissing)
478 return path
503 return path
479
504
480 def clear(self):
505 def clear(self):
481 self._map = {}
506 self._map = {}
482 if "_dirs" in self.__dict__:
507 if "_dirs" in self.__dict__:
483 delattr(self, "_dirs")
508 delattr(self, "_dirs")
484 self._copymap = {}
509 self._copymap = {}
485 self._pl = [nullid, nullid]
510 self._pl = [nullid, nullid]
486 self._lastnormaltime = 0
511 self._lastnormaltime = 0
487 self._dirty = True
512 self._dirty = True
488
513
489 def rebuild(self, parent, allfiles, changedfiles=None):
514 def rebuild(self, parent, allfiles, changedfiles=None):
490 changedfiles = changedfiles or allfiles
515 changedfiles = changedfiles or allfiles
491 oldmap = self._map
516 oldmap = self._map
492 self.clear()
517 self.clear()
493 for f in allfiles:
518 for f in allfiles:
494 if f not in changedfiles:
519 if f not in changedfiles:
495 self._map[f] = oldmap[f]
520 self._map[f] = oldmap[f]
496 else:
521 else:
497 if 'x' in allfiles.flags(f):
522 if 'x' in allfiles.flags(f):
498 self._map[f] = dirstatetuple('n', 0777, -1, 0)
523 self._map[f] = dirstatetuple('n', 0777, -1, 0)
499 else:
524 else:
500 self._map[f] = dirstatetuple('n', 0666, -1, 0)
525 self._map[f] = dirstatetuple('n', 0666, -1, 0)
501 self._pl = (parent, nullid)
526 self._pl = (parent, nullid)
502 self._dirty = True
527 self._dirty = True
503
528
504 def write(self):
529 def write(self):
505 if not self._dirty:
530 if not self._dirty:
506 return
531 return
507
532
508 # enough 'delaywrite' prevents 'pack_dirstate' from dropping
533 # enough 'delaywrite' prevents 'pack_dirstate' from dropping
509 # timestamp of each entries in dirstate, because of 'now > mtime'
534 # timestamp of each entries in dirstate, because of 'now > mtime'
510 delaywrite = self._ui.configint('debug', 'dirstate.delaywrite', 0)
535 delaywrite = self._ui.configint('debug', 'dirstate.delaywrite', 0)
511 if delaywrite:
536 if delaywrite:
512 import time # to avoid useless import
537 import time # to avoid useless import
513 time.sleep(delaywrite)
538 time.sleep(delaywrite)
514
539
515 st = self._opener("dirstate", "w", atomictemp=True)
540 st = self._opener("dirstate", "w", atomictemp=True)
516 # use the modification time of the newly created temporary file as the
541 # use the modification time of the newly created temporary file as the
517 # filesystem's notion of 'now'
542 # filesystem's notion of 'now'
518 now = util.fstat(st).st_mtime
543 now = util.fstat(st).st_mtime
519 st.write(parsers.pack_dirstate(self._map, self._copymap, self._pl, now))
544 st.write(parsers.pack_dirstate(self._map, self._copymap, self._pl, now))
520 st.close()
545 st.close()
521 self._lastnormaltime = 0
546 self._lastnormaltime = 0
522 self._dirty = self._dirtypl = False
547 self._dirty = self._dirtypl = False
523
548
524 def _dirignore(self, f):
549 def _dirignore(self, f):
525 if f == '.':
550 if f == '.':
526 return False
551 return False
527 if self._ignore(f):
552 if self._ignore(f):
528 return True
553 return True
529 for p in scmutil.finddirs(f):
554 for p in scmutil.finddirs(f):
530 if self._ignore(p):
555 if self._ignore(p):
531 return True
556 return True
532 return False
557 return False
533
558
534 def _walkexplicit(self, match, subrepos):
559 def _walkexplicit(self, match, subrepos):
535 '''Get stat data about the files explicitly specified by match.
560 '''Get stat data about the files explicitly specified by match.
536
561
537 Return a triple (results, dirsfound, dirsnotfound).
562 Return a triple (results, dirsfound, dirsnotfound).
538 - results is a mapping from filename to stat result. It also contains
563 - results is a mapping from filename to stat result. It also contains
539 listings mapping subrepos and .hg to None.
564 listings mapping subrepos and .hg to None.
540 - dirsfound is a list of files found to be directories.
565 - dirsfound is a list of files found to be directories.
541 - dirsnotfound is a list of files that the dirstate thinks are
566 - dirsnotfound is a list of files that the dirstate thinks are
542 directories and that were not found.'''
567 directories and that were not found.'''
543
568
544 def badtype(mode):
569 def badtype(mode):
545 kind = _('unknown')
570 kind = _('unknown')
546 if stat.S_ISCHR(mode):
571 if stat.S_ISCHR(mode):
547 kind = _('character device')
572 kind = _('character device')
548 elif stat.S_ISBLK(mode):
573 elif stat.S_ISBLK(mode):
549 kind = _('block device')
574 kind = _('block device')
550 elif stat.S_ISFIFO(mode):
575 elif stat.S_ISFIFO(mode):
551 kind = _('fifo')
576 kind = _('fifo')
552 elif stat.S_ISSOCK(mode):
577 elif stat.S_ISSOCK(mode):
553 kind = _('socket')
578 kind = _('socket')
554 elif stat.S_ISDIR(mode):
579 elif stat.S_ISDIR(mode):
555 kind = _('directory')
580 kind = _('directory')
556 return _('unsupported file type (type is %s)') % kind
581 return _('unsupported file type (type is %s)') % kind
557
582
558 matchedir = match.explicitdir
583 matchedir = match.explicitdir
559 badfn = match.bad
584 badfn = match.bad
560 dmap = self._map
585 dmap = self._map
561 normpath = util.normpath
586 normpath = util.normpath
562 lstat = os.lstat
587 lstat = os.lstat
563 getkind = stat.S_IFMT
588 getkind = stat.S_IFMT
564 dirkind = stat.S_IFDIR
589 dirkind = stat.S_IFDIR
565 regkind = stat.S_IFREG
590 regkind = stat.S_IFREG
566 lnkkind = stat.S_IFLNK
591 lnkkind = stat.S_IFLNK
567 join = self._join
592 join = self._join
568 dirsfound = []
593 dirsfound = []
569 foundadd = dirsfound.append
594 foundadd = dirsfound.append
570 dirsnotfound = []
595 dirsnotfound = []
571 notfoundadd = dirsnotfound.append
596 notfoundadd = dirsnotfound.append
572
597
573 if match.matchfn != match.exact and self._checkcase:
598 if match.matchfn != match.exact and self._checkcase:
574 normalize = self._normalize
599 normalize = self._normalize
575 else:
600 else:
576 normalize = None
601 normalize = None
577
602
578 files = sorted(match.files())
603 files = sorted(match.files())
579 subrepos.sort()
604 subrepos.sort()
580 i, j = 0, 0
605 i, j = 0, 0
581 while i < len(files) and j < len(subrepos):
606 while i < len(files) and j < len(subrepos):
582 subpath = subrepos[j] + "/"
607 subpath = subrepos[j] + "/"
583 if files[i] < subpath:
608 if files[i] < subpath:
584 i += 1
609 i += 1
585 continue
610 continue
586 while i < len(files) and files[i].startswith(subpath):
611 while i < len(files) and files[i].startswith(subpath):
587 del files[i]
612 del files[i]
588 j += 1
613 j += 1
589
614
590 if not files or '.' in files:
615 if not files or '.' in files:
591 files = ['']
616 files = ['']
592 results = dict.fromkeys(subrepos)
617 results = dict.fromkeys(subrepos)
593 results['.hg'] = None
618 results['.hg'] = None
594
619
595 for ff in files:
620 for ff in files:
596 if normalize:
621 if normalize:
597 nf = normalize(normpath(ff), False, True)
622 nf = normalize(normpath(ff), False, True)
598 else:
623 else:
599 nf = normpath(ff)
624 nf = normpath(ff)
600 if nf in results:
625 if nf in results:
601 continue
626 continue
602
627
603 try:
628 try:
604 st = lstat(join(nf))
629 st = lstat(join(nf))
605 kind = getkind(st.st_mode)
630 kind = getkind(st.st_mode)
606 if kind == dirkind:
631 if kind == dirkind:
607 if nf in dmap:
632 if nf in dmap:
608 # file replaced by dir on disk but still in dirstate
633 # file replaced by dir on disk but still in dirstate
609 results[nf] = None
634 results[nf] = None
610 if matchedir:
635 if matchedir:
611 matchedir(nf)
636 matchedir(nf)
612 foundadd(nf)
637 foundadd(nf)
613 elif kind == regkind or kind == lnkkind:
638 elif kind == regkind or kind == lnkkind:
614 results[nf] = st
639 results[nf] = st
615 else:
640 else:
616 badfn(ff, badtype(kind))
641 badfn(ff, badtype(kind))
617 if nf in dmap:
642 if nf in dmap:
618 results[nf] = None
643 results[nf] = None
619 except OSError, inst: # nf not found on disk - it is dirstate only
644 except OSError, inst: # nf not found on disk - it is dirstate only
620 if nf in dmap: # does it exactly match a missing file?
645 if nf in dmap: # does it exactly match a missing file?
621 results[nf] = None
646 results[nf] = None
622 else: # does it match a missing directory?
647 else: # does it match a missing directory?
623 prefix = nf + "/"
648 prefix = nf + "/"
624 for fn in dmap:
649 for fn in dmap:
625 if fn.startswith(prefix):
650 if fn.startswith(prefix):
626 if matchedir:
651 if matchedir:
627 matchedir(nf)
652 matchedir(nf)
628 notfoundadd(nf)
653 notfoundadd(nf)
629 break
654 break
630 else:
655 else:
631 badfn(ff, inst.strerror)
656 badfn(ff, inst.strerror)
632
657
633 return results, dirsfound, dirsnotfound
658 return results, dirsfound, dirsnotfound
634
659
635 def walk(self, match, subrepos, unknown, ignored, full=True):
660 def walk(self, match, subrepos, unknown, ignored, full=True):
636 '''
661 '''
637 Walk recursively through the directory tree, finding all files
662 Walk recursively through the directory tree, finding all files
638 matched by match.
663 matched by match.
639
664
640 If full is False, maybe skip some known-clean files.
665 If full is False, maybe skip some known-clean files.
641
666
642 Return a dict mapping filename to stat-like object (either
667 Return a dict mapping filename to stat-like object (either
643 mercurial.osutil.stat instance or return value of os.stat()).
668 mercurial.osutil.stat instance or return value of os.stat()).
644
669
645 '''
670 '''
646 # full is a flag that extensions that hook into walk can use -- this
671 # full is a flag that extensions that hook into walk can use -- this
647 # implementation doesn't use it at all. This satisfies the contract
672 # implementation doesn't use it at all. This satisfies the contract
648 # because we only guarantee a "maybe".
673 # because we only guarantee a "maybe".
649
674
650 if ignored:
675 if ignored:
651 ignore = util.never
676 ignore = util.never
652 dirignore = util.never
677 dirignore = util.never
653 elif unknown:
678 elif unknown:
654 ignore = self._ignore
679 ignore = self._ignore
655 dirignore = self._dirignore
680 dirignore = self._dirignore
656 else:
681 else:
657 # if not unknown and not ignored, drop dir recursion and step 2
682 # if not unknown and not ignored, drop dir recursion and step 2
658 ignore = util.always
683 ignore = util.always
659 dirignore = util.always
684 dirignore = util.always
660
685
661 matchfn = match.matchfn
686 matchfn = match.matchfn
662 matchalways = match.always()
687 matchalways = match.always()
663 matchtdir = match.traversedir
688 matchtdir = match.traversedir
664 dmap = self._map
689 dmap = self._map
665 listdir = osutil.listdir
690 listdir = osutil.listdir
666 lstat = os.lstat
691 lstat = os.lstat
667 dirkind = stat.S_IFDIR
692 dirkind = stat.S_IFDIR
668 regkind = stat.S_IFREG
693 regkind = stat.S_IFREG
669 lnkkind = stat.S_IFLNK
694 lnkkind = stat.S_IFLNK
670 join = self._join
695 join = self._join
671
696
672 exact = skipstep3 = False
697 exact = skipstep3 = False
673 if matchfn == match.exact: # match.exact
698 if matchfn == match.exact: # match.exact
674 exact = True
699 exact = True
675 dirignore = util.always # skip step 2
700 dirignore = util.always # skip step 2
676 elif match.files() and not match.anypats(): # match.match, no patterns
701 elif match.files() and not match.anypats(): # match.match, no patterns
677 skipstep3 = True
702 skipstep3 = True
678
703
679 if not exact and self._checkcase:
704 if not exact and self._checkcase:
680 normalize = self._normalize
705 normalize = self._normalize
681 skipstep3 = False
706 skipstep3 = False
682 else:
707 else:
683 normalize = None
708 normalize = None
684
709
685 # step 1: find all explicit files
710 # step 1: find all explicit files
686 results, work, dirsnotfound = self._walkexplicit(match, subrepos)
711 results, work, dirsnotfound = self._walkexplicit(match, subrepos)
687
712
688 skipstep3 = skipstep3 and not (work or dirsnotfound)
713 skipstep3 = skipstep3 and not (work or dirsnotfound)
689 work = [d for d in work if not dirignore(d)]
714 work = [d for d in work if not dirignore(d)]
690 wadd = work.append
715 wadd = work.append
691
716
692 # step 2: visit subdirectories
717 # step 2: visit subdirectories
693 while work:
718 while work:
694 nd = work.pop()
719 nd = work.pop()
695 skip = None
720 skip = None
696 if nd == '.':
721 if nd == '.':
697 nd = ''
722 nd = ''
698 else:
723 else:
699 skip = '.hg'
724 skip = '.hg'
700 try:
725 try:
701 entries = listdir(join(nd), stat=True, skip=skip)
726 entries = listdir(join(nd), stat=True, skip=skip)
702 except OSError, inst:
727 except OSError, inst:
703 if inst.errno in (errno.EACCES, errno.ENOENT):
728 if inst.errno in (errno.EACCES, errno.ENOENT):
704 match.bad(self.pathto(nd), inst.strerror)
729 match.bad(self.pathto(nd), inst.strerror)
705 continue
730 continue
706 raise
731 raise
707 for f, kind, st in entries:
732 for f, kind, st in entries:
708 if normalize:
733 if normalize:
709 nf = normalize(nd and (nd + "/" + f) or f, True, True)
734 nf = normalize(nd and (nd + "/" + f) or f, True, True)
710 else:
735 else:
711 nf = nd and (nd + "/" + f) or f
736 nf = nd and (nd + "/" + f) or f
712 if nf not in results:
737 if nf not in results:
713 if kind == dirkind:
738 if kind == dirkind:
714 if not ignore(nf):
739 if not ignore(nf):
715 if matchtdir:
740 if matchtdir:
716 matchtdir(nf)
741 matchtdir(nf)
717 wadd(nf)
742 wadd(nf)
718 if nf in dmap and (matchalways or matchfn(nf)):
743 if nf in dmap and (matchalways or matchfn(nf)):
719 results[nf] = None
744 results[nf] = None
720 elif kind == regkind or kind == lnkkind:
745 elif kind == regkind or kind == lnkkind:
721 if nf in dmap:
746 if nf in dmap:
722 if matchalways or matchfn(nf):
747 if matchalways or matchfn(nf):
723 results[nf] = st
748 results[nf] = st
724 elif (matchalways or matchfn(nf)) and not ignore(nf):
749 elif (matchalways or matchfn(nf)) and not ignore(nf):
725 results[nf] = st
750 results[nf] = st
726 elif nf in dmap and (matchalways or matchfn(nf)):
751 elif nf in dmap and (matchalways or matchfn(nf)):
727 results[nf] = None
752 results[nf] = None
728
753
729 for s in subrepos:
754 for s in subrepos:
730 del results[s]
755 del results[s]
731 del results['.hg']
756 del results['.hg']
732
757
733 # step 3: visit remaining files from dmap
758 # step 3: visit remaining files from dmap
734 if not skipstep3 and not exact:
759 if not skipstep3 and not exact:
735 # If a dmap file is not in results yet, it was either
760 # If a dmap file is not in results yet, it was either
736 # a) not matching matchfn b) ignored, c) missing, or d) under a
761 # a) not matching matchfn b) ignored, c) missing, or d) under a
737 # symlink directory.
762 # symlink directory.
738 if not results and matchalways:
763 if not results and matchalways:
739 visit = dmap.keys()
764 visit = dmap.keys()
740 else:
765 else:
741 visit = [f for f in dmap if f not in results and matchfn(f)]
766 visit = [f for f in dmap if f not in results and matchfn(f)]
742 visit.sort()
767 visit.sort()
743
768
744 if unknown:
769 if unknown:
745 # unknown == True means we walked all dirs under the roots
770 # unknown == True means we walked all dirs under the roots
746 # that wasn't ignored, and everything that matched was stat'ed
771 # that wasn't ignored, and everything that matched was stat'ed
747 # and is already in results.
772 # and is already in results.
748 # The rest must thus be ignored or under a symlink.
773 # The rest must thus be ignored or under a symlink.
749 audit_path = pathutil.pathauditor(self._root)
774 audit_path = pathutil.pathauditor(self._root)
750
775
751 for nf in iter(visit):
776 for nf in iter(visit):
752 # Report ignored items in the dmap as long as they are not
777 # Report ignored items in the dmap as long as they are not
753 # under a symlink directory.
778 # under a symlink directory.
754 if audit_path.check(nf):
779 if audit_path.check(nf):
755 try:
780 try:
756 results[nf] = lstat(join(nf))
781 results[nf] = lstat(join(nf))
757 # file was just ignored, no links, and exists
782 # file was just ignored, no links, and exists
758 except OSError:
783 except OSError:
759 # file doesn't exist
784 # file doesn't exist
760 results[nf] = None
785 results[nf] = None
761 else:
786 else:
762 # It's either missing or under a symlink directory
787 # It's either missing or under a symlink directory
763 # which we in this case report as missing
788 # which we in this case report as missing
764 results[nf] = None
789 results[nf] = None
765 else:
790 else:
766 # We may not have walked the full directory tree above,
791 # We may not have walked the full directory tree above,
767 # so stat and check everything we missed.
792 # so stat and check everything we missed.
768 nf = iter(visit).next
793 nf = iter(visit).next
769 for st in util.statfiles([join(i) for i in visit]):
794 for st in util.statfiles([join(i) for i in visit]):
770 results[nf()] = st
795 results[nf()] = st
771 return results
796 return results
772
797
773 def status(self, match, subrepos, ignored, clean, unknown):
798 def status(self, match, subrepos, ignored, clean, unknown):
774 '''Determine the status of the working copy relative to the
799 '''Determine the status of the working copy relative to the
775 dirstate and return a tuple of lists (unsure, modified, added,
800 dirstate and return a tuple of lists (unsure, modified, added,
776 removed, deleted, unknown, ignored, clean), where:
801 removed, deleted, unknown, ignored, clean), where:
777
802
778 unsure:
803 unsure:
779 files that might have been modified since the dirstate was
804 files that might have been modified since the dirstate was
780 written, but need to be read to be sure (size is the same
805 written, but need to be read to be sure (size is the same
781 but mtime differs)
806 but mtime differs)
782 modified:
807 modified:
783 files that have definitely been modified since the dirstate
808 files that have definitely been modified since the dirstate
784 was written (different size or mode)
809 was written (different size or mode)
785 added:
810 added:
786 files that have been explicitly added with hg add
811 files that have been explicitly added with hg add
787 removed:
812 removed:
788 files that have been explicitly removed with hg remove
813 files that have been explicitly removed with hg remove
789 deleted:
814 deleted:
790 files that have been deleted through other means ("missing")
815 files that have been deleted through other means ("missing")
791 unknown:
816 unknown:
792 files not in the dirstate that are not ignored
817 files not in the dirstate that are not ignored
793 ignored:
818 ignored:
794 files not in the dirstate that are ignored
819 files not in the dirstate that are ignored
795 (by _dirignore())
820 (by _dirignore())
796 clean:
821 clean:
797 files that have definitely not been modified since the
822 files that have definitely not been modified since the
798 dirstate was written
823 dirstate was written
799 '''
824 '''
800 listignored, listclean, listunknown = ignored, clean, unknown
825 listignored, listclean, listunknown = ignored, clean, unknown
801 lookup, modified, added, unknown, ignored = [], [], [], [], []
826 lookup, modified, added, unknown, ignored = [], [], [], [], []
802 removed, deleted, clean = [], [], []
827 removed, deleted, clean = [], [], []
803
828
804 dmap = self._map
829 dmap = self._map
805 ladd = lookup.append # aka "unsure"
830 ladd = lookup.append # aka "unsure"
806 madd = modified.append
831 madd = modified.append
807 aadd = added.append
832 aadd = added.append
808 uadd = unknown.append
833 uadd = unknown.append
809 iadd = ignored.append
834 iadd = ignored.append
810 radd = removed.append
835 radd = removed.append
811 dadd = deleted.append
836 dadd = deleted.append
812 cadd = clean.append
837 cadd = clean.append
813 mexact = match.exact
838 mexact = match.exact
814 dirignore = self._dirignore
839 dirignore = self._dirignore
815 checkexec = self._checkexec
840 checkexec = self._checkexec
816 copymap = self._copymap
841 copymap = self._copymap
817 lastnormaltime = self._lastnormaltime
842 lastnormaltime = self._lastnormaltime
818
843
819 # We need to do full walks when either
844 # We need to do full walks when either
820 # - we're listing all clean files, or
845 # - we're listing all clean files, or
821 # - match.traversedir does something, because match.traversedir should
846 # - match.traversedir does something, because match.traversedir should
822 # be called for every dir in the working dir
847 # be called for every dir in the working dir
823 full = listclean or match.traversedir is not None
848 full = listclean or match.traversedir is not None
824 for fn, st in self.walk(match, subrepos, listunknown, listignored,
849 for fn, st in self.walk(match, subrepos, listunknown, listignored,
825 full=full).iteritems():
850 full=full).iteritems():
826 if fn not in dmap:
851 if fn not in dmap:
827 if (listignored or mexact(fn)) and dirignore(fn):
852 if (listignored or mexact(fn)) and dirignore(fn):
828 if listignored:
853 if listignored:
829 iadd(fn)
854 iadd(fn)
830 else:
855 else:
831 uadd(fn)
856 uadd(fn)
832 continue
857 continue
833
858
834 # This is equivalent to 'state, mode, size, time = dmap[fn]' but not
859 # This is equivalent to 'state, mode, size, time = dmap[fn]' but not
835 # written like that for performance reasons. dmap[fn] is not a
860 # written like that for performance reasons. dmap[fn] is not a
836 # Python tuple in compiled builds. The CPython UNPACK_SEQUENCE
861 # Python tuple in compiled builds. The CPython UNPACK_SEQUENCE
837 # opcode has fast paths when the value to be unpacked is a tuple or
862 # opcode has fast paths when the value to be unpacked is a tuple or
838 # a list, but falls back to creating a full-fledged iterator in
863 # a list, but falls back to creating a full-fledged iterator in
839 # general. That is much slower than simply accessing and storing the
864 # general. That is much slower than simply accessing and storing the
840 # tuple members one by one.
865 # tuple members one by one.
841 t = dmap[fn]
866 t = dmap[fn]
842 state = t[0]
867 state = t[0]
843 mode = t[1]
868 mode = t[1]
844 size = t[2]
869 size = t[2]
845 time = t[3]
870 time = t[3]
846
871
847 if not st and state in "nma":
872 if not st and state in "nma":
848 dadd(fn)
873 dadd(fn)
849 elif state == 'n':
874 elif state == 'n':
850 mtime = int(st.st_mtime)
875 mtime = int(st.st_mtime)
851 if (size >= 0 and
876 if (size >= 0 and
852 ((size != st.st_size and size != st.st_size & _rangemask)
877 ((size != st.st_size and size != st.st_size & _rangemask)
853 or ((mode ^ st.st_mode) & 0100 and checkexec))
878 or ((mode ^ st.st_mode) & 0100 and checkexec))
854 or size == -2 # other parent
879 or size == -2 # other parent
855 or fn in copymap):
880 or fn in copymap):
856 madd(fn)
881 madd(fn)
857 elif time != mtime and time != mtime & _rangemask:
882 elif time != mtime and time != mtime & _rangemask:
858 ladd(fn)
883 ladd(fn)
859 elif mtime == lastnormaltime:
884 elif mtime == lastnormaltime:
860 # fn may have been changed in the same timeslot without
885 # fn may have been changed in the same timeslot without
861 # changing its size. This can happen if we quickly do
886 # changing its size. This can happen if we quickly do
862 # multiple commits in a single transaction.
887 # multiple commits in a single transaction.
863 # Force lookup, so we don't miss such a racy file change.
888 # Force lookup, so we don't miss such a racy file change.
864 ladd(fn)
889 ladd(fn)
865 elif listclean:
890 elif listclean:
866 cadd(fn)
891 cadd(fn)
867 elif state == 'm':
892 elif state == 'm':
868 madd(fn)
893 madd(fn)
869 elif state == 'a':
894 elif state == 'a':
870 aadd(fn)
895 aadd(fn)
871 elif state == 'r':
896 elif state == 'r':
872 radd(fn)
897 radd(fn)
873
898
874 return (lookup, modified, added, removed, deleted, unknown, ignored,
899 return (lookup, modified, added, removed, deleted, unknown, ignored,
875 clean)
900 clean)
876
901
877 def matches(self, match):
902 def matches(self, match):
878 '''
903 '''
879 return files in the dirstate (in whatever state) filtered by match
904 return files in the dirstate (in whatever state) filtered by match
880 '''
905 '''
881 dmap = self._map
906 dmap = self._map
882 if match.always():
907 if match.always():
883 return dmap.keys()
908 return dmap.keys()
884 files = match.files()
909 files = match.files()
885 if match.matchfn == match.exact:
910 if match.matchfn == match.exact:
886 # fast path -- filter the other way around, since typically files is
911 # fast path -- filter the other way around, since typically files is
887 # much smaller than dmap
912 # much smaller than dmap
888 return [f for f in files if f in dmap]
913 return [f for f in files if f in dmap]
889 if not match.anypats() and util.all(fn in dmap for fn in files):
914 if not match.anypats() and util.all(fn in dmap for fn in files):
890 # fast path -- all the values are known to be files, so just return
915 # fast path -- all the values are known to be files, so just return
891 # that
916 # that
892 return list(files)
917 return list(files)
893 return [f for f in dmap if match(f)]
918 return [f for f in dmap if match(f)]
@@ -1,1775 +1,1779 b''
1 # localrepo.py - read/write repository class for mercurial
1 # localrepo.py - read/write repository class for mercurial
2 #
2 #
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 #
4 #
5 # This software may be used and distributed according to the terms of the
5 # This software may be used and distributed according to the terms of the
6 # GNU General Public License version 2 or any later version.
6 # GNU General Public License version 2 or any later version.
7 from node import hex, nullid, short
7 from node import hex, nullid, short
8 from i18n import _
8 from i18n import _
9 import urllib
9 import urllib
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
10 import peer, changegroup, subrepo, pushkey, obsolete, repoview
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
11 import changelog, dirstate, filelog, manifest, context, bookmarks, phases
12 import lock as lockmod
12 import lock as lockmod
13 import transaction, store, encoding, exchange, bundle2
13 import transaction, store, encoding, exchange, bundle2
14 import scmutil, util, extensions, hook, error, revset
14 import scmutil, util, extensions, hook, error, revset
15 import match as matchmod
15 import match as matchmod
16 import merge as mergemod
16 import merge as mergemod
17 import tags as tagsmod
17 import tags as tagsmod
18 from lock import release
18 from lock import release
19 import weakref, errno, os, time, inspect
19 import weakref, errno, os, time, inspect
20 import branchmap, pathutil
20 import branchmap, pathutil
21 propertycache = util.propertycache
21 propertycache = util.propertycache
22 filecache = scmutil.filecache
22 filecache = scmutil.filecache
23
23
24 class repofilecache(filecache):
24 class repofilecache(filecache):
25 """All filecache usage on repo are done for logic that should be unfiltered
25 """All filecache usage on repo are done for logic that should be unfiltered
26 """
26 """
27
27
28 def __get__(self, repo, type=None):
28 def __get__(self, repo, type=None):
29 return super(repofilecache, self).__get__(repo.unfiltered(), type)
29 return super(repofilecache, self).__get__(repo.unfiltered(), type)
30 def __set__(self, repo, value):
30 def __set__(self, repo, value):
31 return super(repofilecache, self).__set__(repo.unfiltered(), value)
31 return super(repofilecache, self).__set__(repo.unfiltered(), value)
32 def __delete__(self, repo):
32 def __delete__(self, repo):
33 return super(repofilecache, self).__delete__(repo.unfiltered())
33 return super(repofilecache, self).__delete__(repo.unfiltered())
34
34
35 class storecache(repofilecache):
35 class storecache(repofilecache):
36 """filecache for files in the store"""
36 """filecache for files in the store"""
37 def join(self, obj, fname):
37 def join(self, obj, fname):
38 return obj.sjoin(fname)
38 return obj.sjoin(fname)
39
39
40 class unfilteredpropertycache(propertycache):
40 class unfilteredpropertycache(propertycache):
41 """propertycache that apply to unfiltered repo only"""
41 """propertycache that apply to unfiltered repo only"""
42
42
43 def __get__(self, repo, type=None):
43 def __get__(self, repo, type=None):
44 unfi = repo.unfiltered()
44 unfi = repo.unfiltered()
45 if unfi is repo:
45 if unfi is repo:
46 return super(unfilteredpropertycache, self).__get__(unfi)
46 return super(unfilteredpropertycache, self).__get__(unfi)
47 return getattr(unfi, self.name)
47 return getattr(unfi, self.name)
48
48
49 class filteredpropertycache(propertycache):
49 class filteredpropertycache(propertycache):
50 """propertycache that must take filtering in account"""
50 """propertycache that must take filtering in account"""
51
51
52 def cachevalue(self, obj, value):
52 def cachevalue(self, obj, value):
53 object.__setattr__(obj, self.name, value)
53 object.__setattr__(obj, self.name, value)
54
54
55
55
56 def hasunfilteredcache(repo, name):
56 def hasunfilteredcache(repo, name):
57 """check if a repo has an unfilteredpropertycache value for <name>"""
57 """check if a repo has an unfilteredpropertycache value for <name>"""
58 return name in vars(repo.unfiltered())
58 return name in vars(repo.unfiltered())
59
59
60 def unfilteredmethod(orig):
60 def unfilteredmethod(orig):
61 """decorate method that always need to be run on unfiltered version"""
61 """decorate method that always need to be run on unfiltered version"""
62 def wrapper(repo, *args, **kwargs):
62 def wrapper(repo, *args, **kwargs):
63 return orig(repo.unfiltered(), *args, **kwargs)
63 return orig(repo.unfiltered(), *args, **kwargs)
64 return wrapper
64 return wrapper
65
65
66 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
66 moderncaps = set(('lookup', 'branchmap', 'pushkey', 'known', 'getbundle',
67 'unbundle'))
67 'unbundle'))
68 legacycaps = moderncaps.union(set(['changegroupsubset']))
68 legacycaps = moderncaps.union(set(['changegroupsubset']))
69
69
70 class localpeer(peer.peerrepository):
70 class localpeer(peer.peerrepository):
71 '''peer for a local repo; reflects only the most recent API'''
71 '''peer for a local repo; reflects only the most recent API'''
72
72
73 def __init__(self, repo, caps=moderncaps):
73 def __init__(self, repo, caps=moderncaps):
74 peer.peerrepository.__init__(self)
74 peer.peerrepository.__init__(self)
75 self._repo = repo.filtered('served')
75 self._repo = repo.filtered('served')
76 self.ui = repo.ui
76 self.ui = repo.ui
77 self._caps = repo._restrictcapabilities(caps)
77 self._caps = repo._restrictcapabilities(caps)
78 self.requirements = repo.requirements
78 self.requirements = repo.requirements
79 self.supportedformats = repo.supportedformats
79 self.supportedformats = repo.supportedformats
80
80
81 def close(self):
81 def close(self):
82 self._repo.close()
82 self._repo.close()
83
83
84 def _capabilities(self):
84 def _capabilities(self):
85 return self._caps
85 return self._caps
86
86
87 def local(self):
87 def local(self):
88 return self._repo
88 return self._repo
89
89
90 def canpush(self):
90 def canpush(self):
91 return True
91 return True
92
92
93 def url(self):
93 def url(self):
94 return self._repo.url()
94 return self._repo.url()
95
95
96 def lookup(self, key):
96 def lookup(self, key):
97 return self._repo.lookup(key)
97 return self._repo.lookup(key)
98
98
99 def branchmap(self):
99 def branchmap(self):
100 return self._repo.branchmap()
100 return self._repo.branchmap()
101
101
102 def heads(self):
102 def heads(self):
103 return self._repo.heads()
103 return self._repo.heads()
104
104
105 def known(self, nodes):
105 def known(self, nodes):
106 return self._repo.known(nodes)
106 return self._repo.known(nodes)
107
107
108 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
108 def getbundle(self, source, heads=None, common=None, bundlecaps=None,
109 format='HG10', **kwargs):
109 format='HG10', **kwargs):
110 cg = exchange.getbundle(self._repo, source, heads=heads,
110 cg = exchange.getbundle(self._repo, source, heads=heads,
111 common=common, bundlecaps=bundlecaps, **kwargs)
111 common=common, bundlecaps=bundlecaps, **kwargs)
112 if bundlecaps is not None and 'HG2X' in bundlecaps:
112 if bundlecaps is not None and 'HG2X' in bundlecaps:
113 # When requesting a bundle2, getbundle returns a stream to make the
113 # When requesting a bundle2, getbundle returns a stream to make the
114 # wire level function happier. We need to build a proper object
114 # wire level function happier. We need to build a proper object
115 # from it in local peer.
115 # from it in local peer.
116 cg = bundle2.unbundle20(self.ui, cg)
116 cg = bundle2.unbundle20(self.ui, cg)
117 return cg
117 return cg
118
118
119 # TODO We might want to move the next two calls into legacypeer and add
119 # TODO We might want to move the next two calls into legacypeer and add
120 # unbundle instead.
120 # unbundle instead.
121
121
122 def unbundle(self, cg, heads, url):
122 def unbundle(self, cg, heads, url):
123 """apply a bundle on a repo
123 """apply a bundle on a repo
124
124
125 This function handles the repo locking itself."""
125 This function handles the repo locking itself."""
126 try:
126 try:
127 cg = exchange.readbundle(self.ui, cg, None)
127 cg = exchange.readbundle(self.ui, cg, None)
128 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
128 ret = exchange.unbundle(self._repo, cg, heads, 'push', url)
129 if util.safehasattr(ret, 'getchunks'):
129 if util.safehasattr(ret, 'getchunks'):
130 # This is a bundle20 object, turn it into an unbundler.
130 # This is a bundle20 object, turn it into an unbundler.
131 # This little dance should be dropped eventually when the API
131 # This little dance should be dropped eventually when the API
132 # is finally improved.
132 # is finally improved.
133 stream = util.chunkbuffer(ret.getchunks())
133 stream = util.chunkbuffer(ret.getchunks())
134 ret = bundle2.unbundle20(self.ui, stream)
134 ret = bundle2.unbundle20(self.ui, stream)
135 return ret
135 return ret
136 except error.PushRaced, exc:
136 except error.PushRaced, exc:
137 raise error.ResponseError(_('push failed:'), str(exc))
137 raise error.ResponseError(_('push failed:'), str(exc))
138
138
139 def lock(self):
139 def lock(self):
140 return self._repo.lock()
140 return self._repo.lock()
141
141
142 def addchangegroup(self, cg, source, url):
142 def addchangegroup(self, cg, source, url):
143 return changegroup.addchangegroup(self._repo, cg, source, url)
143 return changegroup.addchangegroup(self._repo, cg, source, url)
144
144
145 def pushkey(self, namespace, key, old, new):
145 def pushkey(self, namespace, key, old, new):
146 return self._repo.pushkey(namespace, key, old, new)
146 return self._repo.pushkey(namespace, key, old, new)
147
147
148 def listkeys(self, namespace):
148 def listkeys(self, namespace):
149 return self._repo.listkeys(namespace)
149 return self._repo.listkeys(namespace)
150
150
151 def debugwireargs(self, one, two, three=None, four=None, five=None):
151 def debugwireargs(self, one, two, three=None, four=None, five=None):
152 '''used to test argument passing over the wire'''
152 '''used to test argument passing over the wire'''
153 return "%s %s %s %s %s" % (one, two, three, four, five)
153 return "%s %s %s %s %s" % (one, two, three, four, five)
154
154
155 class locallegacypeer(localpeer):
155 class locallegacypeer(localpeer):
156 '''peer extension which implements legacy methods too; used for tests with
156 '''peer extension which implements legacy methods too; used for tests with
157 restricted capabilities'''
157 restricted capabilities'''
158
158
159 def __init__(self, repo):
159 def __init__(self, repo):
160 localpeer.__init__(self, repo, caps=legacycaps)
160 localpeer.__init__(self, repo, caps=legacycaps)
161
161
162 def branches(self, nodes):
162 def branches(self, nodes):
163 return self._repo.branches(nodes)
163 return self._repo.branches(nodes)
164
164
165 def between(self, pairs):
165 def between(self, pairs):
166 return self._repo.between(pairs)
166 return self._repo.between(pairs)
167
167
168 def changegroup(self, basenodes, source):
168 def changegroup(self, basenodes, source):
169 return changegroup.changegroup(self._repo, basenodes, source)
169 return changegroup.changegroup(self._repo, basenodes, source)
170
170
171 def changegroupsubset(self, bases, heads, source):
171 def changegroupsubset(self, bases, heads, source):
172 return changegroup.changegroupsubset(self._repo, bases, heads, source)
172 return changegroup.changegroupsubset(self._repo, bases, heads, source)
173
173
174 class localrepository(object):
174 class localrepository(object):
175
175
176 supportedformats = set(('revlogv1', 'generaldelta'))
176 supportedformats = set(('revlogv1', 'generaldelta'))
177 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
177 _basesupported = supportedformats | set(('store', 'fncache', 'shared',
178 'dotencode'))
178 'dotencode'))
179 openerreqs = set(('revlogv1', 'generaldelta'))
179 openerreqs = set(('revlogv1', 'generaldelta'))
180 requirements = ['revlogv1']
180 requirements = ['revlogv1']
181 filtername = None
181 filtername = None
182
182
183 # a list of (ui, featureset) functions.
183 # a list of (ui, featureset) functions.
184 # only functions defined in module of enabled extensions are invoked
184 # only functions defined in module of enabled extensions are invoked
185 featuresetupfuncs = set()
185 featuresetupfuncs = set()
186
186
187 def _baserequirements(self, create):
187 def _baserequirements(self, create):
188 return self.requirements[:]
188 return self.requirements[:]
189
189
190 def __init__(self, baseui, path=None, create=False):
190 def __init__(self, baseui, path=None, create=False):
191 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
191 self.wvfs = scmutil.vfs(path, expandpath=True, realpath=True)
192 self.wopener = self.wvfs
192 self.wopener = self.wvfs
193 self.root = self.wvfs.base
193 self.root = self.wvfs.base
194 self.path = self.wvfs.join(".hg")
194 self.path = self.wvfs.join(".hg")
195 self.origroot = path
195 self.origroot = path
196 self.auditor = pathutil.pathauditor(self.root, self._checknested)
196 self.auditor = pathutil.pathauditor(self.root, self._checknested)
197 self.vfs = scmutil.vfs(self.path)
197 self.vfs = scmutil.vfs(self.path)
198 self.opener = self.vfs
198 self.opener = self.vfs
199 self.baseui = baseui
199 self.baseui = baseui
200 self.ui = baseui.copy()
200 self.ui = baseui.copy()
201 self.ui.copy = baseui.copy # prevent copying repo configuration
201 self.ui.copy = baseui.copy # prevent copying repo configuration
202 # A list of callback to shape the phase if no data were found.
202 # A list of callback to shape the phase if no data were found.
203 # Callback are in the form: func(repo, roots) --> processed root.
203 # Callback are in the form: func(repo, roots) --> processed root.
204 # This list it to be filled by extension during repo setup
204 # This list it to be filled by extension during repo setup
205 self._phasedefaults = []
205 self._phasedefaults = []
206 try:
206 try:
207 self.ui.readconfig(self.join("hgrc"), self.root)
207 self.ui.readconfig(self.join("hgrc"), self.root)
208 extensions.loadall(self.ui)
208 extensions.loadall(self.ui)
209 except IOError:
209 except IOError:
210 pass
210 pass
211
211
212 if self.featuresetupfuncs:
212 if self.featuresetupfuncs:
213 self.supported = set(self._basesupported) # use private copy
213 self.supported = set(self._basesupported) # use private copy
214 extmods = set(m.__name__ for n, m
214 extmods = set(m.__name__ for n, m
215 in extensions.extensions(self.ui))
215 in extensions.extensions(self.ui))
216 for setupfunc in self.featuresetupfuncs:
216 for setupfunc in self.featuresetupfuncs:
217 if setupfunc.__module__ in extmods:
217 if setupfunc.__module__ in extmods:
218 setupfunc(self.ui, self.supported)
218 setupfunc(self.ui, self.supported)
219 else:
219 else:
220 self.supported = self._basesupported
220 self.supported = self._basesupported
221
221
222 if not self.vfs.isdir():
222 if not self.vfs.isdir():
223 if create:
223 if create:
224 if not self.wvfs.exists():
224 if not self.wvfs.exists():
225 self.wvfs.makedirs()
225 self.wvfs.makedirs()
226 self.vfs.makedir(notindexed=True)
226 self.vfs.makedir(notindexed=True)
227 requirements = self._baserequirements(create)
227 requirements = self._baserequirements(create)
228 if self.ui.configbool('format', 'usestore', True):
228 if self.ui.configbool('format', 'usestore', True):
229 self.vfs.mkdir("store")
229 self.vfs.mkdir("store")
230 requirements.append("store")
230 requirements.append("store")
231 if self.ui.configbool('format', 'usefncache', True):
231 if self.ui.configbool('format', 'usefncache', True):
232 requirements.append("fncache")
232 requirements.append("fncache")
233 if self.ui.configbool('format', 'dotencode', True):
233 if self.ui.configbool('format', 'dotencode', True):
234 requirements.append('dotencode')
234 requirements.append('dotencode')
235 # create an invalid changelog
235 # create an invalid changelog
236 self.vfs.append(
236 self.vfs.append(
237 "00changelog.i",
237 "00changelog.i",
238 '\0\0\0\2' # represents revlogv2
238 '\0\0\0\2' # represents revlogv2
239 ' dummy changelog to prevent using the old repo layout'
239 ' dummy changelog to prevent using the old repo layout'
240 )
240 )
241 if self.ui.configbool('format', 'generaldelta', False):
241 if self.ui.configbool('format', 'generaldelta', False):
242 requirements.append("generaldelta")
242 requirements.append("generaldelta")
243 requirements = set(requirements)
243 requirements = set(requirements)
244 else:
244 else:
245 raise error.RepoError(_("repository %s not found") % path)
245 raise error.RepoError(_("repository %s not found") % path)
246 elif create:
246 elif create:
247 raise error.RepoError(_("repository %s already exists") % path)
247 raise error.RepoError(_("repository %s already exists") % path)
248 else:
248 else:
249 try:
249 try:
250 requirements = scmutil.readrequires(self.vfs, self.supported)
250 requirements = scmutil.readrequires(self.vfs, self.supported)
251 except IOError, inst:
251 except IOError, inst:
252 if inst.errno != errno.ENOENT:
252 if inst.errno != errno.ENOENT:
253 raise
253 raise
254 requirements = set()
254 requirements = set()
255
255
256 self.sharedpath = self.path
256 self.sharedpath = self.path
257 try:
257 try:
258 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
258 vfs = scmutil.vfs(self.vfs.read("sharedpath").rstrip('\n'),
259 realpath=True)
259 realpath=True)
260 s = vfs.base
260 s = vfs.base
261 if not vfs.exists():
261 if not vfs.exists():
262 raise error.RepoError(
262 raise error.RepoError(
263 _('.hg/sharedpath points to nonexistent directory %s') % s)
263 _('.hg/sharedpath points to nonexistent directory %s') % s)
264 self.sharedpath = s
264 self.sharedpath = s
265 except IOError, inst:
265 except IOError, inst:
266 if inst.errno != errno.ENOENT:
266 if inst.errno != errno.ENOENT:
267 raise
267 raise
268
268
269 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
269 self.store = store.store(requirements, self.sharedpath, scmutil.vfs)
270 self.spath = self.store.path
270 self.spath = self.store.path
271 self.svfs = self.store.vfs
271 self.svfs = self.store.vfs
272 self.sopener = self.svfs
272 self.sopener = self.svfs
273 self.sjoin = self.store.join
273 self.sjoin = self.store.join
274 self.vfs.createmode = self.store.createmode
274 self.vfs.createmode = self.store.createmode
275 self._applyrequirements(requirements)
275 self._applyrequirements(requirements)
276 if create:
276 if create:
277 self._writerequirements()
277 self._writerequirements()
278
278
279
279
280 self._branchcaches = {}
280 self._branchcaches = {}
281 self.filterpats = {}
281 self.filterpats = {}
282 self._datafilters = {}
282 self._datafilters = {}
283 self._transref = self._lockref = self._wlockref = None
283 self._transref = self._lockref = self._wlockref = None
284
284
285 # A cache for various files under .hg/ that tracks file changes,
285 # A cache for various files under .hg/ that tracks file changes,
286 # (used by the filecache decorator)
286 # (used by the filecache decorator)
287 #
287 #
288 # Maps a property name to its util.filecacheentry
288 # Maps a property name to its util.filecacheentry
289 self._filecache = {}
289 self._filecache = {}
290
290
291 # hold sets of revision to be filtered
291 # hold sets of revision to be filtered
292 # should be cleared when something might have changed the filter value:
292 # should be cleared when something might have changed the filter value:
293 # - new changesets,
293 # - new changesets,
294 # - phase change,
294 # - phase change,
295 # - new obsolescence marker,
295 # - new obsolescence marker,
296 # - working directory parent change,
296 # - working directory parent change,
297 # - bookmark changes
297 # - bookmark changes
298 self.filteredrevcache = {}
298 self.filteredrevcache = {}
299
299
300 def close(self):
300 def close(self):
301 pass
301 pass
302
302
303 def _restrictcapabilities(self, caps):
303 def _restrictcapabilities(self, caps):
304 # bundle2 is not ready for prime time, drop it unless explicitly
304 # bundle2 is not ready for prime time, drop it unless explicitly
305 # required by the tests (or some brave tester)
305 # required by the tests (or some brave tester)
306 if self.ui.configbool('experimental', 'bundle2-exp', False):
306 if self.ui.configbool('experimental', 'bundle2-exp', False):
307 caps = set(caps)
307 caps = set(caps)
308 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
308 capsblob = bundle2.encodecaps(bundle2.getrepocaps(self))
309 caps.add('bundle2-exp=' + urllib.quote(capsblob))
309 caps.add('bundle2-exp=' + urllib.quote(capsblob))
310 return caps
310 return caps
311
311
312 def _applyrequirements(self, requirements):
312 def _applyrequirements(self, requirements):
313 self.requirements = requirements
313 self.requirements = requirements
314 self.sopener.options = dict((r, 1) for r in requirements
314 self.sopener.options = dict((r, 1) for r in requirements
315 if r in self.openerreqs)
315 if r in self.openerreqs)
316 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
316 chunkcachesize = self.ui.configint('format', 'chunkcachesize')
317 if chunkcachesize is not None:
317 if chunkcachesize is not None:
318 self.sopener.options['chunkcachesize'] = chunkcachesize
318 self.sopener.options['chunkcachesize'] = chunkcachesize
319
319
320 def _writerequirements(self):
320 def _writerequirements(self):
321 reqfile = self.opener("requires", "w")
321 reqfile = self.opener("requires", "w")
322 for r in sorted(self.requirements):
322 for r in sorted(self.requirements):
323 reqfile.write("%s\n" % r)
323 reqfile.write("%s\n" % r)
324 reqfile.close()
324 reqfile.close()
325
325
326 def _checknested(self, path):
326 def _checknested(self, path):
327 """Determine if path is a legal nested repository."""
327 """Determine if path is a legal nested repository."""
328 if not path.startswith(self.root):
328 if not path.startswith(self.root):
329 return False
329 return False
330 subpath = path[len(self.root) + 1:]
330 subpath = path[len(self.root) + 1:]
331 normsubpath = util.pconvert(subpath)
331 normsubpath = util.pconvert(subpath)
332
332
333 # XXX: Checking against the current working copy is wrong in
333 # XXX: Checking against the current working copy is wrong in
334 # the sense that it can reject things like
334 # the sense that it can reject things like
335 #
335 #
336 # $ hg cat -r 10 sub/x.txt
336 # $ hg cat -r 10 sub/x.txt
337 #
337 #
338 # if sub/ is no longer a subrepository in the working copy
338 # if sub/ is no longer a subrepository in the working copy
339 # parent revision.
339 # parent revision.
340 #
340 #
341 # However, it can of course also allow things that would have
341 # However, it can of course also allow things that would have
342 # been rejected before, such as the above cat command if sub/
342 # been rejected before, such as the above cat command if sub/
343 # is a subrepository now, but was a normal directory before.
343 # is a subrepository now, but was a normal directory before.
344 # The old path auditor would have rejected by mistake since it
344 # The old path auditor would have rejected by mistake since it
345 # panics when it sees sub/.hg/.
345 # panics when it sees sub/.hg/.
346 #
346 #
347 # All in all, checking against the working copy seems sensible
347 # All in all, checking against the working copy seems sensible
348 # since we want to prevent access to nested repositories on
348 # since we want to prevent access to nested repositories on
349 # the filesystem *now*.
349 # the filesystem *now*.
350 ctx = self[None]
350 ctx = self[None]
351 parts = util.splitpath(subpath)
351 parts = util.splitpath(subpath)
352 while parts:
352 while parts:
353 prefix = '/'.join(parts)
353 prefix = '/'.join(parts)
354 if prefix in ctx.substate:
354 if prefix in ctx.substate:
355 if prefix == normsubpath:
355 if prefix == normsubpath:
356 return True
356 return True
357 else:
357 else:
358 sub = ctx.sub(prefix)
358 sub = ctx.sub(prefix)
359 return sub.checknested(subpath[len(prefix) + 1:])
359 return sub.checknested(subpath[len(prefix) + 1:])
360 else:
360 else:
361 parts.pop()
361 parts.pop()
362 return False
362 return False
363
363
364 def peer(self):
364 def peer(self):
365 return localpeer(self) # not cached to avoid reference cycle
365 return localpeer(self) # not cached to avoid reference cycle
366
366
367 def unfiltered(self):
367 def unfiltered(self):
368 """Return unfiltered version of the repository
368 """Return unfiltered version of the repository
369
369
370 Intended to be overwritten by filtered repo."""
370 Intended to be overwritten by filtered repo."""
371 return self
371 return self
372
372
373 def filtered(self, name):
373 def filtered(self, name):
374 """Return a filtered version of a repository"""
374 """Return a filtered version of a repository"""
375 # build a new class with the mixin and the current class
375 # build a new class with the mixin and the current class
376 # (possibly subclass of the repo)
376 # (possibly subclass of the repo)
377 class proxycls(repoview.repoview, self.unfiltered().__class__):
377 class proxycls(repoview.repoview, self.unfiltered().__class__):
378 pass
378 pass
379 return proxycls(self, name)
379 return proxycls(self, name)
380
380
381 @repofilecache('bookmarks')
381 @repofilecache('bookmarks')
382 def _bookmarks(self):
382 def _bookmarks(self):
383 return bookmarks.bmstore(self)
383 return bookmarks.bmstore(self)
384
384
385 @repofilecache('bookmarks.current')
385 @repofilecache('bookmarks.current')
386 def _bookmarkcurrent(self):
386 def _bookmarkcurrent(self):
387 return bookmarks.readcurrent(self)
387 return bookmarks.readcurrent(self)
388
388
389 def bookmarkheads(self, bookmark):
389 def bookmarkheads(self, bookmark):
390 name = bookmark.split('@', 1)[0]
390 name = bookmark.split('@', 1)[0]
391 heads = []
391 heads = []
392 for mark, n in self._bookmarks.iteritems():
392 for mark, n in self._bookmarks.iteritems():
393 if mark.split('@', 1)[0] == name:
393 if mark.split('@', 1)[0] == name:
394 heads.append(n)
394 heads.append(n)
395 return heads
395 return heads
396
396
397 @storecache('phaseroots')
397 @storecache('phaseroots')
398 def _phasecache(self):
398 def _phasecache(self):
399 return phases.phasecache(self, self._phasedefaults)
399 return phases.phasecache(self, self._phasedefaults)
400
400
401 @storecache('obsstore')
401 @storecache('obsstore')
402 def obsstore(self):
402 def obsstore(self):
403 store = obsolete.obsstore(self.sopener)
403 store = obsolete.obsstore(self.sopener)
404 if store and not obsolete._enabled:
404 if store and not obsolete._enabled:
405 # message is rare enough to not be translated
405 # message is rare enough to not be translated
406 msg = 'obsolete feature not enabled but %i markers found!\n'
406 msg = 'obsolete feature not enabled but %i markers found!\n'
407 self.ui.warn(msg % len(list(store)))
407 self.ui.warn(msg % len(list(store)))
408 return store
408 return store
409
409
410 @storecache('00changelog.i')
410 @storecache('00changelog.i')
411 def changelog(self):
411 def changelog(self):
412 c = changelog.changelog(self.sopener)
412 c = changelog.changelog(self.sopener)
413 if 'HG_PENDING' in os.environ:
413 if 'HG_PENDING' in os.environ:
414 p = os.environ['HG_PENDING']
414 p = os.environ['HG_PENDING']
415 if p.startswith(self.root):
415 if p.startswith(self.root):
416 c.readpending('00changelog.i.a')
416 c.readpending('00changelog.i.a')
417 return c
417 return c
418
418
419 @storecache('00manifest.i')
419 @storecache('00manifest.i')
420 def manifest(self):
420 def manifest(self):
421 return manifest.manifest(self.sopener)
421 return manifest.manifest(self.sopener)
422
422
423 @repofilecache('dirstate')
423 @repofilecache('dirstate')
424 def dirstate(self):
424 def dirstate(self):
425 warned = [0]
425 warned = [0]
426 def validate(node):
426 def validate(node):
427 try:
427 try:
428 self.changelog.rev(node)
428 self.changelog.rev(node)
429 return node
429 return node
430 except error.LookupError:
430 except error.LookupError:
431 if not warned[0]:
431 if not warned[0]:
432 warned[0] = True
432 warned[0] = True
433 self.ui.warn(_("warning: ignoring unknown"
433 self.ui.warn(_("warning: ignoring unknown"
434 " working parent %s!\n") % short(node))
434 " working parent %s!\n") % short(node))
435 return nullid
435 return nullid
436
436
437 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
437 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
438
438
439 def __getitem__(self, changeid):
439 def __getitem__(self, changeid):
440 if changeid is None:
440 if changeid is None:
441 return context.workingctx(self)
441 return context.workingctx(self)
442 return context.changectx(self, changeid)
442 return context.changectx(self, changeid)
443
443
444 def __contains__(self, changeid):
444 def __contains__(self, changeid):
445 try:
445 try:
446 return bool(self.lookup(changeid))
446 return bool(self.lookup(changeid))
447 except error.RepoLookupError:
447 except error.RepoLookupError:
448 return False
448 return False
449
449
450 def __nonzero__(self):
450 def __nonzero__(self):
451 return True
451 return True
452
452
453 def __len__(self):
453 def __len__(self):
454 return len(self.changelog)
454 return len(self.changelog)
455
455
456 def __iter__(self):
456 def __iter__(self):
457 return iter(self.changelog)
457 return iter(self.changelog)
458
458
459 def revs(self, expr, *args):
459 def revs(self, expr, *args):
460 '''Return a list of revisions matching the given revset'''
460 '''Return a list of revisions matching the given revset'''
461 expr = revset.formatspec(expr, *args)
461 expr = revset.formatspec(expr, *args)
462 m = revset.match(None, expr)
462 m = revset.match(None, expr)
463 return m(self, revset.spanset(self))
463 return m(self, revset.spanset(self))
464
464
465 def set(self, expr, *args):
465 def set(self, expr, *args):
466 '''
466 '''
467 Yield a context for each matching revision, after doing arg
467 Yield a context for each matching revision, after doing arg
468 replacement via revset.formatspec
468 replacement via revset.formatspec
469 '''
469 '''
470 for r in self.revs(expr, *args):
470 for r in self.revs(expr, *args):
471 yield self[r]
471 yield self[r]
472
472
473 def url(self):
473 def url(self):
474 return 'file:' + self.root
474 return 'file:' + self.root
475
475
476 def hook(self, name, throw=False, **args):
476 def hook(self, name, throw=False, **args):
477 """Call a hook, passing this repo instance.
477 """Call a hook, passing this repo instance.
478
478
479 This a convenience method to aid invoking hooks. Extensions likely
479 This a convenience method to aid invoking hooks. Extensions likely
480 won't call this unless they have registered a custom hook or are
480 won't call this unless they have registered a custom hook or are
481 replacing code that is expected to call a hook.
481 replacing code that is expected to call a hook.
482 """
482 """
483 return hook.hook(self.ui, self, name, throw, **args)
483 return hook.hook(self.ui, self, name, throw, **args)
484
484
485 @unfilteredmethod
485 @unfilteredmethod
486 def _tag(self, names, node, message, local, user, date, extra={},
486 def _tag(self, names, node, message, local, user, date, extra={},
487 editor=False):
487 editor=False):
488 if isinstance(names, str):
488 if isinstance(names, str):
489 names = (names,)
489 names = (names,)
490
490
491 branches = self.branchmap()
491 branches = self.branchmap()
492 for name in names:
492 for name in names:
493 self.hook('pretag', throw=True, node=hex(node), tag=name,
493 self.hook('pretag', throw=True, node=hex(node), tag=name,
494 local=local)
494 local=local)
495 if name in branches:
495 if name in branches:
496 self.ui.warn(_("warning: tag %s conflicts with existing"
496 self.ui.warn(_("warning: tag %s conflicts with existing"
497 " branch name\n") % name)
497 " branch name\n") % name)
498
498
499 def writetags(fp, names, munge, prevtags):
499 def writetags(fp, names, munge, prevtags):
500 fp.seek(0, 2)
500 fp.seek(0, 2)
501 if prevtags and prevtags[-1] != '\n':
501 if prevtags and prevtags[-1] != '\n':
502 fp.write('\n')
502 fp.write('\n')
503 for name in names:
503 for name in names:
504 m = munge and munge(name) or name
504 m = munge and munge(name) or name
505 if (self._tagscache.tagtypes and
505 if (self._tagscache.tagtypes and
506 name in self._tagscache.tagtypes):
506 name in self._tagscache.tagtypes):
507 old = self.tags().get(name, nullid)
507 old = self.tags().get(name, nullid)
508 fp.write('%s %s\n' % (hex(old), m))
508 fp.write('%s %s\n' % (hex(old), m))
509 fp.write('%s %s\n' % (hex(node), m))
509 fp.write('%s %s\n' % (hex(node), m))
510 fp.close()
510 fp.close()
511
511
512 prevtags = ''
512 prevtags = ''
513 if local:
513 if local:
514 try:
514 try:
515 fp = self.opener('localtags', 'r+')
515 fp = self.opener('localtags', 'r+')
516 except IOError:
516 except IOError:
517 fp = self.opener('localtags', 'a')
517 fp = self.opener('localtags', 'a')
518 else:
518 else:
519 prevtags = fp.read()
519 prevtags = fp.read()
520
520
521 # local tags are stored in the current charset
521 # local tags are stored in the current charset
522 writetags(fp, names, None, prevtags)
522 writetags(fp, names, None, prevtags)
523 for name in names:
523 for name in names:
524 self.hook('tag', node=hex(node), tag=name, local=local)
524 self.hook('tag', node=hex(node), tag=name, local=local)
525 return
525 return
526
526
527 try:
527 try:
528 fp = self.wfile('.hgtags', 'rb+')
528 fp = self.wfile('.hgtags', 'rb+')
529 except IOError, e:
529 except IOError, e:
530 if e.errno != errno.ENOENT:
530 if e.errno != errno.ENOENT:
531 raise
531 raise
532 fp = self.wfile('.hgtags', 'ab')
532 fp = self.wfile('.hgtags', 'ab')
533 else:
533 else:
534 prevtags = fp.read()
534 prevtags = fp.read()
535
535
536 # committed tags are stored in UTF-8
536 # committed tags are stored in UTF-8
537 writetags(fp, names, encoding.fromlocal, prevtags)
537 writetags(fp, names, encoding.fromlocal, prevtags)
538
538
539 fp.close()
539 fp.close()
540
540
541 self.invalidatecaches()
541 self.invalidatecaches()
542
542
543 if '.hgtags' not in self.dirstate:
543 if '.hgtags' not in self.dirstate:
544 self[None].add(['.hgtags'])
544 self[None].add(['.hgtags'])
545
545
546 m = matchmod.exact(self.root, '', ['.hgtags'])
546 m = matchmod.exact(self.root, '', ['.hgtags'])
547 tagnode = self.commit(message, user, date, extra=extra, match=m,
547 tagnode = self.commit(message, user, date, extra=extra, match=m,
548 editor=editor)
548 editor=editor)
549
549
550 for name in names:
550 for name in names:
551 self.hook('tag', node=hex(node), tag=name, local=local)
551 self.hook('tag', node=hex(node), tag=name, local=local)
552
552
553 return tagnode
553 return tagnode
554
554
555 def tag(self, names, node, message, local, user, date, editor=False):
555 def tag(self, names, node, message, local, user, date, editor=False):
556 '''tag a revision with one or more symbolic names.
556 '''tag a revision with one or more symbolic names.
557
557
558 names is a list of strings or, when adding a single tag, names may be a
558 names is a list of strings or, when adding a single tag, names may be a
559 string.
559 string.
560
560
561 if local is True, the tags are stored in a per-repository file.
561 if local is True, the tags are stored in a per-repository file.
562 otherwise, they are stored in the .hgtags file, and a new
562 otherwise, they are stored in the .hgtags file, and a new
563 changeset is committed with the change.
563 changeset is committed with the change.
564
564
565 keyword arguments:
565 keyword arguments:
566
566
567 local: whether to store tags in non-version-controlled file
567 local: whether to store tags in non-version-controlled file
568 (default False)
568 (default False)
569
569
570 message: commit message to use if committing
570 message: commit message to use if committing
571
571
572 user: name of user to use if committing
572 user: name of user to use if committing
573
573
574 date: date tuple to use if committing'''
574 date: date tuple to use if committing'''
575
575
576 if not local:
576 if not local:
577 for x in self.status()[:5]:
577 for x in self.status()[:5]:
578 if '.hgtags' in x:
578 if '.hgtags' in x:
579 raise util.Abort(_('working copy of .hgtags is changed '
579 raise util.Abort(_('working copy of .hgtags is changed '
580 '(please commit .hgtags manually)'))
580 '(please commit .hgtags manually)'))
581
581
582 self.tags() # instantiate the cache
582 self.tags() # instantiate the cache
583 self._tag(names, node, message, local, user, date, editor=editor)
583 self._tag(names, node, message, local, user, date, editor=editor)
584
584
585 @filteredpropertycache
585 @filteredpropertycache
586 def _tagscache(self):
586 def _tagscache(self):
587 '''Returns a tagscache object that contains various tags related
587 '''Returns a tagscache object that contains various tags related
588 caches.'''
588 caches.'''
589
589
590 # This simplifies its cache management by having one decorated
590 # This simplifies its cache management by having one decorated
591 # function (this one) and the rest simply fetch things from it.
591 # function (this one) and the rest simply fetch things from it.
592 class tagscache(object):
592 class tagscache(object):
593 def __init__(self):
593 def __init__(self):
594 # These two define the set of tags for this repository. tags
594 # These two define the set of tags for this repository. tags
595 # maps tag name to node; tagtypes maps tag name to 'global' or
595 # maps tag name to node; tagtypes maps tag name to 'global' or
596 # 'local'. (Global tags are defined by .hgtags across all
596 # 'local'. (Global tags are defined by .hgtags across all
597 # heads, and local tags are defined in .hg/localtags.)
597 # heads, and local tags are defined in .hg/localtags.)
598 # They constitute the in-memory cache of tags.
598 # They constitute the in-memory cache of tags.
599 self.tags = self.tagtypes = None
599 self.tags = self.tagtypes = None
600
600
601 self.nodetagscache = self.tagslist = None
601 self.nodetagscache = self.tagslist = None
602
602
603 cache = tagscache()
603 cache = tagscache()
604 cache.tags, cache.tagtypes = self._findtags()
604 cache.tags, cache.tagtypes = self._findtags()
605
605
606 return cache
606 return cache
607
607
608 def tags(self):
608 def tags(self):
609 '''return a mapping of tag to node'''
609 '''return a mapping of tag to node'''
610 t = {}
610 t = {}
611 if self.changelog.filteredrevs:
611 if self.changelog.filteredrevs:
612 tags, tt = self._findtags()
612 tags, tt = self._findtags()
613 else:
613 else:
614 tags = self._tagscache.tags
614 tags = self._tagscache.tags
615 for k, v in tags.iteritems():
615 for k, v in tags.iteritems():
616 try:
616 try:
617 # ignore tags to unknown nodes
617 # ignore tags to unknown nodes
618 self.changelog.rev(v)
618 self.changelog.rev(v)
619 t[k] = v
619 t[k] = v
620 except (error.LookupError, ValueError):
620 except (error.LookupError, ValueError):
621 pass
621 pass
622 return t
622 return t
623
623
624 def _findtags(self):
624 def _findtags(self):
625 '''Do the hard work of finding tags. Return a pair of dicts
625 '''Do the hard work of finding tags. Return a pair of dicts
626 (tags, tagtypes) where tags maps tag name to node, and tagtypes
626 (tags, tagtypes) where tags maps tag name to node, and tagtypes
627 maps tag name to a string like \'global\' or \'local\'.
627 maps tag name to a string like \'global\' or \'local\'.
628 Subclasses or extensions are free to add their own tags, but
628 Subclasses or extensions are free to add their own tags, but
629 should be aware that the returned dicts will be retained for the
629 should be aware that the returned dicts will be retained for the
630 duration of the localrepo object.'''
630 duration of the localrepo object.'''
631
631
632 # XXX what tagtype should subclasses/extensions use? Currently
632 # XXX what tagtype should subclasses/extensions use? Currently
633 # mq and bookmarks add tags, but do not set the tagtype at all.
633 # mq and bookmarks add tags, but do not set the tagtype at all.
634 # Should each extension invent its own tag type? Should there
634 # Should each extension invent its own tag type? Should there
635 # be one tagtype for all such "virtual" tags? Or is the status
635 # be one tagtype for all such "virtual" tags? Or is the status
636 # quo fine?
636 # quo fine?
637
637
638 alltags = {} # map tag name to (node, hist)
638 alltags = {} # map tag name to (node, hist)
639 tagtypes = {}
639 tagtypes = {}
640
640
641 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
641 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
642 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
642 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
643
643
644 # Build the return dicts. Have to re-encode tag names because
644 # Build the return dicts. Have to re-encode tag names because
645 # the tags module always uses UTF-8 (in order not to lose info
645 # the tags module always uses UTF-8 (in order not to lose info
646 # writing to the cache), but the rest of Mercurial wants them in
646 # writing to the cache), but the rest of Mercurial wants them in
647 # local encoding.
647 # local encoding.
648 tags = {}
648 tags = {}
649 for (name, (node, hist)) in alltags.iteritems():
649 for (name, (node, hist)) in alltags.iteritems():
650 if node != nullid:
650 if node != nullid:
651 tags[encoding.tolocal(name)] = node
651 tags[encoding.tolocal(name)] = node
652 tags['tip'] = self.changelog.tip()
652 tags['tip'] = self.changelog.tip()
653 tagtypes = dict([(encoding.tolocal(name), value)
653 tagtypes = dict([(encoding.tolocal(name), value)
654 for (name, value) in tagtypes.iteritems()])
654 for (name, value) in tagtypes.iteritems()])
655 return (tags, tagtypes)
655 return (tags, tagtypes)
656
656
657 def tagtype(self, tagname):
657 def tagtype(self, tagname):
658 '''
658 '''
659 return the type of the given tag. result can be:
659 return the type of the given tag. result can be:
660
660
661 'local' : a local tag
661 'local' : a local tag
662 'global' : a global tag
662 'global' : a global tag
663 None : tag does not exist
663 None : tag does not exist
664 '''
664 '''
665
665
666 return self._tagscache.tagtypes.get(tagname)
666 return self._tagscache.tagtypes.get(tagname)
667
667
668 def tagslist(self):
668 def tagslist(self):
669 '''return a list of tags ordered by revision'''
669 '''return a list of tags ordered by revision'''
670 if not self._tagscache.tagslist:
670 if not self._tagscache.tagslist:
671 l = []
671 l = []
672 for t, n in self.tags().iteritems():
672 for t, n in self.tags().iteritems():
673 l.append((self.changelog.rev(n), t, n))
673 l.append((self.changelog.rev(n), t, n))
674 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
674 self._tagscache.tagslist = [(t, n) for r, t, n in sorted(l)]
675
675
676 return self._tagscache.tagslist
676 return self._tagscache.tagslist
677
677
678 def nodetags(self, node):
678 def nodetags(self, node):
679 '''return the tags associated with a node'''
679 '''return the tags associated with a node'''
680 if not self._tagscache.nodetagscache:
680 if not self._tagscache.nodetagscache:
681 nodetagscache = {}
681 nodetagscache = {}
682 for t, n in self._tagscache.tags.iteritems():
682 for t, n in self._tagscache.tags.iteritems():
683 nodetagscache.setdefault(n, []).append(t)
683 nodetagscache.setdefault(n, []).append(t)
684 for tags in nodetagscache.itervalues():
684 for tags in nodetagscache.itervalues():
685 tags.sort()
685 tags.sort()
686 self._tagscache.nodetagscache = nodetagscache
686 self._tagscache.nodetagscache = nodetagscache
687 return self._tagscache.nodetagscache.get(node, [])
687 return self._tagscache.nodetagscache.get(node, [])
688
688
689 def nodebookmarks(self, node):
689 def nodebookmarks(self, node):
690 marks = []
690 marks = []
691 for bookmark, n in self._bookmarks.iteritems():
691 for bookmark, n in self._bookmarks.iteritems():
692 if n == node:
692 if n == node:
693 marks.append(bookmark)
693 marks.append(bookmark)
694 return sorted(marks)
694 return sorted(marks)
695
695
696 def branchmap(self):
696 def branchmap(self):
697 '''returns a dictionary {branch: [branchheads]} with branchheads
697 '''returns a dictionary {branch: [branchheads]} with branchheads
698 ordered by increasing revision number'''
698 ordered by increasing revision number'''
699 branchmap.updatecache(self)
699 branchmap.updatecache(self)
700 return self._branchcaches[self.filtername]
700 return self._branchcaches[self.filtername]
701
701
702 def branchtip(self, branch):
702 def branchtip(self, branch):
703 '''return the tip node for a given branch'''
703 '''return the tip node for a given branch'''
704 try:
704 try:
705 return self.branchmap().branchtip(branch)
705 return self.branchmap().branchtip(branch)
706 except KeyError:
706 except KeyError:
707 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
707 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
708
708
709 def lookup(self, key):
709 def lookup(self, key):
710 return self[key].node()
710 return self[key].node()
711
711
712 def lookupbranch(self, key, remote=None):
712 def lookupbranch(self, key, remote=None):
713 repo = remote or self
713 repo = remote or self
714 if key in repo.branchmap():
714 if key in repo.branchmap():
715 return key
715 return key
716
716
717 repo = (remote and remote.local()) and remote or self
717 repo = (remote and remote.local()) and remote or self
718 return repo[key].branch()
718 return repo[key].branch()
719
719
720 def known(self, nodes):
720 def known(self, nodes):
721 nm = self.changelog.nodemap
721 nm = self.changelog.nodemap
722 pc = self._phasecache
722 pc = self._phasecache
723 result = []
723 result = []
724 for n in nodes:
724 for n in nodes:
725 r = nm.get(n)
725 r = nm.get(n)
726 resp = not (r is None or pc.phase(self, r) >= phases.secret)
726 resp = not (r is None or pc.phase(self, r) >= phases.secret)
727 result.append(resp)
727 result.append(resp)
728 return result
728 return result
729
729
730 def local(self):
730 def local(self):
731 return self
731 return self
732
732
733 def cancopy(self):
733 def cancopy(self):
734 # so statichttprepo's override of local() works
734 # so statichttprepo's override of local() works
735 if not self.local():
735 if not self.local():
736 return False
736 return False
737 if not self.ui.configbool('phases', 'publish', True):
737 if not self.ui.configbool('phases', 'publish', True):
738 return True
738 return True
739 # if publishing we can't copy if there is filtered content
739 # if publishing we can't copy if there is filtered content
740 return not self.filtered('visible').changelog.filteredrevs
740 return not self.filtered('visible').changelog.filteredrevs
741
741
742 def join(self, f, *insidef):
742 def join(self, f, *insidef):
743 return os.path.join(self.path, f, *insidef)
743 return os.path.join(self.path, f, *insidef)
744
744
745 def wjoin(self, f, *insidef):
745 def wjoin(self, f, *insidef):
746 return os.path.join(self.root, f, *insidef)
746 return os.path.join(self.root, f, *insidef)
747
747
748 def file(self, f):
748 def file(self, f):
749 if f[0] == '/':
749 if f[0] == '/':
750 f = f[1:]
750 f = f[1:]
751 return filelog.filelog(self.sopener, f)
751 return filelog.filelog(self.sopener, f)
752
752
753 def changectx(self, changeid):
753 def changectx(self, changeid):
754 return self[changeid]
754 return self[changeid]
755
755
756 def parents(self, changeid=None):
756 def parents(self, changeid=None):
757 '''get list of changectxs for parents of changeid'''
757 '''get list of changectxs for parents of changeid'''
758 return self[changeid].parents()
758 return self[changeid].parents()
759
759
760 def setparents(self, p1, p2=nullid):
760 def setparents(self, p1, p2=nullid):
761 copies = self.dirstate.setparents(p1, p2)
761 copies = self.dirstate.setparents(p1, p2)
762 pctx = self[p1]
762 pctx = self[p1]
763 if copies:
763 if copies:
764 # Adjust copy records, the dirstate cannot do it, it
764 # Adjust copy records, the dirstate cannot do it, it
765 # requires access to parents manifests. Preserve them
765 # requires access to parents manifests. Preserve them
766 # only for entries added to first parent.
766 # only for entries added to first parent.
767 for f in copies:
767 for f in copies:
768 if f not in pctx and copies[f] in pctx:
768 if f not in pctx and copies[f] in pctx:
769 self.dirstate.copy(copies[f], f)
769 self.dirstate.copy(copies[f], f)
770 if p2 == nullid:
770 if p2 == nullid:
771 for f, s in sorted(self.dirstate.copies().items()):
771 for f, s in sorted(self.dirstate.copies().items()):
772 if f not in pctx and s not in pctx:
772 if f not in pctx and s not in pctx:
773 self.dirstate.copy(None, f)
773 self.dirstate.copy(None, f)
774
774
775 def filectx(self, path, changeid=None, fileid=None):
775 def filectx(self, path, changeid=None, fileid=None):
776 """changeid can be a changeset revision, node, or tag.
776 """changeid can be a changeset revision, node, or tag.
777 fileid can be a file revision or node."""
777 fileid can be a file revision or node."""
778 return context.filectx(self, path, changeid, fileid)
778 return context.filectx(self, path, changeid, fileid)
779
779
780 def getcwd(self):
780 def getcwd(self):
781 return self.dirstate.getcwd()
781 return self.dirstate.getcwd()
782
782
783 def pathto(self, f, cwd=None):
783 def pathto(self, f, cwd=None):
784 return self.dirstate.pathto(f, cwd)
784 return self.dirstate.pathto(f, cwd)
785
785
786 def wfile(self, f, mode='r'):
786 def wfile(self, f, mode='r'):
787 return self.wopener(f, mode)
787 return self.wopener(f, mode)
788
788
789 def _link(self, f):
789 def _link(self, f):
790 return self.wvfs.islink(f)
790 return self.wvfs.islink(f)
791
791
792 def _loadfilter(self, filter):
792 def _loadfilter(self, filter):
793 if filter not in self.filterpats:
793 if filter not in self.filterpats:
794 l = []
794 l = []
795 for pat, cmd in self.ui.configitems(filter):
795 for pat, cmd in self.ui.configitems(filter):
796 if cmd == '!':
796 if cmd == '!':
797 continue
797 continue
798 mf = matchmod.match(self.root, '', [pat])
798 mf = matchmod.match(self.root, '', [pat])
799 fn = None
799 fn = None
800 params = cmd
800 params = cmd
801 for name, filterfn in self._datafilters.iteritems():
801 for name, filterfn in self._datafilters.iteritems():
802 if cmd.startswith(name):
802 if cmd.startswith(name):
803 fn = filterfn
803 fn = filterfn
804 params = cmd[len(name):].lstrip()
804 params = cmd[len(name):].lstrip()
805 break
805 break
806 if not fn:
806 if not fn:
807 fn = lambda s, c, **kwargs: util.filter(s, c)
807 fn = lambda s, c, **kwargs: util.filter(s, c)
808 # Wrap old filters not supporting keyword arguments
808 # Wrap old filters not supporting keyword arguments
809 if not inspect.getargspec(fn)[2]:
809 if not inspect.getargspec(fn)[2]:
810 oldfn = fn
810 oldfn = fn
811 fn = lambda s, c, **kwargs: oldfn(s, c)
811 fn = lambda s, c, **kwargs: oldfn(s, c)
812 l.append((mf, fn, params))
812 l.append((mf, fn, params))
813 self.filterpats[filter] = l
813 self.filterpats[filter] = l
814 return self.filterpats[filter]
814 return self.filterpats[filter]
815
815
816 def _filter(self, filterpats, filename, data):
816 def _filter(self, filterpats, filename, data):
817 for mf, fn, cmd in filterpats:
817 for mf, fn, cmd in filterpats:
818 if mf(filename):
818 if mf(filename):
819 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
819 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
820 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
820 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
821 break
821 break
822
822
823 return data
823 return data
824
824
825 @unfilteredpropertycache
825 @unfilteredpropertycache
826 def _encodefilterpats(self):
826 def _encodefilterpats(self):
827 return self._loadfilter('encode')
827 return self._loadfilter('encode')
828
828
829 @unfilteredpropertycache
829 @unfilteredpropertycache
830 def _decodefilterpats(self):
830 def _decodefilterpats(self):
831 return self._loadfilter('decode')
831 return self._loadfilter('decode')
832
832
833 def adddatafilter(self, name, filter):
833 def adddatafilter(self, name, filter):
834 self._datafilters[name] = filter
834 self._datafilters[name] = filter
835
835
836 def wread(self, filename):
836 def wread(self, filename):
837 if self._link(filename):
837 if self._link(filename):
838 data = self.wvfs.readlink(filename)
838 data = self.wvfs.readlink(filename)
839 else:
839 else:
840 data = self.wopener.read(filename)
840 data = self.wopener.read(filename)
841 return self._filter(self._encodefilterpats, filename, data)
841 return self._filter(self._encodefilterpats, filename, data)
842
842
843 def wwrite(self, filename, data, flags):
843 def wwrite(self, filename, data, flags):
844 data = self._filter(self._decodefilterpats, filename, data)
844 data = self._filter(self._decodefilterpats, filename, data)
845 if 'l' in flags:
845 if 'l' in flags:
846 self.wopener.symlink(data, filename)
846 self.wopener.symlink(data, filename)
847 else:
847 else:
848 self.wopener.write(filename, data)
848 self.wopener.write(filename, data)
849 if 'x' in flags:
849 if 'x' in flags:
850 self.wvfs.setflags(filename, False, True)
850 self.wvfs.setflags(filename, False, True)
851
851
852 def wwritedata(self, filename, data):
852 def wwritedata(self, filename, data):
853 return self._filter(self._decodefilterpats, filename, data)
853 return self._filter(self._decodefilterpats, filename, data)
854
854
855 def transaction(self, desc, report=None):
855 def transaction(self, desc, report=None):
856 tr = self._transref and self._transref() or None
856 tr = self._transref and self._transref() or None
857 if tr and tr.running():
857 if tr and tr.running():
858 return tr.nest()
858 return tr.nest()
859
859
860 # abort here if the journal already exists
860 # abort here if the journal already exists
861 if self.svfs.exists("journal"):
861 if self.svfs.exists("journal"):
862 raise error.RepoError(
862 raise error.RepoError(
863 _("abandoned transaction found"),
863 _("abandoned transaction found"),
864 hint=_("run 'hg recover' to clean up transaction"))
864 hint=_("run 'hg recover' to clean up transaction"))
865
865
866 def onclose():
866 def onclose():
867 self.store.write(self._transref())
867 self.store.write(self._transref())
868
868
869 self._writejournal(desc)
869 self._writejournal(desc)
870 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
870 renames = [(vfs, x, undoname(x)) for vfs, x in self._journalfiles()]
871 rp = report and report or self.ui.warn
871 rp = report and report or self.ui.warn
872 tr = transaction.transaction(rp, self.sopener,
872 tr = transaction.transaction(rp, self.sopener,
873 "journal",
873 "journal",
874 aftertrans(renames),
874 aftertrans(renames),
875 self.store.createmode,
875 self.store.createmode,
876 onclose)
876 onclose)
877 self._transref = weakref.ref(tr)
877 self._transref = weakref.ref(tr)
878 return tr
878 return tr
879
879
880 def _journalfiles(self):
880 def _journalfiles(self):
881 return ((self.svfs, 'journal'),
881 return ((self.svfs, 'journal'),
882 (self.vfs, 'journal.dirstate'),
882 (self.vfs, 'journal.dirstate'),
883 (self.vfs, 'journal.branch'),
883 (self.vfs, 'journal.branch'),
884 (self.vfs, 'journal.desc'),
884 (self.vfs, 'journal.desc'),
885 (self.vfs, 'journal.bookmarks'),
885 (self.vfs, 'journal.bookmarks'),
886 (self.svfs, 'journal.phaseroots'))
886 (self.svfs, 'journal.phaseroots'))
887
887
888 def undofiles(self):
888 def undofiles(self):
889 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
889 return [(vfs, undoname(x)) for vfs, x in self._journalfiles()]
890
890
891 def _writejournal(self, desc):
891 def _writejournal(self, desc):
892 self.opener.write("journal.dirstate",
892 self.opener.write("journal.dirstate",
893 self.opener.tryread("dirstate"))
893 self.opener.tryread("dirstate"))
894 self.opener.write("journal.branch",
894 self.opener.write("journal.branch",
895 encoding.fromlocal(self.dirstate.branch()))
895 encoding.fromlocal(self.dirstate.branch()))
896 self.opener.write("journal.desc",
896 self.opener.write("journal.desc",
897 "%d\n%s\n" % (len(self), desc))
897 "%d\n%s\n" % (len(self), desc))
898 self.opener.write("journal.bookmarks",
898 self.opener.write("journal.bookmarks",
899 self.opener.tryread("bookmarks"))
899 self.opener.tryread("bookmarks"))
900 self.sopener.write("journal.phaseroots",
900 self.sopener.write("journal.phaseroots",
901 self.sopener.tryread("phaseroots"))
901 self.sopener.tryread("phaseroots"))
902
902
903 def recover(self):
903 def recover(self):
904 lock = self.lock()
904 lock = self.lock()
905 try:
905 try:
906 if self.svfs.exists("journal"):
906 if self.svfs.exists("journal"):
907 self.ui.status(_("rolling back interrupted transaction\n"))
907 self.ui.status(_("rolling back interrupted transaction\n"))
908 transaction.rollback(self.sopener, "journal",
908 transaction.rollback(self.sopener, "journal",
909 self.ui.warn)
909 self.ui.warn)
910 self.invalidate()
910 self.invalidate()
911 return True
911 return True
912 else:
912 else:
913 self.ui.warn(_("no interrupted transaction available\n"))
913 self.ui.warn(_("no interrupted transaction available\n"))
914 return False
914 return False
915 finally:
915 finally:
916 lock.release()
916 lock.release()
917
917
918 def rollback(self, dryrun=False, force=False):
918 def rollback(self, dryrun=False, force=False):
919 wlock = lock = None
919 wlock = lock = None
920 try:
920 try:
921 wlock = self.wlock()
921 wlock = self.wlock()
922 lock = self.lock()
922 lock = self.lock()
923 if self.svfs.exists("undo"):
923 if self.svfs.exists("undo"):
924 return self._rollback(dryrun, force)
924 return self._rollback(dryrun, force)
925 else:
925 else:
926 self.ui.warn(_("no rollback information available\n"))
926 self.ui.warn(_("no rollback information available\n"))
927 return 1
927 return 1
928 finally:
928 finally:
929 release(lock, wlock)
929 release(lock, wlock)
930
930
931 @unfilteredmethod # Until we get smarter cache management
931 @unfilteredmethod # Until we get smarter cache management
932 def _rollback(self, dryrun, force):
932 def _rollback(self, dryrun, force):
933 ui = self.ui
933 ui = self.ui
934 try:
934 try:
935 args = self.opener.read('undo.desc').splitlines()
935 args = self.opener.read('undo.desc').splitlines()
936 (oldlen, desc, detail) = (int(args[0]), args[1], None)
936 (oldlen, desc, detail) = (int(args[0]), args[1], None)
937 if len(args) >= 3:
937 if len(args) >= 3:
938 detail = args[2]
938 detail = args[2]
939 oldtip = oldlen - 1
939 oldtip = oldlen - 1
940
940
941 if detail and ui.verbose:
941 if detail and ui.verbose:
942 msg = (_('repository tip rolled back to revision %s'
942 msg = (_('repository tip rolled back to revision %s'
943 ' (undo %s: %s)\n')
943 ' (undo %s: %s)\n')
944 % (oldtip, desc, detail))
944 % (oldtip, desc, detail))
945 else:
945 else:
946 msg = (_('repository tip rolled back to revision %s'
946 msg = (_('repository tip rolled back to revision %s'
947 ' (undo %s)\n')
947 ' (undo %s)\n')
948 % (oldtip, desc))
948 % (oldtip, desc))
949 except IOError:
949 except IOError:
950 msg = _('rolling back unknown transaction\n')
950 msg = _('rolling back unknown transaction\n')
951 desc = None
951 desc = None
952
952
953 if not force and self['.'] != self['tip'] and desc == 'commit':
953 if not force and self['.'] != self['tip'] and desc == 'commit':
954 raise util.Abort(
954 raise util.Abort(
955 _('rollback of last commit while not checked out '
955 _('rollback of last commit while not checked out '
956 'may lose data'), hint=_('use -f to force'))
956 'may lose data'), hint=_('use -f to force'))
957
957
958 ui.status(msg)
958 ui.status(msg)
959 if dryrun:
959 if dryrun:
960 return 0
960 return 0
961
961
962 parents = self.dirstate.parents()
962 parents = self.dirstate.parents()
963 self.destroying()
963 self.destroying()
964 transaction.rollback(self.sopener, 'undo', ui.warn)
964 transaction.rollback(self.sopener, 'undo', ui.warn)
965 if self.vfs.exists('undo.bookmarks'):
965 if self.vfs.exists('undo.bookmarks'):
966 self.vfs.rename('undo.bookmarks', 'bookmarks')
966 self.vfs.rename('undo.bookmarks', 'bookmarks')
967 if self.svfs.exists('undo.phaseroots'):
967 if self.svfs.exists('undo.phaseroots'):
968 self.svfs.rename('undo.phaseroots', 'phaseroots')
968 self.svfs.rename('undo.phaseroots', 'phaseroots')
969 self.invalidate()
969 self.invalidate()
970
970
971 parentgone = (parents[0] not in self.changelog.nodemap or
971 parentgone = (parents[0] not in self.changelog.nodemap or
972 parents[1] not in self.changelog.nodemap)
972 parents[1] not in self.changelog.nodemap)
973 if parentgone:
973 if parentgone:
974 self.vfs.rename('undo.dirstate', 'dirstate')
974 self.vfs.rename('undo.dirstate', 'dirstate')
975 try:
975 try:
976 branch = self.opener.read('undo.branch')
976 branch = self.opener.read('undo.branch')
977 self.dirstate.setbranch(encoding.tolocal(branch))
977 self.dirstate.setbranch(encoding.tolocal(branch))
978 except IOError:
978 except IOError:
979 ui.warn(_('named branch could not be reset: '
979 ui.warn(_('named branch could not be reset: '
980 'current branch is still \'%s\'\n')
980 'current branch is still \'%s\'\n')
981 % self.dirstate.branch())
981 % self.dirstate.branch())
982
982
983 self.dirstate.invalidate()
983 self.dirstate.invalidate()
984 parents = tuple([p.rev() for p in self.parents()])
984 parents = tuple([p.rev() for p in self.parents()])
985 if len(parents) > 1:
985 if len(parents) > 1:
986 ui.status(_('working directory now based on '
986 ui.status(_('working directory now based on '
987 'revisions %d and %d\n') % parents)
987 'revisions %d and %d\n') % parents)
988 else:
988 else:
989 ui.status(_('working directory now based on '
989 ui.status(_('working directory now based on '
990 'revision %d\n') % parents)
990 'revision %d\n') % parents)
991 # TODO: if we know which new heads may result from this rollback, pass
991 # TODO: if we know which new heads may result from this rollback, pass
992 # them to destroy(), which will prevent the branchhead cache from being
992 # them to destroy(), which will prevent the branchhead cache from being
993 # invalidated.
993 # invalidated.
994 self.destroyed()
994 self.destroyed()
995 return 0
995 return 0
996
996
997 def invalidatecaches(self):
997 def invalidatecaches(self):
998
998
999 if '_tagscache' in vars(self):
999 if '_tagscache' in vars(self):
1000 # can't use delattr on proxy
1000 # can't use delattr on proxy
1001 del self.__dict__['_tagscache']
1001 del self.__dict__['_tagscache']
1002
1002
1003 self.unfiltered()._branchcaches.clear()
1003 self.unfiltered()._branchcaches.clear()
1004 self.invalidatevolatilesets()
1004 self.invalidatevolatilesets()
1005
1005
1006 def invalidatevolatilesets(self):
1006 def invalidatevolatilesets(self):
1007 self.filteredrevcache.clear()
1007 self.filteredrevcache.clear()
1008 obsolete.clearobscaches(self)
1008 obsolete.clearobscaches(self)
1009
1009
1010 def invalidatedirstate(self):
1010 def invalidatedirstate(self):
1011 '''Invalidates the dirstate, causing the next call to dirstate
1011 '''Invalidates the dirstate, causing the next call to dirstate
1012 to check if it was modified since the last time it was read,
1012 to check if it was modified since the last time it was read,
1013 rereading it if it has.
1013 rereading it if it has.
1014
1014
1015 This is different to dirstate.invalidate() that it doesn't always
1015 This is different to dirstate.invalidate() that it doesn't always
1016 rereads the dirstate. Use dirstate.invalidate() if you want to
1016 rereads the dirstate. Use dirstate.invalidate() if you want to
1017 explicitly read the dirstate again (i.e. restoring it to a previous
1017 explicitly read the dirstate again (i.e. restoring it to a previous
1018 known good state).'''
1018 known good state).'''
1019 if hasunfilteredcache(self, 'dirstate'):
1019 if hasunfilteredcache(self, 'dirstate'):
1020 for k in self.dirstate._filecache:
1020 for k in self.dirstate._filecache:
1021 try:
1021 try:
1022 delattr(self.dirstate, k)
1022 delattr(self.dirstate, k)
1023 except AttributeError:
1023 except AttributeError:
1024 pass
1024 pass
1025 delattr(self.unfiltered(), 'dirstate')
1025 delattr(self.unfiltered(), 'dirstate')
1026
1026
1027 def invalidate(self):
1027 def invalidate(self):
1028 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1028 unfiltered = self.unfiltered() # all file caches are stored unfiltered
1029 for k in self._filecache:
1029 for k in self._filecache:
1030 # dirstate is invalidated separately in invalidatedirstate()
1030 # dirstate is invalidated separately in invalidatedirstate()
1031 if k == 'dirstate':
1031 if k == 'dirstate':
1032 continue
1032 continue
1033
1033
1034 try:
1034 try:
1035 delattr(unfiltered, k)
1035 delattr(unfiltered, k)
1036 except AttributeError:
1036 except AttributeError:
1037 pass
1037 pass
1038 self.invalidatecaches()
1038 self.invalidatecaches()
1039 self.store.invalidatecaches()
1039 self.store.invalidatecaches()
1040
1040
1041 def invalidateall(self):
1041 def invalidateall(self):
1042 '''Fully invalidates both store and non-store parts, causing the
1042 '''Fully invalidates both store and non-store parts, causing the
1043 subsequent operation to reread any outside changes.'''
1043 subsequent operation to reread any outside changes.'''
1044 # extension should hook this to invalidate its caches
1044 # extension should hook this to invalidate its caches
1045 self.invalidate()
1045 self.invalidate()
1046 self.invalidatedirstate()
1046 self.invalidatedirstate()
1047
1047
1048 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1048 def _lock(self, vfs, lockname, wait, releasefn, acquirefn, desc):
1049 try:
1049 try:
1050 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1050 l = lockmod.lock(vfs, lockname, 0, releasefn, desc=desc)
1051 except error.LockHeld, inst:
1051 except error.LockHeld, inst:
1052 if not wait:
1052 if not wait:
1053 raise
1053 raise
1054 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1054 self.ui.warn(_("waiting for lock on %s held by %r\n") %
1055 (desc, inst.locker))
1055 (desc, inst.locker))
1056 # default to 600 seconds timeout
1056 # default to 600 seconds timeout
1057 l = lockmod.lock(vfs, lockname,
1057 l = lockmod.lock(vfs, lockname,
1058 int(self.ui.config("ui", "timeout", "600")),
1058 int(self.ui.config("ui", "timeout", "600")),
1059 releasefn, desc=desc)
1059 releasefn, desc=desc)
1060 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1060 self.ui.warn(_("got lock after %s seconds\n") % l.delay)
1061 if acquirefn:
1061 if acquirefn:
1062 acquirefn()
1062 acquirefn()
1063 return l
1063 return l
1064
1064
1065 def _afterlock(self, callback):
1065 def _afterlock(self, callback):
1066 """add a callback to the current repository lock.
1066 """add a callback to the current repository lock.
1067
1067
1068 The callback will be executed on lock release."""
1068 The callback will be executed on lock release."""
1069 l = self._lockref and self._lockref()
1069 l = self._lockref and self._lockref()
1070 if l:
1070 if l:
1071 l.postrelease.append(callback)
1071 l.postrelease.append(callback)
1072 else:
1072 else:
1073 callback()
1073 callback()
1074
1074
1075 def lock(self, wait=True):
1075 def lock(self, wait=True):
1076 '''Lock the repository store (.hg/store) and return a weak reference
1076 '''Lock the repository store (.hg/store) and return a weak reference
1077 to the lock. Use this before modifying the store (e.g. committing or
1077 to the lock. Use this before modifying the store (e.g. committing or
1078 stripping). If you are opening a transaction, get a lock as well.)'''
1078 stripping). If you are opening a transaction, get a lock as well.)'''
1079 l = self._lockref and self._lockref()
1079 l = self._lockref and self._lockref()
1080 if l is not None and l.held:
1080 if l is not None and l.held:
1081 l.lock()
1081 l.lock()
1082 return l
1082 return l
1083
1083
1084 def unlock():
1084 def unlock():
1085 for k, ce in self._filecache.items():
1085 for k, ce in self._filecache.items():
1086 if k == 'dirstate' or k not in self.__dict__:
1086 if k == 'dirstate' or k not in self.__dict__:
1087 continue
1087 continue
1088 ce.refresh()
1088 ce.refresh()
1089
1089
1090 l = self._lock(self.svfs, "lock", wait, unlock,
1090 l = self._lock(self.svfs, "lock", wait, unlock,
1091 self.invalidate, _('repository %s') % self.origroot)
1091 self.invalidate, _('repository %s') % self.origroot)
1092 self._lockref = weakref.ref(l)
1092 self._lockref = weakref.ref(l)
1093 return l
1093 return l
1094
1094
1095 def wlock(self, wait=True):
1095 def wlock(self, wait=True):
1096 '''Lock the non-store parts of the repository (everything under
1096 '''Lock the non-store parts of the repository (everything under
1097 .hg except .hg/store) and return a weak reference to the lock.
1097 .hg except .hg/store) and return a weak reference to the lock.
1098 Use this before modifying files in .hg.'''
1098 Use this before modifying files in .hg.'''
1099 l = self._wlockref and self._wlockref()
1099 l = self._wlockref and self._wlockref()
1100 if l is not None and l.held:
1100 if l is not None and l.held:
1101 l.lock()
1101 l.lock()
1102 return l
1102 return l
1103
1103
1104 def unlock():
1104 def unlock():
1105 self.dirstate.write()
1105 if self.dirstate.pendingparentchange():
1106 self.dirstate.invalidate()
1107 else:
1108 self.dirstate.write()
1109
1106 self._filecache['dirstate'].refresh()
1110 self._filecache['dirstate'].refresh()
1107
1111
1108 l = self._lock(self.vfs, "wlock", wait, unlock,
1112 l = self._lock(self.vfs, "wlock", wait, unlock,
1109 self.invalidatedirstate, _('working directory of %s') %
1113 self.invalidatedirstate, _('working directory of %s') %
1110 self.origroot)
1114 self.origroot)
1111 self._wlockref = weakref.ref(l)
1115 self._wlockref = weakref.ref(l)
1112 return l
1116 return l
1113
1117
1114 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1118 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
1115 """
1119 """
1116 commit an individual file as part of a larger transaction
1120 commit an individual file as part of a larger transaction
1117 """
1121 """
1118
1122
1119 fname = fctx.path()
1123 fname = fctx.path()
1120 text = fctx.data()
1124 text = fctx.data()
1121 flog = self.file(fname)
1125 flog = self.file(fname)
1122 fparent1 = manifest1.get(fname, nullid)
1126 fparent1 = manifest1.get(fname, nullid)
1123 fparent2 = fparent2o = manifest2.get(fname, nullid)
1127 fparent2 = fparent2o = manifest2.get(fname, nullid)
1124
1128
1125 meta = {}
1129 meta = {}
1126 copy = fctx.renamed()
1130 copy = fctx.renamed()
1127 if copy and copy[0] != fname:
1131 if copy and copy[0] != fname:
1128 # Mark the new revision of this file as a copy of another
1132 # Mark the new revision of this file as a copy of another
1129 # file. This copy data will effectively act as a parent
1133 # file. This copy data will effectively act as a parent
1130 # of this new revision. If this is a merge, the first
1134 # of this new revision. If this is a merge, the first
1131 # parent will be the nullid (meaning "look up the copy data")
1135 # parent will be the nullid (meaning "look up the copy data")
1132 # and the second one will be the other parent. For example:
1136 # and the second one will be the other parent. For example:
1133 #
1137 #
1134 # 0 --- 1 --- 3 rev1 changes file foo
1138 # 0 --- 1 --- 3 rev1 changes file foo
1135 # \ / rev2 renames foo to bar and changes it
1139 # \ / rev2 renames foo to bar and changes it
1136 # \- 2 -/ rev3 should have bar with all changes and
1140 # \- 2 -/ rev3 should have bar with all changes and
1137 # should record that bar descends from
1141 # should record that bar descends from
1138 # bar in rev2 and foo in rev1
1142 # bar in rev2 and foo in rev1
1139 #
1143 #
1140 # this allows this merge to succeed:
1144 # this allows this merge to succeed:
1141 #
1145 #
1142 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1146 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
1143 # \ / merging rev3 and rev4 should use bar@rev2
1147 # \ / merging rev3 and rev4 should use bar@rev2
1144 # \- 2 --- 4 as the merge base
1148 # \- 2 --- 4 as the merge base
1145 #
1149 #
1146
1150
1147 cfname = copy[0]
1151 cfname = copy[0]
1148 crev = manifest1.get(cfname)
1152 crev = manifest1.get(cfname)
1149 newfparent = fparent2
1153 newfparent = fparent2
1150
1154
1151 if manifest2: # branch merge
1155 if manifest2: # branch merge
1152 if fparent2 == nullid or crev is None: # copied on remote side
1156 if fparent2 == nullid or crev is None: # copied on remote side
1153 if cfname in manifest2:
1157 if cfname in manifest2:
1154 crev = manifest2[cfname]
1158 crev = manifest2[cfname]
1155 newfparent = fparent1
1159 newfparent = fparent1
1156
1160
1157 # find source in nearest ancestor if we've lost track
1161 # find source in nearest ancestor if we've lost track
1158 if not crev:
1162 if not crev:
1159 self.ui.debug(" %s: searching for copy revision for %s\n" %
1163 self.ui.debug(" %s: searching for copy revision for %s\n" %
1160 (fname, cfname))
1164 (fname, cfname))
1161 for ancestor in self[None].ancestors():
1165 for ancestor in self[None].ancestors():
1162 if cfname in ancestor:
1166 if cfname in ancestor:
1163 crev = ancestor[cfname].filenode()
1167 crev = ancestor[cfname].filenode()
1164 break
1168 break
1165
1169
1166 if crev:
1170 if crev:
1167 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1171 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
1168 meta["copy"] = cfname
1172 meta["copy"] = cfname
1169 meta["copyrev"] = hex(crev)
1173 meta["copyrev"] = hex(crev)
1170 fparent1, fparent2 = nullid, newfparent
1174 fparent1, fparent2 = nullid, newfparent
1171 else:
1175 else:
1172 self.ui.warn(_("warning: can't find ancestor for '%s' "
1176 self.ui.warn(_("warning: can't find ancestor for '%s' "
1173 "copied from '%s'!\n") % (fname, cfname))
1177 "copied from '%s'!\n") % (fname, cfname))
1174
1178
1175 elif fparent1 == nullid:
1179 elif fparent1 == nullid:
1176 fparent1, fparent2 = fparent2, nullid
1180 fparent1, fparent2 = fparent2, nullid
1177 elif fparent2 != nullid:
1181 elif fparent2 != nullid:
1178 # is one parent an ancestor of the other?
1182 # is one parent an ancestor of the other?
1179 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1183 fparentancestors = flog.commonancestorsheads(fparent1, fparent2)
1180 if fparent1 in fparentancestors:
1184 if fparent1 in fparentancestors:
1181 fparent1, fparent2 = fparent2, nullid
1185 fparent1, fparent2 = fparent2, nullid
1182 elif fparent2 in fparentancestors:
1186 elif fparent2 in fparentancestors:
1183 fparent2 = nullid
1187 fparent2 = nullid
1184
1188
1185 # is the file changed?
1189 # is the file changed?
1186 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1190 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
1187 changelist.append(fname)
1191 changelist.append(fname)
1188 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1192 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
1189
1193
1190 # are just the flags changed during merge?
1194 # are just the flags changed during merge?
1191 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1195 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
1192 changelist.append(fname)
1196 changelist.append(fname)
1193
1197
1194 return fparent1
1198 return fparent1
1195
1199
1196 @unfilteredmethod
1200 @unfilteredmethod
1197 def commit(self, text="", user=None, date=None, match=None, force=False,
1201 def commit(self, text="", user=None, date=None, match=None, force=False,
1198 editor=False, extra={}):
1202 editor=False, extra={}):
1199 """Add a new revision to current repository.
1203 """Add a new revision to current repository.
1200
1204
1201 Revision information is gathered from the working directory,
1205 Revision information is gathered from the working directory,
1202 match can be used to filter the committed files. If editor is
1206 match can be used to filter the committed files. If editor is
1203 supplied, it is called to get a commit message.
1207 supplied, it is called to get a commit message.
1204 """
1208 """
1205
1209
1206 def fail(f, msg):
1210 def fail(f, msg):
1207 raise util.Abort('%s: %s' % (f, msg))
1211 raise util.Abort('%s: %s' % (f, msg))
1208
1212
1209 if not match:
1213 if not match:
1210 match = matchmod.always(self.root, '')
1214 match = matchmod.always(self.root, '')
1211
1215
1212 if not force:
1216 if not force:
1213 vdirs = []
1217 vdirs = []
1214 match.explicitdir = vdirs.append
1218 match.explicitdir = vdirs.append
1215 match.bad = fail
1219 match.bad = fail
1216
1220
1217 wlock = self.wlock()
1221 wlock = self.wlock()
1218 try:
1222 try:
1219 wctx = self[None]
1223 wctx = self[None]
1220 merge = len(wctx.parents()) > 1
1224 merge = len(wctx.parents()) > 1
1221
1225
1222 if (not force and merge and match and
1226 if (not force and merge and match and
1223 (match.files() or match.anypats())):
1227 (match.files() or match.anypats())):
1224 raise util.Abort(_('cannot partially commit a merge '
1228 raise util.Abort(_('cannot partially commit a merge '
1225 '(do not specify files or patterns)'))
1229 '(do not specify files or patterns)'))
1226
1230
1227 changes = self.status(match=match, clean=force)
1231 changes = self.status(match=match, clean=force)
1228 if force:
1232 if force:
1229 changes[0].extend(changes[6]) # mq may commit unchanged files
1233 changes[0].extend(changes[6]) # mq may commit unchanged files
1230
1234
1231 # check subrepos
1235 # check subrepos
1232 subs = []
1236 subs = []
1233 commitsubs = set()
1237 commitsubs = set()
1234 newstate = wctx.substate.copy()
1238 newstate = wctx.substate.copy()
1235 # only manage subrepos and .hgsubstate if .hgsub is present
1239 # only manage subrepos and .hgsubstate if .hgsub is present
1236 if '.hgsub' in wctx:
1240 if '.hgsub' in wctx:
1237 # we'll decide whether to track this ourselves, thanks
1241 # we'll decide whether to track this ourselves, thanks
1238 for c in changes[:3]:
1242 for c in changes[:3]:
1239 if '.hgsubstate' in c:
1243 if '.hgsubstate' in c:
1240 c.remove('.hgsubstate')
1244 c.remove('.hgsubstate')
1241
1245
1242 # compare current state to last committed state
1246 # compare current state to last committed state
1243 # build new substate based on last committed state
1247 # build new substate based on last committed state
1244 oldstate = wctx.p1().substate
1248 oldstate = wctx.p1().substate
1245 for s in sorted(newstate.keys()):
1249 for s in sorted(newstate.keys()):
1246 if not match(s):
1250 if not match(s):
1247 # ignore working copy, use old state if present
1251 # ignore working copy, use old state if present
1248 if s in oldstate:
1252 if s in oldstate:
1249 newstate[s] = oldstate[s]
1253 newstate[s] = oldstate[s]
1250 continue
1254 continue
1251 if not force:
1255 if not force:
1252 raise util.Abort(
1256 raise util.Abort(
1253 _("commit with new subrepo %s excluded") % s)
1257 _("commit with new subrepo %s excluded") % s)
1254 if wctx.sub(s).dirty(True):
1258 if wctx.sub(s).dirty(True):
1255 if not self.ui.configbool('ui', 'commitsubrepos'):
1259 if not self.ui.configbool('ui', 'commitsubrepos'):
1256 raise util.Abort(
1260 raise util.Abort(
1257 _("uncommitted changes in subrepo %s") % s,
1261 _("uncommitted changes in subrepo %s") % s,
1258 hint=_("use --subrepos for recursive commit"))
1262 hint=_("use --subrepos for recursive commit"))
1259 subs.append(s)
1263 subs.append(s)
1260 commitsubs.add(s)
1264 commitsubs.add(s)
1261 else:
1265 else:
1262 bs = wctx.sub(s).basestate()
1266 bs = wctx.sub(s).basestate()
1263 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1267 newstate[s] = (newstate[s][0], bs, newstate[s][2])
1264 if oldstate.get(s, (None, None, None))[1] != bs:
1268 if oldstate.get(s, (None, None, None))[1] != bs:
1265 subs.append(s)
1269 subs.append(s)
1266
1270
1267 # check for removed subrepos
1271 # check for removed subrepos
1268 for p in wctx.parents():
1272 for p in wctx.parents():
1269 r = [s for s in p.substate if s not in newstate]
1273 r = [s for s in p.substate if s not in newstate]
1270 subs += [s for s in r if match(s)]
1274 subs += [s for s in r if match(s)]
1271 if subs:
1275 if subs:
1272 if (not match('.hgsub') and
1276 if (not match('.hgsub') and
1273 '.hgsub' in (wctx.modified() + wctx.added())):
1277 '.hgsub' in (wctx.modified() + wctx.added())):
1274 raise util.Abort(
1278 raise util.Abort(
1275 _("can't commit subrepos without .hgsub"))
1279 _("can't commit subrepos without .hgsub"))
1276 changes[0].insert(0, '.hgsubstate')
1280 changes[0].insert(0, '.hgsubstate')
1277
1281
1278 elif '.hgsub' in changes[2]:
1282 elif '.hgsub' in changes[2]:
1279 # clean up .hgsubstate when .hgsub is removed
1283 # clean up .hgsubstate when .hgsub is removed
1280 if ('.hgsubstate' in wctx and
1284 if ('.hgsubstate' in wctx and
1281 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1285 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
1282 changes[2].insert(0, '.hgsubstate')
1286 changes[2].insert(0, '.hgsubstate')
1283
1287
1284 # make sure all explicit patterns are matched
1288 # make sure all explicit patterns are matched
1285 if not force and match.files():
1289 if not force and match.files():
1286 matched = set(changes[0] + changes[1] + changes[2])
1290 matched = set(changes[0] + changes[1] + changes[2])
1287
1291
1288 for f in match.files():
1292 for f in match.files():
1289 f = self.dirstate.normalize(f)
1293 f = self.dirstate.normalize(f)
1290 if f == '.' or f in matched or f in wctx.substate:
1294 if f == '.' or f in matched or f in wctx.substate:
1291 continue
1295 continue
1292 if f in changes[3]: # missing
1296 if f in changes[3]: # missing
1293 fail(f, _('file not found!'))
1297 fail(f, _('file not found!'))
1294 if f in vdirs: # visited directory
1298 if f in vdirs: # visited directory
1295 d = f + '/'
1299 d = f + '/'
1296 for mf in matched:
1300 for mf in matched:
1297 if mf.startswith(d):
1301 if mf.startswith(d):
1298 break
1302 break
1299 else:
1303 else:
1300 fail(f, _("no match under directory!"))
1304 fail(f, _("no match under directory!"))
1301 elif f not in self.dirstate:
1305 elif f not in self.dirstate:
1302 fail(f, _("file not tracked!"))
1306 fail(f, _("file not tracked!"))
1303
1307
1304 cctx = context.workingctx(self, text, user, date, extra, changes)
1308 cctx = context.workingctx(self, text, user, date, extra, changes)
1305
1309
1306 if (not force and not extra.get("close") and not merge
1310 if (not force and not extra.get("close") and not merge
1307 and not cctx.files()
1311 and not cctx.files()
1308 and wctx.branch() == wctx.p1().branch()):
1312 and wctx.branch() == wctx.p1().branch()):
1309 return None
1313 return None
1310
1314
1311 if merge and cctx.deleted():
1315 if merge and cctx.deleted():
1312 raise util.Abort(_("cannot commit merge with missing files"))
1316 raise util.Abort(_("cannot commit merge with missing files"))
1313
1317
1314 ms = mergemod.mergestate(self)
1318 ms = mergemod.mergestate(self)
1315 for f in changes[0]:
1319 for f in changes[0]:
1316 if f in ms and ms[f] == 'u':
1320 if f in ms and ms[f] == 'u':
1317 raise util.Abort(_("unresolved merge conflicts "
1321 raise util.Abort(_("unresolved merge conflicts "
1318 "(see hg help resolve)"))
1322 "(see hg help resolve)"))
1319
1323
1320 if editor:
1324 if editor:
1321 cctx._text = editor(self, cctx, subs)
1325 cctx._text = editor(self, cctx, subs)
1322 edited = (text != cctx._text)
1326 edited = (text != cctx._text)
1323
1327
1324 # Save commit message in case this transaction gets rolled back
1328 # Save commit message in case this transaction gets rolled back
1325 # (e.g. by a pretxncommit hook). Leave the content alone on
1329 # (e.g. by a pretxncommit hook). Leave the content alone on
1326 # the assumption that the user will use the same editor again.
1330 # the assumption that the user will use the same editor again.
1327 msgfn = self.savecommitmessage(cctx._text)
1331 msgfn = self.savecommitmessage(cctx._text)
1328
1332
1329 # commit subs and write new state
1333 # commit subs and write new state
1330 if subs:
1334 if subs:
1331 for s in sorted(commitsubs):
1335 for s in sorted(commitsubs):
1332 sub = wctx.sub(s)
1336 sub = wctx.sub(s)
1333 self.ui.status(_('committing subrepository %s\n') %
1337 self.ui.status(_('committing subrepository %s\n') %
1334 subrepo.subrelpath(sub))
1338 subrepo.subrelpath(sub))
1335 sr = sub.commit(cctx._text, user, date)
1339 sr = sub.commit(cctx._text, user, date)
1336 newstate[s] = (newstate[s][0], sr)
1340 newstate[s] = (newstate[s][0], sr)
1337 subrepo.writestate(self, newstate)
1341 subrepo.writestate(self, newstate)
1338
1342
1339 p1, p2 = self.dirstate.parents()
1343 p1, p2 = self.dirstate.parents()
1340 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1344 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1341 try:
1345 try:
1342 self.hook("precommit", throw=True, parent1=hookp1,
1346 self.hook("precommit", throw=True, parent1=hookp1,
1343 parent2=hookp2)
1347 parent2=hookp2)
1344 ret = self.commitctx(cctx, True)
1348 ret = self.commitctx(cctx, True)
1345 except: # re-raises
1349 except: # re-raises
1346 if edited:
1350 if edited:
1347 self.ui.write(
1351 self.ui.write(
1348 _('note: commit message saved in %s\n') % msgfn)
1352 _('note: commit message saved in %s\n') % msgfn)
1349 raise
1353 raise
1350
1354
1351 # update bookmarks, dirstate and mergestate
1355 # update bookmarks, dirstate and mergestate
1352 bookmarks.update(self, [p1, p2], ret)
1356 bookmarks.update(self, [p1, p2], ret)
1353 cctx.markcommitted(ret)
1357 cctx.markcommitted(ret)
1354 ms.reset()
1358 ms.reset()
1355 finally:
1359 finally:
1356 wlock.release()
1360 wlock.release()
1357
1361
1358 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1362 def commithook(node=hex(ret), parent1=hookp1, parent2=hookp2):
1359 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1363 self.hook("commit", node=node, parent1=parent1, parent2=parent2)
1360 self._afterlock(commithook)
1364 self._afterlock(commithook)
1361 return ret
1365 return ret
1362
1366
1363 @unfilteredmethod
1367 @unfilteredmethod
1364 def commitctx(self, ctx, error=False):
1368 def commitctx(self, ctx, error=False):
1365 """Add a new revision to current repository.
1369 """Add a new revision to current repository.
1366 Revision information is passed via the context argument.
1370 Revision information is passed via the context argument.
1367 """
1371 """
1368
1372
1369 tr = lock = None
1373 tr = lock = None
1370 removed = list(ctx.removed())
1374 removed = list(ctx.removed())
1371 p1, p2 = ctx.p1(), ctx.p2()
1375 p1, p2 = ctx.p1(), ctx.p2()
1372 user = ctx.user()
1376 user = ctx.user()
1373
1377
1374 lock = self.lock()
1378 lock = self.lock()
1375 try:
1379 try:
1376 tr = self.transaction("commit")
1380 tr = self.transaction("commit")
1377 trp = weakref.proxy(tr)
1381 trp = weakref.proxy(tr)
1378
1382
1379 if ctx.files():
1383 if ctx.files():
1380 m1 = p1.manifest().copy()
1384 m1 = p1.manifest().copy()
1381 m2 = p2.manifest()
1385 m2 = p2.manifest()
1382
1386
1383 # check in files
1387 # check in files
1384 new = {}
1388 new = {}
1385 changed = []
1389 changed = []
1386 linkrev = len(self)
1390 linkrev = len(self)
1387 for f in sorted(ctx.modified() + ctx.added()):
1391 for f in sorted(ctx.modified() + ctx.added()):
1388 self.ui.note(f + "\n")
1392 self.ui.note(f + "\n")
1389 try:
1393 try:
1390 fctx = ctx[f]
1394 fctx = ctx[f]
1391 if fctx is None:
1395 if fctx is None:
1392 removed.append(f)
1396 removed.append(f)
1393 else:
1397 else:
1394 new[f] = self._filecommit(fctx, m1, m2, linkrev,
1398 new[f] = self._filecommit(fctx, m1, m2, linkrev,
1395 trp, changed)
1399 trp, changed)
1396 m1.set(f, fctx.flags())
1400 m1.set(f, fctx.flags())
1397 except OSError, inst:
1401 except OSError, inst:
1398 self.ui.warn(_("trouble committing %s!\n") % f)
1402 self.ui.warn(_("trouble committing %s!\n") % f)
1399 raise
1403 raise
1400 except IOError, inst:
1404 except IOError, inst:
1401 errcode = getattr(inst, 'errno', errno.ENOENT)
1405 errcode = getattr(inst, 'errno', errno.ENOENT)
1402 if error or errcode and errcode != errno.ENOENT:
1406 if error or errcode and errcode != errno.ENOENT:
1403 self.ui.warn(_("trouble committing %s!\n") % f)
1407 self.ui.warn(_("trouble committing %s!\n") % f)
1404 raise
1408 raise
1405
1409
1406 # update manifest
1410 # update manifest
1407 m1.update(new)
1411 m1.update(new)
1408 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1412 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1409 drop = [f for f in removed if f in m1]
1413 drop = [f for f in removed if f in m1]
1410 for f in drop:
1414 for f in drop:
1411 del m1[f]
1415 del m1[f]
1412 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1416 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1413 p2.manifestnode(), (new, drop))
1417 p2.manifestnode(), (new, drop))
1414 files = changed + removed
1418 files = changed + removed
1415 else:
1419 else:
1416 mn = p1.manifestnode()
1420 mn = p1.manifestnode()
1417 files = []
1421 files = []
1418
1422
1419 # update changelog
1423 # update changelog
1420 self.changelog.delayupdate()
1424 self.changelog.delayupdate()
1421 n = self.changelog.add(mn, files, ctx.description(),
1425 n = self.changelog.add(mn, files, ctx.description(),
1422 trp, p1.node(), p2.node(),
1426 trp, p1.node(), p2.node(),
1423 user, ctx.date(), ctx.extra().copy())
1427 user, ctx.date(), ctx.extra().copy())
1424 p = lambda: self.changelog.writepending() and self.root or ""
1428 p = lambda: self.changelog.writepending() and self.root or ""
1425 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1429 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1426 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1430 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1427 parent2=xp2, pending=p)
1431 parent2=xp2, pending=p)
1428 self.changelog.finalize(trp)
1432 self.changelog.finalize(trp)
1429 # set the new commit is proper phase
1433 # set the new commit is proper phase
1430 targetphase = subrepo.newcommitphase(self.ui, ctx)
1434 targetphase = subrepo.newcommitphase(self.ui, ctx)
1431 if targetphase:
1435 if targetphase:
1432 # retract boundary do not alter parent changeset.
1436 # retract boundary do not alter parent changeset.
1433 # if a parent have higher the resulting phase will
1437 # if a parent have higher the resulting phase will
1434 # be compliant anyway
1438 # be compliant anyway
1435 #
1439 #
1436 # if minimal phase was 0 we don't need to retract anything
1440 # if minimal phase was 0 we don't need to retract anything
1437 phases.retractboundary(self, tr, targetphase, [n])
1441 phases.retractboundary(self, tr, targetphase, [n])
1438 tr.close()
1442 tr.close()
1439 branchmap.updatecache(self.filtered('served'))
1443 branchmap.updatecache(self.filtered('served'))
1440 return n
1444 return n
1441 finally:
1445 finally:
1442 if tr:
1446 if tr:
1443 tr.release()
1447 tr.release()
1444 lock.release()
1448 lock.release()
1445
1449
1446 @unfilteredmethod
1450 @unfilteredmethod
1447 def destroying(self):
1451 def destroying(self):
1448 '''Inform the repository that nodes are about to be destroyed.
1452 '''Inform the repository that nodes are about to be destroyed.
1449 Intended for use by strip and rollback, so there's a common
1453 Intended for use by strip and rollback, so there's a common
1450 place for anything that has to be done before destroying history.
1454 place for anything that has to be done before destroying history.
1451
1455
1452 This is mostly useful for saving state that is in memory and waiting
1456 This is mostly useful for saving state that is in memory and waiting
1453 to be flushed when the current lock is released. Because a call to
1457 to be flushed when the current lock is released. Because a call to
1454 destroyed is imminent, the repo will be invalidated causing those
1458 destroyed is imminent, the repo will be invalidated causing those
1455 changes to stay in memory (waiting for the next unlock), or vanish
1459 changes to stay in memory (waiting for the next unlock), or vanish
1456 completely.
1460 completely.
1457 '''
1461 '''
1458 # When using the same lock to commit and strip, the phasecache is left
1462 # When using the same lock to commit and strip, the phasecache is left
1459 # dirty after committing. Then when we strip, the repo is invalidated,
1463 # dirty after committing. Then when we strip, the repo is invalidated,
1460 # causing those changes to disappear.
1464 # causing those changes to disappear.
1461 if '_phasecache' in vars(self):
1465 if '_phasecache' in vars(self):
1462 self._phasecache.write()
1466 self._phasecache.write()
1463
1467
1464 @unfilteredmethod
1468 @unfilteredmethod
1465 def destroyed(self):
1469 def destroyed(self):
1466 '''Inform the repository that nodes have been destroyed.
1470 '''Inform the repository that nodes have been destroyed.
1467 Intended for use by strip and rollback, so there's a common
1471 Intended for use by strip and rollback, so there's a common
1468 place for anything that has to be done after destroying history.
1472 place for anything that has to be done after destroying history.
1469 '''
1473 '''
1470 # When one tries to:
1474 # When one tries to:
1471 # 1) destroy nodes thus calling this method (e.g. strip)
1475 # 1) destroy nodes thus calling this method (e.g. strip)
1472 # 2) use phasecache somewhere (e.g. commit)
1476 # 2) use phasecache somewhere (e.g. commit)
1473 #
1477 #
1474 # then 2) will fail because the phasecache contains nodes that were
1478 # then 2) will fail because the phasecache contains nodes that were
1475 # removed. We can either remove phasecache from the filecache,
1479 # removed. We can either remove phasecache from the filecache,
1476 # causing it to reload next time it is accessed, or simply filter
1480 # causing it to reload next time it is accessed, or simply filter
1477 # the removed nodes now and write the updated cache.
1481 # the removed nodes now and write the updated cache.
1478 self._phasecache.filterunknown(self)
1482 self._phasecache.filterunknown(self)
1479 self._phasecache.write()
1483 self._phasecache.write()
1480
1484
1481 # update the 'served' branch cache to help read only server process
1485 # update the 'served' branch cache to help read only server process
1482 # Thanks to branchcache collaboration this is done from the nearest
1486 # Thanks to branchcache collaboration this is done from the nearest
1483 # filtered subset and it is expected to be fast.
1487 # filtered subset and it is expected to be fast.
1484 branchmap.updatecache(self.filtered('served'))
1488 branchmap.updatecache(self.filtered('served'))
1485
1489
1486 # Ensure the persistent tag cache is updated. Doing it now
1490 # Ensure the persistent tag cache is updated. Doing it now
1487 # means that the tag cache only has to worry about destroyed
1491 # means that the tag cache only has to worry about destroyed
1488 # heads immediately after a strip/rollback. That in turn
1492 # heads immediately after a strip/rollback. That in turn
1489 # guarantees that "cachetip == currenttip" (comparing both rev
1493 # guarantees that "cachetip == currenttip" (comparing both rev
1490 # and node) always means no nodes have been added or destroyed.
1494 # and node) always means no nodes have been added or destroyed.
1491
1495
1492 # XXX this is suboptimal when qrefresh'ing: we strip the current
1496 # XXX this is suboptimal when qrefresh'ing: we strip the current
1493 # head, refresh the tag cache, then immediately add a new head.
1497 # head, refresh the tag cache, then immediately add a new head.
1494 # But I think doing it this way is necessary for the "instant
1498 # But I think doing it this way is necessary for the "instant
1495 # tag cache retrieval" case to work.
1499 # tag cache retrieval" case to work.
1496 self.invalidate()
1500 self.invalidate()
1497
1501
1498 def walk(self, match, node=None):
1502 def walk(self, match, node=None):
1499 '''
1503 '''
1500 walk recursively through the directory tree or a given
1504 walk recursively through the directory tree or a given
1501 changeset, finding all files matched by the match
1505 changeset, finding all files matched by the match
1502 function
1506 function
1503 '''
1507 '''
1504 return self[node].walk(match)
1508 return self[node].walk(match)
1505
1509
1506 def status(self, node1='.', node2=None, match=None,
1510 def status(self, node1='.', node2=None, match=None,
1507 ignored=False, clean=False, unknown=False,
1511 ignored=False, clean=False, unknown=False,
1508 listsubrepos=False):
1512 listsubrepos=False):
1509 '''a convenience method that calls node1.status(node2)'''
1513 '''a convenience method that calls node1.status(node2)'''
1510 return self[node1].status(node2, match, ignored, clean, unknown,
1514 return self[node1].status(node2, match, ignored, clean, unknown,
1511 listsubrepos)
1515 listsubrepos)
1512
1516
1513 def heads(self, start=None):
1517 def heads(self, start=None):
1514 heads = self.changelog.heads(start)
1518 heads = self.changelog.heads(start)
1515 # sort the output in rev descending order
1519 # sort the output in rev descending order
1516 return sorted(heads, key=self.changelog.rev, reverse=True)
1520 return sorted(heads, key=self.changelog.rev, reverse=True)
1517
1521
1518 def branchheads(self, branch=None, start=None, closed=False):
1522 def branchheads(self, branch=None, start=None, closed=False):
1519 '''return a (possibly filtered) list of heads for the given branch
1523 '''return a (possibly filtered) list of heads for the given branch
1520
1524
1521 Heads are returned in topological order, from newest to oldest.
1525 Heads are returned in topological order, from newest to oldest.
1522 If branch is None, use the dirstate branch.
1526 If branch is None, use the dirstate branch.
1523 If start is not None, return only heads reachable from start.
1527 If start is not None, return only heads reachable from start.
1524 If closed is True, return heads that are marked as closed as well.
1528 If closed is True, return heads that are marked as closed as well.
1525 '''
1529 '''
1526 if branch is None:
1530 if branch is None:
1527 branch = self[None].branch()
1531 branch = self[None].branch()
1528 branches = self.branchmap()
1532 branches = self.branchmap()
1529 if branch not in branches:
1533 if branch not in branches:
1530 return []
1534 return []
1531 # the cache returns heads ordered lowest to highest
1535 # the cache returns heads ordered lowest to highest
1532 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1536 bheads = list(reversed(branches.branchheads(branch, closed=closed)))
1533 if start is not None:
1537 if start is not None:
1534 # filter out the heads that cannot be reached from startrev
1538 # filter out the heads that cannot be reached from startrev
1535 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1539 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1536 bheads = [h for h in bheads if h in fbheads]
1540 bheads = [h for h in bheads if h in fbheads]
1537 return bheads
1541 return bheads
1538
1542
1539 def branches(self, nodes):
1543 def branches(self, nodes):
1540 if not nodes:
1544 if not nodes:
1541 nodes = [self.changelog.tip()]
1545 nodes = [self.changelog.tip()]
1542 b = []
1546 b = []
1543 for n in nodes:
1547 for n in nodes:
1544 t = n
1548 t = n
1545 while True:
1549 while True:
1546 p = self.changelog.parents(n)
1550 p = self.changelog.parents(n)
1547 if p[1] != nullid or p[0] == nullid:
1551 if p[1] != nullid or p[0] == nullid:
1548 b.append((t, n, p[0], p[1]))
1552 b.append((t, n, p[0], p[1]))
1549 break
1553 break
1550 n = p[0]
1554 n = p[0]
1551 return b
1555 return b
1552
1556
1553 def between(self, pairs):
1557 def between(self, pairs):
1554 r = []
1558 r = []
1555
1559
1556 for top, bottom in pairs:
1560 for top, bottom in pairs:
1557 n, l, i = top, [], 0
1561 n, l, i = top, [], 0
1558 f = 1
1562 f = 1
1559
1563
1560 while n != bottom and n != nullid:
1564 while n != bottom and n != nullid:
1561 p = self.changelog.parents(n)[0]
1565 p = self.changelog.parents(n)[0]
1562 if i == f:
1566 if i == f:
1563 l.append(n)
1567 l.append(n)
1564 f = f * 2
1568 f = f * 2
1565 n = p
1569 n = p
1566 i += 1
1570 i += 1
1567
1571
1568 r.append(l)
1572 r.append(l)
1569
1573
1570 return r
1574 return r
1571
1575
1572 def pull(self, remote, heads=None, force=False):
1576 def pull(self, remote, heads=None, force=False):
1573 return exchange.pull (self, remote, heads, force)
1577 return exchange.pull (self, remote, heads, force)
1574
1578
1575 def checkpush(self, pushop):
1579 def checkpush(self, pushop):
1576 """Extensions can override this function if additional checks have
1580 """Extensions can override this function if additional checks have
1577 to be performed before pushing, or call it if they override push
1581 to be performed before pushing, or call it if they override push
1578 command.
1582 command.
1579 """
1583 """
1580 pass
1584 pass
1581
1585
1582 @unfilteredpropertycache
1586 @unfilteredpropertycache
1583 def prepushoutgoinghooks(self):
1587 def prepushoutgoinghooks(self):
1584 """Return util.hooks consists of "(repo, remote, outgoing)"
1588 """Return util.hooks consists of "(repo, remote, outgoing)"
1585 functions, which are called before pushing changesets.
1589 functions, which are called before pushing changesets.
1586 """
1590 """
1587 return util.hooks()
1591 return util.hooks()
1588
1592
1589 def push(self, remote, force=False, revs=None, newbranch=False):
1593 def push(self, remote, force=False, revs=None, newbranch=False):
1590 return exchange.push(self, remote, force, revs, newbranch)
1594 return exchange.push(self, remote, force, revs, newbranch)
1591
1595
1592 def stream_in(self, remote, requirements):
1596 def stream_in(self, remote, requirements):
1593 lock = self.lock()
1597 lock = self.lock()
1594 try:
1598 try:
1595 # Save remote branchmap. We will use it later
1599 # Save remote branchmap. We will use it later
1596 # to speed up branchcache creation
1600 # to speed up branchcache creation
1597 rbranchmap = None
1601 rbranchmap = None
1598 if remote.capable("branchmap"):
1602 if remote.capable("branchmap"):
1599 rbranchmap = remote.branchmap()
1603 rbranchmap = remote.branchmap()
1600
1604
1601 fp = remote.stream_out()
1605 fp = remote.stream_out()
1602 l = fp.readline()
1606 l = fp.readline()
1603 try:
1607 try:
1604 resp = int(l)
1608 resp = int(l)
1605 except ValueError:
1609 except ValueError:
1606 raise error.ResponseError(
1610 raise error.ResponseError(
1607 _('unexpected response from remote server:'), l)
1611 _('unexpected response from remote server:'), l)
1608 if resp == 1:
1612 if resp == 1:
1609 raise util.Abort(_('operation forbidden by server'))
1613 raise util.Abort(_('operation forbidden by server'))
1610 elif resp == 2:
1614 elif resp == 2:
1611 raise util.Abort(_('locking the remote repository failed'))
1615 raise util.Abort(_('locking the remote repository failed'))
1612 elif resp != 0:
1616 elif resp != 0:
1613 raise util.Abort(_('the server sent an unknown error code'))
1617 raise util.Abort(_('the server sent an unknown error code'))
1614 self.ui.status(_('streaming all changes\n'))
1618 self.ui.status(_('streaming all changes\n'))
1615 l = fp.readline()
1619 l = fp.readline()
1616 try:
1620 try:
1617 total_files, total_bytes = map(int, l.split(' ', 1))
1621 total_files, total_bytes = map(int, l.split(' ', 1))
1618 except (ValueError, TypeError):
1622 except (ValueError, TypeError):
1619 raise error.ResponseError(
1623 raise error.ResponseError(
1620 _('unexpected response from remote server:'), l)
1624 _('unexpected response from remote server:'), l)
1621 self.ui.status(_('%d files to transfer, %s of data\n') %
1625 self.ui.status(_('%d files to transfer, %s of data\n') %
1622 (total_files, util.bytecount(total_bytes)))
1626 (total_files, util.bytecount(total_bytes)))
1623 handled_bytes = 0
1627 handled_bytes = 0
1624 self.ui.progress(_('clone'), 0, total=total_bytes)
1628 self.ui.progress(_('clone'), 0, total=total_bytes)
1625 start = time.time()
1629 start = time.time()
1626
1630
1627 tr = self.transaction(_('clone'))
1631 tr = self.transaction(_('clone'))
1628 try:
1632 try:
1629 for i in xrange(total_files):
1633 for i in xrange(total_files):
1630 # XXX doesn't support '\n' or '\r' in filenames
1634 # XXX doesn't support '\n' or '\r' in filenames
1631 l = fp.readline()
1635 l = fp.readline()
1632 try:
1636 try:
1633 name, size = l.split('\0', 1)
1637 name, size = l.split('\0', 1)
1634 size = int(size)
1638 size = int(size)
1635 except (ValueError, TypeError):
1639 except (ValueError, TypeError):
1636 raise error.ResponseError(
1640 raise error.ResponseError(
1637 _('unexpected response from remote server:'), l)
1641 _('unexpected response from remote server:'), l)
1638 if self.ui.debugflag:
1642 if self.ui.debugflag:
1639 self.ui.debug('adding %s (%s)\n' %
1643 self.ui.debug('adding %s (%s)\n' %
1640 (name, util.bytecount(size)))
1644 (name, util.bytecount(size)))
1641 # for backwards compat, name was partially encoded
1645 # for backwards compat, name was partially encoded
1642 ofp = self.sopener(store.decodedir(name), 'w')
1646 ofp = self.sopener(store.decodedir(name), 'w')
1643 for chunk in util.filechunkiter(fp, limit=size):
1647 for chunk in util.filechunkiter(fp, limit=size):
1644 handled_bytes += len(chunk)
1648 handled_bytes += len(chunk)
1645 self.ui.progress(_('clone'), handled_bytes,
1649 self.ui.progress(_('clone'), handled_bytes,
1646 total=total_bytes)
1650 total=total_bytes)
1647 ofp.write(chunk)
1651 ofp.write(chunk)
1648 ofp.close()
1652 ofp.close()
1649 tr.close()
1653 tr.close()
1650 finally:
1654 finally:
1651 tr.release()
1655 tr.release()
1652
1656
1653 # Writing straight to files circumvented the inmemory caches
1657 # Writing straight to files circumvented the inmemory caches
1654 self.invalidate()
1658 self.invalidate()
1655
1659
1656 elapsed = time.time() - start
1660 elapsed = time.time() - start
1657 if elapsed <= 0:
1661 if elapsed <= 0:
1658 elapsed = 0.001
1662 elapsed = 0.001
1659 self.ui.progress(_('clone'), None)
1663 self.ui.progress(_('clone'), None)
1660 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1664 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1661 (util.bytecount(total_bytes), elapsed,
1665 (util.bytecount(total_bytes), elapsed,
1662 util.bytecount(total_bytes / elapsed)))
1666 util.bytecount(total_bytes / elapsed)))
1663
1667
1664 # new requirements = old non-format requirements +
1668 # new requirements = old non-format requirements +
1665 # new format-related
1669 # new format-related
1666 # requirements from the streamed-in repository
1670 # requirements from the streamed-in repository
1667 requirements.update(set(self.requirements) - self.supportedformats)
1671 requirements.update(set(self.requirements) - self.supportedformats)
1668 self._applyrequirements(requirements)
1672 self._applyrequirements(requirements)
1669 self._writerequirements()
1673 self._writerequirements()
1670
1674
1671 if rbranchmap:
1675 if rbranchmap:
1672 rbheads = []
1676 rbheads = []
1673 for bheads in rbranchmap.itervalues():
1677 for bheads in rbranchmap.itervalues():
1674 rbheads.extend(bheads)
1678 rbheads.extend(bheads)
1675
1679
1676 if rbheads:
1680 if rbheads:
1677 rtiprev = max((int(self.changelog.rev(node))
1681 rtiprev = max((int(self.changelog.rev(node))
1678 for node in rbheads))
1682 for node in rbheads))
1679 cache = branchmap.branchcache(rbranchmap,
1683 cache = branchmap.branchcache(rbranchmap,
1680 self[rtiprev].node(),
1684 self[rtiprev].node(),
1681 rtiprev)
1685 rtiprev)
1682 # Try to stick it as low as possible
1686 # Try to stick it as low as possible
1683 # filter above served are unlikely to be fetch from a clone
1687 # filter above served are unlikely to be fetch from a clone
1684 for candidate in ('base', 'immutable', 'served'):
1688 for candidate in ('base', 'immutable', 'served'):
1685 rview = self.filtered(candidate)
1689 rview = self.filtered(candidate)
1686 if cache.validfor(rview):
1690 if cache.validfor(rview):
1687 self._branchcaches[candidate] = cache
1691 self._branchcaches[candidate] = cache
1688 cache.write(rview)
1692 cache.write(rview)
1689 break
1693 break
1690 self.invalidate()
1694 self.invalidate()
1691 return len(self.heads()) + 1
1695 return len(self.heads()) + 1
1692 finally:
1696 finally:
1693 lock.release()
1697 lock.release()
1694
1698
1695 def clone(self, remote, heads=[], stream=False):
1699 def clone(self, remote, heads=[], stream=False):
1696 '''clone remote repository.
1700 '''clone remote repository.
1697
1701
1698 keyword arguments:
1702 keyword arguments:
1699 heads: list of revs to clone (forces use of pull)
1703 heads: list of revs to clone (forces use of pull)
1700 stream: use streaming clone if possible'''
1704 stream: use streaming clone if possible'''
1701
1705
1702 # now, all clients that can request uncompressed clones can
1706 # now, all clients that can request uncompressed clones can
1703 # read repo formats supported by all servers that can serve
1707 # read repo formats supported by all servers that can serve
1704 # them.
1708 # them.
1705
1709
1706 # if revlog format changes, client will have to check version
1710 # if revlog format changes, client will have to check version
1707 # and format flags on "stream" capability, and use
1711 # and format flags on "stream" capability, and use
1708 # uncompressed only if compatible.
1712 # uncompressed only if compatible.
1709
1713
1710 if not stream:
1714 if not stream:
1711 # if the server explicitly prefers to stream (for fast LANs)
1715 # if the server explicitly prefers to stream (for fast LANs)
1712 stream = remote.capable('stream-preferred')
1716 stream = remote.capable('stream-preferred')
1713
1717
1714 if stream and not heads:
1718 if stream and not heads:
1715 # 'stream' means remote revlog format is revlogv1 only
1719 # 'stream' means remote revlog format is revlogv1 only
1716 if remote.capable('stream'):
1720 if remote.capable('stream'):
1717 return self.stream_in(remote, set(('revlogv1',)))
1721 return self.stream_in(remote, set(('revlogv1',)))
1718 # otherwise, 'streamreqs' contains the remote revlog format
1722 # otherwise, 'streamreqs' contains the remote revlog format
1719 streamreqs = remote.capable('streamreqs')
1723 streamreqs = remote.capable('streamreqs')
1720 if streamreqs:
1724 if streamreqs:
1721 streamreqs = set(streamreqs.split(','))
1725 streamreqs = set(streamreqs.split(','))
1722 # if we support it, stream in and adjust our requirements
1726 # if we support it, stream in and adjust our requirements
1723 if not streamreqs - self.supportedformats:
1727 if not streamreqs - self.supportedformats:
1724 return self.stream_in(remote, streamreqs)
1728 return self.stream_in(remote, streamreqs)
1725 return self.pull(remote, heads)
1729 return self.pull(remote, heads)
1726
1730
1727 def pushkey(self, namespace, key, old, new):
1731 def pushkey(self, namespace, key, old, new):
1728 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1732 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1729 old=old, new=new)
1733 old=old, new=new)
1730 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1734 self.ui.debug('pushing key for "%s:%s"\n' % (namespace, key))
1731 ret = pushkey.push(self, namespace, key, old, new)
1735 ret = pushkey.push(self, namespace, key, old, new)
1732 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1736 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1733 ret=ret)
1737 ret=ret)
1734 return ret
1738 return ret
1735
1739
1736 def listkeys(self, namespace):
1740 def listkeys(self, namespace):
1737 self.hook('prelistkeys', throw=True, namespace=namespace)
1741 self.hook('prelistkeys', throw=True, namespace=namespace)
1738 self.ui.debug('listing keys for "%s"\n' % namespace)
1742 self.ui.debug('listing keys for "%s"\n' % namespace)
1739 values = pushkey.list(self, namespace)
1743 values = pushkey.list(self, namespace)
1740 self.hook('listkeys', namespace=namespace, values=values)
1744 self.hook('listkeys', namespace=namespace, values=values)
1741 return values
1745 return values
1742
1746
1743 def debugwireargs(self, one, two, three=None, four=None, five=None):
1747 def debugwireargs(self, one, two, three=None, four=None, five=None):
1744 '''used to test argument passing over the wire'''
1748 '''used to test argument passing over the wire'''
1745 return "%s %s %s %s %s" % (one, two, three, four, five)
1749 return "%s %s %s %s %s" % (one, two, three, four, five)
1746
1750
1747 def savecommitmessage(self, text):
1751 def savecommitmessage(self, text):
1748 fp = self.opener('last-message.txt', 'wb')
1752 fp = self.opener('last-message.txt', 'wb')
1749 try:
1753 try:
1750 fp.write(text)
1754 fp.write(text)
1751 finally:
1755 finally:
1752 fp.close()
1756 fp.close()
1753 return self.pathto(fp.name[len(self.root) + 1:])
1757 return self.pathto(fp.name[len(self.root) + 1:])
1754
1758
1755 # used to avoid circular references so destructors work
1759 # used to avoid circular references so destructors work
1756 def aftertrans(files):
1760 def aftertrans(files):
1757 renamefiles = [tuple(t) for t in files]
1761 renamefiles = [tuple(t) for t in files]
1758 def a():
1762 def a():
1759 for vfs, src, dest in renamefiles:
1763 for vfs, src, dest in renamefiles:
1760 try:
1764 try:
1761 vfs.rename(src, dest)
1765 vfs.rename(src, dest)
1762 except OSError: # journal file does not yet exist
1766 except OSError: # journal file does not yet exist
1763 pass
1767 pass
1764 return a
1768 return a
1765
1769
1766 def undoname(fn):
1770 def undoname(fn):
1767 base, name = os.path.split(fn)
1771 base, name = os.path.split(fn)
1768 assert name.startswith('journal')
1772 assert name.startswith('journal')
1769 return os.path.join(base, name.replace('journal', 'undo', 1))
1773 return os.path.join(base, name.replace('journal', 'undo', 1))
1770
1774
1771 def instance(ui, path, create):
1775 def instance(ui, path, create):
1772 return localrepository(ui, util.urllocalpath(path), create)
1776 return localrepository(ui, util.urllocalpath(path), create)
1773
1777
1774 def islocal(path):
1778 def islocal(path):
1775 return True
1779 return True
General Comments 0
You need to be logged in to leave comments. Login now