##// END OF EJS Templates
check-code: catch misspellings of descendant...
Matt Mackall -
r14549:48ec0763 default
parent child Browse files
Show More
@@ -1,379 +1,380 b''
1 1 #!/usr/bin/env python
2 2 #
3 3 # check-code - a style and portability checker for Mercurial
4 4 #
5 5 # Copyright 2010 Matt Mackall <mpm@selenic.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 import re, glob, os, sys
11 11 import keyword
12 12 import optparse
13 13
14 14 def repquote(m):
15 15 t = re.sub(r"\w", "x", m.group('text'))
16 16 t = re.sub(r"[^\sx]", "o", t)
17 17 return m.group('quote') + t + m.group('quote')
18 18
19 19 def reppython(m):
20 20 comment = m.group('comment')
21 21 if comment:
22 22 return "#" * len(comment)
23 23 return repquote(m)
24 24
25 25 def repcomment(m):
26 26 return m.group(1) + "#" * len(m.group(2))
27 27
28 28 def repccomment(m):
29 29 t = re.sub(r"((?<=\n) )|\S", "x", m.group(2))
30 30 return m.group(1) + t + "*/"
31 31
32 32 def repcallspaces(m):
33 33 t = re.sub(r"\n\s+", "\n", m.group(2))
34 34 return m.group(1) + t
35 35
36 36 def repinclude(m):
37 37 return m.group(1) + "<foo>"
38 38
39 39 def rephere(m):
40 40 t = re.sub(r"\S", "x", m.group(2))
41 41 return m.group(1) + t
42 42
43 43
44 44 testpats = [
45 45 [
46 46 (r'(pushd|popd)', "don't use 'pushd' or 'popd', use 'cd'"),
47 47 (r'\W\$?\(\([^\)]*\)\)', "don't use (()) or $(()), use 'expr'"),
48 48 (r'^function', "don't use 'function', use old style"),
49 49 (r'grep.*-q', "don't use 'grep -q', redirect to /dev/null"),
50 50 (r'echo.*\\n', "don't use 'echo \\n', use printf"),
51 51 (r'echo -n', "don't use 'echo -n', use printf"),
52 52 (r'^diff.*-\w*N', "don't use 'diff -N'"),
53 53 (r'(^| )wc[^|]*$', "filter wc output"),
54 54 (r'head -c', "don't use 'head -c', use 'dd'"),
55 55 (r'ls.*-\w*R', "don't use 'ls -R', use 'find'"),
56 56 (r'printf.*\\\d\d\d', "don't use 'printf \NNN', use Python"),
57 57 (r'printf.*\\x', "don't use printf \\x, use Python"),
58 58 (r'\$\(.*\)', "don't use $(expr), use `expr`"),
59 59 (r'rm -rf \*', "don't use naked rm -rf, target a directory"),
60 60 (r'(^|\|\s*)grep (-\w\s+)*[^|]*[(|]\w',
61 61 "use egrep for extended grep syntax"),
62 62 (r'/bin/', "don't use explicit paths for tools"),
63 63 (r'\$PWD', "don't use $PWD, use `pwd`"),
64 64 (r'[^\n]\Z', "no trailing newline"),
65 65 (r'export.*=', "don't export and assign at once"),
66 66 ('^([^"\']|("[^"]*")|(\'[^\']*\'))*\\^', "^ must be quoted"),
67 67 (r'^source\b', "don't use 'source', use '.'"),
68 68 (r'touch -d', "don't use 'touch -d', use 'touch -t' instead"),
69 69 (r'ls\s+[^|-]+\s+-', "options to 'ls' must come before filenames"),
70 70 (r'[^>]>\s*\$HGRCPATH', "don't overwrite $HGRCPATH, append to it"),
71 71 ],
72 72 # warnings
73 73 []
74 74 ]
75 75
76 76 testfilters = [
77 77 (r"( *)(#([^\n]*\S)?)", repcomment),
78 78 (r"<<(\S+)((.|\n)*?\n\1)", rephere),
79 79 ]
80 80
81 81 uprefix = r"^ \$ "
82 82 uprefixc = r"^ > "
83 83 utestpats = [
84 84 [
85 85 (r'^(\S| $ ).*(\S\s+|^\s+)\n', "trailing whitespace on non-output"),
86 86 (uprefix + r'.*\|\s*sed', "use regex test output patterns instead of sed"),
87 87 (uprefix + r'(true|exit 0)', "explicit zero exit unnecessary"),
88 88 (uprefix + r'.*\$\?', "explicit exit code checks unnecessary"),
89 89 (uprefix + r'.*\|\| echo.*(fail|error)',
90 90 "explicit exit code checks unnecessary"),
91 91 (uprefix + r'set -e', "don't use set -e"),
92 92 (uprefixc + r'( *)\t', "don't use tabs to indent"),
93 93 ],
94 94 # warnings
95 95 []
96 96 ]
97 97
98 98 for i in [0, 1]:
99 99 for p, m in testpats[i]:
100 100 if p.startswith('^'):
101 101 p = uprefix + p[1:]
102 102 else:
103 103 p = uprefix + p
104 104 utestpats[i].append((p, m))
105 105
106 106 utestfilters = [
107 107 (r"( *)(#([^\n]*\S)?)", repcomment),
108 108 ]
109 109
110 110 pypats = [
111 111 [
112 112 (r'^\s*def\s*\w+\s*\(.*,\s*\(',
113 113 "tuple parameter unpacking not available in Python 3+"),
114 114 (r'lambda\s*\(.*,.*\)',
115 115 "tuple parameter unpacking not available in Python 3+"),
116 116 (r'(?<!def)\s+(cmp)\(', "cmp is not available in Python 3+"),
117 117 (r'\breduce\s*\(.*', "reduce is not available in Python 3+"),
118 118 (r'\.has_key\b', "dict.has_key is not available in Python 3+"),
119 119 (r'^\s*\t', "don't use tabs"),
120 120 (r'\S;\s*\n', "semicolon"),
121 121 (r'\w,\w', "missing whitespace after ,"),
122 122 (r'\w[+/*\-<>]\w', "missing whitespace in expression"),
123 123 (r'^\s+\w+=\w+[^,)]$', "missing whitespace in assignment"),
124 124 (r'.{85}', "line too long"),
125 125 (r'[^\n]\Z', "no trailing newline"),
126 126 (r'(\S\s+|^\s+)\n', "trailing whitespace"),
127 127 # (r'^\s+[^_ ][^_. ]+_[^_]+\s*=', "don't use underbars in identifiers"),
128 128 # (r'\w*[a-z][A-Z]\w*\s*=', "don't use camelcase in identifiers"),
129 129 (r'^\s*(if|while|def|class|except|try)\s[^[]*:\s*[^\]#\s]+',
130 130 "linebreak after :"),
131 131 (r'class\s[^(]:', "old-style class, use class foo(object)"),
132 132 (r'\b(%s)\(' % '|'.join(keyword.kwlist),
133 133 "Python keyword is not a function"),
134 134 (r',]', "unneeded trailing ',' in list"),
135 135 # (r'class\s[A-Z][^\(]*\((?!Exception)',
136 136 # "don't capitalize non-exception classes"),
137 137 # (r'in range\(', "use xrange"),
138 138 # (r'^\s*print\s+', "avoid using print in core and extensions"),
139 139 (r'[\x80-\xff]', "non-ASCII character literal"),
140 140 (r'("\')\.format\(', "str.format() not available in Python 2.4"),
141 141 (r'^\s*with\s+', "with not available in Python 2.4"),
142 142 (r'\.isdisjoint\(', "set.isdisjoint not available in Python 2.4"),
143 143 (r'^\s*except.* as .*:', "except as not available in Python 2.4"),
144 144 (r'^\s*os\.path\.relpath', "relpath not available in Python 2.4"),
145 145 (r'(?<!def)\s+(any|all|format)\(',
146 146 "any/all/format not available in Python 2.4"),
147 147 (r'(?<!def)\s+(callable)\(',
148 148 "callable not available in Python 3, use hasattr(f, '__call__')"),
149 149 (r'if\s.*\selse', "if ... else form not available in Python 2.4"),
150 150 (r'^\s*(%s)\s\s' % '|'.join(keyword.kwlist),
151 151 "gratuitous whitespace after Python keyword"),
152 152 (r'([\(\[]\s\S)|(\S\s[\)\]])', "gratuitous whitespace in () or []"),
153 153 # (r'\s\s=', "gratuitous whitespace before ="),
154 154 (r'[^>< ](\+=|-=|!=|<>|<=|>=|<<=|>>=)\S',
155 155 "missing whitespace around operator"),
156 156 (r'[^>< ](\+=|-=|!=|<>|<=|>=|<<=|>>=)\s',
157 157 "missing whitespace around operator"),
158 158 (r'\s(\+=|-=|!=|<>|<=|>=|<<=|>>=)\S',
159 159 "missing whitespace around operator"),
160 160 (r'[^+=*/!<>&| -](\s=|=\s)[^= ]',
161 161 "wrong whitespace around ="),
162 162 (r'raise Exception', "don't raise generic exceptions"),
163 163 (r' is\s+(not\s+)?["\'0-9-]', "object comparison with literal"),
164 164 (r' [=!]=\s+(True|False|None)',
165 165 "comparison with singleton, use 'is' or 'is not' instead"),
166 166 (r'^\s*(while|if) [01]:',
167 167 "use True/False for constant Boolean expression"),
168 168 (r'opener\([^)]*\).read\(',
169 169 "use opener.read() instead"),
170 170 (r'opener\([^)]*\).write\(',
171 171 "use opener.write() instead"),
172 172 (r'[\s\(](open|file)\([^)]*\)\.read\(',
173 173 "use util.readfile() instead"),
174 174 (r'[\s\(](open|file)\([^)]*\)\.write\(',
175 175 "use util.readfile() instead"),
176 176 (r'^[\s\(]*(open(er)?|file)\([^)]*\)',
177 177 "always assign an opened file to a variable, and close it afterwards"),
178 178 (r'[\s\(](open|file)\([^)]*\)\.',
179 179 "always assign an opened file to a variable, and close it afterwards"),
180 (r'(?i)descendent', "the proper spelling is descendAnt"),
180 181 ],
181 182 # warnings
182 183 [
183 184 (r'.{81}', "warning: line over 80 characters"),
184 185 (r'^\s*except:$', "warning: naked except clause"),
185 186 (r'ui\.(status|progress|write|note|warn)\([\'\"]x',
186 187 "warning: unwrapped ui message"),
187 188 ]
188 189 ]
189 190
190 191 pyfilters = [
191 192 (r"""(?msx)(?P<comment>\#.*?$)|
192 193 ((?P<quote>('''|\"\"\"|(?<!')'(?!')|(?<!")"(?!")))
193 194 (?P<text>(([^\\]|\\.)*?))
194 195 (?P=quote))""", reppython),
195 196 ]
196 197
197 198 cpats = [
198 199 [
199 200 (r'//', "don't use //-style comments"),
200 201 (r'^ ', "don't use spaces to indent"),
201 202 (r'\S\t', "don't use tabs except for indent"),
202 203 (r'(\S\s+|^\s+)\n', "trailing whitespace"),
203 204 (r'.{85}', "line too long"),
204 205 (r'(while|if|do|for)\(', "use space after while/if/do/for"),
205 206 (r'return\(', "return is not a function"),
206 207 (r' ;', "no space before ;"),
207 208 (r'\w+\* \w+', "use int *foo, not int* foo"),
208 209 (r'\([^\)]+\) \w+', "use (int)foo, not (int) foo"),
209 210 (r'\S+ (\+\+|--)', "use foo++, not foo ++"),
210 211 (r'\w,\w', "missing whitespace after ,"),
211 212 (r'^[^#]\w[+/*]\w', "missing whitespace in expression"),
212 213 (r'^#\s+\w', "use #foo, not # foo"),
213 214 (r'[^\n]\Z', "no trailing newline"),
214 215 (r'^\s*#import\b', "use only #include in standard C code"),
215 216 ],
216 217 # warnings
217 218 []
218 219 ]
219 220
220 221 cfilters = [
221 222 (r'(/\*)(((\*(?!/))|[^*])*)\*/', repccomment),
222 223 (r'''(?P<quote>(?<!")")(?P<text>([^"]|\\")+)"(?!")''', repquote),
223 224 (r'''(#\s*include\s+<)([^>]+)>''', repinclude),
224 225 (r'(\()([^)]+\))', repcallspaces),
225 226 ]
226 227
227 228 inutilpats = [
228 229 [
229 230 (r'\bui\.', "don't use ui in util"),
230 231 ],
231 232 # warnings
232 233 []
233 234 ]
234 235
235 236 inrevlogpats = [
236 237 [
237 238 (r'\brepo\.', "don't use repo in revlog"),
238 239 ],
239 240 # warnings
240 241 []
241 242 ]
242 243
243 244 checks = [
244 245 ('python', r'.*\.(py|cgi)$', pyfilters, pypats),
245 246 ('test script', r'(.*/)?test-[^.~]*$', testfilters, testpats),
246 247 ('c', r'.*\.c$', cfilters, cpats),
247 248 ('unified test', r'.*\.t$', utestfilters, utestpats),
248 249 ('layering violation repo in revlog', r'mercurial/revlog\.py', pyfilters,
249 250 inrevlogpats),
250 251 ('layering violation ui in util', r'mercurial/util\.py', pyfilters,
251 252 inutilpats),
252 253 ]
253 254
254 255 class norepeatlogger(object):
255 256 def __init__(self):
256 257 self._lastseen = None
257 258
258 259 def log(self, fname, lineno, line, msg, blame):
259 260 """print error related a to given line of a given file.
260 261
261 262 The faulty line will also be printed but only once in the case
262 263 of multiple errors.
263 264
264 265 :fname: filename
265 266 :lineno: line number
266 267 :line: actual content of the line
267 268 :msg: error message
268 269 """
269 270 msgid = fname, lineno, line
270 271 if msgid != self._lastseen:
271 272 if blame:
272 273 print "%s:%d (%s):" % (fname, lineno, blame)
273 274 else:
274 275 print "%s:%d:" % (fname, lineno)
275 276 print " > %s" % line
276 277 self._lastseen = msgid
277 278 print " " + msg
278 279
279 280 _defaultlogger = norepeatlogger()
280 281
281 282 def getblame(f):
282 283 lines = []
283 284 for l in os.popen('hg annotate -un %s' % f):
284 285 start, line = l.split(':', 1)
285 286 user, rev = start.split()
286 287 lines.append((line[1:-1], user, rev))
287 288 return lines
288 289
289 290 def checkfile(f, logfunc=_defaultlogger.log, maxerr=None, warnings=False,
290 291 blame=False, debug=False):
291 292 """checks style and portability of a given file
292 293
293 294 :f: filepath
294 295 :logfunc: function used to report error
295 296 logfunc(filename, linenumber, linecontent, errormessage)
296 297 :maxerr: number of error to display before arborting.
297 298 Set to None (default) to report all errors
298 299
299 300 return True if no error is found, False otherwise.
300 301 """
301 302 blamecache = None
302 303 result = True
303 304 for name, match, filters, pats in checks:
304 305 if debug:
305 306 print name, f
306 307 fc = 0
307 308 if not re.match(match, f):
308 309 if debug:
309 310 print "Skipping %s for %s it doesn't match %s" % (
310 311 name, match, f)
311 312 continue
312 313 fp = open(f)
313 314 pre = post = fp.read()
314 315 fp.close()
315 316 if "no-" + "check-code" in pre:
316 317 if debug:
317 318 print "Skipping %s for %s it has no- and check-code" % (
318 319 name, f)
319 320 break
320 321 for p, r in filters:
321 322 post = re.sub(p, r, post)
322 323 if warnings:
323 324 pats = pats[0] + pats[1]
324 325 else:
325 326 pats = pats[0]
326 327 # print post # uncomment to show filtered version
327 328 z = enumerate(zip(pre.splitlines(), post.splitlines(True)))
328 329 if debug:
329 330 print "Checking %s for %s" % (name, f)
330 331 for n, l in z:
331 332 if "check-code" + "-ignore" in l[0]:
332 333 if debug:
333 334 print "Skipping %s for %s:%s (check-code -ignore)" % (
334 335 name, f, n)
335 336 continue
336 337 for p, msg in pats:
337 338 if re.search(p, l[1]):
338 339 bd = ""
339 340 if blame:
340 341 bd = 'working directory'
341 342 if not blamecache:
342 343 blamecache = getblame(f)
343 344 if n < len(blamecache):
344 345 bl, bu, br = blamecache[n]
345 346 if bl == l[0]:
346 347 bd = '%s@%s' % (bu, br)
347 348 logfunc(f, n + 1, l[0], msg, bd)
348 349 fc += 1
349 350 result = False
350 351 if maxerr is not None and fc >= maxerr:
351 352 print " (too many errors, giving up)"
352 353 break
353 354 return result
354 355
355 356 if __name__ == "__main__":
356 357 parser = optparse.OptionParser("%prog [options] [files]")
357 358 parser.add_option("-w", "--warnings", action="store_true",
358 359 help="include warning-level checks")
359 360 parser.add_option("-p", "--per-file", type="int",
360 361 help="max warnings per file")
361 362 parser.add_option("-b", "--blame", action="store_true",
362 363 help="use annotate to generate blame info")
363 364 parser.add_option("", "--debug", action="store_true",
364 365 help="show debug information")
365 366
366 367 parser.set_defaults(per_file=15, warnings=False, blame=False, debug=False)
367 368 (options, args) = parser.parse_args()
368 369
369 370 if len(args) == 0:
370 371 check = glob.glob("*")
371 372 else:
372 373 check = args
373 374
374 375 for f in check:
375 376 ret = 0
376 377 if not checkfile(f, maxerr=options.per_file, warnings=options.warnings,
377 378 blame=options.blame, debug=options.debug):
378 379 ret = 1
379 380 sys.exit(ret)
@@ -1,1989 +1,1989 b''
1 1 # localrepo.py - read/write repository class for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from node import bin, hex, nullid, nullrev, short
9 9 from i18n import _
10 10 import repo, changegroup, subrepo, discovery, pushkey
11 11 import changelog, dirstate, filelog, manifest, context, bookmarks
12 12 import lock, transaction, store, encoding
13 13 import scmutil, util, extensions, hook, error
14 14 import match as matchmod
15 15 import merge as mergemod
16 16 import tags as tagsmod
17 17 from lock import release
18 18 import weakref, errno, os, time, inspect
19 19 propertycache = util.propertycache
20 20
21 21 class localrepository(repo.repository):
22 22 capabilities = set(('lookup', 'changegroupsubset', 'branchmap', 'pushkey',
23 23 'known', 'getbundle'))
24 24 supportedformats = set(('revlogv1', 'generaldelta'))
25 25 supported = supportedformats | set(('store', 'fncache', 'shared',
26 26 'dotencode'))
27 27
28 28 def __init__(self, baseui, path=None, create=False):
29 29 repo.repository.__init__(self)
30 30 self.root = os.path.realpath(util.expandpath(path))
31 31 self.path = os.path.join(self.root, ".hg")
32 32 self.origroot = path
33 33 self.auditor = scmutil.pathauditor(self.root, self._checknested)
34 34 self.opener = scmutil.opener(self.path)
35 35 self.wopener = scmutil.opener(self.root)
36 36 self.baseui = baseui
37 37 self.ui = baseui.copy()
38 38
39 39 try:
40 40 self.ui.readconfig(self.join("hgrc"), self.root)
41 41 extensions.loadall(self.ui)
42 42 except IOError:
43 43 pass
44 44
45 45 if not os.path.isdir(self.path):
46 46 if create:
47 47 if not os.path.exists(path):
48 48 util.makedirs(path)
49 49 util.makedir(self.path, notindexed=True)
50 50 requirements = ["revlogv1"]
51 51 if self.ui.configbool('format', 'usestore', True):
52 52 os.mkdir(os.path.join(self.path, "store"))
53 53 requirements.append("store")
54 54 if self.ui.configbool('format', 'usefncache', True):
55 55 requirements.append("fncache")
56 56 if self.ui.configbool('format', 'dotencode', True):
57 57 requirements.append('dotencode')
58 58 # create an invalid changelog
59 59 self.opener.append(
60 60 "00changelog.i",
61 61 '\0\0\0\2' # represents revlogv2
62 62 ' dummy changelog to prevent using the old repo layout'
63 63 )
64 64 if self.ui.configbool('format', 'generaldelta', False):
65 65 requirements.append("generaldelta")
66 66 else:
67 67 raise error.RepoError(_("repository %s not found") % path)
68 68 elif create:
69 69 raise error.RepoError(_("repository %s already exists") % path)
70 70 else:
71 71 try:
72 72 requirements = scmutil.readrequires(self.opener, self.supported)
73 73 except IOError, inst:
74 74 if inst.errno != errno.ENOENT:
75 75 raise
76 76 requirements = set()
77 77
78 78 self.sharedpath = self.path
79 79 try:
80 80 s = os.path.realpath(self.opener.read("sharedpath"))
81 81 if not os.path.exists(s):
82 82 raise error.RepoError(
83 83 _('.hg/sharedpath points to nonexistent directory %s') % s)
84 84 self.sharedpath = s
85 85 except IOError, inst:
86 86 if inst.errno != errno.ENOENT:
87 87 raise
88 88
89 89 self.store = store.store(requirements, self.sharedpath, scmutil.opener)
90 90 self.spath = self.store.path
91 91 self.sopener = self.store.opener
92 92 self.sjoin = self.store.join
93 93 self.opener.createmode = self.store.createmode
94 94 self._applyrequirements(requirements)
95 95 if create:
96 96 self._writerequirements()
97 97
98 98 # These two define the set of tags for this repository. _tags
99 99 # maps tag name to node; _tagtypes maps tag name to 'global' or
100 100 # 'local'. (Global tags are defined by .hgtags across all
101 101 # heads, and local tags are defined in .hg/localtags.) They
102 102 # constitute the in-memory cache of tags.
103 103 self._tags = None
104 104 self._tagtypes = None
105 105
106 106 self._branchcache = None
107 107 self._branchcachetip = None
108 108 self.nodetagscache = None
109 109 self.filterpats = {}
110 110 self._datafilters = {}
111 111 self._transref = self._lockref = self._wlockref = None
112 112
113 113 def _applyrequirements(self, requirements):
114 114 self.requirements = requirements
115 115 openerreqs = set(('revlogv1', 'generaldelta'))
116 116 self.sopener.options = dict((r, 1) for r in requirements
117 117 if r in openerreqs)
118 118
119 119 def _writerequirements(self):
120 120 reqfile = self.opener("requires", "w")
121 121 for r in self.requirements:
122 122 reqfile.write("%s\n" % r)
123 123 reqfile.close()
124 124
125 125 def _checknested(self, path):
126 126 """Determine if path is a legal nested repository."""
127 127 if not path.startswith(self.root):
128 128 return False
129 129 subpath = path[len(self.root) + 1:]
130 130
131 131 # XXX: Checking against the current working copy is wrong in
132 132 # the sense that it can reject things like
133 133 #
134 134 # $ hg cat -r 10 sub/x.txt
135 135 #
136 136 # if sub/ is no longer a subrepository in the working copy
137 137 # parent revision.
138 138 #
139 139 # However, it can of course also allow things that would have
140 140 # been rejected before, such as the above cat command if sub/
141 141 # is a subrepository now, but was a normal directory before.
142 142 # The old path auditor would have rejected by mistake since it
143 143 # panics when it sees sub/.hg/.
144 144 #
145 145 # All in all, checking against the working copy seems sensible
146 146 # since we want to prevent access to nested repositories on
147 147 # the filesystem *now*.
148 148 ctx = self[None]
149 149 parts = util.splitpath(subpath)
150 150 while parts:
151 151 prefix = os.sep.join(parts)
152 152 if prefix in ctx.substate:
153 153 if prefix == subpath:
154 154 return True
155 155 else:
156 156 sub = ctx.sub(prefix)
157 157 return sub.checknested(subpath[len(prefix) + 1:])
158 158 else:
159 159 parts.pop()
160 160 return False
161 161
162 162 @util.propertycache
163 163 def _bookmarks(self):
164 164 return bookmarks.read(self)
165 165
166 166 @util.propertycache
167 167 def _bookmarkcurrent(self):
168 168 return bookmarks.readcurrent(self)
169 169
170 170 @propertycache
171 171 def changelog(self):
172 172 c = changelog.changelog(self.sopener)
173 173 if 'HG_PENDING' in os.environ:
174 174 p = os.environ['HG_PENDING']
175 175 if p.startswith(self.root):
176 176 c.readpending('00changelog.i.a')
177 177 return c
178 178
179 179 @propertycache
180 180 def manifest(self):
181 181 return manifest.manifest(self.sopener)
182 182
183 183 @propertycache
184 184 def dirstate(self):
185 185 warned = [0]
186 186 def validate(node):
187 187 try:
188 188 self.changelog.rev(node)
189 189 return node
190 190 except error.LookupError:
191 191 if not warned[0]:
192 192 warned[0] = True
193 193 self.ui.warn(_("warning: ignoring unknown"
194 194 " working parent %s!\n") % short(node))
195 195 return nullid
196 196
197 197 return dirstate.dirstate(self.opener, self.ui, self.root, validate)
198 198
199 199 def __getitem__(self, changeid):
200 200 if changeid is None:
201 201 return context.workingctx(self)
202 202 return context.changectx(self, changeid)
203 203
204 204 def __contains__(self, changeid):
205 205 try:
206 206 return bool(self.lookup(changeid))
207 207 except error.RepoLookupError:
208 208 return False
209 209
210 210 def __nonzero__(self):
211 211 return True
212 212
213 213 def __len__(self):
214 214 return len(self.changelog)
215 215
216 216 def __iter__(self):
217 217 for i in xrange(len(self)):
218 218 yield i
219 219
220 220 def url(self):
221 221 return 'file:' + self.root
222 222
223 223 def hook(self, name, throw=False, **args):
224 224 return hook.hook(self.ui, self, name, throw, **args)
225 225
226 226 tag_disallowed = ':\r\n'
227 227
228 228 def _tag(self, names, node, message, local, user, date, extra={}):
229 229 if isinstance(names, str):
230 230 allchars = names
231 231 names = (names,)
232 232 else:
233 233 allchars = ''.join(names)
234 234 for c in self.tag_disallowed:
235 235 if c in allchars:
236 236 raise util.Abort(_('%r cannot be used in a tag name') % c)
237 237
238 238 branches = self.branchmap()
239 239 for name in names:
240 240 self.hook('pretag', throw=True, node=hex(node), tag=name,
241 241 local=local)
242 242 if name in branches:
243 243 self.ui.warn(_("warning: tag %s conflicts with existing"
244 244 " branch name\n") % name)
245 245
246 246 def writetags(fp, names, munge, prevtags):
247 247 fp.seek(0, 2)
248 248 if prevtags and prevtags[-1] != '\n':
249 249 fp.write('\n')
250 250 for name in names:
251 251 m = munge and munge(name) or name
252 252 if self._tagtypes and name in self._tagtypes:
253 253 old = self._tags.get(name, nullid)
254 254 fp.write('%s %s\n' % (hex(old), m))
255 255 fp.write('%s %s\n' % (hex(node), m))
256 256 fp.close()
257 257
258 258 prevtags = ''
259 259 if local:
260 260 try:
261 261 fp = self.opener('localtags', 'r+')
262 262 except IOError:
263 263 fp = self.opener('localtags', 'a')
264 264 else:
265 265 prevtags = fp.read()
266 266
267 267 # local tags are stored in the current charset
268 268 writetags(fp, names, None, prevtags)
269 269 for name in names:
270 270 self.hook('tag', node=hex(node), tag=name, local=local)
271 271 return
272 272
273 273 try:
274 274 fp = self.wfile('.hgtags', 'rb+')
275 275 except IOError:
276 276 fp = self.wfile('.hgtags', 'ab')
277 277 else:
278 278 prevtags = fp.read()
279 279
280 280 # committed tags are stored in UTF-8
281 281 writetags(fp, names, encoding.fromlocal, prevtags)
282 282
283 283 fp.close()
284 284
285 285 if '.hgtags' not in self.dirstate:
286 286 self[None].add(['.hgtags'])
287 287
288 288 m = matchmod.exact(self.root, '', ['.hgtags'])
289 289 tagnode = self.commit(message, user, date, extra=extra, match=m)
290 290
291 291 for name in names:
292 292 self.hook('tag', node=hex(node), tag=name, local=local)
293 293
294 294 return tagnode
295 295
296 296 def tag(self, names, node, message, local, user, date):
297 297 '''tag a revision with one or more symbolic names.
298 298
299 299 names is a list of strings or, when adding a single tag, names may be a
300 300 string.
301 301
302 302 if local is True, the tags are stored in a per-repository file.
303 303 otherwise, they are stored in the .hgtags file, and a new
304 304 changeset is committed with the change.
305 305
306 306 keyword arguments:
307 307
308 308 local: whether to store tags in non-version-controlled file
309 309 (default False)
310 310
311 311 message: commit message to use if committing
312 312
313 313 user: name of user to use if committing
314 314
315 315 date: date tuple to use if committing'''
316 316
317 317 if not local:
318 318 for x in self.status()[:5]:
319 319 if '.hgtags' in x:
320 320 raise util.Abort(_('working copy of .hgtags is changed '
321 321 '(please commit .hgtags manually)'))
322 322
323 323 self.tags() # instantiate the cache
324 324 self._tag(names, node, message, local, user, date)
325 325
326 326 def tags(self):
327 327 '''return a mapping of tag to node'''
328 328 if self._tags is None:
329 329 (self._tags, self._tagtypes) = self._findtags()
330 330
331 331 return self._tags
332 332
333 333 def _findtags(self):
334 334 '''Do the hard work of finding tags. Return a pair of dicts
335 335 (tags, tagtypes) where tags maps tag name to node, and tagtypes
336 336 maps tag name to a string like \'global\' or \'local\'.
337 337 Subclasses or extensions are free to add their own tags, but
338 338 should be aware that the returned dicts will be retained for the
339 339 duration of the localrepo object.'''
340 340
341 341 # XXX what tagtype should subclasses/extensions use? Currently
342 342 # mq and bookmarks add tags, but do not set the tagtype at all.
343 343 # Should each extension invent its own tag type? Should there
344 344 # be one tagtype for all such "virtual" tags? Or is the status
345 345 # quo fine?
346 346
347 347 alltags = {} # map tag name to (node, hist)
348 348 tagtypes = {}
349 349
350 350 tagsmod.findglobaltags(self.ui, self, alltags, tagtypes)
351 351 tagsmod.readlocaltags(self.ui, self, alltags, tagtypes)
352 352
353 353 # Build the return dicts. Have to re-encode tag names because
354 354 # the tags module always uses UTF-8 (in order not to lose info
355 355 # writing to the cache), but the rest of Mercurial wants them in
356 356 # local encoding.
357 357 tags = {}
358 358 for (name, (node, hist)) in alltags.iteritems():
359 359 if node != nullid:
360 360 try:
361 361 # ignore tags to unknown nodes
362 362 self.changelog.lookup(node)
363 363 tags[encoding.tolocal(name)] = node
364 364 except error.LookupError:
365 365 pass
366 366 tags['tip'] = self.changelog.tip()
367 367 tagtypes = dict([(encoding.tolocal(name), value)
368 368 for (name, value) in tagtypes.iteritems()])
369 369 return (tags, tagtypes)
370 370
371 371 def tagtype(self, tagname):
372 372 '''
373 373 return the type of the given tag. result can be:
374 374
375 375 'local' : a local tag
376 376 'global' : a global tag
377 377 None : tag does not exist
378 378 '''
379 379
380 380 self.tags()
381 381
382 382 return self._tagtypes.get(tagname)
383 383
384 384 def tagslist(self):
385 385 '''return a list of tags ordered by revision'''
386 386 l = []
387 387 for t, n in self.tags().iteritems():
388 388 r = self.changelog.rev(n)
389 389 l.append((r, t, n))
390 390 return [(t, n) for r, t, n in sorted(l)]
391 391
392 392 def nodetags(self, node):
393 393 '''return the tags associated with a node'''
394 394 if not self.nodetagscache:
395 395 self.nodetagscache = {}
396 396 for t, n in self.tags().iteritems():
397 397 self.nodetagscache.setdefault(n, []).append(t)
398 398 for tags in self.nodetagscache.itervalues():
399 399 tags.sort()
400 400 return self.nodetagscache.get(node, [])
401 401
402 402 def nodebookmarks(self, node):
403 403 marks = []
404 404 for bookmark, n in self._bookmarks.iteritems():
405 405 if n == node:
406 406 marks.append(bookmark)
407 407 return sorted(marks)
408 408
409 409 def _branchtags(self, partial, lrev):
410 410 # TODO: rename this function?
411 411 tiprev = len(self) - 1
412 412 if lrev != tiprev:
413 413 ctxgen = (self[r] for r in xrange(lrev + 1, tiprev + 1))
414 414 self._updatebranchcache(partial, ctxgen)
415 415 self._writebranchcache(partial, self.changelog.tip(), tiprev)
416 416
417 417 return partial
418 418
419 419 def updatebranchcache(self):
420 420 tip = self.changelog.tip()
421 421 if self._branchcache is not None and self._branchcachetip == tip:
422 422 return self._branchcache
423 423
424 424 oldtip = self._branchcachetip
425 425 self._branchcachetip = tip
426 426 if oldtip is None or oldtip not in self.changelog.nodemap:
427 427 partial, last, lrev = self._readbranchcache()
428 428 else:
429 429 lrev = self.changelog.rev(oldtip)
430 430 partial = self._branchcache
431 431
432 432 self._branchtags(partial, lrev)
433 433 # this private cache holds all heads (not just tips)
434 434 self._branchcache = partial
435 435
436 436 def branchmap(self):
437 437 '''returns a dictionary {branch: [branchheads]}'''
438 438 self.updatebranchcache()
439 439 return self._branchcache
440 440
441 441 def branchtags(self):
442 442 '''return a dict where branch names map to the tipmost head of
443 443 the branch, open heads come before closed'''
444 444 bt = {}
445 445 for bn, heads in self.branchmap().iteritems():
446 446 tip = heads[-1]
447 447 for h in reversed(heads):
448 448 if 'close' not in self.changelog.read(h)[5]:
449 449 tip = h
450 450 break
451 451 bt[bn] = tip
452 452 return bt
453 453
454 454 def _readbranchcache(self):
455 455 partial = {}
456 456 try:
457 457 f = self.opener("cache/branchheads")
458 458 lines = f.read().split('\n')
459 459 f.close()
460 460 except (IOError, OSError):
461 461 return {}, nullid, nullrev
462 462
463 463 try:
464 464 last, lrev = lines.pop(0).split(" ", 1)
465 465 last, lrev = bin(last), int(lrev)
466 466 if lrev >= len(self) or self[lrev].node() != last:
467 467 # invalidate the cache
468 468 raise ValueError('invalidating branch cache (tip differs)')
469 469 for l in lines:
470 470 if not l:
471 471 continue
472 472 node, label = l.split(" ", 1)
473 473 label = encoding.tolocal(label.strip())
474 474 partial.setdefault(label, []).append(bin(node))
475 475 except KeyboardInterrupt:
476 476 raise
477 477 except Exception, inst:
478 478 if self.ui.debugflag:
479 479 self.ui.warn(str(inst), '\n')
480 480 partial, last, lrev = {}, nullid, nullrev
481 481 return partial, last, lrev
482 482
483 483 def _writebranchcache(self, branches, tip, tiprev):
484 484 try:
485 485 f = self.opener("cache/branchheads", "w", atomictemp=True)
486 486 f.write("%s %s\n" % (hex(tip), tiprev))
487 487 for label, nodes in branches.iteritems():
488 488 for node in nodes:
489 489 f.write("%s %s\n" % (hex(node), encoding.fromlocal(label)))
490 490 f.rename()
491 491 except (IOError, OSError):
492 492 pass
493 493
494 494 def _updatebranchcache(self, partial, ctxgen):
495 495 # collect new branch entries
496 496 newbranches = {}
497 497 for c in ctxgen:
498 498 newbranches.setdefault(c.branch(), []).append(c.node())
499 499 # if older branchheads are reachable from new ones, they aren't
500 500 # really branchheads. Note checking parents is insufficient:
501 501 # 1 (branch a) -> 2 (branch b) -> 3 (branch a)
502 502 for branch, newnodes in newbranches.iteritems():
503 503 bheads = partial.setdefault(branch, [])
504 504 bheads.extend(newnodes)
505 505 if len(bheads) <= 1:
506 506 continue
507 507 bheads = sorted(bheads, key=lambda x: self[x].rev())
508 508 # starting from tip means fewer passes over reachable
509 509 while newnodes:
510 510 latest = newnodes.pop()
511 511 if latest not in bheads:
512 512 continue
513 513 minbhrev = self[bheads[0]].node()
514 514 reachable = self.changelog.reachable(latest, minbhrev)
515 515 reachable.remove(latest)
516 516 if reachable:
517 517 bheads = [b for b in bheads if b not in reachable]
518 518 partial[branch] = bheads
519 519
520 520 def lookup(self, key):
521 521 if isinstance(key, int):
522 522 return self.changelog.node(key)
523 523 elif key == '.':
524 524 return self.dirstate.p1()
525 525 elif key == 'null':
526 526 return nullid
527 527 elif key == 'tip':
528 528 return self.changelog.tip()
529 529 n = self.changelog._match(key)
530 530 if n:
531 531 return n
532 532 if key in self._bookmarks:
533 533 return self._bookmarks[key]
534 534 if key in self.tags():
535 535 return self.tags()[key]
536 536 if key in self.branchtags():
537 537 return self.branchtags()[key]
538 538 n = self.changelog._partialmatch(key)
539 539 if n:
540 540 return n
541 541
542 542 # can't find key, check if it might have come from damaged dirstate
543 543 if key in self.dirstate.parents():
544 544 raise error.Abort(_("working directory has unknown parent '%s'!")
545 545 % short(key))
546 546 try:
547 547 if len(key) == 20:
548 548 key = hex(key)
549 549 except TypeError:
550 550 pass
551 551 raise error.RepoLookupError(_("unknown revision '%s'") % key)
552 552
553 553 def lookupbranch(self, key, remote=None):
554 554 repo = remote or self
555 555 if key in repo.branchmap():
556 556 return key
557 557
558 558 repo = (remote and remote.local()) and remote or self
559 559 return repo[key].branch()
560 560
561 561 def known(self, nodes):
562 562 nm = self.changelog.nodemap
563 563 return [(n in nm) for n in nodes]
564 564
565 565 def local(self):
566 566 return True
567 567
568 568 def join(self, f):
569 569 return os.path.join(self.path, f)
570 570
571 571 def wjoin(self, f):
572 572 return os.path.join(self.root, f)
573 573
574 574 def file(self, f):
575 575 if f[0] == '/':
576 576 f = f[1:]
577 577 return filelog.filelog(self.sopener, f)
578 578
579 579 def changectx(self, changeid):
580 580 return self[changeid]
581 581
582 582 def parents(self, changeid=None):
583 583 '''get list of changectxs for parents of changeid'''
584 584 return self[changeid].parents()
585 585
586 586 def filectx(self, path, changeid=None, fileid=None):
587 587 """changeid can be a changeset revision, node, or tag.
588 588 fileid can be a file revision or node."""
589 589 return context.filectx(self, path, changeid, fileid)
590 590
591 591 def getcwd(self):
592 592 return self.dirstate.getcwd()
593 593
594 594 def pathto(self, f, cwd=None):
595 595 return self.dirstate.pathto(f, cwd)
596 596
597 597 def wfile(self, f, mode='r'):
598 598 return self.wopener(f, mode)
599 599
600 600 def _link(self, f):
601 601 return os.path.islink(self.wjoin(f))
602 602
603 603 def _loadfilter(self, filter):
604 604 if filter not in self.filterpats:
605 605 l = []
606 606 for pat, cmd in self.ui.configitems(filter):
607 607 if cmd == '!':
608 608 continue
609 609 mf = matchmod.match(self.root, '', [pat])
610 610 fn = None
611 611 params = cmd
612 612 for name, filterfn in self._datafilters.iteritems():
613 613 if cmd.startswith(name):
614 614 fn = filterfn
615 615 params = cmd[len(name):].lstrip()
616 616 break
617 617 if not fn:
618 618 fn = lambda s, c, **kwargs: util.filter(s, c)
619 619 # Wrap old filters not supporting keyword arguments
620 620 if not inspect.getargspec(fn)[2]:
621 621 oldfn = fn
622 622 fn = lambda s, c, **kwargs: oldfn(s, c)
623 623 l.append((mf, fn, params))
624 624 self.filterpats[filter] = l
625 625 return self.filterpats[filter]
626 626
627 627 def _filter(self, filterpats, filename, data):
628 628 for mf, fn, cmd in filterpats:
629 629 if mf(filename):
630 630 self.ui.debug("filtering %s through %s\n" % (filename, cmd))
631 631 data = fn(data, cmd, ui=self.ui, repo=self, filename=filename)
632 632 break
633 633
634 634 return data
635 635
636 636 @propertycache
637 637 def _encodefilterpats(self):
638 638 return self._loadfilter('encode')
639 639
640 640 @propertycache
641 641 def _decodefilterpats(self):
642 642 return self._loadfilter('decode')
643 643
644 644 def adddatafilter(self, name, filter):
645 645 self._datafilters[name] = filter
646 646
647 647 def wread(self, filename):
648 648 if self._link(filename):
649 649 data = os.readlink(self.wjoin(filename))
650 650 else:
651 651 data = self.wopener.read(filename)
652 652 return self._filter(self._encodefilterpats, filename, data)
653 653
654 654 def wwrite(self, filename, data, flags):
655 655 data = self._filter(self._decodefilterpats, filename, data)
656 656 if 'l' in flags:
657 657 self.wopener.symlink(data, filename)
658 658 else:
659 659 self.wopener.write(filename, data)
660 660 if 'x' in flags:
661 661 util.setflags(self.wjoin(filename), False, True)
662 662
663 663 def wwritedata(self, filename, data):
664 664 return self._filter(self._decodefilterpats, filename, data)
665 665
666 666 def transaction(self, desc):
667 667 tr = self._transref and self._transref() or None
668 668 if tr and tr.running():
669 669 return tr.nest()
670 670
671 671 # abort here if the journal already exists
672 672 if os.path.exists(self.sjoin("journal")):
673 673 raise error.RepoError(
674 674 _("abandoned transaction found - run hg recover"))
675 675
676 676 journalfiles = self._writejournal(desc)
677 677 renames = [(x, undoname(x)) for x in journalfiles]
678 678
679 679 tr = transaction.transaction(self.ui.warn, self.sopener,
680 680 self.sjoin("journal"),
681 681 aftertrans(renames),
682 682 self.store.createmode)
683 683 self._transref = weakref.ref(tr)
684 684 return tr
685 685
686 686 def _writejournal(self, desc):
687 687 # save dirstate for rollback
688 688 try:
689 689 ds = self.opener.read("dirstate")
690 690 except IOError:
691 691 ds = ""
692 692 self.opener.write("journal.dirstate", ds)
693 693 self.opener.write("journal.branch",
694 694 encoding.fromlocal(self.dirstate.branch()))
695 695 self.opener.write("journal.desc",
696 696 "%d\n%s\n" % (len(self), desc))
697 697
698 698 bkname = self.join('bookmarks')
699 699 if os.path.exists(bkname):
700 700 util.copyfile(bkname, self.join('journal.bookmarks'))
701 701 else:
702 702 self.opener.write('journal.bookmarks', '')
703 703
704 704 return (self.sjoin('journal'), self.join('journal.dirstate'),
705 705 self.join('journal.branch'), self.join('journal.desc'),
706 706 self.join('journal.bookmarks'))
707 707
708 708 def recover(self):
709 709 lock = self.lock()
710 710 try:
711 711 if os.path.exists(self.sjoin("journal")):
712 712 self.ui.status(_("rolling back interrupted transaction\n"))
713 713 transaction.rollback(self.sopener, self.sjoin("journal"),
714 714 self.ui.warn)
715 715 self.invalidate()
716 716 return True
717 717 else:
718 718 self.ui.warn(_("no interrupted transaction available\n"))
719 719 return False
720 720 finally:
721 721 lock.release()
722 722
723 723 def rollback(self, dryrun=False):
724 724 wlock = lock = None
725 725 try:
726 726 wlock = self.wlock()
727 727 lock = self.lock()
728 728 if os.path.exists(self.sjoin("undo")):
729 729 try:
730 730 args = self.opener.read("undo.desc").splitlines()
731 731 if len(args) >= 3 and self.ui.verbose:
732 732 desc = _("repository tip rolled back to revision %s"
733 733 " (undo %s: %s)\n") % (
734 734 int(args[0]) - 1, args[1], args[2])
735 735 elif len(args) >= 2:
736 736 desc = _("repository tip rolled back to revision %s"
737 737 " (undo %s)\n") % (
738 738 int(args[0]) - 1, args[1])
739 739 except IOError:
740 740 desc = _("rolling back unknown transaction\n")
741 741 self.ui.status(desc)
742 742 if dryrun:
743 743 return
744 744 transaction.rollback(self.sopener, self.sjoin("undo"),
745 745 self.ui.warn)
746 746 util.rename(self.join("undo.dirstate"), self.join("dirstate"))
747 747 if os.path.exists(self.join('undo.bookmarks')):
748 748 util.rename(self.join('undo.bookmarks'),
749 749 self.join('bookmarks'))
750 750 try:
751 751 branch = self.opener.read("undo.branch")
752 752 self.dirstate.setbranch(branch)
753 753 except IOError:
754 754 self.ui.warn(_("named branch could not be reset, "
755 755 "current branch is still: %s\n")
756 756 % self.dirstate.branch())
757 757 self.invalidate()
758 758 self.dirstate.invalidate()
759 759 self.destroyed()
760 760 parents = tuple([p.rev() for p in self.parents()])
761 761 if len(parents) > 1:
762 762 self.ui.status(_("working directory now based on "
763 763 "revisions %d and %d\n") % parents)
764 764 else:
765 765 self.ui.status(_("working directory now based on "
766 766 "revision %d\n") % parents)
767 767 else:
768 768 self.ui.warn(_("no rollback information available\n"))
769 769 return 1
770 770 finally:
771 771 release(lock, wlock)
772 772
773 773 def invalidatecaches(self):
774 774 self._tags = None
775 775 self._tagtypes = None
776 776 self.nodetagscache = None
777 777 self._branchcache = None # in UTF-8
778 778 self._branchcachetip = None
779 779
780 780 def invalidate(self):
781 781 for a in ("changelog", "manifest", "_bookmarks", "_bookmarkcurrent"):
782 782 if a in self.__dict__:
783 783 delattr(self, a)
784 784 self.invalidatecaches()
785 785
786 786 def _lock(self, lockname, wait, releasefn, acquirefn, desc):
787 787 try:
788 788 l = lock.lock(lockname, 0, releasefn, desc=desc)
789 789 except error.LockHeld, inst:
790 790 if not wait:
791 791 raise
792 792 self.ui.warn(_("waiting for lock on %s held by %r\n") %
793 793 (desc, inst.locker))
794 794 # default to 600 seconds timeout
795 795 l = lock.lock(lockname, int(self.ui.config("ui", "timeout", "600")),
796 796 releasefn, desc=desc)
797 797 if acquirefn:
798 798 acquirefn()
799 799 return l
800 800
801 801 def lock(self, wait=True):
802 802 '''Lock the repository store (.hg/store) and return a weak reference
803 803 to the lock. Use this before modifying the store (e.g. committing or
804 804 stripping). If you are opening a transaction, get a lock as well.)'''
805 805 l = self._lockref and self._lockref()
806 806 if l is not None and l.held:
807 807 l.lock()
808 808 return l
809 809
810 810 l = self._lock(self.sjoin("lock"), wait, self.store.write,
811 811 self.invalidate, _('repository %s') % self.origroot)
812 812 self._lockref = weakref.ref(l)
813 813 return l
814 814
815 815 def wlock(self, wait=True):
816 816 '''Lock the non-store parts of the repository (everything under
817 817 .hg except .hg/store) and return a weak reference to the lock.
818 818 Use this before modifying files in .hg.'''
819 819 l = self._wlockref and self._wlockref()
820 820 if l is not None and l.held:
821 821 l.lock()
822 822 return l
823 823
824 824 l = self._lock(self.join("wlock"), wait, self.dirstate.write,
825 825 self.dirstate.invalidate, _('working directory of %s') %
826 826 self.origroot)
827 827 self._wlockref = weakref.ref(l)
828 828 return l
829 829
830 830 def _filecommit(self, fctx, manifest1, manifest2, linkrev, tr, changelist):
831 831 """
832 832 commit an individual file as part of a larger transaction
833 833 """
834 834
835 835 fname = fctx.path()
836 836 text = fctx.data()
837 837 flog = self.file(fname)
838 838 fparent1 = manifest1.get(fname, nullid)
839 839 fparent2 = fparent2o = manifest2.get(fname, nullid)
840 840
841 841 meta = {}
842 842 copy = fctx.renamed()
843 843 if copy and copy[0] != fname:
844 844 # Mark the new revision of this file as a copy of another
845 845 # file. This copy data will effectively act as a parent
846 846 # of this new revision. If this is a merge, the first
847 847 # parent will be the nullid (meaning "look up the copy data")
848 848 # and the second one will be the other parent. For example:
849 849 #
850 850 # 0 --- 1 --- 3 rev1 changes file foo
851 851 # \ / rev2 renames foo to bar and changes it
852 852 # \- 2 -/ rev3 should have bar with all changes and
853 853 # should record that bar descends from
854 854 # bar in rev2 and foo in rev1
855 855 #
856 856 # this allows this merge to succeed:
857 857 #
858 858 # 0 --- 1 --- 3 rev4 reverts the content change from rev2
859 859 # \ / merging rev3 and rev4 should use bar@rev2
860 860 # \- 2 --- 4 as the merge base
861 861 #
862 862
863 863 cfname = copy[0]
864 864 crev = manifest1.get(cfname)
865 865 newfparent = fparent2
866 866
867 867 if manifest2: # branch merge
868 868 if fparent2 == nullid or crev is None: # copied on remote side
869 869 if cfname in manifest2:
870 870 crev = manifest2[cfname]
871 871 newfparent = fparent1
872 872
873 873 # find source in nearest ancestor if we've lost track
874 874 if not crev:
875 875 self.ui.debug(" %s: searching for copy revision for %s\n" %
876 876 (fname, cfname))
877 877 for ancestor in self[None].ancestors():
878 878 if cfname in ancestor:
879 879 crev = ancestor[cfname].filenode()
880 880 break
881 881
882 882 if crev:
883 883 self.ui.debug(" %s: copy %s:%s\n" % (fname, cfname, hex(crev)))
884 884 meta["copy"] = cfname
885 885 meta["copyrev"] = hex(crev)
886 886 fparent1, fparent2 = nullid, newfparent
887 887 else:
888 888 self.ui.warn(_("warning: can't find ancestor for '%s' "
889 889 "copied from '%s'!\n") % (fname, cfname))
890 890
891 891 elif fparent2 != nullid:
892 892 # is one parent an ancestor of the other?
893 893 fparentancestor = flog.ancestor(fparent1, fparent2)
894 894 if fparentancestor == fparent1:
895 895 fparent1, fparent2 = fparent2, nullid
896 896 elif fparentancestor == fparent2:
897 897 fparent2 = nullid
898 898
899 899 # is the file changed?
900 900 if fparent2 != nullid or flog.cmp(fparent1, text) or meta:
901 901 changelist.append(fname)
902 902 return flog.add(text, meta, tr, linkrev, fparent1, fparent2)
903 903
904 904 # are just the flags changed during merge?
905 905 if fparent1 != fparent2o and manifest1.flags(fname) != fctx.flags():
906 906 changelist.append(fname)
907 907
908 908 return fparent1
909 909
910 910 def commit(self, text="", user=None, date=None, match=None, force=False,
911 911 editor=False, extra={}):
912 912 """Add a new revision to current repository.
913 913
914 914 Revision information is gathered from the working directory,
915 915 match can be used to filter the committed files. If editor is
916 916 supplied, it is called to get a commit message.
917 917 """
918 918
919 919 def fail(f, msg):
920 920 raise util.Abort('%s: %s' % (f, msg))
921 921
922 922 if not match:
923 923 match = matchmod.always(self.root, '')
924 924
925 925 if not force:
926 926 vdirs = []
927 927 match.dir = vdirs.append
928 928 match.bad = fail
929 929
930 930 wlock = self.wlock()
931 931 try:
932 932 wctx = self[None]
933 933 merge = len(wctx.parents()) > 1
934 934
935 935 if (not force and merge and match and
936 936 (match.files() or match.anypats())):
937 937 raise util.Abort(_('cannot partially commit a merge '
938 938 '(do not specify files or patterns)'))
939 939
940 940 changes = self.status(match=match, clean=force)
941 941 if force:
942 942 changes[0].extend(changes[6]) # mq may commit unchanged files
943 943
944 944 # check subrepos
945 945 subs = []
946 946 removedsubs = set()
947 947 if '.hgsub' in wctx:
948 948 # only manage subrepos and .hgsubstate if .hgsub is present
949 949 for p in wctx.parents():
950 950 removedsubs.update(s for s in p.substate if match(s))
951 951 for s in wctx.substate:
952 952 removedsubs.discard(s)
953 953 if match(s) and wctx.sub(s).dirty():
954 954 subs.append(s)
955 955 if (subs or removedsubs):
956 956 if (not match('.hgsub') and
957 957 '.hgsub' in (wctx.modified() + wctx.added())):
958 958 raise util.Abort(
959 959 _("can't commit subrepos without .hgsub"))
960 960 if '.hgsubstate' not in changes[0]:
961 961 changes[0].insert(0, '.hgsubstate')
962 962 if '.hgsubstate' in changes[2]:
963 963 changes[2].remove('.hgsubstate')
964 964 elif '.hgsub' in changes[2]:
965 965 # clean up .hgsubstate when .hgsub is removed
966 966 if ('.hgsubstate' in wctx and
967 967 '.hgsubstate' not in changes[0] + changes[1] + changes[2]):
968 968 changes[2].insert(0, '.hgsubstate')
969 969
970 970 if subs and not self.ui.configbool('ui', 'commitsubrepos', True):
971 971 changedsubs = [s for s in subs if wctx.sub(s).dirty(True)]
972 972 if changedsubs:
973 973 raise util.Abort(_("uncommitted changes in subrepo %s")
974 974 % changedsubs[0])
975 975
976 976 # make sure all explicit patterns are matched
977 977 if not force and match.files():
978 978 matched = set(changes[0] + changes[1] + changes[2])
979 979
980 980 for f in match.files():
981 981 if f == '.' or f in matched or f in wctx.substate:
982 982 continue
983 983 if f in changes[3]: # missing
984 984 fail(f, _('file not found!'))
985 985 if f in vdirs: # visited directory
986 986 d = f + '/'
987 987 for mf in matched:
988 988 if mf.startswith(d):
989 989 break
990 990 else:
991 991 fail(f, _("no match under directory!"))
992 992 elif f not in self.dirstate:
993 993 fail(f, _("file not tracked!"))
994 994
995 995 if (not force and not extra.get("close") and not merge
996 996 and not (changes[0] or changes[1] or changes[2])
997 997 and wctx.branch() == wctx.p1().branch()):
998 998 return None
999 999
1000 1000 ms = mergemod.mergestate(self)
1001 1001 for f in changes[0]:
1002 1002 if f in ms and ms[f] == 'u':
1003 1003 raise util.Abort(_("unresolved merge conflicts "
1004 1004 "(see hg help resolve)"))
1005 1005
1006 1006 cctx = context.workingctx(self, text, user, date, extra, changes)
1007 1007 if editor:
1008 1008 cctx._text = editor(self, cctx, subs)
1009 1009 edited = (text != cctx._text)
1010 1010
1011 1011 # commit subs
1012 1012 if subs or removedsubs:
1013 1013 state = wctx.substate.copy()
1014 1014 for s in sorted(subs):
1015 1015 sub = wctx.sub(s)
1016 1016 self.ui.status(_('committing subrepository %s\n') %
1017 1017 subrepo.subrelpath(sub))
1018 1018 sr = sub.commit(cctx._text, user, date)
1019 1019 state[s] = (state[s][0], sr)
1020 1020 subrepo.writestate(self, state)
1021 1021
1022 1022 # Save commit message in case this transaction gets rolled back
1023 1023 # (e.g. by a pretxncommit hook). Leave the content alone on
1024 1024 # the assumption that the user will use the same editor again.
1025 1025 msgfn = self.savecommitmessage(cctx._text)
1026 1026
1027 1027 p1, p2 = self.dirstate.parents()
1028 1028 hookp1, hookp2 = hex(p1), (p2 != nullid and hex(p2) or '')
1029 1029 try:
1030 1030 self.hook("precommit", throw=True, parent1=hookp1, parent2=hookp2)
1031 1031 ret = self.commitctx(cctx, True)
1032 1032 except:
1033 1033 if edited:
1034 1034 self.ui.write(
1035 1035 _('note: commit message saved in %s\n') % msgfn)
1036 1036 raise
1037 1037
1038 1038 # update bookmarks, dirstate and mergestate
1039 1039 bookmarks.update(self, p1, ret)
1040 1040 for f in changes[0] + changes[1]:
1041 1041 self.dirstate.normal(f)
1042 1042 for f in changes[2]:
1043 1043 self.dirstate.drop(f)
1044 1044 self.dirstate.setparents(ret)
1045 1045 ms.reset()
1046 1046 finally:
1047 1047 wlock.release()
1048 1048
1049 1049 self.hook("commit", node=hex(ret), parent1=hookp1, parent2=hookp2)
1050 1050 return ret
1051 1051
1052 1052 def commitctx(self, ctx, error=False):
1053 1053 """Add a new revision to current repository.
1054 1054 Revision information is passed via the context argument.
1055 1055 """
1056 1056
1057 1057 tr = lock = None
1058 1058 removed = list(ctx.removed())
1059 1059 p1, p2 = ctx.p1(), ctx.p2()
1060 1060 user = ctx.user()
1061 1061
1062 1062 lock = self.lock()
1063 1063 try:
1064 1064 tr = self.transaction("commit")
1065 1065 trp = weakref.proxy(tr)
1066 1066
1067 1067 if ctx.files():
1068 1068 m1 = p1.manifest().copy()
1069 1069 m2 = p2.manifest()
1070 1070
1071 1071 # check in files
1072 1072 new = {}
1073 1073 changed = []
1074 1074 linkrev = len(self)
1075 1075 for f in sorted(ctx.modified() + ctx.added()):
1076 1076 self.ui.note(f + "\n")
1077 1077 try:
1078 1078 fctx = ctx[f]
1079 1079 new[f] = self._filecommit(fctx, m1, m2, linkrev, trp,
1080 1080 changed)
1081 1081 m1.set(f, fctx.flags())
1082 1082 except OSError, inst:
1083 1083 self.ui.warn(_("trouble committing %s!\n") % f)
1084 1084 raise
1085 1085 except IOError, inst:
1086 1086 errcode = getattr(inst, 'errno', errno.ENOENT)
1087 1087 if error or errcode and errcode != errno.ENOENT:
1088 1088 self.ui.warn(_("trouble committing %s!\n") % f)
1089 1089 raise
1090 1090 else:
1091 1091 removed.append(f)
1092 1092
1093 1093 # update manifest
1094 1094 m1.update(new)
1095 1095 removed = [f for f in sorted(removed) if f in m1 or f in m2]
1096 1096 drop = [f for f in removed if f in m1]
1097 1097 for f in drop:
1098 1098 del m1[f]
1099 1099 mn = self.manifest.add(m1, trp, linkrev, p1.manifestnode(),
1100 1100 p2.manifestnode(), (new, drop))
1101 1101 files = changed + removed
1102 1102 else:
1103 1103 mn = p1.manifestnode()
1104 1104 files = []
1105 1105
1106 1106 # update changelog
1107 1107 self.changelog.delayupdate()
1108 1108 n = self.changelog.add(mn, files, ctx.description(),
1109 1109 trp, p1.node(), p2.node(),
1110 1110 user, ctx.date(), ctx.extra().copy())
1111 1111 p = lambda: self.changelog.writepending() and self.root or ""
1112 1112 xp1, xp2 = p1.hex(), p2 and p2.hex() or ''
1113 1113 self.hook('pretxncommit', throw=True, node=hex(n), parent1=xp1,
1114 1114 parent2=xp2, pending=p)
1115 1115 self.changelog.finalize(trp)
1116 1116 tr.close()
1117 1117
1118 1118 if self._branchcache:
1119 1119 self.updatebranchcache()
1120 1120 return n
1121 1121 finally:
1122 1122 if tr:
1123 1123 tr.release()
1124 1124 lock.release()
1125 1125
1126 1126 def destroyed(self):
1127 1127 '''Inform the repository that nodes have been destroyed.
1128 1128 Intended for use by strip and rollback, so there's a common
1129 1129 place for anything that has to be done after destroying history.'''
1130 1130 # XXX it might be nice if we could take the list of destroyed
1131 1131 # nodes, but I don't see an easy way for rollback() to do that
1132 1132
1133 1133 # Ensure the persistent tag cache is updated. Doing it now
1134 1134 # means that the tag cache only has to worry about destroyed
1135 1135 # heads immediately after a strip/rollback. That in turn
1136 1136 # guarantees that "cachetip == currenttip" (comparing both rev
1137 1137 # and node) always means no nodes have been added or destroyed.
1138 1138
1139 1139 # XXX this is suboptimal when qrefresh'ing: we strip the current
1140 1140 # head, refresh the tag cache, then immediately add a new head.
1141 1141 # But I think doing it this way is necessary for the "instant
1142 1142 # tag cache retrieval" case to work.
1143 1143 self.invalidatecaches()
1144 1144
1145 1145 def walk(self, match, node=None):
1146 1146 '''
1147 1147 walk recursively through the directory tree or a given
1148 1148 changeset, finding all files matched by the match
1149 1149 function
1150 1150 '''
1151 1151 return self[node].walk(match)
1152 1152
1153 1153 def status(self, node1='.', node2=None, match=None,
1154 1154 ignored=False, clean=False, unknown=False,
1155 1155 listsubrepos=False):
1156 1156 """return status of files between two nodes or node and working directory
1157 1157
1158 1158 If node1 is None, use the first dirstate parent instead.
1159 1159 If node2 is None, compare node1 with working directory.
1160 1160 """
1161 1161
1162 1162 def mfmatches(ctx):
1163 1163 mf = ctx.manifest().copy()
1164 1164 for fn in mf.keys():
1165 1165 if not match(fn):
1166 1166 del mf[fn]
1167 1167 return mf
1168 1168
1169 1169 if isinstance(node1, context.changectx):
1170 1170 ctx1 = node1
1171 1171 else:
1172 1172 ctx1 = self[node1]
1173 1173 if isinstance(node2, context.changectx):
1174 1174 ctx2 = node2
1175 1175 else:
1176 1176 ctx2 = self[node2]
1177 1177
1178 1178 working = ctx2.rev() is None
1179 1179 parentworking = working and ctx1 == self['.']
1180 1180 match = match or matchmod.always(self.root, self.getcwd())
1181 1181 listignored, listclean, listunknown = ignored, clean, unknown
1182 1182
1183 1183 # load earliest manifest first for caching reasons
1184 1184 if not working and ctx2.rev() < ctx1.rev():
1185 1185 ctx2.manifest()
1186 1186
1187 1187 if not parentworking:
1188 1188 def bad(f, msg):
1189 1189 if f not in ctx1:
1190 1190 self.ui.warn('%s: %s\n' % (self.dirstate.pathto(f), msg))
1191 1191 match.bad = bad
1192 1192
1193 1193 if working: # we need to scan the working dir
1194 1194 subrepos = []
1195 1195 if '.hgsub' in self.dirstate:
1196 1196 subrepos = ctx1.substate.keys()
1197 1197 s = self.dirstate.status(match, subrepos, listignored,
1198 1198 listclean, listunknown)
1199 1199 cmp, modified, added, removed, deleted, unknown, ignored, clean = s
1200 1200
1201 1201 # check for any possibly clean files
1202 1202 if parentworking and cmp:
1203 1203 fixup = []
1204 1204 # do a full compare of any files that might have changed
1205 1205 for f in sorted(cmp):
1206 1206 if (f not in ctx1 or ctx2.flags(f) != ctx1.flags(f)
1207 1207 or ctx1[f].cmp(ctx2[f])):
1208 1208 modified.append(f)
1209 1209 else:
1210 1210 fixup.append(f)
1211 1211
1212 1212 # update dirstate for files that are actually clean
1213 1213 if fixup:
1214 1214 if listclean:
1215 1215 clean += fixup
1216 1216
1217 1217 try:
1218 1218 # updating the dirstate is optional
1219 1219 # so we don't wait on the lock
1220 1220 wlock = self.wlock(False)
1221 1221 try:
1222 1222 for f in fixup:
1223 1223 self.dirstate.normal(f)
1224 1224 finally:
1225 1225 wlock.release()
1226 1226 except error.LockError:
1227 1227 pass
1228 1228
1229 1229 if not parentworking:
1230 1230 mf1 = mfmatches(ctx1)
1231 1231 if working:
1232 1232 # we are comparing working dir against non-parent
1233 1233 # generate a pseudo-manifest for the working dir
1234 1234 mf2 = mfmatches(self['.'])
1235 1235 for f in cmp + modified + added:
1236 1236 mf2[f] = None
1237 1237 mf2.set(f, ctx2.flags(f))
1238 1238 for f in removed:
1239 1239 if f in mf2:
1240 1240 del mf2[f]
1241 1241 else:
1242 1242 # we are comparing two revisions
1243 1243 deleted, unknown, ignored = [], [], []
1244 1244 mf2 = mfmatches(ctx2)
1245 1245
1246 1246 modified, added, clean = [], [], []
1247 1247 for fn in mf2:
1248 1248 if fn in mf1:
1249 1249 if (fn not in deleted and
1250 1250 (mf1.flags(fn) != mf2.flags(fn) or
1251 1251 (mf1[fn] != mf2[fn] and
1252 1252 (mf2[fn] or ctx1[fn].cmp(ctx2[fn]))))):
1253 1253 modified.append(fn)
1254 1254 elif listclean:
1255 1255 clean.append(fn)
1256 1256 del mf1[fn]
1257 1257 elif fn not in deleted:
1258 1258 added.append(fn)
1259 1259 removed = mf1.keys()
1260 1260
1261 1261 r = modified, added, removed, deleted, unknown, ignored, clean
1262 1262
1263 1263 if listsubrepos:
1264 1264 for subpath, sub in subrepo.itersubrepos(ctx1, ctx2):
1265 1265 if working:
1266 1266 rev2 = None
1267 1267 else:
1268 1268 rev2 = ctx2.substate[subpath][1]
1269 1269 try:
1270 1270 submatch = matchmod.narrowmatcher(subpath, match)
1271 1271 s = sub.status(rev2, match=submatch, ignored=listignored,
1272 1272 clean=listclean, unknown=listunknown,
1273 1273 listsubrepos=True)
1274 1274 for rfiles, sfiles in zip(r, s):
1275 1275 rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
1276 1276 except error.LookupError:
1277 1277 self.ui.status(_("skipping missing subrepository: %s\n")
1278 1278 % subpath)
1279 1279
1280 1280 for l in r:
1281 1281 l.sort()
1282 1282 return r
1283 1283
1284 1284 def heads(self, start=None):
1285 1285 heads = self.changelog.heads(start)
1286 1286 # sort the output in rev descending order
1287 1287 return sorted(heads, key=self.changelog.rev, reverse=True)
1288 1288
1289 1289 def branchheads(self, branch=None, start=None, closed=False):
1290 1290 '''return a (possibly filtered) list of heads for the given branch
1291 1291
1292 1292 Heads are returned in topological order, from newest to oldest.
1293 1293 If branch is None, use the dirstate branch.
1294 1294 If start is not None, return only heads reachable from start.
1295 1295 If closed is True, return heads that are marked as closed as well.
1296 1296 '''
1297 1297 if branch is None:
1298 1298 branch = self[None].branch()
1299 1299 branches = self.branchmap()
1300 1300 if branch not in branches:
1301 1301 return []
1302 1302 # the cache returns heads ordered lowest to highest
1303 1303 bheads = list(reversed(branches[branch]))
1304 1304 if start is not None:
1305 1305 # filter out the heads that cannot be reached from startrev
1306 1306 fbheads = set(self.changelog.nodesbetween([start], bheads)[2])
1307 1307 bheads = [h for h in bheads if h in fbheads]
1308 1308 if not closed:
1309 1309 bheads = [h for h in bheads if
1310 1310 ('close' not in self.changelog.read(h)[5])]
1311 1311 return bheads
1312 1312
1313 1313 def branches(self, nodes):
1314 1314 if not nodes:
1315 1315 nodes = [self.changelog.tip()]
1316 1316 b = []
1317 1317 for n in nodes:
1318 1318 t = n
1319 1319 while True:
1320 1320 p = self.changelog.parents(n)
1321 1321 if p[1] != nullid or p[0] == nullid:
1322 1322 b.append((t, n, p[0], p[1]))
1323 1323 break
1324 1324 n = p[0]
1325 1325 return b
1326 1326
1327 1327 def between(self, pairs):
1328 1328 r = []
1329 1329
1330 1330 for top, bottom in pairs:
1331 1331 n, l, i = top, [], 0
1332 1332 f = 1
1333 1333
1334 1334 while n != bottom and n != nullid:
1335 1335 p = self.changelog.parents(n)[0]
1336 1336 if i == f:
1337 1337 l.append(n)
1338 1338 f = f * 2
1339 1339 n = p
1340 1340 i += 1
1341 1341
1342 1342 r.append(l)
1343 1343
1344 1344 return r
1345 1345
1346 1346 def pull(self, remote, heads=None, force=False):
1347 1347 lock = self.lock()
1348 1348 try:
1349 1349 tmp = discovery.findcommonincoming(self, remote, heads=heads,
1350 1350 force=force)
1351 1351 common, fetch, rheads = tmp
1352 1352 if not fetch:
1353 1353 self.ui.status(_("no changes found\n"))
1354 1354 result = 0
1355 1355 else:
1356 1356 if heads is None and list(common) == [nullid]:
1357 1357 self.ui.status(_("requesting all changes\n"))
1358 1358 elif heads is None and remote.capable('changegroupsubset'):
1359 1359 # issue1320, avoid a race if remote changed after discovery
1360 1360 heads = rheads
1361 1361
1362 1362 if remote.capable('getbundle'):
1363 1363 cg = remote.getbundle('pull', common=common,
1364 1364 heads=heads or rheads)
1365 1365 elif heads is None:
1366 1366 cg = remote.changegroup(fetch, 'pull')
1367 1367 elif not remote.capable('changegroupsubset'):
1368 1368 raise util.Abort(_("partial pull cannot be done because "
1369 1369 "other repository doesn't support "
1370 1370 "changegroupsubset."))
1371 1371 else:
1372 1372 cg = remote.changegroupsubset(fetch, heads, 'pull')
1373 1373 result = self.addchangegroup(cg, 'pull', remote.url(),
1374 1374 lock=lock)
1375 1375 finally:
1376 1376 lock.release()
1377 1377
1378 1378 return result
1379 1379
1380 1380 def checkpush(self, force, revs):
1381 1381 """Extensions can override this function if additional checks have
1382 1382 to be performed before pushing, or call it if they override push
1383 1383 command.
1384 1384 """
1385 1385 pass
1386 1386
1387 1387 def push(self, remote, force=False, revs=None, newbranch=False):
1388 1388 '''Push outgoing changesets (limited by revs) from the current
1389 1389 repository to remote. Return an integer:
1390 1390 - 0 means HTTP error *or* nothing to push
1391 1391 - 1 means we pushed and remote head count is unchanged *or*
1392 1392 we have outgoing changesets but refused to push
1393 1393 - other values as described by addchangegroup()
1394 1394 '''
1395 1395 # there are two ways to push to remote repo:
1396 1396 #
1397 1397 # addchangegroup assumes local user can lock remote
1398 1398 # repo (local filesystem, old ssh servers).
1399 1399 #
1400 1400 # unbundle assumes local user cannot lock remote repo (new ssh
1401 1401 # servers, http servers).
1402 1402
1403 1403 self.checkpush(force, revs)
1404 1404 lock = None
1405 1405 unbundle = remote.capable('unbundle')
1406 1406 if not unbundle:
1407 1407 lock = remote.lock()
1408 1408 try:
1409 1409 cg, remote_heads = discovery.prepush(self, remote, force, revs,
1410 1410 newbranch)
1411 1411 ret = remote_heads
1412 1412 if cg is not None:
1413 1413 if unbundle:
1414 1414 # local repo finds heads on server, finds out what
1415 1415 # revs it must push. once revs transferred, if server
1416 1416 # finds it has different heads (someone else won
1417 1417 # commit/push race), server aborts.
1418 1418 if force:
1419 1419 remote_heads = ['force']
1420 1420 # ssh: return remote's addchangegroup()
1421 1421 # http: return remote's addchangegroup() or 0 for error
1422 1422 ret = remote.unbundle(cg, remote_heads, 'push')
1423 1423 else:
1424 1424 # we return an integer indicating remote head count change
1425 1425 ret = remote.addchangegroup(cg, 'push', self.url(),
1426 1426 lock=lock)
1427 1427 finally:
1428 1428 if lock is not None:
1429 1429 lock.release()
1430 1430
1431 1431 self.ui.debug("checking for updated bookmarks\n")
1432 1432 rb = remote.listkeys('bookmarks')
1433 1433 for k in rb.keys():
1434 1434 if k in self._bookmarks:
1435 1435 nr, nl = rb[k], hex(self._bookmarks[k])
1436 1436 if nr in self:
1437 1437 cr = self[nr]
1438 1438 cl = self[nl]
1439 1439 if cl in cr.descendants():
1440 1440 r = remote.pushkey('bookmarks', k, nr, nl)
1441 1441 if r:
1442 1442 self.ui.status(_("updating bookmark %s\n") % k)
1443 1443 else:
1444 1444 self.ui.warn(_('updating bookmark %s'
1445 1445 ' failed!\n') % k)
1446 1446
1447 1447 return ret
1448 1448
1449 1449 def changegroupinfo(self, nodes, source):
1450 1450 if self.ui.verbose or source == 'bundle':
1451 1451 self.ui.status(_("%d changesets found\n") % len(nodes))
1452 1452 if self.ui.debugflag:
1453 1453 self.ui.debug("list of changesets:\n")
1454 1454 for node in nodes:
1455 1455 self.ui.debug("%s\n" % hex(node))
1456 1456
1457 1457 def changegroupsubset(self, bases, heads, source):
1458 1458 """Compute a changegroup consisting of all the nodes that are
1459 descendents of any of the bases and ancestors of any of the heads.
1459 descendants of any of the bases and ancestors of any of the heads.
1460 1460 Return a chunkbuffer object whose read() method will return
1461 1461 successive changegroup chunks.
1462 1462
1463 1463 It is fairly complex as determining which filenodes and which
1464 1464 manifest nodes need to be included for the changeset to be complete
1465 1465 is non-trivial.
1466 1466
1467 1467 Another wrinkle is doing the reverse, figuring out which changeset in
1468 1468 the changegroup a particular filenode or manifestnode belongs to.
1469 1469 """
1470 1470 cl = self.changelog
1471 1471 if not bases:
1472 1472 bases = [nullid]
1473 1473 csets, bases, heads = cl.nodesbetween(bases, heads)
1474 1474 # We assume that all ancestors of bases are known
1475 1475 common = set(cl.ancestors(*[cl.rev(n) for n in bases]))
1476 1476 return self._changegroupsubset(common, csets, heads, source)
1477 1477
1478 1478 def getbundle(self, source, heads=None, common=None):
1479 1479 """Like changegroupsubset, but returns the set difference between the
1480 1480 ancestors of heads and the ancestors common.
1481 1481
1482 1482 If heads is None, use the local heads. If common is None, use [nullid].
1483 1483
1484 1484 The nodes in common might not all be known locally due to the way the
1485 1485 current discovery protocol works.
1486 1486 """
1487 1487 cl = self.changelog
1488 1488 if common:
1489 1489 nm = cl.nodemap
1490 1490 common = [n for n in common if n in nm]
1491 1491 else:
1492 1492 common = [nullid]
1493 1493 if not heads:
1494 1494 heads = cl.heads()
1495 1495 common, missing = cl.findcommonmissing(common, heads)
1496 1496 if not missing:
1497 1497 return None
1498 1498 return self._changegroupsubset(common, missing, heads, source)
1499 1499
1500 1500 def _changegroupsubset(self, commonrevs, csets, heads, source):
1501 1501
1502 1502 cl = self.changelog
1503 1503 mf = self.manifest
1504 1504 mfs = {} # needed manifests
1505 1505 fnodes = {} # needed file nodes
1506 1506 changedfiles = set()
1507 1507 fstate = ['', {}]
1508 1508 count = [0]
1509 1509
1510 1510 # can we go through the fast path ?
1511 1511 heads.sort()
1512 1512 if heads == sorted(self.heads()):
1513 1513 return self._changegroup(csets, source)
1514 1514
1515 1515 # slow path
1516 1516 self.hook('preoutgoing', throw=True, source=source)
1517 1517 self.changegroupinfo(csets, source)
1518 1518
1519 1519 # filter any nodes that claim to be part of the known set
1520 1520 def prune(revlog, missing):
1521 1521 return [n for n in missing
1522 1522 if revlog.linkrev(revlog.rev(n)) not in commonrevs]
1523 1523
1524 1524 def lookup(revlog, x):
1525 1525 if revlog == cl:
1526 1526 c = cl.read(x)
1527 1527 changedfiles.update(c[3])
1528 1528 mfs.setdefault(c[0], x)
1529 1529 count[0] += 1
1530 1530 self.ui.progress(_('bundling'), count[0],
1531 1531 unit=_('changesets'), total=len(csets))
1532 1532 return x
1533 1533 elif revlog == mf:
1534 1534 clnode = mfs[x]
1535 1535 mdata = mf.readfast(x)
1536 1536 for f in changedfiles:
1537 1537 if f in mdata:
1538 1538 fnodes.setdefault(f, {}).setdefault(mdata[f], clnode)
1539 1539 count[0] += 1
1540 1540 self.ui.progress(_('bundling'), count[0],
1541 1541 unit=_('manifests'), total=len(mfs))
1542 1542 return mfs[x]
1543 1543 else:
1544 1544 self.ui.progress(
1545 1545 _('bundling'), count[0], item=fstate[0],
1546 1546 unit=_('files'), total=len(changedfiles))
1547 1547 return fstate[1][x]
1548 1548
1549 1549 bundler = changegroup.bundle10(lookup)
1550 1550 reorder = self.ui.config('bundle', 'reorder', 'auto')
1551 1551 if reorder == 'auto':
1552 1552 reorder = None
1553 1553 else:
1554 1554 reorder = util.parsebool(reorder)
1555 1555
1556 1556 def gengroup():
1557 1557 # Create a changenode group generator that will call our functions
1558 1558 # back to lookup the owning changenode and collect information.
1559 1559 for chunk in cl.group(csets, bundler, reorder=reorder):
1560 1560 yield chunk
1561 1561 self.ui.progress(_('bundling'), None)
1562 1562
1563 1563 # Create a generator for the manifestnodes that calls our lookup
1564 1564 # and data collection functions back.
1565 1565 count[0] = 0
1566 1566 for chunk in mf.group(prune(mf, mfs), bundler, reorder=reorder):
1567 1567 yield chunk
1568 1568 self.ui.progress(_('bundling'), None)
1569 1569
1570 1570 mfs.clear()
1571 1571
1572 1572 # Go through all our files in order sorted by name.
1573 1573 count[0] = 0
1574 1574 for fname in sorted(changedfiles):
1575 1575 filerevlog = self.file(fname)
1576 1576 if not len(filerevlog):
1577 1577 raise util.Abort(_("empty or missing revlog for %s") % fname)
1578 1578 fstate[0] = fname
1579 1579 fstate[1] = fnodes.pop(fname, {})
1580 1580
1581 1581 nodelist = prune(filerevlog, fstate[1])
1582 1582 if nodelist:
1583 1583 count[0] += 1
1584 1584 yield bundler.fileheader(fname)
1585 1585 for chunk in filerevlog.group(nodelist, bundler, reorder):
1586 1586 yield chunk
1587 1587
1588 1588 # Signal that no more groups are left.
1589 1589 yield bundler.close()
1590 1590 self.ui.progress(_('bundling'), None)
1591 1591
1592 1592 if csets:
1593 1593 self.hook('outgoing', node=hex(csets[0]), source=source)
1594 1594
1595 1595 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
1596 1596
1597 1597 def changegroup(self, basenodes, source):
1598 1598 # to avoid a race we use changegroupsubset() (issue1320)
1599 1599 return self.changegroupsubset(basenodes, self.heads(), source)
1600 1600
1601 1601 def _changegroup(self, nodes, source):
1602 1602 """Compute the changegroup of all nodes that we have that a recipient
1603 1603 doesn't. Return a chunkbuffer object whose read() method will return
1604 1604 successive changegroup chunks.
1605 1605
1606 1606 This is much easier than the previous function as we can assume that
1607 1607 the recipient has any changenode we aren't sending them.
1608 1608
1609 1609 nodes is the set of nodes to send"""
1610 1610
1611 1611 cl = self.changelog
1612 1612 mf = self.manifest
1613 1613 mfs = {}
1614 1614 changedfiles = set()
1615 1615 fstate = ['']
1616 1616 count = [0]
1617 1617
1618 1618 self.hook('preoutgoing', throw=True, source=source)
1619 1619 self.changegroupinfo(nodes, source)
1620 1620
1621 1621 revset = set([cl.rev(n) for n in nodes])
1622 1622
1623 1623 def gennodelst(log):
1624 1624 return [log.node(r) for r in log if log.linkrev(r) in revset]
1625 1625
1626 1626 def lookup(revlog, x):
1627 1627 if revlog == cl:
1628 1628 c = cl.read(x)
1629 1629 changedfiles.update(c[3])
1630 1630 mfs.setdefault(c[0], x)
1631 1631 count[0] += 1
1632 1632 self.ui.progress(_('bundling'), count[0],
1633 1633 unit=_('changesets'), total=len(nodes))
1634 1634 return x
1635 1635 elif revlog == mf:
1636 1636 count[0] += 1
1637 1637 self.ui.progress(_('bundling'), count[0],
1638 1638 unit=_('manifests'), total=len(mfs))
1639 1639 return cl.node(revlog.linkrev(revlog.rev(x)))
1640 1640 else:
1641 1641 self.ui.progress(
1642 1642 _('bundling'), count[0], item=fstate[0],
1643 1643 total=len(changedfiles), unit=_('files'))
1644 1644 return cl.node(revlog.linkrev(revlog.rev(x)))
1645 1645
1646 1646 bundler = changegroup.bundle10(lookup)
1647 1647 reorder = self.ui.config('bundle', 'reorder', 'auto')
1648 1648 if reorder == 'auto':
1649 1649 reorder = None
1650 1650 else:
1651 1651 reorder = util.parsebool(reorder)
1652 1652
1653 1653 def gengroup():
1654 1654 '''yield a sequence of changegroup chunks (strings)'''
1655 1655 # construct a list of all changed files
1656 1656
1657 1657 for chunk in cl.group(nodes, bundler, reorder=reorder):
1658 1658 yield chunk
1659 1659 self.ui.progress(_('bundling'), None)
1660 1660
1661 1661 count[0] = 0
1662 1662 for chunk in mf.group(gennodelst(mf), bundler, reorder=reorder):
1663 1663 yield chunk
1664 1664 self.ui.progress(_('bundling'), None)
1665 1665
1666 1666 count[0] = 0
1667 1667 for fname in sorted(changedfiles):
1668 1668 filerevlog = self.file(fname)
1669 1669 if not len(filerevlog):
1670 1670 raise util.Abort(_("empty or missing revlog for %s") % fname)
1671 1671 fstate[0] = fname
1672 1672 nodelist = gennodelst(filerevlog)
1673 1673 if nodelist:
1674 1674 count[0] += 1
1675 1675 yield bundler.fileheader(fname)
1676 1676 for chunk in filerevlog.group(nodelist, bundler, reorder):
1677 1677 yield chunk
1678 1678 yield bundler.close()
1679 1679 self.ui.progress(_('bundling'), None)
1680 1680
1681 1681 if nodes:
1682 1682 self.hook('outgoing', node=hex(nodes[0]), source=source)
1683 1683
1684 1684 return changegroup.unbundle10(util.chunkbuffer(gengroup()), 'UN')
1685 1685
1686 1686 def addchangegroup(self, source, srctype, url, emptyok=False, lock=None):
1687 1687 """Add the changegroup returned by source.read() to this repo.
1688 1688 srctype is a string like 'push', 'pull', or 'unbundle'. url is
1689 1689 the URL of the repo where this changegroup is coming from.
1690 1690 If lock is not None, the function takes ownership of the lock
1691 1691 and releases it after the changegroup is added.
1692 1692
1693 1693 Return an integer summarizing the change to this repo:
1694 1694 - nothing changed or no source: 0
1695 1695 - more heads than before: 1+added heads (2..n)
1696 1696 - fewer heads than before: -1-removed heads (-2..-n)
1697 1697 - number of heads stays the same: 1
1698 1698 """
1699 1699 def csmap(x):
1700 1700 self.ui.debug("add changeset %s\n" % short(x))
1701 1701 return len(cl)
1702 1702
1703 1703 def revmap(x):
1704 1704 return cl.rev(x)
1705 1705
1706 1706 if not source:
1707 1707 return 0
1708 1708
1709 1709 self.hook('prechangegroup', throw=True, source=srctype, url=url)
1710 1710
1711 1711 changesets = files = revisions = 0
1712 1712 efiles = set()
1713 1713
1714 1714 # write changelog data to temp files so concurrent readers will not see
1715 1715 # inconsistent view
1716 1716 cl = self.changelog
1717 1717 cl.delayupdate()
1718 1718 oldheads = cl.heads()
1719 1719
1720 1720 tr = self.transaction("\n".join([srctype, util.hidepassword(url)]))
1721 1721 try:
1722 1722 trp = weakref.proxy(tr)
1723 1723 # pull off the changeset group
1724 1724 self.ui.status(_("adding changesets\n"))
1725 1725 clstart = len(cl)
1726 1726 class prog(object):
1727 1727 step = _('changesets')
1728 1728 count = 1
1729 1729 ui = self.ui
1730 1730 total = None
1731 1731 def __call__(self):
1732 1732 self.ui.progress(self.step, self.count, unit=_('chunks'),
1733 1733 total=self.total)
1734 1734 self.count += 1
1735 1735 pr = prog()
1736 1736 source.callback = pr
1737 1737
1738 1738 source.changelogheader()
1739 1739 if (cl.addgroup(source, csmap, trp) is None
1740 1740 and not emptyok):
1741 1741 raise util.Abort(_("received changelog group is empty"))
1742 1742 clend = len(cl)
1743 1743 changesets = clend - clstart
1744 1744 for c in xrange(clstart, clend):
1745 1745 efiles.update(self[c].files())
1746 1746 efiles = len(efiles)
1747 1747 self.ui.progress(_('changesets'), None)
1748 1748
1749 1749 # pull off the manifest group
1750 1750 self.ui.status(_("adding manifests\n"))
1751 1751 pr.step = _('manifests')
1752 1752 pr.count = 1
1753 1753 pr.total = changesets # manifests <= changesets
1754 1754 # no need to check for empty manifest group here:
1755 1755 # if the result of the merge of 1 and 2 is the same in 3 and 4,
1756 1756 # no new manifest will be created and the manifest group will
1757 1757 # be empty during the pull
1758 1758 source.manifestheader()
1759 1759 self.manifest.addgroup(source, revmap, trp)
1760 1760 self.ui.progress(_('manifests'), None)
1761 1761
1762 1762 needfiles = {}
1763 1763 if self.ui.configbool('server', 'validate', default=False):
1764 1764 # validate incoming csets have their manifests
1765 1765 for cset in xrange(clstart, clend):
1766 1766 mfest = self.changelog.read(self.changelog.node(cset))[0]
1767 1767 mfest = self.manifest.readdelta(mfest)
1768 1768 # store file nodes we must see
1769 1769 for f, n in mfest.iteritems():
1770 1770 needfiles.setdefault(f, set()).add(n)
1771 1771
1772 1772 # process the files
1773 1773 self.ui.status(_("adding file changes\n"))
1774 1774 pr.step = 'files'
1775 1775 pr.count = 1
1776 1776 pr.total = efiles
1777 1777 source.callback = None
1778 1778
1779 1779 while True:
1780 1780 chunkdata = source.filelogheader()
1781 1781 if not chunkdata:
1782 1782 break
1783 1783 f = chunkdata["filename"]
1784 1784 self.ui.debug("adding %s revisions\n" % f)
1785 1785 pr()
1786 1786 fl = self.file(f)
1787 1787 o = len(fl)
1788 1788 if fl.addgroup(source, revmap, trp) is None:
1789 1789 raise util.Abort(_("received file revlog group is empty"))
1790 1790 revisions += len(fl) - o
1791 1791 files += 1
1792 1792 if f in needfiles:
1793 1793 needs = needfiles[f]
1794 1794 for new in xrange(o, len(fl)):
1795 1795 n = fl.node(new)
1796 1796 if n in needs:
1797 1797 needs.remove(n)
1798 1798 if not needs:
1799 1799 del needfiles[f]
1800 1800 self.ui.progress(_('files'), None)
1801 1801
1802 1802 for f, needs in needfiles.iteritems():
1803 1803 fl = self.file(f)
1804 1804 for n in needs:
1805 1805 try:
1806 1806 fl.rev(n)
1807 1807 except error.LookupError:
1808 1808 raise util.Abort(
1809 1809 _('missing file data for %s:%s - run hg verify') %
1810 1810 (f, hex(n)))
1811 1811
1812 1812 dh = 0
1813 1813 if oldheads:
1814 1814 heads = cl.heads()
1815 1815 dh = len(heads) - len(oldheads)
1816 1816 for h in heads:
1817 1817 if h not in oldheads and 'close' in self[h].extra():
1818 1818 dh -= 1
1819 1819 htext = ""
1820 1820 if dh:
1821 1821 htext = _(" (%+d heads)") % dh
1822 1822
1823 1823 self.ui.status(_("added %d changesets"
1824 1824 " with %d changes to %d files%s\n")
1825 1825 % (changesets, revisions, files, htext))
1826 1826
1827 1827 if changesets > 0:
1828 1828 p = lambda: cl.writepending() and self.root or ""
1829 1829 self.hook('pretxnchangegroup', throw=True,
1830 1830 node=hex(cl.node(clstart)), source=srctype,
1831 1831 url=url, pending=p)
1832 1832
1833 1833 # make changelog see real files again
1834 1834 cl.finalize(trp)
1835 1835
1836 1836 tr.close()
1837 1837 finally:
1838 1838 tr.release()
1839 1839 if lock:
1840 1840 lock.release()
1841 1841
1842 1842 if changesets > 0:
1843 1843 # forcefully update the on-disk branch cache
1844 1844 self.ui.debug("updating the branch cache\n")
1845 1845 self.updatebranchcache()
1846 1846 self.hook("changegroup", node=hex(cl.node(clstart)),
1847 1847 source=srctype, url=url)
1848 1848
1849 1849 for i in xrange(clstart, clend):
1850 1850 self.hook("incoming", node=hex(cl.node(i)),
1851 1851 source=srctype, url=url)
1852 1852
1853 1853 # never return 0 here:
1854 1854 if dh < 0:
1855 1855 return dh - 1
1856 1856 else:
1857 1857 return dh + 1
1858 1858
1859 1859 def stream_in(self, remote, requirements):
1860 1860 lock = self.lock()
1861 1861 try:
1862 1862 fp = remote.stream_out()
1863 1863 l = fp.readline()
1864 1864 try:
1865 1865 resp = int(l)
1866 1866 except ValueError:
1867 1867 raise error.ResponseError(
1868 1868 _('Unexpected response from remote server:'), l)
1869 1869 if resp == 1:
1870 1870 raise util.Abort(_('operation forbidden by server'))
1871 1871 elif resp == 2:
1872 1872 raise util.Abort(_('locking the remote repository failed'))
1873 1873 elif resp != 0:
1874 1874 raise util.Abort(_('the server sent an unknown error code'))
1875 1875 self.ui.status(_('streaming all changes\n'))
1876 1876 l = fp.readline()
1877 1877 try:
1878 1878 total_files, total_bytes = map(int, l.split(' ', 1))
1879 1879 except (ValueError, TypeError):
1880 1880 raise error.ResponseError(
1881 1881 _('Unexpected response from remote server:'), l)
1882 1882 self.ui.status(_('%d files to transfer, %s of data\n') %
1883 1883 (total_files, util.bytecount(total_bytes)))
1884 1884 start = time.time()
1885 1885 for i in xrange(total_files):
1886 1886 # XXX doesn't support '\n' or '\r' in filenames
1887 1887 l = fp.readline()
1888 1888 try:
1889 1889 name, size = l.split('\0', 1)
1890 1890 size = int(size)
1891 1891 except (ValueError, TypeError):
1892 1892 raise error.ResponseError(
1893 1893 _('Unexpected response from remote server:'), l)
1894 1894 self.ui.debug('adding %s (%s)\n' % (name, util.bytecount(size)))
1895 1895 # for backwards compat, name was partially encoded
1896 1896 ofp = self.sopener(store.decodedir(name), 'w')
1897 1897 for chunk in util.filechunkiter(fp, limit=size):
1898 1898 ofp.write(chunk)
1899 1899 ofp.close()
1900 1900 elapsed = time.time() - start
1901 1901 if elapsed <= 0:
1902 1902 elapsed = 0.001
1903 1903 self.ui.status(_('transferred %s in %.1f seconds (%s/sec)\n') %
1904 1904 (util.bytecount(total_bytes), elapsed,
1905 1905 util.bytecount(total_bytes / elapsed)))
1906 1906
1907 1907 # new requirements = old non-format requirements + new format-related
1908 1908 # requirements from the streamed-in repository
1909 1909 requirements.update(set(self.requirements) - self.supportedformats)
1910 1910 self._applyrequirements(requirements)
1911 1911 self._writerequirements()
1912 1912
1913 1913 self.invalidate()
1914 1914 return len(self.heads()) + 1
1915 1915 finally:
1916 1916 lock.release()
1917 1917
1918 1918 def clone(self, remote, heads=[], stream=False):
1919 1919 '''clone remote repository.
1920 1920
1921 1921 keyword arguments:
1922 1922 heads: list of revs to clone (forces use of pull)
1923 1923 stream: use streaming clone if possible'''
1924 1924
1925 1925 # now, all clients that can request uncompressed clones can
1926 1926 # read repo formats supported by all servers that can serve
1927 1927 # them.
1928 1928
1929 1929 # if revlog format changes, client will have to check version
1930 1930 # and format flags on "stream" capability, and use
1931 1931 # uncompressed only if compatible.
1932 1932
1933 1933 if stream and not heads:
1934 1934 # 'stream' means remote revlog format is revlogv1 only
1935 1935 if remote.capable('stream'):
1936 1936 return self.stream_in(remote, set(('revlogv1',)))
1937 1937 # otherwise, 'streamreqs' contains the remote revlog format
1938 1938 streamreqs = remote.capable('streamreqs')
1939 1939 if streamreqs:
1940 1940 streamreqs = set(streamreqs.split(','))
1941 1941 # if we support it, stream in and adjust our requirements
1942 1942 if not streamreqs - self.supportedformats:
1943 1943 return self.stream_in(remote, streamreqs)
1944 1944 return self.pull(remote, heads)
1945 1945
1946 1946 def pushkey(self, namespace, key, old, new):
1947 1947 self.hook('prepushkey', throw=True, namespace=namespace, key=key,
1948 1948 old=old, new=new)
1949 1949 ret = pushkey.push(self, namespace, key, old, new)
1950 1950 self.hook('pushkey', namespace=namespace, key=key, old=old, new=new,
1951 1951 ret=ret)
1952 1952 return ret
1953 1953
1954 1954 def listkeys(self, namespace):
1955 1955 self.hook('prelistkeys', throw=True, namespace=namespace)
1956 1956 values = pushkey.list(self, namespace)
1957 1957 self.hook('listkeys', namespace=namespace, values=values)
1958 1958 return values
1959 1959
1960 1960 def debugwireargs(self, one, two, three=None, four=None, five=None):
1961 1961 '''used to test argument passing over the wire'''
1962 1962 return "%s %s %s %s %s" % (one, two, three, four, five)
1963 1963
1964 1964 def savecommitmessage(self, text):
1965 1965 fp = self.opener('last-message.txt', 'wb')
1966 1966 try:
1967 1967 fp.write(text)
1968 1968 finally:
1969 1969 fp.close()
1970 1970 return self.pathto(fp.name[len(self.root)+1:])
1971 1971
1972 1972 # used to avoid circular references so destructors work
1973 1973 def aftertrans(files):
1974 1974 renamefiles = [tuple(t) for t in files]
1975 1975 def a():
1976 1976 for src, dest in renamefiles:
1977 1977 util.rename(src, dest)
1978 1978 return a
1979 1979
1980 1980 def undoname(fn):
1981 1981 base, name = os.path.split(fn)
1982 1982 assert name.startswith('journal')
1983 1983 return os.path.join(base, name.replace('journal', 'undo', 1))
1984 1984
1985 1985 def instance(ui, path, create):
1986 1986 return localrepository(ui, util.localpath(path), create)
1987 1987
1988 1988 def islocal(path):
1989 1989 return True
@@ -1,1278 +1,1278 b''
1 1 # revlog.py - storage back-end for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 """Storage back-end for Mercurial.
9 9
10 10 This provides efficient delta storage with O(1) retrieve and append
11 11 and O(changes) merge between branches.
12 12 """
13 13
14 14 # import stuff from node for others to import from revlog
15 15 from node import bin, hex, nullid, nullrev
16 16 from i18n import _
17 17 import ancestor, mdiff, parsers, error, util, dagutil
18 18 import struct, zlib, errno
19 19
20 20 _pack = struct.pack
21 21 _unpack = struct.unpack
22 22 _compress = zlib.compress
23 23 _decompress = zlib.decompress
24 24 _sha = util.sha1
25 25
26 26 # revlog header flags
27 27 REVLOGV0 = 0
28 28 REVLOGNG = 1
29 29 REVLOGNGINLINEDATA = (1 << 16)
30 30 REVLOGGENERALDELTA = (1 << 17)
31 31 REVLOG_DEFAULT_FLAGS = REVLOGNGINLINEDATA
32 32 REVLOG_DEFAULT_FORMAT = REVLOGNG
33 33 REVLOG_DEFAULT_VERSION = REVLOG_DEFAULT_FORMAT | REVLOG_DEFAULT_FLAGS
34 34 REVLOGNG_FLAGS = REVLOGNGINLINEDATA | REVLOGGENERALDELTA
35 35
36 36 # revlog index flags
37 37 REVIDX_KNOWN_FLAGS = 0
38 38
39 39 # max size of revlog with inline data
40 40 _maxinline = 131072
41 41 _chunksize = 1048576
42 42
43 43 RevlogError = error.RevlogError
44 44 LookupError = error.LookupError
45 45
46 46 def getoffset(q):
47 47 return int(q >> 16)
48 48
49 49 def gettype(q):
50 50 return int(q & 0xFFFF)
51 51
52 52 def offset_type(offset, type):
53 53 return long(long(offset) << 16 | type)
54 54
55 55 nullhash = _sha(nullid)
56 56
57 57 def hash(text, p1, p2):
58 58 """generate a hash from the given text and its parent hashes
59 59
60 60 This hash combines both the current file contents and its history
61 61 in a manner that makes it easy to distinguish nodes with the same
62 62 content in the revision graph.
63 63 """
64 64 # As of now, if one of the parent node is null, p2 is null
65 65 if p2 == nullid:
66 66 # deep copy of a hash is faster than creating one
67 67 s = nullhash.copy()
68 68 s.update(p1)
69 69 else:
70 70 # none of the parent nodes are nullid
71 71 l = [p1, p2]
72 72 l.sort()
73 73 s = _sha(l[0])
74 74 s.update(l[1])
75 75 s.update(text)
76 76 return s.digest()
77 77
78 78 def compress(text):
79 79 """ generate a possibly-compressed representation of text """
80 80 if not text:
81 81 return ("", text)
82 82 l = len(text)
83 83 bin = None
84 84 if l < 44:
85 85 pass
86 86 elif l > 1000000:
87 87 # zlib makes an internal copy, thus doubling memory usage for
88 88 # large files, so lets do this in pieces
89 89 z = zlib.compressobj()
90 90 p = []
91 91 pos = 0
92 92 while pos < l:
93 93 pos2 = pos + 2**20
94 94 p.append(z.compress(text[pos:pos2]))
95 95 pos = pos2
96 96 p.append(z.flush())
97 97 if sum(map(len, p)) < l:
98 98 bin = "".join(p)
99 99 else:
100 100 bin = _compress(text)
101 101 if bin is None or len(bin) > l:
102 102 if text[0] == '\0':
103 103 return ("", text)
104 104 return ('u', text)
105 105 return ("", bin)
106 106
107 107 def decompress(bin):
108 108 """ decompress the given input """
109 109 if not bin:
110 110 return bin
111 111 t = bin[0]
112 112 if t == '\0':
113 113 return bin
114 114 if t == 'x':
115 115 return _decompress(bin)
116 116 if t == 'u':
117 117 return bin[1:]
118 118 raise RevlogError(_("unknown compression type %r") % t)
119 119
120 120 indexformatv0 = ">4l20s20s20s"
121 121 v0shaoffset = 56
122 122
123 123 class revlogoldio(object):
124 124 def __init__(self):
125 125 self.size = struct.calcsize(indexformatv0)
126 126
127 127 def parseindex(self, data, inline):
128 128 s = self.size
129 129 index = []
130 130 nodemap = {nullid: nullrev}
131 131 n = off = 0
132 132 l = len(data)
133 133 while off + s <= l:
134 134 cur = data[off:off + s]
135 135 off += s
136 136 e = _unpack(indexformatv0, cur)
137 137 # transform to revlogv1 format
138 138 e2 = (offset_type(e[0], 0), e[1], -1, e[2], e[3],
139 139 nodemap.get(e[4], nullrev), nodemap.get(e[5], nullrev), e[6])
140 140 index.append(e2)
141 141 nodemap[e[6]] = n
142 142 n += 1
143 143
144 144 # add the magic null revision at -1
145 145 index.append((0, 0, 0, -1, -1, -1, -1, nullid))
146 146
147 147 return index, nodemap, None
148 148
149 149 def packentry(self, entry, node, version, rev):
150 150 if gettype(entry[0]):
151 151 raise RevlogError(_("index entry flags need RevlogNG"))
152 152 e2 = (getoffset(entry[0]), entry[1], entry[3], entry[4],
153 153 node(entry[5]), node(entry[6]), entry[7])
154 154 return _pack(indexformatv0, *e2)
155 155
156 156 # index ng:
157 157 # 6 bytes: offset
158 158 # 2 bytes: flags
159 159 # 4 bytes: compressed length
160 160 # 4 bytes: uncompressed length
161 161 # 4 bytes: base rev
162 162 # 4 bytes: link rev
163 163 # 4 bytes: parent 1 rev
164 164 # 4 bytes: parent 2 rev
165 165 # 32 bytes: nodeid
166 166 indexformatng = ">Qiiiiii20s12x"
167 167 ngshaoffset = 32
168 168 versionformat = ">I"
169 169
170 170 class revlogio(object):
171 171 def __init__(self):
172 172 self.size = struct.calcsize(indexformatng)
173 173
174 174 def parseindex(self, data, inline):
175 175 # call the C implementation to parse the index data
176 176 index, cache = parsers.parse_index2(data, inline)
177 177 return index, None, cache
178 178
179 179 def packentry(self, entry, node, version, rev):
180 180 p = _pack(indexformatng, *entry)
181 181 if rev == 0:
182 182 p = _pack(versionformat, version) + p[4:]
183 183 return p
184 184
185 185 class revlog(object):
186 186 """
187 187 the underlying revision storage object
188 188
189 189 A revlog consists of two parts, an index and the revision data.
190 190
191 191 The index is a file with a fixed record size containing
192 192 information on each revision, including its nodeid (hash), the
193 193 nodeids of its parents, the position and offset of its data within
194 194 the data file, and the revision it's based on. Finally, each entry
195 195 contains a linkrev entry that can serve as a pointer to external
196 196 data.
197 197
198 198 The revision data itself is a linear collection of data chunks.
199 199 Each chunk represents a revision and is usually represented as a
200 200 delta against the previous chunk. To bound lookup time, runs of
201 201 deltas are limited to about 2 times the length of the original
202 202 version data. This makes retrieval of a version proportional to
203 203 its size, or O(1) relative to the number of revisions.
204 204
205 205 Both pieces of the revlog are written to in an append-only
206 206 fashion, which means we never need to rewrite a file to insert or
207 207 remove data, and can use some simple techniques to avoid the need
208 208 for locking while reading.
209 209 """
210 210 def __init__(self, opener, indexfile):
211 211 """
212 212 create a revlog object
213 213
214 214 opener is a function that abstracts the file opening operation
215 215 and can be used to implement COW semantics or the like.
216 216 """
217 217 self.indexfile = indexfile
218 218 self.datafile = indexfile[:-2] + ".d"
219 219 self.opener = opener
220 220 self._cache = None
221 221 self._basecache = (0, 0)
222 222 self._chunkcache = (0, '')
223 223 self.index = []
224 224 self._pcache = {}
225 225 self._nodecache = {nullid: nullrev}
226 226 self._nodepos = None
227 227
228 228 v = REVLOG_DEFAULT_VERSION
229 229 if hasattr(opener, 'options'):
230 230 if 'revlogv1' in opener.options:
231 231 if 'generaldelta' in opener.options:
232 232 v |= REVLOGGENERALDELTA
233 233 else:
234 234 v = 0
235 235
236 236 i = ''
237 237 self._initempty = True
238 238 try:
239 239 f = self.opener(self.indexfile)
240 240 i = f.read()
241 241 f.close()
242 242 if len(i) > 0:
243 243 v = struct.unpack(versionformat, i[:4])[0]
244 244 self._initempty = False
245 245 except IOError, inst:
246 246 if inst.errno != errno.ENOENT:
247 247 raise
248 248
249 249 self.version = v
250 250 self._inline = v & REVLOGNGINLINEDATA
251 251 self._generaldelta = v & REVLOGGENERALDELTA
252 252 flags = v & ~0xFFFF
253 253 fmt = v & 0xFFFF
254 254 if fmt == REVLOGV0 and flags:
255 255 raise RevlogError(_("index %s unknown flags %#04x for format v0")
256 256 % (self.indexfile, flags >> 16))
257 257 elif fmt == REVLOGNG and flags & ~REVLOGNG_FLAGS:
258 258 raise RevlogError(_("index %s unknown flags %#04x for revlogng")
259 259 % (self.indexfile, flags >> 16))
260 260 elif fmt > REVLOGNG:
261 261 raise RevlogError(_("index %s unknown format %d")
262 262 % (self.indexfile, fmt))
263 263
264 264 self._io = revlogio()
265 265 if self.version == REVLOGV0:
266 266 self._io = revlogoldio()
267 267 try:
268 268 d = self._io.parseindex(i, self._inline)
269 269 except (ValueError, IndexError):
270 270 raise RevlogError(_("index %s is corrupted") % (self.indexfile))
271 271 self.index, nodemap, self._chunkcache = d
272 272 if nodemap is not None:
273 273 self.nodemap = self._nodecache = nodemap
274 274 if not self._chunkcache:
275 275 self._chunkclear()
276 276
277 277 def tip(self):
278 278 return self.node(len(self.index) - 2)
279 279 def __len__(self):
280 280 return len(self.index) - 1
281 281 def __iter__(self):
282 282 for i in xrange(len(self)):
283 283 yield i
284 284
285 285 @util.propertycache
286 286 def nodemap(self):
287 287 self.rev(self.node(0))
288 288 return self._nodecache
289 289
290 290 def rev(self, node):
291 291 try:
292 292 return self._nodecache[node]
293 293 except KeyError:
294 294 n = self._nodecache
295 295 i = self.index
296 296 p = self._nodepos
297 297 if p is None:
298 298 p = len(i) - 2
299 299 for r in xrange(p, -1, -1):
300 300 v = i[r][7]
301 301 n[v] = r
302 302 if v == node:
303 303 self._nodepos = r - 1
304 304 return r
305 305 raise LookupError(node, self.indexfile, _('no node'))
306 306
307 307 def node(self, rev):
308 308 return self.index[rev][7]
309 309 def linkrev(self, rev):
310 310 return self.index[rev][4]
311 311 def parents(self, node):
312 312 i = self.index
313 313 d = i[self.rev(node)]
314 314 return i[d[5]][7], i[d[6]][7] # map revisions to nodes inline
315 315 def parentrevs(self, rev):
316 316 return self.index[rev][5:7]
317 317 def start(self, rev):
318 318 return int(self.index[rev][0] >> 16)
319 319 def end(self, rev):
320 320 return self.start(rev) + self.length(rev)
321 321 def length(self, rev):
322 322 return self.index[rev][1]
323 323 def chainbase(self, rev):
324 324 index = self.index
325 325 base = index[rev][3]
326 326 while base != rev:
327 327 rev = base
328 328 base = index[rev][3]
329 329 return base
330 330 def flags(self, rev):
331 331 return self.index[rev][0] & 0xFFFF
332 332 def rawsize(self, rev):
333 333 """return the length of the uncompressed text for a given revision"""
334 334 l = self.index[rev][2]
335 335 if l >= 0:
336 336 return l
337 337
338 338 t = self.revision(self.node(rev))
339 339 return len(t)
340 340 size = rawsize
341 341
342 342 def reachable(self, node, stop=None):
343 343 """return the set of all nodes ancestral to a given node, including
344 344 the node itself, stopping when stop is matched"""
345 345 reachable = set((node,))
346 346 visit = [node]
347 347 if stop:
348 348 stopn = self.rev(stop)
349 349 else:
350 350 stopn = 0
351 351 while visit:
352 352 n = visit.pop(0)
353 353 if n == stop:
354 354 continue
355 355 if n == nullid:
356 356 continue
357 357 for p in self.parents(n):
358 358 if self.rev(p) < stopn:
359 359 continue
360 360 if p not in reachable:
361 361 reachable.add(p)
362 362 visit.append(p)
363 363 return reachable
364 364
365 365 def ancestors(self, *revs):
366 366 """Generate the ancestors of 'revs' in reverse topological order.
367 367
368 368 Yield a sequence of revision numbers starting with the parents
369 369 of each revision in revs, i.e., each revision is *not* considered
370 370 an ancestor of itself. Results are in breadth-first order:
371 371 parents of each rev in revs, then parents of those, etc. Result
372 372 does not include the null revision."""
373 373 visit = list(revs)
374 374 seen = set([nullrev])
375 375 while visit:
376 376 for parent in self.parentrevs(visit.pop(0)):
377 377 if parent not in seen:
378 378 visit.append(parent)
379 379 seen.add(parent)
380 380 yield parent
381 381
382 382 def descendants(self, *revs):
383 383 """Generate the descendants of 'revs' in revision order.
384 384
385 385 Yield a sequence of revision numbers starting with a child of
386 386 some rev in revs, i.e., each revision is *not* considered a
387 387 descendant of itself. Results are ordered by revision number (a
388 388 topological sort)."""
389 389 first = min(revs)
390 390 if first == nullrev:
391 391 for i in self:
392 392 yield i
393 393 return
394 394
395 395 seen = set(revs)
396 396 for i in xrange(first + 1, len(self)):
397 397 for x in self.parentrevs(i):
398 398 if x != nullrev and x in seen:
399 399 seen.add(i)
400 400 yield i
401 401 break
402 402
403 403 def findcommonmissing(self, common=None, heads=None):
404 404 """Return a tuple of the ancestors of common and the ancestors of heads
405 405 that are not ancestors of common.
406 406
407 407 More specifically, the second element is a list of nodes N such that
408 408 every N satisfies the following constraints:
409 409
410 410 1. N is an ancestor of some node in 'heads'
411 411 2. N is not an ancestor of any node in 'common'
412 412
413 413 The list is sorted by revision number, meaning it is
414 414 topologically sorted.
415 415
416 416 'heads' and 'common' are both lists of node IDs. If heads is
417 417 not supplied, uses all of the revlog's heads. If common is not
418 418 supplied, uses nullid."""
419 419 if common is None:
420 420 common = [nullid]
421 421 if heads is None:
422 422 heads = self.heads()
423 423
424 424 common = [self.rev(n) for n in common]
425 425 heads = [self.rev(n) for n in heads]
426 426
427 427 # we want the ancestors, but inclusive
428 428 has = set(self.ancestors(*common))
429 429 has.add(nullrev)
430 430 has.update(common)
431 431
432 432 # take all ancestors from heads that aren't in has
433 433 missing = set()
434 434 visit = [r for r in heads if r not in has]
435 435 while visit:
436 436 r = visit.pop(0)
437 437 if r in missing:
438 438 continue
439 439 else:
440 440 missing.add(r)
441 441 for p in self.parentrevs(r):
442 442 if p not in has:
443 443 visit.append(p)
444 444 missing = list(missing)
445 445 missing.sort()
446 446 return has, [self.node(r) for r in missing]
447 447
448 448 def findmissing(self, common=None, heads=None):
449 449 """Return the ancestors of heads that are not ancestors of common.
450 450
451 451 More specifically, return a list of nodes N such that every N
452 452 satisfies the following constraints:
453 453
454 454 1. N is an ancestor of some node in 'heads'
455 455 2. N is not an ancestor of any node in 'common'
456 456
457 457 The list is sorted by revision number, meaning it is
458 458 topologically sorted.
459 459
460 460 'heads' and 'common' are both lists of node IDs. If heads is
461 461 not supplied, uses all of the revlog's heads. If common is not
462 462 supplied, uses nullid."""
463 463 _common, missing = self.findcommonmissing(common, heads)
464 464 return missing
465 465
466 466 def nodesbetween(self, roots=None, heads=None):
467 467 """Return a topological path from 'roots' to 'heads'.
468 468
469 469 Return a tuple (nodes, outroots, outheads) where 'nodes' is a
470 470 topologically sorted list of all nodes N that satisfy both of
471 471 these constraints:
472 472
473 473 1. N is a descendant of some node in 'roots'
474 474 2. N is an ancestor of some node in 'heads'
475 475
476 476 Every node is considered to be both a descendant and an ancestor
477 477 of itself, so every reachable node in 'roots' and 'heads' will be
478 478 included in 'nodes'.
479 479
480 480 'outroots' is the list of reachable nodes in 'roots', i.e., the
481 481 subset of 'roots' that is returned in 'nodes'. Likewise,
482 482 'outheads' is the subset of 'heads' that is also in 'nodes'.
483 483
484 484 'roots' and 'heads' are both lists of node IDs. If 'roots' is
485 485 unspecified, uses nullid as the only root. If 'heads' is
486 486 unspecified, uses list of all of the revlog's heads."""
487 487 nonodes = ([], [], [])
488 488 if roots is not None:
489 489 roots = list(roots)
490 490 if not roots:
491 491 return nonodes
492 492 lowestrev = min([self.rev(n) for n in roots])
493 493 else:
494 roots = [nullid] # Everybody's a descendent of nullid
494 roots = [nullid] # Everybody's a descendant of nullid
495 495 lowestrev = nullrev
496 496 if (lowestrev == nullrev) and (heads is None):
497 497 # We want _all_ the nodes!
498 498 return ([self.node(r) for r in self], [nullid], list(self.heads()))
499 499 if heads is None:
500 500 # All nodes are ancestors, so the latest ancestor is the last
501 501 # node.
502 502 highestrev = len(self) - 1
503 503 # Set ancestors to None to signal that every node is an ancestor.
504 504 ancestors = None
505 505 # Set heads to an empty dictionary for later discovery of heads
506 506 heads = {}
507 507 else:
508 508 heads = list(heads)
509 509 if not heads:
510 510 return nonodes
511 511 ancestors = set()
512 512 # Turn heads into a dictionary so we can remove 'fake' heads.
513 513 # Also, later we will be using it to filter out the heads we can't
514 514 # find from roots.
515 515 heads = dict.fromkeys(heads, False)
516 516 # Start at the top and keep marking parents until we're done.
517 517 nodestotag = set(heads)
518 518 # Remember where the top was so we can use it as a limit later.
519 519 highestrev = max([self.rev(n) for n in nodestotag])
520 520 while nodestotag:
521 521 # grab a node to tag
522 522 n = nodestotag.pop()
523 523 # Never tag nullid
524 524 if n == nullid:
525 525 continue
526 526 # A node's revision number represents its place in a
527 527 # topologically sorted list of nodes.
528 528 r = self.rev(n)
529 529 if r >= lowestrev:
530 530 if n not in ancestors:
531 # If we are possibly a descendent of one of the roots
531 # If we are possibly a descendant of one of the roots
532 532 # and we haven't already been marked as an ancestor
533 533 ancestors.add(n) # Mark as ancestor
534 534 # Add non-nullid parents to list of nodes to tag.
535 535 nodestotag.update([p for p in self.parents(n) if
536 536 p != nullid])
537 537 elif n in heads: # We've seen it before, is it a fake head?
538 538 # So it is, real heads should not be the ancestors of
539 539 # any other heads.
540 540 heads.pop(n)
541 541 if not ancestors:
542 542 return nonodes
543 543 # Now that we have our set of ancestors, we want to remove any
544 544 # roots that are not ancestors.
545 545
546 546 # If one of the roots was nullid, everything is included anyway.
547 547 if lowestrev > nullrev:
548 548 # But, since we weren't, let's recompute the lowest rev to not
549 549 # include roots that aren't ancestors.
550 550
551 551 # Filter out roots that aren't ancestors of heads
552 552 roots = [n for n in roots if n in ancestors]
553 553 # Recompute the lowest revision
554 554 if roots:
555 555 lowestrev = min([self.rev(n) for n in roots])
556 556 else:
557 557 # No more roots? Return empty list
558 558 return nonodes
559 559 else:
560 560 # We are descending from nullid, and don't need to care about
561 561 # any other roots.
562 562 lowestrev = nullrev
563 563 roots = [nullid]
564 564 # Transform our roots list into a set.
565 descendents = set(roots)
565 descendants = set(roots)
566 566 # Also, keep the original roots so we can filter out roots that aren't
567 567 # 'real' roots (i.e. are descended from other roots).
568 roots = descendents.copy()
568 roots = descendants.copy()
569 569 # Our topologically sorted list of output nodes.
570 570 orderedout = []
571 571 # Don't start at nullid since we don't want nullid in our output list,
572 572 # and if nullid shows up in descedents, empty parents will look like
573 # they're descendents.
573 # they're descendants.
574 574 for r in xrange(max(lowestrev, 0), highestrev + 1):
575 575 n = self.node(r)
576 isdescendent = False
577 if lowestrev == nullrev: # Everybody is a descendent of nullid
578 isdescendent = True
579 elif n in descendents:
580 # n is already a descendent
581 isdescendent = True
576 isdescendant = False
577 if lowestrev == nullrev: # Everybody is a descendant of nullid
578 isdescendant = True
579 elif n in descendants:
580 # n is already a descendant
581 isdescendant = True
582 582 # This check only needs to be done here because all the roots
583 # will start being marked is descendents before the loop.
583 # will start being marked is descendants before the loop.
584 584 if n in roots:
585 585 # If n was a root, check if it's a 'real' root.
586 586 p = tuple(self.parents(n))
587 # If any of its parents are descendents, it's not a root.
588 if (p[0] in descendents) or (p[1] in descendents):
587 # If any of its parents are descendants, it's not a root.
588 if (p[0] in descendants) or (p[1] in descendants):
589 589 roots.remove(n)
590 590 else:
591 591 p = tuple(self.parents(n))
592 # A node is a descendent if either of its parents are
593 # descendents. (We seeded the dependents list with the roots
592 # A node is a descendant if either of its parents are
593 # descendants. (We seeded the dependents list with the roots
594 594 # up there, remember?)
595 if (p[0] in descendents) or (p[1] in descendents):
596 descendents.add(n)
597 isdescendent = True
598 if isdescendent and ((ancestors is None) or (n in ancestors)):
599 # Only include nodes that are both descendents and ancestors.
595 if (p[0] in descendants) or (p[1] in descendants):
596 descendants.add(n)
597 isdescendant = True
598 if isdescendant and ((ancestors is None) or (n in ancestors)):
599 # Only include nodes that are both descendants and ancestors.
600 600 orderedout.append(n)
601 601 if (ancestors is not None) and (n in heads):
602 602 # We're trying to figure out which heads are reachable
603 603 # from roots.
604 604 # Mark this head as having been reached
605 605 heads[n] = True
606 606 elif ancestors is None:
607 607 # Otherwise, we're trying to discover the heads.
608 608 # Assume this is a head because if it isn't, the next step
609 609 # will eventually remove it.
610 610 heads[n] = True
611 611 # But, obviously its parents aren't.
612 612 for p in self.parents(n):
613 613 heads.pop(p, None)
614 614 heads = [n for n, flag in heads.iteritems() if flag]
615 615 roots = list(roots)
616 616 assert orderedout
617 617 assert roots
618 618 assert heads
619 619 return (orderedout, roots, heads)
620 620
621 621 def headrevs(self):
622 622 count = len(self)
623 623 if not count:
624 624 return [nullrev]
625 625 ishead = [1] * (count + 1)
626 626 index = self.index
627 627 for r in xrange(count):
628 628 e = index[r]
629 629 ishead[e[5]] = ishead[e[6]] = 0
630 630 return [r for r in xrange(count) if ishead[r]]
631 631
632 632 def heads(self, start=None, stop=None):
633 633 """return the list of all nodes that have no children
634 634
635 635 if start is specified, only heads that are descendants of
636 636 start will be returned
637 637 if stop is specified, it will consider all the revs from stop
638 638 as if they had no children
639 639 """
640 640 if start is None and stop is None:
641 641 if not len(self):
642 642 return [nullid]
643 643 return [self.node(r) for r in self.headrevs()]
644 644
645 645 if start is None:
646 646 start = nullid
647 647 if stop is None:
648 648 stop = []
649 649 stoprevs = set([self.rev(n) for n in stop])
650 650 startrev = self.rev(start)
651 651 reachable = set((startrev,))
652 652 heads = set((startrev,))
653 653
654 654 parentrevs = self.parentrevs
655 655 for r in xrange(startrev + 1, len(self)):
656 656 for p in parentrevs(r):
657 657 if p in reachable:
658 658 if r not in stoprevs:
659 659 reachable.add(r)
660 660 heads.add(r)
661 661 if p in heads and p not in stoprevs:
662 662 heads.remove(p)
663 663
664 664 return [self.node(r) for r in heads]
665 665
666 666 def children(self, node):
667 667 """find the children of a given node"""
668 668 c = []
669 669 p = self.rev(node)
670 670 for r in range(p + 1, len(self)):
671 671 prevs = [pr for pr in self.parentrevs(r) if pr != nullrev]
672 672 if prevs:
673 673 for pr in prevs:
674 674 if pr == p:
675 675 c.append(self.node(r))
676 676 elif p == nullrev:
677 677 c.append(self.node(r))
678 678 return c
679 679
680 680 def descendant(self, start, end):
681 681 if start == nullrev:
682 682 return True
683 683 for i in self.descendants(start):
684 684 if i == end:
685 685 return True
686 686 elif i > end:
687 687 break
688 688 return False
689 689
690 690 def ancestor(self, a, b):
691 691 """calculate the least common ancestor of nodes a and b"""
692 692
693 693 # fast path, check if it is a descendant
694 694 a, b = self.rev(a), self.rev(b)
695 695 start, end = sorted((a, b))
696 696 if self.descendant(start, end):
697 697 return self.node(start)
698 698
699 699 def parents(rev):
700 700 return [p for p in self.parentrevs(rev) if p != nullrev]
701 701
702 702 c = ancestor.ancestor(a, b, parents)
703 703 if c is None:
704 704 return nullid
705 705
706 706 return self.node(c)
707 707
708 708 def _match(self, id):
709 709 if isinstance(id, (long, int)):
710 710 # rev
711 711 return self.node(id)
712 712 if len(id) == 20:
713 713 # possibly a binary node
714 714 # odds of a binary node being all hex in ASCII are 1 in 10**25
715 715 try:
716 716 node = id
717 717 self.rev(node) # quick search the index
718 718 return node
719 719 except LookupError:
720 720 pass # may be partial hex id
721 721 try:
722 722 # str(rev)
723 723 rev = int(id)
724 724 if str(rev) != id:
725 725 raise ValueError
726 726 if rev < 0:
727 727 rev = len(self) + rev
728 728 if rev < 0 or rev >= len(self):
729 729 raise ValueError
730 730 return self.node(rev)
731 731 except (ValueError, OverflowError):
732 732 pass
733 733 if len(id) == 40:
734 734 try:
735 735 # a full hex nodeid?
736 736 node = bin(id)
737 737 self.rev(node)
738 738 return node
739 739 except (TypeError, LookupError):
740 740 pass
741 741
742 742 def _partialmatch(self, id):
743 743 if id in self._pcache:
744 744 return self._pcache[id]
745 745
746 746 if len(id) < 40:
747 747 try:
748 748 # hex(node)[:...]
749 749 l = len(id) // 2 # grab an even number of digits
750 750 prefix = bin(id[:l * 2])
751 751 nl = [e[7] for e in self.index if e[7].startswith(prefix)]
752 752 nl = [n for n in nl if hex(n).startswith(id)]
753 753 if len(nl) > 0:
754 754 if len(nl) == 1:
755 755 self._pcache[id] = nl[0]
756 756 return nl[0]
757 757 raise LookupError(id, self.indexfile,
758 758 _('ambiguous identifier'))
759 759 return None
760 760 except TypeError:
761 761 pass
762 762
763 763 def lookup(self, id):
764 764 """locate a node based on:
765 765 - revision number or str(revision number)
766 766 - nodeid or subset of hex nodeid
767 767 """
768 768 n = self._match(id)
769 769 if n is not None:
770 770 return n
771 771 n = self._partialmatch(id)
772 772 if n:
773 773 return n
774 774
775 775 raise LookupError(id, self.indexfile, _('no match found'))
776 776
777 777 def cmp(self, node, text):
778 778 """compare text with a given file revision
779 779
780 780 returns True if text is different than what is stored.
781 781 """
782 782 p1, p2 = self.parents(node)
783 783 return hash(text, p1, p2) != node
784 784
785 785 def _addchunk(self, offset, data):
786 786 o, d = self._chunkcache
787 787 # try to add to existing cache
788 788 if o + len(d) == offset and len(d) + len(data) < _chunksize:
789 789 self._chunkcache = o, d + data
790 790 else:
791 791 self._chunkcache = offset, data
792 792
793 793 def _loadchunk(self, offset, length):
794 794 if self._inline:
795 795 df = self.opener(self.indexfile)
796 796 else:
797 797 df = self.opener(self.datafile)
798 798
799 799 readahead = max(65536, length)
800 800 df.seek(offset)
801 801 d = df.read(readahead)
802 802 self._addchunk(offset, d)
803 803 if readahead > length:
804 804 return d[:length]
805 805 return d
806 806
807 807 def _getchunk(self, offset, length):
808 808 o, d = self._chunkcache
809 809 l = len(d)
810 810
811 811 # is it in the cache?
812 812 cachestart = offset - o
813 813 cacheend = cachestart + length
814 814 if cachestart >= 0 and cacheend <= l:
815 815 if cachestart == 0 and cacheend == l:
816 816 return d # avoid a copy
817 817 return d[cachestart:cacheend]
818 818
819 819 return self._loadchunk(offset, length)
820 820
821 821 def _chunkraw(self, startrev, endrev):
822 822 start = self.start(startrev)
823 823 length = self.end(endrev) - start
824 824 if self._inline:
825 825 start += (startrev + 1) * self._io.size
826 826 return self._getchunk(start, length)
827 827
828 828 def _chunk(self, rev):
829 829 return decompress(self._chunkraw(rev, rev))
830 830
831 831 def _chunkbase(self, rev):
832 832 return self._chunk(rev)
833 833
834 834 def _chunkclear(self):
835 835 self._chunkcache = (0, '')
836 836
837 837 def deltaparent(self, rev):
838 838 """return deltaparent of the given revision"""
839 839 base = self.index[rev][3]
840 840 if base == rev:
841 841 return nullrev
842 842 elif self._generaldelta:
843 843 return base
844 844 else:
845 845 return rev - 1
846 846
847 847 def revdiff(self, rev1, rev2):
848 848 """return or calculate a delta between two revisions"""
849 849 if rev1 != nullrev and self.deltaparent(rev2) == rev1:
850 850 return self._chunk(rev2)
851 851
852 852 return mdiff.textdiff(self.revision(self.node(rev1)),
853 853 self.revision(self.node(rev2)))
854 854
855 855 def revision(self, node):
856 856 """return an uncompressed revision of a given node"""
857 857 cachedrev = None
858 858 if node == nullid:
859 859 return ""
860 860 if self._cache:
861 861 if self._cache[0] == node:
862 862 return self._cache[2]
863 863 cachedrev = self._cache[1]
864 864
865 865 # look up what we need to read
866 866 text = None
867 867 rev = self.rev(node)
868 868
869 869 # check rev flags
870 870 if self.flags(rev) & ~REVIDX_KNOWN_FLAGS:
871 871 raise RevlogError(_('incompatible revision flag %x') %
872 872 (self.flags(rev) & ~REVIDX_KNOWN_FLAGS))
873 873
874 874 # build delta chain
875 875 chain = []
876 876 index = self.index # for performance
877 877 generaldelta = self._generaldelta
878 878 iterrev = rev
879 879 e = index[iterrev]
880 880 while iterrev != e[3] and iterrev != cachedrev:
881 881 chain.append(iterrev)
882 882 if generaldelta:
883 883 iterrev = e[3]
884 884 else:
885 885 iterrev -= 1
886 886 e = index[iterrev]
887 887 chain.reverse()
888 888 base = iterrev
889 889
890 890 if iterrev == cachedrev:
891 891 # cache hit
892 892 text = self._cache[2]
893 893
894 894 # drop cache to save memory
895 895 self._cache = None
896 896
897 897 self._chunkraw(base, rev)
898 898 if text is None:
899 899 text = self._chunkbase(base)
900 900
901 901 bins = [self._chunk(r) for r in chain]
902 902 text = mdiff.patches(text, bins)
903 903
904 904 text = self._checkhash(text, node, rev)
905 905
906 906 self._cache = (node, rev, text)
907 907 return text
908 908
909 909 def _checkhash(self, text, node, rev):
910 910 p1, p2 = self.parents(node)
911 911 if node != hash(text, p1, p2):
912 912 raise RevlogError(_("integrity check failed on %s:%d")
913 913 % (self.indexfile, rev))
914 914 return text
915 915
916 916 def checkinlinesize(self, tr, fp=None):
917 917 if not self._inline or (self.start(-2) + self.length(-2)) < _maxinline:
918 918 return
919 919
920 920 trinfo = tr.find(self.indexfile)
921 921 if trinfo is None:
922 922 raise RevlogError(_("%s not found in the transaction")
923 923 % self.indexfile)
924 924
925 925 trindex = trinfo[2]
926 926 dataoff = self.start(trindex)
927 927
928 928 tr.add(self.datafile, dataoff)
929 929
930 930 if fp:
931 931 fp.flush()
932 932 fp.close()
933 933
934 934 df = self.opener(self.datafile, 'w')
935 935 try:
936 936 for r in self:
937 937 df.write(self._chunkraw(r, r))
938 938 finally:
939 939 df.close()
940 940
941 941 fp = self.opener(self.indexfile, 'w', atomictemp=True)
942 942 self.version &= ~(REVLOGNGINLINEDATA)
943 943 self._inline = False
944 944 for i in self:
945 945 e = self._io.packentry(self.index[i], self.node, self.version, i)
946 946 fp.write(e)
947 947
948 948 # if we don't call rename, the temp file will never replace the
949 949 # real index
950 950 fp.rename()
951 951
952 952 tr.replace(self.indexfile, trindex * self._io.size)
953 953 self._chunkclear()
954 954
955 955 def addrevision(self, text, transaction, link, p1, p2, cachedelta=None):
956 956 """add a revision to the log
957 957
958 958 text - the revision data to add
959 959 transaction - the transaction object used for rollback
960 960 link - the linkrev data to add
961 961 p1, p2 - the parent nodeids of the revision
962 962 cachedelta - an optional precomputed delta
963 963 """
964 964 node = hash(text, p1, p2)
965 965 if node in self.nodemap:
966 966 return node
967 967
968 968 dfh = None
969 969 if not self._inline:
970 970 dfh = self.opener(self.datafile, "a")
971 971 ifh = self.opener(self.indexfile, "a+")
972 972 try:
973 973 return self._addrevision(node, text, transaction, link, p1, p2,
974 974 cachedelta, ifh, dfh)
975 975 finally:
976 976 if dfh:
977 977 dfh.close()
978 978 ifh.close()
979 979
980 980 def _addrevision(self, node, text, transaction, link, p1, p2,
981 981 cachedelta, ifh, dfh):
982 982 """internal function to add revisions to the log
983 983
984 984 see addrevision for argument descriptions.
985 985 invariants:
986 986 - text is optional (can be None); if not set, cachedelta must be set.
987 987 if both are set, they must correspond to eachother.
988 988 """
989 989 btext = [text]
990 990 def buildtext():
991 991 if btext[0] is not None:
992 992 return btext[0]
993 993 # flush any pending writes here so we can read it in revision
994 994 if dfh:
995 995 dfh.flush()
996 996 ifh.flush()
997 997 basetext = self.revision(self.node(cachedelta[0]))
998 998 btext[0] = mdiff.patch(basetext, cachedelta[1])
999 999 chk = hash(btext[0], p1, p2)
1000 1000 if chk != node:
1001 1001 raise RevlogError(_("consistency error in delta"))
1002 1002 return btext[0]
1003 1003
1004 1004 def builddelta(rev):
1005 1005 # can we use the cached delta?
1006 1006 if cachedelta and cachedelta[0] == rev:
1007 1007 delta = cachedelta[1]
1008 1008 else:
1009 1009 t = buildtext()
1010 1010 ptext = self.revision(self.node(rev))
1011 1011 delta = mdiff.textdiff(ptext, t)
1012 1012 data = compress(delta)
1013 1013 l = len(data[1]) + len(data[0])
1014 1014 if basecache[0] == rev:
1015 1015 chainbase = basecache[1]
1016 1016 else:
1017 1017 chainbase = self.chainbase(rev)
1018 1018 dist = l + offset - self.start(chainbase)
1019 1019 if self._generaldelta:
1020 1020 base = rev
1021 1021 else:
1022 1022 base = chainbase
1023 1023 return dist, l, data, base, chainbase
1024 1024
1025 1025 curr = len(self)
1026 1026 prev = curr - 1
1027 1027 base = chainbase = curr
1028 1028 offset = self.end(prev)
1029 1029 flags = 0
1030 1030 d = None
1031 1031 basecache = self._basecache
1032 1032 p1r, p2r = self.rev(p1), self.rev(p2)
1033 1033
1034 1034 # should we try to build a delta?
1035 1035 if prev != nullrev:
1036 1036 if self._generaldelta:
1037 1037 if p1r >= basecache[1]:
1038 1038 d = builddelta(p1r)
1039 1039 elif p2r >= basecache[1]:
1040 1040 d = builddelta(p2r)
1041 1041 else:
1042 1042 d = builddelta(prev)
1043 1043 else:
1044 1044 d = builddelta(prev)
1045 1045 dist, l, data, base, chainbase = d
1046 1046
1047 1047 # full versions are inserted when the needed deltas
1048 1048 # become comparable to the uncompressed text
1049 1049 if text is None:
1050 1050 textlen = mdiff.patchedsize(self.rawsize(cachedelta[0]),
1051 1051 cachedelta[1])
1052 1052 else:
1053 1053 textlen = len(text)
1054 1054 if d is None or dist > textlen * 2:
1055 1055 text = buildtext()
1056 1056 data = compress(text)
1057 1057 l = len(data[1]) + len(data[0])
1058 1058 base = chainbase = curr
1059 1059
1060 1060 e = (offset_type(offset, flags), l, textlen,
1061 1061 base, link, p1r, p2r, node)
1062 1062 self.index.insert(-1, e)
1063 1063 self.nodemap[node] = curr
1064 1064
1065 1065 entry = self._io.packentry(e, self.node, self.version, curr)
1066 1066 if not self._inline:
1067 1067 transaction.add(self.datafile, offset)
1068 1068 transaction.add(self.indexfile, curr * len(entry))
1069 1069 if data[0]:
1070 1070 dfh.write(data[0])
1071 1071 dfh.write(data[1])
1072 1072 dfh.flush()
1073 1073 ifh.write(entry)
1074 1074 else:
1075 1075 offset += curr * self._io.size
1076 1076 transaction.add(self.indexfile, offset, curr)
1077 1077 ifh.write(entry)
1078 1078 ifh.write(data[0])
1079 1079 ifh.write(data[1])
1080 1080 self.checkinlinesize(transaction, ifh)
1081 1081
1082 1082 if type(text) == str: # only accept immutable objects
1083 1083 self._cache = (node, curr, text)
1084 1084 self._basecache = (curr, chainbase)
1085 1085 return node
1086 1086
1087 1087 def group(self, nodelist, bundler, reorder=None):
1088 1088 """Calculate a delta group, yielding a sequence of changegroup chunks
1089 1089 (strings).
1090 1090
1091 1091 Given a list of changeset revs, return a set of deltas and
1092 1092 metadata corresponding to nodes. The first delta is
1093 1093 first parent(nodelist[0]) -> nodelist[0], the receiver is
1094 1094 guaranteed to have this parent as it has all history before
1095 1095 these changesets. In the case firstparent is nullrev the
1096 1096 changegroup starts with a full revision.
1097 1097 """
1098 1098
1099 1099 # if we don't have any revisions touched by these changesets, bail
1100 1100 if len(nodelist) == 0:
1101 1101 yield bundler.close()
1102 1102 return
1103 1103
1104 1104 # for generaldelta revlogs, we linearize the revs; this will both be
1105 1105 # much quicker and generate a much smaller bundle
1106 1106 if (self._generaldelta and reorder is not False) or reorder:
1107 1107 dag = dagutil.revlogdag(self)
1108 1108 revs = set(self.rev(n) for n in nodelist)
1109 1109 revs = dag.linearize(revs)
1110 1110 else:
1111 1111 revs = sorted([self.rev(n) for n in nodelist])
1112 1112
1113 1113 # add the parent of the first rev
1114 1114 p = self.parentrevs(revs[0])[0]
1115 1115 revs.insert(0, p)
1116 1116
1117 1117 # build deltas
1118 1118 for r in xrange(len(revs) - 1):
1119 1119 prev, curr = revs[r], revs[r + 1]
1120 1120 for c in bundler.revchunk(self, curr, prev):
1121 1121 yield c
1122 1122
1123 1123 yield bundler.close()
1124 1124
1125 1125 def addgroup(self, bundle, linkmapper, transaction):
1126 1126 """
1127 1127 add a delta group
1128 1128
1129 1129 given a set of deltas, add them to the revision log. the
1130 1130 first delta is against its parent, which should be in our
1131 1131 log, the rest are against the previous delta.
1132 1132 """
1133 1133
1134 1134 # track the base of the current delta log
1135 1135 node = None
1136 1136
1137 1137 r = len(self)
1138 1138 end = 0
1139 1139 if r:
1140 1140 end = self.end(r - 1)
1141 1141 ifh = self.opener(self.indexfile, "a+")
1142 1142 isize = r * self._io.size
1143 1143 if self._inline:
1144 1144 transaction.add(self.indexfile, end + isize, r)
1145 1145 dfh = None
1146 1146 else:
1147 1147 transaction.add(self.indexfile, isize, r)
1148 1148 transaction.add(self.datafile, end)
1149 1149 dfh = self.opener(self.datafile, "a")
1150 1150
1151 1151 try:
1152 1152 # loop through our set of deltas
1153 1153 chain = None
1154 1154 while True:
1155 1155 chunkdata = bundle.deltachunk(chain)
1156 1156 if not chunkdata:
1157 1157 break
1158 1158 node = chunkdata['node']
1159 1159 p1 = chunkdata['p1']
1160 1160 p2 = chunkdata['p2']
1161 1161 cs = chunkdata['cs']
1162 1162 deltabase = chunkdata['deltabase']
1163 1163 delta = chunkdata['delta']
1164 1164
1165 1165 link = linkmapper(cs)
1166 1166 if node in self.nodemap:
1167 1167 # this can happen if two branches make the same change
1168 1168 chain = node
1169 1169 continue
1170 1170
1171 1171 for p in (p1, p2):
1172 1172 if not p in self.nodemap:
1173 1173 raise LookupError(p, self.indexfile,
1174 1174 _('unknown parent'))
1175 1175
1176 1176 if deltabase not in self.nodemap:
1177 1177 raise LookupError(deltabase, self.indexfile,
1178 1178 _('unknown delta base'))
1179 1179
1180 1180 baserev = self.rev(deltabase)
1181 1181 chain = self._addrevision(node, None, transaction, link,
1182 1182 p1, p2, (baserev, delta), ifh, dfh)
1183 1183 if not dfh and not self._inline:
1184 1184 # addrevision switched from inline to conventional
1185 1185 # reopen the index
1186 1186 ifh.close()
1187 1187 dfh = self.opener(self.datafile, "a")
1188 1188 ifh = self.opener(self.indexfile, "a")
1189 1189 finally:
1190 1190 if dfh:
1191 1191 dfh.close()
1192 1192 ifh.close()
1193 1193
1194 1194 return node
1195 1195
1196 1196 def strip(self, minlink, transaction):
1197 1197 """truncate the revlog on the first revision with a linkrev >= minlink
1198 1198
1199 1199 This function is called when we're stripping revision minlink and
1200 1200 its descendants from the repository.
1201 1201
1202 1202 We have to remove all revisions with linkrev >= minlink, because
1203 1203 the equivalent changelog revisions will be renumbered after the
1204 1204 strip.
1205 1205
1206 1206 So we truncate the revlog on the first of these revisions, and
1207 1207 trust that the caller has saved the revisions that shouldn't be
1208 1208 removed and that it'll readd them after this truncation.
1209 1209 """
1210 1210 if len(self) == 0:
1211 1211 return
1212 1212
1213 1213 for rev in self:
1214 1214 if self.index[rev][4] >= minlink:
1215 1215 break
1216 1216 else:
1217 1217 return
1218 1218
1219 1219 # first truncate the files on disk
1220 1220 end = self.start(rev)
1221 1221 if not self._inline:
1222 1222 transaction.add(self.datafile, end)
1223 1223 end = rev * self._io.size
1224 1224 else:
1225 1225 end += rev * self._io.size
1226 1226
1227 1227 transaction.add(self.indexfile, end)
1228 1228
1229 1229 # then reset internal state in memory to forget those revisions
1230 1230 self._cache = None
1231 1231 self._chunkclear()
1232 1232 for x in xrange(rev, len(self)):
1233 1233 del self.nodemap[self.node(x)]
1234 1234
1235 1235 del self.index[rev:-1]
1236 1236
1237 1237 def checksize(self):
1238 1238 expected = 0
1239 1239 if len(self):
1240 1240 expected = max(0, self.end(len(self) - 1))
1241 1241
1242 1242 try:
1243 1243 f = self.opener(self.datafile)
1244 1244 f.seek(0, 2)
1245 1245 actual = f.tell()
1246 1246 f.close()
1247 1247 dd = actual - expected
1248 1248 except IOError, inst:
1249 1249 if inst.errno != errno.ENOENT:
1250 1250 raise
1251 1251 dd = 0
1252 1252
1253 1253 try:
1254 1254 f = self.opener(self.indexfile)
1255 1255 f.seek(0, 2)
1256 1256 actual = f.tell()
1257 1257 f.close()
1258 1258 s = self._io.size
1259 1259 i = max(0, actual // s)
1260 1260 di = actual - (i * s)
1261 1261 if self._inline:
1262 1262 databytes = 0
1263 1263 for r in self:
1264 1264 databytes += max(0, self.length(r))
1265 1265 dd = 0
1266 1266 di = actual - len(self) * s - databytes
1267 1267 except IOError, inst:
1268 1268 if inst.errno != errno.ENOENT:
1269 1269 raise
1270 1270 di = 0
1271 1271
1272 1272 return (dd, di)
1273 1273
1274 1274 def files(self):
1275 1275 res = [self.indexfile]
1276 1276 if not self._inline:
1277 1277 res.append(self.datafile)
1278 1278 return res
@@ -1,453 +1,453 b''
1 1 # Copyright (C) 2004, 2005 Canonical Ltd
2 2 #
3 3 # This program is free software; you can redistribute it and/or modify
4 4 # it under the terms of the GNU General Public License as published by
5 5 # the Free Software Foundation; either version 2 of the License, or
6 6 # (at your option) any later version.
7 7 #
8 8 # This program is distributed in the hope that it will be useful,
9 9 # but WITHOUT ANY WARRANTY; without even the implied warranty of
10 10 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 11 # GNU General Public License for more details.
12 12 #
13 13 # You should have received a copy of the GNU General Public License
14 14 # along with this program; if not, write to the Free Software
15 15 # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
16 16
17 17 # mbp: "you know that thing where cvs gives you conflict markers?"
18 18 # s: "i hate that."
19 19
20 20 from i18n import _
21 21 import scmutil, util, mdiff
22 22 import sys, os
23 23
24 24 class CantReprocessAndShowBase(Exception):
25 25 pass
26 26
27 27 def intersect(ra, rb):
28 28 """Given two ranges return the range where they intersect or None.
29 29
30 30 >>> intersect((0, 10), (0, 6))
31 31 (0, 6)
32 32 >>> intersect((0, 10), (5, 15))
33 33 (5, 10)
34 34 >>> intersect((0, 10), (10, 15))
35 35 >>> intersect((0, 9), (10, 15))
36 36 >>> intersect((0, 9), (7, 15))
37 37 (7, 9)
38 38 """
39 39 assert ra[0] <= ra[1]
40 40 assert rb[0] <= rb[1]
41 41
42 42 sa = max(ra[0], rb[0])
43 43 sb = min(ra[1], rb[1])
44 44 if sa < sb:
45 45 return sa, sb
46 46 else:
47 47 return None
48 48
49 49 def compare_range(a, astart, aend, b, bstart, bend):
50 50 """Compare a[astart:aend] == b[bstart:bend], without slicing.
51 51 """
52 52 if (aend - astart) != (bend - bstart):
53 53 return False
54 54 for ia, ib in zip(xrange(astart, aend), xrange(bstart, bend)):
55 55 if a[ia] != b[ib]:
56 56 return False
57 57 else:
58 58 return True
59 59
60 60 class Merge3Text(object):
61 61 """3-way merge of texts.
62 62
63 63 Given strings BASE, OTHER, THIS, tries to produce a combined text
64 64 incorporating the changes from both BASE->OTHER and BASE->THIS."""
65 65 def __init__(self, basetext, atext, btext, base=None, a=None, b=None):
66 66 self.basetext = basetext
67 67 self.atext = atext
68 68 self.btext = btext
69 69 if base is None:
70 70 base = mdiff.splitnewlines(basetext)
71 71 if a is None:
72 72 a = mdiff.splitnewlines(atext)
73 73 if b is None:
74 74 b = mdiff.splitnewlines(btext)
75 75 self.base = base
76 76 self.a = a
77 77 self.b = b
78 78
79 79 def merge_lines(self,
80 80 name_a=None,
81 81 name_b=None,
82 82 name_base=None,
83 83 start_marker='<<<<<<<',
84 84 mid_marker='=======',
85 85 end_marker='>>>>>>>',
86 86 base_marker=None,
87 87 reprocess=False):
88 88 """Return merge in cvs-like form.
89 89 """
90 90 self.conflicts = False
91 91 newline = '\n'
92 92 if len(self.a) > 0:
93 93 if self.a[0].endswith('\r\n'):
94 94 newline = '\r\n'
95 95 elif self.a[0].endswith('\r'):
96 96 newline = '\r'
97 97 if base_marker and reprocess:
98 98 raise CantReprocessAndShowBase()
99 99 if name_a:
100 100 start_marker = start_marker + ' ' + name_a
101 101 if name_b:
102 102 end_marker = end_marker + ' ' + name_b
103 103 if name_base and base_marker:
104 104 base_marker = base_marker + ' ' + name_base
105 105 merge_regions = self.merge_regions()
106 106 if reprocess is True:
107 107 merge_regions = self.reprocess_merge_regions(merge_regions)
108 108 for t in merge_regions:
109 109 what = t[0]
110 110 if what == 'unchanged':
111 111 for i in range(t[1], t[2]):
112 112 yield self.base[i]
113 113 elif what == 'a' or what == 'same':
114 114 for i in range(t[1], t[2]):
115 115 yield self.a[i]
116 116 elif what == 'b':
117 117 for i in range(t[1], t[2]):
118 118 yield self.b[i]
119 119 elif what == 'conflict':
120 120 self.conflicts = True
121 121 yield start_marker + newline
122 122 for i in range(t[3], t[4]):
123 123 yield self.a[i]
124 124 if base_marker is not None:
125 125 yield base_marker + newline
126 126 for i in range(t[1], t[2]):
127 127 yield self.base[i]
128 128 yield mid_marker + newline
129 129 for i in range(t[5], t[6]):
130 130 yield self.b[i]
131 131 yield end_marker + newline
132 132 else:
133 133 raise ValueError(what)
134 134
135 135 def merge_annotated(self):
136 136 """Return merge with conflicts, showing origin of lines.
137 137
138 138 Most useful for debugging merge.
139 139 """
140 140 for t in self.merge_regions():
141 141 what = t[0]
142 142 if what == 'unchanged':
143 143 for i in range(t[1], t[2]):
144 144 yield 'u | ' + self.base[i]
145 145 elif what == 'a' or what == 'same':
146 146 for i in range(t[1], t[2]):
147 147 yield what[0] + ' | ' + self.a[i]
148 148 elif what == 'b':
149 149 for i in range(t[1], t[2]):
150 150 yield 'b | ' + self.b[i]
151 151 elif what == 'conflict':
152 152 yield '<<<<\n'
153 153 for i in range(t[3], t[4]):
154 154 yield 'A | ' + self.a[i]
155 155 yield '----\n'
156 156 for i in range(t[5], t[6]):
157 157 yield 'B | ' + self.b[i]
158 158 yield '>>>>\n'
159 159 else:
160 160 raise ValueError(what)
161 161
162 162 def merge_groups(self):
163 163 """Yield sequence of line groups. Each one is a tuple:
164 164
165 165 'unchanged', lines
166 166 Lines unchanged from base
167 167
168 168 'a', lines
169 169 Lines taken from a
170 170
171 171 'same', lines
172 172 Lines taken from a (and equal to b)
173 173
174 174 'b', lines
175 175 Lines taken from b
176 176
177 177 'conflict', base_lines, a_lines, b_lines
178 178 Lines from base were changed to either a or b and conflict.
179 179 """
180 180 for t in self.merge_regions():
181 181 what = t[0]
182 182 if what == 'unchanged':
183 183 yield what, self.base[t[1]:t[2]]
184 184 elif what == 'a' or what == 'same':
185 185 yield what, self.a[t[1]:t[2]]
186 186 elif what == 'b':
187 187 yield what, self.b[t[1]:t[2]]
188 188 elif what == 'conflict':
189 189 yield (what,
190 190 self.base[t[1]:t[2]],
191 191 self.a[t[3]:t[4]],
192 192 self.b[t[5]:t[6]])
193 193 else:
194 194 raise ValueError(what)
195 195
196 196 def merge_regions(self):
197 197 """Return sequences of matching and conflicting regions.
198 198
199 199 This returns tuples, where the first value says what kind we
200 200 have:
201 201
202 202 'unchanged', start, end
203 203 Take a region of base[start:end]
204 204
205 205 'same', astart, aend
206 206 b and a are different from base but give the same result
207 207
208 208 'a', start, end
209 209 Non-clashing insertion from a[start:end]
210 210
211 211 Method is as follows:
212 212
213 213 The two sequences align only on regions which match the base
214 and both descendents. These are found by doing a two-way diff
214 and both descendants. These are found by doing a two-way diff
215 215 of each one against the base, and then finding the
216 216 intersections between those regions. These "sync regions"
217 217 are by definition unchanged in both and easily dealt with.
218 218
219 219 The regions in between can be in any of three cases:
220 220 conflicted, or changed on only one side.
221 221 """
222 222
223 223 # section a[0:ia] has been disposed of, etc
224 224 iz = ia = ib = 0
225 225
226 226 for zmatch, zend, amatch, aend, bmatch, bend in self.find_sync_regions():
227 227 #print 'match base [%d:%d]' % (zmatch, zend)
228 228
229 229 matchlen = zend - zmatch
230 230 assert matchlen >= 0
231 231 assert matchlen == (aend - amatch)
232 232 assert matchlen == (bend - bmatch)
233 233
234 234 len_a = amatch - ia
235 235 len_b = bmatch - ib
236 236 len_base = zmatch - iz
237 237 assert len_a >= 0
238 238 assert len_b >= 0
239 239 assert len_base >= 0
240 240
241 241 #print 'unmatched a=%d, b=%d' % (len_a, len_b)
242 242
243 243 if len_a or len_b:
244 244 # try to avoid actually slicing the lists
245 245 equal_a = compare_range(self.a, ia, amatch,
246 246 self.base, iz, zmatch)
247 247 equal_b = compare_range(self.b, ib, bmatch,
248 248 self.base, iz, zmatch)
249 249 same = compare_range(self.a, ia, amatch,
250 250 self.b, ib, bmatch)
251 251
252 252 if same:
253 253 yield 'same', ia, amatch
254 254 elif equal_a and not equal_b:
255 255 yield 'b', ib, bmatch
256 256 elif equal_b and not equal_a:
257 257 yield 'a', ia, amatch
258 258 elif not equal_a and not equal_b:
259 259 yield 'conflict', iz, zmatch, ia, amatch, ib, bmatch
260 260 else:
261 261 raise AssertionError("can't handle a=b=base but unmatched")
262 262
263 263 ia = amatch
264 264 ib = bmatch
265 265 iz = zmatch
266 266
267 267 # if the same part of the base was deleted on both sides
268 268 # that's OK, we can just skip it.
269 269
270 270
271 271 if matchlen > 0:
272 272 assert ia == amatch
273 273 assert ib == bmatch
274 274 assert iz == zmatch
275 275
276 276 yield 'unchanged', zmatch, zend
277 277 iz = zend
278 278 ia = aend
279 279 ib = bend
280 280
281 281 def reprocess_merge_regions(self, merge_regions):
282 282 """Where there are conflict regions, remove the agreed lines.
283 283
284 284 Lines where both A and B have made the same changes are
285 285 eliminated.
286 286 """
287 287 for region in merge_regions:
288 288 if region[0] != "conflict":
289 289 yield region
290 290 continue
291 291 type, iz, zmatch, ia, amatch, ib, bmatch = region
292 292 a_region = self.a[ia:amatch]
293 293 b_region = self.b[ib:bmatch]
294 294 matches = mdiff.get_matching_blocks(''.join(a_region),
295 295 ''.join(b_region))
296 296 next_a = ia
297 297 next_b = ib
298 298 for region_ia, region_ib, region_len in matches[:-1]:
299 299 region_ia += ia
300 300 region_ib += ib
301 301 reg = self.mismatch_region(next_a, region_ia, next_b,
302 302 region_ib)
303 303 if reg is not None:
304 304 yield reg
305 305 yield 'same', region_ia, region_len + region_ia
306 306 next_a = region_ia + region_len
307 307 next_b = region_ib + region_len
308 308 reg = self.mismatch_region(next_a, amatch, next_b, bmatch)
309 309 if reg is not None:
310 310 yield reg
311 311
312 312 def mismatch_region(next_a, region_ia, next_b, region_ib):
313 313 if next_a < region_ia or next_b < region_ib:
314 314 return 'conflict', None, None, next_a, region_ia, next_b, region_ib
315 315 mismatch_region = staticmethod(mismatch_region)
316 316
317 317 def find_sync_regions(self):
318 """Return a list of sync regions, where both descendents match the base.
318 """Return a list of sync regions, where both descendants match the base.
319 319
320 320 Generates a list of (base1, base2, a1, a2, b1, b2). There is
321 321 always a zero-length sync region at the end of all the files.
322 322 """
323 323
324 324 ia = ib = 0
325 325 amatches = mdiff.get_matching_blocks(self.basetext, self.atext)
326 326 bmatches = mdiff.get_matching_blocks(self.basetext, self.btext)
327 327 len_a = len(amatches)
328 328 len_b = len(bmatches)
329 329
330 330 sl = []
331 331
332 332 while ia < len_a and ib < len_b:
333 333 abase, amatch, alen = amatches[ia]
334 334 bbase, bmatch, blen = bmatches[ib]
335 335
336 336 # there is an unconflicted block at i; how long does it
337 337 # extend? until whichever one ends earlier.
338 338 i = intersect((abase, abase + alen), (bbase, bbase + blen))
339 339 if i:
340 340 intbase = i[0]
341 341 intend = i[1]
342 342 intlen = intend - intbase
343 343
344 344 # found a match of base[i[0], i[1]]; this may be less than
345 345 # the region that matches in either one
346 346 assert intlen <= alen
347 347 assert intlen <= blen
348 348 assert abase <= intbase
349 349 assert bbase <= intbase
350 350
351 351 asub = amatch + (intbase - abase)
352 352 bsub = bmatch + (intbase - bbase)
353 353 aend = asub + intlen
354 354 bend = bsub + intlen
355 355
356 356 assert self.base[intbase:intend] == self.a[asub:aend], \
357 357 (self.base[intbase:intend], self.a[asub:aend])
358 358
359 359 assert self.base[intbase:intend] == self.b[bsub:bend]
360 360
361 361 sl.append((intbase, intend,
362 362 asub, aend,
363 363 bsub, bend))
364 364
365 365 # advance whichever one ends first in the base text
366 366 if (abase + alen) < (bbase + blen):
367 367 ia += 1
368 368 else:
369 369 ib += 1
370 370
371 371 intbase = len(self.base)
372 372 abase = len(self.a)
373 373 bbase = len(self.b)
374 374 sl.append((intbase, intbase, abase, abase, bbase, bbase))
375 375
376 376 return sl
377 377
378 378 def find_unconflicted(self):
379 379 """Return a list of ranges in base that are not conflicted."""
380 380 am = mdiff.get_matching_blocks(self.basetext, self.atext)
381 381 bm = mdiff.get_matching_blocks(self.basetext, self.btext)
382 382
383 383 unc = []
384 384
385 385 while am and bm:
386 386 # there is an unconflicted block at i; how long does it
387 387 # extend? until whichever one ends earlier.
388 388 a1 = am[0][0]
389 389 a2 = a1 + am[0][2]
390 390 b1 = bm[0][0]
391 391 b2 = b1 + bm[0][2]
392 392 i = intersect((a1, a2), (b1, b2))
393 393 if i:
394 394 unc.append(i)
395 395
396 396 if a2 < b2:
397 397 del am[0]
398 398 else:
399 399 del bm[0]
400 400
401 401 return unc
402 402
403 403 def simplemerge(ui, local, base, other, **opts):
404 404 def readfile(filename):
405 405 f = open(filename, "rb")
406 406 text = f.read()
407 407 f.close()
408 408 if util.binary(text):
409 409 msg = _("%s looks like a binary file.") % filename
410 410 if not opts.get('quiet'):
411 411 ui.warn(_('warning: %s\n') % msg)
412 412 if not opts.get('text'):
413 413 raise util.Abort(msg)
414 414 return text
415 415
416 416 name_a = local
417 417 name_b = other
418 418 labels = opts.get('label', [])
419 419 if labels:
420 420 name_a = labels.pop(0)
421 421 if labels:
422 422 name_b = labels.pop(0)
423 423 if labels:
424 424 raise util.Abort(_("can only specify two labels."))
425 425
426 426 try:
427 427 localtext = readfile(local)
428 428 basetext = readfile(base)
429 429 othertext = readfile(other)
430 430 except util.Abort:
431 431 return 1
432 432
433 433 local = os.path.realpath(local)
434 434 if not opts.get('print'):
435 435 opener = scmutil.opener(os.path.dirname(local))
436 436 out = opener(os.path.basename(local), "w", atomictemp=True)
437 437 else:
438 438 out = sys.stdout
439 439
440 440 reprocess = not opts.get('no_minimal')
441 441
442 442 m3 = Merge3Text(basetext, localtext, othertext)
443 443 for line in m3.merge_lines(name_a=name_a, name_b=name_b,
444 444 reprocess=reprocess):
445 445 out.write(line)
446 446
447 447 if not opts.get('print'):
448 448 out.rename()
449 449
450 450 if m3.conflicts:
451 451 if not opts.get('quiet'):
452 452 ui.warn(_("warning: conflicts during merge.\n"))
453 453 return 1
General Comments 0
You need to be logged in to leave comments. Login now