##// END OF EJS Templates
base85: switch to policy importer
Yuya Nishihara -
r32368:008d37c4 default
parent child Browse files
Show More
@@ -1,106 +1,105
1 1 #!/usr/bin/env python
2 2 #
3 3 # check-py3-compat - check Python 3 compatibility of Mercurial files
4 4 #
5 5 # Copyright 2015 Gregory Szorc <gregory.szorc@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 from __future__ import absolute_import, print_function
11 11
12 12 import ast
13 13 import importlib
14 14 import os
15 15 import sys
16 16 import traceback
17 17
18 18 # Modules that have both Python and C implementations.
19 19 _dualmodules = (
20 'base85.py',
21 20 'bdiff.py',
22 21 'diffhelpers.py',
23 22 'mpatch.py',
24 23 'parsers.py',
25 24 )
26 25
27 26 def check_compat_py2(f):
28 27 """Check Python 3 compatibility for a file with Python 2"""
29 28 with open(f, 'rb') as fh:
30 29 content = fh.read()
31 30 root = ast.parse(content)
32 31
33 32 # Ignore empty files.
34 33 if not root.body:
35 34 return
36 35
37 36 futures = set()
38 37 haveprint = False
39 38 for node in ast.walk(root):
40 39 if isinstance(node, ast.ImportFrom):
41 40 if node.module == '__future__':
42 41 futures |= set(n.name for n in node.names)
43 42 elif isinstance(node, ast.Print):
44 43 haveprint = True
45 44
46 45 if 'absolute_import' not in futures:
47 46 print('%s not using absolute_import' % f)
48 47 if haveprint and 'print_function' not in futures:
49 48 print('%s requires print_function' % f)
50 49
51 50 def check_compat_py3(f):
52 51 """Check Python 3 compatibility of a file with Python 3."""
53 52 with open(f, 'rb') as fh:
54 53 content = fh.read()
55 54
56 55 try:
57 56 ast.parse(content)
58 57 except SyntaxError as e:
59 58 print('%s: invalid syntax: %s' % (f, e))
60 59 return
61 60
62 61 # Try to import the module.
63 62 # For now we only support mercurial.* and hgext.* modules because figuring
64 63 # out module paths for things not in a package can be confusing.
65 64 if f.startswith(('hgext/', 'mercurial/')) and not f.endswith('__init__.py'):
66 65 assert f.endswith('.py')
67 66 name = f.replace('/', '.')[:-3]
68 67 if f.endswith(_dualmodules):
69 68 name = name.replace('.pure.', '.')
70 69 try:
71 70 importlib.import_module(name)
72 71 except Exception as e:
73 72 exc_type, exc_value, tb = sys.exc_info()
74 73 # We walk the stack and ignore frames from our custom importer,
75 74 # import mechanisms, and stdlib modules. This kinda/sorta
76 75 # emulates CPython behavior in import.c while also attempting
77 76 # to pin blame on a Mercurial file.
78 77 for frame in reversed(traceback.extract_tb(tb)):
79 78 if frame.name == '_call_with_frames_removed':
80 79 continue
81 80 if 'importlib' in frame.filename:
82 81 continue
83 82 if 'mercurial/__init__.py' in frame.filename:
84 83 continue
85 84 if frame.filename.startswith(sys.prefix):
86 85 continue
87 86 break
88 87
89 88 if frame.filename:
90 89 filename = os.path.basename(frame.filename)
91 90 print('%s: error importing: <%s> %s (error at %s:%d)' % (
92 91 f, type(e).__name__, e, filename, frame.lineno))
93 92 else:
94 93 print('%s: error importing module: <%s> %s (line %d)' % (
95 94 f, type(e).__name__, e, frame.lineno))
96 95
97 96 if __name__ == '__main__':
98 97 if sys.version_info[0] == 2:
99 98 fn = check_compat_py2
100 99 else:
101 100 fn = check_compat_py3
102 101
103 102 for f in sys.argv[1:]:
104 103 fn(f)
105 104
106 105 sys.exit(0)
@@ -1,735 +1,734
1 1 #!/usr/bin/env python
2 2
3 3 from __future__ import absolute_import, print_function
4 4
5 5 import ast
6 6 import collections
7 7 import os
8 8 import re
9 9 import sys
10 10
11 11 # Import a minimal set of stdlib modules needed for list_stdlib_modules()
12 12 # to work when run from a virtualenv. The modules were chosen empirically
13 13 # so that the return value matches the return value without virtualenv.
14 14 if True: # disable lexical sorting checks
15 15 import BaseHTTPServer
16 16 import zlib
17 17
18 18 # Whitelist of modules that symbols can be directly imported from.
19 19 allowsymbolimports = (
20 20 '__future__',
21 21 'mercurial.hgweb.common',
22 22 'mercurial.hgweb.request',
23 23 'mercurial.i18n',
24 24 'mercurial.node',
25 25 )
26 26
27 27 # Modules that have both Python and C implementations.
28 28 _dualmodules = (
29 'base85.py',
30 29 'bdiff.py',
31 30 'diffhelpers.py',
32 31 'mpatch.py',
33 32 'parsers.py',
34 33 )
35 34
36 35 # Modules that must be aliased because they are commonly confused with
37 36 # common variables and can create aliasing and readability issues.
38 37 requirealias = {
39 38 'ui': 'uimod',
40 39 }
41 40
42 41 def usingabsolute(root):
43 42 """Whether absolute imports are being used."""
44 43 if sys.version_info[0] >= 3:
45 44 return True
46 45
47 46 for node in ast.walk(root):
48 47 if isinstance(node, ast.ImportFrom):
49 48 if node.module == '__future__':
50 49 for n in node.names:
51 50 if n.name == 'absolute_import':
52 51 return True
53 52
54 53 return False
55 54
56 55 def walklocal(root):
57 56 """Recursively yield all descendant nodes but not in a different scope"""
58 57 todo = collections.deque(ast.iter_child_nodes(root))
59 58 yield root, False
60 59 while todo:
61 60 node = todo.popleft()
62 61 newscope = isinstance(node, ast.FunctionDef)
63 62 if not newscope:
64 63 todo.extend(ast.iter_child_nodes(node))
65 64 yield node, newscope
66 65
67 66 def dotted_name_of_path(path, trimpure=False):
68 67 """Given a relative path to a source file, return its dotted module name.
69 68
70 69 >>> dotted_name_of_path('mercurial/error.py')
71 70 'mercurial.error'
72 71 >>> dotted_name_of_path('mercurial/pure/parsers.py', trimpure=True)
73 72 'mercurial.parsers'
74 73 >>> dotted_name_of_path('zlibmodule.so')
75 74 'zlib'
76 75 """
77 76 parts = path.replace(os.sep, '/').split('/')
78 77 parts[-1] = parts[-1].split('.', 1)[0] # remove .py and .so and .ARCH.so
79 78 if parts[-1].endswith('module'):
80 79 parts[-1] = parts[-1][:-6]
81 80 if trimpure:
82 81 return '.'.join(p for p in parts if p != 'pure')
83 82 return '.'.join(parts)
84 83
85 84 def fromlocalfunc(modulename, localmods):
86 85 """Get a function to examine which locally defined module the
87 86 target source imports via a specified name.
88 87
89 88 `modulename` is an `dotted_name_of_path()`-ed source file path,
90 89 which may have `.__init__` at the end of it, of the target source.
91 90
92 91 `localmods` is a dict (or set), of which key is an absolute
93 92 `dotted_name_of_path()`-ed source file path of locally defined (=
94 93 Mercurial specific) modules.
95 94
96 95 This function assumes that module names not existing in
97 96 `localmods` are from the Python standard library.
98 97
99 98 This function returns the function, which takes `name` argument,
100 99 and returns `(absname, dottedpath, hassubmod)` tuple if `name`
101 100 matches against locally defined module. Otherwise, it returns
102 101 False.
103 102
104 103 It is assumed that `name` doesn't have `.__init__`.
105 104
106 105 `absname` is an absolute module name of specified `name`
107 106 (e.g. "hgext.convert"). This can be used to compose prefix for sub
108 107 modules or so.
109 108
110 109 `dottedpath` is a `dotted_name_of_path()`-ed source file path
111 110 (e.g. "hgext.convert.__init__") of `name`. This is used to look
112 111 module up in `localmods` again.
113 112
114 113 `hassubmod` is whether it may have sub modules under it (for
115 114 convenient, even though this is also equivalent to "absname !=
116 115 dottednpath")
117 116
118 117 >>> localmods = {'foo.__init__': True, 'foo.foo1': True,
119 118 ... 'foo.bar.__init__': True, 'foo.bar.bar1': True,
120 119 ... 'baz.__init__': True, 'baz.baz1': True }
121 120 >>> fromlocal = fromlocalfunc('foo.xxx', localmods)
122 121 >>> # relative
123 122 >>> fromlocal('foo1')
124 123 ('foo.foo1', 'foo.foo1', False)
125 124 >>> fromlocal('bar')
126 125 ('foo.bar', 'foo.bar.__init__', True)
127 126 >>> fromlocal('bar.bar1')
128 127 ('foo.bar.bar1', 'foo.bar.bar1', False)
129 128 >>> # absolute
130 129 >>> fromlocal('baz')
131 130 ('baz', 'baz.__init__', True)
132 131 >>> fromlocal('baz.baz1')
133 132 ('baz.baz1', 'baz.baz1', False)
134 133 >>> # unknown = maybe standard library
135 134 >>> fromlocal('os')
136 135 False
137 136 >>> fromlocal(None, 1)
138 137 ('foo', 'foo.__init__', True)
139 138 >>> fromlocal('foo1', 1)
140 139 ('foo.foo1', 'foo.foo1', False)
141 140 >>> fromlocal2 = fromlocalfunc('foo.xxx.yyy', localmods)
142 141 >>> fromlocal2(None, 2)
143 142 ('foo', 'foo.__init__', True)
144 143 >>> fromlocal2('bar2', 1)
145 144 False
146 145 >>> fromlocal2('bar', 2)
147 146 ('foo.bar', 'foo.bar.__init__', True)
148 147 """
149 148 prefix = '.'.join(modulename.split('.')[:-1])
150 149 if prefix:
151 150 prefix += '.'
152 151 def fromlocal(name, level=0):
153 152 # name is false value when relative imports are used.
154 153 if not name:
155 154 # If relative imports are used, level must not be absolute.
156 155 assert level > 0
157 156 candidates = ['.'.join(modulename.split('.')[:-level])]
158 157 else:
159 158 if not level:
160 159 # Check relative name first.
161 160 candidates = [prefix + name, name]
162 161 else:
163 162 candidates = ['.'.join(modulename.split('.')[:-level]) +
164 163 '.' + name]
165 164
166 165 for n in candidates:
167 166 if n in localmods:
168 167 return (n, n, False)
169 168 dottedpath = n + '.__init__'
170 169 if dottedpath in localmods:
171 170 return (n, dottedpath, True)
172 171 return False
173 172 return fromlocal
174 173
175 174 def list_stdlib_modules():
176 175 """List the modules present in the stdlib.
177 176
178 177 >>> mods = set(list_stdlib_modules())
179 178 >>> 'BaseHTTPServer' in mods
180 179 True
181 180
182 181 os.path isn't really a module, so it's missing:
183 182
184 183 >>> 'os.path' in mods
185 184 False
186 185
187 186 sys requires special treatment, because it's baked into the
188 187 interpreter, but it should still appear:
189 188
190 189 >>> 'sys' in mods
191 190 True
192 191
193 192 >>> 'collections' in mods
194 193 True
195 194
196 195 >>> 'cStringIO' in mods
197 196 True
198 197
199 198 >>> 'cffi' in mods
200 199 True
201 200 """
202 201 for m in sys.builtin_module_names:
203 202 yield m
204 203 # These modules only exist on windows, but we should always
205 204 # consider them stdlib.
206 205 for m in ['msvcrt', '_winreg']:
207 206 yield m
208 207 yield 'builtins' # python3 only
209 208 for m in 'fcntl', 'grp', 'pwd', 'termios': # Unix only
210 209 yield m
211 210 for m in 'cPickle', 'datetime': # in Python (not C) on PyPy
212 211 yield m
213 212 for m in ['cffi']:
214 213 yield m
215 214 stdlib_prefixes = {sys.prefix, sys.exec_prefix}
216 215 # We need to supplement the list of prefixes for the search to work
217 216 # when run from within a virtualenv.
218 217 for mod in (BaseHTTPServer, zlib):
219 218 try:
220 219 # Not all module objects have a __file__ attribute.
221 220 filename = mod.__file__
222 221 except AttributeError:
223 222 continue
224 223 dirname = os.path.dirname(filename)
225 224 for prefix in stdlib_prefixes:
226 225 if dirname.startswith(prefix):
227 226 # Then this directory is redundant.
228 227 break
229 228 else:
230 229 stdlib_prefixes.add(dirname)
231 230 for libpath in sys.path:
232 231 # We want to walk everything in sys.path that starts with
233 232 # something in stdlib_prefixes.
234 233 if not any(libpath.startswith(p) for p in stdlib_prefixes):
235 234 continue
236 235 for top, dirs, files in os.walk(libpath):
237 236 for i, d in reversed(list(enumerate(dirs))):
238 237 if (not os.path.exists(os.path.join(top, d, '__init__.py'))
239 238 or top == libpath and d in ('hgext', 'mercurial')):
240 239 del dirs[i]
241 240 for name in files:
242 241 if not name.endswith(('.py', '.so', '.pyc', '.pyo', '.pyd')):
243 242 continue
244 243 if name.startswith('__init__.py'):
245 244 full_path = top
246 245 else:
247 246 full_path = os.path.join(top, name)
248 247 rel_path = full_path[len(libpath) + 1:]
249 248 mod = dotted_name_of_path(rel_path)
250 249 yield mod
251 250
252 251 stdlib_modules = set(list_stdlib_modules())
253 252
254 253 def imported_modules(source, modulename, f, localmods, ignore_nested=False):
255 254 """Given the source of a file as a string, yield the names
256 255 imported by that file.
257 256
258 257 Args:
259 258 source: The python source to examine as a string.
260 259 modulename: of specified python source (may have `__init__`)
261 260 localmods: dict of locally defined module names (may have `__init__`)
262 261 ignore_nested: If true, import statements that do not start in
263 262 column zero will be ignored.
264 263
265 264 Returns:
266 265 A list of absolute module names imported by the given source.
267 266
268 267 >>> f = 'foo/xxx.py'
269 268 >>> modulename = 'foo.xxx'
270 269 >>> localmods = {'foo.__init__': True,
271 270 ... 'foo.foo1': True, 'foo.foo2': True,
272 271 ... 'foo.bar.__init__': True, 'foo.bar.bar1': True,
273 272 ... 'baz.__init__': True, 'baz.baz1': True }
274 273 >>> # standard library (= not locally defined ones)
275 274 >>> sorted(imported_modules(
276 275 ... 'from stdlib1 import foo, bar; import stdlib2',
277 276 ... modulename, f, localmods))
278 277 []
279 278 >>> # relative importing
280 279 >>> sorted(imported_modules(
281 280 ... 'import foo1; from bar import bar1',
282 281 ... modulename, f, localmods))
283 282 ['foo.bar.bar1', 'foo.foo1']
284 283 >>> sorted(imported_modules(
285 284 ... 'from bar.bar1 import name1, name2, name3',
286 285 ... modulename, f, localmods))
287 286 ['foo.bar.bar1']
288 287 >>> # absolute importing
289 288 >>> sorted(imported_modules(
290 289 ... 'from baz import baz1, name1',
291 290 ... modulename, f, localmods))
292 291 ['baz.__init__', 'baz.baz1']
293 292 >>> # mixed importing, even though it shouldn't be recommended
294 293 >>> sorted(imported_modules(
295 294 ... 'import stdlib, foo1, baz',
296 295 ... modulename, f, localmods))
297 296 ['baz.__init__', 'foo.foo1']
298 297 >>> # ignore_nested
299 298 >>> sorted(imported_modules(
300 299 ... '''import foo
301 300 ... def wat():
302 301 ... import bar
303 302 ... ''', modulename, f, localmods))
304 303 ['foo.__init__', 'foo.bar.__init__']
305 304 >>> sorted(imported_modules(
306 305 ... '''import foo
307 306 ... def wat():
308 307 ... import bar
309 308 ... ''', modulename, f, localmods, ignore_nested=True))
310 309 ['foo.__init__']
311 310 """
312 311 fromlocal = fromlocalfunc(modulename, localmods)
313 312 for node in ast.walk(ast.parse(source, f)):
314 313 if ignore_nested and getattr(node, 'col_offset', 0) > 0:
315 314 continue
316 315 if isinstance(node, ast.Import):
317 316 for n in node.names:
318 317 found = fromlocal(n.name)
319 318 if not found:
320 319 # this should import standard library
321 320 continue
322 321 yield found[1]
323 322 elif isinstance(node, ast.ImportFrom):
324 323 found = fromlocal(node.module, node.level)
325 324 if not found:
326 325 # this should import standard library
327 326 continue
328 327
329 328 absname, dottedpath, hassubmod = found
330 329 if not hassubmod:
331 330 # "dottedpath" is not a package; must be imported
332 331 yield dottedpath
333 332 # examination of "node.names" should be redundant
334 333 # e.g.: from mercurial.node import nullid, nullrev
335 334 continue
336 335
337 336 modnotfound = False
338 337 prefix = absname + '.'
339 338 for n in node.names:
340 339 found = fromlocal(prefix + n.name)
341 340 if not found:
342 341 # this should be a function or a property of "node.module"
343 342 modnotfound = True
344 343 continue
345 344 yield found[1]
346 345 if modnotfound:
347 346 # "dottedpath" is a package, but imported because of non-module
348 347 # lookup
349 348 yield dottedpath
350 349
351 350 def verify_import_convention(module, source, localmods):
352 351 """Verify imports match our established coding convention.
353 352
354 353 We have 2 conventions: legacy and modern. The modern convention is in
355 354 effect when using absolute imports.
356 355
357 356 The legacy convention only looks for mixed imports. The modern convention
358 357 is much more thorough.
359 358 """
360 359 root = ast.parse(source)
361 360 absolute = usingabsolute(root)
362 361
363 362 if absolute:
364 363 return verify_modern_convention(module, root, localmods)
365 364 else:
366 365 return verify_stdlib_on_own_line(root)
367 366
368 367 def verify_modern_convention(module, root, localmods, root_col_offset=0):
369 368 """Verify a file conforms to the modern import convention rules.
370 369
371 370 The rules of the modern convention are:
372 371
373 372 * Ordering is stdlib followed by local imports. Each group is lexically
374 373 sorted.
375 374 * Importing multiple modules via "import X, Y" is not allowed: use
376 375 separate import statements.
377 376 * Importing multiple modules via "from X import ..." is allowed if using
378 377 parenthesis and one entry per line.
379 378 * Only 1 relative import statement per import level ("from .", "from ..")
380 379 is allowed.
381 380 * Relative imports from higher levels must occur before lower levels. e.g.
382 381 "from .." must be before "from .".
383 382 * Imports from peer packages should use relative import (e.g. do not
384 383 "import mercurial.foo" from a "mercurial.*" module).
385 384 * Symbols can only be imported from specific modules (see
386 385 `allowsymbolimports`). For other modules, first import the module then
387 386 assign the symbol to a module-level variable. In addition, these imports
388 387 must be performed before other local imports. This rule only
389 388 applies to import statements outside of any blocks.
390 389 * Relative imports from the standard library are not allowed.
391 390 * Certain modules must be aliased to alternate names to avoid aliasing
392 391 and readability problems. See `requirealias`.
393 392 """
394 393 topmodule = module.split('.')[0]
395 394 fromlocal = fromlocalfunc(module, localmods)
396 395
397 396 # Whether a local/non-stdlib import has been performed.
398 397 seenlocal = None
399 398 # Whether a local/non-stdlib, non-symbol import has been seen.
400 399 seennonsymbollocal = False
401 400 # The last name to be imported (for sorting).
402 401 lastname = None
403 402 laststdlib = None
404 403 # Relative import levels encountered so far.
405 404 seenlevels = set()
406 405
407 406 for node, newscope in walklocal(root):
408 407 def msg(fmt, *args):
409 408 return (fmt % args, node.lineno)
410 409 if newscope:
411 410 # Check for local imports in function
412 411 for r in verify_modern_convention(module, node, localmods,
413 412 node.col_offset + 4):
414 413 yield r
415 414 elif isinstance(node, ast.Import):
416 415 # Disallow "import foo, bar" and require separate imports
417 416 # for each module.
418 417 if len(node.names) > 1:
419 418 yield msg('multiple imported names: %s',
420 419 ', '.join(n.name for n in node.names))
421 420
422 421 name = node.names[0].name
423 422 asname = node.names[0].asname
424 423
425 424 stdlib = name in stdlib_modules
426 425
427 426 # Ignore sorting rules on imports inside blocks.
428 427 if node.col_offset == root_col_offset:
429 428 if lastname and name < lastname and laststdlib == stdlib:
430 429 yield msg('imports not lexically sorted: %s < %s',
431 430 name, lastname)
432 431
433 432 lastname = name
434 433 laststdlib = stdlib
435 434
436 435 # stdlib imports should be before local imports.
437 436 if stdlib and seenlocal and node.col_offset == root_col_offset:
438 437 yield msg('stdlib import "%s" follows local import: %s',
439 438 name, seenlocal)
440 439
441 440 if not stdlib:
442 441 seenlocal = name
443 442
444 443 # Import of sibling modules should use relative imports.
445 444 topname = name.split('.')[0]
446 445 if topname == topmodule:
447 446 yield msg('import should be relative: %s', name)
448 447
449 448 if name in requirealias and asname != requirealias[name]:
450 449 yield msg('%s module must be "as" aliased to %s',
451 450 name, requirealias[name])
452 451
453 452 elif isinstance(node, ast.ImportFrom):
454 453 # Resolve the full imported module name.
455 454 if node.level > 0:
456 455 fullname = '.'.join(module.split('.')[:-node.level])
457 456 if node.module:
458 457 fullname += '.%s' % node.module
459 458 else:
460 459 assert node.module
461 460 fullname = node.module
462 461
463 462 topname = fullname.split('.')[0]
464 463 if topname == topmodule:
465 464 yield msg('import should be relative: %s', fullname)
466 465
467 466 # __future__ is special since it needs to come first and use
468 467 # symbol import.
469 468 if fullname != '__future__':
470 469 if not fullname or fullname in stdlib_modules:
471 470 yield msg('relative import of stdlib module')
472 471 else:
473 472 seenlocal = fullname
474 473
475 474 # Direct symbol import is only allowed from certain modules and
476 475 # must occur before non-symbol imports.
477 476 found = fromlocal(node.module, node.level)
478 477 if found and found[2]: # node.module is a package
479 478 prefix = found[0] + '.'
480 479 symbols = [n.name for n in node.names
481 480 if not fromlocal(prefix + n.name)]
482 481 else:
483 482 symbols = [n.name for n in node.names]
484 483 if node.module and node.col_offset == root_col_offset:
485 484 if symbols and fullname not in allowsymbolimports:
486 485 yield msg('direct symbol import %s from %s',
487 486 ', '.join(symbols), fullname)
488 487
489 488 if symbols and seennonsymbollocal:
490 489 yield msg('symbol import follows non-symbol import: %s',
491 490 fullname)
492 491 if not symbols and fullname not in stdlib_modules:
493 492 seennonsymbollocal = True
494 493
495 494 if not node.module:
496 495 assert node.level
497 496
498 497 # Only allow 1 group per level.
499 498 if (node.level in seenlevels
500 499 and node.col_offset == root_col_offset):
501 500 yield msg('multiple "from %s import" statements',
502 501 '.' * node.level)
503 502
504 503 # Higher-level groups come before lower-level groups.
505 504 if any(node.level > l for l in seenlevels):
506 505 yield msg('higher-level import should come first: %s',
507 506 fullname)
508 507
509 508 seenlevels.add(node.level)
510 509
511 510 # Entries in "from .X import ( ... )" lists must be lexically
512 511 # sorted.
513 512 lastentryname = None
514 513
515 514 for n in node.names:
516 515 if lastentryname and n.name < lastentryname:
517 516 yield msg('imports from %s not lexically sorted: %s < %s',
518 517 fullname, n.name, lastentryname)
519 518
520 519 lastentryname = n.name
521 520
522 521 if n.name in requirealias and n.asname != requirealias[n.name]:
523 522 yield msg('%s from %s must be "as" aliased to %s',
524 523 n.name, fullname, requirealias[n.name])
525 524
526 525 def verify_stdlib_on_own_line(root):
527 526 """Given some python source, verify that stdlib imports are done
528 527 in separate statements from relative local module imports.
529 528
530 529 >>> list(verify_stdlib_on_own_line(ast.parse('import sys, foo')))
531 530 [('mixed imports\\n stdlib: sys\\n relative: foo', 1)]
532 531 >>> list(verify_stdlib_on_own_line(ast.parse('import sys, os')))
533 532 []
534 533 >>> list(verify_stdlib_on_own_line(ast.parse('import foo, bar')))
535 534 []
536 535 """
537 536 for node in ast.walk(root):
538 537 if isinstance(node, ast.Import):
539 538 from_stdlib = {False: [], True: []}
540 539 for n in node.names:
541 540 from_stdlib[n.name in stdlib_modules].append(n.name)
542 541 if from_stdlib[True] and from_stdlib[False]:
543 542 yield ('mixed imports\n stdlib: %s\n relative: %s' %
544 543 (', '.join(sorted(from_stdlib[True])),
545 544 ', '.join(sorted(from_stdlib[False]))), node.lineno)
546 545
547 546 class CircularImport(Exception):
548 547 pass
549 548
550 549 def checkmod(mod, imports):
551 550 shortest = {}
552 551 visit = [[mod]]
553 552 while visit:
554 553 path = visit.pop(0)
555 554 for i in sorted(imports.get(path[-1], [])):
556 555 if len(path) < shortest.get(i, 1000):
557 556 shortest[i] = len(path)
558 557 if i in path:
559 558 if i == path[0]:
560 559 raise CircularImport(path)
561 560 continue
562 561 visit.append(path + [i])
563 562
564 563 def rotatecycle(cycle):
565 564 """arrange a cycle so that the lexicographically first module listed first
566 565
567 566 >>> rotatecycle(['foo', 'bar'])
568 567 ['bar', 'foo', 'bar']
569 568 """
570 569 lowest = min(cycle)
571 570 idx = cycle.index(lowest)
572 571 return cycle[idx:] + cycle[:idx] + [lowest]
573 572
574 573 def find_cycles(imports):
575 574 """Find cycles in an already-loaded import graph.
576 575
577 576 All module names recorded in `imports` should be absolute one.
578 577
579 578 >>> from __future__ import print_function
580 579 >>> imports = {'top.foo': ['top.bar', 'os.path', 'top.qux'],
581 580 ... 'top.bar': ['top.baz', 'sys'],
582 581 ... 'top.baz': ['top.foo'],
583 582 ... 'top.qux': ['top.foo']}
584 583 >>> print('\\n'.join(sorted(find_cycles(imports))))
585 584 top.bar -> top.baz -> top.foo -> top.bar
586 585 top.foo -> top.qux -> top.foo
587 586 """
588 587 cycles = set()
589 588 for mod in sorted(imports.keys()):
590 589 try:
591 590 checkmod(mod, imports)
592 591 except CircularImport as e:
593 592 cycle = e.args[0]
594 593 cycles.add(" -> ".join(rotatecycle(cycle)))
595 594 return cycles
596 595
597 596 def _cycle_sortkey(c):
598 597 return len(c), c
599 598
600 599 def embedded(f, modname, src):
601 600 """Extract embedded python code
602 601
603 602 >>> def test(fn, lines):
604 603 ... for s, m, f, l in embedded(fn, "example", lines):
605 604 ... print("%s %s %s" % (m, f, l))
606 605 ... print(repr(s))
607 606 >>> lines = [
608 607 ... 'comment',
609 608 ... ' >>> from __future__ import print_function',
610 609 ... " >>> ' multiline",
611 610 ... " ... string'",
612 611 ... ' ',
613 612 ... 'comment',
614 613 ... ' $ cat > foo.py <<EOF',
615 614 ... ' > from __future__ import print_function',
616 615 ... ' > EOF',
617 616 ... ]
618 617 >>> test("example.t", lines)
619 618 example[2] doctest.py 2
620 619 "from __future__ import print_function\\n' multiline\\nstring'\\n"
621 620 example[7] foo.py 7
622 621 'from __future__ import print_function\\n'
623 622 """
624 623 inlinepython = 0
625 624 shpython = 0
626 625 script = []
627 626 prefix = 6
628 627 t = ''
629 628 n = 0
630 629 for l in src:
631 630 n += 1
632 631 if not l.endswith(b'\n'):
633 632 l += b'\n'
634 633 if l.startswith(b' >>> '): # python inlines
635 634 if shpython:
636 635 print("%s:%d: Parse Error" % (f, n))
637 636 if not inlinepython:
638 637 # We've just entered a Python block.
639 638 inlinepython = n
640 639 t = 'doctest.py'
641 640 script.append(l[prefix:])
642 641 continue
643 642 if l.startswith(b' ... '): # python inlines
644 643 script.append(l[prefix:])
645 644 continue
646 645 cat = re.search(r"\$ \s*cat\s*>\s*(\S+\.py)\s*<<\s*EOF", l)
647 646 if cat:
648 647 if inlinepython:
649 648 yield ''.join(script), ("%s[%d]" %
650 649 (modname, inlinepython)), t, inlinepython
651 650 script = []
652 651 inlinepython = 0
653 652 shpython = n
654 653 t = cat.group(1)
655 654 continue
656 655 if shpython and l.startswith(b' > '): # sh continuation
657 656 if l == b' > EOF\n':
658 657 yield ''.join(script), ("%s[%d]" %
659 658 (modname, shpython)), t, shpython
660 659 script = []
661 660 shpython = 0
662 661 else:
663 662 script.append(l[4:])
664 663 continue
665 664 if inlinepython and l == b' \n':
666 665 yield ''.join(script), ("%s[%d]" %
667 666 (modname, inlinepython)), t, inlinepython
668 667 script = []
669 668 inlinepython = 0
670 669 continue
671 670
672 671 def sources(f, modname):
673 672 """Yields possibly multiple sources from a filepath
674 673
675 674 input: filepath, modulename
676 675 yields: script(string), modulename, filepath, linenumber
677 676
678 677 For embedded scripts, the modulename and filepath will be different
679 678 from the function arguments. linenumber is an offset relative to
680 679 the input file.
681 680 """
682 681 py = False
683 682 if not f.endswith('.t'):
684 683 with open(f) as src:
685 684 yield src.read(), modname, f, 0
686 685 py = True
687 686 if py or f.endswith('.t'):
688 687 with open(f) as src:
689 688 for script, modname, t, line in embedded(f, modname, src):
690 689 yield script, modname, t, line
691 690
692 691 def main(argv):
693 692 if len(argv) < 2 or (argv[1] == '-' and len(argv) > 2):
694 693 print('Usage: %s {-|file [file] [file] ...}')
695 694 return 1
696 695 if argv[1] == '-':
697 696 argv = argv[:1]
698 697 argv.extend(l.rstrip() for l in sys.stdin.readlines())
699 698 localmods = {}
700 699 used_imports = {}
701 700 any_errors = False
702 701 for source_path in argv[1:]:
703 702 trimpure = source_path.endswith(_dualmodules)
704 703 modname = dotted_name_of_path(source_path, trimpure=trimpure)
705 704 localmods[modname] = source_path
706 705 for localmodname, source_path in sorted(localmods.items()):
707 706 for src, modname, name, line in sources(source_path, localmodname):
708 707 try:
709 708 used_imports[modname] = sorted(
710 709 imported_modules(src, modname, name, localmods,
711 710 ignore_nested=True))
712 711 for error, lineno in verify_import_convention(modname, src,
713 712 localmods):
714 713 any_errors = True
715 714 print('%s:%d: %s' % (source_path, lineno + line, error))
716 715 except SyntaxError as e:
717 716 print('%s:%d: SyntaxError: %s' %
718 717 (source_path, e.lineno + line, e))
719 718 cycles = find_cycles(used_imports)
720 719 if cycles:
721 720 firstmods = set()
722 721 for c in sorted(cycles, key=_cycle_sortkey):
723 722 first = c.split()[0]
724 723 # As a rough cut, ignore any cycle that starts with the
725 724 # same module as some other cycle. Otherwise we see lots
726 725 # of cycles that are effectively duplicates.
727 726 if first in firstmods:
728 727 continue
729 728 print('Import cycle:', c)
730 729 firstmods.add(first)
731 730 any_errors = True
732 731 return any_errors != 0
733 732
734 733 if __name__ == '__main__':
735 734 sys.exit(int(main(sys.argv)))
@@ -1,35 +1,35
1 1 <?xml version="1.0" encoding="utf-8"?>
2 2 <Wix xmlns="http://schemas.microsoft.com/wix/2006/wi">
3 3
4 4 <?include guids.wxi ?>
5 5 <?include defines.wxi ?>
6 6
7 7 <Fragment>
8 8 <DirectoryRef Id="INSTALLDIR" FileSource="$(var.SourceDir)">
9 9 <Component Id="distOutput" Guid="$(var.dist.guid)" Win64='$(var.IsX64)'>
10 10 <File Name="python27.dll" KeyPath="yes" />
11 11 </Component>
12 12 <Directory Id="libdir" Name="lib" FileSource="$(var.SourceDir)/lib">
13 13 <Component Id="libOutput" Guid="$(var.lib.guid)" Win64='$(var.IsX64)'>
14 14 <File Name="library.zip" KeyPath="yes" />
15 <File Name="mercurial.base85.pyd" />
15 <File Name="mercurial.cext.base85.pyd" />
16 16 <File Name="mercurial.bdiff.pyd" />
17 17 <File Name="mercurial.diffhelpers.pyd" />
18 18 <File Name="mercurial.mpatch.pyd" />
19 19 <File Name="mercurial.cext.osutil.pyd" />
20 20 <File Name="mercurial.parsers.pyd" />
21 21 <File Name="pyexpat.pyd" />
22 22 <File Name="bz2.pyd" />
23 23 <File Name="select.pyd" />
24 24 <File Name="unicodedata.pyd" />
25 25 <File Name="_ctypes.pyd" />
26 26 <File Name="_elementtree.pyd" />
27 27 <File Name="_hashlib.pyd" />
28 28 <File Name="_socket.pyd" />
29 29 <File Name="_ssl.pyd" />
30 30 </Component>
31 31 </Directory>
32 32 </DirectoryRef>
33 33 </Fragment>
34 34
35 35 </Wix>
@@ -1,403 +1,402
1 1 # __init__.py - Startup and module loading logic for Mercurial.
2 2 #
3 3 # Copyright 2015 Gregory Szorc <gregory.szorc@gmail.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import imp
11 11 import os
12 12 import sys
13 13 import zipimport
14 14
15 15 from . import (
16 16 policy
17 17 )
18 18
19 19 __all__ = []
20 20
21 21 modulepolicy = policy.policy
22 22
23 23 # Modules that have both Python and C implementations. See also the
24 24 # set of .py files under mercurial/pure/.
25 25 _dualmodules = {
26 'mercurial.base85',
27 26 'mercurial.bdiff',
28 27 'mercurial.diffhelpers',
29 28 'mercurial.mpatch',
30 29 'mercurial.parsers',
31 30 }
32 31
33 32 class hgimporter(object):
34 33 """Object that conforms to import hook interface defined in PEP-302."""
35 34 def find_module(self, name, path=None):
36 35 # We only care about modules that have both C and pure implementations.
37 36 if name in _dualmodules:
38 37 return self
39 38 return None
40 39
41 40 def load_module(self, name):
42 41 mod = sys.modules.get(name, None)
43 42 if mod:
44 43 return mod
45 44
46 45 mercurial = sys.modules['mercurial']
47 46
48 47 # The zip importer behaves sufficiently differently from the default
49 48 # importer to warrant its own code path.
50 49 loader = getattr(mercurial, '__loader__', None)
51 50 if isinstance(loader, zipimport.zipimporter):
52 51 def ziploader(*paths):
53 52 """Obtain a zipimporter for a directory under the main zip."""
54 53 path = os.path.join(loader.archive, *paths)
55 54 zl = sys.path_importer_cache.get(path)
56 55 if not zl:
57 56 zl = zipimport.zipimporter(path)
58 57 return zl
59 58
60 59 try:
61 60 if modulepolicy in policy.policynoc:
62 61 raise ImportError()
63 62
64 63 zl = ziploader('mercurial')
65 64 mod = zl.load_module(name)
66 65 # Unlike imp, ziploader doesn't expose module metadata that
67 66 # indicates the type of module. So just assume what we found
68 67 # is OK (even though it could be a pure Python module).
69 68 except ImportError:
70 69 if modulepolicy == b'c':
71 70 raise
72 71 zl = ziploader('mercurial', 'pure')
73 72 mod = zl.load_module(name)
74 73
75 74 sys.modules[name] = mod
76 75 return mod
77 76
78 77 # Unlike the default importer which searches special locations and
79 78 # sys.path, we only look in the directory where "mercurial" was
80 79 # imported from.
81 80
82 81 # imp.find_module doesn't support submodules (modules with ".").
83 82 # Instead you have to pass the parent package's __path__ attribute
84 83 # as the path argument.
85 84 stem = name.split('.')[-1]
86 85
87 86 try:
88 87 if modulepolicy in policy.policynoc:
89 88 raise ImportError()
90 89
91 90 modinfo = imp.find_module(stem, mercurial.__path__)
92 91
93 92 # The Mercurial installer used to copy files from
94 93 # mercurial/pure/*.py to mercurial/*.py. Therefore, it's possible
95 94 # for some installations to have .py files under mercurial/*.
96 95 # Loading Python modules when we expected C versions could result
97 96 # in a) poor performance b) loading a version from a previous
98 97 # Mercurial version, potentially leading to incompatibility. Either
99 98 # scenario is bad. So we verify that modules loaded from
100 99 # mercurial/* are C extensions. If the current policy allows the
101 100 # loading of .py modules, the module will be re-imported from
102 101 # mercurial/pure/* below.
103 102 if modinfo[2][2] != imp.C_EXTENSION:
104 103 raise ImportError('.py version of %s found where C '
105 104 'version should exist' % name)
106 105
107 106 except ImportError:
108 107 if modulepolicy == b'c':
109 108 raise
110 109
111 110 # Could not load the C extension and pure Python is allowed. So
112 111 # try to load them.
113 112 from . import pure
114 113 modinfo = imp.find_module(stem, pure.__path__)
115 114 if not modinfo:
116 115 raise ImportError('could not find mercurial module %s' %
117 116 name)
118 117
119 118 mod = imp.load_module(name, *modinfo)
120 119 sys.modules[name] = mod
121 120 return mod
122 121
123 122 # Python 3 uses a custom module loader that transforms source code between
124 123 # source file reading and compilation. This is done by registering a custom
125 124 # finder that changes the spec for Mercurial modules to use a custom loader.
126 125 if sys.version_info[0] >= 3:
127 126 from . import pure
128 127 import importlib
129 128 import io
130 129 import token
131 130 import tokenize
132 131
133 132 class hgpathentryfinder(importlib.abc.MetaPathFinder):
134 133 """A sys.meta_path finder that uses a custom module loader."""
135 134 def find_spec(self, fullname, path, target=None):
136 135 # Only handle Mercurial-related modules.
137 136 if not fullname.startswith(('mercurial.', 'hgext.', 'hgext3rd.')):
138 137 return None
139 138 # zstd is already dual-version clean, don't try and mangle it
140 139 if fullname.startswith('mercurial.zstd'):
141 140 return None
142 141
143 142 # This assumes Python 3 doesn't support loading C modules.
144 143 if fullname in _dualmodules:
145 144 stem = fullname.split('.')[-1]
146 145 fullname = 'mercurial.pure.%s' % stem
147 146 target = pure
148 147 assert len(path) == 1
149 148 path = [os.path.join(path[0], 'pure')]
150 149
151 150 # Try to find the module using other registered finders.
152 151 spec = None
153 152 for finder in sys.meta_path:
154 153 if finder == self:
155 154 continue
156 155
157 156 spec = finder.find_spec(fullname, path, target=target)
158 157 if spec:
159 158 break
160 159
161 160 # This is a Mercurial-related module but we couldn't find it
162 161 # using the previously-registered finders. This likely means
163 162 # the module doesn't exist.
164 163 if not spec:
165 164 return None
166 165
167 166 if (fullname.startswith('mercurial.pure.')
168 167 and fullname.replace('.pure.', '.') in _dualmodules):
169 168 spec.name = spec.name.replace('.pure.', '.')
170 169
171 170 # TODO need to support loaders from alternate specs, like zip
172 171 # loaders.
173 172 spec.loader = hgloader(spec.name, spec.origin)
174 173 return spec
175 174
176 175 def replacetokens(tokens, fullname):
177 176 """Transform a stream of tokens from raw to Python 3.
178 177
179 178 It is called by the custom module loading machinery to rewrite
180 179 source/tokens between source decoding and compilation.
181 180
182 181 Returns a generator of possibly rewritten tokens.
183 182
184 183 The input token list may be mutated as part of processing. However,
185 184 its changes do not necessarily match the output token stream.
186 185
187 186 REMEMBER TO CHANGE ``BYTECODEHEADER`` WHEN CHANGING THIS FUNCTION
188 187 OR CACHED FILES WON'T GET INVALIDATED PROPERLY.
189 188 """
190 189 futureimpline = False
191 190
192 191 # The following utility functions access the tokens list and i index of
193 192 # the for i, t enumerate(tokens) loop below
194 193 def _isop(j, *o):
195 194 """Assert that tokens[j] is an OP with one of the given values"""
196 195 try:
197 196 return tokens[j].type == token.OP and tokens[j].string in o
198 197 except IndexError:
199 198 return False
200 199
201 200 def _findargnofcall(n):
202 201 """Find arg n of a call expression (start at 0)
203 202
204 203 Returns index of the first token of that argument, or None if
205 204 there is not that many arguments.
206 205
207 206 Assumes that token[i + 1] is '('.
208 207
209 208 """
210 209 nested = 0
211 210 for j in range(i + 2, len(tokens)):
212 211 if _isop(j, ')', ']', '}'):
213 212 # end of call, tuple, subscription or dict / set
214 213 nested -= 1
215 214 if nested < 0:
216 215 return None
217 216 elif n == 0:
218 217 # this is the starting position of arg
219 218 return j
220 219 elif _isop(j, '(', '[', '{'):
221 220 nested += 1
222 221 elif _isop(j, ',') and nested == 0:
223 222 n -= 1
224 223
225 224 return None
226 225
227 226 def _ensureunicode(j):
228 227 """Make sure the token at j is a unicode string
229 228
230 229 This rewrites a string token to include the unicode literal prefix
231 230 so the string transformer won't add the byte prefix.
232 231
233 232 Ignores tokens that are not strings. Assumes bounds checking has
234 233 already been done.
235 234
236 235 """
237 236 st = tokens[j]
238 237 if st.type == token.STRING and st.string.startswith(("'", '"')):
239 238 tokens[j] = st._replace(string='u%s' % st.string)
240 239
241 240 for i, t in enumerate(tokens):
242 241 # Convert most string literals to byte literals. String literals
243 242 # in Python 2 are bytes. String literals in Python 3 are unicode.
244 243 # Most strings in Mercurial are bytes and unicode strings are rare.
245 244 # Rather than rewrite all string literals to use ``b''`` to indicate
246 245 # byte strings, we apply this token transformer to insert the ``b``
247 246 # prefix nearly everywhere.
248 247 if t.type == token.STRING:
249 248 s = t.string
250 249
251 250 # Preserve docstrings as string literals. This is inconsistent
252 251 # with regular unprefixed strings. However, the
253 252 # "from __future__" parsing (which allows a module docstring to
254 253 # exist before it) doesn't properly handle the docstring if it
255 254 # is b''' prefixed, leading to a SyntaxError. We leave all
256 255 # docstrings as unprefixed to avoid this. This means Mercurial
257 256 # components touching docstrings need to handle unicode,
258 257 # unfortunately.
259 258 if s[0:3] in ("'''", '"""'):
260 259 yield t
261 260 continue
262 261
263 262 # If the first character isn't a quote, it is likely a string
264 263 # prefixing character (such as 'b', 'u', or 'r'. Ignore.
265 264 if s[0] not in ("'", '"'):
266 265 yield t
267 266 continue
268 267
269 268 # String literal. Prefix to make a b'' string.
270 269 yield t._replace(string='b%s' % t.string)
271 270 continue
272 271
273 272 # Insert compatibility imports at "from __future__ import" line.
274 273 # No '\n' should be added to preserve line numbers.
275 274 if (t.type == token.NAME and t.string == 'import' and
276 275 all(u.type == token.NAME for u in tokens[i - 2:i]) and
277 276 [u.string for u in tokens[i - 2:i]] == ['from', '__future__']):
278 277 futureimpline = True
279 278 if t.type == token.NEWLINE and futureimpline:
280 279 futureimpline = False
281 280 if fullname == 'mercurial.pycompat':
282 281 yield t
283 282 continue
284 283 r, c = t.start
285 284 l = (b'; from mercurial.pycompat import '
286 285 b'delattr, getattr, hasattr, setattr, xrange, '
287 286 b'open, unicode\n')
288 287 for u in tokenize.tokenize(io.BytesIO(l).readline):
289 288 if u.type in (tokenize.ENCODING, token.ENDMARKER):
290 289 continue
291 290 yield u._replace(
292 291 start=(r, c + u.start[1]), end=(r, c + u.end[1]))
293 292 continue
294 293
295 294 # This looks like a function call.
296 295 if t.type == token.NAME and _isop(i + 1, '('):
297 296 fn = t.string
298 297
299 298 # *attr() builtins don't accept byte strings to 2nd argument.
300 299 if (fn in ('getattr', 'setattr', 'hasattr', 'safehasattr') and
301 300 not _isop(i - 1, '.')):
302 301 arg1idx = _findargnofcall(1)
303 302 if arg1idx is not None:
304 303 _ensureunicode(arg1idx)
305 304
306 305 # .encode() and .decode() on str/bytes/unicode don't accept
307 306 # byte strings on Python 3.
308 307 elif fn in ('encode', 'decode') and _isop(i - 1, '.'):
309 308 for argn in range(2):
310 309 argidx = _findargnofcall(argn)
311 310 if argidx is not None:
312 311 _ensureunicode(argidx)
313 312
314 313 # It changes iteritems/values to items/values as they are not
315 314 # present in Python 3 world.
316 315 elif fn in ('iteritems', 'itervalues'):
317 316 yield t._replace(string=fn[4:])
318 317 continue
319 318
320 319 # Emit unmodified token.
321 320 yield t
322 321
323 322 # Header to add to bytecode files. This MUST be changed when
324 323 # ``replacetoken`` or any mechanism that changes semantics of module
325 324 # loading is changed. Otherwise cached bytecode may get loaded without
326 325 # the new transformation mechanisms applied.
327 326 BYTECODEHEADER = b'HG\x00\x0a'
328 327
329 328 class hgloader(importlib.machinery.SourceFileLoader):
330 329 """Custom module loader that transforms source code.
331 330
332 331 When the source code is converted to a code object, we transform
333 332 certain patterns to be Python 3 compatible. This allows us to write code
334 333 that is natively Python 2 and compatible with Python 3 without
335 334 making the code excessively ugly.
336 335
337 336 We do this by transforming the token stream between parse and compile.
338 337
339 338 Implementing transformations invalidates caching assumptions made
340 339 by the built-in importer. The built-in importer stores a header on
341 340 saved bytecode files indicating the Python/bytecode version. If the
342 341 version changes, the cached bytecode is ignored. The Mercurial
343 342 transformations could change at any time. This means we need to check
344 343 that cached bytecode was generated with the current transformation
345 344 code or there could be a mismatch between cached bytecode and what
346 345 would be generated from this class.
347 346
348 347 We supplement the bytecode caching layer by wrapping ``get_data``
349 348 and ``set_data``. These functions are called when the
350 349 ``SourceFileLoader`` retrieves and saves bytecode cache files,
351 350 respectively. We simply add an additional header on the file. As
352 351 long as the version in this file is changed when semantics change,
353 352 cached bytecode should be invalidated when transformations change.
354 353
355 354 The added header has the form ``HG<VERSION>``. That is a literal
356 355 ``HG`` with 2 binary bytes indicating the transformation version.
357 356 """
358 357 def get_data(self, path):
359 358 data = super(hgloader, self).get_data(path)
360 359
361 360 if not path.endswith(tuple(importlib.machinery.BYTECODE_SUFFIXES)):
362 361 return data
363 362
364 363 # There should be a header indicating the Mercurial transformation
365 364 # version. If it doesn't exist or doesn't match the current version,
366 365 # we raise an OSError because that is what
367 366 # ``SourceFileLoader.get_code()`` expects when loading bytecode
368 367 # paths to indicate the cached file is "bad."
369 368 if data[0:2] != b'HG':
370 369 raise OSError('no hg header')
371 370 if data[0:4] != BYTECODEHEADER:
372 371 raise OSError('hg header version mismatch')
373 372
374 373 return data[4:]
375 374
376 375 def set_data(self, path, data, *args, **kwargs):
377 376 if path.endswith(tuple(importlib.machinery.BYTECODE_SUFFIXES)):
378 377 data = BYTECODEHEADER + data
379 378
380 379 return super(hgloader, self).set_data(path, data, *args, **kwargs)
381 380
382 381 def source_to_code(self, data, path):
383 382 """Perform token transformation before compilation."""
384 383 buf = io.BytesIO(data)
385 384 tokens = tokenize.tokenize(buf.readline)
386 385 data = tokenize.untokenize(replacetokens(list(tokens), self.name))
387 386 # Python's built-in importer strips frames from exceptions raised
388 387 # for this code. Unfortunately, that mechanism isn't extensible
389 388 # and our frame will be blamed for the import failure. There
390 389 # are extremely hacky ways to do frame stripping. We haven't
391 390 # implemented them because they are very ugly.
392 391 return super(hgloader, self).source_to_code(data, path)
393 392
394 393 # We automagically register our custom importer as a side-effect of loading.
395 394 # This is necessary to ensure that any entry points are able to import
396 395 # mercurial.* modules without having to perform this registration themselves.
397 396 if sys.version_info[0] >= 3:
398 397 _importercls = hgpathentryfinder
399 398 else:
400 399 _importercls = hgimporter
401 400 if not any(isinstance(x, _importercls) for x in sys.meta_path):
402 401 # meta_path is used before any implicit finders and before sys.path.
403 402 sys.meta_path.insert(0, _importercls())
1 NO CONTENT: file renamed from mercurial/base85.c to mercurial/cext/base85.c
@@ -1,2206 +1,2206
1 1 # debugcommands.py - command processing for debug* commands
2 2 #
3 3 # Copyright 2005-2016 Matt Mackall <mpm@selenic.com>
4 4 #
5 5 # This software may be used and distributed according to the terms of the
6 6 # GNU General Public License version 2 or any later version.
7 7
8 8 from __future__ import absolute_import
9 9
10 10 import difflib
11 11 import errno
12 12 import operator
13 13 import os
14 14 import random
15 15 import socket
16 16 import string
17 17 import sys
18 18 import tempfile
19 19 import time
20 20
21 21 from .i18n import _
22 22 from .node import (
23 23 bin,
24 24 hex,
25 25 nullhex,
26 26 nullid,
27 27 nullrev,
28 28 short,
29 29 )
30 30 from . import (
31 31 bundle2,
32 32 changegroup,
33 33 cmdutil,
34 34 color,
35 35 commands,
36 36 context,
37 37 dagparser,
38 38 dagutil,
39 39 encoding,
40 40 error,
41 41 exchange,
42 42 extensions,
43 43 filemerge,
44 44 fileset,
45 45 formatter,
46 46 hg,
47 47 localrepo,
48 48 lock as lockmod,
49 49 merge as mergemod,
50 50 obsolete,
51 51 policy,
52 52 pvec,
53 53 pycompat,
54 54 registrar,
55 55 repair,
56 56 revlog,
57 57 revset,
58 58 revsetlang,
59 59 scmutil,
60 60 setdiscovery,
61 61 simplemerge,
62 62 smartset,
63 63 sslutil,
64 64 streamclone,
65 65 templater,
66 66 treediscovery,
67 67 upgrade,
68 68 util,
69 69 vfs as vfsmod,
70 70 )
71 71
72 72 release = lockmod.release
73 73
74 74 # We reuse the command table from commands because it is easier than
75 75 # teaching dispatch about multiple tables.
76 76 command = registrar.command(commands.table)
77 77
78 78 @command('debugancestor', [], _('[INDEX] REV1 REV2'), optionalrepo=True)
79 79 def debugancestor(ui, repo, *args):
80 80 """find the ancestor revision of two revisions in a given index"""
81 81 if len(args) == 3:
82 82 index, rev1, rev2 = args
83 83 r = revlog.revlog(vfsmod.vfs(pycompat.getcwd(), audit=False), index)
84 84 lookup = r.lookup
85 85 elif len(args) == 2:
86 86 if not repo:
87 87 raise error.Abort(_('there is no Mercurial repository here '
88 88 '(.hg not found)'))
89 89 rev1, rev2 = args
90 90 r = repo.changelog
91 91 lookup = repo.lookup
92 92 else:
93 93 raise error.Abort(_('either two or three arguments required'))
94 94 a = r.ancestor(lookup(rev1), lookup(rev2))
95 95 ui.write('%d:%s\n' % (r.rev(a), hex(a)))
96 96
97 97 @command('debugapplystreamclonebundle', [], 'FILE')
98 98 def debugapplystreamclonebundle(ui, repo, fname):
99 99 """apply a stream clone bundle file"""
100 100 f = hg.openpath(ui, fname)
101 101 gen = exchange.readbundle(ui, f, fname)
102 102 gen.apply(repo)
103 103
104 104 @command('debugbuilddag',
105 105 [('m', 'mergeable-file', None, _('add single file mergeable changes')),
106 106 ('o', 'overwritten-file', None, _('add single file all revs overwrite')),
107 107 ('n', 'new-file', None, _('add new file at each rev'))],
108 108 _('[OPTION]... [TEXT]'))
109 109 def debugbuilddag(ui, repo, text=None,
110 110 mergeable_file=False,
111 111 overwritten_file=False,
112 112 new_file=False):
113 113 """builds a repo with a given DAG from scratch in the current empty repo
114 114
115 115 The description of the DAG is read from stdin if not given on the
116 116 command line.
117 117
118 118 Elements:
119 119
120 120 - "+n" is a linear run of n nodes based on the current default parent
121 121 - "." is a single node based on the current default parent
122 122 - "$" resets the default parent to null (implied at the start);
123 123 otherwise the default parent is always the last node created
124 124 - "<p" sets the default parent to the backref p
125 125 - "*p" is a fork at parent p, which is a backref
126 126 - "*p1/p2" is a merge of parents p1 and p2, which are backrefs
127 127 - "/p2" is a merge of the preceding node and p2
128 128 - ":tag" defines a local tag for the preceding node
129 129 - "@branch" sets the named branch for subsequent nodes
130 130 - "#...\\n" is a comment up to the end of the line
131 131
132 132 Whitespace between the above elements is ignored.
133 133
134 134 A backref is either
135 135
136 136 - a number n, which references the node curr-n, where curr is the current
137 137 node, or
138 138 - the name of a local tag you placed earlier using ":tag", or
139 139 - empty to denote the default parent.
140 140
141 141 All string valued-elements are either strictly alphanumeric, or must
142 142 be enclosed in double quotes ("..."), with "\\" as escape character.
143 143 """
144 144
145 145 if text is None:
146 146 ui.status(_("reading DAG from stdin\n"))
147 147 text = ui.fin.read()
148 148
149 149 cl = repo.changelog
150 150 if len(cl) > 0:
151 151 raise error.Abort(_('repository is not empty'))
152 152
153 153 # determine number of revs in DAG
154 154 total = 0
155 155 for type, data in dagparser.parsedag(text):
156 156 if type == 'n':
157 157 total += 1
158 158
159 159 if mergeable_file:
160 160 linesperrev = 2
161 161 # make a file with k lines per rev
162 162 initialmergedlines = [str(i) for i in xrange(0, total * linesperrev)]
163 163 initialmergedlines.append("")
164 164
165 165 tags = []
166 166
167 167 wlock = lock = tr = None
168 168 try:
169 169 wlock = repo.wlock()
170 170 lock = repo.lock()
171 171 tr = repo.transaction("builddag")
172 172
173 173 at = -1
174 174 atbranch = 'default'
175 175 nodeids = []
176 176 id = 0
177 177 ui.progress(_('building'), id, unit=_('revisions'), total=total)
178 178 for type, data in dagparser.parsedag(text):
179 179 if type == 'n':
180 180 ui.note(('node %s\n' % str(data)))
181 181 id, ps = data
182 182
183 183 files = []
184 184 fctxs = {}
185 185
186 186 p2 = None
187 187 if mergeable_file:
188 188 fn = "mf"
189 189 p1 = repo[ps[0]]
190 190 if len(ps) > 1:
191 191 p2 = repo[ps[1]]
192 192 pa = p1.ancestor(p2)
193 193 base, local, other = [x[fn].data() for x in (pa, p1,
194 194 p2)]
195 195 m3 = simplemerge.Merge3Text(base, local, other)
196 196 ml = [l.strip() for l in m3.merge_lines()]
197 197 ml.append("")
198 198 elif at > 0:
199 199 ml = p1[fn].data().split("\n")
200 200 else:
201 201 ml = initialmergedlines
202 202 ml[id * linesperrev] += " r%i" % id
203 203 mergedtext = "\n".join(ml)
204 204 files.append(fn)
205 205 fctxs[fn] = context.memfilectx(repo, fn, mergedtext)
206 206
207 207 if overwritten_file:
208 208 fn = "of"
209 209 files.append(fn)
210 210 fctxs[fn] = context.memfilectx(repo, fn, "r%i\n" % id)
211 211
212 212 if new_file:
213 213 fn = "nf%i" % id
214 214 files.append(fn)
215 215 fctxs[fn] = context.memfilectx(repo, fn, "r%i\n" % id)
216 216 if len(ps) > 1:
217 217 if not p2:
218 218 p2 = repo[ps[1]]
219 219 for fn in p2:
220 220 if fn.startswith("nf"):
221 221 files.append(fn)
222 222 fctxs[fn] = p2[fn]
223 223
224 224 def fctxfn(repo, cx, path):
225 225 return fctxs.get(path)
226 226
227 227 if len(ps) == 0 or ps[0] < 0:
228 228 pars = [None, None]
229 229 elif len(ps) == 1:
230 230 pars = [nodeids[ps[0]], None]
231 231 else:
232 232 pars = [nodeids[p] for p in ps]
233 233 cx = context.memctx(repo, pars, "r%i" % id, files, fctxfn,
234 234 date=(id, 0),
235 235 user="debugbuilddag",
236 236 extra={'branch': atbranch})
237 237 nodeid = repo.commitctx(cx)
238 238 nodeids.append(nodeid)
239 239 at = id
240 240 elif type == 'l':
241 241 id, name = data
242 242 ui.note(('tag %s\n' % name))
243 243 tags.append("%s %s\n" % (hex(repo.changelog.node(id)), name))
244 244 elif type == 'a':
245 245 ui.note(('branch %s\n' % data))
246 246 atbranch = data
247 247 ui.progress(_('building'), id, unit=_('revisions'), total=total)
248 248 tr.close()
249 249
250 250 if tags:
251 251 repo.vfs.write("localtags", "".join(tags))
252 252 finally:
253 253 ui.progress(_('building'), None)
254 254 release(tr, lock, wlock)
255 255
256 256 def _debugchangegroup(ui, gen, all=None, indent=0, **opts):
257 257 indent_string = ' ' * indent
258 258 if all:
259 259 ui.write(("%sformat: id, p1, p2, cset, delta base, len(delta)\n")
260 260 % indent_string)
261 261
262 262 def showchunks(named):
263 263 ui.write("\n%s%s\n" % (indent_string, named))
264 264 chain = None
265 265 for chunkdata in iter(lambda: gen.deltachunk(chain), {}):
266 266 node = chunkdata['node']
267 267 p1 = chunkdata['p1']
268 268 p2 = chunkdata['p2']
269 269 cs = chunkdata['cs']
270 270 deltabase = chunkdata['deltabase']
271 271 delta = chunkdata['delta']
272 272 ui.write("%s%s %s %s %s %s %s\n" %
273 273 (indent_string, hex(node), hex(p1), hex(p2),
274 274 hex(cs), hex(deltabase), len(delta)))
275 275 chain = node
276 276
277 277 chunkdata = gen.changelogheader()
278 278 showchunks("changelog")
279 279 chunkdata = gen.manifestheader()
280 280 showchunks("manifest")
281 281 for chunkdata in iter(gen.filelogheader, {}):
282 282 fname = chunkdata['filename']
283 283 showchunks(fname)
284 284 else:
285 285 if isinstance(gen, bundle2.unbundle20):
286 286 raise error.Abort(_('use debugbundle2 for this file'))
287 287 chunkdata = gen.changelogheader()
288 288 chain = None
289 289 for chunkdata in iter(lambda: gen.deltachunk(chain), {}):
290 290 node = chunkdata['node']
291 291 ui.write("%s%s\n" % (indent_string, hex(node)))
292 292 chain = node
293 293
294 294 def _debugbundle2(ui, gen, all=None, **opts):
295 295 """lists the contents of a bundle2"""
296 296 if not isinstance(gen, bundle2.unbundle20):
297 297 raise error.Abort(_('not a bundle2 file'))
298 298 ui.write(('Stream params: %s\n' % repr(gen.params)))
299 299 for part in gen.iterparts():
300 300 ui.write('%s -- %r\n' % (part.type, repr(part.params)))
301 301 if part.type == 'changegroup':
302 302 version = part.params.get('version', '01')
303 303 cg = changegroup.getunbundler(version, part, 'UN')
304 304 _debugchangegroup(ui, cg, all=all, indent=4, **opts)
305 305
306 306 @command('debugbundle',
307 307 [('a', 'all', None, _('show all details')),
308 308 ('', 'spec', None, _('print the bundlespec of the bundle'))],
309 309 _('FILE'),
310 310 norepo=True)
311 311 def debugbundle(ui, bundlepath, all=None, spec=None, **opts):
312 312 """lists the contents of a bundle"""
313 313 with hg.openpath(ui, bundlepath) as f:
314 314 if spec:
315 315 spec = exchange.getbundlespec(ui, f)
316 316 ui.write('%s\n' % spec)
317 317 return
318 318
319 319 gen = exchange.readbundle(ui, f, bundlepath)
320 320 if isinstance(gen, bundle2.unbundle20):
321 321 return _debugbundle2(ui, gen, all=all, **opts)
322 322 _debugchangegroup(ui, gen, all=all, **opts)
323 323
324 324 @command('debugcheckstate', [], '')
325 325 def debugcheckstate(ui, repo):
326 326 """validate the correctness of the current dirstate"""
327 327 parent1, parent2 = repo.dirstate.parents()
328 328 m1 = repo[parent1].manifest()
329 329 m2 = repo[parent2].manifest()
330 330 errors = 0
331 331 for f in repo.dirstate:
332 332 state = repo.dirstate[f]
333 333 if state in "nr" and f not in m1:
334 334 ui.warn(_("%s in state %s, but not in manifest1\n") % (f, state))
335 335 errors += 1
336 336 if state in "a" and f in m1:
337 337 ui.warn(_("%s in state %s, but also in manifest1\n") % (f, state))
338 338 errors += 1
339 339 if state in "m" and f not in m1 and f not in m2:
340 340 ui.warn(_("%s in state %s, but not in either manifest\n") %
341 341 (f, state))
342 342 errors += 1
343 343 for f in m1:
344 344 state = repo.dirstate[f]
345 345 if state not in "nrm":
346 346 ui.warn(_("%s in manifest1, but listed as state %s") % (f, state))
347 347 errors += 1
348 348 if errors:
349 349 error = _(".hg/dirstate inconsistent with current parent's manifest")
350 350 raise error.Abort(error)
351 351
352 352 @command('debugcolor',
353 353 [('', 'style', None, _('show all configured styles'))],
354 354 'hg debugcolor')
355 355 def debugcolor(ui, repo, **opts):
356 356 """show available color, effects or style"""
357 357 ui.write(('color mode: %s\n') % ui._colormode)
358 358 if opts.get('style'):
359 359 return _debugdisplaystyle(ui)
360 360 else:
361 361 return _debugdisplaycolor(ui)
362 362
363 363 def _debugdisplaycolor(ui):
364 364 ui = ui.copy()
365 365 ui._styles.clear()
366 366 for effect in color._activeeffects(ui).keys():
367 367 ui._styles[effect] = effect
368 368 if ui._terminfoparams:
369 369 for k, v in ui.configitems('color'):
370 370 if k.startswith('color.'):
371 371 ui._styles[k] = k[6:]
372 372 elif k.startswith('terminfo.'):
373 373 ui._styles[k] = k[9:]
374 374 ui.write(_('available colors:\n'))
375 375 # sort label with a '_' after the other to group '_background' entry.
376 376 items = sorted(ui._styles.items(),
377 377 key=lambda i: ('_' in i[0], i[0], i[1]))
378 378 for colorname, label in items:
379 379 ui.write(('%s\n') % colorname, label=label)
380 380
381 381 def _debugdisplaystyle(ui):
382 382 ui.write(_('available style:\n'))
383 383 width = max(len(s) for s in ui._styles)
384 384 for label, effects in sorted(ui._styles.items()):
385 385 ui.write('%s' % label, label=label)
386 386 if effects:
387 387 # 50
388 388 ui.write(': ')
389 389 ui.write(' ' * (max(0, width - len(label))))
390 390 ui.write(', '.join(ui.label(e, e) for e in effects.split()))
391 391 ui.write('\n')
392 392
393 393 @command('debugcommands', [], _('[COMMAND]'), norepo=True)
394 394 def debugcommands(ui, cmd='', *args):
395 395 """list all available commands and options"""
396 396 for cmd, vals in sorted(commands.table.iteritems()):
397 397 cmd = cmd.split('|')[0].strip('^')
398 398 opts = ', '.join([i[1] for i in vals[1]])
399 399 ui.write('%s: %s\n' % (cmd, opts))
400 400
401 401 @command('debugcomplete',
402 402 [('o', 'options', None, _('show the command options'))],
403 403 _('[-o] CMD'),
404 404 norepo=True)
405 405 def debugcomplete(ui, cmd='', **opts):
406 406 """returns the completion list associated with the given command"""
407 407
408 408 if opts.get('options'):
409 409 options = []
410 410 otables = [commands.globalopts]
411 411 if cmd:
412 412 aliases, entry = cmdutil.findcmd(cmd, commands.table, False)
413 413 otables.append(entry[1])
414 414 for t in otables:
415 415 for o in t:
416 416 if "(DEPRECATED)" in o[3]:
417 417 continue
418 418 if o[0]:
419 419 options.append('-%s' % o[0])
420 420 options.append('--%s' % o[1])
421 421 ui.write("%s\n" % "\n".join(options))
422 422 return
423 423
424 424 cmdlist, unused_allcmds = cmdutil.findpossible(cmd, commands.table)
425 425 if ui.verbose:
426 426 cmdlist = [' '.join(c[0]) for c in cmdlist.values()]
427 427 ui.write("%s\n" % "\n".join(sorted(cmdlist)))
428 428
429 429 @command('debugcreatestreamclonebundle', [], 'FILE')
430 430 def debugcreatestreamclonebundle(ui, repo, fname):
431 431 """create a stream clone bundle file
432 432
433 433 Stream bundles are special bundles that are essentially archives of
434 434 revlog files. They are commonly used for cloning very quickly.
435 435 """
436 436 requirements, gen = streamclone.generatebundlev1(repo)
437 437 changegroup.writechunks(ui, gen, fname)
438 438
439 439 ui.write(_('bundle requirements: %s\n') % ', '.join(sorted(requirements)))
440 440
441 441 @command('debugdag',
442 442 [('t', 'tags', None, _('use tags as labels')),
443 443 ('b', 'branches', None, _('annotate with branch names')),
444 444 ('', 'dots', None, _('use dots for runs')),
445 445 ('s', 'spaces', None, _('separate elements by spaces'))],
446 446 _('[OPTION]... [FILE [REV]...]'),
447 447 optionalrepo=True)
448 448 def debugdag(ui, repo, file_=None, *revs, **opts):
449 449 """format the changelog or an index DAG as a concise textual description
450 450
451 451 If you pass a revlog index, the revlog's DAG is emitted. If you list
452 452 revision numbers, they get labeled in the output as rN.
453 453
454 454 Otherwise, the changelog DAG of the current repo is emitted.
455 455 """
456 456 spaces = opts.get('spaces')
457 457 dots = opts.get('dots')
458 458 if file_:
459 459 rlog = revlog.revlog(vfsmod.vfs(pycompat.getcwd(), audit=False),
460 460 file_)
461 461 revs = set((int(r) for r in revs))
462 462 def events():
463 463 for r in rlog:
464 464 yield 'n', (r, list(p for p in rlog.parentrevs(r)
465 465 if p != -1))
466 466 if r in revs:
467 467 yield 'l', (r, "r%i" % r)
468 468 elif repo:
469 469 cl = repo.changelog
470 470 tags = opts.get('tags')
471 471 branches = opts.get('branches')
472 472 if tags:
473 473 labels = {}
474 474 for l, n in repo.tags().items():
475 475 labels.setdefault(cl.rev(n), []).append(l)
476 476 def events():
477 477 b = "default"
478 478 for r in cl:
479 479 if branches:
480 480 newb = cl.read(cl.node(r))[5]['branch']
481 481 if newb != b:
482 482 yield 'a', newb
483 483 b = newb
484 484 yield 'n', (r, list(p for p in cl.parentrevs(r)
485 485 if p != -1))
486 486 if tags:
487 487 ls = labels.get(r)
488 488 if ls:
489 489 for l in ls:
490 490 yield 'l', (r, l)
491 491 else:
492 492 raise error.Abort(_('need repo for changelog dag'))
493 493
494 494 for line in dagparser.dagtextlines(events(),
495 495 addspaces=spaces,
496 496 wraplabels=True,
497 497 wrapannotations=True,
498 498 wrapnonlinear=dots,
499 499 usedots=dots,
500 500 maxlinewidth=70):
501 501 ui.write(line)
502 502 ui.write("\n")
503 503
504 504 @command('debugdata', commands.debugrevlogopts, _('-c|-m|FILE REV'))
505 505 def debugdata(ui, repo, file_, rev=None, **opts):
506 506 """dump the contents of a data file revision"""
507 507 if opts.get('changelog') or opts.get('manifest') or opts.get('dir'):
508 508 if rev is not None:
509 509 raise error.CommandError('debugdata', _('invalid arguments'))
510 510 file_, rev = None, file_
511 511 elif rev is None:
512 512 raise error.CommandError('debugdata', _('invalid arguments'))
513 513 r = cmdutil.openrevlog(repo, 'debugdata', file_, opts)
514 514 try:
515 515 ui.write(r.revision(r.lookup(rev), raw=True))
516 516 except KeyError:
517 517 raise error.Abort(_('invalid revision identifier %s') % rev)
518 518
519 519 @command('debugdate',
520 520 [('e', 'extended', None, _('try extended date formats'))],
521 521 _('[-e] DATE [RANGE]'),
522 522 norepo=True, optionalrepo=True)
523 523 def debugdate(ui, date, range=None, **opts):
524 524 """parse and display a date"""
525 525 if opts["extended"]:
526 526 d = util.parsedate(date, util.extendeddateformats)
527 527 else:
528 528 d = util.parsedate(date)
529 529 ui.write(("internal: %s %s\n") % d)
530 530 ui.write(("standard: %s\n") % util.datestr(d))
531 531 if range:
532 532 m = util.matchdate(range)
533 533 ui.write(("match: %s\n") % m(d[0]))
534 534
535 535 @command('debugdeltachain',
536 536 commands.debugrevlogopts + commands.formatteropts,
537 537 _('-c|-m|FILE'),
538 538 optionalrepo=True)
539 539 def debugdeltachain(ui, repo, file_=None, **opts):
540 540 """dump information about delta chains in a revlog
541 541
542 542 Output can be templatized. Available template keywords are:
543 543
544 544 :``rev``: revision number
545 545 :``chainid``: delta chain identifier (numbered by unique base)
546 546 :``chainlen``: delta chain length to this revision
547 547 :``prevrev``: previous revision in delta chain
548 548 :``deltatype``: role of delta / how it was computed
549 549 :``compsize``: compressed size of revision
550 550 :``uncompsize``: uncompressed size of revision
551 551 :``chainsize``: total size of compressed revisions in chain
552 552 :``chainratio``: total chain size divided by uncompressed revision size
553 553 (new delta chains typically start at ratio 2.00)
554 554 :``lindist``: linear distance from base revision in delta chain to end
555 555 of this revision
556 556 :``extradist``: total size of revisions not part of this delta chain from
557 557 base of delta chain to end of this revision; a measurement
558 558 of how much extra data we need to read/seek across to read
559 559 the delta chain for this revision
560 560 :``extraratio``: extradist divided by chainsize; another representation of
561 561 how much unrelated data is needed to load this delta chain
562 562 """
563 563 r = cmdutil.openrevlog(repo, 'debugdeltachain', file_, opts)
564 564 index = r.index
565 565 generaldelta = r.version & revlog.FLAG_GENERALDELTA
566 566
567 567 def revinfo(rev):
568 568 e = index[rev]
569 569 compsize = e[1]
570 570 uncompsize = e[2]
571 571 chainsize = 0
572 572
573 573 if generaldelta:
574 574 if e[3] == e[5]:
575 575 deltatype = 'p1'
576 576 elif e[3] == e[6]:
577 577 deltatype = 'p2'
578 578 elif e[3] == rev - 1:
579 579 deltatype = 'prev'
580 580 elif e[3] == rev:
581 581 deltatype = 'base'
582 582 else:
583 583 deltatype = 'other'
584 584 else:
585 585 if e[3] == rev:
586 586 deltatype = 'base'
587 587 else:
588 588 deltatype = 'prev'
589 589
590 590 chain = r._deltachain(rev)[0]
591 591 for iterrev in chain:
592 592 e = index[iterrev]
593 593 chainsize += e[1]
594 594
595 595 return compsize, uncompsize, deltatype, chain, chainsize
596 596
597 597 fm = ui.formatter('debugdeltachain', opts)
598 598
599 599 fm.plain(' rev chain# chainlen prev delta '
600 600 'size rawsize chainsize ratio lindist extradist '
601 601 'extraratio\n')
602 602
603 603 chainbases = {}
604 604 for rev in r:
605 605 comp, uncomp, deltatype, chain, chainsize = revinfo(rev)
606 606 chainbase = chain[0]
607 607 chainid = chainbases.setdefault(chainbase, len(chainbases) + 1)
608 608 basestart = r.start(chainbase)
609 609 revstart = r.start(rev)
610 610 lineardist = revstart + comp - basestart
611 611 extradist = lineardist - chainsize
612 612 try:
613 613 prevrev = chain[-2]
614 614 except IndexError:
615 615 prevrev = -1
616 616
617 617 chainratio = float(chainsize) / float(uncomp)
618 618 extraratio = float(extradist) / float(chainsize)
619 619
620 620 fm.startitem()
621 621 fm.write('rev chainid chainlen prevrev deltatype compsize '
622 622 'uncompsize chainsize chainratio lindist extradist '
623 623 'extraratio',
624 624 '%7d %7d %8d %8d %7s %10d %10d %10d %9.5f %9d %9d %10.5f\n',
625 625 rev, chainid, len(chain), prevrev, deltatype, comp,
626 626 uncomp, chainsize, chainratio, lineardist, extradist,
627 627 extraratio,
628 628 rev=rev, chainid=chainid, chainlen=len(chain),
629 629 prevrev=prevrev, deltatype=deltatype, compsize=comp,
630 630 uncompsize=uncomp, chainsize=chainsize,
631 631 chainratio=chainratio, lindist=lineardist,
632 632 extradist=extradist, extraratio=extraratio)
633 633
634 634 fm.end()
635 635
636 636 @command('debugdirstate|debugstate',
637 637 [('', 'nodates', None, _('do not display the saved mtime')),
638 638 ('', 'datesort', None, _('sort by saved mtime'))],
639 639 _('[OPTION]...'))
640 640 def debugstate(ui, repo, **opts):
641 641 """show the contents of the current dirstate"""
642 642
643 643 nodates = opts.get('nodates')
644 644 datesort = opts.get('datesort')
645 645
646 646 timestr = ""
647 647 if datesort:
648 648 keyfunc = lambda x: (x[1][3], x[0]) # sort by mtime, then by filename
649 649 else:
650 650 keyfunc = None # sort by filename
651 651 for file_, ent in sorted(repo.dirstate._map.iteritems(), key=keyfunc):
652 652 if ent[3] == -1:
653 653 timestr = 'unset '
654 654 elif nodates:
655 655 timestr = 'set '
656 656 else:
657 657 timestr = time.strftime("%Y-%m-%d %H:%M:%S ",
658 658 time.localtime(ent[3]))
659 659 if ent[1] & 0o20000:
660 660 mode = 'lnk'
661 661 else:
662 662 mode = '%3o' % (ent[1] & 0o777 & ~util.umask)
663 663 ui.write("%c %s %10d %s%s\n" % (ent[0], mode, ent[2], timestr, file_))
664 664 for f in repo.dirstate.copies():
665 665 ui.write(_("copy: %s -> %s\n") % (repo.dirstate.copied(f), f))
666 666
667 667 @command('debugdiscovery',
668 668 [('', 'old', None, _('use old-style discovery')),
669 669 ('', 'nonheads', None,
670 670 _('use old-style discovery with non-heads included')),
671 671 ] + commands.remoteopts,
672 672 _('[-l REV] [-r REV] [-b BRANCH]... [OTHER]'))
673 673 def debugdiscovery(ui, repo, remoteurl="default", **opts):
674 674 """runs the changeset discovery protocol in isolation"""
675 675 remoteurl, branches = hg.parseurl(ui.expandpath(remoteurl),
676 676 opts.get('branch'))
677 677 remote = hg.peer(repo, opts, remoteurl)
678 678 ui.status(_('comparing with %s\n') % util.hidepassword(remoteurl))
679 679
680 680 # make sure tests are repeatable
681 681 random.seed(12323)
682 682
683 683 def doit(localheads, remoteheads, remote=remote):
684 684 if opts.get('old'):
685 685 if localheads:
686 686 raise error.Abort('cannot use localheads with old style '
687 687 'discovery')
688 688 if not util.safehasattr(remote, 'branches'):
689 689 # enable in-client legacy support
690 690 remote = localrepo.locallegacypeer(remote.local())
691 691 common, _in, hds = treediscovery.findcommonincoming(repo, remote,
692 692 force=True)
693 693 common = set(common)
694 694 if not opts.get('nonheads'):
695 695 ui.write(("unpruned common: %s\n") %
696 696 " ".join(sorted(short(n) for n in common)))
697 697 dag = dagutil.revlogdag(repo.changelog)
698 698 all = dag.ancestorset(dag.internalizeall(common))
699 699 common = dag.externalizeall(dag.headsetofconnecteds(all))
700 700 else:
701 701 common, any, hds = setdiscovery.findcommonheads(ui, repo, remote)
702 702 common = set(common)
703 703 rheads = set(hds)
704 704 lheads = set(repo.heads())
705 705 ui.write(("common heads: %s\n") %
706 706 " ".join(sorted(short(n) for n in common)))
707 707 if lheads <= common:
708 708 ui.write(("local is subset\n"))
709 709 elif rheads <= common:
710 710 ui.write(("remote is subset\n"))
711 711
712 712 serverlogs = opts.get('serverlog')
713 713 if serverlogs:
714 714 for filename in serverlogs:
715 715 with open(filename, 'r') as logfile:
716 716 line = logfile.readline()
717 717 while line:
718 718 parts = line.strip().split(';')
719 719 op = parts[1]
720 720 if op == 'cg':
721 721 pass
722 722 elif op == 'cgss':
723 723 doit(parts[2].split(' '), parts[3].split(' '))
724 724 elif op == 'unb':
725 725 doit(parts[3].split(' '), parts[2].split(' '))
726 726 line = logfile.readline()
727 727 else:
728 728 remoterevs, _checkout = hg.addbranchrevs(repo, remote, branches,
729 729 opts.get('remote_head'))
730 730 localrevs = opts.get('local_head')
731 731 doit(localrevs, remoterevs)
732 732
733 733 @command('debugextensions', commands.formatteropts, [], norepo=True)
734 734 def debugextensions(ui, **opts):
735 735 '''show information about active extensions'''
736 736 exts = extensions.extensions(ui)
737 737 hgver = util.version()
738 738 fm = ui.formatter('debugextensions', opts)
739 739 for extname, extmod in sorted(exts, key=operator.itemgetter(0)):
740 740 isinternal = extensions.ismoduleinternal(extmod)
741 741 extsource = pycompat.fsencode(extmod.__file__)
742 742 if isinternal:
743 743 exttestedwith = [] # never expose magic string to users
744 744 else:
745 745 exttestedwith = getattr(extmod, 'testedwith', '').split()
746 746 extbuglink = getattr(extmod, 'buglink', None)
747 747
748 748 fm.startitem()
749 749
750 750 if ui.quiet or ui.verbose:
751 751 fm.write('name', '%s\n', extname)
752 752 else:
753 753 fm.write('name', '%s', extname)
754 754 if isinternal or hgver in exttestedwith:
755 755 fm.plain('\n')
756 756 elif not exttestedwith:
757 757 fm.plain(_(' (untested!)\n'))
758 758 else:
759 759 lasttestedversion = exttestedwith[-1]
760 760 fm.plain(' (%s!)\n' % lasttestedversion)
761 761
762 762 fm.condwrite(ui.verbose and extsource, 'source',
763 763 _(' location: %s\n'), extsource or "")
764 764
765 765 if ui.verbose:
766 766 fm.plain(_(' bundled: %s\n') % ['no', 'yes'][isinternal])
767 767 fm.data(bundled=isinternal)
768 768
769 769 fm.condwrite(ui.verbose and exttestedwith, 'testedwith',
770 770 _(' tested with: %s\n'),
771 771 fm.formatlist(exttestedwith, name='ver'))
772 772
773 773 fm.condwrite(ui.verbose and extbuglink, 'buglink',
774 774 _(' bug reporting: %s\n'), extbuglink or "")
775 775
776 776 fm.end()
777 777
778 778 @command('debugfileset',
779 779 [('r', 'rev', '', _('apply the filespec on this revision'), _('REV'))],
780 780 _('[-r REV] FILESPEC'))
781 781 def debugfileset(ui, repo, expr, **opts):
782 782 '''parse and apply a fileset specification'''
783 783 ctx = scmutil.revsingle(repo, opts.get('rev'), None)
784 784 if ui.verbose:
785 785 tree = fileset.parse(expr)
786 786 ui.note(fileset.prettyformat(tree), "\n")
787 787
788 788 for f in ctx.getfileset(expr):
789 789 ui.write("%s\n" % f)
790 790
791 791 @command('debugfsinfo', [], _('[PATH]'), norepo=True)
792 792 def debugfsinfo(ui, path="."):
793 793 """show information detected about current filesystem"""
794 794 ui.write(('exec: %s\n') % (util.checkexec(path) and 'yes' or 'no'))
795 795 ui.write(('fstype: %s\n') % (util.getfstype(path) or '(unknown)'))
796 796 ui.write(('symlink: %s\n') % (util.checklink(path) and 'yes' or 'no'))
797 797 ui.write(('hardlink: %s\n') % (util.checknlink(path) and 'yes' or 'no'))
798 798 casesensitive = '(unknown)'
799 799 try:
800 800 with tempfile.NamedTemporaryFile(prefix='.debugfsinfo', dir=path) as f:
801 801 casesensitive = util.fscasesensitive(f.name) and 'yes' or 'no'
802 802 except OSError:
803 803 pass
804 804 ui.write(('case-sensitive: %s\n') % casesensitive)
805 805
806 806 @command('debuggetbundle',
807 807 [('H', 'head', [], _('id of head node'), _('ID')),
808 808 ('C', 'common', [], _('id of common node'), _('ID')),
809 809 ('t', 'type', 'bzip2', _('bundle compression type to use'), _('TYPE'))],
810 810 _('REPO FILE [-H|-C ID]...'),
811 811 norepo=True)
812 812 def debuggetbundle(ui, repopath, bundlepath, head=None, common=None, **opts):
813 813 """retrieves a bundle from a repo
814 814
815 815 Every ID must be a full-length hex node id string. Saves the bundle to the
816 816 given file.
817 817 """
818 818 repo = hg.peer(ui, opts, repopath)
819 819 if not repo.capable('getbundle'):
820 820 raise error.Abort("getbundle() not supported by target repository")
821 821 args = {}
822 822 if common:
823 823 args['common'] = [bin(s) for s in common]
824 824 if head:
825 825 args['heads'] = [bin(s) for s in head]
826 826 # TODO: get desired bundlecaps from command line.
827 827 args['bundlecaps'] = None
828 828 bundle = repo.getbundle('debug', **args)
829 829
830 830 bundletype = opts.get('type', 'bzip2').lower()
831 831 btypes = {'none': 'HG10UN',
832 832 'bzip2': 'HG10BZ',
833 833 'gzip': 'HG10GZ',
834 834 'bundle2': 'HG20'}
835 835 bundletype = btypes.get(bundletype)
836 836 if bundletype not in bundle2.bundletypes:
837 837 raise error.Abort(_('unknown bundle type specified with --type'))
838 838 bundle2.writebundle(ui, bundle, bundlepath, bundletype)
839 839
840 840 @command('debugignore', [], '[FILE]')
841 841 def debugignore(ui, repo, *files, **opts):
842 842 """display the combined ignore pattern and information about ignored files
843 843
844 844 With no argument display the combined ignore pattern.
845 845
846 846 Given space separated file names, shows if the given file is ignored and
847 847 if so, show the ignore rule (file and line number) that matched it.
848 848 """
849 849 ignore = repo.dirstate._ignore
850 850 if not files:
851 851 # Show all the patterns
852 852 includepat = getattr(ignore, 'includepat', None)
853 853 if includepat is not None:
854 854 ui.write("%s\n" % includepat)
855 855 else:
856 856 raise error.Abort(_("no ignore patterns found"))
857 857 else:
858 858 for f in files:
859 859 nf = util.normpath(f)
860 860 ignored = None
861 861 ignoredata = None
862 862 if nf != '.':
863 863 if ignore(nf):
864 864 ignored = nf
865 865 ignoredata = repo.dirstate._ignorefileandline(nf)
866 866 else:
867 867 for p in util.finddirs(nf):
868 868 if ignore(p):
869 869 ignored = p
870 870 ignoredata = repo.dirstate._ignorefileandline(p)
871 871 break
872 872 if ignored:
873 873 if ignored == nf:
874 874 ui.write(_("%s is ignored\n") % f)
875 875 else:
876 876 ui.write(_("%s is ignored because of "
877 877 "containing folder %s\n")
878 878 % (f, ignored))
879 879 ignorefile, lineno, line = ignoredata
880 880 ui.write(_("(ignore rule in %s, line %d: '%s')\n")
881 881 % (ignorefile, lineno, line))
882 882 else:
883 883 ui.write(_("%s is not ignored\n") % f)
884 884
885 885 @command('debugindex', commands.debugrevlogopts +
886 886 [('f', 'format', 0, _('revlog format'), _('FORMAT'))],
887 887 _('[-f FORMAT] -c|-m|FILE'),
888 888 optionalrepo=True)
889 889 def debugindex(ui, repo, file_=None, **opts):
890 890 """dump the contents of an index file"""
891 891 r = cmdutil.openrevlog(repo, 'debugindex', file_, opts)
892 892 format = opts.get('format', 0)
893 893 if format not in (0, 1):
894 894 raise error.Abort(_("unknown format %d") % format)
895 895
896 896 generaldelta = r.version & revlog.FLAG_GENERALDELTA
897 897 if generaldelta:
898 898 basehdr = ' delta'
899 899 else:
900 900 basehdr = ' base'
901 901
902 902 if ui.debugflag:
903 903 shortfn = hex
904 904 else:
905 905 shortfn = short
906 906
907 907 # There might not be anything in r, so have a sane default
908 908 idlen = 12
909 909 for i in r:
910 910 idlen = len(shortfn(r.node(i)))
911 911 break
912 912
913 913 if format == 0:
914 914 ui.write((" rev offset length " + basehdr + " linkrev"
915 915 " %s %s p2\n") % ("nodeid".ljust(idlen), "p1".ljust(idlen)))
916 916 elif format == 1:
917 917 ui.write((" rev flag offset length"
918 918 " size " + basehdr + " link p1 p2"
919 919 " %s\n") % "nodeid".rjust(idlen))
920 920
921 921 for i in r:
922 922 node = r.node(i)
923 923 if generaldelta:
924 924 base = r.deltaparent(i)
925 925 else:
926 926 base = r.chainbase(i)
927 927 if format == 0:
928 928 try:
929 929 pp = r.parents(node)
930 930 except Exception:
931 931 pp = [nullid, nullid]
932 932 ui.write("% 6d % 9d % 7d % 6d % 7d %s %s %s\n" % (
933 933 i, r.start(i), r.length(i), base, r.linkrev(i),
934 934 shortfn(node), shortfn(pp[0]), shortfn(pp[1])))
935 935 elif format == 1:
936 936 pr = r.parentrevs(i)
937 937 ui.write("% 6d %04x % 8d % 8d % 8d % 6d % 6d % 6d % 6d %s\n" % (
938 938 i, r.flags(i), r.start(i), r.length(i), r.rawsize(i),
939 939 base, r.linkrev(i), pr[0], pr[1], shortfn(node)))
940 940
941 941 @command('debugindexdot', commands.debugrevlogopts,
942 942 _('-c|-m|FILE'), optionalrepo=True)
943 943 def debugindexdot(ui, repo, file_=None, **opts):
944 944 """dump an index DAG as a graphviz dot file"""
945 945 r = cmdutil.openrevlog(repo, 'debugindexdot', file_, opts)
946 946 ui.write(("digraph G {\n"))
947 947 for i in r:
948 948 node = r.node(i)
949 949 pp = r.parents(node)
950 950 ui.write("\t%d -> %d\n" % (r.rev(pp[0]), i))
951 951 if pp[1] != nullid:
952 952 ui.write("\t%d -> %d\n" % (r.rev(pp[1]), i))
953 953 ui.write("}\n")
954 954
955 955 @command('debuginstall', [] + commands.formatteropts, '', norepo=True)
956 956 def debuginstall(ui, **opts):
957 957 '''test Mercurial installation
958 958
959 959 Returns 0 on success.
960 960 '''
961 961
962 962 def writetemp(contents):
963 963 (fd, name) = tempfile.mkstemp(prefix="hg-debuginstall-")
964 964 f = os.fdopen(fd, pycompat.sysstr("wb"))
965 965 f.write(contents)
966 966 f.close()
967 967 return name
968 968
969 969 problems = 0
970 970
971 971 fm = ui.formatter('debuginstall', opts)
972 972 fm.startitem()
973 973
974 974 # encoding
975 975 fm.write('encoding', _("checking encoding (%s)...\n"), encoding.encoding)
976 976 err = None
977 977 try:
978 978 encoding.fromlocal("test")
979 979 except error.Abort as inst:
980 980 err = inst
981 981 problems += 1
982 982 fm.condwrite(err, 'encodingerror', _(" %s\n"
983 983 " (check that your locale is properly set)\n"), err)
984 984
985 985 # Python
986 986 fm.write('pythonexe', _("checking Python executable (%s)\n"),
987 987 pycompat.sysexecutable)
988 988 fm.write('pythonver', _("checking Python version (%s)\n"),
989 989 ("%d.%d.%d" % sys.version_info[:3]))
990 990 fm.write('pythonlib', _("checking Python lib (%s)...\n"),
991 991 os.path.dirname(pycompat.fsencode(os.__file__)))
992 992
993 993 security = set(sslutil.supportedprotocols)
994 994 if sslutil.hassni:
995 995 security.add('sni')
996 996
997 997 fm.write('pythonsecurity', _("checking Python security support (%s)\n"),
998 998 fm.formatlist(sorted(security), name='protocol',
999 999 fmt='%s', sep=','))
1000 1000
1001 1001 # These are warnings, not errors. So don't increment problem count. This
1002 1002 # may change in the future.
1003 1003 if 'tls1.2' not in security:
1004 1004 fm.plain(_(' TLS 1.2 not supported by Python install; '
1005 1005 'network connections lack modern security\n'))
1006 1006 if 'sni' not in security:
1007 1007 fm.plain(_(' SNI not supported by Python install; may have '
1008 1008 'connectivity issues with some servers\n'))
1009 1009
1010 1010 # TODO print CA cert info
1011 1011
1012 1012 # hg version
1013 1013 hgver = util.version()
1014 1014 fm.write('hgver', _("checking Mercurial version (%s)\n"),
1015 1015 hgver.split('+')[0])
1016 1016 fm.write('hgverextra', _("checking Mercurial custom build (%s)\n"),
1017 1017 '+'.join(hgver.split('+')[1:]))
1018 1018
1019 1019 # compiled modules
1020 1020 fm.write('hgmodulepolicy', _("checking module policy (%s)\n"),
1021 1021 policy.policy)
1022 1022 fm.write('hgmodules', _("checking installed modules (%s)...\n"),
1023 1023 os.path.dirname(pycompat.fsencode(__file__)))
1024 1024
1025 1025 if policy.policy in ('c', 'allow'):
1026 1026 err = None
1027 1027 try:
1028 1028 from . import (
1029 base85,
1030 1029 bdiff,
1031 1030 mpatch,
1032 1031 )
1033 1032 from .cext import (
1033 base85,
1034 1034 osutil,
1035 1035 )
1036 1036 dir(bdiff), dir(mpatch), dir(base85), dir(osutil) # quiet pyflakes
1037 1037 except Exception as inst:
1038 1038 err = inst
1039 1039 problems += 1
1040 1040 fm.condwrite(err, 'extensionserror', " %s\n", err)
1041 1041
1042 1042 compengines = util.compengines._engines.values()
1043 1043 fm.write('compengines', _('checking registered compression engines (%s)\n'),
1044 1044 fm.formatlist(sorted(e.name() for e in compengines),
1045 1045 name='compengine', fmt='%s', sep=', '))
1046 1046 fm.write('compenginesavail', _('checking available compression engines '
1047 1047 '(%s)\n'),
1048 1048 fm.formatlist(sorted(e.name() for e in compengines
1049 1049 if e.available()),
1050 1050 name='compengine', fmt='%s', sep=', '))
1051 1051 wirecompengines = util.compengines.supportedwireengines(util.SERVERROLE)
1052 1052 fm.write('compenginesserver', _('checking available compression engines '
1053 1053 'for wire protocol (%s)\n'),
1054 1054 fm.formatlist([e.name() for e in wirecompengines
1055 1055 if e.wireprotosupport()],
1056 1056 name='compengine', fmt='%s', sep=', '))
1057 1057
1058 1058 # templates
1059 1059 p = templater.templatepaths()
1060 1060 fm.write('templatedirs', 'checking templates (%s)...\n', ' '.join(p))
1061 1061 fm.condwrite(not p, '', _(" no template directories found\n"))
1062 1062 if p:
1063 1063 m = templater.templatepath("map-cmdline.default")
1064 1064 if m:
1065 1065 # template found, check if it is working
1066 1066 err = None
1067 1067 try:
1068 1068 templater.templater.frommapfile(m)
1069 1069 except Exception as inst:
1070 1070 err = inst
1071 1071 p = None
1072 1072 fm.condwrite(err, 'defaulttemplateerror', " %s\n", err)
1073 1073 else:
1074 1074 p = None
1075 1075 fm.condwrite(p, 'defaulttemplate',
1076 1076 _("checking default template (%s)\n"), m)
1077 1077 fm.condwrite(not m, 'defaulttemplatenotfound',
1078 1078 _(" template '%s' not found\n"), "default")
1079 1079 if not p:
1080 1080 problems += 1
1081 1081 fm.condwrite(not p, '',
1082 1082 _(" (templates seem to have been installed incorrectly)\n"))
1083 1083
1084 1084 # editor
1085 1085 editor = ui.geteditor()
1086 1086 editor = util.expandpath(editor)
1087 1087 fm.write('editor', _("checking commit editor... (%s)\n"), editor)
1088 1088 cmdpath = util.findexe(pycompat.shlexsplit(editor)[0])
1089 1089 fm.condwrite(not cmdpath and editor == 'vi', 'vinotfound',
1090 1090 _(" No commit editor set and can't find %s in PATH\n"
1091 1091 " (specify a commit editor in your configuration"
1092 1092 " file)\n"), not cmdpath and editor == 'vi' and editor)
1093 1093 fm.condwrite(not cmdpath and editor != 'vi', 'editornotfound',
1094 1094 _(" Can't find editor '%s' in PATH\n"
1095 1095 " (specify a commit editor in your configuration"
1096 1096 " file)\n"), not cmdpath and editor)
1097 1097 if not cmdpath and editor != 'vi':
1098 1098 problems += 1
1099 1099
1100 1100 # check username
1101 1101 username = None
1102 1102 err = None
1103 1103 try:
1104 1104 username = ui.username()
1105 1105 except error.Abort as e:
1106 1106 err = e
1107 1107 problems += 1
1108 1108
1109 1109 fm.condwrite(username, 'username', _("checking username (%s)\n"), username)
1110 1110 fm.condwrite(err, 'usernameerror', _("checking username...\n %s\n"
1111 1111 " (specify a username in your configuration file)\n"), err)
1112 1112
1113 1113 fm.condwrite(not problems, '',
1114 1114 _("no problems detected\n"))
1115 1115 if not problems:
1116 1116 fm.data(problems=problems)
1117 1117 fm.condwrite(problems, 'problems',
1118 1118 _("%d problems detected,"
1119 1119 " please check your install!\n"), problems)
1120 1120 fm.end()
1121 1121
1122 1122 return problems
1123 1123
1124 1124 @command('debugknown', [], _('REPO ID...'), norepo=True)
1125 1125 def debugknown(ui, repopath, *ids, **opts):
1126 1126 """test whether node ids are known to a repo
1127 1127
1128 1128 Every ID must be a full-length hex node id string. Returns a list of 0s
1129 1129 and 1s indicating unknown/known.
1130 1130 """
1131 1131 repo = hg.peer(ui, opts, repopath)
1132 1132 if not repo.capable('known'):
1133 1133 raise error.Abort("known() not supported by target repository")
1134 1134 flags = repo.known([bin(s) for s in ids])
1135 1135 ui.write("%s\n" % ("".join([f and "1" or "0" for f in flags])))
1136 1136
1137 1137 @command('debuglabelcomplete', [], _('LABEL...'))
1138 1138 def debuglabelcomplete(ui, repo, *args):
1139 1139 '''backwards compatibility with old bash completion scripts (DEPRECATED)'''
1140 1140 debugnamecomplete(ui, repo, *args)
1141 1141
1142 1142 @command('debuglocks',
1143 1143 [('L', 'force-lock', None, _('free the store lock (DANGEROUS)')),
1144 1144 ('W', 'force-wlock', None,
1145 1145 _('free the working state lock (DANGEROUS)'))],
1146 1146 _('[OPTION]...'))
1147 1147 def debuglocks(ui, repo, **opts):
1148 1148 """show or modify state of locks
1149 1149
1150 1150 By default, this command will show which locks are held. This
1151 1151 includes the user and process holding the lock, the amount of time
1152 1152 the lock has been held, and the machine name where the process is
1153 1153 running if it's not local.
1154 1154
1155 1155 Locks protect the integrity of Mercurial's data, so should be
1156 1156 treated with care. System crashes or other interruptions may cause
1157 1157 locks to not be properly released, though Mercurial will usually
1158 1158 detect and remove such stale locks automatically.
1159 1159
1160 1160 However, detecting stale locks may not always be possible (for
1161 1161 instance, on a shared filesystem). Removing locks may also be
1162 1162 blocked by filesystem permissions.
1163 1163
1164 1164 Returns 0 if no locks are held.
1165 1165
1166 1166 """
1167 1167
1168 1168 if opts.get('force_lock'):
1169 1169 repo.svfs.unlink('lock')
1170 1170 if opts.get('force_wlock'):
1171 1171 repo.vfs.unlink('wlock')
1172 1172 if opts.get('force_lock') or opts.get('force_lock'):
1173 1173 return 0
1174 1174
1175 1175 now = time.time()
1176 1176 held = 0
1177 1177
1178 1178 def report(vfs, name, method):
1179 1179 # this causes stale locks to get reaped for more accurate reporting
1180 1180 try:
1181 1181 l = method(False)
1182 1182 except error.LockHeld:
1183 1183 l = None
1184 1184
1185 1185 if l:
1186 1186 l.release()
1187 1187 else:
1188 1188 try:
1189 1189 stat = vfs.lstat(name)
1190 1190 age = now - stat.st_mtime
1191 1191 user = util.username(stat.st_uid)
1192 1192 locker = vfs.readlock(name)
1193 1193 if ":" in locker:
1194 1194 host, pid = locker.split(':')
1195 1195 if host == socket.gethostname():
1196 1196 locker = 'user %s, process %s' % (user, pid)
1197 1197 else:
1198 1198 locker = 'user %s, process %s, host %s' \
1199 1199 % (user, pid, host)
1200 1200 ui.write(("%-6s %s (%ds)\n") % (name + ":", locker, age))
1201 1201 return 1
1202 1202 except OSError as e:
1203 1203 if e.errno != errno.ENOENT:
1204 1204 raise
1205 1205
1206 1206 ui.write(("%-6s free\n") % (name + ":"))
1207 1207 return 0
1208 1208
1209 1209 held += report(repo.svfs, "lock", repo.lock)
1210 1210 held += report(repo.vfs, "wlock", repo.wlock)
1211 1211
1212 1212 return held
1213 1213
1214 1214 @command('debugmergestate', [], '')
1215 1215 def debugmergestate(ui, repo, *args):
1216 1216 """print merge state
1217 1217
1218 1218 Use --verbose to print out information about whether v1 or v2 merge state
1219 1219 was chosen."""
1220 1220 def _hashornull(h):
1221 1221 if h == nullhex:
1222 1222 return 'null'
1223 1223 else:
1224 1224 return h
1225 1225
1226 1226 def printrecords(version):
1227 1227 ui.write(('* version %s records\n') % version)
1228 1228 if version == 1:
1229 1229 records = v1records
1230 1230 else:
1231 1231 records = v2records
1232 1232
1233 1233 for rtype, record in records:
1234 1234 # pretty print some record types
1235 1235 if rtype == 'L':
1236 1236 ui.write(('local: %s\n') % record)
1237 1237 elif rtype == 'O':
1238 1238 ui.write(('other: %s\n') % record)
1239 1239 elif rtype == 'm':
1240 1240 driver, mdstate = record.split('\0', 1)
1241 1241 ui.write(('merge driver: %s (state "%s")\n')
1242 1242 % (driver, mdstate))
1243 1243 elif rtype in 'FDC':
1244 1244 r = record.split('\0')
1245 1245 f, state, hash, lfile, afile, anode, ofile = r[0:7]
1246 1246 if version == 1:
1247 1247 onode = 'not stored in v1 format'
1248 1248 flags = r[7]
1249 1249 else:
1250 1250 onode, flags = r[7:9]
1251 1251 ui.write(('file: %s (record type "%s", state "%s", hash %s)\n')
1252 1252 % (f, rtype, state, _hashornull(hash)))
1253 1253 ui.write((' local path: %s (flags "%s")\n') % (lfile, flags))
1254 1254 ui.write((' ancestor path: %s (node %s)\n')
1255 1255 % (afile, _hashornull(anode)))
1256 1256 ui.write((' other path: %s (node %s)\n')
1257 1257 % (ofile, _hashornull(onode)))
1258 1258 elif rtype == 'f':
1259 1259 filename, rawextras = record.split('\0', 1)
1260 1260 extras = rawextras.split('\0')
1261 1261 i = 0
1262 1262 extrastrings = []
1263 1263 while i < len(extras):
1264 1264 extrastrings.append('%s = %s' % (extras[i], extras[i + 1]))
1265 1265 i += 2
1266 1266
1267 1267 ui.write(('file extras: %s (%s)\n')
1268 1268 % (filename, ', '.join(extrastrings)))
1269 1269 elif rtype == 'l':
1270 1270 labels = record.split('\0', 2)
1271 1271 labels = [l for l in labels if len(l) > 0]
1272 1272 ui.write(('labels:\n'))
1273 1273 ui.write((' local: %s\n' % labels[0]))
1274 1274 ui.write((' other: %s\n' % labels[1]))
1275 1275 if len(labels) > 2:
1276 1276 ui.write((' base: %s\n' % labels[2]))
1277 1277 else:
1278 1278 ui.write(('unrecognized entry: %s\t%s\n')
1279 1279 % (rtype, record.replace('\0', '\t')))
1280 1280
1281 1281 # Avoid mergestate.read() since it may raise an exception for unsupported
1282 1282 # merge state records. We shouldn't be doing this, but this is OK since this
1283 1283 # command is pretty low-level.
1284 1284 ms = mergemod.mergestate(repo)
1285 1285
1286 1286 # sort so that reasonable information is on top
1287 1287 v1records = ms._readrecordsv1()
1288 1288 v2records = ms._readrecordsv2()
1289 1289 order = 'LOml'
1290 1290 def key(r):
1291 1291 idx = order.find(r[0])
1292 1292 if idx == -1:
1293 1293 return (1, r[1])
1294 1294 else:
1295 1295 return (0, idx)
1296 1296 v1records.sort(key=key)
1297 1297 v2records.sort(key=key)
1298 1298
1299 1299 if not v1records and not v2records:
1300 1300 ui.write(('no merge state found\n'))
1301 1301 elif not v2records:
1302 1302 ui.note(('no version 2 merge state\n'))
1303 1303 printrecords(1)
1304 1304 elif ms._v1v2match(v1records, v2records):
1305 1305 ui.note(('v1 and v2 states match: using v2\n'))
1306 1306 printrecords(2)
1307 1307 else:
1308 1308 ui.note(('v1 and v2 states mismatch: using v1\n'))
1309 1309 printrecords(1)
1310 1310 if ui.verbose:
1311 1311 printrecords(2)
1312 1312
1313 1313 @command('debugnamecomplete', [], _('NAME...'))
1314 1314 def debugnamecomplete(ui, repo, *args):
1315 1315 '''complete "names" - tags, open branch names, bookmark names'''
1316 1316
1317 1317 names = set()
1318 1318 # since we previously only listed open branches, we will handle that
1319 1319 # specially (after this for loop)
1320 1320 for name, ns in repo.names.iteritems():
1321 1321 if name != 'branches':
1322 1322 names.update(ns.listnames(repo))
1323 1323 names.update(tag for (tag, heads, tip, closed)
1324 1324 in repo.branchmap().iterbranches() if not closed)
1325 1325 completions = set()
1326 1326 if not args:
1327 1327 args = ['']
1328 1328 for a in args:
1329 1329 completions.update(n for n in names if n.startswith(a))
1330 1330 ui.write('\n'.join(sorted(completions)))
1331 1331 ui.write('\n')
1332 1332
1333 1333 @command('debugobsolete',
1334 1334 [('', 'flags', 0, _('markers flag')),
1335 1335 ('', 'record-parents', False,
1336 1336 _('record parent information for the precursor')),
1337 1337 ('r', 'rev', [], _('display markers relevant to REV')),
1338 1338 ('', 'index', False, _('display index of the marker')),
1339 1339 ('', 'delete', [], _('delete markers specified by indices')),
1340 1340 ] + commands.commitopts2 + commands.formatteropts,
1341 1341 _('[OBSOLETED [REPLACEMENT ...]]'))
1342 1342 def debugobsolete(ui, repo, precursor=None, *successors, **opts):
1343 1343 """create arbitrary obsolete marker
1344 1344
1345 1345 With no arguments, displays the list of obsolescence markers."""
1346 1346
1347 1347 def parsenodeid(s):
1348 1348 try:
1349 1349 # We do not use revsingle/revrange functions here to accept
1350 1350 # arbitrary node identifiers, possibly not present in the
1351 1351 # local repository.
1352 1352 n = bin(s)
1353 1353 if len(n) != len(nullid):
1354 1354 raise TypeError()
1355 1355 return n
1356 1356 except TypeError:
1357 1357 raise error.Abort('changeset references must be full hexadecimal '
1358 1358 'node identifiers')
1359 1359
1360 1360 if opts.get('delete'):
1361 1361 indices = []
1362 1362 for v in opts.get('delete'):
1363 1363 try:
1364 1364 indices.append(int(v))
1365 1365 except ValueError:
1366 1366 raise error.Abort(_('invalid index value: %r') % v,
1367 1367 hint=_('use integers for indices'))
1368 1368
1369 1369 if repo.currenttransaction():
1370 1370 raise error.Abort(_('cannot delete obsmarkers in the middle '
1371 1371 'of transaction.'))
1372 1372
1373 1373 with repo.lock():
1374 1374 n = repair.deleteobsmarkers(repo.obsstore, indices)
1375 1375 ui.write(_('deleted %i obsolescence markers\n') % n)
1376 1376
1377 1377 return
1378 1378
1379 1379 if precursor is not None:
1380 1380 if opts['rev']:
1381 1381 raise error.Abort('cannot select revision when creating marker')
1382 1382 metadata = {}
1383 1383 metadata['user'] = opts['user'] or ui.username()
1384 1384 succs = tuple(parsenodeid(succ) for succ in successors)
1385 1385 l = repo.lock()
1386 1386 try:
1387 1387 tr = repo.transaction('debugobsolete')
1388 1388 try:
1389 1389 date = opts.get('date')
1390 1390 if date:
1391 1391 date = util.parsedate(date)
1392 1392 else:
1393 1393 date = None
1394 1394 prec = parsenodeid(precursor)
1395 1395 parents = None
1396 1396 if opts['record_parents']:
1397 1397 if prec not in repo.unfiltered():
1398 1398 raise error.Abort('cannot used --record-parents on '
1399 1399 'unknown changesets')
1400 1400 parents = repo.unfiltered()[prec].parents()
1401 1401 parents = tuple(p.node() for p in parents)
1402 1402 repo.obsstore.create(tr, prec, succs, opts['flags'],
1403 1403 parents=parents, date=date,
1404 1404 metadata=metadata)
1405 1405 tr.close()
1406 1406 except ValueError as exc:
1407 1407 raise error.Abort(_('bad obsmarker input: %s') % exc)
1408 1408 finally:
1409 1409 tr.release()
1410 1410 finally:
1411 1411 l.release()
1412 1412 else:
1413 1413 if opts['rev']:
1414 1414 revs = scmutil.revrange(repo, opts['rev'])
1415 1415 nodes = [repo[r].node() for r in revs]
1416 1416 markers = list(obsolete.getmarkers(repo, nodes=nodes))
1417 1417 markers.sort(key=lambda x: x._data)
1418 1418 else:
1419 1419 markers = obsolete.getmarkers(repo)
1420 1420
1421 1421 markerstoiter = markers
1422 1422 isrelevant = lambda m: True
1423 1423 if opts.get('rev') and opts.get('index'):
1424 1424 markerstoiter = obsolete.getmarkers(repo)
1425 1425 markerset = set(markers)
1426 1426 isrelevant = lambda m: m in markerset
1427 1427
1428 1428 fm = ui.formatter('debugobsolete', opts)
1429 1429 for i, m in enumerate(markerstoiter):
1430 1430 if not isrelevant(m):
1431 1431 # marker can be irrelevant when we're iterating over a set
1432 1432 # of markers (markerstoiter) which is bigger than the set
1433 1433 # of markers we want to display (markers)
1434 1434 # this can happen if both --index and --rev options are
1435 1435 # provided and thus we need to iterate over all of the markers
1436 1436 # to get the correct indices, but only display the ones that
1437 1437 # are relevant to --rev value
1438 1438 continue
1439 1439 fm.startitem()
1440 1440 ind = i if opts.get('index') else None
1441 1441 cmdutil.showmarker(fm, m, index=ind)
1442 1442 fm.end()
1443 1443
1444 1444 @command('debugpathcomplete',
1445 1445 [('f', 'full', None, _('complete an entire path')),
1446 1446 ('n', 'normal', None, _('show only normal files')),
1447 1447 ('a', 'added', None, _('show only added files')),
1448 1448 ('r', 'removed', None, _('show only removed files'))],
1449 1449 _('FILESPEC...'))
1450 1450 def debugpathcomplete(ui, repo, *specs, **opts):
1451 1451 '''complete part or all of a tracked path
1452 1452
1453 1453 This command supports shells that offer path name completion. It
1454 1454 currently completes only files already known to the dirstate.
1455 1455
1456 1456 Completion extends only to the next path segment unless
1457 1457 --full is specified, in which case entire paths are used.'''
1458 1458
1459 1459 def complete(path, acceptable):
1460 1460 dirstate = repo.dirstate
1461 1461 spec = os.path.normpath(os.path.join(pycompat.getcwd(), path))
1462 1462 rootdir = repo.root + pycompat.ossep
1463 1463 if spec != repo.root and not spec.startswith(rootdir):
1464 1464 return [], []
1465 1465 if os.path.isdir(spec):
1466 1466 spec += '/'
1467 1467 spec = spec[len(rootdir):]
1468 1468 fixpaths = pycompat.ossep != '/'
1469 1469 if fixpaths:
1470 1470 spec = spec.replace(pycompat.ossep, '/')
1471 1471 speclen = len(spec)
1472 1472 fullpaths = opts['full']
1473 1473 files, dirs = set(), set()
1474 1474 adddir, addfile = dirs.add, files.add
1475 1475 for f, st in dirstate.iteritems():
1476 1476 if f.startswith(spec) and st[0] in acceptable:
1477 1477 if fixpaths:
1478 1478 f = f.replace('/', pycompat.ossep)
1479 1479 if fullpaths:
1480 1480 addfile(f)
1481 1481 continue
1482 1482 s = f.find(pycompat.ossep, speclen)
1483 1483 if s >= 0:
1484 1484 adddir(f[:s])
1485 1485 else:
1486 1486 addfile(f)
1487 1487 return files, dirs
1488 1488
1489 1489 acceptable = ''
1490 1490 if opts['normal']:
1491 1491 acceptable += 'nm'
1492 1492 if opts['added']:
1493 1493 acceptable += 'a'
1494 1494 if opts['removed']:
1495 1495 acceptable += 'r'
1496 1496 cwd = repo.getcwd()
1497 1497 if not specs:
1498 1498 specs = ['.']
1499 1499
1500 1500 files, dirs = set(), set()
1501 1501 for spec in specs:
1502 1502 f, d = complete(spec, acceptable or 'nmar')
1503 1503 files.update(f)
1504 1504 dirs.update(d)
1505 1505 files.update(dirs)
1506 1506 ui.write('\n'.join(repo.pathto(p, cwd) for p in sorted(files)))
1507 1507 ui.write('\n')
1508 1508
1509 1509 @command('debugpickmergetool',
1510 1510 [('r', 'rev', '', _('check for files in this revision'), _('REV')),
1511 1511 ('', 'changedelete', None, _('emulate merging change and delete')),
1512 1512 ] + commands.walkopts + commands.mergetoolopts,
1513 1513 _('[PATTERN]...'),
1514 1514 inferrepo=True)
1515 1515 def debugpickmergetool(ui, repo, *pats, **opts):
1516 1516 """examine which merge tool is chosen for specified file
1517 1517
1518 1518 As described in :hg:`help merge-tools`, Mercurial examines
1519 1519 configurations below in this order to decide which merge tool is
1520 1520 chosen for specified file.
1521 1521
1522 1522 1. ``--tool`` option
1523 1523 2. ``HGMERGE`` environment variable
1524 1524 3. configurations in ``merge-patterns`` section
1525 1525 4. configuration of ``ui.merge``
1526 1526 5. configurations in ``merge-tools`` section
1527 1527 6. ``hgmerge`` tool (for historical reason only)
1528 1528 7. default tool for fallback (``:merge`` or ``:prompt``)
1529 1529
1530 1530 This command writes out examination result in the style below::
1531 1531
1532 1532 FILE = MERGETOOL
1533 1533
1534 1534 By default, all files known in the first parent context of the
1535 1535 working directory are examined. Use file patterns and/or -I/-X
1536 1536 options to limit target files. -r/--rev is also useful to examine
1537 1537 files in another context without actual updating to it.
1538 1538
1539 1539 With --debug, this command shows warning messages while matching
1540 1540 against ``merge-patterns`` and so on, too. It is recommended to
1541 1541 use this option with explicit file patterns and/or -I/-X options,
1542 1542 because this option increases amount of output per file according
1543 1543 to configurations in hgrc.
1544 1544
1545 1545 With -v/--verbose, this command shows configurations below at
1546 1546 first (only if specified).
1547 1547
1548 1548 - ``--tool`` option
1549 1549 - ``HGMERGE`` environment variable
1550 1550 - configuration of ``ui.merge``
1551 1551
1552 1552 If merge tool is chosen before matching against
1553 1553 ``merge-patterns``, this command can't show any helpful
1554 1554 information, even with --debug. In such case, information above is
1555 1555 useful to know why a merge tool is chosen.
1556 1556 """
1557 1557 overrides = {}
1558 1558 if opts['tool']:
1559 1559 overrides[('ui', 'forcemerge')] = opts['tool']
1560 1560 ui.note(('with --tool %r\n') % (opts['tool']))
1561 1561
1562 1562 with ui.configoverride(overrides, 'debugmergepatterns'):
1563 1563 hgmerge = encoding.environ.get("HGMERGE")
1564 1564 if hgmerge is not None:
1565 1565 ui.note(('with HGMERGE=%r\n') % (hgmerge))
1566 1566 uimerge = ui.config("ui", "merge")
1567 1567 if uimerge:
1568 1568 ui.note(('with ui.merge=%r\n') % (uimerge))
1569 1569
1570 1570 ctx = scmutil.revsingle(repo, opts.get('rev'))
1571 1571 m = scmutil.match(ctx, pats, opts)
1572 1572 changedelete = opts['changedelete']
1573 1573 for path in ctx.walk(m):
1574 1574 fctx = ctx[path]
1575 1575 try:
1576 1576 if not ui.debugflag:
1577 1577 ui.pushbuffer(error=True)
1578 1578 tool, toolpath = filemerge._picktool(repo, ui, path,
1579 1579 fctx.isbinary(),
1580 1580 'l' in fctx.flags(),
1581 1581 changedelete)
1582 1582 finally:
1583 1583 if not ui.debugflag:
1584 1584 ui.popbuffer()
1585 1585 ui.write(('%s = %s\n') % (path, tool))
1586 1586
1587 1587 @command('debugpushkey', [], _('REPO NAMESPACE [KEY OLD NEW]'), norepo=True)
1588 1588 def debugpushkey(ui, repopath, namespace, *keyinfo, **opts):
1589 1589 '''access the pushkey key/value protocol
1590 1590
1591 1591 With two args, list the keys in the given namespace.
1592 1592
1593 1593 With five args, set a key to new if it currently is set to old.
1594 1594 Reports success or failure.
1595 1595 '''
1596 1596
1597 1597 target = hg.peer(ui, {}, repopath)
1598 1598 if keyinfo:
1599 1599 key, old, new = keyinfo
1600 1600 r = target.pushkey(namespace, key, old, new)
1601 1601 ui.status(str(r) + '\n')
1602 1602 return not r
1603 1603 else:
1604 1604 for k, v in sorted(target.listkeys(namespace).iteritems()):
1605 1605 ui.write("%s\t%s\n" % (util.escapestr(k),
1606 1606 util.escapestr(v)))
1607 1607
1608 1608 @command('debugpvec', [], _('A B'))
1609 1609 def debugpvec(ui, repo, a, b=None):
1610 1610 ca = scmutil.revsingle(repo, a)
1611 1611 cb = scmutil.revsingle(repo, b)
1612 1612 pa = pvec.ctxpvec(ca)
1613 1613 pb = pvec.ctxpvec(cb)
1614 1614 if pa == pb:
1615 1615 rel = "="
1616 1616 elif pa > pb:
1617 1617 rel = ">"
1618 1618 elif pa < pb:
1619 1619 rel = "<"
1620 1620 elif pa | pb:
1621 1621 rel = "|"
1622 1622 ui.write(_("a: %s\n") % pa)
1623 1623 ui.write(_("b: %s\n") % pb)
1624 1624 ui.write(_("depth(a): %d depth(b): %d\n") % (pa._depth, pb._depth))
1625 1625 ui.write(_("delta: %d hdist: %d distance: %d relation: %s\n") %
1626 1626 (abs(pa._depth - pb._depth), pvec._hamming(pa._vec, pb._vec),
1627 1627 pa.distance(pb), rel))
1628 1628
1629 1629 @command('debugrebuilddirstate|debugrebuildstate',
1630 1630 [('r', 'rev', '', _('revision to rebuild to'), _('REV')),
1631 1631 ('', 'minimal', None, _('only rebuild files that are inconsistent with '
1632 1632 'the working copy parent')),
1633 1633 ],
1634 1634 _('[-r REV]'))
1635 1635 def debugrebuilddirstate(ui, repo, rev, **opts):
1636 1636 """rebuild the dirstate as it would look like for the given revision
1637 1637
1638 1638 If no revision is specified the first current parent will be used.
1639 1639
1640 1640 The dirstate will be set to the files of the given revision.
1641 1641 The actual working directory content or existing dirstate
1642 1642 information such as adds or removes is not considered.
1643 1643
1644 1644 ``minimal`` will only rebuild the dirstate status for files that claim to be
1645 1645 tracked but are not in the parent manifest, or that exist in the parent
1646 1646 manifest but are not in the dirstate. It will not change adds, removes, or
1647 1647 modified files that are in the working copy parent.
1648 1648
1649 1649 One use of this command is to make the next :hg:`status` invocation
1650 1650 check the actual file content.
1651 1651 """
1652 1652 ctx = scmutil.revsingle(repo, rev)
1653 1653 with repo.wlock():
1654 1654 dirstate = repo.dirstate
1655 1655 changedfiles = None
1656 1656 # See command doc for what minimal does.
1657 1657 if opts.get('minimal'):
1658 1658 manifestfiles = set(ctx.manifest().keys())
1659 1659 dirstatefiles = set(dirstate)
1660 1660 manifestonly = manifestfiles - dirstatefiles
1661 1661 dsonly = dirstatefiles - manifestfiles
1662 1662 dsnotadded = set(f for f in dsonly if dirstate[f] != 'a')
1663 1663 changedfiles = manifestonly | dsnotadded
1664 1664
1665 1665 dirstate.rebuild(ctx.node(), ctx.manifest(), changedfiles)
1666 1666
1667 1667 @command('debugrebuildfncache', [], '')
1668 1668 def debugrebuildfncache(ui, repo):
1669 1669 """rebuild the fncache file"""
1670 1670 repair.rebuildfncache(ui, repo)
1671 1671
1672 1672 @command('debugrename',
1673 1673 [('r', 'rev', '', _('revision to debug'), _('REV'))],
1674 1674 _('[-r REV] FILE'))
1675 1675 def debugrename(ui, repo, file1, *pats, **opts):
1676 1676 """dump rename information"""
1677 1677
1678 1678 ctx = scmutil.revsingle(repo, opts.get('rev'))
1679 1679 m = scmutil.match(ctx, (file1,) + pats, opts)
1680 1680 for abs in ctx.walk(m):
1681 1681 fctx = ctx[abs]
1682 1682 o = fctx.filelog().renamed(fctx.filenode())
1683 1683 rel = m.rel(abs)
1684 1684 if o:
1685 1685 ui.write(_("%s renamed from %s:%s\n") % (rel, o[0], hex(o[1])))
1686 1686 else:
1687 1687 ui.write(_("%s not renamed\n") % rel)
1688 1688
1689 1689 @command('debugrevlog', commands.debugrevlogopts +
1690 1690 [('d', 'dump', False, _('dump index data'))],
1691 1691 _('-c|-m|FILE'),
1692 1692 optionalrepo=True)
1693 1693 def debugrevlog(ui, repo, file_=None, **opts):
1694 1694 """show data and statistics about a revlog"""
1695 1695 r = cmdutil.openrevlog(repo, 'debugrevlog', file_, opts)
1696 1696
1697 1697 if opts.get("dump"):
1698 1698 numrevs = len(r)
1699 1699 ui.write(("# rev p1rev p2rev start end deltastart base p1 p2"
1700 1700 " rawsize totalsize compression heads chainlen\n"))
1701 1701 ts = 0
1702 1702 heads = set()
1703 1703
1704 1704 for rev in xrange(numrevs):
1705 1705 dbase = r.deltaparent(rev)
1706 1706 if dbase == -1:
1707 1707 dbase = rev
1708 1708 cbase = r.chainbase(rev)
1709 1709 clen = r.chainlen(rev)
1710 1710 p1, p2 = r.parentrevs(rev)
1711 1711 rs = r.rawsize(rev)
1712 1712 ts = ts + rs
1713 1713 heads -= set(r.parentrevs(rev))
1714 1714 heads.add(rev)
1715 1715 try:
1716 1716 compression = ts / r.end(rev)
1717 1717 except ZeroDivisionError:
1718 1718 compression = 0
1719 1719 ui.write("%5d %5d %5d %5d %5d %10d %4d %4d %4d %7d %9d "
1720 1720 "%11d %5d %8d\n" %
1721 1721 (rev, p1, p2, r.start(rev), r.end(rev),
1722 1722 r.start(dbase), r.start(cbase),
1723 1723 r.start(p1), r.start(p2),
1724 1724 rs, ts, compression, len(heads), clen))
1725 1725 return 0
1726 1726
1727 1727 v = r.version
1728 1728 format = v & 0xFFFF
1729 1729 flags = []
1730 1730 gdelta = False
1731 1731 if v & revlog.FLAG_INLINE_DATA:
1732 1732 flags.append('inline')
1733 1733 if v & revlog.FLAG_GENERALDELTA:
1734 1734 gdelta = True
1735 1735 flags.append('generaldelta')
1736 1736 if not flags:
1737 1737 flags = ['(none)']
1738 1738
1739 1739 nummerges = 0
1740 1740 numfull = 0
1741 1741 numprev = 0
1742 1742 nump1 = 0
1743 1743 nump2 = 0
1744 1744 numother = 0
1745 1745 nump1prev = 0
1746 1746 nump2prev = 0
1747 1747 chainlengths = []
1748 1748
1749 1749 datasize = [None, 0, 0]
1750 1750 fullsize = [None, 0, 0]
1751 1751 deltasize = [None, 0, 0]
1752 1752 chunktypecounts = {}
1753 1753 chunktypesizes = {}
1754 1754
1755 1755 def addsize(size, l):
1756 1756 if l[0] is None or size < l[0]:
1757 1757 l[0] = size
1758 1758 if size > l[1]:
1759 1759 l[1] = size
1760 1760 l[2] += size
1761 1761
1762 1762 numrevs = len(r)
1763 1763 for rev in xrange(numrevs):
1764 1764 p1, p2 = r.parentrevs(rev)
1765 1765 delta = r.deltaparent(rev)
1766 1766 if format > 0:
1767 1767 addsize(r.rawsize(rev), datasize)
1768 1768 if p2 != nullrev:
1769 1769 nummerges += 1
1770 1770 size = r.length(rev)
1771 1771 if delta == nullrev:
1772 1772 chainlengths.append(0)
1773 1773 numfull += 1
1774 1774 addsize(size, fullsize)
1775 1775 else:
1776 1776 chainlengths.append(chainlengths[delta] + 1)
1777 1777 addsize(size, deltasize)
1778 1778 if delta == rev - 1:
1779 1779 numprev += 1
1780 1780 if delta == p1:
1781 1781 nump1prev += 1
1782 1782 elif delta == p2:
1783 1783 nump2prev += 1
1784 1784 elif delta == p1:
1785 1785 nump1 += 1
1786 1786 elif delta == p2:
1787 1787 nump2 += 1
1788 1788 elif delta != nullrev:
1789 1789 numother += 1
1790 1790
1791 1791 # Obtain data on the raw chunks in the revlog.
1792 1792 segment = r._getsegmentforrevs(rev, rev)[1]
1793 1793 if segment:
1794 1794 chunktype = segment[0]
1795 1795 else:
1796 1796 chunktype = 'empty'
1797 1797
1798 1798 if chunktype not in chunktypecounts:
1799 1799 chunktypecounts[chunktype] = 0
1800 1800 chunktypesizes[chunktype] = 0
1801 1801
1802 1802 chunktypecounts[chunktype] += 1
1803 1803 chunktypesizes[chunktype] += size
1804 1804
1805 1805 # Adjust size min value for empty cases
1806 1806 for size in (datasize, fullsize, deltasize):
1807 1807 if size[0] is None:
1808 1808 size[0] = 0
1809 1809
1810 1810 numdeltas = numrevs - numfull
1811 1811 numoprev = numprev - nump1prev - nump2prev
1812 1812 totalrawsize = datasize[2]
1813 1813 datasize[2] /= numrevs
1814 1814 fulltotal = fullsize[2]
1815 1815 fullsize[2] /= numfull
1816 1816 deltatotal = deltasize[2]
1817 1817 if numrevs - numfull > 0:
1818 1818 deltasize[2] /= numrevs - numfull
1819 1819 totalsize = fulltotal + deltatotal
1820 1820 avgchainlen = sum(chainlengths) / numrevs
1821 1821 maxchainlen = max(chainlengths)
1822 1822 compratio = 1
1823 1823 if totalsize:
1824 1824 compratio = totalrawsize / totalsize
1825 1825
1826 1826 basedfmtstr = '%%%dd\n'
1827 1827 basepcfmtstr = '%%%dd %s(%%5.2f%%%%)\n'
1828 1828
1829 1829 def dfmtstr(max):
1830 1830 return basedfmtstr % len(str(max))
1831 1831 def pcfmtstr(max, padding=0):
1832 1832 return basepcfmtstr % (len(str(max)), ' ' * padding)
1833 1833
1834 1834 def pcfmt(value, total):
1835 1835 if total:
1836 1836 return (value, 100 * float(value) / total)
1837 1837 else:
1838 1838 return value, 100.0
1839 1839
1840 1840 ui.write(('format : %d\n') % format)
1841 1841 ui.write(('flags : %s\n') % ', '.join(flags))
1842 1842
1843 1843 ui.write('\n')
1844 1844 fmt = pcfmtstr(totalsize)
1845 1845 fmt2 = dfmtstr(totalsize)
1846 1846 ui.write(('revisions : ') + fmt2 % numrevs)
1847 1847 ui.write((' merges : ') + fmt % pcfmt(nummerges, numrevs))
1848 1848 ui.write((' normal : ') + fmt % pcfmt(numrevs - nummerges, numrevs))
1849 1849 ui.write(('revisions : ') + fmt2 % numrevs)
1850 1850 ui.write((' full : ') + fmt % pcfmt(numfull, numrevs))
1851 1851 ui.write((' deltas : ') + fmt % pcfmt(numdeltas, numrevs))
1852 1852 ui.write(('revision size : ') + fmt2 % totalsize)
1853 1853 ui.write((' full : ') + fmt % pcfmt(fulltotal, totalsize))
1854 1854 ui.write((' deltas : ') + fmt % pcfmt(deltatotal, totalsize))
1855 1855
1856 1856 def fmtchunktype(chunktype):
1857 1857 if chunktype == 'empty':
1858 1858 return ' %s : ' % chunktype
1859 1859 elif chunktype in string.ascii_letters:
1860 1860 return ' 0x%s (%s) : ' % (hex(chunktype), chunktype)
1861 1861 else:
1862 1862 return ' 0x%s : ' % hex(chunktype)
1863 1863
1864 1864 ui.write('\n')
1865 1865 ui.write(('chunks : ') + fmt2 % numrevs)
1866 1866 for chunktype in sorted(chunktypecounts):
1867 1867 ui.write(fmtchunktype(chunktype))
1868 1868 ui.write(fmt % pcfmt(chunktypecounts[chunktype], numrevs))
1869 1869 ui.write(('chunks size : ') + fmt2 % totalsize)
1870 1870 for chunktype in sorted(chunktypecounts):
1871 1871 ui.write(fmtchunktype(chunktype))
1872 1872 ui.write(fmt % pcfmt(chunktypesizes[chunktype], totalsize))
1873 1873
1874 1874 ui.write('\n')
1875 1875 fmt = dfmtstr(max(avgchainlen, compratio))
1876 1876 ui.write(('avg chain length : ') + fmt % avgchainlen)
1877 1877 ui.write(('max chain length : ') + fmt % maxchainlen)
1878 1878 ui.write(('compression ratio : ') + fmt % compratio)
1879 1879
1880 1880 if format > 0:
1881 1881 ui.write('\n')
1882 1882 ui.write(('uncompressed data size (min/max/avg) : %d / %d / %d\n')
1883 1883 % tuple(datasize))
1884 1884 ui.write(('full revision size (min/max/avg) : %d / %d / %d\n')
1885 1885 % tuple(fullsize))
1886 1886 ui.write(('delta size (min/max/avg) : %d / %d / %d\n')
1887 1887 % tuple(deltasize))
1888 1888
1889 1889 if numdeltas > 0:
1890 1890 ui.write('\n')
1891 1891 fmt = pcfmtstr(numdeltas)
1892 1892 fmt2 = pcfmtstr(numdeltas, 4)
1893 1893 ui.write(('deltas against prev : ') + fmt % pcfmt(numprev, numdeltas))
1894 1894 if numprev > 0:
1895 1895 ui.write((' where prev = p1 : ') + fmt2 % pcfmt(nump1prev,
1896 1896 numprev))
1897 1897 ui.write((' where prev = p2 : ') + fmt2 % pcfmt(nump2prev,
1898 1898 numprev))
1899 1899 ui.write((' other : ') + fmt2 % pcfmt(numoprev,
1900 1900 numprev))
1901 1901 if gdelta:
1902 1902 ui.write(('deltas against p1 : ')
1903 1903 + fmt % pcfmt(nump1, numdeltas))
1904 1904 ui.write(('deltas against p2 : ')
1905 1905 + fmt % pcfmt(nump2, numdeltas))
1906 1906 ui.write(('deltas against other : ') + fmt % pcfmt(numother,
1907 1907 numdeltas))
1908 1908
1909 1909 @command('debugrevspec',
1910 1910 [('', 'optimize', None,
1911 1911 _('print parsed tree after optimizing (DEPRECATED)')),
1912 1912 ('p', 'show-stage', [],
1913 1913 _('print parsed tree at the given stage'), _('NAME')),
1914 1914 ('', 'no-optimized', False, _('evaluate tree without optimization')),
1915 1915 ('', 'verify-optimized', False, _('verify optimized result')),
1916 1916 ],
1917 1917 ('REVSPEC'))
1918 1918 def debugrevspec(ui, repo, expr, **opts):
1919 1919 """parse and apply a revision specification
1920 1920
1921 1921 Use -p/--show-stage option to print the parsed tree at the given stages.
1922 1922 Use -p all to print tree at every stage.
1923 1923
1924 1924 Use --verify-optimized to compare the optimized result with the unoptimized
1925 1925 one. Returns 1 if the optimized result differs.
1926 1926 """
1927 1927 stages = [
1928 1928 ('parsed', lambda tree: tree),
1929 1929 ('expanded', lambda tree: revsetlang.expandaliases(ui, tree)),
1930 1930 ('concatenated', revsetlang.foldconcat),
1931 1931 ('analyzed', revsetlang.analyze),
1932 1932 ('optimized', revsetlang.optimize),
1933 1933 ]
1934 1934 if opts['no_optimized']:
1935 1935 stages = stages[:-1]
1936 1936 if opts['verify_optimized'] and opts['no_optimized']:
1937 1937 raise error.Abort(_('cannot use --verify-optimized with '
1938 1938 '--no-optimized'))
1939 1939 stagenames = set(n for n, f in stages)
1940 1940
1941 1941 showalways = set()
1942 1942 showchanged = set()
1943 1943 if ui.verbose and not opts['show_stage']:
1944 1944 # show parsed tree by --verbose (deprecated)
1945 1945 showalways.add('parsed')
1946 1946 showchanged.update(['expanded', 'concatenated'])
1947 1947 if opts['optimize']:
1948 1948 showalways.add('optimized')
1949 1949 if opts['show_stage'] and opts['optimize']:
1950 1950 raise error.Abort(_('cannot use --optimize with --show-stage'))
1951 1951 if opts['show_stage'] == ['all']:
1952 1952 showalways.update(stagenames)
1953 1953 else:
1954 1954 for n in opts['show_stage']:
1955 1955 if n not in stagenames:
1956 1956 raise error.Abort(_('invalid stage name: %s') % n)
1957 1957 showalways.update(opts['show_stage'])
1958 1958
1959 1959 treebystage = {}
1960 1960 printedtree = None
1961 1961 tree = revsetlang.parse(expr, lookup=repo.__contains__)
1962 1962 for n, f in stages:
1963 1963 treebystage[n] = tree = f(tree)
1964 1964 if n in showalways or (n in showchanged and tree != printedtree):
1965 1965 if opts['show_stage'] or n != 'parsed':
1966 1966 ui.write(("* %s:\n") % n)
1967 1967 ui.write(revsetlang.prettyformat(tree), "\n")
1968 1968 printedtree = tree
1969 1969
1970 1970 if opts['verify_optimized']:
1971 1971 arevs = revset.makematcher(treebystage['analyzed'])(repo)
1972 1972 brevs = revset.makematcher(treebystage['optimized'])(repo)
1973 1973 if ui.verbose:
1974 1974 ui.note(("* analyzed set:\n"), smartset.prettyformat(arevs), "\n")
1975 1975 ui.note(("* optimized set:\n"), smartset.prettyformat(brevs), "\n")
1976 1976 arevs = list(arevs)
1977 1977 brevs = list(brevs)
1978 1978 if arevs == brevs:
1979 1979 return 0
1980 1980 ui.write(('--- analyzed\n'), label='diff.file_a')
1981 1981 ui.write(('+++ optimized\n'), label='diff.file_b')
1982 1982 sm = difflib.SequenceMatcher(None, arevs, brevs)
1983 1983 for tag, alo, ahi, blo, bhi in sm.get_opcodes():
1984 1984 if tag in ('delete', 'replace'):
1985 1985 for c in arevs[alo:ahi]:
1986 1986 ui.write('-%s\n' % c, label='diff.deleted')
1987 1987 if tag in ('insert', 'replace'):
1988 1988 for c in brevs[blo:bhi]:
1989 1989 ui.write('+%s\n' % c, label='diff.inserted')
1990 1990 if tag == 'equal':
1991 1991 for c in arevs[alo:ahi]:
1992 1992 ui.write(' %s\n' % c)
1993 1993 return 1
1994 1994
1995 1995 func = revset.makematcher(tree)
1996 1996 revs = func(repo)
1997 1997 if ui.verbose:
1998 1998 ui.note(("* set:\n"), smartset.prettyformat(revs), "\n")
1999 1999 for c in revs:
2000 2000 ui.write("%s\n" % c)
2001 2001
2002 2002 @command('debugsetparents', [], _('REV1 [REV2]'))
2003 2003 def debugsetparents(ui, repo, rev1, rev2=None):
2004 2004 """manually set the parents of the current working directory
2005 2005
2006 2006 This is useful for writing repository conversion tools, but should
2007 2007 be used with care. For example, neither the working directory nor the
2008 2008 dirstate is updated, so file status may be incorrect after running this
2009 2009 command.
2010 2010
2011 2011 Returns 0 on success.
2012 2012 """
2013 2013
2014 2014 r1 = scmutil.revsingle(repo, rev1).node()
2015 2015 r2 = scmutil.revsingle(repo, rev2, 'null').node()
2016 2016
2017 2017 with repo.wlock():
2018 2018 repo.setparents(r1, r2)
2019 2019
2020 2020 @command('debugsub',
2021 2021 [('r', 'rev', '',
2022 2022 _('revision to check'), _('REV'))],
2023 2023 _('[-r REV] [REV]'))
2024 2024 def debugsub(ui, repo, rev=None):
2025 2025 ctx = scmutil.revsingle(repo, rev, None)
2026 2026 for k, v in sorted(ctx.substate.items()):
2027 2027 ui.write(('path %s\n') % k)
2028 2028 ui.write((' source %s\n') % v[0])
2029 2029 ui.write((' revision %s\n') % v[1])
2030 2030
2031 2031 @command('debugsuccessorssets',
2032 2032 [],
2033 2033 _('[REV]'))
2034 2034 def debugsuccessorssets(ui, repo, *revs):
2035 2035 """show set of successors for revision
2036 2036
2037 2037 A successors set of changeset A is a consistent group of revisions that
2038 2038 succeed A. It contains non-obsolete changesets only.
2039 2039
2040 2040 In most cases a changeset A has a single successors set containing a single
2041 2041 successor (changeset A replaced by A').
2042 2042
2043 2043 A changeset that is made obsolete with no successors are called "pruned".
2044 2044 Such changesets have no successors sets at all.
2045 2045
2046 2046 A changeset that has been "split" will have a successors set containing
2047 2047 more than one successor.
2048 2048
2049 2049 A changeset that has been rewritten in multiple different ways is called
2050 2050 "divergent". Such changesets have multiple successor sets (each of which
2051 2051 may also be split, i.e. have multiple successors).
2052 2052
2053 2053 Results are displayed as follows::
2054 2054
2055 2055 <rev1>
2056 2056 <successors-1A>
2057 2057 <rev2>
2058 2058 <successors-2A>
2059 2059 <successors-2B1> <successors-2B2> <successors-2B3>
2060 2060
2061 2061 Here rev2 has two possible (i.e. divergent) successors sets. The first
2062 2062 holds one element, whereas the second holds three (i.e. the changeset has
2063 2063 been split).
2064 2064 """
2065 2065 # passed to successorssets caching computation from one call to another
2066 2066 cache = {}
2067 2067 ctx2str = str
2068 2068 node2str = short
2069 2069 if ui.debug():
2070 2070 def ctx2str(ctx):
2071 2071 return ctx.hex()
2072 2072 node2str = hex
2073 2073 for rev in scmutil.revrange(repo, revs):
2074 2074 ctx = repo[rev]
2075 2075 ui.write('%s\n'% ctx2str(ctx))
2076 2076 for succsset in obsolete.successorssets(repo, ctx.node(), cache):
2077 2077 if succsset:
2078 2078 ui.write(' ')
2079 2079 ui.write(node2str(succsset[0]))
2080 2080 for node in succsset[1:]:
2081 2081 ui.write(' ')
2082 2082 ui.write(node2str(node))
2083 2083 ui.write('\n')
2084 2084
2085 2085 @command('debugtemplate',
2086 2086 [('r', 'rev', [], _('apply template on changesets'), _('REV')),
2087 2087 ('D', 'define', [], _('define template keyword'), _('KEY=VALUE'))],
2088 2088 _('[-r REV]... [-D KEY=VALUE]... TEMPLATE'),
2089 2089 optionalrepo=True)
2090 2090 def debugtemplate(ui, repo, tmpl, **opts):
2091 2091 """parse and apply a template
2092 2092
2093 2093 If -r/--rev is given, the template is processed as a log template and
2094 2094 applied to the given changesets. Otherwise, it is processed as a generic
2095 2095 template.
2096 2096
2097 2097 Use --verbose to print the parsed tree.
2098 2098 """
2099 2099 revs = None
2100 2100 if opts['rev']:
2101 2101 if repo is None:
2102 2102 raise error.RepoError(_('there is no Mercurial repository here '
2103 2103 '(.hg not found)'))
2104 2104 revs = scmutil.revrange(repo, opts['rev'])
2105 2105
2106 2106 props = {}
2107 2107 for d in opts['define']:
2108 2108 try:
2109 2109 k, v = (e.strip() for e in d.split('=', 1))
2110 2110 if not k or k == 'ui':
2111 2111 raise ValueError
2112 2112 props[k] = v
2113 2113 except ValueError:
2114 2114 raise error.Abort(_('malformed keyword definition: %s') % d)
2115 2115
2116 2116 if ui.verbose:
2117 2117 aliases = ui.configitems('templatealias')
2118 2118 tree = templater.parse(tmpl)
2119 2119 ui.note(templater.prettyformat(tree), '\n')
2120 2120 newtree = templater.expandaliases(tree, aliases)
2121 2121 if newtree != tree:
2122 2122 ui.note(("* expanded:\n"), templater.prettyformat(newtree), '\n')
2123 2123
2124 2124 mapfile = None
2125 2125 if revs is None:
2126 2126 k = 'debugtemplate'
2127 2127 t = formatter.maketemplater(ui, k, tmpl)
2128 2128 ui.write(templater.stringify(t(k, ui=ui, **props)))
2129 2129 else:
2130 2130 displayer = cmdutil.changeset_templater(ui, repo, None, opts, tmpl,
2131 2131 mapfile, buffered=False)
2132 2132 for r in revs:
2133 2133 displayer.show(repo[r], **props)
2134 2134 displayer.close()
2135 2135
2136 2136 @command('debugupdatecaches', [])
2137 2137 def debugupdatecaches(ui, repo, *pats, **opts):
2138 2138 """warm all known caches in the repository"""
2139 2139 with repo.wlock():
2140 2140 with repo.lock():
2141 2141 repo.updatecaches()
2142 2142
2143 2143 @command('debugupgraderepo', [
2144 2144 ('o', 'optimize', [], _('extra optimization to perform'), _('NAME')),
2145 2145 ('', 'run', False, _('performs an upgrade')),
2146 2146 ])
2147 2147 def debugupgraderepo(ui, repo, run=False, optimize=None):
2148 2148 """upgrade a repository to use different features
2149 2149
2150 2150 If no arguments are specified, the repository is evaluated for upgrade
2151 2151 and a list of problems and potential optimizations is printed.
2152 2152
2153 2153 With ``--run``, a repository upgrade is performed. Behavior of the upgrade
2154 2154 can be influenced via additional arguments. More details will be provided
2155 2155 by the command output when run without ``--run``.
2156 2156
2157 2157 During the upgrade, the repository will be locked and no writes will be
2158 2158 allowed.
2159 2159
2160 2160 At the end of the upgrade, the repository may not be readable while new
2161 2161 repository data is swapped in. This window will be as long as it takes to
2162 2162 rename some directories inside the ``.hg`` directory. On most machines, this
2163 2163 should complete almost instantaneously and the chances of a consumer being
2164 2164 unable to access the repository should be low.
2165 2165 """
2166 2166 return upgrade.upgraderepo(ui, repo, run=run, optimize=optimize)
2167 2167
2168 2168 @command('debugwalk', commands.walkopts, _('[OPTION]... [FILE]...'),
2169 2169 inferrepo=True)
2170 2170 def debugwalk(ui, repo, *pats, **opts):
2171 2171 """show how files match on given patterns"""
2172 2172 m = scmutil.match(repo[None], pats, opts)
2173 2173 items = list(repo[None].walk(m))
2174 2174 if not items:
2175 2175 return
2176 2176 f = lambda fn: fn
2177 2177 if ui.configbool('ui', 'slash') and pycompat.ossep != '/':
2178 2178 f = lambda fn: util.normpath(fn)
2179 2179 fmt = 'f %%-%ds %%-%ds %%s' % (
2180 2180 max([len(abs) for abs in items]),
2181 2181 max([len(m.rel(abs)) for abs in items]))
2182 2182 for abs in items:
2183 2183 line = fmt % (abs, f(m.rel(abs)), m.exact(abs) and 'exact' or '')
2184 2184 ui.write("%s\n" % line.rstrip())
2185 2185
2186 2186 @command('debugwireargs',
2187 2187 [('', 'three', '', 'three'),
2188 2188 ('', 'four', '', 'four'),
2189 2189 ('', 'five', '', 'five'),
2190 2190 ] + commands.remoteopts,
2191 2191 _('REPO [OPTIONS]... [ONE [TWO]]'),
2192 2192 norepo=True)
2193 2193 def debugwireargs(ui, repopath, *vals, **opts):
2194 2194 repo = hg.peer(ui, opts, repopath)
2195 2195 for opt in commands.remoteopts:
2196 2196 del opts[opt[1]]
2197 2197 args = {}
2198 2198 for k, v in opts.iteritems():
2199 2199 if v:
2200 2200 args[k] = v
2201 2201 # run twice to check that we don't mess up the stream for the next command
2202 2202 res1 = repo.debugwireargs(*vals, **args)
2203 2203 res2 = repo.debugwireargs(*vals, **args)
2204 2204 ui.write("%s\n" % res1)
2205 2205 if res1 != res2:
2206 2206 ui.warn("%s\n" % res2)
@@ -1,3731 +1,3731
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import
17 17
18 18 import bz2
19 19 import calendar
20 20 import codecs
21 21 import collections
22 22 import datetime
23 23 import errno
24 24 import gc
25 25 import hashlib
26 26 import imp
27 27 import os
28 28 import platform as pyplatform
29 29 import re as remod
30 30 import shutil
31 31 import signal
32 32 import socket
33 33 import stat
34 34 import string
35 35 import subprocess
36 36 import sys
37 37 import tempfile
38 38 import textwrap
39 39 import time
40 40 import traceback
41 41 import warnings
42 42 import zlib
43 43
44 44 from . import (
45 base85,
46 45 encoding,
47 46 error,
48 47 i18n,
49 48 parsers,
50 49 policy,
51 50 pycompat,
52 51 )
53 52
53 base85 = policy.importmod(r'base85')
54 54 osutil = policy.importmod(r'osutil')
55 55
56 56 b85decode = base85.b85decode
57 57 b85encode = base85.b85encode
58 58
59 59 cookielib = pycompat.cookielib
60 60 empty = pycompat.empty
61 61 httplib = pycompat.httplib
62 62 httpserver = pycompat.httpserver
63 63 pickle = pycompat.pickle
64 64 queue = pycompat.queue
65 65 socketserver = pycompat.socketserver
66 66 stderr = pycompat.stderr
67 67 stdin = pycompat.stdin
68 68 stdout = pycompat.stdout
69 69 stringio = pycompat.stringio
70 70 urlerr = pycompat.urlerr
71 71 urlreq = pycompat.urlreq
72 72 xmlrpclib = pycompat.xmlrpclib
73 73
74 74 def isatty(fp):
75 75 try:
76 76 return fp.isatty()
77 77 except AttributeError:
78 78 return False
79 79
80 80 # glibc determines buffering on first write to stdout - if we replace a TTY
81 81 # destined stdout with a pipe destined stdout (e.g. pager), we want line
82 82 # buffering
83 83 if isatty(stdout):
84 84 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
85 85
86 86 if pycompat.osname == 'nt':
87 87 from . import windows as platform
88 88 stdout = platform.winstdout(stdout)
89 89 else:
90 90 from . import posix as platform
91 91
92 92 _ = i18n._
93 93
94 94 bindunixsocket = platform.bindunixsocket
95 95 cachestat = platform.cachestat
96 96 checkexec = platform.checkexec
97 97 checklink = platform.checklink
98 98 copymode = platform.copymode
99 99 executablepath = platform.executablepath
100 100 expandglobs = platform.expandglobs
101 101 explainexit = platform.explainexit
102 102 findexe = platform.findexe
103 103 gethgcmd = platform.gethgcmd
104 104 getuser = platform.getuser
105 105 getpid = os.getpid
106 106 groupmembers = platform.groupmembers
107 107 groupname = platform.groupname
108 108 hidewindow = platform.hidewindow
109 109 isexec = platform.isexec
110 110 isowner = platform.isowner
111 111 listdir = osutil.listdir
112 112 localpath = platform.localpath
113 113 lookupreg = platform.lookupreg
114 114 makedir = platform.makedir
115 115 nlinks = platform.nlinks
116 116 normpath = platform.normpath
117 117 normcase = platform.normcase
118 118 normcasespec = platform.normcasespec
119 119 normcasefallback = platform.normcasefallback
120 120 openhardlinks = platform.openhardlinks
121 121 oslink = platform.oslink
122 122 parsepatchoutput = platform.parsepatchoutput
123 123 pconvert = platform.pconvert
124 124 poll = platform.poll
125 125 popen = platform.popen
126 126 posixfile = platform.posixfile
127 127 quotecommand = platform.quotecommand
128 128 readpipe = platform.readpipe
129 129 rename = platform.rename
130 130 removedirs = platform.removedirs
131 131 samedevice = platform.samedevice
132 132 samefile = platform.samefile
133 133 samestat = platform.samestat
134 134 setbinary = platform.setbinary
135 135 setflags = platform.setflags
136 136 setsignalhandler = platform.setsignalhandler
137 137 shellquote = platform.shellquote
138 138 spawndetached = platform.spawndetached
139 139 split = platform.split
140 140 sshargs = platform.sshargs
141 141 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
142 142 statisexec = platform.statisexec
143 143 statislink = platform.statislink
144 144 testpid = platform.testpid
145 145 umask = platform.umask
146 146 unlink = platform.unlink
147 147 username = platform.username
148 148
149 149 try:
150 150 recvfds = osutil.recvfds
151 151 except AttributeError:
152 152 pass
153 153 try:
154 154 setprocname = osutil.setprocname
155 155 except AttributeError:
156 156 pass
157 157
158 158 # Python compatibility
159 159
160 160 _notset = object()
161 161
162 162 # disable Python's problematic floating point timestamps (issue4836)
163 163 # (Python hypocritically says you shouldn't change this behavior in
164 164 # libraries, and sure enough Mercurial is not a library.)
165 165 os.stat_float_times(False)
166 166
167 167 def safehasattr(thing, attr):
168 168 return getattr(thing, attr, _notset) is not _notset
169 169
170 170 def bitsfrom(container):
171 171 bits = 0
172 172 for bit in container:
173 173 bits |= bit
174 174 return bits
175 175
176 176 # python 2.6 still have deprecation warning enabled by default. We do not want
177 177 # to display anything to standard user so detect if we are running test and
178 178 # only use python deprecation warning in this case.
179 179 _dowarn = bool(encoding.environ.get('HGEMITWARNINGS'))
180 180 if _dowarn:
181 181 # explicitly unfilter our warning for python 2.7
182 182 #
183 183 # The option of setting PYTHONWARNINGS in the test runner was investigated.
184 184 # However, module name set through PYTHONWARNINGS was exactly matched, so
185 185 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
186 186 # makes the whole PYTHONWARNINGS thing useless for our usecase.
187 187 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'mercurial')
188 188 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext')
189 189 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext3rd')
190 190
191 191 def nouideprecwarn(msg, version, stacklevel=1):
192 192 """Issue an python native deprecation warning
193 193
194 194 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
195 195 """
196 196 if _dowarn:
197 197 msg += ("\n(compatibility will be dropped after Mercurial-%s,"
198 198 " update your code.)") % version
199 199 warnings.warn(msg, DeprecationWarning, stacklevel + 1)
200 200
201 201 DIGESTS = {
202 202 'md5': hashlib.md5,
203 203 'sha1': hashlib.sha1,
204 204 'sha512': hashlib.sha512,
205 205 }
206 206 # List of digest types from strongest to weakest
207 207 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
208 208
209 209 for k in DIGESTS_BY_STRENGTH:
210 210 assert k in DIGESTS
211 211
212 212 class digester(object):
213 213 """helper to compute digests.
214 214
215 215 This helper can be used to compute one or more digests given their name.
216 216
217 217 >>> d = digester(['md5', 'sha1'])
218 218 >>> d.update('foo')
219 219 >>> [k for k in sorted(d)]
220 220 ['md5', 'sha1']
221 221 >>> d['md5']
222 222 'acbd18db4cc2f85cedef654fccc4a4d8'
223 223 >>> d['sha1']
224 224 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
225 225 >>> digester.preferred(['md5', 'sha1'])
226 226 'sha1'
227 227 """
228 228
229 229 def __init__(self, digests, s=''):
230 230 self._hashes = {}
231 231 for k in digests:
232 232 if k not in DIGESTS:
233 233 raise Abort(_('unknown digest type: %s') % k)
234 234 self._hashes[k] = DIGESTS[k]()
235 235 if s:
236 236 self.update(s)
237 237
238 238 def update(self, data):
239 239 for h in self._hashes.values():
240 240 h.update(data)
241 241
242 242 def __getitem__(self, key):
243 243 if key not in DIGESTS:
244 244 raise Abort(_('unknown digest type: %s') % k)
245 245 return self._hashes[key].hexdigest()
246 246
247 247 def __iter__(self):
248 248 return iter(self._hashes)
249 249
250 250 @staticmethod
251 251 def preferred(supported):
252 252 """returns the strongest digest type in both supported and DIGESTS."""
253 253
254 254 for k in DIGESTS_BY_STRENGTH:
255 255 if k in supported:
256 256 return k
257 257 return None
258 258
259 259 class digestchecker(object):
260 260 """file handle wrapper that additionally checks content against a given
261 261 size and digests.
262 262
263 263 d = digestchecker(fh, size, {'md5': '...'})
264 264
265 265 When multiple digests are given, all of them are validated.
266 266 """
267 267
268 268 def __init__(self, fh, size, digests):
269 269 self._fh = fh
270 270 self._size = size
271 271 self._got = 0
272 272 self._digests = dict(digests)
273 273 self._digester = digester(self._digests.keys())
274 274
275 275 def read(self, length=-1):
276 276 content = self._fh.read(length)
277 277 self._digester.update(content)
278 278 self._got += len(content)
279 279 return content
280 280
281 281 def validate(self):
282 282 if self._size != self._got:
283 283 raise Abort(_('size mismatch: expected %d, got %d') %
284 284 (self._size, self._got))
285 285 for k, v in self._digests.items():
286 286 if v != self._digester[k]:
287 287 # i18n: first parameter is a digest name
288 288 raise Abort(_('%s mismatch: expected %s, got %s') %
289 289 (k, v, self._digester[k]))
290 290
291 291 try:
292 292 buffer = buffer
293 293 except NameError:
294 294 if not pycompat.ispy3:
295 295 def buffer(sliceable, offset=0, length=None):
296 296 if length is not None:
297 297 return sliceable[offset:offset + length]
298 298 return sliceable[offset:]
299 299 else:
300 300 def buffer(sliceable, offset=0, length=None):
301 301 if length is not None:
302 302 return memoryview(sliceable)[offset:offset + length]
303 303 return memoryview(sliceable)[offset:]
304 304
305 305 closefds = pycompat.osname == 'posix'
306 306
307 307 _chunksize = 4096
308 308
309 309 class bufferedinputpipe(object):
310 310 """a manually buffered input pipe
311 311
312 312 Python will not let us use buffered IO and lazy reading with 'polling' at
313 313 the same time. We cannot probe the buffer state and select will not detect
314 314 that data are ready to read if they are already buffered.
315 315
316 316 This class let us work around that by implementing its own buffering
317 317 (allowing efficient readline) while offering a way to know if the buffer is
318 318 empty from the output (allowing collaboration of the buffer with polling).
319 319
320 320 This class lives in the 'util' module because it makes use of the 'os'
321 321 module from the python stdlib.
322 322 """
323 323
324 324 def __init__(self, input):
325 325 self._input = input
326 326 self._buffer = []
327 327 self._eof = False
328 328 self._lenbuf = 0
329 329
330 330 @property
331 331 def hasbuffer(self):
332 332 """True is any data is currently buffered
333 333
334 334 This will be used externally a pre-step for polling IO. If there is
335 335 already data then no polling should be set in place."""
336 336 return bool(self._buffer)
337 337
338 338 @property
339 339 def closed(self):
340 340 return self._input.closed
341 341
342 342 def fileno(self):
343 343 return self._input.fileno()
344 344
345 345 def close(self):
346 346 return self._input.close()
347 347
348 348 def read(self, size):
349 349 while (not self._eof) and (self._lenbuf < size):
350 350 self._fillbuffer()
351 351 return self._frombuffer(size)
352 352
353 353 def readline(self, *args, **kwargs):
354 354 if 1 < len(self._buffer):
355 355 # this should not happen because both read and readline end with a
356 356 # _frombuffer call that collapse it.
357 357 self._buffer = [''.join(self._buffer)]
358 358 self._lenbuf = len(self._buffer[0])
359 359 lfi = -1
360 360 if self._buffer:
361 361 lfi = self._buffer[-1].find('\n')
362 362 while (not self._eof) and lfi < 0:
363 363 self._fillbuffer()
364 364 if self._buffer:
365 365 lfi = self._buffer[-1].find('\n')
366 366 size = lfi + 1
367 367 if lfi < 0: # end of file
368 368 size = self._lenbuf
369 369 elif 1 < len(self._buffer):
370 370 # we need to take previous chunks into account
371 371 size += self._lenbuf - len(self._buffer[-1])
372 372 return self._frombuffer(size)
373 373
374 374 def _frombuffer(self, size):
375 375 """return at most 'size' data from the buffer
376 376
377 377 The data are removed from the buffer."""
378 378 if size == 0 or not self._buffer:
379 379 return ''
380 380 buf = self._buffer[0]
381 381 if 1 < len(self._buffer):
382 382 buf = ''.join(self._buffer)
383 383
384 384 data = buf[:size]
385 385 buf = buf[len(data):]
386 386 if buf:
387 387 self._buffer = [buf]
388 388 self._lenbuf = len(buf)
389 389 else:
390 390 self._buffer = []
391 391 self._lenbuf = 0
392 392 return data
393 393
394 394 def _fillbuffer(self):
395 395 """read data to the buffer"""
396 396 data = os.read(self._input.fileno(), _chunksize)
397 397 if not data:
398 398 self._eof = True
399 399 else:
400 400 self._lenbuf += len(data)
401 401 self._buffer.append(data)
402 402
403 403 def popen2(cmd, env=None, newlines=False):
404 404 # Setting bufsize to -1 lets the system decide the buffer size.
405 405 # The default for bufsize is 0, meaning unbuffered. This leads to
406 406 # poor performance on Mac OS X: http://bugs.python.org/issue4194
407 407 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
408 408 close_fds=closefds,
409 409 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
410 410 universal_newlines=newlines,
411 411 env=env)
412 412 return p.stdin, p.stdout
413 413
414 414 def popen3(cmd, env=None, newlines=False):
415 415 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
416 416 return stdin, stdout, stderr
417 417
418 418 def popen4(cmd, env=None, newlines=False, bufsize=-1):
419 419 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
420 420 close_fds=closefds,
421 421 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
422 422 stderr=subprocess.PIPE,
423 423 universal_newlines=newlines,
424 424 env=env)
425 425 return p.stdin, p.stdout, p.stderr, p
426 426
427 427 def version():
428 428 """Return version information if available."""
429 429 try:
430 430 from . import __version__
431 431 return __version__.version
432 432 except ImportError:
433 433 return 'unknown'
434 434
435 435 def versiontuple(v=None, n=4):
436 436 """Parses a Mercurial version string into an N-tuple.
437 437
438 438 The version string to be parsed is specified with the ``v`` argument.
439 439 If it isn't defined, the current Mercurial version string will be parsed.
440 440
441 441 ``n`` can be 2, 3, or 4. Here is how some version strings map to
442 442 returned values:
443 443
444 444 >>> v = '3.6.1+190-df9b73d2d444'
445 445 >>> versiontuple(v, 2)
446 446 (3, 6)
447 447 >>> versiontuple(v, 3)
448 448 (3, 6, 1)
449 449 >>> versiontuple(v, 4)
450 450 (3, 6, 1, '190-df9b73d2d444')
451 451
452 452 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
453 453 (3, 6, 1, '190-df9b73d2d444+20151118')
454 454
455 455 >>> v = '3.6'
456 456 >>> versiontuple(v, 2)
457 457 (3, 6)
458 458 >>> versiontuple(v, 3)
459 459 (3, 6, None)
460 460 >>> versiontuple(v, 4)
461 461 (3, 6, None, None)
462 462
463 463 >>> v = '3.9-rc'
464 464 >>> versiontuple(v, 2)
465 465 (3, 9)
466 466 >>> versiontuple(v, 3)
467 467 (3, 9, None)
468 468 >>> versiontuple(v, 4)
469 469 (3, 9, None, 'rc')
470 470
471 471 >>> v = '3.9-rc+2-02a8fea4289b'
472 472 >>> versiontuple(v, 2)
473 473 (3, 9)
474 474 >>> versiontuple(v, 3)
475 475 (3, 9, None)
476 476 >>> versiontuple(v, 4)
477 477 (3, 9, None, 'rc+2-02a8fea4289b')
478 478 """
479 479 if not v:
480 480 v = version()
481 481 parts = remod.split('[\+-]', v, 1)
482 482 if len(parts) == 1:
483 483 vparts, extra = parts[0], None
484 484 else:
485 485 vparts, extra = parts
486 486
487 487 vints = []
488 488 for i in vparts.split('.'):
489 489 try:
490 490 vints.append(int(i))
491 491 except ValueError:
492 492 break
493 493 # (3, 6) -> (3, 6, None)
494 494 while len(vints) < 3:
495 495 vints.append(None)
496 496
497 497 if n == 2:
498 498 return (vints[0], vints[1])
499 499 if n == 3:
500 500 return (vints[0], vints[1], vints[2])
501 501 if n == 4:
502 502 return (vints[0], vints[1], vints[2], extra)
503 503
504 504 # used by parsedate
505 505 defaultdateformats = (
506 506 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
507 507 '%Y-%m-%dT%H:%M', # without seconds
508 508 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
509 509 '%Y-%m-%dT%H%M', # without seconds
510 510 '%Y-%m-%d %H:%M:%S', # our common legal variant
511 511 '%Y-%m-%d %H:%M', # without seconds
512 512 '%Y-%m-%d %H%M%S', # without :
513 513 '%Y-%m-%d %H%M', # without seconds
514 514 '%Y-%m-%d %I:%M:%S%p',
515 515 '%Y-%m-%d %H:%M',
516 516 '%Y-%m-%d %I:%M%p',
517 517 '%Y-%m-%d',
518 518 '%m-%d',
519 519 '%m/%d',
520 520 '%m/%d/%y',
521 521 '%m/%d/%Y',
522 522 '%a %b %d %H:%M:%S %Y',
523 523 '%a %b %d %I:%M:%S%p %Y',
524 524 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
525 525 '%b %d %H:%M:%S %Y',
526 526 '%b %d %I:%M:%S%p %Y',
527 527 '%b %d %H:%M:%S',
528 528 '%b %d %I:%M:%S%p',
529 529 '%b %d %H:%M',
530 530 '%b %d %I:%M%p',
531 531 '%b %d %Y',
532 532 '%b %d',
533 533 '%H:%M:%S',
534 534 '%I:%M:%S%p',
535 535 '%H:%M',
536 536 '%I:%M%p',
537 537 )
538 538
539 539 extendeddateformats = defaultdateformats + (
540 540 "%Y",
541 541 "%Y-%m",
542 542 "%b",
543 543 "%b %Y",
544 544 )
545 545
546 546 def cachefunc(func):
547 547 '''cache the result of function calls'''
548 548 # XXX doesn't handle keywords args
549 549 if func.__code__.co_argcount == 0:
550 550 cache = []
551 551 def f():
552 552 if len(cache) == 0:
553 553 cache.append(func())
554 554 return cache[0]
555 555 return f
556 556 cache = {}
557 557 if func.__code__.co_argcount == 1:
558 558 # we gain a small amount of time because
559 559 # we don't need to pack/unpack the list
560 560 def f(arg):
561 561 if arg not in cache:
562 562 cache[arg] = func(arg)
563 563 return cache[arg]
564 564 else:
565 565 def f(*args):
566 566 if args not in cache:
567 567 cache[args] = func(*args)
568 568 return cache[args]
569 569
570 570 return f
571 571
572 572 class sortdict(collections.OrderedDict):
573 573 '''a simple sorted dictionary
574 574
575 575 >>> d1 = sortdict([('a', 0), ('b', 1)])
576 576 >>> d2 = d1.copy()
577 577 >>> d2
578 578 sortdict([('a', 0), ('b', 1)])
579 579 >>> d2.update([('a', 2)])
580 580 >>> d2.keys() # should still be in last-set order
581 581 ['b', 'a']
582 582 '''
583 583
584 584 def __setitem__(self, key, value):
585 585 if key in self:
586 586 del self[key]
587 587 super(sortdict, self).__setitem__(key, value)
588 588
589 589 class _lrucachenode(object):
590 590 """A node in a doubly linked list.
591 591
592 592 Holds a reference to nodes on either side as well as a key-value
593 593 pair for the dictionary entry.
594 594 """
595 595 __slots__ = (u'next', u'prev', u'key', u'value')
596 596
597 597 def __init__(self):
598 598 self.next = None
599 599 self.prev = None
600 600
601 601 self.key = _notset
602 602 self.value = None
603 603
604 604 def markempty(self):
605 605 """Mark the node as emptied."""
606 606 self.key = _notset
607 607
608 608 class lrucachedict(object):
609 609 """Dict that caches most recent accesses and sets.
610 610
611 611 The dict consists of an actual backing dict - indexed by original
612 612 key - and a doubly linked circular list defining the order of entries in
613 613 the cache.
614 614
615 615 The head node is the newest entry in the cache. If the cache is full,
616 616 we recycle head.prev and make it the new head. Cache accesses result in
617 617 the node being moved to before the existing head and being marked as the
618 618 new head node.
619 619 """
620 620 def __init__(self, max):
621 621 self._cache = {}
622 622
623 623 self._head = head = _lrucachenode()
624 624 head.prev = head
625 625 head.next = head
626 626 self._size = 1
627 627 self._capacity = max
628 628
629 629 def __len__(self):
630 630 return len(self._cache)
631 631
632 632 def __contains__(self, k):
633 633 return k in self._cache
634 634
635 635 def __iter__(self):
636 636 # We don't have to iterate in cache order, but why not.
637 637 n = self._head
638 638 for i in range(len(self._cache)):
639 639 yield n.key
640 640 n = n.next
641 641
642 642 def __getitem__(self, k):
643 643 node = self._cache[k]
644 644 self._movetohead(node)
645 645 return node.value
646 646
647 647 def __setitem__(self, k, v):
648 648 node = self._cache.get(k)
649 649 # Replace existing value and mark as newest.
650 650 if node is not None:
651 651 node.value = v
652 652 self._movetohead(node)
653 653 return
654 654
655 655 if self._size < self._capacity:
656 656 node = self._addcapacity()
657 657 else:
658 658 # Grab the last/oldest item.
659 659 node = self._head.prev
660 660
661 661 # At capacity. Kill the old entry.
662 662 if node.key is not _notset:
663 663 del self._cache[node.key]
664 664
665 665 node.key = k
666 666 node.value = v
667 667 self._cache[k] = node
668 668 # And mark it as newest entry. No need to adjust order since it
669 669 # is already self._head.prev.
670 670 self._head = node
671 671
672 672 def __delitem__(self, k):
673 673 node = self._cache.pop(k)
674 674 node.markempty()
675 675
676 676 # Temporarily mark as newest item before re-adjusting head to make
677 677 # this node the oldest item.
678 678 self._movetohead(node)
679 679 self._head = node.next
680 680
681 681 # Additional dict methods.
682 682
683 683 def get(self, k, default=None):
684 684 try:
685 685 return self._cache[k].value
686 686 except KeyError:
687 687 return default
688 688
689 689 def clear(self):
690 690 n = self._head
691 691 while n.key is not _notset:
692 692 n.markempty()
693 693 n = n.next
694 694
695 695 self._cache.clear()
696 696
697 697 def copy(self):
698 698 result = lrucachedict(self._capacity)
699 699 n = self._head.prev
700 700 # Iterate in oldest-to-newest order, so the copy has the right ordering
701 701 for i in range(len(self._cache)):
702 702 result[n.key] = n.value
703 703 n = n.prev
704 704 return result
705 705
706 706 def _movetohead(self, node):
707 707 """Mark a node as the newest, making it the new head.
708 708
709 709 When a node is accessed, it becomes the freshest entry in the LRU
710 710 list, which is denoted by self._head.
711 711
712 712 Visually, let's make ``N`` the new head node (* denotes head):
713 713
714 714 previous/oldest <-> head <-> next/next newest
715 715
716 716 ----<->--- A* ---<->-----
717 717 | |
718 718 E <-> D <-> N <-> C <-> B
719 719
720 720 To:
721 721
722 722 ----<->--- N* ---<->-----
723 723 | |
724 724 E <-> D <-> C <-> B <-> A
725 725
726 726 This requires the following moves:
727 727
728 728 C.next = D (node.prev.next = node.next)
729 729 D.prev = C (node.next.prev = node.prev)
730 730 E.next = N (head.prev.next = node)
731 731 N.prev = E (node.prev = head.prev)
732 732 N.next = A (node.next = head)
733 733 A.prev = N (head.prev = node)
734 734 """
735 735 head = self._head
736 736 # C.next = D
737 737 node.prev.next = node.next
738 738 # D.prev = C
739 739 node.next.prev = node.prev
740 740 # N.prev = E
741 741 node.prev = head.prev
742 742 # N.next = A
743 743 # It is tempting to do just "head" here, however if node is
744 744 # adjacent to head, this will do bad things.
745 745 node.next = head.prev.next
746 746 # E.next = N
747 747 node.next.prev = node
748 748 # A.prev = N
749 749 node.prev.next = node
750 750
751 751 self._head = node
752 752
753 753 def _addcapacity(self):
754 754 """Add a node to the circular linked list.
755 755
756 756 The new node is inserted before the head node.
757 757 """
758 758 head = self._head
759 759 node = _lrucachenode()
760 760 head.prev.next = node
761 761 node.prev = head.prev
762 762 node.next = head
763 763 head.prev = node
764 764 self._size += 1
765 765 return node
766 766
767 767 def lrucachefunc(func):
768 768 '''cache most recent results of function calls'''
769 769 cache = {}
770 770 order = collections.deque()
771 771 if func.__code__.co_argcount == 1:
772 772 def f(arg):
773 773 if arg not in cache:
774 774 if len(cache) > 20:
775 775 del cache[order.popleft()]
776 776 cache[arg] = func(arg)
777 777 else:
778 778 order.remove(arg)
779 779 order.append(arg)
780 780 return cache[arg]
781 781 else:
782 782 def f(*args):
783 783 if args not in cache:
784 784 if len(cache) > 20:
785 785 del cache[order.popleft()]
786 786 cache[args] = func(*args)
787 787 else:
788 788 order.remove(args)
789 789 order.append(args)
790 790 return cache[args]
791 791
792 792 return f
793 793
794 794 class propertycache(object):
795 795 def __init__(self, func):
796 796 self.func = func
797 797 self.name = func.__name__
798 798 def __get__(self, obj, type=None):
799 799 result = self.func(obj)
800 800 self.cachevalue(obj, result)
801 801 return result
802 802
803 803 def cachevalue(self, obj, value):
804 804 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
805 805 obj.__dict__[self.name] = value
806 806
807 807 def pipefilter(s, cmd):
808 808 '''filter string S through command CMD, returning its output'''
809 809 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
810 810 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
811 811 pout, perr = p.communicate(s)
812 812 return pout
813 813
814 814 def tempfilter(s, cmd):
815 815 '''filter string S through a pair of temporary files with CMD.
816 816 CMD is used as a template to create the real command to be run,
817 817 with the strings INFILE and OUTFILE replaced by the real names of
818 818 the temporary files generated.'''
819 819 inname, outname = None, None
820 820 try:
821 821 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
822 822 fp = os.fdopen(infd, pycompat.sysstr('wb'))
823 823 fp.write(s)
824 824 fp.close()
825 825 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
826 826 os.close(outfd)
827 827 cmd = cmd.replace('INFILE', inname)
828 828 cmd = cmd.replace('OUTFILE', outname)
829 829 code = os.system(cmd)
830 830 if pycompat.sysplatform == 'OpenVMS' and code & 1:
831 831 code = 0
832 832 if code:
833 833 raise Abort(_("command '%s' failed: %s") %
834 834 (cmd, explainexit(code)))
835 835 return readfile(outname)
836 836 finally:
837 837 try:
838 838 if inname:
839 839 os.unlink(inname)
840 840 except OSError:
841 841 pass
842 842 try:
843 843 if outname:
844 844 os.unlink(outname)
845 845 except OSError:
846 846 pass
847 847
848 848 filtertable = {
849 849 'tempfile:': tempfilter,
850 850 'pipe:': pipefilter,
851 851 }
852 852
853 853 def filter(s, cmd):
854 854 "filter a string through a command that transforms its input to its output"
855 855 for name, fn in filtertable.iteritems():
856 856 if cmd.startswith(name):
857 857 return fn(s, cmd[len(name):].lstrip())
858 858 return pipefilter(s, cmd)
859 859
860 860 def binary(s):
861 861 """return true if a string is binary data"""
862 862 return bool(s and '\0' in s)
863 863
864 864 def increasingchunks(source, min=1024, max=65536):
865 865 '''return no less than min bytes per chunk while data remains,
866 866 doubling min after each chunk until it reaches max'''
867 867 def log2(x):
868 868 if not x:
869 869 return 0
870 870 i = 0
871 871 while x:
872 872 x >>= 1
873 873 i += 1
874 874 return i - 1
875 875
876 876 buf = []
877 877 blen = 0
878 878 for chunk in source:
879 879 buf.append(chunk)
880 880 blen += len(chunk)
881 881 if blen >= min:
882 882 if min < max:
883 883 min = min << 1
884 884 nmin = 1 << log2(blen)
885 885 if nmin > min:
886 886 min = nmin
887 887 if min > max:
888 888 min = max
889 889 yield ''.join(buf)
890 890 blen = 0
891 891 buf = []
892 892 if buf:
893 893 yield ''.join(buf)
894 894
895 895 Abort = error.Abort
896 896
897 897 def always(fn):
898 898 return True
899 899
900 900 def never(fn):
901 901 return False
902 902
903 903 def nogc(func):
904 904 """disable garbage collector
905 905
906 906 Python's garbage collector triggers a GC each time a certain number of
907 907 container objects (the number being defined by gc.get_threshold()) are
908 908 allocated even when marked not to be tracked by the collector. Tracking has
909 909 no effect on when GCs are triggered, only on what objects the GC looks
910 910 into. As a workaround, disable GC while building complex (huge)
911 911 containers.
912 912
913 913 This garbage collector issue have been fixed in 2.7.
914 914 """
915 915 if sys.version_info >= (2, 7):
916 916 return func
917 917 def wrapper(*args, **kwargs):
918 918 gcenabled = gc.isenabled()
919 919 gc.disable()
920 920 try:
921 921 return func(*args, **kwargs)
922 922 finally:
923 923 if gcenabled:
924 924 gc.enable()
925 925 return wrapper
926 926
927 927 def pathto(root, n1, n2):
928 928 '''return the relative path from one place to another.
929 929 root should use os.sep to separate directories
930 930 n1 should use os.sep to separate directories
931 931 n2 should use "/" to separate directories
932 932 returns an os.sep-separated path.
933 933
934 934 If n1 is a relative path, it's assumed it's
935 935 relative to root.
936 936 n2 should always be relative to root.
937 937 '''
938 938 if not n1:
939 939 return localpath(n2)
940 940 if os.path.isabs(n1):
941 941 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
942 942 return os.path.join(root, localpath(n2))
943 943 n2 = '/'.join((pconvert(root), n2))
944 944 a, b = splitpath(n1), n2.split('/')
945 945 a.reverse()
946 946 b.reverse()
947 947 while a and b and a[-1] == b[-1]:
948 948 a.pop()
949 949 b.pop()
950 950 b.reverse()
951 951 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
952 952
953 953 def mainfrozen():
954 954 """return True if we are a frozen executable.
955 955
956 956 The code supports py2exe (most common, Windows only) and tools/freeze
957 957 (portable, not much used).
958 958 """
959 959 return (safehasattr(sys, "frozen") or # new py2exe
960 960 safehasattr(sys, "importers") or # old py2exe
961 961 imp.is_frozen(u"__main__")) # tools/freeze
962 962
963 963 # the location of data files matching the source code
964 964 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
965 965 # executable version (py2exe) doesn't support __file__
966 966 datapath = os.path.dirname(pycompat.sysexecutable)
967 967 else:
968 968 datapath = os.path.dirname(pycompat.fsencode(__file__))
969 969
970 970 i18n.setdatapath(datapath)
971 971
972 972 _hgexecutable = None
973 973
974 974 def hgexecutable():
975 975 """return location of the 'hg' executable.
976 976
977 977 Defaults to $HG or 'hg' in the search path.
978 978 """
979 979 if _hgexecutable is None:
980 980 hg = encoding.environ.get('HG')
981 981 mainmod = sys.modules[pycompat.sysstr('__main__')]
982 982 if hg:
983 983 _sethgexecutable(hg)
984 984 elif mainfrozen():
985 985 if getattr(sys, 'frozen', None) == 'macosx_app':
986 986 # Env variable set by py2app
987 987 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
988 988 else:
989 989 _sethgexecutable(pycompat.sysexecutable)
990 990 elif (os.path.basename(
991 991 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
992 992 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
993 993 else:
994 994 exe = findexe('hg') or os.path.basename(sys.argv[0])
995 995 _sethgexecutable(exe)
996 996 return _hgexecutable
997 997
998 998 def _sethgexecutable(path):
999 999 """set location of the 'hg' executable"""
1000 1000 global _hgexecutable
1001 1001 _hgexecutable = path
1002 1002
1003 1003 def _isstdout(f):
1004 1004 fileno = getattr(f, 'fileno', None)
1005 1005 return fileno and fileno() == sys.__stdout__.fileno()
1006 1006
1007 1007 def shellenviron(environ=None):
1008 1008 """return environ with optional override, useful for shelling out"""
1009 1009 def py2shell(val):
1010 1010 'convert python object into string that is useful to shell'
1011 1011 if val is None or val is False:
1012 1012 return '0'
1013 1013 if val is True:
1014 1014 return '1'
1015 1015 return str(val)
1016 1016 env = dict(encoding.environ)
1017 1017 if environ:
1018 1018 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1019 1019 env['HG'] = hgexecutable()
1020 1020 return env
1021 1021
1022 1022 def system(cmd, environ=None, cwd=None, out=None):
1023 1023 '''enhanced shell command execution.
1024 1024 run with environment maybe modified, maybe in different dir.
1025 1025
1026 1026 if out is specified, it is assumed to be a file-like object that has a
1027 1027 write() method. stdout and stderr will be redirected to out.'''
1028 1028 try:
1029 1029 stdout.flush()
1030 1030 except Exception:
1031 1031 pass
1032 1032 cmd = quotecommand(cmd)
1033 1033 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1034 1034 and sys.version_info[1] < 7):
1035 1035 # subprocess kludge to work around issues in half-baked Python
1036 1036 # ports, notably bichued/python:
1037 1037 if not cwd is None:
1038 1038 os.chdir(cwd)
1039 1039 rc = os.system(cmd)
1040 1040 else:
1041 1041 env = shellenviron(environ)
1042 1042 if out is None or _isstdout(out):
1043 1043 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1044 1044 env=env, cwd=cwd)
1045 1045 else:
1046 1046 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1047 1047 env=env, cwd=cwd, stdout=subprocess.PIPE,
1048 1048 stderr=subprocess.STDOUT)
1049 1049 for line in iter(proc.stdout.readline, ''):
1050 1050 out.write(line)
1051 1051 proc.wait()
1052 1052 rc = proc.returncode
1053 1053 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1054 1054 rc = 0
1055 1055 return rc
1056 1056
1057 1057 def checksignature(func):
1058 1058 '''wrap a function with code to check for calling errors'''
1059 1059 def check(*args, **kwargs):
1060 1060 try:
1061 1061 return func(*args, **kwargs)
1062 1062 except TypeError:
1063 1063 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1064 1064 raise error.SignatureError
1065 1065 raise
1066 1066
1067 1067 return check
1068 1068
1069 1069 # a whilelist of known filesystems where hardlink works reliably
1070 1070 _hardlinkfswhitelist = {
1071 1071 'btrfs',
1072 1072 'ext2',
1073 1073 'ext3',
1074 1074 'ext4',
1075 1075 'hfs',
1076 1076 'jfs',
1077 1077 'reiserfs',
1078 1078 'tmpfs',
1079 1079 'ufs',
1080 1080 'xfs',
1081 1081 'zfs',
1082 1082 }
1083 1083
1084 1084 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1085 1085 '''copy a file, preserving mode and optionally other stat info like
1086 1086 atime/mtime
1087 1087
1088 1088 checkambig argument is used with filestat, and is useful only if
1089 1089 destination file is guarded by any lock (e.g. repo.lock or
1090 1090 repo.wlock).
1091 1091
1092 1092 copystat and checkambig should be exclusive.
1093 1093 '''
1094 1094 assert not (copystat and checkambig)
1095 1095 oldstat = None
1096 1096 if os.path.lexists(dest):
1097 1097 if checkambig:
1098 1098 oldstat = checkambig and filestat(dest)
1099 1099 unlink(dest)
1100 1100 if hardlink:
1101 1101 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1102 1102 # unless we are confident that dest is on a whitelisted filesystem.
1103 1103 try:
1104 1104 fstype = getfstype(os.path.dirname(dest))
1105 1105 except OSError:
1106 1106 fstype = None
1107 1107 if fstype not in _hardlinkfswhitelist:
1108 1108 hardlink = False
1109 1109 if hardlink:
1110 1110 try:
1111 1111 oslink(src, dest)
1112 1112 return
1113 1113 except (IOError, OSError):
1114 1114 pass # fall back to normal copy
1115 1115 if os.path.islink(src):
1116 1116 os.symlink(os.readlink(src), dest)
1117 1117 # copytime is ignored for symlinks, but in general copytime isn't needed
1118 1118 # for them anyway
1119 1119 else:
1120 1120 try:
1121 1121 shutil.copyfile(src, dest)
1122 1122 if copystat:
1123 1123 # copystat also copies mode
1124 1124 shutil.copystat(src, dest)
1125 1125 else:
1126 1126 shutil.copymode(src, dest)
1127 1127 if oldstat and oldstat.stat:
1128 1128 newstat = filestat(dest)
1129 1129 if newstat.isambig(oldstat):
1130 1130 # stat of copied file is ambiguous to original one
1131 1131 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1132 1132 os.utime(dest, (advanced, advanced))
1133 1133 except shutil.Error as inst:
1134 1134 raise Abort(str(inst))
1135 1135
1136 1136 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1137 1137 """Copy a directory tree using hardlinks if possible."""
1138 1138 num = 0
1139 1139
1140 1140 gettopic = lambda: hardlink and _('linking') or _('copying')
1141 1141
1142 1142 if os.path.isdir(src):
1143 1143 if hardlink is None:
1144 1144 hardlink = (os.stat(src).st_dev ==
1145 1145 os.stat(os.path.dirname(dst)).st_dev)
1146 1146 topic = gettopic()
1147 1147 os.mkdir(dst)
1148 1148 for name, kind in listdir(src):
1149 1149 srcname = os.path.join(src, name)
1150 1150 dstname = os.path.join(dst, name)
1151 1151 def nprog(t, pos):
1152 1152 if pos is not None:
1153 1153 return progress(t, pos + num)
1154 1154 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1155 1155 num += n
1156 1156 else:
1157 1157 if hardlink is None:
1158 1158 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1159 1159 os.stat(os.path.dirname(dst)).st_dev)
1160 1160 topic = gettopic()
1161 1161
1162 1162 if hardlink:
1163 1163 try:
1164 1164 oslink(src, dst)
1165 1165 except (IOError, OSError):
1166 1166 hardlink = False
1167 1167 shutil.copy(src, dst)
1168 1168 else:
1169 1169 shutil.copy(src, dst)
1170 1170 num += 1
1171 1171 progress(topic, num)
1172 1172 progress(topic, None)
1173 1173
1174 1174 return hardlink, num
1175 1175
1176 1176 _winreservednames = '''con prn aux nul
1177 1177 com1 com2 com3 com4 com5 com6 com7 com8 com9
1178 1178 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1179 1179 _winreservedchars = ':*?"<>|'
1180 1180 def checkwinfilename(path):
1181 1181 r'''Check that the base-relative path is a valid filename on Windows.
1182 1182 Returns None if the path is ok, or a UI string describing the problem.
1183 1183
1184 1184 >>> checkwinfilename("just/a/normal/path")
1185 1185 >>> checkwinfilename("foo/bar/con.xml")
1186 1186 "filename contains 'con', which is reserved on Windows"
1187 1187 >>> checkwinfilename("foo/con.xml/bar")
1188 1188 "filename contains 'con', which is reserved on Windows"
1189 1189 >>> checkwinfilename("foo/bar/xml.con")
1190 1190 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1191 1191 "filename contains 'AUX', which is reserved on Windows"
1192 1192 >>> checkwinfilename("foo/bar/bla:.txt")
1193 1193 "filename contains ':', which is reserved on Windows"
1194 1194 >>> checkwinfilename("foo/bar/b\07la.txt")
1195 1195 "filename contains '\\x07', which is invalid on Windows"
1196 1196 >>> checkwinfilename("foo/bar/bla ")
1197 1197 "filename ends with ' ', which is not allowed on Windows"
1198 1198 >>> checkwinfilename("../bar")
1199 1199 >>> checkwinfilename("foo\\")
1200 1200 "filename ends with '\\', which is invalid on Windows"
1201 1201 >>> checkwinfilename("foo\\/bar")
1202 1202 "directory name ends with '\\', which is invalid on Windows"
1203 1203 '''
1204 1204 if path.endswith('\\'):
1205 1205 return _("filename ends with '\\', which is invalid on Windows")
1206 1206 if '\\/' in path:
1207 1207 return _("directory name ends with '\\', which is invalid on Windows")
1208 1208 for n in path.replace('\\', '/').split('/'):
1209 1209 if not n:
1210 1210 continue
1211 1211 for c in pycompat.bytestr(n):
1212 1212 if c in _winreservedchars:
1213 1213 return _("filename contains '%s', which is reserved "
1214 1214 "on Windows") % c
1215 1215 if ord(c) <= 31:
1216 1216 return _("filename contains %r, which is invalid "
1217 1217 "on Windows") % c
1218 1218 base = n.split('.')[0]
1219 1219 if base and base.lower() in _winreservednames:
1220 1220 return _("filename contains '%s', which is reserved "
1221 1221 "on Windows") % base
1222 1222 t = n[-1]
1223 1223 if t in '. ' and n not in '..':
1224 1224 return _("filename ends with '%s', which is not allowed "
1225 1225 "on Windows") % t
1226 1226
1227 1227 if pycompat.osname == 'nt':
1228 1228 checkosfilename = checkwinfilename
1229 1229 timer = time.clock
1230 1230 else:
1231 1231 checkosfilename = platform.checkosfilename
1232 1232 timer = time.time
1233 1233
1234 1234 if safehasattr(time, "perf_counter"):
1235 1235 timer = time.perf_counter
1236 1236
1237 1237 def makelock(info, pathname):
1238 1238 try:
1239 1239 return os.symlink(info, pathname)
1240 1240 except OSError as why:
1241 1241 if why.errno == errno.EEXIST:
1242 1242 raise
1243 1243 except AttributeError: # no symlink in os
1244 1244 pass
1245 1245
1246 1246 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1247 1247 os.write(ld, info)
1248 1248 os.close(ld)
1249 1249
1250 1250 def readlock(pathname):
1251 1251 try:
1252 1252 return os.readlink(pathname)
1253 1253 except OSError as why:
1254 1254 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1255 1255 raise
1256 1256 except AttributeError: # no symlink in os
1257 1257 pass
1258 1258 fp = posixfile(pathname)
1259 1259 r = fp.read()
1260 1260 fp.close()
1261 1261 return r
1262 1262
1263 1263 def fstat(fp):
1264 1264 '''stat file object that may not have fileno method.'''
1265 1265 try:
1266 1266 return os.fstat(fp.fileno())
1267 1267 except AttributeError:
1268 1268 return os.stat(fp.name)
1269 1269
1270 1270 # File system features
1271 1271
1272 1272 def fscasesensitive(path):
1273 1273 """
1274 1274 Return true if the given path is on a case-sensitive filesystem
1275 1275
1276 1276 Requires a path (like /foo/.hg) ending with a foldable final
1277 1277 directory component.
1278 1278 """
1279 1279 s1 = os.lstat(path)
1280 1280 d, b = os.path.split(path)
1281 1281 b2 = b.upper()
1282 1282 if b == b2:
1283 1283 b2 = b.lower()
1284 1284 if b == b2:
1285 1285 return True # no evidence against case sensitivity
1286 1286 p2 = os.path.join(d, b2)
1287 1287 try:
1288 1288 s2 = os.lstat(p2)
1289 1289 if s2 == s1:
1290 1290 return False
1291 1291 return True
1292 1292 except OSError:
1293 1293 return True
1294 1294
1295 1295 try:
1296 1296 import re2
1297 1297 _re2 = None
1298 1298 except ImportError:
1299 1299 _re2 = False
1300 1300
1301 1301 class _re(object):
1302 1302 def _checkre2(self):
1303 1303 global _re2
1304 1304 try:
1305 1305 # check if match works, see issue3964
1306 1306 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1307 1307 except ImportError:
1308 1308 _re2 = False
1309 1309
1310 1310 def compile(self, pat, flags=0):
1311 1311 '''Compile a regular expression, using re2 if possible
1312 1312
1313 1313 For best performance, use only re2-compatible regexp features. The
1314 1314 only flags from the re module that are re2-compatible are
1315 1315 IGNORECASE and MULTILINE.'''
1316 1316 if _re2 is None:
1317 1317 self._checkre2()
1318 1318 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1319 1319 if flags & remod.IGNORECASE:
1320 1320 pat = '(?i)' + pat
1321 1321 if flags & remod.MULTILINE:
1322 1322 pat = '(?m)' + pat
1323 1323 try:
1324 1324 return re2.compile(pat)
1325 1325 except re2.error:
1326 1326 pass
1327 1327 return remod.compile(pat, flags)
1328 1328
1329 1329 @propertycache
1330 1330 def escape(self):
1331 1331 '''Return the version of escape corresponding to self.compile.
1332 1332
1333 1333 This is imperfect because whether re2 or re is used for a particular
1334 1334 function depends on the flags, etc, but it's the best we can do.
1335 1335 '''
1336 1336 global _re2
1337 1337 if _re2 is None:
1338 1338 self._checkre2()
1339 1339 if _re2:
1340 1340 return re2.escape
1341 1341 else:
1342 1342 return remod.escape
1343 1343
1344 1344 re = _re()
1345 1345
1346 1346 _fspathcache = {}
1347 1347 def fspath(name, root):
1348 1348 '''Get name in the case stored in the filesystem
1349 1349
1350 1350 The name should be relative to root, and be normcase-ed for efficiency.
1351 1351
1352 1352 Note that this function is unnecessary, and should not be
1353 1353 called, for case-sensitive filesystems (simply because it's expensive).
1354 1354
1355 1355 The root should be normcase-ed, too.
1356 1356 '''
1357 1357 def _makefspathcacheentry(dir):
1358 1358 return dict((normcase(n), n) for n in os.listdir(dir))
1359 1359
1360 1360 seps = pycompat.ossep
1361 1361 if pycompat.osaltsep:
1362 1362 seps = seps + pycompat.osaltsep
1363 1363 # Protect backslashes. This gets silly very quickly.
1364 1364 seps.replace('\\','\\\\')
1365 1365 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1366 1366 dir = os.path.normpath(root)
1367 1367 result = []
1368 1368 for part, sep in pattern.findall(name):
1369 1369 if sep:
1370 1370 result.append(sep)
1371 1371 continue
1372 1372
1373 1373 if dir not in _fspathcache:
1374 1374 _fspathcache[dir] = _makefspathcacheentry(dir)
1375 1375 contents = _fspathcache[dir]
1376 1376
1377 1377 found = contents.get(part)
1378 1378 if not found:
1379 1379 # retry "once per directory" per "dirstate.walk" which
1380 1380 # may take place for each patches of "hg qpush", for example
1381 1381 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1382 1382 found = contents.get(part)
1383 1383
1384 1384 result.append(found or part)
1385 1385 dir = os.path.join(dir, part)
1386 1386
1387 1387 return ''.join(result)
1388 1388
1389 1389 def getfstype(dirpath):
1390 1390 '''Get the filesystem type name from a directory (best-effort)
1391 1391
1392 1392 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1393 1393 '''
1394 1394 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1395 1395
1396 1396 def checknlink(testfile):
1397 1397 '''check whether hardlink count reporting works properly'''
1398 1398
1399 1399 # testfile may be open, so we need a separate file for checking to
1400 1400 # work around issue2543 (or testfile may get lost on Samba shares)
1401 1401 f1 = testfile + ".hgtmp1"
1402 1402 if os.path.lexists(f1):
1403 1403 return False
1404 1404 try:
1405 1405 posixfile(f1, 'w').close()
1406 1406 except IOError:
1407 1407 try:
1408 1408 os.unlink(f1)
1409 1409 except OSError:
1410 1410 pass
1411 1411 return False
1412 1412
1413 1413 f2 = testfile + ".hgtmp2"
1414 1414 fd = None
1415 1415 try:
1416 1416 oslink(f1, f2)
1417 1417 # nlinks() may behave differently for files on Windows shares if
1418 1418 # the file is open.
1419 1419 fd = posixfile(f2)
1420 1420 return nlinks(f2) > 1
1421 1421 except OSError:
1422 1422 return False
1423 1423 finally:
1424 1424 if fd is not None:
1425 1425 fd.close()
1426 1426 for f in (f1, f2):
1427 1427 try:
1428 1428 os.unlink(f)
1429 1429 except OSError:
1430 1430 pass
1431 1431
1432 1432 def endswithsep(path):
1433 1433 '''Check path ends with os.sep or os.altsep.'''
1434 1434 return (path.endswith(pycompat.ossep)
1435 1435 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1436 1436
1437 1437 def splitpath(path):
1438 1438 '''Split path by os.sep.
1439 1439 Note that this function does not use os.altsep because this is
1440 1440 an alternative of simple "xxx.split(os.sep)".
1441 1441 It is recommended to use os.path.normpath() before using this
1442 1442 function if need.'''
1443 1443 return path.split(pycompat.ossep)
1444 1444
1445 1445 def gui():
1446 1446 '''Are we running in a GUI?'''
1447 1447 if pycompat.sysplatform == 'darwin':
1448 1448 if 'SSH_CONNECTION' in encoding.environ:
1449 1449 # handle SSH access to a box where the user is logged in
1450 1450 return False
1451 1451 elif getattr(osutil, 'isgui', None):
1452 1452 # check if a CoreGraphics session is available
1453 1453 return osutil.isgui()
1454 1454 else:
1455 1455 # pure build; use a safe default
1456 1456 return True
1457 1457 else:
1458 1458 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1459 1459
1460 1460 def mktempcopy(name, emptyok=False, createmode=None):
1461 1461 """Create a temporary file with the same contents from name
1462 1462
1463 1463 The permission bits are copied from the original file.
1464 1464
1465 1465 If the temporary file is going to be truncated immediately, you
1466 1466 can use emptyok=True as an optimization.
1467 1467
1468 1468 Returns the name of the temporary file.
1469 1469 """
1470 1470 d, fn = os.path.split(name)
1471 1471 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1472 1472 os.close(fd)
1473 1473 # Temporary files are created with mode 0600, which is usually not
1474 1474 # what we want. If the original file already exists, just copy
1475 1475 # its mode. Otherwise, manually obey umask.
1476 1476 copymode(name, temp, createmode)
1477 1477 if emptyok:
1478 1478 return temp
1479 1479 try:
1480 1480 try:
1481 1481 ifp = posixfile(name, "rb")
1482 1482 except IOError as inst:
1483 1483 if inst.errno == errno.ENOENT:
1484 1484 return temp
1485 1485 if not getattr(inst, 'filename', None):
1486 1486 inst.filename = name
1487 1487 raise
1488 1488 ofp = posixfile(temp, "wb")
1489 1489 for chunk in filechunkiter(ifp):
1490 1490 ofp.write(chunk)
1491 1491 ifp.close()
1492 1492 ofp.close()
1493 1493 except: # re-raises
1494 1494 try: os.unlink(temp)
1495 1495 except OSError: pass
1496 1496 raise
1497 1497 return temp
1498 1498
1499 1499 class filestat(object):
1500 1500 """help to exactly detect change of a file
1501 1501
1502 1502 'stat' attribute is result of 'os.stat()' if specified 'path'
1503 1503 exists. Otherwise, it is None. This can avoid preparative
1504 1504 'exists()' examination on client side of this class.
1505 1505 """
1506 1506 def __init__(self, path):
1507 1507 try:
1508 1508 self.stat = os.stat(path)
1509 1509 except OSError as err:
1510 1510 if err.errno != errno.ENOENT:
1511 1511 raise
1512 1512 self.stat = None
1513 1513
1514 1514 __hash__ = object.__hash__
1515 1515
1516 1516 def __eq__(self, old):
1517 1517 try:
1518 1518 # if ambiguity between stat of new and old file is
1519 1519 # avoided, comparison of size, ctime and mtime is enough
1520 1520 # to exactly detect change of a file regardless of platform
1521 1521 return (self.stat.st_size == old.stat.st_size and
1522 1522 self.stat.st_ctime == old.stat.st_ctime and
1523 1523 self.stat.st_mtime == old.stat.st_mtime)
1524 1524 except AttributeError:
1525 1525 return False
1526 1526
1527 1527 def isambig(self, old):
1528 1528 """Examine whether new (= self) stat is ambiguous against old one
1529 1529
1530 1530 "S[N]" below means stat of a file at N-th change:
1531 1531
1532 1532 - S[n-1].ctime < S[n].ctime: can detect change of a file
1533 1533 - S[n-1].ctime == S[n].ctime
1534 1534 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1535 1535 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1536 1536 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1537 1537 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1538 1538
1539 1539 Case (*2) above means that a file was changed twice or more at
1540 1540 same time in sec (= S[n-1].ctime), and comparison of timestamp
1541 1541 is ambiguous.
1542 1542
1543 1543 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1544 1544 timestamp is ambiguous".
1545 1545
1546 1546 But advancing mtime only in case (*2) doesn't work as
1547 1547 expected, because naturally advanced S[n].mtime in case (*1)
1548 1548 might be equal to manually advanced S[n-1 or earlier].mtime.
1549 1549
1550 1550 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1551 1551 treated as ambiguous regardless of mtime, to avoid overlooking
1552 1552 by confliction between such mtime.
1553 1553
1554 1554 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1555 1555 S[n].mtime", even if size of a file isn't changed.
1556 1556 """
1557 1557 try:
1558 1558 return (self.stat.st_ctime == old.stat.st_ctime)
1559 1559 except AttributeError:
1560 1560 return False
1561 1561
1562 1562 def avoidambig(self, path, old):
1563 1563 """Change file stat of specified path to avoid ambiguity
1564 1564
1565 1565 'old' should be previous filestat of 'path'.
1566 1566
1567 1567 This skips avoiding ambiguity, if a process doesn't have
1568 1568 appropriate privileges for 'path'.
1569 1569 """
1570 1570 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1571 1571 try:
1572 1572 os.utime(path, (advanced, advanced))
1573 1573 except OSError as inst:
1574 1574 if inst.errno == errno.EPERM:
1575 1575 # utime() on the file created by another user causes EPERM,
1576 1576 # if a process doesn't have appropriate privileges
1577 1577 return
1578 1578 raise
1579 1579
1580 1580 def __ne__(self, other):
1581 1581 return not self == other
1582 1582
1583 1583 class atomictempfile(object):
1584 1584 '''writable file object that atomically updates a file
1585 1585
1586 1586 All writes will go to a temporary copy of the original file. Call
1587 1587 close() when you are done writing, and atomictempfile will rename
1588 1588 the temporary copy to the original name, making the changes
1589 1589 visible. If the object is destroyed without being closed, all your
1590 1590 writes are discarded.
1591 1591
1592 1592 checkambig argument of constructor is used with filestat, and is
1593 1593 useful only if target file is guarded by any lock (e.g. repo.lock
1594 1594 or repo.wlock).
1595 1595 '''
1596 1596 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1597 1597 self.__name = name # permanent name
1598 1598 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1599 1599 createmode=createmode)
1600 1600 self._fp = posixfile(self._tempname, mode)
1601 1601 self._checkambig = checkambig
1602 1602
1603 1603 # delegated methods
1604 1604 self.read = self._fp.read
1605 1605 self.write = self._fp.write
1606 1606 self.seek = self._fp.seek
1607 1607 self.tell = self._fp.tell
1608 1608 self.fileno = self._fp.fileno
1609 1609
1610 1610 def close(self):
1611 1611 if not self._fp.closed:
1612 1612 self._fp.close()
1613 1613 filename = localpath(self.__name)
1614 1614 oldstat = self._checkambig and filestat(filename)
1615 1615 if oldstat and oldstat.stat:
1616 1616 rename(self._tempname, filename)
1617 1617 newstat = filestat(filename)
1618 1618 if newstat.isambig(oldstat):
1619 1619 # stat of changed file is ambiguous to original one
1620 1620 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1621 1621 os.utime(filename, (advanced, advanced))
1622 1622 else:
1623 1623 rename(self._tempname, filename)
1624 1624
1625 1625 def discard(self):
1626 1626 if not self._fp.closed:
1627 1627 try:
1628 1628 os.unlink(self._tempname)
1629 1629 except OSError:
1630 1630 pass
1631 1631 self._fp.close()
1632 1632
1633 1633 def __del__(self):
1634 1634 if safehasattr(self, '_fp'): # constructor actually did something
1635 1635 self.discard()
1636 1636
1637 1637 def __enter__(self):
1638 1638 return self
1639 1639
1640 1640 def __exit__(self, exctype, excvalue, traceback):
1641 1641 if exctype is not None:
1642 1642 self.discard()
1643 1643 else:
1644 1644 self.close()
1645 1645
1646 1646 def unlinkpath(f, ignoremissing=False):
1647 1647 """unlink and remove the directory if it is empty"""
1648 1648 if ignoremissing:
1649 1649 tryunlink(f)
1650 1650 else:
1651 1651 unlink(f)
1652 1652 # try removing directories that might now be empty
1653 1653 try:
1654 1654 removedirs(os.path.dirname(f))
1655 1655 except OSError:
1656 1656 pass
1657 1657
1658 1658 def tryunlink(f):
1659 1659 """Attempt to remove a file, ignoring ENOENT errors."""
1660 1660 try:
1661 1661 unlink(f)
1662 1662 except OSError as e:
1663 1663 if e.errno != errno.ENOENT:
1664 1664 raise
1665 1665
1666 1666 def makedirs(name, mode=None, notindexed=False):
1667 1667 """recursive directory creation with parent mode inheritance
1668 1668
1669 1669 Newly created directories are marked as "not to be indexed by
1670 1670 the content indexing service", if ``notindexed`` is specified
1671 1671 for "write" mode access.
1672 1672 """
1673 1673 try:
1674 1674 makedir(name, notindexed)
1675 1675 except OSError as err:
1676 1676 if err.errno == errno.EEXIST:
1677 1677 return
1678 1678 if err.errno != errno.ENOENT or not name:
1679 1679 raise
1680 1680 parent = os.path.dirname(os.path.abspath(name))
1681 1681 if parent == name:
1682 1682 raise
1683 1683 makedirs(parent, mode, notindexed)
1684 1684 try:
1685 1685 makedir(name, notindexed)
1686 1686 except OSError as err:
1687 1687 # Catch EEXIST to handle races
1688 1688 if err.errno == errno.EEXIST:
1689 1689 return
1690 1690 raise
1691 1691 if mode is not None:
1692 1692 os.chmod(name, mode)
1693 1693
1694 1694 def readfile(path):
1695 1695 with open(path, 'rb') as fp:
1696 1696 return fp.read()
1697 1697
1698 1698 def writefile(path, text):
1699 1699 with open(path, 'wb') as fp:
1700 1700 fp.write(text)
1701 1701
1702 1702 def appendfile(path, text):
1703 1703 with open(path, 'ab') as fp:
1704 1704 fp.write(text)
1705 1705
1706 1706 class chunkbuffer(object):
1707 1707 """Allow arbitrary sized chunks of data to be efficiently read from an
1708 1708 iterator over chunks of arbitrary size."""
1709 1709
1710 1710 def __init__(self, in_iter):
1711 1711 """in_iter is the iterator that's iterating over the input chunks."""
1712 1712 def splitbig(chunks):
1713 1713 for chunk in chunks:
1714 1714 if len(chunk) > 2**20:
1715 1715 pos = 0
1716 1716 while pos < len(chunk):
1717 1717 end = pos + 2 ** 18
1718 1718 yield chunk[pos:end]
1719 1719 pos = end
1720 1720 else:
1721 1721 yield chunk
1722 1722 self.iter = splitbig(in_iter)
1723 1723 self._queue = collections.deque()
1724 1724 self._chunkoffset = 0
1725 1725
1726 1726 def read(self, l=None):
1727 1727 """Read L bytes of data from the iterator of chunks of data.
1728 1728 Returns less than L bytes if the iterator runs dry.
1729 1729
1730 1730 If size parameter is omitted, read everything"""
1731 1731 if l is None:
1732 1732 return ''.join(self.iter)
1733 1733
1734 1734 left = l
1735 1735 buf = []
1736 1736 queue = self._queue
1737 1737 while left > 0:
1738 1738 # refill the queue
1739 1739 if not queue:
1740 1740 target = 2**18
1741 1741 for chunk in self.iter:
1742 1742 queue.append(chunk)
1743 1743 target -= len(chunk)
1744 1744 if target <= 0:
1745 1745 break
1746 1746 if not queue:
1747 1747 break
1748 1748
1749 1749 # The easy way to do this would be to queue.popleft(), modify the
1750 1750 # chunk (if necessary), then queue.appendleft(). However, for cases
1751 1751 # where we read partial chunk content, this incurs 2 dequeue
1752 1752 # mutations and creates a new str for the remaining chunk in the
1753 1753 # queue. Our code below avoids this overhead.
1754 1754
1755 1755 chunk = queue[0]
1756 1756 chunkl = len(chunk)
1757 1757 offset = self._chunkoffset
1758 1758
1759 1759 # Use full chunk.
1760 1760 if offset == 0 and left >= chunkl:
1761 1761 left -= chunkl
1762 1762 queue.popleft()
1763 1763 buf.append(chunk)
1764 1764 # self._chunkoffset remains at 0.
1765 1765 continue
1766 1766
1767 1767 chunkremaining = chunkl - offset
1768 1768
1769 1769 # Use all of unconsumed part of chunk.
1770 1770 if left >= chunkremaining:
1771 1771 left -= chunkremaining
1772 1772 queue.popleft()
1773 1773 # offset == 0 is enabled by block above, so this won't merely
1774 1774 # copy via ``chunk[0:]``.
1775 1775 buf.append(chunk[offset:])
1776 1776 self._chunkoffset = 0
1777 1777
1778 1778 # Partial chunk needed.
1779 1779 else:
1780 1780 buf.append(chunk[offset:offset + left])
1781 1781 self._chunkoffset += left
1782 1782 left -= chunkremaining
1783 1783
1784 1784 return ''.join(buf)
1785 1785
1786 1786 def filechunkiter(f, size=131072, limit=None):
1787 1787 """Create a generator that produces the data in the file size
1788 1788 (default 131072) bytes at a time, up to optional limit (default is
1789 1789 to read all data). Chunks may be less than size bytes if the
1790 1790 chunk is the last chunk in the file, or the file is a socket or
1791 1791 some other type of file that sometimes reads less data than is
1792 1792 requested."""
1793 1793 assert size >= 0
1794 1794 assert limit is None or limit >= 0
1795 1795 while True:
1796 1796 if limit is None:
1797 1797 nbytes = size
1798 1798 else:
1799 1799 nbytes = min(limit, size)
1800 1800 s = nbytes and f.read(nbytes)
1801 1801 if not s:
1802 1802 break
1803 1803 if limit:
1804 1804 limit -= len(s)
1805 1805 yield s
1806 1806
1807 1807 def makedate(timestamp=None):
1808 1808 '''Return a unix timestamp (or the current time) as a (unixtime,
1809 1809 offset) tuple based off the local timezone.'''
1810 1810 if timestamp is None:
1811 1811 timestamp = time.time()
1812 1812 if timestamp < 0:
1813 1813 hint = _("check your clock")
1814 1814 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1815 1815 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1816 1816 datetime.datetime.fromtimestamp(timestamp))
1817 1817 tz = delta.days * 86400 + delta.seconds
1818 1818 return timestamp, tz
1819 1819
1820 1820 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1821 1821 """represent a (unixtime, offset) tuple as a localized time.
1822 1822 unixtime is seconds since the epoch, and offset is the time zone's
1823 1823 number of seconds away from UTC.
1824 1824
1825 1825 >>> datestr((0, 0))
1826 1826 'Thu Jan 01 00:00:00 1970 +0000'
1827 1827 >>> datestr((42, 0))
1828 1828 'Thu Jan 01 00:00:42 1970 +0000'
1829 1829 >>> datestr((-42, 0))
1830 1830 'Wed Dec 31 23:59:18 1969 +0000'
1831 1831 >>> datestr((0x7fffffff, 0))
1832 1832 'Tue Jan 19 03:14:07 2038 +0000'
1833 1833 >>> datestr((-0x80000000, 0))
1834 1834 'Fri Dec 13 20:45:52 1901 +0000'
1835 1835 """
1836 1836 t, tz = date or makedate()
1837 1837 if "%1" in format or "%2" in format or "%z" in format:
1838 1838 sign = (tz > 0) and "-" or "+"
1839 1839 minutes = abs(tz) // 60
1840 1840 q, r = divmod(minutes, 60)
1841 1841 format = format.replace("%z", "%1%2")
1842 1842 format = format.replace("%1", "%c%02d" % (sign, q))
1843 1843 format = format.replace("%2", "%02d" % r)
1844 1844 d = t - tz
1845 1845 if d > 0x7fffffff:
1846 1846 d = 0x7fffffff
1847 1847 elif d < -0x80000000:
1848 1848 d = -0x80000000
1849 1849 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1850 1850 # because they use the gmtime() system call which is buggy on Windows
1851 1851 # for negative values.
1852 1852 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1853 1853 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1854 1854 return s
1855 1855
1856 1856 def shortdate(date=None):
1857 1857 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1858 1858 return datestr(date, format='%Y-%m-%d')
1859 1859
1860 1860 def parsetimezone(s):
1861 1861 """find a trailing timezone, if any, in string, and return a
1862 1862 (offset, remainder) pair"""
1863 1863
1864 1864 if s.endswith("GMT") or s.endswith("UTC"):
1865 1865 return 0, s[:-3].rstrip()
1866 1866
1867 1867 # Unix-style timezones [+-]hhmm
1868 1868 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1869 1869 sign = (s[-5] == "+") and 1 or -1
1870 1870 hours = int(s[-4:-2])
1871 1871 minutes = int(s[-2:])
1872 1872 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1873 1873
1874 1874 # ISO8601 trailing Z
1875 1875 if s.endswith("Z") and s[-2:-1].isdigit():
1876 1876 return 0, s[:-1]
1877 1877
1878 1878 # ISO8601-style [+-]hh:mm
1879 1879 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1880 1880 s[-5:-3].isdigit() and s[-2:].isdigit()):
1881 1881 sign = (s[-6] == "+") and 1 or -1
1882 1882 hours = int(s[-5:-3])
1883 1883 minutes = int(s[-2:])
1884 1884 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1885 1885
1886 1886 return None, s
1887 1887
1888 1888 def strdate(string, format, defaults=None):
1889 1889 """parse a localized time string and return a (unixtime, offset) tuple.
1890 1890 if the string cannot be parsed, ValueError is raised."""
1891 1891 if defaults is None:
1892 1892 defaults = {}
1893 1893
1894 1894 # NOTE: unixtime = localunixtime + offset
1895 1895 offset, date = parsetimezone(string)
1896 1896
1897 1897 # add missing elements from defaults
1898 1898 usenow = False # default to using biased defaults
1899 1899 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1900 1900 part = pycompat.bytestr(part)
1901 1901 found = [True for p in part if ("%"+p) in format]
1902 1902 if not found:
1903 1903 date += "@" + defaults[part][usenow]
1904 1904 format += "@%" + part[0]
1905 1905 else:
1906 1906 # We've found a specific time element, less specific time
1907 1907 # elements are relative to today
1908 1908 usenow = True
1909 1909
1910 1910 timetuple = time.strptime(encoding.strfromlocal(date),
1911 1911 encoding.strfromlocal(format))
1912 1912 localunixtime = int(calendar.timegm(timetuple))
1913 1913 if offset is None:
1914 1914 # local timezone
1915 1915 unixtime = int(time.mktime(timetuple))
1916 1916 offset = unixtime - localunixtime
1917 1917 else:
1918 1918 unixtime = localunixtime + offset
1919 1919 return unixtime, offset
1920 1920
1921 1921 def parsedate(date, formats=None, bias=None):
1922 1922 """parse a localized date/time and return a (unixtime, offset) tuple.
1923 1923
1924 1924 The date may be a "unixtime offset" string or in one of the specified
1925 1925 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1926 1926
1927 1927 >>> parsedate(' today ') == parsedate(\
1928 1928 datetime.date.today().strftime('%b %d'))
1929 1929 True
1930 1930 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1931 1931 datetime.timedelta(days=1)\
1932 1932 ).strftime('%b %d'))
1933 1933 True
1934 1934 >>> now, tz = makedate()
1935 1935 >>> strnow, strtz = parsedate('now')
1936 1936 >>> (strnow - now) < 1
1937 1937 True
1938 1938 >>> tz == strtz
1939 1939 True
1940 1940 """
1941 1941 if bias is None:
1942 1942 bias = {}
1943 1943 if not date:
1944 1944 return 0, 0
1945 1945 if isinstance(date, tuple) and len(date) == 2:
1946 1946 return date
1947 1947 if not formats:
1948 1948 formats = defaultdateformats
1949 1949 date = date.strip()
1950 1950
1951 1951 if date == 'now' or date == _('now'):
1952 1952 return makedate()
1953 1953 if date == 'today' or date == _('today'):
1954 1954 date = datetime.date.today().strftime('%b %d')
1955 1955 elif date == 'yesterday' or date == _('yesterday'):
1956 1956 date = (datetime.date.today() -
1957 1957 datetime.timedelta(days=1)).strftime('%b %d')
1958 1958
1959 1959 try:
1960 1960 when, offset = map(int, date.split(' '))
1961 1961 except ValueError:
1962 1962 # fill out defaults
1963 1963 now = makedate()
1964 1964 defaults = {}
1965 1965 for part in ("d", "mb", "yY", "HI", "M", "S"):
1966 1966 # this piece is for rounding the specific end of unknowns
1967 1967 b = bias.get(part)
1968 1968 if b is None:
1969 1969 if part[0:1] in "HMS":
1970 1970 b = "00"
1971 1971 else:
1972 1972 b = "0"
1973 1973
1974 1974 # this piece is for matching the generic end to today's date
1975 1975 n = datestr(now, "%" + part[0:1])
1976 1976
1977 1977 defaults[part] = (b, n)
1978 1978
1979 1979 for format in formats:
1980 1980 try:
1981 1981 when, offset = strdate(date, format, defaults)
1982 1982 except (ValueError, OverflowError):
1983 1983 pass
1984 1984 else:
1985 1985 break
1986 1986 else:
1987 1987 raise Abort(_('invalid date: %r') % date)
1988 1988 # validate explicit (probably user-specified) date and
1989 1989 # time zone offset. values must fit in signed 32 bits for
1990 1990 # current 32-bit linux runtimes. timezones go from UTC-12
1991 1991 # to UTC+14
1992 1992 if when < -0x80000000 or when > 0x7fffffff:
1993 1993 raise Abort(_('date exceeds 32 bits: %d') % when)
1994 1994 if offset < -50400 or offset > 43200:
1995 1995 raise Abort(_('impossible time zone offset: %d') % offset)
1996 1996 return when, offset
1997 1997
1998 1998 def matchdate(date):
1999 1999 """Return a function that matches a given date match specifier
2000 2000
2001 2001 Formats include:
2002 2002
2003 2003 '{date}' match a given date to the accuracy provided
2004 2004
2005 2005 '<{date}' on or before a given date
2006 2006
2007 2007 '>{date}' on or after a given date
2008 2008
2009 2009 >>> p1 = parsedate("10:29:59")
2010 2010 >>> p2 = parsedate("10:30:00")
2011 2011 >>> p3 = parsedate("10:30:59")
2012 2012 >>> p4 = parsedate("10:31:00")
2013 2013 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2014 2014 >>> f = matchdate("10:30")
2015 2015 >>> f(p1[0])
2016 2016 False
2017 2017 >>> f(p2[0])
2018 2018 True
2019 2019 >>> f(p3[0])
2020 2020 True
2021 2021 >>> f(p4[0])
2022 2022 False
2023 2023 >>> f(p5[0])
2024 2024 False
2025 2025 """
2026 2026
2027 2027 def lower(date):
2028 2028 d = {'mb': "1", 'd': "1"}
2029 2029 return parsedate(date, extendeddateformats, d)[0]
2030 2030
2031 2031 def upper(date):
2032 2032 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2033 2033 for days in ("31", "30", "29"):
2034 2034 try:
2035 2035 d["d"] = days
2036 2036 return parsedate(date, extendeddateformats, d)[0]
2037 2037 except Abort:
2038 2038 pass
2039 2039 d["d"] = "28"
2040 2040 return parsedate(date, extendeddateformats, d)[0]
2041 2041
2042 2042 date = date.strip()
2043 2043
2044 2044 if not date:
2045 2045 raise Abort(_("dates cannot consist entirely of whitespace"))
2046 2046 elif date[0] == "<":
2047 2047 if not date[1:]:
2048 2048 raise Abort(_("invalid day spec, use '<DATE'"))
2049 2049 when = upper(date[1:])
2050 2050 return lambda x: x <= when
2051 2051 elif date[0] == ">":
2052 2052 if not date[1:]:
2053 2053 raise Abort(_("invalid day spec, use '>DATE'"))
2054 2054 when = lower(date[1:])
2055 2055 return lambda x: x >= when
2056 2056 elif date[0] == "-":
2057 2057 try:
2058 2058 days = int(date[1:])
2059 2059 except ValueError:
2060 2060 raise Abort(_("invalid day spec: %s") % date[1:])
2061 2061 if days < 0:
2062 2062 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2063 2063 % date[1:])
2064 2064 when = makedate()[0] - days * 3600 * 24
2065 2065 return lambda x: x >= when
2066 2066 elif " to " in date:
2067 2067 a, b = date.split(" to ")
2068 2068 start, stop = lower(a), upper(b)
2069 2069 return lambda x: x >= start and x <= stop
2070 2070 else:
2071 2071 start, stop = lower(date), upper(date)
2072 2072 return lambda x: x >= start and x <= stop
2073 2073
2074 2074 def stringmatcher(pattern, casesensitive=True):
2075 2075 """
2076 2076 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2077 2077 returns the matcher name, pattern, and matcher function.
2078 2078 missing or unknown prefixes are treated as literal matches.
2079 2079
2080 2080 helper for tests:
2081 2081 >>> def test(pattern, *tests):
2082 2082 ... kind, pattern, matcher = stringmatcher(pattern)
2083 2083 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2084 2084 >>> def itest(pattern, *tests):
2085 2085 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2086 2086 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2087 2087
2088 2088 exact matching (no prefix):
2089 2089 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2090 2090 ('literal', 'abcdefg', [False, False, True])
2091 2091
2092 2092 regex matching ('re:' prefix)
2093 2093 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2094 2094 ('re', 'a.+b', [False, False, True])
2095 2095
2096 2096 force exact matches ('literal:' prefix)
2097 2097 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2098 2098 ('literal', 're:foobar', [False, True])
2099 2099
2100 2100 unknown prefixes are ignored and treated as literals
2101 2101 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2102 2102 ('literal', 'foo:bar', [False, False, True])
2103 2103
2104 2104 case insensitive regex matches
2105 2105 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2106 2106 ('re', 'A.+b', [False, False, True])
2107 2107
2108 2108 case insensitive literal matches
2109 2109 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2110 2110 ('literal', 'ABCDEFG', [False, False, True])
2111 2111 """
2112 2112 if pattern.startswith('re:'):
2113 2113 pattern = pattern[3:]
2114 2114 try:
2115 2115 flags = 0
2116 2116 if not casesensitive:
2117 2117 flags = remod.I
2118 2118 regex = remod.compile(pattern, flags)
2119 2119 except remod.error as e:
2120 2120 raise error.ParseError(_('invalid regular expression: %s')
2121 2121 % e)
2122 2122 return 're', pattern, regex.search
2123 2123 elif pattern.startswith('literal:'):
2124 2124 pattern = pattern[8:]
2125 2125
2126 2126 match = pattern.__eq__
2127 2127
2128 2128 if not casesensitive:
2129 2129 ipat = encoding.lower(pattern)
2130 2130 match = lambda s: ipat == encoding.lower(s)
2131 2131 return 'literal', pattern, match
2132 2132
2133 2133 def shortuser(user):
2134 2134 """Return a short representation of a user name or email address."""
2135 2135 f = user.find('@')
2136 2136 if f >= 0:
2137 2137 user = user[:f]
2138 2138 f = user.find('<')
2139 2139 if f >= 0:
2140 2140 user = user[f + 1:]
2141 2141 f = user.find(' ')
2142 2142 if f >= 0:
2143 2143 user = user[:f]
2144 2144 f = user.find('.')
2145 2145 if f >= 0:
2146 2146 user = user[:f]
2147 2147 return user
2148 2148
2149 2149 def emailuser(user):
2150 2150 """Return the user portion of an email address."""
2151 2151 f = user.find('@')
2152 2152 if f >= 0:
2153 2153 user = user[:f]
2154 2154 f = user.find('<')
2155 2155 if f >= 0:
2156 2156 user = user[f + 1:]
2157 2157 return user
2158 2158
2159 2159 def email(author):
2160 2160 '''get email of author.'''
2161 2161 r = author.find('>')
2162 2162 if r == -1:
2163 2163 r = None
2164 2164 return author[author.find('<') + 1:r]
2165 2165
2166 2166 def ellipsis(text, maxlength=400):
2167 2167 """Trim string to at most maxlength (default: 400) columns in display."""
2168 2168 return encoding.trim(text, maxlength, ellipsis='...')
2169 2169
2170 2170 def unitcountfn(*unittable):
2171 2171 '''return a function that renders a readable count of some quantity'''
2172 2172
2173 2173 def go(count):
2174 2174 for multiplier, divisor, format in unittable:
2175 2175 if abs(count) >= divisor * multiplier:
2176 2176 return format % (count / float(divisor))
2177 2177 return unittable[-1][2] % count
2178 2178
2179 2179 return go
2180 2180
2181 2181 def processlinerange(fromline, toline):
2182 2182 """Check that linerange <fromline>:<toline> makes sense and return a
2183 2183 0-based range.
2184 2184
2185 2185 >>> processlinerange(10, 20)
2186 2186 (9, 20)
2187 2187 >>> processlinerange(2, 1)
2188 2188 Traceback (most recent call last):
2189 2189 ...
2190 2190 ParseError: line range must be positive
2191 2191 >>> processlinerange(0, 5)
2192 2192 Traceback (most recent call last):
2193 2193 ...
2194 2194 ParseError: fromline must be strictly positive
2195 2195 """
2196 2196 if toline - fromline < 0:
2197 2197 raise error.ParseError(_("line range must be positive"))
2198 2198 if fromline < 1:
2199 2199 raise error.ParseError(_("fromline must be strictly positive"))
2200 2200 return fromline - 1, toline
2201 2201
2202 2202 bytecount = unitcountfn(
2203 2203 (100, 1 << 30, _('%.0f GB')),
2204 2204 (10, 1 << 30, _('%.1f GB')),
2205 2205 (1, 1 << 30, _('%.2f GB')),
2206 2206 (100, 1 << 20, _('%.0f MB')),
2207 2207 (10, 1 << 20, _('%.1f MB')),
2208 2208 (1, 1 << 20, _('%.2f MB')),
2209 2209 (100, 1 << 10, _('%.0f KB')),
2210 2210 (10, 1 << 10, _('%.1f KB')),
2211 2211 (1, 1 << 10, _('%.2f KB')),
2212 2212 (1, 1, _('%.0f bytes')),
2213 2213 )
2214 2214
2215 2215 # Matches a single EOL which can either be a CRLF where repeated CR
2216 2216 # are removed or a LF. We do not care about old Macintosh files, so a
2217 2217 # stray CR is an error.
2218 2218 _eolre = remod.compile(br'\r*\n')
2219 2219
2220 2220 def tolf(s):
2221 2221 return _eolre.sub('\n', s)
2222 2222
2223 2223 def tocrlf(s):
2224 2224 return _eolre.sub('\r\n', s)
2225 2225
2226 2226 if pycompat.oslinesep == '\r\n':
2227 2227 tonativeeol = tocrlf
2228 2228 fromnativeeol = tolf
2229 2229 else:
2230 2230 tonativeeol = pycompat.identity
2231 2231 fromnativeeol = pycompat.identity
2232 2232
2233 2233 def escapestr(s):
2234 2234 # call underlying function of s.encode('string_escape') directly for
2235 2235 # Python 3 compatibility
2236 2236 return codecs.escape_encode(s)[0]
2237 2237
2238 2238 def unescapestr(s):
2239 2239 return codecs.escape_decode(s)[0]
2240 2240
2241 2241 def uirepr(s):
2242 2242 # Avoid double backslash in Windows path repr()
2243 2243 return repr(s).replace('\\\\', '\\')
2244 2244
2245 2245 # delay import of textwrap
2246 2246 def MBTextWrapper(**kwargs):
2247 2247 class tw(textwrap.TextWrapper):
2248 2248 """
2249 2249 Extend TextWrapper for width-awareness.
2250 2250
2251 2251 Neither number of 'bytes' in any encoding nor 'characters' is
2252 2252 appropriate to calculate terminal columns for specified string.
2253 2253
2254 2254 Original TextWrapper implementation uses built-in 'len()' directly,
2255 2255 so overriding is needed to use width information of each characters.
2256 2256
2257 2257 In addition, characters classified into 'ambiguous' width are
2258 2258 treated as wide in East Asian area, but as narrow in other.
2259 2259
2260 2260 This requires use decision to determine width of such characters.
2261 2261 """
2262 2262 def _cutdown(self, ucstr, space_left):
2263 2263 l = 0
2264 2264 colwidth = encoding.ucolwidth
2265 2265 for i in xrange(len(ucstr)):
2266 2266 l += colwidth(ucstr[i])
2267 2267 if space_left < l:
2268 2268 return (ucstr[:i], ucstr[i:])
2269 2269 return ucstr, ''
2270 2270
2271 2271 # overriding of base class
2272 2272 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2273 2273 space_left = max(width - cur_len, 1)
2274 2274
2275 2275 if self.break_long_words:
2276 2276 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2277 2277 cur_line.append(cut)
2278 2278 reversed_chunks[-1] = res
2279 2279 elif not cur_line:
2280 2280 cur_line.append(reversed_chunks.pop())
2281 2281
2282 2282 # this overriding code is imported from TextWrapper of Python 2.6
2283 2283 # to calculate columns of string by 'encoding.ucolwidth()'
2284 2284 def _wrap_chunks(self, chunks):
2285 2285 colwidth = encoding.ucolwidth
2286 2286
2287 2287 lines = []
2288 2288 if self.width <= 0:
2289 2289 raise ValueError("invalid width %r (must be > 0)" % self.width)
2290 2290
2291 2291 # Arrange in reverse order so items can be efficiently popped
2292 2292 # from a stack of chucks.
2293 2293 chunks.reverse()
2294 2294
2295 2295 while chunks:
2296 2296
2297 2297 # Start the list of chunks that will make up the current line.
2298 2298 # cur_len is just the length of all the chunks in cur_line.
2299 2299 cur_line = []
2300 2300 cur_len = 0
2301 2301
2302 2302 # Figure out which static string will prefix this line.
2303 2303 if lines:
2304 2304 indent = self.subsequent_indent
2305 2305 else:
2306 2306 indent = self.initial_indent
2307 2307
2308 2308 # Maximum width for this line.
2309 2309 width = self.width - len(indent)
2310 2310
2311 2311 # First chunk on line is whitespace -- drop it, unless this
2312 2312 # is the very beginning of the text (i.e. no lines started yet).
2313 2313 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2314 2314 del chunks[-1]
2315 2315
2316 2316 while chunks:
2317 2317 l = colwidth(chunks[-1])
2318 2318
2319 2319 # Can at least squeeze this chunk onto the current line.
2320 2320 if cur_len + l <= width:
2321 2321 cur_line.append(chunks.pop())
2322 2322 cur_len += l
2323 2323
2324 2324 # Nope, this line is full.
2325 2325 else:
2326 2326 break
2327 2327
2328 2328 # The current line is full, and the next chunk is too big to
2329 2329 # fit on *any* line (not just this one).
2330 2330 if chunks and colwidth(chunks[-1]) > width:
2331 2331 self._handle_long_word(chunks, cur_line, cur_len, width)
2332 2332
2333 2333 # If the last chunk on this line is all whitespace, drop it.
2334 2334 if (self.drop_whitespace and
2335 2335 cur_line and cur_line[-1].strip() == ''):
2336 2336 del cur_line[-1]
2337 2337
2338 2338 # Convert current line back to a string and store it in list
2339 2339 # of all lines (return value).
2340 2340 if cur_line:
2341 2341 lines.append(indent + ''.join(cur_line))
2342 2342
2343 2343 return lines
2344 2344
2345 2345 global MBTextWrapper
2346 2346 MBTextWrapper = tw
2347 2347 return tw(**kwargs)
2348 2348
2349 2349 def wrap(line, width, initindent='', hangindent=''):
2350 2350 maxindent = max(len(hangindent), len(initindent))
2351 2351 if width <= maxindent:
2352 2352 # adjust for weird terminal size
2353 2353 width = max(78, maxindent + 1)
2354 2354 line = line.decode(pycompat.sysstr(encoding.encoding),
2355 2355 pycompat.sysstr(encoding.encodingmode))
2356 2356 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2357 2357 pycompat.sysstr(encoding.encodingmode))
2358 2358 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2359 2359 pycompat.sysstr(encoding.encodingmode))
2360 2360 wrapper = MBTextWrapper(width=width,
2361 2361 initial_indent=initindent,
2362 2362 subsequent_indent=hangindent)
2363 2363 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2364 2364
2365 2365 if (pyplatform.python_implementation() == 'CPython' and
2366 2366 sys.version_info < (3, 0)):
2367 2367 # There is an issue in CPython that some IO methods do not handle EINTR
2368 2368 # correctly. The following table shows what CPython version (and functions)
2369 2369 # are affected (buggy: has the EINTR bug, okay: otherwise):
2370 2370 #
2371 2371 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2372 2372 # --------------------------------------------------
2373 2373 # fp.__iter__ | buggy | buggy | okay
2374 2374 # fp.read* | buggy | okay [1] | okay
2375 2375 #
2376 2376 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2377 2377 #
2378 2378 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2379 2379 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2380 2380 #
2381 2381 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2382 2382 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2383 2383 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2384 2384 # fp.__iter__ but not other fp.read* methods.
2385 2385 #
2386 2386 # On modern systems like Linux, the "read" syscall cannot be interrupted
2387 2387 # when reading "fast" files like on-disk files. So the EINTR issue only
2388 2388 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2389 2389 # files approximately as "fast" files and use the fast (unsafe) code path,
2390 2390 # to minimize the performance impact.
2391 2391 if sys.version_info >= (2, 7, 4):
2392 2392 # fp.readline deals with EINTR correctly, use it as a workaround.
2393 2393 def _safeiterfile(fp):
2394 2394 return iter(fp.readline, '')
2395 2395 else:
2396 2396 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2397 2397 # note: this may block longer than necessary because of bufsize.
2398 2398 def _safeiterfile(fp, bufsize=4096):
2399 2399 fd = fp.fileno()
2400 2400 line = ''
2401 2401 while True:
2402 2402 try:
2403 2403 buf = os.read(fd, bufsize)
2404 2404 except OSError as ex:
2405 2405 # os.read only raises EINTR before any data is read
2406 2406 if ex.errno == errno.EINTR:
2407 2407 continue
2408 2408 else:
2409 2409 raise
2410 2410 line += buf
2411 2411 if '\n' in buf:
2412 2412 splitted = line.splitlines(True)
2413 2413 line = ''
2414 2414 for l in splitted:
2415 2415 if l[-1] == '\n':
2416 2416 yield l
2417 2417 else:
2418 2418 line = l
2419 2419 if not buf:
2420 2420 break
2421 2421 if line:
2422 2422 yield line
2423 2423
2424 2424 def iterfile(fp):
2425 2425 fastpath = True
2426 2426 if type(fp) is file:
2427 2427 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2428 2428 if fastpath:
2429 2429 return fp
2430 2430 else:
2431 2431 return _safeiterfile(fp)
2432 2432 else:
2433 2433 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2434 2434 def iterfile(fp):
2435 2435 return fp
2436 2436
2437 2437 def iterlines(iterator):
2438 2438 for chunk in iterator:
2439 2439 for line in chunk.splitlines():
2440 2440 yield line
2441 2441
2442 2442 def expandpath(path):
2443 2443 return os.path.expanduser(os.path.expandvars(path))
2444 2444
2445 2445 def hgcmd():
2446 2446 """Return the command used to execute current hg
2447 2447
2448 2448 This is different from hgexecutable() because on Windows we want
2449 2449 to avoid things opening new shell windows like batch files, so we
2450 2450 get either the python call or current executable.
2451 2451 """
2452 2452 if mainfrozen():
2453 2453 if getattr(sys, 'frozen', None) == 'macosx_app':
2454 2454 # Env variable set by py2app
2455 2455 return [encoding.environ['EXECUTABLEPATH']]
2456 2456 else:
2457 2457 return [pycompat.sysexecutable]
2458 2458 return gethgcmd()
2459 2459
2460 2460 def rundetached(args, condfn):
2461 2461 """Execute the argument list in a detached process.
2462 2462
2463 2463 condfn is a callable which is called repeatedly and should return
2464 2464 True once the child process is known to have started successfully.
2465 2465 At this point, the child process PID is returned. If the child
2466 2466 process fails to start or finishes before condfn() evaluates to
2467 2467 True, return -1.
2468 2468 """
2469 2469 # Windows case is easier because the child process is either
2470 2470 # successfully starting and validating the condition or exiting
2471 2471 # on failure. We just poll on its PID. On Unix, if the child
2472 2472 # process fails to start, it will be left in a zombie state until
2473 2473 # the parent wait on it, which we cannot do since we expect a long
2474 2474 # running process on success. Instead we listen for SIGCHLD telling
2475 2475 # us our child process terminated.
2476 2476 terminated = set()
2477 2477 def handler(signum, frame):
2478 2478 terminated.add(os.wait())
2479 2479 prevhandler = None
2480 2480 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2481 2481 if SIGCHLD is not None:
2482 2482 prevhandler = signal.signal(SIGCHLD, handler)
2483 2483 try:
2484 2484 pid = spawndetached(args)
2485 2485 while not condfn():
2486 2486 if ((pid in terminated or not testpid(pid))
2487 2487 and not condfn()):
2488 2488 return -1
2489 2489 time.sleep(0.1)
2490 2490 return pid
2491 2491 finally:
2492 2492 if prevhandler is not None:
2493 2493 signal.signal(signal.SIGCHLD, prevhandler)
2494 2494
2495 2495 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2496 2496 """Return the result of interpolating items in the mapping into string s.
2497 2497
2498 2498 prefix is a single character string, or a two character string with
2499 2499 a backslash as the first character if the prefix needs to be escaped in
2500 2500 a regular expression.
2501 2501
2502 2502 fn is an optional function that will be applied to the replacement text
2503 2503 just before replacement.
2504 2504
2505 2505 escape_prefix is an optional flag that allows using doubled prefix for
2506 2506 its escaping.
2507 2507 """
2508 2508 fn = fn or (lambda s: s)
2509 2509 patterns = '|'.join(mapping.keys())
2510 2510 if escape_prefix:
2511 2511 patterns += '|' + prefix
2512 2512 if len(prefix) > 1:
2513 2513 prefix_char = prefix[1:]
2514 2514 else:
2515 2515 prefix_char = prefix
2516 2516 mapping[prefix_char] = prefix_char
2517 2517 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2518 2518 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2519 2519
2520 2520 def getport(port):
2521 2521 """Return the port for a given network service.
2522 2522
2523 2523 If port is an integer, it's returned as is. If it's a string, it's
2524 2524 looked up using socket.getservbyname(). If there's no matching
2525 2525 service, error.Abort is raised.
2526 2526 """
2527 2527 try:
2528 2528 return int(port)
2529 2529 except ValueError:
2530 2530 pass
2531 2531
2532 2532 try:
2533 2533 return socket.getservbyname(port)
2534 2534 except socket.error:
2535 2535 raise Abort(_("no port number associated with service '%s'") % port)
2536 2536
2537 2537 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2538 2538 '0': False, 'no': False, 'false': False, 'off': False,
2539 2539 'never': False}
2540 2540
2541 2541 def parsebool(s):
2542 2542 """Parse s into a boolean.
2543 2543
2544 2544 If s is not a valid boolean, returns None.
2545 2545 """
2546 2546 return _booleans.get(s.lower(), None)
2547 2547
2548 2548 _hextochr = dict((a + b, chr(int(a + b, 16)))
2549 2549 for a in string.hexdigits for b in string.hexdigits)
2550 2550
2551 2551 class url(object):
2552 2552 r"""Reliable URL parser.
2553 2553
2554 2554 This parses URLs and provides attributes for the following
2555 2555 components:
2556 2556
2557 2557 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2558 2558
2559 2559 Missing components are set to None. The only exception is
2560 2560 fragment, which is set to '' if present but empty.
2561 2561
2562 2562 If parsefragment is False, fragment is included in query. If
2563 2563 parsequery is False, query is included in path. If both are
2564 2564 False, both fragment and query are included in path.
2565 2565
2566 2566 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2567 2567
2568 2568 Note that for backward compatibility reasons, bundle URLs do not
2569 2569 take host names. That means 'bundle://../' has a path of '../'.
2570 2570
2571 2571 Examples:
2572 2572
2573 2573 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2574 2574 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2575 2575 >>> url('ssh://[::1]:2200//home/joe/repo')
2576 2576 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2577 2577 >>> url('file:///home/joe/repo')
2578 2578 <url scheme: 'file', path: '/home/joe/repo'>
2579 2579 >>> url('file:///c:/temp/foo/')
2580 2580 <url scheme: 'file', path: 'c:/temp/foo/'>
2581 2581 >>> url('bundle:foo')
2582 2582 <url scheme: 'bundle', path: 'foo'>
2583 2583 >>> url('bundle://../foo')
2584 2584 <url scheme: 'bundle', path: '../foo'>
2585 2585 >>> url(r'c:\foo\bar')
2586 2586 <url path: 'c:\\foo\\bar'>
2587 2587 >>> url(r'\\blah\blah\blah')
2588 2588 <url path: '\\\\blah\\blah\\blah'>
2589 2589 >>> url(r'\\blah\blah\blah#baz')
2590 2590 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2591 2591 >>> url(r'file:///C:\users\me')
2592 2592 <url scheme: 'file', path: 'C:\\users\\me'>
2593 2593
2594 2594 Authentication credentials:
2595 2595
2596 2596 >>> url('ssh://joe:xyz@x/repo')
2597 2597 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2598 2598 >>> url('ssh://joe@x/repo')
2599 2599 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2600 2600
2601 2601 Query strings and fragments:
2602 2602
2603 2603 >>> url('http://host/a?b#c')
2604 2604 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2605 2605 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2606 2606 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2607 2607
2608 2608 Empty path:
2609 2609
2610 2610 >>> url('')
2611 2611 <url path: ''>
2612 2612 >>> url('#a')
2613 2613 <url path: '', fragment: 'a'>
2614 2614 >>> url('http://host/')
2615 2615 <url scheme: 'http', host: 'host', path: ''>
2616 2616 >>> url('http://host/#a')
2617 2617 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2618 2618
2619 2619 Only scheme:
2620 2620
2621 2621 >>> url('http:')
2622 2622 <url scheme: 'http'>
2623 2623 """
2624 2624
2625 2625 _safechars = "!~*'()+"
2626 2626 _safepchars = "/!~*'()+:\\"
2627 2627 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2628 2628
2629 2629 def __init__(self, path, parsequery=True, parsefragment=True):
2630 2630 # We slowly chomp away at path until we have only the path left
2631 2631 self.scheme = self.user = self.passwd = self.host = None
2632 2632 self.port = self.path = self.query = self.fragment = None
2633 2633 self._localpath = True
2634 2634 self._hostport = ''
2635 2635 self._origpath = path
2636 2636
2637 2637 if parsefragment and '#' in path:
2638 2638 path, self.fragment = path.split('#', 1)
2639 2639
2640 2640 # special case for Windows drive letters and UNC paths
2641 2641 if hasdriveletter(path) or path.startswith('\\\\'):
2642 2642 self.path = path
2643 2643 return
2644 2644
2645 2645 # For compatibility reasons, we can't handle bundle paths as
2646 2646 # normal URLS
2647 2647 if path.startswith('bundle:'):
2648 2648 self.scheme = 'bundle'
2649 2649 path = path[7:]
2650 2650 if path.startswith('//'):
2651 2651 path = path[2:]
2652 2652 self.path = path
2653 2653 return
2654 2654
2655 2655 if self._matchscheme(path):
2656 2656 parts = path.split(':', 1)
2657 2657 if parts[0]:
2658 2658 self.scheme, path = parts
2659 2659 self._localpath = False
2660 2660
2661 2661 if not path:
2662 2662 path = None
2663 2663 if self._localpath:
2664 2664 self.path = ''
2665 2665 return
2666 2666 else:
2667 2667 if self._localpath:
2668 2668 self.path = path
2669 2669 return
2670 2670
2671 2671 if parsequery and '?' in path:
2672 2672 path, self.query = path.split('?', 1)
2673 2673 if not path:
2674 2674 path = None
2675 2675 if not self.query:
2676 2676 self.query = None
2677 2677
2678 2678 # // is required to specify a host/authority
2679 2679 if path and path.startswith('//'):
2680 2680 parts = path[2:].split('/', 1)
2681 2681 if len(parts) > 1:
2682 2682 self.host, path = parts
2683 2683 else:
2684 2684 self.host = parts[0]
2685 2685 path = None
2686 2686 if not self.host:
2687 2687 self.host = None
2688 2688 # path of file:///d is /d
2689 2689 # path of file:///d:/ is d:/, not /d:/
2690 2690 if path and not hasdriveletter(path):
2691 2691 path = '/' + path
2692 2692
2693 2693 if self.host and '@' in self.host:
2694 2694 self.user, self.host = self.host.rsplit('@', 1)
2695 2695 if ':' in self.user:
2696 2696 self.user, self.passwd = self.user.split(':', 1)
2697 2697 if not self.host:
2698 2698 self.host = None
2699 2699
2700 2700 # Don't split on colons in IPv6 addresses without ports
2701 2701 if (self.host and ':' in self.host and
2702 2702 not (self.host.startswith('[') and self.host.endswith(']'))):
2703 2703 self._hostport = self.host
2704 2704 self.host, self.port = self.host.rsplit(':', 1)
2705 2705 if not self.host:
2706 2706 self.host = None
2707 2707
2708 2708 if (self.host and self.scheme == 'file' and
2709 2709 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2710 2710 raise Abort(_('file:// URLs can only refer to localhost'))
2711 2711
2712 2712 self.path = path
2713 2713
2714 2714 # leave the query string escaped
2715 2715 for a in ('user', 'passwd', 'host', 'port',
2716 2716 'path', 'fragment'):
2717 2717 v = getattr(self, a)
2718 2718 if v is not None:
2719 2719 setattr(self, a, urlreq.unquote(v))
2720 2720
2721 2721 def __repr__(self):
2722 2722 attrs = []
2723 2723 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2724 2724 'query', 'fragment'):
2725 2725 v = getattr(self, a)
2726 2726 if v is not None:
2727 2727 attrs.append('%s: %r' % (a, v))
2728 2728 return '<url %s>' % ', '.join(attrs)
2729 2729
2730 2730 def __str__(self):
2731 2731 r"""Join the URL's components back into a URL string.
2732 2732
2733 2733 Examples:
2734 2734
2735 2735 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2736 2736 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2737 2737 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2738 2738 'http://user:pw@host:80/?foo=bar&baz=42'
2739 2739 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2740 2740 'http://user:pw@host:80/?foo=bar%3dbaz'
2741 2741 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2742 2742 'ssh://user:pw@[::1]:2200//home/joe#'
2743 2743 >>> str(url('http://localhost:80//'))
2744 2744 'http://localhost:80//'
2745 2745 >>> str(url('http://localhost:80/'))
2746 2746 'http://localhost:80/'
2747 2747 >>> str(url('http://localhost:80'))
2748 2748 'http://localhost:80/'
2749 2749 >>> str(url('bundle:foo'))
2750 2750 'bundle:foo'
2751 2751 >>> str(url('bundle://../foo'))
2752 2752 'bundle:../foo'
2753 2753 >>> str(url('path'))
2754 2754 'path'
2755 2755 >>> str(url('file:///tmp/foo/bar'))
2756 2756 'file:///tmp/foo/bar'
2757 2757 >>> str(url('file:///c:/tmp/foo/bar'))
2758 2758 'file:///c:/tmp/foo/bar'
2759 2759 >>> print url(r'bundle:foo\bar')
2760 2760 bundle:foo\bar
2761 2761 >>> print url(r'file:///D:\data\hg')
2762 2762 file:///D:\data\hg
2763 2763 """
2764 2764 return encoding.strfromlocal(self.__bytes__())
2765 2765
2766 2766 def __bytes__(self):
2767 2767 if self._localpath:
2768 2768 s = self.path
2769 2769 if self.scheme == 'bundle':
2770 2770 s = 'bundle:' + s
2771 2771 if self.fragment:
2772 2772 s += '#' + self.fragment
2773 2773 return s
2774 2774
2775 2775 s = self.scheme + ':'
2776 2776 if self.user or self.passwd or self.host:
2777 2777 s += '//'
2778 2778 elif self.scheme and (not self.path or self.path.startswith('/')
2779 2779 or hasdriveletter(self.path)):
2780 2780 s += '//'
2781 2781 if hasdriveletter(self.path):
2782 2782 s += '/'
2783 2783 if self.user:
2784 2784 s += urlreq.quote(self.user, safe=self._safechars)
2785 2785 if self.passwd:
2786 2786 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2787 2787 if self.user or self.passwd:
2788 2788 s += '@'
2789 2789 if self.host:
2790 2790 if not (self.host.startswith('[') and self.host.endswith(']')):
2791 2791 s += urlreq.quote(self.host)
2792 2792 else:
2793 2793 s += self.host
2794 2794 if self.port:
2795 2795 s += ':' + urlreq.quote(self.port)
2796 2796 if self.host:
2797 2797 s += '/'
2798 2798 if self.path:
2799 2799 # TODO: similar to the query string, we should not unescape the
2800 2800 # path when we store it, the path might contain '%2f' = '/',
2801 2801 # which we should *not* escape.
2802 2802 s += urlreq.quote(self.path, safe=self._safepchars)
2803 2803 if self.query:
2804 2804 # we store the query in escaped form.
2805 2805 s += '?' + self.query
2806 2806 if self.fragment is not None:
2807 2807 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2808 2808 return s
2809 2809
2810 2810 def authinfo(self):
2811 2811 user, passwd = self.user, self.passwd
2812 2812 try:
2813 2813 self.user, self.passwd = None, None
2814 2814 s = bytes(self)
2815 2815 finally:
2816 2816 self.user, self.passwd = user, passwd
2817 2817 if not self.user:
2818 2818 return (s, None)
2819 2819 # authinfo[1] is passed to urllib2 password manager, and its
2820 2820 # URIs must not contain credentials. The host is passed in the
2821 2821 # URIs list because Python < 2.4.3 uses only that to search for
2822 2822 # a password.
2823 2823 return (s, (None, (s, self.host),
2824 2824 self.user, self.passwd or ''))
2825 2825
2826 2826 def isabs(self):
2827 2827 if self.scheme and self.scheme != 'file':
2828 2828 return True # remote URL
2829 2829 if hasdriveletter(self.path):
2830 2830 return True # absolute for our purposes - can't be joined()
2831 2831 if self.path.startswith(r'\\'):
2832 2832 return True # Windows UNC path
2833 2833 if self.path.startswith('/'):
2834 2834 return True # POSIX-style
2835 2835 return False
2836 2836
2837 2837 def localpath(self):
2838 2838 if self.scheme == 'file' or self.scheme == 'bundle':
2839 2839 path = self.path or '/'
2840 2840 # For Windows, we need to promote hosts containing drive
2841 2841 # letters to paths with drive letters.
2842 2842 if hasdriveletter(self._hostport):
2843 2843 path = self._hostport + '/' + self.path
2844 2844 elif (self.host is not None and self.path
2845 2845 and not hasdriveletter(path)):
2846 2846 path = '/' + path
2847 2847 return path
2848 2848 return self._origpath
2849 2849
2850 2850 def islocal(self):
2851 2851 '''whether localpath will return something that posixfile can open'''
2852 2852 return (not self.scheme or self.scheme == 'file'
2853 2853 or self.scheme == 'bundle')
2854 2854
2855 2855 def hasscheme(path):
2856 2856 return bool(url(path).scheme)
2857 2857
2858 2858 def hasdriveletter(path):
2859 2859 return path and path[1:2] == ':' and path[0:1].isalpha()
2860 2860
2861 2861 def urllocalpath(path):
2862 2862 return url(path, parsequery=False, parsefragment=False).localpath()
2863 2863
2864 2864 def hidepassword(u):
2865 2865 '''hide user credential in a url string'''
2866 2866 u = url(u)
2867 2867 if u.passwd:
2868 2868 u.passwd = '***'
2869 2869 return bytes(u)
2870 2870
2871 2871 def removeauth(u):
2872 2872 '''remove all authentication information from a url string'''
2873 2873 u = url(u)
2874 2874 u.user = u.passwd = None
2875 2875 return str(u)
2876 2876
2877 2877 timecount = unitcountfn(
2878 2878 (1, 1e3, _('%.0f s')),
2879 2879 (100, 1, _('%.1f s')),
2880 2880 (10, 1, _('%.2f s')),
2881 2881 (1, 1, _('%.3f s')),
2882 2882 (100, 0.001, _('%.1f ms')),
2883 2883 (10, 0.001, _('%.2f ms')),
2884 2884 (1, 0.001, _('%.3f ms')),
2885 2885 (100, 0.000001, _('%.1f us')),
2886 2886 (10, 0.000001, _('%.2f us')),
2887 2887 (1, 0.000001, _('%.3f us')),
2888 2888 (100, 0.000000001, _('%.1f ns')),
2889 2889 (10, 0.000000001, _('%.2f ns')),
2890 2890 (1, 0.000000001, _('%.3f ns')),
2891 2891 )
2892 2892
2893 2893 _timenesting = [0]
2894 2894
2895 2895 def timed(func):
2896 2896 '''Report the execution time of a function call to stderr.
2897 2897
2898 2898 During development, use as a decorator when you need to measure
2899 2899 the cost of a function, e.g. as follows:
2900 2900
2901 2901 @util.timed
2902 2902 def foo(a, b, c):
2903 2903 pass
2904 2904 '''
2905 2905
2906 2906 def wrapper(*args, **kwargs):
2907 2907 start = timer()
2908 2908 indent = 2
2909 2909 _timenesting[0] += indent
2910 2910 try:
2911 2911 return func(*args, **kwargs)
2912 2912 finally:
2913 2913 elapsed = timer() - start
2914 2914 _timenesting[0] -= indent
2915 2915 stderr.write('%s%s: %s\n' %
2916 2916 (' ' * _timenesting[0], func.__name__,
2917 2917 timecount(elapsed)))
2918 2918 return wrapper
2919 2919
2920 2920 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2921 2921 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2922 2922
2923 2923 def sizetoint(s):
2924 2924 '''Convert a space specifier to a byte count.
2925 2925
2926 2926 >>> sizetoint('30')
2927 2927 30
2928 2928 >>> sizetoint('2.2kb')
2929 2929 2252
2930 2930 >>> sizetoint('6M')
2931 2931 6291456
2932 2932 '''
2933 2933 t = s.strip().lower()
2934 2934 try:
2935 2935 for k, u in _sizeunits:
2936 2936 if t.endswith(k):
2937 2937 return int(float(t[:-len(k)]) * u)
2938 2938 return int(t)
2939 2939 except ValueError:
2940 2940 raise error.ParseError(_("couldn't parse size: %s") % s)
2941 2941
2942 2942 class hooks(object):
2943 2943 '''A collection of hook functions that can be used to extend a
2944 2944 function's behavior. Hooks are called in lexicographic order,
2945 2945 based on the names of their sources.'''
2946 2946
2947 2947 def __init__(self):
2948 2948 self._hooks = []
2949 2949
2950 2950 def add(self, source, hook):
2951 2951 self._hooks.append((source, hook))
2952 2952
2953 2953 def __call__(self, *args):
2954 2954 self._hooks.sort(key=lambda x: x[0])
2955 2955 results = []
2956 2956 for source, hook in self._hooks:
2957 2957 results.append(hook(*args))
2958 2958 return results
2959 2959
2960 2960 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2961 2961 '''Yields lines for a nicely formatted stacktrace.
2962 2962 Skips the 'skip' last entries, then return the last 'depth' entries.
2963 2963 Each file+linenumber is formatted according to fileline.
2964 2964 Each line is formatted according to line.
2965 2965 If line is None, it yields:
2966 2966 length of longest filepath+line number,
2967 2967 filepath+linenumber,
2968 2968 function
2969 2969
2970 2970 Not be used in production code but very convenient while developing.
2971 2971 '''
2972 2972 entries = [(fileline % (fn, ln), func)
2973 2973 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2974 2974 ][-depth:]
2975 2975 if entries:
2976 2976 fnmax = max(len(entry[0]) for entry in entries)
2977 2977 for fnln, func in entries:
2978 2978 if line is None:
2979 2979 yield (fnmax, fnln, func)
2980 2980 else:
2981 2981 yield line % (fnmax, fnln, func)
2982 2982
2983 2983 def debugstacktrace(msg='stacktrace', skip=0,
2984 2984 f=stderr, otherf=stdout, depth=0):
2985 2985 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2986 2986 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2987 2987 By default it will flush stdout first.
2988 2988 It can be used everywhere and intentionally does not require an ui object.
2989 2989 Not be used in production code but very convenient while developing.
2990 2990 '''
2991 2991 if otherf:
2992 2992 otherf.flush()
2993 2993 f.write('%s at:\n' % msg.rstrip())
2994 2994 for line in getstackframes(skip + 1, depth=depth):
2995 2995 f.write(line)
2996 2996 f.flush()
2997 2997
2998 2998 class dirs(object):
2999 2999 '''a multiset of directory names from a dirstate or manifest'''
3000 3000
3001 3001 def __init__(self, map, skip=None):
3002 3002 self._dirs = {}
3003 3003 addpath = self.addpath
3004 3004 if safehasattr(map, 'iteritems') and skip is not None:
3005 3005 for f, s in map.iteritems():
3006 3006 if s[0] != skip:
3007 3007 addpath(f)
3008 3008 else:
3009 3009 for f in map:
3010 3010 addpath(f)
3011 3011
3012 3012 def addpath(self, path):
3013 3013 dirs = self._dirs
3014 3014 for base in finddirs(path):
3015 3015 if base in dirs:
3016 3016 dirs[base] += 1
3017 3017 return
3018 3018 dirs[base] = 1
3019 3019
3020 3020 def delpath(self, path):
3021 3021 dirs = self._dirs
3022 3022 for base in finddirs(path):
3023 3023 if dirs[base] > 1:
3024 3024 dirs[base] -= 1
3025 3025 return
3026 3026 del dirs[base]
3027 3027
3028 3028 def __iter__(self):
3029 3029 return iter(self._dirs)
3030 3030
3031 3031 def __contains__(self, d):
3032 3032 return d in self._dirs
3033 3033
3034 3034 if safehasattr(parsers, 'dirs'):
3035 3035 dirs = parsers.dirs
3036 3036
3037 3037 def finddirs(path):
3038 3038 pos = path.rfind('/')
3039 3039 while pos != -1:
3040 3040 yield path[:pos]
3041 3041 pos = path.rfind('/', 0, pos)
3042 3042
3043 3043 class ctxmanager(object):
3044 3044 '''A context manager for use in 'with' blocks to allow multiple
3045 3045 contexts to be entered at once. This is both safer and more
3046 3046 flexible than contextlib.nested.
3047 3047
3048 3048 Once Mercurial supports Python 2.7+, this will become mostly
3049 3049 unnecessary.
3050 3050 '''
3051 3051
3052 3052 def __init__(self, *args):
3053 3053 '''Accepts a list of no-argument functions that return context
3054 3054 managers. These will be invoked at __call__ time.'''
3055 3055 self._pending = args
3056 3056 self._atexit = []
3057 3057
3058 3058 def __enter__(self):
3059 3059 return self
3060 3060
3061 3061 def enter(self):
3062 3062 '''Create and enter context managers in the order in which they were
3063 3063 passed to the constructor.'''
3064 3064 values = []
3065 3065 for func in self._pending:
3066 3066 obj = func()
3067 3067 values.append(obj.__enter__())
3068 3068 self._atexit.append(obj.__exit__)
3069 3069 del self._pending
3070 3070 return values
3071 3071
3072 3072 def atexit(self, func, *args, **kwargs):
3073 3073 '''Add a function to call when this context manager exits. The
3074 3074 ordering of multiple atexit calls is unspecified, save that
3075 3075 they will happen before any __exit__ functions.'''
3076 3076 def wrapper(exc_type, exc_val, exc_tb):
3077 3077 func(*args, **kwargs)
3078 3078 self._atexit.append(wrapper)
3079 3079 return func
3080 3080
3081 3081 def __exit__(self, exc_type, exc_val, exc_tb):
3082 3082 '''Context managers are exited in the reverse order from which
3083 3083 they were created.'''
3084 3084 received = exc_type is not None
3085 3085 suppressed = False
3086 3086 pending = None
3087 3087 self._atexit.reverse()
3088 3088 for exitfunc in self._atexit:
3089 3089 try:
3090 3090 if exitfunc(exc_type, exc_val, exc_tb):
3091 3091 suppressed = True
3092 3092 exc_type = None
3093 3093 exc_val = None
3094 3094 exc_tb = None
3095 3095 except BaseException:
3096 3096 pending = sys.exc_info()
3097 3097 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3098 3098 del self._atexit
3099 3099 if pending:
3100 3100 raise exc_val
3101 3101 return received and suppressed
3102 3102
3103 3103 # compression code
3104 3104
3105 3105 SERVERROLE = 'server'
3106 3106 CLIENTROLE = 'client'
3107 3107
3108 3108 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3109 3109 (u'name', u'serverpriority',
3110 3110 u'clientpriority'))
3111 3111
3112 3112 class compressormanager(object):
3113 3113 """Holds registrations of various compression engines.
3114 3114
3115 3115 This class essentially abstracts the differences between compression
3116 3116 engines to allow new compression formats to be added easily, possibly from
3117 3117 extensions.
3118 3118
3119 3119 Compressors are registered against the global instance by calling its
3120 3120 ``register()`` method.
3121 3121 """
3122 3122 def __init__(self):
3123 3123 self._engines = {}
3124 3124 # Bundle spec human name to engine name.
3125 3125 self._bundlenames = {}
3126 3126 # Internal bundle identifier to engine name.
3127 3127 self._bundletypes = {}
3128 3128 # Revlog header to engine name.
3129 3129 self._revlogheaders = {}
3130 3130 # Wire proto identifier to engine name.
3131 3131 self._wiretypes = {}
3132 3132
3133 3133 def __getitem__(self, key):
3134 3134 return self._engines[key]
3135 3135
3136 3136 def __contains__(self, key):
3137 3137 return key in self._engines
3138 3138
3139 3139 def __iter__(self):
3140 3140 return iter(self._engines.keys())
3141 3141
3142 3142 def register(self, engine):
3143 3143 """Register a compression engine with the manager.
3144 3144
3145 3145 The argument must be a ``compressionengine`` instance.
3146 3146 """
3147 3147 if not isinstance(engine, compressionengine):
3148 3148 raise ValueError(_('argument must be a compressionengine'))
3149 3149
3150 3150 name = engine.name()
3151 3151
3152 3152 if name in self._engines:
3153 3153 raise error.Abort(_('compression engine %s already registered') %
3154 3154 name)
3155 3155
3156 3156 bundleinfo = engine.bundletype()
3157 3157 if bundleinfo:
3158 3158 bundlename, bundletype = bundleinfo
3159 3159
3160 3160 if bundlename in self._bundlenames:
3161 3161 raise error.Abort(_('bundle name %s already registered') %
3162 3162 bundlename)
3163 3163 if bundletype in self._bundletypes:
3164 3164 raise error.Abort(_('bundle type %s already registered by %s') %
3165 3165 (bundletype, self._bundletypes[bundletype]))
3166 3166
3167 3167 # No external facing name declared.
3168 3168 if bundlename:
3169 3169 self._bundlenames[bundlename] = name
3170 3170
3171 3171 self._bundletypes[bundletype] = name
3172 3172
3173 3173 wiresupport = engine.wireprotosupport()
3174 3174 if wiresupport:
3175 3175 wiretype = wiresupport.name
3176 3176 if wiretype in self._wiretypes:
3177 3177 raise error.Abort(_('wire protocol compression %s already '
3178 3178 'registered by %s') %
3179 3179 (wiretype, self._wiretypes[wiretype]))
3180 3180
3181 3181 self._wiretypes[wiretype] = name
3182 3182
3183 3183 revlogheader = engine.revlogheader()
3184 3184 if revlogheader and revlogheader in self._revlogheaders:
3185 3185 raise error.Abort(_('revlog header %s already registered by %s') %
3186 3186 (revlogheader, self._revlogheaders[revlogheader]))
3187 3187
3188 3188 if revlogheader:
3189 3189 self._revlogheaders[revlogheader] = name
3190 3190
3191 3191 self._engines[name] = engine
3192 3192
3193 3193 @property
3194 3194 def supportedbundlenames(self):
3195 3195 return set(self._bundlenames.keys())
3196 3196
3197 3197 @property
3198 3198 def supportedbundletypes(self):
3199 3199 return set(self._bundletypes.keys())
3200 3200
3201 3201 def forbundlename(self, bundlename):
3202 3202 """Obtain a compression engine registered to a bundle name.
3203 3203
3204 3204 Will raise KeyError if the bundle type isn't registered.
3205 3205
3206 3206 Will abort if the engine is known but not available.
3207 3207 """
3208 3208 engine = self._engines[self._bundlenames[bundlename]]
3209 3209 if not engine.available():
3210 3210 raise error.Abort(_('compression engine %s could not be loaded') %
3211 3211 engine.name())
3212 3212 return engine
3213 3213
3214 3214 def forbundletype(self, bundletype):
3215 3215 """Obtain a compression engine registered to a bundle type.
3216 3216
3217 3217 Will raise KeyError if the bundle type isn't registered.
3218 3218
3219 3219 Will abort if the engine is known but not available.
3220 3220 """
3221 3221 engine = self._engines[self._bundletypes[bundletype]]
3222 3222 if not engine.available():
3223 3223 raise error.Abort(_('compression engine %s could not be loaded') %
3224 3224 engine.name())
3225 3225 return engine
3226 3226
3227 3227 def supportedwireengines(self, role, onlyavailable=True):
3228 3228 """Obtain compression engines that support the wire protocol.
3229 3229
3230 3230 Returns a list of engines in prioritized order, most desired first.
3231 3231
3232 3232 If ``onlyavailable`` is set, filter out engines that can't be
3233 3233 loaded.
3234 3234 """
3235 3235 assert role in (SERVERROLE, CLIENTROLE)
3236 3236
3237 3237 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3238 3238
3239 3239 engines = [self._engines[e] for e in self._wiretypes.values()]
3240 3240 if onlyavailable:
3241 3241 engines = [e for e in engines if e.available()]
3242 3242
3243 3243 def getkey(e):
3244 3244 # Sort first by priority, highest first. In case of tie, sort
3245 3245 # alphabetically. This is arbitrary, but ensures output is
3246 3246 # stable.
3247 3247 w = e.wireprotosupport()
3248 3248 return -1 * getattr(w, attr), w.name
3249 3249
3250 3250 return list(sorted(engines, key=getkey))
3251 3251
3252 3252 def forwiretype(self, wiretype):
3253 3253 engine = self._engines[self._wiretypes[wiretype]]
3254 3254 if not engine.available():
3255 3255 raise error.Abort(_('compression engine %s could not be loaded') %
3256 3256 engine.name())
3257 3257 return engine
3258 3258
3259 3259 def forrevlogheader(self, header):
3260 3260 """Obtain a compression engine registered to a revlog header.
3261 3261
3262 3262 Will raise KeyError if the revlog header value isn't registered.
3263 3263 """
3264 3264 return self._engines[self._revlogheaders[header]]
3265 3265
3266 3266 compengines = compressormanager()
3267 3267
3268 3268 class compressionengine(object):
3269 3269 """Base class for compression engines.
3270 3270
3271 3271 Compression engines must implement the interface defined by this class.
3272 3272 """
3273 3273 def name(self):
3274 3274 """Returns the name of the compression engine.
3275 3275
3276 3276 This is the key the engine is registered under.
3277 3277
3278 3278 This method must be implemented.
3279 3279 """
3280 3280 raise NotImplementedError()
3281 3281
3282 3282 def available(self):
3283 3283 """Whether the compression engine is available.
3284 3284
3285 3285 The intent of this method is to allow optional compression engines
3286 3286 that may not be available in all installations (such as engines relying
3287 3287 on C extensions that may not be present).
3288 3288 """
3289 3289 return True
3290 3290
3291 3291 def bundletype(self):
3292 3292 """Describes bundle identifiers for this engine.
3293 3293
3294 3294 If this compression engine isn't supported for bundles, returns None.
3295 3295
3296 3296 If this engine can be used for bundles, returns a 2-tuple of strings of
3297 3297 the user-facing "bundle spec" compression name and an internal
3298 3298 identifier used to denote the compression format within bundles. To
3299 3299 exclude the name from external usage, set the first element to ``None``.
3300 3300
3301 3301 If bundle compression is supported, the class must also implement
3302 3302 ``compressstream`` and `decompressorreader``.
3303 3303
3304 3304 The docstring of this method is used in the help system to tell users
3305 3305 about this engine.
3306 3306 """
3307 3307 return None
3308 3308
3309 3309 def wireprotosupport(self):
3310 3310 """Declare support for this compression format on the wire protocol.
3311 3311
3312 3312 If this compression engine isn't supported for compressing wire
3313 3313 protocol payloads, returns None.
3314 3314
3315 3315 Otherwise, returns ``compenginewireprotosupport`` with the following
3316 3316 fields:
3317 3317
3318 3318 * String format identifier
3319 3319 * Integer priority for the server
3320 3320 * Integer priority for the client
3321 3321
3322 3322 The integer priorities are used to order the advertisement of format
3323 3323 support by server and client. The highest integer is advertised
3324 3324 first. Integers with non-positive values aren't advertised.
3325 3325
3326 3326 The priority values are somewhat arbitrary and only used for default
3327 3327 ordering. The relative order can be changed via config options.
3328 3328
3329 3329 If wire protocol compression is supported, the class must also implement
3330 3330 ``compressstream`` and ``decompressorreader``.
3331 3331 """
3332 3332 return None
3333 3333
3334 3334 def revlogheader(self):
3335 3335 """Header added to revlog chunks that identifies this engine.
3336 3336
3337 3337 If this engine can be used to compress revlogs, this method should
3338 3338 return the bytes used to identify chunks compressed with this engine.
3339 3339 Else, the method should return ``None`` to indicate it does not
3340 3340 participate in revlog compression.
3341 3341 """
3342 3342 return None
3343 3343
3344 3344 def compressstream(self, it, opts=None):
3345 3345 """Compress an iterator of chunks.
3346 3346
3347 3347 The method receives an iterator (ideally a generator) of chunks of
3348 3348 bytes to be compressed. It returns an iterator (ideally a generator)
3349 3349 of bytes of chunks representing the compressed output.
3350 3350
3351 3351 Optionally accepts an argument defining how to perform compression.
3352 3352 Each engine treats this argument differently.
3353 3353 """
3354 3354 raise NotImplementedError()
3355 3355
3356 3356 def decompressorreader(self, fh):
3357 3357 """Perform decompression on a file object.
3358 3358
3359 3359 Argument is an object with a ``read(size)`` method that returns
3360 3360 compressed data. Return value is an object with a ``read(size)`` that
3361 3361 returns uncompressed data.
3362 3362 """
3363 3363 raise NotImplementedError()
3364 3364
3365 3365 def revlogcompressor(self, opts=None):
3366 3366 """Obtain an object that can be used to compress revlog entries.
3367 3367
3368 3368 The object has a ``compress(data)`` method that compresses binary
3369 3369 data. This method returns compressed binary data or ``None`` if
3370 3370 the data could not be compressed (too small, not compressible, etc).
3371 3371 The returned data should have a header uniquely identifying this
3372 3372 compression format so decompression can be routed to this engine.
3373 3373 This header should be identified by the ``revlogheader()`` return
3374 3374 value.
3375 3375
3376 3376 The object has a ``decompress(data)`` method that decompresses
3377 3377 data. The method will only be called if ``data`` begins with
3378 3378 ``revlogheader()``. The method should return the raw, uncompressed
3379 3379 data or raise a ``RevlogError``.
3380 3380
3381 3381 The object is reusable but is not thread safe.
3382 3382 """
3383 3383 raise NotImplementedError()
3384 3384
3385 3385 class _zlibengine(compressionengine):
3386 3386 def name(self):
3387 3387 return 'zlib'
3388 3388
3389 3389 def bundletype(self):
3390 3390 """zlib compression using the DEFLATE algorithm.
3391 3391
3392 3392 All Mercurial clients should support this format. The compression
3393 3393 algorithm strikes a reasonable balance between compression ratio
3394 3394 and size.
3395 3395 """
3396 3396 return 'gzip', 'GZ'
3397 3397
3398 3398 def wireprotosupport(self):
3399 3399 return compewireprotosupport('zlib', 20, 20)
3400 3400
3401 3401 def revlogheader(self):
3402 3402 return 'x'
3403 3403
3404 3404 def compressstream(self, it, opts=None):
3405 3405 opts = opts or {}
3406 3406
3407 3407 z = zlib.compressobj(opts.get('level', -1))
3408 3408 for chunk in it:
3409 3409 data = z.compress(chunk)
3410 3410 # Not all calls to compress emit data. It is cheaper to inspect
3411 3411 # here than to feed empty chunks through generator.
3412 3412 if data:
3413 3413 yield data
3414 3414
3415 3415 yield z.flush()
3416 3416
3417 3417 def decompressorreader(self, fh):
3418 3418 def gen():
3419 3419 d = zlib.decompressobj()
3420 3420 for chunk in filechunkiter(fh):
3421 3421 while chunk:
3422 3422 # Limit output size to limit memory.
3423 3423 yield d.decompress(chunk, 2 ** 18)
3424 3424 chunk = d.unconsumed_tail
3425 3425
3426 3426 return chunkbuffer(gen())
3427 3427
3428 3428 class zlibrevlogcompressor(object):
3429 3429 def compress(self, data):
3430 3430 insize = len(data)
3431 3431 # Caller handles empty input case.
3432 3432 assert insize > 0
3433 3433
3434 3434 if insize < 44:
3435 3435 return None
3436 3436
3437 3437 elif insize <= 1000000:
3438 3438 compressed = zlib.compress(data)
3439 3439 if len(compressed) < insize:
3440 3440 return compressed
3441 3441 return None
3442 3442
3443 3443 # zlib makes an internal copy of the input buffer, doubling
3444 3444 # memory usage for large inputs. So do streaming compression
3445 3445 # on large inputs.
3446 3446 else:
3447 3447 z = zlib.compressobj()
3448 3448 parts = []
3449 3449 pos = 0
3450 3450 while pos < insize:
3451 3451 pos2 = pos + 2**20
3452 3452 parts.append(z.compress(data[pos:pos2]))
3453 3453 pos = pos2
3454 3454 parts.append(z.flush())
3455 3455
3456 3456 if sum(map(len, parts)) < insize:
3457 3457 return ''.join(parts)
3458 3458 return None
3459 3459
3460 3460 def decompress(self, data):
3461 3461 try:
3462 3462 return zlib.decompress(data)
3463 3463 except zlib.error as e:
3464 3464 raise error.RevlogError(_('revlog decompress error: %s') %
3465 3465 str(e))
3466 3466
3467 3467 def revlogcompressor(self, opts=None):
3468 3468 return self.zlibrevlogcompressor()
3469 3469
3470 3470 compengines.register(_zlibengine())
3471 3471
3472 3472 class _bz2engine(compressionengine):
3473 3473 def name(self):
3474 3474 return 'bz2'
3475 3475
3476 3476 def bundletype(self):
3477 3477 """An algorithm that produces smaller bundles than ``gzip``.
3478 3478
3479 3479 All Mercurial clients should support this format.
3480 3480
3481 3481 This engine will likely produce smaller bundles than ``gzip`` but
3482 3482 will be significantly slower, both during compression and
3483 3483 decompression.
3484 3484
3485 3485 If available, the ``zstd`` engine can yield similar or better
3486 3486 compression at much higher speeds.
3487 3487 """
3488 3488 return 'bzip2', 'BZ'
3489 3489
3490 3490 # We declare a protocol name but don't advertise by default because
3491 3491 # it is slow.
3492 3492 def wireprotosupport(self):
3493 3493 return compewireprotosupport('bzip2', 0, 0)
3494 3494
3495 3495 def compressstream(self, it, opts=None):
3496 3496 opts = opts or {}
3497 3497 z = bz2.BZ2Compressor(opts.get('level', 9))
3498 3498 for chunk in it:
3499 3499 data = z.compress(chunk)
3500 3500 if data:
3501 3501 yield data
3502 3502
3503 3503 yield z.flush()
3504 3504
3505 3505 def decompressorreader(self, fh):
3506 3506 def gen():
3507 3507 d = bz2.BZ2Decompressor()
3508 3508 for chunk in filechunkiter(fh):
3509 3509 yield d.decompress(chunk)
3510 3510
3511 3511 return chunkbuffer(gen())
3512 3512
3513 3513 compengines.register(_bz2engine())
3514 3514
3515 3515 class _truncatedbz2engine(compressionengine):
3516 3516 def name(self):
3517 3517 return 'bz2truncated'
3518 3518
3519 3519 def bundletype(self):
3520 3520 return None, '_truncatedBZ'
3521 3521
3522 3522 # We don't implement compressstream because it is hackily handled elsewhere.
3523 3523
3524 3524 def decompressorreader(self, fh):
3525 3525 def gen():
3526 3526 # The input stream doesn't have the 'BZ' header. So add it back.
3527 3527 d = bz2.BZ2Decompressor()
3528 3528 d.decompress('BZ')
3529 3529 for chunk in filechunkiter(fh):
3530 3530 yield d.decompress(chunk)
3531 3531
3532 3532 return chunkbuffer(gen())
3533 3533
3534 3534 compengines.register(_truncatedbz2engine())
3535 3535
3536 3536 class _noopengine(compressionengine):
3537 3537 def name(self):
3538 3538 return 'none'
3539 3539
3540 3540 def bundletype(self):
3541 3541 """No compression is performed.
3542 3542
3543 3543 Use this compression engine to explicitly disable compression.
3544 3544 """
3545 3545 return 'none', 'UN'
3546 3546
3547 3547 # Clients always support uncompressed payloads. Servers don't because
3548 3548 # unless you are on a fast network, uncompressed payloads can easily
3549 3549 # saturate your network pipe.
3550 3550 def wireprotosupport(self):
3551 3551 return compewireprotosupport('none', 0, 10)
3552 3552
3553 3553 # We don't implement revlogheader because it is handled specially
3554 3554 # in the revlog class.
3555 3555
3556 3556 def compressstream(self, it, opts=None):
3557 3557 return it
3558 3558
3559 3559 def decompressorreader(self, fh):
3560 3560 return fh
3561 3561
3562 3562 class nooprevlogcompressor(object):
3563 3563 def compress(self, data):
3564 3564 return None
3565 3565
3566 3566 def revlogcompressor(self, opts=None):
3567 3567 return self.nooprevlogcompressor()
3568 3568
3569 3569 compengines.register(_noopengine())
3570 3570
3571 3571 class _zstdengine(compressionengine):
3572 3572 def name(self):
3573 3573 return 'zstd'
3574 3574
3575 3575 @propertycache
3576 3576 def _module(self):
3577 3577 # Not all installs have the zstd module available. So defer importing
3578 3578 # until first access.
3579 3579 try:
3580 3580 from . import zstd
3581 3581 # Force delayed import.
3582 3582 zstd.__version__
3583 3583 return zstd
3584 3584 except ImportError:
3585 3585 return None
3586 3586
3587 3587 def available(self):
3588 3588 return bool(self._module)
3589 3589
3590 3590 def bundletype(self):
3591 3591 """A modern compression algorithm that is fast and highly flexible.
3592 3592
3593 3593 Only supported by Mercurial 4.1 and newer clients.
3594 3594
3595 3595 With the default settings, zstd compression is both faster and yields
3596 3596 better compression than ``gzip``. It also frequently yields better
3597 3597 compression than ``bzip2`` while operating at much higher speeds.
3598 3598
3599 3599 If this engine is available and backwards compatibility is not a
3600 3600 concern, it is likely the best available engine.
3601 3601 """
3602 3602 return 'zstd', 'ZS'
3603 3603
3604 3604 def wireprotosupport(self):
3605 3605 return compewireprotosupport('zstd', 50, 50)
3606 3606
3607 3607 def revlogheader(self):
3608 3608 return '\x28'
3609 3609
3610 3610 def compressstream(self, it, opts=None):
3611 3611 opts = opts or {}
3612 3612 # zstd level 3 is almost always significantly faster than zlib
3613 3613 # while providing no worse compression. It strikes a good balance
3614 3614 # between speed and compression.
3615 3615 level = opts.get('level', 3)
3616 3616
3617 3617 zstd = self._module
3618 3618 z = zstd.ZstdCompressor(level=level).compressobj()
3619 3619 for chunk in it:
3620 3620 data = z.compress(chunk)
3621 3621 if data:
3622 3622 yield data
3623 3623
3624 3624 yield z.flush()
3625 3625
3626 3626 def decompressorreader(self, fh):
3627 3627 zstd = self._module
3628 3628 dctx = zstd.ZstdDecompressor()
3629 3629 return chunkbuffer(dctx.read_from(fh))
3630 3630
3631 3631 class zstdrevlogcompressor(object):
3632 3632 def __init__(self, zstd, level=3):
3633 3633 # Writing the content size adds a few bytes to the output. However,
3634 3634 # it allows decompression to be more optimal since we can
3635 3635 # pre-allocate a buffer to hold the result.
3636 3636 self._cctx = zstd.ZstdCompressor(level=level,
3637 3637 write_content_size=True)
3638 3638 self._dctx = zstd.ZstdDecompressor()
3639 3639 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3640 3640 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3641 3641
3642 3642 def compress(self, data):
3643 3643 insize = len(data)
3644 3644 # Caller handles empty input case.
3645 3645 assert insize > 0
3646 3646
3647 3647 if insize < 50:
3648 3648 return None
3649 3649
3650 3650 elif insize <= 1000000:
3651 3651 compressed = self._cctx.compress(data)
3652 3652 if len(compressed) < insize:
3653 3653 return compressed
3654 3654 return None
3655 3655 else:
3656 3656 z = self._cctx.compressobj()
3657 3657 chunks = []
3658 3658 pos = 0
3659 3659 while pos < insize:
3660 3660 pos2 = pos + self._compinsize
3661 3661 chunk = z.compress(data[pos:pos2])
3662 3662 if chunk:
3663 3663 chunks.append(chunk)
3664 3664 pos = pos2
3665 3665 chunks.append(z.flush())
3666 3666
3667 3667 if sum(map(len, chunks)) < insize:
3668 3668 return ''.join(chunks)
3669 3669 return None
3670 3670
3671 3671 def decompress(self, data):
3672 3672 insize = len(data)
3673 3673
3674 3674 try:
3675 3675 # This was measured to be faster than other streaming
3676 3676 # decompressors.
3677 3677 dobj = self._dctx.decompressobj()
3678 3678 chunks = []
3679 3679 pos = 0
3680 3680 while pos < insize:
3681 3681 pos2 = pos + self._decompinsize
3682 3682 chunk = dobj.decompress(data[pos:pos2])
3683 3683 if chunk:
3684 3684 chunks.append(chunk)
3685 3685 pos = pos2
3686 3686 # Frame should be exhausted, so no finish() API.
3687 3687
3688 3688 return ''.join(chunks)
3689 3689 except Exception as e:
3690 3690 raise error.RevlogError(_('revlog decompress error: %s') %
3691 3691 str(e))
3692 3692
3693 3693 def revlogcompressor(self, opts=None):
3694 3694 opts = opts or {}
3695 3695 return self.zstdrevlogcompressor(self._module,
3696 3696 level=opts.get('level', 3))
3697 3697
3698 3698 compengines.register(_zstdengine())
3699 3699
3700 3700 def bundlecompressiontopics():
3701 3701 """Obtains a list of available bundle compressions for use in help."""
3702 3702 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3703 3703 items = {}
3704 3704
3705 3705 # We need to format the docstring. So use a dummy object/type to hold it
3706 3706 # rather than mutating the original.
3707 3707 class docobject(object):
3708 3708 pass
3709 3709
3710 3710 for name in compengines:
3711 3711 engine = compengines[name]
3712 3712
3713 3713 if not engine.available():
3714 3714 continue
3715 3715
3716 3716 bt = engine.bundletype()
3717 3717 if not bt or not bt[0]:
3718 3718 continue
3719 3719
3720 3720 doc = pycompat.sysstr('``%s``\n %s') % (
3721 3721 bt[0], engine.bundletype.__doc__)
3722 3722
3723 3723 value = docobject()
3724 3724 value.__doc__ = doc
3725 3725
3726 3726 items[bt[0]] = value
3727 3727
3728 3728 return items
3729 3729
3730 3730 # convenient shortcut
3731 3731 dst = debugstacktrace
@@ -1,802 +1,802
1 1 #
2 2 # This is the mercurial setup script.
3 3 #
4 4 # 'python setup.py install', or
5 5 # 'python setup.py --help' for more options
6 6
7 7 import sys, platform
8 8 if sys.version_info < (2, 7, 0, 'final'):
9 9 raise SystemExit('Mercurial requires Python 2.7 or later.')
10 10
11 11 if sys.version_info[0] >= 3:
12 12 printf = eval('print')
13 13 libdir_escape = 'unicode_escape'
14 14 else:
15 15 libdir_escape = 'string_escape'
16 16 def printf(*args, **kwargs):
17 17 f = kwargs.get('file', sys.stdout)
18 18 end = kwargs.get('end', '\n')
19 19 f.write(b' '.join(args) + end)
20 20
21 21 # Solaris Python packaging brain damage
22 22 try:
23 23 import hashlib
24 24 sha = hashlib.sha1()
25 25 except ImportError:
26 26 try:
27 27 import sha
28 28 sha.sha # silence unused import warning
29 29 except ImportError:
30 30 raise SystemExit(
31 31 "Couldn't import standard hashlib (incomplete Python install).")
32 32
33 33 try:
34 34 import zlib
35 35 zlib.compressobj # silence unused import warning
36 36 except ImportError:
37 37 raise SystemExit(
38 38 "Couldn't import standard zlib (incomplete Python install).")
39 39
40 40 # The base IronPython distribution (as of 2.7.1) doesn't support bz2
41 41 isironpython = False
42 42 try:
43 43 isironpython = (platform.python_implementation()
44 44 .lower().find("ironpython") != -1)
45 45 except AttributeError:
46 46 pass
47 47
48 48 if isironpython:
49 49 sys.stderr.write("warning: IronPython detected (no bz2 support)\n")
50 50 else:
51 51 try:
52 52 import bz2
53 53 bz2.BZ2Compressor # silence unused import warning
54 54 except ImportError:
55 55 raise SystemExit(
56 56 "Couldn't import standard bz2 (incomplete Python install).")
57 57
58 58 ispypy = "PyPy" in sys.version
59 59
60 60 import ctypes
61 61 import os, stat, subprocess, time
62 62 import re
63 63 import shutil
64 64 import tempfile
65 65 from distutils import log
66 66 # We have issues with setuptools on some platforms and builders. Until
67 67 # those are resolved, setuptools is opt-in except for platforms where
68 68 # we don't have issues.
69 69 if os.name == 'nt' or 'FORCE_SETUPTOOLS' in os.environ:
70 70 from setuptools import setup
71 71 else:
72 72 from distutils.core import setup
73 73 from distutils.ccompiler import new_compiler
74 74 from distutils.core import Command, Extension
75 75 from distutils.dist import Distribution
76 76 from distutils.command.build import build
77 77 from distutils.command.build_ext import build_ext
78 78 from distutils.command.build_py import build_py
79 79 from distutils.command.build_scripts import build_scripts
80 80 from distutils.command.install_lib import install_lib
81 81 from distutils.command.install_scripts import install_scripts
82 82 from distutils.spawn import spawn, find_executable
83 83 from distutils import file_util
84 84 from distutils.errors import (
85 85 CCompilerError,
86 86 DistutilsError,
87 87 DistutilsExecError,
88 88 )
89 89 from distutils.sysconfig import get_python_inc, get_config_var
90 90 from distutils.version import StrictVersion
91 91
92 92 scripts = ['hg']
93 93 if os.name == 'nt':
94 94 # We remove hg.bat if we are able to build hg.exe.
95 95 scripts.append('contrib/win32/hg.bat')
96 96
97 97 def cancompile(cc, code):
98 98 tmpdir = tempfile.mkdtemp(prefix='hg-install-')
99 99 devnull = oldstderr = None
100 100 try:
101 101 fname = os.path.join(tmpdir, 'testcomp.c')
102 102 f = open(fname, 'w')
103 103 f.write(code)
104 104 f.close()
105 105 # Redirect stderr to /dev/null to hide any error messages
106 106 # from the compiler.
107 107 # This will have to be changed if we ever have to check
108 108 # for a function on Windows.
109 109 devnull = open('/dev/null', 'w')
110 110 oldstderr = os.dup(sys.stderr.fileno())
111 111 os.dup2(devnull.fileno(), sys.stderr.fileno())
112 112 objects = cc.compile([fname], output_dir=tmpdir)
113 113 cc.link_executable(objects, os.path.join(tmpdir, "a.out"))
114 114 return True
115 115 except Exception:
116 116 return False
117 117 finally:
118 118 if oldstderr is not None:
119 119 os.dup2(oldstderr, sys.stderr.fileno())
120 120 if devnull is not None:
121 121 devnull.close()
122 122 shutil.rmtree(tmpdir)
123 123
124 124 # simplified version of distutils.ccompiler.CCompiler.has_function
125 125 # that actually removes its temporary files.
126 126 def hasfunction(cc, funcname):
127 127 code = 'int main(void) { %s(); }\n' % funcname
128 128 return cancompile(cc, code)
129 129
130 130 def hasheader(cc, headername):
131 131 code = '#include <%s>\nint main(void) { return 0; }\n' % headername
132 132 return cancompile(cc, code)
133 133
134 134 # py2exe needs to be installed to work
135 135 try:
136 136 import py2exe
137 137 py2exe.Distribution # silence unused import warning
138 138 py2exeloaded = True
139 139 # import py2exe's patched Distribution class
140 140 from distutils.core import Distribution
141 141 except ImportError:
142 142 py2exeloaded = False
143 143
144 144 def runcmd(cmd, env):
145 145 if (sys.platform == 'plan9'
146 146 and (sys.version_info[0] == 2 and sys.version_info[1] < 7)):
147 147 # subprocess kludge to work around issues in half-baked Python
148 148 # ports, notably bichued/python:
149 149 _, out, err = os.popen3(cmd)
150 150 return str(out), str(err)
151 151 else:
152 152 p = subprocess.Popen(cmd, stdout=subprocess.PIPE,
153 153 stderr=subprocess.PIPE, env=env)
154 154 out, err = p.communicate()
155 155 return out, err
156 156
157 157 def runhg(cmd, env):
158 158 out, err = runcmd(cmd, env)
159 159 # If root is executing setup.py, but the repository is owned by
160 160 # another user (as in "sudo python setup.py install") we will get
161 161 # trust warnings since the .hg/hgrc file is untrusted. That is
162 162 # fine, we don't want to load it anyway. Python may warn about
163 163 # a missing __init__.py in mercurial/locale, we also ignore that.
164 164 err = [e for e in err.splitlines()
165 165 if not e.startswith(b'not trusting file') \
166 166 and not e.startswith(b'warning: Not importing') \
167 167 and not e.startswith(b'obsolete feature not enabled')]
168 168 if err:
169 169 printf("stderr from '%s':" % (' '.join(cmd)), file=sys.stderr)
170 170 printf(b'\n'.join([b' ' + e for e in err]), file=sys.stderr)
171 171 return ''
172 172 return out
173 173
174 174 version = ''
175 175
176 176 # Execute hg out of this directory with a custom environment which takes care
177 177 # to not use any hgrc files and do no localization.
178 178 env = {'HGMODULEPOLICY': 'py',
179 179 'HGRCPATH': '',
180 180 'LANGUAGE': 'C',
181 181 'PATH': ''} # make pypi modules that use os.environ['PATH'] happy
182 182 if 'LD_LIBRARY_PATH' in os.environ:
183 183 env['LD_LIBRARY_PATH'] = os.environ['LD_LIBRARY_PATH']
184 184 if 'SystemRoot' in os.environ:
185 185 # Copy SystemRoot into the custom environment for Python 2.6
186 186 # under Windows. Otherwise, the subprocess will fail with
187 187 # error 0xc0150004. See: http://bugs.python.org/issue3440
188 188 env['SystemRoot'] = os.environ['SystemRoot']
189 189
190 190 if os.path.isdir('.hg'):
191 191 cmd = [sys.executable, 'hg', 'log', '-r', '.', '--template', '{tags}\n']
192 192 numerictags = [t for t in runhg(cmd, env).split() if t[0].isdigit()]
193 193 hgid = runhg([sys.executable, 'hg', 'id', '-i'], env).strip()
194 194 if numerictags: # tag(s) found
195 195 version = numerictags[-1]
196 196 if hgid.endswith('+'): # propagate the dirty status to the tag
197 197 version += '+'
198 198 else: # no tag found
199 199 ltagcmd = [sys.executable, 'hg', 'parents', '--template',
200 200 '{latesttag}']
201 201 ltag = runhg(ltagcmd, env)
202 202 changessincecmd = [sys.executable, 'hg', 'log', '-T', 'x\n', '-r',
203 203 "only(.,'%s')" % ltag]
204 204 changessince = len(runhg(changessincecmd, env).splitlines())
205 205 version = '%s+%s-%s' % (ltag, changessince, hgid)
206 206 if version.endswith('+'):
207 207 version += time.strftime('%Y%m%d')
208 208 elif os.path.exists('.hg_archival.txt'):
209 209 kw = dict([[t.strip() for t in l.split(':', 1)]
210 210 for l in open('.hg_archival.txt')])
211 211 if 'tag' in kw:
212 212 version = kw['tag']
213 213 elif 'latesttag' in kw:
214 214 if 'changessincelatesttag' in kw:
215 215 version = '%(latesttag)s+%(changessincelatesttag)s-%(node).12s' % kw
216 216 else:
217 217 version = '%(latesttag)s+%(latesttagdistance)s-%(node).12s' % kw
218 218 else:
219 219 version = kw.get('node', '')[:12]
220 220
221 221 if version:
222 222 with open("mercurial/__version__.py", "w") as f:
223 223 f.write('# this file is autogenerated by setup.py\n')
224 224 f.write('version = "%s"\n' % version)
225 225
226 226 try:
227 227 oldpolicy = os.environ.get('HGMODULEPOLICY', None)
228 228 os.environ['HGMODULEPOLICY'] = 'py'
229 229 from mercurial import __version__
230 230 version = __version__.version
231 231 except ImportError:
232 232 version = 'unknown'
233 233 finally:
234 234 if oldpolicy is None:
235 235 del os.environ['HGMODULEPOLICY']
236 236 else:
237 237 os.environ['HGMODULEPOLICY'] = oldpolicy
238 238
239 239 class hgbuild(build):
240 240 # Insert hgbuildmo first so that files in mercurial/locale/ are found
241 241 # when build_py is run next.
242 242 sub_commands = [('build_mo', None)] + build.sub_commands
243 243
244 244 class hgbuildmo(build):
245 245
246 246 description = "build translations (.mo files)"
247 247
248 248 def run(self):
249 249 if not find_executable('msgfmt'):
250 250 self.warn("could not find msgfmt executable, no translations "
251 251 "will be built")
252 252 return
253 253
254 254 podir = 'i18n'
255 255 if not os.path.isdir(podir):
256 256 self.warn("could not find %s/ directory" % podir)
257 257 return
258 258
259 259 join = os.path.join
260 260 for po in os.listdir(podir):
261 261 if not po.endswith('.po'):
262 262 continue
263 263 pofile = join(podir, po)
264 264 modir = join('locale', po[:-3], 'LC_MESSAGES')
265 265 mofile = join(modir, 'hg.mo')
266 266 mobuildfile = join('mercurial', mofile)
267 267 cmd = ['msgfmt', '-v', '-o', mobuildfile, pofile]
268 268 if sys.platform != 'sunos5':
269 269 # msgfmt on Solaris does not know about -c
270 270 cmd.append('-c')
271 271 self.mkpath(join('mercurial', modir))
272 272 self.make_file([pofile], mobuildfile, spawn, (cmd,))
273 273
274 274
275 275 class hgdist(Distribution):
276 276 pure = False
277 277 cffi = ispypy
278 278
279 279 global_options = Distribution.global_options + \
280 280 [('pure', None, "use pure (slow) Python "
281 281 "code instead of C extensions"),
282 282 ]
283 283
284 284 def has_ext_modules(self):
285 285 # self.ext_modules is emptied in hgbuildpy.finalize_options which is
286 286 # too late for some cases
287 287 return not self.pure and Distribution.has_ext_modules(self)
288 288
289 289 # This is ugly as a one-liner. So use a variable.
290 290 buildextnegops = dict(getattr(build_ext, 'negative_options', {}))
291 291 buildextnegops['no-zstd'] = 'zstd'
292 292
293 293 class hgbuildext(build_ext):
294 294 user_options = build_ext.user_options + [
295 295 ('zstd', None, 'compile zstd bindings [default]'),
296 296 ('no-zstd', None, 'do not compile zstd bindings'),
297 297 ]
298 298
299 299 boolean_options = build_ext.boolean_options + ['zstd']
300 300 negative_opt = buildextnegops
301 301
302 302 def initialize_options(self):
303 303 self.zstd = True
304 304 return build_ext.initialize_options(self)
305 305
306 306 def build_extensions(self):
307 307 # Filter out zstd if disabled via argument.
308 308 if not self.zstd:
309 309 self.extensions = [e for e in self.extensions
310 310 if e.name != 'mercurial.zstd']
311 311
312 312 return build_ext.build_extensions(self)
313 313
314 314 def build_extension(self, ext):
315 315 try:
316 316 build_ext.build_extension(self, ext)
317 317 except CCompilerError:
318 318 if not getattr(ext, 'optional', False):
319 319 raise
320 320 log.warn("Failed to build optional extension '%s' (skipping)",
321 321 ext.name)
322 322
323 323 class hgbuildscripts(build_scripts):
324 324 def run(self):
325 325 if os.name != 'nt' or self.distribution.pure:
326 326 return build_scripts.run(self)
327 327
328 328 exebuilt = False
329 329 try:
330 330 self.run_command('build_hgexe')
331 331 exebuilt = True
332 332 except (DistutilsError, CCompilerError):
333 333 log.warn('failed to build optional hg.exe')
334 334
335 335 if exebuilt:
336 336 # Copying hg.exe to the scripts build directory ensures it is
337 337 # installed by the install_scripts command.
338 338 hgexecommand = self.get_finalized_command('build_hgexe')
339 339 dest = os.path.join(self.build_dir, 'hg.exe')
340 340 self.mkpath(self.build_dir)
341 341 self.copy_file(hgexecommand.hgexepath, dest)
342 342
343 343 # Remove hg.bat because it is redundant with hg.exe.
344 344 self.scripts.remove('contrib/win32/hg.bat')
345 345
346 346 return build_scripts.run(self)
347 347
348 348 class hgbuildpy(build_py):
349 349 def finalize_options(self):
350 350 build_py.finalize_options(self)
351 351
352 352 if self.distribution.pure:
353 353 self.distribution.ext_modules = []
354 354 elif self.distribution.cffi:
355 355 from mercurial.cffi import (
356 356 bdiff,
357 357 mpatch,
358 358 )
359 359 exts = [mpatch.ffi.distutils_extension(),
360 360 bdiff.ffi.distutils_extension()]
361 361 # cffi modules go here
362 362 if sys.platform == 'darwin':
363 363 from mercurial.cffi import osutil
364 364 exts.append(osutil.ffi.distutils_extension())
365 365 self.distribution.ext_modules = exts
366 366 else:
367 367 h = os.path.join(get_python_inc(), 'Python.h')
368 368 if not os.path.exists(h):
369 369 raise SystemExit('Python headers are required to build '
370 370 'Mercurial but weren\'t found in %s' % h)
371 371
372 372 def run(self):
373 373 if self.distribution.pure:
374 374 modulepolicy = 'py'
375 375 elif self.build_lib == '.':
376 376 # in-place build should run without rebuilding C extensions
377 377 modulepolicy = 'allow'
378 378 else:
379 379 modulepolicy = 'c'
380 380 with open("mercurial/__modulepolicy__.py", "w") as f:
381 381 f.write('# this file is autogenerated by setup.py\n')
382 382 f.write('modulepolicy = b"%s"\n' % modulepolicy)
383 383
384 384 build_py.run(self)
385 385
386 386 class buildhgextindex(Command):
387 387 description = 'generate prebuilt index of hgext (for frozen package)'
388 388 user_options = []
389 389 _indexfilename = 'hgext/__index__.py'
390 390
391 391 def initialize_options(self):
392 392 pass
393 393
394 394 def finalize_options(self):
395 395 pass
396 396
397 397 def run(self):
398 398 if os.path.exists(self._indexfilename):
399 399 with open(self._indexfilename, 'w') as f:
400 400 f.write('# empty\n')
401 401
402 402 # here no extension enabled, disabled() lists up everything
403 403 code = ('import pprint; from mercurial import extensions; '
404 404 'pprint.pprint(extensions.disabled())')
405 405 out, err = runcmd([sys.executable, '-c', code], env)
406 406 if err:
407 407 raise DistutilsExecError(err)
408 408
409 409 with open(self._indexfilename, 'w') as f:
410 410 f.write('# this file is autogenerated by setup.py\n')
411 411 f.write('docs = ')
412 412 f.write(out)
413 413
414 414 class buildhgexe(build_ext):
415 415 description = 'compile hg.exe from mercurial/exewrapper.c'
416 416
417 417 def build_extensions(self):
418 418 if os.name != 'nt':
419 419 return
420 420 if isinstance(self.compiler, HackedMingw32CCompiler):
421 421 self.compiler.compiler_so = self.compiler.compiler # no -mdll
422 422 self.compiler.dll_libraries = [] # no -lmsrvc90
423 423
424 424 # Different Python installs can have different Python library
425 425 # names. e.g. the official CPython distribution uses pythonXY.dll
426 426 # and MinGW uses libpythonX.Y.dll.
427 427 _kernel32 = ctypes.windll.kernel32
428 428 _kernel32.GetModuleFileNameA.argtypes = [ctypes.c_void_p,
429 429 ctypes.c_void_p,
430 430 ctypes.c_ulong]
431 431 _kernel32.GetModuleFileNameA.restype = ctypes.c_ulong
432 432 size = 1000
433 433 buf = ctypes.create_string_buffer(size + 1)
434 434 filelen = _kernel32.GetModuleFileNameA(sys.dllhandle, ctypes.byref(buf),
435 435 size)
436 436
437 437 if filelen > 0 and filelen != size:
438 438 dllbasename = os.path.basename(buf.value)
439 439 if not dllbasename.lower().endswith('.dll'):
440 440 raise SystemExit('Python DLL does not end with .dll: %s' %
441 441 dllbasename)
442 442 pythonlib = dllbasename[:-4]
443 443 else:
444 444 log.warn('could not determine Python DLL filename; '
445 445 'assuming pythonXY')
446 446
447 447 hv = sys.hexversion
448 448 pythonlib = 'python%d%d' % (hv >> 24, (hv >> 16) & 0xff)
449 449
450 450 log.info('using %s as Python library name' % pythonlib)
451 451 with open('mercurial/hgpythonlib.h', 'wb') as f:
452 452 f.write('/* this file is autogenerated by setup.py */\n')
453 453 f.write('#define HGPYTHONLIB "%s"\n' % pythonlib)
454 454 objects = self.compiler.compile(['mercurial/exewrapper.c'],
455 455 output_dir=self.build_temp)
456 456 dir = os.path.dirname(self.get_ext_fullpath('dummy'))
457 457 target = os.path.join(dir, 'hg')
458 458 self.compiler.link_executable(objects, target,
459 459 libraries=[],
460 460 output_dir=self.build_temp)
461 461
462 462 @property
463 463 def hgexepath(self):
464 464 dir = os.path.dirname(self.get_ext_fullpath('dummy'))
465 465 return os.path.join(self.build_temp, dir, 'hg.exe')
466 466
467 467 class hginstalllib(install_lib):
468 468 '''
469 469 This is a specialization of install_lib that replaces the copy_file used
470 470 there so that it supports setting the mode of files after copying them,
471 471 instead of just preserving the mode that the files originally had. If your
472 472 system has a umask of something like 027, preserving the permissions when
473 473 copying will lead to a broken install.
474 474
475 475 Note that just passing keep_permissions=False to copy_file would be
476 476 insufficient, as it might still be applying a umask.
477 477 '''
478 478
479 479 def run(self):
480 480 realcopyfile = file_util.copy_file
481 481 def copyfileandsetmode(*args, **kwargs):
482 482 src, dst = args[0], args[1]
483 483 dst, copied = realcopyfile(*args, **kwargs)
484 484 if copied:
485 485 st = os.stat(src)
486 486 # Persist executable bit (apply it to group and other if user
487 487 # has it)
488 488 if st[stat.ST_MODE] & stat.S_IXUSR:
489 489 setmode = int('0755', 8)
490 490 else:
491 491 setmode = int('0644', 8)
492 492 m = stat.S_IMODE(st[stat.ST_MODE])
493 493 m = (m & ~int('0777', 8)) | setmode
494 494 os.chmod(dst, m)
495 495 file_util.copy_file = copyfileandsetmode
496 496 try:
497 497 install_lib.run(self)
498 498 finally:
499 499 file_util.copy_file = realcopyfile
500 500
501 501 class hginstallscripts(install_scripts):
502 502 '''
503 503 This is a specialization of install_scripts that replaces the @LIBDIR@ with
504 504 the configured directory for modules. If possible, the path is made relative
505 505 to the directory for scripts.
506 506 '''
507 507
508 508 def initialize_options(self):
509 509 install_scripts.initialize_options(self)
510 510
511 511 self.install_lib = None
512 512
513 513 def finalize_options(self):
514 514 install_scripts.finalize_options(self)
515 515 self.set_undefined_options('install',
516 516 ('install_lib', 'install_lib'))
517 517
518 518 def run(self):
519 519 install_scripts.run(self)
520 520
521 521 # It only makes sense to replace @LIBDIR@ with the install path if
522 522 # the install path is known. For wheels, the logic below calculates
523 523 # the libdir to be "../..". This is because the internal layout of a
524 524 # wheel archive looks like:
525 525 #
526 526 # mercurial-3.6.1.data/scripts/hg
527 527 # mercurial/__init__.py
528 528 #
529 529 # When installing wheels, the subdirectories of the "<pkg>.data"
530 530 # directory are translated to system local paths and files therein
531 531 # are copied in place. The mercurial/* files are installed into the
532 532 # site-packages directory. However, the site-packages directory
533 533 # isn't known until wheel install time. This means we have no clue
534 534 # at wheel generation time what the installed site-packages directory
535 535 # will be. And, wheels don't appear to provide the ability to register
536 536 # custom code to run during wheel installation. This all means that
537 537 # we can't reliably set the libdir in wheels: the default behavior
538 538 # of looking in sys.path must do.
539 539
540 540 if (os.path.splitdrive(self.install_dir)[0] !=
541 541 os.path.splitdrive(self.install_lib)[0]):
542 542 # can't make relative paths from one drive to another, so use an
543 543 # absolute path instead
544 544 libdir = self.install_lib
545 545 else:
546 546 common = os.path.commonprefix((self.install_dir, self.install_lib))
547 547 rest = self.install_dir[len(common):]
548 548 uplevel = len([n for n in os.path.split(rest) if n])
549 549
550 550 libdir = uplevel * ('..' + os.sep) + self.install_lib[len(common):]
551 551
552 552 for outfile in self.outfiles:
553 553 with open(outfile, 'rb') as fp:
554 554 data = fp.read()
555 555
556 556 # skip binary files
557 557 if b'\0' in data:
558 558 continue
559 559
560 560 # During local installs, the shebang will be rewritten to the final
561 561 # install path. During wheel packaging, the shebang has a special
562 562 # value.
563 563 if data.startswith(b'#!python'):
564 564 log.info('not rewriting @LIBDIR@ in %s because install path '
565 565 'not known' % outfile)
566 566 continue
567 567
568 568 data = data.replace(b'@LIBDIR@', libdir.encode(libdir_escape))
569 569 with open(outfile, 'wb') as fp:
570 570 fp.write(data)
571 571
572 572 cmdclass = {'build': hgbuild,
573 573 'build_mo': hgbuildmo,
574 574 'build_ext': hgbuildext,
575 575 'build_py': hgbuildpy,
576 576 'build_scripts': hgbuildscripts,
577 577 'build_hgextindex': buildhgextindex,
578 578 'install_lib': hginstalllib,
579 579 'install_scripts': hginstallscripts,
580 580 'build_hgexe': buildhgexe,
581 581 }
582 582
583 583 packages = ['mercurial',
584 584 'mercurial.cext',
585 585 'mercurial.hgweb',
586 586 'mercurial.httpclient',
587 587 'mercurial.pure',
588 588 'hgext', 'hgext.convert', 'hgext.fsmonitor',
589 589 'hgext.fsmonitor.pywatchman', 'hgext.highlight',
590 590 'hgext.largefiles', 'hgext.zeroconf', 'hgext3rd']
591 591
592 592 common_depends = ['mercurial/bitmanipulation.h',
593 593 'mercurial/compat.h',
594 594 'mercurial/util.h']
595 595 common_include_dirs = ['mercurial']
596 596
597 597 osutil_cflags = []
598 598 osutil_ldflags = []
599 599
600 600 # platform specific macros
601 601 for plat, func in [('bsd', 'setproctitle')]:
602 602 if re.search(plat, sys.platform) and hasfunction(new_compiler(), func):
603 603 osutil_cflags.append('-DHAVE_%s' % func.upper())
604 604
605 605 for plat, macro, code in [
606 606 ('bsd|darwin', 'BSD_STATFS', '''
607 607 #include <sys/param.h>
608 608 #include <sys/mount.h>
609 609 int main() { struct statfs s; return sizeof(s.f_fstypename); }
610 610 '''),
611 611 ('linux', 'LINUX_STATFS', '''
612 612 #include <linux/magic.h>
613 613 #include <sys/vfs.h>
614 614 int main() { struct statfs s; return sizeof(s.f_type); }
615 615 '''),
616 616 ]:
617 617 if re.search(plat, sys.platform) and cancompile(new_compiler(), code):
618 618 osutil_cflags.append('-DHAVE_%s' % macro)
619 619
620 620 if sys.platform == 'darwin':
621 621 osutil_ldflags += ['-framework', 'ApplicationServices']
622 622
623 623 extmodules = [
624 Extension('mercurial.base85', ['mercurial/base85.c'],
624 Extension('mercurial.cext.base85', ['mercurial/cext/base85.c'],
625 625 include_dirs=common_include_dirs,
626 626 depends=common_depends),
627 627 Extension('mercurial.bdiff', ['mercurial/bdiff.c',
628 628 'mercurial/bdiff_module.c'],
629 629 include_dirs=common_include_dirs,
630 630 depends=common_depends + ['mercurial/bdiff.h']),
631 631 Extension('mercurial.diffhelpers', ['mercurial/diffhelpers.c'],
632 632 include_dirs=common_include_dirs,
633 633 depends=common_depends),
634 634 Extension('mercurial.mpatch', ['mercurial/mpatch.c',
635 635 'mercurial/mpatch_module.c'],
636 636 include_dirs=common_include_dirs,
637 637 depends=common_depends),
638 638 Extension('mercurial.parsers', ['mercurial/dirs.c',
639 639 'mercurial/manifest.c',
640 640 'mercurial/parsers.c',
641 641 'mercurial/pathencode.c'],
642 642 include_dirs=common_include_dirs,
643 643 depends=common_depends),
644 644 Extension('mercurial.cext.osutil', ['mercurial/cext/osutil.c'],
645 645 include_dirs=common_include_dirs,
646 646 extra_compile_args=osutil_cflags,
647 647 extra_link_args=osutil_ldflags,
648 648 depends=common_depends),
649 649 Extension('hgext.fsmonitor.pywatchman.bser',
650 650 ['hgext/fsmonitor/pywatchman/bser.c']),
651 651 ]
652 652
653 653 sys.path.insert(0, 'contrib/python-zstandard')
654 654 import setup_zstd
655 655 extmodules.append(setup_zstd.get_c_extension(name='mercurial.zstd'))
656 656
657 657 try:
658 658 from distutils import cygwinccompiler
659 659
660 660 # the -mno-cygwin option has been deprecated for years
661 661 compiler = cygwinccompiler.Mingw32CCompiler
662 662
663 663 class HackedMingw32CCompiler(cygwinccompiler.Mingw32CCompiler):
664 664 def __init__(self, *args, **kwargs):
665 665 compiler.__init__(self, *args, **kwargs)
666 666 for i in 'compiler compiler_so linker_exe linker_so'.split():
667 667 try:
668 668 getattr(self, i).remove('-mno-cygwin')
669 669 except ValueError:
670 670 pass
671 671
672 672 cygwinccompiler.Mingw32CCompiler = HackedMingw32CCompiler
673 673 except ImportError:
674 674 # the cygwinccompiler package is not available on some Python
675 675 # distributions like the ones from the optware project for Synology
676 676 # DiskStation boxes
677 677 class HackedMingw32CCompiler(object):
678 678 pass
679 679
680 680 packagedata = {'mercurial': ['locale/*/LC_MESSAGES/hg.mo',
681 681 'help/*.txt',
682 682 'help/internals/*.txt',
683 683 'default.d/*.rc',
684 684 'dummycert.pem']}
685 685
686 686 def ordinarypath(p):
687 687 return p and p[0] != '.' and p[-1] != '~'
688 688
689 689 for root in ('templates',):
690 690 for curdir, dirs, files in os.walk(os.path.join('mercurial', root)):
691 691 curdir = curdir.split(os.sep, 1)[1]
692 692 dirs[:] = filter(ordinarypath, dirs)
693 693 for f in filter(ordinarypath, files):
694 694 f = os.path.join(curdir, f)
695 695 packagedata['mercurial'].append(f)
696 696
697 697 datafiles = []
698 698
699 699 # distutils expects version to be str/unicode. Converting it to
700 700 # unicode on Python 2 still works because it won't contain any
701 701 # non-ascii bytes and will be implicitly converted back to bytes
702 702 # when operated on.
703 703 assert isinstance(version, bytes)
704 704 setupversion = version.decode('ascii')
705 705
706 706 extra = {}
707 707
708 708 if py2exeloaded:
709 709 extra['console'] = [
710 710 {'script':'hg',
711 711 'copyright':'Copyright (C) 2005-2017 Matt Mackall and others',
712 712 'product_version':version}]
713 713 # sub command of 'build' because 'py2exe' does not handle sub_commands
714 714 build.sub_commands.insert(0, ('build_hgextindex', None))
715 715 # put dlls in sub directory so that they won't pollute PATH
716 716 extra['zipfile'] = 'lib/library.zip'
717 717
718 718 if os.name == 'nt':
719 719 # Windows binary file versions for exe/dll files must have the
720 720 # form W.X.Y.Z, where W,X,Y,Z are numbers in the range 0..65535
721 721 setupversion = version.split('+', 1)[0]
722 722
723 723 if sys.platform == 'darwin' and os.path.exists('/usr/bin/xcodebuild'):
724 724 version = runcmd(['/usr/bin/xcodebuild', '-version'], {})[0].splitlines()
725 725 if version:
726 726 version = version[0]
727 727 if sys.version_info[0] == 3:
728 728 version = version.decode('utf-8')
729 729 xcode4 = (version.startswith('Xcode') and
730 730 StrictVersion(version.split()[1]) >= StrictVersion('4.0'))
731 731 xcode51 = re.match(r'^Xcode\s+5\.1', version) is not None
732 732 else:
733 733 # xcodebuild returns empty on OS X Lion with XCode 4.3 not
734 734 # installed, but instead with only command-line tools. Assume
735 735 # that only happens on >= Lion, thus no PPC support.
736 736 xcode4 = True
737 737 xcode51 = False
738 738
739 739 # XCode 4.0 dropped support for ppc architecture, which is hardcoded in
740 740 # distutils.sysconfig
741 741 if xcode4:
742 742 os.environ['ARCHFLAGS'] = ''
743 743
744 744 # XCode 5.1 changes clang such that it now fails to compile if the
745 745 # -mno-fused-madd flag is passed, but the version of Python shipped with
746 746 # OS X 10.9 Mavericks includes this flag. This causes problems in all
747 747 # C extension modules, and a bug has been filed upstream at
748 748 # http://bugs.python.org/issue21244. We also need to patch this here
749 749 # so Mercurial can continue to compile in the meantime.
750 750 if xcode51:
751 751 cflags = get_config_var('CFLAGS')
752 752 if cflags and re.search(r'-mno-fused-madd\b', cflags) is not None:
753 753 os.environ['CFLAGS'] = (
754 754 os.environ.get('CFLAGS', '') + ' -Qunused-arguments')
755 755
756 756 setup(name='mercurial',
757 757 version=setupversion,
758 758 author='Matt Mackall and many others',
759 759 author_email='mercurial@mercurial-scm.org',
760 760 url='https://mercurial-scm.org/',
761 761 download_url='https://mercurial-scm.org/release/',
762 762 description=('Fast scalable distributed SCM (revision control, version '
763 763 'control) system'),
764 764 long_description=('Mercurial is a distributed SCM tool written in Python.'
765 765 ' It is used by a number of large projects that require'
766 766 ' fast, reliable distributed revision control, such as '
767 767 'Mozilla.'),
768 768 license='GNU GPLv2 or any later version',
769 769 classifiers=[
770 770 'Development Status :: 6 - Mature',
771 771 'Environment :: Console',
772 772 'Intended Audience :: Developers',
773 773 'Intended Audience :: System Administrators',
774 774 'License :: OSI Approved :: GNU General Public License (GPL)',
775 775 'Natural Language :: Danish',
776 776 'Natural Language :: English',
777 777 'Natural Language :: German',
778 778 'Natural Language :: Italian',
779 779 'Natural Language :: Japanese',
780 780 'Natural Language :: Portuguese (Brazilian)',
781 781 'Operating System :: Microsoft :: Windows',
782 782 'Operating System :: OS Independent',
783 783 'Operating System :: POSIX',
784 784 'Programming Language :: C',
785 785 'Programming Language :: Python',
786 786 'Topic :: Software Development :: Version Control',
787 787 ],
788 788 scripts=scripts,
789 789 packages=packages,
790 790 ext_modules=extmodules,
791 791 data_files=datafiles,
792 792 package_data=packagedata,
793 793 cmdclass=cmdclass,
794 794 distclass=hgdist,
795 795 options={'py2exe': {'packages': ['hgext', 'email']},
796 796 'bdist_mpkg': {'zipdist': False,
797 797 'license': 'COPYING',
798 798 'readme': 'contrib/macosx/Readme.html',
799 799 'welcome': 'contrib/macosx/Welcome.html',
800 800 },
801 801 },
802 802 **extra)
General Comments 0
You need to be logged in to leave comments. Login now