##// END OF EJS Templates
util: add `nb_bytes` argument to `copyfile` to partially copy a file...
marmoute -
r48192:9b841267 default
parent child Browse files
Show More
@@ -1,3380 +1,3394
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Olivia Mackall <olivia@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import, print_function
17 17
18 18 import abc
19 19 import collections
20 20 import contextlib
21 21 import errno
22 22 import gc
23 23 import hashlib
24 24 import itertools
25 25 import locale
26 26 import mmap
27 27 import os
28 28 import platform as pyplatform
29 29 import re as remod
30 30 import shutil
31 31 import stat
32 32 import sys
33 33 import time
34 34 import traceback
35 35 import warnings
36 36
37 37 from .node import hex
38 38 from .thirdparty import attr
39 39 from .pycompat import (
40 40 delattr,
41 41 getattr,
42 42 open,
43 43 setattr,
44 44 )
45 45 from .node import hex
46 46 from hgdemandimport import tracing
47 47 from . import (
48 48 encoding,
49 49 error,
50 50 i18n,
51 51 policy,
52 52 pycompat,
53 53 urllibcompat,
54 54 )
55 55 from .utils import (
56 56 compression,
57 57 hashutil,
58 58 procutil,
59 59 stringutil,
60 60 urlutil,
61 61 )
62 62
63 63 if pycompat.TYPE_CHECKING:
64 64 from typing import (
65 65 Iterator,
66 66 List,
67 67 Optional,
68 68 Tuple,
69 69 )
70 70
71 71
72 72 base85 = policy.importmod('base85')
73 73 osutil = policy.importmod('osutil')
74 74
75 75 b85decode = base85.b85decode
76 76 b85encode = base85.b85encode
77 77
78 78 cookielib = pycompat.cookielib
79 79 httplib = pycompat.httplib
80 80 pickle = pycompat.pickle
81 81 safehasattr = pycompat.safehasattr
82 82 socketserver = pycompat.socketserver
83 83 bytesio = pycompat.bytesio
84 84 # TODO deprecate stringio name, as it is a lie on Python 3.
85 85 stringio = bytesio
86 86 xmlrpclib = pycompat.xmlrpclib
87 87
88 88 httpserver = urllibcompat.httpserver
89 89 urlerr = urllibcompat.urlerr
90 90 urlreq = urllibcompat.urlreq
91 91
92 92 # workaround for win32mbcs
93 93 _filenamebytestr = pycompat.bytestr
94 94
95 95 if pycompat.iswindows:
96 96 from . import windows as platform
97 97 else:
98 98 from . import posix as platform
99 99
100 100 _ = i18n._
101 101
102 102 bindunixsocket = platform.bindunixsocket
103 103 cachestat = platform.cachestat
104 104 checkexec = platform.checkexec
105 105 checklink = platform.checklink
106 106 copymode = platform.copymode
107 107 expandglobs = platform.expandglobs
108 108 getfsmountpoint = platform.getfsmountpoint
109 109 getfstype = platform.getfstype
110 110 get_password = platform.get_password
111 111 groupmembers = platform.groupmembers
112 112 groupname = platform.groupname
113 113 isexec = platform.isexec
114 114 isowner = platform.isowner
115 115 listdir = osutil.listdir
116 116 localpath = platform.localpath
117 117 lookupreg = platform.lookupreg
118 118 makedir = platform.makedir
119 119 nlinks = platform.nlinks
120 120 normpath = platform.normpath
121 121 normcase = platform.normcase
122 122 normcasespec = platform.normcasespec
123 123 normcasefallback = platform.normcasefallback
124 124 openhardlinks = platform.openhardlinks
125 125 oslink = platform.oslink
126 126 parsepatchoutput = platform.parsepatchoutput
127 127 pconvert = platform.pconvert
128 128 poll = platform.poll
129 129 posixfile = platform.posixfile
130 130 readlink = platform.readlink
131 131 rename = platform.rename
132 132 removedirs = platform.removedirs
133 133 samedevice = platform.samedevice
134 134 samefile = platform.samefile
135 135 samestat = platform.samestat
136 136 setflags = platform.setflags
137 137 split = platform.split
138 138 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
139 139 statisexec = platform.statisexec
140 140 statislink = platform.statislink
141 141 umask = platform.umask
142 142 unlink = platform.unlink
143 143 username = platform.username
144 144
145 145
146 146 def setumask(val):
147 147 # type: (int) -> None
148 148 '''updates the umask. used by chg server'''
149 149 if pycompat.iswindows:
150 150 return
151 151 os.umask(val)
152 152 global umask
153 153 platform.umask = umask = val & 0o777
154 154
155 155
156 156 # small compat layer
157 157 compengines = compression.compengines
158 158 SERVERROLE = compression.SERVERROLE
159 159 CLIENTROLE = compression.CLIENTROLE
160 160
161 161 try:
162 162 recvfds = osutil.recvfds
163 163 except AttributeError:
164 164 pass
165 165
166 166 # Python compatibility
167 167
168 168 _notset = object()
169 169
170 170
171 171 def bitsfrom(container):
172 172 bits = 0
173 173 for bit in container:
174 174 bits |= bit
175 175 return bits
176 176
177 177
178 178 # python 2.6 still have deprecation warning enabled by default. We do not want
179 179 # to display anything to standard user so detect if we are running test and
180 180 # only use python deprecation warning in this case.
181 181 _dowarn = bool(encoding.environ.get(b'HGEMITWARNINGS'))
182 182 if _dowarn:
183 183 # explicitly unfilter our warning for python 2.7
184 184 #
185 185 # The option of setting PYTHONWARNINGS in the test runner was investigated.
186 186 # However, module name set through PYTHONWARNINGS was exactly matched, so
187 187 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
188 188 # makes the whole PYTHONWARNINGS thing useless for our usecase.
189 189 warnings.filterwarnings('default', '', DeprecationWarning, 'mercurial')
190 190 warnings.filterwarnings('default', '', DeprecationWarning, 'hgext')
191 191 warnings.filterwarnings('default', '', DeprecationWarning, 'hgext3rd')
192 192 if _dowarn and pycompat.ispy3:
193 193 # silence warning emitted by passing user string to re.sub()
194 194 warnings.filterwarnings(
195 195 'ignore', 'bad escape', DeprecationWarning, 'mercurial'
196 196 )
197 197 warnings.filterwarnings(
198 198 'ignore', 'invalid escape sequence', DeprecationWarning, 'mercurial'
199 199 )
200 200 # TODO: reinvent imp.is_frozen()
201 201 warnings.filterwarnings(
202 202 'ignore',
203 203 'the imp module is deprecated',
204 204 DeprecationWarning,
205 205 'mercurial',
206 206 )
207 207
208 208
209 209 def nouideprecwarn(msg, version, stacklevel=1):
210 210 """Issue an python native deprecation warning
211 211
212 212 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
213 213 """
214 214 if _dowarn:
215 215 msg += (
216 216 b"\n(compatibility will be dropped after Mercurial-%s,"
217 217 b" update your code.)"
218 218 ) % version
219 219 warnings.warn(pycompat.sysstr(msg), DeprecationWarning, stacklevel + 1)
220 220 # on python 3 with chg, we will need to explicitly flush the output
221 221 sys.stderr.flush()
222 222
223 223
224 224 DIGESTS = {
225 225 b'md5': hashlib.md5,
226 226 b'sha1': hashutil.sha1,
227 227 b'sha512': hashlib.sha512,
228 228 }
229 229 # List of digest types from strongest to weakest
230 230 DIGESTS_BY_STRENGTH = [b'sha512', b'sha1', b'md5']
231 231
232 232 for k in DIGESTS_BY_STRENGTH:
233 233 assert k in DIGESTS
234 234
235 235
236 236 class digester(object):
237 237 """helper to compute digests.
238 238
239 239 This helper can be used to compute one or more digests given their name.
240 240
241 241 >>> d = digester([b'md5', b'sha1'])
242 242 >>> d.update(b'foo')
243 243 >>> [k for k in sorted(d)]
244 244 ['md5', 'sha1']
245 245 >>> d[b'md5']
246 246 'acbd18db4cc2f85cedef654fccc4a4d8'
247 247 >>> d[b'sha1']
248 248 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
249 249 >>> digester.preferred([b'md5', b'sha1'])
250 250 'sha1'
251 251 """
252 252
253 253 def __init__(self, digests, s=b''):
254 254 self._hashes = {}
255 255 for k in digests:
256 256 if k not in DIGESTS:
257 257 raise error.Abort(_(b'unknown digest type: %s') % k)
258 258 self._hashes[k] = DIGESTS[k]()
259 259 if s:
260 260 self.update(s)
261 261
262 262 def update(self, data):
263 263 for h in self._hashes.values():
264 264 h.update(data)
265 265
266 266 def __getitem__(self, key):
267 267 if key not in DIGESTS:
268 268 raise error.Abort(_(b'unknown digest type: %s') % k)
269 269 return hex(self._hashes[key].digest())
270 270
271 271 def __iter__(self):
272 272 return iter(self._hashes)
273 273
274 274 @staticmethod
275 275 def preferred(supported):
276 276 """returns the strongest digest type in both supported and DIGESTS."""
277 277
278 278 for k in DIGESTS_BY_STRENGTH:
279 279 if k in supported:
280 280 return k
281 281 return None
282 282
283 283
284 284 class digestchecker(object):
285 285 """file handle wrapper that additionally checks content against a given
286 286 size and digests.
287 287
288 288 d = digestchecker(fh, size, {'md5': '...'})
289 289
290 290 When multiple digests are given, all of them are validated.
291 291 """
292 292
293 293 def __init__(self, fh, size, digests):
294 294 self._fh = fh
295 295 self._size = size
296 296 self._got = 0
297 297 self._digests = dict(digests)
298 298 self._digester = digester(self._digests.keys())
299 299
300 300 def read(self, length=-1):
301 301 content = self._fh.read(length)
302 302 self._digester.update(content)
303 303 self._got += len(content)
304 304 return content
305 305
306 306 def validate(self):
307 307 if self._size != self._got:
308 308 raise error.Abort(
309 309 _(b'size mismatch: expected %d, got %d')
310 310 % (self._size, self._got)
311 311 )
312 312 for k, v in self._digests.items():
313 313 if v != self._digester[k]:
314 314 # i18n: first parameter is a digest name
315 315 raise error.Abort(
316 316 _(b'%s mismatch: expected %s, got %s')
317 317 % (k, v, self._digester[k])
318 318 )
319 319
320 320
321 321 try:
322 322 buffer = buffer # pytype: disable=name-error
323 323 except NameError:
324 324
325 325 def buffer(sliceable, offset=0, length=None):
326 326 if length is not None:
327 327 return memoryview(sliceable)[offset : offset + length]
328 328 return memoryview(sliceable)[offset:]
329 329
330 330
331 331 _chunksize = 4096
332 332
333 333
334 334 class bufferedinputpipe(object):
335 335 """a manually buffered input pipe
336 336
337 337 Python will not let us use buffered IO and lazy reading with 'polling' at
338 338 the same time. We cannot probe the buffer state and select will not detect
339 339 that data are ready to read if they are already buffered.
340 340
341 341 This class let us work around that by implementing its own buffering
342 342 (allowing efficient readline) while offering a way to know if the buffer is
343 343 empty from the output (allowing collaboration of the buffer with polling).
344 344
345 345 This class lives in the 'util' module because it makes use of the 'os'
346 346 module from the python stdlib.
347 347 """
348 348
349 349 def __new__(cls, fh):
350 350 # If we receive a fileobjectproxy, we need to use a variation of this
351 351 # class that notifies observers about activity.
352 352 if isinstance(fh, fileobjectproxy):
353 353 cls = observedbufferedinputpipe
354 354
355 355 return super(bufferedinputpipe, cls).__new__(cls)
356 356
357 357 def __init__(self, input):
358 358 self._input = input
359 359 self._buffer = []
360 360 self._eof = False
361 361 self._lenbuf = 0
362 362
363 363 @property
364 364 def hasbuffer(self):
365 365 """True is any data is currently buffered
366 366
367 367 This will be used externally a pre-step for polling IO. If there is
368 368 already data then no polling should be set in place."""
369 369 return bool(self._buffer)
370 370
371 371 @property
372 372 def closed(self):
373 373 return self._input.closed
374 374
375 375 def fileno(self):
376 376 return self._input.fileno()
377 377
378 378 def close(self):
379 379 return self._input.close()
380 380
381 381 def read(self, size):
382 382 while (not self._eof) and (self._lenbuf < size):
383 383 self._fillbuffer()
384 384 return self._frombuffer(size)
385 385
386 386 def unbufferedread(self, size):
387 387 if not self._eof and self._lenbuf == 0:
388 388 self._fillbuffer(max(size, _chunksize))
389 389 return self._frombuffer(min(self._lenbuf, size))
390 390
391 391 def readline(self, *args, **kwargs):
392 392 if len(self._buffer) > 1:
393 393 # this should not happen because both read and readline end with a
394 394 # _frombuffer call that collapse it.
395 395 self._buffer = [b''.join(self._buffer)]
396 396 self._lenbuf = len(self._buffer[0])
397 397 lfi = -1
398 398 if self._buffer:
399 399 lfi = self._buffer[-1].find(b'\n')
400 400 while (not self._eof) and lfi < 0:
401 401 self._fillbuffer()
402 402 if self._buffer:
403 403 lfi = self._buffer[-1].find(b'\n')
404 404 size = lfi + 1
405 405 if lfi < 0: # end of file
406 406 size = self._lenbuf
407 407 elif len(self._buffer) > 1:
408 408 # we need to take previous chunks into account
409 409 size += self._lenbuf - len(self._buffer[-1])
410 410 return self._frombuffer(size)
411 411
412 412 def _frombuffer(self, size):
413 413 """return at most 'size' data from the buffer
414 414
415 415 The data are removed from the buffer."""
416 416 if size == 0 or not self._buffer:
417 417 return b''
418 418 buf = self._buffer[0]
419 419 if len(self._buffer) > 1:
420 420 buf = b''.join(self._buffer)
421 421
422 422 data = buf[:size]
423 423 buf = buf[len(data) :]
424 424 if buf:
425 425 self._buffer = [buf]
426 426 self._lenbuf = len(buf)
427 427 else:
428 428 self._buffer = []
429 429 self._lenbuf = 0
430 430 return data
431 431
432 432 def _fillbuffer(self, size=_chunksize):
433 433 """read data to the buffer"""
434 434 data = os.read(self._input.fileno(), size)
435 435 if not data:
436 436 self._eof = True
437 437 else:
438 438 self._lenbuf += len(data)
439 439 self._buffer.append(data)
440 440
441 441 return data
442 442
443 443
444 444 def mmapread(fp, size=None):
445 445 if size == 0:
446 446 # size of 0 to mmap.mmap() means "all data"
447 447 # rather than "zero bytes", so special case that.
448 448 return b''
449 449 elif size is None:
450 450 size = 0
451 451 try:
452 452 fd = getattr(fp, 'fileno', lambda: fp)()
453 453 return mmap.mmap(fd, size, access=mmap.ACCESS_READ)
454 454 except ValueError:
455 455 # Empty files cannot be mmapped, but mmapread should still work. Check
456 456 # if the file is empty, and if so, return an empty buffer.
457 457 if os.fstat(fd).st_size == 0:
458 458 return b''
459 459 raise
460 460
461 461
462 462 class fileobjectproxy(object):
463 463 """A proxy around file objects that tells a watcher when events occur.
464 464
465 465 This type is intended to only be used for testing purposes. Think hard
466 466 before using it in important code.
467 467 """
468 468
469 469 __slots__ = (
470 470 '_orig',
471 471 '_observer',
472 472 )
473 473
474 474 def __init__(self, fh, observer):
475 475 object.__setattr__(self, '_orig', fh)
476 476 object.__setattr__(self, '_observer', observer)
477 477
478 478 def __getattribute__(self, name):
479 479 ours = {
480 480 '_observer',
481 481 # IOBase
482 482 'close',
483 483 # closed if a property
484 484 'fileno',
485 485 'flush',
486 486 'isatty',
487 487 'readable',
488 488 'readline',
489 489 'readlines',
490 490 'seek',
491 491 'seekable',
492 492 'tell',
493 493 'truncate',
494 494 'writable',
495 495 'writelines',
496 496 # RawIOBase
497 497 'read',
498 498 'readall',
499 499 'readinto',
500 500 'write',
501 501 # BufferedIOBase
502 502 # raw is a property
503 503 'detach',
504 504 # read defined above
505 505 'read1',
506 506 # readinto defined above
507 507 # write defined above
508 508 }
509 509
510 510 # We only observe some methods.
511 511 if name in ours:
512 512 return object.__getattribute__(self, name)
513 513
514 514 return getattr(object.__getattribute__(self, '_orig'), name)
515 515
516 516 def __nonzero__(self):
517 517 return bool(object.__getattribute__(self, '_orig'))
518 518
519 519 __bool__ = __nonzero__
520 520
521 521 def __delattr__(self, name):
522 522 return delattr(object.__getattribute__(self, '_orig'), name)
523 523
524 524 def __setattr__(self, name, value):
525 525 return setattr(object.__getattribute__(self, '_orig'), name, value)
526 526
527 527 def __iter__(self):
528 528 return object.__getattribute__(self, '_orig').__iter__()
529 529
530 530 def _observedcall(self, name, *args, **kwargs):
531 531 # Call the original object.
532 532 orig = object.__getattribute__(self, '_orig')
533 533 res = getattr(orig, name)(*args, **kwargs)
534 534
535 535 # Call a method on the observer of the same name with arguments
536 536 # so it can react, log, etc.
537 537 observer = object.__getattribute__(self, '_observer')
538 538 fn = getattr(observer, name, None)
539 539 if fn:
540 540 fn(res, *args, **kwargs)
541 541
542 542 return res
543 543
544 544 def close(self, *args, **kwargs):
545 545 return object.__getattribute__(self, '_observedcall')(
546 546 'close', *args, **kwargs
547 547 )
548 548
549 549 def fileno(self, *args, **kwargs):
550 550 return object.__getattribute__(self, '_observedcall')(
551 551 'fileno', *args, **kwargs
552 552 )
553 553
554 554 def flush(self, *args, **kwargs):
555 555 return object.__getattribute__(self, '_observedcall')(
556 556 'flush', *args, **kwargs
557 557 )
558 558
559 559 def isatty(self, *args, **kwargs):
560 560 return object.__getattribute__(self, '_observedcall')(
561 561 'isatty', *args, **kwargs
562 562 )
563 563
564 564 def readable(self, *args, **kwargs):
565 565 return object.__getattribute__(self, '_observedcall')(
566 566 'readable', *args, **kwargs
567 567 )
568 568
569 569 def readline(self, *args, **kwargs):
570 570 return object.__getattribute__(self, '_observedcall')(
571 571 'readline', *args, **kwargs
572 572 )
573 573
574 574 def readlines(self, *args, **kwargs):
575 575 return object.__getattribute__(self, '_observedcall')(
576 576 'readlines', *args, **kwargs
577 577 )
578 578
579 579 def seek(self, *args, **kwargs):
580 580 return object.__getattribute__(self, '_observedcall')(
581 581 'seek', *args, **kwargs
582 582 )
583 583
584 584 def seekable(self, *args, **kwargs):
585 585 return object.__getattribute__(self, '_observedcall')(
586 586 'seekable', *args, **kwargs
587 587 )
588 588
589 589 def tell(self, *args, **kwargs):
590 590 return object.__getattribute__(self, '_observedcall')(
591 591 'tell', *args, **kwargs
592 592 )
593 593
594 594 def truncate(self, *args, **kwargs):
595 595 return object.__getattribute__(self, '_observedcall')(
596 596 'truncate', *args, **kwargs
597 597 )
598 598
599 599 def writable(self, *args, **kwargs):
600 600 return object.__getattribute__(self, '_observedcall')(
601 601 'writable', *args, **kwargs
602 602 )
603 603
604 604 def writelines(self, *args, **kwargs):
605 605 return object.__getattribute__(self, '_observedcall')(
606 606 'writelines', *args, **kwargs
607 607 )
608 608
609 609 def read(self, *args, **kwargs):
610 610 return object.__getattribute__(self, '_observedcall')(
611 611 'read', *args, **kwargs
612 612 )
613 613
614 614 def readall(self, *args, **kwargs):
615 615 return object.__getattribute__(self, '_observedcall')(
616 616 'readall', *args, **kwargs
617 617 )
618 618
619 619 def readinto(self, *args, **kwargs):
620 620 return object.__getattribute__(self, '_observedcall')(
621 621 'readinto', *args, **kwargs
622 622 )
623 623
624 624 def write(self, *args, **kwargs):
625 625 return object.__getattribute__(self, '_observedcall')(
626 626 'write', *args, **kwargs
627 627 )
628 628
629 629 def detach(self, *args, **kwargs):
630 630 return object.__getattribute__(self, '_observedcall')(
631 631 'detach', *args, **kwargs
632 632 )
633 633
634 634 def read1(self, *args, **kwargs):
635 635 return object.__getattribute__(self, '_observedcall')(
636 636 'read1', *args, **kwargs
637 637 )
638 638
639 639
640 640 class observedbufferedinputpipe(bufferedinputpipe):
641 641 """A variation of bufferedinputpipe that is aware of fileobjectproxy.
642 642
643 643 ``bufferedinputpipe`` makes low-level calls to ``os.read()`` that
644 644 bypass ``fileobjectproxy``. Because of this, we need to make
645 645 ``bufferedinputpipe`` aware of these operations.
646 646
647 647 This variation of ``bufferedinputpipe`` can notify observers about
648 648 ``os.read()`` events. It also re-publishes other events, such as
649 649 ``read()`` and ``readline()``.
650 650 """
651 651
652 652 def _fillbuffer(self):
653 653 res = super(observedbufferedinputpipe, self)._fillbuffer()
654 654
655 655 fn = getattr(self._input._observer, 'osread', None)
656 656 if fn:
657 657 fn(res, _chunksize)
658 658
659 659 return res
660 660
661 661 # We use different observer methods because the operation isn't
662 662 # performed on the actual file object but on us.
663 663 def read(self, size):
664 664 res = super(observedbufferedinputpipe, self).read(size)
665 665
666 666 fn = getattr(self._input._observer, 'bufferedread', None)
667 667 if fn:
668 668 fn(res, size)
669 669
670 670 return res
671 671
672 672 def readline(self, *args, **kwargs):
673 673 res = super(observedbufferedinputpipe, self).readline(*args, **kwargs)
674 674
675 675 fn = getattr(self._input._observer, 'bufferedreadline', None)
676 676 if fn:
677 677 fn(res)
678 678
679 679 return res
680 680
681 681
682 682 PROXIED_SOCKET_METHODS = {
683 683 'makefile',
684 684 'recv',
685 685 'recvfrom',
686 686 'recvfrom_into',
687 687 'recv_into',
688 688 'send',
689 689 'sendall',
690 690 'sendto',
691 691 'setblocking',
692 692 'settimeout',
693 693 'gettimeout',
694 694 'setsockopt',
695 695 }
696 696
697 697
698 698 class socketproxy(object):
699 699 """A proxy around a socket that tells a watcher when events occur.
700 700
701 701 This is like ``fileobjectproxy`` except for sockets.
702 702
703 703 This type is intended to only be used for testing purposes. Think hard
704 704 before using it in important code.
705 705 """
706 706
707 707 __slots__ = (
708 708 '_orig',
709 709 '_observer',
710 710 )
711 711
712 712 def __init__(self, sock, observer):
713 713 object.__setattr__(self, '_orig', sock)
714 714 object.__setattr__(self, '_observer', observer)
715 715
716 716 def __getattribute__(self, name):
717 717 if name in PROXIED_SOCKET_METHODS:
718 718 return object.__getattribute__(self, name)
719 719
720 720 return getattr(object.__getattribute__(self, '_orig'), name)
721 721
722 722 def __delattr__(self, name):
723 723 return delattr(object.__getattribute__(self, '_orig'), name)
724 724
725 725 def __setattr__(self, name, value):
726 726 return setattr(object.__getattribute__(self, '_orig'), name, value)
727 727
728 728 def __nonzero__(self):
729 729 return bool(object.__getattribute__(self, '_orig'))
730 730
731 731 __bool__ = __nonzero__
732 732
733 733 def _observedcall(self, name, *args, **kwargs):
734 734 # Call the original object.
735 735 orig = object.__getattribute__(self, '_orig')
736 736 res = getattr(orig, name)(*args, **kwargs)
737 737
738 738 # Call a method on the observer of the same name with arguments
739 739 # so it can react, log, etc.
740 740 observer = object.__getattribute__(self, '_observer')
741 741 fn = getattr(observer, name, None)
742 742 if fn:
743 743 fn(res, *args, **kwargs)
744 744
745 745 return res
746 746
747 747 def makefile(self, *args, **kwargs):
748 748 res = object.__getattribute__(self, '_observedcall')(
749 749 'makefile', *args, **kwargs
750 750 )
751 751
752 752 # The file object may be used for I/O. So we turn it into a
753 753 # proxy using our observer.
754 754 observer = object.__getattribute__(self, '_observer')
755 755 return makeloggingfileobject(
756 756 observer.fh,
757 757 res,
758 758 observer.name,
759 759 reads=observer.reads,
760 760 writes=observer.writes,
761 761 logdata=observer.logdata,
762 762 logdataapis=observer.logdataapis,
763 763 )
764 764
765 765 def recv(self, *args, **kwargs):
766 766 return object.__getattribute__(self, '_observedcall')(
767 767 'recv', *args, **kwargs
768 768 )
769 769
770 770 def recvfrom(self, *args, **kwargs):
771 771 return object.__getattribute__(self, '_observedcall')(
772 772 'recvfrom', *args, **kwargs
773 773 )
774 774
775 775 def recvfrom_into(self, *args, **kwargs):
776 776 return object.__getattribute__(self, '_observedcall')(
777 777 'recvfrom_into', *args, **kwargs
778 778 )
779 779
780 780 def recv_into(self, *args, **kwargs):
781 781 return object.__getattribute__(self, '_observedcall')(
782 782 'recv_info', *args, **kwargs
783 783 )
784 784
785 785 def send(self, *args, **kwargs):
786 786 return object.__getattribute__(self, '_observedcall')(
787 787 'send', *args, **kwargs
788 788 )
789 789
790 790 def sendall(self, *args, **kwargs):
791 791 return object.__getattribute__(self, '_observedcall')(
792 792 'sendall', *args, **kwargs
793 793 )
794 794
795 795 def sendto(self, *args, **kwargs):
796 796 return object.__getattribute__(self, '_observedcall')(
797 797 'sendto', *args, **kwargs
798 798 )
799 799
800 800 def setblocking(self, *args, **kwargs):
801 801 return object.__getattribute__(self, '_observedcall')(
802 802 'setblocking', *args, **kwargs
803 803 )
804 804
805 805 def settimeout(self, *args, **kwargs):
806 806 return object.__getattribute__(self, '_observedcall')(
807 807 'settimeout', *args, **kwargs
808 808 )
809 809
810 810 def gettimeout(self, *args, **kwargs):
811 811 return object.__getattribute__(self, '_observedcall')(
812 812 'gettimeout', *args, **kwargs
813 813 )
814 814
815 815 def setsockopt(self, *args, **kwargs):
816 816 return object.__getattribute__(self, '_observedcall')(
817 817 'setsockopt', *args, **kwargs
818 818 )
819 819
820 820
821 821 class baseproxyobserver(object):
822 822 def __init__(self, fh, name, logdata, logdataapis):
823 823 self.fh = fh
824 824 self.name = name
825 825 self.logdata = logdata
826 826 self.logdataapis = logdataapis
827 827
828 828 def _writedata(self, data):
829 829 if not self.logdata:
830 830 if self.logdataapis:
831 831 self.fh.write(b'\n')
832 832 self.fh.flush()
833 833 return
834 834
835 835 # Simple case writes all data on a single line.
836 836 if b'\n' not in data:
837 837 if self.logdataapis:
838 838 self.fh.write(b': %s\n' % stringutil.escapestr(data))
839 839 else:
840 840 self.fh.write(
841 841 b'%s> %s\n' % (self.name, stringutil.escapestr(data))
842 842 )
843 843 self.fh.flush()
844 844 return
845 845
846 846 # Data with newlines is written to multiple lines.
847 847 if self.logdataapis:
848 848 self.fh.write(b':\n')
849 849
850 850 lines = data.splitlines(True)
851 851 for line in lines:
852 852 self.fh.write(
853 853 b'%s> %s\n' % (self.name, stringutil.escapestr(line))
854 854 )
855 855 self.fh.flush()
856 856
857 857
858 858 class fileobjectobserver(baseproxyobserver):
859 859 """Logs file object activity."""
860 860
861 861 def __init__(
862 862 self, fh, name, reads=True, writes=True, logdata=False, logdataapis=True
863 863 ):
864 864 super(fileobjectobserver, self).__init__(fh, name, logdata, logdataapis)
865 865 self.reads = reads
866 866 self.writes = writes
867 867
868 868 def read(self, res, size=-1):
869 869 if not self.reads:
870 870 return
871 871 # Python 3 can return None from reads at EOF instead of empty strings.
872 872 if res is None:
873 873 res = b''
874 874
875 875 if size == -1 and res == b'':
876 876 # Suppress pointless read(-1) calls that return
877 877 # nothing. These happen _a lot_ on Python 3, and there
878 878 # doesn't seem to be a better workaround to have matching
879 879 # Python 2 and 3 behavior. :(
880 880 return
881 881
882 882 if self.logdataapis:
883 883 self.fh.write(b'%s> read(%d) -> %d' % (self.name, size, len(res)))
884 884
885 885 self._writedata(res)
886 886
887 887 def readline(self, res, limit=-1):
888 888 if not self.reads:
889 889 return
890 890
891 891 if self.logdataapis:
892 892 self.fh.write(b'%s> readline() -> %d' % (self.name, len(res)))
893 893
894 894 self._writedata(res)
895 895
896 896 def readinto(self, res, dest):
897 897 if not self.reads:
898 898 return
899 899
900 900 if self.logdataapis:
901 901 self.fh.write(
902 902 b'%s> readinto(%d) -> %r' % (self.name, len(dest), res)
903 903 )
904 904
905 905 data = dest[0:res] if res is not None else b''
906 906
907 907 # _writedata() uses "in" operator and is confused by memoryview because
908 908 # characters are ints on Python 3.
909 909 if isinstance(data, memoryview):
910 910 data = data.tobytes()
911 911
912 912 self._writedata(data)
913 913
914 914 def write(self, res, data):
915 915 if not self.writes:
916 916 return
917 917
918 918 # Python 2 returns None from some write() calls. Python 3 (reasonably)
919 919 # returns the integer bytes written.
920 920 if res is None and data:
921 921 res = len(data)
922 922
923 923 if self.logdataapis:
924 924 self.fh.write(b'%s> write(%d) -> %r' % (self.name, len(data), res))
925 925
926 926 self._writedata(data)
927 927
928 928 def flush(self, res):
929 929 if not self.writes:
930 930 return
931 931
932 932 self.fh.write(b'%s> flush() -> %r\n' % (self.name, res))
933 933
934 934 # For observedbufferedinputpipe.
935 935 def bufferedread(self, res, size):
936 936 if not self.reads:
937 937 return
938 938
939 939 if self.logdataapis:
940 940 self.fh.write(
941 941 b'%s> bufferedread(%d) -> %d' % (self.name, size, len(res))
942 942 )
943 943
944 944 self._writedata(res)
945 945
946 946 def bufferedreadline(self, res):
947 947 if not self.reads:
948 948 return
949 949
950 950 if self.logdataapis:
951 951 self.fh.write(
952 952 b'%s> bufferedreadline() -> %d' % (self.name, len(res))
953 953 )
954 954
955 955 self._writedata(res)
956 956
957 957
958 958 def makeloggingfileobject(
959 959 logh, fh, name, reads=True, writes=True, logdata=False, logdataapis=True
960 960 ):
961 961 """Turn a file object into a logging file object."""
962 962
963 963 observer = fileobjectobserver(
964 964 logh,
965 965 name,
966 966 reads=reads,
967 967 writes=writes,
968 968 logdata=logdata,
969 969 logdataapis=logdataapis,
970 970 )
971 971 return fileobjectproxy(fh, observer)
972 972
973 973
974 974 class socketobserver(baseproxyobserver):
975 975 """Logs socket activity."""
976 976
977 977 def __init__(
978 978 self,
979 979 fh,
980 980 name,
981 981 reads=True,
982 982 writes=True,
983 983 states=True,
984 984 logdata=False,
985 985 logdataapis=True,
986 986 ):
987 987 super(socketobserver, self).__init__(fh, name, logdata, logdataapis)
988 988 self.reads = reads
989 989 self.writes = writes
990 990 self.states = states
991 991
992 992 def makefile(self, res, mode=None, bufsize=None):
993 993 if not self.states:
994 994 return
995 995
996 996 self.fh.write(b'%s> makefile(%r, %r)\n' % (self.name, mode, bufsize))
997 997
998 998 def recv(self, res, size, flags=0):
999 999 if not self.reads:
1000 1000 return
1001 1001
1002 1002 if self.logdataapis:
1003 1003 self.fh.write(
1004 1004 b'%s> recv(%d, %d) -> %d' % (self.name, size, flags, len(res))
1005 1005 )
1006 1006 self._writedata(res)
1007 1007
1008 1008 def recvfrom(self, res, size, flags=0):
1009 1009 if not self.reads:
1010 1010 return
1011 1011
1012 1012 if self.logdataapis:
1013 1013 self.fh.write(
1014 1014 b'%s> recvfrom(%d, %d) -> %d'
1015 1015 % (self.name, size, flags, len(res[0]))
1016 1016 )
1017 1017
1018 1018 self._writedata(res[0])
1019 1019
1020 1020 def recvfrom_into(self, res, buf, size, flags=0):
1021 1021 if not self.reads:
1022 1022 return
1023 1023
1024 1024 if self.logdataapis:
1025 1025 self.fh.write(
1026 1026 b'%s> recvfrom_into(%d, %d) -> %d'
1027 1027 % (self.name, size, flags, res[0])
1028 1028 )
1029 1029
1030 1030 self._writedata(buf[0 : res[0]])
1031 1031
1032 1032 def recv_into(self, res, buf, size=0, flags=0):
1033 1033 if not self.reads:
1034 1034 return
1035 1035
1036 1036 if self.logdataapis:
1037 1037 self.fh.write(
1038 1038 b'%s> recv_into(%d, %d) -> %d' % (self.name, size, flags, res)
1039 1039 )
1040 1040
1041 1041 self._writedata(buf[0:res])
1042 1042
1043 1043 def send(self, res, data, flags=0):
1044 1044 if not self.writes:
1045 1045 return
1046 1046
1047 1047 self.fh.write(
1048 1048 b'%s> send(%d, %d) -> %d' % (self.name, len(data), flags, len(res))
1049 1049 )
1050 1050 self._writedata(data)
1051 1051
1052 1052 def sendall(self, res, data, flags=0):
1053 1053 if not self.writes:
1054 1054 return
1055 1055
1056 1056 if self.logdataapis:
1057 1057 # Returns None on success. So don't bother reporting return value.
1058 1058 self.fh.write(
1059 1059 b'%s> sendall(%d, %d)' % (self.name, len(data), flags)
1060 1060 )
1061 1061
1062 1062 self._writedata(data)
1063 1063
1064 1064 def sendto(self, res, data, flagsoraddress, address=None):
1065 1065 if not self.writes:
1066 1066 return
1067 1067
1068 1068 if address:
1069 1069 flags = flagsoraddress
1070 1070 else:
1071 1071 flags = 0
1072 1072
1073 1073 if self.logdataapis:
1074 1074 self.fh.write(
1075 1075 b'%s> sendto(%d, %d, %r) -> %d'
1076 1076 % (self.name, len(data), flags, address, res)
1077 1077 )
1078 1078
1079 1079 self._writedata(data)
1080 1080
1081 1081 def setblocking(self, res, flag):
1082 1082 if not self.states:
1083 1083 return
1084 1084
1085 1085 self.fh.write(b'%s> setblocking(%r)\n' % (self.name, flag))
1086 1086
1087 1087 def settimeout(self, res, value):
1088 1088 if not self.states:
1089 1089 return
1090 1090
1091 1091 self.fh.write(b'%s> settimeout(%r)\n' % (self.name, value))
1092 1092
1093 1093 def gettimeout(self, res):
1094 1094 if not self.states:
1095 1095 return
1096 1096
1097 1097 self.fh.write(b'%s> gettimeout() -> %f\n' % (self.name, res))
1098 1098
1099 1099 def setsockopt(self, res, level, optname, value):
1100 1100 if not self.states:
1101 1101 return
1102 1102
1103 1103 self.fh.write(
1104 1104 b'%s> setsockopt(%r, %r, %r) -> %r\n'
1105 1105 % (self.name, level, optname, value, res)
1106 1106 )
1107 1107
1108 1108
1109 1109 def makeloggingsocket(
1110 1110 logh,
1111 1111 fh,
1112 1112 name,
1113 1113 reads=True,
1114 1114 writes=True,
1115 1115 states=True,
1116 1116 logdata=False,
1117 1117 logdataapis=True,
1118 1118 ):
1119 1119 """Turn a socket into a logging socket."""
1120 1120
1121 1121 observer = socketobserver(
1122 1122 logh,
1123 1123 name,
1124 1124 reads=reads,
1125 1125 writes=writes,
1126 1126 states=states,
1127 1127 logdata=logdata,
1128 1128 logdataapis=logdataapis,
1129 1129 )
1130 1130 return socketproxy(fh, observer)
1131 1131
1132 1132
1133 1133 def version():
1134 1134 """Return version information if available."""
1135 1135 try:
1136 1136 from . import __version__
1137 1137
1138 1138 return __version__.version
1139 1139 except ImportError:
1140 1140 return b'unknown'
1141 1141
1142 1142
1143 1143 def versiontuple(v=None, n=4):
1144 1144 """Parses a Mercurial version string into an N-tuple.
1145 1145
1146 1146 The version string to be parsed is specified with the ``v`` argument.
1147 1147 If it isn't defined, the current Mercurial version string will be parsed.
1148 1148
1149 1149 ``n`` can be 2, 3, or 4. Here is how some version strings map to
1150 1150 returned values:
1151 1151
1152 1152 >>> v = b'3.6.1+190-df9b73d2d444'
1153 1153 >>> versiontuple(v, 2)
1154 1154 (3, 6)
1155 1155 >>> versiontuple(v, 3)
1156 1156 (3, 6, 1)
1157 1157 >>> versiontuple(v, 4)
1158 1158 (3, 6, 1, '190-df9b73d2d444')
1159 1159
1160 1160 >>> versiontuple(b'3.6.1+190-df9b73d2d444+20151118')
1161 1161 (3, 6, 1, '190-df9b73d2d444+20151118')
1162 1162
1163 1163 >>> v = b'3.6'
1164 1164 >>> versiontuple(v, 2)
1165 1165 (3, 6)
1166 1166 >>> versiontuple(v, 3)
1167 1167 (3, 6, None)
1168 1168 >>> versiontuple(v, 4)
1169 1169 (3, 6, None, None)
1170 1170
1171 1171 >>> v = b'3.9-rc'
1172 1172 >>> versiontuple(v, 2)
1173 1173 (3, 9)
1174 1174 >>> versiontuple(v, 3)
1175 1175 (3, 9, None)
1176 1176 >>> versiontuple(v, 4)
1177 1177 (3, 9, None, 'rc')
1178 1178
1179 1179 >>> v = b'3.9-rc+2-02a8fea4289b'
1180 1180 >>> versiontuple(v, 2)
1181 1181 (3, 9)
1182 1182 >>> versiontuple(v, 3)
1183 1183 (3, 9, None)
1184 1184 >>> versiontuple(v, 4)
1185 1185 (3, 9, None, 'rc+2-02a8fea4289b')
1186 1186
1187 1187 >>> versiontuple(b'4.6rc0')
1188 1188 (4, 6, None, 'rc0')
1189 1189 >>> versiontuple(b'4.6rc0+12-425d55e54f98')
1190 1190 (4, 6, None, 'rc0+12-425d55e54f98')
1191 1191 >>> versiontuple(b'.1.2.3')
1192 1192 (None, None, None, '.1.2.3')
1193 1193 >>> versiontuple(b'12.34..5')
1194 1194 (12, 34, None, '..5')
1195 1195 >>> versiontuple(b'1.2.3.4.5.6')
1196 1196 (1, 2, 3, '.4.5.6')
1197 1197 """
1198 1198 if not v:
1199 1199 v = version()
1200 1200 m = remod.match(br'(\d+(?:\.\d+){,2})[+-]?(.*)', v)
1201 1201 if not m:
1202 1202 vparts, extra = b'', v
1203 1203 elif m.group(2):
1204 1204 vparts, extra = m.groups()
1205 1205 else:
1206 1206 vparts, extra = m.group(1), None
1207 1207
1208 1208 assert vparts is not None # help pytype
1209 1209
1210 1210 vints = []
1211 1211 for i in vparts.split(b'.'):
1212 1212 try:
1213 1213 vints.append(int(i))
1214 1214 except ValueError:
1215 1215 break
1216 1216 # (3, 6) -> (3, 6, None)
1217 1217 while len(vints) < 3:
1218 1218 vints.append(None)
1219 1219
1220 1220 if n == 2:
1221 1221 return (vints[0], vints[1])
1222 1222 if n == 3:
1223 1223 return (vints[0], vints[1], vints[2])
1224 1224 if n == 4:
1225 1225 return (vints[0], vints[1], vints[2], extra)
1226 1226
1227 1227
1228 1228 def cachefunc(func):
1229 1229 '''cache the result of function calls'''
1230 1230 # XXX doesn't handle keywords args
1231 1231 if func.__code__.co_argcount == 0:
1232 1232 listcache = []
1233 1233
1234 1234 def f():
1235 1235 if len(listcache) == 0:
1236 1236 listcache.append(func())
1237 1237 return listcache[0]
1238 1238
1239 1239 return f
1240 1240 cache = {}
1241 1241 if func.__code__.co_argcount == 1:
1242 1242 # we gain a small amount of time because
1243 1243 # we don't need to pack/unpack the list
1244 1244 def f(arg):
1245 1245 if arg not in cache:
1246 1246 cache[arg] = func(arg)
1247 1247 return cache[arg]
1248 1248
1249 1249 else:
1250 1250
1251 1251 def f(*args):
1252 1252 if args not in cache:
1253 1253 cache[args] = func(*args)
1254 1254 return cache[args]
1255 1255
1256 1256 return f
1257 1257
1258 1258
1259 1259 class cow(object):
1260 1260 """helper class to make copy-on-write easier
1261 1261
1262 1262 Call preparewrite before doing any writes.
1263 1263 """
1264 1264
1265 1265 def preparewrite(self):
1266 1266 """call this before writes, return self or a copied new object"""
1267 1267 if getattr(self, '_copied', 0):
1268 1268 self._copied -= 1
1269 1269 # Function cow.__init__ expects 1 arg(s), got 2 [wrong-arg-count]
1270 1270 return self.__class__(self) # pytype: disable=wrong-arg-count
1271 1271 return self
1272 1272
1273 1273 def copy(self):
1274 1274 """always do a cheap copy"""
1275 1275 self._copied = getattr(self, '_copied', 0) + 1
1276 1276 return self
1277 1277
1278 1278
1279 1279 class sortdict(collections.OrderedDict):
1280 1280 """a simple sorted dictionary
1281 1281
1282 1282 >>> d1 = sortdict([(b'a', 0), (b'b', 1)])
1283 1283 >>> d2 = d1.copy()
1284 1284 >>> d2
1285 1285 sortdict([('a', 0), ('b', 1)])
1286 1286 >>> d2.update([(b'a', 2)])
1287 1287 >>> list(d2.keys()) # should still be in last-set order
1288 1288 ['b', 'a']
1289 1289 >>> d1.insert(1, b'a.5', 0.5)
1290 1290 >>> d1
1291 1291 sortdict([('a', 0), ('a.5', 0.5), ('b', 1)])
1292 1292 """
1293 1293
1294 1294 def __setitem__(self, key, value):
1295 1295 if key in self:
1296 1296 del self[key]
1297 1297 super(sortdict, self).__setitem__(key, value)
1298 1298
1299 1299 if pycompat.ispypy:
1300 1300 # __setitem__() isn't called as of PyPy 5.8.0
1301 1301 def update(self, src, **f):
1302 1302 if isinstance(src, dict):
1303 1303 src = pycompat.iteritems(src)
1304 1304 for k, v in src:
1305 1305 self[k] = v
1306 1306 for k in f:
1307 1307 self[k] = f[k]
1308 1308
1309 1309 def insert(self, position, key, value):
1310 1310 for (i, (k, v)) in enumerate(list(self.items())):
1311 1311 if i == position:
1312 1312 self[key] = value
1313 1313 if i >= position:
1314 1314 del self[k]
1315 1315 self[k] = v
1316 1316
1317 1317
1318 1318 class cowdict(cow, dict):
1319 1319 """copy-on-write dict
1320 1320
1321 1321 Be sure to call d = d.preparewrite() before writing to d.
1322 1322
1323 1323 >>> a = cowdict()
1324 1324 >>> a is a.preparewrite()
1325 1325 True
1326 1326 >>> b = a.copy()
1327 1327 >>> b is a
1328 1328 True
1329 1329 >>> c = b.copy()
1330 1330 >>> c is a
1331 1331 True
1332 1332 >>> a = a.preparewrite()
1333 1333 >>> b is a
1334 1334 False
1335 1335 >>> a is a.preparewrite()
1336 1336 True
1337 1337 >>> c = c.preparewrite()
1338 1338 >>> b is c
1339 1339 False
1340 1340 >>> b is b.preparewrite()
1341 1341 True
1342 1342 """
1343 1343
1344 1344
1345 1345 class cowsortdict(cow, sortdict):
1346 1346 """copy-on-write sortdict
1347 1347
1348 1348 Be sure to call d = d.preparewrite() before writing to d.
1349 1349 """
1350 1350
1351 1351
1352 1352 class transactional(object): # pytype: disable=ignored-metaclass
1353 1353 """Base class for making a transactional type into a context manager."""
1354 1354
1355 1355 __metaclass__ = abc.ABCMeta
1356 1356
1357 1357 @abc.abstractmethod
1358 1358 def close(self):
1359 1359 """Successfully closes the transaction."""
1360 1360
1361 1361 @abc.abstractmethod
1362 1362 def release(self):
1363 1363 """Marks the end of the transaction.
1364 1364
1365 1365 If the transaction has not been closed, it will be aborted.
1366 1366 """
1367 1367
1368 1368 def __enter__(self):
1369 1369 return self
1370 1370
1371 1371 def __exit__(self, exc_type, exc_val, exc_tb):
1372 1372 try:
1373 1373 if exc_type is None:
1374 1374 self.close()
1375 1375 finally:
1376 1376 self.release()
1377 1377
1378 1378
1379 1379 @contextlib.contextmanager
1380 1380 def acceptintervention(tr=None):
1381 1381 """A context manager that closes the transaction on InterventionRequired
1382 1382
1383 1383 If no transaction was provided, this simply runs the body and returns
1384 1384 """
1385 1385 if not tr:
1386 1386 yield
1387 1387 return
1388 1388 try:
1389 1389 yield
1390 1390 tr.close()
1391 1391 except error.InterventionRequired:
1392 1392 tr.close()
1393 1393 raise
1394 1394 finally:
1395 1395 tr.release()
1396 1396
1397 1397
1398 1398 @contextlib.contextmanager
1399 1399 def nullcontextmanager(enter_result=None):
1400 1400 yield enter_result
1401 1401
1402 1402
1403 1403 class _lrucachenode(object):
1404 1404 """A node in a doubly linked list.
1405 1405
1406 1406 Holds a reference to nodes on either side as well as a key-value
1407 1407 pair for the dictionary entry.
1408 1408 """
1409 1409
1410 1410 __slots__ = ('next', 'prev', 'key', 'value', 'cost')
1411 1411
1412 1412 def __init__(self):
1413 1413 self.next = self
1414 1414 self.prev = self
1415 1415
1416 1416 self.key = _notset
1417 1417 self.value = None
1418 1418 self.cost = 0
1419 1419
1420 1420 def markempty(self):
1421 1421 """Mark the node as emptied."""
1422 1422 self.key = _notset
1423 1423 self.value = None
1424 1424 self.cost = 0
1425 1425
1426 1426
1427 1427 class lrucachedict(object):
1428 1428 """Dict that caches most recent accesses and sets.
1429 1429
1430 1430 The dict consists of an actual backing dict - indexed by original
1431 1431 key - and a doubly linked circular list defining the order of entries in
1432 1432 the cache.
1433 1433
1434 1434 The head node is the newest entry in the cache. If the cache is full,
1435 1435 we recycle head.prev and make it the new head. Cache accesses result in
1436 1436 the node being moved to before the existing head and being marked as the
1437 1437 new head node.
1438 1438
1439 1439 Items in the cache can be inserted with an optional "cost" value. This is
1440 1440 simply an integer that is specified by the caller. The cache can be queried
1441 1441 for the total cost of all items presently in the cache.
1442 1442
1443 1443 The cache can also define a maximum cost. If a cache insertion would
1444 1444 cause the total cost of the cache to go beyond the maximum cost limit,
1445 1445 nodes will be evicted to make room for the new code. This can be used
1446 1446 to e.g. set a max memory limit and associate an estimated bytes size
1447 1447 cost to each item in the cache. By default, no maximum cost is enforced.
1448 1448 """
1449 1449
1450 1450 def __init__(self, max, maxcost=0):
1451 1451 self._cache = {}
1452 1452
1453 1453 self._head = _lrucachenode()
1454 1454 self._size = 1
1455 1455 self.capacity = max
1456 1456 self.totalcost = 0
1457 1457 self.maxcost = maxcost
1458 1458
1459 1459 def __len__(self):
1460 1460 return len(self._cache)
1461 1461
1462 1462 def __contains__(self, k):
1463 1463 return k in self._cache
1464 1464
1465 1465 def __iter__(self):
1466 1466 # We don't have to iterate in cache order, but why not.
1467 1467 n = self._head
1468 1468 for i in range(len(self._cache)):
1469 1469 yield n.key
1470 1470 n = n.next
1471 1471
1472 1472 def __getitem__(self, k):
1473 1473 node = self._cache[k]
1474 1474 self._movetohead(node)
1475 1475 return node.value
1476 1476
1477 1477 def insert(self, k, v, cost=0):
1478 1478 """Insert a new item in the cache with optional cost value."""
1479 1479 node = self._cache.get(k)
1480 1480 # Replace existing value and mark as newest.
1481 1481 if node is not None:
1482 1482 self.totalcost -= node.cost
1483 1483 node.value = v
1484 1484 node.cost = cost
1485 1485 self.totalcost += cost
1486 1486 self._movetohead(node)
1487 1487
1488 1488 if self.maxcost:
1489 1489 self._enforcecostlimit()
1490 1490
1491 1491 return
1492 1492
1493 1493 if self._size < self.capacity:
1494 1494 node = self._addcapacity()
1495 1495 else:
1496 1496 # Grab the last/oldest item.
1497 1497 node = self._head.prev
1498 1498
1499 1499 # At capacity. Kill the old entry.
1500 1500 if node.key is not _notset:
1501 1501 self.totalcost -= node.cost
1502 1502 del self._cache[node.key]
1503 1503
1504 1504 node.key = k
1505 1505 node.value = v
1506 1506 node.cost = cost
1507 1507 self.totalcost += cost
1508 1508 self._cache[k] = node
1509 1509 # And mark it as newest entry. No need to adjust order since it
1510 1510 # is already self._head.prev.
1511 1511 self._head = node
1512 1512
1513 1513 if self.maxcost:
1514 1514 self._enforcecostlimit()
1515 1515
1516 1516 def __setitem__(self, k, v):
1517 1517 self.insert(k, v)
1518 1518
1519 1519 def __delitem__(self, k):
1520 1520 self.pop(k)
1521 1521
1522 1522 def pop(self, k, default=_notset):
1523 1523 try:
1524 1524 node = self._cache.pop(k)
1525 1525 except KeyError:
1526 1526 if default is _notset:
1527 1527 raise
1528 1528 return default
1529 1529
1530 1530 assert node is not None # help pytype
1531 1531 value = node.value
1532 1532 self.totalcost -= node.cost
1533 1533 node.markempty()
1534 1534
1535 1535 # Temporarily mark as newest item before re-adjusting head to make
1536 1536 # this node the oldest item.
1537 1537 self._movetohead(node)
1538 1538 self._head = node.next
1539 1539
1540 1540 return value
1541 1541
1542 1542 # Additional dict methods.
1543 1543
1544 1544 def get(self, k, default=None):
1545 1545 try:
1546 1546 return self.__getitem__(k)
1547 1547 except KeyError:
1548 1548 return default
1549 1549
1550 1550 def peek(self, k, default=_notset):
1551 1551 """Get the specified item without moving it to the head
1552 1552
1553 1553 Unlike get(), this doesn't mutate the internal state. But be aware
1554 1554 that it doesn't mean peek() is thread safe.
1555 1555 """
1556 1556 try:
1557 1557 node = self._cache[k]
1558 1558 assert node is not None # help pytype
1559 1559 return node.value
1560 1560 except KeyError:
1561 1561 if default is _notset:
1562 1562 raise
1563 1563 return default
1564 1564
1565 1565 def clear(self):
1566 1566 n = self._head
1567 1567 while n.key is not _notset:
1568 1568 self.totalcost -= n.cost
1569 1569 n.markempty()
1570 1570 n = n.next
1571 1571
1572 1572 self._cache.clear()
1573 1573
1574 1574 def copy(self, capacity=None, maxcost=0):
1575 1575 """Create a new cache as a copy of the current one.
1576 1576
1577 1577 By default, the new cache has the same capacity as the existing one.
1578 1578 But, the cache capacity can be changed as part of performing the
1579 1579 copy.
1580 1580
1581 1581 Items in the copy have an insertion/access order matching this
1582 1582 instance.
1583 1583 """
1584 1584
1585 1585 capacity = capacity or self.capacity
1586 1586 maxcost = maxcost or self.maxcost
1587 1587 result = lrucachedict(capacity, maxcost=maxcost)
1588 1588
1589 1589 # We copy entries by iterating in oldest-to-newest order so the copy
1590 1590 # has the correct ordering.
1591 1591
1592 1592 # Find the first non-empty entry.
1593 1593 n = self._head.prev
1594 1594 while n.key is _notset and n is not self._head:
1595 1595 n = n.prev
1596 1596
1597 1597 # We could potentially skip the first N items when decreasing capacity.
1598 1598 # But let's keep it simple unless it is a performance problem.
1599 1599 for i in range(len(self._cache)):
1600 1600 result.insert(n.key, n.value, cost=n.cost)
1601 1601 n = n.prev
1602 1602
1603 1603 return result
1604 1604
1605 1605 def popoldest(self):
1606 1606 """Remove the oldest item from the cache.
1607 1607
1608 1608 Returns the (key, value) describing the removed cache entry.
1609 1609 """
1610 1610 if not self._cache:
1611 1611 return
1612 1612
1613 1613 # Walk the linked list backwards starting at tail node until we hit
1614 1614 # a non-empty node.
1615 1615 n = self._head.prev
1616 1616
1617 1617 assert n is not None # help pytype
1618 1618
1619 1619 while n.key is _notset:
1620 1620 n = n.prev
1621 1621
1622 1622 assert n is not None # help pytype
1623 1623
1624 1624 key, value = n.key, n.value
1625 1625
1626 1626 # And remove it from the cache and mark it as empty.
1627 1627 del self._cache[n.key]
1628 1628 self.totalcost -= n.cost
1629 1629 n.markempty()
1630 1630
1631 1631 return key, value
1632 1632
1633 1633 def _movetohead(self, node):
1634 1634 """Mark a node as the newest, making it the new head.
1635 1635
1636 1636 When a node is accessed, it becomes the freshest entry in the LRU
1637 1637 list, which is denoted by self._head.
1638 1638
1639 1639 Visually, let's make ``N`` the new head node (* denotes head):
1640 1640
1641 1641 previous/oldest <-> head <-> next/next newest
1642 1642
1643 1643 ----<->--- A* ---<->-----
1644 1644 | |
1645 1645 E <-> D <-> N <-> C <-> B
1646 1646
1647 1647 To:
1648 1648
1649 1649 ----<->--- N* ---<->-----
1650 1650 | |
1651 1651 E <-> D <-> C <-> B <-> A
1652 1652
1653 1653 This requires the following moves:
1654 1654
1655 1655 C.next = D (node.prev.next = node.next)
1656 1656 D.prev = C (node.next.prev = node.prev)
1657 1657 E.next = N (head.prev.next = node)
1658 1658 N.prev = E (node.prev = head.prev)
1659 1659 N.next = A (node.next = head)
1660 1660 A.prev = N (head.prev = node)
1661 1661 """
1662 1662 head = self._head
1663 1663 # C.next = D
1664 1664 node.prev.next = node.next
1665 1665 # D.prev = C
1666 1666 node.next.prev = node.prev
1667 1667 # N.prev = E
1668 1668 node.prev = head.prev
1669 1669 # N.next = A
1670 1670 # It is tempting to do just "head" here, however if node is
1671 1671 # adjacent to head, this will do bad things.
1672 1672 node.next = head.prev.next
1673 1673 # E.next = N
1674 1674 node.next.prev = node
1675 1675 # A.prev = N
1676 1676 node.prev.next = node
1677 1677
1678 1678 self._head = node
1679 1679
1680 1680 def _addcapacity(self):
1681 1681 """Add a node to the circular linked list.
1682 1682
1683 1683 The new node is inserted before the head node.
1684 1684 """
1685 1685 head = self._head
1686 1686 node = _lrucachenode()
1687 1687 head.prev.next = node
1688 1688 node.prev = head.prev
1689 1689 node.next = head
1690 1690 head.prev = node
1691 1691 self._size += 1
1692 1692 return node
1693 1693
1694 1694 def _enforcecostlimit(self):
1695 1695 # This should run after an insertion. It should only be called if total
1696 1696 # cost limits are being enforced.
1697 1697 # The most recently inserted node is never evicted.
1698 1698 if len(self) <= 1 or self.totalcost <= self.maxcost:
1699 1699 return
1700 1700
1701 1701 # This is logically equivalent to calling popoldest() until we
1702 1702 # free up enough cost. We don't do that since popoldest() needs
1703 1703 # to walk the linked list and doing this in a loop would be
1704 1704 # quadratic. So we find the first non-empty node and then
1705 1705 # walk nodes until we free up enough capacity.
1706 1706 #
1707 1707 # If we only removed the minimum number of nodes to free enough
1708 1708 # cost at insert time, chances are high that the next insert would
1709 1709 # also require pruning. This would effectively constitute quadratic
1710 1710 # behavior for insert-heavy workloads. To mitigate this, we set a
1711 1711 # target cost that is a percentage of the max cost. This will tend
1712 1712 # to free more nodes when the high water mark is reached, which
1713 1713 # lowers the chances of needing to prune on the subsequent insert.
1714 1714 targetcost = int(self.maxcost * 0.75)
1715 1715
1716 1716 n = self._head.prev
1717 1717 while n.key is _notset:
1718 1718 n = n.prev
1719 1719
1720 1720 while len(self) > 1 and self.totalcost > targetcost:
1721 1721 del self._cache[n.key]
1722 1722 self.totalcost -= n.cost
1723 1723 n.markempty()
1724 1724 n = n.prev
1725 1725
1726 1726
1727 1727 def lrucachefunc(func):
1728 1728 '''cache most recent results of function calls'''
1729 1729 cache = {}
1730 1730 order = collections.deque()
1731 1731 if func.__code__.co_argcount == 1:
1732 1732
1733 1733 def f(arg):
1734 1734 if arg not in cache:
1735 1735 if len(cache) > 20:
1736 1736 del cache[order.popleft()]
1737 1737 cache[arg] = func(arg)
1738 1738 else:
1739 1739 order.remove(arg)
1740 1740 order.append(arg)
1741 1741 return cache[arg]
1742 1742
1743 1743 else:
1744 1744
1745 1745 def f(*args):
1746 1746 if args not in cache:
1747 1747 if len(cache) > 20:
1748 1748 del cache[order.popleft()]
1749 1749 cache[args] = func(*args)
1750 1750 else:
1751 1751 order.remove(args)
1752 1752 order.append(args)
1753 1753 return cache[args]
1754 1754
1755 1755 return f
1756 1756
1757 1757
1758 1758 class propertycache(object):
1759 1759 def __init__(self, func):
1760 1760 self.func = func
1761 1761 self.name = func.__name__
1762 1762
1763 1763 def __get__(self, obj, type=None):
1764 1764 result = self.func(obj)
1765 1765 self.cachevalue(obj, result)
1766 1766 return result
1767 1767
1768 1768 def cachevalue(self, obj, value):
1769 1769 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
1770 1770 obj.__dict__[self.name] = value
1771 1771
1772 1772
1773 1773 def clearcachedproperty(obj, prop):
1774 1774 '''clear a cached property value, if one has been set'''
1775 1775 prop = pycompat.sysstr(prop)
1776 1776 if prop in obj.__dict__:
1777 1777 del obj.__dict__[prop]
1778 1778
1779 1779
1780 1780 def increasingchunks(source, min=1024, max=65536):
1781 1781 """return no less than min bytes per chunk while data remains,
1782 1782 doubling min after each chunk until it reaches max"""
1783 1783
1784 1784 def log2(x):
1785 1785 if not x:
1786 1786 return 0
1787 1787 i = 0
1788 1788 while x:
1789 1789 x >>= 1
1790 1790 i += 1
1791 1791 return i - 1
1792 1792
1793 1793 buf = []
1794 1794 blen = 0
1795 1795 for chunk in source:
1796 1796 buf.append(chunk)
1797 1797 blen += len(chunk)
1798 1798 if blen >= min:
1799 1799 if min < max:
1800 1800 min = min << 1
1801 1801 nmin = 1 << log2(blen)
1802 1802 if nmin > min:
1803 1803 min = nmin
1804 1804 if min > max:
1805 1805 min = max
1806 1806 yield b''.join(buf)
1807 1807 blen = 0
1808 1808 buf = []
1809 1809 if buf:
1810 1810 yield b''.join(buf)
1811 1811
1812 1812
1813 1813 def always(fn):
1814 1814 return True
1815 1815
1816 1816
1817 1817 def never(fn):
1818 1818 return False
1819 1819
1820 1820
1821 1821 def nogc(func):
1822 1822 """disable garbage collector
1823 1823
1824 1824 Python's garbage collector triggers a GC each time a certain number of
1825 1825 container objects (the number being defined by gc.get_threshold()) are
1826 1826 allocated even when marked not to be tracked by the collector. Tracking has
1827 1827 no effect on when GCs are triggered, only on what objects the GC looks
1828 1828 into. As a workaround, disable GC while building complex (huge)
1829 1829 containers.
1830 1830
1831 1831 This garbage collector issue have been fixed in 2.7. But it still affect
1832 1832 CPython's performance.
1833 1833 """
1834 1834
1835 1835 def wrapper(*args, **kwargs):
1836 1836 gcenabled = gc.isenabled()
1837 1837 gc.disable()
1838 1838 try:
1839 1839 return func(*args, **kwargs)
1840 1840 finally:
1841 1841 if gcenabled:
1842 1842 gc.enable()
1843 1843
1844 1844 return wrapper
1845 1845
1846 1846
1847 1847 if pycompat.ispypy:
1848 1848 # PyPy runs slower with gc disabled
1849 1849 nogc = lambda x: x
1850 1850
1851 1851
1852 1852 def pathto(root, n1, n2):
1853 1853 # type: (bytes, bytes, bytes) -> bytes
1854 1854 """return the relative path from one place to another.
1855 1855 root should use os.sep to separate directories
1856 1856 n1 should use os.sep to separate directories
1857 1857 n2 should use "/" to separate directories
1858 1858 returns an os.sep-separated path.
1859 1859
1860 1860 If n1 is a relative path, it's assumed it's
1861 1861 relative to root.
1862 1862 n2 should always be relative to root.
1863 1863 """
1864 1864 if not n1:
1865 1865 return localpath(n2)
1866 1866 if os.path.isabs(n1):
1867 1867 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
1868 1868 return os.path.join(root, localpath(n2))
1869 1869 n2 = b'/'.join((pconvert(root), n2))
1870 1870 a, b = splitpath(n1), n2.split(b'/')
1871 1871 a.reverse()
1872 1872 b.reverse()
1873 1873 while a and b and a[-1] == b[-1]:
1874 1874 a.pop()
1875 1875 b.pop()
1876 1876 b.reverse()
1877 1877 return pycompat.ossep.join(([b'..'] * len(a)) + b) or b'.'
1878 1878
1879 1879
1880 1880 def checksignature(func, depth=1):
1881 1881 '''wrap a function with code to check for calling errors'''
1882 1882
1883 1883 def check(*args, **kwargs):
1884 1884 try:
1885 1885 return func(*args, **kwargs)
1886 1886 except TypeError:
1887 1887 if len(traceback.extract_tb(sys.exc_info()[2])) == depth:
1888 1888 raise error.SignatureError
1889 1889 raise
1890 1890
1891 1891 return check
1892 1892
1893 1893
1894 1894 # a whilelist of known filesystems where hardlink works reliably
1895 1895 _hardlinkfswhitelist = {
1896 1896 b'apfs',
1897 1897 b'btrfs',
1898 1898 b'ext2',
1899 1899 b'ext3',
1900 1900 b'ext4',
1901 1901 b'hfs',
1902 1902 b'jfs',
1903 1903 b'NTFS',
1904 1904 b'reiserfs',
1905 1905 b'tmpfs',
1906 1906 b'ufs',
1907 1907 b'xfs',
1908 1908 b'zfs',
1909 1909 }
1910 1910
1911 1911
1912 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1912 def copyfile(
1913 src, dest, hardlink=False, copystat=False, checkambig=False, nb_bytes=None
1914 ):
1913 1915 """copy a file, preserving mode and optionally other stat info like
1914 1916 atime/mtime
1915 1917
1916 1918 checkambig argument is used with filestat, and is useful only if
1917 1919 destination file is guarded by any lock (e.g. repo.lock or
1918 1920 repo.wlock).
1919 1921
1920 1922 copystat and checkambig should be exclusive.
1923
1924 nb_bytes: if set only copy the first `nb_bytes` of the source file.
1921 1925 """
1922 1926 assert not (copystat and checkambig)
1923 1927 oldstat = None
1924 1928 if os.path.lexists(dest):
1925 1929 if checkambig:
1926 1930 oldstat = checkambig and filestat.frompath(dest)
1927 1931 unlink(dest)
1928 1932 if hardlink:
1929 1933 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1930 1934 # unless we are confident that dest is on a whitelisted filesystem.
1931 1935 try:
1932 1936 fstype = getfstype(os.path.dirname(dest))
1933 1937 except OSError:
1934 1938 fstype = None
1935 1939 if fstype not in _hardlinkfswhitelist:
1936 1940 hardlink = False
1937 1941 if hardlink:
1938 1942 try:
1939 1943 oslink(src, dest)
1944 if nb_bytes is not None:
1945 m = "the `nb_bytes` argument is incompatible with `hardlink`"
1946 raise error.ProgrammingError(m)
1940 1947 return
1941 1948 except (IOError, OSError):
1942 1949 pass # fall back to normal copy
1943 1950 if os.path.islink(src):
1944 1951 os.symlink(os.readlink(src), dest)
1945 1952 # copytime is ignored for symlinks, but in general copytime isn't needed
1946 1953 # for them anyway
1954 if nb_bytes is not None:
1955 m = "cannot use `nb_bytes` on a symlink"
1956 raise error.ProgrammingError(m)
1947 1957 else:
1948 1958 try:
1949 1959 shutil.copyfile(src, dest)
1950 1960 if copystat:
1951 1961 # copystat also copies mode
1952 1962 shutil.copystat(src, dest)
1953 1963 else:
1954 1964 shutil.copymode(src, dest)
1955 1965 if oldstat and oldstat.stat:
1956 1966 newstat = filestat.frompath(dest)
1957 1967 if newstat.isambig(oldstat):
1958 1968 # stat of copied file is ambiguous to original one
1959 1969 advanced = (
1960 1970 oldstat.stat[stat.ST_MTIME] + 1
1961 1971 ) & 0x7FFFFFFF
1962 1972 os.utime(dest, (advanced, advanced))
1973 # We could do something smarter using `copy_file_range` call or similar
1974 if nb_bytes is not None:
1975 with open(dest, mode='r+') as f:
1976 f.truncate(nb_bytes)
1963 1977 except shutil.Error as inst:
1964 1978 raise error.Abort(stringutil.forcebytestr(inst))
1965 1979
1966 1980
1967 1981 def copyfiles(src, dst, hardlink=None, progress=None):
1968 1982 """Copy a directory tree using hardlinks if possible."""
1969 1983 num = 0
1970 1984
1971 1985 def settopic():
1972 1986 if progress:
1973 1987 progress.topic = _(b'linking') if hardlink else _(b'copying')
1974 1988
1975 1989 if os.path.isdir(src):
1976 1990 if hardlink is None:
1977 1991 hardlink = (
1978 1992 os.stat(src).st_dev == os.stat(os.path.dirname(dst)).st_dev
1979 1993 )
1980 1994 settopic()
1981 1995 os.mkdir(dst)
1982 1996 for name, kind in listdir(src):
1983 1997 srcname = os.path.join(src, name)
1984 1998 dstname = os.path.join(dst, name)
1985 1999 hardlink, n = copyfiles(srcname, dstname, hardlink, progress)
1986 2000 num += n
1987 2001 else:
1988 2002 if hardlink is None:
1989 2003 hardlink = (
1990 2004 os.stat(os.path.dirname(src)).st_dev
1991 2005 == os.stat(os.path.dirname(dst)).st_dev
1992 2006 )
1993 2007 settopic()
1994 2008
1995 2009 if hardlink:
1996 2010 try:
1997 2011 oslink(src, dst)
1998 2012 except (IOError, OSError):
1999 2013 hardlink = False
2000 2014 shutil.copy(src, dst)
2001 2015 else:
2002 2016 shutil.copy(src, dst)
2003 2017 num += 1
2004 2018 if progress:
2005 2019 progress.increment()
2006 2020
2007 2021 return hardlink, num
2008 2022
2009 2023
2010 2024 _winreservednames = {
2011 2025 b'con',
2012 2026 b'prn',
2013 2027 b'aux',
2014 2028 b'nul',
2015 2029 b'com1',
2016 2030 b'com2',
2017 2031 b'com3',
2018 2032 b'com4',
2019 2033 b'com5',
2020 2034 b'com6',
2021 2035 b'com7',
2022 2036 b'com8',
2023 2037 b'com9',
2024 2038 b'lpt1',
2025 2039 b'lpt2',
2026 2040 b'lpt3',
2027 2041 b'lpt4',
2028 2042 b'lpt5',
2029 2043 b'lpt6',
2030 2044 b'lpt7',
2031 2045 b'lpt8',
2032 2046 b'lpt9',
2033 2047 }
2034 2048 _winreservedchars = b':*?"<>|'
2035 2049
2036 2050
2037 2051 def checkwinfilename(path):
2038 2052 # type: (bytes) -> Optional[bytes]
2039 2053 r"""Check that the base-relative path is a valid filename on Windows.
2040 2054 Returns None if the path is ok, or a UI string describing the problem.
2041 2055
2042 2056 >>> checkwinfilename(b"just/a/normal/path")
2043 2057 >>> checkwinfilename(b"foo/bar/con.xml")
2044 2058 "filename contains 'con', which is reserved on Windows"
2045 2059 >>> checkwinfilename(b"foo/con.xml/bar")
2046 2060 "filename contains 'con', which is reserved on Windows"
2047 2061 >>> checkwinfilename(b"foo/bar/xml.con")
2048 2062 >>> checkwinfilename(b"foo/bar/AUX/bla.txt")
2049 2063 "filename contains 'AUX', which is reserved on Windows"
2050 2064 >>> checkwinfilename(b"foo/bar/bla:.txt")
2051 2065 "filename contains ':', which is reserved on Windows"
2052 2066 >>> checkwinfilename(b"foo/bar/b\07la.txt")
2053 2067 "filename contains '\\x07', which is invalid on Windows"
2054 2068 >>> checkwinfilename(b"foo/bar/bla ")
2055 2069 "filename ends with ' ', which is not allowed on Windows"
2056 2070 >>> checkwinfilename(b"../bar")
2057 2071 >>> checkwinfilename(b"foo\\")
2058 2072 "filename ends with '\\', which is invalid on Windows"
2059 2073 >>> checkwinfilename(b"foo\\/bar")
2060 2074 "directory name ends with '\\', which is invalid on Windows"
2061 2075 """
2062 2076 if path.endswith(b'\\'):
2063 2077 return _(b"filename ends with '\\', which is invalid on Windows")
2064 2078 if b'\\/' in path:
2065 2079 return _(b"directory name ends with '\\', which is invalid on Windows")
2066 2080 for n in path.replace(b'\\', b'/').split(b'/'):
2067 2081 if not n:
2068 2082 continue
2069 2083 for c in _filenamebytestr(n):
2070 2084 if c in _winreservedchars:
2071 2085 return (
2072 2086 _(
2073 2087 b"filename contains '%s', which is reserved "
2074 2088 b"on Windows"
2075 2089 )
2076 2090 % c
2077 2091 )
2078 2092 if ord(c) <= 31:
2079 2093 return _(
2080 2094 b"filename contains '%s', which is invalid on Windows"
2081 2095 ) % stringutil.escapestr(c)
2082 2096 base = n.split(b'.')[0]
2083 2097 if base and base.lower() in _winreservednames:
2084 2098 return (
2085 2099 _(b"filename contains '%s', which is reserved on Windows")
2086 2100 % base
2087 2101 )
2088 2102 t = n[-1:]
2089 2103 if t in b'. ' and n not in b'..':
2090 2104 return (
2091 2105 _(
2092 2106 b"filename ends with '%s', which is not allowed "
2093 2107 b"on Windows"
2094 2108 )
2095 2109 % t
2096 2110 )
2097 2111
2098 2112
2099 2113 timer = getattr(time, "perf_counter", None)
2100 2114
2101 2115 if pycompat.iswindows:
2102 2116 checkosfilename = checkwinfilename
2103 2117 if not timer:
2104 2118 timer = time.clock
2105 2119 else:
2106 2120 # mercurial.windows doesn't have platform.checkosfilename
2107 2121 checkosfilename = platform.checkosfilename # pytype: disable=module-attr
2108 2122 if not timer:
2109 2123 timer = time.time
2110 2124
2111 2125
2112 2126 def makelock(info, pathname):
2113 2127 """Create a lock file atomically if possible
2114 2128
2115 2129 This may leave a stale lock file if symlink isn't supported and signal
2116 2130 interrupt is enabled.
2117 2131 """
2118 2132 try:
2119 2133 return os.symlink(info, pathname)
2120 2134 except OSError as why:
2121 2135 if why.errno == errno.EEXIST:
2122 2136 raise
2123 2137 except AttributeError: # no symlink in os
2124 2138 pass
2125 2139
2126 2140 flags = os.O_CREAT | os.O_WRONLY | os.O_EXCL | getattr(os, 'O_BINARY', 0)
2127 2141 ld = os.open(pathname, flags)
2128 2142 os.write(ld, info)
2129 2143 os.close(ld)
2130 2144
2131 2145
2132 2146 def readlock(pathname):
2133 2147 # type: (bytes) -> bytes
2134 2148 try:
2135 2149 return readlink(pathname)
2136 2150 except OSError as why:
2137 2151 if why.errno not in (errno.EINVAL, errno.ENOSYS):
2138 2152 raise
2139 2153 except AttributeError: # no symlink in os
2140 2154 pass
2141 2155 with posixfile(pathname, b'rb') as fp:
2142 2156 return fp.read()
2143 2157
2144 2158
2145 2159 def fstat(fp):
2146 2160 '''stat file object that may not have fileno method.'''
2147 2161 try:
2148 2162 return os.fstat(fp.fileno())
2149 2163 except AttributeError:
2150 2164 return os.stat(fp.name)
2151 2165
2152 2166
2153 2167 # File system features
2154 2168
2155 2169
2156 2170 def fscasesensitive(path):
2157 2171 # type: (bytes) -> bool
2158 2172 """
2159 2173 Return true if the given path is on a case-sensitive filesystem
2160 2174
2161 2175 Requires a path (like /foo/.hg) ending with a foldable final
2162 2176 directory component.
2163 2177 """
2164 2178 s1 = os.lstat(path)
2165 2179 d, b = os.path.split(path)
2166 2180 b2 = b.upper()
2167 2181 if b == b2:
2168 2182 b2 = b.lower()
2169 2183 if b == b2:
2170 2184 return True # no evidence against case sensitivity
2171 2185 p2 = os.path.join(d, b2)
2172 2186 try:
2173 2187 s2 = os.lstat(p2)
2174 2188 if s2 == s1:
2175 2189 return False
2176 2190 return True
2177 2191 except OSError:
2178 2192 return True
2179 2193
2180 2194
2181 2195 _re2_input = lambda x: x
2182 2196 try:
2183 2197 import re2 # pytype: disable=import-error
2184 2198
2185 2199 _re2 = None
2186 2200 except ImportError:
2187 2201 _re2 = False
2188 2202
2189 2203
2190 2204 class _re(object):
2191 2205 def _checkre2(self):
2192 2206 global _re2
2193 2207 global _re2_input
2194 2208
2195 2209 check_pattern = br'\[([^\[]+)\]'
2196 2210 check_input = b'[ui]'
2197 2211 try:
2198 2212 # check if match works, see issue3964
2199 2213 _re2 = bool(re2.match(check_pattern, check_input))
2200 2214 except ImportError:
2201 2215 _re2 = False
2202 2216 except TypeError:
2203 2217 # the `pyre-2` project provides a re2 module that accept bytes
2204 2218 # the `fb-re2` project provides a re2 module that acccept sysstr
2205 2219 check_pattern = pycompat.sysstr(check_pattern)
2206 2220 check_input = pycompat.sysstr(check_input)
2207 2221 _re2 = bool(re2.match(check_pattern, check_input))
2208 2222 _re2_input = pycompat.sysstr
2209 2223
2210 2224 def compile(self, pat, flags=0):
2211 2225 """Compile a regular expression, using re2 if possible
2212 2226
2213 2227 For best performance, use only re2-compatible regexp features. The
2214 2228 only flags from the re module that are re2-compatible are
2215 2229 IGNORECASE and MULTILINE."""
2216 2230 if _re2 is None:
2217 2231 self._checkre2()
2218 2232 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
2219 2233 if flags & remod.IGNORECASE:
2220 2234 pat = b'(?i)' + pat
2221 2235 if flags & remod.MULTILINE:
2222 2236 pat = b'(?m)' + pat
2223 2237 try:
2224 2238 return re2.compile(_re2_input(pat))
2225 2239 except re2.error:
2226 2240 pass
2227 2241 return remod.compile(pat, flags)
2228 2242
2229 2243 @propertycache
2230 2244 def escape(self):
2231 2245 """Return the version of escape corresponding to self.compile.
2232 2246
2233 2247 This is imperfect because whether re2 or re is used for a particular
2234 2248 function depends on the flags, etc, but it's the best we can do.
2235 2249 """
2236 2250 global _re2
2237 2251 if _re2 is None:
2238 2252 self._checkre2()
2239 2253 if _re2:
2240 2254 return re2.escape
2241 2255 else:
2242 2256 return remod.escape
2243 2257
2244 2258
2245 2259 re = _re()
2246 2260
2247 2261 _fspathcache = {}
2248 2262
2249 2263
2250 2264 def fspath(name, root):
2251 2265 # type: (bytes, bytes) -> bytes
2252 2266 """Get name in the case stored in the filesystem
2253 2267
2254 2268 The name should be relative to root, and be normcase-ed for efficiency.
2255 2269
2256 2270 Note that this function is unnecessary, and should not be
2257 2271 called, for case-sensitive filesystems (simply because it's expensive).
2258 2272
2259 2273 The root should be normcase-ed, too.
2260 2274 """
2261 2275
2262 2276 def _makefspathcacheentry(dir):
2263 2277 return {normcase(n): n for n in os.listdir(dir)}
2264 2278
2265 2279 seps = pycompat.ossep
2266 2280 if pycompat.osaltsep:
2267 2281 seps = seps + pycompat.osaltsep
2268 2282 # Protect backslashes. This gets silly very quickly.
2269 2283 seps.replace(b'\\', b'\\\\')
2270 2284 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
2271 2285 dir = os.path.normpath(root)
2272 2286 result = []
2273 2287 for part, sep in pattern.findall(name):
2274 2288 if sep:
2275 2289 result.append(sep)
2276 2290 continue
2277 2291
2278 2292 if dir not in _fspathcache:
2279 2293 _fspathcache[dir] = _makefspathcacheentry(dir)
2280 2294 contents = _fspathcache[dir]
2281 2295
2282 2296 found = contents.get(part)
2283 2297 if not found:
2284 2298 # retry "once per directory" per "dirstate.walk" which
2285 2299 # may take place for each patches of "hg qpush", for example
2286 2300 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
2287 2301 found = contents.get(part)
2288 2302
2289 2303 result.append(found or part)
2290 2304 dir = os.path.join(dir, part)
2291 2305
2292 2306 return b''.join(result)
2293 2307
2294 2308
2295 2309 def checknlink(testfile):
2296 2310 # type: (bytes) -> bool
2297 2311 '''check whether hardlink count reporting works properly'''
2298 2312
2299 2313 # testfile may be open, so we need a separate file for checking to
2300 2314 # work around issue2543 (or testfile may get lost on Samba shares)
2301 2315 f1, f2, fp = None, None, None
2302 2316 try:
2303 2317 fd, f1 = pycompat.mkstemp(
2304 2318 prefix=b'.%s-' % os.path.basename(testfile),
2305 2319 suffix=b'1~',
2306 2320 dir=os.path.dirname(testfile),
2307 2321 )
2308 2322 os.close(fd)
2309 2323 f2 = b'%s2~' % f1[:-2]
2310 2324
2311 2325 oslink(f1, f2)
2312 2326 # nlinks() may behave differently for files on Windows shares if
2313 2327 # the file is open.
2314 2328 fp = posixfile(f2)
2315 2329 return nlinks(f2) > 1
2316 2330 except OSError:
2317 2331 return False
2318 2332 finally:
2319 2333 if fp is not None:
2320 2334 fp.close()
2321 2335 for f in (f1, f2):
2322 2336 try:
2323 2337 if f is not None:
2324 2338 os.unlink(f)
2325 2339 except OSError:
2326 2340 pass
2327 2341
2328 2342
2329 2343 def endswithsep(path):
2330 2344 # type: (bytes) -> bool
2331 2345 '''Check path ends with os.sep or os.altsep.'''
2332 2346 return bool( # help pytype
2333 2347 path.endswith(pycompat.ossep)
2334 2348 or pycompat.osaltsep
2335 2349 and path.endswith(pycompat.osaltsep)
2336 2350 )
2337 2351
2338 2352
2339 2353 def splitpath(path):
2340 2354 # type: (bytes) -> List[bytes]
2341 2355 """Split path by os.sep.
2342 2356 Note that this function does not use os.altsep because this is
2343 2357 an alternative of simple "xxx.split(os.sep)".
2344 2358 It is recommended to use os.path.normpath() before using this
2345 2359 function if need."""
2346 2360 return path.split(pycompat.ossep)
2347 2361
2348 2362
2349 2363 def mktempcopy(name, emptyok=False, createmode=None, enforcewritable=False):
2350 2364 """Create a temporary file with the same contents from name
2351 2365
2352 2366 The permission bits are copied from the original file.
2353 2367
2354 2368 If the temporary file is going to be truncated immediately, you
2355 2369 can use emptyok=True as an optimization.
2356 2370
2357 2371 Returns the name of the temporary file.
2358 2372 """
2359 2373 d, fn = os.path.split(name)
2360 2374 fd, temp = pycompat.mkstemp(prefix=b'.%s-' % fn, suffix=b'~', dir=d)
2361 2375 os.close(fd)
2362 2376 # Temporary files are created with mode 0600, which is usually not
2363 2377 # what we want. If the original file already exists, just copy
2364 2378 # its mode. Otherwise, manually obey umask.
2365 2379 copymode(name, temp, createmode, enforcewritable)
2366 2380
2367 2381 if emptyok:
2368 2382 return temp
2369 2383 try:
2370 2384 try:
2371 2385 ifp = posixfile(name, b"rb")
2372 2386 except IOError as inst:
2373 2387 if inst.errno == errno.ENOENT:
2374 2388 return temp
2375 2389 if not getattr(inst, 'filename', None):
2376 2390 inst.filename = name
2377 2391 raise
2378 2392 ofp = posixfile(temp, b"wb")
2379 2393 for chunk in filechunkiter(ifp):
2380 2394 ofp.write(chunk)
2381 2395 ifp.close()
2382 2396 ofp.close()
2383 2397 except: # re-raises
2384 2398 try:
2385 2399 os.unlink(temp)
2386 2400 except OSError:
2387 2401 pass
2388 2402 raise
2389 2403 return temp
2390 2404
2391 2405
2392 2406 class filestat(object):
2393 2407 """help to exactly detect change of a file
2394 2408
2395 2409 'stat' attribute is result of 'os.stat()' if specified 'path'
2396 2410 exists. Otherwise, it is None. This can avoid preparative
2397 2411 'exists()' examination on client side of this class.
2398 2412 """
2399 2413
2400 2414 def __init__(self, stat):
2401 2415 self.stat = stat
2402 2416
2403 2417 @classmethod
2404 2418 def frompath(cls, path):
2405 2419 try:
2406 2420 stat = os.stat(path)
2407 2421 except OSError as err:
2408 2422 if err.errno != errno.ENOENT:
2409 2423 raise
2410 2424 stat = None
2411 2425 return cls(stat)
2412 2426
2413 2427 @classmethod
2414 2428 def fromfp(cls, fp):
2415 2429 stat = os.fstat(fp.fileno())
2416 2430 return cls(stat)
2417 2431
2418 2432 __hash__ = object.__hash__
2419 2433
2420 2434 def __eq__(self, old):
2421 2435 try:
2422 2436 # if ambiguity between stat of new and old file is
2423 2437 # avoided, comparison of size, ctime and mtime is enough
2424 2438 # to exactly detect change of a file regardless of platform
2425 2439 return (
2426 2440 self.stat.st_size == old.stat.st_size
2427 2441 and self.stat[stat.ST_CTIME] == old.stat[stat.ST_CTIME]
2428 2442 and self.stat[stat.ST_MTIME] == old.stat[stat.ST_MTIME]
2429 2443 )
2430 2444 except AttributeError:
2431 2445 pass
2432 2446 try:
2433 2447 return self.stat is None and old.stat is None
2434 2448 except AttributeError:
2435 2449 return False
2436 2450
2437 2451 def isambig(self, old):
2438 2452 """Examine whether new (= self) stat is ambiguous against old one
2439 2453
2440 2454 "S[N]" below means stat of a file at N-th change:
2441 2455
2442 2456 - S[n-1].ctime < S[n].ctime: can detect change of a file
2443 2457 - S[n-1].ctime == S[n].ctime
2444 2458 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
2445 2459 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
2446 2460 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
2447 2461 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
2448 2462
2449 2463 Case (*2) above means that a file was changed twice or more at
2450 2464 same time in sec (= S[n-1].ctime), and comparison of timestamp
2451 2465 is ambiguous.
2452 2466
2453 2467 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
2454 2468 timestamp is ambiguous".
2455 2469
2456 2470 But advancing mtime only in case (*2) doesn't work as
2457 2471 expected, because naturally advanced S[n].mtime in case (*1)
2458 2472 might be equal to manually advanced S[n-1 or earlier].mtime.
2459 2473
2460 2474 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
2461 2475 treated as ambiguous regardless of mtime, to avoid overlooking
2462 2476 by confliction between such mtime.
2463 2477
2464 2478 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
2465 2479 S[n].mtime", even if size of a file isn't changed.
2466 2480 """
2467 2481 try:
2468 2482 return self.stat[stat.ST_CTIME] == old.stat[stat.ST_CTIME]
2469 2483 except AttributeError:
2470 2484 return False
2471 2485
2472 2486 def avoidambig(self, path, old):
2473 2487 """Change file stat of specified path to avoid ambiguity
2474 2488
2475 2489 'old' should be previous filestat of 'path'.
2476 2490
2477 2491 This skips avoiding ambiguity, if a process doesn't have
2478 2492 appropriate privileges for 'path'. This returns False in this
2479 2493 case.
2480 2494
2481 2495 Otherwise, this returns True, as "ambiguity is avoided".
2482 2496 """
2483 2497 advanced = (old.stat[stat.ST_MTIME] + 1) & 0x7FFFFFFF
2484 2498 try:
2485 2499 os.utime(path, (advanced, advanced))
2486 2500 except OSError as inst:
2487 2501 if inst.errno == errno.EPERM:
2488 2502 # utime() on the file created by another user causes EPERM,
2489 2503 # if a process doesn't have appropriate privileges
2490 2504 return False
2491 2505 raise
2492 2506 return True
2493 2507
2494 2508 def __ne__(self, other):
2495 2509 return not self == other
2496 2510
2497 2511
2498 2512 class atomictempfile(object):
2499 2513 """writable file object that atomically updates a file
2500 2514
2501 2515 All writes will go to a temporary copy of the original file. Call
2502 2516 close() when you are done writing, and atomictempfile will rename
2503 2517 the temporary copy to the original name, making the changes
2504 2518 visible. If the object is destroyed without being closed, all your
2505 2519 writes are discarded.
2506 2520
2507 2521 checkambig argument of constructor is used with filestat, and is
2508 2522 useful only if target file is guarded by any lock (e.g. repo.lock
2509 2523 or repo.wlock).
2510 2524 """
2511 2525
2512 2526 def __init__(self, name, mode=b'w+b', createmode=None, checkambig=False):
2513 2527 self.__name = name # permanent name
2514 2528 self._tempname = mktempcopy(
2515 2529 name,
2516 2530 emptyok=(b'w' in mode),
2517 2531 createmode=createmode,
2518 2532 enforcewritable=(b'w' in mode),
2519 2533 )
2520 2534
2521 2535 self._fp = posixfile(self._tempname, mode)
2522 2536 self._checkambig = checkambig
2523 2537
2524 2538 # delegated methods
2525 2539 self.read = self._fp.read
2526 2540 self.write = self._fp.write
2527 2541 self.seek = self._fp.seek
2528 2542 self.tell = self._fp.tell
2529 2543 self.fileno = self._fp.fileno
2530 2544
2531 2545 def close(self):
2532 2546 if not self._fp.closed:
2533 2547 self._fp.close()
2534 2548 filename = localpath(self.__name)
2535 2549 oldstat = self._checkambig and filestat.frompath(filename)
2536 2550 if oldstat and oldstat.stat:
2537 2551 rename(self._tempname, filename)
2538 2552 newstat = filestat.frompath(filename)
2539 2553 if newstat.isambig(oldstat):
2540 2554 # stat of changed file is ambiguous to original one
2541 2555 advanced = (oldstat.stat[stat.ST_MTIME] + 1) & 0x7FFFFFFF
2542 2556 os.utime(filename, (advanced, advanced))
2543 2557 else:
2544 2558 rename(self._tempname, filename)
2545 2559
2546 2560 def discard(self):
2547 2561 if not self._fp.closed:
2548 2562 try:
2549 2563 os.unlink(self._tempname)
2550 2564 except OSError:
2551 2565 pass
2552 2566 self._fp.close()
2553 2567
2554 2568 def __del__(self):
2555 2569 if safehasattr(self, '_fp'): # constructor actually did something
2556 2570 self.discard()
2557 2571
2558 2572 def __enter__(self):
2559 2573 return self
2560 2574
2561 2575 def __exit__(self, exctype, excvalue, traceback):
2562 2576 if exctype is not None:
2563 2577 self.discard()
2564 2578 else:
2565 2579 self.close()
2566 2580
2567 2581
2568 2582 def unlinkpath(f, ignoremissing=False, rmdir=True):
2569 2583 # type: (bytes, bool, bool) -> None
2570 2584 """unlink and remove the directory if it is empty"""
2571 2585 if ignoremissing:
2572 2586 tryunlink(f)
2573 2587 else:
2574 2588 unlink(f)
2575 2589 if rmdir:
2576 2590 # try removing directories that might now be empty
2577 2591 try:
2578 2592 removedirs(os.path.dirname(f))
2579 2593 except OSError:
2580 2594 pass
2581 2595
2582 2596
2583 2597 def tryunlink(f):
2584 2598 # type: (bytes) -> None
2585 2599 """Attempt to remove a file, ignoring ENOENT errors."""
2586 2600 try:
2587 2601 unlink(f)
2588 2602 except OSError as e:
2589 2603 if e.errno != errno.ENOENT:
2590 2604 raise
2591 2605
2592 2606
2593 2607 def makedirs(name, mode=None, notindexed=False):
2594 2608 # type: (bytes, Optional[int], bool) -> None
2595 2609 """recursive directory creation with parent mode inheritance
2596 2610
2597 2611 Newly created directories are marked as "not to be indexed by
2598 2612 the content indexing service", if ``notindexed`` is specified
2599 2613 for "write" mode access.
2600 2614 """
2601 2615 try:
2602 2616 makedir(name, notindexed)
2603 2617 except OSError as err:
2604 2618 if err.errno == errno.EEXIST:
2605 2619 return
2606 2620 if err.errno != errno.ENOENT or not name:
2607 2621 raise
2608 2622 parent = os.path.dirname(os.path.abspath(name))
2609 2623 if parent == name:
2610 2624 raise
2611 2625 makedirs(parent, mode, notindexed)
2612 2626 try:
2613 2627 makedir(name, notindexed)
2614 2628 except OSError as err:
2615 2629 # Catch EEXIST to handle races
2616 2630 if err.errno == errno.EEXIST:
2617 2631 return
2618 2632 raise
2619 2633 if mode is not None:
2620 2634 os.chmod(name, mode)
2621 2635
2622 2636
2623 2637 def readfile(path):
2624 2638 # type: (bytes) -> bytes
2625 2639 with open(path, b'rb') as fp:
2626 2640 return fp.read()
2627 2641
2628 2642
2629 2643 def writefile(path, text):
2630 2644 # type: (bytes, bytes) -> None
2631 2645 with open(path, b'wb') as fp:
2632 2646 fp.write(text)
2633 2647
2634 2648
2635 2649 def appendfile(path, text):
2636 2650 # type: (bytes, bytes) -> None
2637 2651 with open(path, b'ab') as fp:
2638 2652 fp.write(text)
2639 2653
2640 2654
2641 2655 class chunkbuffer(object):
2642 2656 """Allow arbitrary sized chunks of data to be efficiently read from an
2643 2657 iterator over chunks of arbitrary size."""
2644 2658
2645 2659 def __init__(self, in_iter):
2646 2660 """in_iter is the iterator that's iterating over the input chunks."""
2647 2661
2648 2662 def splitbig(chunks):
2649 2663 for chunk in chunks:
2650 2664 if len(chunk) > 2 ** 20:
2651 2665 pos = 0
2652 2666 while pos < len(chunk):
2653 2667 end = pos + 2 ** 18
2654 2668 yield chunk[pos:end]
2655 2669 pos = end
2656 2670 else:
2657 2671 yield chunk
2658 2672
2659 2673 self.iter = splitbig(in_iter)
2660 2674 self._queue = collections.deque()
2661 2675 self._chunkoffset = 0
2662 2676
2663 2677 def read(self, l=None):
2664 2678 """Read L bytes of data from the iterator of chunks of data.
2665 2679 Returns less than L bytes if the iterator runs dry.
2666 2680
2667 2681 If size parameter is omitted, read everything"""
2668 2682 if l is None:
2669 2683 return b''.join(self.iter)
2670 2684
2671 2685 left = l
2672 2686 buf = []
2673 2687 queue = self._queue
2674 2688 while left > 0:
2675 2689 # refill the queue
2676 2690 if not queue:
2677 2691 target = 2 ** 18
2678 2692 for chunk in self.iter:
2679 2693 queue.append(chunk)
2680 2694 target -= len(chunk)
2681 2695 if target <= 0:
2682 2696 break
2683 2697 if not queue:
2684 2698 break
2685 2699
2686 2700 # The easy way to do this would be to queue.popleft(), modify the
2687 2701 # chunk (if necessary), then queue.appendleft(). However, for cases
2688 2702 # where we read partial chunk content, this incurs 2 dequeue
2689 2703 # mutations and creates a new str for the remaining chunk in the
2690 2704 # queue. Our code below avoids this overhead.
2691 2705
2692 2706 chunk = queue[0]
2693 2707 chunkl = len(chunk)
2694 2708 offset = self._chunkoffset
2695 2709
2696 2710 # Use full chunk.
2697 2711 if offset == 0 and left >= chunkl:
2698 2712 left -= chunkl
2699 2713 queue.popleft()
2700 2714 buf.append(chunk)
2701 2715 # self._chunkoffset remains at 0.
2702 2716 continue
2703 2717
2704 2718 chunkremaining = chunkl - offset
2705 2719
2706 2720 # Use all of unconsumed part of chunk.
2707 2721 if left >= chunkremaining:
2708 2722 left -= chunkremaining
2709 2723 queue.popleft()
2710 2724 # offset == 0 is enabled by block above, so this won't merely
2711 2725 # copy via ``chunk[0:]``.
2712 2726 buf.append(chunk[offset:])
2713 2727 self._chunkoffset = 0
2714 2728
2715 2729 # Partial chunk needed.
2716 2730 else:
2717 2731 buf.append(chunk[offset : offset + left])
2718 2732 self._chunkoffset += left
2719 2733 left -= chunkremaining
2720 2734
2721 2735 return b''.join(buf)
2722 2736
2723 2737
2724 2738 def filechunkiter(f, size=131072, limit=None):
2725 2739 """Create a generator that produces the data in the file size
2726 2740 (default 131072) bytes at a time, up to optional limit (default is
2727 2741 to read all data). Chunks may be less than size bytes if the
2728 2742 chunk is the last chunk in the file, or the file is a socket or
2729 2743 some other type of file that sometimes reads less data than is
2730 2744 requested."""
2731 2745 assert size >= 0
2732 2746 assert limit is None or limit >= 0
2733 2747 while True:
2734 2748 if limit is None:
2735 2749 nbytes = size
2736 2750 else:
2737 2751 nbytes = min(limit, size)
2738 2752 s = nbytes and f.read(nbytes)
2739 2753 if not s:
2740 2754 break
2741 2755 if limit:
2742 2756 limit -= len(s)
2743 2757 yield s
2744 2758
2745 2759
2746 2760 class cappedreader(object):
2747 2761 """A file object proxy that allows reading up to N bytes.
2748 2762
2749 2763 Given a source file object, instances of this type allow reading up to
2750 2764 N bytes from that source file object. Attempts to read past the allowed
2751 2765 limit are treated as EOF.
2752 2766
2753 2767 It is assumed that I/O is not performed on the original file object
2754 2768 in addition to I/O that is performed by this instance. If there is,
2755 2769 state tracking will get out of sync and unexpected results will ensue.
2756 2770 """
2757 2771
2758 2772 def __init__(self, fh, limit):
2759 2773 """Allow reading up to <limit> bytes from <fh>."""
2760 2774 self._fh = fh
2761 2775 self._left = limit
2762 2776
2763 2777 def read(self, n=-1):
2764 2778 if not self._left:
2765 2779 return b''
2766 2780
2767 2781 if n < 0:
2768 2782 n = self._left
2769 2783
2770 2784 data = self._fh.read(min(n, self._left))
2771 2785 self._left -= len(data)
2772 2786 assert self._left >= 0
2773 2787
2774 2788 return data
2775 2789
2776 2790 def readinto(self, b):
2777 2791 res = self.read(len(b))
2778 2792 if res is None:
2779 2793 return None
2780 2794
2781 2795 b[0 : len(res)] = res
2782 2796 return len(res)
2783 2797
2784 2798
2785 2799 def unitcountfn(*unittable):
2786 2800 '''return a function that renders a readable count of some quantity'''
2787 2801
2788 2802 def go(count):
2789 2803 for multiplier, divisor, format in unittable:
2790 2804 if abs(count) >= divisor * multiplier:
2791 2805 return format % (count / float(divisor))
2792 2806 return unittable[-1][2] % count
2793 2807
2794 2808 return go
2795 2809
2796 2810
2797 2811 def processlinerange(fromline, toline):
2798 2812 # type: (int, int) -> Tuple[int, int]
2799 2813 """Check that linerange <fromline>:<toline> makes sense and return a
2800 2814 0-based range.
2801 2815
2802 2816 >>> processlinerange(10, 20)
2803 2817 (9, 20)
2804 2818 >>> processlinerange(2, 1)
2805 2819 Traceback (most recent call last):
2806 2820 ...
2807 2821 ParseError: line range must be positive
2808 2822 >>> processlinerange(0, 5)
2809 2823 Traceback (most recent call last):
2810 2824 ...
2811 2825 ParseError: fromline must be strictly positive
2812 2826 """
2813 2827 if toline - fromline < 0:
2814 2828 raise error.ParseError(_(b"line range must be positive"))
2815 2829 if fromline < 1:
2816 2830 raise error.ParseError(_(b"fromline must be strictly positive"))
2817 2831 return fromline - 1, toline
2818 2832
2819 2833
2820 2834 bytecount = unitcountfn(
2821 2835 (100, 1 << 30, _(b'%.0f GB')),
2822 2836 (10, 1 << 30, _(b'%.1f GB')),
2823 2837 (1, 1 << 30, _(b'%.2f GB')),
2824 2838 (100, 1 << 20, _(b'%.0f MB')),
2825 2839 (10, 1 << 20, _(b'%.1f MB')),
2826 2840 (1, 1 << 20, _(b'%.2f MB')),
2827 2841 (100, 1 << 10, _(b'%.0f KB')),
2828 2842 (10, 1 << 10, _(b'%.1f KB')),
2829 2843 (1, 1 << 10, _(b'%.2f KB')),
2830 2844 (1, 1, _(b'%.0f bytes')),
2831 2845 )
2832 2846
2833 2847
2834 2848 class transformingwriter(object):
2835 2849 """Writable file wrapper to transform data by function"""
2836 2850
2837 2851 def __init__(self, fp, encode):
2838 2852 self._fp = fp
2839 2853 self._encode = encode
2840 2854
2841 2855 def close(self):
2842 2856 self._fp.close()
2843 2857
2844 2858 def flush(self):
2845 2859 self._fp.flush()
2846 2860
2847 2861 def write(self, data):
2848 2862 return self._fp.write(self._encode(data))
2849 2863
2850 2864
2851 2865 # Matches a single EOL which can either be a CRLF where repeated CR
2852 2866 # are removed or a LF. We do not care about old Macintosh files, so a
2853 2867 # stray CR is an error.
2854 2868 _eolre = remod.compile(br'\r*\n')
2855 2869
2856 2870
2857 2871 def tolf(s):
2858 2872 # type: (bytes) -> bytes
2859 2873 return _eolre.sub(b'\n', s)
2860 2874
2861 2875
2862 2876 def tocrlf(s):
2863 2877 # type: (bytes) -> bytes
2864 2878 return _eolre.sub(b'\r\n', s)
2865 2879
2866 2880
2867 2881 def _crlfwriter(fp):
2868 2882 return transformingwriter(fp, tocrlf)
2869 2883
2870 2884
2871 2885 if pycompat.oslinesep == b'\r\n':
2872 2886 tonativeeol = tocrlf
2873 2887 fromnativeeol = tolf
2874 2888 nativeeolwriter = _crlfwriter
2875 2889 else:
2876 2890 tonativeeol = pycompat.identity
2877 2891 fromnativeeol = pycompat.identity
2878 2892 nativeeolwriter = pycompat.identity
2879 2893
2880 2894 if pyplatform.python_implementation() == b'CPython' and sys.version_info < (
2881 2895 3,
2882 2896 0,
2883 2897 ):
2884 2898 # There is an issue in CPython that some IO methods do not handle EINTR
2885 2899 # correctly. The following table shows what CPython version (and functions)
2886 2900 # are affected (buggy: has the EINTR bug, okay: otherwise):
2887 2901 #
2888 2902 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2889 2903 # --------------------------------------------------
2890 2904 # fp.__iter__ | buggy | buggy | okay
2891 2905 # fp.read* | buggy | okay [1] | okay
2892 2906 #
2893 2907 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2894 2908 #
2895 2909 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2896 2910 # like "read*" work fine, as we do not support Python < 2.7.4.
2897 2911 #
2898 2912 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2899 2913 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2900 2914 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2901 2915 # fp.__iter__ but not other fp.read* methods.
2902 2916 #
2903 2917 # On modern systems like Linux, the "read" syscall cannot be interrupted
2904 2918 # when reading "fast" files like on-disk files. So the EINTR issue only
2905 2919 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2906 2920 # files approximately as "fast" files and use the fast (unsafe) code path,
2907 2921 # to minimize the performance impact.
2908 2922
2909 2923 def iterfile(fp):
2910 2924 fastpath = True
2911 2925 if type(fp) is file:
2912 2926 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2913 2927 if fastpath:
2914 2928 return fp
2915 2929 else:
2916 2930 # fp.readline deals with EINTR correctly, use it as a workaround.
2917 2931 return iter(fp.readline, b'')
2918 2932
2919 2933
2920 2934 else:
2921 2935 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2922 2936 def iterfile(fp):
2923 2937 return fp
2924 2938
2925 2939
2926 2940 def iterlines(iterator):
2927 2941 # type: (Iterator[bytes]) -> Iterator[bytes]
2928 2942 for chunk in iterator:
2929 2943 for line in chunk.splitlines():
2930 2944 yield line
2931 2945
2932 2946
2933 2947 def expandpath(path):
2934 2948 # type: (bytes) -> bytes
2935 2949 return os.path.expanduser(os.path.expandvars(path))
2936 2950
2937 2951
2938 2952 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2939 2953 """Return the result of interpolating items in the mapping into string s.
2940 2954
2941 2955 prefix is a single character string, or a two character string with
2942 2956 a backslash as the first character if the prefix needs to be escaped in
2943 2957 a regular expression.
2944 2958
2945 2959 fn is an optional function that will be applied to the replacement text
2946 2960 just before replacement.
2947 2961
2948 2962 escape_prefix is an optional flag that allows using doubled prefix for
2949 2963 its escaping.
2950 2964 """
2951 2965 fn = fn or (lambda s: s)
2952 2966 patterns = b'|'.join(mapping.keys())
2953 2967 if escape_prefix:
2954 2968 patterns += b'|' + prefix
2955 2969 if len(prefix) > 1:
2956 2970 prefix_char = prefix[1:]
2957 2971 else:
2958 2972 prefix_char = prefix
2959 2973 mapping[prefix_char] = prefix_char
2960 2974 r = remod.compile(br'%s(%s)' % (prefix, patterns))
2961 2975 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2962 2976
2963 2977
2964 2978 def getport(*args, **kwargs):
2965 2979 msg = b'getport(...) moved to mercurial.utils.urlutil'
2966 2980 nouideprecwarn(msg, b'6.0', stacklevel=2)
2967 2981 return urlutil.getport(*args, **kwargs)
2968 2982
2969 2983
2970 2984 def url(*args, **kwargs):
2971 2985 msg = b'url(...) moved to mercurial.utils.urlutil'
2972 2986 nouideprecwarn(msg, b'6.0', stacklevel=2)
2973 2987 return urlutil.url(*args, **kwargs)
2974 2988
2975 2989
2976 2990 def hasscheme(*args, **kwargs):
2977 2991 msg = b'hasscheme(...) moved to mercurial.utils.urlutil'
2978 2992 nouideprecwarn(msg, b'6.0', stacklevel=2)
2979 2993 return urlutil.hasscheme(*args, **kwargs)
2980 2994
2981 2995
2982 2996 def hasdriveletter(*args, **kwargs):
2983 2997 msg = b'hasdriveletter(...) moved to mercurial.utils.urlutil'
2984 2998 nouideprecwarn(msg, b'6.0', stacklevel=2)
2985 2999 return urlutil.hasdriveletter(*args, **kwargs)
2986 3000
2987 3001
2988 3002 def urllocalpath(*args, **kwargs):
2989 3003 msg = b'urllocalpath(...) moved to mercurial.utils.urlutil'
2990 3004 nouideprecwarn(msg, b'6.0', stacklevel=2)
2991 3005 return urlutil.urllocalpath(*args, **kwargs)
2992 3006
2993 3007
2994 3008 def checksafessh(*args, **kwargs):
2995 3009 msg = b'checksafessh(...) moved to mercurial.utils.urlutil'
2996 3010 nouideprecwarn(msg, b'6.0', stacklevel=2)
2997 3011 return urlutil.checksafessh(*args, **kwargs)
2998 3012
2999 3013
3000 3014 def hidepassword(*args, **kwargs):
3001 3015 msg = b'hidepassword(...) moved to mercurial.utils.urlutil'
3002 3016 nouideprecwarn(msg, b'6.0', stacklevel=2)
3003 3017 return urlutil.hidepassword(*args, **kwargs)
3004 3018
3005 3019
3006 3020 def removeauth(*args, **kwargs):
3007 3021 msg = b'removeauth(...) moved to mercurial.utils.urlutil'
3008 3022 nouideprecwarn(msg, b'6.0', stacklevel=2)
3009 3023 return urlutil.removeauth(*args, **kwargs)
3010 3024
3011 3025
3012 3026 timecount = unitcountfn(
3013 3027 (1, 1e3, _(b'%.0f s')),
3014 3028 (100, 1, _(b'%.1f s')),
3015 3029 (10, 1, _(b'%.2f s')),
3016 3030 (1, 1, _(b'%.3f s')),
3017 3031 (100, 0.001, _(b'%.1f ms')),
3018 3032 (10, 0.001, _(b'%.2f ms')),
3019 3033 (1, 0.001, _(b'%.3f ms')),
3020 3034 (100, 0.000001, _(b'%.1f us')),
3021 3035 (10, 0.000001, _(b'%.2f us')),
3022 3036 (1, 0.000001, _(b'%.3f us')),
3023 3037 (100, 0.000000001, _(b'%.1f ns')),
3024 3038 (10, 0.000000001, _(b'%.2f ns')),
3025 3039 (1, 0.000000001, _(b'%.3f ns')),
3026 3040 )
3027 3041
3028 3042
3029 3043 @attr.s
3030 3044 class timedcmstats(object):
3031 3045 """Stats information produced by the timedcm context manager on entering."""
3032 3046
3033 3047 # the starting value of the timer as a float (meaning and resulution is
3034 3048 # platform dependent, see util.timer)
3035 3049 start = attr.ib(default=attr.Factory(lambda: timer()))
3036 3050 # the number of seconds as a floating point value; starts at 0, updated when
3037 3051 # the context is exited.
3038 3052 elapsed = attr.ib(default=0)
3039 3053 # the number of nested timedcm context managers.
3040 3054 level = attr.ib(default=1)
3041 3055
3042 3056 def __bytes__(self):
3043 3057 return timecount(self.elapsed) if self.elapsed else b'<unknown>'
3044 3058
3045 3059 __str__ = encoding.strmethod(__bytes__)
3046 3060
3047 3061
3048 3062 @contextlib.contextmanager
3049 3063 def timedcm(whencefmt, *whenceargs):
3050 3064 """A context manager that produces timing information for a given context.
3051 3065
3052 3066 On entering a timedcmstats instance is produced.
3053 3067
3054 3068 This context manager is reentrant.
3055 3069
3056 3070 """
3057 3071 # track nested context managers
3058 3072 timedcm._nested += 1
3059 3073 timing_stats = timedcmstats(level=timedcm._nested)
3060 3074 try:
3061 3075 with tracing.log(whencefmt, *whenceargs):
3062 3076 yield timing_stats
3063 3077 finally:
3064 3078 timing_stats.elapsed = timer() - timing_stats.start
3065 3079 timedcm._nested -= 1
3066 3080
3067 3081
3068 3082 timedcm._nested = 0
3069 3083
3070 3084
3071 3085 def timed(func):
3072 3086 """Report the execution time of a function call to stderr.
3073 3087
3074 3088 During development, use as a decorator when you need to measure
3075 3089 the cost of a function, e.g. as follows:
3076 3090
3077 3091 @util.timed
3078 3092 def foo(a, b, c):
3079 3093 pass
3080 3094 """
3081 3095
3082 3096 def wrapper(*args, **kwargs):
3083 3097 with timedcm(pycompat.bytestr(func.__name__)) as time_stats:
3084 3098 result = func(*args, **kwargs)
3085 3099 stderr = procutil.stderr
3086 3100 stderr.write(
3087 3101 b'%s%s: %s\n'
3088 3102 % (
3089 3103 b' ' * time_stats.level * 2,
3090 3104 pycompat.bytestr(func.__name__),
3091 3105 time_stats,
3092 3106 )
3093 3107 )
3094 3108 return result
3095 3109
3096 3110 return wrapper
3097 3111
3098 3112
3099 3113 _sizeunits = (
3100 3114 (b'm', 2 ** 20),
3101 3115 (b'k', 2 ** 10),
3102 3116 (b'g', 2 ** 30),
3103 3117 (b'kb', 2 ** 10),
3104 3118 (b'mb', 2 ** 20),
3105 3119 (b'gb', 2 ** 30),
3106 3120 (b'b', 1),
3107 3121 )
3108 3122
3109 3123
3110 3124 def sizetoint(s):
3111 3125 # type: (bytes) -> int
3112 3126 """Convert a space specifier to a byte count.
3113 3127
3114 3128 >>> sizetoint(b'30')
3115 3129 30
3116 3130 >>> sizetoint(b'2.2kb')
3117 3131 2252
3118 3132 >>> sizetoint(b'6M')
3119 3133 6291456
3120 3134 """
3121 3135 t = s.strip().lower()
3122 3136 try:
3123 3137 for k, u in _sizeunits:
3124 3138 if t.endswith(k):
3125 3139 return int(float(t[: -len(k)]) * u)
3126 3140 return int(t)
3127 3141 except ValueError:
3128 3142 raise error.ParseError(_(b"couldn't parse size: %s") % s)
3129 3143
3130 3144
3131 3145 class hooks(object):
3132 3146 """A collection of hook functions that can be used to extend a
3133 3147 function's behavior. Hooks are called in lexicographic order,
3134 3148 based on the names of their sources."""
3135 3149
3136 3150 def __init__(self):
3137 3151 self._hooks = []
3138 3152
3139 3153 def add(self, source, hook):
3140 3154 self._hooks.append((source, hook))
3141 3155
3142 3156 def __call__(self, *args):
3143 3157 self._hooks.sort(key=lambda x: x[0])
3144 3158 results = []
3145 3159 for source, hook in self._hooks:
3146 3160 results.append(hook(*args))
3147 3161 return results
3148 3162
3149 3163
3150 3164 def getstackframes(skip=0, line=b' %-*s in %s\n', fileline=b'%s:%d', depth=0):
3151 3165 """Yields lines for a nicely formatted stacktrace.
3152 3166 Skips the 'skip' last entries, then return the last 'depth' entries.
3153 3167 Each file+linenumber is formatted according to fileline.
3154 3168 Each line is formatted according to line.
3155 3169 If line is None, it yields:
3156 3170 length of longest filepath+line number,
3157 3171 filepath+linenumber,
3158 3172 function
3159 3173
3160 3174 Not be used in production code but very convenient while developing.
3161 3175 """
3162 3176 entries = [
3163 3177 (fileline % (pycompat.sysbytes(fn), ln), pycompat.sysbytes(func))
3164 3178 for fn, ln, func, _text in traceback.extract_stack()[: -skip - 1]
3165 3179 ][-depth:]
3166 3180 if entries:
3167 3181 fnmax = max(len(entry[0]) for entry in entries)
3168 3182 for fnln, func in entries:
3169 3183 if line is None:
3170 3184 yield (fnmax, fnln, func)
3171 3185 else:
3172 3186 yield line % (fnmax, fnln, func)
3173 3187
3174 3188
3175 3189 def debugstacktrace(
3176 3190 msg=b'stacktrace',
3177 3191 skip=0,
3178 3192 f=procutil.stderr,
3179 3193 otherf=procutil.stdout,
3180 3194 depth=0,
3181 3195 prefix=b'',
3182 3196 ):
3183 3197 """Writes a message to f (stderr) with a nicely formatted stacktrace.
3184 3198 Skips the 'skip' entries closest to the call, then show 'depth' entries.
3185 3199 By default it will flush stdout first.
3186 3200 It can be used everywhere and intentionally does not require an ui object.
3187 3201 Not be used in production code but very convenient while developing.
3188 3202 """
3189 3203 if otherf:
3190 3204 otherf.flush()
3191 3205 f.write(b'%s%s at:\n' % (prefix, msg.rstrip()))
3192 3206 for line in getstackframes(skip + 1, depth=depth):
3193 3207 f.write(prefix + line)
3194 3208 f.flush()
3195 3209
3196 3210
3197 3211 # convenient shortcut
3198 3212 dst = debugstacktrace
3199 3213
3200 3214
3201 3215 def safename(f, tag, ctx, others=None):
3202 3216 """
3203 3217 Generate a name that it is safe to rename f to in the given context.
3204 3218
3205 3219 f: filename to rename
3206 3220 tag: a string tag that will be included in the new name
3207 3221 ctx: a context, in which the new name must not exist
3208 3222 others: a set of other filenames that the new name must not be in
3209 3223
3210 3224 Returns a file name of the form oldname~tag[~number] which does not exist
3211 3225 in the provided context and is not in the set of other names.
3212 3226 """
3213 3227 if others is None:
3214 3228 others = set()
3215 3229
3216 3230 fn = b'%s~%s' % (f, tag)
3217 3231 if fn not in ctx and fn not in others:
3218 3232 return fn
3219 3233 for n in itertools.count(1):
3220 3234 fn = b'%s~%s~%s' % (f, tag, n)
3221 3235 if fn not in ctx and fn not in others:
3222 3236 return fn
3223 3237
3224 3238
3225 3239 def readexactly(stream, n):
3226 3240 '''read n bytes from stream.read and abort if less was available'''
3227 3241 s = stream.read(n)
3228 3242 if len(s) < n:
3229 3243 raise error.Abort(
3230 3244 _(b"stream ended unexpectedly (got %d bytes, expected %d)")
3231 3245 % (len(s), n)
3232 3246 )
3233 3247 return s
3234 3248
3235 3249
3236 3250 def uvarintencode(value):
3237 3251 """Encode an unsigned integer value to a varint.
3238 3252
3239 3253 A varint is a variable length integer of 1 or more bytes. Each byte
3240 3254 except the last has the most significant bit set. The lower 7 bits of
3241 3255 each byte store the 2's complement representation, least significant group
3242 3256 first.
3243 3257
3244 3258 >>> uvarintencode(0)
3245 3259 '\\x00'
3246 3260 >>> uvarintencode(1)
3247 3261 '\\x01'
3248 3262 >>> uvarintencode(127)
3249 3263 '\\x7f'
3250 3264 >>> uvarintencode(1337)
3251 3265 '\\xb9\\n'
3252 3266 >>> uvarintencode(65536)
3253 3267 '\\x80\\x80\\x04'
3254 3268 >>> uvarintencode(-1)
3255 3269 Traceback (most recent call last):
3256 3270 ...
3257 3271 ProgrammingError: negative value for uvarint: -1
3258 3272 """
3259 3273 if value < 0:
3260 3274 raise error.ProgrammingError(b'negative value for uvarint: %d' % value)
3261 3275 bits = value & 0x7F
3262 3276 value >>= 7
3263 3277 bytes = []
3264 3278 while value:
3265 3279 bytes.append(pycompat.bytechr(0x80 | bits))
3266 3280 bits = value & 0x7F
3267 3281 value >>= 7
3268 3282 bytes.append(pycompat.bytechr(bits))
3269 3283
3270 3284 return b''.join(bytes)
3271 3285
3272 3286
3273 3287 def uvarintdecodestream(fh):
3274 3288 """Decode an unsigned variable length integer from a stream.
3275 3289
3276 3290 The passed argument is anything that has a ``.read(N)`` method.
3277 3291
3278 3292 >>> try:
3279 3293 ... from StringIO import StringIO as BytesIO
3280 3294 ... except ImportError:
3281 3295 ... from io import BytesIO
3282 3296 >>> uvarintdecodestream(BytesIO(b'\\x00'))
3283 3297 0
3284 3298 >>> uvarintdecodestream(BytesIO(b'\\x01'))
3285 3299 1
3286 3300 >>> uvarintdecodestream(BytesIO(b'\\x7f'))
3287 3301 127
3288 3302 >>> uvarintdecodestream(BytesIO(b'\\xb9\\n'))
3289 3303 1337
3290 3304 >>> uvarintdecodestream(BytesIO(b'\\x80\\x80\\x04'))
3291 3305 65536
3292 3306 >>> uvarintdecodestream(BytesIO(b'\\x80'))
3293 3307 Traceback (most recent call last):
3294 3308 ...
3295 3309 Abort: stream ended unexpectedly (got 0 bytes, expected 1)
3296 3310 """
3297 3311 result = 0
3298 3312 shift = 0
3299 3313 while True:
3300 3314 byte = ord(readexactly(fh, 1))
3301 3315 result |= (byte & 0x7F) << shift
3302 3316 if not (byte & 0x80):
3303 3317 return result
3304 3318 shift += 7
3305 3319
3306 3320
3307 3321 # Passing the '' locale means that the locale should be set according to the
3308 3322 # user settings (environment variables).
3309 3323 # Python sometimes avoids setting the global locale settings. When interfacing
3310 3324 # with C code (e.g. the curses module or the Subversion bindings), the global
3311 3325 # locale settings must be initialized correctly. Python 2 does not initialize
3312 3326 # the global locale settings on interpreter startup. Python 3 sometimes
3313 3327 # initializes LC_CTYPE, but not consistently at least on Windows. Therefore we
3314 3328 # explicitly initialize it to get consistent behavior if it's not already
3315 3329 # initialized. Since CPython commit 177d921c8c03d30daa32994362023f777624b10d,
3316 3330 # LC_CTYPE is always initialized. If we require Python 3.8+, we should re-check
3317 3331 # if we can remove this code.
3318 3332 @contextlib.contextmanager
3319 3333 def with_lc_ctype():
3320 3334 oldloc = locale.setlocale(locale.LC_CTYPE, None)
3321 3335 if oldloc == 'C':
3322 3336 try:
3323 3337 try:
3324 3338 locale.setlocale(locale.LC_CTYPE, '')
3325 3339 except locale.Error:
3326 3340 # The likely case is that the locale from the environment
3327 3341 # variables is unknown.
3328 3342 pass
3329 3343 yield
3330 3344 finally:
3331 3345 locale.setlocale(locale.LC_CTYPE, oldloc)
3332 3346 else:
3333 3347 yield
3334 3348
3335 3349
3336 3350 def _estimatememory():
3337 3351 # type: () -> Optional[int]
3338 3352 """Provide an estimate for the available system memory in Bytes.
3339 3353
3340 3354 If no estimate can be provided on the platform, returns None.
3341 3355 """
3342 3356 if pycompat.sysplatform.startswith(b'win'):
3343 3357 # On Windows, use the GlobalMemoryStatusEx kernel function directly.
3344 3358 from ctypes import c_long as DWORD, c_ulonglong as DWORDLONG
3345 3359 from ctypes.wintypes import ( # pytype: disable=import-error
3346 3360 Structure,
3347 3361 byref,
3348 3362 sizeof,
3349 3363 windll,
3350 3364 )
3351 3365
3352 3366 class MEMORYSTATUSEX(Structure):
3353 3367 _fields_ = [
3354 3368 ('dwLength', DWORD),
3355 3369 ('dwMemoryLoad', DWORD),
3356 3370 ('ullTotalPhys', DWORDLONG),
3357 3371 ('ullAvailPhys', DWORDLONG),
3358 3372 ('ullTotalPageFile', DWORDLONG),
3359 3373 ('ullAvailPageFile', DWORDLONG),
3360 3374 ('ullTotalVirtual', DWORDLONG),
3361 3375 ('ullAvailVirtual', DWORDLONG),
3362 3376 ('ullExtendedVirtual', DWORDLONG),
3363 3377 ]
3364 3378
3365 3379 x = MEMORYSTATUSEX()
3366 3380 x.dwLength = sizeof(x)
3367 3381 windll.kernel32.GlobalMemoryStatusEx(byref(x))
3368 3382 return x.ullAvailPhys
3369 3383
3370 3384 # On newer Unix-like systems and Mac OSX, the sysconf interface
3371 3385 # can be used. _SC_PAGE_SIZE is part of POSIX; _SC_PHYS_PAGES
3372 3386 # seems to be implemented on most systems.
3373 3387 try:
3374 3388 pagesize = os.sysconf(os.sysconf_names['SC_PAGE_SIZE'])
3375 3389 pages = os.sysconf(os.sysconf_names['SC_PHYS_PAGES'])
3376 3390 return pagesize * pages
3377 3391 except OSError: # sysconf can fail
3378 3392 pass
3379 3393 except KeyError: # unknown parameter
3380 3394 pass
General Comments 0
You need to be logged in to leave comments. Login now