##// END OF EJS Templates
util: enable hardlink for copyfile...
Jun Wu -
r31577:e7a02e9a default
parent child Browse files
Show More
@@ -1,3605 +1,3600
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import
17 17
18 18 import bz2
19 19 import calendar
20 20 import codecs
21 21 import collections
22 22 import datetime
23 23 import errno
24 24 import gc
25 25 import hashlib
26 26 import imp
27 27 import os
28 28 import platform as pyplatform
29 29 import re as remod
30 30 import shutil
31 31 import signal
32 32 import socket
33 33 import stat
34 34 import string
35 35 import subprocess
36 36 import sys
37 37 import tempfile
38 38 import textwrap
39 39 import time
40 40 import traceback
41 41 import zlib
42 42
43 43 from . import (
44 44 encoding,
45 45 error,
46 46 i18n,
47 47 osutil,
48 48 parsers,
49 49 pycompat,
50 50 )
51 51
52 52 empty = pycompat.empty
53 53 httplib = pycompat.httplib
54 54 httpserver = pycompat.httpserver
55 55 pickle = pycompat.pickle
56 56 queue = pycompat.queue
57 57 socketserver = pycompat.socketserver
58 58 stderr = pycompat.stderr
59 59 stdin = pycompat.stdin
60 60 stdout = pycompat.stdout
61 61 stringio = pycompat.stringio
62 62 urlerr = pycompat.urlerr
63 63 urlreq = pycompat.urlreq
64 64 xmlrpclib = pycompat.xmlrpclib
65 65
66 66 def isatty(fp):
67 67 try:
68 68 return fp.isatty()
69 69 except AttributeError:
70 70 return False
71 71
72 72 # glibc determines buffering on first write to stdout - if we replace a TTY
73 73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
74 74 # buffering
75 75 if isatty(stdout):
76 76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
77 77
78 78 if pycompat.osname == 'nt':
79 79 from . import windows as platform
80 80 stdout = platform.winstdout(stdout)
81 81 else:
82 82 from . import posix as platform
83 83
84 84 _ = i18n._
85 85
86 86 bindunixsocket = platform.bindunixsocket
87 87 cachestat = platform.cachestat
88 88 checkexec = platform.checkexec
89 89 checklink = platform.checklink
90 90 copymode = platform.copymode
91 91 executablepath = platform.executablepath
92 92 expandglobs = platform.expandglobs
93 93 explainexit = platform.explainexit
94 94 findexe = platform.findexe
95 95 gethgcmd = platform.gethgcmd
96 96 getuser = platform.getuser
97 97 getpid = os.getpid
98 98 groupmembers = platform.groupmembers
99 99 groupname = platform.groupname
100 100 hidewindow = platform.hidewindow
101 101 isexec = platform.isexec
102 102 isowner = platform.isowner
103 103 localpath = platform.localpath
104 104 lookupreg = platform.lookupreg
105 105 makedir = platform.makedir
106 106 nlinks = platform.nlinks
107 107 normpath = platform.normpath
108 108 normcase = platform.normcase
109 109 normcasespec = platform.normcasespec
110 110 normcasefallback = platform.normcasefallback
111 111 openhardlinks = platform.openhardlinks
112 112 oslink = platform.oslink
113 113 parsepatchoutput = platform.parsepatchoutput
114 114 pconvert = platform.pconvert
115 115 poll = platform.poll
116 116 popen = platform.popen
117 117 posixfile = platform.posixfile
118 118 quotecommand = platform.quotecommand
119 119 readpipe = platform.readpipe
120 120 rename = platform.rename
121 121 removedirs = platform.removedirs
122 122 samedevice = platform.samedevice
123 123 samefile = platform.samefile
124 124 samestat = platform.samestat
125 125 setbinary = platform.setbinary
126 126 setflags = platform.setflags
127 127 setsignalhandler = platform.setsignalhandler
128 128 shellquote = platform.shellquote
129 129 spawndetached = platform.spawndetached
130 130 split = platform.split
131 131 sshargs = platform.sshargs
132 132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
133 133 statisexec = platform.statisexec
134 134 statislink = platform.statislink
135 135 testpid = platform.testpid
136 136 umask = platform.umask
137 137 unlink = platform.unlink
138 138 username = platform.username
139 139
140 140 # Python compatibility
141 141
142 142 _notset = object()
143 143
144 144 # disable Python's problematic floating point timestamps (issue4836)
145 145 # (Python hypocritically says you shouldn't change this behavior in
146 146 # libraries, and sure enough Mercurial is not a library.)
147 147 os.stat_float_times(False)
148 148
149 149 def safehasattr(thing, attr):
150 150 return getattr(thing, attr, _notset) is not _notset
151 151
152 152 def bitsfrom(container):
153 153 bits = 0
154 154 for bit in container:
155 155 bits |= bit
156 156 return bits
157 157
158 158 DIGESTS = {
159 159 'md5': hashlib.md5,
160 160 'sha1': hashlib.sha1,
161 161 'sha512': hashlib.sha512,
162 162 }
163 163 # List of digest types from strongest to weakest
164 164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
165 165
166 166 for k in DIGESTS_BY_STRENGTH:
167 167 assert k in DIGESTS
168 168
169 169 class digester(object):
170 170 """helper to compute digests.
171 171
172 172 This helper can be used to compute one or more digests given their name.
173 173
174 174 >>> d = digester(['md5', 'sha1'])
175 175 >>> d.update('foo')
176 176 >>> [k for k in sorted(d)]
177 177 ['md5', 'sha1']
178 178 >>> d['md5']
179 179 'acbd18db4cc2f85cedef654fccc4a4d8'
180 180 >>> d['sha1']
181 181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
182 182 >>> digester.preferred(['md5', 'sha1'])
183 183 'sha1'
184 184 """
185 185
186 186 def __init__(self, digests, s=''):
187 187 self._hashes = {}
188 188 for k in digests:
189 189 if k not in DIGESTS:
190 190 raise Abort(_('unknown digest type: %s') % k)
191 191 self._hashes[k] = DIGESTS[k]()
192 192 if s:
193 193 self.update(s)
194 194
195 195 def update(self, data):
196 196 for h in self._hashes.values():
197 197 h.update(data)
198 198
199 199 def __getitem__(self, key):
200 200 if key not in DIGESTS:
201 201 raise Abort(_('unknown digest type: %s') % k)
202 202 return self._hashes[key].hexdigest()
203 203
204 204 def __iter__(self):
205 205 return iter(self._hashes)
206 206
207 207 @staticmethod
208 208 def preferred(supported):
209 209 """returns the strongest digest type in both supported and DIGESTS."""
210 210
211 211 for k in DIGESTS_BY_STRENGTH:
212 212 if k in supported:
213 213 return k
214 214 return None
215 215
216 216 class digestchecker(object):
217 217 """file handle wrapper that additionally checks content against a given
218 218 size and digests.
219 219
220 220 d = digestchecker(fh, size, {'md5': '...'})
221 221
222 222 When multiple digests are given, all of them are validated.
223 223 """
224 224
225 225 def __init__(self, fh, size, digests):
226 226 self._fh = fh
227 227 self._size = size
228 228 self._got = 0
229 229 self._digests = dict(digests)
230 230 self._digester = digester(self._digests.keys())
231 231
232 232 def read(self, length=-1):
233 233 content = self._fh.read(length)
234 234 self._digester.update(content)
235 235 self._got += len(content)
236 236 return content
237 237
238 238 def validate(self):
239 239 if self._size != self._got:
240 240 raise Abort(_('size mismatch: expected %d, got %d') %
241 241 (self._size, self._got))
242 242 for k, v in self._digests.items():
243 243 if v != self._digester[k]:
244 244 # i18n: first parameter is a digest name
245 245 raise Abort(_('%s mismatch: expected %s, got %s') %
246 246 (k, v, self._digester[k]))
247 247
248 248 try:
249 249 buffer = buffer
250 250 except NameError:
251 251 if not pycompat.ispy3:
252 252 def buffer(sliceable, offset=0, length=None):
253 253 if length is not None:
254 254 return sliceable[offset:offset + length]
255 255 return sliceable[offset:]
256 256 else:
257 257 def buffer(sliceable, offset=0, length=None):
258 258 if length is not None:
259 259 return memoryview(sliceable)[offset:offset + length]
260 260 return memoryview(sliceable)[offset:]
261 261
262 262 closefds = pycompat.osname == 'posix'
263 263
264 264 _chunksize = 4096
265 265
266 266 class bufferedinputpipe(object):
267 267 """a manually buffered input pipe
268 268
269 269 Python will not let us use buffered IO and lazy reading with 'polling' at
270 270 the same time. We cannot probe the buffer state and select will not detect
271 271 that data are ready to read if they are already buffered.
272 272
273 273 This class let us work around that by implementing its own buffering
274 274 (allowing efficient readline) while offering a way to know if the buffer is
275 275 empty from the output (allowing collaboration of the buffer with polling).
276 276
277 277 This class lives in the 'util' module because it makes use of the 'os'
278 278 module from the python stdlib.
279 279 """
280 280
281 281 def __init__(self, input):
282 282 self._input = input
283 283 self._buffer = []
284 284 self._eof = False
285 285 self._lenbuf = 0
286 286
287 287 @property
288 288 def hasbuffer(self):
289 289 """True is any data is currently buffered
290 290
291 291 This will be used externally a pre-step for polling IO. If there is
292 292 already data then no polling should be set in place."""
293 293 return bool(self._buffer)
294 294
295 295 @property
296 296 def closed(self):
297 297 return self._input.closed
298 298
299 299 def fileno(self):
300 300 return self._input.fileno()
301 301
302 302 def close(self):
303 303 return self._input.close()
304 304
305 305 def read(self, size):
306 306 while (not self._eof) and (self._lenbuf < size):
307 307 self._fillbuffer()
308 308 return self._frombuffer(size)
309 309
310 310 def readline(self, *args, **kwargs):
311 311 if 1 < len(self._buffer):
312 312 # this should not happen because both read and readline end with a
313 313 # _frombuffer call that collapse it.
314 314 self._buffer = [''.join(self._buffer)]
315 315 self._lenbuf = len(self._buffer[0])
316 316 lfi = -1
317 317 if self._buffer:
318 318 lfi = self._buffer[-1].find('\n')
319 319 while (not self._eof) and lfi < 0:
320 320 self._fillbuffer()
321 321 if self._buffer:
322 322 lfi = self._buffer[-1].find('\n')
323 323 size = lfi + 1
324 324 if lfi < 0: # end of file
325 325 size = self._lenbuf
326 326 elif 1 < len(self._buffer):
327 327 # we need to take previous chunks into account
328 328 size += self._lenbuf - len(self._buffer[-1])
329 329 return self._frombuffer(size)
330 330
331 331 def _frombuffer(self, size):
332 332 """return at most 'size' data from the buffer
333 333
334 334 The data are removed from the buffer."""
335 335 if size == 0 or not self._buffer:
336 336 return ''
337 337 buf = self._buffer[0]
338 338 if 1 < len(self._buffer):
339 339 buf = ''.join(self._buffer)
340 340
341 341 data = buf[:size]
342 342 buf = buf[len(data):]
343 343 if buf:
344 344 self._buffer = [buf]
345 345 self._lenbuf = len(buf)
346 346 else:
347 347 self._buffer = []
348 348 self._lenbuf = 0
349 349 return data
350 350
351 351 def _fillbuffer(self):
352 352 """read data to the buffer"""
353 353 data = os.read(self._input.fileno(), _chunksize)
354 354 if not data:
355 355 self._eof = True
356 356 else:
357 357 self._lenbuf += len(data)
358 358 self._buffer.append(data)
359 359
360 360 def popen2(cmd, env=None, newlines=False):
361 361 # Setting bufsize to -1 lets the system decide the buffer size.
362 362 # The default for bufsize is 0, meaning unbuffered. This leads to
363 363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
364 364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
365 365 close_fds=closefds,
366 366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
367 367 universal_newlines=newlines,
368 368 env=env)
369 369 return p.stdin, p.stdout
370 370
371 371 def popen3(cmd, env=None, newlines=False):
372 372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
373 373 return stdin, stdout, stderr
374 374
375 375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
376 376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
377 377 close_fds=closefds,
378 378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
379 379 stderr=subprocess.PIPE,
380 380 universal_newlines=newlines,
381 381 env=env)
382 382 return p.stdin, p.stdout, p.stderr, p
383 383
384 384 def version():
385 385 """Return version information if available."""
386 386 try:
387 387 from . import __version__
388 388 return __version__.version
389 389 except ImportError:
390 390 return 'unknown'
391 391
392 392 def versiontuple(v=None, n=4):
393 393 """Parses a Mercurial version string into an N-tuple.
394 394
395 395 The version string to be parsed is specified with the ``v`` argument.
396 396 If it isn't defined, the current Mercurial version string will be parsed.
397 397
398 398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
399 399 returned values:
400 400
401 401 >>> v = '3.6.1+190-df9b73d2d444'
402 402 >>> versiontuple(v, 2)
403 403 (3, 6)
404 404 >>> versiontuple(v, 3)
405 405 (3, 6, 1)
406 406 >>> versiontuple(v, 4)
407 407 (3, 6, 1, '190-df9b73d2d444')
408 408
409 409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
410 410 (3, 6, 1, '190-df9b73d2d444+20151118')
411 411
412 412 >>> v = '3.6'
413 413 >>> versiontuple(v, 2)
414 414 (3, 6)
415 415 >>> versiontuple(v, 3)
416 416 (3, 6, None)
417 417 >>> versiontuple(v, 4)
418 418 (3, 6, None, None)
419 419
420 420 >>> v = '3.9-rc'
421 421 >>> versiontuple(v, 2)
422 422 (3, 9)
423 423 >>> versiontuple(v, 3)
424 424 (3, 9, None)
425 425 >>> versiontuple(v, 4)
426 426 (3, 9, None, 'rc')
427 427
428 428 >>> v = '3.9-rc+2-02a8fea4289b'
429 429 >>> versiontuple(v, 2)
430 430 (3, 9)
431 431 >>> versiontuple(v, 3)
432 432 (3, 9, None)
433 433 >>> versiontuple(v, 4)
434 434 (3, 9, None, 'rc+2-02a8fea4289b')
435 435 """
436 436 if not v:
437 437 v = version()
438 438 parts = remod.split('[\+-]', v, 1)
439 439 if len(parts) == 1:
440 440 vparts, extra = parts[0], None
441 441 else:
442 442 vparts, extra = parts
443 443
444 444 vints = []
445 445 for i in vparts.split('.'):
446 446 try:
447 447 vints.append(int(i))
448 448 except ValueError:
449 449 break
450 450 # (3, 6) -> (3, 6, None)
451 451 while len(vints) < 3:
452 452 vints.append(None)
453 453
454 454 if n == 2:
455 455 return (vints[0], vints[1])
456 456 if n == 3:
457 457 return (vints[0], vints[1], vints[2])
458 458 if n == 4:
459 459 return (vints[0], vints[1], vints[2], extra)
460 460
461 461 # used by parsedate
462 462 defaultdateformats = (
463 463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
464 464 '%Y-%m-%dT%H:%M', # without seconds
465 465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
466 466 '%Y-%m-%dT%H%M', # without seconds
467 467 '%Y-%m-%d %H:%M:%S', # our common legal variant
468 468 '%Y-%m-%d %H:%M', # without seconds
469 469 '%Y-%m-%d %H%M%S', # without :
470 470 '%Y-%m-%d %H%M', # without seconds
471 471 '%Y-%m-%d %I:%M:%S%p',
472 472 '%Y-%m-%d %H:%M',
473 473 '%Y-%m-%d %I:%M%p',
474 474 '%Y-%m-%d',
475 475 '%m-%d',
476 476 '%m/%d',
477 477 '%m/%d/%y',
478 478 '%m/%d/%Y',
479 479 '%a %b %d %H:%M:%S %Y',
480 480 '%a %b %d %I:%M:%S%p %Y',
481 481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
482 482 '%b %d %H:%M:%S %Y',
483 483 '%b %d %I:%M:%S%p %Y',
484 484 '%b %d %H:%M:%S',
485 485 '%b %d %I:%M:%S%p',
486 486 '%b %d %H:%M',
487 487 '%b %d %I:%M%p',
488 488 '%b %d %Y',
489 489 '%b %d',
490 490 '%H:%M:%S',
491 491 '%I:%M:%S%p',
492 492 '%H:%M',
493 493 '%I:%M%p',
494 494 )
495 495
496 496 extendeddateformats = defaultdateformats + (
497 497 "%Y",
498 498 "%Y-%m",
499 499 "%b",
500 500 "%b %Y",
501 501 )
502 502
503 503 def cachefunc(func):
504 504 '''cache the result of function calls'''
505 505 # XXX doesn't handle keywords args
506 506 if func.__code__.co_argcount == 0:
507 507 cache = []
508 508 def f():
509 509 if len(cache) == 0:
510 510 cache.append(func())
511 511 return cache[0]
512 512 return f
513 513 cache = {}
514 514 if func.__code__.co_argcount == 1:
515 515 # we gain a small amount of time because
516 516 # we don't need to pack/unpack the list
517 517 def f(arg):
518 518 if arg not in cache:
519 519 cache[arg] = func(arg)
520 520 return cache[arg]
521 521 else:
522 522 def f(*args):
523 523 if args not in cache:
524 524 cache[args] = func(*args)
525 525 return cache[args]
526 526
527 527 return f
528 528
529 529 class sortdict(dict):
530 530 '''a simple sorted dictionary'''
531 531 def __init__(self, data=None):
532 532 self._list = []
533 533 if data:
534 534 self.update(data)
535 535 def copy(self):
536 536 return sortdict(self)
537 537 def __setitem__(self, key, val):
538 538 if key in self:
539 539 self._list.remove(key)
540 540 self._list.append(key)
541 541 dict.__setitem__(self, key, val)
542 542 def __iter__(self):
543 543 return self._list.__iter__()
544 544 def update(self, src):
545 545 if isinstance(src, dict):
546 546 src = src.iteritems()
547 547 for k, v in src:
548 548 self[k] = v
549 549 def clear(self):
550 550 dict.clear(self)
551 551 self._list = []
552 552 def items(self):
553 553 return [(k, self[k]) for k in self._list]
554 554 def __delitem__(self, key):
555 555 dict.__delitem__(self, key)
556 556 self._list.remove(key)
557 557 def pop(self, key, *args, **kwargs):
558 558 dict.pop(self, key, *args, **kwargs)
559 559 try:
560 560 self._list.remove(key)
561 561 except ValueError:
562 562 pass
563 563 def keys(self):
564 564 return self._list[:]
565 565 def iterkeys(self):
566 566 return self._list.__iter__()
567 567 def iteritems(self):
568 568 for k in self._list:
569 569 yield k, self[k]
570 570 def insert(self, index, key, val):
571 571 self._list.insert(index, key)
572 572 dict.__setitem__(self, key, val)
573 573 def __repr__(self):
574 574 if not self:
575 575 return '%s()' % self.__class__.__name__
576 576 return '%s(%r)' % (self.__class__.__name__, self.items())
577 577
578 578 class _lrucachenode(object):
579 579 """A node in a doubly linked list.
580 580
581 581 Holds a reference to nodes on either side as well as a key-value
582 582 pair for the dictionary entry.
583 583 """
584 584 __slots__ = (u'next', u'prev', u'key', u'value')
585 585
586 586 def __init__(self):
587 587 self.next = None
588 588 self.prev = None
589 589
590 590 self.key = _notset
591 591 self.value = None
592 592
593 593 def markempty(self):
594 594 """Mark the node as emptied."""
595 595 self.key = _notset
596 596
597 597 class lrucachedict(object):
598 598 """Dict that caches most recent accesses and sets.
599 599
600 600 The dict consists of an actual backing dict - indexed by original
601 601 key - and a doubly linked circular list defining the order of entries in
602 602 the cache.
603 603
604 604 The head node is the newest entry in the cache. If the cache is full,
605 605 we recycle head.prev and make it the new head. Cache accesses result in
606 606 the node being moved to before the existing head and being marked as the
607 607 new head node.
608 608 """
609 609 def __init__(self, max):
610 610 self._cache = {}
611 611
612 612 self._head = head = _lrucachenode()
613 613 head.prev = head
614 614 head.next = head
615 615 self._size = 1
616 616 self._capacity = max
617 617
618 618 def __len__(self):
619 619 return len(self._cache)
620 620
621 621 def __contains__(self, k):
622 622 return k in self._cache
623 623
624 624 def __iter__(self):
625 625 # We don't have to iterate in cache order, but why not.
626 626 n = self._head
627 627 for i in range(len(self._cache)):
628 628 yield n.key
629 629 n = n.next
630 630
631 631 def __getitem__(self, k):
632 632 node = self._cache[k]
633 633 self._movetohead(node)
634 634 return node.value
635 635
636 636 def __setitem__(self, k, v):
637 637 node = self._cache.get(k)
638 638 # Replace existing value and mark as newest.
639 639 if node is not None:
640 640 node.value = v
641 641 self._movetohead(node)
642 642 return
643 643
644 644 if self._size < self._capacity:
645 645 node = self._addcapacity()
646 646 else:
647 647 # Grab the last/oldest item.
648 648 node = self._head.prev
649 649
650 650 # At capacity. Kill the old entry.
651 651 if node.key is not _notset:
652 652 del self._cache[node.key]
653 653
654 654 node.key = k
655 655 node.value = v
656 656 self._cache[k] = node
657 657 # And mark it as newest entry. No need to adjust order since it
658 658 # is already self._head.prev.
659 659 self._head = node
660 660
661 661 def __delitem__(self, k):
662 662 node = self._cache.pop(k)
663 663 node.markempty()
664 664
665 665 # Temporarily mark as newest item before re-adjusting head to make
666 666 # this node the oldest item.
667 667 self._movetohead(node)
668 668 self._head = node.next
669 669
670 670 # Additional dict methods.
671 671
672 672 def get(self, k, default=None):
673 673 try:
674 674 return self._cache[k].value
675 675 except KeyError:
676 676 return default
677 677
678 678 def clear(self):
679 679 n = self._head
680 680 while n.key is not _notset:
681 681 n.markempty()
682 682 n = n.next
683 683
684 684 self._cache.clear()
685 685
686 686 def copy(self):
687 687 result = lrucachedict(self._capacity)
688 688 n = self._head.prev
689 689 # Iterate in oldest-to-newest order, so the copy has the right ordering
690 690 for i in range(len(self._cache)):
691 691 result[n.key] = n.value
692 692 n = n.prev
693 693 return result
694 694
695 695 def _movetohead(self, node):
696 696 """Mark a node as the newest, making it the new head.
697 697
698 698 When a node is accessed, it becomes the freshest entry in the LRU
699 699 list, which is denoted by self._head.
700 700
701 701 Visually, let's make ``N`` the new head node (* denotes head):
702 702
703 703 previous/oldest <-> head <-> next/next newest
704 704
705 705 ----<->--- A* ---<->-----
706 706 | |
707 707 E <-> D <-> N <-> C <-> B
708 708
709 709 To:
710 710
711 711 ----<->--- N* ---<->-----
712 712 | |
713 713 E <-> D <-> C <-> B <-> A
714 714
715 715 This requires the following moves:
716 716
717 717 C.next = D (node.prev.next = node.next)
718 718 D.prev = C (node.next.prev = node.prev)
719 719 E.next = N (head.prev.next = node)
720 720 N.prev = E (node.prev = head.prev)
721 721 N.next = A (node.next = head)
722 722 A.prev = N (head.prev = node)
723 723 """
724 724 head = self._head
725 725 # C.next = D
726 726 node.prev.next = node.next
727 727 # D.prev = C
728 728 node.next.prev = node.prev
729 729 # N.prev = E
730 730 node.prev = head.prev
731 731 # N.next = A
732 732 # It is tempting to do just "head" here, however if node is
733 733 # adjacent to head, this will do bad things.
734 734 node.next = head.prev.next
735 735 # E.next = N
736 736 node.next.prev = node
737 737 # A.prev = N
738 738 node.prev.next = node
739 739
740 740 self._head = node
741 741
742 742 def _addcapacity(self):
743 743 """Add a node to the circular linked list.
744 744
745 745 The new node is inserted before the head node.
746 746 """
747 747 head = self._head
748 748 node = _lrucachenode()
749 749 head.prev.next = node
750 750 node.prev = head.prev
751 751 node.next = head
752 752 head.prev = node
753 753 self._size += 1
754 754 return node
755 755
756 756 def lrucachefunc(func):
757 757 '''cache most recent results of function calls'''
758 758 cache = {}
759 759 order = collections.deque()
760 760 if func.__code__.co_argcount == 1:
761 761 def f(arg):
762 762 if arg not in cache:
763 763 if len(cache) > 20:
764 764 del cache[order.popleft()]
765 765 cache[arg] = func(arg)
766 766 else:
767 767 order.remove(arg)
768 768 order.append(arg)
769 769 return cache[arg]
770 770 else:
771 771 def f(*args):
772 772 if args not in cache:
773 773 if len(cache) > 20:
774 774 del cache[order.popleft()]
775 775 cache[args] = func(*args)
776 776 else:
777 777 order.remove(args)
778 778 order.append(args)
779 779 return cache[args]
780 780
781 781 return f
782 782
783 783 class propertycache(object):
784 784 def __init__(self, func):
785 785 self.func = func
786 786 self.name = func.__name__
787 787 def __get__(self, obj, type=None):
788 788 result = self.func(obj)
789 789 self.cachevalue(obj, result)
790 790 return result
791 791
792 792 def cachevalue(self, obj, value):
793 793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
794 794 obj.__dict__[self.name] = value
795 795
796 796 def pipefilter(s, cmd):
797 797 '''filter string S through command CMD, returning its output'''
798 798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
799 799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
800 800 pout, perr = p.communicate(s)
801 801 return pout
802 802
803 803 def tempfilter(s, cmd):
804 804 '''filter string S through a pair of temporary files with CMD.
805 805 CMD is used as a template to create the real command to be run,
806 806 with the strings INFILE and OUTFILE replaced by the real names of
807 807 the temporary files generated.'''
808 808 inname, outname = None, None
809 809 try:
810 810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
811 811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
812 812 fp.write(s)
813 813 fp.close()
814 814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
815 815 os.close(outfd)
816 816 cmd = cmd.replace('INFILE', inname)
817 817 cmd = cmd.replace('OUTFILE', outname)
818 818 code = os.system(cmd)
819 819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
820 820 code = 0
821 821 if code:
822 822 raise Abort(_("command '%s' failed: %s") %
823 823 (cmd, explainexit(code)))
824 824 return readfile(outname)
825 825 finally:
826 826 try:
827 827 if inname:
828 828 os.unlink(inname)
829 829 except OSError:
830 830 pass
831 831 try:
832 832 if outname:
833 833 os.unlink(outname)
834 834 except OSError:
835 835 pass
836 836
837 837 filtertable = {
838 838 'tempfile:': tempfilter,
839 839 'pipe:': pipefilter,
840 840 }
841 841
842 842 def filter(s, cmd):
843 843 "filter a string through a command that transforms its input to its output"
844 844 for name, fn in filtertable.iteritems():
845 845 if cmd.startswith(name):
846 846 return fn(s, cmd[len(name):].lstrip())
847 847 return pipefilter(s, cmd)
848 848
849 849 def binary(s):
850 850 """return true if a string is binary data"""
851 851 return bool(s and '\0' in s)
852 852
853 853 def increasingchunks(source, min=1024, max=65536):
854 854 '''return no less than min bytes per chunk while data remains,
855 855 doubling min after each chunk until it reaches max'''
856 856 def log2(x):
857 857 if not x:
858 858 return 0
859 859 i = 0
860 860 while x:
861 861 x >>= 1
862 862 i += 1
863 863 return i - 1
864 864
865 865 buf = []
866 866 blen = 0
867 867 for chunk in source:
868 868 buf.append(chunk)
869 869 blen += len(chunk)
870 870 if blen >= min:
871 871 if min < max:
872 872 min = min << 1
873 873 nmin = 1 << log2(blen)
874 874 if nmin > min:
875 875 min = nmin
876 876 if min > max:
877 877 min = max
878 878 yield ''.join(buf)
879 879 blen = 0
880 880 buf = []
881 881 if buf:
882 882 yield ''.join(buf)
883 883
884 884 Abort = error.Abort
885 885
886 886 def always(fn):
887 887 return True
888 888
889 889 def never(fn):
890 890 return False
891 891
892 892 def nogc(func):
893 893 """disable garbage collector
894 894
895 895 Python's garbage collector triggers a GC each time a certain number of
896 896 container objects (the number being defined by gc.get_threshold()) are
897 897 allocated even when marked not to be tracked by the collector. Tracking has
898 898 no effect on when GCs are triggered, only on what objects the GC looks
899 899 into. As a workaround, disable GC while building complex (huge)
900 900 containers.
901 901
902 902 This garbage collector issue have been fixed in 2.7.
903 903 """
904 904 if sys.version_info >= (2, 7):
905 905 return func
906 906 def wrapper(*args, **kwargs):
907 907 gcenabled = gc.isenabled()
908 908 gc.disable()
909 909 try:
910 910 return func(*args, **kwargs)
911 911 finally:
912 912 if gcenabled:
913 913 gc.enable()
914 914 return wrapper
915 915
916 916 def pathto(root, n1, n2):
917 917 '''return the relative path from one place to another.
918 918 root should use os.sep to separate directories
919 919 n1 should use os.sep to separate directories
920 920 n2 should use "/" to separate directories
921 921 returns an os.sep-separated path.
922 922
923 923 If n1 is a relative path, it's assumed it's
924 924 relative to root.
925 925 n2 should always be relative to root.
926 926 '''
927 927 if not n1:
928 928 return localpath(n2)
929 929 if os.path.isabs(n1):
930 930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
931 931 return os.path.join(root, localpath(n2))
932 932 n2 = '/'.join((pconvert(root), n2))
933 933 a, b = splitpath(n1), n2.split('/')
934 934 a.reverse()
935 935 b.reverse()
936 936 while a and b and a[-1] == b[-1]:
937 937 a.pop()
938 938 b.pop()
939 939 b.reverse()
940 940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
941 941
942 942 def mainfrozen():
943 943 """return True if we are a frozen executable.
944 944
945 945 The code supports py2exe (most common, Windows only) and tools/freeze
946 946 (portable, not much used).
947 947 """
948 948 return (safehasattr(sys, "frozen") or # new py2exe
949 949 safehasattr(sys, "importers") or # old py2exe
950 950 imp.is_frozen(u"__main__")) # tools/freeze
951 951
952 952 # the location of data files matching the source code
953 953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
954 954 # executable version (py2exe) doesn't support __file__
955 955 datapath = os.path.dirname(pycompat.sysexecutable)
956 956 else:
957 957 datapath = os.path.dirname(pycompat.fsencode(__file__))
958 958
959 959 i18n.setdatapath(datapath)
960 960
961 961 _hgexecutable = None
962 962
963 963 def hgexecutable():
964 964 """return location of the 'hg' executable.
965 965
966 966 Defaults to $HG or 'hg' in the search path.
967 967 """
968 968 if _hgexecutable is None:
969 969 hg = encoding.environ.get('HG')
970 970 mainmod = sys.modules[pycompat.sysstr('__main__')]
971 971 if hg:
972 972 _sethgexecutable(hg)
973 973 elif mainfrozen():
974 974 if getattr(sys, 'frozen', None) == 'macosx_app':
975 975 # Env variable set by py2app
976 976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
977 977 else:
978 978 _sethgexecutable(pycompat.sysexecutable)
979 979 elif (os.path.basename(
980 980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
981 981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
982 982 else:
983 983 exe = findexe('hg') or os.path.basename(sys.argv[0])
984 984 _sethgexecutable(exe)
985 985 return _hgexecutable
986 986
987 987 def _sethgexecutable(path):
988 988 """set location of the 'hg' executable"""
989 989 global _hgexecutable
990 990 _hgexecutable = path
991 991
992 992 def _isstdout(f):
993 993 fileno = getattr(f, 'fileno', None)
994 994 return fileno and fileno() == sys.__stdout__.fileno()
995 995
996 996 def shellenviron(environ=None):
997 997 """return environ with optional override, useful for shelling out"""
998 998 def py2shell(val):
999 999 'convert python object into string that is useful to shell'
1000 1000 if val is None or val is False:
1001 1001 return '0'
1002 1002 if val is True:
1003 1003 return '1'
1004 1004 return str(val)
1005 1005 env = dict(encoding.environ)
1006 1006 if environ:
1007 1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1008 1008 env['HG'] = hgexecutable()
1009 1009 return env
1010 1010
1011 1011 def system(cmd, environ=None, cwd=None, out=None):
1012 1012 '''enhanced shell command execution.
1013 1013 run with environment maybe modified, maybe in different dir.
1014 1014
1015 1015 if out is specified, it is assumed to be a file-like object that has a
1016 1016 write() method. stdout and stderr will be redirected to out.'''
1017 1017 try:
1018 1018 stdout.flush()
1019 1019 except Exception:
1020 1020 pass
1021 1021 cmd = quotecommand(cmd)
1022 1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1023 1023 and sys.version_info[1] < 7):
1024 1024 # subprocess kludge to work around issues in half-baked Python
1025 1025 # ports, notably bichued/python:
1026 1026 if not cwd is None:
1027 1027 os.chdir(cwd)
1028 1028 rc = os.system(cmd)
1029 1029 else:
1030 1030 env = shellenviron(environ)
1031 1031 if out is None or _isstdout(out):
1032 1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1033 1033 env=env, cwd=cwd)
1034 1034 else:
1035 1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1036 1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1037 1037 stderr=subprocess.STDOUT)
1038 1038 for line in iter(proc.stdout.readline, ''):
1039 1039 out.write(line)
1040 1040 proc.wait()
1041 1041 rc = proc.returncode
1042 1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1043 1043 rc = 0
1044 1044 return rc
1045 1045
1046 1046 def checksignature(func):
1047 1047 '''wrap a function with code to check for calling errors'''
1048 1048 def check(*args, **kwargs):
1049 1049 try:
1050 1050 return func(*args, **kwargs)
1051 1051 except TypeError:
1052 1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1053 1053 raise error.SignatureError
1054 1054 raise
1055 1055
1056 1056 return check
1057 1057
1058 # Hardlinks are problematic on CIFS, do not allow hardlinks
1059 # until we find a way to work around it cleanly (issue4546).
1060 # This is a variable so extensions can opt-in to using them.
1061 allowhardlinks = False
1062
1063 1058 # a whilelist of known filesystems where hardlink works reliably
1064 1059 _hardlinkfswhitelist = set([
1065 1060 'btrfs',
1066 1061 'ext2',
1067 1062 'ext3',
1068 1063 'ext4',
1069 1064 'jfs',
1070 1065 'reiserfs',
1071 1066 'tmpfs',
1072 1067 'xfs',
1073 1068 ])
1074 1069
1075 1070 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1076 1071 '''copy a file, preserving mode and optionally other stat info like
1077 1072 atime/mtime
1078 1073
1079 1074 checkambig argument is used with filestat, and is useful only if
1080 1075 destination file is guarded by any lock (e.g. repo.lock or
1081 1076 repo.wlock).
1082 1077
1083 1078 copystat and checkambig should be exclusive.
1084 1079 '''
1085 1080 assert not (copystat and checkambig)
1086 1081 oldstat = None
1087 1082 if os.path.lexists(dest):
1088 1083 if checkambig:
1089 1084 oldstat = checkambig and filestat(dest)
1090 1085 unlink(dest)
1091 1086 if hardlink:
1092 1087 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1093 1088 # unless we are confident that dest is on a whitelisted filesystem.
1094 1089 destdir = os.path.dirname(dest)
1095 1090 fstype = getattr(osutil, 'getfstype', lambda x: None)(destdir)
1096 1091 if fstype not in _hardlinkfswhitelist:
1097 1092 hardlink = False
1098 if allowhardlinks and hardlink:
1093 if hardlink:
1099 1094 try:
1100 1095 oslink(src, dest)
1101 1096 return
1102 1097 except (IOError, OSError):
1103 1098 pass # fall back to normal copy
1104 1099 if os.path.islink(src):
1105 1100 os.symlink(os.readlink(src), dest)
1106 1101 # copytime is ignored for symlinks, but in general copytime isn't needed
1107 1102 # for them anyway
1108 1103 else:
1109 1104 try:
1110 1105 shutil.copyfile(src, dest)
1111 1106 if copystat:
1112 1107 # copystat also copies mode
1113 1108 shutil.copystat(src, dest)
1114 1109 else:
1115 1110 shutil.copymode(src, dest)
1116 1111 if oldstat and oldstat.stat:
1117 1112 newstat = filestat(dest)
1118 1113 if newstat.isambig(oldstat):
1119 1114 # stat of copied file is ambiguous to original one
1120 1115 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1121 1116 os.utime(dest, (advanced, advanced))
1122 1117 except shutil.Error as inst:
1123 1118 raise Abort(str(inst))
1124 1119
1125 1120 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1126 1121 """Copy a directory tree using hardlinks if possible."""
1127 1122 num = 0
1128 1123
1129 1124 if hardlink is None:
1130 1125 hardlink = (os.stat(src).st_dev ==
1131 1126 os.stat(os.path.dirname(dst)).st_dev)
1132 1127 if hardlink:
1133 1128 topic = _('linking')
1134 1129 else:
1135 1130 topic = _('copying')
1136 1131
1137 1132 if os.path.isdir(src):
1138 1133 os.mkdir(dst)
1139 1134 for name, kind in osutil.listdir(src):
1140 1135 srcname = os.path.join(src, name)
1141 1136 dstname = os.path.join(dst, name)
1142 1137 def nprog(t, pos):
1143 1138 if pos is not None:
1144 1139 return progress(t, pos + num)
1145 1140 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1146 1141 num += n
1147 1142 else:
1148 1143 if hardlink:
1149 1144 try:
1150 1145 oslink(src, dst)
1151 1146 except (IOError, OSError):
1152 1147 hardlink = False
1153 1148 shutil.copy(src, dst)
1154 1149 else:
1155 1150 shutil.copy(src, dst)
1156 1151 num += 1
1157 1152 progress(topic, num)
1158 1153 progress(topic, None)
1159 1154
1160 1155 return hardlink, num
1161 1156
1162 1157 _winreservednames = '''con prn aux nul
1163 1158 com1 com2 com3 com4 com5 com6 com7 com8 com9
1164 1159 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1165 1160 _winreservedchars = ':*?"<>|'
1166 1161 def checkwinfilename(path):
1167 1162 r'''Check that the base-relative path is a valid filename on Windows.
1168 1163 Returns None if the path is ok, or a UI string describing the problem.
1169 1164
1170 1165 >>> checkwinfilename("just/a/normal/path")
1171 1166 >>> checkwinfilename("foo/bar/con.xml")
1172 1167 "filename contains 'con', which is reserved on Windows"
1173 1168 >>> checkwinfilename("foo/con.xml/bar")
1174 1169 "filename contains 'con', which is reserved on Windows"
1175 1170 >>> checkwinfilename("foo/bar/xml.con")
1176 1171 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1177 1172 "filename contains 'AUX', which is reserved on Windows"
1178 1173 >>> checkwinfilename("foo/bar/bla:.txt")
1179 1174 "filename contains ':', which is reserved on Windows"
1180 1175 >>> checkwinfilename("foo/bar/b\07la.txt")
1181 1176 "filename contains '\\x07', which is invalid on Windows"
1182 1177 >>> checkwinfilename("foo/bar/bla ")
1183 1178 "filename ends with ' ', which is not allowed on Windows"
1184 1179 >>> checkwinfilename("../bar")
1185 1180 >>> checkwinfilename("foo\\")
1186 1181 "filename ends with '\\', which is invalid on Windows"
1187 1182 >>> checkwinfilename("foo\\/bar")
1188 1183 "directory name ends with '\\', which is invalid on Windows"
1189 1184 '''
1190 1185 if path.endswith('\\'):
1191 1186 return _("filename ends with '\\', which is invalid on Windows")
1192 1187 if '\\/' in path:
1193 1188 return _("directory name ends with '\\', which is invalid on Windows")
1194 1189 for n in path.replace('\\', '/').split('/'):
1195 1190 if not n:
1196 1191 continue
1197 1192 for c in pycompat.bytestr(n):
1198 1193 if c in _winreservedchars:
1199 1194 return _("filename contains '%s', which is reserved "
1200 1195 "on Windows") % c
1201 1196 if ord(c) <= 31:
1202 1197 return _("filename contains %r, which is invalid "
1203 1198 "on Windows") % c
1204 1199 base = n.split('.')[0]
1205 1200 if base and base.lower() in _winreservednames:
1206 1201 return _("filename contains '%s', which is reserved "
1207 1202 "on Windows") % base
1208 1203 t = n[-1]
1209 1204 if t in '. ' and n not in '..':
1210 1205 return _("filename ends with '%s', which is not allowed "
1211 1206 "on Windows") % t
1212 1207
1213 1208 if pycompat.osname == 'nt':
1214 1209 checkosfilename = checkwinfilename
1215 1210 timer = time.clock
1216 1211 else:
1217 1212 checkosfilename = platform.checkosfilename
1218 1213 timer = time.time
1219 1214
1220 1215 if safehasattr(time, "perf_counter"):
1221 1216 timer = time.perf_counter
1222 1217
1223 1218 def makelock(info, pathname):
1224 1219 try:
1225 1220 return os.symlink(info, pathname)
1226 1221 except OSError as why:
1227 1222 if why.errno == errno.EEXIST:
1228 1223 raise
1229 1224 except AttributeError: # no symlink in os
1230 1225 pass
1231 1226
1232 1227 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1233 1228 os.write(ld, info)
1234 1229 os.close(ld)
1235 1230
1236 1231 def readlock(pathname):
1237 1232 try:
1238 1233 return os.readlink(pathname)
1239 1234 except OSError as why:
1240 1235 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1241 1236 raise
1242 1237 except AttributeError: # no symlink in os
1243 1238 pass
1244 1239 fp = posixfile(pathname)
1245 1240 r = fp.read()
1246 1241 fp.close()
1247 1242 return r
1248 1243
1249 1244 def fstat(fp):
1250 1245 '''stat file object that may not have fileno method.'''
1251 1246 try:
1252 1247 return os.fstat(fp.fileno())
1253 1248 except AttributeError:
1254 1249 return os.stat(fp.name)
1255 1250
1256 1251 # File system features
1257 1252
1258 1253 def fscasesensitive(path):
1259 1254 """
1260 1255 Return true if the given path is on a case-sensitive filesystem
1261 1256
1262 1257 Requires a path (like /foo/.hg) ending with a foldable final
1263 1258 directory component.
1264 1259 """
1265 1260 s1 = os.lstat(path)
1266 1261 d, b = os.path.split(path)
1267 1262 b2 = b.upper()
1268 1263 if b == b2:
1269 1264 b2 = b.lower()
1270 1265 if b == b2:
1271 1266 return True # no evidence against case sensitivity
1272 1267 p2 = os.path.join(d, b2)
1273 1268 try:
1274 1269 s2 = os.lstat(p2)
1275 1270 if s2 == s1:
1276 1271 return False
1277 1272 return True
1278 1273 except OSError:
1279 1274 return True
1280 1275
1281 1276 try:
1282 1277 import re2
1283 1278 _re2 = None
1284 1279 except ImportError:
1285 1280 _re2 = False
1286 1281
1287 1282 class _re(object):
1288 1283 def _checkre2(self):
1289 1284 global _re2
1290 1285 try:
1291 1286 # check if match works, see issue3964
1292 1287 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1293 1288 except ImportError:
1294 1289 _re2 = False
1295 1290
1296 1291 def compile(self, pat, flags=0):
1297 1292 '''Compile a regular expression, using re2 if possible
1298 1293
1299 1294 For best performance, use only re2-compatible regexp features. The
1300 1295 only flags from the re module that are re2-compatible are
1301 1296 IGNORECASE and MULTILINE.'''
1302 1297 if _re2 is None:
1303 1298 self._checkre2()
1304 1299 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1305 1300 if flags & remod.IGNORECASE:
1306 1301 pat = '(?i)' + pat
1307 1302 if flags & remod.MULTILINE:
1308 1303 pat = '(?m)' + pat
1309 1304 try:
1310 1305 return re2.compile(pat)
1311 1306 except re2.error:
1312 1307 pass
1313 1308 return remod.compile(pat, flags)
1314 1309
1315 1310 @propertycache
1316 1311 def escape(self):
1317 1312 '''Return the version of escape corresponding to self.compile.
1318 1313
1319 1314 This is imperfect because whether re2 or re is used for a particular
1320 1315 function depends on the flags, etc, but it's the best we can do.
1321 1316 '''
1322 1317 global _re2
1323 1318 if _re2 is None:
1324 1319 self._checkre2()
1325 1320 if _re2:
1326 1321 return re2.escape
1327 1322 else:
1328 1323 return remod.escape
1329 1324
1330 1325 re = _re()
1331 1326
1332 1327 _fspathcache = {}
1333 1328 def fspath(name, root):
1334 1329 '''Get name in the case stored in the filesystem
1335 1330
1336 1331 The name should be relative to root, and be normcase-ed for efficiency.
1337 1332
1338 1333 Note that this function is unnecessary, and should not be
1339 1334 called, for case-sensitive filesystems (simply because it's expensive).
1340 1335
1341 1336 The root should be normcase-ed, too.
1342 1337 '''
1343 1338 def _makefspathcacheentry(dir):
1344 1339 return dict((normcase(n), n) for n in os.listdir(dir))
1345 1340
1346 1341 seps = pycompat.ossep
1347 1342 if pycompat.osaltsep:
1348 1343 seps = seps + pycompat.osaltsep
1349 1344 # Protect backslashes. This gets silly very quickly.
1350 1345 seps.replace('\\','\\\\')
1351 1346 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1352 1347 dir = os.path.normpath(root)
1353 1348 result = []
1354 1349 for part, sep in pattern.findall(name):
1355 1350 if sep:
1356 1351 result.append(sep)
1357 1352 continue
1358 1353
1359 1354 if dir not in _fspathcache:
1360 1355 _fspathcache[dir] = _makefspathcacheentry(dir)
1361 1356 contents = _fspathcache[dir]
1362 1357
1363 1358 found = contents.get(part)
1364 1359 if not found:
1365 1360 # retry "once per directory" per "dirstate.walk" which
1366 1361 # may take place for each patches of "hg qpush", for example
1367 1362 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1368 1363 found = contents.get(part)
1369 1364
1370 1365 result.append(found or part)
1371 1366 dir = os.path.join(dir, part)
1372 1367
1373 1368 return ''.join(result)
1374 1369
1375 1370 def checknlink(testfile):
1376 1371 '''check whether hardlink count reporting works properly'''
1377 1372
1378 1373 # testfile may be open, so we need a separate file for checking to
1379 1374 # work around issue2543 (or testfile may get lost on Samba shares)
1380 1375 f1 = testfile + ".hgtmp1"
1381 1376 if os.path.lexists(f1):
1382 1377 return False
1383 1378 try:
1384 1379 posixfile(f1, 'w').close()
1385 1380 except IOError:
1386 1381 try:
1387 1382 os.unlink(f1)
1388 1383 except OSError:
1389 1384 pass
1390 1385 return False
1391 1386
1392 1387 f2 = testfile + ".hgtmp2"
1393 1388 fd = None
1394 1389 try:
1395 1390 oslink(f1, f2)
1396 1391 # nlinks() may behave differently for files on Windows shares if
1397 1392 # the file is open.
1398 1393 fd = posixfile(f2)
1399 1394 return nlinks(f2) > 1
1400 1395 except OSError:
1401 1396 return False
1402 1397 finally:
1403 1398 if fd is not None:
1404 1399 fd.close()
1405 1400 for f in (f1, f2):
1406 1401 try:
1407 1402 os.unlink(f)
1408 1403 except OSError:
1409 1404 pass
1410 1405
1411 1406 def endswithsep(path):
1412 1407 '''Check path ends with os.sep or os.altsep.'''
1413 1408 return (path.endswith(pycompat.ossep)
1414 1409 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1415 1410
1416 1411 def splitpath(path):
1417 1412 '''Split path by os.sep.
1418 1413 Note that this function does not use os.altsep because this is
1419 1414 an alternative of simple "xxx.split(os.sep)".
1420 1415 It is recommended to use os.path.normpath() before using this
1421 1416 function if need.'''
1422 1417 return path.split(pycompat.ossep)
1423 1418
1424 1419 def gui():
1425 1420 '''Are we running in a GUI?'''
1426 1421 if pycompat.sysplatform == 'darwin':
1427 1422 if 'SSH_CONNECTION' in encoding.environ:
1428 1423 # handle SSH access to a box where the user is logged in
1429 1424 return False
1430 1425 elif getattr(osutil, 'isgui', None):
1431 1426 # check if a CoreGraphics session is available
1432 1427 return osutil.isgui()
1433 1428 else:
1434 1429 # pure build; use a safe default
1435 1430 return True
1436 1431 else:
1437 1432 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1438 1433
1439 1434 def mktempcopy(name, emptyok=False, createmode=None):
1440 1435 """Create a temporary file with the same contents from name
1441 1436
1442 1437 The permission bits are copied from the original file.
1443 1438
1444 1439 If the temporary file is going to be truncated immediately, you
1445 1440 can use emptyok=True as an optimization.
1446 1441
1447 1442 Returns the name of the temporary file.
1448 1443 """
1449 1444 d, fn = os.path.split(name)
1450 1445 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1451 1446 os.close(fd)
1452 1447 # Temporary files are created with mode 0600, which is usually not
1453 1448 # what we want. If the original file already exists, just copy
1454 1449 # its mode. Otherwise, manually obey umask.
1455 1450 copymode(name, temp, createmode)
1456 1451 if emptyok:
1457 1452 return temp
1458 1453 try:
1459 1454 try:
1460 1455 ifp = posixfile(name, "rb")
1461 1456 except IOError as inst:
1462 1457 if inst.errno == errno.ENOENT:
1463 1458 return temp
1464 1459 if not getattr(inst, 'filename', None):
1465 1460 inst.filename = name
1466 1461 raise
1467 1462 ofp = posixfile(temp, "wb")
1468 1463 for chunk in filechunkiter(ifp):
1469 1464 ofp.write(chunk)
1470 1465 ifp.close()
1471 1466 ofp.close()
1472 1467 except: # re-raises
1473 1468 try: os.unlink(temp)
1474 1469 except OSError: pass
1475 1470 raise
1476 1471 return temp
1477 1472
1478 1473 class filestat(object):
1479 1474 """help to exactly detect change of a file
1480 1475
1481 1476 'stat' attribute is result of 'os.stat()' if specified 'path'
1482 1477 exists. Otherwise, it is None. This can avoid preparative
1483 1478 'exists()' examination on client side of this class.
1484 1479 """
1485 1480 def __init__(self, path):
1486 1481 try:
1487 1482 self.stat = os.stat(path)
1488 1483 except OSError as err:
1489 1484 if err.errno != errno.ENOENT:
1490 1485 raise
1491 1486 self.stat = None
1492 1487
1493 1488 __hash__ = object.__hash__
1494 1489
1495 1490 def __eq__(self, old):
1496 1491 try:
1497 1492 # if ambiguity between stat of new and old file is
1498 1493 # avoided, comparison of size, ctime and mtime is enough
1499 1494 # to exactly detect change of a file regardless of platform
1500 1495 return (self.stat.st_size == old.stat.st_size and
1501 1496 self.stat.st_ctime == old.stat.st_ctime and
1502 1497 self.stat.st_mtime == old.stat.st_mtime)
1503 1498 except AttributeError:
1504 1499 return False
1505 1500
1506 1501 def isambig(self, old):
1507 1502 """Examine whether new (= self) stat is ambiguous against old one
1508 1503
1509 1504 "S[N]" below means stat of a file at N-th change:
1510 1505
1511 1506 - S[n-1].ctime < S[n].ctime: can detect change of a file
1512 1507 - S[n-1].ctime == S[n].ctime
1513 1508 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1514 1509 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1515 1510 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1516 1511 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1517 1512
1518 1513 Case (*2) above means that a file was changed twice or more at
1519 1514 same time in sec (= S[n-1].ctime), and comparison of timestamp
1520 1515 is ambiguous.
1521 1516
1522 1517 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1523 1518 timestamp is ambiguous".
1524 1519
1525 1520 But advancing mtime only in case (*2) doesn't work as
1526 1521 expected, because naturally advanced S[n].mtime in case (*1)
1527 1522 might be equal to manually advanced S[n-1 or earlier].mtime.
1528 1523
1529 1524 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1530 1525 treated as ambiguous regardless of mtime, to avoid overlooking
1531 1526 by confliction between such mtime.
1532 1527
1533 1528 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1534 1529 S[n].mtime", even if size of a file isn't changed.
1535 1530 """
1536 1531 try:
1537 1532 return (self.stat.st_ctime == old.stat.st_ctime)
1538 1533 except AttributeError:
1539 1534 return False
1540 1535
1541 1536 def avoidambig(self, path, old):
1542 1537 """Change file stat of specified path to avoid ambiguity
1543 1538
1544 1539 'old' should be previous filestat of 'path'.
1545 1540
1546 1541 This skips avoiding ambiguity, if a process doesn't have
1547 1542 appropriate privileges for 'path'.
1548 1543 """
1549 1544 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1550 1545 try:
1551 1546 os.utime(path, (advanced, advanced))
1552 1547 except OSError as inst:
1553 1548 if inst.errno == errno.EPERM:
1554 1549 # utime() on the file created by another user causes EPERM,
1555 1550 # if a process doesn't have appropriate privileges
1556 1551 return
1557 1552 raise
1558 1553
1559 1554 def __ne__(self, other):
1560 1555 return not self == other
1561 1556
1562 1557 class atomictempfile(object):
1563 1558 '''writable file object that atomically updates a file
1564 1559
1565 1560 All writes will go to a temporary copy of the original file. Call
1566 1561 close() when you are done writing, and atomictempfile will rename
1567 1562 the temporary copy to the original name, making the changes
1568 1563 visible. If the object is destroyed without being closed, all your
1569 1564 writes are discarded.
1570 1565
1571 1566 checkambig argument of constructor is used with filestat, and is
1572 1567 useful only if target file is guarded by any lock (e.g. repo.lock
1573 1568 or repo.wlock).
1574 1569 '''
1575 1570 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1576 1571 self.__name = name # permanent name
1577 1572 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1578 1573 createmode=createmode)
1579 1574 self._fp = posixfile(self._tempname, mode)
1580 1575 self._checkambig = checkambig
1581 1576
1582 1577 # delegated methods
1583 1578 self.read = self._fp.read
1584 1579 self.write = self._fp.write
1585 1580 self.seek = self._fp.seek
1586 1581 self.tell = self._fp.tell
1587 1582 self.fileno = self._fp.fileno
1588 1583
1589 1584 def close(self):
1590 1585 if not self._fp.closed:
1591 1586 self._fp.close()
1592 1587 filename = localpath(self.__name)
1593 1588 oldstat = self._checkambig and filestat(filename)
1594 1589 if oldstat and oldstat.stat:
1595 1590 rename(self._tempname, filename)
1596 1591 newstat = filestat(filename)
1597 1592 if newstat.isambig(oldstat):
1598 1593 # stat of changed file is ambiguous to original one
1599 1594 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1600 1595 os.utime(filename, (advanced, advanced))
1601 1596 else:
1602 1597 rename(self._tempname, filename)
1603 1598
1604 1599 def discard(self):
1605 1600 if not self._fp.closed:
1606 1601 try:
1607 1602 os.unlink(self._tempname)
1608 1603 except OSError:
1609 1604 pass
1610 1605 self._fp.close()
1611 1606
1612 1607 def __del__(self):
1613 1608 if safehasattr(self, '_fp'): # constructor actually did something
1614 1609 self.discard()
1615 1610
1616 1611 def __enter__(self):
1617 1612 return self
1618 1613
1619 1614 def __exit__(self, exctype, excvalue, traceback):
1620 1615 if exctype is not None:
1621 1616 self.discard()
1622 1617 else:
1623 1618 self.close()
1624 1619
1625 1620 def unlinkpath(f, ignoremissing=False):
1626 1621 """unlink and remove the directory if it is empty"""
1627 1622 if ignoremissing:
1628 1623 tryunlink(f)
1629 1624 else:
1630 1625 unlink(f)
1631 1626 # try removing directories that might now be empty
1632 1627 try:
1633 1628 removedirs(os.path.dirname(f))
1634 1629 except OSError:
1635 1630 pass
1636 1631
1637 1632 def tryunlink(f):
1638 1633 """Attempt to remove a file, ignoring ENOENT errors."""
1639 1634 try:
1640 1635 unlink(f)
1641 1636 except OSError as e:
1642 1637 if e.errno != errno.ENOENT:
1643 1638 raise
1644 1639
1645 1640 def makedirs(name, mode=None, notindexed=False):
1646 1641 """recursive directory creation with parent mode inheritance
1647 1642
1648 1643 Newly created directories are marked as "not to be indexed by
1649 1644 the content indexing service", if ``notindexed`` is specified
1650 1645 for "write" mode access.
1651 1646 """
1652 1647 try:
1653 1648 makedir(name, notindexed)
1654 1649 except OSError as err:
1655 1650 if err.errno == errno.EEXIST:
1656 1651 return
1657 1652 if err.errno != errno.ENOENT or not name:
1658 1653 raise
1659 1654 parent = os.path.dirname(os.path.abspath(name))
1660 1655 if parent == name:
1661 1656 raise
1662 1657 makedirs(parent, mode, notindexed)
1663 1658 try:
1664 1659 makedir(name, notindexed)
1665 1660 except OSError as err:
1666 1661 # Catch EEXIST to handle races
1667 1662 if err.errno == errno.EEXIST:
1668 1663 return
1669 1664 raise
1670 1665 if mode is not None:
1671 1666 os.chmod(name, mode)
1672 1667
1673 1668 def readfile(path):
1674 1669 with open(path, 'rb') as fp:
1675 1670 return fp.read()
1676 1671
1677 1672 def writefile(path, text):
1678 1673 with open(path, 'wb') as fp:
1679 1674 fp.write(text)
1680 1675
1681 1676 def appendfile(path, text):
1682 1677 with open(path, 'ab') as fp:
1683 1678 fp.write(text)
1684 1679
1685 1680 class chunkbuffer(object):
1686 1681 """Allow arbitrary sized chunks of data to be efficiently read from an
1687 1682 iterator over chunks of arbitrary size."""
1688 1683
1689 1684 def __init__(self, in_iter):
1690 1685 """in_iter is the iterator that's iterating over the input chunks.
1691 1686 targetsize is how big a buffer to try to maintain."""
1692 1687 def splitbig(chunks):
1693 1688 for chunk in chunks:
1694 1689 if len(chunk) > 2**20:
1695 1690 pos = 0
1696 1691 while pos < len(chunk):
1697 1692 end = pos + 2 ** 18
1698 1693 yield chunk[pos:end]
1699 1694 pos = end
1700 1695 else:
1701 1696 yield chunk
1702 1697 self.iter = splitbig(in_iter)
1703 1698 self._queue = collections.deque()
1704 1699 self._chunkoffset = 0
1705 1700
1706 1701 def read(self, l=None):
1707 1702 """Read L bytes of data from the iterator of chunks of data.
1708 1703 Returns less than L bytes if the iterator runs dry.
1709 1704
1710 1705 If size parameter is omitted, read everything"""
1711 1706 if l is None:
1712 1707 return ''.join(self.iter)
1713 1708
1714 1709 left = l
1715 1710 buf = []
1716 1711 queue = self._queue
1717 1712 while left > 0:
1718 1713 # refill the queue
1719 1714 if not queue:
1720 1715 target = 2**18
1721 1716 for chunk in self.iter:
1722 1717 queue.append(chunk)
1723 1718 target -= len(chunk)
1724 1719 if target <= 0:
1725 1720 break
1726 1721 if not queue:
1727 1722 break
1728 1723
1729 1724 # The easy way to do this would be to queue.popleft(), modify the
1730 1725 # chunk (if necessary), then queue.appendleft(). However, for cases
1731 1726 # where we read partial chunk content, this incurs 2 dequeue
1732 1727 # mutations and creates a new str for the remaining chunk in the
1733 1728 # queue. Our code below avoids this overhead.
1734 1729
1735 1730 chunk = queue[0]
1736 1731 chunkl = len(chunk)
1737 1732 offset = self._chunkoffset
1738 1733
1739 1734 # Use full chunk.
1740 1735 if offset == 0 and left >= chunkl:
1741 1736 left -= chunkl
1742 1737 queue.popleft()
1743 1738 buf.append(chunk)
1744 1739 # self._chunkoffset remains at 0.
1745 1740 continue
1746 1741
1747 1742 chunkremaining = chunkl - offset
1748 1743
1749 1744 # Use all of unconsumed part of chunk.
1750 1745 if left >= chunkremaining:
1751 1746 left -= chunkremaining
1752 1747 queue.popleft()
1753 1748 # offset == 0 is enabled by block above, so this won't merely
1754 1749 # copy via ``chunk[0:]``.
1755 1750 buf.append(chunk[offset:])
1756 1751 self._chunkoffset = 0
1757 1752
1758 1753 # Partial chunk needed.
1759 1754 else:
1760 1755 buf.append(chunk[offset:offset + left])
1761 1756 self._chunkoffset += left
1762 1757 left -= chunkremaining
1763 1758
1764 1759 return ''.join(buf)
1765 1760
1766 1761 def filechunkiter(f, size=131072, limit=None):
1767 1762 """Create a generator that produces the data in the file size
1768 1763 (default 131072) bytes at a time, up to optional limit (default is
1769 1764 to read all data). Chunks may be less than size bytes if the
1770 1765 chunk is the last chunk in the file, or the file is a socket or
1771 1766 some other type of file that sometimes reads less data than is
1772 1767 requested."""
1773 1768 assert size >= 0
1774 1769 assert limit is None or limit >= 0
1775 1770 while True:
1776 1771 if limit is None:
1777 1772 nbytes = size
1778 1773 else:
1779 1774 nbytes = min(limit, size)
1780 1775 s = nbytes and f.read(nbytes)
1781 1776 if not s:
1782 1777 break
1783 1778 if limit:
1784 1779 limit -= len(s)
1785 1780 yield s
1786 1781
1787 1782 def makedate(timestamp=None):
1788 1783 '''Return a unix timestamp (or the current time) as a (unixtime,
1789 1784 offset) tuple based off the local timezone.'''
1790 1785 if timestamp is None:
1791 1786 timestamp = time.time()
1792 1787 if timestamp < 0:
1793 1788 hint = _("check your clock")
1794 1789 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1795 1790 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1796 1791 datetime.datetime.fromtimestamp(timestamp))
1797 1792 tz = delta.days * 86400 + delta.seconds
1798 1793 return timestamp, tz
1799 1794
1800 1795 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1801 1796 """represent a (unixtime, offset) tuple as a localized time.
1802 1797 unixtime is seconds since the epoch, and offset is the time zone's
1803 1798 number of seconds away from UTC.
1804 1799
1805 1800 >>> datestr((0, 0))
1806 1801 'Thu Jan 01 00:00:00 1970 +0000'
1807 1802 >>> datestr((42, 0))
1808 1803 'Thu Jan 01 00:00:42 1970 +0000'
1809 1804 >>> datestr((-42, 0))
1810 1805 'Wed Dec 31 23:59:18 1969 +0000'
1811 1806 >>> datestr((0x7fffffff, 0))
1812 1807 'Tue Jan 19 03:14:07 2038 +0000'
1813 1808 >>> datestr((-0x80000000, 0))
1814 1809 'Fri Dec 13 20:45:52 1901 +0000'
1815 1810 """
1816 1811 t, tz = date or makedate()
1817 1812 if "%1" in format or "%2" in format or "%z" in format:
1818 1813 sign = (tz > 0) and "-" or "+"
1819 1814 minutes = abs(tz) // 60
1820 1815 q, r = divmod(minutes, 60)
1821 1816 format = format.replace("%z", "%1%2")
1822 1817 format = format.replace("%1", "%c%02d" % (sign, q))
1823 1818 format = format.replace("%2", "%02d" % r)
1824 1819 d = t - tz
1825 1820 if d > 0x7fffffff:
1826 1821 d = 0x7fffffff
1827 1822 elif d < -0x80000000:
1828 1823 d = -0x80000000
1829 1824 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1830 1825 # because they use the gmtime() system call which is buggy on Windows
1831 1826 # for negative values.
1832 1827 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1833 1828 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1834 1829 return s
1835 1830
1836 1831 def shortdate(date=None):
1837 1832 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1838 1833 return datestr(date, format='%Y-%m-%d')
1839 1834
1840 1835 def parsetimezone(s):
1841 1836 """find a trailing timezone, if any, in string, and return a
1842 1837 (offset, remainder) pair"""
1843 1838
1844 1839 if s.endswith("GMT") or s.endswith("UTC"):
1845 1840 return 0, s[:-3].rstrip()
1846 1841
1847 1842 # Unix-style timezones [+-]hhmm
1848 1843 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1849 1844 sign = (s[-5] == "+") and 1 or -1
1850 1845 hours = int(s[-4:-2])
1851 1846 minutes = int(s[-2:])
1852 1847 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1853 1848
1854 1849 # ISO8601 trailing Z
1855 1850 if s.endswith("Z") and s[-2:-1].isdigit():
1856 1851 return 0, s[:-1]
1857 1852
1858 1853 # ISO8601-style [+-]hh:mm
1859 1854 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1860 1855 s[-5:-3].isdigit() and s[-2:].isdigit()):
1861 1856 sign = (s[-6] == "+") and 1 or -1
1862 1857 hours = int(s[-5:-3])
1863 1858 minutes = int(s[-2:])
1864 1859 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1865 1860
1866 1861 return None, s
1867 1862
1868 1863 def strdate(string, format, defaults=None):
1869 1864 """parse a localized time string and return a (unixtime, offset) tuple.
1870 1865 if the string cannot be parsed, ValueError is raised."""
1871 1866 if defaults is None:
1872 1867 defaults = {}
1873 1868
1874 1869 # NOTE: unixtime = localunixtime + offset
1875 1870 offset, date = parsetimezone(string)
1876 1871
1877 1872 # add missing elements from defaults
1878 1873 usenow = False # default to using biased defaults
1879 1874 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1880 1875 found = [True for p in part if ("%"+p) in format]
1881 1876 if not found:
1882 1877 date += "@" + defaults[part][usenow]
1883 1878 format += "@%" + part[0]
1884 1879 else:
1885 1880 # We've found a specific time element, less specific time
1886 1881 # elements are relative to today
1887 1882 usenow = True
1888 1883
1889 1884 timetuple = time.strptime(date, format)
1890 1885 localunixtime = int(calendar.timegm(timetuple))
1891 1886 if offset is None:
1892 1887 # local timezone
1893 1888 unixtime = int(time.mktime(timetuple))
1894 1889 offset = unixtime - localunixtime
1895 1890 else:
1896 1891 unixtime = localunixtime + offset
1897 1892 return unixtime, offset
1898 1893
1899 1894 def parsedate(date, formats=None, bias=None):
1900 1895 """parse a localized date/time and return a (unixtime, offset) tuple.
1901 1896
1902 1897 The date may be a "unixtime offset" string or in one of the specified
1903 1898 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1904 1899
1905 1900 >>> parsedate(' today ') == parsedate(\
1906 1901 datetime.date.today().strftime('%b %d'))
1907 1902 True
1908 1903 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1909 1904 datetime.timedelta(days=1)\
1910 1905 ).strftime('%b %d'))
1911 1906 True
1912 1907 >>> now, tz = makedate()
1913 1908 >>> strnow, strtz = parsedate('now')
1914 1909 >>> (strnow - now) < 1
1915 1910 True
1916 1911 >>> tz == strtz
1917 1912 True
1918 1913 """
1919 1914 if bias is None:
1920 1915 bias = {}
1921 1916 if not date:
1922 1917 return 0, 0
1923 1918 if isinstance(date, tuple) and len(date) == 2:
1924 1919 return date
1925 1920 if not formats:
1926 1921 formats = defaultdateformats
1927 1922 date = date.strip()
1928 1923
1929 1924 if date == 'now' or date == _('now'):
1930 1925 return makedate()
1931 1926 if date == 'today' or date == _('today'):
1932 1927 date = datetime.date.today().strftime('%b %d')
1933 1928 elif date == 'yesterday' or date == _('yesterday'):
1934 1929 date = (datetime.date.today() -
1935 1930 datetime.timedelta(days=1)).strftime('%b %d')
1936 1931
1937 1932 try:
1938 1933 when, offset = map(int, date.split(' '))
1939 1934 except ValueError:
1940 1935 # fill out defaults
1941 1936 now = makedate()
1942 1937 defaults = {}
1943 1938 for part in ("d", "mb", "yY", "HI", "M", "S"):
1944 1939 # this piece is for rounding the specific end of unknowns
1945 1940 b = bias.get(part)
1946 1941 if b is None:
1947 1942 if part[0] in "HMS":
1948 1943 b = "00"
1949 1944 else:
1950 1945 b = "0"
1951 1946
1952 1947 # this piece is for matching the generic end to today's date
1953 1948 n = datestr(now, "%" + part[0])
1954 1949
1955 1950 defaults[part] = (b, n)
1956 1951
1957 1952 for format in formats:
1958 1953 try:
1959 1954 when, offset = strdate(date, format, defaults)
1960 1955 except (ValueError, OverflowError):
1961 1956 pass
1962 1957 else:
1963 1958 break
1964 1959 else:
1965 1960 raise Abort(_('invalid date: %r') % date)
1966 1961 # validate explicit (probably user-specified) date and
1967 1962 # time zone offset. values must fit in signed 32 bits for
1968 1963 # current 32-bit linux runtimes. timezones go from UTC-12
1969 1964 # to UTC+14
1970 1965 if when < -0x80000000 or when > 0x7fffffff:
1971 1966 raise Abort(_('date exceeds 32 bits: %d') % when)
1972 1967 if offset < -50400 or offset > 43200:
1973 1968 raise Abort(_('impossible time zone offset: %d') % offset)
1974 1969 return when, offset
1975 1970
1976 1971 def matchdate(date):
1977 1972 """Return a function that matches a given date match specifier
1978 1973
1979 1974 Formats include:
1980 1975
1981 1976 '{date}' match a given date to the accuracy provided
1982 1977
1983 1978 '<{date}' on or before a given date
1984 1979
1985 1980 '>{date}' on or after a given date
1986 1981
1987 1982 >>> p1 = parsedate("10:29:59")
1988 1983 >>> p2 = parsedate("10:30:00")
1989 1984 >>> p3 = parsedate("10:30:59")
1990 1985 >>> p4 = parsedate("10:31:00")
1991 1986 >>> p5 = parsedate("Sep 15 10:30:00 1999")
1992 1987 >>> f = matchdate("10:30")
1993 1988 >>> f(p1[0])
1994 1989 False
1995 1990 >>> f(p2[0])
1996 1991 True
1997 1992 >>> f(p3[0])
1998 1993 True
1999 1994 >>> f(p4[0])
2000 1995 False
2001 1996 >>> f(p5[0])
2002 1997 False
2003 1998 """
2004 1999
2005 2000 def lower(date):
2006 2001 d = {'mb': "1", 'd': "1"}
2007 2002 return parsedate(date, extendeddateformats, d)[0]
2008 2003
2009 2004 def upper(date):
2010 2005 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2011 2006 for days in ("31", "30", "29"):
2012 2007 try:
2013 2008 d["d"] = days
2014 2009 return parsedate(date, extendeddateformats, d)[0]
2015 2010 except Abort:
2016 2011 pass
2017 2012 d["d"] = "28"
2018 2013 return parsedate(date, extendeddateformats, d)[0]
2019 2014
2020 2015 date = date.strip()
2021 2016
2022 2017 if not date:
2023 2018 raise Abort(_("dates cannot consist entirely of whitespace"))
2024 2019 elif date[0] == "<":
2025 2020 if not date[1:]:
2026 2021 raise Abort(_("invalid day spec, use '<DATE'"))
2027 2022 when = upper(date[1:])
2028 2023 return lambda x: x <= when
2029 2024 elif date[0] == ">":
2030 2025 if not date[1:]:
2031 2026 raise Abort(_("invalid day spec, use '>DATE'"))
2032 2027 when = lower(date[1:])
2033 2028 return lambda x: x >= when
2034 2029 elif date[0] == "-":
2035 2030 try:
2036 2031 days = int(date[1:])
2037 2032 except ValueError:
2038 2033 raise Abort(_("invalid day spec: %s") % date[1:])
2039 2034 if days < 0:
2040 2035 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2041 2036 % date[1:])
2042 2037 when = makedate()[0] - days * 3600 * 24
2043 2038 return lambda x: x >= when
2044 2039 elif " to " in date:
2045 2040 a, b = date.split(" to ")
2046 2041 start, stop = lower(a), upper(b)
2047 2042 return lambda x: x >= start and x <= stop
2048 2043 else:
2049 2044 start, stop = lower(date), upper(date)
2050 2045 return lambda x: x >= start and x <= stop
2051 2046
2052 2047 def stringmatcher(pattern, casesensitive=True):
2053 2048 """
2054 2049 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2055 2050 returns the matcher name, pattern, and matcher function.
2056 2051 missing or unknown prefixes are treated as literal matches.
2057 2052
2058 2053 helper for tests:
2059 2054 >>> def test(pattern, *tests):
2060 2055 ... kind, pattern, matcher = stringmatcher(pattern)
2061 2056 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2062 2057 >>> def itest(pattern, *tests):
2063 2058 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2064 2059 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2065 2060
2066 2061 exact matching (no prefix):
2067 2062 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2068 2063 ('literal', 'abcdefg', [False, False, True])
2069 2064
2070 2065 regex matching ('re:' prefix)
2071 2066 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2072 2067 ('re', 'a.+b', [False, False, True])
2073 2068
2074 2069 force exact matches ('literal:' prefix)
2075 2070 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2076 2071 ('literal', 're:foobar', [False, True])
2077 2072
2078 2073 unknown prefixes are ignored and treated as literals
2079 2074 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2080 2075 ('literal', 'foo:bar', [False, False, True])
2081 2076
2082 2077 case insensitive regex matches
2083 2078 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2084 2079 ('re', 'A.+b', [False, False, True])
2085 2080
2086 2081 case insensitive literal matches
2087 2082 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2088 2083 ('literal', 'ABCDEFG', [False, False, True])
2089 2084 """
2090 2085 if pattern.startswith('re:'):
2091 2086 pattern = pattern[3:]
2092 2087 try:
2093 2088 flags = 0
2094 2089 if not casesensitive:
2095 2090 flags = remod.I
2096 2091 regex = remod.compile(pattern, flags)
2097 2092 except remod.error as e:
2098 2093 raise error.ParseError(_('invalid regular expression: %s')
2099 2094 % e)
2100 2095 return 're', pattern, regex.search
2101 2096 elif pattern.startswith('literal:'):
2102 2097 pattern = pattern[8:]
2103 2098
2104 2099 match = pattern.__eq__
2105 2100
2106 2101 if not casesensitive:
2107 2102 ipat = encoding.lower(pattern)
2108 2103 match = lambda s: ipat == encoding.lower(s)
2109 2104 return 'literal', pattern, match
2110 2105
2111 2106 def shortuser(user):
2112 2107 """Return a short representation of a user name or email address."""
2113 2108 f = user.find('@')
2114 2109 if f >= 0:
2115 2110 user = user[:f]
2116 2111 f = user.find('<')
2117 2112 if f >= 0:
2118 2113 user = user[f + 1:]
2119 2114 f = user.find(' ')
2120 2115 if f >= 0:
2121 2116 user = user[:f]
2122 2117 f = user.find('.')
2123 2118 if f >= 0:
2124 2119 user = user[:f]
2125 2120 return user
2126 2121
2127 2122 def emailuser(user):
2128 2123 """Return the user portion of an email address."""
2129 2124 f = user.find('@')
2130 2125 if f >= 0:
2131 2126 user = user[:f]
2132 2127 f = user.find('<')
2133 2128 if f >= 0:
2134 2129 user = user[f + 1:]
2135 2130 return user
2136 2131
2137 2132 def email(author):
2138 2133 '''get email of author.'''
2139 2134 r = author.find('>')
2140 2135 if r == -1:
2141 2136 r = None
2142 2137 return author[author.find('<') + 1:r]
2143 2138
2144 2139 def ellipsis(text, maxlength=400):
2145 2140 """Trim string to at most maxlength (default: 400) columns in display."""
2146 2141 return encoding.trim(text, maxlength, ellipsis='...')
2147 2142
2148 2143 def unitcountfn(*unittable):
2149 2144 '''return a function that renders a readable count of some quantity'''
2150 2145
2151 2146 def go(count):
2152 2147 for multiplier, divisor, format in unittable:
2153 2148 if count >= divisor * multiplier:
2154 2149 return format % (count / float(divisor))
2155 2150 return unittable[-1][2] % count
2156 2151
2157 2152 return go
2158 2153
2159 2154 bytecount = unitcountfn(
2160 2155 (100, 1 << 30, _('%.0f GB')),
2161 2156 (10, 1 << 30, _('%.1f GB')),
2162 2157 (1, 1 << 30, _('%.2f GB')),
2163 2158 (100, 1 << 20, _('%.0f MB')),
2164 2159 (10, 1 << 20, _('%.1f MB')),
2165 2160 (1, 1 << 20, _('%.2f MB')),
2166 2161 (100, 1 << 10, _('%.0f KB')),
2167 2162 (10, 1 << 10, _('%.1f KB')),
2168 2163 (1, 1 << 10, _('%.2f KB')),
2169 2164 (1, 1, _('%.0f bytes')),
2170 2165 )
2171 2166
2172 2167 def escapestr(s):
2173 2168 # call underlying function of s.encode('string_escape') directly for
2174 2169 # Python 3 compatibility
2175 2170 return codecs.escape_encode(s)[0]
2176 2171
2177 2172 def unescapestr(s):
2178 2173 return codecs.escape_decode(s)[0]
2179 2174
2180 2175 def uirepr(s):
2181 2176 # Avoid double backslash in Windows path repr()
2182 2177 return repr(s).replace('\\\\', '\\')
2183 2178
2184 2179 # delay import of textwrap
2185 2180 def MBTextWrapper(**kwargs):
2186 2181 class tw(textwrap.TextWrapper):
2187 2182 """
2188 2183 Extend TextWrapper for width-awareness.
2189 2184
2190 2185 Neither number of 'bytes' in any encoding nor 'characters' is
2191 2186 appropriate to calculate terminal columns for specified string.
2192 2187
2193 2188 Original TextWrapper implementation uses built-in 'len()' directly,
2194 2189 so overriding is needed to use width information of each characters.
2195 2190
2196 2191 In addition, characters classified into 'ambiguous' width are
2197 2192 treated as wide in East Asian area, but as narrow in other.
2198 2193
2199 2194 This requires use decision to determine width of such characters.
2200 2195 """
2201 2196 def _cutdown(self, ucstr, space_left):
2202 2197 l = 0
2203 2198 colwidth = encoding.ucolwidth
2204 2199 for i in xrange(len(ucstr)):
2205 2200 l += colwidth(ucstr[i])
2206 2201 if space_left < l:
2207 2202 return (ucstr[:i], ucstr[i:])
2208 2203 return ucstr, ''
2209 2204
2210 2205 # overriding of base class
2211 2206 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2212 2207 space_left = max(width - cur_len, 1)
2213 2208
2214 2209 if self.break_long_words:
2215 2210 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2216 2211 cur_line.append(cut)
2217 2212 reversed_chunks[-1] = res
2218 2213 elif not cur_line:
2219 2214 cur_line.append(reversed_chunks.pop())
2220 2215
2221 2216 # this overriding code is imported from TextWrapper of Python 2.6
2222 2217 # to calculate columns of string by 'encoding.ucolwidth()'
2223 2218 def _wrap_chunks(self, chunks):
2224 2219 colwidth = encoding.ucolwidth
2225 2220
2226 2221 lines = []
2227 2222 if self.width <= 0:
2228 2223 raise ValueError("invalid width %r (must be > 0)" % self.width)
2229 2224
2230 2225 # Arrange in reverse order so items can be efficiently popped
2231 2226 # from a stack of chucks.
2232 2227 chunks.reverse()
2233 2228
2234 2229 while chunks:
2235 2230
2236 2231 # Start the list of chunks that will make up the current line.
2237 2232 # cur_len is just the length of all the chunks in cur_line.
2238 2233 cur_line = []
2239 2234 cur_len = 0
2240 2235
2241 2236 # Figure out which static string will prefix this line.
2242 2237 if lines:
2243 2238 indent = self.subsequent_indent
2244 2239 else:
2245 2240 indent = self.initial_indent
2246 2241
2247 2242 # Maximum width for this line.
2248 2243 width = self.width - len(indent)
2249 2244
2250 2245 # First chunk on line is whitespace -- drop it, unless this
2251 2246 # is the very beginning of the text (i.e. no lines started yet).
2252 2247 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2253 2248 del chunks[-1]
2254 2249
2255 2250 while chunks:
2256 2251 l = colwidth(chunks[-1])
2257 2252
2258 2253 # Can at least squeeze this chunk onto the current line.
2259 2254 if cur_len + l <= width:
2260 2255 cur_line.append(chunks.pop())
2261 2256 cur_len += l
2262 2257
2263 2258 # Nope, this line is full.
2264 2259 else:
2265 2260 break
2266 2261
2267 2262 # The current line is full, and the next chunk is too big to
2268 2263 # fit on *any* line (not just this one).
2269 2264 if chunks and colwidth(chunks[-1]) > width:
2270 2265 self._handle_long_word(chunks, cur_line, cur_len, width)
2271 2266
2272 2267 # If the last chunk on this line is all whitespace, drop it.
2273 2268 if (self.drop_whitespace and
2274 2269 cur_line and cur_line[-1].strip() == ''):
2275 2270 del cur_line[-1]
2276 2271
2277 2272 # Convert current line back to a string and store it in list
2278 2273 # of all lines (return value).
2279 2274 if cur_line:
2280 2275 lines.append(indent + ''.join(cur_line))
2281 2276
2282 2277 return lines
2283 2278
2284 2279 global MBTextWrapper
2285 2280 MBTextWrapper = tw
2286 2281 return tw(**kwargs)
2287 2282
2288 2283 def wrap(line, width, initindent='', hangindent=''):
2289 2284 maxindent = max(len(hangindent), len(initindent))
2290 2285 if width <= maxindent:
2291 2286 # adjust for weird terminal size
2292 2287 width = max(78, maxindent + 1)
2293 2288 line = line.decode(pycompat.sysstr(encoding.encoding),
2294 2289 pycompat.sysstr(encoding.encodingmode))
2295 2290 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2296 2291 pycompat.sysstr(encoding.encodingmode))
2297 2292 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2298 2293 pycompat.sysstr(encoding.encodingmode))
2299 2294 wrapper = MBTextWrapper(width=width,
2300 2295 initial_indent=initindent,
2301 2296 subsequent_indent=hangindent)
2302 2297 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2303 2298
2304 2299 if (pyplatform.python_implementation() == 'CPython' and
2305 2300 sys.version_info < (3, 0)):
2306 2301 # There is an issue in CPython that some IO methods do not handle EINTR
2307 2302 # correctly. The following table shows what CPython version (and functions)
2308 2303 # are affected (buggy: has the EINTR bug, okay: otherwise):
2309 2304 #
2310 2305 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2311 2306 # --------------------------------------------------
2312 2307 # fp.__iter__ | buggy | buggy | okay
2313 2308 # fp.read* | buggy | okay [1] | okay
2314 2309 #
2315 2310 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2316 2311 #
2317 2312 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2318 2313 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2319 2314 #
2320 2315 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2321 2316 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2322 2317 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2323 2318 # fp.__iter__ but not other fp.read* methods.
2324 2319 #
2325 2320 # On modern systems like Linux, the "read" syscall cannot be interrupted
2326 2321 # when reading "fast" files like on-disk files. So the EINTR issue only
2327 2322 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2328 2323 # files approximately as "fast" files and use the fast (unsafe) code path,
2329 2324 # to minimize the performance impact.
2330 2325 if sys.version_info >= (2, 7, 4):
2331 2326 # fp.readline deals with EINTR correctly, use it as a workaround.
2332 2327 def _safeiterfile(fp):
2333 2328 return iter(fp.readline, '')
2334 2329 else:
2335 2330 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2336 2331 # note: this may block longer than necessary because of bufsize.
2337 2332 def _safeiterfile(fp, bufsize=4096):
2338 2333 fd = fp.fileno()
2339 2334 line = ''
2340 2335 while True:
2341 2336 try:
2342 2337 buf = os.read(fd, bufsize)
2343 2338 except OSError as ex:
2344 2339 # os.read only raises EINTR before any data is read
2345 2340 if ex.errno == errno.EINTR:
2346 2341 continue
2347 2342 else:
2348 2343 raise
2349 2344 line += buf
2350 2345 if '\n' in buf:
2351 2346 splitted = line.splitlines(True)
2352 2347 line = ''
2353 2348 for l in splitted:
2354 2349 if l[-1] == '\n':
2355 2350 yield l
2356 2351 else:
2357 2352 line = l
2358 2353 if not buf:
2359 2354 break
2360 2355 if line:
2361 2356 yield line
2362 2357
2363 2358 def iterfile(fp):
2364 2359 fastpath = True
2365 2360 if type(fp) is file:
2366 2361 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2367 2362 if fastpath:
2368 2363 return fp
2369 2364 else:
2370 2365 return _safeiterfile(fp)
2371 2366 else:
2372 2367 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2373 2368 def iterfile(fp):
2374 2369 return fp
2375 2370
2376 2371 def iterlines(iterator):
2377 2372 for chunk in iterator:
2378 2373 for line in chunk.splitlines():
2379 2374 yield line
2380 2375
2381 2376 def expandpath(path):
2382 2377 return os.path.expanduser(os.path.expandvars(path))
2383 2378
2384 2379 def hgcmd():
2385 2380 """Return the command used to execute current hg
2386 2381
2387 2382 This is different from hgexecutable() because on Windows we want
2388 2383 to avoid things opening new shell windows like batch files, so we
2389 2384 get either the python call or current executable.
2390 2385 """
2391 2386 if mainfrozen():
2392 2387 if getattr(sys, 'frozen', None) == 'macosx_app':
2393 2388 # Env variable set by py2app
2394 2389 return [encoding.environ['EXECUTABLEPATH']]
2395 2390 else:
2396 2391 return [pycompat.sysexecutable]
2397 2392 return gethgcmd()
2398 2393
2399 2394 def rundetached(args, condfn):
2400 2395 """Execute the argument list in a detached process.
2401 2396
2402 2397 condfn is a callable which is called repeatedly and should return
2403 2398 True once the child process is known to have started successfully.
2404 2399 At this point, the child process PID is returned. If the child
2405 2400 process fails to start or finishes before condfn() evaluates to
2406 2401 True, return -1.
2407 2402 """
2408 2403 # Windows case is easier because the child process is either
2409 2404 # successfully starting and validating the condition or exiting
2410 2405 # on failure. We just poll on its PID. On Unix, if the child
2411 2406 # process fails to start, it will be left in a zombie state until
2412 2407 # the parent wait on it, which we cannot do since we expect a long
2413 2408 # running process on success. Instead we listen for SIGCHLD telling
2414 2409 # us our child process terminated.
2415 2410 terminated = set()
2416 2411 def handler(signum, frame):
2417 2412 terminated.add(os.wait())
2418 2413 prevhandler = None
2419 2414 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2420 2415 if SIGCHLD is not None:
2421 2416 prevhandler = signal.signal(SIGCHLD, handler)
2422 2417 try:
2423 2418 pid = spawndetached(args)
2424 2419 while not condfn():
2425 2420 if ((pid in terminated or not testpid(pid))
2426 2421 and not condfn()):
2427 2422 return -1
2428 2423 time.sleep(0.1)
2429 2424 return pid
2430 2425 finally:
2431 2426 if prevhandler is not None:
2432 2427 signal.signal(signal.SIGCHLD, prevhandler)
2433 2428
2434 2429 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2435 2430 """Return the result of interpolating items in the mapping into string s.
2436 2431
2437 2432 prefix is a single character string, or a two character string with
2438 2433 a backslash as the first character if the prefix needs to be escaped in
2439 2434 a regular expression.
2440 2435
2441 2436 fn is an optional function that will be applied to the replacement text
2442 2437 just before replacement.
2443 2438
2444 2439 escape_prefix is an optional flag that allows using doubled prefix for
2445 2440 its escaping.
2446 2441 """
2447 2442 fn = fn or (lambda s: s)
2448 2443 patterns = '|'.join(mapping.keys())
2449 2444 if escape_prefix:
2450 2445 patterns += '|' + prefix
2451 2446 if len(prefix) > 1:
2452 2447 prefix_char = prefix[1:]
2453 2448 else:
2454 2449 prefix_char = prefix
2455 2450 mapping[prefix_char] = prefix_char
2456 2451 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2457 2452 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2458 2453
2459 2454 def getport(port):
2460 2455 """Return the port for a given network service.
2461 2456
2462 2457 If port is an integer, it's returned as is. If it's a string, it's
2463 2458 looked up using socket.getservbyname(). If there's no matching
2464 2459 service, error.Abort is raised.
2465 2460 """
2466 2461 try:
2467 2462 return int(port)
2468 2463 except ValueError:
2469 2464 pass
2470 2465
2471 2466 try:
2472 2467 return socket.getservbyname(port)
2473 2468 except socket.error:
2474 2469 raise Abort(_("no port number associated with service '%s'") % port)
2475 2470
2476 2471 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2477 2472 '0': False, 'no': False, 'false': False, 'off': False,
2478 2473 'never': False}
2479 2474
2480 2475 def parsebool(s):
2481 2476 """Parse s into a boolean.
2482 2477
2483 2478 If s is not a valid boolean, returns None.
2484 2479 """
2485 2480 return _booleans.get(s.lower(), None)
2486 2481
2487 2482 _hextochr = dict((a + b, chr(int(a + b, 16)))
2488 2483 for a in string.hexdigits for b in string.hexdigits)
2489 2484
2490 2485 class url(object):
2491 2486 r"""Reliable URL parser.
2492 2487
2493 2488 This parses URLs and provides attributes for the following
2494 2489 components:
2495 2490
2496 2491 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2497 2492
2498 2493 Missing components are set to None. The only exception is
2499 2494 fragment, which is set to '' if present but empty.
2500 2495
2501 2496 If parsefragment is False, fragment is included in query. If
2502 2497 parsequery is False, query is included in path. If both are
2503 2498 False, both fragment and query are included in path.
2504 2499
2505 2500 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2506 2501
2507 2502 Note that for backward compatibility reasons, bundle URLs do not
2508 2503 take host names. That means 'bundle://../' has a path of '../'.
2509 2504
2510 2505 Examples:
2511 2506
2512 2507 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2513 2508 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2514 2509 >>> url('ssh://[::1]:2200//home/joe/repo')
2515 2510 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2516 2511 >>> url('file:///home/joe/repo')
2517 2512 <url scheme: 'file', path: '/home/joe/repo'>
2518 2513 >>> url('file:///c:/temp/foo/')
2519 2514 <url scheme: 'file', path: 'c:/temp/foo/'>
2520 2515 >>> url('bundle:foo')
2521 2516 <url scheme: 'bundle', path: 'foo'>
2522 2517 >>> url('bundle://../foo')
2523 2518 <url scheme: 'bundle', path: '../foo'>
2524 2519 >>> url(r'c:\foo\bar')
2525 2520 <url path: 'c:\\foo\\bar'>
2526 2521 >>> url(r'\\blah\blah\blah')
2527 2522 <url path: '\\\\blah\\blah\\blah'>
2528 2523 >>> url(r'\\blah\blah\blah#baz')
2529 2524 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2530 2525 >>> url(r'file:///C:\users\me')
2531 2526 <url scheme: 'file', path: 'C:\\users\\me'>
2532 2527
2533 2528 Authentication credentials:
2534 2529
2535 2530 >>> url('ssh://joe:xyz@x/repo')
2536 2531 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2537 2532 >>> url('ssh://joe@x/repo')
2538 2533 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2539 2534
2540 2535 Query strings and fragments:
2541 2536
2542 2537 >>> url('http://host/a?b#c')
2543 2538 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2544 2539 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2545 2540 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2546 2541
2547 2542 Empty path:
2548 2543
2549 2544 >>> url('')
2550 2545 <url path: ''>
2551 2546 >>> url('#a')
2552 2547 <url path: '', fragment: 'a'>
2553 2548 >>> url('http://host/')
2554 2549 <url scheme: 'http', host: 'host', path: ''>
2555 2550 >>> url('http://host/#a')
2556 2551 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2557 2552
2558 2553 Only scheme:
2559 2554
2560 2555 >>> url('http:')
2561 2556 <url scheme: 'http'>
2562 2557 """
2563 2558
2564 2559 _safechars = "!~*'()+"
2565 2560 _safepchars = "/!~*'()+:\\"
2566 2561 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2567 2562
2568 2563 def __init__(self, path, parsequery=True, parsefragment=True):
2569 2564 # We slowly chomp away at path until we have only the path left
2570 2565 self.scheme = self.user = self.passwd = self.host = None
2571 2566 self.port = self.path = self.query = self.fragment = None
2572 2567 self._localpath = True
2573 2568 self._hostport = ''
2574 2569 self._origpath = path
2575 2570
2576 2571 if parsefragment and '#' in path:
2577 2572 path, self.fragment = path.split('#', 1)
2578 2573
2579 2574 # special case for Windows drive letters and UNC paths
2580 2575 if hasdriveletter(path) or path.startswith('\\\\'):
2581 2576 self.path = path
2582 2577 return
2583 2578
2584 2579 # For compatibility reasons, we can't handle bundle paths as
2585 2580 # normal URLS
2586 2581 if path.startswith('bundle:'):
2587 2582 self.scheme = 'bundle'
2588 2583 path = path[7:]
2589 2584 if path.startswith('//'):
2590 2585 path = path[2:]
2591 2586 self.path = path
2592 2587 return
2593 2588
2594 2589 if self._matchscheme(path):
2595 2590 parts = path.split(':', 1)
2596 2591 if parts[0]:
2597 2592 self.scheme, path = parts
2598 2593 self._localpath = False
2599 2594
2600 2595 if not path:
2601 2596 path = None
2602 2597 if self._localpath:
2603 2598 self.path = ''
2604 2599 return
2605 2600 else:
2606 2601 if self._localpath:
2607 2602 self.path = path
2608 2603 return
2609 2604
2610 2605 if parsequery and '?' in path:
2611 2606 path, self.query = path.split('?', 1)
2612 2607 if not path:
2613 2608 path = None
2614 2609 if not self.query:
2615 2610 self.query = None
2616 2611
2617 2612 # // is required to specify a host/authority
2618 2613 if path and path.startswith('//'):
2619 2614 parts = path[2:].split('/', 1)
2620 2615 if len(parts) > 1:
2621 2616 self.host, path = parts
2622 2617 else:
2623 2618 self.host = parts[0]
2624 2619 path = None
2625 2620 if not self.host:
2626 2621 self.host = None
2627 2622 # path of file:///d is /d
2628 2623 # path of file:///d:/ is d:/, not /d:/
2629 2624 if path and not hasdriveletter(path):
2630 2625 path = '/' + path
2631 2626
2632 2627 if self.host and '@' in self.host:
2633 2628 self.user, self.host = self.host.rsplit('@', 1)
2634 2629 if ':' in self.user:
2635 2630 self.user, self.passwd = self.user.split(':', 1)
2636 2631 if not self.host:
2637 2632 self.host = None
2638 2633
2639 2634 # Don't split on colons in IPv6 addresses without ports
2640 2635 if (self.host and ':' in self.host and
2641 2636 not (self.host.startswith('[') and self.host.endswith(']'))):
2642 2637 self._hostport = self.host
2643 2638 self.host, self.port = self.host.rsplit(':', 1)
2644 2639 if not self.host:
2645 2640 self.host = None
2646 2641
2647 2642 if (self.host and self.scheme == 'file' and
2648 2643 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2649 2644 raise Abort(_('file:// URLs can only refer to localhost'))
2650 2645
2651 2646 self.path = path
2652 2647
2653 2648 # leave the query string escaped
2654 2649 for a in ('user', 'passwd', 'host', 'port',
2655 2650 'path', 'fragment'):
2656 2651 v = getattr(self, a)
2657 2652 if v is not None:
2658 2653 setattr(self, a, urlreq.unquote(v))
2659 2654
2660 2655 def __repr__(self):
2661 2656 attrs = []
2662 2657 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2663 2658 'query', 'fragment'):
2664 2659 v = getattr(self, a)
2665 2660 if v is not None:
2666 2661 attrs.append('%s: %r' % (a, v))
2667 2662 return '<url %s>' % ', '.join(attrs)
2668 2663
2669 2664 def __str__(self):
2670 2665 r"""Join the URL's components back into a URL string.
2671 2666
2672 2667 Examples:
2673 2668
2674 2669 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2675 2670 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2676 2671 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2677 2672 'http://user:pw@host:80/?foo=bar&baz=42'
2678 2673 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2679 2674 'http://user:pw@host:80/?foo=bar%3dbaz'
2680 2675 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2681 2676 'ssh://user:pw@[::1]:2200//home/joe#'
2682 2677 >>> str(url('http://localhost:80//'))
2683 2678 'http://localhost:80//'
2684 2679 >>> str(url('http://localhost:80/'))
2685 2680 'http://localhost:80/'
2686 2681 >>> str(url('http://localhost:80'))
2687 2682 'http://localhost:80/'
2688 2683 >>> str(url('bundle:foo'))
2689 2684 'bundle:foo'
2690 2685 >>> str(url('bundle://../foo'))
2691 2686 'bundle:../foo'
2692 2687 >>> str(url('path'))
2693 2688 'path'
2694 2689 >>> str(url('file:///tmp/foo/bar'))
2695 2690 'file:///tmp/foo/bar'
2696 2691 >>> str(url('file:///c:/tmp/foo/bar'))
2697 2692 'file:///c:/tmp/foo/bar'
2698 2693 >>> print url(r'bundle:foo\bar')
2699 2694 bundle:foo\bar
2700 2695 >>> print url(r'file:///D:\data\hg')
2701 2696 file:///D:\data\hg
2702 2697 """
2703 2698 return encoding.strfromlocal(self.__bytes__())
2704 2699
2705 2700 def __bytes__(self):
2706 2701 if self._localpath:
2707 2702 s = self.path
2708 2703 if self.scheme == 'bundle':
2709 2704 s = 'bundle:' + s
2710 2705 if self.fragment:
2711 2706 s += '#' + self.fragment
2712 2707 return s
2713 2708
2714 2709 s = self.scheme + ':'
2715 2710 if self.user or self.passwd or self.host:
2716 2711 s += '//'
2717 2712 elif self.scheme and (not self.path or self.path.startswith('/')
2718 2713 or hasdriveletter(self.path)):
2719 2714 s += '//'
2720 2715 if hasdriveletter(self.path):
2721 2716 s += '/'
2722 2717 if self.user:
2723 2718 s += urlreq.quote(self.user, safe=self._safechars)
2724 2719 if self.passwd:
2725 2720 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2726 2721 if self.user or self.passwd:
2727 2722 s += '@'
2728 2723 if self.host:
2729 2724 if not (self.host.startswith('[') and self.host.endswith(']')):
2730 2725 s += urlreq.quote(self.host)
2731 2726 else:
2732 2727 s += self.host
2733 2728 if self.port:
2734 2729 s += ':' + urlreq.quote(self.port)
2735 2730 if self.host:
2736 2731 s += '/'
2737 2732 if self.path:
2738 2733 # TODO: similar to the query string, we should not unescape the
2739 2734 # path when we store it, the path might contain '%2f' = '/',
2740 2735 # which we should *not* escape.
2741 2736 s += urlreq.quote(self.path, safe=self._safepchars)
2742 2737 if self.query:
2743 2738 # we store the query in escaped form.
2744 2739 s += '?' + self.query
2745 2740 if self.fragment is not None:
2746 2741 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2747 2742 return s
2748 2743
2749 2744 def authinfo(self):
2750 2745 user, passwd = self.user, self.passwd
2751 2746 try:
2752 2747 self.user, self.passwd = None, None
2753 2748 s = str(self)
2754 2749 finally:
2755 2750 self.user, self.passwd = user, passwd
2756 2751 if not self.user:
2757 2752 return (s, None)
2758 2753 # authinfo[1] is passed to urllib2 password manager, and its
2759 2754 # URIs must not contain credentials. The host is passed in the
2760 2755 # URIs list because Python < 2.4.3 uses only that to search for
2761 2756 # a password.
2762 2757 return (s, (None, (s, self.host),
2763 2758 self.user, self.passwd or ''))
2764 2759
2765 2760 def isabs(self):
2766 2761 if self.scheme and self.scheme != 'file':
2767 2762 return True # remote URL
2768 2763 if hasdriveletter(self.path):
2769 2764 return True # absolute for our purposes - can't be joined()
2770 2765 if self.path.startswith(r'\\'):
2771 2766 return True # Windows UNC path
2772 2767 if self.path.startswith('/'):
2773 2768 return True # POSIX-style
2774 2769 return False
2775 2770
2776 2771 def localpath(self):
2777 2772 if self.scheme == 'file' or self.scheme == 'bundle':
2778 2773 path = self.path or '/'
2779 2774 # For Windows, we need to promote hosts containing drive
2780 2775 # letters to paths with drive letters.
2781 2776 if hasdriveletter(self._hostport):
2782 2777 path = self._hostport + '/' + self.path
2783 2778 elif (self.host is not None and self.path
2784 2779 and not hasdriveletter(path)):
2785 2780 path = '/' + path
2786 2781 return path
2787 2782 return self._origpath
2788 2783
2789 2784 def islocal(self):
2790 2785 '''whether localpath will return something that posixfile can open'''
2791 2786 return (not self.scheme or self.scheme == 'file'
2792 2787 or self.scheme == 'bundle')
2793 2788
2794 2789 def hasscheme(path):
2795 2790 return bool(url(path).scheme)
2796 2791
2797 2792 def hasdriveletter(path):
2798 2793 return path and path[1:2] == ':' and path[0:1].isalpha()
2799 2794
2800 2795 def urllocalpath(path):
2801 2796 return url(path, parsequery=False, parsefragment=False).localpath()
2802 2797
2803 2798 def hidepassword(u):
2804 2799 '''hide user credential in a url string'''
2805 2800 u = url(u)
2806 2801 if u.passwd:
2807 2802 u.passwd = '***'
2808 2803 return str(u)
2809 2804
2810 2805 def removeauth(u):
2811 2806 '''remove all authentication information from a url string'''
2812 2807 u = url(u)
2813 2808 u.user = u.passwd = None
2814 2809 return str(u)
2815 2810
2816 2811 timecount = unitcountfn(
2817 2812 (1, 1e3, _('%.0f s')),
2818 2813 (100, 1, _('%.1f s')),
2819 2814 (10, 1, _('%.2f s')),
2820 2815 (1, 1, _('%.3f s')),
2821 2816 (100, 0.001, _('%.1f ms')),
2822 2817 (10, 0.001, _('%.2f ms')),
2823 2818 (1, 0.001, _('%.3f ms')),
2824 2819 (100, 0.000001, _('%.1f us')),
2825 2820 (10, 0.000001, _('%.2f us')),
2826 2821 (1, 0.000001, _('%.3f us')),
2827 2822 (100, 0.000000001, _('%.1f ns')),
2828 2823 (10, 0.000000001, _('%.2f ns')),
2829 2824 (1, 0.000000001, _('%.3f ns')),
2830 2825 )
2831 2826
2832 2827 _timenesting = [0]
2833 2828
2834 2829 def timed(func):
2835 2830 '''Report the execution time of a function call to stderr.
2836 2831
2837 2832 During development, use as a decorator when you need to measure
2838 2833 the cost of a function, e.g. as follows:
2839 2834
2840 2835 @util.timed
2841 2836 def foo(a, b, c):
2842 2837 pass
2843 2838 '''
2844 2839
2845 2840 def wrapper(*args, **kwargs):
2846 2841 start = timer()
2847 2842 indent = 2
2848 2843 _timenesting[0] += indent
2849 2844 try:
2850 2845 return func(*args, **kwargs)
2851 2846 finally:
2852 2847 elapsed = timer() - start
2853 2848 _timenesting[0] -= indent
2854 2849 stderr.write('%s%s: %s\n' %
2855 2850 (' ' * _timenesting[0], func.__name__,
2856 2851 timecount(elapsed)))
2857 2852 return wrapper
2858 2853
2859 2854 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2860 2855 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2861 2856
2862 2857 def sizetoint(s):
2863 2858 '''Convert a space specifier to a byte count.
2864 2859
2865 2860 >>> sizetoint('30')
2866 2861 30
2867 2862 >>> sizetoint('2.2kb')
2868 2863 2252
2869 2864 >>> sizetoint('6M')
2870 2865 6291456
2871 2866 '''
2872 2867 t = s.strip().lower()
2873 2868 try:
2874 2869 for k, u in _sizeunits:
2875 2870 if t.endswith(k):
2876 2871 return int(float(t[:-len(k)]) * u)
2877 2872 return int(t)
2878 2873 except ValueError:
2879 2874 raise error.ParseError(_("couldn't parse size: %s") % s)
2880 2875
2881 2876 class hooks(object):
2882 2877 '''A collection of hook functions that can be used to extend a
2883 2878 function's behavior. Hooks are called in lexicographic order,
2884 2879 based on the names of their sources.'''
2885 2880
2886 2881 def __init__(self):
2887 2882 self._hooks = []
2888 2883
2889 2884 def add(self, source, hook):
2890 2885 self._hooks.append((source, hook))
2891 2886
2892 2887 def __call__(self, *args):
2893 2888 self._hooks.sort(key=lambda x: x[0])
2894 2889 results = []
2895 2890 for source, hook in self._hooks:
2896 2891 results.append(hook(*args))
2897 2892 return results
2898 2893
2899 2894 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2900 2895 '''Yields lines for a nicely formatted stacktrace.
2901 2896 Skips the 'skip' last entries, then return the last 'depth' entries.
2902 2897 Each file+linenumber is formatted according to fileline.
2903 2898 Each line is formatted according to line.
2904 2899 If line is None, it yields:
2905 2900 length of longest filepath+line number,
2906 2901 filepath+linenumber,
2907 2902 function
2908 2903
2909 2904 Not be used in production code but very convenient while developing.
2910 2905 '''
2911 2906 entries = [(fileline % (fn, ln), func)
2912 2907 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2913 2908 ][-depth:]
2914 2909 if entries:
2915 2910 fnmax = max(len(entry[0]) for entry in entries)
2916 2911 for fnln, func in entries:
2917 2912 if line is None:
2918 2913 yield (fnmax, fnln, func)
2919 2914 else:
2920 2915 yield line % (fnmax, fnln, func)
2921 2916
2922 2917 def debugstacktrace(msg='stacktrace', skip=0,
2923 2918 f=stderr, otherf=stdout, depth=0):
2924 2919 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2925 2920 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2926 2921 By default it will flush stdout first.
2927 2922 It can be used everywhere and intentionally does not require an ui object.
2928 2923 Not be used in production code but very convenient while developing.
2929 2924 '''
2930 2925 if otherf:
2931 2926 otherf.flush()
2932 2927 f.write('%s at:\n' % msg.rstrip())
2933 2928 for line in getstackframes(skip + 1, depth=depth):
2934 2929 f.write(line)
2935 2930 f.flush()
2936 2931
2937 2932 class dirs(object):
2938 2933 '''a multiset of directory names from a dirstate or manifest'''
2939 2934
2940 2935 def __init__(self, map, skip=None):
2941 2936 self._dirs = {}
2942 2937 addpath = self.addpath
2943 2938 if safehasattr(map, 'iteritems') and skip is not None:
2944 2939 for f, s in map.iteritems():
2945 2940 if s[0] != skip:
2946 2941 addpath(f)
2947 2942 else:
2948 2943 for f in map:
2949 2944 addpath(f)
2950 2945
2951 2946 def addpath(self, path):
2952 2947 dirs = self._dirs
2953 2948 for base in finddirs(path):
2954 2949 if base in dirs:
2955 2950 dirs[base] += 1
2956 2951 return
2957 2952 dirs[base] = 1
2958 2953
2959 2954 def delpath(self, path):
2960 2955 dirs = self._dirs
2961 2956 for base in finddirs(path):
2962 2957 if dirs[base] > 1:
2963 2958 dirs[base] -= 1
2964 2959 return
2965 2960 del dirs[base]
2966 2961
2967 2962 def __iter__(self):
2968 2963 return iter(self._dirs)
2969 2964
2970 2965 def __contains__(self, d):
2971 2966 return d in self._dirs
2972 2967
2973 2968 if safehasattr(parsers, 'dirs'):
2974 2969 dirs = parsers.dirs
2975 2970
2976 2971 def finddirs(path):
2977 2972 pos = path.rfind('/')
2978 2973 while pos != -1:
2979 2974 yield path[:pos]
2980 2975 pos = path.rfind('/', 0, pos)
2981 2976
2982 2977 class ctxmanager(object):
2983 2978 '''A context manager for use in 'with' blocks to allow multiple
2984 2979 contexts to be entered at once. This is both safer and more
2985 2980 flexible than contextlib.nested.
2986 2981
2987 2982 Once Mercurial supports Python 2.7+, this will become mostly
2988 2983 unnecessary.
2989 2984 '''
2990 2985
2991 2986 def __init__(self, *args):
2992 2987 '''Accepts a list of no-argument functions that return context
2993 2988 managers. These will be invoked at __call__ time.'''
2994 2989 self._pending = args
2995 2990 self._atexit = []
2996 2991
2997 2992 def __enter__(self):
2998 2993 return self
2999 2994
3000 2995 def enter(self):
3001 2996 '''Create and enter context managers in the order in which they were
3002 2997 passed to the constructor.'''
3003 2998 values = []
3004 2999 for func in self._pending:
3005 3000 obj = func()
3006 3001 values.append(obj.__enter__())
3007 3002 self._atexit.append(obj.__exit__)
3008 3003 del self._pending
3009 3004 return values
3010 3005
3011 3006 def atexit(self, func, *args, **kwargs):
3012 3007 '''Add a function to call when this context manager exits. The
3013 3008 ordering of multiple atexit calls is unspecified, save that
3014 3009 they will happen before any __exit__ functions.'''
3015 3010 def wrapper(exc_type, exc_val, exc_tb):
3016 3011 func(*args, **kwargs)
3017 3012 self._atexit.append(wrapper)
3018 3013 return func
3019 3014
3020 3015 def __exit__(self, exc_type, exc_val, exc_tb):
3021 3016 '''Context managers are exited in the reverse order from which
3022 3017 they were created.'''
3023 3018 received = exc_type is not None
3024 3019 suppressed = False
3025 3020 pending = None
3026 3021 self._atexit.reverse()
3027 3022 for exitfunc in self._atexit:
3028 3023 try:
3029 3024 if exitfunc(exc_type, exc_val, exc_tb):
3030 3025 suppressed = True
3031 3026 exc_type = None
3032 3027 exc_val = None
3033 3028 exc_tb = None
3034 3029 except BaseException:
3035 3030 pending = sys.exc_info()
3036 3031 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3037 3032 del self._atexit
3038 3033 if pending:
3039 3034 raise exc_val
3040 3035 return received and suppressed
3041 3036
3042 3037 # compression code
3043 3038
3044 3039 SERVERROLE = 'server'
3045 3040 CLIENTROLE = 'client'
3046 3041
3047 3042 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3048 3043 (u'name', u'serverpriority',
3049 3044 u'clientpriority'))
3050 3045
3051 3046 class compressormanager(object):
3052 3047 """Holds registrations of various compression engines.
3053 3048
3054 3049 This class essentially abstracts the differences between compression
3055 3050 engines to allow new compression formats to be added easily, possibly from
3056 3051 extensions.
3057 3052
3058 3053 Compressors are registered against the global instance by calling its
3059 3054 ``register()`` method.
3060 3055 """
3061 3056 def __init__(self):
3062 3057 self._engines = {}
3063 3058 # Bundle spec human name to engine name.
3064 3059 self._bundlenames = {}
3065 3060 # Internal bundle identifier to engine name.
3066 3061 self._bundletypes = {}
3067 3062 # Revlog header to engine name.
3068 3063 self._revlogheaders = {}
3069 3064 # Wire proto identifier to engine name.
3070 3065 self._wiretypes = {}
3071 3066
3072 3067 def __getitem__(self, key):
3073 3068 return self._engines[key]
3074 3069
3075 3070 def __contains__(self, key):
3076 3071 return key in self._engines
3077 3072
3078 3073 def __iter__(self):
3079 3074 return iter(self._engines.keys())
3080 3075
3081 3076 def register(self, engine):
3082 3077 """Register a compression engine with the manager.
3083 3078
3084 3079 The argument must be a ``compressionengine`` instance.
3085 3080 """
3086 3081 if not isinstance(engine, compressionengine):
3087 3082 raise ValueError(_('argument must be a compressionengine'))
3088 3083
3089 3084 name = engine.name()
3090 3085
3091 3086 if name in self._engines:
3092 3087 raise error.Abort(_('compression engine %s already registered') %
3093 3088 name)
3094 3089
3095 3090 bundleinfo = engine.bundletype()
3096 3091 if bundleinfo:
3097 3092 bundlename, bundletype = bundleinfo
3098 3093
3099 3094 if bundlename in self._bundlenames:
3100 3095 raise error.Abort(_('bundle name %s already registered') %
3101 3096 bundlename)
3102 3097 if bundletype in self._bundletypes:
3103 3098 raise error.Abort(_('bundle type %s already registered by %s') %
3104 3099 (bundletype, self._bundletypes[bundletype]))
3105 3100
3106 3101 # No external facing name declared.
3107 3102 if bundlename:
3108 3103 self._bundlenames[bundlename] = name
3109 3104
3110 3105 self._bundletypes[bundletype] = name
3111 3106
3112 3107 wiresupport = engine.wireprotosupport()
3113 3108 if wiresupport:
3114 3109 wiretype = wiresupport.name
3115 3110 if wiretype in self._wiretypes:
3116 3111 raise error.Abort(_('wire protocol compression %s already '
3117 3112 'registered by %s') %
3118 3113 (wiretype, self._wiretypes[wiretype]))
3119 3114
3120 3115 self._wiretypes[wiretype] = name
3121 3116
3122 3117 revlogheader = engine.revlogheader()
3123 3118 if revlogheader and revlogheader in self._revlogheaders:
3124 3119 raise error.Abort(_('revlog header %s already registered by %s') %
3125 3120 (revlogheader, self._revlogheaders[revlogheader]))
3126 3121
3127 3122 if revlogheader:
3128 3123 self._revlogheaders[revlogheader] = name
3129 3124
3130 3125 self._engines[name] = engine
3131 3126
3132 3127 @property
3133 3128 def supportedbundlenames(self):
3134 3129 return set(self._bundlenames.keys())
3135 3130
3136 3131 @property
3137 3132 def supportedbundletypes(self):
3138 3133 return set(self._bundletypes.keys())
3139 3134
3140 3135 def forbundlename(self, bundlename):
3141 3136 """Obtain a compression engine registered to a bundle name.
3142 3137
3143 3138 Will raise KeyError if the bundle type isn't registered.
3144 3139
3145 3140 Will abort if the engine is known but not available.
3146 3141 """
3147 3142 engine = self._engines[self._bundlenames[bundlename]]
3148 3143 if not engine.available():
3149 3144 raise error.Abort(_('compression engine %s could not be loaded') %
3150 3145 engine.name())
3151 3146 return engine
3152 3147
3153 3148 def forbundletype(self, bundletype):
3154 3149 """Obtain a compression engine registered to a bundle type.
3155 3150
3156 3151 Will raise KeyError if the bundle type isn't registered.
3157 3152
3158 3153 Will abort if the engine is known but not available.
3159 3154 """
3160 3155 engine = self._engines[self._bundletypes[bundletype]]
3161 3156 if not engine.available():
3162 3157 raise error.Abort(_('compression engine %s could not be loaded') %
3163 3158 engine.name())
3164 3159 return engine
3165 3160
3166 3161 def supportedwireengines(self, role, onlyavailable=True):
3167 3162 """Obtain compression engines that support the wire protocol.
3168 3163
3169 3164 Returns a list of engines in prioritized order, most desired first.
3170 3165
3171 3166 If ``onlyavailable`` is set, filter out engines that can't be
3172 3167 loaded.
3173 3168 """
3174 3169 assert role in (SERVERROLE, CLIENTROLE)
3175 3170
3176 3171 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3177 3172
3178 3173 engines = [self._engines[e] for e in self._wiretypes.values()]
3179 3174 if onlyavailable:
3180 3175 engines = [e for e in engines if e.available()]
3181 3176
3182 3177 def getkey(e):
3183 3178 # Sort first by priority, highest first. In case of tie, sort
3184 3179 # alphabetically. This is arbitrary, but ensures output is
3185 3180 # stable.
3186 3181 w = e.wireprotosupport()
3187 3182 return -1 * getattr(w, attr), w.name
3188 3183
3189 3184 return list(sorted(engines, key=getkey))
3190 3185
3191 3186 def forwiretype(self, wiretype):
3192 3187 engine = self._engines[self._wiretypes[wiretype]]
3193 3188 if not engine.available():
3194 3189 raise error.Abort(_('compression engine %s could not be loaded') %
3195 3190 engine.name())
3196 3191 return engine
3197 3192
3198 3193 def forrevlogheader(self, header):
3199 3194 """Obtain a compression engine registered to a revlog header.
3200 3195
3201 3196 Will raise KeyError if the revlog header value isn't registered.
3202 3197 """
3203 3198 return self._engines[self._revlogheaders[header]]
3204 3199
3205 3200 compengines = compressormanager()
3206 3201
3207 3202 class compressionengine(object):
3208 3203 """Base class for compression engines.
3209 3204
3210 3205 Compression engines must implement the interface defined by this class.
3211 3206 """
3212 3207 def name(self):
3213 3208 """Returns the name of the compression engine.
3214 3209
3215 3210 This is the key the engine is registered under.
3216 3211
3217 3212 This method must be implemented.
3218 3213 """
3219 3214 raise NotImplementedError()
3220 3215
3221 3216 def available(self):
3222 3217 """Whether the compression engine is available.
3223 3218
3224 3219 The intent of this method is to allow optional compression engines
3225 3220 that may not be available in all installations (such as engines relying
3226 3221 on C extensions that may not be present).
3227 3222 """
3228 3223 return True
3229 3224
3230 3225 def bundletype(self):
3231 3226 """Describes bundle identifiers for this engine.
3232 3227
3233 3228 If this compression engine isn't supported for bundles, returns None.
3234 3229
3235 3230 If this engine can be used for bundles, returns a 2-tuple of strings of
3236 3231 the user-facing "bundle spec" compression name and an internal
3237 3232 identifier used to denote the compression format within bundles. To
3238 3233 exclude the name from external usage, set the first element to ``None``.
3239 3234
3240 3235 If bundle compression is supported, the class must also implement
3241 3236 ``compressstream`` and `decompressorreader``.
3242 3237 """
3243 3238 return None
3244 3239
3245 3240 def wireprotosupport(self):
3246 3241 """Declare support for this compression format on the wire protocol.
3247 3242
3248 3243 If this compression engine isn't supported for compressing wire
3249 3244 protocol payloads, returns None.
3250 3245
3251 3246 Otherwise, returns ``compenginewireprotosupport`` with the following
3252 3247 fields:
3253 3248
3254 3249 * String format identifier
3255 3250 * Integer priority for the server
3256 3251 * Integer priority for the client
3257 3252
3258 3253 The integer priorities are used to order the advertisement of format
3259 3254 support by server and client. The highest integer is advertised
3260 3255 first. Integers with non-positive values aren't advertised.
3261 3256
3262 3257 The priority values are somewhat arbitrary and only used for default
3263 3258 ordering. The relative order can be changed via config options.
3264 3259
3265 3260 If wire protocol compression is supported, the class must also implement
3266 3261 ``compressstream`` and ``decompressorreader``.
3267 3262 """
3268 3263 return None
3269 3264
3270 3265 def revlogheader(self):
3271 3266 """Header added to revlog chunks that identifies this engine.
3272 3267
3273 3268 If this engine can be used to compress revlogs, this method should
3274 3269 return the bytes used to identify chunks compressed with this engine.
3275 3270 Else, the method should return ``None`` to indicate it does not
3276 3271 participate in revlog compression.
3277 3272 """
3278 3273 return None
3279 3274
3280 3275 def compressstream(self, it, opts=None):
3281 3276 """Compress an iterator of chunks.
3282 3277
3283 3278 The method receives an iterator (ideally a generator) of chunks of
3284 3279 bytes to be compressed. It returns an iterator (ideally a generator)
3285 3280 of bytes of chunks representing the compressed output.
3286 3281
3287 3282 Optionally accepts an argument defining how to perform compression.
3288 3283 Each engine treats this argument differently.
3289 3284 """
3290 3285 raise NotImplementedError()
3291 3286
3292 3287 def decompressorreader(self, fh):
3293 3288 """Perform decompression on a file object.
3294 3289
3295 3290 Argument is an object with a ``read(size)`` method that returns
3296 3291 compressed data. Return value is an object with a ``read(size)`` that
3297 3292 returns uncompressed data.
3298 3293 """
3299 3294 raise NotImplementedError()
3300 3295
3301 3296 def revlogcompressor(self, opts=None):
3302 3297 """Obtain an object that can be used to compress revlog entries.
3303 3298
3304 3299 The object has a ``compress(data)`` method that compresses binary
3305 3300 data. This method returns compressed binary data or ``None`` if
3306 3301 the data could not be compressed (too small, not compressible, etc).
3307 3302 The returned data should have a header uniquely identifying this
3308 3303 compression format so decompression can be routed to this engine.
3309 3304 This header should be identified by the ``revlogheader()`` return
3310 3305 value.
3311 3306
3312 3307 The object has a ``decompress(data)`` method that decompresses
3313 3308 data. The method will only be called if ``data`` begins with
3314 3309 ``revlogheader()``. The method should return the raw, uncompressed
3315 3310 data or raise a ``RevlogError``.
3316 3311
3317 3312 The object is reusable but is not thread safe.
3318 3313 """
3319 3314 raise NotImplementedError()
3320 3315
3321 3316 class _zlibengine(compressionengine):
3322 3317 def name(self):
3323 3318 return 'zlib'
3324 3319
3325 3320 def bundletype(self):
3326 3321 return 'gzip', 'GZ'
3327 3322
3328 3323 def wireprotosupport(self):
3329 3324 return compewireprotosupport('zlib', 20, 20)
3330 3325
3331 3326 def revlogheader(self):
3332 3327 return 'x'
3333 3328
3334 3329 def compressstream(self, it, opts=None):
3335 3330 opts = opts or {}
3336 3331
3337 3332 z = zlib.compressobj(opts.get('level', -1))
3338 3333 for chunk in it:
3339 3334 data = z.compress(chunk)
3340 3335 # Not all calls to compress emit data. It is cheaper to inspect
3341 3336 # here than to feed empty chunks through generator.
3342 3337 if data:
3343 3338 yield data
3344 3339
3345 3340 yield z.flush()
3346 3341
3347 3342 def decompressorreader(self, fh):
3348 3343 def gen():
3349 3344 d = zlib.decompressobj()
3350 3345 for chunk in filechunkiter(fh):
3351 3346 while chunk:
3352 3347 # Limit output size to limit memory.
3353 3348 yield d.decompress(chunk, 2 ** 18)
3354 3349 chunk = d.unconsumed_tail
3355 3350
3356 3351 return chunkbuffer(gen())
3357 3352
3358 3353 class zlibrevlogcompressor(object):
3359 3354 def compress(self, data):
3360 3355 insize = len(data)
3361 3356 # Caller handles empty input case.
3362 3357 assert insize > 0
3363 3358
3364 3359 if insize < 44:
3365 3360 return None
3366 3361
3367 3362 elif insize <= 1000000:
3368 3363 compressed = zlib.compress(data)
3369 3364 if len(compressed) < insize:
3370 3365 return compressed
3371 3366 return None
3372 3367
3373 3368 # zlib makes an internal copy of the input buffer, doubling
3374 3369 # memory usage for large inputs. So do streaming compression
3375 3370 # on large inputs.
3376 3371 else:
3377 3372 z = zlib.compressobj()
3378 3373 parts = []
3379 3374 pos = 0
3380 3375 while pos < insize:
3381 3376 pos2 = pos + 2**20
3382 3377 parts.append(z.compress(data[pos:pos2]))
3383 3378 pos = pos2
3384 3379 parts.append(z.flush())
3385 3380
3386 3381 if sum(map(len, parts)) < insize:
3387 3382 return ''.join(parts)
3388 3383 return None
3389 3384
3390 3385 def decompress(self, data):
3391 3386 try:
3392 3387 return zlib.decompress(data)
3393 3388 except zlib.error as e:
3394 3389 raise error.RevlogError(_('revlog decompress error: %s') %
3395 3390 str(e))
3396 3391
3397 3392 def revlogcompressor(self, opts=None):
3398 3393 return self.zlibrevlogcompressor()
3399 3394
3400 3395 compengines.register(_zlibengine())
3401 3396
3402 3397 class _bz2engine(compressionengine):
3403 3398 def name(self):
3404 3399 return 'bz2'
3405 3400
3406 3401 def bundletype(self):
3407 3402 return 'bzip2', 'BZ'
3408 3403
3409 3404 # We declare a protocol name but don't advertise by default because
3410 3405 # it is slow.
3411 3406 def wireprotosupport(self):
3412 3407 return compewireprotosupport('bzip2', 0, 0)
3413 3408
3414 3409 def compressstream(self, it, opts=None):
3415 3410 opts = opts or {}
3416 3411 z = bz2.BZ2Compressor(opts.get('level', 9))
3417 3412 for chunk in it:
3418 3413 data = z.compress(chunk)
3419 3414 if data:
3420 3415 yield data
3421 3416
3422 3417 yield z.flush()
3423 3418
3424 3419 def decompressorreader(self, fh):
3425 3420 def gen():
3426 3421 d = bz2.BZ2Decompressor()
3427 3422 for chunk in filechunkiter(fh):
3428 3423 yield d.decompress(chunk)
3429 3424
3430 3425 return chunkbuffer(gen())
3431 3426
3432 3427 compengines.register(_bz2engine())
3433 3428
3434 3429 class _truncatedbz2engine(compressionengine):
3435 3430 def name(self):
3436 3431 return 'bz2truncated'
3437 3432
3438 3433 def bundletype(self):
3439 3434 return None, '_truncatedBZ'
3440 3435
3441 3436 # We don't implement compressstream because it is hackily handled elsewhere.
3442 3437
3443 3438 def decompressorreader(self, fh):
3444 3439 def gen():
3445 3440 # The input stream doesn't have the 'BZ' header. So add it back.
3446 3441 d = bz2.BZ2Decompressor()
3447 3442 d.decompress('BZ')
3448 3443 for chunk in filechunkiter(fh):
3449 3444 yield d.decompress(chunk)
3450 3445
3451 3446 return chunkbuffer(gen())
3452 3447
3453 3448 compengines.register(_truncatedbz2engine())
3454 3449
3455 3450 class _noopengine(compressionengine):
3456 3451 def name(self):
3457 3452 return 'none'
3458 3453
3459 3454 def bundletype(self):
3460 3455 return 'none', 'UN'
3461 3456
3462 3457 # Clients always support uncompressed payloads. Servers don't because
3463 3458 # unless you are on a fast network, uncompressed payloads can easily
3464 3459 # saturate your network pipe.
3465 3460 def wireprotosupport(self):
3466 3461 return compewireprotosupport('none', 0, 10)
3467 3462
3468 3463 # We don't implement revlogheader because it is handled specially
3469 3464 # in the revlog class.
3470 3465
3471 3466 def compressstream(self, it, opts=None):
3472 3467 return it
3473 3468
3474 3469 def decompressorreader(self, fh):
3475 3470 return fh
3476 3471
3477 3472 class nooprevlogcompressor(object):
3478 3473 def compress(self, data):
3479 3474 return None
3480 3475
3481 3476 def revlogcompressor(self, opts=None):
3482 3477 return self.nooprevlogcompressor()
3483 3478
3484 3479 compengines.register(_noopengine())
3485 3480
3486 3481 class _zstdengine(compressionengine):
3487 3482 def name(self):
3488 3483 return 'zstd'
3489 3484
3490 3485 @propertycache
3491 3486 def _module(self):
3492 3487 # Not all installs have the zstd module available. So defer importing
3493 3488 # until first access.
3494 3489 try:
3495 3490 from . import zstd
3496 3491 # Force delayed import.
3497 3492 zstd.__version__
3498 3493 return zstd
3499 3494 except ImportError:
3500 3495 return None
3501 3496
3502 3497 def available(self):
3503 3498 return bool(self._module)
3504 3499
3505 3500 def bundletype(self):
3506 3501 return 'zstd', 'ZS'
3507 3502
3508 3503 def wireprotosupport(self):
3509 3504 return compewireprotosupport('zstd', 50, 50)
3510 3505
3511 3506 def revlogheader(self):
3512 3507 return '\x28'
3513 3508
3514 3509 def compressstream(self, it, opts=None):
3515 3510 opts = opts or {}
3516 3511 # zstd level 3 is almost always significantly faster than zlib
3517 3512 # while providing no worse compression. It strikes a good balance
3518 3513 # between speed and compression.
3519 3514 level = opts.get('level', 3)
3520 3515
3521 3516 zstd = self._module
3522 3517 z = zstd.ZstdCompressor(level=level).compressobj()
3523 3518 for chunk in it:
3524 3519 data = z.compress(chunk)
3525 3520 if data:
3526 3521 yield data
3527 3522
3528 3523 yield z.flush()
3529 3524
3530 3525 def decompressorreader(self, fh):
3531 3526 zstd = self._module
3532 3527 dctx = zstd.ZstdDecompressor()
3533 3528 return chunkbuffer(dctx.read_from(fh))
3534 3529
3535 3530 class zstdrevlogcompressor(object):
3536 3531 def __init__(self, zstd, level=3):
3537 3532 # Writing the content size adds a few bytes to the output. However,
3538 3533 # it allows decompression to be more optimal since we can
3539 3534 # pre-allocate a buffer to hold the result.
3540 3535 self._cctx = zstd.ZstdCompressor(level=level,
3541 3536 write_content_size=True)
3542 3537 self._dctx = zstd.ZstdDecompressor()
3543 3538 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3544 3539 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3545 3540
3546 3541 def compress(self, data):
3547 3542 insize = len(data)
3548 3543 # Caller handles empty input case.
3549 3544 assert insize > 0
3550 3545
3551 3546 if insize < 50:
3552 3547 return None
3553 3548
3554 3549 elif insize <= 1000000:
3555 3550 compressed = self._cctx.compress(data)
3556 3551 if len(compressed) < insize:
3557 3552 return compressed
3558 3553 return None
3559 3554 else:
3560 3555 z = self._cctx.compressobj()
3561 3556 chunks = []
3562 3557 pos = 0
3563 3558 while pos < insize:
3564 3559 pos2 = pos + self._compinsize
3565 3560 chunk = z.compress(data[pos:pos2])
3566 3561 if chunk:
3567 3562 chunks.append(chunk)
3568 3563 pos = pos2
3569 3564 chunks.append(z.flush())
3570 3565
3571 3566 if sum(map(len, chunks)) < insize:
3572 3567 return ''.join(chunks)
3573 3568 return None
3574 3569
3575 3570 def decompress(self, data):
3576 3571 insize = len(data)
3577 3572
3578 3573 try:
3579 3574 # This was measured to be faster than other streaming
3580 3575 # decompressors.
3581 3576 dobj = self._dctx.decompressobj()
3582 3577 chunks = []
3583 3578 pos = 0
3584 3579 while pos < insize:
3585 3580 pos2 = pos + self._decompinsize
3586 3581 chunk = dobj.decompress(data[pos:pos2])
3587 3582 if chunk:
3588 3583 chunks.append(chunk)
3589 3584 pos = pos2
3590 3585 # Frame should be exhausted, so no finish() API.
3591 3586
3592 3587 return ''.join(chunks)
3593 3588 except Exception as e:
3594 3589 raise error.RevlogError(_('revlog decompress error: %s') %
3595 3590 str(e))
3596 3591
3597 3592 def revlogcompressor(self, opts=None):
3598 3593 opts = opts or {}
3599 3594 return self.zstdrevlogcompressor(self._module,
3600 3595 level=opts.get('level', 3))
3601 3596
3602 3597 compengines.register(_zstdengine())
3603 3598
3604 3599 # convenient shortcut
3605 3600 dst = debugstacktrace
@@ -1,383 +1,383
1 1 #require hardlink
2 2
3 3 $ cat > nlinks.py <<EOF
4 4 > import sys
5 5 > from mercurial import util
6 6 > for f in sorted(sys.stdin.readlines()):
7 7 > f = f[:-1]
8 8 > print util.nlinks(f), f
9 9 > EOF
10 10
11 11 $ nlinksdir()
12 12 > {
13 13 > find $1 -type f | python $TESTTMP/nlinks.py
14 14 > }
15 15
16 16 Some implementations of cp can't create hardlinks (replaces 'cp -al' on Linux):
17 17
18 18 $ cat > linkcp.py <<EOF
19 19 > from mercurial import util
20 20 > import sys
21 21 > util.copyfiles(sys.argv[1], sys.argv[2], hardlink=True)
22 22 > EOF
23 23
24 24 $ linkcp()
25 25 > {
26 26 > python $TESTTMP/linkcp.py $1 $2
27 27 > }
28 28
29 29 Prepare repo r1:
30 30
31 31 $ hg init r1
32 32 $ cd r1
33 33
34 34 $ echo c1 > f1
35 35 $ hg add f1
36 36 $ hg ci -m0
37 37
38 38 $ mkdir d1
39 39 $ cd d1
40 40 $ echo c2 > f2
41 41 $ hg add f2
42 42 $ hg ci -m1
43 43 $ cd ../..
44 44
45 45 $ nlinksdir r1/.hg/store
46 46 1 r1/.hg/store/00changelog.i
47 47 1 r1/.hg/store/00manifest.i
48 48 1 r1/.hg/store/data/d1/f2.i
49 49 1 r1/.hg/store/data/f1.i
50 50 1 r1/.hg/store/fncache
51 51 1 r1/.hg/store/phaseroots
52 52 1 r1/.hg/store/undo
53 53 1 r1/.hg/store/undo.backup.fncache
54 54 1 r1/.hg/store/undo.backupfiles
55 55 1 r1/.hg/store/undo.phaseroots
56 56
57 57
58 58 Create hardlinked clone r2:
59 59
60 60 $ hg clone -U --debug r1 r2 --config progress.debug=true
61 61 linking: 1
62 62 linking: 2
63 63 linking: 3
64 64 linking: 4
65 65 linking: 5
66 66 linking: 6
67 67 linking: 7
68 68 linked 7 files
69 69
70 70 Create non-hardlinked clone r3:
71 71
72 72 $ hg clone --pull r1 r3
73 73 requesting all changes
74 74 adding changesets
75 75 adding manifests
76 76 adding file changes
77 77 added 2 changesets with 2 changes to 2 files
78 78 updating to branch default
79 79 2 files updated, 0 files merged, 0 files removed, 0 files unresolved
80 80
81 81
82 82 Repos r1 and r2 should now contain hardlinked files:
83 83
84 84 $ nlinksdir r1/.hg/store
85 85 2 r1/.hg/store/00changelog.i
86 86 2 r1/.hg/store/00manifest.i
87 87 2 r1/.hg/store/data/d1/f2.i
88 88 2 r1/.hg/store/data/f1.i
89 89 2 r1/.hg/store/fncache
90 90 1 r1/.hg/store/phaseroots
91 91 1 r1/.hg/store/undo
92 92 1 r1/.hg/store/undo.backup.fncache
93 93 1 r1/.hg/store/undo.backupfiles
94 94 1 r1/.hg/store/undo.phaseroots
95 95
96 96 $ nlinksdir r2/.hg/store
97 97 2 r2/.hg/store/00changelog.i
98 98 2 r2/.hg/store/00manifest.i
99 99 2 r2/.hg/store/data/d1/f2.i
100 100 2 r2/.hg/store/data/f1.i
101 101 2 r2/.hg/store/fncache
102 102
103 103 Repo r3 should not be hardlinked:
104 104
105 105 $ nlinksdir r3/.hg/store
106 106 1 r3/.hg/store/00changelog.i
107 107 1 r3/.hg/store/00manifest.i
108 108 1 r3/.hg/store/data/d1/f2.i
109 109 1 r3/.hg/store/data/f1.i
110 110 1 r3/.hg/store/fncache
111 111 1 r3/.hg/store/phaseroots
112 112 1 r3/.hg/store/undo
113 113 1 r3/.hg/store/undo.backupfiles
114 114 1 r3/.hg/store/undo.phaseroots
115 115
116 116
117 117 Create a non-inlined filelog in r3:
118 118
119 119 $ cd r3/d1
120 120 >>> f = open('data1', 'wb')
121 121 >>> for x in range(10000):
122 122 ... f.write("%s\n" % str(x))
123 123 >>> f.close()
124 124 $ for j in 0 1 2 3 4 5 6 7 8 9; do
125 125 > cat data1 >> f2
126 126 > hg commit -m$j
127 127 > done
128 128 $ cd ../..
129 129
130 130 $ nlinksdir r3/.hg/store
131 131 1 r3/.hg/store/00changelog.i
132 132 1 r3/.hg/store/00manifest.i
133 133 1 r3/.hg/store/data/d1/f2.d
134 134 1 r3/.hg/store/data/d1/f2.i
135 135 1 r3/.hg/store/data/f1.i
136 136 1 r3/.hg/store/fncache
137 137 1 r3/.hg/store/phaseroots
138 138 1 r3/.hg/store/undo
139 139 1 r3/.hg/store/undo.backup.fncache
140 140 1 r3/.hg/store/undo.backup.phaseroots
141 141 1 r3/.hg/store/undo.backupfiles
142 142 1 r3/.hg/store/undo.phaseroots
143 143
144 144 Push to repo r1 should break up most hardlinks in r2:
145 145
146 146 $ hg -R r2 verify
147 147 checking changesets
148 148 checking manifests
149 149 crosschecking files in changesets and manifests
150 150 checking files
151 151 2 files, 2 changesets, 2 total revisions
152 152
153 153 $ cd r3
154 154 $ hg push
155 155 pushing to $TESTTMP/r1 (glob)
156 156 searching for changes
157 157 adding changesets
158 158 adding manifests
159 159 adding file changes
160 160 added 10 changesets with 10 changes to 1 files
161 161
162 162 $ cd ..
163 163
164 164 $ nlinksdir r2/.hg/store
165 165 1 r2/.hg/store/00changelog.i
166 166 1 r2/.hg/store/00manifest.i
167 167 1 r2/.hg/store/data/d1/f2.i
168 168 2 r2/.hg/store/data/f1.i
169 1 r2/.hg/store/fncache
169 [12] r2/\.hg/store/fncache (re)
170 170
171 171 $ hg -R r2 verify
172 172 checking changesets
173 173 checking manifests
174 174 crosschecking files in changesets and manifests
175 175 checking files
176 176 2 files, 2 changesets, 2 total revisions
177 177
178 178
179 179 $ cd r1
180 180 $ hg up
181 181 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
182 182
183 183 Committing a change to f1 in r1 must break up hardlink f1.i in r2:
184 184
185 185 $ echo c1c1 >> f1
186 186 $ hg ci -m00
187 187 $ cd ..
188 188
189 189 $ nlinksdir r2/.hg/store
190 190 1 r2/.hg/store/00changelog.i
191 191 1 r2/.hg/store/00manifest.i
192 192 1 r2/.hg/store/data/d1/f2.i
193 193 1 r2/.hg/store/data/f1.i
194 1 r2/.hg/store/fncache
194 [12] r2/\.hg/store/fncache (re)
195 195
196 196
197 197 $ cd r3
198 198 $ hg tip --template '{rev}:{node|short}\n'
199 199 11:a6451b6bc41f
200 200 $ echo bla > f1
201 201 $ hg ci -m1
202 202 $ cd ..
203 203
204 204 Create hardlinked copy r4 of r3 (on Linux, we would call 'cp -al'):
205 205
206 206 $ linkcp r3 r4
207 207
208 208 r4 has hardlinks in the working dir (not just inside .hg):
209 209
210 210 $ nlinksdir r4
211 211 2 r4/.hg/00changelog.i
212 212 2 r4/.hg/branch
213 213 2 r4/.hg/cache/branch2-served
214 214 2 r4/.hg/cache/checkisexec
215 215 3 r4/.hg/cache/checklink (?)
216 216 ? r4/.hg/cache/checklink-target (glob)
217 217 2 r4/.hg/cache/checknoexec
218 218 2 r4/.hg/cache/rbc-names-v1
219 219 2 r4/.hg/cache/rbc-revs-v1
220 220 2 r4/.hg/dirstate
221 221 2 r4/.hg/hgrc
222 222 2 r4/.hg/last-message.txt
223 223 2 r4/.hg/requires
224 224 2 r4/.hg/store/00changelog.i
225 225 2 r4/.hg/store/00manifest.i
226 226 2 r4/.hg/store/data/d1/f2.d
227 227 2 r4/.hg/store/data/d1/f2.i
228 228 2 r4/.hg/store/data/f1.i
229 229 2 r4/.hg/store/fncache
230 230 2 r4/.hg/store/phaseroots
231 231 2 r4/.hg/store/undo
232 232 2 r4/.hg/store/undo.backup.fncache
233 233 2 r4/.hg/store/undo.backup.phaseroots
234 234 2 r4/.hg/store/undo.backupfiles
235 235 2 r4/.hg/store/undo.phaseroots
236 2 r4/.hg/undo.backup.dirstate
236 [24] r4/\.hg/undo\.backup\.dirstate (re)
237 237 2 r4/.hg/undo.bookmarks
238 238 2 r4/.hg/undo.branch
239 239 2 r4/.hg/undo.desc
240 2 r4/.hg/undo.dirstate
240 [24] r4/\.hg/undo\.dirstate (re)
241 241 2 r4/d1/data1
242 242 2 r4/d1/f2
243 243 2 r4/f1
244 244
245 245 Update back to revision 11 in r4 should break hardlink of file f1:
246 246
247 247 $ hg -R r4 up 11
248 248 1 files updated, 0 files merged, 0 files removed, 0 files unresolved
249 249
250 250 $ nlinksdir r4
251 251 2 r4/.hg/00changelog.i
252 252 1 r4/.hg/branch
253 253 2 r4/.hg/cache/branch2-served
254 254 2 r4/.hg/cache/checkisexec
255 255 2 r4/.hg/cache/checklink-target
256 256 2 r4/.hg/cache/checknoexec
257 257 2 r4/.hg/cache/rbc-names-v1
258 258 2 r4/.hg/cache/rbc-revs-v1
259 259 1 r4/.hg/dirstate
260 260 2 r4/.hg/hgrc
261 261 2 r4/.hg/last-message.txt
262 262 2 r4/.hg/requires
263 263 2 r4/.hg/store/00changelog.i
264 264 2 r4/.hg/store/00manifest.i
265 265 2 r4/.hg/store/data/d1/f2.d
266 266 2 r4/.hg/store/data/d1/f2.i
267 267 2 r4/.hg/store/data/f1.i
268 268 2 r4/.hg/store/fncache
269 269 2 r4/.hg/store/phaseroots
270 270 2 r4/.hg/store/undo
271 271 2 r4/.hg/store/undo.backup.fncache
272 272 2 r4/.hg/store/undo.backup.phaseroots
273 273 2 r4/.hg/store/undo.backupfiles
274 274 2 r4/.hg/store/undo.phaseroots
275 2 r4/.hg/undo.backup.dirstate
275 [24] r4/\.hg/undo\.backup\.dirstate (re)
276 276 2 r4/.hg/undo.bookmarks
277 277 2 r4/.hg/undo.branch
278 278 2 r4/.hg/undo.desc
279 2 r4/.hg/undo.dirstate
279 [24] r4/\.hg/undo\.dirstate (re)
280 280 2 r4/d1/data1
281 281 2 r4/d1/f2
282 282 1 r4/f1
283 283
284 284
285 285 Test hardlinking outside hg:
286 286
287 287 $ mkdir x
288 288 $ echo foo > x/a
289 289
290 290 $ linkcp x y
291 291 $ echo bar >> y/a
292 292
293 293 No diff if hardlink:
294 294
295 295 $ diff x/a y/a
296 296
297 297 Test mq hardlinking:
298 298
299 299 $ echo "[extensions]" >> $HGRCPATH
300 300 $ echo "mq=" >> $HGRCPATH
301 301
302 302 $ hg init a
303 303 $ cd a
304 304
305 305 $ hg qimport -n foo - << EOF
306 306 > # HG changeset patch
307 307 > # Date 1 0
308 308 > diff -r 2588a8b53d66 a
309 309 > --- /dev/null Thu Jan 01 00:00:00 1970 +0000
310 310 > +++ b/a Wed Jul 23 15:54:29 2008 +0200
311 311 > @@ -0,0 +1,1 @@
312 312 > +a
313 313 > EOF
314 314 adding foo to series file
315 315
316 316 $ hg qpush
317 317 applying foo
318 318 now at: foo
319 319
320 320 $ cd ..
321 321 $ linkcp a b
322 322 $ cd b
323 323
324 324 $ hg qimport -n bar - << EOF
325 325 > # HG changeset patch
326 326 > # Date 2 0
327 327 > diff -r 2588a8b53d66 a
328 328 > --- /dev/null Thu Jan 01 00:00:00 1970 +0000
329 329 > +++ b/b Wed Jul 23 15:54:29 2008 +0200
330 330 > @@ -0,0 +1,1 @@
331 331 > +b
332 332 > EOF
333 333 adding bar to series file
334 334
335 335 $ hg qpush
336 336 applying bar
337 337 now at: bar
338 338
339 339 $ cat .hg/patches/status
340 340 430ed4828a74fa4047bc816a25500f7472ab4bfe:foo
341 341 4e7abb4840c46a910f6d7b4d3c3fc7e5209e684c:bar
342 342
343 343 $ cat .hg/patches/series
344 344 foo
345 345 bar
346 346
347 347 $ cat ../a/.hg/patches/status
348 348 430ed4828a74fa4047bc816a25500f7472ab4bfe:foo
349 349
350 350 $ cat ../a/.hg/patches/series
351 351 foo
352 352
353 353 Test tags hardlinking:
354 354
355 355 $ hg qdel -r qbase:qtip
356 356 patch foo finalized without changeset message
357 357 patch bar finalized without changeset message
358 358
359 359 $ hg tag -l lfoo
360 360 $ hg tag foo
361 361
362 362 $ cd ..
363 363 $ linkcp b c
364 364 $ cd c
365 365
366 366 $ hg tag -l -r 0 lbar
367 367 $ hg tag -r 0 bar
368 368
369 369 $ cat .hgtags
370 370 4e7abb4840c46a910f6d7b4d3c3fc7e5209e684c foo
371 371 430ed4828a74fa4047bc816a25500f7472ab4bfe bar
372 372
373 373 $ cat .hg/localtags
374 374 4e7abb4840c46a910f6d7b4d3c3fc7e5209e684c lfoo
375 375 430ed4828a74fa4047bc816a25500f7472ab4bfe lbar
376 376
377 377 $ cat ../b/.hgtags
378 378 4e7abb4840c46a910f6d7b4d3c3fc7e5209e684c foo
379 379
380 380 $ cat ../b/.hg/localtags
381 381 4e7abb4840c46a910f6d7b4d3c3fc7e5209e684c lfoo
382 382
383 383 $ cd ..
General Comments 0
You need to be logged in to leave comments. Login now