##// END OF EJS Templates
util: fix %-formatting on docstring by moving a closing parenthesis...
Augie Fackler -
r31811:d955b02d default
parent child Browse files
Show More
@@ -1,3719 +1,3719
1 # util.py - Mercurial utility functions and platform specific implementations
1 # util.py - Mercurial utility functions and platform specific implementations
2 #
2 #
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 #
6 #
7 # This software may be used and distributed according to the terms of the
7 # This software may be used and distributed according to the terms of the
8 # GNU General Public License version 2 or any later version.
8 # GNU General Public License version 2 or any later version.
9
9
10 """Mercurial utility functions and platform specific implementations.
10 """Mercurial utility functions and platform specific implementations.
11
11
12 This contains helper routines that are independent of the SCM core and
12 This contains helper routines that are independent of the SCM core and
13 hide platform-specific details from the core.
13 hide platform-specific details from the core.
14 """
14 """
15
15
16 from __future__ import absolute_import
16 from __future__ import absolute_import
17
17
18 import bz2
18 import bz2
19 import calendar
19 import calendar
20 import codecs
20 import codecs
21 import collections
21 import collections
22 import datetime
22 import datetime
23 import errno
23 import errno
24 import gc
24 import gc
25 import hashlib
25 import hashlib
26 import imp
26 import imp
27 import os
27 import os
28 import platform as pyplatform
28 import platform as pyplatform
29 import re as remod
29 import re as remod
30 import shutil
30 import shutil
31 import signal
31 import signal
32 import socket
32 import socket
33 import stat
33 import stat
34 import string
34 import string
35 import subprocess
35 import subprocess
36 import sys
36 import sys
37 import tempfile
37 import tempfile
38 import textwrap
38 import textwrap
39 import time
39 import time
40 import traceback
40 import traceback
41 import zlib
41 import zlib
42
42
43 from . import (
43 from . import (
44 encoding,
44 encoding,
45 error,
45 error,
46 i18n,
46 i18n,
47 osutil,
47 osutil,
48 parsers,
48 parsers,
49 pycompat,
49 pycompat,
50 )
50 )
51
51
52 empty = pycompat.empty
52 empty = pycompat.empty
53 httplib = pycompat.httplib
53 httplib = pycompat.httplib
54 httpserver = pycompat.httpserver
54 httpserver = pycompat.httpserver
55 pickle = pycompat.pickle
55 pickle = pycompat.pickle
56 queue = pycompat.queue
56 queue = pycompat.queue
57 socketserver = pycompat.socketserver
57 socketserver = pycompat.socketserver
58 stderr = pycompat.stderr
58 stderr = pycompat.stderr
59 stdin = pycompat.stdin
59 stdin = pycompat.stdin
60 stdout = pycompat.stdout
60 stdout = pycompat.stdout
61 stringio = pycompat.stringio
61 stringio = pycompat.stringio
62 urlerr = pycompat.urlerr
62 urlerr = pycompat.urlerr
63 urlreq = pycompat.urlreq
63 urlreq = pycompat.urlreq
64 xmlrpclib = pycompat.xmlrpclib
64 xmlrpclib = pycompat.xmlrpclib
65
65
66 def isatty(fp):
66 def isatty(fp):
67 try:
67 try:
68 return fp.isatty()
68 return fp.isatty()
69 except AttributeError:
69 except AttributeError:
70 return False
70 return False
71
71
72 # glibc determines buffering on first write to stdout - if we replace a TTY
72 # glibc determines buffering on first write to stdout - if we replace a TTY
73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
74 # buffering
74 # buffering
75 if isatty(stdout):
75 if isatty(stdout):
76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
77
77
78 if pycompat.osname == 'nt':
78 if pycompat.osname == 'nt':
79 from . import windows as platform
79 from . import windows as platform
80 stdout = platform.winstdout(stdout)
80 stdout = platform.winstdout(stdout)
81 else:
81 else:
82 from . import posix as platform
82 from . import posix as platform
83
83
84 _ = i18n._
84 _ = i18n._
85
85
86 bindunixsocket = platform.bindunixsocket
86 bindunixsocket = platform.bindunixsocket
87 cachestat = platform.cachestat
87 cachestat = platform.cachestat
88 checkexec = platform.checkexec
88 checkexec = platform.checkexec
89 checklink = platform.checklink
89 checklink = platform.checklink
90 copymode = platform.copymode
90 copymode = platform.copymode
91 executablepath = platform.executablepath
91 executablepath = platform.executablepath
92 expandglobs = platform.expandglobs
92 expandglobs = platform.expandglobs
93 explainexit = platform.explainexit
93 explainexit = platform.explainexit
94 findexe = platform.findexe
94 findexe = platform.findexe
95 gethgcmd = platform.gethgcmd
95 gethgcmd = platform.gethgcmd
96 getuser = platform.getuser
96 getuser = platform.getuser
97 getpid = os.getpid
97 getpid = os.getpid
98 groupmembers = platform.groupmembers
98 groupmembers = platform.groupmembers
99 groupname = platform.groupname
99 groupname = platform.groupname
100 hidewindow = platform.hidewindow
100 hidewindow = platform.hidewindow
101 isexec = platform.isexec
101 isexec = platform.isexec
102 isowner = platform.isowner
102 isowner = platform.isowner
103 localpath = platform.localpath
103 localpath = platform.localpath
104 lookupreg = platform.lookupreg
104 lookupreg = platform.lookupreg
105 makedir = platform.makedir
105 makedir = platform.makedir
106 nlinks = platform.nlinks
106 nlinks = platform.nlinks
107 normpath = platform.normpath
107 normpath = platform.normpath
108 normcase = platform.normcase
108 normcase = platform.normcase
109 normcasespec = platform.normcasespec
109 normcasespec = platform.normcasespec
110 normcasefallback = platform.normcasefallback
110 normcasefallback = platform.normcasefallback
111 openhardlinks = platform.openhardlinks
111 openhardlinks = platform.openhardlinks
112 oslink = platform.oslink
112 oslink = platform.oslink
113 parsepatchoutput = platform.parsepatchoutput
113 parsepatchoutput = platform.parsepatchoutput
114 pconvert = platform.pconvert
114 pconvert = platform.pconvert
115 poll = platform.poll
115 poll = platform.poll
116 popen = platform.popen
116 popen = platform.popen
117 posixfile = platform.posixfile
117 posixfile = platform.posixfile
118 quotecommand = platform.quotecommand
118 quotecommand = platform.quotecommand
119 readpipe = platform.readpipe
119 readpipe = platform.readpipe
120 rename = platform.rename
120 rename = platform.rename
121 removedirs = platform.removedirs
121 removedirs = platform.removedirs
122 samedevice = platform.samedevice
122 samedevice = platform.samedevice
123 samefile = platform.samefile
123 samefile = platform.samefile
124 samestat = platform.samestat
124 samestat = platform.samestat
125 setbinary = platform.setbinary
125 setbinary = platform.setbinary
126 setflags = platform.setflags
126 setflags = platform.setflags
127 setsignalhandler = platform.setsignalhandler
127 setsignalhandler = platform.setsignalhandler
128 shellquote = platform.shellquote
128 shellquote = platform.shellquote
129 spawndetached = platform.spawndetached
129 spawndetached = platform.spawndetached
130 split = platform.split
130 split = platform.split
131 sshargs = platform.sshargs
131 sshargs = platform.sshargs
132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
133 statisexec = platform.statisexec
133 statisexec = platform.statisexec
134 statislink = platform.statislink
134 statislink = platform.statislink
135 testpid = platform.testpid
135 testpid = platform.testpid
136 umask = platform.umask
136 umask = platform.umask
137 unlink = platform.unlink
137 unlink = platform.unlink
138 username = platform.username
138 username = platform.username
139
139
140 # Python compatibility
140 # Python compatibility
141
141
142 _notset = object()
142 _notset = object()
143
143
144 # disable Python's problematic floating point timestamps (issue4836)
144 # disable Python's problematic floating point timestamps (issue4836)
145 # (Python hypocritically says you shouldn't change this behavior in
145 # (Python hypocritically says you shouldn't change this behavior in
146 # libraries, and sure enough Mercurial is not a library.)
146 # libraries, and sure enough Mercurial is not a library.)
147 os.stat_float_times(False)
147 os.stat_float_times(False)
148
148
149 def safehasattr(thing, attr):
149 def safehasattr(thing, attr):
150 return getattr(thing, attr, _notset) is not _notset
150 return getattr(thing, attr, _notset) is not _notset
151
151
152 def bitsfrom(container):
152 def bitsfrom(container):
153 bits = 0
153 bits = 0
154 for bit in container:
154 for bit in container:
155 bits |= bit
155 bits |= bit
156 return bits
156 return bits
157
157
158 DIGESTS = {
158 DIGESTS = {
159 'md5': hashlib.md5,
159 'md5': hashlib.md5,
160 'sha1': hashlib.sha1,
160 'sha1': hashlib.sha1,
161 'sha512': hashlib.sha512,
161 'sha512': hashlib.sha512,
162 }
162 }
163 # List of digest types from strongest to weakest
163 # List of digest types from strongest to weakest
164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
165
165
166 for k in DIGESTS_BY_STRENGTH:
166 for k in DIGESTS_BY_STRENGTH:
167 assert k in DIGESTS
167 assert k in DIGESTS
168
168
169 class digester(object):
169 class digester(object):
170 """helper to compute digests.
170 """helper to compute digests.
171
171
172 This helper can be used to compute one or more digests given their name.
172 This helper can be used to compute one or more digests given their name.
173
173
174 >>> d = digester(['md5', 'sha1'])
174 >>> d = digester(['md5', 'sha1'])
175 >>> d.update('foo')
175 >>> d.update('foo')
176 >>> [k for k in sorted(d)]
176 >>> [k for k in sorted(d)]
177 ['md5', 'sha1']
177 ['md5', 'sha1']
178 >>> d['md5']
178 >>> d['md5']
179 'acbd18db4cc2f85cedef654fccc4a4d8'
179 'acbd18db4cc2f85cedef654fccc4a4d8'
180 >>> d['sha1']
180 >>> d['sha1']
181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
182 >>> digester.preferred(['md5', 'sha1'])
182 >>> digester.preferred(['md5', 'sha1'])
183 'sha1'
183 'sha1'
184 """
184 """
185
185
186 def __init__(self, digests, s=''):
186 def __init__(self, digests, s=''):
187 self._hashes = {}
187 self._hashes = {}
188 for k in digests:
188 for k in digests:
189 if k not in DIGESTS:
189 if k not in DIGESTS:
190 raise Abort(_('unknown digest type: %s') % k)
190 raise Abort(_('unknown digest type: %s') % k)
191 self._hashes[k] = DIGESTS[k]()
191 self._hashes[k] = DIGESTS[k]()
192 if s:
192 if s:
193 self.update(s)
193 self.update(s)
194
194
195 def update(self, data):
195 def update(self, data):
196 for h in self._hashes.values():
196 for h in self._hashes.values():
197 h.update(data)
197 h.update(data)
198
198
199 def __getitem__(self, key):
199 def __getitem__(self, key):
200 if key not in DIGESTS:
200 if key not in DIGESTS:
201 raise Abort(_('unknown digest type: %s') % k)
201 raise Abort(_('unknown digest type: %s') % k)
202 return self._hashes[key].hexdigest()
202 return self._hashes[key].hexdigest()
203
203
204 def __iter__(self):
204 def __iter__(self):
205 return iter(self._hashes)
205 return iter(self._hashes)
206
206
207 @staticmethod
207 @staticmethod
208 def preferred(supported):
208 def preferred(supported):
209 """returns the strongest digest type in both supported and DIGESTS."""
209 """returns the strongest digest type in both supported and DIGESTS."""
210
210
211 for k in DIGESTS_BY_STRENGTH:
211 for k in DIGESTS_BY_STRENGTH:
212 if k in supported:
212 if k in supported:
213 return k
213 return k
214 return None
214 return None
215
215
216 class digestchecker(object):
216 class digestchecker(object):
217 """file handle wrapper that additionally checks content against a given
217 """file handle wrapper that additionally checks content against a given
218 size and digests.
218 size and digests.
219
219
220 d = digestchecker(fh, size, {'md5': '...'})
220 d = digestchecker(fh, size, {'md5': '...'})
221
221
222 When multiple digests are given, all of them are validated.
222 When multiple digests are given, all of them are validated.
223 """
223 """
224
224
225 def __init__(self, fh, size, digests):
225 def __init__(self, fh, size, digests):
226 self._fh = fh
226 self._fh = fh
227 self._size = size
227 self._size = size
228 self._got = 0
228 self._got = 0
229 self._digests = dict(digests)
229 self._digests = dict(digests)
230 self._digester = digester(self._digests.keys())
230 self._digester = digester(self._digests.keys())
231
231
232 def read(self, length=-1):
232 def read(self, length=-1):
233 content = self._fh.read(length)
233 content = self._fh.read(length)
234 self._digester.update(content)
234 self._digester.update(content)
235 self._got += len(content)
235 self._got += len(content)
236 return content
236 return content
237
237
238 def validate(self):
238 def validate(self):
239 if self._size != self._got:
239 if self._size != self._got:
240 raise Abort(_('size mismatch: expected %d, got %d') %
240 raise Abort(_('size mismatch: expected %d, got %d') %
241 (self._size, self._got))
241 (self._size, self._got))
242 for k, v in self._digests.items():
242 for k, v in self._digests.items():
243 if v != self._digester[k]:
243 if v != self._digester[k]:
244 # i18n: first parameter is a digest name
244 # i18n: first parameter is a digest name
245 raise Abort(_('%s mismatch: expected %s, got %s') %
245 raise Abort(_('%s mismatch: expected %s, got %s') %
246 (k, v, self._digester[k]))
246 (k, v, self._digester[k]))
247
247
248 try:
248 try:
249 buffer = buffer
249 buffer = buffer
250 except NameError:
250 except NameError:
251 if not pycompat.ispy3:
251 if not pycompat.ispy3:
252 def buffer(sliceable, offset=0, length=None):
252 def buffer(sliceable, offset=0, length=None):
253 if length is not None:
253 if length is not None:
254 return sliceable[offset:offset + length]
254 return sliceable[offset:offset + length]
255 return sliceable[offset:]
255 return sliceable[offset:]
256 else:
256 else:
257 def buffer(sliceable, offset=0, length=None):
257 def buffer(sliceable, offset=0, length=None):
258 if length is not None:
258 if length is not None:
259 return memoryview(sliceable)[offset:offset + length]
259 return memoryview(sliceable)[offset:offset + length]
260 return memoryview(sliceable)[offset:]
260 return memoryview(sliceable)[offset:]
261
261
262 closefds = pycompat.osname == 'posix'
262 closefds = pycompat.osname == 'posix'
263
263
264 _chunksize = 4096
264 _chunksize = 4096
265
265
266 class bufferedinputpipe(object):
266 class bufferedinputpipe(object):
267 """a manually buffered input pipe
267 """a manually buffered input pipe
268
268
269 Python will not let us use buffered IO and lazy reading with 'polling' at
269 Python will not let us use buffered IO and lazy reading with 'polling' at
270 the same time. We cannot probe the buffer state and select will not detect
270 the same time. We cannot probe the buffer state and select will not detect
271 that data are ready to read if they are already buffered.
271 that data are ready to read if they are already buffered.
272
272
273 This class let us work around that by implementing its own buffering
273 This class let us work around that by implementing its own buffering
274 (allowing efficient readline) while offering a way to know if the buffer is
274 (allowing efficient readline) while offering a way to know if the buffer is
275 empty from the output (allowing collaboration of the buffer with polling).
275 empty from the output (allowing collaboration of the buffer with polling).
276
276
277 This class lives in the 'util' module because it makes use of the 'os'
277 This class lives in the 'util' module because it makes use of the 'os'
278 module from the python stdlib.
278 module from the python stdlib.
279 """
279 """
280
280
281 def __init__(self, input):
281 def __init__(self, input):
282 self._input = input
282 self._input = input
283 self._buffer = []
283 self._buffer = []
284 self._eof = False
284 self._eof = False
285 self._lenbuf = 0
285 self._lenbuf = 0
286
286
287 @property
287 @property
288 def hasbuffer(self):
288 def hasbuffer(self):
289 """True is any data is currently buffered
289 """True is any data is currently buffered
290
290
291 This will be used externally a pre-step for polling IO. If there is
291 This will be used externally a pre-step for polling IO. If there is
292 already data then no polling should be set in place."""
292 already data then no polling should be set in place."""
293 return bool(self._buffer)
293 return bool(self._buffer)
294
294
295 @property
295 @property
296 def closed(self):
296 def closed(self):
297 return self._input.closed
297 return self._input.closed
298
298
299 def fileno(self):
299 def fileno(self):
300 return self._input.fileno()
300 return self._input.fileno()
301
301
302 def close(self):
302 def close(self):
303 return self._input.close()
303 return self._input.close()
304
304
305 def read(self, size):
305 def read(self, size):
306 while (not self._eof) and (self._lenbuf < size):
306 while (not self._eof) and (self._lenbuf < size):
307 self._fillbuffer()
307 self._fillbuffer()
308 return self._frombuffer(size)
308 return self._frombuffer(size)
309
309
310 def readline(self, *args, **kwargs):
310 def readline(self, *args, **kwargs):
311 if 1 < len(self._buffer):
311 if 1 < len(self._buffer):
312 # this should not happen because both read and readline end with a
312 # this should not happen because both read and readline end with a
313 # _frombuffer call that collapse it.
313 # _frombuffer call that collapse it.
314 self._buffer = [''.join(self._buffer)]
314 self._buffer = [''.join(self._buffer)]
315 self._lenbuf = len(self._buffer[0])
315 self._lenbuf = len(self._buffer[0])
316 lfi = -1
316 lfi = -1
317 if self._buffer:
317 if self._buffer:
318 lfi = self._buffer[-1].find('\n')
318 lfi = self._buffer[-1].find('\n')
319 while (not self._eof) and lfi < 0:
319 while (not self._eof) and lfi < 0:
320 self._fillbuffer()
320 self._fillbuffer()
321 if self._buffer:
321 if self._buffer:
322 lfi = self._buffer[-1].find('\n')
322 lfi = self._buffer[-1].find('\n')
323 size = lfi + 1
323 size = lfi + 1
324 if lfi < 0: # end of file
324 if lfi < 0: # end of file
325 size = self._lenbuf
325 size = self._lenbuf
326 elif 1 < len(self._buffer):
326 elif 1 < len(self._buffer):
327 # we need to take previous chunks into account
327 # we need to take previous chunks into account
328 size += self._lenbuf - len(self._buffer[-1])
328 size += self._lenbuf - len(self._buffer[-1])
329 return self._frombuffer(size)
329 return self._frombuffer(size)
330
330
331 def _frombuffer(self, size):
331 def _frombuffer(self, size):
332 """return at most 'size' data from the buffer
332 """return at most 'size' data from the buffer
333
333
334 The data are removed from the buffer."""
334 The data are removed from the buffer."""
335 if size == 0 or not self._buffer:
335 if size == 0 or not self._buffer:
336 return ''
336 return ''
337 buf = self._buffer[0]
337 buf = self._buffer[0]
338 if 1 < len(self._buffer):
338 if 1 < len(self._buffer):
339 buf = ''.join(self._buffer)
339 buf = ''.join(self._buffer)
340
340
341 data = buf[:size]
341 data = buf[:size]
342 buf = buf[len(data):]
342 buf = buf[len(data):]
343 if buf:
343 if buf:
344 self._buffer = [buf]
344 self._buffer = [buf]
345 self._lenbuf = len(buf)
345 self._lenbuf = len(buf)
346 else:
346 else:
347 self._buffer = []
347 self._buffer = []
348 self._lenbuf = 0
348 self._lenbuf = 0
349 return data
349 return data
350
350
351 def _fillbuffer(self):
351 def _fillbuffer(self):
352 """read data to the buffer"""
352 """read data to the buffer"""
353 data = os.read(self._input.fileno(), _chunksize)
353 data = os.read(self._input.fileno(), _chunksize)
354 if not data:
354 if not data:
355 self._eof = True
355 self._eof = True
356 else:
356 else:
357 self._lenbuf += len(data)
357 self._lenbuf += len(data)
358 self._buffer.append(data)
358 self._buffer.append(data)
359
359
360 def popen2(cmd, env=None, newlines=False):
360 def popen2(cmd, env=None, newlines=False):
361 # Setting bufsize to -1 lets the system decide the buffer size.
361 # Setting bufsize to -1 lets the system decide the buffer size.
362 # The default for bufsize is 0, meaning unbuffered. This leads to
362 # The default for bufsize is 0, meaning unbuffered. This leads to
363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
365 close_fds=closefds,
365 close_fds=closefds,
366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
367 universal_newlines=newlines,
367 universal_newlines=newlines,
368 env=env)
368 env=env)
369 return p.stdin, p.stdout
369 return p.stdin, p.stdout
370
370
371 def popen3(cmd, env=None, newlines=False):
371 def popen3(cmd, env=None, newlines=False):
372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
373 return stdin, stdout, stderr
373 return stdin, stdout, stderr
374
374
375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
377 close_fds=closefds,
377 close_fds=closefds,
378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
379 stderr=subprocess.PIPE,
379 stderr=subprocess.PIPE,
380 universal_newlines=newlines,
380 universal_newlines=newlines,
381 env=env)
381 env=env)
382 return p.stdin, p.stdout, p.stderr, p
382 return p.stdin, p.stdout, p.stderr, p
383
383
384 def version():
384 def version():
385 """Return version information if available."""
385 """Return version information if available."""
386 try:
386 try:
387 from . import __version__
387 from . import __version__
388 return __version__.version
388 return __version__.version
389 except ImportError:
389 except ImportError:
390 return 'unknown'
390 return 'unknown'
391
391
392 def versiontuple(v=None, n=4):
392 def versiontuple(v=None, n=4):
393 """Parses a Mercurial version string into an N-tuple.
393 """Parses a Mercurial version string into an N-tuple.
394
394
395 The version string to be parsed is specified with the ``v`` argument.
395 The version string to be parsed is specified with the ``v`` argument.
396 If it isn't defined, the current Mercurial version string will be parsed.
396 If it isn't defined, the current Mercurial version string will be parsed.
397
397
398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
399 returned values:
399 returned values:
400
400
401 >>> v = '3.6.1+190-df9b73d2d444'
401 >>> v = '3.6.1+190-df9b73d2d444'
402 >>> versiontuple(v, 2)
402 >>> versiontuple(v, 2)
403 (3, 6)
403 (3, 6)
404 >>> versiontuple(v, 3)
404 >>> versiontuple(v, 3)
405 (3, 6, 1)
405 (3, 6, 1)
406 >>> versiontuple(v, 4)
406 >>> versiontuple(v, 4)
407 (3, 6, 1, '190-df9b73d2d444')
407 (3, 6, 1, '190-df9b73d2d444')
408
408
409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
410 (3, 6, 1, '190-df9b73d2d444+20151118')
410 (3, 6, 1, '190-df9b73d2d444+20151118')
411
411
412 >>> v = '3.6'
412 >>> v = '3.6'
413 >>> versiontuple(v, 2)
413 >>> versiontuple(v, 2)
414 (3, 6)
414 (3, 6)
415 >>> versiontuple(v, 3)
415 >>> versiontuple(v, 3)
416 (3, 6, None)
416 (3, 6, None)
417 >>> versiontuple(v, 4)
417 >>> versiontuple(v, 4)
418 (3, 6, None, None)
418 (3, 6, None, None)
419
419
420 >>> v = '3.9-rc'
420 >>> v = '3.9-rc'
421 >>> versiontuple(v, 2)
421 >>> versiontuple(v, 2)
422 (3, 9)
422 (3, 9)
423 >>> versiontuple(v, 3)
423 >>> versiontuple(v, 3)
424 (3, 9, None)
424 (3, 9, None)
425 >>> versiontuple(v, 4)
425 >>> versiontuple(v, 4)
426 (3, 9, None, 'rc')
426 (3, 9, None, 'rc')
427
427
428 >>> v = '3.9-rc+2-02a8fea4289b'
428 >>> v = '3.9-rc+2-02a8fea4289b'
429 >>> versiontuple(v, 2)
429 >>> versiontuple(v, 2)
430 (3, 9)
430 (3, 9)
431 >>> versiontuple(v, 3)
431 >>> versiontuple(v, 3)
432 (3, 9, None)
432 (3, 9, None)
433 >>> versiontuple(v, 4)
433 >>> versiontuple(v, 4)
434 (3, 9, None, 'rc+2-02a8fea4289b')
434 (3, 9, None, 'rc+2-02a8fea4289b')
435 """
435 """
436 if not v:
436 if not v:
437 v = version()
437 v = version()
438 parts = remod.split('[\+-]', v, 1)
438 parts = remod.split('[\+-]', v, 1)
439 if len(parts) == 1:
439 if len(parts) == 1:
440 vparts, extra = parts[0], None
440 vparts, extra = parts[0], None
441 else:
441 else:
442 vparts, extra = parts
442 vparts, extra = parts
443
443
444 vints = []
444 vints = []
445 for i in vparts.split('.'):
445 for i in vparts.split('.'):
446 try:
446 try:
447 vints.append(int(i))
447 vints.append(int(i))
448 except ValueError:
448 except ValueError:
449 break
449 break
450 # (3, 6) -> (3, 6, None)
450 # (3, 6) -> (3, 6, None)
451 while len(vints) < 3:
451 while len(vints) < 3:
452 vints.append(None)
452 vints.append(None)
453
453
454 if n == 2:
454 if n == 2:
455 return (vints[0], vints[1])
455 return (vints[0], vints[1])
456 if n == 3:
456 if n == 3:
457 return (vints[0], vints[1], vints[2])
457 return (vints[0], vints[1], vints[2])
458 if n == 4:
458 if n == 4:
459 return (vints[0], vints[1], vints[2], extra)
459 return (vints[0], vints[1], vints[2], extra)
460
460
461 # used by parsedate
461 # used by parsedate
462 defaultdateformats = (
462 defaultdateformats = (
463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
464 '%Y-%m-%dT%H:%M', # without seconds
464 '%Y-%m-%dT%H:%M', # without seconds
465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
466 '%Y-%m-%dT%H%M', # without seconds
466 '%Y-%m-%dT%H%M', # without seconds
467 '%Y-%m-%d %H:%M:%S', # our common legal variant
467 '%Y-%m-%d %H:%M:%S', # our common legal variant
468 '%Y-%m-%d %H:%M', # without seconds
468 '%Y-%m-%d %H:%M', # without seconds
469 '%Y-%m-%d %H%M%S', # without :
469 '%Y-%m-%d %H%M%S', # without :
470 '%Y-%m-%d %H%M', # without seconds
470 '%Y-%m-%d %H%M', # without seconds
471 '%Y-%m-%d %I:%M:%S%p',
471 '%Y-%m-%d %I:%M:%S%p',
472 '%Y-%m-%d %H:%M',
472 '%Y-%m-%d %H:%M',
473 '%Y-%m-%d %I:%M%p',
473 '%Y-%m-%d %I:%M%p',
474 '%Y-%m-%d',
474 '%Y-%m-%d',
475 '%m-%d',
475 '%m-%d',
476 '%m/%d',
476 '%m/%d',
477 '%m/%d/%y',
477 '%m/%d/%y',
478 '%m/%d/%Y',
478 '%m/%d/%Y',
479 '%a %b %d %H:%M:%S %Y',
479 '%a %b %d %H:%M:%S %Y',
480 '%a %b %d %I:%M:%S%p %Y',
480 '%a %b %d %I:%M:%S%p %Y',
481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
482 '%b %d %H:%M:%S %Y',
482 '%b %d %H:%M:%S %Y',
483 '%b %d %I:%M:%S%p %Y',
483 '%b %d %I:%M:%S%p %Y',
484 '%b %d %H:%M:%S',
484 '%b %d %H:%M:%S',
485 '%b %d %I:%M:%S%p',
485 '%b %d %I:%M:%S%p',
486 '%b %d %H:%M',
486 '%b %d %H:%M',
487 '%b %d %I:%M%p',
487 '%b %d %I:%M%p',
488 '%b %d %Y',
488 '%b %d %Y',
489 '%b %d',
489 '%b %d',
490 '%H:%M:%S',
490 '%H:%M:%S',
491 '%I:%M:%S%p',
491 '%I:%M:%S%p',
492 '%H:%M',
492 '%H:%M',
493 '%I:%M%p',
493 '%I:%M%p',
494 )
494 )
495
495
496 extendeddateformats = defaultdateformats + (
496 extendeddateformats = defaultdateformats + (
497 "%Y",
497 "%Y",
498 "%Y-%m",
498 "%Y-%m",
499 "%b",
499 "%b",
500 "%b %Y",
500 "%b %Y",
501 )
501 )
502
502
503 def cachefunc(func):
503 def cachefunc(func):
504 '''cache the result of function calls'''
504 '''cache the result of function calls'''
505 # XXX doesn't handle keywords args
505 # XXX doesn't handle keywords args
506 if func.__code__.co_argcount == 0:
506 if func.__code__.co_argcount == 0:
507 cache = []
507 cache = []
508 def f():
508 def f():
509 if len(cache) == 0:
509 if len(cache) == 0:
510 cache.append(func())
510 cache.append(func())
511 return cache[0]
511 return cache[0]
512 return f
512 return f
513 cache = {}
513 cache = {}
514 if func.__code__.co_argcount == 1:
514 if func.__code__.co_argcount == 1:
515 # we gain a small amount of time because
515 # we gain a small amount of time because
516 # we don't need to pack/unpack the list
516 # we don't need to pack/unpack the list
517 def f(arg):
517 def f(arg):
518 if arg not in cache:
518 if arg not in cache:
519 cache[arg] = func(arg)
519 cache[arg] = func(arg)
520 return cache[arg]
520 return cache[arg]
521 else:
521 else:
522 def f(*args):
522 def f(*args):
523 if args not in cache:
523 if args not in cache:
524 cache[args] = func(*args)
524 cache[args] = func(*args)
525 return cache[args]
525 return cache[args]
526
526
527 return f
527 return f
528
528
529 class sortdict(dict):
529 class sortdict(dict):
530 '''a simple sorted dictionary'''
530 '''a simple sorted dictionary'''
531 def __init__(self, data=None):
531 def __init__(self, data=None):
532 self._list = []
532 self._list = []
533 if data:
533 if data:
534 self.update(data)
534 self.update(data)
535 def copy(self):
535 def copy(self):
536 return sortdict(self)
536 return sortdict(self)
537 def __setitem__(self, key, val):
537 def __setitem__(self, key, val):
538 if key in self:
538 if key in self:
539 self._list.remove(key)
539 self._list.remove(key)
540 self._list.append(key)
540 self._list.append(key)
541 dict.__setitem__(self, key, val)
541 dict.__setitem__(self, key, val)
542 def __iter__(self):
542 def __iter__(self):
543 return self._list.__iter__()
543 return self._list.__iter__()
544 def update(self, src):
544 def update(self, src):
545 if isinstance(src, dict):
545 if isinstance(src, dict):
546 src = src.iteritems()
546 src = src.iteritems()
547 for k, v in src:
547 for k, v in src:
548 self[k] = v
548 self[k] = v
549 def clear(self):
549 def clear(self):
550 dict.clear(self)
550 dict.clear(self)
551 self._list = []
551 self._list = []
552 def items(self):
552 def items(self):
553 return [(k, self[k]) for k in self._list]
553 return [(k, self[k]) for k in self._list]
554 def __delitem__(self, key):
554 def __delitem__(self, key):
555 dict.__delitem__(self, key)
555 dict.__delitem__(self, key)
556 self._list.remove(key)
556 self._list.remove(key)
557 def pop(self, key, *args, **kwargs):
557 def pop(self, key, *args, **kwargs):
558 dict.pop(self, key, *args, **kwargs)
558 dict.pop(self, key, *args, **kwargs)
559 try:
559 try:
560 self._list.remove(key)
560 self._list.remove(key)
561 except ValueError:
561 except ValueError:
562 pass
562 pass
563 def keys(self):
563 def keys(self):
564 return self._list[:]
564 return self._list[:]
565 def iterkeys(self):
565 def iterkeys(self):
566 return self._list.__iter__()
566 return self._list.__iter__()
567 def iteritems(self):
567 def iteritems(self):
568 for k in self._list:
568 for k in self._list:
569 yield k, self[k]
569 yield k, self[k]
570 def insert(self, index, key, val):
570 def insert(self, index, key, val):
571 self._list.insert(index, key)
571 self._list.insert(index, key)
572 dict.__setitem__(self, key, val)
572 dict.__setitem__(self, key, val)
573 def __repr__(self):
573 def __repr__(self):
574 if not self:
574 if not self:
575 return '%s()' % self.__class__.__name__
575 return '%s()' % self.__class__.__name__
576 return '%s(%r)' % (self.__class__.__name__, self.items())
576 return '%s(%r)' % (self.__class__.__name__, self.items())
577
577
578 class _lrucachenode(object):
578 class _lrucachenode(object):
579 """A node in a doubly linked list.
579 """A node in a doubly linked list.
580
580
581 Holds a reference to nodes on either side as well as a key-value
581 Holds a reference to nodes on either side as well as a key-value
582 pair for the dictionary entry.
582 pair for the dictionary entry.
583 """
583 """
584 __slots__ = (u'next', u'prev', u'key', u'value')
584 __slots__ = (u'next', u'prev', u'key', u'value')
585
585
586 def __init__(self):
586 def __init__(self):
587 self.next = None
587 self.next = None
588 self.prev = None
588 self.prev = None
589
589
590 self.key = _notset
590 self.key = _notset
591 self.value = None
591 self.value = None
592
592
593 def markempty(self):
593 def markempty(self):
594 """Mark the node as emptied."""
594 """Mark the node as emptied."""
595 self.key = _notset
595 self.key = _notset
596
596
597 class lrucachedict(object):
597 class lrucachedict(object):
598 """Dict that caches most recent accesses and sets.
598 """Dict that caches most recent accesses and sets.
599
599
600 The dict consists of an actual backing dict - indexed by original
600 The dict consists of an actual backing dict - indexed by original
601 key - and a doubly linked circular list defining the order of entries in
601 key - and a doubly linked circular list defining the order of entries in
602 the cache.
602 the cache.
603
603
604 The head node is the newest entry in the cache. If the cache is full,
604 The head node is the newest entry in the cache. If the cache is full,
605 we recycle head.prev and make it the new head. Cache accesses result in
605 we recycle head.prev and make it the new head. Cache accesses result in
606 the node being moved to before the existing head and being marked as the
606 the node being moved to before the existing head and being marked as the
607 new head node.
607 new head node.
608 """
608 """
609 def __init__(self, max):
609 def __init__(self, max):
610 self._cache = {}
610 self._cache = {}
611
611
612 self._head = head = _lrucachenode()
612 self._head = head = _lrucachenode()
613 head.prev = head
613 head.prev = head
614 head.next = head
614 head.next = head
615 self._size = 1
615 self._size = 1
616 self._capacity = max
616 self._capacity = max
617
617
618 def __len__(self):
618 def __len__(self):
619 return len(self._cache)
619 return len(self._cache)
620
620
621 def __contains__(self, k):
621 def __contains__(self, k):
622 return k in self._cache
622 return k in self._cache
623
623
624 def __iter__(self):
624 def __iter__(self):
625 # We don't have to iterate in cache order, but why not.
625 # We don't have to iterate in cache order, but why not.
626 n = self._head
626 n = self._head
627 for i in range(len(self._cache)):
627 for i in range(len(self._cache)):
628 yield n.key
628 yield n.key
629 n = n.next
629 n = n.next
630
630
631 def __getitem__(self, k):
631 def __getitem__(self, k):
632 node = self._cache[k]
632 node = self._cache[k]
633 self._movetohead(node)
633 self._movetohead(node)
634 return node.value
634 return node.value
635
635
636 def __setitem__(self, k, v):
636 def __setitem__(self, k, v):
637 node = self._cache.get(k)
637 node = self._cache.get(k)
638 # Replace existing value and mark as newest.
638 # Replace existing value and mark as newest.
639 if node is not None:
639 if node is not None:
640 node.value = v
640 node.value = v
641 self._movetohead(node)
641 self._movetohead(node)
642 return
642 return
643
643
644 if self._size < self._capacity:
644 if self._size < self._capacity:
645 node = self._addcapacity()
645 node = self._addcapacity()
646 else:
646 else:
647 # Grab the last/oldest item.
647 # Grab the last/oldest item.
648 node = self._head.prev
648 node = self._head.prev
649
649
650 # At capacity. Kill the old entry.
650 # At capacity. Kill the old entry.
651 if node.key is not _notset:
651 if node.key is not _notset:
652 del self._cache[node.key]
652 del self._cache[node.key]
653
653
654 node.key = k
654 node.key = k
655 node.value = v
655 node.value = v
656 self._cache[k] = node
656 self._cache[k] = node
657 # And mark it as newest entry. No need to adjust order since it
657 # And mark it as newest entry. No need to adjust order since it
658 # is already self._head.prev.
658 # is already self._head.prev.
659 self._head = node
659 self._head = node
660
660
661 def __delitem__(self, k):
661 def __delitem__(self, k):
662 node = self._cache.pop(k)
662 node = self._cache.pop(k)
663 node.markempty()
663 node.markempty()
664
664
665 # Temporarily mark as newest item before re-adjusting head to make
665 # Temporarily mark as newest item before re-adjusting head to make
666 # this node the oldest item.
666 # this node the oldest item.
667 self._movetohead(node)
667 self._movetohead(node)
668 self._head = node.next
668 self._head = node.next
669
669
670 # Additional dict methods.
670 # Additional dict methods.
671
671
672 def get(self, k, default=None):
672 def get(self, k, default=None):
673 try:
673 try:
674 return self._cache[k].value
674 return self._cache[k].value
675 except KeyError:
675 except KeyError:
676 return default
676 return default
677
677
678 def clear(self):
678 def clear(self):
679 n = self._head
679 n = self._head
680 while n.key is not _notset:
680 while n.key is not _notset:
681 n.markempty()
681 n.markempty()
682 n = n.next
682 n = n.next
683
683
684 self._cache.clear()
684 self._cache.clear()
685
685
686 def copy(self):
686 def copy(self):
687 result = lrucachedict(self._capacity)
687 result = lrucachedict(self._capacity)
688 n = self._head.prev
688 n = self._head.prev
689 # Iterate in oldest-to-newest order, so the copy has the right ordering
689 # Iterate in oldest-to-newest order, so the copy has the right ordering
690 for i in range(len(self._cache)):
690 for i in range(len(self._cache)):
691 result[n.key] = n.value
691 result[n.key] = n.value
692 n = n.prev
692 n = n.prev
693 return result
693 return result
694
694
695 def _movetohead(self, node):
695 def _movetohead(self, node):
696 """Mark a node as the newest, making it the new head.
696 """Mark a node as the newest, making it the new head.
697
697
698 When a node is accessed, it becomes the freshest entry in the LRU
698 When a node is accessed, it becomes the freshest entry in the LRU
699 list, which is denoted by self._head.
699 list, which is denoted by self._head.
700
700
701 Visually, let's make ``N`` the new head node (* denotes head):
701 Visually, let's make ``N`` the new head node (* denotes head):
702
702
703 previous/oldest <-> head <-> next/next newest
703 previous/oldest <-> head <-> next/next newest
704
704
705 ----<->--- A* ---<->-----
705 ----<->--- A* ---<->-----
706 | |
706 | |
707 E <-> D <-> N <-> C <-> B
707 E <-> D <-> N <-> C <-> B
708
708
709 To:
709 To:
710
710
711 ----<->--- N* ---<->-----
711 ----<->--- N* ---<->-----
712 | |
712 | |
713 E <-> D <-> C <-> B <-> A
713 E <-> D <-> C <-> B <-> A
714
714
715 This requires the following moves:
715 This requires the following moves:
716
716
717 C.next = D (node.prev.next = node.next)
717 C.next = D (node.prev.next = node.next)
718 D.prev = C (node.next.prev = node.prev)
718 D.prev = C (node.next.prev = node.prev)
719 E.next = N (head.prev.next = node)
719 E.next = N (head.prev.next = node)
720 N.prev = E (node.prev = head.prev)
720 N.prev = E (node.prev = head.prev)
721 N.next = A (node.next = head)
721 N.next = A (node.next = head)
722 A.prev = N (head.prev = node)
722 A.prev = N (head.prev = node)
723 """
723 """
724 head = self._head
724 head = self._head
725 # C.next = D
725 # C.next = D
726 node.prev.next = node.next
726 node.prev.next = node.next
727 # D.prev = C
727 # D.prev = C
728 node.next.prev = node.prev
728 node.next.prev = node.prev
729 # N.prev = E
729 # N.prev = E
730 node.prev = head.prev
730 node.prev = head.prev
731 # N.next = A
731 # N.next = A
732 # It is tempting to do just "head" here, however if node is
732 # It is tempting to do just "head" here, however if node is
733 # adjacent to head, this will do bad things.
733 # adjacent to head, this will do bad things.
734 node.next = head.prev.next
734 node.next = head.prev.next
735 # E.next = N
735 # E.next = N
736 node.next.prev = node
736 node.next.prev = node
737 # A.prev = N
737 # A.prev = N
738 node.prev.next = node
738 node.prev.next = node
739
739
740 self._head = node
740 self._head = node
741
741
742 def _addcapacity(self):
742 def _addcapacity(self):
743 """Add a node to the circular linked list.
743 """Add a node to the circular linked list.
744
744
745 The new node is inserted before the head node.
745 The new node is inserted before the head node.
746 """
746 """
747 head = self._head
747 head = self._head
748 node = _lrucachenode()
748 node = _lrucachenode()
749 head.prev.next = node
749 head.prev.next = node
750 node.prev = head.prev
750 node.prev = head.prev
751 node.next = head
751 node.next = head
752 head.prev = node
752 head.prev = node
753 self._size += 1
753 self._size += 1
754 return node
754 return node
755
755
756 def lrucachefunc(func):
756 def lrucachefunc(func):
757 '''cache most recent results of function calls'''
757 '''cache most recent results of function calls'''
758 cache = {}
758 cache = {}
759 order = collections.deque()
759 order = collections.deque()
760 if func.__code__.co_argcount == 1:
760 if func.__code__.co_argcount == 1:
761 def f(arg):
761 def f(arg):
762 if arg not in cache:
762 if arg not in cache:
763 if len(cache) > 20:
763 if len(cache) > 20:
764 del cache[order.popleft()]
764 del cache[order.popleft()]
765 cache[arg] = func(arg)
765 cache[arg] = func(arg)
766 else:
766 else:
767 order.remove(arg)
767 order.remove(arg)
768 order.append(arg)
768 order.append(arg)
769 return cache[arg]
769 return cache[arg]
770 else:
770 else:
771 def f(*args):
771 def f(*args):
772 if args not in cache:
772 if args not in cache:
773 if len(cache) > 20:
773 if len(cache) > 20:
774 del cache[order.popleft()]
774 del cache[order.popleft()]
775 cache[args] = func(*args)
775 cache[args] = func(*args)
776 else:
776 else:
777 order.remove(args)
777 order.remove(args)
778 order.append(args)
778 order.append(args)
779 return cache[args]
779 return cache[args]
780
780
781 return f
781 return f
782
782
783 class propertycache(object):
783 class propertycache(object):
784 def __init__(self, func):
784 def __init__(self, func):
785 self.func = func
785 self.func = func
786 self.name = func.__name__
786 self.name = func.__name__
787 def __get__(self, obj, type=None):
787 def __get__(self, obj, type=None):
788 result = self.func(obj)
788 result = self.func(obj)
789 self.cachevalue(obj, result)
789 self.cachevalue(obj, result)
790 return result
790 return result
791
791
792 def cachevalue(self, obj, value):
792 def cachevalue(self, obj, value):
793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
794 obj.__dict__[self.name] = value
794 obj.__dict__[self.name] = value
795
795
796 def pipefilter(s, cmd):
796 def pipefilter(s, cmd):
797 '''filter string S through command CMD, returning its output'''
797 '''filter string S through command CMD, returning its output'''
798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
800 pout, perr = p.communicate(s)
800 pout, perr = p.communicate(s)
801 return pout
801 return pout
802
802
803 def tempfilter(s, cmd):
803 def tempfilter(s, cmd):
804 '''filter string S through a pair of temporary files with CMD.
804 '''filter string S through a pair of temporary files with CMD.
805 CMD is used as a template to create the real command to be run,
805 CMD is used as a template to create the real command to be run,
806 with the strings INFILE and OUTFILE replaced by the real names of
806 with the strings INFILE and OUTFILE replaced by the real names of
807 the temporary files generated.'''
807 the temporary files generated.'''
808 inname, outname = None, None
808 inname, outname = None, None
809 try:
809 try:
810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
812 fp.write(s)
812 fp.write(s)
813 fp.close()
813 fp.close()
814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
815 os.close(outfd)
815 os.close(outfd)
816 cmd = cmd.replace('INFILE', inname)
816 cmd = cmd.replace('INFILE', inname)
817 cmd = cmd.replace('OUTFILE', outname)
817 cmd = cmd.replace('OUTFILE', outname)
818 code = os.system(cmd)
818 code = os.system(cmd)
819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
820 code = 0
820 code = 0
821 if code:
821 if code:
822 raise Abort(_("command '%s' failed: %s") %
822 raise Abort(_("command '%s' failed: %s") %
823 (cmd, explainexit(code)))
823 (cmd, explainexit(code)))
824 return readfile(outname)
824 return readfile(outname)
825 finally:
825 finally:
826 try:
826 try:
827 if inname:
827 if inname:
828 os.unlink(inname)
828 os.unlink(inname)
829 except OSError:
829 except OSError:
830 pass
830 pass
831 try:
831 try:
832 if outname:
832 if outname:
833 os.unlink(outname)
833 os.unlink(outname)
834 except OSError:
834 except OSError:
835 pass
835 pass
836
836
837 filtertable = {
837 filtertable = {
838 'tempfile:': tempfilter,
838 'tempfile:': tempfilter,
839 'pipe:': pipefilter,
839 'pipe:': pipefilter,
840 }
840 }
841
841
842 def filter(s, cmd):
842 def filter(s, cmd):
843 "filter a string through a command that transforms its input to its output"
843 "filter a string through a command that transforms its input to its output"
844 for name, fn in filtertable.iteritems():
844 for name, fn in filtertable.iteritems():
845 if cmd.startswith(name):
845 if cmd.startswith(name):
846 return fn(s, cmd[len(name):].lstrip())
846 return fn(s, cmd[len(name):].lstrip())
847 return pipefilter(s, cmd)
847 return pipefilter(s, cmd)
848
848
849 def binary(s):
849 def binary(s):
850 """return true if a string is binary data"""
850 """return true if a string is binary data"""
851 return bool(s and '\0' in s)
851 return bool(s and '\0' in s)
852
852
853 def increasingchunks(source, min=1024, max=65536):
853 def increasingchunks(source, min=1024, max=65536):
854 '''return no less than min bytes per chunk while data remains,
854 '''return no less than min bytes per chunk while data remains,
855 doubling min after each chunk until it reaches max'''
855 doubling min after each chunk until it reaches max'''
856 def log2(x):
856 def log2(x):
857 if not x:
857 if not x:
858 return 0
858 return 0
859 i = 0
859 i = 0
860 while x:
860 while x:
861 x >>= 1
861 x >>= 1
862 i += 1
862 i += 1
863 return i - 1
863 return i - 1
864
864
865 buf = []
865 buf = []
866 blen = 0
866 blen = 0
867 for chunk in source:
867 for chunk in source:
868 buf.append(chunk)
868 buf.append(chunk)
869 blen += len(chunk)
869 blen += len(chunk)
870 if blen >= min:
870 if blen >= min:
871 if min < max:
871 if min < max:
872 min = min << 1
872 min = min << 1
873 nmin = 1 << log2(blen)
873 nmin = 1 << log2(blen)
874 if nmin > min:
874 if nmin > min:
875 min = nmin
875 min = nmin
876 if min > max:
876 if min > max:
877 min = max
877 min = max
878 yield ''.join(buf)
878 yield ''.join(buf)
879 blen = 0
879 blen = 0
880 buf = []
880 buf = []
881 if buf:
881 if buf:
882 yield ''.join(buf)
882 yield ''.join(buf)
883
883
884 Abort = error.Abort
884 Abort = error.Abort
885
885
886 def always(fn):
886 def always(fn):
887 return True
887 return True
888
888
889 def never(fn):
889 def never(fn):
890 return False
890 return False
891
891
892 def nogc(func):
892 def nogc(func):
893 """disable garbage collector
893 """disable garbage collector
894
894
895 Python's garbage collector triggers a GC each time a certain number of
895 Python's garbage collector triggers a GC each time a certain number of
896 container objects (the number being defined by gc.get_threshold()) are
896 container objects (the number being defined by gc.get_threshold()) are
897 allocated even when marked not to be tracked by the collector. Tracking has
897 allocated even when marked not to be tracked by the collector. Tracking has
898 no effect on when GCs are triggered, only on what objects the GC looks
898 no effect on when GCs are triggered, only on what objects the GC looks
899 into. As a workaround, disable GC while building complex (huge)
899 into. As a workaround, disable GC while building complex (huge)
900 containers.
900 containers.
901
901
902 This garbage collector issue have been fixed in 2.7.
902 This garbage collector issue have been fixed in 2.7.
903 """
903 """
904 if sys.version_info >= (2, 7):
904 if sys.version_info >= (2, 7):
905 return func
905 return func
906 def wrapper(*args, **kwargs):
906 def wrapper(*args, **kwargs):
907 gcenabled = gc.isenabled()
907 gcenabled = gc.isenabled()
908 gc.disable()
908 gc.disable()
909 try:
909 try:
910 return func(*args, **kwargs)
910 return func(*args, **kwargs)
911 finally:
911 finally:
912 if gcenabled:
912 if gcenabled:
913 gc.enable()
913 gc.enable()
914 return wrapper
914 return wrapper
915
915
916 def pathto(root, n1, n2):
916 def pathto(root, n1, n2):
917 '''return the relative path from one place to another.
917 '''return the relative path from one place to another.
918 root should use os.sep to separate directories
918 root should use os.sep to separate directories
919 n1 should use os.sep to separate directories
919 n1 should use os.sep to separate directories
920 n2 should use "/" to separate directories
920 n2 should use "/" to separate directories
921 returns an os.sep-separated path.
921 returns an os.sep-separated path.
922
922
923 If n1 is a relative path, it's assumed it's
923 If n1 is a relative path, it's assumed it's
924 relative to root.
924 relative to root.
925 n2 should always be relative to root.
925 n2 should always be relative to root.
926 '''
926 '''
927 if not n1:
927 if not n1:
928 return localpath(n2)
928 return localpath(n2)
929 if os.path.isabs(n1):
929 if os.path.isabs(n1):
930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
931 return os.path.join(root, localpath(n2))
931 return os.path.join(root, localpath(n2))
932 n2 = '/'.join((pconvert(root), n2))
932 n2 = '/'.join((pconvert(root), n2))
933 a, b = splitpath(n1), n2.split('/')
933 a, b = splitpath(n1), n2.split('/')
934 a.reverse()
934 a.reverse()
935 b.reverse()
935 b.reverse()
936 while a and b and a[-1] == b[-1]:
936 while a and b and a[-1] == b[-1]:
937 a.pop()
937 a.pop()
938 b.pop()
938 b.pop()
939 b.reverse()
939 b.reverse()
940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
941
941
942 def mainfrozen():
942 def mainfrozen():
943 """return True if we are a frozen executable.
943 """return True if we are a frozen executable.
944
944
945 The code supports py2exe (most common, Windows only) and tools/freeze
945 The code supports py2exe (most common, Windows only) and tools/freeze
946 (portable, not much used).
946 (portable, not much used).
947 """
947 """
948 return (safehasattr(sys, "frozen") or # new py2exe
948 return (safehasattr(sys, "frozen") or # new py2exe
949 safehasattr(sys, "importers") or # old py2exe
949 safehasattr(sys, "importers") or # old py2exe
950 imp.is_frozen(u"__main__")) # tools/freeze
950 imp.is_frozen(u"__main__")) # tools/freeze
951
951
952 # the location of data files matching the source code
952 # the location of data files matching the source code
953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
954 # executable version (py2exe) doesn't support __file__
954 # executable version (py2exe) doesn't support __file__
955 datapath = os.path.dirname(pycompat.sysexecutable)
955 datapath = os.path.dirname(pycompat.sysexecutable)
956 else:
956 else:
957 datapath = os.path.dirname(pycompat.fsencode(__file__))
957 datapath = os.path.dirname(pycompat.fsencode(__file__))
958
958
959 i18n.setdatapath(datapath)
959 i18n.setdatapath(datapath)
960
960
961 _hgexecutable = None
961 _hgexecutable = None
962
962
963 def hgexecutable():
963 def hgexecutable():
964 """return location of the 'hg' executable.
964 """return location of the 'hg' executable.
965
965
966 Defaults to $HG or 'hg' in the search path.
966 Defaults to $HG or 'hg' in the search path.
967 """
967 """
968 if _hgexecutable is None:
968 if _hgexecutable is None:
969 hg = encoding.environ.get('HG')
969 hg = encoding.environ.get('HG')
970 mainmod = sys.modules[pycompat.sysstr('__main__')]
970 mainmod = sys.modules[pycompat.sysstr('__main__')]
971 if hg:
971 if hg:
972 _sethgexecutable(hg)
972 _sethgexecutable(hg)
973 elif mainfrozen():
973 elif mainfrozen():
974 if getattr(sys, 'frozen', None) == 'macosx_app':
974 if getattr(sys, 'frozen', None) == 'macosx_app':
975 # Env variable set by py2app
975 # Env variable set by py2app
976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
977 else:
977 else:
978 _sethgexecutable(pycompat.sysexecutable)
978 _sethgexecutable(pycompat.sysexecutable)
979 elif (os.path.basename(
979 elif (os.path.basename(
980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
982 else:
982 else:
983 exe = findexe('hg') or os.path.basename(sys.argv[0])
983 exe = findexe('hg') or os.path.basename(sys.argv[0])
984 _sethgexecutable(exe)
984 _sethgexecutable(exe)
985 return _hgexecutable
985 return _hgexecutable
986
986
987 def _sethgexecutable(path):
987 def _sethgexecutable(path):
988 """set location of the 'hg' executable"""
988 """set location of the 'hg' executable"""
989 global _hgexecutable
989 global _hgexecutable
990 _hgexecutable = path
990 _hgexecutable = path
991
991
992 def _isstdout(f):
992 def _isstdout(f):
993 fileno = getattr(f, 'fileno', None)
993 fileno = getattr(f, 'fileno', None)
994 return fileno and fileno() == sys.__stdout__.fileno()
994 return fileno and fileno() == sys.__stdout__.fileno()
995
995
996 def shellenviron(environ=None):
996 def shellenviron(environ=None):
997 """return environ with optional override, useful for shelling out"""
997 """return environ with optional override, useful for shelling out"""
998 def py2shell(val):
998 def py2shell(val):
999 'convert python object into string that is useful to shell'
999 'convert python object into string that is useful to shell'
1000 if val is None or val is False:
1000 if val is None or val is False:
1001 return '0'
1001 return '0'
1002 if val is True:
1002 if val is True:
1003 return '1'
1003 return '1'
1004 return str(val)
1004 return str(val)
1005 env = dict(encoding.environ)
1005 env = dict(encoding.environ)
1006 if environ:
1006 if environ:
1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1008 env['HG'] = hgexecutable()
1008 env['HG'] = hgexecutable()
1009 return env
1009 return env
1010
1010
1011 def system(cmd, environ=None, cwd=None, out=None):
1011 def system(cmd, environ=None, cwd=None, out=None):
1012 '''enhanced shell command execution.
1012 '''enhanced shell command execution.
1013 run with environment maybe modified, maybe in different dir.
1013 run with environment maybe modified, maybe in different dir.
1014
1014
1015 if out is specified, it is assumed to be a file-like object that has a
1015 if out is specified, it is assumed to be a file-like object that has a
1016 write() method. stdout and stderr will be redirected to out.'''
1016 write() method. stdout and stderr will be redirected to out.'''
1017 try:
1017 try:
1018 stdout.flush()
1018 stdout.flush()
1019 except Exception:
1019 except Exception:
1020 pass
1020 pass
1021 cmd = quotecommand(cmd)
1021 cmd = quotecommand(cmd)
1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1023 and sys.version_info[1] < 7):
1023 and sys.version_info[1] < 7):
1024 # subprocess kludge to work around issues in half-baked Python
1024 # subprocess kludge to work around issues in half-baked Python
1025 # ports, notably bichued/python:
1025 # ports, notably bichued/python:
1026 if not cwd is None:
1026 if not cwd is None:
1027 os.chdir(cwd)
1027 os.chdir(cwd)
1028 rc = os.system(cmd)
1028 rc = os.system(cmd)
1029 else:
1029 else:
1030 env = shellenviron(environ)
1030 env = shellenviron(environ)
1031 if out is None or _isstdout(out):
1031 if out is None or _isstdout(out):
1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1033 env=env, cwd=cwd)
1033 env=env, cwd=cwd)
1034 else:
1034 else:
1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1037 stderr=subprocess.STDOUT)
1037 stderr=subprocess.STDOUT)
1038 for line in iter(proc.stdout.readline, ''):
1038 for line in iter(proc.stdout.readline, ''):
1039 out.write(line)
1039 out.write(line)
1040 proc.wait()
1040 proc.wait()
1041 rc = proc.returncode
1041 rc = proc.returncode
1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1043 rc = 0
1043 rc = 0
1044 return rc
1044 return rc
1045
1045
1046 def checksignature(func):
1046 def checksignature(func):
1047 '''wrap a function with code to check for calling errors'''
1047 '''wrap a function with code to check for calling errors'''
1048 def check(*args, **kwargs):
1048 def check(*args, **kwargs):
1049 try:
1049 try:
1050 return func(*args, **kwargs)
1050 return func(*args, **kwargs)
1051 except TypeError:
1051 except TypeError:
1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1053 raise error.SignatureError
1053 raise error.SignatureError
1054 raise
1054 raise
1055
1055
1056 return check
1056 return check
1057
1057
1058 # a whilelist of known filesystems where hardlink works reliably
1058 # a whilelist of known filesystems where hardlink works reliably
1059 _hardlinkfswhitelist = set([
1059 _hardlinkfswhitelist = set([
1060 'btrfs',
1060 'btrfs',
1061 'ext2',
1061 'ext2',
1062 'ext3',
1062 'ext3',
1063 'ext4',
1063 'ext4',
1064 'hfs',
1064 'hfs',
1065 'jfs',
1065 'jfs',
1066 'reiserfs',
1066 'reiserfs',
1067 'tmpfs',
1067 'tmpfs',
1068 'ufs',
1068 'ufs',
1069 'xfs',
1069 'xfs',
1070 'zfs',
1070 'zfs',
1071 ])
1071 ])
1072
1072
1073 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1073 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1074 '''copy a file, preserving mode and optionally other stat info like
1074 '''copy a file, preserving mode and optionally other stat info like
1075 atime/mtime
1075 atime/mtime
1076
1076
1077 checkambig argument is used with filestat, and is useful only if
1077 checkambig argument is used with filestat, and is useful only if
1078 destination file is guarded by any lock (e.g. repo.lock or
1078 destination file is guarded by any lock (e.g. repo.lock or
1079 repo.wlock).
1079 repo.wlock).
1080
1080
1081 copystat and checkambig should be exclusive.
1081 copystat and checkambig should be exclusive.
1082 '''
1082 '''
1083 assert not (copystat and checkambig)
1083 assert not (copystat and checkambig)
1084 oldstat = None
1084 oldstat = None
1085 if os.path.lexists(dest):
1085 if os.path.lexists(dest):
1086 if checkambig:
1086 if checkambig:
1087 oldstat = checkambig and filestat(dest)
1087 oldstat = checkambig and filestat(dest)
1088 unlink(dest)
1088 unlink(dest)
1089 if hardlink:
1089 if hardlink:
1090 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1090 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1091 # unless we are confident that dest is on a whitelisted filesystem.
1091 # unless we are confident that dest is on a whitelisted filesystem.
1092 try:
1092 try:
1093 fstype = getfstype(os.path.dirname(dest))
1093 fstype = getfstype(os.path.dirname(dest))
1094 except OSError:
1094 except OSError:
1095 fstype = None
1095 fstype = None
1096 if fstype not in _hardlinkfswhitelist:
1096 if fstype not in _hardlinkfswhitelist:
1097 hardlink = False
1097 hardlink = False
1098 if hardlink:
1098 if hardlink:
1099 try:
1099 try:
1100 oslink(src, dest)
1100 oslink(src, dest)
1101 return
1101 return
1102 except (IOError, OSError):
1102 except (IOError, OSError):
1103 pass # fall back to normal copy
1103 pass # fall back to normal copy
1104 if os.path.islink(src):
1104 if os.path.islink(src):
1105 os.symlink(os.readlink(src), dest)
1105 os.symlink(os.readlink(src), dest)
1106 # copytime is ignored for symlinks, but in general copytime isn't needed
1106 # copytime is ignored for symlinks, but in general copytime isn't needed
1107 # for them anyway
1107 # for them anyway
1108 else:
1108 else:
1109 try:
1109 try:
1110 shutil.copyfile(src, dest)
1110 shutil.copyfile(src, dest)
1111 if copystat:
1111 if copystat:
1112 # copystat also copies mode
1112 # copystat also copies mode
1113 shutil.copystat(src, dest)
1113 shutil.copystat(src, dest)
1114 else:
1114 else:
1115 shutil.copymode(src, dest)
1115 shutil.copymode(src, dest)
1116 if oldstat and oldstat.stat:
1116 if oldstat and oldstat.stat:
1117 newstat = filestat(dest)
1117 newstat = filestat(dest)
1118 if newstat.isambig(oldstat):
1118 if newstat.isambig(oldstat):
1119 # stat of copied file is ambiguous to original one
1119 # stat of copied file is ambiguous to original one
1120 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1120 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1121 os.utime(dest, (advanced, advanced))
1121 os.utime(dest, (advanced, advanced))
1122 except shutil.Error as inst:
1122 except shutil.Error as inst:
1123 raise Abort(str(inst))
1123 raise Abort(str(inst))
1124
1124
1125 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1125 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1126 """Copy a directory tree using hardlinks if possible."""
1126 """Copy a directory tree using hardlinks if possible."""
1127 num = 0
1127 num = 0
1128
1128
1129 gettopic = lambda: hardlink and _('linking') or _('copying')
1129 gettopic = lambda: hardlink and _('linking') or _('copying')
1130
1130
1131 if os.path.isdir(src):
1131 if os.path.isdir(src):
1132 if hardlink is None:
1132 if hardlink is None:
1133 hardlink = (os.stat(src).st_dev ==
1133 hardlink = (os.stat(src).st_dev ==
1134 os.stat(os.path.dirname(dst)).st_dev)
1134 os.stat(os.path.dirname(dst)).st_dev)
1135 topic = gettopic()
1135 topic = gettopic()
1136 os.mkdir(dst)
1136 os.mkdir(dst)
1137 for name, kind in osutil.listdir(src):
1137 for name, kind in osutil.listdir(src):
1138 srcname = os.path.join(src, name)
1138 srcname = os.path.join(src, name)
1139 dstname = os.path.join(dst, name)
1139 dstname = os.path.join(dst, name)
1140 def nprog(t, pos):
1140 def nprog(t, pos):
1141 if pos is not None:
1141 if pos is not None:
1142 return progress(t, pos + num)
1142 return progress(t, pos + num)
1143 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1143 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1144 num += n
1144 num += n
1145 else:
1145 else:
1146 if hardlink is None:
1146 if hardlink is None:
1147 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1147 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1148 os.stat(os.path.dirname(dst)).st_dev)
1148 os.stat(os.path.dirname(dst)).st_dev)
1149 topic = gettopic()
1149 topic = gettopic()
1150
1150
1151 if hardlink:
1151 if hardlink:
1152 try:
1152 try:
1153 oslink(src, dst)
1153 oslink(src, dst)
1154 except (IOError, OSError):
1154 except (IOError, OSError):
1155 hardlink = False
1155 hardlink = False
1156 shutil.copy(src, dst)
1156 shutil.copy(src, dst)
1157 else:
1157 else:
1158 shutil.copy(src, dst)
1158 shutil.copy(src, dst)
1159 num += 1
1159 num += 1
1160 progress(topic, num)
1160 progress(topic, num)
1161 progress(topic, None)
1161 progress(topic, None)
1162
1162
1163 return hardlink, num
1163 return hardlink, num
1164
1164
1165 _winreservednames = '''con prn aux nul
1165 _winreservednames = '''con prn aux nul
1166 com1 com2 com3 com4 com5 com6 com7 com8 com9
1166 com1 com2 com3 com4 com5 com6 com7 com8 com9
1167 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1167 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1168 _winreservedchars = ':*?"<>|'
1168 _winreservedchars = ':*?"<>|'
1169 def checkwinfilename(path):
1169 def checkwinfilename(path):
1170 r'''Check that the base-relative path is a valid filename on Windows.
1170 r'''Check that the base-relative path is a valid filename on Windows.
1171 Returns None if the path is ok, or a UI string describing the problem.
1171 Returns None if the path is ok, or a UI string describing the problem.
1172
1172
1173 >>> checkwinfilename("just/a/normal/path")
1173 >>> checkwinfilename("just/a/normal/path")
1174 >>> checkwinfilename("foo/bar/con.xml")
1174 >>> checkwinfilename("foo/bar/con.xml")
1175 "filename contains 'con', which is reserved on Windows"
1175 "filename contains 'con', which is reserved on Windows"
1176 >>> checkwinfilename("foo/con.xml/bar")
1176 >>> checkwinfilename("foo/con.xml/bar")
1177 "filename contains 'con', which is reserved on Windows"
1177 "filename contains 'con', which is reserved on Windows"
1178 >>> checkwinfilename("foo/bar/xml.con")
1178 >>> checkwinfilename("foo/bar/xml.con")
1179 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1179 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1180 "filename contains 'AUX', which is reserved on Windows"
1180 "filename contains 'AUX', which is reserved on Windows"
1181 >>> checkwinfilename("foo/bar/bla:.txt")
1181 >>> checkwinfilename("foo/bar/bla:.txt")
1182 "filename contains ':', which is reserved on Windows"
1182 "filename contains ':', which is reserved on Windows"
1183 >>> checkwinfilename("foo/bar/b\07la.txt")
1183 >>> checkwinfilename("foo/bar/b\07la.txt")
1184 "filename contains '\\x07', which is invalid on Windows"
1184 "filename contains '\\x07', which is invalid on Windows"
1185 >>> checkwinfilename("foo/bar/bla ")
1185 >>> checkwinfilename("foo/bar/bla ")
1186 "filename ends with ' ', which is not allowed on Windows"
1186 "filename ends with ' ', which is not allowed on Windows"
1187 >>> checkwinfilename("../bar")
1187 >>> checkwinfilename("../bar")
1188 >>> checkwinfilename("foo\\")
1188 >>> checkwinfilename("foo\\")
1189 "filename ends with '\\', which is invalid on Windows"
1189 "filename ends with '\\', which is invalid on Windows"
1190 >>> checkwinfilename("foo\\/bar")
1190 >>> checkwinfilename("foo\\/bar")
1191 "directory name ends with '\\', which is invalid on Windows"
1191 "directory name ends with '\\', which is invalid on Windows"
1192 '''
1192 '''
1193 if path.endswith('\\'):
1193 if path.endswith('\\'):
1194 return _("filename ends with '\\', which is invalid on Windows")
1194 return _("filename ends with '\\', which is invalid on Windows")
1195 if '\\/' in path:
1195 if '\\/' in path:
1196 return _("directory name ends with '\\', which is invalid on Windows")
1196 return _("directory name ends with '\\', which is invalid on Windows")
1197 for n in path.replace('\\', '/').split('/'):
1197 for n in path.replace('\\', '/').split('/'):
1198 if not n:
1198 if not n:
1199 continue
1199 continue
1200 for c in pycompat.bytestr(n):
1200 for c in pycompat.bytestr(n):
1201 if c in _winreservedchars:
1201 if c in _winreservedchars:
1202 return _("filename contains '%s', which is reserved "
1202 return _("filename contains '%s', which is reserved "
1203 "on Windows") % c
1203 "on Windows") % c
1204 if ord(c) <= 31:
1204 if ord(c) <= 31:
1205 return _("filename contains %r, which is invalid "
1205 return _("filename contains %r, which is invalid "
1206 "on Windows") % c
1206 "on Windows") % c
1207 base = n.split('.')[0]
1207 base = n.split('.')[0]
1208 if base and base.lower() in _winreservednames:
1208 if base and base.lower() in _winreservednames:
1209 return _("filename contains '%s', which is reserved "
1209 return _("filename contains '%s', which is reserved "
1210 "on Windows") % base
1210 "on Windows") % base
1211 t = n[-1]
1211 t = n[-1]
1212 if t in '. ' and n not in '..':
1212 if t in '. ' and n not in '..':
1213 return _("filename ends with '%s', which is not allowed "
1213 return _("filename ends with '%s', which is not allowed "
1214 "on Windows") % t
1214 "on Windows") % t
1215
1215
1216 if pycompat.osname == 'nt':
1216 if pycompat.osname == 'nt':
1217 checkosfilename = checkwinfilename
1217 checkosfilename = checkwinfilename
1218 timer = time.clock
1218 timer = time.clock
1219 else:
1219 else:
1220 checkosfilename = platform.checkosfilename
1220 checkosfilename = platform.checkosfilename
1221 timer = time.time
1221 timer = time.time
1222
1222
1223 if safehasattr(time, "perf_counter"):
1223 if safehasattr(time, "perf_counter"):
1224 timer = time.perf_counter
1224 timer = time.perf_counter
1225
1225
1226 def makelock(info, pathname):
1226 def makelock(info, pathname):
1227 try:
1227 try:
1228 return os.symlink(info, pathname)
1228 return os.symlink(info, pathname)
1229 except OSError as why:
1229 except OSError as why:
1230 if why.errno == errno.EEXIST:
1230 if why.errno == errno.EEXIST:
1231 raise
1231 raise
1232 except AttributeError: # no symlink in os
1232 except AttributeError: # no symlink in os
1233 pass
1233 pass
1234
1234
1235 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1235 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1236 os.write(ld, info)
1236 os.write(ld, info)
1237 os.close(ld)
1237 os.close(ld)
1238
1238
1239 def readlock(pathname):
1239 def readlock(pathname):
1240 try:
1240 try:
1241 return os.readlink(pathname)
1241 return os.readlink(pathname)
1242 except OSError as why:
1242 except OSError as why:
1243 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1243 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1244 raise
1244 raise
1245 except AttributeError: # no symlink in os
1245 except AttributeError: # no symlink in os
1246 pass
1246 pass
1247 fp = posixfile(pathname)
1247 fp = posixfile(pathname)
1248 r = fp.read()
1248 r = fp.read()
1249 fp.close()
1249 fp.close()
1250 return r
1250 return r
1251
1251
1252 def fstat(fp):
1252 def fstat(fp):
1253 '''stat file object that may not have fileno method.'''
1253 '''stat file object that may not have fileno method.'''
1254 try:
1254 try:
1255 return os.fstat(fp.fileno())
1255 return os.fstat(fp.fileno())
1256 except AttributeError:
1256 except AttributeError:
1257 return os.stat(fp.name)
1257 return os.stat(fp.name)
1258
1258
1259 # File system features
1259 # File system features
1260
1260
1261 def fscasesensitive(path):
1261 def fscasesensitive(path):
1262 """
1262 """
1263 Return true if the given path is on a case-sensitive filesystem
1263 Return true if the given path is on a case-sensitive filesystem
1264
1264
1265 Requires a path (like /foo/.hg) ending with a foldable final
1265 Requires a path (like /foo/.hg) ending with a foldable final
1266 directory component.
1266 directory component.
1267 """
1267 """
1268 s1 = os.lstat(path)
1268 s1 = os.lstat(path)
1269 d, b = os.path.split(path)
1269 d, b = os.path.split(path)
1270 b2 = b.upper()
1270 b2 = b.upper()
1271 if b == b2:
1271 if b == b2:
1272 b2 = b.lower()
1272 b2 = b.lower()
1273 if b == b2:
1273 if b == b2:
1274 return True # no evidence against case sensitivity
1274 return True # no evidence against case sensitivity
1275 p2 = os.path.join(d, b2)
1275 p2 = os.path.join(d, b2)
1276 try:
1276 try:
1277 s2 = os.lstat(p2)
1277 s2 = os.lstat(p2)
1278 if s2 == s1:
1278 if s2 == s1:
1279 return False
1279 return False
1280 return True
1280 return True
1281 except OSError:
1281 except OSError:
1282 return True
1282 return True
1283
1283
1284 try:
1284 try:
1285 import re2
1285 import re2
1286 _re2 = None
1286 _re2 = None
1287 except ImportError:
1287 except ImportError:
1288 _re2 = False
1288 _re2 = False
1289
1289
1290 class _re(object):
1290 class _re(object):
1291 def _checkre2(self):
1291 def _checkre2(self):
1292 global _re2
1292 global _re2
1293 try:
1293 try:
1294 # check if match works, see issue3964
1294 # check if match works, see issue3964
1295 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1295 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1296 except ImportError:
1296 except ImportError:
1297 _re2 = False
1297 _re2 = False
1298
1298
1299 def compile(self, pat, flags=0):
1299 def compile(self, pat, flags=0):
1300 '''Compile a regular expression, using re2 if possible
1300 '''Compile a regular expression, using re2 if possible
1301
1301
1302 For best performance, use only re2-compatible regexp features. The
1302 For best performance, use only re2-compatible regexp features. The
1303 only flags from the re module that are re2-compatible are
1303 only flags from the re module that are re2-compatible are
1304 IGNORECASE and MULTILINE.'''
1304 IGNORECASE and MULTILINE.'''
1305 if _re2 is None:
1305 if _re2 is None:
1306 self._checkre2()
1306 self._checkre2()
1307 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1307 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1308 if flags & remod.IGNORECASE:
1308 if flags & remod.IGNORECASE:
1309 pat = '(?i)' + pat
1309 pat = '(?i)' + pat
1310 if flags & remod.MULTILINE:
1310 if flags & remod.MULTILINE:
1311 pat = '(?m)' + pat
1311 pat = '(?m)' + pat
1312 try:
1312 try:
1313 return re2.compile(pat)
1313 return re2.compile(pat)
1314 except re2.error:
1314 except re2.error:
1315 pass
1315 pass
1316 return remod.compile(pat, flags)
1316 return remod.compile(pat, flags)
1317
1317
1318 @propertycache
1318 @propertycache
1319 def escape(self):
1319 def escape(self):
1320 '''Return the version of escape corresponding to self.compile.
1320 '''Return the version of escape corresponding to self.compile.
1321
1321
1322 This is imperfect because whether re2 or re is used for a particular
1322 This is imperfect because whether re2 or re is used for a particular
1323 function depends on the flags, etc, but it's the best we can do.
1323 function depends on the flags, etc, but it's the best we can do.
1324 '''
1324 '''
1325 global _re2
1325 global _re2
1326 if _re2 is None:
1326 if _re2 is None:
1327 self._checkre2()
1327 self._checkre2()
1328 if _re2:
1328 if _re2:
1329 return re2.escape
1329 return re2.escape
1330 else:
1330 else:
1331 return remod.escape
1331 return remod.escape
1332
1332
1333 re = _re()
1333 re = _re()
1334
1334
1335 _fspathcache = {}
1335 _fspathcache = {}
1336 def fspath(name, root):
1336 def fspath(name, root):
1337 '''Get name in the case stored in the filesystem
1337 '''Get name in the case stored in the filesystem
1338
1338
1339 The name should be relative to root, and be normcase-ed for efficiency.
1339 The name should be relative to root, and be normcase-ed for efficiency.
1340
1340
1341 Note that this function is unnecessary, and should not be
1341 Note that this function is unnecessary, and should not be
1342 called, for case-sensitive filesystems (simply because it's expensive).
1342 called, for case-sensitive filesystems (simply because it's expensive).
1343
1343
1344 The root should be normcase-ed, too.
1344 The root should be normcase-ed, too.
1345 '''
1345 '''
1346 def _makefspathcacheentry(dir):
1346 def _makefspathcacheentry(dir):
1347 return dict((normcase(n), n) for n in os.listdir(dir))
1347 return dict((normcase(n), n) for n in os.listdir(dir))
1348
1348
1349 seps = pycompat.ossep
1349 seps = pycompat.ossep
1350 if pycompat.osaltsep:
1350 if pycompat.osaltsep:
1351 seps = seps + pycompat.osaltsep
1351 seps = seps + pycompat.osaltsep
1352 # Protect backslashes. This gets silly very quickly.
1352 # Protect backslashes. This gets silly very quickly.
1353 seps.replace('\\','\\\\')
1353 seps.replace('\\','\\\\')
1354 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1354 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1355 dir = os.path.normpath(root)
1355 dir = os.path.normpath(root)
1356 result = []
1356 result = []
1357 for part, sep in pattern.findall(name):
1357 for part, sep in pattern.findall(name):
1358 if sep:
1358 if sep:
1359 result.append(sep)
1359 result.append(sep)
1360 continue
1360 continue
1361
1361
1362 if dir not in _fspathcache:
1362 if dir not in _fspathcache:
1363 _fspathcache[dir] = _makefspathcacheentry(dir)
1363 _fspathcache[dir] = _makefspathcacheentry(dir)
1364 contents = _fspathcache[dir]
1364 contents = _fspathcache[dir]
1365
1365
1366 found = contents.get(part)
1366 found = contents.get(part)
1367 if not found:
1367 if not found:
1368 # retry "once per directory" per "dirstate.walk" which
1368 # retry "once per directory" per "dirstate.walk" which
1369 # may take place for each patches of "hg qpush", for example
1369 # may take place for each patches of "hg qpush", for example
1370 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1370 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1371 found = contents.get(part)
1371 found = contents.get(part)
1372
1372
1373 result.append(found or part)
1373 result.append(found or part)
1374 dir = os.path.join(dir, part)
1374 dir = os.path.join(dir, part)
1375
1375
1376 return ''.join(result)
1376 return ''.join(result)
1377
1377
1378 def getfstype(dirpath):
1378 def getfstype(dirpath):
1379 '''Get the filesystem type name from a directory (best-effort)
1379 '''Get the filesystem type name from a directory (best-effort)
1380
1380
1381 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1381 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1382 '''
1382 '''
1383 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1383 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1384
1384
1385 def checknlink(testfile):
1385 def checknlink(testfile):
1386 '''check whether hardlink count reporting works properly'''
1386 '''check whether hardlink count reporting works properly'''
1387
1387
1388 # testfile may be open, so we need a separate file for checking to
1388 # testfile may be open, so we need a separate file for checking to
1389 # work around issue2543 (or testfile may get lost on Samba shares)
1389 # work around issue2543 (or testfile may get lost on Samba shares)
1390 f1 = testfile + ".hgtmp1"
1390 f1 = testfile + ".hgtmp1"
1391 if os.path.lexists(f1):
1391 if os.path.lexists(f1):
1392 return False
1392 return False
1393 try:
1393 try:
1394 posixfile(f1, 'w').close()
1394 posixfile(f1, 'w').close()
1395 except IOError:
1395 except IOError:
1396 try:
1396 try:
1397 os.unlink(f1)
1397 os.unlink(f1)
1398 except OSError:
1398 except OSError:
1399 pass
1399 pass
1400 return False
1400 return False
1401
1401
1402 f2 = testfile + ".hgtmp2"
1402 f2 = testfile + ".hgtmp2"
1403 fd = None
1403 fd = None
1404 try:
1404 try:
1405 oslink(f1, f2)
1405 oslink(f1, f2)
1406 # nlinks() may behave differently for files on Windows shares if
1406 # nlinks() may behave differently for files on Windows shares if
1407 # the file is open.
1407 # the file is open.
1408 fd = posixfile(f2)
1408 fd = posixfile(f2)
1409 return nlinks(f2) > 1
1409 return nlinks(f2) > 1
1410 except OSError:
1410 except OSError:
1411 return False
1411 return False
1412 finally:
1412 finally:
1413 if fd is not None:
1413 if fd is not None:
1414 fd.close()
1414 fd.close()
1415 for f in (f1, f2):
1415 for f in (f1, f2):
1416 try:
1416 try:
1417 os.unlink(f)
1417 os.unlink(f)
1418 except OSError:
1418 except OSError:
1419 pass
1419 pass
1420
1420
1421 def endswithsep(path):
1421 def endswithsep(path):
1422 '''Check path ends with os.sep or os.altsep.'''
1422 '''Check path ends with os.sep or os.altsep.'''
1423 return (path.endswith(pycompat.ossep)
1423 return (path.endswith(pycompat.ossep)
1424 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1424 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1425
1425
1426 def splitpath(path):
1426 def splitpath(path):
1427 '''Split path by os.sep.
1427 '''Split path by os.sep.
1428 Note that this function does not use os.altsep because this is
1428 Note that this function does not use os.altsep because this is
1429 an alternative of simple "xxx.split(os.sep)".
1429 an alternative of simple "xxx.split(os.sep)".
1430 It is recommended to use os.path.normpath() before using this
1430 It is recommended to use os.path.normpath() before using this
1431 function if need.'''
1431 function if need.'''
1432 return path.split(pycompat.ossep)
1432 return path.split(pycompat.ossep)
1433
1433
1434 def gui():
1434 def gui():
1435 '''Are we running in a GUI?'''
1435 '''Are we running in a GUI?'''
1436 if pycompat.sysplatform == 'darwin':
1436 if pycompat.sysplatform == 'darwin':
1437 if 'SSH_CONNECTION' in encoding.environ:
1437 if 'SSH_CONNECTION' in encoding.environ:
1438 # handle SSH access to a box where the user is logged in
1438 # handle SSH access to a box where the user is logged in
1439 return False
1439 return False
1440 elif getattr(osutil, 'isgui', None):
1440 elif getattr(osutil, 'isgui', None):
1441 # check if a CoreGraphics session is available
1441 # check if a CoreGraphics session is available
1442 return osutil.isgui()
1442 return osutil.isgui()
1443 else:
1443 else:
1444 # pure build; use a safe default
1444 # pure build; use a safe default
1445 return True
1445 return True
1446 else:
1446 else:
1447 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1447 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1448
1448
1449 def mktempcopy(name, emptyok=False, createmode=None):
1449 def mktempcopy(name, emptyok=False, createmode=None):
1450 """Create a temporary file with the same contents from name
1450 """Create a temporary file with the same contents from name
1451
1451
1452 The permission bits are copied from the original file.
1452 The permission bits are copied from the original file.
1453
1453
1454 If the temporary file is going to be truncated immediately, you
1454 If the temporary file is going to be truncated immediately, you
1455 can use emptyok=True as an optimization.
1455 can use emptyok=True as an optimization.
1456
1456
1457 Returns the name of the temporary file.
1457 Returns the name of the temporary file.
1458 """
1458 """
1459 d, fn = os.path.split(name)
1459 d, fn = os.path.split(name)
1460 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1460 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1461 os.close(fd)
1461 os.close(fd)
1462 # Temporary files are created with mode 0600, which is usually not
1462 # Temporary files are created with mode 0600, which is usually not
1463 # what we want. If the original file already exists, just copy
1463 # what we want. If the original file already exists, just copy
1464 # its mode. Otherwise, manually obey umask.
1464 # its mode. Otherwise, manually obey umask.
1465 copymode(name, temp, createmode)
1465 copymode(name, temp, createmode)
1466 if emptyok:
1466 if emptyok:
1467 return temp
1467 return temp
1468 try:
1468 try:
1469 try:
1469 try:
1470 ifp = posixfile(name, "rb")
1470 ifp = posixfile(name, "rb")
1471 except IOError as inst:
1471 except IOError as inst:
1472 if inst.errno == errno.ENOENT:
1472 if inst.errno == errno.ENOENT:
1473 return temp
1473 return temp
1474 if not getattr(inst, 'filename', None):
1474 if not getattr(inst, 'filename', None):
1475 inst.filename = name
1475 inst.filename = name
1476 raise
1476 raise
1477 ofp = posixfile(temp, "wb")
1477 ofp = posixfile(temp, "wb")
1478 for chunk in filechunkiter(ifp):
1478 for chunk in filechunkiter(ifp):
1479 ofp.write(chunk)
1479 ofp.write(chunk)
1480 ifp.close()
1480 ifp.close()
1481 ofp.close()
1481 ofp.close()
1482 except: # re-raises
1482 except: # re-raises
1483 try: os.unlink(temp)
1483 try: os.unlink(temp)
1484 except OSError: pass
1484 except OSError: pass
1485 raise
1485 raise
1486 return temp
1486 return temp
1487
1487
1488 class filestat(object):
1488 class filestat(object):
1489 """help to exactly detect change of a file
1489 """help to exactly detect change of a file
1490
1490
1491 'stat' attribute is result of 'os.stat()' if specified 'path'
1491 'stat' attribute is result of 'os.stat()' if specified 'path'
1492 exists. Otherwise, it is None. This can avoid preparative
1492 exists. Otherwise, it is None. This can avoid preparative
1493 'exists()' examination on client side of this class.
1493 'exists()' examination on client side of this class.
1494 """
1494 """
1495 def __init__(self, path):
1495 def __init__(self, path):
1496 try:
1496 try:
1497 self.stat = os.stat(path)
1497 self.stat = os.stat(path)
1498 except OSError as err:
1498 except OSError as err:
1499 if err.errno != errno.ENOENT:
1499 if err.errno != errno.ENOENT:
1500 raise
1500 raise
1501 self.stat = None
1501 self.stat = None
1502
1502
1503 __hash__ = object.__hash__
1503 __hash__ = object.__hash__
1504
1504
1505 def __eq__(self, old):
1505 def __eq__(self, old):
1506 try:
1506 try:
1507 # if ambiguity between stat of new and old file is
1507 # if ambiguity between stat of new and old file is
1508 # avoided, comparison of size, ctime and mtime is enough
1508 # avoided, comparison of size, ctime and mtime is enough
1509 # to exactly detect change of a file regardless of platform
1509 # to exactly detect change of a file regardless of platform
1510 return (self.stat.st_size == old.stat.st_size and
1510 return (self.stat.st_size == old.stat.st_size and
1511 self.stat.st_ctime == old.stat.st_ctime and
1511 self.stat.st_ctime == old.stat.st_ctime and
1512 self.stat.st_mtime == old.stat.st_mtime)
1512 self.stat.st_mtime == old.stat.st_mtime)
1513 except AttributeError:
1513 except AttributeError:
1514 return False
1514 return False
1515
1515
1516 def isambig(self, old):
1516 def isambig(self, old):
1517 """Examine whether new (= self) stat is ambiguous against old one
1517 """Examine whether new (= self) stat is ambiguous against old one
1518
1518
1519 "S[N]" below means stat of a file at N-th change:
1519 "S[N]" below means stat of a file at N-th change:
1520
1520
1521 - S[n-1].ctime < S[n].ctime: can detect change of a file
1521 - S[n-1].ctime < S[n].ctime: can detect change of a file
1522 - S[n-1].ctime == S[n].ctime
1522 - S[n-1].ctime == S[n].ctime
1523 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1523 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1524 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1524 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1525 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1525 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1526 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1526 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1527
1527
1528 Case (*2) above means that a file was changed twice or more at
1528 Case (*2) above means that a file was changed twice or more at
1529 same time in sec (= S[n-1].ctime), and comparison of timestamp
1529 same time in sec (= S[n-1].ctime), and comparison of timestamp
1530 is ambiguous.
1530 is ambiguous.
1531
1531
1532 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1532 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1533 timestamp is ambiguous".
1533 timestamp is ambiguous".
1534
1534
1535 But advancing mtime only in case (*2) doesn't work as
1535 But advancing mtime only in case (*2) doesn't work as
1536 expected, because naturally advanced S[n].mtime in case (*1)
1536 expected, because naturally advanced S[n].mtime in case (*1)
1537 might be equal to manually advanced S[n-1 or earlier].mtime.
1537 might be equal to manually advanced S[n-1 or earlier].mtime.
1538
1538
1539 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1539 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1540 treated as ambiguous regardless of mtime, to avoid overlooking
1540 treated as ambiguous regardless of mtime, to avoid overlooking
1541 by confliction between such mtime.
1541 by confliction between such mtime.
1542
1542
1543 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1543 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1544 S[n].mtime", even if size of a file isn't changed.
1544 S[n].mtime", even if size of a file isn't changed.
1545 """
1545 """
1546 try:
1546 try:
1547 return (self.stat.st_ctime == old.stat.st_ctime)
1547 return (self.stat.st_ctime == old.stat.st_ctime)
1548 except AttributeError:
1548 except AttributeError:
1549 return False
1549 return False
1550
1550
1551 def avoidambig(self, path, old):
1551 def avoidambig(self, path, old):
1552 """Change file stat of specified path to avoid ambiguity
1552 """Change file stat of specified path to avoid ambiguity
1553
1553
1554 'old' should be previous filestat of 'path'.
1554 'old' should be previous filestat of 'path'.
1555
1555
1556 This skips avoiding ambiguity, if a process doesn't have
1556 This skips avoiding ambiguity, if a process doesn't have
1557 appropriate privileges for 'path'.
1557 appropriate privileges for 'path'.
1558 """
1558 """
1559 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1559 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1560 try:
1560 try:
1561 os.utime(path, (advanced, advanced))
1561 os.utime(path, (advanced, advanced))
1562 except OSError as inst:
1562 except OSError as inst:
1563 if inst.errno == errno.EPERM:
1563 if inst.errno == errno.EPERM:
1564 # utime() on the file created by another user causes EPERM,
1564 # utime() on the file created by another user causes EPERM,
1565 # if a process doesn't have appropriate privileges
1565 # if a process doesn't have appropriate privileges
1566 return
1566 return
1567 raise
1567 raise
1568
1568
1569 def __ne__(self, other):
1569 def __ne__(self, other):
1570 return not self == other
1570 return not self == other
1571
1571
1572 class atomictempfile(object):
1572 class atomictempfile(object):
1573 '''writable file object that atomically updates a file
1573 '''writable file object that atomically updates a file
1574
1574
1575 All writes will go to a temporary copy of the original file. Call
1575 All writes will go to a temporary copy of the original file. Call
1576 close() when you are done writing, and atomictempfile will rename
1576 close() when you are done writing, and atomictempfile will rename
1577 the temporary copy to the original name, making the changes
1577 the temporary copy to the original name, making the changes
1578 visible. If the object is destroyed without being closed, all your
1578 visible. If the object is destroyed without being closed, all your
1579 writes are discarded.
1579 writes are discarded.
1580
1580
1581 checkambig argument of constructor is used with filestat, and is
1581 checkambig argument of constructor is used with filestat, and is
1582 useful only if target file is guarded by any lock (e.g. repo.lock
1582 useful only if target file is guarded by any lock (e.g. repo.lock
1583 or repo.wlock).
1583 or repo.wlock).
1584 '''
1584 '''
1585 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1585 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1586 self.__name = name # permanent name
1586 self.__name = name # permanent name
1587 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1587 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1588 createmode=createmode)
1588 createmode=createmode)
1589 self._fp = posixfile(self._tempname, mode)
1589 self._fp = posixfile(self._tempname, mode)
1590 self._checkambig = checkambig
1590 self._checkambig = checkambig
1591
1591
1592 # delegated methods
1592 # delegated methods
1593 self.read = self._fp.read
1593 self.read = self._fp.read
1594 self.write = self._fp.write
1594 self.write = self._fp.write
1595 self.seek = self._fp.seek
1595 self.seek = self._fp.seek
1596 self.tell = self._fp.tell
1596 self.tell = self._fp.tell
1597 self.fileno = self._fp.fileno
1597 self.fileno = self._fp.fileno
1598
1598
1599 def close(self):
1599 def close(self):
1600 if not self._fp.closed:
1600 if not self._fp.closed:
1601 self._fp.close()
1601 self._fp.close()
1602 filename = localpath(self.__name)
1602 filename = localpath(self.__name)
1603 oldstat = self._checkambig and filestat(filename)
1603 oldstat = self._checkambig and filestat(filename)
1604 if oldstat and oldstat.stat:
1604 if oldstat and oldstat.stat:
1605 rename(self._tempname, filename)
1605 rename(self._tempname, filename)
1606 newstat = filestat(filename)
1606 newstat = filestat(filename)
1607 if newstat.isambig(oldstat):
1607 if newstat.isambig(oldstat):
1608 # stat of changed file is ambiguous to original one
1608 # stat of changed file is ambiguous to original one
1609 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1609 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1610 os.utime(filename, (advanced, advanced))
1610 os.utime(filename, (advanced, advanced))
1611 else:
1611 else:
1612 rename(self._tempname, filename)
1612 rename(self._tempname, filename)
1613
1613
1614 def discard(self):
1614 def discard(self):
1615 if not self._fp.closed:
1615 if not self._fp.closed:
1616 try:
1616 try:
1617 os.unlink(self._tempname)
1617 os.unlink(self._tempname)
1618 except OSError:
1618 except OSError:
1619 pass
1619 pass
1620 self._fp.close()
1620 self._fp.close()
1621
1621
1622 def __del__(self):
1622 def __del__(self):
1623 if safehasattr(self, '_fp'): # constructor actually did something
1623 if safehasattr(self, '_fp'): # constructor actually did something
1624 self.discard()
1624 self.discard()
1625
1625
1626 def __enter__(self):
1626 def __enter__(self):
1627 return self
1627 return self
1628
1628
1629 def __exit__(self, exctype, excvalue, traceback):
1629 def __exit__(self, exctype, excvalue, traceback):
1630 if exctype is not None:
1630 if exctype is not None:
1631 self.discard()
1631 self.discard()
1632 else:
1632 else:
1633 self.close()
1633 self.close()
1634
1634
1635 def unlinkpath(f, ignoremissing=False):
1635 def unlinkpath(f, ignoremissing=False):
1636 """unlink and remove the directory if it is empty"""
1636 """unlink and remove the directory if it is empty"""
1637 if ignoremissing:
1637 if ignoremissing:
1638 tryunlink(f)
1638 tryunlink(f)
1639 else:
1639 else:
1640 unlink(f)
1640 unlink(f)
1641 # try removing directories that might now be empty
1641 # try removing directories that might now be empty
1642 try:
1642 try:
1643 removedirs(os.path.dirname(f))
1643 removedirs(os.path.dirname(f))
1644 except OSError:
1644 except OSError:
1645 pass
1645 pass
1646
1646
1647 def tryunlink(f):
1647 def tryunlink(f):
1648 """Attempt to remove a file, ignoring ENOENT errors."""
1648 """Attempt to remove a file, ignoring ENOENT errors."""
1649 try:
1649 try:
1650 unlink(f)
1650 unlink(f)
1651 except OSError as e:
1651 except OSError as e:
1652 if e.errno != errno.ENOENT:
1652 if e.errno != errno.ENOENT:
1653 raise
1653 raise
1654
1654
1655 def makedirs(name, mode=None, notindexed=False):
1655 def makedirs(name, mode=None, notindexed=False):
1656 """recursive directory creation with parent mode inheritance
1656 """recursive directory creation with parent mode inheritance
1657
1657
1658 Newly created directories are marked as "not to be indexed by
1658 Newly created directories are marked as "not to be indexed by
1659 the content indexing service", if ``notindexed`` is specified
1659 the content indexing service", if ``notindexed`` is specified
1660 for "write" mode access.
1660 for "write" mode access.
1661 """
1661 """
1662 try:
1662 try:
1663 makedir(name, notindexed)
1663 makedir(name, notindexed)
1664 except OSError as err:
1664 except OSError as err:
1665 if err.errno == errno.EEXIST:
1665 if err.errno == errno.EEXIST:
1666 return
1666 return
1667 if err.errno != errno.ENOENT or not name:
1667 if err.errno != errno.ENOENT or not name:
1668 raise
1668 raise
1669 parent = os.path.dirname(os.path.abspath(name))
1669 parent = os.path.dirname(os.path.abspath(name))
1670 if parent == name:
1670 if parent == name:
1671 raise
1671 raise
1672 makedirs(parent, mode, notindexed)
1672 makedirs(parent, mode, notindexed)
1673 try:
1673 try:
1674 makedir(name, notindexed)
1674 makedir(name, notindexed)
1675 except OSError as err:
1675 except OSError as err:
1676 # Catch EEXIST to handle races
1676 # Catch EEXIST to handle races
1677 if err.errno == errno.EEXIST:
1677 if err.errno == errno.EEXIST:
1678 return
1678 return
1679 raise
1679 raise
1680 if mode is not None:
1680 if mode is not None:
1681 os.chmod(name, mode)
1681 os.chmod(name, mode)
1682
1682
1683 def readfile(path):
1683 def readfile(path):
1684 with open(path, 'rb') as fp:
1684 with open(path, 'rb') as fp:
1685 return fp.read()
1685 return fp.read()
1686
1686
1687 def writefile(path, text):
1687 def writefile(path, text):
1688 with open(path, 'wb') as fp:
1688 with open(path, 'wb') as fp:
1689 fp.write(text)
1689 fp.write(text)
1690
1690
1691 def appendfile(path, text):
1691 def appendfile(path, text):
1692 with open(path, 'ab') as fp:
1692 with open(path, 'ab') as fp:
1693 fp.write(text)
1693 fp.write(text)
1694
1694
1695 class chunkbuffer(object):
1695 class chunkbuffer(object):
1696 """Allow arbitrary sized chunks of data to be efficiently read from an
1696 """Allow arbitrary sized chunks of data to be efficiently read from an
1697 iterator over chunks of arbitrary size."""
1697 iterator over chunks of arbitrary size."""
1698
1698
1699 def __init__(self, in_iter):
1699 def __init__(self, in_iter):
1700 """in_iter is the iterator that's iterating over the input chunks.
1700 """in_iter is the iterator that's iterating over the input chunks.
1701 targetsize is how big a buffer to try to maintain."""
1701 targetsize is how big a buffer to try to maintain."""
1702 def splitbig(chunks):
1702 def splitbig(chunks):
1703 for chunk in chunks:
1703 for chunk in chunks:
1704 if len(chunk) > 2**20:
1704 if len(chunk) > 2**20:
1705 pos = 0
1705 pos = 0
1706 while pos < len(chunk):
1706 while pos < len(chunk):
1707 end = pos + 2 ** 18
1707 end = pos + 2 ** 18
1708 yield chunk[pos:end]
1708 yield chunk[pos:end]
1709 pos = end
1709 pos = end
1710 else:
1710 else:
1711 yield chunk
1711 yield chunk
1712 self.iter = splitbig(in_iter)
1712 self.iter = splitbig(in_iter)
1713 self._queue = collections.deque()
1713 self._queue = collections.deque()
1714 self._chunkoffset = 0
1714 self._chunkoffset = 0
1715
1715
1716 def read(self, l=None):
1716 def read(self, l=None):
1717 """Read L bytes of data from the iterator of chunks of data.
1717 """Read L bytes of data from the iterator of chunks of data.
1718 Returns less than L bytes if the iterator runs dry.
1718 Returns less than L bytes if the iterator runs dry.
1719
1719
1720 If size parameter is omitted, read everything"""
1720 If size parameter is omitted, read everything"""
1721 if l is None:
1721 if l is None:
1722 return ''.join(self.iter)
1722 return ''.join(self.iter)
1723
1723
1724 left = l
1724 left = l
1725 buf = []
1725 buf = []
1726 queue = self._queue
1726 queue = self._queue
1727 while left > 0:
1727 while left > 0:
1728 # refill the queue
1728 # refill the queue
1729 if not queue:
1729 if not queue:
1730 target = 2**18
1730 target = 2**18
1731 for chunk in self.iter:
1731 for chunk in self.iter:
1732 queue.append(chunk)
1732 queue.append(chunk)
1733 target -= len(chunk)
1733 target -= len(chunk)
1734 if target <= 0:
1734 if target <= 0:
1735 break
1735 break
1736 if not queue:
1736 if not queue:
1737 break
1737 break
1738
1738
1739 # The easy way to do this would be to queue.popleft(), modify the
1739 # The easy way to do this would be to queue.popleft(), modify the
1740 # chunk (if necessary), then queue.appendleft(). However, for cases
1740 # chunk (if necessary), then queue.appendleft(). However, for cases
1741 # where we read partial chunk content, this incurs 2 dequeue
1741 # where we read partial chunk content, this incurs 2 dequeue
1742 # mutations and creates a new str for the remaining chunk in the
1742 # mutations and creates a new str for the remaining chunk in the
1743 # queue. Our code below avoids this overhead.
1743 # queue. Our code below avoids this overhead.
1744
1744
1745 chunk = queue[0]
1745 chunk = queue[0]
1746 chunkl = len(chunk)
1746 chunkl = len(chunk)
1747 offset = self._chunkoffset
1747 offset = self._chunkoffset
1748
1748
1749 # Use full chunk.
1749 # Use full chunk.
1750 if offset == 0 and left >= chunkl:
1750 if offset == 0 and left >= chunkl:
1751 left -= chunkl
1751 left -= chunkl
1752 queue.popleft()
1752 queue.popleft()
1753 buf.append(chunk)
1753 buf.append(chunk)
1754 # self._chunkoffset remains at 0.
1754 # self._chunkoffset remains at 0.
1755 continue
1755 continue
1756
1756
1757 chunkremaining = chunkl - offset
1757 chunkremaining = chunkl - offset
1758
1758
1759 # Use all of unconsumed part of chunk.
1759 # Use all of unconsumed part of chunk.
1760 if left >= chunkremaining:
1760 if left >= chunkremaining:
1761 left -= chunkremaining
1761 left -= chunkremaining
1762 queue.popleft()
1762 queue.popleft()
1763 # offset == 0 is enabled by block above, so this won't merely
1763 # offset == 0 is enabled by block above, so this won't merely
1764 # copy via ``chunk[0:]``.
1764 # copy via ``chunk[0:]``.
1765 buf.append(chunk[offset:])
1765 buf.append(chunk[offset:])
1766 self._chunkoffset = 0
1766 self._chunkoffset = 0
1767
1767
1768 # Partial chunk needed.
1768 # Partial chunk needed.
1769 else:
1769 else:
1770 buf.append(chunk[offset:offset + left])
1770 buf.append(chunk[offset:offset + left])
1771 self._chunkoffset += left
1771 self._chunkoffset += left
1772 left -= chunkremaining
1772 left -= chunkremaining
1773
1773
1774 return ''.join(buf)
1774 return ''.join(buf)
1775
1775
1776 def filechunkiter(f, size=131072, limit=None):
1776 def filechunkiter(f, size=131072, limit=None):
1777 """Create a generator that produces the data in the file size
1777 """Create a generator that produces the data in the file size
1778 (default 131072) bytes at a time, up to optional limit (default is
1778 (default 131072) bytes at a time, up to optional limit (default is
1779 to read all data). Chunks may be less than size bytes if the
1779 to read all data). Chunks may be less than size bytes if the
1780 chunk is the last chunk in the file, or the file is a socket or
1780 chunk is the last chunk in the file, or the file is a socket or
1781 some other type of file that sometimes reads less data than is
1781 some other type of file that sometimes reads less data than is
1782 requested."""
1782 requested."""
1783 assert size >= 0
1783 assert size >= 0
1784 assert limit is None or limit >= 0
1784 assert limit is None or limit >= 0
1785 while True:
1785 while True:
1786 if limit is None:
1786 if limit is None:
1787 nbytes = size
1787 nbytes = size
1788 else:
1788 else:
1789 nbytes = min(limit, size)
1789 nbytes = min(limit, size)
1790 s = nbytes and f.read(nbytes)
1790 s = nbytes and f.read(nbytes)
1791 if not s:
1791 if not s:
1792 break
1792 break
1793 if limit:
1793 if limit:
1794 limit -= len(s)
1794 limit -= len(s)
1795 yield s
1795 yield s
1796
1796
1797 def makedate(timestamp=None):
1797 def makedate(timestamp=None):
1798 '''Return a unix timestamp (or the current time) as a (unixtime,
1798 '''Return a unix timestamp (or the current time) as a (unixtime,
1799 offset) tuple based off the local timezone.'''
1799 offset) tuple based off the local timezone.'''
1800 if timestamp is None:
1800 if timestamp is None:
1801 timestamp = time.time()
1801 timestamp = time.time()
1802 if timestamp < 0:
1802 if timestamp < 0:
1803 hint = _("check your clock")
1803 hint = _("check your clock")
1804 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1804 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1805 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1805 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1806 datetime.datetime.fromtimestamp(timestamp))
1806 datetime.datetime.fromtimestamp(timestamp))
1807 tz = delta.days * 86400 + delta.seconds
1807 tz = delta.days * 86400 + delta.seconds
1808 return timestamp, tz
1808 return timestamp, tz
1809
1809
1810 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1810 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1811 """represent a (unixtime, offset) tuple as a localized time.
1811 """represent a (unixtime, offset) tuple as a localized time.
1812 unixtime is seconds since the epoch, and offset is the time zone's
1812 unixtime is seconds since the epoch, and offset is the time zone's
1813 number of seconds away from UTC.
1813 number of seconds away from UTC.
1814
1814
1815 >>> datestr((0, 0))
1815 >>> datestr((0, 0))
1816 'Thu Jan 01 00:00:00 1970 +0000'
1816 'Thu Jan 01 00:00:00 1970 +0000'
1817 >>> datestr((42, 0))
1817 >>> datestr((42, 0))
1818 'Thu Jan 01 00:00:42 1970 +0000'
1818 'Thu Jan 01 00:00:42 1970 +0000'
1819 >>> datestr((-42, 0))
1819 >>> datestr((-42, 0))
1820 'Wed Dec 31 23:59:18 1969 +0000'
1820 'Wed Dec 31 23:59:18 1969 +0000'
1821 >>> datestr((0x7fffffff, 0))
1821 >>> datestr((0x7fffffff, 0))
1822 'Tue Jan 19 03:14:07 2038 +0000'
1822 'Tue Jan 19 03:14:07 2038 +0000'
1823 >>> datestr((-0x80000000, 0))
1823 >>> datestr((-0x80000000, 0))
1824 'Fri Dec 13 20:45:52 1901 +0000'
1824 'Fri Dec 13 20:45:52 1901 +0000'
1825 """
1825 """
1826 t, tz = date or makedate()
1826 t, tz = date or makedate()
1827 if "%1" in format or "%2" in format or "%z" in format:
1827 if "%1" in format or "%2" in format or "%z" in format:
1828 sign = (tz > 0) and "-" or "+"
1828 sign = (tz > 0) and "-" or "+"
1829 minutes = abs(tz) // 60
1829 minutes = abs(tz) // 60
1830 q, r = divmod(minutes, 60)
1830 q, r = divmod(minutes, 60)
1831 format = format.replace("%z", "%1%2")
1831 format = format.replace("%z", "%1%2")
1832 format = format.replace("%1", "%c%02d" % (sign, q))
1832 format = format.replace("%1", "%c%02d" % (sign, q))
1833 format = format.replace("%2", "%02d" % r)
1833 format = format.replace("%2", "%02d" % r)
1834 d = t - tz
1834 d = t - tz
1835 if d > 0x7fffffff:
1835 if d > 0x7fffffff:
1836 d = 0x7fffffff
1836 d = 0x7fffffff
1837 elif d < -0x80000000:
1837 elif d < -0x80000000:
1838 d = -0x80000000
1838 d = -0x80000000
1839 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1839 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1840 # because they use the gmtime() system call which is buggy on Windows
1840 # because they use the gmtime() system call which is buggy on Windows
1841 # for negative values.
1841 # for negative values.
1842 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1842 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1843 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1843 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1844 return s
1844 return s
1845
1845
1846 def shortdate(date=None):
1846 def shortdate(date=None):
1847 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1847 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1848 return datestr(date, format='%Y-%m-%d')
1848 return datestr(date, format='%Y-%m-%d')
1849
1849
1850 def parsetimezone(s):
1850 def parsetimezone(s):
1851 """find a trailing timezone, if any, in string, and return a
1851 """find a trailing timezone, if any, in string, and return a
1852 (offset, remainder) pair"""
1852 (offset, remainder) pair"""
1853
1853
1854 if s.endswith("GMT") or s.endswith("UTC"):
1854 if s.endswith("GMT") or s.endswith("UTC"):
1855 return 0, s[:-3].rstrip()
1855 return 0, s[:-3].rstrip()
1856
1856
1857 # Unix-style timezones [+-]hhmm
1857 # Unix-style timezones [+-]hhmm
1858 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1858 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1859 sign = (s[-5] == "+") and 1 or -1
1859 sign = (s[-5] == "+") and 1 or -1
1860 hours = int(s[-4:-2])
1860 hours = int(s[-4:-2])
1861 minutes = int(s[-2:])
1861 minutes = int(s[-2:])
1862 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1862 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1863
1863
1864 # ISO8601 trailing Z
1864 # ISO8601 trailing Z
1865 if s.endswith("Z") and s[-2:-1].isdigit():
1865 if s.endswith("Z") and s[-2:-1].isdigit():
1866 return 0, s[:-1]
1866 return 0, s[:-1]
1867
1867
1868 # ISO8601-style [+-]hh:mm
1868 # ISO8601-style [+-]hh:mm
1869 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1869 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1870 s[-5:-3].isdigit() and s[-2:].isdigit()):
1870 s[-5:-3].isdigit() and s[-2:].isdigit()):
1871 sign = (s[-6] == "+") and 1 or -1
1871 sign = (s[-6] == "+") and 1 or -1
1872 hours = int(s[-5:-3])
1872 hours = int(s[-5:-3])
1873 minutes = int(s[-2:])
1873 minutes = int(s[-2:])
1874 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1874 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1875
1875
1876 return None, s
1876 return None, s
1877
1877
1878 def strdate(string, format, defaults=None):
1878 def strdate(string, format, defaults=None):
1879 """parse a localized time string and return a (unixtime, offset) tuple.
1879 """parse a localized time string and return a (unixtime, offset) tuple.
1880 if the string cannot be parsed, ValueError is raised."""
1880 if the string cannot be parsed, ValueError is raised."""
1881 if defaults is None:
1881 if defaults is None:
1882 defaults = {}
1882 defaults = {}
1883
1883
1884 # NOTE: unixtime = localunixtime + offset
1884 # NOTE: unixtime = localunixtime + offset
1885 offset, date = parsetimezone(string)
1885 offset, date = parsetimezone(string)
1886
1886
1887 # add missing elements from defaults
1887 # add missing elements from defaults
1888 usenow = False # default to using biased defaults
1888 usenow = False # default to using biased defaults
1889 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1889 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1890 found = [True for p in part if ("%"+p) in format]
1890 found = [True for p in part if ("%"+p) in format]
1891 if not found:
1891 if not found:
1892 date += "@" + defaults[part][usenow]
1892 date += "@" + defaults[part][usenow]
1893 format += "@%" + part[0]
1893 format += "@%" + part[0]
1894 else:
1894 else:
1895 # We've found a specific time element, less specific time
1895 # We've found a specific time element, less specific time
1896 # elements are relative to today
1896 # elements are relative to today
1897 usenow = True
1897 usenow = True
1898
1898
1899 timetuple = time.strptime(date, format)
1899 timetuple = time.strptime(date, format)
1900 localunixtime = int(calendar.timegm(timetuple))
1900 localunixtime = int(calendar.timegm(timetuple))
1901 if offset is None:
1901 if offset is None:
1902 # local timezone
1902 # local timezone
1903 unixtime = int(time.mktime(timetuple))
1903 unixtime = int(time.mktime(timetuple))
1904 offset = unixtime - localunixtime
1904 offset = unixtime - localunixtime
1905 else:
1905 else:
1906 unixtime = localunixtime + offset
1906 unixtime = localunixtime + offset
1907 return unixtime, offset
1907 return unixtime, offset
1908
1908
1909 def parsedate(date, formats=None, bias=None):
1909 def parsedate(date, formats=None, bias=None):
1910 """parse a localized date/time and return a (unixtime, offset) tuple.
1910 """parse a localized date/time and return a (unixtime, offset) tuple.
1911
1911
1912 The date may be a "unixtime offset" string or in one of the specified
1912 The date may be a "unixtime offset" string or in one of the specified
1913 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1913 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1914
1914
1915 >>> parsedate(' today ') == parsedate(\
1915 >>> parsedate(' today ') == parsedate(\
1916 datetime.date.today().strftime('%b %d'))
1916 datetime.date.today().strftime('%b %d'))
1917 True
1917 True
1918 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1918 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1919 datetime.timedelta(days=1)\
1919 datetime.timedelta(days=1)\
1920 ).strftime('%b %d'))
1920 ).strftime('%b %d'))
1921 True
1921 True
1922 >>> now, tz = makedate()
1922 >>> now, tz = makedate()
1923 >>> strnow, strtz = parsedate('now')
1923 >>> strnow, strtz = parsedate('now')
1924 >>> (strnow - now) < 1
1924 >>> (strnow - now) < 1
1925 True
1925 True
1926 >>> tz == strtz
1926 >>> tz == strtz
1927 True
1927 True
1928 """
1928 """
1929 if bias is None:
1929 if bias is None:
1930 bias = {}
1930 bias = {}
1931 if not date:
1931 if not date:
1932 return 0, 0
1932 return 0, 0
1933 if isinstance(date, tuple) and len(date) == 2:
1933 if isinstance(date, tuple) and len(date) == 2:
1934 return date
1934 return date
1935 if not formats:
1935 if not formats:
1936 formats = defaultdateformats
1936 formats = defaultdateformats
1937 date = date.strip()
1937 date = date.strip()
1938
1938
1939 if date == 'now' or date == _('now'):
1939 if date == 'now' or date == _('now'):
1940 return makedate()
1940 return makedate()
1941 if date == 'today' or date == _('today'):
1941 if date == 'today' or date == _('today'):
1942 date = datetime.date.today().strftime('%b %d')
1942 date = datetime.date.today().strftime('%b %d')
1943 elif date == 'yesterday' or date == _('yesterday'):
1943 elif date == 'yesterday' or date == _('yesterday'):
1944 date = (datetime.date.today() -
1944 date = (datetime.date.today() -
1945 datetime.timedelta(days=1)).strftime('%b %d')
1945 datetime.timedelta(days=1)).strftime('%b %d')
1946
1946
1947 try:
1947 try:
1948 when, offset = map(int, date.split(' '))
1948 when, offset = map(int, date.split(' '))
1949 except ValueError:
1949 except ValueError:
1950 # fill out defaults
1950 # fill out defaults
1951 now = makedate()
1951 now = makedate()
1952 defaults = {}
1952 defaults = {}
1953 for part in ("d", "mb", "yY", "HI", "M", "S"):
1953 for part in ("d", "mb", "yY", "HI", "M", "S"):
1954 # this piece is for rounding the specific end of unknowns
1954 # this piece is for rounding the specific end of unknowns
1955 b = bias.get(part)
1955 b = bias.get(part)
1956 if b is None:
1956 if b is None:
1957 if part[0] in "HMS":
1957 if part[0] in "HMS":
1958 b = "00"
1958 b = "00"
1959 else:
1959 else:
1960 b = "0"
1960 b = "0"
1961
1961
1962 # this piece is for matching the generic end to today's date
1962 # this piece is for matching the generic end to today's date
1963 n = datestr(now, "%" + part[0])
1963 n = datestr(now, "%" + part[0])
1964
1964
1965 defaults[part] = (b, n)
1965 defaults[part] = (b, n)
1966
1966
1967 for format in formats:
1967 for format in formats:
1968 try:
1968 try:
1969 when, offset = strdate(date, format, defaults)
1969 when, offset = strdate(date, format, defaults)
1970 except (ValueError, OverflowError):
1970 except (ValueError, OverflowError):
1971 pass
1971 pass
1972 else:
1972 else:
1973 break
1973 break
1974 else:
1974 else:
1975 raise Abort(_('invalid date: %r') % date)
1975 raise Abort(_('invalid date: %r') % date)
1976 # validate explicit (probably user-specified) date and
1976 # validate explicit (probably user-specified) date and
1977 # time zone offset. values must fit in signed 32 bits for
1977 # time zone offset. values must fit in signed 32 bits for
1978 # current 32-bit linux runtimes. timezones go from UTC-12
1978 # current 32-bit linux runtimes. timezones go from UTC-12
1979 # to UTC+14
1979 # to UTC+14
1980 if when < -0x80000000 or when > 0x7fffffff:
1980 if when < -0x80000000 or when > 0x7fffffff:
1981 raise Abort(_('date exceeds 32 bits: %d') % when)
1981 raise Abort(_('date exceeds 32 bits: %d') % when)
1982 if offset < -50400 or offset > 43200:
1982 if offset < -50400 or offset > 43200:
1983 raise Abort(_('impossible time zone offset: %d') % offset)
1983 raise Abort(_('impossible time zone offset: %d') % offset)
1984 return when, offset
1984 return when, offset
1985
1985
1986 def matchdate(date):
1986 def matchdate(date):
1987 """Return a function that matches a given date match specifier
1987 """Return a function that matches a given date match specifier
1988
1988
1989 Formats include:
1989 Formats include:
1990
1990
1991 '{date}' match a given date to the accuracy provided
1991 '{date}' match a given date to the accuracy provided
1992
1992
1993 '<{date}' on or before a given date
1993 '<{date}' on or before a given date
1994
1994
1995 '>{date}' on or after a given date
1995 '>{date}' on or after a given date
1996
1996
1997 >>> p1 = parsedate("10:29:59")
1997 >>> p1 = parsedate("10:29:59")
1998 >>> p2 = parsedate("10:30:00")
1998 >>> p2 = parsedate("10:30:00")
1999 >>> p3 = parsedate("10:30:59")
1999 >>> p3 = parsedate("10:30:59")
2000 >>> p4 = parsedate("10:31:00")
2000 >>> p4 = parsedate("10:31:00")
2001 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2001 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2002 >>> f = matchdate("10:30")
2002 >>> f = matchdate("10:30")
2003 >>> f(p1[0])
2003 >>> f(p1[0])
2004 False
2004 False
2005 >>> f(p2[0])
2005 >>> f(p2[0])
2006 True
2006 True
2007 >>> f(p3[0])
2007 >>> f(p3[0])
2008 True
2008 True
2009 >>> f(p4[0])
2009 >>> f(p4[0])
2010 False
2010 False
2011 >>> f(p5[0])
2011 >>> f(p5[0])
2012 False
2012 False
2013 """
2013 """
2014
2014
2015 def lower(date):
2015 def lower(date):
2016 d = {'mb': "1", 'd': "1"}
2016 d = {'mb': "1", 'd': "1"}
2017 return parsedate(date, extendeddateformats, d)[0]
2017 return parsedate(date, extendeddateformats, d)[0]
2018
2018
2019 def upper(date):
2019 def upper(date):
2020 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2020 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2021 for days in ("31", "30", "29"):
2021 for days in ("31", "30", "29"):
2022 try:
2022 try:
2023 d["d"] = days
2023 d["d"] = days
2024 return parsedate(date, extendeddateformats, d)[0]
2024 return parsedate(date, extendeddateformats, d)[0]
2025 except Abort:
2025 except Abort:
2026 pass
2026 pass
2027 d["d"] = "28"
2027 d["d"] = "28"
2028 return parsedate(date, extendeddateformats, d)[0]
2028 return parsedate(date, extendeddateformats, d)[0]
2029
2029
2030 date = date.strip()
2030 date = date.strip()
2031
2031
2032 if not date:
2032 if not date:
2033 raise Abort(_("dates cannot consist entirely of whitespace"))
2033 raise Abort(_("dates cannot consist entirely of whitespace"))
2034 elif date[0] == "<":
2034 elif date[0] == "<":
2035 if not date[1:]:
2035 if not date[1:]:
2036 raise Abort(_("invalid day spec, use '<DATE'"))
2036 raise Abort(_("invalid day spec, use '<DATE'"))
2037 when = upper(date[1:])
2037 when = upper(date[1:])
2038 return lambda x: x <= when
2038 return lambda x: x <= when
2039 elif date[0] == ">":
2039 elif date[0] == ">":
2040 if not date[1:]:
2040 if not date[1:]:
2041 raise Abort(_("invalid day spec, use '>DATE'"))
2041 raise Abort(_("invalid day spec, use '>DATE'"))
2042 when = lower(date[1:])
2042 when = lower(date[1:])
2043 return lambda x: x >= when
2043 return lambda x: x >= when
2044 elif date[0] == "-":
2044 elif date[0] == "-":
2045 try:
2045 try:
2046 days = int(date[1:])
2046 days = int(date[1:])
2047 except ValueError:
2047 except ValueError:
2048 raise Abort(_("invalid day spec: %s") % date[1:])
2048 raise Abort(_("invalid day spec: %s") % date[1:])
2049 if days < 0:
2049 if days < 0:
2050 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2050 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2051 % date[1:])
2051 % date[1:])
2052 when = makedate()[0] - days * 3600 * 24
2052 when = makedate()[0] - days * 3600 * 24
2053 return lambda x: x >= when
2053 return lambda x: x >= when
2054 elif " to " in date:
2054 elif " to " in date:
2055 a, b = date.split(" to ")
2055 a, b = date.split(" to ")
2056 start, stop = lower(a), upper(b)
2056 start, stop = lower(a), upper(b)
2057 return lambda x: x >= start and x <= stop
2057 return lambda x: x >= start and x <= stop
2058 else:
2058 else:
2059 start, stop = lower(date), upper(date)
2059 start, stop = lower(date), upper(date)
2060 return lambda x: x >= start and x <= stop
2060 return lambda x: x >= start and x <= stop
2061
2061
2062 def stringmatcher(pattern, casesensitive=True):
2062 def stringmatcher(pattern, casesensitive=True):
2063 """
2063 """
2064 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2064 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2065 returns the matcher name, pattern, and matcher function.
2065 returns the matcher name, pattern, and matcher function.
2066 missing or unknown prefixes are treated as literal matches.
2066 missing or unknown prefixes are treated as literal matches.
2067
2067
2068 helper for tests:
2068 helper for tests:
2069 >>> def test(pattern, *tests):
2069 >>> def test(pattern, *tests):
2070 ... kind, pattern, matcher = stringmatcher(pattern)
2070 ... kind, pattern, matcher = stringmatcher(pattern)
2071 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2071 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2072 >>> def itest(pattern, *tests):
2072 >>> def itest(pattern, *tests):
2073 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2073 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2074 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2074 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2075
2075
2076 exact matching (no prefix):
2076 exact matching (no prefix):
2077 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2077 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2078 ('literal', 'abcdefg', [False, False, True])
2078 ('literal', 'abcdefg', [False, False, True])
2079
2079
2080 regex matching ('re:' prefix)
2080 regex matching ('re:' prefix)
2081 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2081 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2082 ('re', 'a.+b', [False, False, True])
2082 ('re', 'a.+b', [False, False, True])
2083
2083
2084 force exact matches ('literal:' prefix)
2084 force exact matches ('literal:' prefix)
2085 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2085 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2086 ('literal', 're:foobar', [False, True])
2086 ('literal', 're:foobar', [False, True])
2087
2087
2088 unknown prefixes are ignored and treated as literals
2088 unknown prefixes are ignored and treated as literals
2089 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2089 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2090 ('literal', 'foo:bar', [False, False, True])
2090 ('literal', 'foo:bar', [False, False, True])
2091
2091
2092 case insensitive regex matches
2092 case insensitive regex matches
2093 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2093 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2094 ('re', 'A.+b', [False, False, True])
2094 ('re', 'A.+b', [False, False, True])
2095
2095
2096 case insensitive literal matches
2096 case insensitive literal matches
2097 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2097 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2098 ('literal', 'ABCDEFG', [False, False, True])
2098 ('literal', 'ABCDEFG', [False, False, True])
2099 """
2099 """
2100 if pattern.startswith('re:'):
2100 if pattern.startswith('re:'):
2101 pattern = pattern[3:]
2101 pattern = pattern[3:]
2102 try:
2102 try:
2103 flags = 0
2103 flags = 0
2104 if not casesensitive:
2104 if not casesensitive:
2105 flags = remod.I
2105 flags = remod.I
2106 regex = remod.compile(pattern, flags)
2106 regex = remod.compile(pattern, flags)
2107 except remod.error as e:
2107 except remod.error as e:
2108 raise error.ParseError(_('invalid regular expression: %s')
2108 raise error.ParseError(_('invalid regular expression: %s')
2109 % e)
2109 % e)
2110 return 're', pattern, regex.search
2110 return 're', pattern, regex.search
2111 elif pattern.startswith('literal:'):
2111 elif pattern.startswith('literal:'):
2112 pattern = pattern[8:]
2112 pattern = pattern[8:]
2113
2113
2114 match = pattern.__eq__
2114 match = pattern.__eq__
2115
2115
2116 if not casesensitive:
2116 if not casesensitive:
2117 ipat = encoding.lower(pattern)
2117 ipat = encoding.lower(pattern)
2118 match = lambda s: ipat == encoding.lower(s)
2118 match = lambda s: ipat == encoding.lower(s)
2119 return 'literal', pattern, match
2119 return 'literal', pattern, match
2120
2120
2121 def shortuser(user):
2121 def shortuser(user):
2122 """Return a short representation of a user name or email address."""
2122 """Return a short representation of a user name or email address."""
2123 f = user.find('@')
2123 f = user.find('@')
2124 if f >= 0:
2124 if f >= 0:
2125 user = user[:f]
2125 user = user[:f]
2126 f = user.find('<')
2126 f = user.find('<')
2127 if f >= 0:
2127 if f >= 0:
2128 user = user[f + 1:]
2128 user = user[f + 1:]
2129 f = user.find(' ')
2129 f = user.find(' ')
2130 if f >= 0:
2130 if f >= 0:
2131 user = user[:f]
2131 user = user[:f]
2132 f = user.find('.')
2132 f = user.find('.')
2133 if f >= 0:
2133 if f >= 0:
2134 user = user[:f]
2134 user = user[:f]
2135 return user
2135 return user
2136
2136
2137 def emailuser(user):
2137 def emailuser(user):
2138 """Return the user portion of an email address."""
2138 """Return the user portion of an email address."""
2139 f = user.find('@')
2139 f = user.find('@')
2140 if f >= 0:
2140 if f >= 0:
2141 user = user[:f]
2141 user = user[:f]
2142 f = user.find('<')
2142 f = user.find('<')
2143 if f >= 0:
2143 if f >= 0:
2144 user = user[f + 1:]
2144 user = user[f + 1:]
2145 return user
2145 return user
2146
2146
2147 def email(author):
2147 def email(author):
2148 '''get email of author.'''
2148 '''get email of author.'''
2149 r = author.find('>')
2149 r = author.find('>')
2150 if r == -1:
2150 if r == -1:
2151 r = None
2151 r = None
2152 return author[author.find('<') + 1:r]
2152 return author[author.find('<') + 1:r]
2153
2153
2154 def ellipsis(text, maxlength=400):
2154 def ellipsis(text, maxlength=400):
2155 """Trim string to at most maxlength (default: 400) columns in display."""
2155 """Trim string to at most maxlength (default: 400) columns in display."""
2156 return encoding.trim(text, maxlength, ellipsis='...')
2156 return encoding.trim(text, maxlength, ellipsis='...')
2157
2157
2158 def unitcountfn(*unittable):
2158 def unitcountfn(*unittable):
2159 '''return a function that renders a readable count of some quantity'''
2159 '''return a function that renders a readable count of some quantity'''
2160
2160
2161 def go(count):
2161 def go(count):
2162 for multiplier, divisor, format in unittable:
2162 for multiplier, divisor, format in unittable:
2163 if count >= divisor * multiplier:
2163 if count >= divisor * multiplier:
2164 return format % (count / float(divisor))
2164 return format % (count / float(divisor))
2165 return unittable[-1][2] % count
2165 return unittable[-1][2] % count
2166
2166
2167 return go
2167 return go
2168
2168
2169 def processlinerange(fromline, toline):
2169 def processlinerange(fromline, toline):
2170 """Check that linerange <fromline>:<toline> makes sense and return a
2170 """Check that linerange <fromline>:<toline> makes sense and return a
2171 0-based range.
2171 0-based range.
2172
2172
2173 >>> processlinerange(10, 20)
2173 >>> processlinerange(10, 20)
2174 (9, 20)
2174 (9, 20)
2175 >>> processlinerange(2, 1)
2175 >>> processlinerange(2, 1)
2176 Traceback (most recent call last):
2176 Traceback (most recent call last):
2177 ...
2177 ...
2178 ParseError: line range must be positive
2178 ParseError: line range must be positive
2179 >>> processlinerange(0, 5)
2179 >>> processlinerange(0, 5)
2180 Traceback (most recent call last):
2180 Traceback (most recent call last):
2181 ...
2181 ...
2182 ParseError: fromline must be strictly positive
2182 ParseError: fromline must be strictly positive
2183 """
2183 """
2184 if toline - fromline < 0:
2184 if toline - fromline < 0:
2185 raise error.ParseError(_("line range must be positive"))
2185 raise error.ParseError(_("line range must be positive"))
2186 if fromline < 1:
2186 if fromline < 1:
2187 raise error.ParseError(_("fromline must be strictly positive"))
2187 raise error.ParseError(_("fromline must be strictly positive"))
2188 return fromline - 1, toline
2188 return fromline - 1, toline
2189
2189
2190 bytecount = unitcountfn(
2190 bytecount = unitcountfn(
2191 (100, 1 << 30, _('%.0f GB')),
2191 (100, 1 << 30, _('%.0f GB')),
2192 (10, 1 << 30, _('%.1f GB')),
2192 (10, 1 << 30, _('%.1f GB')),
2193 (1, 1 << 30, _('%.2f GB')),
2193 (1, 1 << 30, _('%.2f GB')),
2194 (100, 1 << 20, _('%.0f MB')),
2194 (100, 1 << 20, _('%.0f MB')),
2195 (10, 1 << 20, _('%.1f MB')),
2195 (10, 1 << 20, _('%.1f MB')),
2196 (1, 1 << 20, _('%.2f MB')),
2196 (1, 1 << 20, _('%.2f MB')),
2197 (100, 1 << 10, _('%.0f KB')),
2197 (100, 1 << 10, _('%.0f KB')),
2198 (10, 1 << 10, _('%.1f KB')),
2198 (10, 1 << 10, _('%.1f KB')),
2199 (1, 1 << 10, _('%.2f KB')),
2199 (1, 1 << 10, _('%.2f KB')),
2200 (1, 1, _('%.0f bytes')),
2200 (1, 1, _('%.0f bytes')),
2201 )
2201 )
2202
2202
2203 # Matches a single EOL which can either be a CRLF where repeated CR
2203 # Matches a single EOL which can either be a CRLF where repeated CR
2204 # are removed or a LF. We do not care about old Macintosh files, so a
2204 # are removed or a LF. We do not care about old Macintosh files, so a
2205 # stray CR is an error.
2205 # stray CR is an error.
2206 _eolre = remod.compile(br'\r*\n')
2206 _eolre = remod.compile(br'\r*\n')
2207
2207
2208 def tolf(s):
2208 def tolf(s):
2209 return _eolre.sub('\n', s)
2209 return _eolre.sub('\n', s)
2210
2210
2211 def tocrlf(s):
2211 def tocrlf(s):
2212 return _eolre.sub('\r\n', s)
2212 return _eolre.sub('\r\n', s)
2213
2213
2214 if pycompat.oslinesep == '\r\n':
2214 if pycompat.oslinesep == '\r\n':
2215 tonativeeol = tocrlf
2215 tonativeeol = tocrlf
2216 fromnativeeol = tolf
2216 fromnativeeol = tolf
2217 else:
2217 else:
2218 tonativeeol = pycompat.identity
2218 tonativeeol = pycompat.identity
2219 fromnativeeol = pycompat.identity
2219 fromnativeeol = pycompat.identity
2220
2220
2221 def escapestr(s):
2221 def escapestr(s):
2222 # call underlying function of s.encode('string_escape') directly for
2222 # call underlying function of s.encode('string_escape') directly for
2223 # Python 3 compatibility
2223 # Python 3 compatibility
2224 return codecs.escape_encode(s)[0]
2224 return codecs.escape_encode(s)[0]
2225
2225
2226 def unescapestr(s):
2226 def unescapestr(s):
2227 return codecs.escape_decode(s)[0]
2227 return codecs.escape_decode(s)[0]
2228
2228
2229 def uirepr(s):
2229 def uirepr(s):
2230 # Avoid double backslash in Windows path repr()
2230 # Avoid double backslash in Windows path repr()
2231 return repr(s).replace('\\\\', '\\')
2231 return repr(s).replace('\\\\', '\\')
2232
2232
2233 # delay import of textwrap
2233 # delay import of textwrap
2234 def MBTextWrapper(**kwargs):
2234 def MBTextWrapper(**kwargs):
2235 class tw(textwrap.TextWrapper):
2235 class tw(textwrap.TextWrapper):
2236 """
2236 """
2237 Extend TextWrapper for width-awareness.
2237 Extend TextWrapper for width-awareness.
2238
2238
2239 Neither number of 'bytes' in any encoding nor 'characters' is
2239 Neither number of 'bytes' in any encoding nor 'characters' is
2240 appropriate to calculate terminal columns for specified string.
2240 appropriate to calculate terminal columns for specified string.
2241
2241
2242 Original TextWrapper implementation uses built-in 'len()' directly,
2242 Original TextWrapper implementation uses built-in 'len()' directly,
2243 so overriding is needed to use width information of each characters.
2243 so overriding is needed to use width information of each characters.
2244
2244
2245 In addition, characters classified into 'ambiguous' width are
2245 In addition, characters classified into 'ambiguous' width are
2246 treated as wide in East Asian area, but as narrow in other.
2246 treated as wide in East Asian area, but as narrow in other.
2247
2247
2248 This requires use decision to determine width of such characters.
2248 This requires use decision to determine width of such characters.
2249 """
2249 """
2250 def _cutdown(self, ucstr, space_left):
2250 def _cutdown(self, ucstr, space_left):
2251 l = 0
2251 l = 0
2252 colwidth = encoding.ucolwidth
2252 colwidth = encoding.ucolwidth
2253 for i in xrange(len(ucstr)):
2253 for i in xrange(len(ucstr)):
2254 l += colwidth(ucstr[i])
2254 l += colwidth(ucstr[i])
2255 if space_left < l:
2255 if space_left < l:
2256 return (ucstr[:i], ucstr[i:])
2256 return (ucstr[:i], ucstr[i:])
2257 return ucstr, ''
2257 return ucstr, ''
2258
2258
2259 # overriding of base class
2259 # overriding of base class
2260 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2260 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2261 space_left = max(width - cur_len, 1)
2261 space_left = max(width - cur_len, 1)
2262
2262
2263 if self.break_long_words:
2263 if self.break_long_words:
2264 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2264 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2265 cur_line.append(cut)
2265 cur_line.append(cut)
2266 reversed_chunks[-1] = res
2266 reversed_chunks[-1] = res
2267 elif not cur_line:
2267 elif not cur_line:
2268 cur_line.append(reversed_chunks.pop())
2268 cur_line.append(reversed_chunks.pop())
2269
2269
2270 # this overriding code is imported from TextWrapper of Python 2.6
2270 # this overriding code is imported from TextWrapper of Python 2.6
2271 # to calculate columns of string by 'encoding.ucolwidth()'
2271 # to calculate columns of string by 'encoding.ucolwidth()'
2272 def _wrap_chunks(self, chunks):
2272 def _wrap_chunks(self, chunks):
2273 colwidth = encoding.ucolwidth
2273 colwidth = encoding.ucolwidth
2274
2274
2275 lines = []
2275 lines = []
2276 if self.width <= 0:
2276 if self.width <= 0:
2277 raise ValueError("invalid width %r (must be > 0)" % self.width)
2277 raise ValueError("invalid width %r (must be > 0)" % self.width)
2278
2278
2279 # Arrange in reverse order so items can be efficiently popped
2279 # Arrange in reverse order so items can be efficiently popped
2280 # from a stack of chucks.
2280 # from a stack of chucks.
2281 chunks.reverse()
2281 chunks.reverse()
2282
2282
2283 while chunks:
2283 while chunks:
2284
2284
2285 # Start the list of chunks that will make up the current line.
2285 # Start the list of chunks that will make up the current line.
2286 # cur_len is just the length of all the chunks in cur_line.
2286 # cur_len is just the length of all the chunks in cur_line.
2287 cur_line = []
2287 cur_line = []
2288 cur_len = 0
2288 cur_len = 0
2289
2289
2290 # Figure out which static string will prefix this line.
2290 # Figure out which static string will prefix this line.
2291 if lines:
2291 if lines:
2292 indent = self.subsequent_indent
2292 indent = self.subsequent_indent
2293 else:
2293 else:
2294 indent = self.initial_indent
2294 indent = self.initial_indent
2295
2295
2296 # Maximum width for this line.
2296 # Maximum width for this line.
2297 width = self.width - len(indent)
2297 width = self.width - len(indent)
2298
2298
2299 # First chunk on line is whitespace -- drop it, unless this
2299 # First chunk on line is whitespace -- drop it, unless this
2300 # is the very beginning of the text (i.e. no lines started yet).
2300 # is the very beginning of the text (i.e. no lines started yet).
2301 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2301 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2302 del chunks[-1]
2302 del chunks[-1]
2303
2303
2304 while chunks:
2304 while chunks:
2305 l = colwidth(chunks[-1])
2305 l = colwidth(chunks[-1])
2306
2306
2307 # Can at least squeeze this chunk onto the current line.
2307 # Can at least squeeze this chunk onto the current line.
2308 if cur_len + l <= width:
2308 if cur_len + l <= width:
2309 cur_line.append(chunks.pop())
2309 cur_line.append(chunks.pop())
2310 cur_len += l
2310 cur_len += l
2311
2311
2312 # Nope, this line is full.
2312 # Nope, this line is full.
2313 else:
2313 else:
2314 break
2314 break
2315
2315
2316 # The current line is full, and the next chunk is too big to
2316 # The current line is full, and the next chunk is too big to
2317 # fit on *any* line (not just this one).
2317 # fit on *any* line (not just this one).
2318 if chunks and colwidth(chunks[-1]) > width:
2318 if chunks and colwidth(chunks[-1]) > width:
2319 self._handle_long_word(chunks, cur_line, cur_len, width)
2319 self._handle_long_word(chunks, cur_line, cur_len, width)
2320
2320
2321 # If the last chunk on this line is all whitespace, drop it.
2321 # If the last chunk on this line is all whitespace, drop it.
2322 if (self.drop_whitespace and
2322 if (self.drop_whitespace and
2323 cur_line and cur_line[-1].strip() == ''):
2323 cur_line and cur_line[-1].strip() == ''):
2324 del cur_line[-1]
2324 del cur_line[-1]
2325
2325
2326 # Convert current line back to a string and store it in list
2326 # Convert current line back to a string and store it in list
2327 # of all lines (return value).
2327 # of all lines (return value).
2328 if cur_line:
2328 if cur_line:
2329 lines.append(indent + ''.join(cur_line))
2329 lines.append(indent + ''.join(cur_line))
2330
2330
2331 return lines
2331 return lines
2332
2332
2333 global MBTextWrapper
2333 global MBTextWrapper
2334 MBTextWrapper = tw
2334 MBTextWrapper = tw
2335 return tw(**kwargs)
2335 return tw(**kwargs)
2336
2336
2337 def wrap(line, width, initindent='', hangindent=''):
2337 def wrap(line, width, initindent='', hangindent=''):
2338 maxindent = max(len(hangindent), len(initindent))
2338 maxindent = max(len(hangindent), len(initindent))
2339 if width <= maxindent:
2339 if width <= maxindent:
2340 # adjust for weird terminal size
2340 # adjust for weird terminal size
2341 width = max(78, maxindent + 1)
2341 width = max(78, maxindent + 1)
2342 line = line.decode(pycompat.sysstr(encoding.encoding),
2342 line = line.decode(pycompat.sysstr(encoding.encoding),
2343 pycompat.sysstr(encoding.encodingmode))
2343 pycompat.sysstr(encoding.encodingmode))
2344 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2344 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2345 pycompat.sysstr(encoding.encodingmode))
2345 pycompat.sysstr(encoding.encodingmode))
2346 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2346 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2347 pycompat.sysstr(encoding.encodingmode))
2347 pycompat.sysstr(encoding.encodingmode))
2348 wrapper = MBTextWrapper(width=width,
2348 wrapper = MBTextWrapper(width=width,
2349 initial_indent=initindent,
2349 initial_indent=initindent,
2350 subsequent_indent=hangindent)
2350 subsequent_indent=hangindent)
2351 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2351 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2352
2352
2353 if (pyplatform.python_implementation() == 'CPython' and
2353 if (pyplatform.python_implementation() == 'CPython' and
2354 sys.version_info < (3, 0)):
2354 sys.version_info < (3, 0)):
2355 # There is an issue in CPython that some IO methods do not handle EINTR
2355 # There is an issue in CPython that some IO methods do not handle EINTR
2356 # correctly. The following table shows what CPython version (and functions)
2356 # correctly. The following table shows what CPython version (and functions)
2357 # are affected (buggy: has the EINTR bug, okay: otherwise):
2357 # are affected (buggy: has the EINTR bug, okay: otherwise):
2358 #
2358 #
2359 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2359 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2360 # --------------------------------------------------
2360 # --------------------------------------------------
2361 # fp.__iter__ | buggy | buggy | okay
2361 # fp.__iter__ | buggy | buggy | okay
2362 # fp.read* | buggy | okay [1] | okay
2362 # fp.read* | buggy | okay [1] | okay
2363 #
2363 #
2364 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2364 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2365 #
2365 #
2366 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2366 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2367 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2367 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2368 #
2368 #
2369 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2369 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2370 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2370 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2371 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2371 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2372 # fp.__iter__ but not other fp.read* methods.
2372 # fp.__iter__ but not other fp.read* methods.
2373 #
2373 #
2374 # On modern systems like Linux, the "read" syscall cannot be interrupted
2374 # On modern systems like Linux, the "read" syscall cannot be interrupted
2375 # when reading "fast" files like on-disk files. So the EINTR issue only
2375 # when reading "fast" files like on-disk files. So the EINTR issue only
2376 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2376 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2377 # files approximately as "fast" files and use the fast (unsafe) code path,
2377 # files approximately as "fast" files and use the fast (unsafe) code path,
2378 # to minimize the performance impact.
2378 # to minimize the performance impact.
2379 if sys.version_info >= (2, 7, 4):
2379 if sys.version_info >= (2, 7, 4):
2380 # fp.readline deals with EINTR correctly, use it as a workaround.
2380 # fp.readline deals with EINTR correctly, use it as a workaround.
2381 def _safeiterfile(fp):
2381 def _safeiterfile(fp):
2382 return iter(fp.readline, '')
2382 return iter(fp.readline, '')
2383 else:
2383 else:
2384 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2384 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2385 # note: this may block longer than necessary because of bufsize.
2385 # note: this may block longer than necessary because of bufsize.
2386 def _safeiterfile(fp, bufsize=4096):
2386 def _safeiterfile(fp, bufsize=4096):
2387 fd = fp.fileno()
2387 fd = fp.fileno()
2388 line = ''
2388 line = ''
2389 while True:
2389 while True:
2390 try:
2390 try:
2391 buf = os.read(fd, bufsize)
2391 buf = os.read(fd, bufsize)
2392 except OSError as ex:
2392 except OSError as ex:
2393 # os.read only raises EINTR before any data is read
2393 # os.read only raises EINTR before any data is read
2394 if ex.errno == errno.EINTR:
2394 if ex.errno == errno.EINTR:
2395 continue
2395 continue
2396 else:
2396 else:
2397 raise
2397 raise
2398 line += buf
2398 line += buf
2399 if '\n' in buf:
2399 if '\n' in buf:
2400 splitted = line.splitlines(True)
2400 splitted = line.splitlines(True)
2401 line = ''
2401 line = ''
2402 for l in splitted:
2402 for l in splitted:
2403 if l[-1] == '\n':
2403 if l[-1] == '\n':
2404 yield l
2404 yield l
2405 else:
2405 else:
2406 line = l
2406 line = l
2407 if not buf:
2407 if not buf:
2408 break
2408 break
2409 if line:
2409 if line:
2410 yield line
2410 yield line
2411
2411
2412 def iterfile(fp):
2412 def iterfile(fp):
2413 fastpath = True
2413 fastpath = True
2414 if type(fp) is file:
2414 if type(fp) is file:
2415 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2415 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2416 if fastpath:
2416 if fastpath:
2417 return fp
2417 return fp
2418 else:
2418 else:
2419 return _safeiterfile(fp)
2419 return _safeiterfile(fp)
2420 else:
2420 else:
2421 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2421 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2422 def iterfile(fp):
2422 def iterfile(fp):
2423 return fp
2423 return fp
2424
2424
2425 def iterlines(iterator):
2425 def iterlines(iterator):
2426 for chunk in iterator:
2426 for chunk in iterator:
2427 for line in chunk.splitlines():
2427 for line in chunk.splitlines():
2428 yield line
2428 yield line
2429
2429
2430 def expandpath(path):
2430 def expandpath(path):
2431 return os.path.expanduser(os.path.expandvars(path))
2431 return os.path.expanduser(os.path.expandvars(path))
2432
2432
2433 def hgcmd():
2433 def hgcmd():
2434 """Return the command used to execute current hg
2434 """Return the command used to execute current hg
2435
2435
2436 This is different from hgexecutable() because on Windows we want
2436 This is different from hgexecutable() because on Windows we want
2437 to avoid things opening new shell windows like batch files, so we
2437 to avoid things opening new shell windows like batch files, so we
2438 get either the python call or current executable.
2438 get either the python call or current executable.
2439 """
2439 """
2440 if mainfrozen():
2440 if mainfrozen():
2441 if getattr(sys, 'frozen', None) == 'macosx_app':
2441 if getattr(sys, 'frozen', None) == 'macosx_app':
2442 # Env variable set by py2app
2442 # Env variable set by py2app
2443 return [encoding.environ['EXECUTABLEPATH']]
2443 return [encoding.environ['EXECUTABLEPATH']]
2444 else:
2444 else:
2445 return [pycompat.sysexecutable]
2445 return [pycompat.sysexecutable]
2446 return gethgcmd()
2446 return gethgcmd()
2447
2447
2448 def rundetached(args, condfn):
2448 def rundetached(args, condfn):
2449 """Execute the argument list in a detached process.
2449 """Execute the argument list in a detached process.
2450
2450
2451 condfn is a callable which is called repeatedly and should return
2451 condfn is a callable which is called repeatedly and should return
2452 True once the child process is known to have started successfully.
2452 True once the child process is known to have started successfully.
2453 At this point, the child process PID is returned. If the child
2453 At this point, the child process PID is returned. If the child
2454 process fails to start or finishes before condfn() evaluates to
2454 process fails to start or finishes before condfn() evaluates to
2455 True, return -1.
2455 True, return -1.
2456 """
2456 """
2457 # Windows case is easier because the child process is either
2457 # Windows case is easier because the child process is either
2458 # successfully starting and validating the condition or exiting
2458 # successfully starting and validating the condition or exiting
2459 # on failure. We just poll on its PID. On Unix, if the child
2459 # on failure. We just poll on its PID. On Unix, if the child
2460 # process fails to start, it will be left in a zombie state until
2460 # process fails to start, it will be left in a zombie state until
2461 # the parent wait on it, which we cannot do since we expect a long
2461 # the parent wait on it, which we cannot do since we expect a long
2462 # running process on success. Instead we listen for SIGCHLD telling
2462 # running process on success. Instead we listen for SIGCHLD telling
2463 # us our child process terminated.
2463 # us our child process terminated.
2464 terminated = set()
2464 terminated = set()
2465 def handler(signum, frame):
2465 def handler(signum, frame):
2466 terminated.add(os.wait())
2466 terminated.add(os.wait())
2467 prevhandler = None
2467 prevhandler = None
2468 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2468 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2469 if SIGCHLD is not None:
2469 if SIGCHLD is not None:
2470 prevhandler = signal.signal(SIGCHLD, handler)
2470 prevhandler = signal.signal(SIGCHLD, handler)
2471 try:
2471 try:
2472 pid = spawndetached(args)
2472 pid = spawndetached(args)
2473 while not condfn():
2473 while not condfn():
2474 if ((pid in terminated or not testpid(pid))
2474 if ((pid in terminated or not testpid(pid))
2475 and not condfn()):
2475 and not condfn()):
2476 return -1
2476 return -1
2477 time.sleep(0.1)
2477 time.sleep(0.1)
2478 return pid
2478 return pid
2479 finally:
2479 finally:
2480 if prevhandler is not None:
2480 if prevhandler is not None:
2481 signal.signal(signal.SIGCHLD, prevhandler)
2481 signal.signal(signal.SIGCHLD, prevhandler)
2482
2482
2483 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2483 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2484 """Return the result of interpolating items in the mapping into string s.
2484 """Return the result of interpolating items in the mapping into string s.
2485
2485
2486 prefix is a single character string, or a two character string with
2486 prefix is a single character string, or a two character string with
2487 a backslash as the first character if the prefix needs to be escaped in
2487 a backslash as the first character if the prefix needs to be escaped in
2488 a regular expression.
2488 a regular expression.
2489
2489
2490 fn is an optional function that will be applied to the replacement text
2490 fn is an optional function that will be applied to the replacement text
2491 just before replacement.
2491 just before replacement.
2492
2492
2493 escape_prefix is an optional flag that allows using doubled prefix for
2493 escape_prefix is an optional flag that allows using doubled prefix for
2494 its escaping.
2494 its escaping.
2495 """
2495 """
2496 fn = fn or (lambda s: s)
2496 fn = fn or (lambda s: s)
2497 patterns = '|'.join(mapping.keys())
2497 patterns = '|'.join(mapping.keys())
2498 if escape_prefix:
2498 if escape_prefix:
2499 patterns += '|' + prefix
2499 patterns += '|' + prefix
2500 if len(prefix) > 1:
2500 if len(prefix) > 1:
2501 prefix_char = prefix[1:]
2501 prefix_char = prefix[1:]
2502 else:
2502 else:
2503 prefix_char = prefix
2503 prefix_char = prefix
2504 mapping[prefix_char] = prefix_char
2504 mapping[prefix_char] = prefix_char
2505 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2505 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2506 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2506 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2507
2507
2508 def getport(port):
2508 def getport(port):
2509 """Return the port for a given network service.
2509 """Return the port for a given network service.
2510
2510
2511 If port is an integer, it's returned as is. If it's a string, it's
2511 If port is an integer, it's returned as is. If it's a string, it's
2512 looked up using socket.getservbyname(). If there's no matching
2512 looked up using socket.getservbyname(). If there's no matching
2513 service, error.Abort is raised.
2513 service, error.Abort is raised.
2514 """
2514 """
2515 try:
2515 try:
2516 return int(port)
2516 return int(port)
2517 except ValueError:
2517 except ValueError:
2518 pass
2518 pass
2519
2519
2520 try:
2520 try:
2521 return socket.getservbyname(port)
2521 return socket.getservbyname(port)
2522 except socket.error:
2522 except socket.error:
2523 raise Abort(_("no port number associated with service '%s'") % port)
2523 raise Abort(_("no port number associated with service '%s'") % port)
2524
2524
2525 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2525 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2526 '0': False, 'no': False, 'false': False, 'off': False,
2526 '0': False, 'no': False, 'false': False, 'off': False,
2527 'never': False}
2527 'never': False}
2528
2528
2529 def parsebool(s):
2529 def parsebool(s):
2530 """Parse s into a boolean.
2530 """Parse s into a boolean.
2531
2531
2532 If s is not a valid boolean, returns None.
2532 If s is not a valid boolean, returns None.
2533 """
2533 """
2534 return _booleans.get(s.lower(), None)
2534 return _booleans.get(s.lower(), None)
2535
2535
2536 _hextochr = dict((a + b, chr(int(a + b, 16)))
2536 _hextochr = dict((a + b, chr(int(a + b, 16)))
2537 for a in string.hexdigits for b in string.hexdigits)
2537 for a in string.hexdigits for b in string.hexdigits)
2538
2538
2539 class url(object):
2539 class url(object):
2540 r"""Reliable URL parser.
2540 r"""Reliable URL parser.
2541
2541
2542 This parses URLs and provides attributes for the following
2542 This parses URLs and provides attributes for the following
2543 components:
2543 components:
2544
2544
2545 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2545 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2546
2546
2547 Missing components are set to None. The only exception is
2547 Missing components are set to None. The only exception is
2548 fragment, which is set to '' if present but empty.
2548 fragment, which is set to '' if present but empty.
2549
2549
2550 If parsefragment is False, fragment is included in query. If
2550 If parsefragment is False, fragment is included in query. If
2551 parsequery is False, query is included in path. If both are
2551 parsequery is False, query is included in path. If both are
2552 False, both fragment and query are included in path.
2552 False, both fragment and query are included in path.
2553
2553
2554 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2554 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2555
2555
2556 Note that for backward compatibility reasons, bundle URLs do not
2556 Note that for backward compatibility reasons, bundle URLs do not
2557 take host names. That means 'bundle://../' has a path of '../'.
2557 take host names. That means 'bundle://../' has a path of '../'.
2558
2558
2559 Examples:
2559 Examples:
2560
2560
2561 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2561 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2562 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2562 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2563 >>> url('ssh://[::1]:2200//home/joe/repo')
2563 >>> url('ssh://[::1]:2200//home/joe/repo')
2564 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2564 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2565 >>> url('file:///home/joe/repo')
2565 >>> url('file:///home/joe/repo')
2566 <url scheme: 'file', path: '/home/joe/repo'>
2566 <url scheme: 'file', path: '/home/joe/repo'>
2567 >>> url('file:///c:/temp/foo/')
2567 >>> url('file:///c:/temp/foo/')
2568 <url scheme: 'file', path: 'c:/temp/foo/'>
2568 <url scheme: 'file', path: 'c:/temp/foo/'>
2569 >>> url('bundle:foo')
2569 >>> url('bundle:foo')
2570 <url scheme: 'bundle', path: 'foo'>
2570 <url scheme: 'bundle', path: 'foo'>
2571 >>> url('bundle://../foo')
2571 >>> url('bundle://../foo')
2572 <url scheme: 'bundle', path: '../foo'>
2572 <url scheme: 'bundle', path: '../foo'>
2573 >>> url(r'c:\foo\bar')
2573 >>> url(r'c:\foo\bar')
2574 <url path: 'c:\\foo\\bar'>
2574 <url path: 'c:\\foo\\bar'>
2575 >>> url(r'\\blah\blah\blah')
2575 >>> url(r'\\blah\blah\blah')
2576 <url path: '\\\\blah\\blah\\blah'>
2576 <url path: '\\\\blah\\blah\\blah'>
2577 >>> url(r'\\blah\blah\blah#baz')
2577 >>> url(r'\\blah\blah\blah#baz')
2578 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2578 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2579 >>> url(r'file:///C:\users\me')
2579 >>> url(r'file:///C:\users\me')
2580 <url scheme: 'file', path: 'C:\\users\\me'>
2580 <url scheme: 'file', path: 'C:\\users\\me'>
2581
2581
2582 Authentication credentials:
2582 Authentication credentials:
2583
2583
2584 >>> url('ssh://joe:xyz@x/repo')
2584 >>> url('ssh://joe:xyz@x/repo')
2585 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2585 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2586 >>> url('ssh://joe@x/repo')
2586 >>> url('ssh://joe@x/repo')
2587 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2587 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2588
2588
2589 Query strings and fragments:
2589 Query strings and fragments:
2590
2590
2591 >>> url('http://host/a?b#c')
2591 >>> url('http://host/a?b#c')
2592 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2592 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2593 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2593 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2594 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2594 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2595
2595
2596 Empty path:
2596 Empty path:
2597
2597
2598 >>> url('')
2598 >>> url('')
2599 <url path: ''>
2599 <url path: ''>
2600 >>> url('#a')
2600 >>> url('#a')
2601 <url path: '', fragment: 'a'>
2601 <url path: '', fragment: 'a'>
2602 >>> url('http://host/')
2602 >>> url('http://host/')
2603 <url scheme: 'http', host: 'host', path: ''>
2603 <url scheme: 'http', host: 'host', path: ''>
2604 >>> url('http://host/#a')
2604 >>> url('http://host/#a')
2605 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2605 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2606
2606
2607 Only scheme:
2607 Only scheme:
2608
2608
2609 >>> url('http:')
2609 >>> url('http:')
2610 <url scheme: 'http'>
2610 <url scheme: 'http'>
2611 """
2611 """
2612
2612
2613 _safechars = "!~*'()+"
2613 _safechars = "!~*'()+"
2614 _safepchars = "/!~*'()+:\\"
2614 _safepchars = "/!~*'()+:\\"
2615 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2615 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2616
2616
2617 def __init__(self, path, parsequery=True, parsefragment=True):
2617 def __init__(self, path, parsequery=True, parsefragment=True):
2618 # We slowly chomp away at path until we have only the path left
2618 # We slowly chomp away at path until we have only the path left
2619 self.scheme = self.user = self.passwd = self.host = None
2619 self.scheme = self.user = self.passwd = self.host = None
2620 self.port = self.path = self.query = self.fragment = None
2620 self.port = self.path = self.query = self.fragment = None
2621 self._localpath = True
2621 self._localpath = True
2622 self._hostport = ''
2622 self._hostport = ''
2623 self._origpath = path
2623 self._origpath = path
2624
2624
2625 if parsefragment and '#' in path:
2625 if parsefragment and '#' in path:
2626 path, self.fragment = path.split('#', 1)
2626 path, self.fragment = path.split('#', 1)
2627
2627
2628 # special case for Windows drive letters and UNC paths
2628 # special case for Windows drive letters and UNC paths
2629 if hasdriveletter(path) or path.startswith('\\\\'):
2629 if hasdriveletter(path) or path.startswith('\\\\'):
2630 self.path = path
2630 self.path = path
2631 return
2631 return
2632
2632
2633 # For compatibility reasons, we can't handle bundle paths as
2633 # For compatibility reasons, we can't handle bundle paths as
2634 # normal URLS
2634 # normal URLS
2635 if path.startswith('bundle:'):
2635 if path.startswith('bundle:'):
2636 self.scheme = 'bundle'
2636 self.scheme = 'bundle'
2637 path = path[7:]
2637 path = path[7:]
2638 if path.startswith('//'):
2638 if path.startswith('//'):
2639 path = path[2:]
2639 path = path[2:]
2640 self.path = path
2640 self.path = path
2641 return
2641 return
2642
2642
2643 if self._matchscheme(path):
2643 if self._matchscheme(path):
2644 parts = path.split(':', 1)
2644 parts = path.split(':', 1)
2645 if parts[0]:
2645 if parts[0]:
2646 self.scheme, path = parts
2646 self.scheme, path = parts
2647 self._localpath = False
2647 self._localpath = False
2648
2648
2649 if not path:
2649 if not path:
2650 path = None
2650 path = None
2651 if self._localpath:
2651 if self._localpath:
2652 self.path = ''
2652 self.path = ''
2653 return
2653 return
2654 else:
2654 else:
2655 if self._localpath:
2655 if self._localpath:
2656 self.path = path
2656 self.path = path
2657 return
2657 return
2658
2658
2659 if parsequery and '?' in path:
2659 if parsequery and '?' in path:
2660 path, self.query = path.split('?', 1)
2660 path, self.query = path.split('?', 1)
2661 if not path:
2661 if not path:
2662 path = None
2662 path = None
2663 if not self.query:
2663 if not self.query:
2664 self.query = None
2664 self.query = None
2665
2665
2666 # // is required to specify a host/authority
2666 # // is required to specify a host/authority
2667 if path and path.startswith('//'):
2667 if path and path.startswith('//'):
2668 parts = path[2:].split('/', 1)
2668 parts = path[2:].split('/', 1)
2669 if len(parts) > 1:
2669 if len(parts) > 1:
2670 self.host, path = parts
2670 self.host, path = parts
2671 else:
2671 else:
2672 self.host = parts[0]
2672 self.host = parts[0]
2673 path = None
2673 path = None
2674 if not self.host:
2674 if not self.host:
2675 self.host = None
2675 self.host = None
2676 # path of file:///d is /d
2676 # path of file:///d is /d
2677 # path of file:///d:/ is d:/, not /d:/
2677 # path of file:///d:/ is d:/, not /d:/
2678 if path and not hasdriveletter(path):
2678 if path and not hasdriveletter(path):
2679 path = '/' + path
2679 path = '/' + path
2680
2680
2681 if self.host and '@' in self.host:
2681 if self.host and '@' in self.host:
2682 self.user, self.host = self.host.rsplit('@', 1)
2682 self.user, self.host = self.host.rsplit('@', 1)
2683 if ':' in self.user:
2683 if ':' in self.user:
2684 self.user, self.passwd = self.user.split(':', 1)
2684 self.user, self.passwd = self.user.split(':', 1)
2685 if not self.host:
2685 if not self.host:
2686 self.host = None
2686 self.host = None
2687
2687
2688 # Don't split on colons in IPv6 addresses without ports
2688 # Don't split on colons in IPv6 addresses without ports
2689 if (self.host and ':' in self.host and
2689 if (self.host and ':' in self.host and
2690 not (self.host.startswith('[') and self.host.endswith(']'))):
2690 not (self.host.startswith('[') and self.host.endswith(']'))):
2691 self._hostport = self.host
2691 self._hostport = self.host
2692 self.host, self.port = self.host.rsplit(':', 1)
2692 self.host, self.port = self.host.rsplit(':', 1)
2693 if not self.host:
2693 if not self.host:
2694 self.host = None
2694 self.host = None
2695
2695
2696 if (self.host and self.scheme == 'file' and
2696 if (self.host and self.scheme == 'file' and
2697 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2697 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2698 raise Abort(_('file:// URLs can only refer to localhost'))
2698 raise Abort(_('file:// URLs can only refer to localhost'))
2699
2699
2700 self.path = path
2700 self.path = path
2701
2701
2702 # leave the query string escaped
2702 # leave the query string escaped
2703 for a in ('user', 'passwd', 'host', 'port',
2703 for a in ('user', 'passwd', 'host', 'port',
2704 'path', 'fragment'):
2704 'path', 'fragment'):
2705 v = getattr(self, a)
2705 v = getattr(self, a)
2706 if v is not None:
2706 if v is not None:
2707 setattr(self, a, urlreq.unquote(v))
2707 setattr(self, a, urlreq.unquote(v))
2708
2708
2709 def __repr__(self):
2709 def __repr__(self):
2710 attrs = []
2710 attrs = []
2711 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2711 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2712 'query', 'fragment'):
2712 'query', 'fragment'):
2713 v = getattr(self, a)
2713 v = getattr(self, a)
2714 if v is not None:
2714 if v is not None:
2715 attrs.append('%s: %r' % (a, v))
2715 attrs.append('%s: %r' % (a, v))
2716 return '<url %s>' % ', '.join(attrs)
2716 return '<url %s>' % ', '.join(attrs)
2717
2717
2718 def __str__(self):
2718 def __str__(self):
2719 r"""Join the URL's components back into a URL string.
2719 r"""Join the URL's components back into a URL string.
2720
2720
2721 Examples:
2721 Examples:
2722
2722
2723 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2723 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2724 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2724 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2725 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2725 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2726 'http://user:pw@host:80/?foo=bar&baz=42'
2726 'http://user:pw@host:80/?foo=bar&baz=42'
2727 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2727 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2728 'http://user:pw@host:80/?foo=bar%3dbaz'
2728 'http://user:pw@host:80/?foo=bar%3dbaz'
2729 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2729 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2730 'ssh://user:pw@[::1]:2200//home/joe#'
2730 'ssh://user:pw@[::1]:2200//home/joe#'
2731 >>> str(url('http://localhost:80//'))
2731 >>> str(url('http://localhost:80//'))
2732 'http://localhost:80//'
2732 'http://localhost:80//'
2733 >>> str(url('http://localhost:80/'))
2733 >>> str(url('http://localhost:80/'))
2734 'http://localhost:80/'
2734 'http://localhost:80/'
2735 >>> str(url('http://localhost:80'))
2735 >>> str(url('http://localhost:80'))
2736 'http://localhost:80/'
2736 'http://localhost:80/'
2737 >>> str(url('bundle:foo'))
2737 >>> str(url('bundle:foo'))
2738 'bundle:foo'
2738 'bundle:foo'
2739 >>> str(url('bundle://../foo'))
2739 >>> str(url('bundle://../foo'))
2740 'bundle:../foo'
2740 'bundle:../foo'
2741 >>> str(url('path'))
2741 >>> str(url('path'))
2742 'path'
2742 'path'
2743 >>> str(url('file:///tmp/foo/bar'))
2743 >>> str(url('file:///tmp/foo/bar'))
2744 'file:///tmp/foo/bar'
2744 'file:///tmp/foo/bar'
2745 >>> str(url('file:///c:/tmp/foo/bar'))
2745 >>> str(url('file:///c:/tmp/foo/bar'))
2746 'file:///c:/tmp/foo/bar'
2746 'file:///c:/tmp/foo/bar'
2747 >>> print url(r'bundle:foo\bar')
2747 >>> print url(r'bundle:foo\bar')
2748 bundle:foo\bar
2748 bundle:foo\bar
2749 >>> print url(r'file:///D:\data\hg')
2749 >>> print url(r'file:///D:\data\hg')
2750 file:///D:\data\hg
2750 file:///D:\data\hg
2751 """
2751 """
2752 return encoding.strfromlocal(self.__bytes__())
2752 return encoding.strfromlocal(self.__bytes__())
2753
2753
2754 def __bytes__(self):
2754 def __bytes__(self):
2755 if self._localpath:
2755 if self._localpath:
2756 s = self.path
2756 s = self.path
2757 if self.scheme == 'bundle':
2757 if self.scheme == 'bundle':
2758 s = 'bundle:' + s
2758 s = 'bundle:' + s
2759 if self.fragment:
2759 if self.fragment:
2760 s += '#' + self.fragment
2760 s += '#' + self.fragment
2761 return s
2761 return s
2762
2762
2763 s = self.scheme + ':'
2763 s = self.scheme + ':'
2764 if self.user or self.passwd or self.host:
2764 if self.user or self.passwd or self.host:
2765 s += '//'
2765 s += '//'
2766 elif self.scheme and (not self.path or self.path.startswith('/')
2766 elif self.scheme and (not self.path or self.path.startswith('/')
2767 or hasdriveletter(self.path)):
2767 or hasdriveletter(self.path)):
2768 s += '//'
2768 s += '//'
2769 if hasdriveletter(self.path):
2769 if hasdriveletter(self.path):
2770 s += '/'
2770 s += '/'
2771 if self.user:
2771 if self.user:
2772 s += urlreq.quote(self.user, safe=self._safechars)
2772 s += urlreq.quote(self.user, safe=self._safechars)
2773 if self.passwd:
2773 if self.passwd:
2774 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2774 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2775 if self.user or self.passwd:
2775 if self.user or self.passwd:
2776 s += '@'
2776 s += '@'
2777 if self.host:
2777 if self.host:
2778 if not (self.host.startswith('[') and self.host.endswith(']')):
2778 if not (self.host.startswith('[') and self.host.endswith(']')):
2779 s += urlreq.quote(self.host)
2779 s += urlreq.quote(self.host)
2780 else:
2780 else:
2781 s += self.host
2781 s += self.host
2782 if self.port:
2782 if self.port:
2783 s += ':' + urlreq.quote(self.port)
2783 s += ':' + urlreq.quote(self.port)
2784 if self.host:
2784 if self.host:
2785 s += '/'
2785 s += '/'
2786 if self.path:
2786 if self.path:
2787 # TODO: similar to the query string, we should not unescape the
2787 # TODO: similar to the query string, we should not unescape the
2788 # path when we store it, the path might contain '%2f' = '/',
2788 # path when we store it, the path might contain '%2f' = '/',
2789 # which we should *not* escape.
2789 # which we should *not* escape.
2790 s += urlreq.quote(self.path, safe=self._safepchars)
2790 s += urlreq.quote(self.path, safe=self._safepchars)
2791 if self.query:
2791 if self.query:
2792 # we store the query in escaped form.
2792 # we store the query in escaped form.
2793 s += '?' + self.query
2793 s += '?' + self.query
2794 if self.fragment is not None:
2794 if self.fragment is not None:
2795 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2795 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2796 return s
2796 return s
2797
2797
2798 def authinfo(self):
2798 def authinfo(self):
2799 user, passwd = self.user, self.passwd
2799 user, passwd = self.user, self.passwd
2800 try:
2800 try:
2801 self.user, self.passwd = None, None
2801 self.user, self.passwd = None, None
2802 s = str(self)
2802 s = str(self)
2803 finally:
2803 finally:
2804 self.user, self.passwd = user, passwd
2804 self.user, self.passwd = user, passwd
2805 if not self.user:
2805 if not self.user:
2806 return (s, None)
2806 return (s, None)
2807 # authinfo[1] is passed to urllib2 password manager, and its
2807 # authinfo[1] is passed to urllib2 password manager, and its
2808 # URIs must not contain credentials. The host is passed in the
2808 # URIs must not contain credentials. The host is passed in the
2809 # URIs list because Python < 2.4.3 uses only that to search for
2809 # URIs list because Python < 2.4.3 uses only that to search for
2810 # a password.
2810 # a password.
2811 return (s, (None, (s, self.host),
2811 return (s, (None, (s, self.host),
2812 self.user, self.passwd or ''))
2812 self.user, self.passwd or ''))
2813
2813
2814 def isabs(self):
2814 def isabs(self):
2815 if self.scheme and self.scheme != 'file':
2815 if self.scheme and self.scheme != 'file':
2816 return True # remote URL
2816 return True # remote URL
2817 if hasdriveletter(self.path):
2817 if hasdriveletter(self.path):
2818 return True # absolute for our purposes - can't be joined()
2818 return True # absolute for our purposes - can't be joined()
2819 if self.path.startswith(r'\\'):
2819 if self.path.startswith(r'\\'):
2820 return True # Windows UNC path
2820 return True # Windows UNC path
2821 if self.path.startswith('/'):
2821 if self.path.startswith('/'):
2822 return True # POSIX-style
2822 return True # POSIX-style
2823 return False
2823 return False
2824
2824
2825 def localpath(self):
2825 def localpath(self):
2826 if self.scheme == 'file' or self.scheme == 'bundle':
2826 if self.scheme == 'file' or self.scheme == 'bundle':
2827 path = self.path or '/'
2827 path = self.path or '/'
2828 # For Windows, we need to promote hosts containing drive
2828 # For Windows, we need to promote hosts containing drive
2829 # letters to paths with drive letters.
2829 # letters to paths with drive letters.
2830 if hasdriveletter(self._hostport):
2830 if hasdriveletter(self._hostport):
2831 path = self._hostport + '/' + self.path
2831 path = self._hostport + '/' + self.path
2832 elif (self.host is not None and self.path
2832 elif (self.host is not None and self.path
2833 and not hasdriveletter(path)):
2833 and not hasdriveletter(path)):
2834 path = '/' + path
2834 path = '/' + path
2835 return path
2835 return path
2836 return self._origpath
2836 return self._origpath
2837
2837
2838 def islocal(self):
2838 def islocal(self):
2839 '''whether localpath will return something that posixfile can open'''
2839 '''whether localpath will return something that posixfile can open'''
2840 return (not self.scheme or self.scheme == 'file'
2840 return (not self.scheme or self.scheme == 'file'
2841 or self.scheme == 'bundle')
2841 or self.scheme == 'bundle')
2842
2842
2843 def hasscheme(path):
2843 def hasscheme(path):
2844 return bool(url(path).scheme)
2844 return bool(url(path).scheme)
2845
2845
2846 def hasdriveletter(path):
2846 def hasdriveletter(path):
2847 return path and path[1:2] == ':' and path[0:1].isalpha()
2847 return path and path[1:2] == ':' and path[0:1].isalpha()
2848
2848
2849 def urllocalpath(path):
2849 def urllocalpath(path):
2850 return url(path, parsequery=False, parsefragment=False).localpath()
2850 return url(path, parsequery=False, parsefragment=False).localpath()
2851
2851
2852 def hidepassword(u):
2852 def hidepassword(u):
2853 '''hide user credential in a url string'''
2853 '''hide user credential in a url string'''
2854 u = url(u)
2854 u = url(u)
2855 if u.passwd:
2855 if u.passwd:
2856 u.passwd = '***'
2856 u.passwd = '***'
2857 return str(u)
2857 return str(u)
2858
2858
2859 def removeauth(u):
2859 def removeauth(u):
2860 '''remove all authentication information from a url string'''
2860 '''remove all authentication information from a url string'''
2861 u = url(u)
2861 u = url(u)
2862 u.user = u.passwd = None
2862 u.user = u.passwd = None
2863 return str(u)
2863 return str(u)
2864
2864
2865 timecount = unitcountfn(
2865 timecount = unitcountfn(
2866 (1, 1e3, _('%.0f s')),
2866 (1, 1e3, _('%.0f s')),
2867 (100, 1, _('%.1f s')),
2867 (100, 1, _('%.1f s')),
2868 (10, 1, _('%.2f s')),
2868 (10, 1, _('%.2f s')),
2869 (1, 1, _('%.3f s')),
2869 (1, 1, _('%.3f s')),
2870 (100, 0.001, _('%.1f ms')),
2870 (100, 0.001, _('%.1f ms')),
2871 (10, 0.001, _('%.2f ms')),
2871 (10, 0.001, _('%.2f ms')),
2872 (1, 0.001, _('%.3f ms')),
2872 (1, 0.001, _('%.3f ms')),
2873 (100, 0.000001, _('%.1f us')),
2873 (100, 0.000001, _('%.1f us')),
2874 (10, 0.000001, _('%.2f us')),
2874 (10, 0.000001, _('%.2f us')),
2875 (1, 0.000001, _('%.3f us')),
2875 (1, 0.000001, _('%.3f us')),
2876 (100, 0.000000001, _('%.1f ns')),
2876 (100, 0.000000001, _('%.1f ns')),
2877 (10, 0.000000001, _('%.2f ns')),
2877 (10, 0.000000001, _('%.2f ns')),
2878 (1, 0.000000001, _('%.3f ns')),
2878 (1, 0.000000001, _('%.3f ns')),
2879 )
2879 )
2880
2880
2881 _timenesting = [0]
2881 _timenesting = [0]
2882
2882
2883 def timed(func):
2883 def timed(func):
2884 '''Report the execution time of a function call to stderr.
2884 '''Report the execution time of a function call to stderr.
2885
2885
2886 During development, use as a decorator when you need to measure
2886 During development, use as a decorator when you need to measure
2887 the cost of a function, e.g. as follows:
2887 the cost of a function, e.g. as follows:
2888
2888
2889 @util.timed
2889 @util.timed
2890 def foo(a, b, c):
2890 def foo(a, b, c):
2891 pass
2891 pass
2892 '''
2892 '''
2893
2893
2894 def wrapper(*args, **kwargs):
2894 def wrapper(*args, **kwargs):
2895 start = timer()
2895 start = timer()
2896 indent = 2
2896 indent = 2
2897 _timenesting[0] += indent
2897 _timenesting[0] += indent
2898 try:
2898 try:
2899 return func(*args, **kwargs)
2899 return func(*args, **kwargs)
2900 finally:
2900 finally:
2901 elapsed = timer() - start
2901 elapsed = timer() - start
2902 _timenesting[0] -= indent
2902 _timenesting[0] -= indent
2903 stderr.write('%s%s: %s\n' %
2903 stderr.write('%s%s: %s\n' %
2904 (' ' * _timenesting[0], func.__name__,
2904 (' ' * _timenesting[0], func.__name__,
2905 timecount(elapsed)))
2905 timecount(elapsed)))
2906 return wrapper
2906 return wrapper
2907
2907
2908 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2908 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2909 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2909 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2910
2910
2911 def sizetoint(s):
2911 def sizetoint(s):
2912 '''Convert a space specifier to a byte count.
2912 '''Convert a space specifier to a byte count.
2913
2913
2914 >>> sizetoint('30')
2914 >>> sizetoint('30')
2915 30
2915 30
2916 >>> sizetoint('2.2kb')
2916 >>> sizetoint('2.2kb')
2917 2252
2917 2252
2918 >>> sizetoint('6M')
2918 >>> sizetoint('6M')
2919 6291456
2919 6291456
2920 '''
2920 '''
2921 t = s.strip().lower()
2921 t = s.strip().lower()
2922 try:
2922 try:
2923 for k, u in _sizeunits:
2923 for k, u in _sizeunits:
2924 if t.endswith(k):
2924 if t.endswith(k):
2925 return int(float(t[:-len(k)]) * u)
2925 return int(float(t[:-len(k)]) * u)
2926 return int(t)
2926 return int(t)
2927 except ValueError:
2927 except ValueError:
2928 raise error.ParseError(_("couldn't parse size: %s") % s)
2928 raise error.ParseError(_("couldn't parse size: %s") % s)
2929
2929
2930 class hooks(object):
2930 class hooks(object):
2931 '''A collection of hook functions that can be used to extend a
2931 '''A collection of hook functions that can be used to extend a
2932 function's behavior. Hooks are called in lexicographic order,
2932 function's behavior. Hooks are called in lexicographic order,
2933 based on the names of their sources.'''
2933 based on the names of their sources.'''
2934
2934
2935 def __init__(self):
2935 def __init__(self):
2936 self._hooks = []
2936 self._hooks = []
2937
2937
2938 def add(self, source, hook):
2938 def add(self, source, hook):
2939 self._hooks.append((source, hook))
2939 self._hooks.append((source, hook))
2940
2940
2941 def __call__(self, *args):
2941 def __call__(self, *args):
2942 self._hooks.sort(key=lambda x: x[0])
2942 self._hooks.sort(key=lambda x: x[0])
2943 results = []
2943 results = []
2944 for source, hook in self._hooks:
2944 for source, hook in self._hooks:
2945 results.append(hook(*args))
2945 results.append(hook(*args))
2946 return results
2946 return results
2947
2947
2948 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2948 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2949 '''Yields lines for a nicely formatted stacktrace.
2949 '''Yields lines for a nicely formatted stacktrace.
2950 Skips the 'skip' last entries, then return the last 'depth' entries.
2950 Skips the 'skip' last entries, then return the last 'depth' entries.
2951 Each file+linenumber is formatted according to fileline.
2951 Each file+linenumber is formatted according to fileline.
2952 Each line is formatted according to line.
2952 Each line is formatted according to line.
2953 If line is None, it yields:
2953 If line is None, it yields:
2954 length of longest filepath+line number,
2954 length of longest filepath+line number,
2955 filepath+linenumber,
2955 filepath+linenumber,
2956 function
2956 function
2957
2957
2958 Not be used in production code but very convenient while developing.
2958 Not be used in production code but very convenient while developing.
2959 '''
2959 '''
2960 entries = [(fileline % (fn, ln), func)
2960 entries = [(fileline % (fn, ln), func)
2961 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2961 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2962 ][-depth:]
2962 ][-depth:]
2963 if entries:
2963 if entries:
2964 fnmax = max(len(entry[0]) for entry in entries)
2964 fnmax = max(len(entry[0]) for entry in entries)
2965 for fnln, func in entries:
2965 for fnln, func in entries:
2966 if line is None:
2966 if line is None:
2967 yield (fnmax, fnln, func)
2967 yield (fnmax, fnln, func)
2968 else:
2968 else:
2969 yield line % (fnmax, fnln, func)
2969 yield line % (fnmax, fnln, func)
2970
2970
2971 def debugstacktrace(msg='stacktrace', skip=0,
2971 def debugstacktrace(msg='stacktrace', skip=0,
2972 f=stderr, otherf=stdout, depth=0):
2972 f=stderr, otherf=stdout, depth=0):
2973 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2973 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2974 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2974 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2975 By default it will flush stdout first.
2975 By default it will flush stdout first.
2976 It can be used everywhere and intentionally does not require an ui object.
2976 It can be used everywhere and intentionally does not require an ui object.
2977 Not be used in production code but very convenient while developing.
2977 Not be used in production code but very convenient while developing.
2978 '''
2978 '''
2979 if otherf:
2979 if otherf:
2980 otherf.flush()
2980 otherf.flush()
2981 f.write('%s at:\n' % msg.rstrip())
2981 f.write('%s at:\n' % msg.rstrip())
2982 for line in getstackframes(skip + 1, depth=depth):
2982 for line in getstackframes(skip + 1, depth=depth):
2983 f.write(line)
2983 f.write(line)
2984 f.flush()
2984 f.flush()
2985
2985
2986 class dirs(object):
2986 class dirs(object):
2987 '''a multiset of directory names from a dirstate or manifest'''
2987 '''a multiset of directory names from a dirstate or manifest'''
2988
2988
2989 def __init__(self, map, skip=None):
2989 def __init__(self, map, skip=None):
2990 self._dirs = {}
2990 self._dirs = {}
2991 addpath = self.addpath
2991 addpath = self.addpath
2992 if safehasattr(map, 'iteritems') and skip is not None:
2992 if safehasattr(map, 'iteritems') and skip is not None:
2993 for f, s in map.iteritems():
2993 for f, s in map.iteritems():
2994 if s[0] != skip:
2994 if s[0] != skip:
2995 addpath(f)
2995 addpath(f)
2996 else:
2996 else:
2997 for f in map:
2997 for f in map:
2998 addpath(f)
2998 addpath(f)
2999
2999
3000 def addpath(self, path):
3000 def addpath(self, path):
3001 dirs = self._dirs
3001 dirs = self._dirs
3002 for base in finddirs(path):
3002 for base in finddirs(path):
3003 if base in dirs:
3003 if base in dirs:
3004 dirs[base] += 1
3004 dirs[base] += 1
3005 return
3005 return
3006 dirs[base] = 1
3006 dirs[base] = 1
3007
3007
3008 def delpath(self, path):
3008 def delpath(self, path):
3009 dirs = self._dirs
3009 dirs = self._dirs
3010 for base in finddirs(path):
3010 for base in finddirs(path):
3011 if dirs[base] > 1:
3011 if dirs[base] > 1:
3012 dirs[base] -= 1
3012 dirs[base] -= 1
3013 return
3013 return
3014 del dirs[base]
3014 del dirs[base]
3015
3015
3016 def __iter__(self):
3016 def __iter__(self):
3017 return iter(self._dirs)
3017 return iter(self._dirs)
3018
3018
3019 def __contains__(self, d):
3019 def __contains__(self, d):
3020 return d in self._dirs
3020 return d in self._dirs
3021
3021
3022 if safehasattr(parsers, 'dirs'):
3022 if safehasattr(parsers, 'dirs'):
3023 dirs = parsers.dirs
3023 dirs = parsers.dirs
3024
3024
3025 def finddirs(path):
3025 def finddirs(path):
3026 pos = path.rfind('/')
3026 pos = path.rfind('/')
3027 while pos != -1:
3027 while pos != -1:
3028 yield path[:pos]
3028 yield path[:pos]
3029 pos = path.rfind('/', 0, pos)
3029 pos = path.rfind('/', 0, pos)
3030
3030
3031 class ctxmanager(object):
3031 class ctxmanager(object):
3032 '''A context manager for use in 'with' blocks to allow multiple
3032 '''A context manager for use in 'with' blocks to allow multiple
3033 contexts to be entered at once. This is both safer and more
3033 contexts to be entered at once. This is both safer and more
3034 flexible than contextlib.nested.
3034 flexible than contextlib.nested.
3035
3035
3036 Once Mercurial supports Python 2.7+, this will become mostly
3036 Once Mercurial supports Python 2.7+, this will become mostly
3037 unnecessary.
3037 unnecessary.
3038 '''
3038 '''
3039
3039
3040 def __init__(self, *args):
3040 def __init__(self, *args):
3041 '''Accepts a list of no-argument functions that return context
3041 '''Accepts a list of no-argument functions that return context
3042 managers. These will be invoked at __call__ time.'''
3042 managers. These will be invoked at __call__ time.'''
3043 self._pending = args
3043 self._pending = args
3044 self._atexit = []
3044 self._atexit = []
3045
3045
3046 def __enter__(self):
3046 def __enter__(self):
3047 return self
3047 return self
3048
3048
3049 def enter(self):
3049 def enter(self):
3050 '''Create and enter context managers in the order in which they were
3050 '''Create and enter context managers in the order in which they were
3051 passed to the constructor.'''
3051 passed to the constructor.'''
3052 values = []
3052 values = []
3053 for func in self._pending:
3053 for func in self._pending:
3054 obj = func()
3054 obj = func()
3055 values.append(obj.__enter__())
3055 values.append(obj.__enter__())
3056 self._atexit.append(obj.__exit__)
3056 self._atexit.append(obj.__exit__)
3057 del self._pending
3057 del self._pending
3058 return values
3058 return values
3059
3059
3060 def atexit(self, func, *args, **kwargs):
3060 def atexit(self, func, *args, **kwargs):
3061 '''Add a function to call when this context manager exits. The
3061 '''Add a function to call when this context manager exits. The
3062 ordering of multiple atexit calls is unspecified, save that
3062 ordering of multiple atexit calls is unspecified, save that
3063 they will happen before any __exit__ functions.'''
3063 they will happen before any __exit__ functions.'''
3064 def wrapper(exc_type, exc_val, exc_tb):
3064 def wrapper(exc_type, exc_val, exc_tb):
3065 func(*args, **kwargs)
3065 func(*args, **kwargs)
3066 self._atexit.append(wrapper)
3066 self._atexit.append(wrapper)
3067 return func
3067 return func
3068
3068
3069 def __exit__(self, exc_type, exc_val, exc_tb):
3069 def __exit__(self, exc_type, exc_val, exc_tb):
3070 '''Context managers are exited in the reverse order from which
3070 '''Context managers are exited in the reverse order from which
3071 they were created.'''
3071 they were created.'''
3072 received = exc_type is not None
3072 received = exc_type is not None
3073 suppressed = False
3073 suppressed = False
3074 pending = None
3074 pending = None
3075 self._atexit.reverse()
3075 self._atexit.reverse()
3076 for exitfunc in self._atexit:
3076 for exitfunc in self._atexit:
3077 try:
3077 try:
3078 if exitfunc(exc_type, exc_val, exc_tb):
3078 if exitfunc(exc_type, exc_val, exc_tb):
3079 suppressed = True
3079 suppressed = True
3080 exc_type = None
3080 exc_type = None
3081 exc_val = None
3081 exc_val = None
3082 exc_tb = None
3082 exc_tb = None
3083 except BaseException:
3083 except BaseException:
3084 pending = sys.exc_info()
3084 pending = sys.exc_info()
3085 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3085 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3086 del self._atexit
3086 del self._atexit
3087 if pending:
3087 if pending:
3088 raise exc_val
3088 raise exc_val
3089 return received and suppressed
3089 return received and suppressed
3090
3090
3091 # compression code
3091 # compression code
3092
3092
3093 SERVERROLE = 'server'
3093 SERVERROLE = 'server'
3094 CLIENTROLE = 'client'
3094 CLIENTROLE = 'client'
3095
3095
3096 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3096 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3097 (u'name', u'serverpriority',
3097 (u'name', u'serverpriority',
3098 u'clientpriority'))
3098 u'clientpriority'))
3099
3099
3100 class compressormanager(object):
3100 class compressormanager(object):
3101 """Holds registrations of various compression engines.
3101 """Holds registrations of various compression engines.
3102
3102
3103 This class essentially abstracts the differences between compression
3103 This class essentially abstracts the differences between compression
3104 engines to allow new compression formats to be added easily, possibly from
3104 engines to allow new compression formats to be added easily, possibly from
3105 extensions.
3105 extensions.
3106
3106
3107 Compressors are registered against the global instance by calling its
3107 Compressors are registered against the global instance by calling its
3108 ``register()`` method.
3108 ``register()`` method.
3109 """
3109 """
3110 def __init__(self):
3110 def __init__(self):
3111 self._engines = {}
3111 self._engines = {}
3112 # Bundle spec human name to engine name.
3112 # Bundle spec human name to engine name.
3113 self._bundlenames = {}
3113 self._bundlenames = {}
3114 # Internal bundle identifier to engine name.
3114 # Internal bundle identifier to engine name.
3115 self._bundletypes = {}
3115 self._bundletypes = {}
3116 # Revlog header to engine name.
3116 # Revlog header to engine name.
3117 self._revlogheaders = {}
3117 self._revlogheaders = {}
3118 # Wire proto identifier to engine name.
3118 # Wire proto identifier to engine name.
3119 self._wiretypes = {}
3119 self._wiretypes = {}
3120
3120
3121 def __getitem__(self, key):
3121 def __getitem__(self, key):
3122 return self._engines[key]
3122 return self._engines[key]
3123
3123
3124 def __contains__(self, key):
3124 def __contains__(self, key):
3125 return key in self._engines
3125 return key in self._engines
3126
3126
3127 def __iter__(self):
3127 def __iter__(self):
3128 return iter(self._engines.keys())
3128 return iter(self._engines.keys())
3129
3129
3130 def register(self, engine):
3130 def register(self, engine):
3131 """Register a compression engine with the manager.
3131 """Register a compression engine with the manager.
3132
3132
3133 The argument must be a ``compressionengine`` instance.
3133 The argument must be a ``compressionengine`` instance.
3134 """
3134 """
3135 if not isinstance(engine, compressionengine):
3135 if not isinstance(engine, compressionengine):
3136 raise ValueError(_('argument must be a compressionengine'))
3136 raise ValueError(_('argument must be a compressionengine'))
3137
3137
3138 name = engine.name()
3138 name = engine.name()
3139
3139
3140 if name in self._engines:
3140 if name in self._engines:
3141 raise error.Abort(_('compression engine %s already registered') %
3141 raise error.Abort(_('compression engine %s already registered') %
3142 name)
3142 name)
3143
3143
3144 bundleinfo = engine.bundletype()
3144 bundleinfo = engine.bundletype()
3145 if bundleinfo:
3145 if bundleinfo:
3146 bundlename, bundletype = bundleinfo
3146 bundlename, bundletype = bundleinfo
3147
3147
3148 if bundlename in self._bundlenames:
3148 if bundlename in self._bundlenames:
3149 raise error.Abort(_('bundle name %s already registered') %
3149 raise error.Abort(_('bundle name %s already registered') %
3150 bundlename)
3150 bundlename)
3151 if bundletype in self._bundletypes:
3151 if bundletype in self._bundletypes:
3152 raise error.Abort(_('bundle type %s already registered by %s') %
3152 raise error.Abort(_('bundle type %s already registered by %s') %
3153 (bundletype, self._bundletypes[bundletype]))
3153 (bundletype, self._bundletypes[bundletype]))
3154
3154
3155 # No external facing name declared.
3155 # No external facing name declared.
3156 if bundlename:
3156 if bundlename:
3157 self._bundlenames[bundlename] = name
3157 self._bundlenames[bundlename] = name
3158
3158
3159 self._bundletypes[bundletype] = name
3159 self._bundletypes[bundletype] = name
3160
3160
3161 wiresupport = engine.wireprotosupport()
3161 wiresupport = engine.wireprotosupport()
3162 if wiresupport:
3162 if wiresupport:
3163 wiretype = wiresupport.name
3163 wiretype = wiresupport.name
3164 if wiretype in self._wiretypes:
3164 if wiretype in self._wiretypes:
3165 raise error.Abort(_('wire protocol compression %s already '
3165 raise error.Abort(_('wire protocol compression %s already '
3166 'registered by %s') %
3166 'registered by %s') %
3167 (wiretype, self._wiretypes[wiretype]))
3167 (wiretype, self._wiretypes[wiretype]))
3168
3168
3169 self._wiretypes[wiretype] = name
3169 self._wiretypes[wiretype] = name
3170
3170
3171 revlogheader = engine.revlogheader()
3171 revlogheader = engine.revlogheader()
3172 if revlogheader and revlogheader in self._revlogheaders:
3172 if revlogheader and revlogheader in self._revlogheaders:
3173 raise error.Abort(_('revlog header %s already registered by %s') %
3173 raise error.Abort(_('revlog header %s already registered by %s') %
3174 (revlogheader, self._revlogheaders[revlogheader]))
3174 (revlogheader, self._revlogheaders[revlogheader]))
3175
3175
3176 if revlogheader:
3176 if revlogheader:
3177 self._revlogheaders[revlogheader] = name
3177 self._revlogheaders[revlogheader] = name
3178
3178
3179 self._engines[name] = engine
3179 self._engines[name] = engine
3180
3180
3181 @property
3181 @property
3182 def supportedbundlenames(self):
3182 def supportedbundlenames(self):
3183 return set(self._bundlenames.keys())
3183 return set(self._bundlenames.keys())
3184
3184
3185 @property
3185 @property
3186 def supportedbundletypes(self):
3186 def supportedbundletypes(self):
3187 return set(self._bundletypes.keys())
3187 return set(self._bundletypes.keys())
3188
3188
3189 def forbundlename(self, bundlename):
3189 def forbundlename(self, bundlename):
3190 """Obtain a compression engine registered to a bundle name.
3190 """Obtain a compression engine registered to a bundle name.
3191
3191
3192 Will raise KeyError if the bundle type isn't registered.
3192 Will raise KeyError if the bundle type isn't registered.
3193
3193
3194 Will abort if the engine is known but not available.
3194 Will abort if the engine is known but not available.
3195 """
3195 """
3196 engine = self._engines[self._bundlenames[bundlename]]
3196 engine = self._engines[self._bundlenames[bundlename]]
3197 if not engine.available():
3197 if not engine.available():
3198 raise error.Abort(_('compression engine %s could not be loaded') %
3198 raise error.Abort(_('compression engine %s could not be loaded') %
3199 engine.name())
3199 engine.name())
3200 return engine
3200 return engine
3201
3201
3202 def forbundletype(self, bundletype):
3202 def forbundletype(self, bundletype):
3203 """Obtain a compression engine registered to a bundle type.
3203 """Obtain a compression engine registered to a bundle type.
3204
3204
3205 Will raise KeyError if the bundle type isn't registered.
3205 Will raise KeyError if the bundle type isn't registered.
3206
3206
3207 Will abort if the engine is known but not available.
3207 Will abort if the engine is known but not available.
3208 """
3208 """
3209 engine = self._engines[self._bundletypes[bundletype]]
3209 engine = self._engines[self._bundletypes[bundletype]]
3210 if not engine.available():
3210 if not engine.available():
3211 raise error.Abort(_('compression engine %s could not be loaded') %
3211 raise error.Abort(_('compression engine %s could not be loaded') %
3212 engine.name())
3212 engine.name())
3213 return engine
3213 return engine
3214
3214
3215 def supportedwireengines(self, role, onlyavailable=True):
3215 def supportedwireengines(self, role, onlyavailable=True):
3216 """Obtain compression engines that support the wire protocol.
3216 """Obtain compression engines that support the wire protocol.
3217
3217
3218 Returns a list of engines in prioritized order, most desired first.
3218 Returns a list of engines in prioritized order, most desired first.
3219
3219
3220 If ``onlyavailable`` is set, filter out engines that can't be
3220 If ``onlyavailable`` is set, filter out engines that can't be
3221 loaded.
3221 loaded.
3222 """
3222 """
3223 assert role in (SERVERROLE, CLIENTROLE)
3223 assert role in (SERVERROLE, CLIENTROLE)
3224
3224
3225 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3225 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3226
3226
3227 engines = [self._engines[e] for e in self._wiretypes.values()]
3227 engines = [self._engines[e] for e in self._wiretypes.values()]
3228 if onlyavailable:
3228 if onlyavailable:
3229 engines = [e for e in engines if e.available()]
3229 engines = [e for e in engines if e.available()]
3230
3230
3231 def getkey(e):
3231 def getkey(e):
3232 # Sort first by priority, highest first. In case of tie, sort
3232 # Sort first by priority, highest first. In case of tie, sort
3233 # alphabetically. This is arbitrary, but ensures output is
3233 # alphabetically. This is arbitrary, but ensures output is
3234 # stable.
3234 # stable.
3235 w = e.wireprotosupport()
3235 w = e.wireprotosupport()
3236 return -1 * getattr(w, attr), w.name
3236 return -1 * getattr(w, attr), w.name
3237
3237
3238 return list(sorted(engines, key=getkey))
3238 return list(sorted(engines, key=getkey))
3239
3239
3240 def forwiretype(self, wiretype):
3240 def forwiretype(self, wiretype):
3241 engine = self._engines[self._wiretypes[wiretype]]
3241 engine = self._engines[self._wiretypes[wiretype]]
3242 if not engine.available():
3242 if not engine.available():
3243 raise error.Abort(_('compression engine %s could not be loaded') %
3243 raise error.Abort(_('compression engine %s could not be loaded') %
3244 engine.name())
3244 engine.name())
3245 return engine
3245 return engine
3246
3246
3247 def forrevlogheader(self, header):
3247 def forrevlogheader(self, header):
3248 """Obtain a compression engine registered to a revlog header.
3248 """Obtain a compression engine registered to a revlog header.
3249
3249
3250 Will raise KeyError if the revlog header value isn't registered.
3250 Will raise KeyError if the revlog header value isn't registered.
3251 """
3251 """
3252 return self._engines[self._revlogheaders[header]]
3252 return self._engines[self._revlogheaders[header]]
3253
3253
3254 compengines = compressormanager()
3254 compengines = compressormanager()
3255
3255
3256 class compressionengine(object):
3256 class compressionengine(object):
3257 """Base class for compression engines.
3257 """Base class for compression engines.
3258
3258
3259 Compression engines must implement the interface defined by this class.
3259 Compression engines must implement the interface defined by this class.
3260 """
3260 """
3261 def name(self):
3261 def name(self):
3262 """Returns the name of the compression engine.
3262 """Returns the name of the compression engine.
3263
3263
3264 This is the key the engine is registered under.
3264 This is the key the engine is registered under.
3265
3265
3266 This method must be implemented.
3266 This method must be implemented.
3267 """
3267 """
3268 raise NotImplementedError()
3268 raise NotImplementedError()
3269
3269
3270 def available(self):
3270 def available(self):
3271 """Whether the compression engine is available.
3271 """Whether the compression engine is available.
3272
3272
3273 The intent of this method is to allow optional compression engines
3273 The intent of this method is to allow optional compression engines
3274 that may not be available in all installations (such as engines relying
3274 that may not be available in all installations (such as engines relying
3275 on C extensions that may not be present).
3275 on C extensions that may not be present).
3276 """
3276 """
3277 return True
3277 return True
3278
3278
3279 def bundletype(self):
3279 def bundletype(self):
3280 """Describes bundle identifiers for this engine.
3280 """Describes bundle identifiers for this engine.
3281
3281
3282 If this compression engine isn't supported for bundles, returns None.
3282 If this compression engine isn't supported for bundles, returns None.
3283
3283
3284 If this engine can be used for bundles, returns a 2-tuple of strings of
3284 If this engine can be used for bundles, returns a 2-tuple of strings of
3285 the user-facing "bundle spec" compression name and an internal
3285 the user-facing "bundle spec" compression name and an internal
3286 identifier used to denote the compression format within bundles. To
3286 identifier used to denote the compression format within bundles. To
3287 exclude the name from external usage, set the first element to ``None``.
3287 exclude the name from external usage, set the first element to ``None``.
3288
3288
3289 If bundle compression is supported, the class must also implement
3289 If bundle compression is supported, the class must also implement
3290 ``compressstream`` and `decompressorreader``.
3290 ``compressstream`` and `decompressorreader``.
3291
3291
3292 The docstring of this method is used in the help system to tell users
3292 The docstring of this method is used in the help system to tell users
3293 about this engine.
3293 about this engine.
3294 """
3294 """
3295 return None
3295 return None
3296
3296
3297 def wireprotosupport(self):
3297 def wireprotosupport(self):
3298 """Declare support for this compression format on the wire protocol.
3298 """Declare support for this compression format on the wire protocol.
3299
3299
3300 If this compression engine isn't supported for compressing wire
3300 If this compression engine isn't supported for compressing wire
3301 protocol payloads, returns None.
3301 protocol payloads, returns None.
3302
3302
3303 Otherwise, returns ``compenginewireprotosupport`` with the following
3303 Otherwise, returns ``compenginewireprotosupport`` with the following
3304 fields:
3304 fields:
3305
3305
3306 * String format identifier
3306 * String format identifier
3307 * Integer priority for the server
3307 * Integer priority for the server
3308 * Integer priority for the client
3308 * Integer priority for the client
3309
3309
3310 The integer priorities are used to order the advertisement of format
3310 The integer priorities are used to order the advertisement of format
3311 support by server and client. The highest integer is advertised
3311 support by server and client. The highest integer is advertised
3312 first. Integers with non-positive values aren't advertised.
3312 first. Integers with non-positive values aren't advertised.
3313
3313
3314 The priority values are somewhat arbitrary and only used for default
3314 The priority values are somewhat arbitrary and only used for default
3315 ordering. The relative order can be changed via config options.
3315 ordering. The relative order can be changed via config options.
3316
3316
3317 If wire protocol compression is supported, the class must also implement
3317 If wire protocol compression is supported, the class must also implement
3318 ``compressstream`` and ``decompressorreader``.
3318 ``compressstream`` and ``decompressorreader``.
3319 """
3319 """
3320 return None
3320 return None
3321
3321
3322 def revlogheader(self):
3322 def revlogheader(self):
3323 """Header added to revlog chunks that identifies this engine.
3323 """Header added to revlog chunks that identifies this engine.
3324
3324
3325 If this engine can be used to compress revlogs, this method should
3325 If this engine can be used to compress revlogs, this method should
3326 return the bytes used to identify chunks compressed with this engine.
3326 return the bytes used to identify chunks compressed with this engine.
3327 Else, the method should return ``None`` to indicate it does not
3327 Else, the method should return ``None`` to indicate it does not
3328 participate in revlog compression.
3328 participate in revlog compression.
3329 """
3329 """
3330 return None
3330 return None
3331
3331
3332 def compressstream(self, it, opts=None):
3332 def compressstream(self, it, opts=None):
3333 """Compress an iterator of chunks.
3333 """Compress an iterator of chunks.
3334
3334
3335 The method receives an iterator (ideally a generator) of chunks of
3335 The method receives an iterator (ideally a generator) of chunks of
3336 bytes to be compressed. It returns an iterator (ideally a generator)
3336 bytes to be compressed. It returns an iterator (ideally a generator)
3337 of bytes of chunks representing the compressed output.
3337 of bytes of chunks representing the compressed output.
3338
3338
3339 Optionally accepts an argument defining how to perform compression.
3339 Optionally accepts an argument defining how to perform compression.
3340 Each engine treats this argument differently.
3340 Each engine treats this argument differently.
3341 """
3341 """
3342 raise NotImplementedError()
3342 raise NotImplementedError()
3343
3343
3344 def decompressorreader(self, fh):
3344 def decompressorreader(self, fh):
3345 """Perform decompression on a file object.
3345 """Perform decompression on a file object.
3346
3346
3347 Argument is an object with a ``read(size)`` method that returns
3347 Argument is an object with a ``read(size)`` method that returns
3348 compressed data. Return value is an object with a ``read(size)`` that
3348 compressed data. Return value is an object with a ``read(size)`` that
3349 returns uncompressed data.
3349 returns uncompressed data.
3350 """
3350 """
3351 raise NotImplementedError()
3351 raise NotImplementedError()
3352
3352
3353 def revlogcompressor(self, opts=None):
3353 def revlogcompressor(self, opts=None):
3354 """Obtain an object that can be used to compress revlog entries.
3354 """Obtain an object that can be used to compress revlog entries.
3355
3355
3356 The object has a ``compress(data)`` method that compresses binary
3356 The object has a ``compress(data)`` method that compresses binary
3357 data. This method returns compressed binary data or ``None`` if
3357 data. This method returns compressed binary data or ``None`` if
3358 the data could not be compressed (too small, not compressible, etc).
3358 the data could not be compressed (too small, not compressible, etc).
3359 The returned data should have a header uniquely identifying this
3359 The returned data should have a header uniquely identifying this
3360 compression format so decompression can be routed to this engine.
3360 compression format so decompression can be routed to this engine.
3361 This header should be identified by the ``revlogheader()`` return
3361 This header should be identified by the ``revlogheader()`` return
3362 value.
3362 value.
3363
3363
3364 The object has a ``decompress(data)`` method that decompresses
3364 The object has a ``decompress(data)`` method that decompresses
3365 data. The method will only be called if ``data`` begins with
3365 data. The method will only be called if ``data`` begins with
3366 ``revlogheader()``. The method should return the raw, uncompressed
3366 ``revlogheader()``. The method should return the raw, uncompressed
3367 data or raise a ``RevlogError``.
3367 data or raise a ``RevlogError``.
3368
3368
3369 The object is reusable but is not thread safe.
3369 The object is reusable but is not thread safe.
3370 """
3370 """
3371 raise NotImplementedError()
3371 raise NotImplementedError()
3372
3372
3373 class _zlibengine(compressionengine):
3373 class _zlibengine(compressionengine):
3374 def name(self):
3374 def name(self):
3375 return 'zlib'
3375 return 'zlib'
3376
3376
3377 def bundletype(self):
3377 def bundletype(self):
3378 """zlib compression using the DEFLATE algorithm.
3378 """zlib compression using the DEFLATE algorithm.
3379
3379
3380 All Mercurial clients should support this format. The compression
3380 All Mercurial clients should support this format. The compression
3381 algorithm strikes a reasonable balance between compression ratio
3381 algorithm strikes a reasonable balance between compression ratio
3382 and size.
3382 and size.
3383 """
3383 """
3384 return 'gzip', 'GZ'
3384 return 'gzip', 'GZ'
3385
3385
3386 def wireprotosupport(self):
3386 def wireprotosupport(self):
3387 return compewireprotosupport('zlib', 20, 20)
3387 return compewireprotosupport('zlib', 20, 20)
3388
3388
3389 def revlogheader(self):
3389 def revlogheader(self):
3390 return 'x'
3390 return 'x'
3391
3391
3392 def compressstream(self, it, opts=None):
3392 def compressstream(self, it, opts=None):
3393 opts = opts or {}
3393 opts = opts or {}
3394
3394
3395 z = zlib.compressobj(opts.get('level', -1))
3395 z = zlib.compressobj(opts.get('level', -1))
3396 for chunk in it:
3396 for chunk in it:
3397 data = z.compress(chunk)
3397 data = z.compress(chunk)
3398 # Not all calls to compress emit data. It is cheaper to inspect
3398 # Not all calls to compress emit data. It is cheaper to inspect
3399 # here than to feed empty chunks through generator.
3399 # here than to feed empty chunks through generator.
3400 if data:
3400 if data:
3401 yield data
3401 yield data
3402
3402
3403 yield z.flush()
3403 yield z.flush()
3404
3404
3405 def decompressorreader(self, fh):
3405 def decompressorreader(self, fh):
3406 def gen():
3406 def gen():
3407 d = zlib.decompressobj()
3407 d = zlib.decompressobj()
3408 for chunk in filechunkiter(fh):
3408 for chunk in filechunkiter(fh):
3409 while chunk:
3409 while chunk:
3410 # Limit output size to limit memory.
3410 # Limit output size to limit memory.
3411 yield d.decompress(chunk, 2 ** 18)
3411 yield d.decompress(chunk, 2 ** 18)
3412 chunk = d.unconsumed_tail
3412 chunk = d.unconsumed_tail
3413
3413
3414 return chunkbuffer(gen())
3414 return chunkbuffer(gen())
3415
3415
3416 class zlibrevlogcompressor(object):
3416 class zlibrevlogcompressor(object):
3417 def compress(self, data):
3417 def compress(self, data):
3418 insize = len(data)
3418 insize = len(data)
3419 # Caller handles empty input case.
3419 # Caller handles empty input case.
3420 assert insize > 0
3420 assert insize > 0
3421
3421
3422 if insize < 44:
3422 if insize < 44:
3423 return None
3423 return None
3424
3424
3425 elif insize <= 1000000:
3425 elif insize <= 1000000:
3426 compressed = zlib.compress(data)
3426 compressed = zlib.compress(data)
3427 if len(compressed) < insize:
3427 if len(compressed) < insize:
3428 return compressed
3428 return compressed
3429 return None
3429 return None
3430
3430
3431 # zlib makes an internal copy of the input buffer, doubling
3431 # zlib makes an internal copy of the input buffer, doubling
3432 # memory usage for large inputs. So do streaming compression
3432 # memory usage for large inputs. So do streaming compression
3433 # on large inputs.
3433 # on large inputs.
3434 else:
3434 else:
3435 z = zlib.compressobj()
3435 z = zlib.compressobj()
3436 parts = []
3436 parts = []
3437 pos = 0
3437 pos = 0
3438 while pos < insize:
3438 while pos < insize:
3439 pos2 = pos + 2**20
3439 pos2 = pos + 2**20
3440 parts.append(z.compress(data[pos:pos2]))
3440 parts.append(z.compress(data[pos:pos2]))
3441 pos = pos2
3441 pos = pos2
3442 parts.append(z.flush())
3442 parts.append(z.flush())
3443
3443
3444 if sum(map(len, parts)) < insize:
3444 if sum(map(len, parts)) < insize:
3445 return ''.join(parts)
3445 return ''.join(parts)
3446 return None
3446 return None
3447
3447
3448 def decompress(self, data):
3448 def decompress(self, data):
3449 try:
3449 try:
3450 return zlib.decompress(data)
3450 return zlib.decompress(data)
3451 except zlib.error as e:
3451 except zlib.error as e:
3452 raise error.RevlogError(_('revlog decompress error: %s') %
3452 raise error.RevlogError(_('revlog decompress error: %s') %
3453 str(e))
3453 str(e))
3454
3454
3455 def revlogcompressor(self, opts=None):
3455 def revlogcompressor(self, opts=None):
3456 return self.zlibrevlogcompressor()
3456 return self.zlibrevlogcompressor()
3457
3457
3458 compengines.register(_zlibengine())
3458 compengines.register(_zlibengine())
3459
3459
3460 class _bz2engine(compressionengine):
3460 class _bz2engine(compressionengine):
3461 def name(self):
3461 def name(self):
3462 return 'bz2'
3462 return 'bz2'
3463
3463
3464 def bundletype(self):
3464 def bundletype(self):
3465 """An algorithm that produces smaller bundles than ``gzip``.
3465 """An algorithm that produces smaller bundles than ``gzip``.
3466
3466
3467 All Mercurial clients should support this format.
3467 All Mercurial clients should support this format.
3468
3468
3469 This engine will likely produce smaller bundles than ``gzip`` but
3469 This engine will likely produce smaller bundles than ``gzip`` but
3470 will be significantly slower, both during compression and
3470 will be significantly slower, both during compression and
3471 decompression.
3471 decompression.
3472
3472
3473 If available, the ``zstd`` engine can yield similar or better
3473 If available, the ``zstd`` engine can yield similar or better
3474 compression at much higher speeds.
3474 compression at much higher speeds.
3475 """
3475 """
3476 return 'bzip2', 'BZ'
3476 return 'bzip2', 'BZ'
3477
3477
3478 # We declare a protocol name but don't advertise by default because
3478 # We declare a protocol name but don't advertise by default because
3479 # it is slow.
3479 # it is slow.
3480 def wireprotosupport(self):
3480 def wireprotosupport(self):
3481 return compewireprotosupport('bzip2', 0, 0)
3481 return compewireprotosupport('bzip2', 0, 0)
3482
3482
3483 def compressstream(self, it, opts=None):
3483 def compressstream(self, it, opts=None):
3484 opts = opts or {}
3484 opts = opts or {}
3485 z = bz2.BZ2Compressor(opts.get('level', 9))
3485 z = bz2.BZ2Compressor(opts.get('level', 9))
3486 for chunk in it:
3486 for chunk in it:
3487 data = z.compress(chunk)
3487 data = z.compress(chunk)
3488 if data:
3488 if data:
3489 yield data
3489 yield data
3490
3490
3491 yield z.flush()
3491 yield z.flush()
3492
3492
3493 def decompressorreader(self, fh):
3493 def decompressorreader(self, fh):
3494 def gen():
3494 def gen():
3495 d = bz2.BZ2Decompressor()
3495 d = bz2.BZ2Decompressor()
3496 for chunk in filechunkiter(fh):
3496 for chunk in filechunkiter(fh):
3497 yield d.decompress(chunk)
3497 yield d.decompress(chunk)
3498
3498
3499 return chunkbuffer(gen())
3499 return chunkbuffer(gen())
3500
3500
3501 compengines.register(_bz2engine())
3501 compengines.register(_bz2engine())
3502
3502
3503 class _truncatedbz2engine(compressionengine):
3503 class _truncatedbz2engine(compressionengine):
3504 def name(self):
3504 def name(self):
3505 return 'bz2truncated'
3505 return 'bz2truncated'
3506
3506
3507 def bundletype(self):
3507 def bundletype(self):
3508 return None, '_truncatedBZ'
3508 return None, '_truncatedBZ'
3509
3509
3510 # We don't implement compressstream because it is hackily handled elsewhere.
3510 # We don't implement compressstream because it is hackily handled elsewhere.
3511
3511
3512 def decompressorreader(self, fh):
3512 def decompressorreader(self, fh):
3513 def gen():
3513 def gen():
3514 # The input stream doesn't have the 'BZ' header. So add it back.
3514 # The input stream doesn't have the 'BZ' header. So add it back.
3515 d = bz2.BZ2Decompressor()
3515 d = bz2.BZ2Decompressor()
3516 d.decompress('BZ')
3516 d.decompress('BZ')
3517 for chunk in filechunkiter(fh):
3517 for chunk in filechunkiter(fh):
3518 yield d.decompress(chunk)
3518 yield d.decompress(chunk)
3519
3519
3520 return chunkbuffer(gen())
3520 return chunkbuffer(gen())
3521
3521
3522 compengines.register(_truncatedbz2engine())
3522 compengines.register(_truncatedbz2engine())
3523
3523
3524 class _noopengine(compressionengine):
3524 class _noopengine(compressionengine):
3525 def name(self):
3525 def name(self):
3526 return 'none'
3526 return 'none'
3527
3527
3528 def bundletype(self):
3528 def bundletype(self):
3529 """No compression is performed.
3529 """No compression is performed.
3530
3530
3531 Use this compression engine to explicitly disable compression.
3531 Use this compression engine to explicitly disable compression.
3532 """
3532 """
3533 return 'none', 'UN'
3533 return 'none', 'UN'
3534
3534
3535 # Clients always support uncompressed payloads. Servers don't because
3535 # Clients always support uncompressed payloads. Servers don't because
3536 # unless you are on a fast network, uncompressed payloads can easily
3536 # unless you are on a fast network, uncompressed payloads can easily
3537 # saturate your network pipe.
3537 # saturate your network pipe.
3538 def wireprotosupport(self):
3538 def wireprotosupport(self):
3539 return compewireprotosupport('none', 0, 10)
3539 return compewireprotosupport('none', 0, 10)
3540
3540
3541 # We don't implement revlogheader because it is handled specially
3541 # We don't implement revlogheader because it is handled specially
3542 # in the revlog class.
3542 # in the revlog class.
3543
3543
3544 def compressstream(self, it, opts=None):
3544 def compressstream(self, it, opts=None):
3545 return it
3545 return it
3546
3546
3547 def decompressorreader(self, fh):
3547 def decompressorreader(self, fh):
3548 return fh
3548 return fh
3549
3549
3550 class nooprevlogcompressor(object):
3550 class nooprevlogcompressor(object):
3551 def compress(self, data):
3551 def compress(self, data):
3552 return None
3552 return None
3553
3553
3554 def revlogcompressor(self, opts=None):
3554 def revlogcompressor(self, opts=None):
3555 return self.nooprevlogcompressor()
3555 return self.nooprevlogcompressor()
3556
3556
3557 compengines.register(_noopengine())
3557 compengines.register(_noopengine())
3558
3558
3559 class _zstdengine(compressionengine):
3559 class _zstdengine(compressionengine):
3560 def name(self):
3560 def name(self):
3561 return 'zstd'
3561 return 'zstd'
3562
3562
3563 @propertycache
3563 @propertycache
3564 def _module(self):
3564 def _module(self):
3565 # Not all installs have the zstd module available. So defer importing
3565 # Not all installs have the zstd module available. So defer importing
3566 # until first access.
3566 # until first access.
3567 try:
3567 try:
3568 from . import zstd
3568 from . import zstd
3569 # Force delayed import.
3569 # Force delayed import.
3570 zstd.__version__
3570 zstd.__version__
3571 return zstd
3571 return zstd
3572 except ImportError:
3572 except ImportError:
3573 return None
3573 return None
3574
3574
3575 def available(self):
3575 def available(self):
3576 return bool(self._module)
3576 return bool(self._module)
3577
3577
3578 def bundletype(self):
3578 def bundletype(self):
3579 """A modern compression algorithm that is fast and highly flexible.
3579 """A modern compression algorithm that is fast and highly flexible.
3580
3580
3581 Only supported by Mercurial 4.1 and newer clients.
3581 Only supported by Mercurial 4.1 and newer clients.
3582
3582
3583 With the default settings, zstd compression is both faster and yields
3583 With the default settings, zstd compression is both faster and yields
3584 better compression than ``gzip``. It also frequently yields better
3584 better compression than ``gzip``. It also frequently yields better
3585 compression than ``bzip2`` while operating at much higher speeds.
3585 compression than ``bzip2`` while operating at much higher speeds.
3586
3586
3587 If this engine is available and backwards compatibility is not a
3587 If this engine is available and backwards compatibility is not a
3588 concern, it is likely the best available engine.
3588 concern, it is likely the best available engine.
3589 """
3589 """
3590 return 'zstd', 'ZS'
3590 return 'zstd', 'ZS'
3591
3591
3592 def wireprotosupport(self):
3592 def wireprotosupport(self):
3593 return compewireprotosupport('zstd', 50, 50)
3593 return compewireprotosupport('zstd', 50, 50)
3594
3594
3595 def revlogheader(self):
3595 def revlogheader(self):
3596 return '\x28'
3596 return '\x28'
3597
3597
3598 def compressstream(self, it, opts=None):
3598 def compressstream(self, it, opts=None):
3599 opts = opts or {}
3599 opts = opts or {}
3600 # zstd level 3 is almost always significantly faster than zlib
3600 # zstd level 3 is almost always significantly faster than zlib
3601 # while providing no worse compression. It strikes a good balance
3601 # while providing no worse compression. It strikes a good balance
3602 # between speed and compression.
3602 # between speed and compression.
3603 level = opts.get('level', 3)
3603 level = opts.get('level', 3)
3604
3604
3605 zstd = self._module
3605 zstd = self._module
3606 z = zstd.ZstdCompressor(level=level).compressobj()
3606 z = zstd.ZstdCompressor(level=level).compressobj()
3607 for chunk in it:
3607 for chunk in it:
3608 data = z.compress(chunk)
3608 data = z.compress(chunk)
3609 if data:
3609 if data:
3610 yield data
3610 yield data
3611
3611
3612 yield z.flush()
3612 yield z.flush()
3613
3613
3614 def decompressorreader(self, fh):
3614 def decompressorreader(self, fh):
3615 zstd = self._module
3615 zstd = self._module
3616 dctx = zstd.ZstdDecompressor()
3616 dctx = zstd.ZstdDecompressor()
3617 return chunkbuffer(dctx.read_from(fh))
3617 return chunkbuffer(dctx.read_from(fh))
3618
3618
3619 class zstdrevlogcompressor(object):
3619 class zstdrevlogcompressor(object):
3620 def __init__(self, zstd, level=3):
3620 def __init__(self, zstd, level=3):
3621 # Writing the content size adds a few bytes to the output. However,
3621 # Writing the content size adds a few bytes to the output. However,
3622 # it allows decompression to be more optimal since we can
3622 # it allows decompression to be more optimal since we can
3623 # pre-allocate a buffer to hold the result.
3623 # pre-allocate a buffer to hold the result.
3624 self._cctx = zstd.ZstdCompressor(level=level,
3624 self._cctx = zstd.ZstdCompressor(level=level,
3625 write_content_size=True)
3625 write_content_size=True)
3626 self._dctx = zstd.ZstdDecompressor()
3626 self._dctx = zstd.ZstdDecompressor()
3627 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3627 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3628 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3628 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3629
3629
3630 def compress(self, data):
3630 def compress(self, data):
3631 insize = len(data)
3631 insize = len(data)
3632 # Caller handles empty input case.
3632 # Caller handles empty input case.
3633 assert insize > 0
3633 assert insize > 0
3634
3634
3635 if insize < 50:
3635 if insize < 50:
3636 return None
3636 return None
3637
3637
3638 elif insize <= 1000000:
3638 elif insize <= 1000000:
3639 compressed = self._cctx.compress(data)
3639 compressed = self._cctx.compress(data)
3640 if len(compressed) < insize:
3640 if len(compressed) < insize:
3641 return compressed
3641 return compressed
3642 return None
3642 return None
3643 else:
3643 else:
3644 z = self._cctx.compressobj()
3644 z = self._cctx.compressobj()
3645 chunks = []
3645 chunks = []
3646 pos = 0
3646 pos = 0
3647 while pos < insize:
3647 while pos < insize:
3648 pos2 = pos + self._compinsize
3648 pos2 = pos + self._compinsize
3649 chunk = z.compress(data[pos:pos2])
3649 chunk = z.compress(data[pos:pos2])
3650 if chunk:
3650 if chunk:
3651 chunks.append(chunk)
3651 chunks.append(chunk)
3652 pos = pos2
3652 pos = pos2
3653 chunks.append(z.flush())
3653 chunks.append(z.flush())
3654
3654
3655 if sum(map(len, chunks)) < insize:
3655 if sum(map(len, chunks)) < insize:
3656 return ''.join(chunks)
3656 return ''.join(chunks)
3657 return None
3657 return None
3658
3658
3659 def decompress(self, data):
3659 def decompress(self, data):
3660 insize = len(data)
3660 insize = len(data)
3661
3661
3662 try:
3662 try:
3663 # This was measured to be faster than other streaming
3663 # This was measured to be faster than other streaming
3664 # decompressors.
3664 # decompressors.
3665 dobj = self._dctx.decompressobj()
3665 dobj = self._dctx.decompressobj()
3666 chunks = []
3666 chunks = []
3667 pos = 0
3667 pos = 0
3668 while pos < insize:
3668 while pos < insize:
3669 pos2 = pos + self._decompinsize
3669 pos2 = pos + self._decompinsize
3670 chunk = dobj.decompress(data[pos:pos2])
3670 chunk = dobj.decompress(data[pos:pos2])
3671 if chunk:
3671 if chunk:
3672 chunks.append(chunk)
3672 chunks.append(chunk)
3673 pos = pos2
3673 pos = pos2
3674 # Frame should be exhausted, so no finish() API.
3674 # Frame should be exhausted, so no finish() API.
3675
3675
3676 return ''.join(chunks)
3676 return ''.join(chunks)
3677 except Exception as e:
3677 except Exception as e:
3678 raise error.RevlogError(_('revlog decompress error: %s') %
3678 raise error.RevlogError(_('revlog decompress error: %s') %
3679 str(e))
3679 str(e))
3680
3680
3681 def revlogcompressor(self, opts=None):
3681 def revlogcompressor(self, opts=None):
3682 opts = opts or {}
3682 opts = opts or {}
3683 return self.zstdrevlogcompressor(self._module,
3683 return self.zstdrevlogcompressor(self._module,
3684 level=opts.get('level', 3))
3684 level=opts.get('level', 3))
3685
3685
3686 compengines.register(_zstdengine())
3686 compengines.register(_zstdengine())
3687
3687
3688 def bundlecompressiontopics():
3688 def bundlecompressiontopics():
3689 """Obtains a list of available bundle compressions for use in help."""
3689 """Obtains a list of available bundle compressions for use in help."""
3690 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3690 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3691 items = {}
3691 items = {}
3692
3692
3693 # We need to format the docstring. So use a dummy object/type to hold it
3693 # We need to format the docstring. So use a dummy object/type to hold it
3694 # rather than mutating the original.
3694 # rather than mutating the original.
3695 class docobject(object):
3695 class docobject(object):
3696 pass
3696 pass
3697
3697
3698 for name in compengines:
3698 for name in compengines:
3699 engine = compengines[name]
3699 engine = compengines[name]
3700
3700
3701 if not engine.available():
3701 if not engine.available():
3702 continue
3702 continue
3703
3703
3704 bt = engine.bundletype()
3704 bt = engine.bundletype()
3705 if not bt or not bt[0]:
3705 if not bt or not bt[0]:
3706 continue
3706 continue
3707
3707
3708 doc = pycompat.sysstr('``%s``\n %s' % (
3708 doc = pycompat.sysstr('``%s``\n %s') % (
3709 bt[0], engine.bundletype.__doc__))
3709 bt[0], engine.bundletype.__doc__)
3710
3710
3711 value = docobject()
3711 value = docobject()
3712 value.__doc__ = doc
3712 value.__doc__ = doc
3713
3713
3714 items[bt[0]] = value
3714 items[bt[0]] = value
3715
3715
3716 return items
3716 return items
3717
3717
3718 # convenient shortcut
3718 # convenient shortcut
3719 dst = debugstacktrace
3719 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now