##// END OF EJS Templates
util: enable hardlink for some BSD-family filesystems...
Jun Wu -
r31598:32f09ead default
parent child Browse files
Show More
@@ -1,3606 +1,3609 b''
1 # util.py - Mercurial utility functions and platform specific implementations
1 # util.py - Mercurial utility functions and platform specific implementations
2 #
2 #
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 #
6 #
7 # This software may be used and distributed according to the terms of the
7 # This software may be used and distributed according to the terms of the
8 # GNU General Public License version 2 or any later version.
8 # GNU General Public License version 2 or any later version.
9
9
10 """Mercurial utility functions and platform specific implementations.
10 """Mercurial utility functions and platform specific implementations.
11
11
12 This contains helper routines that are independent of the SCM core and
12 This contains helper routines that are independent of the SCM core and
13 hide platform-specific details from the core.
13 hide platform-specific details from the core.
14 """
14 """
15
15
16 from __future__ import absolute_import
16 from __future__ import absolute_import
17
17
18 import bz2
18 import bz2
19 import calendar
19 import calendar
20 import codecs
20 import codecs
21 import collections
21 import collections
22 import datetime
22 import datetime
23 import errno
23 import errno
24 import gc
24 import gc
25 import hashlib
25 import hashlib
26 import imp
26 import imp
27 import os
27 import os
28 import platform as pyplatform
28 import platform as pyplatform
29 import re as remod
29 import re as remod
30 import shutil
30 import shutil
31 import signal
31 import signal
32 import socket
32 import socket
33 import stat
33 import stat
34 import string
34 import string
35 import subprocess
35 import subprocess
36 import sys
36 import sys
37 import tempfile
37 import tempfile
38 import textwrap
38 import textwrap
39 import time
39 import time
40 import traceback
40 import traceback
41 import zlib
41 import zlib
42
42
43 from . import (
43 from . import (
44 encoding,
44 encoding,
45 error,
45 error,
46 i18n,
46 i18n,
47 osutil,
47 osutil,
48 parsers,
48 parsers,
49 pycompat,
49 pycompat,
50 )
50 )
51
51
52 empty = pycompat.empty
52 empty = pycompat.empty
53 httplib = pycompat.httplib
53 httplib = pycompat.httplib
54 httpserver = pycompat.httpserver
54 httpserver = pycompat.httpserver
55 pickle = pycompat.pickle
55 pickle = pycompat.pickle
56 queue = pycompat.queue
56 queue = pycompat.queue
57 socketserver = pycompat.socketserver
57 socketserver = pycompat.socketserver
58 stderr = pycompat.stderr
58 stderr = pycompat.stderr
59 stdin = pycompat.stdin
59 stdin = pycompat.stdin
60 stdout = pycompat.stdout
60 stdout = pycompat.stdout
61 stringio = pycompat.stringio
61 stringio = pycompat.stringio
62 urlerr = pycompat.urlerr
62 urlerr = pycompat.urlerr
63 urlreq = pycompat.urlreq
63 urlreq = pycompat.urlreq
64 xmlrpclib = pycompat.xmlrpclib
64 xmlrpclib = pycompat.xmlrpclib
65
65
66 def isatty(fp):
66 def isatty(fp):
67 try:
67 try:
68 return fp.isatty()
68 return fp.isatty()
69 except AttributeError:
69 except AttributeError:
70 return False
70 return False
71
71
72 # glibc determines buffering on first write to stdout - if we replace a TTY
72 # glibc determines buffering on first write to stdout - if we replace a TTY
73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
74 # buffering
74 # buffering
75 if isatty(stdout):
75 if isatty(stdout):
76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
77
77
78 if pycompat.osname == 'nt':
78 if pycompat.osname == 'nt':
79 from . import windows as platform
79 from . import windows as platform
80 stdout = platform.winstdout(stdout)
80 stdout = platform.winstdout(stdout)
81 else:
81 else:
82 from . import posix as platform
82 from . import posix as platform
83
83
84 _ = i18n._
84 _ = i18n._
85
85
86 bindunixsocket = platform.bindunixsocket
86 bindunixsocket = platform.bindunixsocket
87 cachestat = platform.cachestat
87 cachestat = platform.cachestat
88 checkexec = platform.checkexec
88 checkexec = platform.checkexec
89 checklink = platform.checklink
89 checklink = platform.checklink
90 copymode = platform.copymode
90 copymode = platform.copymode
91 executablepath = platform.executablepath
91 executablepath = platform.executablepath
92 expandglobs = platform.expandglobs
92 expandglobs = platform.expandglobs
93 explainexit = platform.explainexit
93 explainexit = platform.explainexit
94 findexe = platform.findexe
94 findexe = platform.findexe
95 gethgcmd = platform.gethgcmd
95 gethgcmd = platform.gethgcmd
96 getuser = platform.getuser
96 getuser = platform.getuser
97 getpid = os.getpid
97 getpid = os.getpid
98 groupmembers = platform.groupmembers
98 groupmembers = platform.groupmembers
99 groupname = platform.groupname
99 groupname = platform.groupname
100 hidewindow = platform.hidewindow
100 hidewindow = platform.hidewindow
101 isexec = platform.isexec
101 isexec = platform.isexec
102 isowner = platform.isowner
102 isowner = platform.isowner
103 localpath = platform.localpath
103 localpath = platform.localpath
104 lookupreg = platform.lookupreg
104 lookupreg = platform.lookupreg
105 makedir = platform.makedir
105 makedir = platform.makedir
106 nlinks = platform.nlinks
106 nlinks = platform.nlinks
107 normpath = platform.normpath
107 normpath = platform.normpath
108 normcase = platform.normcase
108 normcase = platform.normcase
109 normcasespec = platform.normcasespec
109 normcasespec = platform.normcasespec
110 normcasefallback = platform.normcasefallback
110 normcasefallback = platform.normcasefallback
111 openhardlinks = platform.openhardlinks
111 openhardlinks = platform.openhardlinks
112 oslink = platform.oslink
112 oslink = platform.oslink
113 parsepatchoutput = platform.parsepatchoutput
113 parsepatchoutput = platform.parsepatchoutput
114 pconvert = platform.pconvert
114 pconvert = platform.pconvert
115 poll = platform.poll
115 poll = platform.poll
116 popen = platform.popen
116 popen = platform.popen
117 posixfile = platform.posixfile
117 posixfile = platform.posixfile
118 quotecommand = platform.quotecommand
118 quotecommand = platform.quotecommand
119 readpipe = platform.readpipe
119 readpipe = platform.readpipe
120 rename = platform.rename
120 rename = platform.rename
121 removedirs = platform.removedirs
121 removedirs = platform.removedirs
122 samedevice = platform.samedevice
122 samedevice = platform.samedevice
123 samefile = platform.samefile
123 samefile = platform.samefile
124 samestat = platform.samestat
124 samestat = platform.samestat
125 setbinary = platform.setbinary
125 setbinary = platform.setbinary
126 setflags = platform.setflags
126 setflags = platform.setflags
127 setsignalhandler = platform.setsignalhandler
127 setsignalhandler = platform.setsignalhandler
128 shellquote = platform.shellquote
128 shellquote = platform.shellquote
129 spawndetached = platform.spawndetached
129 spawndetached = platform.spawndetached
130 split = platform.split
130 split = platform.split
131 sshargs = platform.sshargs
131 sshargs = platform.sshargs
132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
133 statisexec = platform.statisexec
133 statisexec = platform.statisexec
134 statislink = platform.statislink
134 statislink = platform.statislink
135 testpid = platform.testpid
135 testpid = platform.testpid
136 umask = platform.umask
136 umask = platform.umask
137 unlink = platform.unlink
137 unlink = platform.unlink
138 username = platform.username
138 username = platform.username
139
139
140 # Python compatibility
140 # Python compatibility
141
141
142 _notset = object()
142 _notset = object()
143
143
144 # disable Python's problematic floating point timestamps (issue4836)
144 # disable Python's problematic floating point timestamps (issue4836)
145 # (Python hypocritically says you shouldn't change this behavior in
145 # (Python hypocritically says you shouldn't change this behavior in
146 # libraries, and sure enough Mercurial is not a library.)
146 # libraries, and sure enough Mercurial is not a library.)
147 os.stat_float_times(False)
147 os.stat_float_times(False)
148
148
149 def safehasattr(thing, attr):
149 def safehasattr(thing, attr):
150 return getattr(thing, attr, _notset) is not _notset
150 return getattr(thing, attr, _notset) is not _notset
151
151
152 def bitsfrom(container):
152 def bitsfrom(container):
153 bits = 0
153 bits = 0
154 for bit in container:
154 for bit in container:
155 bits |= bit
155 bits |= bit
156 return bits
156 return bits
157
157
158 DIGESTS = {
158 DIGESTS = {
159 'md5': hashlib.md5,
159 'md5': hashlib.md5,
160 'sha1': hashlib.sha1,
160 'sha1': hashlib.sha1,
161 'sha512': hashlib.sha512,
161 'sha512': hashlib.sha512,
162 }
162 }
163 # List of digest types from strongest to weakest
163 # List of digest types from strongest to weakest
164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
165
165
166 for k in DIGESTS_BY_STRENGTH:
166 for k in DIGESTS_BY_STRENGTH:
167 assert k in DIGESTS
167 assert k in DIGESTS
168
168
169 class digester(object):
169 class digester(object):
170 """helper to compute digests.
170 """helper to compute digests.
171
171
172 This helper can be used to compute one or more digests given their name.
172 This helper can be used to compute one or more digests given their name.
173
173
174 >>> d = digester(['md5', 'sha1'])
174 >>> d = digester(['md5', 'sha1'])
175 >>> d.update('foo')
175 >>> d.update('foo')
176 >>> [k for k in sorted(d)]
176 >>> [k for k in sorted(d)]
177 ['md5', 'sha1']
177 ['md5', 'sha1']
178 >>> d['md5']
178 >>> d['md5']
179 'acbd18db4cc2f85cedef654fccc4a4d8'
179 'acbd18db4cc2f85cedef654fccc4a4d8'
180 >>> d['sha1']
180 >>> d['sha1']
181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
182 >>> digester.preferred(['md5', 'sha1'])
182 >>> digester.preferred(['md5', 'sha1'])
183 'sha1'
183 'sha1'
184 """
184 """
185
185
186 def __init__(self, digests, s=''):
186 def __init__(self, digests, s=''):
187 self._hashes = {}
187 self._hashes = {}
188 for k in digests:
188 for k in digests:
189 if k not in DIGESTS:
189 if k not in DIGESTS:
190 raise Abort(_('unknown digest type: %s') % k)
190 raise Abort(_('unknown digest type: %s') % k)
191 self._hashes[k] = DIGESTS[k]()
191 self._hashes[k] = DIGESTS[k]()
192 if s:
192 if s:
193 self.update(s)
193 self.update(s)
194
194
195 def update(self, data):
195 def update(self, data):
196 for h in self._hashes.values():
196 for h in self._hashes.values():
197 h.update(data)
197 h.update(data)
198
198
199 def __getitem__(self, key):
199 def __getitem__(self, key):
200 if key not in DIGESTS:
200 if key not in DIGESTS:
201 raise Abort(_('unknown digest type: %s') % k)
201 raise Abort(_('unknown digest type: %s') % k)
202 return self._hashes[key].hexdigest()
202 return self._hashes[key].hexdigest()
203
203
204 def __iter__(self):
204 def __iter__(self):
205 return iter(self._hashes)
205 return iter(self._hashes)
206
206
207 @staticmethod
207 @staticmethod
208 def preferred(supported):
208 def preferred(supported):
209 """returns the strongest digest type in both supported and DIGESTS."""
209 """returns the strongest digest type in both supported and DIGESTS."""
210
210
211 for k in DIGESTS_BY_STRENGTH:
211 for k in DIGESTS_BY_STRENGTH:
212 if k in supported:
212 if k in supported:
213 return k
213 return k
214 return None
214 return None
215
215
216 class digestchecker(object):
216 class digestchecker(object):
217 """file handle wrapper that additionally checks content against a given
217 """file handle wrapper that additionally checks content against a given
218 size and digests.
218 size and digests.
219
219
220 d = digestchecker(fh, size, {'md5': '...'})
220 d = digestchecker(fh, size, {'md5': '...'})
221
221
222 When multiple digests are given, all of them are validated.
222 When multiple digests are given, all of them are validated.
223 """
223 """
224
224
225 def __init__(self, fh, size, digests):
225 def __init__(self, fh, size, digests):
226 self._fh = fh
226 self._fh = fh
227 self._size = size
227 self._size = size
228 self._got = 0
228 self._got = 0
229 self._digests = dict(digests)
229 self._digests = dict(digests)
230 self._digester = digester(self._digests.keys())
230 self._digester = digester(self._digests.keys())
231
231
232 def read(self, length=-1):
232 def read(self, length=-1):
233 content = self._fh.read(length)
233 content = self._fh.read(length)
234 self._digester.update(content)
234 self._digester.update(content)
235 self._got += len(content)
235 self._got += len(content)
236 return content
236 return content
237
237
238 def validate(self):
238 def validate(self):
239 if self._size != self._got:
239 if self._size != self._got:
240 raise Abort(_('size mismatch: expected %d, got %d') %
240 raise Abort(_('size mismatch: expected %d, got %d') %
241 (self._size, self._got))
241 (self._size, self._got))
242 for k, v in self._digests.items():
242 for k, v in self._digests.items():
243 if v != self._digester[k]:
243 if v != self._digester[k]:
244 # i18n: first parameter is a digest name
244 # i18n: first parameter is a digest name
245 raise Abort(_('%s mismatch: expected %s, got %s') %
245 raise Abort(_('%s mismatch: expected %s, got %s') %
246 (k, v, self._digester[k]))
246 (k, v, self._digester[k]))
247
247
248 try:
248 try:
249 buffer = buffer
249 buffer = buffer
250 except NameError:
250 except NameError:
251 if not pycompat.ispy3:
251 if not pycompat.ispy3:
252 def buffer(sliceable, offset=0, length=None):
252 def buffer(sliceable, offset=0, length=None):
253 if length is not None:
253 if length is not None:
254 return sliceable[offset:offset + length]
254 return sliceable[offset:offset + length]
255 return sliceable[offset:]
255 return sliceable[offset:]
256 else:
256 else:
257 def buffer(sliceable, offset=0, length=None):
257 def buffer(sliceable, offset=0, length=None):
258 if length is not None:
258 if length is not None:
259 return memoryview(sliceable)[offset:offset + length]
259 return memoryview(sliceable)[offset:offset + length]
260 return memoryview(sliceable)[offset:]
260 return memoryview(sliceable)[offset:]
261
261
262 closefds = pycompat.osname == 'posix'
262 closefds = pycompat.osname == 'posix'
263
263
264 _chunksize = 4096
264 _chunksize = 4096
265
265
266 class bufferedinputpipe(object):
266 class bufferedinputpipe(object):
267 """a manually buffered input pipe
267 """a manually buffered input pipe
268
268
269 Python will not let us use buffered IO and lazy reading with 'polling' at
269 Python will not let us use buffered IO and lazy reading with 'polling' at
270 the same time. We cannot probe the buffer state and select will not detect
270 the same time. We cannot probe the buffer state and select will not detect
271 that data are ready to read if they are already buffered.
271 that data are ready to read if they are already buffered.
272
272
273 This class let us work around that by implementing its own buffering
273 This class let us work around that by implementing its own buffering
274 (allowing efficient readline) while offering a way to know if the buffer is
274 (allowing efficient readline) while offering a way to know if the buffer is
275 empty from the output (allowing collaboration of the buffer with polling).
275 empty from the output (allowing collaboration of the buffer with polling).
276
276
277 This class lives in the 'util' module because it makes use of the 'os'
277 This class lives in the 'util' module because it makes use of the 'os'
278 module from the python stdlib.
278 module from the python stdlib.
279 """
279 """
280
280
281 def __init__(self, input):
281 def __init__(self, input):
282 self._input = input
282 self._input = input
283 self._buffer = []
283 self._buffer = []
284 self._eof = False
284 self._eof = False
285 self._lenbuf = 0
285 self._lenbuf = 0
286
286
287 @property
287 @property
288 def hasbuffer(self):
288 def hasbuffer(self):
289 """True is any data is currently buffered
289 """True is any data is currently buffered
290
290
291 This will be used externally a pre-step for polling IO. If there is
291 This will be used externally a pre-step for polling IO. If there is
292 already data then no polling should be set in place."""
292 already data then no polling should be set in place."""
293 return bool(self._buffer)
293 return bool(self._buffer)
294
294
295 @property
295 @property
296 def closed(self):
296 def closed(self):
297 return self._input.closed
297 return self._input.closed
298
298
299 def fileno(self):
299 def fileno(self):
300 return self._input.fileno()
300 return self._input.fileno()
301
301
302 def close(self):
302 def close(self):
303 return self._input.close()
303 return self._input.close()
304
304
305 def read(self, size):
305 def read(self, size):
306 while (not self._eof) and (self._lenbuf < size):
306 while (not self._eof) and (self._lenbuf < size):
307 self._fillbuffer()
307 self._fillbuffer()
308 return self._frombuffer(size)
308 return self._frombuffer(size)
309
309
310 def readline(self, *args, **kwargs):
310 def readline(self, *args, **kwargs):
311 if 1 < len(self._buffer):
311 if 1 < len(self._buffer):
312 # this should not happen because both read and readline end with a
312 # this should not happen because both read and readline end with a
313 # _frombuffer call that collapse it.
313 # _frombuffer call that collapse it.
314 self._buffer = [''.join(self._buffer)]
314 self._buffer = [''.join(self._buffer)]
315 self._lenbuf = len(self._buffer[0])
315 self._lenbuf = len(self._buffer[0])
316 lfi = -1
316 lfi = -1
317 if self._buffer:
317 if self._buffer:
318 lfi = self._buffer[-1].find('\n')
318 lfi = self._buffer[-1].find('\n')
319 while (not self._eof) and lfi < 0:
319 while (not self._eof) and lfi < 0:
320 self._fillbuffer()
320 self._fillbuffer()
321 if self._buffer:
321 if self._buffer:
322 lfi = self._buffer[-1].find('\n')
322 lfi = self._buffer[-1].find('\n')
323 size = lfi + 1
323 size = lfi + 1
324 if lfi < 0: # end of file
324 if lfi < 0: # end of file
325 size = self._lenbuf
325 size = self._lenbuf
326 elif 1 < len(self._buffer):
326 elif 1 < len(self._buffer):
327 # we need to take previous chunks into account
327 # we need to take previous chunks into account
328 size += self._lenbuf - len(self._buffer[-1])
328 size += self._lenbuf - len(self._buffer[-1])
329 return self._frombuffer(size)
329 return self._frombuffer(size)
330
330
331 def _frombuffer(self, size):
331 def _frombuffer(self, size):
332 """return at most 'size' data from the buffer
332 """return at most 'size' data from the buffer
333
333
334 The data are removed from the buffer."""
334 The data are removed from the buffer."""
335 if size == 0 or not self._buffer:
335 if size == 0 or not self._buffer:
336 return ''
336 return ''
337 buf = self._buffer[0]
337 buf = self._buffer[0]
338 if 1 < len(self._buffer):
338 if 1 < len(self._buffer):
339 buf = ''.join(self._buffer)
339 buf = ''.join(self._buffer)
340
340
341 data = buf[:size]
341 data = buf[:size]
342 buf = buf[len(data):]
342 buf = buf[len(data):]
343 if buf:
343 if buf:
344 self._buffer = [buf]
344 self._buffer = [buf]
345 self._lenbuf = len(buf)
345 self._lenbuf = len(buf)
346 else:
346 else:
347 self._buffer = []
347 self._buffer = []
348 self._lenbuf = 0
348 self._lenbuf = 0
349 return data
349 return data
350
350
351 def _fillbuffer(self):
351 def _fillbuffer(self):
352 """read data to the buffer"""
352 """read data to the buffer"""
353 data = os.read(self._input.fileno(), _chunksize)
353 data = os.read(self._input.fileno(), _chunksize)
354 if not data:
354 if not data:
355 self._eof = True
355 self._eof = True
356 else:
356 else:
357 self._lenbuf += len(data)
357 self._lenbuf += len(data)
358 self._buffer.append(data)
358 self._buffer.append(data)
359
359
360 def popen2(cmd, env=None, newlines=False):
360 def popen2(cmd, env=None, newlines=False):
361 # Setting bufsize to -1 lets the system decide the buffer size.
361 # Setting bufsize to -1 lets the system decide the buffer size.
362 # The default for bufsize is 0, meaning unbuffered. This leads to
362 # The default for bufsize is 0, meaning unbuffered. This leads to
363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
365 close_fds=closefds,
365 close_fds=closefds,
366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
367 universal_newlines=newlines,
367 universal_newlines=newlines,
368 env=env)
368 env=env)
369 return p.stdin, p.stdout
369 return p.stdin, p.stdout
370
370
371 def popen3(cmd, env=None, newlines=False):
371 def popen3(cmd, env=None, newlines=False):
372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
373 return stdin, stdout, stderr
373 return stdin, stdout, stderr
374
374
375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
377 close_fds=closefds,
377 close_fds=closefds,
378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
379 stderr=subprocess.PIPE,
379 stderr=subprocess.PIPE,
380 universal_newlines=newlines,
380 universal_newlines=newlines,
381 env=env)
381 env=env)
382 return p.stdin, p.stdout, p.stderr, p
382 return p.stdin, p.stdout, p.stderr, p
383
383
384 def version():
384 def version():
385 """Return version information if available."""
385 """Return version information if available."""
386 try:
386 try:
387 from . import __version__
387 from . import __version__
388 return __version__.version
388 return __version__.version
389 except ImportError:
389 except ImportError:
390 return 'unknown'
390 return 'unknown'
391
391
392 def versiontuple(v=None, n=4):
392 def versiontuple(v=None, n=4):
393 """Parses a Mercurial version string into an N-tuple.
393 """Parses a Mercurial version string into an N-tuple.
394
394
395 The version string to be parsed is specified with the ``v`` argument.
395 The version string to be parsed is specified with the ``v`` argument.
396 If it isn't defined, the current Mercurial version string will be parsed.
396 If it isn't defined, the current Mercurial version string will be parsed.
397
397
398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
399 returned values:
399 returned values:
400
400
401 >>> v = '3.6.1+190-df9b73d2d444'
401 >>> v = '3.6.1+190-df9b73d2d444'
402 >>> versiontuple(v, 2)
402 >>> versiontuple(v, 2)
403 (3, 6)
403 (3, 6)
404 >>> versiontuple(v, 3)
404 >>> versiontuple(v, 3)
405 (3, 6, 1)
405 (3, 6, 1)
406 >>> versiontuple(v, 4)
406 >>> versiontuple(v, 4)
407 (3, 6, 1, '190-df9b73d2d444')
407 (3, 6, 1, '190-df9b73d2d444')
408
408
409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
410 (3, 6, 1, '190-df9b73d2d444+20151118')
410 (3, 6, 1, '190-df9b73d2d444+20151118')
411
411
412 >>> v = '3.6'
412 >>> v = '3.6'
413 >>> versiontuple(v, 2)
413 >>> versiontuple(v, 2)
414 (3, 6)
414 (3, 6)
415 >>> versiontuple(v, 3)
415 >>> versiontuple(v, 3)
416 (3, 6, None)
416 (3, 6, None)
417 >>> versiontuple(v, 4)
417 >>> versiontuple(v, 4)
418 (3, 6, None, None)
418 (3, 6, None, None)
419
419
420 >>> v = '3.9-rc'
420 >>> v = '3.9-rc'
421 >>> versiontuple(v, 2)
421 >>> versiontuple(v, 2)
422 (3, 9)
422 (3, 9)
423 >>> versiontuple(v, 3)
423 >>> versiontuple(v, 3)
424 (3, 9, None)
424 (3, 9, None)
425 >>> versiontuple(v, 4)
425 >>> versiontuple(v, 4)
426 (3, 9, None, 'rc')
426 (3, 9, None, 'rc')
427
427
428 >>> v = '3.9-rc+2-02a8fea4289b'
428 >>> v = '3.9-rc+2-02a8fea4289b'
429 >>> versiontuple(v, 2)
429 >>> versiontuple(v, 2)
430 (3, 9)
430 (3, 9)
431 >>> versiontuple(v, 3)
431 >>> versiontuple(v, 3)
432 (3, 9, None)
432 (3, 9, None)
433 >>> versiontuple(v, 4)
433 >>> versiontuple(v, 4)
434 (3, 9, None, 'rc+2-02a8fea4289b')
434 (3, 9, None, 'rc+2-02a8fea4289b')
435 """
435 """
436 if not v:
436 if not v:
437 v = version()
437 v = version()
438 parts = remod.split('[\+-]', v, 1)
438 parts = remod.split('[\+-]', v, 1)
439 if len(parts) == 1:
439 if len(parts) == 1:
440 vparts, extra = parts[0], None
440 vparts, extra = parts[0], None
441 else:
441 else:
442 vparts, extra = parts
442 vparts, extra = parts
443
443
444 vints = []
444 vints = []
445 for i in vparts.split('.'):
445 for i in vparts.split('.'):
446 try:
446 try:
447 vints.append(int(i))
447 vints.append(int(i))
448 except ValueError:
448 except ValueError:
449 break
449 break
450 # (3, 6) -> (3, 6, None)
450 # (3, 6) -> (3, 6, None)
451 while len(vints) < 3:
451 while len(vints) < 3:
452 vints.append(None)
452 vints.append(None)
453
453
454 if n == 2:
454 if n == 2:
455 return (vints[0], vints[1])
455 return (vints[0], vints[1])
456 if n == 3:
456 if n == 3:
457 return (vints[0], vints[1], vints[2])
457 return (vints[0], vints[1], vints[2])
458 if n == 4:
458 if n == 4:
459 return (vints[0], vints[1], vints[2], extra)
459 return (vints[0], vints[1], vints[2], extra)
460
460
461 # used by parsedate
461 # used by parsedate
462 defaultdateformats = (
462 defaultdateformats = (
463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
464 '%Y-%m-%dT%H:%M', # without seconds
464 '%Y-%m-%dT%H:%M', # without seconds
465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
466 '%Y-%m-%dT%H%M', # without seconds
466 '%Y-%m-%dT%H%M', # without seconds
467 '%Y-%m-%d %H:%M:%S', # our common legal variant
467 '%Y-%m-%d %H:%M:%S', # our common legal variant
468 '%Y-%m-%d %H:%M', # without seconds
468 '%Y-%m-%d %H:%M', # without seconds
469 '%Y-%m-%d %H%M%S', # without :
469 '%Y-%m-%d %H%M%S', # without :
470 '%Y-%m-%d %H%M', # without seconds
470 '%Y-%m-%d %H%M', # without seconds
471 '%Y-%m-%d %I:%M:%S%p',
471 '%Y-%m-%d %I:%M:%S%p',
472 '%Y-%m-%d %H:%M',
472 '%Y-%m-%d %H:%M',
473 '%Y-%m-%d %I:%M%p',
473 '%Y-%m-%d %I:%M%p',
474 '%Y-%m-%d',
474 '%Y-%m-%d',
475 '%m-%d',
475 '%m-%d',
476 '%m/%d',
476 '%m/%d',
477 '%m/%d/%y',
477 '%m/%d/%y',
478 '%m/%d/%Y',
478 '%m/%d/%Y',
479 '%a %b %d %H:%M:%S %Y',
479 '%a %b %d %H:%M:%S %Y',
480 '%a %b %d %I:%M:%S%p %Y',
480 '%a %b %d %I:%M:%S%p %Y',
481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
482 '%b %d %H:%M:%S %Y',
482 '%b %d %H:%M:%S %Y',
483 '%b %d %I:%M:%S%p %Y',
483 '%b %d %I:%M:%S%p %Y',
484 '%b %d %H:%M:%S',
484 '%b %d %H:%M:%S',
485 '%b %d %I:%M:%S%p',
485 '%b %d %I:%M:%S%p',
486 '%b %d %H:%M',
486 '%b %d %H:%M',
487 '%b %d %I:%M%p',
487 '%b %d %I:%M%p',
488 '%b %d %Y',
488 '%b %d %Y',
489 '%b %d',
489 '%b %d',
490 '%H:%M:%S',
490 '%H:%M:%S',
491 '%I:%M:%S%p',
491 '%I:%M:%S%p',
492 '%H:%M',
492 '%H:%M',
493 '%I:%M%p',
493 '%I:%M%p',
494 )
494 )
495
495
496 extendeddateformats = defaultdateformats + (
496 extendeddateformats = defaultdateformats + (
497 "%Y",
497 "%Y",
498 "%Y-%m",
498 "%Y-%m",
499 "%b",
499 "%b",
500 "%b %Y",
500 "%b %Y",
501 )
501 )
502
502
503 def cachefunc(func):
503 def cachefunc(func):
504 '''cache the result of function calls'''
504 '''cache the result of function calls'''
505 # XXX doesn't handle keywords args
505 # XXX doesn't handle keywords args
506 if func.__code__.co_argcount == 0:
506 if func.__code__.co_argcount == 0:
507 cache = []
507 cache = []
508 def f():
508 def f():
509 if len(cache) == 0:
509 if len(cache) == 0:
510 cache.append(func())
510 cache.append(func())
511 return cache[0]
511 return cache[0]
512 return f
512 return f
513 cache = {}
513 cache = {}
514 if func.__code__.co_argcount == 1:
514 if func.__code__.co_argcount == 1:
515 # we gain a small amount of time because
515 # we gain a small amount of time because
516 # we don't need to pack/unpack the list
516 # we don't need to pack/unpack the list
517 def f(arg):
517 def f(arg):
518 if arg not in cache:
518 if arg not in cache:
519 cache[arg] = func(arg)
519 cache[arg] = func(arg)
520 return cache[arg]
520 return cache[arg]
521 else:
521 else:
522 def f(*args):
522 def f(*args):
523 if args not in cache:
523 if args not in cache:
524 cache[args] = func(*args)
524 cache[args] = func(*args)
525 return cache[args]
525 return cache[args]
526
526
527 return f
527 return f
528
528
529 class sortdict(dict):
529 class sortdict(dict):
530 '''a simple sorted dictionary'''
530 '''a simple sorted dictionary'''
531 def __init__(self, data=None):
531 def __init__(self, data=None):
532 self._list = []
532 self._list = []
533 if data:
533 if data:
534 self.update(data)
534 self.update(data)
535 def copy(self):
535 def copy(self):
536 return sortdict(self)
536 return sortdict(self)
537 def __setitem__(self, key, val):
537 def __setitem__(self, key, val):
538 if key in self:
538 if key in self:
539 self._list.remove(key)
539 self._list.remove(key)
540 self._list.append(key)
540 self._list.append(key)
541 dict.__setitem__(self, key, val)
541 dict.__setitem__(self, key, val)
542 def __iter__(self):
542 def __iter__(self):
543 return self._list.__iter__()
543 return self._list.__iter__()
544 def update(self, src):
544 def update(self, src):
545 if isinstance(src, dict):
545 if isinstance(src, dict):
546 src = src.iteritems()
546 src = src.iteritems()
547 for k, v in src:
547 for k, v in src:
548 self[k] = v
548 self[k] = v
549 def clear(self):
549 def clear(self):
550 dict.clear(self)
550 dict.clear(self)
551 self._list = []
551 self._list = []
552 def items(self):
552 def items(self):
553 return [(k, self[k]) for k in self._list]
553 return [(k, self[k]) for k in self._list]
554 def __delitem__(self, key):
554 def __delitem__(self, key):
555 dict.__delitem__(self, key)
555 dict.__delitem__(self, key)
556 self._list.remove(key)
556 self._list.remove(key)
557 def pop(self, key, *args, **kwargs):
557 def pop(self, key, *args, **kwargs):
558 dict.pop(self, key, *args, **kwargs)
558 dict.pop(self, key, *args, **kwargs)
559 try:
559 try:
560 self._list.remove(key)
560 self._list.remove(key)
561 except ValueError:
561 except ValueError:
562 pass
562 pass
563 def keys(self):
563 def keys(self):
564 return self._list[:]
564 return self._list[:]
565 def iterkeys(self):
565 def iterkeys(self):
566 return self._list.__iter__()
566 return self._list.__iter__()
567 def iteritems(self):
567 def iteritems(self):
568 for k in self._list:
568 for k in self._list:
569 yield k, self[k]
569 yield k, self[k]
570 def insert(self, index, key, val):
570 def insert(self, index, key, val):
571 self._list.insert(index, key)
571 self._list.insert(index, key)
572 dict.__setitem__(self, key, val)
572 dict.__setitem__(self, key, val)
573 def __repr__(self):
573 def __repr__(self):
574 if not self:
574 if not self:
575 return '%s()' % self.__class__.__name__
575 return '%s()' % self.__class__.__name__
576 return '%s(%r)' % (self.__class__.__name__, self.items())
576 return '%s(%r)' % (self.__class__.__name__, self.items())
577
577
578 class _lrucachenode(object):
578 class _lrucachenode(object):
579 """A node in a doubly linked list.
579 """A node in a doubly linked list.
580
580
581 Holds a reference to nodes on either side as well as a key-value
581 Holds a reference to nodes on either side as well as a key-value
582 pair for the dictionary entry.
582 pair for the dictionary entry.
583 """
583 """
584 __slots__ = (u'next', u'prev', u'key', u'value')
584 __slots__ = (u'next', u'prev', u'key', u'value')
585
585
586 def __init__(self):
586 def __init__(self):
587 self.next = None
587 self.next = None
588 self.prev = None
588 self.prev = None
589
589
590 self.key = _notset
590 self.key = _notset
591 self.value = None
591 self.value = None
592
592
593 def markempty(self):
593 def markempty(self):
594 """Mark the node as emptied."""
594 """Mark the node as emptied."""
595 self.key = _notset
595 self.key = _notset
596
596
597 class lrucachedict(object):
597 class lrucachedict(object):
598 """Dict that caches most recent accesses and sets.
598 """Dict that caches most recent accesses and sets.
599
599
600 The dict consists of an actual backing dict - indexed by original
600 The dict consists of an actual backing dict - indexed by original
601 key - and a doubly linked circular list defining the order of entries in
601 key - and a doubly linked circular list defining the order of entries in
602 the cache.
602 the cache.
603
603
604 The head node is the newest entry in the cache. If the cache is full,
604 The head node is the newest entry in the cache. If the cache is full,
605 we recycle head.prev and make it the new head. Cache accesses result in
605 we recycle head.prev and make it the new head. Cache accesses result in
606 the node being moved to before the existing head and being marked as the
606 the node being moved to before the existing head and being marked as the
607 new head node.
607 new head node.
608 """
608 """
609 def __init__(self, max):
609 def __init__(self, max):
610 self._cache = {}
610 self._cache = {}
611
611
612 self._head = head = _lrucachenode()
612 self._head = head = _lrucachenode()
613 head.prev = head
613 head.prev = head
614 head.next = head
614 head.next = head
615 self._size = 1
615 self._size = 1
616 self._capacity = max
616 self._capacity = max
617
617
618 def __len__(self):
618 def __len__(self):
619 return len(self._cache)
619 return len(self._cache)
620
620
621 def __contains__(self, k):
621 def __contains__(self, k):
622 return k in self._cache
622 return k in self._cache
623
623
624 def __iter__(self):
624 def __iter__(self):
625 # We don't have to iterate in cache order, but why not.
625 # We don't have to iterate in cache order, but why not.
626 n = self._head
626 n = self._head
627 for i in range(len(self._cache)):
627 for i in range(len(self._cache)):
628 yield n.key
628 yield n.key
629 n = n.next
629 n = n.next
630
630
631 def __getitem__(self, k):
631 def __getitem__(self, k):
632 node = self._cache[k]
632 node = self._cache[k]
633 self._movetohead(node)
633 self._movetohead(node)
634 return node.value
634 return node.value
635
635
636 def __setitem__(self, k, v):
636 def __setitem__(self, k, v):
637 node = self._cache.get(k)
637 node = self._cache.get(k)
638 # Replace existing value and mark as newest.
638 # Replace existing value and mark as newest.
639 if node is not None:
639 if node is not None:
640 node.value = v
640 node.value = v
641 self._movetohead(node)
641 self._movetohead(node)
642 return
642 return
643
643
644 if self._size < self._capacity:
644 if self._size < self._capacity:
645 node = self._addcapacity()
645 node = self._addcapacity()
646 else:
646 else:
647 # Grab the last/oldest item.
647 # Grab the last/oldest item.
648 node = self._head.prev
648 node = self._head.prev
649
649
650 # At capacity. Kill the old entry.
650 # At capacity. Kill the old entry.
651 if node.key is not _notset:
651 if node.key is not _notset:
652 del self._cache[node.key]
652 del self._cache[node.key]
653
653
654 node.key = k
654 node.key = k
655 node.value = v
655 node.value = v
656 self._cache[k] = node
656 self._cache[k] = node
657 # And mark it as newest entry. No need to adjust order since it
657 # And mark it as newest entry. No need to adjust order since it
658 # is already self._head.prev.
658 # is already self._head.prev.
659 self._head = node
659 self._head = node
660
660
661 def __delitem__(self, k):
661 def __delitem__(self, k):
662 node = self._cache.pop(k)
662 node = self._cache.pop(k)
663 node.markempty()
663 node.markempty()
664
664
665 # Temporarily mark as newest item before re-adjusting head to make
665 # Temporarily mark as newest item before re-adjusting head to make
666 # this node the oldest item.
666 # this node the oldest item.
667 self._movetohead(node)
667 self._movetohead(node)
668 self._head = node.next
668 self._head = node.next
669
669
670 # Additional dict methods.
670 # Additional dict methods.
671
671
672 def get(self, k, default=None):
672 def get(self, k, default=None):
673 try:
673 try:
674 return self._cache[k].value
674 return self._cache[k].value
675 except KeyError:
675 except KeyError:
676 return default
676 return default
677
677
678 def clear(self):
678 def clear(self):
679 n = self._head
679 n = self._head
680 while n.key is not _notset:
680 while n.key is not _notset:
681 n.markempty()
681 n.markempty()
682 n = n.next
682 n = n.next
683
683
684 self._cache.clear()
684 self._cache.clear()
685
685
686 def copy(self):
686 def copy(self):
687 result = lrucachedict(self._capacity)
687 result = lrucachedict(self._capacity)
688 n = self._head.prev
688 n = self._head.prev
689 # Iterate in oldest-to-newest order, so the copy has the right ordering
689 # Iterate in oldest-to-newest order, so the copy has the right ordering
690 for i in range(len(self._cache)):
690 for i in range(len(self._cache)):
691 result[n.key] = n.value
691 result[n.key] = n.value
692 n = n.prev
692 n = n.prev
693 return result
693 return result
694
694
695 def _movetohead(self, node):
695 def _movetohead(self, node):
696 """Mark a node as the newest, making it the new head.
696 """Mark a node as the newest, making it the new head.
697
697
698 When a node is accessed, it becomes the freshest entry in the LRU
698 When a node is accessed, it becomes the freshest entry in the LRU
699 list, which is denoted by self._head.
699 list, which is denoted by self._head.
700
700
701 Visually, let's make ``N`` the new head node (* denotes head):
701 Visually, let's make ``N`` the new head node (* denotes head):
702
702
703 previous/oldest <-> head <-> next/next newest
703 previous/oldest <-> head <-> next/next newest
704
704
705 ----<->--- A* ---<->-----
705 ----<->--- A* ---<->-----
706 | |
706 | |
707 E <-> D <-> N <-> C <-> B
707 E <-> D <-> N <-> C <-> B
708
708
709 To:
709 To:
710
710
711 ----<->--- N* ---<->-----
711 ----<->--- N* ---<->-----
712 | |
712 | |
713 E <-> D <-> C <-> B <-> A
713 E <-> D <-> C <-> B <-> A
714
714
715 This requires the following moves:
715 This requires the following moves:
716
716
717 C.next = D (node.prev.next = node.next)
717 C.next = D (node.prev.next = node.next)
718 D.prev = C (node.next.prev = node.prev)
718 D.prev = C (node.next.prev = node.prev)
719 E.next = N (head.prev.next = node)
719 E.next = N (head.prev.next = node)
720 N.prev = E (node.prev = head.prev)
720 N.prev = E (node.prev = head.prev)
721 N.next = A (node.next = head)
721 N.next = A (node.next = head)
722 A.prev = N (head.prev = node)
722 A.prev = N (head.prev = node)
723 """
723 """
724 head = self._head
724 head = self._head
725 # C.next = D
725 # C.next = D
726 node.prev.next = node.next
726 node.prev.next = node.next
727 # D.prev = C
727 # D.prev = C
728 node.next.prev = node.prev
728 node.next.prev = node.prev
729 # N.prev = E
729 # N.prev = E
730 node.prev = head.prev
730 node.prev = head.prev
731 # N.next = A
731 # N.next = A
732 # It is tempting to do just "head" here, however if node is
732 # It is tempting to do just "head" here, however if node is
733 # adjacent to head, this will do bad things.
733 # adjacent to head, this will do bad things.
734 node.next = head.prev.next
734 node.next = head.prev.next
735 # E.next = N
735 # E.next = N
736 node.next.prev = node
736 node.next.prev = node
737 # A.prev = N
737 # A.prev = N
738 node.prev.next = node
738 node.prev.next = node
739
739
740 self._head = node
740 self._head = node
741
741
742 def _addcapacity(self):
742 def _addcapacity(self):
743 """Add a node to the circular linked list.
743 """Add a node to the circular linked list.
744
744
745 The new node is inserted before the head node.
745 The new node is inserted before the head node.
746 """
746 """
747 head = self._head
747 head = self._head
748 node = _lrucachenode()
748 node = _lrucachenode()
749 head.prev.next = node
749 head.prev.next = node
750 node.prev = head.prev
750 node.prev = head.prev
751 node.next = head
751 node.next = head
752 head.prev = node
752 head.prev = node
753 self._size += 1
753 self._size += 1
754 return node
754 return node
755
755
756 def lrucachefunc(func):
756 def lrucachefunc(func):
757 '''cache most recent results of function calls'''
757 '''cache most recent results of function calls'''
758 cache = {}
758 cache = {}
759 order = collections.deque()
759 order = collections.deque()
760 if func.__code__.co_argcount == 1:
760 if func.__code__.co_argcount == 1:
761 def f(arg):
761 def f(arg):
762 if arg not in cache:
762 if arg not in cache:
763 if len(cache) > 20:
763 if len(cache) > 20:
764 del cache[order.popleft()]
764 del cache[order.popleft()]
765 cache[arg] = func(arg)
765 cache[arg] = func(arg)
766 else:
766 else:
767 order.remove(arg)
767 order.remove(arg)
768 order.append(arg)
768 order.append(arg)
769 return cache[arg]
769 return cache[arg]
770 else:
770 else:
771 def f(*args):
771 def f(*args):
772 if args not in cache:
772 if args not in cache:
773 if len(cache) > 20:
773 if len(cache) > 20:
774 del cache[order.popleft()]
774 del cache[order.popleft()]
775 cache[args] = func(*args)
775 cache[args] = func(*args)
776 else:
776 else:
777 order.remove(args)
777 order.remove(args)
778 order.append(args)
778 order.append(args)
779 return cache[args]
779 return cache[args]
780
780
781 return f
781 return f
782
782
783 class propertycache(object):
783 class propertycache(object):
784 def __init__(self, func):
784 def __init__(self, func):
785 self.func = func
785 self.func = func
786 self.name = func.__name__
786 self.name = func.__name__
787 def __get__(self, obj, type=None):
787 def __get__(self, obj, type=None):
788 result = self.func(obj)
788 result = self.func(obj)
789 self.cachevalue(obj, result)
789 self.cachevalue(obj, result)
790 return result
790 return result
791
791
792 def cachevalue(self, obj, value):
792 def cachevalue(self, obj, value):
793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
794 obj.__dict__[self.name] = value
794 obj.__dict__[self.name] = value
795
795
796 def pipefilter(s, cmd):
796 def pipefilter(s, cmd):
797 '''filter string S through command CMD, returning its output'''
797 '''filter string S through command CMD, returning its output'''
798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
800 pout, perr = p.communicate(s)
800 pout, perr = p.communicate(s)
801 return pout
801 return pout
802
802
803 def tempfilter(s, cmd):
803 def tempfilter(s, cmd):
804 '''filter string S through a pair of temporary files with CMD.
804 '''filter string S through a pair of temporary files with CMD.
805 CMD is used as a template to create the real command to be run,
805 CMD is used as a template to create the real command to be run,
806 with the strings INFILE and OUTFILE replaced by the real names of
806 with the strings INFILE and OUTFILE replaced by the real names of
807 the temporary files generated.'''
807 the temporary files generated.'''
808 inname, outname = None, None
808 inname, outname = None, None
809 try:
809 try:
810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
812 fp.write(s)
812 fp.write(s)
813 fp.close()
813 fp.close()
814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
815 os.close(outfd)
815 os.close(outfd)
816 cmd = cmd.replace('INFILE', inname)
816 cmd = cmd.replace('INFILE', inname)
817 cmd = cmd.replace('OUTFILE', outname)
817 cmd = cmd.replace('OUTFILE', outname)
818 code = os.system(cmd)
818 code = os.system(cmd)
819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
820 code = 0
820 code = 0
821 if code:
821 if code:
822 raise Abort(_("command '%s' failed: %s") %
822 raise Abort(_("command '%s' failed: %s") %
823 (cmd, explainexit(code)))
823 (cmd, explainexit(code)))
824 return readfile(outname)
824 return readfile(outname)
825 finally:
825 finally:
826 try:
826 try:
827 if inname:
827 if inname:
828 os.unlink(inname)
828 os.unlink(inname)
829 except OSError:
829 except OSError:
830 pass
830 pass
831 try:
831 try:
832 if outname:
832 if outname:
833 os.unlink(outname)
833 os.unlink(outname)
834 except OSError:
834 except OSError:
835 pass
835 pass
836
836
837 filtertable = {
837 filtertable = {
838 'tempfile:': tempfilter,
838 'tempfile:': tempfilter,
839 'pipe:': pipefilter,
839 'pipe:': pipefilter,
840 }
840 }
841
841
842 def filter(s, cmd):
842 def filter(s, cmd):
843 "filter a string through a command that transforms its input to its output"
843 "filter a string through a command that transforms its input to its output"
844 for name, fn in filtertable.iteritems():
844 for name, fn in filtertable.iteritems():
845 if cmd.startswith(name):
845 if cmd.startswith(name):
846 return fn(s, cmd[len(name):].lstrip())
846 return fn(s, cmd[len(name):].lstrip())
847 return pipefilter(s, cmd)
847 return pipefilter(s, cmd)
848
848
849 def binary(s):
849 def binary(s):
850 """return true if a string is binary data"""
850 """return true if a string is binary data"""
851 return bool(s and '\0' in s)
851 return bool(s and '\0' in s)
852
852
853 def increasingchunks(source, min=1024, max=65536):
853 def increasingchunks(source, min=1024, max=65536):
854 '''return no less than min bytes per chunk while data remains,
854 '''return no less than min bytes per chunk while data remains,
855 doubling min after each chunk until it reaches max'''
855 doubling min after each chunk until it reaches max'''
856 def log2(x):
856 def log2(x):
857 if not x:
857 if not x:
858 return 0
858 return 0
859 i = 0
859 i = 0
860 while x:
860 while x:
861 x >>= 1
861 x >>= 1
862 i += 1
862 i += 1
863 return i - 1
863 return i - 1
864
864
865 buf = []
865 buf = []
866 blen = 0
866 blen = 0
867 for chunk in source:
867 for chunk in source:
868 buf.append(chunk)
868 buf.append(chunk)
869 blen += len(chunk)
869 blen += len(chunk)
870 if blen >= min:
870 if blen >= min:
871 if min < max:
871 if min < max:
872 min = min << 1
872 min = min << 1
873 nmin = 1 << log2(blen)
873 nmin = 1 << log2(blen)
874 if nmin > min:
874 if nmin > min:
875 min = nmin
875 min = nmin
876 if min > max:
876 if min > max:
877 min = max
877 min = max
878 yield ''.join(buf)
878 yield ''.join(buf)
879 blen = 0
879 blen = 0
880 buf = []
880 buf = []
881 if buf:
881 if buf:
882 yield ''.join(buf)
882 yield ''.join(buf)
883
883
884 Abort = error.Abort
884 Abort = error.Abort
885
885
886 def always(fn):
886 def always(fn):
887 return True
887 return True
888
888
889 def never(fn):
889 def never(fn):
890 return False
890 return False
891
891
892 def nogc(func):
892 def nogc(func):
893 """disable garbage collector
893 """disable garbage collector
894
894
895 Python's garbage collector triggers a GC each time a certain number of
895 Python's garbage collector triggers a GC each time a certain number of
896 container objects (the number being defined by gc.get_threshold()) are
896 container objects (the number being defined by gc.get_threshold()) are
897 allocated even when marked not to be tracked by the collector. Tracking has
897 allocated even when marked not to be tracked by the collector. Tracking has
898 no effect on when GCs are triggered, only on what objects the GC looks
898 no effect on when GCs are triggered, only on what objects the GC looks
899 into. As a workaround, disable GC while building complex (huge)
899 into. As a workaround, disable GC while building complex (huge)
900 containers.
900 containers.
901
901
902 This garbage collector issue have been fixed in 2.7.
902 This garbage collector issue have been fixed in 2.7.
903 """
903 """
904 if sys.version_info >= (2, 7):
904 if sys.version_info >= (2, 7):
905 return func
905 return func
906 def wrapper(*args, **kwargs):
906 def wrapper(*args, **kwargs):
907 gcenabled = gc.isenabled()
907 gcenabled = gc.isenabled()
908 gc.disable()
908 gc.disable()
909 try:
909 try:
910 return func(*args, **kwargs)
910 return func(*args, **kwargs)
911 finally:
911 finally:
912 if gcenabled:
912 if gcenabled:
913 gc.enable()
913 gc.enable()
914 return wrapper
914 return wrapper
915
915
916 def pathto(root, n1, n2):
916 def pathto(root, n1, n2):
917 '''return the relative path from one place to another.
917 '''return the relative path from one place to another.
918 root should use os.sep to separate directories
918 root should use os.sep to separate directories
919 n1 should use os.sep to separate directories
919 n1 should use os.sep to separate directories
920 n2 should use "/" to separate directories
920 n2 should use "/" to separate directories
921 returns an os.sep-separated path.
921 returns an os.sep-separated path.
922
922
923 If n1 is a relative path, it's assumed it's
923 If n1 is a relative path, it's assumed it's
924 relative to root.
924 relative to root.
925 n2 should always be relative to root.
925 n2 should always be relative to root.
926 '''
926 '''
927 if not n1:
927 if not n1:
928 return localpath(n2)
928 return localpath(n2)
929 if os.path.isabs(n1):
929 if os.path.isabs(n1):
930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
931 return os.path.join(root, localpath(n2))
931 return os.path.join(root, localpath(n2))
932 n2 = '/'.join((pconvert(root), n2))
932 n2 = '/'.join((pconvert(root), n2))
933 a, b = splitpath(n1), n2.split('/')
933 a, b = splitpath(n1), n2.split('/')
934 a.reverse()
934 a.reverse()
935 b.reverse()
935 b.reverse()
936 while a and b and a[-1] == b[-1]:
936 while a and b and a[-1] == b[-1]:
937 a.pop()
937 a.pop()
938 b.pop()
938 b.pop()
939 b.reverse()
939 b.reverse()
940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
941
941
942 def mainfrozen():
942 def mainfrozen():
943 """return True if we are a frozen executable.
943 """return True if we are a frozen executable.
944
944
945 The code supports py2exe (most common, Windows only) and tools/freeze
945 The code supports py2exe (most common, Windows only) and tools/freeze
946 (portable, not much used).
946 (portable, not much used).
947 """
947 """
948 return (safehasattr(sys, "frozen") or # new py2exe
948 return (safehasattr(sys, "frozen") or # new py2exe
949 safehasattr(sys, "importers") or # old py2exe
949 safehasattr(sys, "importers") or # old py2exe
950 imp.is_frozen(u"__main__")) # tools/freeze
950 imp.is_frozen(u"__main__")) # tools/freeze
951
951
952 # the location of data files matching the source code
952 # the location of data files matching the source code
953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
954 # executable version (py2exe) doesn't support __file__
954 # executable version (py2exe) doesn't support __file__
955 datapath = os.path.dirname(pycompat.sysexecutable)
955 datapath = os.path.dirname(pycompat.sysexecutable)
956 else:
956 else:
957 datapath = os.path.dirname(pycompat.fsencode(__file__))
957 datapath = os.path.dirname(pycompat.fsencode(__file__))
958
958
959 i18n.setdatapath(datapath)
959 i18n.setdatapath(datapath)
960
960
961 _hgexecutable = None
961 _hgexecutable = None
962
962
963 def hgexecutable():
963 def hgexecutable():
964 """return location of the 'hg' executable.
964 """return location of the 'hg' executable.
965
965
966 Defaults to $HG or 'hg' in the search path.
966 Defaults to $HG or 'hg' in the search path.
967 """
967 """
968 if _hgexecutable is None:
968 if _hgexecutable is None:
969 hg = encoding.environ.get('HG')
969 hg = encoding.environ.get('HG')
970 mainmod = sys.modules[pycompat.sysstr('__main__')]
970 mainmod = sys.modules[pycompat.sysstr('__main__')]
971 if hg:
971 if hg:
972 _sethgexecutable(hg)
972 _sethgexecutable(hg)
973 elif mainfrozen():
973 elif mainfrozen():
974 if getattr(sys, 'frozen', None) == 'macosx_app':
974 if getattr(sys, 'frozen', None) == 'macosx_app':
975 # Env variable set by py2app
975 # Env variable set by py2app
976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
977 else:
977 else:
978 _sethgexecutable(pycompat.sysexecutable)
978 _sethgexecutable(pycompat.sysexecutable)
979 elif (os.path.basename(
979 elif (os.path.basename(
980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
982 else:
982 else:
983 exe = findexe('hg') or os.path.basename(sys.argv[0])
983 exe = findexe('hg') or os.path.basename(sys.argv[0])
984 _sethgexecutable(exe)
984 _sethgexecutable(exe)
985 return _hgexecutable
985 return _hgexecutable
986
986
987 def _sethgexecutable(path):
987 def _sethgexecutable(path):
988 """set location of the 'hg' executable"""
988 """set location of the 'hg' executable"""
989 global _hgexecutable
989 global _hgexecutable
990 _hgexecutable = path
990 _hgexecutable = path
991
991
992 def _isstdout(f):
992 def _isstdout(f):
993 fileno = getattr(f, 'fileno', None)
993 fileno = getattr(f, 'fileno', None)
994 return fileno and fileno() == sys.__stdout__.fileno()
994 return fileno and fileno() == sys.__stdout__.fileno()
995
995
996 def shellenviron(environ=None):
996 def shellenviron(environ=None):
997 """return environ with optional override, useful for shelling out"""
997 """return environ with optional override, useful for shelling out"""
998 def py2shell(val):
998 def py2shell(val):
999 'convert python object into string that is useful to shell'
999 'convert python object into string that is useful to shell'
1000 if val is None or val is False:
1000 if val is None or val is False:
1001 return '0'
1001 return '0'
1002 if val is True:
1002 if val is True:
1003 return '1'
1003 return '1'
1004 return str(val)
1004 return str(val)
1005 env = dict(encoding.environ)
1005 env = dict(encoding.environ)
1006 if environ:
1006 if environ:
1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1008 env['HG'] = hgexecutable()
1008 env['HG'] = hgexecutable()
1009 return env
1009 return env
1010
1010
1011 def system(cmd, environ=None, cwd=None, out=None):
1011 def system(cmd, environ=None, cwd=None, out=None):
1012 '''enhanced shell command execution.
1012 '''enhanced shell command execution.
1013 run with environment maybe modified, maybe in different dir.
1013 run with environment maybe modified, maybe in different dir.
1014
1014
1015 if out is specified, it is assumed to be a file-like object that has a
1015 if out is specified, it is assumed to be a file-like object that has a
1016 write() method. stdout and stderr will be redirected to out.'''
1016 write() method. stdout and stderr will be redirected to out.'''
1017 try:
1017 try:
1018 stdout.flush()
1018 stdout.flush()
1019 except Exception:
1019 except Exception:
1020 pass
1020 pass
1021 cmd = quotecommand(cmd)
1021 cmd = quotecommand(cmd)
1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1023 and sys.version_info[1] < 7):
1023 and sys.version_info[1] < 7):
1024 # subprocess kludge to work around issues in half-baked Python
1024 # subprocess kludge to work around issues in half-baked Python
1025 # ports, notably bichued/python:
1025 # ports, notably bichued/python:
1026 if not cwd is None:
1026 if not cwd is None:
1027 os.chdir(cwd)
1027 os.chdir(cwd)
1028 rc = os.system(cmd)
1028 rc = os.system(cmd)
1029 else:
1029 else:
1030 env = shellenviron(environ)
1030 env = shellenviron(environ)
1031 if out is None or _isstdout(out):
1031 if out is None or _isstdout(out):
1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1033 env=env, cwd=cwd)
1033 env=env, cwd=cwd)
1034 else:
1034 else:
1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1037 stderr=subprocess.STDOUT)
1037 stderr=subprocess.STDOUT)
1038 for line in iter(proc.stdout.readline, ''):
1038 for line in iter(proc.stdout.readline, ''):
1039 out.write(line)
1039 out.write(line)
1040 proc.wait()
1040 proc.wait()
1041 rc = proc.returncode
1041 rc = proc.returncode
1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1043 rc = 0
1043 rc = 0
1044 return rc
1044 return rc
1045
1045
1046 def checksignature(func):
1046 def checksignature(func):
1047 '''wrap a function with code to check for calling errors'''
1047 '''wrap a function with code to check for calling errors'''
1048 def check(*args, **kwargs):
1048 def check(*args, **kwargs):
1049 try:
1049 try:
1050 return func(*args, **kwargs)
1050 return func(*args, **kwargs)
1051 except TypeError:
1051 except TypeError:
1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1053 raise error.SignatureError
1053 raise error.SignatureError
1054 raise
1054 raise
1055
1055
1056 return check
1056 return check
1057
1057
1058 # a whilelist of known filesystems where hardlink works reliably
1058 # a whilelist of known filesystems where hardlink works reliably
1059 _hardlinkfswhitelist = set([
1059 _hardlinkfswhitelist = set([
1060 'btrfs',
1060 'btrfs',
1061 'ext2',
1061 'ext2',
1062 'ext3',
1062 'ext3',
1063 'ext4',
1063 'ext4',
1064 'hfs',
1064 'jfs',
1065 'jfs',
1065 'reiserfs',
1066 'reiserfs',
1066 'tmpfs',
1067 'tmpfs',
1068 'ufs',
1067 'xfs',
1069 'xfs',
1070 'zfs',
1068 ])
1071 ])
1069
1072
1070 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1073 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1071 '''copy a file, preserving mode and optionally other stat info like
1074 '''copy a file, preserving mode and optionally other stat info like
1072 atime/mtime
1075 atime/mtime
1073
1076
1074 checkambig argument is used with filestat, and is useful only if
1077 checkambig argument is used with filestat, and is useful only if
1075 destination file is guarded by any lock (e.g. repo.lock or
1078 destination file is guarded by any lock (e.g. repo.lock or
1076 repo.wlock).
1079 repo.wlock).
1077
1080
1078 copystat and checkambig should be exclusive.
1081 copystat and checkambig should be exclusive.
1079 '''
1082 '''
1080 assert not (copystat and checkambig)
1083 assert not (copystat and checkambig)
1081 oldstat = None
1084 oldstat = None
1082 if os.path.lexists(dest):
1085 if os.path.lexists(dest):
1083 if checkambig:
1086 if checkambig:
1084 oldstat = checkambig and filestat(dest)
1087 oldstat = checkambig and filestat(dest)
1085 unlink(dest)
1088 unlink(dest)
1086 if hardlink:
1089 if hardlink:
1087 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1090 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1088 # unless we are confident that dest is on a whitelisted filesystem.
1091 # unless we are confident that dest is on a whitelisted filesystem.
1089 fstype = getfstype(os.path.dirname(dest))
1092 fstype = getfstype(os.path.dirname(dest))
1090 if fstype not in _hardlinkfswhitelist:
1093 if fstype not in _hardlinkfswhitelist:
1091 hardlink = False
1094 hardlink = False
1092 if hardlink:
1095 if hardlink:
1093 try:
1096 try:
1094 oslink(src, dest)
1097 oslink(src, dest)
1095 return
1098 return
1096 except (IOError, OSError):
1099 except (IOError, OSError):
1097 pass # fall back to normal copy
1100 pass # fall back to normal copy
1098 if os.path.islink(src):
1101 if os.path.islink(src):
1099 os.symlink(os.readlink(src), dest)
1102 os.symlink(os.readlink(src), dest)
1100 # copytime is ignored for symlinks, but in general copytime isn't needed
1103 # copytime is ignored for symlinks, but in general copytime isn't needed
1101 # for them anyway
1104 # for them anyway
1102 else:
1105 else:
1103 try:
1106 try:
1104 shutil.copyfile(src, dest)
1107 shutil.copyfile(src, dest)
1105 if copystat:
1108 if copystat:
1106 # copystat also copies mode
1109 # copystat also copies mode
1107 shutil.copystat(src, dest)
1110 shutil.copystat(src, dest)
1108 else:
1111 else:
1109 shutil.copymode(src, dest)
1112 shutil.copymode(src, dest)
1110 if oldstat and oldstat.stat:
1113 if oldstat and oldstat.stat:
1111 newstat = filestat(dest)
1114 newstat = filestat(dest)
1112 if newstat.isambig(oldstat):
1115 if newstat.isambig(oldstat):
1113 # stat of copied file is ambiguous to original one
1116 # stat of copied file is ambiguous to original one
1114 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1117 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1115 os.utime(dest, (advanced, advanced))
1118 os.utime(dest, (advanced, advanced))
1116 except shutil.Error as inst:
1119 except shutil.Error as inst:
1117 raise Abort(str(inst))
1120 raise Abort(str(inst))
1118
1121
1119 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1122 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1120 """Copy a directory tree using hardlinks if possible."""
1123 """Copy a directory tree using hardlinks if possible."""
1121 num = 0
1124 num = 0
1122
1125
1123 if hardlink is None:
1126 if hardlink is None:
1124 hardlink = (os.stat(src).st_dev ==
1127 hardlink = (os.stat(src).st_dev ==
1125 os.stat(os.path.dirname(dst)).st_dev)
1128 os.stat(os.path.dirname(dst)).st_dev)
1126 if hardlink:
1129 if hardlink:
1127 topic = _('linking')
1130 topic = _('linking')
1128 else:
1131 else:
1129 topic = _('copying')
1132 topic = _('copying')
1130
1133
1131 if os.path.isdir(src):
1134 if os.path.isdir(src):
1132 os.mkdir(dst)
1135 os.mkdir(dst)
1133 for name, kind in osutil.listdir(src):
1136 for name, kind in osutil.listdir(src):
1134 srcname = os.path.join(src, name)
1137 srcname = os.path.join(src, name)
1135 dstname = os.path.join(dst, name)
1138 dstname = os.path.join(dst, name)
1136 def nprog(t, pos):
1139 def nprog(t, pos):
1137 if pos is not None:
1140 if pos is not None:
1138 return progress(t, pos + num)
1141 return progress(t, pos + num)
1139 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1142 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1140 num += n
1143 num += n
1141 else:
1144 else:
1142 if hardlink:
1145 if hardlink:
1143 try:
1146 try:
1144 oslink(src, dst)
1147 oslink(src, dst)
1145 except (IOError, OSError):
1148 except (IOError, OSError):
1146 hardlink = False
1149 hardlink = False
1147 shutil.copy(src, dst)
1150 shutil.copy(src, dst)
1148 else:
1151 else:
1149 shutil.copy(src, dst)
1152 shutil.copy(src, dst)
1150 num += 1
1153 num += 1
1151 progress(topic, num)
1154 progress(topic, num)
1152 progress(topic, None)
1155 progress(topic, None)
1153
1156
1154 return hardlink, num
1157 return hardlink, num
1155
1158
1156 _winreservednames = '''con prn aux nul
1159 _winreservednames = '''con prn aux nul
1157 com1 com2 com3 com4 com5 com6 com7 com8 com9
1160 com1 com2 com3 com4 com5 com6 com7 com8 com9
1158 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1161 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1159 _winreservedchars = ':*?"<>|'
1162 _winreservedchars = ':*?"<>|'
1160 def checkwinfilename(path):
1163 def checkwinfilename(path):
1161 r'''Check that the base-relative path is a valid filename on Windows.
1164 r'''Check that the base-relative path is a valid filename on Windows.
1162 Returns None if the path is ok, or a UI string describing the problem.
1165 Returns None if the path is ok, or a UI string describing the problem.
1163
1166
1164 >>> checkwinfilename("just/a/normal/path")
1167 >>> checkwinfilename("just/a/normal/path")
1165 >>> checkwinfilename("foo/bar/con.xml")
1168 >>> checkwinfilename("foo/bar/con.xml")
1166 "filename contains 'con', which is reserved on Windows"
1169 "filename contains 'con', which is reserved on Windows"
1167 >>> checkwinfilename("foo/con.xml/bar")
1170 >>> checkwinfilename("foo/con.xml/bar")
1168 "filename contains 'con', which is reserved on Windows"
1171 "filename contains 'con', which is reserved on Windows"
1169 >>> checkwinfilename("foo/bar/xml.con")
1172 >>> checkwinfilename("foo/bar/xml.con")
1170 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1173 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1171 "filename contains 'AUX', which is reserved on Windows"
1174 "filename contains 'AUX', which is reserved on Windows"
1172 >>> checkwinfilename("foo/bar/bla:.txt")
1175 >>> checkwinfilename("foo/bar/bla:.txt")
1173 "filename contains ':', which is reserved on Windows"
1176 "filename contains ':', which is reserved on Windows"
1174 >>> checkwinfilename("foo/bar/b\07la.txt")
1177 >>> checkwinfilename("foo/bar/b\07la.txt")
1175 "filename contains '\\x07', which is invalid on Windows"
1178 "filename contains '\\x07', which is invalid on Windows"
1176 >>> checkwinfilename("foo/bar/bla ")
1179 >>> checkwinfilename("foo/bar/bla ")
1177 "filename ends with ' ', which is not allowed on Windows"
1180 "filename ends with ' ', which is not allowed on Windows"
1178 >>> checkwinfilename("../bar")
1181 >>> checkwinfilename("../bar")
1179 >>> checkwinfilename("foo\\")
1182 >>> checkwinfilename("foo\\")
1180 "filename ends with '\\', which is invalid on Windows"
1183 "filename ends with '\\', which is invalid on Windows"
1181 >>> checkwinfilename("foo\\/bar")
1184 >>> checkwinfilename("foo\\/bar")
1182 "directory name ends with '\\', which is invalid on Windows"
1185 "directory name ends with '\\', which is invalid on Windows"
1183 '''
1186 '''
1184 if path.endswith('\\'):
1187 if path.endswith('\\'):
1185 return _("filename ends with '\\', which is invalid on Windows")
1188 return _("filename ends with '\\', which is invalid on Windows")
1186 if '\\/' in path:
1189 if '\\/' in path:
1187 return _("directory name ends with '\\', which is invalid on Windows")
1190 return _("directory name ends with '\\', which is invalid on Windows")
1188 for n in path.replace('\\', '/').split('/'):
1191 for n in path.replace('\\', '/').split('/'):
1189 if not n:
1192 if not n:
1190 continue
1193 continue
1191 for c in pycompat.bytestr(n):
1194 for c in pycompat.bytestr(n):
1192 if c in _winreservedchars:
1195 if c in _winreservedchars:
1193 return _("filename contains '%s', which is reserved "
1196 return _("filename contains '%s', which is reserved "
1194 "on Windows") % c
1197 "on Windows") % c
1195 if ord(c) <= 31:
1198 if ord(c) <= 31:
1196 return _("filename contains %r, which is invalid "
1199 return _("filename contains %r, which is invalid "
1197 "on Windows") % c
1200 "on Windows") % c
1198 base = n.split('.')[0]
1201 base = n.split('.')[0]
1199 if base and base.lower() in _winreservednames:
1202 if base and base.lower() in _winreservednames:
1200 return _("filename contains '%s', which is reserved "
1203 return _("filename contains '%s', which is reserved "
1201 "on Windows") % base
1204 "on Windows") % base
1202 t = n[-1]
1205 t = n[-1]
1203 if t in '. ' and n not in '..':
1206 if t in '. ' and n not in '..':
1204 return _("filename ends with '%s', which is not allowed "
1207 return _("filename ends with '%s', which is not allowed "
1205 "on Windows") % t
1208 "on Windows") % t
1206
1209
1207 if pycompat.osname == 'nt':
1210 if pycompat.osname == 'nt':
1208 checkosfilename = checkwinfilename
1211 checkosfilename = checkwinfilename
1209 timer = time.clock
1212 timer = time.clock
1210 else:
1213 else:
1211 checkosfilename = platform.checkosfilename
1214 checkosfilename = platform.checkosfilename
1212 timer = time.time
1215 timer = time.time
1213
1216
1214 if safehasattr(time, "perf_counter"):
1217 if safehasattr(time, "perf_counter"):
1215 timer = time.perf_counter
1218 timer = time.perf_counter
1216
1219
1217 def makelock(info, pathname):
1220 def makelock(info, pathname):
1218 try:
1221 try:
1219 return os.symlink(info, pathname)
1222 return os.symlink(info, pathname)
1220 except OSError as why:
1223 except OSError as why:
1221 if why.errno == errno.EEXIST:
1224 if why.errno == errno.EEXIST:
1222 raise
1225 raise
1223 except AttributeError: # no symlink in os
1226 except AttributeError: # no symlink in os
1224 pass
1227 pass
1225
1228
1226 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1229 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1227 os.write(ld, info)
1230 os.write(ld, info)
1228 os.close(ld)
1231 os.close(ld)
1229
1232
1230 def readlock(pathname):
1233 def readlock(pathname):
1231 try:
1234 try:
1232 return os.readlink(pathname)
1235 return os.readlink(pathname)
1233 except OSError as why:
1236 except OSError as why:
1234 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1237 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1235 raise
1238 raise
1236 except AttributeError: # no symlink in os
1239 except AttributeError: # no symlink in os
1237 pass
1240 pass
1238 fp = posixfile(pathname)
1241 fp = posixfile(pathname)
1239 r = fp.read()
1242 r = fp.read()
1240 fp.close()
1243 fp.close()
1241 return r
1244 return r
1242
1245
1243 def fstat(fp):
1246 def fstat(fp):
1244 '''stat file object that may not have fileno method.'''
1247 '''stat file object that may not have fileno method.'''
1245 try:
1248 try:
1246 return os.fstat(fp.fileno())
1249 return os.fstat(fp.fileno())
1247 except AttributeError:
1250 except AttributeError:
1248 return os.stat(fp.name)
1251 return os.stat(fp.name)
1249
1252
1250 # File system features
1253 # File system features
1251
1254
1252 def fscasesensitive(path):
1255 def fscasesensitive(path):
1253 """
1256 """
1254 Return true if the given path is on a case-sensitive filesystem
1257 Return true if the given path is on a case-sensitive filesystem
1255
1258
1256 Requires a path (like /foo/.hg) ending with a foldable final
1259 Requires a path (like /foo/.hg) ending with a foldable final
1257 directory component.
1260 directory component.
1258 """
1261 """
1259 s1 = os.lstat(path)
1262 s1 = os.lstat(path)
1260 d, b = os.path.split(path)
1263 d, b = os.path.split(path)
1261 b2 = b.upper()
1264 b2 = b.upper()
1262 if b == b2:
1265 if b == b2:
1263 b2 = b.lower()
1266 b2 = b.lower()
1264 if b == b2:
1267 if b == b2:
1265 return True # no evidence against case sensitivity
1268 return True # no evidence against case sensitivity
1266 p2 = os.path.join(d, b2)
1269 p2 = os.path.join(d, b2)
1267 try:
1270 try:
1268 s2 = os.lstat(p2)
1271 s2 = os.lstat(p2)
1269 if s2 == s1:
1272 if s2 == s1:
1270 return False
1273 return False
1271 return True
1274 return True
1272 except OSError:
1275 except OSError:
1273 return True
1276 return True
1274
1277
1275 try:
1278 try:
1276 import re2
1279 import re2
1277 _re2 = None
1280 _re2 = None
1278 except ImportError:
1281 except ImportError:
1279 _re2 = False
1282 _re2 = False
1280
1283
1281 class _re(object):
1284 class _re(object):
1282 def _checkre2(self):
1285 def _checkre2(self):
1283 global _re2
1286 global _re2
1284 try:
1287 try:
1285 # check if match works, see issue3964
1288 # check if match works, see issue3964
1286 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1289 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1287 except ImportError:
1290 except ImportError:
1288 _re2 = False
1291 _re2 = False
1289
1292
1290 def compile(self, pat, flags=0):
1293 def compile(self, pat, flags=0):
1291 '''Compile a regular expression, using re2 if possible
1294 '''Compile a regular expression, using re2 if possible
1292
1295
1293 For best performance, use only re2-compatible regexp features. The
1296 For best performance, use only re2-compatible regexp features. The
1294 only flags from the re module that are re2-compatible are
1297 only flags from the re module that are re2-compatible are
1295 IGNORECASE and MULTILINE.'''
1298 IGNORECASE and MULTILINE.'''
1296 if _re2 is None:
1299 if _re2 is None:
1297 self._checkre2()
1300 self._checkre2()
1298 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1301 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1299 if flags & remod.IGNORECASE:
1302 if flags & remod.IGNORECASE:
1300 pat = '(?i)' + pat
1303 pat = '(?i)' + pat
1301 if flags & remod.MULTILINE:
1304 if flags & remod.MULTILINE:
1302 pat = '(?m)' + pat
1305 pat = '(?m)' + pat
1303 try:
1306 try:
1304 return re2.compile(pat)
1307 return re2.compile(pat)
1305 except re2.error:
1308 except re2.error:
1306 pass
1309 pass
1307 return remod.compile(pat, flags)
1310 return remod.compile(pat, flags)
1308
1311
1309 @propertycache
1312 @propertycache
1310 def escape(self):
1313 def escape(self):
1311 '''Return the version of escape corresponding to self.compile.
1314 '''Return the version of escape corresponding to self.compile.
1312
1315
1313 This is imperfect because whether re2 or re is used for a particular
1316 This is imperfect because whether re2 or re is used for a particular
1314 function depends on the flags, etc, but it's the best we can do.
1317 function depends on the flags, etc, but it's the best we can do.
1315 '''
1318 '''
1316 global _re2
1319 global _re2
1317 if _re2 is None:
1320 if _re2 is None:
1318 self._checkre2()
1321 self._checkre2()
1319 if _re2:
1322 if _re2:
1320 return re2.escape
1323 return re2.escape
1321 else:
1324 else:
1322 return remod.escape
1325 return remod.escape
1323
1326
1324 re = _re()
1327 re = _re()
1325
1328
1326 _fspathcache = {}
1329 _fspathcache = {}
1327 def fspath(name, root):
1330 def fspath(name, root):
1328 '''Get name in the case stored in the filesystem
1331 '''Get name in the case stored in the filesystem
1329
1332
1330 The name should be relative to root, and be normcase-ed for efficiency.
1333 The name should be relative to root, and be normcase-ed for efficiency.
1331
1334
1332 Note that this function is unnecessary, and should not be
1335 Note that this function is unnecessary, and should not be
1333 called, for case-sensitive filesystems (simply because it's expensive).
1336 called, for case-sensitive filesystems (simply because it's expensive).
1334
1337
1335 The root should be normcase-ed, too.
1338 The root should be normcase-ed, too.
1336 '''
1339 '''
1337 def _makefspathcacheentry(dir):
1340 def _makefspathcacheentry(dir):
1338 return dict((normcase(n), n) for n in os.listdir(dir))
1341 return dict((normcase(n), n) for n in os.listdir(dir))
1339
1342
1340 seps = pycompat.ossep
1343 seps = pycompat.ossep
1341 if pycompat.osaltsep:
1344 if pycompat.osaltsep:
1342 seps = seps + pycompat.osaltsep
1345 seps = seps + pycompat.osaltsep
1343 # Protect backslashes. This gets silly very quickly.
1346 # Protect backslashes. This gets silly very quickly.
1344 seps.replace('\\','\\\\')
1347 seps.replace('\\','\\\\')
1345 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1348 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1346 dir = os.path.normpath(root)
1349 dir = os.path.normpath(root)
1347 result = []
1350 result = []
1348 for part, sep in pattern.findall(name):
1351 for part, sep in pattern.findall(name):
1349 if sep:
1352 if sep:
1350 result.append(sep)
1353 result.append(sep)
1351 continue
1354 continue
1352
1355
1353 if dir not in _fspathcache:
1356 if dir not in _fspathcache:
1354 _fspathcache[dir] = _makefspathcacheentry(dir)
1357 _fspathcache[dir] = _makefspathcacheentry(dir)
1355 contents = _fspathcache[dir]
1358 contents = _fspathcache[dir]
1356
1359
1357 found = contents.get(part)
1360 found = contents.get(part)
1358 if not found:
1361 if not found:
1359 # retry "once per directory" per "dirstate.walk" which
1362 # retry "once per directory" per "dirstate.walk" which
1360 # may take place for each patches of "hg qpush", for example
1363 # may take place for each patches of "hg qpush", for example
1361 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1364 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1362 found = contents.get(part)
1365 found = contents.get(part)
1363
1366
1364 result.append(found or part)
1367 result.append(found or part)
1365 dir = os.path.join(dir, part)
1368 dir = os.path.join(dir, part)
1366
1369
1367 return ''.join(result)
1370 return ''.join(result)
1368
1371
1369 def getfstype(dirpath):
1372 def getfstype(dirpath):
1370 '''Get the filesystem type name from a directory (best-effort)
1373 '''Get the filesystem type name from a directory (best-effort)
1371
1374
1372 Returns None if we are unsure, or errors like ENOENT, EPERM happen.
1375 Returns None if we are unsure, or errors like ENOENT, EPERM happen.
1373 '''
1376 '''
1374 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1377 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1375
1378
1376 def checknlink(testfile):
1379 def checknlink(testfile):
1377 '''check whether hardlink count reporting works properly'''
1380 '''check whether hardlink count reporting works properly'''
1378
1381
1379 # testfile may be open, so we need a separate file for checking to
1382 # testfile may be open, so we need a separate file for checking to
1380 # work around issue2543 (or testfile may get lost on Samba shares)
1383 # work around issue2543 (or testfile may get lost on Samba shares)
1381 f1 = testfile + ".hgtmp1"
1384 f1 = testfile + ".hgtmp1"
1382 if os.path.lexists(f1):
1385 if os.path.lexists(f1):
1383 return False
1386 return False
1384 try:
1387 try:
1385 posixfile(f1, 'w').close()
1388 posixfile(f1, 'w').close()
1386 except IOError:
1389 except IOError:
1387 try:
1390 try:
1388 os.unlink(f1)
1391 os.unlink(f1)
1389 except OSError:
1392 except OSError:
1390 pass
1393 pass
1391 return False
1394 return False
1392
1395
1393 f2 = testfile + ".hgtmp2"
1396 f2 = testfile + ".hgtmp2"
1394 fd = None
1397 fd = None
1395 try:
1398 try:
1396 oslink(f1, f2)
1399 oslink(f1, f2)
1397 # nlinks() may behave differently for files on Windows shares if
1400 # nlinks() may behave differently for files on Windows shares if
1398 # the file is open.
1401 # the file is open.
1399 fd = posixfile(f2)
1402 fd = posixfile(f2)
1400 return nlinks(f2) > 1
1403 return nlinks(f2) > 1
1401 except OSError:
1404 except OSError:
1402 return False
1405 return False
1403 finally:
1406 finally:
1404 if fd is not None:
1407 if fd is not None:
1405 fd.close()
1408 fd.close()
1406 for f in (f1, f2):
1409 for f in (f1, f2):
1407 try:
1410 try:
1408 os.unlink(f)
1411 os.unlink(f)
1409 except OSError:
1412 except OSError:
1410 pass
1413 pass
1411
1414
1412 def endswithsep(path):
1415 def endswithsep(path):
1413 '''Check path ends with os.sep or os.altsep.'''
1416 '''Check path ends with os.sep or os.altsep.'''
1414 return (path.endswith(pycompat.ossep)
1417 return (path.endswith(pycompat.ossep)
1415 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1418 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1416
1419
1417 def splitpath(path):
1420 def splitpath(path):
1418 '''Split path by os.sep.
1421 '''Split path by os.sep.
1419 Note that this function does not use os.altsep because this is
1422 Note that this function does not use os.altsep because this is
1420 an alternative of simple "xxx.split(os.sep)".
1423 an alternative of simple "xxx.split(os.sep)".
1421 It is recommended to use os.path.normpath() before using this
1424 It is recommended to use os.path.normpath() before using this
1422 function if need.'''
1425 function if need.'''
1423 return path.split(pycompat.ossep)
1426 return path.split(pycompat.ossep)
1424
1427
1425 def gui():
1428 def gui():
1426 '''Are we running in a GUI?'''
1429 '''Are we running in a GUI?'''
1427 if pycompat.sysplatform == 'darwin':
1430 if pycompat.sysplatform == 'darwin':
1428 if 'SSH_CONNECTION' in encoding.environ:
1431 if 'SSH_CONNECTION' in encoding.environ:
1429 # handle SSH access to a box where the user is logged in
1432 # handle SSH access to a box where the user is logged in
1430 return False
1433 return False
1431 elif getattr(osutil, 'isgui', None):
1434 elif getattr(osutil, 'isgui', None):
1432 # check if a CoreGraphics session is available
1435 # check if a CoreGraphics session is available
1433 return osutil.isgui()
1436 return osutil.isgui()
1434 else:
1437 else:
1435 # pure build; use a safe default
1438 # pure build; use a safe default
1436 return True
1439 return True
1437 else:
1440 else:
1438 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1441 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1439
1442
1440 def mktempcopy(name, emptyok=False, createmode=None):
1443 def mktempcopy(name, emptyok=False, createmode=None):
1441 """Create a temporary file with the same contents from name
1444 """Create a temporary file with the same contents from name
1442
1445
1443 The permission bits are copied from the original file.
1446 The permission bits are copied from the original file.
1444
1447
1445 If the temporary file is going to be truncated immediately, you
1448 If the temporary file is going to be truncated immediately, you
1446 can use emptyok=True as an optimization.
1449 can use emptyok=True as an optimization.
1447
1450
1448 Returns the name of the temporary file.
1451 Returns the name of the temporary file.
1449 """
1452 """
1450 d, fn = os.path.split(name)
1453 d, fn = os.path.split(name)
1451 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1454 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1452 os.close(fd)
1455 os.close(fd)
1453 # Temporary files are created with mode 0600, which is usually not
1456 # Temporary files are created with mode 0600, which is usually not
1454 # what we want. If the original file already exists, just copy
1457 # what we want. If the original file already exists, just copy
1455 # its mode. Otherwise, manually obey umask.
1458 # its mode. Otherwise, manually obey umask.
1456 copymode(name, temp, createmode)
1459 copymode(name, temp, createmode)
1457 if emptyok:
1460 if emptyok:
1458 return temp
1461 return temp
1459 try:
1462 try:
1460 try:
1463 try:
1461 ifp = posixfile(name, "rb")
1464 ifp = posixfile(name, "rb")
1462 except IOError as inst:
1465 except IOError as inst:
1463 if inst.errno == errno.ENOENT:
1466 if inst.errno == errno.ENOENT:
1464 return temp
1467 return temp
1465 if not getattr(inst, 'filename', None):
1468 if not getattr(inst, 'filename', None):
1466 inst.filename = name
1469 inst.filename = name
1467 raise
1470 raise
1468 ofp = posixfile(temp, "wb")
1471 ofp = posixfile(temp, "wb")
1469 for chunk in filechunkiter(ifp):
1472 for chunk in filechunkiter(ifp):
1470 ofp.write(chunk)
1473 ofp.write(chunk)
1471 ifp.close()
1474 ifp.close()
1472 ofp.close()
1475 ofp.close()
1473 except: # re-raises
1476 except: # re-raises
1474 try: os.unlink(temp)
1477 try: os.unlink(temp)
1475 except OSError: pass
1478 except OSError: pass
1476 raise
1479 raise
1477 return temp
1480 return temp
1478
1481
1479 class filestat(object):
1482 class filestat(object):
1480 """help to exactly detect change of a file
1483 """help to exactly detect change of a file
1481
1484
1482 'stat' attribute is result of 'os.stat()' if specified 'path'
1485 'stat' attribute is result of 'os.stat()' if specified 'path'
1483 exists. Otherwise, it is None. This can avoid preparative
1486 exists. Otherwise, it is None. This can avoid preparative
1484 'exists()' examination on client side of this class.
1487 'exists()' examination on client side of this class.
1485 """
1488 """
1486 def __init__(self, path):
1489 def __init__(self, path):
1487 try:
1490 try:
1488 self.stat = os.stat(path)
1491 self.stat = os.stat(path)
1489 except OSError as err:
1492 except OSError as err:
1490 if err.errno != errno.ENOENT:
1493 if err.errno != errno.ENOENT:
1491 raise
1494 raise
1492 self.stat = None
1495 self.stat = None
1493
1496
1494 __hash__ = object.__hash__
1497 __hash__ = object.__hash__
1495
1498
1496 def __eq__(self, old):
1499 def __eq__(self, old):
1497 try:
1500 try:
1498 # if ambiguity between stat of new and old file is
1501 # if ambiguity between stat of new and old file is
1499 # avoided, comparison of size, ctime and mtime is enough
1502 # avoided, comparison of size, ctime and mtime is enough
1500 # to exactly detect change of a file regardless of platform
1503 # to exactly detect change of a file regardless of platform
1501 return (self.stat.st_size == old.stat.st_size and
1504 return (self.stat.st_size == old.stat.st_size and
1502 self.stat.st_ctime == old.stat.st_ctime and
1505 self.stat.st_ctime == old.stat.st_ctime and
1503 self.stat.st_mtime == old.stat.st_mtime)
1506 self.stat.st_mtime == old.stat.st_mtime)
1504 except AttributeError:
1507 except AttributeError:
1505 return False
1508 return False
1506
1509
1507 def isambig(self, old):
1510 def isambig(self, old):
1508 """Examine whether new (= self) stat is ambiguous against old one
1511 """Examine whether new (= self) stat is ambiguous against old one
1509
1512
1510 "S[N]" below means stat of a file at N-th change:
1513 "S[N]" below means stat of a file at N-th change:
1511
1514
1512 - S[n-1].ctime < S[n].ctime: can detect change of a file
1515 - S[n-1].ctime < S[n].ctime: can detect change of a file
1513 - S[n-1].ctime == S[n].ctime
1516 - S[n-1].ctime == S[n].ctime
1514 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1517 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1515 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1518 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1516 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1519 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1517 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1520 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1518
1521
1519 Case (*2) above means that a file was changed twice or more at
1522 Case (*2) above means that a file was changed twice or more at
1520 same time in sec (= S[n-1].ctime), and comparison of timestamp
1523 same time in sec (= S[n-1].ctime), and comparison of timestamp
1521 is ambiguous.
1524 is ambiguous.
1522
1525
1523 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1526 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1524 timestamp is ambiguous".
1527 timestamp is ambiguous".
1525
1528
1526 But advancing mtime only in case (*2) doesn't work as
1529 But advancing mtime only in case (*2) doesn't work as
1527 expected, because naturally advanced S[n].mtime in case (*1)
1530 expected, because naturally advanced S[n].mtime in case (*1)
1528 might be equal to manually advanced S[n-1 or earlier].mtime.
1531 might be equal to manually advanced S[n-1 or earlier].mtime.
1529
1532
1530 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1533 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1531 treated as ambiguous regardless of mtime, to avoid overlooking
1534 treated as ambiguous regardless of mtime, to avoid overlooking
1532 by confliction between such mtime.
1535 by confliction between such mtime.
1533
1536
1534 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1537 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1535 S[n].mtime", even if size of a file isn't changed.
1538 S[n].mtime", even if size of a file isn't changed.
1536 """
1539 """
1537 try:
1540 try:
1538 return (self.stat.st_ctime == old.stat.st_ctime)
1541 return (self.stat.st_ctime == old.stat.st_ctime)
1539 except AttributeError:
1542 except AttributeError:
1540 return False
1543 return False
1541
1544
1542 def avoidambig(self, path, old):
1545 def avoidambig(self, path, old):
1543 """Change file stat of specified path to avoid ambiguity
1546 """Change file stat of specified path to avoid ambiguity
1544
1547
1545 'old' should be previous filestat of 'path'.
1548 'old' should be previous filestat of 'path'.
1546
1549
1547 This skips avoiding ambiguity, if a process doesn't have
1550 This skips avoiding ambiguity, if a process doesn't have
1548 appropriate privileges for 'path'.
1551 appropriate privileges for 'path'.
1549 """
1552 """
1550 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1553 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1551 try:
1554 try:
1552 os.utime(path, (advanced, advanced))
1555 os.utime(path, (advanced, advanced))
1553 except OSError as inst:
1556 except OSError as inst:
1554 if inst.errno == errno.EPERM:
1557 if inst.errno == errno.EPERM:
1555 # utime() on the file created by another user causes EPERM,
1558 # utime() on the file created by another user causes EPERM,
1556 # if a process doesn't have appropriate privileges
1559 # if a process doesn't have appropriate privileges
1557 return
1560 return
1558 raise
1561 raise
1559
1562
1560 def __ne__(self, other):
1563 def __ne__(self, other):
1561 return not self == other
1564 return not self == other
1562
1565
1563 class atomictempfile(object):
1566 class atomictempfile(object):
1564 '''writable file object that atomically updates a file
1567 '''writable file object that atomically updates a file
1565
1568
1566 All writes will go to a temporary copy of the original file. Call
1569 All writes will go to a temporary copy of the original file. Call
1567 close() when you are done writing, and atomictempfile will rename
1570 close() when you are done writing, and atomictempfile will rename
1568 the temporary copy to the original name, making the changes
1571 the temporary copy to the original name, making the changes
1569 visible. If the object is destroyed without being closed, all your
1572 visible. If the object is destroyed without being closed, all your
1570 writes are discarded.
1573 writes are discarded.
1571
1574
1572 checkambig argument of constructor is used with filestat, and is
1575 checkambig argument of constructor is used with filestat, and is
1573 useful only if target file is guarded by any lock (e.g. repo.lock
1576 useful only if target file is guarded by any lock (e.g. repo.lock
1574 or repo.wlock).
1577 or repo.wlock).
1575 '''
1578 '''
1576 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1579 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1577 self.__name = name # permanent name
1580 self.__name = name # permanent name
1578 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1581 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1579 createmode=createmode)
1582 createmode=createmode)
1580 self._fp = posixfile(self._tempname, mode)
1583 self._fp = posixfile(self._tempname, mode)
1581 self._checkambig = checkambig
1584 self._checkambig = checkambig
1582
1585
1583 # delegated methods
1586 # delegated methods
1584 self.read = self._fp.read
1587 self.read = self._fp.read
1585 self.write = self._fp.write
1588 self.write = self._fp.write
1586 self.seek = self._fp.seek
1589 self.seek = self._fp.seek
1587 self.tell = self._fp.tell
1590 self.tell = self._fp.tell
1588 self.fileno = self._fp.fileno
1591 self.fileno = self._fp.fileno
1589
1592
1590 def close(self):
1593 def close(self):
1591 if not self._fp.closed:
1594 if not self._fp.closed:
1592 self._fp.close()
1595 self._fp.close()
1593 filename = localpath(self.__name)
1596 filename = localpath(self.__name)
1594 oldstat = self._checkambig and filestat(filename)
1597 oldstat = self._checkambig and filestat(filename)
1595 if oldstat and oldstat.stat:
1598 if oldstat and oldstat.stat:
1596 rename(self._tempname, filename)
1599 rename(self._tempname, filename)
1597 newstat = filestat(filename)
1600 newstat = filestat(filename)
1598 if newstat.isambig(oldstat):
1601 if newstat.isambig(oldstat):
1599 # stat of changed file is ambiguous to original one
1602 # stat of changed file is ambiguous to original one
1600 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1603 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1601 os.utime(filename, (advanced, advanced))
1604 os.utime(filename, (advanced, advanced))
1602 else:
1605 else:
1603 rename(self._tempname, filename)
1606 rename(self._tempname, filename)
1604
1607
1605 def discard(self):
1608 def discard(self):
1606 if not self._fp.closed:
1609 if not self._fp.closed:
1607 try:
1610 try:
1608 os.unlink(self._tempname)
1611 os.unlink(self._tempname)
1609 except OSError:
1612 except OSError:
1610 pass
1613 pass
1611 self._fp.close()
1614 self._fp.close()
1612
1615
1613 def __del__(self):
1616 def __del__(self):
1614 if safehasattr(self, '_fp'): # constructor actually did something
1617 if safehasattr(self, '_fp'): # constructor actually did something
1615 self.discard()
1618 self.discard()
1616
1619
1617 def __enter__(self):
1620 def __enter__(self):
1618 return self
1621 return self
1619
1622
1620 def __exit__(self, exctype, excvalue, traceback):
1623 def __exit__(self, exctype, excvalue, traceback):
1621 if exctype is not None:
1624 if exctype is not None:
1622 self.discard()
1625 self.discard()
1623 else:
1626 else:
1624 self.close()
1627 self.close()
1625
1628
1626 def unlinkpath(f, ignoremissing=False):
1629 def unlinkpath(f, ignoremissing=False):
1627 """unlink and remove the directory if it is empty"""
1630 """unlink and remove the directory if it is empty"""
1628 if ignoremissing:
1631 if ignoremissing:
1629 tryunlink(f)
1632 tryunlink(f)
1630 else:
1633 else:
1631 unlink(f)
1634 unlink(f)
1632 # try removing directories that might now be empty
1635 # try removing directories that might now be empty
1633 try:
1636 try:
1634 removedirs(os.path.dirname(f))
1637 removedirs(os.path.dirname(f))
1635 except OSError:
1638 except OSError:
1636 pass
1639 pass
1637
1640
1638 def tryunlink(f):
1641 def tryunlink(f):
1639 """Attempt to remove a file, ignoring ENOENT errors."""
1642 """Attempt to remove a file, ignoring ENOENT errors."""
1640 try:
1643 try:
1641 unlink(f)
1644 unlink(f)
1642 except OSError as e:
1645 except OSError as e:
1643 if e.errno != errno.ENOENT:
1646 if e.errno != errno.ENOENT:
1644 raise
1647 raise
1645
1648
1646 def makedirs(name, mode=None, notindexed=False):
1649 def makedirs(name, mode=None, notindexed=False):
1647 """recursive directory creation with parent mode inheritance
1650 """recursive directory creation with parent mode inheritance
1648
1651
1649 Newly created directories are marked as "not to be indexed by
1652 Newly created directories are marked as "not to be indexed by
1650 the content indexing service", if ``notindexed`` is specified
1653 the content indexing service", if ``notindexed`` is specified
1651 for "write" mode access.
1654 for "write" mode access.
1652 """
1655 """
1653 try:
1656 try:
1654 makedir(name, notindexed)
1657 makedir(name, notindexed)
1655 except OSError as err:
1658 except OSError as err:
1656 if err.errno == errno.EEXIST:
1659 if err.errno == errno.EEXIST:
1657 return
1660 return
1658 if err.errno != errno.ENOENT or not name:
1661 if err.errno != errno.ENOENT or not name:
1659 raise
1662 raise
1660 parent = os.path.dirname(os.path.abspath(name))
1663 parent = os.path.dirname(os.path.abspath(name))
1661 if parent == name:
1664 if parent == name:
1662 raise
1665 raise
1663 makedirs(parent, mode, notindexed)
1666 makedirs(parent, mode, notindexed)
1664 try:
1667 try:
1665 makedir(name, notindexed)
1668 makedir(name, notindexed)
1666 except OSError as err:
1669 except OSError as err:
1667 # Catch EEXIST to handle races
1670 # Catch EEXIST to handle races
1668 if err.errno == errno.EEXIST:
1671 if err.errno == errno.EEXIST:
1669 return
1672 return
1670 raise
1673 raise
1671 if mode is not None:
1674 if mode is not None:
1672 os.chmod(name, mode)
1675 os.chmod(name, mode)
1673
1676
1674 def readfile(path):
1677 def readfile(path):
1675 with open(path, 'rb') as fp:
1678 with open(path, 'rb') as fp:
1676 return fp.read()
1679 return fp.read()
1677
1680
1678 def writefile(path, text):
1681 def writefile(path, text):
1679 with open(path, 'wb') as fp:
1682 with open(path, 'wb') as fp:
1680 fp.write(text)
1683 fp.write(text)
1681
1684
1682 def appendfile(path, text):
1685 def appendfile(path, text):
1683 with open(path, 'ab') as fp:
1686 with open(path, 'ab') as fp:
1684 fp.write(text)
1687 fp.write(text)
1685
1688
1686 class chunkbuffer(object):
1689 class chunkbuffer(object):
1687 """Allow arbitrary sized chunks of data to be efficiently read from an
1690 """Allow arbitrary sized chunks of data to be efficiently read from an
1688 iterator over chunks of arbitrary size."""
1691 iterator over chunks of arbitrary size."""
1689
1692
1690 def __init__(self, in_iter):
1693 def __init__(self, in_iter):
1691 """in_iter is the iterator that's iterating over the input chunks.
1694 """in_iter is the iterator that's iterating over the input chunks.
1692 targetsize is how big a buffer to try to maintain."""
1695 targetsize is how big a buffer to try to maintain."""
1693 def splitbig(chunks):
1696 def splitbig(chunks):
1694 for chunk in chunks:
1697 for chunk in chunks:
1695 if len(chunk) > 2**20:
1698 if len(chunk) > 2**20:
1696 pos = 0
1699 pos = 0
1697 while pos < len(chunk):
1700 while pos < len(chunk):
1698 end = pos + 2 ** 18
1701 end = pos + 2 ** 18
1699 yield chunk[pos:end]
1702 yield chunk[pos:end]
1700 pos = end
1703 pos = end
1701 else:
1704 else:
1702 yield chunk
1705 yield chunk
1703 self.iter = splitbig(in_iter)
1706 self.iter = splitbig(in_iter)
1704 self._queue = collections.deque()
1707 self._queue = collections.deque()
1705 self._chunkoffset = 0
1708 self._chunkoffset = 0
1706
1709
1707 def read(self, l=None):
1710 def read(self, l=None):
1708 """Read L bytes of data from the iterator of chunks of data.
1711 """Read L bytes of data from the iterator of chunks of data.
1709 Returns less than L bytes if the iterator runs dry.
1712 Returns less than L bytes if the iterator runs dry.
1710
1713
1711 If size parameter is omitted, read everything"""
1714 If size parameter is omitted, read everything"""
1712 if l is None:
1715 if l is None:
1713 return ''.join(self.iter)
1716 return ''.join(self.iter)
1714
1717
1715 left = l
1718 left = l
1716 buf = []
1719 buf = []
1717 queue = self._queue
1720 queue = self._queue
1718 while left > 0:
1721 while left > 0:
1719 # refill the queue
1722 # refill the queue
1720 if not queue:
1723 if not queue:
1721 target = 2**18
1724 target = 2**18
1722 for chunk in self.iter:
1725 for chunk in self.iter:
1723 queue.append(chunk)
1726 queue.append(chunk)
1724 target -= len(chunk)
1727 target -= len(chunk)
1725 if target <= 0:
1728 if target <= 0:
1726 break
1729 break
1727 if not queue:
1730 if not queue:
1728 break
1731 break
1729
1732
1730 # The easy way to do this would be to queue.popleft(), modify the
1733 # The easy way to do this would be to queue.popleft(), modify the
1731 # chunk (if necessary), then queue.appendleft(). However, for cases
1734 # chunk (if necessary), then queue.appendleft(). However, for cases
1732 # where we read partial chunk content, this incurs 2 dequeue
1735 # where we read partial chunk content, this incurs 2 dequeue
1733 # mutations and creates a new str for the remaining chunk in the
1736 # mutations and creates a new str for the remaining chunk in the
1734 # queue. Our code below avoids this overhead.
1737 # queue. Our code below avoids this overhead.
1735
1738
1736 chunk = queue[0]
1739 chunk = queue[0]
1737 chunkl = len(chunk)
1740 chunkl = len(chunk)
1738 offset = self._chunkoffset
1741 offset = self._chunkoffset
1739
1742
1740 # Use full chunk.
1743 # Use full chunk.
1741 if offset == 0 and left >= chunkl:
1744 if offset == 0 and left >= chunkl:
1742 left -= chunkl
1745 left -= chunkl
1743 queue.popleft()
1746 queue.popleft()
1744 buf.append(chunk)
1747 buf.append(chunk)
1745 # self._chunkoffset remains at 0.
1748 # self._chunkoffset remains at 0.
1746 continue
1749 continue
1747
1750
1748 chunkremaining = chunkl - offset
1751 chunkremaining = chunkl - offset
1749
1752
1750 # Use all of unconsumed part of chunk.
1753 # Use all of unconsumed part of chunk.
1751 if left >= chunkremaining:
1754 if left >= chunkremaining:
1752 left -= chunkremaining
1755 left -= chunkremaining
1753 queue.popleft()
1756 queue.popleft()
1754 # offset == 0 is enabled by block above, so this won't merely
1757 # offset == 0 is enabled by block above, so this won't merely
1755 # copy via ``chunk[0:]``.
1758 # copy via ``chunk[0:]``.
1756 buf.append(chunk[offset:])
1759 buf.append(chunk[offset:])
1757 self._chunkoffset = 0
1760 self._chunkoffset = 0
1758
1761
1759 # Partial chunk needed.
1762 # Partial chunk needed.
1760 else:
1763 else:
1761 buf.append(chunk[offset:offset + left])
1764 buf.append(chunk[offset:offset + left])
1762 self._chunkoffset += left
1765 self._chunkoffset += left
1763 left -= chunkremaining
1766 left -= chunkremaining
1764
1767
1765 return ''.join(buf)
1768 return ''.join(buf)
1766
1769
1767 def filechunkiter(f, size=131072, limit=None):
1770 def filechunkiter(f, size=131072, limit=None):
1768 """Create a generator that produces the data in the file size
1771 """Create a generator that produces the data in the file size
1769 (default 131072) bytes at a time, up to optional limit (default is
1772 (default 131072) bytes at a time, up to optional limit (default is
1770 to read all data). Chunks may be less than size bytes if the
1773 to read all data). Chunks may be less than size bytes if the
1771 chunk is the last chunk in the file, or the file is a socket or
1774 chunk is the last chunk in the file, or the file is a socket or
1772 some other type of file that sometimes reads less data than is
1775 some other type of file that sometimes reads less data than is
1773 requested."""
1776 requested."""
1774 assert size >= 0
1777 assert size >= 0
1775 assert limit is None or limit >= 0
1778 assert limit is None or limit >= 0
1776 while True:
1779 while True:
1777 if limit is None:
1780 if limit is None:
1778 nbytes = size
1781 nbytes = size
1779 else:
1782 else:
1780 nbytes = min(limit, size)
1783 nbytes = min(limit, size)
1781 s = nbytes and f.read(nbytes)
1784 s = nbytes and f.read(nbytes)
1782 if not s:
1785 if not s:
1783 break
1786 break
1784 if limit:
1787 if limit:
1785 limit -= len(s)
1788 limit -= len(s)
1786 yield s
1789 yield s
1787
1790
1788 def makedate(timestamp=None):
1791 def makedate(timestamp=None):
1789 '''Return a unix timestamp (or the current time) as a (unixtime,
1792 '''Return a unix timestamp (or the current time) as a (unixtime,
1790 offset) tuple based off the local timezone.'''
1793 offset) tuple based off the local timezone.'''
1791 if timestamp is None:
1794 if timestamp is None:
1792 timestamp = time.time()
1795 timestamp = time.time()
1793 if timestamp < 0:
1796 if timestamp < 0:
1794 hint = _("check your clock")
1797 hint = _("check your clock")
1795 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1798 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1796 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1799 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1797 datetime.datetime.fromtimestamp(timestamp))
1800 datetime.datetime.fromtimestamp(timestamp))
1798 tz = delta.days * 86400 + delta.seconds
1801 tz = delta.days * 86400 + delta.seconds
1799 return timestamp, tz
1802 return timestamp, tz
1800
1803
1801 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1804 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1802 """represent a (unixtime, offset) tuple as a localized time.
1805 """represent a (unixtime, offset) tuple as a localized time.
1803 unixtime is seconds since the epoch, and offset is the time zone's
1806 unixtime is seconds since the epoch, and offset is the time zone's
1804 number of seconds away from UTC.
1807 number of seconds away from UTC.
1805
1808
1806 >>> datestr((0, 0))
1809 >>> datestr((0, 0))
1807 'Thu Jan 01 00:00:00 1970 +0000'
1810 'Thu Jan 01 00:00:00 1970 +0000'
1808 >>> datestr((42, 0))
1811 >>> datestr((42, 0))
1809 'Thu Jan 01 00:00:42 1970 +0000'
1812 'Thu Jan 01 00:00:42 1970 +0000'
1810 >>> datestr((-42, 0))
1813 >>> datestr((-42, 0))
1811 'Wed Dec 31 23:59:18 1969 +0000'
1814 'Wed Dec 31 23:59:18 1969 +0000'
1812 >>> datestr((0x7fffffff, 0))
1815 >>> datestr((0x7fffffff, 0))
1813 'Tue Jan 19 03:14:07 2038 +0000'
1816 'Tue Jan 19 03:14:07 2038 +0000'
1814 >>> datestr((-0x80000000, 0))
1817 >>> datestr((-0x80000000, 0))
1815 'Fri Dec 13 20:45:52 1901 +0000'
1818 'Fri Dec 13 20:45:52 1901 +0000'
1816 """
1819 """
1817 t, tz = date or makedate()
1820 t, tz = date or makedate()
1818 if "%1" in format or "%2" in format or "%z" in format:
1821 if "%1" in format or "%2" in format or "%z" in format:
1819 sign = (tz > 0) and "-" or "+"
1822 sign = (tz > 0) and "-" or "+"
1820 minutes = abs(tz) // 60
1823 minutes = abs(tz) // 60
1821 q, r = divmod(minutes, 60)
1824 q, r = divmod(minutes, 60)
1822 format = format.replace("%z", "%1%2")
1825 format = format.replace("%z", "%1%2")
1823 format = format.replace("%1", "%c%02d" % (sign, q))
1826 format = format.replace("%1", "%c%02d" % (sign, q))
1824 format = format.replace("%2", "%02d" % r)
1827 format = format.replace("%2", "%02d" % r)
1825 d = t - tz
1828 d = t - tz
1826 if d > 0x7fffffff:
1829 if d > 0x7fffffff:
1827 d = 0x7fffffff
1830 d = 0x7fffffff
1828 elif d < -0x80000000:
1831 elif d < -0x80000000:
1829 d = -0x80000000
1832 d = -0x80000000
1830 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1833 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1831 # because they use the gmtime() system call which is buggy on Windows
1834 # because they use the gmtime() system call which is buggy on Windows
1832 # for negative values.
1835 # for negative values.
1833 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1836 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1834 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1837 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1835 return s
1838 return s
1836
1839
1837 def shortdate(date=None):
1840 def shortdate(date=None):
1838 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1841 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1839 return datestr(date, format='%Y-%m-%d')
1842 return datestr(date, format='%Y-%m-%d')
1840
1843
1841 def parsetimezone(s):
1844 def parsetimezone(s):
1842 """find a trailing timezone, if any, in string, and return a
1845 """find a trailing timezone, if any, in string, and return a
1843 (offset, remainder) pair"""
1846 (offset, remainder) pair"""
1844
1847
1845 if s.endswith("GMT") or s.endswith("UTC"):
1848 if s.endswith("GMT") or s.endswith("UTC"):
1846 return 0, s[:-3].rstrip()
1849 return 0, s[:-3].rstrip()
1847
1850
1848 # Unix-style timezones [+-]hhmm
1851 # Unix-style timezones [+-]hhmm
1849 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1852 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1850 sign = (s[-5] == "+") and 1 or -1
1853 sign = (s[-5] == "+") and 1 or -1
1851 hours = int(s[-4:-2])
1854 hours = int(s[-4:-2])
1852 minutes = int(s[-2:])
1855 minutes = int(s[-2:])
1853 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1856 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1854
1857
1855 # ISO8601 trailing Z
1858 # ISO8601 trailing Z
1856 if s.endswith("Z") and s[-2:-1].isdigit():
1859 if s.endswith("Z") and s[-2:-1].isdigit():
1857 return 0, s[:-1]
1860 return 0, s[:-1]
1858
1861
1859 # ISO8601-style [+-]hh:mm
1862 # ISO8601-style [+-]hh:mm
1860 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1863 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1861 s[-5:-3].isdigit() and s[-2:].isdigit()):
1864 s[-5:-3].isdigit() and s[-2:].isdigit()):
1862 sign = (s[-6] == "+") and 1 or -1
1865 sign = (s[-6] == "+") and 1 or -1
1863 hours = int(s[-5:-3])
1866 hours = int(s[-5:-3])
1864 minutes = int(s[-2:])
1867 minutes = int(s[-2:])
1865 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1868 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1866
1869
1867 return None, s
1870 return None, s
1868
1871
1869 def strdate(string, format, defaults=None):
1872 def strdate(string, format, defaults=None):
1870 """parse a localized time string and return a (unixtime, offset) tuple.
1873 """parse a localized time string and return a (unixtime, offset) tuple.
1871 if the string cannot be parsed, ValueError is raised."""
1874 if the string cannot be parsed, ValueError is raised."""
1872 if defaults is None:
1875 if defaults is None:
1873 defaults = {}
1876 defaults = {}
1874
1877
1875 # NOTE: unixtime = localunixtime + offset
1878 # NOTE: unixtime = localunixtime + offset
1876 offset, date = parsetimezone(string)
1879 offset, date = parsetimezone(string)
1877
1880
1878 # add missing elements from defaults
1881 # add missing elements from defaults
1879 usenow = False # default to using biased defaults
1882 usenow = False # default to using biased defaults
1880 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1883 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1881 found = [True for p in part if ("%"+p) in format]
1884 found = [True for p in part if ("%"+p) in format]
1882 if not found:
1885 if not found:
1883 date += "@" + defaults[part][usenow]
1886 date += "@" + defaults[part][usenow]
1884 format += "@%" + part[0]
1887 format += "@%" + part[0]
1885 else:
1888 else:
1886 # We've found a specific time element, less specific time
1889 # We've found a specific time element, less specific time
1887 # elements are relative to today
1890 # elements are relative to today
1888 usenow = True
1891 usenow = True
1889
1892
1890 timetuple = time.strptime(date, format)
1893 timetuple = time.strptime(date, format)
1891 localunixtime = int(calendar.timegm(timetuple))
1894 localunixtime = int(calendar.timegm(timetuple))
1892 if offset is None:
1895 if offset is None:
1893 # local timezone
1896 # local timezone
1894 unixtime = int(time.mktime(timetuple))
1897 unixtime = int(time.mktime(timetuple))
1895 offset = unixtime - localunixtime
1898 offset = unixtime - localunixtime
1896 else:
1899 else:
1897 unixtime = localunixtime + offset
1900 unixtime = localunixtime + offset
1898 return unixtime, offset
1901 return unixtime, offset
1899
1902
1900 def parsedate(date, formats=None, bias=None):
1903 def parsedate(date, formats=None, bias=None):
1901 """parse a localized date/time and return a (unixtime, offset) tuple.
1904 """parse a localized date/time and return a (unixtime, offset) tuple.
1902
1905
1903 The date may be a "unixtime offset" string or in one of the specified
1906 The date may be a "unixtime offset" string or in one of the specified
1904 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1907 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1905
1908
1906 >>> parsedate(' today ') == parsedate(\
1909 >>> parsedate(' today ') == parsedate(\
1907 datetime.date.today().strftime('%b %d'))
1910 datetime.date.today().strftime('%b %d'))
1908 True
1911 True
1909 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1912 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1910 datetime.timedelta(days=1)\
1913 datetime.timedelta(days=1)\
1911 ).strftime('%b %d'))
1914 ).strftime('%b %d'))
1912 True
1915 True
1913 >>> now, tz = makedate()
1916 >>> now, tz = makedate()
1914 >>> strnow, strtz = parsedate('now')
1917 >>> strnow, strtz = parsedate('now')
1915 >>> (strnow - now) < 1
1918 >>> (strnow - now) < 1
1916 True
1919 True
1917 >>> tz == strtz
1920 >>> tz == strtz
1918 True
1921 True
1919 """
1922 """
1920 if bias is None:
1923 if bias is None:
1921 bias = {}
1924 bias = {}
1922 if not date:
1925 if not date:
1923 return 0, 0
1926 return 0, 0
1924 if isinstance(date, tuple) and len(date) == 2:
1927 if isinstance(date, tuple) and len(date) == 2:
1925 return date
1928 return date
1926 if not formats:
1929 if not formats:
1927 formats = defaultdateformats
1930 formats = defaultdateformats
1928 date = date.strip()
1931 date = date.strip()
1929
1932
1930 if date == 'now' or date == _('now'):
1933 if date == 'now' or date == _('now'):
1931 return makedate()
1934 return makedate()
1932 if date == 'today' or date == _('today'):
1935 if date == 'today' or date == _('today'):
1933 date = datetime.date.today().strftime('%b %d')
1936 date = datetime.date.today().strftime('%b %d')
1934 elif date == 'yesterday' or date == _('yesterday'):
1937 elif date == 'yesterday' or date == _('yesterday'):
1935 date = (datetime.date.today() -
1938 date = (datetime.date.today() -
1936 datetime.timedelta(days=1)).strftime('%b %d')
1939 datetime.timedelta(days=1)).strftime('%b %d')
1937
1940
1938 try:
1941 try:
1939 when, offset = map(int, date.split(' '))
1942 when, offset = map(int, date.split(' '))
1940 except ValueError:
1943 except ValueError:
1941 # fill out defaults
1944 # fill out defaults
1942 now = makedate()
1945 now = makedate()
1943 defaults = {}
1946 defaults = {}
1944 for part in ("d", "mb", "yY", "HI", "M", "S"):
1947 for part in ("d", "mb", "yY", "HI", "M", "S"):
1945 # this piece is for rounding the specific end of unknowns
1948 # this piece is for rounding the specific end of unknowns
1946 b = bias.get(part)
1949 b = bias.get(part)
1947 if b is None:
1950 if b is None:
1948 if part[0] in "HMS":
1951 if part[0] in "HMS":
1949 b = "00"
1952 b = "00"
1950 else:
1953 else:
1951 b = "0"
1954 b = "0"
1952
1955
1953 # this piece is for matching the generic end to today's date
1956 # this piece is for matching the generic end to today's date
1954 n = datestr(now, "%" + part[0])
1957 n = datestr(now, "%" + part[0])
1955
1958
1956 defaults[part] = (b, n)
1959 defaults[part] = (b, n)
1957
1960
1958 for format in formats:
1961 for format in formats:
1959 try:
1962 try:
1960 when, offset = strdate(date, format, defaults)
1963 when, offset = strdate(date, format, defaults)
1961 except (ValueError, OverflowError):
1964 except (ValueError, OverflowError):
1962 pass
1965 pass
1963 else:
1966 else:
1964 break
1967 break
1965 else:
1968 else:
1966 raise Abort(_('invalid date: %r') % date)
1969 raise Abort(_('invalid date: %r') % date)
1967 # validate explicit (probably user-specified) date and
1970 # validate explicit (probably user-specified) date and
1968 # time zone offset. values must fit in signed 32 bits for
1971 # time zone offset. values must fit in signed 32 bits for
1969 # current 32-bit linux runtimes. timezones go from UTC-12
1972 # current 32-bit linux runtimes. timezones go from UTC-12
1970 # to UTC+14
1973 # to UTC+14
1971 if when < -0x80000000 or when > 0x7fffffff:
1974 if when < -0x80000000 or when > 0x7fffffff:
1972 raise Abort(_('date exceeds 32 bits: %d') % when)
1975 raise Abort(_('date exceeds 32 bits: %d') % when)
1973 if offset < -50400 or offset > 43200:
1976 if offset < -50400 or offset > 43200:
1974 raise Abort(_('impossible time zone offset: %d') % offset)
1977 raise Abort(_('impossible time zone offset: %d') % offset)
1975 return when, offset
1978 return when, offset
1976
1979
1977 def matchdate(date):
1980 def matchdate(date):
1978 """Return a function that matches a given date match specifier
1981 """Return a function that matches a given date match specifier
1979
1982
1980 Formats include:
1983 Formats include:
1981
1984
1982 '{date}' match a given date to the accuracy provided
1985 '{date}' match a given date to the accuracy provided
1983
1986
1984 '<{date}' on or before a given date
1987 '<{date}' on or before a given date
1985
1988
1986 '>{date}' on or after a given date
1989 '>{date}' on or after a given date
1987
1990
1988 >>> p1 = parsedate("10:29:59")
1991 >>> p1 = parsedate("10:29:59")
1989 >>> p2 = parsedate("10:30:00")
1992 >>> p2 = parsedate("10:30:00")
1990 >>> p3 = parsedate("10:30:59")
1993 >>> p3 = parsedate("10:30:59")
1991 >>> p4 = parsedate("10:31:00")
1994 >>> p4 = parsedate("10:31:00")
1992 >>> p5 = parsedate("Sep 15 10:30:00 1999")
1995 >>> p5 = parsedate("Sep 15 10:30:00 1999")
1993 >>> f = matchdate("10:30")
1996 >>> f = matchdate("10:30")
1994 >>> f(p1[0])
1997 >>> f(p1[0])
1995 False
1998 False
1996 >>> f(p2[0])
1999 >>> f(p2[0])
1997 True
2000 True
1998 >>> f(p3[0])
2001 >>> f(p3[0])
1999 True
2002 True
2000 >>> f(p4[0])
2003 >>> f(p4[0])
2001 False
2004 False
2002 >>> f(p5[0])
2005 >>> f(p5[0])
2003 False
2006 False
2004 """
2007 """
2005
2008
2006 def lower(date):
2009 def lower(date):
2007 d = {'mb': "1", 'd': "1"}
2010 d = {'mb': "1", 'd': "1"}
2008 return parsedate(date, extendeddateformats, d)[0]
2011 return parsedate(date, extendeddateformats, d)[0]
2009
2012
2010 def upper(date):
2013 def upper(date):
2011 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2014 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2012 for days in ("31", "30", "29"):
2015 for days in ("31", "30", "29"):
2013 try:
2016 try:
2014 d["d"] = days
2017 d["d"] = days
2015 return parsedate(date, extendeddateformats, d)[0]
2018 return parsedate(date, extendeddateformats, d)[0]
2016 except Abort:
2019 except Abort:
2017 pass
2020 pass
2018 d["d"] = "28"
2021 d["d"] = "28"
2019 return parsedate(date, extendeddateformats, d)[0]
2022 return parsedate(date, extendeddateformats, d)[0]
2020
2023
2021 date = date.strip()
2024 date = date.strip()
2022
2025
2023 if not date:
2026 if not date:
2024 raise Abort(_("dates cannot consist entirely of whitespace"))
2027 raise Abort(_("dates cannot consist entirely of whitespace"))
2025 elif date[0] == "<":
2028 elif date[0] == "<":
2026 if not date[1:]:
2029 if not date[1:]:
2027 raise Abort(_("invalid day spec, use '<DATE'"))
2030 raise Abort(_("invalid day spec, use '<DATE'"))
2028 when = upper(date[1:])
2031 when = upper(date[1:])
2029 return lambda x: x <= when
2032 return lambda x: x <= when
2030 elif date[0] == ">":
2033 elif date[0] == ">":
2031 if not date[1:]:
2034 if not date[1:]:
2032 raise Abort(_("invalid day spec, use '>DATE'"))
2035 raise Abort(_("invalid day spec, use '>DATE'"))
2033 when = lower(date[1:])
2036 when = lower(date[1:])
2034 return lambda x: x >= when
2037 return lambda x: x >= when
2035 elif date[0] == "-":
2038 elif date[0] == "-":
2036 try:
2039 try:
2037 days = int(date[1:])
2040 days = int(date[1:])
2038 except ValueError:
2041 except ValueError:
2039 raise Abort(_("invalid day spec: %s") % date[1:])
2042 raise Abort(_("invalid day spec: %s") % date[1:])
2040 if days < 0:
2043 if days < 0:
2041 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2044 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2042 % date[1:])
2045 % date[1:])
2043 when = makedate()[0] - days * 3600 * 24
2046 when = makedate()[0] - days * 3600 * 24
2044 return lambda x: x >= when
2047 return lambda x: x >= when
2045 elif " to " in date:
2048 elif " to " in date:
2046 a, b = date.split(" to ")
2049 a, b = date.split(" to ")
2047 start, stop = lower(a), upper(b)
2050 start, stop = lower(a), upper(b)
2048 return lambda x: x >= start and x <= stop
2051 return lambda x: x >= start and x <= stop
2049 else:
2052 else:
2050 start, stop = lower(date), upper(date)
2053 start, stop = lower(date), upper(date)
2051 return lambda x: x >= start and x <= stop
2054 return lambda x: x >= start and x <= stop
2052
2055
2053 def stringmatcher(pattern, casesensitive=True):
2056 def stringmatcher(pattern, casesensitive=True):
2054 """
2057 """
2055 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2058 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2056 returns the matcher name, pattern, and matcher function.
2059 returns the matcher name, pattern, and matcher function.
2057 missing or unknown prefixes are treated as literal matches.
2060 missing or unknown prefixes are treated as literal matches.
2058
2061
2059 helper for tests:
2062 helper for tests:
2060 >>> def test(pattern, *tests):
2063 >>> def test(pattern, *tests):
2061 ... kind, pattern, matcher = stringmatcher(pattern)
2064 ... kind, pattern, matcher = stringmatcher(pattern)
2062 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2065 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2063 >>> def itest(pattern, *tests):
2066 >>> def itest(pattern, *tests):
2064 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2067 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2065 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2068 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2066
2069
2067 exact matching (no prefix):
2070 exact matching (no prefix):
2068 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2071 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2069 ('literal', 'abcdefg', [False, False, True])
2072 ('literal', 'abcdefg', [False, False, True])
2070
2073
2071 regex matching ('re:' prefix)
2074 regex matching ('re:' prefix)
2072 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2075 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2073 ('re', 'a.+b', [False, False, True])
2076 ('re', 'a.+b', [False, False, True])
2074
2077
2075 force exact matches ('literal:' prefix)
2078 force exact matches ('literal:' prefix)
2076 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2079 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2077 ('literal', 're:foobar', [False, True])
2080 ('literal', 're:foobar', [False, True])
2078
2081
2079 unknown prefixes are ignored and treated as literals
2082 unknown prefixes are ignored and treated as literals
2080 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2083 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2081 ('literal', 'foo:bar', [False, False, True])
2084 ('literal', 'foo:bar', [False, False, True])
2082
2085
2083 case insensitive regex matches
2086 case insensitive regex matches
2084 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2087 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2085 ('re', 'A.+b', [False, False, True])
2088 ('re', 'A.+b', [False, False, True])
2086
2089
2087 case insensitive literal matches
2090 case insensitive literal matches
2088 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2091 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2089 ('literal', 'ABCDEFG', [False, False, True])
2092 ('literal', 'ABCDEFG', [False, False, True])
2090 """
2093 """
2091 if pattern.startswith('re:'):
2094 if pattern.startswith('re:'):
2092 pattern = pattern[3:]
2095 pattern = pattern[3:]
2093 try:
2096 try:
2094 flags = 0
2097 flags = 0
2095 if not casesensitive:
2098 if not casesensitive:
2096 flags = remod.I
2099 flags = remod.I
2097 regex = remod.compile(pattern, flags)
2100 regex = remod.compile(pattern, flags)
2098 except remod.error as e:
2101 except remod.error as e:
2099 raise error.ParseError(_('invalid regular expression: %s')
2102 raise error.ParseError(_('invalid regular expression: %s')
2100 % e)
2103 % e)
2101 return 're', pattern, regex.search
2104 return 're', pattern, regex.search
2102 elif pattern.startswith('literal:'):
2105 elif pattern.startswith('literal:'):
2103 pattern = pattern[8:]
2106 pattern = pattern[8:]
2104
2107
2105 match = pattern.__eq__
2108 match = pattern.__eq__
2106
2109
2107 if not casesensitive:
2110 if not casesensitive:
2108 ipat = encoding.lower(pattern)
2111 ipat = encoding.lower(pattern)
2109 match = lambda s: ipat == encoding.lower(s)
2112 match = lambda s: ipat == encoding.lower(s)
2110 return 'literal', pattern, match
2113 return 'literal', pattern, match
2111
2114
2112 def shortuser(user):
2115 def shortuser(user):
2113 """Return a short representation of a user name or email address."""
2116 """Return a short representation of a user name or email address."""
2114 f = user.find('@')
2117 f = user.find('@')
2115 if f >= 0:
2118 if f >= 0:
2116 user = user[:f]
2119 user = user[:f]
2117 f = user.find('<')
2120 f = user.find('<')
2118 if f >= 0:
2121 if f >= 0:
2119 user = user[f + 1:]
2122 user = user[f + 1:]
2120 f = user.find(' ')
2123 f = user.find(' ')
2121 if f >= 0:
2124 if f >= 0:
2122 user = user[:f]
2125 user = user[:f]
2123 f = user.find('.')
2126 f = user.find('.')
2124 if f >= 0:
2127 if f >= 0:
2125 user = user[:f]
2128 user = user[:f]
2126 return user
2129 return user
2127
2130
2128 def emailuser(user):
2131 def emailuser(user):
2129 """Return the user portion of an email address."""
2132 """Return the user portion of an email address."""
2130 f = user.find('@')
2133 f = user.find('@')
2131 if f >= 0:
2134 if f >= 0:
2132 user = user[:f]
2135 user = user[:f]
2133 f = user.find('<')
2136 f = user.find('<')
2134 if f >= 0:
2137 if f >= 0:
2135 user = user[f + 1:]
2138 user = user[f + 1:]
2136 return user
2139 return user
2137
2140
2138 def email(author):
2141 def email(author):
2139 '''get email of author.'''
2142 '''get email of author.'''
2140 r = author.find('>')
2143 r = author.find('>')
2141 if r == -1:
2144 if r == -1:
2142 r = None
2145 r = None
2143 return author[author.find('<') + 1:r]
2146 return author[author.find('<') + 1:r]
2144
2147
2145 def ellipsis(text, maxlength=400):
2148 def ellipsis(text, maxlength=400):
2146 """Trim string to at most maxlength (default: 400) columns in display."""
2149 """Trim string to at most maxlength (default: 400) columns in display."""
2147 return encoding.trim(text, maxlength, ellipsis='...')
2150 return encoding.trim(text, maxlength, ellipsis='...')
2148
2151
2149 def unitcountfn(*unittable):
2152 def unitcountfn(*unittable):
2150 '''return a function that renders a readable count of some quantity'''
2153 '''return a function that renders a readable count of some quantity'''
2151
2154
2152 def go(count):
2155 def go(count):
2153 for multiplier, divisor, format in unittable:
2156 for multiplier, divisor, format in unittable:
2154 if count >= divisor * multiplier:
2157 if count >= divisor * multiplier:
2155 return format % (count / float(divisor))
2158 return format % (count / float(divisor))
2156 return unittable[-1][2] % count
2159 return unittable[-1][2] % count
2157
2160
2158 return go
2161 return go
2159
2162
2160 bytecount = unitcountfn(
2163 bytecount = unitcountfn(
2161 (100, 1 << 30, _('%.0f GB')),
2164 (100, 1 << 30, _('%.0f GB')),
2162 (10, 1 << 30, _('%.1f GB')),
2165 (10, 1 << 30, _('%.1f GB')),
2163 (1, 1 << 30, _('%.2f GB')),
2166 (1, 1 << 30, _('%.2f GB')),
2164 (100, 1 << 20, _('%.0f MB')),
2167 (100, 1 << 20, _('%.0f MB')),
2165 (10, 1 << 20, _('%.1f MB')),
2168 (10, 1 << 20, _('%.1f MB')),
2166 (1, 1 << 20, _('%.2f MB')),
2169 (1, 1 << 20, _('%.2f MB')),
2167 (100, 1 << 10, _('%.0f KB')),
2170 (100, 1 << 10, _('%.0f KB')),
2168 (10, 1 << 10, _('%.1f KB')),
2171 (10, 1 << 10, _('%.1f KB')),
2169 (1, 1 << 10, _('%.2f KB')),
2172 (1, 1 << 10, _('%.2f KB')),
2170 (1, 1, _('%.0f bytes')),
2173 (1, 1, _('%.0f bytes')),
2171 )
2174 )
2172
2175
2173 def escapestr(s):
2176 def escapestr(s):
2174 # call underlying function of s.encode('string_escape') directly for
2177 # call underlying function of s.encode('string_escape') directly for
2175 # Python 3 compatibility
2178 # Python 3 compatibility
2176 return codecs.escape_encode(s)[0]
2179 return codecs.escape_encode(s)[0]
2177
2180
2178 def unescapestr(s):
2181 def unescapestr(s):
2179 return codecs.escape_decode(s)[0]
2182 return codecs.escape_decode(s)[0]
2180
2183
2181 def uirepr(s):
2184 def uirepr(s):
2182 # Avoid double backslash in Windows path repr()
2185 # Avoid double backslash in Windows path repr()
2183 return repr(s).replace('\\\\', '\\')
2186 return repr(s).replace('\\\\', '\\')
2184
2187
2185 # delay import of textwrap
2188 # delay import of textwrap
2186 def MBTextWrapper(**kwargs):
2189 def MBTextWrapper(**kwargs):
2187 class tw(textwrap.TextWrapper):
2190 class tw(textwrap.TextWrapper):
2188 """
2191 """
2189 Extend TextWrapper for width-awareness.
2192 Extend TextWrapper for width-awareness.
2190
2193
2191 Neither number of 'bytes' in any encoding nor 'characters' is
2194 Neither number of 'bytes' in any encoding nor 'characters' is
2192 appropriate to calculate terminal columns for specified string.
2195 appropriate to calculate terminal columns for specified string.
2193
2196
2194 Original TextWrapper implementation uses built-in 'len()' directly,
2197 Original TextWrapper implementation uses built-in 'len()' directly,
2195 so overriding is needed to use width information of each characters.
2198 so overriding is needed to use width information of each characters.
2196
2199
2197 In addition, characters classified into 'ambiguous' width are
2200 In addition, characters classified into 'ambiguous' width are
2198 treated as wide in East Asian area, but as narrow in other.
2201 treated as wide in East Asian area, but as narrow in other.
2199
2202
2200 This requires use decision to determine width of such characters.
2203 This requires use decision to determine width of such characters.
2201 """
2204 """
2202 def _cutdown(self, ucstr, space_left):
2205 def _cutdown(self, ucstr, space_left):
2203 l = 0
2206 l = 0
2204 colwidth = encoding.ucolwidth
2207 colwidth = encoding.ucolwidth
2205 for i in xrange(len(ucstr)):
2208 for i in xrange(len(ucstr)):
2206 l += colwidth(ucstr[i])
2209 l += colwidth(ucstr[i])
2207 if space_left < l:
2210 if space_left < l:
2208 return (ucstr[:i], ucstr[i:])
2211 return (ucstr[:i], ucstr[i:])
2209 return ucstr, ''
2212 return ucstr, ''
2210
2213
2211 # overriding of base class
2214 # overriding of base class
2212 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2215 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2213 space_left = max(width - cur_len, 1)
2216 space_left = max(width - cur_len, 1)
2214
2217
2215 if self.break_long_words:
2218 if self.break_long_words:
2216 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2219 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2217 cur_line.append(cut)
2220 cur_line.append(cut)
2218 reversed_chunks[-1] = res
2221 reversed_chunks[-1] = res
2219 elif not cur_line:
2222 elif not cur_line:
2220 cur_line.append(reversed_chunks.pop())
2223 cur_line.append(reversed_chunks.pop())
2221
2224
2222 # this overriding code is imported from TextWrapper of Python 2.6
2225 # this overriding code is imported from TextWrapper of Python 2.6
2223 # to calculate columns of string by 'encoding.ucolwidth()'
2226 # to calculate columns of string by 'encoding.ucolwidth()'
2224 def _wrap_chunks(self, chunks):
2227 def _wrap_chunks(self, chunks):
2225 colwidth = encoding.ucolwidth
2228 colwidth = encoding.ucolwidth
2226
2229
2227 lines = []
2230 lines = []
2228 if self.width <= 0:
2231 if self.width <= 0:
2229 raise ValueError("invalid width %r (must be > 0)" % self.width)
2232 raise ValueError("invalid width %r (must be > 0)" % self.width)
2230
2233
2231 # Arrange in reverse order so items can be efficiently popped
2234 # Arrange in reverse order so items can be efficiently popped
2232 # from a stack of chucks.
2235 # from a stack of chucks.
2233 chunks.reverse()
2236 chunks.reverse()
2234
2237
2235 while chunks:
2238 while chunks:
2236
2239
2237 # Start the list of chunks that will make up the current line.
2240 # Start the list of chunks that will make up the current line.
2238 # cur_len is just the length of all the chunks in cur_line.
2241 # cur_len is just the length of all the chunks in cur_line.
2239 cur_line = []
2242 cur_line = []
2240 cur_len = 0
2243 cur_len = 0
2241
2244
2242 # Figure out which static string will prefix this line.
2245 # Figure out which static string will prefix this line.
2243 if lines:
2246 if lines:
2244 indent = self.subsequent_indent
2247 indent = self.subsequent_indent
2245 else:
2248 else:
2246 indent = self.initial_indent
2249 indent = self.initial_indent
2247
2250
2248 # Maximum width for this line.
2251 # Maximum width for this line.
2249 width = self.width - len(indent)
2252 width = self.width - len(indent)
2250
2253
2251 # First chunk on line is whitespace -- drop it, unless this
2254 # First chunk on line is whitespace -- drop it, unless this
2252 # is the very beginning of the text (i.e. no lines started yet).
2255 # is the very beginning of the text (i.e. no lines started yet).
2253 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2256 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2254 del chunks[-1]
2257 del chunks[-1]
2255
2258
2256 while chunks:
2259 while chunks:
2257 l = colwidth(chunks[-1])
2260 l = colwidth(chunks[-1])
2258
2261
2259 # Can at least squeeze this chunk onto the current line.
2262 # Can at least squeeze this chunk onto the current line.
2260 if cur_len + l <= width:
2263 if cur_len + l <= width:
2261 cur_line.append(chunks.pop())
2264 cur_line.append(chunks.pop())
2262 cur_len += l
2265 cur_len += l
2263
2266
2264 # Nope, this line is full.
2267 # Nope, this line is full.
2265 else:
2268 else:
2266 break
2269 break
2267
2270
2268 # The current line is full, and the next chunk is too big to
2271 # The current line is full, and the next chunk is too big to
2269 # fit on *any* line (not just this one).
2272 # fit on *any* line (not just this one).
2270 if chunks and colwidth(chunks[-1]) > width:
2273 if chunks and colwidth(chunks[-1]) > width:
2271 self._handle_long_word(chunks, cur_line, cur_len, width)
2274 self._handle_long_word(chunks, cur_line, cur_len, width)
2272
2275
2273 # If the last chunk on this line is all whitespace, drop it.
2276 # If the last chunk on this line is all whitespace, drop it.
2274 if (self.drop_whitespace and
2277 if (self.drop_whitespace and
2275 cur_line and cur_line[-1].strip() == ''):
2278 cur_line and cur_line[-1].strip() == ''):
2276 del cur_line[-1]
2279 del cur_line[-1]
2277
2280
2278 # Convert current line back to a string and store it in list
2281 # Convert current line back to a string and store it in list
2279 # of all lines (return value).
2282 # of all lines (return value).
2280 if cur_line:
2283 if cur_line:
2281 lines.append(indent + ''.join(cur_line))
2284 lines.append(indent + ''.join(cur_line))
2282
2285
2283 return lines
2286 return lines
2284
2287
2285 global MBTextWrapper
2288 global MBTextWrapper
2286 MBTextWrapper = tw
2289 MBTextWrapper = tw
2287 return tw(**kwargs)
2290 return tw(**kwargs)
2288
2291
2289 def wrap(line, width, initindent='', hangindent=''):
2292 def wrap(line, width, initindent='', hangindent=''):
2290 maxindent = max(len(hangindent), len(initindent))
2293 maxindent = max(len(hangindent), len(initindent))
2291 if width <= maxindent:
2294 if width <= maxindent:
2292 # adjust for weird terminal size
2295 # adjust for weird terminal size
2293 width = max(78, maxindent + 1)
2296 width = max(78, maxindent + 1)
2294 line = line.decode(pycompat.sysstr(encoding.encoding),
2297 line = line.decode(pycompat.sysstr(encoding.encoding),
2295 pycompat.sysstr(encoding.encodingmode))
2298 pycompat.sysstr(encoding.encodingmode))
2296 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2299 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2297 pycompat.sysstr(encoding.encodingmode))
2300 pycompat.sysstr(encoding.encodingmode))
2298 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2301 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2299 pycompat.sysstr(encoding.encodingmode))
2302 pycompat.sysstr(encoding.encodingmode))
2300 wrapper = MBTextWrapper(width=width,
2303 wrapper = MBTextWrapper(width=width,
2301 initial_indent=initindent,
2304 initial_indent=initindent,
2302 subsequent_indent=hangindent)
2305 subsequent_indent=hangindent)
2303 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2306 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2304
2307
2305 if (pyplatform.python_implementation() == 'CPython' and
2308 if (pyplatform.python_implementation() == 'CPython' and
2306 sys.version_info < (3, 0)):
2309 sys.version_info < (3, 0)):
2307 # There is an issue in CPython that some IO methods do not handle EINTR
2310 # There is an issue in CPython that some IO methods do not handle EINTR
2308 # correctly. The following table shows what CPython version (and functions)
2311 # correctly. The following table shows what CPython version (and functions)
2309 # are affected (buggy: has the EINTR bug, okay: otherwise):
2312 # are affected (buggy: has the EINTR bug, okay: otherwise):
2310 #
2313 #
2311 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2314 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2312 # --------------------------------------------------
2315 # --------------------------------------------------
2313 # fp.__iter__ | buggy | buggy | okay
2316 # fp.__iter__ | buggy | buggy | okay
2314 # fp.read* | buggy | okay [1] | okay
2317 # fp.read* | buggy | okay [1] | okay
2315 #
2318 #
2316 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2319 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2317 #
2320 #
2318 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2321 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2319 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2322 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2320 #
2323 #
2321 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2324 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2322 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2325 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2323 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2326 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2324 # fp.__iter__ but not other fp.read* methods.
2327 # fp.__iter__ but not other fp.read* methods.
2325 #
2328 #
2326 # On modern systems like Linux, the "read" syscall cannot be interrupted
2329 # On modern systems like Linux, the "read" syscall cannot be interrupted
2327 # when reading "fast" files like on-disk files. So the EINTR issue only
2330 # when reading "fast" files like on-disk files. So the EINTR issue only
2328 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2331 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2329 # files approximately as "fast" files and use the fast (unsafe) code path,
2332 # files approximately as "fast" files and use the fast (unsafe) code path,
2330 # to minimize the performance impact.
2333 # to minimize the performance impact.
2331 if sys.version_info >= (2, 7, 4):
2334 if sys.version_info >= (2, 7, 4):
2332 # fp.readline deals with EINTR correctly, use it as a workaround.
2335 # fp.readline deals with EINTR correctly, use it as a workaround.
2333 def _safeiterfile(fp):
2336 def _safeiterfile(fp):
2334 return iter(fp.readline, '')
2337 return iter(fp.readline, '')
2335 else:
2338 else:
2336 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2339 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2337 # note: this may block longer than necessary because of bufsize.
2340 # note: this may block longer than necessary because of bufsize.
2338 def _safeiterfile(fp, bufsize=4096):
2341 def _safeiterfile(fp, bufsize=4096):
2339 fd = fp.fileno()
2342 fd = fp.fileno()
2340 line = ''
2343 line = ''
2341 while True:
2344 while True:
2342 try:
2345 try:
2343 buf = os.read(fd, bufsize)
2346 buf = os.read(fd, bufsize)
2344 except OSError as ex:
2347 except OSError as ex:
2345 # os.read only raises EINTR before any data is read
2348 # os.read only raises EINTR before any data is read
2346 if ex.errno == errno.EINTR:
2349 if ex.errno == errno.EINTR:
2347 continue
2350 continue
2348 else:
2351 else:
2349 raise
2352 raise
2350 line += buf
2353 line += buf
2351 if '\n' in buf:
2354 if '\n' in buf:
2352 splitted = line.splitlines(True)
2355 splitted = line.splitlines(True)
2353 line = ''
2356 line = ''
2354 for l in splitted:
2357 for l in splitted:
2355 if l[-1] == '\n':
2358 if l[-1] == '\n':
2356 yield l
2359 yield l
2357 else:
2360 else:
2358 line = l
2361 line = l
2359 if not buf:
2362 if not buf:
2360 break
2363 break
2361 if line:
2364 if line:
2362 yield line
2365 yield line
2363
2366
2364 def iterfile(fp):
2367 def iterfile(fp):
2365 fastpath = True
2368 fastpath = True
2366 if type(fp) is file:
2369 if type(fp) is file:
2367 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2370 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2368 if fastpath:
2371 if fastpath:
2369 return fp
2372 return fp
2370 else:
2373 else:
2371 return _safeiterfile(fp)
2374 return _safeiterfile(fp)
2372 else:
2375 else:
2373 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2376 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2374 def iterfile(fp):
2377 def iterfile(fp):
2375 return fp
2378 return fp
2376
2379
2377 def iterlines(iterator):
2380 def iterlines(iterator):
2378 for chunk in iterator:
2381 for chunk in iterator:
2379 for line in chunk.splitlines():
2382 for line in chunk.splitlines():
2380 yield line
2383 yield line
2381
2384
2382 def expandpath(path):
2385 def expandpath(path):
2383 return os.path.expanduser(os.path.expandvars(path))
2386 return os.path.expanduser(os.path.expandvars(path))
2384
2387
2385 def hgcmd():
2388 def hgcmd():
2386 """Return the command used to execute current hg
2389 """Return the command used to execute current hg
2387
2390
2388 This is different from hgexecutable() because on Windows we want
2391 This is different from hgexecutable() because on Windows we want
2389 to avoid things opening new shell windows like batch files, so we
2392 to avoid things opening new shell windows like batch files, so we
2390 get either the python call or current executable.
2393 get either the python call or current executable.
2391 """
2394 """
2392 if mainfrozen():
2395 if mainfrozen():
2393 if getattr(sys, 'frozen', None) == 'macosx_app':
2396 if getattr(sys, 'frozen', None) == 'macosx_app':
2394 # Env variable set by py2app
2397 # Env variable set by py2app
2395 return [encoding.environ['EXECUTABLEPATH']]
2398 return [encoding.environ['EXECUTABLEPATH']]
2396 else:
2399 else:
2397 return [pycompat.sysexecutable]
2400 return [pycompat.sysexecutable]
2398 return gethgcmd()
2401 return gethgcmd()
2399
2402
2400 def rundetached(args, condfn):
2403 def rundetached(args, condfn):
2401 """Execute the argument list in a detached process.
2404 """Execute the argument list in a detached process.
2402
2405
2403 condfn is a callable which is called repeatedly and should return
2406 condfn is a callable which is called repeatedly and should return
2404 True once the child process is known to have started successfully.
2407 True once the child process is known to have started successfully.
2405 At this point, the child process PID is returned. If the child
2408 At this point, the child process PID is returned. If the child
2406 process fails to start or finishes before condfn() evaluates to
2409 process fails to start or finishes before condfn() evaluates to
2407 True, return -1.
2410 True, return -1.
2408 """
2411 """
2409 # Windows case is easier because the child process is either
2412 # Windows case is easier because the child process is either
2410 # successfully starting and validating the condition or exiting
2413 # successfully starting and validating the condition or exiting
2411 # on failure. We just poll on its PID. On Unix, if the child
2414 # on failure. We just poll on its PID. On Unix, if the child
2412 # process fails to start, it will be left in a zombie state until
2415 # process fails to start, it will be left in a zombie state until
2413 # the parent wait on it, which we cannot do since we expect a long
2416 # the parent wait on it, which we cannot do since we expect a long
2414 # running process on success. Instead we listen for SIGCHLD telling
2417 # running process on success. Instead we listen for SIGCHLD telling
2415 # us our child process terminated.
2418 # us our child process terminated.
2416 terminated = set()
2419 terminated = set()
2417 def handler(signum, frame):
2420 def handler(signum, frame):
2418 terminated.add(os.wait())
2421 terminated.add(os.wait())
2419 prevhandler = None
2422 prevhandler = None
2420 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2423 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2421 if SIGCHLD is not None:
2424 if SIGCHLD is not None:
2422 prevhandler = signal.signal(SIGCHLD, handler)
2425 prevhandler = signal.signal(SIGCHLD, handler)
2423 try:
2426 try:
2424 pid = spawndetached(args)
2427 pid = spawndetached(args)
2425 while not condfn():
2428 while not condfn():
2426 if ((pid in terminated or not testpid(pid))
2429 if ((pid in terminated or not testpid(pid))
2427 and not condfn()):
2430 and not condfn()):
2428 return -1
2431 return -1
2429 time.sleep(0.1)
2432 time.sleep(0.1)
2430 return pid
2433 return pid
2431 finally:
2434 finally:
2432 if prevhandler is not None:
2435 if prevhandler is not None:
2433 signal.signal(signal.SIGCHLD, prevhandler)
2436 signal.signal(signal.SIGCHLD, prevhandler)
2434
2437
2435 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2438 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2436 """Return the result of interpolating items in the mapping into string s.
2439 """Return the result of interpolating items in the mapping into string s.
2437
2440
2438 prefix is a single character string, or a two character string with
2441 prefix is a single character string, or a two character string with
2439 a backslash as the first character if the prefix needs to be escaped in
2442 a backslash as the first character if the prefix needs to be escaped in
2440 a regular expression.
2443 a regular expression.
2441
2444
2442 fn is an optional function that will be applied to the replacement text
2445 fn is an optional function that will be applied to the replacement text
2443 just before replacement.
2446 just before replacement.
2444
2447
2445 escape_prefix is an optional flag that allows using doubled prefix for
2448 escape_prefix is an optional flag that allows using doubled prefix for
2446 its escaping.
2449 its escaping.
2447 """
2450 """
2448 fn = fn or (lambda s: s)
2451 fn = fn or (lambda s: s)
2449 patterns = '|'.join(mapping.keys())
2452 patterns = '|'.join(mapping.keys())
2450 if escape_prefix:
2453 if escape_prefix:
2451 patterns += '|' + prefix
2454 patterns += '|' + prefix
2452 if len(prefix) > 1:
2455 if len(prefix) > 1:
2453 prefix_char = prefix[1:]
2456 prefix_char = prefix[1:]
2454 else:
2457 else:
2455 prefix_char = prefix
2458 prefix_char = prefix
2456 mapping[prefix_char] = prefix_char
2459 mapping[prefix_char] = prefix_char
2457 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2460 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2458 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2461 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2459
2462
2460 def getport(port):
2463 def getport(port):
2461 """Return the port for a given network service.
2464 """Return the port for a given network service.
2462
2465
2463 If port is an integer, it's returned as is. If it's a string, it's
2466 If port is an integer, it's returned as is. If it's a string, it's
2464 looked up using socket.getservbyname(). If there's no matching
2467 looked up using socket.getservbyname(). If there's no matching
2465 service, error.Abort is raised.
2468 service, error.Abort is raised.
2466 """
2469 """
2467 try:
2470 try:
2468 return int(port)
2471 return int(port)
2469 except ValueError:
2472 except ValueError:
2470 pass
2473 pass
2471
2474
2472 try:
2475 try:
2473 return socket.getservbyname(port)
2476 return socket.getservbyname(port)
2474 except socket.error:
2477 except socket.error:
2475 raise Abort(_("no port number associated with service '%s'") % port)
2478 raise Abort(_("no port number associated with service '%s'") % port)
2476
2479
2477 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2480 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2478 '0': False, 'no': False, 'false': False, 'off': False,
2481 '0': False, 'no': False, 'false': False, 'off': False,
2479 'never': False}
2482 'never': False}
2480
2483
2481 def parsebool(s):
2484 def parsebool(s):
2482 """Parse s into a boolean.
2485 """Parse s into a boolean.
2483
2486
2484 If s is not a valid boolean, returns None.
2487 If s is not a valid boolean, returns None.
2485 """
2488 """
2486 return _booleans.get(s.lower(), None)
2489 return _booleans.get(s.lower(), None)
2487
2490
2488 _hextochr = dict((a + b, chr(int(a + b, 16)))
2491 _hextochr = dict((a + b, chr(int(a + b, 16)))
2489 for a in string.hexdigits for b in string.hexdigits)
2492 for a in string.hexdigits for b in string.hexdigits)
2490
2493
2491 class url(object):
2494 class url(object):
2492 r"""Reliable URL parser.
2495 r"""Reliable URL parser.
2493
2496
2494 This parses URLs and provides attributes for the following
2497 This parses URLs and provides attributes for the following
2495 components:
2498 components:
2496
2499
2497 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2500 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2498
2501
2499 Missing components are set to None. The only exception is
2502 Missing components are set to None. The only exception is
2500 fragment, which is set to '' if present but empty.
2503 fragment, which is set to '' if present but empty.
2501
2504
2502 If parsefragment is False, fragment is included in query. If
2505 If parsefragment is False, fragment is included in query. If
2503 parsequery is False, query is included in path. If both are
2506 parsequery is False, query is included in path. If both are
2504 False, both fragment and query are included in path.
2507 False, both fragment and query are included in path.
2505
2508
2506 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2509 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2507
2510
2508 Note that for backward compatibility reasons, bundle URLs do not
2511 Note that for backward compatibility reasons, bundle URLs do not
2509 take host names. That means 'bundle://../' has a path of '../'.
2512 take host names. That means 'bundle://../' has a path of '../'.
2510
2513
2511 Examples:
2514 Examples:
2512
2515
2513 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2516 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2514 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2517 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2515 >>> url('ssh://[::1]:2200//home/joe/repo')
2518 >>> url('ssh://[::1]:2200//home/joe/repo')
2516 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2519 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2517 >>> url('file:///home/joe/repo')
2520 >>> url('file:///home/joe/repo')
2518 <url scheme: 'file', path: '/home/joe/repo'>
2521 <url scheme: 'file', path: '/home/joe/repo'>
2519 >>> url('file:///c:/temp/foo/')
2522 >>> url('file:///c:/temp/foo/')
2520 <url scheme: 'file', path: 'c:/temp/foo/'>
2523 <url scheme: 'file', path: 'c:/temp/foo/'>
2521 >>> url('bundle:foo')
2524 >>> url('bundle:foo')
2522 <url scheme: 'bundle', path: 'foo'>
2525 <url scheme: 'bundle', path: 'foo'>
2523 >>> url('bundle://../foo')
2526 >>> url('bundle://../foo')
2524 <url scheme: 'bundle', path: '../foo'>
2527 <url scheme: 'bundle', path: '../foo'>
2525 >>> url(r'c:\foo\bar')
2528 >>> url(r'c:\foo\bar')
2526 <url path: 'c:\\foo\\bar'>
2529 <url path: 'c:\\foo\\bar'>
2527 >>> url(r'\\blah\blah\blah')
2530 >>> url(r'\\blah\blah\blah')
2528 <url path: '\\\\blah\\blah\\blah'>
2531 <url path: '\\\\blah\\blah\\blah'>
2529 >>> url(r'\\blah\blah\blah#baz')
2532 >>> url(r'\\blah\blah\blah#baz')
2530 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2533 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2531 >>> url(r'file:///C:\users\me')
2534 >>> url(r'file:///C:\users\me')
2532 <url scheme: 'file', path: 'C:\\users\\me'>
2535 <url scheme: 'file', path: 'C:\\users\\me'>
2533
2536
2534 Authentication credentials:
2537 Authentication credentials:
2535
2538
2536 >>> url('ssh://joe:xyz@x/repo')
2539 >>> url('ssh://joe:xyz@x/repo')
2537 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2540 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2538 >>> url('ssh://joe@x/repo')
2541 >>> url('ssh://joe@x/repo')
2539 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2542 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2540
2543
2541 Query strings and fragments:
2544 Query strings and fragments:
2542
2545
2543 >>> url('http://host/a?b#c')
2546 >>> url('http://host/a?b#c')
2544 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2547 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2545 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2548 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2546 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2549 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2547
2550
2548 Empty path:
2551 Empty path:
2549
2552
2550 >>> url('')
2553 >>> url('')
2551 <url path: ''>
2554 <url path: ''>
2552 >>> url('#a')
2555 >>> url('#a')
2553 <url path: '', fragment: 'a'>
2556 <url path: '', fragment: 'a'>
2554 >>> url('http://host/')
2557 >>> url('http://host/')
2555 <url scheme: 'http', host: 'host', path: ''>
2558 <url scheme: 'http', host: 'host', path: ''>
2556 >>> url('http://host/#a')
2559 >>> url('http://host/#a')
2557 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2560 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2558
2561
2559 Only scheme:
2562 Only scheme:
2560
2563
2561 >>> url('http:')
2564 >>> url('http:')
2562 <url scheme: 'http'>
2565 <url scheme: 'http'>
2563 """
2566 """
2564
2567
2565 _safechars = "!~*'()+"
2568 _safechars = "!~*'()+"
2566 _safepchars = "/!~*'()+:\\"
2569 _safepchars = "/!~*'()+:\\"
2567 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2570 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2568
2571
2569 def __init__(self, path, parsequery=True, parsefragment=True):
2572 def __init__(self, path, parsequery=True, parsefragment=True):
2570 # We slowly chomp away at path until we have only the path left
2573 # We slowly chomp away at path until we have only the path left
2571 self.scheme = self.user = self.passwd = self.host = None
2574 self.scheme = self.user = self.passwd = self.host = None
2572 self.port = self.path = self.query = self.fragment = None
2575 self.port = self.path = self.query = self.fragment = None
2573 self._localpath = True
2576 self._localpath = True
2574 self._hostport = ''
2577 self._hostport = ''
2575 self._origpath = path
2578 self._origpath = path
2576
2579
2577 if parsefragment and '#' in path:
2580 if parsefragment and '#' in path:
2578 path, self.fragment = path.split('#', 1)
2581 path, self.fragment = path.split('#', 1)
2579
2582
2580 # special case for Windows drive letters and UNC paths
2583 # special case for Windows drive letters and UNC paths
2581 if hasdriveletter(path) or path.startswith('\\\\'):
2584 if hasdriveletter(path) or path.startswith('\\\\'):
2582 self.path = path
2585 self.path = path
2583 return
2586 return
2584
2587
2585 # For compatibility reasons, we can't handle bundle paths as
2588 # For compatibility reasons, we can't handle bundle paths as
2586 # normal URLS
2589 # normal URLS
2587 if path.startswith('bundle:'):
2590 if path.startswith('bundle:'):
2588 self.scheme = 'bundle'
2591 self.scheme = 'bundle'
2589 path = path[7:]
2592 path = path[7:]
2590 if path.startswith('//'):
2593 if path.startswith('//'):
2591 path = path[2:]
2594 path = path[2:]
2592 self.path = path
2595 self.path = path
2593 return
2596 return
2594
2597
2595 if self._matchscheme(path):
2598 if self._matchscheme(path):
2596 parts = path.split(':', 1)
2599 parts = path.split(':', 1)
2597 if parts[0]:
2600 if parts[0]:
2598 self.scheme, path = parts
2601 self.scheme, path = parts
2599 self._localpath = False
2602 self._localpath = False
2600
2603
2601 if not path:
2604 if not path:
2602 path = None
2605 path = None
2603 if self._localpath:
2606 if self._localpath:
2604 self.path = ''
2607 self.path = ''
2605 return
2608 return
2606 else:
2609 else:
2607 if self._localpath:
2610 if self._localpath:
2608 self.path = path
2611 self.path = path
2609 return
2612 return
2610
2613
2611 if parsequery and '?' in path:
2614 if parsequery and '?' in path:
2612 path, self.query = path.split('?', 1)
2615 path, self.query = path.split('?', 1)
2613 if not path:
2616 if not path:
2614 path = None
2617 path = None
2615 if not self.query:
2618 if not self.query:
2616 self.query = None
2619 self.query = None
2617
2620
2618 # // is required to specify a host/authority
2621 # // is required to specify a host/authority
2619 if path and path.startswith('//'):
2622 if path and path.startswith('//'):
2620 parts = path[2:].split('/', 1)
2623 parts = path[2:].split('/', 1)
2621 if len(parts) > 1:
2624 if len(parts) > 1:
2622 self.host, path = parts
2625 self.host, path = parts
2623 else:
2626 else:
2624 self.host = parts[0]
2627 self.host = parts[0]
2625 path = None
2628 path = None
2626 if not self.host:
2629 if not self.host:
2627 self.host = None
2630 self.host = None
2628 # path of file:///d is /d
2631 # path of file:///d is /d
2629 # path of file:///d:/ is d:/, not /d:/
2632 # path of file:///d:/ is d:/, not /d:/
2630 if path and not hasdriveletter(path):
2633 if path and not hasdriveletter(path):
2631 path = '/' + path
2634 path = '/' + path
2632
2635
2633 if self.host and '@' in self.host:
2636 if self.host and '@' in self.host:
2634 self.user, self.host = self.host.rsplit('@', 1)
2637 self.user, self.host = self.host.rsplit('@', 1)
2635 if ':' in self.user:
2638 if ':' in self.user:
2636 self.user, self.passwd = self.user.split(':', 1)
2639 self.user, self.passwd = self.user.split(':', 1)
2637 if not self.host:
2640 if not self.host:
2638 self.host = None
2641 self.host = None
2639
2642
2640 # Don't split on colons in IPv6 addresses without ports
2643 # Don't split on colons in IPv6 addresses without ports
2641 if (self.host and ':' in self.host and
2644 if (self.host and ':' in self.host and
2642 not (self.host.startswith('[') and self.host.endswith(']'))):
2645 not (self.host.startswith('[') and self.host.endswith(']'))):
2643 self._hostport = self.host
2646 self._hostport = self.host
2644 self.host, self.port = self.host.rsplit(':', 1)
2647 self.host, self.port = self.host.rsplit(':', 1)
2645 if not self.host:
2648 if not self.host:
2646 self.host = None
2649 self.host = None
2647
2650
2648 if (self.host and self.scheme == 'file' and
2651 if (self.host and self.scheme == 'file' and
2649 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2652 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2650 raise Abort(_('file:// URLs can only refer to localhost'))
2653 raise Abort(_('file:// URLs can only refer to localhost'))
2651
2654
2652 self.path = path
2655 self.path = path
2653
2656
2654 # leave the query string escaped
2657 # leave the query string escaped
2655 for a in ('user', 'passwd', 'host', 'port',
2658 for a in ('user', 'passwd', 'host', 'port',
2656 'path', 'fragment'):
2659 'path', 'fragment'):
2657 v = getattr(self, a)
2660 v = getattr(self, a)
2658 if v is not None:
2661 if v is not None:
2659 setattr(self, a, urlreq.unquote(v))
2662 setattr(self, a, urlreq.unquote(v))
2660
2663
2661 def __repr__(self):
2664 def __repr__(self):
2662 attrs = []
2665 attrs = []
2663 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2666 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2664 'query', 'fragment'):
2667 'query', 'fragment'):
2665 v = getattr(self, a)
2668 v = getattr(self, a)
2666 if v is not None:
2669 if v is not None:
2667 attrs.append('%s: %r' % (a, v))
2670 attrs.append('%s: %r' % (a, v))
2668 return '<url %s>' % ', '.join(attrs)
2671 return '<url %s>' % ', '.join(attrs)
2669
2672
2670 def __str__(self):
2673 def __str__(self):
2671 r"""Join the URL's components back into a URL string.
2674 r"""Join the URL's components back into a URL string.
2672
2675
2673 Examples:
2676 Examples:
2674
2677
2675 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2678 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2676 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2679 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2677 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2680 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2678 'http://user:pw@host:80/?foo=bar&baz=42'
2681 'http://user:pw@host:80/?foo=bar&baz=42'
2679 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2682 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2680 'http://user:pw@host:80/?foo=bar%3dbaz'
2683 'http://user:pw@host:80/?foo=bar%3dbaz'
2681 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2684 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2682 'ssh://user:pw@[::1]:2200//home/joe#'
2685 'ssh://user:pw@[::1]:2200//home/joe#'
2683 >>> str(url('http://localhost:80//'))
2686 >>> str(url('http://localhost:80//'))
2684 'http://localhost:80//'
2687 'http://localhost:80//'
2685 >>> str(url('http://localhost:80/'))
2688 >>> str(url('http://localhost:80/'))
2686 'http://localhost:80/'
2689 'http://localhost:80/'
2687 >>> str(url('http://localhost:80'))
2690 >>> str(url('http://localhost:80'))
2688 'http://localhost:80/'
2691 'http://localhost:80/'
2689 >>> str(url('bundle:foo'))
2692 >>> str(url('bundle:foo'))
2690 'bundle:foo'
2693 'bundle:foo'
2691 >>> str(url('bundle://../foo'))
2694 >>> str(url('bundle://../foo'))
2692 'bundle:../foo'
2695 'bundle:../foo'
2693 >>> str(url('path'))
2696 >>> str(url('path'))
2694 'path'
2697 'path'
2695 >>> str(url('file:///tmp/foo/bar'))
2698 >>> str(url('file:///tmp/foo/bar'))
2696 'file:///tmp/foo/bar'
2699 'file:///tmp/foo/bar'
2697 >>> str(url('file:///c:/tmp/foo/bar'))
2700 >>> str(url('file:///c:/tmp/foo/bar'))
2698 'file:///c:/tmp/foo/bar'
2701 'file:///c:/tmp/foo/bar'
2699 >>> print url(r'bundle:foo\bar')
2702 >>> print url(r'bundle:foo\bar')
2700 bundle:foo\bar
2703 bundle:foo\bar
2701 >>> print url(r'file:///D:\data\hg')
2704 >>> print url(r'file:///D:\data\hg')
2702 file:///D:\data\hg
2705 file:///D:\data\hg
2703 """
2706 """
2704 return encoding.strfromlocal(self.__bytes__())
2707 return encoding.strfromlocal(self.__bytes__())
2705
2708
2706 def __bytes__(self):
2709 def __bytes__(self):
2707 if self._localpath:
2710 if self._localpath:
2708 s = self.path
2711 s = self.path
2709 if self.scheme == 'bundle':
2712 if self.scheme == 'bundle':
2710 s = 'bundle:' + s
2713 s = 'bundle:' + s
2711 if self.fragment:
2714 if self.fragment:
2712 s += '#' + self.fragment
2715 s += '#' + self.fragment
2713 return s
2716 return s
2714
2717
2715 s = self.scheme + ':'
2718 s = self.scheme + ':'
2716 if self.user or self.passwd or self.host:
2719 if self.user or self.passwd or self.host:
2717 s += '//'
2720 s += '//'
2718 elif self.scheme and (not self.path or self.path.startswith('/')
2721 elif self.scheme and (not self.path or self.path.startswith('/')
2719 or hasdriveletter(self.path)):
2722 or hasdriveletter(self.path)):
2720 s += '//'
2723 s += '//'
2721 if hasdriveletter(self.path):
2724 if hasdriveletter(self.path):
2722 s += '/'
2725 s += '/'
2723 if self.user:
2726 if self.user:
2724 s += urlreq.quote(self.user, safe=self._safechars)
2727 s += urlreq.quote(self.user, safe=self._safechars)
2725 if self.passwd:
2728 if self.passwd:
2726 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2729 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2727 if self.user or self.passwd:
2730 if self.user or self.passwd:
2728 s += '@'
2731 s += '@'
2729 if self.host:
2732 if self.host:
2730 if not (self.host.startswith('[') and self.host.endswith(']')):
2733 if not (self.host.startswith('[') and self.host.endswith(']')):
2731 s += urlreq.quote(self.host)
2734 s += urlreq.quote(self.host)
2732 else:
2735 else:
2733 s += self.host
2736 s += self.host
2734 if self.port:
2737 if self.port:
2735 s += ':' + urlreq.quote(self.port)
2738 s += ':' + urlreq.quote(self.port)
2736 if self.host:
2739 if self.host:
2737 s += '/'
2740 s += '/'
2738 if self.path:
2741 if self.path:
2739 # TODO: similar to the query string, we should not unescape the
2742 # TODO: similar to the query string, we should not unescape the
2740 # path when we store it, the path might contain '%2f' = '/',
2743 # path when we store it, the path might contain '%2f' = '/',
2741 # which we should *not* escape.
2744 # which we should *not* escape.
2742 s += urlreq.quote(self.path, safe=self._safepchars)
2745 s += urlreq.quote(self.path, safe=self._safepchars)
2743 if self.query:
2746 if self.query:
2744 # we store the query in escaped form.
2747 # we store the query in escaped form.
2745 s += '?' + self.query
2748 s += '?' + self.query
2746 if self.fragment is not None:
2749 if self.fragment is not None:
2747 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2750 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2748 return s
2751 return s
2749
2752
2750 def authinfo(self):
2753 def authinfo(self):
2751 user, passwd = self.user, self.passwd
2754 user, passwd = self.user, self.passwd
2752 try:
2755 try:
2753 self.user, self.passwd = None, None
2756 self.user, self.passwd = None, None
2754 s = str(self)
2757 s = str(self)
2755 finally:
2758 finally:
2756 self.user, self.passwd = user, passwd
2759 self.user, self.passwd = user, passwd
2757 if not self.user:
2760 if not self.user:
2758 return (s, None)
2761 return (s, None)
2759 # authinfo[1] is passed to urllib2 password manager, and its
2762 # authinfo[1] is passed to urllib2 password manager, and its
2760 # URIs must not contain credentials. The host is passed in the
2763 # URIs must not contain credentials. The host is passed in the
2761 # URIs list because Python < 2.4.3 uses only that to search for
2764 # URIs list because Python < 2.4.3 uses only that to search for
2762 # a password.
2765 # a password.
2763 return (s, (None, (s, self.host),
2766 return (s, (None, (s, self.host),
2764 self.user, self.passwd or ''))
2767 self.user, self.passwd or ''))
2765
2768
2766 def isabs(self):
2769 def isabs(self):
2767 if self.scheme and self.scheme != 'file':
2770 if self.scheme and self.scheme != 'file':
2768 return True # remote URL
2771 return True # remote URL
2769 if hasdriveletter(self.path):
2772 if hasdriveletter(self.path):
2770 return True # absolute for our purposes - can't be joined()
2773 return True # absolute for our purposes - can't be joined()
2771 if self.path.startswith(r'\\'):
2774 if self.path.startswith(r'\\'):
2772 return True # Windows UNC path
2775 return True # Windows UNC path
2773 if self.path.startswith('/'):
2776 if self.path.startswith('/'):
2774 return True # POSIX-style
2777 return True # POSIX-style
2775 return False
2778 return False
2776
2779
2777 def localpath(self):
2780 def localpath(self):
2778 if self.scheme == 'file' or self.scheme == 'bundle':
2781 if self.scheme == 'file' or self.scheme == 'bundle':
2779 path = self.path or '/'
2782 path = self.path or '/'
2780 # For Windows, we need to promote hosts containing drive
2783 # For Windows, we need to promote hosts containing drive
2781 # letters to paths with drive letters.
2784 # letters to paths with drive letters.
2782 if hasdriveletter(self._hostport):
2785 if hasdriveletter(self._hostport):
2783 path = self._hostport + '/' + self.path
2786 path = self._hostport + '/' + self.path
2784 elif (self.host is not None and self.path
2787 elif (self.host is not None and self.path
2785 and not hasdriveletter(path)):
2788 and not hasdriveletter(path)):
2786 path = '/' + path
2789 path = '/' + path
2787 return path
2790 return path
2788 return self._origpath
2791 return self._origpath
2789
2792
2790 def islocal(self):
2793 def islocal(self):
2791 '''whether localpath will return something that posixfile can open'''
2794 '''whether localpath will return something that posixfile can open'''
2792 return (not self.scheme or self.scheme == 'file'
2795 return (not self.scheme or self.scheme == 'file'
2793 or self.scheme == 'bundle')
2796 or self.scheme == 'bundle')
2794
2797
2795 def hasscheme(path):
2798 def hasscheme(path):
2796 return bool(url(path).scheme)
2799 return bool(url(path).scheme)
2797
2800
2798 def hasdriveletter(path):
2801 def hasdriveletter(path):
2799 return path and path[1:2] == ':' and path[0:1].isalpha()
2802 return path and path[1:2] == ':' and path[0:1].isalpha()
2800
2803
2801 def urllocalpath(path):
2804 def urllocalpath(path):
2802 return url(path, parsequery=False, parsefragment=False).localpath()
2805 return url(path, parsequery=False, parsefragment=False).localpath()
2803
2806
2804 def hidepassword(u):
2807 def hidepassword(u):
2805 '''hide user credential in a url string'''
2808 '''hide user credential in a url string'''
2806 u = url(u)
2809 u = url(u)
2807 if u.passwd:
2810 if u.passwd:
2808 u.passwd = '***'
2811 u.passwd = '***'
2809 return str(u)
2812 return str(u)
2810
2813
2811 def removeauth(u):
2814 def removeauth(u):
2812 '''remove all authentication information from a url string'''
2815 '''remove all authentication information from a url string'''
2813 u = url(u)
2816 u = url(u)
2814 u.user = u.passwd = None
2817 u.user = u.passwd = None
2815 return str(u)
2818 return str(u)
2816
2819
2817 timecount = unitcountfn(
2820 timecount = unitcountfn(
2818 (1, 1e3, _('%.0f s')),
2821 (1, 1e3, _('%.0f s')),
2819 (100, 1, _('%.1f s')),
2822 (100, 1, _('%.1f s')),
2820 (10, 1, _('%.2f s')),
2823 (10, 1, _('%.2f s')),
2821 (1, 1, _('%.3f s')),
2824 (1, 1, _('%.3f s')),
2822 (100, 0.001, _('%.1f ms')),
2825 (100, 0.001, _('%.1f ms')),
2823 (10, 0.001, _('%.2f ms')),
2826 (10, 0.001, _('%.2f ms')),
2824 (1, 0.001, _('%.3f ms')),
2827 (1, 0.001, _('%.3f ms')),
2825 (100, 0.000001, _('%.1f us')),
2828 (100, 0.000001, _('%.1f us')),
2826 (10, 0.000001, _('%.2f us')),
2829 (10, 0.000001, _('%.2f us')),
2827 (1, 0.000001, _('%.3f us')),
2830 (1, 0.000001, _('%.3f us')),
2828 (100, 0.000000001, _('%.1f ns')),
2831 (100, 0.000000001, _('%.1f ns')),
2829 (10, 0.000000001, _('%.2f ns')),
2832 (10, 0.000000001, _('%.2f ns')),
2830 (1, 0.000000001, _('%.3f ns')),
2833 (1, 0.000000001, _('%.3f ns')),
2831 )
2834 )
2832
2835
2833 _timenesting = [0]
2836 _timenesting = [0]
2834
2837
2835 def timed(func):
2838 def timed(func):
2836 '''Report the execution time of a function call to stderr.
2839 '''Report the execution time of a function call to stderr.
2837
2840
2838 During development, use as a decorator when you need to measure
2841 During development, use as a decorator when you need to measure
2839 the cost of a function, e.g. as follows:
2842 the cost of a function, e.g. as follows:
2840
2843
2841 @util.timed
2844 @util.timed
2842 def foo(a, b, c):
2845 def foo(a, b, c):
2843 pass
2846 pass
2844 '''
2847 '''
2845
2848
2846 def wrapper(*args, **kwargs):
2849 def wrapper(*args, **kwargs):
2847 start = timer()
2850 start = timer()
2848 indent = 2
2851 indent = 2
2849 _timenesting[0] += indent
2852 _timenesting[0] += indent
2850 try:
2853 try:
2851 return func(*args, **kwargs)
2854 return func(*args, **kwargs)
2852 finally:
2855 finally:
2853 elapsed = timer() - start
2856 elapsed = timer() - start
2854 _timenesting[0] -= indent
2857 _timenesting[0] -= indent
2855 stderr.write('%s%s: %s\n' %
2858 stderr.write('%s%s: %s\n' %
2856 (' ' * _timenesting[0], func.__name__,
2859 (' ' * _timenesting[0], func.__name__,
2857 timecount(elapsed)))
2860 timecount(elapsed)))
2858 return wrapper
2861 return wrapper
2859
2862
2860 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2863 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2861 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2864 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2862
2865
2863 def sizetoint(s):
2866 def sizetoint(s):
2864 '''Convert a space specifier to a byte count.
2867 '''Convert a space specifier to a byte count.
2865
2868
2866 >>> sizetoint('30')
2869 >>> sizetoint('30')
2867 30
2870 30
2868 >>> sizetoint('2.2kb')
2871 >>> sizetoint('2.2kb')
2869 2252
2872 2252
2870 >>> sizetoint('6M')
2873 >>> sizetoint('6M')
2871 6291456
2874 6291456
2872 '''
2875 '''
2873 t = s.strip().lower()
2876 t = s.strip().lower()
2874 try:
2877 try:
2875 for k, u in _sizeunits:
2878 for k, u in _sizeunits:
2876 if t.endswith(k):
2879 if t.endswith(k):
2877 return int(float(t[:-len(k)]) * u)
2880 return int(float(t[:-len(k)]) * u)
2878 return int(t)
2881 return int(t)
2879 except ValueError:
2882 except ValueError:
2880 raise error.ParseError(_("couldn't parse size: %s") % s)
2883 raise error.ParseError(_("couldn't parse size: %s") % s)
2881
2884
2882 class hooks(object):
2885 class hooks(object):
2883 '''A collection of hook functions that can be used to extend a
2886 '''A collection of hook functions that can be used to extend a
2884 function's behavior. Hooks are called in lexicographic order,
2887 function's behavior. Hooks are called in lexicographic order,
2885 based on the names of their sources.'''
2888 based on the names of their sources.'''
2886
2889
2887 def __init__(self):
2890 def __init__(self):
2888 self._hooks = []
2891 self._hooks = []
2889
2892
2890 def add(self, source, hook):
2893 def add(self, source, hook):
2891 self._hooks.append((source, hook))
2894 self._hooks.append((source, hook))
2892
2895
2893 def __call__(self, *args):
2896 def __call__(self, *args):
2894 self._hooks.sort(key=lambda x: x[0])
2897 self._hooks.sort(key=lambda x: x[0])
2895 results = []
2898 results = []
2896 for source, hook in self._hooks:
2899 for source, hook in self._hooks:
2897 results.append(hook(*args))
2900 results.append(hook(*args))
2898 return results
2901 return results
2899
2902
2900 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2903 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2901 '''Yields lines for a nicely formatted stacktrace.
2904 '''Yields lines for a nicely formatted stacktrace.
2902 Skips the 'skip' last entries, then return the last 'depth' entries.
2905 Skips the 'skip' last entries, then return the last 'depth' entries.
2903 Each file+linenumber is formatted according to fileline.
2906 Each file+linenumber is formatted according to fileline.
2904 Each line is formatted according to line.
2907 Each line is formatted according to line.
2905 If line is None, it yields:
2908 If line is None, it yields:
2906 length of longest filepath+line number,
2909 length of longest filepath+line number,
2907 filepath+linenumber,
2910 filepath+linenumber,
2908 function
2911 function
2909
2912
2910 Not be used in production code but very convenient while developing.
2913 Not be used in production code but very convenient while developing.
2911 '''
2914 '''
2912 entries = [(fileline % (fn, ln), func)
2915 entries = [(fileline % (fn, ln), func)
2913 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2916 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2914 ][-depth:]
2917 ][-depth:]
2915 if entries:
2918 if entries:
2916 fnmax = max(len(entry[0]) for entry in entries)
2919 fnmax = max(len(entry[0]) for entry in entries)
2917 for fnln, func in entries:
2920 for fnln, func in entries:
2918 if line is None:
2921 if line is None:
2919 yield (fnmax, fnln, func)
2922 yield (fnmax, fnln, func)
2920 else:
2923 else:
2921 yield line % (fnmax, fnln, func)
2924 yield line % (fnmax, fnln, func)
2922
2925
2923 def debugstacktrace(msg='stacktrace', skip=0,
2926 def debugstacktrace(msg='stacktrace', skip=0,
2924 f=stderr, otherf=stdout, depth=0):
2927 f=stderr, otherf=stdout, depth=0):
2925 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2928 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2926 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2929 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2927 By default it will flush stdout first.
2930 By default it will flush stdout first.
2928 It can be used everywhere and intentionally does not require an ui object.
2931 It can be used everywhere and intentionally does not require an ui object.
2929 Not be used in production code but very convenient while developing.
2932 Not be used in production code but very convenient while developing.
2930 '''
2933 '''
2931 if otherf:
2934 if otherf:
2932 otherf.flush()
2935 otherf.flush()
2933 f.write('%s at:\n' % msg.rstrip())
2936 f.write('%s at:\n' % msg.rstrip())
2934 for line in getstackframes(skip + 1, depth=depth):
2937 for line in getstackframes(skip + 1, depth=depth):
2935 f.write(line)
2938 f.write(line)
2936 f.flush()
2939 f.flush()
2937
2940
2938 class dirs(object):
2941 class dirs(object):
2939 '''a multiset of directory names from a dirstate or manifest'''
2942 '''a multiset of directory names from a dirstate or manifest'''
2940
2943
2941 def __init__(self, map, skip=None):
2944 def __init__(self, map, skip=None):
2942 self._dirs = {}
2945 self._dirs = {}
2943 addpath = self.addpath
2946 addpath = self.addpath
2944 if safehasattr(map, 'iteritems') and skip is not None:
2947 if safehasattr(map, 'iteritems') and skip is not None:
2945 for f, s in map.iteritems():
2948 for f, s in map.iteritems():
2946 if s[0] != skip:
2949 if s[0] != skip:
2947 addpath(f)
2950 addpath(f)
2948 else:
2951 else:
2949 for f in map:
2952 for f in map:
2950 addpath(f)
2953 addpath(f)
2951
2954
2952 def addpath(self, path):
2955 def addpath(self, path):
2953 dirs = self._dirs
2956 dirs = self._dirs
2954 for base in finddirs(path):
2957 for base in finddirs(path):
2955 if base in dirs:
2958 if base in dirs:
2956 dirs[base] += 1
2959 dirs[base] += 1
2957 return
2960 return
2958 dirs[base] = 1
2961 dirs[base] = 1
2959
2962
2960 def delpath(self, path):
2963 def delpath(self, path):
2961 dirs = self._dirs
2964 dirs = self._dirs
2962 for base in finddirs(path):
2965 for base in finddirs(path):
2963 if dirs[base] > 1:
2966 if dirs[base] > 1:
2964 dirs[base] -= 1
2967 dirs[base] -= 1
2965 return
2968 return
2966 del dirs[base]
2969 del dirs[base]
2967
2970
2968 def __iter__(self):
2971 def __iter__(self):
2969 return iter(self._dirs)
2972 return iter(self._dirs)
2970
2973
2971 def __contains__(self, d):
2974 def __contains__(self, d):
2972 return d in self._dirs
2975 return d in self._dirs
2973
2976
2974 if safehasattr(parsers, 'dirs'):
2977 if safehasattr(parsers, 'dirs'):
2975 dirs = parsers.dirs
2978 dirs = parsers.dirs
2976
2979
2977 def finddirs(path):
2980 def finddirs(path):
2978 pos = path.rfind('/')
2981 pos = path.rfind('/')
2979 while pos != -1:
2982 while pos != -1:
2980 yield path[:pos]
2983 yield path[:pos]
2981 pos = path.rfind('/', 0, pos)
2984 pos = path.rfind('/', 0, pos)
2982
2985
2983 class ctxmanager(object):
2986 class ctxmanager(object):
2984 '''A context manager for use in 'with' blocks to allow multiple
2987 '''A context manager for use in 'with' blocks to allow multiple
2985 contexts to be entered at once. This is both safer and more
2988 contexts to be entered at once. This is both safer and more
2986 flexible than contextlib.nested.
2989 flexible than contextlib.nested.
2987
2990
2988 Once Mercurial supports Python 2.7+, this will become mostly
2991 Once Mercurial supports Python 2.7+, this will become mostly
2989 unnecessary.
2992 unnecessary.
2990 '''
2993 '''
2991
2994
2992 def __init__(self, *args):
2995 def __init__(self, *args):
2993 '''Accepts a list of no-argument functions that return context
2996 '''Accepts a list of no-argument functions that return context
2994 managers. These will be invoked at __call__ time.'''
2997 managers. These will be invoked at __call__ time.'''
2995 self._pending = args
2998 self._pending = args
2996 self._atexit = []
2999 self._atexit = []
2997
3000
2998 def __enter__(self):
3001 def __enter__(self):
2999 return self
3002 return self
3000
3003
3001 def enter(self):
3004 def enter(self):
3002 '''Create and enter context managers in the order in which they were
3005 '''Create and enter context managers in the order in which they were
3003 passed to the constructor.'''
3006 passed to the constructor.'''
3004 values = []
3007 values = []
3005 for func in self._pending:
3008 for func in self._pending:
3006 obj = func()
3009 obj = func()
3007 values.append(obj.__enter__())
3010 values.append(obj.__enter__())
3008 self._atexit.append(obj.__exit__)
3011 self._atexit.append(obj.__exit__)
3009 del self._pending
3012 del self._pending
3010 return values
3013 return values
3011
3014
3012 def atexit(self, func, *args, **kwargs):
3015 def atexit(self, func, *args, **kwargs):
3013 '''Add a function to call when this context manager exits. The
3016 '''Add a function to call when this context manager exits. The
3014 ordering of multiple atexit calls is unspecified, save that
3017 ordering of multiple atexit calls is unspecified, save that
3015 they will happen before any __exit__ functions.'''
3018 they will happen before any __exit__ functions.'''
3016 def wrapper(exc_type, exc_val, exc_tb):
3019 def wrapper(exc_type, exc_val, exc_tb):
3017 func(*args, **kwargs)
3020 func(*args, **kwargs)
3018 self._atexit.append(wrapper)
3021 self._atexit.append(wrapper)
3019 return func
3022 return func
3020
3023
3021 def __exit__(self, exc_type, exc_val, exc_tb):
3024 def __exit__(self, exc_type, exc_val, exc_tb):
3022 '''Context managers are exited in the reverse order from which
3025 '''Context managers are exited in the reverse order from which
3023 they were created.'''
3026 they were created.'''
3024 received = exc_type is not None
3027 received = exc_type is not None
3025 suppressed = False
3028 suppressed = False
3026 pending = None
3029 pending = None
3027 self._atexit.reverse()
3030 self._atexit.reverse()
3028 for exitfunc in self._atexit:
3031 for exitfunc in self._atexit:
3029 try:
3032 try:
3030 if exitfunc(exc_type, exc_val, exc_tb):
3033 if exitfunc(exc_type, exc_val, exc_tb):
3031 suppressed = True
3034 suppressed = True
3032 exc_type = None
3035 exc_type = None
3033 exc_val = None
3036 exc_val = None
3034 exc_tb = None
3037 exc_tb = None
3035 except BaseException:
3038 except BaseException:
3036 pending = sys.exc_info()
3039 pending = sys.exc_info()
3037 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3040 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3038 del self._atexit
3041 del self._atexit
3039 if pending:
3042 if pending:
3040 raise exc_val
3043 raise exc_val
3041 return received and suppressed
3044 return received and suppressed
3042
3045
3043 # compression code
3046 # compression code
3044
3047
3045 SERVERROLE = 'server'
3048 SERVERROLE = 'server'
3046 CLIENTROLE = 'client'
3049 CLIENTROLE = 'client'
3047
3050
3048 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3051 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3049 (u'name', u'serverpriority',
3052 (u'name', u'serverpriority',
3050 u'clientpriority'))
3053 u'clientpriority'))
3051
3054
3052 class compressormanager(object):
3055 class compressormanager(object):
3053 """Holds registrations of various compression engines.
3056 """Holds registrations of various compression engines.
3054
3057
3055 This class essentially abstracts the differences between compression
3058 This class essentially abstracts the differences between compression
3056 engines to allow new compression formats to be added easily, possibly from
3059 engines to allow new compression formats to be added easily, possibly from
3057 extensions.
3060 extensions.
3058
3061
3059 Compressors are registered against the global instance by calling its
3062 Compressors are registered against the global instance by calling its
3060 ``register()`` method.
3063 ``register()`` method.
3061 """
3064 """
3062 def __init__(self):
3065 def __init__(self):
3063 self._engines = {}
3066 self._engines = {}
3064 # Bundle spec human name to engine name.
3067 # Bundle spec human name to engine name.
3065 self._bundlenames = {}
3068 self._bundlenames = {}
3066 # Internal bundle identifier to engine name.
3069 # Internal bundle identifier to engine name.
3067 self._bundletypes = {}
3070 self._bundletypes = {}
3068 # Revlog header to engine name.
3071 # Revlog header to engine name.
3069 self._revlogheaders = {}
3072 self._revlogheaders = {}
3070 # Wire proto identifier to engine name.
3073 # Wire proto identifier to engine name.
3071 self._wiretypes = {}
3074 self._wiretypes = {}
3072
3075
3073 def __getitem__(self, key):
3076 def __getitem__(self, key):
3074 return self._engines[key]
3077 return self._engines[key]
3075
3078
3076 def __contains__(self, key):
3079 def __contains__(self, key):
3077 return key in self._engines
3080 return key in self._engines
3078
3081
3079 def __iter__(self):
3082 def __iter__(self):
3080 return iter(self._engines.keys())
3083 return iter(self._engines.keys())
3081
3084
3082 def register(self, engine):
3085 def register(self, engine):
3083 """Register a compression engine with the manager.
3086 """Register a compression engine with the manager.
3084
3087
3085 The argument must be a ``compressionengine`` instance.
3088 The argument must be a ``compressionengine`` instance.
3086 """
3089 """
3087 if not isinstance(engine, compressionengine):
3090 if not isinstance(engine, compressionengine):
3088 raise ValueError(_('argument must be a compressionengine'))
3091 raise ValueError(_('argument must be a compressionengine'))
3089
3092
3090 name = engine.name()
3093 name = engine.name()
3091
3094
3092 if name in self._engines:
3095 if name in self._engines:
3093 raise error.Abort(_('compression engine %s already registered') %
3096 raise error.Abort(_('compression engine %s already registered') %
3094 name)
3097 name)
3095
3098
3096 bundleinfo = engine.bundletype()
3099 bundleinfo = engine.bundletype()
3097 if bundleinfo:
3100 if bundleinfo:
3098 bundlename, bundletype = bundleinfo
3101 bundlename, bundletype = bundleinfo
3099
3102
3100 if bundlename in self._bundlenames:
3103 if bundlename in self._bundlenames:
3101 raise error.Abort(_('bundle name %s already registered') %
3104 raise error.Abort(_('bundle name %s already registered') %
3102 bundlename)
3105 bundlename)
3103 if bundletype in self._bundletypes:
3106 if bundletype in self._bundletypes:
3104 raise error.Abort(_('bundle type %s already registered by %s') %
3107 raise error.Abort(_('bundle type %s already registered by %s') %
3105 (bundletype, self._bundletypes[bundletype]))
3108 (bundletype, self._bundletypes[bundletype]))
3106
3109
3107 # No external facing name declared.
3110 # No external facing name declared.
3108 if bundlename:
3111 if bundlename:
3109 self._bundlenames[bundlename] = name
3112 self._bundlenames[bundlename] = name
3110
3113
3111 self._bundletypes[bundletype] = name
3114 self._bundletypes[bundletype] = name
3112
3115
3113 wiresupport = engine.wireprotosupport()
3116 wiresupport = engine.wireprotosupport()
3114 if wiresupport:
3117 if wiresupport:
3115 wiretype = wiresupport.name
3118 wiretype = wiresupport.name
3116 if wiretype in self._wiretypes:
3119 if wiretype in self._wiretypes:
3117 raise error.Abort(_('wire protocol compression %s already '
3120 raise error.Abort(_('wire protocol compression %s already '
3118 'registered by %s') %
3121 'registered by %s') %
3119 (wiretype, self._wiretypes[wiretype]))
3122 (wiretype, self._wiretypes[wiretype]))
3120
3123
3121 self._wiretypes[wiretype] = name
3124 self._wiretypes[wiretype] = name
3122
3125
3123 revlogheader = engine.revlogheader()
3126 revlogheader = engine.revlogheader()
3124 if revlogheader and revlogheader in self._revlogheaders:
3127 if revlogheader and revlogheader in self._revlogheaders:
3125 raise error.Abort(_('revlog header %s already registered by %s') %
3128 raise error.Abort(_('revlog header %s already registered by %s') %
3126 (revlogheader, self._revlogheaders[revlogheader]))
3129 (revlogheader, self._revlogheaders[revlogheader]))
3127
3130
3128 if revlogheader:
3131 if revlogheader:
3129 self._revlogheaders[revlogheader] = name
3132 self._revlogheaders[revlogheader] = name
3130
3133
3131 self._engines[name] = engine
3134 self._engines[name] = engine
3132
3135
3133 @property
3136 @property
3134 def supportedbundlenames(self):
3137 def supportedbundlenames(self):
3135 return set(self._bundlenames.keys())
3138 return set(self._bundlenames.keys())
3136
3139
3137 @property
3140 @property
3138 def supportedbundletypes(self):
3141 def supportedbundletypes(self):
3139 return set(self._bundletypes.keys())
3142 return set(self._bundletypes.keys())
3140
3143
3141 def forbundlename(self, bundlename):
3144 def forbundlename(self, bundlename):
3142 """Obtain a compression engine registered to a bundle name.
3145 """Obtain a compression engine registered to a bundle name.
3143
3146
3144 Will raise KeyError if the bundle type isn't registered.
3147 Will raise KeyError if the bundle type isn't registered.
3145
3148
3146 Will abort if the engine is known but not available.
3149 Will abort if the engine is known but not available.
3147 """
3150 """
3148 engine = self._engines[self._bundlenames[bundlename]]
3151 engine = self._engines[self._bundlenames[bundlename]]
3149 if not engine.available():
3152 if not engine.available():
3150 raise error.Abort(_('compression engine %s could not be loaded') %
3153 raise error.Abort(_('compression engine %s could not be loaded') %
3151 engine.name())
3154 engine.name())
3152 return engine
3155 return engine
3153
3156
3154 def forbundletype(self, bundletype):
3157 def forbundletype(self, bundletype):
3155 """Obtain a compression engine registered to a bundle type.
3158 """Obtain a compression engine registered to a bundle type.
3156
3159
3157 Will raise KeyError if the bundle type isn't registered.
3160 Will raise KeyError if the bundle type isn't registered.
3158
3161
3159 Will abort if the engine is known but not available.
3162 Will abort if the engine is known but not available.
3160 """
3163 """
3161 engine = self._engines[self._bundletypes[bundletype]]
3164 engine = self._engines[self._bundletypes[bundletype]]
3162 if not engine.available():
3165 if not engine.available():
3163 raise error.Abort(_('compression engine %s could not be loaded') %
3166 raise error.Abort(_('compression engine %s could not be loaded') %
3164 engine.name())
3167 engine.name())
3165 return engine
3168 return engine
3166
3169
3167 def supportedwireengines(self, role, onlyavailable=True):
3170 def supportedwireengines(self, role, onlyavailable=True):
3168 """Obtain compression engines that support the wire protocol.
3171 """Obtain compression engines that support the wire protocol.
3169
3172
3170 Returns a list of engines in prioritized order, most desired first.
3173 Returns a list of engines in prioritized order, most desired first.
3171
3174
3172 If ``onlyavailable`` is set, filter out engines that can't be
3175 If ``onlyavailable`` is set, filter out engines that can't be
3173 loaded.
3176 loaded.
3174 """
3177 """
3175 assert role in (SERVERROLE, CLIENTROLE)
3178 assert role in (SERVERROLE, CLIENTROLE)
3176
3179
3177 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3180 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3178
3181
3179 engines = [self._engines[e] for e in self._wiretypes.values()]
3182 engines = [self._engines[e] for e in self._wiretypes.values()]
3180 if onlyavailable:
3183 if onlyavailable:
3181 engines = [e for e in engines if e.available()]
3184 engines = [e for e in engines if e.available()]
3182
3185
3183 def getkey(e):
3186 def getkey(e):
3184 # Sort first by priority, highest first. In case of tie, sort
3187 # Sort first by priority, highest first. In case of tie, sort
3185 # alphabetically. This is arbitrary, but ensures output is
3188 # alphabetically. This is arbitrary, but ensures output is
3186 # stable.
3189 # stable.
3187 w = e.wireprotosupport()
3190 w = e.wireprotosupport()
3188 return -1 * getattr(w, attr), w.name
3191 return -1 * getattr(w, attr), w.name
3189
3192
3190 return list(sorted(engines, key=getkey))
3193 return list(sorted(engines, key=getkey))
3191
3194
3192 def forwiretype(self, wiretype):
3195 def forwiretype(self, wiretype):
3193 engine = self._engines[self._wiretypes[wiretype]]
3196 engine = self._engines[self._wiretypes[wiretype]]
3194 if not engine.available():
3197 if not engine.available():
3195 raise error.Abort(_('compression engine %s could not be loaded') %
3198 raise error.Abort(_('compression engine %s could not be loaded') %
3196 engine.name())
3199 engine.name())
3197 return engine
3200 return engine
3198
3201
3199 def forrevlogheader(self, header):
3202 def forrevlogheader(self, header):
3200 """Obtain a compression engine registered to a revlog header.
3203 """Obtain a compression engine registered to a revlog header.
3201
3204
3202 Will raise KeyError if the revlog header value isn't registered.
3205 Will raise KeyError if the revlog header value isn't registered.
3203 """
3206 """
3204 return self._engines[self._revlogheaders[header]]
3207 return self._engines[self._revlogheaders[header]]
3205
3208
3206 compengines = compressormanager()
3209 compengines = compressormanager()
3207
3210
3208 class compressionengine(object):
3211 class compressionengine(object):
3209 """Base class for compression engines.
3212 """Base class for compression engines.
3210
3213
3211 Compression engines must implement the interface defined by this class.
3214 Compression engines must implement the interface defined by this class.
3212 """
3215 """
3213 def name(self):
3216 def name(self):
3214 """Returns the name of the compression engine.
3217 """Returns the name of the compression engine.
3215
3218
3216 This is the key the engine is registered under.
3219 This is the key the engine is registered under.
3217
3220
3218 This method must be implemented.
3221 This method must be implemented.
3219 """
3222 """
3220 raise NotImplementedError()
3223 raise NotImplementedError()
3221
3224
3222 def available(self):
3225 def available(self):
3223 """Whether the compression engine is available.
3226 """Whether the compression engine is available.
3224
3227
3225 The intent of this method is to allow optional compression engines
3228 The intent of this method is to allow optional compression engines
3226 that may not be available in all installations (such as engines relying
3229 that may not be available in all installations (such as engines relying
3227 on C extensions that may not be present).
3230 on C extensions that may not be present).
3228 """
3231 """
3229 return True
3232 return True
3230
3233
3231 def bundletype(self):
3234 def bundletype(self):
3232 """Describes bundle identifiers for this engine.
3235 """Describes bundle identifiers for this engine.
3233
3236
3234 If this compression engine isn't supported for bundles, returns None.
3237 If this compression engine isn't supported for bundles, returns None.
3235
3238
3236 If this engine can be used for bundles, returns a 2-tuple of strings of
3239 If this engine can be used for bundles, returns a 2-tuple of strings of
3237 the user-facing "bundle spec" compression name and an internal
3240 the user-facing "bundle spec" compression name and an internal
3238 identifier used to denote the compression format within bundles. To
3241 identifier used to denote the compression format within bundles. To
3239 exclude the name from external usage, set the first element to ``None``.
3242 exclude the name from external usage, set the first element to ``None``.
3240
3243
3241 If bundle compression is supported, the class must also implement
3244 If bundle compression is supported, the class must also implement
3242 ``compressstream`` and `decompressorreader``.
3245 ``compressstream`` and `decompressorreader``.
3243 """
3246 """
3244 return None
3247 return None
3245
3248
3246 def wireprotosupport(self):
3249 def wireprotosupport(self):
3247 """Declare support for this compression format on the wire protocol.
3250 """Declare support for this compression format on the wire protocol.
3248
3251
3249 If this compression engine isn't supported for compressing wire
3252 If this compression engine isn't supported for compressing wire
3250 protocol payloads, returns None.
3253 protocol payloads, returns None.
3251
3254
3252 Otherwise, returns ``compenginewireprotosupport`` with the following
3255 Otherwise, returns ``compenginewireprotosupport`` with the following
3253 fields:
3256 fields:
3254
3257
3255 * String format identifier
3258 * String format identifier
3256 * Integer priority for the server
3259 * Integer priority for the server
3257 * Integer priority for the client
3260 * Integer priority for the client
3258
3261
3259 The integer priorities are used to order the advertisement of format
3262 The integer priorities are used to order the advertisement of format
3260 support by server and client. The highest integer is advertised
3263 support by server and client. The highest integer is advertised
3261 first. Integers with non-positive values aren't advertised.
3264 first. Integers with non-positive values aren't advertised.
3262
3265
3263 The priority values are somewhat arbitrary and only used for default
3266 The priority values are somewhat arbitrary and only used for default
3264 ordering. The relative order can be changed via config options.
3267 ordering. The relative order can be changed via config options.
3265
3268
3266 If wire protocol compression is supported, the class must also implement
3269 If wire protocol compression is supported, the class must also implement
3267 ``compressstream`` and ``decompressorreader``.
3270 ``compressstream`` and ``decompressorreader``.
3268 """
3271 """
3269 return None
3272 return None
3270
3273
3271 def revlogheader(self):
3274 def revlogheader(self):
3272 """Header added to revlog chunks that identifies this engine.
3275 """Header added to revlog chunks that identifies this engine.
3273
3276
3274 If this engine can be used to compress revlogs, this method should
3277 If this engine can be used to compress revlogs, this method should
3275 return the bytes used to identify chunks compressed with this engine.
3278 return the bytes used to identify chunks compressed with this engine.
3276 Else, the method should return ``None`` to indicate it does not
3279 Else, the method should return ``None`` to indicate it does not
3277 participate in revlog compression.
3280 participate in revlog compression.
3278 """
3281 """
3279 return None
3282 return None
3280
3283
3281 def compressstream(self, it, opts=None):
3284 def compressstream(self, it, opts=None):
3282 """Compress an iterator of chunks.
3285 """Compress an iterator of chunks.
3283
3286
3284 The method receives an iterator (ideally a generator) of chunks of
3287 The method receives an iterator (ideally a generator) of chunks of
3285 bytes to be compressed. It returns an iterator (ideally a generator)
3288 bytes to be compressed. It returns an iterator (ideally a generator)
3286 of bytes of chunks representing the compressed output.
3289 of bytes of chunks representing the compressed output.
3287
3290
3288 Optionally accepts an argument defining how to perform compression.
3291 Optionally accepts an argument defining how to perform compression.
3289 Each engine treats this argument differently.
3292 Each engine treats this argument differently.
3290 """
3293 """
3291 raise NotImplementedError()
3294 raise NotImplementedError()
3292
3295
3293 def decompressorreader(self, fh):
3296 def decompressorreader(self, fh):
3294 """Perform decompression on a file object.
3297 """Perform decompression on a file object.
3295
3298
3296 Argument is an object with a ``read(size)`` method that returns
3299 Argument is an object with a ``read(size)`` method that returns
3297 compressed data. Return value is an object with a ``read(size)`` that
3300 compressed data. Return value is an object with a ``read(size)`` that
3298 returns uncompressed data.
3301 returns uncompressed data.
3299 """
3302 """
3300 raise NotImplementedError()
3303 raise NotImplementedError()
3301
3304
3302 def revlogcompressor(self, opts=None):
3305 def revlogcompressor(self, opts=None):
3303 """Obtain an object that can be used to compress revlog entries.
3306 """Obtain an object that can be used to compress revlog entries.
3304
3307
3305 The object has a ``compress(data)`` method that compresses binary
3308 The object has a ``compress(data)`` method that compresses binary
3306 data. This method returns compressed binary data or ``None`` if
3309 data. This method returns compressed binary data or ``None`` if
3307 the data could not be compressed (too small, not compressible, etc).
3310 the data could not be compressed (too small, not compressible, etc).
3308 The returned data should have a header uniquely identifying this
3311 The returned data should have a header uniquely identifying this
3309 compression format so decompression can be routed to this engine.
3312 compression format so decompression can be routed to this engine.
3310 This header should be identified by the ``revlogheader()`` return
3313 This header should be identified by the ``revlogheader()`` return
3311 value.
3314 value.
3312
3315
3313 The object has a ``decompress(data)`` method that decompresses
3316 The object has a ``decompress(data)`` method that decompresses
3314 data. The method will only be called if ``data`` begins with
3317 data. The method will only be called if ``data`` begins with
3315 ``revlogheader()``. The method should return the raw, uncompressed
3318 ``revlogheader()``. The method should return the raw, uncompressed
3316 data or raise a ``RevlogError``.
3319 data or raise a ``RevlogError``.
3317
3320
3318 The object is reusable but is not thread safe.
3321 The object is reusable but is not thread safe.
3319 """
3322 """
3320 raise NotImplementedError()
3323 raise NotImplementedError()
3321
3324
3322 class _zlibengine(compressionengine):
3325 class _zlibengine(compressionengine):
3323 def name(self):
3326 def name(self):
3324 return 'zlib'
3327 return 'zlib'
3325
3328
3326 def bundletype(self):
3329 def bundletype(self):
3327 return 'gzip', 'GZ'
3330 return 'gzip', 'GZ'
3328
3331
3329 def wireprotosupport(self):
3332 def wireprotosupport(self):
3330 return compewireprotosupport('zlib', 20, 20)
3333 return compewireprotosupport('zlib', 20, 20)
3331
3334
3332 def revlogheader(self):
3335 def revlogheader(self):
3333 return 'x'
3336 return 'x'
3334
3337
3335 def compressstream(self, it, opts=None):
3338 def compressstream(self, it, opts=None):
3336 opts = opts or {}
3339 opts = opts or {}
3337
3340
3338 z = zlib.compressobj(opts.get('level', -1))
3341 z = zlib.compressobj(opts.get('level', -1))
3339 for chunk in it:
3342 for chunk in it:
3340 data = z.compress(chunk)
3343 data = z.compress(chunk)
3341 # Not all calls to compress emit data. It is cheaper to inspect
3344 # Not all calls to compress emit data. It is cheaper to inspect
3342 # here than to feed empty chunks through generator.
3345 # here than to feed empty chunks through generator.
3343 if data:
3346 if data:
3344 yield data
3347 yield data
3345
3348
3346 yield z.flush()
3349 yield z.flush()
3347
3350
3348 def decompressorreader(self, fh):
3351 def decompressorreader(self, fh):
3349 def gen():
3352 def gen():
3350 d = zlib.decompressobj()
3353 d = zlib.decompressobj()
3351 for chunk in filechunkiter(fh):
3354 for chunk in filechunkiter(fh):
3352 while chunk:
3355 while chunk:
3353 # Limit output size to limit memory.
3356 # Limit output size to limit memory.
3354 yield d.decompress(chunk, 2 ** 18)
3357 yield d.decompress(chunk, 2 ** 18)
3355 chunk = d.unconsumed_tail
3358 chunk = d.unconsumed_tail
3356
3359
3357 return chunkbuffer(gen())
3360 return chunkbuffer(gen())
3358
3361
3359 class zlibrevlogcompressor(object):
3362 class zlibrevlogcompressor(object):
3360 def compress(self, data):
3363 def compress(self, data):
3361 insize = len(data)
3364 insize = len(data)
3362 # Caller handles empty input case.
3365 # Caller handles empty input case.
3363 assert insize > 0
3366 assert insize > 0
3364
3367
3365 if insize < 44:
3368 if insize < 44:
3366 return None
3369 return None
3367
3370
3368 elif insize <= 1000000:
3371 elif insize <= 1000000:
3369 compressed = zlib.compress(data)
3372 compressed = zlib.compress(data)
3370 if len(compressed) < insize:
3373 if len(compressed) < insize:
3371 return compressed
3374 return compressed
3372 return None
3375 return None
3373
3376
3374 # zlib makes an internal copy of the input buffer, doubling
3377 # zlib makes an internal copy of the input buffer, doubling
3375 # memory usage for large inputs. So do streaming compression
3378 # memory usage for large inputs. So do streaming compression
3376 # on large inputs.
3379 # on large inputs.
3377 else:
3380 else:
3378 z = zlib.compressobj()
3381 z = zlib.compressobj()
3379 parts = []
3382 parts = []
3380 pos = 0
3383 pos = 0
3381 while pos < insize:
3384 while pos < insize:
3382 pos2 = pos + 2**20
3385 pos2 = pos + 2**20
3383 parts.append(z.compress(data[pos:pos2]))
3386 parts.append(z.compress(data[pos:pos2]))
3384 pos = pos2
3387 pos = pos2
3385 parts.append(z.flush())
3388 parts.append(z.flush())
3386
3389
3387 if sum(map(len, parts)) < insize:
3390 if sum(map(len, parts)) < insize:
3388 return ''.join(parts)
3391 return ''.join(parts)
3389 return None
3392 return None
3390
3393
3391 def decompress(self, data):
3394 def decompress(self, data):
3392 try:
3395 try:
3393 return zlib.decompress(data)
3396 return zlib.decompress(data)
3394 except zlib.error as e:
3397 except zlib.error as e:
3395 raise error.RevlogError(_('revlog decompress error: %s') %
3398 raise error.RevlogError(_('revlog decompress error: %s') %
3396 str(e))
3399 str(e))
3397
3400
3398 def revlogcompressor(self, opts=None):
3401 def revlogcompressor(self, opts=None):
3399 return self.zlibrevlogcompressor()
3402 return self.zlibrevlogcompressor()
3400
3403
3401 compengines.register(_zlibengine())
3404 compengines.register(_zlibengine())
3402
3405
3403 class _bz2engine(compressionengine):
3406 class _bz2engine(compressionengine):
3404 def name(self):
3407 def name(self):
3405 return 'bz2'
3408 return 'bz2'
3406
3409
3407 def bundletype(self):
3410 def bundletype(self):
3408 return 'bzip2', 'BZ'
3411 return 'bzip2', 'BZ'
3409
3412
3410 # We declare a protocol name but don't advertise by default because
3413 # We declare a protocol name but don't advertise by default because
3411 # it is slow.
3414 # it is slow.
3412 def wireprotosupport(self):
3415 def wireprotosupport(self):
3413 return compewireprotosupport('bzip2', 0, 0)
3416 return compewireprotosupport('bzip2', 0, 0)
3414
3417
3415 def compressstream(self, it, opts=None):
3418 def compressstream(self, it, opts=None):
3416 opts = opts or {}
3419 opts = opts or {}
3417 z = bz2.BZ2Compressor(opts.get('level', 9))
3420 z = bz2.BZ2Compressor(opts.get('level', 9))
3418 for chunk in it:
3421 for chunk in it:
3419 data = z.compress(chunk)
3422 data = z.compress(chunk)
3420 if data:
3423 if data:
3421 yield data
3424 yield data
3422
3425
3423 yield z.flush()
3426 yield z.flush()
3424
3427
3425 def decompressorreader(self, fh):
3428 def decompressorreader(self, fh):
3426 def gen():
3429 def gen():
3427 d = bz2.BZ2Decompressor()
3430 d = bz2.BZ2Decompressor()
3428 for chunk in filechunkiter(fh):
3431 for chunk in filechunkiter(fh):
3429 yield d.decompress(chunk)
3432 yield d.decompress(chunk)
3430
3433
3431 return chunkbuffer(gen())
3434 return chunkbuffer(gen())
3432
3435
3433 compengines.register(_bz2engine())
3436 compengines.register(_bz2engine())
3434
3437
3435 class _truncatedbz2engine(compressionengine):
3438 class _truncatedbz2engine(compressionengine):
3436 def name(self):
3439 def name(self):
3437 return 'bz2truncated'
3440 return 'bz2truncated'
3438
3441
3439 def bundletype(self):
3442 def bundletype(self):
3440 return None, '_truncatedBZ'
3443 return None, '_truncatedBZ'
3441
3444
3442 # We don't implement compressstream because it is hackily handled elsewhere.
3445 # We don't implement compressstream because it is hackily handled elsewhere.
3443
3446
3444 def decompressorreader(self, fh):
3447 def decompressorreader(self, fh):
3445 def gen():
3448 def gen():
3446 # The input stream doesn't have the 'BZ' header. So add it back.
3449 # The input stream doesn't have the 'BZ' header. So add it back.
3447 d = bz2.BZ2Decompressor()
3450 d = bz2.BZ2Decompressor()
3448 d.decompress('BZ')
3451 d.decompress('BZ')
3449 for chunk in filechunkiter(fh):
3452 for chunk in filechunkiter(fh):
3450 yield d.decompress(chunk)
3453 yield d.decompress(chunk)
3451
3454
3452 return chunkbuffer(gen())
3455 return chunkbuffer(gen())
3453
3456
3454 compengines.register(_truncatedbz2engine())
3457 compengines.register(_truncatedbz2engine())
3455
3458
3456 class _noopengine(compressionengine):
3459 class _noopengine(compressionengine):
3457 def name(self):
3460 def name(self):
3458 return 'none'
3461 return 'none'
3459
3462
3460 def bundletype(self):
3463 def bundletype(self):
3461 return 'none', 'UN'
3464 return 'none', 'UN'
3462
3465
3463 # Clients always support uncompressed payloads. Servers don't because
3466 # Clients always support uncompressed payloads. Servers don't because
3464 # unless you are on a fast network, uncompressed payloads can easily
3467 # unless you are on a fast network, uncompressed payloads can easily
3465 # saturate your network pipe.
3468 # saturate your network pipe.
3466 def wireprotosupport(self):
3469 def wireprotosupport(self):
3467 return compewireprotosupport('none', 0, 10)
3470 return compewireprotosupport('none', 0, 10)
3468
3471
3469 # We don't implement revlogheader because it is handled specially
3472 # We don't implement revlogheader because it is handled specially
3470 # in the revlog class.
3473 # in the revlog class.
3471
3474
3472 def compressstream(self, it, opts=None):
3475 def compressstream(self, it, opts=None):
3473 return it
3476 return it
3474
3477
3475 def decompressorreader(self, fh):
3478 def decompressorreader(self, fh):
3476 return fh
3479 return fh
3477
3480
3478 class nooprevlogcompressor(object):
3481 class nooprevlogcompressor(object):
3479 def compress(self, data):
3482 def compress(self, data):
3480 return None
3483 return None
3481
3484
3482 def revlogcompressor(self, opts=None):
3485 def revlogcompressor(self, opts=None):
3483 return self.nooprevlogcompressor()
3486 return self.nooprevlogcompressor()
3484
3487
3485 compengines.register(_noopengine())
3488 compengines.register(_noopengine())
3486
3489
3487 class _zstdengine(compressionengine):
3490 class _zstdengine(compressionengine):
3488 def name(self):
3491 def name(self):
3489 return 'zstd'
3492 return 'zstd'
3490
3493
3491 @propertycache
3494 @propertycache
3492 def _module(self):
3495 def _module(self):
3493 # Not all installs have the zstd module available. So defer importing
3496 # Not all installs have the zstd module available. So defer importing
3494 # until first access.
3497 # until first access.
3495 try:
3498 try:
3496 from . import zstd
3499 from . import zstd
3497 # Force delayed import.
3500 # Force delayed import.
3498 zstd.__version__
3501 zstd.__version__
3499 return zstd
3502 return zstd
3500 except ImportError:
3503 except ImportError:
3501 return None
3504 return None
3502
3505
3503 def available(self):
3506 def available(self):
3504 return bool(self._module)
3507 return bool(self._module)
3505
3508
3506 def bundletype(self):
3509 def bundletype(self):
3507 return 'zstd', 'ZS'
3510 return 'zstd', 'ZS'
3508
3511
3509 def wireprotosupport(self):
3512 def wireprotosupport(self):
3510 return compewireprotosupport('zstd', 50, 50)
3513 return compewireprotosupport('zstd', 50, 50)
3511
3514
3512 def revlogheader(self):
3515 def revlogheader(self):
3513 return '\x28'
3516 return '\x28'
3514
3517
3515 def compressstream(self, it, opts=None):
3518 def compressstream(self, it, opts=None):
3516 opts = opts or {}
3519 opts = opts or {}
3517 # zstd level 3 is almost always significantly faster than zlib
3520 # zstd level 3 is almost always significantly faster than zlib
3518 # while providing no worse compression. It strikes a good balance
3521 # while providing no worse compression. It strikes a good balance
3519 # between speed and compression.
3522 # between speed and compression.
3520 level = opts.get('level', 3)
3523 level = opts.get('level', 3)
3521
3524
3522 zstd = self._module
3525 zstd = self._module
3523 z = zstd.ZstdCompressor(level=level).compressobj()
3526 z = zstd.ZstdCompressor(level=level).compressobj()
3524 for chunk in it:
3527 for chunk in it:
3525 data = z.compress(chunk)
3528 data = z.compress(chunk)
3526 if data:
3529 if data:
3527 yield data
3530 yield data
3528
3531
3529 yield z.flush()
3532 yield z.flush()
3530
3533
3531 def decompressorreader(self, fh):
3534 def decompressorreader(self, fh):
3532 zstd = self._module
3535 zstd = self._module
3533 dctx = zstd.ZstdDecompressor()
3536 dctx = zstd.ZstdDecompressor()
3534 return chunkbuffer(dctx.read_from(fh))
3537 return chunkbuffer(dctx.read_from(fh))
3535
3538
3536 class zstdrevlogcompressor(object):
3539 class zstdrevlogcompressor(object):
3537 def __init__(self, zstd, level=3):
3540 def __init__(self, zstd, level=3):
3538 # Writing the content size adds a few bytes to the output. However,
3541 # Writing the content size adds a few bytes to the output. However,
3539 # it allows decompression to be more optimal since we can
3542 # it allows decompression to be more optimal since we can
3540 # pre-allocate a buffer to hold the result.
3543 # pre-allocate a buffer to hold the result.
3541 self._cctx = zstd.ZstdCompressor(level=level,
3544 self._cctx = zstd.ZstdCompressor(level=level,
3542 write_content_size=True)
3545 write_content_size=True)
3543 self._dctx = zstd.ZstdDecompressor()
3546 self._dctx = zstd.ZstdDecompressor()
3544 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3547 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3545 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3548 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3546
3549
3547 def compress(self, data):
3550 def compress(self, data):
3548 insize = len(data)
3551 insize = len(data)
3549 # Caller handles empty input case.
3552 # Caller handles empty input case.
3550 assert insize > 0
3553 assert insize > 0
3551
3554
3552 if insize < 50:
3555 if insize < 50:
3553 return None
3556 return None
3554
3557
3555 elif insize <= 1000000:
3558 elif insize <= 1000000:
3556 compressed = self._cctx.compress(data)
3559 compressed = self._cctx.compress(data)
3557 if len(compressed) < insize:
3560 if len(compressed) < insize:
3558 return compressed
3561 return compressed
3559 return None
3562 return None
3560 else:
3563 else:
3561 z = self._cctx.compressobj()
3564 z = self._cctx.compressobj()
3562 chunks = []
3565 chunks = []
3563 pos = 0
3566 pos = 0
3564 while pos < insize:
3567 while pos < insize:
3565 pos2 = pos + self._compinsize
3568 pos2 = pos + self._compinsize
3566 chunk = z.compress(data[pos:pos2])
3569 chunk = z.compress(data[pos:pos2])
3567 if chunk:
3570 if chunk:
3568 chunks.append(chunk)
3571 chunks.append(chunk)
3569 pos = pos2
3572 pos = pos2
3570 chunks.append(z.flush())
3573 chunks.append(z.flush())
3571
3574
3572 if sum(map(len, chunks)) < insize:
3575 if sum(map(len, chunks)) < insize:
3573 return ''.join(chunks)
3576 return ''.join(chunks)
3574 return None
3577 return None
3575
3578
3576 def decompress(self, data):
3579 def decompress(self, data):
3577 insize = len(data)
3580 insize = len(data)
3578
3581
3579 try:
3582 try:
3580 # This was measured to be faster than other streaming
3583 # This was measured to be faster than other streaming
3581 # decompressors.
3584 # decompressors.
3582 dobj = self._dctx.decompressobj()
3585 dobj = self._dctx.decompressobj()
3583 chunks = []
3586 chunks = []
3584 pos = 0
3587 pos = 0
3585 while pos < insize:
3588 while pos < insize:
3586 pos2 = pos + self._decompinsize
3589 pos2 = pos + self._decompinsize
3587 chunk = dobj.decompress(data[pos:pos2])
3590 chunk = dobj.decompress(data[pos:pos2])
3588 if chunk:
3591 if chunk:
3589 chunks.append(chunk)
3592 chunks.append(chunk)
3590 pos = pos2
3593 pos = pos2
3591 # Frame should be exhausted, so no finish() API.
3594 # Frame should be exhausted, so no finish() API.
3592
3595
3593 return ''.join(chunks)
3596 return ''.join(chunks)
3594 except Exception as e:
3597 except Exception as e:
3595 raise error.RevlogError(_('revlog decompress error: %s') %
3598 raise error.RevlogError(_('revlog decompress error: %s') %
3596 str(e))
3599 str(e))
3597
3600
3598 def revlogcompressor(self, opts=None):
3601 def revlogcompressor(self, opts=None):
3599 opts = opts or {}
3602 opts = opts or {}
3600 return self.zstdrevlogcompressor(self._module,
3603 return self.zstdrevlogcompressor(self._module,
3601 level=opts.get('level', 3))
3604 level=opts.get('level', 3))
3602
3605
3603 compengines.register(_zstdengine())
3606 compengines.register(_zstdengine())
3604
3607
3605 # convenient shortcut
3608 # convenient shortcut
3606 dst = debugstacktrace
3609 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now