##// END OF EJS Templates
util: checknlink should remove file it creates if an exception occurs...
Tony Tung -
r29832:bac1829e default
parent child Browse files
Show More
@@ -1,2900 +1,2904 b''
1 # util.py - Mercurial utility functions and platform specific implementations
1 # util.py - Mercurial utility functions and platform specific implementations
2 #
2 #
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 #
6 #
7 # This software may be used and distributed according to the terms of the
7 # This software may be used and distributed according to the terms of the
8 # GNU General Public License version 2 or any later version.
8 # GNU General Public License version 2 or any later version.
9
9
10 """Mercurial utility functions and platform specific implementations.
10 """Mercurial utility functions and platform specific implementations.
11
11
12 This contains helper routines that are independent of the SCM core and
12 This contains helper routines that are independent of the SCM core and
13 hide platform-specific details from the core.
13 hide platform-specific details from the core.
14 """
14 """
15
15
16 from __future__ import absolute_import
16 from __future__ import absolute_import
17
17
18 import bz2
18 import bz2
19 import calendar
19 import calendar
20 import collections
20 import collections
21 import datetime
21 import datetime
22 import errno
22 import errno
23 import gc
23 import gc
24 import hashlib
24 import hashlib
25 import imp
25 import imp
26 import os
26 import os
27 import re as remod
27 import re as remod
28 import shutil
28 import shutil
29 import signal
29 import signal
30 import socket
30 import socket
31 import subprocess
31 import subprocess
32 import sys
32 import sys
33 import tempfile
33 import tempfile
34 import textwrap
34 import textwrap
35 import time
35 import time
36 import traceback
36 import traceback
37 import zlib
37 import zlib
38
38
39 from . import (
39 from . import (
40 encoding,
40 encoding,
41 error,
41 error,
42 i18n,
42 i18n,
43 osutil,
43 osutil,
44 parsers,
44 parsers,
45 pycompat,
45 pycompat,
46 )
46 )
47
47
48 for attr in (
48 for attr in (
49 'empty',
49 'empty',
50 'httplib',
50 'httplib',
51 'httpserver',
51 'httpserver',
52 'pickle',
52 'pickle',
53 'queue',
53 'queue',
54 'urlerr',
54 'urlerr',
55 'urlparse',
55 'urlparse',
56 # we do import urlreq, but we do it outside the loop
56 # we do import urlreq, but we do it outside the loop
57 #'urlreq',
57 #'urlreq',
58 'stringio',
58 'stringio',
59 'socketserver',
59 'socketserver',
60 'xmlrpclib',
60 'xmlrpclib',
61 ):
61 ):
62 globals()[attr] = getattr(pycompat, attr)
62 globals()[attr] = getattr(pycompat, attr)
63
63
64 # This line is to make pyflakes happy:
64 # This line is to make pyflakes happy:
65 urlreq = pycompat.urlreq
65 urlreq = pycompat.urlreq
66
66
67 if os.name == 'nt':
67 if os.name == 'nt':
68 from . import windows as platform
68 from . import windows as platform
69 else:
69 else:
70 from . import posix as platform
70 from . import posix as platform
71
71
72 _ = i18n._
72 _ = i18n._
73
73
74 bindunixsocket = platform.bindunixsocket
74 bindunixsocket = platform.bindunixsocket
75 cachestat = platform.cachestat
75 cachestat = platform.cachestat
76 checkexec = platform.checkexec
76 checkexec = platform.checkexec
77 checklink = platform.checklink
77 checklink = platform.checklink
78 copymode = platform.copymode
78 copymode = platform.copymode
79 executablepath = platform.executablepath
79 executablepath = platform.executablepath
80 expandglobs = platform.expandglobs
80 expandglobs = platform.expandglobs
81 explainexit = platform.explainexit
81 explainexit = platform.explainexit
82 findexe = platform.findexe
82 findexe = platform.findexe
83 gethgcmd = platform.gethgcmd
83 gethgcmd = platform.gethgcmd
84 getuser = platform.getuser
84 getuser = platform.getuser
85 getpid = os.getpid
85 getpid = os.getpid
86 groupmembers = platform.groupmembers
86 groupmembers = platform.groupmembers
87 groupname = platform.groupname
87 groupname = platform.groupname
88 hidewindow = platform.hidewindow
88 hidewindow = platform.hidewindow
89 isexec = platform.isexec
89 isexec = platform.isexec
90 isowner = platform.isowner
90 isowner = platform.isowner
91 localpath = platform.localpath
91 localpath = platform.localpath
92 lookupreg = platform.lookupreg
92 lookupreg = platform.lookupreg
93 makedir = platform.makedir
93 makedir = platform.makedir
94 nlinks = platform.nlinks
94 nlinks = platform.nlinks
95 normpath = platform.normpath
95 normpath = platform.normpath
96 normcase = platform.normcase
96 normcase = platform.normcase
97 normcasespec = platform.normcasespec
97 normcasespec = platform.normcasespec
98 normcasefallback = platform.normcasefallback
98 normcasefallback = platform.normcasefallback
99 openhardlinks = platform.openhardlinks
99 openhardlinks = platform.openhardlinks
100 oslink = platform.oslink
100 oslink = platform.oslink
101 parsepatchoutput = platform.parsepatchoutput
101 parsepatchoutput = platform.parsepatchoutput
102 pconvert = platform.pconvert
102 pconvert = platform.pconvert
103 poll = platform.poll
103 poll = platform.poll
104 popen = platform.popen
104 popen = platform.popen
105 posixfile = platform.posixfile
105 posixfile = platform.posixfile
106 quotecommand = platform.quotecommand
106 quotecommand = platform.quotecommand
107 readpipe = platform.readpipe
107 readpipe = platform.readpipe
108 rename = platform.rename
108 rename = platform.rename
109 removedirs = platform.removedirs
109 removedirs = platform.removedirs
110 samedevice = platform.samedevice
110 samedevice = platform.samedevice
111 samefile = platform.samefile
111 samefile = platform.samefile
112 samestat = platform.samestat
112 samestat = platform.samestat
113 setbinary = platform.setbinary
113 setbinary = platform.setbinary
114 setflags = platform.setflags
114 setflags = platform.setflags
115 setsignalhandler = platform.setsignalhandler
115 setsignalhandler = platform.setsignalhandler
116 shellquote = platform.shellquote
116 shellquote = platform.shellquote
117 spawndetached = platform.spawndetached
117 spawndetached = platform.spawndetached
118 split = platform.split
118 split = platform.split
119 sshargs = platform.sshargs
119 sshargs = platform.sshargs
120 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
120 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
121 statisexec = platform.statisexec
121 statisexec = platform.statisexec
122 statislink = platform.statislink
122 statislink = platform.statislink
123 termwidth = platform.termwidth
123 termwidth = platform.termwidth
124 testpid = platform.testpid
124 testpid = platform.testpid
125 umask = platform.umask
125 umask = platform.umask
126 unlink = platform.unlink
126 unlink = platform.unlink
127 unlinkpath = platform.unlinkpath
127 unlinkpath = platform.unlinkpath
128 username = platform.username
128 username = platform.username
129
129
130 # Python compatibility
130 # Python compatibility
131
131
132 _notset = object()
132 _notset = object()
133
133
134 # disable Python's problematic floating point timestamps (issue4836)
134 # disable Python's problematic floating point timestamps (issue4836)
135 # (Python hypocritically says you shouldn't change this behavior in
135 # (Python hypocritically says you shouldn't change this behavior in
136 # libraries, and sure enough Mercurial is not a library.)
136 # libraries, and sure enough Mercurial is not a library.)
137 os.stat_float_times(False)
137 os.stat_float_times(False)
138
138
139 def safehasattr(thing, attr):
139 def safehasattr(thing, attr):
140 return getattr(thing, attr, _notset) is not _notset
140 return getattr(thing, attr, _notset) is not _notset
141
141
142 DIGESTS = {
142 DIGESTS = {
143 'md5': hashlib.md5,
143 'md5': hashlib.md5,
144 'sha1': hashlib.sha1,
144 'sha1': hashlib.sha1,
145 'sha512': hashlib.sha512,
145 'sha512': hashlib.sha512,
146 }
146 }
147 # List of digest types from strongest to weakest
147 # List of digest types from strongest to weakest
148 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
148 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
149
149
150 for k in DIGESTS_BY_STRENGTH:
150 for k in DIGESTS_BY_STRENGTH:
151 assert k in DIGESTS
151 assert k in DIGESTS
152
152
153 class digester(object):
153 class digester(object):
154 """helper to compute digests.
154 """helper to compute digests.
155
155
156 This helper can be used to compute one or more digests given their name.
156 This helper can be used to compute one or more digests given their name.
157
157
158 >>> d = digester(['md5', 'sha1'])
158 >>> d = digester(['md5', 'sha1'])
159 >>> d.update('foo')
159 >>> d.update('foo')
160 >>> [k for k in sorted(d)]
160 >>> [k for k in sorted(d)]
161 ['md5', 'sha1']
161 ['md5', 'sha1']
162 >>> d['md5']
162 >>> d['md5']
163 'acbd18db4cc2f85cedef654fccc4a4d8'
163 'acbd18db4cc2f85cedef654fccc4a4d8'
164 >>> d['sha1']
164 >>> d['sha1']
165 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
165 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
166 >>> digester.preferred(['md5', 'sha1'])
166 >>> digester.preferred(['md5', 'sha1'])
167 'sha1'
167 'sha1'
168 """
168 """
169
169
170 def __init__(self, digests, s=''):
170 def __init__(self, digests, s=''):
171 self._hashes = {}
171 self._hashes = {}
172 for k in digests:
172 for k in digests:
173 if k not in DIGESTS:
173 if k not in DIGESTS:
174 raise Abort(_('unknown digest type: %s') % k)
174 raise Abort(_('unknown digest type: %s') % k)
175 self._hashes[k] = DIGESTS[k]()
175 self._hashes[k] = DIGESTS[k]()
176 if s:
176 if s:
177 self.update(s)
177 self.update(s)
178
178
179 def update(self, data):
179 def update(self, data):
180 for h in self._hashes.values():
180 for h in self._hashes.values():
181 h.update(data)
181 h.update(data)
182
182
183 def __getitem__(self, key):
183 def __getitem__(self, key):
184 if key not in DIGESTS:
184 if key not in DIGESTS:
185 raise Abort(_('unknown digest type: %s') % k)
185 raise Abort(_('unknown digest type: %s') % k)
186 return self._hashes[key].hexdigest()
186 return self._hashes[key].hexdigest()
187
187
188 def __iter__(self):
188 def __iter__(self):
189 return iter(self._hashes)
189 return iter(self._hashes)
190
190
191 @staticmethod
191 @staticmethod
192 def preferred(supported):
192 def preferred(supported):
193 """returns the strongest digest type in both supported and DIGESTS."""
193 """returns the strongest digest type in both supported and DIGESTS."""
194
194
195 for k in DIGESTS_BY_STRENGTH:
195 for k in DIGESTS_BY_STRENGTH:
196 if k in supported:
196 if k in supported:
197 return k
197 return k
198 return None
198 return None
199
199
200 class digestchecker(object):
200 class digestchecker(object):
201 """file handle wrapper that additionally checks content against a given
201 """file handle wrapper that additionally checks content against a given
202 size and digests.
202 size and digests.
203
203
204 d = digestchecker(fh, size, {'md5': '...'})
204 d = digestchecker(fh, size, {'md5': '...'})
205
205
206 When multiple digests are given, all of them are validated.
206 When multiple digests are given, all of them are validated.
207 """
207 """
208
208
209 def __init__(self, fh, size, digests):
209 def __init__(self, fh, size, digests):
210 self._fh = fh
210 self._fh = fh
211 self._size = size
211 self._size = size
212 self._got = 0
212 self._got = 0
213 self._digests = dict(digests)
213 self._digests = dict(digests)
214 self._digester = digester(self._digests.keys())
214 self._digester = digester(self._digests.keys())
215
215
216 def read(self, length=-1):
216 def read(self, length=-1):
217 content = self._fh.read(length)
217 content = self._fh.read(length)
218 self._digester.update(content)
218 self._digester.update(content)
219 self._got += len(content)
219 self._got += len(content)
220 return content
220 return content
221
221
222 def validate(self):
222 def validate(self):
223 if self._size != self._got:
223 if self._size != self._got:
224 raise Abort(_('size mismatch: expected %d, got %d') %
224 raise Abort(_('size mismatch: expected %d, got %d') %
225 (self._size, self._got))
225 (self._size, self._got))
226 for k, v in self._digests.items():
226 for k, v in self._digests.items():
227 if v != self._digester[k]:
227 if v != self._digester[k]:
228 # i18n: first parameter is a digest name
228 # i18n: first parameter is a digest name
229 raise Abort(_('%s mismatch: expected %s, got %s') %
229 raise Abort(_('%s mismatch: expected %s, got %s') %
230 (k, v, self._digester[k]))
230 (k, v, self._digester[k]))
231
231
232 try:
232 try:
233 buffer = buffer
233 buffer = buffer
234 except NameError:
234 except NameError:
235 if sys.version_info[0] < 3:
235 if sys.version_info[0] < 3:
236 def buffer(sliceable, offset=0):
236 def buffer(sliceable, offset=0):
237 return sliceable[offset:]
237 return sliceable[offset:]
238 else:
238 else:
239 def buffer(sliceable, offset=0):
239 def buffer(sliceable, offset=0):
240 return memoryview(sliceable)[offset:]
240 return memoryview(sliceable)[offset:]
241
241
242 closefds = os.name == 'posix'
242 closefds = os.name == 'posix'
243
243
244 _chunksize = 4096
244 _chunksize = 4096
245
245
246 class bufferedinputpipe(object):
246 class bufferedinputpipe(object):
247 """a manually buffered input pipe
247 """a manually buffered input pipe
248
248
249 Python will not let us use buffered IO and lazy reading with 'polling' at
249 Python will not let us use buffered IO and lazy reading with 'polling' at
250 the same time. We cannot probe the buffer state and select will not detect
250 the same time. We cannot probe the buffer state and select will not detect
251 that data are ready to read if they are already buffered.
251 that data are ready to read if they are already buffered.
252
252
253 This class let us work around that by implementing its own buffering
253 This class let us work around that by implementing its own buffering
254 (allowing efficient readline) while offering a way to know if the buffer is
254 (allowing efficient readline) while offering a way to know if the buffer is
255 empty from the output (allowing collaboration of the buffer with polling).
255 empty from the output (allowing collaboration of the buffer with polling).
256
256
257 This class lives in the 'util' module because it makes use of the 'os'
257 This class lives in the 'util' module because it makes use of the 'os'
258 module from the python stdlib.
258 module from the python stdlib.
259 """
259 """
260
260
261 def __init__(self, input):
261 def __init__(self, input):
262 self._input = input
262 self._input = input
263 self._buffer = []
263 self._buffer = []
264 self._eof = False
264 self._eof = False
265 self._lenbuf = 0
265 self._lenbuf = 0
266
266
267 @property
267 @property
268 def hasbuffer(self):
268 def hasbuffer(self):
269 """True is any data is currently buffered
269 """True is any data is currently buffered
270
270
271 This will be used externally a pre-step for polling IO. If there is
271 This will be used externally a pre-step for polling IO. If there is
272 already data then no polling should be set in place."""
272 already data then no polling should be set in place."""
273 return bool(self._buffer)
273 return bool(self._buffer)
274
274
275 @property
275 @property
276 def closed(self):
276 def closed(self):
277 return self._input.closed
277 return self._input.closed
278
278
279 def fileno(self):
279 def fileno(self):
280 return self._input.fileno()
280 return self._input.fileno()
281
281
282 def close(self):
282 def close(self):
283 return self._input.close()
283 return self._input.close()
284
284
285 def read(self, size):
285 def read(self, size):
286 while (not self._eof) and (self._lenbuf < size):
286 while (not self._eof) and (self._lenbuf < size):
287 self._fillbuffer()
287 self._fillbuffer()
288 return self._frombuffer(size)
288 return self._frombuffer(size)
289
289
290 def readline(self, *args, **kwargs):
290 def readline(self, *args, **kwargs):
291 if 1 < len(self._buffer):
291 if 1 < len(self._buffer):
292 # this should not happen because both read and readline end with a
292 # this should not happen because both read and readline end with a
293 # _frombuffer call that collapse it.
293 # _frombuffer call that collapse it.
294 self._buffer = [''.join(self._buffer)]
294 self._buffer = [''.join(self._buffer)]
295 self._lenbuf = len(self._buffer[0])
295 self._lenbuf = len(self._buffer[0])
296 lfi = -1
296 lfi = -1
297 if self._buffer:
297 if self._buffer:
298 lfi = self._buffer[-1].find('\n')
298 lfi = self._buffer[-1].find('\n')
299 while (not self._eof) and lfi < 0:
299 while (not self._eof) and lfi < 0:
300 self._fillbuffer()
300 self._fillbuffer()
301 if self._buffer:
301 if self._buffer:
302 lfi = self._buffer[-1].find('\n')
302 lfi = self._buffer[-1].find('\n')
303 size = lfi + 1
303 size = lfi + 1
304 if lfi < 0: # end of file
304 if lfi < 0: # end of file
305 size = self._lenbuf
305 size = self._lenbuf
306 elif 1 < len(self._buffer):
306 elif 1 < len(self._buffer):
307 # we need to take previous chunks into account
307 # we need to take previous chunks into account
308 size += self._lenbuf - len(self._buffer[-1])
308 size += self._lenbuf - len(self._buffer[-1])
309 return self._frombuffer(size)
309 return self._frombuffer(size)
310
310
311 def _frombuffer(self, size):
311 def _frombuffer(self, size):
312 """return at most 'size' data from the buffer
312 """return at most 'size' data from the buffer
313
313
314 The data are removed from the buffer."""
314 The data are removed from the buffer."""
315 if size == 0 or not self._buffer:
315 if size == 0 or not self._buffer:
316 return ''
316 return ''
317 buf = self._buffer[0]
317 buf = self._buffer[0]
318 if 1 < len(self._buffer):
318 if 1 < len(self._buffer):
319 buf = ''.join(self._buffer)
319 buf = ''.join(self._buffer)
320
320
321 data = buf[:size]
321 data = buf[:size]
322 buf = buf[len(data):]
322 buf = buf[len(data):]
323 if buf:
323 if buf:
324 self._buffer = [buf]
324 self._buffer = [buf]
325 self._lenbuf = len(buf)
325 self._lenbuf = len(buf)
326 else:
326 else:
327 self._buffer = []
327 self._buffer = []
328 self._lenbuf = 0
328 self._lenbuf = 0
329 return data
329 return data
330
330
331 def _fillbuffer(self):
331 def _fillbuffer(self):
332 """read data to the buffer"""
332 """read data to the buffer"""
333 data = os.read(self._input.fileno(), _chunksize)
333 data = os.read(self._input.fileno(), _chunksize)
334 if not data:
334 if not data:
335 self._eof = True
335 self._eof = True
336 else:
336 else:
337 self._lenbuf += len(data)
337 self._lenbuf += len(data)
338 self._buffer.append(data)
338 self._buffer.append(data)
339
339
340 def popen2(cmd, env=None, newlines=False):
340 def popen2(cmd, env=None, newlines=False):
341 # Setting bufsize to -1 lets the system decide the buffer size.
341 # Setting bufsize to -1 lets the system decide the buffer size.
342 # The default for bufsize is 0, meaning unbuffered. This leads to
342 # The default for bufsize is 0, meaning unbuffered. This leads to
343 # poor performance on Mac OS X: http://bugs.python.org/issue4194
343 # poor performance on Mac OS X: http://bugs.python.org/issue4194
344 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
344 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
345 close_fds=closefds,
345 close_fds=closefds,
346 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
346 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
347 universal_newlines=newlines,
347 universal_newlines=newlines,
348 env=env)
348 env=env)
349 return p.stdin, p.stdout
349 return p.stdin, p.stdout
350
350
351 def popen3(cmd, env=None, newlines=False):
351 def popen3(cmd, env=None, newlines=False):
352 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
352 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
353 return stdin, stdout, stderr
353 return stdin, stdout, stderr
354
354
355 def popen4(cmd, env=None, newlines=False, bufsize=-1):
355 def popen4(cmd, env=None, newlines=False, bufsize=-1):
356 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
356 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
357 close_fds=closefds,
357 close_fds=closefds,
358 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
358 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
359 stderr=subprocess.PIPE,
359 stderr=subprocess.PIPE,
360 universal_newlines=newlines,
360 universal_newlines=newlines,
361 env=env)
361 env=env)
362 return p.stdin, p.stdout, p.stderr, p
362 return p.stdin, p.stdout, p.stderr, p
363
363
364 def version():
364 def version():
365 """Return version information if available."""
365 """Return version information if available."""
366 try:
366 try:
367 from . import __version__
367 from . import __version__
368 return __version__.version
368 return __version__.version
369 except ImportError:
369 except ImportError:
370 return 'unknown'
370 return 'unknown'
371
371
372 def versiontuple(v=None, n=4):
372 def versiontuple(v=None, n=4):
373 """Parses a Mercurial version string into an N-tuple.
373 """Parses a Mercurial version string into an N-tuple.
374
374
375 The version string to be parsed is specified with the ``v`` argument.
375 The version string to be parsed is specified with the ``v`` argument.
376 If it isn't defined, the current Mercurial version string will be parsed.
376 If it isn't defined, the current Mercurial version string will be parsed.
377
377
378 ``n`` can be 2, 3, or 4. Here is how some version strings map to
378 ``n`` can be 2, 3, or 4. Here is how some version strings map to
379 returned values:
379 returned values:
380
380
381 >>> v = '3.6.1+190-df9b73d2d444'
381 >>> v = '3.6.1+190-df9b73d2d444'
382 >>> versiontuple(v, 2)
382 >>> versiontuple(v, 2)
383 (3, 6)
383 (3, 6)
384 >>> versiontuple(v, 3)
384 >>> versiontuple(v, 3)
385 (3, 6, 1)
385 (3, 6, 1)
386 >>> versiontuple(v, 4)
386 >>> versiontuple(v, 4)
387 (3, 6, 1, '190-df9b73d2d444')
387 (3, 6, 1, '190-df9b73d2d444')
388
388
389 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
389 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
390 (3, 6, 1, '190-df9b73d2d444+20151118')
390 (3, 6, 1, '190-df9b73d2d444+20151118')
391
391
392 >>> v = '3.6'
392 >>> v = '3.6'
393 >>> versiontuple(v, 2)
393 >>> versiontuple(v, 2)
394 (3, 6)
394 (3, 6)
395 >>> versiontuple(v, 3)
395 >>> versiontuple(v, 3)
396 (3, 6, None)
396 (3, 6, None)
397 >>> versiontuple(v, 4)
397 >>> versiontuple(v, 4)
398 (3, 6, None, None)
398 (3, 6, None, None)
399
399
400 >>> v = '3.9-rc'
400 >>> v = '3.9-rc'
401 >>> versiontuple(v, 2)
401 >>> versiontuple(v, 2)
402 (3, 9)
402 (3, 9)
403 >>> versiontuple(v, 3)
403 >>> versiontuple(v, 3)
404 (3, 9, None)
404 (3, 9, None)
405 >>> versiontuple(v, 4)
405 >>> versiontuple(v, 4)
406 (3, 9, None, 'rc')
406 (3, 9, None, 'rc')
407
407
408 >>> v = '3.9-rc+2-02a8fea4289b'
408 >>> v = '3.9-rc+2-02a8fea4289b'
409 >>> versiontuple(v, 2)
409 >>> versiontuple(v, 2)
410 (3, 9)
410 (3, 9)
411 >>> versiontuple(v, 3)
411 >>> versiontuple(v, 3)
412 (3, 9, None)
412 (3, 9, None)
413 >>> versiontuple(v, 4)
413 >>> versiontuple(v, 4)
414 (3, 9, None, 'rc+2-02a8fea4289b')
414 (3, 9, None, 'rc+2-02a8fea4289b')
415 """
415 """
416 if not v:
416 if not v:
417 v = version()
417 v = version()
418 parts = remod.split('[\+-]', v, 1)
418 parts = remod.split('[\+-]', v, 1)
419 if len(parts) == 1:
419 if len(parts) == 1:
420 vparts, extra = parts[0], None
420 vparts, extra = parts[0], None
421 else:
421 else:
422 vparts, extra = parts
422 vparts, extra = parts
423
423
424 vints = []
424 vints = []
425 for i in vparts.split('.'):
425 for i in vparts.split('.'):
426 try:
426 try:
427 vints.append(int(i))
427 vints.append(int(i))
428 except ValueError:
428 except ValueError:
429 break
429 break
430 # (3, 6) -> (3, 6, None)
430 # (3, 6) -> (3, 6, None)
431 while len(vints) < 3:
431 while len(vints) < 3:
432 vints.append(None)
432 vints.append(None)
433
433
434 if n == 2:
434 if n == 2:
435 return (vints[0], vints[1])
435 return (vints[0], vints[1])
436 if n == 3:
436 if n == 3:
437 return (vints[0], vints[1], vints[2])
437 return (vints[0], vints[1], vints[2])
438 if n == 4:
438 if n == 4:
439 return (vints[0], vints[1], vints[2], extra)
439 return (vints[0], vints[1], vints[2], extra)
440
440
441 # used by parsedate
441 # used by parsedate
442 defaultdateformats = (
442 defaultdateformats = (
443 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
443 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
444 '%Y-%m-%dT%H:%M', # without seconds
444 '%Y-%m-%dT%H:%M', # without seconds
445 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
445 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
446 '%Y-%m-%dT%H%M', # without seconds
446 '%Y-%m-%dT%H%M', # without seconds
447 '%Y-%m-%d %H:%M:%S', # our common legal variant
447 '%Y-%m-%d %H:%M:%S', # our common legal variant
448 '%Y-%m-%d %H:%M', # without seconds
448 '%Y-%m-%d %H:%M', # without seconds
449 '%Y-%m-%d %H%M%S', # without :
449 '%Y-%m-%d %H%M%S', # without :
450 '%Y-%m-%d %H%M', # without seconds
450 '%Y-%m-%d %H%M', # without seconds
451 '%Y-%m-%d %I:%M:%S%p',
451 '%Y-%m-%d %I:%M:%S%p',
452 '%Y-%m-%d %H:%M',
452 '%Y-%m-%d %H:%M',
453 '%Y-%m-%d %I:%M%p',
453 '%Y-%m-%d %I:%M%p',
454 '%Y-%m-%d',
454 '%Y-%m-%d',
455 '%m-%d',
455 '%m-%d',
456 '%m/%d',
456 '%m/%d',
457 '%m/%d/%y',
457 '%m/%d/%y',
458 '%m/%d/%Y',
458 '%m/%d/%Y',
459 '%a %b %d %H:%M:%S %Y',
459 '%a %b %d %H:%M:%S %Y',
460 '%a %b %d %I:%M:%S%p %Y',
460 '%a %b %d %I:%M:%S%p %Y',
461 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
461 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
462 '%b %d %H:%M:%S %Y',
462 '%b %d %H:%M:%S %Y',
463 '%b %d %I:%M:%S%p %Y',
463 '%b %d %I:%M:%S%p %Y',
464 '%b %d %H:%M:%S',
464 '%b %d %H:%M:%S',
465 '%b %d %I:%M:%S%p',
465 '%b %d %I:%M:%S%p',
466 '%b %d %H:%M',
466 '%b %d %H:%M',
467 '%b %d %I:%M%p',
467 '%b %d %I:%M%p',
468 '%b %d %Y',
468 '%b %d %Y',
469 '%b %d',
469 '%b %d',
470 '%H:%M:%S',
470 '%H:%M:%S',
471 '%I:%M:%S%p',
471 '%I:%M:%S%p',
472 '%H:%M',
472 '%H:%M',
473 '%I:%M%p',
473 '%I:%M%p',
474 )
474 )
475
475
476 extendeddateformats = defaultdateformats + (
476 extendeddateformats = defaultdateformats + (
477 "%Y",
477 "%Y",
478 "%Y-%m",
478 "%Y-%m",
479 "%b",
479 "%b",
480 "%b %Y",
480 "%b %Y",
481 )
481 )
482
482
483 def cachefunc(func):
483 def cachefunc(func):
484 '''cache the result of function calls'''
484 '''cache the result of function calls'''
485 # XXX doesn't handle keywords args
485 # XXX doesn't handle keywords args
486 if func.__code__.co_argcount == 0:
486 if func.__code__.co_argcount == 0:
487 cache = []
487 cache = []
488 def f():
488 def f():
489 if len(cache) == 0:
489 if len(cache) == 0:
490 cache.append(func())
490 cache.append(func())
491 return cache[0]
491 return cache[0]
492 return f
492 return f
493 cache = {}
493 cache = {}
494 if func.__code__.co_argcount == 1:
494 if func.__code__.co_argcount == 1:
495 # we gain a small amount of time because
495 # we gain a small amount of time because
496 # we don't need to pack/unpack the list
496 # we don't need to pack/unpack the list
497 def f(arg):
497 def f(arg):
498 if arg not in cache:
498 if arg not in cache:
499 cache[arg] = func(arg)
499 cache[arg] = func(arg)
500 return cache[arg]
500 return cache[arg]
501 else:
501 else:
502 def f(*args):
502 def f(*args):
503 if args not in cache:
503 if args not in cache:
504 cache[args] = func(*args)
504 cache[args] = func(*args)
505 return cache[args]
505 return cache[args]
506
506
507 return f
507 return f
508
508
509 class sortdict(dict):
509 class sortdict(dict):
510 '''a simple sorted dictionary'''
510 '''a simple sorted dictionary'''
511 def __init__(self, data=None):
511 def __init__(self, data=None):
512 self._list = []
512 self._list = []
513 if data:
513 if data:
514 self.update(data)
514 self.update(data)
515 def copy(self):
515 def copy(self):
516 return sortdict(self)
516 return sortdict(self)
517 def __setitem__(self, key, val):
517 def __setitem__(self, key, val):
518 if key in self:
518 if key in self:
519 self._list.remove(key)
519 self._list.remove(key)
520 self._list.append(key)
520 self._list.append(key)
521 dict.__setitem__(self, key, val)
521 dict.__setitem__(self, key, val)
522 def __iter__(self):
522 def __iter__(self):
523 return self._list.__iter__()
523 return self._list.__iter__()
524 def update(self, src):
524 def update(self, src):
525 if isinstance(src, dict):
525 if isinstance(src, dict):
526 src = src.iteritems()
526 src = src.iteritems()
527 for k, v in src:
527 for k, v in src:
528 self[k] = v
528 self[k] = v
529 def clear(self):
529 def clear(self):
530 dict.clear(self)
530 dict.clear(self)
531 self._list = []
531 self._list = []
532 def items(self):
532 def items(self):
533 return [(k, self[k]) for k in self._list]
533 return [(k, self[k]) for k in self._list]
534 def __delitem__(self, key):
534 def __delitem__(self, key):
535 dict.__delitem__(self, key)
535 dict.__delitem__(self, key)
536 self._list.remove(key)
536 self._list.remove(key)
537 def pop(self, key, *args, **kwargs):
537 def pop(self, key, *args, **kwargs):
538 dict.pop(self, key, *args, **kwargs)
538 dict.pop(self, key, *args, **kwargs)
539 try:
539 try:
540 self._list.remove(key)
540 self._list.remove(key)
541 except ValueError:
541 except ValueError:
542 pass
542 pass
543 def keys(self):
543 def keys(self):
544 return self._list
544 return self._list
545 def iterkeys(self):
545 def iterkeys(self):
546 return self._list.__iter__()
546 return self._list.__iter__()
547 def iteritems(self):
547 def iteritems(self):
548 for k in self._list:
548 for k in self._list:
549 yield k, self[k]
549 yield k, self[k]
550 def insert(self, index, key, val):
550 def insert(self, index, key, val):
551 self._list.insert(index, key)
551 self._list.insert(index, key)
552 dict.__setitem__(self, key, val)
552 dict.__setitem__(self, key, val)
553 def __repr__(self):
553 def __repr__(self):
554 if not self:
554 if not self:
555 return '%s()' % self.__class__.__name__
555 return '%s()' % self.__class__.__name__
556 return '%s(%r)' % (self.__class__.__name__, self.items())
556 return '%s(%r)' % (self.__class__.__name__, self.items())
557
557
558 class _lrucachenode(object):
558 class _lrucachenode(object):
559 """A node in a doubly linked list.
559 """A node in a doubly linked list.
560
560
561 Holds a reference to nodes on either side as well as a key-value
561 Holds a reference to nodes on either side as well as a key-value
562 pair for the dictionary entry.
562 pair for the dictionary entry.
563 """
563 """
564 __slots__ = ('next', 'prev', 'key', 'value')
564 __slots__ = ('next', 'prev', 'key', 'value')
565
565
566 def __init__(self):
566 def __init__(self):
567 self.next = None
567 self.next = None
568 self.prev = None
568 self.prev = None
569
569
570 self.key = _notset
570 self.key = _notset
571 self.value = None
571 self.value = None
572
572
573 def markempty(self):
573 def markempty(self):
574 """Mark the node as emptied."""
574 """Mark the node as emptied."""
575 self.key = _notset
575 self.key = _notset
576
576
577 class lrucachedict(object):
577 class lrucachedict(object):
578 """Dict that caches most recent accesses and sets.
578 """Dict that caches most recent accesses and sets.
579
579
580 The dict consists of an actual backing dict - indexed by original
580 The dict consists of an actual backing dict - indexed by original
581 key - and a doubly linked circular list defining the order of entries in
581 key - and a doubly linked circular list defining the order of entries in
582 the cache.
582 the cache.
583
583
584 The head node is the newest entry in the cache. If the cache is full,
584 The head node is the newest entry in the cache. If the cache is full,
585 we recycle head.prev and make it the new head. Cache accesses result in
585 we recycle head.prev and make it the new head. Cache accesses result in
586 the node being moved to before the existing head and being marked as the
586 the node being moved to before the existing head and being marked as the
587 new head node.
587 new head node.
588 """
588 """
589 def __init__(self, max):
589 def __init__(self, max):
590 self._cache = {}
590 self._cache = {}
591
591
592 self._head = head = _lrucachenode()
592 self._head = head = _lrucachenode()
593 head.prev = head
593 head.prev = head
594 head.next = head
594 head.next = head
595 self._size = 1
595 self._size = 1
596 self._capacity = max
596 self._capacity = max
597
597
598 def __len__(self):
598 def __len__(self):
599 return len(self._cache)
599 return len(self._cache)
600
600
601 def __contains__(self, k):
601 def __contains__(self, k):
602 return k in self._cache
602 return k in self._cache
603
603
604 def __iter__(self):
604 def __iter__(self):
605 # We don't have to iterate in cache order, but why not.
605 # We don't have to iterate in cache order, but why not.
606 n = self._head
606 n = self._head
607 for i in range(len(self._cache)):
607 for i in range(len(self._cache)):
608 yield n.key
608 yield n.key
609 n = n.next
609 n = n.next
610
610
611 def __getitem__(self, k):
611 def __getitem__(self, k):
612 node = self._cache[k]
612 node = self._cache[k]
613 self._movetohead(node)
613 self._movetohead(node)
614 return node.value
614 return node.value
615
615
616 def __setitem__(self, k, v):
616 def __setitem__(self, k, v):
617 node = self._cache.get(k)
617 node = self._cache.get(k)
618 # Replace existing value and mark as newest.
618 # Replace existing value and mark as newest.
619 if node is not None:
619 if node is not None:
620 node.value = v
620 node.value = v
621 self._movetohead(node)
621 self._movetohead(node)
622 return
622 return
623
623
624 if self._size < self._capacity:
624 if self._size < self._capacity:
625 node = self._addcapacity()
625 node = self._addcapacity()
626 else:
626 else:
627 # Grab the last/oldest item.
627 # Grab the last/oldest item.
628 node = self._head.prev
628 node = self._head.prev
629
629
630 # At capacity. Kill the old entry.
630 # At capacity. Kill the old entry.
631 if node.key is not _notset:
631 if node.key is not _notset:
632 del self._cache[node.key]
632 del self._cache[node.key]
633
633
634 node.key = k
634 node.key = k
635 node.value = v
635 node.value = v
636 self._cache[k] = node
636 self._cache[k] = node
637 # And mark it as newest entry. No need to adjust order since it
637 # And mark it as newest entry. No need to adjust order since it
638 # is already self._head.prev.
638 # is already self._head.prev.
639 self._head = node
639 self._head = node
640
640
641 def __delitem__(self, k):
641 def __delitem__(self, k):
642 node = self._cache.pop(k)
642 node = self._cache.pop(k)
643 node.markempty()
643 node.markempty()
644
644
645 # Temporarily mark as newest item before re-adjusting head to make
645 # Temporarily mark as newest item before re-adjusting head to make
646 # this node the oldest item.
646 # this node the oldest item.
647 self._movetohead(node)
647 self._movetohead(node)
648 self._head = node.next
648 self._head = node.next
649
649
650 # Additional dict methods.
650 # Additional dict methods.
651
651
652 def get(self, k, default=None):
652 def get(self, k, default=None):
653 try:
653 try:
654 return self._cache[k].value
654 return self._cache[k].value
655 except KeyError:
655 except KeyError:
656 return default
656 return default
657
657
658 def clear(self):
658 def clear(self):
659 n = self._head
659 n = self._head
660 while n.key is not _notset:
660 while n.key is not _notset:
661 n.markempty()
661 n.markempty()
662 n = n.next
662 n = n.next
663
663
664 self._cache.clear()
664 self._cache.clear()
665
665
666 def copy(self):
666 def copy(self):
667 result = lrucachedict(self._capacity)
667 result = lrucachedict(self._capacity)
668 n = self._head.prev
668 n = self._head.prev
669 # Iterate in oldest-to-newest order, so the copy has the right ordering
669 # Iterate in oldest-to-newest order, so the copy has the right ordering
670 for i in range(len(self._cache)):
670 for i in range(len(self._cache)):
671 result[n.key] = n.value
671 result[n.key] = n.value
672 n = n.prev
672 n = n.prev
673 return result
673 return result
674
674
675 def _movetohead(self, node):
675 def _movetohead(self, node):
676 """Mark a node as the newest, making it the new head.
676 """Mark a node as the newest, making it the new head.
677
677
678 When a node is accessed, it becomes the freshest entry in the LRU
678 When a node is accessed, it becomes the freshest entry in the LRU
679 list, which is denoted by self._head.
679 list, which is denoted by self._head.
680
680
681 Visually, let's make ``N`` the new head node (* denotes head):
681 Visually, let's make ``N`` the new head node (* denotes head):
682
682
683 previous/oldest <-> head <-> next/next newest
683 previous/oldest <-> head <-> next/next newest
684
684
685 ----<->--- A* ---<->-----
685 ----<->--- A* ---<->-----
686 | |
686 | |
687 E <-> D <-> N <-> C <-> B
687 E <-> D <-> N <-> C <-> B
688
688
689 To:
689 To:
690
690
691 ----<->--- N* ---<->-----
691 ----<->--- N* ---<->-----
692 | |
692 | |
693 E <-> D <-> C <-> B <-> A
693 E <-> D <-> C <-> B <-> A
694
694
695 This requires the following moves:
695 This requires the following moves:
696
696
697 C.next = D (node.prev.next = node.next)
697 C.next = D (node.prev.next = node.next)
698 D.prev = C (node.next.prev = node.prev)
698 D.prev = C (node.next.prev = node.prev)
699 E.next = N (head.prev.next = node)
699 E.next = N (head.prev.next = node)
700 N.prev = E (node.prev = head.prev)
700 N.prev = E (node.prev = head.prev)
701 N.next = A (node.next = head)
701 N.next = A (node.next = head)
702 A.prev = N (head.prev = node)
702 A.prev = N (head.prev = node)
703 """
703 """
704 head = self._head
704 head = self._head
705 # C.next = D
705 # C.next = D
706 node.prev.next = node.next
706 node.prev.next = node.next
707 # D.prev = C
707 # D.prev = C
708 node.next.prev = node.prev
708 node.next.prev = node.prev
709 # N.prev = E
709 # N.prev = E
710 node.prev = head.prev
710 node.prev = head.prev
711 # N.next = A
711 # N.next = A
712 # It is tempting to do just "head" here, however if node is
712 # It is tempting to do just "head" here, however if node is
713 # adjacent to head, this will do bad things.
713 # adjacent to head, this will do bad things.
714 node.next = head.prev.next
714 node.next = head.prev.next
715 # E.next = N
715 # E.next = N
716 node.next.prev = node
716 node.next.prev = node
717 # A.prev = N
717 # A.prev = N
718 node.prev.next = node
718 node.prev.next = node
719
719
720 self._head = node
720 self._head = node
721
721
722 def _addcapacity(self):
722 def _addcapacity(self):
723 """Add a node to the circular linked list.
723 """Add a node to the circular linked list.
724
724
725 The new node is inserted before the head node.
725 The new node is inserted before the head node.
726 """
726 """
727 head = self._head
727 head = self._head
728 node = _lrucachenode()
728 node = _lrucachenode()
729 head.prev.next = node
729 head.prev.next = node
730 node.prev = head.prev
730 node.prev = head.prev
731 node.next = head
731 node.next = head
732 head.prev = node
732 head.prev = node
733 self._size += 1
733 self._size += 1
734 return node
734 return node
735
735
736 def lrucachefunc(func):
736 def lrucachefunc(func):
737 '''cache most recent results of function calls'''
737 '''cache most recent results of function calls'''
738 cache = {}
738 cache = {}
739 order = collections.deque()
739 order = collections.deque()
740 if func.__code__.co_argcount == 1:
740 if func.__code__.co_argcount == 1:
741 def f(arg):
741 def f(arg):
742 if arg not in cache:
742 if arg not in cache:
743 if len(cache) > 20:
743 if len(cache) > 20:
744 del cache[order.popleft()]
744 del cache[order.popleft()]
745 cache[arg] = func(arg)
745 cache[arg] = func(arg)
746 else:
746 else:
747 order.remove(arg)
747 order.remove(arg)
748 order.append(arg)
748 order.append(arg)
749 return cache[arg]
749 return cache[arg]
750 else:
750 else:
751 def f(*args):
751 def f(*args):
752 if args not in cache:
752 if args not in cache:
753 if len(cache) > 20:
753 if len(cache) > 20:
754 del cache[order.popleft()]
754 del cache[order.popleft()]
755 cache[args] = func(*args)
755 cache[args] = func(*args)
756 else:
756 else:
757 order.remove(args)
757 order.remove(args)
758 order.append(args)
758 order.append(args)
759 return cache[args]
759 return cache[args]
760
760
761 return f
761 return f
762
762
763 class propertycache(object):
763 class propertycache(object):
764 def __init__(self, func):
764 def __init__(self, func):
765 self.func = func
765 self.func = func
766 self.name = func.__name__
766 self.name = func.__name__
767 def __get__(self, obj, type=None):
767 def __get__(self, obj, type=None):
768 result = self.func(obj)
768 result = self.func(obj)
769 self.cachevalue(obj, result)
769 self.cachevalue(obj, result)
770 return result
770 return result
771
771
772 def cachevalue(self, obj, value):
772 def cachevalue(self, obj, value):
773 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
773 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
774 obj.__dict__[self.name] = value
774 obj.__dict__[self.name] = value
775
775
776 def pipefilter(s, cmd):
776 def pipefilter(s, cmd):
777 '''filter string S through command CMD, returning its output'''
777 '''filter string S through command CMD, returning its output'''
778 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
778 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
779 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
779 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
780 pout, perr = p.communicate(s)
780 pout, perr = p.communicate(s)
781 return pout
781 return pout
782
782
783 def tempfilter(s, cmd):
783 def tempfilter(s, cmd):
784 '''filter string S through a pair of temporary files with CMD.
784 '''filter string S through a pair of temporary files with CMD.
785 CMD is used as a template to create the real command to be run,
785 CMD is used as a template to create the real command to be run,
786 with the strings INFILE and OUTFILE replaced by the real names of
786 with the strings INFILE and OUTFILE replaced by the real names of
787 the temporary files generated.'''
787 the temporary files generated.'''
788 inname, outname = None, None
788 inname, outname = None, None
789 try:
789 try:
790 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
790 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
791 fp = os.fdopen(infd, 'wb')
791 fp = os.fdopen(infd, 'wb')
792 fp.write(s)
792 fp.write(s)
793 fp.close()
793 fp.close()
794 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
794 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
795 os.close(outfd)
795 os.close(outfd)
796 cmd = cmd.replace('INFILE', inname)
796 cmd = cmd.replace('INFILE', inname)
797 cmd = cmd.replace('OUTFILE', outname)
797 cmd = cmd.replace('OUTFILE', outname)
798 code = os.system(cmd)
798 code = os.system(cmd)
799 if sys.platform == 'OpenVMS' and code & 1:
799 if sys.platform == 'OpenVMS' and code & 1:
800 code = 0
800 code = 0
801 if code:
801 if code:
802 raise Abort(_("command '%s' failed: %s") %
802 raise Abort(_("command '%s' failed: %s") %
803 (cmd, explainexit(code)))
803 (cmd, explainexit(code)))
804 return readfile(outname)
804 return readfile(outname)
805 finally:
805 finally:
806 try:
806 try:
807 if inname:
807 if inname:
808 os.unlink(inname)
808 os.unlink(inname)
809 except OSError:
809 except OSError:
810 pass
810 pass
811 try:
811 try:
812 if outname:
812 if outname:
813 os.unlink(outname)
813 os.unlink(outname)
814 except OSError:
814 except OSError:
815 pass
815 pass
816
816
817 filtertable = {
817 filtertable = {
818 'tempfile:': tempfilter,
818 'tempfile:': tempfilter,
819 'pipe:': pipefilter,
819 'pipe:': pipefilter,
820 }
820 }
821
821
822 def filter(s, cmd):
822 def filter(s, cmd):
823 "filter a string through a command that transforms its input to its output"
823 "filter a string through a command that transforms its input to its output"
824 for name, fn in filtertable.iteritems():
824 for name, fn in filtertable.iteritems():
825 if cmd.startswith(name):
825 if cmd.startswith(name):
826 return fn(s, cmd[len(name):].lstrip())
826 return fn(s, cmd[len(name):].lstrip())
827 return pipefilter(s, cmd)
827 return pipefilter(s, cmd)
828
828
829 def binary(s):
829 def binary(s):
830 """return true if a string is binary data"""
830 """return true if a string is binary data"""
831 return bool(s and '\0' in s)
831 return bool(s and '\0' in s)
832
832
833 def increasingchunks(source, min=1024, max=65536):
833 def increasingchunks(source, min=1024, max=65536):
834 '''return no less than min bytes per chunk while data remains,
834 '''return no less than min bytes per chunk while data remains,
835 doubling min after each chunk until it reaches max'''
835 doubling min after each chunk until it reaches max'''
836 def log2(x):
836 def log2(x):
837 if not x:
837 if not x:
838 return 0
838 return 0
839 i = 0
839 i = 0
840 while x:
840 while x:
841 x >>= 1
841 x >>= 1
842 i += 1
842 i += 1
843 return i - 1
843 return i - 1
844
844
845 buf = []
845 buf = []
846 blen = 0
846 blen = 0
847 for chunk in source:
847 for chunk in source:
848 buf.append(chunk)
848 buf.append(chunk)
849 blen += len(chunk)
849 blen += len(chunk)
850 if blen >= min:
850 if blen >= min:
851 if min < max:
851 if min < max:
852 min = min << 1
852 min = min << 1
853 nmin = 1 << log2(blen)
853 nmin = 1 << log2(blen)
854 if nmin > min:
854 if nmin > min:
855 min = nmin
855 min = nmin
856 if min > max:
856 if min > max:
857 min = max
857 min = max
858 yield ''.join(buf)
858 yield ''.join(buf)
859 blen = 0
859 blen = 0
860 buf = []
860 buf = []
861 if buf:
861 if buf:
862 yield ''.join(buf)
862 yield ''.join(buf)
863
863
864 Abort = error.Abort
864 Abort = error.Abort
865
865
866 def always(fn):
866 def always(fn):
867 return True
867 return True
868
868
869 def never(fn):
869 def never(fn):
870 return False
870 return False
871
871
872 def nogc(func):
872 def nogc(func):
873 """disable garbage collector
873 """disable garbage collector
874
874
875 Python's garbage collector triggers a GC each time a certain number of
875 Python's garbage collector triggers a GC each time a certain number of
876 container objects (the number being defined by gc.get_threshold()) are
876 container objects (the number being defined by gc.get_threshold()) are
877 allocated even when marked not to be tracked by the collector. Tracking has
877 allocated even when marked not to be tracked by the collector. Tracking has
878 no effect on when GCs are triggered, only on what objects the GC looks
878 no effect on when GCs are triggered, only on what objects the GC looks
879 into. As a workaround, disable GC while building complex (huge)
879 into. As a workaround, disable GC while building complex (huge)
880 containers.
880 containers.
881
881
882 This garbage collector issue have been fixed in 2.7.
882 This garbage collector issue have been fixed in 2.7.
883 """
883 """
884 if sys.version >= (2, 7):
884 if sys.version >= (2, 7):
885 return func
885 return func
886 def wrapper(*args, **kwargs):
886 def wrapper(*args, **kwargs):
887 gcenabled = gc.isenabled()
887 gcenabled = gc.isenabled()
888 gc.disable()
888 gc.disable()
889 try:
889 try:
890 return func(*args, **kwargs)
890 return func(*args, **kwargs)
891 finally:
891 finally:
892 if gcenabled:
892 if gcenabled:
893 gc.enable()
893 gc.enable()
894 return wrapper
894 return wrapper
895
895
896 def pathto(root, n1, n2):
896 def pathto(root, n1, n2):
897 '''return the relative path from one place to another.
897 '''return the relative path from one place to another.
898 root should use os.sep to separate directories
898 root should use os.sep to separate directories
899 n1 should use os.sep to separate directories
899 n1 should use os.sep to separate directories
900 n2 should use "/" to separate directories
900 n2 should use "/" to separate directories
901 returns an os.sep-separated path.
901 returns an os.sep-separated path.
902
902
903 If n1 is a relative path, it's assumed it's
903 If n1 is a relative path, it's assumed it's
904 relative to root.
904 relative to root.
905 n2 should always be relative to root.
905 n2 should always be relative to root.
906 '''
906 '''
907 if not n1:
907 if not n1:
908 return localpath(n2)
908 return localpath(n2)
909 if os.path.isabs(n1):
909 if os.path.isabs(n1):
910 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
910 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
911 return os.path.join(root, localpath(n2))
911 return os.path.join(root, localpath(n2))
912 n2 = '/'.join((pconvert(root), n2))
912 n2 = '/'.join((pconvert(root), n2))
913 a, b = splitpath(n1), n2.split('/')
913 a, b = splitpath(n1), n2.split('/')
914 a.reverse()
914 a.reverse()
915 b.reverse()
915 b.reverse()
916 while a and b and a[-1] == b[-1]:
916 while a and b and a[-1] == b[-1]:
917 a.pop()
917 a.pop()
918 b.pop()
918 b.pop()
919 b.reverse()
919 b.reverse()
920 return os.sep.join((['..'] * len(a)) + b) or '.'
920 return os.sep.join((['..'] * len(a)) + b) or '.'
921
921
922 def mainfrozen():
922 def mainfrozen():
923 """return True if we are a frozen executable.
923 """return True if we are a frozen executable.
924
924
925 The code supports py2exe (most common, Windows only) and tools/freeze
925 The code supports py2exe (most common, Windows only) and tools/freeze
926 (portable, not much used).
926 (portable, not much used).
927 """
927 """
928 return (safehasattr(sys, "frozen") or # new py2exe
928 return (safehasattr(sys, "frozen") or # new py2exe
929 safehasattr(sys, "importers") or # old py2exe
929 safehasattr(sys, "importers") or # old py2exe
930 imp.is_frozen("__main__")) # tools/freeze
930 imp.is_frozen("__main__")) # tools/freeze
931
931
932 # the location of data files matching the source code
932 # the location of data files matching the source code
933 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
933 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
934 # executable version (py2exe) doesn't support __file__
934 # executable version (py2exe) doesn't support __file__
935 datapath = os.path.dirname(sys.executable)
935 datapath = os.path.dirname(sys.executable)
936 else:
936 else:
937 datapath = os.path.dirname(__file__)
937 datapath = os.path.dirname(__file__)
938
938
939 i18n.setdatapath(datapath)
939 i18n.setdatapath(datapath)
940
940
941 _hgexecutable = None
941 _hgexecutable = None
942
942
943 def hgexecutable():
943 def hgexecutable():
944 """return location of the 'hg' executable.
944 """return location of the 'hg' executable.
945
945
946 Defaults to $HG or 'hg' in the search path.
946 Defaults to $HG or 'hg' in the search path.
947 """
947 """
948 if _hgexecutable is None:
948 if _hgexecutable is None:
949 hg = os.environ.get('HG')
949 hg = os.environ.get('HG')
950 mainmod = sys.modules['__main__']
950 mainmod = sys.modules['__main__']
951 if hg:
951 if hg:
952 _sethgexecutable(hg)
952 _sethgexecutable(hg)
953 elif mainfrozen():
953 elif mainfrozen():
954 if getattr(sys, 'frozen', None) == 'macosx_app':
954 if getattr(sys, 'frozen', None) == 'macosx_app':
955 # Env variable set by py2app
955 # Env variable set by py2app
956 _sethgexecutable(os.environ['EXECUTABLEPATH'])
956 _sethgexecutable(os.environ['EXECUTABLEPATH'])
957 else:
957 else:
958 _sethgexecutable(sys.executable)
958 _sethgexecutable(sys.executable)
959 elif os.path.basename(getattr(mainmod, '__file__', '')) == 'hg':
959 elif os.path.basename(getattr(mainmod, '__file__', '')) == 'hg':
960 _sethgexecutable(mainmod.__file__)
960 _sethgexecutable(mainmod.__file__)
961 else:
961 else:
962 exe = findexe('hg') or os.path.basename(sys.argv[0])
962 exe = findexe('hg') or os.path.basename(sys.argv[0])
963 _sethgexecutable(exe)
963 _sethgexecutable(exe)
964 return _hgexecutable
964 return _hgexecutable
965
965
966 def _sethgexecutable(path):
966 def _sethgexecutable(path):
967 """set location of the 'hg' executable"""
967 """set location of the 'hg' executable"""
968 global _hgexecutable
968 global _hgexecutable
969 _hgexecutable = path
969 _hgexecutable = path
970
970
971 def _isstdout(f):
971 def _isstdout(f):
972 fileno = getattr(f, 'fileno', None)
972 fileno = getattr(f, 'fileno', None)
973 return fileno and fileno() == sys.__stdout__.fileno()
973 return fileno and fileno() == sys.__stdout__.fileno()
974
974
975 def system(cmd, environ=None, cwd=None, onerr=None, errprefix=None, out=None):
975 def system(cmd, environ=None, cwd=None, onerr=None, errprefix=None, out=None):
976 '''enhanced shell command execution.
976 '''enhanced shell command execution.
977 run with environment maybe modified, maybe in different dir.
977 run with environment maybe modified, maybe in different dir.
978
978
979 if command fails and onerr is None, return status, else raise onerr
979 if command fails and onerr is None, return status, else raise onerr
980 object as exception.
980 object as exception.
981
981
982 if out is specified, it is assumed to be a file-like object that has a
982 if out is specified, it is assumed to be a file-like object that has a
983 write() method. stdout and stderr will be redirected to out.'''
983 write() method. stdout and stderr will be redirected to out.'''
984 if environ is None:
984 if environ is None:
985 environ = {}
985 environ = {}
986 try:
986 try:
987 sys.stdout.flush()
987 sys.stdout.flush()
988 except Exception:
988 except Exception:
989 pass
989 pass
990 def py2shell(val):
990 def py2shell(val):
991 'convert python object into string that is useful to shell'
991 'convert python object into string that is useful to shell'
992 if val is None or val is False:
992 if val is None or val is False:
993 return '0'
993 return '0'
994 if val is True:
994 if val is True:
995 return '1'
995 return '1'
996 return str(val)
996 return str(val)
997 origcmd = cmd
997 origcmd = cmd
998 cmd = quotecommand(cmd)
998 cmd = quotecommand(cmd)
999 if sys.platform == 'plan9' and (sys.version_info[0] == 2
999 if sys.platform == 'plan9' and (sys.version_info[0] == 2
1000 and sys.version_info[1] < 7):
1000 and sys.version_info[1] < 7):
1001 # subprocess kludge to work around issues in half-baked Python
1001 # subprocess kludge to work around issues in half-baked Python
1002 # ports, notably bichued/python:
1002 # ports, notably bichued/python:
1003 if not cwd is None:
1003 if not cwd is None:
1004 os.chdir(cwd)
1004 os.chdir(cwd)
1005 rc = os.system(cmd)
1005 rc = os.system(cmd)
1006 else:
1006 else:
1007 env = dict(os.environ)
1007 env = dict(os.environ)
1008 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1008 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1009 env['HG'] = hgexecutable()
1009 env['HG'] = hgexecutable()
1010 if out is None or _isstdout(out):
1010 if out is None or _isstdout(out):
1011 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1011 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1012 env=env, cwd=cwd)
1012 env=env, cwd=cwd)
1013 else:
1013 else:
1014 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1014 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1015 env=env, cwd=cwd, stdout=subprocess.PIPE,
1015 env=env, cwd=cwd, stdout=subprocess.PIPE,
1016 stderr=subprocess.STDOUT)
1016 stderr=subprocess.STDOUT)
1017 for line in iter(proc.stdout.readline, ''):
1017 for line in iter(proc.stdout.readline, ''):
1018 out.write(line)
1018 out.write(line)
1019 proc.wait()
1019 proc.wait()
1020 rc = proc.returncode
1020 rc = proc.returncode
1021 if sys.platform == 'OpenVMS' and rc & 1:
1021 if sys.platform == 'OpenVMS' and rc & 1:
1022 rc = 0
1022 rc = 0
1023 if rc and onerr:
1023 if rc and onerr:
1024 errmsg = '%s %s' % (os.path.basename(origcmd.split(None, 1)[0]),
1024 errmsg = '%s %s' % (os.path.basename(origcmd.split(None, 1)[0]),
1025 explainexit(rc)[0])
1025 explainexit(rc)[0])
1026 if errprefix:
1026 if errprefix:
1027 errmsg = '%s: %s' % (errprefix, errmsg)
1027 errmsg = '%s: %s' % (errprefix, errmsg)
1028 raise onerr(errmsg)
1028 raise onerr(errmsg)
1029 return rc
1029 return rc
1030
1030
1031 def checksignature(func):
1031 def checksignature(func):
1032 '''wrap a function with code to check for calling errors'''
1032 '''wrap a function with code to check for calling errors'''
1033 def check(*args, **kwargs):
1033 def check(*args, **kwargs):
1034 try:
1034 try:
1035 return func(*args, **kwargs)
1035 return func(*args, **kwargs)
1036 except TypeError:
1036 except TypeError:
1037 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1037 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1038 raise error.SignatureError
1038 raise error.SignatureError
1039 raise
1039 raise
1040
1040
1041 return check
1041 return check
1042
1042
1043 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1043 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1044 '''copy a file, preserving mode and optionally other stat info like
1044 '''copy a file, preserving mode and optionally other stat info like
1045 atime/mtime
1045 atime/mtime
1046
1046
1047 checkambig argument is used with filestat, and is useful only if
1047 checkambig argument is used with filestat, and is useful only if
1048 destination file is guarded by any lock (e.g. repo.lock or
1048 destination file is guarded by any lock (e.g. repo.lock or
1049 repo.wlock).
1049 repo.wlock).
1050
1050
1051 copystat and checkambig should be exclusive.
1051 copystat and checkambig should be exclusive.
1052 '''
1052 '''
1053 assert not (copystat and checkambig)
1053 assert not (copystat and checkambig)
1054 oldstat = None
1054 oldstat = None
1055 if os.path.lexists(dest):
1055 if os.path.lexists(dest):
1056 if checkambig:
1056 if checkambig:
1057 oldstat = checkambig and filestat(dest)
1057 oldstat = checkambig and filestat(dest)
1058 unlink(dest)
1058 unlink(dest)
1059 # hardlinks are problematic on CIFS, quietly ignore this flag
1059 # hardlinks are problematic on CIFS, quietly ignore this flag
1060 # until we find a way to work around it cleanly (issue4546)
1060 # until we find a way to work around it cleanly (issue4546)
1061 if False and hardlink:
1061 if False and hardlink:
1062 try:
1062 try:
1063 oslink(src, dest)
1063 oslink(src, dest)
1064 return
1064 return
1065 except (IOError, OSError):
1065 except (IOError, OSError):
1066 pass # fall back to normal copy
1066 pass # fall back to normal copy
1067 if os.path.islink(src):
1067 if os.path.islink(src):
1068 os.symlink(os.readlink(src), dest)
1068 os.symlink(os.readlink(src), dest)
1069 # copytime is ignored for symlinks, but in general copytime isn't needed
1069 # copytime is ignored for symlinks, but in general copytime isn't needed
1070 # for them anyway
1070 # for them anyway
1071 else:
1071 else:
1072 try:
1072 try:
1073 shutil.copyfile(src, dest)
1073 shutil.copyfile(src, dest)
1074 if copystat:
1074 if copystat:
1075 # copystat also copies mode
1075 # copystat also copies mode
1076 shutil.copystat(src, dest)
1076 shutil.copystat(src, dest)
1077 else:
1077 else:
1078 shutil.copymode(src, dest)
1078 shutil.copymode(src, dest)
1079 if oldstat and oldstat.stat:
1079 if oldstat and oldstat.stat:
1080 newstat = filestat(dest)
1080 newstat = filestat(dest)
1081 if newstat.isambig(oldstat):
1081 if newstat.isambig(oldstat):
1082 # stat of copied file is ambiguous to original one
1082 # stat of copied file is ambiguous to original one
1083 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1083 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1084 os.utime(dest, (advanced, advanced))
1084 os.utime(dest, (advanced, advanced))
1085 except shutil.Error as inst:
1085 except shutil.Error as inst:
1086 raise Abort(str(inst))
1086 raise Abort(str(inst))
1087
1087
1088 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1088 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1089 """Copy a directory tree using hardlinks if possible."""
1089 """Copy a directory tree using hardlinks if possible."""
1090 num = 0
1090 num = 0
1091
1091
1092 if hardlink is None:
1092 if hardlink is None:
1093 hardlink = (os.stat(src).st_dev ==
1093 hardlink = (os.stat(src).st_dev ==
1094 os.stat(os.path.dirname(dst)).st_dev)
1094 os.stat(os.path.dirname(dst)).st_dev)
1095 if hardlink:
1095 if hardlink:
1096 topic = _('linking')
1096 topic = _('linking')
1097 else:
1097 else:
1098 topic = _('copying')
1098 topic = _('copying')
1099
1099
1100 if os.path.isdir(src):
1100 if os.path.isdir(src):
1101 os.mkdir(dst)
1101 os.mkdir(dst)
1102 for name, kind in osutil.listdir(src):
1102 for name, kind in osutil.listdir(src):
1103 srcname = os.path.join(src, name)
1103 srcname = os.path.join(src, name)
1104 dstname = os.path.join(dst, name)
1104 dstname = os.path.join(dst, name)
1105 def nprog(t, pos):
1105 def nprog(t, pos):
1106 if pos is not None:
1106 if pos is not None:
1107 return progress(t, pos + num)
1107 return progress(t, pos + num)
1108 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1108 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1109 num += n
1109 num += n
1110 else:
1110 else:
1111 if hardlink:
1111 if hardlink:
1112 try:
1112 try:
1113 oslink(src, dst)
1113 oslink(src, dst)
1114 except (IOError, OSError):
1114 except (IOError, OSError):
1115 hardlink = False
1115 hardlink = False
1116 shutil.copy(src, dst)
1116 shutil.copy(src, dst)
1117 else:
1117 else:
1118 shutil.copy(src, dst)
1118 shutil.copy(src, dst)
1119 num += 1
1119 num += 1
1120 progress(topic, num)
1120 progress(topic, num)
1121 progress(topic, None)
1121 progress(topic, None)
1122
1122
1123 return hardlink, num
1123 return hardlink, num
1124
1124
1125 _winreservednames = '''con prn aux nul
1125 _winreservednames = '''con prn aux nul
1126 com1 com2 com3 com4 com5 com6 com7 com8 com9
1126 com1 com2 com3 com4 com5 com6 com7 com8 com9
1127 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1127 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1128 _winreservedchars = ':*?"<>|'
1128 _winreservedchars = ':*?"<>|'
1129 def checkwinfilename(path):
1129 def checkwinfilename(path):
1130 r'''Check that the base-relative path is a valid filename on Windows.
1130 r'''Check that the base-relative path is a valid filename on Windows.
1131 Returns None if the path is ok, or a UI string describing the problem.
1131 Returns None if the path is ok, or a UI string describing the problem.
1132
1132
1133 >>> checkwinfilename("just/a/normal/path")
1133 >>> checkwinfilename("just/a/normal/path")
1134 >>> checkwinfilename("foo/bar/con.xml")
1134 >>> checkwinfilename("foo/bar/con.xml")
1135 "filename contains 'con', which is reserved on Windows"
1135 "filename contains 'con', which is reserved on Windows"
1136 >>> checkwinfilename("foo/con.xml/bar")
1136 >>> checkwinfilename("foo/con.xml/bar")
1137 "filename contains 'con', which is reserved on Windows"
1137 "filename contains 'con', which is reserved on Windows"
1138 >>> checkwinfilename("foo/bar/xml.con")
1138 >>> checkwinfilename("foo/bar/xml.con")
1139 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1139 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1140 "filename contains 'AUX', which is reserved on Windows"
1140 "filename contains 'AUX', which is reserved on Windows"
1141 >>> checkwinfilename("foo/bar/bla:.txt")
1141 >>> checkwinfilename("foo/bar/bla:.txt")
1142 "filename contains ':', which is reserved on Windows"
1142 "filename contains ':', which is reserved on Windows"
1143 >>> checkwinfilename("foo/bar/b\07la.txt")
1143 >>> checkwinfilename("foo/bar/b\07la.txt")
1144 "filename contains '\\x07', which is invalid on Windows"
1144 "filename contains '\\x07', which is invalid on Windows"
1145 >>> checkwinfilename("foo/bar/bla ")
1145 >>> checkwinfilename("foo/bar/bla ")
1146 "filename ends with ' ', which is not allowed on Windows"
1146 "filename ends with ' ', which is not allowed on Windows"
1147 >>> checkwinfilename("../bar")
1147 >>> checkwinfilename("../bar")
1148 >>> checkwinfilename("foo\\")
1148 >>> checkwinfilename("foo\\")
1149 "filename ends with '\\', which is invalid on Windows"
1149 "filename ends with '\\', which is invalid on Windows"
1150 >>> checkwinfilename("foo\\/bar")
1150 >>> checkwinfilename("foo\\/bar")
1151 "directory name ends with '\\', which is invalid on Windows"
1151 "directory name ends with '\\', which is invalid on Windows"
1152 '''
1152 '''
1153 if path.endswith('\\'):
1153 if path.endswith('\\'):
1154 return _("filename ends with '\\', which is invalid on Windows")
1154 return _("filename ends with '\\', which is invalid on Windows")
1155 if '\\/' in path:
1155 if '\\/' in path:
1156 return _("directory name ends with '\\', which is invalid on Windows")
1156 return _("directory name ends with '\\', which is invalid on Windows")
1157 for n in path.replace('\\', '/').split('/'):
1157 for n in path.replace('\\', '/').split('/'):
1158 if not n:
1158 if not n:
1159 continue
1159 continue
1160 for c in n:
1160 for c in n:
1161 if c in _winreservedchars:
1161 if c in _winreservedchars:
1162 return _("filename contains '%s', which is reserved "
1162 return _("filename contains '%s', which is reserved "
1163 "on Windows") % c
1163 "on Windows") % c
1164 if ord(c) <= 31:
1164 if ord(c) <= 31:
1165 return _("filename contains %r, which is invalid "
1165 return _("filename contains %r, which is invalid "
1166 "on Windows") % c
1166 "on Windows") % c
1167 base = n.split('.')[0]
1167 base = n.split('.')[0]
1168 if base and base.lower() in _winreservednames:
1168 if base and base.lower() in _winreservednames:
1169 return _("filename contains '%s', which is reserved "
1169 return _("filename contains '%s', which is reserved "
1170 "on Windows") % base
1170 "on Windows") % base
1171 t = n[-1]
1171 t = n[-1]
1172 if t in '. ' and n not in '..':
1172 if t in '. ' and n not in '..':
1173 return _("filename ends with '%s', which is not allowed "
1173 return _("filename ends with '%s', which is not allowed "
1174 "on Windows") % t
1174 "on Windows") % t
1175
1175
1176 if os.name == 'nt':
1176 if os.name == 'nt':
1177 checkosfilename = checkwinfilename
1177 checkosfilename = checkwinfilename
1178 else:
1178 else:
1179 checkosfilename = platform.checkosfilename
1179 checkosfilename = platform.checkosfilename
1180
1180
1181 def makelock(info, pathname):
1181 def makelock(info, pathname):
1182 try:
1182 try:
1183 return os.symlink(info, pathname)
1183 return os.symlink(info, pathname)
1184 except OSError as why:
1184 except OSError as why:
1185 if why.errno == errno.EEXIST:
1185 if why.errno == errno.EEXIST:
1186 raise
1186 raise
1187 except AttributeError: # no symlink in os
1187 except AttributeError: # no symlink in os
1188 pass
1188 pass
1189
1189
1190 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1190 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1191 os.write(ld, info)
1191 os.write(ld, info)
1192 os.close(ld)
1192 os.close(ld)
1193
1193
1194 def readlock(pathname):
1194 def readlock(pathname):
1195 try:
1195 try:
1196 return os.readlink(pathname)
1196 return os.readlink(pathname)
1197 except OSError as why:
1197 except OSError as why:
1198 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1198 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1199 raise
1199 raise
1200 except AttributeError: # no symlink in os
1200 except AttributeError: # no symlink in os
1201 pass
1201 pass
1202 fp = posixfile(pathname)
1202 fp = posixfile(pathname)
1203 r = fp.read()
1203 r = fp.read()
1204 fp.close()
1204 fp.close()
1205 return r
1205 return r
1206
1206
1207 def fstat(fp):
1207 def fstat(fp):
1208 '''stat file object that may not have fileno method.'''
1208 '''stat file object that may not have fileno method.'''
1209 try:
1209 try:
1210 return os.fstat(fp.fileno())
1210 return os.fstat(fp.fileno())
1211 except AttributeError:
1211 except AttributeError:
1212 return os.stat(fp.name)
1212 return os.stat(fp.name)
1213
1213
1214 # File system features
1214 # File system features
1215
1215
1216 def checkcase(path):
1216 def checkcase(path):
1217 """
1217 """
1218 Return true if the given path is on a case-sensitive filesystem
1218 Return true if the given path is on a case-sensitive filesystem
1219
1219
1220 Requires a path (like /foo/.hg) ending with a foldable final
1220 Requires a path (like /foo/.hg) ending with a foldable final
1221 directory component.
1221 directory component.
1222 """
1222 """
1223 s1 = os.lstat(path)
1223 s1 = os.lstat(path)
1224 d, b = os.path.split(path)
1224 d, b = os.path.split(path)
1225 b2 = b.upper()
1225 b2 = b.upper()
1226 if b == b2:
1226 if b == b2:
1227 b2 = b.lower()
1227 b2 = b.lower()
1228 if b == b2:
1228 if b == b2:
1229 return True # no evidence against case sensitivity
1229 return True # no evidence against case sensitivity
1230 p2 = os.path.join(d, b2)
1230 p2 = os.path.join(d, b2)
1231 try:
1231 try:
1232 s2 = os.lstat(p2)
1232 s2 = os.lstat(p2)
1233 if s2 == s1:
1233 if s2 == s1:
1234 return False
1234 return False
1235 return True
1235 return True
1236 except OSError:
1236 except OSError:
1237 return True
1237 return True
1238
1238
1239 try:
1239 try:
1240 import re2
1240 import re2
1241 _re2 = None
1241 _re2 = None
1242 except ImportError:
1242 except ImportError:
1243 _re2 = False
1243 _re2 = False
1244
1244
1245 class _re(object):
1245 class _re(object):
1246 def _checkre2(self):
1246 def _checkre2(self):
1247 global _re2
1247 global _re2
1248 try:
1248 try:
1249 # check if match works, see issue3964
1249 # check if match works, see issue3964
1250 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1250 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1251 except ImportError:
1251 except ImportError:
1252 _re2 = False
1252 _re2 = False
1253
1253
1254 def compile(self, pat, flags=0):
1254 def compile(self, pat, flags=0):
1255 '''Compile a regular expression, using re2 if possible
1255 '''Compile a regular expression, using re2 if possible
1256
1256
1257 For best performance, use only re2-compatible regexp features. The
1257 For best performance, use only re2-compatible regexp features. The
1258 only flags from the re module that are re2-compatible are
1258 only flags from the re module that are re2-compatible are
1259 IGNORECASE and MULTILINE.'''
1259 IGNORECASE and MULTILINE.'''
1260 if _re2 is None:
1260 if _re2 is None:
1261 self._checkre2()
1261 self._checkre2()
1262 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1262 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1263 if flags & remod.IGNORECASE:
1263 if flags & remod.IGNORECASE:
1264 pat = '(?i)' + pat
1264 pat = '(?i)' + pat
1265 if flags & remod.MULTILINE:
1265 if flags & remod.MULTILINE:
1266 pat = '(?m)' + pat
1266 pat = '(?m)' + pat
1267 try:
1267 try:
1268 return re2.compile(pat)
1268 return re2.compile(pat)
1269 except re2.error:
1269 except re2.error:
1270 pass
1270 pass
1271 return remod.compile(pat, flags)
1271 return remod.compile(pat, flags)
1272
1272
1273 @propertycache
1273 @propertycache
1274 def escape(self):
1274 def escape(self):
1275 '''Return the version of escape corresponding to self.compile.
1275 '''Return the version of escape corresponding to self.compile.
1276
1276
1277 This is imperfect because whether re2 or re is used for a particular
1277 This is imperfect because whether re2 or re is used for a particular
1278 function depends on the flags, etc, but it's the best we can do.
1278 function depends on the flags, etc, but it's the best we can do.
1279 '''
1279 '''
1280 global _re2
1280 global _re2
1281 if _re2 is None:
1281 if _re2 is None:
1282 self._checkre2()
1282 self._checkre2()
1283 if _re2:
1283 if _re2:
1284 return re2.escape
1284 return re2.escape
1285 else:
1285 else:
1286 return remod.escape
1286 return remod.escape
1287
1287
1288 re = _re()
1288 re = _re()
1289
1289
1290 _fspathcache = {}
1290 _fspathcache = {}
1291 def fspath(name, root):
1291 def fspath(name, root):
1292 '''Get name in the case stored in the filesystem
1292 '''Get name in the case stored in the filesystem
1293
1293
1294 The name should be relative to root, and be normcase-ed for efficiency.
1294 The name should be relative to root, and be normcase-ed for efficiency.
1295
1295
1296 Note that this function is unnecessary, and should not be
1296 Note that this function is unnecessary, and should not be
1297 called, for case-sensitive filesystems (simply because it's expensive).
1297 called, for case-sensitive filesystems (simply because it's expensive).
1298
1298
1299 The root should be normcase-ed, too.
1299 The root should be normcase-ed, too.
1300 '''
1300 '''
1301 def _makefspathcacheentry(dir):
1301 def _makefspathcacheentry(dir):
1302 return dict((normcase(n), n) for n in os.listdir(dir))
1302 return dict((normcase(n), n) for n in os.listdir(dir))
1303
1303
1304 seps = os.sep
1304 seps = os.sep
1305 if os.altsep:
1305 if os.altsep:
1306 seps = seps + os.altsep
1306 seps = seps + os.altsep
1307 # Protect backslashes. This gets silly very quickly.
1307 # Protect backslashes. This gets silly very quickly.
1308 seps.replace('\\','\\\\')
1308 seps.replace('\\','\\\\')
1309 pattern = remod.compile(r'([^%s]+)|([%s]+)' % (seps, seps))
1309 pattern = remod.compile(r'([^%s]+)|([%s]+)' % (seps, seps))
1310 dir = os.path.normpath(root)
1310 dir = os.path.normpath(root)
1311 result = []
1311 result = []
1312 for part, sep in pattern.findall(name):
1312 for part, sep in pattern.findall(name):
1313 if sep:
1313 if sep:
1314 result.append(sep)
1314 result.append(sep)
1315 continue
1315 continue
1316
1316
1317 if dir not in _fspathcache:
1317 if dir not in _fspathcache:
1318 _fspathcache[dir] = _makefspathcacheentry(dir)
1318 _fspathcache[dir] = _makefspathcacheentry(dir)
1319 contents = _fspathcache[dir]
1319 contents = _fspathcache[dir]
1320
1320
1321 found = contents.get(part)
1321 found = contents.get(part)
1322 if not found:
1322 if not found:
1323 # retry "once per directory" per "dirstate.walk" which
1323 # retry "once per directory" per "dirstate.walk" which
1324 # may take place for each patches of "hg qpush", for example
1324 # may take place for each patches of "hg qpush", for example
1325 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1325 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1326 found = contents.get(part)
1326 found = contents.get(part)
1327
1327
1328 result.append(found or part)
1328 result.append(found or part)
1329 dir = os.path.join(dir, part)
1329 dir = os.path.join(dir, part)
1330
1330
1331 return ''.join(result)
1331 return ''.join(result)
1332
1332
1333 def checknlink(testfile):
1333 def checknlink(testfile):
1334 '''check whether hardlink count reporting works properly'''
1334 '''check whether hardlink count reporting works properly'''
1335
1335
1336 # testfile may be open, so we need a separate file for checking to
1336 # testfile may be open, so we need a separate file for checking to
1337 # work around issue2543 (or testfile may get lost on Samba shares)
1337 # work around issue2543 (or testfile may get lost on Samba shares)
1338 f1 = testfile + ".hgtmp1"
1338 f1 = testfile + ".hgtmp1"
1339 if os.path.lexists(f1):
1339 if os.path.lexists(f1):
1340 return False
1340 return False
1341 try:
1341 try:
1342 posixfile(f1, 'w').close()
1342 posixfile(f1, 'w').close()
1343 except IOError:
1343 except IOError:
1344 try:
1345 os.unlink(f1)
1346 except OSError:
1347 pass
1344 return False
1348 return False
1345
1349
1346 f2 = testfile + ".hgtmp2"
1350 f2 = testfile + ".hgtmp2"
1347 fd = None
1351 fd = None
1348 try:
1352 try:
1349 oslink(f1, f2)
1353 oslink(f1, f2)
1350 # nlinks() may behave differently for files on Windows shares if
1354 # nlinks() may behave differently for files on Windows shares if
1351 # the file is open.
1355 # the file is open.
1352 fd = posixfile(f2)
1356 fd = posixfile(f2)
1353 return nlinks(f2) > 1
1357 return nlinks(f2) > 1
1354 except OSError:
1358 except OSError:
1355 return False
1359 return False
1356 finally:
1360 finally:
1357 if fd is not None:
1361 if fd is not None:
1358 fd.close()
1362 fd.close()
1359 for f in (f1, f2):
1363 for f in (f1, f2):
1360 try:
1364 try:
1361 os.unlink(f)
1365 os.unlink(f)
1362 except OSError:
1366 except OSError:
1363 pass
1367 pass
1364
1368
1365 def endswithsep(path):
1369 def endswithsep(path):
1366 '''Check path ends with os.sep or os.altsep.'''
1370 '''Check path ends with os.sep or os.altsep.'''
1367 return path.endswith(os.sep) or os.altsep and path.endswith(os.altsep)
1371 return path.endswith(os.sep) or os.altsep and path.endswith(os.altsep)
1368
1372
1369 def splitpath(path):
1373 def splitpath(path):
1370 '''Split path by os.sep.
1374 '''Split path by os.sep.
1371 Note that this function does not use os.altsep because this is
1375 Note that this function does not use os.altsep because this is
1372 an alternative of simple "xxx.split(os.sep)".
1376 an alternative of simple "xxx.split(os.sep)".
1373 It is recommended to use os.path.normpath() before using this
1377 It is recommended to use os.path.normpath() before using this
1374 function if need.'''
1378 function if need.'''
1375 return path.split(os.sep)
1379 return path.split(os.sep)
1376
1380
1377 def gui():
1381 def gui():
1378 '''Are we running in a GUI?'''
1382 '''Are we running in a GUI?'''
1379 if sys.platform == 'darwin':
1383 if sys.platform == 'darwin':
1380 if 'SSH_CONNECTION' in os.environ:
1384 if 'SSH_CONNECTION' in os.environ:
1381 # handle SSH access to a box where the user is logged in
1385 # handle SSH access to a box where the user is logged in
1382 return False
1386 return False
1383 elif getattr(osutil, 'isgui', None):
1387 elif getattr(osutil, 'isgui', None):
1384 # check if a CoreGraphics session is available
1388 # check if a CoreGraphics session is available
1385 return osutil.isgui()
1389 return osutil.isgui()
1386 else:
1390 else:
1387 # pure build; use a safe default
1391 # pure build; use a safe default
1388 return True
1392 return True
1389 else:
1393 else:
1390 return os.name == "nt" or os.environ.get("DISPLAY")
1394 return os.name == "nt" or os.environ.get("DISPLAY")
1391
1395
1392 def mktempcopy(name, emptyok=False, createmode=None):
1396 def mktempcopy(name, emptyok=False, createmode=None):
1393 """Create a temporary file with the same contents from name
1397 """Create a temporary file with the same contents from name
1394
1398
1395 The permission bits are copied from the original file.
1399 The permission bits are copied from the original file.
1396
1400
1397 If the temporary file is going to be truncated immediately, you
1401 If the temporary file is going to be truncated immediately, you
1398 can use emptyok=True as an optimization.
1402 can use emptyok=True as an optimization.
1399
1403
1400 Returns the name of the temporary file.
1404 Returns the name of the temporary file.
1401 """
1405 """
1402 d, fn = os.path.split(name)
1406 d, fn = os.path.split(name)
1403 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1407 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1404 os.close(fd)
1408 os.close(fd)
1405 # Temporary files are created with mode 0600, which is usually not
1409 # Temporary files are created with mode 0600, which is usually not
1406 # what we want. If the original file already exists, just copy
1410 # what we want. If the original file already exists, just copy
1407 # its mode. Otherwise, manually obey umask.
1411 # its mode. Otherwise, manually obey umask.
1408 copymode(name, temp, createmode)
1412 copymode(name, temp, createmode)
1409 if emptyok:
1413 if emptyok:
1410 return temp
1414 return temp
1411 try:
1415 try:
1412 try:
1416 try:
1413 ifp = posixfile(name, "rb")
1417 ifp = posixfile(name, "rb")
1414 except IOError as inst:
1418 except IOError as inst:
1415 if inst.errno == errno.ENOENT:
1419 if inst.errno == errno.ENOENT:
1416 return temp
1420 return temp
1417 if not getattr(inst, 'filename', None):
1421 if not getattr(inst, 'filename', None):
1418 inst.filename = name
1422 inst.filename = name
1419 raise
1423 raise
1420 ofp = posixfile(temp, "wb")
1424 ofp = posixfile(temp, "wb")
1421 for chunk in filechunkiter(ifp):
1425 for chunk in filechunkiter(ifp):
1422 ofp.write(chunk)
1426 ofp.write(chunk)
1423 ifp.close()
1427 ifp.close()
1424 ofp.close()
1428 ofp.close()
1425 except: # re-raises
1429 except: # re-raises
1426 try: os.unlink(temp)
1430 try: os.unlink(temp)
1427 except OSError: pass
1431 except OSError: pass
1428 raise
1432 raise
1429 return temp
1433 return temp
1430
1434
1431 class filestat(object):
1435 class filestat(object):
1432 """help to exactly detect change of a file
1436 """help to exactly detect change of a file
1433
1437
1434 'stat' attribute is result of 'os.stat()' if specified 'path'
1438 'stat' attribute is result of 'os.stat()' if specified 'path'
1435 exists. Otherwise, it is None. This can avoid preparative
1439 exists. Otherwise, it is None. This can avoid preparative
1436 'exists()' examination on client side of this class.
1440 'exists()' examination on client side of this class.
1437 """
1441 """
1438 def __init__(self, path):
1442 def __init__(self, path):
1439 try:
1443 try:
1440 self.stat = os.stat(path)
1444 self.stat = os.stat(path)
1441 except OSError as err:
1445 except OSError as err:
1442 if err.errno != errno.ENOENT:
1446 if err.errno != errno.ENOENT:
1443 raise
1447 raise
1444 self.stat = None
1448 self.stat = None
1445
1449
1446 __hash__ = object.__hash__
1450 __hash__ = object.__hash__
1447
1451
1448 def __eq__(self, old):
1452 def __eq__(self, old):
1449 try:
1453 try:
1450 # if ambiguity between stat of new and old file is
1454 # if ambiguity between stat of new and old file is
1451 # avoided, comparision of size, ctime and mtime is enough
1455 # avoided, comparision of size, ctime and mtime is enough
1452 # to exactly detect change of a file regardless of platform
1456 # to exactly detect change of a file regardless of platform
1453 return (self.stat.st_size == old.stat.st_size and
1457 return (self.stat.st_size == old.stat.st_size and
1454 self.stat.st_ctime == old.stat.st_ctime and
1458 self.stat.st_ctime == old.stat.st_ctime and
1455 self.stat.st_mtime == old.stat.st_mtime)
1459 self.stat.st_mtime == old.stat.st_mtime)
1456 except AttributeError:
1460 except AttributeError:
1457 return False
1461 return False
1458
1462
1459 def isambig(self, old):
1463 def isambig(self, old):
1460 """Examine whether new (= self) stat is ambiguous against old one
1464 """Examine whether new (= self) stat is ambiguous against old one
1461
1465
1462 "S[N]" below means stat of a file at N-th change:
1466 "S[N]" below means stat of a file at N-th change:
1463
1467
1464 - S[n-1].ctime < S[n].ctime: can detect change of a file
1468 - S[n-1].ctime < S[n].ctime: can detect change of a file
1465 - S[n-1].ctime == S[n].ctime
1469 - S[n-1].ctime == S[n].ctime
1466 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1470 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1467 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1471 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1468 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1472 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1469 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1473 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1470
1474
1471 Case (*2) above means that a file was changed twice or more at
1475 Case (*2) above means that a file was changed twice or more at
1472 same time in sec (= S[n-1].ctime), and comparison of timestamp
1476 same time in sec (= S[n-1].ctime), and comparison of timestamp
1473 is ambiguous.
1477 is ambiguous.
1474
1478
1475 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1479 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1476 timestamp is ambiguous".
1480 timestamp is ambiguous".
1477
1481
1478 But advancing mtime only in case (*2) doesn't work as
1482 But advancing mtime only in case (*2) doesn't work as
1479 expected, because naturally advanced S[n].mtime in case (*1)
1483 expected, because naturally advanced S[n].mtime in case (*1)
1480 might be equal to manually advanced S[n-1 or earlier].mtime.
1484 might be equal to manually advanced S[n-1 or earlier].mtime.
1481
1485
1482 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1486 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1483 treated as ambiguous regardless of mtime, to avoid overlooking
1487 treated as ambiguous regardless of mtime, to avoid overlooking
1484 by confliction between such mtime.
1488 by confliction between such mtime.
1485
1489
1486 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1490 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1487 S[n].mtime", even if size of a file isn't changed.
1491 S[n].mtime", even if size of a file isn't changed.
1488 """
1492 """
1489 try:
1493 try:
1490 return (self.stat.st_ctime == old.stat.st_ctime)
1494 return (self.stat.st_ctime == old.stat.st_ctime)
1491 except AttributeError:
1495 except AttributeError:
1492 return False
1496 return False
1493
1497
1494 def __ne__(self, other):
1498 def __ne__(self, other):
1495 return not self == other
1499 return not self == other
1496
1500
1497 class atomictempfile(object):
1501 class atomictempfile(object):
1498 '''writable file object that atomically updates a file
1502 '''writable file object that atomically updates a file
1499
1503
1500 All writes will go to a temporary copy of the original file. Call
1504 All writes will go to a temporary copy of the original file. Call
1501 close() when you are done writing, and atomictempfile will rename
1505 close() when you are done writing, and atomictempfile will rename
1502 the temporary copy to the original name, making the changes
1506 the temporary copy to the original name, making the changes
1503 visible. If the object is destroyed without being closed, all your
1507 visible. If the object is destroyed without being closed, all your
1504 writes are discarded.
1508 writes are discarded.
1505
1509
1506 checkambig argument of constructor is used with filestat, and is
1510 checkambig argument of constructor is used with filestat, and is
1507 useful only if target file is guarded by any lock (e.g. repo.lock
1511 useful only if target file is guarded by any lock (e.g. repo.lock
1508 or repo.wlock).
1512 or repo.wlock).
1509 '''
1513 '''
1510 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1514 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1511 self.__name = name # permanent name
1515 self.__name = name # permanent name
1512 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1516 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1513 createmode=createmode)
1517 createmode=createmode)
1514 self._fp = posixfile(self._tempname, mode)
1518 self._fp = posixfile(self._tempname, mode)
1515 self._checkambig = checkambig
1519 self._checkambig = checkambig
1516
1520
1517 # delegated methods
1521 # delegated methods
1518 self.read = self._fp.read
1522 self.read = self._fp.read
1519 self.write = self._fp.write
1523 self.write = self._fp.write
1520 self.seek = self._fp.seek
1524 self.seek = self._fp.seek
1521 self.tell = self._fp.tell
1525 self.tell = self._fp.tell
1522 self.fileno = self._fp.fileno
1526 self.fileno = self._fp.fileno
1523
1527
1524 def close(self):
1528 def close(self):
1525 if not self._fp.closed:
1529 if not self._fp.closed:
1526 self._fp.close()
1530 self._fp.close()
1527 filename = localpath(self.__name)
1531 filename = localpath(self.__name)
1528 oldstat = self._checkambig and filestat(filename)
1532 oldstat = self._checkambig and filestat(filename)
1529 if oldstat and oldstat.stat:
1533 if oldstat and oldstat.stat:
1530 rename(self._tempname, filename)
1534 rename(self._tempname, filename)
1531 newstat = filestat(filename)
1535 newstat = filestat(filename)
1532 if newstat.isambig(oldstat):
1536 if newstat.isambig(oldstat):
1533 # stat of changed file is ambiguous to original one
1537 # stat of changed file is ambiguous to original one
1534 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1538 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1535 os.utime(filename, (advanced, advanced))
1539 os.utime(filename, (advanced, advanced))
1536 else:
1540 else:
1537 rename(self._tempname, filename)
1541 rename(self._tempname, filename)
1538
1542
1539 def discard(self):
1543 def discard(self):
1540 if not self._fp.closed:
1544 if not self._fp.closed:
1541 try:
1545 try:
1542 os.unlink(self._tempname)
1546 os.unlink(self._tempname)
1543 except OSError:
1547 except OSError:
1544 pass
1548 pass
1545 self._fp.close()
1549 self._fp.close()
1546
1550
1547 def __del__(self):
1551 def __del__(self):
1548 if safehasattr(self, '_fp'): # constructor actually did something
1552 if safehasattr(self, '_fp'): # constructor actually did something
1549 self.discard()
1553 self.discard()
1550
1554
1551 def __enter__(self):
1555 def __enter__(self):
1552 return self
1556 return self
1553
1557
1554 def __exit__(self, exctype, excvalue, traceback):
1558 def __exit__(self, exctype, excvalue, traceback):
1555 if exctype is not None:
1559 if exctype is not None:
1556 self.discard()
1560 self.discard()
1557 else:
1561 else:
1558 self.close()
1562 self.close()
1559
1563
1560 def makedirs(name, mode=None, notindexed=False):
1564 def makedirs(name, mode=None, notindexed=False):
1561 """recursive directory creation with parent mode inheritance
1565 """recursive directory creation with parent mode inheritance
1562
1566
1563 Newly created directories are marked as "not to be indexed by
1567 Newly created directories are marked as "not to be indexed by
1564 the content indexing service", if ``notindexed`` is specified
1568 the content indexing service", if ``notindexed`` is specified
1565 for "write" mode access.
1569 for "write" mode access.
1566 """
1570 """
1567 try:
1571 try:
1568 makedir(name, notindexed)
1572 makedir(name, notindexed)
1569 except OSError as err:
1573 except OSError as err:
1570 if err.errno == errno.EEXIST:
1574 if err.errno == errno.EEXIST:
1571 return
1575 return
1572 if err.errno != errno.ENOENT or not name:
1576 if err.errno != errno.ENOENT or not name:
1573 raise
1577 raise
1574 parent = os.path.dirname(os.path.abspath(name))
1578 parent = os.path.dirname(os.path.abspath(name))
1575 if parent == name:
1579 if parent == name:
1576 raise
1580 raise
1577 makedirs(parent, mode, notindexed)
1581 makedirs(parent, mode, notindexed)
1578 try:
1582 try:
1579 makedir(name, notindexed)
1583 makedir(name, notindexed)
1580 except OSError as err:
1584 except OSError as err:
1581 # Catch EEXIST to handle races
1585 # Catch EEXIST to handle races
1582 if err.errno == errno.EEXIST:
1586 if err.errno == errno.EEXIST:
1583 return
1587 return
1584 raise
1588 raise
1585 if mode is not None:
1589 if mode is not None:
1586 os.chmod(name, mode)
1590 os.chmod(name, mode)
1587
1591
1588 def readfile(path):
1592 def readfile(path):
1589 with open(path, 'rb') as fp:
1593 with open(path, 'rb') as fp:
1590 return fp.read()
1594 return fp.read()
1591
1595
1592 def writefile(path, text):
1596 def writefile(path, text):
1593 with open(path, 'wb') as fp:
1597 with open(path, 'wb') as fp:
1594 fp.write(text)
1598 fp.write(text)
1595
1599
1596 def appendfile(path, text):
1600 def appendfile(path, text):
1597 with open(path, 'ab') as fp:
1601 with open(path, 'ab') as fp:
1598 fp.write(text)
1602 fp.write(text)
1599
1603
1600 class chunkbuffer(object):
1604 class chunkbuffer(object):
1601 """Allow arbitrary sized chunks of data to be efficiently read from an
1605 """Allow arbitrary sized chunks of data to be efficiently read from an
1602 iterator over chunks of arbitrary size."""
1606 iterator over chunks of arbitrary size."""
1603
1607
1604 def __init__(self, in_iter):
1608 def __init__(self, in_iter):
1605 """in_iter is the iterator that's iterating over the input chunks.
1609 """in_iter is the iterator that's iterating over the input chunks.
1606 targetsize is how big a buffer to try to maintain."""
1610 targetsize is how big a buffer to try to maintain."""
1607 def splitbig(chunks):
1611 def splitbig(chunks):
1608 for chunk in chunks:
1612 for chunk in chunks:
1609 if len(chunk) > 2**20:
1613 if len(chunk) > 2**20:
1610 pos = 0
1614 pos = 0
1611 while pos < len(chunk):
1615 while pos < len(chunk):
1612 end = pos + 2 ** 18
1616 end = pos + 2 ** 18
1613 yield chunk[pos:end]
1617 yield chunk[pos:end]
1614 pos = end
1618 pos = end
1615 else:
1619 else:
1616 yield chunk
1620 yield chunk
1617 self.iter = splitbig(in_iter)
1621 self.iter = splitbig(in_iter)
1618 self._queue = collections.deque()
1622 self._queue = collections.deque()
1619 self._chunkoffset = 0
1623 self._chunkoffset = 0
1620
1624
1621 def read(self, l=None):
1625 def read(self, l=None):
1622 """Read L bytes of data from the iterator of chunks of data.
1626 """Read L bytes of data from the iterator of chunks of data.
1623 Returns less than L bytes if the iterator runs dry.
1627 Returns less than L bytes if the iterator runs dry.
1624
1628
1625 If size parameter is omitted, read everything"""
1629 If size parameter is omitted, read everything"""
1626 if l is None:
1630 if l is None:
1627 return ''.join(self.iter)
1631 return ''.join(self.iter)
1628
1632
1629 left = l
1633 left = l
1630 buf = []
1634 buf = []
1631 queue = self._queue
1635 queue = self._queue
1632 while left > 0:
1636 while left > 0:
1633 # refill the queue
1637 # refill the queue
1634 if not queue:
1638 if not queue:
1635 target = 2**18
1639 target = 2**18
1636 for chunk in self.iter:
1640 for chunk in self.iter:
1637 queue.append(chunk)
1641 queue.append(chunk)
1638 target -= len(chunk)
1642 target -= len(chunk)
1639 if target <= 0:
1643 if target <= 0:
1640 break
1644 break
1641 if not queue:
1645 if not queue:
1642 break
1646 break
1643
1647
1644 # The easy way to do this would be to queue.popleft(), modify the
1648 # The easy way to do this would be to queue.popleft(), modify the
1645 # chunk (if necessary), then queue.appendleft(). However, for cases
1649 # chunk (if necessary), then queue.appendleft(). However, for cases
1646 # where we read partial chunk content, this incurs 2 dequeue
1650 # where we read partial chunk content, this incurs 2 dequeue
1647 # mutations and creates a new str for the remaining chunk in the
1651 # mutations and creates a new str for the remaining chunk in the
1648 # queue. Our code below avoids this overhead.
1652 # queue. Our code below avoids this overhead.
1649
1653
1650 chunk = queue[0]
1654 chunk = queue[0]
1651 chunkl = len(chunk)
1655 chunkl = len(chunk)
1652 offset = self._chunkoffset
1656 offset = self._chunkoffset
1653
1657
1654 # Use full chunk.
1658 # Use full chunk.
1655 if offset == 0 and left >= chunkl:
1659 if offset == 0 and left >= chunkl:
1656 left -= chunkl
1660 left -= chunkl
1657 queue.popleft()
1661 queue.popleft()
1658 buf.append(chunk)
1662 buf.append(chunk)
1659 # self._chunkoffset remains at 0.
1663 # self._chunkoffset remains at 0.
1660 continue
1664 continue
1661
1665
1662 chunkremaining = chunkl - offset
1666 chunkremaining = chunkl - offset
1663
1667
1664 # Use all of unconsumed part of chunk.
1668 # Use all of unconsumed part of chunk.
1665 if left >= chunkremaining:
1669 if left >= chunkremaining:
1666 left -= chunkremaining
1670 left -= chunkremaining
1667 queue.popleft()
1671 queue.popleft()
1668 # offset == 0 is enabled by block above, so this won't merely
1672 # offset == 0 is enabled by block above, so this won't merely
1669 # copy via ``chunk[0:]``.
1673 # copy via ``chunk[0:]``.
1670 buf.append(chunk[offset:])
1674 buf.append(chunk[offset:])
1671 self._chunkoffset = 0
1675 self._chunkoffset = 0
1672
1676
1673 # Partial chunk needed.
1677 # Partial chunk needed.
1674 else:
1678 else:
1675 buf.append(chunk[offset:offset + left])
1679 buf.append(chunk[offset:offset + left])
1676 self._chunkoffset += left
1680 self._chunkoffset += left
1677 left -= chunkremaining
1681 left -= chunkremaining
1678
1682
1679 return ''.join(buf)
1683 return ''.join(buf)
1680
1684
1681 def filechunkiter(f, size=65536, limit=None):
1685 def filechunkiter(f, size=65536, limit=None):
1682 """Create a generator that produces the data in the file size
1686 """Create a generator that produces the data in the file size
1683 (default 65536) bytes at a time, up to optional limit (default is
1687 (default 65536) bytes at a time, up to optional limit (default is
1684 to read all data). Chunks may be less than size bytes if the
1688 to read all data). Chunks may be less than size bytes if the
1685 chunk is the last chunk in the file, or the file is a socket or
1689 chunk is the last chunk in the file, or the file is a socket or
1686 some other type of file that sometimes reads less data than is
1690 some other type of file that sometimes reads less data than is
1687 requested."""
1691 requested."""
1688 assert size >= 0
1692 assert size >= 0
1689 assert limit is None or limit >= 0
1693 assert limit is None or limit >= 0
1690 while True:
1694 while True:
1691 if limit is None:
1695 if limit is None:
1692 nbytes = size
1696 nbytes = size
1693 else:
1697 else:
1694 nbytes = min(limit, size)
1698 nbytes = min(limit, size)
1695 s = nbytes and f.read(nbytes)
1699 s = nbytes and f.read(nbytes)
1696 if not s:
1700 if not s:
1697 break
1701 break
1698 if limit:
1702 if limit:
1699 limit -= len(s)
1703 limit -= len(s)
1700 yield s
1704 yield s
1701
1705
1702 def makedate(timestamp=None):
1706 def makedate(timestamp=None):
1703 '''Return a unix timestamp (or the current time) as a (unixtime,
1707 '''Return a unix timestamp (or the current time) as a (unixtime,
1704 offset) tuple based off the local timezone.'''
1708 offset) tuple based off the local timezone.'''
1705 if timestamp is None:
1709 if timestamp is None:
1706 timestamp = time.time()
1710 timestamp = time.time()
1707 if timestamp < 0:
1711 if timestamp < 0:
1708 hint = _("check your clock")
1712 hint = _("check your clock")
1709 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1713 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1710 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1714 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1711 datetime.datetime.fromtimestamp(timestamp))
1715 datetime.datetime.fromtimestamp(timestamp))
1712 tz = delta.days * 86400 + delta.seconds
1716 tz = delta.days * 86400 + delta.seconds
1713 return timestamp, tz
1717 return timestamp, tz
1714
1718
1715 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1719 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1716 """represent a (unixtime, offset) tuple as a localized time.
1720 """represent a (unixtime, offset) tuple as a localized time.
1717 unixtime is seconds since the epoch, and offset is the time zone's
1721 unixtime is seconds since the epoch, and offset is the time zone's
1718 number of seconds away from UTC.
1722 number of seconds away from UTC.
1719
1723
1720 >>> datestr((0, 0))
1724 >>> datestr((0, 0))
1721 'Thu Jan 01 00:00:00 1970 +0000'
1725 'Thu Jan 01 00:00:00 1970 +0000'
1722 >>> datestr((42, 0))
1726 >>> datestr((42, 0))
1723 'Thu Jan 01 00:00:42 1970 +0000'
1727 'Thu Jan 01 00:00:42 1970 +0000'
1724 >>> datestr((-42, 0))
1728 >>> datestr((-42, 0))
1725 'Wed Dec 31 23:59:18 1969 +0000'
1729 'Wed Dec 31 23:59:18 1969 +0000'
1726 >>> datestr((0x7fffffff, 0))
1730 >>> datestr((0x7fffffff, 0))
1727 'Tue Jan 19 03:14:07 2038 +0000'
1731 'Tue Jan 19 03:14:07 2038 +0000'
1728 >>> datestr((-0x80000000, 0))
1732 >>> datestr((-0x80000000, 0))
1729 'Fri Dec 13 20:45:52 1901 +0000'
1733 'Fri Dec 13 20:45:52 1901 +0000'
1730 """
1734 """
1731 t, tz = date or makedate()
1735 t, tz = date or makedate()
1732 if "%1" in format or "%2" in format or "%z" in format:
1736 if "%1" in format or "%2" in format or "%z" in format:
1733 sign = (tz > 0) and "-" or "+"
1737 sign = (tz > 0) and "-" or "+"
1734 minutes = abs(tz) // 60
1738 minutes = abs(tz) // 60
1735 q, r = divmod(minutes, 60)
1739 q, r = divmod(minutes, 60)
1736 format = format.replace("%z", "%1%2")
1740 format = format.replace("%z", "%1%2")
1737 format = format.replace("%1", "%c%02d" % (sign, q))
1741 format = format.replace("%1", "%c%02d" % (sign, q))
1738 format = format.replace("%2", "%02d" % r)
1742 format = format.replace("%2", "%02d" % r)
1739 d = t - tz
1743 d = t - tz
1740 if d > 0x7fffffff:
1744 if d > 0x7fffffff:
1741 d = 0x7fffffff
1745 d = 0x7fffffff
1742 elif d < -0x80000000:
1746 elif d < -0x80000000:
1743 d = -0x80000000
1747 d = -0x80000000
1744 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1748 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1745 # because they use the gmtime() system call which is buggy on Windows
1749 # because they use the gmtime() system call which is buggy on Windows
1746 # for negative values.
1750 # for negative values.
1747 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1751 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1748 s = t.strftime(format)
1752 s = t.strftime(format)
1749 return s
1753 return s
1750
1754
1751 def shortdate(date=None):
1755 def shortdate(date=None):
1752 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1756 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1753 return datestr(date, format='%Y-%m-%d')
1757 return datestr(date, format='%Y-%m-%d')
1754
1758
1755 def parsetimezone(s):
1759 def parsetimezone(s):
1756 """find a trailing timezone, if any, in string, and return a
1760 """find a trailing timezone, if any, in string, and return a
1757 (offset, remainder) pair"""
1761 (offset, remainder) pair"""
1758
1762
1759 if s.endswith("GMT") or s.endswith("UTC"):
1763 if s.endswith("GMT") or s.endswith("UTC"):
1760 return 0, s[:-3].rstrip()
1764 return 0, s[:-3].rstrip()
1761
1765
1762 # Unix-style timezones [+-]hhmm
1766 # Unix-style timezones [+-]hhmm
1763 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1767 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1764 sign = (s[-5] == "+") and 1 or -1
1768 sign = (s[-5] == "+") and 1 or -1
1765 hours = int(s[-4:-2])
1769 hours = int(s[-4:-2])
1766 minutes = int(s[-2:])
1770 minutes = int(s[-2:])
1767 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1771 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1768
1772
1769 # ISO8601 trailing Z
1773 # ISO8601 trailing Z
1770 if s.endswith("Z") and s[-2:-1].isdigit():
1774 if s.endswith("Z") and s[-2:-1].isdigit():
1771 return 0, s[:-1]
1775 return 0, s[:-1]
1772
1776
1773 # ISO8601-style [+-]hh:mm
1777 # ISO8601-style [+-]hh:mm
1774 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1778 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1775 s[-5:-3].isdigit() and s[-2:].isdigit()):
1779 s[-5:-3].isdigit() and s[-2:].isdigit()):
1776 sign = (s[-6] == "+") and 1 or -1
1780 sign = (s[-6] == "+") and 1 or -1
1777 hours = int(s[-5:-3])
1781 hours = int(s[-5:-3])
1778 minutes = int(s[-2:])
1782 minutes = int(s[-2:])
1779 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1783 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1780
1784
1781 return None, s
1785 return None, s
1782
1786
1783 def strdate(string, format, defaults=[]):
1787 def strdate(string, format, defaults=[]):
1784 """parse a localized time string and return a (unixtime, offset) tuple.
1788 """parse a localized time string and return a (unixtime, offset) tuple.
1785 if the string cannot be parsed, ValueError is raised."""
1789 if the string cannot be parsed, ValueError is raised."""
1786 # NOTE: unixtime = localunixtime + offset
1790 # NOTE: unixtime = localunixtime + offset
1787 offset, date = parsetimezone(string)
1791 offset, date = parsetimezone(string)
1788
1792
1789 # add missing elements from defaults
1793 # add missing elements from defaults
1790 usenow = False # default to using biased defaults
1794 usenow = False # default to using biased defaults
1791 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1795 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1792 found = [True for p in part if ("%"+p) in format]
1796 found = [True for p in part if ("%"+p) in format]
1793 if not found:
1797 if not found:
1794 date += "@" + defaults[part][usenow]
1798 date += "@" + defaults[part][usenow]
1795 format += "@%" + part[0]
1799 format += "@%" + part[0]
1796 else:
1800 else:
1797 # We've found a specific time element, less specific time
1801 # We've found a specific time element, less specific time
1798 # elements are relative to today
1802 # elements are relative to today
1799 usenow = True
1803 usenow = True
1800
1804
1801 timetuple = time.strptime(date, format)
1805 timetuple = time.strptime(date, format)
1802 localunixtime = int(calendar.timegm(timetuple))
1806 localunixtime = int(calendar.timegm(timetuple))
1803 if offset is None:
1807 if offset is None:
1804 # local timezone
1808 # local timezone
1805 unixtime = int(time.mktime(timetuple))
1809 unixtime = int(time.mktime(timetuple))
1806 offset = unixtime - localunixtime
1810 offset = unixtime - localunixtime
1807 else:
1811 else:
1808 unixtime = localunixtime + offset
1812 unixtime = localunixtime + offset
1809 return unixtime, offset
1813 return unixtime, offset
1810
1814
1811 def parsedate(date, formats=None, bias=None):
1815 def parsedate(date, formats=None, bias=None):
1812 """parse a localized date/time and return a (unixtime, offset) tuple.
1816 """parse a localized date/time and return a (unixtime, offset) tuple.
1813
1817
1814 The date may be a "unixtime offset" string or in one of the specified
1818 The date may be a "unixtime offset" string or in one of the specified
1815 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1819 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1816
1820
1817 >>> parsedate(' today ') == parsedate(\
1821 >>> parsedate(' today ') == parsedate(\
1818 datetime.date.today().strftime('%b %d'))
1822 datetime.date.today().strftime('%b %d'))
1819 True
1823 True
1820 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1824 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1821 datetime.timedelta(days=1)\
1825 datetime.timedelta(days=1)\
1822 ).strftime('%b %d'))
1826 ).strftime('%b %d'))
1823 True
1827 True
1824 >>> now, tz = makedate()
1828 >>> now, tz = makedate()
1825 >>> strnow, strtz = parsedate('now')
1829 >>> strnow, strtz = parsedate('now')
1826 >>> (strnow - now) < 1
1830 >>> (strnow - now) < 1
1827 True
1831 True
1828 >>> tz == strtz
1832 >>> tz == strtz
1829 True
1833 True
1830 """
1834 """
1831 if bias is None:
1835 if bias is None:
1832 bias = {}
1836 bias = {}
1833 if not date:
1837 if not date:
1834 return 0, 0
1838 return 0, 0
1835 if isinstance(date, tuple) and len(date) == 2:
1839 if isinstance(date, tuple) and len(date) == 2:
1836 return date
1840 return date
1837 if not formats:
1841 if not formats:
1838 formats = defaultdateformats
1842 formats = defaultdateformats
1839 date = date.strip()
1843 date = date.strip()
1840
1844
1841 if date == 'now' or date == _('now'):
1845 if date == 'now' or date == _('now'):
1842 return makedate()
1846 return makedate()
1843 if date == 'today' or date == _('today'):
1847 if date == 'today' or date == _('today'):
1844 date = datetime.date.today().strftime('%b %d')
1848 date = datetime.date.today().strftime('%b %d')
1845 elif date == 'yesterday' or date == _('yesterday'):
1849 elif date == 'yesterday' or date == _('yesterday'):
1846 date = (datetime.date.today() -
1850 date = (datetime.date.today() -
1847 datetime.timedelta(days=1)).strftime('%b %d')
1851 datetime.timedelta(days=1)).strftime('%b %d')
1848
1852
1849 try:
1853 try:
1850 when, offset = map(int, date.split(' '))
1854 when, offset = map(int, date.split(' '))
1851 except ValueError:
1855 except ValueError:
1852 # fill out defaults
1856 # fill out defaults
1853 now = makedate()
1857 now = makedate()
1854 defaults = {}
1858 defaults = {}
1855 for part in ("d", "mb", "yY", "HI", "M", "S"):
1859 for part in ("d", "mb", "yY", "HI", "M", "S"):
1856 # this piece is for rounding the specific end of unknowns
1860 # this piece is for rounding the specific end of unknowns
1857 b = bias.get(part)
1861 b = bias.get(part)
1858 if b is None:
1862 if b is None:
1859 if part[0] in "HMS":
1863 if part[0] in "HMS":
1860 b = "00"
1864 b = "00"
1861 else:
1865 else:
1862 b = "0"
1866 b = "0"
1863
1867
1864 # this piece is for matching the generic end to today's date
1868 # this piece is for matching the generic end to today's date
1865 n = datestr(now, "%" + part[0])
1869 n = datestr(now, "%" + part[0])
1866
1870
1867 defaults[part] = (b, n)
1871 defaults[part] = (b, n)
1868
1872
1869 for format in formats:
1873 for format in formats:
1870 try:
1874 try:
1871 when, offset = strdate(date, format, defaults)
1875 when, offset = strdate(date, format, defaults)
1872 except (ValueError, OverflowError):
1876 except (ValueError, OverflowError):
1873 pass
1877 pass
1874 else:
1878 else:
1875 break
1879 break
1876 else:
1880 else:
1877 raise Abort(_('invalid date: %r') % date)
1881 raise Abort(_('invalid date: %r') % date)
1878 # validate explicit (probably user-specified) date and
1882 # validate explicit (probably user-specified) date and
1879 # time zone offset. values must fit in signed 32 bits for
1883 # time zone offset. values must fit in signed 32 bits for
1880 # current 32-bit linux runtimes. timezones go from UTC-12
1884 # current 32-bit linux runtimes. timezones go from UTC-12
1881 # to UTC+14
1885 # to UTC+14
1882 if when < -0x80000000 or when > 0x7fffffff:
1886 if when < -0x80000000 or when > 0x7fffffff:
1883 raise Abort(_('date exceeds 32 bits: %d') % when)
1887 raise Abort(_('date exceeds 32 bits: %d') % when)
1884 if offset < -50400 or offset > 43200:
1888 if offset < -50400 or offset > 43200:
1885 raise Abort(_('impossible time zone offset: %d') % offset)
1889 raise Abort(_('impossible time zone offset: %d') % offset)
1886 return when, offset
1890 return when, offset
1887
1891
1888 def matchdate(date):
1892 def matchdate(date):
1889 """Return a function that matches a given date match specifier
1893 """Return a function that matches a given date match specifier
1890
1894
1891 Formats include:
1895 Formats include:
1892
1896
1893 '{date}' match a given date to the accuracy provided
1897 '{date}' match a given date to the accuracy provided
1894
1898
1895 '<{date}' on or before a given date
1899 '<{date}' on or before a given date
1896
1900
1897 '>{date}' on or after a given date
1901 '>{date}' on or after a given date
1898
1902
1899 >>> p1 = parsedate("10:29:59")
1903 >>> p1 = parsedate("10:29:59")
1900 >>> p2 = parsedate("10:30:00")
1904 >>> p2 = parsedate("10:30:00")
1901 >>> p3 = parsedate("10:30:59")
1905 >>> p3 = parsedate("10:30:59")
1902 >>> p4 = parsedate("10:31:00")
1906 >>> p4 = parsedate("10:31:00")
1903 >>> p5 = parsedate("Sep 15 10:30:00 1999")
1907 >>> p5 = parsedate("Sep 15 10:30:00 1999")
1904 >>> f = matchdate("10:30")
1908 >>> f = matchdate("10:30")
1905 >>> f(p1[0])
1909 >>> f(p1[0])
1906 False
1910 False
1907 >>> f(p2[0])
1911 >>> f(p2[0])
1908 True
1912 True
1909 >>> f(p3[0])
1913 >>> f(p3[0])
1910 True
1914 True
1911 >>> f(p4[0])
1915 >>> f(p4[0])
1912 False
1916 False
1913 >>> f(p5[0])
1917 >>> f(p5[0])
1914 False
1918 False
1915 """
1919 """
1916
1920
1917 def lower(date):
1921 def lower(date):
1918 d = {'mb': "1", 'd': "1"}
1922 d = {'mb': "1", 'd': "1"}
1919 return parsedate(date, extendeddateformats, d)[0]
1923 return parsedate(date, extendeddateformats, d)[0]
1920
1924
1921 def upper(date):
1925 def upper(date):
1922 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
1926 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
1923 for days in ("31", "30", "29"):
1927 for days in ("31", "30", "29"):
1924 try:
1928 try:
1925 d["d"] = days
1929 d["d"] = days
1926 return parsedate(date, extendeddateformats, d)[0]
1930 return parsedate(date, extendeddateformats, d)[0]
1927 except Abort:
1931 except Abort:
1928 pass
1932 pass
1929 d["d"] = "28"
1933 d["d"] = "28"
1930 return parsedate(date, extendeddateformats, d)[0]
1934 return parsedate(date, extendeddateformats, d)[0]
1931
1935
1932 date = date.strip()
1936 date = date.strip()
1933
1937
1934 if not date:
1938 if not date:
1935 raise Abort(_("dates cannot consist entirely of whitespace"))
1939 raise Abort(_("dates cannot consist entirely of whitespace"))
1936 elif date[0] == "<":
1940 elif date[0] == "<":
1937 if not date[1:]:
1941 if not date[1:]:
1938 raise Abort(_("invalid day spec, use '<DATE'"))
1942 raise Abort(_("invalid day spec, use '<DATE'"))
1939 when = upper(date[1:])
1943 when = upper(date[1:])
1940 return lambda x: x <= when
1944 return lambda x: x <= when
1941 elif date[0] == ">":
1945 elif date[0] == ">":
1942 if not date[1:]:
1946 if not date[1:]:
1943 raise Abort(_("invalid day spec, use '>DATE'"))
1947 raise Abort(_("invalid day spec, use '>DATE'"))
1944 when = lower(date[1:])
1948 when = lower(date[1:])
1945 return lambda x: x >= when
1949 return lambda x: x >= when
1946 elif date[0] == "-":
1950 elif date[0] == "-":
1947 try:
1951 try:
1948 days = int(date[1:])
1952 days = int(date[1:])
1949 except ValueError:
1953 except ValueError:
1950 raise Abort(_("invalid day spec: %s") % date[1:])
1954 raise Abort(_("invalid day spec: %s") % date[1:])
1951 if days < 0:
1955 if days < 0:
1952 raise Abort(_('%s must be nonnegative (see "hg help dates")')
1956 raise Abort(_('%s must be nonnegative (see "hg help dates")')
1953 % date[1:])
1957 % date[1:])
1954 when = makedate()[0] - days * 3600 * 24
1958 when = makedate()[0] - days * 3600 * 24
1955 return lambda x: x >= when
1959 return lambda x: x >= when
1956 elif " to " in date:
1960 elif " to " in date:
1957 a, b = date.split(" to ")
1961 a, b = date.split(" to ")
1958 start, stop = lower(a), upper(b)
1962 start, stop = lower(a), upper(b)
1959 return lambda x: x >= start and x <= stop
1963 return lambda x: x >= start and x <= stop
1960 else:
1964 else:
1961 start, stop = lower(date), upper(date)
1965 start, stop = lower(date), upper(date)
1962 return lambda x: x >= start and x <= stop
1966 return lambda x: x >= start and x <= stop
1963
1967
1964 def stringmatcher(pattern):
1968 def stringmatcher(pattern):
1965 """
1969 """
1966 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1970 accepts a string, possibly starting with 're:' or 'literal:' prefix.
1967 returns the matcher name, pattern, and matcher function.
1971 returns the matcher name, pattern, and matcher function.
1968 missing or unknown prefixes are treated as literal matches.
1972 missing or unknown prefixes are treated as literal matches.
1969
1973
1970 helper for tests:
1974 helper for tests:
1971 >>> def test(pattern, *tests):
1975 >>> def test(pattern, *tests):
1972 ... kind, pattern, matcher = stringmatcher(pattern)
1976 ... kind, pattern, matcher = stringmatcher(pattern)
1973 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1977 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
1974
1978
1975 exact matching (no prefix):
1979 exact matching (no prefix):
1976 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1980 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
1977 ('literal', 'abcdefg', [False, False, True])
1981 ('literal', 'abcdefg', [False, False, True])
1978
1982
1979 regex matching ('re:' prefix)
1983 regex matching ('re:' prefix)
1980 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1984 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
1981 ('re', 'a.+b', [False, False, True])
1985 ('re', 'a.+b', [False, False, True])
1982
1986
1983 force exact matches ('literal:' prefix)
1987 force exact matches ('literal:' prefix)
1984 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1988 >>> test('literal:re:foobar', 'foobar', 're:foobar')
1985 ('literal', 're:foobar', [False, True])
1989 ('literal', 're:foobar', [False, True])
1986
1990
1987 unknown prefixes are ignored and treated as literals
1991 unknown prefixes are ignored and treated as literals
1988 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1992 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
1989 ('literal', 'foo:bar', [False, False, True])
1993 ('literal', 'foo:bar', [False, False, True])
1990 """
1994 """
1991 if pattern.startswith('re:'):
1995 if pattern.startswith('re:'):
1992 pattern = pattern[3:]
1996 pattern = pattern[3:]
1993 try:
1997 try:
1994 regex = remod.compile(pattern)
1998 regex = remod.compile(pattern)
1995 except remod.error as e:
1999 except remod.error as e:
1996 raise error.ParseError(_('invalid regular expression: %s')
2000 raise error.ParseError(_('invalid regular expression: %s')
1997 % e)
2001 % e)
1998 return 're', pattern, regex.search
2002 return 're', pattern, regex.search
1999 elif pattern.startswith('literal:'):
2003 elif pattern.startswith('literal:'):
2000 pattern = pattern[8:]
2004 pattern = pattern[8:]
2001 return 'literal', pattern, pattern.__eq__
2005 return 'literal', pattern, pattern.__eq__
2002
2006
2003 def shortuser(user):
2007 def shortuser(user):
2004 """Return a short representation of a user name or email address."""
2008 """Return a short representation of a user name or email address."""
2005 f = user.find('@')
2009 f = user.find('@')
2006 if f >= 0:
2010 if f >= 0:
2007 user = user[:f]
2011 user = user[:f]
2008 f = user.find('<')
2012 f = user.find('<')
2009 if f >= 0:
2013 if f >= 0:
2010 user = user[f + 1:]
2014 user = user[f + 1:]
2011 f = user.find(' ')
2015 f = user.find(' ')
2012 if f >= 0:
2016 if f >= 0:
2013 user = user[:f]
2017 user = user[:f]
2014 f = user.find('.')
2018 f = user.find('.')
2015 if f >= 0:
2019 if f >= 0:
2016 user = user[:f]
2020 user = user[:f]
2017 return user
2021 return user
2018
2022
2019 def emailuser(user):
2023 def emailuser(user):
2020 """Return the user portion of an email address."""
2024 """Return the user portion of an email address."""
2021 f = user.find('@')
2025 f = user.find('@')
2022 if f >= 0:
2026 if f >= 0:
2023 user = user[:f]
2027 user = user[:f]
2024 f = user.find('<')
2028 f = user.find('<')
2025 if f >= 0:
2029 if f >= 0:
2026 user = user[f + 1:]
2030 user = user[f + 1:]
2027 return user
2031 return user
2028
2032
2029 def email(author):
2033 def email(author):
2030 '''get email of author.'''
2034 '''get email of author.'''
2031 r = author.find('>')
2035 r = author.find('>')
2032 if r == -1:
2036 if r == -1:
2033 r = None
2037 r = None
2034 return author[author.find('<') + 1:r]
2038 return author[author.find('<') + 1:r]
2035
2039
2036 def ellipsis(text, maxlength=400):
2040 def ellipsis(text, maxlength=400):
2037 """Trim string to at most maxlength (default: 400) columns in display."""
2041 """Trim string to at most maxlength (default: 400) columns in display."""
2038 return encoding.trim(text, maxlength, ellipsis='...')
2042 return encoding.trim(text, maxlength, ellipsis='...')
2039
2043
2040 def unitcountfn(*unittable):
2044 def unitcountfn(*unittable):
2041 '''return a function that renders a readable count of some quantity'''
2045 '''return a function that renders a readable count of some quantity'''
2042
2046
2043 def go(count):
2047 def go(count):
2044 for multiplier, divisor, format in unittable:
2048 for multiplier, divisor, format in unittable:
2045 if count >= divisor * multiplier:
2049 if count >= divisor * multiplier:
2046 return format % (count / float(divisor))
2050 return format % (count / float(divisor))
2047 return unittable[-1][2] % count
2051 return unittable[-1][2] % count
2048
2052
2049 return go
2053 return go
2050
2054
2051 bytecount = unitcountfn(
2055 bytecount = unitcountfn(
2052 (100, 1 << 30, _('%.0f GB')),
2056 (100, 1 << 30, _('%.0f GB')),
2053 (10, 1 << 30, _('%.1f GB')),
2057 (10, 1 << 30, _('%.1f GB')),
2054 (1, 1 << 30, _('%.2f GB')),
2058 (1, 1 << 30, _('%.2f GB')),
2055 (100, 1 << 20, _('%.0f MB')),
2059 (100, 1 << 20, _('%.0f MB')),
2056 (10, 1 << 20, _('%.1f MB')),
2060 (10, 1 << 20, _('%.1f MB')),
2057 (1, 1 << 20, _('%.2f MB')),
2061 (1, 1 << 20, _('%.2f MB')),
2058 (100, 1 << 10, _('%.0f KB')),
2062 (100, 1 << 10, _('%.0f KB')),
2059 (10, 1 << 10, _('%.1f KB')),
2063 (10, 1 << 10, _('%.1f KB')),
2060 (1, 1 << 10, _('%.2f KB')),
2064 (1, 1 << 10, _('%.2f KB')),
2061 (1, 1, _('%.0f bytes')),
2065 (1, 1, _('%.0f bytes')),
2062 )
2066 )
2063
2067
2064 def uirepr(s):
2068 def uirepr(s):
2065 # Avoid double backslash in Windows path repr()
2069 # Avoid double backslash in Windows path repr()
2066 return repr(s).replace('\\\\', '\\')
2070 return repr(s).replace('\\\\', '\\')
2067
2071
2068 # delay import of textwrap
2072 # delay import of textwrap
2069 def MBTextWrapper(**kwargs):
2073 def MBTextWrapper(**kwargs):
2070 class tw(textwrap.TextWrapper):
2074 class tw(textwrap.TextWrapper):
2071 """
2075 """
2072 Extend TextWrapper for width-awareness.
2076 Extend TextWrapper for width-awareness.
2073
2077
2074 Neither number of 'bytes' in any encoding nor 'characters' is
2078 Neither number of 'bytes' in any encoding nor 'characters' is
2075 appropriate to calculate terminal columns for specified string.
2079 appropriate to calculate terminal columns for specified string.
2076
2080
2077 Original TextWrapper implementation uses built-in 'len()' directly,
2081 Original TextWrapper implementation uses built-in 'len()' directly,
2078 so overriding is needed to use width information of each characters.
2082 so overriding is needed to use width information of each characters.
2079
2083
2080 In addition, characters classified into 'ambiguous' width are
2084 In addition, characters classified into 'ambiguous' width are
2081 treated as wide in East Asian area, but as narrow in other.
2085 treated as wide in East Asian area, but as narrow in other.
2082
2086
2083 This requires use decision to determine width of such characters.
2087 This requires use decision to determine width of such characters.
2084 """
2088 """
2085 def _cutdown(self, ucstr, space_left):
2089 def _cutdown(self, ucstr, space_left):
2086 l = 0
2090 l = 0
2087 colwidth = encoding.ucolwidth
2091 colwidth = encoding.ucolwidth
2088 for i in xrange(len(ucstr)):
2092 for i in xrange(len(ucstr)):
2089 l += colwidth(ucstr[i])
2093 l += colwidth(ucstr[i])
2090 if space_left < l:
2094 if space_left < l:
2091 return (ucstr[:i], ucstr[i:])
2095 return (ucstr[:i], ucstr[i:])
2092 return ucstr, ''
2096 return ucstr, ''
2093
2097
2094 # overriding of base class
2098 # overriding of base class
2095 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2099 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2096 space_left = max(width - cur_len, 1)
2100 space_left = max(width - cur_len, 1)
2097
2101
2098 if self.break_long_words:
2102 if self.break_long_words:
2099 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2103 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2100 cur_line.append(cut)
2104 cur_line.append(cut)
2101 reversed_chunks[-1] = res
2105 reversed_chunks[-1] = res
2102 elif not cur_line:
2106 elif not cur_line:
2103 cur_line.append(reversed_chunks.pop())
2107 cur_line.append(reversed_chunks.pop())
2104
2108
2105 # this overriding code is imported from TextWrapper of Python 2.6
2109 # this overriding code is imported from TextWrapper of Python 2.6
2106 # to calculate columns of string by 'encoding.ucolwidth()'
2110 # to calculate columns of string by 'encoding.ucolwidth()'
2107 def _wrap_chunks(self, chunks):
2111 def _wrap_chunks(self, chunks):
2108 colwidth = encoding.ucolwidth
2112 colwidth = encoding.ucolwidth
2109
2113
2110 lines = []
2114 lines = []
2111 if self.width <= 0:
2115 if self.width <= 0:
2112 raise ValueError("invalid width %r (must be > 0)" % self.width)
2116 raise ValueError("invalid width %r (must be > 0)" % self.width)
2113
2117
2114 # Arrange in reverse order so items can be efficiently popped
2118 # Arrange in reverse order so items can be efficiently popped
2115 # from a stack of chucks.
2119 # from a stack of chucks.
2116 chunks.reverse()
2120 chunks.reverse()
2117
2121
2118 while chunks:
2122 while chunks:
2119
2123
2120 # Start the list of chunks that will make up the current line.
2124 # Start the list of chunks that will make up the current line.
2121 # cur_len is just the length of all the chunks in cur_line.
2125 # cur_len is just the length of all the chunks in cur_line.
2122 cur_line = []
2126 cur_line = []
2123 cur_len = 0
2127 cur_len = 0
2124
2128
2125 # Figure out which static string will prefix this line.
2129 # Figure out which static string will prefix this line.
2126 if lines:
2130 if lines:
2127 indent = self.subsequent_indent
2131 indent = self.subsequent_indent
2128 else:
2132 else:
2129 indent = self.initial_indent
2133 indent = self.initial_indent
2130
2134
2131 # Maximum width for this line.
2135 # Maximum width for this line.
2132 width = self.width - len(indent)
2136 width = self.width - len(indent)
2133
2137
2134 # First chunk on line is whitespace -- drop it, unless this
2138 # First chunk on line is whitespace -- drop it, unless this
2135 # is the very beginning of the text (i.e. no lines started yet).
2139 # is the very beginning of the text (i.e. no lines started yet).
2136 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2140 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2137 del chunks[-1]
2141 del chunks[-1]
2138
2142
2139 while chunks:
2143 while chunks:
2140 l = colwidth(chunks[-1])
2144 l = colwidth(chunks[-1])
2141
2145
2142 # Can at least squeeze this chunk onto the current line.
2146 # Can at least squeeze this chunk onto the current line.
2143 if cur_len + l <= width:
2147 if cur_len + l <= width:
2144 cur_line.append(chunks.pop())
2148 cur_line.append(chunks.pop())
2145 cur_len += l
2149 cur_len += l
2146
2150
2147 # Nope, this line is full.
2151 # Nope, this line is full.
2148 else:
2152 else:
2149 break
2153 break
2150
2154
2151 # The current line is full, and the next chunk is too big to
2155 # The current line is full, and the next chunk is too big to
2152 # fit on *any* line (not just this one).
2156 # fit on *any* line (not just this one).
2153 if chunks and colwidth(chunks[-1]) > width:
2157 if chunks and colwidth(chunks[-1]) > width:
2154 self._handle_long_word(chunks, cur_line, cur_len, width)
2158 self._handle_long_word(chunks, cur_line, cur_len, width)
2155
2159
2156 # If the last chunk on this line is all whitespace, drop it.
2160 # If the last chunk on this line is all whitespace, drop it.
2157 if (self.drop_whitespace and
2161 if (self.drop_whitespace and
2158 cur_line and cur_line[-1].strip() == ''):
2162 cur_line and cur_line[-1].strip() == ''):
2159 del cur_line[-1]
2163 del cur_line[-1]
2160
2164
2161 # Convert current line back to a string and store it in list
2165 # Convert current line back to a string and store it in list
2162 # of all lines (return value).
2166 # of all lines (return value).
2163 if cur_line:
2167 if cur_line:
2164 lines.append(indent + ''.join(cur_line))
2168 lines.append(indent + ''.join(cur_line))
2165
2169
2166 return lines
2170 return lines
2167
2171
2168 global MBTextWrapper
2172 global MBTextWrapper
2169 MBTextWrapper = tw
2173 MBTextWrapper = tw
2170 return tw(**kwargs)
2174 return tw(**kwargs)
2171
2175
2172 def wrap(line, width, initindent='', hangindent=''):
2176 def wrap(line, width, initindent='', hangindent=''):
2173 maxindent = max(len(hangindent), len(initindent))
2177 maxindent = max(len(hangindent), len(initindent))
2174 if width <= maxindent:
2178 if width <= maxindent:
2175 # adjust for weird terminal size
2179 # adjust for weird terminal size
2176 width = max(78, maxindent + 1)
2180 width = max(78, maxindent + 1)
2177 line = line.decode(encoding.encoding, encoding.encodingmode)
2181 line = line.decode(encoding.encoding, encoding.encodingmode)
2178 initindent = initindent.decode(encoding.encoding, encoding.encodingmode)
2182 initindent = initindent.decode(encoding.encoding, encoding.encodingmode)
2179 hangindent = hangindent.decode(encoding.encoding, encoding.encodingmode)
2183 hangindent = hangindent.decode(encoding.encoding, encoding.encodingmode)
2180 wrapper = MBTextWrapper(width=width,
2184 wrapper = MBTextWrapper(width=width,
2181 initial_indent=initindent,
2185 initial_indent=initindent,
2182 subsequent_indent=hangindent)
2186 subsequent_indent=hangindent)
2183 return wrapper.fill(line).encode(encoding.encoding)
2187 return wrapper.fill(line).encode(encoding.encoding)
2184
2188
2185 def iterlines(iterator):
2189 def iterlines(iterator):
2186 for chunk in iterator:
2190 for chunk in iterator:
2187 for line in chunk.splitlines():
2191 for line in chunk.splitlines():
2188 yield line
2192 yield line
2189
2193
2190 def expandpath(path):
2194 def expandpath(path):
2191 return os.path.expanduser(os.path.expandvars(path))
2195 return os.path.expanduser(os.path.expandvars(path))
2192
2196
2193 def hgcmd():
2197 def hgcmd():
2194 """Return the command used to execute current hg
2198 """Return the command used to execute current hg
2195
2199
2196 This is different from hgexecutable() because on Windows we want
2200 This is different from hgexecutable() because on Windows we want
2197 to avoid things opening new shell windows like batch files, so we
2201 to avoid things opening new shell windows like batch files, so we
2198 get either the python call or current executable.
2202 get either the python call or current executable.
2199 """
2203 """
2200 if mainfrozen():
2204 if mainfrozen():
2201 if getattr(sys, 'frozen', None) == 'macosx_app':
2205 if getattr(sys, 'frozen', None) == 'macosx_app':
2202 # Env variable set by py2app
2206 # Env variable set by py2app
2203 return [os.environ['EXECUTABLEPATH']]
2207 return [os.environ['EXECUTABLEPATH']]
2204 else:
2208 else:
2205 return [sys.executable]
2209 return [sys.executable]
2206 return gethgcmd()
2210 return gethgcmd()
2207
2211
2208 def rundetached(args, condfn):
2212 def rundetached(args, condfn):
2209 """Execute the argument list in a detached process.
2213 """Execute the argument list in a detached process.
2210
2214
2211 condfn is a callable which is called repeatedly and should return
2215 condfn is a callable which is called repeatedly and should return
2212 True once the child process is known to have started successfully.
2216 True once the child process is known to have started successfully.
2213 At this point, the child process PID is returned. If the child
2217 At this point, the child process PID is returned. If the child
2214 process fails to start or finishes before condfn() evaluates to
2218 process fails to start or finishes before condfn() evaluates to
2215 True, return -1.
2219 True, return -1.
2216 """
2220 """
2217 # Windows case is easier because the child process is either
2221 # Windows case is easier because the child process is either
2218 # successfully starting and validating the condition or exiting
2222 # successfully starting and validating the condition or exiting
2219 # on failure. We just poll on its PID. On Unix, if the child
2223 # on failure. We just poll on its PID. On Unix, if the child
2220 # process fails to start, it will be left in a zombie state until
2224 # process fails to start, it will be left in a zombie state until
2221 # the parent wait on it, which we cannot do since we expect a long
2225 # the parent wait on it, which we cannot do since we expect a long
2222 # running process on success. Instead we listen for SIGCHLD telling
2226 # running process on success. Instead we listen for SIGCHLD telling
2223 # us our child process terminated.
2227 # us our child process terminated.
2224 terminated = set()
2228 terminated = set()
2225 def handler(signum, frame):
2229 def handler(signum, frame):
2226 terminated.add(os.wait())
2230 terminated.add(os.wait())
2227 prevhandler = None
2231 prevhandler = None
2228 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2232 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2229 if SIGCHLD is not None:
2233 if SIGCHLD is not None:
2230 prevhandler = signal.signal(SIGCHLD, handler)
2234 prevhandler = signal.signal(SIGCHLD, handler)
2231 try:
2235 try:
2232 pid = spawndetached(args)
2236 pid = spawndetached(args)
2233 while not condfn():
2237 while not condfn():
2234 if ((pid in terminated or not testpid(pid))
2238 if ((pid in terminated or not testpid(pid))
2235 and not condfn()):
2239 and not condfn()):
2236 return -1
2240 return -1
2237 time.sleep(0.1)
2241 time.sleep(0.1)
2238 return pid
2242 return pid
2239 finally:
2243 finally:
2240 if prevhandler is not None:
2244 if prevhandler is not None:
2241 signal.signal(signal.SIGCHLD, prevhandler)
2245 signal.signal(signal.SIGCHLD, prevhandler)
2242
2246
2243 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2247 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2244 """Return the result of interpolating items in the mapping into string s.
2248 """Return the result of interpolating items in the mapping into string s.
2245
2249
2246 prefix is a single character string, or a two character string with
2250 prefix is a single character string, or a two character string with
2247 a backslash as the first character if the prefix needs to be escaped in
2251 a backslash as the first character if the prefix needs to be escaped in
2248 a regular expression.
2252 a regular expression.
2249
2253
2250 fn is an optional function that will be applied to the replacement text
2254 fn is an optional function that will be applied to the replacement text
2251 just before replacement.
2255 just before replacement.
2252
2256
2253 escape_prefix is an optional flag that allows using doubled prefix for
2257 escape_prefix is an optional flag that allows using doubled prefix for
2254 its escaping.
2258 its escaping.
2255 """
2259 """
2256 fn = fn or (lambda s: s)
2260 fn = fn or (lambda s: s)
2257 patterns = '|'.join(mapping.keys())
2261 patterns = '|'.join(mapping.keys())
2258 if escape_prefix:
2262 if escape_prefix:
2259 patterns += '|' + prefix
2263 patterns += '|' + prefix
2260 if len(prefix) > 1:
2264 if len(prefix) > 1:
2261 prefix_char = prefix[1:]
2265 prefix_char = prefix[1:]
2262 else:
2266 else:
2263 prefix_char = prefix
2267 prefix_char = prefix
2264 mapping[prefix_char] = prefix_char
2268 mapping[prefix_char] = prefix_char
2265 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2269 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2266 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2270 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2267
2271
2268 def getport(port):
2272 def getport(port):
2269 """Return the port for a given network service.
2273 """Return the port for a given network service.
2270
2274
2271 If port is an integer, it's returned as is. If it's a string, it's
2275 If port is an integer, it's returned as is. If it's a string, it's
2272 looked up using socket.getservbyname(). If there's no matching
2276 looked up using socket.getservbyname(). If there's no matching
2273 service, error.Abort is raised.
2277 service, error.Abort is raised.
2274 """
2278 """
2275 try:
2279 try:
2276 return int(port)
2280 return int(port)
2277 except ValueError:
2281 except ValueError:
2278 pass
2282 pass
2279
2283
2280 try:
2284 try:
2281 return socket.getservbyname(port)
2285 return socket.getservbyname(port)
2282 except socket.error:
2286 except socket.error:
2283 raise Abort(_("no port number associated with service '%s'") % port)
2287 raise Abort(_("no port number associated with service '%s'") % port)
2284
2288
2285 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2289 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2286 '0': False, 'no': False, 'false': False, 'off': False,
2290 '0': False, 'no': False, 'false': False, 'off': False,
2287 'never': False}
2291 'never': False}
2288
2292
2289 def parsebool(s):
2293 def parsebool(s):
2290 """Parse s into a boolean.
2294 """Parse s into a boolean.
2291
2295
2292 If s is not a valid boolean, returns None.
2296 If s is not a valid boolean, returns None.
2293 """
2297 """
2294 return _booleans.get(s.lower(), None)
2298 return _booleans.get(s.lower(), None)
2295
2299
2296 _hexdig = '0123456789ABCDEFabcdef'
2300 _hexdig = '0123456789ABCDEFabcdef'
2297 _hextochr = dict((a + b, chr(int(a + b, 16)))
2301 _hextochr = dict((a + b, chr(int(a + b, 16)))
2298 for a in _hexdig for b in _hexdig)
2302 for a in _hexdig for b in _hexdig)
2299
2303
2300 def _urlunquote(s):
2304 def _urlunquote(s):
2301 """Decode HTTP/HTML % encoding.
2305 """Decode HTTP/HTML % encoding.
2302
2306
2303 >>> _urlunquote('abc%20def')
2307 >>> _urlunquote('abc%20def')
2304 'abc def'
2308 'abc def'
2305 """
2309 """
2306 res = s.split('%')
2310 res = s.split('%')
2307 # fastpath
2311 # fastpath
2308 if len(res) == 1:
2312 if len(res) == 1:
2309 return s
2313 return s
2310 s = res[0]
2314 s = res[0]
2311 for item in res[1:]:
2315 for item in res[1:]:
2312 try:
2316 try:
2313 s += _hextochr[item[:2]] + item[2:]
2317 s += _hextochr[item[:2]] + item[2:]
2314 except KeyError:
2318 except KeyError:
2315 s += '%' + item
2319 s += '%' + item
2316 except UnicodeDecodeError:
2320 except UnicodeDecodeError:
2317 s += unichr(int(item[:2], 16)) + item[2:]
2321 s += unichr(int(item[:2], 16)) + item[2:]
2318 return s
2322 return s
2319
2323
2320 class url(object):
2324 class url(object):
2321 r"""Reliable URL parser.
2325 r"""Reliable URL parser.
2322
2326
2323 This parses URLs and provides attributes for the following
2327 This parses URLs and provides attributes for the following
2324 components:
2328 components:
2325
2329
2326 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2330 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2327
2331
2328 Missing components are set to None. The only exception is
2332 Missing components are set to None. The only exception is
2329 fragment, which is set to '' if present but empty.
2333 fragment, which is set to '' if present but empty.
2330
2334
2331 If parsefragment is False, fragment is included in query. If
2335 If parsefragment is False, fragment is included in query. If
2332 parsequery is False, query is included in path. If both are
2336 parsequery is False, query is included in path. If both are
2333 False, both fragment and query are included in path.
2337 False, both fragment and query are included in path.
2334
2338
2335 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2339 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2336
2340
2337 Note that for backward compatibility reasons, bundle URLs do not
2341 Note that for backward compatibility reasons, bundle URLs do not
2338 take host names. That means 'bundle://../' has a path of '../'.
2342 take host names. That means 'bundle://../' has a path of '../'.
2339
2343
2340 Examples:
2344 Examples:
2341
2345
2342 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2346 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2343 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2347 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2344 >>> url('ssh://[::1]:2200//home/joe/repo')
2348 >>> url('ssh://[::1]:2200//home/joe/repo')
2345 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2349 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2346 >>> url('file:///home/joe/repo')
2350 >>> url('file:///home/joe/repo')
2347 <url scheme: 'file', path: '/home/joe/repo'>
2351 <url scheme: 'file', path: '/home/joe/repo'>
2348 >>> url('file:///c:/temp/foo/')
2352 >>> url('file:///c:/temp/foo/')
2349 <url scheme: 'file', path: 'c:/temp/foo/'>
2353 <url scheme: 'file', path: 'c:/temp/foo/'>
2350 >>> url('bundle:foo')
2354 >>> url('bundle:foo')
2351 <url scheme: 'bundle', path: 'foo'>
2355 <url scheme: 'bundle', path: 'foo'>
2352 >>> url('bundle://../foo')
2356 >>> url('bundle://../foo')
2353 <url scheme: 'bundle', path: '../foo'>
2357 <url scheme: 'bundle', path: '../foo'>
2354 >>> url(r'c:\foo\bar')
2358 >>> url(r'c:\foo\bar')
2355 <url path: 'c:\\foo\\bar'>
2359 <url path: 'c:\\foo\\bar'>
2356 >>> url(r'\\blah\blah\blah')
2360 >>> url(r'\\blah\blah\blah')
2357 <url path: '\\\\blah\\blah\\blah'>
2361 <url path: '\\\\blah\\blah\\blah'>
2358 >>> url(r'\\blah\blah\blah#baz')
2362 >>> url(r'\\blah\blah\blah#baz')
2359 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2363 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2360 >>> url(r'file:///C:\users\me')
2364 >>> url(r'file:///C:\users\me')
2361 <url scheme: 'file', path: 'C:\\users\\me'>
2365 <url scheme: 'file', path: 'C:\\users\\me'>
2362
2366
2363 Authentication credentials:
2367 Authentication credentials:
2364
2368
2365 >>> url('ssh://joe:xyz@x/repo')
2369 >>> url('ssh://joe:xyz@x/repo')
2366 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2370 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2367 >>> url('ssh://joe@x/repo')
2371 >>> url('ssh://joe@x/repo')
2368 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2372 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2369
2373
2370 Query strings and fragments:
2374 Query strings and fragments:
2371
2375
2372 >>> url('http://host/a?b#c')
2376 >>> url('http://host/a?b#c')
2373 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2377 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2374 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2378 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2375 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2379 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2376 """
2380 """
2377
2381
2378 _safechars = "!~*'()+"
2382 _safechars = "!~*'()+"
2379 _safepchars = "/!~*'()+:\\"
2383 _safepchars = "/!~*'()+:\\"
2380 _matchscheme = remod.compile(r'^[a-zA-Z0-9+.\-]+:').match
2384 _matchscheme = remod.compile(r'^[a-zA-Z0-9+.\-]+:').match
2381
2385
2382 def __init__(self, path, parsequery=True, parsefragment=True):
2386 def __init__(self, path, parsequery=True, parsefragment=True):
2383 # We slowly chomp away at path until we have only the path left
2387 # We slowly chomp away at path until we have only the path left
2384 self.scheme = self.user = self.passwd = self.host = None
2388 self.scheme = self.user = self.passwd = self.host = None
2385 self.port = self.path = self.query = self.fragment = None
2389 self.port = self.path = self.query = self.fragment = None
2386 self._localpath = True
2390 self._localpath = True
2387 self._hostport = ''
2391 self._hostport = ''
2388 self._origpath = path
2392 self._origpath = path
2389
2393
2390 if parsefragment and '#' in path:
2394 if parsefragment and '#' in path:
2391 path, self.fragment = path.split('#', 1)
2395 path, self.fragment = path.split('#', 1)
2392 if not path:
2396 if not path:
2393 path = None
2397 path = None
2394
2398
2395 # special case for Windows drive letters and UNC paths
2399 # special case for Windows drive letters and UNC paths
2396 if hasdriveletter(path) or path.startswith(r'\\'):
2400 if hasdriveletter(path) or path.startswith(r'\\'):
2397 self.path = path
2401 self.path = path
2398 return
2402 return
2399
2403
2400 # For compatibility reasons, we can't handle bundle paths as
2404 # For compatibility reasons, we can't handle bundle paths as
2401 # normal URLS
2405 # normal URLS
2402 if path.startswith('bundle:'):
2406 if path.startswith('bundle:'):
2403 self.scheme = 'bundle'
2407 self.scheme = 'bundle'
2404 path = path[7:]
2408 path = path[7:]
2405 if path.startswith('//'):
2409 if path.startswith('//'):
2406 path = path[2:]
2410 path = path[2:]
2407 self.path = path
2411 self.path = path
2408 return
2412 return
2409
2413
2410 if self._matchscheme(path):
2414 if self._matchscheme(path):
2411 parts = path.split(':', 1)
2415 parts = path.split(':', 1)
2412 if parts[0]:
2416 if parts[0]:
2413 self.scheme, path = parts
2417 self.scheme, path = parts
2414 self._localpath = False
2418 self._localpath = False
2415
2419
2416 if not path:
2420 if not path:
2417 path = None
2421 path = None
2418 if self._localpath:
2422 if self._localpath:
2419 self.path = ''
2423 self.path = ''
2420 return
2424 return
2421 else:
2425 else:
2422 if self._localpath:
2426 if self._localpath:
2423 self.path = path
2427 self.path = path
2424 return
2428 return
2425
2429
2426 if parsequery and '?' in path:
2430 if parsequery and '?' in path:
2427 path, self.query = path.split('?', 1)
2431 path, self.query = path.split('?', 1)
2428 if not path:
2432 if not path:
2429 path = None
2433 path = None
2430 if not self.query:
2434 if not self.query:
2431 self.query = None
2435 self.query = None
2432
2436
2433 # // is required to specify a host/authority
2437 # // is required to specify a host/authority
2434 if path and path.startswith('//'):
2438 if path and path.startswith('//'):
2435 parts = path[2:].split('/', 1)
2439 parts = path[2:].split('/', 1)
2436 if len(parts) > 1:
2440 if len(parts) > 1:
2437 self.host, path = parts
2441 self.host, path = parts
2438 else:
2442 else:
2439 self.host = parts[0]
2443 self.host = parts[0]
2440 path = None
2444 path = None
2441 if not self.host:
2445 if not self.host:
2442 self.host = None
2446 self.host = None
2443 # path of file:///d is /d
2447 # path of file:///d is /d
2444 # path of file:///d:/ is d:/, not /d:/
2448 # path of file:///d:/ is d:/, not /d:/
2445 if path and not hasdriveletter(path):
2449 if path and not hasdriveletter(path):
2446 path = '/' + path
2450 path = '/' + path
2447
2451
2448 if self.host and '@' in self.host:
2452 if self.host and '@' in self.host:
2449 self.user, self.host = self.host.rsplit('@', 1)
2453 self.user, self.host = self.host.rsplit('@', 1)
2450 if ':' in self.user:
2454 if ':' in self.user:
2451 self.user, self.passwd = self.user.split(':', 1)
2455 self.user, self.passwd = self.user.split(':', 1)
2452 if not self.host:
2456 if not self.host:
2453 self.host = None
2457 self.host = None
2454
2458
2455 # Don't split on colons in IPv6 addresses without ports
2459 # Don't split on colons in IPv6 addresses without ports
2456 if (self.host and ':' in self.host and
2460 if (self.host and ':' in self.host and
2457 not (self.host.startswith('[') and self.host.endswith(']'))):
2461 not (self.host.startswith('[') and self.host.endswith(']'))):
2458 self._hostport = self.host
2462 self._hostport = self.host
2459 self.host, self.port = self.host.rsplit(':', 1)
2463 self.host, self.port = self.host.rsplit(':', 1)
2460 if not self.host:
2464 if not self.host:
2461 self.host = None
2465 self.host = None
2462
2466
2463 if (self.host and self.scheme == 'file' and
2467 if (self.host and self.scheme == 'file' and
2464 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2468 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2465 raise Abort(_('file:// URLs can only refer to localhost'))
2469 raise Abort(_('file:// URLs can only refer to localhost'))
2466
2470
2467 self.path = path
2471 self.path = path
2468
2472
2469 # leave the query string escaped
2473 # leave the query string escaped
2470 for a in ('user', 'passwd', 'host', 'port',
2474 for a in ('user', 'passwd', 'host', 'port',
2471 'path', 'fragment'):
2475 'path', 'fragment'):
2472 v = getattr(self, a)
2476 v = getattr(self, a)
2473 if v is not None:
2477 if v is not None:
2474 setattr(self, a, _urlunquote(v))
2478 setattr(self, a, _urlunquote(v))
2475
2479
2476 def __repr__(self):
2480 def __repr__(self):
2477 attrs = []
2481 attrs = []
2478 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2482 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2479 'query', 'fragment'):
2483 'query', 'fragment'):
2480 v = getattr(self, a)
2484 v = getattr(self, a)
2481 if v is not None:
2485 if v is not None:
2482 attrs.append('%s: %r' % (a, v))
2486 attrs.append('%s: %r' % (a, v))
2483 return '<url %s>' % ', '.join(attrs)
2487 return '<url %s>' % ', '.join(attrs)
2484
2488
2485 def __str__(self):
2489 def __str__(self):
2486 r"""Join the URL's components back into a URL string.
2490 r"""Join the URL's components back into a URL string.
2487
2491
2488 Examples:
2492 Examples:
2489
2493
2490 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2494 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2491 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2495 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2492 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2496 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2493 'http://user:pw@host:80/?foo=bar&baz=42'
2497 'http://user:pw@host:80/?foo=bar&baz=42'
2494 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2498 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2495 'http://user:pw@host:80/?foo=bar%3dbaz'
2499 'http://user:pw@host:80/?foo=bar%3dbaz'
2496 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2500 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2497 'ssh://user:pw@[::1]:2200//home/joe#'
2501 'ssh://user:pw@[::1]:2200//home/joe#'
2498 >>> str(url('http://localhost:80//'))
2502 >>> str(url('http://localhost:80//'))
2499 'http://localhost:80//'
2503 'http://localhost:80//'
2500 >>> str(url('http://localhost:80/'))
2504 >>> str(url('http://localhost:80/'))
2501 'http://localhost:80/'
2505 'http://localhost:80/'
2502 >>> str(url('http://localhost:80'))
2506 >>> str(url('http://localhost:80'))
2503 'http://localhost:80/'
2507 'http://localhost:80/'
2504 >>> str(url('bundle:foo'))
2508 >>> str(url('bundle:foo'))
2505 'bundle:foo'
2509 'bundle:foo'
2506 >>> str(url('bundle://../foo'))
2510 >>> str(url('bundle://../foo'))
2507 'bundle:../foo'
2511 'bundle:../foo'
2508 >>> str(url('path'))
2512 >>> str(url('path'))
2509 'path'
2513 'path'
2510 >>> str(url('file:///tmp/foo/bar'))
2514 >>> str(url('file:///tmp/foo/bar'))
2511 'file:///tmp/foo/bar'
2515 'file:///tmp/foo/bar'
2512 >>> str(url('file:///c:/tmp/foo/bar'))
2516 >>> str(url('file:///c:/tmp/foo/bar'))
2513 'file:///c:/tmp/foo/bar'
2517 'file:///c:/tmp/foo/bar'
2514 >>> print url(r'bundle:foo\bar')
2518 >>> print url(r'bundle:foo\bar')
2515 bundle:foo\bar
2519 bundle:foo\bar
2516 >>> print url(r'file:///D:\data\hg')
2520 >>> print url(r'file:///D:\data\hg')
2517 file:///D:\data\hg
2521 file:///D:\data\hg
2518 """
2522 """
2519 if self._localpath:
2523 if self._localpath:
2520 s = self.path
2524 s = self.path
2521 if self.scheme == 'bundle':
2525 if self.scheme == 'bundle':
2522 s = 'bundle:' + s
2526 s = 'bundle:' + s
2523 if self.fragment:
2527 if self.fragment:
2524 s += '#' + self.fragment
2528 s += '#' + self.fragment
2525 return s
2529 return s
2526
2530
2527 s = self.scheme + ':'
2531 s = self.scheme + ':'
2528 if self.user or self.passwd or self.host:
2532 if self.user or self.passwd or self.host:
2529 s += '//'
2533 s += '//'
2530 elif self.scheme and (not self.path or self.path.startswith('/')
2534 elif self.scheme and (not self.path or self.path.startswith('/')
2531 or hasdriveletter(self.path)):
2535 or hasdriveletter(self.path)):
2532 s += '//'
2536 s += '//'
2533 if hasdriveletter(self.path):
2537 if hasdriveletter(self.path):
2534 s += '/'
2538 s += '/'
2535 if self.user:
2539 if self.user:
2536 s += urlreq.quote(self.user, safe=self._safechars)
2540 s += urlreq.quote(self.user, safe=self._safechars)
2537 if self.passwd:
2541 if self.passwd:
2538 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2542 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2539 if self.user or self.passwd:
2543 if self.user or self.passwd:
2540 s += '@'
2544 s += '@'
2541 if self.host:
2545 if self.host:
2542 if not (self.host.startswith('[') and self.host.endswith(']')):
2546 if not (self.host.startswith('[') and self.host.endswith(']')):
2543 s += urlreq.quote(self.host)
2547 s += urlreq.quote(self.host)
2544 else:
2548 else:
2545 s += self.host
2549 s += self.host
2546 if self.port:
2550 if self.port:
2547 s += ':' + urlreq.quote(self.port)
2551 s += ':' + urlreq.quote(self.port)
2548 if self.host:
2552 if self.host:
2549 s += '/'
2553 s += '/'
2550 if self.path:
2554 if self.path:
2551 # TODO: similar to the query string, we should not unescape the
2555 # TODO: similar to the query string, we should not unescape the
2552 # path when we store it, the path might contain '%2f' = '/',
2556 # path when we store it, the path might contain '%2f' = '/',
2553 # which we should *not* escape.
2557 # which we should *not* escape.
2554 s += urlreq.quote(self.path, safe=self._safepchars)
2558 s += urlreq.quote(self.path, safe=self._safepchars)
2555 if self.query:
2559 if self.query:
2556 # we store the query in escaped form.
2560 # we store the query in escaped form.
2557 s += '?' + self.query
2561 s += '?' + self.query
2558 if self.fragment is not None:
2562 if self.fragment is not None:
2559 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2563 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2560 return s
2564 return s
2561
2565
2562 def authinfo(self):
2566 def authinfo(self):
2563 user, passwd = self.user, self.passwd
2567 user, passwd = self.user, self.passwd
2564 try:
2568 try:
2565 self.user, self.passwd = None, None
2569 self.user, self.passwd = None, None
2566 s = str(self)
2570 s = str(self)
2567 finally:
2571 finally:
2568 self.user, self.passwd = user, passwd
2572 self.user, self.passwd = user, passwd
2569 if not self.user:
2573 if not self.user:
2570 return (s, None)
2574 return (s, None)
2571 # authinfo[1] is passed to urllib2 password manager, and its
2575 # authinfo[1] is passed to urllib2 password manager, and its
2572 # URIs must not contain credentials. The host is passed in the
2576 # URIs must not contain credentials. The host is passed in the
2573 # URIs list because Python < 2.4.3 uses only that to search for
2577 # URIs list because Python < 2.4.3 uses only that to search for
2574 # a password.
2578 # a password.
2575 return (s, (None, (s, self.host),
2579 return (s, (None, (s, self.host),
2576 self.user, self.passwd or ''))
2580 self.user, self.passwd or ''))
2577
2581
2578 def isabs(self):
2582 def isabs(self):
2579 if self.scheme and self.scheme != 'file':
2583 if self.scheme and self.scheme != 'file':
2580 return True # remote URL
2584 return True # remote URL
2581 if hasdriveletter(self.path):
2585 if hasdriveletter(self.path):
2582 return True # absolute for our purposes - can't be joined()
2586 return True # absolute for our purposes - can't be joined()
2583 if self.path.startswith(r'\\'):
2587 if self.path.startswith(r'\\'):
2584 return True # Windows UNC path
2588 return True # Windows UNC path
2585 if self.path.startswith('/'):
2589 if self.path.startswith('/'):
2586 return True # POSIX-style
2590 return True # POSIX-style
2587 return False
2591 return False
2588
2592
2589 def localpath(self):
2593 def localpath(self):
2590 if self.scheme == 'file' or self.scheme == 'bundle':
2594 if self.scheme == 'file' or self.scheme == 'bundle':
2591 path = self.path or '/'
2595 path = self.path or '/'
2592 # For Windows, we need to promote hosts containing drive
2596 # For Windows, we need to promote hosts containing drive
2593 # letters to paths with drive letters.
2597 # letters to paths with drive letters.
2594 if hasdriveletter(self._hostport):
2598 if hasdriveletter(self._hostport):
2595 path = self._hostport + '/' + self.path
2599 path = self._hostport + '/' + self.path
2596 elif (self.host is not None and self.path
2600 elif (self.host is not None and self.path
2597 and not hasdriveletter(path)):
2601 and not hasdriveletter(path)):
2598 path = '/' + path
2602 path = '/' + path
2599 return path
2603 return path
2600 return self._origpath
2604 return self._origpath
2601
2605
2602 def islocal(self):
2606 def islocal(self):
2603 '''whether localpath will return something that posixfile can open'''
2607 '''whether localpath will return something that posixfile can open'''
2604 return (not self.scheme or self.scheme == 'file'
2608 return (not self.scheme or self.scheme == 'file'
2605 or self.scheme == 'bundle')
2609 or self.scheme == 'bundle')
2606
2610
2607 def hasscheme(path):
2611 def hasscheme(path):
2608 return bool(url(path).scheme)
2612 return bool(url(path).scheme)
2609
2613
2610 def hasdriveletter(path):
2614 def hasdriveletter(path):
2611 return path and path[1:2] == ':' and path[0:1].isalpha()
2615 return path and path[1:2] == ':' and path[0:1].isalpha()
2612
2616
2613 def urllocalpath(path):
2617 def urllocalpath(path):
2614 return url(path, parsequery=False, parsefragment=False).localpath()
2618 return url(path, parsequery=False, parsefragment=False).localpath()
2615
2619
2616 def hidepassword(u):
2620 def hidepassword(u):
2617 '''hide user credential in a url string'''
2621 '''hide user credential in a url string'''
2618 u = url(u)
2622 u = url(u)
2619 if u.passwd:
2623 if u.passwd:
2620 u.passwd = '***'
2624 u.passwd = '***'
2621 return str(u)
2625 return str(u)
2622
2626
2623 def removeauth(u):
2627 def removeauth(u):
2624 '''remove all authentication information from a url string'''
2628 '''remove all authentication information from a url string'''
2625 u = url(u)
2629 u = url(u)
2626 u.user = u.passwd = None
2630 u.user = u.passwd = None
2627 return str(u)
2631 return str(u)
2628
2632
2629 def isatty(fp):
2633 def isatty(fp):
2630 try:
2634 try:
2631 return fp.isatty()
2635 return fp.isatty()
2632 except AttributeError:
2636 except AttributeError:
2633 return False
2637 return False
2634
2638
2635 timecount = unitcountfn(
2639 timecount = unitcountfn(
2636 (1, 1e3, _('%.0f s')),
2640 (1, 1e3, _('%.0f s')),
2637 (100, 1, _('%.1f s')),
2641 (100, 1, _('%.1f s')),
2638 (10, 1, _('%.2f s')),
2642 (10, 1, _('%.2f s')),
2639 (1, 1, _('%.3f s')),
2643 (1, 1, _('%.3f s')),
2640 (100, 0.001, _('%.1f ms')),
2644 (100, 0.001, _('%.1f ms')),
2641 (10, 0.001, _('%.2f ms')),
2645 (10, 0.001, _('%.2f ms')),
2642 (1, 0.001, _('%.3f ms')),
2646 (1, 0.001, _('%.3f ms')),
2643 (100, 0.000001, _('%.1f us')),
2647 (100, 0.000001, _('%.1f us')),
2644 (10, 0.000001, _('%.2f us')),
2648 (10, 0.000001, _('%.2f us')),
2645 (1, 0.000001, _('%.3f us')),
2649 (1, 0.000001, _('%.3f us')),
2646 (100, 0.000000001, _('%.1f ns')),
2650 (100, 0.000000001, _('%.1f ns')),
2647 (10, 0.000000001, _('%.2f ns')),
2651 (10, 0.000000001, _('%.2f ns')),
2648 (1, 0.000000001, _('%.3f ns')),
2652 (1, 0.000000001, _('%.3f ns')),
2649 )
2653 )
2650
2654
2651 _timenesting = [0]
2655 _timenesting = [0]
2652
2656
2653 def timed(func):
2657 def timed(func):
2654 '''Report the execution time of a function call to stderr.
2658 '''Report the execution time of a function call to stderr.
2655
2659
2656 During development, use as a decorator when you need to measure
2660 During development, use as a decorator when you need to measure
2657 the cost of a function, e.g. as follows:
2661 the cost of a function, e.g. as follows:
2658
2662
2659 @util.timed
2663 @util.timed
2660 def foo(a, b, c):
2664 def foo(a, b, c):
2661 pass
2665 pass
2662 '''
2666 '''
2663
2667
2664 def wrapper(*args, **kwargs):
2668 def wrapper(*args, **kwargs):
2665 start = time.time()
2669 start = time.time()
2666 indent = 2
2670 indent = 2
2667 _timenesting[0] += indent
2671 _timenesting[0] += indent
2668 try:
2672 try:
2669 return func(*args, **kwargs)
2673 return func(*args, **kwargs)
2670 finally:
2674 finally:
2671 elapsed = time.time() - start
2675 elapsed = time.time() - start
2672 _timenesting[0] -= indent
2676 _timenesting[0] -= indent
2673 sys.stderr.write('%s%s: %s\n' %
2677 sys.stderr.write('%s%s: %s\n' %
2674 (' ' * _timenesting[0], func.__name__,
2678 (' ' * _timenesting[0], func.__name__,
2675 timecount(elapsed)))
2679 timecount(elapsed)))
2676 return wrapper
2680 return wrapper
2677
2681
2678 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2682 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2679 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2683 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2680
2684
2681 def sizetoint(s):
2685 def sizetoint(s):
2682 '''Convert a space specifier to a byte count.
2686 '''Convert a space specifier to a byte count.
2683
2687
2684 >>> sizetoint('30')
2688 >>> sizetoint('30')
2685 30
2689 30
2686 >>> sizetoint('2.2kb')
2690 >>> sizetoint('2.2kb')
2687 2252
2691 2252
2688 >>> sizetoint('6M')
2692 >>> sizetoint('6M')
2689 6291456
2693 6291456
2690 '''
2694 '''
2691 t = s.strip().lower()
2695 t = s.strip().lower()
2692 try:
2696 try:
2693 for k, u in _sizeunits:
2697 for k, u in _sizeunits:
2694 if t.endswith(k):
2698 if t.endswith(k):
2695 return int(float(t[:-len(k)]) * u)
2699 return int(float(t[:-len(k)]) * u)
2696 return int(t)
2700 return int(t)
2697 except ValueError:
2701 except ValueError:
2698 raise error.ParseError(_("couldn't parse size: %s") % s)
2702 raise error.ParseError(_("couldn't parse size: %s") % s)
2699
2703
2700 class hooks(object):
2704 class hooks(object):
2701 '''A collection of hook functions that can be used to extend a
2705 '''A collection of hook functions that can be used to extend a
2702 function's behavior. Hooks are called in lexicographic order,
2706 function's behavior. Hooks are called in lexicographic order,
2703 based on the names of their sources.'''
2707 based on the names of their sources.'''
2704
2708
2705 def __init__(self):
2709 def __init__(self):
2706 self._hooks = []
2710 self._hooks = []
2707
2711
2708 def add(self, source, hook):
2712 def add(self, source, hook):
2709 self._hooks.append((source, hook))
2713 self._hooks.append((source, hook))
2710
2714
2711 def __call__(self, *args):
2715 def __call__(self, *args):
2712 self._hooks.sort(key=lambda x: x[0])
2716 self._hooks.sort(key=lambda x: x[0])
2713 results = []
2717 results = []
2714 for source, hook in self._hooks:
2718 for source, hook in self._hooks:
2715 results.append(hook(*args))
2719 results.append(hook(*args))
2716 return results
2720 return results
2717
2721
2718 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s'):
2722 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s'):
2719 '''Yields lines for a nicely formatted stacktrace.
2723 '''Yields lines for a nicely formatted stacktrace.
2720 Skips the 'skip' last entries.
2724 Skips the 'skip' last entries.
2721 Each file+linenumber is formatted according to fileline.
2725 Each file+linenumber is formatted according to fileline.
2722 Each line is formatted according to line.
2726 Each line is formatted according to line.
2723 If line is None, it yields:
2727 If line is None, it yields:
2724 length of longest filepath+line number,
2728 length of longest filepath+line number,
2725 filepath+linenumber,
2729 filepath+linenumber,
2726 function
2730 function
2727
2731
2728 Not be used in production code but very convenient while developing.
2732 Not be used in production code but very convenient while developing.
2729 '''
2733 '''
2730 entries = [(fileline % (fn, ln), func)
2734 entries = [(fileline % (fn, ln), func)
2731 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]]
2735 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]]
2732 if entries:
2736 if entries:
2733 fnmax = max(len(entry[0]) for entry in entries)
2737 fnmax = max(len(entry[0]) for entry in entries)
2734 for fnln, func in entries:
2738 for fnln, func in entries:
2735 if line is None:
2739 if line is None:
2736 yield (fnmax, fnln, func)
2740 yield (fnmax, fnln, func)
2737 else:
2741 else:
2738 yield line % (fnmax, fnln, func)
2742 yield line % (fnmax, fnln, func)
2739
2743
2740 def debugstacktrace(msg='stacktrace', skip=0, f=sys.stderr, otherf=sys.stdout):
2744 def debugstacktrace(msg='stacktrace', skip=0, f=sys.stderr, otherf=sys.stdout):
2741 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2745 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2742 Skips the 'skip' last entries. By default it will flush stdout first.
2746 Skips the 'skip' last entries. By default it will flush stdout first.
2743 It can be used everywhere and intentionally does not require an ui object.
2747 It can be used everywhere and intentionally does not require an ui object.
2744 Not be used in production code but very convenient while developing.
2748 Not be used in production code but very convenient while developing.
2745 '''
2749 '''
2746 if otherf:
2750 if otherf:
2747 otherf.flush()
2751 otherf.flush()
2748 f.write('%s at:\n' % msg)
2752 f.write('%s at:\n' % msg)
2749 for line in getstackframes(skip + 1):
2753 for line in getstackframes(skip + 1):
2750 f.write(line)
2754 f.write(line)
2751 f.flush()
2755 f.flush()
2752
2756
2753 class dirs(object):
2757 class dirs(object):
2754 '''a multiset of directory names from a dirstate or manifest'''
2758 '''a multiset of directory names from a dirstate or manifest'''
2755
2759
2756 def __init__(self, map, skip=None):
2760 def __init__(self, map, skip=None):
2757 self._dirs = {}
2761 self._dirs = {}
2758 addpath = self.addpath
2762 addpath = self.addpath
2759 if safehasattr(map, 'iteritems') and skip is not None:
2763 if safehasattr(map, 'iteritems') and skip is not None:
2760 for f, s in map.iteritems():
2764 for f, s in map.iteritems():
2761 if s[0] != skip:
2765 if s[0] != skip:
2762 addpath(f)
2766 addpath(f)
2763 else:
2767 else:
2764 for f in map:
2768 for f in map:
2765 addpath(f)
2769 addpath(f)
2766
2770
2767 def addpath(self, path):
2771 def addpath(self, path):
2768 dirs = self._dirs
2772 dirs = self._dirs
2769 for base in finddirs(path):
2773 for base in finddirs(path):
2770 if base in dirs:
2774 if base in dirs:
2771 dirs[base] += 1
2775 dirs[base] += 1
2772 return
2776 return
2773 dirs[base] = 1
2777 dirs[base] = 1
2774
2778
2775 def delpath(self, path):
2779 def delpath(self, path):
2776 dirs = self._dirs
2780 dirs = self._dirs
2777 for base in finddirs(path):
2781 for base in finddirs(path):
2778 if dirs[base] > 1:
2782 if dirs[base] > 1:
2779 dirs[base] -= 1
2783 dirs[base] -= 1
2780 return
2784 return
2781 del dirs[base]
2785 del dirs[base]
2782
2786
2783 def __iter__(self):
2787 def __iter__(self):
2784 return self._dirs.iterkeys()
2788 return self._dirs.iterkeys()
2785
2789
2786 def __contains__(self, d):
2790 def __contains__(self, d):
2787 return d in self._dirs
2791 return d in self._dirs
2788
2792
2789 if safehasattr(parsers, 'dirs'):
2793 if safehasattr(parsers, 'dirs'):
2790 dirs = parsers.dirs
2794 dirs = parsers.dirs
2791
2795
2792 def finddirs(path):
2796 def finddirs(path):
2793 pos = path.rfind('/')
2797 pos = path.rfind('/')
2794 while pos != -1:
2798 while pos != -1:
2795 yield path[:pos]
2799 yield path[:pos]
2796 pos = path.rfind('/', 0, pos)
2800 pos = path.rfind('/', 0, pos)
2797
2801
2798 # compression utility
2802 # compression utility
2799
2803
2800 class nocompress(object):
2804 class nocompress(object):
2801 def compress(self, x):
2805 def compress(self, x):
2802 return x
2806 return x
2803 def flush(self):
2807 def flush(self):
2804 return ""
2808 return ""
2805
2809
2806 compressors = {
2810 compressors = {
2807 None: nocompress,
2811 None: nocompress,
2808 # lambda to prevent early import
2812 # lambda to prevent early import
2809 'BZ': lambda: bz2.BZ2Compressor(),
2813 'BZ': lambda: bz2.BZ2Compressor(),
2810 'GZ': lambda: zlib.compressobj(),
2814 'GZ': lambda: zlib.compressobj(),
2811 }
2815 }
2812 # also support the old form by courtesies
2816 # also support the old form by courtesies
2813 compressors['UN'] = compressors[None]
2817 compressors['UN'] = compressors[None]
2814
2818
2815 def _makedecompressor(decompcls):
2819 def _makedecompressor(decompcls):
2816 def generator(f):
2820 def generator(f):
2817 d = decompcls()
2821 d = decompcls()
2818 for chunk in filechunkiter(f):
2822 for chunk in filechunkiter(f):
2819 yield d.decompress(chunk)
2823 yield d.decompress(chunk)
2820 def func(fh):
2824 def func(fh):
2821 return chunkbuffer(generator(fh))
2825 return chunkbuffer(generator(fh))
2822 return func
2826 return func
2823
2827
2824 class ctxmanager(object):
2828 class ctxmanager(object):
2825 '''A context manager for use in 'with' blocks to allow multiple
2829 '''A context manager for use in 'with' blocks to allow multiple
2826 contexts to be entered at once. This is both safer and more
2830 contexts to be entered at once. This is both safer and more
2827 flexible than contextlib.nested.
2831 flexible than contextlib.nested.
2828
2832
2829 Once Mercurial supports Python 2.7+, this will become mostly
2833 Once Mercurial supports Python 2.7+, this will become mostly
2830 unnecessary.
2834 unnecessary.
2831 '''
2835 '''
2832
2836
2833 def __init__(self, *args):
2837 def __init__(self, *args):
2834 '''Accepts a list of no-argument functions that return context
2838 '''Accepts a list of no-argument functions that return context
2835 managers. These will be invoked at __call__ time.'''
2839 managers. These will be invoked at __call__ time.'''
2836 self._pending = args
2840 self._pending = args
2837 self._atexit = []
2841 self._atexit = []
2838
2842
2839 def __enter__(self):
2843 def __enter__(self):
2840 return self
2844 return self
2841
2845
2842 def enter(self):
2846 def enter(self):
2843 '''Create and enter context managers in the order in which they were
2847 '''Create and enter context managers in the order in which they were
2844 passed to the constructor.'''
2848 passed to the constructor.'''
2845 values = []
2849 values = []
2846 for func in self._pending:
2850 for func in self._pending:
2847 obj = func()
2851 obj = func()
2848 values.append(obj.__enter__())
2852 values.append(obj.__enter__())
2849 self._atexit.append(obj.__exit__)
2853 self._atexit.append(obj.__exit__)
2850 del self._pending
2854 del self._pending
2851 return values
2855 return values
2852
2856
2853 def atexit(self, func, *args, **kwargs):
2857 def atexit(self, func, *args, **kwargs):
2854 '''Add a function to call when this context manager exits. The
2858 '''Add a function to call when this context manager exits. The
2855 ordering of multiple atexit calls is unspecified, save that
2859 ordering of multiple atexit calls is unspecified, save that
2856 they will happen before any __exit__ functions.'''
2860 they will happen before any __exit__ functions.'''
2857 def wrapper(exc_type, exc_val, exc_tb):
2861 def wrapper(exc_type, exc_val, exc_tb):
2858 func(*args, **kwargs)
2862 func(*args, **kwargs)
2859 self._atexit.append(wrapper)
2863 self._atexit.append(wrapper)
2860 return func
2864 return func
2861
2865
2862 def __exit__(self, exc_type, exc_val, exc_tb):
2866 def __exit__(self, exc_type, exc_val, exc_tb):
2863 '''Context managers are exited in the reverse order from which
2867 '''Context managers are exited in the reverse order from which
2864 they were created.'''
2868 they were created.'''
2865 received = exc_type is not None
2869 received = exc_type is not None
2866 suppressed = False
2870 suppressed = False
2867 pending = None
2871 pending = None
2868 self._atexit.reverse()
2872 self._atexit.reverse()
2869 for exitfunc in self._atexit:
2873 for exitfunc in self._atexit:
2870 try:
2874 try:
2871 if exitfunc(exc_type, exc_val, exc_tb):
2875 if exitfunc(exc_type, exc_val, exc_tb):
2872 suppressed = True
2876 suppressed = True
2873 exc_type = None
2877 exc_type = None
2874 exc_val = None
2878 exc_val = None
2875 exc_tb = None
2879 exc_tb = None
2876 except BaseException:
2880 except BaseException:
2877 pending = sys.exc_info()
2881 pending = sys.exc_info()
2878 exc_type, exc_val, exc_tb = pending = sys.exc_info()
2882 exc_type, exc_val, exc_tb = pending = sys.exc_info()
2879 del self._atexit
2883 del self._atexit
2880 if pending:
2884 if pending:
2881 raise exc_val
2885 raise exc_val
2882 return received and suppressed
2886 return received and suppressed
2883
2887
2884 def _bz2():
2888 def _bz2():
2885 d = bz2.BZ2Decompressor()
2889 d = bz2.BZ2Decompressor()
2886 # Bzip2 stream start with BZ, but we stripped it.
2890 # Bzip2 stream start with BZ, but we stripped it.
2887 # we put it back for good measure.
2891 # we put it back for good measure.
2888 d.decompress('BZ')
2892 d.decompress('BZ')
2889 return d
2893 return d
2890
2894
2891 decompressors = {None: lambda fh: fh,
2895 decompressors = {None: lambda fh: fh,
2892 '_truncatedBZ': _makedecompressor(_bz2),
2896 '_truncatedBZ': _makedecompressor(_bz2),
2893 'BZ': _makedecompressor(lambda: bz2.BZ2Decompressor()),
2897 'BZ': _makedecompressor(lambda: bz2.BZ2Decompressor()),
2894 'GZ': _makedecompressor(lambda: zlib.decompressobj()),
2898 'GZ': _makedecompressor(lambda: zlib.decompressobj()),
2895 }
2899 }
2896 # also support the old form by courtesies
2900 # also support the old form by courtesies
2897 decompressors['UN'] = decompressors[None]
2901 decompressors['UN'] = decompressors[None]
2898
2902
2899 # convenient shortcut
2903 # convenient shortcut
2900 dst = debugstacktrace
2904 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now