##// END OF EJS Templates
py3: replace str() with bytes()
Pulkit Goyal -
r31841:9ff5a124 default
parent child Browse files
Show More
@@ -1,1052 +1,1052
1 1 # hg.py - repository classes for mercurial
2 2 #
3 3 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
4 4 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
5 5 #
6 6 # This software may be used and distributed according to the terms of the
7 7 # GNU General Public License version 2 or any later version.
8 8
9 9 from __future__ import absolute_import
10 10
11 11 import errno
12 12 import hashlib
13 13 import os
14 14 import shutil
15 15
16 16 from .i18n import _
17 17 from .node import nullid
18 18
19 19 from . import (
20 20 bookmarks,
21 21 bundlerepo,
22 22 cmdutil,
23 23 destutil,
24 24 discovery,
25 25 error,
26 26 exchange,
27 27 extensions,
28 28 httppeer,
29 29 localrepo,
30 30 lock,
31 31 merge as mergemod,
32 32 node,
33 33 phases,
34 34 repoview,
35 35 scmutil,
36 36 sshpeer,
37 37 statichttprepo,
38 38 ui as uimod,
39 39 unionrepo,
40 40 url,
41 41 util,
42 42 verify as verifymod,
43 43 vfs as vfsmod,
44 44 )
45 45
46 46 release = lock.release
47 47
48 48 # shared features
49 49 sharedbookmarks = 'bookmarks'
50 50
51 51 def _local(path):
52 52 path = util.expandpath(util.urllocalpath(path))
53 53 return (os.path.isfile(path) and bundlerepo or localrepo)
54 54
55 55 def addbranchrevs(lrepo, other, branches, revs):
56 56 peer = other.peer() # a courtesy to callers using a localrepo for other
57 57 hashbranch, branches = branches
58 58 if not hashbranch and not branches:
59 59 x = revs or None
60 60 if util.safehasattr(revs, 'first'):
61 61 y = revs.first()
62 62 elif revs:
63 63 y = revs[0]
64 64 else:
65 65 y = None
66 66 return x, y
67 67 if revs:
68 68 revs = list(revs)
69 69 else:
70 70 revs = []
71 71
72 72 if not peer.capable('branchmap'):
73 73 if branches:
74 74 raise error.Abort(_("remote branch lookup not supported"))
75 75 revs.append(hashbranch)
76 76 return revs, revs[0]
77 77 branchmap = peer.branchmap()
78 78
79 79 def primary(branch):
80 80 if branch == '.':
81 81 if not lrepo:
82 82 raise error.Abort(_("dirstate branch not accessible"))
83 83 branch = lrepo.dirstate.branch()
84 84 if branch in branchmap:
85 85 revs.extend(node.hex(r) for r in reversed(branchmap[branch]))
86 86 return True
87 87 else:
88 88 return False
89 89
90 90 for branch in branches:
91 91 if not primary(branch):
92 92 raise error.RepoLookupError(_("unknown branch '%s'") % branch)
93 93 if hashbranch:
94 94 if not primary(hashbranch):
95 95 revs.append(hashbranch)
96 96 return revs, revs[0]
97 97
98 98 def parseurl(path, branches=None):
99 99 '''parse url#branch, returning (url, (branch, branches))'''
100 100
101 101 u = util.url(path)
102 102 branch = None
103 103 if u.fragment:
104 104 branch = u.fragment
105 105 u.fragment = None
106 return str(u), (branch, branches or [])
106 return bytes(u), (branch, branches or [])
107 107
108 108 schemes = {
109 109 'bundle': bundlerepo,
110 110 'union': unionrepo,
111 111 'file': _local,
112 112 'http': httppeer,
113 113 'https': httppeer,
114 114 'ssh': sshpeer,
115 115 'static-http': statichttprepo,
116 116 }
117 117
118 118 def _peerlookup(path):
119 119 u = util.url(path)
120 120 scheme = u.scheme or 'file'
121 121 thing = schemes.get(scheme) or schemes['file']
122 122 try:
123 123 return thing(path)
124 124 except TypeError:
125 125 # we can't test callable(thing) because 'thing' can be an unloaded
126 126 # module that implements __call__
127 127 if not util.safehasattr(thing, 'instance'):
128 128 raise
129 129 return thing
130 130
131 131 def islocal(repo):
132 132 '''return true if repo (or path pointing to repo) is local'''
133 133 if isinstance(repo, str):
134 134 try:
135 135 return _peerlookup(repo).islocal(repo)
136 136 except AttributeError:
137 137 return False
138 138 return repo.local()
139 139
140 140 def openpath(ui, path):
141 141 '''open path with open if local, url.open if remote'''
142 142 pathurl = util.url(path, parsequery=False, parsefragment=False)
143 143 if pathurl.islocal():
144 144 return util.posixfile(pathurl.localpath(), 'rb')
145 145 else:
146 146 return url.open(ui, path)
147 147
148 148 # a list of (ui, repo) functions called for wire peer initialization
149 149 wirepeersetupfuncs = []
150 150
151 151 def _peerorrepo(ui, path, create=False):
152 152 """return a repository object for the specified path"""
153 153 obj = _peerlookup(path).instance(ui, path, create)
154 154 ui = getattr(obj, "ui", ui)
155 155 for name, module in extensions.extensions(ui):
156 156 hook = getattr(module, 'reposetup', None)
157 157 if hook:
158 158 hook(ui, obj)
159 159 if not obj.local():
160 160 for f in wirepeersetupfuncs:
161 161 f(ui, obj)
162 162 return obj
163 163
164 164 def repository(ui, path='', create=False):
165 165 """return a repository object for the specified path"""
166 166 peer = _peerorrepo(ui, path, create)
167 167 repo = peer.local()
168 168 if not repo:
169 169 raise error.Abort(_("repository '%s' is not local") %
170 170 (path or peer.url()))
171 171 return repo.filtered('visible')
172 172
173 173 def peer(uiorrepo, opts, path, create=False):
174 174 '''return a repository peer for the specified path'''
175 175 rui = remoteui(uiorrepo, opts)
176 176 return _peerorrepo(rui, path, create).peer()
177 177
178 178 def defaultdest(source):
179 179 '''return default destination of clone if none is given
180 180
181 181 >>> defaultdest('foo')
182 182 'foo'
183 183 >>> defaultdest('/foo/bar')
184 184 'bar'
185 185 >>> defaultdest('/')
186 186 ''
187 187 >>> defaultdest('')
188 188 ''
189 189 >>> defaultdest('http://example.org/')
190 190 ''
191 191 >>> defaultdest('http://example.org/foo/')
192 192 'foo'
193 193 '''
194 194 path = util.url(source).path
195 195 if not path:
196 196 return ''
197 197 return os.path.basename(os.path.normpath(path))
198 198
199 199 def share(ui, source, dest=None, update=True, bookmarks=True, defaultpath=None,
200 200 relative=False):
201 201 '''create a shared repository'''
202 202
203 203 if not islocal(source):
204 204 raise error.Abort(_('can only share local repositories'))
205 205
206 206 if not dest:
207 207 dest = defaultdest(source)
208 208 else:
209 209 dest = ui.expandpath(dest)
210 210
211 211 if isinstance(source, str):
212 212 origsource = ui.expandpath(source)
213 213 source, branches = parseurl(origsource)
214 214 srcrepo = repository(ui, source)
215 215 rev, checkout = addbranchrevs(srcrepo, srcrepo, branches, None)
216 216 else:
217 217 srcrepo = source.local()
218 218 origsource = source = srcrepo.url()
219 219 checkout = None
220 220
221 221 sharedpath = srcrepo.sharedpath # if our source is already sharing
222 222
223 223 destwvfs = vfsmod.vfs(dest, realpath=True)
224 224 destvfs = vfsmod.vfs(os.path.join(destwvfs.base, '.hg'), realpath=True)
225 225
226 226 if destvfs.lexists():
227 227 raise error.Abort(_('destination already exists'))
228 228
229 229 if not destwvfs.isdir():
230 230 destwvfs.mkdir()
231 231 destvfs.makedir()
232 232
233 233 requirements = ''
234 234 try:
235 235 requirements = srcrepo.vfs.read('requires')
236 236 except IOError as inst:
237 237 if inst.errno != errno.ENOENT:
238 238 raise
239 239
240 240 if relative:
241 241 try:
242 242 sharedpath = os.path.relpath(sharedpath, destvfs.base)
243 243 requirements += 'relshared\n'
244 244 except IOError as e:
245 245 raise error.Abort(_('cannot calculate relative path'),
246 246 hint=str(e))
247 247 else:
248 248 requirements += 'shared\n'
249 249
250 250 destvfs.write('requires', requirements)
251 251 destvfs.write('sharedpath', sharedpath)
252 252
253 253 r = repository(ui, destwvfs.base)
254 254 postshare(srcrepo, r, bookmarks=bookmarks, defaultpath=defaultpath)
255 255 _postshareupdate(r, update, checkout=checkout)
256 256
257 257 def postshare(sourcerepo, destrepo, bookmarks=True, defaultpath=None):
258 258 """Called after a new shared repo is created.
259 259
260 260 The new repo only has a requirements file and pointer to the source.
261 261 This function configures additional shared data.
262 262
263 263 Extensions can wrap this function and write additional entries to
264 264 destrepo/.hg/shared to indicate additional pieces of data to be shared.
265 265 """
266 266 default = defaultpath or sourcerepo.ui.config('paths', 'default')
267 267 if default:
268 268 fp = destrepo.vfs("hgrc", "w", text=True)
269 269 fp.write("[paths]\n")
270 270 fp.write("default = %s\n" % default)
271 271 fp.close()
272 272
273 273 with destrepo.wlock():
274 274 if bookmarks:
275 275 fp = destrepo.vfs('shared', 'w')
276 276 fp.write(sharedbookmarks + '\n')
277 277 fp.close()
278 278
279 279 def _postshareupdate(repo, update, checkout=None):
280 280 """Maybe perform a working directory update after a shared repo is created.
281 281
282 282 ``update`` can be a boolean or a revision to update to.
283 283 """
284 284 if not update:
285 285 return
286 286
287 287 repo.ui.status(_("updating working directory\n"))
288 288 if update is not True:
289 289 checkout = update
290 290 for test in (checkout, 'default', 'tip'):
291 291 if test is None:
292 292 continue
293 293 try:
294 294 uprev = repo.lookup(test)
295 295 break
296 296 except error.RepoLookupError:
297 297 continue
298 298 _update(repo, uprev)
299 299
300 300 def copystore(ui, srcrepo, destpath):
301 301 '''copy files from store of srcrepo in destpath
302 302
303 303 returns destlock
304 304 '''
305 305 destlock = None
306 306 try:
307 307 hardlink = None
308 308 num = 0
309 309 closetopic = [None]
310 310 def prog(topic, pos):
311 311 if pos is None:
312 312 closetopic[0] = topic
313 313 else:
314 314 ui.progress(topic, pos + num)
315 315 srcpublishing = srcrepo.publishing()
316 316 srcvfs = vfsmod.vfs(srcrepo.sharedpath)
317 317 dstvfs = vfsmod.vfs(destpath)
318 318 for f in srcrepo.store.copylist():
319 319 if srcpublishing and f.endswith('phaseroots'):
320 320 continue
321 321 dstbase = os.path.dirname(f)
322 322 if dstbase and not dstvfs.exists(dstbase):
323 323 dstvfs.mkdir(dstbase)
324 324 if srcvfs.exists(f):
325 325 if f.endswith('data'):
326 326 # 'dstbase' may be empty (e.g. revlog format 0)
327 327 lockfile = os.path.join(dstbase, "lock")
328 328 # lock to avoid premature writing to the target
329 329 destlock = lock.lock(dstvfs, lockfile)
330 330 hardlink, n = util.copyfiles(srcvfs.join(f), dstvfs.join(f),
331 331 hardlink, progress=prog)
332 332 num += n
333 333 if hardlink:
334 334 ui.debug("linked %d files\n" % num)
335 335 if closetopic[0]:
336 336 ui.progress(closetopic[0], None)
337 337 else:
338 338 ui.debug("copied %d files\n" % num)
339 339 if closetopic[0]:
340 340 ui.progress(closetopic[0], None)
341 341 return destlock
342 342 except: # re-raises
343 343 release(destlock)
344 344 raise
345 345
346 346 def clonewithshare(ui, peeropts, sharepath, source, srcpeer, dest, pull=False,
347 347 rev=None, update=True, stream=False):
348 348 """Perform a clone using a shared repo.
349 349
350 350 The store for the repository will be located at <sharepath>/.hg. The
351 351 specified revisions will be cloned or pulled from "source". A shared repo
352 352 will be created at "dest" and a working copy will be created if "update" is
353 353 True.
354 354 """
355 355 revs = None
356 356 if rev:
357 357 if not srcpeer.capable('lookup'):
358 358 raise error.Abort(_("src repository does not support "
359 359 "revision lookup and so doesn't "
360 360 "support clone by revision"))
361 361 revs = [srcpeer.lookup(r) for r in rev]
362 362
363 363 # Obtain a lock before checking for or cloning the pooled repo otherwise
364 364 # 2 clients may race creating or populating it.
365 365 pooldir = os.path.dirname(sharepath)
366 366 # lock class requires the directory to exist.
367 367 try:
368 368 util.makedir(pooldir, False)
369 369 except OSError as e:
370 370 if e.errno != errno.EEXIST:
371 371 raise
372 372
373 373 poolvfs = vfsmod.vfs(pooldir)
374 374 basename = os.path.basename(sharepath)
375 375
376 376 with lock.lock(poolvfs, '%s.lock' % basename):
377 377 if os.path.exists(sharepath):
378 378 ui.status(_('(sharing from existing pooled repository %s)\n') %
379 379 basename)
380 380 else:
381 381 ui.status(_('(sharing from new pooled repository %s)\n') % basename)
382 382 # Always use pull mode because hardlinks in share mode don't work
383 383 # well. Never update because working copies aren't necessary in
384 384 # share mode.
385 385 clone(ui, peeropts, source, dest=sharepath, pull=True,
386 386 rev=rev, update=False, stream=stream)
387 387
388 388 # Resolve the value to put in [paths] section for the source.
389 389 if islocal(source):
390 390 defaultpath = os.path.abspath(util.urllocalpath(source))
391 391 else:
392 392 defaultpath = source
393 393
394 394 sharerepo = repository(ui, path=sharepath)
395 395 share(ui, sharerepo, dest=dest, update=False, bookmarks=False,
396 396 defaultpath=defaultpath)
397 397
398 398 # We need to perform a pull against the dest repo to fetch bookmarks
399 399 # and other non-store data that isn't shared by default. In the case of
400 400 # non-existing shared repo, this means we pull from the remote twice. This
401 401 # is a bit weird. But at the time it was implemented, there wasn't an easy
402 402 # way to pull just non-changegroup data.
403 403 destrepo = repository(ui, path=dest)
404 404 exchange.pull(destrepo, srcpeer, heads=revs)
405 405
406 406 _postshareupdate(destrepo, update)
407 407
408 408 return srcpeer, peer(ui, peeropts, dest)
409 409
410 410 def clone(ui, peeropts, source, dest=None, pull=False, rev=None,
411 411 update=True, stream=False, branch=None, shareopts=None):
412 412 """Make a copy of an existing repository.
413 413
414 414 Create a copy of an existing repository in a new directory. The
415 415 source and destination are URLs, as passed to the repository
416 416 function. Returns a pair of repository peers, the source and
417 417 newly created destination.
418 418
419 419 The location of the source is added to the new repository's
420 420 .hg/hgrc file, as the default to be used for future pulls and
421 421 pushes.
422 422
423 423 If an exception is raised, the partly cloned/updated destination
424 424 repository will be deleted.
425 425
426 426 Arguments:
427 427
428 428 source: repository object or URL
429 429
430 430 dest: URL of destination repository to create (defaults to base
431 431 name of source repository)
432 432
433 433 pull: always pull from source repository, even in local case or if the
434 434 server prefers streaming
435 435
436 436 stream: stream raw data uncompressed from repository (fast over
437 437 LAN, slow over WAN)
438 438
439 439 rev: revision to clone up to (implies pull=True)
440 440
441 441 update: update working directory after clone completes, if
442 442 destination is local repository (True means update to default rev,
443 443 anything else is treated as a revision)
444 444
445 445 branch: branches to clone
446 446
447 447 shareopts: dict of options to control auto sharing behavior. The "pool" key
448 448 activates auto sharing mode and defines the directory for stores. The
449 449 "mode" key determines how to construct the directory name of the shared
450 450 repository. "identity" means the name is derived from the node of the first
451 451 changeset in the repository. "remote" means the name is derived from the
452 452 remote's path/URL. Defaults to "identity."
453 453 """
454 454
455 455 if isinstance(source, str):
456 456 origsource = ui.expandpath(source)
457 457 source, branch = parseurl(origsource, branch)
458 458 srcpeer = peer(ui, peeropts, source)
459 459 else:
460 460 srcpeer = source.peer() # in case we were called with a localrepo
461 461 branch = (None, branch or [])
462 462 origsource = source = srcpeer.url()
463 463 rev, checkout = addbranchrevs(srcpeer, srcpeer, branch, rev)
464 464
465 465 if dest is None:
466 466 dest = defaultdest(source)
467 467 if dest:
468 468 ui.status(_("destination directory: %s\n") % dest)
469 469 else:
470 470 dest = ui.expandpath(dest)
471 471
472 472 dest = util.urllocalpath(dest)
473 473 source = util.urllocalpath(source)
474 474
475 475 if not dest:
476 476 raise error.Abort(_("empty destination path is not valid"))
477 477
478 478 destvfs = vfsmod.vfs(dest, expandpath=True)
479 479 if destvfs.lexists():
480 480 if not destvfs.isdir():
481 481 raise error.Abort(_("destination '%s' already exists") % dest)
482 482 elif destvfs.listdir():
483 483 raise error.Abort(_("destination '%s' is not empty") % dest)
484 484
485 485 shareopts = shareopts or {}
486 486 sharepool = shareopts.get('pool')
487 487 sharenamemode = shareopts.get('mode')
488 488 if sharepool and islocal(dest):
489 489 sharepath = None
490 490 if sharenamemode == 'identity':
491 491 # Resolve the name from the initial changeset in the remote
492 492 # repository. This returns nullid when the remote is empty. It
493 493 # raises RepoLookupError if revision 0 is filtered or otherwise
494 494 # not available. If we fail to resolve, sharing is not enabled.
495 495 try:
496 496 rootnode = srcpeer.lookup('0')
497 497 if rootnode != node.nullid:
498 498 sharepath = os.path.join(sharepool, node.hex(rootnode))
499 499 else:
500 500 ui.status(_('(not using pooled storage: '
501 501 'remote appears to be empty)\n'))
502 502 except error.RepoLookupError:
503 503 ui.status(_('(not using pooled storage: '
504 504 'unable to resolve identity of remote)\n'))
505 505 elif sharenamemode == 'remote':
506 506 sharepath = os.path.join(
507 507 sharepool, hashlib.sha1(source).hexdigest())
508 508 else:
509 509 raise error.Abort(_('unknown share naming mode: %s') %
510 510 sharenamemode)
511 511
512 512 if sharepath:
513 513 return clonewithshare(ui, peeropts, sharepath, source, srcpeer,
514 514 dest, pull=pull, rev=rev, update=update,
515 515 stream=stream)
516 516
517 517 srclock = destlock = cleandir = None
518 518 srcrepo = srcpeer.local()
519 519 try:
520 520 abspath = origsource
521 521 if islocal(origsource):
522 522 abspath = os.path.abspath(util.urllocalpath(origsource))
523 523
524 524 if islocal(dest):
525 525 cleandir = dest
526 526
527 527 copy = False
528 528 if (srcrepo and srcrepo.cancopy() and islocal(dest)
529 529 and not phases.hassecret(srcrepo)):
530 530 copy = not pull and not rev
531 531
532 532 if copy:
533 533 try:
534 534 # we use a lock here because if we race with commit, we
535 535 # can end up with extra data in the cloned revlogs that's
536 536 # not pointed to by changesets, thus causing verify to
537 537 # fail
538 538 srclock = srcrepo.lock(wait=False)
539 539 except error.LockError:
540 540 copy = False
541 541
542 542 if copy:
543 543 srcrepo.hook('preoutgoing', throw=True, source='clone')
544 544 hgdir = os.path.realpath(os.path.join(dest, ".hg"))
545 545 if not os.path.exists(dest):
546 546 os.mkdir(dest)
547 547 else:
548 548 # only clean up directories we create ourselves
549 549 cleandir = hgdir
550 550 try:
551 551 destpath = hgdir
552 552 util.makedir(destpath, notindexed=True)
553 553 except OSError as inst:
554 554 if inst.errno == errno.EEXIST:
555 555 cleandir = None
556 556 raise error.Abort(_("destination '%s' already exists")
557 557 % dest)
558 558 raise
559 559
560 560 destlock = copystore(ui, srcrepo, destpath)
561 561 # copy bookmarks over
562 562 srcbookmarks = srcrepo.vfs.join('bookmarks')
563 563 dstbookmarks = os.path.join(destpath, 'bookmarks')
564 564 if os.path.exists(srcbookmarks):
565 565 util.copyfile(srcbookmarks, dstbookmarks)
566 566
567 567 # Recomputing branch cache might be slow on big repos,
568 568 # so just copy it
569 569 def copybranchcache(fname):
570 570 srcbranchcache = srcrepo.vfs.join('cache/%s' % fname)
571 571 dstbranchcache = os.path.join(dstcachedir, fname)
572 572 if os.path.exists(srcbranchcache):
573 573 if not os.path.exists(dstcachedir):
574 574 os.mkdir(dstcachedir)
575 575 util.copyfile(srcbranchcache, dstbranchcache)
576 576
577 577 dstcachedir = os.path.join(destpath, 'cache')
578 578 # In local clones we're copying all nodes, not just served
579 579 # ones. Therefore copy all branch caches over.
580 580 copybranchcache('branch2')
581 581 for cachename in repoview.filtertable:
582 582 copybranchcache('branch2-%s' % cachename)
583 583
584 584 # we need to re-init the repo after manually copying the data
585 585 # into it
586 586 destpeer = peer(srcrepo, peeropts, dest)
587 587 srcrepo.hook('outgoing', source='clone',
588 588 node=node.hex(node.nullid))
589 589 else:
590 590 try:
591 591 destpeer = peer(srcrepo or ui, peeropts, dest, create=True)
592 592 # only pass ui when no srcrepo
593 593 except OSError as inst:
594 594 if inst.errno == errno.EEXIST:
595 595 cleandir = None
596 596 raise error.Abort(_("destination '%s' already exists")
597 597 % dest)
598 598 raise
599 599
600 600 revs = None
601 601 if rev:
602 602 if not srcpeer.capable('lookup'):
603 603 raise error.Abort(_("src repository does not support "
604 604 "revision lookup and so doesn't "
605 605 "support clone by revision"))
606 606 revs = [srcpeer.lookup(r) for r in rev]
607 607 checkout = revs[0]
608 608 local = destpeer.local()
609 609 if local:
610 610 if not stream:
611 611 if pull:
612 612 stream = False
613 613 else:
614 614 stream = None
615 615 # internal config: ui.quietbookmarkmove
616 616 overrides = {('ui', 'quietbookmarkmove'): True}
617 617 with local.ui.configoverride(overrides, 'clone'):
618 618 exchange.pull(local, srcpeer, revs,
619 619 streamclonerequested=stream)
620 620 elif srcrepo:
621 621 exchange.push(srcrepo, destpeer, revs=revs,
622 622 bookmarks=srcrepo._bookmarks.keys())
623 623 else:
624 624 raise error.Abort(_("clone from remote to remote not supported")
625 625 )
626 626
627 627 cleandir = None
628 628
629 629 destrepo = destpeer.local()
630 630 if destrepo:
631 631 template = uimod.samplehgrcs['cloned']
632 632 fp = destrepo.vfs("hgrc", "w", text=True)
633 633 u = util.url(abspath)
634 634 u.passwd = None
635 635 defaulturl = str(u)
636 636 fp.write(template % defaulturl)
637 637 fp.close()
638 638
639 639 destrepo.ui.setconfig('paths', 'default', defaulturl, 'clone')
640 640
641 641 if update:
642 642 if update is not True:
643 643 checkout = srcpeer.lookup(update)
644 644 uprev = None
645 645 status = None
646 646 if checkout is not None:
647 647 try:
648 648 uprev = destrepo.lookup(checkout)
649 649 except error.RepoLookupError:
650 650 if update is not True:
651 651 try:
652 652 uprev = destrepo.lookup(update)
653 653 except error.RepoLookupError:
654 654 pass
655 655 if uprev is None:
656 656 try:
657 657 uprev = destrepo._bookmarks['@']
658 658 update = '@'
659 659 bn = destrepo[uprev].branch()
660 660 if bn == 'default':
661 661 status = _("updating to bookmark @\n")
662 662 else:
663 663 status = (_("updating to bookmark @ on branch %s\n")
664 664 % bn)
665 665 except KeyError:
666 666 try:
667 667 uprev = destrepo.branchtip('default')
668 668 except error.RepoLookupError:
669 669 uprev = destrepo.lookup('tip')
670 670 if not status:
671 671 bn = destrepo[uprev].branch()
672 672 status = _("updating to branch %s\n") % bn
673 673 destrepo.ui.status(status)
674 674 _update(destrepo, uprev)
675 675 if update in destrepo._bookmarks:
676 676 bookmarks.activate(destrepo, update)
677 677 finally:
678 678 release(srclock, destlock)
679 679 if cleandir is not None:
680 680 shutil.rmtree(cleandir, True)
681 681 if srcpeer is not None:
682 682 srcpeer.close()
683 683 return srcpeer, destpeer
684 684
685 685 def _showstats(repo, stats, quietempty=False):
686 686 if quietempty and not any(stats):
687 687 return
688 688 repo.ui.status(_("%d files updated, %d files merged, "
689 689 "%d files removed, %d files unresolved\n") % stats)
690 690
691 691 def updaterepo(repo, node, overwrite, updatecheck=None):
692 692 """Update the working directory to node.
693 693
694 694 When overwrite is set, changes are clobbered, merged else
695 695
696 696 returns stats (see pydoc mercurial.merge.applyupdates)"""
697 697 return mergemod.update(repo, node, False, overwrite,
698 698 labels=['working copy', 'destination'],
699 699 updatecheck=updatecheck)
700 700
701 701 def update(repo, node, quietempty=False, updatecheck=None):
702 702 """update the working directory to node"""
703 703 stats = updaterepo(repo, node, False, updatecheck=updatecheck)
704 704 _showstats(repo, stats, quietempty)
705 705 if stats[3]:
706 706 repo.ui.status(_("use 'hg resolve' to retry unresolved file merges\n"))
707 707 return stats[3] > 0
708 708
709 709 # naming conflict in clone()
710 710 _update = update
711 711
712 712 def clean(repo, node, show_stats=True, quietempty=False):
713 713 """forcibly switch the working directory to node, clobbering changes"""
714 714 stats = updaterepo(repo, node, True)
715 715 repo.vfs.unlinkpath('graftstate', ignoremissing=True)
716 716 if show_stats:
717 717 _showstats(repo, stats, quietempty)
718 718 return stats[3] > 0
719 719
720 720 # naming conflict in updatetotally()
721 721 _clean = clean
722 722
723 723 def updatetotally(ui, repo, checkout, brev, clean=False, updatecheck=None):
724 724 """Update the working directory with extra care for non-file components
725 725
726 726 This takes care of non-file components below:
727 727
728 728 :bookmark: might be advanced or (in)activated
729 729
730 730 This takes arguments below:
731 731
732 732 :checkout: to which revision the working directory is updated
733 733 :brev: a name, which might be a bookmark to be activated after updating
734 734 :clean: whether changes in the working directory can be discarded
735 735 :updatecheck: how to deal with a dirty working directory
736 736
737 737 Valid values for updatecheck are (None => linear):
738 738
739 739 * abort: abort if the working directory is dirty
740 740 * none: don't check (merge working directory changes into destination)
741 741 * linear: check that update is linear before merging working directory
742 742 changes into destination
743 743 * noconflict: check that the update does not result in file merges
744 744
745 745 This returns whether conflict is detected at updating or not.
746 746 """
747 747 if updatecheck is None:
748 748 updatecheck = ui.config('experimental', 'updatecheck')
749 749 if updatecheck not in ('abort', 'none', 'linear', 'noconflict'):
750 750 # If not configured, or invalid value configured
751 751 updatecheck = 'linear'
752 752 with repo.wlock():
753 753 movemarkfrom = None
754 754 warndest = False
755 755 if checkout is None:
756 756 updata = destutil.destupdate(repo, clean=clean)
757 757 checkout, movemarkfrom, brev = updata
758 758 warndest = True
759 759
760 760 if clean:
761 761 ret = _clean(repo, checkout)
762 762 else:
763 763 if updatecheck == 'abort':
764 764 cmdutil.bailifchanged(repo, merge=False)
765 765 updatecheck = 'none'
766 766 ret = _update(repo, checkout, updatecheck=updatecheck)
767 767
768 768 if not ret and movemarkfrom:
769 769 if movemarkfrom == repo['.'].node():
770 770 pass # no-op update
771 771 elif bookmarks.update(repo, [movemarkfrom], repo['.'].node()):
772 772 b = ui.label(repo._activebookmark, 'bookmarks.active')
773 773 ui.status(_("updating bookmark %s\n") % b)
774 774 else:
775 775 # this can happen with a non-linear update
776 776 b = ui.label(repo._activebookmark, 'bookmarks')
777 777 ui.status(_("(leaving bookmark %s)\n") % b)
778 778 bookmarks.deactivate(repo)
779 779 elif brev in repo._bookmarks:
780 780 if brev != repo._activebookmark:
781 781 b = ui.label(brev, 'bookmarks.active')
782 782 ui.status(_("(activating bookmark %s)\n") % b)
783 783 bookmarks.activate(repo, brev)
784 784 elif brev:
785 785 if repo._activebookmark:
786 786 b = ui.label(repo._activebookmark, 'bookmarks')
787 787 ui.status(_("(leaving bookmark %s)\n") % b)
788 788 bookmarks.deactivate(repo)
789 789
790 790 if warndest:
791 791 destutil.statusotherdests(ui, repo)
792 792
793 793 return ret
794 794
795 795 def merge(repo, node, force=None, remind=True, mergeforce=False, labels=None):
796 796 """Branch merge with node, resolving changes. Return true if any
797 797 unresolved conflicts."""
798 798 stats = mergemod.update(repo, node, True, force, mergeforce=mergeforce,
799 799 labels=labels)
800 800 _showstats(repo, stats)
801 801 if stats[3]:
802 802 repo.ui.status(_("use 'hg resolve' to retry unresolved file merges "
803 803 "or 'hg update -C .' to abandon\n"))
804 804 elif remind:
805 805 repo.ui.status(_("(branch merge, don't forget to commit)\n"))
806 806 return stats[3] > 0
807 807
808 808 def _incoming(displaychlist, subreporecurse, ui, repo, source,
809 809 opts, buffered=False):
810 810 """
811 811 Helper for incoming / gincoming.
812 812 displaychlist gets called with
813 813 (remoterepo, incomingchangesetlist, displayer) parameters,
814 814 and is supposed to contain only code that can't be unified.
815 815 """
816 816 source, branches = parseurl(ui.expandpath(source), opts.get('branch'))
817 817 other = peer(repo, opts, source)
818 818 ui.status(_('comparing with %s\n') % util.hidepassword(source))
819 819 revs, checkout = addbranchrevs(repo, other, branches, opts.get('rev'))
820 820
821 821 if revs:
822 822 revs = [other.lookup(rev) for rev in revs]
823 823 other, chlist, cleanupfn = bundlerepo.getremotechanges(ui, repo, other,
824 824 revs, opts["bundle"], opts["force"])
825 825 try:
826 826 if not chlist:
827 827 ui.status(_("no changes found\n"))
828 828 return subreporecurse()
829 829 ui.pager('incoming')
830 830 displayer = cmdutil.show_changeset(ui, other, opts, buffered)
831 831 displaychlist(other, chlist, displayer)
832 832 displayer.close()
833 833 finally:
834 834 cleanupfn()
835 835 subreporecurse()
836 836 return 0 # exit code is zero since we found incoming changes
837 837
838 838 def incoming(ui, repo, source, opts):
839 839 def subreporecurse():
840 840 ret = 1
841 841 if opts.get('subrepos'):
842 842 ctx = repo[None]
843 843 for subpath in sorted(ctx.substate):
844 844 sub = ctx.sub(subpath)
845 845 ret = min(ret, sub.incoming(ui, source, opts))
846 846 return ret
847 847
848 848 def display(other, chlist, displayer):
849 849 limit = cmdutil.loglimit(opts)
850 850 if opts.get('newest_first'):
851 851 chlist.reverse()
852 852 count = 0
853 853 for n in chlist:
854 854 if limit is not None and count >= limit:
855 855 break
856 856 parents = [p for p in other.changelog.parents(n) if p != nullid]
857 857 if opts.get('no_merges') and len(parents) == 2:
858 858 continue
859 859 count += 1
860 860 displayer.show(other[n])
861 861 return _incoming(display, subreporecurse, ui, repo, source, opts)
862 862
863 863 def _outgoing(ui, repo, dest, opts):
864 864 dest = ui.expandpath(dest or 'default-push', dest or 'default')
865 865 dest, branches = parseurl(dest, opts.get('branch'))
866 866 ui.status(_('comparing with %s\n') % util.hidepassword(dest))
867 867 revs, checkout = addbranchrevs(repo, repo, branches, opts.get('rev'))
868 868 if revs:
869 869 revs = [repo.lookup(rev) for rev in scmutil.revrange(repo, revs)]
870 870
871 871 other = peer(repo, opts, dest)
872 872 outgoing = discovery.findcommonoutgoing(repo.unfiltered(), other, revs,
873 873 force=opts.get('force'))
874 874 o = outgoing.missing
875 875 if not o:
876 876 scmutil.nochangesfound(repo.ui, repo, outgoing.excluded)
877 877 return o, other
878 878
879 879 def outgoing(ui, repo, dest, opts):
880 880 def recurse():
881 881 ret = 1
882 882 if opts.get('subrepos'):
883 883 ctx = repo[None]
884 884 for subpath in sorted(ctx.substate):
885 885 sub = ctx.sub(subpath)
886 886 ret = min(ret, sub.outgoing(ui, dest, opts))
887 887 return ret
888 888
889 889 limit = cmdutil.loglimit(opts)
890 890 o, other = _outgoing(ui, repo, dest, opts)
891 891 if not o:
892 892 cmdutil.outgoinghooks(ui, repo, other, opts, o)
893 893 return recurse()
894 894
895 895 if opts.get('newest_first'):
896 896 o.reverse()
897 897 ui.pager('outgoing')
898 898 displayer = cmdutil.show_changeset(ui, repo, opts)
899 899 count = 0
900 900 for n in o:
901 901 if limit is not None and count >= limit:
902 902 break
903 903 parents = [p for p in repo.changelog.parents(n) if p != nullid]
904 904 if opts.get('no_merges') and len(parents) == 2:
905 905 continue
906 906 count += 1
907 907 displayer.show(repo[n])
908 908 displayer.close()
909 909 cmdutil.outgoinghooks(ui, repo, other, opts, o)
910 910 recurse()
911 911 return 0 # exit code is zero since we found outgoing changes
912 912
913 913 def verify(repo):
914 914 """verify the consistency of a repository"""
915 915 ret = verifymod.verify(repo)
916 916
917 917 # Broken subrepo references in hidden csets don't seem worth worrying about,
918 918 # since they can't be pushed/pulled, and --hidden can be used if they are a
919 919 # concern.
920 920
921 921 # pathto() is needed for -R case
922 922 revs = repo.revs("filelog(%s)",
923 923 util.pathto(repo.root, repo.getcwd(), '.hgsubstate'))
924 924
925 925 if revs:
926 926 repo.ui.status(_('checking subrepo links\n'))
927 927 for rev in revs:
928 928 ctx = repo[rev]
929 929 try:
930 930 for subpath in ctx.substate:
931 931 try:
932 932 ret = (ctx.sub(subpath, allowcreate=False).verify()
933 933 or ret)
934 934 except error.RepoError as e:
935 935 repo.ui.warn(('%s: %s\n') % (rev, e))
936 936 except Exception:
937 937 repo.ui.warn(_('.hgsubstate is corrupt in revision %s\n') %
938 938 node.short(ctx.node()))
939 939
940 940 return ret
941 941
942 942 def remoteui(src, opts):
943 943 'build a remote ui from ui or repo and opts'
944 944 if util.safehasattr(src, 'baseui'): # looks like a repository
945 945 dst = src.baseui.copy() # drop repo-specific config
946 946 src = src.ui # copy target options from repo
947 947 else: # assume it's a global ui object
948 948 dst = src.copy() # keep all global options
949 949
950 950 # copy ssh-specific options
951 951 for o in 'ssh', 'remotecmd':
952 952 v = opts.get(o) or src.config('ui', o)
953 953 if v:
954 954 dst.setconfig("ui", o, v, 'copied')
955 955
956 956 # copy bundle-specific options
957 957 r = src.config('bundle', 'mainreporoot')
958 958 if r:
959 959 dst.setconfig('bundle', 'mainreporoot', r, 'copied')
960 960
961 961 # copy selected local settings to the remote ui
962 962 for sect in ('auth', 'hostfingerprints', 'hostsecurity', 'http_proxy'):
963 963 for key, val in src.configitems(sect):
964 964 dst.setconfig(sect, key, val, 'copied')
965 965 v = src.config('web', 'cacerts')
966 966 if v:
967 967 dst.setconfig('web', 'cacerts', util.expandpath(v), 'copied')
968 968
969 969 return dst
970 970
971 971 # Files of interest
972 972 # Used to check if the repository has changed looking at mtime and size of
973 973 # these files.
974 974 foi = [('spath', '00changelog.i'),
975 975 ('spath', 'phaseroots'), # ! phase can change content at the same size
976 976 ('spath', 'obsstore'),
977 977 ('path', 'bookmarks'), # ! bookmark can change content at the same size
978 978 ]
979 979
980 980 class cachedlocalrepo(object):
981 981 """Holds a localrepository that can be cached and reused."""
982 982
983 983 def __init__(self, repo):
984 984 """Create a new cached repo from an existing repo.
985 985
986 986 We assume the passed in repo was recently created. If the
987 987 repo has changed between when it was created and when it was
988 988 turned into a cache, it may not refresh properly.
989 989 """
990 990 assert isinstance(repo, localrepo.localrepository)
991 991 self._repo = repo
992 992 self._state, self.mtime = self._repostate()
993 993 self._filtername = repo.filtername
994 994
995 995 def fetch(self):
996 996 """Refresh (if necessary) and return a repository.
997 997
998 998 If the cached instance is out of date, it will be recreated
999 999 automatically and returned.
1000 1000
1001 1001 Returns a tuple of the repo and a boolean indicating whether a new
1002 1002 repo instance was created.
1003 1003 """
1004 1004 # We compare the mtimes and sizes of some well-known files to
1005 1005 # determine if the repo changed. This is not precise, as mtimes
1006 1006 # are susceptible to clock skew and imprecise filesystems and
1007 1007 # file content can change while maintaining the same size.
1008 1008
1009 1009 state, mtime = self._repostate()
1010 1010 if state == self._state:
1011 1011 return self._repo, False
1012 1012
1013 1013 repo = repository(self._repo.baseui, self._repo.url())
1014 1014 if self._filtername:
1015 1015 self._repo = repo.filtered(self._filtername)
1016 1016 else:
1017 1017 self._repo = repo.unfiltered()
1018 1018 self._state = state
1019 1019 self.mtime = mtime
1020 1020
1021 1021 return self._repo, True
1022 1022
1023 1023 def _repostate(self):
1024 1024 state = []
1025 1025 maxmtime = -1
1026 1026 for attr, fname in foi:
1027 1027 prefix = getattr(self._repo, attr)
1028 1028 p = os.path.join(prefix, fname)
1029 1029 try:
1030 1030 st = os.stat(p)
1031 1031 except OSError:
1032 1032 st = os.stat(prefix)
1033 1033 state.append((st.st_mtime, st.st_size))
1034 1034 maxmtime = max(maxmtime, st.st_mtime)
1035 1035
1036 1036 return tuple(state), maxmtime
1037 1037
1038 1038 def copy(self):
1039 1039 """Obtain a copy of this class instance.
1040 1040
1041 1041 A new localrepository instance is obtained. The new instance should be
1042 1042 completely independent of the original.
1043 1043 """
1044 1044 repo = repository(self._repo.baseui, self._repo.origroot)
1045 1045 if self._filtername:
1046 1046 repo = repo.filtered(self._filtername)
1047 1047 else:
1048 1048 repo = repo.unfiltered()
1049 1049 c = cachedlocalrepo(repo)
1050 1050 c._state = self._state
1051 1051 c.mtime = self.mtime
1052 1052 return c
@@ -1,3719 +1,3719
1 1 # util.py - Mercurial utility functions and platform specific implementations
2 2 #
3 3 # Copyright 2005 K. Thananchayan <thananck@yahoo.com>
4 4 # Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
5 5 # Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
6 6 #
7 7 # This software may be used and distributed according to the terms of the
8 8 # GNU General Public License version 2 or any later version.
9 9
10 10 """Mercurial utility functions and platform specific implementations.
11 11
12 12 This contains helper routines that are independent of the SCM core and
13 13 hide platform-specific details from the core.
14 14 """
15 15
16 16 from __future__ import absolute_import
17 17
18 18 import bz2
19 19 import calendar
20 20 import codecs
21 21 import collections
22 22 import datetime
23 23 import errno
24 24 import gc
25 25 import hashlib
26 26 import imp
27 27 import os
28 28 import platform as pyplatform
29 29 import re as remod
30 30 import shutil
31 31 import signal
32 32 import socket
33 33 import stat
34 34 import string
35 35 import subprocess
36 36 import sys
37 37 import tempfile
38 38 import textwrap
39 39 import time
40 40 import traceback
41 41 import zlib
42 42
43 43 from . import (
44 44 encoding,
45 45 error,
46 46 i18n,
47 47 osutil,
48 48 parsers,
49 49 pycompat,
50 50 )
51 51
52 52 empty = pycompat.empty
53 53 httplib = pycompat.httplib
54 54 httpserver = pycompat.httpserver
55 55 pickle = pycompat.pickle
56 56 queue = pycompat.queue
57 57 socketserver = pycompat.socketserver
58 58 stderr = pycompat.stderr
59 59 stdin = pycompat.stdin
60 60 stdout = pycompat.stdout
61 61 stringio = pycompat.stringio
62 62 urlerr = pycompat.urlerr
63 63 urlreq = pycompat.urlreq
64 64 xmlrpclib = pycompat.xmlrpclib
65 65
66 66 def isatty(fp):
67 67 try:
68 68 return fp.isatty()
69 69 except AttributeError:
70 70 return False
71 71
72 72 # glibc determines buffering on first write to stdout - if we replace a TTY
73 73 # destined stdout with a pipe destined stdout (e.g. pager), we want line
74 74 # buffering
75 75 if isatty(stdout):
76 76 stdout = os.fdopen(stdout.fileno(), pycompat.sysstr('wb'), 1)
77 77
78 78 if pycompat.osname == 'nt':
79 79 from . import windows as platform
80 80 stdout = platform.winstdout(stdout)
81 81 else:
82 82 from . import posix as platform
83 83
84 84 _ = i18n._
85 85
86 86 bindunixsocket = platform.bindunixsocket
87 87 cachestat = platform.cachestat
88 88 checkexec = platform.checkexec
89 89 checklink = platform.checklink
90 90 copymode = platform.copymode
91 91 executablepath = platform.executablepath
92 92 expandglobs = platform.expandglobs
93 93 explainexit = platform.explainexit
94 94 findexe = platform.findexe
95 95 gethgcmd = platform.gethgcmd
96 96 getuser = platform.getuser
97 97 getpid = os.getpid
98 98 groupmembers = platform.groupmembers
99 99 groupname = platform.groupname
100 100 hidewindow = platform.hidewindow
101 101 isexec = platform.isexec
102 102 isowner = platform.isowner
103 103 localpath = platform.localpath
104 104 lookupreg = platform.lookupreg
105 105 makedir = platform.makedir
106 106 nlinks = platform.nlinks
107 107 normpath = platform.normpath
108 108 normcase = platform.normcase
109 109 normcasespec = platform.normcasespec
110 110 normcasefallback = platform.normcasefallback
111 111 openhardlinks = platform.openhardlinks
112 112 oslink = platform.oslink
113 113 parsepatchoutput = platform.parsepatchoutput
114 114 pconvert = platform.pconvert
115 115 poll = platform.poll
116 116 popen = platform.popen
117 117 posixfile = platform.posixfile
118 118 quotecommand = platform.quotecommand
119 119 readpipe = platform.readpipe
120 120 rename = platform.rename
121 121 removedirs = platform.removedirs
122 122 samedevice = platform.samedevice
123 123 samefile = platform.samefile
124 124 samestat = platform.samestat
125 125 setbinary = platform.setbinary
126 126 setflags = platform.setflags
127 127 setsignalhandler = platform.setsignalhandler
128 128 shellquote = platform.shellquote
129 129 spawndetached = platform.spawndetached
130 130 split = platform.split
131 131 sshargs = platform.sshargs
132 132 statfiles = getattr(osutil, 'statfiles', platform.statfiles)
133 133 statisexec = platform.statisexec
134 134 statislink = platform.statislink
135 135 testpid = platform.testpid
136 136 umask = platform.umask
137 137 unlink = platform.unlink
138 138 username = platform.username
139 139
140 140 # Python compatibility
141 141
142 142 _notset = object()
143 143
144 144 # disable Python's problematic floating point timestamps (issue4836)
145 145 # (Python hypocritically says you shouldn't change this behavior in
146 146 # libraries, and sure enough Mercurial is not a library.)
147 147 os.stat_float_times(False)
148 148
149 149 def safehasattr(thing, attr):
150 150 return getattr(thing, attr, _notset) is not _notset
151 151
152 152 def bitsfrom(container):
153 153 bits = 0
154 154 for bit in container:
155 155 bits |= bit
156 156 return bits
157 157
158 158 DIGESTS = {
159 159 'md5': hashlib.md5,
160 160 'sha1': hashlib.sha1,
161 161 'sha512': hashlib.sha512,
162 162 }
163 163 # List of digest types from strongest to weakest
164 164 DIGESTS_BY_STRENGTH = ['sha512', 'sha1', 'md5']
165 165
166 166 for k in DIGESTS_BY_STRENGTH:
167 167 assert k in DIGESTS
168 168
169 169 class digester(object):
170 170 """helper to compute digests.
171 171
172 172 This helper can be used to compute one or more digests given their name.
173 173
174 174 >>> d = digester(['md5', 'sha1'])
175 175 >>> d.update('foo')
176 176 >>> [k for k in sorted(d)]
177 177 ['md5', 'sha1']
178 178 >>> d['md5']
179 179 'acbd18db4cc2f85cedef654fccc4a4d8'
180 180 >>> d['sha1']
181 181 '0beec7b5ea3f0fdbc95d0dd47f3c5bc275da8a33'
182 182 >>> digester.preferred(['md5', 'sha1'])
183 183 'sha1'
184 184 """
185 185
186 186 def __init__(self, digests, s=''):
187 187 self._hashes = {}
188 188 for k in digests:
189 189 if k not in DIGESTS:
190 190 raise Abort(_('unknown digest type: %s') % k)
191 191 self._hashes[k] = DIGESTS[k]()
192 192 if s:
193 193 self.update(s)
194 194
195 195 def update(self, data):
196 196 for h in self._hashes.values():
197 197 h.update(data)
198 198
199 199 def __getitem__(self, key):
200 200 if key not in DIGESTS:
201 201 raise Abort(_('unknown digest type: %s') % k)
202 202 return self._hashes[key].hexdigest()
203 203
204 204 def __iter__(self):
205 205 return iter(self._hashes)
206 206
207 207 @staticmethod
208 208 def preferred(supported):
209 209 """returns the strongest digest type in both supported and DIGESTS."""
210 210
211 211 for k in DIGESTS_BY_STRENGTH:
212 212 if k in supported:
213 213 return k
214 214 return None
215 215
216 216 class digestchecker(object):
217 217 """file handle wrapper that additionally checks content against a given
218 218 size and digests.
219 219
220 220 d = digestchecker(fh, size, {'md5': '...'})
221 221
222 222 When multiple digests are given, all of them are validated.
223 223 """
224 224
225 225 def __init__(self, fh, size, digests):
226 226 self._fh = fh
227 227 self._size = size
228 228 self._got = 0
229 229 self._digests = dict(digests)
230 230 self._digester = digester(self._digests.keys())
231 231
232 232 def read(self, length=-1):
233 233 content = self._fh.read(length)
234 234 self._digester.update(content)
235 235 self._got += len(content)
236 236 return content
237 237
238 238 def validate(self):
239 239 if self._size != self._got:
240 240 raise Abort(_('size mismatch: expected %d, got %d') %
241 241 (self._size, self._got))
242 242 for k, v in self._digests.items():
243 243 if v != self._digester[k]:
244 244 # i18n: first parameter is a digest name
245 245 raise Abort(_('%s mismatch: expected %s, got %s') %
246 246 (k, v, self._digester[k]))
247 247
248 248 try:
249 249 buffer = buffer
250 250 except NameError:
251 251 if not pycompat.ispy3:
252 252 def buffer(sliceable, offset=0, length=None):
253 253 if length is not None:
254 254 return sliceable[offset:offset + length]
255 255 return sliceable[offset:]
256 256 else:
257 257 def buffer(sliceable, offset=0, length=None):
258 258 if length is not None:
259 259 return memoryview(sliceable)[offset:offset + length]
260 260 return memoryview(sliceable)[offset:]
261 261
262 262 closefds = pycompat.osname == 'posix'
263 263
264 264 _chunksize = 4096
265 265
266 266 class bufferedinputpipe(object):
267 267 """a manually buffered input pipe
268 268
269 269 Python will not let us use buffered IO and lazy reading with 'polling' at
270 270 the same time. We cannot probe the buffer state and select will not detect
271 271 that data are ready to read if they are already buffered.
272 272
273 273 This class let us work around that by implementing its own buffering
274 274 (allowing efficient readline) while offering a way to know if the buffer is
275 275 empty from the output (allowing collaboration of the buffer with polling).
276 276
277 277 This class lives in the 'util' module because it makes use of the 'os'
278 278 module from the python stdlib.
279 279 """
280 280
281 281 def __init__(self, input):
282 282 self._input = input
283 283 self._buffer = []
284 284 self._eof = False
285 285 self._lenbuf = 0
286 286
287 287 @property
288 288 def hasbuffer(self):
289 289 """True is any data is currently buffered
290 290
291 291 This will be used externally a pre-step for polling IO. If there is
292 292 already data then no polling should be set in place."""
293 293 return bool(self._buffer)
294 294
295 295 @property
296 296 def closed(self):
297 297 return self._input.closed
298 298
299 299 def fileno(self):
300 300 return self._input.fileno()
301 301
302 302 def close(self):
303 303 return self._input.close()
304 304
305 305 def read(self, size):
306 306 while (not self._eof) and (self._lenbuf < size):
307 307 self._fillbuffer()
308 308 return self._frombuffer(size)
309 309
310 310 def readline(self, *args, **kwargs):
311 311 if 1 < len(self._buffer):
312 312 # this should not happen because both read and readline end with a
313 313 # _frombuffer call that collapse it.
314 314 self._buffer = [''.join(self._buffer)]
315 315 self._lenbuf = len(self._buffer[0])
316 316 lfi = -1
317 317 if self._buffer:
318 318 lfi = self._buffer[-1].find('\n')
319 319 while (not self._eof) and lfi < 0:
320 320 self._fillbuffer()
321 321 if self._buffer:
322 322 lfi = self._buffer[-1].find('\n')
323 323 size = lfi + 1
324 324 if lfi < 0: # end of file
325 325 size = self._lenbuf
326 326 elif 1 < len(self._buffer):
327 327 # we need to take previous chunks into account
328 328 size += self._lenbuf - len(self._buffer[-1])
329 329 return self._frombuffer(size)
330 330
331 331 def _frombuffer(self, size):
332 332 """return at most 'size' data from the buffer
333 333
334 334 The data are removed from the buffer."""
335 335 if size == 0 or not self._buffer:
336 336 return ''
337 337 buf = self._buffer[0]
338 338 if 1 < len(self._buffer):
339 339 buf = ''.join(self._buffer)
340 340
341 341 data = buf[:size]
342 342 buf = buf[len(data):]
343 343 if buf:
344 344 self._buffer = [buf]
345 345 self._lenbuf = len(buf)
346 346 else:
347 347 self._buffer = []
348 348 self._lenbuf = 0
349 349 return data
350 350
351 351 def _fillbuffer(self):
352 352 """read data to the buffer"""
353 353 data = os.read(self._input.fileno(), _chunksize)
354 354 if not data:
355 355 self._eof = True
356 356 else:
357 357 self._lenbuf += len(data)
358 358 self._buffer.append(data)
359 359
360 360 def popen2(cmd, env=None, newlines=False):
361 361 # Setting bufsize to -1 lets the system decide the buffer size.
362 362 # The default for bufsize is 0, meaning unbuffered. This leads to
363 363 # poor performance on Mac OS X: http://bugs.python.org/issue4194
364 364 p = subprocess.Popen(cmd, shell=True, bufsize=-1,
365 365 close_fds=closefds,
366 366 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
367 367 universal_newlines=newlines,
368 368 env=env)
369 369 return p.stdin, p.stdout
370 370
371 371 def popen3(cmd, env=None, newlines=False):
372 372 stdin, stdout, stderr, p = popen4(cmd, env, newlines)
373 373 return stdin, stdout, stderr
374 374
375 375 def popen4(cmd, env=None, newlines=False, bufsize=-1):
376 376 p = subprocess.Popen(cmd, shell=True, bufsize=bufsize,
377 377 close_fds=closefds,
378 378 stdin=subprocess.PIPE, stdout=subprocess.PIPE,
379 379 stderr=subprocess.PIPE,
380 380 universal_newlines=newlines,
381 381 env=env)
382 382 return p.stdin, p.stdout, p.stderr, p
383 383
384 384 def version():
385 385 """Return version information if available."""
386 386 try:
387 387 from . import __version__
388 388 return __version__.version
389 389 except ImportError:
390 390 return 'unknown'
391 391
392 392 def versiontuple(v=None, n=4):
393 393 """Parses a Mercurial version string into an N-tuple.
394 394
395 395 The version string to be parsed is specified with the ``v`` argument.
396 396 If it isn't defined, the current Mercurial version string will be parsed.
397 397
398 398 ``n`` can be 2, 3, or 4. Here is how some version strings map to
399 399 returned values:
400 400
401 401 >>> v = '3.6.1+190-df9b73d2d444'
402 402 >>> versiontuple(v, 2)
403 403 (3, 6)
404 404 >>> versiontuple(v, 3)
405 405 (3, 6, 1)
406 406 >>> versiontuple(v, 4)
407 407 (3, 6, 1, '190-df9b73d2d444')
408 408
409 409 >>> versiontuple('3.6.1+190-df9b73d2d444+20151118')
410 410 (3, 6, 1, '190-df9b73d2d444+20151118')
411 411
412 412 >>> v = '3.6'
413 413 >>> versiontuple(v, 2)
414 414 (3, 6)
415 415 >>> versiontuple(v, 3)
416 416 (3, 6, None)
417 417 >>> versiontuple(v, 4)
418 418 (3, 6, None, None)
419 419
420 420 >>> v = '3.9-rc'
421 421 >>> versiontuple(v, 2)
422 422 (3, 9)
423 423 >>> versiontuple(v, 3)
424 424 (3, 9, None)
425 425 >>> versiontuple(v, 4)
426 426 (3, 9, None, 'rc')
427 427
428 428 >>> v = '3.9-rc+2-02a8fea4289b'
429 429 >>> versiontuple(v, 2)
430 430 (3, 9)
431 431 >>> versiontuple(v, 3)
432 432 (3, 9, None)
433 433 >>> versiontuple(v, 4)
434 434 (3, 9, None, 'rc+2-02a8fea4289b')
435 435 """
436 436 if not v:
437 437 v = version()
438 438 parts = remod.split('[\+-]', v, 1)
439 439 if len(parts) == 1:
440 440 vparts, extra = parts[0], None
441 441 else:
442 442 vparts, extra = parts
443 443
444 444 vints = []
445 445 for i in vparts.split('.'):
446 446 try:
447 447 vints.append(int(i))
448 448 except ValueError:
449 449 break
450 450 # (3, 6) -> (3, 6, None)
451 451 while len(vints) < 3:
452 452 vints.append(None)
453 453
454 454 if n == 2:
455 455 return (vints[0], vints[1])
456 456 if n == 3:
457 457 return (vints[0], vints[1], vints[2])
458 458 if n == 4:
459 459 return (vints[0], vints[1], vints[2], extra)
460 460
461 461 # used by parsedate
462 462 defaultdateformats = (
463 463 '%Y-%m-%dT%H:%M:%S', # the 'real' ISO8601
464 464 '%Y-%m-%dT%H:%M', # without seconds
465 465 '%Y-%m-%dT%H%M%S', # another awful but legal variant without :
466 466 '%Y-%m-%dT%H%M', # without seconds
467 467 '%Y-%m-%d %H:%M:%S', # our common legal variant
468 468 '%Y-%m-%d %H:%M', # without seconds
469 469 '%Y-%m-%d %H%M%S', # without :
470 470 '%Y-%m-%d %H%M', # without seconds
471 471 '%Y-%m-%d %I:%M:%S%p',
472 472 '%Y-%m-%d %H:%M',
473 473 '%Y-%m-%d %I:%M%p',
474 474 '%Y-%m-%d',
475 475 '%m-%d',
476 476 '%m/%d',
477 477 '%m/%d/%y',
478 478 '%m/%d/%Y',
479 479 '%a %b %d %H:%M:%S %Y',
480 480 '%a %b %d %I:%M:%S%p %Y',
481 481 '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822"
482 482 '%b %d %H:%M:%S %Y',
483 483 '%b %d %I:%M:%S%p %Y',
484 484 '%b %d %H:%M:%S',
485 485 '%b %d %I:%M:%S%p',
486 486 '%b %d %H:%M',
487 487 '%b %d %I:%M%p',
488 488 '%b %d %Y',
489 489 '%b %d',
490 490 '%H:%M:%S',
491 491 '%I:%M:%S%p',
492 492 '%H:%M',
493 493 '%I:%M%p',
494 494 )
495 495
496 496 extendeddateformats = defaultdateformats + (
497 497 "%Y",
498 498 "%Y-%m",
499 499 "%b",
500 500 "%b %Y",
501 501 )
502 502
503 503 def cachefunc(func):
504 504 '''cache the result of function calls'''
505 505 # XXX doesn't handle keywords args
506 506 if func.__code__.co_argcount == 0:
507 507 cache = []
508 508 def f():
509 509 if len(cache) == 0:
510 510 cache.append(func())
511 511 return cache[0]
512 512 return f
513 513 cache = {}
514 514 if func.__code__.co_argcount == 1:
515 515 # we gain a small amount of time because
516 516 # we don't need to pack/unpack the list
517 517 def f(arg):
518 518 if arg not in cache:
519 519 cache[arg] = func(arg)
520 520 return cache[arg]
521 521 else:
522 522 def f(*args):
523 523 if args not in cache:
524 524 cache[args] = func(*args)
525 525 return cache[args]
526 526
527 527 return f
528 528
529 529 class sortdict(dict):
530 530 '''a simple sorted dictionary'''
531 531 def __init__(self, data=None):
532 532 self._list = []
533 533 if data:
534 534 self.update(data)
535 535 def copy(self):
536 536 return sortdict(self)
537 537 def __setitem__(self, key, val):
538 538 if key in self:
539 539 self._list.remove(key)
540 540 self._list.append(key)
541 541 dict.__setitem__(self, key, val)
542 542 def __iter__(self):
543 543 return self._list.__iter__()
544 544 def update(self, src):
545 545 if isinstance(src, dict):
546 546 src = src.iteritems()
547 547 for k, v in src:
548 548 self[k] = v
549 549 def clear(self):
550 550 dict.clear(self)
551 551 self._list = []
552 552 def items(self):
553 553 return [(k, self[k]) for k in self._list]
554 554 def __delitem__(self, key):
555 555 dict.__delitem__(self, key)
556 556 self._list.remove(key)
557 557 def pop(self, key, *args, **kwargs):
558 558 dict.pop(self, key, *args, **kwargs)
559 559 try:
560 560 self._list.remove(key)
561 561 except ValueError:
562 562 pass
563 563 def keys(self):
564 564 return self._list[:]
565 565 def iterkeys(self):
566 566 return self._list.__iter__()
567 567 def iteritems(self):
568 568 for k in self._list:
569 569 yield k, self[k]
570 570 def insert(self, index, key, val):
571 571 self._list.insert(index, key)
572 572 dict.__setitem__(self, key, val)
573 573 def __repr__(self):
574 574 if not self:
575 575 return '%s()' % self.__class__.__name__
576 576 return '%s(%r)' % (self.__class__.__name__, self.items())
577 577
578 578 class _lrucachenode(object):
579 579 """A node in a doubly linked list.
580 580
581 581 Holds a reference to nodes on either side as well as a key-value
582 582 pair for the dictionary entry.
583 583 """
584 584 __slots__ = (u'next', u'prev', u'key', u'value')
585 585
586 586 def __init__(self):
587 587 self.next = None
588 588 self.prev = None
589 589
590 590 self.key = _notset
591 591 self.value = None
592 592
593 593 def markempty(self):
594 594 """Mark the node as emptied."""
595 595 self.key = _notset
596 596
597 597 class lrucachedict(object):
598 598 """Dict that caches most recent accesses and sets.
599 599
600 600 The dict consists of an actual backing dict - indexed by original
601 601 key - and a doubly linked circular list defining the order of entries in
602 602 the cache.
603 603
604 604 The head node is the newest entry in the cache. If the cache is full,
605 605 we recycle head.prev and make it the new head. Cache accesses result in
606 606 the node being moved to before the existing head and being marked as the
607 607 new head node.
608 608 """
609 609 def __init__(self, max):
610 610 self._cache = {}
611 611
612 612 self._head = head = _lrucachenode()
613 613 head.prev = head
614 614 head.next = head
615 615 self._size = 1
616 616 self._capacity = max
617 617
618 618 def __len__(self):
619 619 return len(self._cache)
620 620
621 621 def __contains__(self, k):
622 622 return k in self._cache
623 623
624 624 def __iter__(self):
625 625 # We don't have to iterate in cache order, but why not.
626 626 n = self._head
627 627 for i in range(len(self._cache)):
628 628 yield n.key
629 629 n = n.next
630 630
631 631 def __getitem__(self, k):
632 632 node = self._cache[k]
633 633 self._movetohead(node)
634 634 return node.value
635 635
636 636 def __setitem__(self, k, v):
637 637 node = self._cache.get(k)
638 638 # Replace existing value and mark as newest.
639 639 if node is not None:
640 640 node.value = v
641 641 self._movetohead(node)
642 642 return
643 643
644 644 if self._size < self._capacity:
645 645 node = self._addcapacity()
646 646 else:
647 647 # Grab the last/oldest item.
648 648 node = self._head.prev
649 649
650 650 # At capacity. Kill the old entry.
651 651 if node.key is not _notset:
652 652 del self._cache[node.key]
653 653
654 654 node.key = k
655 655 node.value = v
656 656 self._cache[k] = node
657 657 # And mark it as newest entry. No need to adjust order since it
658 658 # is already self._head.prev.
659 659 self._head = node
660 660
661 661 def __delitem__(self, k):
662 662 node = self._cache.pop(k)
663 663 node.markempty()
664 664
665 665 # Temporarily mark as newest item before re-adjusting head to make
666 666 # this node the oldest item.
667 667 self._movetohead(node)
668 668 self._head = node.next
669 669
670 670 # Additional dict methods.
671 671
672 672 def get(self, k, default=None):
673 673 try:
674 674 return self._cache[k].value
675 675 except KeyError:
676 676 return default
677 677
678 678 def clear(self):
679 679 n = self._head
680 680 while n.key is not _notset:
681 681 n.markempty()
682 682 n = n.next
683 683
684 684 self._cache.clear()
685 685
686 686 def copy(self):
687 687 result = lrucachedict(self._capacity)
688 688 n = self._head.prev
689 689 # Iterate in oldest-to-newest order, so the copy has the right ordering
690 690 for i in range(len(self._cache)):
691 691 result[n.key] = n.value
692 692 n = n.prev
693 693 return result
694 694
695 695 def _movetohead(self, node):
696 696 """Mark a node as the newest, making it the new head.
697 697
698 698 When a node is accessed, it becomes the freshest entry in the LRU
699 699 list, which is denoted by self._head.
700 700
701 701 Visually, let's make ``N`` the new head node (* denotes head):
702 702
703 703 previous/oldest <-> head <-> next/next newest
704 704
705 705 ----<->--- A* ---<->-----
706 706 | |
707 707 E <-> D <-> N <-> C <-> B
708 708
709 709 To:
710 710
711 711 ----<->--- N* ---<->-----
712 712 | |
713 713 E <-> D <-> C <-> B <-> A
714 714
715 715 This requires the following moves:
716 716
717 717 C.next = D (node.prev.next = node.next)
718 718 D.prev = C (node.next.prev = node.prev)
719 719 E.next = N (head.prev.next = node)
720 720 N.prev = E (node.prev = head.prev)
721 721 N.next = A (node.next = head)
722 722 A.prev = N (head.prev = node)
723 723 """
724 724 head = self._head
725 725 # C.next = D
726 726 node.prev.next = node.next
727 727 # D.prev = C
728 728 node.next.prev = node.prev
729 729 # N.prev = E
730 730 node.prev = head.prev
731 731 # N.next = A
732 732 # It is tempting to do just "head" here, however if node is
733 733 # adjacent to head, this will do bad things.
734 734 node.next = head.prev.next
735 735 # E.next = N
736 736 node.next.prev = node
737 737 # A.prev = N
738 738 node.prev.next = node
739 739
740 740 self._head = node
741 741
742 742 def _addcapacity(self):
743 743 """Add a node to the circular linked list.
744 744
745 745 The new node is inserted before the head node.
746 746 """
747 747 head = self._head
748 748 node = _lrucachenode()
749 749 head.prev.next = node
750 750 node.prev = head.prev
751 751 node.next = head
752 752 head.prev = node
753 753 self._size += 1
754 754 return node
755 755
756 756 def lrucachefunc(func):
757 757 '''cache most recent results of function calls'''
758 758 cache = {}
759 759 order = collections.deque()
760 760 if func.__code__.co_argcount == 1:
761 761 def f(arg):
762 762 if arg not in cache:
763 763 if len(cache) > 20:
764 764 del cache[order.popleft()]
765 765 cache[arg] = func(arg)
766 766 else:
767 767 order.remove(arg)
768 768 order.append(arg)
769 769 return cache[arg]
770 770 else:
771 771 def f(*args):
772 772 if args not in cache:
773 773 if len(cache) > 20:
774 774 del cache[order.popleft()]
775 775 cache[args] = func(*args)
776 776 else:
777 777 order.remove(args)
778 778 order.append(args)
779 779 return cache[args]
780 780
781 781 return f
782 782
783 783 class propertycache(object):
784 784 def __init__(self, func):
785 785 self.func = func
786 786 self.name = func.__name__
787 787 def __get__(self, obj, type=None):
788 788 result = self.func(obj)
789 789 self.cachevalue(obj, result)
790 790 return result
791 791
792 792 def cachevalue(self, obj, value):
793 793 # __dict__ assignment required to bypass __setattr__ (eg: repoview)
794 794 obj.__dict__[self.name] = value
795 795
796 796 def pipefilter(s, cmd):
797 797 '''filter string S through command CMD, returning its output'''
798 798 p = subprocess.Popen(cmd, shell=True, close_fds=closefds,
799 799 stdin=subprocess.PIPE, stdout=subprocess.PIPE)
800 800 pout, perr = p.communicate(s)
801 801 return pout
802 802
803 803 def tempfilter(s, cmd):
804 804 '''filter string S through a pair of temporary files with CMD.
805 805 CMD is used as a template to create the real command to be run,
806 806 with the strings INFILE and OUTFILE replaced by the real names of
807 807 the temporary files generated.'''
808 808 inname, outname = None, None
809 809 try:
810 810 infd, inname = tempfile.mkstemp(prefix='hg-filter-in-')
811 811 fp = os.fdopen(infd, pycompat.sysstr('wb'))
812 812 fp.write(s)
813 813 fp.close()
814 814 outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-')
815 815 os.close(outfd)
816 816 cmd = cmd.replace('INFILE', inname)
817 817 cmd = cmd.replace('OUTFILE', outname)
818 818 code = os.system(cmd)
819 819 if pycompat.sysplatform == 'OpenVMS' and code & 1:
820 820 code = 0
821 821 if code:
822 822 raise Abort(_("command '%s' failed: %s") %
823 823 (cmd, explainexit(code)))
824 824 return readfile(outname)
825 825 finally:
826 826 try:
827 827 if inname:
828 828 os.unlink(inname)
829 829 except OSError:
830 830 pass
831 831 try:
832 832 if outname:
833 833 os.unlink(outname)
834 834 except OSError:
835 835 pass
836 836
837 837 filtertable = {
838 838 'tempfile:': tempfilter,
839 839 'pipe:': pipefilter,
840 840 }
841 841
842 842 def filter(s, cmd):
843 843 "filter a string through a command that transforms its input to its output"
844 844 for name, fn in filtertable.iteritems():
845 845 if cmd.startswith(name):
846 846 return fn(s, cmd[len(name):].lstrip())
847 847 return pipefilter(s, cmd)
848 848
849 849 def binary(s):
850 850 """return true if a string is binary data"""
851 851 return bool(s and '\0' in s)
852 852
853 853 def increasingchunks(source, min=1024, max=65536):
854 854 '''return no less than min bytes per chunk while data remains,
855 855 doubling min after each chunk until it reaches max'''
856 856 def log2(x):
857 857 if not x:
858 858 return 0
859 859 i = 0
860 860 while x:
861 861 x >>= 1
862 862 i += 1
863 863 return i - 1
864 864
865 865 buf = []
866 866 blen = 0
867 867 for chunk in source:
868 868 buf.append(chunk)
869 869 blen += len(chunk)
870 870 if blen >= min:
871 871 if min < max:
872 872 min = min << 1
873 873 nmin = 1 << log2(blen)
874 874 if nmin > min:
875 875 min = nmin
876 876 if min > max:
877 877 min = max
878 878 yield ''.join(buf)
879 879 blen = 0
880 880 buf = []
881 881 if buf:
882 882 yield ''.join(buf)
883 883
884 884 Abort = error.Abort
885 885
886 886 def always(fn):
887 887 return True
888 888
889 889 def never(fn):
890 890 return False
891 891
892 892 def nogc(func):
893 893 """disable garbage collector
894 894
895 895 Python's garbage collector triggers a GC each time a certain number of
896 896 container objects (the number being defined by gc.get_threshold()) are
897 897 allocated even when marked not to be tracked by the collector. Tracking has
898 898 no effect on when GCs are triggered, only on what objects the GC looks
899 899 into. As a workaround, disable GC while building complex (huge)
900 900 containers.
901 901
902 902 This garbage collector issue have been fixed in 2.7.
903 903 """
904 904 if sys.version_info >= (2, 7):
905 905 return func
906 906 def wrapper(*args, **kwargs):
907 907 gcenabled = gc.isenabled()
908 908 gc.disable()
909 909 try:
910 910 return func(*args, **kwargs)
911 911 finally:
912 912 if gcenabled:
913 913 gc.enable()
914 914 return wrapper
915 915
916 916 def pathto(root, n1, n2):
917 917 '''return the relative path from one place to another.
918 918 root should use os.sep to separate directories
919 919 n1 should use os.sep to separate directories
920 920 n2 should use "/" to separate directories
921 921 returns an os.sep-separated path.
922 922
923 923 If n1 is a relative path, it's assumed it's
924 924 relative to root.
925 925 n2 should always be relative to root.
926 926 '''
927 927 if not n1:
928 928 return localpath(n2)
929 929 if os.path.isabs(n1):
930 930 if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]:
931 931 return os.path.join(root, localpath(n2))
932 932 n2 = '/'.join((pconvert(root), n2))
933 933 a, b = splitpath(n1), n2.split('/')
934 934 a.reverse()
935 935 b.reverse()
936 936 while a and b and a[-1] == b[-1]:
937 937 a.pop()
938 938 b.pop()
939 939 b.reverse()
940 940 return pycompat.ossep.join((['..'] * len(a)) + b) or '.'
941 941
942 942 def mainfrozen():
943 943 """return True if we are a frozen executable.
944 944
945 945 The code supports py2exe (most common, Windows only) and tools/freeze
946 946 (portable, not much used).
947 947 """
948 948 return (safehasattr(sys, "frozen") or # new py2exe
949 949 safehasattr(sys, "importers") or # old py2exe
950 950 imp.is_frozen(u"__main__")) # tools/freeze
951 951
952 952 # the location of data files matching the source code
953 953 if mainfrozen() and getattr(sys, 'frozen', None) != 'macosx_app':
954 954 # executable version (py2exe) doesn't support __file__
955 955 datapath = os.path.dirname(pycompat.sysexecutable)
956 956 else:
957 957 datapath = os.path.dirname(pycompat.fsencode(__file__))
958 958
959 959 i18n.setdatapath(datapath)
960 960
961 961 _hgexecutable = None
962 962
963 963 def hgexecutable():
964 964 """return location of the 'hg' executable.
965 965
966 966 Defaults to $HG or 'hg' in the search path.
967 967 """
968 968 if _hgexecutable is None:
969 969 hg = encoding.environ.get('HG')
970 970 mainmod = sys.modules[pycompat.sysstr('__main__')]
971 971 if hg:
972 972 _sethgexecutable(hg)
973 973 elif mainfrozen():
974 974 if getattr(sys, 'frozen', None) == 'macosx_app':
975 975 # Env variable set by py2app
976 976 _sethgexecutable(encoding.environ['EXECUTABLEPATH'])
977 977 else:
978 978 _sethgexecutable(pycompat.sysexecutable)
979 979 elif (os.path.basename(
980 980 pycompat.fsencode(getattr(mainmod, '__file__', ''))) == 'hg'):
981 981 _sethgexecutable(pycompat.fsencode(mainmod.__file__))
982 982 else:
983 983 exe = findexe('hg') or os.path.basename(sys.argv[0])
984 984 _sethgexecutable(exe)
985 985 return _hgexecutable
986 986
987 987 def _sethgexecutable(path):
988 988 """set location of the 'hg' executable"""
989 989 global _hgexecutable
990 990 _hgexecutable = path
991 991
992 992 def _isstdout(f):
993 993 fileno = getattr(f, 'fileno', None)
994 994 return fileno and fileno() == sys.__stdout__.fileno()
995 995
996 996 def shellenviron(environ=None):
997 997 """return environ with optional override, useful for shelling out"""
998 998 def py2shell(val):
999 999 'convert python object into string that is useful to shell'
1000 1000 if val is None or val is False:
1001 1001 return '0'
1002 1002 if val is True:
1003 1003 return '1'
1004 1004 return str(val)
1005 1005 env = dict(encoding.environ)
1006 1006 if environ:
1007 1007 env.update((k, py2shell(v)) for k, v in environ.iteritems())
1008 1008 env['HG'] = hgexecutable()
1009 1009 return env
1010 1010
1011 1011 def system(cmd, environ=None, cwd=None, out=None):
1012 1012 '''enhanced shell command execution.
1013 1013 run with environment maybe modified, maybe in different dir.
1014 1014
1015 1015 if out is specified, it is assumed to be a file-like object that has a
1016 1016 write() method. stdout and stderr will be redirected to out.'''
1017 1017 try:
1018 1018 stdout.flush()
1019 1019 except Exception:
1020 1020 pass
1021 1021 cmd = quotecommand(cmd)
1022 1022 if pycompat.sysplatform == 'plan9' and (sys.version_info[0] == 2
1023 1023 and sys.version_info[1] < 7):
1024 1024 # subprocess kludge to work around issues in half-baked Python
1025 1025 # ports, notably bichued/python:
1026 1026 if not cwd is None:
1027 1027 os.chdir(cwd)
1028 1028 rc = os.system(cmd)
1029 1029 else:
1030 1030 env = shellenviron(environ)
1031 1031 if out is None or _isstdout(out):
1032 1032 rc = subprocess.call(cmd, shell=True, close_fds=closefds,
1033 1033 env=env, cwd=cwd)
1034 1034 else:
1035 1035 proc = subprocess.Popen(cmd, shell=True, close_fds=closefds,
1036 1036 env=env, cwd=cwd, stdout=subprocess.PIPE,
1037 1037 stderr=subprocess.STDOUT)
1038 1038 for line in iter(proc.stdout.readline, ''):
1039 1039 out.write(line)
1040 1040 proc.wait()
1041 1041 rc = proc.returncode
1042 1042 if pycompat.sysplatform == 'OpenVMS' and rc & 1:
1043 1043 rc = 0
1044 1044 return rc
1045 1045
1046 1046 def checksignature(func):
1047 1047 '''wrap a function with code to check for calling errors'''
1048 1048 def check(*args, **kwargs):
1049 1049 try:
1050 1050 return func(*args, **kwargs)
1051 1051 except TypeError:
1052 1052 if len(traceback.extract_tb(sys.exc_info()[2])) == 1:
1053 1053 raise error.SignatureError
1054 1054 raise
1055 1055
1056 1056 return check
1057 1057
1058 1058 # a whilelist of known filesystems where hardlink works reliably
1059 1059 _hardlinkfswhitelist = set([
1060 1060 'btrfs',
1061 1061 'ext2',
1062 1062 'ext3',
1063 1063 'ext4',
1064 1064 'hfs',
1065 1065 'jfs',
1066 1066 'reiserfs',
1067 1067 'tmpfs',
1068 1068 'ufs',
1069 1069 'xfs',
1070 1070 'zfs',
1071 1071 ])
1072 1072
1073 1073 def copyfile(src, dest, hardlink=False, copystat=False, checkambig=False):
1074 1074 '''copy a file, preserving mode and optionally other stat info like
1075 1075 atime/mtime
1076 1076
1077 1077 checkambig argument is used with filestat, and is useful only if
1078 1078 destination file is guarded by any lock (e.g. repo.lock or
1079 1079 repo.wlock).
1080 1080
1081 1081 copystat and checkambig should be exclusive.
1082 1082 '''
1083 1083 assert not (copystat and checkambig)
1084 1084 oldstat = None
1085 1085 if os.path.lexists(dest):
1086 1086 if checkambig:
1087 1087 oldstat = checkambig and filestat(dest)
1088 1088 unlink(dest)
1089 1089 if hardlink:
1090 1090 # Hardlinks are problematic on CIFS (issue4546), do not allow hardlinks
1091 1091 # unless we are confident that dest is on a whitelisted filesystem.
1092 1092 try:
1093 1093 fstype = getfstype(os.path.dirname(dest))
1094 1094 except OSError:
1095 1095 fstype = None
1096 1096 if fstype not in _hardlinkfswhitelist:
1097 1097 hardlink = False
1098 1098 if hardlink:
1099 1099 try:
1100 1100 oslink(src, dest)
1101 1101 return
1102 1102 except (IOError, OSError):
1103 1103 pass # fall back to normal copy
1104 1104 if os.path.islink(src):
1105 1105 os.symlink(os.readlink(src), dest)
1106 1106 # copytime is ignored for symlinks, but in general copytime isn't needed
1107 1107 # for them anyway
1108 1108 else:
1109 1109 try:
1110 1110 shutil.copyfile(src, dest)
1111 1111 if copystat:
1112 1112 # copystat also copies mode
1113 1113 shutil.copystat(src, dest)
1114 1114 else:
1115 1115 shutil.copymode(src, dest)
1116 1116 if oldstat and oldstat.stat:
1117 1117 newstat = filestat(dest)
1118 1118 if newstat.isambig(oldstat):
1119 1119 # stat of copied file is ambiguous to original one
1120 1120 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1121 1121 os.utime(dest, (advanced, advanced))
1122 1122 except shutil.Error as inst:
1123 1123 raise Abort(str(inst))
1124 1124
1125 1125 def copyfiles(src, dst, hardlink=None, progress=lambda t, pos: None):
1126 1126 """Copy a directory tree using hardlinks if possible."""
1127 1127 num = 0
1128 1128
1129 1129 gettopic = lambda: hardlink and _('linking') or _('copying')
1130 1130
1131 1131 if os.path.isdir(src):
1132 1132 if hardlink is None:
1133 1133 hardlink = (os.stat(src).st_dev ==
1134 1134 os.stat(os.path.dirname(dst)).st_dev)
1135 1135 topic = gettopic()
1136 1136 os.mkdir(dst)
1137 1137 for name, kind in osutil.listdir(src):
1138 1138 srcname = os.path.join(src, name)
1139 1139 dstname = os.path.join(dst, name)
1140 1140 def nprog(t, pos):
1141 1141 if pos is not None:
1142 1142 return progress(t, pos + num)
1143 1143 hardlink, n = copyfiles(srcname, dstname, hardlink, progress=nprog)
1144 1144 num += n
1145 1145 else:
1146 1146 if hardlink is None:
1147 1147 hardlink = (os.stat(os.path.dirname(src)).st_dev ==
1148 1148 os.stat(os.path.dirname(dst)).st_dev)
1149 1149 topic = gettopic()
1150 1150
1151 1151 if hardlink:
1152 1152 try:
1153 1153 oslink(src, dst)
1154 1154 except (IOError, OSError):
1155 1155 hardlink = False
1156 1156 shutil.copy(src, dst)
1157 1157 else:
1158 1158 shutil.copy(src, dst)
1159 1159 num += 1
1160 1160 progress(topic, num)
1161 1161 progress(topic, None)
1162 1162
1163 1163 return hardlink, num
1164 1164
1165 1165 _winreservednames = '''con prn aux nul
1166 1166 com1 com2 com3 com4 com5 com6 com7 com8 com9
1167 1167 lpt1 lpt2 lpt3 lpt4 lpt5 lpt6 lpt7 lpt8 lpt9'''.split()
1168 1168 _winreservedchars = ':*?"<>|'
1169 1169 def checkwinfilename(path):
1170 1170 r'''Check that the base-relative path is a valid filename on Windows.
1171 1171 Returns None if the path is ok, or a UI string describing the problem.
1172 1172
1173 1173 >>> checkwinfilename("just/a/normal/path")
1174 1174 >>> checkwinfilename("foo/bar/con.xml")
1175 1175 "filename contains 'con', which is reserved on Windows"
1176 1176 >>> checkwinfilename("foo/con.xml/bar")
1177 1177 "filename contains 'con', which is reserved on Windows"
1178 1178 >>> checkwinfilename("foo/bar/xml.con")
1179 1179 >>> checkwinfilename("foo/bar/AUX/bla.txt")
1180 1180 "filename contains 'AUX', which is reserved on Windows"
1181 1181 >>> checkwinfilename("foo/bar/bla:.txt")
1182 1182 "filename contains ':', which is reserved on Windows"
1183 1183 >>> checkwinfilename("foo/bar/b\07la.txt")
1184 1184 "filename contains '\\x07', which is invalid on Windows"
1185 1185 >>> checkwinfilename("foo/bar/bla ")
1186 1186 "filename ends with ' ', which is not allowed on Windows"
1187 1187 >>> checkwinfilename("../bar")
1188 1188 >>> checkwinfilename("foo\\")
1189 1189 "filename ends with '\\', which is invalid on Windows"
1190 1190 >>> checkwinfilename("foo\\/bar")
1191 1191 "directory name ends with '\\', which is invalid on Windows"
1192 1192 '''
1193 1193 if path.endswith('\\'):
1194 1194 return _("filename ends with '\\', which is invalid on Windows")
1195 1195 if '\\/' in path:
1196 1196 return _("directory name ends with '\\', which is invalid on Windows")
1197 1197 for n in path.replace('\\', '/').split('/'):
1198 1198 if not n:
1199 1199 continue
1200 1200 for c in pycompat.bytestr(n):
1201 1201 if c in _winreservedchars:
1202 1202 return _("filename contains '%s', which is reserved "
1203 1203 "on Windows") % c
1204 1204 if ord(c) <= 31:
1205 1205 return _("filename contains %r, which is invalid "
1206 1206 "on Windows") % c
1207 1207 base = n.split('.')[0]
1208 1208 if base and base.lower() in _winreservednames:
1209 1209 return _("filename contains '%s', which is reserved "
1210 1210 "on Windows") % base
1211 1211 t = n[-1]
1212 1212 if t in '. ' and n not in '..':
1213 1213 return _("filename ends with '%s', which is not allowed "
1214 1214 "on Windows") % t
1215 1215
1216 1216 if pycompat.osname == 'nt':
1217 1217 checkosfilename = checkwinfilename
1218 1218 timer = time.clock
1219 1219 else:
1220 1220 checkosfilename = platform.checkosfilename
1221 1221 timer = time.time
1222 1222
1223 1223 if safehasattr(time, "perf_counter"):
1224 1224 timer = time.perf_counter
1225 1225
1226 1226 def makelock(info, pathname):
1227 1227 try:
1228 1228 return os.symlink(info, pathname)
1229 1229 except OSError as why:
1230 1230 if why.errno == errno.EEXIST:
1231 1231 raise
1232 1232 except AttributeError: # no symlink in os
1233 1233 pass
1234 1234
1235 1235 ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL)
1236 1236 os.write(ld, info)
1237 1237 os.close(ld)
1238 1238
1239 1239 def readlock(pathname):
1240 1240 try:
1241 1241 return os.readlink(pathname)
1242 1242 except OSError as why:
1243 1243 if why.errno not in (errno.EINVAL, errno.ENOSYS):
1244 1244 raise
1245 1245 except AttributeError: # no symlink in os
1246 1246 pass
1247 1247 fp = posixfile(pathname)
1248 1248 r = fp.read()
1249 1249 fp.close()
1250 1250 return r
1251 1251
1252 1252 def fstat(fp):
1253 1253 '''stat file object that may not have fileno method.'''
1254 1254 try:
1255 1255 return os.fstat(fp.fileno())
1256 1256 except AttributeError:
1257 1257 return os.stat(fp.name)
1258 1258
1259 1259 # File system features
1260 1260
1261 1261 def fscasesensitive(path):
1262 1262 """
1263 1263 Return true if the given path is on a case-sensitive filesystem
1264 1264
1265 1265 Requires a path (like /foo/.hg) ending with a foldable final
1266 1266 directory component.
1267 1267 """
1268 1268 s1 = os.lstat(path)
1269 1269 d, b = os.path.split(path)
1270 1270 b2 = b.upper()
1271 1271 if b == b2:
1272 1272 b2 = b.lower()
1273 1273 if b == b2:
1274 1274 return True # no evidence against case sensitivity
1275 1275 p2 = os.path.join(d, b2)
1276 1276 try:
1277 1277 s2 = os.lstat(p2)
1278 1278 if s2 == s1:
1279 1279 return False
1280 1280 return True
1281 1281 except OSError:
1282 1282 return True
1283 1283
1284 1284 try:
1285 1285 import re2
1286 1286 _re2 = None
1287 1287 except ImportError:
1288 1288 _re2 = False
1289 1289
1290 1290 class _re(object):
1291 1291 def _checkre2(self):
1292 1292 global _re2
1293 1293 try:
1294 1294 # check if match works, see issue3964
1295 1295 _re2 = bool(re2.match(r'\[([^\[]+)\]', '[ui]'))
1296 1296 except ImportError:
1297 1297 _re2 = False
1298 1298
1299 1299 def compile(self, pat, flags=0):
1300 1300 '''Compile a regular expression, using re2 if possible
1301 1301
1302 1302 For best performance, use only re2-compatible regexp features. The
1303 1303 only flags from the re module that are re2-compatible are
1304 1304 IGNORECASE and MULTILINE.'''
1305 1305 if _re2 is None:
1306 1306 self._checkre2()
1307 1307 if _re2 and (flags & ~(remod.IGNORECASE | remod.MULTILINE)) == 0:
1308 1308 if flags & remod.IGNORECASE:
1309 1309 pat = '(?i)' + pat
1310 1310 if flags & remod.MULTILINE:
1311 1311 pat = '(?m)' + pat
1312 1312 try:
1313 1313 return re2.compile(pat)
1314 1314 except re2.error:
1315 1315 pass
1316 1316 return remod.compile(pat, flags)
1317 1317
1318 1318 @propertycache
1319 1319 def escape(self):
1320 1320 '''Return the version of escape corresponding to self.compile.
1321 1321
1322 1322 This is imperfect because whether re2 or re is used for a particular
1323 1323 function depends on the flags, etc, but it's the best we can do.
1324 1324 '''
1325 1325 global _re2
1326 1326 if _re2 is None:
1327 1327 self._checkre2()
1328 1328 if _re2:
1329 1329 return re2.escape
1330 1330 else:
1331 1331 return remod.escape
1332 1332
1333 1333 re = _re()
1334 1334
1335 1335 _fspathcache = {}
1336 1336 def fspath(name, root):
1337 1337 '''Get name in the case stored in the filesystem
1338 1338
1339 1339 The name should be relative to root, and be normcase-ed for efficiency.
1340 1340
1341 1341 Note that this function is unnecessary, and should not be
1342 1342 called, for case-sensitive filesystems (simply because it's expensive).
1343 1343
1344 1344 The root should be normcase-ed, too.
1345 1345 '''
1346 1346 def _makefspathcacheentry(dir):
1347 1347 return dict((normcase(n), n) for n in os.listdir(dir))
1348 1348
1349 1349 seps = pycompat.ossep
1350 1350 if pycompat.osaltsep:
1351 1351 seps = seps + pycompat.osaltsep
1352 1352 # Protect backslashes. This gets silly very quickly.
1353 1353 seps.replace('\\','\\\\')
1354 1354 pattern = remod.compile(br'([^%s]+)|([%s]+)' % (seps, seps))
1355 1355 dir = os.path.normpath(root)
1356 1356 result = []
1357 1357 for part, sep in pattern.findall(name):
1358 1358 if sep:
1359 1359 result.append(sep)
1360 1360 continue
1361 1361
1362 1362 if dir not in _fspathcache:
1363 1363 _fspathcache[dir] = _makefspathcacheentry(dir)
1364 1364 contents = _fspathcache[dir]
1365 1365
1366 1366 found = contents.get(part)
1367 1367 if not found:
1368 1368 # retry "once per directory" per "dirstate.walk" which
1369 1369 # may take place for each patches of "hg qpush", for example
1370 1370 _fspathcache[dir] = contents = _makefspathcacheentry(dir)
1371 1371 found = contents.get(part)
1372 1372
1373 1373 result.append(found or part)
1374 1374 dir = os.path.join(dir, part)
1375 1375
1376 1376 return ''.join(result)
1377 1377
1378 1378 def getfstype(dirpath):
1379 1379 '''Get the filesystem type name from a directory (best-effort)
1380 1380
1381 1381 Returns None if we are unsure. Raises OSError on ENOENT, EPERM, etc.
1382 1382 '''
1383 1383 return getattr(osutil, 'getfstype', lambda x: None)(dirpath)
1384 1384
1385 1385 def checknlink(testfile):
1386 1386 '''check whether hardlink count reporting works properly'''
1387 1387
1388 1388 # testfile may be open, so we need a separate file for checking to
1389 1389 # work around issue2543 (or testfile may get lost on Samba shares)
1390 1390 f1 = testfile + ".hgtmp1"
1391 1391 if os.path.lexists(f1):
1392 1392 return False
1393 1393 try:
1394 1394 posixfile(f1, 'w').close()
1395 1395 except IOError:
1396 1396 try:
1397 1397 os.unlink(f1)
1398 1398 except OSError:
1399 1399 pass
1400 1400 return False
1401 1401
1402 1402 f2 = testfile + ".hgtmp2"
1403 1403 fd = None
1404 1404 try:
1405 1405 oslink(f1, f2)
1406 1406 # nlinks() may behave differently for files on Windows shares if
1407 1407 # the file is open.
1408 1408 fd = posixfile(f2)
1409 1409 return nlinks(f2) > 1
1410 1410 except OSError:
1411 1411 return False
1412 1412 finally:
1413 1413 if fd is not None:
1414 1414 fd.close()
1415 1415 for f in (f1, f2):
1416 1416 try:
1417 1417 os.unlink(f)
1418 1418 except OSError:
1419 1419 pass
1420 1420
1421 1421 def endswithsep(path):
1422 1422 '''Check path ends with os.sep or os.altsep.'''
1423 1423 return (path.endswith(pycompat.ossep)
1424 1424 or pycompat.osaltsep and path.endswith(pycompat.osaltsep))
1425 1425
1426 1426 def splitpath(path):
1427 1427 '''Split path by os.sep.
1428 1428 Note that this function does not use os.altsep because this is
1429 1429 an alternative of simple "xxx.split(os.sep)".
1430 1430 It is recommended to use os.path.normpath() before using this
1431 1431 function if need.'''
1432 1432 return path.split(pycompat.ossep)
1433 1433
1434 1434 def gui():
1435 1435 '''Are we running in a GUI?'''
1436 1436 if pycompat.sysplatform == 'darwin':
1437 1437 if 'SSH_CONNECTION' in encoding.environ:
1438 1438 # handle SSH access to a box where the user is logged in
1439 1439 return False
1440 1440 elif getattr(osutil, 'isgui', None):
1441 1441 # check if a CoreGraphics session is available
1442 1442 return osutil.isgui()
1443 1443 else:
1444 1444 # pure build; use a safe default
1445 1445 return True
1446 1446 else:
1447 1447 return pycompat.osname == "nt" or encoding.environ.get("DISPLAY")
1448 1448
1449 1449 def mktempcopy(name, emptyok=False, createmode=None):
1450 1450 """Create a temporary file with the same contents from name
1451 1451
1452 1452 The permission bits are copied from the original file.
1453 1453
1454 1454 If the temporary file is going to be truncated immediately, you
1455 1455 can use emptyok=True as an optimization.
1456 1456
1457 1457 Returns the name of the temporary file.
1458 1458 """
1459 1459 d, fn = os.path.split(name)
1460 1460 fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d)
1461 1461 os.close(fd)
1462 1462 # Temporary files are created with mode 0600, which is usually not
1463 1463 # what we want. If the original file already exists, just copy
1464 1464 # its mode. Otherwise, manually obey umask.
1465 1465 copymode(name, temp, createmode)
1466 1466 if emptyok:
1467 1467 return temp
1468 1468 try:
1469 1469 try:
1470 1470 ifp = posixfile(name, "rb")
1471 1471 except IOError as inst:
1472 1472 if inst.errno == errno.ENOENT:
1473 1473 return temp
1474 1474 if not getattr(inst, 'filename', None):
1475 1475 inst.filename = name
1476 1476 raise
1477 1477 ofp = posixfile(temp, "wb")
1478 1478 for chunk in filechunkiter(ifp):
1479 1479 ofp.write(chunk)
1480 1480 ifp.close()
1481 1481 ofp.close()
1482 1482 except: # re-raises
1483 1483 try: os.unlink(temp)
1484 1484 except OSError: pass
1485 1485 raise
1486 1486 return temp
1487 1487
1488 1488 class filestat(object):
1489 1489 """help to exactly detect change of a file
1490 1490
1491 1491 'stat' attribute is result of 'os.stat()' if specified 'path'
1492 1492 exists. Otherwise, it is None. This can avoid preparative
1493 1493 'exists()' examination on client side of this class.
1494 1494 """
1495 1495 def __init__(self, path):
1496 1496 try:
1497 1497 self.stat = os.stat(path)
1498 1498 except OSError as err:
1499 1499 if err.errno != errno.ENOENT:
1500 1500 raise
1501 1501 self.stat = None
1502 1502
1503 1503 __hash__ = object.__hash__
1504 1504
1505 1505 def __eq__(self, old):
1506 1506 try:
1507 1507 # if ambiguity between stat of new and old file is
1508 1508 # avoided, comparison of size, ctime and mtime is enough
1509 1509 # to exactly detect change of a file regardless of platform
1510 1510 return (self.stat.st_size == old.stat.st_size and
1511 1511 self.stat.st_ctime == old.stat.st_ctime and
1512 1512 self.stat.st_mtime == old.stat.st_mtime)
1513 1513 except AttributeError:
1514 1514 return False
1515 1515
1516 1516 def isambig(self, old):
1517 1517 """Examine whether new (= self) stat is ambiguous against old one
1518 1518
1519 1519 "S[N]" below means stat of a file at N-th change:
1520 1520
1521 1521 - S[n-1].ctime < S[n].ctime: can detect change of a file
1522 1522 - S[n-1].ctime == S[n].ctime
1523 1523 - S[n-1].ctime < S[n].mtime: means natural advancing (*1)
1524 1524 - S[n-1].ctime == S[n].mtime: is ambiguous (*2)
1525 1525 - S[n-1].ctime > S[n].mtime: never occurs naturally (don't care)
1526 1526 - S[n-1].ctime > S[n].ctime: never occurs naturally (don't care)
1527 1527
1528 1528 Case (*2) above means that a file was changed twice or more at
1529 1529 same time in sec (= S[n-1].ctime), and comparison of timestamp
1530 1530 is ambiguous.
1531 1531
1532 1532 Base idea to avoid such ambiguity is "advance mtime 1 sec, if
1533 1533 timestamp is ambiguous".
1534 1534
1535 1535 But advancing mtime only in case (*2) doesn't work as
1536 1536 expected, because naturally advanced S[n].mtime in case (*1)
1537 1537 might be equal to manually advanced S[n-1 or earlier].mtime.
1538 1538
1539 1539 Therefore, all "S[n-1].ctime == S[n].ctime" cases should be
1540 1540 treated as ambiguous regardless of mtime, to avoid overlooking
1541 1541 by confliction between such mtime.
1542 1542
1543 1543 Advancing mtime "if isambig(oldstat)" ensures "S[n-1].mtime !=
1544 1544 S[n].mtime", even if size of a file isn't changed.
1545 1545 """
1546 1546 try:
1547 1547 return (self.stat.st_ctime == old.stat.st_ctime)
1548 1548 except AttributeError:
1549 1549 return False
1550 1550
1551 1551 def avoidambig(self, path, old):
1552 1552 """Change file stat of specified path to avoid ambiguity
1553 1553
1554 1554 'old' should be previous filestat of 'path'.
1555 1555
1556 1556 This skips avoiding ambiguity, if a process doesn't have
1557 1557 appropriate privileges for 'path'.
1558 1558 """
1559 1559 advanced = (old.stat.st_mtime + 1) & 0x7fffffff
1560 1560 try:
1561 1561 os.utime(path, (advanced, advanced))
1562 1562 except OSError as inst:
1563 1563 if inst.errno == errno.EPERM:
1564 1564 # utime() on the file created by another user causes EPERM,
1565 1565 # if a process doesn't have appropriate privileges
1566 1566 return
1567 1567 raise
1568 1568
1569 1569 def __ne__(self, other):
1570 1570 return not self == other
1571 1571
1572 1572 class atomictempfile(object):
1573 1573 '''writable file object that atomically updates a file
1574 1574
1575 1575 All writes will go to a temporary copy of the original file. Call
1576 1576 close() when you are done writing, and atomictempfile will rename
1577 1577 the temporary copy to the original name, making the changes
1578 1578 visible. If the object is destroyed without being closed, all your
1579 1579 writes are discarded.
1580 1580
1581 1581 checkambig argument of constructor is used with filestat, and is
1582 1582 useful only if target file is guarded by any lock (e.g. repo.lock
1583 1583 or repo.wlock).
1584 1584 '''
1585 1585 def __init__(self, name, mode='w+b', createmode=None, checkambig=False):
1586 1586 self.__name = name # permanent name
1587 1587 self._tempname = mktempcopy(name, emptyok=('w' in mode),
1588 1588 createmode=createmode)
1589 1589 self._fp = posixfile(self._tempname, mode)
1590 1590 self._checkambig = checkambig
1591 1591
1592 1592 # delegated methods
1593 1593 self.read = self._fp.read
1594 1594 self.write = self._fp.write
1595 1595 self.seek = self._fp.seek
1596 1596 self.tell = self._fp.tell
1597 1597 self.fileno = self._fp.fileno
1598 1598
1599 1599 def close(self):
1600 1600 if not self._fp.closed:
1601 1601 self._fp.close()
1602 1602 filename = localpath(self.__name)
1603 1603 oldstat = self._checkambig and filestat(filename)
1604 1604 if oldstat and oldstat.stat:
1605 1605 rename(self._tempname, filename)
1606 1606 newstat = filestat(filename)
1607 1607 if newstat.isambig(oldstat):
1608 1608 # stat of changed file is ambiguous to original one
1609 1609 advanced = (oldstat.stat.st_mtime + 1) & 0x7fffffff
1610 1610 os.utime(filename, (advanced, advanced))
1611 1611 else:
1612 1612 rename(self._tempname, filename)
1613 1613
1614 1614 def discard(self):
1615 1615 if not self._fp.closed:
1616 1616 try:
1617 1617 os.unlink(self._tempname)
1618 1618 except OSError:
1619 1619 pass
1620 1620 self._fp.close()
1621 1621
1622 1622 def __del__(self):
1623 1623 if safehasattr(self, '_fp'): # constructor actually did something
1624 1624 self.discard()
1625 1625
1626 1626 def __enter__(self):
1627 1627 return self
1628 1628
1629 1629 def __exit__(self, exctype, excvalue, traceback):
1630 1630 if exctype is not None:
1631 1631 self.discard()
1632 1632 else:
1633 1633 self.close()
1634 1634
1635 1635 def unlinkpath(f, ignoremissing=False):
1636 1636 """unlink and remove the directory if it is empty"""
1637 1637 if ignoremissing:
1638 1638 tryunlink(f)
1639 1639 else:
1640 1640 unlink(f)
1641 1641 # try removing directories that might now be empty
1642 1642 try:
1643 1643 removedirs(os.path.dirname(f))
1644 1644 except OSError:
1645 1645 pass
1646 1646
1647 1647 def tryunlink(f):
1648 1648 """Attempt to remove a file, ignoring ENOENT errors."""
1649 1649 try:
1650 1650 unlink(f)
1651 1651 except OSError as e:
1652 1652 if e.errno != errno.ENOENT:
1653 1653 raise
1654 1654
1655 1655 def makedirs(name, mode=None, notindexed=False):
1656 1656 """recursive directory creation with parent mode inheritance
1657 1657
1658 1658 Newly created directories are marked as "not to be indexed by
1659 1659 the content indexing service", if ``notindexed`` is specified
1660 1660 for "write" mode access.
1661 1661 """
1662 1662 try:
1663 1663 makedir(name, notindexed)
1664 1664 except OSError as err:
1665 1665 if err.errno == errno.EEXIST:
1666 1666 return
1667 1667 if err.errno != errno.ENOENT or not name:
1668 1668 raise
1669 1669 parent = os.path.dirname(os.path.abspath(name))
1670 1670 if parent == name:
1671 1671 raise
1672 1672 makedirs(parent, mode, notindexed)
1673 1673 try:
1674 1674 makedir(name, notindexed)
1675 1675 except OSError as err:
1676 1676 # Catch EEXIST to handle races
1677 1677 if err.errno == errno.EEXIST:
1678 1678 return
1679 1679 raise
1680 1680 if mode is not None:
1681 1681 os.chmod(name, mode)
1682 1682
1683 1683 def readfile(path):
1684 1684 with open(path, 'rb') as fp:
1685 1685 return fp.read()
1686 1686
1687 1687 def writefile(path, text):
1688 1688 with open(path, 'wb') as fp:
1689 1689 fp.write(text)
1690 1690
1691 1691 def appendfile(path, text):
1692 1692 with open(path, 'ab') as fp:
1693 1693 fp.write(text)
1694 1694
1695 1695 class chunkbuffer(object):
1696 1696 """Allow arbitrary sized chunks of data to be efficiently read from an
1697 1697 iterator over chunks of arbitrary size."""
1698 1698
1699 1699 def __init__(self, in_iter):
1700 1700 """in_iter is the iterator that's iterating over the input chunks.
1701 1701 targetsize is how big a buffer to try to maintain."""
1702 1702 def splitbig(chunks):
1703 1703 for chunk in chunks:
1704 1704 if len(chunk) > 2**20:
1705 1705 pos = 0
1706 1706 while pos < len(chunk):
1707 1707 end = pos + 2 ** 18
1708 1708 yield chunk[pos:end]
1709 1709 pos = end
1710 1710 else:
1711 1711 yield chunk
1712 1712 self.iter = splitbig(in_iter)
1713 1713 self._queue = collections.deque()
1714 1714 self._chunkoffset = 0
1715 1715
1716 1716 def read(self, l=None):
1717 1717 """Read L bytes of data from the iterator of chunks of data.
1718 1718 Returns less than L bytes if the iterator runs dry.
1719 1719
1720 1720 If size parameter is omitted, read everything"""
1721 1721 if l is None:
1722 1722 return ''.join(self.iter)
1723 1723
1724 1724 left = l
1725 1725 buf = []
1726 1726 queue = self._queue
1727 1727 while left > 0:
1728 1728 # refill the queue
1729 1729 if not queue:
1730 1730 target = 2**18
1731 1731 for chunk in self.iter:
1732 1732 queue.append(chunk)
1733 1733 target -= len(chunk)
1734 1734 if target <= 0:
1735 1735 break
1736 1736 if not queue:
1737 1737 break
1738 1738
1739 1739 # The easy way to do this would be to queue.popleft(), modify the
1740 1740 # chunk (if necessary), then queue.appendleft(). However, for cases
1741 1741 # where we read partial chunk content, this incurs 2 dequeue
1742 1742 # mutations and creates a new str for the remaining chunk in the
1743 1743 # queue. Our code below avoids this overhead.
1744 1744
1745 1745 chunk = queue[0]
1746 1746 chunkl = len(chunk)
1747 1747 offset = self._chunkoffset
1748 1748
1749 1749 # Use full chunk.
1750 1750 if offset == 0 and left >= chunkl:
1751 1751 left -= chunkl
1752 1752 queue.popleft()
1753 1753 buf.append(chunk)
1754 1754 # self._chunkoffset remains at 0.
1755 1755 continue
1756 1756
1757 1757 chunkremaining = chunkl - offset
1758 1758
1759 1759 # Use all of unconsumed part of chunk.
1760 1760 if left >= chunkremaining:
1761 1761 left -= chunkremaining
1762 1762 queue.popleft()
1763 1763 # offset == 0 is enabled by block above, so this won't merely
1764 1764 # copy via ``chunk[0:]``.
1765 1765 buf.append(chunk[offset:])
1766 1766 self._chunkoffset = 0
1767 1767
1768 1768 # Partial chunk needed.
1769 1769 else:
1770 1770 buf.append(chunk[offset:offset + left])
1771 1771 self._chunkoffset += left
1772 1772 left -= chunkremaining
1773 1773
1774 1774 return ''.join(buf)
1775 1775
1776 1776 def filechunkiter(f, size=131072, limit=None):
1777 1777 """Create a generator that produces the data in the file size
1778 1778 (default 131072) bytes at a time, up to optional limit (default is
1779 1779 to read all data). Chunks may be less than size bytes if the
1780 1780 chunk is the last chunk in the file, or the file is a socket or
1781 1781 some other type of file that sometimes reads less data than is
1782 1782 requested."""
1783 1783 assert size >= 0
1784 1784 assert limit is None or limit >= 0
1785 1785 while True:
1786 1786 if limit is None:
1787 1787 nbytes = size
1788 1788 else:
1789 1789 nbytes = min(limit, size)
1790 1790 s = nbytes and f.read(nbytes)
1791 1791 if not s:
1792 1792 break
1793 1793 if limit:
1794 1794 limit -= len(s)
1795 1795 yield s
1796 1796
1797 1797 def makedate(timestamp=None):
1798 1798 '''Return a unix timestamp (or the current time) as a (unixtime,
1799 1799 offset) tuple based off the local timezone.'''
1800 1800 if timestamp is None:
1801 1801 timestamp = time.time()
1802 1802 if timestamp < 0:
1803 1803 hint = _("check your clock")
1804 1804 raise Abort(_("negative timestamp: %d") % timestamp, hint=hint)
1805 1805 delta = (datetime.datetime.utcfromtimestamp(timestamp) -
1806 1806 datetime.datetime.fromtimestamp(timestamp))
1807 1807 tz = delta.days * 86400 + delta.seconds
1808 1808 return timestamp, tz
1809 1809
1810 1810 def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'):
1811 1811 """represent a (unixtime, offset) tuple as a localized time.
1812 1812 unixtime is seconds since the epoch, and offset is the time zone's
1813 1813 number of seconds away from UTC.
1814 1814
1815 1815 >>> datestr((0, 0))
1816 1816 'Thu Jan 01 00:00:00 1970 +0000'
1817 1817 >>> datestr((42, 0))
1818 1818 'Thu Jan 01 00:00:42 1970 +0000'
1819 1819 >>> datestr((-42, 0))
1820 1820 'Wed Dec 31 23:59:18 1969 +0000'
1821 1821 >>> datestr((0x7fffffff, 0))
1822 1822 'Tue Jan 19 03:14:07 2038 +0000'
1823 1823 >>> datestr((-0x80000000, 0))
1824 1824 'Fri Dec 13 20:45:52 1901 +0000'
1825 1825 """
1826 1826 t, tz = date or makedate()
1827 1827 if "%1" in format or "%2" in format or "%z" in format:
1828 1828 sign = (tz > 0) and "-" or "+"
1829 1829 minutes = abs(tz) // 60
1830 1830 q, r = divmod(minutes, 60)
1831 1831 format = format.replace("%z", "%1%2")
1832 1832 format = format.replace("%1", "%c%02d" % (sign, q))
1833 1833 format = format.replace("%2", "%02d" % r)
1834 1834 d = t - tz
1835 1835 if d > 0x7fffffff:
1836 1836 d = 0x7fffffff
1837 1837 elif d < -0x80000000:
1838 1838 d = -0x80000000
1839 1839 # Never use time.gmtime() and datetime.datetime.fromtimestamp()
1840 1840 # because they use the gmtime() system call which is buggy on Windows
1841 1841 # for negative values.
1842 1842 t = datetime.datetime(1970, 1, 1) + datetime.timedelta(seconds=d)
1843 1843 s = encoding.strtolocal(t.strftime(encoding.strfromlocal(format)))
1844 1844 return s
1845 1845
1846 1846 def shortdate(date=None):
1847 1847 """turn (timestamp, tzoff) tuple into iso 8631 date."""
1848 1848 return datestr(date, format='%Y-%m-%d')
1849 1849
1850 1850 def parsetimezone(s):
1851 1851 """find a trailing timezone, if any, in string, and return a
1852 1852 (offset, remainder) pair"""
1853 1853
1854 1854 if s.endswith("GMT") or s.endswith("UTC"):
1855 1855 return 0, s[:-3].rstrip()
1856 1856
1857 1857 # Unix-style timezones [+-]hhmm
1858 1858 if len(s) >= 5 and s[-5] in "+-" and s[-4:].isdigit():
1859 1859 sign = (s[-5] == "+") and 1 or -1
1860 1860 hours = int(s[-4:-2])
1861 1861 minutes = int(s[-2:])
1862 1862 return -sign * (hours * 60 + minutes) * 60, s[:-5].rstrip()
1863 1863
1864 1864 # ISO8601 trailing Z
1865 1865 if s.endswith("Z") and s[-2:-1].isdigit():
1866 1866 return 0, s[:-1]
1867 1867
1868 1868 # ISO8601-style [+-]hh:mm
1869 1869 if (len(s) >= 6 and s[-6] in "+-" and s[-3] == ":" and
1870 1870 s[-5:-3].isdigit() and s[-2:].isdigit()):
1871 1871 sign = (s[-6] == "+") and 1 or -1
1872 1872 hours = int(s[-5:-3])
1873 1873 minutes = int(s[-2:])
1874 1874 return -sign * (hours * 60 + minutes) * 60, s[:-6]
1875 1875
1876 1876 return None, s
1877 1877
1878 1878 def strdate(string, format, defaults=None):
1879 1879 """parse a localized time string and return a (unixtime, offset) tuple.
1880 1880 if the string cannot be parsed, ValueError is raised."""
1881 1881 if defaults is None:
1882 1882 defaults = {}
1883 1883
1884 1884 # NOTE: unixtime = localunixtime + offset
1885 1885 offset, date = parsetimezone(string)
1886 1886
1887 1887 # add missing elements from defaults
1888 1888 usenow = False # default to using biased defaults
1889 1889 for part in ("S", "M", "HI", "d", "mb", "yY"): # decreasing specificity
1890 1890 found = [True for p in part if ("%"+p) in format]
1891 1891 if not found:
1892 1892 date += "@" + defaults[part][usenow]
1893 1893 format += "@%" + part[0]
1894 1894 else:
1895 1895 # We've found a specific time element, less specific time
1896 1896 # elements are relative to today
1897 1897 usenow = True
1898 1898
1899 1899 timetuple = time.strptime(date, format)
1900 1900 localunixtime = int(calendar.timegm(timetuple))
1901 1901 if offset is None:
1902 1902 # local timezone
1903 1903 unixtime = int(time.mktime(timetuple))
1904 1904 offset = unixtime - localunixtime
1905 1905 else:
1906 1906 unixtime = localunixtime + offset
1907 1907 return unixtime, offset
1908 1908
1909 1909 def parsedate(date, formats=None, bias=None):
1910 1910 """parse a localized date/time and return a (unixtime, offset) tuple.
1911 1911
1912 1912 The date may be a "unixtime offset" string or in one of the specified
1913 1913 formats. If the date already is a (unixtime, offset) tuple, it is returned.
1914 1914
1915 1915 >>> parsedate(' today ') == parsedate(\
1916 1916 datetime.date.today().strftime('%b %d'))
1917 1917 True
1918 1918 >>> parsedate( 'yesterday ') == parsedate((datetime.date.today() -\
1919 1919 datetime.timedelta(days=1)\
1920 1920 ).strftime('%b %d'))
1921 1921 True
1922 1922 >>> now, tz = makedate()
1923 1923 >>> strnow, strtz = parsedate('now')
1924 1924 >>> (strnow - now) < 1
1925 1925 True
1926 1926 >>> tz == strtz
1927 1927 True
1928 1928 """
1929 1929 if bias is None:
1930 1930 bias = {}
1931 1931 if not date:
1932 1932 return 0, 0
1933 1933 if isinstance(date, tuple) and len(date) == 2:
1934 1934 return date
1935 1935 if not formats:
1936 1936 formats = defaultdateformats
1937 1937 date = date.strip()
1938 1938
1939 1939 if date == 'now' or date == _('now'):
1940 1940 return makedate()
1941 1941 if date == 'today' or date == _('today'):
1942 1942 date = datetime.date.today().strftime('%b %d')
1943 1943 elif date == 'yesterday' or date == _('yesterday'):
1944 1944 date = (datetime.date.today() -
1945 1945 datetime.timedelta(days=1)).strftime('%b %d')
1946 1946
1947 1947 try:
1948 1948 when, offset = map(int, date.split(' '))
1949 1949 except ValueError:
1950 1950 # fill out defaults
1951 1951 now = makedate()
1952 1952 defaults = {}
1953 1953 for part in ("d", "mb", "yY", "HI", "M", "S"):
1954 1954 # this piece is for rounding the specific end of unknowns
1955 1955 b = bias.get(part)
1956 1956 if b is None:
1957 1957 if part[0] in "HMS":
1958 1958 b = "00"
1959 1959 else:
1960 1960 b = "0"
1961 1961
1962 1962 # this piece is for matching the generic end to today's date
1963 1963 n = datestr(now, "%" + part[0])
1964 1964
1965 1965 defaults[part] = (b, n)
1966 1966
1967 1967 for format in formats:
1968 1968 try:
1969 1969 when, offset = strdate(date, format, defaults)
1970 1970 except (ValueError, OverflowError):
1971 1971 pass
1972 1972 else:
1973 1973 break
1974 1974 else:
1975 1975 raise Abort(_('invalid date: %r') % date)
1976 1976 # validate explicit (probably user-specified) date and
1977 1977 # time zone offset. values must fit in signed 32 bits for
1978 1978 # current 32-bit linux runtimes. timezones go from UTC-12
1979 1979 # to UTC+14
1980 1980 if when < -0x80000000 or when > 0x7fffffff:
1981 1981 raise Abort(_('date exceeds 32 bits: %d') % when)
1982 1982 if offset < -50400 or offset > 43200:
1983 1983 raise Abort(_('impossible time zone offset: %d') % offset)
1984 1984 return when, offset
1985 1985
1986 1986 def matchdate(date):
1987 1987 """Return a function that matches a given date match specifier
1988 1988
1989 1989 Formats include:
1990 1990
1991 1991 '{date}' match a given date to the accuracy provided
1992 1992
1993 1993 '<{date}' on or before a given date
1994 1994
1995 1995 '>{date}' on or after a given date
1996 1996
1997 1997 >>> p1 = parsedate("10:29:59")
1998 1998 >>> p2 = parsedate("10:30:00")
1999 1999 >>> p3 = parsedate("10:30:59")
2000 2000 >>> p4 = parsedate("10:31:00")
2001 2001 >>> p5 = parsedate("Sep 15 10:30:00 1999")
2002 2002 >>> f = matchdate("10:30")
2003 2003 >>> f(p1[0])
2004 2004 False
2005 2005 >>> f(p2[0])
2006 2006 True
2007 2007 >>> f(p3[0])
2008 2008 True
2009 2009 >>> f(p4[0])
2010 2010 False
2011 2011 >>> f(p5[0])
2012 2012 False
2013 2013 """
2014 2014
2015 2015 def lower(date):
2016 2016 d = {'mb': "1", 'd': "1"}
2017 2017 return parsedate(date, extendeddateformats, d)[0]
2018 2018
2019 2019 def upper(date):
2020 2020 d = {'mb': "12", 'HI': "23", 'M': "59", 'S': "59"}
2021 2021 for days in ("31", "30", "29"):
2022 2022 try:
2023 2023 d["d"] = days
2024 2024 return parsedate(date, extendeddateformats, d)[0]
2025 2025 except Abort:
2026 2026 pass
2027 2027 d["d"] = "28"
2028 2028 return parsedate(date, extendeddateformats, d)[0]
2029 2029
2030 2030 date = date.strip()
2031 2031
2032 2032 if not date:
2033 2033 raise Abort(_("dates cannot consist entirely of whitespace"))
2034 2034 elif date[0] == "<":
2035 2035 if not date[1:]:
2036 2036 raise Abort(_("invalid day spec, use '<DATE'"))
2037 2037 when = upper(date[1:])
2038 2038 return lambda x: x <= when
2039 2039 elif date[0] == ">":
2040 2040 if not date[1:]:
2041 2041 raise Abort(_("invalid day spec, use '>DATE'"))
2042 2042 when = lower(date[1:])
2043 2043 return lambda x: x >= when
2044 2044 elif date[0] == "-":
2045 2045 try:
2046 2046 days = int(date[1:])
2047 2047 except ValueError:
2048 2048 raise Abort(_("invalid day spec: %s") % date[1:])
2049 2049 if days < 0:
2050 2050 raise Abort(_("%s must be nonnegative (see 'hg help dates')")
2051 2051 % date[1:])
2052 2052 when = makedate()[0] - days * 3600 * 24
2053 2053 return lambda x: x >= when
2054 2054 elif " to " in date:
2055 2055 a, b = date.split(" to ")
2056 2056 start, stop = lower(a), upper(b)
2057 2057 return lambda x: x >= start and x <= stop
2058 2058 else:
2059 2059 start, stop = lower(date), upper(date)
2060 2060 return lambda x: x >= start and x <= stop
2061 2061
2062 2062 def stringmatcher(pattern, casesensitive=True):
2063 2063 """
2064 2064 accepts a string, possibly starting with 're:' or 'literal:' prefix.
2065 2065 returns the matcher name, pattern, and matcher function.
2066 2066 missing or unknown prefixes are treated as literal matches.
2067 2067
2068 2068 helper for tests:
2069 2069 >>> def test(pattern, *tests):
2070 2070 ... kind, pattern, matcher = stringmatcher(pattern)
2071 2071 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2072 2072 >>> def itest(pattern, *tests):
2073 2073 ... kind, pattern, matcher = stringmatcher(pattern, casesensitive=False)
2074 2074 ... return (kind, pattern, [bool(matcher(t)) for t in tests])
2075 2075
2076 2076 exact matching (no prefix):
2077 2077 >>> test('abcdefg', 'abc', 'def', 'abcdefg')
2078 2078 ('literal', 'abcdefg', [False, False, True])
2079 2079
2080 2080 regex matching ('re:' prefix)
2081 2081 >>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
2082 2082 ('re', 'a.+b', [False, False, True])
2083 2083
2084 2084 force exact matches ('literal:' prefix)
2085 2085 >>> test('literal:re:foobar', 'foobar', 're:foobar')
2086 2086 ('literal', 're:foobar', [False, True])
2087 2087
2088 2088 unknown prefixes are ignored and treated as literals
2089 2089 >>> test('foo:bar', 'foo', 'bar', 'foo:bar')
2090 2090 ('literal', 'foo:bar', [False, False, True])
2091 2091
2092 2092 case insensitive regex matches
2093 2093 >>> itest('re:A.+b', 'nomatch', 'fooadef', 'fooadefBar')
2094 2094 ('re', 'A.+b', [False, False, True])
2095 2095
2096 2096 case insensitive literal matches
2097 2097 >>> itest('ABCDEFG', 'abc', 'def', 'abcdefg')
2098 2098 ('literal', 'ABCDEFG', [False, False, True])
2099 2099 """
2100 2100 if pattern.startswith('re:'):
2101 2101 pattern = pattern[3:]
2102 2102 try:
2103 2103 flags = 0
2104 2104 if not casesensitive:
2105 2105 flags = remod.I
2106 2106 regex = remod.compile(pattern, flags)
2107 2107 except remod.error as e:
2108 2108 raise error.ParseError(_('invalid regular expression: %s')
2109 2109 % e)
2110 2110 return 're', pattern, regex.search
2111 2111 elif pattern.startswith('literal:'):
2112 2112 pattern = pattern[8:]
2113 2113
2114 2114 match = pattern.__eq__
2115 2115
2116 2116 if not casesensitive:
2117 2117 ipat = encoding.lower(pattern)
2118 2118 match = lambda s: ipat == encoding.lower(s)
2119 2119 return 'literal', pattern, match
2120 2120
2121 2121 def shortuser(user):
2122 2122 """Return a short representation of a user name or email address."""
2123 2123 f = user.find('@')
2124 2124 if f >= 0:
2125 2125 user = user[:f]
2126 2126 f = user.find('<')
2127 2127 if f >= 0:
2128 2128 user = user[f + 1:]
2129 2129 f = user.find(' ')
2130 2130 if f >= 0:
2131 2131 user = user[:f]
2132 2132 f = user.find('.')
2133 2133 if f >= 0:
2134 2134 user = user[:f]
2135 2135 return user
2136 2136
2137 2137 def emailuser(user):
2138 2138 """Return the user portion of an email address."""
2139 2139 f = user.find('@')
2140 2140 if f >= 0:
2141 2141 user = user[:f]
2142 2142 f = user.find('<')
2143 2143 if f >= 0:
2144 2144 user = user[f + 1:]
2145 2145 return user
2146 2146
2147 2147 def email(author):
2148 2148 '''get email of author.'''
2149 2149 r = author.find('>')
2150 2150 if r == -1:
2151 2151 r = None
2152 2152 return author[author.find('<') + 1:r]
2153 2153
2154 2154 def ellipsis(text, maxlength=400):
2155 2155 """Trim string to at most maxlength (default: 400) columns in display."""
2156 2156 return encoding.trim(text, maxlength, ellipsis='...')
2157 2157
2158 2158 def unitcountfn(*unittable):
2159 2159 '''return a function that renders a readable count of some quantity'''
2160 2160
2161 2161 def go(count):
2162 2162 for multiplier, divisor, format in unittable:
2163 2163 if count >= divisor * multiplier:
2164 2164 return format % (count / float(divisor))
2165 2165 return unittable[-1][2] % count
2166 2166
2167 2167 return go
2168 2168
2169 2169 def processlinerange(fromline, toline):
2170 2170 """Check that linerange <fromline>:<toline> makes sense and return a
2171 2171 0-based range.
2172 2172
2173 2173 >>> processlinerange(10, 20)
2174 2174 (9, 20)
2175 2175 >>> processlinerange(2, 1)
2176 2176 Traceback (most recent call last):
2177 2177 ...
2178 2178 ParseError: line range must be positive
2179 2179 >>> processlinerange(0, 5)
2180 2180 Traceback (most recent call last):
2181 2181 ...
2182 2182 ParseError: fromline must be strictly positive
2183 2183 """
2184 2184 if toline - fromline < 0:
2185 2185 raise error.ParseError(_("line range must be positive"))
2186 2186 if fromline < 1:
2187 2187 raise error.ParseError(_("fromline must be strictly positive"))
2188 2188 return fromline - 1, toline
2189 2189
2190 2190 bytecount = unitcountfn(
2191 2191 (100, 1 << 30, _('%.0f GB')),
2192 2192 (10, 1 << 30, _('%.1f GB')),
2193 2193 (1, 1 << 30, _('%.2f GB')),
2194 2194 (100, 1 << 20, _('%.0f MB')),
2195 2195 (10, 1 << 20, _('%.1f MB')),
2196 2196 (1, 1 << 20, _('%.2f MB')),
2197 2197 (100, 1 << 10, _('%.0f KB')),
2198 2198 (10, 1 << 10, _('%.1f KB')),
2199 2199 (1, 1 << 10, _('%.2f KB')),
2200 2200 (1, 1, _('%.0f bytes')),
2201 2201 )
2202 2202
2203 2203 # Matches a single EOL which can either be a CRLF where repeated CR
2204 2204 # are removed or a LF. We do not care about old Macintosh files, so a
2205 2205 # stray CR is an error.
2206 2206 _eolre = remod.compile(br'\r*\n')
2207 2207
2208 2208 def tolf(s):
2209 2209 return _eolre.sub('\n', s)
2210 2210
2211 2211 def tocrlf(s):
2212 2212 return _eolre.sub('\r\n', s)
2213 2213
2214 2214 if pycompat.oslinesep == '\r\n':
2215 2215 tonativeeol = tocrlf
2216 2216 fromnativeeol = tolf
2217 2217 else:
2218 2218 tonativeeol = pycompat.identity
2219 2219 fromnativeeol = pycompat.identity
2220 2220
2221 2221 def escapestr(s):
2222 2222 # call underlying function of s.encode('string_escape') directly for
2223 2223 # Python 3 compatibility
2224 2224 return codecs.escape_encode(s)[0]
2225 2225
2226 2226 def unescapestr(s):
2227 2227 return codecs.escape_decode(s)[0]
2228 2228
2229 2229 def uirepr(s):
2230 2230 # Avoid double backslash in Windows path repr()
2231 2231 return repr(s).replace('\\\\', '\\')
2232 2232
2233 2233 # delay import of textwrap
2234 2234 def MBTextWrapper(**kwargs):
2235 2235 class tw(textwrap.TextWrapper):
2236 2236 """
2237 2237 Extend TextWrapper for width-awareness.
2238 2238
2239 2239 Neither number of 'bytes' in any encoding nor 'characters' is
2240 2240 appropriate to calculate terminal columns for specified string.
2241 2241
2242 2242 Original TextWrapper implementation uses built-in 'len()' directly,
2243 2243 so overriding is needed to use width information of each characters.
2244 2244
2245 2245 In addition, characters classified into 'ambiguous' width are
2246 2246 treated as wide in East Asian area, but as narrow in other.
2247 2247
2248 2248 This requires use decision to determine width of such characters.
2249 2249 """
2250 2250 def _cutdown(self, ucstr, space_left):
2251 2251 l = 0
2252 2252 colwidth = encoding.ucolwidth
2253 2253 for i in xrange(len(ucstr)):
2254 2254 l += colwidth(ucstr[i])
2255 2255 if space_left < l:
2256 2256 return (ucstr[:i], ucstr[i:])
2257 2257 return ucstr, ''
2258 2258
2259 2259 # overriding of base class
2260 2260 def _handle_long_word(self, reversed_chunks, cur_line, cur_len, width):
2261 2261 space_left = max(width - cur_len, 1)
2262 2262
2263 2263 if self.break_long_words:
2264 2264 cut, res = self._cutdown(reversed_chunks[-1], space_left)
2265 2265 cur_line.append(cut)
2266 2266 reversed_chunks[-1] = res
2267 2267 elif not cur_line:
2268 2268 cur_line.append(reversed_chunks.pop())
2269 2269
2270 2270 # this overriding code is imported from TextWrapper of Python 2.6
2271 2271 # to calculate columns of string by 'encoding.ucolwidth()'
2272 2272 def _wrap_chunks(self, chunks):
2273 2273 colwidth = encoding.ucolwidth
2274 2274
2275 2275 lines = []
2276 2276 if self.width <= 0:
2277 2277 raise ValueError("invalid width %r (must be > 0)" % self.width)
2278 2278
2279 2279 # Arrange in reverse order so items can be efficiently popped
2280 2280 # from a stack of chucks.
2281 2281 chunks.reverse()
2282 2282
2283 2283 while chunks:
2284 2284
2285 2285 # Start the list of chunks that will make up the current line.
2286 2286 # cur_len is just the length of all the chunks in cur_line.
2287 2287 cur_line = []
2288 2288 cur_len = 0
2289 2289
2290 2290 # Figure out which static string will prefix this line.
2291 2291 if lines:
2292 2292 indent = self.subsequent_indent
2293 2293 else:
2294 2294 indent = self.initial_indent
2295 2295
2296 2296 # Maximum width for this line.
2297 2297 width = self.width - len(indent)
2298 2298
2299 2299 # First chunk on line is whitespace -- drop it, unless this
2300 2300 # is the very beginning of the text (i.e. no lines started yet).
2301 2301 if self.drop_whitespace and chunks[-1].strip() == '' and lines:
2302 2302 del chunks[-1]
2303 2303
2304 2304 while chunks:
2305 2305 l = colwidth(chunks[-1])
2306 2306
2307 2307 # Can at least squeeze this chunk onto the current line.
2308 2308 if cur_len + l <= width:
2309 2309 cur_line.append(chunks.pop())
2310 2310 cur_len += l
2311 2311
2312 2312 # Nope, this line is full.
2313 2313 else:
2314 2314 break
2315 2315
2316 2316 # The current line is full, and the next chunk is too big to
2317 2317 # fit on *any* line (not just this one).
2318 2318 if chunks and colwidth(chunks[-1]) > width:
2319 2319 self._handle_long_word(chunks, cur_line, cur_len, width)
2320 2320
2321 2321 # If the last chunk on this line is all whitespace, drop it.
2322 2322 if (self.drop_whitespace and
2323 2323 cur_line and cur_line[-1].strip() == ''):
2324 2324 del cur_line[-1]
2325 2325
2326 2326 # Convert current line back to a string and store it in list
2327 2327 # of all lines (return value).
2328 2328 if cur_line:
2329 2329 lines.append(indent + ''.join(cur_line))
2330 2330
2331 2331 return lines
2332 2332
2333 2333 global MBTextWrapper
2334 2334 MBTextWrapper = tw
2335 2335 return tw(**kwargs)
2336 2336
2337 2337 def wrap(line, width, initindent='', hangindent=''):
2338 2338 maxindent = max(len(hangindent), len(initindent))
2339 2339 if width <= maxindent:
2340 2340 # adjust for weird terminal size
2341 2341 width = max(78, maxindent + 1)
2342 2342 line = line.decode(pycompat.sysstr(encoding.encoding),
2343 2343 pycompat.sysstr(encoding.encodingmode))
2344 2344 initindent = initindent.decode(pycompat.sysstr(encoding.encoding),
2345 2345 pycompat.sysstr(encoding.encodingmode))
2346 2346 hangindent = hangindent.decode(pycompat.sysstr(encoding.encoding),
2347 2347 pycompat.sysstr(encoding.encodingmode))
2348 2348 wrapper = MBTextWrapper(width=width,
2349 2349 initial_indent=initindent,
2350 2350 subsequent_indent=hangindent)
2351 2351 return wrapper.fill(line).encode(pycompat.sysstr(encoding.encoding))
2352 2352
2353 2353 if (pyplatform.python_implementation() == 'CPython' and
2354 2354 sys.version_info < (3, 0)):
2355 2355 # There is an issue in CPython that some IO methods do not handle EINTR
2356 2356 # correctly. The following table shows what CPython version (and functions)
2357 2357 # are affected (buggy: has the EINTR bug, okay: otherwise):
2358 2358 #
2359 2359 # | < 2.7.4 | 2.7.4 to 2.7.12 | >= 3.0
2360 2360 # --------------------------------------------------
2361 2361 # fp.__iter__ | buggy | buggy | okay
2362 2362 # fp.read* | buggy | okay [1] | okay
2363 2363 #
2364 2364 # [1]: fixed by changeset 67dc99a989cd in the cpython hg repo.
2365 2365 #
2366 2366 # Here we workaround the EINTR issue for fileobj.__iter__. Other methods
2367 2367 # like "read*" are ignored for now, as Python < 2.7.4 is a minority.
2368 2368 #
2369 2369 # Although we can workaround the EINTR issue for fp.__iter__, it is slower:
2370 2370 # "for x in fp" is 4x faster than "for x in iter(fp.readline, '')" in
2371 2371 # CPython 2, because CPython 2 maintains an internal readahead buffer for
2372 2372 # fp.__iter__ but not other fp.read* methods.
2373 2373 #
2374 2374 # On modern systems like Linux, the "read" syscall cannot be interrupted
2375 2375 # when reading "fast" files like on-disk files. So the EINTR issue only
2376 2376 # affects things like pipes, sockets, ttys etc. We treat "normal" (S_ISREG)
2377 2377 # files approximately as "fast" files and use the fast (unsafe) code path,
2378 2378 # to minimize the performance impact.
2379 2379 if sys.version_info >= (2, 7, 4):
2380 2380 # fp.readline deals with EINTR correctly, use it as a workaround.
2381 2381 def _safeiterfile(fp):
2382 2382 return iter(fp.readline, '')
2383 2383 else:
2384 2384 # fp.read* are broken too, manually deal with EINTR in a stupid way.
2385 2385 # note: this may block longer than necessary because of bufsize.
2386 2386 def _safeiterfile(fp, bufsize=4096):
2387 2387 fd = fp.fileno()
2388 2388 line = ''
2389 2389 while True:
2390 2390 try:
2391 2391 buf = os.read(fd, bufsize)
2392 2392 except OSError as ex:
2393 2393 # os.read only raises EINTR before any data is read
2394 2394 if ex.errno == errno.EINTR:
2395 2395 continue
2396 2396 else:
2397 2397 raise
2398 2398 line += buf
2399 2399 if '\n' in buf:
2400 2400 splitted = line.splitlines(True)
2401 2401 line = ''
2402 2402 for l in splitted:
2403 2403 if l[-1] == '\n':
2404 2404 yield l
2405 2405 else:
2406 2406 line = l
2407 2407 if not buf:
2408 2408 break
2409 2409 if line:
2410 2410 yield line
2411 2411
2412 2412 def iterfile(fp):
2413 2413 fastpath = True
2414 2414 if type(fp) is file:
2415 2415 fastpath = stat.S_ISREG(os.fstat(fp.fileno()).st_mode)
2416 2416 if fastpath:
2417 2417 return fp
2418 2418 else:
2419 2419 return _safeiterfile(fp)
2420 2420 else:
2421 2421 # PyPy and CPython 3 do not have the EINTR issue thus no workaround needed.
2422 2422 def iterfile(fp):
2423 2423 return fp
2424 2424
2425 2425 def iterlines(iterator):
2426 2426 for chunk in iterator:
2427 2427 for line in chunk.splitlines():
2428 2428 yield line
2429 2429
2430 2430 def expandpath(path):
2431 2431 return os.path.expanduser(os.path.expandvars(path))
2432 2432
2433 2433 def hgcmd():
2434 2434 """Return the command used to execute current hg
2435 2435
2436 2436 This is different from hgexecutable() because on Windows we want
2437 2437 to avoid things opening new shell windows like batch files, so we
2438 2438 get either the python call or current executable.
2439 2439 """
2440 2440 if mainfrozen():
2441 2441 if getattr(sys, 'frozen', None) == 'macosx_app':
2442 2442 # Env variable set by py2app
2443 2443 return [encoding.environ['EXECUTABLEPATH']]
2444 2444 else:
2445 2445 return [pycompat.sysexecutable]
2446 2446 return gethgcmd()
2447 2447
2448 2448 def rundetached(args, condfn):
2449 2449 """Execute the argument list in a detached process.
2450 2450
2451 2451 condfn is a callable which is called repeatedly and should return
2452 2452 True once the child process is known to have started successfully.
2453 2453 At this point, the child process PID is returned. If the child
2454 2454 process fails to start or finishes before condfn() evaluates to
2455 2455 True, return -1.
2456 2456 """
2457 2457 # Windows case is easier because the child process is either
2458 2458 # successfully starting and validating the condition or exiting
2459 2459 # on failure. We just poll on its PID. On Unix, if the child
2460 2460 # process fails to start, it will be left in a zombie state until
2461 2461 # the parent wait on it, which we cannot do since we expect a long
2462 2462 # running process on success. Instead we listen for SIGCHLD telling
2463 2463 # us our child process terminated.
2464 2464 terminated = set()
2465 2465 def handler(signum, frame):
2466 2466 terminated.add(os.wait())
2467 2467 prevhandler = None
2468 2468 SIGCHLD = getattr(signal, 'SIGCHLD', None)
2469 2469 if SIGCHLD is not None:
2470 2470 prevhandler = signal.signal(SIGCHLD, handler)
2471 2471 try:
2472 2472 pid = spawndetached(args)
2473 2473 while not condfn():
2474 2474 if ((pid in terminated or not testpid(pid))
2475 2475 and not condfn()):
2476 2476 return -1
2477 2477 time.sleep(0.1)
2478 2478 return pid
2479 2479 finally:
2480 2480 if prevhandler is not None:
2481 2481 signal.signal(signal.SIGCHLD, prevhandler)
2482 2482
2483 2483 def interpolate(prefix, mapping, s, fn=None, escape_prefix=False):
2484 2484 """Return the result of interpolating items in the mapping into string s.
2485 2485
2486 2486 prefix is a single character string, or a two character string with
2487 2487 a backslash as the first character if the prefix needs to be escaped in
2488 2488 a regular expression.
2489 2489
2490 2490 fn is an optional function that will be applied to the replacement text
2491 2491 just before replacement.
2492 2492
2493 2493 escape_prefix is an optional flag that allows using doubled prefix for
2494 2494 its escaping.
2495 2495 """
2496 2496 fn = fn or (lambda s: s)
2497 2497 patterns = '|'.join(mapping.keys())
2498 2498 if escape_prefix:
2499 2499 patterns += '|' + prefix
2500 2500 if len(prefix) > 1:
2501 2501 prefix_char = prefix[1:]
2502 2502 else:
2503 2503 prefix_char = prefix
2504 2504 mapping[prefix_char] = prefix_char
2505 2505 r = remod.compile(r'%s(%s)' % (prefix, patterns))
2506 2506 return r.sub(lambda x: fn(mapping[x.group()[1:]]), s)
2507 2507
2508 2508 def getport(port):
2509 2509 """Return the port for a given network service.
2510 2510
2511 2511 If port is an integer, it's returned as is. If it's a string, it's
2512 2512 looked up using socket.getservbyname(). If there's no matching
2513 2513 service, error.Abort is raised.
2514 2514 """
2515 2515 try:
2516 2516 return int(port)
2517 2517 except ValueError:
2518 2518 pass
2519 2519
2520 2520 try:
2521 2521 return socket.getservbyname(port)
2522 2522 except socket.error:
2523 2523 raise Abort(_("no port number associated with service '%s'") % port)
2524 2524
2525 2525 _booleans = {'1': True, 'yes': True, 'true': True, 'on': True, 'always': True,
2526 2526 '0': False, 'no': False, 'false': False, 'off': False,
2527 2527 'never': False}
2528 2528
2529 2529 def parsebool(s):
2530 2530 """Parse s into a boolean.
2531 2531
2532 2532 If s is not a valid boolean, returns None.
2533 2533 """
2534 2534 return _booleans.get(s.lower(), None)
2535 2535
2536 2536 _hextochr = dict((a + b, chr(int(a + b, 16)))
2537 2537 for a in string.hexdigits for b in string.hexdigits)
2538 2538
2539 2539 class url(object):
2540 2540 r"""Reliable URL parser.
2541 2541
2542 2542 This parses URLs and provides attributes for the following
2543 2543 components:
2544 2544
2545 2545 <scheme>://<user>:<passwd>@<host>:<port>/<path>?<query>#<fragment>
2546 2546
2547 2547 Missing components are set to None. The only exception is
2548 2548 fragment, which is set to '' if present but empty.
2549 2549
2550 2550 If parsefragment is False, fragment is included in query. If
2551 2551 parsequery is False, query is included in path. If both are
2552 2552 False, both fragment and query are included in path.
2553 2553
2554 2554 See http://www.ietf.org/rfc/rfc2396.txt for more information.
2555 2555
2556 2556 Note that for backward compatibility reasons, bundle URLs do not
2557 2557 take host names. That means 'bundle://../' has a path of '../'.
2558 2558
2559 2559 Examples:
2560 2560
2561 2561 >>> url('http://www.ietf.org/rfc/rfc2396.txt')
2562 2562 <url scheme: 'http', host: 'www.ietf.org', path: 'rfc/rfc2396.txt'>
2563 2563 >>> url('ssh://[::1]:2200//home/joe/repo')
2564 2564 <url scheme: 'ssh', host: '[::1]', port: '2200', path: '/home/joe/repo'>
2565 2565 >>> url('file:///home/joe/repo')
2566 2566 <url scheme: 'file', path: '/home/joe/repo'>
2567 2567 >>> url('file:///c:/temp/foo/')
2568 2568 <url scheme: 'file', path: 'c:/temp/foo/'>
2569 2569 >>> url('bundle:foo')
2570 2570 <url scheme: 'bundle', path: 'foo'>
2571 2571 >>> url('bundle://../foo')
2572 2572 <url scheme: 'bundle', path: '../foo'>
2573 2573 >>> url(r'c:\foo\bar')
2574 2574 <url path: 'c:\\foo\\bar'>
2575 2575 >>> url(r'\\blah\blah\blah')
2576 2576 <url path: '\\\\blah\\blah\\blah'>
2577 2577 >>> url(r'\\blah\blah\blah#baz')
2578 2578 <url path: '\\\\blah\\blah\\blah', fragment: 'baz'>
2579 2579 >>> url(r'file:///C:\users\me')
2580 2580 <url scheme: 'file', path: 'C:\\users\\me'>
2581 2581
2582 2582 Authentication credentials:
2583 2583
2584 2584 >>> url('ssh://joe:xyz@x/repo')
2585 2585 <url scheme: 'ssh', user: 'joe', passwd: 'xyz', host: 'x', path: 'repo'>
2586 2586 >>> url('ssh://joe@x/repo')
2587 2587 <url scheme: 'ssh', user: 'joe', host: 'x', path: 'repo'>
2588 2588
2589 2589 Query strings and fragments:
2590 2590
2591 2591 >>> url('http://host/a?b#c')
2592 2592 <url scheme: 'http', host: 'host', path: 'a', query: 'b', fragment: 'c'>
2593 2593 >>> url('http://host/a?b#c', parsequery=False, parsefragment=False)
2594 2594 <url scheme: 'http', host: 'host', path: 'a?b#c'>
2595 2595
2596 2596 Empty path:
2597 2597
2598 2598 >>> url('')
2599 2599 <url path: ''>
2600 2600 >>> url('#a')
2601 2601 <url path: '', fragment: 'a'>
2602 2602 >>> url('http://host/')
2603 2603 <url scheme: 'http', host: 'host', path: ''>
2604 2604 >>> url('http://host/#a')
2605 2605 <url scheme: 'http', host: 'host', path: '', fragment: 'a'>
2606 2606
2607 2607 Only scheme:
2608 2608
2609 2609 >>> url('http:')
2610 2610 <url scheme: 'http'>
2611 2611 """
2612 2612
2613 2613 _safechars = "!~*'()+"
2614 2614 _safepchars = "/!~*'()+:\\"
2615 2615 _matchscheme = remod.compile('^[a-zA-Z0-9+.\\-]+:').match
2616 2616
2617 2617 def __init__(self, path, parsequery=True, parsefragment=True):
2618 2618 # We slowly chomp away at path until we have only the path left
2619 2619 self.scheme = self.user = self.passwd = self.host = None
2620 2620 self.port = self.path = self.query = self.fragment = None
2621 2621 self._localpath = True
2622 2622 self._hostport = ''
2623 2623 self._origpath = path
2624 2624
2625 2625 if parsefragment and '#' in path:
2626 2626 path, self.fragment = path.split('#', 1)
2627 2627
2628 2628 # special case for Windows drive letters and UNC paths
2629 2629 if hasdriveletter(path) or path.startswith('\\\\'):
2630 2630 self.path = path
2631 2631 return
2632 2632
2633 2633 # For compatibility reasons, we can't handle bundle paths as
2634 2634 # normal URLS
2635 2635 if path.startswith('bundle:'):
2636 2636 self.scheme = 'bundle'
2637 2637 path = path[7:]
2638 2638 if path.startswith('//'):
2639 2639 path = path[2:]
2640 2640 self.path = path
2641 2641 return
2642 2642
2643 2643 if self._matchscheme(path):
2644 2644 parts = path.split(':', 1)
2645 2645 if parts[0]:
2646 2646 self.scheme, path = parts
2647 2647 self._localpath = False
2648 2648
2649 2649 if not path:
2650 2650 path = None
2651 2651 if self._localpath:
2652 2652 self.path = ''
2653 2653 return
2654 2654 else:
2655 2655 if self._localpath:
2656 2656 self.path = path
2657 2657 return
2658 2658
2659 2659 if parsequery and '?' in path:
2660 2660 path, self.query = path.split('?', 1)
2661 2661 if not path:
2662 2662 path = None
2663 2663 if not self.query:
2664 2664 self.query = None
2665 2665
2666 2666 # // is required to specify a host/authority
2667 2667 if path and path.startswith('//'):
2668 2668 parts = path[2:].split('/', 1)
2669 2669 if len(parts) > 1:
2670 2670 self.host, path = parts
2671 2671 else:
2672 2672 self.host = parts[0]
2673 2673 path = None
2674 2674 if not self.host:
2675 2675 self.host = None
2676 2676 # path of file:///d is /d
2677 2677 # path of file:///d:/ is d:/, not /d:/
2678 2678 if path and not hasdriveletter(path):
2679 2679 path = '/' + path
2680 2680
2681 2681 if self.host and '@' in self.host:
2682 2682 self.user, self.host = self.host.rsplit('@', 1)
2683 2683 if ':' in self.user:
2684 2684 self.user, self.passwd = self.user.split(':', 1)
2685 2685 if not self.host:
2686 2686 self.host = None
2687 2687
2688 2688 # Don't split on colons in IPv6 addresses without ports
2689 2689 if (self.host and ':' in self.host and
2690 2690 not (self.host.startswith('[') and self.host.endswith(']'))):
2691 2691 self._hostport = self.host
2692 2692 self.host, self.port = self.host.rsplit(':', 1)
2693 2693 if not self.host:
2694 2694 self.host = None
2695 2695
2696 2696 if (self.host and self.scheme == 'file' and
2697 2697 self.host not in ('localhost', '127.0.0.1', '[::1]')):
2698 2698 raise Abort(_('file:// URLs can only refer to localhost'))
2699 2699
2700 2700 self.path = path
2701 2701
2702 2702 # leave the query string escaped
2703 2703 for a in ('user', 'passwd', 'host', 'port',
2704 2704 'path', 'fragment'):
2705 2705 v = getattr(self, a)
2706 2706 if v is not None:
2707 2707 setattr(self, a, urlreq.unquote(v))
2708 2708
2709 2709 def __repr__(self):
2710 2710 attrs = []
2711 2711 for a in ('scheme', 'user', 'passwd', 'host', 'port', 'path',
2712 2712 'query', 'fragment'):
2713 2713 v = getattr(self, a)
2714 2714 if v is not None:
2715 2715 attrs.append('%s: %r' % (a, v))
2716 2716 return '<url %s>' % ', '.join(attrs)
2717 2717
2718 2718 def __str__(self):
2719 2719 r"""Join the URL's components back into a URL string.
2720 2720
2721 2721 Examples:
2722 2722
2723 2723 >>> str(url('http://user:pw@host:80/c:/bob?fo:oo#ba:ar'))
2724 2724 'http://user:pw@host:80/c:/bob?fo:oo#ba:ar'
2725 2725 >>> str(url('http://user:pw@host:80/?foo=bar&baz=42'))
2726 2726 'http://user:pw@host:80/?foo=bar&baz=42'
2727 2727 >>> str(url('http://user:pw@host:80/?foo=bar%3dbaz'))
2728 2728 'http://user:pw@host:80/?foo=bar%3dbaz'
2729 2729 >>> str(url('ssh://user:pw@[::1]:2200//home/joe#'))
2730 2730 'ssh://user:pw@[::1]:2200//home/joe#'
2731 2731 >>> str(url('http://localhost:80//'))
2732 2732 'http://localhost:80//'
2733 2733 >>> str(url('http://localhost:80/'))
2734 2734 'http://localhost:80/'
2735 2735 >>> str(url('http://localhost:80'))
2736 2736 'http://localhost:80/'
2737 2737 >>> str(url('bundle:foo'))
2738 2738 'bundle:foo'
2739 2739 >>> str(url('bundle://../foo'))
2740 2740 'bundle:../foo'
2741 2741 >>> str(url('path'))
2742 2742 'path'
2743 2743 >>> str(url('file:///tmp/foo/bar'))
2744 2744 'file:///tmp/foo/bar'
2745 2745 >>> str(url('file:///c:/tmp/foo/bar'))
2746 2746 'file:///c:/tmp/foo/bar'
2747 2747 >>> print url(r'bundle:foo\bar')
2748 2748 bundle:foo\bar
2749 2749 >>> print url(r'file:///D:\data\hg')
2750 2750 file:///D:\data\hg
2751 2751 """
2752 2752 return encoding.strfromlocal(self.__bytes__())
2753 2753
2754 2754 def __bytes__(self):
2755 2755 if self._localpath:
2756 2756 s = self.path
2757 2757 if self.scheme == 'bundle':
2758 2758 s = 'bundle:' + s
2759 2759 if self.fragment:
2760 2760 s += '#' + self.fragment
2761 2761 return s
2762 2762
2763 2763 s = self.scheme + ':'
2764 2764 if self.user or self.passwd or self.host:
2765 2765 s += '//'
2766 2766 elif self.scheme and (not self.path or self.path.startswith('/')
2767 2767 or hasdriveletter(self.path)):
2768 2768 s += '//'
2769 2769 if hasdriveletter(self.path):
2770 2770 s += '/'
2771 2771 if self.user:
2772 2772 s += urlreq.quote(self.user, safe=self._safechars)
2773 2773 if self.passwd:
2774 2774 s += ':' + urlreq.quote(self.passwd, safe=self._safechars)
2775 2775 if self.user or self.passwd:
2776 2776 s += '@'
2777 2777 if self.host:
2778 2778 if not (self.host.startswith('[') and self.host.endswith(']')):
2779 2779 s += urlreq.quote(self.host)
2780 2780 else:
2781 2781 s += self.host
2782 2782 if self.port:
2783 2783 s += ':' + urlreq.quote(self.port)
2784 2784 if self.host:
2785 2785 s += '/'
2786 2786 if self.path:
2787 2787 # TODO: similar to the query string, we should not unescape the
2788 2788 # path when we store it, the path might contain '%2f' = '/',
2789 2789 # which we should *not* escape.
2790 2790 s += urlreq.quote(self.path, safe=self._safepchars)
2791 2791 if self.query:
2792 2792 # we store the query in escaped form.
2793 2793 s += '?' + self.query
2794 2794 if self.fragment is not None:
2795 2795 s += '#' + urlreq.quote(self.fragment, safe=self._safepchars)
2796 2796 return s
2797 2797
2798 2798 def authinfo(self):
2799 2799 user, passwd = self.user, self.passwd
2800 2800 try:
2801 2801 self.user, self.passwd = None, None
2802 s = str(self)
2802 s = bytes(self)
2803 2803 finally:
2804 2804 self.user, self.passwd = user, passwd
2805 2805 if not self.user:
2806 2806 return (s, None)
2807 2807 # authinfo[1] is passed to urllib2 password manager, and its
2808 2808 # URIs must not contain credentials. The host is passed in the
2809 2809 # URIs list because Python < 2.4.3 uses only that to search for
2810 2810 # a password.
2811 2811 return (s, (None, (s, self.host),
2812 2812 self.user, self.passwd or ''))
2813 2813
2814 2814 def isabs(self):
2815 2815 if self.scheme and self.scheme != 'file':
2816 2816 return True # remote URL
2817 2817 if hasdriveletter(self.path):
2818 2818 return True # absolute for our purposes - can't be joined()
2819 2819 if self.path.startswith(r'\\'):
2820 2820 return True # Windows UNC path
2821 2821 if self.path.startswith('/'):
2822 2822 return True # POSIX-style
2823 2823 return False
2824 2824
2825 2825 def localpath(self):
2826 2826 if self.scheme == 'file' or self.scheme == 'bundle':
2827 2827 path = self.path or '/'
2828 2828 # For Windows, we need to promote hosts containing drive
2829 2829 # letters to paths with drive letters.
2830 2830 if hasdriveletter(self._hostport):
2831 2831 path = self._hostport + '/' + self.path
2832 2832 elif (self.host is not None and self.path
2833 2833 and not hasdriveletter(path)):
2834 2834 path = '/' + path
2835 2835 return path
2836 2836 return self._origpath
2837 2837
2838 2838 def islocal(self):
2839 2839 '''whether localpath will return something that posixfile can open'''
2840 2840 return (not self.scheme or self.scheme == 'file'
2841 2841 or self.scheme == 'bundle')
2842 2842
2843 2843 def hasscheme(path):
2844 2844 return bool(url(path).scheme)
2845 2845
2846 2846 def hasdriveletter(path):
2847 2847 return path and path[1:2] == ':' and path[0:1].isalpha()
2848 2848
2849 2849 def urllocalpath(path):
2850 2850 return url(path, parsequery=False, parsefragment=False).localpath()
2851 2851
2852 2852 def hidepassword(u):
2853 2853 '''hide user credential in a url string'''
2854 2854 u = url(u)
2855 2855 if u.passwd:
2856 2856 u.passwd = '***'
2857 return str(u)
2857 return bytes(u)
2858 2858
2859 2859 def removeauth(u):
2860 2860 '''remove all authentication information from a url string'''
2861 2861 u = url(u)
2862 2862 u.user = u.passwd = None
2863 2863 return str(u)
2864 2864
2865 2865 timecount = unitcountfn(
2866 2866 (1, 1e3, _('%.0f s')),
2867 2867 (100, 1, _('%.1f s')),
2868 2868 (10, 1, _('%.2f s')),
2869 2869 (1, 1, _('%.3f s')),
2870 2870 (100, 0.001, _('%.1f ms')),
2871 2871 (10, 0.001, _('%.2f ms')),
2872 2872 (1, 0.001, _('%.3f ms')),
2873 2873 (100, 0.000001, _('%.1f us')),
2874 2874 (10, 0.000001, _('%.2f us')),
2875 2875 (1, 0.000001, _('%.3f us')),
2876 2876 (100, 0.000000001, _('%.1f ns')),
2877 2877 (10, 0.000000001, _('%.2f ns')),
2878 2878 (1, 0.000000001, _('%.3f ns')),
2879 2879 )
2880 2880
2881 2881 _timenesting = [0]
2882 2882
2883 2883 def timed(func):
2884 2884 '''Report the execution time of a function call to stderr.
2885 2885
2886 2886 During development, use as a decorator when you need to measure
2887 2887 the cost of a function, e.g. as follows:
2888 2888
2889 2889 @util.timed
2890 2890 def foo(a, b, c):
2891 2891 pass
2892 2892 '''
2893 2893
2894 2894 def wrapper(*args, **kwargs):
2895 2895 start = timer()
2896 2896 indent = 2
2897 2897 _timenesting[0] += indent
2898 2898 try:
2899 2899 return func(*args, **kwargs)
2900 2900 finally:
2901 2901 elapsed = timer() - start
2902 2902 _timenesting[0] -= indent
2903 2903 stderr.write('%s%s: %s\n' %
2904 2904 (' ' * _timenesting[0], func.__name__,
2905 2905 timecount(elapsed)))
2906 2906 return wrapper
2907 2907
2908 2908 _sizeunits = (('m', 2**20), ('k', 2**10), ('g', 2**30),
2909 2909 ('kb', 2**10), ('mb', 2**20), ('gb', 2**30), ('b', 1))
2910 2910
2911 2911 def sizetoint(s):
2912 2912 '''Convert a space specifier to a byte count.
2913 2913
2914 2914 >>> sizetoint('30')
2915 2915 30
2916 2916 >>> sizetoint('2.2kb')
2917 2917 2252
2918 2918 >>> sizetoint('6M')
2919 2919 6291456
2920 2920 '''
2921 2921 t = s.strip().lower()
2922 2922 try:
2923 2923 for k, u in _sizeunits:
2924 2924 if t.endswith(k):
2925 2925 return int(float(t[:-len(k)]) * u)
2926 2926 return int(t)
2927 2927 except ValueError:
2928 2928 raise error.ParseError(_("couldn't parse size: %s") % s)
2929 2929
2930 2930 class hooks(object):
2931 2931 '''A collection of hook functions that can be used to extend a
2932 2932 function's behavior. Hooks are called in lexicographic order,
2933 2933 based on the names of their sources.'''
2934 2934
2935 2935 def __init__(self):
2936 2936 self._hooks = []
2937 2937
2938 2938 def add(self, source, hook):
2939 2939 self._hooks.append((source, hook))
2940 2940
2941 2941 def __call__(self, *args):
2942 2942 self._hooks.sort(key=lambda x: x[0])
2943 2943 results = []
2944 2944 for source, hook in self._hooks:
2945 2945 results.append(hook(*args))
2946 2946 return results
2947 2947
2948 2948 def getstackframes(skip=0, line=' %-*s in %s\n', fileline='%s:%s', depth=0):
2949 2949 '''Yields lines for a nicely formatted stacktrace.
2950 2950 Skips the 'skip' last entries, then return the last 'depth' entries.
2951 2951 Each file+linenumber is formatted according to fileline.
2952 2952 Each line is formatted according to line.
2953 2953 If line is None, it yields:
2954 2954 length of longest filepath+line number,
2955 2955 filepath+linenumber,
2956 2956 function
2957 2957
2958 2958 Not be used in production code but very convenient while developing.
2959 2959 '''
2960 2960 entries = [(fileline % (fn, ln), func)
2961 2961 for fn, ln, func, _text in traceback.extract_stack()[:-skip - 1]
2962 2962 ][-depth:]
2963 2963 if entries:
2964 2964 fnmax = max(len(entry[0]) for entry in entries)
2965 2965 for fnln, func in entries:
2966 2966 if line is None:
2967 2967 yield (fnmax, fnln, func)
2968 2968 else:
2969 2969 yield line % (fnmax, fnln, func)
2970 2970
2971 2971 def debugstacktrace(msg='stacktrace', skip=0,
2972 2972 f=stderr, otherf=stdout, depth=0):
2973 2973 '''Writes a message to f (stderr) with a nicely formatted stacktrace.
2974 2974 Skips the 'skip' entries closest to the call, then show 'depth' entries.
2975 2975 By default it will flush stdout first.
2976 2976 It can be used everywhere and intentionally does not require an ui object.
2977 2977 Not be used in production code but very convenient while developing.
2978 2978 '''
2979 2979 if otherf:
2980 2980 otherf.flush()
2981 2981 f.write('%s at:\n' % msg.rstrip())
2982 2982 for line in getstackframes(skip + 1, depth=depth):
2983 2983 f.write(line)
2984 2984 f.flush()
2985 2985
2986 2986 class dirs(object):
2987 2987 '''a multiset of directory names from a dirstate or manifest'''
2988 2988
2989 2989 def __init__(self, map, skip=None):
2990 2990 self._dirs = {}
2991 2991 addpath = self.addpath
2992 2992 if safehasattr(map, 'iteritems') and skip is not None:
2993 2993 for f, s in map.iteritems():
2994 2994 if s[0] != skip:
2995 2995 addpath(f)
2996 2996 else:
2997 2997 for f in map:
2998 2998 addpath(f)
2999 2999
3000 3000 def addpath(self, path):
3001 3001 dirs = self._dirs
3002 3002 for base in finddirs(path):
3003 3003 if base in dirs:
3004 3004 dirs[base] += 1
3005 3005 return
3006 3006 dirs[base] = 1
3007 3007
3008 3008 def delpath(self, path):
3009 3009 dirs = self._dirs
3010 3010 for base in finddirs(path):
3011 3011 if dirs[base] > 1:
3012 3012 dirs[base] -= 1
3013 3013 return
3014 3014 del dirs[base]
3015 3015
3016 3016 def __iter__(self):
3017 3017 return iter(self._dirs)
3018 3018
3019 3019 def __contains__(self, d):
3020 3020 return d in self._dirs
3021 3021
3022 3022 if safehasattr(parsers, 'dirs'):
3023 3023 dirs = parsers.dirs
3024 3024
3025 3025 def finddirs(path):
3026 3026 pos = path.rfind('/')
3027 3027 while pos != -1:
3028 3028 yield path[:pos]
3029 3029 pos = path.rfind('/', 0, pos)
3030 3030
3031 3031 class ctxmanager(object):
3032 3032 '''A context manager for use in 'with' blocks to allow multiple
3033 3033 contexts to be entered at once. This is both safer and more
3034 3034 flexible than contextlib.nested.
3035 3035
3036 3036 Once Mercurial supports Python 2.7+, this will become mostly
3037 3037 unnecessary.
3038 3038 '''
3039 3039
3040 3040 def __init__(self, *args):
3041 3041 '''Accepts a list of no-argument functions that return context
3042 3042 managers. These will be invoked at __call__ time.'''
3043 3043 self._pending = args
3044 3044 self._atexit = []
3045 3045
3046 3046 def __enter__(self):
3047 3047 return self
3048 3048
3049 3049 def enter(self):
3050 3050 '''Create and enter context managers in the order in which they were
3051 3051 passed to the constructor.'''
3052 3052 values = []
3053 3053 for func in self._pending:
3054 3054 obj = func()
3055 3055 values.append(obj.__enter__())
3056 3056 self._atexit.append(obj.__exit__)
3057 3057 del self._pending
3058 3058 return values
3059 3059
3060 3060 def atexit(self, func, *args, **kwargs):
3061 3061 '''Add a function to call when this context manager exits. The
3062 3062 ordering of multiple atexit calls is unspecified, save that
3063 3063 they will happen before any __exit__ functions.'''
3064 3064 def wrapper(exc_type, exc_val, exc_tb):
3065 3065 func(*args, **kwargs)
3066 3066 self._atexit.append(wrapper)
3067 3067 return func
3068 3068
3069 3069 def __exit__(self, exc_type, exc_val, exc_tb):
3070 3070 '''Context managers are exited in the reverse order from which
3071 3071 they were created.'''
3072 3072 received = exc_type is not None
3073 3073 suppressed = False
3074 3074 pending = None
3075 3075 self._atexit.reverse()
3076 3076 for exitfunc in self._atexit:
3077 3077 try:
3078 3078 if exitfunc(exc_type, exc_val, exc_tb):
3079 3079 suppressed = True
3080 3080 exc_type = None
3081 3081 exc_val = None
3082 3082 exc_tb = None
3083 3083 except BaseException:
3084 3084 pending = sys.exc_info()
3085 3085 exc_type, exc_val, exc_tb = pending = sys.exc_info()
3086 3086 del self._atexit
3087 3087 if pending:
3088 3088 raise exc_val
3089 3089 return received and suppressed
3090 3090
3091 3091 # compression code
3092 3092
3093 3093 SERVERROLE = 'server'
3094 3094 CLIENTROLE = 'client'
3095 3095
3096 3096 compewireprotosupport = collections.namedtuple(u'compenginewireprotosupport',
3097 3097 (u'name', u'serverpriority',
3098 3098 u'clientpriority'))
3099 3099
3100 3100 class compressormanager(object):
3101 3101 """Holds registrations of various compression engines.
3102 3102
3103 3103 This class essentially abstracts the differences between compression
3104 3104 engines to allow new compression formats to be added easily, possibly from
3105 3105 extensions.
3106 3106
3107 3107 Compressors are registered against the global instance by calling its
3108 3108 ``register()`` method.
3109 3109 """
3110 3110 def __init__(self):
3111 3111 self._engines = {}
3112 3112 # Bundle spec human name to engine name.
3113 3113 self._bundlenames = {}
3114 3114 # Internal bundle identifier to engine name.
3115 3115 self._bundletypes = {}
3116 3116 # Revlog header to engine name.
3117 3117 self._revlogheaders = {}
3118 3118 # Wire proto identifier to engine name.
3119 3119 self._wiretypes = {}
3120 3120
3121 3121 def __getitem__(self, key):
3122 3122 return self._engines[key]
3123 3123
3124 3124 def __contains__(self, key):
3125 3125 return key in self._engines
3126 3126
3127 3127 def __iter__(self):
3128 3128 return iter(self._engines.keys())
3129 3129
3130 3130 def register(self, engine):
3131 3131 """Register a compression engine with the manager.
3132 3132
3133 3133 The argument must be a ``compressionengine`` instance.
3134 3134 """
3135 3135 if not isinstance(engine, compressionengine):
3136 3136 raise ValueError(_('argument must be a compressionengine'))
3137 3137
3138 3138 name = engine.name()
3139 3139
3140 3140 if name in self._engines:
3141 3141 raise error.Abort(_('compression engine %s already registered') %
3142 3142 name)
3143 3143
3144 3144 bundleinfo = engine.bundletype()
3145 3145 if bundleinfo:
3146 3146 bundlename, bundletype = bundleinfo
3147 3147
3148 3148 if bundlename in self._bundlenames:
3149 3149 raise error.Abort(_('bundle name %s already registered') %
3150 3150 bundlename)
3151 3151 if bundletype in self._bundletypes:
3152 3152 raise error.Abort(_('bundle type %s already registered by %s') %
3153 3153 (bundletype, self._bundletypes[bundletype]))
3154 3154
3155 3155 # No external facing name declared.
3156 3156 if bundlename:
3157 3157 self._bundlenames[bundlename] = name
3158 3158
3159 3159 self._bundletypes[bundletype] = name
3160 3160
3161 3161 wiresupport = engine.wireprotosupport()
3162 3162 if wiresupport:
3163 3163 wiretype = wiresupport.name
3164 3164 if wiretype in self._wiretypes:
3165 3165 raise error.Abort(_('wire protocol compression %s already '
3166 3166 'registered by %s') %
3167 3167 (wiretype, self._wiretypes[wiretype]))
3168 3168
3169 3169 self._wiretypes[wiretype] = name
3170 3170
3171 3171 revlogheader = engine.revlogheader()
3172 3172 if revlogheader and revlogheader in self._revlogheaders:
3173 3173 raise error.Abort(_('revlog header %s already registered by %s') %
3174 3174 (revlogheader, self._revlogheaders[revlogheader]))
3175 3175
3176 3176 if revlogheader:
3177 3177 self._revlogheaders[revlogheader] = name
3178 3178
3179 3179 self._engines[name] = engine
3180 3180
3181 3181 @property
3182 3182 def supportedbundlenames(self):
3183 3183 return set(self._bundlenames.keys())
3184 3184
3185 3185 @property
3186 3186 def supportedbundletypes(self):
3187 3187 return set(self._bundletypes.keys())
3188 3188
3189 3189 def forbundlename(self, bundlename):
3190 3190 """Obtain a compression engine registered to a bundle name.
3191 3191
3192 3192 Will raise KeyError if the bundle type isn't registered.
3193 3193
3194 3194 Will abort if the engine is known but not available.
3195 3195 """
3196 3196 engine = self._engines[self._bundlenames[bundlename]]
3197 3197 if not engine.available():
3198 3198 raise error.Abort(_('compression engine %s could not be loaded') %
3199 3199 engine.name())
3200 3200 return engine
3201 3201
3202 3202 def forbundletype(self, bundletype):
3203 3203 """Obtain a compression engine registered to a bundle type.
3204 3204
3205 3205 Will raise KeyError if the bundle type isn't registered.
3206 3206
3207 3207 Will abort if the engine is known but not available.
3208 3208 """
3209 3209 engine = self._engines[self._bundletypes[bundletype]]
3210 3210 if not engine.available():
3211 3211 raise error.Abort(_('compression engine %s could not be loaded') %
3212 3212 engine.name())
3213 3213 return engine
3214 3214
3215 3215 def supportedwireengines(self, role, onlyavailable=True):
3216 3216 """Obtain compression engines that support the wire protocol.
3217 3217
3218 3218 Returns a list of engines in prioritized order, most desired first.
3219 3219
3220 3220 If ``onlyavailable`` is set, filter out engines that can't be
3221 3221 loaded.
3222 3222 """
3223 3223 assert role in (SERVERROLE, CLIENTROLE)
3224 3224
3225 3225 attr = 'serverpriority' if role == SERVERROLE else 'clientpriority'
3226 3226
3227 3227 engines = [self._engines[e] for e in self._wiretypes.values()]
3228 3228 if onlyavailable:
3229 3229 engines = [e for e in engines if e.available()]
3230 3230
3231 3231 def getkey(e):
3232 3232 # Sort first by priority, highest first. In case of tie, sort
3233 3233 # alphabetically. This is arbitrary, but ensures output is
3234 3234 # stable.
3235 3235 w = e.wireprotosupport()
3236 3236 return -1 * getattr(w, attr), w.name
3237 3237
3238 3238 return list(sorted(engines, key=getkey))
3239 3239
3240 3240 def forwiretype(self, wiretype):
3241 3241 engine = self._engines[self._wiretypes[wiretype]]
3242 3242 if not engine.available():
3243 3243 raise error.Abort(_('compression engine %s could not be loaded') %
3244 3244 engine.name())
3245 3245 return engine
3246 3246
3247 3247 def forrevlogheader(self, header):
3248 3248 """Obtain a compression engine registered to a revlog header.
3249 3249
3250 3250 Will raise KeyError if the revlog header value isn't registered.
3251 3251 """
3252 3252 return self._engines[self._revlogheaders[header]]
3253 3253
3254 3254 compengines = compressormanager()
3255 3255
3256 3256 class compressionengine(object):
3257 3257 """Base class for compression engines.
3258 3258
3259 3259 Compression engines must implement the interface defined by this class.
3260 3260 """
3261 3261 def name(self):
3262 3262 """Returns the name of the compression engine.
3263 3263
3264 3264 This is the key the engine is registered under.
3265 3265
3266 3266 This method must be implemented.
3267 3267 """
3268 3268 raise NotImplementedError()
3269 3269
3270 3270 def available(self):
3271 3271 """Whether the compression engine is available.
3272 3272
3273 3273 The intent of this method is to allow optional compression engines
3274 3274 that may not be available in all installations (such as engines relying
3275 3275 on C extensions that may not be present).
3276 3276 """
3277 3277 return True
3278 3278
3279 3279 def bundletype(self):
3280 3280 """Describes bundle identifiers for this engine.
3281 3281
3282 3282 If this compression engine isn't supported for bundles, returns None.
3283 3283
3284 3284 If this engine can be used for bundles, returns a 2-tuple of strings of
3285 3285 the user-facing "bundle spec" compression name and an internal
3286 3286 identifier used to denote the compression format within bundles. To
3287 3287 exclude the name from external usage, set the first element to ``None``.
3288 3288
3289 3289 If bundle compression is supported, the class must also implement
3290 3290 ``compressstream`` and `decompressorreader``.
3291 3291
3292 3292 The docstring of this method is used in the help system to tell users
3293 3293 about this engine.
3294 3294 """
3295 3295 return None
3296 3296
3297 3297 def wireprotosupport(self):
3298 3298 """Declare support for this compression format on the wire protocol.
3299 3299
3300 3300 If this compression engine isn't supported for compressing wire
3301 3301 protocol payloads, returns None.
3302 3302
3303 3303 Otherwise, returns ``compenginewireprotosupport`` with the following
3304 3304 fields:
3305 3305
3306 3306 * String format identifier
3307 3307 * Integer priority for the server
3308 3308 * Integer priority for the client
3309 3309
3310 3310 The integer priorities are used to order the advertisement of format
3311 3311 support by server and client. The highest integer is advertised
3312 3312 first. Integers with non-positive values aren't advertised.
3313 3313
3314 3314 The priority values are somewhat arbitrary and only used for default
3315 3315 ordering. The relative order can be changed via config options.
3316 3316
3317 3317 If wire protocol compression is supported, the class must also implement
3318 3318 ``compressstream`` and ``decompressorreader``.
3319 3319 """
3320 3320 return None
3321 3321
3322 3322 def revlogheader(self):
3323 3323 """Header added to revlog chunks that identifies this engine.
3324 3324
3325 3325 If this engine can be used to compress revlogs, this method should
3326 3326 return the bytes used to identify chunks compressed with this engine.
3327 3327 Else, the method should return ``None`` to indicate it does not
3328 3328 participate in revlog compression.
3329 3329 """
3330 3330 return None
3331 3331
3332 3332 def compressstream(self, it, opts=None):
3333 3333 """Compress an iterator of chunks.
3334 3334
3335 3335 The method receives an iterator (ideally a generator) of chunks of
3336 3336 bytes to be compressed. It returns an iterator (ideally a generator)
3337 3337 of bytes of chunks representing the compressed output.
3338 3338
3339 3339 Optionally accepts an argument defining how to perform compression.
3340 3340 Each engine treats this argument differently.
3341 3341 """
3342 3342 raise NotImplementedError()
3343 3343
3344 3344 def decompressorreader(self, fh):
3345 3345 """Perform decompression on a file object.
3346 3346
3347 3347 Argument is an object with a ``read(size)`` method that returns
3348 3348 compressed data. Return value is an object with a ``read(size)`` that
3349 3349 returns uncompressed data.
3350 3350 """
3351 3351 raise NotImplementedError()
3352 3352
3353 3353 def revlogcompressor(self, opts=None):
3354 3354 """Obtain an object that can be used to compress revlog entries.
3355 3355
3356 3356 The object has a ``compress(data)`` method that compresses binary
3357 3357 data. This method returns compressed binary data or ``None`` if
3358 3358 the data could not be compressed (too small, not compressible, etc).
3359 3359 The returned data should have a header uniquely identifying this
3360 3360 compression format so decompression can be routed to this engine.
3361 3361 This header should be identified by the ``revlogheader()`` return
3362 3362 value.
3363 3363
3364 3364 The object has a ``decompress(data)`` method that decompresses
3365 3365 data. The method will only be called if ``data`` begins with
3366 3366 ``revlogheader()``. The method should return the raw, uncompressed
3367 3367 data or raise a ``RevlogError``.
3368 3368
3369 3369 The object is reusable but is not thread safe.
3370 3370 """
3371 3371 raise NotImplementedError()
3372 3372
3373 3373 class _zlibengine(compressionengine):
3374 3374 def name(self):
3375 3375 return 'zlib'
3376 3376
3377 3377 def bundletype(self):
3378 3378 """zlib compression using the DEFLATE algorithm.
3379 3379
3380 3380 All Mercurial clients should support this format. The compression
3381 3381 algorithm strikes a reasonable balance between compression ratio
3382 3382 and size.
3383 3383 """
3384 3384 return 'gzip', 'GZ'
3385 3385
3386 3386 def wireprotosupport(self):
3387 3387 return compewireprotosupport('zlib', 20, 20)
3388 3388
3389 3389 def revlogheader(self):
3390 3390 return 'x'
3391 3391
3392 3392 def compressstream(self, it, opts=None):
3393 3393 opts = opts or {}
3394 3394
3395 3395 z = zlib.compressobj(opts.get('level', -1))
3396 3396 for chunk in it:
3397 3397 data = z.compress(chunk)
3398 3398 # Not all calls to compress emit data. It is cheaper to inspect
3399 3399 # here than to feed empty chunks through generator.
3400 3400 if data:
3401 3401 yield data
3402 3402
3403 3403 yield z.flush()
3404 3404
3405 3405 def decompressorreader(self, fh):
3406 3406 def gen():
3407 3407 d = zlib.decompressobj()
3408 3408 for chunk in filechunkiter(fh):
3409 3409 while chunk:
3410 3410 # Limit output size to limit memory.
3411 3411 yield d.decompress(chunk, 2 ** 18)
3412 3412 chunk = d.unconsumed_tail
3413 3413
3414 3414 return chunkbuffer(gen())
3415 3415
3416 3416 class zlibrevlogcompressor(object):
3417 3417 def compress(self, data):
3418 3418 insize = len(data)
3419 3419 # Caller handles empty input case.
3420 3420 assert insize > 0
3421 3421
3422 3422 if insize < 44:
3423 3423 return None
3424 3424
3425 3425 elif insize <= 1000000:
3426 3426 compressed = zlib.compress(data)
3427 3427 if len(compressed) < insize:
3428 3428 return compressed
3429 3429 return None
3430 3430
3431 3431 # zlib makes an internal copy of the input buffer, doubling
3432 3432 # memory usage for large inputs. So do streaming compression
3433 3433 # on large inputs.
3434 3434 else:
3435 3435 z = zlib.compressobj()
3436 3436 parts = []
3437 3437 pos = 0
3438 3438 while pos < insize:
3439 3439 pos2 = pos + 2**20
3440 3440 parts.append(z.compress(data[pos:pos2]))
3441 3441 pos = pos2
3442 3442 parts.append(z.flush())
3443 3443
3444 3444 if sum(map(len, parts)) < insize:
3445 3445 return ''.join(parts)
3446 3446 return None
3447 3447
3448 3448 def decompress(self, data):
3449 3449 try:
3450 3450 return zlib.decompress(data)
3451 3451 except zlib.error as e:
3452 3452 raise error.RevlogError(_('revlog decompress error: %s') %
3453 3453 str(e))
3454 3454
3455 3455 def revlogcompressor(self, opts=None):
3456 3456 return self.zlibrevlogcompressor()
3457 3457
3458 3458 compengines.register(_zlibengine())
3459 3459
3460 3460 class _bz2engine(compressionengine):
3461 3461 def name(self):
3462 3462 return 'bz2'
3463 3463
3464 3464 def bundletype(self):
3465 3465 """An algorithm that produces smaller bundles than ``gzip``.
3466 3466
3467 3467 All Mercurial clients should support this format.
3468 3468
3469 3469 This engine will likely produce smaller bundles than ``gzip`` but
3470 3470 will be significantly slower, both during compression and
3471 3471 decompression.
3472 3472
3473 3473 If available, the ``zstd`` engine can yield similar or better
3474 3474 compression at much higher speeds.
3475 3475 """
3476 3476 return 'bzip2', 'BZ'
3477 3477
3478 3478 # We declare a protocol name but don't advertise by default because
3479 3479 # it is slow.
3480 3480 def wireprotosupport(self):
3481 3481 return compewireprotosupport('bzip2', 0, 0)
3482 3482
3483 3483 def compressstream(self, it, opts=None):
3484 3484 opts = opts or {}
3485 3485 z = bz2.BZ2Compressor(opts.get('level', 9))
3486 3486 for chunk in it:
3487 3487 data = z.compress(chunk)
3488 3488 if data:
3489 3489 yield data
3490 3490
3491 3491 yield z.flush()
3492 3492
3493 3493 def decompressorreader(self, fh):
3494 3494 def gen():
3495 3495 d = bz2.BZ2Decompressor()
3496 3496 for chunk in filechunkiter(fh):
3497 3497 yield d.decompress(chunk)
3498 3498
3499 3499 return chunkbuffer(gen())
3500 3500
3501 3501 compengines.register(_bz2engine())
3502 3502
3503 3503 class _truncatedbz2engine(compressionengine):
3504 3504 def name(self):
3505 3505 return 'bz2truncated'
3506 3506
3507 3507 def bundletype(self):
3508 3508 return None, '_truncatedBZ'
3509 3509
3510 3510 # We don't implement compressstream because it is hackily handled elsewhere.
3511 3511
3512 3512 def decompressorreader(self, fh):
3513 3513 def gen():
3514 3514 # The input stream doesn't have the 'BZ' header. So add it back.
3515 3515 d = bz2.BZ2Decompressor()
3516 3516 d.decompress('BZ')
3517 3517 for chunk in filechunkiter(fh):
3518 3518 yield d.decompress(chunk)
3519 3519
3520 3520 return chunkbuffer(gen())
3521 3521
3522 3522 compengines.register(_truncatedbz2engine())
3523 3523
3524 3524 class _noopengine(compressionengine):
3525 3525 def name(self):
3526 3526 return 'none'
3527 3527
3528 3528 def bundletype(self):
3529 3529 """No compression is performed.
3530 3530
3531 3531 Use this compression engine to explicitly disable compression.
3532 3532 """
3533 3533 return 'none', 'UN'
3534 3534
3535 3535 # Clients always support uncompressed payloads. Servers don't because
3536 3536 # unless you are on a fast network, uncompressed payloads can easily
3537 3537 # saturate your network pipe.
3538 3538 def wireprotosupport(self):
3539 3539 return compewireprotosupport('none', 0, 10)
3540 3540
3541 3541 # We don't implement revlogheader because it is handled specially
3542 3542 # in the revlog class.
3543 3543
3544 3544 def compressstream(self, it, opts=None):
3545 3545 return it
3546 3546
3547 3547 def decompressorreader(self, fh):
3548 3548 return fh
3549 3549
3550 3550 class nooprevlogcompressor(object):
3551 3551 def compress(self, data):
3552 3552 return None
3553 3553
3554 3554 def revlogcompressor(self, opts=None):
3555 3555 return self.nooprevlogcompressor()
3556 3556
3557 3557 compengines.register(_noopengine())
3558 3558
3559 3559 class _zstdengine(compressionengine):
3560 3560 def name(self):
3561 3561 return 'zstd'
3562 3562
3563 3563 @propertycache
3564 3564 def _module(self):
3565 3565 # Not all installs have the zstd module available. So defer importing
3566 3566 # until first access.
3567 3567 try:
3568 3568 from . import zstd
3569 3569 # Force delayed import.
3570 3570 zstd.__version__
3571 3571 return zstd
3572 3572 except ImportError:
3573 3573 return None
3574 3574
3575 3575 def available(self):
3576 3576 return bool(self._module)
3577 3577
3578 3578 def bundletype(self):
3579 3579 """A modern compression algorithm that is fast and highly flexible.
3580 3580
3581 3581 Only supported by Mercurial 4.1 and newer clients.
3582 3582
3583 3583 With the default settings, zstd compression is both faster and yields
3584 3584 better compression than ``gzip``. It also frequently yields better
3585 3585 compression than ``bzip2`` while operating at much higher speeds.
3586 3586
3587 3587 If this engine is available and backwards compatibility is not a
3588 3588 concern, it is likely the best available engine.
3589 3589 """
3590 3590 return 'zstd', 'ZS'
3591 3591
3592 3592 def wireprotosupport(self):
3593 3593 return compewireprotosupport('zstd', 50, 50)
3594 3594
3595 3595 def revlogheader(self):
3596 3596 return '\x28'
3597 3597
3598 3598 def compressstream(self, it, opts=None):
3599 3599 opts = opts or {}
3600 3600 # zstd level 3 is almost always significantly faster than zlib
3601 3601 # while providing no worse compression. It strikes a good balance
3602 3602 # between speed and compression.
3603 3603 level = opts.get('level', 3)
3604 3604
3605 3605 zstd = self._module
3606 3606 z = zstd.ZstdCompressor(level=level).compressobj()
3607 3607 for chunk in it:
3608 3608 data = z.compress(chunk)
3609 3609 if data:
3610 3610 yield data
3611 3611
3612 3612 yield z.flush()
3613 3613
3614 3614 def decompressorreader(self, fh):
3615 3615 zstd = self._module
3616 3616 dctx = zstd.ZstdDecompressor()
3617 3617 return chunkbuffer(dctx.read_from(fh))
3618 3618
3619 3619 class zstdrevlogcompressor(object):
3620 3620 def __init__(self, zstd, level=3):
3621 3621 # Writing the content size adds a few bytes to the output. However,
3622 3622 # it allows decompression to be more optimal since we can
3623 3623 # pre-allocate a buffer to hold the result.
3624 3624 self._cctx = zstd.ZstdCompressor(level=level,
3625 3625 write_content_size=True)
3626 3626 self._dctx = zstd.ZstdDecompressor()
3627 3627 self._compinsize = zstd.COMPRESSION_RECOMMENDED_INPUT_SIZE
3628 3628 self._decompinsize = zstd.DECOMPRESSION_RECOMMENDED_INPUT_SIZE
3629 3629
3630 3630 def compress(self, data):
3631 3631 insize = len(data)
3632 3632 # Caller handles empty input case.
3633 3633 assert insize > 0
3634 3634
3635 3635 if insize < 50:
3636 3636 return None
3637 3637
3638 3638 elif insize <= 1000000:
3639 3639 compressed = self._cctx.compress(data)
3640 3640 if len(compressed) < insize:
3641 3641 return compressed
3642 3642 return None
3643 3643 else:
3644 3644 z = self._cctx.compressobj()
3645 3645 chunks = []
3646 3646 pos = 0
3647 3647 while pos < insize:
3648 3648 pos2 = pos + self._compinsize
3649 3649 chunk = z.compress(data[pos:pos2])
3650 3650 if chunk:
3651 3651 chunks.append(chunk)
3652 3652 pos = pos2
3653 3653 chunks.append(z.flush())
3654 3654
3655 3655 if sum(map(len, chunks)) < insize:
3656 3656 return ''.join(chunks)
3657 3657 return None
3658 3658
3659 3659 def decompress(self, data):
3660 3660 insize = len(data)
3661 3661
3662 3662 try:
3663 3663 # This was measured to be faster than other streaming
3664 3664 # decompressors.
3665 3665 dobj = self._dctx.decompressobj()
3666 3666 chunks = []
3667 3667 pos = 0
3668 3668 while pos < insize:
3669 3669 pos2 = pos + self._decompinsize
3670 3670 chunk = dobj.decompress(data[pos:pos2])
3671 3671 if chunk:
3672 3672 chunks.append(chunk)
3673 3673 pos = pos2
3674 3674 # Frame should be exhausted, so no finish() API.
3675 3675
3676 3676 return ''.join(chunks)
3677 3677 except Exception as e:
3678 3678 raise error.RevlogError(_('revlog decompress error: %s') %
3679 3679 str(e))
3680 3680
3681 3681 def revlogcompressor(self, opts=None):
3682 3682 opts = opts or {}
3683 3683 return self.zstdrevlogcompressor(self._module,
3684 3684 level=opts.get('level', 3))
3685 3685
3686 3686 compengines.register(_zstdengine())
3687 3687
3688 3688 def bundlecompressiontopics():
3689 3689 """Obtains a list of available bundle compressions for use in help."""
3690 3690 # help.makeitemsdocs() expects a dict of names to items with a .__doc__.
3691 3691 items = {}
3692 3692
3693 3693 # We need to format the docstring. So use a dummy object/type to hold it
3694 3694 # rather than mutating the original.
3695 3695 class docobject(object):
3696 3696 pass
3697 3697
3698 3698 for name in compengines:
3699 3699 engine = compengines[name]
3700 3700
3701 3701 if not engine.available():
3702 3702 continue
3703 3703
3704 3704 bt = engine.bundletype()
3705 3705 if not bt or not bt[0]:
3706 3706 continue
3707 3707
3708 3708 doc = pycompat.sysstr('``%s``\n %s') % (
3709 3709 bt[0], engine.bundletype.__doc__)
3710 3710
3711 3711 value = docobject()
3712 3712 value.__doc__ = doc
3713 3713
3714 3714 items[bt[0]] = value
3715 3715
3716 3716 return items
3717 3717
3718 3718 # convenient shortcut
3719 3719 dst = debugstacktrace
General Comments 0
You need to be logged in to leave comments. Login now