##// END OF EJS Templates
revlog: drop an unused variable assignment...
revlog: drop an unused variable assignment It's assigned again 2 lines later.

File last commit:

r50204:d54b213c default
r50428:8d6c8a9a default
Show More
worker.py
455 lines | 15.4 KiB | text/x-python | PythonLexer
Bryan O'Sullivan
worker: count the number of CPUs...
r18635 # worker.py - master-slave parallelism support
#
# Copyright 2013 Facebook, Inc.
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
Gregory Szorc
worker: use absolute_import
r25992
import os
Gregory Szorc
py3: use pickle directly...
r49725 import pickle
Manuel Jacob
thirdparty: remove Python 2-specific selectors2 copy...
r50175 import selectors
Gregory Szorc
worker: use absolute_import
r25992 import signal
import sys
Wojciech Lis
workers: implemented worker on windows...
r35427 import threading
Wojciech Lis
worker: make windows workers daemons...
r35448 import time
Gregory Szorc
worker: use absolute_import
r25992
from .i18n import _
Jun Wu
worker: migrate to util.iterfile
r30396 from . import (
Pulkit Goyal
py3: replace os.environ with encoding.environ (part 2 of 5)
r30635 encoding,
Jun Wu
worker: migrate to util.iterfile
r30396 error,
Pulkit Goyal
py3: replace os.name with pycompat.osname (part 1 of 2)...
r30639 pycompat,
Jun Wu
worker: use os._exit for posix worker in all cases...
r30521 scmutil,
Jun Wu
worker: migrate to util.iterfile
r30396 )
Bryan O'Sullivan
worker: count the number of CPUs...
r18635
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: count the number of CPUs...
r18635 def countcpus():
'''try to count the number of CPUs on the system'''
Gregory Szorc
worker: restore old countcpus code (issue4869)...
r26568
# posix
Bryan O'Sullivan
worker: count the number of CPUs...
r18635 try:
Augie Fackler
cleanup: remove pointless r-prefixes on single-quoted strings...
r43906 n = int(os.sysconf('SC_NPROCESSORS_ONLN'))
Gregory Szorc
worker: restore old countcpus code (issue4869)...
r26568 if n > 0:
return n
except (AttributeError, ValueError):
pass
# windows
try:
Augie Fackler
formatting: byteify all mercurial/ and hgext/ string literals...
r43347 n = int(encoding.environ[b'NUMBER_OF_PROCESSORS'])
Gregory Szorc
worker: restore old countcpus code (issue4869)...
r26568 if n > 0:
return n
except (KeyError, ValueError):
pass
return 1
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 def _numworkers(ui):
Augie Fackler
formatting: byteify all mercurial/ and hgext/ string literals...
r43347 s = ui.config(b'worker', b'numcpus')
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 if s:
try:
n = int(s)
if n >= 1:
return n
except ValueError:
Augie Fackler
formatting: byteify all mercurial/ and hgext/ string literals...
r43347 raise error.Abort(_(b'number of cpus must be an integer'))
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 return min(max(countcpus(), 4), 32)
Augie Fackler
formatting: blacken the codebase...
r43346
Gregory Szorc
worker: remove Python 2 support code...
r49756 def ismainthread():
return threading.current_thread() == threading.main_thread()
Jan Alexander Steffens (heftig)
worker: manually buffer reads from pickle stream...
r44751
Manuel Jacob
branching: fix wrong merge conflict resolution from 13dfad0f9f7a...
r50119 class _blockingreader:
Manuel Jacob
worker: add docstring to _blockingreader
r50126 """Wrap unbuffered stream such that pickle.load() works with it.
pickle.load() expects that calls to read() and readinto() read as many
bytes as requested. On EOF, it is fine to read fewer bytes. In this case,
pickle.load() raises an EOFError.
"""
Raphaël Gomès
branching: merge stable into default
r50112 def __init__(self, wrapped):
self._wrapped = wrapped
Matt Harbison
worker: adapt _blockingreader to work around a python3.8.[0-1] bug (issue6444)...
r50103
Raphaël Gomès
branching: merge stable into default
r50112 def readline(self):
return self._wrapped.readline()
Manuel Jacob
worker: implement _blockingreader.readinto() (issue6444)...
r50128 def readinto(self, buf):
Raphaël Gomès
branching: merge stable into default
r50112 pos = 0
Manuel Jacob
worker: implement _blockingreader.readinto() (issue6444)...
r50128 size = len(buf)
Raphaël Gomès
branching: merge stable into default
r50112
Manuel Jacob
worker: stop relying on garbage collection to release memoryview...
r50127 with memoryview(buf) as view:
while pos < size:
with view[pos:] as subview:
ret = self._wrapped.readinto(subview)
if not ret:
break
pos += ret
Raphaël Gomès
branching: merge stable into default
r50112
Manuel Jacob
worker: implement _blockingreader.readinto() (issue6444)...
r50128 return pos
# issue multiple reads until size is fulfilled (or EOF is encountered)
def read(self, size=-1):
if size < 0:
return self._wrapped.readall()
buf = bytearray(size)
n_read = self.readinto(buf)
del buf[n_read:]
Raphaël Gomès
branching: merge stable into default
r50112 return bytes(buf)
Jan Alexander Steffens (heftig)
worker: manually buffer reads from pickle stream...
r44751
Wojciech Lis
workers: implemented worker on windows...
r35427 if pycompat.isposix or pycompat.iswindows:
Gregory Szorc
worker: rename variable to reflect constant...
r38753 _STARTUP_COST = 0.01
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754 # The Windows worker is thread based. If tasks are CPU bound, threads
# in the presence of the GIL result in excessive context switching and
# this overhead can slow down execution.
_DISALLOW_THREAD_UNSAFE = pycompat.iswindows
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 else:
Gregory Szorc
worker: rename variable to reflect constant...
r38753 _STARTUP_COST = 1e30
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754 _DISALLOW_THREAD_UNSAFE = False
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636
Augie Fackler
formatting: blacken the codebase...
r43346
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754 def worthwhile(ui, costperop, nops, threadsafe=True):
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """try to determine whether the benefit of multiple processes can
outweigh the cost of starting them"""
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754
if not threadsafe and _DISALLOW_THREAD_UNSAFE:
return False
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 linear = costperop * nops
workers = _numworkers(ui)
Gregory Szorc
worker: rename variable to reflect constant...
r38753 benefit = linear - (_STARTUP_COST * workers + linear / workers)
Bryan O'Sullivan
worker: estimate whether it's worth running a task in parallel...
r18636 return benefit >= 0.15
Bryan O'Sullivan
worker: partition a list (of tasks) into equal-sized chunks
r18637
Augie Fackler
formatting: blacken the codebase...
r43346
def worker(
ui, costperarg, func, staticargs, args, hasretval=False, threadsafe=True
):
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """run a function, possibly in parallel in multiple worker
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 processes.
returns a progress iterator
costperarg - cost of a single task
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 func - function to run. It is expected to return a progress iterator.
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638
staticargs - arguments to pass to every invocation of the function
args - arguments to split into chunks, to pass to individual
workers
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 hasretval - when True, func and the current function return an progress
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 iterator then a dict (encoded as an iterator that yield many (False, ..)
then a (True, dict)). The dicts are joined in some arbitrary order, so
overlapping keys are a bad idea.
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754 threadsafe - whether work items are thread safe and can be executed using
a thread-based worker. Should be disabled for CPU heavy tasks that don't
release the GIL.
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """
Augie Fackler
formatting: byteify all mercurial/ and hgext/ string literals...
r43347 enabled = ui.configbool(b'worker', b'enabled')
Joerg Sonnenberger
worker: POSIX only supports workers from main thread (issue6460)...
r46857 if enabled and _platformworker is _posixworker and not ismainthread():
# The POSIX worker has to install a handler for SIGCHLD.
# Python up to 3.9 only allows this in the main thread.
enabled = False
Gregory Szorc
worker: ability to disable thread unsafe tasks...
r38754 if enabled and worthwhile(ui, costperarg, len(args), threadsafe=threadsafe):
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 return _platformworker(ui, func, staticargs, args, hasretval)
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 return func(*staticargs + (args,))
Augie Fackler
formatting: blacken the codebase...
r43346
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 def _posixworker(ui, func, staticargs, args, hasretval):
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 workers = _numworkers(ui)
Bryan O'Sullivan
worker: fix a race in SIGINT handling...
r18708 oldhandler = signal.getsignal(signal.SIGINT)
signal.signal(signal.SIGINT, signal.SIG_IGN)
Jun Wu
worker: change "pids" to a set...
r30413 pids, problem = set(), [0]
Augie Fackler
formatting: blacken the codebase...
r43346
Jun Wu
worker: move killworkers and waitforworkers up...
r30410 def killworkers():
Yuya Nishihara
worker: make sure killworkers() never be interrupted by another SIGCHLD...
r30423 # unregister SIGCHLD handler as all children will be killed. This
# function shouldn't be interrupted by another SIGCHLD; otherwise pids
# could be updated while iterating, which would cause inconsistency.
signal.signal(signal.SIGCHLD, oldchldhandler)
Jun Wu
worker: move killworkers and waitforworkers up...
r30410 # if one worker bails, there's no good reason to wait for the rest
for p in pids:
try:
os.kill(p, signal.SIGTERM)
Manuel Jacob
py3: catch ProcessLookupError instead of checking errno == ESRCH
r50204 except ProcessLookupError:
pass
Augie Fackler
formatting: blacken the codebase...
r43346
Jun Wu
worker: allow waitforworkers to be non-blocking...
r30412 def waitforworkers(blocking=True):
Jun Wu
worker: make waitforworkers reentrant...
r30414 for pid in pids.copy():
p = st = 0
Manuel Jacob
py3: remove retry on EINTR errno...
r50197 try:
p, st = os.waitpid(pid, (0 if blocking else os.WNOHANG))
Manuel Jacob
py3: catch ChildProcessError instead of checking errno == ECHILD
r50198 except ChildProcessError:
# child would already be reaped, but pids yet been
# updated (maybe interrupted just after waitpid)
pids.discard(pid)
FUJIWARA Katsunori
worker: ignore meaningless exit status indication returned by os.waitpid()...
r31063 if not p:
# skip subsequent steps, because child process should
# be still running in this case
continue
pids.discard(p)
st = _exitstatus(st)
Jun Wu
worker: move killworkers and waitforworkers up...
r30410 if st and not problem[0]:
problem[0] = st
Augie Fackler
formatting: blacken the codebase...
r43346
Jun Wu
worker: add a SIGCHLD handler to collect worker immediately...
r30415 def sigchldhandler(signum, frame):
waitforworkers(blocking=False)
Yuya Nishihara
worker: kill workers after all zombie processes are reaped...
r30424 if problem[0]:
killworkers()
Augie Fackler
formatting: blacken the codebase...
r43346
Jun Wu
worker: add a SIGCHLD handler to collect worker immediately...
r30415 oldchldhandler = signal.signal(signal.SIGCHLD, sigchldhandler)
David Soria Parra
worker: flush ui buffers before running the worker...
r31696 ui.flush()
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 parentpid = os.getpid()
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 pipes = []
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 retval = {}
Martin von Zweigbergk
posixworker: avoid creating workers that end up getting no work...
r45936 for pargs in partition(args, min(workers, len(args))):
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 # Every worker gets its own pipe to send results on, so we don't have to
# implement atomic writes larger than PIPE_BUF. Each forked process has
# its own pipe's descriptors in the local variables, and the parent
# process has the full list of pipe descriptors (and it doesn't really
# care what order they're in).
rfd, wfd = os.pipe()
pipes.append((rfd, wfd))
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 # make sure we use os._exit in all worker code paths. otherwise the
# worker may do some clean-ups which could cause surprises like
# deadlock. see sshpeer.cleanup for example.
# override error handling *before* fork. this is necessary because
# exception (signal) may arrive after fork, before "pid =" assignment
# completes, and other exception handler (dispatch.py) can lead to
# unexpected code path without os._exit.
ret = -1
try:
pid = os.fork()
if pid == 0:
signal.signal(signal.SIGINT, oldhandler)
signal.signal(signal.SIGCHLD, oldchldhandler)
Jun Wu
worker: use os._exit for posix worker in all cases...
r30521
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 def workerfunc():
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 for r, w in pipes[:-1]:
os.close(r)
os.close(w)
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 os.close(rfd)
Manuel Jacob
worker: avoid potential partial write of pickled data...
r50164 with os.fdopen(wfd, 'wb') as wf:
for result in func(*(staticargs + (pargs,))):
pickle.dump(result, wf)
wf.flush()
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 return 0
ret = scmutil.callcatch(ui, workerfunc)
Augie Fackler
formatting: blacken the codebase...
r43346 except: # parent re-raises, child never returns
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 if os.getpid() == parentpid:
raise
exctype = sys.exc_info()[0]
force = not issubclass(exctype, KeyboardInterrupt)
ui.traceback(force=force)
finally:
if os.getpid() != parentpid:
Yuya Nishihara
worker: flush messages written by child processes before exit...
r31118 try:
ui.flush()
Augie Fackler
formatting: blacken the codebase...
r43346 except: # never returns, no re-raises
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 pass
Jun Wu
worker: use os._exit for posix worker in all cases...
r30521 finally:
Jun Wu
worker: rewrite error handling so os._exit covers all cases...
r32112 os._exit(ret & 255)
Jun Wu
worker: change "pids" to a set...
r30413 pids.add(pid)
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 selector = selectors.DefaultSelector()
for rfd, wfd in pipes:
os.close(wfd)
Manuel Jacob
worker: explain why pickle reading stream has to be unbuffered
r50125 # The stream has to be unbuffered. Otherwise, if all data is read from
# the raw file into the buffer, the selector thinks that the FD is not
# ready to read while pickle.load() could read from the buffer. This
# would delay the processing of readable items.
Jan Alexander Steffens (heftig)
worker: manually buffer reads from pickle stream...
r44751 selector.register(os.fdopen(rfd, 'rb', 0), selectors.EVENT_READ)
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 def cleanup():
signal.signal(signal.SIGINT, oldhandler)
Jun Wu
worker: stop using a separate thread waiting for children...
r30416 waitforworkers()
Jun Wu
worker: add a SIGCHLD handler to collect worker immediately...
r30415 signal.signal(signal.SIGCHLD, oldchldhandler)
Yuya Nishihara
worker: call selector.close() to release polling resources
r38763 selector.close()
Yuya Nishihara
worker: do not swallow exception occurred in main process...
r41024 return problem[0]
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 try:
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 openpipes = len(pipes)
while openpipes > 0:
for key, events in selector.select():
try:
Gregory Szorc
worker: silence type error when calling pickle...
r49767 # The pytype error likely goes away on a modern version of
# pytype having a modern typeshed snapshot.
# pytype: disable=wrong-arg-types
Gregory Szorc
py3: use pickle directly...
r49725 res = pickle.load(_blockingreader(key.fileobj))
Gregory Szorc
worker: silence type error when calling pickle...
r49767 # pytype: enable=wrong-arg-types
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 if hasretval and res[0]:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 retval.update(res[1])
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 else:
yield res
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 except EOFError:
selector.unregister(key.fileobj)
Manuel Jacob
thirdparty: remove Python 2-specific selectors2 copy...
r50175 # pytype: disable=attribute-error
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 key.fileobj.close()
Manuel Jacob
thirdparty: remove Python 2-specific selectors2 copy...
r50175 # pytype: enable=attribute-error
Danny Hooper
worker: use one pipe per posix worker and select() in parent process...
r38752 openpipes -= 1
Augie Fackler
formatting: blacken the codebase...
r43346 except: # re-raises
Bryan O'Sullivan
worker: handle worker failures more aggressively...
r18709 killworkers()
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 cleanup()
raise
Yuya Nishihara
worker: do not swallow exception occurred in main process...
r41024 status = cleanup()
if status:
if status < 0:
os.kill(os.getpid(), -status)
Martin von Zweigbergk
worker: raise exception instead of calling sys.exit() with child's code...
r46429 raise error.WorkerError(status)
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 if hasretval:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 yield True, retval
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: on error, exit similarly to the first failing worker...
r18707 def _posixexitstatus(code):
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """convert a posix exit status into the same form returned by
Bryan O'Sullivan
worker: on error, exit similarly to the first failing worker...
r18707 os.spawnv
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 returns None if the process was stopped instead of exiting"""
Bryan O'Sullivan
worker: on error, exit similarly to the first failing worker...
r18707 if os.WIFEXITED(code):
return os.WEXITSTATUS(code)
elif os.WIFSIGNALED(code):
Augie Fackler
formatting: blacken the codebase...
r43346 return -(os.WTERMSIG(code))
Bryan O'Sullivan
worker: on error, exit similarly to the first failing worker...
r18707
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 def _windowsworker(ui, func, staticargs, args, hasretval):
Wojciech Lis
workers: implemented worker on windows...
r35427 class Worker(threading.Thread):
Augie Fackler
formatting: blacken the codebase...
r43346 def __init__(
self, taskqueue, resultqueue, func, staticargs, *args, **kwargs
):
Matt Harbison
py3: roll up threading.Thread constructor args into **kwargs...
r40475 threading.Thread.__init__(self, *args, **kwargs)
Wojciech Lis
workers: implemented worker on windows...
r35427 self._taskqueue = taskqueue
self._resultqueue = resultqueue
self._func = func
self._staticargs = staticargs
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 self._interrupted = False
Wojciech Lis
worker: make windows workers daemons...
r35448 self.daemon = True
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 self.exception = None
def interrupt(self):
self._interrupted = True
Wojciech Lis
workers: implemented worker on windows...
r35427
def run(self):
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 try:
while not self._taskqueue.empty():
try:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 args = self._taskqueue.get_nowait()
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 for res in self._func(*self._staticargs + (args,)):
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 self._resultqueue.put(res)
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 # threading doesn't provide a native way to
# interrupt execution. handle it manually at every
# iteration.
if self._interrupted:
return
Gregory Szorc
pycompat: export queue module instead of symbols in module (API)...
r37863 except pycompat.queue.Empty:
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 break
except Exception as e:
# store the exception such that the main thread can resurface
# it as if the func was running without workers.
self.exception = e
raise
threads = []
Augie Fackler
formatting: blacken the codebase...
r43346
Wojciech Lis
worker: make windows workers daemons...
r35448 def trykillworkers():
# Allow up to 1 second to clean worker threads nicely
cleanupend = time.time() + 1
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 for t in threads:
t.interrupt()
for t in threads:
Wojciech Lis
worker: make windows workers daemons...
r35448 remainingtime = cleanupend - time.time()
t.join(remainingtime)
Wojciech Lis
workers: handling exceptions in windows workers...
r35428 if t.is_alive():
Wojciech Lis
worker: make windows workers daemons...
r35448 # pass over the workers joining failure. it is more
# important to surface the inital exception than the
# fact that one of workers may be processing a large
# task and does not get to handle the interruption.
Augie Fackler
formatting: blacken the codebase...
r43346 ui.warn(
_(
Augie Fackler
formatting: byteify all mercurial/ and hgext/ string literals...
r43347 b"failed to kill worker threads while "
b"handling an exception\n"
Augie Fackler
formatting: blacken the codebase...
r43346 )
)
Wojciech Lis
worker: make windows workers daemons...
r35448 return
Wojciech Lis
workers: implemented worker on windows...
r35427
workers = _numworkers(ui)
Gregory Szorc
pycompat: export queue module instead of symbols in module (API)...
r37863 resultqueue = pycompat.queue.Queue()
taskqueue = pycompat.queue.Queue()
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 retval = {}
Wojciech Lis
workers: implemented worker on windows...
r35427 # partition work to more pieces than workers to minimize the chance
# of uneven distribution of large tasks between the workers
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 for pargs in partition(args, workers * 20):
Wojciech Lis
workers: implemented worker on windows...
r35427 taskqueue.put(pargs)
for _i in range(workers):
t = Worker(taskqueue, resultqueue, func, staticargs)
threads.append(t)
t.start()
Wojciech Lis
worker: make windows workers daemons...
r35448 try:
while len(threads) > 0:
while not resultqueue.empty():
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 res = resultqueue.get()
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 if hasretval and res[0]:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 retval.update(res[1])
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 else:
yield res
Wojciech Lis
worker: make windows workers daemons...
r35448 threads[0].join(0.05)
finishedthreads = [_t for _t in threads if not _t.is_alive()]
for t in finishedthreads:
if t.exception is not None:
raise t.exception
threads.remove(t)
Augie Fackler
formatting: blacken the codebase...
r43346 except (Exception, KeyboardInterrupt): # re-raises
Wojciech Lis
worker: make windows workers daemons...
r35448 trykillworkers()
raise
Wojciech Lis
workers: implemented worker on windows...
r35427 while not resultqueue.empty():
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 res = resultqueue.get()
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 if hasretval and res[0]:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 retval.update(res[1])
Valentin Gatien-Baron
worker: support parallelization of functions with return values...
r42655 else:
yield res
if hasretval:
Valentin Gatien-Baron
update: fix spurious unclean status bug shown by previous commit...
r42722 yield True, retval
Wojciech Lis
workers: implemented worker on windows...
r35427
Augie Fackler
formatting: blacken the codebase...
r43346
Wojciech Lis
workers: implemented worker on windows...
r35427 if pycompat.iswindows:
_platformworker = _windowsworker
else:
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638 _platformworker = _posixworker
Bryan O'Sullivan
worker: on error, exit similarly to the first failing worker...
r18707 _exitstatus = _posixexitstatus
Bryan O'Sullivan
worker: allow a function to be run in multiple worker processes...
r18638
Augie Fackler
formatting: blacken the codebase...
r43346
Bryan O'Sullivan
worker: partition a list (of tasks) into equal-sized chunks
r18637 def partition(lst, nslices):
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """partition a list into N slices of roughly equal size
Gregory Szorc
worker: change partition strategy to every Nth element...
r28181
The current strategy takes every Nth element from the input. If
we ever write workers that need to preserve grouping in input
we should consider allowing callers to specify a partition strategy.
Gregory Szorc
worker: document poor partitioning scheme impact...
r28292
Raphaël Gomès
contributor: change mentions of mpm to olivia...
r47575 olivia is not a fan of this partitioning strategy when files are involved.
Gregory Szorc
worker: document poor partitioning scheme impact...
r28292 In his words:
Single-threaded Mercurial makes a point of creating and visiting
files in a fixed order (alphabetical). When creating files in order,
a typical filesystem is likely to allocate them on nearby regions on
disk. Thus, when revisiting in the same order, locality is maximized
and various forms of OS and disk-level caching and read-ahead get a
chance to work.
This effect can be quite significant on spinning disks. I discovered it
circa Mercurial v0.4 when revlogs were named by hashes of filenames.
Tarring a repo and copying it to another disk effectively randomized
the revlog ordering on disk by sorting the revlogs by hash and suddenly
performance of my kernel checkout benchmark dropped by ~10x because the
"working set" of sectors visited no longer fit in the drive's cache and
the workload switched from streaming to random I/O.
What we should really be doing is have workers read filenames from a
ordered queue. This preserves locality and also keeps any worker from
getting more than one file out of balance.
Augie Fackler
formating: upgrade to black 20.8b1...
r46554 """
Gregory Szorc
worker: change partition strategy to every Nth element...
r28181 for i in range(nslices):
yield lst[i::nslices]