diff --git a/docs/examples/parallel/wave2D/parallelwave-mpi.py b/docs/examples/parallel/wave2D/parallelwave-mpi.py index b3114b8..20509f0 100755 --- a/docs/examples/parallel/wave2D/parallelwave-mpi.py +++ b/docs/examples/parallel/wave2D/parallelwave-mpi.py @@ -8,8 +8,8 @@ are done by an object of class MPIRectPartitioner2D An example of running the program is (8 processors, 4x2 partition, 400x100 grid cells):: - $ ipclusterz start --profile mpi -n 8 # start 8 engines (assuming mpi profile has been configured) - $ ./parallelwave-mpi.py --grid 400 100 --partition 4 2 --profile mpi + $ ipcluster start --engines=MPIExec -n 8 # start 8 engines with mpiexec + $ python parallelwave-mpi.py --grid 400 100 --partition 4 2 See also parallelwave-mpi, which runs the same program, but uses MPI (via mpi4py) for the inter-engine communication. @@ -179,7 +179,7 @@ if __name__ == '__main__': # run again with numpy vectorized inner-implementation t0 = time.time() - ar = view.apply_async(_solve, tstop, dt=0, verbose=True, final_test=final_test)#, user_action=wave_saver) + ar = view.apply_async(_solve, tstop, dt=0, verbose=True, final_test=final_test, user_action=user_action) if final_test: # this sum is performed element-wise as results finish s = sum(ar) diff --git a/docs/examples/parallel/wave2D/parallelwave.py b/docs/examples/parallel/wave2D/parallelwave.py index cb3e055..c79db11 100755 --- a/docs/examples/parallel/wave2D/parallelwave.py +++ b/docs/examples/parallel/wave2D/parallelwave.py @@ -8,8 +8,8 @@ are done by an object of class ZMQRectPartitioner2D An example of running the program is (8 processors, 4x2 partition, 200x200 grid cells):: - $ ipclusterz start -n 8 # start 8 engines - $ ./parallelwave.py --grid 200 200 --partition 4 2 + $ ipcluster start -n 8 # start 8 engines + $ python parallelwave.py --grid 200 200 --partition 4 2 See also parallelwave-mpi, which runs the same program, but uses MPI (via mpi4py) for the inter-engine communication. @@ -188,7 +188,7 @@ if __name__ == '__main__': t0 = time.time() - ar = view.apply_async(_solve, tstop, dt=0, verbose=True, final_test=final_test)#, user_action=wave_saver) + ar = view.apply_async(_solve, tstop, dt=0, verbose=True, final_test=final_test, user_action=user_action) if final_test: # this sum is performed element-wise as results finish s = sum(ar)