Measuring performance with parsl-perf

parsl-perf is tool for making basic performance measurements of Parsl configurations.

It runs increasingly large numbers of no-op apps until a batch takes (by default) 120 seconds, giving a measurement of tasks per second.

This can give a basic measurement of some of the overheads in task execution.

parsl-perf must be invoked with a configuration file, which is a Python file containing a variable config which contains a Config object, or a function fresh_config which returns a Config object. The fresh_config format is the same as used with the pytest test suite.

To specify a parsl_resource_specification for tasks, add a --resources argument.

To change the target runtime from the default of 120 seconds, add a --time parameter.

For example:

$ python -m parsl.benchmark.perf --config parsl/tests/configs/workqueue_ex.py --resources '{"cores":1, "memory":0, "disk":0}'
==== Iteration 1 ====
Will run 10 tasks to target 120 seconds runtime
Submitting tasks / invoking apps
warning: using plain-text when communicating with workers.
warning: use encryption with a key and cert when creating the manager.
All 10 tasks submitted ... waiting for completion
Submission took 0.008 seconds = 1248.676 tasks/second
Runtime: actual 3.668s vs target 120s
Tasks per second: 2.726

[...]

==== Iteration 4 ====
Will run 57640 tasks to target 120 seconds runtime
Submitting tasks / invoking apps
All 57640 tasks submitted ... waiting for completion
Submission took 34.839 seconds = 1654.487 tasks/second
Runtime: actual 364.387s vs target 120s
Tasks per second: 158.184
Cleaning up DFK
The end