Globus Compute Executor¶
The Executor
class, a subclass of Python’s concurrent.futures.Executor
, is the
preferred approach to collecting results from the Globus Compute web services. Over
polling (the historical approach) where the web service must be repeatedly
queried for the status of tasks and results eventually collected in bulk, the
Executor
class instantiates an AMQPS connection that streams results
directly – and immediately – as they arrive at the server. This is a far
more efficient paradigm, simultaneously in terms of bytes over the wire, time
spent waiting for results, and boilerplate code to check for results.
For most “simple” interactions with Globus Compute, this class will likely be the quickest and easiest avenue to submit tasks and acquire results. An example interaction:
from globus_compute_sdk import Executor
def double(x):
return x * 2
tutorial_endpoint_id = '4b116d3c-1703-4f8f-9f6f-39921e5864df'
with Executor(endpoint_id=tutorial_endpoint_id) as gce:
fut = gce.submit(double, 7)
print(fut.result())
This example is only a quick-reference, showing the basic mechanics of how to
use the Executor
class and submitting a single task. However, there
are a number of details to observe. The first is that a Executor
instance is associated with a specific endpoint. We use the “well-known”
tutorial endpoint in this example, but that can point to any endpoint to which
you have access.
Note
A friendly FYI: the tutorial endpoint is public – available for any (authenticated) user. You are welcome to use it, but please limit the size and number of functions you send to this endpoint. It is hosted on a small VM with limited CPU and memory, intentionally underpowered. Its primary intended purpose is for an introduction to the Globus Compute toolset.
This endpoint has been made public by the Globus Compute team for tutorial use, but endpoints created by users can not be shared publicly.
Second, the waiting – or “blocking” – for a result is automatic. The
.submit()
call returns a Future
immediately; the actual HTTP call to the
Globus Compute web-services will not have occurred yet, and neither will the task even
been executed (remotely), much less a result received. The .result()
call
blocks (“waits”) until all of that has completed, and the result has been
received from the upstream services.
Third, Executor
objects can be used as context managers (the with
statement). Underneath the hood, the Executor
class uses threads to
implement the asynchronous interface – a thread to coalesce and submit tasks,
and a thread to watch for incoming results. The Executor
logic cannot
determine when it will no longer receive tasks (i.e., no more .submit()
calls) and so cannot prematurely shutdown. Thus, it must be told, either
explicitly with a call to .shutdown()
, or implicitly when used as a context
manager.
Multiple Function Invocations¶
Building on the asynchronous behavior of .submit()
, we can easily create
multiple tasks and wait for them all. The simplest case is a for-loop over
submission and results. As an example, consider the Collatz conjecture,
alternatively known as the \(3n + 1\) problem. The conjecture is that
given a starting integer and two generation rules, the outcome sequence will
always end with 1. The rules are:
If \(N_i\) is even, then \(N_{i+1} = N_i / 2\)
If \(N_i\) is odd, then \(N_{i+1} = 3 N_i + 1\)
To verify all of the sequences through 100, one brute-force approach is:
from globus_compute_sdk import Executor
def generate_collatz_sequence(N: int, sequence_limit = 10_000):
seq = [N]
while N != 1 and len(seq) < sequence_limit:
if N % 2:
N = 3 * N + 1
else:
N //= 2 # okay because guaranteed integer result
seq.append(N)
return seq
ep_id = "<your_endpoint_id>"
generate_from = 1
generate_through = 100
futs, results, disproof_candidates = [], [], []
with Executor(endpoint_id=ep_id) as gce:
for n in range(generate_from, generate_through + 1):
futs.append(gce.submit(generate_collatz_sequence, n))
print("Tasks all submitted; waiting for results")
# The futures were appended to the `futs` list in order, so one could wait
# for each result in turn to get a submission-ordered set of results:
for f in futs:
r = f.result()
results.append(r)
if r[-1] != 1:
# of course, given the conjecture, we don't expect this branch
disproof_candidates.append(r[0])
print(f"All sequences generated (from {generate_from} to {generate_through})")
for res in results:
print(res)
if disproof_candidates:
print("Possible conjecture disproving integers:", disproof_candidates)
Checking the Status of a Result¶
Sometimes, it is desirable not to wait for a result, but just to check on the
status. Futures make this simple with the .done()
method:
...
future = gce.submit(generate_collatz_sequence, 1234567890)
# Use the .done() method to check the status of the function without
# blocking; this will return a Bool indicating whether the result is ready
print("Status: ", future.done())
Handling Exceptions¶
Assuming that a future will always have a result will lead to broken scripts.
Exceptions happen, whether from a condition the task function does not handle
or from an external execution error. To robustly handle task exceptions, wrap
.result()
calls in a try
block. The following code has updated the
sequence generator to throw an exception after sequence_limit
steps rather
than summarily return, and the specific number chosen starts a sequence that
takes more than 100 steps to complete.
from globus_compute_sdk import Executor
def generate_collatz_sequence(N: int, sequence_limit=100):
seq = [N]
while N != 1 and len(seq) < sequence_limit:
if N % 2:
N = 3 * N + 1
else:
N //= 2 # okay because guaranteed integer result
seq.append(N)
if N != 1:
raise ValueError(f"Sequence not terminated in {sequence_limit} steps")
return seq
with Executor(endpoint_id=ep_id) as gce:
future = gce.submit(generate_collatz_sequence, 1234567890)
try:
print(future.result())
except Exception as exc:
print(f"Oh no! The task raised an exception: {exc})
Receiving Results Out of Order¶
So far, we’ve shown simple iteration through the list of Futures, but that’s not generally the most performant approach for overall workflow completion. In the previous examples, a result may return early at the end of the list, but the script will not recognize it until it “gets there,” waiting in the meantime for the other tasks to complete. (Task functions are not guaranteed to be scheduled in order, nor are they guaranteed to take the same amount of time to finish.) There are a number of ways to work with results as they arrive; this example uses concurrent.futures.as_completed:
import concurrent.futures
def double(x):
return f"{x} -> {x * 2}"
def slow_double(x):
import random, time
time.sleep(x * random.random())
return f"{x} -> {x * 2}"
with Executor(endpoint_id=endpoint_id) as gce:
futs = [gce.submit(double, i) for i in range(10)]
# The futures were appended to the `futs` list in order, so one could
# wait for each result in turn to get an ordered set:
print("Results:", [f.result() for f in futs])
# But often acting on the results *as they arrive* is more desirable
# as results are NOT guaranteed to arrive in the order they were
# submitted.
#
# NOTA BENE: handling results "as they arrive" must happen before the
# executor is shutdown. Since this executor was used in a `with`
# statement, then to stream results, we must *stay* within the `with`
# statement. Otherwise, at the unindent, `.shutdown()` will be
# implicitly invoked (with default arguments) and the script will not
# continue until *all* of the futures complete.
futs = [gce.submit(slow_double, i) for i in range(10, 20)]
for f in concurrent.futures.as_completed(futs):
print("Received:", f.result())
Reloading Tasks¶
Waiting for incoming results with the Executor
requires an active
connection – which is often at odds with closing a laptop clamshell (e.g.,
heading home for the weekend). For longer running jobs like this, the
Executor
offers the .reload_tasks()
method. This method will reach
out to the Globus Compute web-services to collect all of the tasks associated with the
.task_group_id
, create a list of associated futures, finish
(call .set_result()
) any previously finished tasks, and watch the unfinished
futures. Consider the following (contrived) example:
# execute initially as:
# $ python globus_compute_executor_reload_tasks.py
# ... this Task Group ID: <TG_UUID_STR>
# ...
# Then run with the Task Group ID as an argument:
# $ python globus_compute_executor_reload_tasks.py <TG_UUID_STR>
import os, signal, sys, time, typing as t
from globus_compute_sdk import Executor
from globus_compute_sdk.sdk.executor import ComputeFuture
task_group_id = sys.argv[1] if len(sys.argv) > 1 else None
def task_kernel(num):
return f"your Globus Compute logic result, from task: {num}"
ep_id = "<YOUR_ENDPOINT_UUID>"
with Executor(endpoint_id=ep_id) as gce:
futures: t.Iterable[ComputeFuture]
if task_group_id:
print(f"Reloading tasks from Task Group ID: {task_group_id}")
gce.task_group_id = task_group_id
futures = gce.reload_tasks()
else:
# Save the task_group_id somewhere. Perhaps in a file, or less
# robustly "as mere text" on your console:
print(
"New session; creating Globus Compute tasks; if this script dies, rehydrate"
f" futures with this Task Group ID: {gce.task_group_id}"
)
num_tasks = 5
futures = [gce.submit(task_kernel, i + 1) for i in range(num_tasks)]
# Ensure all tasks have been sent upstream ...
while gce.task_count_submitted < num_tasks:
time.sleep(1)
print(f"Tasks submitted upstream: {gce.task_count_submitted}")
# ... before script death for [silly reason; did you lose power!?]
bname = sys.argv[0]
if sys.argv[0] != sys.orig_argv[0]:
bname = f"{sys.orig_argv[0]} {bname}"
print("Simulating unexpected process death! Now reload the session")
print("by rerunning this script with the task_group_id:\n")
print(f" {bname} {gce.task_group_id}\n")
os.kill(os.getpid(), signal.SIGKILL)
exit(1) # In case KILL takes split-second to process
# Get results:
results, exceptions = [], []
for f in futures:
try:
results.append(f.result(timeout=10))
except Exception as exc:
exceptions.append(exc)
print("Results:\n ", "\n ".join(results))
For a slightly more advanced usage, one could manually submit a batch of tasks
with the Client
, and wait for the results at a future time. Submitting
the results might look like:
from globus_compute_sdk import Client
def expensive_task(task_arg):
import time
time.sleep(3600 * 24) # 24 hours
return "All done!"
ep_id = "<endpoint_id>"
gcc = Client()
print(f"Task Group ID for later reloading: {gcc.session_task_group_id}")
fn_id = gcc.register_function(expensive_task)
batch = gcc.create_batch()
for task_i in range(10):
batch.add(fn_id, ep_id, args=(task_i,))
gcc.batch_run(batch)
And ~24 hours later, could reload the tasks with the executor to continue processing:
from globus_compute_sdk import Executor
ep_id = "<endpoint_id>"
tg_id = "Saved task group id from 'yesterday'"
with Executor(endpoint_id=ep_id, task_group_id=tg_id) as gce:
futures = gce.reload_tasks()
for f in concurrent.futures.as_completed(futs):
print("Received:", f.result())
Shell Functions¶
ShellFunction
is the solution to executing commands remotely using Globus Compute.
The ShellFunction
class allows for the specification of a command string, along with
runtime details such as a run directory, per-task sandboxing, walltime, etc and returns a
ShellResult
. ShellResult
encapsulates the outputs from executing the command line string
by wrapping the returnode and snippets from the standard streams (stdout
and stderr
).
Here’s a basic example that demonstrates specifying a ShellFunction
that is to be
formatted with a list of values at launch time.
from globus_compute_sdk import ShellFunction, Executor
ep_id = "<SPECIFY_ENDPOINT_ID>"
# The cmd will be formatted with kwargs at invocation time
bf = ShellFunction("echo '{message}'")
with Executor(endpoint_id=ep_id) as ex:
for msg in ("hello", "hola", "bonjour"):
future = ex.submit(bf, message=msg)
shell_result = future.result() # ShellFunctions return ShellResults
print(shell_result.stdout)
Executing the above prints:
hello
hola
bonjour
The ShellResult
object captures outputs relevant to simplify debugging when execution
failures. By default, ShellFunction
captures 1,000 lines of stdout and stderr, but this
can be changed via the ShellFunction(snippet_lines)
kwarg.
Shell Results¶
The output from a ShellFunction
is encapsulated in a ShellResult
. Here are the various fields made
available through the ShellResult
:
returncode
: The return code from the execution of the command suppliedstdout
: A snippet of upto the last 1K lines captured from the stdout streamstderr
: A snippet of upto the last 1K lines captures form the stderr streamcmd
: The formatted command string executed on the endpoint
Note
Bear in mind that the snippet lines count toward the 10 MiB payload size limit. The
number of lines captured from stdout
and stderr
can be modified by setting
the snippet_lines keyword argument.
Working Directory¶
Since ShellFunctions operate on files, overwriting files unintentionally is a
possibility. To mitigate this, ShellFunction
recommends enabling sandboxing through
the endpoint configuration. By default the working directory is:
~/.globus_compute/<ENDPOINT_NAME>/tasks_working_dir/
. With sandboxing enabled, each
ShellFunction
executes within a task specific directory named:
~/.globus_compute/<ENDPOINT_NAME>/tasks_working_dir/<TASK_UUID>
.
Here’s an example configuration:
display_name: SandboxExample
engine:
type: GlobusComputeEngine
# Enable sandboxing
run_in_sandbox: True
Walltime¶
The walltime
keyword argument to ShellFunction
can be used to specify the maximum duration (in seconds)
after which execution should be interrupted. If the execution was prematurely terminated due to reaching
the walltime, the return code will be set to 124
, which matches the behavior of the
timeout command.
Here’s an example:
# Limit execution to 1s
bf = ShellFunction("sleep 2", walltime=1)
future = executor.submit(bf)
print(future.returncode)
Executing the above prints:
124
MPI Functions¶
MPIFunction
is the solution for executing MPI applications remotely using Globus Compute.
As an extension to ShellFunction
, the MPIFunction
supports the same interface
to specify the command to invoke on the endpoint as well as the capture of output
streams. However, MPIFunction
diverges from ShellFunction
in the following ways:
An
MPIFunction
requires a specification of the resources for its execution. This specification must be set on the executor on the client-side. Theresource_specification: dict
takes the following options:executor.resource_specification = { 'num_nodes': <int>, # Number of nodes required for the application instance 'ranks_per_node': <int>, # Number of ranks / application elements to be launched per node 'num_ranks': <int>, # Number of ranks in total }
MPIFunction
is designed to be used withGlobusMPIEngine
on the endpoint.GlobusMPIEngine
is required for the partitioning of a batch job (blocks) dynamically based on theresource_specification
of theMPIFunction
.MPIFunction
automatically prefixes the supplied command with$PARSL_MPI_PREFIX
which resolves to an appropriate mpi launcher prefix (for e.g,mpiexec -n 4 -host <NODE1,NODE2>
).
Multi-line commands¶
MPIFunction
allows for multi-line commands, however the MPI launcher prefix is applied only to
the entire command. If multiple commands need to be launched, explicitly use the $PARSL_MPI_PREFIX
.
Here’s an example:
MPIFunction("""true; # force the default prefix to launch a no-op
$PARSL_MPI_PREFIX <command_1>
$PARSL_MPI_PREFIX <command_2>
""")
Here is an example configuration for an HPC system that uses PBSPro scheduler:
# Example configuration for a PBSPro based HPC system
display_name: PBSProHPC
engine:
type: GlobusMPIEngine
mpi_launcher: mpiexec
provider:
type: PBSProProvider
# Specify # of nodes per batch job that will be
# shared by multiple MPIFunctions
nodes_per_block: 4
launcher:
type: SimpleLauncher
Here’s another trimmed example for an HPC system that uses Slurm as the scheduler:
# Example configuration for a Slurm based HPC system
display_name: SlurmHPC
engine:
type: GlobusMPIEngine
mpi_launcher: srun
provider:
type: SlurmProvider
launcher:
type: SimpleLauncher
# Specify # of nodes per batch job that will be
# shared by multiple MPIFunctions
nodes_per_block: 4
from globus_compute_sdk import Executor, MPIFunction
ep_id = "<SPECIFY_ENDPOINT_ID>"
func = MPIFunction("hostname")
with Executor(endpoint_id=ep_id) as ex:
for nodes in range(1, 4): # reminder: 1, 2, 3; 4 not included
ex.resource_specification = {
"num_nodes": 2,
"ranks_per_node": nodes
}
fu = ex.submit(func)
mpi_result = fu.result()
print(mpi_result.stdout)
Expect output similar to:
exp-14-08
exp-14-20
exp-14-08
exp-14-20
exp-14-08
exp-14-20
exp-14-08
exp-14-20
exp-14-08
exp-14-08
exp-14-20
exp-14-20
The ShellResult
object captures outputs relevant to simplify debugging when execution
failures. By default, MPIFunction
captures 1,000 lines of stdout and stderr, but this
can be changed via the MPIFunction(snippet_lines:int = <NUM_LINES>)
kwarg.
Results¶
MPIFunction
encapsulates its output in a ShellResult
. Please refer to
our documentation section on shell results for more information.
AMQP Port¶
As of v2.11.0, newly configured endpoints use AMQP over port 443 by default, since firewall rules usually allow outbound HTTPS.
The port that the Executor uses to fetch task results remains the default AMQP
5671, but can be overridden via the amqp_port
parameter during
instantiation. This may be necessary in cases where outbound 5671 is also
unavailable in the task submission environment, such as some cloud VMs
like BinderHub:
>>> from globus_compute_sdk import Executor
>>> gce = Executor(amqp_port=443)