This library isn't ready for release yet. Feedback welcome!
Trio is a friendly Python library for async concurrency and networking. Trimeter is a simple but powerful job scheduler for programs using Trio, released under your choice of the MIT or Apache 2 licenses.
Trimeter's core purpose is to make it easy to execute lots tasks concurrently, with rich options to control the degree of concurrency and to collect the task results.
Say you have 1000 urls that you want to fetch and process somehow:
# Old slow way
for url in urls:
await fetch_and_process(url)
That's slow, so you want to do several at the same time... but to avoid overloading the network, you want to limit it to at most 5 calls at once. Oh, and there's a request quota, so we have to throttle it down to 1 per second. No problem:
# New and fancy way
await trimeter.run_on_each(
fetch_and_process, urls, max_at_once=5, max_per_second=1
)
What if we don't know the whole list of urls up front? No worries, just pass in an async iterable instead, and Trimeter will do the right thing.
What if we want to get the result from each call as it finishes, so we
can do something further with it? Just use amap
(= short for
"async map"):
async with trimeter.amap(fetch_and_process, urls, ...) as results:
# Then iterate over the return values, as they become available
# (i.e., not necessarily in the original order)
async for result in results:
...
Of course amap
also accepts throttling options like
max_at_once
, max_per_second
, etc.
What if we want to use the outcome library to capture exceptions, so one call crashing doesn't terminate the whole program? And also, we want to pass through the original url alongside each result, so we know which result goes with which url?
async with trimeter.amap(
fetch_and_process,
urls,
capture_outcome=True,
include_value=True,
) as outcomes:
# Then iterate over the return values, as they become available
# (i.e., not necessarily in the original order)
async for url, outcome in outcomes:
try:
return_value = outcome.unwrap()
except Exception as exc:
print(f"error while processing {url}: {exc!r}")
What if we just want to call a few functions in parallel and then get the results as a list, like asyncio.gather or Promise.all?
return_values = await trimeter.run_all([
async_fn1,
async_fn2,
functools.partial(async_fn3, extra_arg, kwarg="yeah"),
])
Of course, this takes all the same options as the other functions, so
you can control the degree of parallelism, use capture_outcome
to
capture exceptions, and so forth.
For more details, see the fine manual.
Iambic trimeter? No problem:
Contributors are requested to follow our code of conduct in all project spaces.