prefect icon indicating copy to clipboard operation
prefect copied to clipboard

Running the same subflow concurrently multiple times raises `RuntimeError(

Open marvin-robot opened this issue 2 years ago • 10 comments

Opened from the Prefect Public Slack Community

marcin.grzybowski: Hello again. How can I run the same Flow parallelly? I have used code from https://discourse.prefect.io/t/how-can-i-run-multiple-subflows-or-child-flows-in-parallel/96 and it works.

But when I simplify it and modify to run same Flow (not different 4) then I get

RuntimeError("The task runner is already started!")

Modified code below:

import asyncio
from prefect import flow

@flow
async def subflow_1():
    print("Subflow 1 started!")
    await asyncio.sleep(1)


@flow
async def main_flow():
    parallel_subflows = [subflow_1(), subflow_1()]
    await asyncio.gather(*parallel_subflows)

if __name__ == "__main__":
    main_flow_state = asyncio.run(main_flow())

anna: I can reproduce the issue and I can confirm that it's either a bug or that the tutorial must be updated if there is another way of doing that.

But I'd like to know why would you want to do it this way? Do you really want to run the same subflow twice in parallel with the same parameters? I'm curious what is your use case for that?

As a temporary solution, turning that into a task instead of a subflow will work

marcin.grzybowski: Nah, i would like to run the same flow with different parameters

anna: can you explain your use case a bit more? are those subflows containing a lot of tasks?

if you would build those as tasks, then the issue becomes much simpler, but I can understand why subflows may sometimes be better

marcin.grzybowski: Yep, I'm just checking if we can achieve our goals - and as mentioned in other thread: https://prefect-community.slack.com/archives/C03D12VV4NN/p1654007524220189?thread_ts=1654004474.016759&cid=C03D12VV4NN we will probably need to use combination of nested Flows/Tasks to make it possible that we have reusable code that we can track on graph with desired granularity. If it won't be possible maybe logs will be enough for us, but then there is this problem with logger for DaskTaskRunner :wink:

So I'm just checking what's possible and what is not

marcin.grzybowski: Possibility of seeing details of flow on low level is really nice, but for that, as I understand, I need to use Flow->Taks->Flow->Task... combination

anna: not necessarily - the problem you see here is only when running a single subflow multiple times in parallel - if you don't run those in parallel, it works fine - this works and satisfies your modularity use case:

import asyncio
from prefect import flow

@flow 
async def subflow_1():
    print("Subflow 1 started!")
    await asyncio.sleep(1)


@flow 
async def main_flow():
    for _ in range(5):
        await subflow_1()

if __name__ == "__main__":
    main_flow_state = asyncio.run(main_flow())

anna: <@ULVA73B9P> open "Running the same subflow concurrently multiple times raises RuntimeError("The task runner is already started!")"

Original thread can be found here.

marvin-robot avatar May 31 '22 18:05 marvin-robot

The problem raised by the user:

  • they need modularity, i.e. both tasks and subflows for organization/bookkeeping of work in their flows
  • they sometimes want to start the same subflow multiple times in parallel, but with different parameter values

Doing this sequentially works, only doing this concurrently doesn't

anna-geller avatar May 31 '22 18:05 anna-geller

from the user - deepcopy works:

import asyncio
import copy
from prefect import flow

@flow
async def subflow_1(a):
    print("Subflow 1 started!" + a)
    await asyncio.sleep(1)


@flow
async def main_flow():
    parallel_subflows = [subflow_1('a'), copy.deepcopy(subflow_1)('b')]
    await asyncio.gather(*parallel_subflows)

if __name__ == "__main__":
    main_flow_state = asyncio.run(main_flow())

anna-geller avatar May 31 '22 19:05 anna-geller

unfortunately it works only for flows without tasks... Adding task causes error

import asyncio
import copy
from prefect import flow, get_run_logger, task

@task
def task1(a):
    get_run_logger().info("task 1 started!" + a)

@flow
async def subflow_1(a):
    get_run_logger().info("Subflow 1 started!" + a)
    task1(a)
    await asyncio.sleep(1)


@flow
async def main_flow():
    params = ['a', 'b', 'c']
    parallel_subflows = [copy.deepcopy(subflow_1)(param) for param in params]
    await asyncio.gather(*parallel_subflows)

if __name__ == "__main__":
    main_flow_state = asyncio.run(main_flow())
Traceback (most recent call last):
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 1819, in _execute_context
    self.dialect.do_execute(
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/engine/default.py", line 732, in do_execute
    cursor.execute(statement, parameters)
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/dialects/sqlite/aiosqlite.py", line 100, in execute
    self._adapt_connection._handle_exception(error)
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/dialects/sqlite/aiosqlite.py", line 229, in _handle_exception
    raise error
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/dialects/sqlite/aiosqlite.py", line 82, in execute
    self.await_(_cursor.execute(operation, parameters))
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 68, in await_only
    return current.driver.switch(awaitable)
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 121, in greenlet_spawn
    value = await result
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/aiosqlite/cursor.py", line 37, in execute
    await self._execute(self._cursor.execute, sql, parameters)
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/aiosqlite/cursor.py", line 31, in _execute
    return await self._conn._execute(fn, *args, **kwargs)
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/aiosqlite/core.py", line 129, in _execute
    return await future
  File "/mnt/h/projects/prefect2.0/lib/python3.9/site-packages/aiosqlite/core.py", line 102, in run
    result = function()
sqlite3.OperationalError: database is locked

thesubneo avatar Jun 01 '22 08:06 thesubneo

I'm having a similar problem when trying to run a flow that calls the same subflow multiple times, which itself calls tasks. The outer flow generates and loops over a list of URLs, calling a subflow on each one that checks a cache, and if it's not found, fetches the URL and saves the result (each of those steps being a task).

I have Orion connected to Postgres so I don't get the SQLite issue shown above, but I do get errors like:

RuntimeError: Task <Task pending name='anyio.from_thread.BlockingPortal._call_func' coro=<BlockingPortal._call_func() running at /home/ubuntu/.local/lib/python3.8/site-packages/anyio/from_thread.py:187> cb=[TaskGroup._spawn.<locals>.task_done() at /home/ubuntu/.local/lib/python3.8/site-packages/anyio/_backends/_asyncio.py:629]> got Future <Future pending> attached to a different loop

This can be evaded by making all of the tasks within a (sub-)flow async as well, but that's not ideal as the subflow's tasks (check cache => fetch => save) are all dependent so I'm just awaiting repeatedly. Ideally I could just asynchronously run the subflow over all the URLs, while running the tasks in each subflow run synchronously.

jacobdanovitch avatar Jun 06 '22 16:06 jacobdanovitch

This one is a bit complicated but is on our radar to resolve.

zanieb avatar Jun 06 '22 17:06 zanieb

Had a similar issue with "The task runner is already started!" but in my case that was on retry of a task that failed I think. Not exactly sure what was happening and very hard to pinpoint the exact issue as I had some parallel things running. Was not able to reproduce easily.

Benoss avatar Jun 29 '22 02:06 Benoss

Getting the same issue. RuntimeError: Task <Task pending name='anyio.from_thread.BlockingPortal._call_func' coro=<BlockingPortal._call_func() running at ...> cb=[TaskGroup._spawn.<locals>.task_done() at /Library/Caches/pypoetry/virtualenvs/finance-lGLYToYy-py3.9/lib/python3.9/site-packages/anyio/_backends/_asyncio.py:726]> got Future <Future pending> attached to a different loop

preceding with File "/usr/local/Cellar/[email protected]/3.9.10/Frameworks/Python.framework/Versions/3.9/lib/python3.9/ssl.py", line 888, in read v = self._sslobj.read(len) ssl.SSLWantReadError: The operation did not complete (read) (_ssl.c:2633)

It works locally, but when using prefect cloud it fails

My code looks like

@flow(name="scraping flow")
async def scrape(start_date, future_months_to_scrape, ):
    credentials = get_credentials()
    secrets = get_secrets()
    scraper = Scraper(start_date=start_date
                              , future_months_to_scrape=future_months_to_scrape)
    login_status = await login(scraper, credentials)
    data = await fetch_data(scraper, login_status, wait_for=[login])
    close_state = await close_and_clean_session(scraper, wait_for=[data])
    await load_to_mongo(data, secrets, wait_for=[close_state])
    processed_data = translate_to_mysql_data_model(data)
    await load_to_mysql(processed_data, secrets)


if __name__ == '__main__':
    load_dotenv()
    flow_param = dict(start_date="01/09/2021", future_months_to_scrape=10)
    asyncio.run(scrape(**flow_param))

NoamGit avatar Jul 09 '22 16:07 NoamGit

@NoamGit is it the same issue though? Which of those are subflows in your case? when it comes to Cloud, last Friday we deployed a new version, you may try to upgrade your Prefect client version and try running again with the Cloud backend.

anna-geller avatar Jul 10 '22 02:07 anna-geller

Tried it in 2.0b8. No subflows, just async code @anna-geller .

I've created a minimum example (again running ok without prefect start or cloud)

import asyncio

from dotenv import load_dotenv
from prefect import task, flow

@task
async def print_values(values):
    for value in values:
        await asyncio.sleep(1) # yield
        print(value, end=" ")

@task
def get_numbers():
    return [1, 2]

@flow(name='test-flow')
async def dummy_flow():
    numbers = get_numbers()
    await print_values(numbers)  # runs immediately
    coros = [print_values("abcd"), print_values("6789")]

    # asynchronously gather the tasks
    await asyncio.gather(*coros)



if __name__ == '__main__':
    load_dotenv()
    asyncio.run(dummy_flow())

This fails with

RuntimeError: Task <Task pending name='anyio.from_thread.BlockingPortal._call_func' coro=<BlockingPortal._call_func() running at /Users/noam.cohen/Library/Caches/pypoetry/virtualenvs/finance-lGLYToYy-py3.9/lib/python3.9/site-packages/anyio/from_thread.py:219> cb=[TaskGroup._spawn.<locals>.task_done() at /Users/noam.cohen/Library/Caches/pypoetry/virtualenvs/finance-lGLYToYy-py3.9/lib/python3.9/site-packages/anyio/_backends/_asyncio.py:726]> got Future <Future pending> attached to a different loop

NoamGit avatar Jul 10 '22 17:07 NoamGit

@NoamGit this is unrelated to the error reported in this thread, can you open a new issue?

zanieb avatar Jul 11 '22 14:07 zanieb

I'm at the same point as @jacobdanovitch in https://github.com/PrefectHQ/prefect/issues/5853#issuecomment-1147655272 only I'm struggling to get my code to work by making everything async and awaiting. That is probably due to my own lack of understanding of asyncio so I'll keep looking at it.

I think what we're probably all looking for is a flow.map operator that "just works" instead of having to do the async plumbing, with the deepcopy(subflow), etc.

jeffcarrico avatar Aug 12 '22 04:08 jeffcarrico

@jeffcarrico I wonder whether mapping over flows would solve it though since you would still need to concurrently start multiple runs of the same subflow

@madkinsz would mapping over subflows be even feasible? curious

anna-geller avatar Aug 12 '22 09:08 anna-geller

Yeah this is feasible and something we'll be exploring.

zanieb avatar Aug 12 '22 15:08 zanieb

View in #show-us-what-you-got on Slack

Chris_L. @Chris_L.: Hello Prefect community. :wave: Wanted to share a small code snippet I've been using to resolve this open issue: https://github.com/PrefectHQ/prefect/issues/5853 It's not technically a "fix" as I am creating new functions in the namespace via currying :curry:, but I think it's a cleaner solution compared to the current hack of using "deepcopy". @Jacob_Danovitch @Jeff_Carrico @Anna_Geller Hope this helps others encountering this issue!

import asyncio
from prefect import task, flow

@task
async def print_x(x):
  print(x)
  await asyncio.sleep(1)

def build_subflow(name):
  @flow(name=f"subflow:{name}")
  async def subflow(x):
    await print_x(x)
  return subflow

@flow
async def parent_flow():
  futures = [build_subflow(name=x)(x) for x in ["a", "b", "c"]]
  await asyncio.gather(*futures)

GitHub: Running the same subflow concurrently multiple times raises `RuntimeError( · Issue #5853 · PrefectHQ/prefect

A interesting finding: this pattern works even if @flow(name=f"subflow:{name}") is replaced with just @flow. Prefect gives a warning "A flow named 'subflow'...conflicts with another flow. Consider specifying a unique name parameters in the flow definition", but this does not seem to prevent the curried subflows from running concurrently.

Anna_Geller @Anna_Geller: Thanks for sharing, nice workaround! I'll add it to the GitHub issue too.

When we rename flows, Prefect treats those as separate flows and will assign a different flow ID - that's why it works.

When doing:

import asyncio
from prefect import task, flow, get_run_logger


@task
async def print_x(x):
    logger = get_run_logger()
    <http://logger.info|logger.info>(x)
    await asyncio.sleep(1)


def build_subflow():
    @flow
    async def subflow(x):
        await print_x(x)

    return subflow


@flow
async def parent_flow():
    futures = [build_subflow()(x) for x in ["x1", "x2", "x3"]]
    await asyncio.gather(*futures)


if __name__ == "__main__":
    asyncio.run(parent_flow())

it also works, as you mentioned but with a warning, but this is still nicer as it keeps the same flow ID.

Thanks again, I appreciate that you dive deeper into this topic

anna-geller avatar Aug 21 '22 12:08 anna-geller

I'm getting the same RuntimeError("The task runner is already started!") when trying to run two different subflows in parallel, but each calling the same databricks function jobs_runs_submit_and_wait_for_completion.

prefect 2.4.5 prefect-databricks 0.1.3

import asyncio
import json
from prefect import flow
from prefect_databricks import DatabricksCredentials
from prefect_databricks.flows import jobs_runs_submit_and_wait_for_completion

databricks_credentials = DatabricksCredentials.load("databricks")


@flow(name="databricks_job_1")
async def job_1():
    file = open("flows/databricks/job_1.json")
    tasks = json.load(file)["settings"]["tasks"]
    await jobs_runs_submit_and_wait_for_completion(
        databricks_credentials=databricks_credentials,
        run_name="prefect-job-1",
        tasks=tasks,
    )


@flow(name="databricks_job_2")
async def job_2():
    file = open("flows/databricks/job_2.json")
    tasks = json.load(file)["settings"]["tasks"]
    await jobs_runs_submit_and_wait_for_completion(
        databricks_credentials=databricks_credentials,
        run_name="prefect-job-2",
        tasks=tasks,
    )


@flow
async def async_flow():
    parallel_subflows = [job_1(), job_2()]
    await asyncio.gather(*parallel_subflows)


if __name__ == "__main__":
    main_flow_state = asyncio.run(async_flow())

One subflow starts and keeps running but but the second fails immediately.

17:30:22.401 | INFO    | prefect.engine - Created flow run 'handsome-hog' for flow 'async-flow'
17:30:26.023 | INFO    | Flow run 'handsome-hog' - Created subflow run 'olivine-trout' for flow 'databricks_job_2'
17:30:26.406 | INFO    | Flow run 'handsome-hog' - Created subflow run 'hypnotic-bat' for flow 'databricks_job_1'
17:30:27.267 | INFO    | Flow run 'olivine-trout' - Created subflow run 'mysterious-ara' for flow 'Submit jobs runs and wait for completion'
17:30:28.229 | INFO    | Flow run 'mysterious-ara' - Created task run 'jobs_runs_submit-c00eee75-0' for task 'jobs_runs_submit'
17:30:28.230 | INFO    | Flow run 'mysterious-ara' - Submitted task run 'jobs_runs_submit-c00eee75-0' for execution.
17:30:28.910 | INFO    | Flow run 'hypnotic-bat' - Created subflow run 'silver-numbat' for flow 'Submit jobs runs and wait for completion'
17:30:28.913 | ERROR   | Flow run 'silver-numbat' - Crash detected! Execution was interrupted by an unexpected exception.
17:30:29.666 | ERROR   | Flow run 'hypnotic-bat' - Encountered exception during execution:
Traceback (most recent call last):
  File "/Users/milenko/workspace/rb-prefect-master/venv/lib/python3.10/site-packages/prefect/engine.py", line 589, in orchestrate_flow_run
    result = await flow_call()
  File "/Users/milenko/workspace/rb-prefect-master/flows/async_demo.py", line 14, in job_1
    await jobs_runs_submit_and_wait_for_completion(
  File "/Users/milenko/workspace/rb-prefect-master/venv/lib/python3.10/site-packages/prefect/client/orion.py", line 82, in with_injected_client
    return await fn(*args, **kwargs)
  File "/Users/milenko/workspace/rb-prefect-master/venv/lib/python3.10/site-packages/prefect/engine.py", line 489, in create_and_begin_subflow_run
    task_runner = await stack.enter_async_context(flow.task_runner.start())
  File "/Users/milenko/.pyenv/versions/3.10.5/lib/python3.10/contextlib.py", line 619, in enter_async_context
    result = await _cm_type.__aenter__(cm)
  File "/Users/milenko/.pyenv/versions/3.10.5/lib/python3.10/contextlib.py", line 199, in __aenter__
    return await anext(self.gen)
  File "/Users/milenko/workspace/rb-prefect-master/venv/lib/python3.10/site-packages/prefect/task_runners.py", line 156, in start
    raise RuntimeError("The task runner is already started!")
RuntimeError: The task runner is already started!

milenkobeslic avatar Oct 06 '22 23:10 milenkobeslic

Replaced by https://github.com/PrefectHQ/prefect/issues/7319

zanieb avatar Oct 25 '22 20:10 zanieb

@madkinsz it looks like #7319 is on track to resolve RuntimeError: The task runner is already started! in a clean way. I've opened another issue (#7322) to capture the discussion around subflow.map that started with the suggestion by @jacobdanovitch . Feel free to close it if it is redundant.

https://github.com/PrefectHQ/prefect/issues/5853#issuecomment-1147655272

Ideally I could just asynchronously run the subflow over all the URLs, while running the tasks in each subflow run synchronously.

jeffcarrico avatar Oct 26 '22 02:10 jeffcarrico