Proper async way of waiting for a running container

  aiohttp, docker, python, python-asyncio

I’m using python docker SDK inside of the aiohttp based application. My goal is to start a container when a request comes, wait until it finish all work and return a response. I want to handle many of such requests at the same time.

I run container in the detached mode so the problem isn’t in starting container itself. The problem is how to define when the container finished the work. Container instance of the docker package has a method wait() but it is blocking so I can’t use it "just like that". So I came up with something else: I start a container, create new asyncio task and (in that task) I’m checking whether container changed its state. It looks more or less like that:

import docker
import asyncio
import time

def wait_for_finish(event, container):
        timeout = time.time() + 600
        while time.time() < timeout:
            if container.status == 'running' or container.status == 'created':
                await asyncio.sleep(0.5)
                except requests.exceptions.HTTPError:

docker_client = docker.from_env()
container =**kwargs)
event = asyncio.Event()
            wait_for_finish(event, container)
await event.wait()

It is simple but it works fine. But my question is: is such "dummy" way of waiting for the finish of the container (status different than running or created) is a good method? To be honest I don’t like this solution, because of few reasons and I think something is wrong with such "waiting" but I don’t know what exacly…

I’m aware of the solutions such as aiodocker or the way with run_in_executor/threading or even replacing docker sdk with subprocess but for now I would like to use – if it is possible – combination of asyncio and docker sdk without for example thereading

Source: Docker Questions