Andreas Nigg
04/12/2023, 7:50 AM2023-04-12T00:01:35.802260259Z 00:01:35.800 | WARNING | Task run 'get_details_from_db-2' - Task run '17c77b72-df31-4899-836d-23eb934becc5' received abort during orchestration: The enclosing flow must be running to begin task execution. Task run is in PENDING state.
I'm not sure whether this is related to version 2.10.2 - I've just never seen a run failing due to this specific message.
The full log output please see in the thread.Deceivious
04/12/2023, 8:14 AMJens
04/12/2023, 9:22 AMNic
04/12/2023, 9:24 AMNikhil Joseph
04/12/2023, 10:05 AMCrash detected! Execution was interrupted by an unexpected exception: KeyError: 1267
1267
is random (not sure what it is yet)Peter Styk
04/12/2023, 11:46 AMGosia
04/12/2023, 12:00 PMFailed the last 3 attempts. Please check your environment and configuration.
Examples of recent errors:
Traceback (most recent call last):
File "/opt/conda/envs/prefect/lib/python3.10/site-packages/h2/connection.py",
line 224, in process_input
func, target_state = self._transitions[(self.state, input_)]
KeyError: (<ConnectionState.CLOSED: 3>, <ConnectionInputs.SEND_HEADERS: 0>)
During handling of the above exception, another exception occurred:
The agent doesn’t accept any flows to execute. The only thing I can do is to restart the agent container. Does anyone know what the issue might be, and how to solve it or work around it?Josh Paulin
04/12/2023, 1:21 PMprefect-dask
. We seem to be getting a big number of what looks like the ephemeral cluster shutting down early (Stack trace in the thread) which then leads into the flow crashing.
Any help greatly appreciatedPatrick Tan
04/12/2023, 3:22 PMChoenden Kyirong
04/12/2023, 5:42 PMToluwani Oludare
04/12/2023, 6:41 PM/tmp/
directory as its working directory. Does this directory delete files that enter there or the file never entered the directory.
I would appreciate any help.Aaron Goebel
04/12/2023, 6:45 PMPatrick Tan
04/12/2023, 7:04 PMAlex Beatson
04/12/2023, 7:24 PMx = task_1.submit()
, if we pass that future as an input to another task, y = task_2.submit(x=x)
, Ray will spin up a worker for task_2
with the requested resources, and this worker will wait on x.result()
.
This is obviously undesirable! If we had x1 = task1(); x2 = task2(x1=x1); x3=task3(x2=x2); ...
, all the later tasks in the pipeline would be wasting worker resources while waiting for the earlier workers to complete. We would like the worker for a task to only launch when the task inputs become ready.
Is this known behavior / does anyone know how to solve this?
Here's a working example:
from prefect import flow, task
from prefect_ray import RayTaskRunner
from prefect_ray.context import remote_options
import time
@task
def task_a():
time.sleep(100)
return 'a'
@task
def task_b():
return 'b'
@flow(
task_runner=RayTaskRunner(
address="<ray://ray-cluster-kuberay-head-svc:10001>"
)
)
def multi_test():
with remote_options(num_cpus=1, num_gpus=1):
x = task_a.submit()
for i in range(10):
with remote_options(num_cpus=1, num_gpus=1):
task_b.submit(wait_for=[x])
if __name__ == '__main__':
print(multi_test())
When run, we immediately get 11 workers created, per the image.
CC @Muxin Fang @Haotian LiCharles Leung
04/12/2023, 10:33 PMMatheus Rocha
04/12/2023, 11:51 PMPeter Styk
04/13/2023, 7:26 AMEmma Rizzi
04/13/2023, 8:58 AMprefect deployment build
and I am now trying to migrate to projects to create flows as docker images, but prefect deploy
fails, details in threaddarshan darshan
04/13/2023, 12:01 PMdarshan darshan
04/13/2023, 12:02 PMDeceivious
04/13/2023, 12:28 PMYaron Levi
04/13/2023, 12:53 PMJ
04/13/2023, 1:37 PMPatrick Tan
04/13/2023, 1:54 PMBianca Hoch
04/13/2023, 1:59 PM🤖 Status Bot: [Investigating] Investigating high API latency - We are investigating sporadic periods of high latency and error rates for some Cloud 2 APIs beginning on April 13 around midnight UTC (April 12th, 5pm PT/8pm ET). We are investigating the problem and will provide updates as we learn more. (https://prefect.status.io)
FuETL
04/13/2023, 2:18 PMJosh Paulin
04/13/2023, 3:56 PMDeployment.build_from_flow
. Things have been working fine until I tried adding in work_pool
configuration. Not the tests fail with
except HTTPStatusError as exc:
> raise PrefectHTTPStatusError.from_httpx_error(exc) from exc.__cause__
E prefect.exceptions.PrefectHTTPStatusError: Client error '404 Not Found' for url '<http://ephemeral-prefect/api/deployments/>'
E Response: {'detail': 'Work pool "<work-pool-name>" not found.'}
E For more information check: <https://httpstatuses.com/404>
This works fine if I’m using work_queue
, or leave them both out altogether. Any suggestions?John O'Farrell
04/13/2023, 5:22 PMRRule
if I want that level of control? I know in prefect 1 it was possible to still have flexibility with cron schedulesBianca Hoch
04/13/2023, 5:42 PM🤖 Status Bot: [Monitoring] Investigating high API latency - The issue appears to be related to a routine upgrade of our Kubernetes cluster, and we have rolled back to the previously running version. Our internal metrics show a marked improvement in the p99 latency of our services, and we are still working to identify the underlying cause. We are continuing to monitor the situation, however, we believe that user-impacting issues should have cleared. ✅ (https://prefect.status.io)
Aric Huang
04/13/2023, 6:01 PM