raj
11/05/2025, 12:00 PMRECV_PING) on a closed connection, which is invalid according to the HTTP/2 protocol specification.
At the core, the *`httpx` client*—used internally by *Prefect's Cloud API client*—attempted to send a PING frame (RECV_PING) on an HTTP/2 connection that had already transitioned to a CLOSED state. This caused the underlying httpcore transport layer to raise a LocalProtocolError due to the mismatch in expected connection state.
🧾 Root Cause Analysis (RCA):
• The system attempted to reuse an HTTP/2 connection that had already been closed, leading to an invalid protocol state.
• This is most likely due to a stale connection being kept in the async connection pool or a bug in the connection lifecycle management by httpcore.
• Prefect’s cloud client attempted communication over this dead connection, resulting in failure.Øyvind Monsen
11/05/2025, 3:34 PMArthur Ren
11/05/2025, 5:23 PMSlackbot
11/06/2025, 11:34 AMSergio Luceno
11/06/2025, 11:38 AMVít Brunner
11/06/2025, 1:34 PMArthur Ren
11/07/2025, 2:34 AMKiran
11/07/2025, 5:49 AMPierre L
11/07/2025, 9:41 AMGiacomo Chiarella
11/07/2025, 2:26 PMmy_task = task.submit()
wait for it to finish and
my_task.state
how should I change the last statement to solve the warning? Same question related to my_task.task_run_idAmrit Amar
11/07/2025, 4:56 PMprefect-aws library and I'm excited to start using it, however most of the patterns for running flows seem to evolve around using another EC2 instance.
I was wondering if it was possible to have a Prefect server as an orchestrator (on Fargate/EC2) that would run defined flows and deploy a 'task' on a lambda (I would have already configured the lambda, just need to send a task to it and then retrieve results to pass onto the next task which would be another lambda call). I see that lambda invoke that can happen 'within' a task but is it possible deploy the entire task as a lambda itself rather than invoking it from the main prefect instance?
Also, will the workers, if invoking a lambda from the main instance, actually wait the full 15 minutes for a response from the lambda or do I need to poll the status of the lambda job with another task? And is this scalable to having 1 prefect server instance with 4 defined flows having 10s-100s of concurrent flows running?
My main use case is to have an orchestrator that, upon running a flow with defined tasks, will send info to a lambda, and then get the result from it to pass it on to the next task until the end of a flow.
Thank you! (@Marvin)Quinton Nickum
11/07/2025, 9:46 PMAaron
11/08/2025, 11:56 AMRiya Sinha
11/08/2025, 7:33 PM^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/prefect/flow_engine.py", line 1079, in create_flow_run
return await client.create_flow_run(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/prefect/client/orchestration/_flow_runs/client.py", line 568, in create_flow_run
response = await self.request("POST", "/flow_runs/", json=flow_run_create_json)
full trace in thread to avoid large message — just wanted to ask if this is an issue others are currently facing recentlyHW
11/10/2025, 7:33 AM[error] tests\conftest.py:76: in <module>
from .fixtures.api import *
tests\fixtures\api.py:12: in <module>
from prefect.settings import PREFECT_SERVER_DOCKET_NAME, temporary_settings
E ImportError: cannot import name 'PREFECT_SERVER_DOCKET_NAME' from 'prefect.settings' (...\Lib\site-packages\prefect\settings\__init__.py)
I don't know how settings like PREFECT_SERVER_DOCKET_NAME works, anybody help?Nathan Low
11/10/2025, 2:41 PMpull:
- prefect.deployments.steps.run_shell_script:
id: get-github-token
script: uv tool run prefect-cloud github token owner/repository
- prefect.deployments.steps.git_clone:
id: git-clone
repository: <https://x-access-token>:{{ get-github-token.stdout }}@github.com/owner/repository.git
Then I get this error when running the run_shell_script step when the flow start to run:
`run_shell_script` failed with error code 1: Downloading pygments (1.2MiB)
Downloading uv (20.4MiB)
Downloading pydantic-core (2.0MiB)
Downloading pydantic-core
Downloading uv
Downloading pygments
Installed 35 packages in 126ms
Bytecode compiled 1004 files in 1.11sMichel Lalancette
11/11/2025, 12:20 AMKevin Hu
11/11/2025, 7:16 PMTheom
11/12/2025, 9:41 AM.to_deployment + serve scheme). If I set the environment variable PREFECT_RESULTS_LOCAL_STORAGE_PATH, it's set to the same value for all workflows. I asked Marvin and I was told to set result_storage in the flow decorator, either to LocalFileSystem(basepath=path1) or Path(path1) but it fails. Somehow I need to call .save() on the LocalFileSystem instance beforehand to create a block server-side. I find this a bit convoluted for a fairly simple need, especially since I otherwise don't need to do it when using the PREFECT_RESULTS_LOCAL_STORAGE_PATH variable set to arbitrary values. Marvin also told me that I could pass PREFECT_RESULTS_LOCAL_STORAGE_PATH as a job variable to the deployment, but that it was not the recommended solution. What would you recommend? Also, if there are no other solutions than what I described, I would argue that being allowed to simply provide a Path object to result_storage would be a reasonable behaviour to expect and I would be interested in discussing it or filing an issue on GitHub. Thanks!Phillip Shearin
11/12/2025, 3:11 PMTri
11/13/2025, 4:54 PMhambone johnston
11/13/2025, 8:15 PMhambone johnston
11/13/2025, 8:16 PMalex joyce
11/14/2025, 6:34 AMNathan Low
11/14/2025, 2:57 PMJordan Jones
11/14/2025, 4:18 PMSergio Luceno
11/14/2025, 5:17 PMKevin Hu
11/14/2025, 5:51 PMChu
11/14/2025, 7:42 PMresult_storage from the parent flow, even if the subflow had its own result_storage set?Sơn Lê
11/15/2025, 10:01 AM