Dom
10/20/2025, 6:43 AMIdriss Bellil
10/20/2025, 7:17 AMIdriss Bellil
10/20/2025, 7:22 AMQuinn
10/20/2025, 1:40 PMNathan Low
10/20/2025, 1:56 PMprint(accounts_date_df.to_string())
geneva_data_dfs: ASPrefectGenevaUtil_dataclasses.GenevaExtractData = (
geneva_extract_tasks.extract_geneva_data_with_update_data_and_file.map(
update_row=accounts_date_df.itertuples(),
geneva_tax_lot_args=unmapped(geneva_args),
type_of_pull=type_of_pull,
)
)
geneva_data_tranform: pd.DataFrame = (
geneva_transform_tasks.transform_geneva_extract_data.map(
geneva_data_dfs, type_of_pull=type_of_pull
)
)
geneva_load_tasks.load_geneva_extract_data.map(
geneva_data_tranform, type_of_pull=type_of_pull
)Tom Han
10/20/2025, 9:51 PMLee Trout
10/20/2025, 10:59 PMJessica Smith
10/21/2025, 1:29 PM--limit parameter does when provided to a worker ? The docs say The maximum number of flow runs to start simultaneously. but it seems like it may actually mean something more like "how many flow runs can be running simultaneously" - which is the exact same thing as the existing concurrency limit on the work pool definition.Vincent D. Warmerdam
10/21/2025, 1:53 PMNathan Low
10/21/2025, 4:44 PMAmith M
10/21/2025, 4:54 PMNathan Low
10/22/2025, 12:02 PMMatt Youngberg
10/22/2025, 1:00 PMRick Fanta
10/22/2025, 1:33 PMRick Fanta
10/22/2025, 2:10 PMJagi Natarajan
10/22/2025, 9:32 PMJagi Natarajan
10/22/2025, 9:33 PMthiago
10/23/2025, 1:25 PMSimon Ouellette
10/23/2025, 4:13 PMBen
10/23/2025, 4:20 PMfor loop. E.g.
for object in objects:
data = tasks.operation1(data=object)
tasks.operation2(data=data)
I'm reading the docs which mention using either .submit() or .map() to run tasks concurrently, however i'm not sure how to access the data returned and pass them to a secondary task (which can also be run concurrently if possible)skrawczyk
10/23/2025, 5:24 PMJanet Carson
10/24/2025, 5:45 PMMarcelo de Souza
10/24/2025, 8:07 PMMitch Zink
10/24/2025, 10:52 PMTushar Mahajan
10/25/2025, 8:15 PMkubernetes type work pool and trying to pass env variables using the method described here (via prefect yaml) -> https://github.com/zzstoatzz/prefect-pack/blob/4cfc53ae94c6bd65ff0eefcb1c5ce72b48446179/prefect.yaml#L106
But I always get env variable not found error, anything I might be missing here ?Jeff Rose
10/28/2025, 5:58 PMBen
10/30/2025, 9:23 PMprefect deploy, I get this error:
prefect.exceptions.ScriptError: Script at 'src/providers/Tapology/flows/data_ingestion.py' encountered an exception: TypeError('Result storage configuration must be persisted server-side. Please call `.save()` on your block before passing it in.')
An exception occurred.
The thing is that I have created the storage block directly in the Prefect web UI, and I'm loading it in my code, something like this:
from prefect import flow
from prefect.filesystems import RemoteFileSystem
from prefect.serializers import JSONSerializer
from prefect.futures import as_completed
from prefect.task_runners import ThreadPoolTaskRunner
# Load the storage block you saved previously
r2_storage = RemoteFileSystem.load("r2-persistent-storage")
# Choose serializer for your payloads
json_ser = JSONSerializer()
@flow(
log_prints=True,
task_runner=ThreadPoolTaskRunner(max_workers=3),
persist_result=True,
result_storage=r2_storage,
result_serializer=json_ser,
)
def my_flow_function():
[...]
Another thing to note is that this code is working for already existing deployments, the error is only thrown when trying to deploy a new deployment.
Any idea whats going on?Tom Collom
10/31/2025, 8:32 AMMiguel Moncada
10/31/2025, 9:23 AMUnable to authenticate to the event stream. Please ensure the provided api_key or auth_token you are using is valid for this environment. Reason: Actor is not authenticated
Do you know what could be going on?Pierre L
10/31/2025, 10:57 AMdate.date object. Same error when using a datetime.datetime .
Here is my exact error :
Task run failed with exception: PrefectHTTPStatusError('Client error \'409 Conflict\' for url \'<http://prefect-server.prefectoss.svc.cluster.local:4200/api/deployments/d6150396-f839-4084-8091-401f5a6c53ca/create_flow_run>\'\nResponse:
{\'detail\': "Error creating flow run: Validation failed for field \'time_start\'. Failure reason: 1704067200.0 is not of type \'string\'"}\n
The bug doesnt occur when using Prefect cloud. Why ?