You might be able to map? Just the connection needs to go inside the task rather than passed in. Are these csv files or parquet?
j
Jason Motley
01/12/2022, 10:58 PM
reading out of a data warehouse
Jason Motley
01/12/2022, 10:59 PM
The full extract is too slow, so I'm splitting it apart, running some transforms in python, then running a loop for the load to incrementally load 1/10th of the total data
Jason Motley
01/12/2022, 10:59 PM
But i'd like to speed that extract component if possible
k
Kevin Kho
01/12/2022, 11:03 PM
Ah then the best bet is just to map and create the connection inside the task. Map over a list of start dates and end dates. The limit would be the concurrent connections your warehouse allows
Bring your towel and join one of the fastest growing data communities. Welcome to our second-generation open source orchestration platform, a completely rethought approach to dataflow automation.