Will
09/29/2021, 10:57 AMKevin Kho
CreateContainer
, StartContainer
family of tasks in the task library. In short, you would need to start containers yourself, but I think the limitation here is getting the data in and out of these various containers. They are isolated from each other so if you have a DataFrame in one container, I don’t think another container will readily be able to use it (they are like different machines).
I would imagine packaging all dependencies would be easier, but you can do something like StartContainer1 -> Persist Results1 -> Start Container2 -> Persist Results2 -> Start Container3 to Load and Combine Results1 and Results2. You can then split this up to subflows orchestrated by a main flow. Just persist in a location like S3 for other flows to grab.Will
09/29/2021, 2:40 PMKevin Kho
Bring your towel and join one of the fastest growing data communities. Welcome to our second-generation open source orchestration platform, a completely rethought approach to dataflow automation.
Powered by