Hey team, what is the best practise for accessing ...
# ask-community
e
Hey team, what is the best practise for accessing binary files from a flow? We're processing a fair amount of weather data, stored in a binary format (netcdf). We're running the kubernetes agent, on Azure, with volumes mounted on the pods running our flows with a custom yaml file. It does works but is slightly brittle
d
Hi @Espen Overbye! Your outlined solution seems perfectly fine to me! What aspect of your current setup do you find brittle?
e
f.ex when testing runs, if you forget to add the path to the yaml - no mount ๐Ÿ˜‰. I think we'd ideally like to have a solution where we only accessed services we could attach as task; we're currently running a mix of linux/windows for the devs, and we can't mount the azure volume locally when developing, ending up with local sync issues (not to mention linux/windows folder naming fun)
d
Ahh that totally makes sense
e
we have a (too) complex readme for new devs to follow to get going ..
d
Does Blob Storage work well for binary formats in azure?
If you were willing to trade some I/O time for reliability, that could ensure the same access patterns between local/staging/production
e
jepp, could work
d
Weโ€™re on GCP and I persist everything I use for my Flows (locally or in Production) in GCS and I find it works extremely well
e
We could go down the path of using Hyrax(https://www.opendap.org/), a dedicated distributed solution for this type of data, but yet another piece of complexity to add
๐Ÿ‘€ 1
cool
๐Ÿ‘ 1