I want do analysis on a dataset(like csv file) of 8gb which is in my laptop hard disk. I have already setup a dask kubernetes cluster on AKS with 1 scheduler and 3 worker with 7 gb each.
How can I work on my dataset using this dask kubernetes cluster on AKS? Which file system to share dataset between worker will be best for this purpose?
Any suggestion where I should store this dataset so that I can work on this dataset easily.
The method should work from both a jupyter notebook and from a python file also.
You would probably want to upload your data to an Azure blob store. There is more information about dask remote data (including Azure) here:
The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.