I can't access my files on S3 using a dataframe read : df_read_csv. I get the error: Exception: Unable to locate credentials
This works fine when my dask distributed is running against local worker cores. However, when I import a client with a cluster of additional worker servers it fails. My cluster was created using dask-ec2 on ubuntu with 1 scheduler on the header server and 3 worker servers (all ubuntu).
I am assuming that the failure is because all the workers also need access to S3. I have installed the aws cli on all of them and connected using my key and can list the S3 buckets from the cli. However, for some reason my dataframe read throws an ubuntu error saying that boto can't locate credentials
I have trawled through various posts but can't seem to find anything that helps. This is screenshot of the error: