-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
S3 support #326
base: master
Are you sure you want to change the base?
S3 support #326
Conversation
@tcarval is there any reasons for not having the gz index files on s3 ? |
I am adding the gz indexes (the synchronization gdac - aws is underway) |
- a prototype for the record and benchmark different design
fix access to AWS creds
- and new checker for AWS credentials
New IndexStore ready to work with AWS S3 core index file from argopy import ArgoIndex
idx = ArgoIndex(host='s3://argo-gdac-sandbox/pub/idx').load()
idx.search_wmo_cyc(6903091, 1) poke @tcarval |
… on s3 - refactor argo_index_pa and argo_index_pd to use argostore super init - index_path now a dynamic property - index_path set to use gz file when found on server at instanciation - index property not set when used with s3 store - new search_s3 decorator for some search methods (to be used with s3) - more Path usage - minor reformatting - fix bug in s3index to return appropriate empty pyarrow table if no SQL response is found
- add keywords/shortcuts for hosts
- add a decorator to fix errors raised when pyarrow is not availalble
fix bug for unknown AWS credentials with boto3 client
The Argo ADMT is experiencing with Amazon S3 in order to move the GDAC infrastructure into the cloud.
In order to prepare argopy for this and to be able to access and test the AWS prototype server, we need to develop support for S3.
This would require:
A new data fetcher will be developed in another PR