-
-
Notifications
You must be signed in to change notification settings - Fork 18.8k
Closed
Description
When trying to integrate Pandas CSV reading from s3 for local development (in docker) with containers from LocalStack or Minio we need to be able to define a custom host as well as a port.
PR #12198 introduces the AWS_S3_HOST
environment variable, I propose adding the AWS_S3_PORT
one. Something like:
s3_host = os.environ.get('AWS_S3_HOST','s3.amazonaws.com')
s3_port = os.environ.get('AWS_S3_PORT')
try:
conn = boto.connect_s3(host=s3_host, port=s3_port)
except boto.exception.NoAuthHandlerFound:
conn = boto.connect_s3(host=s3_host,anon=True, port=s3_port)
This would allow to define something like this in the docker-compose.yml
and use Minio for serving the csv files from a local s3 for development and AWS for production:
environment:
- AWS_ACCESS_KEY_ID=supersecret
- AWS_SECRET_ACCESS_KEY=supersecret
- AWS_S3_HOST=s3local
- AWS_S3_PORT=9000
- S3_USE_SIGV4=True
This is only applicable for pandas 0.18.X and 0.19.X since 0.20.X uses s3f. I would be willing to submit a PR for this.
Metadata
Metadata
Assignees
Labels
No labels