Is it possible to use an environmental credentials provider? On Thu, Jan 28, 2021 at 8:35 AM Arvid Heise <ar...@apache.org> wrote:
> Hi Marco, > > afaik you don't need HADOOP_HOME or core-site.xml. > > I'm also not sure from where you got your config keys. (I guess from the > Presto page, which probably all work if you remove hive., maybe we should > also support that) > > All keys with prefix s3 or s3p (and fs.s3, fs.s3p) are routed towards > presto [1]. > > So it should be > s3.access-key: XXX > s3.secret-key: XXX > s3.endpoint: http://aws:4566 > s3.path-style-access: true > s3.path.style.access: true (only one of them is needed, but I don't know > which, so please try out) > > [1] > https://ci.apache.org/projects/flink/flink-docs-stable/deployment/filesystems/s3.html#configure-access-credentials > > On Thu, Jan 28, 2021 at 4:58 PM Marco Villalobos < > mvillalo...@kineteque.com> wrote: > >> Hi, >> >> I got s3a working on localstack. The missing piece of information from >> Flink documentation seems to be that the system requires a HADOOP_HOME and >> core-site.xml. >> >> Flink documentation states that s3p (presto) should be used for file >> checkpointing into s3. I am using RocksDB, which I assume also means that I >> should use s3p (the documentation was not specific about that). Is that >> assumption correct? >> >> However, I cannot get s3p working now. >> >> I did the following so far: >> >> I created the s3-fs-presto plugin directory and copied the jar from the >> opt directory there. >> I am not sure where to put the configuration keys though. The >> documentation states that I can just put in my flink-conf.yaml, but I had >> no success. >> >> Where do I put the presto configuration keys? Are there any other missing >> steps? Is this something that would only work on an EMR environment with a >> real HIVE running? >> >> # The S3 storage endpoint server. This can be used to connect to an >> S3-compatible storage >> # system instead of AWS. When using v4 signatures, it is recommended to >> set this to the >> # AWS region-specific endpoint (e.g., http[s]://<bucket>.s3-<AWS-region>. >> amazonaws.com). >> hive.s3.endpoint: http://aws:4566 >> >> # Use HTTPS to communicate with the S3 API (defaults to true). >> hive.s3.ssl.enabled: false >> >> # Use path-style access for all requests to the S3-compatible storage. >> # This is for S3-compatible storage that doesn’t support >> virtual-hosted-style access. (defaults to false) >> hive.s3.path-style-access: true >> >> But that also did not work. >> >> Any advice would be appreciated. >> >> -Marco Villalobos >> >