Hi,

I got s3a working on localstack. The missing piece of information from
Flink documentation seems to be that the system requires a HADOOP_HOME and
core-site.xml.

Flink documentation states that s3p (presto) should be used for file
checkpointing into s3. I am using RocksDB, which I assume also means that I
should use s3p (the documentation was not specific about that).  Is that
assumption correct?

However, I cannot get s3p working now.

I did the following so far:

I created the s3-fs-presto plugin directory and copied the jar from the opt
directory there.
I am not sure where to put the configuration keys though.  The
documentation states that I can just put in my flink-conf.yaml, but I had
no success.

Where do I put the presto configuration keys? Are there any other missing
steps? Is this something that would only work on an EMR environment with a
real HIVE running?

# The S3 storage endpoint server. This can be used to connect to an
S3-compatible storage
# system instead of AWS. When using v4 signatures, it is recommended to set
this to the
# AWS region-specific endpoint (e.g., http[s]://<bucket>.s3-<AWS-region>.
amazonaws.com).
hive.s3.endpoint: http://aws:4566

# Use HTTPS to communicate with the S3 API (defaults to true).
hive.s3.ssl.enabled: false

# Use path-style access for all requests to the S3-compatible storage.
# This is for S3-compatible storage that doesn’t support
virtual-hosted-style access. (defaults to false)
hive.s3.path-style-access: true

But that also did not work.

Any advice would be appreciated.

-Marco Villalobos

Reply via email to