Fred Teunissen created FLINK-37749:
--------------------------------------
Summary: FileSystem Connector fails writing to S3 bucket
Key: FLINK-37749
URL: https://issues.apache.org/jira/browse/FLINK-37749
Project: Flink
Issue Type: Bug
Components: Connectors / FileSystem
Affects Versions: 1.20.1, 1.19.2, 1.20.0
Environment: Docker running:
- Flink 1.20.1
- minio
Reporter: Fred Teunissen
Attachments: filesystem-connector-s3-issue.tgz
Writing to an output file on S3 using the filesystem connector results in a
crash.
The output file is defined as follows.
{code:java}
CREATE TABLE <tableName>> (
bool BOOLEAN,
`someInt` INT
)
PARTITIONED BY (someInt)
WITH (
'connector' = 'filesystem',
'path' = 's3a://tdata',
'format' = 'csv',
-- 'format' = 'parquet',
'partition.default-name' = 'null'
);{code}
Note that the 'path' refers to an existing bucket that is hosted using Minio.
The job starts OK but as soon as the first checkpoint is being triggered the
job crashes with the following exception:
{code:java}
"Caused by: org.apache.flink.util.SerializedThrowable: java.io.IOException:
Stream closed." (full details in the attached zip file).{code}
The issue can be reproduced by unzipping the attached zip file that contains a
docker-compose file to spin up the Flink cluster and Minio container
and using the instructions in the README.MD to generate test data resulting in
the job crash.
Cause is unknown but I suspect more (or other) .jars need to be added to the
plugins/s3 directory as documented here:
https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/plugins/
--
This message was sent by Atlassian Jira
(v8.20.10#820010)