Hi,
I've been able to sqoop from MySQL into HDFS, but I was wondering if it
was possible to send the data directly to S3 instead. I've read some posts
on this forum and others that indicate that it's not possible to do this -
could someone confirm?
I tried to get it to work by setting:
--warehouse-dir s3n://MYS3APIKEY:MYS3SECRETKEY@bucketname/folder/
or
--target-dir s3n://MYS3APIKEY:MYS3SECRETKEY@bucketname/folder/
options but I get the error:
ERROR tool.ImportTool: Imported Failed: This file system object (hdfs://
10.168.22.133:9000) does not support access to the request path
's3n://****:****@iakbar.emr/new-hive-output/_logs' You possibly called
FileSystem.get(conf) when you should have called FileSystem.get(uri, conf)
to obtain a file system supporting your path
If it's not possible to do this, should I just import to HDFS and then
output to S3? Is there an easy way to do this without having to specify
the schema of the whole table again?
thanks,
imran