[jira] [Created] (FLINK-10268) Document update deployment/aws HADOOP_CLASSPATH

2018-08-30 Thread Andy M (JIRA)
Andy M created FLINK-10268:
--

 Summary: Document update deployment/aws HADOOP_CLASSPATH
 Key: FLINK-10268
 URL: https://issues.apache.org/jira/browse/FLINK-10268
 Project: Flink
  Issue Type: Improvement
  Components: Documentation
Reporter: Andy M


The Deployment/AWS/Custom EMR Installation documents need to be updated.  
Currently the steps will result in a ClassNotFoundException.  A step needs to 
be added to include setting HADOOP_CLASSPATH=`hadoop classpath`



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Cannot run example on EMR

2018-08-30 Thread Andy M.
I am trying to launch a Flink cluster on EMR.  I am following directions
here:
https://ci.apache.org/projects/flink/flink-docs-release-1.6/ops/deployment/aws.html#custom-emr-installation

- I created a cluster with emr-5.16.0 - Hadoop 2.8.4
- I SSH'd into the master node,
downloaded flink-1.6.0-bin-hadoop28-scala_2.11.tgz
- I tried to run HADOOP_CONF_DIR=/etc/hadoop/conf ./bin/flink run -m
yarn-cluster -yn 1 examples/streaming/WordCount.jar and got the an error:
Caused by: java.lang.ClassNotFoundException:
javax.ws.rs.ext.MessageBodyReader

I also get the same error when trying to launch yarn-session.sh

Is anyone able to run 1.6 on EMR and provide me with what I am doing wrong?

Thank you


Re: Unable to write snapshots to S3 on EMR

2017-10-12 Thread Andy M.
Hi Ajoscha,

That didn't seem to do the trick either.  Do the following look correct?

I see 5.9.0 is released with Flink 1.3.2, so I tried that, and got the same
problem, all I did was upload my Scala .jar to the master, updated my
flink-conf.yaml, set my env variables, and ran it with the following
command: HADOOP_CONF_DIR=/etc/hadoop/conf flink run -m yarn-cluster -yn 5
-ys 2 ~/flink-consumer.jar

I am beginning to think something else may be wrong with my configuration.
 Does the following look correct?

In flink-conf.yaml:
state.backend: rocksdb
state.backend.fs.checkpointdir: s3://org/flink-project/state
state.checkpoints.dir: s3://org/flink-project/state

In the code:
env.enableCheckpointing(getCheckpointRate,CheckpointingMode.EXACTLY_ONCE)
env.getCheckpointConfig.enableExternalizedCheckpoints(ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION)
env.getCheckpointConfig.setMinPauseBetweenCheckpoints(1000)
env.getCheckpointConfig.setCheckpointTimeout(5000)
env.getCheckpointConfig.setMaxConcurrentCheckpoints(1)
env.setStateBackend(new RocksDBStateBackend("s3://org/flink-project/state",
true))

Thank you



On Wed, Oct 11, 2017 at 5:19 AM, Aljoscha Krettek <aljos...@apache.org>
wrote:

> Hi Andy,
>
> I remember that I was testing a job with almost exactly the same setup as
> part of the Flink 1.3.2 release testing. The command I used to start my job
> is roughly this:
>
> HADOOP_CONF_DIR=/etc/hadoop/conf bin/flink run -c my.main.Class -m
> yarn-cluster -yn 5 -ys 2 ...
>
> i.e. I export the proper hadoop config dir and I run a per-job YARN
> cluster. I think I also exported the result of "hadoop classpath" as
> HADOOP_CLASSPATH.
>
> Best,
> Aljoscha
>
> > On 10. Oct 2017, at 16:43, Andy M. <ajm2...@gmail.com> wrote:
> >
> > Hello,
> >
> > Bowen:  Unless I am missing something, it says there needs to be no setup
> > on EMR, Each topic says: "You don’t have to configure this manually if
> you
> > are running Flink on EMR."  S3 access from CLI works fine on my clusters.
> >
> > Chen: Thank you for this, I will look into this if I am unable to get
> this
> > running on YARN successfully.
> >
> > Stephan:  Removing the said library causes the flink
> > (flink-1.3.2/bin/flink) bash script to fail.  The underlying Java needs
> > this to work.  I tried explicitly setting the classpath for the java call
> > as well to point to the hadoop library jars.  This is the original java
> > command that I was trying to run:
> >
> > java
> > -Dlog.file=/home/hadoop/flink-1.3.2/log/flink-hadoop-client-
> ip-172-31-19-27.log
> > -Dlog4j.configuration=file:/home/hadoop/flink-1.3.2/conf/
> log4j-cli.properties
> > -Dlogback.configurationFile=file:/home/hadoop/flink-1.3.2/
> conf/logback.xml
> > -classpath
> > /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/
> home/hadoop/flink-1.3.2/lib/flink-shaded-hadoop2-uber-1.3.
> 2.jar:/home/hadoop/flink-1.3.2/lib/log4j-1.2.17.jar:/home/
> hadoop/flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar:/home/
> hadoop/flink-1.3.2/lib/flink-dist_2.11-1.3.2.jar::/etc/hadoop/conf:
> > org.apache.flink.client.CliFrontend run -m yarn-cluster -yn 1
> > /home/hadoop/flink-consumer.jar
> >
> >
> > This is what I changed it too(removing the shadded-hadoop2-uber jar and
> > adding in the hadoop folder):
> >
> > java
> > -Dlog.file=/home/hadoop/flink-1.3.2/log/flink-hadoop-client-
> ip-172-31-19-27.log
> > -Dlog4j.configuration=file:/home/hadoop/flink-1.3.2/conf/
> log4j-cli.properties
> > -Dlogback.configurationFile=file:/home/hadoop/flink-1.3.2/
> conf/logback.xml
> > -classpath
> > /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/
> home/hadoop/flink-1.3.2/lib/log4j-1.2.17.jar:/home/hadoop/
> flink-1.3.2/lib/slf4j-log4j12-1.7.7.jar:/home/hadoop/flink-
> 1.3.2/lib/flink-dist_2.11-1.3.2.jar:/usr/lib/hadoop/lib/
> activation-1.1.jar:/usr/lib/hadoop/lib/commons-io-2.4.jar:
> /usr/lib/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/
> lib/hadoop/lib/log4j-1.2.17.jar:/usr/lib/hadoop/lib/
> apacheds-i18n-2.0.0-M15.jar:/usr/lib/hadoop/lib/commons-
> lang-2.6.jar:/usr/lib/hadoop/lib/jackson-xc-1.9.13.jar:/
> usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/
> lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/lib/hadoop/
> lib/commons-logging-1.1.3.jar:/usr/lib/hadoop/lib/java-
> xmlbuilder-0.4.jar:/usr/lib/hadoop/lib/netty-3.6.2.Final.
> jar:/usr/lib/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/
> lib/hadoop/lib/commons-math3-3.1.1.jar:/usr/lib/hadoop/lib/
> jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/
> usr/lib/hadoop/lib/api-util-1.0.0-M20.jar:/usr/lib/hadoop/
> lib/commons-net-3.1.jar:/usr/lib/

Re: Unable to write snapshots to S3 on EMR

2017-10-10 Thread Andy M.
is of type [org.slf4j.impl.Log4jLoggerFactory]
Exception in thread "main" java.lang.NoClassDefFoundError:
org/apache/hadoop/util/VersionInfo
at
org.apache.flink.runtime.util.EnvironmentInformation.logEnvironmentInfo(EnvironmentInformation.java:283)
at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1124)
Caused by: java.lang.ClassNotFoundException:
org.apache.hadoop.util.VersionInfo
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
... 2 more

Simply removing the .jar from the folder causes the same error.

Thank you




On Mon, Oct 9, 2017 at 5:46 AM, Stephan Ewen <se...@apache.org> wrote:

> Hi!
>
> It looks like multiple Hadoop versions are in the classpath. Flink's hadoop
> jar and the EMR Hadoop jars.
> I would simply drop Flink's own Hadoop dependency and only use the EMR
> Hadoop jars.
>
> Delete the 'flink-shaded-h‌​adoop2-uber' jar from Flink, and make sure the
> setup is such that the Hadoop lib environment variable is set. Then it
> should not have conflicts any more.
>
>
>
> On Sun, Oct 8, 2017 at 12:08 AM, Chen Qin <qinnc...@gmail.com> wrote:
>
> > Attached my side project verified working to deploy jobmanager and
> > taskmanager as stateless service(non yarn/mesos), configuration here
> >
> > https://github.com/chenqin/flink-jar/tree/master/config/hadoop
> >
> > more detail here
> > https://github.com/chenqin/flink-jar/blob/master/src/
> > main/java/FlinkBootstrap.java#L49
> >
> > On Fri, Oct 6, 2017 at 10:26 PM, Bowen Li <bowen...@offerupnow.com>
> wrote:
> >
> > > Hi Andy,
> > >
> > > I believe it's because you didn't set your s3 impl correctly. Try to
> set
> > > your core-site.xml by following https://ci.apache.org/
> > > projects/flink/flink-docs-release-1.4/ops/deployment/
> > > aws.html#s3afilesystem-
> > > recommended
> > >
> > > Bowen
> > >
> > > On Fri, Oct 6, 2017 at 7:59 AM, Andy M. <ajm2...@gmail.com> wrote:
> > >
> > > > Hi Till,
> > > >
> > > > Seems like everything is in line there.  hadoop-common.jar ->
> > > > hadoop-common-2.7.3-amzn-3.jar
> > > >
> > > > And when i decompiled that jar I see  public void
> > > addResource(Configuration
> > > > conf) in org/apache/hadoop/conf/Configuration.java
> > > >
> > > > I agree that an incorrect version of the jar is probably being run,
> is
> > > > there a way to limit the classpath for the TaskManager when starting
> > the
> > > > job?
> > > >
> > > > Thank you
> > > >
> > > > On Fri, Oct 6, 2017 at 6:49 AM, Till Rohrmann <trohrm...@apache.org>
> > > > wrote:
> > > >
> > > > > Hi Andy,
> > > > >
> > > > > could you check which Hadoop version this jar
> > > > > /usr/lib/hadoop/hadoop-common.jar is? Maybe also checking whether
> > the
> > > > > contained hadoop Configuration class has the method
> > > > > Configuration.addResource(Lorg/apache/hadoop/conf/
> Configuration;)V.
> > > > Maybe
> > > > > this jar is the culprit because it comes from a different Hadoop
> > > version.
> > > > >
> > > > > Cheers,
> > > > > Till
> > > > > ​
> > > > >
> > > > > On Thu, Oct 5, 2017 at 4:22 PM, Andy M. <ajm2...@gmail.com> wrote:
> > > > >
> > > > > > Hi Till,
> > > > > >
> > > > > > I believe this is what you are looking for, classpath is much
> > bigger
> > > > for
> > > > > > the task manager.  I can also post the whole log file if needed:
> > > > > >
> > > > > > 2017-10-05 14:17:53,038 INFO  org.apache.flink.yarn.
> > > > > YarnTaskManagerRunner
> > > > > >  -  Classpath:
> > > > > > flink-consumer.jar:lib/flink-dist_2.11-1.3.2.jar:lib/flink-
> > > > > > python_2.11-1.3.2.jar:lib/flink-shaded-hadoop2-uber-1.3.
> > > > > > 2.jar:lib/log4j-1.2.17.jar:lib/slf4j-log4j12-1.7.7.jar:
> > > > > > log4j.properties:logback.xml:flink.jar:flink-conf.yaml::/
> > > > > > etc/hadoop/conf:/usr/lib/hadoop/hadoop-common-2.7.3-
>

Re: Unable to write snapshots to S3 on EMR

2017-10-06 Thread Andy M.
Hi Till,

Seems like everything is in line there.  hadoop-common.jar ->
hadoop-common-2.7.3-amzn-3.jar

And when i decompiled that jar I see  public void addResource(Configuration
conf) in org/apache/hadoop/conf/Configuration.java

I agree that an incorrect version of the jar is probably being run, is
there a way to limit the classpath for the TaskManager when starting the
job?

Thank you

On Fri, Oct 6, 2017 at 6:49 AM, Till Rohrmann <trohrm...@apache.org> wrote:

> Hi Andy,
>
> could you check which Hadoop version this jar
> /usr/lib/hadoop/hadoop-common.jar is? Maybe also checking whether the
> contained hadoop Configuration class has the method
> Configuration.addResource(Lorg/apache/hadoop/conf/Configuration;)V. Maybe
> this jar is the culprit because it comes from a different Hadoop version.
>
> Cheers,
> Till
> ​
>
> On Thu, Oct 5, 2017 at 4:22 PM, Andy M. <ajm2...@gmail.com> wrote:
>
> > Hi Till,
> >
> > I believe this is what you are looking for, classpath is much bigger for
> > the task manager.  I can also post the whole log file if needed:
> >
> > 2017-10-05 14:17:53,038 INFO  org.apache.flink.yarn.
> YarnTaskManagerRunner
> >  -  Classpath:
> > flink-consumer.jar:lib/flink-dist_2.11-1.3.2.jar:lib/flink-
> > python_2.11-1.3.2.jar:lib/flink-shaded-hadoop2-uber-1.3.
> > 2.jar:lib/log4j-1.2.17.jar:lib/slf4j-log4j12-1.7.7.jar:
> > log4j.properties:logback.xml:flink.jar:flink-conf.yaml::/
> > etc/hadoop/conf:/usr/lib/hadoop/hadoop-common-2.7.3-
> > amzn-3-tests.jar:/usr/lib/hadoop/hadoop-annotations-2.7.
> > 3-amzn-3.jar:/usr/lib/hadoop/hadoop-distcp.jar:/usr/lib/
> > hadoop/hadoop-auth-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
> > nfs-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-streaming-2.7.3-
> > amzn-3.jar:/usr/lib/hadoop/hadoop-ant-2.7.3-amzn-3.jar:/
> > usr/lib/hadoop/hadoop-distcp-2.7.3-amzn-3.jar:/usr/lib/
> > hadoop/hadoop-datajoin.jar:/usr/lib/hadoop/hadoop-
> > streaming.jar:/usr/lib/hadoop/hadoop-common.jar:/usr/lib/
> > hadoop/hadoop-ant.jar:/usr/lib/hadoop/hadoop-sls.jar:/
> > usr/lib/hadoop/hadoop-azure-2.7.3-amzn-3.jar:/usr/lib/
> > hadoop/hadoop-nfs.jar:/usr/lib/hadoop/hadoop-extras-2.7.
> > 3-amzn-3.jar:/usr/lib/hadoop/hadoop-gridmix.jar:/usr/lib/
> > hadoop/hadoop-common-2.7.3-amzn-3.jar:/usr/lib/hadoop/
> > hadoop-annotations.jar:/usr/lib/hadoop/hadoop-openstack-2.
> > 7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-archives-2.7.3-
> > amzn-3.jar:/usr/lib/hadoop/hadoop-azure.jar:/usr/lib/
> > hadoop/hadoop-extras.jar:/usr/lib/hadoop/hadoop-openstack.
> > jar:/usr/lib/hadoop/hadoop-rumen.jar:/usr/lib/hadoop/
> > hadoop-aws-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
> > datajoin-2.7.3-amzn-3.jar:/usr/lib/hadoop/hadoop-
> > archives.jar:/usr/lib/hadoop/hadoop-aws.jar:/usr/lib/
> > hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-rumen-2.7.3-
> > amzn-3.jar:/usr/lib/hadoop/hadoop-sls-2.7.3-amzn-3.jar:/
> > usr/lib/hadoop/hadoop-gridmix-2.7.3-amzn-3.jar:/usr/lib/
> > hadoop/lib/jettison-1.1.jar:/usr/lib/hadoop/lib/jaxb-api-2.
> > 2.2.jar:/usr/lib/hadoop/lib/htrace-core-3.1.0-incubating.
> > jar:/usr/lib/hadoop/lib/protobuf-java-2.5.0.jar:/usr/
> > lib/hadoop/lib/httpclient-4.5.3.jar:/usr/lib/hadoop/lib/
> > httpcore-4.4.4.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.
> > 1.jar:/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.
> > jar:/usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/
> > lib/activation-1.1.jar:/usr/lib/hadoop/lib/jersey-server-
> > 1.9.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:/
> > usr/lib/hadoop/lib/guava-11.0.2.jar:/usr/lib/hadoop/lib/
> > gson-2.2.4.jar:/usr/lib/hadoop/lib/commons-digester-1.
> > 8.jar:/usr/lib/hadoop/lib/jackson-xc-1.9.13.jar:/usr/
> > lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/
> > apacheds-i18n-2.0.0-M15.jar:/usr/lib/hadoop/lib/commons-
> > httpclient-3.1.jar:/usr/lib/hadoop/lib/curator-client-2.7.
> > 1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/
> > lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/
> > commons-net-3.1.jar:/usr/lib/hadoop/lib/jets3t-0.9.0.jar:/
> > usr/lib/hadoop/lib/commons-io-2.4.jar:/usr/lib/hadoop/lib/
> > xmlenc-0.52.jar:/usr/lib/hadoop/lib/jersey-json-1.9.
> > jar:/usr/lib/hadoop/lib/avro-1.7.4.jar:/usr/lib/hadoop/lib/
> > commons-math3-3.1.1.jar:/usr/lib/hadoop/lib/log4j-1.2.17.
> > jar:/usr/lib/hadoop/lib/commons-compress-1.4.1.jar:/
> > usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/
> > jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop/lib/netty-3.6.2.
> > Final.jar:/usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/
> > lib/hadoop/lib/asm-3.2.jar:/usr/lib/hado

Re: Unable to write snapshots to S3 on EMR

2017-10-05 Thread Andy M.
/share/aws/aws-java-sdk/aws-java-sdk-dms-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-directory-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-opsworkscm-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-cloudformation-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-cloudfront-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-budgets-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-clouddirectory-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-importexport-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lex-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-marketplaceentitlement-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-dynamodb-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-autoscaling-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-elastictranscoder-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-organizations-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-workspaces-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ssm-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-servermigration-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-events-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-applicationautoscaling-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-health-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-kms-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-logs-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-codestar-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-route53-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-redshift-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-marketplacemeteringservice-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-sns-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-batch-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-waf-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-simpledb-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-codedeploy-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ec2-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-models-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-devicefarm-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-cognitoidentity-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lexmodelbuilding-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-directconnect-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-elasticache-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-costandusagereport-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-discovery-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-resourcegroupstaggingapi-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ses-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lambda-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-workdocs-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-code-generator-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-cognitosync-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-efs-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-sts-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-athena-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-codecommit-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-storagegateway-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-lightsail-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-acm-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-glacier-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-ecr-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-support-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-codegen-maven-plugin-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-elasticsearch-1.11.160.jar:/usr/share/aws/aws-java-sdk/aws-java-sdk-iot-1.11.160.jar

Thank you

On Thu, Oct 5, 2017 at 5:25 AM, Till Rohrmann <trohrm...@apache.org> wrote:

> Hi Andy,
>
> the CliFrontend is not executed via Yarn, thus, it is not affected by
> dependencies which are added due to the underlying Yarn cluster. Therefore,
> it would be helpful to look at the TaskManager logs. Either you have
> enabled log aggregation on your Yarn cluster, then you can obtain the logs
> via `yarn logs -applicationId ` or you have to retrieve
> them from the machines where they were running (either by going directly
> there or via the Yarn web interface).
>
> Cheers,
> Till
>
> On Wed, Oct 4, 2017 at 4:27 PM, Andy M. <ajm2...@gmail.com> wrote:
>
> > Hi Till,
> >
> > That is actually the classpath used by the flink bash script(that
> launches
> > the jar using the java command).  I changed the execute to an echo, and
> > grabbed that for the CLI arguments.
> >
> > I believe this is the class path from the log file(although it might not
> be
> > the taskmanager log, is that any different from what would b

Re: Unable to write snapshots to S3 on EMR

2017-10-02 Thread Andy M.
Hi Fabian,

1) I have looked at the linked docs, and from what I can tell no setup
should really need to be done to get Flink working(Other than downloading
the correct binaries, which I believe I did)
2) I have downloaded the Flink 1.3.2 binaries(flink-1.3.2-bin-
hadoop27-scala_2.11.tgz
<http://apache.claz.org/flink/flink-1.3.2/flink-1.3.2-bin-hadoop27-scala_2.11.tgz>)
This is for hadoop 2.7.X, which matches EMR 5.8.0.

I appreciate any help or guidance you can provide me in fixing my problems,
please let me know if there is anything else I can provide you.

Thank you

On Mon, Oct 2, 2017 at 4:12 PM, Fabian Hueske <fhue...@gmail.com> wrote:

> Hi Andy,
>
> I'm not an AWS expert, so I'll just check on some common issues.
>
> I guess you already had a look at the Flink docs for AWS/EMR but I'll post
> the link just be to sure [1].
>
> Since you are using Flink 1.3.2 (EMR 5.8.0 comes with Flink 1.3.1) did you
> built Flink yourself or did you download the binaries?
> Does the Hadoop version of the Flink build match the Hadoop version of EMR
> 5.8.0, i.e., Hadoop 2.7.x?
>
> Best, Fabian
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/aws.html
>
> 2017-10-02 21:51 GMT+02:00 Andy M. <ajm2...@gmail.com>:
>
> > Hi Fabian,
> >
> > Sorry, I just realized I forgot to include that part.  The error returned
> > is:
> >
> > java.lang.NoSuchMethodError:
> > org.apache.hadoop.conf.Configuration.addResource(
> Lorg/apache/hadoop/conf/
> > Configuration;)V
> > at com.amazon.ws.emr.hadoop.fs.EmrFileSystem.initialize(
> > EmrFileSystem.java:93)
> > at org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.
> > initialize(HadoopFileSystem.java:328)
> > at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(
> > FileSystem.java:350)
> > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389)
> > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293)
> > at org.apache.flink.runtime.state.filesystem.
> > FsCheckpointStreamFactory.(FsCheckpointStreamFactory.java:99)
> > at org.apache.flink.runtime.state.filesystem.FsStateBackend.
> > createStreamFactory(FsStateBackend.java:282)
> > at org.apache.flink.contrib.streaming.state.RocksDBStateBackend.
> > createStreamFactory(RocksDBStateBackend.java:273
> >
> > I believe it has something to do with the classpath, but I am unsure why
> or
> > how to fix it.  The classpath being used during the execution is:
> > /home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/
> > ho‌​me/hadoop/flink-1.3.‌​2/lib/flink-shaded-h‌​adoop2-
> > uber-1.3.2.ja‌​r:/home/hadoop/flink‌​-1.3.2/lib/log4j-1.2‌​.
> > 17.jar:/home/hadoop‌​/flink-1.3.2/lib/slf‌​4j-log4j12-1.7.7.
> > jar‌​:/home/hadoop/flink-‌​1.3.2/lib/flink-dist‌​_2.11-1.3.
> > 2.jar::/et‌​c/hadoop/conf:
> >
> > I decompiled flink-shaded-h‌​adoop2-uber-1.3.2.ja‌​r and it seems the
> > addResource function does seem to be there.
> >
> > Thank you
> >
> > On Mon, Oct 2, 2017 at 3:43 PM, Fabian Hueske <fhue...@gmail.com> wrote:
> >
> > > Hi Andy,
> > >
> > > can you describe in more detail what exactly isn't working?
> > > Do you see error messages in the log files or on the console?
> > >
> > > Thanks, Fabian
> > >
> > > 2017-10-02 15:52 GMT+02:00 Andy M. <ajm2...@gmail.com>:
> > >
> > > > Hello,
> > > >
> > > > I am about to deploy my first Flink projects  to production, but I am
> > > > running into a very big hurdle.  I am unable to launch my project so
> it
> > > can
> > > > write to an S3 bucket.  My project is running on an EMR cluster,
> where
> > I
> > > > have installed Flink 1.3.2.  I am using Yarn to launch the
> application,
> > > and
> > > > it seems to run fine unless I am trying to enable check
> pointing(with a
> > > S3
> > > > target).  I am looking to use RocksDB as my check-pointing backend.
> I
> > > have
> > > > asked a few places, and I am still unable to find a solution to this
> > > > problem.  Here are my steps for creating a cluster, and launching my
> > > > application, perhaps I am missing a step.  I'd be happy to provide
> any
> > > > additional information if needed.
> > > >
> > > > AWS Portal:
> > > >
> > > > 1) EMR -> Create Cluster
> > > > 2) Advanced Options
> > > > 3) Release = emr-5.8.0
> > > > 4) Only sele

Re: Unable to write snapshots to S3 on EMR

2017-10-02 Thread Andy M.
Hi Fabian,

Sorry, I just realized I forgot to include that part.  The error returned
is:

java.lang.NoSuchMethodError:
org.apache.hadoop.conf.Configuration.addResource(Lorg/apache/hadoop/conf/Configuration;)V
at 
com.amazon.ws.emr.hadoop.fs.EmrFileSystem.initialize(EmrFileSystem.java:93)
at 
org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(HadoopFileSystem.java:328)
at 
org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:350)
at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:389)
at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293)
at 
org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.(FsCheckpointStreamFactory.java:99)
at 
org.apache.flink.runtime.state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:282)
at 
org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createStreamFactory(RocksDBStateBackend.java:273

I believe it has something to do with the classpath, but I am unsure why or
how to fix it.  The classpath being used during the execution is:
/home/hadoop/flink-1.3.2/lib/flink-python_2.11-1.3.2.jar:/ho‌​me/hadoop/flink-1.3.‌​2/lib/flink-shaded-h‌​adoop2-uber-1.3.2.ja‌​r:/home/hadoop/flink‌​-1.3.2/lib/log4j-1.2‌​.17.jar:/home/hadoop‌​/flink-1.3.2/lib/slf‌​4j-log4j12-1.7.7.jar‌​:/home/hadoop/flink-‌​1.3.2/lib/flink-dist‌​_2.11-1.3.2.jar::/et‌​c/hadoop/conf:

I decompiled flink-shaded-h‌​adoop2-uber-1.3.2.ja‌​r and it seems the
addResource function does seem to be there.

Thank you

On Mon, Oct 2, 2017 at 3:43 PM, Fabian Hueske <fhue...@gmail.com> wrote:

> Hi Andy,
>
> can you describe in more detail what exactly isn't working?
> Do you see error messages in the log files or on the console?
>
> Thanks, Fabian
>
> 2017-10-02 15:52 GMT+02:00 Andy M. <ajm2...@gmail.com>:
>
> > Hello,
> >
> > I am about to deploy my first Flink projects  to production, but I am
> > running into a very big hurdle.  I am unable to launch my project so it
> can
> > write to an S3 bucket.  My project is running on an EMR cluster, where I
> > have installed Flink 1.3.2.  I am using Yarn to launch the application,
> and
> > it seems to run fine unless I am trying to enable check pointing(with a
> S3
> > target).  I am looking to use RocksDB as my check-pointing backend.  I
> have
> > asked a few places, and I am still unable to find a solution to this
> > problem.  Here are my steps for creating a cluster, and launching my
> > application, perhaps I am missing a step.  I'd be happy to provide any
> > additional information if needed.
> >
> > AWS Portal:
> >
> > 1) EMR -> Create Cluster
> > 2) Advanced Options
> > 3) Release = emr-5.8.0
> > 4) Only select Hadoop 2.7.3
> > 5) Next -> Next -> Next -> Create Cluster ( I do fill out
> > names/keys/etc)
> >
> > Once the cluster is up I ssh into the Master and do the following:
> >
> > 1  wget
> > http://apache.claz.org/flink/flink-1.3.2/flink-1.3.2-bin-
> > hadoop27-scala_2.11.tgz
> > 2  tar -xzf flink-1.3.2-bin-hadoop27-scala_2.11.tgz
> > 3  cd flink-1.3.2
> > 4  ./bin/yarn-session.sh -n 2 -tm 5120 -s 4 -d
> > 5  Change conf/flink-conf.yaml
> > 6  ./bin/flink run -m yarn-cluster -yn 1 ~/flink-consumer.jar
> >
> > My conf/flink-conf.yaml I add the following fields:
> >
> > state.backend: rocksdb
> > state.backend.fs.checkpointdir: s3:/bucket/location
> > state.checkpoints.dir: s3:/bucket/location
> >
> > My program's checkpointing setup:
> >
> >
> > env.enableCheckpointing(getCheckpointRate,CheckpointingMode.EXACTLY_
> ONCE)
> >
> > env.getCheckpointConfig.enableExternalizedCheckpoints(
> > ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION)
> >
> > env.getCheckpointConfig.setMinPauseBetweenCheckpoints(
> > getCheckpointMinPause)
> > env.getCheckpointConfig.setCheckpointTimeout(getCheckpointTimeout)
> > env.getCheckpointConfig.setMaxConcurrentCheckpoints(1)
> > env.setStateBackend(new RocksDBStateBackend("s3://bucket/location",
> > true))
> >
>


Unable to write snapshots to S3 on EMR

2017-10-02 Thread Andy M.
Hello,

I am about to deploy my first Flink projects  to production, but I am
running into a very big hurdle.  I am unable to launch my project so it can
write to an S3 bucket.  My project is running on an EMR cluster, where I
have installed Flink 1.3.2.  I am using Yarn to launch the application, and
it seems to run fine unless I am trying to enable check pointing(with a S3
target).  I am looking to use RocksDB as my check-pointing backend.  I have
asked a few places, and I am still unable to find a solution to this
problem.  Here are my steps for creating a cluster, and launching my
application, perhaps I am missing a step.  I'd be happy to provide any
additional information if needed.

AWS Portal:

1) EMR -> Create Cluster
2) Advanced Options
3) Release = emr-5.8.0
4) Only select Hadoop 2.7.3
5) Next -> Next -> Next -> Create Cluster ( I do fill out
names/keys/etc)

Once the cluster is up I ssh into the Master and do the following:

1  wget
http://apache.claz.org/flink/flink-1.3.2/flink-1.3.2-bin-hadoop27-scala_2.11.tgz
2  tar -xzf flink-1.3.2-bin-hadoop27-scala_2.11.tgz
3  cd flink-1.3.2
4  ./bin/yarn-session.sh -n 2 -tm 5120 -s 4 -d
5  Change conf/flink-conf.yaml
6  ./bin/flink run -m yarn-cluster -yn 1 ~/flink-consumer.jar

My conf/flink-conf.yaml I add the following fields:

state.backend: rocksdb
state.backend.fs.checkpointdir: s3:/bucket/location
state.checkpoints.dir: s3:/bucket/location

My program's checkpointing setup:


env.enableCheckpointing(getCheckpointRate,CheckpointingMode.EXACTLY_ONCE)

env.getCheckpointConfig.enableExternalizedCheckpoints(ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION)

env.getCheckpointConfig.setMinPauseBetweenCheckpoints(getCheckpointMinPause)
env.getCheckpointConfig.setCheckpointTimeout(getCheckpointTimeout)
env.getCheckpointConfig.setMaxConcurrentCheckpoints(1)
env.setStateBackend(new RocksDBStateBackend("s3://bucket/location",
true))