I got this fixed by setting jetty_opts with proxy values. Thanks Peter!!
On Thu, May 17, 2018 at 4:05 PM purna pradeep <purna2prad...@gmail.com> wrote: > Ok I fixed this by adding aws keys in oozie > > But I’m getting below error > > I have tried setting proxy in core-site.xml but no luck > > > 2018-05-17 15:39:20,602 ERROR CoordInputLogicEvaluatorPhaseOne:517 - > SERVER[localhost] USER[-] GROUP[-] TOKEN[-] APP[-] > JOB[0000000-180517144113498-oozie-xjt0-C] ACTION[0000000- > 180517144113498-oozie-xjt0-C@2] > org.apache.oozie.service.HadoopAccessorException: > E0902: Exception occurred: [doesBucketExist on cmsegmentation-qa: > com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect > to mybucket.s3.amazonaws.com:443 > <http://cmsegmentation-qa.s3.amazonaws.com:443/> [mybucket. > s3.amazonaws.com/52.216.165.155 > <http://cmsegmentation-qa.s3.amazonaws.com/52.216.165.155>] failed: > connect timed out] > > org.apache.oozie.service.HadoopAccessorException: E0902: Exception > occurred: [doesBucketExist on cmsegmentation-qa: > com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect > to mybucket.s3.amazonaws.com:443 > <http://cmsegmentation-qa.s3.amazonaws.com:443/> [mybucket > .s3.amazonaws.com > <http://cmsegmentation-qa.s3.amazonaws.com/52.216.165.155> failed: > connect timed out] > > at > org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:630) > > at > org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:594) > at org.apache.oozie.dependency.FSURIHandler.getFileSystem( > FSURIHandler.java:184)-env.sh > > But now I’m getting this error > > > > On Thu, May 17, 2018 at 2:53 PM purna pradeep <purna2prad...@gmail.com> > wrote: > >> Ok I got passed this error >> >> By rebuilding oozie with Dhttpclient.version=4.5.5 >> -Dhttpcore.version=4.4.9 >> >> now getting this error >> >> >> >> ACTION[0000000-180517144113498-oozie-xjt0-C@1] >> org.apache.oozie.service.HadoopAccessorException: E0902: Exception >> occurred: [doesBucketExist on mybucketcom.amazonaws.AmazonClientException: >> No AWS Credentials provided by BasicAWSCredentialsProvider >> EnvironmentVariableCredentialsProvider >> SharedInstanceProfileCredentialsProvider : >> com.amazonaws.SdkClientException: Unable to load credentials from service >> endpoint] >> >> org.apache.oozie.service.HadoopAccessorException: E0902: Exception >> occurred: [doesBucketExist on cmsegmentation-qa: >> com.amazonaws.AmazonClientException: No AWS Credentials provided by >> BasicAWSCredentialsProvider EnvironmentVariableCredentialsProvider >> SharedInstanceProfileCredentialsProvider : >> com.amazonaws.SdkClientException: Unable to load credentials from service >> endpoint] >> >> On Thu, May 17, 2018 at 12:24 PM purna pradeep <purna2prad...@gmail.com> >> wrote: >> >>> >>> Peter, >>> >>> Also When I submit a job with new http client jar, I get >>> >>> ```Error: IO_ERROR : java.io.IOException: Error while connecting Oozie >>> server. No of retries = 1. Exception = Could not authenticate, >>> Authentication failed, status: 500, message: Server Error``` >>> >>> >>> On Thu, May 17, 2018 at 12:14 PM purna pradeep <purna2prad...@gmail.com> >>> wrote: >>> >>>> Ok I have tried this >>>> >>>> It appears that s3a support requires httpclient 4.4.x and oozie is >>>> bundled with httpclient 4.3.6. When httpclient is upgraded, the ext UI >>>> stops loading. >>>> >>>> >>>> >>>> On Thu, May 17, 2018 at 10:28 AM Peter Cseh <gezap...@cloudera.com> >>>> wrote: >>>> >>>>> Purna, >>>>> >>>>> Based on >>>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3 >>>>> you should try to go for s3a. >>>>> You'll have to include the aws-jdk as well if I see it correctly: >>>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3A >>>>> Also, the property names are slightly different so you'll have to >>>>> change the example I've given. >>>>> >>>>> >>>>> >>>>> On Thu, May 17, 2018 at 4:16 PM, purna pradeep < >>>>> purna2prad...@gmail.com> wrote: >>>>> >>>>>> Peter, >>>>>> >>>>>> I’m using latest oozie 5.0.0 and I have tried below changes but no >>>>>> luck >>>>>> >>>>>> Is this for s3 or s3a ? >>>>>> >>>>>> I’m using s3 but if this is for s3a do you know which jar I need to >>>>>> include I mean Hadoop-aws jar or any other jar if required >>>>>> >>>>>> Hadoop-aws-2.8.3.jar is what I’m using >>>>>> >>>>>> On Wed, May 16, 2018 at 5:19 PM Peter Cseh <gezap...@cloudera.com> >>>>>> wrote: >>>>>> >>>>>>> Ok, I've found it: >>>>>>> >>>>>>> If you are using 4.3.0 or newer this is the part which checks for >>>>>>> dependencies: >>>>>>> >>>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/command/coord/CoordCommandUtils.java#L914-L926 >>>>>>> It passes the coordinator action's configuration and even does >>>>>>> impersonation to check for the dependencies: >>>>>>> >>>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/coord/input/logic/CoordInputLogicEvaluatorPhaseOne.java#L159 >>>>>>> >>>>>>> Have you tried the following in the coordinator xml: >>>>>>> >>>>>>> <action> >>>>>>> <workflow> >>>>>>> >>>>>>> <app-path>hdfs://bar:9000/usr/joe/logsprocessor-wf</app-path> >>>>>>> <configuration> >>>>>>> <property> >>>>>>> <name>fs.s3.awsAccessKeyId</name> >>>>>>> <value>[YOURKEYID]</value> >>>>>>> </property> >>>>>>> <property> >>>>>>> <name>fs.s3.awsSecretAccessKey</name> >>>>>>> <value>[YOURKEY]</value> >>>>>>> </property> >>>>>>> </configuration> >>>>>>> </workflow> >>>>>>> </action> >>>>>>> >>>>>>> Based on the source this should be able to poll s3 periodically. >>>>>>> >>>>>>> On Wed, May 16, 2018 at 10:57 PM, purna pradeep < >>>>>>> purna2prad...@gmail.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> I have tried with coordinator's configuration too but no luck ☹️ >>>>>>>> >>>>>>>> On Wed, May 16, 2018 at 3:54 PM Peter Cseh <gezap...@cloudera.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Great progress there purna! :) >>>>>>>>> >>>>>>>>> Have you tried adding these properites to the coordinator's >>>>>>>>> configuration? we usually use the action config to build up >>>>>>>>> connection to >>>>>>>>> the distributed file system. >>>>>>>>> Although I'm not sure we're using these when polling the >>>>>>>>> dependencies for coordinators, but I'm excited about you trying to >>>>>>>>> make it >>>>>>>>> work! >>>>>>>>> >>>>>>>>> I'll get back with a - hopefully - more helpful answer soon, I >>>>>>>>> have to check the code in more depth first. >>>>>>>>> gp >>>>>>>>> >>>>>>>>> On Wed, May 16, 2018 at 9:45 PM, purna pradeep < >>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Peter, >>>>>>>>>> >>>>>>>>>> I got rid of this error by adding >>>>>>>>>> hadoop-aws-2.8.3.jar and jets3t-0.9.4.jar >>>>>>>>>> >>>>>>>>>> But I’m getting below error now >>>>>>>>>> >>>>>>>>>> java.lang.IllegalArgumentException: AWS Access Key ID and Secret >>>>>>>>>> Access Key must be specified by setting the fs.s3.awsAccessKeyId and >>>>>>>>>> fs.s3.awsSecretAccessKey properties (respectively) >>>>>>>>>> >>>>>>>>>> I have tried adding AWS access ,secret keys in >>>>>>>>>> >>>>>>>>>> oozie-site.xml and hadoop core-site.xml , and hadoop-config.xml >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 16, 2018 at 2:30 PM purna pradeep < >>>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> I have tried this ,just added s3 instead of * >>>>>>>>>>> >>>>>>>>>>> <property> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name> >>>>>>>>>>> >>>>>>>>>>> <value>hdfs,hftp,webhdfs,s3</value> >>>>>>>>>>> >>>>>>>>>>> </property> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Getting below error >>>>>>>>>>> >>>>>>>>>>> java.lang.RuntimeException: java.lang.ClassNotFoundException: >>>>>>>>>>> Class org.apache.hadoop.fs.s3a.S3AFileSystem not found >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2369) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2793) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2810) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:100) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2849) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2831) >>>>>>>>>>> >>>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:389) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:625) >>>>>>>>>>> >>>>>>>>>>> at >>>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:623 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, May 16, 2018 at 2:19 PM purna pradeep < >>>>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> This is what is in the logs >>>>>>>>>>>> >>>>>>>>>>>> 2018-05-16 14:06:13,500 INFO URIHandlerService:520 - >>>>>>>>>>>> SERVER[localhost] Loaded urihandlers >>>>>>>>>>>> [org.apache.oozie.dependency.FSURIHandler] >>>>>>>>>>>> >>>>>>>>>>>> 2018-05-16 14:06:13,501 INFO URIHandlerService:520 - >>>>>>>>>>>> SERVER[localhost] Loaded default urihandler >>>>>>>>>>>> org.apache.oozie.dependency.FSURIHandler >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, May 16, 2018 at 12:27 PM Peter Cseh < >>>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> That's strange, this exception should not happen in that case. >>>>>>>>>>>>> Can you check the server logs for messages like this? >>>>>>>>>>>>> LOG.info("Loaded urihandlers {0}", >>>>>>>>>>>>> Arrays.toString(classes)); >>>>>>>>>>>>> LOG.info("Loaded default urihandler {0}", >>>>>>>>>>>>> defaultHandler.getClass().getName()); >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, May 16, 2018 at 5:47 PM, purna pradeep < >>>>>>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> This is what I already have in my oozie-site.xml >>>>>>>>>>>>>> >>>>>>>>>>>>>> <property> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name> >>>>>>>>>>>>>> >>>>>>>>>>>>>> <value>*</value> >>>>>>>>>>>>>> >>>>>>>>>>>>>> </property> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Wed, May 16, 2018 at 11:37 AM Peter Cseh < >>>>>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> You'll have to configure >>>>>>>>>>>>>>> oozie.service.HadoopAccessorService.supported.filesystems >>>>>>>>>>>>>>> hdfs,hftp,webhdfs Enlist >>>>>>>>>>>>>>> the different filesystems supported for federation. If >>>>>>>>>>>>>>> wildcard "*" is >>>>>>>>>>>>>>> specified, then ALL file schemes will be allowed.properly. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> For testing purposes it's ok to put * in there in >>>>>>>>>>>>>>> oozie-site.xml >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Wed, May 16, 2018 at 5:29 PM, purna pradeep < >>>>>>>>>>>>>>> purna2prad...@gmail.com> >>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> > Peter, >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > I have tried to specify dataset with uri starting with >>>>>>>>>>>>>>> s3://, s3a:// and >>>>>>>>>>>>>>> > s3n:// and I am getting exception >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > Exception occurred:E0904: Scheme [s3] not supported in uri >>>>>>>>>>>>>>> > [s3://mybucket/input.data] Making the job failed >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > org.apache.oozie.dependency.URIHandlerException: E0904: >>>>>>>>>>>>>>> Scheme [s3] not >>>>>>>>>>>>>>> > supported in uri [s3:// mybucket /input.data] >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>>> > URIHandlerService.java:185) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>>> > URIHandlerService.java:168) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>>> > URIHandlerService.java:160) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.createEarlyURIs( >>>>>>>>>>>>>>> > CoordCommandUtils.java:465) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils. >>>>>>>>>>>>>>> > separateResolvedAndUnresolved(CoordCommandUtils.java:404) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils. >>>>>>>>>>>>>>> > materializeInputDataEvents(CoordCommandUtils.java:731) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.materializeOneInstance( >>>>>>>>>>>>>>> > CoordCommandUtils.java:546) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordMaterializeTransitionXCom >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> mand.materializeActions(CoordMaterializeTransitionXCommand.java:492) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordMaterializeTransitionXCom >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> mand.materialize(CoordMaterializeTransitionXCommand.java:362) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute( >>>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:73) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute( >>>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:29) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> org.apache.oozie.command.XCommand.call(XCommand.java:290) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> org.apache.oozie.service.CallableQueueService$CallableWrapper.run( >>>>>>>>>>>>>>> > CallableQueueService.java:181) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>>>>>>>>>>> > ThreadPoolExecutor.java:1149) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at >>>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>>>>>>>>>>> > ThreadPoolExecutor.java:624) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > at java.lang.Thread.run(Thread.java:748) >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > Is S3 support specific to CDH distribution or should it >>>>>>>>>>>>>>> work in Apache >>>>>>>>>>>>>>> > Oozie as well? I’m not using CDH yet so >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > On Wed, May 16, 2018 at 10:28 AM Peter Cseh < >>>>>>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > > I think it should be possible for Oozie to poll S3. >>>>>>>>>>>>>>> Check out this >>>>>>>>>>>>>>> > > < >>>>>>>>>>>>>>> > > https://www.cloudera.com/documentation/enterprise/5-9- >>>>>>>>>>>>>>> > x/topics/admin_oozie_s3.html >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > description on how to make it work in jobs, something >>>>>>>>>>>>>>> similar should work >>>>>>>>>>>>>>> > > on the server side as well >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > On Tue, May 15, 2018 at 4:43 PM, purna pradeep < >>>>>>>>>>>>>>> purna2prad...@gmail.com> >>>>>>>>>>>>>>> > > wrote: >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > > Thanks Andras, >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > Also I also would like to know if oozie supports Aws >>>>>>>>>>>>>>> S3 as input events >>>>>>>>>>>>>>> > > to >>>>>>>>>>>>>>> > > > poll for a dependency file before kicking off a spark >>>>>>>>>>>>>>> action >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > For example: I don’t want to kick off a spark action >>>>>>>>>>>>>>> until a file is >>>>>>>>>>>>>>> > > > arrived on a given AWS s3 location >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > On Tue, May 15, 2018 at 10:17 AM Andras Piros < >>>>>>>>>>>>>>> > andras.pi...@cloudera.com >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > wrote: >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > > > Hi, >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > Oozie needs HDFS to store workflow, coordinator, or >>>>>>>>>>>>>>> bundle >>>>>>>>>>>>>>> > definitions, >>>>>>>>>>>>>>> > > > as >>>>>>>>>>>>>>> > > > > well as sharelib files in a safe, distributed and >>>>>>>>>>>>>>> scalable way. Oozie >>>>>>>>>>>>>>> > > > needs >>>>>>>>>>>>>>> > > > > YARN to run almost all of its actions, Spark action >>>>>>>>>>>>>>> being no >>>>>>>>>>>>>>> > exception. >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > At the moment it's not feasible to install Oozie >>>>>>>>>>>>>>> without those Hadoop >>>>>>>>>>>>>>> > > > > components. How to install Oozie please *find here >>>>>>>>>>>>>>> > > > > <https://oozie.apache.org/docs/5.0.0/AG_Install.html >>>>>>>>>>>>>>> >*. >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > Regards, >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > Andras >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > On Tue, May 15, 2018 at 4:11 PM, purna pradeep < >>>>>>>>>>>>>>> > > purna2prad...@gmail.com> >>>>>>>>>>>>>>> > > > > wrote: >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > > > Hi, >>>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>>> > > > > > Would like to know if I can use sparkaction in >>>>>>>>>>>>>>> oozie without having >>>>>>>>>>>>>>> > > > > Hadoop >>>>>>>>>>>>>>> > > > > > cluster? >>>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>>> > > > > > I want to use oozie to schedule spark jobs on >>>>>>>>>>>>>>> Kubernetes cluster >>>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>>> > > > > > I’m a beginner in oozie >>>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>>> > > > > > Thanks >>>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>>> > > > >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > -- >>>>>>>>>>>>>>> > > *Peter Cseh *| Software Engineer >>>>>>>>>>>>>>> > > cloudera.com <https://www.cloudera.com> >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > > [image: Cloudera on Twitter] < >>>>>>>>>>>>>>> https://twitter.com/cloudera> [image: >>>>>>>>>>>>>>> > > Cloudera on Facebook] <https://www.facebook.com/cloudera> >>>>>>>>>>>>>>> [image: >>>>>>>>>>>>>>> > Cloudera >>>>>>>>>>>>>>> > > on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>>>>> > > ------------------------------ >>>>>>>>>>>>>>> > > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>>>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> >>>>>>>>>>>>>>> [image: >>>>>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> >>>>>>>>>>>>>>> [image: Cloudera >>>>>>>>>>>>>>> on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>>>>> ------------------------------ >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>>>>>>> >>>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>>> >>>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> >>>>>>>>>>>>> [image: >>>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>>>>>>>>> Cloudera on LinkedIn] >>>>>>>>>>>>> <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>>> ------------------------------ >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>>> >>>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>> >>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>>> ------------------------------ >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> *Peter Cseh *| Software Engineer >>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>> >>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>> >>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>> ------------------------------ >>>>>>> >>>>>>> >>>>> >>>>> >>>>> -- >>>>> *Peter Cseh *| Software Engineer >>>>> cloudera.com <https://www.cloudera.com> >>>>> >>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>> >>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>> ------------------------------ >>>>> >>>>>