Ok I fixed this by adding aws keys in oozie But I’m getting below error
I have tried setting proxy in core-site.xml but no luck 2018-05-17 15:39:20,602 ERROR CoordInputLogicEvaluatorPhaseOne:517 - SERVER[localhost] USER[-] GROUP[-] TOKEN[-] APP[-] JOB[0000000-180517144113498-oozie-xjt0-C] ACTION[0000000- 180517144113498-oozie-xjt0-C@2] org.apache.oozie.service.HadoopAccessorException: E0902: Exception occurred: [doesBucketExist on cmsegmentation-qa: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to mybucket.s3.amazonaws.com:443 <http://cmsegmentation-qa.s3.amazonaws.com:443/> [mybucket. s3.amazonaws.com/52.216.165.155 <http://cmsegmentation-qa.s3.amazonaws.com/52.216.165.155>] failed: connect timed out] org.apache.oozie.service.HadoopAccessorException: E0902: Exception occurred: [doesBucketExist on cmsegmentation-qa: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to mybucket.s3.amazonaws.com:443 <http://cmsegmentation-qa.s3.amazonaws.com:443/> [mybucket.s3.amazonaws.com <http://cmsegmentation-qa.s3.amazonaws.com/52.216.165.155> failed: connect timed out] at org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:630) at org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:594) at org.apache.oozie.dependency.FSURIHandler.getFileSystem( FSURIHandler.java:184)-env.sh But now I’m getting this error On Thu, May 17, 2018 at 2:53 PM purna pradeep <purna2prad...@gmail.com> wrote: > Ok I got passed this error > > By rebuilding oozie with Dhttpclient.version=4.5.5 -Dhttpcore.version=4.4.9 > > now getting this error > > > > ACTION[0000000-180517144113498-oozie-xjt0-C@1] > org.apache.oozie.service.HadoopAccessorException: E0902: Exception > occurred: [doesBucketExist on mybucketcom.amazonaws.AmazonClientException: > No AWS Credentials provided by BasicAWSCredentialsProvider > EnvironmentVariableCredentialsProvider > SharedInstanceProfileCredentialsProvider : > com.amazonaws.SdkClientException: Unable to load credentials from service > endpoint] > > org.apache.oozie.service.HadoopAccessorException: E0902: Exception > occurred: [doesBucketExist on cmsegmentation-qa: > com.amazonaws.AmazonClientException: No AWS Credentials provided by > BasicAWSCredentialsProvider EnvironmentVariableCredentialsProvider > SharedInstanceProfileCredentialsProvider : > com.amazonaws.SdkClientException: Unable to load credentials from service > endpoint] > > On Thu, May 17, 2018 at 12:24 PM purna pradeep <purna2prad...@gmail.com> > wrote: > >> >> Peter, >> >> Also When I submit a job with new http client jar, I get >> >> ```Error: IO_ERROR : java.io.IOException: Error while connecting Oozie >> server. No of retries = 1. Exception = Could not authenticate, >> Authentication failed, status: 500, message: Server Error``` >> >> >> On Thu, May 17, 2018 at 12:14 PM purna pradeep <purna2prad...@gmail.com> >> wrote: >> >>> Ok I have tried this >>> >>> It appears that s3a support requires httpclient 4.4.x and oozie is >>> bundled with httpclient 4.3.6. When httpclient is upgraded, the ext UI >>> stops loading. >>> >>> >>> >>> On Thu, May 17, 2018 at 10:28 AM Peter Cseh <gezap...@cloudera.com> >>> wrote: >>> >>>> Purna, >>>> >>>> Based on >>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3 >>>> you should try to go for s3a. >>>> You'll have to include the aws-jdk as well if I see it correctly: >>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3A >>>> Also, the property names are slightly different so you'll have to >>>> change the example I've given. >>>> >>>> >>>> >>>> On Thu, May 17, 2018 at 4:16 PM, purna pradeep <purna2prad...@gmail.com >>>> > wrote: >>>> >>>>> Peter, >>>>> >>>>> I’m using latest oozie 5.0.0 and I have tried below changes but no >>>>> luck >>>>> >>>>> Is this for s3 or s3a ? >>>>> >>>>> I’m using s3 but if this is for s3a do you know which jar I need to >>>>> include I mean Hadoop-aws jar or any other jar if required >>>>> >>>>> Hadoop-aws-2.8.3.jar is what I’m using >>>>> >>>>> On Wed, May 16, 2018 at 5:19 PM Peter Cseh <gezap...@cloudera.com> >>>>> wrote: >>>>> >>>>>> Ok, I've found it: >>>>>> >>>>>> If you are using 4.3.0 or newer this is the part which checks for >>>>>> dependencies: >>>>>> >>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/command/coord/CoordCommandUtils.java#L914-L926 >>>>>> It passes the coordinator action's configuration and even does >>>>>> impersonation to check for the dependencies: >>>>>> >>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/coord/input/logic/CoordInputLogicEvaluatorPhaseOne.java#L159 >>>>>> >>>>>> Have you tried the following in the coordinator xml: >>>>>> >>>>>> <action> >>>>>> <workflow> >>>>>> >>>>>> <app-path>hdfs://bar:9000/usr/joe/logsprocessor-wf</app-path> >>>>>> <configuration> >>>>>> <property> >>>>>> <name>fs.s3.awsAccessKeyId</name> >>>>>> <value>[YOURKEYID]</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>fs.s3.awsSecretAccessKey</name> >>>>>> <value>[YOURKEY]</value> >>>>>> </property> >>>>>> </configuration> >>>>>> </workflow> >>>>>> </action> >>>>>> >>>>>> Based on the source this should be able to poll s3 periodically. >>>>>> >>>>>> On Wed, May 16, 2018 at 10:57 PM, purna pradeep < >>>>>> purna2prad...@gmail.com> wrote: >>>>>> >>>>>>> >>>>>>> I have tried with coordinator's configuration too but no luck ☹️ >>>>>>> >>>>>>> On Wed, May 16, 2018 at 3:54 PM Peter Cseh <gezap...@cloudera.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Great progress there purna! :) >>>>>>>> >>>>>>>> Have you tried adding these properites to the coordinator's >>>>>>>> configuration? we usually use the action config to build up connection >>>>>>>> to >>>>>>>> the distributed file system. >>>>>>>> Although I'm not sure we're using these when polling the >>>>>>>> dependencies for coordinators, but I'm excited about you trying to >>>>>>>> make it >>>>>>>> work! >>>>>>>> >>>>>>>> I'll get back with a - hopefully - more helpful answer soon, I have >>>>>>>> to check the code in more depth first. >>>>>>>> gp >>>>>>>> >>>>>>>> On Wed, May 16, 2018 at 9:45 PM, purna pradeep < >>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Peter, >>>>>>>>> >>>>>>>>> I got rid of this error by adding >>>>>>>>> hadoop-aws-2.8.3.jar and jets3t-0.9.4.jar >>>>>>>>> >>>>>>>>> But I’m getting below error now >>>>>>>>> >>>>>>>>> java.lang.IllegalArgumentException: AWS Access Key ID and Secret >>>>>>>>> Access Key must be specified by setting the fs.s3.awsAccessKeyId and >>>>>>>>> fs.s3.awsSecretAccessKey properties (respectively) >>>>>>>>> >>>>>>>>> I have tried adding AWS access ,secret keys in >>>>>>>>> >>>>>>>>> oozie-site.xml and hadoop core-site.xml , and hadoop-config.xml >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, May 16, 2018 at 2:30 PM purna pradeep < >>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> I have tried this ,just added s3 instead of * >>>>>>>>>> >>>>>>>>>> <property> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name> >>>>>>>>>> >>>>>>>>>> <value>hdfs,hftp,webhdfs,s3</value> >>>>>>>>>> >>>>>>>>>> </property> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Getting below error >>>>>>>>>> >>>>>>>>>> java.lang.RuntimeException: java.lang.ClassNotFoundException: >>>>>>>>>> Class org.apache.hadoop.fs.s3a.S3AFileSystem not found >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2369) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2793) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2810) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:100) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2849) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2831) >>>>>>>>>> >>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:389) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:625) >>>>>>>>>> >>>>>>>>>> at >>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:623 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 16, 2018 at 2:19 PM purna pradeep < >>>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> This is what is in the logs >>>>>>>>>>> >>>>>>>>>>> 2018-05-16 14:06:13,500 INFO URIHandlerService:520 - >>>>>>>>>>> SERVER[localhost] Loaded urihandlers >>>>>>>>>>> [org.apache.oozie.dependency.FSURIHandler] >>>>>>>>>>> >>>>>>>>>>> 2018-05-16 14:06:13,501 INFO URIHandlerService:520 - >>>>>>>>>>> SERVER[localhost] Loaded default urihandler >>>>>>>>>>> org.apache.oozie.dependency.FSURIHandler >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, May 16, 2018 at 12:27 PM Peter Cseh < >>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> That's strange, this exception should not happen in that case. >>>>>>>>>>>> Can you check the server logs for messages like this? >>>>>>>>>>>> LOG.info("Loaded urihandlers {0}", >>>>>>>>>>>> Arrays.toString(classes)); >>>>>>>>>>>> LOG.info("Loaded default urihandler {0}", >>>>>>>>>>>> defaultHandler.getClass().getName()); >>>>>>>>>>>> Thanks >>>>>>>>>>>> >>>>>>>>>>>> On Wed, May 16, 2018 at 5:47 PM, purna pradeep < >>>>>>>>>>>> purna2prad...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> This is what I already have in my oozie-site.xml >>>>>>>>>>>>> >>>>>>>>>>>>> <property> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name> >>>>>>>>>>>>> >>>>>>>>>>>>> <value>*</value> >>>>>>>>>>>>> >>>>>>>>>>>>> </property> >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, May 16, 2018 at 11:37 AM Peter Cseh < >>>>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> You'll have to configure >>>>>>>>>>>>>> oozie.service.HadoopAccessorService.supported.filesystems >>>>>>>>>>>>>> hdfs,hftp,webhdfs Enlist >>>>>>>>>>>>>> the different filesystems supported for federation. If >>>>>>>>>>>>>> wildcard "*" is >>>>>>>>>>>>>> specified, then ALL file schemes will be allowed.properly. >>>>>>>>>>>>>> >>>>>>>>>>>>>> For testing purposes it's ok to put * in there in >>>>>>>>>>>>>> oozie-site.xml >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Wed, May 16, 2018 at 5:29 PM, purna pradeep < >>>>>>>>>>>>>> purna2prad...@gmail.com> >>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> > Peter, >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > I have tried to specify dataset with uri starting with >>>>>>>>>>>>>> s3://, s3a:// and >>>>>>>>>>>>>> > s3n:// and I am getting exception >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > Exception occurred:E0904: Scheme [s3] not supported in uri >>>>>>>>>>>>>> > [s3://mybucket/input.data] Making the job failed >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > org.apache.oozie.dependency.URIHandlerException: E0904: >>>>>>>>>>>>>> Scheme [s3] not >>>>>>>>>>>>>> > supported in uri [s3:// mybucket /input.data] >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>> > URIHandlerService.java:185) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>> > URIHandlerService.java:168) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler( >>>>>>>>>>>>>> > URIHandlerService.java:160) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.createEarlyURIs( >>>>>>>>>>>>>> > CoordCommandUtils.java:465) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils. >>>>>>>>>>>>>> > separateResolvedAndUnresolved(CoordCommandUtils.java:404) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils. >>>>>>>>>>>>>> > materializeInputDataEvents(CoordCommandUtils.java:731) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.materializeOneInstance( >>>>>>>>>>>>>> > CoordCommandUtils.java:546) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordMaterializeTransitionXCom >>>>>>>>>>>>>> > >>>>>>>>>>>>>> mand.materializeActions(CoordMaterializeTransitionXCommand.java:492) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordMaterializeTransitionXCom >>>>>>>>>>>>>> > >>>>>>>>>>>>>> mand.materialize(CoordMaterializeTransitionXCommand.java:362) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute( >>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:73) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute( >>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:29) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> org.apache.oozie.command.XCommand.call(XCommand.java:290) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.oozie.service.CallableQueueService$CallableWrapper.run( >>>>>>>>>>>>>> > CallableQueueService.java:181) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>>>>>>>>>> > ThreadPoolExecutor.java:1149) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at >>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>>>>>>>>>> > ThreadPoolExecutor.java:624) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > at java.lang.Thread.run(Thread.java:748) >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > Is S3 support specific to CDH distribution or should it >>>>>>>>>>>>>> work in Apache >>>>>>>>>>>>>> > Oozie as well? I’m not using CDH yet so >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > On Wed, May 16, 2018 at 10:28 AM Peter Cseh < >>>>>>>>>>>>>> gezap...@cloudera.com> wrote: >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > > I think it should be possible for Oozie to poll S3. Check >>>>>>>>>>>>>> out this >>>>>>>>>>>>>> > > < >>>>>>>>>>>>>> > > https://www.cloudera.com/documentation/enterprise/5-9- >>>>>>>>>>>>>> > x/topics/admin_oozie_s3.html >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > description on how to make it work in jobs, something >>>>>>>>>>>>>> similar should work >>>>>>>>>>>>>> > > on the server side as well >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > On Tue, May 15, 2018 at 4:43 PM, purna pradeep < >>>>>>>>>>>>>> purna2prad...@gmail.com> >>>>>>>>>>>>>> > > wrote: >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > > Thanks Andras, >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > Also I also would like to know if oozie supports Aws S3 >>>>>>>>>>>>>> as input events >>>>>>>>>>>>>> > > to >>>>>>>>>>>>>> > > > poll for a dependency file before kicking off a spark >>>>>>>>>>>>>> action >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > For example: I don’t want to kick off a spark action >>>>>>>>>>>>>> until a file is >>>>>>>>>>>>>> > > > arrived on a given AWS s3 location >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > On Tue, May 15, 2018 at 10:17 AM Andras Piros < >>>>>>>>>>>>>> > andras.pi...@cloudera.com >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > wrote: >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > > > Hi, >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > Oozie needs HDFS to store workflow, coordinator, or >>>>>>>>>>>>>> bundle >>>>>>>>>>>>>> > definitions, >>>>>>>>>>>>>> > > > as >>>>>>>>>>>>>> > > > > well as sharelib files in a safe, distributed and >>>>>>>>>>>>>> scalable way. Oozie >>>>>>>>>>>>>> > > > needs >>>>>>>>>>>>>> > > > > YARN to run almost all of its actions, Spark action >>>>>>>>>>>>>> being no >>>>>>>>>>>>>> > exception. >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > At the moment it's not feasible to install Oozie >>>>>>>>>>>>>> without those Hadoop >>>>>>>>>>>>>> > > > > components. How to install Oozie please *find here >>>>>>>>>>>>>> > > > > <https://oozie.apache.org/docs/5.0.0/AG_Install.html >>>>>>>>>>>>>> >*. >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > Regards, >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > Andras >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > On Tue, May 15, 2018 at 4:11 PM, purna pradeep < >>>>>>>>>>>>>> > > purna2prad...@gmail.com> >>>>>>>>>>>>>> > > > > wrote: >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > > > Hi, >>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>> > > > > > Would like to know if I can use sparkaction in >>>>>>>>>>>>>> oozie without having >>>>>>>>>>>>>> > > > > Hadoop >>>>>>>>>>>>>> > > > > > cluster? >>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>> > > > > > I want to use oozie to schedule spark jobs on >>>>>>>>>>>>>> Kubernetes cluster >>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>> > > > > > I’m a beginner in oozie >>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>> > > > > > Thanks >>>>>>>>>>>>>> > > > > > >>>>>>>>>>>>>> > > > > >>>>>>>>>>>>>> > > > >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > -- >>>>>>>>>>>>>> > > *Peter Cseh *| Software Engineer >>>>>>>>>>>>>> > > cloudera.com <https://www.cloudera.com> >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > > [image: Cloudera on Twitter] < >>>>>>>>>>>>>> https://twitter.com/cloudera> [image: >>>>>>>>>>>>>> > > Cloudera on Facebook] <https://www.facebook.com/cloudera> >>>>>>>>>>>>>> [image: >>>>>>>>>>>>>> > Cloudera >>>>>>>>>>>>>> > > on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>>>> > > ------------------------------ >>>>>>>>>>>>>> > > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>>>> >>>>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> >>>>>>>>>>>>>> [image: >>>>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> >>>>>>>>>>>>>> [image: Cloudera >>>>>>>>>>>>>> on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>>>> ------------------------------ >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>>>>>> >>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>>>>>> >>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>>>>>>>> Cloudera on LinkedIn] >>>>>>>>>>>> <https://www.linkedin.com/company/cloudera> >>>>>>>>>>>> ------------------------------ >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> *Peter Cseh *| Software Engineer >>>>>>>> cloudera.com <https://www.cloudera.com> >>>>>>>> >>>>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>>>> >>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>>>> ------------------------------ >>>>>>>> >>>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> *Peter Cseh *| Software Engineer >>>>>> cloudera.com <https://www.cloudera.com> >>>>>> >>>>>> [image: Cloudera] <https://www.cloudera.com/> >>>>>> >>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>>>> ------------------------------ >>>>>> >>>>>> >>>> >>>> >>>> -- >>>> *Peter Cseh *| Software Engineer >>>> cloudera.com <https://www.cloudera.com> >>>> >>>> [image: Cloudera] <https://www.cloudera.com/> >>>> >>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image: >>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image: >>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera> >>>> ------------------------------ >>>> >>>>