Ok I got passed this error

By rebuilding oozie with Dhttpclient.version=4.5.5 -Dhttpcore.version=4.4.9

now getting this error



ACTION[0000000-180517144113498-oozie-xjt0-C@1]
org.apache.oozie.service.HadoopAccessorException: E0902: Exception
occurred: [doesBucketExist on mybucketcom.amazonaws.AmazonClientException:
No AWS Credentials provided by BasicAWSCredentialsProvider
EnvironmentVariableCredentialsProvider
SharedInstanceProfileCredentialsProvider :
com.amazonaws.SdkClientException: Unable to load credentials from service
endpoint]

org.apache.oozie.service.HadoopAccessorException: E0902: Exception
occurred: [doesBucketExist on cmsegmentation-qa:
com.amazonaws.AmazonClientException: No AWS Credentials provided by
BasicAWSCredentialsProvider EnvironmentVariableCredentialsProvider
SharedInstanceProfileCredentialsProvider :
com.amazonaws.SdkClientException: Unable to load credentials from service
endpoint]

On Thu, May 17, 2018 at 12:24 PM purna pradeep <purna2prad...@gmail.com>
wrote:

>
> Peter,
>
> Also When I submit a job with new http client jar, I get
>
> ```Error: IO_ERROR : java.io.IOException: Error while connecting Oozie
> server. No of retries = 1. Exception = Could not authenticate,
> Authentication failed, status: 500, message: Server Error```
>
>
> On Thu, May 17, 2018 at 12:14 PM purna pradeep <purna2prad...@gmail.com>
> wrote:
>
>> Ok I have tried this
>>
>> It appears that s3a support requires httpclient 4.4.x and oozie is
>> bundled with httpclient 4.3.6. When httpclient is upgraded, the ext UI
>> stops loading.
>>
>>
>>
>> On Thu, May 17, 2018 at 10:28 AM Peter Cseh <gezap...@cloudera.com>
>> wrote:
>>
>>> Purna,
>>>
>>> Based on
>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3
>>> you should try to go for s3a.
>>> You'll have to include the aws-jdk as well if I see it correctly:
>>> https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html#S3A
>>> Also, the property names are slightly different so you'll have to change
>>> the example I've given.
>>>
>>>
>>>
>>> On Thu, May 17, 2018 at 4:16 PM, purna pradeep <purna2prad...@gmail.com>
>>> wrote:
>>>
>>>> Peter,
>>>>
>>>> I’m using latest oozie 5.0.0 and I have tried below changes but no luck
>>>>
>>>> Is this for s3 or s3a ?
>>>>
>>>> I’m using s3 but if this is for s3a do you know which jar I need to
>>>> include I mean Hadoop-aws jar or any other jar if required
>>>>
>>>> Hadoop-aws-2.8.3.jar is what I’m using
>>>>
>>>> On Wed, May 16, 2018 at 5:19 PM Peter Cseh <gezap...@cloudera.com>
>>>> wrote:
>>>>
>>>>> Ok, I've found it:
>>>>>
>>>>> If you are using 4.3.0 or newer this is the part which checks for
>>>>> dependencies:
>>>>>
>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/command/coord/CoordCommandUtils.java#L914-L926
>>>>> It passes the coordinator action's configuration and even does
>>>>> impersonation to check for the dependencies:
>>>>>
>>>>> https://github.com/apache/oozie/blob/master/core/src/main/java/org/apache/oozie/coord/input/logic/CoordInputLogicEvaluatorPhaseOne.java#L159
>>>>>
>>>>> Have you tried the following in the coordinator xml:
>>>>>
>>>>>  <action>
>>>>>         <workflow>
>>>>>           <app-path>hdfs://bar:9000/usr/joe/logsprocessor-wf</app-path>
>>>>>           <configuration>
>>>>>             <property>
>>>>>               <name>fs.s3.awsAccessKeyId</name>
>>>>>               <value>[YOURKEYID]</value>
>>>>>             </property>
>>>>>             <property>
>>>>>               <name>fs.s3.awsSecretAccessKey</name>
>>>>>               <value>[YOURKEY]</value>
>>>>>             </property>
>>>>>          </configuration>
>>>>>        </workflow>
>>>>>       </action>
>>>>>
>>>>> Based on the source this should be able to poll s3 periodically.
>>>>>
>>>>> On Wed, May 16, 2018 at 10:57 PM, purna pradeep <
>>>>> purna2prad...@gmail.com> wrote:
>>>>>
>>>>>>
>>>>>> I have tried with coordinator's configuration too but no luck ☹️
>>>>>>
>>>>>> On Wed, May 16, 2018 at 3:54 PM Peter Cseh <gezap...@cloudera.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Great progress there purna! :)
>>>>>>>
>>>>>>> Have you tried adding these properites to the coordinator's
>>>>>>> configuration? we usually use the action config to build up connection 
>>>>>>> to
>>>>>>> the distributed file system.
>>>>>>> Although I'm not sure we're using these when polling the
>>>>>>> dependencies for coordinators, but I'm excited about you trying to make 
>>>>>>> it
>>>>>>> work!
>>>>>>>
>>>>>>> I'll get back with a - hopefully - more helpful answer soon, I have
>>>>>>> to check the code in more depth first.
>>>>>>> gp
>>>>>>>
>>>>>>> On Wed, May 16, 2018 at 9:45 PM, purna pradeep <
>>>>>>> purna2prad...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Peter,
>>>>>>>>
>>>>>>>> I got rid of this error by adding
>>>>>>>> hadoop-aws-2.8.3.jar and jets3t-0.9.4.jar
>>>>>>>>
>>>>>>>> But I’m getting below error now
>>>>>>>>
>>>>>>>> java.lang.IllegalArgumentException: AWS Access Key ID and Secret
>>>>>>>> Access Key must be specified by setting the fs.s3.awsAccessKeyId and
>>>>>>>> fs.s3.awsSecretAccessKey properties (respectively)
>>>>>>>>
>>>>>>>> I have tried adding AWS access ,secret keys in
>>>>>>>>
>>>>>>>> oozie-site.xml and hadoop core-site.xml , and hadoop-config.xml
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, May 16, 2018 at 2:30 PM purna pradeep <
>>>>>>>> purna2prad...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> I have tried this ,just added s3 instead of *
>>>>>>>>>
>>>>>>>>> <property>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name>
>>>>>>>>>
>>>>>>>>>     <value>hdfs,hftp,webhdfs,s3</value>
>>>>>>>>>
>>>>>>>>> </property>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Getting below error
>>>>>>>>>
>>>>>>>>> java.lang.RuntimeException: java.lang.ClassNotFoundException:
>>>>>>>>> Class org.apache.hadoop.fs.s3a.S3AFileSystem not found
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2369)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2793)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2810)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:100)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2849)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2831)
>>>>>>>>>
>>>>>>>>>     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:389)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:625)
>>>>>>>>>
>>>>>>>>>     at
>>>>>>>>> org.apache.oozie.service.HadoopAccessorService$5.run(HadoopAccessorService.java:623
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Wed, May 16, 2018 at 2:19 PM purna pradeep <
>>>>>>>>> purna2prad...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> This is what is in the logs
>>>>>>>>>>
>>>>>>>>>> 2018-05-16 14:06:13,500  INFO URIHandlerService:520 -
>>>>>>>>>> SERVER[localhost] Loaded urihandlers
>>>>>>>>>> [org.apache.oozie.dependency.FSURIHandler]
>>>>>>>>>>
>>>>>>>>>> 2018-05-16 14:06:13,501  INFO URIHandlerService:520 -
>>>>>>>>>> SERVER[localhost] Loaded default urihandler
>>>>>>>>>> org.apache.oozie.dependency.FSURIHandler
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Wed, May 16, 2018 at 12:27 PM Peter Cseh <
>>>>>>>>>> gezap...@cloudera.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> That's strange, this exception should not happen in that case.
>>>>>>>>>>> Can you check the server logs for messages like this?
>>>>>>>>>>>         LOG.info("Loaded urihandlers {0}",
>>>>>>>>>>> Arrays.toString(classes));
>>>>>>>>>>>         LOG.info("Loaded default urihandler {0}",
>>>>>>>>>>> defaultHandler.getClass().getName());
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>> On Wed, May 16, 2018 at 5:47 PM, purna pradeep <
>>>>>>>>>>> purna2prad...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> This is what I already have in my oozie-site.xml
>>>>>>>>>>>>
>>>>>>>>>>>> <property>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> <name>oozie.service.HadoopAccessorService.supported.filesystems</name>
>>>>>>>>>>>>
>>>>>>>>>>>>         <value>*</value>
>>>>>>>>>>>>
>>>>>>>>>>>> </property>
>>>>>>>>>>>>
>>>>>>>>>>>> On Wed, May 16, 2018 at 11:37 AM Peter Cseh <
>>>>>>>>>>>> gezap...@cloudera.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> You'll have to configure
>>>>>>>>>>>>> oozie.service.HadoopAccessorService.supported.filesystems
>>>>>>>>>>>>> hdfs,hftp,webhdfs Enlist
>>>>>>>>>>>>> the different filesystems supported for federation. If
>>>>>>>>>>>>> wildcard "*" is
>>>>>>>>>>>>> specified, then ALL file schemes will be allowed.properly.
>>>>>>>>>>>>>
>>>>>>>>>>>>> For testing purposes it's ok to put * in there in
>>>>>>>>>>>>> oozie-site.xml
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wed, May 16, 2018 at 5:29 PM, purna pradeep <
>>>>>>>>>>>>> purna2prad...@gmail.com>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> > Peter,
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > I have tried to specify dataset with uri starting with
>>>>>>>>>>>>> s3://, s3a:// and
>>>>>>>>>>>>> > s3n:// and I am getting exception
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > Exception occurred:E0904: Scheme [s3] not supported in uri
>>>>>>>>>>>>> > [s3://mybucket/input.data] Making the job failed
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > org.apache.oozie.dependency.URIHandlerException: E0904:
>>>>>>>>>>>>> Scheme [s3] not
>>>>>>>>>>>>> > supported in uri [s3:// mybucket /input.data]
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler(
>>>>>>>>>>>>> > URIHandlerService.java:185)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler(
>>>>>>>>>>>>> > URIHandlerService.java:168)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.service.URIHandlerService.getURIHandler(
>>>>>>>>>>>>> > URIHandlerService.java:160)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> >
>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.createEarlyURIs(
>>>>>>>>>>>>> > CoordCommandUtils.java:465)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils.
>>>>>>>>>>>>> > separateResolvedAndUnresolved(CoordCommandUtils.java:404)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordCommandUtils.
>>>>>>>>>>>>> > materializeInputDataEvents(CoordCommandUtils.java:731)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> >
>>>>>>>>>>>>> org.apache.oozie.command.coord.CoordCommandUtils.materializeOneInstance(
>>>>>>>>>>>>> > CoordCommandUtils.java:546)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordMaterializeTransitionXCom
>>>>>>>>>>>>> >
>>>>>>>>>>>>> mand.materializeActions(CoordMaterializeTransitionXCommand.java:492)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > org.apache.oozie.command.coord.CoordMaterializeTransitionXCom
>>>>>>>>>>>>> > mand.materialize(CoordMaterializeTransitionXCommand.java:362)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> >
>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute(
>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:73)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> >
>>>>>>>>>>>>> org.apache.oozie.command.MaterializeTransitionXCommand.execute(
>>>>>>>>>>>>> > MaterializeTransitionXCommand.java:29)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> org.apache.oozie.command.XCommand.call(XCommand.java:290)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> >
>>>>>>>>>>>>> org.apache.oozie.service.CallableQueueService$CallableWrapper.run(
>>>>>>>>>>>>> > CallableQueueService.java:181)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>>>>>>>> > ThreadPoolExecutor.java:1149)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at
>>>>>>>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>>>>>>>>>> > ThreadPoolExecutor.java:624)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >     at java.lang.Thread.run(Thread.java:748)
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > Is S3 support specific to CDH distribution or should it work
>>>>>>>>>>>>> in Apache
>>>>>>>>>>>>> > Oozie as well? I’m not using CDH yet so
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > On Wed, May 16, 2018 at 10:28 AM Peter Cseh <
>>>>>>>>>>>>> gezap...@cloudera.com> wrote:
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > > I think it should be possible for Oozie to poll S3. Check
>>>>>>>>>>>>> out this
>>>>>>>>>>>>> > > <
>>>>>>>>>>>>> > > https://www.cloudera.com/documentation/enterprise/5-9-
>>>>>>>>>>>>> > x/topics/admin_oozie_s3.html
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > description on how to make it work in jobs, something
>>>>>>>>>>>>> similar should work
>>>>>>>>>>>>> > > on the server side as well
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > > On Tue, May 15, 2018 at 4:43 PM, purna pradeep <
>>>>>>>>>>>>> purna2prad...@gmail.com>
>>>>>>>>>>>>> > > wrote:
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > > > Thanks Andras,
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > > Also I also would like to know if oozie supports Aws S3
>>>>>>>>>>>>> as input events
>>>>>>>>>>>>> > > to
>>>>>>>>>>>>> > > > poll for a dependency file before kicking off a spark
>>>>>>>>>>>>> action
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > > For example: I don’t want to kick off a spark action
>>>>>>>>>>>>> until a file is
>>>>>>>>>>>>> > > > arrived on a given AWS s3 location
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > > On Tue, May 15, 2018 at 10:17 AM Andras Piros <
>>>>>>>>>>>>> > andras.pi...@cloudera.com
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > > wrote:
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > > > > Hi,
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > Oozie needs HDFS to store workflow, coordinator, or
>>>>>>>>>>>>> bundle
>>>>>>>>>>>>> > definitions,
>>>>>>>>>>>>> > > > as
>>>>>>>>>>>>> > > > > well as sharelib files in a safe, distributed and
>>>>>>>>>>>>> scalable way. Oozie
>>>>>>>>>>>>> > > > needs
>>>>>>>>>>>>> > > > > YARN to run almost all of its actions, Spark action
>>>>>>>>>>>>> being no
>>>>>>>>>>>>> > exception.
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > At the moment it's not feasible to install Oozie
>>>>>>>>>>>>> without those Hadoop
>>>>>>>>>>>>> > > > > components. How to install Oozie please *find here
>>>>>>>>>>>>> > > > > <https://oozie.apache.org/docs/5.0.0/AG_Install.html
>>>>>>>>>>>>> >*.
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > Regards,
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > Andras
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > On Tue, May 15, 2018 at 4:11 PM, purna pradeep <
>>>>>>>>>>>>> > > purna2prad...@gmail.com>
>>>>>>>>>>>>> > > > > wrote:
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > > > > Hi,
>>>>>>>>>>>>> > > > > >
>>>>>>>>>>>>> > > > > > Would like to know if I can use sparkaction in oozie
>>>>>>>>>>>>> without having
>>>>>>>>>>>>> > > > > Hadoop
>>>>>>>>>>>>> > > > > > cluster?
>>>>>>>>>>>>> > > > > >
>>>>>>>>>>>>> > > > > > I want to use oozie to schedule spark jobs on
>>>>>>>>>>>>> Kubernetes cluster
>>>>>>>>>>>>> > > > > >
>>>>>>>>>>>>> > > > > > I’m a beginner in oozie
>>>>>>>>>>>>> > > > > >
>>>>>>>>>>>>> > > > > > Thanks
>>>>>>>>>>>>> > > > > >
>>>>>>>>>>>>> > > > >
>>>>>>>>>>>>> > > >
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > > --
>>>>>>>>>>>>> > > *Peter Cseh *| Software Engineer
>>>>>>>>>>>>> > > cloudera.com <https://www.cloudera.com>
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > > [image: Cloudera] <https://www.cloudera.com/>
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> > > [image: Cloudera on Twitter] <https://twitter.com/cloudera>
>>>>>>>>>>>>> [image:
>>>>>>>>>>>>> > > Cloudera on Facebook] <https://www.facebook.com/cloudera>
>>>>>>>>>>>>> [image:
>>>>>>>>>>>>> > Cloudera
>>>>>>>>>>>>> > > on LinkedIn] <https://www.linkedin.com/company/cloudera>
>>>>>>>>>>>>> > > ------------------------------
>>>>>>>>>>>>> > >
>>>>>>>>>>>>> >
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> *Peter Cseh *| Software Engineer
>>>>>>>>>>>>> cloudera.com <https://www.cloudera.com>
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/>
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera>
>>>>>>>>>>>>> [image:
>>>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera>
>>>>>>>>>>>>> [image: Cloudera
>>>>>>>>>>>>> on LinkedIn] <https://www.linkedin.com/company/cloudera>
>>>>>>>>>>>>> ------------------------------
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> *Peter Cseh *| Software Engineer
>>>>>>>>>>> cloudera.com <https://www.cloudera.com>
>>>>>>>>>>>
>>>>>>>>>>> [image: Cloudera] <https://www.cloudera.com/>
>>>>>>>>>>>
>>>>>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image:
>>>>>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image:
>>>>>>>>>>> Cloudera on LinkedIn]
>>>>>>>>>>> <https://www.linkedin.com/company/cloudera>
>>>>>>>>>>> ------------------------------
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> *Peter Cseh *| Software Engineer
>>>>>>> cloudera.com <https://www.cloudera.com>
>>>>>>>
>>>>>>> [image: Cloudera] <https://www.cloudera.com/>
>>>>>>>
>>>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image:
>>>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image:
>>>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera>
>>>>>>> ------------------------------
>>>>>>>
>>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> *Peter Cseh *| Software Engineer
>>>>> cloudera.com <https://www.cloudera.com>
>>>>>
>>>>> [image: Cloudera] <https://www.cloudera.com/>
>>>>>
>>>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image:
>>>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image:
>>>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera>
>>>>> ------------------------------
>>>>>
>>>>>
>>>
>>>
>>> --
>>> *Peter Cseh *| Software Engineer
>>> cloudera.com <https://www.cloudera.com>
>>>
>>> [image: Cloudera] <https://www.cloudera.com/>
>>>
>>> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image:
>>> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image:
>>> Cloudera on LinkedIn] <https://www.linkedin.com/company/cloudera>
>>> ------------------------------
>>>
>>>

Reply via email to