Ravi,

You likely need to build the Hbase connectors with the MapR libraries as
well.

Can you please change your ~/.m2/settings.xml so that you have:

    <profile>
      <id>mapr</id>
      <repositories>
        <repository>
          <id>mapr-repo</id>
          <name>MapR Repository</name>
          <url>http://repository.mapr.com/maven/</url>
          <releases>
            <enabled>true</enabled>
          </releases>
          <snapshots>
            <enabled>false</enabled>
          </snapshots>
        </repository>
      </repositories>
    </profile>

and then compile nifi (without Sumo's patch) using:

mvn <whatever you usually use> -Pmapr
-Dhadoop.version=2.7.0-mapr-1602 (assuming you are running 5.1 - otherwise
check the docs for the appropriate version)


This should cause hbase connector to be compiled with the MapR libraries as
dependencies.

Cheers

On Thu, Jun 16, 2016 at 9:32 AM, Ravi Papisetti (rpapiset) <
rpapi...@cisco.com> wrote:

> Thanks Sumo. We are able to use 0.6.1.nar file to connect with maps 5.1.
>
> Can we use HBase_1_1_2_ClientService to connect with Hbase 0.98.12? Also,
> again here we are using mapr cluster (mapr db). Connection fails with below
> error. Any thoughts?
>
> 2016-06-15 18:22:33,131 ERROR [StandardProcessScheduler Thread-8]
> o.a.n.c.s.StandardControllerServiceNode Failed to invoke @OnEnabled method
> of HBase_1_1_2_ClientService[id=076c05f1-a2b9-4e6a-803b-c5eb76da1c6d] due
> to org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after
> attempts=1, exceptions:
>
> Wed Jun 15 18:22:33 CDT 2016,
> RpcRetryingCaller{globalStartTime=1466032935495, pause=100, retries=1},
> org.apache.hadoop.hbase.MasterNotRunningException:
> org.apache.hadoop.hbase.MasterNotRunningException: Can't get connection to
> ZooKeeper: KeeperErrorCode = ConnectionLoss for /hbase
>
>
> Thanks,
>
> Ravi Papisetti
>
> Technical Leader
>
> Services Technology Incubation Center
> <http://wwwin.cisco.com/CustAdv/ts/cstg/stic/>
>
> rpapi...@cisco.com
>
> Phone: +1 512 340 3377
>
>
> [image: stic-logo-email-blue]
>
> From: Sumanth Chinthagunta <xmlk...@gmail.com>
> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
> Date: Wednesday, June 15, 2016 at 5:16 PM
> To: "users@nifi.apache.org" <users@nifi.apache.org>
> Subject: Re: Writing files to MapR File system using putHDFS
>
> Just want to conform that:
> After fresh MapR Client (Secure) installation on NiFi host, I was able to
> use putHDFS from NiFi *without* any core-site.xml changes I described in
> mapr-client.md
> we are using custom MapR nifi-hadoop-libraries-nar-0.6.1.nar
> <https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/releases/download/0.6.2/nifi-hadoop-libraries-nar-0.6.1.nar>
>  bundle
> and /opt/mapr/conf/mapr-clusters.conf now.
> If Matt’s PR is released, ideally we don’t have to do any
> custom configuration  to make NiFi work with MapR
>
> PS: my MapR env is 4.0.2
> -Sumo
>
> On Jun 14, 2016, at 3:55 PM, Andre <andre-li...@fucs.org> wrote:
>
> Ravi,
>
> "1.
> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md
>  –
> has a step to configure fs.defaultFS, should I configure the cluster
> resource manage here? Or should I get maprfs url from our IT who is
> supporting the cluster?"
>
>
> Depends on your configuration, but generally the fs.defaultFS can be
> configured just by using your own settings (and pointing the NiFi processor
> to your settings).
>
> Ideally you should be connecting to a cluster defined under 
> /opt/mapr/conf/mapr-clusters.conf
> (assuming you haven't played with client configs)
>
> Also have in mind that mapr-client.md file seems to be focused on
> insecure MapR clusters. Are you using insecure (terrible choice), MapR
> Hybrid security (not ideal IMHO) or Kerberos (recommended IMHO)?
>
> Cheers
>
> On Wed, Jun 15, 2016 at 3:36 AM, Ravi Papisetti (rpapiset) <
> rpapi...@cisco.com> wrote:
>
>> Hi,
>>
>> I have configure the way it is mentioned in below e-mail, still no
>> luck:-(.
>>
>> I have two questions:
>> 1.
>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md
>>  –
>> has a step to configure fs.defaultFS, should I configure the cluster
>> resource manage here? Or should I get maprfs url from our IT who is
>> supporting the cluster?
>> 2. Matt: What is MaprPR? Can you please elaborate
>>
>> Appreciate all your responses.
>>
>> Thanks,
>> Ravi Papisetti
>> Technical Leader
>> Services Technology Incubation Center
>> <http://wwwin.cisco.com/CustAdv/ts/cstg/stic/>
>> rpapi...@cisco.com
>> Phone: +1 512 340 3377
>>
>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[28].png>
>>
>> From: Matt Burgess <mattyb...@gmail.com>
>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>> Date: Monday, June 13, 2016 at 8:26 PM
>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>> Subject: Re: Writing files to MapR File system using putHDFS
>>
>> Sumo,
>>
>> I'll try the MapR PR with your additional settings below. If they work,
>> they'll need to be added to the doc (or ideally, the profile if possible).
>> That's what I suspected had been missing but haven't had a chance to try
>> yet, will do that shortly :)
>>
>> Thanks,
>> Matt
>>
>> On Jun 13, 2016, at 9:17 PM, Sumanth Chinthagunta <xmlk...@gmail.com>
>> wrote:
>>
>>
>> I had been using custom build nifi-hadoop-libraries-nar-0.6.1.nar that
>> worked with MapR 4.02
>> make sure you add java.security.auth.login.config and follow mapR client
>> setup on the NiFi server (
>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md)
>>
>>
>> $NIFI_HOME/conf/bootstrap.conf
>>
>>
>> java.arg.15=-Djava.security.auth.login.config=/opt/mapr/conf/mapr.login.conf
>> I just build *Nar* with MapR 2.7.0-mapr-1602 libs. I haven’t tested with
>> MapR 5.1 but you can try and let us know.
>>
>> Hadoop bundle for NiFi v0.6.1 and MapR 2.7.0-mapr-1602
>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/releases
>>
>> -Sumo
>>
>>
>> On Jun 13, 2016, at 5:57 PM, Bryan Bende <bbe...@gmail.com> wrote:
>>
>> I'm not sure if this would make a difference, but typically the
>> configuration resources would be the full paths to core-site.xml and
>> hdfs-site.xml. Wondering if using those instead of yarn-site.xml changes
>> anything.
>>
>> On Monday, June 13, 2016, Ravi Papisetti (rpapiset) <rpapi...@cisco.com>
>> wrote:
>>
>>> Yes, Aldrin. Tried listHDFS, gets the similar error complaining
>>> directory doesn't exist.
>>>
>>> NiFi – 0.6.1
>>> MapR – 5.1
>>>
>>> NiFi is local standalone instance. Target cluster is enabled with token
>>> based authentication. I am able to execute "hadoop fs –ls <path>" from cli
>>> on the node with NiFi installed.
>>>
>>> Thanks,
>>> Ravi Papisetti
>>> Technical Leader
>>> Services Technology Incubation Center
>>> <http://wwwin.cisco.com/CustAdv/ts/cstg/stic/>
>>> rpapi...@cisco.com
>>> Phone: +1 512 340 3377
>>>
>>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[22].png>
>>>
>>> From: Aldrin Piri <aldrinp...@gmail.com>
>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>> Date: Monday, June 13, 2016 at 6:24 PM
>>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>>> Subject: Re: Writing files to MapR File system using putHDFS
>>>
>>> Hi Ravi,
>>>
>>> Could you provide some additional details in terms of both your NiFi
>>> environment and the MapR destination?
>>>
>>> Is your NiFi a single instance or clustered?  In the case of the latter,
>>> is security established for your ZooKeeper ensemble?
>>>
>>> Is your target cluster Kerberized?  What version are you running?  Have
>>> you attempted to use the List/GetHDFS processors?  Do they also have errors
>>> in reading?
>>>
>>> Thanks!
>>> --aldrin
>>>
>>> On Mon, Jun 13, 2016 at 5:19 PM, Ravi Papisetti (rpapiset) <
>>> rpapi...@cisco.com> wrote:
>>>
>>>> Thanks Conrad for your reply.
>>>>
>>>> Yes, I have configured putHDFS with "Remove Owner" and "Renive Group"
>>>> with same values as on HDFS. Also, nifi service is started under the same
>>>> user.
>>>>
>>>>
>>>>
>>>> Thanks,
>>>> Ravi Papisetti
>>>> Technical Leader
>>>> Services Technology Incubation Center
>>>> <http://wwwin.cisco.com/CustAdv/ts/cstg/stic/>
>>>> rpapi...@cisco.com
>>>> Phone: +1 512 340 3377
>>>>
>>>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[21].png>
>>>>
>>>> From: Conrad Crampton <conrad.cramp...@secdata.com>
>>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>> Date: Monday, June 13, 2016 at 4:01 PM
>>>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>> Subject: Re: Writing files to MapR File system using putHDFS
>>>>
>>>> Hi,
>>>>
>>>> Sounds like a permissions problem. Have you set the Remote Owner and
>>>> Remote Groups settings in the processor appropriate for HDFS permissions?
>>>>
>>>> Conrad
>>>>
>>>>
>>>>
>>>> *From: *"Ravi Papisetti (rpapiset)" <rpapi...@cisco.com>
>>>> *Reply-To: *"users@nifi.apache.org" <users@nifi.apache.org>
>>>> *Date: *Monday, 13 June 2016 at 21:25
>>>> *To: *"users@nifi.apache.org" <users@nifi.apache.org>, "
>>>> d...@nifi.apache.org" <d...@nifi.apache.org>
>>>> *Subject: *Writing files to MapR File system using putHDFS
>>>>
>>>>
>>>>
>>>> Hi,
>>>>
>>>>
>>>>
>>>> We just started exploring apache nifi for data onboarding into MapR
>>>> distribution. Have configured putHDFS with yarn-site.xml from on local mapr
>>>> client where cluster information is provided, configured the "Directory"
>>>> with mapr fs directory to write the files, configured nifi to run as user
>>>> has permission to write to mapr fs, inspie of that we are getting below
>>>> error while writing the file into given file system path. I am doubting,
>>>> nifi is not talking to the cluster or talking with wrong user, appreciate
>>>> if you some can guide me to troubleshoot this issue or any solutions if we
>>>> are doing something wrong:
>>>>
>>>> Nifi workflow is very simple: GetFile is configure to read from locla
>>>> file system, connected to PutHDFS with yarn-site.xml and directory
>>>> information configured.
>>>>
>>>> 2016-06-13 15:14:36,305 INFO [Timer-Driven Process Thread-2]
>>>> o.apache.nifi.processors.hadoop.PutHDFS
>>>> PutHDFS[id=07abcfaa-fa8d-496b-81f0-b1b770672719] Kerberos relogin
>>>> successful or ticket still valid
>>>> 2016-06-13 15:14:36,324 ERROR [Timer-Driven Process Thread-2]
>>>> o.apache.nifi.processors.hadoop.PutHDFS
>>>> PutHDFS[id=07abcfaa-fa8d-496b-81f0-b1b770672719] Failed to write to HDFS
>>>> due to java.io.IOException: /app/DataAnalyticsFramework/catalog/nifi could
>>>> not be created: java.io.IOException:
>>>> /app/DataAnalyticsFramework/catalog/nifi could not be created
>>>> 2016-06-13 15:14:36,330 ERROR [Timer-Driven Process Thread-2]
>>>> o.apache.nifi.processors.hadoop.PutHDFS
>>>> java.io.IOException: /app/DataAnalyticsFramework/catalog/nifi could not
>>>> be created
>>>>         at
>>>> org.apache.nifi.processors.hadoop.PutHDFS.onTrigger(PutHDFS.java:238)
>>>> ~[nifi-hdfs-processors-0.6.1.jar:0.6.1]
>>>>         at
>>>> org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27)
>>>> [nifi-api-0.6.1.jar:0.6.1]
>>>>         at
>>>> org.apache.nifi.controller.StandardProcessorNode.onTrigger(StandardProcessorNode.java:1059)
>>>> [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>         at
>>>> org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:136)
>>>> [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>         at
>>>> org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:47)
>>>> [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>         at
>>>> org.apache.nifi.controller.scheduling.TimerDrivenSchedulingAgent$1.run(TimerDrivenSchedulingAgent.java:123)
>>>> [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>         at
>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>>>> [na:1.7.0_101]
>>>>         at
>>>> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304)
>>>> [na:1.7.0_101]
>>>>         at
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
>>>> [na:1.7.0_101]
>>>>         at
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>> [na:1.7.0_101]
>>>>         at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> [na:1.7.0_101]
>>>>         at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> [na:1.7.0_101]
>>>>         at java.lang.Thread.run(Thread.java:745) [na:1.7.0_101]
>>>>
>>>> Appreciate any help.
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Ravi Papisetti
>>>>
>>>> Technical Leader
>>>>
>>>> Services Technology Incubation Center
>>>> <http://wwwin.cisco.com/CustAdv/ts/cstg/stic/>
>>>>
>>>> rpapi...@cisco.com
>>>>
>>>> Phone: +1 512 340 3377
>>>>
>>>>
>>>>
>>>> <image001.png>
>>>>
>>>>
>>>>
>>>> ***This email originated outside SecureData***
>>>>
>>>> Click here <https://www.mailcontrol.com/sr/MZbqvYs5QwJvpeaetUwhCQ==>
>>>> to report this email as spam.
>>>>
>>>>
>>>> SecureData, combating cyber threats
>>>>
>>>> ------------------------------
>>>>
>>>> The information contained in this message or any of its attachments may
>>>> be privileged and confidential and intended for the exclusive use of the
>>>> intended recipient. If you are not the intended recipient any disclosure,
>>>> reproduction, distribution or other dissemination or use of this
>>>> communications is strictly prohibited. The views expressed in this email
>>>> are those of the individual and not necessarily of SecureData Europe Ltd.
>>>> Any prices quoted are only valid if followed up by a formal written quote.
>>>>
>>>> SecureData Europe Limited. Registered in England & Wales 04365896.
>>>> Registered Address: SecureData House, Hermitage Court, Hermitage Lane,
>>>> Maidstone, Kent, ME16 9NT
>>>>
>>>
>>>
>>
>> --
>> Sent from Gmail Mobile
>>
>>
>>
>
>

Reply via email to