Thanks for instructions Andre.
Should we build with master branch from github or 0.6.1 ?
Hope this helps me to use kites and hbase processors with MapR
-Sumo
Sent from my iPhone

> On Jun 15, 2016, at 4:53 PM, Andre <andre-li...@fucs.org> wrote:
> 
> Ravi,
> 
> You likely need to build the Hbase connectors with the MapR libraries as well.
> 
> Can you please change your ~/.m2/settings.xml so that you have:
> 
>     <profile>
>       <id>mapr</id>
>       <repositories>
>         <repository>
>           <id>mapr-repo</id>
>           <name>MapR Repository</name>
>           <url>http://repository.mapr.com/maven/</url>
>           <releases>
>             <enabled>true</enabled>
>           </releases>
>           <snapshots>
>             <enabled>false</enabled>
>           </snapshots>
>         </repository>
>       </repositories>
>     </profile>
> 
> and then compile nifi (without Sumo's patch) using:
> 
> mvn <whatever you usually use> -Pmapr -Dhadoop.version=2.7.0-mapr-1602 
> (assuming you are running 5.1 - otherwise check the docs for the appropriate 
> version)
> 
> 
> This should cause hbase connector to be compiled with the MapR libraries as 
> dependencies.
> 
> Cheers
> 
>> On Thu, Jun 16, 2016 at 9:32 AM, Ravi Papisetti (rpapiset) 
>> <rpapi...@cisco.com> wrote:
>> Thanks Sumo. We are able to use 0.6.1.nar file to connect with maps 5.1.
>> 
>> Can we use HBase_1_1_2_ClientService to connect with Hbase 0.98.12? Also, 
>> again here we are using mapr cluster (mapr db). Connection fails with below 
>> error. Any thoughts?
>> 
>> 2016-06-15 18:22:33,131 ERROR [StandardProcessScheduler Thread-8] 
>> o.a.n.c.s.StandardControllerServiceNode Failed to invoke @OnEnabled method 
>> of HBase_1_1_2_ClientService[id=076c05f1-a2b9-4e6a-803b-c5eb76da1c6d] due to 
>> org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after 
>> attempts=1, exceptions:
>> Wed Jun 15 18:22:33 CDT 2016, 
>> RpcRetryingCaller{globalStartTime=1466032935495, pause=100, retries=1}, 
>> org.apache.hadoop.hbase.MasterNotRunningException: 
>> org.apache.hadoop.hbase.MasterNotRunningException: Can't get connection to 
>> ZooKeeper: KeeperErrorCode = ConnectionLoss for /hbase
>> 
>> Thanks,
>> Ravi Papisetti
>> Technical Leader
>> Services Technology Incubation Center
>> rpapi...@cisco.com
>> Phone: +1 512 340 3377
>> 
>> 
>> 
>> From: Sumanth Chinthagunta <xmlk...@gmail.com>
>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>> Date: Wednesday, June 15, 2016 at 5:16 PM
>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>> Subject: Re: Writing files to MapR File system using putHDFS
>> 
>> Just want to conform that: 
>> After fresh MapR Client (Secure) installation on NiFi host, I was able to 
>> use putHDFS from NiFi without any core-site.xml changes I described in 
>> mapr-client.md
>> we are using custom MapR nifi-hadoop-libraries-nar-0.6.1.nar bundle and 
>> /opt/mapr/conf/mapr-clusters.conf now. 
>> If Matt’s PR is released, ideally we don’t have to do any custom 
>> configuration  to make NiFi work with MapR 
>> 
>> PS: my MapR env is 4.0.2 
>> -Sumo 
>> 
>>> On Jun 14, 2016, at 3:55 PM, Andre <andre-li...@fucs.org> wrote:
>>> 
>>> Ravi,
>>> 
>>> "1. 
>>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md
>>>  – has a step to configure fs.defaultFS, should I configure the cluster 
>>> resource manage here? Or should I get maprfs url from our IT who is 
>>> supporting the cluster?"
>>> 
>>> Depends on your configuration, but generally the fs.defaultFS can be 
>>> configured just by using your own settings (and pointing the NiFi processor 
>>> to your settings).
>>> 
>>> Ideally you should be connecting to a cluster defined under 
>>> /opt/mapr/conf/mapr-clusters.conf (assuming you haven't played with client 
>>> configs)
>>> 
>>> Also have in mind that mapr-client.md file seems to be focused on insecure 
>>> MapR clusters. Are you using insecure (terrible choice), MapR Hybrid 
>>> security (not ideal IMHO) or Kerberos (recommended IMHO)?
>>> 
>>> Cheers
>>> 
>>>> On Wed, Jun 15, 2016 at 3:36 AM, Ravi Papisetti (rpapiset) 
>>>> <rpapi...@cisco.com> wrote:
>>>> Hi,
>>>> 
>>>> I have configure the way it is mentioned in below e-mail, still no luck:-(.
>>>> 
>>>> I have two questions:
>>>> 1. 
>>>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md
>>>>  – has a step to configure fs.defaultFS, should I configure the cluster 
>>>> resource manage here? Or should I get maprfs url from our IT who is 
>>>> supporting the cluster?
>>>> 2. Matt: What is MaprPR? Can you please elaborate
>>>> 
>>>> Appreciate all your responses.
>>>> 
>>>> Thanks,
>>>> Ravi Papisetti
>>>> Technical Leader
>>>> Services Technology Incubation Center
>>>> rpapi...@cisco.com
>>>> Phone: +1 512 340 3377
>>>> 
>>>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[28].png>
>>>> 
>>>> From: Matt Burgess <mattyb...@gmail.com>
>>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>> Date: Monday, June 13, 2016 at 8:26 PM
>>>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>> Subject: Re: Writing files to MapR File system using putHDFS
>>>> 
>>>> Sumo,
>>>> 
>>>> I'll try the MapR PR with your additional settings below. If they work, 
>>>> they'll need to be added to the doc (or ideally, the profile if possible). 
>>>> That's what I suspected had been missing but haven't had a chance to try 
>>>> yet, will do that shortly :)
>>>> 
>>>> Thanks,
>>>> Matt
>>>> 
>>>> On Jun 13, 2016, at 9:17 PM, Sumanth Chinthagunta <xmlk...@gmail.com> 
>>>> wrote:
>>>> 
>>>>> 
>>>>> I had been using custom build nifi-hadoop-libraries-nar-0.6.1.nar that 
>>>>> worked with MapR 4.02
>>>>> make sure you add java.security.auth.login.config and follow mapR client 
>>>>> setup on the NiFi server 
>>>>> (https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/blob/master/mapr-client.md)
>>>>>   
>>>>> $NIFI_HOME/conf/bootstrap.conf
>>>>> 
>>>>> java.arg.15=-Djava.security.auth.login.config=/opt/mapr/conf/mapr.login.conf
>>>>> 
>>>>> I just build Nar with MapR 2.7.0-mapr-1602 libs. I haven’t tested with 
>>>>> MapR 5.1 but you can try and let us know. 
>>>>>  
>>>>> Hadoop bundle for NiFi v0.6.1 and MapR 2.7.0-mapr-1602
>>>>> https://github.com/xmlking/mapr-nifi-hadoop-libraries-bundle/releases
>>>>> 
>>>>> -Sumo
>>>>> 
>>>>> 
>>>>>> On Jun 13, 2016, at 5:57 PM, Bryan Bende <bbe...@gmail.com> wrote:
>>>>>> 
>>>>>> I'm not sure if this would make a difference, but typically the 
>>>>>> configuration resources would be the full paths to core-site.xml and 
>>>>>> hdfs-site.xml. Wondering if using those instead of yarn-site.xml changes 
>>>>>> anything.
>>>>>> 
>>>>>>> On Monday, June 13, 2016, Ravi Papisetti (rpapiset) 
>>>>>>> <rpapi...@cisco.com> wrote:
>>>>>>> Yes, Aldrin. Tried listHDFS, gets the similar error complaining 
>>>>>>> directory doesn't exist.
>>>>>>> 
>>>>>>> NiFi – 0.6.1
>>>>>>> MapR – 5.1
>>>>>>> 
>>>>>>> NiFi is local standalone instance. Target cluster is enabled with token 
>>>>>>> based authentication. I am able to execute "hadoop fs –ls <path>" from 
>>>>>>> cli on the node with NiFi installed.
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Ravi Papisetti
>>>>>>> Technical Leader
>>>>>>> Services Technology Incubation Center
>>>>>>> rpapi...@cisco.com
>>>>>>> Phone: +1 512 340 3377
>>>>>>> 
>>>>>>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[22].png>
>>>>>>> 
>>>>>>> From: Aldrin Piri <aldrinp...@gmail.com>
>>>>>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>>>>> Date: Monday, June 13, 2016 at 6:24 PM
>>>>>>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>>>>> Subject: Re: Writing files to MapR File system using putHDFS
>>>>>>> 
>>>>>>> Hi Ravi,
>>>>>>> 
>>>>>>> Could you provide some additional details in terms of both your NiFi 
>>>>>>> environment and the MapR destination?
>>>>>>> 
>>>>>>> Is your NiFi a single instance or clustered?  In the case of the 
>>>>>>> latter, is security established for your ZooKeeper ensemble? 
>>>>>>> 
>>>>>>> Is your target cluster Kerberized?  What version are you running?  Have 
>>>>>>> you attempted to use the List/GetHDFS processors?  Do they also have 
>>>>>>> errors in reading? 
>>>>>>> 
>>>>>>> Thanks!
>>>>>>> --aldrin
>>>>>>> 
>>>>>>>> On Mon, Jun 13, 2016 at 5:19 PM, Ravi Papisetti (rpapiset) 
>>>>>>>> <rpapi...@cisco.com> wrote:
>>>>>>>> Thanks Conrad for your reply.
>>>>>>>> 
>>>>>>>> Yes, I have configured putHDFS with "Remove Owner" and "Renive Group" 
>>>>>>>> with same values as on HDFS. Also, nifi service is started under the 
>>>>>>>> same user.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> Ravi Papisetti
>>>>>>>> Technical Leader
>>>>>>>> Services Technology Incubation Center
>>>>>>>> rpapi...@cisco.com
>>>>>>>> Phone: +1 512 340 3377
>>>>>>>> 
>>>>>>>> <0B65E9AA-485A-49EE-9FB4-8600F4D55880[21].png>
>>>>>>>> 
>>>>>>>> From: Conrad Crampton <conrad.cramp...@secdata.com>
>>>>>>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>>>>>> Date: Monday, June 13, 2016 at 4:01 PM
>>>>>>>> To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>>>>>> Subject: Re: Writing files to MapR File system using putHDFS
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> 
>>>>>>>> Sounds like a permissions problem. Have you set the Remote Owner and 
>>>>>>>> Remote Groups settings in the processor appropriate for HDFS 
>>>>>>>> permissions?
>>>>>>>> 
>>>>>>>> Conrad
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> From: "Ravi Papisetti (rpapiset)" <rpapi...@cisco.com>
>>>>>>>> Reply-To: "users@nifi.apache.org" <users@nifi.apache.org>
>>>>>>>> Date: Monday, 13 June 2016 at 21:25
>>>>>>>> To: "users@nifi.apache.org" <users@nifi.apache.org>, 
>>>>>>>> "d...@nifi.apache.org" <d...@nifi.apache.org>
>>>>>>>> Subject: Writing files to MapR File system using putHDFS
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> We just started exploring apache nifi for data onboarding into MapR 
>>>>>>>> distribution. Have configured putHDFS with yarn-site.xml from on local 
>>>>>>>> mapr client where cluster information is provided, configured the 
>>>>>>>> "Directory" with mapr fs directory to write the files, configured nifi 
>>>>>>>> to run as user has permission to write to mapr fs, inspie of that we 
>>>>>>>> are getting below error while writing the  file into given file system 
>>>>>>>> path. I am doubting, nifi is not talking to the cluster or talking 
>>>>>>>> with wrong user, appreciate if you some can guide me to troubleshoot 
>>>>>>>> this issue or any solutions if we are doing something wrong:
>>>>>>>> 
>>>>>>>> Nifi workflow is very simple: GetFile is configure to read from locla 
>>>>>>>> file system, connected to PutHDFS with yarn-site.xml and directory 
>>>>>>>> information configured.
>>>>>>>> 
>>>>>>>> 2016-06-13 15:14:36,305 INFO [Timer-Driven Process Thread-2] 
>>>>>>>> o.apache.nifi.processors.hadoop.PutHDFS 
>>>>>>>> PutHDFS[id=07abcfaa-fa8d-496b-81f0-b1b770672719] Kerberos relogin 
>>>>>>>> successful or ticket still valid
>>>>>>>> 2016-06-13 15:14:36,324 ERROR [Timer-Driven Process Thread-2] 
>>>>>>>> o.apache.nifi.processors.hadoop.PutHDFS 
>>>>>>>> PutHDFS[id=07abcfaa-fa8d-496b-81f0-b1b770672719] Failed to write to 
>>>>>>>> HDFS due to java.io.IOException: 
>>>>>>>> /app/DataAnalyticsFramework/catalog/nifi could not be created: 
>>>>>>>> java.io.IOException: /app/DataAnalyticsFramework/catalog/nifi could 
>>>>>>>> not be created
>>>>>>>> 2016-06-13 15:14:36,330 ERROR [Timer-Driven Process Thread-2] 
>>>>>>>> o.apache.nifi.processors.hadoop.PutHDFS
>>>>>>>> java.io.IOException: /app/DataAnalyticsFramework/catalog/nifi could 
>>>>>>>> not be created
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.processors.hadoop.PutHDFS.onTrigger(PutHDFS.java:238) 
>>>>>>>> ~[nifi-hdfs-processors-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27)
>>>>>>>>  [nifi-api-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.controller.StandardProcessorNode.onTrigger(StandardProcessorNode.java:1059)
>>>>>>>>  [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:136)
>>>>>>>>  [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:47)
>>>>>>>>  [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> org.apache.nifi.controller.scheduling.TimerDrivenSchedulingAgent$1.run(TimerDrivenSchedulingAgent.java:123)
>>>>>>>>  [nifi-framework-core-0.6.1.jar:0.6.1]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>>>>>>>>  [na:1.7.0_101]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304) 
>>>>>>>> [na:1.7.0_101]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
>>>>>>>>  [na:1.7.0_101]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>  [na:1.7.0_101]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>>>  [na:1.7.0_101]
>>>>>>>>         at 
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>>>  [na:1.7.0_101]
>>>>>>>>         at java.lang.Thread.run(Thread.java:745) [na:1.7.0_101]
>>>>>>>> 
>>>>>>>> Appreciate any help.
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> 
>>>>>>>> Ravi Papisetti
>>>>>>>> 
>>>>>>>> Technical Leader
>>>>>>>> 
>>>>>>>> Services Technology Incubation Center
>>>>>>>> 
>>>>>>>> rpapi...@cisco.com
>>>>>>>> 
>>>>>>>> Phone: +1 512 340 3377
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> <image001.png>
>>>>>>>> 
>>>>>>>>  
>>>>>>>> 
>>>>>>>> ***This email originated outside SecureData***
>>>>>>>> 
>>>>>>>> Click here to report this email as spam.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> SecureData, combating cyber threats
>>>>>>>> 
>>>>>>>> The information contained in this message or any of its attachments 
>>>>>>>> may be privileged and confidential and intended for the exclusive use 
>>>>>>>> of the intended recipient. If you are not the intended recipient any 
>>>>>>>> disclosure, reproduction, distribution or other dissemination or use 
>>>>>>>> of this communications is strictly prohibited. The views expressed in 
>>>>>>>> this email are those of the individual and not necessarily of 
>>>>>>>> SecureData Europe Ltd. Any prices quoted are only valid if followed up 
>>>>>>>> by a formal written quote.
>>>>>>>> 
>>>>>>>> SecureData Europe Limited. Registered in England & Wales 04365896. 
>>>>>>>> Registered Address: SecureData House, Hermitage Court, Hermitage Lane, 
>>>>>>>> Maidstone, Kent, ME16 9NT
>>>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> -- 
>>>>>> Sent from Gmail Mobile
> 

Reply via email to