Re: Processing rows in parallel with MapReduce jobs.

2016-04-18 Thread Ted Yu
Did you see the "Message to log?" log ? Can you pastebin the error / exception you got ? On Mon, Apr 18, 2016 at 1:54 AM, Ivan Cores gonzalez wrote: > > > Hi Ted, > So, If I understand the behaviour of getSplits(), I can create "virtual" > splits overriding the

Re: Java or Scala Namespace collisions using multiple versions of the hadoop or spark hbase api

2016-04-18 Thread Stack
Tell us more Colin? You want to launch the two clients inside the same process? Use asynchbase if you want to go against the two different versions using the same client; it allows you do this. St.Ack On Fri, Apr 15, 2016 at 2:19 PM, Colin Kincaid Williams wrote: > Hello, > >

Re: Java or Scala Namespace collisions using multiple versions of the hadoop or spark hbase api

2016-04-18 Thread Colin Kincaid Williams
bump! On Fri, Apr 15, 2016 at 9:19 PM, Colin Kincaid Williams wrote: > Hello, > > I'm in the process of transferring data between Hbase .92 and Hbase > .98 . I'd like to write some test comparing my databases to ensure I > have not lost content in the transformation process. From

Re: maven dependency resources for hbase .98 and apache-spark

2016-04-18 Thread Colin Kincaid Williams
After some googling I found some dependency links. I will give them a shot, and ask elsewhere if I'm still confused. On Mon, Apr 18, 2016 at 7:48 PM, Ted Yu wrote: > The referenced link is from a specific vendor. > > Mind posting on the vendor's mailing list ? > > On Mon,

Re: maven dependency resources for hbase .98 and apache-spark

2016-04-18 Thread Ted Yu
The referenced link is from a specific vendor. Mind posting on the vendor's mailing list ? On Mon, Apr 18, 2016 at 12:45 PM, Colin Kincaid Williams wrote: > I would like to insert some data from Spark and or Spark streaming > into Hbase, on v .98. I found this section of the

maven dependency resources for hbase .98 and apache-spark

2016-04-18 Thread Colin Kincaid Williams
I would like to insert some data from Spark and or Spark streaming into Hbase, on v .98. I found this section of the book which shows examples of using the apis: https://archive.cloudera.com/cdh5/cdh/5/hbase/book.html#spark . However I'm unsure what dependency sections I need to add to my maven

Re: Best way to pass configuration properties to MRv2 jobs

2016-04-18 Thread Dima Spivak
Probably better off asking on the Hadoop user mailing list ( u...@hadoop.apache.org) than the HBase oneā€¦ :) -Dima On Mon, Apr 18, 2016 at 2:57 AM, Henning Blohm wrote: > Hi, > > in our Hadoop 2.6.0 cluster, we need to pass some properties to all Hadoop > processes so

Best way to pass configuration properties to MRv2 jobs

2016-04-18 Thread Henning Blohm
Hi, in our Hadoop 2.6.0 cluster, we need to pass some properties to all Hadoop processes so they can be referenced using ${...} syntax in configuration files. This works reasonably well using HADOOP_NAMENODE_OPTS and the like. For Map/Reduce jobs however, we need to speficy not only

Re: Processing rows in parallel with MapReduce jobs.

2016-04-18 Thread Ivan Cores gonzalez
Hi Ted, So, If I understand the behaviour of getSplits(), I can create "virtual" splits overriding the getSplits function. I was performing some tests, but my code crash in runtime and I cannot found the problem. Any help? I didn't find examples. public class SimpleRowCounter extends