Thanks Ravi for the update. If you can share some more info on the expected date for PhoenixHbaseLoader that will be great.
I am eager to use Phoenix in my current project but I want to know about the extra work done by phoenix while inserting records in HBase table so that I can freely use other batch analysis tools such as pig, impala, hive. *In crux, I want to know if Phoenix and other tools are inter-operable.* On Fri, Mar 21, 2014 at 7:37 PM, Ravi Kiran <[email protected]>wrote: > Hi > > We are currently working on having a PhoenixHbaseLoader to load data > from HBase using Pig. > > Regards > Ravi > > > On Sat, Mar 22, 2014 at 5:19 AM, local host <[email protected] > > wrote: > >> Hey All, >> >> *How can I analyze the Hbase data, which was inserted by jdbc phoenix, >> using Pig?* >> I wish to do batch processing on Hbase data using pig and correct the >> maintained counters. >> >> In crux, I want to know what extra work phoenix is doing in a HBase table >> at the time of insertion that requires some extra steps when I am analyzing >> it from other mapreduce tools such as hive, pig, dril etc. >> >> >> --UniLocal >> > >
