Hey Ravi and James,

Do you have any updates on PhoenixHbaseLoader?
I am eager to test it out.

Thanks


On Mon, Mar 24, 2014 at 9:21 AM, local host
<[email protected]>wrote:

> Thanks Ravi for the update.
>
> If you can share some more info on the expected date for PhoenixHbaseLoader
> that will be great.
>
> I am eager to use Phoenix in my current project but I want to know about
> the extra work done by phoenix while inserting records in HBase table so
> that I can freely use other batch analysis tools such as pig, impala, hive.
> *In crux, I want to know if Phoenix and other tools are inter-operable.*
>
>
>
>
> On Fri, Mar 21, 2014 at 7:37 PM, Ravi Kiran <[email protected]>wrote:
>
>> Hi
>>
>>    We are currently working on having a PhoenixHbaseLoader to load data
>> from HBase using Pig.
>>
>> Regards
>> Ravi
>>
>>
>> On Sat, Mar 22, 2014 at 5:19 AM, local host <
>> [email protected]> wrote:
>>
>>> Hey All,
>>>
>>> *How can I analyze the Hbase data, which was inserted by jdbc phoenix,
>>> using Pig?*
>>> I wish to do batch processing on Hbase data using pig and correct the
>>> maintained counters.
>>>
>>> In crux, I want to know what extra work phoenix is doing in a HBase
>>> table at the time of insertion that requires some extra steps when I am
>>> analyzing it from other mapreduce tools such as hive, pig, dril etc.
>>>
>>>
>>> --UniLocal
>>>
>>
>>
>

Reply via email to