Setting the below configuration in the MR made it work

conf.set("hbase.coprocessor.region.classes", "");

Thanks everyone!

On Tue, Apr 26, 2016 at 8:53 AM, Sanooj Padmakumar <p.san...@gmail.com>
wrote:

> Apologies if its against the convention to re-open an old discussion.
>
> The exception was resolved after I added
> "conf.set("fs.permissions.umask-mode", "000");" as per the suggestion
> provided by Gabriel.
>
> However the same exception has re-surfaced again in a new kerberized
> cluster setup and I am unable to get it going. Some differences between the
> earlier try and the current one is that earlier I used 4.3.0 version of
> Phoenix and now I am on 4.5.2. I tried the same program in a non-kerberized
> cluster with the 4.5.2 of Phoenix and that works.
>
> I already tried setting file permission as shown above, and also running
> as hbase user, but nothing helped.
>
> Pasting the exception below.
>
> 16/04/25 11:25:02 INFO mapreduce.LoadIncrementalHFiles: Trying to load
> hfile=hdfs://nameservice1/tmp/output/account7/ACCOUNT/0/f26e809c916a45198e19cd63a4971d65
> first=<<masked>> last=<<masked>>
> 16/04/25 11:26:11 INFO client.RpcRetryingCaller: Call exception, tries=10,
> retries=35, started=68213 ms ago, cancelled=false, msg=row '' on table
> 'ACCOUNT' at
> region=ACCOUNT,,1461124051363.cbb334b331f00e23a6b83b85685cf793.,
> hostname=<<masked>>,60020,1461243982079, seqNum=10
> 16/04/25 11:26:31 INFO client.RpcRetryingCaller: Call exception, tries=11,
> retries=35, started=88415 ms ago, cancelled=false, msg=row '' on table
> 'ACCOUNT' at
> region=ACCOUNT,,1461124051363.cbb334b331f00e23a6b83b85685cf793.,
> hostname=<<masked>>,60020,1461243982079, seqNum=10
>
>
> Any points to debug this will help.
>
> Thanks
> Sanooj
>
> On Wed, Nov 18, 2015 at 8:26 PM, Gabriel Reid <gabriel.r...@gmail.com>
> wrote:
>
>> Re-adding the user list, which I accidentally left off.
>>
>> On Wed, Nov 18, 2015 at 3:55 PM, Gabriel Reid <gabriel.r...@gmail.com>
>> wrote:
>> > Yes, I believe that's correct, if you change the umask you make the
>> > HFiles readable to all during creation.
>> >
>> > I believe that the alternate solutions listed on the jira ticket
>> > (running the tool as the hbase user or using the alternate HBase
>> > coprocessor for loading HFiles) won't have this drawback.
>> >
>> > - Gabriel
>> >
>> >
>> > On Wed, Nov 18, 2015 at 3:49 PM, Sanooj Padmakumar <p.san...@gmail.com>
>> wrote:
>> >> Thank you Gabriel..
>> >>
>> >> Does it mean that the generated hfile be read/modified by any user
>> other
>> >> than "hbase" user?
>> >>
>> >> Regards
>> >> Sanooj
>> >>
>> >> On 18 Nov 2015 02:39, "Gabriel Reid" <gabriel.r...@gmail.com> wrote:
>> >>>
>> >>> Hi Sanooj,
>> >>>
>> >>> Yes, I think that should do it, or you can pass that config parameter
>> >>> as a command line parameter.
>> >>>
>> >>> - Gabriel
>> >>>
>> >>> On Tue, Nov 17, 2015 at 8:16 PM, Sanooj Padmakumar <
>> p.san...@gmail.com>
>> >>> wrote:
>> >>> > Hi Gabriel
>> >>> >
>> >>> > Thank you so much
>> >>> >
>> >>> > I set the below property and it worked now.. I hope this is the
>> correct
>> >>> > thing to do ?
>> >>> >
>> >>> >  conf.set("fs.permissions.umask-mode", "000");
>> >>> >
>> >>> >
>> >>> > Thanks Again
>> >>> >
>> >>> > Sanooj
>> >>> >
>> >>> > On Wed, Nov 18, 2015 at 12:29 AM, Gabriel Reid <
>> gabriel.r...@gmail.com>
>> >>> > wrote:
>> >>> >>
>> >>> >> Hi Sanooj,
>> >>> >>
>> >>> >> I believe that this is related to the issue described in
>> PHOENIX-976
>> >>> >> [1]. In that case, it's not strictly related to Kerberos, but
>> instead
>> >>> >> to file permissions (could it be that your dev environment also
>> >>> >> doesn't have file permissions turned on?)
>> >>> >>
>> >>> >> If you look at the comments on that jira ticket, there are a
>> couple of
>> >>> >> things that you could try doing to resolve this (running the import
>> >>> >> job as the hbase user, or using custom file permissions, or using
>> an
>> >>> >> alternate incremental load coprocessor).
>> >>> >>
>> >>> >> - Gabriel
>> >>> >>
>> >>> >>
>> >>> >> 1. https://issues.apache.org/jira/browse/PHOENIX-976
>> >>> >>
>> >>> >> On Tue, Nov 17, 2015 at 7:14 PM, Sanooj Padmakumar <
>> p.san...@gmail.com>
>> >>> >> wrote:
>> >>> >> > Hello -
>> >>> >> >
>> >>> >> > I am using the bulkload of Phoenix on a cluster secured with
>> >>> >> > Kerberos.
>> >>> >> > The
>> >>> >> > mapper runs fine, reducer runs fine .. and then the counters are
>> >>> >> > printed
>> >>> >> > fine.. finally the LoadIncrementalHFiles steps fails.. A portion
>> of
>> >>> >> > the
>> >>> >> > log
>> >>> >> > is given below..
>> >>> >> >
>> >>> >> >
>> >>> >> > 15/11/17 09:44:48 INFO mapreduce.LoadIncrementalHFiles: Trying to
>> >>> >> > load
>> >>> >> > hfile=hdfs://..........<<masked>>>
>> >>> >> > 15/11/17 09:45:56 INFO client.RpcRetryingCaller: Call exception,
>> >>> >> > tries=10,
>> >>> >> > retries=35, started=68220 ms ago, cancelled=false, msg=row '' on
>> >>> >> > table
>> >>> >> > 'TABLE1' at region=TABLE1,<<<masked>>>>, seqNum=26
>> >>> >> > 15/11/17 09:46:16 INFO client.RpcRetryingCaller: Call exception,
>> >>> >> > tries=11,
>> >>> >> > retries=35, started=88315 ms ago, cancelled=false, msg=row '' on
>> >>> >> > table
>> >>> >> > 'TABLE1' at region=TABLE1,<<<masked>>>>, seqNum=26
>> >>> >> >
>> >>> >> > Is there any setting I should make inorder to make the program
>> work
>> >>> >> > on
>> >>> >> > Kerberos secured environment ?
>> >>> >> >
>> >>> >> > Please note , our DEV environment doesnt use Kerberos and things
>> are
>> >>> >> > working
>> >>> >> > just fine
>> >>> >> >
>> >>> >> > --
>> >>> >> > Thanks in advance,
>> >>> >> > Sanooj Padmakumar
>> >>> >
>> >>> >
>> >>> >
>> >>> >
>> >>> > --
>> >>> > Thanks,
>> >>> > Sanooj Padmakumar
>>
>
>
>
> --
> Thanks,
> Sanooj Padmakumar
>



-- 
Thanks,
Sanooj Padmakumar

Reply via email to