Can anyone help on how to set  custom parameters  after ranger setup ?


On Thu, Jan 5, 2017 at 12:43 AM, Anandha L Ranganathan <
analog.s...@gmail.com> wrote:

>
> We are just facing another problem to set  custom parameters.  How do we
> set these parameters in beeline at runtime ?    These are out custom
> parameters.
>
> SET airflow_cluster=${env:CLUSTER};
> SET default_date=unix_timestamp('1970-01-01 00:00:00');
> SET default_timestamp=CAST('1970-01-01 00:00:00' AS TIMESTAMP);
> SET default_future_date=unix_timestamp('2099-12-31 00:00:00');
>
> We get these errors when we set these parameters.
>
> 0: jdbc:hive2://usw2dbdpmn01:10000/> SET default_timestamp=CAST('1970-01-01
> 00:00:00' AS TIMESTAMP);
> Error: Error while processing statement: Cannot modify default_timestamp
> at runtime. It is not in list of params that are allowed to be modified at
> runtime (state=42000,code=1)
>
>
> Thanks
> Anand
>
>
> On Mon, Dec 19, 2016 at 5:34 PM, Anandha L Ranganathan <
> analog.s...@gmail.com> wrote:
>
>> Cool. After adding  the configuration it is working fine.
>>
>> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.authorization.sq
>> lstd.confwhitelist.append;
>> +-----------------------------------------------------------
>> -------------------------+--+
>> |                                        set
>> |
>> +-----------------------------------------------------------
>> -------------------------+--+
>> | 
>> hive.security.authorization.sqlstd.confwhitelist.append=|fs\.s3a\..*|fs\.s3n\..*
>> |  |
>> +-----------------------------------------------------------
>> -------------------------+--+
>>
>>
>> Thanks Selva for the quick help.
>>
>>
>>
>> On Mon, Dec 19, 2016 at 5:29 PM, Selvamohan Neethiraj <
>> sneet...@apache.org> wrote:
>>
>>> Hi,
>>>
>>> Can you try appending the following string to the  existing value of
>>>  hive.security.authorization.sqlstd.confwhitelist
>>>
>>> |fs\.s3a\..*
>>>
>>> And restart the HiveServer2 to see if this fixes this issue ?
>>>
>>> Thanks,
>>> Selva-
>>> From: Anandha L Ranganathan <analog.s...@gmail.com>
>>> Reply-To: "user@ranger.incubator.apache.org" <
>>> user@ranger.incubator.apache.org>
>>> Date: Monday, December 19, 2016 at 6:27 PM
>>>
>>> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org
>>> >
>>> Subject: Re: Unable to connect to S3 after enabling Ranger with Hive
>>>
>>> Selva,
>>>
>>> Please find the results.
>>>
>>> set hive.security.authorization.sqlstd.confwhitelist;
>>>
>>> | hive.security.authorization.sqlstd.confwhitelist=hive\.auto\
>>> ..*|hive\.cbo\..*|hive\.convert\..*|hive\.exec\.dynamic\.par
>>> tition.*|hive\.exec\..*\.dynamic\.partitions\..*|hive\.exec\
>>> .compress\..*|hive\.exec\.infer\..*|hive\.exec\.mode.loc
>>> al\..*|hive\.exec\.orc\..*|hive\.exec\.parallel.*|hive\.expl
>>> ain\..*|hive\.fetch.task\..*|hive\.groupby\..*|hive\.hbase\
>>> ..*|hive\.index\..*|hive\.index\..*|hive\.intermediate\.
>>> .*|hive\.join\..*|hive\.limit\..*|hive\.log\..*|hive\.
>>> mapjoin\..*|hive\.merge\..*|hive\.optimize\..*|hive\.orc\.
>>> .*|hive\.outerjoin\..*|hive\.parquet\..*|hive\.ppd\..*|
>>> hive\.prewarm\..*|hive\.server2\.proxy\.user|hive\.
>>> skewjoin\..*|hive\.smbjoin\..*|hive\.stats\..*|hive\.tez\..*
>>> |hive\.vectorized\..*|mapred\.map\..*|mapred\.reduce\..*|
>>> mapred\.output\.compression\.codec|mapred\.job\.queuename|
>>> mapred\.output\.compression\.type|mapred\.min\.split\.size|
>>> mapreduce\.job\.reduce\.slowstart\.completedmaps|
>>> mapreduce\.job\.queuename|mapreduce\.job\.tags|
>>> mapreduce\.input\.fileinputformat\.split\.minsize|mapreduce\
>>> .map\..*|mapreduce\.reduce\..*|mapreduce\.output\.fileoutput
>>> format\.compress\.codec|mapreduce\.output\.fileoutputf
>>> ormat\.compress\.type|tez\.am\..*|tez\.task\..*|tez\.runtime\..*|
>>> tez.queue.name|hive\.exec\.reducers\.bytes\.per\.
>>> reducer|hive\.client\.stats\.counters|hive\.exec\.default\.
>>> partition\.name|hive\.exec\.drop\.ignorenonexistent|hive\.
>>> counters\.group\.name|hive\.default\.fileformat\.managed|
>>> hive\.enforce\.bucketing|hive\.enforce\.bucketmapjoin|hive\.
>>> enforce\.sorting|hive\.enforce\.sortmergebucketmapjoi
>>> n|hive\.cache\.expr\.evaluation|hive\.hashtable\.
>>> loadfactor|hive\.hashtable\.initialCapacity|hive\.ignore\.
>>> mapjoin\.hint|hive\.limit\.row\.max\.size|hive\.mapred\.
>>> mode|hive\.map\.aggr|hive\.compute\.query\.using\.stats|
>>> hive\.exec\.rowoffset|hive\.variable\.substitute|hive\.
>>> variable\.substitute\.depth|hive\.autogen\.columnalias\.
>>> prefix\.includefuncname|hive\.autogen\.columnalias\.prefix\.
>>> label|hive\.exec\.check\.crossproducts|hive\.compat|
>>> hive\.exec\.concatenate\.check\.index|hive\.display\.
>>> partition\.cols\.separately|hive\.error\.on\.empty\.
>>> partition|hive\.execution\.engine|hive\.exim\.uri\.
>>> scheme\.whitelist|hive\.file\.max\.footer|hive\.mapred\.
>>> supports\.subdirectories|hive\.insert\.into\.multilevel\.
>>> dirs|hive\.localize\.resource\.num\.wait\.attempts|hive\.
>>> multi\.insert\.move\.tasks\.share\.dependencies|hive\.
>>> support\.quoted\.identifiers|hive\.resultset\.use\.unique\.
>>> column\.names|hive\.analyze\.stmt\.collect\.partlevel\.
>>> stats|hive\.server2\.logging\.operation\.level|hive\.
>>> support\.sql11\.reserved\.keywords|hive\.exec\.job\.
>>> debug\.capture\.stacktraces|hive\.exec\.job\.debug\.
>>> timeout|hive\.exec\.max\.created\.files|hive\.exec\.
>>> reducers\.max|hive\.reorder\.nway\.joins|hive\.output\.
>>> file\.extension|hive\.exec\.show\.job\.failure\.debug\.
>>> info|hive\.exec\.tasklog\.debug\.timeout  |
>>>
>>>
>>>
>>> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.authorization.sq
>>> lstd.confwhitelist.append;
>>> +-----------------------------------------------------------
>>> ------------+--+
>>> |                                  set
>>> |
>>> +-----------------------------------------------------------
>>> ------------+--+
>>> | hive.security.authorization.sqlstd.confwhitelist.append is undefined
>>> |
>>> +-----------------------------------------------------------
>>> ------------+--+
>>>
>>>
>>> On Mon, Dec 19, 2016 at 3:12 PM, Selvamohan Neethiraj <
>>> sneet...@apache.org> wrote:
>>>
>>>> Hi,
>>>>
>>>> Can you also post here the value for the following two parameters:
>>>>
>>>> hive.security.authorization.sqlstd.confwhitelist
>>>>
>>>> hive.security.authorization.sqlstd.confwhitelist.append
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Selva-
>>>>
>>>> From: Anandha L Ranganathan <analog.s...@gmail.com>
>>>> Reply-To: "user@ranger.incubator.apache.org" <
>>>> user@ranger.incubator.apache.org>
>>>> Date: Monday, December 19, 2016 at 5:54 PM
>>>> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.
>>>> org>
>>>> Subject: Re: Unable to connect to S3 after enabling Ranger with Hive
>>>>
>>>> Selva,
>>>>
>>>> We are using HDP and here are versions and results.
>>>>
>>>> Hive :  1.2.1.2.4
>>>> Ranger: 0.5.0.2.4
>>>>
>>>>
>>>>
>>>> 0: jdbc:hive2://usw2dxdpmn01:10010> set  hive.conf.restricted.list;
>>>> +-----------------------------------------------------------
>>>> ------------------------------------------------------------
>>>> -----------------+--+
>>>> |
>>>> set                                                                   |
>>>> +-----------------------------------------------------------
>>>> ------------------------------------------------------------
>>>> -----------------+--+
>>>> | hive.conf.restricted.list=hive.security.authorization.enable
>>>> d,hive.security.authorization.manager,hive.security.authenticator.manager
>>>> |
>>>> +-----------------------------------------------------------
>>>> ------------------------------------------------------------
>>>> -----------------+--+
>>>> 1 row selected (0.006 seconds)
>>>> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.command.whitelis
>>>> t;
>>>> +-----------------------------------------------------------
>>>> --------------------+--+
>>>> |                                      set
>>>> |
>>>> +-----------------------------------------------------------
>>>> --------------------+--+
>>>> | 
>>>> hive.security.command.whitelist=set,reset,dfs,add,list,delete,reload,compile
>>>> |
>>>> +-----------------------------------------------------------
>>>> --------------------+--+
>>>> 1 row selected (0.008 seconds)
>>>>
>>>>
>>>>
>>>>
>>>> 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key=xxxxxxxxxxxx
>>>> xxx;
>>>> Error: Error while processing statement: Cannot modify
>>>> fs.s3a.access.key at runtime. It is not in list of params that are allowed
>>>> to be modified at runtime (state=42000,code=1)
>>>>
>>>> On Mon, Dec 19, 2016 at 2:47 PM, Selvamohan Neethiraj <
>>>> sneet...@apache.org> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Which version of Hive and Ranger are you using ? Can you check if
>>>>> Ranger has added  hiveserver2 parameters  hive.conf.restricted.list,
>>>>> hive.security.command.whitelist  in the hive configuration file(s) ?
>>>>> Can you please list out these parameter values here ?
>>>>>
>>>>> Thanks,
>>>>> Selva-
>>>>>
>>>>> From: Anandha L Ranganathan <analog.s...@gmail.com>
>>>>> Reply-To: "user@ranger.incubator.apache.org" <
>>>>> user@ranger.incubator.apache.org>
>>>>> Date: Monday, December 19, 2016 at 5:30 PM
>>>>> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.
>>>>> org>
>>>>> Subject: Unable to connect to S3 after enabling Ranger with Hive
>>>>>
>>>>> Hi,
>>>>>
>>>>>
>>>>> Unable to create table pointing to S3 after enabling Ranger.
>>>>>
>>>>> This is database we created before enabling Ranger.
>>>>>
>>>>>
>>>>>    1. SET fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>>>>>    2. SET fs.s3a.access.key=xxxxxxx;
>>>>>    3. SET fs.s3a.secret.key=yyyyyyyyyyyyyyy;
>>>>>    4.
>>>>>    5.
>>>>>    6. CREATE DATABASE IF NOT EXISTS backup_s3a1
>>>>>    7. COMMENT "s3a schema test"
>>>>>    8. LOCATION "s3a://gd-de-dp-db-hcat-backup-schema/";
>>>>>
>>>>> After Ranger was enabled, we try to create another database but it is
>>>>> throwing error.
>>>>>
>>>>>
>>>>>    1. 0: jdbc:hive2://usw2dxdpmn01.local:> SET 
>>>>> fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>>>>>    2. Error: Error while processing statement: Cannot modify fs.s3a.impl 
>>>>> at runtime. It is not in list of params that are allowed to be modified 
>>>>> at runtime (state=42000,code=1)
>>>>>    3.
>>>>>
>>>>>
>>>>>
>>>>> I configured the credentials in the core-site.xml and always returns
>>>>> "undefined" when I am trying to see the values for  below commands. This 
>>>>> is
>>>>> in our " dev" environment where Ranger is enabled. In   other environment
>>>>> where Ranger is not installed , we are not facing this problem.
>>>>>
>>>>>
>>>>>    1. 0: jdbc:hive2://usw2dxdpmn01:10010> set  fs.s3a.impl;
>>>>>    2. +-----------------------------------------------------+--+
>>>>>    3. |                         set                         |
>>>>>    4. +-----------------------------------------------------+--+
>>>>>    5. | fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem  |
>>>>>    6. +-----------------------------------------------------+--+
>>>>>    7. 1 row selected (0.006 seconds)
>>>>>    8. 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key;
>>>>>    9. +---------------------------------+--+
>>>>>    10. |               set               |
>>>>>    11. +---------------------------------+--+
>>>>>    12. | fs.s3a.access.key is undefined  |
>>>>>    13. +---------------------------------+--+
>>>>>    14. 1 row selected (0.005 seconds)
>>>>>    15. 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.secret.key;
>>>>>    16. +---------------------------------+--+
>>>>>    17. |               set               |
>>>>>    18. +---------------------------------+--+
>>>>>    19. | fs.s3a.secret.key is undefined  |
>>>>>    20. +---------------------------------+--+
>>>>>    21. 1 row selected (0.005 seconds)
>>>>>
>>>>>
>>>>> Any help or pointers is appreciated.
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to