Yes. It is ranger for hive.

On Thu, Jan 5, 2017 at 12:58 PM, Don Bosco Durai <bo...@apache.org> wrote:

> Is this for Ranger or Hive?
>
>
>
> Bosco
>
>
>
>
>
> *From: *Anandha L Ranganathan <analog.s...@gmail.com>
> *Reply-To: *<user@ranger.incubator.apache.org>
> *Date: *Thursday, January 5, 2017 at 12:16 PM
> *To: *<user@ranger.incubator.apache.org>
>
> *Subject: *Re: Unable to connect to S3 after enabling Ranger with Hive
>
>
>
> Can anyone help on how to set  custom parameters  after ranger setup ?
>
>
>
> On Thu, Jan 5, 2017 at 12:43 AM, Anandha L Ranganathan <
> analog.s...@gmail.com> wrote:
>
>
>
> We are just facing another problem to set  custom parameters.  How do we
> set these parameters in beeline at runtime ?    These are out custom
> parameters.
>
> SET airflow_cluster=${env:CLUSTER};
> SET default_date=unix_timestamp('1970-01-01 00:00:00');
> SET default_timestamp=CAST('1970-01-01 00:00:00' AS TIMESTAMP);
> SET default_future_date=unix_timestamp('2099-12-31 00:00:00');
>
> We get these errors when we set these parameters.
>
>
> 0: jdbc:hive2://usw2dbdpmn01:10000/> SET default_timestamp=CAST('1970-01-01
> 00:00:00' AS TIMESTAMP);
> Error: Error while processing statement: Cannot modify default_timestamp
> at runtime. It is not in list of params that are allowed to be modified at
> runtime (state=42000,code=1)
>
> Thanks
>
> Anand
>
>
>
>
>
> On Mon, Dec 19, 2016 at 5:34 PM, Anandha L Ranganathan <
> analog.s...@gmail.com> wrote:
>
> Cool. After adding  the configuration it is working fine.
>
> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.authorization.
> sqlstd.confwhitelist.append;
> +-----------------------------------------------------------
> -------------------------+--+
> |                                        set
> |
> +-----------------------------------------------------------
> -------------------------+--+
> | 
> hive.security.authorization.sqlstd.confwhitelist.append=|fs\.s3a\..*|fs\.s3n\..*
> |  |
> +-----------------------------------------------------------
> -------------------------+--+
>
> Thanks Selva for the quick help.
>
>
>
>
>
> On Mon, Dec 19, 2016 at 5:29 PM, Selvamohan Neethiraj <sneet...@apache.org>
> wrote:
>
> Hi,
>
>
>
> Can you try appending the following string to the  existing value of
>  hive.security.authorization.sqlstd.confwhitelist
>
>
>
> |fs\.s3a\..*
>
>
>
> And restart the HiveServer2 to see if this fixes this issue ?
>
>
>
> Thanks,
>
> Selva-
>
> *From: *Anandha L Ranganathan <analog.s...@gmail.com>
> *Reply-To: *"user@ranger.incubator.apache.org" <
> user@ranger.incubator.apache.org>
> *Date: *Monday, December 19, 2016 at 6:27 PM
>
>
> *To: *"user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org
> >
> *Subject: *Re: Unable to connect to S3 after enabling Ranger with Hive
>
>
>
> Selva,
>
> Please find the results.
>
>
> set hive.security.authorization.sqlstd.confwhitelist;
>
> | hive.security.authorization.sqlstd.confwhitelist=hive\.
> auto\..*|hive\.cbo\..*|hive\.convert\..*|hive\.exec\.
> dynamic\.partition.*|hive\.exec\..*\.dynamic\.partitions\
> ..*|hive\.exec\.compress\..*|hive\.exec\.infer\..*|hive\.
> exec\.mode.local\..*|hive\.exec\.orc\..*|hive\.exec\.
> parallel.*|hive\.explain\..*|hive\.fetch.task\..*|hive\.
> groupby\..*|hive\.hbase\..*|hive\.index\..*|hive\.index\..
> *|hive\.intermediate\..*|hive\.join\..*|hive\.limit\..*|
> hive\.log\..*|hive\.mapjoin\..*|hive\.merge\..*|hive\.
> optimize\..*|hive\.orc\..*|hive\.outerjoin\..*|hive\.
> parquet\..*|hive\.ppd\..*|hive\.prewarm\..*|hive\.
> server2\.proxy\.user|hive\.skewjoin\..*|hive\.smbjoin\..*
> |hive\.stats\..*|hive\.tez\..*|hive\.vectorized\..*|mapred\.
> map\..*|mapred\.reduce\..*|mapred\.output\.compression\.
> codec|mapred\.job\.queuename|mapred\.output\.compression\.
> type|mapred\.min\.split\.size|mapreduce\.job\.reduce\.
> slowstart\.completedmaps|mapreduce\.job\.queuename|mapreduce\.job\.tags|
> mapreduce\.input\.fileinputformat\.split\.minsize|mapreduce\.map\..*|
> mapreduce\.reduce\..*|mapreduce\.output\.fileoutputformat\.compress\.
> codec|mapreduce\.output\.fileoutputformat\.compress\.
> type|tez\.am\..*|tez\.task\..*|tez\.runtime\..*|tez.queue.name
> |hive\.exec\.reducers\.bytes\.per\.reducer|hive\.
> client\.stats\.counters|hive\.exec\.default\.partition\.
> name|hive\.exec\.drop\.ignorenonexistent|hive\.
> counters\.group\.name|hive\.default\.fileformat\.managed|
> hive\.enforce\.bucketing|hive\.enforce\.bucketmapjoin|hive\.
> enforce\.sorting|hive\.enforce\.sortmergebucketmapjoin|hive\.
> cache\.expr\.evaluation|hive\.hashtable\.loadfactor|hive\.
> hashtable\.initialCapacity|hive\.ignore\.mapjoin\.hint|
> hive\.limit\.row\.max\.size|hive\.mapred\.mode|hive\.map\.
> aggr|hive\.compute\.query\.using\.stats|hive\.exec\.
> rowoffset|hive\.variable\.substitute|hive\.variable\.
> substitute\.depth|hive\.autogen\.columnalias\.prefix\.
> includefuncname|hive\.autogen\.columnalias\.prefix\.label|
> hive\.exec\.check\.crossproducts|hive\.compat|hive\.exec\.concatenate\.
> check\.index|hive\.display\.partition\.cols\.separately|
> hive\.error\.on\.empty\.partition|hive\.execution\.
> engine|hive\.exim\.uri\.scheme\.whitelist|hive\.file\.
> max\.footer|hive\.mapred\.supports\.subdirectories|hive\
> .insert\.into\.multilevel\.dirs|hive\.localize\.resource\
> .num\.wait\.attempts|hive\.multi\.insert\.move\.tasks\.
> share\.dependencies|hive\.support\.quoted\.identifiers|
> hive\.resultset\.use\.unique\.column\.names|hive\.analyze\.
> stmt\.collect\.partlevel\.stats|hive\.server2\.logging\.
> operation\.level|hive\.support\.sql11\.reserved\.
> keywords|hive\.exec\.job\.debug\.capture\.stacktraces|
> hive\.exec\.job\.debug\.timeout|hive\.exec\.max\.
> created\.files|hive\.exec\.reducers\.max|hive\.reorder\.
> nway\.joins|hive\.output\.file\.extension|hive\.exec\.
> show\.job\.failure\.debug\.info|hive\.exec\.tasklog\.debug\.timeout  |
>
>
>
> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.authorization.
> sqlstd.confwhitelist.append;
> +-----------------------------------------------------------
> ------------+--+
> |                                  set                                  |
> +-----------------------------------------------------------
> ------------+--+
> | hive.security.authorization.sqlstd.confwhitelist.append is undefined  |
> +-----------------------------------------------------------
> ------------+--+
>
>
>
> On Mon, Dec 19, 2016 at 3:12 PM, Selvamohan Neethiraj <sneet...@apache.org>
> wrote:
>
> Hi,
>
>
>
> Can you also post here the value for the following two parameters:
>
>
>
> hive.security.authorization.sqlstd.confwhitelist
>
> hive.security.authorization.sqlstd.confwhitelist.append
>
>
>
>
>
> Thanks,
>
> Selva-
>
>
>
> *From: *Anandha L Ranganathan <analog.s...@gmail.com>
> *Reply-To: *"user@ranger.incubator.apache.org" <
> user@ranger.incubator.apache.org>
> *Date: *Monday, December 19, 2016 at 5:54 PM
> *To: *"user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org
> >
> *Subject: *Re: Unable to connect to S3 after enabling Ranger with Hive
>
>
>
> Selva,
>
> We are using HDP and here are versions and results.
>
> Hive :  1.2.1.2.4
>
> Ranger: 0.5.0.2.4
>
>
>
>
>
> 0: jdbc:hive2://usw2dxdpmn01:10010> set  hive.conf.restricted.list;
> +-----------------------------------------------------------
> ------------------------------------------------------------
> -----------------+--+
> |
> set                                                                   |
> +-----------------------------------------------------------
> ------------------------------------------------------------
> -----------------+--+
> | hive.conf.restricted.list=hive.security.authorization.
> enabled,hive.security.authorization.manager,hive.security.authenticator.manager
> |
> +-----------------------------------------------------------
> ------------------------------------------------------------
> -----------------+--+
> 1 row selected (0.006 seconds)
> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.command.whitelist;
> +-----------------------------------------------------------
> --------------------+--+
> |                                      set
> |
> +-----------------------------------------------------------
> --------------------+--+
> | hive.security.command.whitelist=set,reset,dfs,add,list,delete,reload,compile
> |
> +-----------------------------------------------------------
> --------------------+--+
> 1 row selected (0.008 seconds)
>
>
>
>
> 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key=xxxxxxxxxxxxxxx;
> Error: Error while processing statement: Cannot modify fs.s3a.access.key
> at runtime. It is not in list of params that are allowed to be modified at
> runtime (state=42000,code=1)
>
>
>
> On Mon, Dec 19, 2016 at 2:47 PM, Selvamohan Neethiraj <sneet...@apache.org>
> wrote:
>
> Hi,
>
>
>
> Which version of Hive and Ranger are you using ? Can you check if Ranger
> has added  hiveserver2 parameters  
> hive.conf.restricted.list,hive.security.command.whitelist
>  in the hive configuration file(s) ?
>
> Can you please list out these parameter values here ?
>
>
>
> Thanks,
>
> Selva-
>
>
>
> *From: *Anandha L Ranganathan <analog.s...@gmail.com>
> *Reply-To: *"user@ranger.incubator.apache.org" <
> user@ranger.incubator.apache.org>
> *Date: *Monday, December 19, 2016 at 5:30 PM
> *To: *"user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org
> >
> *Subject: *Unable to connect to S3 after enabling Ranger with Hive
>
>
>
> Hi,
>
>
> Unable to create table pointing to S3 after enabling Ranger.
>
> This is database we created before enabling Ranger.
>
> 1.  SET fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>
> 2.  SET fs.s3a.access.key=xxxxxxx;
>
> 3.  SET fs.s3a.secret.key=yyyyyyyyyyyyyyy;
>
> 4.
>
> 5.
>
> 6.  CREATE DATABASE IF NOT EXISTS backup_s3a1
>
> 7.  COMMENT "s3a schema test"
>
> 8.  LOCATION "s3a://gd-de-dp-db-hcat-backup-schema/";
>
> After Ranger was enabled, we try to create another database but it is
> throwing error.
>
> 1.  0: jdbc:hive2://usw2dxdpmn01.local:> SET 
> fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>
> 2.  Error: Error while processing statement: Cannot modify fs.s3a.impl at 
> runtime. It is not in list of params that are allowed to be modified at 
> runtime (state=42000,code=1)
>
> 3.
>
>
>
> I configured the credentials in the core-site.xml and always returns
> "undefined" when I am trying to see the values for  below commands. This is
> in our " dev" environment where Ranger is enabled. In   other environment
> where Ranger is not installed , we are not facing this problem.
>
> 1.  0: jdbc:hive2://usw2dxdpmn01:10010> set  fs.s3a.impl;
>
> 2.  +-----------------------------------------------------+--+
>
> 3.  |                         set                         |
>
> 4.  +-----------------------------------------------------+--+
>
> 5.  | fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem  |
>
> 6.  +-----------------------------------------------------+--+
>
> 7.  1 row selected (0.006 seconds)
>
> 8.  0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key;
>
> 9.  +---------------------------------+--+
>
> 10.|               set               |
>
> 11.+---------------------------------+--+
>
> 12.| fs.s3a.access.key is undefined  |
>
> 13.+---------------------------------+--+
>
> 14.1 row selected (0.005 seconds)
>
> 15.0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.secret.key;
>
> 16.+---------------------------------+--+
>
> 17.|               set               |
>
> 18.+---------------------------------+--+
>
> 19.| fs.s3a.secret.key is undefined  |
>
> 20.+---------------------------------+--+
>
> 21.1 row selected (0.005 seconds)
>
>
>
> Any help or pointers is appreciated.
>
>
>
>
>
>
>
>
>
>
>
>
>

Reply via email to