Cool. After adding  the configuration it is working fine.

0: jdbc:hive2://usw2dxdpmn01:10010> set
hive.security.authorization.sqlstd.confwhitelist.append;
+------------------------------------------------------------------------------------+--+
|
set                                         |
+------------------------------------------------------------------------------------+--+
|
hive.security.authorization.sqlstd.confwhitelist.append=|fs\.s3a\..*|fs\.s3n\..*
|  |
+------------------------------------------------------------------------------------+--+


Thanks Selva for the quick help.



On Mon, Dec 19, 2016 at 5:29 PM, Selvamohan Neethiraj <sneet...@apache.org>
wrote:

> Hi,
>
> Can you try appending the following string to the  existing value of
>  hive.security.authorization.sqlstd.confwhitelist
>
> |fs\.s3a\..*
>
> And restart the HiveServer2 to see if this fixes this issue ?
>
> Thanks,
> Selva-
> From: Anandha L Ranganathan <analog.s...@gmail.com>
> Reply-To: "user@ranger.incubator.apache.org" <
> user@ranger.incubator.apache.org>
> Date: Monday, December 19, 2016 at 6:27 PM
>
> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org>
> Subject: Re: Unable to connect to S3 after enabling Ranger with Hive
>
> Selva,
>
> Please find the results.
>
> set hive.security.authorization.sqlstd.confwhitelist;
>
> | hive.security.authorization.sqlstd.confwhitelist=hive\.
> auto\..*|hive\.cbo\..*|hive\.convert\..*|hive\.exec\.
> dynamic\.partition.*|hive\.exec\..*\.dynamic\.partitions\
> ..*|hive\.exec\.compress\..*|hive\.exec\.infer\..*|hive\.
> exec\.mode.local\..*|hive\.exec\.orc\..*|hive\.exec\.
> parallel.*|hive\.explain\..*|hive\.fetch.task\..*|hive\.
> groupby\..*|hive\.hbase\..*|hive\.index\..*|hive\.index\..
> *|hive\.intermediate\..*|hive\.join\..*|hive\.limit\..*|
> hive\.log\..*|hive\.mapjoin\..*|hive\.merge\..*|hive\.
> optimize\..*|hive\.orc\..*|hive\.outerjoin\..*|hive\.
> parquet\..*|hive\.ppd\..*|hive\.prewarm\..*|hive\.
> server2\.proxy\.user|hive\.skewjoin\..*|hive\.smbjoin\..*
> |hive\.stats\..*|hive\.tez\..*|hive\.vectorized\..*|mapred\.
> map\..*|mapred\.reduce\..*|mapred\.output\.compression\.
> codec|mapred\.job\.queuename|mapred\.output\.compression\.
> type|mapred\.min\.split\.size|mapreduce\.job\.reduce\.
> slowstart\.completedmaps|mapreduce\.job\.queuename|mapreduce\.job\.tags|
> mapreduce\.input\.fileinputformat\.split\.minsize|mapreduce\.map\..*|
> mapreduce\.reduce\..*|mapreduce\.output\.fileoutputformat\.compress\.
> codec|mapreduce\.output\.fileoutputformat\.compress\.
> type|tez\.am\..*|tez\.task\..*|tez\.runtime\..*|tez.queue.name
> |hive\.exec\.reducers\.bytes\.per\.reducer|hive\.
> client\.stats\.counters|hive\.exec\.default\.partition\.
> name|hive\.exec\.drop\.ignorenonexistent|hive\.
> counters\.group\.name|hive\.default\.fileformat\.managed|
> hive\.enforce\.bucketing|hive\.enforce\.bucketmapjoin|hive\.
> enforce\.sorting|hive\.enforce\.sortmergebucketmapjoin|hive\.
> cache\.expr\.evaluation|hive\.hashtable\.loadfactor|hive\.
> hashtable\.initialCapacity|hive\.ignore\.mapjoin\.hint|
> hive\.limit\.row\.max\.size|hive\.mapred\.mode|hive\.map\.
> aggr|hive\.compute\.query\.using\.stats|hive\.exec\.
> rowoffset|hive\.variable\.substitute|hive\.variable\.
> substitute\.depth|hive\.autogen\.columnalias\.prefix\.
> includefuncname|hive\.autogen\.columnalias\.prefix\.label|
> hive\.exec\.check\.crossproducts|hive\.compat|hive\.exec\.concatenate\.
> check\.index|hive\.display\.partition\.cols\.separately|
> hive\.error\.on\.empty\.partition|hive\.execution\.
> engine|hive\.exim\.uri\.scheme\.whitelist|hive\.file\.
> max\.footer|hive\.mapred\.supports\.subdirectories|hive\
> .insert\.into\.multilevel\.dirs|hive\.localize\.resource\
> .num\.wait\.attempts|hive\.multi\.insert\.move\.tasks\.
> share\.dependencies|hive\.support\.quoted\.identifiers|
> hive\.resultset\.use\.unique\.column\.names|hive\.analyze\.
> stmt\.collect\.partlevel\.stats|hive\.server2\.logging\.
> operation\.level|hive\.support\.sql11\.reserved\.
> keywords|hive\.exec\.job\.debug\.capture\.stacktraces|
> hive\.exec\.job\.debug\.timeout|hive\.exec\.max\.
> created\.files|hive\.exec\.reducers\.max|hive\.reorder\.
> nway\.joins|hive\.output\.file\.extension|hive\.exec\.
> show\.job\.failure\.debug\.info|hive\.exec\.tasklog\.debug\.timeout  |
>
>
>
> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.authorization.
> sqlstd.confwhitelist.append;
> +-----------------------------------------------------------
> ------------+--+
> |                                  set                                  |
> +-----------------------------------------------------------
> ------------+--+
> | hive.security.authorization.sqlstd.confwhitelist.append is undefined  |
> +-----------------------------------------------------------
> ------------+--+
>
>
> On Mon, Dec 19, 2016 at 3:12 PM, Selvamohan Neethiraj <sneet...@apache.org
> > wrote:
>
>> Hi,
>>
>> Can you also post here the value for the following two parameters:
>>
>> hive.security.authorization.sqlstd.confwhitelist
>>
>> hive.security.authorization.sqlstd.confwhitelist.append
>>
>>
>>
>> Thanks,
>>
>> Selva-
>>
>> From: Anandha L Ranganathan <analog.s...@gmail.com>
>> Reply-To: "user@ranger.incubator.apache.org" <
>> user@ranger.incubator.apache.org>
>> Date: Monday, December 19, 2016 at 5:54 PM
>> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org>
>> Subject: Re: Unable to connect to S3 after enabling Ranger with Hive
>>
>> Selva,
>>
>> We are using HDP and here are versions and results.
>>
>> Hive :  1.2.1.2.4
>> Ranger: 0.5.0.2.4
>>
>>
>>
>> 0: jdbc:hive2://usw2dxdpmn01:10010> set  hive.conf.restricted.list;
>> +-----------------------------------------------------------
>> ------------------------------------------------------------
>> -----------------+--+
>> |
>> set                                                                   |
>> +-----------------------------------------------------------
>> ------------------------------------------------------------
>> -----------------+--+
>> | hive.conf.restricted.list=hive.security.authorization.enable
>> d,hive.security.authorization.manager,hive.security.authenticator.manager
>> |
>> +-----------------------------------------------------------
>> ------------------------------------------------------------
>> -----------------+--+
>> 1 row selected (0.006 seconds)
>> 0: jdbc:hive2://usw2dxdpmn01:10010> set hive.security.command.whitelist;
>> +-----------------------------------------------------------
>> --------------------+--+
>> |                                      set
>> |
>> +-----------------------------------------------------------
>> --------------------+--+
>> | 
>> hive.security.command.whitelist=set,reset,dfs,add,list,delete,reload,compile
>> |
>> +-----------------------------------------------------------
>> --------------------+--+
>> 1 row selected (0.008 seconds)
>>
>>
>>
>>
>> 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key=xxxxxxxxxxxx
>> xxx;
>> Error: Error while processing statement: Cannot modify fs.s3a.access.key
>> at runtime. It is not in list of params that are allowed to be modified at
>> runtime (state=42000,code=1)
>>
>> On Mon, Dec 19, 2016 at 2:47 PM, Selvamohan Neethiraj <
>> sneet...@apache.org> wrote:
>>
>>> Hi,
>>>
>>> Which version of Hive and Ranger are you using ? Can you check if Ranger
>>> has added  hiveserver2 parameters  
>>> hive.conf.restricted.list,hive.security.command.whitelist
>>>  in the hive configuration file(s) ?
>>> Can you please list out these parameter values here ?
>>>
>>> Thanks,
>>> Selva-
>>>
>>> From: Anandha L Ranganathan <analog.s...@gmail.com>
>>> Reply-To: "user@ranger.incubator.apache.org" <
>>> user@ranger.incubator.apache.org>
>>> Date: Monday, December 19, 2016 at 5:30 PM
>>> To: "user@ranger.incubator.apache.org" <user@ranger.incubator.apache.org
>>> >
>>> Subject: Unable to connect to S3 after enabling Ranger with Hive
>>>
>>> Hi,
>>>
>>>
>>> Unable to create table pointing to S3 after enabling Ranger.
>>>
>>> This is database we created before enabling Ranger.
>>>
>>>
>>>    1. SET fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>>>    2. SET fs.s3a.access.key=xxxxxxx;
>>>    3. SET fs.s3a.secret.key=yyyyyyyyyyyyyyy;
>>>    4.
>>>    5.
>>>    6. CREATE DATABASE IF NOT EXISTS backup_s3a1
>>>    7. COMMENT "s3a schema test"
>>>    8. LOCATION "s3a://gd-de-dp-db-hcat-backup-schema/";
>>>
>>> After Ranger was enabled, we try to create another database but it is
>>> throwing error.
>>>
>>>
>>>    1. 0: jdbc:hive2://usw2dxdpmn01.local:> SET 
>>> fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem;
>>>    2. Error: Error while processing statement: Cannot modify fs.s3a.impl at 
>>> runtime. It is not in list of params that are allowed to be modified at 
>>> runtime (state=42000,code=1)
>>>    3.
>>>
>>>
>>>
>>> I configured the credentials in the core-site.xml and always returns
>>> "undefined" when I am trying to see the values for  below commands. This is
>>> in our " dev" environment where Ranger is enabled. In   other environment
>>> where Ranger is not installed , we are not facing this problem.
>>>
>>>
>>>    1. 0: jdbc:hive2://usw2dxdpmn01:10010> set  fs.s3a.impl;
>>>    2. +-----------------------------------------------------+--+
>>>    3. |                         set                         |
>>>    4. +-----------------------------------------------------+--+
>>>    5. | fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem  |
>>>    6. +-----------------------------------------------------+--+
>>>    7. 1 row selected (0.006 seconds)
>>>    8. 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.access.key;
>>>    9. +---------------------------------+--+
>>>    10. |               set               |
>>>    11. +---------------------------------+--+
>>>    12. | fs.s3a.access.key is undefined  |
>>>    13. +---------------------------------+--+
>>>    14. 1 row selected (0.005 seconds)
>>>    15. 0: jdbc:hive2://usw2dxdpmn01:10010> set fs.s3a.secret.key;
>>>    16. +---------------------------------+--+
>>>    17. |               set               |
>>>    18. +---------------------------------+--+
>>>    19. | fs.s3a.secret.key is undefined  |
>>>    20. +---------------------------------+--+
>>>    21. 1 row selected (0.005 seconds)
>>>
>>>
>>> Any help or pointers is appreciated.
>>>
>>>
>>
>

Reply via email to