[
https://issues.apache.org/jira/browse/SPARK-19526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Mark Grover updated SPARK-19526:
--------------------------------
Summary: Spark should raise an exception when it tries to read a Hive view
but it doesn't have read access on the corresponding table(s) (was: Spark
should rise an exception when it tries to read a Hive view but it doesn't have
read access on the corresponding table(s))
> Spark should raise an exception when it tries to read a Hive view but it
> doesn't have read access on the corresponding table(s)
> -------------------------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-19526
> URL: https://issues.apache.org/jira/browse/SPARK-19526
> Project: Spark
> Issue Type: Improvement
> Components: SQL
> Affects Versions: 1.6.4, 2.0.3, 2.2.0, 2.3.0
> Reporter: Reza Safi
>
> Spark sees a Hive views as a set of hdfs "files". So to read anything from a
> Hive view, Spark needs access to all of the files that belongs to the
> table(s) that the view queries them. In other words a Spark user cannot be
> granted fine grained permissions at the levels of Hive columns or records.
> Consider that there is a Spark job that contains a SQL query that tries to
> read a Hive view. Currently the Spark job will finish successfully if the
> user that runs the Spark job doesn't have proper read access permissions to
> the tables that the Hive view has been built on top of them. It will just
> return an empty result set. This can be confusing for the users, since the
> job will be finishes without any exception or error.
> Spark should raise an exception like AccessDenied when it tries to run a
> Hive view query and its user doesn't have proper permissions to the tables
> that the Hive view is created on top of them.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]