Hi,

My understanding is that the most common (perhaps the only?) way to let
users run Hive queries on datasets stored in HDFS, is to configure Hive as
a proxy user in the namenodes config.

I'm wondering if, instead of using proxy user privileges, a Hive client
could be configured to first collect HDFS delegation tokens for the user
and then pass those tokens to the Hive server. That way, the Hive server
would use the tokens to authenticate with HDFS on behalf of the user.

Spark offers something similar to that with the
spark.yarn.access.hadoopFileSystems
<https://spark.apache.org/docs/latest/running-on-yarn.html#kerberos>
property. By chance, is there a way to achieve the same thing for Hive when
using a client like Beeline?

Thank you,

Julien

Reply via email to