Hi, My understanding is that the most common (perhaps the only?) way to let users run Hive queries on datasets stored in HDFS, is to configure Hive as a proxy user in the namenodes config.
I'm wondering if, instead of using proxy user privileges, a Hive client could be configured to first collect HDFS delegation tokens for the user and then pass those tokens to the Hive server. That way, the Hive server would use the tokens to authenticate with HDFS on behalf of the user. Spark offers something similar to that with the spark.yarn.access.hadoopFileSystems <https://spark.apache.org/docs/latest/running-on-yarn.html#kerberos> property. By chance, is there a way to achieve the same thing for Hive when using a client like Beeline? Thank you, Julien