> On 30 Mar 2016, at 21:02, Sean Busbey <bus...@cloudera.com> wrote:
> 
> On Wed, Mar 30, 2016 at 4:33 AM, Steve Loughran <ste...@hortonworks.com> 
> wrote:
>> 
>>> On 29 Mar 2016, at 22:19, Michael Segel <msegel_had...@hotmail.com> wrote:
>>> 
>>> Hi,
>>> 
>>> So yeah, I know that Spark jobs running on a Hadoop cluster will inherit 
>>> its security from the underlying YARN job.
>>> However… that’s not really saying much when you think about some use cases.
>>> 
>>> Like using the thrift service …
>>> 
>>> I’m wondering what else is new and what people have been thinking about how 
>>> to enhance spark’s security.
>>> 
>> 
>> Been thinking a bit.
>> 
>> One thing to look at is renewal of hbase and hive tokens on long-lived 
>> services, alongside hdfs
>> 
>> 
> 
> I've been looking at this as well. The current work-around I'm using
> is to use keytab logins on the executors, which is less than
> desirable.


OK, let's work together on this ... the current spark renewal code assumes its 
only for HDFS (indeed, that the filesystem is HDFS and therefore the #of tokens 
> 0); there' s no fundamental reason why the code in YarnSparkHadoopUtils can't 
run in the AM too.

> 
> Since the HBase project maintains Spark integration points, it'd be
> great if there were just a hook for services to provide "here's how to
> renew" to a common renewal service.
> 

1. Wittenauer is doing some work on a tool for doing this; I'm pushing for it 
to be a fairly generic API. Even if Spark has to use reflection to get at it, 
at least it would be consistent across services. See 
https://issues.apache.org/jira/browse/HADOOP-12563

2. The topic of HTTPS based acquisition/use of HDFS tokens has arisen 
elsewhere; needed for long-haul job submission when  you don' t have a keytab 
to hand. This could be useful as it'd avoid actually needing hbase-*.jar on the 
classpath at submit time.


> 
> 
> -- 
> busbey
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
> 
> 

Reply via email to