Yeah, I'd just add a spark-util that has these things.

Matei

On Jul 14, 2014, at 1:04 PM, Michael Armbrust <mich...@databricks.com> wrote:

> Yeah, sadly this dependency was introduced when someone consolidated the 
> logging infrastructure.  However, the dependency should be very small and 
> thus easy to remove, and I would like catalyst to be usable outside of Spark. 
>  A pull request to make this possible would be welcome.
> 
> Ideally, we'd create some sort of spark common package that has things like 
> logging.  That way catalyst could depend on that, without pulling in all of 
> Hadoop, etc.  Maybe others have opinions though, so I'm cc-ing the dev list.
> 
> 
> On Mon, Jul 14, 2014 at 12:21 AM, Yanbo Liang <yanboha...@gmail.com> wrote:
> Make Catalyst independent of Spark is the goal of Catalyst, maybe need time 
> and evolution.
> I awared that package org.apache.spark.sql.catalyst.util embraced 
> org.apache.spark.util.{Utils => SparkUtils},
> so that Catalyst has a dependency on Spark core. 
> I'm not sure whether it will be replaced by other component independent of 
> Spark in later release.
> 
> 
> 2014-07-14 11:51 GMT+08:00 Aniket Bhatnagar <aniket.bhatna...@gmail.com>:
> 
> As per the recent presentation given in Scala days 
> (http://people.apache.org/~marmbrus/talks/SparkSQLScalaDays2014.pdf), it was 
> mentioned that Catalyst is independent of Spark. But on inspecting pom.xml of 
> sql/catalyst module, it seems it has a dependency on Spark Core. Any 
> particular reason for the dependency? I would love to use Catalyst outside 
> Spark
> 
> (reposted as previous email bounced. Sorry if this is a duplicate).
> 
> 

Reply via email to