You will need to compile spark with SPARK_HIVE=true.
On Wed, Jun 11, 2014 at 5:37 PM, Stephen Boesch <java...@gmail.com> wrote: > Hi, > The documentation of Catalyst describes using HiveContext; however, the > scala classes do not exist in Master or 1.0.0 Branch. What is the > replacement/equivalent in Master? > > Package does not exist: > org.apache.spark.sql.hive > > Here is code from SQL on Spark meetup slides referencing that > package/classes: > > val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc) > import hiveContext._ > > hql("CREATE TABLE IF NOT EXISTS src (key INT, value STRING)") > > > >