And the code is right here: https://github.com/apache/spark/blob/master/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala
On Wed, Jun 11, 2014 at 5:38 PM, Michael Armbrust <mich...@databricks.com> wrote: > You will need to compile spark with SPARK_HIVE=true. > > > On Wed, Jun 11, 2014 at 5:37 PM, Stephen Boesch <java...@gmail.com> wrote: > >> Hi, >> The documentation of Catalyst describes using HiveContext; however, the >> scala classes do not exist in Master or 1.0.0 Branch. What is the >> replacement/equivalent in Master? >> >> Package does not exist: >> org.apache.spark.sql.hive >> >> Here is code from SQL on Spark meetup slides referencing that >> package/classes: >> >> val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc) >> import hiveContext._ >> >> hql("CREATE TABLE IF NOT EXISTS src (key INT, value STRING)") >> >> >> >> >