You will need to compile spark with SPARK_HIVE=true.

On Wed, Jun 11, 2014 at 5:37 PM, Stephen Boesch <java...@gmail.com> wrote:

> Hi,
>   The documentation of Catalyst describes using HiveContext; however, the
> scala classes do not exist in Master or 1.0.0 Branch.  What is the
> replacement/equivalent in Master?
>
> Package does not exist:
> org.apache.spark.sql.hive
>
> Here is code from SQL on Spark meetup slides referencing that
> package/classes:
>
> val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
> import hiveContext._
>
> hql("CREATE TABLE IF NOT EXISTS src (key INT, value STRING)")
>
>
>
>

Reply via email to