oh. i missed Gopal's reply. oy... that sounds foreboding. I'll keep you posted on my progress.
On Tue, Sep 26, 2017 at 4:40 PM, Gopal Vijayaraghavan <gop...@apache.org> wrote: > Hi, > > > org.apache.hadoop.hive.ql.parse.SemanticException: Failed to get a > spark session: org.apache.hadoop.hive.ql.metadata.HiveException: Failed > to create spark client. > > I get inexplicable errors with Hive-on-Spark unless I do a three step > build. > > Build Hive first, use that version to build Spark, use that Spark version > to rebuild Hive. > > I have to do this to make it work because Spark contains Hive jars and > Hive contains Spark jars in the class-path. > > And specifically I have to edit the pom.xml files, instead of passing in > params with -Dspark.version, because the installed pom files don't get > replacements from the build args. > > Cheers, > Gopal > > >