Thanks for reporting back. I was pretty confused trying to reproduce the error :)
On Thu, Jul 24, 2014 at 1:09 PM, Stephen Boesch <java...@gmail.com> wrote: > OK I did find my error. The missing step: > > mvn install > > I should have republished (mvn install) all of the other modules . > > The mvn -pl will rely on the modules locally published and so the latest > code that I had git pull'ed was not being used (except the sql/core module > code). > > The tests are passing after having properly performed the mvn install > before running with the mvn -pl sql/core. > > > > > 2014-07-24 12:04 GMT-07:00 Stephen Boesch <java...@gmail.com>: > > > > > Are other developers seeing the following error for the recently added > > substr() method? If not, any ideas why the following invocation of tests > > would be failing for me - i.e. how the given invocation would need to be > > tweaked? > > > > mvn -Pyarn -Pcdh5 test -pl sql/core > > -DwildcardSuites=org.apache.spark.sql.SQLQuerySuite > > > > (note cdh5 is a custom profile for cdh5.0.0 but should not be affecting > > these results) > > > > Only the test("SPARK-2407 Added Parser of SQL SUBSTR()") fails: all of > the > > other 33 tests pass. > > > > SQLQuerySuite: > > - SPARK-2041 column name equals tablename > > - SPARK-2407 Added Parser of SQL SUBSTR() *** FAILED *** > > Exception thrown while executing query: > > == Logical Plan == > > java.lang.UnsupportedOperationException > > == Optimized Logical Plan == > > java.lang.UnsupportedOperationException > > == Physical Plan == > > java.lang.UnsupportedOperationException > > == Exception == > > java.lang.UnsupportedOperationException > > java.lang.UnsupportedOperationException > > at > > > org.apache.spark.sql.catalyst.analysis.EmptyFunctionRegistry$.lookupFunction(FunctionRegistry.scala:33) > > at > > > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5$$anonfun$applyOrElse$3.applyOrElse(Analyzer.scala:131) > > at > > > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveFunctions$$anonfun$apply$5$$anonfun$applyOrElse$3.applyOrElse(Analyzer.scala:129) > > at > > > org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:165) > > at > > > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:183) > > at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) > > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) > > at > > scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) > > at > > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) > > at > > scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) > > at scala.collection.TraversableOnce$class.to > (TraversableOnce.scala:273) > > at scala.collection.AbstractIterator.to(Iterator.scala:1157) > > at > > > scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) > > at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) > > at > > scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) > > at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) > > at > > > org.apache.spark.sql.catalyst.trees.TreeNode.transformChildrenDown(TreeNode.scala:212) > > at > > > org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:168) > > at org.apache.spark.sql.catalyst.plans.QueryPlan.org > > > $apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1(QueryPlan.scala:52) > > at > > > org.apache.spark.sql.catalyst.plans.QueryPlan$$anonfun$1$$anonfun$apply$1.apply(QueryPlan.scala:66) > > at > > > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > > at > > > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > > at scala.collection.immutable.List.foreach(List.scala:318) > > at > scala.collection.TraversableLike$class.map(TraversableLike.scala:244) > > at scala.collection.AbstractTraversable.map(Traversable.scala:105) > > at > > >