Sean, Ok makes sense, im using a cloned repo. I built with Scala 2.12 profile using the related tag v2.4.0-rc4:
./dev/change-scala-version.sh 2.12 ./dev/make-distribution.sh --name test --r --tgz -Pscala-2.12 -Psparkr -Phadoop-2.7 -Pkubernetes -Phive Pushed images to dockerhub (previous email) since I didnt use the minikube daemon (default behavior). Then run tests successfully against minikube: TGZ_PATH=$(pwd)/spark-2.4.0-bin-test.gz cd resource-managers/kubernetes/integration-tests ./dev/dev-run-integration-tests.sh --spark-tgz $TGZ_PATH --service-account default --namespace default --image-tag k8s-scala-12 --image-repo skonto [INFO] [INFO] --- scalatest-maven-plugin:1.0:test (integration-test) @ spark-kubernetes-integration-tests_2.12 --- Discovery starting. Discovery completed in 229 milliseconds. Run starting. Expected test count is: 14 KubernetesSuite: - Run SparkPi with no resources - Run SparkPi with a very long application name. - Use SparkLauncher.NO_RESOURCE - Run SparkPi with a master URL without a scheme. - Run SparkPi with an argument. - Run SparkPi with custom labels, annotations, and environment variables. - Run extraJVMOptions check on driver - Run SparkRemoteFileTest using a remote data file - Run SparkPi with env and mount secrets. - Run PySpark on simple pi.py example - Run PySpark with Python2 to test a pyfiles example - Run PySpark with Python3 to test a pyfiles example - Run PySpark with memory customization - Run in client mode. Run completed in 5 minutes, 24 seconds. Total number of tests run: 14 Suites: completed 2, aborted 0 Tests: succeeded 14, failed 0, canceled 0, ignored 0, pending 0 All tests passed. [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Spark Project Parent POM 2.4.0 ..................... SUCCESS [ 4.491 s] [INFO] Spark Project Tags ................................. SUCCESS [ 3.833 s] [INFO] Spark Project Local DB ............................. SUCCESS [ 2.680 s] [INFO] Spark Project Networking ........................... SUCCESS [ 4.817 s] [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [ 2.541 s] [INFO] Spark Project Unsafe ............................... SUCCESS [ 2.795 s] [INFO] Spark Project Launcher ............................. SUCCESS [ 5.593 s] [INFO] Spark Project Core ................................. SUCCESS [ 25.160 s] [INFO] Spark Project Kubernetes Integration Tests 2.4.0 ... SUCCESS [05:30 min] [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 06:23 min [INFO] Finished at: 2018-10-23T18:39:11Z [INFO] ------------------------------------------------------------------------ but had to modify this line <https://github.com/apache/spark/blob/master/resource-managers/kubernetes/integration-tests/dev/dev-run-integration-tests.sh#L106> and added -Pscala-2.12 , otherwise it fails (these tests inherit from the parent pom but the profile is not propagated to the mvn command that launches the tests, I can create a PR to fix that). On Tue, Oct 23, 2018 at 7:44 PM, Hyukjin Kwon <gurwls...@gmail.com> wrote: > https://github.com/apache/spark/pull/22514 sounds like a regression that > affects Hive CTAS in write path (by not replacing them into Spark internal > datasources; therefore performance regression). > but yea I suspect if we should block the release by this. > > https://github.com/apache/spark/pull/22144 is just being discussed if I > am not mistaken. > > Thanks. > > 2018년 10월 24일 (수) 오전 12:27, Xiao Li <gatorsm...@gmail.com>님이 작성: > >> https://github.com/apache/spark/pull/22144 is also not a blocker of >> Spark 2.4 release, as discussed in the PR. >> >> Thanks, >> >> Xiao >> >> Xiao Li <gatorsm...@gmail.com> 于2018年10月23日周二 上午9:20写道: >> >>> Thanks for reporting this. https://github.com/apache/spark/pull/22514 >>> is not a blocker. We can fix it in the next minor release, if we are unable >>> to make it in this release. >>> >>> Thanks, >>> >>> Xiao >>> >>> Sean Owen <sro...@gmail.com> 于2018年10月23日周二 上午9:14写道: >>> >>>> (I should add, I only observed this with the Scala 2.12 build. It all >>>> seemed to work with 2.11. Therefore I'm not too worried about it. I >>>> don't think it's a Scala version issue, but perhaps something looking >>>> for a spark 2.11 tarball and not finding it. See >>>> https://github.com/apache/spark/pull/22805#issuecomment-432304622 for >>>> a change that might address this kind of thing.) >>>> >>>> On Tue, Oct 23, 2018 at 11:05 AM Sean Owen <sro...@gmail.com> wrote: >>>> > >>>> > Yeah, that's maybe the issue here. This is a source release, not a >>>> git checkout, and it still needs to work in this context. >>>> > >>>> > I just added -Pkubernetes to my build and didn't do anything else. I >>>> think the ideal is that a "mvn -P... -P... install" to work from a source >>>> release; that's a good expectation and consistent with docs. >>>> > >>>> > Maybe these tests simply don't need to run with the normal suite of >>>> tests, and can be considered tests run manually by developers running these >>>> scripts? Basically, KubernetesSuite shouldn't run in a normal mvn install? >>>> > >>>> > I don't think this has to block the release even if so, just trying >>>> to get to the bottom of it. >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>> >>>> -- Stavros Kontopoulos *Senior Software Engineer* *Lightbend, Inc.* *p: +30 6977967274 <%2B1%20650%20678%200020>* *e: stavros.kontopou...@lightbend.com* <dave.mar...@lightbend.com>