BTW, for that integration suite, I saw the related artifacts in the RC4 staging directory.
Does Spark 2.4.0 need to start to release these `spark-kubernetes-integration-tests` artifacts? - https://repository.apache.org/content/repositories/orgapachespark-1290/org/apache/spark/spark-kubernetes-integration-tests_2.11/ - https://repository.apache.org/content/repositories/orgapachespark-1290/org/apache/spark/spark-kubernetes-integration-tests_2.12/ Historically, Spark released `spark-docker-integration-tests` at Spark 1.6.x era and stopped since Spark 2.0.0. - http://central.maven.org/maven2/org/apache/spark/spark-docker-integration-tests_2.10/ - http://central.maven.org/maven2/org/apache/spark/spark-docker-integration-tests_2.11/ Bests, Dongjoon. On Tue, Oct 23, 2018 at 11:43 AM Stavros Kontopoulos < stavros.kontopou...@lightbend.com> wrote: > Sean, > > Ok makes sense, im using a cloned repo. I built with Scala 2.12 profile > using the related tag v2.4.0-rc4: > > ./dev/change-scala-version.sh 2.12 > ./dev/make-distribution.sh --name test --r --tgz -Pscala-2.12 -Psparkr > -Phadoop-2.7 -Pkubernetes -Phive > Pushed images to dockerhub (previous email) since I didnt use the minikube > daemon (default behavior). > > Then run tests successfully against minikube: > > TGZ_PATH=$(pwd)/spark-2.4.0-bin-test.gz > cd resource-managers/kubernetes/integration-tests > > ./dev/dev-run-integration-tests.sh --spark-tgz $TGZ_PATH --service-account > default --namespace default --image-tag k8s-scala-12 --image-repo skonto > > > [INFO] > [INFO] --- scalatest-maven-plugin:1.0:test (integration-test) @ > spark-kubernetes-integration-tests_2.12 --- > Discovery starting. > Discovery completed in 229 milliseconds. > Run starting. Expected test count is: 14 > KubernetesSuite: > - Run SparkPi with no resources > - Run SparkPi with a very long application name. > - Use SparkLauncher.NO_RESOURCE > - Run SparkPi with a master URL without a scheme. > - Run SparkPi with an argument. > - Run SparkPi with custom labels, annotations, and environment variables. > - Run extraJVMOptions check on driver > - Run SparkRemoteFileTest using a remote data file > - Run SparkPi with env and mount secrets. > - Run PySpark on simple pi.py example > - Run PySpark with Python2 to test a pyfiles example > - Run PySpark with Python3 to test a pyfiles example > - Run PySpark with memory customization > - Run in client mode. > Run completed in 5 minutes, 24 seconds. > Total number of tests run: 14 > Suites: completed 2, aborted 0 > Tests: succeeded 14, failed 0, canceled 0, ignored 0, pending 0 > All tests passed. > [INFO] > ------------------------------------------------------------------------ > [INFO] Reactor Summary: > [INFO] > [INFO] Spark Project Parent POM 2.4.0 ..................... SUCCESS [ > 4.491 s] > [INFO] Spark Project Tags ................................. SUCCESS [ > 3.833 s] > [INFO] Spark Project Local DB ............................. SUCCESS [ > 2.680 s] > [INFO] Spark Project Networking ........................... SUCCESS [ > 4.817 s] > [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [ > 2.541 s] > [INFO] Spark Project Unsafe ............................... SUCCESS [ > 2.795 s] > [INFO] Spark Project Launcher ............................. SUCCESS [ > 5.593 s] > [INFO] Spark Project Core ................................. SUCCESS [ > 25.160 s] > [INFO] Spark Project Kubernetes Integration Tests 2.4.0 ... SUCCESS [05:30 > min] > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD SUCCESS > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 06:23 min > [INFO] Finished at: 2018-10-23T18:39:11Z > [INFO] > ------------------------------------------------------------------------ > > > but had to modify this line > <https://github.com/apache/spark/blob/master/resource-managers/kubernetes/integration-tests/dev/dev-run-integration-tests.sh#L106> > and > added -Pscala-2.12 , otherwise it fails (these tests inherit from the > parent pom but the profile is not propagated to the mvn command that > launches the tests, I can create a PR to fix that). > > > On Tue, Oct 23, 2018 at 7:44 PM, Hyukjin Kwon <gurwls...@gmail.com> wrote: > >> https://github.com/apache/spark/pull/22514 sounds like a regression that >> affects Hive CTAS in write path (by not replacing them into Spark internal >> datasources; therefore performance regression). >> but yea I suspect if we should block the release by this. >> >> https://github.com/apache/spark/pull/22144 is just being discussed if I >> am not mistaken. >> >> Thanks. >> >> 2018년 10월 24일 (수) 오전 12:27, Xiao Li <gatorsm...@gmail.com>님이 작성: >> >>> https://github.com/apache/spark/pull/22144 is also not a blocker of >>> Spark 2.4 release, as discussed in the PR. >>> >>> Thanks, >>> >>> Xiao >>> >>> Xiao Li <gatorsm...@gmail.com> 于2018年10月23日周二 上午9:20写道: >>> >>>> Thanks for reporting this. https://github.com/apache/spark/pull/22514 >>>> is not a blocker. We can fix it in the next minor release, if we are unable >>>> to make it in this release. >>>> >>>> Thanks, >>>> >>>> Xiao >>>> >>>> Sean Owen <sro...@gmail.com> 于2018年10月23日周二 上午9:14写道: >>>> >>>>> (I should add, I only observed this with the Scala 2.12 build. It all >>>>> seemed to work with 2.11. Therefore I'm not too worried about it. I >>>>> don't think it's a Scala version issue, but perhaps something looking >>>>> for a spark 2.11 tarball and not finding it. See >>>>> https://github.com/apache/spark/pull/22805#issuecomment-432304622 for >>>>> a change that might address this kind of thing.) >>>>> >>>>> On Tue, Oct 23, 2018 at 11:05 AM Sean Owen <sro...@gmail.com> wrote: >>>>> > >>>>> > Yeah, that's maybe the issue here. This is a source release, not a >>>>> git checkout, and it still needs to work in this context. >>>>> > >>>>> > I just added -Pkubernetes to my build and didn't do anything else. I >>>>> think the ideal is that a "mvn -P... -P... install" to work from a source >>>>> release; that's a good expectation and consistent with docs. >>>>> > >>>>> > Maybe these tests simply don't need to run with the normal suite of >>>>> tests, and can be considered tests run manually by developers running >>>>> these >>>>> scripts? Basically, KubernetesSuite shouldn't run in a normal mvn install? >>>>> > >>>>> > I don't think this has to block the release even if so, just trying >>>>> to get to the bottom of it. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>> >>>>> > > > -- > Stavros Kontopoulos > > *Senior Software Engineer* > *Lightbend, Inc.* > > *p: +30 6977967274 <%2B1%20650%20678%200020>* > *e: stavros.kontopou...@lightbend.com* <dave.mar...@lightbend.com> > > >