I see, this must be because of hadoop version you are selecting by using "-Dhadoop.version=2.6.0-cdh5.13.1". Spark 3.1.1 only support hadoop-2.7 and hadoop-3.2, at least these two can be given via profiles: -Phadoop-2.7 and -Phadoop-3.2 (the default).
On Wed, Mar 10, 2021 at 12:26 PM jiahong li <monkeyboy....@gmail.com> wrote: > i use ./build/mvn to compile ,and after execute command > :./build/zinc-0.3.15/bin/zinc > -shutdown > and execute command like this: /dev/make-distribution.sh --name > custom-spark --pip --tgz -Phive -Phive-thriftserver -Pyarn > -Dhadoop.version=2.6.0-cdh5.13.1 -DskipTests > same error appear. > and execute command: ps -ef |grep zinc, there is nothing containe zinc > > Attila Zsolt Piros <piros.attila.zs...@gmail.com> 于2021年3月10日周三 下午6:55写道: > >> hi! >> >> Are you compiling Spark itself? >> Do you use "./build/mvn" from the project root? >> If you compiled an other version of Spark before and there the scala >> version was different then zinc/nailgun could cached the old classes which >> can cause similar troubles. >> In that case this could help: >> >> ./build/zinc-0.3.15/bin/zinc -shutdown >> >> Best Regards, >> Attila >> >> On Wed, Mar 10, 2021 at 11:27 AM jiahong li <monkeyboy....@gmail.com> >> wrote: >> >>> hi, everybody, when i compile spark 3.1.1 from tag v3.1.1 ,encounter >>> error like this: >>> >>> INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @ >>> spark-core_2.12 --- >>> [INFO] Using incremental compilation using Mixed compile order >>> [INFO] Compiler bridge file: >>> .sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar >>> [INFO] compiler plugin: >>> BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null) >>> [INFO] Compiling 560 Scala sources and 99 Java sources to >>> git/spark/core/target/scala-2.12/classes ... >>> [ERROR] [Error] >>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: >>> type mismatch; >>> found : K where type K >>> required: String >>> [ERROR] [Error] >>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: >>> value map is not a member of V >>> [ERROR] [Error] >>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: >>> missing argument list for method stripXSS in class XssSafeRequest >>> Unapplied methods are only converted to functions when a function type >>> is expected. >>> You can make this conversion explicit by writing `stripXSS _` or >>> `stripXSS(_)` instead of `stripXSS`. >>> [ERROR] [Error] >>> git/spark/core/src/main/scala/org/apache/spark/ui/PagedTable.scala:307: >>> value startsWith is not a member of K >>> [ERROR] [Error] >>> git/spark/core/src/main/scala/org/apache/spark/util/Utils.scala:580: value >>> toLowerCase is not a member of object org.apache.hadoop.util.StringUtils >>> [ERROR] 5 errors found >>> >>> anybody encounter error like this? >>> >>> >>