I see, this must be because of hadoop version you are selecting by using
"-Dhadoop.version=2.6.0-cdh5.13.1".
Spark 3.1.1 only support hadoop-2.7 and hadoop-3.2, at least these two can
be given via profiles:  -Phadoop-2.7  and -Phadoop-3.2 (the default).


On Wed, Mar 10, 2021 at 12:26 PM jiahong li <monkeyboy....@gmail.com> wrote:

> i use ./build/mvn to compile ,and after execute command 
> :./build/zinc-0.3.15/bin/zinc
> -shutdown
> and execute command like this: /dev/make-distribution.sh --name
> custom-spark --pip  --tgz -Phive -Phive-thriftserver -Pyarn
> -Dhadoop.version=2.6.0-cdh5.13.1 -DskipTests
> same error appear.
> and execute command: ps -ef |grep zinc, there is nothing containe zinc
>
> Attila Zsolt Piros <piros.attila.zs...@gmail.com> 于2021年3月10日周三 下午6:55写道:
>
>> hi!
>>
>> Are you compiling Spark itself?
>> Do you use "./build/mvn" from the project root?
>> If you compiled an other version of Spark before and there the scala
>> version was different then zinc/nailgun could cached the old classes which
>> can cause similar troubles.
>> In that case this could help:
>>
>> ./build/zinc-0.3.15/bin/zinc -shutdown
>>
>> Best Regards,
>> Attila
>>
>> On Wed, Mar 10, 2021 at 11:27 AM jiahong li <monkeyboy....@gmail.com>
>> wrote:
>>
>>> hi, everybody, when i compile spark 3.1.1 from tag v3.1.1 ,encounter
>>> error like this:
>>>
>>> INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @
>>> spark-core_2.12 ---
>>> [INFO] Using incremental compilation using Mixed compile order
>>> [INFO] Compiler bridge file:
>>> .sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar
>>> [INFO] compiler plugin:
>>> BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null)
>>> [INFO] Compiling 560 Scala sources and 99 Java sources to
>>> git/spark/core/target/scala-2.12/classes ...
>>> [ERROR] [Error]
>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>> type mismatch;
>>>  found   : K where type K
>>>  required: String
>>> [ERROR] [Error]
>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>> value map is not a member of V
>>> [ERROR] [Error]
>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>> missing argument list for method stripXSS in class XssSafeRequest
>>> Unapplied methods are only converted to functions when a function type
>>> is expected.
>>> You can make this conversion explicit by writing `stripXSS _` or
>>> `stripXSS(_)` instead of `stripXSS`.
>>> [ERROR] [Error]
>>> git/spark/core/src/main/scala/org/apache/spark/ui/PagedTable.scala:307:
>>> value startsWith is not a member of K
>>> [ERROR] [Error]
>>> git/spark/core/src/main/scala/org/apache/spark/util/Utils.scala:580: value
>>> toLowerCase is not a member of object org.apache.hadoop.util.StringUtils
>>> [ERROR] 5 errors found
>>>
>>> anybody encounter error like this?
>>>
>>>
>>

Reply via email to