Thanks. But, I am required to do a maven release to Nexus on spark 2.0.2 built against scala 2.10. How can I go about with this? Is this a bug that I need to open in Spark jira?
On Mon, Jun 19, 2017 at 12:12 PM, Shixiong(Ryan) Zhu < shixi...@databricks.com> wrote: > Some of projects (such as spark-tags) are Java projects. Spark doesn't > fix the artifact name and just hard-core 2.11. > > For your issue, try to use `install` rather than `package`. > > On Sat, Jun 17, 2017 at 7:20 PM, Kanagha Kumar <kpra...@salesforce.com> > wrote: > >> Hi, >> >> Bumping up again! Why does spark modules depend upon scala2.11 versions >> inspite of changing pom.xmls using ./dev/change-scala-version.sh 2.10. >> Appreciate any quick help!! >> >> Thanks >> >> On Fri, Jun 16, 2017 at 2:59 PM, Kanagha Kumar <kpra...@salesforce.com> >> wrote: >> >>> Hey all, >>> >>> >>> I'm trying to use Spark 2.0.2 with scala 2.10 by following this >>> https://spark.apache.org/docs/2.0.2/building-spark.html >>> #building-for-scala-210 >>> >>> ./dev/change-scala-version.sh 2.10 >>> ./build/mvn -Pyarn -Phadoop-2.4 -Dscala-2.10 -DskipTests clean package >>> >>> >>> I could build the distribution successfully using >>> bash -xv dev/make-distribution.sh --tgz -Dscala-2.10 -DskipTests >>> >>> But, when I am trying to maven release, it keeps failing with the error >>> using the command: >>> >>> >>> Executing Maven: -B -f pom.xml -DscmCommentPrefix=[maven-release-plugin] >>> -e -Dscala-2.10 -Pyarn -Phadoop-2.7 -Phadoop-provided -DskipTests >>> -Dresume=false -U -X *release:prepare release:perform* >>> >>> Failed to execute goal on project spark-sketch_2.10: Could not resolve >>> dependencies for project >>> org.apache.spark:spark-sketch_2.10:jar:2.0.2-sfdc-3.0.0: >>> *Failure to find org.apache.spark:spark-tags_2.11:jar:2.0.2-sfdc-3.0.0* >>> in <a .. nexus repo...> was cached in the local repository, resolution will >>> not be reattempted until the update interval of nexus has elapsed or >>> updates are forced -> [Help 1] >>> >>> >>> Why does spark-sketch depend upon spark-tags_2.11 when I have already >>> compiled against scala 2.10?? Any pointers would be helpful. >>> Thanks >>> Kanagha >>> >> >> >