What if you don't run zinc? I.e. just download maven and run that "mvn package...". It might take longer, but I wonder if it will work.
On Mon, Apr 6, 2015 at 10:26 PM, mjhb <sp...@mjhb.com> wrote: > Similar problem on 1.2 branch: > > [ERROR] Failed to execute goal on project spark-core_2.11: Could not resolve > dependencies for project > org.apache.spark:spark-core_2.11:jar:1.2.3-SNAPSHOT: The following artifacts > could not be resolved: > org.apache.spark:spark-network-common_2.10:jar:1.2.3-SNAPSHOT, > org.apache.spark:spark-network-shuffle_2.10:jar:1.2.3-SNAPSHOT: Failure to > find org.apache.spark:spark-network-common_2.10:jar:1.2.3-SNAPSHOT in > http://repository.apache.org/snapshots was cached in the local repository, > resolution will not be reattempted until the update interval of > apache.snapshots has elapsed or updates are forced -> [Help 1] > org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute > goal on project spark-core_2.11: Could not resolve dependencies for project > org.apache.spark:spark-core_2.11:jar:1.2.3-SNAPSHOT: The following artifacts > could not be resolved: > org.apache.spark:spark-network-common_2.10:jar:1.2.3-SNAPSHOT, > org.apache.spark:spark-network-shuffle_2.10:jar:1.2.3-SNAPSHOT: Failure to > find org.apache.spark:spark-network-common_2.10:jar:1.2.3-SNAPSHOT in > http://repository.apache.org/snapshots was cached in the local repository, > resolution will not be reattempted until the update interval of > apache.snapshots has elapsed or updates are forced > > > > > -- > View this message in context: > http://apache-spark-developers-list.1001551.n3.nabble.com/1-3-Build-Error-with-Scala-2-11-tp11441p11442.html > Sent from the Apache Spark Developers List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org