There are plenty of spark / akka / scala / elasticsearch-hadoop dependencies to keep track of.
Is it true that elasticsearch-hadoop needs to be compiled for a specific spark version to run correctly on the cluster? I'm also trying to keep track of the akka version and scala version. i.e, wil es-hadoop compiled for spark 1.2 work with Spark 1.3 ? When the elasticsearch-hadoop versions are released, as v2.0 v2.1, v2.1.0.Beta3, at what point do we need to keep in mind what spark version it was also compiled against? i.e., is it safe to assume the es-hadoop versions are tied to a specific spark core version? I've been keeping the following chart in my notes to see what all the versions are with all dependencies ============================= Akka Version Dependencies Current Akka Stable Release: 2.3.9 Elasticsearch-Hadoop: 2.1.0Beta3 = Spark 1.1.0 Elasticsearch-Hadoop: 2.1.0Beta3-SNAPSHOT = Spark 1.2.1 Elasticsearch-Hadoop: what about spark 1.3 ? Spark: 1.3, Akka: 2.3.4-spark Spark: 1.2, Akka: 2.3.4-spark Spark: 1.1, Akka: 2.2.3-shaded-protobuf Activator 1.2.12 comes with with Akka 2.3.4 Play 2.3.8, akka 2.3.4, scala 2.11.1 (will also work with 2.10.4 ) Play 2.2.x, akka 2.2.0 Spark Job Server 0.4.1, Spark Core 1.1.0, Akka, 2.2.4 Spark Job Server Master as of Feb 22, 2015, Spark Core 1.2.0, Akka 2.3.4, Scala 2.10.4 Akka persistence latest 2.3.4 or later Akka 2.3.9 is released for Scala 2.10.4 and 2.11.5 -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/28ad3f78-8b3d-450a-a29d-06d3e6636cfd%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.
