Greetings Spark users/devs! I'm interested in using Spark to process large volumes of data with a geospatial component, and I haven't been able to find much information on Spark's ability to handle this kind of operation. I don't need anything too complex; just distance between two points, point-in-polygon and the like.
Does Spark (or possibly Shark) support this kind of query? Has anyone written a plugin/extension along these lines? If there isn't anything like this so far, then it seems like I have two options. I can either abandon Spark and fall back on Hadoop and Hive with the ESRI Tools extension, or I can stick with Spark and try to write/port a GIS toolkit. Which option do you think I should pursue? How hard is it for someone that's new to the Spark codebase to write an extension? Is there anyone else in the community that would be interested in having geospatial capability in Spark? Thanks for your help! -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Is-Spark-a-good-choice-for-geospatial-GIS-applications-Is-a-community-volunteer-needed-in-this-area-tp4685.html Sent from the Apache Spark User List mailing list archive at Nabble.com.