Hi All,
I was able to resolve this matter with a simple fix. It seems that in order
to process a reduceByKey and the flat map operations at the same time, the
only way to resolve was to increase the number of threads to 1.
Since I'm developing on my personal machine for speed, I simply updated the
sparkURL argument to:
private static String sparkURL = local[2]; //Instead of local
,which is then used by the JavaStreamingContext method as a parameter.
After I made this change, I was able to see the reduceByKey values properly
aggregated and counted.
Best Regards,
D
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/reduceByKey-Not-Being-Called-by-Spark-Streaming-tp8684p8739.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.