Hi All,
Spark Shell doesnt seem to use spark workers but Spark Submit does. I had the workers ips listed under conf/slaves file. I am trying to count number of rows in Cassandra using spark-shell so I do the following on spark master val df = spark.sql("SELECT test from hello") // This has about billion rows scala> df.count [Stage 0:=> (686 + 2) / 24686] // What are these numbers precisely? This is taking forever so I checked the I/O, CPU, Network usage using dstat, iostat and so on it looks like nothing is going on in worker machines but for master I can see it. I am using spark 2.0.2 Any ideas on what is going on? and how to fix this? Thanks, kant