Hi,
i am trying to read cassandra table as Dataframe but got the below issue
spark-shell --packages com.datastax.spark:spark-cassandra-connector_2.10:1.3.0
--conf spark.cassandra.connection.host=******
val df = sqlContext.read.
| format("org.apache.spark.sql.cassandra").
| options(Map( "table" -> "****", "keyspace" -> "***")).
| load()
java.util.NoSuchElementException: key not found: c_table
at scala.collection.MapLike$class.default(MapLike.scala:228)
at org.apache.spark.sql.execution.datasources.
CaseInsensitiveMap.default(ddl.scala:151)
at scala.collection.MapLike$class.apply(MapLike.scala:141)
at org.apache.spark.sql.execution.datasources.
CaseInsensitiveMap.apply(ddl.scala:151)
at org.apache.spark.sql.cassandra.DefaultSource$.TableRefAndOptions(
DefaultSource.scala:120)
at org.apache.spark.sql.cassandra.DefaultSource.
createRelation(DefaultSource.scala:56)
at org.apache.spark.sql.execution.datasources.
ResolvedDataSource$.apply(ResolvedDataSource.scala:125)
When i am using
SC.CassandraTable(tablename,keyspace) it went fine. but when i crate action
it throws plenty of error
example:
com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured
columnfamily size_estimates
--
Selvam Raman
"லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"