This would also be a better question for the SCC user list :) https://groups.google.com/a/lists.datastax.com/forum/#!forum/spark-connector-user
On Sun, Sep 4, 2016 at 9:31 AM Russell Spitzer <russell.spit...@gmail.com> wrote: > > https://github.com/datastax/spark-cassandra-connector/blob/v1.3.1/doc/14_data_frames.md > In Spark 1.3 it was illegal to use "table" as a key in Spark SQL so in > that version of Spark the connector needed to use the option "c_table" > > > val df = sqlContext.read. > | format("org.apache.spark.sql.cassandra"). > | options(Map( "c_table" -> "****", "keyspace" -> "***")). > | load() > > > On Sun, Sep 4, 2016 at 8:32 AM Mich Talebzadeh <mich.talebza...@gmail.com> > wrote: > >> and your Cassandra table is there etc? >> >> >> >> Dr Mich Talebzadeh >> >> >> >> LinkedIn * >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >> >> >> >> http://talebzadehmich.wordpress.com >> >> >> *Disclaimer:* Use it at your own risk. Any and all responsibility for >> any loss, damage or destruction of data or any other property which may >> arise from relying on this email's technical content is explicitly >> disclaimed. The author will in no case be liable for any monetary damages >> arising from such loss, damage or destruction. >> >> >> >> On 4 September 2016 at 16:20, Selvam Raman <sel...@gmail.com> wrote: >> >>> Hey Mich, >>> >>> I am using the same one right now. Thanks for the reply. >>> import org.apache.spark.sql.cassandra._ >>> import com.datastax.spark.connector._ //Loads implicit functions >>> sc.cassandraTable("keyspace name", "table name") >>> >>> >>> On Sun, Sep 4, 2016 at 8:48 PM, Mich Talebzadeh < >>> mich.talebza...@gmail.com> wrote: >>> >>>> Hi Selvan. >>>> >>>> I don't deal with Cassandra but have you tried other options as >>>> described here >>>> >>>> >>>> https://github.com/datastax/spark-cassandra-connector/blob/master/doc/2_loading.md >>>> >>>> To get a Spark RDD that represents a Cassandra table, call the >>>> cassandraTable method on the SparkContext object. >>>> >>>> import com.datastax.spark.connector._ //Loads implicit functions >>>> sc.cassandraTable("keyspace name", "table name") >>>> >>>> >>>> >>>> HTH >>>> >>>> >>>> Dr Mich Talebzadeh >>>> >>>> >>>> >>>> LinkedIn * >>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>>> >>>> >>>> >>>> http://talebzadehmich.wordpress.com >>>> >>>> >>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for >>>> any loss, damage or destruction of data or any other property which may >>>> arise from relying on this email's technical content is explicitly >>>> disclaimed. The author will in no case be liable for any monetary damages >>>> arising from such loss, damage or destruction. >>>> >>>> >>>> >>>> On 4 September 2016 at 15:52, Selvam Raman <sel...@gmail.com> wrote: >>>> >>>>> its very urgent. please help me guys. >>>>> >>>>> On Sun, Sep 4, 2016 at 8:05 PM, Selvam Raman <sel...@gmail.com> wrote: >>>>> >>>>>> Please help me to solve the issue. >>>>>> >>>>>> spark-shell --packages >>>>>> com.datastax.spark:spark-cassandra-connector_2.10:1.3.0 --conf >>>>>> spark.cassandra.connection.host=****** >>>>>> >>>>>> val df = sqlContext.read. >>>>>> | format("org.apache.spark.sql.cassandra"). >>>>>> | options(Map( "table" -> "****", "keyspace" -> "***")). >>>>>> | load() >>>>>> java.util.NoSuchElementException: key not found: c_table >>>>>> at scala.collection.MapLike$class.default(MapLike.scala:228) >>>>>> at >>>>>> org.apache.spark.sql.execution.datasources.CaseInsensitiveMap.default(ddl.scala:151) >>>>>> at scala.collection.MapLike$class.apply(MapLike.scala:141) >>>>>> at >>>>>> org.apache.spark.sql.execution.datasources.CaseInsensitiveMap.apply(ddl.scala:151) >>>>>> at >>>>>> org.apache.spark.sql.cassandra.DefaultSource$.TableRefAndOptions(DefaultSource.scala:120) >>>>>> at >>>>>> org.apache.spark.sql.cassandra.DefaultSource.createRelation(DefaultSource.scala:56) >>>>>> at >>>>>> org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:125) >>>>>> a >>>>>> >>>>>> -- >>>>>> Selvam Raman >>>>>> "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து" >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Selvam Raman >>>>> "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து" >>>>> >>>> >>>> >>> >>> >>> -- >>> Selvam Raman >>> "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து" >>> >> >>