Hi Cleosson Jose, First of all, if you think this is a caused by a cassandra-lucene-index <https://github.com/Stratio/cassandra-lucene-index> bug, this user list is not the best way to report it. Please use github issues <https://github.com/Stratio/cassandra-lucene-index/issues> for this.
Second, in order to reproduce this error, i need to know which versions of cassandra, cassandra-lucene-index, spark and spark-cassandra-connector you are using Regards Eduardo Alonso Vía de las dos Castillas, 33, Ática 4, 3ª Planta 28224 Pozuelo de Alarcón, Madrid Tel: +34 91 828 6473 // www.stratio.com // *@stratiobd <https://twitter.com/StratioBD>* 2016-03-28 23:43 GMT+02:00 Cleosson José Pirani de Souza < cso...@daitangroup.com>: > Hi Jack, > > > Yes, I used the exact same commands in the Stratio readme. > > > Thanks, > > Cleososn > > > ------------------------------ > *From:* Jack Krupansky <jack.krupan...@gmail.com> > *Sent:* Monday, March 28, 2016 6:06 PM > *To:* user@cassandra.apache.org > > *Subject:* Re: Does saveToCassandra work with Cassandra Lucene plugin ? > > The exception message has an empty column name. Odd. Not sure if that is a > bug in the exception code or whether you actually have an empty column name > somewhere. > > Did you use the absolutely exact same commands to create the keyspace, > table, and custom index as in the Stratio readme? > > -- Jack Krupansky > > On Mon, Mar 28, 2016 at 4:57 PM, Cleosson José Pirani de Souza < > cso...@daitangroup.com> wrote: > >> Hi, >> >> One important thing, if I remove the custom index using Lucene, >> saveToCassandra works. >> >> >> Thanks >> >> Cleosson >> >> >> ------------------------------ >> *From:* Anuj Wadehra <anujw_2...@yahoo.co.in> >> *Sent:* Monday, March 28, 2016 3:27 PM >> *To:* user@cassandra.apache.org; Cleosson José Pirani de Souza; >> user@cassandra.apache.org >> *Subject:* Re: Does saveToCassandra work with Cassandra Lucene plugin ? >> >> I used it with Java and there, every field of Pojo must map to column >> names of the table. I think someone with Scala syntax knowledge can help >> you better. >> >> >> Thanks >> Anuj >> >> Sent from Yahoo Mail on Android >> <https://overview.mail.yahoo.com/mobile/?.src=Android> >> >> On Mon, 28 Mar, 2016 at 11:47 pm, Anuj Wadehra >> <anujw_2...@yahoo.co.in> wrote: >> With my limited experience with Spark, I can tell you that you need to >> make sure that all columns mentioned in somecolumns must be part of CQL >> schema of table. >> >> >> Thanks >> Anuj >> >> Sent from Yahoo Mail on Android >> <https://overview.mail.yahoo.com/mobile/?.src=Android> >> >> On Mon, 28 Mar, 2016 at 11:38 pm, Cleosson José Pirani de Souza >> <cso...@daitangroup.com> wrote: >> >> >> >> Hello, >> >> >> >> I am implementing the example on the github ( >> https://github.com/Stratio/cassandra-lucene-index) and when I try to >> save the data using saveToCassandra I get the exception >> NoSuchElementException. >> If I use CassandraConnector.withSessionDo I am able to add elements into >> Cassandra and no exception is raised. >> >> >> The code : >> import org.apache.spark.{SparkConf, SparkContext, Logging} >> import com.datastax.spark.connector.cql.CassandraConnector >> import com.datastax.spark.connector._ >> >> object App extends Logging{ >> def main(args: Array[String]) { >> >> // Get the cassandra IP and create the spark context >> val cassandraIP = System.getenv("CASSANDRA_IP"); >> val sparkConf = new SparkConf(true) >> .set("spark.cassandra.connection.host", >> cassandraIP) >> .set("spark.cleaner.ttl", "3600") >> .setAppName("Simple Spark Cassandra Example") >> >> >> * val sc = new SparkContext(sparkConf)* >> >> * // Works* >> * CassandraConnector(sparkConf).withSessionDo { session =>* >> * session.execute("INSERT INTO demo.tweets(id, user, body, >> time, latitude, longitude) VALUES (19, 'Name', 'Body', '2016-03-19 >> 09:00:00-0300', 39, 39)")* >> * }* >> >> * // Does not work* >> * val demo = sc.parallelize(Seq((9, "Name", "Body", "2016-03-29 >> 19:00:00-0300", 29, 29)))* >> * // Raises the exception* >> * demo.saveToCassandra("demo", "tweets", SomeColumns("id", "user", >> "body", "time", "latitude", "longitude"))* >> >> >> * } * >> *}* >> >> >> >> >> The exception: >> *16/03/28 14:15:41 INFO CassandraConnector: Connected to Cassandra >> cluster: Test Cluster* >> *Exception in thread "main" java.util.NoSuchElementException: Column not >> found in demo.tweets* >> at >> com.datastax.spark.connector.cql.StructDef$$anonfun$columnByName$2.apply(Schema.scala:60) >> at >> com.datastax.spark.connector.cql.StructDef$$anonfun$columnByName$2.apply(Schema.scala:60) >> at scala.collection.Map$WithDefault.default(Map.scala:52) >> at scala.collection.MapLike$class.apply(MapLike.scala:141) >> at scala.collection.AbstractMap.apply(Map.scala:58) >> at >> com.datastax.spark.connector.cql.TableDef$$anonfun$9.apply(Schema.scala:153) >> at >> com.datastax.spark.connector.cql.TableDef$$anonfun$9.apply(Schema.scala:152) >> at >> scala.collection.TraversableLike$WithFilter$$anonfun$map$2.apply(TraversableLike.scala:722) >> at scala.collection.immutable.Map$Map1.foreach(Map.scala:109) >> at >> scala.collection.TraversableLike$WithFilter.map(TraversableLike.scala:721) >> at com.datastax.spark.connector.cql.TableDef.<init>(Schema.scala:152) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$com$datastax$spark$connector$cql$Schema$$fetchTables$1$2.apply(Schema.scala:283) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$com$datastax$spark$connector$cql$Schema$$fetchTables$1$2.apply(Schema.scala:271) >> at >> scala.collection.TraversableLike$WithFilter$$anonfun$map$2.apply(TraversableLike.scala:722) >> at scala.collection.immutable.Set$Set4.foreach(Set.scala:137) >> at >> scala.collection.TraversableLike$WithFilter.map(TraversableLike.scala:721) >> at >> com.datastax.spark.connector.cql.Schema$.com$datastax$spark$connector$cql$Schema$$fetchTables$1(Schema.scala:271) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$com$datastax$spark$connector$cql$Schema$$fetchKeyspaces$1$2.apply(Schema.scala:295) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$com$datastax$spark$connector$cql$Schema$$fetchKeyspaces$1$2.apply(Schema.scala:294) >> at >> scala.collection.TraversableLike$WithFilter$$anonfun$map$2.apply(TraversableLike.scala:722) >> at scala.collection.immutable.HashSet$HashSet1.foreach(HashSet.scala:153) >> at >> scala.collection.immutable.HashSet$HashTrieSet.foreach(HashSet.scala:306) >> at >> scala.collection.TraversableLike$WithFilter.map(TraversableLike.scala:721) >> at >> com.datastax.spark.connector.cql.Schema$.com$datastax$spark$connector$cql$Schema$$fetchKeyspaces$1(Schema.scala:294) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$fromCassandra$1.apply(Schema.scala:307) >> at >> com.datastax.spark.connector.cql.Schema$$anonfun$fromCassandra$1.apply(Schema.scala:304) >> at >> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withClusterDo$1.apply(CassandraConnector.scala:121) >> at >> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withClusterDo$1.apply(CassandraConnector.scala:120) >> at >> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:110) >> at >> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:109) >> at >> com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:139) >> at >> com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109) >> at >> com.datastax.spark.connector.cql.CassandraConnector.withClusterDo(CassandraConnector.scala:120) >> at >> com.datastax.spark.connector.cql.Schema$.fromCassandra(Schema.scala:304) >> at >> com.datastax.spark.connector.writer.TableWriter$.apply(TableWriter.scala:275) >> at >> com.datastax.spark.connector.RDDFunctions.saveToCassandra(RDDFunctions.scala:36) >> at com.webradar.spci.spark.cassandra.App$.main(App.scala:27) >> at com.webradar.spci.spark.cassandra.App.main(App.scala) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:497) >> at >> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) >> at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) >> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) >> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >> >> >> Regards, >> Cleosson >> >> >> >> >> >