I removed core and streaming jar. And the exception still same.

I tried what you said then results:

~/cassandra/apache-cassandra-2.1.5$ bin/cassandra-cli -h localhost -p 9160
Connected to: "Test Cluster" on localhost/9160
Welcome to Cassandra CLI version 2.1.5

The CLI is deprecated and will be removed in Cassandra 3.0.  Consider
migrating to cqlsh.
CQL is fully backwards compatible with Thrift data; see
http://www.datastax.com/dev/blog/thrift-to-cql3

Type 'help;' or '?' for help.
Type 'quit;' or 'exit;' to quit.

[default@unknown]

and

~/cassandra/apache-cassandra-2.1.5$ bin/cqlsh
Connected to Test Cluster at 127.0.0.1:9042.
[cqlsh 5.0.1 | Cassandra 2.1.5 | CQL spec 3.2.0 | Native protocol v3]
Use HELP for help.
cqlsh>

Thank you for your kind responses ...


2015-06-09 20:59 GMT+03:00 Yana Kadiyska <yana.kadiy...@gmail.com>:

> Hm, jars look ok, although it's a bit of a mess -- you have spark-assembly
> 1.3.0 but then core and streaming 1.3.1...It's generally a bad idea to mix
> versions. Spark-assembly bundless all spark packages, so either do them
> separately or use spark-assembly but don't mix like you've shown.
>
> As to the port issue -- what about this:
>
> $bin/cassandra-cli -h localhost -p 9160
> Connected to: "Test Cluster" on localhost/9160
> Welcome to Cassandra CLI version 2.1.5
>
>
> On Tue, Jun 9, 2015 at 1:29 PM, Yasemin Kaya <godo...@gmail.com> wrote:
>
>> My jar files are:
>>
>> cassandra-driver-core-2.1.5.jar
>> cassandra-thrift-2.1.3.jar
>> guava-18.jar
>> jsr166e-1.1.0.jar
>> spark-assembly-1.3.0.jar
>> spark-cassandra-connector_2.10-1.3.0-M1.jar
>> spark-cassandra-connector-java_2.10-1.3.0-M1.jar
>> spark-core_2.10-1.3.1.jar
>> spark-streaming_2.10-1.3.1.jar
>>
>> And my code from datastax spark-cassandra-connector
>> <https://github.com/datastax/spark-cassandra-connector/blob/master/spark-cassandra-connector-demos/simple-demos/src/main/java/com/datastax/spark/connector/demo/JavaApiDemo.java>
>> .
>>
>> Thanx alot.
>> yasemin
>>
>> 2015-06-09 18:58 GMT+03:00 Yana Kadiyska <yana.kadiy...@gmail.com>:
>>
>>> hm. Yeah, your port is good...have you seen this thread:
>>> http://stackoverflow.com/questions/27288380/fail-to-use-spark-cassandra-connector
>>> ? It seems that you might be running into version mis-match issues?
>>>
>>> What versions of Spark/Cassandra-connector are you trying to use?
>>>
>>> On Tue, Jun 9, 2015 at 10:18 AM, Yasemin Kaya <godo...@gmail.com> wrote:
>>>
>>>> Sorry my answer I hit terminal lsof -i:9160: result is
>>>>
>>>> lsof -i:9160
>>>> COMMAND  PID    USER   FD   TYPE DEVICE SIZE/OFF NODE NAME
>>>> java    7597 inosens  101u  IPv4  85754      0t0  TCP localhost:9160
>>>> (LISTEN)
>>>>
>>>> so 9160 port is available or not ?
>>>>
>>>> 2015-06-09 17:16 GMT+03:00 Yasemin Kaya <godo...@gmail.com>:
>>>>
>>>>> Yes my cassandra is listening on 9160 I think. Actually I know from
>>>>> yaml file. The file includes :
>>>>>
>>>>> rpc_address: localhost
>>>>> # port for Thrift to listen for clients on
>>>>> rpc_port: 9160
>>>>>
>>>>> I check the port "nc -z localhost 9160; echo $?" it returns me "0". I
>>>>> think it close, should I open this port ?
>>>>>
>>>>> 2015-06-09 16:55 GMT+03:00 Yana Kadiyska <yana.kadiy...@gmail.com>:
>>>>>
>>>>>> Is your cassandra installation actually listening on 9160?
>>>>>>
>>>>>> lsof -i :9160COMMAND   PID     USER   FD   TYPE   DEVICE SIZE/OFF NODE 
>>>>>> NAME
>>>>>> java    29232 ykadiysk   69u  IPv4 42152497      0t0  TCP localhost:9160 
>>>>>> (LISTEN)
>>>>>>
>>>>>> ​
>>>>>> I am running an out-of-the box cassandra conf where
>>>>>>
>>>>>> rpc_address: localhost
>>>>>> # port for Thrift to listen for clients on
>>>>>> rpc_port: 9160
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Jun 9, 2015 at 7:36 AM, Yasemin Kaya <godo...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I couldn't find any solution. I can write but I can't read from
>>>>>>> Cassandra.
>>>>>>>
>>>>>>> 2015-06-09 8:52 GMT+03:00 Yasemin Kaya <godo...@gmail.com>:
>>>>>>>
>>>>>>>> Thanks alot Mohammed, Gerard and Yana.
>>>>>>>> I can write to table, but exception returns me. It says "*Exception
>>>>>>>> in thread "main" java.io.IOException: Failed to open thrift connection 
>>>>>>>> to
>>>>>>>> Cassandra at 127.0.0.1:9160 <http://127.0.0.1:9160>*"
>>>>>>>>
>>>>>>>> In yaml file :
>>>>>>>> rpc_address: localhost
>>>>>>>> rpc_port: 9160
>>>>>>>>
>>>>>>>> And at project :
>>>>>>>>
>>>>>>>> .set("spark.cassandra.connection.host", "127.0.0.1")
>>>>>>>> .set("spark.cassandra.connection.rpc.port", "9160");
>>>>>>>>
>>>>>>>> or
>>>>>>>>
>>>>>>>> .set("spark.cassandra.connection.host", "localhost")
>>>>>>>> .set("spark.cassandra.connection.rpc.port", "9160");
>>>>>>>>
>>>>>>>> whatever I write setting,  I get same exception. Any help ??
>>>>>>>>
>>>>>>>>
>>>>>>>> 2015-06-08 18:23 GMT+03:00 Yana Kadiyska <yana.kadiy...@gmail.com>:
>>>>>>>>
>>>>>>>>> yes, whatever you put for listen_address in cassandra.yaml. Also,
>>>>>>>>> you should try to connect to your cassandra cluster via bin/cqlsh to 
>>>>>>>>> make
>>>>>>>>> sure you have connectivity before you try to make a a connection via 
>>>>>>>>> spark.
>>>>>>>>>
>>>>>>>>> On Mon, Jun 8, 2015 at 4:43 AM, Yasemin Kaya <godo...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>> I run my project on local. How can find ip address of my
>>>>>>>>>> cassandra host ? From cassandra.yaml or ??
>>>>>>>>>>
>>>>>>>>>> yasemin
>>>>>>>>>>
>>>>>>>>>> 2015-06-08 11:27 GMT+03:00 Gerard Maas <gerard.m...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>>> ????? = <ip address of your cassandra host>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Jun 8, 2015 at 10:12 AM, Yasemin Kaya <godo...@gmail.com
>>>>>>>>>>> > wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi ,
>>>>>>>>>>>>
>>>>>>>>>>>> How can I find spark.cassandra.connection.host? And what should
>>>>>>>>>>>> I change ? Should I change cassandra.yaml ?
>>>>>>>>>>>>
>>>>>>>>>>>> Error says me *"Exception in thread "main"
>>>>>>>>>>>> java.io.IOException: Failed to open native connection to Cassandra 
>>>>>>>>>>>> at
>>>>>>>>>>>> {127.0.1.1}:9042"*
>>>>>>>>>>>>
>>>>>>>>>>>> What should I add *SparkConf sparkConf = new
>>>>>>>>>>>> SparkConf().setAppName("JavaApiDemo").set(**"spark.driver.allowMultipleContexts",
>>>>>>>>>>>> "true").set("spark.cassandra.connection.host", ?????);*
>>>>>>>>>>>>
>>>>>>>>>>>> Best
>>>>>>>>>>>> yasemin
>>>>>>>>>>>>
>>>>>>>>>>>> 2015-06-06 3:04 GMT+03:00 Mohammed Guller <
>>>>>>>>>>>> moham...@glassbeam.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>>  Check your spark.cassandra.connection.host setting. It
>>>>>>>>>>>>> should be pointing to one of your Cassandra nodes.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Mohammed
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> *From:* Yasemin Kaya [mailto:godo...@gmail.com]
>>>>>>>>>>>>> *Sent:* Friday, June 5, 2015 7:31 AM
>>>>>>>>>>>>> *To:* user@spark.apache.org
>>>>>>>>>>>>> *Subject:* Cassandra Submit
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> I am using cassandraDB in my project. I had that error *Exception
>>>>>>>>>>>>> in thread "main" java.io.IOException: Failed to open native 
>>>>>>>>>>>>> connection to
>>>>>>>>>>>>> Cassandra at {127.0.1.1}:9042*
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> I think I have to modify the submit line. What should I add or
>>>>>>>>>>>>> remove when I submit my project?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Best,
>>>>>>>>>>>>>
>>>>>>>>>>>>> yasemin
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>>
>>>>>>>>>>>>> hiç ender hiç
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> hiç ender hiç
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> hiç ender hiç
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> hiç ender hiç
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> hiç ender hiç
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> hiç ender hiç
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> hiç ender hiç
>>>>
>>>
>>>
>>
>>
>> --
>> hiç ender hiç
>>
>
>


-- 
hiç ender hiç

Reply via email to