Re: SnappyCompressor and Cassandra 1.1.1

2012-07-02 Thread Sylvain Lebresne
Yeah, we tried to catch when Snappy couldn't load (and deactivate it
when that happens) but apparently we've missed a few of the exceptions
that can be thrown in that case. I've created
https://issues.apache.org/jira/browse/CASSANDRA-4400 to fix that. If
you could try the patch on that issue and check it works for you that
would be awesome since I don't have a Raspberry Pi myself to test.

--
Sylvain

On Sun, Jul 1, 2012 at 3:26 PM, Andy Cobley
acob...@computing.dundee.ac.uk wrote:
 I'm running Cassandra on Raspberry Pi (for educational reason) and have been 
 successfully running 1.1.0 for some time.  However there is no native build 
 of SnappyCompressor for the platform (I'm currently working n rectifying that 
 if I can) so that compression is unavailable.  When I try and start 1.1.1 on 
 the platform I'm getting the following error which looks to me like 1.1.1 is 
 trying to load snappy compressor at startup and falls over when to can't find 
 it.  Thats not been the case with 1.1.0:

 INFO 14:22:07,600 Global memtable threshold is enabled at 35MB
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:616)
 at 
 org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor.java:45)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompressor.java:55)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompressor.java:37)
 at org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
 org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:439)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCassandraDaemon.java:353)
 at 
 org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
 at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1681)
 at java.lang.Runtime.loadLibrary0(Runtime.java:840)
 at java.lang.System.loadLibrary(System.java:1047)
 at 
 org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
 ... 17 more
 ERROR 14:22:09,934 Exception encountered during startup
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor.java:45)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompressor.java:55)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompressor.java:37)
 at org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
 org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:439)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCassandraDaemon.java:353)
 at 
 org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106)
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor.java:45)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompressor.java:55)
 at 
 org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompressor.java:37)
 at org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
 org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:439)
 at 
 

Re: Bootstrap code path

2012-07-02 Thread Sylvain Lebresne
On Sun, Jul 1, 2012 at 3:31 PM, Bill Hastings bllhasti...@gmail.com wrote:
 Could someone please tell me where I should start looking at code to
 understand how cassandra bootstrap process works?

The bootstrap phase pretty much start in the bootstrap() method in
StorageService.java, though a good part of the code is in
Bootstrapper.java.


 Also is my understanding correct that the new nodes that are added
 are not joining the ring till the bootstrap process is complete i.e do
 not receive any read or write requests from outside?

It is true for reads but not for writes. Writes are send early in the
bootstrap process otherwise the node wouldn't be up to date once the
bootstrap is completed.

--
Sylvain


Re: BulkLoading SSTables and compression

2012-07-02 Thread Sylvain Lebresne
On Mon, Jul 2, 2012 at 12:14 AM, jmodha jmo...@gmail.com wrote:
 Sure, before I create a ticket, is there a way I can confirm that the
 sstables are indeed not compressed other than running the rebuildsstables
 nodetool command (and observing the live size go down)?

If the sstable is compressed it will have a -CompressionInfo.db
component (i.e. one of the file composing the sstable with end with
--CompressionInfo.db). If there is no such file then it'ss not
compressed.

--
Sylvain


 Thanks.

 --
 View this message in context: 
 http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/BulkLoading-SSTables-and-compression-tp7580849p7580922.html
 Sent from the cassandra-u...@incubator.apache.org mailing list archive at 
 Nabble.com.


Re: Memtable tuning in 1.0 and higher

2012-07-02 Thread Jonathan Ellis
I'm afraid not. It's too much change for an oldstable release series,
and the bulk of the change is to AtomicSortedColumns which doesn't
exist in 1.0, so even if we wanted to take a maybe it's okay if we
release it first in 1.1.3 and then backport approach it wouldn't
improve our safety margin since you'd basically need to rewrite the
patch.

On Sun, Jul 1, 2012 at 6:40 AM, Joost Van De Wijgerd jwijg...@gmail.com wrote:
 Hi Jonathan,

 Looks good, any chance of porting this fix to the 1.0 branch?

 Kind regards

 Joost

 Sent from my iPhone


 On 1 jul. 2012, at 09:25, Jonathan Ellis jbel...@gmail.com wrote:

 On Thu, Jun 28, 2012 at 1:39 PM, Joost van de Wijgerd
 jwijg...@gmail.com wrote:
 the currentThoughput is increased even before the data is merged into the
 memtable so it is actually measuring the throughput afaik.

 You're right.  I've attached a patch to
 https://issues.apache.org/jira/browse/CASSANDRA-4399 to fix this.

 --
 Jonathan Ellis
 Project Chair, Apache Cassandra
 co-founder of DataStax, the source for professional Cassandra support
 http://www.datastax.com



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of DataStax, the source for professional Cassandra support
http://www.datastax.com


Re: SnappyCompressor and Cassandra 1.1.1

2012-07-02 Thread Andy Cobley
Thanks ,

Will give it a go now.

Andy

On 02/07/2012 08:13, Sylvain Lebresne sylv...@datastax.com wrote:

Yeah, we tried to catch when Snappy couldn't load (and deactivate it
when that happens) but apparently we've missed a few of the exceptions
that can be thrown in that case. I've created
https://issues.apache.org/jira/browse/CASSANDRA-4400 to fix that. If
you could try the patch on that issue and check it works for you that
would be awesome since I don't have a Raspberry Pi myself to test.

--
Sylvain

On Sun, Jul 1, 2012 at 3:26 PM, Andy Cobley
acob...@computing.dundee.ac.uk wrote:
 I'm running Cassandra on Raspberry Pi (for educational reason) and have
been successfully running 1.1.0 for some time.  However there is no
native build of SnappyCompressor for the platform (I'm currently working
n rectifying that if I can) so that compression is unavailable.  When I
try and start 1.1.1 on the platform I'm getting the following error
which looks to me like 1.1.1 is trying to load snappy compressor at
startup and falls over when to can't find it.  Thats not been the case
with 1.1.0:

 INFO 14:22:07,600 Global memtable threshold is enabled at 35MB
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
mpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:616)
 at 
org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescripto
r.java:439)
 at 
org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescripto
r.java:118)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassan
draDaemon.java:126)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCas
sandraDaemon.java:353)
 at 
org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106
)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in
java.library.path
 at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1681)
 at java.lang.Runtime.loadLibrary0(Runtime.java:840)
 at java.lang.System.loadLibrary(System.java:1047)
 at 
org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:
52)
 ... 17 more
 ERROR 14:22:09,934 Exception encountered during startup
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescripto
r.java:439)
 at 
org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescripto
r.java:118)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassan
draDaemon.java:126)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCas
sandraDaemon.java:353)
 at 
org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106
)
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 

Re: ClassCastException during Cassandra server startup

2012-07-02 Thread Thierry Templier

Hello Yuki,

Could you give me hints about where to find these files. I have a look 
in the installation folder of Cassandra and in the /var/lib/cassandra 
folder?


Thanks very much for your help.
Thierry
That was bug in 1.1.1 and fixed in 
https://issues.apache.org/jira/browse/CASSANDRA-4331.
Workaround is deleting the key cache files for your index CFs should 
fix this.



Yuki


Nodes marked dead…. leap second?

2012-07-02 Thread Filippo Diotalevi
Hi,  
we had some really weird issues during the weekend, with our cassandra nodes 
starting marking as dead other (working) nodes in the cluster. That happened 
all Sunday, and it's still happening. Node are marked dead and up all the time….

Some example logs:

INFO [GossipTasks:1] 2012-07-02 06:55:01,804 Gossiper.java (line 818) 
InetAddress /xx.xx.xx.233 is now dead.
INFO [GossipTasks:1] 2012-07-02 06:55:01,805 Gossiper.java (line 818) 
InetAddress /xx.xx.xx.235 is now dead.
INFO [GossipStage:1] 2012-07-02 06:55:21,748 Gossiper.java (line 804) 
InetAddress /xx.xx.xx.233 is now UP
INFO [GossipStage:1] 2012-07-02 06:55:21,893 Gossiper.java (line 804) 
InetAddress /xx.xx.xx.235 is now UP
INFO [GossipTasks:1] 2012-07-02 06:56:03,877 Gossiper.java (line 818) 
InetAddress /xx.xx.xx.235 is now dead.
INFO [GossipTasks:1] 2012-07-02 06:57:58,537 Gossiper.java (line 818) 
InetAddress /xx.xx.xx.233 is now dead.
INFO [GossipStage:1] 2012-07-02 06:59:06,444 Gossiper.java (line 804) 
InetAddress /xx.xx.xx.233 is now UP



I couldn't find any real exception in the logs, but I noticed that the first 
error occurred at  
 INFO [GossipTasks:1] 2012-07-01 02:00:31,169 Gossiper.java (line 818) 
InetAddress /xx.xx.xx.234 is now dead.

2012-07-01 02:00:31,169, in the German timezone were the machine is hosted, is 
June 30th 23:59:60 UTC, the leap second that caused quite a few issues this 
weekend.  

Can it be the cause of the cluster failure? Has anybody noticed similar issues? 
( also see https://twitter.com/redditstatus/status/219244389044731904 )

I'm running Ubuntu 10.04.3 LTS.  

Many thanks,
--  
Filippo Diotalevi



cql 3.0.0

2012-07-02 Thread Abhijit Chanda
Hi All,I am using cassandra 1.0.8. How can i use cql 3.0.0 in this?


-- 
Abhijit Chanda
Analyst
VeHere Interactive Pvt. Ltd.
+91-974395


Re: BulkLoading SSTables and compression

2012-07-02 Thread jmodha
Thanks Sylvain.

I had a look at a node where we streamed data to and I do indeed see the
..-CompressionInfo.db files..

However, prior to running the upgradesstables command, the total size of
all the SSTables was 27GB and afterwards its 12GB.

So even though the CompressionInfo files were there immediately after bulk
loading the data, it wasn't really compressed..?

Can you think of anything else I can try to confirm this is indeed a bug?

Out of interest, we're not specifying a specific chunk size on the schema
(in the hope that it would just use the default of 64kb), so it reads
something like:

create column family test
  with column_type = 'Standard'
  and comparator = 'BytesType'
  and default_validation_class = 'UTF8Type'
  and key_validation_class = 'BytesType'
  and compaction_strategy =
'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'
  and compression_options = {'sstable_compression' :
'org.apache.cassandra.io.compress.SnappyCompressor'};

Would this cause any issues? 



--
View this message in context: 
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/BulkLoading-SSTables-and-compression-tp7580849p7580933.html
Sent from the cassandra-u...@incubator.apache.org mailing list archive at 
Nabble.com.


Re: cql 3.0.0

2012-07-02 Thread Sylvain Lebresne
On Mon, Jul 2, 2012 at 1:09 PM, Abhijit Chanda
abhijit.chan...@gmail.com wrote:
 Hi All,

 I am using cassandra 1.0.8. How can i use cql 3.0.0 in this?

You can't, CQL 3 is only in Cassandra 1.1 onwards.

--
Sylvain



 --
 Abhijit Chanda
 Analyst
 VeHere Interactive Pvt. Ltd.
 +91-974395



Re: BulkLoading SSTables and compression

2012-07-02 Thread Sylvain Lebresne
On Mon, Jul 2, 2012 at 1:24 PM, jmodha jmo...@gmail.com wrote:
 Out of interest, we're not specifying a specific chunk size on the schema
 (in the hope that it would just use the default of 64kb), so it reads
 something like:

Actually, we've had bugs related to the handling of the chunk size in early 1.0
releases and if I remember correctly that was fully fixed only in
1.0.5. So it is
possible that your 1.0.3 sstables are cmpressed but with a ridiculously low
chunk size (like 64 bytes instead of 64K), which would amount to no
compression.

In that case then yes it is a bug, but one that has been fixed by
upgrading to 1.1.

--
Sylvain

 create column family test
   with column_type = 'Standard'
   and comparator = 'BytesType'
   and default_validation_class = 'UTF8Type'
   and key_validation_class = 'BytesType'
   and compaction_strategy =
 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy'
   and compression_options = {'sstable_compression' :
 'org.apache.cassandra.io.compress.SnappyCompressor'};

 Would this cause any issues?



 --
 View this message in context: 
 http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/BulkLoading-SSTables-and-compression-tp7580849p7580933.html
 Sent from the cassandra-u...@incubator.apache.org mailing list archive at 
 Nabble.com.


Re: Nodes marked dead…. leap second?

2012-07-02 Thread Jean Paul Adant
Hi,

I did have the same problem with cassandra 1.1.1 on Ubuntu 11.10
I had to reboot all nodes
I'm interested in any information about this.

Thanks

Jean Paul

2012/7/2 Filippo Diotalevi fili...@ntoklo.com

  Hi,
 we had some really weird issues during the weekend, with our cassandra
 nodes starting marking as dead other (working) nodes in the cluster. That
 happened all Sunday, and it's still happening. Node are marked dead and up
 all the time….

 Some example logs:

 INFO [GossipTasks:1] 2012-07-02 06:55:01,804 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.233 is now dead.
 INFO [GossipTasks:1] 2012-07-02 06:55:01,805 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.235 is now dead.
 INFO [GossipStage:1] 2012-07-02 06:55:21,748 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.233 is now UP
 INFO [GossipStage:1] 2012-07-02 06:55:21,893 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.235 is now UP
 INFO [GossipTasks:1] 2012-07-02 06:56:03,877 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.235 is now dead.
 INFO [GossipTasks:1] 2012-07-02 06:57:58,537 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.233 is now dead.
 INFO [GossipStage:1] 2012-07-02 06:59:06,444 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.233 is now UP


 I couldn't find any real exception in the logs, but I noticed that the
 first error occurred at
  INFO [GossipTasks:1] 2012-07-01 02:00:31,169 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.234 is now dead.

 2012-07-01 02:00:31,169, in the German timezone were the machine is
 hosted, is June 30th 23:59:60 UTC, the leap second that caused quite a few
 issues this weekend.

 Can it be the cause of the cluster failure? Has anybody noticed similar
 issues? ( also see
 https://twitter.com/redditstatus/status/219244389044731904 )

 I'm running Ubuntu 10.04.3 LTS.

 Many thanks,
 --
 Filippo Diotalevi




-- 
-
Jean Paul Adant - Créative-Ingénierie
jean.paul.ad...@gmail.com


Re: Nodes marked dead…. leap second?

2012-07-02 Thread Henrik Schröder
Bug: https://lkml.org/lkml/2012/6/30/122

Simple fix to reset the leap second flag: date; date `date
+%m%d%H%M%C%y.%S`; date;


/Henrik

On Mon, Jul 2, 2012 at 1:56 PM, Jean Paul Adant
jean.paul.ad...@gmail.comwrote:

 Hi,

 I did have the same problem with cassandra 1.1.1 on Ubuntu 11.10
 I had to reboot all nodes
 I'm interested in any information about this.

 Thanks

 Jean Paul

 2012/7/2 Filippo Diotalevi fili...@ntoklo.com

  Hi,
 we had some really weird issues during the weekend, with our cassandra
 nodes starting marking as dead other (working) nodes in the cluster. That
 happened all Sunday, and it's still happening. Node are marked dead and up
 all the time….

 Some example logs:

 INFO [GossipTasks:1] 2012-07-02 06:55:01,804 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.233 is now dead.
 INFO [GossipTasks:1] 2012-07-02 06:55:01,805 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.235 is now dead.
 INFO [GossipStage:1] 2012-07-02 06:55:21,748 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.233 is now UP
 INFO [GossipStage:1] 2012-07-02 06:55:21,893 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.235 is now UP
 INFO [GossipTasks:1] 2012-07-02 06:56:03,877 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.235 is now dead.
 INFO [GossipTasks:1] 2012-07-02 06:57:58,537 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.233 is now dead.
 INFO [GossipStage:1] 2012-07-02 06:59:06,444 Gossiper.java (line 804)
 InetAddress /xx.xx.xx.233 is now UP


 I couldn't find any real exception in the logs, but I noticed that the
 first error occurred at
  INFO [GossipTasks:1] 2012-07-01 02:00:31,169 Gossiper.java (line 818)
 InetAddress /xx.xx.xx.234 is now dead.

 2012-07-01 02:00:31,169, in the German timezone were the machine is
 hosted, is June 30th 23:59:60 UTC, the leap second that caused quite a few
 issues this weekend.

 Can it be the cause of the cluster failure? Has anybody noticed similar
 issues? ( also see
 https://twitter.com/redditstatus/status/219244389044731904 )

 I'm running Ubuntu 10.04.3 LTS.

 Many thanks,
 --
 Filippo Diotalevi




 --
 -
 Jean Paul Adant - Créative-Ingénierie
 jean.paul.ad...@gmail.com






Re: BulkLoading SSTables and compression

2012-07-02 Thread jmodha
Just to clarify, the data that we're loading SSTables from (v1.0.3) doesn't
have compression enabled on any of the CF's. 

So in theory the compression should occur on the receiving end (v1.1.1) as
we're going from uncompressed data to compressed data.

So I'm not sure if the bug you mention is causing the behaviour we're seeing
here.

The only thing I can think of is that the upgradesstables option follows a
slightly different path to the bulk uploader when it comes to generating the
sstables that have been flushed to disk?

--
View this message in context: 
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/BulkLoading-SSTables-and-compression-tp7580849p7580938.html
Sent from the cassandra-u...@incubator.apache.org mailing list archive at 
Nabble.com.


cassandra on re-Start

2012-07-02 Thread puneet loya
Hi,

I have set up cassandra on my linux system.

When I restarted the system , it is showing the keyspace does not exist.

Not even letting me to create the keyspace with the same name again.

Did anybody find this issue?

Any solutions on it?


Cheers,

Puneet


forceUserDefinedCompaction in 1.1.0

2012-07-02 Thread Pieter Callewaert
Hi guys,

We have a 6-node 1.0.9 cluster for production and 3-node 1.1.0 cluster for 
testing the new version of Cassandra.
In both we insert data in a particular CF with always a TTL of 31 days.
To clean up the files faster  we use the forceUserDefinedCompaction to manually 
force compaction on the old sstables which we are sure the data has expired.
In 1.0 this works perfect, but in the 1.1 the command executes without error, 
but in the log of the node I see the following:

INFO [CompactionExecutor:546] 2012-07-02 15:05:11,837 CompactionManager.java 
(line 337) Will not compact MapData024CTwT/MapData024CTwT-HOS-hc-21150: it is 
not an active sstable
INFO [CompactionExecutor:546] 2012-07-02 15:05:11,838 CompactionManager.java 
(line 350) No file to compact for user defined compaction

I am pretty sure the sstable because it should contain partly expired data en 
partly data which is still active.
Does this have to do something with the new directory structure from 1.1 ? Or 
are the parameters changed from the function?

Kind regards,
Pieter Callewaert



RE: forceUserDefinedCompaction in 1.1.0

2012-07-02 Thread Pieter Callewaert
Hi,

While I was typing my mail I had the idea to try with the new directory layout.
It seems you have to change the parameter settings from 1.0 to 1.1
In 1.0:
Param 1: keyspace name
Param 2: datafile name
In 1.1:
Param 1: keyspace name
Param 2: column family name/datafile name

Don't know if this is a bug or a breaking change ?

Kind regards,
Pieter Callewaert

From: Pieter Callewaert [mailto:pieter.callewa...@be-mobile.be]
Sent: maandag 2 juli 2012 15:10
To: user@cassandra.apache.org
Subject: forceUserDefinedCompaction in 1.1.0

Hi guys,

We have a 6-node 1.0.9 cluster for production and 3-node 1.1.0 cluster for 
testing the new version of Cassandra.
In both we insert data in a particular CF with always a TTL of 31 days.
To clean up the files faster  we use the forceUserDefinedCompaction to manually 
force compaction on the old sstables which we are sure the data has expired.
In 1.0 this works perfect, but in the 1.1 the command executes without error, 
but in the log of the node I see the following:

INFO [CompactionExecutor:546] 2012-07-02 15:05:11,837 CompactionManager.java 
(line 337) Will not compact MapData024CTwT/MapData024CTwT-HOS-hc-21150: it is 
not an active sstable
INFO [CompactionExecutor:546] 2012-07-02 15:05:11,838 CompactionManager.java 
(line 350) No file to compact for user defined compaction

I am pretty sure the sstable because it should contain partly expired data en 
partly data which is still active.
Does this have to do something with the new directory structure from 1.1 ? Or 
are the parameters changed from the function?

Kind regards,
Pieter Callewaert



Any reason to limit one's self to a single high level java client?

2012-07-02 Thread David Leimbach
I recognize that behind the scenes there's connection pooling and all kinds
of nice asynchronous dispatch of requests to cassandra, but is there any
sort of reason to avoid using different Java clients in the same
application?

I'm noticing that some are better suited to certain kinds activity than
others.

Dave


Re: SnappyCompressor and Cassandra 1.1.1

2012-07-02 Thread Andy Cobley
I've tested this and added a not to issue 4400.  Hopefully that¹s the
correct thing to do.

Andy

On 02/07/2012 08:13, Sylvain Lebresne sylv...@datastax.com wrote:

Yeah, we tried to catch when Snappy couldn't load (and deactivate it
when that happens) but apparently we've missed a few of the exceptions
that can be thrown in that case. I've created
https://issues.apache.org/jira/browse/CASSANDRA-4400 to fix that. If
you could try the patch on that issue and check it works for you that
would be awesome since I don't have a Raspberry Pi myself to test.

--
Sylvain

On Sun, Jul 1, 2012 at 3:26 PM, Andy Cobley
acob...@computing.dundee.ac.uk wrote:
 I'm running Cassandra on Raspberry Pi (for educational reason) and have
been successfully running 1.1.0 for some time.  However there is no
native build of SnappyCompressor for the platform (I'm currently working
n rectifying that if I can) so that compression is unavailable.  When I
try and start 1.1.1 on the platform I'm getting the following error
which looks to me like 1.1.1 is trying to load snappy compressor at
startup and falls over when to can't find it.  Thats not been the case
with 1.1.0:

 INFO 14:22:07,600 Global memtable threshold is enabled at 35MB
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
:57)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
mpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:616)
 at 
org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescripto
r.java:439)
 at 
org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescripto
r.java:118)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassan
draDaemon.java:126)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCas
sandraDaemon.java:353)
 at 
org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106
)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in
java.library.path
 at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1681)
 at java.lang.Runtime.loadLibrary0(Runtime.java:840)
 at java.lang.System.loadLibrary(System.java:1047)
 at 
org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:
52)
 ... 17 more
 ERROR 14:22:09,934 Exception encountered during startup
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 
org.apache.cassandra.config.KSMetaData.systemKeyspace(KSMetaData.java:79)
 at 
org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescripto
r.java:439)
 at 
org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescripto
r.java:118)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassan
draDaemon.java:126)
 at 
org.apache.cassandra.service.AbstractCassandraDaemon.activate(AbstractCas
sandraDaemon.java:353)
 at 
org.apache.cassandra.thrift.CassandraDaemon.main(CassandraDaemon.java:106
)
 org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.create(SnappyCompressor
.java:45)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.isAvailable(SnappyCompr
essor.java:55)
 at 
org.apache.cassandra.io.compress.SnappyCompressor.clinit(SnappyCompress
or.java:37)
 at 
org.apache.cassandra.config.CFMetaData.clinit(CFMetaData.java:76)
 at 

frequent node up/downs

2012-07-02 Thread feedly team
Hello,
   I recently set up a 2 node cassandra cluster on dedicated hardware. In
the logs there have been a lot of InetAddress xxx is now dead' or UP
messages. Comparing the log messages between the 2 nodes, they seem to
coincide with extremely long ParNew collections. I have seem some of up to
50 seconds. The installation is pretty vanilla, I didn't change any
settings and the machines don't seem particularly busy - cassandra is the
only thing running on the machine with an 8GB heap. The machine has 64GB of
RAM and CPU/IO usage looks pretty light. I do see a lot of 'Heap is xxx
full. You may need to reduce memtable and/or cache sizes' messages. Would
this help with the long ParNew collections? That message seems to be
triggered on a full collection.


RE: frequent node up/downs

2012-07-02 Thread Pieter Callewaert
Hi,

Had the same problem this morning, seems related to the leap second bug.
Rebooting the nodes fixed it for me, but there seems to be a fix also without 
rebooting the server.

Kind regards,
Pieter

From: feedly team [mailto:feedly...@gmail.com]
Sent: maandag 2 juli 2012 17:09
To: user@cassandra.apache.org
Subject: frequent node up/downs

Hello,
   I recently set up a 2 node cassandra cluster on dedicated hardware. In the 
logs there have been a lot of InetAddress xxx is now dead' or UP messages. 
Comparing the log messages between the 2 nodes, they seem to coincide with 
extremely long ParNew collections. I have seem some of up to 50 seconds. The 
installation is pretty vanilla, I didn't change any settings and the machines 
don't seem particularly busy - cassandra is the only thing running on the 
machine with an 8GB heap. The machine has 64GB of RAM and CPU/IO usage looks 
pretty light. I do see a lot of 'Heap is xxx full. You may need to reduce 
memtable and/or cache sizes' messages. Would this help with the long ParNew 
collections? That message seems to be triggered on a full collection.


Re: frequent node up/downs

2012-07-02 Thread Marcus Both
Yeah! Look that.
http://arstechnica.com/business/2012/07/one-day-later-the-leap-second-v-the-internet-scorecard/
I had the same problem. The solution was rebooting.

On Mon, 2 Jul 2012 11:08:57 -0400
feedly team feedly...@gmail.com wrote:

 Hello,
I recently set up a 2 node cassandra cluster on dedicated hardware. In
 the logs there have been a lot of InetAddress xxx is now dead' or UP
 messages. Comparing the log messages between the 2 nodes, they seem to
 coincide with extremely long ParNew collections. I have seem some of up to
 50 seconds. The installation is pretty vanilla, I didn't change any
 settings and the machines don't seem particularly busy - cassandra is the
 only thing running on the machine with an 8GB heap. The machine has 64GB of
 RAM and CPU/IO usage looks pretty light. I do see a lot of 'Heap is xxx
 full. You may need to reduce memtable and/or cache sizes' messages. Would
 this help with the long ParNew collections? That message seems to be
 triggered on a full collection.

-- 
Marcus Both 



Re: ClassCastException during Cassandra server startup

2012-07-02 Thread Yuki Morishita
Thierry,

Key cache files are stored inside your saved_caches_directory defined in 
cassandra.yaml, which has default value of /var/lib/cassandra/saved_caches. 


Yuki


On Monday, July 2, 2012 at 4:00 AM, Thierry Templier wrote:

 Hello Yuki,
 
 Could you give me hints about where to find these files. I have a look in the 
 installation folder of Cassandra and in the /var/lib/cassandra folder?
 
 Thanks very much for your help.
 Thierry
  That was bug in 1.1.1 and fixed in 
  https://issues.apache.org/jira/browse/CASSANDRA-4331. 
  Workaround is deleting the key cache files for your index CFs should fix 
  this.
  
  
  
  Yuki



bulk load glitch

2012-07-02 Thread Brian Jeltema
I'm attempting to perform a bulk load by calling the jmx:bulkLoad method on 
several nodes in parallel. In a Casssandra log
file I see a few occurrences of the following:

 INFO [GossipTasks:1] 2012-07-02 10:12:33,626 Gossiper.java (line 759) 
InetAddress /10.4.0.3 is now dead.
ERROR [GossipTasks:1] 2012-07-02 10:12:53,630 StreamOutSession.java (line 232) 
StreamOutSession /10.4.0.3 failed because {} died or was restarted/removed
 INFO [GossipStage:1] 2012-07-02 10:18:40,008 Gossiper.java (line 745) 
InetAddress /10.4.0.3 is now UP

I can't tell whether the bulk load process recovered from the transient dead 
node, or whether I need to start over.

Does anybody know?

Brian

This is Cassanrda 1.0.5, by the way.

faillout from AWS outage - HELP

2012-07-02 Thread Deno Vichas

all,

my 4 node cluster seems pretty screwed up after the AWS outage.  we 
found all our machines with their cpu stuck at 100%.  so i went to 
restart each cassandra node one by one.  i did node with token id 0 
first.  i came back but doesn't look like it doing anything.  once i 
thought it was up i went and restarted the next.  this one got stuck on 
the AMI init startup.  i cancelled it, rebooted againg and now it't suck 
with [INFO] 07/02/12-15:42:29 Received 0 of 1 responses from: 


suggestions on how to fix this?


thanks,
deno


Re: jscv CPU Consumption

2012-07-02 Thread Olivier Mallassi
Thanks for the answer. In fact this is maybe simpler.

We do a lot of inserts and a couple of reads.
- tpstats show us the pending read stage was growing.
- *It seems* there was an issue with one of our request. We have 4 indexed
columns (UT8Type)
First, one of the indexed value was ---. we replaced this value
by and empty string and it works better. (running the query in CLI)

Yet, we still observed pending Read Stage after 10 minutes of run (on a
couple of nodes).

Do you have any ideas, advices on how to investigate in more details?

Any help would be appreciate.


On Sun, Jul 1, 2012 at 8:56 AM, Jonathan Ellis jbel...@gmail.com wrote:

 Sounds like http://wiki.apache.org/cassandra/FAQ#ubuntu_ec2_hangs to me.

 On Fri, Jun 29, 2012 at 1:45 AM, Olivier Mallassi omalla...@octo.com
 wrote:
  Hi all
 
  We have a 12 servers clusters (8 cores by machines..).
  OS is Ubuntu 10.04.2.
 
  On one of the machine (only one) and without any load (no inserts, no
  reads), we have a huge CPU Load whereas there is no activities (no
  compaction in progress etc...)
  A top on the machine show us the process jscv is using all the available
  CPUs.
 
  Is that link to JNA? do you have any ideas?
 
  Cheers
 
  --
  
  Olivier Mallassi
  OCTO Technology
  
  50, Avenue des Champs-Elysées
  75008 Paris
 
  Mobile: (33) 6 28 70 26 61
  Tél: (33) 1 58 56 10 00
  Fax: (33) 1 58 56 10 01
 
  http://www.octo.com
  Octo Talks! http://blog.octo.com
 
 



 --
 Jonathan Ellis
 Project Chair, Apache Cassandra
 co-founder of DataStax, the source for professional Cassandra support
 http://www.datastax.com




-- 

Olivier Mallassi
OCTO Technology

50, Avenue des Champs-Elysées
75008 Paris

Mobile: (33) 6 28 70 26 61
Tél: (33) 1 58 56 10 00
Fax: (33) 1 58 56 10 01

http://www.octo.com
Octo Talks! http://blog.octo.com


Secondary Index, performance , data type

2012-07-02 Thread Olivier Mallassi
Hi all

We have 4 indexed columns; all configured in UT8Type even if one columns is
a date and the other an integer).

1/ the read query we run can have up to 4 criteria
*select my_cf where columnA = a and columnB = b and columnC = c and columnD
= d*

This query, is fast (500ms) up to 3 criterias but when we add the fourth
one, the exection time is 9,5s.
Do you have any ideas? is there any way to understand how cassandra
internally run the query (a kind of explain plan)?

2/ Is there any limitations on the number of criterias we can usually have?

3/ Even if we have different data type (date, string, int), we have all
stored them as UTF8Type. Could we expect performance improvements if we use
DateType, LongType?

Many thx for all your answers.

-- 

Olivier Mallassi
OCTO Technology

50, Avenue des Champs-Elysées
75008 Paris

Mobile: (33) 6 28 70 26 61
Tél: (33) 1 58 56 10 00
Fax: (33) 1 58 56 10 01

http://www.octo.com
Octo Talks! http://blog.octo.com


Re: faillout from AWS outage - HELP

2012-07-02 Thread Joost van de Wijgerd
Could be related to the leap second bug..

try this:

Bug: https://lkml.org/lkml/2012/6/30/122

Simple fix to reset the leap second flag: date; date `date
+%m%d%H%M%C%y.%S`; date;

mvg

Joost

On Mon, Jul 2, 2012 at 5:43 PM, Deno Vichas d...@syncopated.net wrote:
 all,

 my 4 node cluster seems pretty screwed up after the AWS outage.  we found
 all our machines with their cpu stuck at 100%.  so i went to restart each
 cassandra node one by one.  i did node with token id 0 first.  i came back
 but doesn't look like it doing anything.  once i thought it was up i went
 and restarted the next.  this one got stuck on the AMI init startup.  i
 cancelled it, rebooted againg and now it't suck with [INFO]
 07/02/12-15:42:29 Received 0 of 1 responses from: 

 suggestions on how to fix this?


 thanks,
 deno



-- 
Joost van de Wijgerd
Visseringstraat 21B
1051KH Amsterdam
+31624111401
joost.van.de.wijgerd@Skype
http://www.linkedin.com/in/jwijgerd


Re: faillout from AWS outage - HELP

2012-07-02 Thread Deno Vichas

the node that doesn't want to start just spit out -

   /EC2 is experiencing some issues and has not allocated all of the
   resources in under 10 minutes.
   Aborting the clustering of this reservation. Please try again.

   Please visit http://datastax.com/ami for this AMI's feature set.
   /




On 7/2/2012 8:43 AM, Deno Vichas wrote:

all,

my 4 node cluster seems pretty screwed up after the AWS outage. we 
found all our machines with their cpu stuck at 100%.  so i went to 
restart each cassandra node one by one.  i did node with token id 0 
first.  i came back but doesn't look like it doing anything. once i 
thought it was up i went and restarted the next.  this one got stuck 
on the AMI init startup.  i cancelled it, rebooted againg and now it't 
suck with [INFO] 07/02/12-15:42:29 Received 0 of 1 responses from: 


suggestions on how to fix this?


thanks,
deno





MultigetSliceQuery count 1 returning multiple results

2012-07-02 Thread Owen Davies
We are using hector.rb to do a MultigetSliceQuery over one row, with count 1. 
Most of the time it works, but occasionally we are getting back 2 columns. Has 
anyone seen this before?

Cassandra: 1.1.0-SNAPSHOT
Hector: 1.0.2
Hector.rb: 0.0.1

Gem Code:

https://github.com/jashmenn/hector.rb/blob/master/lib/hector/hector.rb#L126

Example Query:

client.get_range(column_family, row_id, row_id, :start = time, :finish = 
(time - 1.day), :count = 1, :reversed = true)
client.get_range(column_family, row_id, row_id, :start = time, :finish = 
(time + 1.day), :count = 1)

Thanks,

Owen

Re: Node won't start (faillout from AWS outage) - HELP

2012-07-02 Thread Robin Verlangen
Your missing the sliced_buffer_size_in_kb property in your cassandra.yaml
file.

2012/7/2 Deno Vichas d...@syncopated.net

  i'm seeing the following exception in my log now.  this is even after
 re-creating my config file.


  INFO [main] 2012-07-02 16:58:17,850 AbstractCassandraDaemon.java (line
 121) JVM vendor/version: Java HotSpot(TM) 64-Bit Server VM/1.6.0_26
  INFO [main] 2012-07-02 16:58:17,853 AbstractCassandraDaemon.java (line
 122) Heap size: 1936719872/1937768448
  INFO [main] 2012-07-02 16:58:17,854 AbstractCassandraDaemon.java (line
 123) Classpath:
 /usr/share/cassandra/lib/antlr-3.2.jar:/usr/share/cassandra/lib/avro-1.4.0-fixes.jar:/usr/share/cassandra/lib/avro-1.4.0-sources-fixes.jar:/usr/share/cassandra/lib/commons-cli-1.1.jar:/usr/share/cassandra/lib/commons-codec-1.2.jar:/usr/share/cassandra/lib/commons-lang-2.4.jar:/usr/share/cassandra/lib/compress-lzf-0.8.4.jar:/usr/share/cassandra/lib/concurrentlinkedhashmap-lru-1.3.jar:/usr/share/cassandra/lib/guava-r08.jar:/usr/share/cassandra/lib/high-scale-lib-1.1.2.jar:/usr/share/cassandra/lib/jackson-core-asl-1.9.2.jar:/usr/share/cassandra/lib/jackson-mapper-asl-1.9.2.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar:/usr/share/cassandra/lib/jline-0.9.94.jar:/usr/share/cassandra/lib/joda-time-1.6.2.jar:/usr/share/cassandra/lib/json-simple-1.1.jar:/usr/share/cassandra/lib/libthrift-0.7.0.jar:/usr/share/cassandra/lib/log4j-1.2.16.jar:/usr/share/cassandra/lib/metrics-core-2.0.3.jar:/usr/share/cassandra/lib/servlet-api-2.5-20081211.jar:/usr/share/cassandra/lib/slf4j-api-1.6.1.jar:/usr/share
 /cassandra/lib/slf4j-log4j12-1.6.1.jar:/usr/share/cassandra/lib/snakeyaml-1.6.jar:/usr/share/cassandra/lib/snappy-java-1.0.4.1.jar:/usr/share/cassandra/lib/snaptree-0.1.jar:/usr/share/cassandra/apache-cassandra-1.1.1.jar:/usr/share/cassandra/apache-cassandra-thrift-1.1.1.jar:/usr/share/cassandra/apache-cassandra.jar:/usr/share/cassandra/stress.jar:/usr/share/java/jna.jar:/etc/cassandra:/usr/share/java/commons-daemon.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar
  INFO [main] 2012-07-02 16:58:19,309 CLibrary.java (line 111) JNA mlockall
 successful
  INFO [main] 2012-07-02 16:58:19,319 DatabaseDescriptor.java (line 127)
 Loading settings from file:/etc/cassandra/cassandra.yaml
 ERROR [main] 2012-07-02 16:58:19,517 DatabaseDescriptor.java (line 481)
 Fatal configuration error error
 Can't construct a java object for 
 tag:yaml.org,2002:org.apache.cassandra.config.Config;
 exception=Cannot create property=sliced_buffer_size_in_kb for
 JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to find
 property 'sliced_buffer_size_in_kb' on class:
 org.apache.cassandra.config.Config
  in reader, line 4, column 1:
 cluster_name: 'StockTouchCluster'
 ^

 at
 org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:372)
 at
 org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:177)
 at
 org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:136)
 at
 org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:122)
 at org.yaml.snakeyaml.Loader.load(Loader.java:52)
 at org.yaml.snakeyaml.Yaml.load(Yaml.java:166)
 at
 org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:143)
 at
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
 at
 org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
 at
 org.apache.cassandra.service.AbstractCassandraDaemon.init(AbstractCassandraDaemon.java:254)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at
 org.apache.commons.daemon.support.DaemonLoader.load(DaemonLoader.java:160)
 Caused by: org.yaml.snakeyaml.error.YAMLException: Cannot create
 property=sliced_buffer_size_in_kb for
 JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to find
 property 'sliced_buffer_size_in_kb' on class:
 org.apache.cassandra.config.Config
 at
 org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.constructJavaBean2ndStep(Constructor.java:305)
 at
 org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:184)
 at
 org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:370)
 ... 14 more
 Caused by: org.yaml.snakeyaml.error.YAMLException: Unable to find property
 'sliced_buffer_size_in_kb' on class: org.apache.cassandra.config.Config
 at
 org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.getProperty(Constructor.java:342)
 at
 

Re: Node won't start (faillout from AWS outage) - HELP

2012-07-02 Thread Brian Jeltema
Just the opposite,  I think. The property value exists in the yaml file but 
does not have a corresponding definition in the Config class.
Typically caused by a version mismatch in my experience.

On Jul 2, 2012, at 1:20 PM, Robin Verlangen wrote:

 Your missing the sliced_buffer_size_in_kb property in your cassandra.yaml 
 file.
 
 2012/7/2 Deno Vichas d...@syncopated.net
 i'm seeing the following exception in my log now.  this is even after 
 re-creating my config file.
 
 
  INFO [main] 2012-07-02 16:58:17,850 AbstractCassandraDaemon.java (line 121) 
 JVM vendor/version: Java HotSpot(TM) 64-Bit Server VM/1.6.0_26
  INFO [main] 2012-07-02 16:58:17,853 AbstractCassandraDaemon.java (line 122) 
 Heap size: 1936719872/1937768448
  INFO [main] 2012-07-02 16:58:17,854 AbstractCassandraDaemon.java (line 123) 
 Classpath: 
 /usr/share/cassandra/lib/antlr-3.2.jar:/usr/share/cassandra/lib/avro-1.4.0-fixes.jar:/usr/share/cassandra/lib/avro-1.4.0-sources-fixes.jar:/usr/share/cassandra/lib/commons-cli-1.1.jar:/usr/share/cassandra/lib/commons-codec-1.2.jar:/usr/share/cassandra/lib/commons-lang-2.4.jar:/usr/share/cassandra/lib/compress-lzf-0.8.4.jar:/usr/share/cassandra/lib/concurrentlinkedhashmap-lru-1.3.jar:/usr/share/cassandra/lib/guava-r08.jar:/usr/share/cassandra/lib/high-scale-lib-1.1.2.jar:/usr/share/cassandra/lib/jackson-core-asl-1.9.2.jar:/usr/share/cassandra/lib/jackson-mapper-asl-1.9.2.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar:/usr/share/cassandra/lib/jline-0.9.94.jar:/usr/share/cassandra/lib/joda-time-1.6.2.jar:/usr/share/cassandra/lib/json-simple-1.1.jar:/usr/share/cassandra/lib/libthrift-0.7.0.jar:/usr/share/cassandra/lib/log4j-1.2.16.jar:/usr/share/cassandra/lib/metrics-core-2.0.3.jar:/usr/share/cassandra/lib/servlet-api-2.5-20081211.jar:/usr/share/cassandra/lib/slf4j-api-1.6.1.jar:/usr/share
  
 /cassandra/lib/slf4j-log4j12-1.6.1.jar:/usr/share/cassandra/lib/snakeyaml-1.6.jar:/usr/share/cassandra/lib/snappy-java-1.0.4.1.jar:/usr/share/cassandra/lib/snaptree-0.1.jar:/usr/share/cassandra/apache-cassandra-1.1.1.jar:/usr/share/cassandra/apache-cassandra-thrift-1.1.1.jar:/usr/share/cassandra/apache-cassandra.jar:/usr/share/cassandra/stress.jar:/usr/share/java/jna.jar:/etc/cassandra:/usr/share/java/commons-daemon.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar
  INFO [main] 2012-07-02 16:58:19,309 CLibrary.java (line 111) JNA mlockall 
 successful
  INFO [main] 2012-07-02 16:58:19,319 DatabaseDescriptor.java (line 127) 
 Loading settings from file:/etc/cassandra/cassandra.yaml
 ERROR [main] 2012-07-02 16:58:19,517 DatabaseDescriptor.java (line 481) Fatal 
 configuration error error
 Can't construct a java object for 
 tag:yaml.org,2002:org.apache.cassandra.config.Config; exception=Cannot create 
 property=sliced_buffer_size_in_kb for 
 JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to find property 
 'sliced_buffer_size_in_kb' on class: org.apache.cassandra.config.Config
  in reader, line 4, column 1:
 cluster_name: 'StockTouchCluster'
 ^
 
 at 
 org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:372)
 at 
 org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:177)
 at 
 org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:136)
 at 
 org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:122)
 at org.yaml.snakeyaml.Loader.load(Loader.java:52)
 at org.yaml.snakeyaml.Yaml.load(Yaml.java:166)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:143)
 at 
 org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
 at 
 org.apache.cassandra.service.AbstractCassandraDaemon.init(AbstractCassandraDaemon.java:254)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.commons.daemon.support.DaemonLoader.load(DaemonLoader.java:160)
 Caused by: org.yaml.snakeyaml.error.YAMLException: Cannot create 
 property=sliced_buffer_size_in_kb for 
 JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to find property 
 'sliced_buffer_size_in_kb' on class: org.apache.cassandra.config.Config
 at 
 org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.constructJavaBean2ndStep(Constructor.java:305)
 at 
 org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:184)
 at 
 org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:370)
 ... 14 more
 Caused by: 

Re: Node won't start (faillout from AWS outage) - HELP

2012-07-02 Thread Deno Vichas
no, it's there... after editing the first couple lines of the config 
file there spot where it complains changes.




On 7/2/2012 10:20 AM, Robin Verlangen wrote:
Your missing the sliced_buffer_size_in_kb property in your 
cassandra.yaml file.


2012/7/2 Deno Vichas d...@syncopated.net mailto:d...@syncopated.net

i'm seeing the following exception in my log now.  this is even
after re-creating my config file.


 INFO [main] 2012-07-02 16:58:17,850 AbstractCassandraDaemon.java
(line 121) JVM vendor/version: Java HotSpot(TM) 64-Bit Server
VM/1.6.0_26
 INFO [main] 2012-07-02 16:58:17,853 AbstractCassandraDaemon.java
(line 122) Heap size: 1936719872/1937768448
 INFO [main] 2012-07-02 16:58:17,854 AbstractCassandraDaemon.java
(line 123) Classpath:

/usr/share/cassandra/lib/antlr-3.2.jar:/usr/share/cassandra/lib/avro-1.4.0-fixes.jar:/usr/share/cassandra/lib/avro-1.4.0-sources-fixes.jar:/usr/share/cassandra/lib/commons-cli-1.1.jar:/usr/share/cassandra/lib/commons-codec-1.2.jar:/usr/share/cassandra/lib/commons-lang-2.4.jar:/usr/share/cassandra/lib/compress-lzf-0.8.4.jar:/usr/share/cassandra/lib/concurrentlinkedhashmap-lru-1.3.jar:/usr/share/cassandra/lib/guava-r08.jar:/usr/share/cassandra/lib/high-scale-lib-1.1.2.jar:/usr/share/cassandra/lib/jackson-core-asl-1.9.2.jar:/usr/share/cassandra/lib/jackson-mapper-asl-1.9.2.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar:/usr/share/cassandra/lib/jline-0.9.94.jar:/usr/share/cassandra/lib/joda-time-1.6.2.jar:/usr/share/cassandra/lib/json-simple-1.1.jar:/usr/share/cassandra/lib/libthrift-0.7.0.jar:/usr/share/cassandra/lib/log4j-1.2.16.jar:/usr/share/cassandra/lib/metrics-core-2.0.3.jar:/usr/share/cassandra/lib/servlet-api-2.5-20081211.jar:/usr/share/cassandra/lib/slf4j-api-1.6.1.jar:/usr/share

/cassandra/lib/slf4j-log4j12-1.6.1.jar:/usr/share/cassandra/lib/snakeyaml-1.6.jar:/usr/share/cassandra/lib/snappy-java-1.0.4.1.jar:/usr/share/cassandra/lib/snaptree-0.1.jar:/usr/share/cassandra/apache-cassandra-1.1.1.jar:/usr/share/cassandra/apache-cassandra-thrift-1.1.1.jar:/usr/share/cassandra/apache-cassandra.jar:/usr/share/cassandra/stress.jar:/usr/share/java/jna.jar:/etc/cassandra:/usr/share/java/commons-daemon.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar
 INFO [main] 2012-07-02 16:58:19,309 CLibrary.java (line 111) JNA
mlockall successful
 INFO [main] 2012-07-02 16:58:19,319 DatabaseDescriptor.java (line
127) Loading settings from file:/etc/cassandra/cassandra.yaml
ERROR [main] 2012-07-02 16:58:19,517 DatabaseDescriptor.java (line
481) Fatal configuration error error
Can't construct a java object for tag:yaml.org
http://yaml.org,2002:org.apache.cassandra.config.Config;
exception=Cannot create property=sliced_buffer_size_in_kb for
JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to
find property 'sliced_buffer_size_in_kb' on class:
org.apache.cassandra.config.Config
 in reader, line 4, column 1:
cluster_name: 'StockTouchCluster'
^

at

org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:372)
at

org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:177)
at

org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:136)
at

org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:122)
at org.yaml.snakeyaml.Loader.load(Loader.java:52)
at org.yaml.snakeyaml.Yaml.load(Yaml.java:166)
at

org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:143)
at

org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
at

org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
at

org.apache.cassandra.service.AbstractCassandraDaemon.init(AbstractCassandraDaemon.java:254)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at

sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at

sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at
org.apache.commons.daemon.support.DaemonLoader.load(DaemonLoader.java:160)
Caused by: org.yaml.snakeyaml.error.YAMLException: Cannot create
property=sliced_buffer_size_in_kb for
JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to
find property 'sliced_buffer_size_in_kb' on class:
org.apache.cassandra.config.Config
at

org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.constructJavaBean2ndStep(Constructor.java:305)
at

org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:184)
at


Re: cassandra on re-Start

2012-07-02 Thread Rob Coli
On Mon, Jul 2, 2012 at 5:43 AM, puneet loya puneetl...@gmail.com wrote:
 When I restarted the system , it is showing the keyspace does not exist.

 Not even letting me to create the keyspace with the same name again.

Paste the error you get.

=Rob

-- 
=Robert Coli
AIMGTALK - rc...@palominodb.com
YAHOO - rcoli.palominob
SKYPE - rcoli_palominodb


Re: Node won't start (faillout from AWS outage) - HELP

2012-07-02 Thread Deno Vichas
should be the same version. i did notice some extra files in my etc 
dir.  there's a few dpkg-new files.



ubuntu@ip-10-40-207-151:~/datastax_ami$ ll /etc/cassandra/
total 100
drwxr-xr-x  2 root root  4096 2012-07-02 18:13 ./
drwxr-xr-x 99 root root  4096 2012-07-02 16:31 ../
-rw-r--r--  1 root root  7649 2012-02-21 07:10 cassandra-env.sh
-rw-r--r--  1 root root  7690 2012-06-01 17:54 cassandra-env.sh.dpkg-new
-rw-r--r--  1 root root 18424 2012-07-02 18:11 cassandra.yaml
-rw-r--r--  1 root root 19234 2012-07-02 16:54 cassandra.yaml.bak
-rw-r--r--  1 root root 24911 2012-07-02 17:11 cassandra.yaml.dpkg-new
-rw-r--r--  1 root root  1849 2012-07-02 17:12 log4j-server.properties
-rw-r--r--  1 root root  1850 2012-06-01 17:54 
log4j-server.properties.dpkg-new*

*

On 7/2/2012 10:25 AM, Brian Jeltema wrote:
Just the opposite,  I think. The property value exists in the yaml 
file but does not have a corresponding definition in the Config class.

Typically caused by a version mismatch in my experience.

On Jul 2, 2012, at 1:20 PM, Robin Verlangen wrote:

Your missing the sliced_buffer_size_in_kb property in your 
cassandra.yaml file.


2012/7/2 Deno Vichas d...@syncopated.net mailto:d...@syncopated.net

i'm seeing the following exception in my log now.  this is even
after re-creating my config file.


 INFO [main] 2012-07-02 16:58:17,850 AbstractCassandraDaemon.java
(line 121) JVM vendor/version: Java HotSpot(TM) 64-Bit Server
VM/1.6.0_26
 INFO [main] 2012-07-02 16:58:17,853 AbstractCassandraDaemon.java
(line 122) Heap size: 1936719872/1937768448
 INFO [main] 2012-07-02 16:58:17,854 AbstractCassandraDaemon.java
(line 123) Classpath:

/usr/share/cassandra/lib/antlr-3.2.jar:/usr/share/cassandra/lib/avro-1.4.0-fixes.jar:/usr/share/cassandra/lib/avro-1.4.0-sources-fixes.jar:/usr/share/cassandra/lib/commons-cli-1.1.jar:/usr/share/cassandra/lib/commons-codec-1.2.jar:/usr/share/cassandra/lib/commons-lang-2.4.jar:/usr/share/cassandra/lib/compress-lzf-0.8.4.jar:/usr/share/cassandra/lib/concurrentlinkedhashmap-lru-1.3.jar:/usr/share/cassandra/lib/guava-r08.jar:/usr/share/cassandra/lib/high-scale-lib-1.1.2.jar:/usr/share/cassandra/lib/jackson-core-asl-1.9.2.jar:/usr/share/cassandra/lib/jackson-mapper-asl-1.9.2.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar:/usr/share/cassandra/lib/jline-0.9.94.jar:/usr/share/cassandra/lib/joda-time-1.6.2.jar:/usr/share/cassandra/lib/json-simple-1.1.jar:/usr/share/cassandra/lib/libthrift-0.7.0.jar:/usr/share/cassandra/lib/log4j-1.2.16.jar:/usr/share/cassandra/lib/metrics-core-2.0.3.jar:/usr/share/cassandra/lib/servlet-api-2.5-20081211.jar:/usr/share/cassandra/lib/slf4j-api-1.6.1.jar:/usr/share

/cassandra/lib/slf4j-log4j12-1.6.1.jar:/usr/share/cassandra/lib/snakeyaml-1.6.jar:/usr/share/cassandra/lib/snappy-java-1.0.4.1.jar:/usr/share/cassandra/lib/snaptree-0.1.jar:/usr/share/cassandra/apache-cassandra-1.1.1.jar:/usr/share/cassandra/apache-cassandra-thrift-1.1.1.jar:/usr/share/cassandra/apache-cassandra.jar:/usr/share/cassandra/stress.jar:/usr/share/java/jna.jar:/etc/cassandra:/usr/share/java/commons-daemon.jar:/usr/share/cassandra/lib/jamm-0.2.5.jar
 INFO [main] 2012-07-02 16:58:19,309 CLibrary.java (line 111) JNA
mlockall successful
 INFO [main] 2012-07-02 16:58:19,319 DatabaseDescriptor.java
(line 127) Loading settings from file:/etc/cassandra/cassandra.yaml
ERROR [main] 2012-07-02 16:58:19,517 DatabaseDescriptor.java
(line 481) Fatal configuration error error
Can't construct a java object for tag:yaml.org
http://yaml.org/,2002:org.apache.cassandra.config.Config;
exception=Cannot create property=sliced_buffer_size_in_kb for
JavaBean=org.apache.cassandra.config.Config@371bed71; Unable to
find property 'sliced_buffer_size_in_kb' on class:
org.apache.cassandra.config.Config
 in reader, line 4, column 1:
cluster_name: 'StockTouchCluster'
^

at

org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:372)
at

org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:177)
at

org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:136)
at

org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:122)
at org.yaml.snakeyaml.Loader.load(Loader.java:52)
at org.yaml.snakeyaml.Yaml.load(Yaml.java:166)
at

org.apache.cassandra.config.DatabaseDescriptor.loadYaml(DatabaseDescriptor.java:143)
at

org.apache.cassandra.config.DatabaseDescriptor.clinit(DatabaseDescriptor.java:118)
at

org.apache.cassandra.service.AbstractCassandraDaemon.setup(AbstractCassandraDaemon.java:126)
at

org.apache.cassandra.service.AbstractCassandraDaemon.init(AbstractCassandraDaemon.java:254)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
   

datastax aws ami

2012-07-02 Thread Deno Vichas

all,

i've got a datastax 2.1 ami instance that's screwed up.  for some reason 
it won't read the config file.  what's the recommended way to replace 
this node with a new one?  it doesn't seem like you can use the ami to 
bring up single nodes as it want to do whole clusters.



thanks,
deno




[RELEASE] Apache Cassandra 1.1.2 released

2012-07-02 Thread Sylvain Lebresne
The Cassandra team is pleased to announce the release of Apache Cassandra
version 1.1.2.

Cassandra is a highly scalable second-generation distributed database,
bringing together Dynamo's fully distributed design and Bigtable's
ColumnFamily-based data model. You can read more here:

 http://cassandra.apache.org/

Downloads of source and binary distributions are listed in our download
section:

 http://cassandra.apache.org/download/

This version is a maintenance/bug fix release[1] on the 1.1 series. As always,
please pay attention to the release notes[2] and Let us know[3] if you were to
encounter any problem.

Enjoy!

[1]: http://goo.gl/kt5e9 (CHANGES.txt)
[2]: http://goo.gl/nfE2w (NEWS.txt)
[3]: https://issues.apache.org/jira/browse/CASSANDRA


Re: datastax aws ami

2012-07-02 Thread Deno Vichas

is the 2.1 image still around?

On 7/2/2012 11:24 AM, Deno Vichas wrote:

all,

i've got a datastax 2.1 ami instance that's screwed up.  for some 
reason it won't read the config file.  what's the recommended way to 
replace this node with a new one?  it doesn't seem like you can use 
the ami to bring up single nodes as it want to do whole clusters.



thanks,
deno







Thrift version and OOM errors

2012-07-02 Thread Vasileios Vlachos
Hello All,

We are using Cassandra 1.0.7 on AWS on mediums (that is 3.8G RAM, 1 Core),
running Ubuntu 12.04. We have three nodes in the cluster and we hit only
one node from our application. Thrift version is 0.6.1 (we changed from 0.8
because we thought there was a compatibility problem between thrift and
Cassandra ('old client' according to the output.log). We are still not sure
what version of thrift to use with Cassandra 1.0.7 (we are still getting
the same message regarding the 'old client'). I would appreciate any help
on that please.

Below, I am sharing the errors we are getting from the output.log file.
First three errors are not responsible for the crash, only the OOM error
is, but something seems to be really wrong there...

Error #1

ERROR 14:00:12,057 Thrift error occurred during processing of message.
org.apache.thrift.TException: Message length exceeded: 1970238464
at
org.apache.thrift.protocol.TBinaryProtocol.checkReadLength(TBinaryProtocol.java:393)
at
org.apache.thrift.protocol.TBinaryProtocol.readBinary(TBinaryProtocol.java:363)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:102)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:112)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:112)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:112)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:121)
at org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:60)
at org.apache.cassandra.thrift.Mutation.read(Mutation.java:355)
at
org.apache.cassandra.thrift.Cassandra$batch_mutate_args.read(Cassandra.java:18966)
at
org.apache.cassandra.thrift.Cassandra$Processor$batch_mutate.process(Cassandra.java:3441)
at
org.apache.cassandra.thrift.Cassandra$Processor.process(Cassandra.java:2889)
at
org.apache.cassandra.thrift.CustomTThreadPoolServer$WorkerProcess.run(CustomTThreadPoolServer.java:187)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

Error #2

ERROR 14:03:48,004 Error occurred during processing of message.
java.lang.StringIndexOutOfBoundsException: String index out of range:
-2147418111
at java.lang.String.checkBounds(String.java:397)
at java.lang.String.init(String.java:442)
at
org.apache.thrift.protocol.TBinaryProtocol.readString(TBinaryProtocol.java:339)
at
org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:210)
at
org.apache.cassandra.thrift.Cassandra$Processor.process(Cassandra.java:2877)
at
org.apache.cassandra.thrift.CustomTThreadPoolServer$WorkerProcess.run(CustomTThreadPoolServer.java:187)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

Error #3

ERROR 14:07:24,415 Thrift error occurred during processing of message.
org.apache.thrift.protocol.TProtocolException: Missing version in
readMessageBegin, old client?
at
org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:213)
at
org.apache.cassandra.thrift.Cassandra$Processor.process(Cassandra.java:2877)
at
org.apache.cassandra.thrift.CustomTThreadPoolServer$WorkerProcess.run(CustomTThreadPoolServer.java:187)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

Error #4

ERROR 16:07:10,168 Thrift error occurred during processing of message.
org.apache.thrift.TException: Message length exceeded: 218104076
at
org.apache.thrift.protocol.TBinaryProtocol.checkReadLength(TBinaryProtocol.java:393)
at
org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:352)
at
org.apache.thrift.protocol.TBinaryProtocol.readString(TBinaryProtocol.java:347)
at
org.apache.cassandra.thrift.Cassandra$batch_mutate_args.read(Cassandra.java:18958)
at
org.apache.cassandra.thrift.Cassandra$Processor$batch_mutate.process(Cassandra.java:3441)
at
org.apache.cassandra.thrift.Cassandra$Processor.process(Cassandra.java:2889)
at
org.apache.cassandra.thrift.CustomTThreadPoolServer$WorkerProcess.run(CustomTThreadPoolServer.java:187)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
java.lang.OutOfMemoryError: Java heap space
Dumping heap to /var/lib/cassandra/java_1341224307.hprof ...
INFO 16:07:18,882 GC for Copy: 886 ms for 1 collections, 2242700896 used;
max is 2670985216
Java HotSpot(TM) 64-Bit Server VM warning: record is too large
Heap dump file created [4429997807 bytes in 95.755 secs]
INFO 16:08:54,749 GC for ConcurrentMarkSweep: 1157 ms for 4 collections,

Re: Cassandra and massive TTL expirations cause HEAP issue

2012-07-02 Thread aaron morton
  After 10 days my cluster crashes due to a java.lang.OutOfMemoryError during 
 compaction of the big column family that contains roughly 95% of the data. 

Does this column family have very wide rows ? 

  simply some tweaks I need to make in the yaml file.  I have tried:
The main things that reduce the impact compaction has on memory are:

in_memory_compaction_limit_in_mb
concurrent_compactors

Of the top of my head I cannot think of any shortcuts taken by compaction 
if/when all data in an SSTable is past TTL. I think there was talk of something 
like that though. 

Hope that helps.

-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 27/06/2012, at 2:38 AM, Nils Pommerien wrote:

 Hello,
 I am evaluating Cassandra in a log retrieval application.  My ring conists 
 of3 m2.xlarge instances (17.1 GB memory, 6.5 ECU (2 virtual cores with 3.25 
 EC2 Compute Units each), 420 GB of local instance storage, 64-bit platform) 
 and I am writing at roughly 220 writes/sec.  Per day I am adding roughly 60GB 
 of data.  All of this sounds simple and easy and all three nodes are humming 
 along with basically no load.  
 
 The issue is that I am writing all my data with a TTL of 10 days.  After 10 
 days my cluster crashes due to a java.lang.OutOfMemoryError during compaction 
 of the big column family that contains roughly 95% of the data.  So basically 
 after 10 days my data set is 600GB and after 10 days Cassandra would have to 
 tombstone and purge 60GB of data at the same rate of roughly 220 
 deletes/second.  I am not sure if Cassandra should be able to do it, whether 
 I should take a partitioning approach (one CF per day), or if there is simply 
 some tweaks I need to make in the yaml file.  I have tried:
 Decrease flush-largest-memtables-at to .4 
 reduce_cache_sizes_at and reduce_cache_capacity_to set to 1
 Now, the issue remains the same:
 
 WARN [ScheduledTasks:1] 2012-06-11 19:39:42,017 GCInspector.java (line 145) 
 Heap is 0.9920103380107628 full.  You may need to reduce memtable and/or 
 cache sizes.  Cassandra will now flush up to the two largest memtables to 
 free up memory.  Adjust flush_largest_memtables_at threshold in 
 cassandra.yaml if you don't want Cassandra to do this automatically.
 
 Eventually it will just die with this message.  This affects all nodes in the 
 cluster, not just one. 
  
 Dump file is incomplete: file size limit
 ERROR 19:39:39,695 Exception in thread Thread[ReadStage:134,5,main]
 java.lang.OutOfMemoryError: Java heap space
 ERROR 19:39:39,724 Exception in thread Thread[MutationStage:57,5,main]
 java.lang.OutOfMemoryError: Java heap space
   at 
 org.apache.cassandra.utils.FBUtilities.hashToBigInteger(FBUtilities.java:213)
   at 
 org.apache.cassandra.dht.RandomPartitioner.getToken(RandomPartitioner.java:154)
   at 
 org.apache.cassandra.dht.RandomPartitioner.decorateKey(RandomPartitioner.java:47)
   at org.apache.cassandra.db.RowPosition.forKey(RowPosition.java:54)
  
 Any help is highly appreciated.  It would be cool to tweak it in a way that I 
 can have a moving window of 10 days in Cassandra while dropping the old data… 
 Or, if there is any other recommended way to deal with such sliding time 
 windows I am open for ideas.
 
 Thank you for your help!  



Re: cassandra 1.0.9 error - Read an invalid frame size of 0

2012-07-02 Thread aaron morton
Gurpreet ,
If you can reproduce this please create a ticket on 
https://issues.apache.org/jira/browse/CASSANDRA

Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 19/06/2012, at 6:38 AM, Gurpreet Singh wrote:

 
 I found a fix for this one, rather a workaround.
 
 I changed the rpc_server_type in cassandra.yaml, from hsha to sync, and the 
 error went away. I guess, there is some issue with the thrift nonblocking 
 server.
 
 Thanks
 Gurpreet
 
 On Wed, May 16, 2012 at 7:04 PM, Gurpreet Singh gurpreet.si...@gmail.com 
 wrote:
 Thanks Aaron. will do!
 
 
 On Mon, May 14, 2012 at 1:14 PM, aaron morton aa...@thelastpickle.com wrote:
 Are you using framed transport on the client side ? 
 
 Try the Hector user list for hector specific help 
 https://groups.google.com/forum/?fromgroups#!searchin/hector-users
 
 Cheers
 
 -
 Aaron Morton
 Freelance Developer
 @aaronmorton
 http://www.thelastpickle.com
 
 On 12/05/2012, at 5:44 AM, Gurpreet Singh wrote:
 
 This is hampering our testing of cassandra a lot, and our move to cassandra 
 1.0.9.
 Has anyone seen this before? Should I be trying a different version of 
 cassandra?
 
 /G
 
 On Thu, May 10, 2012 at 11:29 PM, Gurpreet Singh gurpreet.si...@gmail.com 
 wrote:
 Hi,
 i have created 1 node cluster of cassandra 1.0.9. I am setting this up for 
 testing reads/writes.
 
 I am seeing the following error in the server system.log
 
 ERROR [Selector-Thread-7] 2012-05-10 22:44:02,607 TNonblockingServer.java 
 (line 467) Read an invalid frame size of 0. Are you using TFramedTransport 
 on the client side?
 
 Initially i was using a old hector 0.7.x, but even after switching to hector 
 1.0-5 and thrift version 0.6.1, i still see this error. 
 I am using 20 threads writing/reading from cassandra. The max write batch 
 size is 10 with payload size constant per key to be 600 bytes.
 
 On the client side, i see Hector exceptions happenning coinciding with these 
 messages on the server. 
 
 Any ideas why these errors are happenning?
 
 Thanks
 Gurpreet
 
 
 
 
 



Re: repair never finishing 1.0.7

2012-07-02 Thread aaron morton
The nodes in DC1 need to be able to reach the nodes in DC2 on the public 
(NAT'd) IP.  

Others may be able to provide some more details . 

Cheers

-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 27/06/2012, at 9:51 PM, Andras Szerdahelyi wrote:

 Aaron,
 
 The broadcast_address allows a node to broadcast an address that is 
 different to the ones it's bound to on the local interfaces 
 https://github.com/apache/cassandra/blob/trunk/conf/cassandra.yaml#L270
 
 
 Yes and thats not where the problem is IMO.. If you broadcast your translated 
 address ( say 1.2.3.4, a public ip ) , nodes outside your VPN'd network will 
 have no problems connecting as long as they can route to this address ( which 
 they should ), but any other nodes on the local net ( e.g. 10.0.1.2 ) won't 
 be able to connect/route to their neighbor who's telling them to open the 
 return socket to 1.2.3.4
 
 Am i getting this right? At least this is what i have experienced not so long 
 ago:
 
 DC1 nodes
 a) 10.0.1.1 translated to 1.2.3.4 on NAT
 b) 10.0.1.2 translated to 1.2.3.5 on NAT 
 
 DC2 nodes
 a) 10.0.2.1 translated to 1.2.4.4 on NAT
 b) 10.0.2.2 translated to 1.2.4.5 on NAT
 
 Let's assume DC2 nodes' broadcast_addresses are their public addresses.
 
 if, DC1:a and DC1:b broadcast their public address, 1.2.3.4 and 1.2.3.5, they 
 are advertising an address that is not routable on their network ( loopback ) 
 but DC2:a and DC2:b can connect/route to them just fine. Nodetool ring on any 
 DC1 node says the others in DC1 are down, everything else is up . Nodetool 
 ring on any DC2 node says everything is up.
 
 if DC1:a and DC1:b broadcast their private address, they can connect to each 
 other fine, but  DC2:a and DC2:b will have no chance to route to them. 
 Nodetool ring on any DC1 node says everything is up. Nodetool ring on any DC2 
 node says DC1 nodes are down.
 
 regards,
 Andras
 
 
 
 
 On 27 Jun 2012, at 11:29, aaron morton wrote:
 
 Setting up a Cassandra ring across NAT ( without a VPN ) is impossible in 
 my experience. 
 The broadcast_address allows a node to broadcast an address that is 
 different to the ones it's bound to on the local interfaces 
 https://github.com/apache/cassandra/blob/trunk/conf/cassandra.yaml#L270
 
  1) How can I make sure that the JIRA issue above is my real problem? (I see 
 no errors or warns in the logs; no other activity)
 
 
 If the errors are not there it is not your problem. 
 
 - a full cluster restart allows the first attempted repair to complete 
 (haven't tested yet; this is not practical even if it works)
 Rolling restart of the nodes involved in the repair is sufficient. 
 
 Double checking the networking and check the logs on both sides of the 
 transfer for errors or warnings. The code around streaming is better at 
 failing loudly now days. 
 
 If you dont see anything set DEBUG logging on 
 org.apache.cassandra.streaming.FileStreamTask. That will let you know if 
 things start and progress. 
 
 Hope that helps. 
 
 
 -
 Aaron Morton
 Freelance Developer
 @aaronmorton
 http://www.thelastpickle.com
 
 On 26/06/2012, at 6:16 PM, Alexandru Sicoe wrote:
 
 Hi Andras,
 
 I am not using a VPN. The system has been running successfully in this 
 configuration for a couple of weeks until I noticed the repair is not 
 working.
 
 What happens is that I configure the IP Tables of the machine on each 
 Cassandra node to forward packets that are sent to any of the IPs in the 
 other DC (on ports 7000, 9160 and 7199)  to be sent to the gateway IP. The 
 gateway does the NAT sending the packets on the other side to the real 
 destination IP, having replaced the source IP with the initial sender's IP 
 (at least in my understanding of it). 
 
 What might be the problem given the configuration? How to fix this?
 
 Cheers,
 Alex
 
 On Mon, Jun 25, 2012 at 12:47 PM, Andras Szerdahelyi 
 andras.szerdahe...@ignitionone.com wrote:
 
  The DCs are communicating over a gateway where I do NAT for ports 7000, 
 9160 and 7199.
 
 
 Ah, that sounds familiar. You don't mention if you are VPN'd or not. I'll 
 assume you are not.
 
 So, your nodes are behind network address translation - is that to say they 
 advertise ( broadcast ) their internal or translated/forwarded IP to each 
 other? Setting up a Cassandra ring across NAT ( without a VPN ) is 
 impossible in my experience. Either the nodes on your local network won't 
 be able to communicate with each other, because they broadcast their 
 translated ( public ) address which is normally ( router configuration ) 
 not routable from within the local network, or the nodes broadcast their 
 internal IP, in which case the outside nodes are helpless in trying to 
 connect to a local net. On DC2 nodes/the node you issue the repair on, 
 check for any sockets being opened to the internal addresses of the nodes 
 in DC1.
 
 
 regards,
 Andras
 
 
 
 On 25 Jun 2012, at 11:57, Alexandru Sicoe 

Re: CQL / ASSUME for keys

2012-07-02 Thread aaron morton
 I'm trying to do the following : update keyspace.CF set '2' = '2' + 12
 WHERE KEY = 'mykey';
CQL does not support expressions in the SET clause. 

http://www.datastax.com/docs/1.0/references/cql/UPDATE

Cheers


-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 28/06/2012, at 2:08 AM, Alain RODRIGUEZ wrote:

 Hi,
 
 I'm trying to do the following : update keyspace.CF set '2' = '2' + 12
 WHERE KEY = 'mykey';
 And got this answer: Bad Request: cannot parse 'mykey' as
 hex bytes
 
 Using this doesn't help: assume keyspace.CF(KEY) VALUES ARE
 text; (Found here
 http://www.datastax.com/docs/1.0/references/cql/ASSUME and I'm using
 C* 1.0.9)
 
 Show schema in cli gives :
 
 create column family CF
  with column_type = 'Standard'
  and comparator = 'UTF8Type'
  and default_validation_class = 'CounterColumnType'
  and key_validation_class = 'BytesType'
  and rows_cached = 0.0
  and row_cache_save_period = 0
  and row_cache_keys_to_save = 2147483647
  and keys_cached = 2000.0
  and key_cache_save_period = 14400
  and read_repair_chance = 1.0
  and gc_grace = 864000
  and min_compaction_threshold = 4
  and max_compaction_threshold = 32
  and replicate_on_write = true
  and row_cache_provider = 'ConcurrentLinkedHashCacheProvider'
  and compaction_strategy =
 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy';
 
 What would be the consequences of changing the key_validation_class
 from 'BytesType' to 'UTF8Type'  (in production)?
 
 Shouldn't my assume command allow me to update my data even if I don't
 give the key as Bytes ?
 
 Alain



Re: Enable CQL3 from Astyanax

2012-07-02 Thread aaron morton
Thanks for contributing. 

I'm behind the curve on CQL 3, but here is a post about some of the changes 
http://www.datastax.com/dev/blog/whats-new-in-cql-3-0

Cheers


-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 28/06/2012, at 2:30 AM, Thierry Templier wrote:

 Hello Aaron,
 
 I created an issue on the Astyanax github for this problem. I added a fix to 
 support CQL3 in the tool.
 See the link https://github.com/Netflix/astyanax/issues/75.
 
 Thierry
 
 Had a quick look, the current master does not appear to support it. 
 
 Cheers



Re: Node crashing during read repair

2012-07-02 Thread aaron morton
Is this still an issue ? 

It looks like something shut down the messaging service. Was there anything 
else in the logs ? 

Cheers

-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 28/06/2012, at 3:49 AM, Robin Verlangen wrote:

 Hi there,
 
 Today I found one node (running 1.1.1 in a 3 node cluster) being dead for the 
 third time this week, it died with the following message:
 
 ERROR [ReadRepairStage:3] 2012-06-27 14:28:30,929 
 AbstractCassandraDaemon.java (line 134) Exception in thread 
 Thread[ReadRepairStage:3,5,main]
 java.util.concurrent.RejectedExecutionException: ThreadPoolExecutor has shut 
 down
 at 
 org.apache.cassandra.concurrent.DebuggableThreadPoolExecutor$1.rejectedExecution(DebuggableThreadPoolExecutor.java:60)
 at java.util.concurrent.ThreadPoolExecutor.reject(Unknown Source)
 at java.util.concurrent.ThreadPoolExecutor.execute(Unknown Source)
 at 
 org.apache.cassandra.net.MessagingService.receive(MessagingService.java:566)
 at 
 org.apache.cassandra.net.MessagingService.sendOneWay(MessagingService.java:439)
 at 
 org.apache.cassandra.net.MessagingService.sendRR(MessagingService.java:391)
 at 
 org.apache.cassandra.net.MessagingService.sendRR(MessagingService.java:372)
 at 
 org.apache.cassandra.net.MessagingService.sendRR(MessagingService.java:460)
 at 
 org.apache.cassandra.service.RowRepairResolver.scheduleRepairs(RowRepairResolver.java:136)
 at 
 org.apache.cassandra.service.RowRepairResolver.resolve(RowRepairResolver.java:94)
 at 
 org.apache.cassandra.service.AsyncRepairCallback$1.runMayThrow(AsyncRepairCallback.java:54)
 at 
 org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30)
 at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(Unknown 
 Source)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
 at java.lang.Thread.run(Unknown Source)
 
 Is this a common bug in 1.1.1, or did I get a race condition?
 
 -- 
 With kind regards,
 
 Robin Verlangen
 Software engineer
 
 W http://www.robinverlangen.nl
 E ro...@us2.nl
 
 Disclaimer: The information contained in this message and attachments is 
 intended solely for the attention and use of the named addressee and may be 
 confidential. If you are not the intended recipient, you are reminded that 
 the information remains the property of the sender. You must not use, 
 disclose, distribute, copy, print or rely on this e-mail. If you have 
 received this message in error, please contact the sender immediately and 
 irrevocably delete this message and any copies.
 



Re: bulk load glitch

2012-07-02 Thread Brandon Williams
On Mon, Jul 2, 2012 at 10:35 AM, Brian Jeltema
brian.jelt...@digitalenvoy.net wrote:
 I can't tell whether the bulk load process recovered from the transient dead 
 node, or whether I need to start over.

 Does anybody know?

You need to start over if the failure detector tripped, but it will
retry a few times for regular network errors.

-Brandon


Re: frequent node up/downs

2012-07-02 Thread feedly team
Yeah I noticed the leap second problem and ran the suggested fix, but I
have been facing these problems before Saturday and still see the
occasional failures after running the fix.

Thanks.

On Mon, Jul 2, 2012 at 11:17 AM, Marcus Both mb...@terra.com.br wrote:

 Yeah! Look that.

 http://arstechnica.com/business/2012/07/one-day-later-the-leap-second-v-the-internet-scorecard/
 I had the same problem. The solution was rebooting.

 On Mon, 2 Jul 2012 11:08:57 -0400
 feedly team feedly...@gmail.com wrote:

  Hello,
 I recently set up a 2 node cassandra cluster on dedicated hardware. In
  the logs there have been a lot of InetAddress xxx is now dead' or UP
  messages. Comparing the log messages between the 2 nodes, they seem to
  coincide with extremely long ParNew collections. I have seem some of up
 to
  50 seconds. The installation is pretty vanilla, I didn't change any
  settings and the machines don't seem particularly busy - cassandra is the
  only thing running on the machine with an 8GB heap. The machine has 64GB
 of
  RAM and CPU/IO usage looks pretty light. I do see a lot of 'Heap is xxx
  full. You may need to reduce memtable and/or cache sizes' messages. Would
  this help with the long ParNew collections? That message seems to be
  triggered on a full collection.

 --
 Marcus Both




Re: faillout from AWS outage - HELP

2012-07-02 Thread Deno Vichas
yes, you where correct.  resetting the flag and bouncing cassandra fixed 
it.  thx

On 7/2/2012 8:59 AM, Joost van de Wijgerd wrote:

Could be related to the leap second bug..

try this:

Bug: https://lkml.org/lkml/2012/6/30/122

Simple fix to reset the leap second flag: date; date `date
+%m%d%H%M%C%y.%S`; date;

mvg

Joost

On Mon, Jul 2, 2012 at 5:43 PM, Deno Vichas d...@syncopated.net wrote:

all,

my 4 node cluster seems pretty screwed up after the AWS outage.  we found
all our machines with their cpu stuck at 100%.  so i went to restart each
cassandra node one by one.  i did node with token id 0 first.  i came back
but doesn't look like it doing anything.  once i thought it was up i went
and restarted the next.  this one got stuck on the AMI init startup.  i
cancelled it, rebooted againg and now it't suck with [INFO]
07/02/12-15:42:29 Received 0 of 1 responses from: 

suggestions on how to fix this?


thanks,
deno








Re: hector timeouts

2012-07-02 Thread Deno Vichas
is anybody using kafka?  what other options is there?  currently i need 
to do around 50,000 (is that a lot?) a minute.



On 7/1/2012 11:39 AM, aaron morton wrote:
Using Cassandra as a queue is generally thought of as a bas idea, 
owing to the high delete workload. Levelled compaction handles it 
better but it is still no the best approach.


Depending on your needs consider running 
http://incubator.apache.org/kafka/


could you share some details on this?  we're using hector and we see 
random timeout warns in the logs and not sure how to address them.
First determine if they are server side or client side timeouts. Then 
determine what the query was.


Cheers

-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 29/06/2012, at 7:02 AM, Deno Vichas wrote:


On 6/28/2012 9:37 AM, David Leimbach wrote:


That coupled with Hector timeout issues became a real problem for us.


could you share some details on this?  we're using hector and we see 
random timeout warns in the logs and not sure how to address them.



thanks,
deno







Re: hector timeouts

2012-07-02 Thread Joe Stein
lots of folks use Apache Kafka, check out
https://cwiki.apache.org/confluence/display/KAFKA/Powered+By just to name a
few

you can read about the performance for yourself
http://incubator.apache.org/kafka/performance.html

@ http://www.medialets.com we use Kafka upstream of Cassandra acting like a
queue so our workers can do their business logic prior to storing their
results in Cassandra, Hadoop  MySQL

this decouples our backend analytics from our forwarding facing system
keeping our forward facing system (ad serving to mobile devices) as fast as
possible and our backend results near realtime (seconds from data coming in)

here are some papers and presentations
https://cwiki.apache.org/confluence/display/KAFKA/Kafka+papers+and+presentations

On Mon, Jul 2, 2012 at 10:09 PM, Deno Vichas d...@syncopated.net wrote:

  is anybody using kafka?  what other options is there?  currently i need
 to do around 50,000 (is that a lot?) a minute.


 On 7/1/2012 11:39 AM, aaron morton wrote:

 Using Cassandra as a queue is generally thought of as a bas idea, owing to
 the high delete workload. Levelled compaction handles it better but it is
 still no the best approach.

  Depending on your needs consider running
 http://incubator.apache.org/kafka/

   could you share some details on this?  we're using hector and we see
 random timeout warns in the logs and not sure how to address them.

 First determine if they are server side or client side timeouts. Then
 determine what the query was.

  Cheers

 -
 Aaron Morton
 Freelance Developer
 @aaronmorton
 http://www.thelastpickle.com

  On 29/06/2012, at 7:02 AM, Deno Vichas wrote:

  On 6/28/2012 9:37 AM, David Leimbach wrote:


  That coupled with Hector timeout issues became a real problem for us.


 could you share some details on this?  we're using hector and we see
 random timeout warns in the logs and not sure how to address them.


 thanks,
 deno







-- 

/*
Joe Stein
http://www.linkedin.com/in/charmalloc
Twitter: @allthingshadoop http://www.twitter.com/allthingshadoop
*/


Re: frequent node up/downs

2012-07-02 Thread feedly team
Couple more details. I confirmed that swap space is not being used (free -m
shows 0 swap) and cassandra.log has a message like JNA mlockall
successful. top shows the process having 9g in resident memory but 21.6g
in virtual...What accounts for the much larger virtual number? some kind of
off-heap memory?

I'm a little puzzled as to why I would get such long pauses without
swapping. I uncommented all the gc logging options in cassandra-env.sh to
try to see what is going on when the node freezes.

Thanks
Kireet

On Mon, Jul 2, 2012 at 9:51 PM, feedly team feedly...@gmail.com wrote:

 Yeah I noticed the leap second problem and ran the suggested fix, but I
 have been facing these problems before Saturday and still see the
 occasional failures after running the fix.

 Thanks.


 On Mon, Jul 2, 2012 at 11:17 AM, Marcus Both mb...@terra.com.br wrote:

 Yeah! Look that.

 http://arstechnica.com/business/2012/07/one-day-later-the-leap-second-v-the-internet-scorecard/
 I had the same problem. The solution was rebooting.

 On Mon, 2 Jul 2012 11:08:57 -0400
 feedly team feedly...@gmail.com wrote:

  Hello,
 I recently set up a 2 node cassandra cluster on dedicated hardware.
 In
  the logs there have been a lot of InetAddress xxx is now dead' or UP
  messages. Comparing the log messages between the 2 nodes, they seem to
  coincide with extremely long ParNew collections. I have seem some of up
 to
  50 seconds. The installation is pretty vanilla, I didn't change any
  settings and the machines don't seem particularly busy - cassandra is
 the
  only thing running on the machine with an 8GB heap. The machine has
 64GB of
  RAM and CPU/IO usage looks pretty light. I do see a lot of 'Heap is xxx
  full. You may need to reduce memtable and/or cache sizes' messages.
 Would
  this help with the long ParNew collections? That message seems to be
  triggered on a full collection.

 --
 Marcus Both