Re: Question about how to remove data

2015-08-22 Thread Analia Lorenzatto
Thanks guys for the answers!

Saludos / Regards.

Analía Lorenzatto.

Hapiness is not something really made. It comes from your own actions by
Dalai Lama


On 21 Aug 2015 2:31 pm, Sebastian Estevez sebastian.este...@datastax.com
wrote:

 To clarify, you do not need a ttl for deletes to be compacted away in
 Cassandra. When you delete, we create a tombstone which will remain in the
 system __at least__ gc grace seconds. We wait this long to give the
 tombstone a chance to make it to all replica nodes, the best practice is to
 run repairs as often as gc grace seconds in order to ensure edge cases
 where data comes back to life (i.e. the tombstone was never sent to one of
 your replicas and when the tombstones and data are removed from the other
 two replicas, all that is left is the old value.

 __at least__ are the key words in the previous paragraph, there are more
 conditions that need to be met in order for a tombstone to actually get
 cleaned up. As most things in Cassandra, these conditions are configurable
 (via the following compaction sub-properties):


 http://docs.datastax.com/en/cassandra/2.1/cassandra/operations/ops_configure_compaction_t.html

 All the best,


 [image: datastax_logo.png] http://www.datastax.com/

 Sebastián Estévez

 Solutions Architect | 954 905 8615 | sebastian.este...@datastax.com

 [image: linkedin.png] https://www.linkedin.com/company/datastax [image:
 facebook.png] https://www.facebook.com/datastax [image: twitter.png]
 https://twitter.com/datastax [image: g+.png]
 https://plus.google.com/+Datastax/about
 http://feeds.feedburner.com/datastax


 http://cassandrasummit-datastax.com/?utm_campaign=summit15utm_medium=summiticonutm_source=emailsignature

 DataStax is the fastest, most scalable distributed database technology,
 delivering Apache Cassandra to the world’s most innovative enterprises.
 Datastax is built to be agile, always-on, and predictably scalable to any
 size. With more than 500 customers in 45 countries, DataStax is the
 database technology and transactional backbone of choice for the worlds
 most innovative companies such as Netflix, Adobe, Intuit, and eBay.

 On Thu, Aug 20, 2015 at 4:13 PM, Daniel Chia danc...@coursera.org wrote:

 The TTL shouldn't matter if you deleted the data, since to my
 understanding the delete should shadow the data signaling to C* that the
 data is a candidate for removal on compaction.

 Others might know better, but it could very well be the fact that
 gc_grace_seconds is 0 that is causing your problems. Others might have
 other suggestions, but you could potentially use sstable2json to see the
 raw contents of the sstable on disk and see why data is still there.

 Thanks,
 Daniel

 On Thu, Aug 20, 2015 at 12:55 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello,

 Daniel, I am using Size Tiered compaction.

 My concern is that as I do not have a TTL defined on the Column family,
 and I do not have the possibility to create it.   Perhaps, the deleted
 data is never actually going to be removed?

 Thanks a lot!


 On Thu, Aug 20, 2015 at 4:24 AM, Daniel Chia danc...@coursera.org
 wrote:

 Is this a LCS family, or Size Tiered? Manually running compaction on
 LCS doesn't do anything until C* 2.2 (
 https://issues.apache.org/jira/browse/CASSANDRA-7272)

 Thanks,
 Daniel

 On Wed, Aug 19, 2015 at 6:56 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello Michael,

 Thanks for responding!

 I do not have snapshots on any node of the cluster.

 Saludos / Regards.

 Analía Lorenzatto.

 Hapiness is not something really made. It comes from your own
 actions by Dalai Lama


 On 19 Aug 2015 6:19 pm, Laing, Michael michael.la...@nytimes.com
 wrote:

 Possibly you have snapshots? If so, use nodetool to clear them.

 On Wed, Aug 19, 2015 at 4:54 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello guys,

 I have a cassandra cluster 2.1 comprised of 4 nodes.

 I removed a lot of data in a Column Family, then I ran manually a
 compaction on this Column family on every node.   After doing that, If I
 query that data, cassandra correctly says this data is not there.  But 
 the
 space on disk is exactly the same before removing that data.

 Also, I realized that  gc_grace_seconds = 0.  Some people on the
 internet say that it could produce zombie data, what do you think?

 I do not have a TTL defined on the Column family, and I do not have
 the possibility to create it.   So my questions is, given that I do not
 have a TTL defined is data going to be removed?  or the deleted data is
 never actually going to be deleted due to I do not have a TTL?


 Thanks in advance!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not
 weakness.  That is life.  By Captain Jean-Luc Picard.






 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not
 weakness.  That is life.  By Captain Jean-Luc Picard.






Re: Question about how to remove data

2015-08-20 Thread Analia Lorenzatto
Hello,

Daniel, I am using Size Tiered compaction.

My concern is that as I do not have a TTL defined on the Column family, and
I do not have the possibility to create it.   Perhaps, the deleted data
is never actually going to be removed?

Thanks a lot!


On Thu, Aug 20, 2015 at 4:24 AM, Daniel Chia danc...@coursera.org wrote:

 Is this a LCS family, or Size Tiered? Manually running compaction on LCS
 doesn't do anything until C* 2.2 (
 https://issues.apache.org/jira/browse/CASSANDRA-7272)

 Thanks,
 Daniel

 On Wed, Aug 19, 2015 at 6:56 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello Michael,

 Thanks for responding!

 I do not have snapshots on any node of the cluster.

 Saludos / Regards.

 Analía Lorenzatto.

 Hapiness is not something really made. It comes from your own actions
 by Dalai Lama


 On 19 Aug 2015 6:19 pm, Laing, Michael michael.la...@nytimes.com
 wrote:

 Possibly you have snapshots? If so, use nodetool to clear them.

 On Wed, Aug 19, 2015 at 4:54 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello guys,

 I have a cassandra cluster 2.1 comprised of 4 nodes.

 I removed a lot of data in a Column Family, then I ran manually a
 compaction on this Column family on every node.   After doing that, If I
 query that data, cassandra correctly says this data is not there.  But the
 space on disk is exactly the same before removing that data.

 Also, I realized that  gc_grace_seconds = 0.  Some people on the
 internet say that it could produce zombie data, what do you think?

 I do not have a TTL defined on the Column family, and I do not have the
 possibility to create it.   So my questions is, given that I do not have a
 TTL defined is data going to be removed?  or the deleted data is never
 actually going to be deleted due to I do not have a TTL?


 Thanks in advance!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not
 weakness.  That is life.  By Captain Jean-Luc Picard.






-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Question about how to remove data

2015-08-19 Thread Analia Lorenzatto
Hello guys,

I have a cassandra cluster 2.1 comprised of 4 nodes.

I removed a lot of data in a Column Family, then I ran manually a
compaction on this Column family on every node.   After doing that, If I
query that data, cassandra correctly says this data is not there.  But the
space on disk is exactly the same before removing that data.

Also, I realized that  gc_grace_seconds = 0.  Some people on the internet
say that it could produce zombie data, what do you think?

I do not have a TTL defined on the Column family, and I do not have the
possibility to create it.   So my questions is, given that I do not have a
TTL defined is data going to be removed?  or the deleted data is never
actually going to be deleted due to I do not have a TTL?


Thanks in advance!

-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Re: Question about how to remove data

2015-08-19 Thread Analia Lorenzatto
Hello Michael,

Thanks for responding!

I do not have snapshots on any node of the cluster.

Saludos / Regards.

Analía Lorenzatto.

Hapiness is not something really made. It comes from your own actions by
Dalai Lama


On 19 Aug 2015 6:19 pm, Laing, Michael michael.la...@nytimes.com wrote:

 Possibly you have snapshots? If so, use nodetool to clear them.

 On Wed, Aug 19, 2015 at 4:54 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello guys,

 I have a cassandra cluster 2.1 comprised of 4 nodes.

 I removed a lot of data in a Column Family, then I ran manually a
 compaction on this Column family on every node.   After doing that, If I
 query that data, cassandra correctly says this data is not there.  But the
 space on disk is exactly the same before removing that data.

 Also, I realized that  gc_grace_seconds = 0.  Some people on the internet
 say that it could produce zombie data, what do you think?

 I do not have a TTL defined on the Column family, and I do not have the
 possibility to create it.   So my questions is, given that I do not have a
 TTL defined is data going to be removed?  or the deleted data is never
 actually going to be deleted due to I do not have a TTL?


 Thanks in advance!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not weakness.
 That is life.  By Captain Jean-Luc Picard.





Re: After running nodetool clean up, the used disk space was increased

2015-05-15 Thread Analia Lorenzatto
Yes Robert, I already cleared the snapshots.  After that, the used disk
space is:

10.x.x.b@$ df -h /mnt/cassandra
Filesystem  Size  Used Avail Use% Mounted on
/dev/xvdb1  745G  174G  572G  24% /mnt/cassandra

But, the cluster shows me a different thing:

$ nodetool status
Datacenter: us-east
===
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
--  Address Load   Tokens  Owns (effective)  Host ID
Rack
UN  10.x.x.a  420.74 GB  256 66.7%
eed9e9f5-f279-4b2f-b521-c056cbf65b52  1c
UN  10.x.x.b  416.42 GB  256 68.3%
19492c26-4458-4a0b-af04-72e0aab6598e  1c
UN  10.x.x.c  165.15 GB  256 64.9%
b8da952c-24b3-444a-a34e-7a1804eee6e6  1c

I do not understand why the cluster still sees even more data before adding
the third node.

Thanks a lot!!


On Fri, May 15, 2015 at 6:27 PM, Robert Wille rwi...@fold3.com wrote:

  Have you cleared snapshots?

  On May 15, 2015, at 2:24 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

  The Replication Factor = 2.  The RP is the default, but not sure how to
 check it.
 I am attaching the output of: nodetool ring

  Thanks a lot!

 On Fri, May 15, 2015 at 4:17 PM, Kiran mk coolkiran2...@gmail.com wrote:

 run cleanup on all the nodes and wait till it completes.
 On May 15, 2015 10:47 PM, Analia Lorenzatto analialorenza...@gmail.com
 wrote:

 Hello guys,

  I have a cassandra cluster = 2.1.0-2 comprised of 3 nodes.  I
 successfully added the third node last week.  After that, I ran nodetool
 cleanup on one of the other two nodes, and it finished well but it
 increased the used disk space.
 Before running the clean up the node was 197 GB of used space, and after
 that it is 329GB used.  It is my understanding that the clean up frees up
 some space, but in this case it was highly increased.

  I am running out of space, that's why I added a third node.  Do you
 have any clue on how to proceed with that situation?

  Thanks in advance!!

  --
  Saludos / Regards.

  Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not
 weakness.  That is life.  By Captain Jean-Luc Picard.




  --
  Saludos / Regards.

  Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not weakness.
 That is life.  By Captain Jean-Luc Picard.
   list





-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Re: After running nodetool clean up, the used disk space was increased

2015-05-15 Thread Analia Lorenzatto
The Replication Factor = 2.  The RP is the default, but not sure how to
check it.
I am attaching the output of: nodetool ring

Thanks a lot!

On Fri, May 15, 2015 at 4:17 PM, Kiran mk coolkiran2...@gmail.com wrote:

 run cleanup on all the nodes and wait till it completes.
 On May 15, 2015 10:47 PM, Analia Lorenzatto analialorenza...@gmail.com
 wrote:

 Hello guys,

 I have a cassandra cluster = 2.1.0-2 comprised of 3 nodes.  I
 successfully added the third node last week.  After that, I ran nodetool
 cleanup on one of the other two nodes, and it finished well but it
 increased the used disk space.
 Before running the clean up the node was 197 GB of used space, and after
 that it is 329GB used.  It is my understanding that the clean up frees up
 some space, but in this case it was highly increased.

 I am running out of space, that's why I added a third node.  Do you have
 any clue on how to proceed with that situation?

 Thanks in advance!!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not weakness.
 That is life.  By Captain Jean-Luc Picard.




-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


list
Description: Binary data


Re: After running nodetool clean up, the used disk space was increased

2015-05-15 Thread Analia Lorenzatto
Thanks Kiran for answering!

I already ran cleanup on just one node.  At this moment, I am running on
the second one, but it did not finish there.


On Fri, May 15, 2015 at 3:37 PM, Kiran mk coolkiran2...@gmail.com wrote:

 Did you try running nodetool cleanup on all the nodes ?
 On May 15, 2015 10:47 PM, Analia Lorenzatto analialorenza...@gmail.com
 wrote:

 Hello guys,

 I have a cassandra cluster = 2.1.0-2 comprised of 3 nodes.  I
 successfully added the third node last week.  After that, I ran nodetool
 cleanup on one of the other two nodes, and it finished well but it
 increased the used disk space.
 Before running the clean up the node was 197 GB of used space, and after
 that it is 329GB used.  It is my understanding that the clean up frees up
 some space, but in this case it was highly increased.

 I am running out of space, that's why I added a third node.  Do you have
 any clue on how to proceed with that situation?

 Thanks in advance!!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not weakness.
 That is life.  By Captain Jean-Luc Picard.




-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Re: After running nodetool clean up, the used disk space was increased

2015-05-15 Thread Analia Lorenzatto
The cluster is comprised of 3 nodes, with a RP=2:

# nodetool status Datacenter: us-east
===
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
--  Address Load   Tokens  Owns (effective)  Host ID
Rack
UN  10.x.x.a  420.05 GB  256 66.7%
eed9e9f5-f279-4b2f-b521-c056cbf65b52  1c
UN  10.x.x.b  415.68 GB  256 68.3%
19492c26-4458-4a0b-af04-72e0aab6598e  1c
UN  10.x.x.c  164.45 GB  256 64.9%
b8da952c-24b3-444a-a34e-7a1804eee6e6  1c

10.x.x.c is the last node added.

Thanks!

On Fri, May 15, 2015 at 4:02 PM, Kiran mk coolkiran2...@gmail.com wrote:

 What is the data distribution status across nodes ? What is the RP ?
 On May 16, 2015 12:30 AM, Analia Lorenzatto analialorenza...@gmail.com
 wrote:

 Thanks Kiran for answering!

 I already ran cleanup on just one node.  At this moment, I am running on
 the second one, but it did not finish there.


 On Fri, May 15, 2015 at 3:37 PM, Kiran mk coolkiran2...@gmail.com
 wrote:

 Did you try running nodetool cleanup on all the nodes ?
 On May 15, 2015 10:47 PM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:

 Hello guys,

 I have a cassandra cluster = 2.1.0-2 comprised of 3 nodes.  I
 successfully added the third node last week.  After that, I ran nodetool
 cleanup on one of the other two nodes, and it finished well but it
 increased the used disk space.
 Before running the clean up the node was 197 GB of used space, and
 after that it is 329GB used.  It is my understanding that the clean up
 frees up some space, but in this case it was highly increased.

 I am running out of space, that's why I added a third node.  Do you
 have any clue on how to proceed with that situation?

 Thanks in advance!!

 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not
 weakness.  That is life.  By Captain Jean-Luc Picard.




 --
 Saludos / Regards.

 Analía Lorenzatto.

 “It's possible to commit no errors and still lose. That is not weakness.
 That is life.  By Captain Jean-Luc Picard.




-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


After running nodetool clean up, the used disk space was increased

2015-05-15 Thread Analia Lorenzatto
Hello guys,

I have a cassandra cluster = 2.1.0-2 comprised of 3 nodes.  I successfully
added the third node last week.  After that, I ran nodetool cleanup on one
of the other two nodes, and it finished well but it increased the used disk
space.
Before running the clean up the node was 197 GB of used space, and after
that it is 329GB used.  It is my understanding that the clean up frees up
some space, but in this case it was highly increased.

I am running out of space, that's why I added a third node.  Do you have
any clue on how to proceed with that situation?

Thanks in advance!!

-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Re: Unexpected behavior after adding successffully a new node

2015-05-12 Thread Analia Lorenzatto
Just in case I want to clarify that after bootstrapping the third node, it
got data and seemed to be working fine.  But it was the last night when
 the cluster started behaving in a weird way.  The last node (successfully
added last week) were being reported up and down all the time. After
restarting cassandra service, the node got empty but reported as Up and
Normal.

Also, the amount of data reported by nodetool on the other nodes is much
more than the amount already used in the device, is that normal?

root@10.0.0.a:~# df -h /mnt
Filesystem  Size  Used Avail Use% Mounted on
/dev/xvdb1  745G  163G  583G  22% /mnt/cassandra

root@:10.0.0.b~$ df -h /mnt
Filesystem  Size  Used Avail Use% Mounted on
/dev/xvdb1  745G  196G  550G  27% /mnt/cassandra


# nodetool status
Datacenter: us-east
===
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
--  Address Load   Tokens  Owns (effective)  Host ID
Rack
UN  10.0.0.a  391.34 GB  256 66.7%
eed9e9f5-f279-4b2f-b521-c056cbf65b52  1c
UN  10.0.0.b  383.38 GB  256 68.3%
19492c26-4458-4a0b-af04-72e0aab6598e  1c

Thanks in advance!




On Tue, May 12, 2015 at 2:35 PM, Robert Coli rc...@eventbrite.com wrote:

 On Tue, May 12, 2015 at 9:59 AM, arun sirimalla arunsi...@gmail.com
 wrote:

 Try running repair on node 3.


 Mostly disagree. If a node is empty after a bootstrap, remove it and
 re-bootstrap it.

 =Rob





-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Unexpected behavior after adding successffully a new node

2015-05-12 Thread Analia Lorenzatto
Hello guys,


I have a cluster 2.1.0-2 comprised of 3 nodes.  The replication factor=2.
We successfully added the third node last week.  After that, We ran clean
ups on one node at that time.  Then we ran repairs on all the nodes, and
finally compactions on all the CFs.

Last night, I noticed the cluster started behaving in a weird way.  The
last node (successfully added last week) were being reported up and down
all the time.  I could see a lot of messages like this on logs:

WARN  [SharedPool-Worker-33] 2015-05-11 21:31:45,125
AbstractTracingAwareExecutorService.java:167 - Uncaught exception on thread
Thread[SharedPool-Worker-33,5,main]: {}
java.lang.RuntimeException: java.io.FileNotFoundException:
/mnt/cassandra/data/matchings-85b4929048e211e4a949a3ed319cbedc/matchings-ka-3914-Data.db
(No such file or directory)

At the same time the consumption of heap used was on the top, up to the
point the rest of the cluster saw this node as down.  After that, I just
restarted the cassandra service with no problems on that node.

Now, I can see the three nodes on the cluster Up and Normal, but this last
node (which was rebooted) does not have data.  But it has all the structure
of cassandra data.

I can query against the new node and I get the same result as if do the
query against the others nodes.  But, on this new node I do not have any
SStables:



root@prd-rtbkit-cassandra-03:/var/log/cassandra# nodetool status
Datacenter: us-east
===
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
--  Address Load   Tokens  Owns (effective)  Host ID
Rack
UN  10.0.0.a  390.28 GB  256 66.7%
eed9e9f5-f279-4b2f-b521-c056cbf65b52  1c
UN  10.0.0.b  382.36 GB  256 68.3%
19492c26-4458-4a0b-af04-72e0aab6598e  1c
UN  10.0.0.c  40.61 MB   256 64.9%
b8da952c-24b3-444a-a34e-7a1804eee6e6  1c

What do you recommend to do? Leave this as if, remove it and try to join
this or a new one?
Thanks in advance!!

-- 
Saludos / Regards.

Analía Lorenzatto.

“It's possible to commit no errors and still lose. That is not weakness.
That is life.  By Captain Jean-Luc Picard.


Re: New node got stuck joining the cluster after a while

2015-04-28 Thread Analia Lorenzatto
Thanks very much for answering!

Do you think that after failing the joining task of a node to the cluster
should I run some repairs and cleanups?

Thanks!

On Tue, Apr 28, 2015 at 5:13 AM, Carlos Rolo r...@pythian.com wrote:

 Hi,

 The 2.1.x series is not recommeded for use, especially the first versions.
 I would downgrade to 2.0.14 or if must stay on 2.1 upgrade your cluster to
 2.1.4 or the imminent release of 2.1.5.

 This mailing list as a few tips how to deal with the 2.1.x releases, but
 the best way is indeed a downgrade or wait for the 2.1.5.

 Regards,

 Carlos Juzarte Rolo
 Cassandra Consultant

 Pythian - Love your data

 rolo@pythian | Twitter: cjrolo | Linkedin: *linkedin.com/in/carlosjuzarterolo
 http://linkedin.com/in/carlosjuzarterolo*
 Mobile: +31 6 159 61 814 | Tel: +1 613 565 8696 x1649
 www.pythian.com

 On Tue, Apr 28, 2015 at 3:30 AM, Analia Lorenzatto 
 analialorenza...@gmail.com wrote:


 Hello guys,

 I have a cluster comprised of 2 nodes, configured with vnodes.  Using
 2.1.0-2 version of cassandra.

 And I am facing an issue when I want to joing a new node to the cluster.

 At first starting joining but then it got stuck:

 UN  1x.x.x.x  348.11 GB  256 100.0%  1c
 UN  1x.x.x.x  342.74 GB  256 100.0%  1c
 UJ  1x.x.x.x  26.86 GB   256 ?   1c


 I can see some errors on the already working nodes:

 *WARN  [SharedPool-Worker-7] 2015-04-27 17:41:16,060
 SliceQueryFilter.java:236 - Read 5001 live and 66548 tombstoned cells in
 usmc.userpixel (see tombstone_warn_threshol*
 *d). 5000 columns was requested, slices=[-],
 delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647
 2147483647}*
 *WARN  [SharedPool-Worker-32] 2015-04-27 17:41:16,668
 SliceQueryFilter.java:236 - Read 2012 live and 30440 tombstoned cells in
 usmc.userpixel (see tombstone_warn_thresho*
 *ld). 5001 columns was requested,
 slices=[b6d051df-0a8f-4c13-b93c-1b4ff0d82b8d:date-],
 delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647}*

 *ERROR [CompactionExecutor:35638] 2015-04-27 19:06:07,613
 CassandraDaemon.java:166 - Exception in thread
 Thread[CompactionExecutor:35638,1,main]*
 *java.lang.AssertionError: Memory was freed*
 *at
 org.apache.cassandra.io.util.Memory.checkPosition(Memory.java:281)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at org.apache.cassandra.io.util.Memory.getInt(Memory.java:233)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.io.sstable.IndexSummary.getPositionInSummary(IndexSummary.java:118)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.io.sstable.IndexSummary.getKey(IndexSummary.java:123)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 * at
 org.apache.cassandra.io.sstable.IndexSummary.binarySearch(IndexSummary.java:92)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.io.sstable.SSTableReader.getSampleIndexesForRanges(SSTableReader.java:1209)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.io.sstable.SSTableReader.estimatedKeysForRanges(SSTableReader.java:1165)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.db.compaction.AbstractCompactionStrategy.worthDroppingTombstones(AbstractCompactionStrategy.java:328)
 ~[apache-cassandra-2.1.0.jar:2.1.0*
 *]*
 *at
 org.apache.cassandra.db.compaction.LeveledCompactionStrategy.findDroppableSSTable(LeveledCompactionStrategy.java:365)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.db.compaction.LeveledCompactionStrategy.getMaximalTask(LeveledCompactionStrategy.java:127)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.db.compaction.LeveledCompactionStrategy.getNextBackgroundTask(LeveledCompactionStrategy.java:112)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:229)
 ~[apache-cassandra-2.1.0.jar:2.1.0]*
 *at
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
 ~[na:1.7.0_51]*
 *at java.util.concurrent.FutureTask.run(FutureTask.java:262)
 ~[na:1.7.0_51]*
 *at
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 ~[na:1.7.0_51]*
 *at
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 [na:1.7.0_51]*
 *at java.lang.Thread.run(Thread.java:744) [na:1.7.0_51]*

 But I do not see any warning or error message in logs of the joining
 nodes.  I just see an exception there when I run: nodetool info:

 root@:~# nodetool info
 ID   : f5e49647-59fa-474f-b6af-9f65abc43581
 Gossip active: true
 Thrift active: false
 Native Transport active: false
 Load : 26.86 GB
 Generation No: 1430163258
 Uptime (seconds) : 18799
 Heap Memory (MB) : 4185.15 / 7566.00
 error: null
 -- StackTrace --
 java.lang.AssertionError

New node got stuck joining the cluster after a while

2015-04-27 Thread Analia Lorenzatto
Hello guys,

I have a cluster comprised of 2 nodes, configured with vnodes.  Using
2.1.0-2 version of cassandra.

And I am facing an issue when I want to joing a new node to the cluster.

At first starting joining but then it got stuck:

UN  1x.x.x.x  348.11 GB  256 100.0%  1c
UN  1x.x.x.x  342.74 GB  256 100.0%  1c
UJ  1x.x.x.x  26.86 GB   256 ?   1c


I can see some errors on the already working nodes:

*WARN  [SharedPool-Worker-7] 2015-04-27 17:41:16,060
SliceQueryFilter.java:236 - Read 5001 live and 66548 tombstoned cells in
usmc.userpixel (see tombstone_warn_threshol*
*d). 5000 columns was requested, slices=[-],
delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647
2147483647}*
*WARN  [SharedPool-Worker-32] 2015-04-27 17:41:16,668
SliceQueryFilter.java:236 - Read 2012 live and 30440 tombstoned cells in
usmc.userpixel (see tombstone_warn_thresho*
*ld). 5001 columns was requested,
slices=[b6d051df-0a8f-4c13-b93c-1b4ff0d82b8d:date-],
delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647}*

*ERROR [CompactionExecutor:35638] 2015-04-27 19:06:07,613
CassandraDaemon.java:166 - Exception in thread
Thread[CompactionExecutor:35638,1,main]*
*java.lang.AssertionError: Memory was freed*
*at
org.apache.cassandra.io.util.Memory.checkPosition(Memory.java:281)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at org.apache.cassandra.io.util.Memory.getInt(Memory.java:233)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.io.sstable.IndexSummary.getPositionInSummary(IndexSummary.java:118)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.io.sstable.IndexSummary.getKey(IndexSummary.java:123)
~[apache-cassandra-2.1.0.jar:2.1.0]*
* at
org.apache.cassandra.io.sstable.IndexSummary.binarySearch(IndexSummary.java:92)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.io.sstable.SSTableReader.getSampleIndexesForRanges(SSTableReader.java:1209)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.io.sstable.SSTableReader.estimatedKeysForRanges(SSTableReader.java:1165)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.db.compaction.AbstractCompactionStrategy.worthDroppingTombstones(AbstractCompactionStrategy.java:328)
~[apache-cassandra-2.1.0.jar:2.1.0*
*]*
*at
org.apache.cassandra.db.compaction.LeveledCompactionStrategy.findDroppableSSTable(LeveledCompactionStrategy.java:365)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.db.compaction.LeveledCompactionStrategy.getMaximalTask(LeveledCompactionStrategy.java:127)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.db.compaction.LeveledCompactionStrategy.getNextBackgroundTask(LeveledCompactionStrategy.java:112)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:229)
~[apache-cassandra-2.1.0.jar:2.1.0]*
*at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
~[na:1.7.0_51]*
*at java.util.concurrent.FutureTask.run(FutureTask.java:262)
~[na:1.7.0_51]*
*at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
~[na:1.7.0_51]*
*at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
[na:1.7.0_51]*
*at java.lang.Thread.run(Thread.java:744) [na:1.7.0_51]*

But I do not see any warning or error message in logs of the joining
nodes.  I just see an exception there when I run: nodetool info:

root@:~# nodetool info
ID   : f5e49647-59fa-474f-b6af-9f65abc43581
Gossip active: true
Thrift active: false
Native Transport active: false
Load : 26.86 GB
Generation No: 1430163258
Uptime (seconds) : 18799
Heap Memory (MB) : 4185.15 / 7566.00
error: null
-- StackTrace --
java.lang.AssertionError
at
org.apache.cassandra.locator.TokenMetadata.getTokens(TokenMetadata.java:440)
at
org.apache.cassandra.service.StorageService.getTokens(StorageService.java:2079)
at
org.apache.cassandra.service.StorageService.getTokens(StorageService.java:2068)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.Trampoline.invoke(MethodUtil.java:75)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.MethodUtil.invoke(MethodUtil.java:279)
at
com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:112)
at