Re: can I have a mix of 32 and 64 bit machines in a cluster?

2012-10-10 Thread Mateusz Korniak
On Tuesday 09 of October 2012, Brian Tarbox wrote:
 I can't imagine why this would be a problem but I wonder if anyone has
 experience with running a mix of 32 and 64 bit nodes in a cluster.

We are running mixed userspace 64/32bit (all kernels 64bit) linux 1.0.10  
cluster for our daily operations for months now without issue.

Regards,
-- 
Mateusz Korniak


[0.8.x] Node join stuck with all network transfers done

2012-01-08 Thread Mateusz Korniak
Hi !
I have problem with 0.8.7 node joining cluster of two 0.8.9s (RF=2).
Seems all transfers ware done but joining node(.17) does not change it's state 
[3].
Strange is Nothing streaming from /192.168.3.8 netstats result [2] and still 
raising number of pending tasks [1], while .8 not transferring anything [4].
I tried to restart each of nodes, it didn't help except joining process 
started again, transferring all data again, stuck in same moment but having 
doubled load on joining node.

Any hints ? Thanks in advance, regards.

[1]:
@192.168.3.17 ~]$ nodetool -h localhost compactionstats
pending tasks: 836

[2]:
@192.168.3.17 ~]$ nodetool -h localhost netstats
Mode: Bootstrapping
Not sending any streams.
 Nothing streaming from /192.168.3.8
Pool NameActive   Pending  Completed
Commandsn/a 0171
Responses   n/a 09387965

[3]:
@192.168.3.17 ~]$ nodetool -h localhost ring
Address DC  RackStatus State   LoadOwns
Token
113427455640312821154458202477256070485
192.168.3.8 datacenter1 rack1   Up Normal  128.51 GB   33.33%  
0
192.168.3.7 datacenter1 rack1   Up Normal  137.65 GB   50.00%  
85070591730234615865843651857942052864
192.168.3.17datacenter1 rack1   Up Joining 127.02 GB   16.67%  
113427455640312821154458202477256070485

[4]:   
@192.168.3.8 ~]$ nodetool -h localhost netstats
Mode: Normal
Not sending any streams.
Not receiving any streams.
Pool NameActive   Pending  Completed
Commandsn/a 05261062
Responses   n/a 02963742


-- 
Mateusz Korniak


Re: [0.8.x] Node join stuck with all network transfers done

2012-01-09 Thread Mateusz Korniak
On Monday 09 of January 2012, aaron morton wrote:
 (...) Is there a reason you are not adding 0.8.9 ?
Only my mistake, I repeated procedure with 0.8.9 node joining and now, after 
finishing net transfers, node was busy compacting, finally switching to 
Normal state.  

 Check the logs on .17 for errors. 
No exceptions, just Finished streaming session, netstats empty, node idling 
(net, cpu, io ) and rising number of pending tasks without doing anything.

Big thanks for caring to answer, regards !
 
 http://www.thelastpickle.com
 
 On 9/01/2012, at 8:20 PM, Mateusz Korniak wrote:
  Hi !
  I have problem with 0.8.7 node joining cluster of two 0.8.9s (RF=2).
  Seems all transfers ware done but joining node(.17) does not change it's
  state [3].
  Strange is Nothing streaming from /192.168.3.8 netstats result [2] and
  still raising number of pending tasks [1], while .8 not transferring
  anything [4]. I tried to restart each of nodes, it didn't help except
  joining process started again, transferring all data again, stuck in
  same moment but having doubled load on joining node.
  
  Any hints ? Thanks in advance, regards.
  
  [1]:
  @192.168.3.17 ~]$ nodetool -h localhost compactionstats
  pending tasks: 836
  
  [2]:
  @192.168.3.17 ~]$ nodetool -h localhost netstats
  Mode: Bootstrapping
  Not sending any streams.
  Nothing streaming from /192.168.3.8
  Pool NameActive   Pending  Completed
  Commandsn/a 0171
  Responses   n/a 09387965
  
  [3]:
  @192.168.3.17 ~]$ nodetool -h localhost ring
  Address DC  RackStatus State   Load   
  Owns Token
  
 113427455640312821154458202477256070485
  
  192.168.3.8 datacenter1 rack1   Up Normal  128.51 GB  
  33.33% 0
  192.168.3.7 datacenter1 rack1   Up Normal  137.65 GB  
  50.00% 85070591730234615865843651857942052864
  192.168.3.17datacenter1 rack1   Up Joining 127.02 GB  
  16.67% 113427455640312821154458202477256070485
  
  [4]:
  @192.168.3.8 ~]$ nodetool -h localhost netstats
  Mode: Normal
  Not sending any streams.
  Not receiving any streams.
  Pool NameActive   Pending  Completed
  Commandsn/a 05261062
  Responses   n/a 02963742


-- 
Mateusz Korniak


Re: Doubt regarding CQL

2012-02-22 Thread Mateusz Korniak
On Wednesday 22 of February 2012, Rishabh Agrawal wrote:
 I have installed CQL drivers for python. When I try execute cqlsh I get
 following error
 cql-1.0.3$ cqlsh localhost 9160
 (...)
   File /usr/local/lib/python2.7/dist-packages/cql/cassandra/ttypes.py,
 line 7, in module from thrift.Thrift import *
 ImportError: No module named thrift.Thrift

Seems you do not have installed python thrift module.

In my distro (PLD) it is:
Package:python-thrift-0.5.0-4.i686
/usr/lib/python2.7/site-packages:  Thrift-0.1-py2.7.egg-info,
/usr/lib/python2.7/site-packages/thrift:  TSCons.pyc, TSCons.pyo, 
TSerialization.pyc, TSerialization.pyo, Thrift.pyc, Thrift.pyo, __init__.pyc, 
__init__.pyo,
/usr/lib/python2.7/site-packages/thrift/protocol:  TBinaryProtocol.pyc, 
TBinaryProtocol.pyo, TCompactProtocol.pyc, TCompactProtocol.pyo, 
TProtocol.pyc, TProtocol.pyo, __init__.pyc, __init__.pyo, fastbinary.so
/usr/lib/python2.7/site-packages/thrift/server:  THttpServer.pyc, 
THttpServer.pyo, TNonblockingServer.pyc, TNonblockingServer.pyo, TServer.pyc, 
TServer.pyo, __init__.pyc, __init__.pyo
/usr/lib/python2.7/site-packages/thrift/transport:  THttpClient.pyc, 
THttpClient.pyo, TSocket.pyc, TSocket.pyo, TTransport.pyc, TTransport.pyo, 
TTwisted.pyc, TTwisted.pyo, __init__.pyc, __init__.pyo


Regards,

-- 
Mateusz Korniak


Re: Bootstrapping a new node to a running cluster (setting RF N)

2012-03-15 Thread Mateusz Korniak
On Thursday 15 of March 2012, aaron morton wrote:
  1. a running cluster of N=3,  R=3
  2. upgrade R to 4
 
 You should not be allowed to set the RF higher than the number of nodes.

I wonder why is that restriction ?
It would be easier to increase RF first (similar as having new node down), add 
node (not responding to client reads), repair new node, turn on clients.

Would it be simpler than changing CL ?

 I'm going to assume that clients on the web server only talk to the local
 cassandra. So that when you add the new cassandra node it will not have
 any clients until you serve pages off the node.
 
 0. Personally I would run a repair before doing this to ensure the data is
 fully distributed. 1. Optionally, increase the CL  QUOURM. See step 3.
 2. Add the new node with auto_bootstrap off. It will join the ring, write
 requests will be sent to it (from other cassandra nodes), but it should
 not get any direct client reads. It will not stream data from other nodes.
 3. It is now possible for a READ to be received at an old node where it is
 no longer a replica for the row. It has to send the request to another
 node. If it is sent to the new node (at CL ONE) the read will fail. If you
 are running at a high CL it will always involve the old nodes. 4. Update
 the RF to 4. Every node is now a replica for every key.
 5. Roll back the CL change.
 6. Repair the new node.
 7. Turn on the clients for the new node.

Regards,
-- 
Mateusz Korniak


Re: Advice on architecture

2012-03-28 Thread Mateusz Korniak
On Wednesday 28 of March 2012, Igor wrote:
 I'm also trying to evaluate different strategies for RAID0 as drive for
 cassandra data storage. If I need 2T space to keep node tables, which
 drive configuration is better: 1T x 2drives or 500G x 4drives? 

Having _similar_ family of HDDs 4x smaller should be twice faster in reads 
than 2x bigger.


 Should I use hardware raid or linux raid is ok?
Instead of buying hardware raids buy more disks/nodes - should give more 
performance gain.

 I mostly concerned with read performance.


-- 
Mateusz Korniak
(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa.
Nikos Kazantzakis - Grek Zorba

-- 
Mateusz Korniak


Meaning of compression chunk_length_kb

2012-06-14 Thread Mateusz Korniak
Hi !
What is meaning  of
chunk_length_kb: sets the compression chunk size in kilobytes. 
?

It means that uncompressed sstable data is compressed to approximately 
chunk_length_kb and every read needs to read approximately chunk_length_kb and 
decompress it to read any value from compressed range  ?

Or it means approximately chunk_length_kb of sstable data is compressed and 
stored on disk, so similar values must be in chunk_length_kb range to make 
compression efficient  ?

Or something else ?

Thanks in advance, regards, 

-- 
Mateusz Korniak


How schema disagreement can be fixed faster on 1.0.10 cluster ?

2012-07-26 Thread Mateusz Korniak
Hi !
We got into schema disagreement situation on 1.0.10 having 250GB of compressed 
data per node.

Following
http://wiki.apache.org/cassandra/FAQ#schema_disagreement
after node restart looks like it is replaying all schema changes one be one , 
right ? 
As we did a lot of them during cluster lifetime, now node is busy creating 
long time ago dropped secondary indexes which looks like gonna take hours.
Can it be done faster ?

1. Can we move all data SStables out of data/*/ directories,
2. follow FAQ#schema_disagreement (it should be faster on no data node) until 
we reach schema agreement. 
3. Than stop cassandra,
4. Copy files back.
5. Start cassandra.


Will it work ?

Extra option is to disable thrift during above process (can it be done in 
config ? In cassandra.yaml rpc_port: 0 ? )



Thanks in advance for any hints, regards,

-- 
Mateusz Korniak


Re: Cassandra 1.0 hangs during GC

2012-07-30 Thread Mateusz Korniak
On Monday 30 of July 2012, Nikolay Kоvshov wrote:
 What I plan to compare between 'bad' cluster and 'good' cluster:
 
 - Configs, schemas, data etc: same
 - java version : same
 - RAM and CPU : 'bad' cluster has more
 - Ubuntu version
 - Networking
 - What else???

- JNA. If it's working and swapping is disabled for cassandra binary.
- IO subsystem. kernel, io schedulers, performance of IO, HDDs 
performance/errors.


-- 
Mateusz Korniak


Re: Cassandra 1.0 hangs during GC

2012-07-30 Thread Mateusz Korniak
On Monday 30 of July 2012, Nikolay Kоvshov wrote:
 -  JNA is not installed on both machines

So your GC times may be strongly [1] affected by swapping. 
IIRC, also snapshotting is more expensive and may trigger more swapping.
I would start with turning JNA mlockall on [2].


[1]:
Not sure if up to numbers you presented ( many seconds)...

[2]:
INFO [main] 2012-07-27 12:18:27,135 CLibrary.java (line 109) JNA mlockall 
successful



-- 
Mateusz Korniak


Re: lost keyspace

2012-08-08 Thread Mateusz Korniak
On Wednesday 08 of August 2012, Arend-Jan Wijtzes wrote:
 Forgot to mention that the keyspace 'twitter' was created, then droppend
 and re-created a couple of days ago.
 
 How about if I create a new keyspace with the same definition and then
 copy the existing tables into the proper place and call nodetool
 refresh on each node. Would that work or are there references in the
 tables to the keyspace name?

Worth to try, should work.  Perhaps use  _copy_ of SStables.
I did that during schema version desync solving few times on 1.0.x.
Make sure you have same schema version on all nodes first.

Regards,
-- 
Mateusz Korniak


Reading data problems during bootstrap. [pycassa 0.7.0 rc4]

2011-01-04 Thread Mateusz Korniak
hi ! 
As cassandra newbie, I am trying to convert my single node cluster to cluster 
with two nodes with RF=2.

I have one node cluster , RF=1  all data accessible:
nodetool -h 192.168.3.8  ring
Address Status State   LoadOwnsToken
192.168.3.8 Up Normal  1.59 GB 100.00% 
150705614882854895881815284349323762700

  Replication Strategy: org.apache.cassandra.locator.SimpleStrategy
Replication Factor: 1


If I switch to RF=2 here my cluster will refuse updates and reads. So I 
bootstrap 2nd node (192.168.3.4) :

$ nodetool -h 192.168.3.8  ring
Address Status State   LoadOwnsToken
   
150705614882854895881815284349323762700
192.168.3.4 Up Joining 120.94 KB   49.80%  
65291865063116528976449321853009633660
192.168.3.8 Up Normal  1.59 GB 50.20%  
150705614882854895881815284349323762700

question 1: Why in that point I am unable to read data from part of keys ? 

Even when node is up:

nodetool -h 192.168.3.8  ring
Address Status State   LoadOwnsToken
   
150705614882854895881815284349323762700
192.168.3.4 Up Normal  721.37 MB   49.80%  
65291865063116528976449321853009633660
192.168.3.8 Up Normal  1.59 GB 50.20%  
150705614882854895881815284349323762700

I am still unable to read part of my original set of data with 
ConsistencyLevel.ONE  (NotFoundException in pycassa) :/

question 2: Why is that ? And what should I do to have cluster with full data 
?

Next I planned to do:
update keyspacewith replication_factor=2;
repair both nodes,
and this point have fully working 2 node cluster with RF=2

question 3: Is this proper approach or there is better one ?


question 4: I hoped that during above operations I would be able to _read_ 
whole dataset as it was at beginning in one node cluster, is it possible ?


Thanks in advance for any answers, regards,
-- 
Mateusz Korniak


Re: Cassandra from unprivileged user

2011-02-15 Thread Mateusz Korniak
On Tuesday 15 of February 2011, ruslan usifov wrote:
 Is it possible to launch cassandra from unprivileged user?
On linux - yes.


-- 
Mateusz Korniak


[0.7.2] Compacting exception

2011-03-18 Thread Mateusz Korniak
Hi !
I have run with Cassandra 0.7.2 out of disc space and after moving to bigger 
partition I experience compaction failures[1].

1) I suspect one of SSTables is broken. If I am right how can I find which one 
exactly ?

2) Knowing which one is broken is it safe to stop Cassandra, remove 
-Data.db -Filter.db -Index.db -Statistics.db of broken SSTable and restart ? 
Sure I will loose data but it's one of RF=3 nodes so it's not big problem.

Any suggestions, hints ?
Thanks in advance, regards,

[1]
INFO [CompactionExecutor:1] 2011-03-18 12:27:17,103 CompactionManager.java 
(line 458) Compacted to /var/lib/cassandra/data/foo/bar-tmp-f-707-Data.db.  
42,276,185 to 42,103,628 (~99% of original) bytes for 33,869 keys.  Time: 
25,924ms.
 INFO [CompactionExecutor:1] 2011-03-18 12:27:17,133 CompactionManager.java 
(line 373) Compacting [org.apache.cassandra.io.sstable.SSTableReader(path=
'/var/lib/cassandra/data/foo/bar-f-238-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-243-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-361-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-635-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-643-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-680-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-684-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-685-Data.db'), 
org.apache.cassandra.io.sstable.SSTableReader( 
path='/var/lib/cassandra/data/foo/bar-f-687-Data.db')]
ERROR [CompactionExecutor:1] 2011-03-18 12:27:18,241 
AbstractCassandraDaemon.java (line 114) Fatal exception in thread 
Thread[CompactionExecutor:1,1,main]
java.io.IOError: java.io.EOFException
at 
org.apache.cassandra.io.sstable.SSTableIdentityIterator.init(SSTableIdentityIterator.java:78)
at 
org.apache.cassandra.io.sstable.SSTableScanner$KeyScanningIterator.next(SSTableScanner.java:179)
at 
org.apache.cassandra.io.sstable.SSTableScanner$KeyScanningIterator.next(SSTableScanner.java:144)
at 
org.apache.cassandra.io.sstable.SSTableScanner.next(SSTableScanner.java:136)
at 
org.apache.cassandra.io.sstable.SSTableScanner.next(SSTableScanner.java:39)
at 
org.apache.commons.collections.iterators.CollatingIterator.set(CollatingIterator.java:284)
at 
org.apache.commons.collections.iterators.CollatingIterator.least(CollatingIterator.java:326)
at 
org.apache.commons.collections.iterators.CollatingIterator.next(CollatingIterator.java:230)
at 
org.apache.cassandra.utils.ReducingIterator.computeNext(ReducingIterator.java:68)
at 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:136)
at 
com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:131)
at 
org.apache.commons.collections.iterators.FilterIterator.setNextObject(FilterIterator.java:183)
at 
org.apache.commons.collections.iterators.FilterIterator.hasNext(FilterIterator.java:94)
at 
org.apache.cassandra.db.CompactionManager.doCompaction(CompactionManager.java:427)
at 
org.apache.cassandra.db.CompactionManager$1.call(CompactionManager.java:123)
at 
org.apache.cassandra.db.CompactionManager$1.call(CompactionManager.java:93)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
Caused by: java.io.EOFException
at 
org.apache.cassandra.io.sstable.IndexHelper.skipIndex(IndexHelper.java:65)
at 
org.apache.cassandra.io.sstable.SSTableIdentityIterator.init(SSTableIdentityIterator.java:70)
... 20 more


-- 
Mateusz Korniak


Re: [0.7.2] Compacting exception

2011-03-18 Thread Mateusz Korniak
On Friday 18 of March 2011, Jonathan Ellis wrote:
 Upgrade to 0.7.4 and run nodetool scrub (then watch the log).

Unfortunately nodetool scrub fails with:

 WARN [CompactionExecutor:1] 2011-03-19 00:34:53,511 CompactionManager.java 
(line 607) Non-fatal error reading row (stacktrace follows)
java.io.IOError: java.io.IOException: Impossible row size 3458764513820542976
at 
org.apache.cassandra.db.CompactionManager.doScrub(CompactionManager.java:589)
at 
org.apache.cassandra.db.CompactionManager.access$600(CompactionManager.java:56)
at 
org.apache.cassandra.db.CompactionManager$3.call(CompactionManager.java:195)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
Caused by: java.io.IOException: Impossible row size 3458764513820542976
... 8 more
 INFO [CompactionExecutor:1] 2011-03-19 00:34:53,512 CompactionManager.java 
(line 613) Retrying from row index; data is 10296 bytes starting at 2648761
ERROR [CompactionExecutor:1] 2011-03-19 00:34:53,512 
AbstractCassandraDaemon.java (line 112) Fatal exception in thread 
Thread[CompactionExecutor:1,1,main]
java.lang.OutOfMemoryError: Requested array size exceeds VM limit
at 
org.apache.cassandra.utils.BloomFilterSerializer.deserialize(BloomFilterSerializer.java:49)
at 
org.apache.cassandra.utils.BloomFilterSerializer.deserialize(BloomFilterSerializer.java:30)
at 
org.apache.cassandra.io.sstable.IndexHelper.defreezeBloomFilter(IndexHelper.java:117)
at 
org.apache.cassandra.io.sstable.SSTableIdentityIterator.init(SSTableIdentityIterator.java:87)
at 
org.apache.cassandra.db.CompactionManager.doScrub(CompactionManager.java:618)
at 
org.apache.cassandra.db.CompactionManager.access$600(CompactionManager.java:56)
at 
org.apache.cassandra.db.CompactionManager$3.call(CompactionManager.java:195)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

 If it
 reports bad rows then run repair post-scrub.

You mean:
 nodetool  repair
?

Thanks for help, regards,
 
 On Fri, Mar 18, 2011 at 6:44 AM, Mateusz Korniak
 
 mateusz-li...@ant.gliwice.pl wrote:
  Hi !
  I have run with Cassandra 0.7.2 out of disc space and after moving to
  bigger partition I experience compaction failures[1].
  
  1) I suspect one of SSTables is broken. If I am right how can I find
  which one exactly ?
  
  2) Knowing which one is broken is it safe to stop Cassandra, remove
  -Data.db -Filter.db -Index.db -Statistics.db of broken SSTable and
  restart ? Sure I will loose data but it's one of RF=3 nodes so it's not
  big problem.
  
  Any suggestions, hints ?
  Thanks in advance, regards,
  
  [1]
  INFO [CompactionExecutor:1] 2011-03-18 12:27:17,103
  CompactionManager.java (line 458) Compacted to
  /var/lib/cassandra/data/foo/bar-tmp-f-707-Data.db. 42,276,185 to
  42,103,628 (~99% of original) bytes for 33,869 keys.  Time: 25,924ms.
   INFO [CompactionExecutor:1] 2011-03-18 12:27:17,133
  CompactionManager.java (line 373) Compacting
  [org.apache.cassandra.io.sstable.SSTableReader(path=
  '/var/lib/cassandra/data/foo/bar-f-238-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-243-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-361-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-635-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-643-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-680-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-684-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-685-Data.db'),
  org.apache.cassandra.io.sstable.SSTableReader(
  path='/var/lib/cassandra/data/foo/bar-f-687-Data.db')]
  ERROR [CompactionExecutor:1] 2011-03-18 12:27:18,241
  AbstractCassandraDaemon.java (line 114) Fatal exception in thread
  Thread[CompactionExecutor:1,1,main]
  java.io.IOError: java.io.EOFException
 at
  org.apache.cassandra.io.sstable.SSTableIdentityIterator.init(SSTableIde
  ntityIterator.java:78) at
  org.apache.cassandra.io.sstable.SSTableScanner$KeyScanningIterator.next(S

Re: Strange nodetool repair behaviour

2011-04-04 Thread Mateusz Korniak
On Monday 04 of April 2011, Jonas Borgström wrote:
 I have a 6 node 0.7.4 cluster with replication_factor=3 where nodetool
 repair keyspace behaves really strange.
I think I am observing similar issue.
I have three 0.7.4 nodes with RF=3. 

After compaction I see about 7GB load in node but after running repair few 
times I see load rising [1] and enormouos data (larger few times than whole 
data) is being transferred [2].
Also I see lot of compactions, compacting a lot ( 33% or less) which should 
not happen, as my data loading pattern are only insertions (no updates, no 
deletions).
I have about 200k rows representing observable and columns representing daily 
values (now around 250 days maximum). Once once per day I insert new values to 
few CF.

Not sure if matters but:
Two nodes are in DC1 while 3rd in DC2. DCs are separated by slow internet/VPN 
link and thus size of data transferred is very important to me.
I see no problems with transferring inserted data over that link during 
insertions, only repair makes transfers so big, that repair of single node 
single CF is not able to finish in day.

I do not know how generation of merkle tree works. Is Cassandra implementation 
able to detect and transfer only whole rows ? Or row with new column added 
must be transfered as whole ? 


[1]:
192.168.3.5 Up Normal  7.85 GB 50.00%  0
10.20.1.66  Up Normal  13.7 GB 25.00%  
42535295865117307932921825928971026432
192.168.3.4 Up Normal  12.13 GB25.00%  
85070591730234615865843651857942052864

[2]:
 like twenty or more:
 progress=0/2086059022 - 0%
 progress=0/2085191380 - 0%


-- 
Mateusz Korniak


Re: cassandra disks cache on SSD

2016-04-01 Thread Mateusz Korniak
On Friday 01 April 2016 13:16:53 vincent gromakowski wrote:
> (...)  looking
> for a way to use some kind of tiering with few SSD caching hot data from
> HDD.
> I have identified two solutions (...)

We are using lvmcache for that.
Regards,
-- 
Mateusz Korniak
"(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa."
Nikos Kazantzakis - "Grek Zorba"



Re: cassandra disks cache on SSD

2016-04-03 Thread Mateusz Korniak
On Friday 01 April 2016 21:48:08 vincent gromakowski wrote:
> 2016-04-01 19:27 GMT+02:00 Mateusz Korniak <mateusz-li...@ant.gliwice.pl>:
> > We are using lvmcache for that.
>
> Can you provide me a approximate estimation of performance gain ?
 
I suspect it depends on frequently read data size. If it fits into caching 
SSDs.
In our scenario, adding one 250GB SSD to nodes owning 1.2 TB resulted in 75% 
read cache hit ratio.
It was enough to stop nodes being I/O bound and application to work without 
read timeouts.  

-- 
Mateusz Korniak
"(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa."
Nikos Kazantzakis - "Grek Zorba"



Re: Is it safe to upgrade between minor releases without first performing a repair?

2017-07-21 Thread Mateusz Korniak
On Thursday 20 of July 2017 23:27:55 Jerry wrote:
> In general it seems that repairs should be done prior to every upgrade (in
> fact they should be done at least weekly), but with a minor upgrade like
> this is it safe to upgrade without first repairing?

Depends on meaning of "is it safe" ... 

> Specifically, I'm looking to upgrade from Cassandra 2.0.11 to 2.0.17.

This is bugfix only upgrade, so in general Cassandra should perform only 
better _after_ upgrade.

On other hand, _during_ upgrade, you have to restart all nodes. Assuming 
having data inconsistency among nodes, you may get different query results 
(old/missing data, etc...) during node restarts.


-- 
Mateusz Korniak
"(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa."
Nikos Kazantzakis - "Grek Zorba"


-
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org



Re: Bootstraping a Node With a Newer Version

2017-05-16 Thread Mateusz Korniak
On Tuesday 16 of May 2017 15:27:11 Shalom Sagges wrote:
> My question is this, after reinstalling the server with the new kernel, can
> I first install the upgraded Cassandra version and then bootstrap it to the
> cluster?
No.

Bootstrap/repair may/will not work between nodes with different major 
versions.


Regards,
-- 
Mateusz Korniak
"(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa."
Nikos Kazantzakis - "Grek Zorba"


-
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org



Re: storing indexes on ssd

2018-02-12 Thread Mateusz Korniak
On Saturday 10 of February 2018 23:09:40 Dan Kinder wrote:
> We're optimizing Cassandra right now for fairly random reads on a large
> dataset. In this dataset, the values are much larger than the keys. I was
> wondering, is it possible to have Cassandra write the *index* files
> (*-Index.db) to one drive (SSD), but write the *data* files (*-Data.db) to
> another (HDD)? This would be an overall win for us since it's
> cost-prohibitive to store the data itself all on SSD, but we hit the limits
> if we just use HDD; effectively we would need to buy double, since we are
> doing 2 random reads (index + data).

Considered putting cassandra data on lvmcache?
We are using this on small (3x2TB compressed data, 128/256MB cache) clusters 
since reaching I/O limits of 2xHDD in RAID10.


-- 
Mateusz Korniak
"(...) mam brata - poważny, domator, liczykrupa, hipokryta, pobożniś,
krótko mówiąc - podpora społeczeństwa."
Nikos Kazantzakis - "Grek Zorba"


-
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org