thanks for the response Aaron. Our cluster has 6 nodes with 10 GB load on
each. RF=3.AMD 64 bit Blades, Quad Core, 8 GB ram, running Debian Linux.
Swap off. Cassandra 0.7.4
On Apr 6, 2011, at 2:40 AM, aaron morton wrote:
> Not that I know of, may be useful to be able to throttle thi
Cassandra 0.7.4
# nodetool -h localhost getcompactionthreshold Keyspace1 Standard1
min=4 max=32
# nodetool -h localhost setcompactionthreshold Keyspace1 Standard1 0 0
# nodetool -h localhost getcompactionthreshold Keyspace1 Standard1
min=0 max=0
Now the thresholds have changed on the JMX pannel,
I think you need to specify the port in the JMXServiceURL. The exception
indicates there is no service listening on given host and port. Also, I
guess, based on 127.0.0.1, you are running the client on same m/c as
Cassandra. If that is not the case then fix the host as well. You might want
to look
Hi Aaron,
I failed to mention that my queries have very high frequency so this is not
a data analytics problem.
That's the reason I wanted to reduce transferring a lot of data to client or
using standard map/reduce solutions that can do the job but add too much
latency.
Pig or hive are appropria
On Tue, Apr 5, 2011 at 10:45 PM, Yudong Gao wrote:
>> A better solution would be to just push the DecoratedKey into the
>> ReplicationStrategy so it can make its decision before information is
>> thrown away.
>
> I agree. So in this case, I guess the hashed based token ring is still
> preserved to
On Tue, Apr 5, 2011 at 9:59 PM, Jonathan Ellis wrote:
> On Tue, Apr 5, 2011 at 8:37 PM, Yudong Gao wrote:
>> One thing I am worrying about is how to maintain the location
>> information for each row. The current partitioner maps a key to MD5
>> hash, and it is almost impossible to control the has
Thanks Sylvain, it's very clear.
But should I still need to force major compaction regularly to clear tombstones?
I know that minor compaction clear the tombstones after 0.7, but
maximumCompactionThreshold limits the maximum number of sstable which
will be merged at once, so to GC all tombstones in
On Tue, Apr 5, 2011 at 8:37 PM, Yudong Gao wrote:
> One thing I am worrying about is how to maintain the location
> information for each row. The current partitioner maps a key to MD5
> hash, and it is almost impossible to control the hashed token by
> manipulating the value of the key. Also, main
Thanks for the reply, Jonathan!
This per-row control is exactly what I need. I will be happy to help
tackle it in the long term. Is there some further information or plan
for this issues?
One thing I am worrying about is how to maintain the location
information for each row. The current partition
Hi All,
I had written code for cassandra 0.6.3 using JMX to call
compaction,when I try to use that code to connect to 0.7.3 I get the
following
error
Exception in thread "main" java.rmi.ConnectException: Connection refused to
host: 127.0.0.1; nested exception is:
java.net.ConnectExcep
Not that I know of, may be useful to be able to throttle things. But if the
receiving node has little head room it may still be overwhelmed.
Currently there is a single thread for streaming. If we were to throttle it may
be best to make it multi threaded with a single concurrent stream per end
Couple of thoughts...
- Use only the custom secondary indexes you have built using your categories.
Pull back all items that match and then do additional filtering and sorting
client side.
- If you really need very flexible query semantics and the use case allows for
it consider using somethi
Pelops raises a RuntimeException? Can you provide more info please?
--
Dan Washusen
Make big files fly
visit digitalpigeon.com
On Tuesday, 5 April 2011 at 11:43 PM, Héctor Izquierdo Seliva wrote:
El mar, 05-04-2011 a las 09:35 -0400, Dan Hendry escribió:
> > I too have seen the out of sequence r
You'd really want https://issues.apache.org/jira/browse/CASSANDRA-2369
to control per-row. Let me know if you'd like to help tackle that.
On Tue, Apr 5, 2011 at 5:05 PM, Yudong Gao wrote:
>
> Hi,
>
> I am thinking about using Cassandra for our research project, and we
> are thinking about one int
0.6 should be wire-compatible with 0.5 -- i.e., go ahead and build
your clients with it.
there's a few changes needed though if you want to build the server
against 0.6: https://issues.apache.org/jira/browse/CASSANDRA-2412
On Tue, Apr 5, 2011 at 4:28 PM, Jeremiah Jordan
wrote:
> Anyone know if 0
Hi,
I am thinking about using Cassandra for our research project, and we
are thinking about one interesting feature.
Our setup has multiple datacenters located in different geography
locations. Data is accessed with predictable patterns. Think of
something like Craigslist, data objects correspon
Anyone know if 0.7.4 will work with thirft 0.6? Or do I have to keep
thrift 0.5 around to use it?
Thanks!
Jeremiah Jordan
Application Developer
Morningstar, Inc.
Morningstar. Illuminating investing worldwide.
+1 312 696-6128 voice
jeremiah.jor...@morningstar.com
If you can reach your jmx ip/port, you can use any jmx client to start a
compaction. Use jconsole to connect to your jmx ip/port and then navigate to
mbeans->org.apache.cassandra.db->columnfamilies-> ->operations
Underneath there you can invoke a bunch of methods including compaction.
Sridha
Well, since my last post, about 10 minutes later, the node goes into bootstrap
mode. It's kind of worrying that a lot of time goes by where it seems like
nothing is happening, then all of a sudden things get going again.
22,584 keys. Time: 20,276ms.
INFO [HintedHandoff:1] 2011-04-05 22:29:2
I added a node to the cluster and I am having a difficult time reassigning the
new tokens.
It seems after a while nothing shows up in the new node's logs and it just
stays in status "Leaving". "nodetool netstats" on all nodes shows "Nothing
streaming to/from".
There is no activity in the
> Is there a way I can run compaction on the cassandra cluster from a
> machine where cassandra is not installed.I have a cluster of 6 machines
> but I want to run compaction on them from a different machine which does not
> have cassandra installed.
"nodetool -h your-remote-host compact"
> - Different collectors: -XX:+UseParallelGC -XX:+UseParallelOldGC
Unless you also removed the -XX:+UseConcMarkSweepGC I *think* it takes
precedence, so that the above options would have no effect. I didn't
test. In either case, did you definitely confirm CMS was no longer
being used? (Should be p
Hi All,
Is there a way I can run compaction on the cassandra cluster from a
machine where cassandra is not installed.I have a cluster of 6 machines
but I want to run compaction on them from a different machine which does not
have cassandra installed.
Thanks
Anurag
> note the virtual mem used 20.6 GB ?! and Shared 8.4 GB ?!
http://wiki.apache.org/cassandra/FAQ#mmap
--
/ Peter Schuller
I've seen the other posts about memory consumption, but I'm seeing some weird
behavior with 0.7.4 with 5 GB heap size (64 bit system with 8 GB ram
total)...
note the virtual mem used 20.6 GB ?! and Shared 8.4 GB ?!
PID USER PR NI VIRT RES SHR S %CPU %MEMTIME+ COMMAND
> Would you recommend to disable system swap as a rule? I'm running on Debian
> 64bit and am seeing light swapping:
I'm not Jonathan, but *yes*. I would go so far as to say that
disabling swap is a good rule of thumb for *most* production systems
that serve latency sensitive traffic. For a mach
Hi Jonathan -
Would you recommend to disable system swap as a rule? I'm running on Debian
64bit and am seeing light swapping:
total used free sharedbuffers cached
Mem: 8003 7969 33 0 0 4254
-/+ buffers/cache:
This is a minor followup to this thread which includes required context:
http://www.mail-archive.com/user@cassandra.apache.org/msg09279.html
I haven't solved the problem, but since negative results can also be
useful I thought I would share them. Things I tried unsuccessfully (on
individual node
On Tue, Apr 5, 2011 at 9:50 AM, Jonathan Colby wrote:
> Hi experts,
>
> We have a cluster of 4 nodes across 2 data centers and we are using
> OldNetworkTopology to balance the replicas (RF=3).
>
> Question: Can we afford to lose 1 node in a data center and still have
> Quorum work?
Insomuch
Other than you will have to watch log to know when it is done, no.
On Tue, Apr 5, 2011 at 9:54 AM, Jonathan Colby wrote:
>
> when doing a "nodetool move ", after about 15 minutes I got the below
> exception. The cassandra log seems to indicate that the move is still
> ongoing. Is this anyth
Step 1: disable swap.
2011/4/5 Héctor Izquierdo Seliva :
> Update with more info:
>
> I'm still running into problems. Now I don't write more than 100 columns
> at a time, and I'm having lots of Stop-the-world gc pauses.
>
> I'm writing into three column families, with memtable_operations = 0.3
>
> Can someone comment on this ? Or is the question too vague ?
Honestly yeah I couldn't figure out what you were asking ;) What
specifically about the diagram are you trying to convey?
--
/ Peter Schuller
AJ,
One issue that I found in using load balancer in front of cassandra nodes is
that a single node might become bogged down by compaction, or other actions
unrelated to the client. If the load balancer does not pick this up in
time, it might route client requests to the node that is temporarily
Can someone comment on this ? Or is the question too vague ?
Thanks.
On Wed, Mar 30, 2011 at 3:58 PM, A J wrote:
> Does the following load balancing scenario look reasonable with cassandra ?
> I will not be having any app servers.
>
> http://dl.dropbox.com/u/7258508/2011-03-30_1542.png
>
> Thank
On Tue, Apr 5, 2011 at 12:01 AM, Maki Watanabe wrote:
> Hello,
> On reading O'Reilly's Cassandra book and wiki, I'm a bit confusing on
> nodetool repair and compact.
> I believe we need to run nodetool repair regularly, and it synchronize
> all replica nodes at the end.
> According to the document
Update with more info:
I'm still running into problems. Now I don't write more than 100 columns
at a time, and I'm having lots of Stop-the-world gc pauses.
I'm writing into three column families, with memtable_operations = 0.3
and memtable_throughput = 64. There is now swapping, and full GCs are
When doing a move, decommission, loadbalance, etc. data is streamed to the
next node in such a way that it really strains the receiving node - to the
point where it has a problem serving requests.
Any way to throttle the streaming of data?
I'm still running into problems. Now I don't write more than 100 columns
at a time, and I'm having lots of Stop-the-world gc pauses.
I'm writing into three column families, with memtable_operations = 0.3
and memtable_throughput = 64.
Is any of this wrong?
> > -Original Message-
> > From
On 04/05/2011 09:57 AM, Jonathan Ellis wrote:
> On Tue, Apr 5, 2011 at 8:54 AM, Jonathan Ellis wrote:
>> Adjusting indexinterval is unlikely to be useful on very narrow rows.
>> (Its purpose is to make random access to _large_ rows doable.)
>
> Whoops, that's column_index_size_in_kb.
>
> I'd pla
when doing a "nodetool move ", after about 15 minutes I got the below
exception. The cassandra log seems to indicate that the move is still
ongoing. Is this anything to worry about?
Exception in thread "main" java.rmi.UnmarshalException: Error unmarshaling
return header; nested exception
Hi experts,
We have a cluster of 4 nodes across 2 data centers and we are using
OldNetworkTopology to balance the replicas (RF=3).
Question:Can we afford to lose 1 node in a data center and still have
Quorum work?
My understanding is that OldNetworkTopology writes 2 to the "local" data ce
On 04/05/2011 03:49 PM, Jonathan Ellis wrote:
> Sounds like https://issues.apache.org/jira/browse/CASSANDRA-2324
Yes, that sounds like the issue I'm having. Any chance for a fix for
this being backported to 0.7.x?
Anyway, I guess I might as well share the test case I've used to
reproduce this pr
Try the patch at https://issues.apache.org/jira/browse/CASSANDRA-2417
On Mon, Apr 4, 2011 at 11:18 AM, Or Yanay wrote:
> Hi All,
>
>
>
> I have upgraded from 0.7.0 to 0.7.4, and while running scrub I get the
> following exception quite a lot:
>
>
>
> java.lang.AssertionError: mmap segment underfl
On Tue, Apr 5, 2011 at 8:54 AM, Jonathan Ellis wrote:
> Adjusting indexinterval is unlikely to be useful on very narrow rows.
> (Its purpose is to make random access to _large_ rows doable.)
Whoops, that's column_index_size_in_kb.
I'd play w/ keycache before index_interval personally. (If you c
I am considering treating the node as a dead node, delete it's data files and
bootstrap from scratch.
Is that a reasonable approach? Is there a way for me to identify the bad file/s
and remove it/them?
From: Or Yanay [mailto:o...@peer39.com]
Sent: Monday, April 04, 2011 7:19 PM
To: user@cassandr
I've been seeing this EOF in my system.log file occasionally as well.
Doesn't seem to be causing harm:
ERROR [Thread-22] 2011-04-05 20:37:22,562 AbstractCassandraDaemon.java
(line 112) Fatal exception in thread Thread[Thread-22,5,main]
java.io.IOError: java.io.EOFException
at
org.apache.c
Adjusting indexinterval is unlikely to be useful on very narrow rows.
(Its purpose is to make random access to _large_ rows doable.)
On Mon, Apr 4, 2011 at 8:38 PM, Chris Burroughs
wrote:
> I have a case with very narrow rows. As such I have a large row cache
> that does nicely handles > 50% of
You're right, the advice there is not helpful.
On Mon, Apr 4, 2011 at 10:13 AM, shimi wrote:
> It make sense to me that compaction should solved this as well since
> compaction creates new index files.
> Am I missing something here?
> WARN [CompactionExecutor:1] 2011-04-04 14:50:54,105 Compaction
Sounds like https://issues.apache.org/jira/browse/CASSANDRA-2324
On Mon, Apr 4, 2011 at 7:46 AM, aaron morton wrote:
> Jonas, AFAIK if repair completed successfully there should be no streaming
> the next time round. This sounds odd please look into it if you can.
>
> Can you run at DEBUG loggin
Oops, I saw "EOFException" and jumped to "scrub."
But your EOF is coming from TCP. Something (almost certainly a
non-cassandra process) is connecting to the internal Cassandra
communication port (the one that defaults to 7000) and disconnecting.
On Mon, Apr 4, 2011 at 4:14 AM, Kazuo YAGI wrote:
El mar, 05-04-2011 a las 09:35 -0400, Dan Hendry escribió:
> I too have seen the out of sequence response problem. My solution has just
> been to retry and it seems to work. None of my mutations are THAT large (<
> 200 columns).
>
> The only related information I could find points to a thrift/u
sounds like they haven't been munmapped yet. try forcing a GC.
On Sat, Apr 2, 2011 at 5:38 AM, Roland Gude wrote:
>
> Hi,
>
> The open file limit is 1024
> Sstable count is somewhere around 20 or so thread count is in the same order
> of magnitude I guess
> But lsof shows that deleted sstables
I too have seen the out of sequence response problem. My solution has just been
to retry and it seems to work. None of my mutations are THAT large (< 200
columns).
The only related information I could find points to a thrift/ubuntu bug of some
kind (http://markmail.org/message/xc3tskhhvsf5awz7
That should work, but "nodetool drain" before shutdown to avoid
confusing the commitlog.
On Tue, Apr 5, 2011 at 3:56 AM, Jean-Yves LEBLEU wrote:
> Hi all,
>
> We are running .6.12, is there any particular precaution to rename the
> keyspace, is it enough to shutdown cassandra, update storag-conf.
Hi everyone. I'm having trouble while inserting big amounts of data into
cassandra. I'm getting this exception:
batch_mutate failed: out of sequence response
I'm gessing is due to very big mutates. I have made the batch mutates
smaller and it seems to be behaving. Can somebody shed some light?
T
Hi all,
We are running .6.12, is there any particular precaution to rename the
keyspace, is it enough to shutdown cassandra, update storag-conf.xml,
rename data directory and start cassandra again.
Thanks for your help.
Jean-Yves
I think the key thing to remember is that compaction is performed on
*similar* sized sstables. so it makes sense that over time this will have a
cascading effect. I think by default it starts out with compacting 4
flushed sstables, then the cycle begins.
On Apr 4, 2011 3:42pm, shimi wrote:
> Okay, I see. But isn't there a big issue for scaling here ?
> Imagine that I am the developper of a certain very successful website : At
> year 1 I need 20 CF. I might need to have 8Gb of RAM. Year 2 I need 50 CF
> because I added functionalities to my wonderful webiste will I need 20 Gb of
> RAM
58 matches
Mail list logo