Have checked it's all in the 0.6 branch and asked the devs for a 0.6.12 release. Will let you know how it goes.cheersAaronOn 16 Feb, 2011,at 08:38 AM, Aaron Morton aa...@thelastpickle.com wrote:I worked on that ticket, will try to chase it up.AaronOn 15/02/2011, at 2:01 PM, Gregory Szorc
Hi,
just started an upgrade on a single node of a live production cluster,
and did a nodetool repair/compact/cleanup in the logs I see
exceptions, is this normal?
ERROR [ReadStage:31] 2011-02-16 08:18:38,094
DebuggableThreadPoolExecutor.java (line 103) Error in ThreadPoolExecutor
is there a limit or a factor to take into account when the number of rows in
a CF exceeds a certain number? i see the columns for a row can get upwards
of 2 billion ... can i have 2 billion rows without much issue?
--
Sasha Dolgy
sasha.do...@gmail.com
I have same question.
I read the source code of NetworkTopologyStrategy, seems it always put replica
on the first nodes on the ring of the DC.
If I am misunderstand, It seems those nodes will became hot spot.
Why NetworkTopologyStrategy works that way? is there some alternative can avoid
this
Sky is the limit.
Columns in a row are limited to 2 billion because the size of a row is
recorded in a java int. A row must also fit on one node, so this also limit
in a way the size of a row (if you have large values, you could be limited
by this factor much before reaching 2 billions columns).
Thanks Sylvain
On Wed, Feb 16, 2011 at 10:05 AM, Sylvain Lebresne sylv...@datastax.comwrote:
Sky is the limit.
Columns in a row are limited to 2 billion because the size of a row is
recorded in a java int. A row must also fit on one node, so this also limit
in a way the size of a row (if
Hello
Here is my output from ps aux:
root 737 0.0 0.0 8940 804 ?S04:34 0:00 nimbus(hdb)
root 768 0.0 0.0 0 0 ?S04:34 0:00
[flush-202:0]
root 1368 43.1 70.5 8248976 5935852 ? Sl 04:45 297:47
/usr/local/jdk1.6.0_23/bin/java -ea
On 02/15/2011 07:15 PM, Sylvain Lebresne wrote:
On Tue, Feb 15, 2011 at 7:10 PM, ruslan usifov ruslan.usi...@gmail.com
mailto:ruslan.usi...@gmail.com wrote:
It will be great if patch appear very quick
patch attached here: https://issues.apache.org/jira/browse/CASSANDRA-2165
Hi again
The first thing you should try is to restart the node that is not
syncing. That will reset the version it gossips, which was not always
getting updated (fixed in CASSANDRA-2083).
Gary.
On Tue, Feb 15, 2011 at 18:25, Shu Zhang szh...@mediosystems.com wrote:
Hi, a node in my cassandra cluster
Yes, I read the same and it sounded weird.
*Note that with RackAwareStrategy, succeeding nodes along the ring should
alternate data centers to avoid hot spots. For instance, if you have nodes
A, B, C, and D in increasing Token order, and instead of alternating you
place A and B in DC1, and C and
Each of your 21 column families will have its own memtable if you have
the default memtable settings your memory usage will grow quite large
over time. Have you tuned down your memtable size?
Which config parameter make this? binary_memtable_throughput_in_mb?
Dropping a CF requires the compaction lock, so if a compaction is in
progress it needs to wait for it to finish.
On Tue, Feb 15, 2011 at 8:19 PM, William R Speirs bill.spe...@gmail.com wrote:
What would/could take so long for the nodes to agree? It's a small cluster
(7 nodes) all on local LAN
Yes I didn't see there was 2 different parameters. I was personally setting
( in cassandra 0.6.6 ) MemTableThoughputInMB, but I don't know what
BinaryMemtableThroughtputInMB is.
And I take this opportunity to ask a question :
If you have a small amount of data per key so that your memtable is
It does look a lot like 1932. Make sure everything is really running
0.7.2, 0.7.0 can't read data files created by 0.7.1+.
If the versions are ok, take a snapshot, then compact, and see if the
problem still occurs with all files on the [compacted] new format.
2011/2/16 Jonas Borgström
2011/2/16 Victor Kabdebon victor.kabde...@gmail.com
Ruslan I have seen your question in the other mail and I have the same
problem. How many CF do you have ?
16
Someone please correct me if I am wrong, but I think the overhead you can
expect is something like :
16* MemTableThroughtPutInMB
but I don't know when BinaryMemTableThroughputInMb come into account..
2011/2/16 ruslan usifov ruslan.usi...@gmail.com
2011/2/16 Victor Kabdebon
Just wanted to let people who follow the user list know that if there is
interest in something like plugins, triggers, or coprocessors on the
server-side with Cassandra, the ticket to follow or get involved with (code,
comments, etc) is CASSANDRA-1311:
Thanks for the confirmation. Interesting alternatives to avoid random
coordinator.
Are there any blogs/writeups of they (primary node as co-ordinator) been
used in production scenarios. I googled but could not find anything
relevant.
On Wed, Feb 16, 2011 at 3:25 AM, Oleg Anastasyev
Hi,
We are interested in a multi-tenancy environment, that may consist of up to
hundreds of data centers. The current design requires cross rack and cross
DC replication. Specifically, the per-tenant CFs will be replicated 6 times:
in three racks, with 2 copies inside a rack, the racks will be
On 02/16/2011 03:54 PM, Jonathan Ellis wrote:
It does look a lot like 1932. Make sure everything is really running
0.7.2, 0.7.0 can't read data files created by 0.7.1+.
All nodes are running 0.7.2
If the versions are ok, take a snapshot, then compact, and see if the
problem still occurs
Hi,
unfortiunately i made a copy paste error and created two indexes called
myindex on different columnfamilies.
What can I do to fix this?
Below the output from describe keyspace
ColumnFamily: A
Columns sorted by: org.apache.cassandra.db.marshal.TimeUUIDType
Row cache size /
On Wed, Feb 16, 2011 at 7:12 AM, Victor Kabdebon
victor.kabde...@gmail.com wrote:
Someone please correct me if I am wrong, but I think the overhead you can
expect is something like :
MemTableThroughtPutInMB * JavaOverheadFudgeFactor * maximum likely
number of such memtables which might exist
Hi Gaurav,
Thanks for the reply...I did look at the cages framework and I see that it
provides some functionality for locking and atomic writes for multiple keys.
My question was that do people rely on these kind of frameworks - if so, is
cages the only one or are there others as well...and if
Adding to the above message of mine, bulk atomic writes (or transaction
blocks) tend to be a common pattern in rich enterprise applications where
business logic requires all or no writes on set of entities. There may not
be a need for all ACID properties but atleast atomicity and durability are
a
Ritesh,
There don't seem to be any common best practices to do this. I think the
reason is that by adding transaction semantics on top of Cassandra you're
throwing away the most important properties of Cassandra. The effects of a
transaction/locking layer:
- A centralized performance bottleneck
0.7.2 is just about to be released.
Aaron
On 16/02/2011, at 11:42 AM, buddhasystem potek...@bnl.gov wrote:
Thank you Attila!
We will indeed have a few months of breaking in. I suppose I'll
keep my fingers crossed and see that 0.7.X is very stable. So I'll
deploy 0.7.1 -- I will need to
Hi Tyler,
Thanks for you input, but we have a set key that consists of
name:timestamp that we are using.. and we need to also retrieve the
oldest data as well..
Thanks
On 2/15/2011 9:07 PM, Tyler Hobbs wrote:
But wouldn't using timestamp as row keys cause conflicts?
Depending on client
Thanks robert, and do you know if there is a way to control the maximum
likely number of memtables ? (I'd like to cap it at 2)
2011/2/16 Robert Coli rc...@digg.com
On Wed, Feb 16, 2011 at 7:12 AM, Victor Kabdebon
victor.kabde...@gmail.com wrote:
Someone please correct me if I am wrong, but I
JVM heap memory is controlled by the settings in conf/Cassandra-env.sh
Memory mapped files will use additional virtual memory, is controlled in
conf/Cassandra.yaml disk_access_mode
Aaron
On 17/02/2011, at 2:19 AM, ruslan usifov ruslan.usi...@gmail.com wrote:
Hello
Here is my output from
thanks Dave. So general purpose mechanism would be use RDBMS for data that
requires locking semantics or use something like Cages on top of Cassandra
and then use Cassandra for data mining/high throughput read queries and
writable data that does not require transactions?
Are there any sample
Ritesh,
The gist of Dave's contention is that Casandra adds value in spite of the
lack of transactions. However, that need not mean that it can be used for
Enterprise applications. Transaction semantics needs to be re-imagined
within the capabilities of this new kind of database infrastructure,
On Wed, Feb 16, 2011 at 11:22 AM, Victor Kabdebon
victor.kabde...@gmail.com wrote:
Thanks robert, and do you know if there is a way to control the maximum
likely number of memtables ? (I'd like to cap it at 2)
That likely number of memtables is the number of memtables which :
a) have been
There is no rename, but update column family though the cli or api with just
the renamed index should work.
The code says it will remove old and add new indexes based on their name.
I'm not sure if the name is used for anything other than identifying the index
inside the CF. Are the duplicate
Thanks a lot Anthony. That does help me think on possible options...
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Patterns-for-writing-enterprise-applications-on-cassandra-tp6030077p6033316.html
Sent from the
Hi,
I'm trying to add a new secondary index on a CF which already have a secondary
index by calling update column family.
But the CF was left with just the new secondary index and the old index
disappeared. AFAIK, the only way to keep both indices was to include the old
secondary index in the
Thanks for you input, but we have a set key that consists of name:timestamp
that we are using.. and we need to also retrieve the oldest data as well..
Then you'll need to denormalize and store every row three ways: timestamp,
inverted timestamp, and normal, if you want to be able to access
+1 on avoiding OPP
On Wed, Feb 16, 2011 at 3:27 PM, Tyler Hobbs ty...@datastax.com wrote:
Thanks for you input, but we have a set key that consists of name:timestamp
that we are using.. and we need to also retrieve the oldest data as well..
Then you'll need to denormalize and store every
I sidestep this problem by using a Python script (pycassa-based) where I
configure my CFs. This way, it's reproducible and documented.
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/create-additional-secondary-index-tp6033574p6033683.html
Sent
I found the following paper (PDF) very helpful in shaping my thoughts
about what it means to build systems without transactions.
http://www.ics.uci.edu/~cs223/papers/cidr07p15.pdf
LIfe Beyond Distributed Transactions: an Apostate's Opinion by Pat Helland
On Wed, Feb 16, 2011 at 2:00 PM,
Thank you guys ...
On 2/16/2011 1:36 PM, Matthew Dennis wrote:
+1 on avoiding OPP
On Wed, Feb 16, 2011 at 3:27 PM, Tyler Hobbsty...@datastax.com wrote:
Thanks for you input, but we have a set key that consists of name:timestamp
that we are using.. and we need to also retrieve the oldest
By any chance are there companies that provide support for Cassandra ?
Consult on setup and configuration and annual support packages ?
riptano - contact matt pfeil
mike
2011/2/17 A J s5a...@gmail.com
By any chance are there companies that provide support for Cassandra ?
Consult on setup and configuration and annual support packages ?
--
bayoda.com - Professional Online Backup Solutions for Small and Medium Sized
They also offer great training sessions. Have a look at their site for more
information: http://www.datastax.com/about-us
On Wed, Feb 16, 2011 at 3:13 PM, Michael Widmann
michael.widm...@gmail.comwrote:
riptano - contact matt pfeil
mike
2011/2/17 A J s5a...@gmail.com
By any chance are
Re Anthony's statement:
So it can be done and frameworks like CAGES are showing a way forward. At
the heart of it, there will need to be a Two-Phase commit type protocol
coordinator that sits in front of Cassandra. Of which - one can be sure -
there
will be many implementations / best
Dave,
I agree with you, mostly ;) !!
While the reference to 2PC is a tad misplaced here - the idea is that the
paradigm of transactions might have to get redefined or - better still -
broadened to include protocols that the provide similar guarantees in an
eventually consistent dispensation.
CASSANDRA-2165[1] became evident almost as soon as 0.7.1 released, and
it's ugly enough that we didn't want to wait.
Be sure you've read the changelog[2] and release notes[3], and let us
know[4] if you encounter any problems.
Thanks!
[1]: https://issues.apache.org/jira/browse/CASSANDRA-2165
I do the range query by the hector client when the state and birht_date
columns are both indexed, it show the error:
The method addGtExpression(String,String) in the type
IndexedSlicesQueryString,String,String is not applicable for the
arguments(String,long)
The code is as below:
That's because in createIndexedSlicesQuery(), you have used SS as the
valueSerializer, and then in addGtExpression(birth_date, 1970L), you pass
a long as the value.
@see:
https://groups.google.com/group/hector-users/browse_thread/thread/520c5332dedfca4a
2011/2/17 Raoyixuan (Shandy)
Not really.
I also tried the code as below:
StringSerializer ss = StringSerializer.get();
LongSerializer ls = LongSerializer.get();
IndexedSlicesQueryString, String, Long indexedSlicesQuery =
HFactory.createIndexedSlicesQuery(keyspace, ss, ss, ls);
indexedSlicesQuery.setColumnNames(full_name,
addGtEqual???
You are still making the same mistake.
You are now saying that your values are going to be long (by passing LS as
the valueSerializer), but in *addEqualsExpression*(), you are passing a
String value.
2011/2/17 Raoyixuan (Shandy) raoyix...@huawei.com
Not really.
I also tried
But my requirement is get the data by birth_date and state , how to achieve it
by hector client?
From: Roshan Dawrani [mailto:roshandawr...@gmail.com]
Sent: Thursday, February 17, 2011 10:48 AM
To: user@cassandra.apache.org
Subject: Re: indexedslicesQuery class error
addGtEqual???
You are
You didn't see the previously given link?
@see:
https://groups.google.com/group/hector-users/browse_thread/thread/520c5332dedfca4a
2011/2/17 Raoyixuan (Shandy) raoyix...@huawei.com
But my requirement is get the data by birth_date and state , how to
achieve it by hector client?
*From:*
I modify it as below
StringSerializer ss = StringSerializer.get();
IndexedSlicesQueryString, String, String indexedSlicesQuery =
HFactory.createIndexedSlicesQuery(keyspace, ss, ss, ss);
indexedSlicesQuery.setColumnNames(full_name, birth_date, state);
indexedSlicesQuery.addGtExpression(birth_date,
Thanks .
I change ss to bytebufferserializer. It works fine.
From: Raoyixuan (Shandy) [mailto:raoyix...@huawei.com]
Sent: Thursday, February 17, 2011 11:16 AM
To: user@cassandra.apache.org
Subject: RE: indexedslicesQuery class error
I modify it as below
StringSerializer ss =
See the following mail thread:
http://www.mail-archive.com/user@cassandra.apache.org/msg10183.html
In short, running nodetool compact should clear it up.
On Wed, Feb 16, 2011 at 11:23 PM, Casey Deccio ca...@deccio.net wrote:
I recently upgraded to 0.7.2 from 0.7.0, and now when I run my
On Wed, Feb 16, 2011 at 10:01 PM, Nate McCall n...@datastax.com wrote:
See the following mail thread:
http://www.mail-archive.com/user@cassandra.apache.org/msg10183.html
In short, running nodetool compact should clear it up.
Thanks for the pointer! I ran nodetool compact on my nodes, and
Hello,
We were occasionally experiencing client exceptions with 0.6.3, so we
upgraded to 0.7.0 a couple weeks ago, but unfortunately we now get
more client exceptions, and more frequently. Also, occasionally
nodetool ring will show a node Down even though cassandra is still
running and the node
57 matches
Mail list logo