Whyy ?
2013/3/18 Jeffrey Fass jeffreyf...@lineardesign.net
Unsubscribe
It succeeds but returns nothing as my columnfamily has only data for columns
appearing in the CREATE TABLE order. It you want to keep it you should provide
the CREATE order with sample data
--
Cyril SCETBON
On Mar 14, 2013, at 2:16 PM, aaron morton
The way compression is implemented, it is oblivious to the CF being
wide-row or narrow-row. There is nothing intrinsically less efficient in
the compression for wide-rows.
--
Sylvain
On Fri, Mar 15, 2013 at 11:53 PM, Drew Kutcharian d...@venarc.com wrote:
Hey Guys,
I remember reading
CQL can't work correctly if 2 (CQL) columns have the same name. Now, to
allow upgrade from thrift, CQL does use some default names like key for
the Row key when there isn't anything else.
Honestly I think the easiest workaround here is probably to disambiguate
things manually. Typically, you
If this is the case, Why can't we restrict key as a keyword and not to be
used as a column name?
-Vivek
On Mon, Mar 18, 2013 at 2:37 PM, Sylvain Lebresne sylv...@datastax.comwrote:
CQL can't work correctly if 2 (CQL) columns have the same name. Now, to
allow upgrade from thrift, CQL does use
If this is the case, Why can't we restrict key as a keyword and not to
be used as a column name?
This is only a problem when upgrading from thrift to CQL. Forbidding key
as a column name in thrift would be weird to say the least.
What could be done is that CQL could, when it picks the default
Imho it is probably more efficient for wide. When you decompress 8k blocks
to get at a 200 byte row you create overhead , particularly young gen.
On Monday, March 18, 2013, Sylvain Lebresne sylv...@datastax.com wrote:
The way compression is implemented, it is oblivious to the CF being
wide-row or
Hi,
I have following schema:
TimeStamp
MACAddress
Data Transfer
Data Rate
LocationID
PKEY is (TimeStamp, MACAddress). That means partitioning is on TimeStamp,
and data is ordered by MACAddress, and stored together physically (let me
know if my understanding is wrong). I have 1000
Hi Aaron,
Thank you for your answer!
My idea was to do the snapshots in the backup DC only. That way the backup
procedure will not affect the live DC. However I'm afraid that a
point-in-time recovery via the snapshots in the second DC (first restore
backup on backup DC and then repair live DC)
Hi,
I have following schema:
TimeStamp
MACAddress
Data Transfer
Data Rate
LocationID
PKEY is (TimeStamp, MACAddress). That means partitioning is on TimeStamp,
and data is ordered by MACAddress, and stored together physically (let me
know if my understanding is wrong). I have 1000
On Mon, Mar 18, 2013 at 7:33 AM, Rene Kochen
rene.koc...@emea.schange.com wrote:
Hi Aaron,
Thank you for your answer!
My idea was to do the snapshots in the backup DC only. That way the backup
procedure will not affect the live DC. However I'm afraid that a
point-in-time recovery via the
You can check which nodes hints are being held for using the JMX api. Look for
the org.apache.cassandra.db:type=HintedHandoffManager MBean and call the
listEndpointsPendingHints() function.
There are two points where hints may be stored, if the node is down when the
request started or if the
If you have the chance, could you expand on m1.xlarge being the much better
choice?
For background
http://perfcap.blogspot.co.nz/2011/03/understanding-and-using-amazon-ebs.html
it's an older post, but the idea is the same. On smaller machines you share
more, and sharing is bad. Many times
This strategy, however, doesn't seem to work with LongFamily. I've tried
various combinations of key ranges (0, 0), (1, 0), (0, Long.MaxValue), but
none iterate over the entire family. Any ideas?
What results do you see?
Have you tried Long.MinValue to Long.MaxValue ?
If you are just
What are the supported platforms for Cassandra?
Linux is the most used, there are people using Windows.
we have run into issues with some of the native calls used by Cassandra.
Could you provide some more information and/or file a ticket on
https://issues.apache.org/jira/browse/CASSANDRA
You
Brett,
Do you have some steps to reproduce the problem ? If so please create a
ticket on jira.
Cheers
-
Aaron Morton
Freelance Cassandra Consultant
New Zealand
@aaronmorton
http://www.thelastpickle.com
On 16/03/2013, at 11:40 AM, Janne Jalkanen
1. With a ConsistencyLevel of quorum, does
FBUtilities.waitForFutures() wait for read repair to complete before
returning?
No
That's just a utility method.
Nothing on the read path waits for Read Repair, and controlled by
read_repair_chance CF property, it's all async to the client request.
If a node is a seed node it will not bootstrap data from others the first time
it starts.
You can always run a nodetool repair when you think data is not where it should
be.
Cheers
-
Aaron Morton
Freelance Cassandra Consultant
New Zealand
@aaronmorton
As you see, this node thinks lots of ranges are out of sync which shouldn't
be the case as successful repairs where done every night prior to the
upgrade.
Could this be explained by writes occurring during the upgrade process ?
I found this bug which touches timestamp and tomstones which
Hi Aaron:
Thanks for your attention.
The cluster in question is a 4 node sandbox cluster we have that does not
have much traffic. I was able to chase down this issue on a CF that doesn't
change much.
That bug was flagged as fixed on 1.1.10.
They were row level deletes.
We use the nanosecond
I'm having trouble completing a repair on several of my nodes due to
errors during compaction. This is a 6 node cluster using the simple
replication strategy, rf=3, with each node assigned a single token.
I'm running nodetool repair -pr on node1, which progresses until a
specific keyspace then
Aaron,
No recipe yet. It pops up randomly and, i think due to the nature of our app,
goes away.
Seems like when we have updates that are large (10k rows in one mutate) the
problem is more likely to occur.
I'll try to workout a repro...
-Brett
On Mar 18, 2013, at 10:18 AM, aaron morton
Thanks for the info, got a couple of follow up questions, and just as
a note, this is on Cassandra 1.2.0.
I checked the flushwriter thread pool stats and saw this:
Pool NameActive Pending Completed Blocked
All time blocked
FlushWriter 1
That's what I originally thought but the OOYALA presentation from C*2012 got me
confused. Do you guys know what's going on here?
The video:
http://www.youtube.com/watch?v=r2nGBUuvVmcfeature=player_detailpage#t=790s
The slides: Slide 22 @
I feel this has come up before. I believe the compression is block based,
so just because no two column names are the same does not mean the
compression will not be effective. Possibly in their case the compression
was not effective.
On Mon, Mar 18, 2013 at 9:08 PM, Drew Kutcharian
Edward/Sylvain,
I also came across this post on DataStax's blog:
When to use compression
Compression is best suited for ColumnFamilies where there are many rows, with
each row having the same columns, or at least many columns in common. For
example, a ColumnFamily containing user data such
Hi, all:
I am a beginner of cassandra. I have a four node cassandra group.
One of my cassandra group node had been running for a week. Recently
because of too much writing and reading, it crashed.
I want to restart the node.
While I start up cassandra, it ends up with throwing
On 03/19/2013 11:22 AM, 杨辉强 wrote:
Hi, all:
I am a beginner of cassandra. I have a four node cassandra group.
One of my cassandra group node had been running for a week.
Recently because of too much writing and reading, it crashed.
I want to restart the node.
While I start up
28 matches
Mail list logo