Re: Unsubscribe

2013-03-18 Thread oualid ait wafli
Whyy ? 2013/3/18 Jeffrey Fass jeffreyf...@lineardesign.net Unsubscribe

Re: Pig / Map Reduce on Cassandra

2013-03-18 Thread cscetbon.ext
It succeeds but returns nothing as my columnfamily has only data for columns appearing in the CREATE TABLE order. It you want to keep it you should provide the CREATE order with sample data -- Cyril SCETBON On Mar 14, 2013, at 2:16 PM, aaron morton

Re: Cassandra Compression and Wide Rows

2013-03-18 Thread Sylvain Lebresne
The way compression is implemented, it is oblivious to the CF being wide-row or narrow-row. There is nothing intrinsically less efficient in the compression for wide-rows. -- Sylvain On Fri, Mar 15, 2013 at 11:53 PM, Drew Kutcharian d...@venarc.com wrote: Hey Guys, I remember reading

Re: cql query not giving any result.

2013-03-18 Thread Sylvain Lebresne
CQL can't work correctly if 2 (CQL) columns have the same name. Now, to allow upgrade from thrift, CQL does use some default names like key for the Row key when there isn't anything else. Honestly I think the easiest workaround here is probably to disambiguate things manually. Typically, you

Re: cql query not giving any result.

2013-03-18 Thread Vivek Mishra
If this is the case, Why can't we restrict key as a keyword and not to be used as a column name? -Vivek On Mon, Mar 18, 2013 at 2:37 PM, Sylvain Lebresne sylv...@datastax.comwrote: CQL can't work correctly if 2 (CQL) columns have the same name. Now, to allow upgrade from thrift, CQL does use

Re: cql query not giving any result.

2013-03-18 Thread Sylvain Lebresne
If this is the case, Why can't we restrict key as a keyword and not to be used as a column name? This is only a problem when upgrading from thrift to CQL. Forbidding key as a column name in thrift would be weird to say the least. What could be done is that CQL could, when it picks the default

Re: Cassandra Compression and Wide Rows

2013-03-18 Thread Edward Capriolo
Imho it is probably more efficient for wide. When you decompress 8k blocks to get at a 200 byte row you create overhead , particularly young gen. On Monday, March 18, 2013, Sylvain Lebresne sylv...@datastax.com wrote: The way compression is implemented, it is oblivious to the CF being wide-row or

Unable to fetch large amount of rows

2013-03-18 Thread Pushkar Prasad
Hi, I have following schema: TimeStamp MACAddress Data Transfer Data Rate LocationID PKEY is (TimeStamp, MACAddress). That means partitioning is on TimeStamp, and data is ordered by MACAddress, and stored together physically (let me know if my understanding is wrong). I have 1000

Re: Backup solution

2013-03-18 Thread Rene Kochen
Hi Aaron, Thank you for your answer! My idea was to do the snapshots in the backup DC only. That way the backup procedure will not affect the live DC. However I'm afraid that a point-in-time recovery via the snapshots in the second DC (first restore backup on backup DC and then repair live DC)

Unable to fetch large amount of rows

2013-03-18 Thread Pushkar Prasad
Hi, I have following schema: TimeStamp MACAddress Data Transfer Data Rate LocationID PKEY is (TimeStamp, MACAddress). That means partitioning is on TimeStamp, and data is ordered by MACAddress, and stored together physically (let me know if my understanding is wrong). I have 1000

Re: Backup solution

2013-03-18 Thread Aaron Turner
On Mon, Mar 18, 2013 at 7:33 AM, Rene Kochen rene.koc...@emea.schange.com wrote: Hi Aaron, Thank you for your answer! My idea was to do the snapshots in the backup DC only. That way the backup procedure will not affect the live DC. However I'm afraid that a point-in-time recovery via the

Re: 33million hinted handoffs from nowhere

2013-03-18 Thread aaron morton
You can check which nodes hints are being held for using the JMX api. Look for the org.apache.cassandra.db:type=HintedHandoffManager MBean and call the listEndpointsPendingHints() function. There are two points where hints may be stored, if the node is down when the request started or if the

Re: About the heap

2013-03-18 Thread aaron morton
If you have the chance, could you expand on m1.xlarge being the much better choice? For background http://perfcap.blogspot.co.nz/2011/03/understanding-and-using-amazon-ebs.html it's an older post, but the idea is the same. On smaller machines you share more, and sharing is bad. Many times

Re: Iterating over all keys in column family with LongType key

2013-03-18 Thread aaron morton
This strategy, however, doesn't seem to work with LongFamily. I've tried various combinations of key ranges (0, 0), (1, 0), (0, Long.MaxValue), but none iterate over the entire family. Any ideas? What results do you see? Have you tried Long.MinValue to Long.MaxValue ? If you are just

Re: Cassandra Platform Support

2013-03-18 Thread aaron morton
What are the supported platforms for Cassandra? Linux is the most used, there are people using Windows. we have run into issues with some of the native calls used by Cassandra. Could you provide some more information and/or file a ticket on https://issues.apache.org/jira/browse/CASSANDRA You

Re: secondary index problem

2013-03-18 Thread aaron morton
Brett, Do you have some steps to reproduce the problem ? If so please create a ticket on jira. Cheers - Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton http://www.thelastpickle.com On 16/03/2013, at 11:40 AM, Janne Jalkanen

Re: Waiting on read repair?

2013-03-18 Thread aaron morton
1. With a ConsistencyLevel of quorum, does FBUtilities.waitForFutures() wait for read repair to complete before returning? No That's just a utility method. Nothing on the read path waits for Read Repair, and controlled by read_repair_chance CF property, it's all async to the client request.

Re: Can't replace dead node

2013-03-18 Thread aaron morton
If a node is a seed node it will not bootstrap data from others the first time it starts. You can always run a nodetool repair when you think data is not where it should be. Cheers - Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton

Re: Lots of Deleted Rows Came back after upgrade 1.1.6 to 1.1.10

2013-03-18 Thread aaron morton
As you see, this node thinks lots of ranges are out of sync which shouldn't be the case as successful repairs where done every night prior to the upgrade. Could this be explained by writes occurring during the upgrade process ? I found this bug which touches timestamp and tomstones which

Re: Lots of Deleted Rows Came back after upgrade 1.1.6 to 1.1.10

2013-03-18 Thread Arya Goudarzi
Hi Aaron: Thanks for your attention. The cluster in question is a 4 node sandbox cluster we have that does not have much traffic. I was able to chase down this issue on a CF that doesn't change much. That bug was flagged as fixed on 1.1.10. They were row level deletes. We use the nanosecond

Errors on replica nodes halt repair

2013-03-18 Thread Dane Miller
I'm having trouble completing a repair on several of my nodes due to errors during compaction. This is a 6 node cluster using the simple replication strategy, rf=3, with each node assigned a single token. I'm running nodetool repair -pr on node1, which progresses until a specific keyspace then

Re: secondary index problem

2013-03-18 Thread Brett Tinling
Aaron, No recipe yet. It pops up randomly and, i think due to the nature of our app, goes away. Seems like when we have updates that are large (10k rows in one mutate) the problem is more likely to occur. I'll try to workout a repro... -Brett On Mar 18, 2013, at 10:18 AM, aaron morton

Re: Waiting on read repair?

2013-03-18 Thread Jasdeep Hundal
Thanks for the info, got a couple of follow up questions, and just as a note, this is on Cassandra 1.2.0. I checked the flushwriter thread pool stats and saw this: Pool NameActive Pending Completed Blocked All time blocked FlushWriter 1

Re: Cassandra Compression and Wide Rows

2013-03-18 Thread Drew Kutcharian
That's what I originally thought but the OOYALA presentation from C*2012 got me confused. Do you guys know what's going on here? The video: http://www.youtube.com/watch?v=r2nGBUuvVmcfeature=player_detailpage#t=790s The slides: Slide 22 @

Re: Cassandra Compression and Wide Rows

2013-03-18 Thread Edward Capriolo
I feel this has come up before. I believe the compression is block based, so just because no two column names are the same does not mean the compression will not be effective. Possibly in their case the compression was not effective. On Mon, Mar 18, 2013 at 9:08 PM, Drew Kutcharian

Re: Cassandra Compression and Wide Rows

2013-03-18 Thread Drew Kutcharian
Edward/Sylvain, I also came across this post on DataStax's blog: When to use compression Compression is best suited for ColumnFamilies where there are many rows, with each row having the same columns, or at least many columns in common. For example, a ColumnFamily containing user data such

Cassandra 1.2.0H startup exception

2013-03-18 Thread 杨辉强
Hi, all: I am a beginner of cassandra. I have a four node cassandra group. One of my cassandra group node had been running for a week. Recently because of too much writing and reading, it crashed. I want to restart the node. While I start up cassandra, it ends up with throwing

Re: Cassandra 1.2.0H startup exception

2013-03-18 Thread 杨辉强
On 03/19/2013 11:22 AM, 杨辉强 wrote: Hi, all: I am a beginner of cassandra. I have a four node cassandra group. One of my cassandra group node had been running for a week. Recently because of too much writing and reading, it crashed. I want to restart the node. While I start up