Re: Starting Cassandra Fauna

2010-04-14 Thread Nirmala Agadgar
Hi, Can anyone please list steps to install and run cassandra in centos. It can help me to follow and check where i missed and run correctly. Also, if i wanted to insert some data programmatically, where i need to do place the code in Fauna.Can anyone help me on this? On Mon, Apr 12, 2010 at

Re: GC options

2010-04-14 Thread Benjamin Black
FYI, G1 has been in 1.6 since u14. 2010/4/13 Peter Schüller sc...@spotify.com: I'm working on getting our latency as consistent as possible, and the gc likes to kick off 60+ms periods of unavailability for a node, which for my application leads to a reasonable number of timed out requests.

Re: GC options

2010-04-14 Thread Benjamin Black
Got it, thanks 2010/4/13 Peter Schüller sc...@spotify.com: FYI, G1 has been in 1.6 since u14. Yes, but (last time I checked) in a considerably older form. The JDK 1.7 one is more mature. -- / Peter Schuller aka scode

Re: History values

2010-04-14 Thread Sylvain Lebresne
I am new to using cassandra. In the documentation I have read, understand, that as in other non-documentary databases, to update the value of a key-value tuple, this new value is stored with a timestamp different but without entirely losing the old value. I wonder, as I can restore the

Re: History values

2010-04-14 Thread Yésica Rey
Ok, thank you very much for your reply. I have another question may seem stupid ... Cassandra has a graphical console, such as mysql for SQL databases? Regards!

Re: History values

2010-04-14 Thread Bertil Chapuis
I'm also new to cassandra and about the same question I asked me if using super columns with one key per version was feasible. Is there limitations to this use case (or better practices)? Thank you and best regards, Bertil Chapuis On 14 April 2010 09:45, Sylvain Lebresne sylv...@yakaz.com

Re: History values

2010-04-14 Thread Zhiguo Zhang
I think it is still to young, and have to wait or write your self the graphical console, at least, I don't find any until now. On Wed, Apr 14, 2010 at 10:04 AM, Bertil Chapuis bchap...@gmail.com wrote: I'm also new to cassandra and about the same question I asked me if using super columns with

Re: History values

2010-04-14 Thread aXqd
On Wed, Apr 14, 2010 at 5:13 PM, Zhiguo Zhang mikewolfx...@gmail.com wrote: I think it is still to young, and have to wait or write your self the graphical console, at least, I don't find any until now. Frankly speaking, I'm OK to be without GUI...But I am really disappointed by those so-called

server crash - how to invertigate

2010-04-14 Thread Ran Tavory
I'm running a 0.6.0 cluster with four nodes and one of them just crashed. The logs all seem normal and I haven't seen anything special in the jmx counters before the crash. I have one client writing and reading using 10 threads and using 3 different column families: KvAds, KvImpressions and

Re: RE : Re: RE : Re: Two dimensional matrices

2010-04-14 Thread Philippe
I'm confused : don't range queries such as the ones we've been discussing require using an orderedpartitionner ? Alright, so distribution depends on your choice of token. Ah yes, I get it now : with a naive orderedpartitioner, the key is associated with the node whose token is the closest

Re: History values

2010-04-14 Thread Jonathan Ellis
The closest is http://github.com/driftx/chiton On Wed, Apr 14, 2010 at 2:57 AM, Yésica Rey yes...@gdtic.es wrote: Ok, thank you very much for your reply. I have another question may seem stupid ... Cassandra has a graphical console, such as mysql for SQL databases? Regards!

Time-series data model

2010-04-14 Thread Jean-Pierre Bergamin
Hello everyone We are currently evaluating a new DB system (replacing MySQL) to store massive amounts of time-series data. The data are various metrics from various network and IT devices and systems. Metrics i.e. could be CPU usage of the server xy in percent, memory usage of server xy in MB,

Re: Time-series data model

2010-04-14 Thread Zhiguo Zhang
first of all I am a new bee by Non-SQL. I try write down my opinions as references: If I were you, I will use 2 columnfamilys: 1.CF, key is devices 2.CF, key is timeuuid how do u think about that? Mike On Wed, Apr 14, 2010 at 3:02 PM, Jean-Pierre Bergamin ja...@ractive.chwrote: Hello

Re: Time-series data model

2010-04-14 Thread Ted Zlatanov
On Wed, 14 Apr 2010 15:02:29 +0200 Jean-Pierre Bergamin ja...@ractive.ch wrote: JB The metrics are stored together with a timestamp. The queries we want to JB perform are: JB * The last value of a specific metric of a device JB * The values of a specific metric of a device between two

Re: Reading thousands of columns

2010-04-14 Thread Gautam Singaraju
Yes, I find that get_range_slices takes an incredibly long time return the results. --- Gautam On Tue, Apr 13, 2010 at 2:00 PM, James Golick jamesgol...@gmail.com wrote: Hi All, I'm seeing about 35-50ms to read 1000 columns from a CF using get_range_slices. The columns are TimeUUIDType with

Re: Reading thousands of columns

2010-04-14 Thread Jonathan Ellis
35-50ms for how many rows of 1000 columns each? get_range_slices does not use the row cache, for the same reason that oracle doesn't cache tuples from sequential scans -- blowing away 1000s of rows worth of recently used rows queried by key, for a swath of rows from the scan, is the wrong call

Re: [RELEASE] 0.6.0

2010-04-14 Thread Ted Zlatanov
On Tue, 13 Apr 2010 15:54:39 -0500 Eric Evans eev...@rackspace.com wrote: EE I leaned into it. An updated package has been uploaded to the Cassandra EE repo (see: http://wiki.apache.org/cassandra/DebianPackaging). Thank you for providing the release to the repository. Can it support a non-root

KeysCached and sstable

2010-04-14 Thread Paul Prescod
The inline docs say: ~ The optional KeysCached attribute specifies ~ the number of keys per sstable whose locations we keep in ~ memory in mostly LRU order. There are a few confusing bits in that sentence. 1. Why is keys per sstable rather than keys per column family. If I

Re: Reading thousands of columns

2010-04-14 Thread James Golick
Right - that make sense. I'm only fetching one row. I'll give it a try with get_slice(). Thanks, -James On Wed, Apr 14, 2010 at 7:45 AM, Jonathan Ellis jbel...@gmail.com wrote: 35-50ms for how many rows of 1000 columns each? get_range_slices does not use the row cache, for the same reason

Re: Lucandra or some way to query

2010-04-14 Thread Eric Evans
On Wed, 2010-04-14 at 06:45 -0300, Jesus Ibanez wrote: Option 1 - insert data in all different ways I need in order to be able to query? Rolling your own indexes is fairly common with Cassandra. Option 2 - implement Lucandra? Can you link me to a blog or an article that guides me on how to

Re: [RELEASE] 0.6.0

2010-04-14 Thread Eric Evans
On Wed, 2010-04-14 at 10:16 -0500, Ted Zlatanov wrote: Can it support a non-root user through /etc/default/cassandra? I've been patching the init script myself but was hoping this would be standard. It's the first item on debian/TODO, but, you know, patches welcome and all that. -- Eric

Re: Reading thousands of columns

2010-04-14 Thread Mike Malone
On Wed, Apr 14, 2010 at 7:45 AM, Jonathan Ellis jbel...@gmail.com wrote: 35-50ms for how many rows of 1000 columns each? get_range_slices does not use the row cache, for the same reason that oracle doesn't cache tuples from sequential scans -- blowing away 1000s of rows worth of recently

Re: History values

2010-04-14 Thread Paul Prescod
If you want to use Cassandra, you should probably store each historical value as a new column in the row. On Wed, Apr 14, 2010 at 12:34 AM, Yésica Rey yes...@gdtic.es wrote: I am new to using cassandra. In the documentation I have read, understand, that as in other non-documentary databases, to

Re: Reading thousands of columns

2010-04-14 Thread Paul Prescod
On Wed, Apr 14, 2010 at 10:31 AM, Mike Malone m...@simplegeo.com wrote: ... Couldn't you cache a list of keys that were returned for the key range, then cache individual rows separately or not at all? By blowing away rows queried by key I'm guessing you mean pushing them out of the LRU

Re: Reading thousands of columns

2010-04-14 Thread James Golick
The values are empty. It's 3000 UUIDs. On Wed, Apr 14, 2010 at 12:40 PM, Avinash Lakshman avinash.laksh...@gmail.com wrote: How large are the values? How much data on disk? On Wednesday, April 14, 2010, James Golick jamesgol...@gmail.com wrote: Just for the record, I am able to repeat this

Re: Lucandra or some way to query

2010-04-14 Thread Jake Luciani
Hi, What doesn't work with lucandra exactly? Feel free to msg me. -Jake On Wed, Apr 14, 2010 at 9:30 PM, Jesus Ibanez jesusiba...@gmail.com wrote: I will explore Lucandra a little more and if I can't get it to work today, I will go for Option 2. Using SQL will not be efficient in the

Re: Lucandra or some way to query

2010-04-14 Thread HubertChang
If you worked with Lucandra in a dedicated searching-purposed cluster, you could balanced the data very well with some effort. I think Lucandra is really a great idea, but since it needs order-preserving-partitioner, does that mean there may be some 'hot-spot' during searching? -- View this

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Ken Sandney
Large files can be split into small blocks, and the size of block can be tuned. It may increase the complexity of writing such a file system, but can be for general purpose (not only for relative small files) On Thu, Apr 15, 2010 at 10:08 AM, Tatu Saloranta tsalora...@gmail.comwrote: On Wed,

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Miguel Verde
On Wed, Apr 14, 2010 at 9:15 PM, Ken Sandney bluefl...@gmail.com wrote: Large files can be split into small blocks, and the size of block can be tuned. It may increase the complexity of writing such a file system, but can be for general purpose (not only for relative small files) Right,

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Avinash Lakshman
Exactly. You can split a file into blocks of any size and you can actually distribute the metadata across a large set of machines. You wouldn't have the issue of having small files in this approach. The issue maybe the eventual consistency - not sure that is a paradigm that would be acceptable for

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Avinash Lakshman
OPP is not required here. You would be better off using a Random partitioner because you want to get a random distribution of the metadata. Avinash On Wed, Apr 14, 2010 at 7:25 PM, Avinash Lakshman avinash.laksh...@gmail.com wrote: Exactly. You can split a file into blocks of any size and you

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Miguel Verde
On Wed, Apr 14, 2010 at 9:26 PM, Avinash Lakshman avinash.laksh...@gmail.com wrote: OPP is not required here. You would be better off using a Random partitioner because you want to get a random distribution of the metadata. Not required, certainly. However, it strikes me that 1 cluster is

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread HubertChang
Note: there are glusterfs, ceph, brtfs and luster. there is drbd. -- View this message in context: http://n2.nabble.com/Is-that-possible-to-write-a-file-system-over-Cassandra-tp4905111p4905312.html Sent from the cassandra-u...@incubator.apache.org mailing list archive at Nabble.com.

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Michael Greene
On Wed, Apr 14, 2010 at 11:01 PM, Ken Sandney bluefl...@gmail.com wrote: a fuse based FS maybe better I guess This has been done, for better or worse, by jdarcy of http://pl.atyp.us/: http://github.com/jdarcy/CassFS

Re: Is that possible to write a file system over Cassandra?

2010-04-14 Thread Ken Sandney
tried CassFS, but not stable yet, may be a good prototype to start On Thu, Apr 15, 2010 at 12:15 PM, Michael Greene michael.gre...@gmail.comwrote: On Wed, Apr 14, 2010 at 11:01 PM, Ken Sandney bluefl...@gmail.com wrote: a fuse based FS maybe better I guess This has been done, for better

Re: Starting Cassandra Fauna

2010-04-14 Thread Nirmala Agadgar
Hi, I want to insert data into Cassandra programmatically in a loop. Also i'm a newbie to Linux world and Github. Started to work on Linux for only reason to implement Cassandra.Digging Cassandra for last on week.How to insert data in cassandra and test it? Can anyone help me out on this? -

Re: Starting Cassandra Fauna

2010-04-14 Thread richard yao
try this https://wiki.fourkitchens.com/display/PF/Using+Cassandra+with+PHP On Thu, Apr 15, 2010 at 12:23 PM, Nirmala Agadgar nirmala...@gmail.comwrote: Hi, I want to insert data into Cassandra programmatically in a loop. Also i'm a newbie to Linux world and Github. Started to work on

Re: Starting Cassandra Fauna

2010-04-14 Thread Paul Prescod
There is a tutorial here: * http://www.sodeso.nl/?p=80 This page includes data inserts: * http://www.sodeso.nl/?p=251 Like: c.setColumn(new Column(email.getBytes(utf-8), ronald (at) sodeso.nl.getBytes(utf-8), timestamp)) columns.add(c); The Sample code is attached to that blog post. On

Re: Starting Cassandra Fauna

2010-04-14 Thread Nirmala Agadgar
Hi, I'm using ruby client as of now. Can u give details for ruby client.Also if possible java client. Thanks for reply. - Nirmala On Thu, Apr 15, 2010 at 10:02 AM, richard yao richard.yao2...@gmail.comwrote: try this https://wiki.fourkitchens.com/display/PF/Using+Cassandra+with+PHP On

TException: Error: TSocket: timed out reading 1024 bytes from 10.1.1.27:9160

2010-04-14 Thread richard yao
I am having a try on cassandra, and I use php to access cassandra by thrift API. I got an error like this: TException: Error: TSocket: timed out reading 1024 bytes from 10.1.1.27:9160 What's wrong? Thanks.

Re: Lucandra or some way to query

2010-04-14 Thread Jake Luciani
Lucandra spreads the data randomly by index + field combination so you do get some distribution for free. Otherwise you can use nodetool loadbalance to alter the token ring to alleviate hotspots. On Thu, Apr 15, 2010 at 2:04 AM, HubertChang hui...@gmail.com wrote: If you worked with Lucandra