A general thought, if you're using AWS for this, I'd strongly recommend you
consider using OpsWorks and custom chef recipes for your node deployment if
its an option for you. The easy of provisioning new nodes without the need
for snapshotting is certainly worth the hassle, and there are already
Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com
On 30/12/2013, at 1:28 pm, David McNelis dmcne...@gmail.com wrote:
I am currently running a cluster with 1.2.8. One of my larger column
families on one of my nodes has keyspace-tablename-ic--Data.db
I am currently running a cluster with 1.2.8. One of my larger column
families on one of my nodes has keyspace-tablename-ic--Data.db with a
modify date in August.
Since august we have added several nodes (with vnodes), with the same
number of vnodes as all the existing nodes.
As a result,
It is a little more involved than just changing the heap size. Every
cluster is different, so there isn't much of a set formula. Some areas to
look into, though:
**Caveat, we're still running in the 1.2 branch and 2.0 has some
differences in what is on versus off heap memory usage, but the
Stable loader is the way to go to load up the new cluster.
On Tuesday, September 17, 2013, Juan Manuel Formoso wrote:
If your shuffle succeeds, you will be the first reported case of
shuffle succeeding on a non-test cluster.
Awesome! :O
I'll try to migrate to a new cluster then.
Any
As Rob mentioned, no one (myself included) has successfully used shuffle in
the wild (that I've heard of).
Shuffle is *supposed* to be a transparent background process... and is
designed, in theory, to take a long time to run (weeks is the right way to
think of it).
Be sure to keep an eye on
Morning,
I'm getting the following error (21 node cluster running 1.2.8)
FSReadError in
/var/cassandra/data/et/http_request/ks-mycql3table-ic-1799-Data.db
at
org.apache.cassandra.io.compress.CompressedRandomAccessReader.reBuffer(CompressedRandomAccessReader.java:93)
at
Looks to be the case, getting an IO error when trying to cp the file. That
is unfortunate. On the bright side, now we at least have a more narrow
scope of the problem's source.
On Mon, Sep 9, 2013 at 12:54 PM, Robert Coli rc...@eventbrite.com wrote:
On Mon, Sep 9, 2013 at 6:15 AM, David
You would, however, want to clear the snapshot folder afterword, right? I
thought that truncate, like drop table, created a snapshot (unless that
feature had been disabled in your yaml.
On Thu, Aug 29, 2013 at 6:51 PM, Robert Coli rc...@eventbrite.com wrote:
On Thu, Aug 29, 2013 at 3:48 PM, S
Completely understandable. Thank you for all this work, Alex, et. al.
On Fri, Aug 9, 2013 at 3:27 PM, Alex Popescu al...@datastax.com wrote:
On Fri, Aug 9, 2013 at 10:12 AM, David McNelis dmcne...@gmail.com wrote:
Is DevCenter a project that might end up open sourced? The original blog
Morning folks,
For the last couple of days all of my nodes (17, all running 1.2.8) have
been stuck at various percentages of completion for compacting
system.hints. I've tried restarting the nodes (including a full rolling
restart of the cluster) to no avail.
When I turn on Debugging I am
rpc_server_type are you using?
On Wed, Aug 7, 2013 at 6:14 AM, David McNelis dmcne...@gmail.com wrote:
Morning folks,
For the last couple of days all of my nodes (17, all running 1.2.8) have
been stuck at various percentages of completion for compacting
system.hints.
I've tried restarting
-n 'm' | nc localhost 9160
On Wed, Aug 7, 2013 at 11:11 AM, David McNelis dmcne...@gmail.com wrote:
Nate,
We had a node that was flaking on us last week and had a lot of handoffs
fail to that node. We ended up decommissioning that node entirely. I
can't
find the actual error we were
Hey folks,
Because 1.2.8 hasn't been pushed to the repo yet, I see that I can pick
up the package at http://people.apache.org/~eevans/ and install it
manually. This is great. I'm wondering though, is there a place where
I can pick up Debian packages for older releases? I definitely prefer
the
Thanks, fwiw, did I just blatantly miss some documentation saying those
existed there?
On Thu, Aug 1, 2013 at 3:32 PM, Blair Zajac bl...@orcaware.com wrote:
On 08/01/2013 12:27 PM, David McNelis wrote:
Hey folks,
Because 1.2.8 hasn't been pushed to the repo yet, I see that I can pick
up
I second Romain, do the upgrade and make sure the health is good first.
If you have or plan to have a large number of nodes, you might consider
using fewer than 256 as your initial vnodes amount. I think that number is
inflated from reasonable in the docs, as we've had some people talk about
Eric,
Unfortunately if you've got a non-vnode cluster and are trying to convert,
you are likely going to at least want, if not have to, run shuffle. It
isn't a pleasant situation when you run into that because in order for the
shuffle to execute safely and successfully you need to have
I think you'd just be better served with just a little different primary
key.
If your primary key was (user_id, created_at) or (user_id, created_at,
question_id), then you'd be able to run the above query without a problem.
This will mean that the entire pantheon of a specific user_id will be
) row? I thought having fat or big rows
was bad. I worked with Cassandra 0.6 at my previous job and given the
nature of our work, we would sometimes generate these fat rows... at
which point Cassandra would basically shit the bed.
Thanks for the help.
On Wed, Jun 19, 2013 at 12:26 PM, David
://www.thelastpickle.com
On 18/06/2013, at 2:59 PM, David McNelis dmcne...@gmail.com wrote:
I have a node in my ring (1.2.5) that when it was set up, had the wrong
number of vnodes assigned (double the amount it should have had).
As a result, and because we can't reduce the number of vnodes
I have a node in my ring (1.2.5) that when it was set up, had the wrong
number of vnodes assigned (double the amount it should have had).
As a result, and because we can't reduce the number of vnodes on a machine
(at least at this point), I need to decommission the node.
The problem is that
Another thing to keep in mind when doing this with CQL is to take into
account the ordering partitioner you may or may not be using. If you're
using one you'll need to make sure that if you have a larger number of rows
for the partitioner key than your query limit, then you can end up in a
So, I had 7 nodes that I set up using vnodes, 256 tokens each, no problem.
I added two 512 token nodes, no problem, things seemed to balance.
The next 3 nodes I added, all at 256 tokens, and they have a cumulative
load of 116mb (where as the other nodes are at ~100GB and ~200GB (256 and
512
thing to do is use COMPACT STORAGE but that may not suite all
use cases http://www.datastax.com/docs/1.2/cql_cli/cql/CREATE_TABLE
Cheers
-
Aaron Morton
Freelance Cassandra Consultant
New Zealand
@aaronmorton
http://www.thelastpickle.com
On 20/04/2013, at 4:36 PM, David
In order to do a query like that you'll need to have a timestamp/date as
the second portion of the primary key.
You'll only be able to do queries where you already know the key. Unless
you're using an OrderPreservingPartitioner, there is no way to get a
continuous set of information back based
Was trying to do a test of writing SSTs for a CQL3 table. So I created the
following table:
CREATE TABLE test_sst_load (
mykey1 ascii,
mykey2 ascii,
value1 ascii,
PRIMARY KEY (mykey1, mykey2)
)
I then set up my writer like so: (moved to gist:
https://gist.github.com/dmcnelis/5424756 )
I had a situation earlier where my shuffle failed after a hard disk drive
filled up. I went through and disabled shuffle on the machines while
trying to get the situation resolved. Now, while I can re-enable shuffle
on the machines, when trying to do an ls, I get a timeout.
Looking at the
I'm running 1.2.3 and have both CQL3 tabels and old school style CFs in my
cluster.
I'd had a large insert job running the last several days which just
ended it had been inserting using cql3 insert statements in a cql3
table.
Now, I show no compactions going on in my cluster but for some
Appears that restarting a node makes CQL available on that node again, but
only that node.
Looks like I'll be doing a rolling restart.
On Fri, Mar 29, 2013 at 10:26 AM, David McNelis dmcne...@gmail.com wrote:
I'm running 1.2.3 and have both CQL3 tabels and old school style CFs in my
cluster
, if you wanted
to change from sync to hsha in a cluster you'd have to entirely restart the
cluster (not a big deal), but CQL would apparently not work at all until
all of your nodes had been restarted.
On Fri, Mar 29, 2013 at 10:35 AM, David McNelis dmcne...@gmail.com wrote:
Appears that restarting
In using CQL (the python library, at least), I didn't see a way to pass in
multiple nodes as hosts. With other libraries (like Hector and Pycassa) I
can set multiple hosts and my app will work with anyone on that list. Is
there something similar going on in the background with CQL?
If not, then
Morning,
Was reading up on secondary indexes and on the Datastax post about them, it
mentions the additional management overhead, and also that if you alter an
existing column family, that data will be updated in the background. But
how do secondary indexes affect write performance?
If the
, and another to the index column family, where in this index
column family the key is the value of the secondary column, and the value
is the key of the original row.
On 08/04/2012 11:40 AM, David McNelis wrote:
Morning,
Was reading up on secondary indexes and on the Datastax post about them
I would use something other than the page itself as the key. Maybe a
filename, something smaller.
Then you could use a LongType comparator for the columns and use the page
number for the column name, the value being the contents of the files.
On Wed, Jul 11, 2012 at 1:34 PM, Tomek Hankus
That article is a good starting point. To make your life a bit easier,
consider checking out CassandraUnit that provides facilities to load
example data in a variety of ways.
https://github.com/jsevellec/cassandra-unit
Then you just need to be able to pass in which cassandra instance to
connect
The documentation for that section needs to be updated...
What happens is that if you just autobootstrap without setting a token it
will by default bisect the range of the largest node.
So if you go through several iterations of adding nodes, then this is what
you would see:
Gen 1:
Node A:
Daning,
You can see how to do this basic sort of thing on the Wiki's operations
page ( http://wiki.apache.org/cassandra/Operations )
In short, you'll want to run:
nodetool -h hostname move newtoken
Then, once you've update each of your tokens that you want to move, you'll
want to run
nodetool
Stax make their chef cook books available here
https://github.com/riptano/chef
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 11/01/2012, at 9:53 AM, David McNelis wrote:
Is anyone familiar with any tools that are already available
.
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 12/01/2012, at 8:47 AM, David McNelis wrote:
Not currently using any of those tools (though certainly an option, just
never looked into them).
Those tools seem more based around
Is anyone familiar with any tools that are already available to allow for
configurable synchronization of different clusters?
Specifically for purposes of development, i.e. Dev, staging, test, and
production cassandra environments, so that you can easily plug in the
information that you want to
of data from our cluster we'd certainly want to run it, or after added a
new node and adjusting the tokens.
So I want to make sure I'm not missing something here and that there would
be other reasons to run cleanup regularly?
--
*David McNelis*
Lead Software Engineer
Agentis Energy
an ideal doing it this way versus creating
separate CF per sector?how do you create Super CF inside of Super CF
via the CLI?
thanks,
deno
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
*A Smart Grid technology company focused
a
column name value that have a timeuuid of later than x minutes? i need to
be able to find all symbols that have not been fetch in x minutes by
sector. i know i get list of symbol by sector from my sector CF.
thanks,
deno
On 11/30/2011 1:07 PM, David McNelis wrote:
Then I would have a column
that if we were to ever purge a segment
of data from our cluster we'd certainly want to run it, or after added a
new node and adjusting the tokens.
So I want to make sure I'm not missing something here and that there
would be other reasons to run cleanup regularly?
--
David McNelis
Lead
it to v1 once
we migrate there (I don't know what JMX calls have changed at this
point)...if someone wants to send me a list of updates to the JMX calls,
I'll add them in and update it to handle multiple versions.
https://github.com/dmcnelis/NagiosCassandraMonitor
--
*David McNelis*
Lead Software
that
purpose.
Initialized, RPCServerRunning, OperationMode, Joined, and perhaps others
Note that some of those may not exist depending on your version of
cassandra, pick one appropriate for your version.
On Tue, Nov 22, 2011 at 1:02 PM, David McNelis
dmcne...@agentisenergy.com wrote
).
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
In that case, I think that the documentation is incorrect, as it has
Service listed as the package related to the StorageService.
I apologize for the lack of the rest of the thread, everything is getting
bounced when I try to send it for some reason.
--
*David McNelis*
Lead Software Engineer
to think its an issue with not being able to connect to
JMX in general.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
/writes from the CLI and
you should see the MBean afterwards.
This also means your monitoring application should handle this error
in the case of nodes restarting.
On Tue, Nov 22, 2011 at 7:51 AM, David McNelis
dmcne...@agentisenergy.com wrote:
Good morning,
I'm trying to set up a simple
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
for monitoring, trend analysis, etc.?
** **
**JConsole is useful for single node monitoring/etc but not scalable
data obviously doesn't persist between sessions...**
**
**
Many thanks,
Brian
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
lead to problems with concurrent inserts in a distributed
environment, but this is handled by our application logic.
What are your ideas on that?
Thanks
Matthias
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
c: 219.384.5143
*A Smart Grid technology company
to be evenly distributed
across my nodes or do I have to do something ?
Thanks in advance.
L. Aufrechter
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
10s
of gigs of ram on each node just to handle that overhead...at least as of
v1.
On Fri, Oct 7, 2011 at 9:40 AM, Jonathan Ellis jbel...@gmail.com wrote:
On Fri, Oct 7, 2011 at 9:36 AM, David McNelis
dmcne...@agentisenergy.com wrote:
In some documentation I've read it says that
keyspace's
our storage needs consistently.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
.
Cheers
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 13/09/2011, at 9:32 AM, David McNelis wrote:
Auto-bootstrapping is turned on and the node had been started several
hours ago. Since the node already shows up as part of the ring
So we tried decommissionning the 100.5 node, then re added it to the ring.
It now appears to be streaming data properly to that node.
On Tue, Sep 13, 2011 at 6:56 AM, David McNelis
dmcne...@agentisenergy.comwrote:
I ran a repair on 100.5. It returned back almost immediately and netstats
run
after?
We have a node that is almost full and need to move it so that we can shift
its loadbut it already has a cleanup process running which, instead of
causing less data usage as expected, is actually growing the amount of
space taken at a pretty fast rate.
--
*David McNelis*
Lead
streaming on the thought
that something may have failed, but that didn't yield any appreciable
results.
Are we seeing completely abnormal behavior? Should I consider making the
token for the fourth node considerably smaller? We calculated the node's
tokens using the standard python script.
--
*David
...@gmail.com wrote:
Looks kind of like the 4th node was added to the cluster w/o bootstrapping.
On Mon, Sep 12, 2011 at 3:59 PM, David McNelis
dmcne...@agentisenergy.com wrote:
We are running the datastax .8 rpm distro. We have a situation where we
have 4 nodes and each owns 25% of the keys
set correctly each would own 33.33%.
Thanks.
On Tue, Aug 16, 2011 at 3:33 PM, Jonathan Ellis jbel...@gmail.com wrote:
Yes, that looks about right.
Totally baffled how the wiki script could spit out those tokens for a
3-node cluster.
On Tue, Aug 16, 2011 at 2:04 PM, David McNelis
dmcne
throughput.
Anyone have any thoughts on a Blade v. Rackable solution for spinning up a
cassandra cluster?
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often
on
machine X from machine Y?
On Fri, Jun 24, 2011 at 8:20 AM, David McNelis
dmcne...@agentisenergy.com wrote:
Running on Centos.
We had a massive power failure and our UPS wasn't up to 48 hours without
power...
In this situation the IP addresses have all stayed the same. I can still
Just to close this out, in case anyone was interested... my problem was
firewall related, in that I didn't have my messaging/data port (7000) open
on my seed node. Allowing traffic on this port resolved my issues.
On Fri, Jun 3, 2011 at 1:43 PM, David McNelis dmcne...@agentisenergy.comwrote
, from the documentation, that
both of my nodes would show up in the ring if I ran 'ring' in nodetool.
This is a new cluster.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping
something mis-configured in
my cassandra instance (which is stock, outside of having data upgraded from
0.7.2.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
)
at CqlTest.main(CqlTest.java:25)
On Wed, Apr 27, 2011 at 4:27 PM, Jonathan Ellis jbel...@gmail.com wrote:
What's the stacktrace?
On Wed, Apr 27, 2011 at 9:45 AM, David McNelis
dmcne...@agentisenergy.com wrote:
I have a feeling that I'm likely doing something dumb. I have the
following code
the old Cassandra jar is no
longer on your classpath?
On Wed, Apr 27, 2011 at 4:29 PM, David McNelis
dmcne...@agentisenergy.com wrote:
Attached:
21 [main] INFO org.apache.cassandra.cql.jdbc.Connection - Connected to
localhost:9160
Exception in thread main
with the
changed ip comes up, it's fine ... but other members in the ring don't see
it and keep the old ip address regardless of the yaml changing
-sd
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology
to my server's
IP address instead of 127.0.0.1 and that caused me not to be able to connect
in general. My rpc_address is 0.0.0.0.
Has anyone else experienced this or have an inclination where I'm going
wrong?
Thanks,
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
,
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
,
it looks like our main problem is with the Thrift PHP module.
On Wed, Mar 2, 2011 at 9:04 AM, David McNelis dmcne...@agentisenergy.comwrote:
We are able to telnet to port 9160, and didn't have any issues resolving
the target along those lines. So at this point, I don't think we're looking
In case anyone is interested. Our problem revolved around one machine
having the phpcassa thrift patch, and the other did not. Its resolved now.
On Wed, Mar 2, 2011 at 10:25 AM, David McNelis
dmcne...@agentisenergy.comwrote:
It looks like we are having an issue with the Thrift installation
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused on helping consumers of energy
control an often under-managed resource.*
smart enough to get the columns in that range.
Thanks!
Bill-
On Wed, Jan 26, 2011 at 4:12 PM, David McNelis
dmcne...@agentisenergy.com wrote:
I would say in that case you might want to try a single column family
where the key to the column is the system name.
Then, you could name
contained in this message that do not relate to the official
business of the WorldPay Group shall not be understood as endorsed or given
by it.
--
*David McNelis*
Lead Software Engineer
Agentis Energy
www.agentisenergy.com
o: 630.359.6395
c: 219.384.5143
*A Smart Grid technology company focused
Silly question, M us thousand or million? In print, thousand is M, fwiw
Sent from my Droid
On Jan 23, 2011 7:26 PM, Maxim Potekhin potek...@bnl.gov wrote:
Aaron -- thanks!
I don't have examples like Timo.
But,
I'm keen to use multiple indices over a database
of 300M rows.
Maxim
On
80 matches
Mail list logo