, Charles charles.sibb...@bskyb.com wrote:
Hi Oleg,
What is the minor version of 1.2? I am looking to do the same for 1.2.14
in a very large cluster.
Regards
Charles
On 02/02/2015 13:33, Oleg Dulin oleg.du...@gmail.com wrote:
Dear Distinguished Colleagues:
We'd like to upgrade our cluster from
, Charles said:
Hi Oleg,
What is the minor version of 1.2? I am looking to do the same for 1.2.14
in a very large cluster.
Regards
Charles
On 02/02/2015 13:33, Oleg Dulin oleg.du...@gmail.com wrote:
Dear Distinguished Colleagues:
We'd like to upgrade our cluster from 1.2 to 2.0
Dear Distinguished Colleagues:
We'd like to upgrade our cluster from 1.2 to 2.0 and then to 2.1 .
We are using Pelops Thrift client, which has long been abandoned by its
authors. I've read that 2.x has changes to the Thrift protocol making
it incompatible with 1.2 (and of course now the link
right now).
Regards,
Carlos Juzarte Rolo
Cassandra Consultant
Pythian - Love your data
rolo@pythian | Twitter: cjrolo | Linkedin: linkedin.com/in/carlosjuzarterolo
Tel: 1649
www.pythian.com
On Mon, Feb 2, 2015 at 3:21 PM, Oleg Dulin oleg.du...@gmail.com wrote:
Sure but the question is really
I have a setup with 6 cassandra nodes (1.2.18), using RandomPartition,
not using vnodes -- this is a legacy cluster.
We went from 3 nodes to 6 in the last few days to add capacity.
However, there appears to be an imbalance:
Datacenter: us-east
==
Replicas: 2
Address Rack
/jira/browse/CASSANDRA-7173
Regards,
Mark
On 28 October 2014 10:35, Oleg Dulin oleg.du...@gmail.com wrote:
I have a setup with 6 cassandra nodes (1.2.18), using RandomPartition,
not using vnodes -- this is a legacy cluster.
We went from 3 nodes to 6 in the last few days to add capacity
Dear Colleagues:
I need to move Cassandra from EC2 classic into VPC.
What I was thinking is that I can create a new data center within VPC
and rebuild it from my existing one (switching to vnodes while I am at
it). However, I don't understand how the ec2-snitch will deal with this.
Another
.
There should be no need to use public IP's if whatever talks to
cassandra is also within your VPC.
Hope this helps.
Bram
On Mon, Sep 8, 2014 at 3:34 PM, Oleg Dulin oleg.du...@gmail.com wrote:
Dear Colleagues:
I need to move Cassandra from EC2 classic into VPC.
What I was thinking is that I can
Distinguished Colleagues:
Our current Cassandra cluster on AWS looks like this:
3 nodes in N. Virginia, one per zone.
RF=3
Each node is a c3.4xlarge with 2x160G SSDs in RAID-0 (~300 Gig SSD on
each node). Works great, I find it the most optimal configuration for a
Cassandra node.
But the
Colleagues:
I'd like to announce a pet project I started:
https://github.com/olegdulin/cassandra-aws
What I would like to accomplish as an end-goal is an Amazon marketplace
AMI that makes it easy to configure a new Cassandra cluster or add new
nodes to an existing Cassandra cluster, w/o
On 2014-06-06 15:25:44 +, Michael Shuler said:
On 06/06/2014 09:57 AM, Oleg Dulin wrote:
I'd like to announce a pet project I started:
https://github.com/olegdulin/cassandra-aws
Cool :)
https://github.com/riptano/ComboAMI is the DataStax AMI repo.
What I would like to accomplish as an end
I have a cluster that looks like this:
Datacenter: us-east
==
Replicas: 2
Address RackStatus State LoadOwns
Token
113427455640312821154458202477256070484
*.*.*.1
I keep asking same question it seems -- sign of insanity.
Cassandra version 1.2, not using vnodes (legacy).
On 2014-03-07 19:37:48 +, Robert Coli said:
On Fri, Mar 7, 2014 at 6:00 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I have the following situation:
10.194.2.5 RAC1 Up
I need to rebalance my cluster. I am sure this question has been asked
before -- will 1.2 continue to serve reads and writes correctly while
move is in progress ?
Need this for my sanity.
--
Regards,
Oleg Dulin
http://www.olegdulin.com
utilization in line ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
,
Oleg Dulin
http://www.olegdulin.com
62.72 GB100.00% 100
--
Regards,
Oleg Dulin
http://www.olegdulin.com
, Apr 7, 2014 at 8:08 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I added two more nodes on Friday, and moved tokens around.
For four nodes, the tokesn should be:
Node #1:0
Node #2: 42535295865117307932921825928971026432
Node #3
, there is no even
way to distribute your data among those four nodes.
tl;dr Switch node 2 to rack 1a.
-Tupshin
On Mon, Apr 7, 2014 at 8:08 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I added two more nodes on Friday, and moved tokens around.
For four nodes, the tokesn should be:
Node #1
Sigh, so I am back to where I started from...
I did lower gc_grace...
jmap -histo:live shows heap is stuffed with DeletedColumn and
ExpiringColumn
This is extremely frustrating.
On 2014-03-11 19:24:50 +, Oleg Dulin said:
Good news is that since I lowered gc_grace period
I came across something on the cassandra it that made me concerned.
Default value for hinted_handoff_throttle_in_kb is 1024, one Meg per
second. I have four nodes and rf=2. I have hints timeout set to 24, to
avoid having to do repairs if I took longer than that to reboot a node.
What got me
,
Oleg Dulin
http://www.olegdulin.com
Robert Coli rc...@eventbrite.com wrote:
On Tue, Mar 11, 2014 at 4:30 PM, ziju feng pkdog...@gmail.com wrote:
Is there any way to guarantee a counter's value
no.
=Rob
I wouldn't use cassandra for counters... Use something like redis if that
is what you want.
/browse/CASSANDRA-6541
Hope this helps.
Jonathan
Jonathan Lacefield
Solutions Architect, DataStax
(404) 822 3487
image
image
On Mon, Mar 10, 2014 at 6:41 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I get that :)
What I'd like to know is how to fix that :)
On 2014-03-09 20:24:54 +, Takenori
query.
On Sun, Mar 9, 2014 at 10:55 PM, Oleg Dulin oleg.du...@gmail.com wrote:
I am trying to understand why one of my nodes keeps full GC.
I have Xmx set to 8gigs, memtable total size is 2 gigs.
Consider the top entries from jmap -histo:live @ http://pastebin.com/UaatHfpJ
--
Regards,
Oleg
I am trying to understand why one of my nodes keeps full GC.
I have Xmx set to 8gigs, memtable total size is 2 gigs.
Consider the top entries from jmap -histo:live @ http://pastebin.com/UaatHfpJ
--
Regards,
Oleg Dulin
http://www.olegdulin.com
to be the heaviest
loaded node by CPU usage.
What would be a clean way to rebalance ? If I use move operation
follwoed by cleanup, would it require a repair afterwards ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
Bumping this up -- anything ? anyone ?
On 2014-02-13 16:01:50 +, Oleg Dulin said:
I am getting these exceptions on one of the nodes, quite often, during
compactions:
java.lang.AssertionError: originally calculated column size of
84562492 but now it is 84562600
Usually
, and
subsequently causing other performance issues for me.
Is there a way to fix that ? Would nodetool scrub take care of this ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
checklist what I would need to do.
--
Regards,
Oleg Dulin
http://www.olegdulin.com
?
Regards,
Oleg
On 2014-02-13 22:01:10 +, Oleg Dulin said:
Dear Distinguished Colleagues:
I have a situation where in the production environment one of the
machines is overheating and needs to be serviced. Now, the landscape
looks like this:
4 machines in primary DC, 4 machiens in DR DC
.
Can someone help me on this?
--
Regards,
Mahesh Rajamani
Read the column first...
--
Regards,
Oleg Dulin
http://www.olegdulin.com
, Dec 16, 2013 at 6:53 AM, Oleg Dulin oleg.du...@gmail.com wrote:
Hi guys!
I beleive my limits settings are correct. Here is the output of ulimits -a:
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size
(Thread.java:722)
Any input is greatly appreciated.
--
Regards,
Oleg Dulin
http://www.olegdulin.com
,
uuid uuid,
reserved_by int,
reserved_until bigint,
primary key(start,end));
Will having an app server mark a block as reserved (QUOROM) and then
reading it back (QUOROM) be enough to for an app server to know it owns
that block of numbers?
Best regards,
Jacob
--
Regards,
Oleg
What happens if they are not being successfully delivered ? Will they
eventually TTL-out ?
Also, do I need to truncate hints on every node or is it replicated ?
Oleg
On 2013-11-04 21:34:55 +, Robert Coli said:
On Mon, Nov 4, 2013 at 11:34 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I
.
On Sun, Nov 3, 2013 at 11:54 PM, Oleg Dulin oleg.du...@gmail.com wrote:
Cass 1.1.11 ran out of memory on me with this exception (see below).
My parameters are 8gig heap, new gen is 1200M.
ERROR [ReadStage:55887] 2013-11-02 23:35:18,419
AbstractCassandraDaemon.java (line 132) Exception
I have a dual DC setup, 4 nodes, RF=4 in each.
The one that is used as primary has its system keyspace fill up with
200 gigs of data, majority of which is hints.
Why does this happen ?
How can I clean it up ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:722)
Any thoughts ?
This is a dual data center set up, with 4 nodes in each DC and RF=2 in each.
--
Regards,
Oleg Dulin
http://www.olegdulin.com
to me how the repair works such that I don't
totally trash my cluster or spill into work week ?
Is there any improvement and clarity in 1.2 ? How about 2.0 ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
could cause this ?
Any help is greatly apprecaited.
--
Regards,
Oleg Dulin
http://www.olegdulin.com
two options:
1) Treat this as a new data center and let the nodes sync from scratch, or
2) Bring the nodes back up with all the data in place and do a repair.
We are talking about 30-40Gigs per node. There are 4 nodes in both data
centers, with RF=2.
--
Regards,
Oleg Dulin
http
of this. Is this because DC1 is not fully
compacted ? Is it because DC2 is not fully synced and I am not checking
correctly ? How can I tell that there is still replication going on in
progress (note, I started my load yesterday at 9:50am).
--
Regards,
Oleg Dulin
http://www.olegdulin.com
and trigger a read-repair. But
why would nodetool repair logs show that everything is in sync ?
Regards,
Oleg
On 2013-09-27 10:23:45 +, Oleg Dulin said:
Consider this output from nodetool ring:
Address DC RackStatus State Load
Effective-Ownership
handoff traffic ?
Regards,
Oleg
On 2013-09-27 10:35:34 +, Oleg Dulin said:
Wanted to add one more thing:
I can also tell that the numbers are not consistent across DRs this way
-- I have a column family with really wide rows (a couple million
columns).
DC1 reports higher column counts than
-- it'll just hog
the 45Megabit pipe…
Somehow I need away to load all the data into primary cluster, then
ship it over to the backup in a more timely fashion…
Any suggestions are greatly appreciated.
Also, I need a way to know if the replication is up to date or not.
--
Regards,
Oleg Dulin
http
name as parsed into a tuple of two parts.
Am I doing something wrong here ?
--
Regards,
Oleg Dulin
http://www.olegdulin.com
on any port. Is there an easy
way?
Thanks!
Mark
--
Regards,
Oleg Dulin
http://www.olegdulin.com
a correct data model.
Can you describe your use case ?
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
. But it seems like I can
view the contents like this:
strings *-Data.db
At what point does compression start ? How can I confirm it is working ?
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
: 339882.089: [GC
10856891K-9179680K(16567552K), 0.0264210 secs]
41550 2013-05-03T07:18:25.062-0400: 339885.900: [GC
10857504K-9178985K(16567552K), 0.0267200 secs]
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
?
-Bryan
On Fri, May 3, 2013 at 4:27 AM, Oleg Dulin oleg.du...@gmail.com wrote:
Here is my question. It can't possibly be a good set up to use 16gig
heap space, but this is the best I can do. Setting it to default never
worked well for me, setting it to 8g doesn't work well either. It can't
keep
)
10432 at java.lang.Thread.run(Thread.java:722)
ANy thoughts ? Should I be concerned about data being lost ?
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
400GB.
Compactions/Repairs/Move operations etc become a nightmare otherwise. How
much data do you expect to have on each node? Also depends on caches,
bloom filters etc
On 11/5/12 8:57 AM, Oleg Dulin oleg.du...@gmail.com wrote:
I have 4 nodes at my disposal.
I can configure them like this:
1) RF=1
It works pretty fast.
Cool.
Just keep an eye out for how big the lucene token row gets.
Cheers
Indeed, it may get out of hand, but for now we are ok -- for the
foreseable future I would say.
Should it get larger, I can split it up into rows -- i.e. all tokens
that start with a, all
) and it is
causing slowness with read requests.
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
...@gmail.com wrote:
Some one did search on Lucene, but for very fresh data they build
search index in memory so data become available for search without
delays.
On 3 September 2012 22:25, Oleg Dulin oleg.du...@gmail.com wrote:
Dear Distinguished Colleagues:
--
Regards,
Oleg Dulin
NYC Java Big
jdk 1.7 on several clusters on 1.1 for a while now.
--
/ Peter Schuller (@scode, http://worldmodscode.wordpress.com)
--
Ric Dong Newegg Ecommerce, MIS department --
The downside of being better than everyone else is that people tend to
assume you're pretentious.
--
Regards,
Oleg Dulin
NYC
You should create an index where you store references to your records.
You can use composite column names where column
name=composite(timestamp,key)
then you would get a slice of all columns where timestamp part of the
composite is = TTL in the past, and then iterate through them and
delete
is going om here?
Many thanks!
Rene
image
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
I am currently profiling a Cassandra 1.1.1 set up using G1 and JVM 7.
It is my feeble attempt to reduce Full GC pauses.
Has anyone had any experience with this ? Anyone tried it ?
--
Regards,
Oleg Dulin
NYC Java Big Data Engineer
http://www.olegdulin.com/
Has anyone tried running 1.1.1 on Java 7?
I know Datastax does not recommend it for DSE, is there a reason why ?
Regards,
Oleg
Dear Distinguished Colleagues:
I need to add full-text search and somewhat free form queries to my
application. Our data is made up of items that are stored in a single
column family, and we have a bunch of secondary indices for look ups.
An item has header fields and data fields, and the
I get this:
InvalidRequestException(why:invalid operation for commutative columnfamily
Any thoughts ?
We use Pelops...
Here are my requirements.
We use Cassandra.
I get millions of invoice line items into the system. As I load them I
need to build up some data structures.
* Invoice line items by invoice id (each line item has an invoice id on
it ), with total dollar value
* Invoice line items by customer
Here is my flow:
One process write a really wide row (250K+ supercolumns, each one with
5 subcolumns, for the total of 1K or so per supercolumn)
Second process comes in literally 2-3 seconds later and starts reading from it.
My observation is that nothing good happens. It is ridiculously
Does a write to a sub column involve deserialization of the entire super
column ?
Thanks,
Oleg
I am using cassandra 1.1.0 on a 3 node environment. I just truncated a
few column families then restarted the nodes. now when I restarted them
it says my keyspace doesn't exist. The data for the keyspace is still
in the data directory. Does anyone know what could have caused this?
I can't quite describe what happened, but essentially one day I found
that my column values that are supposed to be UTF-8 strings started
getting bogus characters.
Is there a known data corruption issue with 1.1 ?
We have a 3-node cluster. We use RF of 3 and CL of ONE for both reads
and writes…. Is there a reason I should schedule a regular nodetool
repair job ?
Thanks,
Oleg
We are using Cassandra 1.1.0 with an older Pelops version, but I don't
think that in itself is a problem here.
I am getting this exception:
TimedOutException()
at
org.apache.cassandra.thrift.Cassandra$get_slice_result.read(Cassandra.java:7660)
at
Tyler Hobbs ty...@datastax.com wrote:
On Fri, Jun 1, 2012 at 9:39 AM, Oleg Dulin oleg.du...@gmail.com wrote:
Is my understanding correct that this is where cassandra is telling us it
can't accomplish something within that timeout value -- as opposed to
network timeout ? Where is it set
Is it possible ? How ?
Dear distinguished colleagues:
I am trying to come up with a data model that lets me do aggregations, such
as sums and averages.
Here are my requirements:
1. Data may be updated concurrently
2. I want to avoid changing schema; we have a multitennant cloud solution
that is driven by
of 3 million, the chance of
having 2 items from the same bin is pretty low). We will experiment
with row cache, and hopefully it will help, not the opposite (the
tuning guide says row cache could be detrimental in some circumstances).
-- Y.
On Wed, May 16, 2012 at 4:25 PM, Oleg Dulin oleg.du
I am running my processes on the same nodes as Cassandra.
What I'd like to do is when I get a connection from Pelops, it gives preference
to the Cassandra node local to the host my process is on.
Is it possible ? How ?
Regards,
Oleg Dulin
Please note my new office #: 732-917-0159
76 matches
Mail list logo