hi, what happens if unloader meets blob field?
2015-04-20 23:43 GMT+02:00 Sebastian Estevez sebastian.este...@datastax.com
:
Try Brian's cassandra-unloader
https://github.com/brianmhess/cassandra-loader#cassandra-unloader
All the best,
[image: datastax_logo.png] http://www.datastax.com/
Try Brian's cassandra-unloader
https://github.com/brianmhess/cassandra-loader#cassandra-unloader
All the best,
[image: datastax_logo.png] http://www.datastax.com/
Sebastián Estévez
Solutions Architect | 954 905 8615 | sebastian.este...@datastax.com
[image: linkedin.png]
Hi guys,
We have a 100+ nodes cluster, each node has about 400G data, and is running
on a flash disk. We are running 2.1.2.
When I bring in a new node into the cluster, it introduces significant load
to the cluster. For the new node, the cpu usage is 100%, but disk write io
is only around
Can you grep for GCInspector in your system.log? Maybe you have long GC
pauses.
All the best,
[image: datastax_logo.png] http://www.datastax.com/
Sebastián Estévez
Solutions Architect | 954 905 8615 | sebastian.este...@datastax.com
[image: linkedin.png]
Blobs are ByteBuffer s it calls getBytes().toString:
https://github.com/brianmhess/cassandra-loader/blob/master/src/main/java/com/datastax/loader/parser/ByteBufferParser.java#L35
All the best,
[image: datastax_logo.png] http://www.datastax.com/
Sebastián Estévez
Solutions Architect | 954
Tyler,
I can consider trying out light weight transactions, but here are my
concerns
#1. We have 2 data centers located close by with plans to expand to more
data centers which are even further away geographically.
#2. How will this impact light weight transactions when there is high level
of
Thanks Sebastian, I will try it out.
But I am also curious why is the COPY command failing with Out of Memory
Error.
regards
Neha
On Tue, Apr 21, 2015 at 4:35 AM, Sebastian Estevez
sebastian.este...@datastax.com wrote:
Blobs are ByteBuffer s it calls getBytes().toString:
hi,
there were only a few (4 of them across 4 minutes with around 200ms), so
shouldn't be the reason
The system log has tons of
INFO [MigrationStage:1] 2015-04-20 11:03:21,880 ColumnFamilyStore.java
(line 633) Enqueuing flush of Memtable-schema_keyspaces@2079381036(138/1215
serialized/live
On Mon, Apr 20, 2015 at 6:08 PM, Dikang Gu dikan...@gmail.com wrote:
When I bring in a new node into the cluster, it introduces significant
load to the cluster. For the new node, the cpu usage is 100%, but disk
write io is only around 50MB/s, while we have 10G network.
Does it sound normal
Hi Rob,
Why do you say steaming is single threaded? I see a lot of background
streaming threads running, for example:
STREAM-IN-/10.210.165.49 daemon prio=10 tid=0x7f81fc001000
nid=0x107075 runnable [0x7f836b256000]
STREAM-IN-/10.213.51.57 daemon prio=10 tid=0x7f81f0002000
Values in /etc/security/limits.d/cassandra.conf
# Provided by the cassandra package
cassandra - memlock unlimited
cassandra - nofile 10
On Mon, Apr 20, 2015 at 12:21 PM, Kiran mk coolkiran2...@gmail.com wrote:
Hi,
Thanks for the info,
Does the nproc,nofile,memlock settings in
Start one node at a time. Wait 2 minutes before starting each node.
How much data and nodes you have already? Depending on that, the streaming
of data can stress on the resources you have.
I would recommend to start one and monitor, if things are ok, add another
one. And so on.
Regards,
Carlos
unsubscribe
On Apr 20, 2015, at 8:08 AM, Carlos Rolo r...@pythian.com wrote:
Independent of the snitch, data needs to travel to the new nodes (plus all
the keyspace information that goes via gossip). So I won't bootstrap them all
at once, even if it is only for network traffic generated.
Hi Colin,
To remove your address from the list, send a message to:
user-unsubscr...@cassandra.apache.org
Cheers,
Matt
*From:* Colin Clark [mailto:co...@clark.ws]
*Sent:* 20 April 2015 14:10
*To:* user@cassandra.apache.org
*Subject:* Re: Adding nodes to existing cluster
Thanks for the response.
Sure we'll monitor as we're adding nodes.
We're now using 6 nodes on each DC. (We have 2 DCs)
Each node contains ~800GB
Do you know how rack configurations are relevant here?
Do you see any reason to bootstrap them one by one if we're not using
rack awareness?
On Mon,
Independent of the snitch, data needs to travel to the new nodes (plus all
the keyspace information that goes via gossip). So I won't bootstrap them
all at once, even if it is only for network traffic generated.
Don't forget to run cleanup on the old nodes once all nodes are in place to
reclaim
Jimmy,
What's the exact command that produced this trace? Are you saying that
the 16-second wait in your trace what times out in your CREATE TABLE
statements?
Jim Witschey
Software Engineer in Test | jim.witsc...@datastax.com
On Sun, Apr 19, 2015 at 7:13 PM, Jimmy Lin y2klyf+w...@gmail.com
Hi all,
In the near future I'll need to add more than 10 nodes to a 2.0.9
cluster (using vnodes).
I read this documentation on datastax website:
http://docs.datastax.com/en/cassandra/2.0/cassandra/operations/ops_add_node_to_cluster_t.html
In one point it says:
If you are using racks, you can
Hi,
Recently, we discovered that millions of mutations were getting dropped on our
cluster. Eventually, we solved this problem by increasing the value of
memtable_flush_writers from 1 to 3. We usually write 3 CFs simultaneously an
one of them has 4 Secondary Indexes.
New changes also
The documentation is referring to Consistent Range Movements.
There is a change in 2.1 that won't allow you to bootstrap multiple nodes
at the same time unless you explicitly turn off consistent range movements.
Check out the jira:
https://issues.apache.org/jira/browse/CASSANDRA-2434
All the
OK.
Thanks.
I'll monitor the resources status (network, memory, cpu, io) as I go
and try to bootsrap them at chunks which seems not to have a bad
impact.
Will do regarding the cleanup.
Thanks!
On Mon, Apr 20, 2015 at 4:08 PM, Carlos Rolo r...@pythian.com wrote:
Independent of the snitch, data
I think this is just saying that young gen collection using Par new collector
took 248 seconds. This is quite normal with CMS unless it happens too
frequenltly several times in a sec. I think query time has more to do with read
timeout in yaml. Try increasing it. If its a range query then
You'll have to configure your nodes to:
1. use AWS internal IPs for inter-node connection (check listen_address)
and
2. use the AWS public IP for client-to-node connections (check rpc_address)
Depending on the setup, there might be other interesting conf options in
cassandra.yaml
I would like to note that this will require all clients connect over the
external IP address. If you have clients within Amazon that need to connect
over the private IP address, this would not be possible. If you have a mix of
clients that need to connect over private IP address and public,
Ideally you'll be on the same network, but if you can't be, you'll need to
use the public ip in listen_address.
On Mon, Apr 20, 2015 at 9:47 AM Matthew Johnson matt.john...@algomi.com
wrote:
Hi all,
I have set up a Cassandra cluster with 2.1.4 on some existing AWS boxes,
just as a POC.
Hi all,
I have set up a Cassandra cluster with 2.1.4 on some existing AWS boxes,
just as a POC. Cassandra servers connect to each other over their internal
AWS IP addresses (172.x.x.x) aliased in /etc/hosts as sales1, sales2 and
sales3.
I connect to it from my local dev environment using the
There are a couple options here. You can use the built in address translator,
or, write a new load balancing policy. See
https://datastax-oss.atlassian.net/browse/JAVA-145 for more information.
From: Jonathan Haddad
Reply-To: user@cassandra.apache.org
Date: Monday, April 20, 2015 at 12:50
Hi,
TechEmpower Web Framework Benchmarks (
https://www.techempower.com/benchmarks/) is a collaborative effort for
measuring performance of a large number of contemporary web development
platforms. Benchmarking and test implementation code is published as
open-source.
I've contributed a test
Hi,
I am keep getting following line in the cassandra logs, apparently
something related to Garbage Collection. And I guess this is one of the
signs why i do not get any response (i get time-out) when I query large
volume of data ?!!!
ParNew GC in 248ms. CMS Old Gen: 453244264 - 570471312; Par
Yes, sometimes it is create table and sometime it is create index.
It doesn't happen all the time, but feel like if multiple tests trying to
do schema change(create or drop), Cassandra has a long delay on the schema
change statements.
I also just read about auto_snapshot, and I turn it off but
Does the nproc,nofile,memlock settings in
/etc/security/limits.d/cassandra.conf are set to optimum value ?
it's all default.
What is the consistency level ?
CL = Qurom
Is there any other way to export a table to CSV?
regards
Neha
On Mon, Apr 20, 2015 at 12:21 PM, Kiran mk
Small correction: we are making writes in 5 cf an reading frm one at high
speeds.
Thanks
Anuj Wadehra
Sent from Yahoo Mail on Android
From:Anuj Wadehra anujw_2...@yahoo.co.in
Date:Mon, 20 Apr, 2015 at 7:53 pm
Subject:Handle Write Heavy Loads in Cassandra 2.0.3
Hi,
Recently, we
I meant 248 milli seconds
Sent from Yahoo Mail on Android
From:Anuj Wadehra anujw_2...@yahoo.co.in
Date:Mon, 20 Apr, 2015 at 11:41 pm
Subject:Re: Getting ParNew GC in ... CMS Old Gen ... in logs
I think this is just saying that young gen collection using Par new collector
took 248 seconds.
Hi,
check the MAX_HEAP_SIZE configuration in cassandra-env.sh environment file
Also HEAP_NEWSIZE ?
What is the Consistency Level you are using ?
Best REgards,
Kiran.M.K.
On Mon, Apr 20, 2015 at 11:13 AM, Kiran mk coolkiran2...@gmail.com wrote:
Seems like the is related to JAVA HEAP Memory.
hi,
What is the count of records in the column-family ?
We have about 38,000 Rows in the column-family for which we are
trying to export
What is the Cassandra Version ?
We are using Cassandra 2.0.11
MAX_HEAP_SIZE and HEAP_NEWSIZE is the default .
The Server is 8 GB.
regards
Neha
On
Hi,
Thanks for the info,
Does the nproc,nofile,memlock settings in
/etc/security/limits.d/cassandra.conf are set to optimum value ?
What is the consistency level ?
Best Regardds,
Kiran.M.K.
On Mon, Apr 20, 2015 at 11:55 AM, Neha Trivedi nehajtriv...@gmail.com
wrote:
hi,
What is the count
37 matches
Mail list logo