Re: monitor and alert tool for open source cassandra

2018-09-26 Thread Joaquin Casares
Hello James,

This project has provides a Docker Compose environment for using Prometheus
and Grafana along with preconfigured files for Prometheus and Graphite
consumption. I would, however, prefer Prometheus over Graphite as Grafana's
backend.

https://github.com/thelastpickle/docker-cassandra-bootstrap


Cheers,

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com


On Mon, Sep 24, 2018 at 8:44 PM James Shaw  wrote:

> Adam:
>Thanks!
> Very helpful. I will take a look.
>
> James
>
> On Mon, Sep 24, 2018 at 6:59 PM Adam Zegelin  wrote:
>
>> Hi James,
>>
>> Prometheus is the most common monitoring solution for K8s-managed
>> applications.
>>
>> There are a number of options to get Cassandra metrics into Prometheus.
>> One of which, shameless plug, is something I've been working on for the
>> past few months -- cassandra-exporter, a JVM agent that aims to be the
>> easiest & fastest way to get Cassandra metrics into Prometheus.
>>
>> Check it out on GitHub: https://github.com/zegelin/cassandra-exporter
>>
>> cassandra-exporter is part of Instaclustrs current work-in-progress
>> cassandra-operator for Kubernetes.
>> If you're interested in running Cassandra on a K8s cluster and would like
>> easy scale up/scale down, automatic Prometheus integration, automatic
>> backups, etc, then an operator makes it a lot easier to manage.
>> Check it out on GitHub: https://github.com/instaclustr/cassandra-operator
>> The project is approaching MVP status, and we would certainly appreciate
>> any feedback or contributions.
>>
>> Regards,
>> Adam
>>
>>
>>
>> On Mon, 24 Sep 2018 at 14:32, James Shaw  wrote:
>>
>>> Hi, there:
>>>What are latest good tools for monitoring open source cassandra
>>> ?
>>> I was used to Datastax opscenter tool, felt all tasks quite easy. Now on
>>> new project, open source cassandra, on Kubernetes container/docker, logs in
>>> Splunk,  feel very challenge.
>>> Most wanted metrics are read / write latency, read / write time out.
>>>
>>> Any advice is welcome.  I appreciate your help!
>>>
>>> Thanks,
>>>
>>> James
>>>
>>


Re: EXT: Cassandra Monitoring tool

2018-05-25 Thread Joaquin Casares
Hello Aneesh,

While this doesn't provide a GUI, tablesnap is a community tool that does a
great job at handling backups:

https://github.com/JeremyGrosser/tablesnap


Cheers,

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, May 25, 2018 at 11:36 AM, ANEESH KUMAR K.M <anee...@gmail.com>
wrote:

> Thank you Hari for the details.
>
> One more question, please suggest me some cluster managing too for
> Cassandra cluster. Looking for opensource tools that will support to take
> snapshot and restore via GUI.
>
> Regards,
> Aneesh
>
> On Fri, May 25, 2018 at 10:00 PM, Harikrishnan Pillai <
> hpil...@walmartlabs.com> wrote:
>
>> I assume you are using open source cassandra and you can look at
>> Prometheus grafana for cassandra monitoring and lot information available
>> in internet regarding how to setup the Prometheus monitoring for cassandra .
>>
>>
>> Sent from my iPhone
>>
>> > On May 25, 2018, at 9:23 AM, ANEESH KUMAR K.M <anee...@gmail.com>
>> wrote:
>> >
>> > Please suggest me some good cluster monitoring tool for cassandra multi
>> region cluster.
>> >
>>
>> -
>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
>> For additional commands, e-mail: user-h...@cassandra.apache.org
>>
>>
>


Re: Cassandra Monitoring tool

2018-05-25 Thread Joaquin Casares
Hello Aneesh,

Alain and I also worked on this project that you might find useful:

https://github.com/thelastpickle/docker-cassandra-bootstrap


It is a Docker Compose environment to showcase the Prometheus graphs that
we worked on. There's still a missing update, so do please check back soon,
but the majority of what we've worked on is housed in that repo. If you
decide to move forward with Prometheus, you can use the curl scripts
provided in the repo to upload the charts and all preset alerts into your
Grafana environment. If you use the GUI to upload the graphs, you'll lose
all preset visual thresholds and alerts, but still keep all completed
tooltips/descriptions.

Now that I remembered it was missing an update, we should push the new
features out sooner rather than later. The new update includes better
descriptive text, alert text, and alert fixes.

Hope it helps!

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, May 25, 2018 at 12:54 PM, Alain RODRIGUEZ <arodr...@gmail.com>
wrote:

> Hello,
>
> With more details on your criteria, it would be easier.
>
> I used:
>
> - SPM from Sematext (Commercial - dashboards out of the box)
> - Datastax OpsCenter (Commercial, With DSE only nowadays)
> - Grafana/Graphite or Prometheus (Open source - 'do it yourself' - but
> some templates exist)
> - Datadog (Commercial - dashboards out of the box)
>
> It depends what you want really. All those worked for me at some point in
> time :).
>
> Note: We built the Dashboards that are out of the box for Datadog. I
> believe it should put anyone starting with Cassandra in good conditions to
> operate and troubleshot Apache Cassandra. My opinion here is biased most
> definitely :). I liked the interfaces of Datadog (using regularly) and
> Sematext (old experience) the most.
>
> C*heers,
> ---
> Alain Rodriguez - @arodream - al...@thelastpickle.com
> France / Spain
>
> The Last Pickle - Apache Cassandra Consulting
> http://www.thelastpickle.com
>
> 2018-05-25 17:23 GMT+01:00 ANEESH KUMAR K.M <anee...@gmail.com>:
>
>> Please suggest me some good cluster monitoring tool for cassandra multi
>> region cluster.
>>
>>
>


Re: Cassandra reaper

2018-04-24 Thread Joaquin Casares
Sure thing Abdul,

That's great to hear! Unfortunately, the JMX authentication needs to be in
the config file currently. And even if the JMX authentication was stored
within Cassandra, we would still need to store connection details within
the yaml and storing the JMX authentication credentials within Cassandra
may not be ideal from a security standpoint.

The UI keeps logs of all the previous repairs, to the best of my knowledge.
If you want to completely uninstall Reaper, you can perform a DROP KEYSPACE
reaper_db; from within cqlsh, but that would remove all schedules as well.

Cheers,

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Tue, Apr 24, 2018 at 7:49 PM, Abdul Patel <abd786...@gmail.com> wrote:

> Thanks Joaquin,
>
> Yes i used the same and worked fine ..only thing is i had to add userid
> password ..which is somewhat annoyoing to keep in comfig file ..can i get
> reed of it and still store on reaper_db keyspace?
> Also how to clean reaper_db by deleting completed reaper information from
> gui? Or any other cleanup is required?
>
>
> On Tuesday, April 24, 2018, Joaquin Casares <joaq...@thelastpickle.com>
> wrote:
>
>> Hello Abdul,
>>
>> Depending on what you want your backend to be stored on, you'll want to
>> use a different file.
>>
>> So if you want your Reaper state to be stored within a Cassandra cluster,
>> which I would recommend, use this file as your base file:
>>
>> https://github.com/thelastpickle/cassandra-reaper/blob/
>> master/src/packaging/resource/cassandra-reaper-cassandra.yaml
>>
>> Make a copy of the yaml and include your system-specific settings. Then
>> symlink it to the following location:
>>
>> /etc/cassandra-reaper/cassandra-reaper.yaml
>>
>> For completeness, this file is an example of how to use a Postgres server
>> to store the Reaper state:
>>
>> https://github.com/thelastpickle/cassandra-reaper/blob/
>> master/src/packaging/resource/cassandra-reaper.yaml
>>
>>
>> Hope that helped!
>>
>> Joaquin Casares
>> Consultant
>> Austin, TX
>>
>> Apache Cassandra Consulting
>> http://www.thelastpickle.com
>>
>> On Tue, Apr 24, 2018 at 7:07 PM, Abdul Patel <abd786...@gmail.com> wrote:
>>
>>> Thanks
>>>
>>> But the differnce here is cassandra-reaper-caasandra has more paramters
>>> than the cassandra-reaper.yaml
>>> Can i just use the 1 file with all details or it looks for one specific
>>> file?
>>>
>>>
>>> On Tuesday, April 24, 2018, Joaquin Casares <joaq...@thelastpickle.com>
>>> wrote:
>>>
>>>> Hello Abdul,
>>>>
>>>> You'll only want one:
>>>>
>>>> The yaml file used by the service is located at
>>>> /etc/cassandra-reaper/cassandra-reaper.yaml and alternate config
>>>> templates can be found under /etc/cassandra-reaper/configs. It is
>>>> recommended to create a new file with your specific configuration and
>>>> symlink it as /etc/cassandra-reaper/cassandra-reaper.yaml to avoid
>>>> your configuration from being overwritten during upgrades.
>>>>
>>>> Adapt the config file to suit your setup and then run `sudo service
>>>> cassandra-reaper start`.
>>>>
>>>>
>>>> Source: http://cassandra-reaper.io/docs/download/install/#service-co
>>>> nfiguration
>>>>
>>>>
>>>> Hope that helps!
>>>>
>>>> Joaquin Casares
>>>> Consultant
>>>> Austin, TX
>>>>
>>>> Apache Cassandra Consulting
>>>> http://www.thelastpickle.com
>>>>
>>>> On Tue, Apr 24, 2018 at 6:51 PM, Abdul Patel <abd786...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi All,
>>>>>
>>>>> For reaper do we need both file or only one?
>>>>>
>>>>> Cassandra.reaper.yaml
>>>>> Cassandra-reaper-cassandra.yaml
>>>>>
>>>>
>>>>
>>


Re: Cassandra reaper

2018-04-24 Thread Joaquin Casares
Hello Abdul,

Depending on what you want your backend to be stored on, you'll want to use
a different file.

So if you want your Reaper state to be stored within a Cassandra cluster,
which I would recommend, use this file as your base file:

https://github.com/thelastpickle/cassandra-reaper/blob/master/src/packaging/resource/cassandra-reaper-cassandra.yaml

Make a copy of the yaml and include your system-specific settings. Then
symlink it to the following location:

/etc/cassandra-reaper/cassandra-reaper.yaml

For completeness, this file is an example of how to use a Postgres server
to store the Reaper state:

https://github.com/thelastpickle/cassandra-reaper/blob/master/src/packaging/resource/cassandra-reaper.yaml


Hope that helped!

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Tue, Apr 24, 2018 at 7:07 PM, Abdul Patel <abd786...@gmail.com> wrote:

> Thanks
>
> But the differnce here is cassandra-reaper-caasandra has more paramters
> than the cassandra-reaper.yaml
> Can i just use the 1 file with all details or it looks for one specific
> file?
>
>
> On Tuesday, April 24, 2018, Joaquin Casares <joaq...@thelastpickle.com>
> wrote:
>
>> Hello Abdul,
>>
>> You'll only want one:
>>
>> The yaml file used by the service is located at
>> /etc/cassandra-reaper/cassandra-reaper.yaml and alternate config
>> templates can be found under /etc/cassandra-reaper/configs. It is
>> recommended to create a new file with your specific configuration and
>> symlink it as /etc/cassandra-reaper/cassandra-reaper.yaml to avoid your
>> configuration from being overwritten during upgrades.
>>
>> Adapt the config file to suit your setup and then run `sudo service
>> cassandra-reaper start`.
>>
>>
>> Source: http://cassandra-reaper.io/docs/download/install/#service-
>> configuration
>>
>>
>> Hope that helps!
>>
>> Joaquin Casares
>> Consultant
>> Austin, TX
>>
>> Apache Cassandra Consulting
>> http://www.thelastpickle.com
>>
>> On Tue, Apr 24, 2018 at 6:51 PM, Abdul Patel <abd786...@gmail.com> wrote:
>>
>>> Hi All,
>>>
>>> For reaper do we need both file or only one?
>>>
>>> Cassandra.reaper.yaml
>>> Cassandra-reaper-cassandra.yaml
>>>
>>
>>


Re: Cassandra reaper

2018-04-24 Thread Joaquin Casares
Hello Abdul,

You'll only want one:

The yaml file used by the service is located at
/etc/cassandra-reaper/cassandra-reaper.yaml and alternate config templates
can be found under /etc/cassandra-reaper/configs. It is recommended to
create a new file with your specific configuration and symlink it as
/etc/cassandra-reaper/cassandra-reaper.yaml to avoid your configuration
from being overwritten during upgrades.

Adapt the config file to suit your setup and then run `sudo service
cassandra-reaper start`.


Source:
http://cassandra-reaper.io/docs/download/install/#service-configuration


Hope that helps!

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Tue, Apr 24, 2018 at 6:51 PM, Abdul Patel <abd786...@gmail.com> wrote:

> Hi All,
>
> For reaper do we need both file or only one?
>
> Cassandra.reaper.yaml
> Cassandra-reaper-cassandra.yaml
>


Re: [EXTERNAL] Cassandra vs MySQL

2018-03-20 Thread Joaquin Casares
Hello Oliver,

The first thing that I check when seeing if a workload will work well
within Cassandra is by looking at it's read patterns. Once the read
patterns can be written down on paper, we need to figure out how the write
patterns will populate the required tables. Since you know enough about
CQL, it's mainly about checking to see how denormalization is going to work
out for on-disk read access requests.

Once the read and write patterns are known, we can see if Cassandra will be
a good fit for denormalizing your workflow and thereby benefiting from a
datastore that can scale out horizontally. If your datastore can scale out
horizontally then Cassandra should be faster than a single node MySQL
cluster. If your datastore has too many relational requirements, is built
in for a queue-like purpose, or other edge cases, then it doesn't matter
how fast Cassandra is if it's not the correct tool for the job.

I hope that helps align your discovery/investigation process. :)

Cheers,

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Tue, Mar 20, 2018 at 1:44 PM, Oliver Ruebenacker <cur...@gmail.com>
wrote:

>
>  Hello,
>
>   Thanks for all the responses.
>
>   I do know some SQL and CQL, so I know the main differences. You can do
> joins in MySQL, but the bigger your data, the less likely you want to do
> that.
>
>   If you are a team that wants to consider migrating from MySQL to
> Cassandra, you need some reason to believe that it is going to be faster.
> What evidence is there?
>
>   Even the Cassandra home page has references to benchmarks to make the
> case for Cassandra. Unfortunately, they seem to be about five to six years
> old. It doesn't make sense to keep them there if you just can't compare.
>
>  Best, Oliver
>
> On Tue, Mar 20, 2018 at 1:13 PM, Durity, Sean R <
> sean_r_dur...@homedepot.com> wrote:
>
>> I’m not sure there is a fair comparison. MySQL and Cassandra have
>> different ways of solving related (but not necessarily the same) problems
>> of storing and retrieving data.
>>
>>
>>
>> The data model between MySQL and Cassandra is likely to be very
>> different. The key for Cassandra is that you need to model for the queries
>> that will be executed. If you cannot know the queries ahead of time,
>> Cassandra is not the best choice. If table scans are typically required,
>> Cassandra is not a good choice. If you need more than a few hundred tables
>> in a cluster, Cassandra is not a good choice.
>>
>>
>>
>> If multi-datacenter replication is required, Cassandra is an awesome
>> choice. If you are going to always query by a partition key (or primary
>> key), Cassandra is a great choice. The nice thing is that the performance
>> scales linearly, so additional data is fine (as long as you add nodes) –
>> again, if your data model is designed for Cassandra. If you like
>> no-downtime upgrades and extreme reliability and availability, Cassandra is
>> a great choice.
>>
>>
>>
>> Personally, I hope to never have to use/support MySQL again, and I love
>> working with Cassandra. But, Cassandra is not the choice for all data
>> problems.
>>
>>
>>
>>
>>
>> Sean Durity
>>
>>
>>
>> *From:* Oliver Ruebenacker [mailto:cur...@gmail.com]
>> *Sent:* Monday, March 12, 2018 3:58 PM
>> *To:* user@cassandra.apache.org
>> *Subject:* [EXTERNAL] Cassandra vs MySQL
>>
>>
>>
>>
>>
>>  Hello,
>>
>>   We have a project currently using MySQL single-node with 5-6TB of data
>> and some performance issues, and we plan to add data up to a total size of
>> maybe 25-30TB.
>>
>>   We are thinking of migrating to Cassandra. I have been trying to find
>> benchmarks or other guidelines to compare MySQL and Cassandra, but most of
>> them seem to be five years old or older.
>>
>>   Is there some good more recent material?
>>
>>   Thanks!
>>
>>  Best, Oliver
>>
>>
>> --
>>
>> Oliver Ruebenacker
>>
>> Senior Software Engineer, Diabetes Portal
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__www.type2diabetesgenetics.org_=DwMFaQ=MtgQEAMQGqekjTjiAhkudQ=aC_gxC6z_4f9GLlbWiKzHm1vucZTtVYWDDvyLkh8IaQ=j3Lz6pcGNV-FgBKxSeA0Lj6Jh2PC7f53PrXNjGYOPiU=1qS6jO1gSrBpPz6yc33IUcVUA-Q0jKm6jmjJr1u89Tc=>,
>> Broad Institute
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__www.broadinstitute.org_=DwMFaQ=MtgQEAMQGqekjTjiAhkudQ=aC_gxC6z_4f9GLlbWiKzHm1vucZTtVYWDDvyLkh8IaQ=j3Lz6pcGNV-FgBKxSeA0Lj6Jh2PC7f53PrXNjGYOPiU=bzHFcavS9i7dzp6ahF4aLzSmH_LukAHXbiiLk03LeD8=>
>&g

Re: Whch version is the best version to run now?

2018-03-02 Thread Joaquin Casares
Hello Kenneth,

We've been recommending 3.11.2 for new projects since it comes with quite a
few performance improvements and bug fixes not available in 3.0.16.

Upgrading to newer versions is taken on a case-by-case basis, but if 2.x is
pretty important to the customer, 2.2.12 is our go-to recommendation.

Cheers,

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, Mar 2, 2018 at 4:37 PM, Kenneth Brotman <
kenbrot...@yahoo.com.invalid> wrote:

> Seems like a lot of people are running old versions of Cassandra.  What is
> the best version, most reliable stable version to use now?
>
>
>
> Kenneth Brotman
>


Re: Indexing for lots of numbers

2018-01-05 Thread Joaquin Casares
Hello Oliver,

I don't see this being a particularly good fit for Cassandra, but I hope
someone confirms this.

However, your use case did look interesting for another project I've
interacted with indirectly, Pilosa, which used to have a Cassandra backend
before a complete Golang rewrite: https://github.com/pilosa/pilosa.

It seems as though the intersect PQL command may work for your use case:
https://www.pilosa.com/docs/latest/query-language/#intersect.

Cheers,

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, Jan 5, 2018 at 5:15 PM, Oliver Ruebenacker <cur...@gmail.com> wrote:

>
>  Hello,
>
>   Let's say I have a table that has one column with a unique id as a
> primary key, and then hundreds of columns of floats, although a large
> fraction of cells are empty.
>
>   I want to create an application that allows users to pick one or more
> number columns, specify a condition (equality or inequality) for each
> number column picked, and then get all rows that satisfy all the conditions
> provided. An empty cell would mean condition is not fulfilled.
>
>   (for example, find all rows for which column35 < 0.5 AND column72 > 1.5
> AND column236 < 0.1)
>
>   What's the best way to enable this? A secondary index for each number
> column? Would I have to obtain all columns that satisfy one condition and
> filter them down, or is there a better way? Does Cassandra support finding
> out which index best to start with?
>
>   Thanks!
>
>  Best, Oliver
>
> --
> Oliver Ruebenacker
> Senior Software Engineer, Diabetes Portal
> <http://www.type2diabetesgenetics.org/>, Broad Institute
> <http://www.broadinstitute.org/>
>
>


Re: Difference between yum and git

2017-05-10 Thread Joaquin Casares
Hi Yuji,

Were both tests run from the same machine at close the same time? If not,
noisy neighbors may be affecting your performance on different AWS
instances.

You should verify that you're using the same version of Java during both
tests.

Also, ensure that you're using the same test instance (that is not running
Cassandra) to connect to both Cassandra clusters.

Cheers,

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Wed, May 10, 2017 at 5:01 AM, Yuji Ito <y...@imagine-orb.com> wrote:

> Hi all,
>
> I'm trying a simple performance test.
> The test requests select operations (CL.SERIAL or CL.QUORUM) by increasing
> the number of threads.
> There is the difference of the performance between C* installed by yum and
> C* which I built by myself.
> What causes the difference?
>
> I use C* 2.2.8.
> One of them was installed by yum (# yum install cassandra22).
> Another was acquired by git from https://github.com/apache/cass
> andra/tree/cassandra-2.2.8 and built it by myself.
> I changed cassandra.yaml to set `commitlog_sync: batch` and
> `commitlog_sync_batch_window_in_ms: 2`.
>
> My environment:
> - a cluster has 3 nodes
> - node: AWS EC2 m4.large with 200 IOPS EBS volume
> - Replication Factor: 3
> - 1 rows
>
> Result:
> ** yum
>  select (CL.SERIAL) 
> threads  operations/sec
> 1   188
> 2   156
> 4   434
> 8   396
> 16  837
> 32  1176
> 64  2206
> 128 4115
> 256 7272
>
> ** git
>  select (CL.SERIAL) 
> threads  operations/sec
> 1   192
> 2   162
> 4   264
> 8   446
> 16  733
> 32  1114
> 64  1715
> 128 2776
> 256 3920
>
> ** yum
>  select (CL.QUORUM) 
> threads  operations/sec
> 1   434
> 2   909
> 4   1481
> 8   1904
> 16  2666
> 32  3106
> 64  3555
> 128 5000
> 256 9014
>
> ** git
>  select (CL.QUORUM) 
> threads  operations/sec
> 1   666
> 2   1538
> 4   2500
> 8   
> 16  4210
> 32  5333
> 64  6597
> 128 7356
> 256 8075
>
> Thanks,
> Yuji
>
>


Re: OOM on Apache Cassandra on 30 Plus node at the same time

2017-03-03 Thread Joaquin Casares
Hello Shravan,

Typically asynchronous requests are recommended over batch statements since
batch statements will cause more work on the coordinator node while
individual requests, when using a TokenAwarePolicy, will hit a specific
coordinator, perform a local disk seek, and return the requested
information.

The only times that using batch statements are ideal is if writing to the
same partition key, even if it's across multiple tables when using the same
hashing algorithm (like murmur3).

Could you provide a bit of insight into what the batch statement was trying
to accomplish and how many child statements were bundled up within that
batch?

Cheers,

Joaquin

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, Mar 3, 2017 at 11:18 AM, Shravan Ch <chall...@outlook.com> wrote:

> Hello,
>
> More than 30 plus Cassandra servers in the primary DC went down OOM
> exception below. What puzzles me is the scale at which it happened (at the
> same minute). I will share some more details below.
>
> System Log: http://pastebin.com/iPeYrWVR
> GC Log: http://pastebin.com/CzNNGs0r
>
> <http://pastebin.com/CzNNGs0r>During the OOM I saw lot of WARNings like
> the below (these were there for quite sometime may be weeks)
> *WARN  [SharedPool-Worker-81] 2017-03-01 19:55:41,209
> BatchStatement.java:252 - Batch of prepared statements for [keyspace.table]
> is of size 225455, exceeding specified threshold of 65536 by 159919.*
>
> * Environment:*
> We are using ApacheCassandra-2.1.9 on Multi DC cluster. Primary DC (more
> C* nodes on SSD and apps run here)  and secondary DC (geographically remote
> and more like a DR to primary) on SAS drives.
> Cassandra config:
>
> Java 1.8.0_65
> Garbage Collector: G1GC
> memtable_allocation_type: offheap_objects
>
> Post this OOM I am seeing huge hints pile up on majority of the nodes and
> the pending hints keep going up. I have increased HintedHandoff
> CoreThreads to 6 but that did not help (I admit that I tried this on one
> node to try).
>
> nodetool compactionstats -H
> pending tasks: 3
> compaction typekeyspace  table
> completed  totalunit   progress
> Compaction  system  hints
> 28.5 GB   92.38 GB   bytes 30.85%
>
>
> Appreciate your inputs here.
>
> Thanks,
> Shravan
>


Re: Cassandra data model right definition

2016-09-30 Thread Joaquin Casares
Hi Mehdi,

I can help clarify a few things.

As Carlos said, Cassandra is a Wide Column Store. Theoretically a row can
have 2 billion columns, but in practice it shouldn't have more than 100
million columns.

Cassandra partitions data to certain nodes based on the partition key(s),
but does provide the option of setting zero or more clustering keys.
Together, the partition key(s) and clustering key(s) form the primary key.

When writing to Cassandra, you will need to provide the full primary key,
however, when reading from Cassandra, you only need to provide the full
partition key.

When you only provide the partition key for a read operation, you're able
to return all columns that exist on that partition with low latency. These
columns are displayed as "CQL rows" to make it easier to reason about.

Consider the schema:

CREATE TABLE foo (
  bar uuid,

  boz uuid,

  baz timeuuid,
  data1 text,

  data2 text,

  PRIMARY KEY ((bar, boz), baz)

);


When you write to Cassandra you will need to send bar, boz, and baz and
optionally data*, if it's relevant for that CQL row. If you chose not to
define a data* field for a particular CQL row, then nothing is stored nor
allocated on disk. But I wouldn't consider that caveat to be "schema-less".

However, all writes to the same bar/boz will end up on the same Cassandra
replica set (a configurable number of nodes) and be stored on the same
place(s) on disk within the SSTable(s). And on disk, each field that's not
a partition key is stored as a column, including clustering keys (this is
optimized in Cassandra 3+, but now we're getting deep into internals).

In this way you can get fast responses for all activity for bar/boz either
over time, or for a specific time, with roughly the same number of disk
seeks, with varying lengths on the disk scans.

Hope that helps!

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, Sep 30, 2016 at 11:40 AM, Carlos Alonso <i...@mrcalonso.com> wrote:

> Cassandra is a Wide Column Store http://db-engines.com/en/system/Cassandra
>
> Carlos Alonso | Software Engineer | @calonso <https://twitter.com/calonso>
>
> On 30 September 2016 at 18:24, Mehdi Bada <mehdi.b...@dbi-services.com>
> wrote:
>
>> Hi all,
>>
>> I have a theoritical question:
>> - Is Apache Cassandra really a column store?
>> Column store mean storing the data as column rather than as a rows.
>>
>> In fact C* store the data as row, and data is partionned with row key.
>>
>> Finally, for me, Cassandra is a row oriented schema less DBMS Is it
>> true for you also???
>>
>> Many thanks in advance for your reply
>>
>> Best Regards
>> Mehdi Bada
>> 
>>
>> *Mehdi Bada* | Consultant
>> Phone: +41 32 422 96 00 | Mobile: +41 79 928 75 48 | Fax: +41 32 422 96
>> 15
>> dbi services, Rue de la Jeunesse 2, CH-2800 Delémont
>> mehdi.b...@dbi-services.com
>> www.dbi-services.com
>>
>>
>>
>>
>> *⇒ dbi services is recruiting Oracle & SQL Server experts ! – Join the
>> team
>> <http://www.dbi-services.com/fr/dbi-services-et-ses-collaborateurs/offres-emplois-opportunites-carrieres/>*
>>
>
>


Re: TRUNCATE throws OperationTimedOut randomly

2016-09-28 Thread Joaquin Casares
Hi George,

Try grepping for WARN and ERROR on the system.logs across all nodes when
you run the command. Could you post any of the recent stacktraces that you
see?

Cheers,

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Wed, Sep 28, 2016 at 12:43 PM, George Sigletos <sigle...@textkernel.nl>
wrote:

> Thanks a lot for your reply.
>
> I understand that truncate is an expensive operation. But throwing a
> timeout while truncating a table that is already empty?
>
> A workaround is to set a high --request-timeout when connecting. Even 20
> seconds is not always enough
>
> Kind regards,
> George
>
>
> On Wed, Sep 28, 2016 at 6:59 PM, Edward Capriolo <edlinuxg...@gmail.com>
> wrote:
>
>> Truncate does a few things (based on version)
>>   truncate takes snapshots
>>   truncate causes a flush
>>   in very old versions truncate causes a schema migration.
>>
>> In newer versions like cassandra 3.4 you have this knob.
>>
>> # How long the coordinator should wait for truncates to complete
>> # (This can be much longer, because unless auto_snapshot is disabled
>> # we need to flush first so we can snapshot before removing the data.)
>> truncate_request_timeout_in_ms: 6
>>
>>
>> In older versions you can not control when this call will timeout, it is
>> fairly normal that it does!
>>
>>
>> On Wed, Sep 28, 2016 at 12:50 PM, George Sigletos <sigle...@textkernel.nl
>> > wrote:
>>
>>> Hello,
>>>
>>> I keep executing a TRUNCATE command on an empty table and it throws
>>> OperationTimedOut randomly:
>>>
>>> cassandra@cqlsh> truncate test.mytable;
>>> OperationTimedOut: errors={}, last_host=cassiebeta-01
>>> cassandra@cqlsh> truncate test.mytable;
>>> OperationTimedOut: errors={}, last_host=cassiebeta-01
>>>
>>> Having a 3 node cluster running 2.1.14. No connectivity problems. Has
>>> anybody come across the same error?
>>>
>>> Thanks,
>>> George
>>>
>>>
>>
>


Re: How long/how many days 'nodetool gossipinfo' will have decommissioned nodes info

2016-09-26 Thread Joaquin Casares
Hello Techpyassa,

Sometimes old gossip information tends to echo around for quite a bit
longer than intended. I'm unsure how long the LEFT messages are supposed to
be echoed for, but if you want to force the removal of a removed node from
gossip, you can use the Assassinate Endpoint JMX command. On larger
clusters, running this command synchronously across all machines may be
required. Instructions on Assassinate Endpoint can be found here:

https://gist.github.com/justenwalker/8338334

If you're planning on recommissioning the same node, upon bootstrapping the
gossiped message should change to a JOINING message overwriting the LEFT
message.

I've personally never checked `nodetool gossipinfo` before recommissioning
a node and typically only ensure the node does not appear in `nodetool
status`.

Hope that helps,

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Sun, Sep 25, 2016 at 2:17 PM, Laxmikanth S <techpya...@gmail.com> wrote:

> Hi,
>
> Recently we have decommissioned nodes from Cassandra cluster , but even
> after nearly 48 hours 'nodetool gossipinfo' still shows the removed nodes(
> as LEFT).
>
> I just wanted to recommission the same node again. So just wanted to know
> , will it create a problem if I recommission the same node(same IP)  again
> while its state is maintained as LEFT in 'nodetool gossipnfo'.
>
>
> Thanks,
> Techpyaasa
>


Re: Upgrading from Cassandra 2.1.12 to 3.0.9

2016-09-23 Thread Joaquin Casares
Hello Razi,

Since you were using a highly stable version of 2.1.x, you may want to
stick with using 3.0.9. 3.7 has introduced many great features, but has not
been as heavily tested in production as 3.0.9.

Running heterogenous clusters, even when using the same major version (e.g.
3.0.8 and 3.0.9), is never recommended. Running a cluster that spans major
releases, for longer than the timespan of a routine upgrade, is strongly
not advised.

Hope that helps!

Joaquin Casares
Consultant
Austin, TX

Apache Cassandra Consulting
http://www.thelastpickle.com

On Fri, Sep 23, 2016 at 8:17 AM, Khaja, Raziuddin (NIH/NLM/NCBI) [C] <
raziuddin.kh...@nih.gov> wrote:

> Hello all,
>
>
>
> I would like to upgrade my Cassandra cluster from 2.1.12 to 3.7.
>
>
>
> I have read the following documentation:
>
> · http://docs.datastax.com/en/latest-upgrade/upgrade/
> cassandra/upgrdCassandra.html
>
> · https://github.com/apache/cassandra/blob/cassandra-3.7/NEWS.txt
>
>
>
> but still had some questions:
>
> · The upgrdCassandra.html page says : “Upgrade from Cassandra 2.1
> versions later or equal to 2.1.9 directly to Cassandra 3.0.x”, and I am not
> sure if this includes 3.x, my question is: Can I upgrade directly to 3.7?
>
> · Can I run a heterogeneous cluster, with one DC running 2.1.12
> and another DC running 3.7?
>
>
>
> Thanks,
>
> -Razi
>


Re: datastax community ami -- broken? --- datastax-agent conflicts with opscenter-agent

2013-12-06 Thread Joaquin Casares
Hello again John,

The AMI has been patched and tested for both DSE and C* and works for the
standard 3 node test. The new code has been pushed to the 2.4 branch so
launching a new set of instances will give you an updated AMI.

You should now have the newest version of OpsCenter installed, along with
the new DataStax Agents (that replace the OpsCenter Agents). Also, I've
patched the two bugs for the motd and for allowing the other nodes to join.

The issue came from a new release of nodetool that contained some
unexpected text that choked up the AMI code as it waited for nodes to come
online.

Let me know if you see any further issues.

Thanks,

Joaquin Casares
DataStax
Software Engineer in Test


http://www.datastax.com/what-we-offer/products-services/training/virtual-training


On Fri, Dec 6, 2013 at 2:02 PM, Joaquin Casares joaq...@datastax.comwrote:

 Hey John,

 Thanks for letting us know. I'm also seeing that the motd gets stuck, but
 if I ctrl-c during the message and try a `nodetool status` there doesn't
 appear to be an issue.

 I'm currently investigating why it's getting stuck. Are you seeing
 something similar?

 What happens if you try to run a `sudo service cassandra restart`? Could
 you send me your /var/log/cassandra/system.log if this still fails?

 Also, I realize now that the package name for the newest version of
 OpsCenter changed from opscenter-free to opscenter. I committed that change
 to our dev AMI and am testing it now. Once this change is made you will no
 longer have to install agents via OpsCenter since they should already be on
 the system. That being said, you won't hit the current OpsCenter/DataStax
 Agent version mismatch you've been hitting.

 Also, we currently only have one AMI. Each time an instance is launched
 the newest version of the code is pulled down from
 https://github.com/riptano/ComboAMI to ensure the code never gets stale
 and can easily keep up with DSE/C* releases as well as AMI code fixes.

 I'll reply again as soon as I figure out and patch this motd issue.

 Thanks,

 Joaquin Casares
 DataStax
 Software Engineer in Test



 http://www.datastax.com/what-we-offer/products-services/training/virtual-training


 On Fri, Dec 6, 2013 at 7:16 AM, John R. Frank j...@mit.edu wrote:

 Hi C* experts,

 In the last 18hrs or so, I have been having trouble getting cassandra
 instances to launch using the datastax community AMI.  Has anyone else seen
 this?

 The instance comes up but then cassandra fails to run.  The most
 informative error message that I've seen so far is in the obscenter agent
 install log (below) --- see especially this line:

 datastax-agent conflicts with opscenter-agent

 I have also attached the ami.log

 I run into these issues when launching with either the browser page:
 https://aws.amazon.com/amis/datastax-auto-clustering-ami-2-2

 Or command line, like this:

 ec2-run-instances ami-814ec2e8 -t m1.large --region us-east-1 --key
 buildbot -n $num_instances -d --clustername Foo --totalnodes
 $num_instances  --version community  --opscenter yes


 Is there a newer AMI?


 Any advice?
 jrf




 Some agent installations failed:

 - 10.6.133.241: Failure installing agent on 10.6.133.241.
 Error output:
 Unable to install the opscenter-agent package. Please check your apt-get
 configuration as well as the agent install log (/var/log/opscenter-agent/
 installer.log).

 Standard output:
 Removing old opscenter-agent files.
 opscenter-agent: unrecognized service
 Reading package lists...
 Building dependency tree...
 Reading state information...
 0 upgraded, 0 newly installed, 0 to remove and 171 not upgraded.
 Starting agent installation process for version 3.2.2
 Reading package lists...
 Building dependency tree...
 Reading state information...
 sysstat is already the newest version.
 0 upgraded, 0 newly installed, 0 to remove and 171 not upgraded.
 Selecting previously unselected package opscenter-agent.
 dpkg: regarding .../opscenter-agent.deb containing opscenter-agent:
 datastax-agent conflicts with opscenter-agent
 opscenter-agent (version 3.2.2) is to be installed.
 opscenter-agent provides opscenter-agent and is to be installed.
 dpkg: error processing opscenter_agent_setup.vYRzL0Tevn/opscenter-agent.deb
 (--install):
 conflicting packages - not installing opscenter-agent
 Errors were encountered while processing:
 opscenter_agent_setup.vYRzL0Tevn/opscenter-agent.deb
 FAILURE: Unable to install the opscenter-agent package. Please check your
 apt-get configuration as well as the agent install log
 (/var/log/opscenter-agent/installer.log).

 Exit code: 1





Re: Questions regarding DataStax AMI

2012-07-27 Thread Joaquin Casares
Oh you're right, sorry about that. The concept of keeping older packages
was recently implemented and while using --version community, you would
need --release 1.0 in order to get 1.0.10.

If you are using --version enterprise, you can use --release 2.0 to get
DataStax Enterprise 2.0 which comes bundled with 1.0.8. As long as you
don't include --analyticsnodes or --searchnodes, you will get vanilla
Cassandra.

As of now, those are the only options available.

Thanks for pointing that out and sorry about the confusion,

Joaquin Casares
DataStax
Software Engineer/Support



On Fri, Jul 27, 2012 at 3:50 AM, Tamar Fraenkel ta...@tok-media.com wrote:

 HI!
 I tried starting a cluster with
 Cluster started with these options:
 --clustername Name --totalnodes 3 --version community --release 1.0.8

 But Cassandra's version is 1.1.2
 Thanks

 *Tamar Fraenkel *
 Senior Software Engineer, TOK Media

 [image: Inline image 1]


 ta...@tok-media.com
 Tel:   +972 2 6409736
 Mob:  +972 54 8356490
 Fax:   +972 2 5612956





 On Thu, Jul 26, 2012 at 9:56 PM, Tamar Fraenkel ta...@tok-media.comwrote:

 What should be the value to create it with Cassandra 1.0.8
 Tamar

 Sent from my iPod

 On Jul 26, 2012, at 7:06 PM, Joaquin Casares joaq...@datastax.com
 wrote:

 Yes, you can easily do this by using the --release version switch as
 found here:
 http://www.datastax.com/docs/1.0/install/install_ami

 Thanks,

 Joaquin Casares
 DataStax
 Software Engineer/Support



 On Thu, Jul 26, 2012 at 12:44 AM, Tamar Fraenkel ta...@tok-media.comwrote:

 Hi!
 Is there a way to launch EC2 cluster from DataStax latest community AMI
 that will run Cassandra 1.0.8 and not 1.1.2?
 Thanks
  *Tamar Fraenkel *
 Senior Software Engineer, TOK Media

 tokLogo.png


 ta...@tok-media.com
 Tel:   +972 2 6409736
 Mob:  +972 54 8356490
 Fax:   +972 2 5612956







tokLogo.png

Re: Questions regarding DataStax AMI

2012-07-26 Thread Joaquin Casares
Yes, you can easily do this by using the --release version switch as
found here:
http://www.datastax.com/docs/1.0/install/install_ami

Thanks,

Joaquin Casares
DataStax
Software Engineer/Support



On Thu, Jul 26, 2012 at 12:44 AM, Tamar Fraenkel ta...@tok-media.comwrote:

 Hi!
 Is there a way to launch EC2 cluster from DataStax latest community AMI
 that will run Cassandra 1.0.8 and not 1.1.2?
 Thanks
 *Tamar Fraenkel *
 Senior Software Engineer, TOK Media

 [image: Inline image 1]

 ta...@tok-media.com
 Tel:   +972 2 6409736
 Mob:  +972 54 8356490
 Fax:   +972 2 5612956





tokLogo.png

Re: Snapshot failing on JSON files in 1.1.0

2012-06-13 Thread Joaquin Casares
Hello Alain,

Yes, the AMI is geared to working out of the box for most dev purposes.

We recently spotted an issue with JNA 3.2.7 on Ubuntu 10.10 not being
picked up.

You can try running `apt-get install libjna-java` but in order for the
change to be activated, you must restart your Cassandra service.

Could you try removing the older version in the /usr/share/java/ folder and
restarting your Cassandra service? You should see a 'JNA mlockall successful'
message in your system.log. If not, then remove the remaining jna.jar and
replace it with the newest jna.jar as found here:
https://github.com/twall/jna.

Upon another restart of your node, you should see the 'JNA mlockall
successful' message as well as the ability to run the snapshots.

I was unable to replicate this on a new instance so all new launches should
include the patched code.

Do let me know if anyone else sees this issue.

Thanks,

Joaquin Casares
DataStax
Software Engineer/Support



On Wed, Jun 13, 2012 at 10:28 AM, Alain RODRIGUEZ arodr...@gmail.comwrote:

 if I do : locate jna

 /opt/java/64/jdk1.6.0_31/db/docs/html/ref/rrefsqljnaturaljoin.html
 /root/.m2/repository/net/java/dev/jna
 /root/.m2/repository/net/java/dev/jna/jna
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7-sources.jar
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7-sources.jar.sha1
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7.jar
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7.jar.sha1
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7.pom
 /root/.m2/repository/net/java/dev/jna/jna/3.2.7/jna-3.2.7.pom.sha1
 /usr/share/doc/libjna-java
 /usr/share/doc/libjna-java/README.Debian
 /usr/share/doc/libjna-java/changelog.Debian.gz
 /usr/share/doc/libjna-java/copyright
 /usr/share/java/jna-3.2.4.jar
 /usr/share/java/jna.jar
 /usr/share/maven-repo/net/java/dev/jna
 /usr/share/maven-repo/net/java/dev/jna/jna
 /usr/share/maven-repo/net/java/dev/jna/jna/3.2.4
 /usr/share/maven-repo/net/java/dev/jna/jna/debian
 /usr/share/maven-repo/net/java/dev/jna/jna/3.2.4/jna-3.2.4.jar
 /usr/share/maven-repo/net/java/dev/jna/jna/3.2.4/jna-3.2.4.pom
 /usr/share/maven-repo/net/java/dev/jna/jna/debian/jna-debian.jar
 /usr/share/maven-repo/net/java/dev/jna/jna/debian/jna-debian.pom
 /var/cache/apt/archives/libjna-java_3.2.4-2_amd64.deb
 /var/lib/dpkg/info/libjna-java.list
 /var/lib/dpkg/info/libjna-java.md5sums

 So what version am I using (jna 3.2.7 or 3.2.4 ?)

 Should I do an apt-get install libjna-java and does this need a restart ?

 Alain

 2012/6/13 Alain RODRIGUEZ arodr...@gmail.com:
  Hi Aaron,
 
  We are using Ubuntu (AMI Datastax 1.0.9 as I said).
 
  Release:10.10
  Codename:   maverick
 
  ERROR [RMI TCP Connection(37732)-10.248.10.94] 2012-06-13 15:00:17,157
  CLibrary.java (line 153) Unable to create hard link
  com.sun.jna.LastErrorException: errno was 1
 at org.apache.cassandra.utils.CLibrary.link(Native Method)
 at
 org.apache.cassandra.utils.CLibrary.createHardLink(CLibrary.java:145)
 at
 org.apache.cassandra.io.sstable.SSTableReader.createLinks(SSTableReader.java:857)
 at
 org.apache.cassandra.db.ColumnFamilyStore.snapshotWithoutFlush(ColumnFamilyStore.java:1412)
 at
 org.apache.cassandra.db.ColumnFamilyStore.snapshot(ColumnFamilyStore.java:1462)
 at org.apache.cassandra.db.Table.snapshot(Table.java:210)
 at
 org.apache.cassandra.service.StorageService.takeSnapshot(StorageService.java:1710)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at
 com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:93)
 at
 com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:27)
 at
 com.sun.jmx.mbeanserver.MBeanIntrospector.invokeM(MBeanIntrospector.java:208)
 at
 com.sun.jmx.mbeanserver.PerInterface.invoke(PerInterface.java:120)
 at
 com.sun.jmx.mbeanserver.MBeanSupport.invoke(MBeanSupport.java:262)
 at
 com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:836)
 at
 com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:761)
 at
 javax.management.remote.rmi.RMIConnectionImpl.doOperation(RMIConnectionImpl.java:1427)
 at
 javax.management.remote.rmi.RMIConnectionImpl.access$200(RMIConnectionImpl.java:72)
 at
 javax.management.remote.rmi.RMIConnectionImpl$PrivilegedOperation.run(RMIConnectionImpl.java:1265)
 at
 javax.management.remote.rmi.RMIConnectionImpl.doPrivilegedOperation(RMIConnectionImpl.java:1360

Re: EC2 raid0 disks ?

2011-10-04 Thread Joaquin Casares
Correct. Not with ephemeral storage.

Here's a complete list of the drives that are attached:
http://docs.amazonwebservices.com/AWSEC2/latest/UserGuide/index.html?InstanceStorage.html#StorageOnInstanceTypes

Joaquin Casares
DataStax
Software Engineer/Support



On Tue, Oct 4, 2011 at 4:01 PM, Yang tedd...@gmail.com wrote:

 it seems that how many virtual disks you can have is fixed:

 on m2.4xlarge you have 2 disks, while on m2.2xlarge you have only 1,
 so I can't setup a raid0 on m2.2xlarge

 am I correct?

 Thanks
 Yang



Re: EC2 raid0 disks ?

2011-10-04 Thread Joaquin Casares
Hello again,

Also, EBS volumes can be attached, but the performance issues cause other
issues when running a healthy cluster. From experience running clusters on
EBS volumes bring their own set of unique problems and are harder to debug.

Here's a quick link that provides a bit more background information on why
it's not the best fit for Cassandra.
http://www.mail-archive.com/user@cassandra.apache.org/msg11022.html

Thanks,
Joaquin Casares
DataStax
Software Engineer/Support



2011/10/4 Yi Yang i...@iyyang.com

 AFAIK it's around 450G per ephemeral disk.
 BTW randomly you can get high performance EBS drives as well. Performance
 are good for DB but are random in IOps.
 --Original Message--
 From: Yang
 To: user@cassandra.apache.org
 ReplyTo: user@cassandra.apache.org
 Subject: EC2 raid0 disks ?
 Sent: Oct 5, 2011 5:01 AM

 it seems that how many virtual disks you can have is fixed:

 on m2.4xlarge you have 2 disks, while on m2.2xlarge you have only 1,
 so I can't setup a raid0 on m2.2xlarge

 am I correct?

 Thanks
 Yang

 從我的 BlackBerry(R) 無線裝置


Re: EC2 raid0 disks ?

2011-10-04 Thread Joaquin Casares
Not a problem!

We ran a test a few months back and know it's better to use RAID0 vs. just
one mount directory. It slips my mind whether the test was also run for
multiple directories vs. RAID0 as well.

We chose the RAID0 method for the AMI as to avoid confusion and allow for
all the sstables to be on one drive and easier to find. This is also in line
with many setups that we see in the wild.

Thanks,

Joaquin Casares
DataStax
Software Engineer/Support



2011/10/4 Yang tedd...@gmail.com

 Thanks guys.

 btw, what is the performance difference between doing a raid0 on the
 multiple ephemeral drives available, and then assign it to cassandra
 data directory, vs creating a mount on each of these drives, and then
 specify all of these to cassandra's data directory list?

 since these drives are all virtual, would there be any benefit at all
 in doing a raid0 ?

 Yang

 2011/10/4 Joaquin Casares joaq...@datastax.com:
  Hello again,
  Also, EBS volumes can be attached, but the performance issues cause other
  issues when running a healthy cluster. From experience running clusters
 on
  EBS volumes bring their own set of unique problems and are harder to
 debug.
  Here's a quick link that provides a bit more background information on
 why
  it's not the best fit for Cassandra.
  http://www.mail-archive.com/user@cassandra.apache.org/msg11022.html
  Thanks,
  Joaquin Casares
  DataStax
  Software Engineer/Support
 
 
  2011/10/4 Yi Yang i...@iyyang.com
 
  AFAIK it's around 450G per ephemeral disk.
  BTW randomly you can get high performance EBS drives as well.
 Performance
  are good for DB but are random in IOps.
  --Original Message--
  From: Yang
  To: user@cassandra.apache.org
  ReplyTo: user@cassandra.apache.org
  Subject: EC2 raid0 disks ?
  Sent: Oct 5, 2011 5:01 AM
 
  it seems that how many virtual disks you can have is fixed:
 
  on m2.4xlarge you have 2 disks, while on m2.2xlarge you have only 1,
  so I can't setup a raid0 on m2.2xlarge
 
  am I correct?
 
  Thanks
  Yang
 
  從我的 BlackBerry(R) 無線裝置
 



Re: Cassandra start/stop scripts

2011-07-26 Thread Joaquin Casares
Did you install via a package or tarball binaries?

Packages allow you to run cassandra as a service with
sudo service cassandra start|stop

But if you are running via tarballs, then yes, running a kill command
against Cassandra is the way to do it since Cassandra runs in crash-only
mode. Kill pid would work however.

Thanks,

Joaquin Casares
DataStax
Software Engineer/Support



On Tue, Jul 26, 2011 at 12:19 PM, Priyanka priya...@gmail.com wrote:

 I do the same way...

 On Tue, Jul 26, 2011 at 1:07 PM, mcasandra [via [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=6622997i=0]
 [hidden email] http://user/SendEmail.jtp?type=nodenode=6622997i=1wrote:

 I need to write cassandra start/stop script. Currently I run cassandra
 to start and kill -9 to stop.

 Is this the best way? kill -9 doesn't sound right :) Wondering how others
 do it.

 --
  If you reply to this email, your message will be added to the discussion
 below:

 http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Cassandra-start-stop-scripts-tp6622977p6622977.html
  To start a new topic under [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=6622997i=2,
 email [hidden email]http://user/SendEmail.jtp?type=nodenode=6622997i=3
 To unsubscribe from [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=6622997i=4,
 click here.



 --
 View this message in context: Re: Cassandra start/stop 
 scriptshttp://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Cassandra-start-stop-scripts-tp6622977p6622997.html

 Sent from the cassandra-u...@incubator.apache.org mailing list 
 archivehttp://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/at 
 Nabble.com.



Re: Possible bug in Cassandra 0.8.1?

2011-07-18 Thread Joaquin Casares
For the Too many open files error see:
http://www.datastax.com/docs/0.8/troubleshooting/index#java-reports-an-error-saying-there-are-too-many-open-files

Restart the node and see if the node is able to complete the pending repair
this time. Your node may have just been stuck on this error that caused
everything else to halt.

Joaquin Casares
DataStax
Software Engineer/Support



On Mon, Jul 18, 2011 at 3:12 PM, Sameer Farooqui cassandral...@gmail.comwrote:

 I'm running into a quirky issue with Brisk 1.0 Beta 2 (w/ Cassandra 0.8.1).

 I think the last node in our cluster is having problems (10.201.x.x).
 OpsCenter and nodetool ring (run from that node) show the node as down, but
 the rest of the cluster sees it as up.

 If I run nodetool ring from one of the first 11 nodes, I get this output...
 everything is up:

 ubuntu@ip-10-85-x-x:~/brisk/resources/cassandra$ bin/nodetool -h localhost
 ring
 Address DC  RackStatus State   Load
 OwnsToken

 148873535527910577765226390751398592512
 10.2.x.xDC1 RAC1Up Normal  901.57 GB
 12.50%  0
 10.116.x.xDC2 RAC1Up Normal  258.22 GB
 6.25%   10633823966279326983230456482242756608
 10.110.x.xDC1 RAC1Up Normal  129.07 GB
 6.25%   21267647932558653966460912964485513216
 10.2.x.x  DC1 RAC1Up Normal  128.5 GB
 12.50%  42535295865117307932921825928971026432
 10.114.x.x   DC2 RAC1Up Normal  257.31 GB
 6.25%   53169119831396634916152282411213783040
 10.210.x.x   DC1 RAC1Up Normal  128.66 GB
 6.25%   63802943797675961899382738893456539648
 10.207.x.x   DC1 RAC2Up Normal  643.12 GB
 12.50%  85070591730234615865843651857942052864
 10.85.x.xDC2 RAC1Up Normal  256.76 GB
 6.25%   95704415696513942849074108340184809472
 10.2.x.xDC1 RAC2Up Normal  128.95 GB
 6.25%   106338239662793269832304564822427566080
 10.96.x.xDC1 RAC2Up Normal  128.29 GB   12.50%
 127605887595351923798765477786913079296
 10.194.x.xDC2 RAC1Up Normal  257.14 GB
 6.25%   138239711561631250781995934269155835904
 10.201.x.xDC1 RAC2Up Normal  129.45 GB
 6.25%   148873535527910577765226390751398592512

 However, OpsCenter shows the last node (10.201.x.x) as unresponsive:
 http://blueplastic.com/accenture/unresponsive.PNG

 And if I try to run nodetool ring from the 10.201.x.x node, I get
 connection errors like this:

 ubuntu@ip-10-194-x-x:~/brisk/resources/cassandra$ bin/nodetool -h
 localhost ring
 Error connection to remote JMX agent!
 java.io.IOException: Failed to retrieve RMIServer stub:
 javax.naming.CommunicationException [Root exception is
 java.rmi.ConnectIOException: error during JRMP connection establishment;
 nested exception is:
 java.net.SocketTimeoutException: Read timed out]
 at
 javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:338)
 at
 javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:248)
 at org.apache.cassandra.tools.NodeProbe.connect(NodeProbe.java:141)
 at org.apache.cassandra.tools.NodeProbe.init(NodeProbe.java:111)
 at org.apache.cassandra.tools.NodeCmd.main(NodeCmd.java:559)
 Caused by: javax.naming.CommunicationException [Root exception is
 java.rmi.ConnectIOException: error during JRMP connection establishment;
 nested exception is:
 java.net.SocketTimeoutException: Read timed out]
 at
 com.sun.jndi.rmi.registry.RegistryContext.lookup(RegistryContext.java:101)


 tpstats command also didn't work:

 ubuntu@ip-10-194-x-x:~/brisk/resources/cassandra$ bin/nodetool -h
 localhost tpstats
 Error connection to remote JMX agent!
 java.io.IOException: Failed to retrieve RMIServer stub:
 javax.naming.CommunicationException [Root exception is java.rmi.Co
 nnectIOException: error during JRMP connection establishment; nested
 exception is:
 java.net.SocketTimeoutException: Read timed out]
 at
 javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:338)


 Looking at what's listening on 7199 on the node shows a bunch of results:

 ubuntu@ip-10-194-x-x:~/brisk/resources/cassandra$ sudo netstat -anp | grep
 7199
 tcp0  0 0.0.0.0:71990.0.0.0:*
 LISTEN  1459/java
 tcp8  0 10.194.x.x:7199 10.2.x.x:40135  CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:49835
 CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:55087
 CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:49837
 CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:55647
 CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:49833
 CLOSE_WAIT  -
 tcp8  0 127.0.0.1:7199  127.0.0.1:52935
 CLOSE_WAIT  -
 tcp8  0

Re: Problem with PropertyFileSnitch in Amazon EC2

2011-06-21 Thread Joaquin Casares
Could you verify any security settings that may come into play with Elastic
IPs? You should make sure the appropriate ports are open.

See: http://www.datastax.com/docs/0.8/brisk/install_brisk_ami
for a list of ports in the first chart.

Joaquin Casares
DataStax
Software Engineer/Support



On Mon, Jun 20, 2011 at 7:41 PM, Sameer Farooqui cassandral...@gmail.comwrote:

 Quick update...

 I'm trying to get a 3-node cluster defined the following way in the
 topology.properties file to work first:
 10.68.x.x=DC1:RAC1
 10.198.x.x=DC1:RAC2
 10.204.x.x=DC1:RAC3

 I'll split up the 3rd node into a separate data center later.

 Also, ignore that comment I made about the $BRISK_HOME/lib/ folder not
 existing. When you run ANT, I believe it populates correctly, but I'll have
 to confirm/test later.

 Based on Joaquin @ DataStax's suggestion, I tried changing the Seed IP in
 all 3 nodes' YAML file to the Amazon Private IP, instead of the Elastic IP.
 After this change, all three nodes joined the ring correctly:

 ubuntu@ip-10-68-x-x:~/brisk-1.0~beta1.2/resources/cassandra/conf$
 ../bin/nodetool -h localhost ring
 Address Status State   LoadOwnsToken

  113427455640312821154458202477256070485
 10.68.x.x Up Normal  10.9 KB 33.33%  0
 10.198.x.x  Up Normal  15.21 KB33.33%
  56713727820156410577229101238628035242
 10.204.x.x  Up Normal  6.55 KB 33.33%
  113427455640312821154458202477256070485

 PasteBin is down and is showing me a diligent cat typing on a keyboard, so
 I uploaded some relevant DEBUG level log files here:

 http://blueplastic.com/accenture/N1-system-seed_is_ElasticIP.log (problem
 exists)
 http://blueplastic.com/accenture/N2-system-seed_is_ElasticIP.log (problem
 exists)

 http://blueplastic.com/accenture/N1-system-seed_is_privateIP.log(everything 
 works)
 http://blueplastic.com/accenture/N2-system-seed_is_privateIP.log(everything 
 works)


 But if I want to set up the Brisk cluster across Amazon regions, I have to
 be able to use the Elastic IP for the seed. Also, using v 0.7.4 of Cassandra
 in Amazon, we successfully set up a 30+ node cluster using 3 seed nodes
 which were declared in the YAML file using Elastic IPs. All 30 nodes were in
 the same region and availability zone. So, in an older version of Cassandra,
 providing the Seeds as Elastic IP used to work.

 In my current setup, even though nodes 1  2 are in the same region 
 availability zone, I can't seem to get them to join the same ring correctly.


 Here is what the system log file shows when I declare the Seed using
 Elastic IP:
 INFO [Thread-4] 2011-06-21 00:10:30,849 BriskDaemon.java (line 187)
 Listening for thrift clients...
 DEBUG [GossipTasks:1] 2011-06-21 00:10:31,608 Gossiper.java (line 201)
 Assuming current protocol version for /50.17.x.x
 DEBUG [WRITE-/50.17.212.84] 2011-06-21 00:10:31,610
 OutboundTcpConnection.java (line 161) attempting to connect to /50.17.x.x
 DEBUG [GossipTasks:1] 2011-06-21 00:10:32,610 Gossiper.java (line 201)
 Assuming current protocol version for /50.17.x.x
 DEBUG [ScheduledTasks:1] 2011-06-21 00:10:32,613 StorageLoadBalancer.java
 (line 334) Disseminating load info ...
 DEBUG [GossipTasks:1] 2011-06-21 00:10:33,611 Gossiper.java (line 201)
 Assuming current protocol version for /50.17.x.x
 DEBUG [GossipTasks:1] 2011-06-21 00:10:34,612 Gossiper.java (line 201)
 Assuming current protocol version for /50.17.x.x


 But when I use private IP, the log shows:

 INFO [Thread-4] 2011-06-21 00:19:47,993 BriskDaemon.java (line 187)
 Listening for thrift clients...
 DEBUG [ScheduledTasks:1] 2011-06-21 00:19:49,769 StorageLoadBalancer.java
 (line 334) Disseminating load info ...
 DEBUG [WRITE-/10.198.126.193] 2011-06-21 00:20:09,658
 OutboundTcpConnection.java (line 161) attempting to connect to /10.198.x.x
  INFO [GossipStage:1] 2011-06-21 00:20:09,690 Gossiper.java (line 637) Node
 /10.198.x.x is now part of the cluster
 DEBUG [GossipStage:1] 2011-06-21 00:20:09,691 MessagingService.java (line
 158) Resetting pool for /10.198.x.x
  INFO [GossipStage:1] 2011-06-21 00:20:09,691 Gossiper.java (line 605)
 InetAddress /10.198.x.x is now UP
 DEBUG [HintedHandoff:1] 2011-06-21 00:20:09,692 HintedHandOffManager.java
 (line 282) Checking remote schema before delivering hints
 DEBUG [HintedHandoff:1] 2011-06-21 00:20:09,692 HintedHandOffManager.java
 (line 274) schema for /10.198.x.x matches local schema
 DEBUG [HintedHandoff:1] 2011-06-21 00:20:09,692 HintedHandOffManager.java
 (line 288) Sleeping 11662ms to stagger hint delivery

 - Sameer


 On Mon, Jun 20, 2011 at 2:28 PM, Sameer Farooqui 
 cassandral...@gmail.comwrote:

 Hi,

 I'm setting up a 3 node test cluster in multiple Amazon Availability Zones
 to test cross-zone internode communication (and eventually cross-region
 communications).

 But I wanted to start with a cross-zone setup and am having trouble
 getting the nodes to connect to each other and join one 3-node ring. All
 nodes just seem

Re: Ec2 Stress Results

2011-04-25 Thread Joaquin Casares
Did the images have EBS storage or Instance Store storage?

Typically EBS volumes aren't the best to be benchmarking against:
http://www.mail-archive.com/user@cassandra.apache.org/msg11022.html

Joaquin Casares
DataStax
Software Engineer/Support



On Wed, Apr 20, 2011 at 5:12 PM, Jonathan Ellis jbel...@gmail.com wrote:

 A few months ago I was seeing 12k writes/s on a single EC2 XL. So
 something is wrong.

 My first suspicion is that your client node may be the bottleneck.

 On Wed, Apr 20, 2011 at 2:56 PM, Alex Araujo
 cassandra-us...@alex.otherinbox.com wrote:
  Does anyone have any Ec2 benchmarks/experiences they can share?  I am
 trying
  to get a sense for what to expect from a production cluster on Ec2 so
 that I
  can compare my application's performance against a sane baseline.  What I
  have done so far is:
 
  1. Lunched a 4 node cluster of m1.xlarge instances in the same
 availability
  zone using PyStratus (https://github.com/digitalreasoning/PyStratus).
 Each
  node has the following specs (according to Amazon):
  15 GB memory
  8 EC2 Compute Units (4 virtual cores with 2 EC2 Compute Units each)
  1,690 GB instance storage
  64-bit platform
 
  2. Changed the default PyStratus directories in order to have commit logs
 on
  the root partition and data files on ephemeral storage:
  commitlog_directory: /var/cassandra-logs
  data_file_directories: [/mnt/cassandra-data]
 
  2. Gave each node 10GB of MAX_HEAP; 1GB HEAP_NEWSIZE in
  conf/cassandra-env.sh
 
  3. Ran `contrib/stress/bin/stress -d node1,..,node4 -n 1000 -t 100`
 on a
  separate m1.large instance:
  total,interval_op_rate,interval_key_rate,avg_latency,elapsed_time
  ...
  9832712,7120,7120,0.004948514851485148,842
  9907616,7490,7490,0.0043189949802413755,852
  9978357,7074,7074,0.004560353967289125,863
  1000,2164,2164,0.004065933558194335,867
 
  4. Truncated Keyspace1.Standard1:
  # /usr/local/apache-cassandra/bin/cassandra-cli -host localhost -port
 9160
  Connected to: Test Cluster on x.x.x.x/9160
  Welcome to cassandra CLI.
 
  Type 'help;' or '?' for help. Type 'quit;' or 'exit;' to quit.
  [default@unknown] use Keyspace1;
  Authenticated to keyspace: Keyspace1
  [default@Keyspace1] truncate Standard1;
  null
 
  5. Expanded the cluster to 8 nodes using PyStratus and sanity checked
 using
  nodetool:
  # /usr/local/apache-cassandra/bin/nodetool -h localhost ring
  Address Status State   LoadOwns
  Token
  x.x.x.x  Up Normal  1.3 GB  12.50%
  21267647932558653966460912964485513216
  x.x.x.x   Up Normal  3.06 GB 12.50%
  42535295865117307932921825928971026432
  x.x.x.x Up Normal  1.16 GB 12.50%
  63802943797675961899382738893456539648
  x.x.x.x   Up Normal  2.43 GB 12.50%
  85070591730234615865843651857942052864
  x.x.x.x   Up Normal  1.22 GB 12.50%
  106338239662793269832304564822427566080
  x.x.x.xUp Normal  2.74 GB 12.50%
  127605887595351923798765477786913079296
  x.x.x.xUp Normal  1.22 GB 12.50%
  148873535527910577765226390751398592512
  x.x.x.x   Up Normal  2.57 GB 12.50%
  170141183460469231731687303715884105728
 
  6. Ran `contrib/stress/bin/stress -d node1,..,node8 -n 1000 -t 100`
 on a
  separate m1.large instance again:
  total,interval_op_rate,interval_key_rate,avg_latency,elapsed_time
  ...
  9880360,9649,9649,0.003210443956226165,720
  9942718,6235,6235,0.003206934154398794,731
  9997035,5431,5431,0.0032615939761032457,741
  1000,296,296,0.002660033726812816,742
 
  In a nutshell, 4 nodes inserted at 11,534 writes/sec and 8 nodes inserted
 at
  13,477 writes/sec.
 
  Those numbers seem a little low to me, but I don't have anything to
 compare
  to.  I'd like to hear others' opinions before I spin my wheels with with
  number of nodes, threads,  memtable, memory, and/or GC settings.  Cheers,
  Alex.
 



 --
 Jonathan Ellis
 Project Chair, Apache Cassandra
 co-founder of DataStax, the source for professional Cassandra support
 http://www.datastax.com



Re: help

2011-04-12 Thread Joaquin Casares
http://wiki.apache.org/cassandra/FAQ#unsubscribe

http://wiki.apache.org/cassandra/FAQ#unsubscribeIs this what you're
looking for?

Joaquin Casares
DataStax
Software Engineer/Support



On Tue, Apr 12, 2011 at 2:03 PM, Denis Kirpichenkov 
den.doki.kirpichen...@gmail.com wrote: