Re: EBS SSD - Cassandra ?

2014-06-19 Thread Alain RODRIGUEZ
Ok, looks fair enough.

Thanks guys. I would be great to be able to add disks when amount of data
raises and add nodes when throughput increases... :)


2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:


 http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

 From the link:

 EBS volumes are not recommended for Cassandra data volumes for the
 following reasons:

 * EBS volumes contend directly for network throughput with standard
 packets. This means that EBS throughput is likely to fail if you saturate a
 network link.
  * EBS volumes have unreliable performance. I/O performance can be
 exceptionally slow, causing the system to back load reads and writes until
 the entire cluster becomes unresponsive.
  * Adding capacity by increasing the number of EBS volumes per host does
 not scale. You can easily surpass the ability of the system to keep
 effective buffer caches and concurrently serve requests for all of the data
 it is responsible for managing.

 Still applies, especially the network contention and latency issues.

 Ben Bromhead
 Instaclustr | www.instaclustr.com | @instaclustr
 http://twitter.com/instaclustr | +61 415 936 359

 On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

 While they guarantee IOPS, they don't really make any guarantees about
 latency. Since EBS goes over the network, there's so many things in the
 path of getting at your data, I would be concerned with random latency
 spikes, unless proven otherwise.

 Thanks,
 Daniel


 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 In this document it is said :


- Provisioned IOPS (SSD) - Volumes of this type are ideal for the
most demanding I/O intensive, transactional workloads and large relational
or NoSQL databases. This volume type provides the most consistent
performance and allows you to provision the exact level of performance you
need with the most predictable and consistent performance. With this type
of volume you provision exactly what you need, and pay for what you
provision. Once again, you can achieve up to 48,000 IOPS by connecting
multiple volumes together using RAID.



 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

 Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that this
 hardware architecture might be useful alongside Cassandra, right ?

 Alain







Re: EBS SSD - Cassandra ?

2014-06-19 Thread Benedict Elliott Smith
I would say this is worth benchmarking before jumping to conclusions. The
network being a bottleneck (or latency causing) for EBS is, to my
knowledge, supposition, and instances can be started with direct
connections to EBS if this is a concern. The blog post below shows that
even without SSDs the EBS-optimised provisioned-IOPS instances show pretty
consistent latency numbers, although those latencies are higher than you
would typically expect from locally attached storage.

http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/

Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up
with number of nodes, not with depth of nodes (as Ben mentions, saturating
a single node's data capacity is pretty easy these days. CPUs rapidly
become the bottleneck as you try to go deep). However the argument that EBS
cannot provide consistent performance seems overly pessimistic, and should
probably be empirically determined for your use case.


On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:

 Ok, looks fair enough.

 Thanks guys. I would be great to be able to add disks when amount of data
 raises and add nodes when throughput increases... :)


 2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:


 http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

 From the link:

 EBS volumes are not recommended for Cassandra data volumes for the
 following reasons:

 • EBS volumes contend directly for network throughput with standard
 packets. This means that EBS throughput is likely to fail if you saturate a
 network link.
  • EBS volumes have unreliable performance. I/O performance can be
 exceptionally slow, causing the system to back load reads and writes until
 the entire cluster becomes unresponsive.
  • Adding capacity by increasing the number of EBS volumes per host does
 not scale. You can easily surpass the ability of the system to keep
 effective buffer caches and concurrently serve requests for all of the data
 it is responsible for managing.

 Still applies, especially the network contention and latency issues.

 Ben Bromhead
 Instaclustr | www.instaclustr.com | @instaclustr
 http://twitter.com/instaclustr | +61 415 936 359

 On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

 While they guarantee IOPS, they don't really make any guarantees about
 latency. Since EBS goes over the network, there's so many things in the
 path of getting at your data, I would be concerned with random latency
 spikes, unless proven otherwise.

 Thanks,
 Daniel


 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 In this document it is said :


- Provisioned IOPS (SSD) - Volumes of this type are ideal for the
most demanding I/O intensive, transactional workloads and large 
 relational
or NoSQL databases. This volume type provides the most consistent
performance and allows you to provision the exact level of performance 
 you
need with the most predictable and consistent performance. With this type
of volume you provision exactly what you need, and pay for what you
provision. Once again, you can achieve up to 48,000 IOPS by connecting
multiple volumes together using RAID.



 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

 Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that
 this hardware architecture might be useful alongside Cassandra, right ?

 Alain








Re: EBS SSD - Cassandra ?

2014-06-19 Thread Nate McCall
If someone really wanted to try this it, I recommend adding an Elastic
Network Interface or two for gossip and client/API traffic. This lets EBS
and management traffic have the pre-configured network.


On Thu, Jun 19, 2014 at 6:54 AM, Benedict Elliott Smith 
belliottsm...@datastax.com wrote:

 I would say this is worth benchmarking before jumping to conclusions. The
 network being a bottleneck (or latency causing) for EBS is, to my
 knowledge, supposition, and instances can be started with direct
 connections to EBS if this is a concern. The blog post below shows that
 even without SSDs the EBS-optimised provisioned-IOPS instances show pretty
 consistent latency numbers, although those latencies are higher than you
 would typically expect from locally attached storage.


 http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/

 Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up
 with number of nodes, not with depth of nodes (as Ben mentions, saturating
 a single node's data capacity is pretty easy these days. CPUs rapidly
 become the bottleneck as you try to go deep). However the argument that EBS
 cannot provide consistent performance seems overly pessimistic, and should
 probably be empirically determined for your use case.


 On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 Ok, looks fair enough.

 Thanks guys. I would be great to be able to add disks when amount of data
 raises and add nodes when throughput increases... :)


 2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:


 http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

 From the link:

 EBS volumes are not recommended for Cassandra data volumes for the
 following reasons:

 • EBS volumes contend directly for network throughput with standard
 packets. This means that EBS throughput is likely to fail if you saturate a
 network link.
  • EBS volumes have unreliable performance. I/O performance can be
 exceptionally slow, causing the system to back load reads and writes until
 the entire cluster becomes unresponsive.
  • Adding capacity by increasing the number of EBS volumes per host
 does not scale. You can easily surpass the ability of the system to keep
 effective buffer caches and concurrently serve requests for all of the data
 it is responsible for managing.

 Still applies, especially the network contention and latency issues.

 Ben Bromhead
 Instaclustr | www.instaclustr.com | @instaclustr
 http://twitter.com/instaclustr | +61 415 936 359

 On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

 While they guarantee IOPS, they don't really make any guarantees about
 latency. Since EBS goes over the network, there's so many things in the
 path of getting at your data, I would be concerned with random latency
 spikes, unless proven otherwise.

 Thanks,
 Daniel


 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 In this document it is said :


- Provisioned IOPS (SSD) - Volumes of this type are ideal for the
most demanding I/O intensive, transactional workloads and large 
 relational
or NoSQL databases. This volume type provides the most consistent
performance and allows you to provision the exact level of performance 
 you
need with the most predictable and consistent performance. With this 
 type
of volume you provision exactly what you need, and pay for what you
provision. Once again, you can achieve up to 48,000 IOPS by connecting
multiple volumes together using RAID.



 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

 Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that
 this hardware architecture might be useful alongside Cassandra, right ?

 Alain









-- 
-
Nate McCall
Austin, TX
@zznate

Co-Founder  Sr. Technical Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com


Re: EBS SSD - Cassandra ?

2014-06-19 Thread Russell Bradberry
does an elastic network interface really use a different physical network 
interface? or is it just to give the ability for multiple ip addresses?



On June 19, 2014 at 3:56:34 PM, Nate McCall (n...@thelastpickle.com) wrote:

If someone really wanted to try this it, I recommend adding an Elastic Network 
Interface or two for gossip and client/API traffic. This lets EBS and 
management traffic have the pre-configured network. 


On Thu, Jun 19, 2014 at 6:54 AM, Benedict Elliott Smith 
belliottsm...@datastax.com wrote:
I would say this is worth benchmarking before jumping to conclusions. The 
network being a bottleneck (or latency causing) for EBS is, to my knowledge, 
supposition, and instances can be started with direct connections to EBS if 
this is a concern. The blog post below shows that even without SSDs the 
EBS-optimised provisioned-IOPS instances show pretty consistent latency 
numbers, although those latencies are higher than you would typically expect 
from locally attached storage.

http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/

Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up with 
number of nodes, not with depth of nodes (as Ben mentions, saturating a single 
node's data capacity is pretty easy these days. CPUs rapidly become the 
bottleneck as you try to go deep). However the argument that EBS cannot provide 
consistent performance seems overly pessimistic, and should probably be 
empirically determined for your use case.


On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:
Ok, looks fair enough.

Thanks guys. I would be great to be able to add disks when amount of data 
raises and add nodes when throughput increases... :)


2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:

http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

From the link:

EBS volumes are not recommended for Cassandra data volumes for the following 
reasons:

• EBS volumes contend directly for network throughput with standard packets. 
This means that EBS throughput is likely to fail if you saturate a network link.
• EBS volumes have unreliable performance. I/O performance can be exceptionally 
slow, causing the system to back load reads and writes until the entire cluster 
becomes unresponsive.
• Adding capacity by increasing the number of EBS volumes per host does not 
scale. You can easily surpass the ability of the system to keep effective 
buffer caches and concurrently serve requests for all of the data it is 
responsible for managing.

Still applies, especially the network contention and latency issues. 

Ben Bromhead
Instaclustr | www.instaclustr.com | @instaclustr | +61 415 936 359

On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

While they guarantee IOPS, they don't really make any guarantees about latency. 
Since EBS goes over the network, there's so many things in the path of getting 
at your data, I would be concerned with random latency spikes, unless proven 
otherwise.

Thanks,
Daniel


On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:
In this document it is said :

Provisioned IOPS (SSD) - Volumes of this type are ideal for the most demanding 
I/O intensive, transactional workloads and large relational or NoSQL databases. 
This volume type provides the most consistent performance and allows you to 
provision the exact level of performance you need with the most predictable and 
consistent performance. With this type of volume you provision exactly what you 
need, and pay for what you provision. Once again, you can achieve up to 48,000 
IOPS by connecting multiple volumes together using RAID.


2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

Hi,

I just saw this : 
http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

Since the problem with EBS was the network, there is no chance that this 
hardware architecture might be useful alongside Cassandra, right ?

Alain








--
-
Nate McCall
Austin, TX
@zznate

Co-Founder  Sr. Technical Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

Re: EBS SSD - Cassandra ?

2014-06-19 Thread Nate McCall
Sorry - should have been clear I was speaking in terms of route optimizing,
not bandwidth. No idea as to the implementation (probably instance
specific) and I doubt it actually doubles bandwidth.

Specifically: having an ENI dedicated to API traffic did smooth out some
recent load tests we did for a client. It could be that overall throughput
increases where more a function of cleaner traffic segmentation/smoother
routing. We werent being terribly scientific - was more an artifact of
testing network segmentation.

I'm just going to say that using an ENI will make things better (since
traffic segmentation is always good practice anyway :)  YMMV.



On Thu, Jun 19, 2014 at 3:39 PM, Russell Bradberry rbradbe...@gmail.com
wrote:

 does an elastic network interface really use a different physical network
 interface? or is it just to give the ability for multiple ip addresses?



 On June 19, 2014 at 3:56:34 PM, Nate McCall (n...@thelastpickle.com)
 wrote:

 If someone really wanted to try this it, I recommend adding an Elastic
 Network Interface or two for gossip and client/API traffic. This lets EBS
 and management traffic have the pre-configured network.


 On Thu, Jun 19, 2014 at 6:54 AM, Benedict Elliott Smith 
 belliottsm...@datastax.com wrote:

 I would say this is worth benchmarking before jumping to conclusions. The
 network being a bottleneck (or latency causing) for EBS is, to my
 knowledge, supposition, and instances can be started with direct
 connections to EBS if this is a concern. The blog post below shows that
 even without SSDs the EBS-optimised provisioned-IOPS instances show pretty
 consistent latency numbers, although those latencies are higher than you
 would typically expect from locally attached storage.


 http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/

 Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up
 with number of nodes, not with depth of nodes (as Ben mentions, saturating
 a single node's data capacity is pretty easy these days. CPUs rapidly
 become the bottleneck as you try to go deep). However the argument that EBS
 cannot provide consistent performance seems overly pessimistic, and should
 probably be empirically determined for your use case.


 On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 Ok, looks fair enough.

 Thanks guys. I would be great to be able to add disks when amount of
 data raises and add nodes when throughput increases... :)


 2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:


 http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

 From the link:

 EBS volumes are not recommended for Cassandra data volumes for the
 following reasons:

 • EBS volumes contend directly for network throughput with standard
 packets. This means that EBS throughput is likely to fail if you saturate a
 network link.
 • EBS volumes have unreliable performance. I/O performance can be
 exceptionally slow, causing the system to back load reads and writes until
 the entire cluster becomes unresponsive.
 • Adding capacity by increasing the number of EBS volumes per host does
 not scale. You can easily surpass the ability of the system to keep
 effective buffer caches and concurrently serve requests for all of the data
 it is responsible for managing.

 Still applies, especially the network contention and latency issues.

   Ben Bromhead
  Instaclustr | www.instaclustr.com | @instaclustr
 http://twitter.com/instaclustr | +61 415 936 359

  On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

  While they guarantee IOPS, they don't really make any guarantees
 about latency. Since EBS goes over the network, there's so many things in
 the path of getting at your data, I would be concerned with random latency
 spikes, unless proven otherwise.

 Thanks,
 Daniel


 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com
 wrote:

 In this document it is said :


- Provisioned IOPS (SSD) - Volumes of this type are ideal for the
most demanding I/O intensive, transactional workloads and large 
 relational
or NoSQL databases. This volume type provides the most consistent
performance and allows you to provision the exact level of performance 
 you
need with the most predictable and consistent performance. With this 
 type
of volume you provision exactly what you need, and pay for what you
provision. Once again, you can achieve up to 48,000 IOPS by connecting
multiple volumes together using RAID.



 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

  Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that
 this hardware architecture might be useful alongside Cassandra, right ?

 Alain









 --
 -
 Nate McCall
 Austin, TX
 @zznate

 Co-Founder  

Re: EBS SSD - Cassandra ?

2014-06-19 Thread Ben Bromhead
Irrespective of performance and latency numbers there are fundamental flaws 
with using EBS/NAS and Cassandra, particularly around bandwidth contention and 
what happens when the shared storage medium breaks. Also obligatory reference 
to http://www.eecs.berkeley.edu/~rcs/research/interactive_latency.html.

Regarding ENI

AWS are pretty explicit about it’s impact on bandwidth:

http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/using-eni.html
Attaching another network interface to an instance is not a method to increase 
or double the network bandwidth to or from the dual-homed instance.

So Nate you are right in that it is a function of logical separation helps for 
some reason. 
 

Ben Bromhead
Instaclustr | www.instaclustr.com | @instaclustr | +61 415 936 359

On 20 Jun 2014, at 8:17 am, Nate McCall n...@thelastpickle.com wrote:

 Sorry - should have been clear I was speaking in terms of route optimizing, 
 not bandwidth. No idea as to the implementation (probably instance specific) 
 and I doubt it actually doubles bandwidth. 
 
 Specifically: having an ENI dedicated to API traffic did smooth out some 
 recent load tests we did for a client. It could be that overall throughput 
 increases where more a function of cleaner traffic segmentation/smoother 
 routing. We werent being terribly scientific - was more an artifact of 
 testing network segmentation. 
 
 I'm just going to say that using an ENI will make things better (since 
 traffic segmentation is always good practice anyway :)  YMMV. 
 
 
 
 On Thu, Jun 19, 2014 at 3:39 PM, Russell Bradberry rbradbe...@gmail.com 
 wrote:
 does an elastic network interface really use a different physical network 
 interface? or is it just to give the ability for multiple ip addresses?
 
 
 
 On June 19, 2014 at 3:56:34 PM, Nate McCall (n...@thelastpickle.com) wrote:
 
 If someone really wanted to try this it, I recommend adding an Elastic 
 Network Interface or two for gossip and client/API traffic. This lets EBS 
 and management traffic have the pre-configured network. 
 
 
 On Thu, Jun 19, 2014 at 6:54 AM, Benedict Elliott Smith 
 belliottsm...@datastax.com wrote:
 I would say this is worth benchmarking before jumping to conclusions. The 
 network being a bottleneck (or latency causing) for EBS is, to my knowledge, 
 supposition, and instances can be started with direct connections to EBS if 
 this is a concern. The blog post below shows that even without SSDs the 
 EBS-optimised provisioned-IOPS instances show pretty consistent latency 
 numbers, although those latencies are higher than you would typically expect 
 from locally attached storage.
 
 http://blog.parse.com/2012/09/17/parse-databases-upgraded-to-amazon-provisioned-iops/
 
 Note, I'm not endorsing the use of EBS. Cassandra is designed to scale up 
 with number of nodes, not with depth of nodes (as Ben mentions, saturating a 
 single node's data capacity is pretty easy these days. CPUs rapidly become 
 the bottleneck as you try to go deep). However the argument that EBS cannot 
 provide consistent performance seems overly pessimistic, and should probably 
 be empirically determined for your use case.
 
 
 On Thu, Jun 19, 2014 at 9:50 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:
 Ok, looks fair enough.
 
 Thanks guys. I would be great to be able to add disks when amount of data 
 raises and add nodes when throughput increases... :)
 
 
 2014-06-19 5:27 GMT+02:00 Ben Bromhead b...@instaclustr.com:
 
 http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html
 
 From the link:
 
 EBS volumes are not recommended for Cassandra data volumes for the following 
 reasons:
 
 • EBS volumes contend directly for network throughput with standard packets. 
 This means that EBS throughput is likely to fail if you saturate a network 
 link.
 • EBS volumes have unreliable performance. I/O performance can be 
 exceptionally slow, causing the system to back load reads and writes until 
 the entire cluster becomes unresponsive.
 • Adding capacity by increasing the number of EBS volumes per host does not 
 scale. You can easily surpass the ability of the system to keep effective 
 buffer caches and concurrently serve requests for all of the data it is 
 responsible for managing.
 
 Still applies, especially the network contention and latency issues. 
 
 Ben Bromhead
 Instaclustr | www.instaclustr.com | @instaclustr | +61 415 936 359
 
 On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:
 
 While they guarantee IOPS, they don't really make any guarantees about 
 latency. Since EBS goes over the network, there's so many things in the 
 path of getting at your data, I would be concerned with random latency 
 spikes, unless proven otherwise.
 
 Thanks,
 Daniel
 
 
 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:
 In this document it is said :
 
 Provisioned IOPS (SSD) - Volumes of this type are ideal for the most 
 demanding I/O 

EBS SSD - Cassandra ?

2014-06-18 Thread Alain RODRIGUEZ
Hi,

I just saw this :
http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

Since the problem with EBS was the network, there is no chance that this
hardware architecture might be useful alongside Cassandra, right ?

Alain


Re: EBS SSD - Cassandra ?

2014-06-18 Thread Alain RODRIGUEZ
In this document it is said :


   - Provisioned IOPS (SSD) - Volumes of this type are ideal for the most
   demanding I/O intensive, transactional workloads and large relational or
   NoSQL databases. This volume type provides the most consistent performance
   and allows you to provision the exact level of performance you need with
   the most predictable and consistent performance. With this type of volume
   you provision exactly what you need, and pay for what you provision. Once
   again, you can achieve up to 48,000 IOPS by connecting multiple volumes
   together using RAID.



2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

 Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that this
 hardware architecture might be useful alongside Cassandra, right ?

 Alain



Re: EBS SSD - Cassandra ?

2014-06-18 Thread Daniel Chia
While they guarantee IOPS, they don't really make any guarantees about
latency. Since EBS goes over the network, there's so many things in the
path of getting at your data, I would be concerned with random latency
spikes, unless proven otherwise.

Thanks,
Daniel


On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:

 In this document it is said :


- Provisioned IOPS (SSD) - Volumes of this type are ideal for the most
demanding I/O intensive, transactional workloads and large relational or
NoSQL databases. This volume type provides the most consistent performance
and allows you to provision the exact level of performance you need with
the most predictable and consistent performance. With this type of volume
you provision exactly what you need, and pay for what you provision. Once
again, you can achieve up to 48,000 IOPS by connecting multiple volumes
together using RAID.



 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:

 Hi,

 I just saw this :
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/

 Since the problem with EBS was the network, there is no chance that this
 hardware architecture might be useful alongside Cassandra, right ?

 Alain





Re: EBS SSD - Cassandra ?

2014-06-18 Thread Ben Bromhead
http://www.datastax.com/documentation/cassandra/1.2/cassandra/architecture/architecturePlanningEC2_c.html

From the link:

EBS volumes are not recommended for Cassandra data volumes for the following 
reasons:

• EBS volumes contend directly for network throughput with standard 
packets. This means that EBS throughput is likely to fail if you saturate a 
network link.
• EBS volumes have unreliable performance. I/O performance can be 
exceptionally slow, causing the system to back load reads and writes until the 
entire cluster becomes unresponsive.
• Adding capacity by increasing the number of EBS volumes per host does 
not scale. You can easily surpass the ability of the system to keep effective 
buffer caches and concurrently serve requests for all of the data it is 
responsible for managing.

Still applies, especially the network contention and latency issues. 

Ben Bromhead
Instaclustr | www.instaclustr.com | @instaclustr | +61 415 936 359

On 18 Jun 2014, at 7:18 pm, Daniel Chia danc...@coursera.org wrote:

 While they guarantee IOPS, they don't really make any guarantees about 
 latency. Since EBS goes over the network, there's so many things in the path 
 of getting at your data, I would be concerned with random latency spikes, 
 unless proven otherwise.
 
 Thanks,
 Daniel
 
 
 On Wed, Jun 18, 2014 at 1:58 AM, Alain RODRIGUEZ arodr...@gmail.com wrote:
 In this document it is said :
 
 Provisioned IOPS (SSD) - Volumes of this type are ideal for the most 
 demanding I/O intensive, transactional workloads and large relational or 
 NoSQL databases. This volume type provides the most consistent performance 
 and allows you to provision the exact level of performance you need with the 
 most predictable and consistent performance. With this type of volume you 
 provision exactly what you need, and pay for what you provision. Once again, 
 you can achieve up to 48,000 IOPS by connecting multiple volumes together 
 using RAID.
 
 
 2014-06-18 10:57 GMT+02:00 Alain RODRIGUEZ arodr...@gmail.com:
 
 Hi,
 
 I just saw this : 
 http://aws.amazon.com/fr/blogs/aws/new-ssd-backed-elastic-block-storage/
 
 Since the problem with EBS was the network, there is no chance that this 
 hardware architecture might be useful alongside Cassandra, right ?
 
 Alain