[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062678#comment-13062678 ] Stu Hood commented on CASSANDRA-47: --- bq. example ./bin/stress -S 1024 -n 100 -C 250 -V [...] I see 3.8GB compressed into 781MB in my tests. Why would the uncompressed size for -C 250 be different from the uncompressed size for -C 50: 3.8 GB vs the 1.7 GB from before? On a side note: a cardinality of 250 makes for less variance in the average size of the random generated values, but I still see relatively large size differences between consecutive runs. It might be worth opening a separate ticket for stress.java to make the generated random values a fixed size. With the understanding that there is a lot of variance in the results, here are some preliminary results for the _bin/stress -S 1024 -n 100 -C 250 -V_ workload: || build || disk volume (bytes) || write runtime (s) || read runtime (s) || read ops/s || | trunk | 4,015,004,187 | 372 | 1000s | ~216 | | #674 + #2319 | 594,796,624 | 273 | 255 | 3845 | | #47* | | | | | \* need to figure out the problem I was having above SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062709#comment-13062709 ] Pavel Yaskevich commented on CASSANDRA-47: -- bq. I haven't had any luck seeing actual compression with this patch... is there a manual step to enable it? On OSX, the patch slowed the server down to a crawl, but did not result in compression. Performance seems to be reasonable on Linux, but without any effect: running bin/stress -S 1024 -n 100 -C 250 -V resulted in 3.3 GB of data. As I wrote previously - right now on linux you can see compressed data size by running ls with ahs and taking a size in blocks, because current implementations uses file holes (reserves free space for the future chunk changes). I'm currently working on the way to eliminate that need. SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (CASSANDRA-2877) git-cassandra-angosso-angosso.html
git-cassandra-angosso-angosso.html -- Key: CASSANDRA-2877 URL: https://issues.apache.org/jira/browse/CASSANDRA-2877 Project: Cassandra Issue Type: Improvement Components: API Affects Versions: 0.8.0 Environment: https://ango...@github.com/angosso/git-cassandra-angosso-angosso.html.git http://angosso.git.sourceforge.net/git/gitweb-index.cgi configure include/config.h dlls/Makefile.in programs/Makefile.in */Makefile Reporter: Roger Mbiama Priority: Critical Fix For: 0.8.0 Requirements * Java = 1.6 (OpenJDK and Sun have been tested) Getting started --- This short guide will walk you through getting a basic one node cluster up and running, and demonstrate some simple reads and writes. * tar -zxvf apache-cassandra-$VERSION.tar.gz * cd apache-cassandra-$VERSION * sudo mkdir -p /var/log/cassandra * sudo chown -R `angosso` /var/log/cassandra * sudo mkdir -p /var/lib/cassandra * sudo chown -R `angosso` /var/lib/cassandra Note: The sample configuration files in conf/ determine the file-system locations Cassandra uses for logging and data storage. You are free to change these to suit your own environment and adjust the path names used here accordingly. Now that we're ready, let's start it up! * bin/cassandra -f Running the startup script with the -f argument will cause Cassandra to remain in the foreground and log to standard out. Now let's try to read and write some data using the command line client. * bin/cassandra-cli --host http://angosso1.w02.winhost.com --port 9160 The command line client is interactive so if everything worked you should be sitting in front of a prompt... Connected to http://angosso1.w02.winhost.com/9160 Welcome to cassandra CLI. Type 'help' or '?' for help. Type 'quit' or 'exit' to quit. cassandra As the banner says, you can use 'help' or '?' to see what the CLI has to offer, and 'quit' or 'exit' when you've had enough fun. But lets try something slightly more interesting... cassandra set Keyspace1.Standard2['rmbiama']['first'] = 'Roger' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['last'] = 'mbiama' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['age'] = '54' Value inserted. cassandra get Keyspace1.Standard2['rmbiama'] (column=age, value=42; timestamp=1249930062801) (column=first, value=Roger; timestamp=1249930053103) (column=last, value=Mbiama; timestamp=1249930058345) Returned 3 rows. cassandra If your session looks similar to what's above, congrats, your single node cluster is operational! But what exactly was all of that? Let's break it down into pieces and see. set Keyspace1.Standard2['rmbiama']['angosso / git-cassandra-angosso-angosso.html'] = 'Roger' \\ \\ \ \\ \_ key \ \_ value \\ \_ column \_ keyspace \_ column family Data stored in Cassandra is associated with a column family (Standard2), which in turn is associated with a keyspace (Keyspace1). In the example above, we set the value 'Roger' in the 'first' column for key 'rmbiama'. Mirror of Apache Cassandra (incubating); install schematool in debian package A commit object contains a (possibly empty) list of the logical predecessor(s) in the line of development, i.e. GIT-cassandra/angosso/angosso.html its parents.[rogerM] -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (CASSANDRA-2877) git-cassandra-angosso-angosso.html
[ https://issues.apache.org/jira/browse/CASSANDRA-2877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roger Mbiama resolved CASSANDRA-2877. - Resolution: Fixed git-cassandra-angosso-angosso.html -- Key: CASSANDRA-2877 URL: https://issues.apache.org/jira/browse/CASSANDRA-2877 Project: Cassandra Issue Type: Improvement Components: API Affects Versions: 0.8.0 Environment: https://ango...@github.com/angosso/git-cassandra-angosso-angosso.html.git http://angosso.git.sourceforge.net/git/gitweb-index.cgi configure include/config.h dlls/Makefile.in programs/Makefile.in */Makefile Reporter: Roger Mbiama Priority: Critical Labels: features Fix For: 0.8.0 Original Estimate: 504h Remaining Estimate: 504h Requirements * Java = 1.6 (OpenJDK and Sun have been tested) Getting started --- This short guide will walk you through getting a basic one node cluster up and running, and demonstrate some simple reads and writes. * tar -zxvf apache-cassandra-$VERSION.tar.gz * cd apache-cassandra-$VERSION * sudo mkdir -p /var/log/cassandra * sudo chown -R `angosso` /var/log/cassandra * sudo mkdir -p /var/lib/cassandra * sudo chown -R `angosso` /var/lib/cassandra Note: The sample configuration files in conf/ determine the file-system locations Cassandra uses for logging and data storage. You are free to change these to suit your own environment and adjust the path names used here accordingly. Now that we're ready, let's start it up! * bin/cassandra -f Running the startup script with the -f argument will cause Cassandra to remain in the foreground and log to standard out. Now let's try to read and write some data using the command line client. * bin/cassandra-cli --host http://angosso1.w02.winhost.com --port 9160 The command line client is interactive so if everything worked you should be sitting in front of a prompt... Connected to http://angosso1.w02.winhost.com/9160 Welcome to cassandra CLI. Type 'help' or '?' for help. Type 'quit' or 'exit' to quit. cassandra As the banner says, you can use 'help' or '?' to see what the CLI has to offer, and 'quit' or 'exit' when you've had enough fun. But lets try something slightly more interesting... cassandra set Keyspace1.Standard2['rmbiama']['first'] = 'Roger' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['last'] = 'mbiama' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['age'] = '54' Value inserted. cassandra get Keyspace1.Standard2['rmbiama'] (column=age, value=42; timestamp=1249930062801) (column=first, value=Roger; timestamp=1249930053103) (column=last, value=Mbiama; timestamp=1249930058345) Returned 3 rows. cassandra If your session looks similar to what's above, congrats, your single node cluster is operational! But what exactly was all of that? Let's break it down into pieces and see. set Keyspace1.Standard2['rmbiama']['angosso / git-cassandra-angosso-angosso.html'] = 'Roger' \\ \\ \ \\ \_ key \ \_ value \\ \_ column \_ keyspace \_ column family Data stored in Cassandra is associated with a column family (Standard2), which in turn is associated with a keyspace (Keyspace1). In the example above, we set the value 'Roger' in the 'first' column for key 'rmbiama'. Mirror of Apache Cassandra (incubating); install schematool in debian package A commit object contains a (possibly empty) list of the logical predecessor(s) in the line of development, i.e. GIT-cassandra/angosso/angosso.html its parents.[rogerM] -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (CASSANDRA-2877) git-cassandra-angosso-angosso.html
[ https://issues.apache.org/jira/browse/CASSANDRA-2877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roger Mbiama updated CASSANDRA-2877: Description: Requirements * Java = 1.6 (OpenJDK and Sun have been tested) Getting started --- This short guide will walk you through getting a basic one node cluster up and running, and demonstrate some simple reads and writes. * tar -zxvf apache-cassandra-$VERSION.tar.gz * cd apache-cassandra-$VERSION * sudo mkdir -p /var/log/cassandra * sudo chown -R `angosso` /var/log/cassandra * sudo mkdir -p /var/lib/cassandra * sudo chown -R `angosso` /var/lib/cassandra Note: The sample configuration files in conf/ determine the file-system locations Cassandra uses for logging and data storage. You are free to change these to suit your own environment and adjust the path names used here accordingly. Now that we're ready, let's start it up! * bin/cassandra -f Running the startup script with the -f argument will cause Cassandra to remain in the foreground and log to standard out. Now let's try to read and write some data using the command line client. * bin/cassandra-cli --host http://angosso1.w02.winhost.com --port 9160 The command line client is interactive so if everything worked you should be sitting in front of a prompt... Connected to http://angosso1.w02.winhost.com/9160 Welcome to cassandra CLI. Type 'help' or '?' for help. Type 'quit' or 'exit' to quit. cassandra As the banner says, you can use 'help' or '?' to see what the CLI has to offer, and 'quit' or 'exit' when you've had enough fun. But lets try something slightly more interesting... cassandra set Keyspace1.Standard2['rmbiama']['first'] = 'Roger' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['last'] = 'mbiama' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['age'] = '54' Value inserted. cassandra get Keyspace1.Standard2['rmbiama'] (column=age, value=54; timestamp=1249930062801) (column=first, value=Roger; timestamp=1249930053103) (column=last, value=Mbiama; timestamp=1249930058345) Returned 3 rows. cassandra If your session looks similar to what's above, congrats, your single node cluster is operational! But what exactly was all of that? Let's break it down into pieces and see. set Keyspace1.Standard2['rmbiama']['angosso / git-cassandra-angosso-angosso.html'] = 'Roger' \\ \\ \ \\ \_ key \ \_ value \\ \_ column \_ keyspace \_ column family Data stored in Cassandra is associated with a column family (Standard2), which in turn is associated with a keyspace (Keyspace1). In the example above, we set the value 'Roger' in the 'first' column for key 'rmbiama'. Mirror of Apache Cassandra (incubating); install schematool in debian package A commit object contains a (possibly empty) list of the logical predecessor(s) in the line of development, i.e. GIT-cassandra/angosso/angosso.html its parents.[rogerM] was: Requirements * Java = 1.6 (OpenJDK and Sun have been tested) Getting started --- This short guide will walk you through getting a basic one node cluster up and running, and demonstrate some simple reads and writes. * tar -zxvf apache-cassandra-$VERSION.tar.gz * cd apache-cassandra-$VERSION * sudo mkdir -p /var/log/cassandra * sudo chown -R `angosso` /var/log/cassandra * sudo mkdir -p /var/lib/cassandra * sudo chown -R `angosso` /var/lib/cassandra Note: The sample configuration files in conf/ determine the file-system locations Cassandra uses for logging and data storage. You are free to change these to suit your own environment and adjust the path names used here accordingly. Now that we're ready, let's start it up! * bin/cassandra -f Running the startup script with the -f argument will cause Cassandra to remain in the foreground and log to standard out. Now let's try to read and write some data using the command line client. * bin/cassandra-cli --host http://angosso1.w02.winhost.com --port 9160 The command line client is interactive so if everything worked you should be sitting in front of a prompt... Connected to http://angosso1.w02.winhost.com/9160 Welcome to cassandra CLI. Type 'help' or '?' for help. Type 'quit' or 'exit' to quit. cassandra As the banner says, you can use 'help' or '?' to see what the CLI has to offer, and 'quit' or 'exit' when you've had enough fun. But lets try something slightly more interesting... cassandra set Keyspace1.Standard2['rmbiama']['first'] = 'Roger' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['last'] = 'mbiama' Value inserted. cassandra set Keyspace1.Standard2['rmbiama']['age'] = '54' Value inserted. cassandra get Keyspace1.Standard2['rmbiama']
[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062729#comment-13062729 ] Jonathan Ellis commented on CASSANDRA-47: - bq. current implementations uses file holes (reserves free space for the future chunk changes) Is this for where we have to seek back to write the row size on large rows? We're already making two passes when compacting large rows (first to compute the indexes), we could make the first pass compute the serialized size so we don't have to seek after the 2nd pass that writes the data. SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062730#comment-13062730 ] Pavel Yaskevich commented on CASSANDRA-47: -- bq. Why would the uncompressed size for -C 250 be different from the uncompressed size for -C 50: 3.8 GB vs the 1.7 GB from before? I did simply choose the largest files from individual compactions. Also I don't know why you getting only 3.7GB of data after `-S 1024 -n 100 -C 250 -V` because in all of my tests I get about 5.1GB which current trunk code and with patch applied. My results: ||build||disk volume (bytes)||write runtime (s)||read runtime (s)||read ops/s|| ||trunk||5,241,718,144||166||2210||~450|| ||#47||5,090,003,162 (1.2GB of blocks aka real size)||156||480||~2100|| Both sizes are after last major compaction, cassandra with default configuration running on Quad-Core AMD Opteron(tm) Processor 2374 HE with 4229730MHz on each core, 2GB RAM - Debian 5.0 (Lenny) (kernel 2.6.35.4-rscloud) hosted on rackspace. SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062731#comment-13062731 ] Pavel Yaskevich commented on CASSANDRA-47: -- bq. Is this for where we have to seek back to write the row size on large rows? Exactly. bq. We're already making two passes when compacting large rows (first to compute the indexes), we could make the first pass compute the serialized size so we don't have to seek after the 2nd pass that writes the data. I will look into it. SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-47) SSTable compression
[ https://issues.apache.org/jira/browse/CASSANDRA-47?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062739#comment-13062739 ] Jonathan Ellis commented on CASSANDRA-47: - My preference would be to do seekless 2-pass as a separate patch before this one, btw. SSTable compression --- Key: CASSANDRA-47 URL: https://issues.apache.org/jira/browse/CASSANDRA-47 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Pavel Yaskevich Labels: compression Fix For: 1.0 Attachments: CASSANDRA-47.patch, snappy-java-1.0.3-rc4.jar We should be able to do SSTable compression which would trade CPU for I/O (almost always a good trade). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-1125) Filter out ColumnFamily rows that aren't part of the query
[ https://issues.apache.org/jira/browse/CASSANDRA-1125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062812#comment-13062812 ] Mck SembWever commented on CASSANDRA-1125: -- +1 (tested) on 1125-v3.txt Filter out ColumnFamily rows that aren't part of the query -- Key: CASSANDRA-1125 URL: https://issues.apache.org/jira/browse/CASSANDRA-1125 Project: Cassandra Issue Type: New Feature Components: Hadoop Reporter: Jeremy Hanna Assignee: Mck SembWever Priority: Minor Fix For: 1.0 Attachments: 1125-formatted.txt, 1125-v3.txt, CASSANDRA-1125.patch, CASSANDRA-1125.patch Currently, when running a MapReduce job against data in a Cassandra data store, it reads through all the data for a particular ColumnFamily. This could be optimized to only read through those rows that have to do with the query. It's a small change but wanted to put it in Jira so that it didn't fall through the cracks. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (CASSANDRA-1125) Filter out ColumnFamily rows that aren't part of the query (using a KeyRange)
[ https://issues.apache.org/jira/browse/CASSANDRA-1125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mck SembWever updated CASSANDRA-1125: - Summary: Filter out ColumnFamily rows that aren't part of the query (using a KeyRange) (was: Filter out ColumnFamily rows that aren't part of the query) Filter out ColumnFamily rows that aren't part of the query (using a KeyRange) - Key: CASSANDRA-1125 URL: https://issues.apache.org/jira/browse/CASSANDRA-1125 Project: Cassandra Issue Type: New Feature Components: Hadoop Reporter: Jeremy Hanna Assignee: Mck SembWever Priority: Minor Fix For: 1.0 Attachments: 1125-formatted.txt, 1125-v3.txt, CASSANDRA-1125.patch, CASSANDRA-1125.patch Currently, when running a MapReduce job against data in a Cassandra data store, it reads through all the data for a particular ColumnFamily. This could be optimized to only read through those rows that have to do with the query. It's a small change but wanted to put it in Jira so that it didn't fall through the cracks. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (CASSANDRA-2878) Filter out ColumnFamily rows that aren't part of the query (using a IndexClause)
Filter out ColumnFamily rows that aren't part of the query (using a IndexClause) Key: CASSANDRA-2878 URL: https://issues.apache.org/jira/browse/CASSANDRA-2878 Project: Cassandra Issue Type: New Feature Components: Hadoop Reporter: Mck SembWever Assignee: Mck SembWever Priority: Minor Fix For: 1.0 Currently, when running a MapReduce job against data in a Cassandra data store, it reads through all the data for a particular ColumnFamily. This could be optimized to only read through those rows that have to do with the query. It's a small change but wanted to put it in Jira so that it didn't fall through the cracks. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-1125) Filter out ColumnFamily rows that aren't part of the query (using a KeyRange)
[ https://issues.apache.org/jira/browse/CASSANDRA-1125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062814#comment-13062814 ] Mck SembWever commented on CASSANDRA-1125: -- Created CASSANDRA-2878 for the better solution using a IndexClause Filter out ColumnFamily rows that aren't part of the query (using a KeyRange) - Key: CASSANDRA-1125 URL: https://issues.apache.org/jira/browse/CASSANDRA-1125 Project: Cassandra Issue Type: New Feature Components: Hadoop Reporter: Jeremy Hanna Assignee: Mck SembWever Priority: Minor Fix For: 1.0 Attachments: 1125-formatted.txt, 1125-v3.txt, CASSANDRA-1125.patch, CASSANDRA-1125.patch Currently, when running a MapReduce job against data in a Cassandra data store, it reads through all the data for a particular ColumnFamily. This could be optimized to only read through those rows that have to do with the query. It's a small change but wanted to put it in Jira so that it didn't fall through the cracks. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (CASSANDRA-2879) Make SSTableWriter.append(...) methods seekless.
Make SSTableWriter.append(...) methods seekless. Key: CASSANDRA-2879 URL: https://issues.apache.org/jira/browse/CASSANDRA-2879 Project: Cassandra Issue Type: Improvement Components: Core Reporter: Pavel Yaskevich Assignee: Pavel Yaskevich Fix For: 1.0 as we already have a CF.serializedSize() method we don't need to reserve a place to store data size when we write data to SSTable. Compaction should be seekless too because we calculate data size before we write actual content. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (CASSANDRA-2771) Remove commitlog_rotation_threshold_in_mb
[ https://issues.apache.org/jira/browse/CASSANDRA-2771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13062850#comment-13062850 ] Kirk True commented on CASSANDRA-2771: -- There's still a reference in examples/client_only/conf/cassandra.yaml Remove commitlog_rotation_threshold_in_mb -- Key: CASSANDRA-2771 URL: https://issues.apache.org/jira/browse/CASSANDRA-2771 Project: Cassandra Issue Type: Improvement Reporter: Patricio Echague Assignee: Patricio Echague Priority: Minor Labels: commitlog Fix For: 1.0 Attachments: CASSANDRA-2771-2-trunk.txt, CASSANDRA-2771-3-trunk.txt Remove the commitlog segment size config setting, nobody has ever changed it. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira