On Wed, Jun 11, 2014 at 10:12 AM, Jeremy Jongsma
wrote:
> The big problem seems to have been requesting a large number of row keys
> combined with a large number of named columns in a query. 20K rows with 20K
> columns destroyed my cluster. Splitting it into slices of 100 sequential
On Wed, Jun 11, 2014 at 12:43 PM, Phil Luckhurst <
phil.luckhu...@powerassure.com> wrote:
> It just seems that what we are trying to do here is
> such basic functionality of an index that I thought we must be doing
> something wrong for it to appear to be this broken.
>
To be clear, I did not rea
must be doing
something wrong for it to appear to be this broken.
Phil
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/RPC-timeout-paging-secondary-index-query-results-tp7595078p7595092.html
Sent from the cassandra-u...@incubator.apache.org
I like the "- Provides the illusion that you are using a RDBMS." part ;-)
On Wed, Jun 11, 2014 at 8:52 PM, Robert Coli wrote:
> On Wed, Jun 11, 2014 at 2:24 AM, Phil Luckhurst <
> phil.luckhu...@powerassure.com> wrote:
>
>> Is paging through the results of a
On Wed, Jun 11, 2014 at 2:24 AM, Phil Luckhurst <
phil.luckhu...@powerassure.com> wrote:
> Is paging through the results of a secondary index query broken in
> Cassandra
> 2.0.7 or are we doing something wrong?
>
General feedback on questions of this type :
http://mail-a
like a bug to me - it shouldn't be possible to
> completely lock up a cluster with a valid query that isn't doing a table
> scan, should it?
>
There's lots of valid SQL queries which will "lock up" your server, for
some values of "lock up"?
=Rob
The big problem seems to have been requesting a large number of row keys
combined with a large number of named columns in a query. 20K rows with 20K
columns destroyed my cluster. Splitting it into slices of 100 sequential
queries fixed the performance issue.
When updating 20K rows at a time, I
I'm using Astyanax with a query like this:
clusterContext
.getClient()
.getKeyspace("instruments")
.prepareQuery(INSTRUMENTS_CF)
.setConsistencyLevel(ConsistencyLevel.CL_LOCAL_QUORUM)
.getKeySlice(new String[] {
"ROW1",
"ROW2",
/
Is paging through the results of a secondary index query broken in Cassandra
2.0.7 or are we doing something wrong?
We have table with a few hundred thousand records and an indexed
low-cardinality column. The relevant bits of the table definition are shown
below
CREATE TABLE measurement
Perhaps if you described both the schema and the query in more detail, we
could help... e.g. did the query have an IN clause with 2 keys? Or is
the key compound? More detail will help.
On Tue, Jun 10, 2014 at 7:15 PM, Jeremy Jongsma wrote:
> I didn't explain clearly - I'm n
distributed full
> scan on all the partitions across the cluster, it's normal that the nodes
> are somehow stressed.
>
> How did you make the query? Are you using Thrift or CQL3 API?
>
> Please note that there is another way to get all partition keys : SELECT
> DISTIN
Hello Jeremy
Basically what you are doing is to ask Cassandra to do a distributed full
scan on all the partitions across the cluster, it's normal that the nodes
are somehow stressed.
How did you make the query? Are you using Thrift or CQL3 API?
Please note that there is another way t
I ran an application today that attempted to fetch 20,000+ unique row keys
in one query against a set of completely empty column families. On a 4-node
cluster (EC2 m1.large instances) with the recommended memory settings (2 GB
heap), every single node immediately ran out of memory and became
tables, then drop old tables, ala
>"audit_table_20140610", "audit_table_20140609", etc..
>But then I run into the issue of having to query every table--I would
>have to execute queries against every day to get the data, and then merge
>the data myself. Unless,
be
implemented easily. I've seen the following:
1. Use date-based tables, then drop old tables, ala
"audit_table_20140610", "audit_table_20140609", etc..
But then I run into the issue of having to query every table--I would
have to execute queries against e
e, pid)
> );
Data is queried on a specific object ID and region. Optionally, users can
restrict their query to a specific date range, which the above data model
provides.
However, we generate quite a bit of data, and we want a convenient way to
get rid of the oldest data. Since our system
t not going to
> work.
>
> So here's a question--for these date-based tables (i.e., a table per
> day/week/month/whatever), how are they queried? If I keep 60 days worth of
> auditing data, for example, I'd need to query all 60 tables--can I do that
> smoothly? Or do I hav
If I keep 60 days worth of
auditing data, for example, I'd need to query all 60 tables--can I do that
smoothly? Or do I have to have 60 different select statements? Is there a
way for me to run the same query against all the tables?
On Mon, Jun 9, 2014 at 3:42 PM, Redmumba wrote:
>
PM, Michal Michalski <
michal.michal...@boxever.com> wrote:
> Secondary indexes internally are just CFs that map the indexed value to a
> row key which that value belongs to, so you can only query these indexes
> using "=", not ">", ">=" etc.
>
> Howe
Secondary indexes internally are just CFs that map the indexed value to a
row key which that value belongs to, so you can only query these indexes
using "=", not ">", ">=" etc.
However, your query does not require index *IF* you provide a row key - you
can use
Hello,
You are receiving this item because you are not passing in the Partition
Key as part of your query. Cassandra is telling you it doesn't know which
node to find the data and you haven't explicitly told it to search across
all your nodes for the data. The ALLOW FILTERING claus
I have a table with a timestamp column on it; however, when I try to query
based on it, it fails saying that I must use ALLOW FILTERING--which to me,
means its not using the secondary index. Table definition is (snipping out
irrelevant parts)...
CREATE TABLE audit (
> id bigint,
>
tml
3. PrestoDB http://prestodb.io/
Thanks
Bobby
On May 30, 2014, at 12:09 PM, "cbert...@libero.it" wrote:
Hello,
I have a working cluster of Cassandra that performs very well on a high
traffic web application.
Now I need to build a backend web application to query Cassandra on ma
Hello,
I have a working cluster of Cassandra that performs very well on a high
traffic web application.
Now I need to build a backend web application to query Cassandra on many non
indexed columns ... what is the best way to do that? Apache hive? Pig?
Cassandra 2
Thanks
9:36 AM
> To: user@cassandra.apache.org
> Subject: Possible to Add multiple columns in one query ?
>
> I’m sure this is a CQL 101 question, but.
>
> Is it possible to add MULTIPLE Rows/Columns to a single Partition in a
> single CQL 3 Query / Call.
>
>
, May 25, 2014 9:36 AM
> To: user@cassandra.apache.org
> Subject: Possible to Add multiple columns in one query ?
>
> I’m sure this is a CQL 101 question, but.
>
> Is it possible to add MULTIPLE Rows/Columns to a single Partition in a
> single CQL 3 Query / Call.
>
nan
Sent: Sunday, May 25, 2014 9:36 AM
To: user@cassandra.apache.org
Subject: Possible to Add multiple columns in one query ?
I’m sure this is a CQL 101 question, but.
Is it possible to add MULTIPLE Rows/Columns to a single Partition in a
single CQL 3 Query / Call.
Need:
I’m try
I'm sure this is a CQL 101 question, but.
Is it possible to add MULTIPLE Rows/Columns to a single Partition in a
single CQL 3 Query / Call.
Need:
I'm trying to find the most efficient way to add multiple time series events
to a table in a single call.
Whilst most t
I think there are several issues in your schema and queries.
First, the schema can't efficiently return the single newest post for every
author. It can efficiently return the newest N posts for a particular
author.
On Fri, May 16, 2014 at 11:53 PM, 後藤 泰陽 wrote:
>
> But I consider LIMIT to be a
Calling execute the second time runs the query a second time, and it looks like
the query mutates instance state during the pagination.
What happens if you only call execute() once ?
Cheers
Aaron
-
Aaron Morton
New Zealand
@aaronmorton
Co-Founder & Principal Consul
;
>>
>> 2014/05/16 23:54、Jonathan Lacefield のメール:
>>
>> Hello,
>>
>> Have you looked at using the CLUSTERING ORDER BY and LIMIT features of
>> CQL3?
>>
>> These may help you achieve your goals.
>>
>>
>> http://www.data
LIMIT features of
> CQL3?
>
> These may help you achieve your goals.
>
>
> http://www.datastax.com/documentation/cql/3.1/cql/cql_reference/refClstrOrdr.html
>
> http://www.datastax.com/documentation/cql/3.1/cql/cql_reference/select_r.html
>
> Jonathan Lacefield
> So
documentation/cql/3.1/cql/cql_reference/select_r.html
>
> Jonathan Lacefield
> Solutions Architect, DataStax
> (404) 822 3487
>
>
>
>
>
>
> On Fri, May 16, 2014 at 12:23 AM, Matope Ono wrote:
> Hi, I'm modeling some queries in CQL3.
>
> I
Hi, I'm modeling some queries in CQL3.
I'd like to query first 1 columns for each partitioning keys in CQL3.
For example:
create table posts(
> author ascii,
> created_at timeuuid,
> entry text,
> primary key(author,created_at)
> );
> insert into posts(author,creat
Jonathan Lacefield
Solutions Architect, DataStax
(404) 822 3487
<http://www.linkedin.com/in/jlacefield>
<http://www.datastax.com/cassandrasummit14>
On Fri, May 16, 2014 at 12:23 AM, Matope Ono wrote:
> Hi, I'm modeling some queries in CQL3.
>
> I'd like to query first
Hi, All,
I use the astyanax 1.56.48 + Cassandra 2.0.6 in my test codes and do some query
like this:
query = keyspace.prepareQuery(..).getKey(...)
.autoPaginate(true)
.withColumnRange(new RangeBuilder().setLimit(pageSize).build());
ColumnList result;
result= query.execute().getResult();
while
Hi all,
I'm using Cassandra 2.0.6 and I have 8 nodes. I'm doing some tests by using
operations below:
disable gossip of node A;
check the status by nodetool in other node, node A is Down now;
use cqlsh connecting an "Up" node and create a table;
enable gossip of node A;
check the status, all node
This just happened, is this fixed in 2.0.7?
cqlsh:tap> select * from setting;
Bad Request: unconfigured columnfamily settings
cqlsh:tap> select * from settings;
name | value
--+--
ld
Consider filing a jira. Cql is the standard interface to cassandra
everything is heavily tested.
On Thursday, March 13, 2014, Katsutoshi Nagaoka
wrote:
> Hi.
>
> I am using Cassandra 2.0.6 version. There is a case that select query
returns wrong value if use DESC option. My test proced
Hi.
I am using Cassandra 2.0.6 version. There is a case that select query
returns wrong value if use DESC option. My test procedure is as follows:
--
cqlsh:test> CREATE TABLE mytable (key int, range int, PRIMARY KEY (key,
range));
cqlsh:test> INSERT INTO mytable (key,
Hi, experts,
I need to query all columns of a row in a column family that meet some
conditions (see below). The column is composite column and has following
format:
... where componentN has String type.
What I want to do is to find out all columns that meet following conditions:
1
at 12:14 PM, Senthil, Athinanthny X. -ND <
athinanthny.x.senthil@disney.com> wrote:
> Anyone can suggest how to query on blob column via CQL3. I get bad
> request error saying cannot parse data. I want to lookup on key column
> which is defined as blob.
>
> But I am able to loo
Did you try http://cassandra.apache.org/doc/cql3/CQL.html#blobFun ?
On 2/28/14, 9:14, Senthil, Athinanthny X. -ND wrote:
Anyone can suggest how to query on blob column via CQL3. I get bad
request error saying cannot parse data. I want to lookup on key column
which is defined as blob.
But I
Anyone can suggest how to query on blob column via CQL3. I get bad request
error saying cannot parse data. I want to lookup on key column which is defined
as blob.
But I am able to lookup data via opscenter data explorer. Is there a
conversion functions I need to use?
Sent from my Galaxy
Hi Steve,
It looks like it will be pretty easy for us to do some testing with the
new client version. I'm going to give it a shot and keep my fingers
crossed.
Thanks again,
Chap
On 5 Feb 2014, at 18:10, Steven A Robenalt wrote:
Hi Chap,
If you have the ability to test the 2.0.0rc2 driver,
Hi Chap,
If you have the ability to test the 2.0.0rc2 driver, I would recommend
doing so, even from a dedicated test client or a JUnit test case. There are
other benefits to the change, such as being able to use BatchStatements,
aside from possible impact on your read timeouts.
Steve
On Wed, F
Hi Steve,
Thanks for the reply. After all that information in my initial message I
would forget one of the most important bits. We're running Cassandra
2.0.3 with the 1.0.4 version of the DataStax driver. I'd seen mention
of those timeouts under earlier 2.x versions and really hoped they were
raging 30-40 writes/sec and about 8 reads/sec according to OpsCenter.
> The failures don't seem to be related to any changes in load. A single
> query repeated from CQLSH (about once a second or so) will fail
> approximately one out of ten times. I do see an increase in the average
> read
changes in
load. A single query repeated from CQLSH (about once a second or so)
will fail approximately one out of ten times. I do see an increase in
the average read latency around the time of the failure, though it's
unclear if that's from the single failed request or if others are
I'm guessing its just a coincident.. As far as I know, seeds have nothing
to do with where the data should be located.
I think there could be couple of reasons why you wouldn't see SSTables on a
specific column family folder, these are some of them:
- You're using a few distinct keys which non of t
Hi ,
I have a 4 node cassandra cluster with one node marked as seed node. When i
checked the data directory of seed node , it has two folders
/keyspace/columnfamily.
But sstable db files are not available.the folder is empty.The db files are
available in remaining nodes.
I want to know th
b-2010
(1 rows)
*SELECT * FROM testcollectionindex WHERE periodid='Period2' AND
unit='Number' AND datatagsset CONTAINS 'Store1';*
THIS QUERY DO NOT WORK..I get RPC timeout error and server logs showing
indexoutofbound exception (http://pastebin.com/f
anks what version of Cassandra?
>
> -Roger
>
>
>
>
>
> *From:* Vladimir Prudnikov [mailto:v.prudni...@gmail.com]
> *Sent:* Monday, January 13, 2014 11:57 AM
> *To:* user
> *Subject:* Problem inserting set when query contains IF NOT EXISTS.
>
>
>
> Hi all
Validimir,
Thanks what version of Cassandra?
-Roger
From: Vladimir Prudnikov [mailto:v.prudni...@gmail.com]
Sent: Monday, January 13, 2014 11:57 AM
To: user
Subject: Problem inserting set when query contains IF NOT EXISTS.
Hi all,
I've spend a lot of time finding a bug in system, but it
Hi all,
I've spend a lot of time finding a bug in system, but it turns out that the
problem is in Cassandra.
Here is how to reproduce.
=
CREATE KEYSPACE IF NOT EXISTS test_set WITH REPLICATION = { 'class' :
'SimpleStrategy', 'replication_factor' : 1 };
USE test_set;
CREATE TABLE IF N
You CAN only supply some of the components for a slice.
On Fri, Dec 20, 2013 at 2:13 PM, Josh Dzielak wrote:
> Is there a way to include *multiple* column names in a slice query where
> one only component of the composite column name key needs to match?
>
> For example, if this
Not directly, no.
>
> > 2) If not, is there clever data modeling or indexing that could accomplish
> > this use case? 1 single-row round-trip to get these columns?
> >
>
>
> If this is a query done frequently you could prefix both columns with a
> static
>
>
> My questions –
>
> 1) Is this supported in the Thrift interface or CQL?
>
Not directly, no.
> 2) If not, is there clever data modeling or indexing that could accomplish
> this use case? 1 single-row round-trip to get these columns?
>
If this is a query done
Is there a way to include *multiple* column names in a slice query where one
only component of the composite column name key needs to match?
For example, if this was a single row -
username:0 | username:1 | city:0 | city:1 | other:0|
other:1
that this akin to a "range-of-range" query via composite columns. Is
something like this possible in cassandra, may be in latest versions?
--
Ravi
x27;1' and bucket_id in ('global_props', 'test_bucket')
>
> But that gives the error in the subject.
>
> There's pretty interesting thing is that if I query for text column then the
> query works, while does not work for the map column. Check the two
27; and bucket_id in ('global_props', 'test_bucket')
But that gives the error in the subject.
There's pretty interesting thing is that if I query for text column then
the query works, while does not work for the map column. Check the two
queries at the bottom http://pastie.org/private
Is there any way i can validate the count of data loaded.
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Cassandra-Data-Query-tp7591180.html
Sent from the cassandra-u...@incubator.apache.org mailing list archive at
Nabble.com.
Thanks Rob that helps !
On Fri, Oct 25, 2013 at 7:34 PM, Robert Coli wrote:
> On Fri, Oct 25, 2013 at 2:47 PM, srmore wrote:
>
>> I don't know whether this is possible but was just curious, can you query
>> for the data in the remote datacenter with a CL.ONE ?
>>
On Fri, Oct 25, 2013 at 2:47 PM, srmore wrote:
> I don't know whether this is possible but was just curious, can you query
> for the data in the remote datacenter with a CL.ONE ?
>
A coordinator at CL.ONE picks which replica(s) to query based in large part
on the dynamic snitch.
I don't know whether this is possible but was just curious, can you query
for the data in the remote datacenter with a CL.ONE ?
There could be a case where one might not have a QUORUM and would like to
read the most recent data which includes the data from the other
datacenter. AFAIK to rel
Hi all, when using the java-driver I see this error on the client, for
reads (as well as for writes).
Many of the ops succeed, however I do see a significant amount of errors.
com.datastax.driver.core.exceptions.WriteTimeoutException: Cassandra
timeout during write query at consistency ONE (1
> Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer
> private though?
>
>
Yes but I thought you had it in your examples, was included for completeness.
use the official overloads.
Cheers
-
Aaron Morton
New Zealand
@aaronmorton
Co-Founder & Principa
Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer
private though? I know what to do now in order to construct a column with a
TTL now. Thanks.
On Sep 26, 2013 9:00 PM, "Aaron Morton" wrote:
> > org.apache.cassandra.thrift.Column column; // initialize this with name,
> va
> org.apache.cassandra.thrift.Column column; // initialize this with name,
> value, timestamp, TTL
This is the wrong object to use.
one overload of addColumn() accepts IColumn which is from
org.apache.cassanda.db . The thrift classes are only use for the thrift API.
> What is the difference b
Can someone answer this doubt reg. SSTableSimpleWriter ? I'd asked about
this earlier but it probably missed. Apologies for repeating the question
(with minor additions) :
"""
Let's say I've initialized a *SSTableSimpleWriter* instance and a new
column with TTL set :
*org.apache.cassandra.io.sst
Let's say I've initialized a *SSTableSimpleWriter* instance and a new
column with TTL set :
*SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here
*/);*
*Column column;*
What is the difference between calling *writer.addColumn()* on the column's
name and value, and *writer.addE
I would like to use select count query.
Although it was work at Cassandra 1.2.9, but there is a situation which
does not work at Cassandra 2.0.0.
so, If some row is deleted, 'select count query' seems to return the wrong
value.
Did anything change by Cassandra 2.0.0 ? or Have I made
ilto:kohlisank...@gmail.com]
> *Sent:* Monday, September 16, 2013 1:10 PM
>
> *To:* user@cassandra.apache.org
> *Subject:* Re: Read query slows down when a node goes down
>
> ** **
>
> For how long does the read latencies go up once a machine is down? It
> takes a c
isank...@gmail.com]
> *Sent:* Sunday, September 15, 2013 4:52 PM
> *To:* user@cassandra.apache.org
> *Subject:* Re: Read query slows down when a node goes down
>
> ** **
>
> What is your replication factor? DO you have multi-DC deployment? Also are
> u using v nodes
red I was doing something wrong).
From: sankalp kohli [mailto:kohlisank...@gmail.com]
Sent: Monday, September 16, 2013 1:10 PM
To: user@cassandra.apache.org
Subject: Re: Read query slows down when a node goes down
For how long does the read latencies go up once a machine is down? It takes a
configurable
From: sankalp kohli [mailto:kohlisank...@gmail.com]
Sent: Sunday, September 15, 2013 4:52 PM
To: user@cassandra.apache.org
Subject: Re: Read query slows down when a node goes down
What is your replication factor? DO you have multi-DC deployment? Also are u
using v nodes?
On Sun, Sep 15, 2013 at
What is your replication factor? DO you have multi-DC deployment? Also are
u using v nodes?
On Sun, Sep 15, 2013 at 7:54 AM, Parag Patel wrote:
> Hi,
>
> ** **
>
> We have a six node cluster running DataStax Community Edition 1.2.9. From
> our app, we use the Netflix Astyanax library to re
Hi,
We have a six node cluster running DataStax Community Edition 1.2.9. From our
app, we use the Netflix Astyanax library to read and write records into our
cluster. We read and write with QUARUM. We're experiencing an issue where
when a node goes down, we see our read queries slowing down
SlicePredicate only support “N” columns. So, you need to query one facet at a
time OR you can query m columns such that it returns n revisions. You may need
intelligence to increase or decrease m columns heuristically.
From: ravi prasad
Sent: Tuesday, July 30, 2013 8:11 PM
To
Hi,
I have a data modelling question. I'm modelling for an use case where, an
object can have multiple facets and each facet can have multiple revisions and
the query pattern looks like "get latest 'n' revisions for all facets for an
object (n=1,2,3)". With a t
> Too bad Rainbird isn't open sourced yet!
It's been 2 years, I would not hold your breath.
Remembered there are two time series open source projects out there
https://github.com/deanhiller/databus
https://github.com/Pardot/Rhombus
Cheers
-
Aaron Morton
Cassandra Consultant
New
Thanks Aaron.
Too bad Rainbird isn't open sourced yet!
On Tue, Jul 23, 2013 at 4:48 AM, aaron morton wrote:
> For background on rollup analytics:
>
> Twitter Rainbird
> http://www.slideshare.net/kevinweil/rainbird-realtime-analytics-at-twitter-strata-2011
> Acunu http://www.acunu.com/
>
> Cheer
For background on rollup analytics:
Twitter Rainbird
http://www.slideshare.net/kevinweil/rainbird-realtime-analytics-at-twitter-strata-2011
Acunu http://www.acunu.com/
Cheers
-
Aaron Morton
Cassandra Consultant
New Zealand
@aaronmorton
http://www.thelastpickle.com
On 22/07/20
for 1000 columns.
> If you consider it depends also on the number of nodes in the cluster, the
> memory available and the number of rows and column the query needs, the
> problem of how optimally divide a request becomes quite complex.
It sounds like you are targeting single read threa
This can be done easily,
Use normal column family to store the sequence of events where key is
session #ID identifying one use interaction with a website, column names
are TimeUUID values and column value id of the event (do not write
something like "user added product to shopping cart", something
Would cassandra be a good choice for creating a funnel analytics type
product similar to mixpanel?
e.g. You create a set of events and store them in cassandra for things
like:
event#1 user visited product page
event#2 user added product to shopping cart
event#3 user clicked on checkout page
even
higher
overhead.
If you consider it depends also on the number of nodes in the cluster, the
memory available and the number of rows and column the query needs, the
problem of how optimally divide a request becomes quite complex.
Does these numbers make sense for you?
Cheers
2013/7/17 aaron
and
@aaronmorton
http://www.thelastpickle.com
On 17/07/2013, at 8:24 PM, cesare cugnasco wrote:
> Hi Rob,
> of course, we could issue multiple requests, but then we should consider
> which is the optimal way to split the query in smaller ones. Moreover, we
> should choose how many
> This would seem to conflict with the advice to only use secondary indexes on
> fields with low cardinality, not high cardinality. I guess low cardinality is
> good, as long as it isn't /too/ low?
My concern is seeing people in the wild create secondary indexes with low
cardinality that genera
Hi Rob,
of course, we could issue multiple requests, but then we should consider
which is the optimal way to split the query in smaller ones. Moreover, we
should choose how many of sub-query run in parallel.
In ours tests, we found there's a significant performance difference
between va
On Tue, Jul 16, 2013 at 4:46 AM, cesare cugnasco
wrote:
> We are working on porting some life science applications to Cassandra,
> but we have to deal with its limits managing huge queries. Our queries are
> usually multiget_slice ones: many rows with many columns each.
>
You are not getting muc
when
increasing the amount of rows and columns required in a single query .
Where does this limit come from?
Giving a fast look to the code seems like the entry point is stressed
because it has to keep all the responses in memory. Only after it has
received all the responses, from the nodes, Then
Couple of questions about the test setup:
- are you running the tests in parallel (via threadCount in surefire
or failsafe for example?)
- is the instance of cassandra per-class for per jvm? (or is fork=true?)
On Sun, Jul 14, 2013 at 5:52 PM, Tristan Seligmann
wrote:
> On Mon, Jul 15, 2013 at 12
On Mon, Jul 15, 2013 at 12:26 AM, aaron morton wrote:
> Aaron Morton can confirm but I think one problem could be that to create
> an index on a field with small number of possible values is not good.
>
> Yes.
> In cassandra each value in the index becomes a single row in the internal
> secondary
> Aaron Morton can confirm but I think one problem could be that to create an
> index on a field with small number of possible values is not good.
Yes.
In cassandra each value in the index becomes a single row in the internal
secondary index CF. You will end up with a huge row for all the values
Aaron Morton can confirm but I think one problem could be that to create an
index on a field with small number of possible values is not good.
Regards,
Shahab
On Sat, Jul 13, 2013 at 9:14 AM, Tristan Seligmann
wrote:
> On Fri, Jul 12, 2013 at 10:38 AM, aaron morton wrote:
>
>> CREATE INDEX ON c
On Fri, Jul 12, 2013 at 10:38 AM, aaron morton wrote:
> CREATE INDEX ON conv_msgdata_by_participant_cql(msgReadFlag);
>
> On general this is a bad idea in Cassandra (also in a relational DB IMHO).
> You will get poor performance from it.
>
Could you elaborate on why this is a bad idea?
--
mi
t;Pruner, Anne (Anne)" wrote:
> Hi,
> I’ve been tearing my hair out trying to figure out why this
> query fails. In fact, it only fails on machines with slower CPUs and after
> having previously run some other junit tests. I’m running junits to an
> embedded
Hi,
I've been tearing my hair out trying to figure out why this
query fails. In fact, it only fails on machines with slower CPUs and after
having previously run some other junit tests. I'm running junits to an
embedded Cassandra server, which works well in prett
701 - 800 of 1183 matches
Mail list logo