We use Apache Cassandra 3.10-13
On Jun 26, 2017 8:41 PM, "Michael Shuler" wrote:
What version of Cassandra?
--
Michael
On 06/26/2017 09:53 PM, Balaji Venkatesan wrote:
> Hi All,
>
> When I run nodetool repair on a keyspace I constantly get "Some repair
> failed"
Good point.
> On Jun 27, 2017, at 1:34 PM, Anuj Wadehra wrote:
>
> Also, if you restore exactly same data with different IP, you may need to
> clear gossip state on the node.
>
> Anuj
>
> Sent from Yahoo Mail on Android
>
Also, if you restore exactly same data with different IP, you may need to clear
gossip state on the node.
Anuj
Sent from Yahoo Mail on Android
On Tue, Jun 27, 2017 at 11:56 PM, Anuj Wadehra wrote:
Hi Nitan,
I asked for adding autobootstrap false to avoid
Then the partition is too big or has too many sstables that contain data
for that partition so that the query times out. You can run a manual
compaction on that table. That helped me several times.
+ I hope you are not trying to read that parition at once. Please use
paging to query large
Query it at consistency ALL and let read repair do its thing.
On Tue, Jun 27, 2017 at 11:48 AM Pranay akula
wrote:
> I have a CF with composite partition key, partition key consists of blob
> and text data types.
> The select query against this particular partition is
I have a CF with composite partition key, partition key consists of blob
and text data types.
The select query against this particular partition is timing out so to
debug it further i ran nodetool getendpoints i am getting error like below
error: Non-hex characters in
Hi Jonathan,
I tried multiple times running the query at consistency all, every time i
run it i getting the same output that co-ordinator is timing out. I can see
it wasn't helping much there were multiple read repair drops. Is there any
other way to get that partition fixed ??
Thanks
Pranay.
Hi Cassandra-Users,
C* 3.0.13 RELEASE HAS A CORNER CASE BUG ON SCHEMA UPDATE, WHICH CORRUPTS
THE DATA. PLEASE DO NOT UPDATE SCHEMA. OUR TEAM IS WORKING ON FIXING THE
ISSUE!
Thanks, Varun
https://lists.apache.org/thread.html/e381a172bd4ebe3b4fbbf6065389114ebf503f646bb9ae2aa1d8c925@%3Cuser.cassandra.apache.org%3E
3.0.14 NEWS.txt from the release announcement: https://goo.gl/Z2QzQm
--
Kind regards,
Michael
On 06/27/2017 02:54 PM, Varun Gupta wrote:
> Hi Cassandra-Users,
>
> C*
Yeah i am running with paging on, and the output rows returned are not
consistent the max rows i got returned is 7 and i am running with tracing
on and i don't see any tombstones as well. Hpw can i determine if my
partition is large or not ??
i was able to figure out the token from one node and
To clarify, are you talking about the same issue fixed in 3.0.14 for
CASSANDRA-13004 or something different?
Could you please, submit a JIRA ticket for whatever you are seeing,
Varun? I do not see one from your JIRA user, but perhaps someone else did?
Thanks
[0]
Hannu,
Thanks for your response. Yes our gc grace period is at default setting, so you
are right, I do not need hints which are a month old. Especially since I have
scheduled repairs running on the cluster almost every day and cross DC repair
running every week. The network connection is slow
Hi,
I am curious to know how people practically use Snapshot restore provided that
snapshot restore may lead to inconsistent reads until full repair is run on the
node being restored ( if you have dropped mutations in your cluster).
Example:
9 am snapshot taken on all 3 nodes10 am mutation
Anuj,
Thank you for your explanation. I had misread the documentation and was under
the impression all these days that Hints_window refers to ‘time until
deletion’. Also, thanks for the information about TCP settings, I will have to
recheck the existing configuration. I will also test the
I think you missed the part where I said I already tried
> On Jun 26, 2017, at 10:48 PM, Jonathan Haddad wrote:
>
> While someone here might know the answer, I think you'll want to ask on the
> DC/OS mailing list for best chance of getting a response. The Apache
>
Hi,
First of all, I don’t know why they get delivered so slowly.
However, if your gc grace seconds is the default 10 days then those hints
from May are not needed and could/should be truncated. If the hint delivery
is causing problems, then one option is that you could just disable it and
rely
Hi Mark,
Please ensure that the node is not defined as seed node in the yaml. Seed nodes
don't bootstrap.
ThanksAnuj
On Tue, Jun 27, 2017 at 9:56 PM, Mark Furlong wrote:
I have a node that has been decommissioned and it showed ‘UL’, the data volume
and the
Hi Meg,
max_hint_window_in_ms =3 hrs means that if a node is down/unresponsive for more
than 3 hrs, hints would not be stored for it any further until it becomes
responsive again. It should not mean that already stored hints would be
truncated after 3 hours.
Regarding connection timeouts
I have a node that has been decommissioned and it showed ‘UL’, the data volume
and the commitlogs have been removed, and I now want to add that node back into
my ring. When I add this node, (bootstrap=true, start cassandra service) it
comes back up in the ring as an existing node and shows as
Hello,
I am facing an issue with Hinted Handoff files in Cassandra v3.0.10. A DC1 node
is storing large number of hints for DC2 nodes (we are facing connection
timeout issues). The problem is that the hint files which are created on DC1
are not getting deleted after the 3 hour window. Hints
Hi Nitan,
I think it would be simpler to take one node down at a time and replace it by
bringing the new node up after linux upgrade, doing same cassandra setup, using
replace_address option and setting autobootstrap=false ( as data is already
there). No downtime as it would be a rolling
Anuj,
We did that in past, even if data was not removed, replace_node caused data
streaming. So changing IP is simplest and safest option.
> On Jun 27, 2017, at 12:43 PM, Anuj Wadehra wrote:
>
> replace_address_first_boot
So,
If you are just moving data to another node, you can just start the new C*
node up without any replace_address parameters. Cassandra will treat it as
existing node with changed IP address (assuming all files within data are
moved).
Hannu
On 27 June 2017 at 20:43:18, Anuj Wadehra
Thats right Hannu. We are planning to go with this approach instead of
replace_address.
> On Jun 27, 2017, at 12:48 PM, Hannu Kröger wrote:
>
> So,
>
> If you are just moving data to another node, you can just start the new C*
> node up without any replace_address
Hi,
In CQL binary protocol v4, is it guaranteed by the protocol that a fetch of
the next page will see writes which were ACKed before that fetch, but after
the whole query was started?
Regards,
Tomek
What version of Cassandra? What state do the other nodes see this node as?
You could force a rebuild of the data with nodetool rebuild
Regards Cameron
Senior Operations & Support Engineer at Instaclustr
26 matches
Mail list logo