CDC from Cassandra works using Oracle Goldengate for Bigdata, we are doing that
and publishing to kafka. But one of the downstream need batch files with
complete dataset.
I am evaluating some options based on previous responses.
Thanks
Bibin John
From: Peter Corless
Sent: Friday, February 21,
Question: would daily deltas be a good use of CDC? (Rather than export
entire tables.)
(I can understand that this might make analytics hard if you need to span
multiple resultant daily files.)
Perhaps along with CDC, maybe set up the tables for export via a Kafka
topic?
(https://docs.lenses.io/
Batches are for atomicity, not performance.
I would do single deletes with a prepared statement. An IN clause causes extra
work for the coordinator because multiple partitions are being impacted. So,
the coordinator has to coordinate all nodes involved in those writes (up to the
whole cluster).
I would consider building a lookup table instead. Something like:
Create table new_lookup (
new-lookup-partition text,
existing-key text
PRIMARY KEY (new-lookup-partition)
)
For me, these are easier to understand and reason through for Cassandra
performance and availability. I would use
I would also push for something besides a full refresh, if at all possible. It
feels like a waste of resources to me – and not predictably scalable.
Suggestions: use a queue to send writes to both systems. If the downstream
system doesn’t handle TTL, perhaps set an expiration date and a purge qu
On 2/21/20 10:28 AM, Michael Shuler wrote:
So a little extra background:
The ASF INFRA team made a change last week to the required release
location and redirected all requests for the previous required release
URL to:
https://downloads.apache.org/cassandra//
(previous: https://www.ap
So a little extra background:
The ASF INFRA team made a change last week to the required release
location and redirected all requests for the previous required release
URL to:
https://downloads.apache.org/cassandra//
(previous: https://www.apache.org/dist/cassandra// )
This is where projec
Well, that's where they've all been before. All our automation is pulling them
from there. Can you point me to the details of this change you mentioned, so
we can look at what changes we'll need to make and additional thing's we'll
have to proxy behind our firewall.
All the 4.0-alpha releases
Why?
This release adjusted the location of the tar artifacts, so they were
published to the normal recommended dist/release location and not
included in maven, where I understand they are not desired to be there.
Kind regards,
Michael
On 2/21/20 8:18 AM, Chad Helms wrote:
Can we get "apache
Can we get "apache-cassandra:3.11.6:bin.tar.gz" artifact published to maven
central too, please?
On 2/14/20, 5:28 PM, "Michael Shuler" wrote:
The Cassandra team is pleased to announce the release of Apache
Cassandra version 3.11.6.
Apache Cassandra is a fully distributed data
On Fri, Feb 21, 2020 at 2:12 PM Deepak Sharma
wrote:
>
> We have a use case where we need to have two separate PreparedStatement
> objects (one with RetryPolicy and the other without any retry policy) for
> the same query string. And when we try to create two separate
> PreparedStatements, we see
Hi There,
We have a use case where we need to have two separate PreparedStatement
objects (one with RetryPolicy and the other without any retry policy) for
the same query string. And when we try to create two separate
PreparedStatements, we see only one PreparedStatement getting retained (the
prev
12 matches
Mail list logo