thank you On Thu, Jul 16, 2020 at 12:29 PM Alex Ott <alex...@gmail.com> wrote:
> look into a series of the blog posts that I sent, I think that it should > be in the 4th post > > On Thu, Jul 16, 2020 at 8:27 PM Jai Bheemsen Rao Dhanwada < > jaibheem...@gmail.com> wrote: > >> okay, is there a way to export the TTL using CQLsh or DSBulk? >> >> On Thu, Jul 16, 2020 at 11:20 AM Alex Ott <alex...@gmail.com> wrote: >> >>> if you didn't export TTL explicitly, and didn't load it back, then >>> you'll get not expirable data. >>> >>> On Thu, Jul 16, 2020 at 7:48 PM Jai Bheemsen Rao Dhanwada < >>> jaibheem...@gmail.com> wrote: >>> >>>> In tried verify metadata, In case of writetime it is setting it as >>>> insert time but the TTL value is showing as null. Is this expected? Does >>>> this mean this record will never expire after the insert? >>>> Is there any alternative to preserve the TTL ? >>>> >>>> In the new Table inserted with Cqlsh and Dsbulk >>>> cqlsh > SELECT ttl(secret) from ks_blah.cf_blah ; >>>> >>>> ttl(secret) >>>> -------------- >>>> null >>>> null >>>> >>>> (2 rows) >>>> >>>> In the old table where the data was written from application >>>> >>>> cqlsh > SELECT ttl(secret) from ks_old.cf_old ; >>>> >>>> ttl(secret) >>>> -------------------- >>>> 4517461 >>>> 4525958 >>>> >>>> (2 rows) >>>> >>>> On Wed, Jul 15, 2020 at 1:17 PM Jai Bheemsen Rao Dhanwada < >>>> jaibheem...@gmail.com> wrote: >>>> >>>>> thank you >>>>> >>>>> On Wed, Jul 15, 2020 at 1:11 PM Russell Spitzer < >>>>> russell.spit...@gmail.com> wrote: >>>>> >>>>>> Alex is referring to the "writetime" and "tttl" values for each cell. >>>>>> Most tools copy via CQL writes and don't by default copy those previous >>>>>> writetime and ttl values and instead just give a new writetime value >>>>>> which >>>>>> matches the copy time rather than initial insert time. >>>>>> >>>>>> On Wed, Jul 15, 2020 at 3:01 PM Jai Bheemsen Rao Dhanwada < >>>>>> jaibheem...@gmail.com> wrote: >>>>>> >>>>>>> Hello Alex, >>>>>>> >>>>>>> >>>>>>> - use DSBulk - it's a very effective tool for unloading & >>>>>>> loading data from/to Cassandra/DSE. Use zstd compression for >>>>>>> offloaded data >>>>>>> to save disk space (see blog links below for more details). But the >>>>>>> *preserving >>>>>>> metadata* could be a problem. >>>>>>> >>>>>>> Here what exactly do you mean by "preserving metadata" ? would you >>>>>>> mind explaining? >>>>>>> >>>>>>> On Tue, Jul 14, 2020 at 8:50 AM Jai Bheemsen Rao Dhanwada < >>>>>>> jaibheem...@gmail.com> wrote: >>>>>>> >>>>>>>> Thank you for the suggestions >>>>>>>> >>>>>>>> On Tue, Jul 14, 2020 at 1:42 AM Alex Ott <alex...@gmail.com> wrote: >>>>>>>> >>>>>>>>> CQLSH definitely won't work for that amount of data, so you need >>>>>>>>> to use other tools. >>>>>>>>> >>>>>>>>> But before selecting them, you need to define requirements. For >>>>>>>>> example: >>>>>>>>> >>>>>>>>> 1. Are you copying the data into tables with exactly the same >>>>>>>>> structure? >>>>>>>>> 2. Do you need to preserve metadata, like, writetime & TTL? >>>>>>>>> >>>>>>>>> Depending on that, you may have following choices: >>>>>>>>> >>>>>>>>> - use sstableloader - it will preserve all metadata, like, ttl >>>>>>>>> and writetime. You just need to copy SSTable files, or stream >>>>>>>>> directly from >>>>>>>>> the source cluster. But this will require copying of data into >>>>>>>>> tables with >>>>>>>>> exactly same structure (and in case of UDTs, the keyspace names >>>>>>>>> should be >>>>>>>>> the same) >>>>>>>>> - use DSBulk - it's a very effective tool for unloading & >>>>>>>>> loading data from/to Cassandra/DSE. Use zstd compression for >>>>>>>>> offloaded data >>>>>>>>> to save disk space (see blog links below for more details). But >>>>>>>>> the >>>>>>>>> preserving metadata could be a problem. >>>>>>>>> - use Spark + Spark Cassandra Connector. But also, preserving >>>>>>>>> the metadata is not an easy task, and requires programming to >>>>>>>>> handle all >>>>>>>>> edge cases (see >>>>>>>>> https://datastax-oss.atlassian.net/browse/SPARKC-596 for >>>>>>>>> details) >>>>>>>>> >>>>>>>>> >>>>>>>>> blog series on DSBulk: >>>>>>>>> >>>>>>>>> - >>>>>>>>> >>>>>>>>> https://www.datastax.com/blog/2019/03/datastax-bulk-loader-introduction-and-loading >>>>>>>>> - >>>>>>>>> >>>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-more-loading >>>>>>>>> - >>>>>>>>> >>>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-common-settings >>>>>>>>> - >>>>>>>>> >>>>>>>>> https://www.datastax.com/blog/2019/06/datastax-bulk-loader-unloading >>>>>>>>> - >>>>>>>>> https://www.datastax.com/blog/2019/07/datastax-bulk-loader-counting >>>>>>>>> - >>>>>>>>> >>>>>>>>> https://www.datastax.com/blog/2019/12/datastax-bulk-loader-examples-loading-other-locations >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Jul 14, 2020 at 1:47 AM Jai Bheemsen Rao Dhanwada < >>>>>>>>> jaibheem...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hello, >>>>>>>>>> >>>>>>>>>> I would like to copy some data from one cassandra cluster to >>>>>>>>>> another cassandra cluster using the CQLSH copy command. Is this the >>>>>>>>>> good >>>>>>>>>> approach if the dataset size on the source cluster is very high(500G >>>>>>>>>> - >>>>>>>>>> 1TB)? If not what is the safe approach? and are there any >>>>>>>>>> limitations/known >>>>>>>>>> issues to keep in mind before attempting this? >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> With best wishes, Alex Ott >>>>>>>>> http://alexott.net/ >>>>>>>>> Twitter: alexott_en (English), alexott (Russian) >>>>>>>>> >>>>>>>> >>> >>> -- >>> With best wishes, Alex Ott >>> http://alexott.net/ >>> Twitter: alexott_en (English), alexott (Russian) >>> >> > > -- > With best wishes, Alex Ott > http://alexott.net/ > Twitter: alexott_en (English), alexott (Russian) >