if you didn't export TTL explicitly, and didn't load it back, then you'll get not expirable data.
On Thu, Jul 16, 2020 at 7:48 PM Jai Bheemsen Rao Dhanwada < jaibheem...@gmail.com> wrote: > In tried verify metadata, In case of writetime it is setting it as insert > time but the TTL value is showing as null. Is this expected? Does this mean > this record will never expire after the insert? > Is there any alternative to preserve the TTL ? > > In the new Table inserted with Cqlsh and Dsbulk > cqlsh > SELECT ttl(secret) from ks_blah.cf_blah ; > > ttl(secret) > -------------- > null > null > > (2 rows) > > In the old table where the data was written from application > > cqlsh > SELECT ttl(secret) from ks_old.cf_old ; > > ttl(secret) > -------------------- > 4517461 > 4525958 > > (2 rows) > > On Wed, Jul 15, 2020 at 1:17 PM Jai Bheemsen Rao Dhanwada < > jaibheem...@gmail.com> wrote: > >> thank you >> >> On Wed, Jul 15, 2020 at 1:11 PM Russell Spitzer < >> russell.spit...@gmail.com> wrote: >> >>> Alex is referring to the "writetime" and "tttl" values for each cell. >>> Most tools copy via CQL writes and don't by default copy those previous >>> writetime and ttl values and instead just give a new writetime value which >>> matches the copy time rather than initial insert time. >>> >>> On Wed, Jul 15, 2020 at 3:01 PM Jai Bheemsen Rao Dhanwada < >>> jaibheem...@gmail.com> wrote: >>> >>>> Hello Alex, >>>> >>>> >>>> - use DSBulk - it's a very effective tool for unloading & loading >>>> data from/to Cassandra/DSE. Use zstd compression for offloaded data to >>>> save >>>> disk space (see blog links below for more details). But the *preserving >>>> metadata* could be a problem. >>>> >>>> Here what exactly do you mean by "preserving metadata" ? would you >>>> mind explaining? >>>> >>>> On Tue, Jul 14, 2020 at 8:50 AM Jai Bheemsen Rao Dhanwada < >>>> jaibheem...@gmail.com> wrote: >>>> >>>>> Thank you for the suggestions >>>>> >>>>> On Tue, Jul 14, 2020 at 1:42 AM Alex Ott <alex...@gmail.com> wrote: >>>>> >>>>>> CQLSH definitely won't work for that amount of data, so you need to >>>>>> use other tools. >>>>>> >>>>>> But before selecting them, you need to define requirements. For >>>>>> example: >>>>>> >>>>>> 1. Are you copying the data into tables with exactly the same >>>>>> structure? >>>>>> 2. Do you need to preserve metadata, like, writetime & TTL? >>>>>> >>>>>> Depending on that, you may have following choices: >>>>>> >>>>>> - use sstableloader - it will preserve all metadata, like, ttl >>>>>> and writetime. You just need to copy SSTable files, or stream >>>>>> directly from >>>>>> the source cluster. But this will require copying of data into >>>>>> tables with >>>>>> exactly same structure (and in case of UDTs, the keyspace names >>>>>> should be >>>>>> the same) >>>>>> - use DSBulk - it's a very effective tool for unloading & loading >>>>>> data from/to Cassandra/DSE. Use zstd compression for offloaded data >>>>>> to save >>>>>> disk space (see blog links below for more details). But the >>>>>> preserving >>>>>> metadata could be a problem. >>>>>> - use Spark + Spark Cassandra Connector. But also, preserving the >>>>>> metadata is not an easy task, and requires programming to handle all >>>>>> edge >>>>>> cases (see https://datastax-oss.atlassian.net/browse/SPARKC-596 >>>>>> for details) >>>>>> >>>>>> >>>>>> blog series on DSBulk: >>>>>> >>>>>> - >>>>>> >>>>>> https://www.datastax.com/blog/2019/03/datastax-bulk-loader-introduction-and-loading >>>>>> - >>>>>> >>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-more-loading >>>>>> - >>>>>> >>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-common-settings >>>>>> - >>>>>> https://www.datastax.com/blog/2019/06/datastax-bulk-loader-unloading >>>>>> - >>>>>> https://www.datastax.com/blog/2019/07/datastax-bulk-loader-counting >>>>>> - >>>>>> >>>>>> https://www.datastax.com/blog/2019/12/datastax-bulk-loader-examples-loading-other-locations >>>>>> >>>>>> >>>>>> On Tue, Jul 14, 2020 at 1:47 AM Jai Bheemsen Rao Dhanwada < >>>>>> jaibheem...@gmail.com> wrote: >>>>>> >>>>>>> Hello, >>>>>>> >>>>>>> I would like to copy some data from one cassandra cluster to another >>>>>>> cassandra cluster using the CQLSH copy command. Is this the good >>>>>>> approach >>>>>>> if the dataset size on the source cluster is very high(500G - 1TB)? If >>>>>>> not >>>>>>> what is the safe approach? and are there any limitations/known issues to >>>>>>> keep in mind before attempting this? >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> With best wishes, Alex Ott >>>>>> http://alexott.net/ >>>>>> Twitter: alexott_en (English), alexott (Russian) >>>>>> >>>>> -- With best wishes, Alex Ott http://alexott.net/ Twitter: alexott_en (English), alexott (Russian)