okay, is there a way to export the TTL using CQLsh or DSBulk?

On Thu, Jul 16, 2020 at 11:20 AM Alex Ott <alex...@gmail.com> wrote:

> if you didn't export TTL explicitly, and didn't load it back, then you'll
> get not expirable data.
>
> On Thu, Jul 16, 2020 at 7:48 PM Jai Bheemsen Rao Dhanwada <
> jaibheem...@gmail.com> wrote:
>
>> In tried verify metadata, In case of writetime it is setting it as insert
>> time but the TTL value is showing as null. Is this expected? Does this mean
>> this record will never expire after the insert?
>> Is there any alternative to preserve the TTL ?
>>
>> In the new Table inserted with Cqlsh and Dsbulk
>> cqlsh > SELECT ttl(secret) from ks_blah.cf_blah ;
>>
>>  ttl(secret)
>> --------------
>>          null
>>          null
>>
>> (2 rows)
>>
>> In the old table where the data was written from application
>>
>> cqlsh > SELECT ttl(secret) from ks_old.cf_old ;
>>
>>  ttl(secret)
>> --------------------
>>          4517461
>>          4525958
>>
>> (2 rows)
>>
>> On Wed, Jul 15, 2020 at 1:17 PM Jai Bheemsen Rao Dhanwada <
>> jaibheem...@gmail.com> wrote:
>>
>>> thank you
>>>
>>> On Wed, Jul 15, 2020 at 1:11 PM Russell Spitzer <
>>> russell.spit...@gmail.com> wrote:
>>>
>>>> Alex is referring to the "writetime" and "tttl" values for each cell.
>>>> Most tools copy via CQL writes and don't by default copy those previous
>>>> writetime and ttl values and instead just give a new writetime value which
>>>> matches the copy time rather than initial insert time.
>>>>
>>>> On Wed, Jul 15, 2020 at 3:01 PM Jai Bheemsen Rao Dhanwada <
>>>> jaibheem...@gmail.com> wrote:
>>>>
>>>>> Hello Alex,
>>>>>
>>>>>
>>>>>    - use DSBulk - it's a very effective tool for unloading & loading
>>>>>    data from/to Cassandra/DSE. Use zstd compression for offloaded data to 
>>>>> save
>>>>>    disk space (see blog links below for more details).  But the 
>>>>> *preserving
>>>>>    metadata* could be a problem.
>>>>>
>>>>> Here what exactly do you mean by "preserving metadata" ? would you
>>>>> mind explaining?
>>>>>
>>>>> On Tue, Jul 14, 2020 at 8:50 AM Jai Bheemsen Rao Dhanwada <
>>>>> jaibheem...@gmail.com> wrote:
>>>>>
>>>>>> Thank you for the suggestions
>>>>>>
>>>>>> On Tue, Jul 14, 2020 at 1:42 AM Alex Ott <alex...@gmail.com> wrote:
>>>>>>
>>>>>>> CQLSH definitely won't work for that amount of data, so you need to
>>>>>>> use other tools.
>>>>>>>
>>>>>>> But before selecting them, you need to define requirements. For
>>>>>>> example:
>>>>>>>
>>>>>>>    1. Are you copying the data into tables with exactly the same
>>>>>>>    structure?
>>>>>>>    2. Do you need to preserve metadata, like, writetime & TTL?
>>>>>>>
>>>>>>> Depending on that, you may have following choices:
>>>>>>>
>>>>>>>    - use sstableloader - it will preserve all metadata, like, ttl
>>>>>>>    and writetime. You just need to copy SSTable files, or stream 
>>>>>>> directly from
>>>>>>>    the source cluster.  But this will require copying of data into 
>>>>>>> tables with
>>>>>>>    exactly same structure (and in case of UDTs, the keyspace names 
>>>>>>> should be
>>>>>>>    the same)
>>>>>>>    - use DSBulk - it's a very effective tool for unloading &
>>>>>>>    loading data from/to Cassandra/DSE. Use zstd compression for 
>>>>>>> offloaded data
>>>>>>>    to save disk space (see blog links below for more details).  But the
>>>>>>>    preserving metadata could be a problem.
>>>>>>>    - use Spark + Spark Cassandra Connector. But also, preserving
>>>>>>>    the metadata is not an easy task, and requires programming to handle 
>>>>>>> all
>>>>>>>    edge cases (see
>>>>>>>    https://datastax-oss.atlassian.net/browse/SPARKC-596 for details)
>>>>>>>
>>>>>>>
>>>>>>> blog series on DSBulk:
>>>>>>>
>>>>>>>    -
>>>>>>>    
>>>>>>> https://www.datastax.com/blog/2019/03/datastax-bulk-loader-introduction-and-loading
>>>>>>>    -
>>>>>>>    
>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-more-loading
>>>>>>>    -
>>>>>>>    
>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-common-settings
>>>>>>>    -
>>>>>>>    https://www.datastax.com/blog/2019/06/datastax-bulk-loader-unloading
>>>>>>>    -
>>>>>>>    https://www.datastax.com/blog/2019/07/datastax-bulk-loader-counting
>>>>>>>    -
>>>>>>>    
>>>>>>> https://www.datastax.com/blog/2019/12/datastax-bulk-loader-examples-loading-other-locations
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Jul 14, 2020 at 1:47 AM Jai Bheemsen Rao Dhanwada <
>>>>>>> jaibheem...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hello,
>>>>>>>>
>>>>>>>> I would like to copy some data from one cassandra cluster to
>>>>>>>> another cassandra cluster using the CQLSH copy command. Is this the 
>>>>>>>> good
>>>>>>>> approach if the dataset size on the source cluster is very high(500G -
>>>>>>>> 1TB)? If not what is the safe approach? and are there any 
>>>>>>>> limitations/known
>>>>>>>> issues to keep in mind before attempting this?
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> With best wishes,                    Alex Ott
>>>>>>> http://alexott.net/
>>>>>>> Twitter: alexott_en (English), alexott (Russian)
>>>>>>>
>>>>>>
>
> --
> With best wishes,                    Alex Ott
> http://alexott.net/
> Twitter: alexott_en (English), alexott (Russian)
>

Reply via email to