look into a series of the blog posts that I sent, I think that it should be
in the 4th post

On Thu, Jul 16, 2020 at 8:27 PM Jai Bheemsen Rao Dhanwada <
jaibheem...@gmail.com> wrote:

> okay, is there a way to export the TTL using CQLsh or DSBulk?
>
> On Thu, Jul 16, 2020 at 11:20 AM Alex Ott <alex...@gmail.com> wrote:
>
>> if you didn't export TTL explicitly, and didn't load it back, then you'll
>> get not expirable data.
>>
>> On Thu, Jul 16, 2020 at 7:48 PM Jai Bheemsen Rao Dhanwada <
>> jaibheem...@gmail.com> wrote:
>>
>>> In tried verify metadata, In case of writetime it is setting it as
>>> insert time but the TTL value is showing as null. Is this expected? Does
>>> this mean this record will never expire after the insert?
>>> Is there any alternative to preserve the TTL ?
>>>
>>> In the new Table inserted with Cqlsh and Dsbulk
>>> cqlsh > SELECT ttl(secret) from ks_blah.cf_blah ;
>>>
>>>  ttl(secret)
>>> --------------
>>>          null
>>>          null
>>>
>>> (2 rows)
>>>
>>> In the old table where the data was written from application
>>>
>>> cqlsh > SELECT ttl(secret) from ks_old.cf_old ;
>>>
>>>  ttl(secret)
>>> --------------------
>>>          4517461
>>>          4525958
>>>
>>> (2 rows)
>>>
>>> On Wed, Jul 15, 2020 at 1:17 PM Jai Bheemsen Rao Dhanwada <
>>> jaibheem...@gmail.com> wrote:
>>>
>>>> thank you
>>>>
>>>> On Wed, Jul 15, 2020 at 1:11 PM Russell Spitzer <
>>>> russell.spit...@gmail.com> wrote:
>>>>
>>>>> Alex is referring to the "writetime" and "tttl" values for each cell.
>>>>> Most tools copy via CQL writes and don't by default copy those previous
>>>>> writetime and ttl values and instead just give a new writetime value which
>>>>> matches the copy time rather than initial insert time.
>>>>>
>>>>> On Wed, Jul 15, 2020 at 3:01 PM Jai Bheemsen Rao Dhanwada <
>>>>> jaibheem...@gmail.com> wrote:
>>>>>
>>>>>> Hello Alex,
>>>>>>
>>>>>>
>>>>>>    - use DSBulk - it's a very effective tool for unloading & loading
>>>>>>    data from/to Cassandra/DSE. Use zstd compression for offloaded data 
>>>>>> to save
>>>>>>    disk space (see blog links below for more details).  But the 
>>>>>> *preserving
>>>>>>    metadata* could be a problem.
>>>>>>
>>>>>> Here what exactly do you mean by "preserving metadata" ? would you
>>>>>> mind explaining?
>>>>>>
>>>>>> On Tue, Jul 14, 2020 at 8:50 AM Jai Bheemsen Rao Dhanwada <
>>>>>> jaibheem...@gmail.com> wrote:
>>>>>>
>>>>>>> Thank you for the suggestions
>>>>>>>
>>>>>>> On Tue, Jul 14, 2020 at 1:42 AM Alex Ott <alex...@gmail.com> wrote:
>>>>>>>
>>>>>>>> CQLSH definitely won't work for that amount of data, so you need to
>>>>>>>> use other tools.
>>>>>>>>
>>>>>>>> But before selecting them, you need to define requirements. For
>>>>>>>> example:
>>>>>>>>
>>>>>>>>    1. Are you copying the data into tables with exactly the same
>>>>>>>>    structure?
>>>>>>>>    2. Do you need to preserve metadata, like, writetime & TTL?
>>>>>>>>
>>>>>>>> Depending on that, you may have following choices:
>>>>>>>>
>>>>>>>>    - use sstableloader - it will preserve all metadata, like, ttl
>>>>>>>>    and writetime. You just need to copy SSTable files, or stream 
>>>>>>>> directly from
>>>>>>>>    the source cluster.  But this will require copying of data into 
>>>>>>>> tables with
>>>>>>>>    exactly same structure (and in case of UDTs, the keyspace names 
>>>>>>>> should be
>>>>>>>>    the same)
>>>>>>>>    - use DSBulk - it's a very effective tool for unloading &
>>>>>>>>    loading data from/to Cassandra/DSE. Use zstd compression for 
>>>>>>>> offloaded data
>>>>>>>>    to save disk space (see blog links below for more details).  But the
>>>>>>>>    preserving metadata could be a problem.
>>>>>>>>    - use Spark + Spark Cassandra Connector. But also, preserving
>>>>>>>>    the metadata is not an easy task, and requires programming to 
>>>>>>>> handle all
>>>>>>>>    edge cases (see
>>>>>>>>    https://datastax-oss.atlassian.net/browse/SPARKC-596 for
>>>>>>>>    details)
>>>>>>>>
>>>>>>>>
>>>>>>>> blog series on DSBulk:
>>>>>>>>
>>>>>>>>    -
>>>>>>>>    
>>>>>>>> https://www.datastax.com/blog/2019/03/datastax-bulk-loader-introduction-and-loading
>>>>>>>>    -
>>>>>>>>    
>>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-more-loading
>>>>>>>>    -
>>>>>>>>    
>>>>>>>> https://www.datastax.com/blog/2019/04/datastax-bulk-loader-common-settings
>>>>>>>>    -
>>>>>>>>    https://www.datastax.com/blog/2019/06/datastax-bulk-loader-unloading
>>>>>>>>    -
>>>>>>>>    https://www.datastax.com/blog/2019/07/datastax-bulk-loader-counting
>>>>>>>>    -
>>>>>>>>    
>>>>>>>> https://www.datastax.com/blog/2019/12/datastax-bulk-loader-examples-loading-other-locations
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Jul 14, 2020 at 1:47 AM Jai Bheemsen Rao Dhanwada <
>>>>>>>> jaibheem...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> I would like to copy some data from one cassandra cluster to
>>>>>>>>> another cassandra cluster using the CQLSH copy command. Is this the 
>>>>>>>>> good
>>>>>>>>> approach if the dataset size on the source cluster is very high(500G -
>>>>>>>>> 1TB)? If not what is the safe approach? and are there any 
>>>>>>>>> limitations/known
>>>>>>>>> issues to keep in mind before attempting this?
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> With best wishes,                    Alex Ott
>>>>>>>> http://alexott.net/
>>>>>>>> Twitter: alexott_en (English), alexott (Russian)
>>>>>>>>
>>>>>>>
>>
>> --
>> With best wishes,                    Alex Ott
>> http://alexott.net/
>> Twitter: alexott_en (English), alexott (Russian)
>>
>

-- 
With best wishes,                    Alex Ott
http://alexott.net/
Twitter: alexott_en (English), alexott (Russian)

Reply via email to