@ChrisK , it’s not advised to use Extended retention on WORM cloud backups. Essentially, we cannot do any micro pruning with cloud worm. Instead, we need to seal the dedupe store and macro prune at regular intervals as the only method of pruning data (and mixed/Extended Retention complicate that to say the least).
With that said, I would set a single overall retention to meet your business requirements.
@Mike Struening Thank you for the feedback. I see the issue with using extended retention on WORM, I’m not a fan of it either..
Do you know if Commvault shows the Usage on the S3 Cloud Library based on the amount of data of all active jobs or the actual used space on the S3 Bucket itselt? Because from my understanding once a job is aged / has reached the retention, it gets flagged but can’t be deleted/pruned on the bucket itself yet as WORM Lock is still active for x days. Meaning there is potentially more data on the Bucket than Commvault “knows about”. So from my understanding it would make sense that Commvault only counts the data amount that’s linked to actual active jobs.
Alternatively - and I’m just thinking out loud here - what would the behaviour be like if we used non-deduplicated copies to enable WORM on? Would the WORM lock still be twice the retention of the storage policy copy?
From the documentation (Link) it’s not clear how non-deduplicated cloud libraries are behaving when WORM is enabled there, it only mentions for ones that have dedup enabled.
@ChrisK, COMM vault knows how much data is physically present, as we keep track of whether or not any physical pruning has occurred or not (whether we're talking about WORM or not). We can report on the actual size, not just based on the active jobs.
The 2x storage policy retention rule doesn't apply for non-dedupe, as there is no DDB that needs to be sealed. Aassuming you are using a cloud library that supports object-based retention lock (as opposed to only bucket / container level retention lock), we can prune the volumes as soon as that retention time is met. There is very little if any actual storage consumption foot print when using storage level worm without any dedupe.
The 2x storage policy retention logic comes into play due to our requirement to macro prune all data associated with a DDB. in a dedupe scenario, the following needs to take place:
-
all jobs need to logically age (storage policy retention)
-
the last job(s) associated with that DDB need to satisfy the object level lock on storage
With non-dedupe, the data associated with each job can be pruned as soon as the object lock retention requirement is met.
Let me know if that helps!
perfect, thank you @Mike Struening !