-
Notifications
You must be signed in to change notification settings - Fork 268
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
S3 Snapshot Retention policy is too aggressive #5216
Comments
That is correct - the S3 retention count applies to ALL snapshots stored on S3, not to the number of snapshots per node. On earlier releases of RKE2, snapshots were pruned from S3 only by the node that uploaded them, but this lead to snapshots never being pruned when their source node was deleted or otherwise removed from the cluster. For clusters where nodes are upgraded or patched by replacement, this lead to S3 snapshot counts growing without bound due to the orphaned snapshots never being cleaned up. One option would be to provide separate settings and commands for pruning snapshots from nodes that are no longer cluster members, but this would probably have to go through the RFE process. |
OK, so it sounds like default behavior was changed and this is affecting our users as it was not clearly communicated through to them. Perhaps we need to look at our process to see how to avoid this in the future. cc: @snasovich @cwayne18 |
It was changed based on complaints from users. Ref: |
But right now the behaviour is not consistent with what the UI says and also what you'd expect. We've worked around by increasing the retention count from 5 to 15. This will lead to 15 snapshots per master node but at least we have 5 usable snapshots in S3 instead of 2... |
That would be the rancher UI, not the RKE2 UI. It sounds like the rancher UI text needs to be updated.
That is not correct. RKE2 stores If you are looking at just s3, then yes you would probably want to align your retention count so that you keep the desired number of snapshots on s3 - if you have 3 nodes and want 2 cycles worth of snapshots, then you should set retention to 6, for example. We will discuss how to enhance this to provide more granular control over snapshot retention, but it will probably require splitting the settings to allow separate counts for local and s3. |
splitting the settings is probably the most sensible approach. |
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 45 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
Unstale |
I'm not sure this is exactly related, but even I set the option |
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 45 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
This is still relevant. S3 backups are messed up and don't work as expected |
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 45 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
still very much an issue - workaround is to enable versioning and/or soft-delete on your s3 storage. Not sure if rke2 registers recovered backups automatically though. |
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 45 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
Unstale |
This repository uses a bot to automatically label issues which have not had any activity (commit/comment/label) for 45 days. This helps us manage the community issues better. If the issue is still relevant, please add a comment to the issue so the bot can remove the label and we know it is still valid. If it is no longer relevant (or possibly fixed in the latest release), the bot will automatically close the issue in 14 days. Thank you for your contributions. |
Still important. |
Environmental Info:
RKE2 Version:
v1.25.15+rke2r2
,v1.27.7+rke2r2
Node(s) CPU architecture, OS, and Version:
N/A
Cluster Configuration:
Multi-master
Describe the bug:
According to the
r/r
issue, RKE2 is too aggressive at pruning S3 snapshots.Steps To Reproduce:
See Rancher issue.
Additional context / logs:
rancher/rancher#43872
The text was updated successfully, but these errors were encountered: