-
Notifications
You must be signed in to change notification settings - Fork 157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use namespace as subject instead of etcd pod for k8s etcd backup and restore #742
Conversation
42478f9
to
f6b0639
Compare
examples/etcd/etcd-in-cluster/k8s/etcd-incluster-blueprint.yaml
Outdated
Show resolved
Hide resolved
examples/etcd/etcd-in-cluster/k8s/etcd-incluster-blueprint.yaml
Outdated
Show resolved
Hide resolved
@@ -139,7 +139,8 @@ No resources found. | |||
|
|||
To restore the ETCD cluster we can follow the [documentation](https://docs.openshift.com/container-platform/4.5/backup_and_restore/disaster_recovery/scenario-2-restoring-cluster-state.html) that is provided by the OpenShift team. | |||
But we will have to make some modification in the restore script (`cluster-restore.sh`) because default | |||
restore script expects the static pods manifest as well, and in our case we didn't backup the static pod manifests. | |||
restore script expects the static pods manifest as well, and in our case we didn't backup the static pod manifests. We have made respective changes | |||
in the script that is provided by OpenShift, that can be found in this repo. | |||
|
|||
You can follow the steps that are mentioned below along with the documentation that is mentioned above, most of the steps that are mentioned here are either directly taken from the documentation or are modified version of it. Among all the running leader nodes choose one node to be the restore node, make sure you have SSH connectivity to all of the leader nodes including the one that you have chosen to be restore node. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You can follow the steps that are mentioned below along with the documentation that is mentioned above, most of the steps that are mentioned here are either directly taken from the documentation or are modified version of it. Among all the running leader nodes choose one node to be the restore node, make sure you have SSH connectivity to all of the leader nodes including the one that you have chosen to be restore node. | |
Additional steps to be followed are mentioned below. Among all the running leader nodes, choose one node to be the restore node. Ensure there is SSH connectivity to all the leader nodes including the one the chosen restore node. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
so I wanted to clearly specify that the steps that are mentioned below are not our steps (steps that we have figured) but they have just been taken from the above documentation. And I think that is an important distinction to make. WDYT
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @tdmanv
Can you please have a look into above comment. And suggest what should be done.
Change Overview
Per current blueprint we were using etcd pod as subject, which would result in issues because of static pods. This PR changes that to have namespace as subject.
Pull request type
Please check the type of change your PR introduces:
Issues
Test Plan