-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue 579: Fixing leader election failure after node reboot #581
Conversation
Signed-off-by: anishakj <[email protected]>
Signed-off-by: anishakj <[email protected]>
Codecov Report
@@ Coverage Diff @@
## master #581 +/- ##
==========================================
- Coverage 74.68% 74.17% -0.51%
==========================================
Files 15 16 +1
Lines 4164 4229 +65
==========================================
+ Hits 3110 3137 +27
- Misses 932 962 +30
- Partials 122 130 +8
Continue to review full report at Codecov.
|
Signed-off-by: anishakj <[email protected]>
Signed-off-by: anishakj <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Signed-off-by: anishakj <[email protected]>
Change log description
In VMware cluster, some pods are stuck in ProviderFailed state, and leader election function, provided by operator SDK, is unable to process that, so new pods are stuck in wait cycle.
Purpose of the change
Fixes #579
What the code does
Customise the leader.Become() function of operator-sdk and if the pod is in
ProviderFailed
state, delete the pod and configmap so that new pod can come up.How to verify it
Verify in Vmware setup pods are coming up successfully after node reboot.