Skip to content
This repository has been archived by the owner on Apr 17, 2019. It is now read-only.

[GCE L7LB controller] watch cluster uid config map so updates propogate without restart #1362

Closed
wittrock opened this issue Jul 13, 2016 · 9 comments
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@wittrock
Copy link

wittrock commented Jul 13, 2016

After upgrading two of my clusters in the same project from versions 1.2.5 to version 1.3 of kubernetes, their GCE objects are clobbering each other, and it seems to be because my manually-set cluster-uid got deleted during the upgrade.

➜  ~ kubectl describe configMap --namespace kube-system
Name:       ingress-uid
Namespace:  kube-system
Labels:     <none>
Annotations:    <none>

Data
===
uid:    0 bytes

Even if I edit this configMap and add a cluster-uid, my GCE resources still clobber each other. For instance, all load balancers created by creating ingresses point to the same GCE instance group, called k8s-ig. Previously, they pointed to k8s-ig--my-manually-set-uid.

Based on #680 by @bprashanth, it looks like new objects should be created with whatever is in that configMap first, but that's not happening.

This is fixed if I spin up a new cluster and create a test ingress--those objects are properly namespaced.

Given that I can't fix the issue by editing ingress-uid manually, is there another workaround?

@bprashanth
Copy link

If this is on GCE you can restart your controller by ssh-ing into the master.
Assuming this is on GKE, I can restart your controller (I also just mailed you).
Deleting the pod won't restart the actual container, since we're now running it as a static pod.

@bprashanth bprashanth self-assigned this Jul 13, 2016
@bprashanth
Copy link

This is also a bug because we should allow modification of cluster-uid by watching the configmap, we currently don't do so, which is why the restart is necessary. Thanks for the report.

@bprashanth
Copy link

I think this was fixed, feel free to reopen if that's not the case

@bprashanth bprashanth changed the title [GCE L7LB controller] Upgrade from 1.2.5 to 1.3 didn't propagate cluster-uid [GCE L7LB controller] watch cluster uid config map so updates propogate without restart Jul 15, 2016
@bprashanth
Copy link

No, that doesn't make sense. The issue isn't fixed, a single occurrance of the issue was worked around. reopening.

@bprashanth bprashanth reopened this Jul 15, 2016
@DanielHeckrath
Copy link

@bprashanth I can confirm this issue is still present after updating my cluster from 1.2.4 to 1.3.5 on GKE. Controller also doesn't seem to pick up recently set uid because of the static pod issue you mentioned. How can I restart the controller on GKE so that it's uses the updated id?

@bprashanth
Copy link

@DanielHeckrath sorry for the hassle, please do: https://groups.google.com/forum/#!searchin/kubernetes-users/ingress%7Csort:relevance/kubernetes-users/QpsRZcaVmks/ltJYsdr7BQAJ
A fix has been checked in but it will only be in 1.3.6.

@fejta-bot
Copy link

Issues go stale after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.

Prevent issues from auto-closing with an /lifecycle frozen comment.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or @fejta.
/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Dec 17, 2017
@fejta-bot
Copy link

Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.

If this issue is safe to close now please do so with /close.

Send feedback to sig-testing, kubernetes/test-infra and/or @fejta.
/lifecycle rotten
/remove-lifecycle stale

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Jan 16, 2018
@fejta-bot
Copy link

Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen.
Mark the issue as fresh with /remove-lifecycle rotten.

Send feedback to sig-testing, kubernetes/test-infra and/or fejta.
/close

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

5 participants