Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Failed to check for stale credentials #813

Open
srmars opened this issue Jan 23, 2025 · 4 comments
Open

[BUG] Failed to check for stale credentials #813

srmars opened this issue Jan 23, 2025 · 4 comments
Labels
bug Something isn't working

Comments

@srmars
Copy link

srmars commented Jan 23, 2025

Components and versions
App version 1.7.3

Describe the bug

After upgrading the akv2k8s to 1.73, noticed pod getting restarted with error “ main.go:267] “failed to get credentials” error=“Failed to check for stale credentials …….. dial 10.0.1.80: connect : connection refused”

But after 1 or 2 restart the pod running without any issues.

To Reproduce

Create deployment with environment injection and delete pod or rollout restart

Expected behavior
No restart to the pod

@srmars srmars added the bug Something isn't working label Jan 23, 2025
@srmars
Copy link
Author

srmars commented Mar 19, 2025

Any update on this issue, please share if there is any workaround

@c4tz
Copy link

c4tz commented Mar 19, 2025

I also had this problem the day before yesterday, using the Helm chart. The only difference was that I got a timeout instead of a refused connection.

I tried multiple things:

  • Downgrading the chart to 2.7.1 (App version 1.7.1) and 2.6.0 (App version 1.6.0)
  • curling the URL logged from the same namespace and from another namespace. Always had no problem and got 200 back
  • Setting the loglevel to debug first and trace afterwards, but this didn't change the output at all

Then, at some point it randomly started working again. But this was after I had undone all my changes and about an hour had passed after that.

During my whole investigation, the controller was able to sync secrets to k8s. It was just the env injector having trouble. Because of this, I don't think there was an issue with Azure at the time and (as mentioned before), I also tested the network connectivity.

@srmars
Copy link
Author

srmars commented Apr 3, 2025

It’s one of the blocker for the upgrade, for now we have downgraded the versions, it would great if we get a fix for this

@c4tz
Copy link

c4tz commented Apr 7, 2025

We had this problem again last week, but it turned out to be a faulty node which was missing some network connectivity.

When I replaced it with a new one, the akv2k8s timeouts were gone and everything worked normally. So this might not be a bug in this repo at all.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants