-
Notifications
You must be signed in to change notification settings - Fork 10
KF1.7 Applications not Syncing #1762
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
With I can't remember if it was degraded before / when we first upgraded it, I didnt think it was but I may be mistaken. Maybe this case happened where during some cluster operation it kinda dies. |
Monitoring System InvestigationPort-forwarding
I'm trying to use the cURL described here but I don't seem to have access? provides
If I look at the date of According to Pat;
Trying to edit what I got from |
More talking; found this in the docs about reducing replicas. We might not need to specify
After which though we may need to go and delete the extra shards (so in our case this The other thing is we're unsure "if we want to set it to 1 or 0 (is replica an extra copy of the original?"
and with what Pat said before, I feel like we can just go down to zero? The shard data that was provided from Steps:
Argocd is now all green, there are no more unassigned_shards either and attempting the previous query of
|
Regarding This is because kubeflow removed it from their stack and will not exist in future releases. We can instead get a standalone install like what has been done for cloud main. I will close this ticket when we have a team discussion just to confirm that it is ok / create a new issue for spark Just had to delete the application from argocd as well and its gone |
Kserve stuck in Progressing" -- solved
Original ticket were the issue was found : #1752
Getting imagepullbackoff, probably because gatekeeper is blocking image: 'kserve/kserve-controller:v0.10.0'... Having said that the error does say, "Back-off restarting failed container"
Getting: {"level":"error","ts":1688561752.6363223,"logger":"entrypoint","msg":"unable to get ingress config.","error":"unable to parse ingress config json: invalid character '}' looking for beginning of object key string","stacktrace":"runtime.main\n\t/usr/local/go/src/runtime/proc.go:250"} from the manager container
We are currently using
kserve/kserve-controller:v0.8.0
in prod, hence I don't think the gatekeeper is the reason it doesn't workLooking at the pods log we see
According to an article on stackoverflow, this would be most likely malformed json file.
The other application is
Monitoring-system stuck at Degraded Health
Monitoring-system one became degraded. It's for the ElasticSearch reousrce in monitoring-system and it's because the Health/Status is yellow, have we encountered this before / how did we fix it? (I can't seem to find any issues in our repos that mention it aside from our new one)
The text was updated successfully, but these errors were encountered: