-
Notifications
You must be signed in to change notification settings - Fork 38
gcsfuse gcs-fuse-csi-driver increase default memory capacity #372
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi @raed114, The connection refused error tends to be related to unavailability from the
|
Do you happen to have any updates on this? I have a similar issue; the problem occurs when copying any file over 1Gib in size. |
Hi @Ahmed-Alhameedawi, please refer to our Public Documentation where you can adjust sidecar resources according to your workload needs. |
Hi @hime I'm also facing similar issues. We are using Argo Workflows that will mount GCS buckets via the CSI driver. We encountered this issue when submitting many workflow requests in parallel. Here are some more info from our side:
|
@roelschr-ft could you provide GKE version that you are using? |
cluster: 1.30.8-gke.1261000 |
What you would like to accomplish:
Increasing the default amount of the GCSFuse sidecar's memory from 100Mi to 200Mi instead of using a mutation webhook.
How this should work:
Implemented either automatically and/or will have a native option to increase the memory of the sidecar container.*
Explanation of the problem:
The gcs-fuse-csi-driver sidecar container seems to be repeatedly restarting and shows as OOMKilled all the while the node doesn't seem to run out of resources. or doesn't have any memory pressure issues. While the restart of this container eventually happened, it took 3 hours in this case (instead of a few minutes), which made the workload unresponsive, it became stuck until the eventual eviction of the pod
Several error messages after a deep investigation into it:
MountVolume.SetUp failed for volume "<VOLUME_NAME>" : kubernetes.io/csi: mounter.SetUpAt failed to determine if the node service has VOLUME_MOUNT_GROUP capability: rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial unix /var/lib/kubelet/plugins/gcsfuse.csi.storage.gke.io/csi.sock: connect: connection refused"
(combined from similar events): Memory cgroup out of memory: Killed process 2960932 (gcs-fuse-csi-dr) total-vm:1340248kB, anon-rss:100796kB, file-rss:26844kB, shmem-rss:0kB, UID:0 pgtables:372kB oom_score_adj:-997
Since this sidecar is created automatically and is managed by default, problems like these can cause serious downtime for the workload.
The text was updated successfully, but these errors were encountered: