-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
loading seccomp filter: invalid argument #2865
Comments
Seeing this too with containerd. I can repro easily in an AKS cluster with rc93. rc92 works just fine. Calls look like this:
I truncated the |
Worth noting, |
Bisected to 7a8d716 |
/cc @cyphar |
I believe #2871 fixes this. |
There is a bug with rc93 that may be causing CI instability (opencontainers/runc#2865). Downgrading to rc92 to see if we get better runs. Signed-off-by: Brian Goff <[email protected]>
maybe same issue containerd/containerd#5280 |
we get this error pretty consistently on 2 machines running ubuntu 16.04 downgrading containerd.io to 1.4.3 doesn't fix it as it seems that still uses runc93 and not 92 it was plaguing us for weeks, sometimes locking the system up requiring a reboot was the only fix, and deploys were getting stuck we found in the end, after reading this thread, that to unstick them is to run:
that will strace each one until they exit, and then the system works |
This might be fixed by #2871, which was just merged @oppianmatt @wu0407 @areed @cpuguy83 can you please test the runc tip and report back if the bug is fixed? |
we were only experiencing this on prod, until today that is. I tried removing an image from our staging server and it got stuck. So using it as an opportunity to test. Downloaded runc 92 and replaced the binary on the system and restarted containerd service and docker service. When restarting docker in this state we see huge stackdumps in the log all about mutex locks can see some were quite stuck for a while so much dumping going into syslog that the logs are 5 minutes behind after that though can remove the stuck containers of course won't know for sure until I get another stuck instance, if I do I will report right back, so take no news as good news |
@kolyshkin Verified both by cherry-picking that commit onto rc93 and testing HEAD directly. |
Okay, so it looks like it's fixed then. Please comment if this still isn't fixed after testing on tip. We will do a new release soon. |
Signed-off-by: Sebastian Hasler <[email protected]>
Signed-off-by: Sebastian Hasler <[email protected]>
Signed-off-by: Sebastian Hasler <[email protected]> Co-authored-by: Sebastian Hasler <[email protected]>
This reverts commit 87b0c75.
This reverts commit 87b0c75. Co-authored-by: Sebastian Hasler <[email protected]>
We're seeing machines with several
runc init
processes blocked writing the same message to stderr:This appears to cause a chain reaction on Kubernetes nodes where a lock acquired during
docker start
for the pause container of a Pod blocks PLEG and the node flaps between Ready and NotReady.The text was updated successfully, but these errors were encountered: