Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Chromium builds broken from v0.5.0 with filesystem store #1565

Open
chrisstaite-menlo opened this issue Jan 23, 2025 · 1 comment
Open

Chromium builds broken from v0.5.0 with filesystem store #1565

chrisstaite-menlo opened this issue Jan 23, 2025 · 1 comment
Assignees
Labels
bug Something isn't working

Comments

@chrisstaite-menlo
Copy link
Collaborator

When running a CAS with the following configuration, Chromium/Reclient builds stall workers. It appears that they are sat waiting for something but they have low CPU and network usage.

Running v0.5.3 worker and scheduler with a v0.4.0 CAS with the following configuration works just fine. So something weird is happening.

{
  "stores": {
    "CAS_MAIN_STORE": {
      "fast_slow": {
        "fast": {
          // Don't in-memory cache over 100kb, and cache up to 100mb
          "size_partitioning": {
            "size": 100000,
            "lower_store": {
              "memory": {
                "eviction_policy": {
                  "max_bytes": 100000000
                }
              }
            },
            "upper_store": {
              "noop": {}
            }
          }
        },
        "slow": {
          "filesystem": {
            "content_path": "/root/.cache/turbo-cache/content_path-cas",
            "temp_path": "/root/.cache/turbo-cache/tmp_path-cas",
            "eviction_policy": {
              // This uses around 8Gb of RAM for metadata.  This is the actual size
              // of the data ignoring inode sizing, so actual space used is higher.
              // 110gb
              "max_bytes": 110000000000,
              // 2gb
              "evict_bytes": 2000000000
            }
          }
        }
      }
    },
    "AC_MAIN_STORE": {
      "fast_slow": {
        "fast": {
          // Don't in-memory cache over 4kb, and cache up to 10mb
          "size_partitioning": {
            "size": 4000,
            "lower_store": {
              "memory": {
                "eviction_policy": {
                  "max_bytes": 10000000
                }
              }
            },
            "upper_store": {
              "noop": {}
            }
          }
        },
        "slow": {
          "filesystem": {
            "content_path": "/root/.cache/turbo-cache/content_path-cas_ac",
            "temp_path": "/root/.cache/turbo-cache/tmp_path-cas_ac",
            "eviction_policy": {
              // 10gb.
              "max_bytes": 10000000000
            }
          }
        }
      }
    }
  },
  "servers": [{
    "listener": {
      "http": {
        "socket_address": "0.0.0.0:50052",
        "advanced_http": {
          "experimental_http2_max_pending_accept_reset_streams": 300
        }
      }
    },
    "services": {
      "cas": {
        "main": {
          "cas_store": "CAS_MAIN_STORE"
        }
      },
      "ac": {
        "main": {
          "ac_store": "AC_MAIN_STORE"
        }
      },
      "capabilities": {},
      "bytestream": {
        "cas_stores": {
          "main": "CAS_MAIN_STORE"
        },
        // According to https://github.com/grpc/grpc.github.io/issues/371 16KiB - 64KiB is optimal.
        "max_bytes_per_stream": 65536
      }
    }
  }]
}

@aaronmondal
Copy link
Member

I have a hunch that this is caused by the same root cause that makes (Bazel) builds like this fail if they're started, canceled on the client side and then restarted:

[
    genrule(
        name = "dummy_test_sh_{}".format(n),
        outs = ["dummy_test_{}.sh".format(n)],
        cmd = "sleep 70; echo \"sleep .1;   echo $$(printf '=%.0s' {1..100})\" > \"$@\"",
    )
    for n in range(100)
]

What happens is that the loop that updates subscriptions has some bug (likely here

async fn changed(&mut self) -> Result<AwaitedAction, Error> {
).

@aaronmondal aaronmondal self-assigned this Jan 23, 2025
@aaronmondal aaronmondal added the bug Something isn't working label Jan 23, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants