toolkit icon indicating copy to clipboard operation
toolkit copied to clipboard

Cache Stuck

Open ScottPierce opened this issue 4 years ago • 19 comments

Describe the bug All my actions are running, but are refusing to cache with the same error:

Post job cleanup.
Unable to reserve cache with key Linux-modules-0c1e633988f0463573ae79cec4e4e741153037a6a270a901c8bf040b3d674a4f, another job may be creating this cache.

There must be some sort of race condition, and the cache is stuck, because they all have the same precache log:

Run actions/cache@v2
  with:
    path: **/node_modules
    key: Linux-modules-0c1e633988f0463573ae79cec4e4e741153037a6a270a901c8bf040b3d674a4f
Cache not found for input keys: Linux-modules-0c1e633988f0463573ae79cec4e4e741153037a6a270a901c8bf040b3d674a4f

I imagine that if I changed my yarn lock file, which the cache is based on, this issue would resolve itself.

Expected behavior That the cache always works, and when it gets stuck like this, it would automatically fix itself.

Additional context

  • My github runner is running inside a docker container. Not deferring to a docker container, but are running inside a container where the runner is installed. I don't think this impacts things as my cache normally works.
  • I don't know how to reproduce this. Normally the cache works. There just is clearly some sort of race condition where it can get stuck.

ScottPierce avatar Dec 02 '20 04:12 ScottPierce

I've run into the same issue. Each time it starts with Cache not found for input keys: windows-master-N002 and ends with Unable to reserve cache with key windows-master-N002, another job may be creating this cache., so no cache ends up actually getting created.

Pentarctagon avatar Dec 04 '20 05:12 Pentarctagon

I'm getting this too, on a GitHub-hosted runner. Logs are here: https://github.com/rcowsill/NodeGoat/runs/1504051837 (see the Post Use cache (docker layers) section)

In my case it's happening when the cache with the matching key already exists. Normally that would fail with the message "Cache already exists." instead of "Unable to reserve cache with key [...]"

I'm using satackey/[email protected], which imports @actions/[email protected]. In case it's important, that action is running up to four cache uploads in parallel, but all caches have unique keys within a single build. I was only running that one build using those cache keys at the time.

EDIT: I tested with parallel uploads switched off and got the same result. I'm wondering if the serverside response has been changed for the "cache exists" case to avoid retries which will never succeed.

rcowsill avatar Dec 05 '20 18:12 rcowsill

I've seen this happen again. This might be the initial error:

Error: Cache upload failed because file read failed with EBADF: bad file descriptor, read

ScottPierce avatar Dec 11 '20 05:12 ScottPierce

I've also noticed this is extremely reproducible for some reason. This failure happens a lot for my docker runners when things are in parallel, especially now that I have 8 runners now.

ScottPierce avatar Dec 11 '20 06:12 ScottPierce

This is now blocking the build system. I'm seeing this build error constantly:

Warning: uploadChunk (start: 100663296, end: 134217727) failed: Cache service responded with 503
/_work/_actions/actions/cache/v2/dist/save/index.js:3305
                        throw new Error(`Cache upload failed because file read failed with ${error.message}`);
                        ^

Error: Cache upload failed because file read failed with EBADF: bad file descriptor, read
    at ReadStream.<anonymous> (/_work/_actions/actions/cache/v2/dist/save/index.js:3305:31)
    at ReadStream.emit (events.js:210:5)
    at internal/fs/streams.js:167:12
    at FSReqCallback.wrapper [as oncomplete] (fs.js:470:5)

I'm going to remove caching to fix it.

ScottPierce avatar Dec 12 '20 19:12 ScottPierce

Same issue, macOS runner. Works on GitHub runner, but not local. Our key is key: ${{ runner.os }}-yarn-${{ hashFiles('web/yarn.lock') }}.

Unable to reserve cache with key macOS-yarn-5967645e7e3b673f1a9f0792c61e21863dc66df3a592de7561b991f867052c7b, another job may be creating this cache

klausbadelt avatar Jan 17 '21 02:01 klausbadelt

I think we encountered this problem by manually cancelling a run that had begun caching with a certain key before it completed the caching step. It seemed to never released the key, and subsequent runs that try to use the same key fail to reserve it:

Run actions/cache@v2 Cache not found for input keys: <key>

...

Post Run actions/cache@v2 Post job cleanup. Unable to reserve cache with key <key>, another job may be creating this cache.

Our only workaround was to change the key.

christianchownsan avatar Jan 26 '21 15:01 christianchownsan

I'm also experiencing this issue with my custom github action, it seems that it really insists on the cache key being "unique"? otherwise it keeps failing with Error: Failed to save cache: ReserveCacheError: Unable to reserve cache with key flatpak-builder-0f51bddf4f15c39e39b55b6c92c8249d99d44205, another job may be creating this cache. Except there are 0 other jobs running at the same time. The issue can be triggered by re-running a workflow job that just created the cache.

bilelmoussaoui avatar Apr 01 '21 15:04 bilelmoussaoui

Same here, or some Warning: uploadChunk (start: 67108864, end: 100663295) failed: Cache service responded with 503

RSickenberg avatar Apr 08 '21 16:04 RSickenberg

Post job cleanup.
/usr/bin/tar --posix --use-compress-program zstd -T0 -cf cache.tzst -P -C /home/runner/work/***/*** --files-from manifest.txt
Warning: uploadChunk (start: 33554432, end: 67108863) failed: Cache service responded with 503
/home/runner/work/_actions/actions/cache/v2/dist/save/index.js:4043
                        throw new Error(`Cache upload failed because file read failed with ${error.message}`);
                        ^

Error: Cache upload failed because file read failed with EBADF: bad file descriptor, read
    at ReadStream.<anonymous> (/home/runner/work/_actions/actions/cache/v2/dist/save/index.js:4043:31)
    at ReadStream.emit (events.js:210:5)
    at internal/fs/streams.js:167:12
    at FSReqCallback.wrapper [as oncomplete] (fs.js:470:5)

After reruning the workflow error is gone.

rafis avatar May 09 '21 05:05 rafis

I had a relative path in the list of files to cache. But it seems that relative paths are not supported. Removing it fixed the error.

Mubelotix avatar Nov 05 '21 12:11 Mubelotix

I think I am hitting this on two workflows, as rust-cache uses this underneath. Is there any way to view/delete the caches if they are in a 'funky' state?

Manually overriding cache keys, and doing a commit seems like a bit of whack-a-mole approach. Happy to manually delete the cache from a UI (or API) feels like a bit of a cleaner workaround until the root cause is fixed.

btrepp avatar Dec 11 '21 02:12 btrepp

I also experienced Warning: uploadChunk (start: 67108864, end: 100663295) failed: Cache service responded with 503

MichaelTamm avatar Dec 17 '21 16:12 MichaelTamm

Faced the same on macOS matrix.

Error: uploadChunk (start: 0, end: 15553027) failed: Cache service responded with 503

Though everything worked on the very next run.

mabdullahadeel avatar Jan 20 '22 23:01 mabdullahadeel

I ran into this issue and think I figured out the most common cause. It seems like there's two causes for this:

  1. Transient uploading issues (bad file descriptor, chunk, 503s, etc)
  2. trying to save the cache for a cache key that was already saved (and restored)

If you're running into this problem and don't have any errors uploading, I'm guessing the root cause is the last one.

Basically I compared what this cache library was doing with how the official actions/cache action actually uses it and here's the key part: https://github.com/actions/cache/blob/611465405cc89ab98caca2c42504d9289fb5f22e/src/save.ts#L39-L54

The official action does not try to save the cache if there was previously an exact key match on a cache hit. If you just naively do a restore cache + save cache (like I tried), you'll run into this error every time there's a cache hit (meaning you're trying to save a cache key which is already cached). Ideally saveCache was an atomic operation, but since it's not we have to replicate that behaviour.

So unfortunately the solution is to replicate all the logic within https://github.com/actions/cache/blob/611465405cc89ab98caca2c42504d9289fb5f22e/src/save.ts.

Here's a utility function to wrap a cacheable function (like calling exec.exec('some command')) which works for me:

async function withCache(cacheable, paths, baseKey, hashPattern) {
  const keyPrefix = `${process.env.RUNNER_OS}-${baseKey}-`;
  const hash = await glob.hashFiles(hashPattern);
  const primaryKey = `${keyPrefix}${hash}`;
  const restoreKeys = [keyPrefix];

  const cacheKey = await cache.restoreCache(paths, primaryKey, restoreKeys);

  if (!cacheKey) {
    core.info(`Cache not found for keys: ${[primaryKey, ...restoreKeys].join(", ")}`);
  }

  core.info(`Cache restored from key: ${cacheKey}`);

  await cacheable();

  if (isExactCacheKeyMatch(primaryKey, cacheKey)) {
    core.info(`Cache hit occurred on the primary key ${primaryKey}, not saving cache.`);
    return;
  }

  await cache.saveCache(paths, primaryKey);
}

await withCache(async () => {
  await exec.exec('npm install')
}, ['node_modules], 'npm', '**/package.json');

You might want to customize the arguments and how the keys are built (maybe accept a list of restore keys too).

swalkinshaw avatar Jan 23 '22 02:01 swalkinshaw

A build fails with Unable to reserve cache with key 3.1.10-x64-master, another job may be creating this cache. First time I have seen this about 1-2 weeks ago on different repo. Now I see it at least once in every repo per week.

filipworksdev avatar Jun 14 '22 20:06 filipworksdev

the issue is still reproducible, there's not a workaround at the moment

Francesco146 avatar Oct 08 '23 12:10 Francesco146

Had a similar issue with ruff:

ruff failed
  Cause: Failed to create cache file '/home/runner/work/aoc_2023/aoc_2023/.ruff_cache/0.1.12/1323824952410372998'
  Cause: No such file or directory (os error 2)

Emptying the workflow file, committing, then pasting it back in fixed the issue.

natemcintosh avatar Jan 13 '24 15:01 natemcintosh

ruff failed Cause: Failed to create cache file '/home/runner/work/aoc_2023/aoc_2023/.ruff_cache/0.1.12/1323824952410372998' Cause: No such file or directory (os error 2)

Alfmac22 avatar Jan 18 '24 18:01 Alfmac22