mirror of
https://code.forgejo.org/actions/cache.git
synced 2024-11-30 02:10:56 -05:00
Move workarounds to a different file
This commit is contained in:
parent
c38f55d874
commit
471fb0c87e
2 changed files with 42 additions and 38 deletions
42
README.md
42
README.md
|
@ -227,44 +227,10 @@ jobs:
|
||||||
## Known practices and workarounds
|
## Known practices and workarounds
|
||||||
Following are some of the known practices/workarounds which community has used to fulfill specific requirements. You may choose to use them if suits your use case. Note these are not necessarily the only or the recommended solution.
|
Following are some of the known practices/workarounds which community has used to fulfill specific requirements. You may choose to use them if suits your use case. Note these are not necessarily the only or the recommended solution.
|
||||||
|
|
||||||
#### Cache segment restore timeout
|
- [Cache segment restore timeout](./workarounds.md#cache-segment-restore-timeout)
|
||||||
A cache gets downloaded in multiple segments of fixed sizes (`1GB` for a `32-bit` runner and `2GB` for a `64-bit` runner). Sometimes, a segment download gets stuck which causes the workflow job to be stuck forever and fail. Version `v3.0.8` of `actions/cache` introduces a segment download timeout. The segment download timeout will allow the segment download to get aborted and hence allow the job to proceed with a cache miss.
|
- [Update a cache](./workarounds.md#update-a-cache)
|
||||||
|
- [Use cache across feature branches](./workarounds.md#use-cache-across-feature-branches)
|
||||||
Default value of this timeout is 60 minutes and can be customized by specifying an [environment variable](https://docs.github.com/en/actions/learn-github-actions/environment-variables) named `SEGMENT_DOWNLOAD_TIMEOUT_MINS` with timeout value in minutes.
|
- [Improving cache restore performance on Windows/Using cross-os caching](./workarounds.md#improving-cache-restore-performance-on-windows-using-cross-os-caching)
|
||||||
|
|
||||||
#### Update a cache
|
|
||||||
A cache today is immutable and cannot be updated. But some use cases require the cache to be saved even though there was a "hit" during restore. To do so, use a `key` which is unique for every run and use `restore-keys` to restore the nearest cache. For example:
|
|
||||||
```yaml
|
|
||||||
- name: update cache on every commit
|
|
||||||
uses: actions/cache@v3
|
|
||||||
with:
|
|
||||||
path: prime-numbers
|
|
||||||
key: primes-${{ runner.os }}-${{ github.run_id }} # Can use time based key as well
|
|
||||||
restore-keys: |
|
|
||||||
primes-${{ runner.os }}
|
|
||||||
```
|
|
||||||
Please note that this will create a new cache on every run and hence will consume the cache [quota](#cache-limits).
|
|
||||||
|
|
||||||
#### Use cache across feature branches
|
|
||||||
Reusing cache across feature branches is not allowed today to provide cache [isolation](https://docs.github.com/en/actions/using-workflows/caching-dependencies-to-speed-up-workflows#restrictions-for-accessing-a-cache). However if both feature branches are from the default branch, a good way to achieve this is to ensure that the default branch has a cache. This cache will then be consumable by both feature branches.
|
|
||||||
|
|
||||||
#### Improving cache restore performance on Windows/Using cross-os caching
|
|
||||||
Currently, cache restore is slow on Windows due to tar being inherently slow and the compression algorithm `gzip` in use. `zstd` is the default algorithm in use on linux and macos. It was disabled on Windows due to issues with bsd tar(libarchive), the tar implementation in use on Windows.
|
|
||||||
|
|
||||||
To improve cache restore performance, we can re-enable `zstd` as the compression algorithm using the following workaround. Add the following step to your workflow before the cache step:
|
|
||||||
|
|
||||||
```yaml
|
|
||||||
- if: ${{ runner.os == 'Windows' }}
|
|
||||||
name: Use GNU tar
|
|
||||||
shell: cmd
|
|
||||||
run: |
|
|
||||||
echo "Adding GNU tar to PATH"
|
|
||||||
echo C:\Program Files\Git\usr\bin>>"%GITHUB_PATH%"
|
|
||||||
```
|
|
||||||
|
|
||||||
The `cache` action will use GNU tar instead of bsd tar on Windows. This should work on all Github Hosted runners as it is. For self-hosted runners, please ensure you have GNU tar and `zstd` installed.
|
|
||||||
|
|
||||||
The above workaround is also needed if you wish to use cross-os caching since difference of compression algorithms will result in different cache versions for the same cache key. So the above workaround will ensure `zstd` is used for caching on all platforms thus resulting in the same cache version for the same cache key.
|
|
||||||
|
|
||||||
## Contributing
|
## Contributing
|
||||||
We would love for you to contribute to `actions/cache`, pull requests are welcome! Please see the [CONTRIBUTING.md](CONTRIBUTING.md) for more information.
|
We would love for you to contribute to `actions/cache`, pull requests are welcome! Please see the [CONTRIBUTING.md](CONTRIBUTING.md) for more information.
|
||||||
|
|
38
workarounds.md
Normal file
38
workarounds.md
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
#### Cache segment restore timeout
|
||||||
|
A cache gets downloaded in multiple segments of fixed sizes (`1GB` for a `32-bit` runner and `2GB` for a `64-bit` runner). Sometimes, a segment download gets stuck which causes the workflow job to be stuck forever and fail. Version `v3.0.8` of `actions/cache` introduces a segment download timeout. The segment download timeout will allow the segment download to get aborted and hence allow the job to proceed with a cache miss.
|
||||||
|
|
||||||
|
Default value of this timeout is 60 minutes and can be customized by specifying an [environment variable](https://docs.github.com/en/actions/learn-github-actions/environment-variables) named `SEGMENT_DOWNLOAD_TIMEOUT_MINS` with timeout value in minutes.
|
||||||
|
|
||||||
|
#### Update a cache
|
||||||
|
A cache today is immutable and cannot be updated. But some use cases require the cache to be saved even though there was a "hit" during restore. To do so, use a `key` which is unique for every run and use `restore-keys` to restore the nearest cache. For example:
|
||||||
|
```yaml
|
||||||
|
- name: update cache on every commit
|
||||||
|
uses: actions/cache@v3
|
||||||
|
with:
|
||||||
|
path: prime-numbers
|
||||||
|
key: primes-${{ runner.os }}-${{ github.run_id }} # Can use time based key as well
|
||||||
|
restore-keys: |
|
||||||
|
primes-${{ runner.os }}
|
||||||
|
```
|
||||||
|
Please note that this will create a new cache on every run and hence will consume the cache [quota](#cache-limits).
|
||||||
|
|
||||||
|
#### Use cache across feature branches
|
||||||
|
Reusing cache across feature branches is not allowed today to provide cache [isolation](https://docs.github.com/en/actions/using-workflows/caching-dependencies-to-speed-up-workflows#restrictions-for-accessing-a-cache). However if both feature branches are from the default branch, a good way to achieve this is to ensure that the default branch has a cache. This cache will then be consumable by both feature branches.
|
||||||
|
|
||||||
|
#### Improving cache restore performance on Windows/Using cross-os caching
|
||||||
|
Currently, cache restore is slow on Windows due to tar being inherently slow and the compression algorithm `gzip` in use. `zstd` is the default algorithm in use on linux and macos. It was disabled on Windows due to issues with bsd tar(libarchive), the tar implementation in use on Windows.
|
||||||
|
|
||||||
|
To improve cache restore performance, we can re-enable `zstd` as the compression algorithm using the following workaround. Add the following step to your workflow before the cache step:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
- if: ${{ runner.os == 'Windows' }}
|
||||||
|
name: Use GNU tar
|
||||||
|
shell: cmd
|
||||||
|
run: |
|
||||||
|
echo "Adding GNU tar to PATH"
|
||||||
|
echo C:\Program Files\Git\usr\bin>>"%GITHUB_PATH%"
|
||||||
|
```
|
||||||
|
|
||||||
|
The `cache` action will use GNU tar instead of bsd tar on Windows. This should work on all Github Hosted runners as it is. For self-hosted runners, please ensure you have GNU tar and `zstd` installed.
|
||||||
|
|
||||||
|
The above workaround is also needed if you wish to use cross-os caching since difference of compression algorithms will result in different cache versions for the same cache key. So the above workaround will ensure `zstd` is used for caching on all platforms thus resulting in the same cache version for the same cache key.
|
Loading…
Reference in a new issue