GitHub Actions I/O performance
![GitHub Actions I/O performance](/_astro/disk-perf.BcXuUAIQ_Z1tb1DP.webp)
This page offers a detailed comparison of disk I/O performance for various providers of GitHub Actions self-hosted runners. The aim is to assist in identifying the best GitHub Actions self-hosted runner provider for your projects, based on your specific needs.
Providers included in the benchmark
Third-party alternatives:
- Buildjet (Hetzner)
- Blacksmith (Hetzner / Secured Servers LLC)
- Namespace (Own infrastructure, US/Europe)
- Ubicloud (Hetzner)
- Cirrus (Hetzner)
Note: Warpbuild and Depot are third-party providers that resell AWS instances, but they forbid benchmarking their platform 🤷.
Disk performance
While CPU speed is the most important factor for most workflows, disk performance can be the limiting factor in specific scenarios that require a high number of IOPS. That’s why it is important to be able to pick runner types that match your needs.
The table below compares the disk performance of various providers. Sequential read and write performance is compared, as well as random read and write performance. Sequential performance is more important when dealing with large files, while random performance is more important for small, many files. Benchmark is run as per this Google Cloud article ↗ and measures the performance on the disk where the GITHUB_WORKSPACE
resides. I performed 3 runs and picked the best result for each provider, although I didn’t notice much variance across runs.
For RunsOn, multiple instance types are compared, to highlight the performance difference between:
-
instance types with locally attached NVMe volumes vs EBS-only instances. When an instance has local storage in addition to EBS, the results are for the local storage only (EBS results stay the same).
-
storage bandwidth scaling of locally attached NVMe volumes with instance size, since on AWS larger instances are allocated a higher percentage of the storage bandwidth ↗ of the underlying host.
Note that RunsOn automatically mounts and formats the local instance storage volumes for you. If multiple volumes are available, they are automatically mounted in a RAID-0 configuration, to maximize disk space and performance.
The default sorting of the table is by Random writes
, because this is where the network-attached EBS volumes show their main limitation. You can click on the column header to sort by a different metric.
Results
Provider | Type | Configuration | Seq Writes (MiB/s) | Rand Writes (MiB/s) | Seq Reads (MiB/s) | Rand Reads (MiB/s) | Infrastructure |
---|---|---|---|---|---|---|---|
RunsOn | c6id.24xlarge | 5312 GiB (4 * 1328 GiB NVMe SSD) | 3435 | 1495 | 7332 | 1564 | AWS |
RunsOn | c5ad.24xlarge | 3540 GiB (2 * 1770 GiB NVMe SSD) | 2966 | 1465 | 6353 | 1285 | AWS |
RunsOn | c7gd.12xlarge | 2656 GiB (2 * 1328 GiB NVMe SSD) | 1699 | 1402 | 3664 | 1648 | AWS |
RunsOn | c5ad.12xlarge | 1676 GiB (2 * 838 GiB NVMe SSD) | 1479 | 1315 | 3177 | 1313 | AWS |
RunsOn | c7gd.metal (equivalent to c7gd.16xlarge) | 3540 GiB (2 * 1770 GiB NVMe SSD) | 2285 | 1286 | 4899 | 1483 | AWS |
Cirrus | ghcr.io/cirruslabs/ubuntu-runner-arm64:22.04-md | 47 GiB SSD | 2680 | 1162 | 34500 | 1140 | Hetzner |
RunsOn | c7gd.8xlarge | 1770 GiB NVMe SSD | 1135 | 1038 | 2448 | 1688 | AWS |
Namespace | nscloud-ubuntu-22.04-arm64-2x8 | 98 GiB SSD | 2501 | 638 | 4960 | 831 | Hetzner |
RunsOn | c7gd.4xlarge | 884 GiB NVMe SSD | 574 | 508 | 1205 | 1051 | AWS |
Ubicloud | ubicloud-standard-2-arm | 86 GiB SSD | 1079 | 316 | 1149 | 443 | Hetzner |
RunsOn | c5ad.4xlarge | 560 GiB (2 * 280 GiB NVMe SSD) | 242 | 219 | 604 | 511 | AWS |
Buildjet | buildjet-2vcpu-ubuntu-2204-arm | 118 GiB SSD | 1819 | 165 | 8746 | 320 | Hetzner |
RunsOn | c6id.xlarge | 220 GiB NVMe SSD | 146 | 131 | 303 | 262 | AWS |
RunsOn | c7gd.xlarge | 220 GiB NVMe SSD | 147 | 131 | 303 | 262 | AWS |
Blacksmith | blacksmith-2vcpu-ubuntu-2204-arm | 130 GiB SSD | 143 | 93 | 44 | 1271 | Hetzner |
RunsOn | c7gd.large | 110 GiB NVMe SSD | 75 | 66 | 153 | 131 | AWS |
RunsOn | c7gd.medium | 55 GiB NVMe SSD | 41 | 33 | 80 | 66 | AWS |
GitHub | 2cpu-arm64 | 75 GiB, Network volume | 199 | 17 | 203 | 28 | Azure |
RunsOn | c7g.medium | 40 GiB, EBS only, 400MiB provisioned throughput | 404 | 16 | 406 | 16 | AWS |
RunsOn | c7g.medium | 80 GiB, EBS only, 750MiB provisioned throughput | 730 | 16 | 755 | 16 | AWS |
Observations
Network-attached volumes
- Sequential read/write of network-attached volumes is actually pretty good, and you get what you pay for in terms of volume throughput provisioning. Interestingly it can be higher than some providers with local SSDs (gp3 volumes can provision up to 1000MiB/s of throughput).
- At the same time, random read/write performance of network-attached volumes is… not great, reaching only up to 20MiB/s. This is the case for default runner types at GitHub, RunsOn, and other providers at AWS.
The good news is that poor random read/write performance has a surprisingly low impact in the context of most CI workflows, where the sequential performance is often more relevant, as you’ll be pulling / writing large cache files, tarballs, artefacts, or docker images from the network, and then executing mostly CPU-intensive tasks.
However some workflows will require very large amounts of space and/or much improved random performance, and that’s where it’s important to be able to also pick a runner with local SSDs when you need it.
Local SSDs
- If you want the fastest sequential AND random write performance, AWS with locally-attached NVMe volumes is your friend. Although you will need to choose larger instance types due to the linear scaling of the storage bandwidth with the instance size. You also get access to terabytes of local storage if you need it.
- Performance of AWS locally-attached NVMe volumes is identical whether you select an x64 or arm64 instance.
- Cirrus has really high sequential read performance, and good performance overall. Namespace as well. I suppose (reach out if I’m wrong!) that contrary to AWS they are not throttling bandwidth so performance could suffer from noisy neighbours on the same host. Probably same story with Buildjet and Ubicloud, which have lower performance as well. Blacksmith is probably suffering from this issue, since the disk performance is much lower than the other providers at Hetzner (except for random reads, for some reason).
- Among Hetzner providers, max disk space is currently limited to 130GiB (Blacksmith). AWS allows up to 64 TiB for EBS volumes, and offers instance types that can go up to 305 TiB GB of local storage.
RunsOn allows to select instances with local SSDs (of variable sizes and performance), and automatically mount them.