Dump info about installed packages #17
Workflow file for this run
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Benchmarks | |
on: | |
push: | |
branches: | |
- main | |
pull_request: | |
branches: | |
- '*' | |
workflow_dispatch: # allows to trigger manually | |
schedule: | |
- cron: "0 5 * * 4" # Once a week at 0500 on Thursday | |
# Sets permissions of the GITHUB_TOKEN to allow writing back to `master` | |
permissions: | |
contents: write | |
pages: write | |
id-token: write | |
# Allow only one concurrent deployment, skipping runs queued between the run in-progress and latest queued. | |
# Cancel in-progress deployment jobs so only the latest one succeeds. | |
concurrency: | |
group: "pages" | |
cancel-in-progress: true | |
env: | |
MACHINE_NAME: gcp-n1-standard-8 | |
jobs: | |
run-benchmarks: | |
runs-on: "cirun-benchmark-runner--${{ github.run_id }}" | |
steps: | |
# Install git first; otherwise actions/checkout silently falls back | |
# to github REST API for downloading the repo | |
- name: Install dependencies | |
run: | | |
sudo apt update -y | |
sudo apt install git zlib1g-dev build-essential pkg-config rsync -y | |
- uses: actions/checkout@v4 | |
with: | |
repository: deshaw/versioned-hdf5 | |
path: ./versioned-hdf5/ | |
fetch-depth: 0 # Needed for asv to be able to run benchmarks on old commits | |
- uses: actions/checkout@v4 | |
with: | |
path: ./versioned-hdf5-benchmarks/ | |
- name: Setup python with miniconda | |
uses: conda-incubator/setup-miniconda@v3 | |
with: | |
python-version: 3.11 | |
channels: conda-forge | |
- name: Install versioned-hdf5 | |
working-directory: ./versioned-hdf5 | |
run: | | |
echo "PROJECT_SHA=$(git rev-parse HEAD)" >> $GITHUB_ENV | |
conda install -n test pip hdf5 openmpi h5py ndindex -c conda-forge -y | |
conda run -n test pip install '.[bench]' | |
- name: Dump conda environment | |
run: conda list | |
- name: Print hdf5 configuration | |
run: | | |
set -xe | |
conda run -n test python -c 'import utils; utils.debug_libhdf5_so()' | |
h5cc -showconfig | |
- name: Run benchmarks | |
working-directory: ./versioned-hdf5 | |
run: | | |
# Copy the old asv results back into ./versioned-hdf5 to avoid | |
# running old benchmarks again | |
rsync -r ../versioned-hdf5-benchmarks/.asv ./ | |
# set the machine name; otherwise this is unique to each run | |
echo "Setting machine name to $MACHINE_NAME" | |
conda run -n test asv machine --machine $MACHINE_NAME --yes -v | |
cat ~/.asv-machine.json | |
# Don't return exit code 1 if results are slower | |
conda run -n test asv run --skip-existing --machine $MACHINE_NAME 1.7.0.. || true; | |
# Copy the new benchmark results to the benchmarks repo to commit them | |
rsync -r .asv ../versioned-hdf5-benchmarks/ | |
- name: Add and commit benchmarks | |
uses: EndBug/add-and-commit@v9 | |
with: | |
cwd: ./versioned-hdf5-benchmarks | |
add: .asv/ | |
message: "Update benchmarks for commit ${{ env.PROJECT_SHA }}" | |
- name: Generate html | |
working-directory: ./versioned-hdf5 | |
run: | | |
conda run -n test asv publish -o ./html | |
- name: Setup Pages | |
uses: actions/configure-pages@v5 | |
- name: Upload artifact | |
uses: actions/upload-pages-artifact@v3 | |
with: | |
path: ./versioned-hdf5/html | |
retention-days: 30 | |
- name: Deploy to GitHub Pages | |
if: github.event_name != 'pull_request' | |
id: deployment | |
uses: actions/deploy-pages@v4 |