Skip to content

Latest commit

 

History

History
720 lines (483 loc) · 35.6 KB

CONTRIBUTING.adoc

File metadata and controls

720 lines (483 loc) · 35.6 KB

Contributing to Plutus

Setting up and working with our development tools

This project relies on Nix to provision the complete toolchain needed to build all repository artifacts.

Thanks to Nix we ensure that everyone has consistent versions of the tools that we use.

Please use Nix since problems due to mismatched versions of tools are particularly annoying to fix!

If you really cannot use Nix and still want to contribute to Plutus then skip to the relevant section.

Installing and setting up Nix

This project uses Flakes, so you must have Nix version >= 2.4 installed.

Either upgrade or install Nix following the instructions on its website.

On Linux this command should suffice: sh <(curl -L https://nixos.org/nix/install) --daemon

How to get a shell environment with tools

To enter the development shell run the following command in the root directory:

nix develop

If you receive this message:

error: experimental Nix feature 'nix-command' is disabled; use '--extra-experimental-features nix-command' to override

Then try this:

nix develop --extra-experimental-features 'nix-command flakes'

You may now choose to edit your /etc/nix/nix.conf (global) or ~/.config/nix/nix.conf (you must be a trusted user to do this) by appending the following line:

extra-experimental-features = nix-command flakes

If you do so you will be able to enter the development shell by running nix develop.

If you are a VSCode user, you may want to start your development session by running:

nix develop --command code

If this is your first time running nix develop, you may be asked the following:

do you want to allow configuration setting 'allow-import-from-derivation' to be set to 'true' (y/N)? y
do you want to permanently mark this value as trusted (y/N)? y

Please type y to this and all subsequent prompts to make life easy for yourself.

Important
It is particularly important to accept the extra-substituters and extra-trusted-public-keys settings, which will grant access to our Nix binary cache.

At this point and if this is your first time running nix develop then Nix will download, build and install a copious amout of dependencies.

It is not uncommon for this process to take a couple of hours and to write tens of GBs to the /nix/store.

Rest assured that this only happens the very first time you run nix develop.

However you should pay attention to the output of nix develop: if you notice that you are building GHC, it is likely that your caches have not been configured correctly.

Accepting the configuration settings as outlined above should be sufficient to avoid this, however if your caches are still broken, you may try to append the following lines to /etc/nix/nix.conf and/or ~/.config/nix/nix.conf:

substituters = https://cache.zw3rk.com https://cache.iog.io https://cache.nixos.org/
trusted-public-keys = loony-tools:pr9m4BkM/5/eSTZlkQyRt57Jz7OMBxNSUiMC4FkcNfk= hydra.iohk.io:f/Ea+s+dFdN+3Y/G+FDgSq+a5NEWhJGzdjvKNGv0/EQ= cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY=
extra-experimental-features = nix-command flakes

On NixOS, set the following NixOS options:

nix = {
  binaryCaches          = [ "https://cache.zw3rk.com" "https://cache.iog.io" ];
  binaryCachePublicKeys = [ "loony-tools:pr9m4BkM/5/eSTZlkQyRt57Jz7OMBxNSUiMC4FkcNfk=" "hydra.iohk.io:f/Ea+s+dFdN+3Y/G+FDgSq+a5NEWhJGzdjvKNGv0/EQ=" ];
};

You will need to restart the nix-daemon on non-NixOS for the changes to take effect, or just restart your machine.

Note on pre-commit hooks

If you are committing code outside nix develop, you will get this error:

pre-commit not found. Did you forget to activate your virtualenv?

In that case, you may either pass the flag --no-verify to git commit, or pip install pre-commit.

The pre-commit checks will be run by CI anyway, so any formatting errors will be caught when submitting your PR.

What to do once inside the nix develop shell

Your prompt will change to [plutus-shell] and you will be presented with a menu of available commands.

Please read that menu carefully.

How to build the code during development

The nix develop environment has the correct GHC with all the external Haskell dependencies of the project.

From here you can build the project packages directly with cabal.

Note
You may need to run cabal update so that cabal knows about the index state we have pinned.

Run cabal build plutus-core from the root to build the Plutus Core library.

See the cabal project file for a list of other packages that you can build with cabal.

How to setup haskell-language-server

The nix develop environment has a haskell-language-server binary for the right version of GHC.

Important
This binary is called haskell-language-server, rather than haskell-language-server-wrapper, which is what some of the editor integrations expect.

We don’t have a hie.yaml, the implicit cradle support in HLS seems to work fine these days.

How to build the project’s artifacts with Nix

Haskell components are provisioned by Nix via Haskell.nix

In general you can run nix build .#SYSTEM.plutus.library.plutus-project.hsPkgs.PACKAGE.components.COMPONENT

For example nix build .#x86_64-linux.plutus.library.plutus-project.hsPkgs.plutus-core.components.library

For full documentation about the Nix code see the Nix README.

There you will find how to build all other artifacts, which are mostly related to documentation.

How to develop without Nix

You can build some of the Haskell packages without Nix, but this is not recommended and we don’t guarantee that these prerequisites are sufficient.

If you use Nix, these tools are provided for you via nix develop, and you do not need to install them yourself.

  • If you want to build our Agda code, then install Agda and the standard library.

  • If you want to build our Haskell packages with cabal, then install it.

Warning
  • We rely on forked or new versions of some system libraries.

  • You may get different versions of packages.

    • This shouldn’t happen, but we can’t guarantee it.

  • We are not currently enabling the Nix integration for these tools, so they will use your system GHC and libraries, rather than that ones that will be used by Nix.

    • We sometimes patch the GHC that we use in Nix, so this can at least potentially cause problems or cause you to be missing bug workarounds.

How to add a new Haskell package

You need to do a few things when adding a new package, in the following order:

  • Add the cabal file for the new package.

  • Add the package to cabal.project.

  • Check that you can build the package with nix as well (see How to build with Nix) or wait for CI to check this for you.

How to update our Haskell dependencies

Our Haskell packages come from two package repositories: - Hackage - CHaP (which is essentially another Hackage)

The "index state" of each repository is pinned to a particular time in cabal.project. This tells Cabal to treat the repository "as if" it was the specified time, ensuring reproducibility. If you want to use a package version from repository X which was added after the pinned index state time, you need to bump the index state for X. This is not a big deal, since all it does is change what packages cabal considers to be available when doing solving, but it will change what package versions cabal picks for the plan, and so will likely result in significant recompilation, and potentially some breakage. That typically just means that we need to fix the breakage (and add a lower-bound on the problematic package), or add an upper-bound on the problematic package.

Note that cabal itself keeps track of what index states it knows about, so when you bump the pinned index state you may need call cabal update in order for cabal to be happy.

The Nix code which builds our packages also cares about the index state. This is represented by some pinned inputs in our flake (see here for more details) You can update these by running: - nix flake lock --update-input hackage-nix for Hackage - nix flake lock --update-input CHaP for CHaP

Use of `source-repository-package`s

We can use Cabal’s source-repository-package mechanism to pull in un-released package versions. However, we should try and avoid this. In particular, we should not release our packages while we depend on a source-repository-package.

If we are stuck in a situation where we need a long-running fork of a package, we should release it to CHaP instead (see the CHaP README for more).

If you do add a source-repository-package, you need to update the sha256 mapping in nix/cells/plutus/library/make-plutus-project.nix. For the moment you have to do this by hand, using the following command to get the sha: nix-prefetch-git --quiet <repo-url> <rev> | jq .sha256, or by just getting it wrong and trying to build it, in which case Nix will give you the right value.

How to update our pinned Nix dependencies

We pin versions of some git repositories that are used by Nix, for example nixpkgs.

Specifically, you will probably want to say nix flake lock --update-input <input-name>.

Do not use nix flake update, as that will update all the inputs, which we typically don’t want to do.

How to build the code with profiling

TODO: Currently not available, coming soon

If you launch nix develop .#profiled-plutus you will get a shell where all the dependencies have been built with profiling.

Warning

The shell with profiling dependencies is not currently cached, so this will result in you rebuilding all of our dependencies with profiling on your machine. This will take a long time.

Once you have a shell with profiling libraries for our dependencies, add profiling: true to cabal.project.local, which will tell cabal that you want profiling (in particular, that will cause it to build our libraries with profiling).

Alternatively, you can pass the --enable-profiling option to cabal on an ad-hoc basis, but adding the option to cabal.project.local will make it apply to everything, which is probably what you want when you’re doing profiling work.

At this point you need to configure which cost centres you want GHC to insert.

The GHC user guide explains this very well.

A typical way of doing this is to add -fprof-auto to either the ghc-options in the .cabal file for the project, or in an OPTIONS_GHC pragma in the module you care about.

Warning

Do not set the -prof option yourself! This will enable profiling libraries unconditionally, which interferes with what cabal wants. Setting profiling: true already sorts this out properly.

Then you can use the RTS -p option to dump a profile e.g. cabal run plc …​ — +RTS -p.

Warning

When building plutus-core, you might get a compilation error similar to the following:

ghc: ^^ Could not load 'recursionzmschemeszm5zi2zm8KxPjFseRtMJfccAAVODSC_DataziFunctorziFoldableziTH_zdfMakeBaseFunctorNamezuzdcmakeBaseFunctor_closure', dependency unresolved. See top entry above.

ByteCodeLink.lookupCE
During interactive linking, GHCi couldn't find the following symbol:
  recursionzmschemeszm5zi2zm8KxPjFseRtMJfccAAVODSC_DataziFunctorziFoldableziTH_zdfMakeBaseFunctorNamezuzdcmakeBaseFunctor_closure

To resolve it, simply add the following lines in your cabal.project.local:

package plutus-core
  ghc-options: -fexternal-interpreter

This issue is tracked upstream at https://gitlab.haskell.org/ghc/ghc/-/issues/18320

There are various tools for visualizing the resulting profile, e.g. https://hackage.haskell.org/package/ghc-prof-flamegraph.

GHC versions

We have a set of supported GHC major versions, which are built in CI and which we commit to keeping working.

We also have a primary development version. This is the version we use in the default dev shell, and the one we use day-to-day. The primary development version should be whichever version the Cardano node is using (or migrating towards using). Supported versions older than the primary development version are deprecated, and we can drop them as soon as the Cardano node has moved to a later GHC version.

At the moment, our supported GHC versions are: - 8.10 (deprecated) - 9.2 (primary) - 9.6

Plugin support

Making plutus-tx-plugin work on multiple GHC versions can be painful, because it uses a lot of GHC internals. Since the plugin is not in the dependency closure of the Cardano node, we don’t need to support all the GHC versions, which is helpful. We mostly commit to supporting our primary development version, but we may support other versions as well if it is easy.

The plugin and its dependents should be marked as unbuildable with cabal conditionals on any non-supported major GHC versions. This makes cabal behave reasonably well (it will ignore them and not try and fail to build them).

We currently support the plugin on: - 9.2 - 9.6

Per-version tooling

We have a dev shell for each supported GHC version which you can use if you need to fix issues on that specific version. You can access them like this: nix develop .#plutus-shell-810.

Note that HLS in particular won’t work in the non-primary dev shell as it will still be built with the primary GHC version, but other tooling should work fine.

Per-version testing

As far as we know, given an exact GHC compiler version the GHC compiler will generate the same (reproducible) GHC-Core, from any Haskell code, running under any platform (linux/osx/windows/etc).

However, when comparing the GHC-Core output coming from different GHC versions, this might not still hold. Because the PlutusTx language heavily relies on GHC-Core, any (slight) GHC-Core differences will change the resulting plutus core (and its archetypes, plutus ir and typed plutus core). Since our (test) code must be reproducible with multiple GHC versions (see Plugin Support) at the same time in the repository, we keep multiple test output directories (containg golden plutus files), one for each GHC version that the plugin supports.

To test and/or regenerate test golden test output, one has to run for each GHC version supported by the plugin:

nix develop ".#plutus-shell-$GHC_MAJOR_VERSION" --command cabal test plutus-tx-plugin --test-options=--accept

Working conventions

Style guide

Please follow our Haskell style guide, which documents most of our conventions for working on Haskell code.

Code is communication

We are a relatively large team working on sometimes quite abstruse problems. As such, it’s important that future people who work on the project know how things work, and just as importantly, why. These future people may even be yourself - we forget things very quickly!

When writing, try to put yourself in the position of someone coming to this code for the first time. What do they need to do to understand it and do their job? Write it down!

Code review is a good lens for this: if you have to explain something to a reviewer, then it is probably not clear in the code and should have a note.

This applies both to the code itself (structure, naming, etc.) and also to comments. How to write useful comments is a large topic which we don’t attempt to cover here, but Antirez is good. If in doubt: write more!

"Notes"

One special kind of comment is worth drawing attention to. We adopt a convention (stolen from GHC) of writing fairly substantial notes in our code with a particular structure. These correspond to what Antirez calls "design comments", with some conventions about cross-referencing them.

The structure is:

  • The Note should be in a multiline comment (i.e. {- -})

  • The first line of the Note should be Note [Name of note]

  • Refer to a Note from where it is relevant with a comment saying See Note [Name of note]

For example:

{- Note [How to write a note]
A note should look a bit like this.

Go wild, write lots of stuff!

Here's a small diagram:
A ----> B >> C

And of course, you should see Note [Another note].
-}

Notes are a great place to put substantial discussion that you need to refer to from multiple places. For example, if you used an encoding trick to fit more data into an output format, you could write a Note describing the trick (and justifying its usage!), and then refer to it from the encoder and the decoder.

Code formatting

We use stylish-haskell for Haskell code formatting, and cabal-fmt for cabal files. They are run automatically as pre-commit hooks, but CI will run them again and expect that to be a no-op, so if you somehow don’t apply them your PR will not go green.

To run stylish-haskell or cabal-fmt manually over your tree, type fix-stylish-haskell or fix-cabal-fmt respectively. They are provided by the nix develop environment.

Code linting

There are two .hlint.yaml files, one in ./ and the other in .github/. The one in ./ is the default hint file used by editors, and the one in .github/ is used by CI. Think of the former as suggested hints, and the latter as enforced hints.

Compiler warnings

The CI builds Haskell code with -Werror, so will fail if there are any compiler warnings. So fix your own warnings!

If the warnings are stupid, we can turn them off, e.g. sometimes it makes sense to add -Wno-orphans to a file where we know it’s safe.

Commit messages

Please make informative commit messages! It makes it much easier to work out why things are the way they are when you’re debugging things later.

A commit message is communication, so as usual, put yourself in the position of the reader: what does a reviewer, or someone reading the commit message later need to do their job? Write it down! It is even better to include this information in the code itself, but sometimes it doesn’t belong there (e.g. ticket info).

Also, include any relevant meta-information, such as ticket numbers. If a commit completely addresses a ticket, you can put that in the headline if you want, but it’s fine to just put it in the body.

There is plenty to say on this topic, but broadly the guidelines in this post are good.

Commit signing

Set it up if you can, it’s relatively easy to do.

Making and reviewing changes

Changelogs

We write changelogs for our versioned and published packages, which currently include:

  • plutus-core

  • plutus-ledger-api

  • plutus-tx

  • plutus-tx-plugin

  • prettyprinter-configurable

To do this we use scriv, a changelog management tool.

When to write a changelog entry

We have no clear policy here, it is up to the judgement of the contributor. Not all PRs need changelog entries. However our CI will check for changelog entries unless the 'No Changelog Required' label is applied.

The broad heuristic is to put yourself in the position of the consumer of the piece of software in question and ask if you would want to know about this change. If the answer is yes, then write a quick changelog entry.

How to write a changelog entry

The basic idea is that you write a changelog "fragment" in the changelog.d directory. When we do a release, these will be collected into the main CHANGELOG.md. Usually we don’t edit CHANGELOG.md directly.

You can make a changelog fragment using scriv create in the package directory, but you can also just create the fragment directly with an editor. A fragment is a markdown file beginning with a header giving the category of change. Currently these are:

  • Removed

  • Added

  • Changed

  • Deprecated

  • Fixed

  • Security

We can change these categories if we want.

A fragment can contain multiple sections if there are multiple changes. For example:

# Changed

Updated foo to take a bar.

# Deprecated

Deprecated baz.

Opening a pull request

A pull request is a change to the codebase, but it is also an artifact which goes through a change acceptance process. There are a bunch of things which we can do to make this process smooth which may have nothing to do with the code itself.

The key bottleneck in getting a PR merged is code review. Code review is great (see below), but it can slow you down if you don’t take the time to make it easy.

The amount of time it’s worth spending doing this is probably much more than you think.

What branch to target

PRs should target master unless there is a very good reason not to. The only PRs to release branches should be backport PRs which should consist only of cherry-picks of commits from master (and any fixups that are needed). For more details, see Plutus Release Process.

What changes to include, and pull request sizes

When developing a non-trivial new feature, usually the best way to get the code reviewed is to break the implementation down to a chain of small diffs, each representing a meaningful, logical and reviewable step. Unfortunately GitHub doesn’t have good support for this. You basically have three options:

  • Open the first PR against master, the second PR against the first PR’s branch, and so on. Merging a stack of PRs created this way into master can be non-trivial.

  • Wait until one PR is merged before opening the next PR.

  • Use a single PR for the whole feature that contains multiple small commits. The problem is that Github doesn’t support approving, rejecting or merging individual commits in a PR. You can look at each individual commit, but it’s not necessarily useful or even appropriate - many PRs have quite messy commits, and commits are sometimes overwritten via force push.

The first two options are often referred to as trunk-based development, while the third "long-lived feature branches". There is no single best option for all cases, although in general we encourage adopting trunk-based development styles. Long-lived feature branches with too many commits are harmful because

  1. they are difficult to review - the PR can be quite large, and it is hard to review it incrementally;

  2. it can be difficult to resolve merge conflicts;

  3. they make it more likely that other people need to depend on your unmerged changes.

It is fine to have partially implemented features or not well-tested features in master. You can simply not turn them on until they are ready, or guard them with conditional flags.

But this is not a hard rule and should be determined on a case-by-case basis. Sometimes for a small or medium-sized piece of work, you may not want to break it into multiple PRs, and wait till each PR is merged before creating the next one. You’d rather put all your code out quickly in a single PR for review. And that’s fine. Or maybe it’s a piece of performance improvement work, and you don’t know whether or not it actually improves the performance, until you finish implementing and testing the whole thing.

Whichever option you choose, please keep each of your PR to a single topic. Do not mix business logic with such things as reformatting and refactoring in a single PR.

Pull request descriptions

A pull request is communication, so as usual, put yourself in the position of the reader: what does your audience (the reviewer) need to know to do their job? This information is easy for you to access, but hard for them to figure out, so write it down!

However, better to put information in the code, commit messages, or changelog if possible: these persist but PR descriptions do not. It’s okay to repeat information from such places, or simply to point to it. For one-commit PRs, Github will automatically populate the PR description with the commit message, so if you’ve written a good commit message you’re done! Sometimes there is "change-related" information that doesn’t belong in a commit message but is useful ("Kris I think this will fix the issue you had yesterday").

Misc PR tips

  • Review the diff of your own PR at the last minute before hitting "create". It’s amazing how many obvious things you spot here, and it stops the reviewer having to point them all out.

  • It’s fine to make WIP PRs if you just want to show your code to someone else or have the CI check it. Use the Github "draft" feature for this.

Rebasing and force pushing

Force pushing to master (or any other protected branch) is never allowed. There is no exception to this rule.

Rebasing and force pushing to other branches you own is fine, even when you have an open PR on the branch. Indeed, if you need to update your branch with changes from master, rebasing is typically better than merging.

Some projects do not allow force pushing to any remote branch. This is not a popular policy and we do not adopt it, because

  • This means you must only ever use the "merge commit" merge method (or occasionally, fast forward merge, which GitHub doesn’t support).

  • This means you aren’t even allowed to clean up commits in your own PR, and must eventually merge everything into master. It discourages people from pushing commits frequently when developing. We should instead encourage cleaning up commits in PRs, at least before merging.

  • The argument that this will cause massive pain for those who merge other people’s PR branch into their branch is questionable. This should be rare to begin with, if we adopt trunk-based development in general, instead of long-lived feature branches. And even if you do need to depend on other people’s unmerged work, you can instead rebase your branch on theirs, and if their branch changes, just rebase again.

Rebasing and force pushing can be used to your advantage, for example:

  • Add low-effort or WIP commits to fix review comments, and then squash them away before merging the PR.

  • If you have already had a PR review, don’t rebase away the old commits until the PR is ready to merge, so that the reviewer only has to look at the "new" commits.

  • Rewrite the commits to make the story clearer where possible.

It is advisable to always prefer git push --force-with-lease instead of git push --force to ensure that no work gets accidentally deleted.

Code review

All pull-requests should be approved by at least one other person. We don’t enforce this, though: a PR fixing a typo is fine to self-merge, beyond that use your judgement.

As an author, code review is an opportunity for you to get feedback from clear eyes. As a reviewer, code review is an opportunity for you to help your colleagues and learn about what they are doing. Make the best use of it you can!

For the author

  • Pick the right reviewer(s). If you don’t know who to pick, ask!

  • Respect your reviewers' time. Their time is as valuable as yours, and it’s typically more efficient for you to spend time explaining or clarifying something in advance than for them to puzzle it out or pose a question.

  • If someone had to ask about your code, it wasn’t clear enough so change it or add a comment.

Read this blog post for more good tips: https://mtlynch.io/code-review-love/

For the reviewer

  • Respond to review requests as quickly as you can. If you can’t review it all, say what you can and come back to it. Waiting for review is often a blocker for other people, so prioritize it.

  • If you don’t understand something, ask. You are as clever as any person who will read this in the future, if it confuses you it’s confusing.

  • Do spend the time to understand the code. This will help you make more useful comments, help you review future changes more easily, and help you if you ever need to work on it yourself.

  • More reviewing is usually helpful. If you think a PR is interesting, you can review it even if nobody asked you to, you will probably have things to contribute and you’ll learn something.

Benchmarking PRs

Sometimes it is useful to benchmark a PR, and we have some automation for this. To trigger it, make a comment on the PR with this form: /benchmark <benchmark-component>, where <benchmark-component> is as you would provide it to cabal. For example, if you would run cabal bench plutus-benchmark:validation locally, then write /benchmark plutus-benchmark:validation in the comment.

This will trigger a benchmarking job on a stable machine. The job will:

  1. Run the specified benchmark on the base of the PR branch.

  2. Run the specified benchmark on the tip of the PR branch.

  3. Compare the two runs.

  4. Post the comparison as a comment to the PR.

Merging PRs

Merge method and commit history

All 3 Github merge methods (merge commit, squash and merge, and rebase and merge) are allowed. Use whichever you deem appropriate. As said before, sometimes people use a single PR with multiple commits for their work; other times they create multiple small PRs. The best merge method is different for different cases.

That being said, there are not many cases where "rebase and merge" is appropriate, and you might as well rebase it yourself. And if you use this method, your PR must have a clean commit history: every commit should have a meaningful message, and should be buildable. You don’t want to have commits like "fix a typo", "this may work" or "wip, done for the day" in master with a linear history. And if some of these commits are non-buildable, it can create problems for "git bisect".

This is slightly less of a problem when you use the "merge commit" method. While these interim commits would still be unpleasant, at least the merge commits and the non-linear history clearly indicate where they come from.

The best thing to do, of course, is to not have those interim commits. If you think merging multiple commits makes more sense, clean up the history. If you don’t, squash. The option chosen can vary from PR to PR.

Beware divergence of master and PR branch

Merging a PR can break master, if the PR branch has diverged from master, even if CI on the PR is green. This happens because the PRs conflict in a way that isn’t obvious to git, e.g. one adds a usage of a function and the other removes that function. The problems with a broken master include inconveniencing other developers, and causing problems for "git bisect". There are ways to guarantee master never breaks, such as GitHub’s merge queue.

We don’t use the merge queue because

  • A broken master has historically been quite infrequent.

  • The merge queue increases the time it takes to merge a PR, which causes productivity loss if you are waiting to create the next PR after merging the current one (which happens often).

However, if your PR branch has diverged too much from master, it is recommended that you rebase or merge master into the PR branch before merging. And whenever you notice a broken master, please fix it ASAP.

External contributors

The Plutus team welcomes contributions from external contributors. However, it can be difficult for the Plutus team to quickly review contributions from people where we don’t have an existing relationship. For that reason, we ask you to follow these additional guidelines (the rest of the document also applies!), which will make it easier for us to review your work, and therefore make the contributing process smoother for you.

All changes must have an issue

Make sure that any change you make has a corresponding GitHub issue. The issue should describe the problem and describe your proposed solutiion. Before you start working on implementing it, you must get a comment from the Plutus team that the solution seems sensible. This functions as a light "design review" before you get too stuck into doing a PR.

Reviewing the issue makes things easier for the Plutus team (it’s easier to read an issue than a PR); and less frustrating for the contributor (it’s nicer to get design feeback before you have done lots of work on the implementation). We can also offer advice on implementation, or let you know that we’re already planning to fix the issue (or that there is a good reason not to!).

Keep changes well-scoped

Try to keep your PR focussed on one change. This is a pratice we try to follow generally, but especially for external contributions where reviews tend to be more laborious, it’s good to keep things focussed. If your PR contains a dozen drive-by refactorings, it’s unlikely to be merged as such!

Supporting systems

Continuous integration

We have a few sources of CI checks at the moment:

  • Hydra

  • ReadTheDocs

  • Github Actions

  • Buildkite

The CI will report statuses on your PRs with links to the logs in case of failure. Pull requests cannot be merged without at least the Hydra CI check being green.

Note
This isn’t strictly true: repository admins (notably Michael) can force-merge PRs without the checks being green If you really need this, ask.

Hydra

Hydra is the "standard" CI builder for Nix-based projects. It builds everything in the project, including all the tests, documentation, etc.

Hydra builds jobs based on the hydraJobs flake output.

Hydra can be a bit flaky, unfortunately: - If evaluation fails saying "out of memory" or "unexpected EOF reading line", then this is likely a transient failure. These will be automatically retried, but if you’re in a hurry Michael has permissions to force a new evaluation. - If a build fails spuriously, this is a problem: please report it to whoever is responsible for that build and we should try and iron it out. Nondeterministic failures are very annoying. Michael also has permissions to restart failed builds.

ReadTheDocs

The documentation site is built on ReadTheDocs. It will build a preview for each PR which is linked from the PR status. It’s useful to take a look if you’re changing any of the documentation.

Enter the development shell using nix develop. If you get a segfault, run GC_DONT_GC=1 nix develop instead.

Then you can run serve-docs to host a local instance at http://0.0.0.0:8002 (Haddock is at http://0.0.0.0:8002/haddock).

Github Actions

These perform some of the same checks as Hydra, but Github Actions is often more available, so they return faster and act as a "smoke check".

Project roles and responsibilities

  • The regular contributors to the Haskell code, all of whom can review and merge PRs are:

  • @michaelpj

  • @effectfully

  • @kwxm

  • @bezirg

  • @thealmarty

  • @zliu41

  • The maintainer of the documentation is @joseph-fajen.

  • The maintainer of the Agda code is @jmchapman, @effectfully can help with small issues.

  • If you have a technical dispute that you need help resolving, you can ask @michaelpj.

  • For problems with the developer environment setup, builds, or CI, you can ask @zeme-iohk, @Pacman99, or @michaelpj.

  • The regular contributors take turns releasing our software, but if you have a specific problem ask @zliu41 or @michaelpj.