This is the repository of the conformance test suite for TUF clients. The goal of this repository is to allow client developers to
- test their clients against the TUF specification
- Achieve better practical compatibility with other implementations
- Collaborate on tests with other client developers
The conformance test suite provides a GitHub action that can be used to test a TUF client. There are two required steps:
- Include an executable in the client project that implements the client-under-test CLI protocol.
- Use the
theupdateframework/tuf-conformance
action in your test workflow:jobs: tuf-conformance: runs-on: ubuntu-latest steps: - uses: actions/checkout@v4 # insert possible client compilation/installation steps here - uses: theupdateframework/tuf-conformance@v1 with: entrypoint: path/to/my/test/executable
This repository contains two client-under-test CLI protocol implementations
to enable easy development and testing. The test suite depends on various
python modules which will be installed by the make commands into a virtual environment.
The suite also depends on faketime
tool which needs to be available.
# run test suite against both included clients or just one of them:
make test-all
make test-python-tuf
make test-go-tuf
It's also possible to locally run the test suite with a client-under-test CLI that is locally installed elsewhere:
make dev
./env/bin/pytest tuf_conformance --entrypoint path/to/my/client-under-test/cli
linters can also be invoked with make:
# Run linters
make lint
# Fix issues that can be automatically fixed
make fix
Let's say we want to test that clients will not accept a decreasing targets version. We start with a simple skeleton test that doesn't assert anything but sets up a repository and runs client refresh against it:
def test_targets_version(
client: ClientRunner, server: SimulatorServer
) -> None:
# Initialize our repository: modify targets version and make sure the version is included in snapshot
init_data, repo = server.new_test(client.test_name)
repo.targets.version = 7
repo.update_snapshot() # snapshot v2
# initialize client-under-test, run refresh
client.init_client(init_data)
client.refresh(init_data)
we can now run the test:
./env/bin/pytest tuf_conformance \
-k test_targets_version # run a specific test only
--entrypoint "./clients/go-tuf/go-tuf" # use the included go-tuf client as client-under-test
--repository-dump-dir /tmp/test-repos # dump repository contents
# take a look at the repository targets.json that got debug dumped
cat /tmp/test-repos/test_targets_version/refresh-1/targets.json
cat /tmp/test-repos/test_targets_version/refresh-1/snapshot.json
The metadata looks as expected (targets version is 7) so we can add a modification to the end of the test:
# Make an non-compliant change in repository
repo.targets.version = 6
repo.update_snapshot() # snapshot v3
# refresh client again
client.refresh(init_data)
Running the test again results in a second repository version being dumped (each client refresh leads to a dump):
./env/bin/pytest tuf_conformance \
-k test_targets_version # run a specific test only
--entrypoint "./clients/go-tuf/go-tuf" # use the included go-tuf client as client-under-test
--repository-dump-dir /tmp/test-repos # dump repository contents
# take a look at targets versions in both snapshot versions that got debug dumped
cat /tmp/test-repos/test_targets_version/refresh-1/snapshot.json
cat /tmp/test-repos/test_targets_version/refresh-2/snapshot.json
The repository metadata looks as expected (but not spec-compliant) so we can add some asserts for client behaviour now. The final test looks like this:
def test_targets_version(
client: ClientRunner, server: SimulatorServer
) -> None:
# Initialize our repository: modify targets version and make sure it's included in snapshot
init_data, repo = server.new_test(client.test_name)
repo.targets.version = 7
repo.update_snapshot() # snapshot v2
# initialize client-under-test
client.init_client(init_data)
# Run refresh on client-under-test, expect success
assert client.refresh(init_data) == 0
# Make a non-compliant change in repository
repo.targets.version = 6
repo.update_snapshot() # snapshot v3
# refresh client again, expect failure and refusal to accept snapshot and targets
assert client.refresh(init_data) == 1
assert client.version(Snapshot.type) == 2
assert client.version(Targets.type) == 7
Checklist for making a new tuf-conformance release
- Review changes since last release, decide on version number
- If the client-under-test CLI has changed or client workflows are otherwise likely to break, bump major version
- otherwise if new tests were added, bump minor version
- otherwise bump patch version
- Create and merge PR with README changes if needed (the workflow example contains the major version number)
- tag the new version from a commit in main branch. Example when releasing v1.1.1:
git tag --sign v1.1.1 -m "v1.1.1" git push origin v1.1.1 # now rewrite the major tag: yes rewriting tags is awful, it's also what GitHub recommends... git tag --delete v1 git push --force origin v1 git tag --sign v1 -m "v1.1.1" git push origin v1
- Add release notes to GitHub release in the web UI: this will be shown to action users in the dependabot update. Release notes must mention all breaking changes.
- pytest is used as the test infrastructure, the client-under-test executable is given with
--entrypoint
- A single web server is started. Individual tests can create a simulated TUF repository that will be served in subdirectories of the web server
- Each test sets up a simulated repository, attaches it to the server, runs the client-under-test against that repository. It can then modify the repository state and run the client-under-test again
- the idea is that a test can run the client multiple times while modifying the repository state. After each client
execution the test can verify
- client success/failure
- clients internal metadata state (what it considers currently valid metadata) and
- that the requests client made were the expected ones
- There should be helpers to make these verifications simple in the tests