Infrastructure Testing with Testinfra


An often overlooked and admittedly cumbersome topic when developing infrastructure as code (IaC), is the testing of the resources that were created from said infrastructure code.

Even a simple setup like a VM that is spun up in some cloud to serve HTTP requests, already confronts us with some obstacles to overcome. A pragmatic approach for testing such a setup could be to ensure that after deployment of the VM the HTTP port answers with a success HTTP code (2xx).

Although it would look like the deployment was successful, there might still be a hidden misconfiguration that is invisible from the outside, putting the whole deployment at risk.

For example, this could be a set of wrong permissions introducing a security risk. It might also be a faulty configuration value for a service that could cause harm in some special cases, e.g. some debug setting that is only meant for local debugging.

With the inherent complexity of more advanced stacks like Kubernetes, the problem only gets more difficult, since the single component we might want to test could be even harder to reach than the content of a simple VM.

A library that can help us with that is testinfra which is a plugin for the Python testing framework pytest and enables us to run Python tests in and on infrastructure components.

This post will show you how to use testinfra to make your infrastructure deployments more predictable and shorten the feedback cycle for infrastructure code. It will also cover three common test scenarios (Docker images, Kubernetes Helm Deployments, Cloud VMs) and its pitfalls.

As always the full code is available online here github.com/pellepelster/kitchen-sink/infrastructure-testing-testinfra. The folder contains a do file to start all tasks needed to follow the examples in this post.

Bootstrapping

As usual when working with Python code the first step is to bootstrap a Python environment dedicated to our project. First create the file requirements.txt

pytest-testinfra==10.1.0
requests==2.31.0
kubernetes==29.0.0

containing all needed Python dependencies and then init the Python venv with:

python3 -m venv "${DIR}/venv/"
"${DIR}/venv/bin/pip" install -r "${DIR}/requirements.txt"

In the provided example project you can bootstrap everything like this:

cd infrastructure-testing-testinfra
./do bootstrap

That’s it. We now have a Python environment with testinfra in it, and are ready to write our first test.

Docker

As a first step to learn how testinfra is supposed to work, we will use a Docker image as our test target. In a real-world scenario this could be used to ensure that certain tools are installed in a docker image, or that permissions are correctly set.

We will use the following Dockerfile as our test target:

FROM alpine:3.19.1

RUN apk --no-cache add caddy

RUN addgroup \
    --gid "10000" \
    "www" \
&&  adduser \
    --disabled-password \
    --gecos "" \
    --home "/www/public_html" \
    --ingroup www \
    --uid 10000 \
    "www"

RUN mkdir -p /www/public_html
ADD index.html /www/public_html/
RUN chown www:www /www/public_html
ADD Caddyfile /www/Caddyfile
USER www
CMD ["caddy", "run", "--config", "/www/Caddyfile"]

To test the image we need to build it first, so run:

./do docker-build

We can see the author of this Dockerfile spent some time on getting the permissions for the HTTP server right, and also went the extra mile to make sure the www user in the container does not clash with any user ids on the host system. Let’s make sure it stays that way by adding a test for it.

Since testinfra is a pytest plugin everything is executed in the scope of pytest. The special feature of testinfra is that it passes a host object into the test methods that can be used to make assertions on the target; in our case the Docker image.

def test_caddy_runs_as_non_root_user(host):
    caddy = host.process.get(comm="caddy")
    assert caddy.user == 'www'

The example above attempts to the get the process with the name caddy and asserts it is running with the correct user id (wwww). For more available modules on the host object see the modules part of the documentation.

Testinfra comes with a wide range of connection backends, that can be used to talk to different infrastructure types. In our case we want to test a locally running Docker container, using the docker connection backend.

Common to all backends is, that we have to provide a connection string, telling testinfra how to connect to the backend, in the Docker case this looks like this:

py.test --hosts='docker://[user@]container_id_or_name'

Now all that’s left to do to run the test, is to start the Docker container, and provide the name of the started container to testinfra, so it can connect to the container and run the assertions:

docker run --detach --name "${DOCKER_IMAGE_NAME}" "${DOCKER_IMAGE_NAME}:${DOCKER_IMAGE_TAG}"
"${DIR}/venv/bin/py.test" --verbose --capture=tee-sys --hosts="docker://${DOCKER_IMAGE_NAME}" "${DIR}/test/test_docker.py"

Note that because testinfra is a pytest plugin, it can be started directly via the py.test binary from our previously created Python environment. In the example project you can run:

./do docker-test

to start the tests on your local machine.

============================================================================= test session starts =============================================================================
platform linux -- Python 3.10.12, pytest-8.1.1, pluggy-1.4.0 -- /home/pelle/git/kitchen-sink/infrastructure-testing-testinfra/venv/bin/python3
cachedir: .pytest_cache
rootdir: /home/pelle/git/kitchen-sink/infrastructure-testing-testinfra
plugins: testinfra-10.1.0
collected 4 items                                                                                                                                                             

test/test_docker.py::test_caddy_runs_as_non_root_user[docker://testinfra] PASSED                                                                                        [ 25%]
test/test_docker.py::test_caddy_listening_on_8080[docker://testinfra] PASSED                                                                                            [ 50%]
test/test_docker.py::test_www_user[docker://testinfra] PASSED                                                                                                           [ 75%]
test/test_docker.py::test_www_public_html[docker://testinfra] PASSED                                                                                                    [100%]

============================================================================== 4 passed in 1.42s ==============================================================================

As we can see, all tests from test_docker.py were successfully executed, and we can now be sure that our docker container is fit for deployment in the next step.

Kubernetes

Prerequisites

For the next step we want to deploy the docker image to a Kubernetes cluster using a self-written Helm chart that lives in k8s/testinfra. We will be using minikube to spin up a local Kubernetes cluster for testing, please refer to this guide for information on how to install minikube.

Once installed you can run the following command:

./do k8s-start

To start the minikube cluster and import the already built docker image.

Testing

Like in the previous docker example, we need to provide a connection string to pytest to specify how to connect to Kubernetes and which pod we want it to test. The syntax to do so looks like this:

py.test --hosts="kubectl://${POD}?namespace=${NAMESPACE}"

For more details on the connection string and its options see the kubectl backend documentation.

Once the helm chart is deployed via:

helm upgrade --install --namespace ${TEST_NAMESPACE} --wait --create-namespace testinfra ${DIR}/k8s/testinfra

or using the task:

./do k8s-deploy

we can use kubectl to retrieve the information needed for the connection string.

export TEST_POD="$(kubectl get pods --namespace ${TEST_NAMESPACE} --output jsonpath="{.items[0].metadata.name}")"
echo "selected pod '${TEST_POD}' for test execution"
"${DIR}/venv/bin/py.test" --verbose --capture=tee-sys --hosts="kubectl://${TEST_POD}?namespace=${TEST_NAMESPACE}" "${DIR}/test/test_k8s.py"

Once the test is started with the correct kubectl connection string, everything feels and behaves like it does in the Docker example, and we can use all testinfra modules to make assertions on the pods. In our example we assert that the Helm chart does set the DEBUG flag to false.

def test_k8s_environment_variables(host):
    environment = host.environment()
    assert environment['DEBUG'] == "false"

You can run all tests as usual using a do task:

./do k8s-test

Kubernetes Resources

Unfortunately testinfra, lacks a module to assert Kubernetes resources like pod metadata. Let’s for example assume, that we want to assert the pod has the label log-service-name set, which might be used by some logging system that ingests data from Kubernetes.

Although Python has a Kubernetes API like when using kubectl to get pod information we need a pod name and namespace. To avoid funneling this into the tests from the do file, we can use a little trick to determine this information from within the pod.

hostname = host.environment()['HOSTNAME']
namespace = host.run('cat /run/secrets/kubernetes.io/serviceaccount/namespace').stdout

Using the Kubernetes API we can retrieve the information we need and write our assertions as usual.

config.load_kube_config()
v1 = client.CoreV1Api()
pod = v1.read_namespaced_pod(name=hostname, namespace=namespace)
assert pod.metadata.labels['log-service-name'] == 'infrastructure-testing-testinfra'

Virtual Machines (Hetzner)

Prerequisites

For the VM example we will use Terraform to deploy everything we need to the Hetzner Cloud. For this to work you need a Hetzner Cloud account and a valid API token, which needs to be provided to the deployment task via the HCLOUD_TOKEN environment variable.

HCLOUD_TOKEN="...." ./do hetzner-deploy 

Testing

For the final part of this post, we will have a look on how to test a full VM. The focus of this part is less how to write tests (we now know how to do this), but more the wiring needed to seamlessly connect testinfra to a cloud VM.

Testinfra provides the ssh backend for this purpose and expects a connection string like this:

py.test --ssh-config=${ssh_config} --hosts='ssh://${ssh_server}'

When connecting to a VM via SSH, we need to solve two problems:

  • interactive logins using username and password are not a good fit for automated tests
  • since this is a test, the VM will likely be re-provisioned several times, and we don’t want to verify the host identity each time or, even worse, blindly connect to any host without checking its identity

The answer to both problems is: certificates!

We are using the Terraform module at infrastructure-testing-testinfra/hetzner to provision all resources related to the VM. Using Terraform we can also generate an SSH key that we provide when provisioning the server:

resource "tls_private_key" "ssh_client_identity" {
  algorithm = "RSA"
}

resource "hcloud_ssh_key" "ssh_key" {
  name       = "infrastructure-testing-testinfra"
  public_key = tls_private_key.ssh_client_identity.public_key_openssh
}

This allows us to log in to the server using the private part of the key. Writing it to the local filesystem allows us (and testinfra) to use it:

resource "local_file" "ssh_client_identity" {
  content         = tls_private_key.ssh_client_identity.private_key_openssh
  filename        = "${path.module}/ssh/ssh_client_identity"
  file_permission = "0600"
}

Secondly to avoid the host identity-check, we generate another key, which we use as SSH host identity:

resource "tls_private_key" "ssh_host_identity" {
  algorithm = "RSA"
}

We then insert the generated host identity into the OpenSSH server, using the cloud init script:

function sshd_config {
cat <<-EOF
HostKey /etc/ssh/ssh_host_identity_key

LoginGraceTime 2m
PermitRootLogin yes

PasswordAuthentication no
PubkeyAuthentication yes
PermitEmptyPasswords no

ChallengeResponseAuthentication no
UsePAM yes
X11Forwarding no
PrintMotd no

AcceptEnv LANG LC_*
EOF
}

function sshd_setup() {
  local ssh_host_identity_key_base64="${1:-}"
  rm -rf /etc/ssh/ssh_host_*_key.pub

  touch /etc/ssh/ssh_host_identity_key
  chmod 600 /etc/ssh/ssh_host_identity_key
  echo "${ssh_host_identity_key_base64}" | base64 -d > /etc/ssh/ssh_host_identity_key

  sshd_config > /etc/ssh/sshd_config
  service ssh restart
}

Using the public key part of the host identity key, we can then move forward and generate a knownhosts file for our test VM

data "template_file" "known_hosts" {
  template = file("${path.module}/templates/known_hosts.template")

  vars = {
    ip_address            = hcloud_server.server.ipv4_address
    ssh_host_identity_pub = tls_private_key.ssh_host_identity.public_key_openssh
  }
}

Finally, we tie everything together with an ssh_config that we generate with Terraform using the following template:

Host infrastructure-testing-testinfra
    Hostname ${ip_address}
    UserKnownHostsFile ${known_hosts_file}
    IdentityFile ${client_identity_file}
    IdentitiesOnly yes

This defines a config for the host infrastructure-testing-testinfra with the current ${ip_address} of the virtual machine. The ${client_identity_file} we used to provision the host will let us log in without a password and the ${known_hosts_file} containing the public key for the host identity will make sure that we trust the right host. This is because the public key in the file matches the private key used as SSH host identity.

Now we have everything we need to provide testinfra with a full connection string that allows it to automatically connect via SSH using the generated config file.

  (
    cd "${DIR}/hetzner"
    terraform init -upgrade
    terraform apply -auto-approve
  )

  while [[ "$(ssh -F "${DIR}/hetzner/ssh/ssh_config" root@infrastructure-testing-testinfra whoami)" != "root" ]]; do
    echo "waiting for ssh"
    sleep 2
  done

The while-wait-loop ensures that the SSH server is ready to accept connections using the pre-generated SSH keys, because the SSH host key provisioning may take some time.

At this point we are almost ready to run our testinfra tests the way we do in Docker or Kubernetes. Do keep in mind though, that if we want to test anything that is a result of the user-data execution, we need to wait until cloud-init is finished executing the user data part.

We can achieve this by adding a global wait to the pytest module which will wait for the cloud-init result file, and only continue once it exists without any errors.


def wait_until(predicate, timeout, period=2, *args, **kwargs):
    end = time.time() + timeout
    while time.time() < end:
        if predicate(*args, **kwargs):
            return True
        time.sleep(period)
    return False


def host_is_initialized(host):
    result_file = host.file("/run/cloud-init/result.json")
    return result_file.exists


@pytest.fixture(scope="module", autouse=True)
def wait_for_host_is_initialized(host):
    wait_until(lambda: host_is_initialized(host), 60)
    result = json.loads(host.file("/run/cloud-init/result.json").content)
    assert len(result['v1']['errors']) == 0, f"cloud init finished with errors '{result['v1']['errors']}'"


def test_hetzner_caddy_runs_as_non_root_user(host):
    caddy = host.process.get(comm="caddy")
    assert caddy.user == 'caddy'

Closing thoughts

While testinfra-based integration tests are a very good way to assert important details on distinct infrastructure components, in my opinion its main purpose is to speed up the feedback cycle during development, and to make sure we have no regressions on issues that can be tested on component level. It also helps to assert that the interface, which in this case means settings or certain behaviours, is implemented correctly to other components. In the scope of the whole system, still only a full integration-test can make sure that all components work together like they should.

Let's work together!