Running tests

Fetching the test data.

The input data that is needed to run the tests will need to be downloaded from the openpipelines Amazon AWS s3 bucket first. To do so, the download/sync_test_resource component can be used, which will download the data to the correct location (resources_test) by default.

./bin/viash run ./src/download/sync_test_resources/config.vsh.yaml -p docker -- --input s3://openpipelines-data

Or, if you do not want to use Docker and have aws-cli tools installed natively:

./bin/viash run ./src/download/sync_test_resources/config.vsh.yaml -p native -- --input s3://openpipelines-data

Running component unittests

To build and run tests for individual component that you are working on, use viash test with the config.vsh.yaml of the component you would like to test. For example:

./bin/viash test ./src/convert/from_10xh5_to_h5mu/config.vsh.yaml

Keep in mind that when no platform is passed to viash test, it will use the first platform that is specified in the config, which is docker for most of the components in openpipelines. Use -p native for example if you do not want to use docker.

It is also possible to execute the tests for all components in each namespace using ./bin/viash_test (note the underscore instead of a space).


Integration tests

Individual integration tests can be run by using the scripts for a pipeline, located next to the in the workflows folder.


Running all integration tests is also possible using a helper script that can be found at workflows/test/ Using this script requires a working R installation with tidyverse installed. However, as pipelines are implemented by combining individual components